mirror of
				https://git.kernel.org/pub/scm/linux/kernel/git/chenhuacai/linux-loongson
				synced 2025-10-31 18:53:24 +00:00 
			
		
		
		
	 12e36b2f41
			
		
	
	
		12e36b2f41
		
	
	
	
	
		
			
			* git://git.kernel.org/pub/scm/linux/kernel/git/sfrench/cifs-2.6: (27 commits) [CIFS] Missing flags2 for DFS [CIFS] Workaround incomplete byte length returned by some [CIFS] cifs Kconfig: don't select CONNECTOR [CIFS] Level 1 QPathInfo needed for proper OS2 support [CIFS] fix typo in previous patch [CIFS] Fix old DOS time conversion to handle timezone [CIFS] Do not need to adjust for Jan/Feb for leap day [CIFS] Fix leaps year calculation for years after 2100 [CIFS] readdir (ffirst) enablement of accurate timestamps from legacy servers [CIFS] Fix compiler warning with previous patch [CIFS] Fix typo [CIFS] Allow for 15 minute TZs (e.g. Nepal) and be more explicit about [CIFS] Fix readdir of large directories for backlevel servers [CIFS] Allow LANMAN21 support even in both POSIX non-POSIX path [CIFS] Make use of newer QFSInfo dependent on capability bit instead of [CIFS] Do not send newer QFSInfo to legacy servers which can not support it [CIFS] Fix typo in name of new cifs_show_stats [CIFS] Rename server time zone field [CIFS] Handle legacy servers which return undefined time zone [CIFS] CIFS support for /proc/<pid>/mountstats part 1 ... Manual conflict resolution in fs/cifs/connect.c
		
			
				
	
	
		
			1031 lines
		
	
	
		
			27 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			1031 lines
		
	
	
		
			27 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
| /*
 | |
|  *   fs/cifs/cifsfs.c
 | |
|  *
 | |
|  *   Copyright (C) International Business Machines  Corp., 2002,2004
 | |
|  *   Author(s): Steve French (sfrench@us.ibm.com)
 | |
|  *
 | |
|  *   Common Internet FileSystem (CIFS) client
 | |
|  *
 | |
|  *   This library is free software; you can redistribute it and/or modify
 | |
|  *   it under the terms of the GNU Lesser General Public License as published
 | |
|  *   by the Free Software Foundation; either version 2.1 of the License, or
 | |
|  *   (at your option) any later version.
 | |
|  *
 | |
|  *   This library is distributed in the hope that it will be useful,
 | |
|  *   but WITHOUT ANY WARRANTY; without even the implied warranty of
 | |
|  *   MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See
 | |
|  *   the GNU Lesser General Public License for more details.
 | |
|  *
 | |
|  *   You should have received a copy of the GNU Lesser General Public License
 | |
|  *   along with this library; if not, write to the Free Software
 | |
|  *   Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA
 | |
|  */
 | |
| 
 | |
| /* Note that BB means BUGBUG (ie something to fix eventually) */
 | |
| 
 | |
| #include <linux/module.h>
 | |
| #include <linux/fs.h>
 | |
| #include <linux/mount.h>
 | |
| #include <linux/slab.h>
 | |
| #include <linux/init.h>
 | |
| #include <linux/list.h>
 | |
| #include <linux/seq_file.h>
 | |
| #include <linux/vfs.h>
 | |
| #include <linux/mempool.h>
 | |
| #include <linux/delay.h>
 | |
| #include <linux/kthread.h>
 | |
| #include "cifsfs.h"
 | |
| #include "cifspdu.h"
 | |
| #define DECLARE_GLOBALS_HERE
 | |
| #include "cifsglob.h"
 | |
| #include "cifsproto.h"
 | |
| #include "cifs_debug.h"
 | |
| #include "cifs_fs_sb.h"
 | |
| #include <linux/mm.h>
 | |
| #define CIFS_MAGIC_NUMBER 0xFF534D42	/* the first four bytes of SMB PDUs */
 | |
| 
 | |
| #ifdef CONFIG_CIFS_QUOTA
 | |
| static struct quotactl_ops cifs_quotactl_ops;
 | |
| #endif
 | |
| 
 | |
| int cifsFYI = 0;
 | |
| int cifsERROR = 1;
 | |
| int traceSMB = 0;
 | |
| unsigned int oplockEnabled = 1;
 | |
| unsigned int experimEnabled = 0;
 | |
| unsigned int linuxExtEnabled = 1;
 | |
| unsigned int lookupCacheEnabled = 1;
 | |
| unsigned int multiuser_mount = 0;
 | |
| unsigned int extended_security = CIFSSEC_DEF;
 | |
| /* unsigned int ntlmv2_support = 0; */
 | |
| unsigned int sign_CIFS_PDUs = 1;
 | |
| extern struct task_struct * oplockThread; /* remove sparse warning */
 | |
| struct task_struct * oplockThread = NULL;
 | |
| extern struct task_struct * dnotifyThread; /* remove sparse warning */
 | |
| struct task_struct * dnotifyThread = NULL;
 | |
| static struct super_operations cifs_super_ops; 
 | |
| unsigned int CIFSMaxBufSize = CIFS_MAX_MSGSIZE;
 | |
| module_param(CIFSMaxBufSize, int, 0);
 | |
| MODULE_PARM_DESC(CIFSMaxBufSize,"Network buffer size (not including header). Default: 16384 Range: 8192 to 130048");
 | |
| unsigned int cifs_min_rcv = CIFS_MIN_RCV_POOL;
 | |
| module_param(cifs_min_rcv, int, 0);
 | |
| MODULE_PARM_DESC(cifs_min_rcv,"Network buffers in pool. Default: 4 Range: 1 to 64");
 | |
| unsigned int cifs_min_small = 30;
 | |
| module_param(cifs_min_small, int, 0);
 | |
| MODULE_PARM_DESC(cifs_min_small,"Small network buffers in pool. Default: 30 Range: 2 to 256");
 | |
| unsigned int cifs_max_pending = CIFS_MAX_REQ;
 | |
| module_param(cifs_max_pending, int, 0);
 | |
| MODULE_PARM_DESC(cifs_max_pending,"Simultaneous requests to server. Default: 50 Range: 2 to 256");
 | |
| 
 | |
| extern mempool_t *cifs_sm_req_poolp;
 | |
| extern mempool_t *cifs_req_poolp;
 | |
| extern mempool_t *cifs_mid_poolp;
 | |
| 
 | |
| extern kmem_cache_t *cifs_oplock_cachep;
 | |
| 
 | |
| static int
 | |
| cifs_read_super(struct super_block *sb, void *data,
 | |
| 		const char *devname, int silent)
 | |
| {
 | |
| 	struct inode *inode;
 | |
| 	struct cifs_sb_info *cifs_sb;
 | |
| 	int rc = 0;
 | |
| 
 | |
| 	sb->s_flags |= MS_NODIRATIME; /* and probably even noatime */
 | |
| 	sb->s_fs_info = kzalloc(sizeof(struct cifs_sb_info),GFP_KERNEL);
 | |
| 	cifs_sb = CIFS_SB(sb);
 | |
| 	if(cifs_sb == NULL)
 | |
| 		return -ENOMEM;
 | |
| 
 | |
| 	rc = cifs_mount(sb, cifs_sb, data, devname);
 | |
| 
 | |
| 	if (rc) {
 | |
| 		if (!silent)
 | |
| 			cERROR(1,
 | |
| 			       ("cifs_mount failed w/return code = %d", rc));
 | |
| 		goto out_mount_failed;
 | |
| 	}
 | |
| 
 | |
| 	sb->s_magic = CIFS_MAGIC_NUMBER;
 | |
| 	sb->s_op = &cifs_super_ops;
 | |
| /*	if(cifs_sb->tcon->ses->server->maxBuf > MAX_CIFS_HDR_SIZE + 512)
 | |
| 	    sb->s_blocksize = cifs_sb->tcon->ses->server->maxBuf - MAX_CIFS_HDR_SIZE; */
 | |
| #ifdef CONFIG_CIFS_QUOTA
 | |
| 	sb->s_qcop = &cifs_quotactl_ops;
 | |
| #endif
 | |
| 	sb->s_blocksize = CIFS_MAX_MSGSIZE;
 | |
| 	sb->s_blocksize_bits = 14;	/* default 2**14 = CIFS_MAX_MSGSIZE */
 | |
| 	inode = iget(sb, ROOT_I);
 | |
| 
 | |
| 	if (!inode) {
 | |
| 		rc = -ENOMEM;
 | |
| 		goto out_no_root;
 | |
| 	}
 | |
| 
 | |
| 	sb->s_root = d_alloc_root(inode);
 | |
| 
 | |
| 	if (!sb->s_root) {
 | |
| 		rc = -ENOMEM;
 | |
| 		goto out_no_root;
 | |
| 	}
 | |
| 
 | |
| 	return 0;
 | |
| 
 | |
| out_no_root:
 | |
| 	cERROR(1, ("cifs_read_super: get root inode failed"));
 | |
| 	if (inode)
 | |
| 		iput(inode);
 | |
| 
 | |
| out_mount_failed:
 | |
| 	if(cifs_sb) {
 | |
| 		if(cifs_sb->local_nls)
 | |
| 			unload_nls(cifs_sb->local_nls);	
 | |
| 		kfree(cifs_sb);
 | |
| 	}
 | |
| 	return rc;
 | |
| }
 | |
| 
 | |
| static void
 | |
| cifs_put_super(struct super_block *sb)
 | |
| {
 | |
| 	int rc = 0;
 | |
| 	struct cifs_sb_info *cifs_sb;
 | |
| 
 | |
| 	cFYI(1, ("In cifs_put_super"));
 | |
| 	cifs_sb = CIFS_SB(sb);
 | |
| 	if(cifs_sb == NULL) {
 | |
| 		cFYI(1,("Empty cifs superblock info passed to unmount"));
 | |
| 		return;
 | |
| 	}
 | |
| 	rc = cifs_umount(sb, cifs_sb); 
 | |
| 	if (rc) {
 | |
| 		cERROR(1, ("cifs_umount failed with return code %d", rc));
 | |
| 	}
 | |
| 	unload_nls(cifs_sb->local_nls);
 | |
| 	kfree(cifs_sb);
 | |
| 	return;
 | |
| }
 | |
| 
 | |
| static int
 | |
| cifs_statfs(struct dentry *dentry, struct kstatfs *buf)
 | |
| {
 | |
| 	struct super_block *sb = dentry->d_sb;
 | |
| 	int xid; 
 | |
| 	int rc = -EOPNOTSUPP;
 | |
| 	struct cifs_sb_info *cifs_sb;
 | |
| 	struct cifsTconInfo *pTcon;
 | |
| 
 | |
| 	xid = GetXid();
 | |
| 
 | |
| 	cifs_sb = CIFS_SB(sb);
 | |
| 	pTcon = cifs_sb->tcon;
 | |
| 
 | |
| 	buf->f_type = CIFS_MAGIC_NUMBER;
 | |
| 
 | |
| 	/* instead could get the real value via SMB_QUERY_FS_ATTRIBUTE_INFO */
 | |
| 	buf->f_namelen = PATH_MAX; /* PATH_MAX may be too long - it would 
 | |
| 				      presumably be total path, but note
 | |
| 				      that some servers (includinng Samba 3)
 | |
| 				      have a shorter maximum path */
 | |
| 	buf->f_files = 0;	/* undefined */
 | |
| 	buf->f_ffree = 0;	/* unlimited */
 | |
| 
 | |
| /* BB we could add a second check for a QFS Unix capability bit */
 | |
| /* BB FIXME check CIFS_POSIX_EXTENSIONS Unix cap first FIXME BB */
 | |
|     if ((pTcon->ses->capabilities & CAP_UNIX) && (CIFS_POSIX_EXTENSIONS &
 | |
| 			le64_to_cpu(pTcon->fsUnixInfo.Capability)))
 | |
| 	    rc = CIFSSMBQFSPosixInfo(xid, pTcon, buf);
 | |
| 
 | |
|     /* Only need to call the old QFSInfo if failed
 | |
|     on newer one */
 | |
|     if(rc)
 | |
| 	if(pTcon->ses->capabilities & CAP_NT_SMBS)
 | |
| 		rc = CIFSSMBQFSInfo(xid, pTcon, buf); /* not supported by OS2 */
 | |
| 
 | |
| 	/* Some old Windows servers also do not support level 103, retry with
 | |
| 	   older level one if old server failed the previous call or we
 | |
| 	   bypassed it because we detected that this was an older LANMAN sess */
 | |
| 	if(rc)
 | |
| 		rc = SMBOldQFSInfo(xid, pTcon, buf);
 | |
| 	/*     
 | |
| 	   int f_type;
 | |
| 	   __fsid_t f_fsid;
 | |
| 	   int f_namelen;  */
 | |
| 	/* BB get from info in tcon struct at mount time call to QFSAttrInfo */
 | |
| 	FreeXid(xid);
 | |
| 	return 0;		/* always return success? what if volume is no
 | |
| 				   longer available? */
 | |
| }
 | |
| 
 | |
| static int cifs_permission(struct inode * inode, int mask, struct nameidata *nd)
 | |
| {
 | |
| 	struct cifs_sb_info *cifs_sb;
 | |
| 
 | |
| 	cifs_sb = CIFS_SB(inode->i_sb);
 | |
| 
 | |
| 	if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_NO_PERM) {
 | |
| 		return 0;
 | |
| 	} else /* file mode might have been restricted at mount time 
 | |
| 		on the client (above and beyond ACL on servers) for  
 | |
| 		servers which do not support setting and viewing mode bits,
 | |
| 		so allowing client to check permissions is useful */ 
 | |
| 		return generic_permission(inode, mask, NULL);
 | |
| }
 | |
| 
 | |
| static kmem_cache_t *cifs_inode_cachep;
 | |
| static kmem_cache_t *cifs_req_cachep;
 | |
| static kmem_cache_t *cifs_mid_cachep;
 | |
| kmem_cache_t *cifs_oplock_cachep;
 | |
| static kmem_cache_t *cifs_sm_req_cachep;
 | |
| mempool_t *cifs_sm_req_poolp;
 | |
| mempool_t *cifs_req_poolp;
 | |
| mempool_t *cifs_mid_poolp;
 | |
| 
 | |
| static struct inode *
 | |
| cifs_alloc_inode(struct super_block *sb)
 | |
| {
 | |
| 	struct cifsInodeInfo *cifs_inode;
 | |
| 	cifs_inode = kmem_cache_alloc(cifs_inode_cachep, SLAB_KERNEL);
 | |
| 	if (!cifs_inode)
 | |
| 		return NULL;
 | |
| 	cifs_inode->cifsAttrs = 0x20;	/* default */
 | |
| 	atomic_set(&cifs_inode->inUse, 0);
 | |
| 	cifs_inode->time = 0;
 | |
| 	/* Until the file is open and we have gotten oplock
 | |
| 	info back from the server, can not assume caching of
 | |
| 	file data or metadata */
 | |
| 	cifs_inode->clientCanCacheRead = FALSE;
 | |
| 	cifs_inode->clientCanCacheAll = FALSE;
 | |
| 	cifs_inode->vfs_inode.i_blkbits = 14;  /* 2**14 = CIFS_MAX_MSGSIZE */
 | |
| 	cifs_inode->vfs_inode.i_flags = S_NOATIME | S_NOCMTIME;
 | |
| 	INIT_LIST_HEAD(&cifs_inode->openFileList);
 | |
| 	return &cifs_inode->vfs_inode;
 | |
| }
 | |
| 
 | |
| static void
 | |
| cifs_destroy_inode(struct inode *inode)
 | |
| {
 | |
| 	kmem_cache_free(cifs_inode_cachep, CIFS_I(inode));
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * cifs_show_options() is for displaying mount options in /proc/mounts.
 | |
|  * Not all settable options are displayed but most of the important
 | |
|  * ones are.
 | |
|  */
 | |
| static int
 | |
| cifs_show_options(struct seq_file *s, struct vfsmount *m)
 | |
| {
 | |
| 	struct cifs_sb_info *cifs_sb;
 | |
| 
 | |
| 	cifs_sb = CIFS_SB(m->mnt_sb);
 | |
| 
 | |
| 	if (cifs_sb) {
 | |
| 		if (cifs_sb->tcon) {
 | |
| 			seq_printf(s, ",unc=%s", cifs_sb->tcon->treeName);
 | |
| 			if (cifs_sb->tcon->ses) {
 | |
| 				if (cifs_sb->tcon->ses->userName)
 | |
| 					seq_printf(s, ",username=%s",
 | |
| 					   cifs_sb->tcon->ses->userName);
 | |
| 				if(cifs_sb->tcon->ses->domainName)
 | |
| 					seq_printf(s, ",domain=%s",
 | |
| 					   cifs_sb->tcon->ses->domainName);
 | |
| 			}
 | |
| 		}
 | |
| 		seq_printf(s, ",rsize=%d",cifs_sb->rsize);
 | |
| 		seq_printf(s, ",wsize=%d",cifs_sb->wsize);
 | |
| 	}
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| #ifdef CONFIG_CIFS_QUOTA
 | |
| int cifs_xquota_set(struct super_block * sb, int quota_type, qid_t qid,
 | |
| 		struct fs_disk_quota * pdquota)
 | |
| {
 | |
| 	int xid;
 | |
| 	int rc = 0;
 | |
| 	struct cifs_sb_info *cifs_sb = CIFS_SB(sb);
 | |
| 	struct cifsTconInfo *pTcon;
 | |
| 	
 | |
| 	if(cifs_sb)
 | |
| 		pTcon = cifs_sb->tcon;
 | |
| 	else
 | |
| 		return -EIO;
 | |
| 
 | |
| 
 | |
| 	xid = GetXid();
 | |
| 	if(pTcon) {
 | |
| 		cFYI(1,("set type: 0x%x id: %d",quota_type,qid));		
 | |
| 	} else {
 | |
| 		return -EIO;
 | |
| 	}
 | |
| 
 | |
| 	FreeXid(xid);
 | |
| 	return rc;
 | |
| }
 | |
| 
 | |
| int cifs_xquota_get(struct super_block * sb, int quota_type, qid_t qid,
 | |
|                 struct fs_disk_quota * pdquota)
 | |
| {
 | |
| 	int xid;
 | |
| 	int rc = 0;
 | |
| 	struct cifs_sb_info *cifs_sb = CIFS_SB(sb);
 | |
| 	struct cifsTconInfo *pTcon;
 | |
| 
 | |
| 	if(cifs_sb)
 | |
| 		pTcon = cifs_sb->tcon;
 | |
| 	else
 | |
| 		return -EIO;
 | |
| 
 | |
| 	xid = GetXid();
 | |
| 	if(pTcon) {
 | |
|                 cFYI(1,("set type: 0x%x id: %d",quota_type,qid));
 | |
| 	} else {
 | |
| 		rc = -EIO;
 | |
| 	}
 | |
| 
 | |
| 	FreeXid(xid);
 | |
| 	return rc;
 | |
| }
 | |
| 
 | |
| int cifs_xstate_set(struct super_block * sb, unsigned int flags, int operation)
 | |
| {
 | |
| 	int xid; 
 | |
| 	int rc = 0;
 | |
| 	struct cifs_sb_info *cifs_sb = CIFS_SB(sb);
 | |
| 	struct cifsTconInfo *pTcon;
 | |
| 
 | |
| 	if(cifs_sb)
 | |
| 		pTcon = cifs_sb->tcon;
 | |
| 	else
 | |
| 		return -EIO;
 | |
| 
 | |
| 	xid = GetXid();
 | |
| 	if(pTcon) {
 | |
|                 cFYI(1,("flags: 0x%x operation: 0x%x",flags,operation));
 | |
| 	} else {
 | |
| 		rc = -EIO;
 | |
| 	}
 | |
| 
 | |
| 	FreeXid(xid);
 | |
| 	return rc;
 | |
| }
 | |
| 
 | |
| int cifs_xstate_get(struct super_block * sb, struct fs_quota_stat *qstats)
 | |
| {
 | |
| 	int xid;
 | |
| 	int rc = 0;
 | |
| 	struct cifs_sb_info *cifs_sb = CIFS_SB(sb);
 | |
| 	struct cifsTconInfo *pTcon;
 | |
| 
 | |
| 	if(cifs_sb) {
 | |
| 		pTcon = cifs_sb->tcon;
 | |
| 	} else {
 | |
| 		return -EIO;
 | |
| 	}
 | |
| 	xid = GetXid();
 | |
| 	if(pTcon) {
 | |
| 		cFYI(1,("pqstats %p",qstats));		
 | |
| 	} else {
 | |
| 		rc = -EIO;
 | |
| 	}
 | |
| 
 | |
| 	FreeXid(xid);
 | |
| 	return rc;
 | |
| }
 | |
| 
 | |
| static struct quotactl_ops cifs_quotactl_ops = {
 | |
| 	.set_xquota	= cifs_xquota_set,
 | |
| 	.get_xquota	= cifs_xquota_set,
 | |
| 	.set_xstate	= cifs_xstate_set,
 | |
| 	.get_xstate	= cifs_xstate_get,
 | |
| };
 | |
| #endif
 | |
| 
 | |
| static void cifs_umount_begin(struct vfsmount * vfsmnt, int flags)
 | |
| {
 | |
| 	struct cifs_sb_info *cifs_sb;
 | |
| 	struct cifsTconInfo * tcon;
 | |
| 
 | |
| 	if (!(flags & MNT_FORCE))
 | |
| 		return;
 | |
| 	cifs_sb = CIFS_SB(vfsmnt->mnt_sb);
 | |
| 	if(cifs_sb == NULL)
 | |
| 		return;
 | |
| 
 | |
| 	tcon = cifs_sb->tcon;
 | |
| 	if(tcon == NULL)
 | |
| 		return;
 | |
| 	down(&tcon->tconSem);
 | |
| 	if (atomic_read(&tcon->useCount) == 1)
 | |
| 		tcon->tidStatus = CifsExiting;
 | |
| 	up(&tcon->tconSem);
 | |
| 
 | |
| 	/* cancel_brl_requests(tcon); */ /* BB mark all brl mids as exiting */
 | |
| 	/* cancel_notify_requests(tcon); */
 | |
| 	if(tcon->ses && tcon->ses->server)
 | |
| 	{
 | |
| 		cFYI(1,("wake up tasks now - umount begin not complete"));
 | |
| 		wake_up_all(&tcon->ses->server->request_q);
 | |
| 		wake_up_all(&tcon->ses->server->response_q);
 | |
| 		msleep(1); /* yield */
 | |
| 		/* we have to kick the requests once more */
 | |
| 		wake_up_all(&tcon->ses->server->response_q);
 | |
| 		msleep(1);
 | |
| 	}
 | |
| /* BB FIXME - finish add checks for tidStatus BB */
 | |
| 
 | |
| 	return;
 | |
| }
 | |
| 
 | |
| #ifdef CONFIG_CIFS_STATS2
 | |
| static int cifs_show_stats(struct seq_file *s, struct vfsmount *mnt)
 | |
| {
 | |
| 	/* BB FIXME */
 | |
| 	return 0;
 | |
| }
 | |
| #endif
 | |
| 
 | |
| static int cifs_remount(struct super_block *sb, int *flags, char *data)
 | |
| {
 | |
| 	*flags |= MS_NODIRATIME;
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| static struct super_operations cifs_super_ops = {
 | |
| 	.read_inode = cifs_read_inode,
 | |
| 	.put_super = cifs_put_super,
 | |
| 	.statfs = cifs_statfs,
 | |
| 	.alloc_inode = cifs_alloc_inode,
 | |
| 	.destroy_inode = cifs_destroy_inode,
 | |
| /*	.drop_inode	    = generic_delete_inode, 
 | |
| 	.delete_inode	= cifs_delete_inode,  *//* Do not need the above two functions     
 | |
|    unless later we add lazy close of inodes or unless the kernel forgets to call
 | |
|    us with the same number of releases (closes) as opens */
 | |
| 	.show_options = cifs_show_options,
 | |
| 	.umount_begin   = cifs_umount_begin,
 | |
| 	.remount_fs = cifs_remount,
 | |
| #ifdef CONFIG_CIFS_STATS2
 | |
| 	.show_stats = cifs_show_stats,
 | |
| #endif
 | |
| };
 | |
| 
 | |
| static int
 | |
| cifs_get_sb(struct file_system_type *fs_type,
 | |
| 	    int flags, const char *dev_name, void *data, struct vfsmount *mnt)
 | |
| {
 | |
| 	int rc;
 | |
| 	struct super_block *sb = sget(fs_type, NULL, set_anon_super, NULL);
 | |
| 
 | |
| 	cFYI(1, ("Devname: %s flags: %d ", dev_name, flags));
 | |
| 
 | |
| 	if (IS_ERR(sb))
 | |
| 		return PTR_ERR(sb);
 | |
| 
 | |
| 	sb->s_flags = flags;
 | |
| 
 | |
| 	rc = cifs_read_super(sb, data, dev_name, flags & MS_SILENT ? 1 : 0);
 | |
| 	if (rc) {
 | |
| 		up_write(&sb->s_umount);
 | |
| 		deactivate_super(sb);
 | |
| 		return rc;
 | |
| 	}
 | |
| 	sb->s_flags |= MS_ACTIVE;
 | |
| 	return simple_set_mnt(mnt, sb);
 | |
| }
 | |
| 
 | |
| static ssize_t cifs_file_aio_write(struct kiocb *iocb, const struct iovec *iov,
 | |
| 				   unsigned long nr_segs, loff_t pos)
 | |
| {
 | |
| 	struct inode *inode = iocb->ki_filp->f_dentry->d_inode;
 | |
| 	ssize_t written;
 | |
| 
 | |
| 	written = generic_file_aio_write(iocb, iov, nr_segs, pos);
 | |
| 	if (!CIFS_I(inode)->clientCanCacheAll)
 | |
| 		filemap_fdatawrite(inode->i_mapping);
 | |
| 	return written;
 | |
| }
 | |
| 
 | |
| static loff_t cifs_llseek(struct file *file, loff_t offset, int origin)
 | |
| {
 | |
| 	/* origin == SEEK_END => we must revalidate the cached file length */
 | |
| 	if (origin == SEEK_END) {
 | |
| 		int retval = cifs_revalidate(file->f_dentry);
 | |
| 		if (retval < 0)
 | |
| 			return (loff_t)retval;
 | |
| 	}
 | |
| 	return remote_llseek(file, offset, origin);
 | |
| }
 | |
| 
 | |
| static struct file_system_type cifs_fs_type = {
 | |
| 	.owner = THIS_MODULE,
 | |
| 	.name = "cifs",
 | |
| 	.get_sb = cifs_get_sb,
 | |
| 	.kill_sb = kill_anon_super,
 | |
| 	/*  .fs_flags */
 | |
| };
 | |
| struct inode_operations cifs_dir_inode_ops = {
 | |
| 	.create = cifs_create,
 | |
| 	.lookup = cifs_lookup,
 | |
| 	.getattr = cifs_getattr,
 | |
| 	.unlink = cifs_unlink,
 | |
| 	.link = cifs_hardlink,
 | |
| 	.mkdir = cifs_mkdir,
 | |
| 	.rmdir = cifs_rmdir,
 | |
| 	.rename = cifs_rename,
 | |
| 	.permission = cifs_permission,
 | |
| /*	revalidate:cifs_revalidate,   */
 | |
| 	.setattr = cifs_setattr,
 | |
| 	.symlink = cifs_symlink,
 | |
| 	.mknod   = cifs_mknod,
 | |
| #ifdef CONFIG_CIFS_XATTR
 | |
| 	.setxattr = cifs_setxattr,
 | |
| 	.getxattr = cifs_getxattr,
 | |
| 	.listxattr = cifs_listxattr,
 | |
| 	.removexattr = cifs_removexattr,
 | |
| #endif
 | |
| };
 | |
| 
 | |
| struct inode_operations cifs_file_inode_ops = {
 | |
| /*	revalidate:cifs_revalidate, */
 | |
| 	.setattr = cifs_setattr,
 | |
| 	.getattr = cifs_getattr, /* do we need this anymore? */
 | |
| 	.rename = cifs_rename,
 | |
| 	.permission = cifs_permission,
 | |
| #ifdef CONFIG_CIFS_XATTR
 | |
| 	.setxattr = cifs_setxattr,
 | |
| 	.getxattr = cifs_getxattr,
 | |
| 	.listxattr = cifs_listxattr,
 | |
| 	.removexattr = cifs_removexattr,
 | |
| #endif 
 | |
| };
 | |
| 
 | |
| struct inode_operations cifs_symlink_inode_ops = {
 | |
| 	.readlink = generic_readlink, 
 | |
| 	.follow_link = cifs_follow_link,
 | |
| 	.put_link = cifs_put_link,
 | |
| 	.permission = cifs_permission,
 | |
| 	/* BB add the following two eventually */
 | |
| 	/* revalidate: cifs_revalidate,
 | |
| 	   setattr:    cifs_notify_change, *//* BB do we need notify change */
 | |
| #ifdef CONFIG_CIFS_XATTR
 | |
| 	.setxattr = cifs_setxattr,
 | |
| 	.getxattr = cifs_getxattr,
 | |
| 	.listxattr = cifs_listxattr,
 | |
| 	.removexattr = cifs_removexattr,
 | |
| #endif 
 | |
| };
 | |
| 
 | |
| const struct file_operations cifs_file_ops = {
 | |
| 	.read = do_sync_read,
 | |
| 	.write = do_sync_write,
 | |
| 	.aio_read = generic_file_aio_read,
 | |
| 	.aio_write = cifs_file_aio_write,
 | |
| 	.open = cifs_open,
 | |
| 	.release = cifs_close,
 | |
| 	.lock = cifs_lock,
 | |
| 	.fsync = cifs_fsync,
 | |
| 	.flush = cifs_flush,
 | |
| 	.mmap  = cifs_file_mmap,
 | |
| 	.sendfile = generic_file_sendfile,
 | |
| 	.llseek = cifs_llseek,
 | |
| #ifdef CONFIG_CIFS_POSIX
 | |
| 	.ioctl	= cifs_ioctl,
 | |
| #endif /* CONFIG_CIFS_POSIX */
 | |
| 
 | |
| #ifdef CONFIG_CIFS_EXPERIMENTAL
 | |
| 	.dir_notify = cifs_dir_notify,
 | |
| #endif /* CONFIG_CIFS_EXPERIMENTAL */
 | |
| };
 | |
| 
 | |
| const struct file_operations cifs_file_direct_ops = {
 | |
| 	/* no mmap, no aio, no readv - 
 | |
| 	   BB reevaluate whether they can be done with directio, no cache */
 | |
| 	.read = cifs_user_read,
 | |
| 	.write = cifs_user_write,
 | |
| 	.open = cifs_open,
 | |
| 	.release = cifs_close,
 | |
| 	.lock = cifs_lock,
 | |
| 	.fsync = cifs_fsync,
 | |
| 	.flush = cifs_flush,
 | |
| 	.sendfile = generic_file_sendfile, /* BB removeme BB */
 | |
| #ifdef CONFIG_CIFS_POSIX
 | |
| 	.ioctl  = cifs_ioctl,
 | |
| #endif /* CONFIG_CIFS_POSIX */
 | |
| 	.llseek = cifs_llseek,
 | |
| #ifdef CONFIG_CIFS_EXPERIMENTAL
 | |
| 	.dir_notify = cifs_dir_notify,
 | |
| #endif /* CONFIG_CIFS_EXPERIMENTAL */
 | |
| };
 | |
| const struct file_operations cifs_file_nobrl_ops = {
 | |
| 	.read = do_sync_read,
 | |
| 	.write = do_sync_write,
 | |
| 	.aio_read = generic_file_aio_read,
 | |
| 	.aio_write = cifs_file_aio_write,
 | |
| 	.open = cifs_open,
 | |
| 	.release = cifs_close,
 | |
| 	.fsync = cifs_fsync,
 | |
| 	.flush = cifs_flush,
 | |
| 	.mmap  = cifs_file_mmap,
 | |
| 	.sendfile = generic_file_sendfile,
 | |
| 	.llseek = cifs_llseek,
 | |
| #ifdef CONFIG_CIFS_POSIX
 | |
| 	.ioctl	= cifs_ioctl,
 | |
| #endif /* CONFIG_CIFS_POSIX */
 | |
| 
 | |
| #ifdef CONFIG_CIFS_EXPERIMENTAL
 | |
| 	.dir_notify = cifs_dir_notify,
 | |
| #endif /* CONFIG_CIFS_EXPERIMENTAL */
 | |
| };
 | |
| 
 | |
| const struct file_operations cifs_file_direct_nobrl_ops = {
 | |
| 	/* no mmap, no aio, no readv - 
 | |
| 	   BB reevaluate whether they can be done with directio, no cache */
 | |
| 	.read = cifs_user_read,
 | |
| 	.write = cifs_user_write,
 | |
| 	.open = cifs_open,
 | |
| 	.release = cifs_close,
 | |
| 	.fsync = cifs_fsync,
 | |
| 	.flush = cifs_flush,
 | |
| 	.sendfile = generic_file_sendfile, /* BB removeme BB */
 | |
| #ifdef CONFIG_CIFS_POSIX
 | |
| 	.ioctl  = cifs_ioctl,
 | |
| #endif /* CONFIG_CIFS_POSIX */
 | |
| 	.llseek = cifs_llseek,
 | |
| #ifdef CONFIG_CIFS_EXPERIMENTAL
 | |
| 	.dir_notify = cifs_dir_notify,
 | |
| #endif /* CONFIG_CIFS_EXPERIMENTAL */
 | |
| };
 | |
| 
 | |
| const struct file_operations cifs_dir_ops = {
 | |
| 	.readdir = cifs_readdir,
 | |
| 	.release = cifs_closedir,
 | |
| 	.read    = generic_read_dir,
 | |
| #ifdef CONFIG_CIFS_EXPERIMENTAL
 | |
| 	.dir_notify = cifs_dir_notify,
 | |
| #endif /* CONFIG_CIFS_EXPERIMENTAL */
 | |
|         .ioctl  = cifs_ioctl,
 | |
| };
 | |
| 
 | |
| static void
 | |
| cifs_init_once(void *inode, kmem_cache_t * cachep, unsigned long flags)
 | |
| {
 | |
| 	struct cifsInodeInfo *cifsi = inode;
 | |
| 
 | |
| 	if ((flags & (SLAB_CTOR_VERIFY | SLAB_CTOR_CONSTRUCTOR)) ==
 | |
| 	    SLAB_CTOR_CONSTRUCTOR) {
 | |
| 		inode_init_once(&cifsi->vfs_inode);
 | |
| 		INIT_LIST_HEAD(&cifsi->lockList);
 | |
| 	}
 | |
| }
 | |
| 
 | |
| static int
 | |
| cifs_init_inodecache(void)
 | |
| {
 | |
| 	cifs_inode_cachep = kmem_cache_create("cifs_inode_cache",
 | |
| 					      sizeof (struct cifsInodeInfo),
 | |
| 					      0, (SLAB_RECLAIM_ACCOUNT|
 | |
| 						SLAB_MEM_SPREAD),
 | |
| 					      cifs_init_once, NULL);
 | |
| 	if (cifs_inode_cachep == NULL)
 | |
| 		return -ENOMEM;
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| static void
 | |
| cifs_destroy_inodecache(void)
 | |
| {
 | |
| 	kmem_cache_destroy(cifs_inode_cachep);
 | |
| }
 | |
| 
 | |
| static int
 | |
| cifs_init_request_bufs(void)
 | |
| {
 | |
| 	if(CIFSMaxBufSize < 8192) {
 | |
| 	/* Buffer size can not be smaller than 2 * PATH_MAX since maximum
 | |
| 	Unicode path name has to fit in any SMB/CIFS path based frames */
 | |
| 		CIFSMaxBufSize = 8192;
 | |
| 	} else if (CIFSMaxBufSize > 1024*127) {
 | |
| 		CIFSMaxBufSize = 1024 * 127;
 | |
| 	} else {
 | |
| 		CIFSMaxBufSize &= 0x1FE00; /* Round size to even 512 byte mult*/
 | |
| 	}
 | |
| /*	cERROR(1,("CIFSMaxBufSize %d 0x%x",CIFSMaxBufSize,CIFSMaxBufSize)); */
 | |
| 	cifs_req_cachep = kmem_cache_create("cifs_request",
 | |
| 					    CIFSMaxBufSize +
 | |
| 					    MAX_CIFS_HDR_SIZE, 0,
 | |
| 					    SLAB_HWCACHE_ALIGN, NULL, NULL);
 | |
| 	if (cifs_req_cachep == NULL)
 | |
| 		return -ENOMEM;
 | |
| 
 | |
| 	if(cifs_min_rcv < 1)
 | |
| 		cifs_min_rcv = 1;
 | |
| 	else if (cifs_min_rcv > 64) {
 | |
| 		cifs_min_rcv = 64;
 | |
| 		cERROR(1,("cifs_min_rcv set to maximum (64)"));
 | |
| 	}
 | |
| 
 | |
| 	cifs_req_poolp = mempool_create_slab_pool(cifs_min_rcv,
 | |
| 						  cifs_req_cachep);
 | |
| 
 | |
| 	if(cifs_req_poolp == NULL) {
 | |
| 		kmem_cache_destroy(cifs_req_cachep);
 | |
| 		return -ENOMEM;
 | |
| 	}
 | |
| 	/* MAX_CIFS_SMALL_BUFFER_SIZE bytes is enough for most SMB responses and
 | |
| 	almost all handle based requests (but not write response, nor is it
 | |
| 	sufficient for path based requests).  A smaller size would have
 | |
| 	been more efficient (compacting multiple slab items on one 4k page) 
 | |
| 	for the case in which debug was on, but this larger size allows
 | |
| 	more SMBs to use small buffer alloc and is still much more
 | |
| 	efficient to alloc 1 per page off the slab compared to 17K (5page) 
 | |
| 	alloc of large cifs buffers even when page debugging is on */
 | |
| 	cifs_sm_req_cachep = kmem_cache_create("cifs_small_rq",
 | |
| 			MAX_CIFS_SMALL_BUFFER_SIZE, 0, SLAB_HWCACHE_ALIGN, 
 | |
| 			NULL, NULL);
 | |
| 	if (cifs_sm_req_cachep == NULL) {
 | |
| 		mempool_destroy(cifs_req_poolp);
 | |
| 		kmem_cache_destroy(cifs_req_cachep);
 | |
| 		return -ENOMEM;              
 | |
| 	}
 | |
| 
 | |
| 	if(cifs_min_small < 2)
 | |
| 		cifs_min_small = 2;
 | |
| 	else if (cifs_min_small > 256) {
 | |
| 		cifs_min_small = 256;
 | |
| 		cFYI(1,("cifs_min_small set to maximum (256)"));
 | |
| 	}
 | |
| 
 | |
| 	cifs_sm_req_poolp = mempool_create_slab_pool(cifs_min_small,
 | |
| 						     cifs_sm_req_cachep);
 | |
| 
 | |
| 	if(cifs_sm_req_poolp == NULL) {
 | |
| 		mempool_destroy(cifs_req_poolp);
 | |
| 		kmem_cache_destroy(cifs_req_cachep);
 | |
| 		kmem_cache_destroy(cifs_sm_req_cachep);
 | |
| 		return -ENOMEM;
 | |
| 	}
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| static void
 | |
| cifs_destroy_request_bufs(void)
 | |
| {
 | |
| 	mempool_destroy(cifs_req_poolp);
 | |
| 	kmem_cache_destroy(cifs_req_cachep);
 | |
| 	mempool_destroy(cifs_sm_req_poolp);
 | |
| 	kmem_cache_destroy(cifs_sm_req_cachep);
 | |
| }
 | |
| 
 | |
| static int
 | |
| cifs_init_mids(void)
 | |
| {
 | |
| 	cifs_mid_cachep = kmem_cache_create("cifs_mpx_ids",
 | |
| 				sizeof (struct mid_q_entry), 0,
 | |
| 				SLAB_HWCACHE_ALIGN, NULL, NULL);
 | |
| 	if (cifs_mid_cachep == NULL)
 | |
| 		return -ENOMEM;
 | |
| 
 | |
| 	/* 3 is a reasonable minimum number of simultaneous operations */
 | |
| 	cifs_mid_poolp = mempool_create_slab_pool(3, cifs_mid_cachep);
 | |
| 	if(cifs_mid_poolp == NULL) {
 | |
| 		kmem_cache_destroy(cifs_mid_cachep);
 | |
| 		return -ENOMEM;
 | |
| 	}
 | |
| 
 | |
| 	cifs_oplock_cachep = kmem_cache_create("cifs_oplock_structs",
 | |
| 				sizeof (struct oplock_q_entry), 0,
 | |
| 				SLAB_HWCACHE_ALIGN, NULL, NULL);
 | |
| 	if (cifs_oplock_cachep == NULL) {
 | |
| 		kmem_cache_destroy(cifs_mid_cachep);
 | |
| 		mempool_destroy(cifs_mid_poolp);
 | |
| 		return -ENOMEM;
 | |
| 	}
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| static void
 | |
| cifs_destroy_mids(void)
 | |
| {
 | |
| 	mempool_destroy(cifs_mid_poolp);
 | |
| 	kmem_cache_destroy(cifs_mid_cachep);
 | |
| 	kmem_cache_destroy(cifs_oplock_cachep);
 | |
| }
 | |
| 
 | |
| static int cifs_oplock_thread(void * dummyarg)
 | |
| {
 | |
| 	struct oplock_q_entry * oplock_item;
 | |
| 	struct cifsTconInfo *pTcon;
 | |
| 	struct inode * inode;
 | |
| 	__u16  netfid;
 | |
| 	int rc;
 | |
| 
 | |
| 	do {
 | |
| 		if (try_to_freeze()) 
 | |
| 			continue;
 | |
| 		
 | |
| 		spin_lock(&GlobalMid_Lock);
 | |
| 		if(list_empty(&GlobalOplock_Q)) {
 | |
| 			spin_unlock(&GlobalMid_Lock);
 | |
| 			set_current_state(TASK_INTERRUPTIBLE);
 | |
| 			schedule_timeout(39*HZ);
 | |
| 		} else {
 | |
| 			oplock_item = list_entry(GlobalOplock_Q.next, 
 | |
| 				struct oplock_q_entry, qhead);
 | |
| 			if(oplock_item) {
 | |
| 				cFYI(1,("found oplock item to write out")); 
 | |
| 				pTcon = oplock_item->tcon;
 | |
| 				inode = oplock_item->pinode;
 | |
| 				netfid = oplock_item->netfid;
 | |
| 				spin_unlock(&GlobalMid_Lock);
 | |
| 				DeleteOplockQEntry(oplock_item);
 | |
| 				/* can not grab inode sem here since it would
 | |
| 				deadlock when oplock received on delete 
 | |
| 				since vfs_unlink holds the i_mutex across
 | |
| 				the call */
 | |
| 				/* mutex_lock(&inode->i_mutex);*/
 | |
| 				if (S_ISREG(inode->i_mode)) {
 | |
| 					rc = filemap_fdatawrite(inode->i_mapping);
 | |
| 					if(CIFS_I(inode)->clientCanCacheRead == 0) {
 | |
| 						filemap_fdatawait(inode->i_mapping);
 | |
| 						invalidate_remote_inode(inode);
 | |
| 					}
 | |
| 				} else
 | |
| 					rc = 0;
 | |
| 				/* mutex_unlock(&inode->i_mutex);*/
 | |
| 				if (rc)
 | |
| 					CIFS_I(inode)->write_behind_rc = rc;
 | |
| 				cFYI(1,("Oplock flush inode %p rc %d",inode,rc));
 | |
| 
 | |
| 				/* releasing a stale oplock after recent reconnection 
 | |
| 				of smb session using a now incorrect file 
 | |
| 				handle is not a data integrity issue but do  
 | |
| 				not bother sending an oplock release if session 
 | |
| 				to server still is disconnected since oplock 
 | |
| 				already released by the server in that case */
 | |
| 				if(pTcon->tidStatus != CifsNeedReconnect) {
 | |
| 				    rc = CIFSSMBLock(0, pTcon, netfid,
 | |
| 					    0 /* len */ , 0 /* offset */, 0, 
 | |
| 					    0, LOCKING_ANDX_OPLOCK_RELEASE,
 | |
| 					    0 /* wait flag */);
 | |
| 					cFYI(1,("Oplock release rc = %d ",rc));
 | |
| 				}
 | |
| 			} else
 | |
| 				spin_unlock(&GlobalMid_Lock);
 | |
| 			set_current_state(TASK_INTERRUPTIBLE);
 | |
| 			schedule_timeout(1);  /* yield in case q were corrupt */
 | |
| 		}
 | |
| 	} while (!kthread_should_stop());
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| static int cifs_dnotify_thread(void * dummyarg)
 | |
| {
 | |
| 	struct list_head *tmp;
 | |
| 	struct cifsSesInfo *ses;
 | |
| 
 | |
| 	do {
 | |
| 		if (try_to_freeze())
 | |
| 			continue;
 | |
| 		set_current_state(TASK_INTERRUPTIBLE);
 | |
| 		schedule_timeout(15*HZ);
 | |
| 		read_lock(&GlobalSMBSeslock);
 | |
| 		/* check if any stuck requests that need
 | |
| 		   to be woken up and wakeq so the
 | |
| 		   thread can wake up and error out */
 | |
| 		list_for_each(tmp, &GlobalSMBSessionList) {
 | |
| 			ses = list_entry(tmp, struct cifsSesInfo, 
 | |
| 				cifsSessionList);
 | |
| 			if(ses && ses->server && 
 | |
| 			     atomic_read(&ses->server->inFlight))
 | |
| 				wake_up_all(&ses->server->response_q);
 | |
| 		}
 | |
| 		read_unlock(&GlobalSMBSeslock);
 | |
| 	} while (!kthread_should_stop());
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| static int __init
 | |
| init_cifs(void)
 | |
| {
 | |
| 	int rc = 0;
 | |
| #ifdef CONFIG_PROC_FS
 | |
| 	cifs_proc_init();
 | |
| #endif
 | |
| /*	INIT_LIST_HEAD(&GlobalServerList);*/	/* BB not implemented yet */
 | |
| 	INIT_LIST_HEAD(&GlobalSMBSessionList);
 | |
| 	INIT_LIST_HEAD(&GlobalTreeConnectionList);
 | |
| 	INIT_LIST_HEAD(&GlobalOplock_Q);
 | |
| #ifdef CONFIG_CIFS_EXPERIMENTAL
 | |
| 	INIT_LIST_HEAD(&GlobalDnotifyReqList);
 | |
| 	INIT_LIST_HEAD(&GlobalDnotifyRsp_Q);
 | |
| #endif	
 | |
| /*
 | |
|  *  Initialize Global counters
 | |
|  */
 | |
| 	atomic_set(&sesInfoAllocCount, 0);
 | |
| 	atomic_set(&tconInfoAllocCount, 0);
 | |
| 	atomic_set(&tcpSesAllocCount,0);
 | |
| 	atomic_set(&tcpSesReconnectCount, 0);
 | |
| 	atomic_set(&tconInfoReconnectCount, 0);
 | |
| 
 | |
| 	atomic_set(&bufAllocCount, 0);
 | |
| 	atomic_set(&smBufAllocCount, 0);
 | |
| #ifdef CONFIG_CIFS_STATS2
 | |
| 	atomic_set(&totBufAllocCount, 0);
 | |
| 	atomic_set(&totSmBufAllocCount, 0);
 | |
| #endif /* CONFIG_CIFS_STATS2 */
 | |
| 
 | |
| 	atomic_set(&midCount, 0);
 | |
| 	GlobalCurrentXid = 0;
 | |
| 	GlobalTotalActiveXid = 0;
 | |
| 	GlobalMaxActiveXid = 0;
 | |
| 	memset(Local_System_Name, 0, 15);
 | |
| 	rwlock_init(&GlobalSMBSeslock);
 | |
| 	spin_lock_init(&GlobalMid_Lock);
 | |
| 
 | |
| 	if(cifs_max_pending < 2) {
 | |
| 		cifs_max_pending = 2;
 | |
| 		cFYI(1,("cifs_max_pending set to min of 2"));
 | |
| 	} else if(cifs_max_pending > 256) {
 | |
| 		cifs_max_pending = 256;
 | |
| 		cFYI(1,("cifs_max_pending set to max of 256"));
 | |
| 	}
 | |
| 
 | |
| 	rc = cifs_init_inodecache();
 | |
| 	if (rc)
 | |
| 		goto out_clean_proc;
 | |
| 
 | |
| 	rc = cifs_init_mids();
 | |
| 	if (rc)
 | |
| 		goto out_destroy_inodecache;
 | |
| 
 | |
| 	rc = cifs_init_request_bufs();
 | |
| 	if (rc)
 | |
| 		goto out_destroy_mids;
 | |
| 
 | |
| 	rc = register_filesystem(&cifs_fs_type);
 | |
| 	if (rc)
 | |
| 		goto out_destroy_request_bufs;
 | |
| 
 | |
| 	oplockThread = kthread_run(cifs_oplock_thread, NULL, "cifsoplockd");
 | |
| 	if (IS_ERR(oplockThread)) {
 | |
| 		rc = PTR_ERR(oplockThread);
 | |
| 		cERROR(1,("error %d create oplock thread", rc));
 | |
| 		goto out_unregister_filesystem;
 | |
| 	}
 | |
| 
 | |
| 	dnotifyThread = kthread_run(cifs_dnotify_thread, NULL, "cifsdnotifyd");
 | |
| 	if (IS_ERR(dnotifyThread)) {
 | |
| 		rc = PTR_ERR(dnotifyThread);
 | |
| 		cERROR(1,("error %d create dnotify thread", rc));
 | |
| 		goto out_stop_oplock_thread;
 | |
| 	}
 | |
| 
 | |
| 	return 0;
 | |
| 
 | |
|  out_stop_oplock_thread:
 | |
| 	kthread_stop(oplockThread);
 | |
|  out_unregister_filesystem:
 | |
| 	unregister_filesystem(&cifs_fs_type);
 | |
|  out_destroy_request_bufs:
 | |
| 	cifs_destroy_request_bufs();
 | |
|  out_destroy_mids:
 | |
| 	cifs_destroy_mids();
 | |
|  out_destroy_inodecache:
 | |
| 	cifs_destroy_inodecache();
 | |
|  out_clean_proc:
 | |
| #ifdef CONFIG_PROC_FS
 | |
| 	cifs_proc_clean();
 | |
| #endif
 | |
| 	return rc;
 | |
| }
 | |
| 
 | |
| static void __exit
 | |
| exit_cifs(void)
 | |
| {
 | |
| 	cFYI(0, ("In unregister ie exit_cifs"));
 | |
| #ifdef CONFIG_PROC_FS
 | |
| 	cifs_proc_clean();
 | |
| #endif
 | |
| 	unregister_filesystem(&cifs_fs_type);
 | |
| 	cifs_destroy_inodecache();
 | |
| 	cifs_destroy_mids();
 | |
| 	cifs_destroy_request_bufs();
 | |
| 	kthread_stop(oplockThread);
 | |
| 	kthread_stop(dnotifyThread);
 | |
| }
 | |
| 
 | |
| MODULE_AUTHOR("Steve French <sfrench@us.ibm.com>");
 | |
| MODULE_LICENSE("GPL");		/* combination of LGPL + GPL source behaves as GPL */
 | |
| MODULE_DESCRIPTION
 | |
|     ("VFS to access servers complying with the SNIA CIFS Specification e.g. Samba and Windows");
 | |
| MODULE_VERSION(CIFS_VERSION);
 | |
| module_init(init_cifs)
 | |
| module_exit(exit_cifs)
 |