1 // SPDX-License-Identifier: LGPL-2.1 2 /* 3 * 4 * Copyright (C) International Business Machines Corp., 2002,2008 5 * Author(s): Steve French (sfrench@us.ibm.com) 6 * 7 * Common Internet FileSystem (CIFS) client 8 * 9 */ 10 11 /* Note that BB means BUGBUG (ie something to fix eventually) */ 12 13 #include <linux/module.h> 14 #include <linux/fs.h> 15 #include <linux/filelock.h> 16 #include <linux/mount.h> 17 #include <linux/slab.h> 18 #include <linux/init.h> 19 #include <linux/list.h> 20 #include <linux/seq_file.h> 21 #include <linux/vfs.h> 22 #include <linux/mempool.h> 23 #include <linux/delay.h> 24 #include <linux/kthread.h> 25 #include <linux/freezer.h> 26 #include <linux/namei.h> 27 #include <linux/random.h> 28 #include <linux/uuid.h> 29 #include <linux/xattr.h> 30 #include <uapi/linux/magic.h> 31 #include <net/ipv6.h> 32 #include "cifsfs.h" 33 #include "cifspdu.h" 34 #define DECLARE_GLOBALS_HERE 35 #include "cifsglob.h" 36 #include "cifsproto.h" 37 #include "cifs_debug.h" 38 #include "cifs_fs_sb.h" 39 #include <linux/mm.h> 40 #include <linux/key-type.h> 41 #include "cifs_spnego.h" 42 #include "fscache.h" 43 #ifdef CONFIG_CIFS_DFS_UPCALL 44 #include "dfs_cache.h" 45 #endif 46 #ifdef CONFIG_CIFS_SWN_UPCALL 47 #include "netlink.h" 48 #endif 49 #include "fs_context.h" 50 #include "cached_dir.h" 51 52 /* 53 * DOS dates from 1980/1/1 through 2107/12/31 54 * Protocol specifications indicate the range should be to 119, which 55 * limits maximum year to 2099. But this range has not been checked. 56 */ 57 #define SMB_DATE_MAX (127<<9 | 12<<5 | 31) 58 #define SMB_DATE_MIN (0<<9 | 1<<5 | 1) 59 #define SMB_TIME_MAX (23<<11 | 59<<5 | 29) 60 61 int cifsFYI = 0; 62 bool traceSMB; 63 bool enable_oplocks = true; 64 bool linuxExtEnabled = true; 65 bool lookupCacheEnabled = true; 66 bool disable_legacy_dialects; /* false by default */ 67 bool enable_gcm_256 = true; 68 bool require_gcm_256; /* false by default */ 69 bool enable_negotiate_signing; /* false by default */ 70 unsigned int global_secflags = CIFSSEC_DEF; 71 /* unsigned int ntlmv2_support = 0; */ 72 unsigned int sign_CIFS_PDUs = 1; 73 74 /* 75 * Global transaction id (XID) information 76 */ 77 unsigned int GlobalCurrentXid; /* protected by GlobalMid_Sem */ 78 unsigned int GlobalTotalActiveXid; /* prot by GlobalMid_Sem */ 79 unsigned int GlobalMaxActiveXid; /* prot by GlobalMid_Sem */ 80 spinlock_t GlobalMid_Lock; /* protects above & list operations on midQ entries */ 81 82 /* 83 * Global counters, updated atomically 84 */ 85 atomic_t sesInfoAllocCount; 86 atomic_t tconInfoAllocCount; 87 atomic_t tcpSesNextId; 88 atomic_t tcpSesAllocCount; 89 atomic_t tcpSesReconnectCount; 90 atomic_t tconInfoReconnectCount; 91 92 atomic_t mid_count; 93 atomic_t buf_alloc_count; 94 atomic_t small_buf_alloc_count; 95 #ifdef CONFIG_CIFS_STATS2 96 atomic_t total_buf_alloc_count; 97 atomic_t total_small_buf_alloc_count; 98 #endif/* STATS2 */ 99 struct list_head cifs_tcp_ses_list; 100 spinlock_t cifs_tcp_ses_lock; 101 static const struct super_operations cifs_super_ops; 102 unsigned int CIFSMaxBufSize = CIFS_MAX_MSGSIZE; 103 module_param(CIFSMaxBufSize, uint, 0444); 104 MODULE_PARM_DESC(CIFSMaxBufSize, "Network buffer size (not including header) " 105 "for CIFS requests. " 106 "Default: 16384 Range: 8192 to 130048"); 107 unsigned int cifs_min_rcv = CIFS_MIN_RCV_POOL; 108 module_param(cifs_min_rcv, uint, 0444); 109 MODULE_PARM_DESC(cifs_min_rcv, "Network buffers in pool. Default: 4 Range: " 110 "1 to 64"); 111 unsigned int cifs_min_small = 30; 112 module_param(cifs_min_small, uint, 0444); 113 MODULE_PARM_DESC(cifs_min_small, "Small network buffers in pool. Default: 30 " 114 "Range: 2 to 256"); 115 unsigned int cifs_max_pending = CIFS_MAX_REQ; 116 module_param(cifs_max_pending, uint, 0444); 117 MODULE_PARM_DESC(cifs_max_pending, "Simultaneous requests to server for " 118 "CIFS/SMB1 dialect (N/A for SMB3) " 119 "Default: 32767 Range: 2 to 32767."); 120 #ifdef CONFIG_CIFS_STATS2 121 unsigned int slow_rsp_threshold = 1; 122 module_param(slow_rsp_threshold, uint, 0644); 123 MODULE_PARM_DESC(slow_rsp_threshold, "Amount of time (in seconds) to wait " 124 "before logging that a response is delayed. " 125 "Default: 1 (if set to 0 disables msg)."); 126 #endif /* STATS2 */ 127 128 module_param(enable_oplocks, bool, 0644); 129 MODULE_PARM_DESC(enable_oplocks, "Enable or disable oplocks. Default: y/Y/1"); 130 131 module_param(enable_gcm_256, bool, 0644); 132 MODULE_PARM_DESC(enable_gcm_256, "Enable requesting strongest (256 bit) GCM encryption. Default: n/N/0"); 133 134 module_param(require_gcm_256, bool, 0644); 135 MODULE_PARM_DESC(require_gcm_256, "Require strongest (256 bit) GCM encryption. Default: n/N/0"); 136 137 module_param(enable_negotiate_signing, bool, 0644); 138 MODULE_PARM_DESC(enable_negotiate_signing, "Enable negotiating packet signing algorithm with server. Default: n/N/0"); 139 140 module_param(disable_legacy_dialects, bool, 0644); 141 MODULE_PARM_DESC(disable_legacy_dialects, "To improve security it may be " 142 "helpful to restrict the ability to " 143 "override the default dialects (SMB2.1, " 144 "SMB3 and SMB3.02) on mount with old " 145 "dialects (CIFS/SMB1 and SMB2) since " 146 "vers=1.0 (CIFS/SMB1) and vers=2.0 are weaker" 147 " and less secure. Default: n/N/0"); 148 149 extern mempool_t *cifs_sm_req_poolp; 150 extern mempool_t *cifs_req_poolp; 151 extern mempool_t *cifs_mid_poolp; 152 153 struct workqueue_struct *cifsiod_wq; 154 struct workqueue_struct *decrypt_wq; 155 struct workqueue_struct *fileinfo_put_wq; 156 struct workqueue_struct *cifsoplockd_wq; 157 struct workqueue_struct *deferredclose_wq; 158 __u32 cifs_lock_secret; 159 160 /* 161 * Bumps refcount for cifs super block. 162 * Note that it should be only called if a referece to VFS super block is 163 * already held, e.g. in open-type syscalls context. Otherwise it can race with 164 * atomic_dec_and_test in deactivate_locked_super. 165 */ 166 void 167 cifs_sb_active(struct super_block *sb) 168 { 169 struct cifs_sb_info *server = CIFS_SB(sb); 170 171 if (atomic_inc_return(&server->active) == 1) 172 atomic_inc(&sb->s_active); 173 } 174 175 void 176 cifs_sb_deactive(struct super_block *sb) 177 { 178 struct cifs_sb_info *server = CIFS_SB(sb); 179 180 if (atomic_dec_and_test(&server->active)) 181 deactivate_super(sb); 182 } 183 184 static int 185 cifs_read_super(struct super_block *sb) 186 { 187 struct inode *inode; 188 struct cifs_sb_info *cifs_sb; 189 struct cifs_tcon *tcon; 190 struct timespec64 ts; 191 int rc = 0; 192 193 cifs_sb = CIFS_SB(sb); 194 tcon = cifs_sb_master_tcon(cifs_sb); 195 196 if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_POSIXACL) 197 sb->s_flags |= SB_POSIXACL; 198 199 if (tcon->snapshot_time) 200 sb->s_flags |= SB_RDONLY; 201 202 if (tcon->ses->capabilities & tcon->ses->server->vals->cap_large_files) 203 sb->s_maxbytes = MAX_LFS_FILESIZE; 204 else 205 sb->s_maxbytes = MAX_NON_LFS; 206 207 /* 208 * Some very old servers like DOS and OS/2 used 2 second granularity 209 * (while all current servers use 100ns granularity - see MS-DTYP) 210 * but 1 second is the maximum allowed granularity for the VFS 211 * so for old servers set time granularity to 1 second while for 212 * everything else (current servers) set it to 100ns. 213 */ 214 if ((tcon->ses->server->vals->protocol_id == SMB10_PROT_ID) && 215 ((tcon->ses->capabilities & 216 tcon->ses->server->vals->cap_nt_find) == 0) && 217 !tcon->unix_ext) { 218 sb->s_time_gran = 1000000000; /* 1 second is max allowed gran */ 219 ts = cnvrtDosUnixTm(cpu_to_le16(SMB_DATE_MIN), 0, 0); 220 sb->s_time_min = ts.tv_sec; 221 ts = cnvrtDosUnixTm(cpu_to_le16(SMB_DATE_MAX), 222 cpu_to_le16(SMB_TIME_MAX), 0); 223 sb->s_time_max = ts.tv_sec; 224 } else { 225 /* 226 * Almost every server, including all SMB2+, uses DCE TIME 227 * ie 100 nanosecond units, since 1601. See MS-DTYP and MS-FSCC 228 */ 229 sb->s_time_gran = 100; 230 ts = cifs_NTtimeToUnix(0); 231 sb->s_time_min = ts.tv_sec; 232 ts = cifs_NTtimeToUnix(cpu_to_le64(S64_MAX)); 233 sb->s_time_max = ts.tv_sec; 234 } 235 236 sb->s_magic = CIFS_SUPER_MAGIC; 237 sb->s_op = &cifs_super_ops; 238 sb->s_xattr = cifs_xattr_handlers; 239 rc = super_setup_bdi(sb); 240 if (rc) 241 goto out_no_root; 242 /* tune readahead according to rsize if readahead size not set on mount */ 243 if (cifs_sb->ctx->rsize == 0) 244 cifs_sb->ctx->rsize = 245 tcon->ses->server->ops->negotiate_rsize(tcon, cifs_sb->ctx); 246 if (cifs_sb->ctx->rasize) 247 sb->s_bdi->ra_pages = cifs_sb->ctx->rasize / PAGE_SIZE; 248 else 249 sb->s_bdi->ra_pages = 2 * (cifs_sb->ctx->rsize / PAGE_SIZE); 250 251 sb->s_blocksize = CIFS_MAX_MSGSIZE; 252 sb->s_blocksize_bits = 14; /* default 2**14 = CIFS_MAX_MSGSIZE */ 253 inode = cifs_root_iget(sb); 254 255 if (IS_ERR(inode)) { 256 rc = PTR_ERR(inode); 257 goto out_no_root; 258 } 259 260 if (tcon->nocase) 261 sb->s_d_op = &cifs_ci_dentry_ops; 262 else 263 sb->s_d_op = &cifs_dentry_ops; 264 265 sb->s_root = d_make_root(inode); 266 if (!sb->s_root) { 267 rc = -ENOMEM; 268 goto out_no_root; 269 } 270 271 #ifdef CONFIG_CIFS_NFSD_EXPORT 272 if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_SERVER_INUM) { 273 cifs_dbg(FYI, "export ops supported\n"); 274 sb->s_export_op = &cifs_export_ops; 275 } 276 #endif /* CONFIG_CIFS_NFSD_EXPORT */ 277 278 return 0; 279 280 out_no_root: 281 cifs_dbg(VFS, "%s: get root inode failed\n", __func__); 282 return rc; 283 } 284 285 static void cifs_kill_sb(struct super_block *sb) 286 { 287 struct cifs_sb_info *cifs_sb = CIFS_SB(sb); 288 289 /* 290 * We ned to release all dentries for the cached directories 291 * before we kill the sb. 292 */ 293 if (cifs_sb->root) { 294 close_all_cached_dirs(cifs_sb); 295 296 /* finally release root dentry */ 297 dput(cifs_sb->root); 298 cifs_sb->root = NULL; 299 } 300 301 kill_anon_super(sb); 302 cifs_umount(cifs_sb); 303 } 304 305 static int 306 cifs_statfs(struct dentry *dentry, struct kstatfs *buf) 307 { 308 struct super_block *sb = dentry->d_sb; 309 struct cifs_sb_info *cifs_sb = CIFS_SB(sb); 310 struct cifs_tcon *tcon = cifs_sb_master_tcon(cifs_sb); 311 struct TCP_Server_Info *server = tcon->ses->server; 312 unsigned int xid; 313 int rc = 0; 314 315 xid = get_xid(); 316 317 if (le32_to_cpu(tcon->fsAttrInfo.MaxPathNameComponentLength) > 0) 318 buf->f_namelen = 319 le32_to_cpu(tcon->fsAttrInfo.MaxPathNameComponentLength); 320 else 321 buf->f_namelen = PATH_MAX; 322 323 buf->f_fsid.val[0] = tcon->vol_serial_number; 324 /* are using part of create time for more randomness, see man statfs */ 325 buf->f_fsid.val[1] = (int)le64_to_cpu(tcon->vol_create_time); 326 327 buf->f_files = 0; /* undefined */ 328 buf->f_ffree = 0; /* unlimited */ 329 330 if (server->ops->queryfs) 331 rc = server->ops->queryfs(xid, tcon, cifs_sb, buf); 332 333 free_xid(xid); 334 return rc; 335 } 336 337 static long cifs_fallocate(struct file *file, int mode, loff_t off, loff_t len) 338 { 339 struct cifs_sb_info *cifs_sb = CIFS_FILE_SB(file); 340 struct cifs_tcon *tcon = cifs_sb_master_tcon(cifs_sb); 341 struct TCP_Server_Info *server = tcon->ses->server; 342 343 if (server->ops->fallocate) 344 return server->ops->fallocate(file, tcon, mode, off, len); 345 346 return -EOPNOTSUPP; 347 } 348 349 static int cifs_permission(struct mnt_idmap *idmap, 350 struct inode *inode, int mask) 351 { 352 struct cifs_sb_info *cifs_sb; 353 354 cifs_sb = CIFS_SB(inode->i_sb); 355 356 if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_NO_PERM) { 357 if ((mask & MAY_EXEC) && !execute_ok(inode)) 358 return -EACCES; 359 else 360 return 0; 361 } else /* file mode might have been restricted at mount time 362 on the client (above and beyond ACL on servers) for 363 servers which do not support setting and viewing mode bits, 364 so allowing client to check permissions is useful */ 365 return generic_permission(&nop_mnt_idmap, inode, mask); 366 } 367 368 static struct kmem_cache *cifs_inode_cachep; 369 static struct kmem_cache *cifs_req_cachep; 370 static struct kmem_cache *cifs_mid_cachep; 371 static struct kmem_cache *cifs_sm_req_cachep; 372 mempool_t *cifs_sm_req_poolp; 373 mempool_t *cifs_req_poolp; 374 mempool_t *cifs_mid_poolp; 375 376 static struct inode * 377 cifs_alloc_inode(struct super_block *sb) 378 { 379 struct cifsInodeInfo *cifs_inode; 380 cifs_inode = alloc_inode_sb(sb, cifs_inode_cachep, GFP_KERNEL); 381 if (!cifs_inode) 382 return NULL; 383 cifs_inode->cifsAttrs = 0x20; /* default */ 384 cifs_inode->time = 0; 385 /* 386 * Until the file is open and we have gotten oplock info back from the 387 * server, can not assume caching of file data or metadata. 388 */ 389 cifs_set_oplock_level(cifs_inode, 0); 390 cifs_inode->flags = 0; 391 spin_lock_init(&cifs_inode->writers_lock); 392 cifs_inode->writers = 0; 393 cifs_inode->netfs.inode.i_blkbits = 14; /* 2**14 = CIFS_MAX_MSGSIZE */ 394 cifs_inode->server_eof = 0; 395 cifs_inode->uniqueid = 0; 396 cifs_inode->createtime = 0; 397 cifs_inode->epoch = 0; 398 spin_lock_init(&cifs_inode->open_file_lock); 399 generate_random_uuid(cifs_inode->lease_key); 400 cifs_inode->symlink_target = NULL; 401 402 /* 403 * Can not set i_flags here - they get immediately overwritten to zero 404 * by the VFS. 405 */ 406 /* cifs_inode->netfs.inode.i_flags = S_NOATIME | S_NOCMTIME; */ 407 INIT_LIST_HEAD(&cifs_inode->openFileList); 408 INIT_LIST_HEAD(&cifs_inode->llist); 409 INIT_LIST_HEAD(&cifs_inode->deferred_closes); 410 spin_lock_init(&cifs_inode->deferred_lock); 411 return &cifs_inode->netfs.inode; 412 } 413 414 static void 415 cifs_free_inode(struct inode *inode) 416 { 417 struct cifsInodeInfo *cinode = CIFS_I(inode); 418 419 if (S_ISLNK(inode->i_mode)) 420 kfree(cinode->symlink_target); 421 kmem_cache_free(cifs_inode_cachep, cinode); 422 } 423 424 static void 425 cifs_evict_inode(struct inode *inode) 426 { 427 truncate_inode_pages_final(&inode->i_data); 428 if (inode->i_state & I_PINNING_FSCACHE_WB) 429 cifs_fscache_unuse_inode_cookie(inode, true); 430 cifs_fscache_release_inode_cookie(inode); 431 clear_inode(inode); 432 } 433 434 static void 435 cifs_show_address(struct seq_file *s, struct TCP_Server_Info *server) 436 { 437 struct sockaddr_in *sa = (struct sockaddr_in *) &server->dstaddr; 438 struct sockaddr_in6 *sa6 = (struct sockaddr_in6 *) &server->dstaddr; 439 440 seq_puts(s, ",addr="); 441 442 switch (server->dstaddr.ss_family) { 443 case AF_INET: 444 seq_printf(s, "%pI4", &sa->sin_addr.s_addr); 445 break; 446 case AF_INET6: 447 seq_printf(s, "%pI6", &sa6->sin6_addr.s6_addr); 448 if (sa6->sin6_scope_id) 449 seq_printf(s, "%%%u", sa6->sin6_scope_id); 450 break; 451 default: 452 seq_puts(s, "(unknown)"); 453 } 454 if (server->rdma) 455 seq_puts(s, ",rdma"); 456 } 457 458 static void 459 cifs_show_security(struct seq_file *s, struct cifs_ses *ses) 460 { 461 if (ses->sectype == Unspecified) { 462 if (ses->user_name == NULL) 463 seq_puts(s, ",sec=none"); 464 return; 465 } 466 467 seq_puts(s, ",sec="); 468 469 switch (ses->sectype) { 470 case NTLMv2: 471 seq_puts(s, "ntlmv2"); 472 break; 473 case Kerberos: 474 seq_puts(s, "krb5"); 475 break; 476 case RawNTLMSSP: 477 seq_puts(s, "ntlmssp"); 478 break; 479 default: 480 /* shouldn't ever happen */ 481 seq_puts(s, "unknown"); 482 break; 483 } 484 485 if (ses->sign) 486 seq_puts(s, "i"); 487 488 if (ses->sectype == Kerberos) 489 seq_printf(s, ",cruid=%u", 490 from_kuid_munged(&init_user_ns, ses->cred_uid)); 491 } 492 493 static void 494 cifs_show_cache_flavor(struct seq_file *s, struct cifs_sb_info *cifs_sb) 495 { 496 seq_puts(s, ",cache="); 497 498 if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_STRICT_IO) 499 seq_puts(s, "strict"); 500 else if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_DIRECT_IO) 501 seq_puts(s, "none"); 502 else if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_RW_CACHE) 503 seq_puts(s, "singleclient"); /* assume only one client access */ 504 else if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_RO_CACHE) 505 seq_puts(s, "ro"); /* read only caching assumed */ 506 else 507 seq_puts(s, "loose"); 508 } 509 510 /* 511 * cifs_show_devname() is used so we show the mount device name with correct 512 * format (e.g. forward slashes vs. back slashes) in /proc/mounts 513 */ 514 static int cifs_show_devname(struct seq_file *m, struct dentry *root) 515 { 516 struct cifs_sb_info *cifs_sb = CIFS_SB(root->d_sb); 517 char *devname = kstrdup(cifs_sb->ctx->source, GFP_KERNEL); 518 519 if (devname == NULL) 520 seq_puts(m, "none"); 521 else { 522 convert_delimiter(devname, '/'); 523 /* escape all spaces in share names */ 524 seq_escape(m, devname, " \t"); 525 kfree(devname); 526 } 527 return 0; 528 } 529 530 /* 531 * cifs_show_options() is for displaying mount options in /proc/mounts. 532 * Not all settable options are displayed but most of the important 533 * ones are. 534 */ 535 static int 536 cifs_show_options(struct seq_file *s, struct dentry *root) 537 { 538 struct cifs_sb_info *cifs_sb = CIFS_SB(root->d_sb); 539 struct cifs_tcon *tcon = cifs_sb_master_tcon(cifs_sb); 540 struct sockaddr *srcaddr; 541 srcaddr = (struct sockaddr *)&tcon->ses->server->srcaddr; 542 543 seq_show_option(s, "vers", tcon->ses->server->vals->version_string); 544 cifs_show_security(s, tcon->ses); 545 cifs_show_cache_flavor(s, cifs_sb); 546 547 if (tcon->no_lease) 548 seq_puts(s, ",nolease"); 549 if (cifs_sb->ctx->multiuser) 550 seq_puts(s, ",multiuser"); 551 else if (tcon->ses->user_name) 552 seq_show_option(s, "username", tcon->ses->user_name); 553 554 if (tcon->ses->domainName && tcon->ses->domainName[0] != 0) 555 seq_show_option(s, "domain", tcon->ses->domainName); 556 557 if (srcaddr->sa_family != AF_UNSPEC) { 558 struct sockaddr_in *saddr4; 559 struct sockaddr_in6 *saddr6; 560 saddr4 = (struct sockaddr_in *)srcaddr; 561 saddr6 = (struct sockaddr_in6 *)srcaddr; 562 if (srcaddr->sa_family == AF_INET6) 563 seq_printf(s, ",srcaddr=%pI6c", 564 &saddr6->sin6_addr); 565 else if (srcaddr->sa_family == AF_INET) 566 seq_printf(s, ",srcaddr=%pI4", 567 &saddr4->sin_addr.s_addr); 568 else 569 seq_printf(s, ",srcaddr=BAD-AF:%i", 570 (int)(srcaddr->sa_family)); 571 } 572 573 seq_printf(s, ",uid=%u", 574 from_kuid_munged(&init_user_ns, cifs_sb->ctx->linux_uid)); 575 if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_OVERR_UID) 576 seq_puts(s, ",forceuid"); 577 else 578 seq_puts(s, ",noforceuid"); 579 580 seq_printf(s, ",gid=%u", 581 from_kgid_munged(&init_user_ns, cifs_sb->ctx->linux_gid)); 582 if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_OVERR_GID) 583 seq_puts(s, ",forcegid"); 584 else 585 seq_puts(s, ",noforcegid"); 586 587 cifs_show_address(s, tcon->ses->server); 588 589 if (!tcon->unix_ext) 590 seq_printf(s, ",file_mode=0%ho,dir_mode=0%ho", 591 cifs_sb->ctx->file_mode, 592 cifs_sb->ctx->dir_mode); 593 if (cifs_sb->ctx->iocharset) 594 seq_printf(s, ",iocharset=%s", cifs_sb->ctx->iocharset); 595 if (tcon->seal) 596 seq_puts(s, ",seal"); 597 else if (tcon->ses->server->ignore_signature) 598 seq_puts(s, ",signloosely"); 599 if (tcon->nocase) 600 seq_puts(s, ",nocase"); 601 if (tcon->nodelete) 602 seq_puts(s, ",nodelete"); 603 if (cifs_sb->ctx->no_sparse) 604 seq_puts(s, ",nosparse"); 605 if (tcon->local_lease) 606 seq_puts(s, ",locallease"); 607 if (tcon->retry) 608 seq_puts(s, ",hard"); 609 else 610 seq_puts(s, ",soft"); 611 if (tcon->use_persistent) 612 seq_puts(s, ",persistenthandles"); 613 else if (tcon->use_resilient) 614 seq_puts(s, ",resilienthandles"); 615 if (tcon->posix_extensions) 616 seq_puts(s, ",posix"); 617 else if (tcon->unix_ext) 618 seq_puts(s, ",unix"); 619 else 620 seq_puts(s, ",nounix"); 621 if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_NO_DFS) 622 seq_puts(s, ",nodfs"); 623 if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_POSIX_PATHS) 624 seq_puts(s, ",posixpaths"); 625 if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_SET_UID) 626 seq_puts(s, ",setuids"); 627 if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_UID_FROM_ACL) 628 seq_puts(s, ",idsfromsid"); 629 if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_SERVER_INUM) 630 seq_puts(s, ",serverino"); 631 if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_RWPIDFORWARD) 632 seq_puts(s, ",rwpidforward"); 633 if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_NOPOSIXBRL) 634 seq_puts(s, ",forcemand"); 635 if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_NO_XATTR) 636 seq_puts(s, ",nouser_xattr"); 637 if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_MAP_SPECIAL_CHR) 638 seq_puts(s, ",mapchars"); 639 if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_MAP_SFM_CHR) 640 seq_puts(s, ",mapposix"); 641 if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_UNX_EMUL) 642 seq_puts(s, ",sfu"); 643 if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_NO_BRL) 644 seq_puts(s, ",nobrl"); 645 if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_NO_HANDLE_CACHE) 646 seq_puts(s, ",nohandlecache"); 647 if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_MODE_FROM_SID) 648 seq_puts(s, ",modefromsid"); 649 if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_CIFS_ACL) 650 seq_puts(s, ",cifsacl"); 651 if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_DYNPERM) 652 seq_puts(s, ",dynperm"); 653 if (root->d_sb->s_flags & SB_POSIXACL) 654 seq_puts(s, ",acl"); 655 if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_MF_SYMLINKS) 656 seq_puts(s, ",mfsymlinks"); 657 if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_FSCACHE) 658 seq_puts(s, ",fsc"); 659 if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_NOSSYNC) 660 seq_puts(s, ",nostrictsync"); 661 if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_NO_PERM) 662 seq_puts(s, ",noperm"); 663 if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_CIFS_BACKUPUID) 664 seq_printf(s, ",backupuid=%u", 665 from_kuid_munged(&init_user_ns, 666 cifs_sb->ctx->backupuid)); 667 if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_CIFS_BACKUPGID) 668 seq_printf(s, ",backupgid=%u", 669 from_kgid_munged(&init_user_ns, 670 cifs_sb->ctx->backupgid)); 671 672 seq_printf(s, ",rsize=%u", cifs_sb->ctx->rsize); 673 seq_printf(s, ",wsize=%u", cifs_sb->ctx->wsize); 674 seq_printf(s, ",bsize=%u", cifs_sb->ctx->bsize); 675 if (cifs_sb->ctx->rasize) 676 seq_printf(s, ",rasize=%u", cifs_sb->ctx->rasize); 677 if (tcon->ses->server->min_offload) 678 seq_printf(s, ",esize=%u", tcon->ses->server->min_offload); 679 seq_printf(s, ",echo_interval=%lu", 680 tcon->ses->server->echo_interval / HZ); 681 682 /* Only display the following if overridden on mount */ 683 if (tcon->ses->server->max_credits != SMB2_MAX_CREDITS_AVAILABLE) 684 seq_printf(s, ",max_credits=%u", tcon->ses->server->max_credits); 685 if (tcon->ses->server->tcp_nodelay) 686 seq_puts(s, ",tcpnodelay"); 687 if (tcon->ses->server->noautotune) 688 seq_puts(s, ",noautotune"); 689 if (tcon->ses->server->noblocksnd) 690 seq_puts(s, ",noblocksend"); 691 if (tcon->ses->server->nosharesock) 692 seq_puts(s, ",nosharesock"); 693 694 if (tcon->snapshot_time) 695 seq_printf(s, ",snapshot=%llu", tcon->snapshot_time); 696 if (tcon->handle_timeout) 697 seq_printf(s, ",handletimeout=%u", tcon->handle_timeout); 698 699 /* 700 * Display file and directory attribute timeout in seconds. 701 * If file and directory attribute timeout the same then actimeo 702 * was likely specified on mount 703 */ 704 if (cifs_sb->ctx->acdirmax == cifs_sb->ctx->acregmax) 705 seq_printf(s, ",actimeo=%lu", cifs_sb->ctx->acregmax / HZ); 706 else { 707 seq_printf(s, ",acdirmax=%lu", cifs_sb->ctx->acdirmax / HZ); 708 seq_printf(s, ",acregmax=%lu", cifs_sb->ctx->acregmax / HZ); 709 } 710 seq_printf(s, ",closetimeo=%lu", cifs_sb->ctx->closetimeo / HZ); 711 712 if (tcon->ses->chan_max > 1) 713 seq_printf(s, ",multichannel,max_channels=%zu", 714 tcon->ses->chan_max); 715 716 if (tcon->use_witness) 717 seq_puts(s, ",witness"); 718 719 return 0; 720 } 721 722 static void cifs_umount_begin(struct super_block *sb) 723 { 724 struct cifs_sb_info *cifs_sb = CIFS_SB(sb); 725 struct cifs_tcon *tcon; 726 727 if (cifs_sb == NULL) 728 return; 729 730 tcon = cifs_sb_master_tcon(cifs_sb); 731 732 spin_lock(&cifs_tcp_ses_lock); 733 spin_lock(&tcon->tc_lock); 734 if ((tcon->tc_count > 1) || (tcon->status == TID_EXITING)) { 735 /* we have other mounts to same share or we have 736 already tried to umount this and woken up 737 all waiting network requests, nothing to do */ 738 spin_unlock(&tcon->tc_lock); 739 spin_unlock(&cifs_tcp_ses_lock); 740 return; 741 } 742 /* 743 * can not set tcon->status to TID_EXITING yet since we don't know if umount -f will 744 * fail later (e.g. due to open files). TID_EXITING will be set just before tdis req sent 745 */ 746 spin_unlock(&tcon->tc_lock); 747 spin_unlock(&cifs_tcp_ses_lock); 748 749 cifs_close_all_deferred_files(tcon); 750 /* cancel_brl_requests(tcon); */ /* BB mark all brl mids as exiting */ 751 /* cancel_notify_requests(tcon); */ 752 if (tcon->ses && tcon->ses->server) { 753 cifs_dbg(FYI, "wake up tasks now - umount begin not complete\n"); 754 wake_up_all(&tcon->ses->server->request_q); 755 wake_up_all(&tcon->ses->server->response_q); 756 msleep(1); /* yield */ 757 /* we have to kick the requests once more */ 758 wake_up_all(&tcon->ses->server->response_q); 759 msleep(1); 760 } 761 762 return; 763 } 764 765 static int cifs_freeze(struct super_block *sb) 766 { 767 struct cifs_sb_info *cifs_sb = CIFS_SB(sb); 768 struct cifs_tcon *tcon; 769 770 if (cifs_sb == NULL) 771 return 0; 772 773 tcon = cifs_sb_master_tcon(cifs_sb); 774 775 cifs_close_all_deferred_files(tcon); 776 return 0; 777 } 778 779 #ifdef CONFIG_CIFS_STATS2 780 static int cifs_show_stats(struct seq_file *s, struct dentry *root) 781 { 782 /* BB FIXME */ 783 return 0; 784 } 785 #endif 786 787 static int cifs_write_inode(struct inode *inode, struct writeback_control *wbc) 788 { 789 fscache_unpin_writeback(wbc, cifs_inode_cookie(inode)); 790 return 0; 791 } 792 793 static int cifs_drop_inode(struct inode *inode) 794 { 795 struct cifs_sb_info *cifs_sb = CIFS_SB(inode->i_sb); 796 797 /* no serverino => unconditional eviction */ 798 return !(cifs_sb->mnt_cifs_flags & CIFS_MOUNT_SERVER_INUM) || 799 generic_drop_inode(inode); 800 } 801 802 static const struct super_operations cifs_super_ops = { 803 .statfs = cifs_statfs, 804 .alloc_inode = cifs_alloc_inode, 805 .write_inode = cifs_write_inode, 806 .free_inode = cifs_free_inode, 807 .drop_inode = cifs_drop_inode, 808 .evict_inode = cifs_evict_inode, 809 /* .show_path = cifs_show_path, */ /* Would we ever need show path? */ 810 .show_devname = cifs_show_devname, 811 /* .delete_inode = cifs_delete_inode, */ /* Do not need above 812 function unless later we add lazy close of inodes or unless the 813 kernel forgets to call us with the same number of releases (closes) 814 as opens */ 815 .show_options = cifs_show_options, 816 .umount_begin = cifs_umount_begin, 817 .freeze_fs = cifs_freeze, 818 #ifdef CONFIG_CIFS_STATS2 819 .show_stats = cifs_show_stats, 820 #endif 821 }; 822 823 /* 824 * Get root dentry from superblock according to prefix path mount option. 825 * Return dentry with refcount + 1 on success and NULL otherwise. 826 */ 827 static struct dentry * 828 cifs_get_root(struct smb3_fs_context *ctx, struct super_block *sb) 829 { 830 struct dentry *dentry; 831 struct cifs_sb_info *cifs_sb = CIFS_SB(sb); 832 char *full_path = NULL; 833 char *s, *p; 834 char sep; 835 836 if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_USE_PREFIX_PATH) 837 return dget(sb->s_root); 838 839 full_path = cifs_build_path_to_root(ctx, cifs_sb, 840 cifs_sb_master_tcon(cifs_sb), 0); 841 if (full_path == NULL) 842 return ERR_PTR(-ENOMEM); 843 844 cifs_dbg(FYI, "Get root dentry for %s\n", full_path); 845 846 sep = CIFS_DIR_SEP(cifs_sb); 847 dentry = dget(sb->s_root); 848 s = full_path; 849 850 do { 851 struct inode *dir = d_inode(dentry); 852 struct dentry *child; 853 854 if (!S_ISDIR(dir->i_mode)) { 855 dput(dentry); 856 dentry = ERR_PTR(-ENOTDIR); 857 break; 858 } 859 860 /* skip separators */ 861 while (*s == sep) 862 s++; 863 if (!*s) 864 break; 865 p = s++; 866 /* next separator */ 867 while (*s && *s != sep) 868 s++; 869 870 child = lookup_positive_unlocked(p, dentry, s - p); 871 dput(dentry); 872 dentry = child; 873 } while (!IS_ERR(dentry)); 874 kfree(full_path); 875 return dentry; 876 } 877 878 static int cifs_set_super(struct super_block *sb, void *data) 879 { 880 struct cifs_mnt_data *mnt_data = data; 881 sb->s_fs_info = mnt_data->cifs_sb; 882 return set_anon_super(sb, NULL); 883 } 884 885 struct dentry * 886 cifs_smb3_do_mount(struct file_system_type *fs_type, 887 int flags, struct smb3_fs_context *old_ctx) 888 { 889 struct cifs_mnt_data mnt_data; 890 struct cifs_sb_info *cifs_sb; 891 struct super_block *sb; 892 struct dentry *root; 893 int rc; 894 895 if (cifsFYI) { 896 cifs_dbg(FYI, "%s: devname=%s flags=0x%x\n", __func__, 897 old_ctx->source, flags); 898 } else { 899 cifs_info("Attempting to mount %s\n", old_ctx->source); 900 } 901 902 cifs_sb = kzalloc(sizeof(*cifs_sb), GFP_KERNEL); 903 if (!cifs_sb) 904 return ERR_PTR(-ENOMEM); 905 906 cifs_sb->ctx = kzalloc(sizeof(struct smb3_fs_context), GFP_KERNEL); 907 if (!cifs_sb->ctx) { 908 root = ERR_PTR(-ENOMEM); 909 goto out; 910 } 911 rc = smb3_fs_context_dup(cifs_sb->ctx, old_ctx); 912 if (rc) { 913 root = ERR_PTR(rc); 914 goto out; 915 } 916 917 rc = cifs_setup_cifs_sb(cifs_sb); 918 if (rc) { 919 root = ERR_PTR(rc); 920 goto out; 921 } 922 923 rc = cifs_mount(cifs_sb, cifs_sb->ctx); 924 if (rc) { 925 if (!(flags & SB_SILENT)) 926 cifs_dbg(VFS, "cifs_mount failed w/return code = %d\n", 927 rc); 928 root = ERR_PTR(rc); 929 goto out; 930 } 931 932 mnt_data.ctx = cifs_sb->ctx; 933 mnt_data.cifs_sb = cifs_sb; 934 mnt_data.flags = flags; 935 936 /* BB should we make this contingent on mount parm? */ 937 flags |= SB_NODIRATIME | SB_NOATIME; 938 939 sb = sget(fs_type, cifs_match_super, cifs_set_super, flags, &mnt_data); 940 if (IS_ERR(sb)) { 941 cifs_umount(cifs_sb); 942 return ERR_CAST(sb); 943 } 944 945 if (sb->s_root) { 946 cifs_dbg(FYI, "Use existing superblock\n"); 947 cifs_umount(cifs_sb); 948 cifs_sb = NULL; 949 } else { 950 rc = cifs_read_super(sb); 951 if (rc) { 952 root = ERR_PTR(rc); 953 goto out_super; 954 } 955 956 sb->s_flags |= SB_ACTIVE; 957 } 958 959 root = cifs_get_root(cifs_sb ? cifs_sb->ctx : old_ctx, sb); 960 if (IS_ERR(root)) 961 goto out_super; 962 963 if (cifs_sb) 964 cifs_sb->root = dget(root); 965 966 cifs_dbg(FYI, "dentry root is: %p\n", root); 967 return root; 968 969 out_super: 970 deactivate_locked_super(sb); 971 return root; 972 out: 973 kfree(cifs_sb->prepath); 974 smb3_cleanup_fs_context(cifs_sb->ctx); 975 kfree(cifs_sb); 976 return root; 977 } 978 979 980 static ssize_t 981 cifs_loose_read_iter(struct kiocb *iocb, struct iov_iter *iter) 982 { 983 ssize_t rc; 984 struct inode *inode = file_inode(iocb->ki_filp); 985 986 if (iocb->ki_flags & IOCB_DIRECT) 987 return cifs_user_readv(iocb, iter); 988 989 rc = cifs_revalidate_mapping(inode); 990 if (rc) 991 return rc; 992 993 return generic_file_read_iter(iocb, iter); 994 } 995 996 static ssize_t cifs_file_write_iter(struct kiocb *iocb, struct iov_iter *from) 997 { 998 struct inode *inode = file_inode(iocb->ki_filp); 999 struct cifsInodeInfo *cinode = CIFS_I(inode); 1000 ssize_t written; 1001 int rc; 1002 1003 if (iocb->ki_filp->f_flags & O_DIRECT) { 1004 written = cifs_user_writev(iocb, from); 1005 if (written > 0 && CIFS_CACHE_READ(cinode)) { 1006 cifs_zap_mapping(inode); 1007 cifs_dbg(FYI, 1008 "Set no oplock for inode=%p after a write operation\n", 1009 inode); 1010 cinode->oplock = 0; 1011 } 1012 return written; 1013 } 1014 1015 written = cifs_get_writer(cinode); 1016 if (written) 1017 return written; 1018 1019 written = generic_file_write_iter(iocb, from); 1020 1021 if (CIFS_CACHE_WRITE(CIFS_I(inode))) 1022 goto out; 1023 1024 rc = filemap_fdatawrite(inode->i_mapping); 1025 if (rc) 1026 cifs_dbg(FYI, "cifs_file_write_iter: %d rc on %p inode\n", 1027 rc, inode); 1028 1029 out: 1030 cifs_put_writer(cinode); 1031 return written; 1032 } 1033 1034 static loff_t cifs_llseek(struct file *file, loff_t offset, int whence) 1035 { 1036 struct cifsFileInfo *cfile = file->private_data; 1037 struct cifs_tcon *tcon; 1038 1039 /* 1040 * whence == SEEK_END || SEEK_DATA || SEEK_HOLE => we must revalidate 1041 * the cached file length 1042 */ 1043 if (whence != SEEK_SET && whence != SEEK_CUR) { 1044 int rc; 1045 struct inode *inode = file_inode(file); 1046 1047 /* 1048 * We need to be sure that all dirty pages are written and the 1049 * server has the newest file length. 1050 */ 1051 if (!CIFS_CACHE_READ(CIFS_I(inode)) && inode->i_mapping && 1052 inode->i_mapping->nrpages != 0) { 1053 rc = filemap_fdatawait(inode->i_mapping); 1054 if (rc) { 1055 mapping_set_error(inode->i_mapping, rc); 1056 return rc; 1057 } 1058 } 1059 /* 1060 * Some applications poll for the file length in this strange 1061 * way so we must seek to end on non-oplocked files by 1062 * setting the revalidate time to zero. 1063 */ 1064 CIFS_I(inode)->time = 0; 1065 1066 rc = cifs_revalidate_file_attr(file); 1067 if (rc < 0) 1068 return (loff_t)rc; 1069 } 1070 if (cfile && cfile->tlink) { 1071 tcon = tlink_tcon(cfile->tlink); 1072 if (tcon->ses->server->ops->llseek) 1073 return tcon->ses->server->ops->llseek(file, tcon, 1074 offset, whence); 1075 } 1076 return generic_file_llseek(file, offset, whence); 1077 } 1078 1079 static int 1080 cifs_setlease(struct file *file, long arg, struct file_lock **lease, void **priv) 1081 { 1082 /* 1083 * Note that this is called by vfs setlease with i_lock held to 1084 * protect *lease from going away. 1085 */ 1086 struct inode *inode = file_inode(file); 1087 struct cifsFileInfo *cfile = file->private_data; 1088 1089 if (!(S_ISREG(inode->i_mode))) 1090 return -EINVAL; 1091 1092 /* Check if file is oplocked if this is request for new lease */ 1093 if (arg == F_UNLCK || 1094 ((arg == F_RDLCK) && CIFS_CACHE_READ(CIFS_I(inode))) || 1095 ((arg == F_WRLCK) && CIFS_CACHE_WRITE(CIFS_I(inode)))) 1096 return generic_setlease(file, arg, lease, priv); 1097 else if (tlink_tcon(cfile->tlink)->local_lease && 1098 !CIFS_CACHE_READ(CIFS_I(inode))) 1099 /* 1100 * If the server claims to support oplock on this file, then we 1101 * still need to check oplock even if the local_lease mount 1102 * option is set, but there are servers which do not support 1103 * oplock for which this mount option may be useful if the user 1104 * knows that the file won't be changed on the server by anyone 1105 * else. 1106 */ 1107 return generic_setlease(file, arg, lease, priv); 1108 else 1109 return -EAGAIN; 1110 } 1111 1112 struct file_system_type cifs_fs_type = { 1113 .owner = THIS_MODULE, 1114 .name = "cifs", 1115 .init_fs_context = smb3_init_fs_context, 1116 .parameters = smb3_fs_parameters, 1117 .kill_sb = cifs_kill_sb, 1118 .fs_flags = FS_RENAME_DOES_D_MOVE, 1119 }; 1120 MODULE_ALIAS_FS("cifs"); 1121 1122 struct file_system_type smb3_fs_type = { 1123 .owner = THIS_MODULE, 1124 .name = "smb3", 1125 .init_fs_context = smb3_init_fs_context, 1126 .parameters = smb3_fs_parameters, 1127 .kill_sb = cifs_kill_sb, 1128 .fs_flags = FS_RENAME_DOES_D_MOVE, 1129 }; 1130 MODULE_ALIAS_FS("smb3"); 1131 MODULE_ALIAS("smb3"); 1132 1133 const struct inode_operations cifs_dir_inode_ops = { 1134 .create = cifs_create, 1135 .atomic_open = cifs_atomic_open, 1136 .lookup = cifs_lookup, 1137 .getattr = cifs_getattr, 1138 .unlink = cifs_unlink, 1139 .link = cifs_hardlink, 1140 .mkdir = cifs_mkdir, 1141 .rmdir = cifs_rmdir, 1142 .rename = cifs_rename2, 1143 .permission = cifs_permission, 1144 .setattr = cifs_setattr, 1145 .symlink = cifs_symlink, 1146 .mknod = cifs_mknod, 1147 .listxattr = cifs_listxattr, 1148 .get_acl = cifs_get_acl, 1149 .set_acl = cifs_set_acl, 1150 }; 1151 1152 const struct inode_operations cifs_file_inode_ops = { 1153 .setattr = cifs_setattr, 1154 .getattr = cifs_getattr, 1155 .permission = cifs_permission, 1156 .listxattr = cifs_listxattr, 1157 .fiemap = cifs_fiemap, 1158 .get_acl = cifs_get_acl, 1159 .set_acl = cifs_set_acl, 1160 }; 1161 1162 const char *cifs_get_link(struct dentry *dentry, struct inode *inode, 1163 struct delayed_call *done) 1164 { 1165 char *target_path; 1166 1167 target_path = kmalloc(PATH_MAX, GFP_KERNEL); 1168 if (!target_path) 1169 return ERR_PTR(-ENOMEM); 1170 1171 spin_lock(&inode->i_lock); 1172 if (likely(CIFS_I(inode)->symlink_target)) { 1173 strscpy(target_path, CIFS_I(inode)->symlink_target, PATH_MAX); 1174 } else { 1175 kfree(target_path); 1176 target_path = ERR_PTR(-EOPNOTSUPP); 1177 } 1178 spin_unlock(&inode->i_lock); 1179 1180 if (!IS_ERR(target_path)) 1181 set_delayed_call(done, kfree_link, target_path); 1182 1183 return target_path; 1184 } 1185 1186 const struct inode_operations cifs_symlink_inode_ops = { 1187 .get_link = cifs_get_link, 1188 .permission = cifs_permission, 1189 .listxattr = cifs_listxattr, 1190 }; 1191 1192 static loff_t cifs_remap_file_range(struct file *src_file, loff_t off, 1193 struct file *dst_file, loff_t destoff, loff_t len, 1194 unsigned int remap_flags) 1195 { 1196 struct inode *src_inode = file_inode(src_file); 1197 struct inode *target_inode = file_inode(dst_file); 1198 struct cifsFileInfo *smb_file_src = src_file->private_data; 1199 struct cifsFileInfo *smb_file_target; 1200 struct cifs_tcon *target_tcon; 1201 unsigned int xid; 1202 int rc; 1203 1204 if (remap_flags & ~(REMAP_FILE_DEDUP | REMAP_FILE_ADVISORY)) 1205 return -EINVAL; 1206 1207 cifs_dbg(FYI, "clone range\n"); 1208 1209 xid = get_xid(); 1210 1211 if (!src_file->private_data || !dst_file->private_data) { 1212 rc = -EBADF; 1213 cifs_dbg(VFS, "missing cifsFileInfo on copy range src file\n"); 1214 goto out; 1215 } 1216 1217 smb_file_target = dst_file->private_data; 1218 target_tcon = tlink_tcon(smb_file_target->tlink); 1219 1220 /* 1221 * Note: cifs case is easier than btrfs since server responsible for 1222 * checks for proper open modes and file type and if it wants 1223 * server could even support copy of range where source = target 1224 */ 1225 lock_two_nondirectories(target_inode, src_inode); 1226 1227 if (len == 0) 1228 len = src_inode->i_size - off; 1229 1230 cifs_dbg(FYI, "about to flush pages\n"); 1231 /* should we flush first and last page first */ 1232 truncate_inode_pages_range(&target_inode->i_data, destoff, 1233 PAGE_ALIGN(destoff + len)-1); 1234 1235 if (target_tcon->ses->server->ops->duplicate_extents) 1236 rc = target_tcon->ses->server->ops->duplicate_extents(xid, 1237 smb_file_src, smb_file_target, off, len, destoff); 1238 else 1239 rc = -EOPNOTSUPP; 1240 1241 /* force revalidate of size and timestamps of target file now 1242 that target is updated on the server */ 1243 CIFS_I(target_inode)->time = 0; 1244 /* although unlocking in the reverse order from locking is not 1245 strictly necessary here it is a little cleaner to be consistent */ 1246 unlock_two_nondirectories(src_inode, target_inode); 1247 out: 1248 free_xid(xid); 1249 return rc < 0 ? rc : len; 1250 } 1251 1252 ssize_t cifs_file_copychunk_range(unsigned int xid, 1253 struct file *src_file, loff_t off, 1254 struct file *dst_file, loff_t destoff, 1255 size_t len, unsigned int flags) 1256 { 1257 struct inode *src_inode = file_inode(src_file); 1258 struct inode *target_inode = file_inode(dst_file); 1259 struct cifsFileInfo *smb_file_src; 1260 struct cifsFileInfo *smb_file_target; 1261 struct cifs_tcon *src_tcon; 1262 struct cifs_tcon *target_tcon; 1263 ssize_t rc; 1264 1265 cifs_dbg(FYI, "copychunk range\n"); 1266 1267 if (!src_file->private_data || !dst_file->private_data) { 1268 rc = -EBADF; 1269 cifs_dbg(VFS, "missing cifsFileInfo on copy range src file\n"); 1270 goto out; 1271 } 1272 1273 rc = -EXDEV; 1274 smb_file_target = dst_file->private_data; 1275 smb_file_src = src_file->private_data; 1276 src_tcon = tlink_tcon(smb_file_src->tlink); 1277 target_tcon = tlink_tcon(smb_file_target->tlink); 1278 1279 if (src_tcon->ses != target_tcon->ses) { 1280 cifs_dbg(VFS, "source and target of copy not on same server\n"); 1281 goto out; 1282 } 1283 1284 rc = -EOPNOTSUPP; 1285 if (!target_tcon->ses->server->ops->copychunk_range) 1286 goto out; 1287 1288 /* 1289 * Note: cifs case is easier than btrfs since server responsible for 1290 * checks for proper open modes and file type and if it wants 1291 * server could even support copy of range where source = target 1292 */ 1293 lock_two_nondirectories(target_inode, src_inode); 1294 1295 cifs_dbg(FYI, "about to flush pages\n"); 1296 1297 rc = filemap_write_and_wait_range(src_inode->i_mapping, off, 1298 off + len - 1); 1299 if (rc) 1300 goto unlock; 1301 1302 /* should we flush first and last page first */ 1303 truncate_inode_pages(&target_inode->i_data, 0); 1304 1305 rc = file_modified(dst_file); 1306 if (!rc) 1307 rc = target_tcon->ses->server->ops->copychunk_range(xid, 1308 smb_file_src, smb_file_target, off, len, destoff); 1309 1310 file_accessed(src_file); 1311 1312 /* force revalidate of size and timestamps of target file now 1313 * that target is updated on the server 1314 */ 1315 CIFS_I(target_inode)->time = 0; 1316 1317 unlock: 1318 /* although unlocking in the reverse order from locking is not 1319 * strictly necessary here it is a little cleaner to be consistent 1320 */ 1321 unlock_two_nondirectories(src_inode, target_inode); 1322 1323 out: 1324 return rc; 1325 } 1326 1327 /* 1328 * Directory operations under CIFS/SMB2/SMB3 are synchronous, so fsync() 1329 * is a dummy operation. 1330 */ 1331 static int cifs_dir_fsync(struct file *file, loff_t start, loff_t end, int datasync) 1332 { 1333 cifs_dbg(FYI, "Sync directory - name: %pD datasync: 0x%x\n", 1334 file, datasync); 1335 1336 return 0; 1337 } 1338 1339 static ssize_t cifs_copy_file_range(struct file *src_file, loff_t off, 1340 struct file *dst_file, loff_t destoff, 1341 size_t len, unsigned int flags) 1342 { 1343 unsigned int xid = get_xid(); 1344 ssize_t rc; 1345 struct cifsFileInfo *cfile = dst_file->private_data; 1346 1347 if (cfile->swapfile) { 1348 rc = -EOPNOTSUPP; 1349 free_xid(xid); 1350 return rc; 1351 } 1352 1353 rc = cifs_file_copychunk_range(xid, src_file, off, dst_file, destoff, 1354 len, flags); 1355 free_xid(xid); 1356 1357 if (rc == -EOPNOTSUPP || rc == -EXDEV) 1358 rc = generic_copy_file_range(src_file, off, dst_file, 1359 destoff, len, flags); 1360 return rc; 1361 } 1362 1363 const struct file_operations cifs_file_ops = { 1364 .read_iter = cifs_loose_read_iter, 1365 .write_iter = cifs_file_write_iter, 1366 .open = cifs_open, 1367 .release = cifs_close, 1368 .lock = cifs_lock, 1369 .flock = cifs_flock, 1370 .fsync = cifs_fsync, 1371 .flush = cifs_flush, 1372 .mmap = cifs_file_mmap, 1373 .splice_read = filemap_splice_read, 1374 .splice_write = iter_file_splice_write, 1375 .llseek = cifs_llseek, 1376 .unlocked_ioctl = cifs_ioctl, 1377 .copy_file_range = cifs_copy_file_range, 1378 .remap_file_range = cifs_remap_file_range, 1379 .setlease = cifs_setlease, 1380 .fallocate = cifs_fallocate, 1381 }; 1382 1383 const struct file_operations cifs_file_strict_ops = { 1384 .read_iter = cifs_strict_readv, 1385 .write_iter = cifs_strict_writev, 1386 .open = cifs_open, 1387 .release = cifs_close, 1388 .lock = cifs_lock, 1389 .flock = cifs_flock, 1390 .fsync = cifs_strict_fsync, 1391 .flush = cifs_flush, 1392 .mmap = cifs_file_strict_mmap, 1393 .splice_read = filemap_splice_read, 1394 .splice_write = iter_file_splice_write, 1395 .llseek = cifs_llseek, 1396 .unlocked_ioctl = cifs_ioctl, 1397 .copy_file_range = cifs_copy_file_range, 1398 .remap_file_range = cifs_remap_file_range, 1399 .setlease = cifs_setlease, 1400 .fallocate = cifs_fallocate, 1401 }; 1402 1403 const struct file_operations cifs_file_direct_ops = { 1404 .read_iter = cifs_direct_readv, 1405 .write_iter = cifs_direct_writev, 1406 .open = cifs_open, 1407 .release = cifs_close, 1408 .lock = cifs_lock, 1409 .flock = cifs_flock, 1410 .fsync = cifs_fsync, 1411 .flush = cifs_flush, 1412 .mmap = cifs_file_mmap, 1413 .splice_read = copy_splice_read, 1414 .splice_write = iter_file_splice_write, 1415 .unlocked_ioctl = cifs_ioctl, 1416 .copy_file_range = cifs_copy_file_range, 1417 .remap_file_range = cifs_remap_file_range, 1418 .llseek = cifs_llseek, 1419 .setlease = cifs_setlease, 1420 .fallocate = cifs_fallocate, 1421 }; 1422 1423 const struct file_operations cifs_file_nobrl_ops = { 1424 .read_iter = cifs_loose_read_iter, 1425 .write_iter = cifs_file_write_iter, 1426 .open = cifs_open, 1427 .release = cifs_close, 1428 .fsync = cifs_fsync, 1429 .flush = cifs_flush, 1430 .mmap = cifs_file_mmap, 1431 .splice_read = filemap_splice_read, 1432 .splice_write = iter_file_splice_write, 1433 .llseek = cifs_llseek, 1434 .unlocked_ioctl = cifs_ioctl, 1435 .copy_file_range = cifs_copy_file_range, 1436 .remap_file_range = cifs_remap_file_range, 1437 .setlease = cifs_setlease, 1438 .fallocate = cifs_fallocate, 1439 }; 1440 1441 const struct file_operations cifs_file_strict_nobrl_ops = { 1442 .read_iter = cifs_strict_readv, 1443 .write_iter = cifs_strict_writev, 1444 .open = cifs_open, 1445 .release = cifs_close, 1446 .fsync = cifs_strict_fsync, 1447 .flush = cifs_flush, 1448 .mmap = cifs_file_strict_mmap, 1449 .splice_read = filemap_splice_read, 1450 .splice_write = iter_file_splice_write, 1451 .llseek = cifs_llseek, 1452 .unlocked_ioctl = cifs_ioctl, 1453 .copy_file_range = cifs_copy_file_range, 1454 .remap_file_range = cifs_remap_file_range, 1455 .setlease = cifs_setlease, 1456 .fallocate = cifs_fallocate, 1457 }; 1458 1459 const struct file_operations cifs_file_direct_nobrl_ops = { 1460 .read_iter = cifs_direct_readv, 1461 .write_iter = cifs_direct_writev, 1462 .open = cifs_open, 1463 .release = cifs_close, 1464 .fsync = cifs_fsync, 1465 .flush = cifs_flush, 1466 .mmap = cifs_file_mmap, 1467 .splice_read = copy_splice_read, 1468 .splice_write = iter_file_splice_write, 1469 .unlocked_ioctl = cifs_ioctl, 1470 .copy_file_range = cifs_copy_file_range, 1471 .remap_file_range = cifs_remap_file_range, 1472 .llseek = cifs_llseek, 1473 .setlease = cifs_setlease, 1474 .fallocate = cifs_fallocate, 1475 }; 1476 1477 const struct file_operations cifs_dir_ops = { 1478 .iterate_shared = cifs_readdir, 1479 .release = cifs_closedir, 1480 .read = generic_read_dir, 1481 .unlocked_ioctl = cifs_ioctl, 1482 .copy_file_range = cifs_copy_file_range, 1483 .remap_file_range = cifs_remap_file_range, 1484 .llseek = generic_file_llseek, 1485 .fsync = cifs_dir_fsync, 1486 }; 1487 1488 static void 1489 cifs_init_once(void *inode) 1490 { 1491 struct cifsInodeInfo *cifsi = inode; 1492 1493 inode_init_once(&cifsi->netfs.inode); 1494 init_rwsem(&cifsi->lock_sem); 1495 } 1496 1497 static int __init 1498 cifs_init_inodecache(void) 1499 { 1500 cifs_inode_cachep = kmem_cache_create("cifs_inode_cache", 1501 sizeof(struct cifsInodeInfo), 1502 0, (SLAB_RECLAIM_ACCOUNT| 1503 SLAB_MEM_SPREAD|SLAB_ACCOUNT), 1504 cifs_init_once); 1505 if (cifs_inode_cachep == NULL) 1506 return -ENOMEM; 1507 1508 return 0; 1509 } 1510 1511 static void 1512 cifs_destroy_inodecache(void) 1513 { 1514 /* 1515 * Make sure all delayed rcu free inodes are flushed before we 1516 * destroy cache. 1517 */ 1518 rcu_barrier(); 1519 kmem_cache_destroy(cifs_inode_cachep); 1520 } 1521 1522 static int 1523 cifs_init_request_bufs(void) 1524 { 1525 /* 1526 * SMB2 maximum header size is bigger than CIFS one - no problems to 1527 * allocate some more bytes for CIFS. 1528 */ 1529 size_t max_hdr_size = MAX_SMB2_HDR_SIZE; 1530 1531 if (CIFSMaxBufSize < 8192) { 1532 /* Buffer size can not be smaller than 2 * PATH_MAX since maximum 1533 Unicode path name has to fit in any SMB/CIFS path based frames */ 1534 CIFSMaxBufSize = 8192; 1535 } else if (CIFSMaxBufSize > 1024*127) { 1536 CIFSMaxBufSize = 1024 * 127; 1537 } else { 1538 CIFSMaxBufSize &= 0x1FE00; /* Round size to even 512 byte mult*/ 1539 } 1540 /* 1541 cifs_dbg(VFS, "CIFSMaxBufSize %d 0x%x\n", 1542 CIFSMaxBufSize, CIFSMaxBufSize); 1543 */ 1544 cifs_req_cachep = kmem_cache_create_usercopy("cifs_request", 1545 CIFSMaxBufSize + max_hdr_size, 0, 1546 SLAB_HWCACHE_ALIGN, 0, 1547 CIFSMaxBufSize + max_hdr_size, 1548 NULL); 1549 if (cifs_req_cachep == NULL) 1550 return -ENOMEM; 1551 1552 if (cifs_min_rcv < 1) 1553 cifs_min_rcv = 1; 1554 else if (cifs_min_rcv > 64) { 1555 cifs_min_rcv = 64; 1556 cifs_dbg(VFS, "cifs_min_rcv set to maximum (64)\n"); 1557 } 1558 1559 cifs_req_poolp = mempool_create_slab_pool(cifs_min_rcv, 1560 cifs_req_cachep); 1561 1562 if (cifs_req_poolp == NULL) { 1563 kmem_cache_destroy(cifs_req_cachep); 1564 return -ENOMEM; 1565 } 1566 /* MAX_CIFS_SMALL_BUFFER_SIZE bytes is enough for most SMB responses and 1567 almost all handle based requests (but not write response, nor is it 1568 sufficient for path based requests). A smaller size would have 1569 been more efficient (compacting multiple slab items on one 4k page) 1570 for the case in which debug was on, but this larger size allows 1571 more SMBs to use small buffer alloc and is still much more 1572 efficient to alloc 1 per page off the slab compared to 17K (5page) 1573 alloc of large cifs buffers even when page debugging is on */ 1574 cifs_sm_req_cachep = kmem_cache_create_usercopy("cifs_small_rq", 1575 MAX_CIFS_SMALL_BUFFER_SIZE, 0, SLAB_HWCACHE_ALIGN, 1576 0, MAX_CIFS_SMALL_BUFFER_SIZE, NULL); 1577 if (cifs_sm_req_cachep == NULL) { 1578 mempool_destroy(cifs_req_poolp); 1579 kmem_cache_destroy(cifs_req_cachep); 1580 return -ENOMEM; 1581 } 1582 1583 if (cifs_min_small < 2) 1584 cifs_min_small = 2; 1585 else if (cifs_min_small > 256) { 1586 cifs_min_small = 256; 1587 cifs_dbg(FYI, "cifs_min_small set to maximum (256)\n"); 1588 } 1589 1590 cifs_sm_req_poolp = mempool_create_slab_pool(cifs_min_small, 1591 cifs_sm_req_cachep); 1592 1593 if (cifs_sm_req_poolp == NULL) { 1594 mempool_destroy(cifs_req_poolp); 1595 kmem_cache_destroy(cifs_req_cachep); 1596 kmem_cache_destroy(cifs_sm_req_cachep); 1597 return -ENOMEM; 1598 } 1599 1600 return 0; 1601 } 1602 1603 static void 1604 cifs_destroy_request_bufs(void) 1605 { 1606 mempool_destroy(cifs_req_poolp); 1607 kmem_cache_destroy(cifs_req_cachep); 1608 mempool_destroy(cifs_sm_req_poolp); 1609 kmem_cache_destroy(cifs_sm_req_cachep); 1610 } 1611 1612 static int init_mids(void) 1613 { 1614 cifs_mid_cachep = kmem_cache_create("cifs_mpx_ids", 1615 sizeof(struct mid_q_entry), 0, 1616 SLAB_HWCACHE_ALIGN, NULL); 1617 if (cifs_mid_cachep == NULL) 1618 return -ENOMEM; 1619 1620 /* 3 is a reasonable minimum number of simultaneous operations */ 1621 cifs_mid_poolp = mempool_create_slab_pool(3, cifs_mid_cachep); 1622 if (cifs_mid_poolp == NULL) { 1623 kmem_cache_destroy(cifs_mid_cachep); 1624 return -ENOMEM; 1625 } 1626 1627 return 0; 1628 } 1629 1630 static void destroy_mids(void) 1631 { 1632 mempool_destroy(cifs_mid_poolp); 1633 kmem_cache_destroy(cifs_mid_cachep); 1634 } 1635 1636 static int __init 1637 init_cifs(void) 1638 { 1639 int rc = 0; 1640 cifs_proc_init(); 1641 INIT_LIST_HEAD(&cifs_tcp_ses_list); 1642 /* 1643 * Initialize Global counters 1644 */ 1645 atomic_set(&sesInfoAllocCount, 0); 1646 atomic_set(&tconInfoAllocCount, 0); 1647 atomic_set(&tcpSesNextId, 0); 1648 atomic_set(&tcpSesAllocCount, 0); 1649 atomic_set(&tcpSesReconnectCount, 0); 1650 atomic_set(&tconInfoReconnectCount, 0); 1651 1652 atomic_set(&buf_alloc_count, 0); 1653 atomic_set(&small_buf_alloc_count, 0); 1654 #ifdef CONFIG_CIFS_STATS2 1655 atomic_set(&total_buf_alloc_count, 0); 1656 atomic_set(&total_small_buf_alloc_count, 0); 1657 if (slow_rsp_threshold < 1) 1658 cifs_dbg(FYI, "slow_response_threshold msgs disabled\n"); 1659 else if (slow_rsp_threshold > 32767) 1660 cifs_dbg(VFS, 1661 "slow response threshold set higher than recommended (0 to 32767)\n"); 1662 #endif /* CONFIG_CIFS_STATS2 */ 1663 1664 atomic_set(&mid_count, 0); 1665 GlobalCurrentXid = 0; 1666 GlobalTotalActiveXid = 0; 1667 GlobalMaxActiveXid = 0; 1668 spin_lock_init(&cifs_tcp_ses_lock); 1669 spin_lock_init(&GlobalMid_Lock); 1670 1671 cifs_lock_secret = get_random_u32(); 1672 1673 if (cifs_max_pending < 2) { 1674 cifs_max_pending = 2; 1675 cifs_dbg(FYI, "cifs_max_pending set to min of 2\n"); 1676 } else if (cifs_max_pending > CIFS_MAX_REQ) { 1677 cifs_max_pending = CIFS_MAX_REQ; 1678 cifs_dbg(FYI, "cifs_max_pending set to max of %u\n", 1679 CIFS_MAX_REQ); 1680 } 1681 1682 cifsiod_wq = alloc_workqueue("cifsiod", WQ_FREEZABLE|WQ_MEM_RECLAIM, 0); 1683 if (!cifsiod_wq) { 1684 rc = -ENOMEM; 1685 goto out_clean_proc; 1686 } 1687 1688 /* 1689 * Consider in future setting limit!=0 maybe to min(num_of_cores - 1, 3) 1690 * so that we don't launch too many worker threads but 1691 * Documentation/core-api/workqueue.rst recommends setting it to 0 1692 */ 1693 1694 /* WQ_UNBOUND allows decrypt tasks to run on any CPU */ 1695 decrypt_wq = alloc_workqueue("smb3decryptd", 1696 WQ_UNBOUND|WQ_FREEZABLE|WQ_MEM_RECLAIM, 0); 1697 if (!decrypt_wq) { 1698 rc = -ENOMEM; 1699 goto out_destroy_cifsiod_wq; 1700 } 1701 1702 fileinfo_put_wq = alloc_workqueue("cifsfileinfoput", 1703 WQ_UNBOUND|WQ_FREEZABLE|WQ_MEM_RECLAIM, 0); 1704 if (!fileinfo_put_wq) { 1705 rc = -ENOMEM; 1706 goto out_destroy_decrypt_wq; 1707 } 1708 1709 cifsoplockd_wq = alloc_workqueue("cifsoplockd", 1710 WQ_FREEZABLE|WQ_MEM_RECLAIM, 0); 1711 if (!cifsoplockd_wq) { 1712 rc = -ENOMEM; 1713 goto out_destroy_fileinfo_put_wq; 1714 } 1715 1716 deferredclose_wq = alloc_workqueue("deferredclose", 1717 WQ_FREEZABLE|WQ_MEM_RECLAIM, 0); 1718 if (!deferredclose_wq) { 1719 rc = -ENOMEM; 1720 goto out_destroy_cifsoplockd_wq; 1721 } 1722 1723 rc = cifs_init_inodecache(); 1724 if (rc) 1725 goto out_destroy_deferredclose_wq; 1726 1727 rc = init_mids(); 1728 if (rc) 1729 goto out_destroy_inodecache; 1730 1731 rc = cifs_init_request_bufs(); 1732 if (rc) 1733 goto out_destroy_mids; 1734 1735 #ifdef CONFIG_CIFS_DFS_UPCALL 1736 rc = dfs_cache_init(); 1737 if (rc) 1738 goto out_destroy_request_bufs; 1739 #endif /* CONFIG_CIFS_DFS_UPCALL */ 1740 #ifdef CONFIG_CIFS_UPCALL 1741 rc = init_cifs_spnego(); 1742 if (rc) 1743 goto out_destroy_dfs_cache; 1744 #endif /* CONFIG_CIFS_UPCALL */ 1745 #ifdef CONFIG_CIFS_SWN_UPCALL 1746 rc = cifs_genl_init(); 1747 if (rc) 1748 goto out_register_key_type; 1749 #endif /* CONFIG_CIFS_SWN_UPCALL */ 1750 1751 rc = init_cifs_idmap(); 1752 if (rc) 1753 goto out_cifs_swn_init; 1754 1755 rc = register_filesystem(&cifs_fs_type); 1756 if (rc) 1757 goto out_init_cifs_idmap; 1758 1759 rc = register_filesystem(&smb3_fs_type); 1760 if (rc) { 1761 unregister_filesystem(&cifs_fs_type); 1762 goto out_init_cifs_idmap; 1763 } 1764 1765 return 0; 1766 1767 out_init_cifs_idmap: 1768 exit_cifs_idmap(); 1769 out_cifs_swn_init: 1770 #ifdef CONFIG_CIFS_SWN_UPCALL 1771 cifs_genl_exit(); 1772 out_register_key_type: 1773 #endif 1774 #ifdef CONFIG_CIFS_UPCALL 1775 exit_cifs_spnego(); 1776 out_destroy_dfs_cache: 1777 #endif 1778 #ifdef CONFIG_CIFS_DFS_UPCALL 1779 dfs_cache_destroy(); 1780 out_destroy_request_bufs: 1781 #endif 1782 cifs_destroy_request_bufs(); 1783 out_destroy_mids: 1784 destroy_mids(); 1785 out_destroy_inodecache: 1786 cifs_destroy_inodecache(); 1787 out_destroy_deferredclose_wq: 1788 destroy_workqueue(deferredclose_wq); 1789 out_destroy_cifsoplockd_wq: 1790 destroy_workqueue(cifsoplockd_wq); 1791 out_destroy_fileinfo_put_wq: 1792 destroy_workqueue(fileinfo_put_wq); 1793 out_destroy_decrypt_wq: 1794 destroy_workqueue(decrypt_wq); 1795 out_destroy_cifsiod_wq: 1796 destroy_workqueue(cifsiod_wq); 1797 out_clean_proc: 1798 cifs_proc_clean(); 1799 return rc; 1800 } 1801 1802 static void __exit 1803 exit_cifs(void) 1804 { 1805 cifs_dbg(NOISY, "exit_smb3\n"); 1806 unregister_filesystem(&cifs_fs_type); 1807 unregister_filesystem(&smb3_fs_type); 1808 cifs_dfs_release_automount_timer(); 1809 exit_cifs_idmap(); 1810 #ifdef CONFIG_CIFS_SWN_UPCALL 1811 cifs_genl_exit(); 1812 #endif 1813 #ifdef CONFIG_CIFS_UPCALL 1814 exit_cifs_spnego(); 1815 #endif 1816 #ifdef CONFIG_CIFS_DFS_UPCALL 1817 dfs_cache_destroy(); 1818 #endif 1819 cifs_destroy_request_bufs(); 1820 destroy_mids(); 1821 cifs_destroy_inodecache(); 1822 destroy_workqueue(deferredclose_wq); 1823 destroy_workqueue(cifsoplockd_wq); 1824 destroy_workqueue(decrypt_wq); 1825 destroy_workqueue(fileinfo_put_wq); 1826 destroy_workqueue(cifsiod_wq); 1827 cifs_proc_clean(); 1828 } 1829 1830 MODULE_AUTHOR("Steve French"); 1831 MODULE_LICENSE("GPL"); /* combination of LGPL + GPL source behaves as GPL */ 1832 MODULE_DESCRIPTION 1833 ("VFS to access SMB3 servers e.g. Samba, Macs, Azure and Windows (and " 1834 "also older servers complying with the SNIA CIFS Specification)"); 1835 MODULE_VERSION(CIFS_VERSION); 1836 MODULE_SOFTDEP("ecb"); 1837 MODULE_SOFTDEP("hmac"); 1838 MODULE_SOFTDEP("md5"); 1839 MODULE_SOFTDEP("nls"); 1840 MODULE_SOFTDEP("aes"); 1841 MODULE_SOFTDEP("cmac"); 1842 MODULE_SOFTDEP("sha256"); 1843 MODULE_SOFTDEP("sha512"); 1844 MODULE_SOFTDEP("aead2"); 1845 MODULE_SOFTDEP("ccm"); 1846 MODULE_SOFTDEP("gcm"); 1847 module_init(init_cifs) 1848 module_exit(exit_cifs) 1849