1 // SPDX-License-Identifier: LGPL-2.1 2 /* 3 * 4 * Copyright (C) International Business Machines Corp., 2002,2011 5 * Author(s): Steve French (sfrench@us.ibm.com) 6 * 7 */ 8 #include <linux/fs.h> 9 #include <linux/net.h> 10 #include <linux/string.h> 11 #include <linux/sched/mm.h> 12 #include <linux/sched/signal.h> 13 #include <linux/list.h> 14 #include <linux/wait.h> 15 #include <linux/slab.h> 16 #include <linux/pagemap.h> 17 #include <linux/ctype.h> 18 #include <linux/utsname.h> 19 #include <linux/mempool.h> 20 #include <linux/delay.h> 21 #include <linux/completion.h> 22 #include <linux/kthread.h> 23 #include <linux/pagevec.h> 24 #include <linux/freezer.h> 25 #include <linux/namei.h> 26 #include <linux/uuid.h> 27 #include <linux/uaccess.h> 28 #include <asm/processor.h> 29 #include <linux/inet.h> 30 #include <linux/module.h> 31 #include <keys/user-type.h> 32 #include <net/ipv6.h> 33 #include <linux/parser.h> 34 #include <linux/bvec.h> 35 #include "cifspdu.h" 36 #include "cifsglob.h" 37 #include "cifsproto.h" 38 #include "cifs_unicode.h" 39 #include "cifs_debug.h" 40 #include "cifs_fs_sb.h" 41 #include "ntlmssp.h" 42 #include "nterr.h" 43 #include "rfc1002pdu.h" 44 #include "fscache.h" 45 #include "smb2proto.h" 46 #include "smbdirect.h" 47 #include "dns_resolve.h" 48 #ifdef CONFIG_CIFS_DFS_UPCALL 49 #include "dfs.h" 50 #include "dfs_cache.h" 51 #endif 52 #include "fs_context.h" 53 #include "cifs_swn.h" 54 55 /* FIXME: should these be tunable? */ 56 #define TLINK_ERROR_EXPIRE (1 * HZ) 57 #define TLINK_IDLE_EXPIRE (600 * HZ) 58 59 /* Drop the connection to not overload the server */ 60 #define MAX_STATUS_IO_TIMEOUT 5 61 62 static int ip_connect(struct TCP_Server_Info *server); 63 static int generic_ip_connect(struct TCP_Server_Info *server); 64 static void tlink_rb_insert(struct rb_root *root, struct tcon_link *new_tlink); 65 static void cifs_prune_tlinks(struct work_struct *work); 66 67 /* 68 * Resolve hostname and set ip addr in tcp ses. Useful for hostnames that may 69 * get their ip addresses changed at some point. 70 * 71 * This should be called with server->srv_mutex held. 72 */ 73 static int reconn_set_ipaddr_from_hostname(struct TCP_Server_Info *server) 74 { 75 int rc; 76 int len; 77 char *unc; 78 struct sockaddr_storage ss; 79 80 if (!server->hostname) 81 return -EINVAL; 82 83 /* if server hostname isn't populated, there's nothing to do here */ 84 if (server->hostname[0] == '\0') 85 return 0; 86 87 len = strlen(server->hostname) + 3; 88 89 unc = kmalloc(len, GFP_KERNEL); 90 if (!unc) { 91 cifs_dbg(FYI, "%s: failed to create UNC path\n", __func__); 92 return -ENOMEM; 93 } 94 scnprintf(unc, len, "\\\\%s", server->hostname); 95 96 spin_lock(&server->srv_lock); 97 ss = server->dstaddr; 98 spin_unlock(&server->srv_lock); 99 100 rc = dns_resolve_server_name_to_ip(unc, (struct sockaddr *)&ss, NULL); 101 kfree(unc); 102 103 if (rc < 0) { 104 cifs_dbg(FYI, "%s: failed to resolve server part of %s to IP: %d\n", 105 __func__, server->hostname, rc); 106 } else { 107 spin_lock(&server->srv_lock); 108 memcpy(&server->dstaddr, &ss, sizeof(server->dstaddr)); 109 spin_unlock(&server->srv_lock); 110 rc = 0; 111 } 112 113 return rc; 114 } 115 116 static void smb2_query_server_interfaces(struct work_struct *work) 117 { 118 int rc; 119 int xid; 120 struct cifs_tcon *tcon = container_of(work, 121 struct cifs_tcon, 122 query_interfaces.work); 123 struct TCP_Server_Info *server = tcon->ses->server; 124 125 /* 126 * query server network interfaces, in case they change 127 */ 128 if (!server->ops->query_server_interfaces) 129 return; 130 131 xid = get_xid(); 132 rc = server->ops->query_server_interfaces(xid, tcon, false); 133 free_xid(xid); 134 135 if (rc) { 136 if (rc == -EOPNOTSUPP) 137 return; 138 139 cifs_dbg(FYI, "%s: failed to query server interfaces: %d\n", 140 __func__, rc); 141 } 142 143 queue_delayed_work(cifsiod_wq, &tcon->query_interfaces, 144 (SMB_INTERFACE_POLL_INTERVAL * HZ)); 145 } 146 147 /* 148 * Update the tcpStatus for the server. 149 * This is used to signal the cifsd thread to call cifs_reconnect 150 * ONLY cifsd thread should call cifs_reconnect. For any other 151 * thread, use this function 152 * 153 * @server: the tcp ses for which reconnect is needed 154 * @all_channels: if this needs to be done for all channels 155 */ 156 void 157 cifs_signal_cifsd_for_reconnect(struct TCP_Server_Info *server, 158 bool all_channels) 159 { 160 struct TCP_Server_Info *pserver; 161 struct cifs_ses *ses; 162 int i; 163 164 /* If server is a channel, select the primary channel */ 165 pserver = SERVER_IS_CHAN(server) ? server->primary_server : server; 166 167 /* if we need to signal just this channel */ 168 if (!all_channels) { 169 spin_lock(&server->srv_lock); 170 if (server->tcpStatus != CifsExiting) 171 server->tcpStatus = CifsNeedReconnect; 172 spin_unlock(&server->srv_lock); 173 return; 174 } 175 176 spin_lock(&cifs_tcp_ses_lock); 177 list_for_each_entry(ses, &pserver->smb_ses_list, smb_ses_list) { 178 if (cifs_ses_exiting(ses)) 179 continue; 180 spin_lock(&ses->chan_lock); 181 for (i = 0; i < ses->chan_count; i++) { 182 if (!ses->chans[i].server) 183 continue; 184 185 spin_lock(&ses->chans[i].server->srv_lock); 186 if (ses->chans[i].server->tcpStatus != CifsExiting) 187 ses->chans[i].server->tcpStatus = CifsNeedReconnect; 188 spin_unlock(&ses->chans[i].server->srv_lock); 189 } 190 spin_unlock(&ses->chan_lock); 191 } 192 spin_unlock(&cifs_tcp_ses_lock); 193 } 194 195 /* 196 * Mark all sessions and tcons for reconnect. 197 * IMPORTANT: make sure that this gets called only from 198 * cifsd thread. For any other thread, use 199 * cifs_signal_cifsd_for_reconnect 200 * 201 * @server: the tcp ses for which reconnect is needed 202 * @server needs to be previously set to CifsNeedReconnect. 203 * @mark_smb_session: whether even sessions need to be marked 204 */ 205 void 206 cifs_mark_tcp_ses_conns_for_reconnect(struct TCP_Server_Info *server, 207 bool mark_smb_session) 208 { 209 struct TCP_Server_Info *pserver; 210 struct cifs_ses *ses, *nses; 211 struct cifs_tcon *tcon; 212 213 /* 214 * before reconnecting the tcp session, mark the smb session (uid) and the tid bad so they 215 * are not used until reconnected. 216 */ 217 cifs_dbg(FYI, "%s: marking necessary sessions and tcons for reconnect\n", __func__); 218 219 /* If server is a channel, select the primary channel */ 220 pserver = SERVER_IS_CHAN(server) ? server->primary_server : server; 221 222 /* 223 * if the server has been marked for termination, there is a 224 * chance that the remaining channels all need reconnect. To be 225 * on the safer side, mark the session and trees for reconnect 226 * for this scenario. This might cause a few redundant session 227 * setup and tree connect requests, but it is better than not doing 228 * a tree connect when needed, and all following requests failing 229 */ 230 if (server->terminate) { 231 mark_smb_session = true; 232 server = pserver; 233 } 234 235 spin_lock(&cifs_tcp_ses_lock); 236 list_for_each_entry_safe(ses, nses, &pserver->smb_ses_list, smb_ses_list) { 237 spin_lock(&ses->ses_lock); 238 if (ses->ses_status == SES_EXITING) { 239 spin_unlock(&ses->ses_lock); 240 continue; 241 } 242 spin_unlock(&ses->ses_lock); 243 244 spin_lock(&ses->chan_lock); 245 if (cifs_ses_get_chan_index(ses, server) == 246 CIFS_INVAL_CHAN_INDEX) { 247 spin_unlock(&ses->chan_lock); 248 continue; 249 } 250 251 if (!cifs_chan_is_iface_active(ses, server)) { 252 spin_unlock(&ses->chan_lock); 253 cifs_chan_update_iface(ses, server); 254 spin_lock(&ses->chan_lock); 255 } 256 257 if (!mark_smb_session && cifs_chan_needs_reconnect(ses, server)) { 258 spin_unlock(&ses->chan_lock); 259 continue; 260 } 261 262 if (mark_smb_session) 263 CIFS_SET_ALL_CHANS_NEED_RECONNECT(ses); 264 else 265 cifs_chan_set_need_reconnect(ses, server); 266 267 cifs_dbg(FYI, "%s: channel connect bitmap: 0x%lx\n", 268 __func__, ses->chans_need_reconnect); 269 270 /* If all channels need reconnect, then tcon needs reconnect */ 271 if (!mark_smb_session && !CIFS_ALL_CHANS_NEED_RECONNECT(ses)) { 272 spin_unlock(&ses->chan_lock); 273 continue; 274 } 275 spin_unlock(&ses->chan_lock); 276 277 spin_lock(&ses->ses_lock); 278 ses->ses_status = SES_NEED_RECON; 279 spin_unlock(&ses->ses_lock); 280 281 list_for_each_entry(tcon, &ses->tcon_list, tcon_list) { 282 tcon->need_reconnect = true; 283 spin_lock(&tcon->tc_lock); 284 tcon->status = TID_NEED_RECON; 285 spin_unlock(&tcon->tc_lock); 286 287 cancel_delayed_work(&tcon->query_interfaces); 288 } 289 if (ses->tcon_ipc) { 290 ses->tcon_ipc->need_reconnect = true; 291 spin_lock(&ses->tcon_ipc->tc_lock); 292 ses->tcon_ipc->status = TID_NEED_RECON; 293 spin_unlock(&ses->tcon_ipc->tc_lock); 294 } 295 } 296 spin_unlock(&cifs_tcp_ses_lock); 297 } 298 299 static void 300 cifs_abort_connection(struct TCP_Server_Info *server) 301 { 302 struct mid_q_entry *mid, *nmid; 303 struct list_head retry_list; 304 305 server->maxBuf = 0; 306 server->max_read = 0; 307 308 /* do not want to be sending data on a socket we are freeing */ 309 cifs_dbg(FYI, "%s: tearing down socket\n", __func__); 310 cifs_server_lock(server); 311 if (server->ssocket) { 312 cifs_dbg(FYI, "State: 0x%x Flags: 0x%lx\n", server->ssocket->state, 313 server->ssocket->flags); 314 kernel_sock_shutdown(server->ssocket, SHUT_WR); 315 cifs_dbg(FYI, "Post shutdown state: 0x%x Flags: 0x%lx\n", server->ssocket->state, 316 server->ssocket->flags); 317 sock_release(server->ssocket); 318 server->ssocket = NULL; 319 } 320 server->sequence_number = 0; 321 server->session_estab = false; 322 kfree_sensitive(server->session_key.response); 323 server->session_key.response = NULL; 324 server->session_key.len = 0; 325 server->lstrp = jiffies; 326 327 /* mark submitted MIDs for retry and issue callback */ 328 INIT_LIST_HEAD(&retry_list); 329 cifs_dbg(FYI, "%s: moving mids to private list\n", __func__); 330 spin_lock(&server->mid_lock); 331 list_for_each_entry_safe(mid, nmid, &server->pending_mid_q, qhead) { 332 kref_get(&mid->refcount); 333 if (mid->mid_state == MID_REQUEST_SUBMITTED) 334 mid->mid_state = MID_RETRY_NEEDED; 335 list_move(&mid->qhead, &retry_list); 336 mid->mid_flags |= MID_DELETED; 337 } 338 spin_unlock(&server->mid_lock); 339 cifs_server_unlock(server); 340 341 cifs_dbg(FYI, "%s: issuing mid callbacks\n", __func__); 342 list_for_each_entry_safe(mid, nmid, &retry_list, qhead) { 343 list_del_init(&mid->qhead); 344 mid->callback(mid); 345 release_mid(mid); 346 } 347 348 if (cifs_rdma_enabled(server)) { 349 cifs_server_lock(server); 350 smbd_destroy(server); 351 cifs_server_unlock(server); 352 } 353 } 354 355 static bool cifs_tcp_ses_needs_reconnect(struct TCP_Server_Info *server, int num_targets) 356 { 357 spin_lock(&server->srv_lock); 358 server->nr_targets = num_targets; 359 if (server->tcpStatus == CifsExiting) { 360 /* the demux thread will exit normally next time through the loop */ 361 spin_unlock(&server->srv_lock); 362 wake_up(&server->response_q); 363 return false; 364 } 365 366 cifs_dbg(FYI, "Mark tcp session as need reconnect\n"); 367 trace_smb3_reconnect(server->CurrentMid, server->conn_id, 368 server->hostname); 369 server->tcpStatus = CifsNeedReconnect; 370 371 spin_unlock(&server->srv_lock); 372 return true; 373 } 374 375 /* 376 * cifs tcp session reconnection 377 * 378 * mark tcp session as reconnecting so temporarily locked 379 * mark all smb sessions as reconnecting for tcp session 380 * reconnect tcp session 381 * wake up waiters on reconnection? - (not needed currently) 382 * 383 * if mark_smb_session is passed as true, unconditionally mark 384 * the smb session (and tcon) for reconnect as well. This value 385 * doesn't really matter for non-multichannel scenario. 386 * 387 */ 388 static int __cifs_reconnect(struct TCP_Server_Info *server, 389 bool mark_smb_session) 390 { 391 int rc = 0; 392 393 if (!cifs_tcp_ses_needs_reconnect(server, 1)) 394 return 0; 395 396 cifs_mark_tcp_ses_conns_for_reconnect(server, mark_smb_session); 397 398 cifs_abort_connection(server); 399 400 do { 401 try_to_freeze(); 402 cifs_server_lock(server); 403 404 if (!cifs_swn_set_server_dstaddr(server)) { 405 /* resolve the hostname again to make sure that IP address is up-to-date */ 406 rc = reconn_set_ipaddr_from_hostname(server); 407 cifs_dbg(FYI, "%s: reconn_set_ipaddr_from_hostname: rc=%d\n", __func__, rc); 408 } 409 410 if (cifs_rdma_enabled(server)) 411 rc = smbd_reconnect(server); 412 else 413 rc = generic_ip_connect(server); 414 if (rc) { 415 cifs_server_unlock(server); 416 cifs_dbg(FYI, "%s: reconnect error %d\n", __func__, rc); 417 msleep(3000); 418 } else { 419 atomic_inc(&tcpSesReconnectCount); 420 set_credits(server, 1); 421 spin_lock(&server->srv_lock); 422 if (server->tcpStatus != CifsExiting) 423 server->tcpStatus = CifsNeedNegotiate; 424 spin_unlock(&server->srv_lock); 425 cifs_swn_reset_server_dstaddr(server); 426 cifs_server_unlock(server); 427 mod_delayed_work(cifsiod_wq, &server->reconnect, 0); 428 } 429 } while (server->tcpStatus == CifsNeedReconnect); 430 431 spin_lock(&server->srv_lock); 432 if (server->tcpStatus == CifsNeedNegotiate) 433 mod_delayed_work(cifsiod_wq, &server->echo, 0); 434 spin_unlock(&server->srv_lock); 435 436 wake_up(&server->response_q); 437 return rc; 438 } 439 440 #ifdef CONFIG_CIFS_DFS_UPCALL 441 static int __reconnect_target_unlocked(struct TCP_Server_Info *server, const char *target) 442 { 443 int rc; 444 char *hostname; 445 446 if (!cifs_swn_set_server_dstaddr(server)) { 447 if (server->hostname != target) { 448 hostname = extract_hostname(target); 449 if (!IS_ERR(hostname)) { 450 spin_lock(&server->srv_lock); 451 kfree(server->hostname); 452 server->hostname = hostname; 453 spin_unlock(&server->srv_lock); 454 } else { 455 cifs_dbg(FYI, "%s: couldn't extract hostname or address from dfs target: %ld\n", 456 __func__, PTR_ERR(hostname)); 457 cifs_dbg(FYI, "%s: default to last target server: %s\n", __func__, 458 server->hostname); 459 } 460 } 461 /* resolve the hostname again to make sure that IP address is up-to-date. */ 462 rc = reconn_set_ipaddr_from_hostname(server); 463 cifs_dbg(FYI, "%s: reconn_set_ipaddr_from_hostname: rc=%d\n", __func__, rc); 464 } 465 /* Reconnect the socket */ 466 if (cifs_rdma_enabled(server)) 467 rc = smbd_reconnect(server); 468 else 469 rc = generic_ip_connect(server); 470 471 return rc; 472 } 473 474 static int reconnect_target_unlocked(struct TCP_Server_Info *server, struct dfs_cache_tgt_list *tl, 475 struct dfs_cache_tgt_iterator **target_hint) 476 { 477 int rc; 478 struct dfs_cache_tgt_iterator *tit; 479 480 *target_hint = NULL; 481 482 /* If dfs target list is empty, then reconnect to last server */ 483 tit = dfs_cache_get_tgt_iterator(tl); 484 if (!tit) 485 return __reconnect_target_unlocked(server, server->hostname); 486 487 /* Otherwise, try every dfs target in @tl */ 488 for (; tit; tit = dfs_cache_get_next_tgt(tl, tit)) { 489 rc = __reconnect_target_unlocked(server, dfs_cache_get_tgt_name(tit)); 490 if (!rc) { 491 *target_hint = tit; 492 break; 493 } 494 } 495 return rc; 496 } 497 498 static int reconnect_dfs_server(struct TCP_Server_Info *server) 499 { 500 struct dfs_cache_tgt_iterator *target_hint = NULL; 501 502 DFS_CACHE_TGT_LIST(tl); 503 int num_targets = 0; 504 int rc = 0; 505 506 /* 507 * Determine the number of dfs targets the referral path in @cifs_sb resolves to. 508 * 509 * smb2_reconnect() needs to know how long it should wait based upon the number of dfs 510 * targets (server->nr_targets). It's also possible that the cached referral was cleared 511 * through /proc/fs/cifs/dfscache or the target list is empty due to server settings after 512 * refreshing the referral, so, in this case, default it to 1. 513 */ 514 mutex_lock(&server->refpath_lock); 515 if (!dfs_cache_noreq_find(server->leaf_fullpath + 1, NULL, &tl)) 516 num_targets = dfs_cache_get_nr_tgts(&tl); 517 mutex_unlock(&server->refpath_lock); 518 if (!num_targets) 519 num_targets = 1; 520 521 if (!cifs_tcp_ses_needs_reconnect(server, num_targets)) 522 return 0; 523 524 /* 525 * Unconditionally mark all sessions & tcons for reconnect as we might be connecting to a 526 * different server or share during failover. It could be improved by adding some logic to 527 * only do that in case it connects to a different server or share, though. 528 */ 529 cifs_mark_tcp_ses_conns_for_reconnect(server, true); 530 531 cifs_abort_connection(server); 532 533 do { 534 try_to_freeze(); 535 cifs_server_lock(server); 536 537 rc = reconnect_target_unlocked(server, &tl, &target_hint); 538 if (rc) { 539 /* Failed to reconnect socket */ 540 cifs_server_unlock(server); 541 cifs_dbg(FYI, "%s: reconnect error %d\n", __func__, rc); 542 msleep(3000); 543 continue; 544 } 545 /* 546 * Socket was created. Update tcp session status to CifsNeedNegotiate so that a 547 * process waiting for reconnect will know it needs to re-establish session and tcon 548 * through the reconnected target server. 549 */ 550 atomic_inc(&tcpSesReconnectCount); 551 set_credits(server, 1); 552 spin_lock(&server->srv_lock); 553 if (server->tcpStatus != CifsExiting) 554 server->tcpStatus = CifsNeedNegotiate; 555 spin_unlock(&server->srv_lock); 556 cifs_swn_reset_server_dstaddr(server); 557 cifs_server_unlock(server); 558 mod_delayed_work(cifsiod_wq, &server->reconnect, 0); 559 } while (server->tcpStatus == CifsNeedReconnect); 560 561 mutex_lock(&server->refpath_lock); 562 dfs_cache_noreq_update_tgthint(server->leaf_fullpath + 1, target_hint); 563 mutex_unlock(&server->refpath_lock); 564 dfs_cache_free_tgts(&tl); 565 566 /* Need to set up echo worker again once connection has been established */ 567 spin_lock(&server->srv_lock); 568 if (server->tcpStatus == CifsNeedNegotiate) 569 mod_delayed_work(cifsiod_wq, &server->echo, 0); 570 spin_unlock(&server->srv_lock); 571 572 wake_up(&server->response_q); 573 return rc; 574 } 575 576 int cifs_reconnect(struct TCP_Server_Info *server, bool mark_smb_session) 577 { 578 mutex_lock(&server->refpath_lock); 579 if (!server->leaf_fullpath) { 580 mutex_unlock(&server->refpath_lock); 581 return __cifs_reconnect(server, mark_smb_session); 582 } 583 mutex_unlock(&server->refpath_lock); 584 585 return reconnect_dfs_server(server); 586 } 587 #else 588 int cifs_reconnect(struct TCP_Server_Info *server, bool mark_smb_session) 589 { 590 return __cifs_reconnect(server, mark_smb_session); 591 } 592 #endif 593 594 static void 595 cifs_echo_request(struct work_struct *work) 596 { 597 int rc; 598 struct TCP_Server_Info *server = container_of(work, 599 struct TCP_Server_Info, echo.work); 600 601 /* 602 * We cannot send an echo if it is disabled. 603 * Also, no need to ping if we got a response recently. 604 */ 605 606 if (server->tcpStatus == CifsNeedReconnect || 607 server->tcpStatus == CifsExiting || 608 server->tcpStatus == CifsNew || 609 (server->ops->can_echo && !server->ops->can_echo(server)) || 610 time_before(jiffies, server->lstrp + server->echo_interval - HZ)) 611 goto requeue_echo; 612 613 rc = server->ops->echo ? server->ops->echo(server) : -ENOSYS; 614 cifs_server_dbg(FYI, "send echo request: rc = %d\n", rc); 615 616 /* Check witness registrations */ 617 cifs_swn_check(); 618 619 requeue_echo: 620 queue_delayed_work(cifsiod_wq, &server->echo, server->echo_interval); 621 } 622 623 static bool 624 allocate_buffers(struct TCP_Server_Info *server) 625 { 626 if (!server->bigbuf) { 627 server->bigbuf = (char *)cifs_buf_get(); 628 if (!server->bigbuf) { 629 cifs_server_dbg(VFS, "No memory for large SMB response\n"); 630 msleep(3000); 631 /* retry will check if exiting */ 632 return false; 633 } 634 } else if (server->large_buf) { 635 /* we are reusing a dirty large buf, clear its start */ 636 memset(server->bigbuf, 0, HEADER_SIZE(server)); 637 } 638 639 if (!server->smallbuf) { 640 server->smallbuf = (char *)cifs_small_buf_get(); 641 if (!server->smallbuf) { 642 cifs_server_dbg(VFS, "No memory for SMB response\n"); 643 msleep(1000); 644 /* retry will check if exiting */ 645 return false; 646 } 647 /* beginning of smb buffer is cleared in our buf_get */ 648 } else { 649 /* if existing small buf clear beginning */ 650 memset(server->smallbuf, 0, HEADER_SIZE(server)); 651 } 652 653 return true; 654 } 655 656 static bool 657 server_unresponsive(struct TCP_Server_Info *server) 658 { 659 /* 660 * If we're in the process of mounting a share or reconnecting a session 661 * and the server abruptly shut down (e.g. socket wasn't closed, packet 662 * had been ACK'ed but no SMB response), don't wait longer than 20s to 663 * negotiate protocol. 664 */ 665 spin_lock(&server->srv_lock); 666 if (server->tcpStatus == CifsInNegotiate && 667 time_after(jiffies, server->lstrp + 20 * HZ)) { 668 spin_unlock(&server->srv_lock); 669 cifs_reconnect(server, false); 670 return true; 671 } 672 /* 673 * We need to wait 3 echo intervals to make sure we handle such 674 * situations right: 675 * 1s client sends a normal SMB request 676 * 2s client gets a response 677 * 30s echo workqueue job pops, and decides we got a response recently 678 * and don't need to send another 679 * ... 680 * 65s kernel_recvmsg times out, and we see that we haven't gotten 681 * a response in >60s. 682 */ 683 if ((server->tcpStatus == CifsGood || 684 server->tcpStatus == CifsNeedNegotiate) && 685 (!server->ops->can_echo || server->ops->can_echo(server)) && 686 time_after(jiffies, server->lstrp + 3 * server->echo_interval)) { 687 spin_unlock(&server->srv_lock); 688 cifs_server_dbg(VFS, "has not responded in %lu seconds. Reconnecting...\n", 689 (3 * server->echo_interval) / HZ); 690 cifs_reconnect(server, false); 691 return true; 692 } 693 spin_unlock(&server->srv_lock); 694 695 return false; 696 } 697 698 static inline bool 699 zero_credits(struct TCP_Server_Info *server) 700 { 701 int val; 702 703 spin_lock(&server->req_lock); 704 val = server->credits + server->echo_credits + server->oplock_credits; 705 if (server->in_flight == 0 && val == 0) { 706 spin_unlock(&server->req_lock); 707 return true; 708 } 709 spin_unlock(&server->req_lock); 710 return false; 711 } 712 713 static int 714 cifs_readv_from_socket(struct TCP_Server_Info *server, struct msghdr *smb_msg) 715 { 716 int length = 0; 717 int total_read; 718 719 for (total_read = 0; msg_data_left(smb_msg); total_read += length) { 720 try_to_freeze(); 721 722 /* reconnect if no credits and no requests in flight */ 723 if (zero_credits(server)) { 724 cifs_reconnect(server, false); 725 return -ECONNABORTED; 726 } 727 728 if (server_unresponsive(server)) 729 return -ECONNABORTED; 730 if (cifs_rdma_enabled(server) && server->smbd_conn) 731 length = smbd_recv(server->smbd_conn, smb_msg); 732 else 733 length = sock_recvmsg(server->ssocket, smb_msg, 0); 734 735 spin_lock(&server->srv_lock); 736 if (server->tcpStatus == CifsExiting) { 737 spin_unlock(&server->srv_lock); 738 return -ESHUTDOWN; 739 } 740 741 if (server->tcpStatus == CifsNeedReconnect) { 742 spin_unlock(&server->srv_lock); 743 cifs_reconnect(server, false); 744 return -ECONNABORTED; 745 } 746 spin_unlock(&server->srv_lock); 747 748 if (length == -ERESTARTSYS || 749 length == -EAGAIN || 750 length == -EINTR) { 751 /* 752 * Minimum sleep to prevent looping, allowing socket 753 * to clear and app threads to set tcpStatus 754 * CifsNeedReconnect if server hung. 755 */ 756 usleep_range(1000, 2000); 757 length = 0; 758 continue; 759 } 760 761 if (length <= 0) { 762 cifs_dbg(FYI, "Received no data or error: %d\n", length); 763 cifs_reconnect(server, false); 764 return -ECONNABORTED; 765 } 766 } 767 return total_read; 768 } 769 770 int 771 cifs_read_from_socket(struct TCP_Server_Info *server, char *buf, 772 unsigned int to_read) 773 { 774 struct msghdr smb_msg = {}; 775 struct kvec iov = {.iov_base = buf, .iov_len = to_read}; 776 777 iov_iter_kvec(&smb_msg.msg_iter, ITER_DEST, &iov, 1, to_read); 778 779 return cifs_readv_from_socket(server, &smb_msg); 780 } 781 782 ssize_t 783 cifs_discard_from_socket(struct TCP_Server_Info *server, size_t to_read) 784 { 785 struct msghdr smb_msg = {}; 786 787 /* 788 * iov_iter_discard already sets smb_msg.type and count and iov_offset 789 * and cifs_readv_from_socket sets msg_control and msg_controllen 790 * so little to initialize in struct msghdr 791 */ 792 iov_iter_discard(&smb_msg.msg_iter, ITER_DEST, to_read); 793 794 return cifs_readv_from_socket(server, &smb_msg); 795 } 796 797 int 798 cifs_read_page_from_socket(struct TCP_Server_Info *server, struct page *page, 799 unsigned int page_offset, unsigned int to_read) 800 { 801 struct msghdr smb_msg = {}; 802 struct bio_vec bv; 803 804 bvec_set_page(&bv, page, to_read, page_offset); 805 iov_iter_bvec(&smb_msg.msg_iter, ITER_DEST, &bv, 1, to_read); 806 return cifs_readv_from_socket(server, &smb_msg); 807 } 808 809 int 810 cifs_read_iter_from_socket(struct TCP_Server_Info *server, struct iov_iter *iter, 811 unsigned int to_read) 812 { 813 struct msghdr smb_msg = { .msg_iter = *iter }; 814 int ret; 815 816 iov_iter_truncate(&smb_msg.msg_iter, to_read); 817 ret = cifs_readv_from_socket(server, &smb_msg); 818 if (ret > 0) 819 iov_iter_advance(iter, ret); 820 return ret; 821 } 822 823 static bool 824 is_smb_response(struct TCP_Server_Info *server, unsigned char type) 825 { 826 /* 827 * The first byte big endian of the length field, 828 * is actually not part of the length but the type 829 * with the most common, zero, as regular data. 830 */ 831 switch (type) { 832 case RFC1002_SESSION_MESSAGE: 833 /* Regular SMB response */ 834 return true; 835 case RFC1002_SESSION_KEEP_ALIVE: 836 cifs_dbg(FYI, "RFC 1002 session keep alive\n"); 837 break; 838 case RFC1002_POSITIVE_SESSION_RESPONSE: 839 cifs_dbg(FYI, "RFC 1002 positive session response\n"); 840 break; 841 case RFC1002_NEGATIVE_SESSION_RESPONSE: 842 /* 843 * We get this from Windows 98 instead of an error on 844 * SMB negprot response. 845 */ 846 cifs_dbg(FYI, "RFC 1002 negative session response\n"); 847 /* give server a second to clean up */ 848 msleep(1000); 849 /* 850 * Always try 445 first on reconnect since we get NACK 851 * on some if we ever connected to port 139 (the NACK 852 * is since we do not begin with RFC1001 session 853 * initialize frame). 854 */ 855 cifs_set_port((struct sockaddr *)&server->dstaddr, CIFS_PORT); 856 cifs_reconnect(server, true); 857 break; 858 default: 859 cifs_server_dbg(VFS, "RFC 1002 unknown response type 0x%x\n", type); 860 cifs_reconnect(server, true); 861 } 862 863 return false; 864 } 865 866 void 867 dequeue_mid(struct mid_q_entry *mid, bool malformed) 868 { 869 #ifdef CONFIG_CIFS_STATS2 870 mid->when_received = jiffies; 871 #endif 872 spin_lock(&mid->server->mid_lock); 873 if (!malformed) 874 mid->mid_state = MID_RESPONSE_RECEIVED; 875 else 876 mid->mid_state = MID_RESPONSE_MALFORMED; 877 /* 878 * Trying to handle/dequeue a mid after the send_recv() 879 * function has finished processing it is a bug. 880 */ 881 if (mid->mid_flags & MID_DELETED) { 882 spin_unlock(&mid->server->mid_lock); 883 pr_warn_once("trying to dequeue a deleted mid\n"); 884 } else { 885 list_del_init(&mid->qhead); 886 mid->mid_flags |= MID_DELETED; 887 spin_unlock(&mid->server->mid_lock); 888 } 889 } 890 891 static unsigned int 892 smb2_get_credits_from_hdr(char *buffer, struct TCP_Server_Info *server) 893 { 894 struct smb2_hdr *shdr = (struct smb2_hdr *)buffer; 895 896 /* 897 * SMB1 does not use credits. 898 */ 899 if (is_smb1(server)) 900 return 0; 901 902 return le16_to_cpu(shdr->CreditRequest); 903 } 904 905 static void 906 handle_mid(struct mid_q_entry *mid, struct TCP_Server_Info *server, 907 char *buf, int malformed) 908 { 909 if (server->ops->check_trans2 && 910 server->ops->check_trans2(mid, server, buf, malformed)) 911 return; 912 mid->credits_received = smb2_get_credits_from_hdr(buf, server); 913 mid->resp_buf = buf; 914 mid->large_buf = server->large_buf; 915 /* Was previous buf put in mpx struct for multi-rsp? */ 916 if (!mid->multiRsp) { 917 /* smb buffer will be freed by user thread */ 918 if (server->large_buf) 919 server->bigbuf = NULL; 920 else 921 server->smallbuf = NULL; 922 } 923 dequeue_mid(mid, malformed); 924 } 925 926 int 927 cifs_enable_signing(struct TCP_Server_Info *server, bool mnt_sign_required) 928 { 929 bool srv_sign_required = server->sec_mode & server->vals->signing_required; 930 bool srv_sign_enabled = server->sec_mode & server->vals->signing_enabled; 931 bool mnt_sign_enabled; 932 933 /* 934 * Is signing required by mnt options? If not then check 935 * global_secflags to see if it is there. 936 */ 937 if (!mnt_sign_required) 938 mnt_sign_required = ((global_secflags & CIFSSEC_MUST_SIGN) == 939 CIFSSEC_MUST_SIGN); 940 941 /* 942 * If signing is required then it's automatically enabled too, 943 * otherwise, check to see if the secflags allow it. 944 */ 945 mnt_sign_enabled = mnt_sign_required ? mnt_sign_required : 946 (global_secflags & CIFSSEC_MAY_SIGN); 947 948 /* If server requires signing, does client allow it? */ 949 if (srv_sign_required) { 950 if (!mnt_sign_enabled) { 951 cifs_dbg(VFS, "Server requires signing, but it's disabled in SecurityFlags!\n"); 952 return -EOPNOTSUPP; 953 } 954 server->sign = true; 955 } 956 957 /* If client requires signing, does server allow it? */ 958 if (mnt_sign_required) { 959 if (!srv_sign_enabled) { 960 cifs_dbg(VFS, "Server does not support signing!\n"); 961 return -EOPNOTSUPP; 962 } 963 server->sign = true; 964 } 965 966 if (cifs_rdma_enabled(server) && server->sign) 967 cifs_dbg(VFS, "Signing is enabled, and RDMA read/write will be disabled\n"); 968 969 return 0; 970 } 971 972 static noinline_for_stack void 973 clean_demultiplex_info(struct TCP_Server_Info *server) 974 { 975 int length; 976 977 /* take it off the list, if it's not already */ 978 spin_lock(&server->srv_lock); 979 list_del_init(&server->tcp_ses_list); 980 spin_unlock(&server->srv_lock); 981 982 cancel_delayed_work_sync(&server->echo); 983 984 spin_lock(&server->srv_lock); 985 server->tcpStatus = CifsExiting; 986 spin_unlock(&server->srv_lock); 987 wake_up_all(&server->response_q); 988 989 /* check if we have blocked requests that need to free */ 990 spin_lock(&server->req_lock); 991 if (server->credits <= 0) 992 server->credits = 1; 993 spin_unlock(&server->req_lock); 994 /* 995 * Although there should not be any requests blocked on this queue it 996 * can not hurt to be paranoid and try to wake up requests that may 997 * haven been blocked when more than 50 at time were on the wire to the 998 * same server - they now will see the session is in exit state and get 999 * out of SendReceive. 1000 */ 1001 wake_up_all(&server->request_q); 1002 /* give those requests time to exit */ 1003 msleep(125); 1004 if (cifs_rdma_enabled(server)) 1005 smbd_destroy(server); 1006 if (server->ssocket) { 1007 sock_release(server->ssocket); 1008 server->ssocket = NULL; 1009 } 1010 1011 if (!list_empty(&server->pending_mid_q)) { 1012 struct list_head dispose_list; 1013 struct mid_q_entry *mid_entry; 1014 struct list_head *tmp, *tmp2; 1015 1016 INIT_LIST_HEAD(&dispose_list); 1017 spin_lock(&server->mid_lock); 1018 list_for_each_safe(tmp, tmp2, &server->pending_mid_q) { 1019 mid_entry = list_entry(tmp, struct mid_q_entry, qhead); 1020 cifs_dbg(FYI, "Clearing mid %llu\n", mid_entry->mid); 1021 kref_get(&mid_entry->refcount); 1022 mid_entry->mid_state = MID_SHUTDOWN; 1023 list_move(&mid_entry->qhead, &dispose_list); 1024 mid_entry->mid_flags |= MID_DELETED; 1025 } 1026 spin_unlock(&server->mid_lock); 1027 1028 /* now walk dispose list and issue callbacks */ 1029 list_for_each_safe(tmp, tmp2, &dispose_list) { 1030 mid_entry = list_entry(tmp, struct mid_q_entry, qhead); 1031 cifs_dbg(FYI, "Callback mid %llu\n", mid_entry->mid); 1032 list_del_init(&mid_entry->qhead); 1033 mid_entry->callback(mid_entry); 1034 release_mid(mid_entry); 1035 } 1036 /* 1/8th of sec is more than enough time for them to exit */ 1037 msleep(125); 1038 } 1039 1040 if (!list_empty(&server->pending_mid_q)) { 1041 /* 1042 * mpx threads have not exited yet give them at least the smb 1043 * send timeout time for long ops. 1044 * 1045 * Due to delays on oplock break requests, we need to wait at 1046 * least 45 seconds before giving up on a request getting a 1047 * response and going ahead and killing cifsd. 1048 */ 1049 cifs_dbg(FYI, "Wait for exit from demultiplex thread\n"); 1050 msleep(46000); 1051 /* 1052 * If threads still have not exited they are probably never 1053 * coming home not much else we can do but free the memory. 1054 */ 1055 } 1056 1057 put_net(cifs_net_ns(server)); 1058 kfree(server->leaf_fullpath); 1059 kfree(server); 1060 1061 length = atomic_dec_return(&tcpSesAllocCount); 1062 if (length > 0) 1063 mempool_resize(cifs_req_poolp, length + cifs_min_rcv); 1064 } 1065 1066 static int 1067 standard_receive3(struct TCP_Server_Info *server, struct mid_q_entry *mid) 1068 { 1069 int length; 1070 char *buf = server->smallbuf; 1071 unsigned int pdu_length = server->pdu_size; 1072 1073 /* make sure this will fit in a large buffer */ 1074 if (pdu_length > CIFSMaxBufSize + MAX_HEADER_SIZE(server) - 1075 HEADER_PREAMBLE_SIZE(server)) { 1076 cifs_server_dbg(VFS, "SMB response too long (%u bytes)\n", pdu_length); 1077 cifs_reconnect(server, true); 1078 return -ECONNABORTED; 1079 } 1080 1081 /* switch to large buffer if too big for a small one */ 1082 if (pdu_length > MAX_CIFS_SMALL_BUFFER_SIZE - 4) { 1083 server->large_buf = true; 1084 memcpy(server->bigbuf, buf, server->total_read); 1085 buf = server->bigbuf; 1086 } 1087 1088 /* now read the rest */ 1089 length = cifs_read_from_socket(server, buf + HEADER_SIZE(server) - 1, 1090 pdu_length - MID_HEADER_SIZE(server)); 1091 1092 if (length < 0) 1093 return length; 1094 server->total_read += length; 1095 1096 dump_smb(buf, server->total_read); 1097 1098 return cifs_handle_standard(server, mid); 1099 } 1100 1101 int 1102 cifs_handle_standard(struct TCP_Server_Info *server, struct mid_q_entry *mid) 1103 { 1104 char *buf = server->large_buf ? server->bigbuf : server->smallbuf; 1105 int rc; 1106 1107 /* 1108 * We know that we received enough to get to the MID as we 1109 * checked the pdu_length earlier. Now check to see 1110 * if the rest of the header is OK. 1111 * 1112 * 48 bytes is enough to display the header and a little bit 1113 * into the payload for debugging purposes. 1114 */ 1115 rc = server->ops->check_message(buf, server->total_read, server); 1116 if (rc) 1117 cifs_dump_mem("Bad SMB: ", buf, 1118 min_t(unsigned int, server->total_read, 48)); 1119 1120 if (server->ops->is_session_expired && 1121 server->ops->is_session_expired(buf)) { 1122 cifs_reconnect(server, true); 1123 return -1; 1124 } 1125 1126 if (server->ops->is_status_pending && 1127 server->ops->is_status_pending(buf, server)) 1128 return -1; 1129 1130 if (!mid) 1131 return rc; 1132 1133 handle_mid(mid, server, buf, rc); 1134 return 0; 1135 } 1136 1137 static void 1138 smb2_add_credits_from_hdr(char *buffer, struct TCP_Server_Info *server) 1139 { 1140 struct smb2_hdr *shdr = (struct smb2_hdr *)buffer; 1141 int scredits, in_flight; 1142 1143 /* 1144 * SMB1 does not use credits. 1145 */ 1146 if (is_smb1(server)) 1147 return; 1148 1149 if (shdr->CreditRequest) { 1150 spin_lock(&server->req_lock); 1151 server->credits += le16_to_cpu(shdr->CreditRequest); 1152 scredits = server->credits; 1153 in_flight = server->in_flight; 1154 spin_unlock(&server->req_lock); 1155 wake_up(&server->request_q); 1156 1157 trace_smb3_hdr_credits(server->CurrentMid, 1158 server->conn_id, server->hostname, scredits, 1159 le16_to_cpu(shdr->CreditRequest), in_flight); 1160 cifs_server_dbg(FYI, "%s: added %u credits total=%d\n", 1161 __func__, le16_to_cpu(shdr->CreditRequest), 1162 scredits); 1163 } 1164 } 1165 1166 1167 static int 1168 cifs_demultiplex_thread(void *p) 1169 { 1170 int i, num_mids, length; 1171 struct TCP_Server_Info *server = p; 1172 unsigned int pdu_length; 1173 unsigned int next_offset; 1174 char *buf = NULL; 1175 struct task_struct *task_to_wake = NULL; 1176 struct mid_q_entry *mids[MAX_COMPOUND]; 1177 char *bufs[MAX_COMPOUND]; 1178 unsigned int noreclaim_flag, num_io_timeout = 0; 1179 bool pending_reconnect = false; 1180 1181 noreclaim_flag = memalloc_noreclaim_save(); 1182 cifs_dbg(FYI, "Demultiplex PID: %d\n", task_pid_nr(current)); 1183 1184 length = atomic_inc_return(&tcpSesAllocCount); 1185 if (length > 1) 1186 mempool_resize(cifs_req_poolp, length + cifs_min_rcv); 1187 1188 set_freezable(); 1189 allow_kernel_signal(SIGKILL); 1190 while (server->tcpStatus != CifsExiting) { 1191 if (try_to_freeze()) 1192 continue; 1193 1194 if (!allocate_buffers(server)) 1195 continue; 1196 1197 server->large_buf = false; 1198 buf = server->smallbuf; 1199 pdu_length = 4; /* enough to get RFC1001 header */ 1200 1201 length = cifs_read_from_socket(server, buf, pdu_length); 1202 if (length < 0) 1203 continue; 1204 1205 if (is_smb1(server)) 1206 server->total_read = length; 1207 else 1208 server->total_read = 0; 1209 1210 /* 1211 * The right amount was read from socket - 4 bytes, 1212 * so we can now interpret the length field. 1213 */ 1214 pdu_length = get_rfc1002_length(buf); 1215 1216 cifs_dbg(FYI, "RFC1002 header 0x%x\n", pdu_length); 1217 if (!is_smb_response(server, buf[0])) 1218 continue; 1219 1220 pending_reconnect = false; 1221 next_pdu: 1222 server->pdu_size = pdu_length; 1223 1224 /* make sure we have enough to get to the MID */ 1225 if (server->pdu_size < MID_HEADER_SIZE(server)) { 1226 cifs_server_dbg(VFS, "SMB response too short (%u bytes)\n", 1227 server->pdu_size); 1228 cifs_reconnect(server, true); 1229 continue; 1230 } 1231 1232 /* read down to the MID */ 1233 length = cifs_read_from_socket(server, 1234 buf + HEADER_PREAMBLE_SIZE(server), 1235 MID_HEADER_SIZE(server)); 1236 if (length < 0) 1237 continue; 1238 server->total_read += length; 1239 1240 if (server->ops->next_header) { 1241 if (server->ops->next_header(server, buf, &next_offset)) { 1242 cifs_dbg(VFS, "%s: malformed response (next_offset=%u)\n", 1243 __func__, next_offset); 1244 cifs_reconnect(server, true); 1245 continue; 1246 } 1247 if (next_offset) 1248 server->pdu_size = next_offset; 1249 } 1250 1251 memset(mids, 0, sizeof(mids)); 1252 memset(bufs, 0, sizeof(bufs)); 1253 num_mids = 0; 1254 1255 if (server->ops->is_transform_hdr && 1256 server->ops->receive_transform && 1257 server->ops->is_transform_hdr(buf)) { 1258 length = server->ops->receive_transform(server, 1259 mids, 1260 bufs, 1261 &num_mids); 1262 } else { 1263 mids[0] = server->ops->find_mid(server, buf); 1264 bufs[0] = buf; 1265 num_mids = 1; 1266 1267 if (!mids[0] || !mids[0]->receive) 1268 length = standard_receive3(server, mids[0]); 1269 else 1270 length = mids[0]->receive(server, mids[0]); 1271 } 1272 1273 if (length < 0) { 1274 for (i = 0; i < num_mids; i++) 1275 if (mids[i]) 1276 release_mid(mids[i]); 1277 continue; 1278 } 1279 1280 if (server->ops->is_status_io_timeout && 1281 server->ops->is_status_io_timeout(buf)) { 1282 num_io_timeout++; 1283 if (num_io_timeout > MAX_STATUS_IO_TIMEOUT) { 1284 cifs_server_dbg(VFS, 1285 "Number of request timeouts exceeded %d. Reconnecting", 1286 MAX_STATUS_IO_TIMEOUT); 1287 1288 pending_reconnect = true; 1289 num_io_timeout = 0; 1290 } 1291 } 1292 1293 server->lstrp = jiffies; 1294 1295 for (i = 0; i < num_mids; i++) { 1296 if (mids[i] != NULL) { 1297 mids[i]->resp_buf_size = server->pdu_size; 1298 1299 if (bufs[i] != NULL) { 1300 if (server->ops->is_network_name_deleted && 1301 server->ops->is_network_name_deleted(bufs[i], 1302 server)) { 1303 cifs_server_dbg(FYI, 1304 "Share deleted. Reconnect needed"); 1305 } 1306 } 1307 1308 if (!mids[i]->multiRsp || mids[i]->multiEnd) 1309 mids[i]->callback(mids[i]); 1310 1311 release_mid(mids[i]); 1312 } else if (server->ops->is_oplock_break && 1313 server->ops->is_oplock_break(bufs[i], 1314 server)) { 1315 smb2_add_credits_from_hdr(bufs[i], server); 1316 cifs_dbg(FYI, "Received oplock break\n"); 1317 } else { 1318 cifs_server_dbg(VFS, "No task to wake, unknown frame received! NumMids %d\n", 1319 atomic_read(&mid_count)); 1320 cifs_dump_mem("Received Data is: ", bufs[i], 1321 HEADER_SIZE(server)); 1322 smb2_add_credits_from_hdr(bufs[i], server); 1323 #ifdef CONFIG_CIFS_DEBUG2 1324 if (server->ops->dump_detail) 1325 server->ops->dump_detail(bufs[i], 1326 server); 1327 cifs_dump_mids(server); 1328 #endif /* CIFS_DEBUG2 */ 1329 } 1330 } 1331 1332 if (pdu_length > server->pdu_size) { 1333 if (!allocate_buffers(server)) 1334 continue; 1335 pdu_length -= server->pdu_size; 1336 server->total_read = 0; 1337 server->large_buf = false; 1338 buf = server->smallbuf; 1339 goto next_pdu; 1340 } 1341 1342 /* do this reconnect at the very end after processing all MIDs */ 1343 if (pending_reconnect) 1344 cifs_reconnect(server, true); 1345 1346 } /* end while !EXITING */ 1347 1348 /* buffer usually freed in free_mid - need to free it here on exit */ 1349 cifs_buf_release(server->bigbuf); 1350 if (server->smallbuf) /* no sense logging a debug message if NULL */ 1351 cifs_small_buf_release(server->smallbuf); 1352 1353 task_to_wake = xchg(&server->tsk, NULL); 1354 clean_demultiplex_info(server); 1355 1356 /* if server->tsk was NULL then wait for a signal before exiting */ 1357 if (!task_to_wake) { 1358 set_current_state(TASK_INTERRUPTIBLE); 1359 while (!signal_pending(current)) { 1360 schedule(); 1361 set_current_state(TASK_INTERRUPTIBLE); 1362 } 1363 set_current_state(TASK_RUNNING); 1364 } 1365 1366 memalloc_noreclaim_restore(noreclaim_flag); 1367 module_put_and_kthread_exit(0); 1368 } 1369 1370 int 1371 cifs_ipaddr_cmp(struct sockaddr *srcaddr, struct sockaddr *rhs) 1372 { 1373 struct sockaddr_in *saddr4 = (struct sockaddr_in *)srcaddr; 1374 struct sockaddr_in *vaddr4 = (struct sockaddr_in *)rhs; 1375 struct sockaddr_in6 *saddr6 = (struct sockaddr_in6 *)srcaddr; 1376 struct sockaddr_in6 *vaddr6 = (struct sockaddr_in6 *)rhs; 1377 1378 switch (srcaddr->sa_family) { 1379 case AF_UNSPEC: 1380 switch (rhs->sa_family) { 1381 case AF_UNSPEC: 1382 return 0; 1383 case AF_INET: 1384 case AF_INET6: 1385 return 1; 1386 default: 1387 return -1; 1388 } 1389 case AF_INET: { 1390 switch (rhs->sa_family) { 1391 case AF_UNSPEC: 1392 return -1; 1393 case AF_INET: 1394 return memcmp(saddr4, vaddr4, 1395 sizeof(struct sockaddr_in)); 1396 case AF_INET6: 1397 return 1; 1398 default: 1399 return -1; 1400 } 1401 } 1402 case AF_INET6: { 1403 switch (rhs->sa_family) { 1404 case AF_UNSPEC: 1405 case AF_INET: 1406 return -1; 1407 case AF_INET6: 1408 return memcmp(saddr6, 1409 vaddr6, 1410 sizeof(struct sockaddr_in6)); 1411 default: 1412 return -1; 1413 } 1414 } 1415 default: 1416 return -1; /* don't expect to be here */ 1417 } 1418 } 1419 1420 /* 1421 * Returns true if srcaddr isn't specified and rhs isn't specified, or 1422 * if srcaddr is specified and matches the IP address of the rhs argument 1423 */ 1424 bool 1425 cifs_match_ipaddr(struct sockaddr *srcaddr, struct sockaddr *rhs) 1426 { 1427 switch (srcaddr->sa_family) { 1428 case AF_UNSPEC: 1429 return (rhs->sa_family == AF_UNSPEC); 1430 case AF_INET: { 1431 struct sockaddr_in *saddr4 = (struct sockaddr_in *)srcaddr; 1432 struct sockaddr_in *vaddr4 = (struct sockaddr_in *)rhs; 1433 1434 return (saddr4->sin_addr.s_addr == vaddr4->sin_addr.s_addr); 1435 } 1436 case AF_INET6: { 1437 struct sockaddr_in6 *saddr6 = (struct sockaddr_in6 *)srcaddr; 1438 struct sockaddr_in6 *vaddr6 = (struct sockaddr_in6 *)rhs; 1439 1440 return (ipv6_addr_equal(&saddr6->sin6_addr, &vaddr6->sin6_addr) 1441 && saddr6->sin6_scope_id == vaddr6->sin6_scope_id); 1442 } 1443 default: 1444 WARN_ON(1); 1445 return false; /* don't expect to be here */ 1446 } 1447 } 1448 1449 /* 1450 * If no port is specified in addr structure, we try to match with 445 port 1451 * and if it fails - with 139 ports. It should be called only if address 1452 * families of server and addr are equal. 1453 */ 1454 static bool 1455 match_port(struct TCP_Server_Info *server, struct sockaddr *addr) 1456 { 1457 __be16 port, *sport; 1458 1459 /* SMBDirect manages its own ports, don't match it here */ 1460 if (server->rdma) 1461 return true; 1462 1463 switch (addr->sa_family) { 1464 case AF_INET: 1465 sport = &((struct sockaddr_in *) &server->dstaddr)->sin_port; 1466 port = ((struct sockaddr_in *) addr)->sin_port; 1467 break; 1468 case AF_INET6: 1469 sport = &((struct sockaddr_in6 *) &server->dstaddr)->sin6_port; 1470 port = ((struct sockaddr_in6 *) addr)->sin6_port; 1471 break; 1472 default: 1473 WARN_ON(1); 1474 return false; 1475 } 1476 1477 if (!port) { 1478 port = htons(CIFS_PORT); 1479 if (port == *sport) 1480 return true; 1481 1482 port = htons(RFC1001_PORT); 1483 } 1484 1485 return port == *sport; 1486 } 1487 1488 static bool match_server_address(struct TCP_Server_Info *server, struct sockaddr *addr) 1489 { 1490 if (!cifs_match_ipaddr(addr, (struct sockaddr *)&server->dstaddr)) 1491 return false; 1492 1493 return true; 1494 } 1495 1496 static bool 1497 match_security(struct TCP_Server_Info *server, struct smb3_fs_context *ctx) 1498 { 1499 /* 1500 * The select_sectype function should either return the ctx->sectype 1501 * that was specified, or "Unspecified" if that sectype was not 1502 * compatible with the given NEGOTIATE request. 1503 */ 1504 if (server->ops->select_sectype(server, ctx->sectype) 1505 == Unspecified) 1506 return false; 1507 1508 /* 1509 * Now check if signing mode is acceptable. No need to check 1510 * global_secflags at this point since if MUST_SIGN is set then 1511 * the server->sign had better be too. 1512 */ 1513 if (ctx->sign && !server->sign) 1514 return false; 1515 1516 return true; 1517 } 1518 1519 /* this function must be called with srv_lock held */ 1520 static int match_server(struct TCP_Server_Info *server, 1521 struct smb3_fs_context *ctx, 1522 bool match_super) 1523 { 1524 struct sockaddr *addr = (struct sockaddr *)&ctx->dstaddr; 1525 1526 lockdep_assert_held(&server->srv_lock); 1527 1528 if (ctx->nosharesock) 1529 return 0; 1530 1531 /* this server does not share socket */ 1532 if (server->nosharesock) 1533 return 0; 1534 1535 /* If multidialect negotiation see if existing sessions match one */ 1536 if (strcmp(ctx->vals->version_string, SMB3ANY_VERSION_STRING) == 0) { 1537 if (server->vals->protocol_id < SMB30_PROT_ID) 1538 return 0; 1539 } else if (strcmp(ctx->vals->version_string, 1540 SMBDEFAULT_VERSION_STRING) == 0) { 1541 if (server->vals->protocol_id < SMB21_PROT_ID) 1542 return 0; 1543 } else if ((server->vals != ctx->vals) || (server->ops != ctx->ops)) 1544 return 0; 1545 1546 if (!net_eq(cifs_net_ns(server), current->nsproxy->net_ns)) 1547 return 0; 1548 1549 if (!cifs_match_ipaddr((struct sockaddr *)&ctx->srcaddr, 1550 (struct sockaddr *)&server->srcaddr)) 1551 return 0; 1552 /* 1553 * When matching cifs.ko superblocks (@match_super == true), we can't 1554 * really match either @server->leaf_fullpath or @server->dstaddr 1555 * directly since this @server might belong to a completely different 1556 * server -- in case of domain-based DFS referrals or DFS links -- as 1557 * provided earlier by mount(2) through 'source' and 'ip' options. 1558 * 1559 * Otherwise, match the DFS referral in @server->leaf_fullpath or the 1560 * destination address in @server->dstaddr. 1561 * 1562 * When using 'nodfs' mount option, we avoid sharing it with DFS 1563 * connections as they might failover. 1564 */ 1565 if (!match_super) { 1566 if (!ctx->nodfs) { 1567 if (server->leaf_fullpath) { 1568 if (!ctx->leaf_fullpath || 1569 strcasecmp(server->leaf_fullpath, 1570 ctx->leaf_fullpath)) 1571 return 0; 1572 } else if (ctx->leaf_fullpath) { 1573 return 0; 1574 } 1575 } else if (server->leaf_fullpath) { 1576 return 0; 1577 } 1578 } 1579 1580 /* 1581 * Match for a regular connection (address/hostname/port) which has no 1582 * DFS referrals set. 1583 */ 1584 if (!server->leaf_fullpath && 1585 (strcasecmp(server->hostname, ctx->server_hostname) || 1586 !match_server_address(server, addr) || 1587 !match_port(server, addr))) 1588 return 0; 1589 1590 if (!match_security(server, ctx)) 1591 return 0; 1592 1593 if (server->echo_interval != ctx->echo_interval * HZ) 1594 return 0; 1595 1596 if (server->rdma != ctx->rdma) 1597 return 0; 1598 1599 if (server->ignore_signature != ctx->ignore_signature) 1600 return 0; 1601 1602 if (server->min_offload != ctx->min_offload) 1603 return 0; 1604 1605 if (server->retrans != ctx->retrans) 1606 return 0; 1607 1608 return 1; 1609 } 1610 1611 struct TCP_Server_Info * 1612 cifs_find_tcp_session(struct smb3_fs_context *ctx) 1613 { 1614 struct TCP_Server_Info *server; 1615 1616 spin_lock(&cifs_tcp_ses_lock); 1617 list_for_each_entry(server, &cifs_tcp_ses_list, tcp_ses_list) { 1618 spin_lock(&server->srv_lock); 1619 /* 1620 * Skip ses channels since they're only handled in lower layers 1621 * (e.g. cifs_send_recv). 1622 */ 1623 if (SERVER_IS_CHAN(server) || 1624 !match_server(server, ctx, false)) { 1625 spin_unlock(&server->srv_lock); 1626 continue; 1627 } 1628 spin_unlock(&server->srv_lock); 1629 1630 ++server->srv_count; 1631 spin_unlock(&cifs_tcp_ses_lock); 1632 cifs_dbg(FYI, "Existing tcp session with server found\n"); 1633 return server; 1634 } 1635 spin_unlock(&cifs_tcp_ses_lock); 1636 return NULL; 1637 } 1638 1639 void 1640 cifs_put_tcp_session(struct TCP_Server_Info *server, int from_reconnect) 1641 { 1642 struct task_struct *task; 1643 1644 spin_lock(&cifs_tcp_ses_lock); 1645 if (--server->srv_count > 0) { 1646 spin_unlock(&cifs_tcp_ses_lock); 1647 return; 1648 } 1649 1650 /* srv_count can never go negative */ 1651 WARN_ON(server->srv_count < 0); 1652 1653 list_del_init(&server->tcp_ses_list); 1654 spin_unlock(&cifs_tcp_ses_lock); 1655 1656 cancel_delayed_work_sync(&server->echo); 1657 1658 if (from_reconnect) 1659 /* 1660 * Avoid deadlock here: reconnect work calls 1661 * cifs_put_tcp_session() at its end. Need to be sure 1662 * that reconnect work does nothing with server pointer after 1663 * that step. 1664 */ 1665 cancel_delayed_work(&server->reconnect); 1666 else 1667 cancel_delayed_work_sync(&server->reconnect); 1668 1669 /* For secondary channels, we pick up ref-count on the primary server */ 1670 if (SERVER_IS_CHAN(server)) 1671 cifs_put_tcp_session(server->primary_server, from_reconnect); 1672 1673 spin_lock(&server->srv_lock); 1674 server->tcpStatus = CifsExiting; 1675 spin_unlock(&server->srv_lock); 1676 1677 cifs_crypto_secmech_release(server); 1678 1679 kfree_sensitive(server->session_key.response); 1680 server->session_key.response = NULL; 1681 server->session_key.len = 0; 1682 kfree(server->hostname); 1683 server->hostname = NULL; 1684 1685 task = xchg(&server->tsk, NULL); 1686 if (task) 1687 send_sig(SIGKILL, task, 1); 1688 } 1689 1690 struct TCP_Server_Info * 1691 cifs_get_tcp_session(struct smb3_fs_context *ctx, 1692 struct TCP_Server_Info *primary_server) 1693 { 1694 struct TCP_Server_Info *tcp_ses = NULL; 1695 int rc; 1696 1697 cifs_dbg(FYI, "UNC: %s\n", ctx->UNC); 1698 1699 /* see if we already have a matching tcp_ses */ 1700 tcp_ses = cifs_find_tcp_session(ctx); 1701 if (tcp_ses) 1702 return tcp_ses; 1703 1704 tcp_ses = kzalloc(sizeof(struct TCP_Server_Info), GFP_KERNEL); 1705 if (!tcp_ses) { 1706 rc = -ENOMEM; 1707 goto out_err; 1708 } 1709 1710 tcp_ses->hostname = kstrdup(ctx->server_hostname, GFP_KERNEL); 1711 if (!tcp_ses->hostname) { 1712 rc = -ENOMEM; 1713 goto out_err; 1714 } 1715 1716 if (ctx->leaf_fullpath) { 1717 tcp_ses->leaf_fullpath = kstrdup(ctx->leaf_fullpath, GFP_KERNEL); 1718 if (!tcp_ses->leaf_fullpath) { 1719 rc = -ENOMEM; 1720 goto out_err; 1721 } 1722 } 1723 1724 if (ctx->nosharesock) 1725 tcp_ses->nosharesock = true; 1726 1727 tcp_ses->ops = ctx->ops; 1728 tcp_ses->vals = ctx->vals; 1729 cifs_set_net_ns(tcp_ses, get_net(current->nsproxy->net_ns)); 1730 1731 tcp_ses->conn_id = atomic_inc_return(&tcpSesNextId); 1732 tcp_ses->noblockcnt = ctx->rootfs; 1733 tcp_ses->noblocksnd = ctx->noblocksnd || ctx->rootfs; 1734 tcp_ses->noautotune = ctx->noautotune; 1735 tcp_ses->tcp_nodelay = ctx->sockopt_tcp_nodelay; 1736 tcp_ses->rdma = ctx->rdma; 1737 tcp_ses->in_flight = 0; 1738 tcp_ses->max_in_flight = 0; 1739 tcp_ses->credits = 1; 1740 if (primary_server) { 1741 spin_lock(&cifs_tcp_ses_lock); 1742 ++primary_server->srv_count; 1743 spin_unlock(&cifs_tcp_ses_lock); 1744 tcp_ses->primary_server = primary_server; 1745 } 1746 init_waitqueue_head(&tcp_ses->response_q); 1747 init_waitqueue_head(&tcp_ses->request_q); 1748 INIT_LIST_HEAD(&tcp_ses->pending_mid_q); 1749 mutex_init(&tcp_ses->_srv_mutex); 1750 memcpy(tcp_ses->workstation_RFC1001_name, 1751 ctx->source_rfc1001_name, RFC1001_NAME_LEN_WITH_NULL); 1752 memcpy(tcp_ses->server_RFC1001_name, 1753 ctx->target_rfc1001_name, RFC1001_NAME_LEN_WITH_NULL); 1754 tcp_ses->session_estab = false; 1755 tcp_ses->sequence_number = 0; 1756 tcp_ses->channel_sequence_num = 0; /* only tracked for primary channel */ 1757 tcp_ses->reconnect_instance = 1; 1758 tcp_ses->lstrp = jiffies; 1759 tcp_ses->compression.requested = ctx->compress; 1760 spin_lock_init(&tcp_ses->req_lock); 1761 spin_lock_init(&tcp_ses->srv_lock); 1762 spin_lock_init(&tcp_ses->mid_lock); 1763 INIT_LIST_HEAD(&tcp_ses->tcp_ses_list); 1764 INIT_LIST_HEAD(&tcp_ses->smb_ses_list); 1765 INIT_DELAYED_WORK(&tcp_ses->echo, cifs_echo_request); 1766 INIT_DELAYED_WORK(&tcp_ses->reconnect, smb2_reconnect_server); 1767 mutex_init(&tcp_ses->reconnect_mutex); 1768 #ifdef CONFIG_CIFS_DFS_UPCALL 1769 mutex_init(&tcp_ses->refpath_lock); 1770 #endif 1771 memcpy(&tcp_ses->srcaddr, &ctx->srcaddr, 1772 sizeof(tcp_ses->srcaddr)); 1773 memcpy(&tcp_ses->dstaddr, &ctx->dstaddr, 1774 sizeof(tcp_ses->dstaddr)); 1775 if (ctx->use_client_guid) 1776 memcpy(tcp_ses->client_guid, ctx->client_guid, 1777 SMB2_CLIENT_GUID_SIZE); 1778 else 1779 generate_random_uuid(tcp_ses->client_guid); 1780 /* 1781 * at this point we are the only ones with the pointer 1782 * to the struct since the kernel thread not created yet 1783 * no need to spinlock this init of tcpStatus or srv_count 1784 */ 1785 tcp_ses->tcpStatus = CifsNew; 1786 ++tcp_ses->srv_count; 1787 1788 if (ctx->echo_interval >= SMB_ECHO_INTERVAL_MIN && 1789 ctx->echo_interval <= SMB_ECHO_INTERVAL_MAX) 1790 tcp_ses->echo_interval = ctx->echo_interval * HZ; 1791 else 1792 tcp_ses->echo_interval = SMB_ECHO_INTERVAL_DEFAULT * HZ; 1793 if (tcp_ses->rdma) { 1794 #ifndef CONFIG_CIFS_SMB_DIRECT 1795 cifs_dbg(VFS, "CONFIG_CIFS_SMB_DIRECT is not enabled\n"); 1796 rc = -ENOENT; 1797 goto out_err_crypto_release; 1798 #endif 1799 tcp_ses->smbd_conn = smbd_get_connection( 1800 tcp_ses, (struct sockaddr *)&ctx->dstaddr); 1801 if (tcp_ses->smbd_conn) { 1802 cifs_dbg(VFS, "RDMA transport established\n"); 1803 rc = 0; 1804 goto smbd_connected; 1805 } else { 1806 rc = -ENOENT; 1807 goto out_err_crypto_release; 1808 } 1809 } 1810 rc = ip_connect(tcp_ses); 1811 if (rc < 0) { 1812 cifs_dbg(VFS, "Error connecting to socket. Aborting operation.\n"); 1813 goto out_err_crypto_release; 1814 } 1815 smbd_connected: 1816 /* 1817 * since we're in a cifs function already, we know that 1818 * this will succeed. No need for try_module_get(). 1819 */ 1820 __module_get(THIS_MODULE); 1821 tcp_ses->tsk = kthread_run(cifs_demultiplex_thread, 1822 tcp_ses, "cifsd"); 1823 if (IS_ERR(tcp_ses->tsk)) { 1824 rc = PTR_ERR(tcp_ses->tsk); 1825 cifs_dbg(VFS, "error %d create cifsd thread\n", rc); 1826 module_put(THIS_MODULE); 1827 goto out_err_crypto_release; 1828 } 1829 tcp_ses->min_offload = ctx->min_offload; 1830 tcp_ses->retrans = ctx->retrans; 1831 /* 1832 * at this point we are the only ones with the pointer 1833 * to the struct since the kernel thread not created yet 1834 * no need to spinlock this update of tcpStatus 1835 */ 1836 spin_lock(&tcp_ses->srv_lock); 1837 tcp_ses->tcpStatus = CifsNeedNegotiate; 1838 spin_unlock(&tcp_ses->srv_lock); 1839 1840 if ((ctx->max_credits < 20) || (ctx->max_credits > 60000)) 1841 tcp_ses->max_credits = SMB2_MAX_CREDITS_AVAILABLE; 1842 else 1843 tcp_ses->max_credits = ctx->max_credits; 1844 1845 tcp_ses->nr_targets = 1; 1846 tcp_ses->ignore_signature = ctx->ignore_signature; 1847 /* thread spawned, put it on the list */ 1848 spin_lock(&cifs_tcp_ses_lock); 1849 list_add(&tcp_ses->tcp_ses_list, &cifs_tcp_ses_list); 1850 spin_unlock(&cifs_tcp_ses_lock); 1851 1852 /* queue echo request delayed work */ 1853 queue_delayed_work(cifsiod_wq, &tcp_ses->echo, tcp_ses->echo_interval); 1854 1855 return tcp_ses; 1856 1857 out_err_crypto_release: 1858 cifs_crypto_secmech_release(tcp_ses); 1859 1860 put_net(cifs_net_ns(tcp_ses)); 1861 1862 out_err: 1863 if (tcp_ses) { 1864 if (SERVER_IS_CHAN(tcp_ses)) 1865 cifs_put_tcp_session(tcp_ses->primary_server, false); 1866 kfree(tcp_ses->hostname); 1867 kfree(tcp_ses->leaf_fullpath); 1868 if (tcp_ses->ssocket) 1869 sock_release(tcp_ses->ssocket); 1870 kfree(tcp_ses); 1871 } 1872 return ERR_PTR(rc); 1873 } 1874 1875 /* this function must be called with ses_lock and chan_lock held */ 1876 static int match_session(struct cifs_ses *ses, struct smb3_fs_context *ctx) 1877 { 1878 if (ctx->sectype != Unspecified && 1879 ctx->sectype != ses->sectype) 1880 return 0; 1881 1882 if (ctx->dfs_root_ses != ses->dfs_root_ses) 1883 return 0; 1884 1885 /* 1886 * If an existing session is limited to less channels than 1887 * requested, it should not be reused 1888 */ 1889 if (ses->chan_max < ctx->max_channels) 1890 return 0; 1891 1892 switch (ses->sectype) { 1893 case Kerberos: 1894 if (!uid_eq(ctx->cred_uid, ses->cred_uid)) 1895 return 0; 1896 break; 1897 default: 1898 /* NULL username means anonymous session */ 1899 if (ses->user_name == NULL) { 1900 if (!ctx->nullauth) 1901 return 0; 1902 break; 1903 } 1904 1905 /* anything else takes username/password */ 1906 if (strncmp(ses->user_name, 1907 ctx->username ? ctx->username : "", 1908 CIFS_MAX_USERNAME_LEN)) 1909 return 0; 1910 if ((ctx->username && strlen(ctx->username) != 0) && 1911 ses->password != NULL && 1912 strncmp(ses->password, 1913 ctx->password ? ctx->password : "", 1914 CIFS_MAX_PASSWORD_LEN)) 1915 return 0; 1916 } 1917 1918 if (strcmp(ctx->local_nls->charset, ses->local_nls->charset)) 1919 return 0; 1920 1921 return 1; 1922 } 1923 1924 /** 1925 * cifs_setup_ipc - helper to setup the IPC tcon for the session 1926 * @ses: smb session to issue the request on 1927 * @ctx: the superblock configuration context to use for building the 1928 * new tree connection for the IPC (interprocess communication RPC) 1929 * 1930 * A new IPC connection is made and stored in the session 1931 * tcon_ipc. The IPC tcon has the same lifetime as the session. 1932 */ 1933 static int 1934 cifs_setup_ipc(struct cifs_ses *ses, struct smb3_fs_context *ctx) 1935 { 1936 int rc = 0, xid; 1937 struct cifs_tcon *tcon; 1938 char unc[SERVER_NAME_LENGTH + sizeof("//x/IPC$")] = {0}; 1939 bool seal = false; 1940 struct TCP_Server_Info *server = ses->server; 1941 1942 /* 1943 * If the mount request that resulted in the creation of the 1944 * session requires encryption, force IPC to be encrypted too. 1945 */ 1946 if (ctx->seal) { 1947 if (server->capabilities & SMB2_GLOBAL_CAP_ENCRYPTION) 1948 seal = true; 1949 else { 1950 cifs_server_dbg(VFS, 1951 "IPC: server doesn't support encryption\n"); 1952 return -EOPNOTSUPP; 1953 } 1954 } 1955 1956 /* no need to setup directory caching on IPC share, so pass in false */ 1957 tcon = tcon_info_alloc(false, netfs_trace_tcon_ref_new_ipc); 1958 if (tcon == NULL) 1959 return -ENOMEM; 1960 1961 spin_lock(&server->srv_lock); 1962 scnprintf(unc, sizeof(unc), "\\\\%s\\IPC$", server->hostname); 1963 spin_unlock(&server->srv_lock); 1964 1965 xid = get_xid(); 1966 tcon->ses = ses; 1967 tcon->ipc = true; 1968 tcon->seal = seal; 1969 rc = server->ops->tree_connect(xid, ses, unc, tcon, ctx->local_nls); 1970 free_xid(xid); 1971 1972 if (rc) { 1973 cifs_server_dbg(VFS, "failed to connect to IPC (rc=%d)\n", rc); 1974 tconInfoFree(tcon, netfs_trace_tcon_ref_free_ipc_fail); 1975 goto out; 1976 } 1977 1978 cifs_dbg(FYI, "IPC tcon rc=%d ipc tid=0x%x\n", rc, tcon->tid); 1979 1980 spin_lock(&tcon->tc_lock); 1981 tcon->status = TID_GOOD; 1982 spin_unlock(&tcon->tc_lock); 1983 ses->tcon_ipc = tcon; 1984 out: 1985 return rc; 1986 } 1987 1988 static struct cifs_ses * 1989 cifs_find_smb_ses(struct TCP_Server_Info *server, struct smb3_fs_context *ctx) 1990 { 1991 struct cifs_ses *ses, *ret = NULL; 1992 1993 spin_lock(&cifs_tcp_ses_lock); 1994 list_for_each_entry(ses, &server->smb_ses_list, smb_ses_list) { 1995 spin_lock(&ses->ses_lock); 1996 if (ses->ses_status == SES_EXITING) { 1997 spin_unlock(&ses->ses_lock); 1998 continue; 1999 } 2000 spin_lock(&ses->chan_lock); 2001 if (match_session(ses, ctx)) { 2002 spin_unlock(&ses->chan_lock); 2003 spin_unlock(&ses->ses_lock); 2004 ret = ses; 2005 break; 2006 } 2007 spin_unlock(&ses->chan_lock); 2008 spin_unlock(&ses->ses_lock); 2009 } 2010 if (ret) 2011 cifs_smb_ses_inc_refcount(ret); 2012 spin_unlock(&cifs_tcp_ses_lock); 2013 return ret; 2014 } 2015 2016 void __cifs_put_smb_ses(struct cifs_ses *ses) 2017 { 2018 struct TCP_Server_Info *server = ses->server; 2019 struct cifs_tcon *tcon; 2020 unsigned int xid; 2021 size_t i; 2022 bool do_logoff; 2023 int rc; 2024 2025 spin_lock(&cifs_tcp_ses_lock); 2026 spin_lock(&ses->ses_lock); 2027 cifs_dbg(FYI, "%s: id=0x%llx ses_count=%d ses_status=%u ipc=%s\n", 2028 __func__, ses->Suid, ses->ses_count, ses->ses_status, 2029 ses->tcon_ipc ? ses->tcon_ipc->tree_name : "none"); 2030 if (ses->ses_status == SES_EXITING || --ses->ses_count > 0) { 2031 spin_unlock(&ses->ses_lock); 2032 spin_unlock(&cifs_tcp_ses_lock); 2033 return; 2034 } 2035 /* ses_count can never go negative */ 2036 WARN_ON(ses->ses_count < 0); 2037 2038 spin_lock(&ses->chan_lock); 2039 cifs_chan_clear_need_reconnect(ses, server); 2040 spin_unlock(&ses->chan_lock); 2041 2042 do_logoff = ses->ses_status == SES_GOOD && server->ops->logoff; 2043 ses->ses_status = SES_EXITING; 2044 tcon = ses->tcon_ipc; 2045 ses->tcon_ipc = NULL; 2046 spin_unlock(&ses->ses_lock); 2047 spin_unlock(&cifs_tcp_ses_lock); 2048 2049 /* 2050 * On session close, the IPC is closed and the server must release all 2051 * tcons of the session. No need to send a tree disconnect here. 2052 * 2053 * Besides, it will make the server to not close durable and resilient 2054 * files on session close, as specified in MS-SMB2 3.3.5.6 Receiving an 2055 * SMB2 LOGOFF Request. 2056 */ 2057 tconInfoFree(tcon, netfs_trace_tcon_ref_free_ipc); 2058 if (do_logoff) { 2059 xid = get_xid(); 2060 rc = server->ops->logoff(xid, ses); 2061 if (rc) 2062 cifs_server_dbg(VFS, "%s: Session Logoff failure rc=%d\n", 2063 __func__, rc); 2064 _free_xid(xid); 2065 } 2066 2067 spin_lock(&cifs_tcp_ses_lock); 2068 list_del_init(&ses->smb_ses_list); 2069 spin_unlock(&cifs_tcp_ses_lock); 2070 2071 /* close any extra channels */ 2072 for (i = 1; i < ses->chan_count; i++) { 2073 if (ses->chans[i].iface) { 2074 kref_put(&ses->chans[i].iface->refcount, release_iface); 2075 ses->chans[i].iface = NULL; 2076 } 2077 cifs_put_tcp_session(ses->chans[i].server, 0); 2078 ses->chans[i].server = NULL; 2079 } 2080 2081 /* we now account for primary channel in iface->refcount */ 2082 if (ses->chans[0].iface) { 2083 kref_put(&ses->chans[0].iface->refcount, release_iface); 2084 ses->chans[0].server = NULL; 2085 } 2086 2087 sesInfoFree(ses); 2088 cifs_put_tcp_session(server, 0); 2089 } 2090 2091 #ifdef CONFIG_KEYS 2092 2093 /* strlen("cifs:a:") + CIFS_MAX_DOMAINNAME_LEN + 1 */ 2094 #define CIFSCREDS_DESC_SIZE (7 + CIFS_MAX_DOMAINNAME_LEN + 1) 2095 2096 /* Populate username and pw fields from keyring if possible */ 2097 static int 2098 cifs_set_cifscreds(struct smb3_fs_context *ctx, struct cifs_ses *ses) 2099 { 2100 int rc = 0; 2101 int is_domain = 0; 2102 const char *delim, *payload; 2103 char *desc; 2104 ssize_t len; 2105 struct key *key; 2106 struct TCP_Server_Info *server = ses->server; 2107 struct sockaddr_in *sa; 2108 struct sockaddr_in6 *sa6; 2109 const struct user_key_payload *upayload; 2110 2111 desc = kmalloc(CIFSCREDS_DESC_SIZE, GFP_KERNEL); 2112 if (!desc) 2113 return -ENOMEM; 2114 2115 /* try to find an address key first */ 2116 switch (server->dstaddr.ss_family) { 2117 case AF_INET: 2118 sa = (struct sockaddr_in *)&server->dstaddr; 2119 sprintf(desc, "cifs:a:%pI4", &sa->sin_addr.s_addr); 2120 break; 2121 case AF_INET6: 2122 sa6 = (struct sockaddr_in6 *)&server->dstaddr; 2123 sprintf(desc, "cifs:a:%pI6c", &sa6->sin6_addr.s6_addr); 2124 break; 2125 default: 2126 cifs_dbg(FYI, "Bad ss_family (%hu)\n", 2127 server->dstaddr.ss_family); 2128 rc = -EINVAL; 2129 goto out_err; 2130 } 2131 2132 cifs_dbg(FYI, "%s: desc=%s\n", __func__, desc); 2133 key = request_key(&key_type_logon, desc, ""); 2134 if (IS_ERR(key)) { 2135 if (!ses->domainName) { 2136 cifs_dbg(FYI, "domainName is NULL\n"); 2137 rc = PTR_ERR(key); 2138 goto out_err; 2139 } 2140 2141 /* didn't work, try to find a domain key */ 2142 sprintf(desc, "cifs:d:%s", ses->domainName); 2143 cifs_dbg(FYI, "%s: desc=%s\n", __func__, desc); 2144 key = request_key(&key_type_logon, desc, ""); 2145 if (IS_ERR(key)) { 2146 rc = PTR_ERR(key); 2147 goto out_err; 2148 } 2149 is_domain = 1; 2150 } 2151 2152 down_read(&key->sem); 2153 upayload = user_key_payload_locked(key); 2154 if (IS_ERR_OR_NULL(upayload)) { 2155 rc = upayload ? PTR_ERR(upayload) : -EINVAL; 2156 goto out_key_put; 2157 } 2158 2159 /* find first : in payload */ 2160 payload = upayload->data; 2161 delim = strnchr(payload, upayload->datalen, ':'); 2162 cifs_dbg(FYI, "payload=%s\n", payload); 2163 if (!delim) { 2164 cifs_dbg(FYI, "Unable to find ':' in payload (datalen=%d)\n", 2165 upayload->datalen); 2166 rc = -EINVAL; 2167 goto out_key_put; 2168 } 2169 2170 len = delim - payload; 2171 if (len > CIFS_MAX_USERNAME_LEN || len <= 0) { 2172 cifs_dbg(FYI, "Bad value from username search (len=%zd)\n", 2173 len); 2174 rc = -EINVAL; 2175 goto out_key_put; 2176 } 2177 2178 ctx->username = kstrndup(payload, len, GFP_KERNEL); 2179 if (!ctx->username) { 2180 cifs_dbg(FYI, "Unable to allocate %zd bytes for username\n", 2181 len); 2182 rc = -ENOMEM; 2183 goto out_key_put; 2184 } 2185 cifs_dbg(FYI, "%s: username=%s\n", __func__, ctx->username); 2186 2187 len = key->datalen - (len + 1); 2188 if (len > CIFS_MAX_PASSWORD_LEN || len <= 0) { 2189 cifs_dbg(FYI, "Bad len for password search (len=%zd)\n", len); 2190 rc = -EINVAL; 2191 kfree(ctx->username); 2192 ctx->username = NULL; 2193 goto out_key_put; 2194 } 2195 2196 ++delim; 2197 /* BB consider adding support for password2 (Key Rotation) for multiuser in future */ 2198 ctx->password = kstrndup(delim, len, GFP_KERNEL); 2199 if (!ctx->password) { 2200 cifs_dbg(FYI, "Unable to allocate %zd bytes for password\n", 2201 len); 2202 rc = -ENOMEM; 2203 kfree(ctx->username); 2204 ctx->username = NULL; 2205 goto out_key_put; 2206 } 2207 2208 /* 2209 * If we have a domain key then we must set the domainName in the 2210 * for the request. 2211 */ 2212 if (is_domain && ses->domainName) { 2213 ctx->domainname = kstrdup(ses->domainName, GFP_KERNEL); 2214 if (!ctx->domainname) { 2215 cifs_dbg(FYI, "Unable to allocate %zd bytes for domain\n", 2216 len); 2217 rc = -ENOMEM; 2218 kfree(ctx->username); 2219 ctx->username = NULL; 2220 kfree_sensitive(ctx->password); 2221 /* no need to free ctx->password2 since not allocated in this path */ 2222 ctx->password = NULL; 2223 goto out_key_put; 2224 } 2225 } 2226 2227 strscpy(ctx->workstation_name, ses->workstation_name, sizeof(ctx->workstation_name)); 2228 2229 out_key_put: 2230 up_read(&key->sem); 2231 key_put(key); 2232 out_err: 2233 kfree(desc); 2234 cifs_dbg(FYI, "%s: returning %d\n", __func__, rc); 2235 return rc; 2236 } 2237 #else /* ! CONFIG_KEYS */ 2238 static inline int 2239 cifs_set_cifscreds(struct smb3_fs_context *ctx __attribute__((unused)), 2240 struct cifs_ses *ses __attribute__((unused))) 2241 { 2242 return -ENOSYS; 2243 } 2244 #endif /* CONFIG_KEYS */ 2245 2246 /** 2247 * cifs_get_smb_ses - get a session matching @ctx data from @server 2248 * @server: server to setup the session to 2249 * @ctx: superblock configuration context to use to setup the session 2250 * 2251 * This function assumes it is being called from cifs_mount() where we 2252 * already got a server reference (server refcount +1). See 2253 * cifs_get_tcon() for refcount explanations. 2254 */ 2255 struct cifs_ses * 2256 cifs_get_smb_ses(struct TCP_Server_Info *server, struct smb3_fs_context *ctx) 2257 { 2258 int rc = 0; 2259 unsigned int xid; 2260 struct cifs_ses *ses; 2261 struct sockaddr_in *addr = (struct sockaddr_in *)&server->dstaddr; 2262 struct sockaddr_in6 *addr6 = (struct sockaddr_in6 *)&server->dstaddr; 2263 2264 xid = get_xid(); 2265 2266 ses = cifs_find_smb_ses(server, ctx); 2267 if (ses) { 2268 cifs_dbg(FYI, "Existing smb sess found (status=%d)\n", 2269 ses->ses_status); 2270 2271 spin_lock(&ses->chan_lock); 2272 if (cifs_chan_needs_reconnect(ses, server)) { 2273 spin_unlock(&ses->chan_lock); 2274 cifs_dbg(FYI, "Session needs reconnect\n"); 2275 2276 mutex_lock(&ses->session_mutex); 2277 rc = cifs_negotiate_protocol(xid, ses, server); 2278 if (rc) { 2279 mutex_unlock(&ses->session_mutex); 2280 /* problem -- put our ses reference */ 2281 cifs_put_smb_ses(ses); 2282 free_xid(xid); 2283 return ERR_PTR(rc); 2284 } 2285 2286 rc = cifs_setup_session(xid, ses, server, 2287 ctx->local_nls); 2288 if (rc) { 2289 mutex_unlock(&ses->session_mutex); 2290 /* problem -- put our reference */ 2291 cifs_put_smb_ses(ses); 2292 free_xid(xid); 2293 return ERR_PTR(rc); 2294 } 2295 mutex_unlock(&ses->session_mutex); 2296 2297 spin_lock(&ses->chan_lock); 2298 } 2299 spin_unlock(&ses->chan_lock); 2300 2301 /* existing SMB ses has a server reference already */ 2302 cifs_put_tcp_session(server, 0); 2303 free_xid(xid); 2304 return ses; 2305 } 2306 2307 rc = -ENOMEM; 2308 2309 cifs_dbg(FYI, "Existing smb sess not found\n"); 2310 ses = sesInfoAlloc(); 2311 if (ses == NULL) 2312 goto get_ses_fail; 2313 2314 /* new SMB session uses our server ref */ 2315 ses->server = server; 2316 if (server->dstaddr.ss_family == AF_INET6) 2317 sprintf(ses->ip_addr, "%pI6", &addr6->sin6_addr); 2318 else 2319 sprintf(ses->ip_addr, "%pI4", &addr->sin_addr); 2320 2321 if (ctx->username) { 2322 ses->user_name = kstrdup(ctx->username, GFP_KERNEL); 2323 if (!ses->user_name) 2324 goto get_ses_fail; 2325 } 2326 2327 /* ctx->password freed at unmount */ 2328 if (ctx->password) { 2329 ses->password = kstrdup(ctx->password, GFP_KERNEL); 2330 if (!ses->password) 2331 goto get_ses_fail; 2332 } 2333 /* ctx->password freed at unmount */ 2334 if (ctx->password2) { 2335 ses->password2 = kstrdup(ctx->password2, GFP_KERNEL); 2336 if (!ses->password2) 2337 goto get_ses_fail; 2338 } 2339 if (ctx->domainname) { 2340 ses->domainName = kstrdup(ctx->domainname, GFP_KERNEL); 2341 if (!ses->domainName) 2342 goto get_ses_fail; 2343 } 2344 2345 strscpy(ses->workstation_name, ctx->workstation_name, sizeof(ses->workstation_name)); 2346 2347 if (ctx->domainauto) 2348 ses->domainAuto = ctx->domainauto; 2349 ses->cred_uid = ctx->cred_uid; 2350 ses->linux_uid = ctx->linux_uid; 2351 2352 ses->sectype = ctx->sectype; 2353 ses->sign = ctx->sign; 2354 ses->local_nls = load_nls(ctx->local_nls->charset); 2355 2356 /* add server as first channel */ 2357 spin_lock(&ses->chan_lock); 2358 ses->chans[0].server = server; 2359 ses->chan_count = 1; 2360 ses->chan_max = ctx->multichannel ? ctx->max_channels:1; 2361 ses->chans_need_reconnect = 1; 2362 spin_unlock(&ses->chan_lock); 2363 2364 mutex_lock(&ses->session_mutex); 2365 rc = cifs_negotiate_protocol(xid, ses, server); 2366 if (!rc) 2367 rc = cifs_setup_session(xid, ses, server, ctx->local_nls); 2368 mutex_unlock(&ses->session_mutex); 2369 2370 /* each channel uses a different signing key */ 2371 spin_lock(&ses->chan_lock); 2372 memcpy(ses->chans[0].signkey, ses->smb3signingkey, 2373 sizeof(ses->smb3signingkey)); 2374 spin_unlock(&ses->chan_lock); 2375 2376 if (rc) 2377 goto get_ses_fail; 2378 2379 /* 2380 * success, put it on the list and add it as first channel 2381 * note: the session becomes active soon after this. So you'll 2382 * need to lock before changing something in the session. 2383 */ 2384 spin_lock(&cifs_tcp_ses_lock); 2385 if (ctx->dfs_root_ses) 2386 cifs_smb_ses_inc_refcount(ctx->dfs_root_ses); 2387 ses->dfs_root_ses = ctx->dfs_root_ses; 2388 list_add(&ses->smb_ses_list, &server->smb_ses_list); 2389 spin_unlock(&cifs_tcp_ses_lock); 2390 2391 cifs_setup_ipc(ses, ctx); 2392 2393 free_xid(xid); 2394 2395 return ses; 2396 2397 get_ses_fail: 2398 sesInfoFree(ses); 2399 free_xid(xid); 2400 return ERR_PTR(rc); 2401 } 2402 2403 /* this function must be called with tc_lock held */ 2404 static int match_tcon(struct cifs_tcon *tcon, struct smb3_fs_context *ctx) 2405 { 2406 struct TCP_Server_Info *server = tcon->ses->server; 2407 2408 if (tcon->status == TID_EXITING) 2409 return 0; 2410 2411 if (tcon->origin_fullpath) { 2412 if (!ctx->source || 2413 !dfs_src_pathname_equal(ctx->source, 2414 tcon->origin_fullpath)) 2415 return 0; 2416 } else if (!server->leaf_fullpath && 2417 strncmp(tcon->tree_name, ctx->UNC, MAX_TREE_SIZE)) { 2418 return 0; 2419 } 2420 if (tcon->seal != ctx->seal) 2421 return 0; 2422 if (tcon->snapshot_time != ctx->snapshot_time) 2423 return 0; 2424 if (tcon->handle_timeout != ctx->handle_timeout) 2425 return 0; 2426 if (tcon->no_lease != ctx->no_lease) 2427 return 0; 2428 if (tcon->nodelete != ctx->nodelete) 2429 return 0; 2430 return 1; 2431 } 2432 2433 static struct cifs_tcon * 2434 cifs_find_tcon(struct cifs_ses *ses, struct smb3_fs_context *ctx) 2435 { 2436 struct cifs_tcon *tcon; 2437 2438 spin_lock(&cifs_tcp_ses_lock); 2439 list_for_each_entry(tcon, &ses->tcon_list, tcon_list) { 2440 spin_lock(&tcon->tc_lock); 2441 if (!match_tcon(tcon, ctx)) { 2442 spin_unlock(&tcon->tc_lock); 2443 continue; 2444 } 2445 ++tcon->tc_count; 2446 trace_smb3_tcon_ref(tcon->debug_id, tcon->tc_count, 2447 netfs_trace_tcon_ref_get_find); 2448 spin_unlock(&tcon->tc_lock); 2449 spin_unlock(&cifs_tcp_ses_lock); 2450 return tcon; 2451 } 2452 spin_unlock(&cifs_tcp_ses_lock); 2453 return NULL; 2454 } 2455 2456 void 2457 cifs_put_tcon(struct cifs_tcon *tcon, enum smb3_tcon_ref_trace trace) 2458 { 2459 unsigned int xid; 2460 struct cifs_ses *ses; 2461 2462 /* 2463 * IPC tcon share the lifetime of their session and are 2464 * destroyed in the session put function 2465 */ 2466 if (tcon == NULL || tcon->ipc) 2467 return; 2468 2469 ses = tcon->ses; 2470 cifs_dbg(FYI, "%s: tc_count=%d\n", __func__, tcon->tc_count); 2471 spin_lock(&cifs_tcp_ses_lock); 2472 spin_lock(&tcon->tc_lock); 2473 trace_smb3_tcon_ref(tcon->debug_id, tcon->tc_count - 1, trace); 2474 if (--tcon->tc_count > 0) { 2475 spin_unlock(&tcon->tc_lock); 2476 spin_unlock(&cifs_tcp_ses_lock); 2477 return; 2478 } 2479 2480 /* tc_count can never go negative */ 2481 WARN_ON(tcon->tc_count < 0); 2482 2483 list_del_init(&tcon->tcon_list); 2484 tcon->status = TID_EXITING; 2485 spin_unlock(&tcon->tc_lock); 2486 spin_unlock(&cifs_tcp_ses_lock); 2487 2488 /* cancel polling of interfaces */ 2489 cancel_delayed_work_sync(&tcon->query_interfaces); 2490 #ifdef CONFIG_CIFS_DFS_UPCALL 2491 cancel_delayed_work_sync(&tcon->dfs_cache_work); 2492 #endif 2493 2494 if (tcon->use_witness) { 2495 int rc; 2496 2497 rc = cifs_swn_unregister(tcon); 2498 if (rc < 0) { 2499 cifs_dbg(VFS, "%s: Failed to unregister for witness notifications: %d\n", 2500 __func__, rc); 2501 } 2502 } 2503 2504 xid = get_xid(); 2505 if (ses->server->ops->tree_disconnect) 2506 ses->server->ops->tree_disconnect(xid, tcon); 2507 _free_xid(xid); 2508 2509 cifs_fscache_release_super_cookie(tcon); 2510 tconInfoFree(tcon, netfs_trace_tcon_ref_free); 2511 cifs_put_smb_ses(ses); 2512 } 2513 2514 /** 2515 * cifs_get_tcon - get a tcon matching @ctx data from @ses 2516 * @ses: smb session to issue the request on 2517 * @ctx: the superblock configuration context to use for building the 2518 * 2519 * - tcon refcount is the number of mount points using the tcon. 2520 * - ses refcount is the number of tcon using the session. 2521 * 2522 * 1. This function assumes it is being called from cifs_mount() where 2523 * we already got a session reference (ses refcount +1). 2524 * 2525 * 2. Since we're in the context of adding a mount point, the end 2526 * result should be either: 2527 * 2528 * a) a new tcon already allocated with refcount=1 (1 mount point) and 2529 * its session refcount incremented (1 new tcon). This +1 was 2530 * already done in (1). 2531 * 2532 * b) an existing tcon with refcount+1 (add a mount point to it) and 2533 * identical ses refcount (no new tcon). Because of (1) we need to 2534 * decrement the ses refcount. 2535 */ 2536 static struct cifs_tcon * 2537 cifs_get_tcon(struct cifs_ses *ses, struct smb3_fs_context *ctx) 2538 { 2539 struct cifs_tcon *tcon; 2540 bool nohandlecache; 2541 int rc, xid; 2542 2543 tcon = cifs_find_tcon(ses, ctx); 2544 if (tcon) { 2545 /* 2546 * tcon has refcount already incremented but we need to 2547 * decrement extra ses reference gotten by caller (case b) 2548 */ 2549 cifs_dbg(FYI, "Found match on UNC path\n"); 2550 cifs_put_smb_ses(ses); 2551 return tcon; 2552 } 2553 2554 if (!ses->server->ops->tree_connect) { 2555 rc = -ENOSYS; 2556 goto out_fail; 2557 } 2558 2559 if (ses->server->dialect >= SMB20_PROT_ID && 2560 (ses->server->capabilities & SMB2_GLOBAL_CAP_DIRECTORY_LEASING)) 2561 nohandlecache = ctx->nohandlecache; 2562 else 2563 nohandlecache = true; 2564 tcon = tcon_info_alloc(!nohandlecache, netfs_trace_tcon_ref_new); 2565 if (tcon == NULL) { 2566 rc = -ENOMEM; 2567 goto out_fail; 2568 } 2569 tcon->nohandlecache = nohandlecache; 2570 2571 if (ctx->snapshot_time) { 2572 if (ses->server->vals->protocol_id == 0) { 2573 cifs_dbg(VFS, 2574 "Use SMB2 or later for snapshot mount option\n"); 2575 rc = -EOPNOTSUPP; 2576 goto out_fail; 2577 } else 2578 tcon->snapshot_time = ctx->snapshot_time; 2579 } 2580 2581 if (ctx->handle_timeout) { 2582 if (ses->server->vals->protocol_id == 0) { 2583 cifs_dbg(VFS, 2584 "Use SMB2.1 or later for handle timeout option\n"); 2585 rc = -EOPNOTSUPP; 2586 goto out_fail; 2587 } else 2588 tcon->handle_timeout = ctx->handle_timeout; 2589 } 2590 2591 tcon->ses = ses; 2592 if (ctx->password) { 2593 tcon->password = kstrdup(ctx->password, GFP_KERNEL); 2594 if (!tcon->password) { 2595 rc = -ENOMEM; 2596 goto out_fail; 2597 } 2598 } 2599 2600 if (ctx->seal) { 2601 if (ses->server->vals->protocol_id == 0) { 2602 cifs_dbg(VFS, 2603 "SMB3 or later required for encryption\n"); 2604 rc = -EOPNOTSUPP; 2605 goto out_fail; 2606 } else if (tcon->ses->server->capabilities & 2607 SMB2_GLOBAL_CAP_ENCRYPTION) 2608 tcon->seal = true; 2609 else { 2610 cifs_dbg(VFS, "Encryption is not supported on share\n"); 2611 rc = -EOPNOTSUPP; 2612 goto out_fail; 2613 } 2614 } 2615 2616 if (ctx->linux_ext) { 2617 if (ses->server->posix_ext_supported) { 2618 tcon->posix_extensions = true; 2619 pr_warn_once("SMB3.11 POSIX Extensions are experimental\n"); 2620 } else if ((ses->server->vals->protocol_id == SMB311_PROT_ID) || 2621 (strcmp(ses->server->vals->version_string, 2622 SMB3ANY_VERSION_STRING) == 0) || 2623 (strcmp(ses->server->vals->version_string, 2624 SMBDEFAULT_VERSION_STRING) == 0)) { 2625 cifs_dbg(VFS, "Server does not support mounting with posix SMB3.11 extensions\n"); 2626 rc = -EOPNOTSUPP; 2627 goto out_fail; 2628 } else if (ses->server->vals->protocol_id == SMB10_PROT_ID) 2629 if (cap_unix(ses)) 2630 cifs_dbg(FYI, "Unix Extensions requested on SMB1 mount\n"); 2631 else { 2632 cifs_dbg(VFS, "SMB1 Unix Extensions not supported by server\n"); 2633 rc = -EOPNOTSUPP; 2634 goto out_fail; 2635 } else { 2636 cifs_dbg(VFS, 2637 "Check vers= mount option. SMB3.11 disabled but required for POSIX extensions\n"); 2638 rc = -EOPNOTSUPP; 2639 goto out_fail; 2640 } 2641 } 2642 2643 xid = get_xid(); 2644 rc = ses->server->ops->tree_connect(xid, ses, ctx->UNC, tcon, 2645 ctx->local_nls); 2646 free_xid(xid); 2647 cifs_dbg(FYI, "Tcon rc = %d\n", rc); 2648 if (rc) 2649 goto out_fail; 2650 2651 tcon->use_persistent = false; 2652 /* check if SMB2 or later, CIFS does not support persistent handles */ 2653 if (ctx->persistent) { 2654 if (ses->server->vals->protocol_id == 0) { 2655 cifs_dbg(VFS, 2656 "SMB3 or later required for persistent handles\n"); 2657 rc = -EOPNOTSUPP; 2658 goto out_fail; 2659 } else if (ses->server->capabilities & 2660 SMB2_GLOBAL_CAP_PERSISTENT_HANDLES) 2661 tcon->use_persistent = true; 2662 else /* persistent handles requested but not supported */ { 2663 cifs_dbg(VFS, 2664 "Persistent handles not supported on share\n"); 2665 rc = -EOPNOTSUPP; 2666 goto out_fail; 2667 } 2668 } else if ((tcon->capabilities & SMB2_SHARE_CAP_CONTINUOUS_AVAILABILITY) 2669 && (ses->server->capabilities & SMB2_GLOBAL_CAP_PERSISTENT_HANDLES) 2670 && (ctx->nopersistent == false)) { 2671 cifs_dbg(FYI, "enabling persistent handles\n"); 2672 tcon->use_persistent = true; 2673 } else if (ctx->resilient) { 2674 if (ses->server->vals->protocol_id == 0) { 2675 cifs_dbg(VFS, 2676 "SMB2.1 or later required for resilient handles\n"); 2677 rc = -EOPNOTSUPP; 2678 goto out_fail; 2679 } 2680 tcon->use_resilient = true; 2681 } 2682 2683 tcon->use_witness = false; 2684 if (IS_ENABLED(CONFIG_CIFS_SWN_UPCALL) && ctx->witness) { 2685 if (ses->server->vals->protocol_id >= SMB30_PROT_ID) { 2686 if (tcon->capabilities & SMB2_SHARE_CAP_CLUSTER) { 2687 /* 2688 * Set witness in use flag in first place 2689 * to retry registration in the echo task 2690 */ 2691 tcon->use_witness = true; 2692 /* And try to register immediately */ 2693 rc = cifs_swn_register(tcon); 2694 if (rc < 0) { 2695 cifs_dbg(VFS, "Failed to register for witness notifications: %d\n", rc); 2696 goto out_fail; 2697 } 2698 } else { 2699 /* TODO: try to extend for non-cluster uses (eg multichannel) */ 2700 cifs_dbg(VFS, "witness requested on mount but no CLUSTER capability on share\n"); 2701 rc = -EOPNOTSUPP; 2702 goto out_fail; 2703 } 2704 } else { 2705 cifs_dbg(VFS, "SMB3 or later required for witness option\n"); 2706 rc = -EOPNOTSUPP; 2707 goto out_fail; 2708 } 2709 } 2710 2711 /* If the user really knows what they are doing they can override */ 2712 if (tcon->share_flags & SMB2_SHAREFLAG_NO_CACHING) { 2713 if (ctx->cache_ro) 2714 cifs_dbg(VFS, "cache=ro requested on mount but NO_CACHING flag set on share\n"); 2715 else if (ctx->cache_rw) 2716 cifs_dbg(VFS, "cache=singleclient requested on mount but NO_CACHING flag set on share\n"); 2717 } 2718 2719 if (ctx->no_lease) { 2720 if (ses->server->vals->protocol_id == 0) { 2721 cifs_dbg(VFS, 2722 "SMB2 or later required for nolease option\n"); 2723 rc = -EOPNOTSUPP; 2724 goto out_fail; 2725 } else 2726 tcon->no_lease = ctx->no_lease; 2727 } 2728 2729 /* 2730 * We can have only one retry value for a connection to a share so for 2731 * resources mounted more than once to the same server share the last 2732 * value passed in for the retry flag is used. 2733 */ 2734 tcon->retry = ctx->retry; 2735 tcon->nocase = ctx->nocase; 2736 tcon->broken_sparse_sup = ctx->no_sparse; 2737 tcon->max_cached_dirs = ctx->max_cached_dirs; 2738 tcon->nodelete = ctx->nodelete; 2739 tcon->local_lease = ctx->local_lease; 2740 INIT_LIST_HEAD(&tcon->pending_opens); 2741 tcon->status = TID_GOOD; 2742 2743 INIT_DELAYED_WORK(&tcon->query_interfaces, 2744 smb2_query_server_interfaces); 2745 if (ses->server->dialect >= SMB30_PROT_ID && 2746 (ses->server->capabilities & SMB2_GLOBAL_CAP_MULTI_CHANNEL)) { 2747 /* schedule query interfaces poll */ 2748 queue_delayed_work(cifsiod_wq, &tcon->query_interfaces, 2749 (SMB_INTERFACE_POLL_INTERVAL * HZ)); 2750 } 2751 #ifdef CONFIG_CIFS_DFS_UPCALL 2752 INIT_DELAYED_WORK(&tcon->dfs_cache_work, dfs_cache_refresh); 2753 #endif 2754 spin_lock(&cifs_tcp_ses_lock); 2755 list_add(&tcon->tcon_list, &ses->tcon_list); 2756 spin_unlock(&cifs_tcp_ses_lock); 2757 2758 return tcon; 2759 2760 out_fail: 2761 tconInfoFree(tcon, netfs_trace_tcon_ref_free_fail); 2762 return ERR_PTR(rc); 2763 } 2764 2765 void 2766 cifs_put_tlink(struct tcon_link *tlink) 2767 { 2768 if (!tlink || IS_ERR(tlink)) 2769 return; 2770 2771 if (!atomic_dec_and_test(&tlink->tl_count) || 2772 test_bit(TCON_LINK_IN_TREE, &tlink->tl_flags)) { 2773 tlink->tl_time = jiffies; 2774 return; 2775 } 2776 2777 if (!IS_ERR(tlink_tcon(tlink))) 2778 cifs_put_tcon(tlink_tcon(tlink), netfs_trace_tcon_ref_put_tlink); 2779 kfree(tlink); 2780 } 2781 2782 static int 2783 compare_mount_options(struct super_block *sb, struct cifs_mnt_data *mnt_data) 2784 { 2785 struct cifs_sb_info *old = CIFS_SB(sb); 2786 struct cifs_sb_info *new = mnt_data->cifs_sb; 2787 unsigned int oldflags = old->mnt_cifs_flags & CIFS_MOUNT_MASK; 2788 unsigned int newflags = new->mnt_cifs_flags & CIFS_MOUNT_MASK; 2789 2790 if ((sb->s_flags & CIFS_MS_MASK) != (mnt_data->flags & CIFS_MS_MASK)) 2791 return 0; 2792 2793 if (old->mnt_cifs_serverino_autodisabled) 2794 newflags &= ~CIFS_MOUNT_SERVER_INUM; 2795 2796 if (oldflags != newflags) 2797 return 0; 2798 2799 /* 2800 * We want to share sb only if we don't specify an r/wsize or 2801 * specified r/wsize is greater than or equal to existing one. 2802 */ 2803 if (new->ctx->wsize && new->ctx->wsize < old->ctx->wsize) 2804 return 0; 2805 2806 if (new->ctx->rsize && new->ctx->rsize < old->ctx->rsize) 2807 return 0; 2808 2809 if (!uid_eq(old->ctx->linux_uid, new->ctx->linux_uid) || 2810 !gid_eq(old->ctx->linux_gid, new->ctx->linux_gid)) 2811 return 0; 2812 2813 if (old->ctx->file_mode != new->ctx->file_mode || 2814 old->ctx->dir_mode != new->ctx->dir_mode) 2815 return 0; 2816 2817 if (strcmp(old->local_nls->charset, new->local_nls->charset)) 2818 return 0; 2819 2820 if (old->ctx->acregmax != new->ctx->acregmax) 2821 return 0; 2822 if (old->ctx->acdirmax != new->ctx->acdirmax) 2823 return 0; 2824 if (old->ctx->closetimeo != new->ctx->closetimeo) 2825 return 0; 2826 if (old->ctx->reparse_type != new->ctx->reparse_type) 2827 return 0; 2828 2829 return 1; 2830 } 2831 2832 static int match_prepath(struct super_block *sb, 2833 struct cifs_tcon *tcon, 2834 struct cifs_mnt_data *mnt_data) 2835 { 2836 struct smb3_fs_context *ctx = mnt_data->ctx; 2837 struct cifs_sb_info *old = CIFS_SB(sb); 2838 struct cifs_sb_info *new = mnt_data->cifs_sb; 2839 bool old_set = (old->mnt_cifs_flags & CIFS_MOUNT_USE_PREFIX_PATH) && 2840 old->prepath; 2841 bool new_set = (new->mnt_cifs_flags & CIFS_MOUNT_USE_PREFIX_PATH) && 2842 new->prepath; 2843 2844 if (tcon->origin_fullpath && 2845 dfs_src_pathname_equal(tcon->origin_fullpath, ctx->source)) 2846 return 1; 2847 2848 if (old_set && new_set && !strcmp(new->prepath, old->prepath)) 2849 return 1; 2850 else if (!old_set && !new_set) 2851 return 1; 2852 2853 return 0; 2854 } 2855 2856 int 2857 cifs_match_super(struct super_block *sb, void *data) 2858 { 2859 struct cifs_mnt_data *mnt_data = data; 2860 struct smb3_fs_context *ctx; 2861 struct cifs_sb_info *cifs_sb; 2862 struct TCP_Server_Info *tcp_srv; 2863 struct cifs_ses *ses; 2864 struct cifs_tcon *tcon; 2865 struct tcon_link *tlink; 2866 int rc = 0; 2867 2868 spin_lock(&cifs_tcp_ses_lock); 2869 cifs_sb = CIFS_SB(sb); 2870 2871 /* We do not want to use a superblock that has been shutdown */ 2872 if (CIFS_MOUNT_SHUTDOWN & cifs_sb->mnt_cifs_flags) { 2873 spin_unlock(&cifs_tcp_ses_lock); 2874 return 0; 2875 } 2876 2877 tlink = cifs_get_tlink(cifs_sb_master_tlink(cifs_sb)); 2878 if (IS_ERR_OR_NULL(tlink)) { 2879 pr_warn_once("%s: skip super matching due to bad tlink(%p)\n", 2880 __func__, tlink); 2881 spin_unlock(&cifs_tcp_ses_lock); 2882 return 0; 2883 } 2884 tcon = tlink_tcon(tlink); 2885 ses = tcon->ses; 2886 tcp_srv = ses->server; 2887 2888 ctx = mnt_data->ctx; 2889 2890 spin_lock(&tcp_srv->srv_lock); 2891 spin_lock(&ses->ses_lock); 2892 spin_lock(&ses->chan_lock); 2893 spin_lock(&tcon->tc_lock); 2894 if (!match_server(tcp_srv, ctx, true) || 2895 !match_session(ses, ctx) || 2896 !match_tcon(tcon, ctx) || 2897 !match_prepath(sb, tcon, mnt_data)) { 2898 rc = 0; 2899 goto out; 2900 } 2901 2902 rc = compare_mount_options(sb, mnt_data); 2903 out: 2904 spin_unlock(&tcon->tc_lock); 2905 spin_unlock(&ses->chan_lock); 2906 spin_unlock(&ses->ses_lock); 2907 spin_unlock(&tcp_srv->srv_lock); 2908 2909 spin_unlock(&cifs_tcp_ses_lock); 2910 cifs_put_tlink(tlink); 2911 return rc; 2912 } 2913 2914 #ifdef CONFIG_DEBUG_LOCK_ALLOC 2915 static struct lock_class_key cifs_key[2]; 2916 static struct lock_class_key cifs_slock_key[2]; 2917 2918 static inline void 2919 cifs_reclassify_socket4(struct socket *sock) 2920 { 2921 struct sock *sk = sock->sk; 2922 2923 BUG_ON(!sock_allow_reclassification(sk)); 2924 sock_lock_init_class_and_name(sk, "slock-AF_INET-CIFS", 2925 &cifs_slock_key[0], "sk_lock-AF_INET-CIFS", &cifs_key[0]); 2926 } 2927 2928 static inline void 2929 cifs_reclassify_socket6(struct socket *sock) 2930 { 2931 struct sock *sk = sock->sk; 2932 2933 BUG_ON(!sock_allow_reclassification(sk)); 2934 sock_lock_init_class_and_name(sk, "slock-AF_INET6-CIFS", 2935 &cifs_slock_key[1], "sk_lock-AF_INET6-CIFS", &cifs_key[1]); 2936 } 2937 #else 2938 static inline void 2939 cifs_reclassify_socket4(struct socket *sock) 2940 { 2941 } 2942 2943 static inline void 2944 cifs_reclassify_socket6(struct socket *sock) 2945 { 2946 } 2947 #endif 2948 2949 /* See RFC1001 section 14 on representation of Netbios names */ 2950 static void rfc1002mangle(char *target, char *source, unsigned int length) 2951 { 2952 unsigned int i, j; 2953 2954 for (i = 0, j = 0; i < (length); i++) { 2955 /* mask a nibble at a time and encode */ 2956 target[j] = 'A' + (0x0F & (source[i] >> 4)); 2957 target[j+1] = 'A' + (0x0F & source[i]); 2958 j += 2; 2959 } 2960 2961 } 2962 2963 static int 2964 bind_socket(struct TCP_Server_Info *server) 2965 { 2966 int rc = 0; 2967 2968 if (server->srcaddr.ss_family != AF_UNSPEC) { 2969 /* Bind to the specified local IP address */ 2970 struct socket *socket = server->ssocket; 2971 2972 rc = kernel_bind(socket, 2973 (struct sockaddr *) &server->srcaddr, 2974 sizeof(server->srcaddr)); 2975 if (rc < 0) { 2976 struct sockaddr_in *saddr4; 2977 struct sockaddr_in6 *saddr6; 2978 2979 saddr4 = (struct sockaddr_in *)&server->srcaddr; 2980 saddr6 = (struct sockaddr_in6 *)&server->srcaddr; 2981 if (saddr6->sin6_family == AF_INET6) 2982 cifs_server_dbg(VFS, "Failed to bind to: %pI6c, error: %d\n", 2983 &saddr6->sin6_addr, rc); 2984 else 2985 cifs_server_dbg(VFS, "Failed to bind to: %pI4, error: %d\n", 2986 &saddr4->sin_addr.s_addr, rc); 2987 } 2988 } 2989 return rc; 2990 } 2991 2992 static int 2993 ip_rfc1001_connect(struct TCP_Server_Info *server) 2994 { 2995 int rc = 0; 2996 /* 2997 * some servers require RFC1001 sessinit before sending 2998 * negprot - BB check reconnection in case where second 2999 * sessinit is sent but no second negprot 3000 */ 3001 struct rfc1002_session_packet req = {}; 3002 struct smb_hdr *smb_buf = (struct smb_hdr *)&req; 3003 unsigned int len; 3004 3005 req.trailer.session_req.called_len = sizeof(req.trailer.session_req.called_name); 3006 3007 if (server->server_RFC1001_name[0] != 0) 3008 rfc1002mangle(req.trailer.session_req.called_name, 3009 server->server_RFC1001_name, 3010 RFC1001_NAME_LEN_WITH_NULL); 3011 else 3012 rfc1002mangle(req.trailer.session_req.called_name, 3013 DEFAULT_CIFS_CALLED_NAME, 3014 RFC1001_NAME_LEN_WITH_NULL); 3015 3016 req.trailer.session_req.calling_len = sizeof(req.trailer.session_req.calling_name); 3017 3018 /* calling name ends in null (byte 16) from old smb convention */ 3019 if (server->workstation_RFC1001_name[0] != 0) 3020 rfc1002mangle(req.trailer.session_req.calling_name, 3021 server->workstation_RFC1001_name, 3022 RFC1001_NAME_LEN_WITH_NULL); 3023 else 3024 rfc1002mangle(req.trailer.session_req.calling_name, 3025 "LINUX_CIFS_CLNT", 3026 RFC1001_NAME_LEN_WITH_NULL); 3027 3028 /* 3029 * As per rfc1002, @len must be the number of bytes that follows the 3030 * length field of a rfc1002 session request payload. 3031 */ 3032 len = sizeof(req) - offsetof(struct rfc1002_session_packet, trailer.session_req); 3033 3034 smb_buf->smb_buf_length = cpu_to_be32((RFC1002_SESSION_REQUEST << 24) | len); 3035 rc = smb_send(server, smb_buf, len); 3036 /* 3037 * RFC1001 layer in at least one server requires very short break before 3038 * negprot presumably because not expecting negprot to follow so fast. 3039 * This is a simple solution that works without complicating the code 3040 * and causes no significant slowing down on mount for everyone else 3041 */ 3042 usleep_range(1000, 2000); 3043 3044 return rc; 3045 } 3046 3047 static int 3048 generic_ip_connect(struct TCP_Server_Info *server) 3049 { 3050 struct sockaddr *saddr; 3051 struct socket *socket; 3052 int slen, sfamily; 3053 __be16 sport; 3054 int rc = 0; 3055 3056 saddr = (struct sockaddr *) &server->dstaddr; 3057 3058 if (server->dstaddr.ss_family == AF_INET6) { 3059 struct sockaddr_in6 *ipv6 = (struct sockaddr_in6 *)&server->dstaddr; 3060 3061 sport = ipv6->sin6_port; 3062 slen = sizeof(struct sockaddr_in6); 3063 sfamily = AF_INET6; 3064 cifs_dbg(FYI, "%s: connecting to [%pI6]:%d\n", __func__, &ipv6->sin6_addr, 3065 ntohs(sport)); 3066 } else { 3067 struct sockaddr_in *ipv4 = (struct sockaddr_in *)&server->dstaddr; 3068 3069 sport = ipv4->sin_port; 3070 slen = sizeof(struct sockaddr_in); 3071 sfamily = AF_INET; 3072 cifs_dbg(FYI, "%s: connecting to %pI4:%d\n", __func__, &ipv4->sin_addr, 3073 ntohs(sport)); 3074 } 3075 3076 if (server->ssocket) { 3077 socket = server->ssocket; 3078 } else { 3079 struct net *net = cifs_net_ns(server); 3080 struct sock *sk; 3081 3082 rc = __sock_create(net, sfamily, SOCK_STREAM, 3083 IPPROTO_TCP, &server->ssocket, 1); 3084 if (rc < 0) { 3085 cifs_server_dbg(VFS, "Error %d creating socket\n", rc); 3086 return rc; 3087 } 3088 3089 sk = server->ssocket->sk; 3090 __netns_tracker_free(net, &sk->ns_tracker, false); 3091 sk->sk_net_refcnt = 1; 3092 get_net_track(net, &sk->ns_tracker, GFP_KERNEL); 3093 sock_inuse_add(net, 1); 3094 3095 /* BB other socket options to set KEEPALIVE, NODELAY? */ 3096 cifs_dbg(FYI, "Socket created\n"); 3097 socket = server->ssocket; 3098 socket->sk->sk_allocation = GFP_NOFS; 3099 socket->sk->sk_use_task_frag = false; 3100 if (sfamily == AF_INET6) 3101 cifs_reclassify_socket6(socket); 3102 else 3103 cifs_reclassify_socket4(socket); 3104 } 3105 3106 rc = bind_socket(server); 3107 if (rc < 0) 3108 return rc; 3109 3110 /* 3111 * Eventually check for other socket options to change from 3112 * the default. sock_setsockopt not used because it expects 3113 * user space buffer 3114 */ 3115 socket->sk->sk_rcvtimeo = 7 * HZ; 3116 socket->sk->sk_sndtimeo = 5 * HZ; 3117 3118 /* make the bufsizes depend on wsize/rsize and max requests */ 3119 if (server->noautotune) { 3120 if (socket->sk->sk_sndbuf < (200 * 1024)) 3121 socket->sk->sk_sndbuf = 200 * 1024; 3122 if (socket->sk->sk_rcvbuf < (140 * 1024)) 3123 socket->sk->sk_rcvbuf = 140 * 1024; 3124 } 3125 3126 if (server->tcp_nodelay) 3127 tcp_sock_set_nodelay(socket->sk); 3128 3129 cifs_dbg(FYI, "sndbuf %d rcvbuf %d rcvtimeo 0x%lx\n", 3130 socket->sk->sk_sndbuf, 3131 socket->sk->sk_rcvbuf, socket->sk->sk_rcvtimeo); 3132 3133 rc = kernel_connect(socket, saddr, slen, 3134 server->noblockcnt ? O_NONBLOCK : 0); 3135 /* 3136 * When mounting SMB root file systems, we do not want to block in 3137 * connect. Otherwise bail out and then let cifs_reconnect() perform 3138 * reconnect failover - if possible. 3139 */ 3140 if (server->noblockcnt && rc == -EINPROGRESS) 3141 rc = 0; 3142 if (rc < 0) { 3143 cifs_dbg(FYI, "Error %d connecting to server\n", rc); 3144 trace_smb3_connect_err(server->hostname, server->conn_id, &server->dstaddr, rc); 3145 sock_release(socket); 3146 server->ssocket = NULL; 3147 return rc; 3148 } 3149 trace_smb3_connect_done(server->hostname, server->conn_id, &server->dstaddr); 3150 if (sport == htons(RFC1001_PORT)) 3151 rc = ip_rfc1001_connect(server); 3152 3153 return rc; 3154 } 3155 3156 static int 3157 ip_connect(struct TCP_Server_Info *server) 3158 { 3159 __be16 *sport; 3160 struct sockaddr_in6 *addr6 = (struct sockaddr_in6 *)&server->dstaddr; 3161 struct sockaddr_in *addr = (struct sockaddr_in *)&server->dstaddr; 3162 3163 if (server->dstaddr.ss_family == AF_INET6) 3164 sport = &addr6->sin6_port; 3165 else 3166 sport = &addr->sin_port; 3167 3168 if (*sport == 0) { 3169 int rc; 3170 3171 /* try with 445 port at first */ 3172 *sport = htons(CIFS_PORT); 3173 3174 rc = generic_ip_connect(server); 3175 if (rc >= 0) 3176 return rc; 3177 3178 /* if it failed, try with 139 port */ 3179 *sport = htons(RFC1001_PORT); 3180 } 3181 3182 return generic_ip_connect(server); 3183 } 3184 3185 #ifdef CONFIG_CIFS_ALLOW_INSECURE_LEGACY 3186 void reset_cifs_unix_caps(unsigned int xid, struct cifs_tcon *tcon, 3187 struct cifs_sb_info *cifs_sb, struct smb3_fs_context *ctx) 3188 { 3189 /* 3190 * If we are reconnecting then should we check to see if 3191 * any requested capabilities changed locally e.g. via 3192 * remount but we can not do much about it here 3193 * if they have (even if we could detect it by the following) 3194 * Perhaps we could add a backpointer to array of sb from tcon 3195 * or if we change to make all sb to same share the same 3196 * sb as NFS - then we only have one backpointer to sb. 3197 * What if we wanted to mount the server share twice once with 3198 * and once without posixacls or posix paths? 3199 */ 3200 __u64 saved_cap = le64_to_cpu(tcon->fsUnixInfo.Capability); 3201 3202 if (ctx && ctx->no_linux_ext) { 3203 tcon->fsUnixInfo.Capability = 0; 3204 tcon->unix_ext = 0; /* Unix Extensions disabled */ 3205 cifs_dbg(FYI, "Linux protocol extensions disabled\n"); 3206 return; 3207 } else if (ctx) 3208 tcon->unix_ext = 1; /* Unix Extensions supported */ 3209 3210 if (!tcon->unix_ext) { 3211 cifs_dbg(FYI, "Unix extensions disabled so not set on reconnect\n"); 3212 return; 3213 } 3214 3215 if (!CIFSSMBQFSUnixInfo(xid, tcon)) { 3216 __u64 cap = le64_to_cpu(tcon->fsUnixInfo.Capability); 3217 3218 cifs_dbg(FYI, "unix caps which server supports %lld\n", cap); 3219 /* 3220 * check for reconnect case in which we do not 3221 * want to change the mount behavior if we can avoid it 3222 */ 3223 if (ctx == NULL) { 3224 /* 3225 * turn off POSIX ACL and PATHNAMES if not set 3226 * originally at mount time 3227 */ 3228 if ((saved_cap & CIFS_UNIX_POSIX_ACL_CAP) == 0) 3229 cap &= ~CIFS_UNIX_POSIX_ACL_CAP; 3230 if ((saved_cap & CIFS_UNIX_POSIX_PATHNAMES_CAP) == 0) { 3231 if (cap & CIFS_UNIX_POSIX_PATHNAMES_CAP) 3232 cifs_dbg(VFS, "POSIXPATH support change\n"); 3233 cap &= ~CIFS_UNIX_POSIX_PATHNAMES_CAP; 3234 } else if ((cap & CIFS_UNIX_POSIX_PATHNAMES_CAP) == 0) { 3235 cifs_dbg(VFS, "possible reconnect error\n"); 3236 cifs_dbg(VFS, "server disabled POSIX path support\n"); 3237 } 3238 } 3239 3240 if (cap & CIFS_UNIX_TRANSPORT_ENCRYPTION_MANDATORY_CAP) 3241 cifs_dbg(VFS, "per-share encryption not supported yet\n"); 3242 3243 cap &= CIFS_UNIX_CAP_MASK; 3244 if (ctx && ctx->no_psx_acl) 3245 cap &= ~CIFS_UNIX_POSIX_ACL_CAP; 3246 else if (CIFS_UNIX_POSIX_ACL_CAP & cap) { 3247 cifs_dbg(FYI, "negotiated posix acl support\n"); 3248 if (cifs_sb) 3249 cifs_sb->mnt_cifs_flags |= 3250 CIFS_MOUNT_POSIXACL; 3251 } 3252 3253 if (ctx && ctx->posix_paths == 0) 3254 cap &= ~CIFS_UNIX_POSIX_PATHNAMES_CAP; 3255 else if (cap & CIFS_UNIX_POSIX_PATHNAMES_CAP) { 3256 cifs_dbg(FYI, "negotiate posix pathnames\n"); 3257 if (cifs_sb) 3258 cifs_sb->mnt_cifs_flags |= 3259 CIFS_MOUNT_POSIX_PATHS; 3260 } 3261 3262 cifs_dbg(FYI, "Negotiate caps 0x%x\n", (int)cap); 3263 #ifdef CONFIG_CIFS_DEBUG2 3264 if (cap & CIFS_UNIX_FCNTL_CAP) 3265 cifs_dbg(FYI, "FCNTL cap\n"); 3266 if (cap & CIFS_UNIX_EXTATTR_CAP) 3267 cifs_dbg(FYI, "EXTATTR cap\n"); 3268 if (cap & CIFS_UNIX_POSIX_PATHNAMES_CAP) 3269 cifs_dbg(FYI, "POSIX path cap\n"); 3270 if (cap & CIFS_UNIX_XATTR_CAP) 3271 cifs_dbg(FYI, "XATTR cap\n"); 3272 if (cap & CIFS_UNIX_POSIX_ACL_CAP) 3273 cifs_dbg(FYI, "POSIX ACL cap\n"); 3274 if (cap & CIFS_UNIX_LARGE_READ_CAP) 3275 cifs_dbg(FYI, "very large read cap\n"); 3276 if (cap & CIFS_UNIX_LARGE_WRITE_CAP) 3277 cifs_dbg(FYI, "very large write cap\n"); 3278 if (cap & CIFS_UNIX_TRANSPORT_ENCRYPTION_CAP) 3279 cifs_dbg(FYI, "transport encryption cap\n"); 3280 if (cap & CIFS_UNIX_TRANSPORT_ENCRYPTION_MANDATORY_CAP) 3281 cifs_dbg(FYI, "mandatory transport encryption cap\n"); 3282 #endif /* CIFS_DEBUG2 */ 3283 if (CIFSSMBSetFSUnixInfo(xid, tcon, cap)) { 3284 if (ctx == NULL) 3285 cifs_dbg(FYI, "resetting capabilities failed\n"); 3286 else 3287 cifs_dbg(VFS, "Negotiating Unix capabilities with the server failed. Consider mounting with the Unix Extensions disabled if problems are found by specifying the nounix mount option.\n"); 3288 3289 } 3290 } 3291 } 3292 #endif /* CONFIG_CIFS_ALLOW_INSECURE_LEGACY */ 3293 3294 int cifs_setup_cifs_sb(struct cifs_sb_info *cifs_sb) 3295 { 3296 struct smb3_fs_context *ctx = cifs_sb->ctx; 3297 3298 INIT_DELAYED_WORK(&cifs_sb->prune_tlinks, cifs_prune_tlinks); 3299 3300 spin_lock_init(&cifs_sb->tlink_tree_lock); 3301 cifs_sb->tlink_tree = RB_ROOT; 3302 3303 cifs_dbg(FYI, "file mode: %04ho dir mode: %04ho\n", 3304 ctx->file_mode, ctx->dir_mode); 3305 3306 /* this is needed for ASCII cp to Unicode converts */ 3307 if (ctx->iocharset == NULL) { 3308 /* load_nls_default cannot return null */ 3309 cifs_sb->local_nls = load_nls_default(); 3310 } else { 3311 cifs_sb->local_nls = load_nls(ctx->iocharset); 3312 if (cifs_sb->local_nls == NULL) { 3313 cifs_dbg(VFS, "CIFS mount error: iocharset %s not found\n", 3314 ctx->iocharset); 3315 return -ELIBACC; 3316 } 3317 } 3318 ctx->local_nls = cifs_sb->local_nls; 3319 3320 smb3_update_mnt_flags(cifs_sb); 3321 3322 if (ctx->direct_io) 3323 cifs_dbg(FYI, "mounting share using direct i/o\n"); 3324 if (ctx->cache_ro) { 3325 cifs_dbg(VFS, "mounting share with read only caching. Ensure that the share will not be modified while in use.\n"); 3326 cifs_sb->mnt_cifs_flags |= CIFS_MOUNT_RO_CACHE; 3327 } else if (ctx->cache_rw) { 3328 cifs_dbg(VFS, "mounting share in single client RW caching mode. Ensure that no other systems will be accessing the share.\n"); 3329 cifs_sb->mnt_cifs_flags |= (CIFS_MOUNT_RO_CACHE | 3330 CIFS_MOUNT_RW_CACHE); 3331 } 3332 3333 if ((ctx->cifs_acl) && (ctx->dynperm)) 3334 cifs_dbg(VFS, "mount option dynperm ignored if cifsacl mount option supported\n"); 3335 3336 if (ctx->prepath) { 3337 cifs_sb->prepath = kstrdup(ctx->prepath, GFP_KERNEL); 3338 if (cifs_sb->prepath == NULL) 3339 return -ENOMEM; 3340 cifs_sb->mnt_cifs_flags |= CIFS_MOUNT_USE_PREFIX_PATH; 3341 } 3342 3343 return 0; 3344 } 3345 3346 /* Release all succeed connections */ 3347 void cifs_mount_put_conns(struct cifs_mount_ctx *mnt_ctx) 3348 { 3349 int rc = 0; 3350 3351 if (mnt_ctx->tcon) 3352 cifs_put_tcon(mnt_ctx->tcon, netfs_trace_tcon_ref_put_mnt_ctx); 3353 else if (mnt_ctx->ses) 3354 cifs_put_smb_ses(mnt_ctx->ses); 3355 else if (mnt_ctx->server) 3356 cifs_put_tcp_session(mnt_ctx->server, 0); 3357 mnt_ctx->ses = NULL; 3358 mnt_ctx->tcon = NULL; 3359 mnt_ctx->server = NULL; 3360 mnt_ctx->cifs_sb->mnt_cifs_flags &= ~CIFS_MOUNT_POSIX_PATHS; 3361 free_xid(mnt_ctx->xid); 3362 } 3363 3364 int cifs_mount_get_session(struct cifs_mount_ctx *mnt_ctx) 3365 { 3366 struct TCP_Server_Info *server = NULL; 3367 struct smb3_fs_context *ctx; 3368 struct cifs_ses *ses = NULL; 3369 unsigned int xid; 3370 int rc = 0; 3371 3372 xid = get_xid(); 3373 3374 if (WARN_ON_ONCE(!mnt_ctx || !mnt_ctx->fs_ctx)) { 3375 rc = -EINVAL; 3376 goto out; 3377 } 3378 ctx = mnt_ctx->fs_ctx; 3379 3380 /* get a reference to a tcp session */ 3381 server = cifs_get_tcp_session(ctx, NULL); 3382 if (IS_ERR(server)) { 3383 rc = PTR_ERR(server); 3384 server = NULL; 3385 goto out; 3386 } 3387 3388 /* get a reference to a SMB session */ 3389 ses = cifs_get_smb_ses(server, ctx); 3390 if (IS_ERR(ses)) { 3391 rc = PTR_ERR(ses); 3392 ses = NULL; 3393 goto out; 3394 } 3395 3396 if ((ctx->persistent == true) && (!(ses->server->capabilities & 3397 SMB2_GLOBAL_CAP_PERSISTENT_HANDLES))) { 3398 cifs_server_dbg(VFS, "persistent handles not supported by server\n"); 3399 rc = -EOPNOTSUPP; 3400 } 3401 3402 out: 3403 mnt_ctx->xid = xid; 3404 mnt_ctx->server = server; 3405 mnt_ctx->ses = ses; 3406 mnt_ctx->tcon = NULL; 3407 3408 return rc; 3409 } 3410 3411 int cifs_mount_get_tcon(struct cifs_mount_ctx *mnt_ctx) 3412 { 3413 struct TCP_Server_Info *server; 3414 struct cifs_sb_info *cifs_sb; 3415 struct smb3_fs_context *ctx; 3416 struct cifs_tcon *tcon = NULL; 3417 int rc = 0; 3418 3419 if (WARN_ON_ONCE(!mnt_ctx || !mnt_ctx->server || !mnt_ctx->ses || !mnt_ctx->fs_ctx || 3420 !mnt_ctx->cifs_sb)) { 3421 rc = -EINVAL; 3422 goto out; 3423 } 3424 server = mnt_ctx->server; 3425 ctx = mnt_ctx->fs_ctx; 3426 cifs_sb = mnt_ctx->cifs_sb; 3427 3428 /* search for existing tcon to this server share */ 3429 tcon = cifs_get_tcon(mnt_ctx->ses, ctx); 3430 if (IS_ERR(tcon)) { 3431 rc = PTR_ERR(tcon); 3432 tcon = NULL; 3433 goto out; 3434 } 3435 3436 /* if new SMB3.11 POSIX extensions are supported do not remap / and \ */ 3437 if (tcon->posix_extensions) 3438 cifs_sb->mnt_cifs_flags |= CIFS_MOUNT_POSIX_PATHS; 3439 3440 #ifdef CONFIG_CIFS_ALLOW_INSECURE_LEGACY 3441 /* tell server which Unix caps we support */ 3442 if (cap_unix(tcon->ses)) { 3443 /* 3444 * reset of caps checks mount to see if unix extensions disabled 3445 * for just this mount. 3446 */ 3447 reset_cifs_unix_caps(mnt_ctx->xid, tcon, cifs_sb, ctx); 3448 spin_lock(&tcon->ses->server->srv_lock); 3449 if ((tcon->ses->server->tcpStatus == CifsNeedReconnect) && 3450 (le64_to_cpu(tcon->fsUnixInfo.Capability) & 3451 CIFS_UNIX_TRANSPORT_ENCRYPTION_MANDATORY_CAP)) { 3452 spin_unlock(&tcon->ses->server->srv_lock); 3453 rc = -EACCES; 3454 goto out; 3455 } 3456 spin_unlock(&tcon->ses->server->srv_lock); 3457 } else 3458 #endif /* CONFIG_CIFS_ALLOW_INSECURE_LEGACY */ 3459 tcon->unix_ext = 0; /* server does not support them */ 3460 3461 /* do not care if a following call succeed - informational */ 3462 if (!tcon->pipe && server->ops->qfs_tcon) { 3463 server->ops->qfs_tcon(mnt_ctx->xid, tcon, cifs_sb); 3464 if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_RO_CACHE) { 3465 if (tcon->fsDevInfo.DeviceCharacteristics & 3466 cpu_to_le32(FILE_READ_ONLY_DEVICE)) 3467 cifs_dbg(VFS, "mounted to read only share\n"); 3468 else if ((cifs_sb->mnt_cifs_flags & 3469 CIFS_MOUNT_RW_CACHE) == 0) 3470 cifs_dbg(VFS, "read only mount of RW share\n"); 3471 /* no need to log a RW mount of a typical RW share */ 3472 } 3473 } 3474 3475 /* 3476 * Clamp the rsize/wsize mount arguments if they are too big for the server 3477 * and set the rsize/wsize to the negotiated values if not passed in by 3478 * the user on mount 3479 */ 3480 if ((cifs_sb->ctx->wsize == 0) || 3481 (cifs_sb->ctx->wsize > server->ops->negotiate_wsize(tcon, ctx))) { 3482 cifs_sb->ctx->wsize = 3483 round_down(server->ops->negotiate_wsize(tcon, ctx), PAGE_SIZE); 3484 /* 3485 * in the very unlikely event that the server sent a max write size under PAGE_SIZE, 3486 * (which would get rounded down to 0) then reset wsize to absolute minimum eg 4096 3487 */ 3488 if (cifs_sb->ctx->wsize == 0) { 3489 cifs_sb->ctx->wsize = PAGE_SIZE; 3490 cifs_dbg(VFS, "wsize too small, reset to minimum ie PAGE_SIZE, usually 4096\n"); 3491 } 3492 } 3493 if ((cifs_sb->ctx->rsize == 0) || 3494 (cifs_sb->ctx->rsize > server->ops->negotiate_rsize(tcon, ctx))) 3495 cifs_sb->ctx->rsize = server->ops->negotiate_rsize(tcon, ctx); 3496 3497 /* 3498 * The cookie is initialized from volume info returned above. 3499 * Inside cifs_fscache_get_super_cookie it checks 3500 * that we do not get super cookie twice. 3501 */ 3502 if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_FSCACHE) 3503 cifs_fscache_get_super_cookie(tcon); 3504 3505 out: 3506 mnt_ctx->tcon = tcon; 3507 return rc; 3508 } 3509 3510 static int mount_setup_tlink(struct cifs_sb_info *cifs_sb, struct cifs_ses *ses, 3511 struct cifs_tcon *tcon) 3512 { 3513 struct tcon_link *tlink; 3514 3515 /* hang the tcon off of the superblock */ 3516 tlink = kzalloc(sizeof(*tlink), GFP_KERNEL); 3517 if (tlink == NULL) 3518 return -ENOMEM; 3519 3520 tlink->tl_uid = ses->linux_uid; 3521 tlink->tl_tcon = tcon; 3522 tlink->tl_time = jiffies; 3523 set_bit(TCON_LINK_MASTER, &tlink->tl_flags); 3524 set_bit(TCON_LINK_IN_TREE, &tlink->tl_flags); 3525 3526 cifs_sb->master_tlink = tlink; 3527 spin_lock(&cifs_sb->tlink_tree_lock); 3528 tlink_rb_insert(&cifs_sb->tlink_tree, tlink); 3529 spin_unlock(&cifs_sb->tlink_tree_lock); 3530 3531 queue_delayed_work(cifsiod_wq, &cifs_sb->prune_tlinks, 3532 TLINK_IDLE_EXPIRE); 3533 return 0; 3534 } 3535 3536 static int 3537 cifs_are_all_path_components_accessible(struct TCP_Server_Info *server, 3538 unsigned int xid, 3539 struct cifs_tcon *tcon, 3540 struct cifs_sb_info *cifs_sb, 3541 char *full_path, 3542 int added_treename) 3543 { 3544 int rc; 3545 char *s; 3546 char sep, tmp; 3547 int skip = added_treename ? 1 : 0; 3548 3549 sep = CIFS_DIR_SEP(cifs_sb); 3550 s = full_path; 3551 3552 rc = server->ops->is_path_accessible(xid, tcon, cifs_sb, ""); 3553 while (rc == 0) { 3554 /* skip separators */ 3555 while (*s == sep) 3556 s++; 3557 if (!*s) 3558 break; 3559 /* next separator */ 3560 while (*s && *s != sep) 3561 s++; 3562 /* 3563 * if the treename is added, we then have to skip the first 3564 * part within the separators 3565 */ 3566 if (skip) { 3567 skip = 0; 3568 continue; 3569 } 3570 /* 3571 * temporarily null-terminate the path at the end of 3572 * the current component 3573 */ 3574 tmp = *s; 3575 *s = 0; 3576 rc = server->ops->is_path_accessible(xid, tcon, cifs_sb, 3577 full_path); 3578 *s = tmp; 3579 } 3580 return rc; 3581 } 3582 3583 /* 3584 * Check if path is remote (i.e. a DFS share). 3585 * 3586 * Return -EREMOTE if it is, otherwise 0 or -errno. 3587 */ 3588 int cifs_is_path_remote(struct cifs_mount_ctx *mnt_ctx) 3589 { 3590 int rc; 3591 struct cifs_sb_info *cifs_sb = mnt_ctx->cifs_sb; 3592 struct TCP_Server_Info *server = mnt_ctx->server; 3593 unsigned int xid = mnt_ctx->xid; 3594 struct cifs_tcon *tcon = mnt_ctx->tcon; 3595 struct smb3_fs_context *ctx = mnt_ctx->fs_ctx; 3596 char *full_path; 3597 3598 if (!server->ops->is_path_accessible) 3599 return -EOPNOTSUPP; 3600 3601 /* 3602 * cifs_build_path_to_root works only when we have a valid tcon 3603 */ 3604 full_path = cifs_build_path_to_root(ctx, cifs_sb, tcon, 3605 tcon->Flags & SMB_SHARE_IS_IN_DFS); 3606 if (full_path == NULL) 3607 return -ENOMEM; 3608 3609 cifs_dbg(FYI, "%s: full_path: %s\n", __func__, full_path); 3610 3611 rc = server->ops->is_path_accessible(xid, tcon, cifs_sb, 3612 full_path); 3613 if (rc != 0 && rc != -EREMOTE) 3614 goto out; 3615 3616 if (rc != -EREMOTE) { 3617 rc = cifs_are_all_path_components_accessible(server, xid, tcon, 3618 cifs_sb, full_path, tcon->Flags & SMB_SHARE_IS_IN_DFS); 3619 if (rc != 0) { 3620 cifs_server_dbg(VFS, "cannot query dirs between root and final path, enabling CIFS_MOUNT_USE_PREFIX_PATH\n"); 3621 cifs_sb->mnt_cifs_flags |= CIFS_MOUNT_USE_PREFIX_PATH; 3622 rc = 0; 3623 } 3624 } 3625 3626 out: 3627 kfree(full_path); 3628 return rc; 3629 } 3630 3631 #ifdef CONFIG_CIFS_DFS_UPCALL 3632 int cifs_mount(struct cifs_sb_info *cifs_sb, struct smb3_fs_context *ctx) 3633 { 3634 struct cifs_mount_ctx mnt_ctx = { .cifs_sb = cifs_sb, .fs_ctx = ctx, }; 3635 bool isdfs; 3636 int rc; 3637 3638 rc = dfs_mount_share(&mnt_ctx, &isdfs); 3639 if (rc) 3640 goto error; 3641 if (!isdfs) 3642 goto out; 3643 3644 /* 3645 * After reconnecting to a different server, unique ids won't match anymore, so we disable 3646 * serverino. This prevents dentry revalidation to think the dentry are stale (ESTALE). 3647 */ 3648 cifs_autodisable_serverino(cifs_sb); 3649 /* 3650 * Force the use of prefix path to support failover on DFS paths that resolve to targets 3651 * that have different prefix paths. 3652 */ 3653 cifs_sb->mnt_cifs_flags |= CIFS_MOUNT_USE_PREFIX_PATH; 3654 kfree(cifs_sb->prepath); 3655 cifs_sb->prepath = ctx->prepath; 3656 ctx->prepath = NULL; 3657 3658 out: 3659 cifs_try_adding_channels(mnt_ctx.ses); 3660 rc = mount_setup_tlink(cifs_sb, mnt_ctx.ses, mnt_ctx.tcon); 3661 if (rc) 3662 goto error; 3663 3664 free_xid(mnt_ctx.xid); 3665 return rc; 3666 3667 error: 3668 cifs_mount_put_conns(&mnt_ctx); 3669 return rc; 3670 } 3671 #else 3672 int cifs_mount(struct cifs_sb_info *cifs_sb, struct smb3_fs_context *ctx) 3673 { 3674 int rc = 0; 3675 struct cifs_mount_ctx mnt_ctx = { .cifs_sb = cifs_sb, .fs_ctx = ctx, }; 3676 3677 rc = cifs_mount_get_session(&mnt_ctx); 3678 if (rc) 3679 goto error; 3680 3681 rc = cifs_mount_get_tcon(&mnt_ctx); 3682 if (!rc) { 3683 /* 3684 * Prevent superblock from being created with any missing 3685 * connections. 3686 */ 3687 if (WARN_ON(!mnt_ctx.server)) 3688 rc = -EHOSTDOWN; 3689 else if (WARN_ON(!mnt_ctx.ses)) 3690 rc = -EACCES; 3691 else if (WARN_ON(!mnt_ctx.tcon)) 3692 rc = -ENOENT; 3693 } 3694 if (rc) 3695 goto error; 3696 3697 rc = cifs_is_path_remote(&mnt_ctx); 3698 if (rc == -EREMOTE) 3699 rc = -EOPNOTSUPP; 3700 if (rc) 3701 goto error; 3702 3703 rc = mount_setup_tlink(cifs_sb, mnt_ctx.ses, mnt_ctx.tcon); 3704 if (rc) 3705 goto error; 3706 3707 free_xid(mnt_ctx.xid); 3708 return rc; 3709 3710 error: 3711 cifs_mount_put_conns(&mnt_ctx); 3712 return rc; 3713 } 3714 #endif 3715 3716 #ifdef CONFIG_CIFS_ALLOW_INSECURE_LEGACY 3717 /* 3718 * Issue a TREE_CONNECT request. 3719 */ 3720 int 3721 CIFSTCon(const unsigned int xid, struct cifs_ses *ses, 3722 const char *tree, struct cifs_tcon *tcon, 3723 const struct nls_table *nls_codepage) 3724 { 3725 struct smb_hdr *smb_buffer; 3726 struct smb_hdr *smb_buffer_response; 3727 TCONX_REQ *pSMB; 3728 TCONX_RSP *pSMBr; 3729 unsigned char *bcc_ptr; 3730 int rc = 0; 3731 int length; 3732 __u16 bytes_left, count; 3733 3734 if (ses == NULL) 3735 return -EIO; 3736 3737 smb_buffer = cifs_buf_get(); 3738 if (smb_buffer == NULL) 3739 return -ENOMEM; 3740 3741 smb_buffer_response = smb_buffer; 3742 3743 header_assemble(smb_buffer, SMB_COM_TREE_CONNECT_ANDX, 3744 NULL /*no tid */, 4 /*wct */); 3745 3746 smb_buffer->Mid = get_next_mid(ses->server); 3747 smb_buffer->Uid = ses->Suid; 3748 pSMB = (TCONX_REQ *) smb_buffer; 3749 pSMBr = (TCONX_RSP *) smb_buffer_response; 3750 3751 pSMB->AndXCommand = 0xFF; 3752 pSMB->Flags = cpu_to_le16(TCON_EXTENDED_SECINFO); 3753 bcc_ptr = &pSMB->Password[0]; 3754 3755 pSMB->PasswordLength = cpu_to_le16(1); /* minimum */ 3756 *bcc_ptr = 0; /* password is null byte */ 3757 bcc_ptr++; /* skip password */ 3758 /* already aligned so no need to do it below */ 3759 3760 if (ses->server->sign) 3761 smb_buffer->Flags2 |= SMBFLG2_SECURITY_SIGNATURE; 3762 3763 if (ses->capabilities & CAP_STATUS32) 3764 smb_buffer->Flags2 |= SMBFLG2_ERR_STATUS; 3765 3766 if (ses->capabilities & CAP_DFS) 3767 smb_buffer->Flags2 |= SMBFLG2_DFS; 3768 3769 if (ses->capabilities & CAP_UNICODE) { 3770 smb_buffer->Flags2 |= SMBFLG2_UNICODE; 3771 length = 3772 cifs_strtoUTF16((__le16 *) bcc_ptr, tree, 3773 6 /* max utf8 char length in bytes */ * 3774 (/* server len*/ + 256 /* share len */), nls_codepage); 3775 bcc_ptr += 2 * length; /* convert num 16 bit words to bytes */ 3776 bcc_ptr += 2; /* skip trailing null */ 3777 } else { /* ASCII */ 3778 strcpy(bcc_ptr, tree); 3779 bcc_ptr += strlen(tree) + 1; 3780 } 3781 strcpy(bcc_ptr, "?????"); 3782 bcc_ptr += strlen("?????"); 3783 bcc_ptr += 1; 3784 count = bcc_ptr - &pSMB->Password[0]; 3785 be32_add_cpu(&pSMB->hdr.smb_buf_length, count); 3786 pSMB->ByteCount = cpu_to_le16(count); 3787 3788 rc = SendReceive(xid, ses, smb_buffer, smb_buffer_response, &length, 3789 0); 3790 3791 /* above now done in SendReceive */ 3792 if (rc == 0) { 3793 bool is_unicode; 3794 3795 tcon->tid = smb_buffer_response->Tid; 3796 bcc_ptr = pByteArea(smb_buffer_response); 3797 bytes_left = get_bcc(smb_buffer_response); 3798 length = strnlen(bcc_ptr, bytes_left - 2); 3799 if (smb_buffer->Flags2 & SMBFLG2_UNICODE) 3800 is_unicode = true; 3801 else 3802 is_unicode = false; 3803 3804 3805 /* skip service field (NB: this field is always ASCII) */ 3806 if (length == 3) { 3807 if ((bcc_ptr[0] == 'I') && (bcc_ptr[1] == 'P') && 3808 (bcc_ptr[2] == 'C')) { 3809 cifs_dbg(FYI, "IPC connection\n"); 3810 tcon->ipc = true; 3811 tcon->pipe = true; 3812 } 3813 } else if (length == 2) { 3814 if ((bcc_ptr[0] == 'A') && (bcc_ptr[1] == ':')) { 3815 /* the most common case */ 3816 cifs_dbg(FYI, "disk share connection\n"); 3817 } 3818 } 3819 bcc_ptr += length + 1; 3820 bytes_left -= (length + 1); 3821 strscpy(tcon->tree_name, tree, sizeof(tcon->tree_name)); 3822 3823 /* mostly informational -- no need to fail on error here */ 3824 kfree(tcon->nativeFileSystem); 3825 tcon->nativeFileSystem = cifs_strndup_from_utf16(bcc_ptr, 3826 bytes_left, is_unicode, 3827 nls_codepage); 3828 3829 cifs_dbg(FYI, "nativeFileSystem=%s\n", tcon->nativeFileSystem); 3830 3831 if ((smb_buffer_response->WordCount == 3) || 3832 (smb_buffer_response->WordCount == 7)) 3833 /* field is in same location */ 3834 tcon->Flags = le16_to_cpu(pSMBr->OptionalSupport); 3835 else 3836 tcon->Flags = 0; 3837 cifs_dbg(FYI, "Tcon flags: 0x%x\n", tcon->Flags); 3838 3839 /* 3840 * reset_cifs_unix_caps calls QFSInfo which requires 3841 * need_reconnect to be false, but we would not need to call 3842 * reset_caps if this were not a reconnect case so must check 3843 * need_reconnect flag here. The caller will also clear 3844 * need_reconnect when tcon was successful but needed to be 3845 * cleared earlier in the case of unix extensions reconnect 3846 */ 3847 if (tcon->need_reconnect && tcon->unix_ext) { 3848 cifs_dbg(FYI, "resetting caps for %s\n", tcon->tree_name); 3849 tcon->need_reconnect = false; 3850 reset_cifs_unix_caps(xid, tcon, NULL, NULL); 3851 } 3852 } 3853 cifs_buf_release(smb_buffer); 3854 return rc; 3855 } 3856 #endif /* CONFIG_CIFS_ALLOW_INSECURE_LEGACY */ 3857 3858 static void delayed_free(struct rcu_head *p) 3859 { 3860 struct cifs_sb_info *cifs_sb = container_of(p, struct cifs_sb_info, rcu); 3861 3862 unload_nls(cifs_sb->local_nls); 3863 smb3_cleanup_fs_context(cifs_sb->ctx); 3864 kfree(cifs_sb); 3865 } 3866 3867 void 3868 cifs_umount(struct cifs_sb_info *cifs_sb) 3869 { 3870 struct rb_root *root = &cifs_sb->tlink_tree; 3871 struct rb_node *node; 3872 struct tcon_link *tlink; 3873 3874 cancel_delayed_work_sync(&cifs_sb->prune_tlinks); 3875 3876 spin_lock(&cifs_sb->tlink_tree_lock); 3877 while ((node = rb_first(root))) { 3878 tlink = rb_entry(node, struct tcon_link, tl_rbnode); 3879 cifs_get_tlink(tlink); 3880 clear_bit(TCON_LINK_IN_TREE, &tlink->tl_flags); 3881 rb_erase(node, root); 3882 3883 spin_unlock(&cifs_sb->tlink_tree_lock); 3884 cifs_put_tlink(tlink); 3885 spin_lock(&cifs_sb->tlink_tree_lock); 3886 } 3887 spin_unlock(&cifs_sb->tlink_tree_lock); 3888 3889 kfree(cifs_sb->prepath); 3890 call_rcu(&cifs_sb->rcu, delayed_free); 3891 } 3892 3893 int 3894 cifs_negotiate_protocol(const unsigned int xid, struct cifs_ses *ses, 3895 struct TCP_Server_Info *server) 3896 { 3897 int rc = 0; 3898 3899 if (!server->ops->need_neg || !server->ops->negotiate) 3900 return -ENOSYS; 3901 3902 /* only send once per connect */ 3903 spin_lock(&server->srv_lock); 3904 if (server->tcpStatus != CifsGood && 3905 server->tcpStatus != CifsNew && 3906 server->tcpStatus != CifsNeedNegotiate) { 3907 spin_unlock(&server->srv_lock); 3908 return -EHOSTDOWN; 3909 } 3910 3911 if (!server->ops->need_neg(server) && 3912 server->tcpStatus == CifsGood) { 3913 spin_unlock(&server->srv_lock); 3914 return 0; 3915 } 3916 3917 server->tcpStatus = CifsInNegotiate; 3918 spin_unlock(&server->srv_lock); 3919 3920 rc = server->ops->negotiate(xid, ses, server); 3921 if (rc == 0) { 3922 spin_lock(&server->srv_lock); 3923 if (server->tcpStatus == CifsInNegotiate) 3924 server->tcpStatus = CifsGood; 3925 else 3926 rc = -EHOSTDOWN; 3927 spin_unlock(&server->srv_lock); 3928 } else { 3929 spin_lock(&server->srv_lock); 3930 if (server->tcpStatus == CifsInNegotiate) 3931 server->tcpStatus = CifsNeedNegotiate; 3932 spin_unlock(&server->srv_lock); 3933 } 3934 3935 return rc; 3936 } 3937 3938 int 3939 cifs_setup_session(const unsigned int xid, struct cifs_ses *ses, 3940 struct TCP_Server_Info *server, 3941 struct nls_table *nls_info) 3942 { 3943 int rc = -ENOSYS; 3944 struct TCP_Server_Info *pserver = SERVER_IS_CHAN(server) ? server->primary_server : server; 3945 struct sockaddr_in6 *addr6 = (struct sockaddr_in6 *)&pserver->dstaddr; 3946 struct sockaddr_in *addr = (struct sockaddr_in *)&pserver->dstaddr; 3947 bool is_binding = false; 3948 3949 spin_lock(&ses->ses_lock); 3950 cifs_dbg(FYI, "%s: channel connect bitmap: 0x%lx\n", 3951 __func__, ses->chans_need_reconnect); 3952 3953 if (ses->ses_status != SES_GOOD && 3954 ses->ses_status != SES_NEW && 3955 ses->ses_status != SES_NEED_RECON) { 3956 spin_unlock(&ses->ses_lock); 3957 return -EHOSTDOWN; 3958 } 3959 3960 /* only send once per connect */ 3961 spin_lock(&ses->chan_lock); 3962 if (CIFS_ALL_CHANS_GOOD(ses)) { 3963 if (ses->ses_status == SES_NEED_RECON) 3964 ses->ses_status = SES_GOOD; 3965 spin_unlock(&ses->chan_lock); 3966 spin_unlock(&ses->ses_lock); 3967 return 0; 3968 } 3969 3970 cifs_chan_set_in_reconnect(ses, server); 3971 is_binding = !CIFS_ALL_CHANS_NEED_RECONNECT(ses); 3972 spin_unlock(&ses->chan_lock); 3973 3974 if (!is_binding) { 3975 ses->ses_status = SES_IN_SETUP; 3976 3977 /* force iface_list refresh */ 3978 ses->iface_last_update = 0; 3979 } 3980 spin_unlock(&ses->ses_lock); 3981 3982 /* update ses ip_addr only for primary chan */ 3983 if (server == pserver) { 3984 if (server->dstaddr.ss_family == AF_INET6) 3985 scnprintf(ses->ip_addr, sizeof(ses->ip_addr), "%pI6", &addr6->sin6_addr); 3986 else 3987 scnprintf(ses->ip_addr, sizeof(ses->ip_addr), "%pI4", &addr->sin_addr); 3988 } 3989 3990 if (!is_binding) { 3991 ses->capabilities = server->capabilities; 3992 if (!linuxExtEnabled) 3993 ses->capabilities &= (~server->vals->cap_unix); 3994 3995 if (ses->auth_key.response) { 3996 cifs_dbg(FYI, "Free previous auth_key.response = %p\n", 3997 ses->auth_key.response); 3998 kfree_sensitive(ses->auth_key.response); 3999 ses->auth_key.response = NULL; 4000 ses->auth_key.len = 0; 4001 } 4002 } 4003 4004 cifs_dbg(FYI, "Security Mode: 0x%x Capabilities: 0x%x TimeAdjust: %d\n", 4005 server->sec_mode, server->capabilities, server->timeAdj); 4006 4007 if (server->ops->sess_setup) 4008 rc = server->ops->sess_setup(xid, ses, server, nls_info); 4009 4010 if (rc) { 4011 cifs_server_dbg(VFS, "Send error in SessSetup = %d\n", rc); 4012 spin_lock(&ses->ses_lock); 4013 if (ses->ses_status == SES_IN_SETUP) 4014 ses->ses_status = SES_NEED_RECON; 4015 spin_lock(&ses->chan_lock); 4016 cifs_chan_clear_in_reconnect(ses, server); 4017 spin_unlock(&ses->chan_lock); 4018 spin_unlock(&ses->ses_lock); 4019 } else { 4020 spin_lock(&ses->ses_lock); 4021 if (ses->ses_status == SES_IN_SETUP) 4022 ses->ses_status = SES_GOOD; 4023 spin_lock(&ses->chan_lock); 4024 cifs_chan_clear_in_reconnect(ses, server); 4025 cifs_chan_clear_need_reconnect(ses, server); 4026 spin_unlock(&ses->chan_lock); 4027 spin_unlock(&ses->ses_lock); 4028 } 4029 4030 return rc; 4031 } 4032 4033 static int 4034 cifs_set_vol_auth(struct smb3_fs_context *ctx, struct cifs_ses *ses) 4035 { 4036 ctx->sectype = ses->sectype; 4037 4038 /* krb5 is special, since we don't need username or pw */ 4039 if (ctx->sectype == Kerberos) 4040 return 0; 4041 4042 return cifs_set_cifscreds(ctx, ses); 4043 } 4044 4045 static struct cifs_tcon * 4046 __cifs_construct_tcon(struct cifs_sb_info *cifs_sb, kuid_t fsuid) 4047 { 4048 int rc; 4049 struct cifs_tcon *master_tcon = cifs_sb_master_tcon(cifs_sb); 4050 struct cifs_ses *ses; 4051 struct cifs_tcon *tcon = NULL; 4052 struct smb3_fs_context *ctx; 4053 char *origin_fullpath = NULL; 4054 4055 ctx = kzalloc(sizeof(*ctx), GFP_KERNEL); 4056 if (ctx == NULL) 4057 return ERR_PTR(-ENOMEM); 4058 4059 ctx->local_nls = cifs_sb->local_nls; 4060 ctx->linux_uid = fsuid; 4061 ctx->cred_uid = fsuid; 4062 ctx->UNC = master_tcon->tree_name; 4063 ctx->retry = master_tcon->retry; 4064 ctx->nocase = master_tcon->nocase; 4065 ctx->nohandlecache = master_tcon->nohandlecache; 4066 ctx->local_lease = master_tcon->local_lease; 4067 ctx->no_lease = master_tcon->no_lease; 4068 ctx->resilient = master_tcon->use_resilient; 4069 ctx->persistent = master_tcon->use_persistent; 4070 ctx->handle_timeout = master_tcon->handle_timeout; 4071 ctx->no_linux_ext = !master_tcon->unix_ext; 4072 ctx->linux_ext = master_tcon->posix_extensions; 4073 ctx->sectype = master_tcon->ses->sectype; 4074 ctx->sign = master_tcon->ses->sign; 4075 ctx->seal = master_tcon->seal; 4076 ctx->witness = master_tcon->use_witness; 4077 ctx->dfs_root_ses = master_tcon->ses->dfs_root_ses; 4078 4079 rc = cifs_set_vol_auth(ctx, master_tcon->ses); 4080 if (rc) { 4081 tcon = ERR_PTR(rc); 4082 goto out; 4083 } 4084 4085 /* get a reference for the same TCP session */ 4086 spin_lock(&cifs_tcp_ses_lock); 4087 ++master_tcon->ses->server->srv_count; 4088 spin_unlock(&cifs_tcp_ses_lock); 4089 4090 ses = cifs_get_smb_ses(master_tcon->ses->server, ctx); 4091 if (IS_ERR(ses)) { 4092 tcon = (struct cifs_tcon *)ses; 4093 cifs_put_tcp_session(master_tcon->ses->server, 0); 4094 goto out; 4095 } 4096 4097 #ifdef CONFIG_CIFS_DFS_UPCALL 4098 spin_lock(&master_tcon->tc_lock); 4099 if (master_tcon->origin_fullpath) { 4100 spin_unlock(&master_tcon->tc_lock); 4101 origin_fullpath = dfs_get_path(cifs_sb, cifs_sb->ctx->source); 4102 if (IS_ERR(origin_fullpath)) { 4103 tcon = ERR_CAST(origin_fullpath); 4104 origin_fullpath = NULL; 4105 cifs_put_smb_ses(ses); 4106 goto out; 4107 } 4108 } else { 4109 spin_unlock(&master_tcon->tc_lock); 4110 } 4111 #endif 4112 4113 tcon = cifs_get_tcon(ses, ctx); 4114 if (IS_ERR(tcon)) { 4115 cifs_put_smb_ses(ses); 4116 goto out; 4117 } 4118 4119 #ifdef CONFIG_CIFS_DFS_UPCALL 4120 if (origin_fullpath) { 4121 spin_lock(&tcon->tc_lock); 4122 tcon->origin_fullpath = origin_fullpath; 4123 spin_unlock(&tcon->tc_lock); 4124 origin_fullpath = NULL; 4125 queue_delayed_work(dfscache_wq, &tcon->dfs_cache_work, 4126 dfs_cache_get_ttl() * HZ); 4127 } 4128 #endif 4129 4130 #ifdef CONFIG_CIFS_ALLOW_INSECURE_LEGACY 4131 if (cap_unix(ses)) 4132 reset_cifs_unix_caps(0, tcon, NULL, ctx); 4133 #endif /* CONFIG_CIFS_ALLOW_INSECURE_LEGACY */ 4134 4135 out: 4136 kfree(ctx->username); 4137 kfree_sensitive(ctx->password); 4138 kfree(origin_fullpath); 4139 kfree(ctx); 4140 4141 return tcon; 4142 } 4143 4144 static struct cifs_tcon * 4145 cifs_construct_tcon(struct cifs_sb_info *cifs_sb, kuid_t fsuid) 4146 { 4147 struct cifs_tcon *ret; 4148 4149 cifs_mount_lock(); 4150 ret = __cifs_construct_tcon(cifs_sb, fsuid); 4151 cifs_mount_unlock(); 4152 return ret; 4153 } 4154 4155 struct cifs_tcon * 4156 cifs_sb_master_tcon(struct cifs_sb_info *cifs_sb) 4157 { 4158 return tlink_tcon(cifs_sb_master_tlink(cifs_sb)); 4159 } 4160 4161 /* find and return a tlink with given uid */ 4162 static struct tcon_link * 4163 tlink_rb_search(struct rb_root *root, kuid_t uid) 4164 { 4165 struct rb_node *node = root->rb_node; 4166 struct tcon_link *tlink; 4167 4168 while (node) { 4169 tlink = rb_entry(node, struct tcon_link, tl_rbnode); 4170 4171 if (uid_gt(tlink->tl_uid, uid)) 4172 node = node->rb_left; 4173 else if (uid_lt(tlink->tl_uid, uid)) 4174 node = node->rb_right; 4175 else 4176 return tlink; 4177 } 4178 return NULL; 4179 } 4180 4181 /* insert a tcon_link into the tree */ 4182 static void 4183 tlink_rb_insert(struct rb_root *root, struct tcon_link *new_tlink) 4184 { 4185 struct rb_node **new = &(root->rb_node), *parent = NULL; 4186 struct tcon_link *tlink; 4187 4188 while (*new) { 4189 tlink = rb_entry(*new, struct tcon_link, tl_rbnode); 4190 parent = *new; 4191 4192 if (uid_gt(tlink->tl_uid, new_tlink->tl_uid)) 4193 new = &((*new)->rb_left); 4194 else 4195 new = &((*new)->rb_right); 4196 } 4197 4198 rb_link_node(&new_tlink->tl_rbnode, parent, new); 4199 rb_insert_color(&new_tlink->tl_rbnode, root); 4200 } 4201 4202 /* 4203 * Find or construct an appropriate tcon given a cifs_sb and the fsuid of the 4204 * current task. 4205 * 4206 * If the superblock doesn't refer to a multiuser mount, then just return 4207 * the master tcon for the mount. 4208 * 4209 * First, search the rbtree for an existing tcon for this fsuid. If one 4210 * exists, then check to see if it's pending construction. If it is then wait 4211 * for construction to complete. Once it's no longer pending, check to see if 4212 * it failed and either return an error or retry construction, depending on 4213 * the timeout. 4214 * 4215 * If one doesn't exist then insert a new tcon_link struct into the tree and 4216 * try to construct a new one. 4217 */ 4218 struct tcon_link * 4219 cifs_sb_tlink(struct cifs_sb_info *cifs_sb) 4220 { 4221 int ret; 4222 kuid_t fsuid = current_fsuid(); 4223 struct tcon_link *tlink, *newtlink; 4224 4225 if (!(cifs_sb->mnt_cifs_flags & CIFS_MOUNT_MULTIUSER)) 4226 return cifs_get_tlink(cifs_sb_master_tlink(cifs_sb)); 4227 4228 spin_lock(&cifs_sb->tlink_tree_lock); 4229 tlink = tlink_rb_search(&cifs_sb->tlink_tree, fsuid); 4230 if (tlink) 4231 cifs_get_tlink(tlink); 4232 spin_unlock(&cifs_sb->tlink_tree_lock); 4233 4234 if (tlink == NULL) { 4235 newtlink = kzalloc(sizeof(*tlink), GFP_KERNEL); 4236 if (newtlink == NULL) 4237 return ERR_PTR(-ENOMEM); 4238 newtlink->tl_uid = fsuid; 4239 newtlink->tl_tcon = ERR_PTR(-EACCES); 4240 set_bit(TCON_LINK_PENDING, &newtlink->tl_flags); 4241 set_bit(TCON_LINK_IN_TREE, &newtlink->tl_flags); 4242 cifs_get_tlink(newtlink); 4243 4244 spin_lock(&cifs_sb->tlink_tree_lock); 4245 /* was one inserted after previous search? */ 4246 tlink = tlink_rb_search(&cifs_sb->tlink_tree, fsuid); 4247 if (tlink) { 4248 cifs_get_tlink(tlink); 4249 spin_unlock(&cifs_sb->tlink_tree_lock); 4250 kfree(newtlink); 4251 goto wait_for_construction; 4252 } 4253 tlink = newtlink; 4254 tlink_rb_insert(&cifs_sb->tlink_tree, tlink); 4255 spin_unlock(&cifs_sb->tlink_tree_lock); 4256 } else { 4257 wait_for_construction: 4258 ret = wait_on_bit(&tlink->tl_flags, TCON_LINK_PENDING, 4259 TASK_INTERRUPTIBLE); 4260 if (ret) { 4261 cifs_put_tlink(tlink); 4262 return ERR_PTR(-ERESTARTSYS); 4263 } 4264 4265 /* if it's good, return it */ 4266 if (!IS_ERR(tlink->tl_tcon)) 4267 return tlink; 4268 4269 /* return error if we tried this already recently */ 4270 if (time_before(jiffies, tlink->tl_time + TLINK_ERROR_EXPIRE)) { 4271 cifs_put_tlink(tlink); 4272 return ERR_PTR(-EACCES); 4273 } 4274 4275 if (test_and_set_bit(TCON_LINK_PENDING, &tlink->tl_flags)) 4276 goto wait_for_construction; 4277 } 4278 4279 tlink->tl_tcon = cifs_construct_tcon(cifs_sb, fsuid); 4280 clear_bit(TCON_LINK_PENDING, &tlink->tl_flags); 4281 wake_up_bit(&tlink->tl_flags, TCON_LINK_PENDING); 4282 4283 if (IS_ERR(tlink->tl_tcon)) { 4284 cifs_put_tlink(tlink); 4285 return ERR_PTR(-EACCES); 4286 } 4287 4288 return tlink; 4289 } 4290 4291 /* 4292 * periodic workqueue job that scans tcon_tree for a superblock and closes 4293 * out tcons. 4294 */ 4295 static void 4296 cifs_prune_tlinks(struct work_struct *work) 4297 { 4298 struct cifs_sb_info *cifs_sb = container_of(work, struct cifs_sb_info, 4299 prune_tlinks.work); 4300 struct rb_root *root = &cifs_sb->tlink_tree; 4301 struct rb_node *node; 4302 struct rb_node *tmp; 4303 struct tcon_link *tlink; 4304 4305 /* 4306 * Because we drop the spinlock in the loop in order to put the tlink 4307 * it's not guarded against removal of links from the tree. The only 4308 * places that remove entries from the tree are this function and 4309 * umounts. Because this function is non-reentrant and is canceled 4310 * before umount can proceed, this is safe. 4311 */ 4312 spin_lock(&cifs_sb->tlink_tree_lock); 4313 node = rb_first(root); 4314 while (node != NULL) { 4315 tmp = node; 4316 node = rb_next(tmp); 4317 tlink = rb_entry(tmp, struct tcon_link, tl_rbnode); 4318 4319 if (test_bit(TCON_LINK_MASTER, &tlink->tl_flags) || 4320 atomic_read(&tlink->tl_count) != 0 || 4321 time_after(tlink->tl_time + TLINK_IDLE_EXPIRE, jiffies)) 4322 continue; 4323 4324 cifs_get_tlink(tlink); 4325 clear_bit(TCON_LINK_IN_TREE, &tlink->tl_flags); 4326 rb_erase(tmp, root); 4327 4328 spin_unlock(&cifs_sb->tlink_tree_lock); 4329 cifs_put_tlink(tlink); 4330 spin_lock(&cifs_sb->tlink_tree_lock); 4331 } 4332 spin_unlock(&cifs_sb->tlink_tree_lock); 4333 4334 queue_delayed_work(cifsiod_wq, &cifs_sb->prune_tlinks, 4335 TLINK_IDLE_EXPIRE); 4336 } 4337 4338 #ifndef CONFIG_CIFS_DFS_UPCALL 4339 int cifs_tree_connect(const unsigned int xid, struct cifs_tcon *tcon, const struct nls_table *nlsc) 4340 { 4341 int rc; 4342 const struct smb_version_operations *ops = tcon->ses->server->ops; 4343 4344 /* only send once per connect */ 4345 spin_lock(&tcon->tc_lock); 4346 4347 /* if tcon is marked for needing reconnect, update state */ 4348 if (tcon->need_reconnect) 4349 tcon->status = TID_NEED_TCON; 4350 4351 if (tcon->status == TID_GOOD) { 4352 spin_unlock(&tcon->tc_lock); 4353 return 0; 4354 } 4355 4356 if (tcon->status != TID_NEW && 4357 tcon->status != TID_NEED_TCON) { 4358 spin_unlock(&tcon->tc_lock); 4359 return -EHOSTDOWN; 4360 } 4361 4362 tcon->status = TID_IN_TCON; 4363 spin_unlock(&tcon->tc_lock); 4364 4365 rc = ops->tree_connect(xid, tcon->ses, tcon->tree_name, tcon, nlsc); 4366 if (rc) { 4367 spin_lock(&tcon->tc_lock); 4368 if (tcon->status == TID_IN_TCON) 4369 tcon->status = TID_NEED_TCON; 4370 spin_unlock(&tcon->tc_lock); 4371 } else { 4372 spin_lock(&tcon->tc_lock); 4373 if (tcon->status == TID_IN_TCON) 4374 tcon->status = TID_GOOD; 4375 tcon->need_reconnect = false; 4376 spin_unlock(&tcon->tc_lock); 4377 } 4378 4379 return rc; 4380 } 4381 #endif 4382