1 // SPDX-License-Identifier: LGPL-2.1 2 /* 3 * 4 * Copyright (C) International Business Machines Corp., 2002,2011 5 * Author(s): Steve French (sfrench@us.ibm.com) 6 * 7 */ 8 #include <linux/fs.h> 9 #include <linux/net.h> 10 #include <linux/string.h> 11 #include <linux/sched/mm.h> 12 #include <linux/sched/signal.h> 13 #include <linux/list.h> 14 #include <linux/wait.h> 15 #include <linux/slab.h> 16 #include <linux/pagemap.h> 17 #include <linux/ctype.h> 18 #include <linux/utsname.h> 19 #include <linux/mempool.h> 20 #include <linux/delay.h> 21 #include <linux/completion.h> 22 #include <linux/kthread.h> 23 #include <linux/pagevec.h> 24 #include <linux/freezer.h> 25 #include <linux/namei.h> 26 #include <linux/uuid.h> 27 #include <linux/uaccess.h> 28 #include <asm/processor.h> 29 #include <linux/inet.h> 30 #include <linux/module.h> 31 #include <keys/user-type.h> 32 #include <net/ipv6.h> 33 #include <linux/parser.h> 34 #include <linux/bvec.h> 35 #include "cifspdu.h" 36 #include "cifsglob.h" 37 #include "cifsproto.h" 38 #include "cifs_unicode.h" 39 #include "cifs_debug.h" 40 #include "cifs_fs_sb.h" 41 #include "ntlmssp.h" 42 #include "nterr.h" 43 #include "rfc1002pdu.h" 44 #include "fscache.h" 45 #include "smb2proto.h" 46 #include "smbdirect.h" 47 #include "dns_resolve.h" 48 #ifdef CONFIG_CIFS_DFS_UPCALL 49 #include "dfs.h" 50 #include "dfs_cache.h" 51 #endif 52 #include "fs_context.h" 53 #include "cifs_swn.h" 54 55 extern mempool_t *cifs_req_poolp; 56 extern bool disable_legacy_dialects; 57 58 /* FIXME: should these be tunable? */ 59 #define TLINK_ERROR_EXPIRE (1 * HZ) 60 #define TLINK_IDLE_EXPIRE (600 * HZ) 61 62 /* Drop the connection to not overload the server */ 63 #define MAX_STATUS_IO_TIMEOUT 5 64 65 static int ip_connect(struct TCP_Server_Info *server); 66 static int generic_ip_connect(struct TCP_Server_Info *server); 67 static void tlink_rb_insert(struct rb_root *root, struct tcon_link *new_tlink); 68 static void cifs_prune_tlinks(struct work_struct *work); 69 70 /* 71 * Resolve hostname and set ip addr in tcp ses. Useful for hostnames that may 72 * get their ip addresses changed at some point. 73 * 74 * This should be called with server->srv_mutex held. 75 */ 76 static int reconn_set_ipaddr_from_hostname(struct TCP_Server_Info *server) 77 { 78 int rc; 79 int len; 80 char *unc; 81 struct sockaddr_storage ss; 82 83 if (!server->hostname) 84 return -EINVAL; 85 86 /* if server hostname isn't populated, there's nothing to do here */ 87 if (server->hostname[0] == '\0') 88 return 0; 89 90 len = strlen(server->hostname) + 3; 91 92 unc = kmalloc(len, GFP_KERNEL); 93 if (!unc) { 94 cifs_dbg(FYI, "%s: failed to create UNC path\n", __func__); 95 return -ENOMEM; 96 } 97 scnprintf(unc, len, "\\\\%s", server->hostname); 98 99 spin_lock(&server->srv_lock); 100 ss = server->dstaddr; 101 spin_unlock(&server->srv_lock); 102 103 rc = dns_resolve_server_name_to_ip(unc, (struct sockaddr *)&ss, NULL); 104 kfree(unc); 105 106 if (rc < 0) { 107 cifs_dbg(FYI, "%s: failed to resolve server part of %s to IP: %d\n", 108 __func__, server->hostname, rc); 109 } else { 110 spin_lock(&server->srv_lock); 111 memcpy(&server->dstaddr, &ss, sizeof(server->dstaddr)); 112 spin_unlock(&server->srv_lock); 113 rc = 0; 114 } 115 116 return rc; 117 } 118 119 static void smb2_query_server_interfaces(struct work_struct *work) 120 { 121 int rc; 122 struct cifs_tcon *tcon = container_of(work, 123 struct cifs_tcon, 124 query_interfaces.work); 125 126 /* 127 * query server network interfaces, in case they change 128 */ 129 rc = SMB3_request_interfaces(0, tcon, false); 130 if (rc) { 131 cifs_dbg(FYI, "%s: failed to query server interfaces: %d\n", 132 __func__, rc); 133 } 134 135 queue_delayed_work(cifsiod_wq, &tcon->query_interfaces, 136 (SMB_INTERFACE_POLL_INTERVAL * HZ)); 137 } 138 139 /* 140 * Update the tcpStatus for the server. 141 * This is used to signal the cifsd thread to call cifs_reconnect 142 * ONLY cifsd thread should call cifs_reconnect. For any other 143 * thread, use this function 144 * 145 * @server: the tcp ses for which reconnect is needed 146 * @all_channels: if this needs to be done for all channels 147 */ 148 void 149 cifs_signal_cifsd_for_reconnect(struct TCP_Server_Info *server, 150 bool all_channels) 151 { 152 struct TCP_Server_Info *pserver; 153 struct cifs_ses *ses; 154 int i; 155 156 /* If server is a channel, select the primary channel */ 157 pserver = CIFS_SERVER_IS_CHAN(server) ? server->primary_server : server; 158 159 spin_lock(&pserver->srv_lock); 160 if (!all_channels) { 161 pserver->tcpStatus = CifsNeedReconnect; 162 spin_unlock(&pserver->srv_lock); 163 return; 164 } 165 spin_unlock(&pserver->srv_lock); 166 167 spin_lock(&cifs_tcp_ses_lock); 168 list_for_each_entry(ses, &pserver->smb_ses_list, smb_ses_list) { 169 spin_lock(&ses->chan_lock); 170 for (i = 0; i < ses->chan_count; i++) { 171 spin_lock(&ses->chans[i].server->srv_lock); 172 ses->chans[i].server->tcpStatus = CifsNeedReconnect; 173 spin_unlock(&ses->chans[i].server->srv_lock); 174 } 175 spin_unlock(&ses->chan_lock); 176 } 177 spin_unlock(&cifs_tcp_ses_lock); 178 } 179 180 /* 181 * Mark all sessions and tcons for reconnect. 182 * IMPORTANT: make sure that this gets called only from 183 * cifsd thread. For any other thread, use 184 * cifs_signal_cifsd_for_reconnect 185 * 186 * @server: the tcp ses for which reconnect is needed 187 * @server needs to be previously set to CifsNeedReconnect. 188 * @mark_smb_session: whether even sessions need to be marked 189 */ 190 void 191 cifs_mark_tcp_ses_conns_for_reconnect(struct TCP_Server_Info *server, 192 bool mark_smb_session) 193 { 194 struct TCP_Server_Info *pserver; 195 struct cifs_ses *ses, *nses; 196 struct cifs_tcon *tcon; 197 198 /* 199 * before reconnecting the tcp session, mark the smb session (uid) and the tid bad so they 200 * are not used until reconnected. 201 */ 202 cifs_dbg(FYI, "%s: marking necessary sessions and tcons for reconnect\n", __func__); 203 204 /* If server is a channel, select the primary channel */ 205 pserver = CIFS_SERVER_IS_CHAN(server) ? server->primary_server : server; 206 207 208 spin_lock(&cifs_tcp_ses_lock); 209 list_for_each_entry_safe(ses, nses, &pserver->smb_ses_list, smb_ses_list) { 210 /* check if iface is still active */ 211 if (!cifs_chan_is_iface_active(ses, server)) 212 cifs_chan_update_iface(ses, server); 213 214 spin_lock(&ses->chan_lock); 215 if (!mark_smb_session && cifs_chan_needs_reconnect(ses, server)) { 216 spin_unlock(&ses->chan_lock); 217 continue; 218 } 219 220 if (mark_smb_session) 221 CIFS_SET_ALL_CHANS_NEED_RECONNECT(ses); 222 else 223 cifs_chan_set_need_reconnect(ses, server); 224 225 cifs_dbg(FYI, "%s: channel connect bitmap: 0x%lx\n", 226 __func__, ses->chans_need_reconnect); 227 228 /* If all channels need reconnect, then tcon needs reconnect */ 229 if (!mark_smb_session && !CIFS_ALL_CHANS_NEED_RECONNECT(ses)) { 230 spin_unlock(&ses->chan_lock); 231 continue; 232 } 233 spin_unlock(&ses->chan_lock); 234 235 spin_lock(&ses->ses_lock); 236 ses->ses_status = SES_NEED_RECON; 237 spin_unlock(&ses->ses_lock); 238 239 list_for_each_entry(tcon, &ses->tcon_list, tcon_list) { 240 tcon->need_reconnect = true; 241 spin_lock(&tcon->tc_lock); 242 tcon->status = TID_NEED_RECON; 243 spin_unlock(&tcon->tc_lock); 244 } 245 if (ses->tcon_ipc) { 246 ses->tcon_ipc->need_reconnect = true; 247 spin_lock(&ses->tcon_ipc->tc_lock); 248 ses->tcon_ipc->status = TID_NEED_RECON; 249 spin_unlock(&ses->tcon_ipc->tc_lock); 250 } 251 } 252 spin_unlock(&cifs_tcp_ses_lock); 253 } 254 255 static void 256 cifs_abort_connection(struct TCP_Server_Info *server) 257 { 258 struct mid_q_entry *mid, *nmid; 259 struct list_head retry_list; 260 261 server->maxBuf = 0; 262 server->max_read = 0; 263 264 /* do not want to be sending data on a socket we are freeing */ 265 cifs_dbg(FYI, "%s: tearing down socket\n", __func__); 266 cifs_server_lock(server); 267 if (server->ssocket) { 268 cifs_dbg(FYI, "State: 0x%x Flags: 0x%lx\n", server->ssocket->state, 269 server->ssocket->flags); 270 kernel_sock_shutdown(server->ssocket, SHUT_WR); 271 cifs_dbg(FYI, "Post shutdown state: 0x%x Flags: 0x%lx\n", server->ssocket->state, 272 server->ssocket->flags); 273 sock_release(server->ssocket); 274 server->ssocket = NULL; 275 } 276 server->sequence_number = 0; 277 server->session_estab = false; 278 kfree_sensitive(server->session_key.response); 279 server->session_key.response = NULL; 280 server->session_key.len = 0; 281 server->lstrp = jiffies; 282 283 /* mark submitted MIDs for retry and issue callback */ 284 INIT_LIST_HEAD(&retry_list); 285 cifs_dbg(FYI, "%s: moving mids to private list\n", __func__); 286 spin_lock(&server->mid_lock); 287 list_for_each_entry_safe(mid, nmid, &server->pending_mid_q, qhead) { 288 kref_get(&mid->refcount); 289 if (mid->mid_state == MID_REQUEST_SUBMITTED) 290 mid->mid_state = MID_RETRY_NEEDED; 291 list_move(&mid->qhead, &retry_list); 292 mid->mid_flags |= MID_DELETED; 293 } 294 spin_unlock(&server->mid_lock); 295 cifs_server_unlock(server); 296 297 cifs_dbg(FYI, "%s: issuing mid callbacks\n", __func__); 298 list_for_each_entry_safe(mid, nmid, &retry_list, qhead) { 299 list_del_init(&mid->qhead); 300 mid->callback(mid); 301 release_mid(mid); 302 } 303 304 if (cifs_rdma_enabled(server)) { 305 cifs_server_lock(server); 306 smbd_destroy(server); 307 cifs_server_unlock(server); 308 } 309 } 310 311 static bool cifs_tcp_ses_needs_reconnect(struct TCP_Server_Info *server, int num_targets) 312 { 313 spin_lock(&server->srv_lock); 314 server->nr_targets = num_targets; 315 if (server->tcpStatus == CifsExiting) { 316 /* the demux thread will exit normally next time through the loop */ 317 spin_unlock(&server->srv_lock); 318 wake_up(&server->response_q); 319 return false; 320 } 321 322 cifs_dbg(FYI, "Mark tcp session as need reconnect\n"); 323 trace_smb3_reconnect(server->CurrentMid, server->conn_id, 324 server->hostname); 325 server->tcpStatus = CifsNeedReconnect; 326 327 spin_unlock(&server->srv_lock); 328 return true; 329 } 330 331 /* 332 * cifs tcp session reconnection 333 * 334 * mark tcp session as reconnecting so temporarily locked 335 * mark all smb sessions as reconnecting for tcp session 336 * reconnect tcp session 337 * wake up waiters on reconnection? - (not needed currently) 338 * 339 * if mark_smb_session is passed as true, unconditionally mark 340 * the smb session (and tcon) for reconnect as well. This value 341 * doesn't really matter for non-multichannel scenario. 342 * 343 */ 344 static int __cifs_reconnect(struct TCP_Server_Info *server, 345 bool mark_smb_session) 346 { 347 int rc = 0; 348 349 if (!cifs_tcp_ses_needs_reconnect(server, 1)) 350 return 0; 351 352 cifs_mark_tcp_ses_conns_for_reconnect(server, mark_smb_session); 353 354 cifs_abort_connection(server); 355 356 do { 357 try_to_freeze(); 358 cifs_server_lock(server); 359 360 if (!cifs_swn_set_server_dstaddr(server)) { 361 /* resolve the hostname again to make sure that IP address is up-to-date */ 362 rc = reconn_set_ipaddr_from_hostname(server); 363 cifs_dbg(FYI, "%s: reconn_set_ipaddr_from_hostname: rc=%d\n", __func__, rc); 364 } 365 366 if (cifs_rdma_enabled(server)) 367 rc = smbd_reconnect(server); 368 else 369 rc = generic_ip_connect(server); 370 if (rc) { 371 cifs_server_unlock(server); 372 cifs_dbg(FYI, "%s: reconnect error %d\n", __func__, rc); 373 msleep(3000); 374 } else { 375 atomic_inc(&tcpSesReconnectCount); 376 set_credits(server, 1); 377 spin_lock(&server->srv_lock); 378 if (server->tcpStatus != CifsExiting) 379 server->tcpStatus = CifsNeedNegotiate; 380 spin_unlock(&server->srv_lock); 381 cifs_swn_reset_server_dstaddr(server); 382 cifs_server_unlock(server); 383 mod_delayed_work(cifsiod_wq, &server->reconnect, 0); 384 } 385 } while (server->tcpStatus == CifsNeedReconnect); 386 387 spin_lock(&server->srv_lock); 388 if (server->tcpStatus == CifsNeedNegotiate) 389 mod_delayed_work(cifsiod_wq, &server->echo, 0); 390 spin_unlock(&server->srv_lock); 391 392 wake_up(&server->response_q); 393 return rc; 394 } 395 396 #ifdef CONFIG_CIFS_DFS_UPCALL 397 static int __reconnect_target_unlocked(struct TCP_Server_Info *server, const char *target) 398 { 399 int rc; 400 char *hostname; 401 402 if (!cifs_swn_set_server_dstaddr(server)) { 403 if (server->hostname != target) { 404 hostname = extract_hostname(target); 405 if (!IS_ERR(hostname)) { 406 spin_lock(&server->srv_lock); 407 kfree(server->hostname); 408 server->hostname = hostname; 409 spin_unlock(&server->srv_lock); 410 } else { 411 cifs_dbg(FYI, "%s: couldn't extract hostname or address from dfs target: %ld\n", 412 __func__, PTR_ERR(hostname)); 413 cifs_dbg(FYI, "%s: default to last target server: %s\n", __func__, 414 server->hostname); 415 } 416 } 417 /* resolve the hostname again to make sure that IP address is up-to-date. */ 418 rc = reconn_set_ipaddr_from_hostname(server); 419 cifs_dbg(FYI, "%s: reconn_set_ipaddr_from_hostname: rc=%d\n", __func__, rc); 420 } 421 /* Reconnect the socket */ 422 if (cifs_rdma_enabled(server)) 423 rc = smbd_reconnect(server); 424 else 425 rc = generic_ip_connect(server); 426 427 return rc; 428 } 429 430 static int reconnect_target_unlocked(struct TCP_Server_Info *server, struct dfs_cache_tgt_list *tl, 431 struct dfs_cache_tgt_iterator **target_hint) 432 { 433 int rc; 434 struct dfs_cache_tgt_iterator *tit; 435 436 *target_hint = NULL; 437 438 /* If dfs target list is empty, then reconnect to last server */ 439 tit = dfs_cache_get_tgt_iterator(tl); 440 if (!tit) 441 return __reconnect_target_unlocked(server, server->hostname); 442 443 /* Otherwise, try every dfs target in @tl */ 444 for (; tit; tit = dfs_cache_get_next_tgt(tl, tit)) { 445 rc = __reconnect_target_unlocked(server, dfs_cache_get_tgt_name(tit)); 446 if (!rc) { 447 *target_hint = tit; 448 break; 449 } 450 } 451 return rc; 452 } 453 454 static int reconnect_dfs_server(struct TCP_Server_Info *server) 455 { 456 int rc = 0; 457 struct dfs_cache_tgt_list tl = DFS_CACHE_TGT_LIST_INIT(tl); 458 struct dfs_cache_tgt_iterator *target_hint = NULL; 459 int num_targets = 0; 460 461 /* 462 * Determine the number of dfs targets the referral path in @cifs_sb resolves to. 463 * 464 * smb2_reconnect() needs to know how long it should wait based upon the number of dfs 465 * targets (server->nr_targets). It's also possible that the cached referral was cleared 466 * through /proc/fs/cifs/dfscache or the target list is empty due to server settings after 467 * refreshing the referral, so, in this case, default it to 1. 468 */ 469 mutex_lock(&server->refpath_lock); 470 if (!dfs_cache_noreq_find(server->leaf_fullpath + 1, NULL, &tl)) 471 num_targets = dfs_cache_get_nr_tgts(&tl); 472 mutex_unlock(&server->refpath_lock); 473 if (!num_targets) 474 num_targets = 1; 475 476 if (!cifs_tcp_ses_needs_reconnect(server, num_targets)) 477 return 0; 478 479 /* 480 * Unconditionally mark all sessions & tcons for reconnect as we might be connecting to a 481 * different server or share during failover. It could be improved by adding some logic to 482 * only do that in case it connects to a different server or share, though. 483 */ 484 cifs_mark_tcp_ses_conns_for_reconnect(server, true); 485 486 cifs_abort_connection(server); 487 488 do { 489 try_to_freeze(); 490 cifs_server_lock(server); 491 492 rc = reconnect_target_unlocked(server, &tl, &target_hint); 493 if (rc) { 494 /* Failed to reconnect socket */ 495 cifs_server_unlock(server); 496 cifs_dbg(FYI, "%s: reconnect error %d\n", __func__, rc); 497 msleep(3000); 498 continue; 499 } 500 /* 501 * Socket was created. Update tcp session status to CifsNeedNegotiate so that a 502 * process waiting for reconnect will know it needs to re-establish session and tcon 503 * through the reconnected target server. 504 */ 505 atomic_inc(&tcpSesReconnectCount); 506 set_credits(server, 1); 507 spin_lock(&server->srv_lock); 508 if (server->tcpStatus != CifsExiting) 509 server->tcpStatus = CifsNeedNegotiate; 510 spin_unlock(&server->srv_lock); 511 cifs_swn_reset_server_dstaddr(server); 512 cifs_server_unlock(server); 513 mod_delayed_work(cifsiod_wq, &server->reconnect, 0); 514 } while (server->tcpStatus == CifsNeedReconnect); 515 516 mutex_lock(&server->refpath_lock); 517 dfs_cache_noreq_update_tgthint(server->leaf_fullpath + 1, target_hint); 518 mutex_unlock(&server->refpath_lock); 519 dfs_cache_free_tgts(&tl); 520 521 /* Need to set up echo worker again once connection has been established */ 522 spin_lock(&server->srv_lock); 523 if (server->tcpStatus == CifsNeedNegotiate) 524 mod_delayed_work(cifsiod_wq, &server->echo, 0); 525 spin_unlock(&server->srv_lock); 526 527 wake_up(&server->response_q); 528 return rc; 529 } 530 531 int cifs_reconnect(struct TCP_Server_Info *server, bool mark_smb_session) 532 { 533 mutex_lock(&server->refpath_lock); 534 if (!server->leaf_fullpath) { 535 mutex_unlock(&server->refpath_lock); 536 return __cifs_reconnect(server, mark_smb_session); 537 } 538 mutex_unlock(&server->refpath_lock); 539 540 return reconnect_dfs_server(server); 541 } 542 #else 543 int cifs_reconnect(struct TCP_Server_Info *server, bool mark_smb_session) 544 { 545 return __cifs_reconnect(server, mark_smb_session); 546 } 547 #endif 548 549 static void 550 cifs_echo_request(struct work_struct *work) 551 { 552 int rc; 553 struct TCP_Server_Info *server = container_of(work, 554 struct TCP_Server_Info, echo.work); 555 556 /* 557 * We cannot send an echo if it is disabled. 558 * Also, no need to ping if we got a response recently. 559 */ 560 561 if (server->tcpStatus == CifsNeedReconnect || 562 server->tcpStatus == CifsExiting || 563 server->tcpStatus == CifsNew || 564 (server->ops->can_echo && !server->ops->can_echo(server)) || 565 time_before(jiffies, server->lstrp + server->echo_interval - HZ)) 566 goto requeue_echo; 567 568 rc = server->ops->echo ? server->ops->echo(server) : -ENOSYS; 569 cifs_server_dbg(FYI, "send echo request: rc = %d\n", rc); 570 571 /* Check witness registrations */ 572 cifs_swn_check(); 573 574 requeue_echo: 575 queue_delayed_work(cifsiod_wq, &server->echo, server->echo_interval); 576 } 577 578 static bool 579 allocate_buffers(struct TCP_Server_Info *server) 580 { 581 if (!server->bigbuf) { 582 server->bigbuf = (char *)cifs_buf_get(); 583 if (!server->bigbuf) { 584 cifs_server_dbg(VFS, "No memory for large SMB response\n"); 585 msleep(3000); 586 /* retry will check if exiting */ 587 return false; 588 } 589 } else if (server->large_buf) { 590 /* we are reusing a dirty large buf, clear its start */ 591 memset(server->bigbuf, 0, HEADER_SIZE(server)); 592 } 593 594 if (!server->smallbuf) { 595 server->smallbuf = (char *)cifs_small_buf_get(); 596 if (!server->smallbuf) { 597 cifs_server_dbg(VFS, "No memory for SMB response\n"); 598 msleep(1000); 599 /* retry will check if exiting */ 600 return false; 601 } 602 /* beginning of smb buffer is cleared in our buf_get */ 603 } else { 604 /* if existing small buf clear beginning */ 605 memset(server->smallbuf, 0, HEADER_SIZE(server)); 606 } 607 608 return true; 609 } 610 611 static bool 612 server_unresponsive(struct TCP_Server_Info *server) 613 { 614 /* 615 * We need to wait 3 echo intervals to make sure we handle such 616 * situations right: 617 * 1s client sends a normal SMB request 618 * 2s client gets a response 619 * 30s echo workqueue job pops, and decides we got a response recently 620 * and don't need to send another 621 * ... 622 * 65s kernel_recvmsg times out, and we see that we haven't gotten 623 * a response in >60s. 624 */ 625 spin_lock(&server->srv_lock); 626 if ((server->tcpStatus == CifsGood || 627 server->tcpStatus == CifsNeedNegotiate) && 628 (!server->ops->can_echo || server->ops->can_echo(server)) && 629 time_after(jiffies, server->lstrp + 3 * server->echo_interval)) { 630 spin_unlock(&server->srv_lock); 631 cifs_server_dbg(VFS, "has not responded in %lu seconds. Reconnecting...\n", 632 (3 * server->echo_interval) / HZ); 633 cifs_reconnect(server, false); 634 return true; 635 } 636 spin_unlock(&server->srv_lock); 637 638 return false; 639 } 640 641 static inline bool 642 zero_credits(struct TCP_Server_Info *server) 643 { 644 int val; 645 646 spin_lock(&server->req_lock); 647 val = server->credits + server->echo_credits + server->oplock_credits; 648 if (server->in_flight == 0 && val == 0) { 649 spin_unlock(&server->req_lock); 650 return true; 651 } 652 spin_unlock(&server->req_lock); 653 return false; 654 } 655 656 static int 657 cifs_readv_from_socket(struct TCP_Server_Info *server, struct msghdr *smb_msg) 658 { 659 int length = 0; 660 int total_read; 661 662 for (total_read = 0; msg_data_left(smb_msg); total_read += length) { 663 try_to_freeze(); 664 665 /* reconnect if no credits and no requests in flight */ 666 if (zero_credits(server)) { 667 cifs_reconnect(server, false); 668 return -ECONNABORTED; 669 } 670 671 if (server_unresponsive(server)) 672 return -ECONNABORTED; 673 if (cifs_rdma_enabled(server) && server->smbd_conn) 674 length = smbd_recv(server->smbd_conn, smb_msg); 675 else 676 length = sock_recvmsg(server->ssocket, smb_msg, 0); 677 678 spin_lock(&server->srv_lock); 679 if (server->tcpStatus == CifsExiting) { 680 spin_unlock(&server->srv_lock); 681 return -ESHUTDOWN; 682 } 683 684 if (server->tcpStatus == CifsNeedReconnect) { 685 spin_unlock(&server->srv_lock); 686 cifs_reconnect(server, false); 687 return -ECONNABORTED; 688 } 689 spin_unlock(&server->srv_lock); 690 691 if (length == -ERESTARTSYS || 692 length == -EAGAIN || 693 length == -EINTR) { 694 /* 695 * Minimum sleep to prevent looping, allowing socket 696 * to clear and app threads to set tcpStatus 697 * CifsNeedReconnect if server hung. 698 */ 699 usleep_range(1000, 2000); 700 length = 0; 701 continue; 702 } 703 704 if (length <= 0) { 705 cifs_dbg(FYI, "Received no data or error: %d\n", length); 706 cifs_reconnect(server, false); 707 return -ECONNABORTED; 708 } 709 } 710 return total_read; 711 } 712 713 int 714 cifs_read_from_socket(struct TCP_Server_Info *server, char *buf, 715 unsigned int to_read) 716 { 717 struct msghdr smb_msg = {}; 718 struct kvec iov = {.iov_base = buf, .iov_len = to_read}; 719 iov_iter_kvec(&smb_msg.msg_iter, ITER_DEST, &iov, 1, to_read); 720 721 return cifs_readv_from_socket(server, &smb_msg); 722 } 723 724 ssize_t 725 cifs_discard_from_socket(struct TCP_Server_Info *server, size_t to_read) 726 { 727 struct msghdr smb_msg = {}; 728 729 /* 730 * iov_iter_discard already sets smb_msg.type and count and iov_offset 731 * and cifs_readv_from_socket sets msg_control and msg_controllen 732 * so little to initialize in struct msghdr 733 */ 734 iov_iter_discard(&smb_msg.msg_iter, ITER_DEST, to_read); 735 736 return cifs_readv_from_socket(server, &smb_msg); 737 } 738 739 int 740 cifs_read_page_from_socket(struct TCP_Server_Info *server, struct page *page, 741 unsigned int page_offset, unsigned int to_read) 742 { 743 struct msghdr smb_msg = {}; 744 struct bio_vec bv; 745 746 bvec_set_page(&bv, page, to_read, page_offset); 747 iov_iter_bvec(&smb_msg.msg_iter, ITER_DEST, &bv, 1, to_read); 748 return cifs_readv_from_socket(server, &smb_msg); 749 } 750 751 int 752 cifs_read_iter_from_socket(struct TCP_Server_Info *server, struct iov_iter *iter, 753 unsigned int to_read) 754 { 755 struct msghdr smb_msg = { .msg_iter = *iter }; 756 int ret; 757 758 iov_iter_truncate(&smb_msg.msg_iter, to_read); 759 ret = cifs_readv_from_socket(server, &smb_msg); 760 if (ret > 0) 761 iov_iter_advance(iter, ret); 762 return ret; 763 } 764 765 static bool 766 is_smb_response(struct TCP_Server_Info *server, unsigned char type) 767 { 768 /* 769 * The first byte big endian of the length field, 770 * is actually not part of the length but the type 771 * with the most common, zero, as regular data. 772 */ 773 switch (type) { 774 case RFC1002_SESSION_MESSAGE: 775 /* Regular SMB response */ 776 return true; 777 case RFC1002_SESSION_KEEP_ALIVE: 778 cifs_dbg(FYI, "RFC 1002 session keep alive\n"); 779 break; 780 case RFC1002_POSITIVE_SESSION_RESPONSE: 781 cifs_dbg(FYI, "RFC 1002 positive session response\n"); 782 break; 783 case RFC1002_NEGATIVE_SESSION_RESPONSE: 784 /* 785 * We get this from Windows 98 instead of an error on 786 * SMB negprot response. 787 */ 788 cifs_dbg(FYI, "RFC 1002 negative session response\n"); 789 /* give server a second to clean up */ 790 msleep(1000); 791 /* 792 * Always try 445 first on reconnect since we get NACK 793 * on some if we ever connected to port 139 (the NACK 794 * is since we do not begin with RFC1001 session 795 * initialize frame). 796 */ 797 cifs_set_port((struct sockaddr *)&server->dstaddr, CIFS_PORT); 798 cifs_reconnect(server, true); 799 break; 800 default: 801 cifs_server_dbg(VFS, "RFC 1002 unknown response type 0x%x\n", type); 802 cifs_reconnect(server, true); 803 } 804 805 return false; 806 } 807 808 void 809 dequeue_mid(struct mid_q_entry *mid, bool malformed) 810 { 811 #ifdef CONFIG_CIFS_STATS2 812 mid->when_received = jiffies; 813 #endif 814 spin_lock(&mid->server->mid_lock); 815 if (!malformed) 816 mid->mid_state = MID_RESPONSE_RECEIVED; 817 else 818 mid->mid_state = MID_RESPONSE_MALFORMED; 819 /* 820 * Trying to handle/dequeue a mid after the send_recv() 821 * function has finished processing it is a bug. 822 */ 823 if (mid->mid_flags & MID_DELETED) { 824 spin_unlock(&mid->server->mid_lock); 825 pr_warn_once("trying to dequeue a deleted mid\n"); 826 } else { 827 list_del_init(&mid->qhead); 828 mid->mid_flags |= MID_DELETED; 829 spin_unlock(&mid->server->mid_lock); 830 } 831 } 832 833 static unsigned int 834 smb2_get_credits_from_hdr(char *buffer, struct TCP_Server_Info *server) 835 { 836 struct smb2_hdr *shdr = (struct smb2_hdr *)buffer; 837 838 /* 839 * SMB1 does not use credits. 840 */ 841 if (is_smb1(server)) 842 return 0; 843 844 return le16_to_cpu(shdr->CreditRequest); 845 } 846 847 static void 848 handle_mid(struct mid_q_entry *mid, struct TCP_Server_Info *server, 849 char *buf, int malformed) 850 { 851 if (server->ops->check_trans2 && 852 server->ops->check_trans2(mid, server, buf, malformed)) 853 return; 854 mid->credits_received = smb2_get_credits_from_hdr(buf, server); 855 mid->resp_buf = buf; 856 mid->large_buf = server->large_buf; 857 /* Was previous buf put in mpx struct for multi-rsp? */ 858 if (!mid->multiRsp) { 859 /* smb buffer will be freed by user thread */ 860 if (server->large_buf) 861 server->bigbuf = NULL; 862 else 863 server->smallbuf = NULL; 864 } 865 dequeue_mid(mid, malformed); 866 } 867 868 int 869 cifs_enable_signing(struct TCP_Server_Info *server, bool mnt_sign_required) 870 { 871 bool srv_sign_required = server->sec_mode & server->vals->signing_required; 872 bool srv_sign_enabled = server->sec_mode & server->vals->signing_enabled; 873 bool mnt_sign_enabled; 874 875 /* 876 * Is signing required by mnt options? If not then check 877 * global_secflags to see if it is there. 878 */ 879 if (!mnt_sign_required) 880 mnt_sign_required = ((global_secflags & CIFSSEC_MUST_SIGN) == 881 CIFSSEC_MUST_SIGN); 882 883 /* 884 * If signing is required then it's automatically enabled too, 885 * otherwise, check to see if the secflags allow it. 886 */ 887 mnt_sign_enabled = mnt_sign_required ? mnt_sign_required : 888 (global_secflags & CIFSSEC_MAY_SIGN); 889 890 /* If server requires signing, does client allow it? */ 891 if (srv_sign_required) { 892 if (!mnt_sign_enabled) { 893 cifs_dbg(VFS, "Server requires signing, but it's disabled in SecurityFlags!\n"); 894 return -EOPNOTSUPP; 895 } 896 server->sign = true; 897 } 898 899 /* If client requires signing, does server allow it? */ 900 if (mnt_sign_required) { 901 if (!srv_sign_enabled) { 902 cifs_dbg(VFS, "Server does not support signing!\n"); 903 return -EOPNOTSUPP; 904 } 905 server->sign = true; 906 } 907 908 if (cifs_rdma_enabled(server) && server->sign) 909 cifs_dbg(VFS, "Signing is enabled, and RDMA read/write will be disabled\n"); 910 911 return 0; 912 } 913 914 915 static void clean_demultiplex_info(struct TCP_Server_Info *server) 916 { 917 int length; 918 919 /* take it off the list, if it's not already */ 920 spin_lock(&server->srv_lock); 921 list_del_init(&server->tcp_ses_list); 922 spin_unlock(&server->srv_lock); 923 924 cancel_delayed_work_sync(&server->echo); 925 926 spin_lock(&server->srv_lock); 927 server->tcpStatus = CifsExiting; 928 spin_unlock(&server->srv_lock); 929 wake_up_all(&server->response_q); 930 931 /* check if we have blocked requests that need to free */ 932 spin_lock(&server->req_lock); 933 if (server->credits <= 0) 934 server->credits = 1; 935 spin_unlock(&server->req_lock); 936 /* 937 * Although there should not be any requests blocked on this queue it 938 * can not hurt to be paranoid and try to wake up requests that may 939 * haven been blocked when more than 50 at time were on the wire to the 940 * same server - they now will see the session is in exit state and get 941 * out of SendReceive. 942 */ 943 wake_up_all(&server->request_q); 944 /* give those requests time to exit */ 945 msleep(125); 946 if (cifs_rdma_enabled(server)) 947 smbd_destroy(server); 948 if (server->ssocket) { 949 sock_release(server->ssocket); 950 server->ssocket = NULL; 951 } 952 953 if (!list_empty(&server->pending_mid_q)) { 954 struct list_head dispose_list; 955 struct mid_q_entry *mid_entry; 956 struct list_head *tmp, *tmp2; 957 958 INIT_LIST_HEAD(&dispose_list); 959 spin_lock(&server->mid_lock); 960 list_for_each_safe(tmp, tmp2, &server->pending_mid_q) { 961 mid_entry = list_entry(tmp, struct mid_q_entry, qhead); 962 cifs_dbg(FYI, "Clearing mid %llu\n", mid_entry->mid); 963 kref_get(&mid_entry->refcount); 964 mid_entry->mid_state = MID_SHUTDOWN; 965 list_move(&mid_entry->qhead, &dispose_list); 966 mid_entry->mid_flags |= MID_DELETED; 967 } 968 spin_unlock(&server->mid_lock); 969 970 /* now walk dispose list and issue callbacks */ 971 list_for_each_safe(tmp, tmp2, &dispose_list) { 972 mid_entry = list_entry(tmp, struct mid_q_entry, qhead); 973 cifs_dbg(FYI, "Callback mid %llu\n", mid_entry->mid); 974 list_del_init(&mid_entry->qhead); 975 mid_entry->callback(mid_entry); 976 release_mid(mid_entry); 977 } 978 /* 1/8th of sec is more than enough time for them to exit */ 979 msleep(125); 980 } 981 982 if (!list_empty(&server->pending_mid_q)) { 983 /* 984 * mpx threads have not exited yet give them at least the smb 985 * send timeout time for long ops. 986 * 987 * Due to delays on oplock break requests, we need to wait at 988 * least 45 seconds before giving up on a request getting a 989 * response and going ahead and killing cifsd. 990 */ 991 cifs_dbg(FYI, "Wait for exit from demultiplex thread\n"); 992 msleep(46000); 993 /* 994 * If threads still have not exited they are probably never 995 * coming home not much else we can do but free the memory. 996 */ 997 } 998 999 kfree(server->leaf_fullpath); 1000 kfree(server); 1001 1002 length = atomic_dec_return(&tcpSesAllocCount); 1003 if (length > 0) 1004 mempool_resize(cifs_req_poolp, length + cifs_min_rcv); 1005 } 1006 1007 static int 1008 standard_receive3(struct TCP_Server_Info *server, struct mid_q_entry *mid) 1009 { 1010 int length; 1011 char *buf = server->smallbuf; 1012 unsigned int pdu_length = server->pdu_size; 1013 1014 /* make sure this will fit in a large buffer */ 1015 if (pdu_length > CIFSMaxBufSize + MAX_HEADER_SIZE(server) - 1016 HEADER_PREAMBLE_SIZE(server)) { 1017 cifs_server_dbg(VFS, "SMB response too long (%u bytes)\n", pdu_length); 1018 cifs_reconnect(server, true); 1019 return -ECONNABORTED; 1020 } 1021 1022 /* switch to large buffer if too big for a small one */ 1023 if (pdu_length > MAX_CIFS_SMALL_BUFFER_SIZE - 4) { 1024 server->large_buf = true; 1025 memcpy(server->bigbuf, buf, server->total_read); 1026 buf = server->bigbuf; 1027 } 1028 1029 /* now read the rest */ 1030 length = cifs_read_from_socket(server, buf + HEADER_SIZE(server) - 1, 1031 pdu_length - MID_HEADER_SIZE(server)); 1032 1033 if (length < 0) 1034 return length; 1035 server->total_read += length; 1036 1037 dump_smb(buf, server->total_read); 1038 1039 return cifs_handle_standard(server, mid); 1040 } 1041 1042 int 1043 cifs_handle_standard(struct TCP_Server_Info *server, struct mid_q_entry *mid) 1044 { 1045 char *buf = server->large_buf ? server->bigbuf : server->smallbuf; 1046 int rc; 1047 1048 /* 1049 * We know that we received enough to get to the MID as we 1050 * checked the pdu_length earlier. Now check to see 1051 * if the rest of the header is OK. 1052 * 1053 * 48 bytes is enough to display the header and a little bit 1054 * into the payload for debugging purposes. 1055 */ 1056 rc = server->ops->check_message(buf, server->total_read, server); 1057 if (rc) 1058 cifs_dump_mem("Bad SMB: ", buf, 1059 min_t(unsigned int, server->total_read, 48)); 1060 1061 if (server->ops->is_session_expired && 1062 server->ops->is_session_expired(buf)) { 1063 cifs_reconnect(server, true); 1064 return -1; 1065 } 1066 1067 if (server->ops->is_status_pending && 1068 server->ops->is_status_pending(buf, server)) 1069 return -1; 1070 1071 if (!mid) 1072 return rc; 1073 1074 handle_mid(mid, server, buf, rc); 1075 return 0; 1076 } 1077 1078 static void 1079 smb2_add_credits_from_hdr(char *buffer, struct TCP_Server_Info *server) 1080 { 1081 struct smb2_hdr *shdr = (struct smb2_hdr *)buffer; 1082 int scredits, in_flight; 1083 1084 /* 1085 * SMB1 does not use credits. 1086 */ 1087 if (is_smb1(server)) 1088 return; 1089 1090 if (shdr->CreditRequest) { 1091 spin_lock(&server->req_lock); 1092 server->credits += le16_to_cpu(shdr->CreditRequest); 1093 scredits = server->credits; 1094 in_flight = server->in_flight; 1095 spin_unlock(&server->req_lock); 1096 wake_up(&server->request_q); 1097 1098 trace_smb3_hdr_credits(server->CurrentMid, 1099 server->conn_id, server->hostname, scredits, 1100 le16_to_cpu(shdr->CreditRequest), in_flight); 1101 cifs_server_dbg(FYI, "%s: added %u credits total=%d\n", 1102 __func__, le16_to_cpu(shdr->CreditRequest), 1103 scredits); 1104 } 1105 } 1106 1107 1108 static int 1109 cifs_demultiplex_thread(void *p) 1110 { 1111 int i, num_mids, length; 1112 struct TCP_Server_Info *server = p; 1113 unsigned int pdu_length; 1114 unsigned int next_offset; 1115 char *buf = NULL; 1116 struct task_struct *task_to_wake = NULL; 1117 struct mid_q_entry *mids[MAX_COMPOUND]; 1118 char *bufs[MAX_COMPOUND]; 1119 unsigned int noreclaim_flag, num_io_timeout = 0; 1120 bool pending_reconnect = false; 1121 1122 noreclaim_flag = memalloc_noreclaim_save(); 1123 cifs_dbg(FYI, "Demultiplex PID: %d\n", task_pid_nr(current)); 1124 1125 length = atomic_inc_return(&tcpSesAllocCount); 1126 if (length > 1) 1127 mempool_resize(cifs_req_poolp, length + cifs_min_rcv); 1128 1129 set_freezable(); 1130 allow_kernel_signal(SIGKILL); 1131 while (server->tcpStatus != CifsExiting) { 1132 if (try_to_freeze()) 1133 continue; 1134 1135 if (!allocate_buffers(server)) 1136 continue; 1137 1138 server->large_buf = false; 1139 buf = server->smallbuf; 1140 pdu_length = 4; /* enough to get RFC1001 header */ 1141 1142 length = cifs_read_from_socket(server, buf, pdu_length); 1143 if (length < 0) 1144 continue; 1145 1146 if (is_smb1(server)) 1147 server->total_read = length; 1148 else 1149 server->total_read = 0; 1150 1151 /* 1152 * The right amount was read from socket - 4 bytes, 1153 * so we can now interpret the length field. 1154 */ 1155 pdu_length = get_rfc1002_length(buf); 1156 1157 cifs_dbg(FYI, "RFC1002 header 0x%x\n", pdu_length); 1158 if (!is_smb_response(server, buf[0])) 1159 continue; 1160 1161 pending_reconnect = false; 1162 next_pdu: 1163 server->pdu_size = pdu_length; 1164 1165 /* make sure we have enough to get to the MID */ 1166 if (server->pdu_size < MID_HEADER_SIZE(server)) { 1167 cifs_server_dbg(VFS, "SMB response too short (%u bytes)\n", 1168 server->pdu_size); 1169 cifs_reconnect(server, true); 1170 continue; 1171 } 1172 1173 /* read down to the MID */ 1174 length = cifs_read_from_socket(server, 1175 buf + HEADER_PREAMBLE_SIZE(server), 1176 MID_HEADER_SIZE(server)); 1177 if (length < 0) 1178 continue; 1179 server->total_read += length; 1180 1181 if (server->ops->next_header) { 1182 next_offset = server->ops->next_header(buf); 1183 if (next_offset) 1184 server->pdu_size = next_offset; 1185 } 1186 1187 memset(mids, 0, sizeof(mids)); 1188 memset(bufs, 0, sizeof(bufs)); 1189 num_mids = 0; 1190 1191 if (server->ops->is_transform_hdr && 1192 server->ops->receive_transform && 1193 server->ops->is_transform_hdr(buf)) { 1194 length = server->ops->receive_transform(server, 1195 mids, 1196 bufs, 1197 &num_mids); 1198 } else { 1199 mids[0] = server->ops->find_mid(server, buf); 1200 bufs[0] = buf; 1201 num_mids = 1; 1202 1203 if (!mids[0] || !mids[0]->receive) 1204 length = standard_receive3(server, mids[0]); 1205 else 1206 length = mids[0]->receive(server, mids[0]); 1207 } 1208 1209 if (length < 0) { 1210 for (i = 0; i < num_mids; i++) 1211 if (mids[i]) 1212 release_mid(mids[i]); 1213 continue; 1214 } 1215 1216 if (server->ops->is_status_io_timeout && 1217 server->ops->is_status_io_timeout(buf)) { 1218 num_io_timeout++; 1219 if (num_io_timeout > MAX_STATUS_IO_TIMEOUT) { 1220 cifs_server_dbg(VFS, 1221 "Number of request timeouts exceeded %d. Reconnecting", 1222 MAX_STATUS_IO_TIMEOUT); 1223 1224 pending_reconnect = true; 1225 num_io_timeout = 0; 1226 } 1227 } 1228 1229 server->lstrp = jiffies; 1230 1231 for (i = 0; i < num_mids; i++) { 1232 if (mids[i] != NULL) { 1233 mids[i]->resp_buf_size = server->pdu_size; 1234 1235 if (bufs[i] != NULL) { 1236 if (server->ops->is_network_name_deleted && 1237 server->ops->is_network_name_deleted(bufs[i], 1238 server)) { 1239 cifs_server_dbg(FYI, 1240 "Share deleted. Reconnect needed"); 1241 } 1242 } 1243 1244 if (!mids[i]->multiRsp || mids[i]->multiEnd) 1245 mids[i]->callback(mids[i]); 1246 1247 release_mid(mids[i]); 1248 } else if (server->ops->is_oplock_break && 1249 server->ops->is_oplock_break(bufs[i], 1250 server)) { 1251 smb2_add_credits_from_hdr(bufs[i], server); 1252 cifs_dbg(FYI, "Received oplock break\n"); 1253 } else { 1254 cifs_server_dbg(VFS, "No task to wake, unknown frame received! NumMids %d\n", 1255 atomic_read(&mid_count)); 1256 cifs_dump_mem("Received Data is: ", bufs[i], 1257 HEADER_SIZE(server)); 1258 smb2_add_credits_from_hdr(bufs[i], server); 1259 #ifdef CONFIG_CIFS_DEBUG2 1260 if (server->ops->dump_detail) 1261 server->ops->dump_detail(bufs[i], 1262 server); 1263 cifs_dump_mids(server); 1264 #endif /* CIFS_DEBUG2 */ 1265 } 1266 } 1267 1268 if (pdu_length > server->pdu_size) { 1269 if (!allocate_buffers(server)) 1270 continue; 1271 pdu_length -= server->pdu_size; 1272 server->total_read = 0; 1273 server->large_buf = false; 1274 buf = server->smallbuf; 1275 goto next_pdu; 1276 } 1277 1278 /* do this reconnect at the very end after processing all MIDs */ 1279 if (pending_reconnect) 1280 cifs_reconnect(server, true); 1281 1282 } /* end while !EXITING */ 1283 1284 /* buffer usually freed in free_mid - need to free it here on exit */ 1285 cifs_buf_release(server->bigbuf); 1286 if (server->smallbuf) /* no sense logging a debug message if NULL */ 1287 cifs_small_buf_release(server->smallbuf); 1288 1289 task_to_wake = xchg(&server->tsk, NULL); 1290 clean_demultiplex_info(server); 1291 1292 /* if server->tsk was NULL then wait for a signal before exiting */ 1293 if (!task_to_wake) { 1294 set_current_state(TASK_INTERRUPTIBLE); 1295 while (!signal_pending(current)) { 1296 schedule(); 1297 set_current_state(TASK_INTERRUPTIBLE); 1298 } 1299 set_current_state(TASK_RUNNING); 1300 } 1301 1302 memalloc_noreclaim_restore(noreclaim_flag); 1303 module_put_and_kthread_exit(0); 1304 } 1305 1306 int 1307 cifs_ipaddr_cmp(struct sockaddr *srcaddr, struct sockaddr *rhs) 1308 { 1309 struct sockaddr_in *saddr4 = (struct sockaddr_in *)srcaddr; 1310 struct sockaddr_in *vaddr4 = (struct sockaddr_in *)rhs; 1311 struct sockaddr_in6 *saddr6 = (struct sockaddr_in6 *)srcaddr; 1312 struct sockaddr_in6 *vaddr6 = (struct sockaddr_in6 *)rhs; 1313 1314 switch (srcaddr->sa_family) { 1315 case AF_UNSPEC: 1316 switch (rhs->sa_family) { 1317 case AF_UNSPEC: 1318 return 0; 1319 case AF_INET: 1320 case AF_INET6: 1321 return 1; 1322 default: 1323 return -1; 1324 } 1325 case AF_INET: { 1326 switch (rhs->sa_family) { 1327 case AF_UNSPEC: 1328 return -1; 1329 case AF_INET: 1330 return memcmp(saddr4, vaddr4, 1331 sizeof(struct sockaddr_in)); 1332 case AF_INET6: 1333 return 1; 1334 default: 1335 return -1; 1336 } 1337 } 1338 case AF_INET6: { 1339 switch (rhs->sa_family) { 1340 case AF_UNSPEC: 1341 case AF_INET: 1342 return -1; 1343 case AF_INET6: 1344 return memcmp(saddr6, 1345 vaddr6, 1346 sizeof(struct sockaddr_in6)); 1347 default: 1348 return -1; 1349 } 1350 } 1351 default: 1352 return -1; /* don't expect to be here */ 1353 } 1354 } 1355 1356 /* 1357 * Returns true if srcaddr isn't specified and rhs isn't specified, or 1358 * if srcaddr is specified and matches the IP address of the rhs argument 1359 */ 1360 bool 1361 cifs_match_ipaddr(struct sockaddr *srcaddr, struct sockaddr *rhs) 1362 { 1363 switch (srcaddr->sa_family) { 1364 case AF_UNSPEC: 1365 return (rhs->sa_family == AF_UNSPEC); 1366 case AF_INET: { 1367 struct sockaddr_in *saddr4 = (struct sockaddr_in *)srcaddr; 1368 struct sockaddr_in *vaddr4 = (struct sockaddr_in *)rhs; 1369 return (saddr4->sin_addr.s_addr == vaddr4->sin_addr.s_addr); 1370 } 1371 case AF_INET6: { 1372 struct sockaddr_in6 *saddr6 = (struct sockaddr_in6 *)srcaddr; 1373 struct sockaddr_in6 *vaddr6 = (struct sockaddr_in6 *)rhs; 1374 return (ipv6_addr_equal(&saddr6->sin6_addr, &vaddr6->sin6_addr) 1375 && saddr6->sin6_scope_id == vaddr6->sin6_scope_id); 1376 } 1377 default: 1378 WARN_ON(1); 1379 return false; /* don't expect to be here */ 1380 } 1381 } 1382 1383 /* 1384 * If no port is specified in addr structure, we try to match with 445 port 1385 * and if it fails - with 139 ports. It should be called only if address 1386 * families of server and addr are equal. 1387 */ 1388 static bool 1389 match_port(struct TCP_Server_Info *server, struct sockaddr *addr) 1390 { 1391 __be16 port, *sport; 1392 1393 /* SMBDirect manages its own ports, don't match it here */ 1394 if (server->rdma) 1395 return true; 1396 1397 switch (addr->sa_family) { 1398 case AF_INET: 1399 sport = &((struct sockaddr_in *) &server->dstaddr)->sin_port; 1400 port = ((struct sockaddr_in *) addr)->sin_port; 1401 break; 1402 case AF_INET6: 1403 sport = &((struct sockaddr_in6 *) &server->dstaddr)->sin6_port; 1404 port = ((struct sockaddr_in6 *) addr)->sin6_port; 1405 break; 1406 default: 1407 WARN_ON(1); 1408 return false; 1409 } 1410 1411 if (!port) { 1412 port = htons(CIFS_PORT); 1413 if (port == *sport) 1414 return true; 1415 1416 port = htons(RFC1001_PORT); 1417 } 1418 1419 return port == *sport; 1420 } 1421 1422 static bool match_server_address(struct TCP_Server_Info *server, struct sockaddr *addr) 1423 { 1424 if (!cifs_match_ipaddr(addr, (struct sockaddr *)&server->dstaddr)) 1425 return false; 1426 1427 return true; 1428 } 1429 1430 static bool 1431 match_security(struct TCP_Server_Info *server, struct smb3_fs_context *ctx) 1432 { 1433 /* 1434 * The select_sectype function should either return the ctx->sectype 1435 * that was specified, or "Unspecified" if that sectype was not 1436 * compatible with the given NEGOTIATE request. 1437 */ 1438 if (server->ops->select_sectype(server, ctx->sectype) 1439 == Unspecified) 1440 return false; 1441 1442 /* 1443 * Now check if signing mode is acceptable. No need to check 1444 * global_secflags at this point since if MUST_SIGN is set then 1445 * the server->sign had better be too. 1446 */ 1447 if (ctx->sign && !server->sign) 1448 return false; 1449 1450 return true; 1451 } 1452 1453 /* this function must be called with srv_lock held */ 1454 static int match_server(struct TCP_Server_Info *server, 1455 struct smb3_fs_context *ctx, 1456 bool match_super) 1457 { 1458 struct sockaddr *addr = (struct sockaddr *)&ctx->dstaddr; 1459 1460 lockdep_assert_held(&server->srv_lock); 1461 1462 if (ctx->nosharesock) 1463 return 0; 1464 1465 /* this server does not share socket */ 1466 if (server->nosharesock) 1467 return 0; 1468 1469 /* If multidialect negotiation see if existing sessions match one */ 1470 if (strcmp(ctx->vals->version_string, SMB3ANY_VERSION_STRING) == 0) { 1471 if (server->vals->protocol_id < SMB30_PROT_ID) 1472 return 0; 1473 } else if (strcmp(ctx->vals->version_string, 1474 SMBDEFAULT_VERSION_STRING) == 0) { 1475 if (server->vals->protocol_id < SMB21_PROT_ID) 1476 return 0; 1477 } else if ((server->vals != ctx->vals) || (server->ops != ctx->ops)) 1478 return 0; 1479 1480 if (!net_eq(cifs_net_ns(server), current->nsproxy->net_ns)) 1481 return 0; 1482 1483 if (!cifs_match_ipaddr((struct sockaddr *)&ctx->srcaddr, 1484 (struct sockaddr *)&server->srcaddr)) 1485 return 0; 1486 /* 1487 * When matching cifs.ko superblocks (@match_super == true), we can't 1488 * really match either @server->leaf_fullpath or @server->dstaddr 1489 * directly since this @server might belong to a completely different 1490 * server -- in case of domain-based DFS referrals or DFS links -- as 1491 * provided earlier by mount(2) through 'source' and 'ip' options. 1492 * 1493 * Otherwise, match the DFS referral in @server->leaf_fullpath or the 1494 * destination address in @server->dstaddr. 1495 * 1496 * When using 'nodfs' mount option, we avoid sharing it with DFS 1497 * connections as they might failover. 1498 */ 1499 if (!match_super) { 1500 if (!ctx->nodfs) { 1501 if (server->leaf_fullpath) { 1502 if (!ctx->leaf_fullpath || 1503 strcasecmp(server->leaf_fullpath, 1504 ctx->leaf_fullpath)) 1505 return 0; 1506 } else if (ctx->leaf_fullpath) { 1507 return 0; 1508 } 1509 } else if (server->leaf_fullpath) { 1510 return 0; 1511 } 1512 } 1513 1514 /* 1515 * Match for a regular connection (address/hostname/port) which has no 1516 * DFS referrals set. 1517 */ 1518 if (!server->leaf_fullpath && 1519 (strcasecmp(server->hostname, ctx->server_hostname) || 1520 !match_server_address(server, addr) || 1521 !match_port(server, addr))) 1522 return 0; 1523 1524 if (!match_security(server, ctx)) 1525 return 0; 1526 1527 if (server->echo_interval != ctx->echo_interval * HZ) 1528 return 0; 1529 1530 if (server->rdma != ctx->rdma) 1531 return 0; 1532 1533 if (server->ignore_signature != ctx->ignore_signature) 1534 return 0; 1535 1536 if (server->min_offload != ctx->min_offload) 1537 return 0; 1538 1539 return 1; 1540 } 1541 1542 struct TCP_Server_Info * 1543 cifs_find_tcp_session(struct smb3_fs_context *ctx) 1544 { 1545 struct TCP_Server_Info *server; 1546 1547 spin_lock(&cifs_tcp_ses_lock); 1548 list_for_each_entry(server, &cifs_tcp_ses_list, tcp_ses_list) { 1549 spin_lock(&server->srv_lock); 1550 /* 1551 * Skip ses channels since they're only handled in lower layers 1552 * (e.g. cifs_send_recv). 1553 */ 1554 if (CIFS_SERVER_IS_CHAN(server) || 1555 !match_server(server, ctx, false)) { 1556 spin_unlock(&server->srv_lock); 1557 continue; 1558 } 1559 spin_unlock(&server->srv_lock); 1560 1561 ++server->srv_count; 1562 spin_unlock(&cifs_tcp_ses_lock); 1563 cifs_dbg(FYI, "Existing tcp session with server found\n"); 1564 return server; 1565 } 1566 spin_unlock(&cifs_tcp_ses_lock); 1567 return NULL; 1568 } 1569 1570 void 1571 cifs_put_tcp_session(struct TCP_Server_Info *server, int from_reconnect) 1572 { 1573 struct task_struct *task; 1574 1575 spin_lock(&cifs_tcp_ses_lock); 1576 if (--server->srv_count > 0) { 1577 spin_unlock(&cifs_tcp_ses_lock); 1578 return; 1579 } 1580 1581 /* srv_count can never go negative */ 1582 WARN_ON(server->srv_count < 0); 1583 1584 put_net(cifs_net_ns(server)); 1585 1586 list_del_init(&server->tcp_ses_list); 1587 spin_unlock(&cifs_tcp_ses_lock); 1588 1589 /* For secondary channels, we pick up ref-count on the primary server */ 1590 if (CIFS_SERVER_IS_CHAN(server)) 1591 cifs_put_tcp_session(server->primary_server, from_reconnect); 1592 1593 cancel_delayed_work_sync(&server->echo); 1594 1595 if (from_reconnect) 1596 /* 1597 * Avoid deadlock here: reconnect work calls 1598 * cifs_put_tcp_session() at its end. Need to be sure 1599 * that reconnect work does nothing with server pointer after 1600 * that step. 1601 */ 1602 cancel_delayed_work(&server->reconnect); 1603 else 1604 cancel_delayed_work_sync(&server->reconnect); 1605 1606 spin_lock(&server->srv_lock); 1607 server->tcpStatus = CifsExiting; 1608 spin_unlock(&server->srv_lock); 1609 1610 cifs_crypto_secmech_release(server); 1611 1612 kfree_sensitive(server->session_key.response); 1613 server->session_key.response = NULL; 1614 server->session_key.len = 0; 1615 kfree(server->hostname); 1616 server->hostname = NULL; 1617 1618 task = xchg(&server->tsk, NULL); 1619 if (task) 1620 send_sig(SIGKILL, task, 1); 1621 } 1622 1623 struct TCP_Server_Info * 1624 cifs_get_tcp_session(struct smb3_fs_context *ctx, 1625 struct TCP_Server_Info *primary_server) 1626 { 1627 struct TCP_Server_Info *tcp_ses = NULL; 1628 int rc; 1629 1630 cifs_dbg(FYI, "UNC: %s\n", ctx->UNC); 1631 1632 /* see if we already have a matching tcp_ses */ 1633 tcp_ses = cifs_find_tcp_session(ctx); 1634 if (tcp_ses) 1635 return tcp_ses; 1636 1637 tcp_ses = kzalloc(sizeof(struct TCP_Server_Info), GFP_KERNEL); 1638 if (!tcp_ses) { 1639 rc = -ENOMEM; 1640 goto out_err; 1641 } 1642 1643 tcp_ses->hostname = kstrdup(ctx->server_hostname, GFP_KERNEL); 1644 if (!tcp_ses->hostname) { 1645 rc = -ENOMEM; 1646 goto out_err; 1647 } 1648 1649 if (ctx->leaf_fullpath) { 1650 tcp_ses->leaf_fullpath = kstrdup(ctx->leaf_fullpath, GFP_KERNEL); 1651 if (!tcp_ses->leaf_fullpath) { 1652 rc = -ENOMEM; 1653 goto out_err; 1654 } 1655 } 1656 1657 if (ctx->nosharesock) 1658 tcp_ses->nosharesock = true; 1659 1660 tcp_ses->ops = ctx->ops; 1661 tcp_ses->vals = ctx->vals; 1662 cifs_set_net_ns(tcp_ses, get_net(current->nsproxy->net_ns)); 1663 1664 tcp_ses->conn_id = atomic_inc_return(&tcpSesNextId); 1665 tcp_ses->noblockcnt = ctx->rootfs; 1666 tcp_ses->noblocksnd = ctx->noblocksnd || ctx->rootfs; 1667 tcp_ses->noautotune = ctx->noautotune; 1668 tcp_ses->tcp_nodelay = ctx->sockopt_tcp_nodelay; 1669 tcp_ses->rdma = ctx->rdma; 1670 tcp_ses->in_flight = 0; 1671 tcp_ses->max_in_flight = 0; 1672 tcp_ses->credits = 1; 1673 if (primary_server) { 1674 spin_lock(&cifs_tcp_ses_lock); 1675 ++primary_server->srv_count; 1676 spin_unlock(&cifs_tcp_ses_lock); 1677 tcp_ses->primary_server = primary_server; 1678 } 1679 init_waitqueue_head(&tcp_ses->response_q); 1680 init_waitqueue_head(&tcp_ses->request_q); 1681 INIT_LIST_HEAD(&tcp_ses->pending_mid_q); 1682 mutex_init(&tcp_ses->_srv_mutex); 1683 memcpy(tcp_ses->workstation_RFC1001_name, 1684 ctx->source_rfc1001_name, RFC1001_NAME_LEN_WITH_NULL); 1685 memcpy(tcp_ses->server_RFC1001_name, 1686 ctx->target_rfc1001_name, RFC1001_NAME_LEN_WITH_NULL); 1687 tcp_ses->session_estab = false; 1688 tcp_ses->sequence_number = 0; 1689 tcp_ses->reconnect_instance = 1; 1690 tcp_ses->lstrp = jiffies; 1691 tcp_ses->compress_algorithm = cpu_to_le16(ctx->compression); 1692 spin_lock_init(&tcp_ses->req_lock); 1693 spin_lock_init(&tcp_ses->srv_lock); 1694 spin_lock_init(&tcp_ses->mid_lock); 1695 INIT_LIST_HEAD(&tcp_ses->tcp_ses_list); 1696 INIT_LIST_HEAD(&tcp_ses->smb_ses_list); 1697 INIT_DELAYED_WORK(&tcp_ses->echo, cifs_echo_request); 1698 INIT_DELAYED_WORK(&tcp_ses->reconnect, smb2_reconnect_server); 1699 mutex_init(&tcp_ses->reconnect_mutex); 1700 #ifdef CONFIG_CIFS_DFS_UPCALL 1701 mutex_init(&tcp_ses->refpath_lock); 1702 #endif 1703 memcpy(&tcp_ses->srcaddr, &ctx->srcaddr, 1704 sizeof(tcp_ses->srcaddr)); 1705 memcpy(&tcp_ses->dstaddr, &ctx->dstaddr, 1706 sizeof(tcp_ses->dstaddr)); 1707 if (ctx->use_client_guid) 1708 memcpy(tcp_ses->client_guid, ctx->client_guid, 1709 SMB2_CLIENT_GUID_SIZE); 1710 else 1711 generate_random_uuid(tcp_ses->client_guid); 1712 /* 1713 * at this point we are the only ones with the pointer 1714 * to the struct since the kernel thread not created yet 1715 * no need to spinlock this init of tcpStatus or srv_count 1716 */ 1717 tcp_ses->tcpStatus = CifsNew; 1718 ++tcp_ses->srv_count; 1719 1720 if (ctx->echo_interval >= SMB_ECHO_INTERVAL_MIN && 1721 ctx->echo_interval <= SMB_ECHO_INTERVAL_MAX) 1722 tcp_ses->echo_interval = ctx->echo_interval * HZ; 1723 else 1724 tcp_ses->echo_interval = SMB_ECHO_INTERVAL_DEFAULT * HZ; 1725 if (tcp_ses->rdma) { 1726 #ifndef CONFIG_CIFS_SMB_DIRECT 1727 cifs_dbg(VFS, "CONFIG_CIFS_SMB_DIRECT is not enabled\n"); 1728 rc = -ENOENT; 1729 goto out_err_crypto_release; 1730 #endif 1731 tcp_ses->smbd_conn = smbd_get_connection( 1732 tcp_ses, (struct sockaddr *)&ctx->dstaddr); 1733 if (tcp_ses->smbd_conn) { 1734 cifs_dbg(VFS, "RDMA transport established\n"); 1735 rc = 0; 1736 goto smbd_connected; 1737 } else { 1738 rc = -ENOENT; 1739 goto out_err_crypto_release; 1740 } 1741 } 1742 rc = ip_connect(tcp_ses); 1743 if (rc < 0) { 1744 cifs_dbg(VFS, "Error connecting to socket. Aborting operation.\n"); 1745 goto out_err_crypto_release; 1746 } 1747 smbd_connected: 1748 /* 1749 * since we're in a cifs function already, we know that 1750 * this will succeed. No need for try_module_get(). 1751 */ 1752 __module_get(THIS_MODULE); 1753 tcp_ses->tsk = kthread_run(cifs_demultiplex_thread, 1754 tcp_ses, "cifsd"); 1755 if (IS_ERR(tcp_ses->tsk)) { 1756 rc = PTR_ERR(tcp_ses->tsk); 1757 cifs_dbg(VFS, "error %d create cifsd thread\n", rc); 1758 module_put(THIS_MODULE); 1759 goto out_err_crypto_release; 1760 } 1761 tcp_ses->min_offload = ctx->min_offload; 1762 /* 1763 * at this point we are the only ones with the pointer 1764 * to the struct since the kernel thread not created yet 1765 * no need to spinlock this update of tcpStatus 1766 */ 1767 spin_lock(&tcp_ses->srv_lock); 1768 tcp_ses->tcpStatus = CifsNeedNegotiate; 1769 spin_unlock(&tcp_ses->srv_lock); 1770 1771 if ((ctx->max_credits < 20) || (ctx->max_credits > 60000)) 1772 tcp_ses->max_credits = SMB2_MAX_CREDITS_AVAILABLE; 1773 else 1774 tcp_ses->max_credits = ctx->max_credits; 1775 1776 tcp_ses->nr_targets = 1; 1777 tcp_ses->ignore_signature = ctx->ignore_signature; 1778 /* thread spawned, put it on the list */ 1779 spin_lock(&cifs_tcp_ses_lock); 1780 list_add(&tcp_ses->tcp_ses_list, &cifs_tcp_ses_list); 1781 spin_unlock(&cifs_tcp_ses_lock); 1782 1783 /* queue echo request delayed work */ 1784 queue_delayed_work(cifsiod_wq, &tcp_ses->echo, tcp_ses->echo_interval); 1785 1786 return tcp_ses; 1787 1788 out_err_crypto_release: 1789 cifs_crypto_secmech_release(tcp_ses); 1790 1791 put_net(cifs_net_ns(tcp_ses)); 1792 1793 out_err: 1794 if (tcp_ses) { 1795 if (CIFS_SERVER_IS_CHAN(tcp_ses)) 1796 cifs_put_tcp_session(tcp_ses->primary_server, false); 1797 kfree(tcp_ses->hostname); 1798 kfree(tcp_ses->leaf_fullpath); 1799 if (tcp_ses->ssocket) 1800 sock_release(tcp_ses->ssocket); 1801 kfree(tcp_ses); 1802 } 1803 return ERR_PTR(rc); 1804 } 1805 1806 /* this function must be called with ses_lock and chan_lock held */ 1807 static int match_session(struct cifs_ses *ses, struct smb3_fs_context *ctx) 1808 { 1809 if (ctx->sectype != Unspecified && 1810 ctx->sectype != ses->sectype) 1811 return 0; 1812 1813 /* 1814 * If an existing session is limited to less channels than 1815 * requested, it should not be reused 1816 */ 1817 if (ses->chan_max < ctx->max_channels) 1818 return 0; 1819 1820 switch (ses->sectype) { 1821 case Kerberos: 1822 if (!uid_eq(ctx->cred_uid, ses->cred_uid)) 1823 return 0; 1824 break; 1825 default: 1826 /* NULL username means anonymous session */ 1827 if (ses->user_name == NULL) { 1828 if (!ctx->nullauth) 1829 return 0; 1830 break; 1831 } 1832 1833 /* anything else takes username/password */ 1834 if (strncmp(ses->user_name, 1835 ctx->username ? ctx->username : "", 1836 CIFS_MAX_USERNAME_LEN)) 1837 return 0; 1838 if ((ctx->username && strlen(ctx->username) != 0) && 1839 ses->password != NULL && 1840 strncmp(ses->password, 1841 ctx->password ? ctx->password : "", 1842 CIFS_MAX_PASSWORD_LEN)) 1843 return 0; 1844 } 1845 1846 if (strcmp(ctx->local_nls->charset, ses->local_nls->charset)) 1847 return 0; 1848 1849 return 1; 1850 } 1851 1852 /** 1853 * cifs_setup_ipc - helper to setup the IPC tcon for the session 1854 * @ses: smb session to issue the request on 1855 * @ctx: the superblock configuration context to use for building the 1856 * new tree connection for the IPC (interprocess communication RPC) 1857 * 1858 * A new IPC connection is made and stored in the session 1859 * tcon_ipc. The IPC tcon has the same lifetime as the session. 1860 */ 1861 static int 1862 cifs_setup_ipc(struct cifs_ses *ses, struct smb3_fs_context *ctx) 1863 { 1864 int rc = 0, xid; 1865 struct cifs_tcon *tcon; 1866 char unc[SERVER_NAME_LENGTH + sizeof("//x/IPC$")] = {0}; 1867 bool seal = false; 1868 struct TCP_Server_Info *server = ses->server; 1869 1870 /* 1871 * If the mount request that resulted in the creation of the 1872 * session requires encryption, force IPC to be encrypted too. 1873 */ 1874 if (ctx->seal) { 1875 if (server->capabilities & SMB2_GLOBAL_CAP_ENCRYPTION) 1876 seal = true; 1877 else { 1878 cifs_server_dbg(VFS, 1879 "IPC: server doesn't support encryption\n"); 1880 return -EOPNOTSUPP; 1881 } 1882 } 1883 1884 tcon = tconInfoAlloc(); 1885 if (tcon == NULL) 1886 return -ENOMEM; 1887 1888 spin_lock(&server->srv_lock); 1889 scnprintf(unc, sizeof(unc), "\\\\%s\\IPC$", server->hostname); 1890 spin_unlock(&server->srv_lock); 1891 1892 xid = get_xid(); 1893 tcon->ses = ses; 1894 tcon->ipc = true; 1895 tcon->seal = seal; 1896 rc = server->ops->tree_connect(xid, ses, unc, tcon, ctx->local_nls); 1897 free_xid(xid); 1898 1899 if (rc) { 1900 cifs_server_dbg(VFS, "failed to connect to IPC (rc=%d)\n", rc); 1901 tconInfoFree(tcon); 1902 goto out; 1903 } 1904 1905 cifs_dbg(FYI, "IPC tcon rc=%d ipc tid=0x%x\n", rc, tcon->tid); 1906 1907 spin_lock(&tcon->tc_lock); 1908 tcon->status = TID_GOOD; 1909 spin_unlock(&tcon->tc_lock); 1910 ses->tcon_ipc = tcon; 1911 out: 1912 return rc; 1913 } 1914 1915 /** 1916 * cifs_free_ipc - helper to release the session IPC tcon 1917 * @ses: smb session to unmount the IPC from 1918 * 1919 * Needs to be called everytime a session is destroyed. 1920 * 1921 * On session close, the IPC is closed and the server must release all tcons of the session. 1922 * No need to send a tree disconnect here. 1923 * 1924 * Besides, it will make the server to not close durable and resilient files on session close, as 1925 * specified in MS-SMB2 3.3.5.6 Receiving an SMB2 LOGOFF Request. 1926 */ 1927 static int 1928 cifs_free_ipc(struct cifs_ses *ses) 1929 { 1930 struct cifs_tcon *tcon = ses->tcon_ipc; 1931 1932 if (tcon == NULL) 1933 return 0; 1934 1935 tconInfoFree(tcon); 1936 ses->tcon_ipc = NULL; 1937 return 0; 1938 } 1939 1940 static struct cifs_ses * 1941 cifs_find_smb_ses(struct TCP_Server_Info *server, struct smb3_fs_context *ctx) 1942 { 1943 struct cifs_ses *ses, *ret = NULL; 1944 1945 spin_lock(&cifs_tcp_ses_lock); 1946 list_for_each_entry(ses, &server->smb_ses_list, smb_ses_list) { 1947 spin_lock(&ses->ses_lock); 1948 if (ses->ses_status == SES_EXITING) { 1949 spin_unlock(&ses->ses_lock); 1950 continue; 1951 } 1952 spin_lock(&ses->chan_lock); 1953 if (match_session(ses, ctx)) { 1954 spin_unlock(&ses->chan_lock); 1955 spin_unlock(&ses->ses_lock); 1956 ret = ses; 1957 break; 1958 } 1959 spin_unlock(&ses->chan_lock); 1960 spin_unlock(&ses->ses_lock); 1961 } 1962 if (ret) 1963 cifs_smb_ses_inc_refcount(ret); 1964 spin_unlock(&cifs_tcp_ses_lock); 1965 return ret; 1966 } 1967 1968 void __cifs_put_smb_ses(struct cifs_ses *ses) 1969 { 1970 unsigned int rc, xid; 1971 unsigned int chan_count; 1972 struct TCP_Server_Info *server = ses->server; 1973 1974 spin_lock(&ses->ses_lock); 1975 if (ses->ses_status == SES_EXITING) { 1976 spin_unlock(&ses->ses_lock); 1977 return; 1978 } 1979 spin_unlock(&ses->ses_lock); 1980 1981 cifs_dbg(FYI, "%s: ses_count=%d\n", __func__, ses->ses_count); 1982 cifs_dbg(FYI, 1983 "%s: ses ipc: %s\n", __func__, ses->tcon_ipc ? ses->tcon_ipc->tree_name : "NONE"); 1984 1985 spin_lock(&cifs_tcp_ses_lock); 1986 if (--ses->ses_count > 0) { 1987 spin_unlock(&cifs_tcp_ses_lock); 1988 return; 1989 } 1990 spin_lock(&ses->ses_lock); 1991 if (ses->ses_status == SES_GOOD) 1992 ses->ses_status = SES_EXITING; 1993 spin_unlock(&ses->ses_lock); 1994 spin_unlock(&cifs_tcp_ses_lock); 1995 1996 /* ses_count can never go negative */ 1997 WARN_ON(ses->ses_count < 0); 1998 1999 spin_lock(&ses->ses_lock); 2000 if (ses->ses_status == SES_EXITING && server->ops->logoff) { 2001 spin_unlock(&ses->ses_lock); 2002 cifs_free_ipc(ses); 2003 xid = get_xid(); 2004 rc = server->ops->logoff(xid, ses); 2005 if (rc) 2006 cifs_server_dbg(VFS, "%s: Session Logoff failure rc=%d\n", 2007 __func__, rc); 2008 _free_xid(xid); 2009 } else { 2010 spin_unlock(&ses->ses_lock); 2011 cifs_free_ipc(ses); 2012 } 2013 2014 spin_lock(&cifs_tcp_ses_lock); 2015 list_del_init(&ses->smb_ses_list); 2016 spin_unlock(&cifs_tcp_ses_lock); 2017 2018 chan_count = ses->chan_count; 2019 2020 /* close any extra channels */ 2021 if (chan_count > 1) { 2022 int i; 2023 2024 for (i = 1; i < chan_count; i++) { 2025 if (ses->chans[i].iface) { 2026 kref_put(&ses->chans[i].iface->refcount, release_iface); 2027 ses->chans[i].iface = NULL; 2028 } 2029 cifs_put_tcp_session(ses->chans[i].server, 0); 2030 ses->chans[i].server = NULL; 2031 } 2032 } 2033 2034 sesInfoFree(ses); 2035 cifs_put_tcp_session(server, 0); 2036 } 2037 2038 #ifdef CONFIG_KEYS 2039 2040 /* strlen("cifs:a:") + CIFS_MAX_DOMAINNAME_LEN + 1 */ 2041 #define CIFSCREDS_DESC_SIZE (7 + CIFS_MAX_DOMAINNAME_LEN + 1) 2042 2043 /* Populate username and pw fields from keyring if possible */ 2044 static int 2045 cifs_set_cifscreds(struct smb3_fs_context *ctx, struct cifs_ses *ses) 2046 { 2047 int rc = 0; 2048 int is_domain = 0; 2049 const char *delim, *payload; 2050 char *desc; 2051 ssize_t len; 2052 struct key *key; 2053 struct TCP_Server_Info *server = ses->server; 2054 struct sockaddr_in *sa; 2055 struct sockaddr_in6 *sa6; 2056 const struct user_key_payload *upayload; 2057 2058 desc = kmalloc(CIFSCREDS_DESC_SIZE, GFP_KERNEL); 2059 if (!desc) 2060 return -ENOMEM; 2061 2062 /* try to find an address key first */ 2063 switch (server->dstaddr.ss_family) { 2064 case AF_INET: 2065 sa = (struct sockaddr_in *)&server->dstaddr; 2066 sprintf(desc, "cifs:a:%pI4", &sa->sin_addr.s_addr); 2067 break; 2068 case AF_INET6: 2069 sa6 = (struct sockaddr_in6 *)&server->dstaddr; 2070 sprintf(desc, "cifs:a:%pI6c", &sa6->sin6_addr.s6_addr); 2071 break; 2072 default: 2073 cifs_dbg(FYI, "Bad ss_family (%hu)\n", 2074 server->dstaddr.ss_family); 2075 rc = -EINVAL; 2076 goto out_err; 2077 } 2078 2079 cifs_dbg(FYI, "%s: desc=%s\n", __func__, desc); 2080 key = request_key(&key_type_logon, desc, ""); 2081 if (IS_ERR(key)) { 2082 if (!ses->domainName) { 2083 cifs_dbg(FYI, "domainName is NULL\n"); 2084 rc = PTR_ERR(key); 2085 goto out_err; 2086 } 2087 2088 /* didn't work, try to find a domain key */ 2089 sprintf(desc, "cifs:d:%s", ses->domainName); 2090 cifs_dbg(FYI, "%s: desc=%s\n", __func__, desc); 2091 key = request_key(&key_type_logon, desc, ""); 2092 if (IS_ERR(key)) { 2093 rc = PTR_ERR(key); 2094 goto out_err; 2095 } 2096 is_domain = 1; 2097 } 2098 2099 down_read(&key->sem); 2100 upayload = user_key_payload_locked(key); 2101 if (IS_ERR_OR_NULL(upayload)) { 2102 rc = upayload ? PTR_ERR(upayload) : -EINVAL; 2103 goto out_key_put; 2104 } 2105 2106 /* find first : in payload */ 2107 payload = upayload->data; 2108 delim = strnchr(payload, upayload->datalen, ':'); 2109 cifs_dbg(FYI, "payload=%s\n", payload); 2110 if (!delim) { 2111 cifs_dbg(FYI, "Unable to find ':' in payload (datalen=%d)\n", 2112 upayload->datalen); 2113 rc = -EINVAL; 2114 goto out_key_put; 2115 } 2116 2117 len = delim - payload; 2118 if (len > CIFS_MAX_USERNAME_LEN || len <= 0) { 2119 cifs_dbg(FYI, "Bad value from username search (len=%zd)\n", 2120 len); 2121 rc = -EINVAL; 2122 goto out_key_put; 2123 } 2124 2125 ctx->username = kstrndup(payload, len, GFP_KERNEL); 2126 if (!ctx->username) { 2127 cifs_dbg(FYI, "Unable to allocate %zd bytes for username\n", 2128 len); 2129 rc = -ENOMEM; 2130 goto out_key_put; 2131 } 2132 cifs_dbg(FYI, "%s: username=%s\n", __func__, ctx->username); 2133 2134 len = key->datalen - (len + 1); 2135 if (len > CIFS_MAX_PASSWORD_LEN || len <= 0) { 2136 cifs_dbg(FYI, "Bad len for password search (len=%zd)\n", len); 2137 rc = -EINVAL; 2138 kfree(ctx->username); 2139 ctx->username = NULL; 2140 goto out_key_put; 2141 } 2142 2143 ++delim; 2144 ctx->password = kstrndup(delim, len, GFP_KERNEL); 2145 if (!ctx->password) { 2146 cifs_dbg(FYI, "Unable to allocate %zd bytes for password\n", 2147 len); 2148 rc = -ENOMEM; 2149 kfree(ctx->username); 2150 ctx->username = NULL; 2151 goto out_key_put; 2152 } 2153 2154 /* 2155 * If we have a domain key then we must set the domainName in the 2156 * for the request. 2157 */ 2158 if (is_domain && ses->domainName) { 2159 ctx->domainname = kstrdup(ses->domainName, GFP_KERNEL); 2160 if (!ctx->domainname) { 2161 cifs_dbg(FYI, "Unable to allocate %zd bytes for domain\n", 2162 len); 2163 rc = -ENOMEM; 2164 kfree(ctx->username); 2165 ctx->username = NULL; 2166 kfree_sensitive(ctx->password); 2167 ctx->password = NULL; 2168 goto out_key_put; 2169 } 2170 } 2171 2172 strscpy(ctx->workstation_name, ses->workstation_name, sizeof(ctx->workstation_name)); 2173 2174 out_key_put: 2175 up_read(&key->sem); 2176 key_put(key); 2177 out_err: 2178 kfree(desc); 2179 cifs_dbg(FYI, "%s: returning %d\n", __func__, rc); 2180 return rc; 2181 } 2182 #else /* ! CONFIG_KEYS */ 2183 static inline int 2184 cifs_set_cifscreds(struct smb3_fs_context *ctx __attribute__((unused)), 2185 struct cifs_ses *ses __attribute__((unused))) 2186 { 2187 return -ENOSYS; 2188 } 2189 #endif /* CONFIG_KEYS */ 2190 2191 /** 2192 * cifs_get_smb_ses - get a session matching @ctx data from @server 2193 * @server: server to setup the session to 2194 * @ctx: superblock configuration context to use to setup the session 2195 * 2196 * This function assumes it is being called from cifs_mount() where we 2197 * already got a server reference (server refcount +1). See 2198 * cifs_get_tcon() for refcount explanations. 2199 */ 2200 struct cifs_ses * 2201 cifs_get_smb_ses(struct TCP_Server_Info *server, struct smb3_fs_context *ctx) 2202 { 2203 int rc = 0; 2204 unsigned int xid; 2205 struct cifs_ses *ses; 2206 struct sockaddr_in *addr = (struct sockaddr_in *)&server->dstaddr; 2207 struct sockaddr_in6 *addr6 = (struct sockaddr_in6 *)&server->dstaddr; 2208 2209 xid = get_xid(); 2210 2211 ses = cifs_find_smb_ses(server, ctx); 2212 if (ses) { 2213 cifs_dbg(FYI, "Existing smb sess found (status=%d)\n", 2214 ses->ses_status); 2215 2216 spin_lock(&ses->chan_lock); 2217 if (cifs_chan_needs_reconnect(ses, server)) { 2218 spin_unlock(&ses->chan_lock); 2219 cifs_dbg(FYI, "Session needs reconnect\n"); 2220 2221 mutex_lock(&ses->session_mutex); 2222 rc = cifs_negotiate_protocol(xid, ses, server); 2223 if (rc) { 2224 mutex_unlock(&ses->session_mutex); 2225 /* problem -- put our ses reference */ 2226 cifs_put_smb_ses(ses); 2227 free_xid(xid); 2228 return ERR_PTR(rc); 2229 } 2230 2231 rc = cifs_setup_session(xid, ses, server, 2232 ctx->local_nls); 2233 if (rc) { 2234 mutex_unlock(&ses->session_mutex); 2235 /* problem -- put our reference */ 2236 cifs_put_smb_ses(ses); 2237 free_xid(xid); 2238 return ERR_PTR(rc); 2239 } 2240 mutex_unlock(&ses->session_mutex); 2241 2242 spin_lock(&ses->chan_lock); 2243 } 2244 spin_unlock(&ses->chan_lock); 2245 2246 /* existing SMB ses has a server reference already */ 2247 cifs_put_tcp_session(server, 0); 2248 free_xid(xid); 2249 return ses; 2250 } 2251 2252 rc = -ENOMEM; 2253 2254 cifs_dbg(FYI, "Existing smb sess not found\n"); 2255 ses = sesInfoAlloc(); 2256 if (ses == NULL) 2257 goto get_ses_fail; 2258 2259 /* new SMB session uses our server ref */ 2260 ses->server = server; 2261 if (server->dstaddr.ss_family == AF_INET6) 2262 sprintf(ses->ip_addr, "%pI6", &addr6->sin6_addr); 2263 else 2264 sprintf(ses->ip_addr, "%pI4", &addr->sin_addr); 2265 2266 if (ctx->username) { 2267 ses->user_name = kstrdup(ctx->username, GFP_KERNEL); 2268 if (!ses->user_name) 2269 goto get_ses_fail; 2270 } 2271 2272 /* ctx->password freed at unmount */ 2273 if (ctx->password) { 2274 ses->password = kstrdup(ctx->password, GFP_KERNEL); 2275 if (!ses->password) 2276 goto get_ses_fail; 2277 } 2278 if (ctx->domainname) { 2279 ses->domainName = kstrdup(ctx->domainname, GFP_KERNEL); 2280 if (!ses->domainName) 2281 goto get_ses_fail; 2282 } 2283 2284 strscpy(ses->workstation_name, ctx->workstation_name, sizeof(ses->workstation_name)); 2285 2286 if (ctx->domainauto) 2287 ses->domainAuto = ctx->domainauto; 2288 ses->cred_uid = ctx->cred_uid; 2289 ses->linux_uid = ctx->linux_uid; 2290 2291 ses->sectype = ctx->sectype; 2292 ses->sign = ctx->sign; 2293 ses->local_nls = load_nls(ctx->local_nls->charset); 2294 2295 /* add server as first channel */ 2296 spin_lock(&ses->chan_lock); 2297 ses->chans[0].server = server; 2298 ses->chan_count = 1; 2299 ses->chan_max = ctx->multichannel ? ctx->max_channels:1; 2300 ses->chans_need_reconnect = 1; 2301 spin_unlock(&ses->chan_lock); 2302 2303 mutex_lock(&ses->session_mutex); 2304 rc = cifs_negotiate_protocol(xid, ses, server); 2305 if (!rc) 2306 rc = cifs_setup_session(xid, ses, server, ctx->local_nls); 2307 mutex_unlock(&ses->session_mutex); 2308 2309 /* each channel uses a different signing key */ 2310 spin_lock(&ses->chan_lock); 2311 memcpy(ses->chans[0].signkey, ses->smb3signingkey, 2312 sizeof(ses->smb3signingkey)); 2313 spin_unlock(&ses->chan_lock); 2314 2315 if (rc) 2316 goto get_ses_fail; 2317 2318 /* 2319 * success, put it on the list and add it as first channel 2320 * note: the session becomes active soon after this. So you'll 2321 * need to lock before changing something in the session. 2322 */ 2323 spin_lock(&cifs_tcp_ses_lock); 2324 ses->dfs_root_ses = ctx->dfs_root_ses; 2325 if (ses->dfs_root_ses) 2326 ses->dfs_root_ses->ses_count++; 2327 list_add(&ses->smb_ses_list, &server->smb_ses_list); 2328 spin_unlock(&cifs_tcp_ses_lock); 2329 2330 cifs_setup_ipc(ses, ctx); 2331 2332 free_xid(xid); 2333 2334 return ses; 2335 2336 get_ses_fail: 2337 sesInfoFree(ses); 2338 free_xid(xid); 2339 return ERR_PTR(rc); 2340 } 2341 2342 /* this function must be called with tc_lock held */ 2343 static int match_tcon(struct cifs_tcon *tcon, struct smb3_fs_context *ctx) 2344 { 2345 struct TCP_Server_Info *server = tcon->ses->server; 2346 2347 if (tcon->status == TID_EXITING) 2348 return 0; 2349 2350 if (tcon->origin_fullpath) { 2351 if (!ctx->source || 2352 !dfs_src_pathname_equal(ctx->source, 2353 tcon->origin_fullpath)) 2354 return 0; 2355 } else if (!server->leaf_fullpath && 2356 strncmp(tcon->tree_name, ctx->UNC, MAX_TREE_SIZE)) { 2357 return 0; 2358 } 2359 if (tcon->seal != ctx->seal) 2360 return 0; 2361 if (tcon->snapshot_time != ctx->snapshot_time) 2362 return 0; 2363 if (tcon->handle_timeout != ctx->handle_timeout) 2364 return 0; 2365 if (tcon->no_lease != ctx->no_lease) 2366 return 0; 2367 if (tcon->nodelete != ctx->nodelete) 2368 return 0; 2369 return 1; 2370 } 2371 2372 static struct cifs_tcon * 2373 cifs_find_tcon(struct cifs_ses *ses, struct smb3_fs_context *ctx) 2374 { 2375 struct cifs_tcon *tcon; 2376 2377 spin_lock(&cifs_tcp_ses_lock); 2378 list_for_each_entry(tcon, &ses->tcon_list, tcon_list) { 2379 spin_lock(&tcon->tc_lock); 2380 if (!match_tcon(tcon, ctx)) { 2381 spin_unlock(&tcon->tc_lock); 2382 continue; 2383 } 2384 ++tcon->tc_count; 2385 spin_unlock(&tcon->tc_lock); 2386 spin_unlock(&cifs_tcp_ses_lock); 2387 return tcon; 2388 } 2389 spin_unlock(&cifs_tcp_ses_lock); 2390 return NULL; 2391 } 2392 2393 void 2394 cifs_put_tcon(struct cifs_tcon *tcon) 2395 { 2396 unsigned int xid; 2397 struct cifs_ses *ses; 2398 2399 /* 2400 * IPC tcon share the lifetime of their session and are 2401 * destroyed in the session put function 2402 */ 2403 if (tcon == NULL || tcon->ipc) 2404 return; 2405 2406 ses = tcon->ses; 2407 cifs_dbg(FYI, "%s: tc_count=%d\n", __func__, tcon->tc_count); 2408 spin_lock(&cifs_tcp_ses_lock); 2409 spin_lock(&tcon->tc_lock); 2410 if (--tcon->tc_count > 0) { 2411 spin_unlock(&tcon->tc_lock); 2412 spin_unlock(&cifs_tcp_ses_lock); 2413 return; 2414 } 2415 2416 /* tc_count can never go negative */ 2417 WARN_ON(tcon->tc_count < 0); 2418 2419 list_del_init(&tcon->tcon_list); 2420 tcon->status = TID_EXITING; 2421 spin_unlock(&tcon->tc_lock); 2422 spin_unlock(&cifs_tcp_ses_lock); 2423 2424 /* cancel polling of interfaces */ 2425 cancel_delayed_work_sync(&tcon->query_interfaces); 2426 #ifdef CONFIG_CIFS_DFS_UPCALL 2427 cancel_delayed_work_sync(&tcon->dfs_cache_work); 2428 #endif 2429 2430 if (tcon->use_witness) { 2431 int rc; 2432 2433 rc = cifs_swn_unregister(tcon); 2434 if (rc < 0) { 2435 cifs_dbg(VFS, "%s: Failed to unregister for witness notifications: %d\n", 2436 __func__, rc); 2437 } 2438 } 2439 2440 xid = get_xid(); 2441 if (ses->server->ops->tree_disconnect) 2442 ses->server->ops->tree_disconnect(xid, tcon); 2443 _free_xid(xid); 2444 2445 cifs_fscache_release_super_cookie(tcon); 2446 tconInfoFree(tcon); 2447 cifs_put_smb_ses(ses); 2448 } 2449 2450 /** 2451 * cifs_get_tcon - get a tcon matching @ctx data from @ses 2452 * @ses: smb session to issue the request on 2453 * @ctx: the superblock configuration context to use for building the 2454 * 2455 * - tcon refcount is the number of mount points using the tcon. 2456 * - ses refcount is the number of tcon using the session. 2457 * 2458 * 1. This function assumes it is being called from cifs_mount() where 2459 * we already got a session reference (ses refcount +1). 2460 * 2461 * 2. Since we're in the context of adding a mount point, the end 2462 * result should be either: 2463 * 2464 * a) a new tcon already allocated with refcount=1 (1 mount point) and 2465 * its session refcount incremented (1 new tcon). This +1 was 2466 * already done in (1). 2467 * 2468 * b) an existing tcon with refcount+1 (add a mount point to it) and 2469 * identical ses refcount (no new tcon). Because of (1) we need to 2470 * decrement the ses refcount. 2471 */ 2472 static struct cifs_tcon * 2473 cifs_get_tcon(struct cifs_ses *ses, struct smb3_fs_context *ctx) 2474 { 2475 int rc, xid; 2476 struct cifs_tcon *tcon; 2477 2478 tcon = cifs_find_tcon(ses, ctx); 2479 if (tcon) { 2480 /* 2481 * tcon has refcount already incremented but we need to 2482 * decrement extra ses reference gotten by caller (case b) 2483 */ 2484 cifs_dbg(FYI, "Found match on UNC path\n"); 2485 cifs_put_smb_ses(ses); 2486 return tcon; 2487 } 2488 2489 if (!ses->server->ops->tree_connect) { 2490 rc = -ENOSYS; 2491 goto out_fail; 2492 } 2493 2494 tcon = tconInfoAlloc(); 2495 if (tcon == NULL) { 2496 rc = -ENOMEM; 2497 goto out_fail; 2498 } 2499 2500 if (ctx->snapshot_time) { 2501 if (ses->server->vals->protocol_id == 0) { 2502 cifs_dbg(VFS, 2503 "Use SMB2 or later for snapshot mount option\n"); 2504 rc = -EOPNOTSUPP; 2505 goto out_fail; 2506 } else 2507 tcon->snapshot_time = ctx->snapshot_time; 2508 } 2509 2510 if (ctx->handle_timeout) { 2511 if (ses->server->vals->protocol_id == 0) { 2512 cifs_dbg(VFS, 2513 "Use SMB2.1 or later for handle timeout option\n"); 2514 rc = -EOPNOTSUPP; 2515 goto out_fail; 2516 } else 2517 tcon->handle_timeout = ctx->handle_timeout; 2518 } 2519 2520 tcon->ses = ses; 2521 if (ctx->password) { 2522 tcon->password = kstrdup(ctx->password, GFP_KERNEL); 2523 if (!tcon->password) { 2524 rc = -ENOMEM; 2525 goto out_fail; 2526 } 2527 } 2528 2529 if (ctx->seal) { 2530 if (ses->server->vals->protocol_id == 0) { 2531 cifs_dbg(VFS, 2532 "SMB3 or later required for encryption\n"); 2533 rc = -EOPNOTSUPP; 2534 goto out_fail; 2535 } else if (tcon->ses->server->capabilities & 2536 SMB2_GLOBAL_CAP_ENCRYPTION) 2537 tcon->seal = true; 2538 else { 2539 cifs_dbg(VFS, "Encryption is not supported on share\n"); 2540 rc = -EOPNOTSUPP; 2541 goto out_fail; 2542 } 2543 } 2544 2545 if (ctx->linux_ext) { 2546 if (ses->server->posix_ext_supported) { 2547 tcon->posix_extensions = true; 2548 pr_warn_once("SMB3.11 POSIX Extensions are experimental\n"); 2549 } else if ((ses->server->vals->protocol_id == SMB311_PROT_ID) || 2550 (strcmp(ses->server->vals->version_string, 2551 SMB3ANY_VERSION_STRING) == 0) || 2552 (strcmp(ses->server->vals->version_string, 2553 SMBDEFAULT_VERSION_STRING) == 0)) { 2554 cifs_dbg(VFS, "Server does not support mounting with posix SMB3.11 extensions\n"); 2555 rc = -EOPNOTSUPP; 2556 goto out_fail; 2557 } else { 2558 cifs_dbg(VFS, "Check vers= mount option. SMB3.11 " 2559 "disabled but required for POSIX extensions\n"); 2560 rc = -EOPNOTSUPP; 2561 goto out_fail; 2562 } 2563 } 2564 2565 xid = get_xid(); 2566 rc = ses->server->ops->tree_connect(xid, ses, ctx->UNC, tcon, 2567 ctx->local_nls); 2568 free_xid(xid); 2569 cifs_dbg(FYI, "Tcon rc = %d\n", rc); 2570 if (rc) 2571 goto out_fail; 2572 2573 tcon->use_persistent = false; 2574 /* check if SMB2 or later, CIFS does not support persistent handles */ 2575 if (ctx->persistent) { 2576 if (ses->server->vals->protocol_id == 0) { 2577 cifs_dbg(VFS, 2578 "SMB3 or later required for persistent handles\n"); 2579 rc = -EOPNOTSUPP; 2580 goto out_fail; 2581 } else if (ses->server->capabilities & 2582 SMB2_GLOBAL_CAP_PERSISTENT_HANDLES) 2583 tcon->use_persistent = true; 2584 else /* persistent handles requested but not supported */ { 2585 cifs_dbg(VFS, 2586 "Persistent handles not supported on share\n"); 2587 rc = -EOPNOTSUPP; 2588 goto out_fail; 2589 } 2590 } else if ((tcon->capabilities & SMB2_SHARE_CAP_CONTINUOUS_AVAILABILITY) 2591 && (ses->server->capabilities & SMB2_GLOBAL_CAP_PERSISTENT_HANDLES) 2592 && (ctx->nopersistent == false)) { 2593 cifs_dbg(FYI, "enabling persistent handles\n"); 2594 tcon->use_persistent = true; 2595 } else if (ctx->resilient) { 2596 if (ses->server->vals->protocol_id == 0) { 2597 cifs_dbg(VFS, 2598 "SMB2.1 or later required for resilient handles\n"); 2599 rc = -EOPNOTSUPP; 2600 goto out_fail; 2601 } 2602 tcon->use_resilient = true; 2603 } 2604 2605 tcon->use_witness = false; 2606 if (IS_ENABLED(CONFIG_CIFS_SWN_UPCALL) && ctx->witness) { 2607 if (ses->server->vals->protocol_id >= SMB30_PROT_ID) { 2608 if (tcon->capabilities & SMB2_SHARE_CAP_CLUSTER) { 2609 /* 2610 * Set witness in use flag in first place 2611 * to retry registration in the echo task 2612 */ 2613 tcon->use_witness = true; 2614 /* And try to register immediately */ 2615 rc = cifs_swn_register(tcon); 2616 if (rc < 0) { 2617 cifs_dbg(VFS, "Failed to register for witness notifications: %d\n", rc); 2618 goto out_fail; 2619 } 2620 } else { 2621 /* TODO: try to extend for non-cluster uses (eg multichannel) */ 2622 cifs_dbg(VFS, "witness requested on mount but no CLUSTER capability on share\n"); 2623 rc = -EOPNOTSUPP; 2624 goto out_fail; 2625 } 2626 } else { 2627 cifs_dbg(VFS, "SMB3 or later required for witness option\n"); 2628 rc = -EOPNOTSUPP; 2629 goto out_fail; 2630 } 2631 } 2632 2633 /* If the user really knows what they are doing they can override */ 2634 if (tcon->share_flags & SMB2_SHAREFLAG_NO_CACHING) { 2635 if (ctx->cache_ro) 2636 cifs_dbg(VFS, "cache=ro requested on mount but NO_CACHING flag set on share\n"); 2637 else if (ctx->cache_rw) 2638 cifs_dbg(VFS, "cache=singleclient requested on mount but NO_CACHING flag set on share\n"); 2639 } 2640 2641 if (ctx->no_lease) { 2642 if (ses->server->vals->protocol_id == 0) { 2643 cifs_dbg(VFS, 2644 "SMB2 or later required for nolease option\n"); 2645 rc = -EOPNOTSUPP; 2646 goto out_fail; 2647 } else 2648 tcon->no_lease = ctx->no_lease; 2649 } 2650 2651 /* 2652 * We can have only one retry value for a connection to a share so for 2653 * resources mounted more than once to the same server share the last 2654 * value passed in for the retry flag is used. 2655 */ 2656 tcon->retry = ctx->retry; 2657 tcon->nocase = ctx->nocase; 2658 tcon->broken_sparse_sup = ctx->no_sparse; 2659 if (ses->server->capabilities & SMB2_GLOBAL_CAP_DIRECTORY_LEASING) 2660 tcon->nohandlecache = ctx->nohandlecache; 2661 else 2662 tcon->nohandlecache = true; 2663 tcon->nodelete = ctx->nodelete; 2664 tcon->local_lease = ctx->local_lease; 2665 INIT_LIST_HEAD(&tcon->pending_opens); 2666 tcon->status = TID_GOOD; 2667 2668 INIT_DELAYED_WORK(&tcon->query_interfaces, 2669 smb2_query_server_interfaces); 2670 if (ses->server->dialect >= SMB30_PROT_ID && 2671 (ses->server->capabilities & SMB2_GLOBAL_CAP_MULTI_CHANNEL)) { 2672 /* schedule query interfaces poll */ 2673 queue_delayed_work(cifsiod_wq, &tcon->query_interfaces, 2674 (SMB_INTERFACE_POLL_INTERVAL * HZ)); 2675 } 2676 #ifdef CONFIG_CIFS_DFS_UPCALL 2677 INIT_DELAYED_WORK(&tcon->dfs_cache_work, dfs_cache_refresh); 2678 #endif 2679 spin_lock(&cifs_tcp_ses_lock); 2680 list_add(&tcon->tcon_list, &ses->tcon_list); 2681 spin_unlock(&cifs_tcp_ses_lock); 2682 2683 return tcon; 2684 2685 out_fail: 2686 tconInfoFree(tcon); 2687 return ERR_PTR(rc); 2688 } 2689 2690 void 2691 cifs_put_tlink(struct tcon_link *tlink) 2692 { 2693 if (!tlink || IS_ERR(tlink)) 2694 return; 2695 2696 if (!atomic_dec_and_test(&tlink->tl_count) || 2697 test_bit(TCON_LINK_IN_TREE, &tlink->tl_flags)) { 2698 tlink->tl_time = jiffies; 2699 return; 2700 } 2701 2702 if (!IS_ERR(tlink_tcon(tlink))) 2703 cifs_put_tcon(tlink_tcon(tlink)); 2704 kfree(tlink); 2705 return; 2706 } 2707 2708 static int 2709 compare_mount_options(struct super_block *sb, struct cifs_mnt_data *mnt_data) 2710 { 2711 struct cifs_sb_info *old = CIFS_SB(sb); 2712 struct cifs_sb_info *new = mnt_data->cifs_sb; 2713 unsigned int oldflags = old->mnt_cifs_flags & CIFS_MOUNT_MASK; 2714 unsigned int newflags = new->mnt_cifs_flags & CIFS_MOUNT_MASK; 2715 2716 if ((sb->s_flags & CIFS_MS_MASK) != (mnt_data->flags & CIFS_MS_MASK)) 2717 return 0; 2718 2719 if (old->mnt_cifs_serverino_autodisabled) 2720 newflags &= ~CIFS_MOUNT_SERVER_INUM; 2721 2722 if (oldflags != newflags) 2723 return 0; 2724 2725 /* 2726 * We want to share sb only if we don't specify an r/wsize or 2727 * specified r/wsize is greater than or equal to existing one. 2728 */ 2729 if (new->ctx->wsize && new->ctx->wsize < old->ctx->wsize) 2730 return 0; 2731 2732 if (new->ctx->rsize && new->ctx->rsize < old->ctx->rsize) 2733 return 0; 2734 2735 if (!uid_eq(old->ctx->linux_uid, new->ctx->linux_uid) || 2736 !gid_eq(old->ctx->linux_gid, new->ctx->linux_gid)) 2737 return 0; 2738 2739 if (old->ctx->file_mode != new->ctx->file_mode || 2740 old->ctx->dir_mode != new->ctx->dir_mode) 2741 return 0; 2742 2743 if (strcmp(old->local_nls->charset, new->local_nls->charset)) 2744 return 0; 2745 2746 if (old->ctx->acregmax != new->ctx->acregmax) 2747 return 0; 2748 if (old->ctx->acdirmax != new->ctx->acdirmax) 2749 return 0; 2750 if (old->ctx->closetimeo != new->ctx->closetimeo) 2751 return 0; 2752 2753 return 1; 2754 } 2755 2756 static int match_prepath(struct super_block *sb, 2757 struct cifs_tcon *tcon, 2758 struct cifs_mnt_data *mnt_data) 2759 { 2760 struct smb3_fs_context *ctx = mnt_data->ctx; 2761 struct cifs_sb_info *old = CIFS_SB(sb); 2762 struct cifs_sb_info *new = mnt_data->cifs_sb; 2763 bool old_set = (old->mnt_cifs_flags & CIFS_MOUNT_USE_PREFIX_PATH) && 2764 old->prepath; 2765 bool new_set = (new->mnt_cifs_flags & CIFS_MOUNT_USE_PREFIX_PATH) && 2766 new->prepath; 2767 2768 if (tcon->origin_fullpath && 2769 dfs_src_pathname_equal(tcon->origin_fullpath, ctx->source)) 2770 return 1; 2771 2772 if (old_set && new_set && !strcmp(new->prepath, old->prepath)) 2773 return 1; 2774 else if (!old_set && !new_set) 2775 return 1; 2776 2777 return 0; 2778 } 2779 2780 int 2781 cifs_match_super(struct super_block *sb, void *data) 2782 { 2783 struct cifs_mnt_data *mnt_data = data; 2784 struct smb3_fs_context *ctx; 2785 struct cifs_sb_info *cifs_sb; 2786 struct TCP_Server_Info *tcp_srv; 2787 struct cifs_ses *ses; 2788 struct cifs_tcon *tcon; 2789 struct tcon_link *tlink; 2790 int rc = 0; 2791 2792 spin_lock(&cifs_tcp_ses_lock); 2793 cifs_sb = CIFS_SB(sb); 2794 2795 /* We do not want to use a superblock that has been shutdown */ 2796 if (CIFS_MOUNT_SHUTDOWN & cifs_sb->mnt_cifs_flags) { 2797 spin_unlock(&cifs_tcp_ses_lock); 2798 return 0; 2799 } 2800 2801 tlink = cifs_get_tlink(cifs_sb_master_tlink(cifs_sb)); 2802 if (IS_ERR_OR_NULL(tlink)) { 2803 pr_warn_once("%s: skip super matching due to bad tlink(%p)\n", 2804 __func__, tlink); 2805 spin_unlock(&cifs_tcp_ses_lock); 2806 return 0; 2807 } 2808 tcon = tlink_tcon(tlink); 2809 ses = tcon->ses; 2810 tcp_srv = ses->server; 2811 2812 ctx = mnt_data->ctx; 2813 2814 spin_lock(&tcp_srv->srv_lock); 2815 spin_lock(&ses->ses_lock); 2816 spin_lock(&ses->chan_lock); 2817 spin_lock(&tcon->tc_lock); 2818 if (!match_server(tcp_srv, ctx, true) || 2819 !match_session(ses, ctx) || 2820 !match_tcon(tcon, ctx) || 2821 !match_prepath(sb, tcon, mnt_data)) { 2822 rc = 0; 2823 goto out; 2824 } 2825 2826 rc = compare_mount_options(sb, mnt_data); 2827 out: 2828 spin_unlock(&tcon->tc_lock); 2829 spin_unlock(&ses->chan_lock); 2830 spin_unlock(&ses->ses_lock); 2831 spin_unlock(&tcp_srv->srv_lock); 2832 2833 spin_unlock(&cifs_tcp_ses_lock); 2834 cifs_put_tlink(tlink); 2835 return rc; 2836 } 2837 2838 #ifdef CONFIG_DEBUG_LOCK_ALLOC 2839 static struct lock_class_key cifs_key[2]; 2840 static struct lock_class_key cifs_slock_key[2]; 2841 2842 static inline void 2843 cifs_reclassify_socket4(struct socket *sock) 2844 { 2845 struct sock *sk = sock->sk; 2846 BUG_ON(!sock_allow_reclassification(sk)); 2847 sock_lock_init_class_and_name(sk, "slock-AF_INET-CIFS", 2848 &cifs_slock_key[0], "sk_lock-AF_INET-CIFS", &cifs_key[0]); 2849 } 2850 2851 static inline void 2852 cifs_reclassify_socket6(struct socket *sock) 2853 { 2854 struct sock *sk = sock->sk; 2855 BUG_ON(!sock_allow_reclassification(sk)); 2856 sock_lock_init_class_and_name(sk, "slock-AF_INET6-CIFS", 2857 &cifs_slock_key[1], "sk_lock-AF_INET6-CIFS", &cifs_key[1]); 2858 } 2859 #else 2860 static inline void 2861 cifs_reclassify_socket4(struct socket *sock) 2862 { 2863 } 2864 2865 static inline void 2866 cifs_reclassify_socket6(struct socket *sock) 2867 { 2868 } 2869 #endif 2870 2871 /* See RFC1001 section 14 on representation of Netbios names */ 2872 static void rfc1002mangle(char *target, char *source, unsigned int length) 2873 { 2874 unsigned int i, j; 2875 2876 for (i = 0, j = 0; i < (length); i++) { 2877 /* mask a nibble at a time and encode */ 2878 target[j] = 'A' + (0x0F & (source[i] >> 4)); 2879 target[j+1] = 'A' + (0x0F & source[i]); 2880 j += 2; 2881 } 2882 2883 } 2884 2885 static int 2886 bind_socket(struct TCP_Server_Info *server) 2887 { 2888 int rc = 0; 2889 if (server->srcaddr.ss_family != AF_UNSPEC) { 2890 /* Bind to the specified local IP address */ 2891 struct socket *socket = server->ssocket; 2892 rc = socket->ops->bind(socket, 2893 (struct sockaddr *) &server->srcaddr, 2894 sizeof(server->srcaddr)); 2895 if (rc < 0) { 2896 struct sockaddr_in *saddr4; 2897 struct sockaddr_in6 *saddr6; 2898 saddr4 = (struct sockaddr_in *)&server->srcaddr; 2899 saddr6 = (struct sockaddr_in6 *)&server->srcaddr; 2900 if (saddr6->sin6_family == AF_INET6) 2901 cifs_server_dbg(VFS, "Failed to bind to: %pI6c, error: %d\n", 2902 &saddr6->sin6_addr, rc); 2903 else 2904 cifs_server_dbg(VFS, "Failed to bind to: %pI4, error: %d\n", 2905 &saddr4->sin_addr.s_addr, rc); 2906 } 2907 } 2908 return rc; 2909 } 2910 2911 static int 2912 ip_rfc1001_connect(struct TCP_Server_Info *server) 2913 { 2914 int rc = 0; 2915 /* 2916 * some servers require RFC1001 sessinit before sending 2917 * negprot - BB check reconnection in case where second 2918 * sessinit is sent but no second negprot 2919 */ 2920 struct rfc1002_session_packet req = {}; 2921 struct smb_hdr *smb_buf = (struct smb_hdr *)&req; 2922 unsigned int len; 2923 2924 req.trailer.session_req.called_len = sizeof(req.trailer.session_req.called_name); 2925 2926 if (server->server_RFC1001_name[0] != 0) 2927 rfc1002mangle(req.trailer.session_req.called_name, 2928 server->server_RFC1001_name, 2929 RFC1001_NAME_LEN_WITH_NULL); 2930 else 2931 rfc1002mangle(req.trailer.session_req.called_name, 2932 DEFAULT_CIFS_CALLED_NAME, 2933 RFC1001_NAME_LEN_WITH_NULL); 2934 2935 req.trailer.session_req.calling_len = sizeof(req.trailer.session_req.calling_name); 2936 2937 /* calling name ends in null (byte 16) from old smb convention */ 2938 if (server->workstation_RFC1001_name[0] != 0) 2939 rfc1002mangle(req.trailer.session_req.calling_name, 2940 server->workstation_RFC1001_name, 2941 RFC1001_NAME_LEN_WITH_NULL); 2942 else 2943 rfc1002mangle(req.trailer.session_req.calling_name, 2944 "LINUX_CIFS_CLNT", 2945 RFC1001_NAME_LEN_WITH_NULL); 2946 2947 /* 2948 * As per rfc1002, @len must be the number of bytes that follows the 2949 * length field of a rfc1002 session request payload. 2950 */ 2951 len = sizeof(req) - offsetof(struct rfc1002_session_packet, trailer.session_req); 2952 2953 smb_buf->smb_buf_length = cpu_to_be32((RFC1002_SESSION_REQUEST << 24) | len); 2954 rc = smb_send(server, smb_buf, len); 2955 /* 2956 * RFC1001 layer in at least one server requires very short break before 2957 * negprot presumably because not expecting negprot to follow so fast. 2958 * This is a simple solution that works without complicating the code 2959 * and causes no significant slowing down on mount for everyone else 2960 */ 2961 usleep_range(1000, 2000); 2962 2963 return rc; 2964 } 2965 2966 static int 2967 generic_ip_connect(struct TCP_Server_Info *server) 2968 { 2969 struct sockaddr *saddr; 2970 struct socket *socket; 2971 int slen, sfamily; 2972 __be16 sport; 2973 int rc = 0; 2974 2975 saddr = (struct sockaddr *) &server->dstaddr; 2976 2977 if (server->dstaddr.ss_family == AF_INET6) { 2978 struct sockaddr_in6 *ipv6 = (struct sockaddr_in6 *)&server->dstaddr; 2979 2980 sport = ipv6->sin6_port; 2981 slen = sizeof(struct sockaddr_in6); 2982 sfamily = AF_INET6; 2983 cifs_dbg(FYI, "%s: connecting to [%pI6]:%d\n", __func__, &ipv6->sin6_addr, 2984 ntohs(sport)); 2985 } else { 2986 struct sockaddr_in *ipv4 = (struct sockaddr_in *)&server->dstaddr; 2987 2988 sport = ipv4->sin_port; 2989 slen = sizeof(struct sockaddr_in); 2990 sfamily = AF_INET; 2991 cifs_dbg(FYI, "%s: connecting to %pI4:%d\n", __func__, &ipv4->sin_addr, 2992 ntohs(sport)); 2993 } 2994 2995 if (server->ssocket) { 2996 socket = server->ssocket; 2997 } else { 2998 rc = __sock_create(cifs_net_ns(server), sfamily, SOCK_STREAM, 2999 IPPROTO_TCP, &server->ssocket, 1); 3000 if (rc < 0) { 3001 cifs_server_dbg(VFS, "Error %d creating socket\n", rc); 3002 return rc; 3003 } 3004 3005 /* BB other socket options to set KEEPALIVE, NODELAY? */ 3006 cifs_dbg(FYI, "Socket created\n"); 3007 socket = server->ssocket; 3008 socket->sk->sk_allocation = GFP_NOFS; 3009 socket->sk->sk_use_task_frag = false; 3010 if (sfamily == AF_INET6) 3011 cifs_reclassify_socket6(socket); 3012 else 3013 cifs_reclassify_socket4(socket); 3014 } 3015 3016 rc = bind_socket(server); 3017 if (rc < 0) 3018 return rc; 3019 3020 /* 3021 * Eventually check for other socket options to change from 3022 * the default. sock_setsockopt not used because it expects 3023 * user space buffer 3024 */ 3025 socket->sk->sk_rcvtimeo = 7 * HZ; 3026 socket->sk->sk_sndtimeo = 5 * HZ; 3027 3028 /* make the bufsizes depend on wsize/rsize and max requests */ 3029 if (server->noautotune) { 3030 if (socket->sk->sk_sndbuf < (200 * 1024)) 3031 socket->sk->sk_sndbuf = 200 * 1024; 3032 if (socket->sk->sk_rcvbuf < (140 * 1024)) 3033 socket->sk->sk_rcvbuf = 140 * 1024; 3034 } 3035 3036 if (server->tcp_nodelay) 3037 tcp_sock_set_nodelay(socket->sk); 3038 3039 cifs_dbg(FYI, "sndbuf %d rcvbuf %d rcvtimeo 0x%lx\n", 3040 socket->sk->sk_sndbuf, 3041 socket->sk->sk_rcvbuf, socket->sk->sk_rcvtimeo); 3042 3043 rc = socket->ops->connect(socket, saddr, slen, 3044 server->noblockcnt ? O_NONBLOCK : 0); 3045 /* 3046 * When mounting SMB root file systems, we do not want to block in 3047 * connect. Otherwise bail out and then let cifs_reconnect() perform 3048 * reconnect failover - if possible. 3049 */ 3050 if (server->noblockcnt && rc == -EINPROGRESS) 3051 rc = 0; 3052 if (rc < 0) { 3053 cifs_dbg(FYI, "Error %d connecting to server\n", rc); 3054 trace_smb3_connect_err(server->hostname, server->conn_id, &server->dstaddr, rc); 3055 sock_release(socket); 3056 server->ssocket = NULL; 3057 return rc; 3058 } 3059 trace_smb3_connect_done(server->hostname, server->conn_id, &server->dstaddr); 3060 if (sport == htons(RFC1001_PORT)) 3061 rc = ip_rfc1001_connect(server); 3062 3063 return rc; 3064 } 3065 3066 static int 3067 ip_connect(struct TCP_Server_Info *server) 3068 { 3069 __be16 *sport; 3070 struct sockaddr_in6 *addr6 = (struct sockaddr_in6 *)&server->dstaddr; 3071 struct sockaddr_in *addr = (struct sockaddr_in *)&server->dstaddr; 3072 3073 if (server->dstaddr.ss_family == AF_INET6) 3074 sport = &addr6->sin6_port; 3075 else 3076 sport = &addr->sin_port; 3077 3078 if (*sport == 0) { 3079 int rc; 3080 3081 /* try with 445 port at first */ 3082 *sport = htons(CIFS_PORT); 3083 3084 rc = generic_ip_connect(server); 3085 if (rc >= 0) 3086 return rc; 3087 3088 /* if it failed, try with 139 port */ 3089 *sport = htons(RFC1001_PORT); 3090 } 3091 3092 return generic_ip_connect(server); 3093 } 3094 3095 #ifdef CONFIG_CIFS_ALLOW_INSECURE_LEGACY 3096 void reset_cifs_unix_caps(unsigned int xid, struct cifs_tcon *tcon, 3097 struct cifs_sb_info *cifs_sb, struct smb3_fs_context *ctx) 3098 { 3099 /* 3100 * If we are reconnecting then should we check to see if 3101 * any requested capabilities changed locally e.g. via 3102 * remount but we can not do much about it here 3103 * if they have (even if we could detect it by the following) 3104 * Perhaps we could add a backpointer to array of sb from tcon 3105 * or if we change to make all sb to same share the same 3106 * sb as NFS - then we only have one backpointer to sb. 3107 * What if we wanted to mount the server share twice once with 3108 * and once without posixacls or posix paths? 3109 */ 3110 __u64 saved_cap = le64_to_cpu(tcon->fsUnixInfo.Capability); 3111 3112 if (ctx && ctx->no_linux_ext) { 3113 tcon->fsUnixInfo.Capability = 0; 3114 tcon->unix_ext = 0; /* Unix Extensions disabled */ 3115 cifs_dbg(FYI, "Linux protocol extensions disabled\n"); 3116 return; 3117 } else if (ctx) 3118 tcon->unix_ext = 1; /* Unix Extensions supported */ 3119 3120 if (!tcon->unix_ext) { 3121 cifs_dbg(FYI, "Unix extensions disabled so not set on reconnect\n"); 3122 return; 3123 } 3124 3125 if (!CIFSSMBQFSUnixInfo(xid, tcon)) { 3126 __u64 cap = le64_to_cpu(tcon->fsUnixInfo.Capability); 3127 cifs_dbg(FYI, "unix caps which server supports %lld\n", cap); 3128 /* 3129 * check for reconnect case in which we do not 3130 * want to change the mount behavior if we can avoid it 3131 */ 3132 if (ctx == NULL) { 3133 /* 3134 * turn off POSIX ACL and PATHNAMES if not set 3135 * originally at mount time 3136 */ 3137 if ((saved_cap & CIFS_UNIX_POSIX_ACL_CAP) == 0) 3138 cap &= ~CIFS_UNIX_POSIX_ACL_CAP; 3139 if ((saved_cap & CIFS_UNIX_POSIX_PATHNAMES_CAP) == 0) { 3140 if (cap & CIFS_UNIX_POSIX_PATHNAMES_CAP) 3141 cifs_dbg(VFS, "POSIXPATH support change\n"); 3142 cap &= ~CIFS_UNIX_POSIX_PATHNAMES_CAP; 3143 } else if ((cap & CIFS_UNIX_POSIX_PATHNAMES_CAP) == 0) { 3144 cifs_dbg(VFS, "possible reconnect error\n"); 3145 cifs_dbg(VFS, "server disabled POSIX path support\n"); 3146 } 3147 } 3148 3149 if (cap & CIFS_UNIX_TRANSPORT_ENCRYPTION_MANDATORY_CAP) 3150 cifs_dbg(VFS, "per-share encryption not supported yet\n"); 3151 3152 cap &= CIFS_UNIX_CAP_MASK; 3153 if (ctx && ctx->no_psx_acl) 3154 cap &= ~CIFS_UNIX_POSIX_ACL_CAP; 3155 else if (CIFS_UNIX_POSIX_ACL_CAP & cap) { 3156 cifs_dbg(FYI, "negotiated posix acl support\n"); 3157 if (cifs_sb) 3158 cifs_sb->mnt_cifs_flags |= 3159 CIFS_MOUNT_POSIXACL; 3160 } 3161 3162 if (ctx && ctx->posix_paths == 0) 3163 cap &= ~CIFS_UNIX_POSIX_PATHNAMES_CAP; 3164 else if (cap & CIFS_UNIX_POSIX_PATHNAMES_CAP) { 3165 cifs_dbg(FYI, "negotiate posix pathnames\n"); 3166 if (cifs_sb) 3167 cifs_sb->mnt_cifs_flags |= 3168 CIFS_MOUNT_POSIX_PATHS; 3169 } 3170 3171 cifs_dbg(FYI, "Negotiate caps 0x%x\n", (int)cap); 3172 #ifdef CONFIG_CIFS_DEBUG2 3173 if (cap & CIFS_UNIX_FCNTL_CAP) 3174 cifs_dbg(FYI, "FCNTL cap\n"); 3175 if (cap & CIFS_UNIX_EXTATTR_CAP) 3176 cifs_dbg(FYI, "EXTATTR cap\n"); 3177 if (cap & CIFS_UNIX_POSIX_PATHNAMES_CAP) 3178 cifs_dbg(FYI, "POSIX path cap\n"); 3179 if (cap & CIFS_UNIX_XATTR_CAP) 3180 cifs_dbg(FYI, "XATTR cap\n"); 3181 if (cap & CIFS_UNIX_POSIX_ACL_CAP) 3182 cifs_dbg(FYI, "POSIX ACL cap\n"); 3183 if (cap & CIFS_UNIX_LARGE_READ_CAP) 3184 cifs_dbg(FYI, "very large read cap\n"); 3185 if (cap & CIFS_UNIX_LARGE_WRITE_CAP) 3186 cifs_dbg(FYI, "very large write cap\n"); 3187 if (cap & CIFS_UNIX_TRANSPORT_ENCRYPTION_CAP) 3188 cifs_dbg(FYI, "transport encryption cap\n"); 3189 if (cap & CIFS_UNIX_TRANSPORT_ENCRYPTION_MANDATORY_CAP) 3190 cifs_dbg(FYI, "mandatory transport encryption cap\n"); 3191 #endif /* CIFS_DEBUG2 */ 3192 if (CIFSSMBSetFSUnixInfo(xid, tcon, cap)) { 3193 if (ctx == NULL) 3194 cifs_dbg(FYI, "resetting capabilities failed\n"); 3195 else 3196 cifs_dbg(VFS, "Negotiating Unix capabilities with the server failed. Consider mounting with the Unix Extensions disabled if problems are found by specifying the nounix mount option.\n"); 3197 3198 } 3199 } 3200 } 3201 #endif /* CONFIG_CIFS_ALLOW_INSECURE_LEGACY */ 3202 3203 int cifs_setup_cifs_sb(struct cifs_sb_info *cifs_sb) 3204 { 3205 struct smb3_fs_context *ctx = cifs_sb->ctx; 3206 3207 INIT_DELAYED_WORK(&cifs_sb->prune_tlinks, cifs_prune_tlinks); 3208 3209 spin_lock_init(&cifs_sb->tlink_tree_lock); 3210 cifs_sb->tlink_tree = RB_ROOT; 3211 3212 cifs_dbg(FYI, "file mode: %04ho dir mode: %04ho\n", 3213 ctx->file_mode, ctx->dir_mode); 3214 3215 /* this is needed for ASCII cp to Unicode converts */ 3216 if (ctx->iocharset == NULL) { 3217 /* load_nls_default cannot return null */ 3218 cifs_sb->local_nls = load_nls_default(); 3219 } else { 3220 cifs_sb->local_nls = load_nls(ctx->iocharset); 3221 if (cifs_sb->local_nls == NULL) { 3222 cifs_dbg(VFS, "CIFS mount error: iocharset %s not found\n", 3223 ctx->iocharset); 3224 return -ELIBACC; 3225 } 3226 } 3227 ctx->local_nls = cifs_sb->local_nls; 3228 3229 smb3_update_mnt_flags(cifs_sb); 3230 3231 if (ctx->direct_io) 3232 cifs_dbg(FYI, "mounting share using direct i/o\n"); 3233 if (ctx->cache_ro) { 3234 cifs_dbg(VFS, "mounting share with read only caching. Ensure that the share will not be modified while in use.\n"); 3235 cifs_sb->mnt_cifs_flags |= CIFS_MOUNT_RO_CACHE; 3236 } else if (ctx->cache_rw) { 3237 cifs_dbg(VFS, "mounting share in single client RW caching mode. Ensure that no other systems will be accessing the share.\n"); 3238 cifs_sb->mnt_cifs_flags |= (CIFS_MOUNT_RO_CACHE | 3239 CIFS_MOUNT_RW_CACHE); 3240 } 3241 3242 if ((ctx->cifs_acl) && (ctx->dynperm)) 3243 cifs_dbg(VFS, "mount option dynperm ignored if cifsacl mount option supported\n"); 3244 3245 if (ctx->prepath) { 3246 cifs_sb->prepath = kstrdup(ctx->prepath, GFP_KERNEL); 3247 if (cifs_sb->prepath == NULL) 3248 return -ENOMEM; 3249 cifs_sb->mnt_cifs_flags |= CIFS_MOUNT_USE_PREFIX_PATH; 3250 } 3251 3252 return 0; 3253 } 3254 3255 /* Release all succeed connections */ 3256 void cifs_mount_put_conns(struct cifs_mount_ctx *mnt_ctx) 3257 { 3258 int rc = 0; 3259 3260 if (mnt_ctx->tcon) 3261 cifs_put_tcon(mnt_ctx->tcon); 3262 else if (mnt_ctx->ses) 3263 cifs_put_smb_ses(mnt_ctx->ses); 3264 else if (mnt_ctx->server) 3265 cifs_put_tcp_session(mnt_ctx->server, 0); 3266 mnt_ctx->cifs_sb->mnt_cifs_flags &= ~CIFS_MOUNT_POSIX_PATHS; 3267 free_xid(mnt_ctx->xid); 3268 } 3269 3270 int cifs_mount_get_session(struct cifs_mount_ctx *mnt_ctx) 3271 { 3272 struct TCP_Server_Info *server = NULL; 3273 struct smb3_fs_context *ctx; 3274 struct cifs_ses *ses = NULL; 3275 unsigned int xid; 3276 int rc = 0; 3277 3278 xid = get_xid(); 3279 3280 if (WARN_ON_ONCE(!mnt_ctx || !mnt_ctx->fs_ctx)) { 3281 rc = -EINVAL; 3282 goto out; 3283 } 3284 ctx = mnt_ctx->fs_ctx; 3285 3286 /* get a reference to a tcp session */ 3287 server = cifs_get_tcp_session(ctx, NULL); 3288 if (IS_ERR(server)) { 3289 rc = PTR_ERR(server); 3290 server = NULL; 3291 goto out; 3292 } 3293 3294 /* get a reference to a SMB session */ 3295 ses = cifs_get_smb_ses(server, ctx); 3296 if (IS_ERR(ses)) { 3297 rc = PTR_ERR(ses); 3298 ses = NULL; 3299 goto out; 3300 } 3301 3302 if ((ctx->persistent == true) && (!(ses->server->capabilities & 3303 SMB2_GLOBAL_CAP_PERSISTENT_HANDLES))) { 3304 cifs_server_dbg(VFS, "persistent handles not supported by server\n"); 3305 rc = -EOPNOTSUPP; 3306 } 3307 3308 out: 3309 mnt_ctx->xid = xid; 3310 mnt_ctx->server = server; 3311 mnt_ctx->ses = ses; 3312 mnt_ctx->tcon = NULL; 3313 3314 return rc; 3315 } 3316 3317 int cifs_mount_get_tcon(struct cifs_mount_ctx *mnt_ctx) 3318 { 3319 struct TCP_Server_Info *server; 3320 struct cifs_sb_info *cifs_sb; 3321 struct smb3_fs_context *ctx; 3322 struct cifs_tcon *tcon = NULL; 3323 int rc = 0; 3324 3325 if (WARN_ON_ONCE(!mnt_ctx || !mnt_ctx->server || !mnt_ctx->ses || !mnt_ctx->fs_ctx || 3326 !mnt_ctx->cifs_sb)) { 3327 rc = -EINVAL; 3328 goto out; 3329 } 3330 server = mnt_ctx->server; 3331 ctx = mnt_ctx->fs_ctx; 3332 cifs_sb = mnt_ctx->cifs_sb; 3333 3334 /* search for existing tcon to this server share */ 3335 tcon = cifs_get_tcon(mnt_ctx->ses, ctx); 3336 if (IS_ERR(tcon)) { 3337 rc = PTR_ERR(tcon); 3338 tcon = NULL; 3339 goto out; 3340 } 3341 3342 /* if new SMB3.11 POSIX extensions are supported do not remap / and \ */ 3343 if (tcon->posix_extensions) 3344 cifs_sb->mnt_cifs_flags |= CIFS_MOUNT_POSIX_PATHS; 3345 3346 #ifdef CONFIG_CIFS_ALLOW_INSECURE_LEGACY 3347 /* tell server which Unix caps we support */ 3348 if (cap_unix(tcon->ses)) { 3349 /* 3350 * reset of caps checks mount to see if unix extensions disabled 3351 * for just this mount. 3352 */ 3353 reset_cifs_unix_caps(mnt_ctx->xid, tcon, cifs_sb, ctx); 3354 spin_lock(&tcon->ses->server->srv_lock); 3355 if ((tcon->ses->server->tcpStatus == CifsNeedReconnect) && 3356 (le64_to_cpu(tcon->fsUnixInfo.Capability) & 3357 CIFS_UNIX_TRANSPORT_ENCRYPTION_MANDATORY_CAP)) { 3358 spin_unlock(&tcon->ses->server->srv_lock); 3359 rc = -EACCES; 3360 goto out; 3361 } 3362 spin_unlock(&tcon->ses->server->srv_lock); 3363 } else 3364 #endif /* CONFIG_CIFS_ALLOW_INSECURE_LEGACY */ 3365 tcon->unix_ext = 0; /* server does not support them */ 3366 3367 /* do not care if a following call succeed - informational */ 3368 if (!tcon->pipe && server->ops->qfs_tcon) { 3369 server->ops->qfs_tcon(mnt_ctx->xid, tcon, cifs_sb); 3370 if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_RO_CACHE) { 3371 if (tcon->fsDevInfo.DeviceCharacteristics & 3372 cpu_to_le32(FILE_READ_ONLY_DEVICE)) 3373 cifs_dbg(VFS, "mounted to read only share\n"); 3374 else if ((cifs_sb->mnt_cifs_flags & 3375 CIFS_MOUNT_RW_CACHE) == 0) 3376 cifs_dbg(VFS, "read only mount of RW share\n"); 3377 /* no need to log a RW mount of a typical RW share */ 3378 } 3379 } 3380 3381 /* 3382 * Clamp the rsize/wsize mount arguments if they are too big for the server 3383 * and set the rsize/wsize to the negotiated values if not passed in by 3384 * the user on mount 3385 */ 3386 if ((cifs_sb->ctx->wsize == 0) || 3387 (cifs_sb->ctx->wsize > server->ops->negotiate_wsize(tcon, ctx))) 3388 cifs_sb->ctx->wsize = server->ops->negotiate_wsize(tcon, ctx); 3389 if ((cifs_sb->ctx->rsize == 0) || 3390 (cifs_sb->ctx->rsize > server->ops->negotiate_rsize(tcon, ctx))) 3391 cifs_sb->ctx->rsize = server->ops->negotiate_rsize(tcon, ctx); 3392 3393 /* 3394 * The cookie is initialized from volume info returned above. 3395 * Inside cifs_fscache_get_super_cookie it checks 3396 * that we do not get super cookie twice. 3397 */ 3398 if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_FSCACHE) 3399 cifs_fscache_get_super_cookie(tcon); 3400 3401 out: 3402 mnt_ctx->tcon = tcon; 3403 return rc; 3404 } 3405 3406 static int mount_setup_tlink(struct cifs_sb_info *cifs_sb, struct cifs_ses *ses, 3407 struct cifs_tcon *tcon) 3408 { 3409 struct tcon_link *tlink; 3410 3411 /* hang the tcon off of the superblock */ 3412 tlink = kzalloc(sizeof(*tlink), GFP_KERNEL); 3413 if (tlink == NULL) 3414 return -ENOMEM; 3415 3416 tlink->tl_uid = ses->linux_uid; 3417 tlink->tl_tcon = tcon; 3418 tlink->tl_time = jiffies; 3419 set_bit(TCON_LINK_MASTER, &tlink->tl_flags); 3420 set_bit(TCON_LINK_IN_TREE, &tlink->tl_flags); 3421 3422 cifs_sb->master_tlink = tlink; 3423 spin_lock(&cifs_sb->tlink_tree_lock); 3424 tlink_rb_insert(&cifs_sb->tlink_tree, tlink); 3425 spin_unlock(&cifs_sb->tlink_tree_lock); 3426 3427 queue_delayed_work(cifsiod_wq, &cifs_sb->prune_tlinks, 3428 TLINK_IDLE_EXPIRE); 3429 return 0; 3430 } 3431 3432 static int 3433 cifs_are_all_path_components_accessible(struct TCP_Server_Info *server, 3434 unsigned int xid, 3435 struct cifs_tcon *tcon, 3436 struct cifs_sb_info *cifs_sb, 3437 char *full_path, 3438 int added_treename) 3439 { 3440 int rc; 3441 char *s; 3442 char sep, tmp; 3443 int skip = added_treename ? 1 : 0; 3444 3445 sep = CIFS_DIR_SEP(cifs_sb); 3446 s = full_path; 3447 3448 rc = server->ops->is_path_accessible(xid, tcon, cifs_sb, ""); 3449 while (rc == 0) { 3450 /* skip separators */ 3451 while (*s == sep) 3452 s++; 3453 if (!*s) 3454 break; 3455 /* next separator */ 3456 while (*s && *s != sep) 3457 s++; 3458 /* 3459 * if the treename is added, we then have to skip the first 3460 * part within the separators 3461 */ 3462 if (skip) { 3463 skip = 0; 3464 continue; 3465 } 3466 /* 3467 * temporarily null-terminate the path at the end of 3468 * the current component 3469 */ 3470 tmp = *s; 3471 *s = 0; 3472 rc = server->ops->is_path_accessible(xid, tcon, cifs_sb, 3473 full_path); 3474 *s = tmp; 3475 } 3476 return rc; 3477 } 3478 3479 /* 3480 * Check if path is remote (i.e. a DFS share). 3481 * 3482 * Return -EREMOTE if it is, otherwise 0 or -errno. 3483 */ 3484 int cifs_is_path_remote(struct cifs_mount_ctx *mnt_ctx) 3485 { 3486 int rc; 3487 struct cifs_sb_info *cifs_sb = mnt_ctx->cifs_sb; 3488 struct TCP_Server_Info *server = mnt_ctx->server; 3489 unsigned int xid = mnt_ctx->xid; 3490 struct cifs_tcon *tcon = mnt_ctx->tcon; 3491 struct smb3_fs_context *ctx = mnt_ctx->fs_ctx; 3492 char *full_path; 3493 3494 if (!server->ops->is_path_accessible) 3495 return -EOPNOTSUPP; 3496 3497 /* 3498 * cifs_build_path_to_root works only when we have a valid tcon 3499 */ 3500 full_path = cifs_build_path_to_root(ctx, cifs_sb, tcon, 3501 tcon->Flags & SMB_SHARE_IS_IN_DFS); 3502 if (full_path == NULL) 3503 return -ENOMEM; 3504 3505 cifs_dbg(FYI, "%s: full_path: %s\n", __func__, full_path); 3506 3507 rc = server->ops->is_path_accessible(xid, tcon, cifs_sb, 3508 full_path); 3509 if (rc != 0 && rc != -EREMOTE) 3510 goto out; 3511 3512 if (rc != -EREMOTE) { 3513 rc = cifs_are_all_path_components_accessible(server, xid, tcon, 3514 cifs_sb, full_path, tcon->Flags & SMB_SHARE_IS_IN_DFS); 3515 if (rc != 0) { 3516 cifs_server_dbg(VFS, "cannot query dirs between root and final path, enabling CIFS_MOUNT_USE_PREFIX_PATH\n"); 3517 cifs_sb->mnt_cifs_flags |= CIFS_MOUNT_USE_PREFIX_PATH; 3518 rc = 0; 3519 } 3520 } 3521 3522 out: 3523 kfree(full_path); 3524 return rc; 3525 } 3526 3527 #ifdef CONFIG_CIFS_DFS_UPCALL 3528 int cifs_mount(struct cifs_sb_info *cifs_sb, struct smb3_fs_context *ctx) 3529 { 3530 struct cifs_mount_ctx mnt_ctx = { .cifs_sb = cifs_sb, .fs_ctx = ctx, }; 3531 bool isdfs; 3532 int rc; 3533 3534 INIT_LIST_HEAD(&mnt_ctx.dfs_ses_list); 3535 3536 rc = dfs_mount_share(&mnt_ctx, &isdfs); 3537 if (rc) 3538 goto error; 3539 if (!isdfs) 3540 goto out; 3541 3542 /* 3543 * After reconnecting to a different server, unique ids won't match anymore, so we disable 3544 * serverino. This prevents dentry revalidation to think the dentry are stale (ESTALE). 3545 */ 3546 cifs_autodisable_serverino(cifs_sb); 3547 /* 3548 * Force the use of prefix path to support failover on DFS paths that resolve to targets 3549 * that have different prefix paths. 3550 */ 3551 cifs_sb->mnt_cifs_flags |= CIFS_MOUNT_USE_PREFIX_PATH; 3552 kfree(cifs_sb->prepath); 3553 cifs_sb->prepath = ctx->prepath; 3554 ctx->prepath = NULL; 3555 3556 out: 3557 cifs_try_adding_channels(cifs_sb, mnt_ctx.ses); 3558 rc = mount_setup_tlink(cifs_sb, mnt_ctx.ses, mnt_ctx.tcon); 3559 if (rc) 3560 goto error; 3561 3562 free_xid(mnt_ctx.xid); 3563 return rc; 3564 3565 error: 3566 dfs_put_root_smb_sessions(&mnt_ctx.dfs_ses_list); 3567 cifs_mount_put_conns(&mnt_ctx); 3568 return rc; 3569 } 3570 #else 3571 int cifs_mount(struct cifs_sb_info *cifs_sb, struct smb3_fs_context *ctx) 3572 { 3573 int rc = 0; 3574 struct cifs_mount_ctx mnt_ctx = { .cifs_sb = cifs_sb, .fs_ctx = ctx, }; 3575 3576 rc = cifs_mount_get_session(&mnt_ctx); 3577 if (rc) 3578 goto error; 3579 3580 rc = cifs_mount_get_tcon(&mnt_ctx); 3581 if (rc) 3582 goto error; 3583 3584 rc = cifs_is_path_remote(&mnt_ctx); 3585 if (rc == -EREMOTE) 3586 rc = -EOPNOTSUPP; 3587 if (rc) 3588 goto error; 3589 3590 rc = mount_setup_tlink(cifs_sb, mnt_ctx.ses, mnt_ctx.tcon); 3591 if (rc) 3592 goto error; 3593 3594 free_xid(mnt_ctx.xid); 3595 return rc; 3596 3597 error: 3598 cifs_mount_put_conns(&mnt_ctx); 3599 return rc; 3600 } 3601 #endif 3602 3603 /* 3604 * Issue a TREE_CONNECT request. 3605 */ 3606 int 3607 CIFSTCon(const unsigned int xid, struct cifs_ses *ses, 3608 const char *tree, struct cifs_tcon *tcon, 3609 const struct nls_table *nls_codepage) 3610 { 3611 struct smb_hdr *smb_buffer; 3612 struct smb_hdr *smb_buffer_response; 3613 TCONX_REQ *pSMB; 3614 TCONX_RSP *pSMBr; 3615 unsigned char *bcc_ptr; 3616 int rc = 0; 3617 int length; 3618 __u16 bytes_left, count; 3619 3620 if (ses == NULL) 3621 return -EIO; 3622 3623 smb_buffer = cifs_buf_get(); 3624 if (smb_buffer == NULL) 3625 return -ENOMEM; 3626 3627 smb_buffer_response = smb_buffer; 3628 3629 header_assemble(smb_buffer, SMB_COM_TREE_CONNECT_ANDX, 3630 NULL /*no tid */ , 4 /*wct */ ); 3631 3632 smb_buffer->Mid = get_next_mid(ses->server); 3633 smb_buffer->Uid = ses->Suid; 3634 pSMB = (TCONX_REQ *) smb_buffer; 3635 pSMBr = (TCONX_RSP *) smb_buffer_response; 3636 3637 pSMB->AndXCommand = 0xFF; 3638 pSMB->Flags = cpu_to_le16(TCON_EXTENDED_SECINFO); 3639 bcc_ptr = &pSMB->Password[0]; 3640 3641 pSMB->PasswordLength = cpu_to_le16(1); /* minimum */ 3642 *bcc_ptr = 0; /* password is null byte */ 3643 bcc_ptr++; /* skip password */ 3644 /* already aligned so no need to do it below */ 3645 3646 if (ses->server->sign) 3647 smb_buffer->Flags2 |= SMBFLG2_SECURITY_SIGNATURE; 3648 3649 if (ses->capabilities & CAP_STATUS32) { 3650 smb_buffer->Flags2 |= SMBFLG2_ERR_STATUS; 3651 } 3652 if (ses->capabilities & CAP_DFS) { 3653 smb_buffer->Flags2 |= SMBFLG2_DFS; 3654 } 3655 if (ses->capabilities & CAP_UNICODE) { 3656 smb_buffer->Flags2 |= SMBFLG2_UNICODE; 3657 length = 3658 cifs_strtoUTF16((__le16 *) bcc_ptr, tree, 3659 6 /* max utf8 char length in bytes */ * 3660 (/* server len*/ + 256 /* share len */), nls_codepage); 3661 bcc_ptr += 2 * length; /* convert num 16 bit words to bytes */ 3662 bcc_ptr += 2; /* skip trailing null */ 3663 } else { /* ASCII */ 3664 strcpy(bcc_ptr, tree); 3665 bcc_ptr += strlen(tree) + 1; 3666 } 3667 strcpy(bcc_ptr, "?????"); 3668 bcc_ptr += strlen("?????"); 3669 bcc_ptr += 1; 3670 count = bcc_ptr - &pSMB->Password[0]; 3671 be32_add_cpu(&pSMB->hdr.smb_buf_length, count); 3672 pSMB->ByteCount = cpu_to_le16(count); 3673 3674 rc = SendReceive(xid, ses, smb_buffer, smb_buffer_response, &length, 3675 0); 3676 3677 /* above now done in SendReceive */ 3678 if (rc == 0) { 3679 bool is_unicode; 3680 3681 tcon->tid = smb_buffer_response->Tid; 3682 bcc_ptr = pByteArea(smb_buffer_response); 3683 bytes_left = get_bcc(smb_buffer_response); 3684 length = strnlen(bcc_ptr, bytes_left - 2); 3685 if (smb_buffer->Flags2 & SMBFLG2_UNICODE) 3686 is_unicode = true; 3687 else 3688 is_unicode = false; 3689 3690 3691 /* skip service field (NB: this field is always ASCII) */ 3692 if (length == 3) { 3693 if ((bcc_ptr[0] == 'I') && (bcc_ptr[1] == 'P') && 3694 (bcc_ptr[2] == 'C')) { 3695 cifs_dbg(FYI, "IPC connection\n"); 3696 tcon->ipc = true; 3697 tcon->pipe = true; 3698 } 3699 } else if (length == 2) { 3700 if ((bcc_ptr[0] == 'A') && (bcc_ptr[1] == ':')) { 3701 /* the most common case */ 3702 cifs_dbg(FYI, "disk share connection\n"); 3703 } 3704 } 3705 bcc_ptr += length + 1; 3706 bytes_left -= (length + 1); 3707 strscpy(tcon->tree_name, tree, sizeof(tcon->tree_name)); 3708 3709 /* mostly informational -- no need to fail on error here */ 3710 kfree(tcon->nativeFileSystem); 3711 tcon->nativeFileSystem = cifs_strndup_from_utf16(bcc_ptr, 3712 bytes_left, is_unicode, 3713 nls_codepage); 3714 3715 cifs_dbg(FYI, "nativeFileSystem=%s\n", tcon->nativeFileSystem); 3716 3717 if ((smb_buffer_response->WordCount == 3) || 3718 (smb_buffer_response->WordCount == 7)) 3719 /* field is in same location */ 3720 tcon->Flags = le16_to_cpu(pSMBr->OptionalSupport); 3721 else 3722 tcon->Flags = 0; 3723 cifs_dbg(FYI, "Tcon flags: 0x%x\n", tcon->Flags); 3724 } 3725 3726 cifs_buf_release(smb_buffer); 3727 return rc; 3728 } 3729 3730 static void delayed_free(struct rcu_head *p) 3731 { 3732 struct cifs_sb_info *cifs_sb = container_of(p, struct cifs_sb_info, rcu); 3733 3734 unload_nls(cifs_sb->local_nls); 3735 smb3_cleanup_fs_context(cifs_sb->ctx); 3736 kfree(cifs_sb); 3737 } 3738 3739 void 3740 cifs_umount(struct cifs_sb_info *cifs_sb) 3741 { 3742 struct rb_root *root = &cifs_sb->tlink_tree; 3743 struct rb_node *node; 3744 struct tcon_link *tlink; 3745 3746 cancel_delayed_work_sync(&cifs_sb->prune_tlinks); 3747 3748 spin_lock(&cifs_sb->tlink_tree_lock); 3749 while ((node = rb_first(root))) { 3750 tlink = rb_entry(node, struct tcon_link, tl_rbnode); 3751 cifs_get_tlink(tlink); 3752 clear_bit(TCON_LINK_IN_TREE, &tlink->tl_flags); 3753 rb_erase(node, root); 3754 3755 spin_unlock(&cifs_sb->tlink_tree_lock); 3756 cifs_put_tlink(tlink); 3757 spin_lock(&cifs_sb->tlink_tree_lock); 3758 } 3759 spin_unlock(&cifs_sb->tlink_tree_lock); 3760 3761 kfree(cifs_sb->prepath); 3762 call_rcu(&cifs_sb->rcu, delayed_free); 3763 } 3764 3765 int 3766 cifs_negotiate_protocol(const unsigned int xid, struct cifs_ses *ses, 3767 struct TCP_Server_Info *server) 3768 { 3769 int rc = 0; 3770 3771 if (!server->ops->need_neg || !server->ops->negotiate) 3772 return -ENOSYS; 3773 3774 /* only send once per connect */ 3775 spin_lock(&server->srv_lock); 3776 if (server->tcpStatus != CifsGood && 3777 server->tcpStatus != CifsNew && 3778 server->tcpStatus != CifsNeedNegotiate) { 3779 spin_unlock(&server->srv_lock); 3780 return -EHOSTDOWN; 3781 } 3782 3783 if (!server->ops->need_neg(server) && 3784 server->tcpStatus == CifsGood) { 3785 spin_unlock(&server->srv_lock); 3786 return 0; 3787 } 3788 3789 server->tcpStatus = CifsInNegotiate; 3790 spin_unlock(&server->srv_lock); 3791 3792 rc = server->ops->negotiate(xid, ses, server); 3793 if (rc == 0) { 3794 spin_lock(&server->srv_lock); 3795 if (server->tcpStatus == CifsInNegotiate) 3796 server->tcpStatus = CifsGood; 3797 else 3798 rc = -EHOSTDOWN; 3799 spin_unlock(&server->srv_lock); 3800 } else { 3801 spin_lock(&server->srv_lock); 3802 if (server->tcpStatus == CifsInNegotiate) 3803 server->tcpStatus = CifsNeedNegotiate; 3804 spin_unlock(&server->srv_lock); 3805 } 3806 3807 return rc; 3808 } 3809 3810 int 3811 cifs_setup_session(const unsigned int xid, struct cifs_ses *ses, 3812 struct TCP_Server_Info *server, 3813 struct nls_table *nls_info) 3814 { 3815 int rc = -ENOSYS; 3816 struct TCP_Server_Info *pserver = CIFS_SERVER_IS_CHAN(server) ? server->primary_server : server; 3817 struct sockaddr_in6 *addr6 = (struct sockaddr_in6 *)&pserver->dstaddr; 3818 struct sockaddr_in *addr = (struct sockaddr_in *)&pserver->dstaddr; 3819 bool is_binding = false; 3820 3821 spin_lock(&ses->ses_lock); 3822 cifs_dbg(FYI, "%s: channel connect bitmap: 0x%lx\n", 3823 __func__, ses->chans_need_reconnect); 3824 3825 if (ses->ses_status != SES_GOOD && 3826 ses->ses_status != SES_NEW && 3827 ses->ses_status != SES_NEED_RECON) { 3828 spin_unlock(&ses->ses_lock); 3829 return -EHOSTDOWN; 3830 } 3831 3832 /* only send once per connect */ 3833 spin_lock(&ses->chan_lock); 3834 if (CIFS_ALL_CHANS_GOOD(ses)) { 3835 if (ses->ses_status == SES_NEED_RECON) 3836 ses->ses_status = SES_GOOD; 3837 spin_unlock(&ses->chan_lock); 3838 spin_unlock(&ses->ses_lock); 3839 return 0; 3840 } 3841 3842 cifs_chan_set_in_reconnect(ses, server); 3843 is_binding = !CIFS_ALL_CHANS_NEED_RECONNECT(ses); 3844 spin_unlock(&ses->chan_lock); 3845 3846 if (!is_binding) 3847 ses->ses_status = SES_IN_SETUP; 3848 spin_unlock(&ses->ses_lock); 3849 3850 /* update ses ip_addr only for primary chan */ 3851 if (server == pserver) { 3852 if (server->dstaddr.ss_family == AF_INET6) 3853 scnprintf(ses->ip_addr, sizeof(ses->ip_addr), "%pI6", &addr6->sin6_addr); 3854 else 3855 scnprintf(ses->ip_addr, sizeof(ses->ip_addr), "%pI4", &addr->sin_addr); 3856 } 3857 3858 if (!is_binding) { 3859 ses->capabilities = server->capabilities; 3860 if (!linuxExtEnabled) 3861 ses->capabilities &= (~server->vals->cap_unix); 3862 3863 if (ses->auth_key.response) { 3864 cifs_dbg(FYI, "Free previous auth_key.response = %p\n", 3865 ses->auth_key.response); 3866 kfree_sensitive(ses->auth_key.response); 3867 ses->auth_key.response = NULL; 3868 ses->auth_key.len = 0; 3869 } 3870 } 3871 3872 cifs_dbg(FYI, "Security Mode: 0x%x Capabilities: 0x%x TimeAdjust: %d\n", 3873 server->sec_mode, server->capabilities, server->timeAdj); 3874 3875 if (server->ops->sess_setup) 3876 rc = server->ops->sess_setup(xid, ses, server, nls_info); 3877 3878 if (rc) { 3879 cifs_server_dbg(VFS, "Send error in SessSetup = %d\n", rc); 3880 spin_lock(&ses->ses_lock); 3881 if (ses->ses_status == SES_IN_SETUP) 3882 ses->ses_status = SES_NEED_RECON; 3883 spin_lock(&ses->chan_lock); 3884 cifs_chan_clear_in_reconnect(ses, server); 3885 spin_unlock(&ses->chan_lock); 3886 spin_unlock(&ses->ses_lock); 3887 } else { 3888 spin_lock(&ses->ses_lock); 3889 if (ses->ses_status == SES_IN_SETUP) 3890 ses->ses_status = SES_GOOD; 3891 spin_lock(&ses->chan_lock); 3892 cifs_chan_clear_in_reconnect(ses, server); 3893 cifs_chan_clear_need_reconnect(ses, server); 3894 spin_unlock(&ses->chan_lock); 3895 spin_unlock(&ses->ses_lock); 3896 } 3897 3898 return rc; 3899 } 3900 3901 static int 3902 cifs_set_vol_auth(struct smb3_fs_context *ctx, struct cifs_ses *ses) 3903 { 3904 ctx->sectype = ses->sectype; 3905 3906 /* krb5 is special, since we don't need username or pw */ 3907 if (ctx->sectype == Kerberos) 3908 return 0; 3909 3910 return cifs_set_cifscreds(ctx, ses); 3911 } 3912 3913 static struct cifs_tcon * 3914 cifs_construct_tcon(struct cifs_sb_info *cifs_sb, kuid_t fsuid) 3915 { 3916 int rc; 3917 struct cifs_tcon *master_tcon = cifs_sb_master_tcon(cifs_sb); 3918 struct cifs_ses *ses; 3919 struct cifs_tcon *tcon = NULL; 3920 struct smb3_fs_context *ctx; 3921 3922 ctx = kzalloc(sizeof(*ctx), GFP_KERNEL); 3923 if (ctx == NULL) 3924 return ERR_PTR(-ENOMEM); 3925 3926 ctx->local_nls = cifs_sb->local_nls; 3927 ctx->linux_uid = fsuid; 3928 ctx->cred_uid = fsuid; 3929 ctx->UNC = master_tcon->tree_name; 3930 ctx->retry = master_tcon->retry; 3931 ctx->nocase = master_tcon->nocase; 3932 ctx->nohandlecache = master_tcon->nohandlecache; 3933 ctx->local_lease = master_tcon->local_lease; 3934 ctx->no_lease = master_tcon->no_lease; 3935 ctx->resilient = master_tcon->use_resilient; 3936 ctx->persistent = master_tcon->use_persistent; 3937 ctx->handle_timeout = master_tcon->handle_timeout; 3938 ctx->no_linux_ext = !master_tcon->unix_ext; 3939 ctx->linux_ext = master_tcon->posix_extensions; 3940 ctx->sectype = master_tcon->ses->sectype; 3941 ctx->sign = master_tcon->ses->sign; 3942 ctx->seal = master_tcon->seal; 3943 ctx->witness = master_tcon->use_witness; 3944 3945 rc = cifs_set_vol_auth(ctx, master_tcon->ses); 3946 if (rc) { 3947 tcon = ERR_PTR(rc); 3948 goto out; 3949 } 3950 3951 /* get a reference for the same TCP session */ 3952 spin_lock(&cifs_tcp_ses_lock); 3953 ++master_tcon->ses->server->srv_count; 3954 spin_unlock(&cifs_tcp_ses_lock); 3955 3956 ses = cifs_get_smb_ses(master_tcon->ses->server, ctx); 3957 if (IS_ERR(ses)) { 3958 tcon = (struct cifs_tcon *)ses; 3959 cifs_put_tcp_session(master_tcon->ses->server, 0); 3960 goto out; 3961 } 3962 3963 tcon = cifs_get_tcon(ses, ctx); 3964 if (IS_ERR(tcon)) { 3965 cifs_put_smb_ses(ses); 3966 goto out; 3967 } 3968 3969 #ifdef CONFIG_CIFS_ALLOW_INSECURE_LEGACY 3970 if (cap_unix(ses)) 3971 reset_cifs_unix_caps(0, tcon, NULL, ctx); 3972 #endif /* CONFIG_CIFS_ALLOW_INSECURE_LEGACY */ 3973 3974 out: 3975 kfree(ctx->username); 3976 kfree_sensitive(ctx->password); 3977 kfree(ctx); 3978 3979 return tcon; 3980 } 3981 3982 struct cifs_tcon * 3983 cifs_sb_master_tcon(struct cifs_sb_info *cifs_sb) 3984 { 3985 return tlink_tcon(cifs_sb_master_tlink(cifs_sb)); 3986 } 3987 3988 /* find and return a tlink with given uid */ 3989 static struct tcon_link * 3990 tlink_rb_search(struct rb_root *root, kuid_t uid) 3991 { 3992 struct rb_node *node = root->rb_node; 3993 struct tcon_link *tlink; 3994 3995 while (node) { 3996 tlink = rb_entry(node, struct tcon_link, tl_rbnode); 3997 3998 if (uid_gt(tlink->tl_uid, uid)) 3999 node = node->rb_left; 4000 else if (uid_lt(tlink->tl_uid, uid)) 4001 node = node->rb_right; 4002 else 4003 return tlink; 4004 } 4005 return NULL; 4006 } 4007 4008 /* insert a tcon_link into the tree */ 4009 static void 4010 tlink_rb_insert(struct rb_root *root, struct tcon_link *new_tlink) 4011 { 4012 struct rb_node **new = &(root->rb_node), *parent = NULL; 4013 struct tcon_link *tlink; 4014 4015 while (*new) { 4016 tlink = rb_entry(*new, struct tcon_link, tl_rbnode); 4017 parent = *new; 4018 4019 if (uid_gt(tlink->tl_uid, new_tlink->tl_uid)) 4020 new = &((*new)->rb_left); 4021 else 4022 new = &((*new)->rb_right); 4023 } 4024 4025 rb_link_node(&new_tlink->tl_rbnode, parent, new); 4026 rb_insert_color(&new_tlink->tl_rbnode, root); 4027 } 4028 4029 /* 4030 * Find or construct an appropriate tcon given a cifs_sb and the fsuid of the 4031 * current task. 4032 * 4033 * If the superblock doesn't refer to a multiuser mount, then just return 4034 * the master tcon for the mount. 4035 * 4036 * First, search the rbtree for an existing tcon for this fsuid. If one 4037 * exists, then check to see if it's pending construction. If it is then wait 4038 * for construction to complete. Once it's no longer pending, check to see if 4039 * it failed and either return an error or retry construction, depending on 4040 * the timeout. 4041 * 4042 * If one doesn't exist then insert a new tcon_link struct into the tree and 4043 * try to construct a new one. 4044 */ 4045 struct tcon_link * 4046 cifs_sb_tlink(struct cifs_sb_info *cifs_sb) 4047 { 4048 int ret; 4049 kuid_t fsuid = current_fsuid(); 4050 struct tcon_link *tlink, *newtlink; 4051 4052 if (!(cifs_sb->mnt_cifs_flags & CIFS_MOUNT_MULTIUSER)) 4053 return cifs_get_tlink(cifs_sb_master_tlink(cifs_sb)); 4054 4055 spin_lock(&cifs_sb->tlink_tree_lock); 4056 tlink = tlink_rb_search(&cifs_sb->tlink_tree, fsuid); 4057 if (tlink) 4058 cifs_get_tlink(tlink); 4059 spin_unlock(&cifs_sb->tlink_tree_lock); 4060 4061 if (tlink == NULL) { 4062 newtlink = kzalloc(sizeof(*tlink), GFP_KERNEL); 4063 if (newtlink == NULL) 4064 return ERR_PTR(-ENOMEM); 4065 newtlink->tl_uid = fsuid; 4066 newtlink->tl_tcon = ERR_PTR(-EACCES); 4067 set_bit(TCON_LINK_PENDING, &newtlink->tl_flags); 4068 set_bit(TCON_LINK_IN_TREE, &newtlink->tl_flags); 4069 cifs_get_tlink(newtlink); 4070 4071 spin_lock(&cifs_sb->tlink_tree_lock); 4072 /* was one inserted after previous search? */ 4073 tlink = tlink_rb_search(&cifs_sb->tlink_tree, fsuid); 4074 if (tlink) { 4075 cifs_get_tlink(tlink); 4076 spin_unlock(&cifs_sb->tlink_tree_lock); 4077 kfree(newtlink); 4078 goto wait_for_construction; 4079 } 4080 tlink = newtlink; 4081 tlink_rb_insert(&cifs_sb->tlink_tree, tlink); 4082 spin_unlock(&cifs_sb->tlink_tree_lock); 4083 } else { 4084 wait_for_construction: 4085 ret = wait_on_bit(&tlink->tl_flags, TCON_LINK_PENDING, 4086 TASK_INTERRUPTIBLE); 4087 if (ret) { 4088 cifs_put_tlink(tlink); 4089 return ERR_PTR(-ERESTARTSYS); 4090 } 4091 4092 /* if it's good, return it */ 4093 if (!IS_ERR(tlink->tl_tcon)) 4094 return tlink; 4095 4096 /* return error if we tried this already recently */ 4097 if (time_before(jiffies, tlink->tl_time + TLINK_ERROR_EXPIRE)) { 4098 cifs_put_tlink(tlink); 4099 return ERR_PTR(-EACCES); 4100 } 4101 4102 if (test_and_set_bit(TCON_LINK_PENDING, &tlink->tl_flags)) 4103 goto wait_for_construction; 4104 } 4105 4106 tlink->tl_tcon = cifs_construct_tcon(cifs_sb, fsuid); 4107 clear_bit(TCON_LINK_PENDING, &tlink->tl_flags); 4108 wake_up_bit(&tlink->tl_flags, TCON_LINK_PENDING); 4109 4110 if (IS_ERR(tlink->tl_tcon)) { 4111 cifs_put_tlink(tlink); 4112 return ERR_PTR(-EACCES); 4113 } 4114 4115 return tlink; 4116 } 4117 4118 /* 4119 * periodic workqueue job that scans tcon_tree for a superblock and closes 4120 * out tcons. 4121 */ 4122 static void 4123 cifs_prune_tlinks(struct work_struct *work) 4124 { 4125 struct cifs_sb_info *cifs_sb = container_of(work, struct cifs_sb_info, 4126 prune_tlinks.work); 4127 struct rb_root *root = &cifs_sb->tlink_tree; 4128 struct rb_node *node; 4129 struct rb_node *tmp; 4130 struct tcon_link *tlink; 4131 4132 /* 4133 * Because we drop the spinlock in the loop in order to put the tlink 4134 * it's not guarded against removal of links from the tree. The only 4135 * places that remove entries from the tree are this function and 4136 * umounts. Because this function is non-reentrant and is canceled 4137 * before umount can proceed, this is safe. 4138 */ 4139 spin_lock(&cifs_sb->tlink_tree_lock); 4140 node = rb_first(root); 4141 while (node != NULL) { 4142 tmp = node; 4143 node = rb_next(tmp); 4144 tlink = rb_entry(tmp, struct tcon_link, tl_rbnode); 4145 4146 if (test_bit(TCON_LINK_MASTER, &tlink->tl_flags) || 4147 atomic_read(&tlink->tl_count) != 0 || 4148 time_after(tlink->tl_time + TLINK_IDLE_EXPIRE, jiffies)) 4149 continue; 4150 4151 cifs_get_tlink(tlink); 4152 clear_bit(TCON_LINK_IN_TREE, &tlink->tl_flags); 4153 rb_erase(tmp, root); 4154 4155 spin_unlock(&cifs_sb->tlink_tree_lock); 4156 cifs_put_tlink(tlink); 4157 spin_lock(&cifs_sb->tlink_tree_lock); 4158 } 4159 spin_unlock(&cifs_sb->tlink_tree_lock); 4160 4161 queue_delayed_work(cifsiod_wq, &cifs_sb->prune_tlinks, 4162 TLINK_IDLE_EXPIRE); 4163 } 4164 4165 #ifndef CONFIG_CIFS_DFS_UPCALL 4166 int cifs_tree_connect(const unsigned int xid, struct cifs_tcon *tcon, const struct nls_table *nlsc) 4167 { 4168 int rc; 4169 const struct smb_version_operations *ops = tcon->ses->server->ops; 4170 4171 /* only send once per connect */ 4172 spin_lock(&tcon->tc_lock); 4173 if (tcon->status == TID_GOOD) { 4174 spin_unlock(&tcon->tc_lock); 4175 return 0; 4176 } 4177 4178 if (tcon->status != TID_NEW && 4179 tcon->status != TID_NEED_TCON) { 4180 spin_unlock(&tcon->tc_lock); 4181 return -EHOSTDOWN; 4182 } 4183 4184 tcon->status = TID_IN_TCON; 4185 spin_unlock(&tcon->tc_lock); 4186 4187 rc = ops->tree_connect(xid, tcon->ses, tcon->tree_name, tcon, nlsc); 4188 if (rc) { 4189 spin_lock(&tcon->tc_lock); 4190 if (tcon->status == TID_IN_TCON) 4191 tcon->status = TID_NEED_TCON; 4192 spin_unlock(&tcon->tc_lock); 4193 } else { 4194 spin_lock(&tcon->tc_lock); 4195 if (tcon->status == TID_IN_TCON) 4196 tcon->status = TID_GOOD; 4197 tcon->need_reconnect = false; 4198 spin_unlock(&tcon->tc_lock); 4199 } 4200 4201 return rc; 4202 } 4203 #endif 4204