1 // SPDX-License-Identifier: GPL-2.0
2 /*
3 * SMB2 version specific operations
4 *
5 * Copyright (c) 2012, Jeff Layton <jlayton@redhat.com>
6 */
7
8 #include <linux/pagemap.h>
9 #include <linux/vfs.h>
10 #include <linux/falloc.h>
11 #include <linux/scatterlist.h>
12 #include <linux/uuid.h>
13 #include <linux/sort.h>
14 #include <crypto/aead.h>
15 #include <linux/fiemap.h>
16 #include <uapi/linux/magic.h>
17 #include "cifsfs.h"
18 #include "cifsglob.h"
19 #include "smb2pdu.h"
20 #include "smb2proto.h"
21 #include "cifsproto.h"
22 #include "cifs_debug.h"
23 #include "cifs_unicode.h"
24 #include "smb2status.h"
25 #include "smb2glob.h"
26 #include "cifs_ioctl.h"
27 #include "smbdirect.h"
28 #include "fscache.h"
29 #include "fs_context.h"
30 #include "cached_dir.h"
31 #include "reparse.h"
32
33 /* Change credits for different ops and return the total number of credits */
34 static int
change_conf(struct TCP_Server_Info * server)35 change_conf(struct TCP_Server_Info *server)
36 {
37 server->credits += server->echo_credits + server->oplock_credits;
38 if (server->credits > server->max_credits)
39 server->credits = server->max_credits;
40 server->oplock_credits = server->echo_credits = 0;
41 switch (server->credits) {
42 case 0:
43 return 0;
44 case 1:
45 server->echoes = false;
46 server->oplocks = false;
47 break;
48 case 2:
49 server->echoes = true;
50 server->oplocks = false;
51 server->echo_credits = 1;
52 break;
53 default:
54 server->echoes = true;
55 if (enable_oplocks) {
56 server->oplocks = true;
57 server->oplock_credits = 1;
58 } else
59 server->oplocks = false;
60
61 server->echo_credits = 1;
62 }
63 server->credits -= server->echo_credits + server->oplock_credits;
64 return server->credits + server->echo_credits + server->oplock_credits;
65 }
66
67 static void
smb2_add_credits(struct TCP_Server_Info * server,const struct cifs_credits * credits,const int optype)68 smb2_add_credits(struct TCP_Server_Info *server,
69 const struct cifs_credits *credits, const int optype)
70 {
71 int *val, rc = -1;
72 int scredits, in_flight;
73 unsigned int add = credits->value;
74 unsigned int instance = credits->instance;
75 bool reconnect_detected = false;
76 bool reconnect_with_invalid_credits = false;
77
78 spin_lock(&server->req_lock);
79 val = server->ops->get_credits_field(server, optype);
80
81 /* eg found case where write overlapping reconnect messed up credits */
82 if (((optype & CIFS_OP_MASK) == CIFS_NEG_OP) && (*val != 0))
83 reconnect_with_invalid_credits = true;
84
85 if ((instance == 0) || (instance == server->reconnect_instance))
86 *val += add;
87 else
88 reconnect_detected = true;
89
90 if (*val > 65000) {
91 *val = 65000; /* Don't get near 64K credits, avoid srv bugs */
92 pr_warn_once("server overflowed SMB3 credits\n");
93 trace_smb3_overflow_credits(server->CurrentMid,
94 server->conn_id, server->hostname, *val,
95 add, server->in_flight);
96 }
97 WARN_ON_ONCE(server->in_flight == 0);
98 server->in_flight--;
99 if (server->in_flight == 0 &&
100 ((optype & CIFS_OP_MASK) != CIFS_NEG_OP) &&
101 ((optype & CIFS_OP_MASK) != CIFS_SESS_OP))
102 rc = change_conf(server);
103 /*
104 * Sometimes server returns 0 credits on oplock break ack - we need to
105 * rebalance credits in this case.
106 */
107 else if (server->in_flight > 0 && server->oplock_credits == 0 &&
108 server->oplocks) {
109 if (server->credits > 1) {
110 server->credits--;
111 server->oplock_credits++;
112 }
113 } else if ((server->in_flight > 0) && (server->oplock_credits > 3) &&
114 ((optype & CIFS_OP_MASK) == CIFS_OBREAK_OP))
115 /* if now have too many oplock credits, rebalance so don't starve normal ops */
116 change_conf(server);
117
118 scredits = *val;
119 in_flight = server->in_flight;
120 spin_unlock(&server->req_lock);
121 wake_up(&server->request_q);
122
123 if (reconnect_detected) {
124 trace_smb3_reconnect_detected(server->CurrentMid,
125 server->conn_id, server->hostname, scredits, add, in_flight);
126
127 cifs_dbg(FYI, "trying to put %d credits from the old server instance %d\n",
128 add, instance);
129 }
130
131 if (reconnect_with_invalid_credits) {
132 trace_smb3_reconnect_with_invalid_credits(server->CurrentMid,
133 server->conn_id, server->hostname, scredits, add, in_flight);
134 cifs_dbg(FYI, "Negotiate operation when server credits is non-zero. Optype: %d, server credits: %d, credits added: %d\n",
135 optype, scredits, add);
136 }
137
138 spin_lock(&server->srv_lock);
139 if (server->tcpStatus == CifsNeedReconnect
140 || server->tcpStatus == CifsExiting) {
141 spin_unlock(&server->srv_lock);
142 return;
143 }
144 spin_unlock(&server->srv_lock);
145
146 switch (rc) {
147 case -1:
148 /* change_conf hasn't been executed */
149 break;
150 case 0:
151 cifs_server_dbg(VFS, "Possible client or server bug - zero credits\n");
152 break;
153 case 1:
154 cifs_server_dbg(VFS, "disabling echoes and oplocks\n");
155 break;
156 case 2:
157 cifs_dbg(FYI, "disabling oplocks\n");
158 break;
159 default:
160 /* change_conf rebalanced credits for different types */
161 break;
162 }
163
164 trace_smb3_add_credits(server->CurrentMid,
165 server->conn_id, server->hostname, scredits, add, in_flight);
166 cifs_dbg(FYI, "%s: added %u credits total=%d\n", __func__, add, scredits);
167 }
168
169 static void
smb2_set_credits(struct TCP_Server_Info * server,const int val)170 smb2_set_credits(struct TCP_Server_Info *server, const int val)
171 {
172 int scredits, in_flight;
173
174 spin_lock(&server->req_lock);
175 server->credits = val;
176 if (val == 1) {
177 server->reconnect_instance++;
178 /*
179 * ChannelSequence updated for all channels in primary channel so that consistent
180 * across SMB3 requests sent on any channel. See MS-SMB2 3.2.4.1 and 3.2.7.1
181 */
182 if (SERVER_IS_CHAN(server))
183 server->primary_server->channel_sequence_num++;
184 else
185 server->channel_sequence_num++;
186 }
187 scredits = server->credits;
188 in_flight = server->in_flight;
189 spin_unlock(&server->req_lock);
190
191 trace_smb3_set_credits(server->CurrentMid,
192 server->conn_id, server->hostname, scredits, val, in_flight);
193 cifs_dbg(FYI, "%s: set %u credits\n", __func__, val);
194
195 /* don't log while holding the lock */
196 if (val == 1)
197 cifs_dbg(FYI, "set credits to 1 due to smb2 reconnect\n");
198 }
199
200 static int *
smb2_get_credits_field(struct TCP_Server_Info * server,const int optype)201 smb2_get_credits_field(struct TCP_Server_Info *server, const int optype)
202 {
203 switch (optype) {
204 case CIFS_ECHO_OP:
205 return &server->echo_credits;
206 case CIFS_OBREAK_OP:
207 return &server->oplock_credits;
208 default:
209 return &server->credits;
210 }
211 }
212
213 static unsigned int
smb2_get_credits(struct mid_q_entry * mid)214 smb2_get_credits(struct mid_q_entry *mid)
215 {
216 return mid->credits_received;
217 }
218
219 static int
smb2_wait_mtu_credits(struct TCP_Server_Info * server,unsigned int size,unsigned int * num,struct cifs_credits * credits)220 smb2_wait_mtu_credits(struct TCP_Server_Info *server, unsigned int size,
221 unsigned int *num, struct cifs_credits *credits)
222 {
223 int rc = 0;
224 unsigned int scredits, in_flight;
225
226 spin_lock(&server->req_lock);
227 while (1) {
228 spin_unlock(&server->req_lock);
229
230 spin_lock(&server->srv_lock);
231 if (server->tcpStatus == CifsExiting) {
232 spin_unlock(&server->srv_lock);
233 return -ENOENT;
234 }
235 spin_unlock(&server->srv_lock);
236
237 spin_lock(&server->req_lock);
238 if (server->credits <= 0) {
239 spin_unlock(&server->req_lock);
240 cifs_num_waiters_inc(server);
241 rc = wait_event_killable(server->request_q,
242 has_credits(server, &server->credits, 1));
243 cifs_num_waiters_dec(server);
244 if (rc)
245 return rc;
246 spin_lock(&server->req_lock);
247 } else {
248 scredits = server->credits;
249 /* can deadlock with reopen */
250 if (scredits <= 8) {
251 *num = SMB2_MAX_BUFFER_SIZE;
252 credits->value = 0;
253 credits->instance = 0;
254 break;
255 }
256
257 /* leave some credits for reopen and other ops */
258 scredits -= 8;
259 *num = min_t(unsigned int, size,
260 scredits * SMB2_MAX_BUFFER_SIZE);
261
262 credits->value =
263 DIV_ROUND_UP(*num, SMB2_MAX_BUFFER_SIZE);
264 credits->instance = server->reconnect_instance;
265 server->credits -= credits->value;
266 server->in_flight++;
267 if (server->in_flight > server->max_in_flight)
268 server->max_in_flight = server->in_flight;
269 break;
270 }
271 }
272 scredits = server->credits;
273 in_flight = server->in_flight;
274 spin_unlock(&server->req_lock);
275
276 trace_smb3_wait_credits(server->CurrentMid,
277 server->conn_id, server->hostname, scredits, -(credits->value), in_flight);
278 cifs_dbg(FYI, "%s: removed %u credits total=%d\n",
279 __func__, credits->value, scredits);
280
281 return rc;
282 }
283
284 static int
smb2_adjust_credits(struct TCP_Server_Info * server,struct cifs_credits * credits,const unsigned int payload_size)285 smb2_adjust_credits(struct TCP_Server_Info *server,
286 struct cifs_credits *credits,
287 const unsigned int payload_size)
288 {
289 int new_val = DIV_ROUND_UP(payload_size, SMB2_MAX_BUFFER_SIZE);
290 int scredits, in_flight;
291
292 if (!credits->value || credits->value == new_val)
293 return 0;
294
295 if (credits->value < new_val) {
296 trace_smb3_too_many_credits(server->CurrentMid,
297 server->conn_id, server->hostname, 0, credits->value - new_val, 0);
298 cifs_server_dbg(VFS, "request has less credits (%d) than required (%d)",
299 credits->value, new_val);
300
301 return -EOPNOTSUPP;
302 }
303
304 spin_lock(&server->req_lock);
305
306 if (server->reconnect_instance != credits->instance) {
307 scredits = server->credits;
308 in_flight = server->in_flight;
309 spin_unlock(&server->req_lock);
310
311 trace_smb3_reconnect_detected(server->CurrentMid,
312 server->conn_id, server->hostname, scredits,
313 credits->value - new_val, in_flight);
314 cifs_server_dbg(VFS, "trying to return %d credits to old session\n",
315 credits->value - new_val);
316 return -EAGAIN;
317 }
318
319 server->credits += credits->value - new_val;
320 scredits = server->credits;
321 in_flight = server->in_flight;
322 spin_unlock(&server->req_lock);
323 wake_up(&server->request_q);
324
325 trace_smb3_adj_credits(server->CurrentMid,
326 server->conn_id, server->hostname, scredits,
327 credits->value - new_val, in_flight);
328 cifs_dbg(FYI, "%s: adjust added %u credits total=%d\n",
329 __func__, credits->value - new_val, scredits);
330
331 credits->value = new_val;
332
333 return 0;
334 }
335
336 static __u64
smb2_get_next_mid(struct TCP_Server_Info * server)337 smb2_get_next_mid(struct TCP_Server_Info *server)
338 {
339 __u64 mid;
340 /* for SMB2 we need the current value */
341 spin_lock(&server->mid_lock);
342 mid = server->CurrentMid++;
343 spin_unlock(&server->mid_lock);
344 return mid;
345 }
346
347 static void
smb2_revert_current_mid(struct TCP_Server_Info * server,const unsigned int val)348 smb2_revert_current_mid(struct TCP_Server_Info *server, const unsigned int val)
349 {
350 spin_lock(&server->mid_lock);
351 if (server->CurrentMid >= val)
352 server->CurrentMid -= val;
353 spin_unlock(&server->mid_lock);
354 }
355
356 static struct mid_q_entry *
__smb2_find_mid(struct TCP_Server_Info * server,char * buf,bool dequeue)357 __smb2_find_mid(struct TCP_Server_Info *server, char *buf, bool dequeue)
358 {
359 struct mid_q_entry *mid;
360 struct smb2_hdr *shdr = (struct smb2_hdr *)buf;
361 __u64 wire_mid = le64_to_cpu(shdr->MessageId);
362
363 if (shdr->ProtocolId == SMB2_TRANSFORM_PROTO_NUM) {
364 cifs_server_dbg(VFS, "Encrypted frame parsing not supported yet\n");
365 return NULL;
366 }
367
368 spin_lock(&server->mid_lock);
369 list_for_each_entry(mid, &server->pending_mid_q, qhead) {
370 if ((mid->mid == wire_mid) &&
371 (mid->mid_state == MID_REQUEST_SUBMITTED) &&
372 (mid->command == shdr->Command)) {
373 kref_get(&mid->refcount);
374 if (dequeue) {
375 list_del_init(&mid->qhead);
376 mid->mid_flags |= MID_DELETED;
377 }
378 spin_unlock(&server->mid_lock);
379 return mid;
380 }
381 }
382 spin_unlock(&server->mid_lock);
383 return NULL;
384 }
385
386 static struct mid_q_entry *
smb2_find_mid(struct TCP_Server_Info * server,char * buf)387 smb2_find_mid(struct TCP_Server_Info *server, char *buf)
388 {
389 return __smb2_find_mid(server, buf, false);
390 }
391
392 static struct mid_q_entry *
smb2_find_dequeue_mid(struct TCP_Server_Info * server,char * buf)393 smb2_find_dequeue_mid(struct TCP_Server_Info *server, char *buf)
394 {
395 return __smb2_find_mid(server, buf, true);
396 }
397
398 static void
smb2_dump_detail(void * buf,struct TCP_Server_Info * server)399 smb2_dump_detail(void *buf, struct TCP_Server_Info *server)
400 {
401 #ifdef CONFIG_CIFS_DEBUG2
402 struct smb2_hdr *shdr = (struct smb2_hdr *)buf;
403
404 cifs_server_dbg(VFS, "Cmd: %d Err: 0x%x Flags: 0x%x Mid: %llu Pid: %d\n",
405 shdr->Command, shdr->Status, shdr->Flags, shdr->MessageId,
406 shdr->Id.SyncId.ProcessId);
407 if (!server->ops->check_message(buf, server->total_read, server)) {
408 cifs_server_dbg(VFS, "smb buf %p len %u\n", buf,
409 server->ops->calc_smb_size(buf));
410 }
411 #endif
412 }
413
414 static bool
smb2_need_neg(struct TCP_Server_Info * server)415 smb2_need_neg(struct TCP_Server_Info *server)
416 {
417 return server->max_read == 0;
418 }
419
420 static int
smb2_negotiate(const unsigned int xid,struct cifs_ses * ses,struct TCP_Server_Info * server)421 smb2_negotiate(const unsigned int xid,
422 struct cifs_ses *ses,
423 struct TCP_Server_Info *server)
424 {
425 int rc;
426
427 spin_lock(&server->mid_lock);
428 server->CurrentMid = 0;
429 spin_unlock(&server->mid_lock);
430 rc = SMB2_negotiate(xid, ses, server);
431 /* BB we probably don't need to retry with modern servers */
432 if (rc == -EAGAIN)
433 rc = -EHOSTDOWN;
434 return rc;
435 }
436
437 static unsigned int
smb2_negotiate_wsize(struct cifs_tcon * tcon,struct smb3_fs_context * ctx)438 smb2_negotiate_wsize(struct cifs_tcon *tcon, struct smb3_fs_context *ctx)
439 {
440 struct TCP_Server_Info *server = tcon->ses->server;
441 unsigned int wsize;
442
443 /* start with specified wsize, or default */
444 wsize = ctx->wsize ? ctx->wsize : CIFS_DEFAULT_IOSIZE;
445 wsize = min_t(unsigned int, wsize, server->max_write);
446 if (!(server->capabilities & SMB2_GLOBAL_CAP_LARGE_MTU))
447 wsize = min_t(unsigned int, wsize, SMB2_MAX_BUFFER_SIZE);
448
449 return wsize;
450 }
451
452 static unsigned int
smb3_negotiate_wsize(struct cifs_tcon * tcon,struct smb3_fs_context * ctx)453 smb3_negotiate_wsize(struct cifs_tcon *tcon, struct smb3_fs_context *ctx)
454 {
455 struct TCP_Server_Info *server = tcon->ses->server;
456 unsigned int wsize;
457
458 /* start with specified wsize, or default */
459 wsize = ctx->wsize ? ctx->wsize : SMB3_DEFAULT_IOSIZE;
460 wsize = min_t(unsigned int, wsize, server->max_write);
461 #ifdef CONFIG_CIFS_SMB_DIRECT
462 if (server->rdma) {
463 if (server->sign)
464 /*
465 * Account for SMB2 data transfer packet header and
466 * possible encryption header
467 */
468 wsize = min_t(unsigned int,
469 wsize,
470 server->smbd_conn->max_fragmented_send_size -
471 SMB2_READWRITE_PDU_HEADER_SIZE -
472 sizeof(struct smb2_transform_hdr));
473 else
474 wsize = min_t(unsigned int,
475 wsize, server->smbd_conn->max_readwrite_size);
476 }
477 #endif
478 if (!(server->capabilities & SMB2_GLOBAL_CAP_LARGE_MTU))
479 wsize = min_t(unsigned int, wsize, SMB2_MAX_BUFFER_SIZE);
480
481 return wsize;
482 }
483
484 static unsigned int
smb2_negotiate_rsize(struct cifs_tcon * tcon,struct smb3_fs_context * ctx)485 smb2_negotiate_rsize(struct cifs_tcon *tcon, struct smb3_fs_context *ctx)
486 {
487 struct TCP_Server_Info *server = tcon->ses->server;
488 unsigned int rsize;
489
490 /* start with specified rsize, or default */
491 rsize = ctx->rsize ? ctx->rsize : CIFS_DEFAULT_IOSIZE;
492 rsize = min_t(unsigned int, rsize, server->max_read);
493
494 if (!(server->capabilities & SMB2_GLOBAL_CAP_LARGE_MTU))
495 rsize = min_t(unsigned int, rsize, SMB2_MAX_BUFFER_SIZE);
496
497 return rsize;
498 }
499
500 static unsigned int
smb3_negotiate_rsize(struct cifs_tcon * tcon,struct smb3_fs_context * ctx)501 smb3_negotiate_rsize(struct cifs_tcon *tcon, struct smb3_fs_context *ctx)
502 {
503 struct TCP_Server_Info *server = tcon->ses->server;
504 unsigned int rsize;
505
506 /* start with specified rsize, or default */
507 rsize = ctx->rsize ? ctx->rsize : SMB3_DEFAULT_IOSIZE;
508 rsize = min_t(unsigned int, rsize, server->max_read);
509 #ifdef CONFIG_CIFS_SMB_DIRECT
510 if (server->rdma) {
511 if (server->sign)
512 /*
513 * Account for SMB2 data transfer packet header and
514 * possible encryption header
515 */
516 rsize = min_t(unsigned int,
517 rsize,
518 server->smbd_conn->max_fragmented_recv_size -
519 SMB2_READWRITE_PDU_HEADER_SIZE -
520 sizeof(struct smb2_transform_hdr));
521 else
522 rsize = min_t(unsigned int,
523 rsize, server->smbd_conn->max_readwrite_size);
524 }
525 #endif
526
527 if (!(server->capabilities & SMB2_GLOBAL_CAP_LARGE_MTU))
528 rsize = min_t(unsigned int, rsize, SMB2_MAX_BUFFER_SIZE);
529
530 return rsize;
531 }
532
533 /*
534 * compare two interfaces a and b
535 * return 0 if everything matches.
536 * return 1 if a is rdma capable, or rss capable, or has higher link speed
537 * return -1 otherwise.
538 */
539 static int
iface_cmp(struct cifs_server_iface * a,struct cifs_server_iface * b)540 iface_cmp(struct cifs_server_iface *a, struct cifs_server_iface *b)
541 {
542 int cmp_ret = 0;
543
544 WARN_ON(!a || !b);
545 if (a->rdma_capable == b->rdma_capable) {
546 if (a->rss_capable == b->rss_capable) {
547 if (a->speed == b->speed) {
548 cmp_ret = cifs_ipaddr_cmp((struct sockaddr *) &a->sockaddr,
549 (struct sockaddr *) &b->sockaddr);
550 if (!cmp_ret)
551 return 0;
552 else if (cmp_ret > 0)
553 return 1;
554 else
555 return -1;
556 } else if (a->speed > b->speed)
557 return 1;
558 else
559 return -1;
560 } else if (a->rss_capable > b->rss_capable)
561 return 1;
562 else
563 return -1;
564 } else if (a->rdma_capable > b->rdma_capable)
565 return 1;
566 else
567 return -1;
568 }
569
570 static int
parse_server_interfaces(struct network_interface_info_ioctl_rsp * buf,size_t buf_len,struct cifs_ses * ses,bool in_mount)571 parse_server_interfaces(struct network_interface_info_ioctl_rsp *buf,
572 size_t buf_len, struct cifs_ses *ses, bool in_mount)
573 {
574 struct network_interface_info_ioctl_rsp *p;
575 struct sockaddr_in *addr4;
576 struct sockaddr_in6 *addr6;
577 struct iface_info_ipv4 *p4;
578 struct iface_info_ipv6 *p6;
579 struct cifs_server_iface *info = NULL, *iface = NULL, *niface = NULL;
580 struct cifs_server_iface tmp_iface;
581 ssize_t bytes_left;
582 size_t next = 0;
583 int nb_iface = 0;
584 int rc = 0, ret = 0;
585
586 bytes_left = buf_len;
587 p = buf;
588
589 spin_lock(&ses->iface_lock);
590 /* do not query too frequently, this time with lock held */
591 if (ses->iface_last_update &&
592 time_before(jiffies, ses->iface_last_update +
593 (SMB_INTERFACE_POLL_INTERVAL * HZ))) {
594 spin_unlock(&ses->iface_lock);
595 return 0;
596 }
597
598 /*
599 * Go through iface_list and mark them as inactive
600 */
601 list_for_each_entry_safe(iface, niface, &ses->iface_list,
602 iface_head)
603 iface->is_active = 0;
604
605 spin_unlock(&ses->iface_lock);
606
607 /*
608 * Samba server e.g. can return an empty interface list in some cases,
609 * which would only be a problem if we were requesting multichannel
610 */
611 if (bytes_left == 0) {
612 /* avoid spamming logs every 10 minutes, so log only in mount */
613 if ((ses->chan_max > 1) && in_mount)
614 cifs_dbg(VFS,
615 "multichannel not available\n"
616 "Empty network interface list returned by server %s\n",
617 ses->server->hostname);
618 rc = -EOPNOTSUPP;
619 ses->iface_last_update = jiffies;
620 goto out;
621 }
622
623 while (bytes_left >= (ssize_t)sizeof(*p)) {
624 memset(&tmp_iface, 0, sizeof(tmp_iface));
625 tmp_iface.speed = le64_to_cpu(p->LinkSpeed);
626 tmp_iface.rdma_capable = le32_to_cpu(p->Capability & RDMA_CAPABLE) ? 1 : 0;
627 tmp_iface.rss_capable = le32_to_cpu(p->Capability & RSS_CAPABLE) ? 1 : 0;
628
629 switch (p->Family) {
630 /*
631 * The kernel and wire socket structures have the same
632 * layout and use network byte order but make the
633 * conversion explicit in case either one changes.
634 */
635 case INTERNETWORK:
636 addr4 = (struct sockaddr_in *)&tmp_iface.sockaddr;
637 p4 = (struct iface_info_ipv4 *)p->Buffer;
638 addr4->sin_family = AF_INET;
639 memcpy(&addr4->sin_addr, &p4->IPv4Address, 4);
640
641 /* [MS-SMB2] 2.2.32.5.1.1 Clients MUST ignore these */
642 addr4->sin_port = cpu_to_be16(CIFS_PORT);
643
644 cifs_dbg(FYI, "%s: ipv4 %pI4\n", __func__,
645 &addr4->sin_addr);
646 break;
647 case INTERNETWORKV6:
648 addr6 = (struct sockaddr_in6 *)&tmp_iface.sockaddr;
649 p6 = (struct iface_info_ipv6 *)p->Buffer;
650 addr6->sin6_family = AF_INET6;
651 memcpy(&addr6->sin6_addr, &p6->IPv6Address, 16);
652
653 /* [MS-SMB2] 2.2.32.5.1.2 Clients MUST ignore these */
654 addr6->sin6_flowinfo = 0;
655 addr6->sin6_scope_id = 0;
656 addr6->sin6_port = cpu_to_be16(CIFS_PORT);
657
658 cifs_dbg(FYI, "%s: ipv6 %pI6\n", __func__,
659 &addr6->sin6_addr);
660 break;
661 default:
662 cifs_dbg(VFS,
663 "%s: skipping unsupported socket family\n",
664 __func__);
665 goto next_iface;
666 }
667
668 /*
669 * The iface_list is assumed to be sorted by speed.
670 * Check if the new interface exists in that list.
671 * NEVER change iface. it could be in use.
672 * Add a new one instead
673 */
674 spin_lock(&ses->iface_lock);
675 list_for_each_entry_safe(iface, niface, &ses->iface_list,
676 iface_head) {
677 ret = iface_cmp(iface, &tmp_iface);
678 if (!ret) {
679 iface->is_active = 1;
680 spin_unlock(&ses->iface_lock);
681 goto next_iface;
682 } else if (ret < 0) {
683 /* all remaining ifaces are slower */
684 kref_get(&iface->refcount);
685 break;
686 }
687 }
688 spin_unlock(&ses->iface_lock);
689
690 /* no match. insert the entry in the list */
691 info = kmalloc(sizeof(struct cifs_server_iface),
692 GFP_KERNEL);
693 if (!info) {
694 rc = -ENOMEM;
695 goto out;
696 }
697 memcpy(info, &tmp_iface, sizeof(tmp_iface));
698
699 /* add this new entry to the list */
700 kref_init(&info->refcount);
701 info->is_active = 1;
702
703 cifs_dbg(FYI, "%s: adding iface %zu\n", __func__, ses->iface_count);
704 cifs_dbg(FYI, "%s: speed %zu bps\n", __func__, info->speed);
705 cifs_dbg(FYI, "%s: capabilities 0x%08x\n", __func__,
706 le32_to_cpu(p->Capability));
707
708 spin_lock(&ses->iface_lock);
709 if (!list_entry_is_head(iface, &ses->iface_list, iface_head)) {
710 list_add_tail(&info->iface_head, &iface->iface_head);
711 kref_put(&iface->refcount, release_iface);
712 } else
713 list_add_tail(&info->iface_head, &ses->iface_list);
714
715 ses->iface_count++;
716 spin_unlock(&ses->iface_lock);
717 next_iface:
718 nb_iface++;
719 next = le32_to_cpu(p->Next);
720 if (!next) {
721 bytes_left -= sizeof(*p);
722 break;
723 }
724 p = (struct network_interface_info_ioctl_rsp *)((u8 *)p+next);
725 bytes_left -= next;
726 }
727
728 if (!nb_iface) {
729 cifs_dbg(VFS, "%s: malformed interface info\n", __func__);
730 rc = -EINVAL;
731 goto out;
732 }
733
734 /* Azure rounds the buffer size up 8, to a 16 byte boundary */
735 if ((bytes_left > 8) || p->Next)
736 cifs_dbg(VFS, "%s: incomplete interface info\n", __func__);
737
738 ses->iface_last_update = jiffies;
739
740 out:
741 /*
742 * Go through the list again and put the inactive entries
743 */
744 spin_lock(&ses->iface_lock);
745 list_for_each_entry_safe(iface, niface, &ses->iface_list,
746 iface_head) {
747 if (!iface->is_active) {
748 list_del(&iface->iface_head);
749 kref_put(&iface->refcount, release_iface);
750 ses->iface_count--;
751 }
752 }
753 spin_unlock(&ses->iface_lock);
754
755 return rc;
756 }
757
758 int
SMB3_request_interfaces(const unsigned int xid,struct cifs_tcon * tcon,bool in_mount)759 SMB3_request_interfaces(const unsigned int xid, struct cifs_tcon *tcon, bool in_mount)
760 {
761 int rc;
762 unsigned int ret_data_len = 0;
763 struct network_interface_info_ioctl_rsp *out_buf = NULL;
764 struct cifs_ses *ses = tcon->ses;
765 struct TCP_Server_Info *pserver;
766
767 /* do not query too frequently */
768 if (ses->iface_last_update &&
769 time_before(jiffies, ses->iface_last_update +
770 (SMB_INTERFACE_POLL_INTERVAL * HZ)))
771 return 0;
772
773 rc = SMB2_ioctl(xid, tcon, NO_FILE_ID, NO_FILE_ID,
774 FSCTL_QUERY_NETWORK_INTERFACE_INFO,
775 NULL /* no data input */, 0 /* no data input */,
776 CIFSMaxBufSize, (char **)&out_buf, &ret_data_len);
777 if (rc == -EOPNOTSUPP) {
778 cifs_dbg(FYI,
779 "server does not support query network interfaces\n");
780 ret_data_len = 0;
781 } else if (rc != 0) {
782 cifs_tcon_dbg(VFS, "error %d on ioctl to get interface list\n", rc);
783 goto out;
784 }
785
786 rc = parse_server_interfaces(out_buf, ret_data_len, ses, in_mount);
787 if (rc)
788 goto out;
789
790 /* check if iface is still active */
791 spin_lock(&ses->chan_lock);
792 pserver = ses->chans[0].server;
793 if (pserver && !cifs_chan_is_iface_active(ses, pserver)) {
794 spin_unlock(&ses->chan_lock);
795 cifs_chan_update_iface(ses, pserver);
796 spin_lock(&ses->chan_lock);
797 }
798 spin_unlock(&ses->chan_lock);
799
800 out:
801 kfree(out_buf);
802 return rc;
803 }
804
805 static void
smb3_qfs_tcon(const unsigned int xid,struct cifs_tcon * tcon,struct cifs_sb_info * cifs_sb)806 smb3_qfs_tcon(const unsigned int xid, struct cifs_tcon *tcon,
807 struct cifs_sb_info *cifs_sb)
808 {
809 int rc;
810 __le16 srch_path = 0; /* Null - open root of share */
811 u8 oplock = SMB2_OPLOCK_LEVEL_NONE;
812 struct cifs_open_parms oparms;
813 struct cifs_fid fid;
814 struct cached_fid *cfid = NULL;
815
816 oparms = (struct cifs_open_parms) {
817 .tcon = tcon,
818 .path = "",
819 .desired_access = FILE_READ_ATTRIBUTES,
820 .disposition = FILE_OPEN,
821 .create_options = cifs_create_options(cifs_sb, 0),
822 .fid = &fid,
823 };
824
825 rc = open_cached_dir(xid, tcon, "", cifs_sb, false, &cfid);
826 if (rc == 0)
827 memcpy(&fid, &cfid->fid, sizeof(struct cifs_fid));
828 else
829 rc = SMB2_open(xid, &oparms, &srch_path, &oplock, NULL, NULL,
830 NULL, NULL);
831 if (rc)
832 return;
833
834 SMB3_request_interfaces(xid, tcon, true /* called during mount */);
835
836 SMB2_QFS_attr(xid, tcon, fid.persistent_fid, fid.volatile_fid,
837 FS_ATTRIBUTE_INFORMATION);
838 SMB2_QFS_attr(xid, tcon, fid.persistent_fid, fid.volatile_fid,
839 FS_DEVICE_INFORMATION);
840 SMB2_QFS_attr(xid, tcon, fid.persistent_fid, fid.volatile_fid,
841 FS_VOLUME_INFORMATION);
842 SMB2_QFS_attr(xid, tcon, fid.persistent_fid, fid.volatile_fid,
843 FS_SECTOR_SIZE_INFORMATION); /* SMB3 specific */
844 if (cfid == NULL)
845 SMB2_close(xid, tcon, fid.persistent_fid, fid.volatile_fid);
846 else
847 close_cached_dir(cfid);
848 }
849
850 static void
smb2_qfs_tcon(const unsigned int xid,struct cifs_tcon * tcon,struct cifs_sb_info * cifs_sb)851 smb2_qfs_tcon(const unsigned int xid, struct cifs_tcon *tcon,
852 struct cifs_sb_info *cifs_sb)
853 {
854 int rc;
855 __le16 srch_path = 0; /* Null - open root of share */
856 u8 oplock = SMB2_OPLOCK_LEVEL_NONE;
857 struct cifs_open_parms oparms;
858 struct cifs_fid fid;
859
860 oparms = (struct cifs_open_parms) {
861 .tcon = tcon,
862 .path = "",
863 .desired_access = FILE_READ_ATTRIBUTES,
864 .disposition = FILE_OPEN,
865 .create_options = cifs_create_options(cifs_sb, 0),
866 .fid = &fid,
867 };
868
869 rc = SMB2_open(xid, &oparms, &srch_path, &oplock, NULL, NULL,
870 NULL, NULL);
871 if (rc)
872 return;
873
874 SMB2_QFS_attr(xid, tcon, fid.persistent_fid, fid.volatile_fid,
875 FS_ATTRIBUTE_INFORMATION);
876 SMB2_QFS_attr(xid, tcon, fid.persistent_fid, fid.volatile_fid,
877 FS_DEVICE_INFORMATION);
878 SMB2_close(xid, tcon, fid.persistent_fid, fid.volatile_fid);
879 }
880
881 static int
smb2_is_path_accessible(const unsigned int xid,struct cifs_tcon * tcon,struct cifs_sb_info * cifs_sb,const char * full_path)882 smb2_is_path_accessible(const unsigned int xid, struct cifs_tcon *tcon,
883 struct cifs_sb_info *cifs_sb, const char *full_path)
884 {
885 __le16 *utf16_path;
886 __u8 oplock = SMB2_OPLOCK_LEVEL_NONE;
887 int err_buftype = CIFS_NO_BUFFER;
888 struct cifs_open_parms oparms;
889 struct kvec err_iov = {};
890 struct cifs_fid fid;
891 struct cached_fid *cfid;
892 bool islink;
893 int rc, rc2;
894
895 rc = open_cached_dir(xid, tcon, full_path, cifs_sb, true, &cfid);
896 if (!rc) {
897 if (cfid->has_lease) {
898 close_cached_dir(cfid);
899 return 0;
900 }
901 close_cached_dir(cfid);
902 }
903
904 utf16_path = cifs_convert_path_to_utf16(full_path, cifs_sb);
905 if (!utf16_path)
906 return -ENOMEM;
907
908 oparms = (struct cifs_open_parms) {
909 .tcon = tcon,
910 .path = full_path,
911 .desired_access = FILE_READ_ATTRIBUTES,
912 .disposition = FILE_OPEN,
913 .create_options = cifs_create_options(cifs_sb, 0),
914 .fid = &fid,
915 };
916
917 rc = SMB2_open(xid, &oparms, utf16_path, &oplock, NULL, NULL,
918 &err_iov, &err_buftype);
919 if (rc) {
920 struct smb2_hdr *hdr = err_iov.iov_base;
921
922 if (unlikely(!hdr || err_buftype == CIFS_NO_BUFFER))
923 goto out;
924
925 if (rc != -EREMOTE && hdr->Status == STATUS_OBJECT_NAME_INVALID) {
926 rc2 = cifs_inval_name_dfs_link_error(xid, tcon, cifs_sb,
927 full_path, &islink);
928 if (rc2) {
929 rc = rc2;
930 goto out;
931 }
932 if (islink)
933 rc = -EREMOTE;
934 }
935 if (rc == -EREMOTE && IS_ENABLED(CONFIG_CIFS_DFS_UPCALL) && cifs_sb &&
936 (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_NO_DFS))
937 rc = -EOPNOTSUPP;
938 goto out;
939 }
940
941 rc = SMB2_close(xid, tcon, fid.persistent_fid, fid.volatile_fid);
942
943 out:
944 free_rsp_buf(err_buftype, err_iov.iov_base);
945 kfree(utf16_path);
946 return rc;
947 }
948
smb2_get_srv_inum(const unsigned int xid,struct cifs_tcon * tcon,struct cifs_sb_info * cifs_sb,const char * full_path,u64 * uniqueid,struct cifs_open_info_data * data)949 static int smb2_get_srv_inum(const unsigned int xid, struct cifs_tcon *tcon,
950 struct cifs_sb_info *cifs_sb, const char *full_path,
951 u64 *uniqueid, struct cifs_open_info_data *data)
952 {
953 *uniqueid = le64_to_cpu(data->fi.IndexNumber);
954 return 0;
955 }
956
smb2_query_file_info(const unsigned int xid,struct cifs_tcon * tcon,struct cifsFileInfo * cfile,struct cifs_open_info_data * data)957 static int smb2_query_file_info(const unsigned int xid, struct cifs_tcon *tcon,
958 struct cifsFileInfo *cfile, struct cifs_open_info_data *data)
959 {
960 struct cifs_fid *fid = &cfile->fid;
961
962 if (cfile->symlink_target) {
963 data->symlink_target = kstrdup(cfile->symlink_target, GFP_KERNEL);
964 if (!data->symlink_target)
965 return -ENOMEM;
966 }
967 return SMB2_query_info(xid, tcon, fid->persistent_fid, fid->volatile_fid, &data->fi);
968 }
969
970 #ifdef CONFIG_CIFS_XATTR
971 static ssize_t
move_smb2_ea_to_cifs(char * dst,size_t dst_size,struct smb2_file_full_ea_info * src,size_t src_size,const unsigned char * ea_name)972 move_smb2_ea_to_cifs(char *dst, size_t dst_size,
973 struct smb2_file_full_ea_info *src, size_t src_size,
974 const unsigned char *ea_name)
975 {
976 int rc = 0;
977 unsigned int ea_name_len = ea_name ? strlen(ea_name) : 0;
978 char *name, *value;
979 size_t buf_size = dst_size;
980 size_t name_len, value_len, user_name_len;
981
982 while (src_size > 0) {
983 name_len = (size_t)src->ea_name_length;
984 value_len = (size_t)le16_to_cpu(src->ea_value_length);
985
986 if (name_len == 0)
987 break;
988
989 if (src_size < 8 + name_len + 1 + value_len) {
990 cifs_dbg(FYI, "EA entry goes beyond length of list\n");
991 rc = -EIO;
992 goto out;
993 }
994
995 name = &src->ea_data[0];
996 value = &src->ea_data[src->ea_name_length + 1];
997
998 if (ea_name) {
999 if (ea_name_len == name_len &&
1000 memcmp(ea_name, name, name_len) == 0) {
1001 rc = value_len;
1002 if (dst_size == 0)
1003 goto out;
1004 if (dst_size < value_len) {
1005 rc = -ERANGE;
1006 goto out;
1007 }
1008 memcpy(dst, value, value_len);
1009 goto out;
1010 }
1011 } else {
1012 /* 'user.' plus a terminating null */
1013 user_name_len = 5 + 1 + name_len;
1014
1015 if (buf_size == 0) {
1016 /* skip copy - calc size only */
1017 rc += user_name_len;
1018 } else if (dst_size >= user_name_len) {
1019 dst_size -= user_name_len;
1020 memcpy(dst, "user.", 5);
1021 dst += 5;
1022 memcpy(dst, src->ea_data, name_len);
1023 dst += name_len;
1024 *dst = 0;
1025 ++dst;
1026 rc += user_name_len;
1027 } else {
1028 /* stop before overrun buffer */
1029 rc = -ERANGE;
1030 break;
1031 }
1032 }
1033
1034 if (!src->next_entry_offset)
1035 break;
1036
1037 if (src_size < le32_to_cpu(src->next_entry_offset)) {
1038 /* stop before overrun buffer */
1039 rc = -ERANGE;
1040 break;
1041 }
1042 src_size -= le32_to_cpu(src->next_entry_offset);
1043 src = (void *)((char *)src +
1044 le32_to_cpu(src->next_entry_offset));
1045 }
1046
1047 /* didn't find the named attribute */
1048 if (ea_name)
1049 rc = -ENODATA;
1050
1051 out:
1052 return (ssize_t)rc;
1053 }
1054
1055 static ssize_t
smb2_query_eas(const unsigned int xid,struct cifs_tcon * tcon,const unsigned char * path,const unsigned char * ea_name,char * ea_data,size_t buf_size,struct cifs_sb_info * cifs_sb)1056 smb2_query_eas(const unsigned int xid, struct cifs_tcon *tcon,
1057 const unsigned char *path, const unsigned char *ea_name,
1058 char *ea_data, size_t buf_size,
1059 struct cifs_sb_info *cifs_sb)
1060 {
1061 int rc;
1062 struct kvec rsp_iov = {NULL, 0};
1063 int buftype = CIFS_NO_BUFFER;
1064 struct smb2_query_info_rsp *rsp;
1065 struct smb2_file_full_ea_info *info = NULL;
1066
1067 rc = smb2_query_info_compound(xid, tcon, path,
1068 FILE_READ_EA,
1069 FILE_FULL_EA_INFORMATION,
1070 SMB2_O_INFO_FILE,
1071 CIFSMaxBufSize -
1072 MAX_SMB2_CREATE_RESPONSE_SIZE -
1073 MAX_SMB2_CLOSE_RESPONSE_SIZE,
1074 &rsp_iov, &buftype, cifs_sb);
1075 if (rc) {
1076 /*
1077 * If ea_name is NULL (listxattr) and there are no EAs,
1078 * return 0 as it's not an error. Otherwise, the specified
1079 * ea_name was not found.
1080 */
1081 if (!ea_name && rc == -ENODATA)
1082 rc = 0;
1083 goto qeas_exit;
1084 }
1085
1086 rsp = (struct smb2_query_info_rsp *)rsp_iov.iov_base;
1087 rc = smb2_validate_iov(le16_to_cpu(rsp->OutputBufferOffset),
1088 le32_to_cpu(rsp->OutputBufferLength),
1089 &rsp_iov,
1090 sizeof(struct smb2_file_full_ea_info));
1091 if (rc)
1092 goto qeas_exit;
1093
1094 info = (struct smb2_file_full_ea_info *)(
1095 le16_to_cpu(rsp->OutputBufferOffset) + (char *)rsp);
1096 rc = move_smb2_ea_to_cifs(ea_data, buf_size, info,
1097 le32_to_cpu(rsp->OutputBufferLength), ea_name);
1098
1099 qeas_exit:
1100 free_rsp_buf(buftype, rsp_iov.iov_base);
1101 return rc;
1102 }
1103
1104 static int
smb2_set_ea(const unsigned int xid,struct cifs_tcon * tcon,const char * path,const char * ea_name,const void * ea_value,const __u16 ea_value_len,const struct nls_table * nls_codepage,struct cifs_sb_info * cifs_sb)1105 smb2_set_ea(const unsigned int xid, struct cifs_tcon *tcon,
1106 const char *path, const char *ea_name, const void *ea_value,
1107 const __u16 ea_value_len, const struct nls_table *nls_codepage,
1108 struct cifs_sb_info *cifs_sb)
1109 {
1110 struct smb2_compound_vars *vars;
1111 struct cifs_ses *ses = tcon->ses;
1112 struct TCP_Server_Info *server;
1113 struct smb_rqst *rqst;
1114 struct kvec *rsp_iov;
1115 __le16 *utf16_path = NULL;
1116 int ea_name_len = strlen(ea_name);
1117 int flags = CIFS_CP_CREATE_CLOSE_OP;
1118 int len;
1119 int resp_buftype[3];
1120 struct cifs_open_parms oparms;
1121 __u8 oplock = SMB2_OPLOCK_LEVEL_NONE;
1122 struct cifs_fid fid;
1123 unsigned int size[1];
1124 void *data[1];
1125 struct smb2_file_full_ea_info *ea = NULL;
1126 struct smb2_query_info_rsp *rsp;
1127 int rc, used_len = 0;
1128 int retries = 0, cur_sleep = 1;
1129
1130 replay_again:
1131 /* reinitialize for possible replay */
1132 flags = CIFS_CP_CREATE_CLOSE_OP;
1133 oplock = SMB2_OPLOCK_LEVEL_NONE;
1134 server = cifs_pick_channel(ses);
1135
1136 if (smb3_encryption_required(tcon))
1137 flags |= CIFS_TRANSFORM_REQ;
1138
1139 if (ea_name_len > 255)
1140 return -EINVAL;
1141
1142 utf16_path = cifs_convert_path_to_utf16(path, cifs_sb);
1143 if (!utf16_path)
1144 return -ENOMEM;
1145
1146 resp_buftype[0] = resp_buftype[1] = resp_buftype[2] = CIFS_NO_BUFFER;
1147 vars = kzalloc(sizeof(*vars), GFP_KERNEL);
1148 if (!vars) {
1149 rc = -ENOMEM;
1150 goto out_free_path;
1151 }
1152 rqst = vars->rqst;
1153 rsp_iov = vars->rsp_iov;
1154
1155 if (ses->server->ops->query_all_EAs) {
1156 if (!ea_value) {
1157 rc = ses->server->ops->query_all_EAs(xid, tcon, path,
1158 ea_name, NULL, 0,
1159 cifs_sb);
1160 if (rc == -ENODATA)
1161 goto sea_exit;
1162 } else {
1163 /* If we are adding a attribute we should first check
1164 * if there will be enough space available to store
1165 * the new EA. If not we should not add it since we
1166 * would not be able to even read the EAs back.
1167 */
1168 rc = smb2_query_info_compound(xid, tcon, path,
1169 FILE_READ_EA,
1170 FILE_FULL_EA_INFORMATION,
1171 SMB2_O_INFO_FILE,
1172 CIFSMaxBufSize -
1173 MAX_SMB2_CREATE_RESPONSE_SIZE -
1174 MAX_SMB2_CLOSE_RESPONSE_SIZE,
1175 &rsp_iov[1], &resp_buftype[1], cifs_sb);
1176 if (rc == 0) {
1177 rsp = (struct smb2_query_info_rsp *)rsp_iov[1].iov_base;
1178 used_len = le32_to_cpu(rsp->OutputBufferLength);
1179 }
1180 free_rsp_buf(resp_buftype[1], rsp_iov[1].iov_base);
1181 resp_buftype[1] = CIFS_NO_BUFFER;
1182 memset(&rsp_iov[1], 0, sizeof(rsp_iov[1]));
1183 rc = 0;
1184
1185 /* Use a fudge factor of 256 bytes in case we collide
1186 * with a different set_EAs command.
1187 */
1188 if (CIFSMaxBufSize - MAX_SMB2_CREATE_RESPONSE_SIZE -
1189 MAX_SMB2_CLOSE_RESPONSE_SIZE - 256 <
1190 used_len + ea_name_len + ea_value_len + 1) {
1191 rc = -ENOSPC;
1192 goto sea_exit;
1193 }
1194 }
1195 }
1196
1197 /* Open */
1198 rqst[0].rq_iov = vars->open_iov;
1199 rqst[0].rq_nvec = SMB2_CREATE_IOV_SIZE;
1200
1201 oparms = (struct cifs_open_parms) {
1202 .tcon = tcon,
1203 .path = path,
1204 .desired_access = FILE_WRITE_EA,
1205 .disposition = FILE_OPEN,
1206 .create_options = cifs_create_options(cifs_sb, 0),
1207 .fid = &fid,
1208 .replay = !!(retries),
1209 };
1210
1211 rc = SMB2_open_init(tcon, server,
1212 &rqst[0], &oplock, &oparms, utf16_path);
1213 if (rc)
1214 goto sea_exit;
1215 smb2_set_next_command(tcon, &rqst[0]);
1216
1217
1218 /* Set Info */
1219 rqst[1].rq_iov = vars->si_iov;
1220 rqst[1].rq_nvec = 1;
1221
1222 len = sizeof(*ea) + ea_name_len + ea_value_len + 1;
1223 ea = kzalloc(len, GFP_KERNEL);
1224 if (ea == NULL) {
1225 rc = -ENOMEM;
1226 goto sea_exit;
1227 }
1228
1229 ea->ea_name_length = ea_name_len;
1230 ea->ea_value_length = cpu_to_le16(ea_value_len);
1231 memcpy(ea->ea_data, ea_name, ea_name_len + 1);
1232 memcpy(ea->ea_data + ea_name_len + 1, ea_value, ea_value_len);
1233
1234 size[0] = len;
1235 data[0] = ea;
1236
1237 rc = SMB2_set_info_init(tcon, server,
1238 &rqst[1], COMPOUND_FID,
1239 COMPOUND_FID, current->tgid,
1240 FILE_FULL_EA_INFORMATION,
1241 SMB2_O_INFO_FILE, 0, data, size);
1242 if (rc)
1243 goto sea_exit;
1244 smb2_set_next_command(tcon, &rqst[1]);
1245 smb2_set_related(&rqst[1]);
1246
1247 /* Close */
1248 rqst[2].rq_iov = &vars->close_iov;
1249 rqst[2].rq_nvec = 1;
1250 rc = SMB2_close_init(tcon, server,
1251 &rqst[2], COMPOUND_FID, COMPOUND_FID, false);
1252 if (rc)
1253 goto sea_exit;
1254 smb2_set_related(&rqst[2]);
1255
1256 if (retries) {
1257 smb2_set_replay(server, &rqst[0]);
1258 smb2_set_replay(server, &rqst[1]);
1259 smb2_set_replay(server, &rqst[2]);
1260 }
1261
1262 rc = compound_send_recv(xid, ses, server,
1263 flags, 3, rqst,
1264 resp_buftype, rsp_iov);
1265 /* no need to bump num_remote_opens because handle immediately closed */
1266
1267 sea_exit:
1268 kfree(ea);
1269 SMB2_open_free(&rqst[0]);
1270 SMB2_set_info_free(&rqst[1]);
1271 SMB2_close_free(&rqst[2]);
1272 free_rsp_buf(resp_buftype[0], rsp_iov[0].iov_base);
1273 free_rsp_buf(resp_buftype[1], rsp_iov[1].iov_base);
1274 free_rsp_buf(resp_buftype[2], rsp_iov[2].iov_base);
1275 kfree(vars);
1276 out_free_path:
1277 kfree(utf16_path);
1278
1279 if (is_replayable_error(rc) &&
1280 smb2_should_replay(tcon, &retries, &cur_sleep))
1281 goto replay_again;
1282
1283 return rc;
1284 }
1285 #endif
1286
1287 static bool
smb2_can_echo(struct TCP_Server_Info * server)1288 smb2_can_echo(struct TCP_Server_Info *server)
1289 {
1290 return server->echoes;
1291 }
1292
1293 static void
smb2_clear_stats(struct cifs_tcon * tcon)1294 smb2_clear_stats(struct cifs_tcon *tcon)
1295 {
1296 int i;
1297
1298 for (i = 0; i < NUMBER_OF_SMB2_COMMANDS; i++) {
1299 atomic_set(&tcon->stats.smb2_stats.smb2_com_sent[i], 0);
1300 atomic_set(&tcon->stats.smb2_stats.smb2_com_failed[i], 0);
1301 }
1302 }
1303
1304 static void
smb2_dump_share_caps(struct seq_file * m,struct cifs_tcon * tcon)1305 smb2_dump_share_caps(struct seq_file *m, struct cifs_tcon *tcon)
1306 {
1307 seq_puts(m, "\n\tShare Capabilities:");
1308 if (tcon->capabilities & SMB2_SHARE_CAP_DFS)
1309 seq_puts(m, " DFS,");
1310 if (tcon->capabilities & SMB2_SHARE_CAP_CONTINUOUS_AVAILABILITY)
1311 seq_puts(m, " CONTINUOUS AVAILABILITY,");
1312 if (tcon->capabilities & SMB2_SHARE_CAP_SCALEOUT)
1313 seq_puts(m, " SCALEOUT,");
1314 if (tcon->capabilities & SMB2_SHARE_CAP_CLUSTER)
1315 seq_puts(m, " CLUSTER,");
1316 if (tcon->capabilities & SMB2_SHARE_CAP_ASYMMETRIC)
1317 seq_puts(m, " ASYMMETRIC,");
1318 if (tcon->capabilities == 0)
1319 seq_puts(m, " None");
1320 if (tcon->ss_flags & SSINFO_FLAGS_ALIGNED_DEVICE)
1321 seq_puts(m, " Aligned,");
1322 if (tcon->ss_flags & SSINFO_FLAGS_PARTITION_ALIGNED_ON_DEVICE)
1323 seq_puts(m, " Partition Aligned,");
1324 if (tcon->ss_flags & SSINFO_FLAGS_NO_SEEK_PENALTY)
1325 seq_puts(m, " SSD,");
1326 if (tcon->ss_flags & SSINFO_FLAGS_TRIM_ENABLED)
1327 seq_puts(m, " TRIM-support,");
1328
1329 seq_printf(m, "\tShare Flags: 0x%x", tcon->share_flags);
1330 seq_printf(m, "\n\ttid: 0x%x", tcon->tid);
1331 if (tcon->perf_sector_size)
1332 seq_printf(m, "\tOptimal sector size: 0x%x",
1333 tcon->perf_sector_size);
1334 seq_printf(m, "\tMaximal Access: 0x%x", tcon->maximal_access);
1335 }
1336
1337 static void
smb2_print_stats(struct seq_file * m,struct cifs_tcon * tcon)1338 smb2_print_stats(struct seq_file *m, struct cifs_tcon *tcon)
1339 {
1340 atomic_t *sent = tcon->stats.smb2_stats.smb2_com_sent;
1341 atomic_t *failed = tcon->stats.smb2_stats.smb2_com_failed;
1342
1343 /*
1344 * Can't display SMB2_NEGOTIATE, SESSION_SETUP, LOGOFF, CANCEL and ECHO
1345 * totals (requests sent) since those SMBs are per-session not per tcon
1346 */
1347 seq_printf(m, "\nBytes read: %llu Bytes written: %llu",
1348 (long long)(tcon->bytes_read),
1349 (long long)(tcon->bytes_written));
1350 seq_printf(m, "\nOpen files: %d total (local), %d open on server",
1351 atomic_read(&tcon->num_local_opens),
1352 atomic_read(&tcon->num_remote_opens));
1353 seq_printf(m, "\nTreeConnects: %d total %d failed",
1354 atomic_read(&sent[SMB2_TREE_CONNECT_HE]),
1355 atomic_read(&failed[SMB2_TREE_CONNECT_HE]));
1356 seq_printf(m, "\nTreeDisconnects: %d total %d failed",
1357 atomic_read(&sent[SMB2_TREE_DISCONNECT_HE]),
1358 atomic_read(&failed[SMB2_TREE_DISCONNECT_HE]));
1359 seq_printf(m, "\nCreates: %d total %d failed",
1360 atomic_read(&sent[SMB2_CREATE_HE]),
1361 atomic_read(&failed[SMB2_CREATE_HE]));
1362 seq_printf(m, "\nCloses: %d total %d failed",
1363 atomic_read(&sent[SMB2_CLOSE_HE]),
1364 atomic_read(&failed[SMB2_CLOSE_HE]));
1365 seq_printf(m, "\nFlushes: %d total %d failed",
1366 atomic_read(&sent[SMB2_FLUSH_HE]),
1367 atomic_read(&failed[SMB2_FLUSH_HE]));
1368 seq_printf(m, "\nReads: %d total %d failed",
1369 atomic_read(&sent[SMB2_READ_HE]),
1370 atomic_read(&failed[SMB2_READ_HE]));
1371 seq_printf(m, "\nWrites: %d total %d failed",
1372 atomic_read(&sent[SMB2_WRITE_HE]),
1373 atomic_read(&failed[SMB2_WRITE_HE]));
1374 seq_printf(m, "\nLocks: %d total %d failed",
1375 atomic_read(&sent[SMB2_LOCK_HE]),
1376 atomic_read(&failed[SMB2_LOCK_HE]));
1377 seq_printf(m, "\nIOCTLs: %d total %d failed",
1378 atomic_read(&sent[SMB2_IOCTL_HE]),
1379 atomic_read(&failed[SMB2_IOCTL_HE]));
1380 seq_printf(m, "\nQueryDirectories: %d total %d failed",
1381 atomic_read(&sent[SMB2_QUERY_DIRECTORY_HE]),
1382 atomic_read(&failed[SMB2_QUERY_DIRECTORY_HE]));
1383 seq_printf(m, "\nChangeNotifies: %d total %d failed",
1384 atomic_read(&sent[SMB2_CHANGE_NOTIFY_HE]),
1385 atomic_read(&failed[SMB2_CHANGE_NOTIFY_HE]));
1386 seq_printf(m, "\nQueryInfos: %d total %d failed",
1387 atomic_read(&sent[SMB2_QUERY_INFO_HE]),
1388 atomic_read(&failed[SMB2_QUERY_INFO_HE]));
1389 seq_printf(m, "\nSetInfos: %d total %d failed",
1390 atomic_read(&sent[SMB2_SET_INFO_HE]),
1391 atomic_read(&failed[SMB2_SET_INFO_HE]));
1392 seq_printf(m, "\nOplockBreaks: %d sent %d failed",
1393 atomic_read(&sent[SMB2_OPLOCK_BREAK_HE]),
1394 atomic_read(&failed[SMB2_OPLOCK_BREAK_HE]));
1395 }
1396
1397 static void
smb2_set_fid(struct cifsFileInfo * cfile,struct cifs_fid * fid,__u32 oplock)1398 smb2_set_fid(struct cifsFileInfo *cfile, struct cifs_fid *fid, __u32 oplock)
1399 {
1400 struct cifsInodeInfo *cinode = CIFS_I(d_inode(cfile->dentry));
1401 struct TCP_Server_Info *server = tlink_tcon(cfile->tlink)->ses->server;
1402
1403 cfile->fid.persistent_fid = fid->persistent_fid;
1404 cfile->fid.volatile_fid = fid->volatile_fid;
1405 cfile->fid.access = fid->access;
1406 #ifdef CONFIG_CIFS_DEBUG2
1407 cfile->fid.mid = fid->mid;
1408 #endif /* CIFS_DEBUG2 */
1409 server->ops->set_oplock_level(cinode, oplock, fid->epoch,
1410 &fid->purge_cache);
1411 cinode->can_cache_brlcks = CIFS_CACHE_WRITE(cinode);
1412 memcpy(cfile->fid.create_guid, fid->create_guid, 16);
1413 }
1414
1415 static int
smb2_close_file(const unsigned int xid,struct cifs_tcon * tcon,struct cifs_fid * fid)1416 smb2_close_file(const unsigned int xid, struct cifs_tcon *tcon,
1417 struct cifs_fid *fid)
1418 {
1419 return SMB2_close(xid, tcon, fid->persistent_fid, fid->volatile_fid);
1420 }
1421
1422 static int
smb2_close_getattr(const unsigned int xid,struct cifs_tcon * tcon,struct cifsFileInfo * cfile)1423 smb2_close_getattr(const unsigned int xid, struct cifs_tcon *tcon,
1424 struct cifsFileInfo *cfile)
1425 {
1426 struct smb2_file_network_open_info file_inf;
1427 struct inode *inode;
1428 int rc;
1429
1430 rc = __SMB2_close(xid, tcon, cfile->fid.persistent_fid,
1431 cfile->fid.volatile_fid, &file_inf);
1432 if (rc)
1433 return rc;
1434
1435 inode = d_inode(cfile->dentry);
1436
1437 spin_lock(&inode->i_lock);
1438 CIFS_I(inode)->time = jiffies;
1439
1440 /* Creation time should not need to be updated on close */
1441 if (file_inf.LastWriteTime)
1442 inode_set_mtime_to_ts(inode,
1443 cifs_NTtimeToUnix(file_inf.LastWriteTime));
1444 if (file_inf.ChangeTime)
1445 inode_set_ctime_to_ts(inode,
1446 cifs_NTtimeToUnix(file_inf.ChangeTime));
1447 if (file_inf.LastAccessTime)
1448 inode_set_atime_to_ts(inode,
1449 cifs_NTtimeToUnix(file_inf.LastAccessTime));
1450
1451 /*
1452 * i_blocks is not related to (i_size / i_blksize),
1453 * but instead 512 byte (2**9) size is required for
1454 * calculating num blocks.
1455 */
1456 if (le64_to_cpu(file_inf.AllocationSize) > 4096)
1457 inode->i_blocks =
1458 (512 - 1 + le64_to_cpu(file_inf.AllocationSize)) >> 9;
1459
1460 /* End of file and Attributes should not have to be updated on close */
1461 spin_unlock(&inode->i_lock);
1462 return rc;
1463 }
1464
1465 static int
SMB2_request_res_key(const unsigned int xid,struct cifs_tcon * tcon,u64 persistent_fid,u64 volatile_fid,struct copychunk_ioctl * pcchunk)1466 SMB2_request_res_key(const unsigned int xid, struct cifs_tcon *tcon,
1467 u64 persistent_fid, u64 volatile_fid,
1468 struct copychunk_ioctl *pcchunk)
1469 {
1470 int rc;
1471 unsigned int ret_data_len;
1472 struct resume_key_req *res_key;
1473
1474 rc = SMB2_ioctl(xid, tcon, persistent_fid, volatile_fid,
1475 FSCTL_SRV_REQUEST_RESUME_KEY, NULL, 0 /* no input */,
1476 CIFSMaxBufSize, (char **)&res_key, &ret_data_len);
1477
1478 if (rc == -EOPNOTSUPP) {
1479 pr_warn_once("Server share %s does not support copy range\n", tcon->tree_name);
1480 goto req_res_key_exit;
1481 } else if (rc) {
1482 cifs_tcon_dbg(VFS, "refcpy ioctl error %d getting resume key\n", rc);
1483 goto req_res_key_exit;
1484 }
1485 if (ret_data_len < sizeof(struct resume_key_req)) {
1486 cifs_tcon_dbg(VFS, "Invalid refcopy resume key length\n");
1487 rc = -EINVAL;
1488 goto req_res_key_exit;
1489 }
1490 memcpy(pcchunk->SourceKey, res_key->ResumeKey, COPY_CHUNK_RES_KEY_SIZE);
1491
1492 req_res_key_exit:
1493 kfree(res_key);
1494 return rc;
1495 }
1496
1497 static int
smb2_ioctl_query_info(const unsigned int xid,struct cifs_tcon * tcon,struct cifs_sb_info * cifs_sb,__le16 * path,int is_dir,unsigned long p)1498 smb2_ioctl_query_info(const unsigned int xid,
1499 struct cifs_tcon *tcon,
1500 struct cifs_sb_info *cifs_sb,
1501 __le16 *path, int is_dir,
1502 unsigned long p)
1503 {
1504 struct smb2_compound_vars *vars;
1505 struct smb_rqst *rqst;
1506 struct kvec *rsp_iov;
1507 struct cifs_ses *ses = tcon->ses;
1508 struct TCP_Server_Info *server;
1509 char __user *arg = (char __user *)p;
1510 struct smb_query_info qi;
1511 struct smb_query_info __user *pqi;
1512 int rc = 0;
1513 int flags = CIFS_CP_CREATE_CLOSE_OP;
1514 struct smb2_query_info_rsp *qi_rsp = NULL;
1515 struct smb2_ioctl_rsp *io_rsp = NULL;
1516 void *buffer = NULL;
1517 int resp_buftype[3];
1518 struct cifs_open_parms oparms;
1519 u8 oplock = SMB2_OPLOCK_LEVEL_NONE;
1520 struct cifs_fid fid;
1521 unsigned int size[2];
1522 void *data[2];
1523 int create_options = is_dir ? CREATE_NOT_FILE : CREATE_NOT_DIR;
1524 void (*free_req1_func)(struct smb_rqst *r);
1525 int retries = 0, cur_sleep = 1;
1526
1527 replay_again:
1528 /* reinitialize for possible replay */
1529 flags = CIFS_CP_CREATE_CLOSE_OP;
1530 oplock = SMB2_OPLOCK_LEVEL_NONE;
1531 server = cifs_pick_channel(ses);
1532
1533 vars = kzalloc(sizeof(*vars), GFP_ATOMIC);
1534 if (vars == NULL)
1535 return -ENOMEM;
1536 rqst = &vars->rqst[0];
1537 rsp_iov = &vars->rsp_iov[0];
1538
1539 resp_buftype[0] = resp_buftype[1] = resp_buftype[2] = CIFS_NO_BUFFER;
1540
1541 if (copy_from_user(&qi, arg, sizeof(struct smb_query_info))) {
1542 rc = -EFAULT;
1543 goto free_vars;
1544 }
1545 if (qi.output_buffer_length > 1024) {
1546 rc = -EINVAL;
1547 goto free_vars;
1548 }
1549
1550 if (!ses || !server) {
1551 rc = -EIO;
1552 goto free_vars;
1553 }
1554
1555 if (smb3_encryption_required(tcon))
1556 flags |= CIFS_TRANSFORM_REQ;
1557
1558 if (qi.output_buffer_length) {
1559 buffer = memdup_user(arg + sizeof(struct smb_query_info), qi.output_buffer_length);
1560 if (IS_ERR(buffer)) {
1561 rc = PTR_ERR(buffer);
1562 goto free_vars;
1563 }
1564 }
1565
1566 /* Open */
1567 rqst[0].rq_iov = &vars->open_iov[0];
1568 rqst[0].rq_nvec = SMB2_CREATE_IOV_SIZE;
1569
1570 oparms = (struct cifs_open_parms) {
1571 .tcon = tcon,
1572 .disposition = FILE_OPEN,
1573 .create_options = cifs_create_options(cifs_sb, create_options),
1574 .fid = &fid,
1575 .replay = !!(retries),
1576 };
1577
1578 if (qi.flags & PASSTHRU_FSCTL) {
1579 switch (qi.info_type & FSCTL_DEVICE_ACCESS_MASK) {
1580 case FSCTL_DEVICE_ACCESS_FILE_READ_WRITE_ACCESS:
1581 oparms.desired_access = FILE_READ_DATA | FILE_WRITE_DATA | FILE_READ_ATTRIBUTES | SYNCHRONIZE;
1582 break;
1583 case FSCTL_DEVICE_ACCESS_FILE_ANY_ACCESS:
1584 oparms.desired_access = GENERIC_ALL;
1585 break;
1586 case FSCTL_DEVICE_ACCESS_FILE_READ_ACCESS:
1587 oparms.desired_access = GENERIC_READ;
1588 break;
1589 case FSCTL_DEVICE_ACCESS_FILE_WRITE_ACCESS:
1590 oparms.desired_access = GENERIC_WRITE;
1591 break;
1592 }
1593 } else if (qi.flags & PASSTHRU_SET_INFO) {
1594 oparms.desired_access = GENERIC_WRITE;
1595 } else {
1596 oparms.desired_access = FILE_READ_ATTRIBUTES | READ_CONTROL;
1597 }
1598
1599 rc = SMB2_open_init(tcon, server,
1600 &rqst[0], &oplock, &oparms, path);
1601 if (rc)
1602 goto free_output_buffer;
1603 smb2_set_next_command(tcon, &rqst[0]);
1604
1605 /* Query */
1606 if (qi.flags & PASSTHRU_FSCTL) {
1607 /* Can eventually relax perm check since server enforces too */
1608 if (!capable(CAP_SYS_ADMIN)) {
1609 rc = -EPERM;
1610 goto free_open_req;
1611 }
1612 rqst[1].rq_iov = &vars->io_iov[0];
1613 rqst[1].rq_nvec = SMB2_IOCTL_IOV_SIZE;
1614
1615 rc = SMB2_ioctl_init(tcon, server, &rqst[1], COMPOUND_FID, COMPOUND_FID,
1616 qi.info_type, buffer, qi.output_buffer_length,
1617 CIFSMaxBufSize - MAX_SMB2_CREATE_RESPONSE_SIZE -
1618 MAX_SMB2_CLOSE_RESPONSE_SIZE);
1619 free_req1_func = SMB2_ioctl_free;
1620 } else if (qi.flags == PASSTHRU_SET_INFO) {
1621 /* Can eventually relax perm check since server enforces too */
1622 if (!capable(CAP_SYS_ADMIN)) {
1623 rc = -EPERM;
1624 goto free_open_req;
1625 }
1626 if (qi.output_buffer_length < 8) {
1627 rc = -EINVAL;
1628 goto free_open_req;
1629 }
1630 rqst[1].rq_iov = vars->si_iov;
1631 rqst[1].rq_nvec = 1;
1632
1633 /* MS-FSCC 2.4.13 FileEndOfFileInformation */
1634 size[0] = 8;
1635 data[0] = buffer;
1636
1637 rc = SMB2_set_info_init(tcon, server, &rqst[1], COMPOUND_FID, COMPOUND_FID,
1638 current->tgid, FILE_END_OF_FILE_INFORMATION,
1639 SMB2_O_INFO_FILE, 0, data, size);
1640 free_req1_func = SMB2_set_info_free;
1641 } else if (qi.flags == PASSTHRU_QUERY_INFO) {
1642 rqst[1].rq_iov = &vars->qi_iov;
1643 rqst[1].rq_nvec = 1;
1644
1645 rc = SMB2_query_info_init(tcon, server,
1646 &rqst[1], COMPOUND_FID,
1647 COMPOUND_FID, qi.file_info_class,
1648 qi.info_type, qi.additional_information,
1649 qi.input_buffer_length,
1650 qi.output_buffer_length, buffer);
1651 free_req1_func = SMB2_query_info_free;
1652 } else { /* unknown flags */
1653 cifs_tcon_dbg(VFS, "Invalid passthru query flags: 0x%x\n",
1654 qi.flags);
1655 rc = -EINVAL;
1656 }
1657
1658 if (rc)
1659 goto free_open_req;
1660 smb2_set_next_command(tcon, &rqst[1]);
1661 smb2_set_related(&rqst[1]);
1662
1663 /* Close */
1664 rqst[2].rq_iov = &vars->close_iov;
1665 rqst[2].rq_nvec = 1;
1666
1667 rc = SMB2_close_init(tcon, server,
1668 &rqst[2], COMPOUND_FID, COMPOUND_FID, false);
1669 if (rc)
1670 goto free_req_1;
1671 smb2_set_related(&rqst[2]);
1672
1673 if (retries) {
1674 smb2_set_replay(server, &rqst[0]);
1675 smb2_set_replay(server, &rqst[1]);
1676 smb2_set_replay(server, &rqst[2]);
1677 }
1678
1679 rc = compound_send_recv(xid, ses, server,
1680 flags, 3, rqst,
1681 resp_buftype, rsp_iov);
1682 if (rc)
1683 goto out;
1684
1685 /* No need to bump num_remote_opens since handle immediately closed */
1686 if (qi.flags & PASSTHRU_FSCTL) {
1687 pqi = (struct smb_query_info __user *)arg;
1688 io_rsp = (struct smb2_ioctl_rsp *)rsp_iov[1].iov_base;
1689 if (le32_to_cpu(io_rsp->OutputCount) < qi.input_buffer_length)
1690 qi.input_buffer_length = le32_to_cpu(io_rsp->OutputCount);
1691 if (qi.input_buffer_length > 0 &&
1692 le32_to_cpu(io_rsp->OutputOffset) + qi.input_buffer_length
1693 > rsp_iov[1].iov_len) {
1694 rc = -EFAULT;
1695 goto out;
1696 }
1697
1698 if (copy_to_user(&pqi->input_buffer_length,
1699 &qi.input_buffer_length,
1700 sizeof(qi.input_buffer_length))) {
1701 rc = -EFAULT;
1702 goto out;
1703 }
1704
1705 if (copy_to_user((void __user *)pqi + sizeof(struct smb_query_info),
1706 (const void *)io_rsp + le32_to_cpu(io_rsp->OutputOffset),
1707 qi.input_buffer_length))
1708 rc = -EFAULT;
1709 } else {
1710 pqi = (struct smb_query_info __user *)arg;
1711 qi_rsp = (struct smb2_query_info_rsp *)rsp_iov[1].iov_base;
1712 if (le32_to_cpu(qi_rsp->OutputBufferLength) < qi.input_buffer_length)
1713 qi.input_buffer_length = le32_to_cpu(qi_rsp->OutputBufferLength);
1714 if (copy_to_user(&pqi->input_buffer_length,
1715 &qi.input_buffer_length,
1716 sizeof(qi.input_buffer_length))) {
1717 rc = -EFAULT;
1718 goto out;
1719 }
1720
1721 if (copy_to_user(pqi + 1, qi_rsp->Buffer,
1722 qi.input_buffer_length))
1723 rc = -EFAULT;
1724 }
1725
1726 out:
1727 free_rsp_buf(resp_buftype[0], rsp_iov[0].iov_base);
1728 free_rsp_buf(resp_buftype[1], rsp_iov[1].iov_base);
1729 free_rsp_buf(resp_buftype[2], rsp_iov[2].iov_base);
1730 SMB2_close_free(&rqst[2]);
1731 free_req_1:
1732 free_req1_func(&rqst[1]);
1733 free_open_req:
1734 SMB2_open_free(&rqst[0]);
1735 free_output_buffer:
1736 kfree(buffer);
1737 free_vars:
1738 kfree(vars);
1739
1740 if (is_replayable_error(rc) &&
1741 smb2_should_replay(tcon, &retries, &cur_sleep))
1742 goto replay_again;
1743
1744 return rc;
1745 }
1746
1747 static ssize_t
smb2_copychunk_range(const unsigned int xid,struct cifsFileInfo * srcfile,struct cifsFileInfo * trgtfile,u64 src_off,u64 len,u64 dest_off)1748 smb2_copychunk_range(const unsigned int xid,
1749 struct cifsFileInfo *srcfile,
1750 struct cifsFileInfo *trgtfile, u64 src_off,
1751 u64 len, u64 dest_off)
1752 {
1753 int rc;
1754 unsigned int ret_data_len;
1755 struct copychunk_ioctl *pcchunk;
1756 struct copychunk_ioctl_rsp *retbuf = NULL;
1757 struct cifs_tcon *tcon;
1758 int chunks_copied = 0;
1759 bool chunk_sizes_updated = false;
1760 ssize_t bytes_written, total_bytes_written = 0;
1761
1762 pcchunk = kmalloc(sizeof(struct copychunk_ioctl), GFP_KERNEL);
1763 if (pcchunk == NULL)
1764 return -ENOMEM;
1765
1766 cifs_dbg(FYI, "%s: about to call request res key\n", __func__);
1767 /* Request a key from the server to identify the source of the copy */
1768 rc = SMB2_request_res_key(xid, tlink_tcon(srcfile->tlink),
1769 srcfile->fid.persistent_fid,
1770 srcfile->fid.volatile_fid, pcchunk);
1771
1772 /* Note: request_res_key sets res_key null only if rc !=0 */
1773 if (rc)
1774 goto cchunk_out;
1775
1776 /* For now array only one chunk long, will make more flexible later */
1777 pcchunk->ChunkCount = cpu_to_le32(1);
1778 pcchunk->Reserved = 0;
1779 pcchunk->Reserved2 = 0;
1780
1781 tcon = tlink_tcon(trgtfile->tlink);
1782
1783 while (len > 0) {
1784 pcchunk->SourceOffset = cpu_to_le64(src_off);
1785 pcchunk->TargetOffset = cpu_to_le64(dest_off);
1786 pcchunk->Length =
1787 cpu_to_le32(min_t(u64, len, tcon->max_bytes_chunk));
1788
1789 /* Request server copy to target from src identified by key */
1790 kfree(retbuf);
1791 retbuf = NULL;
1792 rc = SMB2_ioctl(xid, tcon, trgtfile->fid.persistent_fid,
1793 trgtfile->fid.volatile_fid, FSCTL_SRV_COPYCHUNK_WRITE,
1794 (char *)pcchunk, sizeof(struct copychunk_ioctl),
1795 CIFSMaxBufSize, (char **)&retbuf, &ret_data_len);
1796 if (rc == 0) {
1797 if (ret_data_len !=
1798 sizeof(struct copychunk_ioctl_rsp)) {
1799 cifs_tcon_dbg(VFS, "Invalid cchunk response size\n");
1800 rc = -EIO;
1801 goto cchunk_out;
1802 }
1803 if (retbuf->TotalBytesWritten == 0) {
1804 cifs_dbg(FYI, "no bytes copied\n");
1805 rc = -EIO;
1806 goto cchunk_out;
1807 }
1808 /*
1809 * Check if server claimed to write more than we asked
1810 */
1811 if (le32_to_cpu(retbuf->TotalBytesWritten) >
1812 le32_to_cpu(pcchunk->Length)) {
1813 cifs_tcon_dbg(VFS, "Invalid copy chunk response\n");
1814 rc = -EIO;
1815 goto cchunk_out;
1816 }
1817 if (le32_to_cpu(retbuf->ChunksWritten) != 1) {
1818 cifs_tcon_dbg(VFS, "Invalid num chunks written\n");
1819 rc = -EIO;
1820 goto cchunk_out;
1821 }
1822 chunks_copied++;
1823
1824 bytes_written = le32_to_cpu(retbuf->TotalBytesWritten);
1825 src_off += bytes_written;
1826 dest_off += bytes_written;
1827 len -= bytes_written;
1828 total_bytes_written += bytes_written;
1829
1830 cifs_dbg(FYI, "Chunks %d PartialChunk %d Total %zu\n",
1831 le32_to_cpu(retbuf->ChunksWritten),
1832 le32_to_cpu(retbuf->ChunkBytesWritten),
1833 bytes_written);
1834 } else if (rc == -EINVAL) {
1835 if (ret_data_len != sizeof(struct copychunk_ioctl_rsp))
1836 goto cchunk_out;
1837
1838 cifs_dbg(FYI, "MaxChunks %d BytesChunk %d MaxCopy %d\n",
1839 le32_to_cpu(retbuf->ChunksWritten),
1840 le32_to_cpu(retbuf->ChunkBytesWritten),
1841 le32_to_cpu(retbuf->TotalBytesWritten));
1842
1843 /*
1844 * Check if this is the first request using these sizes,
1845 * (ie check if copy succeed once with original sizes
1846 * and check if the server gave us different sizes after
1847 * we already updated max sizes on previous request).
1848 * if not then why is the server returning an error now
1849 */
1850 if ((chunks_copied != 0) || chunk_sizes_updated)
1851 goto cchunk_out;
1852
1853 /* Check that server is not asking us to grow size */
1854 if (le32_to_cpu(retbuf->ChunkBytesWritten) <
1855 tcon->max_bytes_chunk)
1856 tcon->max_bytes_chunk =
1857 le32_to_cpu(retbuf->ChunkBytesWritten);
1858 else
1859 goto cchunk_out; /* server gave us bogus size */
1860
1861 /* No need to change MaxChunks since already set to 1 */
1862 chunk_sizes_updated = true;
1863 } else
1864 goto cchunk_out;
1865 }
1866
1867 cchunk_out:
1868 kfree(pcchunk);
1869 kfree(retbuf);
1870 if (rc)
1871 return rc;
1872 else
1873 return total_bytes_written;
1874 }
1875
1876 static int
smb2_flush_file(const unsigned int xid,struct cifs_tcon * tcon,struct cifs_fid * fid)1877 smb2_flush_file(const unsigned int xid, struct cifs_tcon *tcon,
1878 struct cifs_fid *fid)
1879 {
1880 return SMB2_flush(xid, tcon, fid->persistent_fid, fid->volatile_fid);
1881 }
1882
1883 static unsigned int
smb2_read_data_offset(char * buf)1884 smb2_read_data_offset(char *buf)
1885 {
1886 struct smb2_read_rsp *rsp = (struct smb2_read_rsp *)buf;
1887
1888 return rsp->DataOffset;
1889 }
1890
1891 static unsigned int
smb2_read_data_length(char * buf,bool in_remaining)1892 smb2_read_data_length(char *buf, bool in_remaining)
1893 {
1894 struct smb2_read_rsp *rsp = (struct smb2_read_rsp *)buf;
1895
1896 if (in_remaining)
1897 return le32_to_cpu(rsp->DataRemaining);
1898
1899 return le32_to_cpu(rsp->DataLength);
1900 }
1901
1902
1903 static int
smb2_sync_read(const unsigned int xid,struct cifs_fid * pfid,struct cifs_io_parms * parms,unsigned int * bytes_read,char ** buf,int * buf_type)1904 smb2_sync_read(const unsigned int xid, struct cifs_fid *pfid,
1905 struct cifs_io_parms *parms, unsigned int *bytes_read,
1906 char **buf, int *buf_type)
1907 {
1908 parms->persistent_fid = pfid->persistent_fid;
1909 parms->volatile_fid = pfid->volatile_fid;
1910 return SMB2_read(xid, parms, bytes_read, buf, buf_type);
1911 }
1912
1913 static int
smb2_sync_write(const unsigned int xid,struct cifs_fid * pfid,struct cifs_io_parms * parms,unsigned int * written,struct kvec * iov,unsigned long nr_segs)1914 smb2_sync_write(const unsigned int xid, struct cifs_fid *pfid,
1915 struct cifs_io_parms *parms, unsigned int *written,
1916 struct kvec *iov, unsigned long nr_segs)
1917 {
1918
1919 parms->persistent_fid = pfid->persistent_fid;
1920 parms->volatile_fid = pfid->volatile_fid;
1921 return SMB2_write(xid, parms, written, iov, nr_segs);
1922 }
1923
1924 /* Set or clear the SPARSE_FILE attribute based on value passed in setsparse */
smb2_set_sparse(const unsigned int xid,struct cifs_tcon * tcon,struct cifsFileInfo * cfile,struct inode * inode,__u8 setsparse)1925 static bool smb2_set_sparse(const unsigned int xid, struct cifs_tcon *tcon,
1926 struct cifsFileInfo *cfile, struct inode *inode, __u8 setsparse)
1927 {
1928 struct cifsInodeInfo *cifsi;
1929 int rc;
1930
1931 cifsi = CIFS_I(inode);
1932
1933 /* if file already sparse don't bother setting sparse again */
1934 if ((cifsi->cifsAttrs & FILE_ATTRIBUTE_SPARSE_FILE) && setsparse)
1935 return true; /* already sparse */
1936
1937 if (!(cifsi->cifsAttrs & FILE_ATTRIBUTE_SPARSE_FILE) && !setsparse)
1938 return true; /* already not sparse */
1939
1940 /*
1941 * Can't check for sparse support on share the usual way via the
1942 * FS attribute info (FILE_SUPPORTS_SPARSE_FILES) on the share
1943 * since Samba server doesn't set the flag on the share, yet
1944 * supports the set sparse FSCTL and returns sparse correctly
1945 * in the file attributes. If we fail setting sparse though we
1946 * mark that server does not support sparse files for this share
1947 * to avoid repeatedly sending the unsupported fsctl to server
1948 * if the file is repeatedly extended.
1949 */
1950 if (tcon->broken_sparse_sup)
1951 return false;
1952
1953 rc = SMB2_ioctl(xid, tcon, cfile->fid.persistent_fid,
1954 cfile->fid.volatile_fid, FSCTL_SET_SPARSE,
1955 &setsparse, 1, CIFSMaxBufSize, NULL, NULL);
1956 if (rc) {
1957 tcon->broken_sparse_sup = true;
1958 cifs_dbg(FYI, "set sparse rc = %d\n", rc);
1959 return false;
1960 }
1961
1962 if (setsparse)
1963 cifsi->cifsAttrs |= FILE_ATTRIBUTE_SPARSE_FILE;
1964 else
1965 cifsi->cifsAttrs &= (~FILE_ATTRIBUTE_SPARSE_FILE);
1966
1967 return true;
1968 }
1969
1970 static int
smb2_set_file_size(const unsigned int xid,struct cifs_tcon * tcon,struct cifsFileInfo * cfile,__u64 size,bool set_alloc)1971 smb2_set_file_size(const unsigned int xid, struct cifs_tcon *tcon,
1972 struct cifsFileInfo *cfile, __u64 size, bool set_alloc)
1973 {
1974 struct inode *inode;
1975
1976 /*
1977 * If extending file more than one page make sparse. Many Linux fs
1978 * make files sparse by default when extending via ftruncate
1979 */
1980 inode = d_inode(cfile->dentry);
1981
1982 if (!set_alloc && (size > inode->i_size + 8192)) {
1983 __u8 set_sparse = 1;
1984
1985 /* whether set sparse succeeds or not, extend the file */
1986 smb2_set_sparse(xid, tcon, cfile, inode, set_sparse);
1987 }
1988
1989 return SMB2_set_eof(xid, tcon, cfile->fid.persistent_fid,
1990 cfile->fid.volatile_fid, cfile->pid, size);
1991 }
1992
1993 static int
smb2_duplicate_extents(const unsigned int xid,struct cifsFileInfo * srcfile,struct cifsFileInfo * trgtfile,u64 src_off,u64 len,u64 dest_off)1994 smb2_duplicate_extents(const unsigned int xid,
1995 struct cifsFileInfo *srcfile,
1996 struct cifsFileInfo *trgtfile, u64 src_off,
1997 u64 len, u64 dest_off)
1998 {
1999 int rc;
2000 unsigned int ret_data_len;
2001 struct inode *inode;
2002 struct duplicate_extents_to_file dup_ext_buf;
2003 struct cifs_tcon *tcon = tlink_tcon(trgtfile->tlink);
2004
2005 /* server fileays advertise duplicate extent support with this flag */
2006 if ((le32_to_cpu(tcon->fsAttrInfo.Attributes) &
2007 FILE_SUPPORTS_BLOCK_REFCOUNTING) == 0)
2008 return -EOPNOTSUPP;
2009
2010 dup_ext_buf.VolatileFileHandle = srcfile->fid.volatile_fid;
2011 dup_ext_buf.PersistentFileHandle = srcfile->fid.persistent_fid;
2012 dup_ext_buf.SourceFileOffset = cpu_to_le64(src_off);
2013 dup_ext_buf.TargetFileOffset = cpu_to_le64(dest_off);
2014 dup_ext_buf.ByteCount = cpu_to_le64(len);
2015 cifs_dbg(FYI, "Duplicate extents: src off %lld dst off %lld len %lld\n",
2016 src_off, dest_off, len);
2017
2018 inode = d_inode(trgtfile->dentry);
2019 if (inode->i_size < dest_off + len) {
2020 rc = smb2_set_file_size(xid, tcon, trgtfile, dest_off + len, false);
2021 if (rc)
2022 goto duplicate_extents_out;
2023
2024 /*
2025 * Although also could set plausible allocation size (i_blocks)
2026 * here in addition to setting the file size, in reflink
2027 * it is likely that the target file is sparse. Its allocation
2028 * size will be queried on next revalidate, but it is important
2029 * to make sure that file's cached size is updated immediately
2030 */
2031 cifs_setsize(inode, dest_off + len);
2032 }
2033 rc = SMB2_ioctl(xid, tcon, trgtfile->fid.persistent_fid,
2034 trgtfile->fid.volatile_fid,
2035 FSCTL_DUPLICATE_EXTENTS_TO_FILE,
2036 (char *)&dup_ext_buf,
2037 sizeof(struct duplicate_extents_to_file),
2038 CIFSMaxBufSize, NULL,
2039 &ret_data_len);
2040
2041 if (ret_data_len > 0)
2042 cifs_dbg(FYI, "Non-zero response length in duplicate extents\n");
2043
2044 duplicate_extents_out:
2045 return rc;
2046 }
2047
2048 static int
smb2_set_compression(const unsigned int xid,struct cifs_tcon * tcon,struct cifsFileInfo * cfile)2049 smb2_set_compression(const unsigned int xid, struct cifs_tcon *tcon,
2050 struct cifsFileInfo *cfile)
2051 {
2052 return SMB2_set_compression(xid, tcon, cfile->fid.persistent_fid,
2053 cfile->fid.volatile_fid);
2054 }
2055
2056 static int
smb3_set_integrity(const unsigned int xid,struct cifs_tcon * tcon,struct cifsFileInfo * cfile)2057 smb3_set_integrity(const unsigned int xid, struct cifs_tcon *tcon,
2058 struct cifsFileInfo *cfile)
2059 {
2060 struct fsctl_set_integrity_information_req integr_info;
2061 unsigned int ret_data_len;
2062
2063 integr_info.ChecksumAlgorithm = cpu_to_le16(CHECKSUM_TYPE_UNCHANGED);
2064 integr_info.Flags = 0;
2065 integr_info.Reserved = 0;
2066
2067 return SMB2_ioctl(xid, tcon, cfile->fid.persistent_fid,
2068 cfile->fid.volatile_fid,
2069 FSCTL_SET_INTEGRITY_INFORMATION,
2070 (char *)&integr_info,
2071 sizeof(struct fsctl_set_integrity_information_req),
2072 CIFSMaxBufSize, NULL,
2073 &ret_data_len);
2074
2075 }
2076
2077 /* GMT Token is @GMT-YYYY.MM.DD-HH.MM.SS Unicode which is 48 bytes + null */
2078 #define GMT_TOKEN_SIZE 50
2079
2080 #define MIN_SNAPSHOT_ARRAY_SIZE 16 /* See MS-SMB2 section 3.3.5.15.1 */
2081
2082 /*
2083 * Input buffer contains (empty) struct smb_snapshot array with size filled in
2084 * For output see struct SRV_SNAPSHOT_ARRAY in MS-SMB2 section 2.2.32.2
2085 */
2086 static int
smb3_enum_snapshots(const unsigned int xid,struct cifs_tcon * tcon,struct cifsFileInfo * cfile,void __user * ioc_buf)2087 smb3_enum_snapshots(const unsigned int xid, struct cifs_tcon *tcon,
2088 struct cifsFileInfo *cfile, void __user *ioc_buf)
2089 {
2090 char *retbuf = NULL;
2091 unsigned int ret_data_len = 0;
2092 int rc;
2093 u32 max_response_size;
2094 struct smb_snapshot_array snapshot_in;
2095
2096 /*
2097 * On the first query to enumerate the list of snapshots available
2098 * for this volume the buffer begins with 0 (number of snapshots
2099 * which can be returned is zero since at that point we do not know
2100 * how big the buffer needs to be). On the second query,
2101 * it (ret_data_len) is set to number of snapshots so we can
2102 * know to set the maximum response size larger (see below).
2103 */
2104 if (get_user(ret_data_len, (unsigned int __user *)ioc_buf))
2105 return -EFAULT;
2106
2107 /*
2108 * Note that for snapshot queries that servers like Azure expect that
2109 * the first query be minimal size (and just used to get the number/size
2110 * of previous versions) so response size must be specified as EXACTLY
2111 * sizeof(struct snapshot_array) which is 16 when rounded up to multiple
2112 * of eight bytes.
2113 */
2114 if (ret_data_len == 0)
2115 max_response_size = MIN_SNAPSHOT_ARRAY_SIZE;
2116 else
2117 max_response_size = CIFSMaxBufSize;
2118
2119 rc = SMB2_ioctl(xid, tcon, cfile->fid.persistent_fid,
2120 cfile->fid.volatile_fid,
2121 FSCTL_SRV_ENUMERATE_SNAPSHOTS,
2122 NULL, 0 /* no input data */, max_response_size,
2123 (char **)&retbuf,
2124 &ret_data_len);
2125 cifs_dbg(FYI, "enum snaphots ioctl returned %d and ret buflen is %d\n",
2126 rc, ret_data_len);
2127 if (rc)
2128 return rc;
2129
2130 if (ret_data_len && (ioc_buf != NULL) && (retbuf != NULL)) {
2131 /* Fixup buffer */
2132 if (copy_from_user(&snapshot_in, ioc_buf,
2133 sizeof(struct smb_snapshot_array))) {
2134 rc = -EFAULT;
2135 kfree(retbuf);
2136 return rc;
2137 }
2138
2139 /*
2140 * Check for min size, ie not large enough to fit even one GMT
2141 * token (snapshot). On the first ioctl some users may pass in
2142 * smaller size (or zero) to simply get the size of the array
2143 * so the user space caller can allocate sufficient memory
2144 * and retry the ioctl again with larger array size sufficient
2145 * to hold all of the snapshot GMT tokens on the second try.
2146 */
2147 if (snapshot_in.snapshot_array_size < GMT_TOKEN_SIZE)
2148 ret_data_len = sizeof(struct smb_snapshot_array);
2149
2150 /*
2151 * We return struct SRV_SNAPSHOT_ARRAY, followed by
2152 * the snapshot array (of 50 byte GMT tokens) each
2153 * representing an available previous version of the data
2154 */
2155 if (ret_data_len > (snapshot_in.snapshot_array_size +
2156 sizeof(struct smb_snapshot_array)))
2157 ret_data_len = snapshot_in.snapshot_array_size +
2158 sizeof(struct smb_snapshot_array);
2159
2160 if (copy_to_user(ioc_buf, retbuf, ret_data_len))
2161 rc = -EFAULT;
2162 }
2163
2164 kfree(retbuf);
2165 return rc;
2166 }
2167
2168
2169
2170 static int
smb3_notify(const unsigned int xid,struct file * pfile,void __user * ioc_buf,bool return_changes)2171 smb3_notify(const unsigned int xid, struct file *pfile,
2172 void __user *ioc_buf, bool return_changes)
2173 {
2174 struct smb3_notify_info notify;
2175 struct smb3_notify_info __user *pnotify_buf;
2176 struct dentry *dentry = pfile->f_path.dentry;
2177 struct inode *inode = file_inode(pfile);
2178 struct cifs_sb_info *cifs_sb = CIFS_SB(inode->i_sb);
2179 struct cifs_open_parms oparms;
2180 struct cifs_fid fid;
2181 struct cifs_tcon *tcon;
2182 const unsigned char *path;
2183 char *returned_ioctl_info = NULL;
2184 void *page = alloc_dentry_path();
2185 __le16 *utf16_path = NULL;
2186 u8 oplock = SMB2_OPLOCK_LEVEL_NONE;
2187 int rc = 0;
2188 __u32 ret_len = 0;
2189
2190 path = build_path_from_dentry(dentry, page);
2191 if (IS_ERR(path)) {
2192 rc = PTR_ERR(path);
2193 goto notify_exit;
2194 }
2195
2196 utf16_path = cifs_convert_path_to_utf16(path, cifs_sb);
2197 if (utf16_path == NULL) {
2198 rc = -ENOMEM;
2199 goto notify_exit;
2200 }
2201
2202 if (return_changes) {
2203 if (copy_from_user(¬ify, ioc_buf, sizeof(struct smb3_notify_info))) {
2204 rc = -EFAULT;
2205 goto notify_exit;
2206 }
2207 } else {
2208 if (copy_from_user(¬ify, ioc_buf, sizeof(struct smb3_notify))) {
2209 rc = -EFAULT;
2210 goto notify_exit;
2211 }
2212 notify.data_len = 0;
2213 }
2214
2215 tcon = cifs_sb_master_tcon(cifs_sb);
2216 oparms = (struct cifs_open_parms) {
2217 .tcon = tcon,
2218 .path = path,
2219 .desired_access = FILE_READ_ATTRIBUTES | FILE_READ_DATA,
2220 .disposition = FILE_OPEN,
2221 .create_options = cifs_create_options(cifs_sb, 0),
2222 .fid = &fid,
2223 };
2224
2225 rc = SMB2_open(xid, &oparms, utf16_path, &oplock, NULL, NULL, NULL,
2226 NULL);
2227 if (rc)
2228 goto notify_exit;
2229
2230 rc = SMB2_change_notify(xid, tcon, fid.persistent_fid, fid.volatile_fid,
2231 notify.watch_tree, notify.completion_filter,
2232 notify.data_len, &returned_ioctl_info, &ret_len);
2233
2234 SMB2_close(xid, tcon, fid.persistent_fid, fid.volatile_fid);
2235
2236 cifs_dbg(FYI, "change notify for path %s rc %d\n", path, rc);
2237 if (return_changes && (ret_len > 0) && (notify.data_len > 0)) {
2238 if (ret_len > notify.data_len)
2239 ret_len = notify.data_len;
2240 pnotify_buf = (struct smb3_notify_info __user *)ioc_buf;
2241 if (copy_to_user(pnotify_buf->notify_data, returned_ioctl_info, ret_len))
2242 rc = -EFAULT;
2243 else if (copy_to_user(&pnotify_buf->data_len, &ret_len, sizeof(ret_len)))
2244 rc = -EFAULT;
2245 }
2246 kfree(returned_ioctl_info);
2247 notify_exit:
2248 free_dentry_path(page);
2249 kfree(utf16_path);
2250 return rc;
2251 }
2252
2253 static int
smb2_query_dir_first(const unsigned int xid,struct cifs_tcon * tcon,const char * path,struct cifs_sb_info * cifs_sb,struct cifs_fid * fid,__u16 search_flags,struct cifs_search_info * srch_inf)2254 smb2_query_dir_first(const unsigned int xid, struct cifs_tcon *tcon,
2255 const char *path, struct cifs_sb_info *cifs_sb,
2256 struct cifs_fid *fid, __u16 search_flags,
2257 struct cifs_search_info *srch_inf)
2258 {
2259 __le16 *utf16_path;
2260 struct smb_rqst rqst[2];
2261 struct kvec rsp_iov[2];
2262 int resp_buftype[2];
2263 struct kvec open_iov[SMB2_CREATE_IOV_SIZE];
2264 struct kvec qd_iov[SMB2_QUERY_DIRECTORY_IOV_SIZE];
2265 int rc, flags = 0;
2266 u8 oplock = SMB2_OPLOCK_LEVEL_NONE;
2267 struct cifs_open_parms oparms;
2268 struct smb2_query_directory_rsp *qd_rsp = NULL;
2269 struct smb2_create_rsp *op_rsp = NULL;
2270 struct TCP_Server_Info *server;
2271 int retries = 0, cur_sleep = 1;
2272
2273 replay_again:
2274 /* reinitialize for possible replay */
2275 flags = 0;
2276 oplock = SMB2_OPLOCK_LEVEL_NONE;
2277 server = cifs_pick_channel(tcon->ses);
2278
2279 utf16_path = cifs_convert_path_to_utf16(path, cifs_sb);
2280 if (!utf16_path)
2281 return -ENOMEM;
2282
2283 if (smb3_encryption_required(tcon))
2284 flags |= CIFS_TRANSFORM_REQ;
2285
2286 memset(rqst, 0, sizeof(rqst));
2287 resp_buftype[0] = resp_buftype[1] = CIFS_NO_BUFFER;
2288 memset(rsp_iov, 0, sizeof(rsp_iov));
2289
2290 /* Open */
2291 memset(&open_iov, 0, sizeof(open_iov));
2292 rqst[0].rq_iov = open_iov;
2293 rqst[0].rq_nvec = SMB2_CREATE_IOV_SIZE;
2294
2295 oparms = (struct cifs_open_parms) {
2296 .tcon = tcon,
2297 .path = path,
2298 .desired_access = FILE_READ_ATTRIBUTES | FILE_READ_DATA,
2299 .disposition = FILE_OPEN,
2300 .create_options = cifs_create_options(cifs_sb, 0),
2301 .fid = fid,
2302 .replay = !!(retries),
2303 };
2304
2305 rc = SMB2_open_init(tcon, server,
2306 &rqst[0], &oplock, &oparms, utf16_path);
2307 if (rc)
2308 goto qdf_free;
2309 smb2_set_next_command(tcon, &rqst[0]);
2310
2311 /* Query directory */
2312 srch_inf->entries_in_buffer = 0;
2313 srch_inf->index_of_last_entry = 2;
2314
2315 memset(&qd_iov, 0, sizeof(qd_iov));
2316 rqst[1].rq_iov = qd_iov;
2317 rqst[1].rq_nvec = SMB2_QUERY_DIRECTORY_IOV_SIZE;
2318
2319 rc = SMB2_query_directory_init(xid, tcon, server,
2320 &rqst[1],
2321 COMPOUND_FID, COMPOUND_FID,
2322 0, srch_inf->info_level);
2323 if (rc)
2324 goto qdf_free;
2325
2326 smb2_set_related(&rqst[1]);
2327
2328 if (retries) {
2329 smb2_set_replay(server, &rqst[0]);
2330 smb2_set_replay(server, &rqst[1]);
2331 }
2332
2333 rc = compound_send_recv(xid, tcon->ses, server,
2334 flags, 2, rqst,
2335 resp_buftype, rsp_iov);
2336
2337 /* If the open failed there is nothing to do */
2338 op_rsp = (struct smb2_create_rsp *)rsp_iov[0].iov_base;
2339 if (op_rsp == NULL || op_rsp->hdr.Status != STATUS_SUCCESS) {
2340 cifs_dbg(FYI, "query_dir_first: open failed rc=%d\n", rc);
2341 goto qdf_free;
2342 }
2343 fid->persistent_fid = op_rsp->PersistentFileId;
2344 fid->volatile_fid = op_rsp->VolatileFileId;
2345
2346 /* Anything else than ENODATA means a genuine error */
2347 if (rc && rc != -ENODATA) {
2348 SMB2_close(xid, tcon, fid->persistent_fid, fid->volatile_fid);
2349 cifs_dbg(FYI, "query_dir_first: query directory failed rc=%d\n", rc);
2350 trace_smb3_query_dir_err(xid, fid->persistent_fid,
2351 tcon->tid, tcon->ses->Suid, 0, 0, rc);
2352 goto qdf_free;
2353 }
2354
2355 atomic_inc(&tcon->num_remote_opens);
2356
2357 qd_rsp = (struct smb2_query_directory_rsp *)rsp_iov[1].iov_base;
2358 if (qd_rsp->hdr.Status == STATUS_NO_MORE_FILES) {
2359 trace_smb3_query_dir_done(xid, fid->persistent_fid,
2360 tcon->tid, tcon->ses->Suid, 0, 0);
2361 srch_inf->endOfSearch = true;
2362 rc = 0;
2363 goto qdf_free;
2364 }
2365
2366 rc = smb2_parse_query_directory(tcon, &rsp_iov[1], resp_buftype[1],
2367 srch_inf);
2368 if (rc) {
2369 trace_smb3_query_dir_err(xid, fid->persistent_fid, tcon->tid,
2370 tcon->ses->Suid, 0, 0, rc);
2371 goto qdf_free;
2372 }
2373 resp_buftype[1] = CIFS_NO_BUFFER;
2374
2375 trace_smb3_query_dir_done(xid, fid->persistent_fid, tcon->tid,
2376 tcon->ses->Suid, 0, srch_inf->entries_in_buffer);
2377
2378 qdf_free:
2379 kfree(utf16_path);
2380 SMB2_open_free(&rqst[0]);
2381 SMB2_query_directory_free(&rqst[1]);
2382 free_rsp_buf(resp_buftype[0], rsp_iov[0].iov_base);
2383 free_rsp_buf(resp_buftype[1], rsp_iov[1].iov_base);
2384
2385 if (is_replayable_error(rc) &&
2386 smb2_should_replay(tcon, &retries, &cur_sleep))
2387 goto replay_again;
2388
2389 return rc;
2390 }
2391
2392 static int
smb2_query_dir_next(const unsigned int xid,struct cifs_tcon * tcon,struct cifs_fid * fid,__u16 search_flags,struct cifs_search_info * srch_inf)2393 smb2_query_dir_next(const unsigned int xid, struct cifs_tcon *tcon,
2394 struct cifs_fid *fid, __u16 search_flags,
2395 struct cifs_search_info *srch_inf)
2396 {
2397 return SMB2_query_directory(xid, tcon, fid->persistent_fid,
2398 fid->volatile_fid, 0, srch_inf);
2399 }
2400
2401 static int
smb2_close_dir(const unsigned int xid,struct cifs_tcon * tcon,struct cifs_fid * fid)2402 smb2_close_dir(const unsigned int xid, struct cifs_tcon *tcon,
2403 struct cifs_fid *fid)
2404 {
2405 return SMB2_close(xid, tcon, fid->persistent_fid, fid->volatile_fid);
2406 }
2407
2408 /*
2409 * If we negotiate SMB2 protocol and get STATUS_PENDING - update
2410 * the number of credits and return true. Otherwise - return false.
2411 */
2412 static bool
smb2_is_status_pending(char * buf,struct TCP_Server_Info * server)2413 smb2_is_status_pending(char *buf, struct TCP_Server_Info *server)
2414 {
2415 struct smb2_hdr *shdr = (struct smb2_hdr *)buf;
2416 int scredits, in_flight;
2417
2418 if (shdr->Status != STATUS_PENDING)
2419 return false;
2420
2421 if (shdr->CreditRequest) {
2422 spin_lock(&server->req_lock);
2423 server->credits += le16_to_cpu(shdr->CreditRequest);
2424 scredits = server->credits;
2425 in_flight = server->in_flight;
2426 spin_unlock(&server->req_lock);
2427 wake_up(&server->request_q);
2428
2429 trace_smb3_pend_credits(server->CurrentMid,
2430 server->conn_id, server->hostname, scredits,
2431 le16_to_cpu(shdr->CreditRequest), in_flight);
2432 cifs_dbg(FYI, "%s: status pending add %u credits total=%d\n",
2433 __func__, le16_to_cpu(shdr->CreditRequest), scredits);
2434 }
2435
2436 return true;
2437 }
2438
2439 static bool
smb2_is_session_expired(char * buf)2440 smb2_is_session_expired(char *buf)
2441 {
2442 struct smb2_hdr *shdr = (struct smb2_hdr *)buf;
2443
2444 if (shdr->Status != STATUS_NETWORK_SESSION_EXPIRED &&
2445 shdr->Status != STATUS_USER_SESSION_DELETED)
2446 return false;
2447
2448 trace_smb3_ses_expired(le32_to_cpu(shdr->Id.SyncId.TreeId),
2449 le64_to_cpu(shdr->SessionId),
2450 le16_to_cpu(shdr->Command),
2451 le64_to_cpu(shdr->MessageId));
2452 cifs_dbg(FYI, "Session expired or deleted\n");
2453
2454 return true;
2455 }
2456
2457 static bool
smb2_is_status_io_timeout(char * buf)2458 smb2_is_status_io_timeout(char *buf)
2459 {
2460 struct smb2_hdr *shdr = (struct smb2_hdr *)buf;
2461
2462 if (shdr->Status == STATUS_IO_TIMEOUT)
2463 return true;
2464 else
2465 return false;
2466 }
2467
2468 static bool
smb2_is_network_name_deleted(char * buf,struct TCP_Server_Info * server)2469 smb2_is_network_name_deleted(char *buf, struct TCP_Server_Info *server)
2470 {
2471 struct smb2_hdr *shdr = (struct smb2_hdr *)buf;
2472 struct TCP_Server_Info *pserver;
2473 struct cifs_ses *ses;
2474 struct cifs_tcon *tcon;
2475
2476 if (shdr->Status != STATUS_NETWORK_NAME_DELETED)
2477 return false;
2478
2479 /* If server is a channel, select the primary channel */
2480 pserver = SERVER_IS_CHAN(server) ? server->primary_server : server;
2481
2482 spin_lock(&cifs_tcp_ses_lock);
2483 list_for_each_entry(ses, &pserver->smb_ses_list, smb_ses_list) {
2484 if (cifs_ses_exiting(ses))
2485 continue;
2486 list_for_each_entry(tcon, &ses->tcon_list, tcon_list) {
2487 if (tcon->tid == le32_to_cpu(shdr->Id.SyncId.TreeId)) {
2488 spin_lock(&tcon->tc_lock);
2489 tcon->need_reconnect = true;
2490 spin_unlock(&tcon->tc_lock);
2491 spin_unlock(&cifs_tcp_ses_lock);
2492 pr_warn_once("Server share %s deleted.\n",
2493 tcon->tree_name);
2494 return true;
2495 }
2496 }
2497 }
2498 spin_unlock(&cifs_tcp_ses_lock);
2499
2500 return false;
2501 }
2502
2503 static int
smb2_oplock_response(struct cifs_tcon * tcon,__u64 persistent_fid,__u64 volatile_fid,__u16 net_fid,struct cifsInodeInfo * cinode)2504 smb2_oplock_response(struct cifs_tcon *tcon, __u64 persistent_fid,
2505 __u64 volatile_fid, __u16 net_fid, struct cifsInodeInfo *cinode)
2506 {
2507 if (tcon->ses->server->capabilities & SMB2_GLOBAL_CAP_LEASING)
2508 return SMB2_lease_break(0, tcon, cinode->lease_key,
2509 smb2_get_lease_state(cinode));
2510
2511 return SMB2_oplock_break(0, tcon, persistent_fid, volatile_fid,
2512 CIFS_CACHE_READ(cinode) ? 1 : 0);
2513 }
2514
2515 void
smb2_set_replay(struct TCP_Server_Info * server,struct smb_rqst * rqst)2516 smb2_set_replay(struct TCP_Server_Info *server, struct smb_rqst *rqst)
2517 {
2518 struct smb2_hdr *shdr;
2519
2520 if (server->dialect < SMB30_PROT_ID)
2521 return;
2522
2523 shdr = (struct smb2_hdr *)(rqst->rq_iov[0].iov_base);
2524 if (shdr == NULL) {
2525 cifs_dbg(FYI, "shdr NULL in smb2_set_related\n");
2526 return;
2527 }
2528 shdr->Flags |= SMB2_FLAGS_REPLAY_OPERATION;
2529 }
2530
2531 void
smb2_set_related(struct smb_rqst * rqst)2532 smb2_set_related(struct smb_rqst *rqst)
2533 {
2534 struct smb2_hdr *shdr;
2535
2536 shdr = (struct smb2_hdr *)(rqst->rq_iov[0].iov_base);
2537 if (shdr == NULL) {
2538 cifs_dbg(FYI, "shdr NULL in smb2_set_related\n");
2539 return;
2540 }
2541 shdr->Flags |= SMB2_FLAGS_RELATED_OPERATIONS;
2542 }
2543
2544 char smb2_padding[7] = {0, 0, 0, 0, 0, 0, 0};
2545
2546 void
smb2_set_next_command(struct cifs_tcon * tcon,struct smb_rqst * rqst)2547 smb2_set_next_command(struct cifs_tcon *tcon, struct smb_rqst *rqst)
2548 {
2549 struct smb2_hdr *shdr;
2550 struct cifs_ses *ses = tcon->ses;
2551 struct TCP_Server_Info *server = ses->server;
2552 unsigned long len = smb_rqst_len(server, rqst);
2553 int i, num_padding;
2554
2555 shdr = (struct smb2_hdr *)(rqst->rq_iov[0].iov_base);
2556 if (shdr == NULL) {
2557 cifs_dbg(FYI, "shdr NULL in smb2_set_next_command\n");
2558 return;
2559 }
2560
2561 /* SMB headers in a compound are 8 byte aligned. */
2562
2563 /* No padding needed */
2564 if (!(len & 7))
2565 goto finished;
2566
2567 num_padding = 8 - (len & 7);
2568 if (!smb3_encryption_required(tcon)) {
2569 /*
2570 * If we do not have encryption then we can just add an extra
2571 * iov for the padding.
2572 */
2573 rqst->rq_iov[rqst->rq_nvec].iov_base = smb2_padding;
2574 rqst->rq_iov[rqst->rq_nvec].iov_len = num_padding;
2575 rqst->rq_nvec++;
2576 len += num_padding;
2577 } else {
2578 /*
2579 * We can not add a small padding iov for the encryption case
2580 * because the encryption framework can not handle the padding
2581 * iovs.
2582 * We have to flatten this into a single buffer and add
2583 * the padding to it.
2584 */
2585 for (i = 1; i < rqst->rq_nvec; i++) {
2586 memcpy(rqst->rq_iov[0].iov_base +
2587 rqst->rq_iov[0].iov_len,
2588 rqst->rq_iov[i].iov_base,
2589 rqst->rq_iov[i].iov_len);
2590 rqst->rq_iov[0].iov_len += rqst->rq_iov[i].iov_len;
2591 }
2592 memset(rqst->rq_iov[0].iov_base + rqst->rq_iov[0].iov_len,
2593 0, num_padding);
2594 rqst->rq_iov[0].iov_len += num_padding;
2595 len += num_padding;
2596 rqst->rq_nvec = 1;
2597 }
2598
2599 finished:
2600 shdr->NextCommand = cpu_to_le32(len);
2601 }
2602
2603 /*
2604 * helper function for exponential backoff and check if replayable
2605 */
smb2_should_replay(struct cifs_tcon * tcon,int * pretries,int * pcur_sleep)2606 bool smb2_should_replay(struct cifs_tcon *tcon,
2607 int *pretries,
2608 int *pcur_sleep)
2609 {
2610 if (!pretries || !pcur_sleep)
2611 return false;
2612
2613 if (tcon->retry || (*pretries)++ < tcon->ses->server->retrans) {
2614 msleep(*pcur_sleep);
2615 (*pcur_sleep) = ((*pcur_sleep) << 1);
2616 if ((*pcur_sleep) > CIFS_MAX_SLEEP)
2617 (*pcur_sleep) = CIFS_MAX_SLEEP;
2618 return true;
2619 }
2620
2621 return false;
2622 }
2623
2624 /*
2625 * Passes the query info response back to the caller on success.
2626 * Caller need to free this with free_rsp_buf().
2627 */
2628 int
smb2_query_info_compound(const unsigned int xid,struct cifs_tcon * tcon,const char * path,u32 desired_access,u32 class,u32 type,u32 output_len,struct kvec * rsp,int * buftype,struct cifs_sb_info * cifs_sb)2629 smb2_query_info_compound(const unsigned int xid, struct cifs_tcon *tcon,
2630 const char *path, u32 desired_access,
2631 u32 class, u32 type, u32 output_len,
2632 struct kvec *rsp, int *buftype,
2633 struct cifs_sb_info *cifs_sb)
2634 {
2635 struct smb2_compound_vars *vars;
2636 struct cifs_ses *ses = tcon->ses;
2637 struct TCP_Server_Info *server;
2638 int flags = CIFS_CP_CREATE_CLOSE_OP;
2639 struct smb_rqst *rqst;
2640 int resp_buftype[3];
2641 struct kvec *rsp_iov;
2642 u8 oplock = SMB2_OPLOCK_LEVEL_NONE;
2643 struct cifs_open_parms oparms;
2644 struct cifs_fid fid;
2645 int rc;
2646 __le16 *utf16_path;
2647 struct cached_fid *cfid = NULL;
2648 int retries = 0, cur_sleep = 1;
2649
2650 replay_again:
2651 /* reinitialize for possible replay */
2652 flags = CIFS_CP_CREATE_CLOSE_OP;
2653 oplock = SMB2_OPLOCK_LEVEL_NONE;
2654 server = cifs_pick_channel(ses);
2655
2656 if (!path)
2657 path = "";
2658 utf16_path = cifs_convert_path_to_utf16(path, cifs_sb);
2659 if (!utf16_path)
2660 return -ENOMEM;
2661
2662 if (smb3_encryption_required(tcon))
2663 flags |= CIFS_TRANSFORM_REQ;
2664
2665 resp_buftype[0] = resp_buftype[1] = resp_buftype[2] = CIFS_NO_BUFFER;
2666 vars = kzalloc(sizeof(*vars), GFP_KERNEL);
2667 if (!vars) {
2668 rc = -ENOMEM;
2669 goto out_free_path;
2670 }
2671 rqst = vars->rqst;
2672 rsp_iov = vars->rsp_iov;
2673
2674 /*
2675 * We can only call this for things we know are directories.
2676 */
2677 if (!strcmp(path, ""))
2678 open_cached_dir(xid, tcon, path, cifs_sb, false,
2679 &cfid); /* cfid null if open dir failed */
2680
2681 rqst[0].rq_iov = vars->open_iov;
2682 rqst[0].rq_nvec = SMB2_CREATE_IOV_SIZE;
2683
2684 oparms = (struct cifs_open_parms) {
2685 .tcon = tcon,
2686 .path = path,
2687 .desired_access = desired_access,
2688 .disposition = FILE_OPEN,
2689 .create_options = cifs_create_options(cifs_sb, 0),
2690 .fid = &fid,
2691 .replay = !!(retries),
2692 };
2693
2694 rc = SMB2_open_init(tcon, server,
2695 &rqst[0], &oplock, &oparms, utf16_path);
2696 if (rc)
2697 goto qic_exit;
2698 smb2_set_next_command(tcon, &rqst[0]);
2699
2700 rqst[1].rq_iov = &vars->qi_iov;
2701 rqst[1].rq_nvec = 1;
2702
2703 if (cfid) {
2704 rc = SMB2_query_info_init(tcon, server,
2705 &rqst[1],
2706 cfid->fid.persistent_fid,
2707 cfid->fid.volatile_fid,
2708 class, type, 0,
2709 output_len, 0,
2710 NULL);
2711 } else {
2712 rc = SMB2_query_info_init(tcon, server,
2713 &rqst[1],
2714 COMPOUND_FID,
2715 COMPOUND_FID,
2716 class, type, 0,
2717 output_len, 0,
2718 NULL);
2719 }
2720 if (rc)
2721 goto qic_exit;
2722 if (!cfid) {
2723 smb2_set_next_command(tcon, &rqst[1]);
2724 smb2_set_related(&rqst[1]);
2725 }
2726
2727 rqst[2].rq_iov = &vars->close_iov;
2728 rqst[2].rq_nvec = 1;
2729
2730 rc = SMB2_close_init(tcon, server,
2731 &rqst[2], COMPOUND_FID, COMPOUND_FID, false);
2732 if (rc)
2733 goto qic_exit;
2734 smb2_set_related(&rqst[2]);
2735
2736 if (retries) {
2737 if (!cfid) {
2738 smb2_set_replay(server, &rqst[0]);
2739 smb2_set_replay(server, &rqst[2]);
2740 }
2741 smb2_set_replay(server, &rqst[1]);
2742 }
2743
2744 if (cfid) {
2745 rc = compound_send_recv(xid, ses, server,
2746 flags, 1, &rqst[1],
2747 &resp_buftype[1], &rsp_iov[1]);
2748 } else {
2749 rc = compound_send_recv(xid, ses, server,
2750 flags, 3, rqst,
2751 resp_buftype, rsp_iov);
2752 }
2753 if (rc) {
2754 free_rsp_buf(resp_buftype[1], rsp_iov[1].iov_base);
2755 if (rc == -EREMCHG) {
2756 tcon->need_reconnect = true;
2757 pr_warn_once("server share %s deleted\n",
2758 tcon->tree_name);
2759 }
2760 goto qic_exit;
2761 }
2762 *rsp = rsp_iov[1];
2763 *buftype = resp_buftype[1];
2764
2765 qic_exit:
2766 SMB2_open_free(&rqst[0]);
2767 SMB2_query_info_free(&rqst[1]);
2768 SMB2_close_free(&rqst[2]);
2769 free_rsp_buf(resp_buftype[0], rsp_iov[0].iov_base);
2770 free_rsp_buf(resp_buftype[2], rsp_iov[2].iov_base);
2771 if (cfid)
2772 close_cached_dir(cfid);
2773 kfree(vars);
2774 out_free_path:
2775 kfree(utf16_path);
2776
2777 if (is_replayable_error(rc) &&
2778 smb2_should_replay(tcon, &retries, &cur_sleep))
2779 goto replay_again;
2780
2781 return rc;
2782 }
2783
2784 static int
smb2_queryfs(const unsigned int xid,struct cifs_tcon * tcon,const char * path,struct cifs_sb_info * cifs_sb,struct kstatfs * buf)2785 smb2_queryfs(const unsigned int xid, struct cifs_tcon *tcon,
2786 const char *path, struct cifs_sb_info *cifs_sb, struct kstatfs *buf)
2787 {
2788 struct smb2_query_info_rsp *rsp;
2789 struct smb2_fs_full_size_info *info = NULL;
2790 struct kvec rsp_iov = {NULL, 0};
2791 int buftype = CIFS_NO_BUFFER;
2792 int rc;
2793
2794
2795 rc = smb2_query_info_compound(xid, tcon, path,
2796 FILE_READ_ATTRIBUTES,
2797 FS_FULL_SIZE_INFORMATION,
2798 SMB2_O_INFO_FILESYSTEM,
2799 sizeof(struct smb2_fs_full_size_info),
2800 &rsp_iov, &buftype, cifs_sb);
2801 if (rc)
2802 goto qfs_exit;
2803
2804 rsp = (struct smb2_query_info_rsp *)rsp_iov.iov_base;
2805 buf->f_type = SMB2_SUPER_MAGIC;
2806 info = (struct smb2_fs_full_size_info *)(
2807 le16_to_cpu(rsp->OutputBufferOffset) + (char *)rsp);
2808 rc = smb2_validate_iov(le16_to_cpu(rsp->OutputBufferOffset),
2809 le32_to_cpu(rsp->OutputBufferLength),
2810 &rsp_iov,
2811 sizeof(struct smb2_fs_full_size_info));
2812 if (!rc)
2813 smb2_copy_fs_info_to_kstatfs(info, buf);
2814
2815 qfs_exit:
2816 trace_smb3_qfs_done(xid, tcon->tid, tcon->ses->Suid, tcon->tree_name, rc);
2817 free_rsp_buf(buftype, rsp_iov.iov_base);
2818 return rc;
2819 }
2820
2821 static int
smb311_queryfs(const unsigned int xid,struct cifs_tcon * tcon,const char * path,struct cifs_sb_info * cifs_sb,struct kstatfs * buf)2822 smb311_queryfs(const unsigned int xid, struct cifs_tcon *tcon,
2823 const char *path, struct cifs_sb_info *cifs_sb, struct kstatfs *buf)
2824 {
2825 int rc;
2826 __le16 *utf16_path = NULL;
2827 u8 oplock = SMB2_OPLOCK_LEVEL_NONE;
2828 struct cifs_open_parms oparms;
2829 struct cifs_fid fid;
2830
2831 if (!tcon->posix_extensions)
2832 return smb2_queryfs(xid, tcon, path, cifs_sb, buf);
2833
2834 oparms = (struct cifs_open_parms) {
2835 .tcon = tcon,
2836 .path = path,
2837 .desired_access = FILE_READ_ATTRIBUTES,
2838 .disposition = FILE_OPEN,
2839 .create_options = cifs_create_options(cifs_sb, 0),
2840 .fid = &fid,
2841 };
2842
2843 utf16_path = cifs_convert_path_to_utf16(path, cifs_sb);
2844 if (utf16_path == NULL)
2845 return -ENOMEM;
2846
2847 rc = SMB2_open(xid, &oparms, utf16_path, &oplock, NULL, NULL,
2848 NULL, NULL);
2849 kfree(utf16_path);
2850 if (rc)
2851 return rc;
2852
2853 rc = SMB311_posix_qfs_info(xid, tcon, fid.persistent_fid,
2854 fid.volatile_fid, buf);
2855 buf->f_type = SMB2_SUPER_MAGIC;
2856 SMB2_close(xid, tcon, fid.persistent_fid, fid.volatile_fid);
2857 return rc;
2858 }
2859
2860 static bool
smb2_compare_fids(struct cifsFileInfo * ob1,struct cifsFileInfo * ob2)2861 smb2_compare_fids(struct cifsFileInfo *ob1, struct cifsFileInfo *ob2)
2862 {
2863 return ob1->fid.persistent_fid == ob2->fid.persistent_fid &&
2864 ob1->fid.volatile_fid == ob2->fid.volatile_fid;
2865 }
2866
2867 static int
smb2_mand_lock(const unsigned int xid,struct cifsFileInfo * cfile,__u64 offset,__u64 length,__u32 type,int lock,int unlock,bool wait)2868 smb2_mand_lock(const unsigned int xid, struct cifsFileInfo *cfile, __u64 offset,
2869 __u64 length, __u32 type, int lock, int unlock, bool wait)
2870 {
2871 if (unlock && !lock)
2872 type = SMB2_LOCKFLAG_UNLOCK;
2873 return SMB2_lock(xid, tlink_tcon(cfile->tlink),
2874 cfile->fid.persistent_fid, cfile->fid.volatile_fid,
2875 current->tgid, length, offset, type, wait);
2876 }
2877
2878 static void
smb2_get_lease_key(struct inode * inode,struct cifs_fid * fid)2879 smb2_get_lease_key(struct inode *inode, struct cifs_fid *fid)
2880 {
2881 memcpy(fid->lease_key, CIFS_I(inode)->lease_key, SMB2_LEASE_KEY_SIZE);
2882 }
2883
2884 static void
smb2_set_lease_key(struct inode * inode,struct cifs_fid * fid)2885 smb2_set_lease_key(struct inode *inode, struct cifs_fid *fid)
2886 {
2887 memcpy(CIFS_I(inode)->lease_key, fid->lease_key, SMB2_LEASE_KEY_SIZE);
2888 }
2889
2890 static void
smb2_new_lease_key(struct cifs_fid * fid)2891 smb2_new_lease_key(struct cifs_fid *fid)
2892 {
2893 generate_random_uuid(fid->lease_key);
2894 }
2895
2896 static int
smb2_get_dfs_refer(const unsigned int xid,struct cifs_ses * ses,const char * search_name,struct dfs_info3_param ** target_nodes,unsigned int * num_of_nodes,const struct nls_table * nls_codepage,int remap)2897 smb2_get_dfs_refer(const unsigned int xid, struct cifs_ses *ses,
2898 const char *search_name,
2899 struct dfs_info3_param **target_nodes,
2900 unsigned int *num_of_nodes,
2901 const struct nls_table *nls_codepage, int remap)
2902 {
2903 int rc;
2904 __le16 *utf16_path = NULL;
2905 int utf16_path_len = 0;
2906 struct cifs_tcon *tcon;
2907 struct fsctl_get_dfs_referral_req *dfs_req = NULL;
2908 struct get_dfs_referral_rsp *dfs_rsp = NULL;
2909 u32 dfs_req_size = 0, dfs_rsp_size = 0;
2910 int retry_count = 0;
2911
2912 cifs_dbg(FYI, "%s: path: %s\n", __func__, search_name);
2913
2914 /*
2915 * Try to use the IPC tcon, otherwise just use any
2916 */
2917 tcon = ses->tcon_ipc;
2918 if (tcon == NULL) {
2919 spin_lock(&cifs_tcp_ses_lock);
2920 tcon = list_first_entry_or_null(&ses->tcon_list,
2921 struct cifs_tcon,
2922 tcon_list);
2923 if (tcon) {
2924 tcon->tc_count++;
2925 trace_smb3_tcon_ref(tcon->debug_id, tcon->tc_count,
2926 netfs_trace_tcon_ref_get_dfs_refer);
2927 }
2928 spin_unlock(&cifs_tcp_ses_lock);
2929 }
2930
2931 if (tcon == NULL) {
2932 cifs_dbg(VFS, "session %p has no tcon available for a dfs referral request\n",
2933 ses);
2934 rc = -ENOTCONN;
2935 goto out;
2936 }
2937
2938 utf16_path = cifs_strndup_to_utf16(search_name, PATH_MAX,
2939 &utf16_path_len,
2940 nls_codepage, remap);
2941 if (!utf16_path) {
2942 rc = -ENOMEM;
2943 goto out;
2944 }
2945
2946 dfs_req_size = sizeof(*dfs_req) + utf16_path_len;
2947 dfs_req = kzalloc(dfs_req_size, GFP_KERNEL);
2948 if (!dfs_req) {
2949 rc = -ENOMEM;
2950 goto out;
2951 }
2952
2953 /* Highest DFS referral version understood */
2954 dfs_req->MaxReferralLevel = DFS_VERSION;
2955
2956 /* Path to resolve in an UTF-16 null-terminated string */
2957 memcpy(dfs_req->RequestFileName, utf16_path, utf16_path_len);
2958
2959 do {
2960 rc = SMB2_ioctl(xid, tcon, NO_FILE_ID, NO_FILE_ID,
2961 FSCTL_DFS_GET_REFERRALS,
2962 (char *)dfs_req, dfs_req_size, CIFSMaxBufSize,
2963 (char **)&dfs_rsp, &dfs_rsp_size);
2964 if (!is_retryable_error(rc))
2965 break;
2966 usleep_range(512, 2048);
2967 } while (++retry_count < 5);
2968
2969 if (!rc && !dfs_rsp)
2970 rc = -EIO;
2971 if (rc) {
2972 if (!is_retryable_error(rc) && rc != -ENOENT && rc != -EOPNOTSUPP)
2973 cifs_tcon_dbg(VFS, "%s: ioctl error: rc=%d\n", __func__, rc);
2974 goto out;
2975 }
2976
2977 rc = parse_dfs_referrals(dfs_rsp, dfs_rsp_size,
2978 num_of_nodes, target_nodes,
2979 nls_codepage, remap, search_name,
2980 true /* is_unicode */);
2981 if (rc) {
2982 cifs_tcon_dbg(VFS, "parse error in %s rc=%d\n", __func__, rc);
2983 goto out;
2984 }
2985
2986 out:
2987 if (tcon && !tcon->ipc) {
2988 /* ipc tcons are not refcounted */
2989 spin_lock(&cifs_tcp_ses_lock);
2990 tcon->tc_count--;
2991 trace_smb3_tcon_ref(tcon->debug_id, tcon->tc_count,
2992 netfs_trace_tcon_ref_dec_dfs_refer);
2993 /* tc_count can never go negative */
2994 WARN_ON(tcon->tc_count < 0);
2995 spin_unlock(&cifs_tcp_ses_lock);
2996 }
2997 kfree(utf16_path);
2998 kfree(dfs_req);
2999 kfree(dfs_rsp);
3000 return rc;
3001 }
3002
3003 static struct cifs_ntsd *
get_smb2_acl_by_fid(struct cifs_sb_info * cifs_sb,const struct cifs_fid * cifsfid,u32 * pacllen,u32 info)3004 get_smb2_acl_by_fid(struct cifs_sb_info *cifs_sb,
3005 const struct cifs_fid *cifsfid, u32 *pacllen, u32 info)
3006 {
3007 struct cifs_ntsd *pntsd = NULL;
3008 unsigned int xid;
3009 int rc = -EOPNOTSUPP;
3010 struct tcon_link *tlink = cifs_sb_tlink(cifs_sb);
3011
3012 if (IS_ERR(tlink))
3013 return ERR_CAST(tlink);
3014
3015 xid = get_xid();
3016 cifs_dbg(FYI, "trying to get acl\n");
3017
3018 rc = SMB2_query_acl(xid, tlink_tcon(tlink), cifsfid->persistent_fid,
3019 cifsfid->volatile_fid, (void **)&pntsd, pacllen,
3020 info);
3021 free_xid(xid);
3022
3023 cifs_put_tlink(tlink);
3024
3025 cifs_dbg(FYI, "%s: rc = %d ACL len %d\n", __func__, rc, *pacllen);
3026 if (rc)
3027 return ERR_PTR(rc);
3028 return pntsd;
3029
3030 }
3031
3032 static struct cifs_ntsd *
get_smb2_acl_by_path(struct cifs_sb_info * cifs_sb,const char * path,u32 * pacllen,u32 info)3033 get_smb2_acl_by_path(struct cifs_sb_info *cifs_sb,
3034 const char *path, u32 *pacllen, u32 info)
3035 {
3036 struct cifs_ntsd *pntsd = NULL;
3037 u8 oplock = SMB2_OPLOCK_LEVEL_NONE;
3038 unsigned int xid;
3039 int rc;
3040 struct cifs_tcon *tcon;
3041 struct tcon_link *tlink = cifs_sb_tlink(cifs_sb);
3042 struct cifs_fid fid;
3043 struct cifs_open_parms oparms;
3044 __le16 *utf16_path;
3045
3046 cifs_dbg(FYI, "get smb3 acl for path %s\n", path);
3047 if (IS_ERR(tlink))
3048 return ERR_CAST(tlink);
3049
3050 tcon = tlink_tcon(tlink);
3051 xid = get_xid();
3052
3053 utf16_path = cifs_convert_path_to_utf16(path, cifs_sb);
3054 if (!utf16_path) {
3055 rc = -ENOMEM;
3056 free_xid(xid);
3057 return ERR_PTR(rc);
3058 }
3059
3060 oparms = (struct cifs_open_parms) {
3061 .tcon = tcon,
3062 .path = path,
3063 .desired_access = READ_CONTROL,
3064 .disposition = FILE_OPEN,
3065 /*
3066 * When querying an ACL, even if the file is a symlink
3067 * we want to open the source not the target, and so
3068 * the protocol requires that the client specify this
3069 * flag when opening a reparse point
3070 */
3071 .create_options = cifs_create_options(cifs_sb, 0) |
3072 OPEN_REPARSE_POINT,
3073 .fid = &fid,
3074 };
3075
3076 if (info & SACL_SECINFO)
3077 oparms.desired_access |= SYSTEM_SECURITY;
3078
3079 rc = SMB2_open(xid, &oparms, utf16_path, &oplock, NULL, NULL, NULL,
3080 NULL);
3081 kfree(utf16_path);
3082 if (!rc) {
3083 rc = SMB2_query_acl(xid, tlink_tcon(tlink), fid.persistent_fid,
3084 fid.volatile_fid, (void **)&pntsd, pacllen,
3085 info);
3086 SMB2_close(xid, tcon, fid.persistent_fid, fid.volatile_fid);
3087 }
3088
3089 cifs_put_tlink(tlink);
3090 free_xid(xid);
3091
3092 cifs_dbg(FYI, "%s: rc = %d ACL len %d\n", __func__, rc, *pacllen);
3093 if (rc)
3094 return ERR_PTR(rc);
3095 return pntsd;
3096 }
3097
3098 static int
set_smb2_acl(struct cifs_ntsd * pnntsd,__u32 acllen,struct inode * inode,const char * path,int aclflag)3099 set_smb2_acl(struct cifs_ntsd *pnntsd, __u32 acllen,
3100 struct inode *inode, const char *path, int aclflag)
3101 {
3102 u8 oplock = SMB2_OPLOCK_LEVEL_NONE;
3103 unsigned int xid;
3104 int rc, access_flags = 0;
3105 struct cifs_tcon *tcon;
3106 struct cifs_sb_info *cifs_sb = CIFS_SB(inode->i_sb);
3107 struct tcon_link *tlink = cifs_sb_tlink(cifs_sb);
3108 struct cifs_fid fid;
3109 struct cifs_open_parms oparms;
3110 __le16 *utf16_path;
3111
3112 cifs_dbg(FYI, "set smb3 acl for path %s\n", path);
3113 if (IS_ERR(tlink))
3114 return PTR_ERR(tlink);
3115
3116 tcon = tlink_tcon(tlink);
3117 xid = get_xid();
3118
3119 if (aclflag & CIFS_ACL_OWNER || aclflag & CIFS_ACL_GROUP)
3120 access_flags |= WRITE_OWNER;
3121 if (aclflag & CIFS_ACL_SACL)
3122 access_flags |= SYSTEM_SECURITY;
3123 if (aclflag & CIFS_ACL_DACL)
3124 access_flags |= WRITE_DAC;
3125
3126 utf16_path = cifs_convert_path_to_utf16(path, cifs_sb);
3127 if (!utf16_path) {
3128 rc = -ENOMEM;
3129 free_xid(xid);
3130 return rc;
3131 }
3132
3133 oparms = (struct cifs_open_parms) {
3134 .tcon = tcon,
3135 .desired_access = access_flags,
3136 .create_options = cifs_create_options(cifs_sb, 0),
3137 .disposition = FILE_OPEN,
3138 .path = path,
3139 .fid = &fid,
3140 };
3141
3142 rc = SMB2_open(xid, &oparms, utf16_path, &oplock, NULL, NULL,
3143 NULL, NULL);
3144 kfree(utf16_path);
3145 if (!rc) {
3146 rc = SMB2_set_acl(xid, tlink_tcon(tlink), fid.persistent_fid,
3147 fid.volatile_fid, pnntsd, acllen, aclflag);
3148 SMB2_close(xid, tcon, fid.persistent_fid, fid.volatile_fid);
3149 }
3150
3151 cifs_put_tlink(tlink);
3152 free_xid(xid);
3153 return rc;
3154 }
3155
3156 /* Retrieve an ACL from the server */
3157 static struct cifs_ntsd *
get_smb2_acl(struct cifs_sb_info * cifs_sb,struct inode * inode,const char * path,u32 * pacllen,u32 info)3158 get_smb2_acl(struct cifs_sb_info *cifs_sb,
3159 struct inode *inode, const char *path,
3160 u32 *pacllen, u32 info)
3161 {
3162 struct cifs_ntsd *pntsd = NULL;
3163 struct cifsFileInfo *open_file = NULL;
3164
3165 if (inode && !(info & SACL_SECINFO))
3166 open_file = find_readable_file(CIFS_I(inode), true);
3167 if (!open_file || (info & SACL_SECINFO))
3168 return get_smb2_acl_by_path(cifs_sb, path, pacllen, info);
3169
3170 pntsd = get_smb2_acl_by_fid(cifs_sb, &open_file->fid, pacllen, info);
3171 cifsFileInfo_put(open_file);
3172 return pntsd;
3173 }
3174
smb3_zero_data(struct file * file,struct cifs_tcon * tcon,loff_t offset,loff_t len,unsigned int xid)3175 static long smb3_zero_data(struct file *file, struct cifs_tcon *tcon,
3176 loff_t offset, loff_t len, unsigned int xid)
3177 {
3178 struct cifsFileInfo *cfile = file->private_data;
3179 struct file_zero_data_information fsctl_buf;
3180
3181 cifs_dbg(FYI, "Offset %lld len %lld\n", offset, len);
3182
3183 fsctl_buf.FileOffset = cpu_to_le64(offset);
3184 fsctl_buf.BeyondFinalZero = cpu_to_le64(offset + len);
3185
3186 return SMB2_ioctl(xid, tcon, cfile->fid.persistent_fid,
3187 cfile->fid.volatile_fid, FSCTL_SET_ZERO_DATA,
3188 (char *)&fsctl_buf,
3189 sizeof(struct file_zero_data_information),
3190 0, NULL, NULL);
3191 }
3192
smb3_zero_range(struct file * file,struct cifs_tcon * tcon,unsigned long long offset,unsigned long long len,bool keep_size)3193 static long smb3_zero_range(struct file *file, struct cifs_tcon *tcon,
3194 unsigned long long offset, unsigned long long len,
3195 bool keep_size)
3196 {
3197 struct cifs_ses *ses = tcon->ses;
3198 struct inode *inode = file_inode(file);
3199 struct cifsInodeInfo *cifsi = CIFS_I(inode);
3200 struct cifsFileInfo *cfile = file->private_data;
3201 struct netfs_inode *ictx = netfs_inode(inode);
3202 unsigned long long i_size, new_size, remote_size;
3203 long rc;
3204 unsigned int xid;
3205
3206 xid = get_xid();
3207
3208 trace_smb3_zero_enter(xid, cfile->fid.persistent_fid, tcon->tid,
3209 ses->Suid, offset, len);
3210
3211 inode_lock(inode);
3212 filemap_invalidate_lock(inode->i_mapping);
3213
3214 i_size = i_size_read(inode);
3215 remote_size = ictx->remote_i_size;
3216 if (offset + len >= remote_size && offset < i_size) {
3217 unsigned long long top = umin(offset + len, i_size);
3218
3219 rc = filemap_write_and_wait_range(inode->i_mapping, offset, top - 1);
3220 if (rc < 0)
3221 goto zero_range_exit;
3222 }
3223
3224 /*
3225 * We zero the range through ioctl, so we need remove the page caches
3226 * first, otherwise the data may be inconsistent with the server.
3227 */
3228 truncate_pagecache_range(inode, offset, offset + len - 1);
3229
3230 /* if file not oplocked can't be sure whether asking to extend size */
3231 rc = -EOPNOTSUPP;
3232 if (keep_size == false && !CIFS_CACHE_READ(cifsi))
3233 goto zero_range_exit;
3234
3235 rc = smb3_zero_data(file, tcon, offset, len, xid);
3236 if (rc < 0)
3237 goto zero_range_exit;
3238
3239 /*
3240 * do we also need to change the size of the file?
3241 */
3242 new_size = offset + len;
3243 if (keep_size == false && (unsigned long long)i_size_read(inode) < new_size) {
3244 rc = SMB2_set_eof(xid, tcon, cfile->fid.persistent_fid,
3245 cfile->fid.volatile_fid, cfile->pid, new_size);
3246 if (rc >= 0) {
3247 truncate_setsize(inode, new_size);
3248 fscache_resize_cookie(cifs_inode_cookie(inode), new_size);
3249 }
3250 }
3251
3252 zero_range_exit:
3253 filemap_invalidate_unlock(inode->i_mapping);
3254 inode_unlock(inode);
3255 free_xid(xid);
3256 if (rc)
3257 trace_smb3_zero_err(xid, cfile->fid.persistent_fid, tcon->tid,
3258 ses->Suid, offset, len, rc);
3259 else
3260 trace_smb3_zero_done(xid, cfile->fid.persistent_fid, tcon->tid,
3261 ses->Suid, offset, len);
3262 return rc;
3263 }
3264
smb3_punch_hole(struct file * file,struct cifs_tcon * tcon,loff_t offset,loff_t len)3265 static long smb3_punch_hole(struct file *file, struct cifs_tcon *tcon,
3266 loff_t offset, loff_t len)
3267 {
3268 struct inode *inode = file_inode(file);
3269 struct cifsFileInfo *cfile = file->private_data;
3270 struct file_zero_data_information fsctl_buf;
3271 unsigned long long end = offset + len, i_size, remote_i_size;
3272 long rc;
3273 unsigned int xid;
3274 __u8 set_sparse = 1;
3275
3276 xid = get_xid();
3277
3278 inode_lock(inode);
3279 /* Need to make file sparse, if not already, before freeing range. */
3280 /* Consider adding equivalent for compressed since it could also work */
3281 if (!smb2_set_sparse(xid, tcon, cfile, inode, set_sparse)) {
3282 rc = -EOPNOTSUPP;
3283 goto out;
3284 }
3285
3286 filemap_invalidate_lock(inode->i_mapping);
3287 /*
3288 * We implement the punch hole through ioctl, so we need remove the page
3289 * caches first, otherwise the data may be inconsistent with the server.
3290 */
3291 truncate_pagecache_range(inode, offset, offset + len - 1);
3292
3293 cifs_dbg(FYI, "Offset %lld len %lld\n", offset, len);
3294
3295 fsctl_buf.FileOffset = cpu_to_le64(offset);
3296 fsctl_buf.BeyondFinalZero = cpu_to_le64(offset + len);
3297
3298 rc = SMB2_ioctl(xid, tcon, cfile->fid.persistent_fid,
3299 cfile->fid.volatile_fid, FSCTL_SET_ZERO_DATA,
3300 (char *)&fsctl_buf,
3301 sizeof(struct file_zero_data_information),
3302 CIFSMaxBufSize, NULL, NULL);
3303
3304 if (rc)
3305 goto unlock;
3306
3307 /* If there's dirty data in the buffer that would extend the EOF if it
3308 * were written, then we need to move the EOF marker over to the lower
3309 * of the high end of the hole and the proposed EOF. The problem is
3310 * that we locally hole-punch the tail of the dirty data, the proposed
3311 * EOF update will end up in the wrong place.
3312 */
3313 i_size = i_size_read(inode);
3314 remote_i_size = netfs_inode(inode)->remote_i_size;
3315 if (end > remote_i_size && i_size > remote_i_size) {
3316 unsigned long long extend_to = umin(end, i_size);
3317 rc = SMB2_set_eof(xid, tcon, cfile->fid.persistent_fid,
3318 cfile->fid.volatile_fid, cfile->pid, extend_to);
3319 if (rc >= 0)
3320 netfs_inode(inode)->remote_i_size = extend_to;
3321 }
3322
3323 unlock:
3324 filemap_invalidate_unlock(inode->i_mapping);
3325 out:
3326 inode_unlock(inode);
3327 free_xid(xid);
3328 return rc;
3329 }
3330
smb3_simple_fallocate_write_range(unsigned int xid,struct cifs_tcon * tcon,struct cifsFileInfo * cfile,loff_t off,loff_t len,char * buf)3331 static int smb3_simple_fallocate_write_range(unsigned int xid,
3332 struct cifs_tcon *tcon,
3333 struct cifsFileInfo *cfile,
3334 loff_t off, loff_t len,
3335 char *buf)
3336 {
3337 struct cifs_io_parms io_parms = {0};
3338 int nbytes;
3339 int rc = 0;
3340 struct kvec iov[2];
3341
3342 io_parms.netfid = cfile->fid.netfid;
3343 io_parms.pid = current->tgid;
3344 io_parms.tcon = tcon;
3345 io_parms.persistent_fid = cfile->fid.persistent_fid;
3346 io_parms.volatile_fid = cfile->fid.volatile_fid;
3347
3348 while (len) {
3349 io_parms.offset = off;
3350 io_parms.length = len;
3351 if (io_parms.length > SMB2_MAX_BUFFER_SIZE)
3352 io_parms.length = SMB2_MAX_BUFFER_SIZE;
3353 /* iov[0] is reserved for smb header */
3354 iov[1].iov_base = buf;
3355 iov[1].iov_len = io_parms.length;
3356 rc = SMB2_write(xid, &io_parms, &nbytes, iov, 1);
3357 if (rc)
3358 break;
3359 if (nbytes > len)
3360 return -EINVAL;
3361 buf += nbytes;
3362 off += nbytes;
3363 len -= nbytes;
3364 }
3365 return rc;
3366 }
3367
smb3_simple_fallocate_range(unsigned int xid,struct cifs_tcon * tcon,struct cifsFileInfo * cfile,loff_t off,loff_t len)3368 static int smb3_simple_fallocate_range(unsigned int xid,
3369 struct cifs_tcon *tcon,
3370 struct cifsFileInfo *cfile,
3371 loff_t off, loff_t len)
3372 {
3373 struct file_allocated_range_buffer in_data, *out_data = NULL, *tmp_data;
3374 u32 out_data_len;
3375 char *buf = NULL;
3376 loff_t l;
3377 int rc;
3378
3379 in_data.file_offset = cpu_to_le64(off);
3380 in_data.length = cpu_to_le64(len);
3381 rc = SMB2_ioctl(xid, tcon, cfile->fid.persistent_fid,
3382 cfile->fid.volatile_fid,
3383 FSCTL_QUERY_ALLOCATED_RANGES,
3384 (char *)&in_data, sizeof(in_data),
3385 1024 * sizeof(struct file_allocated_range_buffer),
3386 (char **)&out_data, &out_data_len);
3387 if (rc)
3388 goto out;
3389
3390 buf = kzalloc(1024 * 1024, GFP_KERNEL);
3391 if (buf == NULL) {
3392 rc = -ENOMEM;
3393 goto out;
3394 }
3395
3396 tmp_data = out_data;
3397 while (len) {
3398 /*
3399 * The rest of the region is unmapped so write it all.
3400 */
3401 if (out_data_len == 0) {
3402 rc = smb3_simple_fallocate_write_range(xid, tcon,
3403 cfile, off, len, buf);
3404 goto out;
3405 }
3406
3407 if (out_data_len < sizeof(struct file_allocated_range_buffer)) {
3408 rc = -EINVAL;
3409 goto out;
3410 }
3411
3412 if (off < le64_to_cpu(tmp_data->file_offset)) {
3413 /*
3414 * We are at a hole. Write until the end of the region
3415 * or until the next allocated data,
3416 * whichever comes next.
3417 */
3418 l = le64_to_cpu(tmp_data->file_offset) - off;
3419 if (len < l)
3420 l = len;
3421 rc = smb3_simple_fallocate_write_range(xid, tcon,
3422 cfile, off, l, buf);
3423 if (rc)
3424 goto out;
3425 off = off + l;
3426 len = len - l;
3427 if (len == 0)
3428 goto out;
3429 }
3430 /*
3431 * We are at a section of allocated data, just skip forward
3432 * until the end of the data or the end of the region
3433 * we are supposed to fallocate, whichever comes first.
3434 */
3435 l = le64_to_cpu(tmp_data->length);
3436 if (len < l)
3437 l = len;
3438 off += l;
3439 len -= l;
3440
3441 tmp_data = &tmp_data[1];
3442 out_data_len -= sizeof(struct file_allocated_range_buffer);
3443 }
3444
3445 out:
3446 kfree(out_data);
3447 kfree(buf);
3448 return rc;
3449 }
3450
3451
smb3_simple_falloc(struct file * file,struct cifs_tcon * tcon,loff_t off,loff_t len,bool keep_size)3452 static long smb3_simple_falloc(struct file *file, struct cifs_tcon *tcon,
3453 loff_t off, loff_t len, bool keep_size)
3454 {
3455 struct inode *inode;
3456 struct cifsInodeInfo *cifsi;
3457 struct cifsFileInfo *cfile = file->private_data;
3458 long rc = -EOPNOTSUPP;
3459 unsigned int xid;
3460 loff_t new_eof;
3461
3462 xid = get_xid();
3463
3464 inode = d_inode(cfile->dentry);
3465 cifsi = CIFS_I(inode);
3466
3467 trace_smb3_falloc_enter(xid, cfile->fid.persistent_fid, tcon->tid,
3468 tcon->ses->Suid, off, len);
3469 /* if file not oplocked can't be sure whether asking to extend size */
3470 if (!CIFS_CACHE_READ(cifsi))
3471 if (keep_size == false) {
3472 trace_smb3_falloc_err(xid, cfile->fid.persistent_fid,
3473 tcon->tid, tcon->ses->Suid, off, len, rc);
3474 free_xid(xid);
3475 return rc;
3476 }
3477
3478 /*
3479 * Extending the file
3480 */
3481 if ((keep_size == false) && i_size_read(inode) < off + len) {
3482 rc = inode_newsize_ok(inode, off + len);
3483 if (rc)
3484 goto out;
3485
3486 if (cifsi->cifsAttrs & FILE_ATTRIBUTE_SPARSE_FILE)
3487 smb2_set_sparse(xid, tcon, cfile, inode, false);
3488
3489 new_eof = off + len;
3490 rc = SMB2_set_eof(xid, tcon, cfile->fid.persistent_fid,
3491 cfile->fid.volatile_fid, cfile->pid, new_eof);
3492 if (rc == 0) {
3493 cifsi->server_eof = new_eof;
3494 cifs_setsize(inode, new_eof);
3495 cifs_truncate_page(inode->i_mapping, inode->i_size);
3496 truncate_setsize(inode, new_eof);
3497 }
3498 goto out;
3499 }
3500
3501 /*
3502 * Files are non-sparse by default so falloc may be a no-op
3503 * Must check if file sparse. If not sparse, and since we are not
3504 * extending then no need to do anything since file already allocated
3505 */
3506 if ((cifsi->cifsAttrs & FILE_ATTRIBUTE_SPARSE_FILE) == 0) {
3507 rc = 0;
3508 goto out;
3509 }
3510
3511 if (keep_size == true) {
3512 /*
3513 * We can not preallocate pages beyond the end of the file
3514 * in SMB2
3515 */
3516 if (off >= i_size_read(inode)) {
3517 rc = 0;
3518 goto out;
3519 }
3520 /*
3521 * For fallocates that are partially beyond the end of file,
3522 * clamp len so we only fallocate up to the end of file.
3523 */
3524 if (off + len > i_size_read(inode)) {
3525 len = i_size_read(inode) - off;
3526 }
3527 }
3528
3529 if ((keep_size == true) || (i_size_read(inode) >= off + len)) {
3530 /*
3531 * At this point, we are trying to fallocate an internal
3532 * regions of a sparse file. Since smb2 does not have a
3533 * fallocate command we have two otions on how to emulate this.
3534 * We can either turn the entire file to become non-sparse
3535 * which we only do if the fallocate is for virtually
3536 * the whole file, or we can overwrite the region with zeroes
3537 * using SMB2_write, which could be prohibitevly expensive
3538 * if len is large.
3539 */
3540 /*
3541 * We are only trying to fallocate a small region so
3542 * just write it with zero.
3543 */
3544 if (len <= 1024 * 1024) {
3545 rc = smb3_simple_fallocate_range(xid, tcon, cfile,
3546 off, len);
3547 goto out;
3548 }
3549
3550 /*
3551 * Check if falloc starts within first few pages of file
3552 * and ends within a few pages of the end of file to
3553 * ensure that most of file is being forced to be
3554 * fallocated now. If so then setting whole file sparse
3555 * ie potentially making a few extra pages at the beginning
3556 * or end of the file non-sparse via set_sparse is harmless.
3557 */
3558 if ((off > 8192) || (off + len + 8192 < i_size_read(inode))) {
3559 rc = -EOPNOTSUPP;
3560 goto out;
3561 }
3562 }
3563
3564 smb2_set_sparse(xid, tcon, cfile, inode, false);
3565 rc = 0;
3566
3567 out:
3568 if (rc)
3569 trace_smb3_falloc_err(xid, cfile->fid.persistent_fid, tcon->tid,
3570 tcon->ses->Suid, off, len, rc);
3571 else
3572 trace_smb3_falloc_done(xid, cfile->fid.persistent_fid, tcon->tid,
3573 tcon->ses->Suid, off, len);
3574
3575 free_xid(xid);
3576 return rc;
3577 }
3578
smb3_collapse_range(struct file * file,struct cifs_tcon * tcon,loff_t off,loff_t len)3579 static long smb3_collapse_range(struct file *file, struct cifs_tcon *tcon,
3580 loff_t off, loff_t len)
3581 {
3582 int rc;
3583 unsigned int xid;
3584 struct inode *inode = file_inode(file);
3585 struct cifsFileInfo *cfile = file->private_data;
3586 struct cifsInodeInfo *cifsi = CIFS_I(inode);
3587 loff_t old_eof, new_eof;
3588
3589 xid = get_xid();
3590
3591 inode_lock(inode);
3592
3593 old_eof = i_size_read(inode);
3594 if ((off >= old_eof) ||
3595 off + len >= old_eof) {
3596 rc = -EINVAL;
3597 goto out;
3598 }
3599
3600 filemap_invalidate_lock(inode->i_mapping);
3601 rc = filemap_write_and_wait_range(inode->i_mapping, off, old_eof - 1);
3602 if (rc < 0)
3603 goto out_2;
3604
3605 truncate_pagecache_range(inode, off, old_eof);
3606
3607 rc = smb2_copychunk_range(xid, cfile, cfile, off + len,
3608 old_eof - off - len, off);
3609 if (rc < 0)
3610 goto out_2;
3611
3612 new_eof = old_eof - len;
3613 rc = SMB2_set_eof(xid, tcon, cfile->fid.persistent_fid,
3614 cfile->fid.volatile_fid, cfile->pid, new_eof);
3615 if (rc < 0)
3616 goto out_2;
3617
3618 rc = 0;
3619
3620 cifsi->server_eof = i_size_read(inode) - len;
3621 truncate_setsize(inode, cifsi->server_eof);
3622 fscache_resize_cookie(cifs_inode_cookie(inode), cifsi->server_eof);
3623 out_2:
3624 filemap_invalidate_unlock(inode->i_mapping);
3625 out:
3626 inode_unlock(inode);
3627 free_xid(xid);
3628 return rc;
3629 }
3630
smb3_insert_range(struct file * file,struct cifs_tcon * tcon,loff_t off,loff_t len)3631 static long smb3_insert_range(struct file *file, struct cifs_tcon *tcon,
3632 loff_t off, loff_t len)
3633 {
3634 int rc;
3635 unsigned int xid;
3636 struct cifsFileInfo *cfile = file->private_data;
3637 struct inode *inode = file_inode(file);
3638 __u64 count, old_eof, new_eof;
3639
3640 xid = get_xid();
3641
3642 inode_lock(inode);
3643
3644 old_eof = i_size_read(inode);
3645 if (off >= old_eof) {
3646 rc = -EINVAL;
3647 goto out;
3648 }
3649
3650 count = old_eof - off;
3651 new_eof = old_eof + len;
3652
3653 filemap_invalidate_lock(inode->i_mapping);
3654 rc = filemap_write_and_wait_range(inode->i_mapping, off, new_eof - 1);
3655 if (rc < 0)
3656 goto out_2;
3657 truncate_pagecache_range(inode, off, old_eof);
3658
3659 rc = SMB2_set_eof(xid, tcon, cfile->fid.persistent_fid,
3660 cfile->fid.volatile_fid, cfile->pid, new_eof);
3661 if (rc < 0)
3662 goto out_2;
3663
3664 truncate_setsize(inode, new_eof);
3665 fscache_resize_cookie(cifs_inode_cookie(inode), i_size_read(inode));
3666
3667 rc = smb2_copychunk_range(xid, cfile, cfile, off, count, off + len);
3668 if (rc < 0)
3669 goto out_2;
3670
3671 rc = smb3_zero_data(file, tcon, off, len, xid);
3672 if (rc < 0)
3673 goto out_2;
3674
3675 rc = 0;
3676 out_2:
3677 filemap_invalidate_unlock(inode->i_mapping);
3678 out:
3679 inode_unlock(inode);
3680 free_xid(xid);
3681 return rc;
3682 }
3683
smb3_llseek(struct file * file,struct cifs_tcon * tcon,loff_t offset,int whence)3684 static loff_t smb3_llseek(struct file *file, struct cifs_tcon *tcon, loff_t offset, int whence)
3685 {
3686 struct cifsFileInfo *wrcfile, *cfile = file->private_data;
3687 struct cifsInodeInfo *cifsi;
3688 struct inode *inode;
3689 int rc = 0;
3690 struct file_allocated_range_buffer in_data, *out_data = NULL;
3691 u32 out_data_len;
3692 unsigned int xid;
3693
3694 if (whence != SEEK_HOLE && whence != SEEK_DATA)
3695 return generic_file_llseek(file, offset, whence);
3696
3697 inode = d_inode(cfile->dentry);
3698 cifsi = CIFS_I(inode);
3699
3700 if (offset < 0 || offset >= i_size_read(inode))
3701 return -ENXIO;
3702
3703 xid = get_xid();
3704 /*
3705 * We need to be sure that all dirty pages are written as they
3706 * might fill holes on the server.
3707 * Note that we also MUST flush any written pages since at least
3708 * some servers (Windows2016) will not reflect recent writes in
3709 * QUERY_ALLOCATED_RANGES until SMB2_flush is called.
3710 */
3711 wrcfile = find_writable_file(cifsi, FIND_WR_ANY);
3712 if (wrcfile) {
3713 filemap_write_and_wait(inode->i_mapping);
3714 smb2_flush_file(xid, tcon, &wrcfile->fid);
3715 cifsFileInfo_put(wrcfile);
3716 }
3717
3718 if (!(cifsi->cifsAttrs & FILE_ATTRIBUTE_SPARSE_FILE)) {
3719 if (whence == SEEK_HOLE)
3720 offset = i_size_read(inode);
3721 goto lseek_exit;
3722 }
3723
3724 in_data.file_offset = cpu_to_le64(offset);
3725 in_data.length = cpu_to_le64(i_size_read(inode));
3726
3727 rc = SMB2_ioctl(xid, tcon, cfile->fid.persistent_fid,
3728 cfile->fid.volatile_fid,
3729 FSCTL_QUERY_ALLOCATED_RANGES,
3730 (char *)&in_data, sizeof(in_data),
3731 sizeof(struct file_allocated_range_buffer),
3732 (char **)&out_data, &out_data_len);
3733 if (rc == -E2BIG)
3734 rc = 0;
3735 if (rc)
3736 goto lseek_exit;
3737
3738 if (whence == SEEK_HOLE && out_data_len == 0)
3739 goto lseek_exit;
3740
3741 if (whence == SEEK_DATA && out_data_len == 0) {
3742 rc = -ENXIO;
3743 goto lseek_exit;
3744 }
3745
3746 if (out_data_len < sizeof(struct file_allocated_range_buffer)) {
3747 rc = -EINVAL;
3748 goto lseek_exit;
3749 }
3750 if (whence == SEEK_DATA) {
3751 offset = le64_to_cpu(out_data->file_offset);
3752 goto lseek_exit;
3753 }
3754 if (offset < le64_to_cpu(out_data->file_offset))
3755 goto lseek_exit;
3756
3757 offset = le64_to_cpu(out_data->file_offset) + le64_to_cpu(out_data->length);
3758
3759 lseek_exit:
3760 free_xid(xid);
3761 kfree(out_data);
3762 if (!rc)
3763 return vfs_setpos(file, offset, inode->i_sb->s_maxbytes);
3764 else
3765 return rc;
3766 }
3767
smb3_fiemap(struct cifs_tcon * tcon,struct cifsFileInfo * cfile,struct fiemap_extent_info * fei,u64 start,u64 len)3768 static int smb3_fiemap(struct cifs_tcon *tcon,
3769 struct cifsFileInfo *cfile,
3770 struct fiemap_extent_info *fei, u64 start, u64 len)
3771 {
3772 unsigned int xid;
3773 struct file_allocated_range_buffer in_data, *out_data;
3774 u32 out_data_len;
3775 int i, num, rc, flags, last_blob;
3776 u64 next;
3777
3778 rc = fiemap_prep(d_inode(cfile->dentry), fei, start, &len, 0);
3779 if (rc)
3780 return rc;
3781
3782 xid = get_xid();
3783 again:
3784 in_data.file_offset = cpu_to_le64(start);
3785 in_data.length = cpu_to_le64(len);
3786
3787 rc = SMB2_ioctl(xid, tcon, cfile->fid.persistent_fid,
3788 cfile->fid.volatile_fid,
3789 FSCTL_QUERY_ALLOCATED_RANGES,
3790 (char *)&in_data, sizeof(in_data),
3791 1024 * sizeof(struct file_allocated_range_buffer),
3792 (char **)&out_data, &out_data_len);
3793 if (rc == -E2BIG) {
3794 last_blob = 0;
3795 rc = 0;
3796 } else
3797 last_blob = 1;
3798 if (rc)
3799 goto out;
3800
3801 if (out_data_len && out_data_len < sizeof(struct file_allocated_range_buffer)) {
3802 rc = -EINVAL;
3803 goto out;
3804 }
3805 if (out_data_len % sizeof(struct file_allocated_range_buffer)) {
3806 rc = -EINVAL;
3807 goto out;
3808 }
3809
3810 num = out_data_len / sizeof(struct file_allocated_range_buffer);
3811 for (i = 0; i < num; i++) {
3812 flags = 0;
3813 if (i == num - 1 && last_blob)
3814 flags |= FIEMAP_EXTENT_LAST;
3815
3816 rc = fiemap_fill_next_extent(fei,
3817 le64_to_cpu(out_data[i].file_offset),
3818 le64_to_cpu(out_data[i].file_offset),
3819 le64_to_cpu(out_data[i].length),
3820 flags);
3821 if (rc < 0)
3822 goto out;
3823 if (rc == 1) {
3824 rc = 0;
3825 goto out;
3826 }
3827 }
3828
3829 if (!last_blob) {
3830 next = le64_to_cpu(out_data[num - 1].file_offset) +
3831 le64_to_cpu(out_data[num - 1].length);
3832 len = len - (next - start);
3833 start = next;
3834 goto again;
3835 }
3836
3837 out:
3838 free_xid(xid);
3839 kfree(out_data);
3840 return rc;
3841 }
3842
smb3_fallocate(struct file * file,struct cifs_tcon * tcon,int mode,loff_t off,loff_t len)3843 static long smb3_fallocate(struct file *file, struct cifs_tcon *tcon, int mode,
3844 loff_t off, loff_t len)
3845 {
3846 /* KEEP_SIZE already checked for by do_fallocate */
3847 if (mode & FALLOC_FL_PUNCH_HOLE)
3848 return smb3_punch_hole(file, tcon, off, len);
3849 else if (mode & FALLOC_FL_ZERO_RANGE) {
3850 if (mode & FALLOC_FL_KEEP_SIZE)
3851 return smb3_zero_range(file, tcon, off, len, true);
3852 return smb3_zero_range(file, tcon, off, len, false);
3853 } else if (mode == FALLOC_FL_KEEP_SIZE)
3854 return smb3_simple_falloc(file, tcon, off, len, true);
3855 else if (mode == FALLOC_FL_COLLAPSE_RANGE)
3856 return smb3_collapse_range(file, tcon, off, len);
3857 else if (mode == FALLOC_FL_INSERT_RANGE)
3858 return smb3_insert_range(file, tcon, off, len);
3859 else if (mode == 0)
3860 return smb3_simple_falloc(file, tcon, off, len, false);
3861
3862 return -EOPNOTSUPP;
3863 }
3864
3865 static void
smb2_downgrade_oplock(struct TCP_Server_Info * server,struct cifsInodeInfo * cinode,__u32 oplock,unsigned int epoch,bool * purge_cache)3866 smb2_downgrade_oplock(struct TCP_Server_Info *server,
3867 struct cifsInodeInfo *cinode, __u32 oplock,
3868 unsigned int epoch, bool *purge_cache)
3869 {
3870 server->ops->set_oplock_level(cinode, oplock, 0, NULL);
3871 }
3872
3873 static void
3874 smb21_set_oplock_level(struct cifsInodeInfo *cinode, __u32 oplock,
3875 unsigned int epoch, bool *purge_cache);
3876
3877 static void
smb3_downgrade_oplock(struct TCP_Server_Info * server,struct cifsInodeInfo * cinode,__u32 oplock,unsigned int epoch,bool * purge_cache)3878 smb3_downgrade_oplock(struct TCP_Server_Info *server,
3879 struct cifsInodeInfo *cinode, __u32 oplock,
3880 unsigned int epoch, bool *purge_cache)
3881 {
3882 unsigned int old_state = cinode->oplock;
3883 unsigned int old_epoch = cinode->epoch;
3884 unsigned int new_state;
3885
3886 if (epoch > old_epoch) {
3887 smb21_set_oplock_level(cinode, oplock, 0, NULL);
3888 cinode->epoch = epoch;
3889 }
3890
3891 new_state = cinode->oplock;
3892 *purge_cache = false;
3893
3894 if ((old_state & CIFS_CACHE_READ_FLG) != 0 &&
3895 (new_state & CIFS_CACHE_READ_FLG) == 0)
3896 *purge_cache = true;
3897 else if (old_state == new_state && (epoch - old_epoch > 1))
3898 *purge_cache = true;
3899 }
3900
3901 static void
smb2_set_oplock_level(struct cifsInodeInfo * cinode,__u32 oplock,unsigned int epoch,bool * purge_cache)3902 smb2_set_oplock_level(struct cifsInodeInfo *cinode, __u32 oplock,
3903 unsigned int epoch, bool *purge_cache)
3904 {
3905 oplock &= 0xFF;
3906 cinode->lease_granted = false;
3907 if (oplock == SMB2_OPLOCK_LEVEL_NOCHANGE)
3908 return;
3909 if (oplock == SMB2_OPLOCK_LEVEL_BATCH) {
3910 cinode->oplock = CIFS_CACHE_RHW_FLG;
3911 cifs_dbg(FYI, "Batch Oplock granted on inode %p\n",
3912 &cinode->netfs.inode);
3913 } else if (oplock == SMB2_OPLOCK_LEVEL_EXCLUSIVE) {
3914 cinode->oplock = CIFS_CACHE_RW_FLG;
3915 cifs_dbg(FYI, "Exclusive Oplock granted on inode %p\n",
3916 &cinode->netfs.inode);
3917 } else if (oplock == SMB2_OPLOCK_LEVEL_II) {
3918 cinode->oplock = CIFS_CACHE_READ_FLG;
3919 cifs_dbg(FYI, "Level II Oplock granted on inode %p\n",
3920 &cinode->netfs.inode);
3921 } else
3922 cinode->oplock = 0;
3923 }
3924
3925 static void
smb21_set_oplock_level(struct cifsInodeInfo * cinode,__u32 oplock,unsigned int epoch,bool * purge_cache)3926 smb21_set_oplock_level(struct cifsInodeInfo *cinode, __u32 oplock,
3927 unsigned int epoch, bool *purge_cache)
3928 {
3929 char message[5] = {0};
3930 unsigned int new_oplock = 0;
3931
3932 oplock &= 0xFF;
3933 cinode->lease_granted = true;
3934 if (oplock == SMB2_OPLOCK_LEVEL_NOCHANGE)
3935 return;
3936
3937 /* Check if the server granted an oplock rather than a lease */
3938 if (oplock & SMB2_OPLOCK_LEVEL_EXCLUSIVE)
3939 return smb2_set_oplock_level(cinode, oplock, epoch,
3940 purge_cache);
3941
3942 if (oplock & SMB2_LEASE_READ_CACHING_HE) {
3943 new_oplock |= CIFS_CACHE_READ_FLG;
3944 strcat(message, "R");
3945 }
3946 if (oplock & SMB2_LEASE_HANDLE_CACHING_HE) {
3947 new_oplock |= CIFS_CACHE_HANDLE_FLG;
3948 strcat(message, "H");
3949 }
3950 if (oplock & SMB2_LEASE_WRITE_CACHING_HE) {
3951 new_oplock |= CIFS_CACHE_WRITE_FLG;
3952 strcat(message, "W");
3953 }
3954 if (!new_oplock)
3955 strncpy(message, "None", sizeof(message));
3956
3957 cinode->oplock = new_oplock;
3958 cifs_dbg(FYI, "%s Lease granted on inode %p\n", message,
3959 &cinode->netfs.inode);
3960 }
3961
3962 static void
smb3_set_oplock_level(struct cifsInodeInfo * cinode,__u32 oplock,unsigned int epoch,bool * purge_cache)3963 smb3_set_oplock_level(struct cifsInodeInfo *cinode, __u32 oplock,
3964 unsigned int epoch, bool *purge_cache)
3965 {
3966 unsigned int old_oplock = cinode->oplock;
3967
3968 smb21_set_oplock_level(cinode, oplock, epoch, purge_cache);
3969
3970 if (purge_cache) {
3971 *purge_cache = false;
3972 if (old_oplock == CIFS_CACHE_READ_FLG) {
3973 if (cinode->oplock == CIFS_CACHE_READ_FLG &&
3974 (epoch - cinode->epoch > 0))
3975 *purge_cache = true;
3976 else if (cinode->oplock == CIFS_CACHE_RH_FLG &&
3977 (epoch - cinode->epoch > 1))
3978 *purge_cache = true;
3979 else if (cinode->oplock == CIFS_CACHE_RHW_FLG &&
3980 (epoch - cinode->epoch > 1))
3981 *purge_cache = true;
3982 else if (cinode->oplock == 0 &&
3983 (epoch - cinode->epoch > 0))
3984 *purge_cache = true;
3985 } else if (old_oplock == CIFS_CACHE_RH_FLG) {
3986 if (cinode->oplock == CIFS_CACHE_RH_FLG &&
3987 (epoch - cinode->epoch > 0))
3988 *purge_cache = true;
3989 else if (cinode->oplock == CIFS_CACHE_RHW_FLG &&
3990 (epoch - cinode->epoch > 1))
3991 *purge_cache = true;
3992 }
3993 cinode->epoch = epoch;
3994 }
3995 }
3996
3997 #ifdef CONFIG_CIFS_ALLOW_INSECURE_LEGACY
3998 static bool
smb2_is_read_op(__u32 oplock)3999 smb2_is_read_op(__u32 oplock)
4000 {
4001 return oplock == SMB2_OPLOCK_LEVEL_II;
4002 }
4003 #endif /* CIFS_ALLOW_INSECURE_LEGACY */
4004
4005 static bool
smb21_is_read_op(__u32 oplock)4006 smb21_is_read_op(__u32 oplock)
4007 {
4008 return (oplock & SMB2_LEASE_READ_CACHING_HE) &&
4009 !(oplock & SMB2_LEASE_WRITE_CACHING_HE);
4010 }
4011
4012 static __le32
map_oplock_to_lease(u8 oplock)4013 map_oplock_to_lease(u8 oplock)
4014 {
4015 if (oplock == SMB2_OPLOCK_LEVEL_EXCLUSIVE)
4016 return SMB2_LEASE_WRITE_CACHING_LE | SMB2_LEASE_READ_CACHING_LE;
4017 else if (oplock == SMB2_OPLOCK_LEVEL_II)
4018 return SMB2_LEASE_READ_CACHING_LE;
4019 else if (oplock == SMB2_OPLOCK_LEVEL_BATCH)
4020 return SMB2_LEASE_HANDLE_CACHING_LE | SMB2_LEASE_READ_CACHING_LE |
4021 SMB2_LEASE_WRITE_CACHING_LE;
4022 return 0;
4023 }
4024
4025 static char *
smb2_create_lease_buf(u8 * lease_key,u8 oplock)4026 smb2_create_lease_buf(u8 *lease_key, u8 oplock)
4027 {
4028 struct create_lease *buf;
4029
4030 buf = kzalloc(sizeof(struct create_lease), GFP_KERNEL);
4031 if (!buf)
4032 return NULL;
4033
4034 memcpy(&buf->lcontext.LeaseKey, lease_key, SMB2_LEASE_KEY_SIZE);
4035 buf->lcontext.LeaseState = map_oplock_to_lease(oplock);
4036
4037 buf->ccontext.DataOffset = cpu_to_le16(offsetof
4038 (struct create_lease, lcontext));
4039 buf->ccontext.DataLength = cpu_to_le32(sizeof(struct lease_context));
4040 buf->ccontext.NameOffset = cpu_to_le16(offsetof
4041 (struct create_lease, Name));
4042 buf->ccontext.NameLength = cpu_to_le16(4);
4043 /* SMB2_CREATE_REQUEST_LEASE is "RqLs" */
4044 buf->Name[0] = 'R';
4045 buf->Name[1] = 'q';
4046 buf->Name[2] = 'L';
4047 buf->Name[3] = 's';
4048 return (char *)buf;
4049 }
4050
4051 static char *
smb3_create_lease_buf(u8 * lease_key,u8 oplock)4052 smb3_create_lease_buf(u8 *lease_key, u8 oplock)
4053 {
4054 struct create_lease_v2 *buf;
4055
4056 buf = kzalloc(sizeof(struct create_lease_v2), GFP_KERNEL);
4057 if (!buf)
4058 return NULL;
4059
4060 memcpy(&buf->lcontext.LeaseKey, lease_key, SMB2_LEASE_KEY_SIZE);
4061 buf->lcontext.LeaseState = map_oplock_to_lease(oplock);
4062
4063 buf->ccontext.DataOffset = cpu_to_le16(offsetof
4064 (struct create_lease_v2, lcontext));
4065 buf->ccontext.DataLength = cpu_to_le32(sizeof(struct lease_context_v2));
4066 buf->ccontext.NameOffset = cpu_to_le16(offsetof
4067 (struct create_lease_v2, Name));
4068 buf->ccontext.NameLength = cpu_to_le16(4);
4069 /* SMB2_CREATE_REQUEST_LEASE is "RqLs" */
4070 buf->Name[0] = 'R';
4071 buf->Name[1] = 'q';
4072 buf->Name[2] = 'L';
4073 buf->Name[3] = 's';
4074 return (char *)buf;
4075 }
4076
4077 static __u8
smb2_parse_lease_buf(void * buf,unsigned int * epoch,char * lease_key)4078 smb2_parse_lease_buf(void *buf, unsigned int *epoch, char *lease_key)
4079 {
4080 struct create_lease *lc = (struct create_lease *)buf;
4081
4082 *epoch = 0; /* not used */
4083 if (lc->lcontext.LeaseFlags & SMB2_LEASE_FLAG_BREAK_IN_PROGRESS_LE)
4084 return SMB2_OPLOCK_LEVEL_NOCHANGE;
4085 return le32_to_cpu(lc->lcontext.LeaseState);
4086 }
4087
4088 static __u8
smb3_parse_lease_buf(void * buf,unsigned int * epoch,char * lease_key)4089 smb3_parse_lease_buf(void *buf, unsigned int *epoch, char *lease_key)
4090 {
4091 struct create_lease_v2 *lc = (struct create_lease_v2 *)buf;
4092
4093 *epoch = le16_to_cpu(lc->lcontext.Epoch);
4094 if (lc->lcontext.LeaseFlags & SMB2_LEASE_FLAG_BREAK_IN_PROGRESS_LE)
4095 return SMB2_OPLOCK_LEVEL_NOCHANGE;
4096 if (lease_key)
4097 memcpy(lease_key, &lc->lcontext.LeaseKey, SMB2_LEASE_KEY_SIZE);
4098 return le32_to_cpu(lc->lcontext.LeaseState);
4099 }
4100
4101 static unsigned int
smb2_wp_retry_size(struct inode * inode)4102 smb2_wp_retry_size(struct inode *inode)
4103 {
4104 return min_t(unsigned int, CIFS_SB(inode->i_sb)->ctx->wsize,
4105 SMB2_MAX_BUFFER_SIZE);
4106 }
4107
4108 static bool
smb2_dir_needs_close(struct cifsFileInfo * cfile)4109 smb2_dir_needs_close(struct cifsFileInfo *cfile)
4110 {
4111 return !cfile->invalidHandle;
4112 }
4113
4114 static void
fill_transform_hdr(struct smb2_transform_hdr * tr_hdr,unsigned int orig_len,struct smb_rqst * old_rq,__le16 cipher_type)4115 fill_transform_hdr(struct smb2_transform_hdr *tr_hdr, unsigned int orig_len,
4116 struct smb_rqst *old_rq, __le16 cipher_type)
4117 {
4118 struct smb2_hdr *shdr =
4119 (struct smb2_hdr *)old_rq->rq_iov[0].iov_base;
4120
4121 memset(tr_hdr, 0, sizeof(struct smb2_transform_hdr));
4122 tr_hdr->ProtocolId = SMB2_TRANSFORM_PROTO_NUM;
4123 tr_hdr->OriginalMessageSize = cpu_to_le32(orig_len);
4124 tr_hdr->Flags = cpu_to_le16(0x01);
4125 if ((cipher_type == SMB2_ENCRYPTION_AES128_GCM) ||
4126 (cipher_type == SMB2_ENCRYPTION_AES256_GCM))
4127 get_random_bytes(&tr_hdr->Nonce, SMB3_AES_GCM_NONCE);
4128 else
4129 get_random_bytes(&tr_hdr->Nonce, SMB3_AES_CCM_NONCE);
4130 memcpy(&tr_hdr->SessionId, &shdr->SessionId, 8);
4131 }
4132
smb2_aead_req_alloc(struct crypto_aead * tfm,const struct smb_rqst * rqst,int num_rqst,const u8 * sig,u8 ** iv,struct aead_request ** req,struct sg_table * sgt,unsigned int * num_sgs,size_t * sensitive_size)4133 static void *smb2_aead_req_alloc(struct crypto_aead *tfm, const struct smb_rqst *rqst,
4134 int num_rqst, const u8 *sig, u8 **iv,
4135 struct aead_request **req, struct sg_table *sgt,
4136 unsigned int *num_sgs, size_t *sensitive_size)
4137 {
4138 unsigned int req_size = sizeof(**req) + crypto_aead_reqsize(tfm);
4139 unsigned int iv_size = crypto_aead_ivsize(tfm);
4140 unsigned int len;
4141 u8 *p;
4142
4143 *num_sgs = cifs_get_num_sgs(rqst, num_rqst, sig);
4144 if (IS_ERR_VALUE((long)(int)*num_sgs))
4145 return ERR_PTR(*num_sgs);
4146
4147 len = iv_size;
4148 len += crypto_aead_alignmask(tfm) & ~(crypto_tfm_ctx_alignment() - 1);
4149 len = ALIGN(len, crypto_tfm_ctx_alignment());
4150 len += req_size;
4151 len = ALIGN(len, __alignof__(struct scatterlist));
4152 len += array_size(*num_sgs, sizeof(struct scatterlist));
4153 *sensitive_size = len;
4154
4155 p = kvzalloc(len, GFP_NOFS);
4156 if (!p)
4157 return ERR_PTR(-ENOMEM);
4158
4159 *iv = (u8 *)PTR_ALIGN(p, crypto_aead_alignmask(tfm) + 1);
4160 *req = (struct aead_request *)PTR_ALIGN(*iv + iv_size,
4161 crypto_tfm_ctx_alignment());
4162 sgt->sgl = (struct scatterlist *)PTR_ALIGN((u8 *)*req + req_size,
4163 __alignof__(struct scatterlist));
4164 return p;
4165 }
4166
smb2_get_aead_req(struct crypto_aead * tfm,struct smb_rqst * rqst,int num_rqst,const u8 * sig,u8 ** iv,struct aead_request ** req,struct scatterlist ** sgl,size_t * sensitive_size)4167 static void *smb2_get_aead_req(struct crypto_aead *tfm, struct smb_rqst *rqst,
4168 int num_rqst, const u8 *sig, u8 **iv,
4169 struct aead_request **req, struct scatterlist **sgl,
4170 size_t *sensitive_size)
4171 {
4172 struct sg_table sgtable = {};
4173 unsigned int skip, num_sgs, i, j;
4174 ssize_t rc;
4175 void *p;
4176
4177 p = smb2_aead_req_alloc(tfm, rqst, num_rqst, sig, iv, req, &sgtable,
4178 &num_sgs, sensitive_size);
4179 if (IS_ERR(p))
4180 return ERR_CAST(p);
4181
4182 sg_init_marker(sgtable.sgl, num_sgs);
4183
4184 /*
4185 * The first rqst has a transform header where the
4186 * first 20 bytes are not part of the encrypted blob.
4187 */
4188 skip = 20;
4189
4190 for (i = 0; i < num_rqst; i++) {
4191 struct iov_iter *iter = &rqst[i].rq_iter;
4192 size_t count = iov_iter_count(iter);
4193
4194 for (j = 0; j < rqst[i].rq_nvec; j++) {
4195 cifs_sg_set_buf(&sgtable,
4196 rqst[i].rq_iov[j].iov_base + skip,
4197 rqst[i].rq_iov[j].iov_len - skip);
4198
4199 /* See the above comment on the 'skip' assignment */
4200 skip = 0;
4201 }
4202 sgtable.orig_nents = sgtable.nents;
4203
4204 rc = extract_iter_to_sg(iter, count, &sgtable,
4205 num_sgs - sgtable.nents, 0);
4206 iov_iter_revert(iter, rc);
4207 sgtable.orig_nents = sgtable.nents;
4208 }
4209
4210 cifs_sg_set_buf(&sgtable, sig, SMB2_SIGNATURE_SIZE);
4211 sg_mark_end(&sgtable.sgl[sgtable.nents - 1]);
4212 *sgl = sgtable.sgl;
4213 return p;
4214 }
4215
4216 static int
smb2_get_enc_key(struct TCP_Server_Info * server,__u64 ses_id,int enc,u8 * key)4217 smb2_get_enc_key(struct TCP_Server_Info *server, __u64 ses_id, int enc, u8 *key)
4218 {
4219 struct TCP_Server_Info *pserver;
4220 struct cifs_ses *ses;
4221 u8 *ses_enc_key;
4222
4223 /* If server is a channel, select the primary channel */
4224 pserver = SERVER_IS_CHAN(server) ? server->primary_server : server;
4225
4226 spin_lock(&cifs_tcp_ses_lock);
4227 list_for_each_entry(ses, &pserver->smb_ses_list, smb_ses_list) {
4228 if (ses->Suid == ses_id) {
4229 spin_lock(&ses->ses_lock);
4230 ses_enc_key = enc ? ses->smb3encryptionkey :
4231 ses->smb3decryptionkey;
4232 memcpy(key, ses_enc_key, SMB3_ENC_DEC_KEY_SIZE);
4233 spin_unlock(&ses->ses_lock);
4234 spin_unlock(&cifs_tcp_ses_lock);
4235 return 0;
4236 }
4237 }
4238 spin_unlock(&cifs_tcp_ses_lock);
4239
4240 trace_smb3_ses_not_found(ses_id);
4241
4242 return -EAGAIN;
4243 }
4244 /*
4245 * Encrypt or decrypt @rqst message. @rqst[0] has the following format:
4246 * iov[0] - transform header (associate data),
4247 * iov[1-N] - SMB2 header and pages - data to encrypt.
4248 * On success return encrypted data in iov[1-N] and pages, leave iov[0]
4249 * untouched.
4250 */
4251 static int
crypt_message(struct TCP_Server_Info * server,int num_rqst,struct smb_rqst * rqst,int enc,struct crypto_aead * tfm)4252 crypt_message(struct TCP_Server_Info *server, int num_rqst,
4253 struct smb_rqst *rqst, int enc, struct crypto_aead *tfm)
4254 {
4255 struct smb2_transform_hdr *tr_hdr =
4256 (struct smb2_transform_hdr *)rqst[0].rq_iov[0].iov_base;
4257 unsigned int assoc_data_len = sizeof(struct smb2_transform_hdr) - 20;
4258 int rc = 0;
4259 struct scatterlist *sg;
4260 u8 sign[SMB2_SIGNATURE_SIZE] = {};
4261 u8 key[SMB3_ENC_DEC_KEY_SIZE];
4262 struct aead_request *req;
4263 u8 *iv;
4264 unsigned int crypt_len = le32_to_cpu(tr_hdr->OriginalMessageSize);
4265 void *creq;
4266 size_t sensitive_size;
4267
4268 rc = smb2_get_enc_key(server, le64_to_cpu(tr_hdr->SessionId), enc, key);
4269 if (rc) {
4270 cifs_server_dbg(FYI, "%s: Could not get %scryption key. sid: 0x%llx\n", __func__,
4271 enc ? "en" : "de", le64_to_cpu(tr_hdr->SessionId));
4272 return rc;
4273 }
4274
4275 if ((server->cipher_type == SMB2_ENCRYPTION_AES256_CCM) ||
4276 (server->cipher_type == SMB2_ENCRYPTION_AES256_GCM))
4277 rc = crypto_aead_setkey(tfm, key, SMB3_GCM256_CRYPTKEY_SIZE);
4278 else
4279 rc = crypto_aead_setkey(tfm, key, SMB3_GCM128_CRYPTKEY_SIZE);
4280
4281 if (rc) {
4282 cifs_server_dbg(VFS, "%s: Failed to set aead key %d\n", __func__, rc);
4283 return rc;
4284 }
4285
4286 rc = crypto_aead_setauthsize(tfm, SMB2_SIGNATURE_SIZE);
4287 if (rc) {
4288 cifs_server_dbg(VFS, "%s: Failed to set authsize %d\n", __func__, rc);
4289 return rc;
4290 }
4291
4292 creq = smb2_get_aead_req(tfm, rqst, num_rqst, sign, &iv, &req, &sg,
4293 &sensitive_size);
4294 if (IS_ERR(creq))
4295 return PTR_ERR(creq);
4296
4297 if (!enc) {
4298 memcpy(sign, &tr_hdr->Signature, SMB2_SIGNATURE_SIZE);
4299 crypt_len += SMB2_SIGNATURE_SIZE;
4300 }
4301
4302 if ((server->cipher_type == SMB2_ENCRYPTION_AES128_GCM) ||
4303 (server->cipher_type == SMB2_ENCRYPTION_AES256_GCM))
4304 memcpy(iv, (char *)tr_hdr->Nonce, SMB3_AES_GCM_NONCE);
4305 else {
4306 iv[0] = 3;
4307 memcpy(iv + 1, (char *)tr_hdr->Nonce, SMB3_AES_CCM_NONCE);
4308 }
4309
4310 aead_request_set_tfm(req, tfm);
4311 aead_request_set_crypt(req, sg, sg, crypt_len, iv);
4312 aead_request_set_ad(req, assoc_data_len);
4313
4314 rc = enc ? crypto_aead_encrypt(req) : crypto_aead_decrypt(req);
4315
4316 if (!rc && enc)
4317 memcpy(&tr_hdr->Signature, sign, SMB2_SIGNATURE_SIZE);
4318
4319 kvfree_sensitive(creq, sensitive_size);
4320 return rc;
4321 }
4322
4323 /*
4324 * Clear a read buffer, discarding the folios which have XA_MARK_0 set.
4325 */
cifs_clear_xarray_buffer(struct xarray * buffer)4326 static void cifs_clear_xarray_buffer(struct xarray *buffer)
4327 {
4328 struct folio *folio;
4329
4330 XA_STATE(xas, buffer, 0);
4331
4332 rcu_read_lock();
4333 xas_for_each_marked(&xas, folio, ULONG_MAX, XA_MARK_0) {
4334 folio_put(folio);
4335 }
4336 rcu_read_unlock();
4337 xa_destroy(buffer);
4338 }
4339
4340 void
smb3_free_compound_rqst(int num_rqst,struct smb_rqst * rqst)4341 smb3_free_compound_rqst(int num_rqst, struct smb_rqst *rqst)
4342 {
4343 int i;
4344
4345 for (i = 0; i < num_rqst; i++)
4346 if (!xa_empty(&rqst[i].rq_buffer))
4347 cifs_clear_xarray_buffer(&rqst[i].rq_buffer);
4348 }
4349
4350 /*
4351 * This function will initialize new_rq and encrypt the content.
4352 * The first entry, new_rq[0], only contains a single iov which contains
4353 * a smb2_transform_hdr and is pre-allocated by the caller.
4354 * This function then populates new_rq[1+] with the content from olq_rq[0+].
4355 *
4356 * The end result is an array of smb_rqst structures where the first structure
4357 * only contains a single iov for the transform header which we then can pass
4358 * to crypt_message().
4359 *
4360 * new_rq[0].rq_iov[0] : smb2_transform_hdr pre-allocated by the caller
4361 * new_rq[1+].rq_iov[*] == old_rq[0+].rq_iov[*] : SMB2/3 requests
4362 */
4363 static int
smb3_init_transform_rq(struct TCP_Server_Info * server,int num_rqst,struct smb_rqst * new_rq,struct smb_rqst * old_rq)4364 smb3_init_transform_rq(struct TCP_Server_Info *server, int num_rqst,
4365 struct smb_rqst *new_rq, struct smb_rqst *old_rq)
4366 {
4367 struct smb2_transform_hdr *tr_hdr = new_rq[0].rq_iov[0].iov_base;
4368 struct page *page;
4369 unsigned int orig_len = 0;
4370 int i, j;
4371 int rc = -ENOMEM;
4372
4373 for (i = 1; i < num_rqst; i++) {
4374 struct smb_rqst *old = &old_rq[i - 1];
4375 struct smb_rqst *new = &new_rq[i];
4376 struct xarray *buffer = &new->rq_buffer;
4377 size_t size = iov_iter_count(&old->rq_iter), seg, copied = 0;
4378
4379 orig_len += smb_rqst_len(server, old);
4380 new->rq_iov = old->rq_iov;
4381 new->rq_nvec = old->rq_nvec;
4382
4383 xa_init(buffer);
4384
4385 if (size > 0) {
4386 unsigned int npages = DIV_ROUND_UP(size, PAGE_SIZE);
4387
4388 for (j = 0; j < npages; j++) {
4389 void *o;
4390
4391 rc = -ENOMEM;
4392 page = alloc_page(GFP_KERNEL|__GFP_HIGHMEM);
4393 if (!page)
4394 goto err_free;
4395 page->index = j;
4396 o = xa_store(buffer, j, page, GFP_KERNEL);
4397 if (xa_is_err(o)) {
4398 rc = xa_err(o);
4399 put_page(page);
4400 goto err_free;
4401 }
4402
4403 xa_set_mark(buffer, j, XA_MARK_0);
4404
4405 seg = min_t(size_t, size - copied, PAGE_SIZE);
4406 if (copy_page_from_iter(page, 0, seg, &old->rq_iter) != seg) {
4407 rc = -EFAULT;
4408 goto err_free;
4409 }
4410 copied += seg;
4411 }
4412 iov_iter_xarray(&new->rq_iter, ITER_SOURCE,
4413 buffer, 0, size);
4414 new->rq_iter_size = size;
4415 }
4416 }
4417
4418 /* fill the 1st iov with a transform header */
4419 fill_transform_hdr(tr_hdr, orig_len, old_rq, server->cipher_type);
4420
4421 rc = crypt_message(server, num_rqst, new_rq, 1, server->secmech.enc);
4422 cifs_dbg(FYI, "Encrypt message returned %d\n", rc);
4423 if (rc)
4424 goto err_free;
4425
4426 return rc;
4427
4428 err_free:
4429 smb3_free_compound_rqst(num_rqst - 1, &new_rq[1]);
4430 return rc;
4431 }
4432
4433 static int
smb3_is_transform_hdr(void * buf)4434 smb3_is_transform_hdr(void *buf)
4435 {
4436 struct smb2_transform_hdr *trhdr = buf;
4437
4438 return trhdr->ProtocolId == SMB2_TRANSFORM_PROTO_NUM;
4439 }
4440
4441 static int
decrypt_raw_data(struct TCP_Server_Info * server,char * buf,unsigned int buf_data_size,struct iov_iter * iter,bool is_offloaded)4442 decrypt_raw_data(struct TCP_Server_Info *server, char *buf,
4443 unsigned int buf_data_size, struct iov_iter *iter,
4444 bool is_offloaded)
4445 {
4446 struct crypto_aead *tfm;
4447 struct smb_rqst rqst = {NULL};
4448 struct kvec iov[2];
4449 size_t iter_size = 0;
4450 int rc;
4451
4452 iov[0].iov_base = buf;
4453 iov[0].iov_len = sizeof(struct smb2_transform_hdr);
4454 iov[1].iov_base = buf + sizeof(struct smb2_transform_hdr);
4455 iov[1].iov_len = buf_data_size;
4456
4457 rqst.rq_iov = iov;
4458 rqst.rq_nvec = 2;
4459 if (iter) {
4460 rqst.rq_iter = *iter;
4461 rqst.rq_iter_size = iov_iter_count(iter);
4462 iter_size = iov_iter_count(iter);
4463 }
4464
4465 if (is_offloaded) {
4466 if ((server->cipher_type == SMB2_ENCRYPTION_AES128_GCM) ||
4467 (server->cipher_type == SMB2_ENCRYPTION_AES256_GCM))
4468 tfm = crypto_alloc_aead("gcm(aes)", 0, 0);
4469 else
4470 tfm = crypto_alloc_aead("ccm(aes)", 0, 0);
4471 if (IS_ERR(tfm)) {
4472 rc = PTR_ERR(tfm);
4473 cifs_server_dbg(VFS, "%s: Failed alloc decrypt TFM, rc=%d\n", __func__, rc);
4474
4475 return rc;
4476 }
4477 } else {
4478 if (unlikely(!server->secmech.dec))
4479 return -EIO;
4480
4481 tfm = server->secmech.dec;
4482 }
4483
4484 rc = crypt_message(server, 1, &rqst, 0, tfm);
4485 cifs_dbg(FYI, "Decrypt message returned %d\n", rc);
4486
4487 if (is_offloaded)
4488 crypto_free_aead(tfm);
4489
4490 if (rc)
4491 return rc;
4492
4493 memmove(buf, iov[1].iov_base, buf_data_size);
4494
4495 if (!is_offloaded)
4496 server->total_read = buf_data_size + iter_size;
4497
4498 return rc;
4499 }
4500
4501 static int
cifs_copy_pages_to_iter(struct xarray * pages,unsigned int data_size,unsigned int skip,struct iov_iter * iter)4502 cifs_copy_pages_to_iter(struct xarray *pages, unsigned int data_size,
4503 unsigned int skip, struct iov_iter *iter)
4504 {
4505 struct page *page;
4506 unsigned long index;
4507
4508 xa_for_each(pages, index, page) {
4509 size_t n, len = min_t(unsigned int, PAGE_SIZE - skip, data_size);
4510
4511 n = copy_page_to_iter(page, skip, len, iter);
4512 if (n != len) {
4513 cifs_dbg(VFS, "%s: something went wrong\n", __func__);
4514 return -EIO;
4515 }
4516 data_size -= n;
4517 skip = 0;
4518 }
4519
4520 return 0;
4521 }
4522
4523 static int
handle_read_data(struct TCP_Server_Info * server,struct mid_q_entry * mid,char * buf,unsigned int buf_len,struct xarray * pages,unsigned int pages_len,bool is_offloaded)4524 handle_read_data(struct TCP_Server_Info *server, struct mid_q_entry *mid,
4525 char *buf, unsigned int buf_len, struct xarray *pages,
4526 unsigned int pages_len, bool is_offloaded)
4527 {
4528 unsigned int data_offset;
4529 unsigned int data_len;
4530 unsigned int cur_off;
4531 unsigned int cur_page_idx;
4532 unsigned int pad_len;
4533 struct cifs_readdata *rdata = mid->callback_data;
4534 struct smb2_hdr *shdr = (struct smb2_hdr *)buf;
4535 int length;
4536 bool use_rdma_mr = false;
4537
4538 if (shdr->Command != SMB2_READ) {
4539 cifs_server_dbg(VFS, "only big read responses are supported\n");
4540 return -EOPNOTSUPP;
4541 }
4542
4543 if (server->ops->is_session_expired &&
4544 server->ops->is_session_expired(buf)) {
4545 if (!is_offloaded)
4546 cifs_reconnect(server, true);
4547 return -1;
4548 }
4549
4550 if (server->ops->is_status_pending &&
4551 server->ops->is_status_pending(buf, server))
4552 return -1;
4553
4554 /* set up first two iov to get credits */
4555 rdata->iov[0].iov_base = buf;
4556 rdata->iov[0].iov_len = 0;
4557 rdata->iov[1].iov_base = buf;
4558 rdata->iov[1].iov_len =
4559 min_t(unsigned int, buf_len, server->vals->read_rsp_size);
4560 cifs_dbg(FYI, "0: iov_base=%p iov_len=%zu\n",
4561 rdata->iov[0].iov_base, rdata->iov[0].iov_len);
4562 cifs_dbg(FYI, "1: iov_base=%p iov_len=%zu\n",
4563 rdata->iov[1].iov_base, rdata->iov[1].iov_len);
4564
4565 rdata->result = server->ops->map_error(buf, true);
4566 if (rdata->result != 0) {
4567 cifs_dbg(FYI, "%s: server returned error %d\n",
4568 __func__, rdata->result);
4569 /* normal error on read response */
4570 if (is_offloaded)
4571 mid->mid_state = MID_RESPONSE_RECEIVED;
4572 else
4573 dequeue_mid(mid, false);
4574 return 0;
4575 }
4576
4577 data_offset = server->ops->read_data_offset(buf);
4578 #ifdef CONFIG_CIFS_SMB_DIRECT
4579 use_rdma_mr = rdata->mr;
4580 #endif
4581 data_len = server->ops->read_data_length(buf, use_rdma_mr);
4582
4583 if (data_offset < server->vals->read_rsp_size) {
4584 /*
4585 * win2k8 sometimes sends an offset of 0 when the read
4586 * is beyond the EOF. Treat it as if the data starts just after
4587 * the header.
4588 */
4589 cifs_dbg(FYI, "%s: data offset (%u) inside read response header\n",
4590 __func__, data_offset);
4591 data_offset = server->vals->read_rsp_size;
4592 } else if (data_offset > MAX_CIFS_SMALL_BUFFER_SIZE) {
4593 /* data_offset is beyond the end of smallbuf */
4594 cifs_dbg(FYI, "%s: data offset (%u) beyond end of smallbuf\n",
4595 __func__, data_offset);
4596 rdata->result = -EIO;
4597 if (is_offloaded)
4598 mid->mid_state = MID_RESPONSE_MALFORMED;
4599 else
4600 dequeue_mid(mid, rdata->result);
4601 return 0;
4602 }
4603
4604 pad_len = data_offset - server->vals->read_rsp_size;
4605
4606 if (buf_len <= data_offset) {
4607 /* read response payload is in pages */
4608 cur_page_idx = pad_len / PAGE_SIZE;
4609 cur_off = pad_len % PAGE_SIZE;
4610
4611 if (cur_page_idx != 0) {
4612 /* data offset is beyond the 1st page of response */
4613 cifs_dbg(FYI, "%s: data offset (%u) beyond 1st page of response\n",
4614 __func__, data_offset);
4615 rdata->result = -EIO;
4616 if (is_offloaded)
4617 mid->mid_state = MID_RESPONSE_MALFORMED;
4618 else
4619 dequeue_mid(mid, rdata->result);
4620 return 0;
4621 }
4622
4623 if (data_len > pages_len - pad_len) {
4624 /* data_len is corrupt -- discard frame */
4625 rdata->result = -EIO;
4626 if (is_offloaded)
4627 mid->mid_state = MID_RESPONSE_MALFORMED;
4628 else
4629 dequeue_mid(mid, rdata->result);
4630 return 0;
4631 }
4632
4633 /* Copy the data to the output I/O iterator. */
4634 rdata->result = cifs_copy_pages_to_iter(pages, pages_len,
4635 cur_off, &rdata->iter);
4636 if (rdata->result != 0) {
4637 if (is_offloaded)
4638 mid->mid_state = MID_RESPONSE_MALFORMED;
4639 else
4640 dequeue_mid(mid, rdata->result);
4641 return 0;
4642 }
4643 rdata->got_bytes = pages_len;
4644
4645 } else if (buf_len >= data_offset + data_len) {
4646 /* read response payload is in buf */
4647 WARN_ONCE(pages && !xa_empty(pages),
4648 "read data can be either in buf or in pages");
4649 length = copy_to_iter(buf + data_offset, data_len, &rdata->iter);
4650 if (length < 0)
4651 return length;
4652 rdata->got_bytes = data_len;
4653 } else {
4654 /* read response payload cannot be in both buf and pages */
4655 WARN_ONCE(1, "buf can not contain only a part of read data");
4656 rdata->result = -EIO;
4657 if (is_offloaded)
4658 mid->mid_state = MID_RESPONSE_MALFORMED;
4659 else
4660 dequeue_mid(mid, rdata->result);
4661 return 0;
4662 }
4663
4664 if (is_offloaded)
4665 mid->mid_state = MID_RESPONSE_RECEIVED;
4666 else
4667 dequeue_mid(mid, false);
4668 return 0;
4669 }
4670
4671 struct smb2_decrypt_work {
4672 struct work_struct decrypt;
4673 struct TCP_Server_Info *server;
4674 struct xarray buffer;
4675 char *buf;
4676 unsigned int len;
4677 };
4678
4679
smb2_decrypt_offload(struct work_struct * work)4680 static void smb2_decrypt_offload(struct work_struct *work)
4681 {
4682 struct smb2_decrypt_work *dw = container_of(work,
4683 struct smb2_decrypt_work, decrypt);
4684 int rc;
4685 struct mid_q_entry *mid;
4686 struct iov_iter iter;
4687
4688 iov_iter_xarray(&iter, ITER_DEST, &dw->buffer, 0, dw->len);
4689 rc = decrypt_raw_data(dw->server, dw->buf, dw->server->vals->read_rsp_size,
4690 &iter, true);
4691 if (rc) {
4692 cifs_dbg(VFS, "error decrypting rc=%d\n", rc);
4693 goto free_pages;
4694 }
4695
4696 dw->server->lstrp = jiffies;
4697 mid = smb2_find_dequeue_mid(dw->server, dw->buf);
4698 if (mid == NULL)
4699 cifs_dbg(FYI, "mid not found\n");
4700 else {
4701 mid->decrypted = true;
4702 rc = handle_read_data(dw->server, mid, dw->buf,
4703 dw->server->vals->read_rsp_size,
4704 &dw->buffer, dw->len,
4705 true);
4706 if (rc >= 0) {
4707 #ifdef CONFIG_CIFS_STATS2
4708 mid->when_received = jiffies;
4709 #endif
4710 if (dw->server->ops->is_network_name_deleted)
4711 dw->server->ops->is_network_name_deleted(dw->buf,
4712 dw->server);
4713
4714 mid->callback(mid);
4715 } else {
4716 spin_lock(&dw->server->srv_lock);
4717 if (dw->server->tcpStatus == CifsNeedReconnect) {
4718 spin_lock(&dw->server->mid_lock);
4719 mid->mid_state = MID_RETRY_NEEDED;
4720 spin_unlock(&dw->server->mid_lock);
4721 spin_unlock(&dw->server->srv_lock);
4722 mid->callback(mid);
4723 } else {
4724 spin_lock(&dw->server->mid_lock);
4725 mid->mid_state = MID_REQUEST_SUBMITTED;
4726 mid->mid_flags &= ~(MID_DELETED);
4727 list_add_tail(&mid->qhead,
4728 &dw->server->pending_mid_q);
4729 spin_unlock(&dw->server->mid_lock);
4730 spin_unlock(&dw->server->srv_lock);
4731 }
4732 }
4733 release_mid(mid);
4734 }
4735
4736 free_pages:
4737 cifs_clear_xarray_buffer(&dw->buffer);
4738 cifs_small_buf_release(dw->buf);
4739 kfree(dw);
4740 }
4741
4742
4743 static int
receive_encrypted_read(struct TCP_Server_Info * server,struct mid_q_entry ** mid,int * num_mids)4744 receive_encrypted_read(struct TCP_Server_Info *server, struct mid_q_entry **mid,
4745 int *num_mids)
4746 {
4747 struct page *page;
4748 char *buf = server->smallbuf;
4749 struct smb2_transform_hdr *tr_hdr = (struct smb2_transform_hdr *)buf;
4750 struct iov_iter iter;
4751 unsigned int len, npages;
4752 unsigned int buflen = server->pdu_size;
4753 int rc;
4754 int i = 0;
4755 struct smb2_decrypt_work *dw;
4756
4757 dw = kzalloc(sizeof(struct smb2_decrypt_work), GFP_KERNEL);
4758 if (!dw)
4759 return -ENOMEM;
4760 xa_init(&dw->buffer);
4761 INIT_WORK(&dw->decrypt, smb2_decrypt_offload);
4762 dw->server = server;
4763
4764 *num_mids = 1;
4765 len = min_t(unsigned int, buflen, server->vals->read_rsp_size +
4766 sizeof(struct smb2_transform_hdr)) - HEADER_SIZE(server) + 1;
4767
4768 rc = cifs_read_from_socket(server, buf + HEADER_SIZE(server) - 1, len);
4769 if (rc < 0)
4770 goto free_dw;
4771 server->total_read += rc;
4772
4773 len = le32_to_cpu(tr_hdr->OriginalMessageSize) -
4774 server->vals->read_rsp_size;
4775 dw->len = len;
4776 npages = DIV_ROUND_UP(len, PAGE_SIZE);
4777
4778 rc = -ENOMEM;
4779 for (; i < npages; i++) {
4780 void *old;
4781
4782 page = alloc_page(GFP_KERNEL|__GFP_HIGHMEM);
4783 if (!page)
4784 goto discard_data;
4785 page->index = i;
4786 old = xa_store(&dw->buffer, i, page, GFP_KERNEL);
4787 if (xa_is_err(old)) {
4788 rc = xa_err(old);
4789 put_page(page);
4790 goto discard_data;
4791 }
4792 xa_set_mark(&dw->buffer, i, XA_MARK_0);
4793 }
4794
4795 iov_iter_xarray(&iter, ITER_DEST, &dw->buffer, 0, npages * PAGE_SIZE);
4796
4797 /* Read the data into the buffer and clear excess bufferage. */
4798 rc = cifs_read_iter_from_socket(server, &iter, dw->len);
4799 if (rc < 0)
4800 goto discard_data;
4801
4802 server->total_read += rc;
4803 if (rc < npages * PAGE_SIZE)
4804 iov_iter_zero(npages * PAGE_SIZE - rc, &iter);
4805 iov_iter_revert(&iter, npages * PAGE_SIZE);
4806 iov_iter_truncate(&iter, dw->len);
4807
4808 rc = cifs_discard_remaining_data(server);
4809 if (rc)
4810 goto free_pages;
4811
4812 /*
4813 * For large reads, offload to different thread for better performance,
4814 * use more cores decrypting which can be expensive
4815 */
4816
4817 if ((server->min_offload) && (server->in_flight > 1) &&
4818 (server->pdu_size >= server->min_offload)) {
4819 dw->buf = server->smallbuf;
4820 server->smallbuf = (char *)cifs_small_buf_get();
4821
4822 queue_work(decrypt_wq, &dw->decrypt);
4823 *num_mids = 0; /* worker thread takes care of finding mid */
4824 return -1;
4825 }
4826
4827 rc = decrypt_raw_data(server, buf, server->vals->read_rsp_size,
4828 &iter, false);
4829 if (rc)
4830 goto free_pages;
4831
4832 *mid = smb2_find_mid(server, buf);
4833 if (*mid == NULL) {
4834 cifs_dbg(FYI, "mid not found\n");
4835 } else {
4836 cifs_dbg(FYI, "mid found\n");
4837 (*mid)->decrypted = true;
4838 rc = handle_read_data(server, *mid, buf,
4839 server->vals->read_rsp_size,
4840 &dw->buffer, dw->len, false);
4841 if (rc >= 0) {
4842 if (server->ops->is_network_name_deleted) {
4843 server->ops->is_network_name_deleted(buf,
4844 server);
4845 }
4846 }
4847 }
4848
4849 free_pages:
4850 cifs_clear_xarray_buffer(&dw->buffer);
4851 free_dw:
4852 kfree(dw);
4853 return rc;
4854 discard_data:
4855 cifs_discard_remaining_data(server);
4856 goto free_pages;
4857 }
4858
4859 static int
receive_encrypted_standard(struct TCP_Server_Info * server,struct mid_q_entry ** mids,char ** bufs,int * num_mids)4860 receive_encrypted_standard(struct TCP_Server_Info *server,
4861 struct mid_q_entry **mids, char **bufs,
4862 int *num_mids)
4863 {
4864 int ret, length;
4865 char *buf = server->smallbuf;
4866 struct smb2_hdr *shdr;
4867 unsigned int pdu_length = server->pdu_size;
4868 unsigned int buf_size;
4869 unsigned int next_cmd;
4870 struct mid_q_entry *mid_entry;
4871 int next_is_large;
4872 char *next_buffer = NULL;
4873
4874 *num_mids = 0;
4875
4876 /* switch to large buffer if too big for a small one */
4877 if (pdu_length > MAX_CIFS_SMALL_BUFFER_SIZE) {
4878 server->large_buf = true;
4879 memcpy(server->bigbuf, buf, server->total_read);
4880 buf = server->bigbuf;
4881 }
4882
4883 /* now read the rest */
4884 length = cifs_read_from_socket(server, buf + HEADER_SIZE(server) - 1,
4885 pdu_length - HEADER_SIZE(server) + 1);
4886 if (length < 0)
4887 return length;
4888 server->total_read += length;
4889
4890 buf_size = pdu_length - sizeof(struct smb2_transform_hdr);
4891 length = decrypt_raw_data(server, buf, buf_size, NULL, false);
4892 if (length)
4893 return length;
4894
4895 next_is_large = server->large_buf;
4896 one_more:
4897 shdr = (struct smb2_hdr *)buf;
4898 next_cmd = le32_to_cpu(shdr->NextCommand);
4899 if (next_cmd) {
4900 if (WARN_ON_ONCE(next_cmd > pdu_length))
4901 return -1;
4902 if (next_is_large)
4903 next_buffer = (char *)cifs_buf_get();
4904 else
4905 next_buffer = (char *)cifs_small_buf_get();
4906 memcpy(next_buffer, buf + next_cmd, pdu_length - next_cmd);
4907 }
4908
4909 mid_entry = smb2_find_mid(server, buf);
4910 if (mid_entry == NULL)
4911 cifs_dbg(FYI, "mid not found\n");
4912 else {
4913 cifs_dbg(FYI, "mid found\n");
4914 mid_entry->decrypted = true;
4915 mid_entry->resp_buf_size = server->pdu_size;
4916 }
4917
4918 if (*num_mids >= MAX_COMPOUND) {
4919 cifs_server_dbg(VFS, "too many PDUs in compound\n");
4920 return -1;
4921 }
4922 bufs[*num_mids] = buf;
4923 mids[(*num_mids)++] = mid_entry;
4924
4925 if (mid_entry && mid_entry->handle)
4926 ret = mid_entry->handle(server, mid_entry);
4927 else
4928 ret = cifs_handle_standard(server, mid_entry);
4929
4930 if (ret == 0 && next_cmd) {
4931 pdu_length -= next_cmd;
4932 server->large_buf = next_is_large;
4933 if (next_is_large)
4934 server->bigbuf = buf = next_buffer;
4935 else
4936 server->smallbuf = buf = next_buffer;
4937 goto one_more;
4938 } else if (ret != 0) {
4939 /*
4940 * ret != 0 here means that we didn't get to handle_mid() thus
4941 * server->smallbuf and server->bigbuf are still valid. We need
4942 * to free next_buffer because it is not going to be used
4943 * anywhere.
4944 */
4945 if (next_is_large)
4946 free_rsp_buf(CIFS_LARGE_BUFFER, next_buffer);
4947 else
4948 free_rsp_buf(CIFS_SMALL_BUFFER, next_buffer);
4949 }
4950
4951 return ret;
4952 }
4953
4954 static int
smb3_receive_transform(struct TCP_Server_Info * server,struct mid_q_entry ** mids,char ** bufs,int * num_mids)4955 smb3_receive_transform(struct TCP_Server_Info *server,
4956 struct mid_q_entry **mids, char **bufs, int *num_mids)
4957 {
4958 char *buf = server->smallbuf;
4959 unsigned int pdu_length = server->pdu_size;
4960 struct smb2_transform_hdr *tr_hdr = (struct smb2_transform_hdr *)buf;
4961 unsigned int orig_len = le32_to_cpu(tr_hdr->OriginalMessageSize);
4962
4963 if (pdu_length < sizeof(struct smb2_transform_hdr) +
4964 sizeof(struct smb2_hdr)) {
4965 cifs_server_dbg(VFS, "Transform message is too small (%u)\n",
4966 pdu_length);
4967 cifs_reconnect(server, true);
4968 return -ECONNABORTED;
4969 }
4970
4971 if (pdu_length < orig_len + sizeof(struct smb2_transform_hdr)) {
4972 cifs_server_dbg(VFS, "Transform message is broken\n");
4973 cifs_reconnect(server, true);
4974 return -ECONNABORTED;
4975 }
4976
4977 /* TODO: add support for compounds containing READ. */
4978 if (pdu_length > CIFSMaxBufSize + MAX_HEADER_SIZE(server)) {
4979 return receive_encrypted_read(server, &mids[0], num_mids);
4980 }
4981
4982 return receive_encrypted_standard(server, mids, bufs, num_mids);
4983 }
4984
4985 int
smb3_handle_read_data(struct TCP_Server_Info * server,struct mid_q_entry * mid)4986 smb3_handle_read_data(struct TCP_Server_Info *server, struct mid_q_entry *mid)
4987 {
4988 char *buf = server->large_buf ? server->bigbuf : server->smallbuf;
4989
4990 return handle_read_data(server, mid, buf, server->pdu_size,
4991 NULL, 0, false);
4992 }
4993
smb2_next_header(struct TCP_Server_Info * server,char * buf,unsigned int * noff)4994 static int smb2_next_header(struct TCP_Server_Info *server, char *buf,
4995 unsigned int *noff)
4996 {
4997 struct smb2_hdr *hdr = (struct smb2_hdr *)buf;
4998 struct smb2_transform_hdr *t_hdr = (struct smb2_transform_hdr *)buf;
4999
5000 if (hdr->ProtocolId == SMB2_TRANSFORM_PROTO_NUM) {
5001 *noff = le32_to_cpu(t_hdr->OriginalMessageSize);
5002 if (unlikely(check_add_overflow(*noff, sizeof(*t_hdr), noff)))
5003 return -EINVAL;
5004 } else {
5005 *noff = le32_to_cpu(hdr->NextCommand);
5006 }
5007 if (unlikely(*noff && *noff < MID_HEADER_SIZE(server)))
5008 return -EINVAL;
5009 return 0;
5010 }
5011
__cifs_sfu_make_node(unsigned int xid,struct inode * inode,struct dentry * dentry,struct cifs_tcon * tcon,const char * full_path,umode_t mode,dev_t dev)5012 static int __cifs_sfu_make_node(unsigned int xid, struct inode *inode,
5013 struct dentry *dentry, struct cifs_tcon *tcon,
5014 const char *full_path, umode_t mode, dev_t dev)
5015 {
5016 struct TCP_Server_Info *server = tcon->ses->server;
5017 struct cifs_open_parms oparms;
5018 struct cifs_io_parms io_parms = {};
5019 struct cifs_sb_info *cifs_sb = CIFS_SB(inode->i_sb);
5020 struct cifs_fid fid;
5021 unsigned int bytes_written;
5022 struct win_dev pdev = {};
5023 struct kvec iov[2];
5024 __u32 oplock = server->oplocks ? REQ_OPLOCK : 0;
5025 int rc;
5026
5027 switch (mode & S_IFMT) {
5028 case S_IFCHR:
5029 strscpy(pdev.type, "IntxCHR", strlen("IntxChr"));
5030 pdev.major = cpu_to_le64(MAJOR(dev));
5031 pdev.minor = cpu_to_le64(MINOR(dev));
5032 break;
5033 case S_IFBLK:
5034 strscpy(pdev.type, "IntxBLK", strlen("IntxBLK"));
5035 pdev.major = cpu_to_le64(MAJOR(dev));
5036 pdev.minor = cpu_to_le64(MINOR(dev));
5037 break;
5038 case S_IFIFO:
5039 strscpy(pdev.type, "LnxFIFO", strlen("LnxFIFO"));
5040 break;
5041 default:
5042 return -EPERM;
5043 }
5044
5045 oparms = CIFS_OPARMS(cifs_sb, tcon, full_path, GENERIC_WRITE,
5046 FILE_CREATE, CREATE_NOT_DIR |
5047 CREATE_OPTION_SPECIAL, ACL_NO_MODE);
5048 oparms.fid = &fid;
5049
5050 rc = server->ops->open(xid, &oparms, &oplock, NULL);
5051 if (rc)
5052 return rc;
5053
5054 io_parms.pid = current->tgid;
5055 io_parms.tcon = tcon;
5056 io_parms.length = sizeof(pdev);
5057 iov[1].iov_base = &pdev;
5058 iov[1].iov_len = sizeof(pdev);
5059
5060 rc = server->ops->sync_write(xid, &fid, &io_parms,
5061 &bytes_written, iov, 1);
5062 server->ops->close(xid, tcon, &fid);
5063 return rc;
5064 }
5065
cifs_sfu_make_node(unsigned int xid,struct inode * inode,struct dentry * dentry,struct cifs_tcon * tcon,const char * full_path,umode_t mode,dev_t dev)5066 int cifs_sfu_make_node(unsigned int xid, struct inode *inode,
5067 struct dentry *dentry, struct cifs_tcon *tcon,
5068 const char *full_path, umode_t mode, dev_t dev)
5069 {
5070 struct inode *new = NULL;
5071 int rc;
5072
5073 rc = __cifs_sfu_make_node(xid, inode, dentry, tcon,
5074 full_path, mode, dev);
5075 if (rc)
5076 return rc;
5077
5078 if (tcon->posix_extensions) {
5079 rc = smb311_posix_get_inode_info(&new, full_path, NULL,
5080 inode->i_sb, xid);
5081 } else if (tcon->unix_ext) {
5082 rc = cifs_get_inode_info_unix(&new, full_path,
5083 inode->i_sb, xid);
5084 } else {
5085 rc = cifs_get_inode_info(&new, full_path, NULL,
5086 inode->i_sb, xid, NULL);
5087 }
5088 if (!rc)
5089 d_instantiate(dentry, new);
5090 return rc;
5091 }
5092
smb2_make_node(unsigned int xid,struct inode * inode,struct dentry * dentry,struct cifs_tcon * tcon,const char * full_path,umode_t mode,dev_t dev)5093 static int smb2_make_node(unsigned int xid, struct inode *inode,
5094 struct dentry *dentry, struct cifs_tcon *tcon,
5095 const char *full_path, umode_t mode, dev_t dev)
5096 {
5097 struct cifs_sb_info *cifs_sb = CIFS_SB(inode->i_sb);
5098 int rc;
5099
5100 /*
5101 * Check if mounted with mount parm 'sfu' mount parm.
5102 * SFU emulation should work with all servers, but only
5103 * supports block and char device (no socket & fifo),
5104 * and was used by default in earlier versions of Windows
5105 */
5106 if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_UNX_EMUL) {
5107 rc = cifs_sfu_make_node(xid, inode, dentry, tcon,
5108 full_path, mode, dev);
5109 } else {
5110 rc = smb2_mknod_reparse(xid, inode, dentry, tcon,
5111 full_path, mode, dev);
5112 }
5113 return rc;
5114 }
5115
5116 #ifdef CONFIG_CIFS_ALLOW_INSECURE_LEGACY
5117 struct smb_version_operations smb20_operations = {
5118 .compare_fids = smb2_compare_fids,
5119 .setup_request = smb2_setup_request,
5120 .setup_async_request = smb2_setup_async_request,
5121 .check_receive = smb2_check_receive,
5122 .add_credits = smb2_add_credits,
5123 .set_credits = smb2_set_credits,
5124 .get_credits_field = smb2_get_credits_field,
5125 .get_credits = smb2_get_credits,
5126 .wait_mtu_credits = cifs_wait_mtu_credits,
5127 .get_next_mid = smb2_get_next_mid,
5128 .revert_current_mid = smb2_revert_current_mid,
5129 .read_data_offset = smb2_read_data_offset,
5130 .read_data_length = smb2_read_data_length,
5131 .map_error = map_smb2_to_linux_error,
5132 .find_mid = smb2_find_mid,
5133 .check_message = smb2_check_message,
5134 .dump_detail = smb2_dump_detail,
5135 .clear_stats = smb2_clear_stats,
5136 .print_stats = smb2_print_stats,
5137 .is_oplock_break = smb2_is_valid_oplock_break,
5138 .handle_cancelled_mid = smb2_handle_cancelled_mid,
5139 .downgrade_oplock = smb2_downgrade_oplock,
5140 .need_neg = smb2_need_neg,
5141 .negotiate = smb2_negotiate,
5142 .negotiate_wsize = smb2_negotiate_wsize,
5143 .negotiate_rsize = smb2_negotiate_rsize,
5144 .sess_setup = SMB2_sess_setup,
5145 .logoff = SMB2_logoff,
5146 .tree_connect = SMB2_tcon,
5147 .tree_disconnect = SMB2_tdis,
5148 .qfs_tcon = smb2_qfs_tcon,
5149 .is_path_accessible = smb2_is_path_accessible,
5150 .can_echo = smb2_can_echo,
5151 .echo = SMB2_echo,
5152 .query_path_info = smb2_query_path_info,
5153 .query_reparse_point = smb2_query_reparse_point,
5154 .get_srv_inum = smb2_get_srv_inum,
5155 .query_file_info = smb2_query_file_info,
5156 .set_path_size = smb2_set_path_size,
5157 .set_file_size = smb2_set_file_size,
5158 .set_file_info = smb2_set_file_info,
5159 .set_compression = smb2_set_compression,
5160 .mkdir = smb2_mkdir,
5161 .mkdir_setinfo = smb2_mkdir_setinfo,
5162 .rmdir = smb2_rmdir,
5163 .unlink = smb2_unlink,
5164 .rename = smb2_rename_path,
5165 .create_hardlink = smb2_create_hardlink,
5166 .parse_reparse_point = smb2_parse_reparse_point,
5167 .query_mf_symlink = smb3_query_mf_symlink,
5168 .create_mf_symlink = smb3_create_mf_symlink,
5169 .create_reparse_symlink = smb2_create_reparse_symlink,
5170 .open = smb2_open_file,
5171 .set_fid = smb2_set_fid,
5172 .close = smb2_close_file,
5173 .flush = smb2_flush_file,
5174 .async_readv = smb2_async_readv,
5175 .async_writev = smb2_async_writev,
5176 .sync_read = smb2_sync_read,
5177 .sync_write = smb2_sync_write,
5178 .query_dir_first = smb2_query_dir_first,
5179 .query_dir_next = smb2_query_dir_next,
5180 .close_dir = smb2_close_dir,
5181 .calc_smb_size = smb2_calc_size,
5182 .is_status_pending = smb2_is_status_pending,
5183 .is_session_expired = smb2_is_session_expired,
5184 .oplock_response = smb2_oplock_response,
5185 .queryfs = smb2_queryfs,
5186 .mand_lock = smb2_mand_lock,
5187 .mand_unlock_range = smb2_unlock_range,
5188 .push_mand_locks = smb2_push_mandatory_locks,
5189 .get_lease_key = smb2_get_lease_key,
5190 .set_lease_key = smb2_set_lease_key,
5191 .new_lease_key = smb2_new_lease_key,
5192 .calc_signature = smb2_calc_signature,
5193 .is_read_op = smb2_is_read_op,
5194 .set_oplock_level = smb2_set_oplock_level,
5195 .create_lease_buf = smb2_create_lease_buf,
5196 .parse_lease_buf = smb2_parse_lease_buf,
5197 .copychunk_range = smb2_copychunk_range,
5198 .wp_retry_size = smb2_wp_retry_size,
5199 .dir_needs_close = smb2_dir_needs_close,
5200 .get_dfs_refer = smb2_get_dfs_refer,
5201 .select_sectype = smb2_select_sectype,
5202 #ifdef CONFIG_CIFS_XATTR
5203 .query_all_EAs = smb2_query_eas,
5204 .set_EA = smb2_set_ea,
5205 #endif /* CIFS_XATTR */
5206 .get_acl = get_smb2_acl,
5207 .get_acl_by_fid = get_smb2_acl_by_fid,
5208 .set_acl = set_smb2_acl,
5209 .next_header = smb2_next_header,
5210 .ioctl_query_info = smb2_ioctl_query_info,
5211 .make_node = smb2_make_node,
5212 .fiemap = smb3_fiemap,
5213 .llseek = smb3_llseek,
5214 .is_status_io_timeout = smb2_is_status_io_timeout,
5215 .is_network_name_deleted = smb2_is_network_name_deleted,
5216 };
5217 #endif /* CIFS_ALLOW_INSECURE_LEGACY */
5218
5219 struct smb_version_operations smb21_operations = {
5220 .compare_fids = smb2_compare_fids,
5221 .setup_request = smb2_setup_request,
5222 .setup_async_request = smb2_setup_async_request,
5223 .check_receive = smb2_check_receive,
5224 .add_credits = smb2_add_credits,
5225 .set_credits = smb2_set_credits,
5226 .get_credits_field = smb2_get_credits_field,
5227 .get_credits = smb2_get_credits,
5228 .wait_mtu_credits = smb2_wait_mtu_credits,
5229 .adjust_credits = smb2_adjust_credits,
5230 .get_next_mid = smb2_get_next_mid,
5231 .revert_current_mid = smb2_revert_current_mid,
5232 .read_data_offset = smb2_read_data_offset,
5233 .read_data_length = smb2_read_data_length,
5234 .map_error = map_smb2_to_linux_error,
5235 .find_mid = smb2_find_mid,
5236 .check_message = smb2_check_message,
5237 .dump_detail = smb2_dump_detail,
5238 .clear_stats = smb2_clear_stats,
5239 .print_stats = smb2_print_stats,
5240 .is_oplock_break = smb2_is_valid_oplock_break,
5241 .handle_cancelled_mid = smb2_handle_cancelled_mid,
5242 .downgrade_oplock = smb2_downgrade_oplock,
5243 .need_neg = smb2_need_neg,
5244 .negotiate = smb2_negotiate,
5245 .negotiate_wsize = smb2_negotiate_wsize,
5246 .negotiate_rsize = smb2_negotiate_rsize,
5247 .sess_setup = SMB2_sess_setup,
5248 .logoff = SMB2_logoff,
5249 .tree_connect = SMB2_tcon,
5250 .tree_disconnect = SMB2_tdis,
5251 .qfs_tcon = smb2_qfs_tcon,
5252 .is_path_accessible = smb2_is_path_accessible,
5253 .can_echo = smb2_can_echo,
5254 .echo = SMB2_echo,
5255 .query_path_info = smb2_query_path_info,
5256 .query_reparse_point = smb2_query_reparse_point,
5257 .get_srv_inum = smb2_get_srv_inum,
5258 .query_file_info = smb2_query_file_info,
5259 .set_path_size = smb2_set_path_size,
5260 .set_file_size = smb2_set_file_size,
5261 .set_file_info = smb2_set_file_info,
5262 .set_compression = smb2_set_compression,
5263 .mkdir = smb2_mkdir,
5264 .mkdir_setinfo = smb2_mkdir_setinfo,
5265 .rmdir = smb2_rmdir,
5266 .unlink = smb2_unlink,
5267 .rename = smb2_rename_path,
5268 .create_hardlink = smb2_create_hardlink,
5269 .parse_reparse_point = smb2_parse_reparse_point,
5270 .query_mf_symlink = smb3_query_mf_symlink,
5271 .create_mf_symlink = smb3_create_mf_symlink,
5272 .create_reparse_symlink = smb2_create_reparse_symlink,
5273 .open = smb2_open_file,
5274 .set_fid = smb2_set_fid,
5275 .close = smb2_close_file,
5276 .flush = smb2_flush_file,
5277 .async_readv = smb2_async_readv,
5278 .async_writev = smb2_async_writev,
5279 .sync_read = smb2_sync_read,
5280 .sync_write = smb2_sync_write,
5281 .query_dir_first = smb2_query_dir_first,
5282 .query_dir_next = smb2_query_dir_next,
5283 .close_dir = smb2_close_dir,
5284 .calc_smb_size = smb2_calc_size,
5285 .is_status_pending = smb2_is_status_pending,
5286 .is_session_expired = smb2_is_session_expired,
5287 .oplock_response = smb2_oplock_response,
5288 .queryfs = smb2_queryfs,
5289 .mand_lock = smb2_mand_lock,
5290 .mand_unlock_range = smb2_unlock_range,
5291 .push_mand_locks = smb2_push_mandatory_locks,
5292 .get_lease_key = smb2_get_lease_key,
5293 .set_lease_key = smb2_set_lease_key,
5294 .new_lease_key = smb2_new_lease_key,
5295 .calc_signature = smb2_calc_signature,
5296 .is_read_op = smb21_is_read_op,
5297 .set_oplock_level = smb21_set_oplock_level,
5298 .create_lease_buf = smb2_create_lease_buf,
5299 .parse_lease_buf = smb2_parse_lease_buf,
5300 .copychunk_range = smb2_copychunk_range,
5301 .wp_retry_size = smb2_wp_retry_size,
5302 .dir_needs_close = smb2_dir_needs_close,
5303 .enum_snapshots = smb3_enum_snapshots,
5304 .notify = smb3_notify,
5305 .get_dfs_refer = smb2_get_dfs_refer,
5306 .select_sectype = smb2_select_sectype,
5307 #ifdef CONFIG_CIFS_XATTR
5308 .query_all_EAs = smb2_query_eas,
5309 .set_EA = smb2_set_ea,
5310 #endif /* CIFS_XATTR */
5311 .get_acl = get_smb2_acl,
5312 .get_acl_by_fid = get_smb2_acl_by_fid,
5313 .set_acl = set_smb2_acl,
5314 .next_header = smb2_next_header,
5315 .ioctl_query_info = smb2_ioctl_query_info,
5316 .make_node = smb2_make_node,
5317 .fiemap = smb3_fiemap,
5318 .llseek = smb3_llseek,
5319 .is_status_io_timeout = smb2_is_status_io_timeout,
5320 .is_network_name_deleted = smb2_is_network_name_deleted,
5321 };
5322
5323 struct smb_version_operations smb30_operations = {
5324 .compare_fids = smb2_compare_fids,
5325 .setup_request = smb2_setup_request,
5326 .setup_async_request = smb2_setup_async_request,
5327 .check_receive = smb2_check_receive,
5328 .add_credits = smb2_add_credits,
5329 .set_credits = smb2_set_credits,
5330 .get_credits_field = smb2_get_credits_field,
5331 .get_credits = smb2_get_credits,
5332 .wait_mtu_credits = smb2_wait_mtu_credits,
5333 .adjust_credits = smb2_adjust_credits,
5334 .get_next_mid = smb2_get_next_mid,
5335 .revert_current_mid = smb2_revert_current_mid,
5336 .read_data_offset = smb2_read_data_offset,
5337 .read_data_length = smb2_read_data_length,
5338 .map_error = map_smb2_to_linux_error,
5339 .find_mid = smb2_find_mid,
5340 .check_message = smb2_check_message,
5341 .dump_detail = smb2_dump_detail,
5342 .clear_stats = smb2_clear_stats,
5343 .print_stats = smb2_print_stats,
5344 .dump_share_caps = smb2_dump_share_caps,
5345 .is_oplock_break = smb2_is_valid_oplock_break,
5346 .handle_cancelled_mid = smb2_handle_cancelled_mid,
5347 .downgrade_oplock = smb3_downgrade_oplock,
5348 .need_neg = smb2_need_neg,
5349 .negotiate = smb2_negotiate,
5350 .negotiate_wsize = smb3_negotiate_wsize,
5351 .negotiate_rsize = smb3_negotiate_rsize,
5352 .sess_setup = SMB2_sess_setup,
5353 .logoff = SMB2_logoff,
5354 .tree_connect = SMB2_tcon,
5355 .tree_disconnect = SMB2_tdis,
5356 .qfs_tcon = smb3_qfs_tcon,
5357 .query_server_interfaces = SMB3_request_interfaces,
5358 .is_path_accessible = smb2_is_path_accessible,
5359 .can_echo = smb2_can_echo,
5360 .echo = SMB2_echo,
5361 .query_path_info = smb2_query_path_info,
5362 /* WSL tags introduced long after smb2.1, enable for SMB3, 3.11 only */
5363 .query_reparse_point = smb2_query_reparse_point,
5364 .get_srv_inum = smb2_get_srv_inum,
5365 .query_file_info = smb2_query_file_info,
5366 .set_path_size = smb2_set_path_size,
5367 .set_file_size = smb2_set_file_size,
5368 .set_file_info = smb2_set_file_info,
5369 .set_compression = smb2_set_compression,
5370 .mkdir = smb2_mkdir,
5371 .mkdir_setinfo = smb2_mkdir_setinfo,
5372 .rmdir = smb2_rmdir,
5373 .unlink = smb2_unlink,
5374 .rename = smb2_rename_path,
5375 .create_hardlink = smb2_create_hardlink,
5376 .parse_reparse_point = smb2_parse_reparse_point,
5377 .query_mf_symlink = smb3_query_mf_symlink,
5378 .create_mf_symlink = smb3_create_mf_symlink,
5379 .create_reparse_symlink = smb2_create_reparse_symlink,
5380 .open = smb2_open_file,
5381 .set_fid = smb2_set_fid,
5382 .close = smb2_close_file,
5383 .close_getattr = smb2_close_getattr,
5384 .flush = smb2_flush_file,
5385 .async_readv = smb2_async_readv,
5386 .async_writev = smb2_async_writev,
5387 .sync_read = smb2_sync_read,
5388 .sync_write = smb2_sync_write,
5389 .query_dir_first = smb2_query_dir_first,
5390 .query_dir_next = smb2_query_dir_next,
5391 .close_dir = smb2_close_dir,
5392 .calc_smb_size = smb2_calc_size,
5393 .is_status_pending = smb2_is_status_pending,
5394 .is_session_expired = smb2_is_session_expired,
5395 .oplock_response = smb2_oplock_response,
5396 .queryfs = smb2_queryfs,
5397 .mand_lock = smb2_mand_lock,
5398 .mand_unlock_range = smb2_unlock_range,
5399 .push_mand_locks = smb2_push_mandatory_locks,
5400 .get_lease_key = smb2_get_lease_key,
5401 .set_lease_key = smb2_set_lease_key,
5402 .new_lease_key = smb2_new_lease_key,
5403 .generate_signingkey = generate_smb30signingkey,
5404 .calc_signature = smb3_calc_signature,
5405 .set_integrity = smb3_set_integrity,
5406 .is_read_op = smb21_is_read_op,
5407 .set_oplock_level = smb3_set_oplock_level,
5408 .create_lease_buf = smb3_create_lease_buf,
5409 .parse_lease_buf = smb3_parse_lease_buf,
5410 .copychunk_range = smb2_copychunk_range,
5411 .duplicate_extents = smb2_duplicate_extents,
5412 .validate_negotiate = smb3_validate_negotiate,
5413 .wp_retry_size = smb2_wp_retry_size,
5414 .dir_needs_close = smb2_dir_needs_close,
5415 .fallocate = smb3_fallocate,
5416 .enum_snapshots = smb3_enum_snapshots,
5417 .notify = smb3_notify,
5418 .init_transform_rq = smb3_init_transform_rq,
5419 .is_transform_hdr = smb3_is_transform_hdr,
5420 .receive_transform = smb3_receive_transform,
5421 .get_dfs_refer = smb2_get_dfs_refer,
5422 .select_sectype = smb2_select_sectype,
5423 #ifdef CONFIG_CIFS_XATTR
5424 .query_all_EAs = smb2_query_eas,
5425 .set_EA = smb2_set_ea,
5426 #endif /* CIFS_XATTR */
5427 .get_acl = get_smb2_acl,
5428 .get_acl_by_fid = get_smb2_acl_by_fid,
5429 .set_acl = set_smb2_acl,
5430 .next_header = smb2_next_header,
5431 .ioctl_query_info = smb2_ioctl_query_info,
5432 .make_node = smb2_make_node,
5433 .fiemap = smb3_fiemap,
5434 .llseek = smb3_llseek,
5435 .is_status_io_timeout = smb2_is_status_io_timeout,
5436 .is_network_name_deleted = smb2_is_network_name_deleted,
5437 };
5438
5439 struct smb_version_operations smb311_operations = {
5440 .compare_fids = smb2_compare_fids,
5441 .setup_request = smb2_setup_request,
5442 .setup_async_request = smb2_setup_async_request,
5443 .check_receive = smb2_check_receive,
5444 .add_credits = smb2_add_credits,
5445 .set_credits = smb2_set_credits,
5446 .get_credits_field = smb2_get_credits_field,
5447 .get_credits = smb2_get_credits,
5448 .wait_mtu_credits = smb2_wait_mtu_credits,
5449 .adjust_credits = smb2_adjust_credits,
5450 .get_next_mid = smb2_get_next_mid,
5451 .revert_current_mid = smb2_revert_current_mid,
5452 .read_data_offset = smb2_read_data_offset,
5453 .read_data_length = smb2_read_data_length,
5454 .map_error = map_smb2_to_linux_error,
5455 .find_mid = smb2_find_mid,
5456 .check_message = smb2_check_message,
5457 .dump_detail = smb2_dump_detail,
5458 .clear_stats = smb2_clear_stats,
5459 .print_stats = smb2_print_stats,
5460 .dump_share_caps = smb2_dump_share_caps,
5461 .is_oplock_break = smb2_is_valid_oplock_break,
5462 .handle_cancelled_mid = smb2_handle_cancelled_mid,
5463 .downgrade_oplock = smb3_downgrade_oplock,
5464 .need_neg = smb2_need_neg,
5465 .negotiate = smb2_negotiate,
5466 .negotiate_wsize = smb3_negotiate_wsize,
5467 .negotiate_rsize = smb3_negotiate_rsize,
5468 .sess_setup = SMB2_sess_setup,
5469 .logoff = SMB2_logoff,
5470 .tree_connect = SMB2_tcon,
5471 .tree_disconnect = SMB2_tdis,
5472 .qfs_tcon = smb3_qfs_tcon,
5473 .query_server_interfaces = SMB3_request_interfaces,
5474 .is_path_accessible = smb2_is_path_accessible,
5475 .can_echo = smb2_can_echo,
5476 .echo = SMB2_echo,
5477 .query_path_info = smb2_query_path_info,
5478 .query_reparse_point = smb2_query_reparse_point,
5479 .get_srv_inum = smb2_get_srv_inum,
5480 .query_file_info = smb2_query_file_info,
5481 .set_path_size = smb2_set_path_size,
5482 .set_file_size = smb2_set_file_size,
5483 .set_file_info = smb2_set_file_info,
5484 .set_compression = smb2_set_compression,
5485 .mkdir = smb2_mkdir,
5486 .mkdir_setinfo = smb2_mkdir_setinfo,
5487 .posix_mkdir = smb311_posix_mkdir,
5488 .rmdir = smb2_rmdir,
5489 .unlink = smb2_unlink,
5490 .rename = smb2_rename_path,
5491 .create_hardlink = smb2_create_hardlink,
5492 .parse_reparse_point = smb2_parse_reparse_point,
5493 .query_mf_symlink = smb3_query_mf_symlink,
5494 .create_mf_symlink = smb3_create_mf_symlink,
5495 .create_reparse_symlink = smb2_create_reparse_symlink,
5496 .open = smb2_open_file,
5497 .set_fid = smb2_set_fid,
5498 .close = smb2_close_file,
5499 .close_getattr = smb2_close_getattr,
5500 .flush = smb2_flush_file,
5501 .async_readv = smb2_async_readv,
5502 .async_writev = smb2_async_writev,
5503 .sync_read = smb2_sync_read,
5504 .sync_write = smb2_sync_write,
5505 .query_dir_first = smb2_query_dir_first,
5506 .query_dir_next = smb2_query_dir_next,
5507 .close_dir = smb2_close_dir,
5508 .calc_smb_size = smb2_calc_size,
5509 .is_status_pending = smb2_is_status_pending,
5510 .is_session_expired = smb2_is_session_expired,
5511 .oplock_response = smb2_oplock_response,
5512 .queryfs = smb311_queryfs,
5513 .mand_lock = smb2_mand_lock,
5514 .mand_unlock_range = smb2_unlock_range,
5515 .push_mand_locks = smb2_push_mandatory_locks,
5516 .get_lease_key = smb2_get_lease_key,
5517 .set_lease_key = smb2_set_lease_key,
5518 .new_lease_key = smb2_new_lease_key,
5519 .generate_signingkey = generate_smb311signingkey,
5520 .calc_signature = smb3_calc_signature,
5521 .set_integrity = smb3_set_integrity,
5522 .is_read_op = smb21_is_read_op,
5523 .set_oplock_level = smb3_set_oplock_level,
5524 .create_lease_buf = smb3_create_lease_buf,
5525 .parse_lease_buf = smb3_parse_lease_buf,
5526 .copychunk_range = smb2_copychunk_range,
5527 .duplicate_extents = smb2_duplicate_extents,
5528 /* .validate_negotiate = smb3_validate_negotiate, */ /* not used in 3.11 */
5529 .wp_retry_size = smb2_wp_retry_size,
5530 .dir_needs_close = smb2_dir_needs_close,
5531 .fallocate = smb3_fallocate,
5532 .enum_snapshots = smb3_enum_snapshots,
5533 .notify = smb3_notify,
5534 .init_transform_rq = smb3_init_transform_rq,
5535 .is_transform_hdr = smb3_is_transform_hdr,
5536 .receive_transform = smb3_receive_transform,
5537 .get_dfs_refer = smb2_get_dfs_refer,
5538 .select_sectype = smb2_select_sectype,
5539 #ifdef CONFIG_CIFS_XATTR
5540 .query_all_EAs = smb2_query_eas,
5541 .set_EA = smb2_set_ea,
5542 #endif /* CIFS_XATTR */
5543 .get_acl = get_smb2_acl,
5544 .get_acl_by_fid = get_smb2_acl_by_fid,
5545 .set_acl = set_smb2_acl,
5546 .next_header = smb2_next_header,
5547 .ioctl_query_info = smb2_ioctl_query_info,
5548 .make_node = smb2_make_node,
5549 .fiemap = smb3_fiemap,
5550 .llseek = smb3_llseek,
5551 .is_status_io_timeout = smb2_is_status_io_timeout,
5552 .is_network_name_deleted = smb2_is_network_name_deleted,
5553 };
5554
5555 #ifdef CONFIG_CIFS_ALLOW_INSECURE_LEGACY
5556 struct smb_version_values smb20_values = {
5557 .version_string = SMB20_VERSION_STRING,
5558 .protocol_id = SMB20_PROT_ID,
5559 .req_capabilities = 0, /* MBZ */
5560 .large_lock_type = 0,
5561 .exclusive_lock_type = SMB2_LOCKFLAG_EXCLUSIVE,
5562 .shared_lock_type = SMB2_LOCKFLAG_SHARED,
5563 .unlock_lock_type = SMB2_LOCKFLAG_UNLOCK,
5564 .header_size = sizeof(struct smb2_hdr),
5565 .header_preamble_size = 0,
5566 .max_header_size = MAX_SMB2_HDR_SIZE,
5567 .read_rsp_size = sizeof(struct smb2_read_rsp),
5568 .lock_cmd = SMB2_LOCK,
5569 .cap_unix = 0,
5570 .cap_nt_find = SMB2_NT_FIND,
5571 .cap_large_files = SMB2_LARGE_FILES,
5572 .signing_enabled = SMB2_NEGOTIATE_SIGNING_ENABLED | SMB2_NEGOTIATE_SIGNING_REQUIRED,
5573 .signing_required = SMB2_NEGOTIATE_SIGNING_REQUIRED,
5574 .create_lease_size = sizeof(struct create_lease),
5575 };
5576 #endif /* ALLOW_INSECURE_LEGACY */
5577
5578 struct smb_version_values smb21_values = {
5579 .version_string = SMB21_VERSION_STRING,
5580 .protocol_id = SMB21_PROT_ID,
5581 .req_capabilities = 0, /* MBZ on negotiate req until SMB3 dialect */
5582 .large_lock_type = 0,
5583 .exclusive_lock_type = SMB2_LOCKFLAG_EXCLUSIVE,
5584 .shared_lock_type = SMB2_LOCKFLAG_SHARED,
5585 .unlock_lock_type = SMB2_LOCKFLAG_UNLOCK,
5586 .header_size = sizeof(struct smb2_hdr),
5587 .header_preamble_size = 0,
5588 .max_header_size = MAX_SMB2_HDR_SIZE,
5589 .read_rsp_size = sizeof(struct smb2_read_rsp),
5590 .lock_cmd = SMB2_LOCK,
5591 .cap_unix = 0,
5592 .cap_nt_find = SMB2_NT_FIND,
5593 .cap_large_files = SMB2_LARGE_FILES,
5594 .signing_enabled = SMB2_NEGOTIATE_SIGNING_ENABLED | SMB2_NEGOTIATE_SIGNING_REQUIRED,
5595 .signing_required = SMB2_NEGOTIATE_SIGNING_REQUIRED,
5596 .create_lease_size = sizeof(struct create_lease),
5597 };
5598
5599 struct smb_version_values smb3any_values = {
5600 .version_string = SMB3ANY_VERSION_STRING,
5601 .protocol_id = SMB302_PROT_ID, /* doesn't matter, send protocol array */
5602 .req_capabilities = SMB2_GLOBAL_CAP_DFS | SMB2_GLOBAL_CAP_LEASING | SMB2_GLOBAL_CAP_LARGE_MTU | SMB2_GLOBAL_CAP_PERSISTENT_HANDLES | SMB2_GLOBAL_CAP_ENCRYPTION | SMB2_GLOBAL_CAP_DIRECTORY_LEASING,
5603 .large_lock_type = 0,
5604 .exclusive_lock_type = SMB2_LOCKFLAG_EXCLUSIVE,
5605 .shared_lock_type = SMB2_LOCKFLAG_SHARED,
5606 .unlock_lock_type = SMB2_LOCKFLAG_UNLOCK,
5607 .header_size = sizeof(struct smb2_hdr),
5608 .header_preamble_size = 0,
5609 .max_header_size = MAX_SMB2_HDR_SIZE,
5610 .read_rsp_size = sizeof(struct smb2_read_rsp),
5611 .lock_cmd = SMB2_LOCK,
5612 .cap_unix = 0,
5613 .cap_nt_find = SMB2_NT_FIND,
5614 .cap_large_files = SMB2_LARGE_FILES,
5615 .signing_enabled = SMB2_NEGOTIATE_SIGNING_ENABLED | SMB2_NEGOTIATE_SIGNING_REQUIRED,
5616 .signing_required = SMB2_NEGOTIATE_SIGNING_REQUIRED,
5617 .create_lease_size = sizeof(struct create_lease_v2),
5618 };
5619
5620 struct smb_version_values smbdefault_values = {
5621 .version_string = SMBDEFAULT_VERSION_STRING,
5622 .protocol_id = SMB302_PROT_ID, /* doesn't matter, send protocol array */
5623 .req_capabilities = SMB2_GLOBAL_CAP_DFS | SMB2_GLOBAL_CAP_LEASING | SMB2_GLOBAL_CAP_LARGE_MTU | SMB2_GLOBAL_CAP_PERSISTENT_HANDLES | SMB2_GLOBAL_CAP_ENCRYPTION | SMB2_GLOBAL_CAP_DIRECTORY_LEASING,
5624 .large_lock_type = 0,
5625 .exclusive_lock_type = SMB2_LOCKFLAG_EXCLUSIVE,
5626 .shared_lock_type = SMB2_LOCKFLAG_SHARED,
5627 .unlock_lock_type = SMB2_LOCKFLAG_UNLOCK,
5628 .header_size = sizeof(struct smb2_hdr),
5629 .header_preamble_size = 0,
5630 .max_header_size = MAX_SMB2_HDR_SIZE,
5631 .read_rsp_size = sizeof(struct smb2_read_rsp),
5632 .lock_cmd = SMB2_LOCK,
5633 .cap_unix = 0,
5634 .cap_nt_find = SMB2_NT_FIND,
5635 .cap_large_files = SMB2_LARGE_FILES,
5636 .signing_enabled = SMB2_NEGOTIATE_SIGNING_ENABLED | SMB2_NEGOTIATE_SIGNING_REQUIRED,
5637 .signing_required = SMB2_NEGOTIATE_SIGNING_REQUIRED,
5638 .create_lease_size = sizeof(struct create_lease_v2),
5639 };
5640
5641 struct smb_version_values smb30_values = {
5642 .version_string = SMB30_VERSION_STRING,
5643 .protocol_id = SMB30_PROT_ID,
5644 .req_capabilities = SMB2_GLOBAL_CAP_DFS | SMB2_GLOBAL_CAP_LEASING | SMB2_GLOBAL_CAP_LARGE_MTU | SMB2_GLOBAL_CAP_PERSISTENT_HANDLES | SMB2_GLOBAL_CAP_ENCRYPTION | SMB2_GLOBAL_CAP_DIRECTORY_LEASING,
5645 .large_lock_type = 0,
5646 .exclusive_lock_type = SMB2_LOCKFLAG_EXCLUSIVE,
5647 .shared_lock_type = SMB2_LOCKFLAG_SHARED,
5648 .unlock_lock_type = SMB2_LOCKFLAG_UNLOCK,
5649 .header_size = sizeof(struct smb2_hdr),
5650 .header_preamble_size = 0,
5651 .max_header_size = MAX_SMB2_HDR_SIZE,
5652 .read_rsp_size = sizeof(struct smb2_read_rsp),
5653 .lock_cmd = SMB2_LOCK,
5654 .cap_unix = 0,
5655 .cap_nt_find = SMB2_NT_FIND,
5656 .cap_large_files = SMB2_LARGE_FILES,
5657 .signing_enabled = SMB2_NEGOTIATE_SIGNING_ENABLED | SMB2_NEGOTIATE_SIGNING_REQUIRED,
5658 .signing_required = SMB2_NEGOTIATE_SIGNING_REQUIRED,
5659 .create_lease_size = sizeof(struct create_lease_v2),
5660 };
5661
5662 struct smb_version_values smb302_values = {
5663 .version_string = SMB302_VERSION_STRING,
5664 .protocol_id = SMB302_PROT_ID,
5665 .req_capabilities = SMB2_GLOBAL_CAP_DFS | SMB2_GLOBAL_CAP_LEASING | SMB2_GLOBAL_CAP_LARGE_MTU | SMB2_GLOBAL_CAP_PERSISTENT_HANDLES | SMB2_GLOBAL_CAP_ENCRYPTION | SMB2_GLOBAL_CAP_DIRECTORY_LEASING,
5666 .large_lock_type = 0,
5667 .exclusive_lock_type = SMB2_LOCKFLAG_EXCLUSIVE,
5668 .shared_lock_type = SMB2_LOCKFLAG_SHARED,
5669 .unlock_lock_type = SMB2_LOCKFLAG_UNLOCK,
5670 .header_size = sizeof(struct smb2_hdr),
5671 .header_preamble_size = 0,
5672 .max_header_size = MAX_SMB2_HDR_SIZE,
5673 .read_rsp_size = sizeof(struct smb2_read_rsp),
5674 .lock_cmd = SMB2_LOCK,
5675 .cap_unix = 0,
5676 .cap_nt_find = SMB2_NT_FIND,
5677 .cap_large_files = SMB2_LARGE_FILES,
5678 .signing_enabled = SMB2_NEGOTIATE_SIGNING_ENABLED | SMB2_NEGOTIATE_SIGNING_REQUIRED,
5679 .signing_required = SMB2_NEGOTIATE_SIGNING_REQUIRED,
5680 .create_lease_size = sizeof(struct create_lease_v2),
5681 };
5682
5683 struct smb_version_values smb311_values = {
5684 .version_string = SMB311_VERSION_STRING,
5685 .protocol_id = SMB311_PROT_ID,
5686 .req_capabilities = SMB2_GLOBAL_CAP_DFS | SMB2_GLOBAL_CAP_LEASING | SMB2_GLOBAL_CAP_LARGE_MTU | SMB2_GLOBAL_CAP_PERSISTENT_HANDLES | SMB2_GLOBAL_CAP_ENCRYPTION | SMB2_GLOBAL_CAP_DIRECTORY_LEASING,
5687 .large_lock_type = 0,
5688 .exclusive_lock_type = SMB2_LOCKFLAG_EXCLUSIVE,
5689 .shared_lock_type = SMB2_LOCKFLAG_SHARED,
5690 .unlock_lock_type = SMB2_LOCKFLAG_UNLOCK,
5691 .header_size = sizeof(struct smb2_hdr),
5692 .header_preamble_size = 0,
5693 .max_header_size = MAX_SMB2_HDR_SIZE,
5694 .read_rsp_size = sizeof(struct smb2_read_rsp),
5695 .lock_cmd = SMB2_LOCK,
5696 .cap_unix = 0,
5697 .cap_nt_find = SMB2_NT_FIND,
5698 .cap_large_files = SMB2_LARGE_FILES,
5699 .signing_enabled = SMB2_NEGOTIATE_SIGNING_ENABLED | SMB2_NEGOTIATE_SIGNING_REQUIRED,
5700 .signing_required = SMB2_NEGOTIATE_SIGNING_REQUIRED,
5701 .create_lease_size = sizeof(struct create_lease_v2),
5702 };
5703