xref: /openbmc/linux/fs/smb/client/file.c (revision d6b6592ac6d11eab91e6758d224eac35f4122aca)
1  // SPDX-License-Identifier: LGPL-2.1
2  /*
3   *
4   *   vfs operations that deal with files
5   *
6   *   Copyright (C) International Business Machines  Corp., 2002,2010
7   *   Author(s): Steve French (sfrench@us.ibm.com)
8   *              Jeremy Allison (jra@samba.org)
9   *
10   */
11  #include <linux/fs.h>
12  #include <linux/filelock.h>
13  #include <linux/backing-dev.h>
14  #include <linux/stat.h>
15  #include <linux/fcntl.h>
16  #include <linux/pagemap.h>
17  #include <linux/pagevec.h>
18  #include <linux/writeback.h>
19  #include <linux/task_io_accounting_ops.h>
20  #include <linux/delay.h>
21  #include <linux/mount.h>
22  #include <linux/slab.h>
23  #include <linux/swap.h>
24  #include <linux/mm.h>
25  #include <asm/div64.h>
26  #include "cifsfs.h"
27  #include "cifspdu.h"
28  #include "cifsglob.h"
29  #include "cifsproto.h"
30  #include "smb2proto.h"
31  #include "cifs_unicode.h"
32  #include "cifs_debug.h"
33  #include "cifs_fs_sb.h"
34  #include "fscache.h"
35  #include "smbdirect.h"
36  #include "fs_context.h"
37  #include "cifs_ioctl.h"
38  #include "cached_dir.h"
39  
40  /*
41   * Remove the dirty flags from a span of pages.
42   */
cifs_undirty_folios(struct inode * inode,loff_t start,unsigned int len)43  static void cifs_undirty_folios(struct inode *inode, loff_t start, unsigned int len)
44  {
45  	struct address_space *mapping = inode->i_mapping;
46  	struct folio *folio;
47  	pgoff_t end;
48  
49  	XA_STATE(xas, &mapping->i_pages, start / PAGE_SIZE);
50  
51  	rcu_read_lock();
52  
53  	end = (start + len - 1) / PAGE_SIZE;
54  	xas_for_each_marked(&xas, folio, end, PAGECACHE_TAG_DIRTY) {
55  		if (xas_retry(&xas, folio))
56  			continue;
57  		xas_pause(&xas);
58  		rcu_read_unlock();
59  		folio_lock(folio);
60  		folio_clear_dirty_for_io(folio);
61  		folio_unlock(folio);
62  		rcu_read_lock();
63  	}
64  
65  	rcu_read_unlock();
66  }
67  
68  /*
69   * Completion of write to server.
70   */
cifs_pages_written_back(struct inode * inode,loff_t start,unsigned int len)71  void cifs_pages_written_back(struct inode *inode, loff_t start, unsigned int len)
72  {
73  	struct address_space *mapping = inode->i_mapping;
74  	struct folio *folio;
75  	pgoff_t end;
76  
77  	XA_STATE(xas, &mapping->i_pages, start / PAGE_SIZE);
78  
79  	if (!len)
80  		return;
81  
82  	rcu_read_lock();
83  
84  	end = (start + len - 1) / PAGE_SIZE;
85  	xas_for_each(&xas, folio, end) {
86  		if (xas_retry(&xas, folio))
87  			continue;
88  		if (!folio_test_writeback(folio)) {
89  			WARN_ONCE(1, "bad %x @%llx page %lx %lx\n",
90  				  len, start, folio->index, end);
91  			continue;
92  		}
93  
94  		folio_detach_private(folio);
95  		folio_end_writeback(folio);
96  	}
97  
98  	rcu_read_unlock();
99  }
100  
101  /*
102   * Failure of write to server.
103   */
cifs_pages_write_failed(struct inode * inode,loff_t start,unsigned int len)104  void cifs_pages_write_failed(struct inode *inode, loff_t start, unsigned int len)
105  {
106  	struct address_space *mapping = inode->i_mapping;
107  	struct folio *folio;
108  	pgoff_t end;
109  
110  	XA_STATE(xas, &mapping->i_pages, start / PAGE_SIZE);
111  
112  	if (!len)
113  		return;
114  
115  	rcu_read_lock();
116  
117  	end = (start + len - 1) / PAGE_SIZE;
118  	xas_for_each(&xas, folio, end) {
119  		if (xas_retry(&xas, folio))
120  			continue;
121  		if (!folio_test_writeback(folio)) {
122  			WARN_ONCE(1, "bad %x @%llx page %lx %lx\n",
123  				  len, start, folio->index, end);
124  			continue;
125  		}
126  
127  		folio_set_error(folio);
128  		folio_end_writeback(folio);
129  	}
130  
131  	rcu_read_unlock();
132  }
133  
134  /*
135   * Redirty pages after a temporary failure.
136   */
cifs_pages_write_redirty(struct inode * inode,loff_t start,unsigned int len)137  void cifs_pages_write_redirty(struct inode *inode, loff_t start, unsigned int len)
138  {
139  	struct address_space *mapping = inode->i_mapping;
140  	struct folio *folio;
141  	pgoff_t end;
142  
143  	XA_STATE(xas, &mapping->i_pages, start / PAGE_SIZE);
144  
145  	if (!len)
146  		return;
147  
148  	rcu_read_lock();
149  
150  	end = (start + len - 1) / PAGE_SIZE;
151  	xas_for_each(&xas, folio, end) {
152  		if (!folio_test_writeback(folio)) {
153  			WARN_ONCE(1, "bad %x @%llx page %lx %lx\n",
154  				  len, start, folio->index, end);
155  			continue;
156  		}
157  
158  		filemap_dirty_folio(folio->mapping, folio);
159  		folio_end_writeback(folio);
160  	}
161  
162  	rcu_read_unlock();
163  }
164  
165  /*
166   * Mark as invalid, all open files on tree connections since they
167   * were closed when session to server was lost.
168   */
169  void
cifs_mark_open_files_invalid(struct cifs_tcon * tcon)170  cifs_mark_open_files_invalid(struct cifs_tcon *tcon)
171  {
172  	struct cifsFileInfo *open_file = NULL;
173  	struct list_head *tmp;
174  	struct list_head *tmp1;
175  
176  	/* only send once per connect */
177  	spin_lock(&tcon->tc_lock);
178  	if (tcon->need_reconnect)
179  		tcon->status = TID_NEED_RECON;
180  
181  	if (tcon->status != TID_NEED_RECON) {
182  		spin_unlock(&tcon->tc_lock);
183  		return;
184  	}
185  	tcon->status = TID_IN_FILES_INVALIDATE;
186  	spin_unlock(&tcon->tc_lock);
187  
188  	/* list all files open on tree connection and mark them invalid */
189  	spin_lock(&tcon->open_file_lock);
190  	list_for_each_safe(tmp, tmp1, &tcon->openFileList) {
191  		open_file = list_entry(tmp, struct cifsFileInfo, tlist);
192  		open_file->invalidHandle = true;
193  		open_file->oplock_break_cancelled = true;
194  	}
195  	spin_unlock(&tcon->open_file_lock);
196  
197  	invalidate_all_cached_dirs(tcon);
198  	spin_lock(&tcon->tc_lock);
199  	if (tcon->status == TID_IN_FILES_INVALIDATE)
200  		tcon->status = TID_NEED_TCON;
201  	spin_unlock(&tcon->tc_lock);
202  
203  	/*
204  	 * BB Add call to invalidate_inodes(sb) for all superblocks mounted
205  	 * to this tcon.
206  	 */
207  }
208  
cifs_convert_flags(unsigned int flags,int rdwr_for_fscache)209  static inline int cifs_convert_flags(unsigned int flags, int rdwr_for_fscache)
210  {
211  	if ((flags & O_ACCMODE) == O_RDONLY)
212  		return GENERIC_READ;
213  	else if ((flags & O_ACCMODE) == O_WRONLY)
214  		return rdwr_for_fscache == 1 ? (GENERIC_READ | GENERIC_WRITE) : GENERIC_WRITE;
215  	else if ((flags & O_ACCMODE) == O_RDWR) {
216  		/* GENERIC_ALL is too much permission to request
217  		   can cause unnecessary access denied on create */
218  		/* return GENERIC_ALL; */
219  		return (GENERIC_READ | GENERIC_WRITE);
220  	}
221  
222  	return (READ_CONTROL | FILE_WRITE_ATTRIBUTES | FILE_READ_ATTRIBUTES |
223  		FILE_WRITE_EA | FILE_APPEND_DATA | FILE_WRITE_DATA |
224  		FILE_READ_DATA);
225  }
226  
227  #ifdef CONFIG_CIFS_ALLOW_INSECURE_LEGACY
cifs_posix_convert_flags(unsigned int flags)228  static u32 cifs_posix_convert_flags(unsigned int flags)
229  {
230  	u32 posix_flags = 0;
231  
232  	if ((flags & O_ACCMODE) == O_RDONLY)
233  		posix_flags = SMB_O_RDONLY;
234  	else if ((flags & O_ACCMODE) == O_WRONLY)
235  		posix_flags = SMB_O_WRONLY;
236  	else if ((flags & O_ACCMODE) == O_RDWR)
237  		posix_flags = SMB_O_RDWR;
238  
239  	if (flags & O_CREAT) {
240  		posix_flags |= SMB_O_CREAT;
241  		if (flags & O_EXCL)
242  			posix_flags |= SMB_O_EXCL;
243  	} else if (flags & O_EXCL)
244  		cifs_dbg(FYI, "Application %s pid %d has incorrectly set O_EXCL flag but not O_CREAT on file open. Ignoring O_EXCL\n",
245  			 current->comm, current->tgid);
246  
247  	if (flags & O_TRUNC)
248  		posix_flags |= SMB_O_TRUNC;
249  	/* be safe and imply O_SYNC for O_DSYNC */
250  	if (flags & O_DSYNC)
251  		posix_flags |= SMB_O_SYNC;
252  	if (flags & O_DIRECTORY)
253  		posix_flags |= SMB_O_DIRECTORY;
254  	if (flags & O_NOFOLLOW)
255  		posix_flags |= SMB_O_NOFOLLOW;
256  	if (flags & O_DIRECT)
257  		posix_flags |= SMB_O_DIRECT;
258  
259  	return posix_flags;
260  }
261  #endif /* CONFIG_CIFS_ALLOW_INSECURE_LEGACY */
262  
cifs_get_disposition(unsigned int flags)263  static inline int cifs_get_disposition(unsigned int flags)
264  {
265  	if ((flags & (O_CREAT | O_EXCL)) == (O_CREAT | O_EXCL))
266  		return FILE_CREATE;
267  	else if ((flags & (O_CREAT | O_TRUNC)) == (O_CREAT | O_TRUNC))
268  		return FILE_OVERWRITE_IF;
269  	else if ((flags & O_CREAT) == O_CREAT)
270  		return FILE_OPEN_IF;
271  	else if ((flags & O_TRUNC) == O_TRUNC)
272  		return FILE_OVERWRITE;
273  	else
274  		return FILE_OPEN;
275  }
276  
277  #ifdef CONFIG_CIFS_ALLOW_INSECURE_LEGACY
cifs_posix_open(const char * full_path,struct inode ** pinode,struct super_block * sb,int mode,unsigned int f_flags,__u32 * poplock,__u16 * pnetfid,unsigned int xid)278  int cifs_posix_open(const char *full_path, struct inode **pinode,
279  			struct super_block *sb, int mode, unsigned int f_flags,
280  			__u32 *poplock, __u16 *pnetfid, unsigned int xid)
281  {
282  	int rc;
283  	FILE_UNIX_BASIC_INFO *presp_data;
284  	__u32 posix_flags = 0;
285  	struct cifs_sb_info *cifs_sb = CIFS_SB(sb);
286  	struct cifs_fattr fattr;
287  	struct tcon_link *tlink;
288  	struct cifs_tcon *tcon;
289  
290  	cifs_dbg(FYI, "posix open %s\n", full_path);
291  
292  	presp_data = kzalloc(sizeof(FILE_UNIX_BASIC_INFO), GFP_KERNEL);
293  	if (presp_data == NULL)
294  		return -ENOMEM;
295  
296  	tlink = cifs_sb_tlink(cifs_sb);
297  	if (IS_ERR(tlink)) {
298  		rc = PTR_ERR(tlink);
299  		goto posix_open_ret;
300  	}
301  
302  	tcon = tlink_tcon(tlink);
303  	mode &= ~current_umask();
304  
305  	posix_flags = cifs_posix_convert_flags(f_flags);
306  	rc = CIFSPOSIXCreate(xid, tcon, posix_flags, mode, pnetfid, presp_data,
307  			     poplock, full_path, cifs_sb->local_nls,
308  			     cifs_remap(cifs_sb));
309  	cifs_put_tlink(tlink);
310  
311  	if (rc)
312  		goto posix_open_ret;
313  
314  	if (presp_data->Type == cpu_to_le32(-1))
315  		goto posix_open_ret; /* open ok, caller does qpathinfo */
316  
317  	if (!pinode)
318  		goto posix_open_ret; /* caller does not need info */
319  
320  	cifs_unix_basic_to_fattr(&fattr, presp_data, cifs_sb);
321  
322  	/* get new inode and set it up */
323  	if (*pinode == NULL) {
324  		cifs_fill_uniqueid(sb, &fattr);
325  		*pinode = cifs_iget(sb, &fattr);
326  		if (!*pinode) {
327  			rc = -ENOMEM;
328  			goto posix_open_ret;
329  		}
330  	} else {
331  		cifs_revalidate_mapping(*pinode);
332  		rc = cifs_fattr_to_inode(*pinode, &fattr, false);
333  	}
334  
335  posix_open_ret:
336  	kfree(presp_data);
337  	return rc;
338  }
339  #endif /* CONFIG_CIFS_ALLOW_INSECURE_LEGACY */
340  
cifs_nt_open(const char * full_path,struct inode * inode,struct cifs_sb_info * cifs_sb,struct cifs_tcon * tcon,unsigned int f_flags,__u32 * oplock,struct cifs_fid * fid,unsigned int xid,struct cifs_open_info_data * buf)341  static int cifs_nt_open(const char *full_path, struct inode *inode, struct cifs_sb_info *cifs_sb,
342  			struct cifs_tcon *tcon, unsigned int f_flags, __u32 *oplock,
343  			struct cifs_fid *fid, unsigned int xid, struct cifs_open_info_data *buf)
344  {
345  	int rc;
346  	int desired_access;
347  	int disposition;
348  	int create_options = CREATE_NOT_DIR;
349  	struct TCP_Server_Info *server = tcon->ses->server;
350  	struct cifs_open_parms oparms;
351  	int rdwr_for_fscache = 0;
352  
353  	if (!server->ops->open)
354  		return -ENOSYS;
355  
356  	/* If we're caching, we need to be able to fill in around partial writes. */
357  	if (cifs_fscache_enabled(inode) && (f_flags & O_ACCMODE) == O_WRONLY)
358  		rdwr_for_fscache = 1;
359  
360  	desired_access = cifs_convert_flags(f_flags, rdwr_for_fscache);
361  
362  /*********************************************************************
363   *  open flag mapping table:
364   *
365   *	POSIX Flag            CIFS Disposition
366   *	----------            ----------------
367   *	O_CREAT               FILE_OPEN_IF
368   *	O_CREAT | O_EXCL      FILE_CREATE
369   *	O_CREAT | O_TRUNC     FILE_OVERWRITE_IF
370   *	O_TRUNC               FILE_OVERWRITE
371   *	none of the above     FILE_OPEN
372   *
373   *	Note that there is not a direct match between disposition
374   *	FILE_SUPERSEDE (ie create whether or not file exists although
375   *	O_CREAT | O_TRUNC is similar but truncates the existing
376   *	file rather than creating a new file as FILE_SUPERSEDE does
377   *	(which uses the attributes / metadata passed in on open call)
378   *?
379   *?  O_SYNC is a reasonable match to CIFS writethrough flag
380   *?  and the read write flags match reasonably.  O_LARGEFILE
381   *?  is irrelevant because largefile support is always used
382   *?  by this client. Flags O_APPEND, O_DIRECT, O_DIRECTORY,
383   *	 O_FASYNC, O_NOFOLLOW, O_NONBLOCK need further investigation
384   *********************************************************************/
385  
386  	disposition = cifs_get_disposition(f_flags);
387  
388  	/* BB pass O_SYNC flag through on file attributes .. BB */
389  
390  	/* O_SYNC also has bit for O_DSYNC so following check picks up either */
391  	if (f_flags & O_SYNC)
392  		create_options |= CREATE_WRITE_THROUGH;
393  
394  	if (f_flags & O_DIRECT)
395  		create_options |= CREATE_NO_BUFFER;
396  
397  retry_open:
398  	oparms = (struct cifs_open_parms) {
399  		.tcon = tcon,
400  		.cifs_sb = cifs_sb,
401  		.desired_access = desired_access,
402  		.create_options = cifs_create_options(cifs_sb, create_options),
403  		.disposition = disposition,
404  		.path = full_path,
405  		.fid = fid,
406  	};
407  
408  	rc = server->ops->open(xid, &oparms, oplock, buf);
409  	if (rc) {
410  		if (rc == -EACCES && rdwr_for_fscache == 1) {
411  			desired_access = cifs_convert_flags(f_flags, 0);
412  			rdwr_for_fscache = 2;
413  			goto retry_open;
414  		}
415  		return rc;
416  	}
417  	if (rdwr_for_fscache == 2)
418  		cifs_invalidate_cache(inode, FSCACHE_INVAL_DIO_WRITE);
419  
420  	/* TODO: Add support for calling posix query info but with passing in fid */
421  	if (tcon->unix_ext)
422  		rc = cifs_get_inode_info_unix(&inode, full_path, inode->i_sb,
423  					      xid);
424  	else
425  		rc = cifs_get_inode_info(&inode, full_path, buf, inode->i_sb,
426  					 xid, fid);
427  
428  	if (rc) {
429  		server->ops->close(xid, tcon, fid);
430  		if (rc == -ESTALE)
431  			rc = -EOPENSTALE;
432  	}
433  
434  	return rc;
435  }
436  
437  static bool
cifs_has_mand_locks(struct cifsInodeInfo * cinode)438  cifs_has_mand_locks(struct cifsInodeInfo *cinode)
439  {
440  	struct cifs_fid_locks *cur;
441  	bool has_locks = false;
442  
443  	down_read(&cinode->lock_sem);
444  	list_for_each_entry(cur, &cinode->llist, llist) {
445  		if (!list_empty(&cur->locks)) {
446  			has_locks = true;
447  			break;
448  		}
449  	}
450  	up_read(&cinode->lock_sem);
451  	return has_locks;
452  }
453  
454  void
cifs_down_write(struct rw_semaphore * sem)455  cifs_down_write(struct rw_semaphore *sem)
456  {
457  	while (!down_write_trylock(sem))
458  		msleep(10);
459  }
460  
461  static void cifsFileInfo_put_work(struct work_struct *work);
462  void serverclose_work(struct work_struct *work);
463  
cifs_new_fileinfo(struct cifs_fid * fid,struct file * file,struct tcon_link * tlink,__u32 oplock,const char * symlink_target)464  struct cifsFileInfo *cifs_new_fileinfo(struct cifs_fid *fid, struct file *file,
465  				       struct tcon_link *tlink, __u32 oplock,
466  				       const char *symlink_target)
467  {
468  	struct dentry *dentry = file_dentry(file);
469  	struct inode *inode = d_inode(dentry);
470  	struct cifsInodeInfo *cinode = CIFS_I(inode);
471  	struct cifsFileInfo *cfile;
472  	struct cifs_fid_locks *fdlocks;
473  	struct cifs_tcon *tcon = tlink_tcon(tlink);
474  	struct TCP_Server_Info *server = tcon->ses->server;
475  
476  	cfile = kzalloc(sizeof(struct cifsFileInfo), GFP_KERNEL);
477  	if (cfile == NULL)
478  		return cfile;
479  
480  	fdlocks = kzalloc(sizeof(struct cifs_fid_locks), GFP_KERNEL);
481  	if (!fdlocks) {
482  		kfree(cfile);
483  		return NULL;
484  	}
485  
486  	if (symlink_target) {
487  		cfile->symlink_target = kstrdup(symlink_target, GFP_KERNEL);
488  		if (!cfile->symlink_target) {
489  			kfree(fdlocks);
490  			kfree(cfile);
491  			return NULL;
492  		}
493  	}
494  
495  	INIT_LIST_HEAD(&fdlocks->locks);
496  	fdlocks->cfile = cfile;
497  	cfile->llist = fdlocks;
498  
499  	cfile->count = 1;
500  	cfile->pid = current->tgid;
501  	cfile->uid = current_fsuid();
502  	cfile->dentry = dget(dentry);
503  	cfile->f_flags = file->f_flags;
504  	cfile->invalidHandle = false;
505  	cfile->deferred_close_scheduled = false;
506  	cfile->tlink = cifs_get_tlink(tlink);
507  	INIT_WORK(&cfile->oplock_break, cifs_oplock_break);
508  	INIT_WORK(&cfile->put, cifsFileInfo_put_work);
509  	INIT_WORK(&cfile->serverclose, serverclose_work);
510  	INIT_DELAYED_WORK(&cfile->deferred, smb2_deferred_work_close);
511  	mutex_init(&cfile->fh_mutex);
512  	spin_lock_init(&cfile->file_info_lock);
513  
514  	cifs_sb_active(inode->i_sb);
515  
516  	/*
517  	 * If the server returned a read oplock and we have mandatory brlocks,
518  	 * set oplock level to None.
519  	 */
520  	if (server->ops->is_read_op(oplock) && cifs_has_mand_locks(cinode)) {
521  		cifs_dbg(FYI, "Reset oplock val from read to None due to mand locks\n");
522  		oplock = 0;
523  	}
524  
525  	cifs_down_write(&cinode->lock_sem);
526  	list_add(&fdlocks->llist, &cinode->llist);
527  	up_write(&cinode->lock_sem);
528  
529  	spin_lock(&tcon->open_file_lock);
530  	if (fid->pending_open->oplock != CIFS_OPLOCK_NO_CHANGE && oplock)
531  		oplock = fid->pending_open->oplock;
532  	list_del(&fid->pending_open->olist);
533  
534  	fid->purge_cache = false;
535  	server->ops->set_fid(cfile, fid, oplock);
536  
537  	list_add(&cfile->tlist, &tcon->openFileList);
538  	atomic_inc(&tcon->num_local_opens);
539  
540  	/* if readable file instance put first in list*/
541  	spin_lock(&cinode->open_file_lock);
542  	if (file->f_mode & FMODE_READ)
543  		list_add(&cfile->flist, &cinode->openFileList);
544  	else
545  		list_add_tail(&cfile->flist, &cinode->openFileList);
546  	spin_unlock(&cinode->open_file_lock);
547  	spin_unlock(&tcon->open_file_lock);
548  
549  	if (fid->purge_cache)
550  		cifs_zap_mapping(inode);
551  
552  	file->private_data = cfile;
553  	return cfile;
554  }
555  
556  struct cifsFileInfo *
cifsFileInfo_get(struct cifsFileInfo * cifs_file)557  cifsFileInfo_get(struct cifsFileInfo *cifs_file)
558  {
559  	spin_lock(&cifs_file->file_info_lock);
560  	cifsFileInfo_get_locked(cifs_file);
561  	spin_unlock(&cifs_file->file_info_lock);
562  	return cifs_file;
563  }
564  
cifsFileInfo_put_final(struct cifsFileInfo * cifs_file)565  static void cifsFileInfo_put_final(struct cifsFileInfo *cifs_file)
566  {
567  	struct inode *inode = d_inode(cifs_file->dentry);
568  	struct cifsInodeInfo *cifsi = CIFS_I(inode);
569  	struct cifsLockInfo *li, *tmp;
570  	struct super_block *sb = inode->i_sb;
571  
572  	/*
573  	 * Delete any outstanding lock records. We'll lose them when the file
574  	 * is closed anyway.
575  	 */
576  	cifs_down_write(&cifsi->lock_sem);
577  	list_for_each_entry_safe(li, tmp, &cifs_file->llist->locks, llist) {
578  		list_del(&li->llist);
579  		cifs_del_lock_waiters(li);
580  		kfree(li);
581  	}
582  	list_del(&cifs_file->llist->llist);
583  	kfree(cifs_file->llist);
584  	up_write(&cifsi->lock_sem);
585  
586  	cifs_put_tlink(cifs_file->tlink);
587  	dput(cifs_file->dentry);
588  	cifs_sb_deactive(sb);
589  	kfree(cifs_file->symlink_target);
590  	kfree(cifs_file);
591  }
592  
cifsFileInfo_put_work(struct work_struct * work)593  static void cifsFileInfo_put_work(struct work_struct *work)
594  {
595  	struct cifsFileInfo *cifs_file = container_of(work,
596  			struct cifsFileInfo, put);
597  
598  	cifsFileInfo_put_final(cifs_file);
599  }
600  
serverclose_work(struct work_struct * work)601  void serverclose_work(struct work_struct *work)
602  {
603  	struct cifsFileInfo *cifs_file = container_of(work,
604  			struct cifsFileInfo, serverclose);
605  
606  	struct cifs_tcon *tcon = tlink_tcon(cifs_file->tlink);
607  
608  	struct TCP_Server_Info *server = tcon->ses->server;
609  	int rc = 0;
610  	int retries = 0;
611  	int MAX_RETRIES = 4;
612  
613  	do {
614  		if (server->ops->close_getattr)
615  			rc = server->ops->close_getattr(0, tcon, cifs_file);
616  		else if (server->ops->close)
617  			rc = server->ops->close(0, tcon, &cifs_file->fid);
618  
619  		if (rc == -EBUSY || rc == -EAGAIN) {
620  			retries++;
621  			msleep(250);
622  		}
623  	} while ((rc == -EBUSY || rc == -EAGAIN) && (retries < MAX_RETRIES)
624  	);
625  
626  	if (retries == MAX_RETRIES)
627  		pr_warn("Serverclose failed %d times, giving up\n", MAX_RETRIES);
628  
629  	if (cifs_file->offload)
630  		queue_work(fileinfo_put_wq, &cifs_file->put);
631  	else
632  		cifsFileInfo_put_final(cifs_file);
633  }
634  
635  /**
636   * cifsFileInfo_put - release a reference of file priv data
637   *
638   * Always potentially wait for oplock handler. See _cifsFileInfo_put().
639   *
640   * @cifs_file:	cifs/smb3 specific info (eg refcounts) for an open file
641   */
cifsFileInfo_put(struct cifsFileInfo * cifs_file)642  void cifsFileInfo_put(struct cifsFileInfo *cifs_file)
643  {
644  	_cifsFileInfo_put(cifs_file, true, true);
645  }
646  
647  /**
648   * _cifsFileInfo_put - release a reference of file priv data
649   *
650   * This may involve closing the filehandle @cifs_file out on the
651   * server. Must be called without holding tcon->open_file_lock,
652   * cinode->open_file_lock and cifs_file->file_info_lock.
653   *
654   * If @wait_for_oplock_handler is true and we are releasing the last
655   * reference, wait for any running oplock break handler of the file
656   * and cancel any pending one.
657   *
658   * @cifs_file:	cifs/smb3 specific info (eg refcounts) for an open file
659   * @wait_oplock_handler: must be false if called from oplock_break_handler
660   * @offload:	not offloaded on close and oplock breaks
661   *
662   */
_cifsFileInfo_put(struct cifsFileInfo * cifs_file,bool wait_oplock_handler,bool offload)663  void _cifsFileInfo_put(struct cifsFileInfo *cifs_file,
664  		       bool wait_oplock_handler, bool offload)
665  {
666  	struct inode *inode = d_inode(cifs_file->dentry);
667  	struct cifs_tcon *tcon = tlink_tcon(cifs_file->tlink);
668  	struct TCP_Server_Info *server = tcon->ses->server;
669  	struct cifsInodeInfo *cifsi = CIFS_I(inode);
670  	struct super_block *sb = inode->i_sb;
671  	struct cifs_sb_info *cifs_sb = CIFS_SB(sb);
672  	struct cifs_fid fid = {};
673  	struct cifs_pending_open open;
674  	bool oplock_break_cancelled;
675  	bool serverclose_offloaded = false;
676  
677  	spin_lock(&tcon->open_file_lock);
678  	spin_lock(&cifsi->open_file_lock);
679  	spin_lock(&cifs_file->file_info_lock);
680  
681  	cifs_file->offload = offload;
682  	if (--cifs_file->count > 0) {
683  		spin_unlock(&cifs_file->file_info_lock);
684  		spin_unlock(&cifsi->open_file_lock);
685  		spin_unlock(&tcon->open_file_lock);
686  		return;
687  	}
688  	spin_unlock(&cifs_file->file_info_lock);
689  
690  	if (server->ops->get_lease_key)
691  		server->ops->get_lease_key(inode, &fid);
692  
693  	/* store open in pending opens to make sure we don't miss lease break */
694  	cifs_add_pending_open_locked(&fid, cifs_file->tlink, &open);
695  
696  	/* remove it from the lists */
697  	list_del(&cifs_file->flist);
698  	list_del(&cifs_file->tlist);
699  	atomic_dec(&tcon->num_local_opens);
700  
701  	if (list_empty(&cifsi->openFileList)) {
702  		cifs_dbg(FYI, "closing last open instance for inode %p\n",
703  			 d_inode(cifs_file->dentry));
704  		/*
705  		 * In strict cache mode we need invalidate mapping on the last
706  		 * close  because it may cause a error when we open this file
707  		 * again and get at least level II oplock.
708  		 */
709  		if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_STRICT_IO)
710  			set_bit(CIFS_INO_INVALID_MAPPING, &cifsi->flags);
711  		cifs_set_oplock_level(cifsi, 0);
712  	}
713  
714  	spin_unlock(&cifsi->open_file_lock);
715  	spin_unlock(&tcon->open_file_lock);
716  
717  	oplock_break_cancelled = wait_oplock_handler ?
718  		cancel_work_sync(&cifs_file->oplock_break) : false;
719  
720  	if (!tcon->need_reconnect && !cifs_file->invalidHandle) {
721  		struct TCP_Server_Info *server = tcon->ses->server;
722  		unsigned int xid;
723  		int rc = 0;
724  
725  		xid = get_xid();
726  		if (server->ops->close_getattr)
727  			rc = server->ops->close_getattr(xid, tcon, cifs_file);
728  		else if (server->ops->close)
729  			rc = server->ops->close(xid, tcon, &cifs_file->fid);
730  		_free_xid(xid);
731  
732  		if (rc == -EBUSY || rc == -EAGAIN) {
733  			// Server close failed, hence offloading it as an async op
734  			queue_work(serverclose_wq, &cifs_file->serverclose);
735  			serverclose_offloaded = true;
736  		}
737  	}
738  
739  	if (oplock_break_cancelled)
740  		cifs_done_oplock_break(cifsi);
741  
742  	cifs_del_pending_open(&open);
743  
744  	// if serverclose has been offloaded to wq (on failure), it will
745  	// handle offloading put as well. If serverclose not offloaded,
746  	// we need to handle offloading put here.
747  	if (!serverclose_offloaded) {
748  		if (offload)
749  			queue_work(fileinfo_put_wq, &cifs_file->put);
750  		else
751  			cifsFileInfo_put_final(cifs_file);
752  	}
753  }
754  
cifs_open(struct inode * inode,struct file * file)755  int cifs_open(struct inode *inode, struct file *file)
756  
757  {
758  	int rc = -EACCES;
759  	unsigned int xid;
760  	__u32 oplock;
761  	struct cifs_sb_info *cifs_sb;
762  	struct TCP_Server_Info *server;
763  	struct cifs_tcon *tcon;
764  	struct tcon_link *tlink;
765  	struct cifsFileInfo *cfile = NULL;
766  	void *page;
767  	const char *full_path;
768  	bool posix_open_ok = false;
769  	struct cifs_fid fid = {};
770  	struct cifs_pending_open open;
771  	struct cifs_open_info_data data = {};
772  
773  	xid = get_xid();
774  
775  	cifs_sb = CIFS_SB(inode->i_sb);
776  	if (unlikely(cifs_forced_shutdown(cifs_sb))) {
777  		free_xid(xid);
778  		return -EIO;
779  	}
780  
781  	tlink = cifs_sb_tlink(cifs_sb);
782  	if (IS_ERR(tlink)) {
783  		free_xid(xid);
784  		return PTR_ERR(tlink);
785  	}
786  	tcon = tlink_tcon(tlink);
787  	server = tcon->ses->server;
788  
789  	page = alloc_dentry_path();
790  	full_path = build_path_from_dentry(file_dentry(file), page);
791  	if (IS_ERR(full_path)) {
792  		rc = PTR_ERR(full_path);
793  		goto out;
794  	}
795  
796  	cifs_dbg(FYI, "inode = 0x%p file flags are 0x%x for %s\n",
797  		 inode, file->f_flags, full_path);
798  
799  	if (file->f_flags & O_DIRECT &&
800  	    cifs_sb->mnt_cifs_flags & CIFS_MOUNT_STRICT_IO) {
801  		if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_NO_BRL)
802  			file->f_op = &cifs_file_direct_nobrl_ops;
803  		else
804  			file->f_op = &cifs_file_direct_ops;
805  	}
806  
807  	/* Get the cached handle as SMB2 close is deferred */
808  	rc = cifs_get_readable_path(tcon, full_path, &cfile);
809  	if (rc == 0) {
810  		if (file->f_flags == cfile->f_flags) {
811  			file->private_data = cfile;
812  			spin_lock(&CIFS_I(inode)->deferred_lock);
813  			cifs_del_deferred_close(cfile);
814  			spin_unlock(&CIFS_I(inode)->deferred_lock);
815  			goto use_cache;
816  		} else {
817  			_cifsFileInfo_put(cfile, true, false);
818  		}
819  	}
820  
821  	if (server->oplocks)
822  		oplock = REQ_OPLOCK;
823  	else
824  		oplock = 0;
825  
826  #ifdef CONFIG_CIFS_ALLOW_INSECURE_LEGACY
827  	if (!tcon->broken_posix_open && tcon->unix_ext &&
828  	    cap_unix(tcon->ses) && (CIFS_UNIX_POSIX_PATH_OPS_CAP &
829  				le64_to_cpu(tcon->fsUnixInfo.Capability))) {
830  		/* can not refresh inode info since size could be stale */
831  		rc = cifs_posix_open(full_path, &inode, inode->i_sb,
832  				cifs_sb->ctx->file_mode /* ignored */,
833  				file->f_flags, &oplock, &fid.netfid, xid);
834  		if (rc == 0) {
835  			cifs_dbg(FYI, "posix open succeeded\n");
836  			posix_open_ok = true;
837  		} else if ((rc == -EINVAL) || (rc == -EOPNOTSUPP)) {
838  			if (tcon->ses->serverNOS)
839  				cifs_dbg(VFS, "server %s of type %s returned unexpected error on SMB posix open, disabling posix open support. Check if server update available.\n",
840  					 tcon->ses->ip_addr,
841  					 tcon->ses->serverNOS);
842  			tcon->broken_posix_open = true;
843  		} else if ((rc != -EIO) && (rc != -EREMOTE) &&
844  			 (rc != -EOPNOTSUPP)) /* path not found or net err */
845  			goto out;
846  		/*
847  		 * Else fallthrough to retry open the old way on network i/o
848  		 * or DFS errors.
849  		 */
850  	}
851  #endif /* CONFIG_CIFS_ALLOW_INSECURE_LEGACY */
852  
853  	if (server->ops->get_lease_key)
854  		server->ops->get_lease_key(inode, &fid);
855  
856  	cifs_add_pending_open(&fid, tlink, &open);
857  
858  	if (!posix_open_ok) {
859  		if (server->ops->get_lease_key)
860  			server->ops->get_lease_key(inode, &fid);
861  
862  		rc = cifs_nt_open(full_path, inode, cifs_sb, tcon, file->f_flags, &oplock, &fid,
863  				  xid, &data);
864  		if (rc) {
865  			cifs_del_pending_open(&open);
866  			goto out;
867  		}
868  	}
869  
870  	cfile = cifs_new_fileinfo(&fid, file, tlink, oplock, data.symlink_target);
871  	if (cfile == NULL) {
872  		if (server->ops->close)
873  			server->ops->close(xid, tcon, &fid);
874  		cifs_del_pending_open(&open);
875  		rc = -ENOMEM;
876  		goto out;
877  	}
878  
879  #ifdef CONFIG_CIFS_ALLOW_INSECURE_LEGACY
880  	if ((oplock & CIFS_CREATE_ACTION) && !posix_open_ok && tcon->unix_ext) {
881  		/*
882  		 * Time to set mode which we can not set earlier due to
883  		 * problems creating new read-only files.
884  		 */
885  		struct cifs_unix_set_info_args args = {
886  			.mode	= inode->i_mode,
887  			.uid	= INVALID_UID, /* no change */
888  			.gid	= INVALID_GID, /* no change */
889  			.ctime	= NO_CHANGE_64,
890  			.atime	= NO_CHANGE_64,
891  			.mtime	= NO_CHANGE_64,
892  			.device	= 0,
893  		};
894  		CIFSSMBUnixSetFileInfo(xid, tcon, &args, fid.netfid,
895  				       cfile->pid);
896  	}
897  #endif /* CONFIG_CIFS_ALLOW_INSECURE_LEGACY */
898  
899  use_cache:
900  	fscache_use_cookie(cifs_inode_cookie(file_inode(file)),
901  			   file->f_mode & FMODE_WRITE);
902  	if (!(file->f_flags & O_DIRECT))
903  		goto out;
904  	if ((file->f_flags & (O_ACCMODE | O_APPEND)) == O_RDONLY)
905  		goto out;
906  	cifs_invalidate_cache(file_inode(file), FSCACHE_INVAL_DIO_WRITE);
907  
908  out:
909  	free_dentry_path(page);
910  	free_xid(xid);
911  	cifs_put_tlink(tlink);
912  	cifs_free_open_info(&data);
913  	return rc;
914  }
915  
916  #ifdef CONFIG_CIFS_ALLOW_INSECURE_LEGACY
917  static int cifs_push_posix_locks(struct cifsFileInfo *cfile);
918  #endif /* CONFIG_CIFS_ALLOW_INSECURE_LEGACY */
919  
920  /*
921   * Try to reacquire byte range locks that were released when session
922   * to server was lost.
923   */
924  static int
cifs_relock_file(struct cifsFileInfo * cfile)925  cifs_relock_file(struct cifsFileInfo *cfile)
926  {
927  	struct cifsInodeInfo *cinode = CIFS_I(d_inode(cfile->dentry));
928  	struct cifs_tcon *tcon = tlink_tcon(cfile->tlink);
929  	int rc = 0;
930  #ifdef CONFIG_CIFS_ALLOW_INSECURE_LEGACY
931  	struct cifs_sb_info *cifs_sb = CIFS_SB(cfile->dentry->d_sb);
932  #endif /* CONFIG_CIFS_ALLOW_INSECURE_LEGACY */
933  
934  	down_read_nested(&cinode->lock_sem, SINGLE_DEPTH_NESTING);
935  	if (cinode->can_cache_brlcks) {
936  		/* can cache locks - no need to relock */
937  		up_read(&cinode->lock_sem);
938  		return rc;
939  	}
940  
941  #ifdef CONFIG_CIFS_ALLOW_INSECURE_LEGACY
942  	if (cap_unix(tcon->ses) &&
943  	    (CIFS_UNIX_FCNTL_CAP & le64_to_cpu(tcon->fsUnixInfo.Capability)) &&
944  	    ((cifs_sb->mnt_cifs_flags & CIFS_MOUNT_NOPOSIXBRL) == 0))
945  		rc = cifs_push_posix_locks(cfile);
946  	else
947  #endif /* CONFIG_CIFS_ALLOW_INSECURE_LEGACY */
948  		rc = tcon->ses->server->ops->push_mand_locks(cfile);
949  
950  	up_read(&cinode->lock_sem);
951  	return rc;
952  }
953  
954  static int
cifs_reopen_file(struct cifsFileInfo * cfile,bool can_flush)955  cifs_reopen_file(struct cifsFileInfo *cfile, bool can_flush)
956  {
957  	int rc = -EACCES;
958  	unsigned int xid;
959  	__u32 oplock;
960  	struct cifs_sb_info *cifs_sb;
961  	struct cifs_tcon *tcon;
962  	struct TCP_Server_Info *server;
963  	struct cifsInodeInfo *cinode;
964  	struct inode *inode;
965  	void *page;
966  	const char *full_path;
967  	int desired_access;
968  	int disposition = FILE_OPEN;
969  	int create_options = CREATE_NOT_DIR;
970  	struct cifs_open_parms oparms;
971  	int rdwr_for_fscache = 0;
972  
973  	xid = get_xid();
974  	mutex_lock(&cfile->fh_mutex);
975  	if (!cfile->invalidHandle) {
976  		mutex_unlock(&cfile->fh_mutex);
977  		free_xid(xid);
978  		return 0;
979  	}
980  
981  	inode = d_inode(cfile->dentry);
982  	cifs_sb = CIFS_SB(inode->i_sb);
983  	tcon = tlink_tcon(cfile->tlink);
984  	server = tcon->ses->server;
985  
986  	/*
987  	 * Can not grab rename sem here because various ops, including those
988  	 * that already have the rename sem can end up causing writepage to get
989  	 * called and if the server was down that means we end up here, and we
990  	 * can never tell if the caller already has the rename_sem.
991  	 */
992  	page = alloc_dentry_path();
993  	full_path = build_path_from_dentry(cfile->dentry, page);
994  	if (IS_ERR(full_path)) {
995  		mutex_unlock(&cfile->fh_mutex);
996  		free_dentry_path(page);
997  		free_xid(xid);
998  		return PTR_ERR(full_path);
999  	}
1000  
1001  	cifs_dbg(FYI, "inode = 0x%p file flags 0x%x for %s\n",
1002  		 inode, cfile->f_flags, full_path);
1003  
1004  	if (tcon->ses->server->oplocks)
1005  		oplock = REQ_OPLOCK;
1006  	else
1007  		oplock = 0;
1008  
1009  #ifdef CONFIG_CIFS_ALLOW_INSECURE_LEGACY
1010  	if (tcon->unix_ext && cap_unix(tcon->ses) &&
1011  	    (CIFS_UNIX_POSIX_PATH_OPS_CAP &
1012  				le64_to_cpu(tcon->fsUnixInfo.Capability))) {
1013  		/*
1014  		 * O_CREAT, O_EXCL and O_TRUNC already had their effect on the
1015  		 * original open. Must mask them off for a reopen.
1016  		 */
1017  		unsigned int oflags = cfile->f_flags &
1018  						~(O_CREAT | O_EXCL | O_TRUNC);
1019  
1020  		rc = cifs_posix_open(full_path, NULL, inode->i_sb,
1021  				     cifs_sb->ctx->file_mode /* ignored */,
1022  				     oflags, &oplock, &cfile->fid.netfid, xid);
1023  		if (rc == 0) {
1024  			cifs_dbg(FYI, "posix reopen succeeded\n");
1025  			oparms.reconnect = true;
1026  			goto reopen_success;
1027  		}
1028  		/*
1029  		 * fallthrough to retry open the old way on errors, especially
1030  		 * in the reconnect path it is important to retry hard
1031  		 */
1032  	}
1033  #endif /* CONFIG_CIFS_ALLOW_INSECURE_LEGACY */
1034  
1035  	/* If we're caching, we need to be able to fill in around partial writes. */
1036  	if (cifs_fscache_enabled(inode) && (cfile->f_flags & O_ACCMODE) == O_WRONLY)
1037  		rdwr_for_fscache = 1;
1038  
1039  	desired_access = cifs_convert_flags(cfile->f_flags, rdwr_for_fscache);
1040  
1041  	/* O_SYNC also has bit for O_DSYNC so following check picks up either */
1042  	if (cfile->f_flags & O_SYNC)
1043  		create_options |= CREATE_WRITE_THROUGH;
1044  
1045  	if (cfile->f_flags & O_DIRECT)
1046  		create_options |= CREATE_NO_BUFFER;
1047  
1048  	if (server->ops->get_lease_key)
1049  		server->ops->get_lease_key(inode, &cfile->fid);
1050  
1051  retry_open:
1052  	oparms = (struct cifs_open_parms) {
1053  		.tcon = tcon,
1054  		.cifs_sb = cifs_sb,
1055  		.desired_access = desired_access,
1056  		.create_options = cifs_create_options(cifs_sb, create_options),
1057  		.disposition = disposition,
1058  		.path = full_path,
1059  		.fid = &cfile->fid,
1060  		.reconnect = true,
1061  	};
1062  
1063  	/*
1064  	 * Can not refresh inode by passing in file_info buf to be returned by
1065  	 * ops->open and then calling get_inode_info with returned buf since
1066  	 * file might have write behind data that needs to be flushed and server
1067  	 * version of file size can be stale. If we knew for sure that inode was
1068  	 * not dirty locally we could do this.
1069  	 */
1070  	rc = server->ops->open(xid, &oparms, &oplock, NULL);
1071  	if (rc == -ENOENT && oparms.reconnect == false) {
1072  		/* durable handle timeout is expired - open the file again */
1073  		rc = server->ops->open(xid, &oparms, &oplock, NULL);
1074  		/* indicate that we need to relock the file */
1075  		oparms.reconnect = true;
1076  	}
1077  	if (rc == -EACCES && rdwr_for_fscache == 1) {
1078  		desired_access = cifs_convert_flags(cfile->f_flags, 0);
1079  		rdwr_for_fscache = 2;
1080  		goto retry_open;
1081  	}
1082  
1083  	if (rc) {
1084  		mutex_unlock(&cfile->fh_mutex);
1085  		cifs_dbg(FYI, "cifs_reopen returned 0x%x\n", rc);
1086  		cifs_dbg(FYI, "oplock: %d\n", oplock);
1087  		goto reopen_error_exit;
1088  	}
1089  
1090  	if (rdwr_for_fscache == 2)
1091  		cifs_invalidate_cache(inode, FSCACHE_INVAL_DIO_WRITE);
1092  
1093  #ifdef CONFIG_CIFS_ALLOW_INSECURE_LEGACY
1094  reopen_success:
1095  #endif /* CONFIG_CIFS_ALLOW_INSECURE_LEGACY */
1096  	cfile->invalidHandle = false;
1097  	mutex_unlock(&cfile->fh_mutex);
1098  	cinode = CIFS_I(inode);
1099  
1100  	if (can_flush) {
1101  		rc = filemap_write_and_wait(inode->i_mapping);
1102  		if (!is_interrupt_error(rc))
1103  			mapping_set_error(inode->i_mapping, rc);
1104  
1105  		if (tcon->posix_extensions) {
1106  			rc = smb311_posix_get_inode_info(&inode, full_path,
1107  							 NULL, inode->i_sb, xid);
1108  		} else if (tcon->unix_ext) {
1109  			rc = cifs_get_inode_info_unix(&inode, full_path,
1110  						      inode->i_sb, xid);
1111  		} else {
1112  			rc = cifs_get_inode_info(&inode, full_path, NULL,
1113  						 inode->i_sb, xid, NULL);
1114  		}
1115  	}
1116  	/*
1117  	 * Else we are writing out data to server already and could deadlock if
1118  	 * we tried to flush data, and since we do not know if we have data that
1119  	 * would invalidate the current end of file on the server we can not go
1120  	 * to the server to get the new inode info.
1121  	 */
1122  
1123  	/*
1124  	 * If the server returned a read oplock and we have mandatory brlocks,
1125  	 * set oplock level to None.
1126  	 */
1127  	if (server->ops->is_read_op(oplock) && cifs_has_mand_locks(cinode)) {
1128  		cifs_dbg(FYI, "Reset oplock val from read to None due to mand locks\n");
1129  		oplock = 0;
1130  	}
1131  
1132  	server->ops->set_fid(cfile, &cfile->fid, oplock);
1133  	if (oparms.reconnect)
1134  		cifs_relock_file(cfile);
1135  
1136  reopen_error_exit:
1137  	free_dentry_path(page);
1138  	free_xid(xid);
1139  	return rc;
1140  }
1141  
smb2_deferred_work_close(struct work_struct * work)1142  void smb2_deferred_work_close(struct work_struct *work)
1143  {
1144  	struct cifsFileInfo *cfile = container_of(work,
1145  			struct cifsFileInfo, deferred.work);
1146  
1147  	spin_lock(&CIFS_I(d_inode(cfile->dentry))->deferred_lock);
1148  	cifs_del_deferred_close(cfile);
1149  	cfile->deferred_close_scheduled = false;
1150  	spin_unlock(&CIFS_I(d_inode(cfile->dentry))->deferred_lock);
1151  	_cifsFileInfo_put(cfile, true, false);
1152  }
1153  
1154  static bool
smb2_can_defer_close(struct inode * inode,struct cifs_deferred_close * dclose)1155  smb2_can_defer_close(struct inode *inode, struct cifs_deferred_close *dclose)
1156  {
1157  	struct cifs_sb_info *cifs_sb = CIFS_SB(inode->i_sb);
1158  	struct cifsInodeInfo *cinode = CIFS_I(inode);
1159  
1160  	return (cifs_sb->ctx->closetimeo && cinode->lease_granted && dclose &&
1161  			(cinode->oplock == CIFS_CACHE_RHW_FLG ||
1162  			 cinode->oplock == CIFS_CACHE_RH_FLG) &&
1163  			!test_bit(CIFS_INO_CLOSE_ON_LOCK, &cinode->flags));
1164  
1165  }
1166  
cifs_close(struct inode * inode,struct file * file)1167  int cifs_close(struct inode *inode, struct file *file)
1168  {
1169  	struct cifsFileInfo *cfile;
1170  	struct cifsInodeInfo *cinode = CIFS_I(inode);
1171  	struct cifs_sb_info *cifs_sb = CIFS_SB(inode->i_sb);
1172  	struct cifs_deferred_close *dclose;
1173  
1174  	cifs_fscache_unuse_inode_cookie(inode, file->f_mode & FMODE_WRITE);
1175  
1176  	if (file->private_data != NULL) {
1177  		cfile = file->private_data;
1178  		file->private_data = NULL;
1179  		dclose = kmalloc(sizeof(struct cifs_deferred_close), GFP_KERNEL);
1180  		if ((cfile->status_file_deleted == false) &&
1181  		    (smb2_can_defer_close(inode, dclose))) {
1182  			if (test_and_clear_bit(CIFS_INO_MODIFIED_ATTR, &cinode->flags)) {
1183  				inode_set_mtime_to_ts(inode,
1184  						      inode_set_ctime_current(inode));
1185  			}
1186  			spin_lock(&cinode->deferred_lock);
1187  			cifs_add_deferred_close(cfile, dclose);
1188  			if (cfile->deferred_close_scheduled &&
1189  			    delayed_work_pending(&cfile->deferred)) {
1190  				/*
1191  				 * If there is no pending work, mod_delayed_work queues new work.
1192  				 * So, Increase the ref count to avoid use-after-free.
1193  				 */
1194  				if (!mod_delayed_work(deferredclose_wq,
1195  						&cfile->deferred, cifs_sb->ctx->closetimeo))
1196  					cifsFileInfo_get(cfile);
1197  			} else {
1198  				/* Deferred close for files */
1199  				queue_delayed_work(deferredclose_wq,
1200  						&cfile->deferred, cifs_sb->ctx->closetimeo);
1201  				cfile->deferred_close_scheduled = true;
1202  				spin_unlock(&cinode->deferred_lock);
1203  				return 0;
1204  			}
1205  			spin_unlock(&cinode->deferred_lock);
1206  			_cifsFileInfo_put(cfile, true, false);
1207  		} else {
1208  			_cifsFileInfo_put(cfile, true, false);
1209  			kfree(dclose);
1210  		}
1211  	}
1212  
1213  	/* return code from the ->release op is always ignored */
1214  	return 0;
1215  }
1216  
1217  void
cifs_reopen_persistent_handles(struct cifs_tcon * tcon)1218  cifs_reopen_persistent_handles(struct cifs_tcon *tcon)
1219  {
1220  	struct cifsFileInfo *open_file, *tmp;
1221  	struct list_head tmp_list;
1222  
1223  	if (!tcon->use_persistent || !tcon->need_reopen_files)
1224  		return;
1225  
1226  	tcon->need_reopen_files = false;
1227  
1228  	cifs_dbg(FYI, "Reopen persistent handles\n");
1229  	INIT_LIST_HEAD(&tmp_list);
1230  
1231  	/* list all files open on tree connection, reopen resilient handles  */
1232  	spin_lock(&tcon->open_file_lock);
1233  	list_for_each_entry(open_file, &tcon->openFileList, tlist) {
1234  		if (!open_file->invalidHandle)
1235  			continue;
1236  		cifsFileInfo_get(open_file);
1237  		list_add_tail(&open_file->rlist, &tmp_list);
1238  	}
1239  	spin_unlock(&tcon->open_file_lock);
1240  
1241  	list_for_each_entry_safe(open_file, tmp, &tmp_list, rlist) {
1242  		if (cifs_reopen_file(open_file, false /* do not flush */))
1243  			tcon->need_reopen_files = true;
1244  		list_del_init(&open_file->rlist);
1245  		cifsFileInfo_put(open_file);
1246  	}
1247  }
1248  
cifs_closedir(struct inode * inode,struct file * file)1249  int cifs_closedir(struct inode *inode, struct file *file)
1250  {
1251  	int rc = 0;
1252  	unsigned int xid;
1253  	struct cifsFileInfo *cfile = file->private_data;
1254  	struct cifs_tcon *tcon;
1255  	struct TCP_Server_Info *server;
1256  	char *buf;
1257  
1258  	cifs_dbg(FYI, "Closedir inode = 0x%p\n", inode);
1259  
1260  	if (cfile == NULL)
1261  		return rc;
1262  
1263  	xid = get_xid();
1264  	tcon = tlink_tcon(cfile->tlink);
1265  	server = tcon->ses->server;
1266  
1267  	cifs_dbg(FYI, "Freeing private data in close dir\n");
1268  	spin_lock(&cfile->file_info_lock);
1269  	if (server->ops->dir_needs_close(cfile)) {
1270  		cfile->invalidHandle = true;
1271  		spin_unlock(&cfile->file_info_lock);
1272  		if (server->ops->close_dir)
1273  			rc = server->ops->close_dir(xid, tcon, &cfile->fid);
1274  		else
1275  			rc = -ENOSYS;
1276  		cifs_dbg(FYI, "Closing uncompleted readdir with rc %d\n", rc);
1277  		/* not much we can do if it fails anyway, ignore rc */
1278  		rc = 0;
1279  	} else
1280  		spin_unlock(&cfile->file_info_lock);
1281  
1282  	buf = cfile->srch_inf.ntwrk_buf_start;
1283  	if (buf) {
1284  		cifs_dbg(FYI, "closedir free smb buf in srch struct\n");
1285  		cfile->srch_inf.ntwrk_buf_start = NULL;
1286  		if (cfile->srch_inf.smallBuf)
1287  			cifs_small_buf_release(buf);
1288  		else
1289  			cifs_buf_release(buf);
1290  	}
1291  
1292  	cifs_put_tlink(cfile->tlink);
1293  	kfree(file->private_data);
1294  	file->private_data = NULL;
1295  	/* BB can we lock the filestruct while this is going on? */
1296  	free_xid(xid);
1297  	return rc;
1298  }
1299  
1300  static struct cifsLockInfo *
cifs_lock_init(__u64 offset,__u64 length,__u8 type,__u16 flags)1301  cifs_lock_init(__u64 offset, __u64 length, __u8 type, __u16 flags)
1302  {
1303  	struct cifsLockInfo *lock =
1304  		kmalloc(sizeof(struct cifsLockInfo), GFP_KERNEL);
1305  	if (!lock)
1306  		return lock;
1307  	lock->offset = offset;
1308  	lock->length = length;
1309  	lock->type = type;
1310  	lock->pid = current->tgid;
1311  	lock->flags = flags;
1312  	INIT_LIST_HEAD(&lock->blist);
1313  	init_waitqueue_head(&lock->block_q);
1314  	return lock;
1315  }
1316  
1317  void
cifs_del_lock_waiters(struct cifsLockInfo * lock)1318  cifs_del_lock_waiters(struct cifsLockInfo *lock)
1319  {
1320  	struct cifsLockInfo *li, *tmp;
1321  	list_for_each_entry_safe(li, tmp, &lock->blist, blist) {
1322  		list_del_init(&li->blist);
1323  		wake_up(&li->block_q);
1324  	}
1325  }
1326  
1327  #define CIFS_LOCK_OP	0
1328  #define CIFS_READ_OP	1
1329  #define CIFS_WRITE_OP	2
1330  
1331  /* @rw_check : 0 - no op, 1 - read, 2 - write */
1332  static bool
cifs_find_fid_lock_conflict(struct cifs_fid_locks * fdlocks,__u64 offset,__u64 length,__u8 type,__u16 flags,struct cifsFileInfo * cfile,struct cifsLockInfo ** conf_lock,int rw_check)1333  cifs_find_fid_lock_conflict(struct cifs_fid_locks *fdlocks, __u64 offset,
1334  			    __u64 length, __u8 type, __u16 flags,
1335  			    struct cifsFileInfo *cfile,
1336  			    struct cifsLockInfo **conf_lock, int rw_check)
1337  {
1338  	struct cifsLockInfo *li;
1339  	struct cifsFileInfo *cur_cfile = fdlocks->cfile;
1340  	struct TCP_Server_Info *server = tlink_tcon(cfile->tlink)->ses->server;
1341  
1342  	list_for_each_entry(li, &fdlocks->locks, llist) {
1343  		if (offset + length <= li->offset ||
1344  		    offset >= li->offset + li->length)
1345  			continue;
1346  		if (rw_check != CIFS_LOCK_OP && current->tgid == li->pid &&
1347  		    server->ops->compare_fids(cfile, cur_cfile)) {
1348  			/* shared lock prevents write op through the same fid */
1349  			if (!(li->type & server->vals->shared_lock_type) ||
1350  			    rw_check != CIFS_WRITE_OP)
1351  				continue;
1352  		}
1353  		if ((type & server->vals->shared_lock_type) &&
1354  		    ((server->ops->compare_fids(cfile, cur_cfile) &&
1355  		     current->tgid == li->pid) || type == li->type))
1356  			continue;
1357  		if (rw_check == CIFS_LOCK_OP &&
1358  		    (flags & FL_OFDLCK) && (li->flags & FL_OFDLCK) &&
1359  		    server->ops->compare_fids(cfile, cur_cfile))
1360  			continue;
1361  		if (conf_lock)
1362  			*conf_lock = li;
1363  		return true;
1364  	}
1365  	return false;
1366  }
1367  
1368  bool
cifs_find_lock_conflict(struct cifsFileInfo * cfile,__u64 offset,__u64 length,__u8 type,__u16 flags,struct cifsLockInfo ** conf_lock,int rw_check)1369  cifs_find_lock_conflict(struct cifsFileInfo *cfile, __u64 offset, __u64 length,
1370  			__u8 type, __u16 flags,
1371  			struct cifsLockInfo **conf_lock, int rw_check)
1372  {
1373  	bool rc = false;
1374  	struct cifs_fid_locks *cur;
1375  	struct cifsInodeInfo *cinode = CIFS_I(d_inode(cfile->dentry));
1376  
1377  	list_for_each_entry(cur, &cinode->llist, llist) {
1378  		rc = cifs_find_fid_lock_conflict(cur, offset, length, type,
1379  						 flags, cfile, conf_lock,
1380  						 rw_check);
1381  		if (rc)
1382  			break;
1383  	}
1384  
1385  	return rc;
1386  }
1387  
1388  /*
1389   * Check if there is another lock that prevents us to set the lock (mandatory
1390   * style). If such a lock exists, update the flock structure with its
1391   * properties. Otherwise, set the flock type to F_UNLCK if we can cache brlocks
1392   * or leave it the same if we can't. Returns 0 if we don't need to request to
1393   * the server or 1 otherwise.
1394   */
1395  static int
cifs_lock_test(struct cifsFileInfo * cfile,__u64 offset,__u64 length,__u8 type,struct file_lock * flock)1396  cifs_lock_test(struct cifsFileInfo *cfile, __u64 offset, __u64 length,
1397  	       __u8 type, struct file_lock *flock)
1398  {
1399  	int rc = 0;
1400  	struct cifsLockInfo *conf_lock;
1401  	struct cifsInodeInfo *cinode = CIFS_I(d_inode(cfile->dentry));
1402  	struct TCP_Server_Info *server = tlink_tcon(cfile->tlink)->ses->server;
1403  	bool exist;
1404  
1405  	down_read(&cinode->lock_sem);
1406  
1407  	exist = cifs_find_lock_conflict(cfile, offset, length, type,
1408  					flock->fl_flags, &conf_lock,
1409  					CIFS_LOCK_OP);
1410  	if (exist) {
1411  		flock->fl_start = conf_lock->offset;
1412  		flock->fl_end = conf_lock->offset + conf_lock->length - 1;
1413  		flock->fl_pid = conf_lock->pid;
1414  		if (conf_lock->type & server->vals->shared_lock_type)
1415  			flock->fl_type = F_RDLCK;
1416  		else
1417  			flock->fl_type = F_WRLCK;
1418  	} else if (!cinode->can_cache_brlcks)
1419  		rc = 1;
1420  	else
1421  		flock->fl_type = F_UNLCK;
1422  
1423  	up_read(&cinode->lock_sem);
1424  	return rc;
1425  }
1426  
1427  static void
cifs_lock_add(struct cifsFileInfo * cfile,struct cifsLockInfo * lock)1428  cifs_lock_add(struct cifsFileInfo *cfile, struct cifsLockInfo *lock)
1429  {
1430  	struct cifsInodeInfo *cinode = CIFS_I(d_inode(cfile->dentry));
1431  	cifs_down_write(&cinode->lock_sem);
1432  	list_add_tail(&lock->llist, &cfile->llist->locks);
1433  	up_write(&cinode->lock_sem);
1434  }
1435  
1436  /*
1437   * Set the byte-range lock (mandatory style). Returns:
1438   * 1) 0, if we set the lock and don't need to request to the server;
1439   * 2) 1, if no locks prevent us but we need to request to the server;
1440   * 3) -EACCES, if there is a lock that prevents us and wait is false.
1441   */
1442  static int
cifs_lock_add_if(struct cifsFileInfo * cfile,struct cifsLockInfo * lock,bool wait)1443  cifs_lock_add_if(struct cifsFileInfo *cfile, struct cifsLockInfo *lock,
1444  		 bool wait)
1445  {
1446  	struct cifsLockInfo *conf_lock;
1447  	struct cifsInodeInfo *cinode = CIFS_I(d_inode(cfile->dentry));
1448  	bool exist;
1449  	int rc = 0;
1450  
1451  try_again:
1452  	exist = false;
1453  	cifs_down_write(&cinode->lock_sem);
1454  
1455  	exist = cifs_find_lock_conflict(cfile, lock->offset, lock->length,
1456  					lock->type, lock->flags, &conf_lock,
1457  					CIFS_LOCK_OP);
1458  	if (!exist && cinode->can_cache_brlcks) {
1459  		list_add_tail(&lock->llist, &cfile->llist->locks);
1460  		up_write(&cinode->lock_sem);
1461  		return rc;
1462  	}
1463  
1464  	if (!exist)
1465  		rc = 1;
1466  	else if (!wait)
1467  		rc = -EACCES;
1468  	else {
1469  		list_add_tail(&lock->blist, &conf_lock->blist);
1470  		up_write(&cinode->lock_sem);
1471  		rc = wait_event_interruptible(lock->block_q,
1472  					(lock->blist.prev == &lock->blist) &&
1473  					(lock->blist.next == &lock->blist));
1474  		if (!rc)
1475  			goto try_again;
1476  		cifs_down_write(&cinode->lock_sem);
1477  		list_del_init(&lock->blist);
1478  	}
1479  
1480  	up_write(&cinode->lock_sem);
1481  	return rc;
1482  }
1483  
1484  #ifdef CONFIG_CIFS_ALLOW_INSECURE_LEGACY
1485  /*
1486   * Check if there is another lock that prevents us to set the lock (posix
1487   * style). If such a lock exists, update the flock structure with its
1488   * properties. Otherwise, set the flock type to F_UNLCK if we can cache brlocks
1489   * or leave it the same if we can't. Returns 0 if we don't need to request to
1490   * the server or 1 otherwise.
1491   */
1492  static int
cifs_posix_lock_test(struct file * file,struct file_lock * flock)1493  cifs_posix_lock_test(struct file *file, struct file_lock *flock)
1494  {
1495  	int rc = 0;
1496  	struct cifsInodeInfo *cinode = CIFS_I(file_inode(file));
1497  	unsigned char saved_type = flock->fl_type;
1498  
1499  	if ((flock->fl_flags & FL_POSIX) == 0)
1500  		return 1;
1501  
1502  	down_read(&cinode->lock_sem);
1503  	posix_test_lock(file, flock);
1504  
1505  	if (flock->fl_type == F_UNLCK && !cinode->can_cache_brlcks) {
1506  		flock->fl_type = saved_type;
1507  		rc = 1;
1508  	}
1509  
1510  	up_read(&cinode->lock_sem);
1511  	return rc;
1512  }
1513  
1514  /*
1515   * Set the byte-range lock (posix style). Returns:
1516   * 1) <0, if the error occurs while setting the lock;
1517   * 2) 0, if we set the lock and don't need to request to the server;
1518   * 3) FILE_LOCK_DEFERRED, if we will wait for some other file_lock;
1519   * 4) FILE_LOCK_DEFERRED + 1, if we need to request to the server.
1520   */
1521  static int
cifs_posix_lock_set(struct file * file,struct file_lock * flock)1522  cifs_posix_lock_set(struct file *file, struct file_lock *flock)
1523  {
1524  	struct cifsInodeInfo *cinode = CIFS_I(file_inode(file));
1525  	int rc = FILE_LOCK_DEFERRED + 1;
1526  
1527  	if ((flock->fl_flags & FL_POSIX) == 0)
1528  		return rc;
1529  
1530  	cifs_down_write(&cinode->lock_sem);
1531  	if (!cinode->can_cache_brlcks) {
1532  		up_write(&cinode->lock_sem);
1533  		return rc;
1534  	}
1535  
1536  	rc = posix_lock_file(file, flock, NULL);
1537  	up_write(&cinode->lock_sem);
1538  	return rc;
1539  }
1540  
1541  int
cifs_push_mandatory_locks(struct cifsFileInfo * cfile)1542  cifs_push_mandatory_locks(struct cifsFileInfo *cfile)
1543  {
1544  	unsigned int xid;
1545  	int rc = 0, stored_rc;
1546  	struct cifsLockInfo *li, *tmp;
1547  	struct cifs_tcon *tcon;
1548  	unsigned int num, max_num, max_buf;
1549  	LOCKING_ANDX_RANGE *buf, *cur;
1550  	static const int types[] = {
1551  		LOCKING_ANDX_LARGE_FILES,
1552  		LOCKING_ANDX_SHARED_LOCK | LOCKING_ANDX_LARGE_FILES
1553  	};
1554  	int i;
1555  
1556  	xid = get_xid();
1557  	tcon = tlink_tcon(cfile->tlink);
1558  
1559  	/*
1560  	 * Accessing maxBuf is racy with cifs_reconnect - need to store value
1561  	 * and check it before using.
1562  	 */
1563  	max_buf = tcon->ses->server->maxBuf;
1564  	if (max_buf < (sizeof(struct smb_hdr) + sizeof(LOCKING_ANDX_RANGE))) {
1565  		free_xid(xid);
1566  		return -EINVAL;
1567  	}
1568  
1569  	BUILD_BUG_ON(sizeof(struct smb_hdr) + sizeof(LOCKING_ANDX_RANGE) >
1570  		     PAGE_SIZE);
1571  	max_buf = min_t(unsigned int, max_buf - sizeof(struct smb_hdr),
1572  			PAGE_SIZE);
1573  	max_num = (max_buf - sizeof(struct smb_hdr)) /
1574  						sizeof(LOCKING_ANDX_RANGE);
1575  	buf = kcalloc(max_num, sizeof(LOCKING_ANDX_RANGE), GFP_KERNEL);
1576  	if (!buf) {
1577  		free_xid(xid);
1578  		return -ENOMEM;
1579  	}
1580  
1581  	for (i = 0; i < 2; i++) {
1582  		cur = buf;
1583  		num = 0;
1584  		list_for_each_entry_safe(li, tmp, &cfile->llist->locks, llist) {
1585  			if (li->type != types[i])
1586  				continue;
1587  			cur->Pid = cpu_to_le16(li->pid);
1588  			cur->LengthLow = cpu_to_le32((u32)li->length);
1589  			cur->LengthHigh = cpu_to_le32((u32)(li->length>>32));
1590  			cur->OffsetLow = cpu_to_le32((u32)li->offset);
1591  			cur->OffsetHigh = cpu_to_le32((u32)(li->offset>>32));
1592  			if (++num == max_num) {
1593  				stored_rc = cifs_lockv(xid, tcon,
1594  						       cfile->fid.netfid,
1595  						       (__u8)li->type, 0, num,
1596  						       buf);
1597  				if (stored_rc)
1598  					rc = stored_rc;
1599  				cur = buf;
1600  				num = 0;
1601  			} else
1602  				cur++;
1603  		}
1604  
1605  		if (num) {
1606  			stored_rc = cifs_lockv(xid, tcon, cfile->fid.netfid,
1607  					       (__u8)types[i], 0, num, buf);
1608  			if (stored_rc)
1609  				rc = stored_rc;
1610  		}
1611  	}
1612  
1613  	kfree(buf);
1614  	free_xid(xid);
1615  	return rc;
1616  }
1617  
1618  static __u32
hash_lockowner(fl_owner_t owner)1619  hash_lockowner(fl_owner_t owner)
1620  {
1621  	return cifs_lock_secret ^ hash32_ptr((const void *)owner);
1622  }
1623  #endif /* CONFIG_CIFS_ALLOW_INSECURE_LEGACY */
1624  
1625  struct lock_to_push {
1626  	struct list_head llist;
1627  	__u64 offset;
1628  	__u64 length;
1629  	__u32 pid;
1630  	__u16 netfid;
1631  	__u8 type;
1632  };
1633  
1634  #ifdef CONFIG_CIFS_ALLOW_INSECURE_LEGACY
1635  static int
cifs_push_posix_locks(struct cifsFileInfo * cfile)1636  cifs_push_posix_locks(struct cifsFileInfo *cfile)
1637  {
1638  	struct inode *inode = d_inode(cfile->dentry);
1639  	struct cifs_tcon *tcon = tlink_tcon(cfile->tlink);
1640  	struct file_lock *flock;
1641  	struct file_lock_context *flctx = locks_inode_context(inode);
1642  	unsigned int count = 0, i;
1643  	int rc = 0, xid, type;
1644  	struct list_head locks_to_send, *el;
1645  	struct lock_to_push *lck, *tmp;
1646  	__u64 length;
1647  
1648  	xid = get_xid();
1649  
1650  	if (!flctx)
1651  		goto out;
1652  
1653  	spin_lock(&flctx->flc_lock);
1654  	list_for_each(el, &flctx->flc_posix) {
1655  		count++;
1656  	}
1657  	spin_unlock(&flctx->flc_lock);
1658  
1659  	INIT_LIST_HEAD(&locks_to_send);
1660  
1661  	/*
1662  	 * Allocating count locks is enough because no FL_POSIX locks can be
1663  	 * added to the list while we are holding cinode->lock_sem that
1664  	 * protects locking operations of this inode.
1665  	 */
1666  	for (i = 0; i < count; i++) {
1667  		lck = kmalloc(sizeof(struct lock_to_push), GFP_KERNEL);
1668  		if (!lck) {
1669  			rc = -ENOMEM;
1670  			goto err_out;
1671  		}
1672  		list_add_tail(&lck->llist, &locks_to_send);
1673  	}
1674  
1675  	el = locks_to_send.next;
1676  	spin_lock(&flctx->flc_lock);
1677  	list_for_each_entry(flock, &flctx->flc_posix, fl_list) {
1678  		if (el == &locks_to_send) {
1679  			/*
1680  			 * The list ended. We don't have enough allocated
1681  			 * structures - something is really wrong.
1682  			 */
1683  			cifs_dbg(VFS, "Can't push all brlocks!\n");
1684  			break;
1685  		}
1686  		length = cifs_flock_len(flock);
1687  		if (flock->fl_type == F_RDLCK || flock->fl_type == F_SHLCK)
1688  			type = CIFS_RDLCK;
1689  		else
1690  			type = CIFS_WRLCK;
1691  		lck = list_entry(el, struct lock_to_push, llist);
1692  		lck->pid = hash_lockowner(flock->fl_owner);
1693  		lck->netfid = cfile->fid.netfid;
1694  		lck->length = length;
1695  		lck->type = type;
1696  		lck->offset = flock->fl_start;
1697  	}
1698  	spin_unlock(&flctx->flc_lock);
1699  
1700  	list_for_each_entry_safe(lck, tmp, &locks_to_send, llist) {
1701  		int stored_rc;
1702  
1703  		stored_rc = CIFSSMBPosixLock(xid, tcon, lck->netfid, lck->pid,
1704  					     lck->offset, lck->length, NULL,
1705  					     lck->type, 0);
1706  		if (stored_rc)
1707  			rc = stored_rc;
1708  		list_del(&lck->llist);
1709  		kfree(lck);
1710  	}
1711  
1712  out:
1713  	free_xid(xid);
1714  	return rc;
1715  err_out:
1716  	list_for_each_entry_safe(lck, tmp, &locks_to_send, llist) {
1717  		list_del(&lck->llist);
1718  		kfree(lck);
1719  	}
1720  	goto out;
1721  }
1722  #endif /* CONFIG_CIFS_ALLOW_INSECURE_LEGACY */
1723  
1724  static int
cifs_push_locks(struct cifsFileInfo * cfile)1725  cifs_push_locks(struct cifsFileInfo *cfile)
1726  {
1727  	struct cifsInodeInfo *cinode = CIFS_I(d_inode(cfile->dentry));
1728  	struct cifs_tcon *tcon = tlink_tcon(cfile->tlink);
1729  	int rc = 0;
1730  #ifdef CONFIG_CIFS_ALLOW_INSECURE_LEGACY
1731  	struct cifs_sb_info *cifs_sb = CIFS_SB(cfile->dentry->d_sb);
1732  #endif /* CONFIG_CIFS_ALLOW_INSECURE_LEGACY */
1733  
1734  	/* we are going to update can_cache_brlcks here - need a write access */
1735  	cifs_down_write(&cinode->lock_sem);
1736  	if (!cinode->can_cache_brlcks) {
1737  		up_write(&cinode->lock_sem);
1738  		return rc;
1739  	}
1740  
1741  #ifdef CONFIG_CIFS_ALLOW_INSECURE_LEGACY
1742  	if (cap_unix(tcon->ses) &&
1743  	    (CIFS_UNIX_FCNTL_CAP & le64_to_cpu(tcon->fsUnixInfo.Capability)) &&
1744  	    ((cifs_sb->mnt_cifs_flags & CIFS_MOUNT_NOPOSIXBRL) == 0))
1745  		rc = cifs_push_posix_locks(cfile);
1746  	else
1747  #endif /* CONFIG_CIFS_ALLOW_INSECURE_LEGACY */
1748  		rc = tcon->ses->server->ops->push_mand_locks(cfile);
1749  
1750  	cinode->can_cache_brlcks = false;
1751  	up_write(&cinode->lock_sem);
1752  	return rc;
1753  }
1754  
1755  static void
cifs_read_flock(struct file_lock * flock,__u32 * type,int * lock,int * unlock,bool * wait_flag,struct TCP_Server_Info * server)1756  cifs_read_flock(struct file_lock *flock, __u32 *type, int *lock, int *unlock,
1757  		bool *wait_flag, struct TCP_Server_Info *server)
1758  {
1759  	if (flock->fl_flags & FL_POSIX)
1760  		cifs_dbg(FYI, "Posix\n");
1761  	if (flock->fl_flags & FL_FLOCK)
1762  		cifs_dbg(FYI, "Flock\n");
1763  	if (flock->fl_flags & FL_SLEEP) {
1764  		cifs_dbg(FYI, "Blocking lock\n");
1765  		*wait_flag = true;
1766  	}
1767  	if (flock->fl_flags & FL_ACCESS)
1768  		cifs_dbg(FYI, "Process suspended by mandatory locking - not implemented yet\n");
1769  	if (flock->fl_flags & FL_LEASE)
1770  		cifs_dbg(FYI, "Lease on file - not implemented yet\n");
1771  	if (flock->fl_flags &
1772  	    (~(FL_POSIX | FL_FLOCK | FL_SLEEP |
1773  	       FL_ACCESS | FL_LEASE | FL_CLOSE | FL_OFDLCK)))
1774  		cifs_dbg(FYI, "Unknown lock flags 0x%x\n", flock->fl_flags);
1775  
1776  	*type = server->vals->large_lock_type;
1777  	if (flock->fl_type == F_WRLCK) {
1778  		cifs_dbg(FYI, "F_WRLCK\n");
1779  		*type |= server->vals->exclusive_lock_type;
1780  		*lock = 1;
1781  	} else if (flock->fl_type == F_UNLCK) {
1782  		cifs_dbg(FYI, "F_UNLCK\n");
1783  		*type |= server->vals->unlock_lock_type;
1784  		*unlock = 1;
1785  		/* Check if unlock includes more than one lock range */
1786  	} else if (flock->fl_type == F_RDLCK) {
1787  		cifs_dbg(FYI, "F_RDLCK\n");
1788  		*type |= server->vals->shared_lock_type;
1789  		*lock = 1;
1790  	} else if (flock->fl_type == F_EXLCK) {
1791  		cifs_dbg(FYI, "F_EXLCK\n");
1792  		*type |= server->vals->exclusive_lock_type;
1793  		*lock = 1;
1794  	} else if (flock->fl_type == F_SHLCK) {
1795  		cifs_dbg(FYI, "F_SHLCK\n");
1796  		*type |= server->vals->shared_lock_type;
1797  		*lock = 1;
1798  	} else
1799  		cifs_dbg(FYI, "Unknown type of lock\n");
1800  }
1801  
1802  static int
cifs_getlk(struct file * file,struct file_lock * flock,__u32 type,bool wait_flag,bool posix_lck,unsigned int xid)1803  cifs_getlk(struct file *file, struct file_lock *flock, __u32 type,
1804  	   bool wait_flag, bool posix_lck, unsigned int xid)
1805  {
1806  	int rc = 0;
1807  	__u64 length = cifs_flock_len(flock);
1808  	struct cifsFileInfo *cfile = (struct cifsFileInfo *)file->private_data;
1809  	struct cifs_tcon *tcon = tlink_tcon(cfile->tlink);
1810  	struct TCP_Server_Info *server = tcon->ses->server;
1811  #ifdef CONFIG_CIFS_ALLOW_INSECURE_LEGACY
1812  	__u16 netfid = cfile->fid.netfid;
1813  
1814  	if (posix_lck) {
1815  		int posix_lock_type;
1816  
1817  		rc = cifs_posix_lock_test(file, flock);
1818  		if (!rc)
1819  			return rc;
1820  
1821  		if (type & server->vals->shared_lock_type)
1822  			posix_lock_type = CIFS_RDLCK;
1823  		else
1824  			posix_lock_type = CIFS_WRLCK;
1825  		rc = CIFSSMBPosixLock(xid, tcon, netfid,
1826  				      hash_lockowner(flock->fl_owner),
1827  				      flock->fl_start, length, flock,
1828  				      posix_lock_type, wait_flag);
1829  		return rc;
1830  	}
1831  #endif /* CONFIG_CIFS_ALLOW_INSECURE_LEGACY */
1832  
1833  	rc = cifs_lock_test(cfile, flock->fl_start, length, type, flock);
1834  	if (!rc)
1835  		return rc;
1836  
1837  	/* BB we could chain these into one lock request BB */
1838  	rc = server->ops->mand_lock(xid, cfile, flock->fl_start, length, type,
1839  				    1, 0, false);
1840  	if (rc == 0) {
1841  		rc = server->ops->mand_lock(xid, cfile, flock->fl_start, length,
1842  					    type, 0, 1, false);
1843  		flock->fl_type = F_UNLCK;
1844  		if (rc != 0)
1845  			cifs_dbg(VFS, "Error unlocking previously locked range %d during test of lock\n",
1846  				 rc);
1847  		return 0;
1848  	}
1849  
1850  	if (type & server->vals->shared_lock_type) {
1851  		flock->fl_type = F_WRLCK;
1852  		return 0;
1853  	}
1854  
1855  	type &= ~server->vals->exclusive_lock_type;
1856  
1857  	rc = server->ops->mand_lock(xid, cfile, flock->fl_start, length,
1858  				    type | server->vals->shared_lock_type,
1859  				    1, 0, false);
1860  	if (rc == 0) {
1861  		rc = server->ops->mand_lock(xid, cfile, flock->fl_start, length,
1862  			type | server->vals->shared_lock_type, 0, 1, false);
1863  		flock->fl_type = F_RDLCK;
1864  		if (rc != 0)
1865  			cifs_dbg(VFS, "Error unlocking previously locked range %d during test of lock\n",
1866  				 rc);
1867  	} else
1868  		flock->fl_type = F_WRLCK;
1869  
1870  	return 0;
1871  }
1872  
1873  void
cifs_move_llist(struct list_head * source,struct list_head * dest)1874  cifs_move_llist(struct list_head *source, struct list_head *dest)
1875  {
1876  	struct list_head *li, *tmp;
1877  	list_for_each_safe(li, tmp, source)
1878  		list_move(li, dest);
1879  }
1880  
1881  void
cifs_free_llist(struct list_head * llist)1882  cifs_free_llist(struct list_head *llist)
1883  {
1884  	struct cifsLockInfo *li, *tmp;
1885  	list_for_each_entry_safe(li, tmp, llist, llist) {
1886  		cifs_del_lock_waiters(li);
1887  		list_del(&li->llist);
1888  		kfree(li);
1889  	}
1890  }
1891  
1892  #ifdef CONFIG_CIFS_ALLOW_INSECURE_LEGACY
1893  int
cifs_unlock_range(struct cifsFileInfo * cfile,struct file_lock * flock,unsigned int xid)1894  cifs_unlock_range(struct cifsFileInfo *cfile, struct file_lock *flock,
1895  		  unsigned int xid)
1896  {
1897  	int rc = 0, stored_rc;
1898  	static const int types[] = {
1899  		LOCKING_ANDX_LARGE_FILES,
1900  		LOCKING_ANDX_SHARED_LOCK | LOCKING_ANDX_LARGE_FILES
1901  	};
1902  	unsigned int i;
1903  	unsigned int max_num, num, max_buf;
1904  	LOCKING_ANDX_RANGE *buf, *cur;
1905  	struct cifs_tcon *tcon = tlink_tcon(cfile->tlink);
1906  	struct cifsInodeInfo *cinode = CIFS_I(d_inode(cfile->dentry));
1907  	struct cifsLockInfo *li, *tmp;
1908  	__u64 length = cifs_flock_len(flock);
1909  	struct list_head tmp_llist;
1910  
1911  	INIT_LIST_HEAD(&tmp_llist);
1912  
1913  	/*
1914  	 * Accessing maxBuf is racy with cifs_reconnect - need to store value
1915  	 * and check it before using.
1916  	 */
1917  	max_buf = tcon->ses->server->maxBuf;
1918  	if (max_buf < (sizeof(struct smb_hdr) + sizeof(LOCKING_ANDX_RANGE)))
1919  		return -EINVAL;
1920  
1921  	BUILD_BUG_ON(sizeof(struct smb_hdr) + sizeof(LOCKING_ANDX_RANGE) >
1922  		     PAGE_SIZE);
1923  	max_buf = min_t(unsigned int, max_buf - sizeof(struct smb_hdr),
1924  			PAGE_SIZE);
1925  	max_num = (max_buf - sizeof(struct smb_hdr)) /
1926  						sizeof(LOCKING_ANDX_RANGE);
1927  	buf = kcalloc(max_num, sizeof(LOCKING_ANDX_RANGE), GFP_KERNEL);
1928  	if (!buf)
1929  		return -ENOMEM;
1930  
1931  	cifs_down_write(&cinode->lock_sem);
1932  	for (i = 0; i < 2; i++) {
1933  		cur = buf;
1934  		num = 0;
1935  		list_for_each_entry_safe(li, tmp, &cfile->llist->locks, llist) {
1936  			if (flock->fl_start > li->offset ||
1937  			    (flock->fl_start + length) <
1938  			    (li->offset + li->length))
1939  				continue;
1940  			if (current->tgid != li->pid)
1941  				continue;
1942  			if (types[i] != li->type)
1943  				continue;
1944  			if (cinode->can_cache_brlcks) {
1945  				/*
1946  				 * We can cache brlock requests - simply remove
1947  				 * a lock from the file's list.
1948  				 */
1949  				list_del(&li->llist);
1950  				cifs_del_lock_waiters(li);
1951  				kfree(li);
1952  				continue;
1953  			}
1954  			cur->Pid = cpu_to_le16(li->pid);
1955  			cur->LengthLow = cpu_to_le32((u32)li->length);
1956  			cur->LengthHigh = cpu_to_le32((u32)(li->length>>32));
1957  			cur->OffsetLow = cpu_to_le32((u32)li->offset);
1958  			cur->OffsetHigh = cpu_to_le32((u32)(li->offset>>32));
1959  			/*
1960  			 * We need to save a lock here to let us add it again to
1961  			 * the file's list if the unlock range request fails on
1962  			 * the server.
1963  			 */
1964  			list_move(&li->llist, &tmp_llist);
1965  			if (++num == max_num) {
1966  				stored_rc = cifs_lockv(xid, tcon,
1967  						       cfile->fid.netfid,
1968  						       li->type, num, 0, buf);
1969  				if (stored_rc) {
1970  					/*
1971  					 * We failed on the unlock range
1972  					 * request - add all locks from the tmp
1973  					 * list to the head of the file's list.
1974  					 */
1975  					cifs_move_llist(&tmp_llist,
1976  							&cfile->llist->locks);
1977  					rc = stored_rc;
1978  				} else
1979  					/*
1980  					 * The unlock range request succeed -
1981  					 * free the tmp list.
1982  					 */
1983  					cifs_free_llist(&tmp_llist);
1984  				cur = buf;
1985  				num = 0;
1986  			} else
1987  				cur++;
1988  		}
1989  		if (num) {
1990  			stored_rc = cifs_lockv(xid, tcon, cfile->fid.netfid,
1991  					       types[i], num, 0, buf);
1992  			if (stored_rc) {
1993  				cifs_move_llist(&tmp_llist,
1994  						&cfile->llist->locks);
1995  				rc = stored_rc;
1996  			} else
1997  				cifs_free_llist(&tmp_llist);
1998  		}
1999  	}
2000  
2001  	up_write(&cinode->lock_sem);
2002  	kfree(buf);
2003  	return rc;
2004  }
2005  #endif /* CONFIG_CIFS_ALLOW_INSECURE_LEGACY */
2006  
2007  static int
cifs_setlk(struct file * file,struct file_lock * flock,__u32 type,bool wait_flag,bool posix_lck,int lock,int unlock,unsigned int xid)2008  cifs_setlk(struct file *file, struct file_lock *flock, __u32 type,
2009  	   bool wait_flag, bool posix_lck, int lock, int unlock,
2010  	   unsigned int xid)
2011  {
2012  	int rc = 0;
2013  	__u64 length = cifs_flock_len(flock);
2014  	struct cifsFileInfo *cfile = (struct cifsFileInfo *)file->private_data;
2015  	struct cifs_tcon *tcon = tlink_tcon(cfile->tlink);
2016  	struct TCP_Server_Info *server = tcon->ses->server;
2017  	struct inode *inode = d_inode(cfile->dentry);
2018  
2019  #ifdef CONFIG_CIFS_ALLOW_INSECURE_LEGACY
2020  	if (posix_lck) {
2021  		int posix_lock_type;
2022  
2023  		rc = cifs_posix_lock_set(file, flock);
2024  		if (rc <= FILE_LOCK_DEFERRED)
2025  			return rc;
2026  
2027  		if (type & server->vals->shared_lock_type)
2028  			posix_lock_type = CIFS_RDLCK;
2029  		else
2030  			posix_lock_type = CIFS_WRLCK;
2031  
2032  		if (unlock == 1)
2033  			posix_lock_type = CIFS_UNLCK;
2034  
2035  		rc = CIFSSMBPosixLock(xid, tcon, cfile->fid.netfid,
2036  				      hash_lockowner(flock->fl_owner),
2037  				      flock->fl_start, length,
2038  				      NULL, posix_lock_type, wait_flag);
2039  		goto out;
2040  	}
2041  #endif /* CONFIG_CIFS_ALLOW_INSECURE_LEGACY */
2042  	if (lock) {
2043  		struct cifsLockInfo *lock;
2044  
2045  		lock = cifs_lock_init(flock->fl_start, length, type,
2046  				      flock->fl_flags);
2047  		if (!lock)
2048  			return -ENOMEM;
2049  
2050  		rc = cifs_lock_add_if(cfile, lock, wait_flag);
2051  		if (rc < 0) {
2052  			kfree(lock);
2053  			return rc;
2054  		}
2055  		if (!rc)
2056  			goto out;
2057  
2058  		/*
2059  		 * Windows 7 server can delay breaking lease from read to None
2060  		 * if we set a byte-range lock on a file - break it explicitly
2061  		 * before sending the lock to the server to be sure the next
2062  		 * read won't conflict with non-overlapted locks due to
2063  		 * pagereading.
2064  		 */
2065  		if (!CIFS_CACHE_WRITE(CIFS_I(inode)) &&
2066  					CIFS_CACHE_READ(CIFS_I(inode))) {
2067  			cifs_zap_mapping(inode);
2068  			cifs_dbg(FYI, "Set no oplock for inode=%p due to mand locks\n",
2069  				 inode);
2070  			CIFS_I(inode)->oplock = 0;
2071  		}
2072  
2073  		rc = server->ops->mand_lock(xid, cfile, flock->fl_start, length,
2074  					    type, 1, 0, wait_flag);
2075  		if (rc) {
2076  			kfree(lock);
2077  			return rc;
2078  		}
2079  
2080  		cifs_lock_add(cfile, lock);
2081  	} else if (unlock)
2082  		rc = server->ops->mand_unlock_range(cfile, flock, xid);
2083  
2084  out:
2085  	if ((flock->fl_flags & FL_POSIX) || (flock->fl_flags & FL_FLOCK)) {
2086  		/*
2087  		 * If this is a request to remove all locks because we
2088  		 * are closing the file, it doesn't matter if the
2089  		 * unlocking failed as both cifs.ko and the SMB server
2090  		 * remove the lock on file close
2091  		 */
2092  		if (rc) {
2093  			cifs_dbg(VFS, "%s failed rc=%d\n", __func__, rc);
2094  			if (!(flock->fl_flags & FL_CLOSE))
2095  				return rc;
2096  		}
2097  		rc = locks_lock_file_wait(file, flock);
2098  	}
2099  	return rc;
2100  }
2101  
cifs_flock(struct file * file,int cmd,struct file_lock * fl)2102  int cifs_flock(struct file *file, int cmd, struct file_lock *fl)
2103  {
2104  	int rc, xid;
2105  	int lock = 0, unlock = 0;
2106  	bool wait_flag = false;
2107  	bool posix_lck = false;
2108  	struct cifs_sb_info *cifs_sb;
2109  	struct cifs_tcon *tcon;
2110  	struct cifsFileInfo *cfile;
2111  	__u32 type;
2112  
2113  	xid = get_xid();
2114  
2115  	if (!(fl->fl_flags & FL_FLOCK)) {
2116  		rc = -ENOLCK;
2117  		free_xid(xid);
2118  		return rc;
2119  	}
2120  
2121  	cfile = (struct cifsFileInfo *)file->private_data;
2122  	tcon = tlink_tcon(cfile->tlink);
2123  
2124  	cifs_read_flock(fl, &type, &lock, &unlock, &wait_flag,
2125  			tcon->ses->server);
2126  	cifs_sb = CIFS_FILE_SB(file);
2127  
2128  	if (cap_unix(tcon->ses) &&
2129  	    (CIFS_UNIX_FCNTL_CAP & le64_to_cpu(tcon->fsUnixInfo.Capability)) &&
2130  	    ((cifs_sb->mnt_cifs_flags & CIFS_MOUNT_NOPOSIXBRL) == 0))
2131  		posix_lck = true;
2132  
2133  	if (!lock && !unlock) {
2134  		/*
2135  		 * if no lock or unlock then nothing to do since we do not
2136  		 * know what it is
2137  		 */
2138  		rc = -EOPNOTSUPP;
2139  		free_xid(xid);
2140  		return rc;
2141  	}
2142  
2143  	rc = cifs_setlk(file, fl, type, wait_flag, posix_lck, lock, unlock,
2144  			xid);
2145  	free_xid(xid);
2146  	return rc;
2147  
2148  
2149  }
2150  
cifs_lock(struct file * file,int cmd,struct file_lock * flock)2151  int cifs_lock(struct file *file, int cmd, struct file_lock *flock)
2152  {
2153  	int rc, xid;
2154  	int lock = 0, unlock = 0;
2155  	bool wait_flag = false;
2156  	bool posix_lck = false;
2157  	struct cifs_sb_info *cifs_sb;
2158  	struct cifs_tcon *tcon;
2159  	struct cifsFileInfo *cfile;
2160  	__u32 type;
2161  
2162  	rc = -EACCES;
2163  	xid = get_xid();
2164  
2165  	cifs_dbg(FYI, "%s: %pD2 cmd=0x%x type=0x%x flags=0x%x r=%lld:%lld\n", __func__, file, cmd,
2166  		 flock->fl_flags, flock->fl_type, (long long)flock->fl_start,
2167  		 (long long)flock->fl_end);
2168  
2169  	cfile = (struct cifsFileInfo *)file->private_data;
2170  	tcon = tlink_tcon(cfile->tlink);
2171  
2172  	cifs_read_flock(flock, &type, &lock, &unlock, &wait_flag,
2173  			tcon->ses->server);
2174  	cifs_sb = CIFS_FILE_SB(file);
2175  	set_bit(CIFS_INO_CLOSE_ON_LOCK, &CIFS_I(d_inode(cfile->dentry))->flags);
2176  
2177  	if (cap_unix(tcon->ses) &&
2178  	    (CIFS_UNIX_FCNTL_CAP & le64_to_cpu(tcon->fsUnixInfo.Capability)) &&
2179  	    ((cifs_sb->mnt_cifs_flags & CIFS_MOUNT_NOPOSIXBRL) == 0))
2180  		posix_lck = true;
2181  	/*
2182  	 * BB add code here to normalize offset and length to account for
2183  	 * negative length which we can not accept over the wire.
2184  	 */
2185  	if (IS_GETLK(cmd)) {
2186  		rc = cifs_getlk(file, flock, type, wait_flag, posix_lck, xid);
2187  		free_xid(xid);
2188  		return rc;
2189  	}
2190  
2191  	if (!lock && !unlock) {
2192  		/*
2193  		 * if no lock or unlock then nothing to do since we do not
2194  		 * know what it is
2195  		 */
2196  		free_xid(xid);
2197  		return -EOPNOTSUPP;
2198  	}
2199  
2200  	rc = cifs_setlk(file, flock, type, wait_flag, posix_lck, lock, unlock,
2201  			xid);
2202  	free_xid(xid);
2203  	return rc;
2204  }
2205  
2206  /*
2207   * update the file size (if needed) after a write. Should be called with
2208   * the inode->i_lock held
2209   */
2210  void
cifs_update_eof(struct cifsInodeInfo * cifsi,loff_t offset,unsigned int bytes_written)2211  cifs_update_eof(struct cifsInodeInfo *cifsi, loff_t offset,
2212  		      unsigned int bytes_written)
2213  {
2214  	loff_t end_of_write = offset + bytes_written;
2215  
2216  	if (end_of_write > cifsi->server_eof)
2217  		cifsi->server_eof = end_of_write;
2218  }
2219  
2220  static ssize_t
cifs_write(struct cifsFileInfo * open_file,__u32 pid,const char * write_data,size_t write_size,loff_t * offset)2221  cifs_write(struct cifsFileInfo *open_file, __u32 pid, const char *write_data,
2222  	   size_t write_size, loff_t *offset)
2223  {
2224  	int rc = 0;
2225  	unsigned int bytes_written = 0;
2226  	unsigned int total_written;
2227  	struct cifs_tcon *tcon;
2228  	struct TCP_Server_Info *server;
2229  	unsigned int xid;
2230  	struct dentry *dentry = open_file->dentry;
2231  	struct cifsInodeInfo *cifsi = CIFS_I(d_inode(dentry));
2232  	struct cifs_io_parms io_parms = {0};
2233  
2234  	cifs_dbg(FYI, "write %zd bytes to offset %lld of %pd\n",
2235  		 write_size, *offset, dentry);
2236  
2237  	tcon = tlink_tcon(open_file->tlink);
2238  	server = tcon->ses->server;
2239  
2240  	if (!server->ops->sync_write)
2241  		return -ENOSYS;
2242  
2243  	xid = get_xid();
2244  
2245  	for (total_written = 0; write_size > total_written;
2246  	     total_written += bytes_written) {
2247  		rc = -EAGAIN;
2248  		while (rc == -EAGAIN) {
2249  			struct kvec iov[2];
2250  			unsigned int len;
2251  
2252  			if (open_file->invalidHandle) {
2253  				/* we could deadlock if we called
2254  				   filemap_fdatawait from here so tell
2255  				   reopen_file not to flush data to
2256  				   server now */
2257  				rc = cifs_reopen_file(open_file, false);
2258  				if (rc != 0)
2259  					break;
2260  			}
2261  
2262  			len = min(server->ops->wp_retry_size(d_inode(dentry)),
2263  				  (unsigned int)write_size - total_written);
2264  			/* iov[0] is reserved for smb header */
2265  			iov[1].iov_base = (char *)write_data + total_written;
2266  			iov[1].iov_len = len;
2267  			io_parms.pid = pid;
2268  			io_parms.tcon = tcon;
2269  			io_parms.offset = *offset;
2270  			io_parms.length = len;
2271  			rc = server->ops->sync_write(xid, &open_file->fid,
2272  					&io_parms, &bytes_written, iov, 1);
2273  		}
2274  		if (rc || (bytes_written == 0)) {
2275  			if (total_written)
2276  				break;
2277  			else {
2278  				free_xid(xid);
2279  				return rc;
2280  			}
2281  		} else {
2282  			spin_lock(&d_inode(dentry)->i_lock);
2283  			cifs_update_eof(cifsi, *offset, bytes_written);
2284  			spin_unlock(&d_inode(dentry)->i_lock);
2285  			*offset += bytes_written;
2286  		}
2287  	}
2288  
2289  	cifs_stats_bytes_written(tcon, total_written);
2290  
2291  	if (total_written > 0) {
2292  		spin_lock(&d_inode(dentry)->i_lock);
2293  		if (*offset > d_inode(dentry)->i_size) {
2294  			i_size_write(d_inode(dentry), *offset);
2295  			d_inode(dentry)->i_blocks = (512 - 1 + *offset) >> 9;
2296  		}
2297  		spin_unlock(&d_inode(dentry)->i_lock);
2298  	}
2299  	mark_inode_dirty_sync(d_inode(dentry));
2300  	free_xid(xid);
2301  	return total_written;
2302  }
2303  
find_readable_file(struct cifsInodeInfo * cifs_inode,bool fsuid_only)2304  struct cifsFileInfo *find_readable_file(struct cifsInodeInfo *cifs_inode,
2305  					bool fsuid_only)
2306  {
2307  	struct cifsFileInfo *open_file = NULL;
2308  	struct cifs_sb_info *cifs_sb = CIFS_SB(cifs_inode->netfs.inode.i_sb);
2309  
2310  	/* only filter by fsuid on multiuser mounts */
2311  	if (!(cifs_sb->mnt_cifs_flags & CIFS_MOUNT_MULTIUSER))
2312  		fsuid_only = false;
2313  
2314  	spin_lock(&cifs_inode->open_file_lock);
2315  	/* we could simply get the first_list_entry since write-only entries
2316  	   are always at the end of the list but since the first entry might
2317  	   have a close pending, we go through the whole list */
2318  	list_for_each_entry(open_file, &cifs_inode->openFileList, flist) {
2319  		if (fsuid_only && !uid_eq(open_file->uid, current_fsuid()))
2320  			continue;
2321  		if (OPEN_FMODE(open_file->f_flags) & FMODE_READ) {
2322  			if ((!open_file->invalidHandle)) {
2323  				/* found a good file */
2324  				/* lock it so it will not be closed on us */
2325  				cifsFileInfo_get(open_file);
2326  				spin_unlock(&cifs_inode->open_file_lock);
2327  				return open_file;
2328  			} /* else might as well continue, and look for
2329  			     another, or simply have the caller reopen it
2330  			     again rather than trying to fix this handle */
2331  		} else /* write only file */
2332  			break; /* write only files are last so must be done */
2333  	}
2334  	spin_unlock(&cifs_inode->open_file_lock);
2335  	return NULL;
2336  }
2337  
2338  /* Return -EBADF if no handle is found and general rc otherwise */
2339  int
cifs_get_writable_file(struct cifsInodeInfo * cifs_inode,int flags,struct cifsFileInfo ** ret_file)2340  cifs_get_writable_file(struct cifsInodeInfo *cifs_inode, int flags,
2341  		       struct cifsFileInfo **ret_file)
2342  {
2343  	struct cifsFileInfo *open_file, *inv_file = NULL;
2344  	struct cifs_sb_info *cifs_sb;
2345  	bool any_available = false;
2346  	int rc = -EBADF;
2347  	unsigned int refind = 0;
2348  	bool fsuid_only = flags & FIND_WR_FSUID_ONLY;
2349  	bool with_delete = flags & FIND_WR_WITH_DELETE;
2350  	*ret_file = NULL;
2351  
2352  	/*
2353  	 * Having a null inode here (because mapping->host was set to zero by
2354  	 * the VFS or MM) should not happen but we had reports of on oops (due
2355  	 * to it being zero) during stress testcases so we need to check for it
2356  	 */
2357  
2358  	if (cifs_inode == NULL) {
2359  		cifs_dbg(VFS, "Null inode passed to cifs_writeable_file\n");
2360  		dump_stack();
2361  		return rc;
2362  	}
2363  
2364  	cifs_sb = CIFS_SB(cifs_inode->netfs.inode.i_sb);
2365  
2366  	/* only filter by fsuid on multiuser mounts */
2367  	if (!(cifs_sb->mnt_cifs_flags & CIFS_MOUNT_MULTIUSER))
2368  		fsuid_only = false;
2369  
2370  	spin_lock(&cifs_inode->open_file_lock);
2371  refind_writable:
2372  	if (refind > MAX_REOPEN_ATT) {
2373  		spin_unlock(&cifs_inode->open_file_lock);
2374  		return rc;
2375  	}
2376  	list_for_each_entry(open_file, &cifs_inode->openFileList, flist) {
2377  		if (!any_available && open_file->pid != current->tgid)
2378  			continue;
2379  		if (fsuid_only && !uid_eq(open_file->uid, current_fsuid()))
2380  			continue;
2381  		if (with_delete && !(open_file->fid.access & DELETE))
2382  			continue;
2383  		if (OPEN_FMODE(open_file->f_flags) & FMODE_WRITE) {
2384  			if (!open_file->invalidHandle) {
2385  				/* found a good writable file */
2386  				cifsFileInfo_get(open_file);
2387  				spin_unlock(&cifs_inode->open_file_lock);
2388  				*ret_file = open_file;
2389  				return 0;
2390  			} else {
2391  				if (!inv_file)
2392  					inv_file = open_file;
2393  			}
2394  		}
2395  	}
2396  	/* couldn't find useable FH with same pid, try any available */
2397  	if (!any_available) {
2398  		any_available = true;
2399  		goto refind_writable;
2400  	}
2401  
2402  	if (inv_file) {
2403  		any_available = false;
2404  		cifsFileInfo_get(inv_file);
2405  	}
2406  
2407  	spin_unlock(&cifs_inode->open_file_lock);
2408  
2409  	if (inv_file) {
2410  		rc = cifs_reopen_file(inv_file, false);
2411  		if (!rc) {
2412  			*ret_file = inv_file;
2413  			return 0;
2414  		}
2415  
2416  		spin_lock(&cifs_inode->open_file_lock);
2417  		list_move_tail(&inv_file->flist, &cifs_inode->openFileList);
2418  		spin_unlock(&cifs_inode->open_file_lock);
2419  		cifsFileInfo_put(inv_file);
2420  		++refind;
2421  		inv_file = NULL;
2422  		spin_lock(&cifs_inode->open_file_lock);
2423  		goto refind_writable;
2424  	}
2425  
2426  	return rc;
2427  }
2428  
2429  struct cifsFileInfo *
find_writable_file(struct cifsInodeInfo * cifs_inode,int flags)2430  find_writable_file(struct cifsInodeInfo *cifs_inode, int flags)
2431  {
2432  	struct cifsFileInfo *cfile;
2433  	int rc;
2434  
2435  	rc = cifs_get_writable_file(cifs_inode, flags, &cfile);
2436  	if (rc)
2437  		cifs_dbg(FYI, "Couldn't find writable handle rc=%d\n", rc);
2438  
2439  	return cfile;
2440  }
2441  
2442  int
cifs_get_writable_path(struct cifs_tcon * tcon,const char * name,int flags,struct cifsFileInfo ** ret_file)2443  cifs_get_writable_path(struct cifs_tcon *tcon, const char *name,
2444  		       int flags,
2445  		       struct cifsFileInfo **ret_file)
2446  {
2447  	struct cifsFileInfo *cfile;
2448  	void *page = alloc_dentry_path();
2449  
2450  	*ret_file = NULL;
2451  
2452  	spin_lock(&tcon->open_file_lock);
2453  	list_for_each_entry(cfile, &tcon->openFileList, tlist) {
2454  		struct cifsInodeInfo *cinode;
2455  		const char *full_path = build_path_from_dentry(cfile->dentry, page);
2456  		if (IS_ERR(full_path)) {
2457  			spin_unlock(&tcon->open_file_lock);
2458  			free_dentry_path(page);
2459  			return PTR_ERR(full_path);
2460  		}
2461  		if (strcmp(full_path, name))
2462  			continue;
2463  
2464  		cinode = CIFS_I(d_inode(cfile->dentry));
2465  		spin_unlock(&tcon->open_file_lock);
2466  		free_dentry_path(page);
2467  		return cifs_get_writable_file(cinode, flags, ret_file);
2468  	}
2469  
2470  	spin_unlock(&tcon->open_file_lock);
2471  	free_dentry_path(page);
2472  	return -ENOENT;
2473  }
2474  
2475  int
cifs_get_readable_path(struct cifs_tcon * tcon,const char * name,struct cifsFileInfo ** ret_file)2476  cifs_get_readable_path(struct cifs_tcon *tcon, const char *name,
2477  		       struct cifsFileInfo **ret_file)
2478  {
2479  	struct cifsFileInfo *cfile;
2480  	void *page = alloc_dentry_path();
2481  
2482  	*ret_file = NULL;
2483  
2484  	spin_lock(&tcon->open_file_lock);
2485  	list_for_each_entry(cfile, &tcon->openFileList, tlist) {
2486  		struct cifsInodeInfo *cinode;
2487  		const char *full_path = build_path_from_dentry(cfile->dentry, page);
2488  		if (IS_ERR(full_path)) {
2489  			spin_unlock(&tcon->open_file_lock);
2490  			free_dentry_path(page);
2491  			return PTR_ERR(full_path);
2492  		}
2493  		if (strcmp(full_path, name))
2494  			continue;
2495  
2496  		cinode = CIFS_I(d_inode(cfile->dentry));
2497  		spin_unlock(&tcon->open_file_lock);
2498  		free_dentry_path(page);
2499  		*ret_file = find_readable_file(cinode, 0);
2500  		return *ret_file ? 0 : -ENOENT;
2501  	}
2502  
2503  	spin_unlock(&tcon->open_file_lock);
2504  	free_dentry_path(page);
2505  	return -ENOENT;
2506  }
2507  
2508  void
cifs_writedata_release(struct kref * refcount)2509  cifs_writedata_release(struct kref *refcount)
2510  {
2511  	struct cifs_writedata *wdata = container_of(refcount,
2512  					struct cifs_writedata, refcount);
2513  #ifdef CONFIG_CIFS_SMB_DIRECT
2514  	if (wdata->mr) {
2515  		smbd_deregister_mr(wdata->mr);
2516  		wdata->mr = NULL;
2517  	}
2518  #endif
2519  
2520  	if (wdata->cfile)
2521  		cifsFileInfo_put(wdata->cfile);
2522  
2523  	kfree(wdata);
2524  }
2525  
2526  /*
2527   * Write failed with a retryable error. Resend the write request. It's also
2528   * possible that the page was redirtied so re-clean the page.
2529   */
2530  static void
cifs_writev_requeue(struct cifs_writedata * wdata)2531  cifs_writev_requeue(struct cifs_writedata *wdata)
2532  {
2533  	int rc = 0;
2534  	struct inode *inode = d_inode(wdata->cfile->dentry);
2535  	struct TCP_Server_Info *server;
2536  	unsigned int rest_len = wdata->bytes;
2537  	loff_t fpos = wdata->offset;
2538  
2539  	server = tlink_tcon(wdata->cfile->tlink)->ses->server;
2540  	do {
2541  		struct cifs_writedata *wdata2;
2542  		unsigned int wsize, cur_len;
2543  
2544  		wsize = server->ops->wp_retry_size(inode);
2545  		if (wsize < rest_len) {
2546  			if (wsize < PAGE_SIZE) {
2547  				rc = -EOPNOTSUPP;
2548  				break;
2549  			}
2550  			cur_len = min(round_down(wsize, PAGE_SIZE), rest_len);
2551  		} else {
2552  			cur_len = rest_len;
2553  		}
2554  
2555  		wdata2 = cifs_writedata_alloc(cifs_writev_complete);
2556  		if (!wdata2) {
2557  			rc = -ENOMEM;
2558  			break;
2559  		}
2560  
2561  		wdata2->sync_mode = wdata->sync_mode;
2562  		wdata2->offset	= fpos;
2563  		wdata2->bytes	= cur_len;
2564  		wdata2->iter	= wdata->iter;
2565  
2566  		iov_iter_advance(&wdata2->iter, fpos - wdata->offset);
2567  		iov_iter_truncate(&wdata2->iter, wdata2->bytes);
2568  
2569  		if (iov_iter_is_xarray(&wdata2->iter))
2570  			/* Check for pages having been redirtied and clean
2571  			 * them.  We can do this by walking the xarray.  If
2572  			 * it's not an xarray, then it's a DIO and we shouldn't
2573  			 * be mucking around with the page bits.
2574  			 */
2575  			cifs_undirty_folios(inode, fpos, cur_len);
2576  
2577  		rc = cifs_get_writable_file(CIFS_I(inode), FIND_WR_ANY,
2578  					    &wdata2->cfile);
2579  		if (!wdata2->cfile) {
2580  			cifs_dbg(VFS, "No writable handle to retry writepages rc=%d\n",
2581  				 rc);
2582  			if (!is_retryable_error(rc))
2583  				rc = -EBADF;
2584  		} else {
2585  			wdata2->pid = wdata2->cfile->pid;
2586  			rc = server->ops->async_writev(wdata2,
2587  						       cifs_writedata_release);
2588  		}
2589  
2590  		kref_put(&wdata2->refcount, cifs_writedata_release);
2591  		if (rc) {
2592  			if (is_retryable_error(rc))
2593  				continue;
2594  			fpos += cur_len;
2595  			rest_len -= cur_len;
2596  			break;
2597  		}
2598  
2599  		fpos += cur_len;
2600  		rest_len -= cur_len;
2601  	} while (rest_len > 0);
2602  
2603  	/* Clean up remaining pages from the original wdata */
2604  	if (iov_iter_is_xarray(&wdata->iter))
2605  		cifs_pages_write_failed(inode, fpos, rest_len);
2606  
2607  	if (rc != 0 && !is_retryable_error(rc))
2608  		mapping_set_error(inode->i_mapping, rc);
2609  	kref_put(&wdata->refcount, cifs_writedata_release);
2610  }
2611  
2612  void
cifs_writev_complete(struct work_struct * work)2613  cifs_writev_complete(struct work_struct *work)
2614  {
2615  	struct cifs_writedata *wdata = container_of(work,
2616  						struct cifs_writedata, work);
2617  	struct inode *inode = d_inode(wdata->cfile->dentry);
2618  
2619  	if (wdata->result == 0) {
2620  		spin_lock(&inode->i_lock);
2621  		cifs_update_eof(CIFS_I(inode), wdata->offset, wdata->bytes);
2622  		spin_unlock(&inode->i_lock);
2623  		cifs_stats_bytes_written(tlink_tcon(wdata->cfile->tlink),
2624  					 wdata->bytes);
2625  	} else if (wdata->sync_mode == WB_SYNC_ALL && wdata->result == -EAGAIN)
2626  		return cifs_writev_requeue(wdata);
2627  
2628  	if (wdata->result == -EAGAIN)
2629  		cifs_pages_write_redirty(inode, wdata->offset, wdata->bytes);
2630  	else if (wdata->result < 0)
2631  		cifs_pages_write_failed(inode, wdata->offset, wdata->bytes);
2632  	else
2633  		cifs_pages_written_back(inode, wdata->offset, wdata->bytes);
2634  
2635  	if (wdata->result != -EAGAIN)
2636  		mapping_set_error(inode->i_mapping, wdata->result);
2637  	kref_put(&wdata->refcount, cifs_writedata_release);
2638  }
2639  
cifs_writedata_alloc(work_func_t complete)2640  struct cifs_writedata *cifs_writedata_alloc(work_func_t complete)
2641  {
2642  	struct cifs_writedata *wdata;
2643  
2644  	wdata = kzalloc(sizeof(*wdata), GFP_NOFS);
2645  	if (wdata != NULL) {
2646  		kref_init(&wdata->refcount);
2647  		INIT_LIST_HEAD(&wdata->list);
2648  		init_completion(&wdata->done);
2649  		INIT_WORK(&wdata->work, complete);
2650  	}
2651  	return wdata;
2652  }
2653  
cifs_partialpagewrite(struct page * page,unsigned from,unsigned to)2654  static int cifs_partialpagewrite(struct page *page, unsigned from, unsigned to)
2655  {
2656  	struct address_space *mapping = page->mapping;
2657  	loff_t offset = (loff_t)page->index << PAGE_SHIFT;
2658  	char *write_data;
2659  	int rc = -EFAULT;
2660  	int bytes_written = 0;
2661  	struct inode *inode;
2662  	struct cifsFileInfo *open_file;
2663  
2664  	if (!mapping || !mapping->host)
2665  		return -EFAULT;
2666  
2667  	inode = page->mapping->host;
2668  
2669  	offset += (loff_t)from;
2670  	write_data = kmap(page);
2671  	write_data += from;
2672  
2673  	if ((to > PAGE_SIZE) || (from > to)) {
2674  		kunmap(page);
2675  		return -EIO;
2676  	}
2677  
2678  	/* racing with truncate? */
2679  	if (offset > mapping->host->i_size) {
2680  		kunmap(page);
2681  		return 0; /* don't care */
2682  	}
2683  
2684  	/* check to make sure that we are not extending the file */
2685  	if (mapping->host->i_size - offset < (loff_t)to)
2686  		to = (unsigned)(mapping->host->i_size - offset);
2687  
2688  	rc = cifs_get_writable_file(CIFS_I(mapping->host), FIND_WR_ANY,
2689  				    &open_file);
2690  	if (!rc) {
2691  		bytes_written = cifs_write(open_file, open_file->pid,
2692  					   write_data, to - from, &offset);
2693  		cifsFileInfo_put(open_file);
2694  		/* Does mm or vfs already set times? */
2695  		simple_inode_init_ts(inode);
2696  		if ((bytes_written > 0) && (offset))
2697  			rc = 0;
2698  		else if (bytes_written < 0)
2699  			rc = bytes_written;
2700  		else
2701  			rc = -EFAULT;
2702  	} else {
2703  		cifs_dbg(FYI, "No writable handle for write page rc=%d\n", rc);
2704  		if (!is_retryable_error(rc))
2705  			rc = -EIO;
2706  	}
2707  
2708  	kunmap(page);
2709  	return rc;
2710  }
2711  
2712  /*
2713   * Extend the region to be written back to include subsequent contiguously
2714   * dirty pages if possible, but don't sleep while doing so.
2715   */
cifs_extend_writeback(struct address_space * mapping,struct xa_state * xas,long * _count,loff_t start,int max_pages,loff_t max_len,size_t * _len)2716  static void cifs_extend_writeback(struct address_space *mapping,
2717  				  struct xa_state *xas,
2718  				  long *_count,
2719  				  loff_t start,
2720  				  int max_pages,
2721  				  loff_t max_len,
2722  				  size_t *_len)
2723  {
2724  	struct folio_batch batch;
2725  	struct folio *folio;
2726  	unsigned int nr_pages;
2727  	pgoff_t index = (start + *_len) / PAGE_SIZE;
2728  	size_t len;
2729  	bool stop = true;
2730  	unsigned int i;
2731  
2732  	folio_batch_init(&batch);
2733  
2734  	do {
2735  		/* Firstly, we gather up a batch of contiguous dirty pages
2736  		 * under the RCU read lock - but we can't clear the dirty flags
2737  		 * there if any of those pages are mapped.
2738  		 */
2739  		rcu_read_lock();
2740  
2741  		xas_for_each(xas, folio, ULONG_MAX) {
2742  			stop = true;
2743  			if (xas_retry(xas, folio))
2744  				continue;
2745  			if (xa_is_value(folio))
2746  				break;
2747  			if (folio->index != index) {
2748  				xas_reset(xas);
2749  				break;
2750  			}
2751  
2752  			if (!folio_try_get(folio)) {
2753  				xas_reset(xas);
2754  				continue;
2755  			}
2756  			nr_pages = folio_nr_pages(folio);
2757  			if (nr_pages > max_pages) {
2758  				xas_reset(xas);
2759  				break;
2760  			}
2761  
2762  			/* Has the page moved or been split? */
2763  			if (unlikely(folio != xas_reload(xas))) {
2764  				folio_put(folio);
2765  				xas_reset(xas);
2766  				break;
2767  			}
2768  
2769  			if (!folio_trylock(folio)) {
2770  				folio_put(folio);
2771  				xas_reset(xas);
2772  				break;
2773  			}
2774  			if (!folio_test_dirty(folio) ||
2775  			    folio_test_writeback(folio)) {
2776  				folio_unlock(folio);
2777  				folio_put(folio);
2778  				xas_reset(xas);
2779  				break;
2780  			}
2781  
2782  			max_pages -= nr_pages;
2783  			len = folio_size(folio);
2784  			stop = false;
2785  
2786  			index += nr_pages;
2787  			*_count -= nr_pages;
2788  			*_len += len;
2789  			if (max_pages <= 0 || *_len >= max_len || *_count <= 0)
2790  				stop = true;
2791  
2792  			if (!folio_batch_add(&batch, folio))
2793  				break;
2794  			if (stop)
2795  				break;
2796  		}
2797  
2798  		xas_pause(xas);
2799  		rcu_read_unlock();
2800  
2801  		/* Now, if we obtained any pages, we can shift them to being
2802  		 * writable and mark them for caching.
2803  		 */
2804  		if (!folio_batch_count(&batch))
2805  			break;
2806  
2807  		for (i = 0; i < folio_batch_count(&batch); i++) {
2808  			folio = batch.folios[i];
2809  			/* The folio should be locked, dirty and not undergoing
2810  			 * writeback from the loop above.
2811  			 */
2812  			if (!folio_clear_dirty_for_io(folio))
2813  				WARN_ON(1);
2814  			folio_start_writeback(folio);
2815  			folio_unlock(folio);
2816  		}
2817  
2818  		folio_batch_release(&batch);
2819  		cond_resched();
2820  	} while (!stop);
2821  }
2822  
2823  /*
2824   * Write back the locked page and any subsequent non-locked dirty pages.
2825   */
cifs_write_back_from_locked_folio(struct address_space * mapping,struct writeback_control * wbc,struct xa_state * xas,struct folio * folio,unsigned long long start,unsigned long long end)2826  static ssize_t cifs_write_back_from_locked_folio(struct address_space *mapping,
2827  						 struct writeback_control *wbc,
2828  						 struct xa_state *xas,
2829  						 struct folio *folio,
2830  						 unsigned long long start,
2831  						 unsigned long long end)
2832  {
2833  	struct inode *inode = mapping->host;
2834  	struct TCP_Server_Info *server;
2835  	struct cifs_writedata *wdata;
2836  	struct cifs_sb_info *cifs_sb = CIFS_SB(inode->i_sb);
2837  	struct cifs_credits credits_on_stack;
2838  	struct cifs_credits *credits = &credits_on_stack;
2839  	struct cifsFileInfo *cfile = NULL;
2840  	unsigned long long i_size = i_size_read(inode), max_len;
2841  	unsigned int xid, wsize;
2842  	size_t len = folio_size(folio);
2843  	long count = wbc->nr_to_write;
2844  	int rc;
2845  
2846  	/* The folio should be locked, dirty and not undergoing writeback. */
2847  	if (!folio_clear_dirty_for_io(folio))
2848  		WARN_ON_ONCE(1);
2849  	folio_start_writeback(folio);
2850  
2851  	count -= folio_nr_pages(folio);
2852  
2853  	xid = get_xid();
2854  	server = cifs_pick_channel(cifs_sb_master_tcon(cifs_sb)->ses);
2855  
2856  	rc = cifs_get_writable_file(CIFS_I(inode), FIND_WR_ANY, &cfile);
2857  	if (rc) {
2858  		cifs_dbg(VFS, "No writable handle in writepages rc=%d\n", rc);
2859  		goto err_xid;
2860  	}
2861  
2862  	rc = server->ops->wait_mtu_credits(server, cifs_sb->ctx->wsize,
2863  					   &wsize, credits);
2864  	if (rc != 0)
2865  		goto err_close;
2866  
2867  	wdata = cifs_writedata_alloc(cifs_writev_complete);
2868  	if (!wdata) {
2869  		rc = -ENOMEM;
2870  		goto err_uncredit;
2871  	}
2872  
2873  	wdata->sync_mode = wbc->sync_mode;
2874  	wdata->offset = folio_pos(folio);
2875  	wdata->pid = cfile->pid;
2876  	wdata->credits = credits_on_stack;
2877  	wdata->cfile = cfile;
2878  	wdata->server = server;
2879  	cfile = NULL;
2880  
2881  	/* Find all consecutive lockable dirty pages that have contiguous
2882  	 * written regions, stopping when we find a page that is not
2883  	 * immediately lockable, is not dirty or is missing, or we reach the
2884  	 * end of the range.
2885  	 */
2886  	if (start < i_size) {
2887  		/* Trim the write to the EOF; the extra data is ignored.  Also
2888  		 * put an upper limit on the size of a single storedata op.
2889  		 */
2890  		max_len = wsize;
2891  		max_len = min_t(unsigned long long, max_len, end - start + 1);
2892  		max_len = min_t(unsigned long long, max_len, i_size - start);
2893  
2894  		if (len < max_len) {
2895  			int max_pages = INT_MAX;
2896  
2897  #ifdef CONFIG_CIFS_SMB_DIRECT
2898  			if (server->smbd_conn)
2899  				max_pages = server->smbd_conn->max_frmr_depth;
2900  #endif
2901  			max_pages -= folio_nr_pages(folio);
2902  
2903  			if (max_pages > 0)
2904  				cifs_extend_writeback(mapping, xas, &count, start,
2905  						      max_pages, max_len, &len);
2906  		}
2907  	}
2908  	len = min_t(unsigned long long, len, i_size - start);
2909  
2910  	/* We now have a contiguous set of dirty pages, each with writeback
2911  	 * set; the first page is still locked at this point, but all the rest
2912  	 * have been unlocked.
2913  	 */
2914  	folio_unlock(folio);
2915  	wdata->bytes = len;
2916  
2917  	if (start < i_size) {
2918  		iov_iter_xarray(&wdata->iter, ITER_SOURCE, &mapping->i_pages,
2919  				start, len);
2920  
2921  		rc = adjust_credits(wdata->server, &wdata->credits, wdata->bytes);
2922  		if (rc)
2923  			goto err_wdata;
2924  
2925  		if (wdata->cfile->invalidHandle)
2926  			rc = -EAGAIN;
2927  		else
2928  			rc = wdata->server->ops->async_writev(wdata,
2929  							      cifs_writedata_release);
2930  		if (rc >= 0) {
2931  			kref_put(&wdata->refcount, cifs_writedata_release);
2932  			goto err_close;
2933  		}
2934  	} else {
2935  		/* The dirty region was entirely beyond the EOF. */
2936  		cifs_pages_written_back(inode, start, len);
2937  		rc = 0;
2938  	}
2939  
2940  err_wdata:
2941  	kref_put(&wdata->refcount, cifs_writedata_release);
2942  err_uncredit:
2943  	add_credits_and_wake_if(server, credits, 0);
2944  err_close:
2945  	if (cfile)
2946  		cifsFileInfo_put(cfile);
2947  err_xid:
2948  	free_xid(xid);
2949  	if (rc == 0) {
2950  		wbc->nr_to_write = count;
2951  		rc = len;
2952  	} else if (is_retryable_error(rc)) {
2953  		cifs_pages_write_redirty(inode, start, len);
2954  	} else {
2955  		cifs_pages_write_failed(inode, start, len);
2956  		mapping_set_error(mapping, rc);
2957  	}
2958  	/* Indication to update ctime and mtime as close is deferred */
2959  	set_bit(CIFS_INO_MODIFIED_ATTR, &CIFS_I(inode)->flags);
2960  	return rc;
2961  }
2962  
2963  /*
2964   * write a region of pages back to the server
2965   */
cifs_writepages_begin(struct address_space * mapping,struct writeback_control * wbc,struct xa_state * xas,unsigned long long * _start,unsigned long long end)2966  static ssize_t cifs_writepages_begin(struct address_space *mapping,
2967  				     struct writeback_control *wbc,
2968  				     struct xa_state *xas,
2969  				     unsigned long long *_start,
2970  				     unsigned long long end)
2971  {
2972  	struct folio *folio;
2973  	unsigned long long start = *_start;
2974  	ssize_t ret;
2975  	int skips = 0;
2976  
2977  search_again:
2978  	/* Find the first dirty page. */
2979  	rcu_read_lock();
2980  
2981  	for (;;) {
2982  		folio = xas_find_marked(xas, end / PAGE_SIZE, PAGECACHE_TAG_DIRTY);
2983  		if (xas_retry(xas, folio) || xa_is_value(folio))
2984  			continue;
2985  		if (!folio)
2986  			break;
2987  
2988  		if (!folio_try_get(folio)) {
2989  			xas_reset(xas);
2990  			continue;
2991  		}
2992  
2993  		if (unlikely(folio != xas_reload(xas))) {
2994  			folio_put(folio);
2995  			xas_reset(xas);
2996  			continue;
2997  		}
2998  
2999  		xas_pause(xas);
3000  		break;
3001  	}
3002  	rcu_read_unlock();
3003  	if (!folio)
3004  		return 0;
3005  
3006  	start = folio_pos(folio); /* May regress with THPs */
3007  
3008  	/* At this point we hold neither the i_pages lock nor the page lock:
3009  	 * the page may be truncated or invalidated (changing page->mapping to
3010  	 * NULL), or even swizzled back from swapper_space to tmpfs file
3011  	 * mapping
3012  	 */
3013  lock_again:
3014  	if (wbc->sync_mode != WB_SYNC_NONE) {
3015  		ret = folio_lock_killable(folio);
3016  		if (ret < 0)
3017  			return ret;
3018  	} else {
3019  		if (!folio_trylock(folio))
3020  			goto search_again;
3021  	}
3022  
3023  	if (folio->mapping != mapping ||
3024  	    !folio_test_dirty(folio)) {
3025  		start += folio_size(folio);
3026  		folio_unlock(folio);
3027  		goto search_again;
3028  	}
3029  
3030  	if (folio_test_writeback(folio) ||
3031  	    folio_test_fscache(folio)) {
3032  		folio_unlock(folio);
3033  		if (wbc->sync_mode != WB_SYNC_NONE) {
3034  			folio_wait_writeback(folio);
3035  #ifdef CONFIG_CIFS_FSCACHE
3036  			folio_wait_fscache(folio);
3037  #endif
3038  			goto lock_again;
3039  		}
3040  
3041  		start += folio_size(folio);
3042  		if (wbc->sync_mode == WB_SYNC_NONE) {
3043  			if (skips >= 5 || need_resched()) {
3044  				ret = 0;
3045  				goto out;
3046  			}
3047  			skips++;
3048  		}
3049  		goto search_again;
3050  	}
3051  
3052  	ret = cifs_write_back_from_locked_folio(mapping, wbc, xas, folio, start, end);
3053  out:
3054  	if (ret > 0)
3055  		*_start = start + ret;
3056  	return ret;
3057  }
3058  
3059  /*
3060   * Write a region of pages back to the server
3061   */
cifs_writepages_region(struct address_space * mapping,struct writeback_control * wbc,unsigned long long * _start,unsigned long long end)3062  static int cifs_writepages_region(struct address_space *mapping,
3063  				  struct writeback_control *wbc,
3064  				  unsigned long long *_start,
3065  				  unsigned long long end)
3066  {
3067  	ssize_t ret;
3068  
3069  	XA_STATE(xas, &mapping->i_pages, *_start / PAGE_SIZE);
3070  
3071  	do {
3072  		ret = cifs_writepages_begin(mapping, wbc, &xas, _start, end);
3073  		if (ret > 0 && wbc->nr_to_write > 0)
3074  			cond_resched();
3075  	} while (ret > 0 && wbc->nr_to_write > 0);
3076  
3077  	return ret > 0 ? 0 : ret;
3078  }
3079  
3080  /*
3081   * Write some of the pending data back to the server
3082   */
cifs_writepages(struct address_space * mapping,struct writeback_control * wbc)3083  static int cifs_writepages(struct address_space *mapping,
3084  			   struct writeback_control *wbc)
3085  {
3086  	loff_t start, end;
3087  	int ret;
3088  
3089  	/* We have to be careful as we can end up racing with setattr()
3090  	 * truncating the pagecache since the caller doesn't take a lock here
3091  	 * to prevent it.
3092  	 */
3093  
3094  	if (wbc->range_cyclic && mapping->writeback_index) {
3095  		start = mapping->writeback_index * PAGE_SIZE;
3096  		ret = cifs_writepages_region(mapping, wbc, &start, LLONG_MAX);
3097  		if (ret < 0)
3098  			goto out;
3099  
3100  		if (wbc->nr_to_write <= 0) {
3101  			mapping->writeback_index = start / PAGE_SIZE;
3102  			goto out;
3103  		}
3104  
3105  		start = 0;
3106  		end = mapping->writeback_index * PAGE_SIZE;
3107  		mapping->writeback_index = 0;
3108  		ret = cifs_writepages_region(mapping, wbc, &start, end);
3109  		if (ret == 0)
3110  			mapping->writeback_index = start / PAGE_SIZE;
3111  	} else if (wbc->range_start == 0 && wbc->range_end == LLONG_MAX) {
3112  		start = 0;
3113  		ret = cifs_writepages_region(mapping, wbc, &start, LLONG_MAX);
3114  		if (wbc->nr_to_write > 0 && ret == 0)
3115  			mapping->writeback_index = start / PAGE_SIZE;
3116  	} else {
3117  		start = wbc->range_start;
3118  		ret = cifs_writepages_region(mapping, wbc, &start, wbc->range_end);
3119  	}
3120  
3121  out:
3122  	return ret;
3123  }
3124  
3125  static int
cifs_writepage_locked(struct page * page,struct writeback_control * wbc)3126  cifs_writepage_locked(struct page *page, struct writeback_control *wbc)
3127  {
3128  	int rc;
3129  	unsigned int xid;
3130  
3131  	xid = get_xid();
3132  /* BB add check for wbc flags */
3133  	get_page(page);
3134  	if (!PageUptodate(page))
3135  		cifs_dbg(FYI, "ppw - page not up to date\n");
3136  
3137  	/*
3138  	 * Set the "writeback" flag, and clear "dirty" in the radix tree.
3139  	 *
3140  	 * A writepage() implementation always needs to do either this,
3141  	 * or re-dirty the page with "redirty_page_for_writepage()" in
3142  	 * the case of a failure.
3143  	 *
3144  	 * Just unlocking the page will cause the radix tree tag-bits
3145  	 * to fail to update with the state of the page correctly.
3146  	 */
3147  	set_page_writeback(page);
3148  retry_write:
3149  	rc = cifs_partialpagewrite(page, 0, PAGE_SIZE);
3150  	if (is_retryable_error(rc)) {
3151  		if (wbc->sync_mode == WB_SYNC_ALL && rc == -EAGAIN)
3152  			goto retry_write;
3153  		redirty_page_for_writepage(wbc, page);
3154  	} else if (rc != 0) {
3155  		SetPageError(page);
3156  		mapping_set_error(page->mapping, rc);
3157  	} else {
3158  		SetPageUptodate(page);
3159  	}
3160  	end_page_writeback(page);
3161  	put_page(page);
3162  	free_xid(xid);
3163  	return rc;
3164  }
3165  
cifs_write_end(struct file * file,struct address_space * mapping,loff_t pos,unsigned len,unsigned copied,struct page * page,void * fsdata)3166  static int cifs_write_end(struct file *file, struct address_space *mapping,
3167  			loff_t pos, unsigned len, unsigned copied,
3168  			struct page *page, void *fsdata)
3169  {
3170  	int rc;
3171  	struct inode *inode = mapping->host;
3172  	struct cifsFileInfo *cfile = file->private_data;
3173  	struct cifs_sb_info *cifs_sb = CIFS_SB(cfile->dentry->d_sb);
3174  	struct folio *folio = page_folio(page);
3175  	__u32 pid;
3176  
3177  	if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_RWPIDFORWARD)
3178  		pid = cfile->pid;
3179  	else
3180  		pid = current->tgid;
3181  
3182  	cifs_dbg(FYI, "write_end for page %p from pos %lld with %d bytes\n",
3183  		 page, pos, copied);
3184  
3185  	if (folio_test_checked(folio)) {
3186  		if (copied == len)
3187  			folio_mark_uptodate(folio);
3188  		folio_clear_checked(folio);
3189  	} else if (!folio_test_uptodate(folio) && copied == PAGE_SIZE)
3190  		folio_mark_uptodate(folio);
3191  
3192  	if (!folio_test_uptodate(folio)) {
3193  		char *page_data;
3194  		unsigned offset = pos & (PAGE_SIZE - 1);
3195  		unsigned int xid;
3196  
3197  		xid = get_xid();
3198  		/* this is probably better than directly calling
3199  		   partialpage_write since in this function the file handle is
3200  		   known which we might as well	leverage */
3201  		/* BB check if anything else missing out of ppw
3202  		   such as updating last write time */
3203  		page_data = kmap(page);
3204  		rc = cifs_write(cfile, pid, page_data + offset, copied, &pos);
3205  		/* if (rc < 0) should we set writebehind rc? */
3206  		kunmap(page);
3207  
3208  		free_xid(xid);
3209  	} else {
3210  		rc = copied;
3211  		pos += copied;
3212  		set_page_dirty(page);
3213  	}
3214  
3215  	if (rc > 0) {
3216  		spin_lock(&inode->i_lock);
3217  		if (pos > inode->i_size) {
3218  			loff_t additional_blocks = (512 - 1 + copied) >> 9;
3219  
3220  			i_size_write(inode, pos);
3221  			/*
3222  			 * Estimate new allocation size based on the amount written.
3223  			 * This will be updated from server on close (and on queryinfo)
3224  			 */
3225  			inode->i_blocks = min_t(blkcnt_t, (512 - 1 + pos) >> 9,
3226  						inode->i_blocks + additional_blocks);
3227  		}
3228  		spin_unlock(&inode->i_lock);
3229  	}
3230  
3231  	unlock_page(page);
3232  	put_page(page);
3233  	/* Indication to update ctime and mtime as close is deferred */
3234  	set_bit(CIFS_INO_MODIFIED_ATTR, &CIFS_I(inode)->flags);
3235  
3236  	return rc;
3237  }
3238  
cifs_strict_fsync(struct file * file,loff_t start,loff_t end,int datasync)3239  int cifs_strict_fsync(struct file *file, loff_t start, loff_t end,
3240  		      int datasync)
3241  {
3242  	unsigned int xid;
3243  	int rc = 0;
3244  	struct cifs_tcon *tcon;
3245  	struct TCP_Server_Info *server;
3246  	struct cifsFileInfo *smbfile = file->private_data;
3247  	struct inode *inode = file_inode(file);
3248  	struct cifs_sb_info *cifs_sb = CIFS_SB(inode->i_sb);
3249  
3250  	rc = file_write_and_wait_range(file, start, end);
3251  	if (rc) {
3252  		trace_cifs_fsync_err(inode->i_ino, rc);
3253  		return rc;
3254  	}
3255  
3256  	xid = get_xid();
3257  
3258  	cifs_dbg(FYI, "Sync file - name: %pD datasync: 0x%x\n",
3259  		 file, datasync);
3260  
3261  	if (!CIFS_CACHE_READ(CIFS_I(inode))) {
3262  		rc = cifs_zap_mapping(inode);
3263  		if (rc) {
3264  			cifs_dbg(FYI, "rc: %d during invalidate phase\n", rc);
3265  			rc = 0; /* don't care about it in fsync */
3266  		}
3267  	}
3268  
3269  	tcon = tlink_tcon(smbfile->tlink);
3270  	if (!(cifs_sb->mnt_cifs_flags & CIFS_MOUNT_NOSSYNC)) {
3271  		server = tcon->ses->server;
3272  		if (server->ops->flush == NULL) {
3273  			rc = -ENOSYS;
3274  			goto strict_fsync_exit;
3275  		}
3276  
3277  		if ((OPEN_FMODE(smbfile->f_flags) & FMODE_WRITE) == 0) {
3278  			smbfile = find_writable_file(CIFS_I(inode), FIND_WR_ANY);
3279  			if (smbfile) {
3280  				rc = server->ops->flush(xid, tcon, &smbfile->fid);
3281  				cifsFileInfo_put(smbfile);
3282  			} else
3283  				cifs_dbg(FYI, "ignore fsync for file not open for write\n");
3284  		} else
3285  			rc = server->ops->flush(xid, tcon, &smbfile->fid);
3286  	}
3287  
3288  strict_fsync_exit:
3289  	free_xid(xid);
3290  	return rc;
3291  }
3292  
cifs_fsync(struct file * file,loff_t start,loff_t end,int datasync)3293  int cifs_fsync(struct file *file, loff_t start, loff_t end, int datasync)
3294  {
3295  	unsigned int xid;
3296  	int rc = 0;
3297  	struct cifs_tcon *tcon;
3298  	struct TCP_Server_Info *server;
3299  	struct cifsFileInfo *smbfile = file->private_data;
3300  	struct inode *inode = file_inode(file);
3301  	struct cifs_sb_info *cifs_sb = CIFS_FILE_SB(file);
3302  
3303  	rc = file_write_and_wait_range(file, start, end);
3304  	if (rc) {
3305  		trace_cifs_fsync_err(file_inode(file)->i_ino, rc);
3306  		return rc;
3307  	}
3308  
3309  	xid = get_xid();
3310  
3311  	cifs_dbg(FYI, "Sync file - name: %pD datasync: 0x%x\n",
3312  		 file, datasync);
3313  
3314  	tcon = tlink_tcon(smbfile->tlink);
3315  	if (!(cifs_sb->mnt_cifs_flags & CIFS_MOUNT_NOSSYNC)) {
3316  		server = tcon->ses->server;
3317  		if (server->ops->flush == NULL) {
3318  			rc = -ENOSYS;
3319  			goto fsync_exit;
3320  		}
3321  
3322  		if ((OPEN_FMODE(smbfile->f_flags) & FMODE_WRITE) == 0) {
3323  			smbfile = find_writable_file(CIFS_I(inode), FIND_WR_ANY);
3324  			if (smbfile) {
3325  				rc = server->ops->flush(xid, tcon, &smbfile->fid);
3326  				cifsFileInfo_put(smbfile);
3327  			} else
3328  				cifs_dbg(FYI, "ignore fsync for file not open for write\n");
3329  		} else
3330  			rc = server->ops->flush(xid, tcon, &smbfile->fid);
3331  	}
3332  
3333  fsync_exit:
3334  	free_xid(xid);
3335  	return rc;
3336  }
3337  
3338  /*
3339   * As file closes, flush all cached write data for this inode checking
3340   * for write behind errors.
3341   */
cifs_flush(struct file * file,fl_owner_t id)3342  int cifs_flush(struct file *file, fl_owner_t id)
3343  {
3344  	struct inode *inode = file_inode(file);
3345  	int rc = 0;
3346  
3347  	if (file->f_mode & FMODE_WRITE)
3348  		rc = filemap_write_and_wait(inode->i_mapping);
3349  
3350  	cifs_dbg(FYI, "Flush inode %p file %p rc %d\n", inode, file, rc);
3351  	if (rc) {
3352  		/* get more nuanced writeback errors */
3353  		rc = filemap_check_wb_err(file->f_mapping, 0);
3354  		trace_cifs_flush_err(inode->i_ino, rc);
3355  	}
3356  	return rc;
3357  }
3358  
3359  static void
cifs_uncached_writedata_release(struct kref * refcount)3360  cifs_uncached_writedata_release(struct kref *refcount)
3361  {
3362  	struct cifs_writedata *wdata = container_of(refcount,
3363  					struct cifs_writedata, refcount);
3364  
3365  	kref_put(&wdata->ctx->refcount, cifs_aio_ctx_release);
3366  	cifs_writedata_release(refcount);
3367  }
3368  
3369  static void collect_uncached_write_data(struct cifs_aio_ctx *ctx);
3370  
3371  static void
cifs_uncached_writev_complete(struct work_struct * work)3372  cifs_uncached_writev_complete(struct work_struct *work)
3373  {
3374  	struct cifs_writedata *wdata = container_of(work,
3375  					struct cifs_writedata, work);
3376  	struct inode *inode = d_inode(wdata->cfile->dentry);
3377  	struct cifsInodeInfo *cifsi = CIFS_I(inode);
3378  
3379  	spin_lock(&inode->i_lock);
3380  	cifs_update_eof(cifsi, wdata->offset, wdata->bytes);
3381  	if (cifsi->server_eof > inode->i_size)
3382  		i_size_write(inode, cifsi->server_eof);
3383  	spin_unlock(&inode->i_lock);
3384  
3385  	complete(&wdata->done);
3386  	collect_uncached_write_data(wdata->ctx);
3387  	/* the below call can possibly free the last ref to aio ctx */
3388  	kref_put(&wdata->refcount, cifs_uncached_writedata_release);
3389  }
3390  
3391  static int
cifs_resend_wdata(struct cifs_writedata * wdata,struct list_head * wdata_list,struct cifs_aio_ctx * ctx)3392  cifs_resend_wdata(struct cifs_writedata *wdata, struct list_head *wdata_list,
3393  	struct cifs_aio_ctx *ctx)
3394  {
3395  	unsigned int wsize;
3396  	struct cifs_credits credits;
3397  	int rc;
3398  	struct TCP_Server_Info *server = wdata->server;
3399  
3400  	do {
3401  		if (wdata->cfile->invalidHandle) {
3402  			rc = cifs_reopen_file(wdata->cfile, false);
3403  			if (rc == -EAGAIN)
3404  				continue;
3405  			else if (rc)
3406  				break;
3407  		}
3408  
3409  
3410  		/*
3411  		 * Wait for credits to resend this wdata.
3412  		 * Note: we are attempting to resend the whole wdata not in
3413  		 * segments
3414  		 */
3415  		do {
3416  			rc = server->ops->wait_mtu_credits(server, wdata->bytes,
3417  						&wsize, &credits);
3418  			if (rc)
3419  				goto fail;
3420  
3421  			if (wsize < wdata->bytes) {
3422  				add_credits_and_wake_if(server, &credits, 0);
3423  				msleep(1000);
3424  			}
3425  		} while (wsize < wdata->bytes);
3426  		wdata->credits = credits;
3427  
3428  		rc = adjust_credits(server, &wdata->credits, wdata->bytes);
3429  
3430  		if (!rc) {
3431  			if (wdata->cfile->invalidHandle)
3432  				rc = -EAGAIN;
3433  			else {
3434  				wdata->replay = true;
3435  #ifdef CONFIG_CIFS_SMB_DIRECT
3436  				if (wdata->mr) {
3437  					wdata->mr->need_invalidate = true;
3438  					smbd_deregister_mr(wdata->mr);
3439  					wdata->mr = NULL;
3440  				}
3441  #endif
3442  				rc = server->ops->async_writev(wdata,
3443  					cifs_uncached_writedata_release);
3444  			}
3445  		}
3446  
3447  		/* If the write was successfully sent, we are done */
3448  		if (!rc) {
3449  			list_add_tail(&wdata->list, wdata_list);
3450  			return 0;
3451  		}
3452  
3453  		/* Roll back credits and retry if needed */
3454  		add_credits_and_wake_if(server, &wdata->credits, 0);
3455  	} while (rc == -EAGAIN);
3456  
3457  fail:
3458  	kref_put(&wdata->refcount, cifs_uncached_writedata_release);
3459  	return rc;
3460  }
3461  
3462  /*
3463   * Select span of a bvec iterator we're going to use.  Limit it by both maximum
3464   * size and maximum number of segments.
3465   */
cifs_limit_bvec_subset(const struct iov_iter * iter,size_t max_size,size_t max_segs,unsigned int * _nsegs)3466  static size_t cifs_limit_bvec_subset(const struct iov_iter *iter, size_t max_size,
3467  				     size_t max_segs, unsigned int *_nsegs)
3468  {
3469  	const struct bio_vec *bvecs = iter->bvec;
3470  	unsigned int nbv = iter->nr_segs, ix = 0, nsegs = 0;
3471  	size_t len, span = 0, n = iter->count;
3472  	size_t skip = iter->iov_offset;
3473  
3474  	if (WARN_ON(!iov_iter_is_bvec(iter)) || n == 0)
3475  		return 0;
3476  
3477  	while (n && ix < nbv && skip) {
3478  		len = bvecs[ix].bv_len;
3479  		if (skip < len)
3480  			break;
3481  		skip -= len;
3482  		n -= len;
3483  		ix++;
3484  	}
3485  
3486  	while (n && ix < nbv) {
3487  		len = min3(n, bvecs[ix].bv_len - skip, max_size);
3488  		span += len;
3489  		max_size -= len;
3490  		nsegs++;
3491  		ix++;
3492  		if (max_size == 0 || nsegs >= max_segs)
3493  			break;
3494  		skip = 0;
3495  		n -= len;
3496  	}
3497  
3498  	*_nsegs = nsegs;
3499  	return span;
3500  }
3501  
3502  static int
cifs_write_from_iter(loff_t fpos,size_t len,struct iov_iter * from,struct cifsFileInfo * open_file,struct cifs_sb_info * cifs_sb,struct list_head * wdata_list,struct cifs_aio_ctx * ctx)3503  cifs_write_from_iter(loff_t fpos, size_t len, struct iov_iter *from,
3504  		     struct cifsFileInfo *open_file,
3505  		     struct cifs_sb_info *cifs_sb, struct list_head *wdata_list,
3506  		     struct cifs_aio_ctx *ctx)
3507  {
3508  	int rc = 0;
3509  	size_t cur_len, max_len;
3510  	struct cifs_writedata *wdata;
3511  	pid_t pid;
3512  	struct TCP_Server_Info *server;
3513  	unsigned int xid, max_segs = INT_MAX;
3514  
3515  	if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_RWPIDFORWARD)
3516  		pid = open_file->pid;
3517  	else
3518  		pid = current->tgid;
3519  
3520  	server = cifs_pick_channel(tlink_tcon(open_file->tlink)->ses);
3521  	xid = get_xid();
3522  
3523  #ifdef CONFIG_CIFS_SMB_DIRECT
3524  	if (server->smbd_conn)
3525  		max_segs = server->smbd_conn->max_frmr_depth;
3526  #endif
3527  
3528  	do {
3529  		struct cifs_credits credits_on_stack;
3530  		struct cifs_credits *credits = &credits_on_stack;
3531  		unsigned int wsize, nsegs = 0;
3532  
3533  		if (signal_pending(current)) {
3534  			rc = -EINTR;
3535  			break;
3536  		}
3537  
3538  		if (open_file->invalidHandle) {
3539  			rc = cifs_reopen_file(open_file, false);
3540  			if (rc == -EAGAIN)
3541  				continue;
3542  			else if (rc)
3543  				break;
3544  		}
3545  
3546  		rc = server->ops->wait_mtu_credits(server, cifs_sb->ctx->wsize,
3547  						   &wsize, credits);
3548  		if (rc)
3549  			break;
3550  
3551  		max_len = min_t(const size_t, len, wsize);
3552  		if (!max_len) {
3553  			rc = -EAGAIN;
3554  			add_credits_and_wake_if(server, credits, 0);
3555  			break;
3556  		}
3557  
3558  		cur_len = cifs_limit_bvec_subset(from, max_len, max_segs, &nsegs);
3559  		cifs_dbg(FYI, "write_from_iter len=%zx/%zx nsegs=%u/%lu/%u\n",
3560  			 cur_len, max_len, nsegs, from->nr_segs, max_segs);
3561  		if (cur_len == 0) {
3562  			rc = -EIO;
3563  			add_credits_and_wake_if(server, credits, 0);
3564  			break;
3565  		}
3566  
3567  		wdata = cifs_writedata_alloc(cifs_uncached_writev_complete);
3568  		if (!wdata) {
3569  			rc = -ENOMEM;
3570  			add_credits_and_wake_if(server, credits, 0);
3571  			break;
3572  		}
3573  
3574  		wdata->sync_mode = WB_SYNC_ALL;
3575  		wdata->offset	= (__u64)fpos;
3576  		wdata->cfile	= cifsFileInfo_get(open_file);
3577  		wdata->server	= server;
3578  		wdata->pid	= pid;
3579  		wdata->bytes	= cur_len;
3580  		wdata->credits	= credits_on_stack;
3581  		wdata->iter	= *from;
3582  		wdata->ctx	= ctx;
3583  		kref_get(&ctx->refcount);
3584  
3585  		iov_iter_truncate(&wdata->iter, cur_len);
3586  
3587  		rc = adjust_credits(server, &wdata->credits, wdata->bytes);
3588  
3589  		if (!rc) {
3590  			if (wdata->cfile->invalidHandle)
3591  				rc = -EAGAIN;
3592  			else
3593  				rc = server->ops->async_writev(wdata,
3594  					cifs_uncached_writedata_release);
3595  		}
3596  
3597  		if (rc) {
3598  			add_credits_and_wake_if(server, &wdata->credits, 0);
3599  			kref_put(&wdata->refcount,
3600  				 cifs_uncached_writedata_release);
3601  			if (rc == -EAGAIN)
3602  				continue;
3603  			break;
3604  		}
3605  
3606  		list_add_tail(&wdata->list, wdata_list);
3607  		iov_iter_advance(from, cur_len);
3608  		fpos += cur_len;
3609  		len -= cur_len;
3610  	} while (len > 0);
3611  
3612  	free_xid(xid);
3613  	return rc;
3614  }
3615  
collect_uncached_write_data(struct cifs_aio_ctx * ctx)3616  static void collect_uncached_write_data(struct cifs_aio_ctx *ctx)
3617  {
3618  	struct cifs_writedata *wdata, *tmp;
3619  	struct cifs_tcon *tcon;
3620  	struct cifs_sb_info *cifs_sb;
3621  	struct dentry *dentry = ctx->cfile->dentry;
3622  	ssize_t rc;
3623  
3624  	tcon = tlink_tcon(ctx->cfile->tlink);
3625  	cifs_sb = CIFS_SB(dentry->d_sb);
3626  
3627  	mutex_lock(&ctx->aio_mutex);
3628  
3629  	if (list_empty(&ctx->list)) {
3630  		mutex_unlock(&ctx->aio_mutex);
3631  		return;
3632  	}
3633  
3634  	rc = ctx->rc;
3635  	/*
3636  	 * Wait for and collect replies for any successful sends in order of
3637  	 * increasing offset. Once an error is hit, then return without waiting
3638  	 * for any more replies.
3639  	 */
3640  restart_loop:
3641  	list_for_each_entry_safe(wdata, tmp, &ctx->list, list) {
3642  		if (!rc) {
3643  			if (!try_wait_for_completion(&wdata->done)) {
3644  				mutex_unlock(&ctx->aio_mutex);
3645  				return;
3646  			}
3647  
3648  			if (wdata->result)
3649  				rc = wdata->result;
3650  			else
3651  				ctx->total_len += wdata->bytes;
3652  
3653  			/* resend call if it's a retryable error */
3654  			if (rc == -EAGAIN) {
3655  				struct list_head tmp_list;
3656  				struct iov_iter tmp_from = ctx->iter;
3657  
3658  				INIT_LIST_HEAD(&tmp_list);
3659  				list_del_init(&wdata->list);
3660  
3661  				if (ctx->direct_io)
3662  					rc = cifs_resend_wdata(
3663  						wdata, &tmp_list, ctx);
3664  				else {
3665  					iov_iter_advance(&tmp_from,
3666  						 wdata->offset - ctx->pos);
3667  
3668  					rc = cifs_write_from_iter(wdata->offset,
3669  						wdata->bytes, &tmp_from,
3670  						ctx->cfile, cifs_sb, &tmp_list,
3671  						ctx);
3672  
3673  					kref_put(&wdata->refcount,
3674  						cifs_uncached_writedata_release);
3675  				}
3676  
3677  				list_splice(&tmp_list, &ctx->list);
3678  				goto restart_loop;
3679  			}
3680  		}
3681  		list_del_init(&wdata->list);
3682  		kref_put(&wdata->refcount, cifs_uncached_writedata_release);
3683  	}
3684  
3685  	cifs_stats_bytes_written(tcon, ctx->total_len);
3686  	set_bit(CIFS_INO_INVALID_MAPPING, &CIFS_I(dentry->d_inode)->flags);
3687  
3688  	ctx->rc = (rc == 0) ? ctx->total_len : rc;
3689  
3690  	mutex_unlock(&ctx->aio_mutex);
3691  
3692  	if (ctx->iocb && ctx->iocb->ki_complete)
3693  		ctx->iocb->ki_complete(ctx->iocb, ctx->rc);
3694  	else
3695  		complete(&ctx->done);
3696  }
3697  
__cifs_writev(struct kiocb * iocb,struct iov_iter * from,bool direct)3698  static ssize_t __cifs_writev(
3699  	struct kiocb *iocb, struct iov_iter *from, bool direct)
3700  {
3701  	struct file *file = iocb->ki_filp;
3702  	ssize_t total_written = 0;
3703  	struct cifsFileInfo *cfile;
3704  	struct cifs_tcon *tcon;
3705  	struct cifs_sb_info *cifs_sb;
3706  	struct cifs_aio_ctx *ctx;
3707  	int rc;
3708  
3709  	rc = generic_write_checks(iocb, from);
3710  	if (rc <= 0)
3711  		return rc;
3712  
3713  	cifs_sb = CIFS_FILE_SB(file);
3714  	cfile = file->private_data;
3715  	tcon = tlink_tcon(cfile->tlink);
3716  
3717  	if (!tcon->ses->server->ops->async_writev)
3718  		return -ENOSYS;
3719  
3720  	ctx = cifs_aio_ctx_alloc();
3721  	if (!ctx)
3722  		return -ENOMEM;
3723  
3724  	ctx->cfile = cifsFileInfo_get(cfile);
3725  
3726  	if (!is_sync_kiocb(iocb))
3727  		ctx->iocb = iocb;
3728  
3729  	ctx->pos = iocb->ki_pos;
3730  	ctx->direct_io = direct;
3731  	ctx->nr_pinned_pages = 0;
3732  
3733  	if (user_backed_iter(from)) {
3734  		/*
3735  		 * Extract IOVEC/UBUF-type iterators to a BVEC-type iterator as
3736  		 * they contain references to the calling process's virtual
3737  		 * memory layout which won't be available in an async worker
3738  		 * thread.  This also takes a pin on every folio involved.
3739  		 */
3740  		rc = netfs_extract_user_iter(from, iov_iter_count(from),
3741  					     &ctx->iter, 0);
3742  		if (rc < 0) {
3743  			kref_put(&ctx->refcount, cifs_aio_ctx_release);
3744  			return rc;
3745  		}
3746  
3747  		ctx->nr_pinned_pages = rc;
3748  		ctx->bv = (void *)ctx->iter.bvec;
3749  		ctx->bv_need_unpin = iov_iter_extract_will_pin(from);
3750  	} else if ((iov_iter_is_bvec(from) || iov_iter_is_kvec(from)) &&
3751  		   !is_sync_kiocb(iocb)) {
3752  		/*
3753  		 * If the op is asynchronous, we need to copy the list attached
3754  		 * to a BVEC/KVEC-type iterator, but we assume that the storage
3755  		 * will be pinned by the caller; in any case, we may or may not
3756  		 * be able to pin the pages, so we don't try.
3757  		 */
3758  		ctx->bv = (void *)dup_iter(&ctx->iter, from, GFP_KERNEL);
3759  		if (!ctx->bv) {
3760  			kref_put(&ctx->refcount, cifs_aio_ctx_release);
3761  			return -ENOMEM;
3762  		}
3763  	} else {
3764  		/*
3765  		 * Otherwise, we just pass the iterator down as-is and rely on
3766  		 * the caller to make sure the pages referred to by the
3767  		 * iterator don't evaporate.
3768  		 */
3769  		ctx->iter = *from;
3770  	}
3771  
3772  	ctx->len = iov_iter_count(&ctx->iter);
3773  
3774  	/* grab a lock here due to read response handlers can access ctx */
3775  	mutex_lock(&ctx->aio_mutex);
3776  
3777  	rc = cifs_write_from_iter(iocb->ki_pos, ctx->len, &ctx->iter,
3778  				  cfile, cifs_sb, &ctx->list, ctx);
3779  
3780  	/*
3781  	 * If at least one write was successfully sent, then discard any rc
3782  	 * value from the later writes. If the other write succeeds, then
3783  	 * we'll end up returning whatever was written. If it fails, then
3784  	 * we'll get a new rc value from that.
3785  	 */
3786  	if (!list_empty(&ctx->list))
3787  		rc = 0;
3788  
3789  	mutex_unlock(&ctx->aio_mutex);
3790  
3791  	if (rc) {
3792  		kref_put(&ctx->refcount, cifs_aio_ctx_release);
3793  		return rc;
3794  	}
3795  
3796  	if (!is_sync_kiocb(iocb)) {
3797  		kref_put(&ctx->refcount, cifs_aio_ctx_release);
3798  		return -EIOCBQUEUED;
3799  	}
3800  
3801  	rc = wait_for_completion_killable(&ctx->done);
3802  	if (rc) {
3803  		mutex_lock(&ctx->aio_mutex);
3804  		ctx->rc = rc = -EINTR;
3805  		total_written = ctx->total_len;
3806  		mutex_unlock(&ctx->aio_mutex);
3807  	} else {
3808  		rc = ctx->rc;
3809  		total_written = ctx->total_len;
3810  	}
3811  
3812  	kref_put(&ctx->refcount, cifs_aio_ctx_release);
3813  
3814  	if (unlikely(!total_written))
3815  		return rc;
3816  
3817  	iocb->ki_pos += total_written;
3818  	return total_written;
3819  }
3820  
cifs_direct_writev(struct kiocb * iocb,struct iov_iter * from)3821  ssize_t cifs_direct_writev(struct kiocb *iocb, struct iov_iter *from)
3822  {
3823  	struct file *file = iocb->ki_filp;
3824  
3825  	cifs_revalidate_mapping(file->f_inode);
3826  	return __cifs_writev(iocb, from, true);
3827  }
3828  
cifs_user_writev(struct kiocb * iocb,struct iov_iter * from)3829  ssize_t cifs_user_writev(struct kiocb *iocb, struct iov_iter *from)
3830  {
3831  	return __cifs_writev(iocb, from, false);
3832  }
3833  
3834  static ssize_t
cifs_writev(struct kiocb * iocb,struct iov_iter * from)3835  cifs_writev(struct kiocb *iocb, struct iov_iter *from)
3836  {
3837  	struct file *file = iocb->ki_filp;
3838  	struct cifsFileInfo *cfile = (struct cifsFileInfo *)file->private_data;
3839  	struct inode *inode = file->f_mapping->host;
3840  	struct cifsInodeInfo *cinode = CIFS_I(inode);
3841  	struct TCP_Server_Info *server = tlink_tcon(cfile->tlink)->ses->server;
3842  	ssize_t rc;
3843  
3844  	inode_lock(inode);
3845  	/*
3846  	 * We need to hold the sem to be sure nobody modifies lock list
3847  	 * with a brlock that prevents writing.
3848  	 */
3849  	down_read(&cinode->lock_sem);
3850  
3851  	rc = generic_write_checks(iocb, from);
3852  	if (rc <= 0)
3853  		goto out;
3854  
3855  	if (!cifs_find_lock_conflict(cfile, iocb->ki_pos, iov_iter_count(from),
3856  				     server->vals->exclusive_lock_type, 0,
3857  				     NULL, CIFS_WRITE_OP))
3858  		rc = __generic_file_write_iter(iocb, from);
3859  	else
3860  		rc = -EACCES;
3861  out:
3862  	up_read(&cinode->lock_sem);
3863  	inode_unlock(inode);
3864  
3865  	if (rc > 0)
3866  		rc = generic_write_sync(iocb, rc);
3867  	return rc;
3868  }
3869  
3870  ssize_t
cifs_strict_writev(struct kiocb * iocb,struct iov_iter * from)3871  cifs_strict_writev(struct kiocb *iocb, struct iov_iter *from)
3872  {
3873  	struct inode *inode = file_inode(iocb->ki_filp);
3874  	struct cifsInodeInfo *cinode = CIFS_I(inode);
3875  	struct cifs_sb_info *cifs_sb = CIFS_SB(inode->i_sb);
3876  	struct cifsFileInfo *cfile = (struct cifsFileInfo *)
3877  						iocb->ki_filp->private_data;
3878  	struct cifs_tcon *tcon = tlink_tcon(cfile->tlink);
3879  	ssize_t written;
3880  
3881  	written = cifs_get_writer(cinode);
3882  	if (written)
3883  		return written;
3884  
3885  	if (CIFS_CACHE_WRITE(cinode)) {
3886  		if (cap_unix(tcon->ses) &&
3887  		(CIFS_UNIX_FCNTL_CAP & le64_to_cpu(tcon->fsUnixInfo.Capability))
3888  		  && ((cifs_sb->mnt_cifs_flags & CIFS_MOUNT_NOPOSIXBRL) == 0)) {
3889  			written = generic_file_write_iter(iocb, from);
3890  			goto out;
3891  		}
3892  		written = cifs_writev(iocb, from);
3893  		goto out;
3894  	}
3895  	/*
3896  	 * For non-oplocked files in strict cache mode we need to write the data
3897  	 * to the server exactly from the pos to pos+len-1 rather than flush all
3898  	 * affected pages because it may cause a error with mandatory locks on
3899  	 * these pages but not on the region from pos to ppos+len-1.
3900  	 */
3901  	written = cifs_user_writev(iocb, from);
3902  	if (CIFS_CACHE_READ(cinode)) {
3903  		/*
3904  		 * We have read level caching and we have just sent a write
3905  		 * request to the server thus making data in the cache stale.
3906  		 * Zap the cache and set oplock/lease level to NONE to avoid
3907  		 * reading stale data from the cache. All subsequent read
3908  		 * operations will read new data from the server.
3909  		 */
3910  		cifs_zap_mapping(inode);
3911  		cifs_dbg(FYI, "Set Oplock/Lease to NONE for inode=%p after write\n",
3912  			 inode);
3913  		cinode->oplock = 0;
3914  	}
3915  out:
3916  	cifs_put_writer(cinode);
3917  	return written;
3918  }
3919  
cifs_readdata_alloc(work_func_t complete)3920  static struct cifs_readdata *cifs_readdata_alloc(work_func_t complete)
3921  {
3922  	struct cifs_readdata *rdata;
3923  
3924  	rdata = kzalloc(sizeof(*rdata), GFP_KERNEL);
3925  	if (rdata) {
3926  		kref_init(&rdata->refcount);
3927  		INIT_LIST_HEAD(&rdata->list);
3928  		init_completion(&rdata->done);
3929  		INIT_WORK(&rdata->work, complete);
3930  	}
3931  
3932  	return rdata;
3933  }
3934  
3935  void
cifs_readdata_release(struct kref * refcount)3936  cifs_readdata_release(struct kref *refcount)
3937  {
3938  	struct cifs_readdata *rdata = container_of(refcount,
3939  					struct cifs_readdata, refcount);
3940  
3941  	if (rdata->ctx)
3942  		kref_put(&rdata->ctx->refcount, cifs_aio_ctx_release);
3943  #ifdef CONFIG_CIFS_SMB_DIRECT
3944  	if (rdata->mr) {
3945  		smbd_deregister_mr(rdata->mr);
3946  		rdata->mr = NULL;
3947  	}
3948  #endif
3949  	if (rdata->cfile)
3950  		cifsFileInfo_put(rdata->cfile);
3951  
3952  	kfree(rdata);
3953  }
3954  
3955  static void collect_uncached_read_data(struct cifs_aio_ctx *ctx);
3956  
3957  static void
cifs_uncached_readv_complete(struct work_struct * work)3958  cifs_uncached_readv_complete(struct work_struct *work)
3959  {
3960  	struct cifs_readdata *rdata = container_of(work,
3961  						struct cifs_readdata, work);
3962  
3963  	complete(&rdata->done);
3964  	collect_uncached_read_data(rdata->ctx);
3965  	/* the below call can possibly free the last ref to aio ctx */
3966  	kref_put(&rdata->refcount, cifs_readdata_release);
3967  }
3968  
cifs_resend_rdata(struct cifs_readdata * rdata,struct list_head * rdata_list,struct cifs_aio_ctx * ctx)3969  static int cifs_resend_rdata(struct cifs_readdata *rdata,
3970  			struct list_head *rdata_list,
3971  			struct cifs_aio_ctx *ctx)
3972  {
3973  	unsigned int rsize;
3974  	struct cifs_credits credits;
3975  	int rc;
3976  	struct TCP_Server_Info *server;
3977  
3978  	/* XXX: should we pick a new channel here? */
3979  	server = rdata->server;
3980  
3981  	do {
3982  		if (rdata->cfile->invalidHandle) {
3983  			rc = cifs_reopen_file(rdata->cfile, true);
3984  			if (rc == -EAGAIN)
3985  				continue;
3986  			else if (rc)
3987  				break;
3988  		}
3989  
3990  		/*
3991  		 * Wait for credits to resend this rdata.
3992  		 * Note: we are attempting to resend the whole rdata not in
3993  		 * segments
3994  		 */
3995  		do {
3996  			rc = server->ops->wait_mtu_credits(server, rdata->bytes,
3997  						&rsize, &credits);
3998  
3999  			if (rc)
4000  				goto fail;
4001  
4002  			if (rsize < rdata->bytes) {
4003  				add_credits_and_wake_if(server, &credits, 0);
4004  				msleep(1000);
4005  			}
4006  		} while (rsize < rdata->bytes);
4007  		rdata->credits = credits;
4008  
4009  		rc = adjust_credits(server, &rdata->credits, rdata->bytes);
4010  		if (!rc) {
4011  			if (rdata->cfile->invalidHandle)
4012  				rc = -EAGAIN;
4013  			else {
4014  #ifdef CONFIG_CIFS_SMB_DIRECT
4015  				if (rdata->mr) {
4016  					rdata->mr->need_invalidate = true;
4017  					smbd_deregister_mr(rdata->mr);
4018  					rdata->mr = NULL;
4019  				}
4020  #endif
4021  				rc = server->ops->async_readv(rdata);
4022  			}
4023  		}
4024  
4025  		/* If the read was successfully sent, we are done */
4026  		if (!rc) {
4027  			/* Add to aio pending list */
4028  			list_add_tail(&rdata->list, rdata_list);
4029  			return 0;
4030  		}
4031  
4032  		/* Roll back credits and retry if needed */
4033  		add_credits_and_wake_if(server, &rdata->credits, 0);
4034  	} while (rc == -EAGAIN);
4035  
4036  fail:
4037  	kref_put(&rdata->refcount, cifs_readdata_release);
4038  	return rc;
4039  }
4040  
4041  static int
cifs_send_async_read(loff_t fpos,size_t len,struct cifsFileInfo * open_file,struct cifs_sb_info * cifs_sb,struct list_head * rdata_list,struct cifs_aio_ctx * ctx)4042  cifs_send_async_read(loff_t fpos, size_t len, struct cifsFileInfo *open_file,
4043  		     struct cifs_sb_info *cifs_sb, struct list_head *rdata_list,
4044  		     struct cifs_aio_ctx *ctx)
4045  {
4046  	struct cifs_readdata *rdata;
4047  	unsigned int rsize, nsegs, max_segs = INT_MAX;
4048  	struct cifs_credits credits_on_stack;
4049  	struct cifs_credits *credits = &credits_on_stack;
4050  	size_t cur_len, max_len;
4051  	int rc;
4052  	pid_t pid;
4053  	struct TCP_Server_Info *server;
4054  
4055  	server = cifs_pick_channel(tlink_tcon(open_file->tlink)->ses);
4056  
4057  #ifdef CONFIG_CIFS_SMB_DIRECT
4058  	if (server->smbd_conn)
4059  		max_segs = server->smbd_conn->max_frmr_depth;
4060  #endif
4061  
4062  	if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_RWPIDFORWARD)
4063  		pid = open_file->pid;
4064  	else
4065  		pid = current->tgid;
4066  
4067  	do {
4068  		if (open_file->invalidHandle) {
4069  			rc = cifs_reopen_file(open_file, true);
4070  			if (rc == -EAGAIN)
4071  				continue;
4072  			else if (rc)
4073  				break;
4074  		}
4075  
4076  		if (cifs_sb->ctx->rsize == 0)
4077  			cifs_sb->ctx->rsize =
4078  				server->ops->negotiate_rsize(tlink_tcon(open_file->tlink),
4079  							     cifs_sb->ctx);
4080  
4081  		rc = server->ops->wait_mtu_credits(server, cifs_sb->ctx->rsize,
4082  						   &rsize, credits);
4083  		if (rc)
4084  			break;
4085  
4086  		max_len = min_t(size_t, len, rsize);
4087  
4088  		cur_len = cifs_limit_bvec_subset(&ctx->iter, max_len,
4089  						 max_segs, &nsegs);
4090  		cifs_dbg(FYI, "read-to-iter len=%zx/%zx nsegs=%u/%lu/%u\n",
4091  			 cur_len, max_len, nsegs, ctx->iter.nr_segs, max_segs);
4092  		if (cur_len == 0) {
4093  			rc = -EIO;
4094  			add_credits_and_wake_if(server, credits, 0);
4095  			break;
4096  		}
4097  
4098  		rdata = cifs_readdata_alloc(cifs_uncached_readv_complete);
4099  		if (!rdata) {
4100  			add_credits_and_wake_if(server, credits, 0);
4101  			rc = -ENOMEM;
4102  			break;
4103  		}
4104  
4105  		rdata->server	= server;
4106  		rdata->cfile	= cifsFileInfo_get(open_file);
4107  		rdata->offset	= fpos;
4108  		rdata->bytes	= cur_len;
4109  		rdata->pid	= pid;
4110  		rdata->credits	= credits_on_stack;
4111  		rdata->ctx	= ctx;
4112  		kref_get(&ctx->refcount);
4113  
4114  		rdata->iter	= ctx->iter;
4115  		iov_iter_truncate(&rdata->iter, cur_len);
4116  
4117  		rc = adjust_credits(server, &rdata->credits, rdata->bytes);
4118  
4119  		if (!rc) {
4120  			if (rdata->cfile->invalidHandle)
4121  				rc = -EAGAIN;
4122  			else
4123  				rc = server->ops->async_readv(rdata);
4124  		}
4125  
4126  		if (rc) {
4127  			add_credits_and_wake_if(server, &rdata->credits, 0);
4128  			kref_put(&rdata->refcount, cifs_readdata_release);
4129  			if (rc == -EAGAIN)
4130  				continue;
4131  			break;
4132  		}
4133  
4134  		list_add_tail(&rdata->list, rdata_list);
4135  		iov_iter_advance(&ctx->iter, cur_len);
4136  		fpos += cur_len;
4137  		len -= cur_len;
4138  	} while (len > 0);
4139  
4140  	return rc;
4141  }
4142  
4143  static void
collect_uncached_read_data(struct cifs_aio_ctx * ctx)4144  collect_uncached_read_data(struct cifs_aio_ctx *ctx)
4145  {
4146  	struct cifs_readdata *rdata, *tmp;
4147  	struct cifs_sb_info *cifs_sb;
4148  	int rc;
4149  
4150  	cifs_sb = CIFS_SB(ctx->cfile->dentry->d_sb);
4151  
4152  	mutex_lock(&ctx->aio_mutex);
4153  
4154  	if (list_empty(&ctx->list)) {
4155  		mutex_unlock(&ctx->aio_mutex);
4156  		return;
4157  	}
4158  
4159  	rc = ctx->rc;
4160  	/* the loop below should proceed in the order of increasing offsets */
4161  again:
4162  	list_for_each_entry_safe(rdata, tmp, &ctx->list, list) {
4163  		if (!rc) {
4164  			if (!try_wait_for_completion(&rdata->done)) {
4165  				mutex_unlock(&ctx->aio_mutex);
4166  				return;
4167  			}
4168  
4169  			if (rdata->result == -EAGAIN) {
4170  				/* resend call if it's a retryable error */
4171  				struct list_head tmp_list;
4172  				unsigned int got_bytes = rdata->got_bytes;
4173  
4174  				list_del_init(&rdata->list);
4175  				INIT_LIST_HEAD(&tmp_list);
4176  
4177  				if (ctx->direct_io) {
4178  					/*
4179  					 * Re-use rdata as this is a
4180  					 * direct I/O
4181  					 */
4182  					rc = cifs_resend_rdata(
4183  						rdata,
4184  						&tmp_list, ctx);
4185  				} else {
4186  					rc = cifs_send_async_read(
4187  						rdata->offset + got_bytes,
4188  						rdata->bytes - got_bytes,
4189  						rdata->cfile, cifs_sb,
4190  						&tmp_list, ctx);
4191  
4192  					kref_put(&rdata->refcount,
4193  						cifs_readdata_release);
4194  				}
4195  
4196  				list_splice(&tmp_list, &ctx->list);
4197  
4198  				goto again;
4199  			} else if (rdata->result)
4200  				rc = rdata->result;
4201  
4202  			/* if there was a short read -- discard anything left */
4203  			if (rdata->got_bytes && rdata->got_bytes < rdata->bytes)
4204  				rc = -ENODATA;
4205  
4206  			ctx->total_len += rdata->got_bytes;
4207  		}
4208  		list_del_init(&rdata->list);
4209  		kref_put(&rdata->refcount, cifs_readdata_release);
4210  	}
4211  
4212  	/* mask nodata case */
4213  	if (rc == -ENODATA)
4214  		rc = 0;
4215  
4216  	ctx->rc = (rc == 0) ? (ssize_t)ctx->total_len : rc;
4217  
4218  	mutex_unlock(&ctx->aio_mutex);
4219  
4220  	if (ctx->iocb && ctx->iocb->ki_complete)
4221  		ctx->iocb->ki_complete(ctx->iocb, ctx->rc);
4222  	else
4223  		complete(&ctx->done);
4224  }
4225  
__cifs_readv(struct kiocb * iocb,struct iov_iter * to,bool direct)4226  static ssize_t __cifs_readv(
4227  	struct kiocb *iocb, struct iov_iter *to, bool direct)
4228  {
4229  	size_t len;
4230  	struct file *file = iocb->ki_filp;
4231  	struct cifs_sb_info *cifs_sb;
4232  	struct cifsFileInfo *cfile;
4233  	struct cifs_tcon *tcon;
4234  	ssize_t rc, total_read = 0;
4235  	loff_t offset = iocb->ki_pos;
4236  	struct cifs_aio_ctx *ctx;
4237  
4238  	len = iov_iter_count(to);
4239  	if (!len)
4240  		return 0;
4241  
4242  	cifs_sb = CIFS_FILE_SB(file);
4243  	cfile = file->private_data;
4244  	tcon = tlink_tcon(cfile->tlink);
4245  
4246  	if (!tcon->ses->server->ops->async_readv)
4247  		return -ENOSYS;
4248  
4249  	if ((file->f_flags & O_ACCMODE) == O_WRONLY)
4250  		cifs_dbg(FYI, "attempting read on write only file instance\n");
4251  
4252  	ctx = cifs_aio_ctx_alloc();
4253  	if (!ctx)
4254  		return -ENOMEM;
4255  
4256  	ctx->pos	= offset;
4257  	ctx->direct_io	= direct;
4258  	ctx->len	= len;
4259  	ctx->cfile	= cifsFileInfo_get(cfile);
4260  	ctx->nr_pinned_pages = 0;
4261  
4262  	if (!is_sync_kiocb(iocb))
4263  		ctx->iocb = iocb;
4264  
4265  	if (user_backed_iter(to)) {
4266  		/*
4267  		 * Extract IOVEC/UBUF-type iterators to a BVEC-type iterator as
4268  		 * they contain references to the calling process's virtual
4269  		 * memory layout which won't be available in an async worker
4270  		 * thread.  This also takes a pin on every folio involved.
4271  		 */
4272  		rc = netfs_extract_user_iter(to, iov_iter_count(to),
4273  					     &ctx->iter, 0);
4274  		if (rc < 0) {
4275  			kref_put(&ctx->refcount, cifs_aio_ctx_release);
4276  			return rc;
4277  		}
4278  
4279  		ctx->nr_pinned_pages = rc;
4280  		ctx->bv = (void *)ctx->iter.bvec;
4281  		ctx->bv_need_unpin = iov_iter_extract_will_pin(to);
4282  		ctx->should_dirty = true;
4283  	} else if ((iov_iter_is_bvec(to) || iov_iter_is_kvec(to)) &&
4284  		   !is_sync_kiocb(iocb)) {
4285  		/*
4286  		 * If the op is asynchronous, we need to copy the list attached
4287  		 * to a BVEC/KVEC-type iterator, but we assume that the storage
4288  		 * will be retained by the caller; in any case, we may or may
4289  		 * not be able to pin the pages, so we don't try.
4290  		 */
4291  		ctx->bv = (void *)dup_iter(&ctx->iter, to, GFP_KERNEL);
4292  		if (!ctx->bv) {
4293  			kref_put(&ctx->refcount, cifs_aio_ctx_release);
4294  			return -ENOMEM;
4295  		}
4296  	} else {
4297  		/*
4298  		 * Otherwise, we just pass the iterator down as-is and rely on
4299  		 * the caller to make sure the pages referred to by the
4300  		 * iterator don't evaporate.
4301  		 */
4302  		ctx->iter = *to;
4303  	}
4304  
4305  	if (direct) {
4306  		rc = filemap_write_and_wait_range(file->f_inode->i_mapping,
4307  						  offset, offset + len - 1);
4308  		if (rc) {
4309  			kref_put(&ctx->refcount, cifs_aio_ctx_release);
4310  			return -EAGAIN;
4311  		}
4312  	}
4313  
4314  	/* grab a lock here due to read response handlers can access ctx */
4315  	mutex_lock(&ctx->aio_mutex);
4316  
4317  	rc = cifs_send_async_read(offset, len, cfile, cifs_sb, &ctx->list, ctx);
4318  
4319  	/* if at least one read request send succeeded, then reset rc */
4320  	if (!list_empty(&ctx->list))
4321  		rc = 0;
4322  
4323  	mutex_unlock(&ctx->aio_mutex);
4324  
4325  	if (rc) {
4326  		kref_put(&ctx->refcount, cifs_aio_ctx_release);
4327  		return rc;
4328  	}
4329  
4330  	if (!is_sync_kiocb(iocb)) {
4331  		kref_put(&ctx->refcount, cifs_aio_ctx_release);
4332  		return -EIOCBQUEUED;
4333  	}
4334  
4335  	rc = wait_for_completion_killable(&ctx->done);
4336  	if (rc) {
4337  		mutex_lock(&ctx->aio_mutex);
4338  		ctx->rc = rc = -EINTR;
4339  		total_read = ctx->total_len;
4340  		mutex_unlock(&ctx->aio_mutex);
4341  	} else {
4342  		rc = ctx->rc;
4343  		total_read = ctx->total_len;
4344  	}
4345  
4346  	kref_put(&ctx->refcount, cifs_aio_ctx_release);
4347  
4348  	if (total_read) {
4349  		iocb->ki_pos += total_read;
4350  		return total_read;
4351  	}
4352  	return rc;
4353  }
4354  
cifs_direct_readv(struct kiocb * iocb,struct iov_iter * to)4355  ssize_t cifs_direct_readv(struct kiocb *iocb, struct iov_iter *to)
4356  {
4357  	return __cifs_readv(iocb, to, true);
4358  }
4359  
cifs_user_readv(struct kiocb * iocb,struct iov_iter * to)4360  ssize_t cifs_user_readv(struct kiocb *iocb, struct iov_iter *to)
4361  {
4362  	return __cifs_readv(iocb, to, false);
4363  }
4364  
4365  ssize_t
cifs_strict_readv(struct kiocb * iocb,struct iov_iter * to)4366  cifs_strict_readv(struct kiocb *iocb, struct iov_iter *to)
4367  {
4368  	struct inode *inode = file_inode(iocb->ki_filp);
4369  	struct cifsInodeInfo *cinode = CIFS_I(inode);
4370  	struct cifs_sb_info *cifs_sb = CIFS_SB(inode->i_sb);
4371  	struct cifsFileInfo *cfile = (struct cifsFileInfo *)
4372  						iocb->ki_filp->private_data;
4373  	struct cifs_tcon *tcon = tlink_tcon(cfile->tlink);
4374  	int rc = -EACCES;
4375  
4376  	/*
4377  	 * In strict cache mode we need to read from the server all the time
4378  	 * if we don't have level II oplock because the server can delay mtime
4379  	 * change - so we can't make a decision about inode invalidating.
4380  	 * And we can also fail with pagereading if there are mandatory locks
4381  	 * on pages affected by this read but not on the region from pos to
4382  	 * pos+len-1.
4383  	 */
4384  	if (!CIFS_CACHE_READ(cinode))
4385  		return cifs_user_readv(iocb, to);
4386  
4387  	if (cap_unix(tcon->ses) &&
4388  	    (CIFS_UNIX_FCNTL_CAP & le64_to_cpu(tcon->fsUnixInfo.Capability)) &&
4389  	    ((cifs_sb->mnt_cifs_flags & CIFS_MOUNT_NOPOSIXBRL) == 0))
4390  		return generic_file_read_iter(iocb, to);
4391  
4392  	/*
4393  	 * We need to hold the sem to be sure nobody modifies lock list
4394  	 * with a brlock that prevents reading.
4395  	 */
4396  	down_read(&cinode->lock_sem);
4397  	if (!cifs_find_lock_conflict(cfile, iocb->ki_pos, iov_iter_count(to),
4398  				     tcon->ses->server->vals->shared_lock_type,
4399  				     0, NULL, CIFS_READ_OP))
4400  		rc = generic_file_read_iter(iocb, to);
4401  	up_read(&cinode->lock_sem);
4402  	return rc;
4403  }
4404  
4405  static ssize_t
cifs_read(struct file * file,char * read_data,size_t read_size,loff_t * offset)4406  cifs_read(struct file *file, char *read_data, size_t read_size, loff_t *offset)
4407  {
4408  	int rc = -EACCES;
4409  	unsigned int bytes_read = 0;
4410  	unsigned int total_read;
4411  	unsigned int current_read_size;
4412  	unsigned int rsize;
4413  	struct cifs_sb_info *cifs_sb;
4414  	struct cifs_tcon *tcon;
4415  	struct TCP_Server_Info *server;
4416  	unsigned int xid;
4417  	char *cur_offset;
4418  	struct cifsFileInfo *open_file;
4419  	struct cifs_io_parms io_parms = {0};
4420  	int buf_type = CIFS_NO_BUFFER;
4421  	__u32 pid;
4422  
4423  	xid = get_xid();
4424  	cifs_sb = CIFS_FILE_SB(file);
4425  
4426  	/* FIXME: set up handlers for larger reads and/or convert to async */
4427  	rsize = min_t(unsigned int, cifs_sb->ctx->rsize, CIFSMaxBufSize);
4428  
4429  	if (file->private_data == NULL) {
4430  		rc = -EBADF;
4431  		free_xid(xid);
4432  		return rc;
4433  	}
4434  	open_file = file->private_data;
4435  	tcon = tlink_tcon(open_file->tlink);
4436  	server = cifs_pick_channel(tcon->ses);
4437  
4438  	if (!server->ops->sync_read) {
4439  		free_xid(xid);
4440  		return -ENOSYS;
4441  	}
4442  
4443  	if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_RWPIDFORWARD)
4444  		pid = open_file->pid;
4445  	else
4446  		pid = current->tgid;
4447  
4448  	if ((file->f_flags & O_ACCMODE) == O_WRONLY)
4449  		cifs_dbg(FYI, "attempting read on write only file instance\n");
4450  
4451  	for (total_read = 0, cur_offset = read_data; read_size > total_read;
4452  	     total_read += bytes_read, cur_offset += bytes_read) {
4453  		do {
4454  			current_read_size = min_t(uint, read_size - total_read,
4455  						  rsize);
4456  			/*
4457  			 * For windows me and 9x we do not want to request more
4458  			 * than it negotiated since it will refuse the read
4459  			 * then.
4460  			 */
4461  			if (!(tcon->ses->capabilities &
4462  				tcon->ses->server->vals->cap_large_files)) {
4463  				current_read_size = min_t(uint,
4464  					current_read_size, CIFSMaxBufSize);
4465  			}
4466  			if (open_file->invalidHandle) {
4467  				rc = cifs_reopen_file(open_file, true);
4468  				if (rc != 0)
4469  					break;
4470  			}
4471  			io_parms.pid = pid;
4472  			io_parms.tcon = tcon;
4473  			io_parms.offset = *offset;
4474  			io_parms.length = current_read_size;
4475  			io_parms.server = server;
4476  			rc = server->ops->sync_read(xid, &open_file->fid, &io_parms,
4477  						    &bytes_read, &cur_offset,
4478  						    &buf_type);
4479  		} while (rc == -EAGAIN);
4480  
4481  		if (rc || (bytes_read == 0)) {
4482  			if (total_read) {
4483  				break;
4484  			} else {
4485  				free_xid(xid);
4486  				return rc;
4487  			}
4488  		} else {
4489  			cifs_stats_bytes_read(tcon, total_read);
4490  			*offset += bytes_read;
4491  		}
4492  	}
4493  	free_xid(xid);
4494  	return total_read;
4495  }
4496  
4497  /*
4498   * If the page is mmap'ed into a process' page tables, then we need to make
4499   * sure that it doesn't change while being written back.
4500   */
cifs_page_mkwrite(struct vm_fault * vmf)4501  static vm_fault_t cifs_page_mkwrite(struct vm_fault *vmf)
4502  {
4503  	struct folio *folio = page_folio(vmf->page);
4504  
4505  	/* Wait for the folio to be written to the cache before we allow it to
4506  	 * be modified.  We then assume the entire folio will need writing back.
4507  	 */
4508  #ifdef CONFIG_CIFS_FSCACHE
4509  	if (folio_test_fscache(folio) &&
4510  	    folio_wait_fscache_killable(folio) < 0)
4511  		return VM_FAULT_RETRY;
4512  #endif
4513  
4514  	folio_wait_writeback(folio);
4515  
4516  	if (folio_lock_killable(folio) < 0)
4517  		return VM_FAULT_RETRY;
4518  	return VM_FAULT_LOCKED;
4519  }
4520  
4521  static const struct vm_operations_struct cifs_file_vm_ops = {
4522  	.fault = filemap_fault,
4523  	.map_pages = filemap_map_pages,
4524  	.page_mkwrite = cifs_page_mkwrite,
4525  };
4526  
cifs_file_strict_mmap(struct file * file,struct vm_area_struct * vma)4527  int cifs_file_strict_mmap(struct file *file, struct vm_area_struct *vma)
4528  {
4529  	int xid, rc = 0;
4530  	struct inode *inode = file_inode(file);
4531  
4532  	xid = get_xid();
4533  
4534  	if (!CIFS_CACHE_READ(CIFS_I(inode)))
4535  		rc = cifs_zap_mapping(inode);
4536  	if (!rc)
4537  		rc = generic_file_mmap(file, vma);
4538  	if (!rc)
4539  		vma->vm_ops = &cifs_file_vm_ops;
4540  
4541  	free_xid(xid);
4542  	return rc;
4543  }
4544  
cifs_file_mmap(struct file * file,struct vm_area_struct * vma)4545  int cifs_file_mmap(struct file *file, struct vm_area_struct *vma)
4546  {
4547  	int rc, xid;
4548  
4549  	xid = get_xid();
4550  
4551  	rc = cifs_revalidate_file(file);
4552  	if (rc)
4553  		cifs_dbg(FYI, "Validation prior to mmap failed, error=%d\n",
4554  			 rc);
4555  	if (!rc)
4556  		rc = generic_file_mmap(file, vma);
4557  	if (!rc)
4558  		vma->vm_ops = &cifs_file_vm_ops;
4559  
4560  	free_xid(xid);
4561  	return rc;
4562  }
4563  
4564  /*
4565   * Unlock a bunch of folios in the pagecache.
4566   */
cifs_unlock_folios(struct address_space * mapping,pgoff_t first,pgoff_t last)4567  static void cifs_unlock_folios(struct address_space *mapping, pgoff_t first, pgoff_t last)
4568  {
4569  	struct folio *folio;
4570  	XA_STATE(xas, &mapping->i_pages, first);
4571  
4572  	rcu_read_lock();
4573  	xas_for_each(&xas, folio, last) {
4574  		folio_unlock(folio);
4575  	}
4576  	rcu_read_unlock();
4577  }
4578  
cifs_readahead_complete(struct work_struct * work)4579  static void cifs_readahead_complete(struct work_struct *work)
4580  {
4581  	struct cifs_readdata *rdata = container_of(work,
4582  						   struct cifs_readdata, work);
4583  	struct folio *folio;
4584  	pgoff_t last;
4585  	bool good = rdata->result == 0 || (rdata->result == -EAGAIN && rdata->got_bytes);
4586  
4587  	XA_STATE(xas, &rdata->mapping->i_pages, rdata->offset / PAGE_SIZE);
4588  
4589  	if (good)
4590  		cifs_readahead_to_fscache(rdata->mapping->host,
4591  					  rdata->offset, rdata->bytes);
4592  
4593  	if (iov_iter_count(&rdata->iter) > 0)
4594  		iov_iter_zero(iov_iter_count(&rdata->iter), &rdata->iter);
4595  
4596  	last = (rdata->offset + rdata->bytes - 1) / PAGE_SIZE;
4597  
4598  	rcu_read_lock();
4599  	xas_for_each(&xas, folio, last) {
4600  		if (good) {
4601  			flush_dcache_folio(folio);
4602  			folio_mark_uptodate(folio);
4603  		}
4604  		folio_unlock(folio);
4605  	}
4606  	rcu_read_unlock();
4607  
4608  	kref_put(&rdata->refcount, cifs_readdata_release);
4609  }
4610  
cifs_readahead(struct readahead_control * ractl)4611  static void cifs_readahead(struct readahead_control *ractl)
4612  {
4613  	struct cifsFileInfo *open_file = ractl->file->private_data;
4614  	struct cifs_sb_info *cifs_sb = CIFS_FILE_SB(ractl->file);
4615  	struct TCP_Server_Info *server;
4616  	unsigned int xid, nr_pages, cache_nr_pages = 0;
4617  	unsigned int ra_pages;
4618  	pgoff_t next_cached = ULONG_MAX, ra_index;
4619  	bool caching = fscache_cookie_enabled(cifs_inode_cookie(ractl->mapping->host)) &&
4620  		cifs_inode_cookie(ractl->mapping->host)->cache_priv;
4621  	bool check_cache = caching;
4622  	pid_t pid;
4623  	int rc = 0;
4624  
4625  	/* Note that readahead_count() lags behind our dequeuing of pages from
4626  	 * the ractl, wo we have to keep track for ourselves.
4627  	 */
4628  	ra_pages = readahead_count(ractl);
4629  	ra_index = readahead_index(ractl);
4630  
4631  	xid = get_xid();
4632  
4633  	if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_RWPIDFORWARD)
4634  		pid = open_file->pid;
4635  	else
4636  		pid = current->tgid;
4637  
4638  	server = cifs_pick_channel(tlink_tcon(open_file->tlink)->ses);
4639  
4640  	cifs_dbg(FYI, "%s: file=%p mapping=%p num_pages=%u\n",
4641  		 __func__, ractl->file, ractl->mapping, ra_pages);
4642  
4643  	/*
4644  	 * Chop the readahead request up into rsize-sized read requests.
4645  	 */
4646  	while ((nr_pages = ra_pages)) {
4647  		unsigned int i, rsize;
4648  		struct cifs_readdata *rdata;
4649  		struct cifs_credits credits_on_stack;
4650  		struct cifs_credits *credits = &credits_on_stack;
4651  		struct folio *folio;
4652  		pgoff_t fsize;
4653  
4654  		/*
4655  		 * Find out if we have anything cached in the range of
4656  		 * interest, and if so, where the next chunk of cached data is.
4657  		 */
4658  		if (caching) {
4659  			if (check_cache) {
4660  				rc = cifs_fscache_query_occupancy(
4661  					ractl->mapping->host, ra_index, nr_pages,
4662  					&next_cached, &cache_nr_pages);
4663  				if (rc < 0)
4664  					caching = false;
4665  				check_cache = false;
4666  			}
4667  
4668  			if (ra_index == next_cached) {
4669  				/*
4670  				 * TODO: Send a whole batch of pages to be read
4671  				 * by the cache.
4672  				 */
4673  				folio = readahead_folio(ractl);
4674  				fsize = folio_nr_pages(folio);
4675  				ra_pages -= fsize;
4676  				ra_index += fsize;
4677  				if (cifs_readpage_from_fscache(ractl->mapping->host,
4678  							       &folio->page) < 0) {
4679  					/*
4680  					 * TODO: Deal with cache read failure
4681  					 * here, but for the moment, delegate
4682  					 * that to readpage.
4683  					 */
4684  					caching = false;
4685  				}
4686  				folio_unlock(folio);
4687  				next_cached += fsize;
4688  				cache_nr_pages -= fsize;
4689  				if (cache_nr_pages == 0)
4690  					check_cache = true;
4691  				continue;
4692  			}
4693  		}
4694  
4695  		if (open_file->invalidHandle) {
4696  			rc = cifs_reopen_file(open_file, true);
4697  			if (rc) {
4698  				if (rc == -EAGAIN)
4699  					continue;
4700  				break;
4701  			}
4702  		}
4703  
4704  		if (cifs_sb->ctx->rsize == 0)
4705  			cifs_sb->ctx->rsize =
4706  				server->ops->negotiate_rsize(tlink_tcon(open_file->tlink),
4707  							     cifs_sb->ctx);
4708  
4709  		rc = server->ops->wait_mtu_credits(server, cifs_sb->ctx->rsize,
4710  						   &rsize, credits);
4711  		if (rc)
4712  			break;
4713  		nr_pages = min_t(size_t, rsize / PAGE_SIZE, ra_pages);
4714  		if (next_cached != ULONG_MAX)
4715  			nr_pages = min_t(size_t, nr_pages, next_cached - ra_index);
4716  
4717  		/*
4718  		 * Give up immediately if rsize is too small to read an entire
4719  		 * page. The VFS will fall back to readpage. We should never
4720  		 * reach this point however since we set ra_pages to 0 when the
4721  		 * rsize is smaller than a cache page.
4722  		 */
4723  		if (unlikely(!nr_pages)) {
4724  			add_credits_and_wake_if(server, credits, 0);
4725  			break;
4726  		}
4727  
4728  		rdata = cifs_readdata_alloc(cifs_readahead_complete);
4729  		if (!rdata) {
4730  			/* best to give up if we're out of mem */
4731  			add_credits_and_wake_if(server, credits, 0);
4732  			break;
4733  		}
4734  
4735  		rdata->offset	= ra_index * PAGE_SIZE;
4736  		rdata->bytes	= nr_pages * PAGE_SIZE;
4737  		rdata->cfile	= cifsFileInfo_get(open_file);
4738  		rdata->server	= server;
4739  		rdata->mapping	= ractl->mapping;
4740  		rdata->pid	= pid;
4741  		rdata->credits	= credits_on_stack;
4742  
4743  		for (i = 0; i < nr_pages; i++) {
4744  			if (!readahead_folio(ractl))
4745  				WARN_ON(1);
4746  		}
4747  		ra_pages -= nr_pages;
4748  		ra_index += nr_pages;
4749  
4750  		iov_iter_xarray(&rdata->iter, ITER_DEST, &rdata->mapping->i_pages,
4751  				rdata->offset, rdata->bytes);
4752  
4753  		rc = adjust_credits(server, &rdata->credits, rdata->bytes);
4754  		if (!rc) {
4755  			if (rdata->cfile->invalidHandle)
4756  				rc = -EAGAIN;
4757  			else
4758  				rc = server->ops->async_readv(rdata);
4759  		}
4760  
4761  		if (rc) {
4762  			add_credits_and_wake_if(server, &rdata->credits, 0);
4763  			cifs_unlock_folios(rdata->mapping,
4764  					   rdata->offset / PAGE_SIZE,
4765  					   (rdata->offset + rdata->bytes - 1) / PAGE_SIZE);
4766  			/* Fallback to the readpage in error/reconnect cases */
4767  			kref_put(&rdata->refcount, cifs_readdata_release);
4768  			break;
4769  		}
4770  
4771  		kref_put(&rdata->refcount, cifs_readdata_release);
4772  	}
4773  
4774  	free_xid(xid);
4775  }
4776  
4777  /*
4778   * cifs_readpage_worker must be called with the page pinned
4779   */
cifs_readpage_worker(struct file * file,struct page * page,loff_t * poffset)4780  static int cifs_readpage_worker(struct file *file, struct page *page,
4781  	loff_t *poffset)
4782  {
4783  	struct inode *inode = file_inode(file);
4784  	struct timespec64 atime, mtime;
4785  	char *read_data;
4786  	int rc;
4787  
4788  	/* Is the page cached? */
4789  	rc = cifs_readpage_from_fscache(inode, page);
4790  	if (rc == 0)
4791  		goto read_complete;
4792  
4793  	read_data = kmap(page);
4794  	/* for reads over a certain size could initiate async read ahead */
4795  
4796  	rc = cifs_read(file, read_data, PAGE_SIZE, poffset);
4797  
4798  	if (rc < 0)
4799  		goto io_error;
4800  	else
4801  		cifs_dbg(FYI, "Bytes read %d\n", rc);
4802  
4803  	/* we do not want atime to be less than mtime, it broke some apps */
4804  	atime = inode_set_atime_to_ts(inode, current_time(inode));
4805  	mtime = inode_get_mtime(inode);
4806  	if (timespec64_compare(&atime, &mtime) < 0)
4807  		inode_set_atime_to_ts(inode, inode_get_mtime(inode));
4808  
4809  	if (PAGE_SIZE > rc)
4810  		memset(read_data + rc, 0, PAGE_SIZE - rc);
4811  
4812  	flush_dcache_page(page);
4813  	SetPageUptodate(page);
4814  	rc = 0;
4815  
4816  io_error:
4817  	kunmap(page);
4818  
4819  read_complete:
4820  	unlock_page(page);
4821  	return rc;
4822  }
4823  
cifs_read_folio(struct file * file,struct folio * folio)4824  static int cifs_read_folio(struct file *file, struct folio *folio)
4825  {
4826  	struct page *page = &folio->page;
4827  	loff_t offset = page_file_offset(page);
4828  	int rc = -EACCES;
4829  	unsigned int xid;
4830  
4831  	xid = get_xid();
4832  
4833  	if (file->private_data == NULL) {
4834  		rc = -EBADF;
4835  		free_xid(xid);
4836  		return rc;
4837  	}
4838  
4839  	cifs_dbg(FYI, "read_folio %p at offset %d 0x%x\n",
4840  		 page, (int)offset, (int)offset);
4841  
4842  	rc = cifs_readpage_worker(file, page, &offset);
4843  
4844  	free_xid(xid);
4845  	return rc;
4846  }
4847  
is_inode_writable(struct cifsInodeInfo * cifs_inode)4848  static int is_inode_writable(struct cifsInodeInfo *cifs_inode)
4849  {
4850  	struct cifsFileInfo *open_file;
4851  
4852  	spin_lock(&cifs_inode->open_file_lock);
4853  	list_for_each_entry(open_file, &cifs_inode->openFileList, flist) {
4854  		if (OPEN_FMODE(open_file->f_flags) & FMODE_WRITE) {
4855  			spin_unlock(&cifs_inode->open_file_lock);
4856  			return 1;
4857  		}
4858  	}
4859  	spin_unlock(&cifs_inode->open_file_lock);
4860  	return 0;
4861  }
4862  
4863  /* We do not want to update the file size from server for inodes
4864     open for write - to avoid races with writepage extending
4865     the file - in the future we could consider allowing
4866     refreshing the inode only on increases in the file size
4867     but this is tricky to do without racing with writebehind
4868     page caching in the current Linux kernel design */
is_size_safe_to_change(struct cifsInodeInfo * cifsInode,__u64 end_of_file,bool from_readdir)4869  bool is_size_safe_to_change(struct cifsInodeInfo *cifsInode, __u64 end_of_file,
4870  			    bool from_readdir)
4871  {
4872  	if (!cifsInode)
4873  		return true;
4874  
4875  	if (is_inode_writable(cifsInode) ||
4876  		((cifsInode->oplock & CIFS_CACHE_RW_FLG) != 0 && from_readdir)) {
4877  		/* This inode is open for write at least once */
4878  		struct cifs_sb_info *cifs_sb;
4879  
4880  		cifs_sb = CIFS_SB(cifsInode->netfs.inode.i_sb);
4881  		if (cifs_sb->mnt_cifs_flags & CIFS_MOUNT_DIRECT_IO) {
4882  			/* since no page cache to corrupt on directio
4883  			we can change size safely */
4884  			return true;
4885  		}
4886  
4887  		if (i_size_read(&cifsInode->netfs.inode) < end_of_file)
4888  			return true;
4889  
4890  		return false;
4891  	} else
4892  		return true;
4893  }
4894  
cifs_write_begin(struct file * file,struct address_space * mapping,loff_t pos,unsigned len,struct page ** pagep,void ** fsdata)4895  static int cifs_write_begin(struct file *file, struct address_space *mapping,
4896  			loff_t pos, unsigned len,
4897  			struct page **pagep, void **fsdata)
4898  {
4899  	int oncethru = 0;
4900  	pgoff_t index = pos >> PAGE_SHIFT;
4901  	loff_t offset = pos & (PAGE_SIZE - 1);
4902  	loff_t page_start = pos & PAGE_MASK;
4903  	loff_t i_size;
4904  	struct page *page;
4905  	int rc = 0;
4906  
4907  	cifs_dbg(FYI, "write_begin from %lld len %d\n", (long long)pos, len);
4908  
4909  start:
4910  	page = grab_cache_page_write_begin(mapping, index);
4911  	if (!page) {
4912  		rc = -ENOMEM;
4913  		goto out;
4914  	}
4915  
4916  	if (PageUptodate(page))
4917  		goto out;
4918  
4919  	/*
4920  	 * If we write a full page it will be up to date, no need to read from
4921  	 * the server. If the write is short, we'll end up doing a sync write
4922  	 * instead.
4923  	 */
4924  	if (len == PAGE_SIZE)
4925  		goto out;
4926  
4927  	/*
4928  	 * optimize away the read when we have an oplock, and we're not
4929  	 * expecting to use any of the data we'd be reading in. That
4930  	 * is, when the page lies beyond the EOF, or straddles the EOF
4931  	 * and the write will cover all of the existing data.
4932  	 */
4933  	if (CIFS_CACHE_READ(CIFS_I(mapping->host))) {
4934  		i_size = i_size_read(mapping->host);
4935  		if (page_start >= i_size ||
4936  		    (offset == 0 && (pos + len) >= i_size)) {
4937  			zero_user_segments(page, 0, offset,
4938  					   offset + len,
4939  					   PAGE_SIZE);
4940  			/*
4941  			 * PageChecked means that the parts of the page
4942  			 * to which we're not writing are considered up
4943  			 * to date. Once the data is copied to the
4944  			 * page, it can be set uptodate.
4945  			 */
4946  			SetPageChecked(page);
4947  			goto out;
4948  		}
4949  	}
4950  
4951  	if ((file->f_flags & O_ACCMODE) != O_WRONLY && !oncethru) {
4952  		/*
4953  		 * might as well read a page, it is fast enough. If we get
4954  		 * an error, we don't need to return it. cifs_write_end will
4955  		 * do a sync write instead since PG_uptodate isn't set.
4956  		 */
4957  		cifs_readpage_worker(file, page, &page_start);
4958  		put_page(page);
4959  		oncethru = 1;
4960  		goto start;
4961  	} else {
4962  		/* we could try using another file handle if there is one -
4963  		   but how would we lock it to prevent close of that handle
4964  		   racing with this read? In any case
4965  		   this will be written out by write_end so is fine */
4966  	}
4967  out:
4968  	*pagep = page;
4969  	return rc;
4970  }
4971  
cifs_release_folio(struct folio * folio,gfp_t gfp)4972  static bool cifs_release_folio(struct folio *folio, gfp_t gfp)
4973  {
4974  	if (folio_test_private(folio))
4975  		return 0;
4976  	if (folio_test_fscache(folio)) {
4977  		if (current_is_kswapd() || !(gfp & __GFP_FS))
4978  			return false;
4979  		folio_wait_fscache(folio);
4980  	}
4981  	fscache_note_page_release(cifs_inode_cookie(folio->mapping->host));
4982  	return true;
4983  }
4984  
cifs_invalidate_folio(struct folio * folio,size_t offset,size_t length)4985  static void cifs_invalidate_folio(struct folio *folio, size_t offset,
4986  				 size_t length)
4987  {
4988  	folio_wait_fscache(folio);
4989  }
4990  
cifs_launder_folio(struct folio * folio)4991  static int cifs_launder_folio(struct folio *folio)
4992  {
4993  	int rc = 0;
4994  	loff_t range_start = folio_pos(folio);
4995  	loff_t range_end = range_start + folio_size(folio);
4996  	struct writeback_control wbc = {
4997  		.sync_mode = WB_SYNC_ALL,
4998  		.nr_to_write = 0,
4999  		.range_start = range_start,
5000  		.range_end = range_end,
5001  	};
5002  
5003  	cifs_dbg(FYI, "Launder page: %lu\n", folio->index);
5004  
5005  	if (folio_clear_dirty_for_io(folio))
5006  		rc = cifs_writepage_locked(&folio->page, &wbc);
5007  
5008  	folio_wait_fscache(folio);
5009  	return rc;
5010  }
5011  
cifs_oplock_break(struct work_struct * work)5012  void cifs_oplock_break(struct work_struct *work)
5013  {
5014  	struct cifsFileInfo *cfile = container_of(work, struct cifsFileInfo,
5015  						  oplock_break);
5016  	struct inode *inode = d_inode(cfile->dentry);
5017  	struct cifs_sb_info *cifs_sb = CIFS_SB(inode->i_sb);
5018  	struct cifsInodeInfo *cinode = CIFS_I(inode);
5019  	struct cifs_tcon *tcon;
5020  	struct TCP_Server_Info *server;
5021  	struct tcon_link *tlink;
5022  	int rc = 0;
5023  	bool purge_cache = false, oplock_break_cancelled;
5024  	__u64 persistent_fid, volatile_fid;
5025  	__u16 net_fid;
5026  
5027  	wait_on_bit(&cinode->flags, CIFS_INODE_PENDING_WRITERS,
5028  			TASK_UNINTERRUPTIBLE);
5029  
5030  	tlink = cifs_sb_tlink(cifs_sb);
5031  	if (IS_ERR(tlink))
5032  		goto out;
5033  	tcon = tlink_tcon(tlink);
5034  	server = tcon->ses->server;
5035  
5036  	server->ops->downgrade_oplock(server, cinode, cfile->oplock_level,
5037  				      cfile->oplock_epoch, &purge_cache);
5038  
5039  	if (!CIFS_CACHE_WRITE(cinode) && CIFS_CACHE_READ(cinode) &&
5040  						cifs_has_mand_locks(cinode)) {
5041  		cifs_dbg(FYI, "Reset oplock to None for inode=%p due to mand locks\n",
5042  			 inode);
5043  		cinode->oplock = 0;
5044  	}
5045  
5046  	if (inode && S_ISREG(inode->i_mode)) {
5047  		if (CIFS_CACHE_READ(cinode))
5048  			break_lease(inode, O_RDONLY);
5049  		else
5050  			break_lease(inode, O_WRONLY);
5051  		rc = filemap_fdatawrite(inode->i_mapping);
5052  		if (!CIFS_CACHE_READ(cinode) || purge_cache) {
5053  			rc = filemap_fdatawait(inode->i_mapping);
5054  			mapping_set_error(inode->i_mapping, rc);
5055  			cifs_zap_mapping(inode);
5056  		}
5057  		cifs_dbg(FYI, "Oplock flush inode %p rc %d\n", inode, rc);
5058  		if (CIFS_CACHE_WRITE(cinode))
5059  			goto oplock_break_ack;
5060  	}
5061  
5062  	rc = cifs_push_locks(cfile);
5063  	if (rc)
5064  		cifs_dbg(VFS, "Push locks rc = %d\n", rc);
5065  
5066  oplock_break_ack:
5067  	/*
5068  	 * When oplock break is received and there are no active
5069  	 * file handles but cached, then schedule deferred close immediately.
5070  	 * So, new open will not use cached handle.
5071  	 */
5072  
5073  	if (!CIFS_CACHE_HANDLE(cinode) && !list_empty(&cinode->deferred_closes))
5074  		cifs_close_deferred_file(cinode);
5075  
5076  	persistent_fid = cfile->fid.persistent_fid;
5077  	volatile_fid = cfile->fid.volatile_fid;
5078  	net_fid = cfile->fid.netfid;
5079  	oplock_break_cancelled = cfile->oplock_break_cancelled;
5080  
5081  	_cifsFileInfo_put(cfile, false /* do not wait for ourself */, false);
5082  	/*
5083  	 * MS-SMB2 3.2.5.19.1 and 3.2.5.19.2 (and MS-CIFS 3.2.5.42) do not require
5084  	 * an acknowledgment to be sent when the file has already been closed.
5085  	 */
5086  	spin_lock(&cinode->open_file_lock);
5087  	/* check list empty since can race with kill_sb calling tree disconnect */
5088  	if (!oplock_break_cancelled && !list_empty(&cinode->openFileList)) {
5089  		spin_unlock(&cinode->open_file_lock);
5090  		rc = server->ops->oplock_response(tcon, persistent_fid,
5091  						  volatile_fid, net_fid, cinode);
5092  		cifs_dbg(FYI, "Oplock release rc = %d\n", rc);
5093  	} else
5094  		spin_unlock(&cinode->open_file_lock);
5095  
5096  	cifs_put_tlink(tlink);
5097  out:
5098  	cifs_done_oplock_break(cinode);
5099  }
5100  
5101  /*
5102   * The presence of cifs_direct_io() in the address space ops vector
5103   * allowes open() O_DIRECT flags which would have failed otherwise.
5104   *
5105   * In the non-cached mode (mount with cache=none), we shunt off direct read and write requests
5106   * so this method should never be called.
5107   *
5108   * Direct IO is not yet supported in the cached mode.
5109   */
5110  static ssize_t
cifs_direct_io(struct kiocb * iocb,struct iov_iter * iter)5111  cifs_direct_io(struct kiocb *iocb, struct iov_iter *iter)
5112  {
5113          /*
5114           * FIXME
5115           * Eventually need to support direct IO for non forcedirectio mounts
5116           */
5117          return -EINVAL;
5118  }
5119  
cifs_swap_activate(struct swap_info_struct * sis,struct file * swap_file,sector_t * span)5120  static int cifs_swap_activate(struct swap_info_struct *sis,
5121  			      struct file *swap_file, sector_t *span)
5122  {
5123  	struct cifsFileInfo *cfile = swap_file->private_data;
5124  	struct inode *inode = swap_file->f_mapping->host;
5125  	unsigned long blocks;
5126  	long long isize;
5127  
5128  	cifs_dbg(FYI, "swap activate\n");
5129  
5130  	if (!swap_file->f_mapping->a_ops->swap_rw)
5131  		/* Cannot support swap */
5132  		return -EINVAL;
5133  
5134  	spin_lock(&inode->i_lock);
5135  	blocks = inode->i_blocks;
5136  	isize = inode->i_size;
5137  	spin_unlock(&inode->i_lock);
5138  	if (blocks*512 < isize) {
5139  		pr_warn("swap activate: swapfile has holes\n");
5140  		return -EINVAL;
5141  	}
5142  	*span = sis->pages;
5143  
5144  	pr_warn_once("Swap support over SMB3 is experimental\n");
5145  
5146  	/*
5147  	 * TODO: consider adding ACL (or documenting how) to prevent other
5148  	 * users (on this or other systems) from reading it
5149  	 */
5150  
5151  
5152  	/* TODO: add sk_set_memalloc(inet) or similar */
5153  
5154  	if (cfile)
5155  		cfile->swapfile = true;
5156  	/*
5157  	 * TODO: Since file already open, we can't open with DENY_ALL here
5158  	 * but we could add call to grab a byte range lock to prevent others
5159  	 * from reading or writing the file
5160  	 */
5161  
5162  	sis->flags |= SWP_FS_OPS;
5163  	return add_swap_extent(sis, 0, sis->max, 0);
5164  }
5165  
cifs_swap_deactivate(struct file * file)5166  static void cifs_swap_deactivate(struct file *file)
5167  {
5168  	struct cifsFileInfo *cfile = file->private_data;
5169  
5170  	cifs_dbg(FYI, "swap deactivate\n");
5171  
5172  	/* TODO: undo sk_set_memalloc(inet) will eventually be needed */
5173  
5174  	if (cfile)
5175  		cfile->swapfile = false;
5176  
5177  	/* do we need to unpin (or unlock) the file */
5178  }
5179  
5180  /*
5181   * Mark a page as having been made dirty and thus needing writeback.  We also
5182   * need to pin the cache object to write back to.
5183   */
5184  #ifdef CONFIG_CIFS_FSCACHE
cifs_dirty_folio(struct address_space * mapping,struct folio * folio)5185  static bool cifs_dirty_folio(struct address_space *mapping, struct folio *folio)
5186  {
5187  	return fscache_dirty_folio(mapping, folio,
5188  					cifs_inode_cookie(mapping->host));
5189  }
5190  #else
5191  #define cifs_dirty_folio filemap_dirty_folio
5192  #endif
5193  
5194  const struct address_space_operations cifs_addr_ops = {
5195  	.read_folio = cifs_read_folio,
5196  	.readahead = cifs_readahead,
5197  	.writepages = cifs_writepages,
5198  	.write_begin = cifs_write_begin,
5199  	.write_end = cifs_write_end,
5200  	.dirty_folio = cifs_dirty_folio,
5201  	.release_folio = cifs_release_folio,
5202  	.direct_IO = cifs_direct_io,
5203  	.invalidate_folio = cifs_invalidate_folio,
5204  	.launder_folio = cifs_launder_folio,
5205  	.migrate_folio = filemap_migrate_folio,
5206  	/*
5207  	 * TODO: investigate and if useful we could add an is_dirty_writeback
5208  	 * helper if needed
5209  	 */
5210  	.swap_activate = cifs_swap_activate,
5211  	.swap_deactivate = cifs_swap_deactivate,
5212  };
5213  
5214  /*
5215   * cifs_readahead requires the server to support a buffer large enough to
5216   * contain the header plus one complete page of data.  Otherwise, we need
5217   * to leave cifs_readahead out of the address space operations.
5218   */
5219  const struct address_space_operations cifs_addr_ops_smallbuf = {
5220  	.read_folio = cifs_read_folio,
5221  	.writepages = cifs_writepages,
5222  	.write_begin = cifs_write_begin,
5223  	.write_end = cifs_write_end,
5224  	.dirty_folio = cifs_dirty_folio,
5225  	.release_folio = cifs_release_folio,
5226  	.invalidate_folio = cifs_invalidate_folio,
5227  	.launder_folio = cifs_launder_folio,
5228  	.migrate_folio = filemap_migrate_folio,
5229  };
5230