xref: /openbmc/linux/fs/9p/vfs_file.c (revision 5a170e9e)
1 /*
2  *  linux/fs/9p/vfs_file.c
3  *
4  * This file contians vfs file ops for 9P2000.
5  *
6  *  Copyright (C) 2004 by Eric Van Hensbergen <ericvh@gmail.com>
7  *  Copyright (C) 2002 by Ron Minnich <rminnich@lanl.gov>
8  *
9  *  This program is free software; you can redistribute it and/or modify
10  *  it under the terms of the GNU General Public License version 2
11  *  as published by the Free Software Foundation.
12  *
13  *  This program is distributed in the hope that it will be useful,
14  *  but WITHOUT ANY WARRANTY; without even the implied warranty of
15  *  MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
16  *  GNU General Public License for more details.
17  *
18  *  You should have received a copy of the GNU General Public License
19  *  along with this program; if not, write to:
20  *  Free Software Foundation
21  *  51 Franklin Street, Fifth Floor
22  *  Boston, MA  02111-1301  USA
23  *
24  */
25 
26 #include <linux/module.h>
27 #include <linux/errno.h>
28 #include <linux/fs.h>
29 #include <linux/sched.h>
30 #include <linux/file.h>
31 #include <linux/stat.h>
32 #include <linux/string.h>
33 #include <linux/inet.h>
34 #include <linux/list.h>
35 #include <linux/pagemap.h>
36 #include <linux/utsname.h>
37 #include <linux/uaccess.h>
38 #include <linux/idr.h>
39 #include <linux/uio.h>
40 #include <linux/slab.h>
41 #include <net/9p/9p.h>
42 #include <net/9p/client.h>
43 
44 #include "v9fs.h"
45 #include "v9fs_vfs.h"
46 #include "fid.h"
47 #include "cache.h"
48 
49 static const struct vm_operations_struct v9fs_file_vm_ops;
50 static const struct vm_operations_struct v9fs_mmap_file_vm_ops;
51 
52 /**
53  * v9fs_file_open - open a file (or directory)
54  * @inode: inode to be opened
55  * @file: file being opened
56  *
57  */
58 
59 int v9fs_file_open(struct inode *inode, struct file *file)
60 {
61 	int err;
62 	struct v9fs_inode *v9inode;
63 	struct v9fs_session_info *v9ses;
64 	struct p9_fid *fid;
65 	int omode;
66 
67 	p9_debug(P9_DEBUG_VFS, "inode: %p file: %p\n", inode, file);
68 	v9inode = V9FS_I(inode);
69 	v9ses = v9fs_inode2v9ses(inode);
70 	if (v9fs_proto_dotl(v9ses))
71 		omode = v9fs_open_to_dotl_flags(file->f_flags);
72 	else
73 		omode = v9fs_uflags2omode(file->f_flags,
74 					v9fs_proto_dotu(v9ses));
75 	fid = file->private_data;
76 	if (!fid) {
77 		fid = v9fs_fid_clone(file_dentry(file));
78 		if (IS_ERR(fid))
79 			return PTR_ERR(fid);
80 
81 		err = p9_client_open(fid, omode);
82 		if (err < 0) {
83 			p9_client_clunk(fid);
84 			return err;
85 		}
86 		if ((file->f_flags & O_APPEND) &&
87 			(!v9fs_proto_dotu(v9ses) && !v9fs_proto_dotl(v9ses)))
88 			generic_file_llseek(file, 0, SEEK_END);
89 	}
90 
91 	file->private_data = fid;
92 	mutex_lock(&v9inode->v_mutex);
93 	if ((v9ses->cache == CACHE_LOOSE || v9ses->cache == CACHE_FSCACHE) &&
94 	    !v9inode->writeback_fid &&
95 	    ((file->f_flags & O_ACCMODE) != O_RDONLY)) {
96 		/*
97 		 * clone a fid and add it to writeback_fid
98 		 * we do it during open time instead of
99 		 * page dirty time via write_begin/page_mkwrite
100 		 * because we want write after unlink usecase
101 		 * to work.
102 		 */
103 		fid = v9fs_writeback_fid(file_dentry(file));
104 		if (IS_ERR(fid)) {
105 			err = PTR_ERR(fid);
106 			mutex_unlock(&v9inode->v_mutex);
107 			goto out_error;
108 		}
109 		v9inode->writeback_fid = (void *) fid;
110 	}
111 	mutex_unlock(&v9inode->v_mutex);
112 	if (v9ses->cache == CACHE_LOOSE || v9ses->cache == CACHE_FSCACHE)
113 		v9fs_cache_inode_set_cookie(inode, file);
114 	return 0;
115 out_error:
116 	p9_client_clunk(file->private_data);
117 	file->private_data = NULL;
118 	return err;
119 }
120 
121 /**
122  * v9fs_file_lock - lock a file (or directory)
123  * @filp: file to be locked
124  * @cmd: lock command
125  * @fl: file lock structure
126  *
127  * Bugs: this looks like a local only lock, we should extend into 9P
128  *       by using open exclusive
129  */
130 
131 static int v9fs_file_lock(struct file *filp, int cmd, struct file_lock *fl)
132 {
133 	int res = 0;
134 	struct inode *inode = file_inode(filp);
135 
136 	p9_debug(P9_DEBUG_VFS, "filp: %p lock: %p\n", filp, fl);
137 
138 	/* No mandatory locks */
139 	if (__mandatory_lock(inode) && fl->fl_type != F_UNLCK)
140 		return -ENOLCK;
141 
142 	if ((IS_SETLK(cmd) || IS_SETLKW(cmd)) && fl->fl_type != F_UNLCK) {
143 		filemap_write_and_wait(inode->i_mapping);
144 		invalidate_mapping_pages(&inode->i_data, 0, -1);
145 	}
146 
147 	return res;
148 }
149 
150 static int v9fs_file_do_lock(struct file *filp, int cmd, struct file_lock *fl)
151 {
152 	struct p9_flock flock;
153 	struct p9_fid *fid;
154 	uint8_t status = P9_LOCK_ERROR;
155 	int res = 0;
156 	unsigned char fl_type;
157 	struct v9fs_session_info *v9ses;
158 
159 	fid = filp->private_data;
160 	BUG_ON(fid == NULL);
161 
162 	if ((fl->fl_flags & FL_POSIX) != FL_POSIX)
163 		BUG();
164 
165 	res = locks_lock_file_wait(filp, fl);
166 	if (res < 0)
167 		goto out;
168 
169 	/* convert posix lock to p9 tlock args */
170 	memset(&flock, 0, sizeof(flock));
171 	/* map the lock type */
172 	switch (fl->fl_type) {
173 	case F_RDLCK:
174 		flock.type = P9_LOCK_TYPE_RDLCK;
175 		break;
176 	case F_WRLCK:
177 		flock.type = P9_LOCK_TYPE_WRLCK;
178 		break;
179 	case F_UNLCK:
180 		flock.type = P9_LOCK_TYPE_UNLCK;
181 		break;
182 	}
183 	flock.start = fl->fl_start;
184 	if (fl->fl_end == OFFSET_MAX)
185 		flock.length = 0;
186 	else
187 		flock.length = fl->fl_end - fl->fl_start + 1;
188 	flock.proc_id = fl->fl_pid;
189 	flock.client_id = fid->clnt->name;
190 	if (IS_SETLKW(cmd))
191 		flock.flags = P9_LOCK_FLAGS_BLOCK;
192 
193 	v9ses = v9fs_inode2v9ses(file_inode(filp));
194 
195 	/*
196 	 * if its a blocked request and we get P9_LOCK_BLOCKED as the status
197 	 * for lock request, keep on trying
198 	 */
199 	for (;;) {
200 		res = p9_client_lock_dotl(fid, &flock, &status);
201 		if (res < 0)
202 			goto out_unlock;
203 
204 		if (status != P9_LOCK_BLOCKED)
205 			break;
206 		if (status == P9_LOCK_BLOCKED && !IS_SETLKW(cmd))
207 			break;
208 		if (schedule_timeout_interruptible(v9ses->session_lock_timeout)
209 				!= 0)
210 			break;
211 		/*
212 		 * p9_client_lock_dotl overwrites flock.client_id with the
213 		 * server message, free and reuse the client name
214 		 */
215 		if (flock.client_id != fid->clnt->name) {
216 			kfree(flock.client_id);
217 			flock.client_id = fid->clnt->name;
218 		}
219 	}
220 
221 	/* map 9p status to VFS status */
222 	switch (status) {
223 	case P9_LOCK_SUCCESS:
224 		res = 0;
225 		break;
226 	case P9_LOCK_BLOCKED:
227 		res = -EAGAIN;
228 		break;
229 	default:
230 		WARN_ONCE(1, "unknown lock status code: %d\n", status);
231 		/* fall through */
232 	case P9_LOCK_ERROR:
233 	case P9_LOCK_GRACE:
234 		res = -ENOLCK;
235 		break;
236 	}
237 
238 out_unlock:
239 	/*
240 	 * incase server returned error for lock request, revert
241 	 * it locally
242 	 */
243 	if (res < 0 && fl->fl_type != F_UNLCK) {
244 		fl_type = fl->fl_type;
245 		fl->fl_type = F_UNLCK;
246 		/* Even if this fails we want to return the remote error */
247 		locks_lock_file_wait(filp, fl);
248 		fl->fl_type = fl_type;
249 	}
250 	if (flock.client_id != fid->clnt->name)
251 		kfree(flock.client_id);
252 out:
253 	return res;
254 }
255 
256 static int v9fs_file_getlock(struct file *filp, struct file_lock *fl)
257 {
258 	struct p9_getlock glock;
259 	struct p9_fid *fid;
260 	int res = 0;
261 
262 	fid = filp->private_data;
263 	BUG_ON(fid == NULL);
264 
265 	posix_test_lock(filp, fl);
266 	/*
267 	 * if we have a conflicting lock locally, no need to validate
268 	 * with server
269 	 */
270 	if (fl->fl_type != F_UNLCK)
271 		return res;
272 
273 	/* convert posix lock to p9 tgetlock args */
274 	memset(&glock, 0, sizeof(glock));
275 	glock.type  = P9_LOCK_TYPE_UNLCK;
276 	glock.start = fl->fl_start;
277 	if (fl->fl_end == OFFSET_MAX)
278 		glock.length = 0;
279 	else
280 		glock.length = fl->fl_end - fl->fl_start + 1;
281 	glock.proc_id = fl->fl_pid;
282 	glock.client_id = fid->clnt->name;
283 
284 	res = p9_client_getlock_dotl(fid, &glock);
285 	if (res < 0)
286 		goto out;
287 	/* map 9p lock type to os lock type */
288 	switch (glock.type) {
289 	case P9_LOCK_TYPE_RDLCK:
290 		fl->fl_type = F_RDLCK;
291 		break;
292 	case P9_LOCK_TYPE_WRLCK:
293 		fl->fl_type = F_WRLCK;
294 		break;
295 	case P9_LOCK_TYPE_UNLCK:
296 		fl->fl_type = F_UNLCK;
297 		break;
298 	}
299 	if (glock.type != P9_LOCK_TYPE_UNLCK) {
300 		fl->fl_start = glock.start;
301 		if (glock.length == 0)
302 			fl->fl_end = OFFSET_MAX;
303 		else
304 			fl->fl_end = glock.start + glock.length - 1;
305 		fl->fl_pid = -glock.proc_id;
306 	}
307 out:
308 	if (glock.client_id != fid->clnt->name)
309 		kfree(glock.client_id);
310 	return res;
311 }
312 
313 /**
314  * v9fs_file_lock_dotl - lock a file (or directory)
315  * @filp: file to be locked
316  * @cmd: lock command
317  * @fl: file lock structure
318  *
319  */
320 
321 static int v9fs_file_lock_dotl(struct file *filp, int cmd, struct file_lock *fl)
322 {
323 	struct inode *inode = file_inode(filp);
324 	int ret = -ENOLCK;
325 
326 	p9_debug(P9_DEBUG_VFS, "filp: %p cmd:%d lock: %p name: %pD\n",
327 		 filp, cmd, fl, filp);
328 
329 	/* No mandatory locks */
330 	if (__mandatory_lock(inode) && fl->fl_type != F_UNLCK)
331 		goto out_err;
332 
333 	if ((IS_SETLK(cmd) || IS_SETLKW(cmd)) && fl->fl_type != F_UNLCK) {
334 		filemap_write_and_wait(inode->i_mapping);
335 		invalidate_mapping_pages(&inode->i_data, 0, -1);
336 	}
337 
338 	if (IS_SETLK(cmd) || IS_SETLKW(cmd))
339 		ret = v9fs_file_do_lock(filp, cmd, fl);
340 	else if (IS_GETLK(cmd))
341 		ret = v9fs_file_getlock(filp, fl);
342 	else
343 		ret = -EINVAL;
344 out_err:
345 	return ret;
346 }
347 
348 /**
349  * v9fs_file_flock_dotl - lock a file
350  * @filp: file to be locked
351  * @cmd: lock command
352  * @fl: file lock structure
353  *
354  */
355 
356 static int v9fs_file_flock_dotl(struct file *filp, int cmd,
357 	struct file_lock *fl)
358 {
359 	struct inode *inode = file_inode(filp);
360 	int ret = -ENOLCK;
361 
362 	p9_debug(P9_DEBUG_VFS, "filp: %p cmd:%d lock: %p name: %pD\n",
363 		 filp, cmd, fl, filp);
364 
365 	/* No mandatory locks */
366 	if (__mandatory_lock(inode) && fl->fl_type != F_UNLCK)
367 		goto out_err;
368 
369 	if (!(fl->fl_flags & FL_FLOCK))
370 		goto out_err;
371 
372 	if ((IS_SETLK(cmd) || IS_SETLKW(cmd)) && fl->fl_type != F_UNLCK) {
373 		filemap_write_and_wait(inode->i_mapping);
374 		invalidate_mapping_pages(&inode->i_data, 0, -1);
375 	}
376 	/* Convert flock to posix lock */
377 	fl->fl_flags |= FL_POSIX;
378 	fl->fl_flags ^= FL_FLOCK;
379 
380 	if (IS_SETLK(cmd) | IS_SETLKW(cmd))
381 		ret = v9fs_file_do_lock(filp, cmd, fl);
382 	else
383 		ret = -EINVAL;
384 out_err:
385 	return ret;
386 }
387 
388 /**
389  * v9fs_file_read - read from a file
390  * @filp: file pointer to read
391  * @udata: user data buffer to read data into
392  * @count: size of buffer
393  * @offset: offset at which to read data
394  *
395  */
396 
397 static ssize_t
398 v9fs_file_read_iter(struct kiocb *iocb, struct iov_iter *to)
399 {
400 	struct p9_fid *fid = iocb->ki_filp->private_data;
401 	int ret, err = 0;
402 
403 	p9_debug(P9_DEBUG_VFS, "count %zu offset %lld\n",
404 		 iov_iter_count(to), iocb->ki_pos);
405 
406 	ret = p9_client_read(fid, iocb->ki_pos, to, &err);
407 	if (!ret)
408 		return err;
409 
410 	iocb->ki_pos += ret;
411 	return ret;
412 }
413 
414 /**
415  * v9fs_file_write - write to a file
416  * @filp: file pointer to write
417  * @data: data buffer to write data from
418  * @count: size of buffer
419  * @offset: offset at which to write data
420  *
421  */
422 static ssize_t
423 v9fs_file_write_iter(struct kiocb *iocb, struct iov_iter *from)
424 {
425 	struct file *file = iocb->ki_filp;
426 	ssize_t retval;
427 	loff_t origin;
428 	int err = 0;
429 
430 	retval = generic_write_checks(iocb, from);
431 	if (retval <= 0)
432 		return retval;
433 
434 	origin = iocb->ki_pos;
435 	retval = p9_client_write(file->private_data, iocb->ki_pos, from, &err);
436 	if (retval > 0) {
437 		struct inode *inode = file_inode(file);
438 		loff_t i_size;
439 		unsigned long pg_start, pg_end;
440 		pg_start = origin >> PAGE_SHIFT;
441 		pg_end = (origin + retval - 1) >> PAGE_SHIFT;
442 		if (inode->i_mapping && inode->i_mapping->nrpages)
443 			invalidate_inode_pages2_range(inode->i_mapping,
444 						      pg_start, pg_end);
445 		iocb->ki_pos += retval;
446 		i_size = i_size_read(inode);
447 		if (iocb->ki_pos > i_size) {
448 			inode_add_bytes(inode, iocb->ki_pos - i_size);
449 			i_size_write(inode, iocb->ki_pos);
450 		}
451 		return retval;
452 	}
453 	return err;
454 }
455 
456 static int v9fs_file_fsync(struct file *filp, loff_t start, loff_t end,
457 			   int datasync)
458 {
459 	struct p9_fid *fid;
460 	struct inode *inode = filp->f_mapping->host;
461 	struct p9_wstat wstat;
462 	int retval;
463 
464 	retval = file_write_and_wait_range(filp, start, end);
465 	if (retval)
466 		return retval;
467 
468 	inode_lock(inode);
469 	p9_debug(P9_DEBUG_VFS, "filp %p datasync %x\n", filp, datasync);
470 
471 	fid = filp->private_data;
472 	v9fs_blank_wstat(&wstat);
473 
474 	retval = p9_client_wstat(fid, &wstat);
475 	inode_unlock(inode);
476 
477 	return retval;
478 }
479 
480 int v9fs_file_fsync_dotl(struct file *filp, loff_t start, loff_t end,
481 			 int datasync)
482 {
483 	struct p9_fid *fid;
484 	struct inode *inode = filp->f_mapping->host;
485 	int retval;
486 
487 	retval = file_write_and_wait_range(filp, start, end);
488 	if (retval)
489 		return retval;
490 
491 	inode_lock(inode);
492 	p9_debug(P9_DEBUG_VFS, "filp %p datasync %x\n", filp, datasync);
493 
494 	fid = filp->private_data;
495 
496 	retval = p9_client_fsync(fid, datasync);
497 	inode_unlock(inode);
498 
499 	return retval;
500 }
501 
502 static int
503 v9fs_file_mmap(struct file *filp, struct vm_area_struct *vma)
504 {
505 	int retval;
506 
507 
508 	retval = generic_file_mmap(filp, vma);
509 	if (!retval)
510 		vma->vm_ops = &v9fs_file_vm_ops;
511 
512 	return retval;
513 }
514 
515 static int
516 v9fs_mmap_file_mmap(struct file *filp, struct vm_area_struct *vma)
517 {
518 	int retval;
519 	struct inode *inode;
520 	struct v9fs_inode *v9inode;
521 	struct p9_fid *fid;
522 
523 	inode = file_inode(filp);
524 	v9inode = V9FS_I(inode);
525 	mutex_lock(&v9inode->v_mutex);
526 	if (!v9inode->writeback_fid &&
527 	    (vma->vm_flags & VM_WRITE)) {
528 		/*
529 		 * clone a fid and add it to writeback_fid
530 		 * we do it during mmap instead of
531 		 * page dirty time via write_begin/page_mkwrite
532 		 * because we want write after unlink usecase
533 		 * to work.
534 		 */
535 		fid = v9fs_writeback_fid(file_dentry(filp));
536 		if (IS_ERR(fid)) {
537 			retval = PTR_ERR(fid);
538 			mutex_unlock(&v9inode->v_mutex);
539 			return retval;
540 		}
541 		v9inode->writeback_fid = (void *) fid;
542 	}
543 	mutex_unlock(&v9inode->v_mutex);
544 
545 	retval = generic_file_mmap(filp, vma);
546 	if (!retval)
547 		vma->vm_ops = &v9fs_mmap_file_vm_ops;
548 
549 	return retval;
550 }
551 
552 static vm_fault_t
553 v9fs_vm_page_mkwrite(struct vm_fault *vmf)
554 {
555 	struct v9fs_inode *v9inode;
556 	struct page *page = vmf->page;
557 	struct file *filp = vmf->vma->vm_file;
558 	struct inode *inode = file_inode(filp);
559 
560 
561 	p9_debug(P9_DEBUG_VFS, "page %p fid %lx\n",
562 		 page, (unsigned long)filp->private_data);
563 
564 	/* Update file times before taking page lock */
565 	file_update_time(filp);
566 
567 	v9inode = V9FS_I(inode);
568 	/* make sure the cache has finished storing the page */
569 	v9fs_fscache_wait_on_page_write(inode, page);
570 	BUG_ON(!v9inode->writeback_fid);
571 	lock_page(page);
572 	if (page->mapping != inode->i_mapping)
573 		goto out_unlock;
574 	wait_for_stable_page(page);
575 
576 	return VM_FAULT_LOCKED;
577 out_unlock:
578 	unlock_page(page);
579 	return VM_FAULT_NOPAGE;
580 }
581 
582 /**
583  * v9fs_mmap_file_read - read from a file
584  * @filp: file pointer to read
585  * @data: user data buffer to read data into
586  * @count: size of buffer
587  * @offset: offset at which to read data
588  *
589  */
590 static ssize_t
591 v9fs_mmap_file_read_iter(struct kiocb *iocb, struct iov_iter *to)
592 {
593 	/* TODO: Check if there are dirty pages */
594 	return v9fs_file_read_iter(iocb, to);
595 }
596 
597 /**
598  * v9fs_mmap_file_write - write to a file
599  * @filp: file pointer to write
600  * @data: data buffer to write data from
601  * @count: size of buffer
602  * @offset: offset at which to write data
603  *
604  */
605 static ssize_t
606 v9fs_mmap_file_write_iter(struct kiocb *iocb, struct iov_iter *from)
607 {
608 	/*
609 	 * TODO: invalidate mmaps on filp's inode between
610 	 * offset and offset+count
611 	 */
612 	return v9fs_file_write_iter(iocb, from);
613 }
614 
615 static void v9fs_mmap_vm_close(struct vm_area_struct *vma)
616 {
617 	struct inode *inode;
618 
619 	struct writeback_control wbc = {
620 		.nr_to_write = LONG_MAX,
621 		.sync_mode = WB_SYNC_ALL,
622 		.range_start = vma->vm_pgoff * PAGE_SIZE,
623 		 /* absolute end, byte at end included */
624 		.range_end = vma->vm_pgoff * PAGE_SIZE +
625 			(vma->vm_end - vma->vm_start - 1),
626 	};
627 
628 
629 	p9_debug(P9_DEBUG_VFS, "9p VMA close, %p, flushing", vma);
630 
631 	inode = file_inode(vma->vm_file);
632 
633 	if (!mapping_cap_writeback_dirty(inode->i_mapping))
634 		wbc.nr_to_write = 0;
635 
636 	might_sleep();
637 	sync_inode(inode, &wbc);
638 }
639 
640 
641 static const struct vm_operations_struct v9fs_file_vm_ops = {
642 	.fault = filemap_fault,
643 	.map_pages = filemap_map_pages,
644 	.page_mkwrite = v9fs_vm_page_mkwrite,
645 };
646 
647 static const struct vm_operations_struct v9fs_mmap_file_vm_ops = {
648 	.close = v9fs_mmap_vm_close,
649 	.fault = filemap_fault,
650 	.map_pages = filemap_map_pages,
651 	.page_mkwrite = v9fs_vm_page_mkwrite,
652 };
653 
654 
655 const struct file_operations v9fs_cached_file_operations = {
656 	.llseek = generic_file_llseek,
657 	.read_iter = generic_file_read_iter,
658 	.write_iter = generic_file_write_iter,
659 	.open = v9fs_file_open,
660 	.release = v9fs_dir_release,
661 	.lock = v9fs_file_lock,
662 	.mmap = v9fs_file_mmap,
663 	.fsync = v9fs_file_fsync,
664 };
665 
666 const struct file_operations v9fs_cached_file_operations_dotl = {
667 	.llseek = generic_file_llseek,
668 	.read_iter = generic_file_read_iter,
669 	.write_iter = generic_file_write_iter,
670 	.open = v9fs_file_open,
671 	.release = v9fs_dir_release,
672 	.lock = v9fs_file_lock_dotl,
673 	.flock = v9fs_file_flock_dotl,
674 	.mmap = v9fs_file_mmap,
675 	.fsync = v9fs_file_fsync_dotl,
676 };
677 
678 const struct file_operations v9fs_file_operations = {
679 	.llseek = generic_file_llseek,
680 	.read_iter = v9fs_file_read_iter,
681 	.write_iter = v9fs_file_write_iter,
682 	.open = v9fs_file_open,
683 	.release = v9fs_dir_release,
684 	.lock = v9fs_file_lock,
685 	.mmap = generic_file_readonly_mmap,
686 	.fsync = v9fs_file_fsync,
687 };
688 
689 const struct file_operations v9fs_file_operations_dotl = {
690 	.llseek = generic_file_llseek,
691 	.read_iter = v9fs_file_read_iter,
692 	.write_iter = v9fs_file_write_iter,
693 	.open = v9fs_file_open,
694 	.release = v9fs_dir_release,
695 	.lock = v9fs_file_lock_dotl,
696 	.flock = v9fs_file_flock_dotl,
697 	.mmap = generic_file_readonly_mmap,
698 	.fsync = v9fs_file_fsync_dotl,
699 };
700 
701 const struct file_operations v9fs_mmap_file_operations = {
702 	.llseek = generic_file_llseek,
703 	.read_iter = v9fs_mmap_file_read_iter,
704 	.write_iter = v9fs_mmap_file_write_iter,
705 	.open = v9fs_file_open,
706 	.release = v9fs_dir_release,
707 	.lock = v9fs_file_lock,
708 	.mmap = v9fs_mmap_file_mmap,
709 	.fsync = v9fs_file_fsync,
710 };
711 
712 const struct file_operations v9fs_mmap_file_operations_dotl = {
713 	.llseek = generic_file_llseek,
714 	.read_iter = v9fs_mmap_file_read_iter,
715 	.write_iter = v9fs_mmap_file_write_iter,
716 	.open = v9fs_file_open,
717 	.release = v9fs_dir_release,
718 	.lock = v9fs_file_lock_dotl,
719 	.flock = v9fs_file_flock_dotl,
720 	.mmap = v9fs_mmap_file_mmap,
721 	.fsync = v9fs_file_fsync_dotl,
722 };
723