xref: /openbmc/linux/fs/afs/file.c (revision 6cd70754)
1 // SPDX-License-Identifier: GPL-2.0-or-later
2 /* AFS filesystem file handling
3  *
4  * Copyright (C) 2002, 2007 Red Hat, Inc. All Rights Reserved.
5  * Written by David Howells (dhowells@redhat.com)
6  */
7 
8 #include <linux/kernel.h>
9 #include <linux/module.h>
10 #include <linux/init.h>
11 #include <linux/fs.h>
12 #include <linux/pagemap.h>
13 #include <linux/writeback.h>
14 #include <linux/gfp.h>
15 #include <linux/task_io_accounting_ops.h>
16 #include <linux/mm.h>
17 #include "internal.h"
18 
19 static int afs_file_mmap(struct file *file, struct vm_area_struct *vma);
20 static int afs_readpage(struct file *file, struct page *page);
21 static void afs_invalidatepage(struct page *page, unsigned int offset,
22 			       unsigned int length);
23 static int afs_releasepage(struct page *page, gfp_t gfp_flags);
24 
25 static int afs_readpages(struct file *filp, struct address_space *mapping,
26 			 struct list_head *pages, unsigned nr_pages);
27 
28 const struct file_operations afs_file_operations = {
29 	.open		= afs_open,
30 	.release	= afs_release,
31 	.llseek		= generic_file_llseek,
32 	.read_iter	= generic_file_read_iter,
33 	.write_iter	= afs_file_write,
34 	.mmap		= afs_file_mmap,
35 	.splice_read	= generic_file_splice_read,
36 	.splice_write	= iter_file_splice_write,
37 	.fsync		= afs_fsync,
38 	.lock		= afs_lock,
39 	.flock		= afs_flock,
40 };
41 
42 const struct inode_operations afs_file_inode_operations = {
43 	.getattr	= afs_getattr,
44 	.setattr	= afs_setattr,
45 	.permission	= afs_permission,
46 };
47 
48 const struct address_space_operations afs_fs_aops = {
49 	.readpage	= afs_readpage,
50 	.readpages	= afs_readpages,
51 	.set_page_dirty	= afs_set_page_dirty,
52 	.launder_page	= afs_launder_page,
53 	.releasepage	= afs_releasepage,
54 	.invalidatepage	= afs_invalidatepage,
55 	.write_begin	= afs_write_begin,
56 	.write_end	= afs_write_end,
57 	.writepage	= afs_writepage,
58 	.writepages	= afs_writepages,
59 };
60 
61 static const struct vm_operations_struct afs_vm_ops = {
62 	.fault		= filemap_fault,
63 	.map_pages	= filemap_map_pages,
64 	.page_mkwrite	= afs_page_mkwrite,
65 };
66 
67 /*
68  * Discard a pin on a writeback key.
69  */
70 void afs_put_wb_key(struct afs_wb_key *wbk)
71 {
72 	if (wbk && refcount_dec_and_test(&wbk->usage)) {
73 		key_put(wbk->key);
74 		kfree(wbk);
75 	}
76 }
77 
78 /*
79  * Cache key for writeback.
80  */
81 int afs_cache_wb_key(struct afs_vnode *vnode, struct afs_file *af)
82 {
83 	struct afs_wb_key *wbk, *p;
84 
85 	wbk = kzalloc(sizeof(struct afs_wb_key), GFP_KERNEL);
86 	if (!wbk)
87 		return -ENOMEM;
88 	refcount_set(&wbk->usage, 2);
89 	wbk->key = af->key;
90 
91 	spin_lock(&vnode->wb_lock);
92 	list_for_each_entry(p, &vnode->wb_keys, vnode_link) {
93 		if (p->key == wbk->key)
94 			goto found;
95 	}
96 
97 	key_get(wbk->key);
98 	list_add_tail(&wbk->vnode_link, &vnode->wb_keys);
99 	spin_unlock(&vnode->wb_lock);
100 	af->wb = wbk;
101 	return 0;
102 
103 found:
104 	refcount_inc(&p->usage);
105 	spin_unlock(&vnode->wb_lock);
106 	af->wb = p;
107 	kfree(wbk);
108 	return 0;
109 }
110 
111 /*
112  * open an AFS file or directory and attach a key to it
113  */
114 int afs_open(struct inode *inode, struct file *file)
115 {
116 	struct afs_vnode *vnode = AFS_FS_I(inode);
117 	struct afs_file *af;
118 	struct key *key;
119 	int ret;
120 
121 	_enter("{%llx:%llu},", vnode->fid.vid, vnode->fid.vnode);
122 
123 	key = afs_request_key(vnode->volume->cell);
124 	if (IS_ERR(key)) {
125 		ret = PTR_ERR(key);
126 		goto error;
127 	}
128 
129 	af = kzalloc(sizeof(*af), GFP_KERNEL);
130 	if (!af) {
131 		ret = -ENOMEM;
132 		goto error_key;
133 	}
134 	af->key = key;
135 
136 	ret = afs_validate(vnode, key);
137 	if (ret < 0)
138 		goto error_af;
139 
140 	if (file->f_mode & FMODE_WRITE) {
141 		ret = afs_cache_wb_key(vnode, af);
142 		if (ret < 0)
143 			goto error_af;
144 	}
145 
146 	if (file->f_flags & O_TRUNC)
147 		set_bit(AFS_VNODE_NEW_CONTENT, &vnode->flags);
148 
149 	file->private_data = af;
150 	_leave(" = 0");
151 	return 0;
152 
153 error_af:
154 	kfree(af);
155 error_key:
156 	key_put(key);
157 error:
158 	_leave(" = %d", ret);
159 	return ret;
160 }
161 
162 /*
163  * release an AFS file or directory and discard its key
164  */
165 int afs_release(struct inode *inode, struct file *file)
166 {
167 	struct afs_vnode *vnode = AFS_FS_I(inode);
168 	struct afs_file *af = file->private_data;
169 	int ret = 0;
170 
171 	_enter("{%llx:%llu},", vnode->fid.vid, vnode->fid.vnode);
172 
173 	if ((file->f_mode & FMODE_WRITE))
174 		ret = vfs_fsync(file, 0);
175 
176 	file->private_data = NULL;
177 	if (af->wb)
178 		afs_put_wb_key(af->wb);
179 	key_put(af->key);
180 	kfree(af);
181 	afs_prune_wb_keys(vnode);
182 	_leave(" = %d", ret);
183 	return ret;
184 }
185 
186 /*
187  * Dispose of a ref to a read record.
188  */
189 void afs_put_read(struct afs_read *req)
190 {
191 	int i;
192 
193 	if (refcount_dec_and_test(&req->usage)) {
194 		if (req->pages) {
195 			for (i = 0; i < req->nr_pages; i++)
196 				if (req->pages[i])
197 					put_page(req->pages[i]);
198 			if (req->pages != req->array)
199 				kfree(req->pages);
200 		}
201 		kfree(req);
202 	}
203 }
204 
205 #ifdef CONFIG_AFS_FSCACHE
206 /*
207  * deal with notification that a page was read from the cache
208  */
209 static void afs_file_readpage_read_complete(struct page *page,
210 					    void *data,
211 					    int error)
212 {
213 	_enter("%p,%p,%d", page, data, error);
214 
215 	/* if the read completes with an error, we just unlock the page and let
216 	 * the VM reissue the readpage */
217 	if (!error)
218 		SetPageUptodate(page);
219 	unlock_page(page);
220 }
221 #endif
222 
223 static void afs_fetch_data_success(struct afs_operation *op)
224 {
225 	struct afs_vnode *vnode = op->file[0].vnode;
226 
227 	_enter("op=%08x", op->debug_id);
228 	afs_vnode_commit_status(op, &op->file[0]);
229 	afs_stat_v(vnode, n_fetches);
230 	atomic_long_add(op->fetch.req->actual_len, &op->net->n_fetch_bytes);
231 }
232 
233 static void afs_fetch_data_put(struct afs_operation *op)
234 {
235 	afs_put_read(op->fetch.req);
236 }
237 
238 static const struct afs_operation_ops afs_fetch_data_operation = {
239 	.issue_afs_rpc	= afs_fs_fetch_data,
240 	.issue_yfs_rpc	= yfs_fs_fetch_data,
241 	.success	= afs_fetch_data_success,
242 	.aborted	= afs_check_for_remote_deletion,
243 	.put		= afs_fetch_data_put,
244 };
245 
246 /*
247  * Fetch file data from the volume.
248  */
249 int afs_fetch_data(struct afs_vnode *vnode, struct key *key, struct afs_read *req)
250 {
251 	struct afs_operation *op;
252 
253 	_enter("%s{%llx:%llu.%u},%x,,,",
254 	       vnode->volume->name,
255 	       vnode->fid.vid,
256 	       vnode->fid.vnode,
257 	       vnode->fid.unique,
258 	       key_serial(key));
259 
260 	op = afs_alloc_operation(key, vnode->volume);
261 	if (IS_ERR(op))
262 		return PTR_ERR(op);
263 
264 	afs_op_set_vnode(op, 0, vnode);
265 
266 	op->fetch.req	= afs_get_read(req);
267 	op->ops		= &afs_fetch_data_operation;
268 	return afs_do_sync_operation(op);
269 }
270 
271 /*
272  * read page from file, directory or symlink, given a key to use
273  */
274 int afs_page_filler(void *data, struct page *page)
275 {
276 	struct inode *inode = page->mapping->host;
277 	struct afs_vnode *vnode = AFS_FS_I(inode);
278 	struct afs_read *req;
279 	struct key *key = data;
280 	int ret;
281 
282 	_enter("{%x},{%lu},{%lu}", key_serial(key), inode->i_ino, page->index);
283 
284 	BUG_ON(!PageLocked(page));
285 
286 	ret = -ESTALE;
287 	if (test_bit(AFS_VNODE_DELETED, &vnode->flags))
288 		goto error;
289 
290 	/* is it cached? */
291 #ifdef CONFIG_AFS_FSCACHE
292 	ret = fscache_read_or_alloc_page(vnode->cache,
293 					 page,
294 					 afs_file_readpage_read_complete,
295 					 NULL,
296 					 GFP_KERNEL);
297 #else
298 	ret = -ENOBUFS;
299 #endif
300 	switch (ret) {
301 		/* read BIO submitted (page in cache) */
302 	case 0:
303 		break;
304 
305 		/* page not yet cached */
306 	case -ENODATA:
307 		_debug("cache said ENODATA");
308 		goto go_on;
309 
310 		/* page will not be cached */
311 	case -ENOBUFS:
312 		_debug("cache said ENOBUFS");
313 
314 		fallthrough;
315 	default:
316 	go_on:
317 		req = kzalloc(struct_size(req, array, 1), GFP_KERNEL);
318 		if (!req)
319 			goto enomem;
320 
321 		/* We request a full page.  If the page is a partial one at the
322 		 * end of the file, the server will return a short read and the
323 		 * unmarshalling code will clear the unfilled space.
324 		 */
325 		refcount_set(&req->usage, 1);
326 		req->pos = (loff_t)page->index << PAGE_SHIFT;
327 		req->len = PAGE_SIZE;
328 		req->nr_pages = 1;
329 		req->pages = req->array;
330 		req->pages[0] = page;
331 		get_page(page);
332 
333 		/* read the contents of the file from the server into the
334 		 * page */
335 		ret = afs_fetch_data(vnode, key, req);
336 		afs_put_read(req);
337 
338 		if (ret < 0) {
339 			if (ret == -ENOENT) {
340 				_debug("got NOENT from server"
341 				       " - marking file deleted and stale");
342 				set_bit(AFS_VNODE_DELETED, &vnode->flags);
343 				ret = -ESTALE;
344 			}
345 
346 #ifdef CONFIG_AFS_FSCACHE
347 			fscache_uncache_page(vnode->cache, page);
348 #endif
349 			BUG_ON(PageFsCache(page));
350 
351 			if (ret == -EINTR ||
352 			    ret == -ENOMEM ||
353 			    ret == -ERESTARTSYS ||
354 			    ret == -EAGAIN)
355 				goto error;
356 			goto io_error;
357 		}
358 
359 		SetPageUptodate(page);
360 
361 		/* send the page to the cache */
362 #ifdef CONFIG_AFS_FSCACHE
363 		if (PageFsCache(page) &&
364 		    fscache_write_page(vnode->cache, page, vnode->status.size,
365 				       GFP_KERNEL) != 0) {
366 			fscache_uncache_page(vnode->cache, page);
367 			BUG_ON(PageFsCache(page));
368 		}
369 #endif
370 		unlock_page(page);
371 	}
372 
373 	_leave(" = 0");
374 	return 0;
375 
376 io_error:
377 	SetPageError(page);
378 	goto error;
379 enomem:
380 	ret = -ENOMEM;
381 error:
382 	unlock_page(page);
383 	_leave(" = %d", ret);
384 	return ret;
385 }
386 
387 /*
388  * read page from file, directory or symlink, given a file to nominate the key
389  * to be used
390  */
391 static int afs_readpage(struct file *file, struct page *page)
392 {
393 	struct key *key;
394 	int ret;
395 
396 	if (file) {
397 		key = afs_file_key(file);
398 		ASSERT(key != NULL);
399 		ret = afs_page_filler(key, page);
400 	} else {
401 		struct inode *inode = page->mapping->host;
402 		key = afs_request_key(AFS_FS_S(inode->i_sb)->cell);
403 		if (IS_ERR(key)) {
404 			ret = PTR_ERR(key);
405 		} else {
406 			ret = afs_page_filler(key, page);
407 			key_put(key);
408 		}
409 	}
410 	return ret;
411 }
412 
413 /*
414  * Make pages available as they're filled.
415  */
416 static void afs_readpages_page_done(struct afs_read *req)
417 {
418 #ifdef CONFIG_AFS_FSCACHE
419 	struct afs_vnode *vnode = req->vnode;
420 #endif
421 	struct page *page = req->pages[req->index];
422 
423 	req->pages[req->index] = NULL;
424 	SetPageUptodate(page);
425 
426 	/* send the page to the cache */
427 #ifdef CONFIG_AFS_FSCACHE
428 	if (PageFsCache(page) &&
429 	    fscache_write_page(vnode->cache, page, vnode->status.size,
430 			       GFP_KERNEL) != 0) {
431 		fscache_uncache_page(vnode->cache, page);
432 		BUG_ON(PageFsCache(page));
433 	}
434 #endif
435 	unlock_page(page);
436 	put_page(page);
437 }
438 
439 /*
440  * Read a contiguous set of pages.
441  */
442 static int afs_readpages_one(struct file *file, struct address_space *mapping,
443 			     struct list_head *pages)
444 {
445 	struct afs_vnode *vnode = AFS_FS_I(mapping->host);
446 	struct afs_read *req;
447 	struct list_head *p;
448 	struct page *first, *page;
449 	struct key *key = afs_file_key(file);
450 	pgoff_t index;
451 	int ret, n, i;
452 
453 	/* Count the number of contiguous pages at the front of the list.  Note
454 	 * that the list goes prev-wards rather than next-wards.
455 	 */
456 	first = lru_to_page(pages);
457 	index = first->index + 1;
458 	n = 1;
459 	for (p = first->lru.prev; p != pages; p = p->prev) {
460 		page = list_entry(p, struct page, lru);
461 		if (page->index != index)
462 			break;
463 		index++;
464 		n++;
465 	}
466 
467 	req = kzalloc(struct_size(req, array, n), GFP_NOFS);
468 	if (!req)
469 		return -ENOMEM;
470 
471 	refcount_set(&req->usage, 1);
472 	req->vnode = vnode;
473 	req->page_done = afs_readpages_page_done;
474 	req->pos = first->index;
475 	req->pos <<= PAGE_SHIFT;
476 	req->pages = req->array;
477 
478 	/* Transfer the pages to the request.  We add them in until one fails
479 	 * to add to the LRU and then we stop (as that'll make a hole in the
480 	 * contiguous run.
481 	 *
482 	 * Note that it's possible for the file size to change whilst we're
483 	 * doing this, but we rely on the server returning less than we asked
484 	 * for if the file shrank.  We also rely on this to deal with a partial
485 	 * page at the end of the file.
486 	 */
487 	do {
488 		page = lru_to_page(pages);
489 		list_del(&page->lru);
490 		index = page->index;
491 		if (add_to_page_cache_lru(page, mapping, index,
492 					  readahead_gfp_mask(mapping))) {
493 #ifdef CONFIG_AFS_FSCACHE
494 			fscache_uncache_page(vnode->cache, page);
495 #endif
496 			put_page(page);
497 			break;
498 		}
499 
500 		req->pages[req->nr_pages++] = page;
501 		req->len += PAGE_SIZE;
502 	} while (req->nr_pages < n);
503 
504 	if (req->nr_pages == 0) {
505 		kfree(req);
506 		return 0;
507 	}
508 
509 	ret = afs_fetch_data(vnode, key, req);
510 	if (ret < 0)
511 		goto error;
512 
513 	task_io_account_read(PAGE_SIZE * req->nr_pages);
514 	afs_put_read(req);
515 	return 0;
516 
517 error:
518 	if (ret == -ENOENT) {
519 		_debug("got NOENT from server"
520 		       " - marking file deleted and stale");
521 		set_bit(AFS_VNODE_DELETED, &vnode->flags);
522 		ret = -ESTALE;
523 	}
524 
525 	for (i = 0; i < req->nr_pages; i++) {
526 		page = req->pages[i];
527 		if (page) {
528 #ifdef CONFIG_AFS_FSCACHE
529 			fscache_uncache_page(vnode->cache, page);
530 #endif
531 			SetPageError(page);
532 			unlock_page(page);
533 		}
534 	}
535 
536 	afs_put_read(req);
537 	return ret;
538 }
539 
540 /*
541  * read a set of pages
542  */
543 static int afs_readpages(struct file *file, struct address_space *mapping,
544 			 struct list_head *pages, unsigned nr_pages)
545 {
546 	struct key *key = afs_file_key(file);
547 	struct afs_vnode *vnode;
548 	int ret = 0;
549 
550 	_enter("{%d},{%lu},,%d",
551 	       key_serial(key), mapping->host->i_ino, nr_pages);
552 
553 	ASSERT(key != NULL);
554 
555 	vnode = AFS_FS_I(mapping->host);
556 	if (test_bit(AFS_VNODE_DELETED, &vnode->flags)) {
557 		_leave(" = -ESTALE");
558 		return -ESTALE;
559 	}
560 
561 	/* attempt to read as many of the pages as possible */
562 #ifdef CONFIG_AFS_FSCACHE
563 	ret = fscache_read_or_alloc_pages(vnode->cache,
564 					  mapping,
565 					  pages,
566 					  &nr_pages,
567 					  afs_file_readpage_read_complete,
568 					  NULL,
569 					  mapping_gfp_mask(mapping));
570 #else
571 	ret = -ENOBUFS;
572 #endif
573 
574 	switch (ret) {
575 		/* all pages are being read from the cache */
576 	case 0:
577 		BUG_ON(!list_empty(pages));
578 		BUG_ON(nr_pages != 0);
579 		_leave(" = 0 [reading all]");
580 		return 0;
581 
582 		/* there were pages that couldn't be read from the cache */
583 	case -ENODATA:
584 	case -ENOBUFS:
585 		break;
586 
587 		/* other error */
588 	default:
589 		_leave(" = %d", ret);
590 		return ret;
591 	}
592 
593 	while (!list_empty(pages)) {
594 		ret = afs_readpages_one(file, mapping, pages);
595 		if (ret < 0)
596 			break;
597 	}
598 
599 	_leave(" = %d [netting]", ret);
600 	return ret;
601 }
602 
603 /*
604  * Adjust the dirty region of the page on truncation or full invalidation,
605  * getting rid of the markers altogether if the region is entirely invalidated.
606  */
607 static void afs_invalidate_dirty(struct page *page, unsigned int offset,
608 				 unsigned int length)
609 {
610 	struct afs_vnode *vnode = AFS_FS_I(page->mapping->host);
611 	unsigned long priv;
612 	unsigned int f, t, end = offset + length;
613 
614 	priv = page_private(page);
615 
616 	/* we clean up only if the entire page is being invalidated */
617 	if (offset == 0 && length == thp_size(page))
618 		goto full_invalidate;
619 
620 	 /* If the page was dirtied by page_mkwrite(), the PTE stays writable
621 	  * and we don't get another notification to tell us to expand it
622 	  * again.
623 	  */
624 	if (afs_is_page_dirty_mmapped(priv))
625 		return;
626 
627 	/* We may need to shorten the dirty region */
628 	f = afs_page_dirty_from(priv);
629 	t = afs_page_dirty_to(priv);
630 
631 	if (t <= offset || f >= end)
632 		return; /* Doesn't overlap */
633 
634 	if (f < offset && t > end)
635 		return; /* Splits the dirty region - just absorb it */
636 
637 	if (f >= offset && t <= end)
638 		goto undirty;
639 
640 	if (f < offset)
641 		t = offset;
642 	else
643 		f = end;
644 	if (f == t)
645 		goto undirty;
646 
647 	priv = afs_page_dirty(f, t);
648 	set_page_private(page, priv);
649 	trace_afs_page_dirty(vnode, tracepoint_string("trunc"), page->index, priv);
650 	return;
651 
652 undirty:
653 	trace_afs_page_dirty(vnode, tracepoint_string("undirty"), page->index, priv);
654 	clear_page_dirty_for_io(page);
655 full_invalidate:
656 	priv = (unsigned long)detach_page_private(page);
657 	trace_afs_page_dirty(vnode, tracepoint_string("inval"), page->index, priv);
658 }
659 
660 /*
661  * invalidate part or all of a page
662  * - release a page and clean up its private data if offset is 0 (indicating
663  *   the entire page)
664  */
665 static void afs_invalidatepage(struct page *page, unsigned int offset,
666 			       unsigned int length)
667 {
668 	_enter("{%lu},%u,%u", page->index, offset, length);
669 
670 	BUG_ON(!PageLocked(page));
671 
672 #ifdef CONFIG_AFS_FSCACHE
673 	/* we clean up only if the entire page is being invalidated */
674 	if (offset == 0 && length == PAGE_SIZE) {
675 		if (PageFsCache(page)) {
676 			struct afs_vnode *vnode = AFS_FS_I(page->mapping->host);
677 			fscache_wait_on_page_write(vnode->cache, page);
678 			fscache_uncache_page(vnode->cache, page);
679 		}
680 	}
681 #endif
682 
683 	if (PagePrivate(page))
684 		afs_invalidate_dirty(page, offset, length);
685 
686 	_leave("");
687 }
688 
689 /*
690  * release a page and clean up its private state if it's not busy
691  * - return true if the page can now be released, false if not
692  */
693 static int afs_releasepage(struct page *page, gfp_t gfp_flags)
694 {
695 	struct afs_vnode *vnode = AFS_FS_I(page->mapping->host);
696 	unsigned long priv;
697 
698 	_enter("{{%llx:%llu}[%lu],%lx},%x",
699 	       vnode->fid.vid, vnode->fid.vnode, page->index, page->flags,
700 	       gfp_flags);
701 
702 	/* deny if page is being written to the cache and the caller hasn't
703 	 * elected to wait */
704 #ifdef CONFIG_AFS_FSCACHE
705 	if (!fscache_maybe_release_page(vnode->cache, page, gfp_flags)) {
706 		_leave(" = F [cache busy]");
707 		return 0;
708 	}
709 #endif
710 
711 	if (PagePrivate(page)) {
712 		priv = (unsigned long)detach_page_private(page);
713 		trace_afs_page_dirty(vnode, tracepoint_string("rel"),
714 				     page->index, priv);
715 	}
716 
717 	/* indicate that the page can be released */
718 	_leave(" = T");
719 	return 1;
720 }
721 
722 /*
723  * Handle setting up a memory mapping on an AFS file.
724  */
725 static int afs_file_mmap(struct file *file, struct vm_area_struct *vma)
726 {
727 	int ret;
728 
729 	ret = generic_file_mmap(file, vma);
730 	if (ret == 0)
731 		vma->vm_ops = &afs_vm_ops;
732 	return ret;
733 }
734