xref: /openbmc/linux/fs/nfs/read.c (revision 000dbe0bec058cbf2ca9e156e4a5584f5158b0f9)
1457c8996SThomas Gleixner // SPDX-License-Identifier: GPL-2.0-only
21da177e4SLinus Torvalds /*
31da177e4SLinus Torvalds  * linux/fs/nfs/read.c
41da177e4SLinus Torvalds  *
51da177e4SLinus Torvalds  * Block I/O for NFS
61da177e4SLinus Torvalds  *
71da177e4SLinus Torvalds  * Partial copy of Linus' read cache modifications to fs/nfs/file.c
81da177e4SLinus Torvalds  * modified for async RPC by okir@monad.swb.de
91da177e4SLinus Torvalds  */
101da177e4SLinus Torvalds 
111da177e4SLinus Torvalds #include <linux/time.h>
121da177e4SLinus Torvalds #include <linux/kernel.h>
131da177e4SLinus Torvalds #include <linux/errno.h>
141da177e4SLinus Torvalds #include <linux/fcntl.h>
151da177e4SLinus Torvalds #include <linux/stat.h>
161da177e4SLinus Torvalds #include <linux/mm.h>
171da177e4SLinus Torvalds #include <linux/slab.h>
189c88ea00SDave Wysochanski #include <linux/task_io_accounting_ops.h>
191da177e4SLinus Torvalds #include <linux/pagemap.h>
201da177e4SLinus Torvalds #include <linux/sunrpc/clnt.h>
211da177e4SLinus Torvalds #include <linux/nfs_fs.h>
221da177e4SLinus Torvalds #include <linux/nfs_page.h>
2364419a9bSAndy Adamson #include <linux/module.h>
241da177e4SLinus Torvalds 
25f11c88afSAndy Adamson #include "nfs4_fs.h"
2649a70f27STrond Myklebust #include "internal.h"
2791d5b470SChuck Lever #include "iostat.h"
289a9fc1c0SDavid Howells #include "fscache.h"
29fab5fc25SChristoph Hellwig #include "pnfs.h"
308224b273SChuck Lever #include "nfstrace.h"
3191d5b470SChuck Lever 
321da177e4SLinus Torvalds #define NFSDBG_FACILITY		NFSDBG_PAGECACHE
331da177e4SLinus Torvalds 
34*000dbe0bSDave Wysochanski const struct nfs_pgio_completion_ops nfs_async_read_completion_ops;
354a0de55cSAnna Schumaker static const struct nfs_rw_ops nfs_rw_read_ops;
361da177e4SLinus Torvalds 
37e18b890bSChristoph Lameter static struct kmem_cache *nfs_rdata_cachep;
381da177e4SLinus Torvalds 
391e7f3a48SWeston Andros Adamson static struct nfs_pgio_header *nfs_readhdr_alloc(void)
403feb2d49STrond Myklebust {
41fbe77c30SBenjamin Coddington 	struct nfs_pgio_header *p = kmem_cache_zalloc(nfs_rdata_cachep, GFP_KERNEL);
42fbe77c30SBenjamin Coddington 
43fbe77c30SBenjamin Coddington 	if (p)
44fbe77c30SBenjamin Coddington 		p->rw_mode = FMODE_READ;
45fbe77c30SBenjamin Coddington 	return p;
464db6e0b7SFred Isaman }
474db6e0b7SFred Isaman 
481e7f3a48SWeston Andros Adamson static void nfs_readhdr_free(struct nfs_pgio_header *rhdr)
493feb2d49STrond Myklebust {
50cd841605SFred Isaman 	kmem_cache_free(nfs_rdata_cachep, rhdr);
513feb2d49STrond Myklebust }
523feb2d49STrond Myklebust 
53ab75bff1STrond Myklebust static int nfs_return_empty_folio(struct folio *folio)
541da177e4SLinus Torvalds {
55ab75bff1STrond Myklebust 	folio_zero_segment(folio, 0, folio_size(folio));
56ab75bff1STrond Myklebust 	folio_mark_uptodate(folio);
57ab75bff1STrond Myklebust 	folio_unlock(folio);
581da177e4SLinus Torvalds 	return 0;
591da177e4SLinus Torvalds }
601da177e4SLinus Torvalds 
611abb5088SBryan Schumaker void nfs_pageio_init_read(struct nfs_pageio_descriptor *pgio,
62fab5fc25SChristoph Hellwig 			      struct inode *inode, bool force_mds,
63061ae2edSFred Isaman 			      const struct nfs_pgio_completion_ops *compl_ops)
641751c363STrond Myklebust {
65fab5fc25SChristoph Hellwig 	struct nfs_server *server = NFS_SERVER(inode);
6641d8d5b7SAnna Schumaker 	const struct nfs_pageio_ops *pg_ops = &nfs_pgio_rw_ops;
67fab5fc25SChristoph Hellwig 
68fab5fc25SChristoph Hellwig #ifdef CONFIG_NFS_V4_1
69fab5fc25SChristoph Hellwig 	if (server->pnfs_curr_ld && !force_mds)
70fab5fc25SChristoph Hellwig 		pg_ops = server->pnfs_curr_ld->pg_read_ops;
71fab5fc25SChristoph Hellwig #endif
724a0de55cSAnna Schumaker 	nfs_pageio_init(pgio, inode, pg_ops, compl_ops, &nfs_rw_read_ops,
733bde7afdSTrond Myklebust 			server->rsize, 0);
741751c363STrond Myklebust }
75ddda8e0aSBryan Schumaker EXPORT_SYMBOL_GPL(nfs_pageio_init_read);
761751c363STrond Myklebust 
77*000dbe0bSDave Wysochanski void nfs_pageio_complete_read(struct nfs_pageio_descriptor *pgio)
781e83b173SDave Wysochanski {
791e83b173SDave Wysochanski 	struct nfs_pgio_mirror *pgm;
801e83b173SDave Wysochanski 	unsigned long npages;
811e83b173SDave Wysochanski 
821e83b173SDave Wysochanski 	nfs_pageio_complete(pgio);
831e83b173SDave Wysochanski 
841e83b173SDave Wysochanski 	/* It doesn't make sense to do mirrored reads! */
851e83b173SDave Wysochanski 	WARN_ON_ONCE(pgio->pg_mirror_count != 1);
861e83b173SDave Wysochanski 
871e83b173SDave Wysochanski 	pgm = &pgio->pg_mirrors[0];
88b42ad64fSDave Wysochanski 	NFS_I(pgio->pg_inode)->read_io += pgm->pg_bytes_written;
891e83b173SDave Wysochanski 	npages = (pgm->pg_bytes_written + PAGE_SIZE - 1) >> PAGE_SHIFT;
90b42ad64fSDave Wysochanski 	nfs_add_stats(pgio->pg_inode, NFSIOS_READPAGES, npages);
911e83b173SDave Wysochanski }
921e83b173SDave Wysochanski 
931e83b173SDave Wysochanski 
94493292ddSTrond Myklebust void nfs_pageio_reset_read_mds(struct nfs_pageio_descriptor *pgio)
95493292ddSTrond Myklebust {
96a7d42ddbSWeston Andros Adamson 	struct nfs_pgio_mirror *mirror;
97a7d42ddbSWeston Andros Adamson 
986f29b9bbSKinglong Mee 	if (pgio->pg_ops && pgio->pg_ops->pg_cleanup)
996f29b9bbSKinglong Mee 		pgio->pg_ops->pg_cleanup(pgio);
1006f29b9bbSKinglong Mee 
10141d8d5b7SAnna Schumaker 	pgio->pg_ops = &nfs_pgio_rw_ops;
102a7d42ddbSWeston Andros Adamson 
103a7d42ddbSWeston Andros Adamson 	/* read path should never have more than one mirror */
104a7d42ddbSWeston Andros Adamson 	WARN_ON_ONCE(pgio->pg_mirror_count != 1);
105a7d42ddbSWeston Andros Adamson 
106a7d42ddbSWeston Andros Adamson 	mirror = &pgio->pg_mirrors[0];
107a7d42ddbSWeston Andros Adamson 	mirror->pg_bsize = NFS_SERVER(pgio->pg_inode)->rsize;
108493292ddSTrond Myklebust }
1091f945357STrond Myklebust EXPORT_SYMBOL_GPL(nfs_pageio_reset_read_mds);
110493292ddSTrond Myklebust 
1118f54c7a4STrond Myklebust static void nfs_readpage_release(struct nfs_page *req, int error)
1120bcbf039SPeng Tao {
113ab75bff1STrond Myklebust 	struct folio *folio = nfs_page_to_folio(req);
1140bcbf039SPeng Tao 
1158f54c7a4STrond Myklebust 	if (nfs_error_is_fatal_on_server(error) && error != -ETIMEDOUT)
116ab75bff1STrond Myklebust 		folio_set_error(folio);
117*000dbe0bSDave Wysochanski 	if (nfs_page_group_sync_on_bit(req, PG_UNLOCKPAGE))
118*000dbe0bSDave Wysochanski 		if (nfs_netfs_folio_unlock(folio))
119ab75bff1STrond Myklebust 			folio_unlock(folio);
120*000dbe0bSDave Wysochanski 
1210bcbf039SPeng Tao 	nfs_release_request(req);
1220bcbf039SPeng Tao }
1230bcbf039SPeng Tao 
12467d0338eSWeston Andros Adamson static void nfs_page_group_set_uptodate(struct nfs_page *req)
12567d0338eSWeston Andros Adamson {
12667d0338eSWeston Andros Adamson 	if (nfs_page_group_sync_on_bit(req, PG_UPTODATE))
127ab75bff1STrond Myklebust 		folio_mark_uptodate(nfs_page_to_folio(req));
12867d0338eSWeston Andros Adamson }
12967d0338eSWeston Andros Adamson 
130061ae2edSFred Isaman static void nfs_read_completion(struct nfs_pgio_header *hdr)
1314db6e0b7SFred Isaman {
1324db6e0b7SFred Isaman 	unsigned long bytes = 0;
1338f54c7a4STrond Myklebust 	int error;
1344db6e0b7SFred Isaman 
1354db6e0b7SFred Isaman 	if (test_bit(NFS_IOHDR_REDO, &hdr->flags))
1364db6e0b7SFred Isaman 		goto out;
1374db6e0b7SFred Isaman 	while (!list_empty(&hdr->pages)) {
1384db6e0b7SFred Isaman 		struct nfs_page *req = nfs_list_entry(hdr->pages.next);
139ab75bff1STrond Myklebust 		struct folio *folio = nfs_page_to_folio(req);
14068072992SWeston Andros Adamson 		unsigned long start = req->wb_pgbase;
14168072992SWeston Andros Adamson 		unsigned long end = req->wb_pgbase + req->wb_bytes;
1424db6e0b7SFred Isaman 
1434db6e0b7SFred Isaman 		if (test_bit(NFS_IOHDR_EOF, &hdr->flags)) {
14468072992SWeston Andros Adamson 			/* note: regions of the page not covered by a
14501c3a400SDave Wysochanski 			 * request are zeroed in nfs_read_add_folio
14601c3a400SDave Wysochanski 			 */
14768072992SWeston Andros Adamson 			if (bytes > hdr->good_bytes) {
14868072992SWeston Andros Adamson 				/* nothing in this request was good, so zero
14968072992SWeston Andros Adamson 				 * the full extent of the request */
150ab75bff1STrond Myklebust 				folio_zero_segment(folio, start, end);
15168072992SWeston Andros Adamson 
15268072992SWeston Andros Adamson 			} else if (hdr->good_bytes - bytes < req->wb_bytes) {
15368072992SWeston Andros Adamson 				/* part of this request has good bytes, but
15468072992SWeston Andros Adamson 				 * not all. zero the bad bytes */
15568072992SWeston Andros Adamson 				start += hdr->good_bytes - bytes;
15668072992SWeston Andros Adamson 				WARN_ON(start < req->wb_pgbase);
157ab75bff1STrond Myklebust 				folio_zero_segment(folio, start, end);
15868072992SWeston Andros Adamson 			}
1594db6e0b7SFred Isaman 		}
1608f54c7a4STrond Myklebust 		error = 0;
1614bd8b010STrond Myklebust 		bytes += req->wb_bytes;
1624bd8b010STrond Myklebust 		if (test_bit(NFS_IOHDR_ERROR, &hdr->flags)) {
1634bd8b010STrond Myklebust 			if (bytes <= hdr->good_bytes)
16467d0338eSWeston Andros Adamson 				nfs_page_group_set_uptodate(req);
1658f54c7a4STrond Myklebust 			else {
1668f54c7a4STrond Myklebust 				error = hdr->error;
1678f54c7a4STrond Myklebust 				xchg(&nfs_req_openctx(req)->error, error);
1688f54c7a4STrond Myklebust 			}
1694bd8b010STrond Myklebust 		} else
17067d0338eSWeston Andros Adamson 			nfs_page_group_set_uptodate(req);
1714db6e0b7SFred Isaman 		nfs_list_remove_request(req);
1728f54c7a4STrond Myklebust 		nfs_readpage_release(req, error);
1734db6e0b7SFred Isaman 	}
174*000dbe0bSDave Wysochanski 	nfs_netfs_read_completion(hdr);
175*000dbe0bSDave Wysochanski 
1764db6e0b7SFred Isaman out:
1774db6e0b7SFred Isaman 	hdr->release(hdr);
1784db6e0b7SFred Isaman }
1794db6e0b7SFred Isaman 
180d45f60c6SWeston Andros Adamson static void nfs_initiate_read(struct nfs_pgio_header *hdr,
181d45f60c6SWeston Andros Adamson 			      struct rpc_message *msg,
182abde71f4STom Haynes 			      const struct nfs_rpc_ops *rpc_ops,
1831ed26f33SAnna Schumaker 			      struct rpc_task_setup *task_setup_data, int how)
18464419a9bSAndy Adamson {
185abde71f4STom Haynes 	rpc_ops->read_setup(hdr, msg);
186*000dbe0bSDave Wysochanski 	nfs_netfs_initiate_read(hdr);
1872343172dSTrond Myklebust 	trace_nfs_initiate_read(hdr);
18864419a9bSAndy Adamson }
18964419a9bSAndy Adamson 
190061ae2edSFred Isaman static void
191df3accb8STrond Myklebust nfs_async_read_error(struct list_head *head, int error)
1921da177e4SLinus Torvalds {
1931da177e4SLinus Torvalds 	struct nfs_page	*req;
1941da177e4SLinus Torvalds 
1951da177e4SLinus Torvalds 	while (!list_empty(head)) {
1961da177e4SLinus Torvalds 		req = nfs_list_entry(head->next);
1971da177e4SLinus Torvalds 		nfs_list_remove_request(req);
1988f54c7a4STrond Myklebust 		nfs_readpage_release(req, error);
1991da177e4SLinus Torvalds 	}
2001da177e4SLinus Torvalds }
2011da177e4SLinus Torvalds 
202*000dbe0bSDave Wysochanski const struct nfs_pgio_completion_ops nfs_async_read_completion_ops = {
203061ae2edSFred Isaman 	.error_cleanup = nfs_async_read_error,
204061ae2edSFred Isaman 	.completion = nfs_read_completion,
205061ae2edSFred Isaman };
206061ae2edSFred Isaman 
2071da177e4SLinus Torvalds /*
2080b671301STrond Myklebust  * This is the callback from RPC telling us whether a reply was
2090b671301STrond Myklebust  * received or some error occurred (timeout or socket shutdown).
2100b671301STrond Myklebust  */
211d45f60c6SWeston Andros Adamson static int nfs_readpage_done(struct rpc_task *task,
212d45f60c6SWeston Andros Adamson 			     struct nfs_pgio_header *hdr,
2130eecb214SAnna Schumaker 			     struct inode *inode)
2140b671301STrond Myklebust {
215d45f60c6SWeston Andros Adamson 	int status = NFS_PROTO(inode)->read_done(task, hdr);
2160b671301STrond Myklebust 	if (status != 0)
2170b671301STrond Myklebust 		return status;
2180b671301STrond Myklebust 
219d45f60c6SWeston Andros Adamson 	nfs_add_stats(inode, NFSIOS_SERVERREADBYTES, hdr->res.count);
2202343172dSTrond Myklebust 	trace_nfs_readpage_done(task, hdr);
2210b671301STrond Myklebust 
2220b671301STrond Myklebust 	if (task->tk_status == -ESTALE) {
22393ce4af7STrond Myklebust 		nfs_set_inode_stale(inode);
224cd841605SFred Isaman 		nfs_mark_for_revalidate(inode);
2250b671301STrond Myklebust 	}
2260b671301STrond Myklebust 	return 0;
2270b671301STrond Myklebust }
2280b671301STrond Myklebust 
229d45f60c6SWeston Andros Adamson static void nfs_readpage_retry(struct rpc_task *task,
230d45f60c6SWeston Andros Adamson 			       struct nfs_pgio_header *hdr)
2310b671301STrond Myklebust {
232d45f60c6SWeston Andros Adamson 	struct nfs_pgio_args *argp = &hdr->args;
233d45f60c6SWeston Andros Adamson 	struct nfs_pgio_res  *resp = &hdr->res;
2340b671301STrond Myklebust 
2350b671301STrond Myklebust 	/* This is a short read! */
236d45f60c6SWeston Andros Adamson 	nfs_inc_stats(hdr->inode, NFSIOS_SHORTREAD);
237fd2b6121SChuck Lever 	trace_nfs_readpage_short(task, hdr);
238fd2b6121SChuck Lever 
2390b671301STrond Myklebust 	/* Has the server at least made some progress? */
2404db6e0b7SFred Isaman 	if (resp->count == 0) {
241d45f60c6SWeston Andros Adamson 		nfs_set_pgio_error(hdr, -EIO, argp->offset);
242d61e612aSTrond Myklebust 		return;
2434db6e0b7SFred Isaman 	}
244f8417b48SKinglong Mee 
245f8417b48SKinglong Mee 	/* For non rpc-based layout drivers, retry-through-MDS */
246f8417b48SKinglong Mee 	if (!task->tk_ops) {
247f8417b48SKinglong Mee 		hdr->pnfs_error = -EAGAIN;
248f8417b48SKinglong Mee 		return;
249f8417b48SKinglong Mee 	}
250f8417b48SKinglong Mee 
251d45f60c6SWeston Andros Adamson 	/* Yes, so retry the read at the end of the hdr */
252d45f60c6SWeston Andros Adamson 	hdr->mds_offset += resp->count;
2530b671301STrond Myklebust 	argp->offset += resp->count;
2540b671301STrond Myklebust 	argp->pgbase += resp->count;
2550b671301STrond Myklebust 	argp->count -= resp->count;
2568c9cb714STrond Myklebust 	resp->count = 0;
2578c9cb714STrond Myklebust 	resp->eof = 0;
258d00c5d43STrond Myklebust 	rpc_restart_call_prepare(task);
2590b671301STrond Myklebust }
2600b671301STrond Myklebust 
261d45f60c6SWeston Andros Adamson static void nfs_readpage_result(struct rpc_task *task,
262d45f60c6SWeston Andros Adamson 				struct nfs_pgio_header *hdr)
2631da177e4SLinus Torvalds {
264d45f60c6SWeston Andros Adamson 	if (hdr->res.eof) {
2651c6c4b74STrond Myklebust 		loff_t pos = hdr->args.offset + hdr->res.count;
2661c6c4b74STrond Myklebust 		unsigned int new = pos - hdr->io_start;
267fdd1e74cSTrond Myklebust 
2681c6c4b74STrond Myklebust 		if (hdr->good_bytes > new) {
2691c6c4b74STrond Myklebust 			hdr->good_bytes = new;
2704db6e0b7SFred Isaman 			set_bit(NFS_IOHDR_EOF, &hdr->flags);
2714db6e0b7SFred Isaman 			clear_bit(NFS_IOHDR_ERROR, &hdr->flags);
2724db6e0b7SFred Isaman 		}
273f8417b48SKinglong Mee 	} else if (hdr->res.count < hdr->args.count)
274d45f60c6SWeston Andros Adamson 		nfs_readpage_retry(task, hdr);
2750b671301STrond Myklebust }
276fdd1e74cSTrond Myklebust 
277*000dbe0bSDave Wysochanski int nfs_read_add_folio(struct nfs_pageio_descriptor *pgio,
27801c3a400SDave Wysochanski 		       struct nfs_open_context *ctx,
27901c3a400SDave Wysochanski 		       struct folio *folio)
2801e83b173SDave Wysochanski {
281ab75bff1STrond Myklebust 	struct inode *inode = folio_file_mapping(folio)->host;
282ab75bff1STrond Myklebust 	struct nfs_server *server = NFS_SERVER(inode);
283ab75bff1STrond Myklebust 	size_t fsize = folio_size(folio);
284ab75bff1STrond Myklebust 	unsigned int rsize = server->rsize;
2851e83b173SDave Wysochanski 	struct nfs_page *new;
2868cfb9015STrond Myklebust 	unsigned int len, aligned_len;
2871e83b173SDave Wysochanski 	int error;
2881e83b173SDave Wysochanski 
289ab75bff1STrond Myklebust 	len = nfs_folio_length(folio);
2901e83b173SDave Wysochanski 	if (len == 0)
291ab75bff1STrond Myklebust 		return nfs_return_empty_folio(folio);
2921e83b173SDave Wysochanski 
293ab75bff1STrond Myklebust 	aligned_len = min_t(unsigned int, ALIGN(len, rsize), fsize);
2948cfb9015STrond Myklebust 
29501c3a400SDave Wysochanski 	new = nfs_page_create_from_folio(ctx, folio, 0, aligned_len);
296*000dbe0bSDave Wysochanski 	if (IS_ERR(new)) {
297*000dbe0bSDave Wysochanski 		error = PTR_ERR(new);
298*000dbe0bSDave Wysochanski 		goto out;
299*000dbe0bSDave Wysochanski 	}
3001e83b173SDave Wysochanski 
301ab75bff1STrond Myklebust 	if (len < fsize)
302ab75bff1STrond Myklebust 		folio_zero_segment(folio, len, fsize);
30301c3a400SDave Wysochanski 	if (!nfs_pageio_add_request(pgio, new)) {
3041e83b173SDave Wysochanski 		nfs_list_remove_request(new);
30501c3a400SDave Wysochanski 		error = pgio->pg_error;
3061e83b173SDave Wysochanski 		nfs_readpage_release(new, error);
3071e83b173SDave Wysochanski 		goto out;
3081e83b173SDave Wysochanski 	}
3091e83b173SDave Wysochanski 	return 0;
3101e83b173SDave Wysochanski out:
3111e83b173SDave Wysochanski 	return error;
3121e83b173SDave Wysochanski }
3131e83b173SDave Wysochanski 
3141da177e4SLinus Torvalds /*
3151da177e4SLinus Torvalds  * Read a page over NFS.
3161da177e4SLinus Torvalds  * We read the page synchronously in the following case:
3171da177e4SLinus Torvalds  *  -	The error flag is set for this page. This happens only when a
3181da177e4SLinus Torvalds  *	previous async read operation failed.
3191da177e4SLinus Torvalds  */
32065d023afSMatthew Wilcox (Oracle) int nfs_read_folio(struct file *file, struct folio *folio)
3211da177e4SLinus Torvalds {
322ab75bff1STrond Myklebust 	struct inode *inode = file_inode(file);
32301c3a400SDave Wysochanski 	struct nfs_pageio_descriptor pgio;
32401c3a400SDave Wysochanski 	struct nfs_open_context *ctx;
32549dee700SDave Wysochanski 	int ret;
3261da177e4SLinus Torvalds 
327ab75bff1STrond Myklebust 	trace_nfs_aop_readpage(inode, folio);
32891d5b470SChuck Lever 	nfs_inc_stats(inode, NFSIOS_VFSREADPAGE);
3299c88ea00SDave Wysochanski 	task_io_account_read(folio_size(folio));
33091d5b470SChuck Lever 
3311da177e4SLinus Torvalds 	/*
3321da177e4SLinus Torvalds 	 * Try to flush any pending writes to the file..
3331da177e4SLinus Torvalds 	 *
334ab75bff1STrond Myklebust 	 * NOTE! Because we own the folio lock, there cannot
3351da177e4SLinus Torvalds 	 * be any new pending writes generated at this point
336ab75bff1STrond Myklebust 	 * for this folio (other folios can be written to).
3371da177e4SLinus Torvalds 	 */
338ab75bff1STrond Myklebust 	ret = nfs_wb_folio(inode, folio);
33949dee700SDave Wysochanski 	if (ret)
340de05a0ccSTrond Myklebust 		goto out_unlock;
341ab75bff1STrond Myklebust 	if (folio_test_uptodate(folio))
342de05a0ccSTrond Myklebust 		goto out_unlock;
3431da177e4SLinus Torvalds 
34449dee700SDave Wysochanski 	ret = -ESTALE;
3455f004cf2STrond Myklebust 	if (NFS_STALE(inode))
346de05a0ccSTrond Myklebust 		goto out_unlock;
3475f004cf2STrond Myklebust 
348*000dbe0bSDave Wysochanski 	ret = nfs_netfs_read_folio(file, folio);
349*000dbe0bSDave Wysochanski 	if (!ret)
350*000dbe0bSDave Wysochanski 		goto out;
351*000dbe0bSDave Wysochanski 
35201c3a400SDave Wysochanski 	ctx = get_nfs_open_context(nfs_file_open_context(file));
3531da177e4SLinus Torvalds 
35401c3a400SDave Wysochanski 	xchg(&ctx->error, 0);
35501c3a400SDave Wysochanski 	nfs_pageio_init_read(&pgio, inode, false,
3561e83b173SDave Wysochanski 			     &nfs_async_read_completion_ops);
3571e83b173SDave Wysochanski 
35801c3a400SDave Wysochanski 	ret = nfs_read_add_folio(&pgio, ctx, folio);
359e0340f16SDave Wysochanski 	if (ret)
360*000dbe0bSDave Wysochanski 		goto out_put;
3611e83b173SDave Wysochanski 
36201c3a400SDave Wysochanski 	nfs_pageio_complete_read(&pgio);
36301c3a400SDave Wysochanski 	ret = pgio.pg_error < 0 ? pgio.pg_error : 0;
36449dee700SDave Wysochanski 	if (!ret) {
365ab75bff1STrond Myklebust 		ret = folio_wait_locked_killable(folio);
366ab75bff1STrond Myklebust 		if (!folio_test_uptodate(folio) && !ret)
36701c3a400SDave Wysochanski 			ret = xchg(&ctx->error, 0);
3688f54c7a4STrond Myklebust 	}
369*000dbe0bSDave Wysochanski out_put:
37001c3a400SDave Wysochanski 	put_nfs_open_context(ctx);
371*000dbe0bSDave Wysochanski out:
372ab75bff1STrond Myklebust 	trace_nfs_aop_readpage_done(inode, folio, ret);
37349dee700SDave Wysochanski 	return ret;
374de05a0ccSTrond Myklebust out_unlock:
375ab75bff1STrond Myklebust 	folio_unlock(folio);
376*000dbe0bSDave Wysochanski 	goto out;
3771da177e4SLinus Torvalds }
3781da177e4SLinus Torvalds 
3798786fde8SMatthew Wilcox (Oracle) void nfs_readahead(struct readahead_control *ractl)
3801da177e4SLinus Torvalds {
38101c3a400SDave Wysochanski 	struct nfs_pageio_descriptor pgio;
38201c3a400SDave Wysochanski 	struct nfs_open_context *ctx;
3838786fde8SMatthew Wilcox (Oracle) 	unsigned int nr_pages = readahead_count(ractl);
3848786fde8SMatthew Wilcox (Oracle) 	struct file *file = ractl->file;
3858786fde8SMatthew Wilcox (Oracle) 	struct inode *inode = ractl->mapping->host;
386ab75bff1STrond Myklebust 	struct folio *folio;
38749dee700SDave Wysochanski 	int ret;
3881da177e4SLinus Torvalds 
3898786fde8SMatthew Wilcox (Oracle) 	trace_nfs_aop_readahead(inode, readahead_pos(ractl), nr_pages);
39091d5b470SChuck Lever 	nfs_inc_stats(inode, NFSIOS_VFSREADPAGES);
3919c88ea00SDave Wysochanski 	task_io_account_read(readahead_length(ractl));
3921da177e4SLinus Torvalds 
39349dee700SDave Wysochanski 	ret = -ESTALE;
3945f004cf2STrond Myklebust 	if (NFS_STALE(inode))
3955f004cf2STrond Myklebust 		goto out;
3965f004cf2STrond Myklebust 
397*000dbe0bSDave Wysochanski 	ret = nfs_netfs_readahead(ractl);
398*000dbe0bSDave Wysochanski 	if (!ret)
399*000dbe0bSDave Wysochanski 		goto out;
400*000dbe0bSDave Wysochanski 
40149dee700SDave Wysochanski 	if (file == NULL) {
40249dee700SDave Wysochanski 		ret = -EBADF;
40301c3a400SDave Wysochanski 		ctx = nfs_find_open_context(inode, NULL, FMODE_READ);
40401c3a400SDave Wysochanski 		if (ctx == NULL)
40549dee700SDave Wysochanski 			goto out;
4061da177e4SLinus Torvalds 	} else
40701c3a400SDave Wysochanski 		ctx = get_nfs_open_context(nfs_file_open_context(file));
4089a9fc1c0SDavid Howells 
40901c3a400SDave Wysochanski 	nfs_pageio_init_read(&pgio, inode, false,
410fab5fc25SChristoph Hellwig 			     &nfs_async_read_completion_ops);
4118b09bee3STrond Myklebust 
412ab75bff1STrond Myklebust 	while ((folio = readahead_folio(ractl)) != NULL) {
41301c3a400SDave Wysochanski 		ret = nfs_read_add_folio(&pgio, ctx, folio);
4148786fde8SMatthew Wilcox (Oracle) 		if (ret)
4158786fde8SMatthew Wilcox (Oracle) 			break;
4168786fde8SMatthew Wilcox (Oracle) 	}
417a7d42ddbSWeston Andros Adamson 
41801c3a400SDave Wysochanski 	nfs_pageio_complete_read(&pgio);
419a7d42ddbSWeston Andros Adamson 
42001c3a400SDave Wysochanski 	put_nfs_open_context(ctx);
4215f004cf2STrond Myklebust out:
422d9f87743SChuck Lever 	trace_nfs_aop_readahead_done(inode, nr_pages, ret);
4231da177e4SLinus Torvalds }
4241da177e4SLinus Torvalds 
425f7b422b1SDavid Howells int __init nfs_init_readpagecache(void)
4261da177e4SLinus Torvalds {
4271da177e4SLinus Torvalds 	nfs_rdata_cachep = kmem_cache_create("nfs_read_data",
4281e7f3a48SWeston Andros Adamson 					     sizeof(struct nfs_pgio_header),
4291da177e4SLinus Torvalds 					     0, SLAB_HWCACHE_ALIGN,
43020c2df83SPaul Mundt 					     NULL);
4311da177e4SLinus Torvalds 	if (nfs_rdata_cachep == NULL)
4321da177e4SLinus Torvalds 		return -ENOMEM;
4331da177e4SLinus Torvalds 
4341da177e4SLinus Torvalds 	return 0;
4351da177e4SLinus Torvalds }
4361da177e4SLinus Torvalds 
437266bee88SDavid Brownell void nfs_destroy_readpagecache(void)
4381da177e4SLinus Torvalds {
4391a1d92c1SAlexey Dobriyan 	kmem_cache_destroy(nfs_rdata_cachep);
4401da177e4SLinus Torvalds }
4414a0de55cSAnna Schumaker 
4424a0de55cSAnna Schumaker static const struct nfs_rw_ops nfs_rw_read_ops = {
4434a0de55cSAnna Schumaker 	.rw_alloc_header	= nfs_readhdr_alloc,
4444a0de55cSAnna Schumaker 	.rw_free_header		= nfs_readhdr_free,
4450eecb214SAnna Schumaker 	.rw_done		= nfs_readpage_done,
4460eecb214SAnna Schumaker 	.rw_result		= nfs_readpage_result,
4471ed26f33SAnna Schumaker 	.rw_initiate		= nfs_initiate_read,
4484a0de55cSAnna Schumaker };
449