1 /* NFS filesystem cache interface 2 * 3 * Copyright (C) 2008 Red Hat, Inc. All Rights Reserved. 4 * Written by David Howells (dhowells@redhat.com) 5 * 6 * This program is free software; you can redistribute it and/or 7 * modify it under the terms of the GNU General Public Licence 8 * as published by the Free Software Foundation; either version 9 * 2 of the Licence, or (at your option) any later version. 10 */ 11 12 #include <linux/init.h> 13 #include <linux/kernel.h> 14 #include <linux/sched.h> 15 #include <linux/mm.h> 16 #include <linux/nfs_fs.h> 17 #include <linux/nfs_fs_sb.h> 18 #include <linux/in6.h> 19 #include <linux/seq_file.h> 20 #include <linux/slab.h> 21 22 #include "internal.h" 23 #include "iostat.h" 24 #include "fscache.h" 25 26 #define NFSDBG_FACILITY NFSDBG_FSCACHE 27 28 static struct rb_root nfs_fscache_keys = RB_ROOT; 29 static DEFINE_SPINLOCK(nfs_fscache_keys_lock); 30 31 /* 32 * Get the per-client index cookie for an NFS client if the appropriate mount 33 * flag was set 34 * - We always try and get an index cookie for the client, but get filehandle 35 * cookies on a per-superblock basis, depending on the mount flags 36 */ 37 void nfs_fscache_get_client_cookie(struct nfs_client *clp) 38 { 39 /* create a cache index for looking up filehandles */ 40 clp->fscache = fscache_acquire_cookie(nfs_fscache_netfs.primary_index, 41 &nfs_fscache_server_index_def, 42 clp, true); 43 dfprintk(FSCACHE, "NFS: get client cookie (0x%p/0x%p)\n", 44 clp, clp->fscache); 45 } 46 47 /* 48 * Dispose of a per-client cookie 49 */ 50 void nfs_fscache_release_client_cookie(struct nfs_client *clp) 51 { 52 dfprintk(FSCACHE, "NFS: releasing client cookie (0x%p/0x%p)\n", 53 clp, clp->fscache); 54 55 fscache_relinquish_cookie(clp->fscache, 0); 56 clp->fscache = NULL; 57 } 58 59 /* 60 * Get the cache cookie for an NFS superblock. We have to handle 61 * uniquification here because the cache doesn't do it for us. 62 * 63 * The default uniquifier is just an empty string, but it may be overridden 64 * either by the 'fsc=xxx' option to mount, or by inheriting it from the parent 65 * superblock across an automount point of some nature. 66 */ 67 void nfs_fscache_get_super_cookie(struct super_block *sb, const char *uniq, int ulen) 68 { 69 struct nfs_fscache_key *key, *xkey; 70 struct nfs_server *nfss = NFS_SB(sb); 71 struct rb_node **p, *parent; 72 int diff; 73 74 if (!uniq) { 75 uniq = ""; 76 ulen = 1; 77 } 78 79 key = kzalloc(sizeof(*key) + ulen, GFP_KERNEL); 80 if (!key) 81 return; 82 83 key->nfs_client = nfss->nfs_client; 84 key->key.super.s_flags = sb->s_flags & NFS_MS_MASK; 85 key->key.nfs_server.flags = nfss->flags; 86 key->key.nfs_server.rsize = nfss->rsize; 87 key->key.nfs_server.wsize = nfss->wsize; 88 key->key.nfs_server.acregmin = nfss->acregmin; 89 key->key.nfs_server.acregmax = nfss->acregmax; 90 key->key.nfs_server.acdirmin = nfss->acdirmin; 91 key->key.nfs_server.acdirmax = nfss->acdirmax; 92 key->key.nfs_server.fsid = nfss->fsid; 93 key->key.rpc_auth.au_flavor = nfss->client->cl_auth->au_flavor; 94 95 key->key.uniq_len = ulen; 96 memcpy(key->key.uniquifier, uniq, ulen); 97 98 spin_lock(&nfs_fscache_keys_lock); 99 p = &nfs_fscache_keys.rb_node; 100 parent = NULL; 101 while (*p) { 102 parent = *p; 103 xkey = rb_entry(parent, struct nfs_fscache_key, node); 104 105 if (key->nfs_client < xkey->nfs_client) 106 goto go_left; 107 if (key->nfs_client > xkey->nfs_client) 108 goto go_right; 109 110 diff = memcmp(&key->key, &xkey->key, sizeof(key->key)); 111 if (diff < 0) 112 goto go_left; 113 if (diff > 0) 114 goto go_right; 115 116 if (key->key.uniq_len == 0) 117 goto non_unique; 118 diff = memcmp(key->key.uniquifier, 119 xkey->key.uniquifier, 120 key->key.uniq_len); 121 if (diff < 0) 122 goto go_left; 123 if (diff > 0) 124 goto go_right; 125 goto non_unique; 126 127 go_left: 128 p = &(*p)->rb_left; 129 continue; 130 go_right: 131 p = &(*p)->rb_right; 132 } 133 134 rb_link_node(&key->node, parent, p); 135 rb_insert_color(&key->node, &nfs_fscache_keys); 136 spin_unlock(&nfs_fscache_keys_lock); 137 nfss->fscache_key = key; 138 139 /* create a cache index for looking up filehandles */ 140 nfss->fscache = fscache_acquire_cookie(nfss->nfs_client->fscache, 141 &nfs_fscache_super_index_def, 142 nfss, true); 143 dfprintk(FSCACHE, "NFS: get superblock cookie (0x%p/0x%p)\n", 144 nfss, nfss->fscache); 145 return; 146 147 non_unique: 148 spin_unlock(&nfs_fscache_keys_lock); 149 kfree(key); 150 nfss->fscache_key = NULL; 151 nfss->fscache = NULL; 152 printk(KERN_WARNING "NFS:" 153 " Cache request denied due to non-unique superblock keys\n"); 154 } 155 156 /* 157 * release a per-superblock cookie 158 */ 159 void nfs_fscache_release_super_cookie(struct super_block *sb) 160 { 161 struct nfs_server *nfss = NFS_SB(sb); 162 163 dfprintk(FSCACHE, "NFS: releasing superblock cookie (0x%p/0x%p)\n", 164 nfss, nfss->fscache); 165 166 fscache_relinquish_cookie(nfss->fscache, 0); 167 nfss->fscache = NULL; 168 169 if (nfss->fscache_key) { 170 spin_lock(&nfs_fscache_keys_lock); 171 rb_erase(&nfss->fscache_key->node, &nfs_fscache_keys); 172 spin_unlock(&nfs_fscache_keys_lock); 173 kfree(nfss->fscache_key); 174 nfss->fscache_key = NULL; 175 } 176 } 177 178 /* 179 * Initialise the per-inode cache cookie pointer for an NFS inode. 180 */ 181 void nfs_fscache_init_inode(struct inode *inode) 182 { 183 struct nfs_inode *nfsi = NFS_I(inode); 184 185 nfsi->fscache = NULL; 186 if (!S_ISREG(inode->i_mode)) 187 return; 188 nfsi->fscache = fscache_acquire_cookie(NFS_SB(inode->i_sb)->fscache, 189 &nfs_fscache_inode_object_def, 190 nfsi, false); 191 } 192 193 /* 194 * Release a per-inode cookie. 195 */ 196 void nfs_fscache_clear_inode(struct inode *inode) 197 { 198 struct nfs_inode *nfsi = NFS_I(inode); 199 struct fscache_cookie *cookie = nfs_i_fscache(inode); 200 201 dfprintk(FSCACHE, "NFS: clear cookie (0x%p/0x%p)\n", nfsi, cookie); 202 203 fscache_relinquish_cookie(cookie, false); 204 nfsi->fscache = NULL; 205 } 206 207 static bool nfs_fscache_can_enable(void *data) 208 { 209 struct inode *inode = data; 210 211 return !inode_is_open_for_write(inode); 212 } 213 214 /* 215 * Enable or disable caching for a file that is being opened as appropriate. 216 * The cookie is allocated when the inode is initialised, but is not enabled at 217 * that time. Enablement is deferred to file-open time to avoid stat() and 218 * access() thrashing the cache. 219 * 220 * For now, with NFS, only regular files that are open read-only will be able 221 * to use the cache. 222 * 223 * We enable the cache for an inode if we open it read-only and it isn't 224 * currently open for writing. We disable the cache if the inode is open 225 * write-only. 226 * 227 * The caller uses the file struct to pin i_writecount on the inode before 228 * calling us when a file is opened for writing, so we can make use of that. 229 * 230 * Note that this may be invoked multiple times in parallel by parallel 231 * nfs_open() functions. 232 */ 233 void nfs_fscache_open_file(struct inode *inode, struct file *filp) 234 { 235 struct nfs_inode *nfsi = NFS_I(inode); 236 struct fscache_cookie *cookie = nfs_i_fscache(inode); 237 238 if (!fscache_cookie_valid(cookie)) 239 return; 240 241 if (inode_is_open_for_write(inode)) { 242 dfprintk(FSCACHE, "NFS: nfsi 0x%p disabling cache\n", nfsi); 243 clear_bit(NFS_INO_FSCACHE, &nfsi->flags); 244 fscache_disable_cookie(cookie, true); 245 fscache_uncache_all_inode_pages(cookie, inode); 246 } else { 247 dfprintk(FSCACHE, "NFS: nfsi 0x%p enabling cache\n", nfsi); 248 fscache_enable_cookie(cookie, nfs_fscache_can_enable, inode); 249 if (fscache_cookie_enabled(cookie)) 250 set_bit(NFS_INO_FSCACHE, &NFS_I(inode)->flags); 251 } 252 } 253 EXPORT_SYMBOL_GPL(nfs_fscache_open_file); 254 255 /* 256 * Release the caching state associated with a page, if the page isn't busy 257 * interacting with the cache. 258 * - Returns true (can release page) or false (page busy). 259 */ 260 int nfs_fscache_release_page(struct page *page, gfp_t gfp) 261 { 262 if (PageFsCache(page)) { 263 struct fscache_cookie *cookie = nfs_i_fscache(page->mapping->host); 264 265 BUG_ON(!cookie); 266 dfprintk(FSCACHE, "NFS: fscache releasepage (0x%p/0x%p/0x%p)\n", 267 cookie, page, NFS_I(page->mapping->host)); 268 269 if (!fscache_maybe_release_page(cookie, page, gfp)) 270 return 0; 271 272 nfs_inc_fscache_stats(page->mapping->host, 273 NFSIOS_FSCACHE_PAGES_UNCACHED); 274 } 275 276 return 1; 277 } 278 279 /* 280 * Release the caching state associated with a page if undergoing complete page 281 * invalidation. 282 */ 283 void __nfs_fscache_invalidate_page(struct page *page, struct inode *inode) 284 { 285 struct fscache_cookie *cookie = nfs_i_fscache(inode); 286 287 BUG_ON(!cookie); 288 289 dfprintk(FSCACHE, "NFS: fscache invalidatepage (0x%p/0x%p/0x%p)\n", 290 cookie, page, NFS_I(inode)); 291 292 fscache_wait_on_page_write(cookie, page); 293 294 BUG_ON(!PageLocked(page)); 295 fscache_uncache_page(cookie, page); 296 nfs_inc_fscache_stats(page->mapping->host, 297 NFSIOS_FSCACHE_PAGES_UNCACHED); 298 } 299 300 /* 301 * Handle completion of a page being read from the cache. 302 * - Called in process (keventd) context. 303 */ 304 static void nfs_readpage_from_fscache_complete(struct page *page, 305 void *context, 306 int error) 307 { 308 dfprintk(FSCACHE, 309 "NFS: readpage_from_fscache_complete (0x%p/0x%p/%d)\n", 310 page, context, error); 311 312 /* if the read completes with an error, we just unlock the page and let 313 * the VM reissue the readpage */ 314 if (!error) { 315 SetPageUptodate(page); 316 unlock_page(page); 317 } else { 318 error = nfs_readpage_async(context, page->mapping->host, page); 319 if (error) 320 unlock_page(page); 321 } 322 } 323 324 /* 325 * Retrieve a page from fscache 326 */ 327 int __nfs_readpage_from_fscache(struct nfs_open_context *ctx, 328 struct inode *inode, struct page *page) 329 { 330 int ret; 331 332 dfprintk(FSCACHE, 333 "NFS: readpage_from_fscache(fsc:%p/p:%p(i:%lx f:%lx)/0x%p)\n", 334 nfs_i_fscache(inode), page, page->index, page->flags, inode); 335 336 ret = fscache_read_or_alloc_page(nfs_i_fscache(inode), 337 page, 338 nfs_readpage_from_fscache_complete, 339 ctx, 340 GFP_KERNEL); 341 342 switch (ret) { 343 case 0: /* read BIO submitted (page in fscache) */ 344 dfprintk(FSCACHE, 345 "NFS: readpage_from_fscache: BIO submitted\n"); 346 nfs_inc_fscache_stats(inode, NFSIOS_FSCACHE_PAGES_READ_OK); 347 return ret; 348 349 case -ENOBUFS: /* inode not in cache */ 350 case -ENODATA: /* page not in cache */ 351 nfs_inc_fscache_stats(inode, NFSIOS_FSCACHE_PAGES_READ_FAIL); 352 dfprintk(FSCACHE, 353 "NFS: readpage_from_fscache %d\n", ret); 354 return 1; 355 356 default: 357 dfprintk(FSCACHE, "NFS: readpage_from_fscache %d\n", ret); 358 nfs_inc_fscache_stats(inode, NFSIOS_FSCACHE_PAGES_READ_FAIL); 359 } 360 return ret; 361 } 362 363 /* 364 * Retrieve a set of pages from fscache 365 */ 366 int __nfs_readpages_from_fscache(struct nfs_open_context *ctx, 367 struct inode *inode, 368 struct address_space *mapping, 369 struct list_head *pages, 370 unsigned *nr_pages) 371 { 372 unsigned npages = *nr_pages; 373 int ret; 374 375 dfprintk(FSCACHE, "NFS: nfs_getpages_from_fscache (0x%p/%u/0x%p)\n", 376 nfs_i_fscache(inode), npages, inode); 377 378 ret = fscache_read_or_alloc_pages(nfs_i_fscache(inode), 379 mapping, pages, nr_pages, 380 nfs_readpage_from_fscache_complete, 381 ctx, 382 mapping_gfp_mask(mapping)); 383 if (*nr_pages < npages) 384 nfs_add_fscache_stats(inode, NFSIOS_FSCACHE_PAGES_READ_OK, 385 npages); 386 if (*nr_pages > 0) 387 nfs_add_fscache_stats(inode, NFSIOS_FSCACHE_PAGES_READ_FAIL, 388 *nr_pages); 389 390 switch (ret) { 391 case 0: /* read submitted to the cache for all pages */ 392 BUG_ON(!list_empty(pages)); 393 BUG_ON(*nr_pages != 0); 394 dfprintk(FSCACHE, 395 "NFS: nfs_getpages_from_fscache: submitted\n"); 396 397 return ret; 398 399 case -ENOBUFS: /* some pages aren't cached and can't be */ 400 case -ENODATA: /* some pages aren't cached */ 401 dfprintk(FSCACHE, 402 "NFS: nfs_getpages_from_fscache: no page: %d\n", ret); 403 return 1; 404 405 default: 406 dfprintk(FSCACHE, 407 "NFS: nfs_getpages_from_fscache: ret %d\n", ret); 408 } 409 410 return ret; 411 } 412 413 /* 414 * Store a newly fetched page in fscache 415 * - PG_fscache must be set on the page 416 */ 417 void __nfs_readpage_to_fscache(struct inode *inode, struct page *page, int sync) 418 { 419 int ret; 420 421 dfprintk(FSCACHE, 422 "NFS: readpage_to_fscache(fsc:%p/p:%p(i:%lx f:%lx)/%d)\n", 423 nfs_i_fscache(inode), page, page->index, page->flags, sync); 424 425 ret = fscache_write_page(nfs_i_fscache(inode), page, GFP_KERNEL); 426 dfprintk(FSCACHE, 427 "NFS: readpage_to_fscache: p:%p(i:%lu f:%lx) ret %d\n", 428 page, page->index, page->flags, ret); 429 430 if (ret != 0) { 431 fscache_uncache_page(nfs_i_fscache(inode), page); 432 nfs_inc_fscache_stats(inode, 433 NFSIOS_FSCACHE_PAGES_WRITTEN_FAIL); 434 nfs_inc_fscache_stats(inode, NFSIOS_FSCACHE_PAGES_UNCACHED); 435 } else { 436 nfs_inc_fscache_stats(inode, 437 NFSIOS_FSCACHE_PAGES_WRITTEN_OK); 438 } 439 } 440