xref: /openbmc/linux/fs/fscache/cache.c (revision f3dfffb3)
1 // SPDX-License-Identifier: GPL-2.0-or-later
2 /* FS-Cache cache handling
3  *
4  * Copyright (C) 2021 Red Hat, Inc. All Rights Reserved.
5  * Written by David Howells (dhowells@redhat.com)
6  */
7 
8 #define FSCACHE_DEBUG_LEVEL CACHE
9 #include <linux/export.h>
10 #include <linux/slab.h>
11 #include "internal.h"
12 
13 static LIST_HEAD(fscache_caches);
14 DECLARE_RWSEM(fscache_addremove_sem);
15 EXPORT_SYMBOL(fscache_addremove_sem);
16 DECLARE_WAIT_QUEUE_HEAD(fscache_clearance_waiters);
17 EXPORT_SYMBOL(fscache_clearance_waiters);
18 
19 static atomic_t fscache_cache_debug_id;
20 
21 /*
22  * Allocate a cache cookie.
23  */
24 static struct fscache_cache *fscache_alloc_cache(const char *name)
25 {
26 	struct fscache_cache *cache;
27 
28 	cache = kzalloc(sizeof(*cache), GFP_KERNEL);
29 	if (cache) {
30 		if (name) {
31 			cache->name = kstrdup(name, GFP_KERNEL);
32 			if (!cache->name) {
33 				kfree(cache);
34 				return NULL;
35 			}
36 		}
37 		refcount_set(&cache->ref, 1);
38 		INIT_LIST_HEAD(&cache->cache_link);
39 		cache->debug_id = atomic_inc_return(&fscache_cache_debug_id);
40 	}
41 	return cache;
42 }
43 
44 static bool fscache_get_cache_maybe(struct fscache_cache *cache,
45 				    enum fscache_cache_trace where)
46 {
47 	bool success;
48 	int ref;
49 
50 	success = __refcount_inc_not_zero(&cache->ref, &ref);
51 	if (success)
52 		trace_fscache_cache(cache->debug_id, ref + 1, where);
53 	return success;
54 }
55 
56 /*
57  * Look up a cache cookie.
58  */
59 struct fscache_cache *fscache_lookup_cache(const char *name, bool is_cache)
60 {
61 	struct fscache_cache *candidate, *cache, *unnamed = NULL;
62 
63 	/* firstly check for the existence of the cache under read lock */
64 	down_read(&fscache_addremove_sem);
65 
66 	list_for_each_entry(cache, &fscache_caches, cache_link) {
67 		if (cache->name && name && strcmp(cache->name, name) == 0 &&
68 		    fscache_get_cache_maybe(cache, fscache_cache_get_acquire))
69 			goto got_cache_r;
70 		if (!cache->name && !name &&
71 		    fscache_get_cache_maybe(cache, fscache_cache_get_acquire))
72 			goto got_cache_r;
73 	}
74 
75 	if (!name) {
76 		list_for_each_entry(cache, &fscache_caches, cache_link) {
77 			if (cache->name &&
78 			    fscache_get_cache_maybe(cache, fscache_cache_get_acquire))
79 				goto got_cache_r;
80 		}
81 	}
82 
83 	up_read(&fscache_addremove_sem);
84 
85 	/* the cache does not exist - create a candidate */
86 	candidate = fscache_alloc_cache(name);
87 	if (!candidate)
88 		return ERR_PTR(-ENOMEM);
89 
90 	/* write lock, search again and add if still not present */
91 	down_write(&fscache_addremove_sem);
92 
93 	list_for_each_entry(cache, &fscache_caches, cache_link) {
94 		if (cache->name && name && strcmp(cache->name, name) == 0 &&
95 		    fscache_get_cache_maybe(cache, fscache_cache_get_acquire))
96 			goto got_cache_w;
97 		if (!cache->name) {
98 			unnamed = cache;
99 			if (!name &&
100 			    fscache_get_cache_maybe(cache, fscache_cache_get_acquire))
101 				goto got_cache_w;
102 		}
103 	}
104 
105 	if (unnamed && is_cache &&
106 	    fscache_get_cache_maybe(unnamed, fscache_cache_get_acquire))
107 		goto use_unnamed_cache;
108 
109 	if (!name) {
110 		list_for_each_entry(cache, &fscache_caches, cache_link) {
111 			if (cache->name &&
112 			    fscache_get_cache_maybe(cache, fscache_cache_get_acquire))
113 				goto got_cache_w;
114 		}
115 	}
116 
117 	list_add_tail(&candidate->cache_link, &fscache_caches);
118 	trace_fscache_cache(candidate->debug_id,
119 			    refcount_read(&candidate->ref),
120 			    fscache_cache_new_acquire);
121 	up_write(&fscache_addremove_sem);
122 	return candidate;
123 
124 got_cache_r:
125 	up_read(&fscache_addremove_sem);
126 	return cache;
127 use_unnamed_cache:
128 	cache = unnamed;
129 	cache->name = candidate->name;
130 	candidate->name = NULL;
131 got_cache_w:
132 	up_write(&fscache_addremove_sem);
133 	kfree(candidate->name);
134 	kfree(candidate);
135 	return cache;
136 }
137 
138 /**
139  * fscache_acquire_cache - Acquire a cache-level cookie.
140  * @name: The name of the cache.
141  *
142  * Get a cookie to represent an actual cache.  If a name is given and there is
143  * a nameless cache record available, this will acquire that and set its name,
144  * directing all the volumes using it to this cache.
145  *
146  * The cache will be switched over to the preparing state if not currently in
147  * use, otherwise -EBUSY will be returned.
148  */
149 struct fscache_cache *fscache_acquire_cache(const char *name)
150 {
151 	struct fscache_cache *cache;
152 
153 	ASSERT(name);
154 	cache = fscache_lookup_cache(name, true);
155 	if (IS_ERR(cache))
156 		return cache;
157 
158 	if (!fscache_set_cache_state_maybe(cache,
159 					   FSCACHE_CACHE_IS_NOT_PRESENT,
160 					   FSCACHE_CACHE_IS_PREPARING)) {
161 		pr_warn("Cache tag %s in use\n", name);
162 		fscache_put_cache(cache, fscache_cache_put_cache);
163 		return ERR_PTR(-EBUSY);
164 	}
165 
166 	return cache;
167 }
168 EXPORT_SYMBOL(fscache_acquire_cache);
169 
170 /**
171  * fscache_put_cache - Release a cache-level cookie.
172  * @cache: The cache cookie to be released
173  * @where: An indication of where the release happened
174  *
175  * Release the caller's reference on a cache-level cookie.  The @where
176  * indication should give information about the circumstances in which the call
177  * occurs and will be logged through a tracepoint.
178  */
179 void fscache_put_cache(struct fscache_cache *cache,
180 		       enum fscache_cache_trace where)
181 {
182 	unsigned int debug_id;
183 	bool zero;
184 	int ref;
185 
186 	if (IS_ERR_OR_NULL(cache))
187 		return;
188 
189 	debug_id = cache->debug_id;
190 	zero = __refcount_dec_and_test(&cache->ref, &ref);
191 	trace_fscache_cache(debug_id, ref - 1, where);
192 
193 	if (zero) {
194 		down_write(&fscache_addremove_sem);
195 		list_del_init(&cache->cache_link);
196 		up_write(&fscache_addremove_sem);
197 		kfree(cache->name);
198 		kfree(cache);
199 	}
200 }
201 
202 /**
203  * fscache_relinquish_cache - Reset cache state and release cookie
204  * @cache: The cache cookie to be released
205  *
206  * Reset the state of a cache and release the caller's reference on a cache
207  * cookie.
208  */
209 void fscache_relinquish_cache(struct fscache_cache *cache)
210 {
211 	enum fscache_cache_trace where =
212 		(cache->state == FSCACHE_CACHE_IS_PREPARING) ?
213 		fscache_cache_put_prep_failed :
214 		fscache_cache_put_relinquish;
215 
216 	cache->ops = NULL;
217 	cache->cache_priv = NULL;
218 	fscache_set_cache_state(cache, FSCACHE_CACHE_IS_NOT_PRESENT);
219 	fscache_put_cache(cache, where);
220 }
221 EXPORT_SYMBOL(fscache_relinquish_cache);
222 
223 /**
224  * fscache_add_cache - Declare a cache as being open for business
225  * @cache: The cache-level cookie representing the cache
226  * @ops: Table of cache operations to use
227  * @cache_priv: Private data for the cache record
228  *
229  * Add a cache to the system, making it available for netfs's to use.
230  *
231  * See Documentation/filesystems/caching/backend-api.rst for a complete
232  * description.
233  */
234 int fscache_add_cache(struct fscache_cache *cache,
235 		      const struct fscache_cache_ops *ops,
236 		      void *cache_priv)
237 {
238 	int n_accesses;
239 
240 	_enter("{%s,%s}", ops->name, cache->name);
241 
242 	BUG_ON(fscache_cache_state(cache) != FSCACHE_CACHE_IS_PREPARING);
243 
244 	/* Get a ref on the cache cookie and keep its n_accesses counter raised
245 	 * by 1 to prevent wakeups from transitioning it to 0 until we're
246 	 * withdrawing caching services from it.
247 	 */
248 	n_accesses = atomic_inc_return(&cache->n_accesses);
249 	trace_fscache_access_cache(cache->debug_id, refcount_read(&cache->ref),
250 				   n_accesses, fscache_access_cache_pin);
251 
252 	down_write(&fscache_addremove_sem);
253 
254 	cache->ops = ops;
255 	cache->cache_priv = cache_priv;
256 	fscache_set_cache_state(cache, FSCACHE_CACHE_IS_ACTIVE);
257 
258 	up_write(&fscache_addremove_sem);
259 	pr_notice("Cache \"%s\" added (type %s)\n", cache->name, ops->name);
260 	_leave(" = 0 [%s]", cache->name);
261 	return 0;
262 }
263 EXPORT_SYMBOL(fscache_add_cache);
264 
265 /**
266  * fscache_begin_cache_access - Pin a cache so it can be accessed
267  * @cache: The cache-level cookie
268  * @why: An indication of the circumstances of the access for tracing
269  *
270  * Attempt to pin the cache to prevent it from going away whilst we're
271  * accessing it and returns true if successful.  This works as follows:
272  *
273  *  (1) If the cache tests as not live (state is not FSCACHE_CACHE_IS_ACTIVE),
274  *      then we return false to indicate access was not permitted.
275  *
276  *  (2) If the cache tests as live, then we increment the n_accesses count and
277  *      then recheck the liveness, ending the access if it ceased to be live.
278  *
279  *  (3) When we end the access, we decrement n_accesses and wake up the any
280  *      waiters if it reaches 0.
281  *
282  *  (4) Whilst the cache is caching, n_accesses is kept artificially
283  *      incremented to prevent wakeups from happening.
284  *
285  *  (5) When the cache is taken offline, the state is changed to prevent new
286  *      accesses, n_accesses is decremented and we wait for n_accesses to
287  *      become 0.
288  */
289 bool fscache_begin_cache_access(struct fscache_cache *cache, enum fscache_access_trace why)
290 {
291 	int n_accesses;
292 
293 	if (!fscache_cache_is_live(cache))
294 		return false;
295 
296 	n_accesses = atomic_inc_return(&cache->n_accesses);
297 	smp_mb__after_atomic(); /* Reread live flag after n_accesses */
298 	trace_fscache_access_cache(cache->debug_id, refcount_read(&cache->ref),
299 				   n_accesses, why);
300 	if (!fscache_cache_is_live(cache)) {
301 		fscache_end_cache_access(cache, fscache_access_unlive);
302 		return false;
303 	}
304 	return true;
305 }
306 
307 /**
308  * fscache_end_cache_access - Unpin a cache at the end of an access.
309  * @cache: The cache-level cookie
310  * @why: An indication of the circumstances of the access for tracing
311  *
312  * Unpin a cache after we've accessed it.  The @why indicator is merely
313  * provided for tracing purposes.
314  */
315 void fscache_end_cache_access(struct fscache_cache *cache, enum fscache_access_trace why)
316 {
317 	int n_accesses;
318 
319 	smp_mb__before_atomic();
320 	n_accesses = atomic_dec_return(&cache->n_accesses);
321 	trace_fscache_access_cache(cache->debug_id, refcount_read(&cache->ref),
322 				   n_accesses, why);
323 	if (n_accesses == 0)
324 		wake_up_var(&cache->n_accesses);
325 }
326 
327 /**
328  * fscache_io_error - Note a cache I/O error
329  * @cache: The record describing the cache
330  *
331  * Note that an I/O error occurred in a cache and that it should no longer be
332  * used for anything.  This also reports the error into the kernel log.
333  *
334  * See Documentation/filesystems/caching/backend-api.rst for a complete
335  * description.
336  */
337 void fscache_io_error(struct fscache_cache *cache)
338 {
339 	if (fscache_set_cache_state_maybe(cache,
340 					  FSCACHE_CACHE_IS_ACTIVE,
341 					  FSCACHE_CACHE_GOT_IOERROR))
342 		pr_err("Cache '%s' stopped due to I/O error\n",
343 		       cache->name);
344 }
345 EXPORT_SYMBOL(fscache_io_error);
346 
347 /**
348  * fscache_withdraw_cache - Withdraw a cache from the active service
349  * @cache: The cache cookie
350  *
351  * Begin the process of withdrawing a cache from service.  This stops new
352  * cache-level and volume-level accesses from taking place and waits for
353  * currently ongoing cache-level accesses to end.
354  */
355 void fscache_withdraw_cache(struct fscache_cache *cache)
356 {
357 	int n_accesses;
358 
359 	pr_notice("Withdrawing cache \"%s\" (%u objs)\n",
360 		  cache->name, atomic_read(&cache->object_count));
361 
362 	fscache_set_cache_state(cache, FSCACHE_CACHE_IS_WITHDRAWN);
363 
364 	/* Allow wakeups on dec-to-0 */
365 	n_accesses = atomic_dec_return(&cache->n_accesses);
366 	trace_fscache_access_cache(cache->debug_id, refcount_read(&cache->ref),
367 				   n_accesses, fscache_access_cache_unpin);
368 
369 	wait_var_event(&cache->n_accesses,
370 		       atomic_read(&cache->n_accesses) == 0);
371 }
372 EXPORT_SYMBOL(fscache_withdraw_cache);
373 
374 #ifdef CONFIG_PROC_FS
375 static const char fscache_cache_states[NR__FSCACHE_CACHE_STATE] = "-PAEW";
376 
377 /*
378  * Generate a list of caches in /proc/fs/fscache/caches
379  */
380 static int fscache_caches_seq_show(struct seq_file *m, void *v)
381 {
382 	struct fscache_cache *cache;
383 
384 	if (v == &fscache_caches) {
385 		seq_puts(m,
386 			 "CACHE    REF   VOLS  OBJS  ACCES S NAME\n"
387 			 "======== ===== ===== ===== ===== = ===============\n"
388 			 );
389 		return 0;
390 	}
391 
392 	cache = list_entry(v, struct fscache_cache, cache_link);
393 	seq_printf(m,
394 		   "%08x %5d %5d %5d %5d %c %s\n",
395 		   cache->debug_id,
396 		   refcount_read(&cache->ref),
397 		   atomic_read(&cache->n_volumes),
398 		   atomic_read(&cache->object_count),
399 		   atomic_read(&cache->n_accesses),
400 		   fscache_cache_states[cache->state],
401 		   cache->name ?: "-");
402 	return 0;
403 }
404 
405 static void *fscache_caches_seq_start(struct seq_file *m, loff_t *_pos)
406 	__acquires(fscache_addremove_sem)
407 {
408 	down_read(&fscache_addremove_sem);
409 	return seq_list_start_head(&fscache_caches, *_pos);
410 }
411 
412 static void *fscache_caches_seq_next(struct seq_file *m, void *v, loff_t *_pos)
413 {
414 	return seq_list_next(v, &fscache_caches, _pos);
415 }
416 
417 static void fscache_caches_seq_stop(struct seq_file *m, void *v)
418 	__releases(fscache_addremove_sem)
419 {
420 	up_read(&fscache_addremove_sem);
421 }
422 
423 const struct seq_operations fscache_caches_seq_ops = {
424 	.start  = fscache_caches_seq_start,
425 	.next   = fscache_caches_seq_next,
426 	.stop   = fscache_caches_seq_stop,
427 	.show   = fscache_caches_seq_show,
428 };
429 #endif /* CONFIG_PROC_FS */
430