xref: /openbmc/linux/fs/afs/server.c (revision cb1aaebe)
1 /* AFS server record management
2  *
3  * Copyright (C) 2002, 2007 Red Hat, Inc. All Rights Reserved.
4  * Written by David Howells (dhowells@redhat.com)
5  *
6  * This program is free software; you can redistribute it and/or
7  * modify it under the terms of the GNU General Public License
8  * as published by the Free Software Foundation; either version
9  * 2 of the License, or (at your option) any later version.
10  */
11 
12 #include <linux/sched.h>
13 #include <linux/slab.h>
14 #include "afs_fs.h"
15 #include "internal.h"
16 #include "protocol_yfs.h"
17 
18 static unsigned afs_server_gc_delay = 10;	/* Server record timeout in seconds */
19 static unsigned afs_server_update_delay = 30;	/* Time till VLDB recheck in secs */
20 
21 static void afs_inc_servers_outstanding(struct afs_net *net)
22 {
23 	atomic_inc(&net->servers_outstanding);
24 }
25 
26 static void afs_dec_servers_outstanding(struct afs_net *net)
27 {
28 	if (atomic_dec_and_test(&net->servers_outstanding))
29 		wake_up_var(&net->servers_outstanding);
30 }
31 
32 /*
33  * Find a server by one of its addresses.
34  */
35 struct afs_server *afs_find_server(struct afs_net *net,
36 				   const struct sockaddr_rxrpc *srx)
37 {
38 	const struct sockaddr_in6 *a = &srx->transport.sin6, *b;
39 	const struct afs_addr_list *alist;
40 	struct afs_server *server = NULL;
41 	unsigned int i;
42 	bool ipv6 = true;
43 	int seq = 0, diff;
44 
45 	if (srx->transport.sin6.sin6_addr.s6_addr32[0] == 0 ||
46 	    srx->transport.sin6.sin6_addr.s6_addr32[1] == 0 ||
47 	    srx->transport.sin6.sin6_addr.s6_addr32[2] == htonl(0xffff))
48 		ipv6 = false;
49 
50 	rcu_read_lock();
51 
52 	do {
53 		if (server)
54 			afs_put_server(net, server);
55 		server = NULL;
56 		read_seqbegin_or_lock(&net->fs_addr_lock, &seq);
57 
58 		if (ipv6) {
59 			hlist_for_each_entry_rcu(server, &net->fs_addresses6, addr6_link) {
60 				alist = rcu_dereference(server->addresses);
61 				for (i = alist->nr_ipv4; i < alist->nr_addrs; i++) {
62 					b = &alist->addrs[i].transport.sin6;
63 					diff = ((u16 __force)a->sin6_port -
64 						(u16 __force)b->sin6_port);
65 					if (diff == 0)
66 						diff = memcmp(&a->sin6_addr,
67 							      &b->sin6_addr,
68 							      sizeof(struct in6_addr));
69 					if (diff == 0)
70 						goto found;
71 				}
72 			}
73 		} else {
74 			hlist_for_each_entry_rcu(server, &net->fs_addresses4, addr4_link) {
75 				alist = rcu_dereference(server->addresses);
76 				for (i = 0; i < alist->nr_ipv4; i++) {
77 					b = &alist->addrs[i].transport.sin6;
78 					diff = ((u16 __force)a->sin6_port -
79 						(u16 __force)b->sin6_port);
80 					if (diff == 0)
81 						diff = ((u32 __force)a->sin6_addr.s6_addr32[3] -
82 							(u32 __force)b->sin6_addr.s6_addr32[3]);
83 					if (diff == 0)
84 						goto found;
85 				}
86 			}
87 		}
88 
89 		server = NULL;
90 	found:
91 		if (server && !atomic_inc_not_zero(&server->usage))
92 			server = NULL;
93 
94 	} while (need_seqretry(&net->fs_addr_lock, seq));
95 
96 	done_seqretry(&net->fs_addr_lock, seq);
97 
98 	rcu_read_unlock();
99 	return server;
100 }
101 
102 /*
103  * Look up a server by its UUID
104  */
105 struct afs_server *afs_find_server_by_uuid(struct afs_net *net, const uuid_t *uuid)
106 {
107 	struct afs_server *server = NULL;
108 	struct rb_node *p;
109 	int diff, seq = 0;
110 
111 	_enter("%pU", uuid);
112 
113 	do {
114 		/* Unfortunately, rbtree walking doesn't give reliable results
115 		 * under just the RCU read lock, so we have to check for
116 		 * changes.
117 		 */
118 		if (server)
119 			afs_put_server(net, server);
120 		server = NULL;
121 
122 		read_seqbegin_or_lock(&net->fs_lock, &seq);
123 
124 		p = net->fs_servers.rb_node;
125 		while (p) {
126 			server = rb_entry(p, struct afs_server, uuid_rb);
127 
128 			diff = memcmp(uuid, &server->uuid, sizeof(*uuid));
129 			if (diff < 0) {
130 				p = p->rb_left;
131 			} else if (diff > 0) {
132 				p = p->rb_right;
133 			} else {
134 				afs_get_server(server);
135 				break;
136 			}
137 
138 			server = NULL;
139 		}
140 	} while (need_seqretry(&net->fs_lock, seq));
141 
142 	done_seqretry(&net->fs_lock, seq);
143 
144 	_leave(" = %p", server);
145 	return server;
146 }
147 
148 /*
149  * Install a server record in the namespace tree
150  */
151 static struct afs_server *afs_install_server(struct afs_net *net,
152 					     struct afs_server *candidate)
153 {
154 	const struct afs_addr_list *alist;
155 	struct afs_server *server;
156 	struct rb_node **pp, *p;
157 	int ret = -EEXIST, diff;
158 
159 	_enter("%p", candidate);
160 
161 	write_seqlock(&net->fs_lock);
162 
163 	/* Firstly install the server in the UUID lookup tree */
164 	pp = &net->fs_servers.rb_node;
165 	p = NULL;
166 	while (*pp) {
167 		p = *pp;
168 		_debug("- consider %p", p);
169 		server = rb_entry(p, struct afs_server, uuid_rb);
170 		diff = memcmp(&candidate->uuid, &server->uuid, sizeof(uuid_t));
171 		if (diff < 0)
172 			pp = &(*pp)->rb_left;
173 		else if (diff > 0)
174 			pp = &(*pp)->rb_right;
175 		else
176 			goto exists;
177 	}
178 
179 	server = candidate;
180 	rb_link_node(&server->uuid_rb, p, pp);
181 	rb_insert_color(&server->uuid_rb, &net->fs_servers);
182 	hlist_add_head_rcu(&server->proc_link, &net->fs_proc);
183 
184 	write_seqlock(&net->fs_addr_lock);
185 	alist = rcu_dereference_protected(server->addresses,
186 					  lockdep_is_held(&net->fs_addr_lock.lock));
187 
188 	/* Secondly, if the server has any IPv4 and/or IPv6 addresses, install
189 	 * it in the IPv4 and/or IPv6 reverse-map lists.
190 	 *
191 	 * TODO: For speed we want to use something other than a flat list
192 	 * here; even sorting the list in terms of lowest address would help a
193 	 * bit, but anything we might want to do gets messy and memory
194 	 * intensive.
195 	 */
196 	if (alist->nr_ipv4 > 0)
197 		hlist_add_head_rcu(&server->addr4_link, &net->fs_addresses4);
198 	if (alist->nr_addrs > alist->nr_ipv4)
199 		hlist_add_head_rcu(&server->addr6_link, &net->fs_addresses6);
200 
201 	write_sequnlock(&net->fs_addr_lock);
202 	ret = 0;
203 
204 exists:
205 	afs_get_server(server);
206 	write_sequnlock(&net->fs_lock);
207 	return server;
208 }
209 
210 /*
211  * allocate a new server record
212  */
213 static struct afs_server *afs_alloc_server(struct afs_net *net,
214 					   const uuid_t *uuid,
215 					   struct afs_addr_list *alist)
216 {
217 	struct afs_server *server;
218 
219 	_enter("");
220 
221 	server = kzalloc(sizeof(struct afs_server), GFP_KERNEL);
222 	if (!server)
223 		goto enomem;
224 
225 	atomic_set(&server->usage, 1);
226 	RCU_INIT_POINTER(server->addresses, alist);
227 	server->addr_version = alist->version;
228 	server->uuid = *uuid;
229 	server->update_at = ktime_get_real_seconds() + afs_server_update_delay;
230 	rwlock_init(&server->fs_lock);
231 	INIT_HLIST_HEAD(&server->cb_volumes);
232 	rwlock_init(&server->cb_break_lock);
233 	init_waitqueue_head(&server->probe_wq);
234 	spin_lock_init(&server->probe_lock);
235 
236 	afs_inc_servers_outstanding(net);
237 	_leave(" = %p", server);
238 	return server;
239 
240 enomem:
241 	_leave(" = NULL [nomem]");
242 	return NULL;
243 }
244 
245 /*
246  * Look up an address record for a server
247  */
248 static struct afs_addr_list *afs_vl_lookup_addrs(struct afs_cell *cell,
249 						 struct key *key, const uuid_t *uuid)
250 {
251 	struct afs_vl_cursor vc;
252 	struct afs_addr_list *alist = NULL;
253 	int ret;
254 
255 	ret = -ERESTARTSYS;
256 	if (afs_begin_vlserver_operation(&vc, cell, key)) {
257 		while (afs_select_vlserver(&vc)) {
258 			if (test_bit(AFS_VLSERVER_FL_IS_YFS, &vc.server->flags))
259 				alist = afs_yfsvl_get_endpoints(&vc, uuid);
260 			else
261 				alist = afs_vl_get_addrs_u(&vc, uuid);
262 		}
263 
264 		ret = afs_end_vlserver_operation(&vc);
265 	}
266 
267 	return ret < 0 ? ERR_PTR(ret) : alist;
268 }
269 
270 /*
271  * Get or create a fileserver record.
272  */
273 struct afs_server *afs_lookup_server(struct afs_cell *cell, struct key *key,
274 				     const uuid_t *uuid)
275 {
276 	struct afs_addr_list *alist;
277 	struct afs_server *server, *candidate;
278 
279 	_enter("%p,%pU", cell->net, uuid);
280 
281 	server = afs_find_server_by_uuid(cell->net, uuid);
282 	if (server)
283 		return server;
284 
285 	alist = afs_vl_lookup_addrs(cell, key, uuid);
286 	if (IS_ERR(alist))
287 		return ERR_CAST(alist);
288 
289 	candidate = afs_alloc_server(cell->net, uuid, alist);
290 	if (!candidate) {
291 		afs_put_addrlist(alist);
292 		return ERR_PTR(-ENOMEM);
293 	}
294 
295 	server = afs_install_server(cell->net, candidate);
296 	if (server != candidate) {
297 		afs_put_addrlist(alist);
298 		kfree(candidate);
299 	}
300 
301 	_leave(" = %p{%d}", server, atomic_read(&server->usage));
302 	return server;
303 }
304 
305 /*
306  * Set the server timer to fire after a given delay, assuming it's not already
307  * set for an earlier time.
308  */
309 static void afs_set_server_timer(struct afs_net *net, time64_t delay)
310 {
311 	if (net->live) {
312 		afs_inc_servers_outstanding(net);
313 		if (timer_reduce(&net->fs_timer, jiffies + delay * HZ))
314 			afs_dec_servers_outstanding(net);
315 	}
316 }
317 
318 /*
319  * Server management timer.  We have an increment on fs_outstanding that we
320  * need to pass along to the work item.
321  */
322 void afs_servers_timer(struct timer_list *timer)
323 {
324 	struct afs_net *net = container_of(timer, struct afs_net, fs_timer);
325 
326 	_enter("");
327 	if (!queue_work(afs_wq, &net->fs_manager))
328 		afs_dec_servers_outstanding(net);
329 }
330 
331 /*
332  * Release a reference on a server record.
333  */
334 void afs_put_server(struct afs_net *net, struct afs_server *server)
335 {
336 	unsigned int usage;
337 
338 	if (!server)
339 		return;
340 
341 	server->put_time = ktime_get_real_seconds();
342 
343 	usage = atomic_dec_return(&server->usage);
344 
345 	_enter("{%u}", usage);
346 
347 	if (likely(usage > 0))
348 		return;
349 
350 	afs_set_server_timer(net, afs_server_gc_delay);
351 }
352 
353 static void afs_server_rcu(struct rcu_head *rcu)
354 {
355 	struct afs_server *server = container_of(rcu, struct afs_server, rcu);
356 
357 	afs_put_addrlist(rcu_access_pointer(server->addresses));
358 	kfree(server);
359 }
360 
361 /*
362  * destroy a dead server
363  */
364 static void afs_destroy_server(struct afs_net *net, struct afs_server *server)
365 {
366 	struct afs_addr_list *alist = rcu_access_pointer(server->addresses);
367 	struct afs_addr_cursor ac = {
368 		.alist	= alist,
369 		.index	= alist->preferred,
370 		.error	= 0,
371 	};
372 	_enter("%p", server);
373 
374 	if (test_bit(AFS_SERVER_FL_MAY_HAVE_CB, &server->flags))
375 		afs_fs_give_up_all_callbacks(net, server, &ac, NULL);
376 
377 	wait_var_event(&server->probe_outstanding,
378 		       atomic_read(&server->probe_outstanding) == 0);
379 
380 	call_rcu(&server->rcu, afs_server_rcu);
381 	afs_dec_servers_outstanding(net);
382 }
383 
384 /*
385  * Garbage collect any expired servers.
386  */
387 static void afs_gc_servers(struct afs_net *net, struct afs_server *gc_list)
388 {
389 	struct afs_server *server;
390 	bool deleted;
391 	int usage;
392 
393 	while ((server = gc_list)) {
394 		gc_list = server->gc_next;
395 
396 		write_seqlock(&net->fs_lock);
397 		usage = 1;
398 		deleted = atomic_try_cmpxchg(&server->usage, &usage, 0);
399 		if (deleted) {
400 			rb_erase(&server->uuid_rb, &net->fs_servers);
401 			hlist_del_rcu(&server->proc_link);
402 		}
403 		write_sequnlock(&net->fs_lock);
404 
405 		if (deleted) {
406 			write_seqlock(&net->fs_addr_lock);
407 			if (!hlist_unhashed(&server->addr4_link))
408 				hlist_del_rcu(&server->addr4_link);
409 			if (!hlist_unhashed(&server->addr6_link))
410 				hlist_del_rcu(&server->addr6_link);
411 			write_sequnlock(&net->fs_addr_lock);
412 			afs_destroy_server(net, server);
413 		}
414 	}
415 }
416 
417 /*
418  * Manage the records of servers known to be within a network namespace.  This
419  * includes garbage collecting unused servers.
420  *
421  * Note also that we were given an increment on net->servers_outstanding by
422  * whoever queued us that we need to deal with before returning.
423  */
424 void afs_manage_servers(struct work_struct *work)
425 {
426 	struct afs_net *net = container_of(work, struct afs_net, fs_manager);
427 	struct afs_server *gc_list = NULL;
428 	struct rb_node *cursor;
429 	time64_t now = ktime_get_real_seconds(), next_manage = TIME64_MAX;
430 	bool purging = !net->live;
431 
432 	_enter("");
433 
434 	/* Trawl the server list looking for servers that have expired from
435 	 * lack of use.
436 	 */
437 	read_seqlock_excl(&net->fs_lock);
438 
439 	for (cursor = rb_first(&net->fs_servers); cursor; cursor = rb_next(cursor)) {
440 		struct afs_server *server =
441 			rb_entry(cursor, struct afs_server, uuid_rb);
442 		int usage = atomic_read(&server->usage);
443 
444 		_debug("manage %pU %u", &server->uuid, usage);
445 
446 		ASSERTCMP(usage, >=, 1);
447 		ASSERTIFCMP(purging, usage, ==, 1);
448 
449 		if (usage == 1) {
450 			time64_t expire_at = server->put_time;
451 
452 			if (!test_bit(AFS_SERVER_FL_VL_FAIL, &server->flags) &&
453 			    !test_bit(AFS_SERVER_FL_NOT_FOUND, &server->flags))
454 				expire_at += afs_server_gc_delay;
455 			if (purging || expire_at <= now) {
456 				server->gc_next = gc_list;
457 				gc_list = server;
458 			} else if (expire_at < next_manage) {
459 				next_manage = expire_at;
460 			}
461 		}
462 	}
463 
464 	read_sequnlock_excl(&net->fs_lock);
465 
466 	/* Update the timer on the way out.  We have to pass an increment on
467 	 * servers_outstanding in the namespace that we are in to the timer or
468 	 * the work scheduler.
469 	 */
470 	if (!purging && next_manage < TIME64_MAX) {
471 		now = ktime_get_real_seconds();
472 
473 		if (next_manage - now <= 0) {
474 			if (queue_work(afs_wq, &net->fs_manager))
475 				afs_inc_servers_outstanding(net);
476 		} else {
477 			afs_set_server_timer(net, next_manage - now);
478 		}
479 	}
480 
481 	afs_gc_servers(net, gc_list);
482 
483 	afs_dec_servers_outstanding(net);
484 	_leave(" [%d]", atomic_read(&net->servers_outstanding));
485 }
486 
487 static void afs_queue_server_manager(struct afs_net *net)
488 {
489 	afs_inc_servers_outstanding(net);
490 	if (!queue_work(afs_wq, &net->fs_manager))
491 		afs_dec_servers_outstanding(net);
492 }
493 
494 /*
495  * Purge list of servers.
496  */
497 void afs_purge_servers(struct afs_net *net)
498 {
499 	_enter("");
500 
501 	if (del_timer_sync(&net->fs_timer))
502 		atomic_dec(&net->servers_outstanding);
503 
504 	afs_queue_server_manager(net);
505 
506 	_debug("wait");
507 	wait_var_event(&net->servers_outstanding,
508 		       !atomic_read(&net->servers_outstanding));
509 	_leave("");
510 }
511 
512 /*
513  * Get an update for a server's address list.
514  */
515 static noinline bool afs_update_server_record(struct afs_fs_cursor *fc, struct afs_server *server)
516 {
517 	struct afs_addr_list *alist, *discard;
518 
519 	_enter("");
520 
521 	alist = afs_vl_lookup_addrs(fc->vnode->volume->cell, fc->key,
522 				    &server->uuid);
523 	if (IS_ERR(alist)) {
524 		if ((PTR_ERR(alist) == -ERESTARTSYS ||
525 		     PTR_ERR(alist) == -EINTR) &&
526 		    !(fc->flags & AFS_FS_CURSOR_INTR) &&
527 		    server->addresses) {
528 			_leave(" = t [intr]");
529 			return true;
530 		}
531 		fc->error = PTR_ERR(alist);
532 		_leave(" = f [%d]", fc->error);
533 		return false;
534 	}
535 
536 	discard = alist;
537 	if (server->addr_version != alist->version) {
538 		write_lock(&server->fs_lock);
539 		discard = rcu_dereference_protected(server->addresses,
540 						    lockdep_is_held(&server->fs_lock));
541 		rcu_assign_pointer(server->addresses, alist);
542 		server->addr_version = alist->version;
543 		write_unlock(&server->fs_lock);
544 	}
545 
546 	server->update_at = ktime_get_real_seconds() + afs_server_update_delay;
547 	afs_put_addrlist(discard);
548 	_leave(" = t");
549 	return true;
550 }
551 
552 /*
553  * See if a server's address list needs updating.
554  */
555 bool afs_check_server_record(struct afs_fs_cursor *fc, struct afs_server *server)
556 {
557 	time64_t now = ktime_get_real_seconds();
558 	long diff;
559 	bool success;
560 	int ret, retries = 0;
561 
562 	_enter("");
563 
564 	ASSERT(server);
565 
566 retry:
567 	diff = READ_ONCE(server->update_at) - now;
568 	if (diff > 0) {
569 		_leave(" = t [not now %ld]", diff);
570 		return true;
571 	}
572 
573 	if (!test_and_set_bit_lock(AFS_SERVER_FL_UPDATING, &server->flags)) {
574 		success = afs_update_server_record(fc, server);
575 		clear_bit_unlock(AFS_SERVER_FL_UPDATING, &server->flags);
576 		wake_up_bit(&server->flags, AFS_SERVER_FL_UPDATING);
577 		_leave(" = %d", success);
578 		return success;
579 	}
580 
581 	ret = wait_on_bit(&server->flags, AFS_SERVER_FL_UPDATING,
582 			  TASK_INTERRUPTIBLE);
583 	if (ret == -ERESTARTSYS) {
584 		if (!(fc->flags & AFS_FS_CURSOR_INTR) && server->addresses) {
585 			_leave(" = t [intr]");
586 			return true;
587 		}
588 		fc->error = ret;
589 		_leave(" = f [intr]");
590 		return false;
591 	}
592 
593 	retries++;
594 	if (retries == 4) {
595 		_leave(" = f [stale]");
596 		ret = -ESTALE;
597 		return false;
598 	}
599 	goto retry;
600 }
601