xref: /openbmc/linux/net/core/neighbour.c (revision 827634ad)
1 /*
2  *	Generic address resolution entity
3  *
4  *	Authors:
5  *	Pedro Roque		<roque@di.fc.ul.pt>
6  *	Alexey Kuznetsov	<kuznet@ms2.inr.ac.ru>
7  *
8  *	This program is free software; you can redistribute it and/or
9  *      modify it under the terms of the GNU General Public License
10  *      as published by the Free Software Foundation; either version
11  *      2 of the License, or (at your option) any later version.
12  *
13  *	Fixes:
14  *	Vitaly E. Lavrov	releasing NULL neighbor in neigh_add.
15  *	Harald Welte		Add neighbour cache statistics like rtstat
16  */
17 
18 #define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
19 
20 #include <linux/slab.h>
21 #include <linux/types.h>
22 #include <linux/kernel.h>
23 #include <linux/module.h>
24 #include <linux/socket.h>
25 #include <linux/netdevice.h>
26 #include <linux/proc_fs.h>
27 #ifdef CONFIG_SYSCTL
28 #include <linux/sysctl.h>
29 #endif
30 #include <linux/times.h>
31 #include <net/net_namespace.h>
32 #include <net/neighbour.h>
33 #include <net/dst.h>
34 #include <net/sock.h>
35 #include <net/netevent.h>
36 #include <net/netlink.h>
37 #include <linux/rtnetlink.h>
38 #include <linux/random.h>
39 #include <linux/string.h>
40 #include <linux/log2.h>
41 #include <linux/inetdevice.h>
42 #include <net/addrconf.h>
43 
44 #define DEBUG
45 #define NEIGH_DEBUG 1
46 #define neigh_dbg(level, fmt, ...)		\
47 do {						\
48 	if (level <= NEIGH_DEBUG)		\
49 		pr_debug(fmt, ##__VA_ARGS__);	\
50 } while (0)
51 
52 #define PNEIGH_HASHMASK		0xF
53 
54 static void neigh_timer_handler(unsigned long arg);
55 static void __neigh_notify(struct neighbour *n, int type, int flags);
56 static void neigh_update_notify(struct neighbour *neigh);
57 static int pneigh_ifdown(struct neigh_table *tbl, struct net_device *dev);
58 
59 #ifdef CONFIG_PROC_FS
60 static const struct file_operations neigh_stat_seq_fops;
61 #endif
62 
63 /*
64    Neighbour hash table buckets are protected with rwlock tbl->lock.
65 
66    - All the scans/updates to hash buckets MUST be made under this lock.
67    - NOTHING clever should be made under this lock: no callbacks
68      to protocol backends, no attempts to send something to network.
69      It will result in deadlocks, if backend/driver wants to use neighbour
70      cache.
71    - If the entry requires some non-trivial actions, increase
72      its reference count and release table lock.
73 
74    Neighbour entries are protected:
75    - with reference count.
76    - with rwlock neigh->lock
77 
78    Reference count prevents destruction.
79 
80    neigh->lock mainly serializes ll address data and its validity state.
81    However, the same lock is used to protect another entry fields:
82     - timer
83     - resolution queue
84 
85    Again, nothing clever shall be made under neigh->lock,
86    the most complicated procedure, which we allow is dev->hard_header.
87    It is supposed, that dev->hard_header is simplistic and does
88    not make callbacks to neighbour tables.
89  */
90 
91 static int neigh_blackhole(struct neighbour *neigh, struct sk_buff *skb)
92 {
93 	kfree_skb(skb);
94 	return -ENETDOWN;
95 }
96 
97 static void neigh_cleanup_and_release(struct neighbour *neigh)
98 {
99 	if (neigh->parms->neigh_cleanup)
100 		neigh->parms->neigh_cleanup(neigh);
101 
102 	__neigh_notify(neigh, RTM_DELNEIGH, 0);
103 	neigh_release(neigh);
104 }
105 
106 /*
107  * It is random distribution in the interval (1/2)*base...(3/2)*base.
108  * It corresponds to default IPv6 settings and is not overridable,
109  * because it is really reasonable choice.
110  */
111 
112 unsigned long neigh_rand_reach_time(unsigned long base)
113 {
114 	return base ? (prandom_u32() % base) + (base >> 1) : 0;
115 }
116 EXPORT_SYMBOL(neigh_rand_reach_time);
117 
118 
119 static int neigh_forced_gc(struct neigh_table *tbl)
120 {
121 	int shrunk = 0;
122 	int i;
123 	struct neigh_hash_table *nht;
124 
125 	NEIGH_CACHE_STAT_INC(tbl, forced_gc_runs);
126 
127 	write_lock_bh(&tbl->lock);
128 	nht = rcu_dereference_protected(tbl->nht,
129 					lockdep_is_held(&tbl->lock));
130 	for (i = 0; i < (1 << nht->hash_shift); i++) {
131 		struct neighbour *n;
132 		struct neighbour __rcu **np;
133 
134 		np = &nht->hash_buckets[i];
135 		while ((n = rcu_dereference_protected(*np,
136 					lockdep_is_held(&tbl->lock))) != NULL) {
137 			/* Neighbour record may be discarded if:
138 			 * - nobody refers to it.
139 			 * - it is not permanent
140 			 */
141 			write_lock(&n->lock);
142 			if (atomic_read(&n->refcnt) == 1 &&
143 			    !(n->nud_state & NUD_PERMANENT)) {
144 				rcu_assign_pointer(*np,
145 					rcu_dereference_protected(n->next,
146 						  lockdep_is_held(&tbl->lock)));
147 				n->dead = 1;
148 				shrunk	= 1;
149 				write_unlock(&n->lock);
150 				neigh_cleanup_and_release(n);
151 				continue;
152 			}
153 			write_unlock(&n->lock);
154 			np = &n->next;
155 		}
156 	}
157 
158 	tbl->last_flush = jiffies;
159 
160 	write_unlock_bh(&tbl->lock);
161 
162 	return shrunk;
163 }
164 
165 static void neigh_add_timer(struct neighbour *n, unsigned long when)
166 {
167 	neigh_hold(n);
168 	if (unlikely(mod_timer(&n->timer, when))) {
169 		printk("NEIGH: BUG, double timer add, state is %x\n",
170 		       n->nud_state);
171 		dump_stack();
172 	}
173 }
174 
175 static int neigh_del_timer(struct neighbour *n)
176 {
177 	if ((n->nud_state & NUD_IN_TIMER) &&
178 	    del_timer(&n->timer)) {
179 		neigh_release(n);
180 		return 1;
181 	}
182 	return 0;
183 }
184 
185 static void pneigh_queue_purge(struct sk_buff_head *list)
186 {
187 	struct sk_buff *skb;
188 
189 	while ((skb = skb_dequeue(list)) != NULL) {
190 		dev_put(skb->dev);
191 		kfree_skb(skb);
192 	}
193 }
194 
195 static void neigh_flush_dev(struct neigh_table *tbl, struct net_device *dev)
196 {
197 	int i;
198 	struct neigh_hash_table *nht;
199 
200 	nht = rcu_dereference_protected(tbl->nht,
201 					lockdep_is_held(&tbl->lock));
202 
203 	for (i = 0; i < (1 << nht->hash_shift); i++) {
204 		struct neighbour *n;
205 		struct neighbour __rcu **np = &nht->hash_buckets[i];
206 
207 		while ((n = rcu_dereference_protected(*np,
208 					lockdep_is_held(&tbl->lock))) != NULL) {
209 			if (dev && n->dev != dev) {
210 				np = &n->next;
211 				continue;
212 			}
213 			rcu_assign_pointer(*np,
214 				   rcu_dereference_protected(n->next,
215 						lockdep_is_held(&tbl->lock)));
216 			write_lock(&n->lock);
217 			neigh_del_timer(n);
218 			n->dead = 1;
219 
220 			if (atomic_read(&n->refcnt) != 1) {
221 				/* The most unpleasant situation.
222 				   We must destroy neighbour entry,
223 				   but someone still uses it.
224 
225 				   The destroy will be delayed until
226 				   the last user releases us, but
227 				   we must kill timers etc. and move
228 				   it to safe state.
229 				 */
230 				__skb_queue_purge(&n->arp_queue);
231 				n->arp_queue_len_bytes = 0;
232 				n->output = neigh_blackhole;
233 				if (n->nud_state & NUD_VALID)
234 					n->nud_state = NUD_NOARP;
235 				else
236 					n->nud_state = NUD_NONE;
237 				neigh_dbg(2, "neigh %p is stray\n", n);
238 			}
239 			write_unlock(&n->lock);
240 			neigh_cleanup_and_release(n);
241 		}
242 	}
243 }
244 
245 void neigh_changeaddr(struct neigh_table *tbl, struct net_device *dev)
246 {
247 	write_lock_bh(&tbl->lock);
248 	neigh_flush_dev(tbl, dev);
249 	write_unlock_bh(&tbl->lock);
250 }
251 EXPORT_SYMBOL(neigh_changeaddr);
252 
253 int neigh_ifdown(struct neigh_table *tbl, struct net_device *dev)
254 {
255 	write_lock_bh(&tbl->lock);
256 	neigh_flush_dev(tbl, dev);
257 	pneigh_ifdown(tbl, dev);
258 	write_unlock_bh(&tbl->lock);
259 
260 	del_timer_sync(&tbl->proxy_timer);
261 	pneigh_queue_purge(&tbl->proxy_queue);
262 	return 0;
263 }
264 EXPORT_SYMBOL(neigh_ifdown);
265 
266 static struct neighbour *neigh_alloc(struct neigh_table *tbl, struct net_device *dev)
267 {
268 	struct neighbour *n = NULL;
269 	unsigned long now = jiffies;
270 	int entries;
271 
272 	entries = atomic_inc_return(&tbl->entries) - 1;
273 	if (entries >= tbl->gc_thresh3 ||
274 	    (entries >= tbl->gc_thresh2 &&
275 	     time_after(now, tbl->last_flush + 5 * HZ))) {
276 		if (!neigh_forced_gc(tbl) &&
277 		    entries >= tbl->gc_thresh3)
278 			goto out_entries;
279 	}
280 
281 	n = kzalloc(tbl->entry_size + dev->neigh_priv_len, GFP_ATOMIC);
282 	if (!n)
283 		goto out_entries;
284 
285 	__skb_queue_head_init(&n->arp_queue);
286 	rwlock_init(&n->lock);
287 	seqlock_init(&n->ha_lock);
288 	n->updated	  = n->used = now;
289 	n->nud_state	  = NUD_NONE;
290 	n->output	  = neigh_blackhole;
291 	seqlock_init(&n->hh.hh_lock);
292 	n->parms	  = neigh_parms_clone(&tbl->parms);
293 	setup_timer(&n->timer, neigh_timer_handler, (unsigned long)n);
294 
295 	NEIGH_CACHE_STAT_INC(tbl, allocs);
296 	n->tbl		  = tbl;
297 	atomic_set(&n->refcnt, 1);
298 	n->dead		  = 1;
299 out:
300 	return n;
301 
302 out_entries:
303 	atomic_dec(&tbl->entries);
304 	goto out;
305 }
306 
307 static void neigh_get_hash_rnd(u32 *x)
308 {
309 	get_random_bytes(x, sizeof(*x));
310 	*x |= 1;
311 }
312 
313 static struct neigh_hash_table *neigh_hash_alloc(unsigned int shift)
314 {
315 	size_t size = (1 << shift) * sizeof(struct neighbour *);
316 	struct neigh_hash_table *ret;
317 	struct neighbour __rcu **buckets;
318 	int i;
319 
320 	ret = kmalloc(sizeof(*ret), GFP_ATOMIC);
321 	if (!ret)
322 		return NULL;
323 	if (size <= PAGE_SIZE)
324 		buckets = kzalloc(size, GFP_ATOMIC);
325 	else
326 		buckets = (struct neighbour __rcu **)
327 			  __get_free_pages(GFP_ATOMIC | __GFP_ZERO,
328 					   get_order(size));
329 	if (!buckets) {
330 		kfree(ret);
331 		return NULL;
332 	}
333 	ret->hash_buckets = buckets;
334 	ret->hash_shift = shift;
335 	for (i = 0; i < NEIGH_NUM_HASH_RND; i++)
336 		neigh_get_hash_rnd(&ret->hash_rnd[i]);
337 	return ret;
338 }
339 
340 static void neigh_hash_free_rcu(struct rcu_head *head)
341 {
342 	struct neigh_hash_table *nht = container_of(head,
343 						    struct neigh_hash_table,
344 						    rcu);
345 	size_t size = (1 << nht->hash_shift) * sizeof(struct neighbour *);
346 	struct neighbour __rcu **buckets = nht->hash_buckets;
347 
348 	if (size <= PAGE_SIZE)
349 		kfree(buckets);
350 	else
351 		free_pages((unsigned long)buckets, get_order(size));
352 	kfree(nht);
353 }
354 
355 static struct neigh_hash_table *neigh_hash_grow(struct neigh_table *tbl,
356 						unsigned long new_shift)
357 {
358 	unsigned int i, hash;
359 	struct neigh_hash_table *new_nht, *old_nht;
360 
361 	NEIGH_CACHE_STAT_INC(tbl, hash_grows);
362 
363 	old_nht = rcu_dereference_protected(tbl->nht,
364 					    lockdep_is_held(&tbl->lock));
365 	new_nht = neigh_hash_alloc(new_shift);
366 	if (!new_nht)
367 		return old_nht;
368 
369 	for (i = 0; i < (1 << old_nht->hash_shift); i++) {
370 		struct neighbour *n, *next;
371 
372 		for (n = rcu_dereference_protected(old_nht->hash_buckets[i],
373 						   lockdep_is_held(&tbl->lock));
374 		     n != NULL;
375 		     n = next) {
376 			hash = tbl->hash(n->primary_key, n->dev,
377 					 new_nht->hash_rnd);
378 
379 			hash >>= (32 - new_nht->hash_shift);
380 			next = rcu_dereference_protected(n->next,
381 						lockdep_is_held(&tbl->lock));
382 
383 			rcu_assign_pointer(n->next,
384 					   rcu_dereference_protected(
385 						new_nht->hash_buckets[hash],
386 						lockdep_is_held(&tbl->lock)));
387 			rcu_assign_pointer(new_nht->hash_buckets[hash], n);
388 		}
389 	}
390 
391 	rcu_assign_pointer(tbl->nht, new_nht);
392 	call_rcu(&old_nht->rcu, neigh_hash_free_rcu);
393 	return new_nht;
394 }
395 
396 struct neighbour *neigh_lookup(struct neigh_table *tbl, const void *pkey,
397 			       struct net_device *dev)
398 {
399 	struct neighbour *n;
400 
401 	NEIGH_CACHE_STAT_INC(tbl, lookups);
402 
403 	rcu_read_lock_bh();
404 	n = __neigh_lookup_noref(tbl, pkey, dev);
405 	if (n) {
406 		if (!atomic_inc_not_zero(&n->refcnt))
407 			n = NULL;
408 		NEIGH_CACHE_STAT_INC(tbl, hits);
409 	}
410 
411 	rcu_read_unlock_bh();
412 	return n;
413 }
414 EXPORT_SYMBOL(neigh_lookup);
415 
416 struct neighbour *neigh_lookup_nodev(struct neigh_table *tbl, struct net *net,
417 				     const void *pkey)
418 {
419 	struct neighbour *n;
420 	int key_len = tbl->key_len;
421 	u32 hash_val;
422 	struct neigh_hash_table *nht;
423 
424 	NEIGH_CACHE_STAT_INC(tbl, lookups);
425 
426 	rcu_read_lock_bh();
427 	nht = rcu_dereference_bh(tbl->nht);
428 	hash_val = tbl->hash(pkey, NULL, nht->hash_rnd) >> (32 - nht->hash_shift);
429 
430 	for (n = rcu_dereference_bh(nht->hash_buckets[hash_val]);
431 	     n != NULL;
432 	     n = rcu_dereference_bh(n->next)) {
433 		if (!memcmp(n->primary_key, pkey, key_len) &&
434 		    net_eq(dev_net(n->dev), net)) {
435 			if (!atomic_inc_not_zero(&n->refcnt))
436 				n = NULL;
437 			NEIGH_CACHE_STAT_INC(tbl, hits);
438 			break;
439 		}
440 	}
441 
442 	rcu_read_unlock_bh();
443 	return n;
444 }
445 EXPORT_SYMBOL(neigh_lookup_nodev);
446 
447 struct neighbour *__neigh_create(struct neigh_table *tbl, const void *pkey,
448 				 struct net_device *dev, bool want_ref)
449 {
450 	u32 hash_val;
451 	int key_len = tbl->key_len;
452 	int error;
453 	struct neighbour *n1, *rc, *n = neigh_alloc(tbl, dev);
454 	struct neigh_hash_table *nht;
455 
456 	if (!n) {
457 		rc = ERR_PTR(-ENOBUFS);
458 		goto out;
459 	}
460 
461 	memcpy(n->primary_key, pkey, key_len);
462 	n->dev = dev;
463 	dev_hold(dev);
464 
465 	/* Protocol specific setup. */
466 	if (tbl->constructor &&	(error = tbl->constructor(n)) < 0) {
467 		rc = ERR_PTR(error);
468 		goto out_neigh_release;
469 	}
470 
471 	if (dev->netdev_ops->ndo_neigh_construct) {
472 		error = dev->netdev_ops->ndo_neigh_construct(n);
473 		if (error < 0) {
474 			rc = ERR_PTR(error);
475 			goto out_neigh_release;
476 		}
477 	}
478 
479 	/* Device specific setup. */
480 	if (n->parms->neigh_setup &&
481 	    (error = n->parms->neigh_setup(n)) < 0) {
482 		rc = ERR_PTR(error);
483 		goto out_neigh_release;
484 	}
485 
486 	n->confirmed = jiffies - (NEIGH_VAR(n->parms, BASE_REACHABLE_TIME) << 1);
487 
488 	write_lock_bh(&tbl->lock);
489 	nht = rcu_dereference_protected(tbl->nht,
490 					lockdep_is_held(&tbl->lock));
491 
492 	if (atomic_read(&tbl->entries) > (1 << nht->hash_shift))
493 		nht = neigh_hash_grow(tbl, nht->hash_shift + 1);
494 
495 	hash_val = tbl->hash(pkey, dev, nht->hash_rnd) >> (32 - nht->hash_shift);
496 
497 	if (n->parms->dead) {
498 		rc = ERR_PTR(-EINVAL);
499 		goto out_tbl_unlock;
500 	}
501 
502 	for (n1 = rcu_dereference_protected(nht->hash_buckets[hash_val],
503 					    lockdep_is_held(&tbl->lock));
504 	     n1 != NULL;
505 	     n1 = rcu_dereference_protected(n1->next,
506 			lockdep_is_held(&tbl->lock))) {
507 		if (dev == n1->dev && !memcmp(n1->primary_key, pkey, key_len)) {
508 			if (want_ref)
509 				neigh_hold(n1);
510 			rc = n1;
511 			goto out_tbl_unlock;
512 		}
513 	}
514 
515 	n->dead = 0;
516 	if (want_ref)
517 		neigh_hold(n);
518 	rcu_assign_pointer(n->next,
519 			   rcu_dereference_protected(nht->hash_buckets[hash_val],
520 						     lockdep_is_held(&tbl->lock)));
521 	rcu_assign_pointer(nht->hash_buckets[hash_val], n);
522 	write_unlock_bh(&tbl->lock);
523 	neigh_dbg(2, "neigh %p is created\n", n);
524 	rc = n;
525 out:
526 	return rc;
527 out_tbl_unlock:
528 	write_unlock_bh(&tbl->lock);
529 out_neigh_release:
530 	neigh_release(n);
531 	goto out;
532 }
533 EXPORT_SYMBOL(__neigh_create);
534 
535 static u32 pneigh_hash(const void *pkey, int key_len)
536 {
537 	u32 hash_val = *(u32 *)(pkey + key_len - 4);
538 	hash_val ^= (hash_val >> 16);
539 	hash_val ^= hash_val >> 8;
540 	hash_val ^= hash_val >> 4;
541 	hash_val &= PNEIGH_HASHMASK;
542 	return hash_val;
543 }
544 
545 static struct pneigh_entry *__pneigh_lookup_1(struct pneigh_entry *n,
546 					      struct net *net,
547 					      const void *pkey,
548 					      int key_len,
549 					      struct net_device *dev)
550 {
551 	while (n) {
552 		if (!memcmp(n->key, pkey, key_len) &&
553 		    net_eq(pneigh_net(n), net) &&
554 		    (n->dev == dev || !n->dev))
555 			return n;
556 		n = n->next;
557 	}
558 	return NULL;
559 }
560 
561 struct pneigh_entry *__pneigh_lookup(struct neigh_table *tbl,
562 		struct net *net, const void *pkey, struct net_device *dev)
563 {
564 	int key_len = tbl->key_len;
565 	u32 hash_val = pneigh_hash(pkey, key_len);
566 
567 	return __pneigh_lookup_1(tbl->phash_buckets[hash_val],
568 				 net, pkey, key_len, dev);
569 }
570 EXPORT_SYMBOL_GPL(__pneigh_lookup);
571 
572 struct pneigh_entry * pneigh_lookup(struct neigh_table *tbl,
573 				    struct net *net, const void *pkey,
574 				    struct net_device *dev, int creat)
575 {
576 	struct pneigh_entry *n;
577 	int key_len = tbl->key_len;
578 	u32 hash_val = pneigh_hash(pkey, key_len);
579 
580 	read_lock_bh(&tbl->lock);
581 	n = __pneigh_lookup_1(tbl->phash_buckets[hash_val],
582 			      net, pkey, key_len, dev);
583 	read_unlock_bh(&tbl->lock);
584 
585 	if (n || !creat)
586 		goto out;
587 
588 	ASSERT_RTNL();
589 
590 	n = kmalloc(sizeof(*n) + key_len, GFP_KERNEL);
591 	if (!n)
592 		goto out;
593 
594 	write_pnet(&n->net, net);
595 	memcpy(n->key, pkey, key_len);
596 	n->dev = dev;
597 	if (dev)
598 		dev_hold(dev);
599 
600 	if (tbl->pconstructor && tbl->pconstructor(n)) {
601 		if (dev)
602 			dev_put(dev);
603 		kfree(n);
604 		n = NULL;
605 		goto out;
606 	}
607 
608 	write_lock_bh(&tbl->lock);
609 	n->next = tbl->phash_buckets[hash_val];
610 	tbl->phash_buckets[hash_val] = n;
611 	write_unlock_bh(&tbl->lock);
612 out:
613 	return n;
614 }
615 EXPORT_SYMBOL(pneigh_lookup);
616 
617 
618 int pneigh_delete(struct neigh_table *tbl, struct net *net, const void *pkey,
619 		  struct net_device *dev)
620 {
621 	struct pneigh_entry *n, **np;
622 	int key_len = tbl->key_len;
623 	u32 hash_val = pneigh_hash(pkey, key_len);
624 
625 	write_lock_bh(&tbl->lock);
626 	for (np = &tbl->phash_buckets[hash_val]; (n = *np) != NULL;
627 	     np = &n->next) {
628 		if (!memcmp(n->key, pkey, key_len) && n->dev == dev &&
629 		    net_eq(pneigh_net(n), net)) {
630 			*np = n->next;
631 			write_unlock_bh(&tbl->lock);
632 			if (tbl->pdestructor)
633 				tbl->pdestructor(n);
634 			if (n->dev)
635 				dev_put(n->dev);
636 			kfree(n);
637 			return 0;
638 		}
639 	}
640 	write_unlock_bh(&tbl->lock);
641 	return -ENOENT;
642 }
643 
644 static int pneigh_ifdown(struct neigh_table *tbl, struct net_device *dev)
645 {
646 	struct pneigh_entry *n, **np;
647 	u32 h;
648 
649 	for (h = 0; h <= PNEIGH_HASHMASK; h++) {
650 		np = &tbl->phash_buckets[h];
651 		while ((n = *np) != NULL) {
652 			if (!dev || n->dev == dev) {
653 				*np = n->next;
654 				if (tbl->pdestructor)
655 					tbl->pdestructor(n);
656 				if (n->dev)
657 					dev_put(n->dev);
658 				kfree(n);
659 				continue;
660 			}
661 			np = &n->next;
662 		}
663 	}
664 	return -ENOENT;
665 }
666 
667 static void neigh_parms_destroy(struct neigh_parms *parms);
668 
669 static inline void neigh_parms_put(struct neigh_parms *parms)
670 {
671 	if (atomic_dec_and_test(&parms->refcnt))
672 		neigh_parms_destroy(parms);
673 }
674 
675 /*
676  *	neighbour must already be out of the table;
677  *
678  */
679 void neigh_destroy(struct neighbour *neigh)
680 {
681 	struct net_device *dev = neigh->dev;
682 
683 	NEIGH_CACHE_STAT_INC(neigh->tbl, destroys);
684 
685 	if (!neigh->dead) {
686 		pr_warn("Destroying alive neighbour %p\n", neigh);
687 		dump_stack();
688 		return;
689 	}
690 
691 	if (neigh_del_timer(neigh))
692 		pr_warn("Impossible event\n");
693 
694 	write_lock_bh(&neigh->lock);
695 	__skb_queue_purge(&neigh->arp_queue);
696 	write_unlock_bh(&neigh->lock);
697 	neigh->arp_queue_len_bytes = 0;
698 
699 	if (dev->netdev_ops->ndo_neigh_destroy)
700 		dev->netdev_ops->ndo_neigh_destroy(neigh);
701 
702 	dev_put(dev);
703 	neigh_parms_put(neigh->parms);
704 
705 	neigh_dbg(2, "neigh %p is destroyed\n", neigh);
706 
707 	atomic_dec(&neigh->tbl->entries);
708 	kfree_rcu(neigh, rcu);
709 }
710 EXPORT_SYMBOL(neigh_destroy);
711 
712 /* Neighbour state is suspicious;
713    disable fast path.
714 
715    Called with write_locked neigh.
716  */
717 static void neigh_suspect(struct neighbour *neigh)
718 {
719 	neigh_dbg(2, "neigh %p is suspected\n", neigh);
720 
721 	neigh->output = neigh->ops->output;
722 }
723 
724 /* Neighbour state is OK;
725    enable fast path.
726 
727    Called with write_locked neigh.
728  */
729 static void neigh_connect(struct neighbour *neigh)
730 {
731 	neigh_dbg(2, "neigh %p is connected\n", neigh);
732 
733 	neigh->output = neigh->ops->connected_output;
734 }
735 
736 static void neigh_periodic_work(struct work_struct *work)
737 {
738 	struct neigh_table *tbl = container_of(work, struct neigh_table, gc_work.work);
739 	struct neighbour *n;
740 	struct neighbour __rcu **np;
741 	unsigned int i;
742 	struct neigh_hash_table *nht;
743 
744 	NEIGH_CACHE_STAT_INC(tbl, periodic_gc_runs);
745 
746 	write_lock_bh(&tbl->lock);
747 	nht = rcu_dereference_protected(tbl->nht,
748 					lockdep_is_held(&tbl->lock));
749 
750 	/*
751 	 *	periodically recompute ReachableTime from random function
752 	 */
753 
754 	if (time_after(jiffies, tbl->last_rand + 300 * HZ)) {
755 		struct neigh_parms *p;
756 		tbl->last_rand = jiffies;
757 		list_for_each_entry(p, &tbl->parms_list, list)
758 			p->reachable_time =
759 				neigh_rand_reach_time(NEIGH_VAR(p, BASE_REACHABLE_TIME));
760 	}
761 
762 	if (atomic_read(&tbl->entries) < tbl->gc_thresh1)
763 		goto out;
764 
765 	for (i = 0 ; i < (1 << nht->hash_shift); i++) {
766 		np = &nht->hash_buckets[i];
767 
768 		while ((n = rcu_dereference_protected(*np,
769 				lockdep_is_held(&tbl->lock))) != NULL) {
770 			unsigned int state;
771 
772 			write_lock(&n->lock);
773 
774 			state = n->nud_state;
775 			if (state & (NUD_PERMANENT | NUD_IN_TIMER)) {
776 				write_unlock(&n->lock);
777 				goto next_elt;
778 			}
779 
780 			if (time_before(n->used, n->confirmed))
781 				n->used = n->confirmed;
782 
783 			if (atomic_read(&n->refcnt) == 1 &&
784 			    (state == NUD_FAILED ||
785 			     time_after(jiffies, n->used + NEIGH_VAR(n->parms, GC_STALETIME)))) {
786 				*np = n->next;
787 				n->dead = 1;
788 				write_unlock(&n->lock);
789 				neigh_cleanup_and_release(n);
790 				continue;
791 			}
792 			write_unlock(&n->lock);
793 
794 next_elt:
795 			np = &n->next;
796 		}
797 		/*
798 		 * It's fine to release lock here, even if hash table
799 		 * grows while we are preempted.
800 		 */
801 		write_unlock_bh(&tbl->lock);
802 		cond_resched();
803 		write_lock_bh(&tbl->lock);
804 		nht = rcu_dereference_protected(tbl->nht,
805 						lockdep_is_held(&tbl->lock));
806 	}
807 out:
808 	/* Cycle through all hash buckets every BASE_REACHABLE_TIME/2 ticks.
809 	 * ARP entry timeouts range from 1/2 BASE_REACHABLE_TIME to 3/2
810 	 * BASE_REACHABLE_TIME.
811 	 */
812 	queue_delayed_work(system_power_efficient_wq, &tbl->gc_work,
813 			      NEIGH_VAR(&tbl->parms, BASE_REACHABLE_TIME) >> 1);
814 	write_unlock_bh(&tbl->lock);
815 }
816 
817 static __inline__ int neigh_max_probes(struct neighbour *n)
818 {
819 	struct neigh_parms *p = n->parms;
820 	return NEIGH_VAR(p, UCAST_PROBES) + NEIGH_VAR(p, APP_PROBES) +
821 	       (n->nud_state & NUD_PROBE ? NEIGH_VAR(p, MCAST_REPROBES) :
822 	        NEIGH_VAR(p, MCAST_PROBES));
823 }
824 
825 static void neigh_invalidate(struct neighbour *neigh)
826 	__releases(neigh->lock)
827 	__acquires(neigh->lock)
828 {
829 	struct sk_buff *skb;
830 
831 	NEIGH_CACHE_STAT_INC(neigh->tbl, res_failed);
832 	neigh_dbg(2, "neigh %p is failed\n", neigh);
833 	neigh->updated = jiffies;
834 
835 	/* It is very thin place. report_unreachable is very complicated
836 	   routine. Particularly, it can hit the same neighbour entry!
837 
838 	   So that, we try to be accurate and avoid dead loop. --ANK
839 	 */
840 	while (neigh->nud_state == NUD_FAILED &&
841 	       (skb = __skb_dequeue(&neigh->arp_queue)) != NULL) {
842 		write_unlock(&neigh->lock);
843 		neigh->ops->error_report(neigh, skb);
844 		write_lock(&neigh->lock);
845 	}
846 	__skb_queue_purge(&neigh->arp_queue);
847 	neigh->arp_queue_len_bytes = 0;
848 }
849 
850 static void neigh_probe(struct neighbour *neigh)
851 	__releases(neigh->lock)
852 {
853 	struct sk_buff *skb = skb_peek_tail(&neigh->arp_queue);
854 	/* keep skb alive even if arp_queue overflows */
855 	if (skb)
856 		skb = skb_copy(skb, GFP_ATOMIC);
857 	write_unlock(&neigh->lock);
858 	neigh->ops->solicit(neigh, skb);
859 	atomic_inc(&neigh->probes);
860 	kfree_skb(skb);
861 }
862 
863 /* Called when a timer expires for a neighbour entry. */
864 
865 static void neigh_timer_handler(unsigned long arg)
866 {
867 	unsigned long now, next;
868 	struct neighbour *neigh = (struct neighbour *)arg;
869 	unsigned int state;
870 	int notify = 0;
871 
872 	write_lock(&neigh->lock);
873 
874 	state = neigh->nud_state;
875 	now = jiffies;
876 	next = now + HZ;
877 
878 	if (!(state & NUD_IN_TIMER))
879 		goto out;
880 
881 	if (state & NUD_REACHABLE) {
882 		if (time_before_eq(now,
883 				   neigh->confirmed + neigh->parms->reachable_time)) {
884 			neigh_dbg(2, "neigh %p is still alive\n", neigh);
885 			next = neigh->confirmed + neigh->parms->reachable_time;
886 		} else if (time_before_eq(now,
887 					  neigh->used +
888 					  NEIGH_VAR(neigh->parms, DELAY_PROBE_TIME))) {
889 			neigh_dbg(2, "neigh %p is delayed\n", neigh);
890 			neigh->nud_state = NUD_DELAY;
891 			neigh->updated = jiffies;
892 			neigh_suspect(neigh);
893 			next = now + NEIGH_VAR(neigh->parms, DELAY_PROBE_TIME);
894 		} else {
895 			neigh_dbg(2, "neigh %p is suspected\n", neigh);
896 			neigh->nud_state = NUD_STALE;
897 			neigh->updated = jiffies;
898 			neigh_suspect(neigh);
899 			notify = 1;
900 		}
901 	} else if (state & NUD_DELAY) {
902 		if (time_before_eq(now,
903 				   neigh->confirmed +
904 				   NEIGH_VAR(neigh->parms, DELAY_PROBE_TIME))) {
905 			neigh_dbg(2, "neigh %p is now reachable\n", neigh);
906 			neigh->nud_state = NUD_REACHABLE;
907 			neigh->updated = jiffies;
908 			neigh_connect(neigh);
909 			notify = 1;
910 			next = neigh->confirmed + neigh->parms->reachable_time;
911 		} else {
912 			neigh_dbg(2, "neigh %p is probed\n", neigh);
913 			neigh->nud_state = NUD_PROBE;
914 			neigh->updated = jiffies;
915 			atomic_set(&neigh->probes, 0);
916 			next = now + NEIGH_VAR(neigh->parms, RETRANS_TIME);
917 		}
918 	} else {
919 		/* NUD_PROBE|NUD_INCOMPLETE */
920 		next = now + NEIGH_VAR(neigh->parms, RETRANS_TIME);
921 	}
922 
923 	if ((neigh->nud_state & (NUD_INCOMPLETE | NUD_PROBE)) &&
924 	    atomic_read(&neigh->probes) >= neigh_max_probes(neigh)) {
925 		neigh->nud_state = NUD_FAILED;
926 		notify = 1;
927 		neigh_invalidate(neigh);
928 		goto out;
929 	}
930 
931 	if (neigh->nud_state & NUD_IN_TIMER) {
932 		if (time_before(next, jiffies + HZ/2))
933 			next = jiffies + HZ/2;
934 		if (!mod_timer(&neigh->timer, next))
935 			neigh_hold(neigh);
936 	}
937 	if (neigh->nud_state & (NUD_INCOMPLETE | NUD_PROBE)) {
938 		neigh_probe(neigh);
939 	} else {
940 out:
941 		write_unlock(&neigh->lock);
942 	}
943 
944 	if (notify)
945 		neigh_update_notify(neigh);
946 
947 	neigh_release(neigh);
948 }
949 
950 int __neigh_event_send(struct neighbour *neigh, struct sk_buff *skb)
951 {
952 	int rc;
953 	bool immediate_probe = false;
954 
955 	write_lock_bh(&neigh->lock);
956 
957 	rc = 0;
958 	if (neigh->nud_state & (NUD_CONNECTED | NUD_DELAY | NUD_PROBE))
959 		goto out_unlock_bh;
960 
961 	if (!(neigh->nud_state & (NUD_STALE | NUD_INCOMPLETE))) {
962 		if (NEIGH_VAR(neigh->parms, MCAST_PROBES) +
963 		    NEIGH_VAR(neigh->parms, APP_PROBES)) {
964 			unsigned long next, now = jiffies;
965 
966 			atomic_set(&neigh->probes,
967 				   NEIGH_VAR(neigh->parms, UCAST_PROBES));
968 			neigh->nud_state     = NUD_INCOMPLETE;
969 			neigh->updated = now;
970 			next = now + max(NEIGH_VAR(neigh->parms, RETRANS_TIME),
971 					 HZ/2);
972 			neigh_add_timer(neigh, next);
973 			immediate_probe = true;
974 		} else {
975 			neigh->nud_state = NUD_FAILED;
976 			neigh->updated = jiffies;
977 			write_unlock_bh(&neigh->lock);
978 
979 			kfree_skb(skb);
980 			return 1;
981 		}
982 	} else if (neigh->nud_state & NUD_STALE) {
983 		neigh_dbg(2, "neigh %p is delayed\n", neigh);
984 		neigh->nud_state = NUD_DELAY;
985 		neigh->updated = jiffies;
986 		neigh_add_timer(neigh, jiffies +
987 				NEIGH_VAR(neigh->parms, DELAY_PROBE_TIME));
988 	}
989 
990 	if (neigh->nud_state == NUD_INCOMPLETE) {
991 		if (skb) {
992 			while (neigh->arp_queue_len_bytes + skb->truesize >
993 			       NEIGH_VAR(neigh->parms, QUEUE_LEN_BYTES)) {
994 				struct sk_buff *buff;
995 
996 				buff = __skb_dequeue(&neigh->arp_queue);
997 				if (!buff)
998 					break;
999 				neigh->arp_queue_len_bytes -= buff->truesize;
1000 				kfree_skb(buff);
1001 				NEIGH_CACHE_STAT_INC(neigh->tbl, unres_discards);
1002 			}
1003 			skb_dst_force(skb);
1004 			__skb_queue_tail(&neigh->arp_queue, skb);
1005 			neigh->arp_queue_len_bytes += skb->truesize;
1006 		}
1007 		rc = 1;
1008 	}
1009 out_unlock_bh:
1010 	if (immediate_probe)
1011 		neigh_probe(neigh);
1012 	else
1013 		write_unlock(&neigh->lock);
1014 	local_bh_enable();
1015 	return rc;
1016 }
1017 EXPORT_SYMBOL(__neigh_event_send);
1018 
1019 static void neigh_update_hhs(struct neighbour *neigh)
1020 {
1021 	struct hh_cache *hh;
1022 	void (*update)(struct hh_cache*, const struct net_device*, const unsigned char *)
1023 		= NULL;
1024 
1025 	if (neigh->dev->header_ops)
1026 		update = neigh->dev->header_ops->cache_update;
1027 
1028 	if (update) {
1029 		hh = &neigh->hh;
1030 		if (hh->hh_len) {
1031 			write_seqlock_bh(&hh->hh_lock);
1032 			update(hh, neigh->dev, neigh->ha);
1033 			write_sequnlock_bh(&hh->hh_lock);
1034 		}
1035 	}
1036 }
1037 
1038 
1039 
1040 /* Generic update routine.
1041    -- lladdr is new lladdr or NULL, if it is not supplied.
1042    -- new    is new state.
1043    -- flags
1044 	NEIGH_UPDATE_F_OVERRIDE allows to override existing lladdr,
1045 				if it is different.
1046 	NEIGH_UPDATE_F_WEAK_OVERRIDE will suspect existing "connected"
1047 				lladdr instead of overriding it
1048 				if it is different.
1049 				It also allows to retain current state
1050 				if lladdr is unchanged.
1051 	NEIGH_UPDATE_F_ADMIN	means that the change is administrative.
1052 
1053 	NEIGH_UPDATE_F_OVERRIDE_ISROUTER allows to override existing
1054 				NTF_ROUTER flag.
1055 	NEIGH_UPDATE_F_ISROUTER	indicates if the neighbour is known as
1056 				a router.
1057 
1058    Caller MUST hold reference count on the entry.
1059  */
1060 
1061 int neigh_update(struct neighbour *neigh, const u8 *lladdr, u8 new,
1062 		 u32 flags)
1063 {
1064 	u8 old;
1065 	int err;
1066 	int notify = 0;
1067 	struct net_device *dev;
1068 	int update_isrouter = 0;
1069 
1070 	write_lock_bh(&neigh->lock);
1071 
1072 	dev    = neigh->dev;
1073 	old    = neigh->nud_state;
1074 	err    = -EPERM;
1075 
1076 	if (!(flags & NEIGH_UPDATE_F_ADMIN) &&
1077 	    (old & (NUD_NOARP | NUD_PERMANENT)))
1078 		goto out;
1079 
1080 	if (!(new & NUD_VALID)) {
1081 		neigh_del_timer(neigh);
1082 		if (old & NUD_CONNECTED)
1083 			neigh_suspect(neigh);
1084 		neigh->nud_state = new;
1085 		err = 0;
1086 		notify = old & NUD_VALID;
1087 		if ((old & (NUD_INCOMPLETE | NUD_PROBE)) &&
1088 		    (new & NUD_FAILED)) {
1089 			neigh_invalidate(neigh);
1090 			notify = 1;
1091 		}
1092 		goto out;
1093 	}
1094 
1095 	/* Compare new lladdr with cached one */
1096 	if (!dev->addr_len) {
1097 		/* First case: device needs no address. */
1098 		lladdr = neigh->ha;
1099 	} else if (lladdr) {
1100 		/* The second case: if something is already cached
1101 		   and a new address is proposed:
1102 		   - compare new & old
1103 		   - if they are different, check override flag
1104 		 */
1105 		if ((old & NUD_VALID) &&
1106 		    !memcmp(lladdr, neigh->ha, dev->addr_len))
1107 			lladdr = neigh->ha;
1108 	} else {
1109 		/* No address is supplied; if we know something,
1110 		   use it, otherwise discard the request.
1111 		 */
1112 		err = -EINVAL;
1113 		if (!(old & NUD_VALID))
1114 			goto out;
1115 		lladdr = neigh->ha;
1116 	}
1117 
1118 	if (new & NUD_CONNECTED)
1119 		neigh->confirmed = jiffies;
1120 	neigh->updated = jiffies;
1121 
1122 	/* If entry was valid and address is not changed,
1123 	   do not change entry state, if new one is STALE.
1124 	 */
1125 	err = 0;
1126 	update_isrouter = flags & NEIGH_UPDATE_F_OVERRIDE_ISROUTER;
1127 	if (old & NUD_VALID) {
1128 		if (lladdr != neigh->ha && !(flags & NEIGH_UPDATE_F_OVERRIDE)) {
1129 			update_isrouter = 0;
1130 			if ((flags & NEIGH_UPDATE_F_WEAK_OVERRIDE) &&
1131 			    (old & NUD_CONNECTED)) {
1132 				lladdr = neigh->ha;
1133 				new = NUD_STALE;
1134 			} else
1135 				goto out;
1136 		} else {
1137 			if (lladdr == neigh->ha && new == NUD_STALE &&
1138 			    ((flags & NEIGH_UPDATE_F_WEAK_OVERRIDE) ||
1139 			     (old & NUD_CONNECTED))
1140 			    )
1141 				new = old;
1142 		}
1143 	}
1144 
1145 	if (new != old) {
1146 		neigh_del_timer(neigh);
1147 		if (new & NUD_IN_TIMER)
1148 			neigh_add_timer(neigh, (jiffies +
1149 						((new & NUD_REACHABLE) ?
1150 						 neigh->parms->reachable_time :
1151 						 0)));
1152 		neigh->nud_state = new;
1153 		notify = 1;
1154 	}
1155 
1156 	if (lladdr != neigh->ha) {
1157 		write_seqlock(&neigh->ha_lock);
1158 		memcpy(&neigh->ha, lladdr, dev->addr_len);
1159 		write_sequnlock(&neigh->ha_lock);
1160 		neigh_update_hhs(neigh);
1161 		if (!(new & NUD_CONNECTED))
1162 			neigh->confirmed = jiffies -
1163 				      (NEIGH_VAR(neigh->parms, BASE_REACHABLE_TIME) << 1);
1164 		notify = 1;
1165 	}
1166 	if (new == old)
1167 		goto out;
1168 	if (new & NUD_CONNECTED)
1169 		neigh_connect(neigh);
1170 	else
1171 		neigh_suspect(neigh);
1172 	if (!(old & NUD_VALID)) {
1173 		struct sk_buff *skb;
1174 
1175 		/* Again: avoid dead loop if something went wrong */
1176 
1177 		while (neigh->nud_state & NUD_VALID &&
1178 		       (skb = __skb_dequeue(&neigh->arp_queue)) != NULL) {
1179 			struct dst_entry *dst = skb_dst(skb);
1180 			struct neighbour *n2, *n1 = neigh;
1181 			write_unlock_bh(&neigh->lock);
1182 
1183 			rcu_read_lock();
1184 
1185 			/* Why not just use 'neigh' as-is?  The problem is that
1186 			 * things such as shaper, eql, and sch_teql can end up
1187 			 * using alternative, different, neigh objects to output
1188 			 * the packet in the output path.  So what we need to do
1189 			 * here is re-lookup the top-level neigh in the path so
1190 			 * we can reinject the packet there.
1191 			 */
1192 			n2 = NULL;
1193 			if (dst) {
1194 				n2 = dst_neigh_lookup_skb(dst, skb);
1195 				if (n2)
1196 					n1 = n2;
1197 			}
1198 			n1->output(n1, skb);
1199 			if (n2)
1200 				neigh_release(n2);
1201 			rcu_read_unlock();
1202 
1203 			write_lock_bh(&neigh->lock);
1204 		}
1205 		__skb_queue_purge(&neigh->arp_queue);
1206 		neigh->arp_queue_len_bytes = 0;
1207 	}
1208 out:
1209 	if (update_isrouter) {
1210 		neigh->flags = (flags & NEIGH_UPDATE_F_ISROUTER) ?
1211 			(neigh->flags | NTF_ROUTER) :
1212 			(neigh->flags & ~NTF_ROUTER);
1213 	}
1214 	write_unlock_bh(&neigh->lock);
1215 
1216 	if (notify)
1217 		neigh_update_notify(neigh);
1218 
1219 	return err;
1220 }
1221 EXPORT_SYMBOL(neigh_update);
1222 
1223 /* Update the neigh to listen temporarily for probe responses, even if it is
1224  * in a NUD_FAILED state. The caller has to hold neigh->lock for writing.
1225  */
1226 void __neigh_set_probe_once(struct neighbour *neigh)
1227 {
1228 	neigh->updated = jiffies;
1229 	if (!(neigh->nud_state & NUD_FAILED))
1230 		return;
1231 	neigh->nud_state = NUD_INCOMPLETE;
1232 	atomic_set(&neigh->probes, neigh_max_probes(neigh));
1233 	neigh_add_timer(neigh,
1234 			jiffies + NEIGH_VAR(neigh->parms, RETRANS_TIME));
1235 }
1236 EXPORT_SYMBOL(__neigh_set_probe_once);
1237 
1238 struct neighbour *neigh_event_ns(struct neigh_table *tbl,
1239 				 u8 *lladdr, void *saddr,
1240 				 struct net_device *dev)
1241 {
1242 	struct neighbour *neigh = __neigh_lookup(tbl, saddr, dev,
1243 						 lladdr || !dev->addr_len);
1244 	if (neigh)
1245 		neigh_update(neigh, lladdr, NUD_STALE,
1246 			     NEIGH_UPDATE_F_OVERRIDE);
1247 	return neigh;
1248 }
1249 EXPORT_SYMBOL(neigh_event_ns);
1250 
1251 /* called with read_lock_bh(&n->lock); */
1252 static void neigh_hh_init(struct neighbour *n)
1253 {
1254 	struct net_device *dev = n->dev;
1255 	__be16 prot = n->tbl->protocol;
1256 	struct hh_cache	*hh = &n->hh;
1257 
1258 	write_lock_bh(&n->lock);
1259 
1260 	/* Only one thread can come in here and initialize the
1261 	 * hh_cache entry.
1262 	 */
1263 	if (!hh->hh_len)
1264 		dev->header_ops->cache(n, hh, prot);
1265 
1266 	write_unlock_bh(&n->lock);
1267 }
1268 
1269 /* Slow and careful. */
1270 
1271 int neigh_resolve_output(struct neighbour *neigh, struct sk_buff *skb)
1272 {
1273 	int rc = 0;
1274 
1275 	if (!neigh_event_send(neigh, skb)) {
1276 		int err;
1277 		struct net_device *dev = neigh->dev;
1278 		unsigned int seq;
1279 
1280 		if (dev->header_ops->cache && !neigh->hh.hh_len)
1281 			neigh_hh_init(neigh);
1282 
1283 		do {
1284 			__skb_pull(skb, skb_network_offset(skb));
1285 			seq = read_seqbegin(&neigh->ha_lock);
1286 			err = dev_hard_header(skb, dev, ntohs(skb->protocol),
1287 					      neigh->ha, NULL, skb->len);
1288 		} while (read_seqretry(&neigh->ha_lock, seq));
1289 
1290 		if (err >= 0)
1291 			rc = dev_queue_xmit(skb);
1292 		else
1293 			goto out_kfree_skb;
1294 	}
1295 out:
1296 	return rc;
1297 out_kfree_skb:
1298 	rc = -EINVAL;
1299 	kfree_skb(skb);
1300 	goto out;
1301 }
1302 EXPORT_SYMBOL(neigh_resolve_output);
1303 
1304 /* As fast as possible without hh cache */
1305 
1306 int neigh_connected_output(struct neighbour *neigh, struct sk_buff *skb)
1307 {
1308 	struct net_device *dev = neigh->dev;
1309 	unsigned int seq;
1310 	int err;
1311 
1312 	do {
1313 		__skb_pull(skb, skb_network_offset(skb));
1314 		seq = read_seqbegin(&neigh->ha_lock);
1315 		err = dev_hard_header(skb, dev, ntohs(skb->protocol),
1316 				      neigh->ha, NULL, skb->len);
1317 	} while (read_seqretry(&neigh->ha_lock, seq));
1318 
1319 	if (err >= 0)
1320 		err = dev_queue_xmit(skb);
1321 	else {
1322 		err = -EINVAL;
1323 		kfree_skb(skb);
1324 	}
1325 	return err;
1326 }
1327 EXPORT_SYMBOL(neigh_connected_output);
1328 
1329 int neigh_direct_output(struct neighbour *neigh, struct sk_buff *skb)
1330 {
1331 	return dev_queue_xmit(skb);
1332 }
1333 EXPORT_SYMBOL(neigh_direct_output);
1334 
1335 static void neigh_proxy_process(unsigned long arg)
1336 {
1337 	struct neigh_table *tbl = (struct neigh_table *)arg;
1338 	long sched_next = 0;
1339 	unsigned long now = jiffies;
1340 	struct sk_buff *skb, *n;
1341 
1342 	spin_lock(&tbl->proxy_queue.lock);
1343 
1344 	skb_queue_walk_safe(&tbl->proxy_queue, skb, n) {
1345 		long tdif = NEIGH_CB(skb)->sched_next - now;
1346 
1347 		if (tdif <= 0) {
1348 			struct net_device *dev = skb->dev;
1349 
1350 			__skb_unlink(skb, &tbl->proxy_queue);
1351 			if (tbl->proxy_redo && netif_running(dev)) {
1352 				rcu_read_lock();
1353 				tbl->proxy_redo(skb);
1354 				rcu_read_unlock();
1355 			} else {
1356 				kfree_skb(skb);
1357 			}
1358 
1359 			dev_put(dev);
1360 		} else if (!sched_next || tdif < sched_next)
1361 			sched_next = tdif;
1362 	}
1363 	del_timer(&tbl->proxy_timer);
1364 	if (sched_next)
1365 		mod_timer(&tbl->proxy_timer, jiffies + sched_next);
1366 	spin_unlock(&tbl->proxy_queue.lock);
1367 }
1368 
1369 void pneigh_enqueue(struct neigh_table *tbl, struct neigh_parms *p,
1370 		    struct sk_buff *skb)
1371 {
1372 	unsigned long now = jiffies;
1373 
1374 	unsigned long sched_next = now + (prandom_u32() %
1375 					  NEIGH_VAR(p, PROXY_DELAY));
1376 
1377 	if (tbl->proxy_queue.qlen > NEIGH_VAR(p, PROXY_QLEN)) {
1378 		kfree_skb(skb);
1379 		return;
1380 	}
1381 
1382 	NEIGH_CB(skb)->sched_next = sched_next;
1383 	NEIGH_CB(skb)->flags |= LOCALLY_ENQUEUED;
1384 
1385 	spin_lock(&tbl->proxy_queue.lock);
1386 	if (del_timer(&tbl->proxy_timer)) {
1387 		if (time_before(tbl->proxy_timer.expires, sched_next))
1388 			sched_next = tbl->proxy_timer.expires;
1389 	}
1390 	skb_dst_drop(skb);
1391 	dev_hold(skb->dev);
1392 	__skb_queue_tail(&tbl->proxy_queue, skb);
1393 	mod_timer(&tbl->proxy_timer, sched_next);
1394 	spin_unlock(&tbl->proxy_queue.lock);
1395 }
1396 EXPORT_SYMBOL(pneigh_enqueue);
1397 
1398 static inline struct neigh_parms *lookup_neigh_parms(struct neigh_table *tbl,
1399 						      struct net *net, int ifindex)
1400 {
1401 	struct neigh_parms *p;
1402 
1403 	list_for_each_entry(p, &tbl->parms_list, list) {
1404 		if ((p->dev && p->dev->ifindex == ifindex && net_eq(neigh_parms_net(p), net)) ||
1405 		    (!p->dev && !ifindex && net_eq(net, &init_net)))
1406 			return p;
1407 	}
1408 
1409 	return NULL;
1410 }
1411 
1412 struct neigh_parms *neigh_parms_alloc(struct net_device *dev,
1413 				      struct neigh_table *tbl)
1414 {
1415 	struct neigh_parms *p;
1416 	struct net *net = dev_net(dev);
1417 	const struct net_device_ops *ops = dev->netdev_ops;
1418 
1419 	p = kmemdup(&tbl->parms, sizeof(*p), GFP_KERNEL);
1420 	if (p) {
1421 		p->tbl		  = tbl;
1422 		atomic_set(&p->refcnt, 1);
1423 		p->reachable_time =
1424 				neigh_rand_reach_time(NEIGH_VAR(p, BASE_REACHABLE_TIME));
1425 		dev_hold(dev);
1426 		p->dev = dev;
1427 		write_pnet(&p->net, net);
1428 		p->sysctl_table = NULL;
1429 
1430 		if (ops->ndo_neigh_setup && ops->ndo_neigh_setup(dev, p)) {
1431 			dev_put(dev);
1432 			kfree(p);
1433 			return NULL;
1434 		}
1435 
1436 		write_lock_bh(&tbl->lock);
1437 		list_add(&p->list, &tbl->parms.list);
1438 		write_unlock_bh(&tbl->lock);
1439 
1440 		neigh_parms_data_state_cleanall(p);
1441 	}
1442 	return p;
1443 }
1444 EXPORT_SYMBOL(neigh_parms_alloc);
1445 
1446 static void neigh_rcu_free_parms(struct rcu_head *head)
1447 {
1448 	struct neigh_parms *parms =
1449 		container_of(head, struct neigh_parms, rcu_head);
1450 
1451 	neigh_parms_put(parms);
1452 }
1453 
1454 void neigh_parms_release(struct neigh_table *tbl, struct neigh_parms *parms)
1455 {
1456 	if (!parms || parms == &tbl->parms)
1457 		return;
1458 	write_lock_bh(&tbl->lock);
1459 	list_del(&parms->list);
1460 	parms->dead = 1;
1461 	write_unlock_bh(&tbl->lock);
1462 	if (parms->dev)
1463 		dev_put(parms->dev);
1464 	call_rcu(&parms->rcu_head, neigh_rcu_free_parms);
1465 }
1466 EXPORT_SYMBOL(neigh_parms_release);
1467 
1468 static void neigh_parms_destroy(struct neigh_parms *parms)
1469 {
1470 	kfree(parms);
1471 }
1472 
1473 static struct lock_class_key neigh_table_proxy_queue_class;
1474 
1475 static struct neigh_table *neigh_tables[NEIGH_NR_TABLES] __read_mostly;
1476 
1477 void neigh_table_init(int index, struct neigh_table *tbl)
1478 {
1479 	unsigned long now = jiffies;
1480 	unsigned long phsize;
1481 
1482 	INIT_LIST_HEAD(&tbl->parms_list);
1483 	list_add(&tbl->parms.list, &tbl->parms_list);
1484 	write_pnet(&tbl->parms.net, &init_net);
1485 	atomic_set(&tbl->parms.refcnt, 1);
1486 	tbl->parms.reachable_time =
1487 			  neigh_rand_reach_time(NEIGH_VAR(&tbl->parms, BASE_REACHABLE_TIME));
1488 
1489 	tbl->stats = alloc_percpu(struct neigh_statistics);
1490 	if (!tbl->stats)
1491 		panic("cannot create neighbour cache statistics");
1492 
1493 #ifdef CONFIG_PROC_FS
1494 	if (!proc_create_data(tbl->id, 0, init_net.proc_net_stat,
1495 			      &neigh_stat_seq_fops, tbl))
1496 		panic("cannot create neighbour proc dir entry");
1497 #endif
1498 
1499 	RCU_INIT_POINTER(tbl->nht, neigh_hash_alloc(3));
1500 
1501 	phsize = (PNEIGH_HASHMASK + 1) * sizeof(struct pneigh_entry *);
1502 	tbl->phash_buckets = kzalloc(phsize, GFP_KERNEL);
1503 
1504 	if (!tbl->nht || !tbl->phash_buckets)
1505 		panic("cannot allocate neighbour cache hashes");
1506 
1507 	if (!tbl->entry_size)
1508 		tbl->entry_size = ALIGN(offsetof(struct neighbour, primary_key) +
1509 					tbl->key_len, NEIGH_PRIV_ALIGN);
1510 	else
1511 		WARN_ON(tbl->entry_size % NEIGH_PRIV_ALIGN);
1512 
1513 	rwlock_init(&tbl->lock);
1514 	INIT_DEFERRABLE_WORK(&tbl->gc_work, neigh_periodic_work);
1515 	queue_delayed_work(system_power_efficient_wq, &tbl->gc_work,
1516 			tbl->parms.reachable_time);
1517 	setup_timer(&tbl->proxy_timer, neigh_proxy_process, (unsigned long)tbl);
1518 	skb_queue_head_init_class(&tbl->proxy_queue,
1519 			&neigh_table_proxy_queue_class);
1520 
1521 	tbl->last_flush = now;
1522 	tbl->last_rand	= now + tbl->parms.reachable_time * 20;
1523 
1524 	neigh_tables[index] = tbl;
1525 }
1526 EXPORT_SYMBOL(neigh_table_init);
1527 
1528 int neigh_table_clear(int index, struct neigh_table *tbl)
1529 {
1530 	neigh_tables[index] = NULL;
1531 	/* It is not clean... Fix it to unload IPv6 module safely */
1532 	cancel_delayed_work_sync(&tbl->gc_work);
1533 	del_timer_sync(&tbl->proxy_timer);
1534 	pneigh_queue_purge(&tbl->proxy_queue);
1535 	neigh_ifdown(tbl, NULL);
1536 	if (atomic_read(&tbl->entries))
1537 		pr_crit("neighbour leakage\n");
1538 
1539 	call_rcu(&rcu_dereference_protected(tbl->nht, 1)->rcu,
1540 		 neigh_hash_free_rcu);
1541 	tbl->nht = NULL;
1542 
1543 	kfree(tbl->phash_buckets);
1544 	tbl->phash_buckets = NULL;
1545 
1546 	remove_proc_entry(tbl->id, init_net.proc_net_stat);
1547 
1548 	free_percpu(tbl->stats);
1549 	tbl->stats = NULL;
1550 
1551 	return 0;
1552 }
1553 EXPORT_SYMBOL(neigh_table_clear);
1554 
1555 static struct neigh_table *neigh_find_table(int family)
1556 {
1557 	struct neigh_table *tbl = NULL;
1558 
1559 	switch (family) {
1560 	case AF_INET:
1561 		tbl = neigh_tables[NEIGH_ARP_TABLE];
1562 		break;
1563 	case AF_INET6:
1564 		tbl = neigh_tables[NEIGH_ND_TABLE];
1565 		break;
1566 	case AF_DECnet:
1567 		tbl = neigh_tables[NEIGH_DN_TABLE];
1568 		break;
1569 	}
1570 
1571 	return tbl;
1572 }
1573 
1574 static int neigh_delete(struct sk_buff *skb, struct nlmsghdr *nlh)
1575 {
1576 	struct net *net = sock_net(skb->sk);
1577 	struct ndmsg *ndm;
1578 	struct nlattr *dst_attr;
1579 	struct neigh_table *tbl;
1580 	struct neighbour *neigh;
1581 	struct net_device *dev = NULL;
1582 	int err = -EINVAL;
1583 
1584 	ASSERT_RTNL();
1585 	if (nlmsg_len(nlh) < sizeof(*ndm))
1586 		goto out;
1587 
1588 	dst_attr = nlmsg_find_attr(nlh, sizeof(*ndm), NDA_DST);
1589 	if (dst_attr == NULL)
1590 		goto out;
1591 
1592 	ndm = nlmsg_data(nlh);
1593 	if (ndm->ndm_ifindex) {
1594 		dev = __dev_get_by_index(net, ndm->ndm_ifindex);
1595 		if (dev == NULL) {
1596 			err = -ENODEV;
1597 			goto out;
1598 		}
1599 	}
1600 
1601 	tbl = neigh_find_table(ndm->ndm_family);
1602 	if (tbl == NULL)
1603 		return -EAFNOSUPPORT;
1604 
1605 	if (nla_len(dst_attr) < tbl->key_len)
1606 		goto out;
1607 
1608 	if (ndm->ndm_flags & NTF_PROXY) {
1609 		err = pneigh_delete(tbl, net, nla_data(dst_attr), dev);
1610 		goto out;
1611 	}
1612 
1613 	if (dev == NULL)
1614 		goto out;
1615 
1616 	neigh = neigh_lookup(tbl, nla_data(dst_attr), dev);
1617 	if (neigh == NULL) {
1618 		err = -ENOENT;
1619 		goto out;
1620 	}
1621 
1622 	err = neigh_update(neigh, NULL, NUD_FAILED,
1623 			   NEIGH_UPDATE_F_OVERRIDE |
1624 			   NEIGH_UPDATE_F_ADMIN);
1625 	neigh_release(neigh);
1626 
1627 out:
1628 	return err;
1629 }
1630 
1631 static int neigh_add(struct sk_buff *skb, struct nlmsghdr *nlh)
1632 {
1633 	int flags = NEIGH_UPDATE_F_ADMIN | NEIGH_UPDATE_F_OVERRIDE;
1634 	struct net *net = sock_net(skb->sk);
1635 	struct ndmsg *ndm;
1636 	struct nlattr *tb[NDA_MAX+1];
1637 	struct neigh_table *tbl;
1638 	struct net_device *dev = NULL;
1639 	struct neighbour *neigh;
1640 	void *dst, *lladdr;
1641 	int err;
1642 
1643 	ASSERT_RTNL();
1644 	err = nlmsg_parse(nlh, sizeof(*ndm), tb, NDA_MAX, NULL);
1645 	if (err < 0)
1646 		goto out;
1647 
1648 	err = -EINVAL;
1649 	if (tb[NDA_DST] == NULL)
1650 		goto out;
1651 
1652 	ndm = nlmsg_data(nlh);
1653 	if (ndm->ndm_ifindex) {
1654 		dev = __dev_get_by_index(net, ndm->ndm_ifindex);
1655 		if (dev == NULL) {
1656 			err = -ENODEV;
1657 			goto out;
1658 		}
1659 
1660 		if (tb[NDA_LLADDR] && nla_len(tb[NDA_LLADDR]) < dev->addr_len)
1661 			goto out;
1662 	}
1663 
1664 	tbl = neigh_find_table(ndm->ndm_family);
1665 	if (tbl == NULL)
1666 		return -EAFNOSUPPORT;
1667 
1668 	if (nla_len(tb[NDA_DST]) < tbl->key_len)
1669 		goto out;
1670 	dst = nla_data(tb[NDA_DST]);
1671 	lladdr = tb[NDA_LLADDR] ? nla_data(tb[NDA_LLADDR]) : NULL;
1672 
1673 	if (ndm->ndm_flags & NTF_PROXY) {
1674 		struct pneigh_entry *pn;
1675 
1676 		err = -ENOBUFS;
1677 		pn = pneigh_lookup(tbl, net, dst, dev, 1);
1678 		if (pn) {
1679 			pn->flags = ndm->ndm_flags;
1680 			err = 0;
1681 		}
1682 		goto out;
1683 	}
1684 
1685 	if (dev == NULL)
1686 		goto out;
1687 
1688 	neigh = neigh_lookup(tbl, dst, dev);
1689 	if (neigh == NULL) {
1690 		if (!(nlh->nlmsg_flags & NLM_F_CREATE)) {
1691 			err = -ENOENT;
1692 			goto out;
1693 		}
1694 
1695 		neigh = __neigh_lookup_errno(tbl, dst, dev);
1696 		if (IS_ERR(neigh)) {
1697 			err = PTR_ERR(neigh);
1698 			goto out;
1699 		}
1700 	} else {
1701 		if (nlh->nlmsg_flags & NLM_F_EXCL) {
1702 			err = -EEXIST;
1703 			neigh_release(neigh);
1704 			goto out;
1705 		}
1706 
1707 		if (!(nlh->nlmsg_flags & NLM_F_REPLACE))
1708 			flags &= ~NEIGH_UPDATE_F_OVERRIDE;
1709 	}
1710 
1711 	if (ndm->ndm_flags & NTF_USE) {
1712 		neigh_event_send(neigh, NULL);
1713 		err = 0;
1714 	} else
1715 		err = neigh_update(neigh, lladdr, ndm->ndm_state, flags);
1716 	neigh_release(neigh);
1717 
1718 out:
1719 	return err;
1720 }
1721 
1722 static int neightbl_fill_parms(struct sk_buff *skb, struct neigh_parms *parms)
1723 {
1724 	struct nlattr *nest;
1725 
1726 	nest = nla_nest_start(skb, NDTA_PARMS);
1727 	if (nest == NULL)
1728 		return -ENOBUFS;
1729 
1730 	if ((parms->dev &&
1731 	     nla_put_u32(skb, NDTPA_IFINDEX, parms->dev->ifindex)) ||
1732 	    nla_put_u32(skb, NDTPA_REFCNT, atomic_read(&parms->refcnt)) ||
1733 	    nla_put_u32(skb, NDTPA_QUEUE_LENBYTES,
1734 			NEIGH_VAR(parms, QUEUE_LEN_BYTES)) ||
1735 	    /* approximative value for deprecated QUEUE_LEN (in packets) */
1736 	    nla_put_u32(skb, NDTPA_QUEUE_LEN,
1737 			NEIGH_VAR(parms, QUEUE_LEN_BYTES) / SKB_TRUESIZE(ETH_FRAME_LEN)) ||
1738 	    nla_put_u32(skb, NDTPA_PROXY_QLEN, NEIGH_VAR(parms, PROXY_QLEN)) ||
1739 	    nla_put_u32(skb, NDTPA_APP_PROBES, NEIGH_VAR(parms, APP_PROBES)) ||
1740 	    nla_put_u32(skb, NDTPA_UCAST_PROBES,
1741 			NEIGH_VAR(parms, UCAST_PROBES)) ||
1742 	    nla_put_u32(skb, NDTPA_MCAST_PROBES,
1743 			NEIGH_VAR(parms, MCAST_PROBES)) ||
1744 	    nla_put_u32(skb, NDTPA_MCAST_REPROBES,
1745 			NEIGH_VAR(parms, MCAST_REPROBES)) ||
1746 	    nla_put_msecs(skb, NDTPA_REACHABLE_TIME, parms->reachable_time) ||
1747 	    nla_put_msecs(skb, NDTPA_BASE_REACHABLE_TIME,
1748 			  NEIGH_VAR(parms, BASE_REACHABLE_TIME)) ||
1749 	    nla_put_msecs(skb, NDTPA_GC_STALETIME,
1750 			  NEIGH_VAR(parms, GC_STALETIME)) ||
1751 	    nla_put_msecs(skb, NDTPA_DELAY_PROBE_TIME,
1752 			  NEIGH_VAR(parms, DELAY_PROBE_TIME)) ||
1753 	    nla_put_msecs(skb, NDTPA_RETRANS_TIME,
1754 			  NEIGH_VAR(parms, RETRANS_TIME)) ||
1755 	    nla_put_msecs(skb, NDTPA_ANYCAST_DELAY,
1756 			  NEIGH_VAR(parms, ANYCAST_DELAY)) ||
1757 	    nla_put_msecs(skb, NDTPA_PROXY_DELAY,
1758 			  NEIGH_VAR(parms, PROXY_DELAY)) ||
1759 	    nla_put_msecs(skb, NDTPA_LOCKTIME,
1760 			  NEIGH_VAR(parms, LOCKTIME)))
1761 		goto nla_put_failure;
1762 	return nla_nest_end(skb, nest);
1763 
1764 nla_put_failure:
1765 	nla_nest_cancel(skb, nest);
1766 	return -EMSGSIZE;
1767 }
1768 
1769 static int neightbl_fill_info(struct sk_buff *skb, struct neigh_table *tbl,
1770 			      u32 pid, u32 seq, int type, int flags)
1771 {
1772 	struct nlmsghdr *nlh;
1773 	struct ndtmsg *ndtmsg;
1774 
1775 	nlh = nlmsg_put(skb, pid, seq, type, sizeof(*ndtmsg), flags);
1776 	if (nlh == NULL)
1777 		return -EMSGSIZE;
1778 
1779 	ndtmsg = nlmsg_data(nlh);
1780 
1781 	read_lock_bh(&tbl->lock);
1782 	ndtmsg->ndtm_family = tbl->family;
1783 	ndtmsg->ndtm_pad1   = 0;
1784 	ndtmsg->ndtm_pad2   = 0;
1785 
1786 	if (nla_put_string(skb, NDTA_NAME, tbl->id) ||
1787 	    nla_put_msecs(skb, NDTA_GC_INTERVAL, tbl->gc_interval) ||
1788 	    nla_put_u32(skb, NDTA_THRESH1, tbl->gc_thresh1) ||
1789 	    nla_put_u32(skb, NDTA_THRESH2, tbl->gc_thresh2) ||
1790 	    nla_put_u32(skb, NDTA_THRESH3, tbl->gc_thresh3))
1791 		goto nla_put_failure;
1792 	{
1793 		unsigned long now = jiffies;
1794 		unsigned int flush_delta = now - tbl->last_flush;
1795 		unsigned int rand_delta = now - tbl->last_rand;
1796 		struct neigh_hash_table *nht;
1797 		struct ndt_config ndc = {
1798 			.ndtc_key_len		= tbl->key_len,
1799 			.ndtc_entry_size	= tbl->entry_size,
1800 			.ndtc_entries		= atomic_read(&tbl->entries),
1801 			.ndtc_last_flush	= jiffies_to_msecs(flush_delta),
1802 			.ndtc_last_rand		= jiffies_to_msecs(rand_delta),
1803 			.ndtc_proxy_qlen	= tbl->proxy_queue.qlen,
1804 		};
1805 
1806 		rcu_read_lock_bh();
1807 		nht = rcu_dereference_bh(tbl->nht);
1808 		ndc.ndtc_hash_rnd = nht->hash_rnd[0];
1809 		ndc.ndtc_hash_mask = ((1 << nht->hash_shift) - 1);
1810 		rcu_read_unlock_bh();
1811 
1812 		if (nla_put(skb, NDTA_CONFIG, sizeof(ndc), &ndc))
1813 			goto nla_put_failure;
1814 	}
1815 
1816 	{
1817 		int cpu;
1818 		struct ndt_stats ndst;
1819 
1820 		memset(&ndst, 0, sizeof(ndst));
1821 
1822 		for_each_possible_cpu(cpu) {
1823 			struct neigh_statistics	*st;
1824 
1825 			st = per_cpu_ptr(tbl->stats, cpu);
1826 			ndst.ndts_allocs		+= st->allocs;
1827 			ndst.ndts_destroys		+= st->destroys;
1828 			ndst.ndts_hash_grows		+= st->hash_grows;
1829 			ndst.ndts_res_failed		+= st->res_failed;
1830 			ndst.ndts_lookups		+= st->lookups;
1831 			ndst.ndts_hits			+= st->hits;
1832 			ndst.ndts_rcv_probes_mcast	+= st->rcv_probes_mcast;
1833 			ndst.ndts_rcv_probes_ucast	+= st->rcv_probes_ucast;
1834 			ndst.ndts_periodic_gc_runs	+= st->periodic_gc_runs;
1835 			ndst.ndts_forced_gc_runs	+= st->forced_gc_runs;
1836 		}
1837 
1838 		if (nla_put(skb, NDTA_STATS, sizeof(ndst), &ndst))
1839 			goto nla_put_failure;
1840 	}
1841 
1842 	BUG_ON(tbl->parms.dev);
1843 	if (neightbl_fill_parms(skb, &tbl->parms) < 0)
1844 		goto nla_put_failure;
1845 
1846 	read_unlock_bh(&tbl->lock);
1847 	nlmsg_end(skb, nlh);
1848 	return 0;
1849 
1850 nla_put_failure:
1851 	read_unlock_bh(&tbl->lock);
1852 	nlmsg_cancel(skb, nlh);
1853 	return -EMSGSIZE;
1854 }
1855 
1856 static int neightbl_fill_param_info(struct sk_buff *skb,
1857 				    struct neigh_table *tbl,
1858 				    struct neigh_parms *parms,
1859 				    u32 pid, u32 seq, int type,
1860 				    unsigned int flags)
1861 {
1862 	struct ndtmsg *ndtmsg;
1863 	struct nlmsghdr *nlh;
1864 
1865 	nlh = nlmsg_put(skb, pid, seq, type, sizeof(*ndtmsg), flags);
1866 	if (nlh == NULL)
1867 		return -EMSGSIZE;
1868 
1869 	ndtmsg = nlmsg_data(nlh);
1870 
1871 	read_lock_bh(&tbl->lock);
1872 	ndtmsg->ndtm_family = tbl->family;
1873 	ndtmsg->ndtm_pad1   = 0;
1874 	ndtmsg->ndtm_pad2   = 0;
1875 
1876 	if (nla_put_string(skb, NDTA_NAME, tbl->id) < 0 ||
1877 	    neightbl_fill_parms(skb, parms) < 0)
1878 		goto errout;
1879 
1880 	read_unlock_bh(&tbl->lock);
1881 	nlmsg_end(skb, nlh);
1882 	return 0;
1883 errout:
1884 	read_unlock_bh(&tbl->lock);
1885 	nlmsg_cancel(skb, nlh);
1886 	return -EMSGSIZE;
1887 }
1888 
1889 static const struct nla_policy nl_neightbl_policy[NDTA_MAX+1] = {
1890 	[NDTA_NAME]		= { .type = NLA_STRING },
1891 	[NDTA_THRESH1]		= { .type = NLA_U32 },
1892 	[NDTA_THRESH2]		= { .type = NLA_U32 },
1893 	[NDTA_THRESH3]		= { .type = NLA_U32 },
1894 	[NDTA_GC_INTERVAL]	= { .type = NLA_U64 },
1895 	[NDTA_PARMS]		= { .type = NLA_NESTED },
1896 };
1897 
1898 static const struct nla_policy nl_ntbl_parm_policy[NDTPA_MAX+1] = {
1899 	[NDTPA_IFINDEX]			= { .type = NLA_U32 },
1900 	[NDTPA_QUEUE_LEN]		= { .type = NLA_U32 },
1901 	[NDTPA_PROXY_QLEN]		= { .type = NLA_U32 },
1902 	[NDTPA_APP_PROBES]		= { .type = NLA_U32 },
1903 	[NDTPA_UCAST_PROBES]		= { .type = NLA_U32 },
1904 	[NDTPA_MCAST_PROBES]		= { .type = NLA_U32 },
1905 	[NDTPA_MCAST_REPROBES]		= { .type = NLA_U32 },
1906 	[NDTPA_BASE_REACHABLE_TIME]	= { .type = NLA_U64 },
1907 	[NDTPA_GC_STALETIME]		= { .type = NLA_U64 },
1908 	[NDTPA_DELAY_PROBE_TIME]	= { .type = NLA_U64 },
1909 	[NDTPA_RETRANS_TIME]		= { .type = NLA_U64 },
1910 	[NDTPA_ANYCAST_DELAY]		= { .type = NLA_U64 },
1911 	[NDTPA_PROXY_DELAY]		= { .type = NLA_U64 },
1912 	[NDTPA_LOCKTIME]		= { .type = NLA_U64 },
1913 };
1914 
1915 static int neightbl_set(struct sk_buff *skb, struct nlmsghdr *nlh)
1916 {
1917 	struct net *net = sock_net(skb->sk);
1918 	struct neigh_table *tbl;
1919 	struct ndtmsg *ndtmsg;
1920 	struct nlattr *tb[NDTA_MAX+1];
1921 	bool found = false;
1922 	int err, tidx;
1923 
1924 	err = nlmsg_parse(nlh, sizeof(*ndtmsg), tb, NDTA_MAX,
1925 			  nl_neightbl_policy);
1926 	if (err < 0)
1927 		goto errout;
1928 
1929 	if (tb[NDTA_NAME] == NULL) {
1930 		err = -EINVAL;
1931 		goto errout;
1932 	}
1933 
1934 	ndtmsg = nlmsg_data(nlh);
1935 
1936 	for (tidx = 0; tidx < NEIGH_NR_TABLES; tidx++) {
1937 		tbl = neigh_tables[tidx];
1938 		if (!tbl)
1939 			continue;
1940 		if (ndtmsg->ndtm_family && tbl->family != ndtmsg->ndtm_family)
1941 			continue;
1942 		if (nla_strcmp(tb[NDTA_NAME], tbl->id) == 0) {
1943 			found = true;
1944 			break;
1945 		}
1946 	}
1947 
1948 	if (!found)
1949 		return -ENOENT;
1950 
1951 	/*
1952 	 * We acquire tbl->lock to be nice to the periodic timers and
1953 	 * make sure they always see a consistent set of values.
1954 	 */
1955 	write_lock_bh(&tbl->lock);
1956 
1957 	if (tb[NDTA_PARMS]) {
1958 		struct nlattr *tbp[NDTPA_MAX+1];
1959 		struct neigh_parms *p;
1960 		int i, ifindex = 0;
1961 
1962 		err = nla_parse_nested(tbp, NDTPA_MAX, tb[NDTA_PARMS],
1963 				       nl_ntbl_parm_policy);
1964 		if (err < 0)
1965 			goto errout_tbl_lock;
1966 
1967 		if (tbp[NDTPA_IFINDEX])
1968 			ifindex = nla_get_u32(tbp[NDTPA_IFINDEX]);
1969 
1970 		p = lookup_neigh_parms(tbl, net, ifindex);
1971 		if (p == NULL) {
1972 			err = -ENOENT;
1973 			goto errout_tbl_lock;
1974 		}
1975 
1976 		for (i = 1; i <= NDTPA_MAX; i++) {
1977 			if (tbp[i] == NULL)
1978 				continue;
1979 
1980 			switch (i) {
1981 			case NDTPA_QUEUE_LEN:
1982 				NEIGH_VAR_SET(p, QUEUE_LEN_BYTES,
1983 					      nla_get_u32(tbp[i]) *
1984 					      SKB_TRUESIZE(ETH_FRAME_LEN));
1985 				break;
1986 			case NDTPA_QUEUE_LENBYTES:
1987 				NEIGH_VAR_SET(p, QUEUE_LEN_BYTES,
1988 					      nla_get_u32(tbp[i]));
1989 				break;
1990 			case NDTPA_PROXY_QLEN:
1991 				NEIGH_VAR_SET(p, PROXY_QLEN,
1992 					      nla_get_u32(tbp[i]));
1993 				break;
1994 			case NDTPA_APP_PROBES:
1995 				NEIGH_VAR_SET(p, APP_PROBES,
1996 					      nla_get_u32(tbp[i]));
1997 				break;
1998 			case NDTPA_UCAST_PROBES:
1999 				NEIGH_VAR_SET(p, UCAST_PROBES,
2000 					      nla_get_u32(tbp[i]));
2001 				break;
2002 			case NDTPA_MCAST_PROBES:
2003 				NEIGH_VAR_SET(p, MCAST_PROBES,
2004 					      nla_get_u32(tbp[i]));
2005 				break;
2006 			case NDTPA_MCAST_REPROBES:
2007 				NEIGH_VAR_SET(p, MCAST_REPROBES,
2008 					      nla_get_u32(tbp[i]));
2009 				break;
2010 			case NDTPA_BASE_REACHABLE_TIME:
2011 				NEIGH_VAR_SET(p, BASE_REACHABLE_TIME,
2012 					      nla_get_msecs(tbp[i]));
2013 				/* update reachable_time as well, otherwise, the change will
2014 				 * only be effective after the next time neigh_periodic_work
2015 				 * decides to recompute it (can be multiple minutes)
2016 				 */
2017 				p->reachable_time =
2018 					neigh_rand_reach_time(NEIGH_VAR(p, BASE_REACHABLE_TIME));
2019 				break;
2020 			case NDTPA_GC_STALETIME:
2021 				NEIGH_VAR_SET(p, GC_STALETIME,
2022 					      nla_get_msecs(tbp[i]));
2023 				break;
2024 			case NDTPA_DELAY_PROBE_TIME:
2025 				NEIGH_VAR_SET(p, DELAY_PROBE_TIME,
2026 					      nla_get_msecs(tbp[i]));
2027 				break;
2028 			case NDTPA_RETRANS_TIME:
2029 				NEIGH_VAR_SET(p, RETRANS_TIME,
2030 					      nla_get_msecs(tbp[i]));
2031 				break;
2032 			case NDTPA_ANYCAST_DELAY:
2033 				NEIGH_VAR_SET(p, ANYCAST_DELAY,
2034 					      nla_get_msecs(tbp[i]));
2035 				break;
2036 			case NDTPA_PROXY_DELAY:
2037 				NEIGH_VAR_SET(p, PROXY_DELAY,
2038 					      nla_get_msecs(tbp[i]));
2039 				break;
2040 			case NDTPA_LOCKTIME:
2041 				NEIGH_VAR_SET(p, LOCKTIME,
2042 					      nla_get_msecs(tbp[i]));
2043 				break;
2044 			}
2045 		}
2046 	}
2047 
2048 	err = -ENOENT;
2049 	if ((tb[NDTA_THRESH1] || tb[NDTA_THRESH2] ||
2050 	     tb[NDTA_THRESH3] || tb[NDTA_GC_INTERVAL]) &&
2051 	    !net_eq(net, &init_net))
2052 		goto errout_tbl_lock;
2053 
2054 	if (tb[NDTA_THRESH1])
2055 		tbl->gc_thresh1 = nla_get_u32(tb[NDTA_THRESH1]);
2056 
2057 	if (tb[NDTA_THRESH2])
2058 		tbl->gc_thresh2 = nla_get_u32(tb[NDTA_THRESH2]);
2059 
2060 	if (tb[NDTA_THRESH3])
2061 		tbl->gc_thresh3 = nla_get_u32(tb[NDTA_THRESH3]);
2062 
2063 	if (tb[NDTA_GC_INTERVAL])
2064 		tbl->gc_interval = nla_get_msecs(tb[NDTA_GC_INTERVAL]);
2065 
2066 	err = 0;
2067 
2068 errout_tbl_lock:
2069 	write_unlock_bh(&tbl->lock);
2070 errout:
2071 	return err;
2072 }
2073 
2074 static int neightbl_dump_info(struct sk_buff *skb, struct netlink_callback *cb)
2075 {
2076 	struct net *net = sock_net(skb->sk);
2077 	int family, tidx, nidx = 0;
2078 	int tbl_skip = cb->args[0];
2079 	int neigh_skip = cb->args[1];
2080 	struct neigh_table *tbl;
2081 
2082 	family = ((struct rtgenmsg *) nlmsg_data(cb->nlh))->rtgen_family;
2083 
2084 	for (tidx = 0; tidx < NEIGH_NR_TABLES; tidx++) {
2085 		struct neigh_parms *p;
2086 
2087 		tbl = neigh_tables[tidx];
2088 		if (!tbl)
2089 			continue;
2090 
2091 		if (tidx < tbl_skip || (family && tbl->family != family))
2092 			continue;
2093 
2094 		if (neightbl_fill_info(skb, tbl, NETLINK_CB(cb->skb).portid,
2095 				       cb->nlh->nlmsg_seq, RTM_NEWNEIGHTBL,
2096 				       NLM_F_MULTI) < 0)
2097 			break;
2098 
2099 		nidx = 0;
2100 		p = list_next_entry(&tbl->parms, list);
2101 		list_for_each_entry_from(p, &tbl->parms_list, list) {
2102 			if (!net_eq(neigh_parms_net(p), net))
2103 				continue;
2104 
2105 			if (nidx < neigh_skip)
2106 				goto next;
2107 
2108 			if (neightbl_fill_param_info(skb, tbl, p,
2109 						     NETLINK_CB(cb->skb).portid,
2110 						     cb->nlh->nlmsg_seq,
2111 						     RTM_NEWNEIGHTBL,
2112 						     NLM_F_MULTI) < 0)
2113 				goto out;
2114 		next:
2115 			nidx++;
2116 		}
2117 
2118 		neigh_skip = 0;
2119 	}
2120 out:
2121 	cb->args[0] = tidx;
2122 	cb->args[1] = nidx;
2123 
2124 	return skb->len;
2125 }
2126 
2127 static int neigh_fill_info(struct sk_buff *skb, struct neighbour *neigh,
2128 			   u32 pid, u32 seq, int type, unsigned int flags)
2129 {
2130 	unsigned long now = jiffies;
2131 	struct nda_cacheinfo ci;
2132 	struct nlmsghdr *nlh;
2133 	struct ndmsg *ndm;
2134 
2135 	nlh = nlmsg_put(skb, pid, seq, type, sizeof(*ndm), flags);
2136 	if (nlh == NULL)
2137 		return -EMSGSIZE;
2138 
2139 	ndm = nlmsg_data(nlh);
2140 	ndm->ndm_family	 = neigh->ops->family;
2141 	ndm->ndm_pad1    = 0;
2142 	ndm->ndm_pad2    = 0;
2143 	ndm->ndm_flags	 = neigh->flags;
2144 	ndm->ndm_type	 = neigh->type;
2145 	ndm->ndm_ifindex = neigh->dev->ifindex;
2146 
2147 	if (nla_put(skb, NDA_DST, neigh->tbl->key_len, neigh->primary_key))
2148 		goto nla_put_failure;
2149 
2150 	read_lock_bh(&neigh->lock);
2151 	ndm->ndm_state	 = neigh->nud_state;
2152 	if (neigh->nud_state & NUD_VALID) {
2153 		char haddr[MAX_ADDR_LEN];
2154 
2155 		neigh_ha_snapshot(haddr, neigh, neigh->dev);
2156 		if (nla_put(skb, NDA_LLADDR, neigh->dev->addr_len, haddr) < 0) {
2157 			read_unlock_bh(&neigh->lock);
2158 			goto nla_put_failure;
2159 		}
2160 	}
2161 
2162 	ci.ndm_used	 = jiffies_to_clock_t(now - neigh->used);
2163 	ci.ndm_confirmed = jiffies_to_clock_t(now - neigh->confirmed);
2164 	ci.ndm_updated	 = jiffies_to_clock_t(now - neigh->updated);
2165 	ci.ndm_refcnt	 = atomic_read(&neigh->refcnt) - 1;
2166 	read_unlock_bh(&neigh->lock);
2167 
2168 	if (nla_put_u32(skb, NDA_PROBES, atomic_read(&neigh->probes)) ||
2169 	    nla_put(skb, NDA_CACHEINFO, sizeof(ci), &ci))
2170 		goto nla_put_failure;
2171 
2172 	nlmsg_end(skb, nlh);
2173 	return 0;
2174 
2175 nla_put_failure:
2176 	nlmsg_cancel(skb, nlh);
2177 	return -EMSGSIZE;
2178 }
2179 
2180 static int pneigh_fill_info(struct sk_buff *skb, struct pneigh_entry *pn,
2181 			    u32 pid, u32 seq, int type, unsigned int flags,
2182 			    struct neigh_table *tbl)
2183 {
2184 	struct nlmsghdr *nlh;
2185 	struct ndmsg *ndm;
2186 
2187 	nlh = nlmsg_put(skb, pid, seq, type, sizeof(*ndm), flags);
2188 	if (nlh == NULL)
2189 		return -EMSGSIZE;
2190 
2191 	ndm = nlmsg_data(nlh);
2192 	ndm->ndm_family	 = tbl->family;
2193 	ndm->ndm_pad1    = 0;
2194 	ndm->ndm_pad2    = 0;
2195 	ndm->ndm_flags	 = pn->flags | NTF_PROXY;
2196 	ndm->ndm_type	 = RTN_UNICAST;
2197 	ndm->ndm_ifindex = pn->dev->ifindex;
2198 	ndm->ndm_state	 = NUD_NONE;
2199 
2200 	if (nla_put(skb, NDA_DST, tbl->key_len, pn->key))
2201 		goto nla_put_failure;
2202 
2203 	nlmsg_end(skb, nlh);
2204 	return 0;
2205 
2206 nla_put_failure:
2207 	nlmsg_cancel(skb, nlh);
2208 	return -EMSGSIZE;
2209 }
2210 
2211 static void neigh_update_notify(struct neighbour *neigh)
2212 {
2213 	call_netevent_notifiers(NETEVENT_NEIGH_UPDATE, neigh);
2214 	__neigh_notify(neigh, RTM_NEWNEIGH, 0);
2215 }
2216 
2217 static int neigh_dump_table(struct neigh_table *tbl, struct sk_buff *skb,
2218 			    struct netlink_callback *cb)
2219 {
2220 	struct net *net = sock_net(skb->sk);
2221 	struct neighbour *n;
2222 	int rc, h, s_h = cb->args[1];
2223 	int idx, s_idx = idx = cb->args[2];
2224 	struct neigh_hash_table *nht;
2225 
2226 	rcu_read_lock_bh();
2227 	nht = rcu_dereference_bh(tbl->nht);
2228 
2229 	for (h = s_h; h < (1 << nht->hash_shift); h++) {
2230 		if (h > s_h)
2231 			s_idx = 0;
2232 		for (n = rcu_dereference_bh(nht->hash_buckets[h]), idx = 0;
2233 		     n != NULL;
2234 		     n = rcu_dereference_bh(n->next)) {
2235 			if (!net_eq(dev_net(n->dev), net))
2236 				continue;
2237 			if (idx < s_idx)
2238 				goto next;
2239 			if (neigh_fill_info(skb, n, NETLINK_CB(cb->skb).portid,
2240 					    cb->nlh->nlmsg_seq,
2241 					    RTM_NEWNEIGH,
2242 					    NLM_F_MULTI) < 0) {
2243 				rc = -1;
2244 				goto out;
2245 			}
2246 next:
2247 			idx++;
2248 		}
2249 	}
2250 	rc = skb->len;
2251 out:
2252 	rcu_read_unlock_bh();
2253 	cb->args[1] = h;
2254 	cb->args[2] = idx;
2255 	return rc;
2256 }
2257 
2258 static int pneigh_dump_table(struct neigh_table *tbl, struct sk_buff *skb,
2259 			     struct netlink_callback *cb)
2260 {
2261 	struct pneigh_entry *n;
2262 	struct net *net = sock_net(skb->sk);
2263 	int rc, h, s_h = cb->args[3];
2264 	int idx, s_idx = idx = cb->args[4];
2265 
2266 	read_lock_bh(&tbl->lock);
2267 
2268 	for (h = s_h; h <= PNEIGH_HASHMASK; h++) {
2269 		if (h > s_h)
2270 			s_idx = 0;
2271 		for (n = tbl->phash_buckets[h], idx = 0; n; n = n->next) {
2272 			if (dev_net(n->dev) != net)
2273 				continue;
2274 			if (idx < s_idx)
2275 				goto next;
2276 			if (pneigh_fill_info(skb, n, NETLINK_CB(cb->skb).portid,
2277 					    cb->nlh->nlmsg_seq,
2278 					    RTM_NEWNEIGH,
2279 					    NLM_F_MULTI, tbl) < 0) {
2280 				read_unlock_bh(&tbl->lock);
2281 				rc = -1;
2282 				goto out;
2283 			}
2284 		next:
2285 			idx++;
2286 		}
2287 	}
2288 
2289 	read_unlock_bh(&tbl->lock);
2290 	rc = skb->len;
2291 out:
2292 	cb->args[3] = h;
2293 	cb->args[4] = idx;
2294 	return rc;
2295 
2296 }
2297 
2298 static int neigh_dump_info(struct sk_buff *skb, struct netlink_callback *cb)
2299 {
2300 	struct neigh_table *tbl;
2301 	int t, family, s_t;
2302 	int proxy = 0;
2303 	int err;
2304 
2305 	family = ((struct rtgenmsg *) nlmsg_data(cb->nlh))->rtgen_family;
2306 
2307 	/* check for full ndmsg structure presence, family member is
2308 	 * the same for both structures
2309 	 */
2310 	if (nlmsg_len(cb->nlh) >= sizeof(struct ndmsg) &&
2311 	    ((struct ndmsg *) nlmsg_data(cb->nlh))->ndm_flags == NTF_PROXY)
2312 		proxy = 1;
2313 
2314 	s_t = cb->args[0];
2315 
2316 	for (t = 0; t < NEIGH_NR_TABLES; t++) {
2317 		tbl = neigh_tables[t];
2318 
2319 		if (!tbl)
2320 			continue;
2321 		if (t < s_t || (family && tbl->family != family))
2322 			continue;
2323 		if (t > s_t)
2324 			memset(&cb->args[1], 0, sizeof(cb->args) -
2325 						sizeof(cb->args[0]));
2326 		if (proxy)
2327 			err = pneigh_dump_table(tbl, skb, cb);
2328 		else
2329 			err = neigh_dump_table(tbl, skb, cb);
2330 		if (err < 0)
2331 			break;
2332 	}
2333 
2334 	cb->args[0] = t;
2335 	return skb->len;
2336 }
2337 
2338 void neigh_for_each(struct neigh_table *tbl, void (*cb)(struct neighbour *, void *), void *cookie)
2339 {
2340 	int chain;
2341 	struct neigh_hash_table *nht;
2342 
2343 	rcu_read_lock_bh();
2344 	nht = rcu_dereference_bh(tbl->nht);
2345 
2346 	read_lock(&tbl->lock); /* avoid resizes */
2347 	for (chain = 0; chain < (1 << nht->hash_shift); chain++) {
2348 		struct neighbour *n;
2349 
2350 		for (n = rcu_dereference_bh(nht->hash_buckets[chain]);
2351 		     n != NULL;
2352 		     n = rcu_dereference_bh(n->next))
2353 			cb(n, cookie);
2354 	}
2355 	read_unlock(&tbl->lock);
2356 	rcu_read_unlock_bh();
2357 }
2358 EXPORT_SYMBOL(neigh_for_each);
2359 
2360 /* The tbl->lock must be held as a writer and BH disabled. */
2361 void __neigh_for_each_release(struct neigh_table *tbl,
2362 			      int (*cb)(struct neighbour *))
2363 {
2364 	int chain;
2365 	struct neigh_hash_table *nht;
2366 
2367 	nht = rcu_dereference_protected(tbl->nht,
2368 					lockdep_is_held(&tbl->lock));
2369 	for (chain = 0; chain < (1 << nht->hash_shift); chain++) {
2370 		struct neighbour *n;
2371 		struct neighbour __rcu **np;
2372 
2373 		np = &nht->hash_buckets[chain];
2374 		while ((n = rcu_dereference_protected(*np,
2375 					lockdep_is_held(&tbl->lock))) != NULL) {
2376 			int release;
2377 
2378 			write_lock(&n->lock);
2379 			release = cb(n);
2380 			if (release) {
2381 				rcu_assign_pointer(*np,
2382 					rcu_dereference_protected(n->next,
2383 						lockdep_is_held(&tbl->lock)));
2384 				n->dead = 1;
2385 			} else
2386 				np = &n->next;
2387 			write_unlock(&n->lock);
2388 			if (release)
2389 				neigh_cleanup_and_release(n);
2390 		}
2391 	}
2392 }
2393 EXPORT_SYMBOL(__neigh_for_each_release);
2394 
2395 int neigh_xmit(int index, struct net_device *dev,
2396 	       const void *addr, struct sk_buff *skb)
2397 {
2398 	int err = -EAFNOSUPPORT;
2399 	if (likely(index < NEIGH_NR_TABLES)) {
2400 		struct neigh_table *tbl;
2401 		struct neighbour *neigh;
2402 
2403 		tbl = neigh_tables[index];
2404 		if (!tbl)
2405 			goto out;
2406 		neigh = __neigh_lookup_noref(tbl, addr, dev);
2407 		if (!neigh)
2408 			neigh = __neigh_create(tbl, addr, dev, false);
2409 		err = PTR_ERR(neigh);
2410 		if (IS_ERR(neigh))
2411 			goto out_kfree_skb;
2412 		err = neigh->output(neigh, skb);
2413 	}
2414 	else if (index == NEIGH_LINK_TABLE) {
2415 		err = dev_hard_header(skb, dev, ntohs(skb->protocol),
2416 				      addr, NULL, skb->len);
2417 		if (err < 0)
2418 			goto out_kfree_skb;
2419 		err = dev_queue_xmit(skb);
2420 	}
2421 out:
2422 	return err;
2423 out_kfree_skb:
2424 	kfree_skb(skb);
2425 	goto out;
2426 }
2427 EXPORT_SYMBOL(neigh_xmit);
2428 
2429 #ifdef CONFIG_PROC_FS
2430 
2431 static struct neighbour *neigh_get_first(struct seq_file *seq)
2432 {
2433 	struct neigh_seq_state *state = seq->private;
2434 	struct net *net = seq_file_net(seq);
2435 	struct neigh_hash_table *nht = state->nht;
2436 	struct neighbour *n = NULL;
2437 	int bucket = state->bucket;
2438 
2439 	state->flags &= ~NEIGH_SEQ_IS_PNEIGH;
2440 	for (bucket = 0; bucket < (1 << nht->hash_shift); bucket++) {
2441 		n = rcu_dereference_bh(nht->hash_buckets[bucket]);
2442 
2443 		while (n) {
2444 			if (!net_eq(dev_net(n->dev), net))
2445 				goto next;
2446 			if (state->neigh_sub_iter) {
2447 				loff_t fakep = 0;
2448 				void *v;
2449 
2450 				v = state->neigh_sub_iter(state, n, &fakep);
2451 				if (!v)
2452 					goto next;
2453 			}
2454 			if (!(state->flags & NEIGH_SEQ_SKIP_NOARP))
2455 				break;
2456 			if (n->nud_state & ~NUD_NOARP)
2457 				break;
2458 next:
2459 			n = rcu_dereference_bh(n->next);
2460 		}
2461 
2462 		if (n)
2463 			break;
2464 	}
2465 	state->bucket = bucket;
2466 
2467 	return n;
2468 }
2469 
2470 static struct neighbour *neigh_get_next(struct seq_file *seq,
2471 					struct neighbour *n,
2472 					loff_t *pos)
2473 {
2474 	struct neigh_seq_state *state = seq->private;
2475 	struct net *net = seq_file_net(seq);
2476 	struct neigh_hash_table *nht = state->nht;
2477 
2478 	if (state->neigh_sub_iter) {
2479 		void *v = state->neigh_sub_iter(state, n, pos);
2480 		if (v)
2481 			return n;
2482 	}
2483 	n = rcu_dereference_bh(n->next);
2484 
2485 	while (1) {
2486 		while (n) {
2487 			if (!net_eq(dev_net(n->dev), net))
2488 				goto next;
2489 			if (state->neigh_sub_iter) {
2490 				void *v = state->neigh_sub_iter(state, n, pos);
2491 				if (v)
2492 					return n;
2493 				goto next;
2494 			}
2495 			if (!(state->flags & NEIGH_SEQ_SKIP_NOARP))
2496 				break;
2497 
2498 			if (n->nud_state & ~NUD_NOARP)
2499 				break;
2500 next:
2501 			n = rcu_dereference_bh(n->next);
2502 		}
2503 
2504 		if (n)
2505 			break;
2506 
2507 		if (++state->bucket >= (1 << nht->hash_shift))
2508 			break;
2509 
2510 		n = rcu_dereference_bh(nht->hash_buckets[state->bucket]);
2511 	}
2512 
2513 	if (n && pos)
2514 		--(*pos);
2515 	return n;
2516 }
2517 
2518 static struct neighbour *neigh_get_idx(struct seq_file *seq, loff_t *pos)
2519 {
2520 	struct neighbour *n = neigh_get_first(seq);
2521 
2522 	if (n) {
2523 		--(*pos);
2524 		while (*pos) {
2525 			n = neigh_get_next(seq, n, pos);
2526 			if (!n)
2527 				break;
2528 		}
2529 	}
2530 	return *pos ? NULL : n;
2531 }
2532 
2533 static struct pneigh_entry *pneigh_get_first(struct seq_file *seq)
2534 {
2535 	struct neigh_seq_state *state = seq->private;
2536 	struct net *net = seq_file_net(seq);
2537 	struct neigh_table *tbl = state->tbl;
2538 	struct pneigh_entry *pn = NULL;
2539 	int bucket = state->bucket;
2540 
2541 	state->flags |= NEIGH_SEQ_IS_PNEIGH;
2542 	for (bucket = 0; bucket <= PNEIGH_HASHMASK; bucket++) {
2543 		pn = tbl->phash_buckets[bucket];
2544 		while (pn && !net_eq(pneigh_net(pn), net))
2545 			pn = pn->next;
2546 		if (pn)
2547 			break;
2548 	}
2549 	state->bucket = bucket;
2550 
2551 	return pn;
2552 }
2553 
2554 static struct pneigh_entry *pneigh_get_next(struct seq_file *seq,
2555 					    struct pneigh_entry *pn,
2556 					    loff_t *pos)
2557 {
2558 	struct neigh_seq_state *state = seq->private;
2559 	struct net *net = seq_file_net(seq);
2560 	struct neigh_table *tbl = state->tbl;
2561 
2562 	do {
2563 		pn = pn->next;
2564 	} while (pn && !net_eq(pneigh_net(pn), net));
2565 
2566 	while (!pn) {
2567 		if (++state->bucket > PNEIGH_HASHMASK)
2568 			break;
2569 		pn = tbl->phash_buckets[state->bucket];
2570 		while (pn && !net_eq(pneigh_net(pn), net))
2571 			pn = pn->next;
2572 		if (pn)
2573 			break;
2574 	}
2575 
2576 	if (pn && pos)
2577 		--(*pos);
2578 
2579 	return pn;
2580 }
2581 
2582 static struct pneigh_entry *pneigh_get_idx(struct seq_file *seq, loff_t *pos)
2583 {
2584 	struct pneigh_entry *pn = pneigh_get_first(seq);
2585 
2586 	if (pn) {
2587 		--(*pos);
2588 		while (*pos) {
2589 			pn = pneigh_get_next(seq, pn, pos);
2590 			if (!pn)
2591 				break;
2592 		}
2593 	}
2594 	return *pos ? NULL : pn;
2595 }
2596 
2597 static void *neigh_get_idx_any(struct seq_file *seq, loff_t *pos)
2598 {
2599 	struct neigh_seq_state *state = seq->private;
2600 	void *rc;
2601 	loff_t idxpos = *pos;
2602 
2603 	rc = neigh_get_idx(seq, &idxpos);
2604 	if (!rc && !(state->flags & NEIGH_SEQ_NEIGH_ONLY))
2605 		rc = pneigh_get_idx(seq, &idxpos);
2606 
2607 	return rc;
2608 }
2609 
2610 void *neigh_seq_start(struct seq_file *seq, loff_t *pos, struct neigh_table *tbl, unsigned int neigh_seq_flags)
2611 	__acquires(rcu_bh)
2612 {
2613 	struct neigh_seq_state *state = seq->private;
2614 
2615 	state->tbl = tbl;
2616 	state->bucket = 0;
2617 	state->flags = (neigh_seq_flags & ~NEIGH_SEQ_IS_PNEIGH);
2618 
2619 	rcu_read_lock_bh();
2620 	state->nht = rcu_dereference_bh(tbl->nht);
2621 
2622 	return *pos ? neigh_get_idx_any(seq, pos) : SEQ_START_TOKEN;
2623 }
2624 EXPORT_SYMBOL(neigh_seq_start);
2625 
2626 void *neigh_seq_next(struct seq_file *seq, void *v, loff_t *pos)
2627 {
2628 	struct neigh_seq_state *state;
2629 	void *rc;
2630 
2631 	if (v == SEQ_START_TOKEN) {
2632 		rc = neigh_get_first(seq);
2633 		goto out;
2634 	}
2635 
2636 	state = seq->private;
2637 	if (!(state->flags & NEIGH_SEQ_IS_PNEIGH)) {
2638 		rc = neigh_get_next(seq, v, NULL);
2639 		if (rc)
2640 			goto out;
2641 		if (!(state->flags & NEIGH_SEQ_NEIGH_ONLY))
2642 			rc = pneigh_get_first(seq);
2643 	} else {
2644 		BUG_ON(state->flags & NEIGH_SEQ_NEIGH_ONLY);
2645 		rc = pneigh_get_next(seq, v, NULL);
2646 	}
2647 out:
2648 	++(*pos);
2649 	return rc;
2650 }
2651 EXPORT_SYMBOL(neigh_seq_next);
2652 
2653 void neigh_seq_stop(struct seq_file *seq, void *v)
2654 	__releases(rcu_bh)
2655 {
2656 	rcu_read_unlock_bh();
2657 }
2658 EXPORT_SYMBOL(neigh_seq_stop);
2659 
2660 /* statistics via seq_file */
2661 
2662 static void *neigh_stat_seq_start(struct seq_file *seq, loff_t *pos)
2663 {
2664 	struct neigh_table *tbl = seq->private;
2665 	int cpu;
2666 
2667 	if (*pos == 0)
2668 		return SEQ_START_TOKEN;
2669 
2670 	for (cpu = *pos-1; cpu < nr_cpu_ids; ++cpu) {
2671 		if (!cpu_possible(cpu))
2672 			continue;
2673 		*pos = cpu+1;
2674 		return per_cpu_ptr(tbl->stats, cpu);
2675 	}
2676 	return NULL;
2677 }
2678 
2679 static void *neigh_stat_seq_next(struct seq_file *seq, void *v, loff_t *pos)
2680 {
2681 	struct neigh_table *tbl = seq->private;
2682 	int cpu;
2683 
2684 	for (cpu = *pos; cpu < nr_cpu_ids; ++cpu) {
2685 		if (!cpu_possible(cpu))
2686 			continue;
2687 		*pos = cpu+1;
2688 		return per_cpu_ptr(tbl->stats, cpu);
2689 	}
2690 	return NULL;
2691 }
2692 
2693 static void neigh_stat_seq_stop(struct seq_file *seq, void *v)
2694 {
2695 
2696 }
2697 
2698 static int neigh_stat_seq_show(struct seq_file *seq, void *v)
2699 {
2700 	struct neigh_table *tbl = seq->private;
2701 	struct neigh_statistics *st = v;
2702 
2703 	if (v == SEQ_START_TOKEN) {
2704 		seq_printf(seq, "entries  allocs destroys hash_grows  lookups hits  res_failed  rcv_probes_mcast rcv_probes_ucast  periodic_gc_runs forced_gc_runs unresolved_discards\n");
2705 		return 0;
2706 	}
2707 
2708 	seq_printf(seq, "%08x  %08lx %08lx %08lx  %08lx %08lx  %08lx  "
2709 			"%08lx %08lx  %08lx %08lx %08lx\n",
2710 		   atomic_read(&tbl->entries),
2711 
2712 		   st->allocs,
2713 		   st->destroys,
2714 		   st->hash_grows,
2715 
2716 		   st->lookups,
2717 		   st->hits,
2718 
2719 		   st->res_failed,
2720 
2721 		   st->rcv_probes_mcast,
2722 		   st->rcv_probes_ucast,
2723 
2724 		   st->periodic_gc_runs,
2725 		   st->forced_gc_runs,
2726 		   st->unres_discards
2727 		   );
2728 
2729 	return 0;
2730 }
2731 
2732 static const struct seq_operations neigh_stat_seq_ops = {
2733 	.start	= neigh_stat_seq_start,
2734 	.next	= neigh_stat_seq_next,
2735 	.stop	= neigh_stat_seq_stop,
2736 	.show	= neigh_stat_seq_show,
2737 };
2738 
2739 static int neigh_stat_seq_open(struct inode *inode, struct file *file)
2740 {
2741 	int ret = seq_open(file, &neigh_stat_seq_ops);
2742 
2743 	if (!ret) {
2744 		struct seq_file *sf = file->private_data;
2745 		sf->private = PDE_DATA(inode);
2746 	}
2747 	return ret;
2748 };
2749 
2750 static const struct file_operations neigh_stat_seq_fops = {
2751 	.owner	 = THIS_MODULE,
2752 	.open 	 = neigh_stat_seq_open,
2753 	.read	 = seq_read,
2754 	.llseek	 = seq_lseek,
2755 	.release = seq_release,
2756 };
2757 
2758 #endif /* CONFIG_PROC_FS */
2759 
2760 static inline size_t neigh_nlmsg_size(void)
2761 {
2762 	return NLMSG_ALIGN(sizeof(struct ndmsg))
2763 	       + nla_total_size(MAX_ADDR_LEN) /* NDA_DST */
2764 	       + nla_total_size(MAX_ADDR_LEN) /* NDA_LLADDR */
2765 	       + nla_total_size(sizeof(struct nda_cacheinfo))
2766 	       + nla_total_size(4); /* NDA_PROBES */
2767 }
2768 
2769 static void __neigh_notify(struct neighbour *n, int type, int flags)
2770 {
2771 	struct net *net = dev_net(n->dev);
2772 	struct sk_buff *skb;
2773 	int err = -ENOBUFS;
2774 
2775 	skb = nlmsg_new(neigh_nlmsg_size(), GFP_ATOMIC);
2776 	if (skb == NULL)
2777 		goto errout;
2778 
2779 	err = neigh_fill_info(skb, n, 0, 0, type, flags);
2780 	if (err < 0) {
2781 		/* -EMSGSIZE implies BUG in neigh_nlmsg_size() */
2782 		WARN_ON(err == -EMSGSIZE);
2783 		kfree_skb(skb);
2784 		goto errout;
2785 	}
2786 	rtnl_notify(skb, net, 0, RTNLGRP_NEIGH, NULL, GFP_ATOMIC);
2787 	return;
2788 errout:
2789 	if (err < 0)
2790 		rtnl_set_sk_err(net, RTNLGRP_NEIGH, err);
2791 }
2792 
2793 void neigh_app_ns(struct neighbour *n)
2794 {
2795 	__neigh_notify(n, RTM_GETNEIGH, NLM_F_REQUEST);
2796 }
2797 EXPORT_SYMBOL(neigh_app_ns);
2798 
2799 #ifdef CONFIG_SYSCTL
2800 static int zero;
2801 static int int_max = INT_MAX;
2802 static int unres_qlen_max = INT_MAX / SKB_TRUESIZE(ETH_FRAME_LEN);
2803 
2804 static int proc_unres_qlen(struct ctl_table *ctl, int write,
2805 			   void __user *buffer, size_t *lenp, loff_t *ppos)
2806 {
2807 	int size, ret;
2808 	struct ctl_table tmp = *ctl;
2809 
2810 	tmp.extra1 = &zero;
2811 	tmp.extra2 = &unres_qlen_max;
2812 	tmp.data = &size;
2813 
2814 	size = *(int *)ctl->data / SKB_TRUESIZE(ETH_FRAME_LEN);
2815 	ret = proc_dointvec_minmax(&tmp, write, buffer, lenp, ppos);
2816 
2817 	if (write && !ret)
2818 		*(int *)ctl->data = size * SKB_TRUESIZE(ETH_FRAME_LEN);
2819 	return ret;
2820 }
2821 
2822 static struct neigh_parms *neigh_get_dev_parms_rcu(struct net_device *dev,
2823 						   int family)
2824 {
2825 	switch (family) {
2826 	case AF_INET:
2827 		return __in_dev_arp_parms_get_rcu(dev);
2828 	case AF_INET6:
2829 		return __in6_dev_nd_parms_get_rcu(dev);
2830 	}
2831 	return NULL;
2832 }
2833 
2834 static void neigh_copy_dflt_parms(struct net *net, struct neigh_parms *p,
2835 				  int index)
2836 {
2837 	struct net_device *dev;
2838 	int family = neigh_parms_family(p);
2839 
2840 	rcu_read_lock();
2841 	for_each_netdev_rcu(net, dev) {
2842 		struct neigh_parms *dst_p =
2843 				neigh_get_dev_parms_rcu(dev, family);
2844 
2845 		if (dst_p && !test_bit(index, dst_p->data_state))
2846 			dst_p->data[index] = p->data[index];
2847 	}
2848 	rcu_read_unlock();
2849 }
2850 
2851 static void neigh_proc_update(struct ctl_table *ctl, int write)
2852 {
2853 	struct net_device *dev = ctl->extra1;
2854 	struct neigh_parms *p = ctl->extra2;
2855 	struct net *net = neigh_parms_net(p);
2856 	int index = (int *) ctl->data - p->data;
2857 
2858 	if (!write)
2859 		return;
2860 
2861 	set_bit(index, p->data_state);
2862 	if (!dev) /* NULL dev means this is default value */
2863 		neigh_copy_dflt_parms(net, p, index);
2864 }
2865 
2866 static int neigh_proc_dointvec_zero_intmax(struct ctl_table *ctl, int write,
2867 					   void __user *buffer,
2868 					   size_t *lenp, loff_t *ppos)
2869 {
2870 	struct ctl_table tmp = *ctl;
2871 	int ret;
2872 
2873 	tmp.extra1 = &zero;
2874 	tmp.extra2 = &int_max;
2875 
2876 	ret = proc_dointvec_minmax(&tmp, write, buffer, lenp, ppos);
2877 	neigh_proc_update(ctl, write);
2878 	return ret;
2879 }
2880 
2881 int neigh_proc_dointvec(struct ctl_table *ctl, int write,
2882 			void __user *buffer, size_t *lenp, loff_t *ppos)
2883 {
2884 	int ret = proc_dointvec(ctl, write, buffer, lenp, ppos);
2885 
2886 	neigh_proc_update(ctl, write);
2887 	return ret;
2888 }
2889 EXPORT_SYMBOL(neigh_proc_dointvec);
2890 
2891 int neigh_proc_dointvec_jiffies(struct ctl_table *ctl, int write,
2892 				void __user *buffer,
2893 				size_t *lenp, loff_t *ppos)
2894 {
2895 	int ret = proc_dointvec_jiffies(ctl, write, buffer, lenp, ppos);
2896 
2897 	neigh_proc_update(ctl, write);
2898 	return ret;
2899 }
2900 EXPORT_SYMBOL(neigh_proc_dointvec_jiffies);
2901 
2902 static int neigh_proc_dointvec_userhz_jiffies(struct ctl_table *ctl, int write,
2903 					      void __user *buffer,
2904 					      size_t *lenp, loff_t *ppos)
2905 {
2906 	int ret = proc_dointvec_userhz_jiffies(ctl, write, buffer, lenp, ppos);
2907 
2908 	neigh_proc_update(ctl, write);
2909 	return ret;
2910 }
2911 
2912 int neigh_proc_dointvec_ms_jiffies(struct ctl_table *ctl, int write,
2913 				   void __user *buffer,
2914 				   size_t *lenp, loff_t *ppos)
2915 {
2916 	int ret = proc_dointvec_ms_jiffies(ctl, write, buffer, lenp, ppos);
2917 
2918 	neigh_proc_update(ctl, write);
2919 	return ret;
2920 }
2921 EXPORT_SYMBOL(neigh_proc_dointvec_ms_jiffies);
2922 
2923 static int neigh_proc_dointvec_unres_qlen(struct ctl_table *ctl, int write,
2924 					  void __user *buffer,
2925 					  size_t *lenp, loff_t *ppos)
2926 {
2927 	int ret = proc_unres_qlen(ctl, write, buffer, lenp, ppos);
2928 
2929 	neigh_proc_update(ctl, write);
2930 	return ret;
2931 }
2932 
2933 static int neigh_proc_base_reachable_time(struct ctl_table *ctl, int write,
2934 					  void __user *buffer,
2935 					  size_t *lenp, loff_t *ppos)
2936 {
2937 	struct neigh_parms *p = ctl->extra2;
2938 	int ret;
2939 
2940 	if (strcmp(ctl->procname, "base_reachable_time") == 0)
2941 		ret = neigh_proc_dointvec_jiffies(ctl, write, buffer, lenp, ppos);
2942 	else if (strcmp(ctl->procname, "base_reachable_time_ms") == 0)
2943 		ret = neigh_proc_dointvec_ms_jiffies(ctl, write, buffer, lenp, ppos);
2944 	else
2945 		ret = -1;
2946 
2947 	if (write && ret == 0) {
2948 		/* update reachable_time as well, otherwise, the change will
2949 		 * only be effective after the next time neigh_periodic_work
2950 		 * decides to recompute it
2951 		 */
2952 		p->reachable_time =
2953 			neigh_rand_reach_time(NEIGH_VAR(p, BASE_REACHABLE_TIME));
2954 	}
2955 	return ret;
2956 }
2957 
2958 #define NEIGH_PARMS_DATA_OFFSET(index)	\
2959 	(&((struct neigh_parms *) 0)->data[index])
2960 
2961 #define NEIGH_SYSCTL_ENTRY(attr, data_attr, name, mval, proc) \
2962 	[NEIGH_VAR_ ## attr] = { \
2963 		.procname	= name, \
2964 		.data		= NEIGH_PARMS_DATA_OFFSET(NEIGH_VAR_ ## data_attr), \
2965 		.maxlen		= sizeof(int), \
2966 		.mode		= mval, \
2967 		.proc_handler	= proc, \
2968 	}
2969 
2970 #define NEIGH_SYSCTL_ZERO_INTMAX_ENTRY(attr, name) \
2971 	NEIGH_SYSCTL_ENTRY(attr, attr, name, 0644, neigh_proc_dointvec_zero_intmax)
2972 
2973 #define NEIGH_SYSCTL_JIFFIES_ENTRY(attr, name) \
2974 	NEIGH_SYSCTL_ENTRY(attr, attr, name, 0644, neigh_proc_dointvec_jiffies)
2975 
2976 #define NEIGH_SYSCTL_USERHZ_JIFFIES_ENTRY(attr, name) \
2977 	NEIGH_SYSCTL_ENTRY(attr, attr, name, 0644, neigh_proc_dointvec_userhz_jiffies)
2978 
2979 #define NEIGH_SYSCTL_MS_JIFFIES_ENTRY(attr, name) \
2980 	NEIGH_SYSCTL_ENTRY(attr, attr, name, 0644, neigh_proc_dointvec_ms_jiffies)
2981 
2982 #define NEIGH_SYSCTL_MS_JIFFIES_REUSED_ENTRY(attr, data_attr, name) \
2983 	NEIGH_SYSCTL_ENTRY(attr, data_attr, name, 0644, neigh_proc_dointvec_ms_jiffies)
2984 
2985 #define NEIGH_SYSCTL_UNRES_QLEN_REUSED_ENTRY(attr, data_attr, name) \
2986 	NEIGH_SYSCTL_ENTRY(attr, data_attr, name, 0644, neigh_proc_dointvec_unres_qlen)
2987 
2988 static struct neigh_sysctl_table {
2989 	struct ctl_table_header *sysctl_header;
2990 	struct ctl_table neigh_vars[NEIGH_VAR_MAX + 1];
2991 } neigh_sysctl_template __read_mostly = {
2992 	.neigh_vars = {
2993 		NEIGH_SYSCTL_ZERO_INTMAX_ENTRY(MCAST_PROBES, "mcast_solicit"),
2994 		NEIGH_SYSCTL_ZERO_INTMAX_ENTRY(UCAST_PROBES, "ucast_solicit"),
2995 		NEIGH_SYSCTL_ZERO_INTMAX_ENTRY(APP_PROBES, "app_solicit"),
2996 		NEIGH_SYSCTL_ZERO_INTMAX_ENTRY(MCAST_REPROBES, "mcast_resolicit"),
2997 		NEIGH_SYSCTL_USERHZ_JIFFIES_ENTRY(RETRANS_TIME, "retrans_time"),
2998 		NEIGH_SYSCTL_JIFFIES_ENTRY(BASE_REACHABLE_TIME, "base_reachable_time"),
2999 		NEIGH_SYSCTL_JIFFIES_ENTRY(DELAY_PROBE_TIME, "delay_first_probe_time"),
3000 		NEIGH_SYSCTL_JIFFIES_ENTRY(GC_STALETIME, "gc_stale_time"),
3001 		NEIGH_SYSCTL_ZERO_INTMAX_ENTRY(QUEUE_LEN_BYTES, "unres_qlen_bytes"),
3002 		NEIGH_SYSCTL_ZERO_INTMAX_ENTRY(PROXY_QLEN, "proxy_qlen"),
3003 		NEIGH_SYSCTL_USERHZ_JIFFIES_ENTRY(ANYCAST_DELAY, "anycast_delay"),
3004 		NEIGH_SYSCTL_USERHZ_JIFFIES_ENTRY(PROXY_DELAY, "proxy_delay"),
3005 		NEIGH_SYSCTL_USERHZ_JIFFIES_ENTRY(LOCKTIME, "locktime"),
3006 		NEIGH_SYSCTL_UNRES_QLEN_REUSED_ENTRY(QUEUE_LEN, QUEUE_LEN_BYTES, "unres_qlen"),
3007 		NEIGH_SYSCTL_MS_JIFFIES_REUSED_ENTRY(RETRANS_TIME_MS, RETRANS_TIME, "retrans_time_ms"),
3008 		NEIGH_SYSCTL_MS_JIFFIES_REUSED_ENTRY(BASE_REACHABLE_TIME_MS, BASE_REACHABLE_TIME, "base_reachable_time_ms"),
3009 		[NEIGH_VAR_GC_INTERVAL] = {
3010 			.procname	= "gc_interval",
3011 			.maxlen		= sizeof(int),
3012 			.mode		= 0644,
3013 			.proc_handler	= proc_dointvec_jiffies,
3014 		},
3015 		[NEIGH_VAR_GC_THRESH1] = {
3016 			.procname	= "gc_thresh1",
3017 			.maxlen		= sizeof(int),
3018 			.mode		= 0644,
3019 			.extra1 	= &zero,
3020 			.extra2		= &int_max,
3021 			.proc_handler	= proc_dointvec_minmax,
3022 		},
3023 		[NEIGH_VAR_GC_THRESH2] = {
3024 			.procname	= "gc_thresh2",
3025 			.maxlen		= sizeof(int),
3026 			.mode		= 0644,
3027 			.extra1 	= &zero,
3028 			.extra2		= &int_max,
3029 			.proc_handler	= proc_dointvec_minmax,
3030 		},
3031 		[NEIGH_VAR_GC_THRESH3] = {
3032 			.procname	= "gc_thresh3",
3033 			.maxlen		= sizeof(int),
3034 			.mode		= 0644,
3035 			.extra1 	= &zero,
3036 			.extra2		= &int_max,
3037 			.proc_handler	= proc_dointvec_minmax,
3038 		},
3039 		{},
3040 	},
3041 };
3042 
3043 int neigh_sysctl_register(struct net_device *dev, struct neigh_parms *p,
3044 			  proc_handler *handler)
3045 {
3046 	int i;
3047 	struct neigh_sysctl_table *t;
3048 	const char *dev_name_source;
3049 	char neigh_path[ sizeof("net//neigh/") + IFNAMSIZ + IFNAMSIZ ];
3050 	char *p_name;
3051 
3052 	t = kmemdup(&neigh_sysctl_template, sizeof(*t), GFP_KERNEL);
3053 	if (!t)
3054 		goto err;
3055 
3056 	for (i = 0; i < NEIGH_VAR_GC_INTERVAL; i++) {
3057 		t->neigh_vars[i].data += (long) p;
3058 		t->neigh_vars[i].extra1 = dev;
3059 		t->neigh_vars[i].extra2 = p;
3060 	}
3061 
3062 	if (dev) {
3063 		dev_name_source = dev->name;
3064 		/* Terminate the table early */
3065 		memset(&t->neigh_vars[NEIGH_VAR_GC_INTERVAL], 0,
3066 		       sizeof(t->neigh_vars[NEIGH_VAR_GC_INTERVAL]));
3067 	} else {
3068 		struct neigh_table *tbl = p->tbl;
3069 		dev_name_source = "default";
3070 		t->neigh_vars[NEIGH_VAR_GC_INTERVAL].data = &tbl->gc_interval;
3071 		t->neigh_vars[NEIGH_VAR_GC_THRESH1].data = &tbl->gc_thresh1;
3072 		t->neigh_vars[NEIGH_VAR_GC_THRESH2].data = &tbl->gc_thresh2;
3073 		t->neigh_vars[NEIGH_VAR_GC_THRESH3].data = &tbl->gc_thresh3;
3074 	}
3075 
3076 	if (handler) {
3077 		/* RetransTime */
3078 		t->neigh_vars[NEIGH_VAR_RETRANS_TIME].proc_handler = handler;
3079 		/* ReachableTime */
3080 		t->neigh_vars[NEIGH_VAR_BASE_REACHABLE_TIME].proc_handler = handler;
3081 		/* RetransTime (in milliseconds)*/
3082 		t->neigh_vars[NEIGH_VAR_RETRANS_TIME_MS].proc_handler = handler;
3083 		/* ReachableTime (in milliseconds) */
3084 		t->neigh_vars[NEIGH_VAR_BASE_REACHABLE_TIME_MS].proc_handler = handler;
3085 	} else {
3086 		/* Those handlers will update p->reachable_time after
3087 		 * base_reachable_time(_ms) is set to ensure the new timer starts being
3088 		 * applied after the next neighbour update instead of waiting for
3089 		 * neigh_periodic_work to update its value (can be multiple minutes)
3090 		 * So any handler that replaces them should do this as well
3091 		 */
3092 		/* ReachableTime */
3093 		t->neigh_vars[NEIGH_VAR_BASE_REACHABLE_TIME].proc_handler =
3094 			neigh_proc_base_reachable_time;
3095 		/* ReachableTime (in milliseconds) */
3096 		t->neigh_vars[NEIGH_VAR_BASE_REACHABLE_TIME_MS].proc_handler =
3097 			neigh_proc_base_reachable_time;
3098 	}
3099 
3100 	/* Don't export sysctls to unprivileged users */
3101 	if (neigh_parms_net(p)->user_ns != &init_user_ns)
3102 		t->neigh_vars[0].procname = NULL;
3103 
3104 	switch (neigh_parms_family(p)) {
3105 	case AF_INET:
3106 	      p_name = "ipv4";
3107 	      break;
3108 	case AF_INET6:
3109 	      p_name = "ipv6";
3110 	      break;
3111 	default:
3112 	      BUG();
3113 	}
3114 
3115 	snprintf(neigh_path, sizeof(neigh_path), "net/%s/neigh/%s",
3116 		p_name, dev_name_source);
3117 	t->sysctl_header =
3118 		register_net_sysctl(neigh_parms_net(p), neigh_path, t->neigh_vars);
3119 	if (!t->sysctl_header)
3120 		goto free;
3121 
3122 	p->sysctl_table = t;
3123 	return 0;
3124 
3125 free:
3126 	kfree(t);
3127 err:
3128 	return -ENOBUFS;
3129 }
3130 EXPORT_SYMBOL(neigh_sysctl_register);
3131 
3132 void neigh_sysctl_unregister(struct neigh_parms *p)
3133 {
3134 	if (p->sysctl_table) {
3135 		struct neigh_sysctl_table *t = p->sysctl_table;
3136 		p->sysctl_table = NULL;
3137 		unregister_net_sysctl_table(t->sysctl_header);
3138 		kfree(t);
3139 	}
3140 }
3141 EXPORT_SYMBOL(neigh_sysctl_unregister);
3142 
3143 #endif	/* CONFIG_SYSCTL */
3144 
3145 static int __init neigh_init(void)
3146 {
3147 	rtnl_register(PF_UNSPEC, RTM_NEWNEIGH, neigh_add, NULL, NULL);
3148 	rtnl_register(PF_UNSPEC, RTM_DELNEIGH, neigh_delete, NULL, NULL);
3149 	rtnl_register(PF_UNSPEC, RTM_GETNEIGH, NULL, neigh_dump_info, NULL);
3150 
3151 	rtnl_register(PF_UNSPEC, RTM_GETNEIGHTBL, NULL, neightbl_dump_info,
3152 		      NULL);
3153 	rtnl_register(PF_UNSPEC, RTM_SETNEIGHTBL, neightbl_set, NULL, NULL);
3154 
3155 	return 0;
3156 }
3157 
3158 subsys_initcall(neigh_init);
3159 
3160