xref: /openbmc/linux/net/8021q/vlan.c (revision fca3aa16)
1 /*
2  * INET		802.1Q VLAN
3  *		Ethernet-type device handling.
4  *
5  * Authors:	Ben Greear <greearb@candelatech.com>
6  *              Please send support related email to: netdev@vger.kernel.org
7  *              VLAN Home Page: http://www.candelatech.com/~greear/vlan.html
8  *
9  * Fixes:
10  *              Fix for packet capture - Nick Eggleston <nick@dccinc.com>;
11  *		Add HW acceleration hooks - David S. Miller <davem@redhat.com>;
12  *		Correct all the locking - David S. Miller <davem@redhat.com>;
13  *		Use hash table for VLAN groups - David S. Miller <davem@redhat.com>
14  *
15  *		This program is free software; you can redistribute it and/or
16  *		modify it under the terms of the GNU General Public License
17  *		as published by the Free Software Foundation; either version
18  *		2 of the License, or (at your option) any later version.
19  */
20 
21 #define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
22 
23 #include <linux/capability.h>
24 #include <linux/module.h>
25 #include <linux/netdevice.h>
26 #include <linux/skbuff.h>
27 #include <linux/slab.h>
28 #include <linux/init.h>
29 #include <linux/rculist.h>
30 #include <net/p8022.h>
31 #include <net/arp.h>
32 #include <linux/rtnetlink.h>
33 #include <linux/notifier.h>
34 #include <net/rtnetlink.h>
35 #include <net/net_namespace.h>
36 #include <net/netns/generic.h>
37 #include <linux/uaccess.h>
38 
39 #include <linux/if_vlan.h>
40 #include "vlan.h"
41 #include "vlanproc.h"
42 
43 #define DRV_VERSION "1.8"
44 
45 /* Global VLAN variables */
46 
47 unsigned int vlan_net_id __read_mostly;
48 
49 const char vlan_fullname[] = "802.1Q VLAN Support";
50 const char vlan_version[] = DRV_VERSION;
51 
52 /* End of global variables definitions. */
53 
54 static int vlan_group_prealloc_vid(struct vlan_group *vg,
55 				   __be16 vlan_proto, u16 vlan_id)
56 {
57 	struct net_device **array;
58 	unsigned int pidx, vidx;
59 	unsigned int size;
60 
61 	ASSERT_RTNL();
62 
63 	pidx  = vlan_proto_idx(vlan_proto);
64 	vidx  = vlan_id / VLAN_GROUP_ARRAY_PART_LEN;
65 	array = vg->vlan_devices_arrays[pidx][vidx];
66 	if (array != NULL)
67 		return 0;
68 
69 	size = sizeof(struct net_device *) * VLAN_GROUP_ARRAY_PART_LEN;
70 	array = kzalloc(size, GFP_KERNEL);
71 	if (array == NULL)
72 		return -ENOBUFS;
73 
74 	vg->vlan_devices_arrays[pidx][vidx] = array;
75 	return 0;
76 }
77 
78 void unregister_vlan_dev(struct net_device *dev, struct list_head *head)
79 {
80 	struct vlan_dev_priv *vlan = vlan_dev_priv(dev);
81 	struct net_device *real_dev = vlan->real_dev;
82 	struct vlan_info *vlan_info;
83 	struct vlan_group *grp;
84 	u16 vlan_id = vlan->vlan_id;
85 
86 	ASSERT_RTNL();
87 
88 	vlan_info = rtnl_dereference(real_dev->vlan_info);
89 	BUG_ON(!vlan_info);
90 
91 	grp = &vlan_info->grp;
92 
93 	grp->nr_vlan_devs--;
94 
95 	if (vlan->flags & VLAN_FLAG_MVRP)
96 		vlan_mvrp_request_leave(dev);
97 	if (vlan->flags & VLAN_FLAG_GVRP)
98 		vlan_gvrp_request_leave(dev);
99 
100 	vlan_group_set_device(grp, vlan->vlan_proto, vlan_id, NULL);
101 
102 	netdev_upper_dev_unlink(real_dev, dev);
103 	/* Because unregister_netdevice_queue() makes sure at least one rcu
104 	 * grace period is respected before device freeing,
105 	 * we dont need to call synchronize_net() here.
106 	 */
107 	unregister_netdevice_queue(dev, head);
108 
109 	if (grp->nr_vlan_devs == 0) {
110 		vlan_mvrp_uninit_applicant(real_dev);
111 		vlan_gvrp_uninit_applicant(real_dev);
112 	}
113 
114 	vlan_vid_del(real_dev, vlan->vlan_proto, vlan_id);
115 
116 	/* Get rid of the vlan's reference to real_dev */
117 	dev_put(real_dev);
118 }
119 
120 int vlan_check_real_dev(struct net_device *real_dev,
121 			__be16 protocol, u16 vlan_id)
122 {
123 	const char *name = real_dev->name;
124 
125 	if (real_dev->features & NETIF_F_VLAN_CHALLENGED) {
126 		pr_info("VLANs not supported on %s\n", name);
127 		return -EOPNOTSUPP;
128 	}
129 
130 	if (vlan_find_dev(real_dev, protocol, vlan_id) != NULL)
131 		return -EEXIST;
132 
133 	return 0;
134 }
135 
136 int register_vlan_dev(struct net_device *dev, struct netlink_ext_ack *extack)
137 {
138 	struct vlan_dev_priv *vlan = vlan_dev_priv(dev);
139 	struct net_device *real_dev = vlan->real_dev;
140 	u16 vlan_id = vlan->vlan_id;
141 	struct vlan_info *vlan_info;
142 	struct vlan_group *grp;
143 	int err;
144 
145 	err = vlan_vid_add(real_dev, vlan->vlan_proto, vlan_id);
146 	if (err)
147 		return err;
148 
149 	vlan_info = rtnl_dereference(real_dev->vlan_info);
150 	/* vlan_info should be there now. vlan_vid_add took care of it */
151 	BUG_ON(!vlan_info);
152 
153 	grp = &vlan_info->grp;
154 	if (grp->nr_vlan_devs == 0) {
155 		err = vlan_gvrp_init_applicant(real_dev);
156 		if (err < 0)
157 			goto out_vid_del;
158 		err = vlan_mvrp_init_applicant(real_dev);
159 		if (err < 0)
160 			goto out_uninit_gvrp;
161 	}
162 
163 	err = vlan_group_prealloc_vid(grp, vlan->vlan_proto, vlan_id);
164 	if (err < 0)
165 		goto out_uninit_mvrp;
166 
167 	vlan->nest_level = dev_get_nest_level(real_dev) + 1;
168 	err = register_netdevice(dev);
169 	if (err < 0)
170 		goto out_uninit_mvrp;
171 
172 	err = netdev_upper_dev_link(real_dev, dev, extack);
173 	if (err)
174 		goto out_unregister_netdev;
175 
176 	/* Account for reference in struct vlan_dev_priv */
177 	dev_hold(real_dev);
178 
179 	netif_stacked_transfer_operstate(real_dev, dev);
180 	linkwatch_fire_event(dev); /* _MUST_ call rfc2863_policy() */
181 
182 	/* So, got the sucker initialized, now lets place
183 	 * it into our local structure.
184 	 */
185 	vlan_group_set_device(grp, vlan->vlan_proto, vlan_id, dev);
186 	grp->nr_vlan_devs++;
187 
188 	return 0;
189 
190 out_unregister_netdev:
191 	unregister_netdevice(dev);
192 out_uninit_mvrp:
193 	if (grp->nr_vlan_devs == 0)
194 		vlan_mvrp_uninit_applicant(real_dev);
195 out_uninit_gvrp:
196 	if (grp->nr_vlan_devs == 0)
197 		vlan_gvrp_uninit_applicant(real_dev);
198 out_vid_del:
199 	vlan_vid_del(real_dev, vlan->vlan_proto, vlan_id);
200 	return err;
201 }
202 
203 /*  Attach a VLAN device to a mac address (ie Ethernet Card).
204  *  Returns 0 if the device was created or a negative error code otherwise.
205  */
206 static int register_vlan_device(struct net_device *real_dev, u16 vlan_id)
207 {
208 	struct net_device *new_dev;
209 	struct vlan_dev_priv *vlan;
210 	struct net *net = dev_net(real_dev);
211 	struct vlan_net *vn = net_generic(net, vlan_net_id);
212 	char name[IFNAMSIZ];
213 	int err;
214 
215 	if (vlan_id >= VLAN_VID_MASK)
216 		return -ERANGE;
217 
218 	err = vlan_check_real_dev(real_dev, htons(ETH_P_8021Q), vlan_id);
219 	if (err < 0)
220 		return err;
221 
222 	/* Gotta set up the fields for the device. */
223 	switch (vn->name_type) {
224 	case VLAN_NAME_TYPE_RAW_PLUS_VID:
225 		/* name will look like:	 eth1.0005 */
226 		snprintf(name, IFNAMSIZ, "%s.%.4i", real_dev->name, vlan_id);
227 		break;
228 	case VLAN_NAME_TYPE_PLUS_VID_NO_PAD:
229 		/* Put our vlan.VID in the name.
230 		 * Name will look like:	 vlan5
231 		 */
232 		snprintf(name, IFNAMSIZ, "vlan%i", vlan_id);
233 		break;
234 	case VLAN_NAME_TYPE_RAW_PLUS_VID_NO_PAD:
235 		/* Put our vlan.VID in the name.
236 		 * Name will look like:	 eth0.5
237 		 */
238 		snprintf(name, IFNAMSIZ, "%s.%i", real_dev->name, vlan_id);
239 		break;
240 	case VLAN_NAME_TYPE_PLUS_VID:
241 		/* Put our vlan.VID in the name.
242 		 * Name will look like:	 vlan0005
243 		 */
244 	default:
245 		snprintf(name, IFNAMSIZ, "vlan%.4i", vlan_id);
246 	}
247 
248 	new_dev = alloc_netdev(sizeof(struct vlan_dev_priv), name,
249 			       NET_NAME_UNKNOWN, vlan_setup);
250 
251 	if (new_dev == NULL)
252 		return -ENOBUFS;
253 
254 	dev_net_set(new_dev, net);
255 	/* need 4 bytes for extra VLAN header info,
256 	 * hope the underlying device can handle it.
257 	 */
258 	new_dev->mtu = real_dev->mtu;
259 
260 	vlan = vlan_dev_priv(new_dev);
261 	vlan->vlan_proto = htons(ETH_P_8021Q);
262 	vlan->vlan_id = vlan_id;
263 	vlan->real_dev = real_dev;
264 	vlan->dent = NULL;
265 	vlan->flags = VLAN_FLAG_REORDER_HDR;
266 
267 	new_dev->rtnl_link_ops = &vlan_link_ops;
268 	err = register_vlan_dev(new_dev, NULL);
269 	if (err < 0)
270 		goto out_free_newdev;
271 
272 	return 0;
273 
274 out_free_newdev:
275 	if (new_dev->reg_state == NETREG_UNINITIALIZED)
276 		free_netdev(new_dev);
277 	return err;
278 }
279 
280 static void vlan_sync_address(struct net_device *dev,
281 			      struct net_device *vlandev)
282 {
283 	struct vlan_dev_priv *vlan = vlan_dev_priv(vlandev);
284 
285 	/* May be called without an actual change */
286 	if (ether_addr_equal(vlan->real_dev_addr, dev->dev_addr))
287 		return;
288 
289 	/* vlan continues to inherit address of lower device */
290 	if (vlan_dev_inherit_address(vlandev, dev))
291 		goto out;
292 
293 	/* vlan address was different from the old address and is equal to
294 	 * the new address */
295 	if (!ether_addr_equal(vlandev->dev_addr, vlan->real_dev_addr) &&
296 	    ether_addr_equal(vlandev->dev_addr, dev->dev_addr))
297 		dev_uc_del(dev, vlandev->dev_addr);
298 
299 	/* vlan address was equal to the old address and is different from
300 	 * the new address */
301 	if (ether_addr_equal(vlandev->dev_addr, vlan->real_dev_addr) &&
302 	    !ether_addr_equal(vlandev->dev_addr, dev->dev_addr))
303 		dev_uc_add(dev, vlandev->dev_addr);
304 
305 out:
306 	ether_addr_copy(vlan->real_dev_addr, dev->dev_addr);
307 }
308 
309 static void vlan_transfer_features(struct net_device *dev,
310 				   struct net_device *vlandev)
311 {
312 	struct vlan_dev_priv *vlan = vlan_dev_priv(vlandev);
313 
314 	vlandev->gso_max_size = dev->gso_max_size;
315 	vlandev->gso_max_segs = dev->gso_max_segs;
316 
317 	if (vlan_hw_offload_capable(dev->features, vlan->vlan_proto))
318 		vlandev->hard_header_len = dev->hard_header_len;
319 	else
320 		vlandev->hard_header_len = dev->hard_header_len + VLAN_HLEN;
321 
322 #if IS_ENABLED(CONFIG_FCOE)
323 	vlandev->fcoe_ddp_xid = dev->fcoe_ddp_xid;
324 #endif
325 
326 	vlandev->priv_flags &= ~IFF_XMIT_DST_RELEASE;
327 	vlandev->priv_flags |= (vlan->real_dev->priv_flags & IFF_XMIT_DST_RELEASE);
328 
329 	netdev_update_features(vlandev);
330 }
331 
332 static int __vlan_device_event(struct net_device *dev, unsigned long event)
333 {
334 	int err = 0;
335 
336 	switch (event) {
337 	case NETDEV_CHANGENAME:
338 		vlan_proc_rem_dev(dev);
339 		err = vlan_proc_add_dev(dev);
340 		break;
341 	case NETDEV_REGISTER:
342 		err = vlan_proc_add_dev(dev);
343 		break;
344 	case NETDEV_UNREGISTER:
345 		vlan_proc_rem_dev(dev);
346 		break;
347 	}
348 
349 	return err;
350 }
351 
352 static int vlan_device_event(struct notifier_block *unused, unsigned long event,
353 			     void *ptr)
354 {
355 	struct net_device *dev = netdev_notifier_info_to_dev(ptr);
356 	struct vlan_group *grp;
357 	struct vlan_info *vlan_info;
358 	int i, flgs;
359 	struct net_device *vlandev;
360 	struct vlan_dev_priv *vlan;
361 	bool last = false;
362 	LIST_HEAD(list);
363 	int err;
364 
365 	if (is_vlan_dev(dev)) {
366 		int err = __vlan_device_event(dev, event);
367 
368 		if (err)
369 			return notifier_from_errno(err);
370 	}
371 
372 	if ((event == NETDEV_UP) &&
373 	    (dev->features & NETIF_F_HW_VLAN_CTAG_FILTER)) {
374 		pr_info("adding VLAN 0 to HW filter on device %s\n",
375 			dev->name);
376 		vlan_vid_add(dev, htons(ETH_P_8021Q), 0);
377 	}
378 	if (event == NETDEV_DOWN &&
379 	    (dev->features & NETIF_F_HW_VLAN_CTAG_FILTER))
380 		vlan_vid_del(dev, htons(ETH_P_8021Q), 0);
381 
382 	vlan_info = rtnl_dereference(dev->vlan_info);
383 	if (!vlan_info)
384 		goto out;
385 	grp = &vlan_info->grp;
386 
387 	/* It is OK that we do not hold the group lock right now,
388 	 * as we run under the RTNL lock.
389 	 */
390 
391 	switch (event) {
392 	case NETDEV_CHANGE:
393 		/* Propagate real device state to vlan devices */
394 		vlan_group_for_each_dev(grp, i, vlandev)
395 			netif_stacked_transfer_operstate(dev, vlandev);
396 		break;
397 
398 	case NETDEV_CHANGEADDR:
399 		/* Adjust unicast filters on underlying device */
400 		vlan_group_for_each_dev(grp, i, vlandev) {
401 			flgs = vlandev->flags;
402 			if (!(flgs & IFF_UP))
403 				continue;
404 
405 			vlan_sync_address(dev, vlandev);
406 		}
407 		break;
408 
409 	case NETDEV_CHANGEMTU:
410 		vlan_group_for_each_dev(grp, i, vlandev) {
411 			if (vlandev->mtu <= dev->mtu)
412 				continue;
413 
414 			dev_set_mtu(vlandev, dev->mtu);
415 		}
416 		break;
417 
418 	case NETDEV_FEAT_CHANGE:
419 		/* Propagate device features to underlying device */
420 		vlan_group_for_each_dev(grp, i, vlandev)
421 			vlan_transfer_features(dev, vlandev);
422 		break;
423 
424 	case NETDEV_DOWN: {
425 		struct net_device *tmp;
426 		LIST_HEAD(close_list);
427 
428 		/* Put all VLANs for this dev in the down state too.  */
429 		vlan_group_for_each_dev(grp, i, vlandev) {
430 			flgs = vlandev->flags;
431 			if (!(flgs & IFF_UP))
432 				continue;
433 
434 			vlan = vlan_dev_priv(vlandev);
435 			if (!(vlan->flags & VLAN_FLAG_LOOSE_BINDING))
436 				list_add(&vlandev->close_list, &close_list);
437 		}
438 
439 		dev_close_many(&close_list, false);
440 
441 		list_for_each_entry_safe(vlandev, tmp, &close_list, close_list) {
442 			netif_stacked_transfer_operstate(dev, vlandev);
443 			list_del_init(&vlandev->close_list);
444 		}
445 		list_del(&close_list);
446 		break;
447 	}
448 	case NETDEV_UP:
449 		/* Put all VLANs for this dev in the up state too.  */
450 		vlan_group_for_each_dev(grp, i, vlandev) {
451 			flgs = dev_get_flags(vlandev);
452 			if (flgs & IFF_UP)
453 				continue;
454 
455 			vlan = vlan_dev_priv(vlandev);
456 			if (!(vlan->flags & VLAN_FLAG_LOOSE_BINDING))
457 				dev_change_flags(vlandev, flgs | IFF_UP);
458 			netif_stacked_transfer_operstate(dev, vlandev);
459 		}
460 		break;
461 
462 	case NETDEV_UNREGISTER:
463 		/* twiddle thumbs on netns device moves */
464 		if (dev->reg_state != NETREG_UNREGISTERING)
465 			break;
466 
467 		vlan_group_for_each_dev(grp, i, vlandev) {
468 			/* removal of last vid destroys vlan_info, abort
469 			 * afterwards */
470 			if (vlan_info->nr_vids == 1)
471 				last = true;
472 
473 			unregister_vlan_dev(vlandev, &list);
474 			if (last)
475 				break;
476 		}
477 		unregister_netdevice_many(&list);
478 		break;
479 
480 	case NETDEV_PRE_TYPE_CHANGE:
481 		/* Forbid underlaying device to change its type. */
482 		if (vlan_uses_dev(dev))
483 			return NOTIFY_BAD;
484 		break;
485 
486 	case NETDEV_NOTIFY_PEERS:
487 	case NETDEV_BONDING_FAILOVER:
488 	case NETDEV_RESEND_IGMP:
489 		/* Propagate to vlan devices */
490 		vlan_group_for_each_dev(grp, i, vlandev)
491 			call_netdevice_notifiers(event, vlandev);
492 		break;
493 
494 	case NETDEV_CVLAN_FILTER_PUSH_INFO:
495 		err = vlan_filter_push_vids(vlan_info, htons(ETH_P_8021Q));
496 		if (err)
497 			return notifier_from_errno(err);
498 		break;
499 
500 	case NETDEV_CVLAN_FILTER_DROP_INFO:
501 		vlan_filter_drop_vids(vlan_info, htons(ETH_P_8021Q));
502 		break;
503 
504 	case NETDEV_SVLAN_FILTER_PUSH_INFO:
505 		err = vlan_filter_push_vids(vlan_info, htons(ETH_P_8021AD));
506 		if (err)
507 			return notifier_from_errno(err);
508 		break;
509 
510 	case NETDEV_SVLAN_FILTER_DROP_INFO:
511 		vlan_filter_drop_vids(vlan_info, htons(ETH_P_8021AD));
512 		break;
513 	}
514 
515 out:
516 	return NOTIFY_DONE;
517 }
518 
519 static struct notifier_block vlan_notifier_block __read_mostly = {
520 	.notifier_call = vlan_device_event,
521 };
522 
523 /*
524  *	VLAN IOCTL handler.
525  *	o execute requested action or pass command to the device driver
526  *   arg is really a struct vlan_ioctl_args __user *.
527  */
528 static int vlan_ioctl_handler(struct net *net, void __user *arg)
529 {
530 	int err;
531 	struct vlan_ioctl_args args;
532 	struct net_device *dev = NULL;
533 
534 	if (copy_from_user(&args, arg, sizeof(struct vlan_ioctl_args)))
535 		return -EFAULT;
536 
537 	/* Null terminate this sucker, just in case. */
538 	args.device1[sizeof(args.device1) - 1] = 0;
539 	args.u.device2[sizeof(args.u.device2) - 1] = 0;
540 
541 	rtnl_lock();
542 
543 	switch (args.cmd) {
544 	case SET_VLAN_INGRESS_PRIORITY_CMD:
545 	case SET_VLAN_EGRESS_PRIORITY_CMD:
546 	case SET_VLAN_FLAG_CMD:
547 	case ADD_VLAN_CMD:
548 	case DEL_VLAN_CMD:
549 	case GET_VLAN_REALDEV_NAME_CMD:
550 	case GET_VLAN_VID_CMD:
551 		err = -ENODEV;
552 		dev = __dev_get_by_name(net, args.device1);
553 		if (!dev)
554 			goto out;
555 
556 		err = -EINVAL;
557 		if (args.cmd != ADD_VLAN_CMD && !is_vlan_dev(dev))
558 			goto out;
559 	}
560 
561 	switch (args.cmd) {
562 	case SET_VLAN_INGRESS_PRIORITY_CMD:
563 		err = -EPERM;
564 		if (!ns_capable(net->user_ns, CAP_NET_ADMIN))
565 			break;
566 		vlan_dev_set_ingress_priority(dev,
567 					      args.u.skb_priority,
568 					      args.vlan_qos);
569 		err = 0;
570 		break;
571 
572 	case SET_VLAN_EGRESS_PRIORITY_CMD:
573 		err = -EPERM;
574 		if (!ns_capable(net->user_ns, CAP_NET_ADMIN))
575 			break;
576 		err = vlan_dev_set_egress_priority(dev,
577 						   args.u.skb_priority,
578 						   args.vlan_qos);
579 		break;
580 
581 	case SET_VLAN_FLAG_CMD:
582 		err = -EPERM;
583 		if (!ns_capable(net->user_ns, CAP_NET_ADMIN))
584 			break;
585 		err = vlan_dev_change_flags(dev,
586 					    args.vlan_qos ? args.u.flag : 0,
587 					    args.u.flag);
588 		break;
589 
590 	case SET_VLAN_NAME_TYPE_CMD:
591 		err = -EPERM;
592 		if (!ns_capable(net->user_ns, CAP_NET_ADMIN))
593 			break;
594 		if (args.u.name_type < VLAN_NAME_TYPE_HIGHEST) {
595 			struct vlan_net *vn;
596 
597 			vn = net_generic(net, vlan_net_id);
598 			vn->name_type = args.u.name_type;
599 			err = 0;
600 		} else {
601 			err = -EINVAL;
602 		}
603 		break;
604 
605 	case ADD_VLAN_CMD:
606 		err = -EPERM;
607 		if (!ns_capable(net->user_ns, CAP_NET_ADMIN))
608 			break;
609 		err = register_vlan_device(dev, args.u.VID);
610 		break;
611 
612 	case DEL_VLAN_CMD:
613 		err = -EPERM;
614 		if (!ns_capable(net->user_ns, CAP_NET_ADMIN))
615 			break;
616 		unregister_vlan_dev(dev, NULL);
617 		err = 0;
618 		break;
619 
620 	case GET_VLAN_REALDEV_NAME_CMD:
621 		err = 0;
622 		vlan_dev_get_realdev_name(dev, args.u.device2);
623 		if (copy_to_user(arg, &args,
624 				 sizeof(struct vlan_ioctl_args)))
625 			err = -EFAULT;
626 		break;
627 
628 	case GET_VLAN_VID_CMD:
629 		err = 0;
630 		args.u.VID = vlan_dev_vlan_id(dev);
631 		if (copy_to_user(arg, &args,
632 				 sizeof(struct vlan_ioctl_args)))
633 		      err = -EFAULT;
634 		break;
635 
636 	default:
637 		err = -EOPNOTSUPP;
638 		break;
639 	}
640 out:
641 	rtnl_unlock();
642 	return err;
643 }
644 
645 static struct sk_buff **vlan_gro_receive(struct sk_buff **head,
646 					 struct sk_buff *skb)
647 {
648 	struct sk_buff *p, **pp = NULL;
649 	struct vlan_hdr *vhdr;
650 	unsigned int hlen, off_vlan;
651 	const struct packet_offload *ptype;
652 	__be16 type;
653 	int flush = 1;
654 
655 	off_vlan = skb_gro_offset(skb);
656 	hlen = off_vlan + sizeof(*vhdr);
657 	vhdr = skb_gro_header_fast(skb, off_vlan);
658 	if (skb_gro_header_hard(skb, hlen)) {
659 		vhdr = skb_gro_header_slow(skb, hlen, off_vlan);
660 		if (unlikely(!vhdr))
661 			goto out;
662 	}
663 
664 	type = vhdr->h_vlan_encapsulated_proto;
665 
666 	rcu_read_lock();
667 	ptype = gro_find_receive_by_type(type);
668 	if (!ptype)
669 		goto out_unlock;
670 
671 	flush = 0;
672 
673 	for (p = *head; p; p = p->next) {
674 		struct vlan_hdr *vhdr2;
675 
676 		if (!NAPI_GRO_CB(p)->same_flow)
677 			continue;
678 
679 		vhdr2 = (struct vlan_hdr *)(p->data + off_vlan);
680 		if (compare_vlan_header(vhdr, vhdr2))
681 			NAPI_GRO_CB(p)->same_flow = 0;
682 	}
683 
684 	skb_gro_pull(skb, sizeof(*vhdr));
685 	skb_gro_postpull_rcsum(skb, vhdr, sizeof(*vhdr));
686 	pp = call_gro_receive(ptype->callbacks.gro_receive, head, skb);
687 
688 out_unlock:
689 	rcu_read_unlock();
690 out:
691 	NAPI_GRO_CB(skb)->flush |= flush;
692 
693 	return pp;
694 }
695 
696 static int vlan_gro_complete(struct sk_buff *skb, int nhoff)
697 {
698 	struct vlan_hdr *vhdr = (struct vlan_hdr *)(skb->data + nhoff);
699 	__be16 type = vhdr->h_vlan_encapsulated_proto;
700 	struct packet_offload *ptype;
701 	int err = -ENOENT;
702 
703 	rcu_read_lock();
704 	ptype = gro_find_complete_by_type(type);
705 	if (ptype)
706 		err = ptype->callbacks.gro_complete(skb, nhoff + sizeof(*vhdr));
707 
708 	rcu_read_unlock();
709 	return err;
710 }
711 
712 static struct packet_offload vlan_packet_offloads[] __read_mostly = {
713 	{
714 		.type = cpu_to_be16(ETH_P_8021Q),
715 		.priority = 10,
716 		.callbacks = {
717 			.gro_receive = vlan_gro_receive,
718 			.gro_complete = vlan_gro_complete,
719 		},
720 	},
721 	{
722 		.type = cpu_to_be16(ETH_P_8021AD),
723 		.priority = 10,
724 		.callbacks = {
725 			.gro_receive = vlan_gro_receive,
726 			.gro_complete = vlan_gro_complete,
727 		},
728 	},
729 };
730 
731 static int __net_init vlan_init_net(struct net *net)
732 {
733 	struct vlan_net *vn = net_generic(net, vlan_net_id);
734 	int err;
735 
736 	vn->name_type = VLAN_NAME_TYPE_RAW_PLUS_VID_NO_PAD;
737 
738 	err = vlan_proc_init(net);
739 
740 	return err;
741 }
742 
743 static void __net_exit vlan_exit_net(struct net *net)
744 {
745 	vlan_proc_cleanup(net);
746 }
747 
748 static struct pernet_operations vlan_net_ops = {
749 	.init = vlan_init_net,
750 	.exit = vlan_exit_net,
751 	.id   = &vlan_net_id,
752 	.size = sizeof(struct vlan_net),
753 };
754 
755 static int __init vlan_proto_init(void)
756 {
757 	int err;
758 	unsigned int i;
759 
760 	pr_info("%s v%s\n", vlan_fullname, vlan_version);
761 
762 	err = register_pernet_subsys(&vlan_net_ops);
763 	if (err < 0)
764 		goto err0;
765 
766 	err = register_netdevice_notifier(&vlan_notifier_block);
767 	if (err < 0)
768 		goto err2;
769 
770 	err = vlan_gvrp_init();
771 	if (err < 0)
772 		goto err3;
773 
774 	err = vlan_mvrp_init();
775 	if (err < 0)
776 		goto err4;
777 
778 	err = vlan_netlink_init();
779 	if (err < 0)
780 		goto err5;
781 
782 	for (i = 0; i < ARRAY_SIZE(vlan_packet_offloads); i++)
783 		dev_add_offload(&vlan_packet_offloads[i]);
784 
785 	vlan_ioctl_set(vlan_ioctl_handler);
786 	return 0;
787 
788 err5:
789 	vlan_mvrp_uninit();
790 err4:
791 	vlan_gvrp_uninit();
792 err3:
793 	unregister_netdevice_notifier(&vlan_notifier_block);
794 err2:
795 	unregister_pernet_subsys(&vlan_net_ops);
796 err0:
797 	return err;
798 }
799 
800 static void __exit vlan_cleanup_module(void)
801 {
802 	unsigned int i;
803 
804 	vlan_ioctl_set(NULL);
805 
806 	for (i = 0; i < ARRAY_SIZE(vlan_packet_offloads); i++)
807 		dev_remove_offload(&vlan_packet_offloads[i]);
808 
809 	vlan_netlink_fini();
810 
811 	unregister_netdevice_notifier(&vlan_notifier_block);
812 
813 	unregister_pernet_subsys(&vlan_net_ops);
814 	rcu_barrier(); /* Wait for completion of call_rcu()'s */
815 
816 	vlan_mvrp_uninit();
817 	vlan_gvrp_uninit();
818 }
819 
820 module_init(vlan_proto_init);
821 module_exit(vlan_cleanup_module);
822 
823 MODULE_LICENSE("GPL");
824 MODULE_VERSION(DRV_VERSION);
825