1 /* 2 * INET 802.1Q VLAN 3 * Ethernet-type device handling. 4 * 5 * Authors: Ben Greear <greearb@candelatech.com> 6 * Please send support related email to: netdev@vger.kernel.org 7 * VLAN Home Page: http://www.candelatech.com/~greear/vlan.html 8 * 9 * Fixes: 10 * Fix for packet capture - Nick Eggleston <nick@dccinc.com>; 11 * Add HW acceleration hooks - David S. Miller <davem@redhat.com>; 12 * Correct all the locking - David S. Miller <davem@redhat.com>; 13 * Use hash table for VLAN groups - David S. Miller <davem@redhat.com> 14 * 15 * This program is free software; you can redistribute it and/or 16 * modify it under the terms of the GNU General Public License 17 * as published by the Free Software Foundation; either version 18 * 2 of the License, or (at your option) any later version. 19 */ 20 21 #include <linux/capability.h> 22 #include <linux/module.h> 23 #include <linux/netdevice.h> 24 #include <linux/skbuff.h> 25 #include <linux/init.h> 26 #include <linux/rculist.h> 27 #include <net/p8022.h> 28 #include <net/arp.h> 29 #include <linux/rtnetlink.h> 30 #include <linux/notifier.h> 31 #include <net/rtnetlink.h> 32 #include <net/net_namespace.h> 33 #include <net/netns/generic.h> 34 #include <asm/uaccess.h> 35 36 #include <linux/if_vlan.h> 37 #include "vlan.h" 38 #include "vlanproc.h" 39 40 #define DRV_VERSION "1.8" 41 42 /* Global VLAN variables */ 43 44 int vlan_net_id __read_mostly; 45 46 /* Our listing of VLAN group(s) */ 47 static struct hlist_head vlan_group_hash[VLAN_GRP_HASH_SIZE]; 48 49 const char vlan_fullname[] = "802.1Q VLAN Support"; 50 const char vlan_version[] = DRV_VERSION; 51 static const char vlan_copyright[] = "Ben Greear <greearb@candelatech.com>"; 52 static const char vlan_buggyright[] = "David S. Miller <davem@redhat.com>"; 53 54 static struct packet_type vlan_packet_type __read_mostly = { 55 .type = cpu_to_be16(ETH_P_8021Q), 56 .func = vlan_skb_recv, /* VLAN receive method */ 57 }; 58 59 /* End of global variables definitions. */ 60 61 static inline unsigned int vlan_grp_hashfn(unsigned int idx) 62 { 63 return ((idx >> VLAN_GRP_HASH_SHIFT) ^ idx) & VLAN_GRP_HASH_MASK; 64 } 65 66 /* Must be invoked with RCU read lock (no preempt) */ 67 static struct vlan_group *__vlan_find_group(struct net_device *real_dev) 68 { 69 struct vlan_group *grp; 70 struct hlist_node *n; 71 int hash = vlan_grp_hashfn(real_dev->ifindex); 72 73 hlist_for_each_entry_rcu(grp, n, &vlan_group_hash[hash], hlist) { 74 if (grp->real_dev == real_dev) 75 return grp; 76 } 77 78 return NULL; 79 } 80 81 /* Find the protocol handler. Assumes VID < VLAN_VID_MASK. 82 * 83 * Must be invoked with RCU read lock (no preempt) 84 */ 85 struct net_device *__find_vlan_dev(struct net_device *real_dev, u16 vlan_id) 86 { 87 struct vlan_group *grp = __vlan_find_group(real_dev); 88 89 if (grp) 90 return vlan_group_get_device(grp, vlan_id); 91 92 return NULL; 93 } 94 95 static void vlan_group_free(struct vlan_group *grp) 96 { 97 int i; 98 99 for (i = 0; i < VLAN_GROUP_ARRAY_SPLIT_PARTS; i++) 100 kfree(grp->vlan_devices_arrays[i]); 101 kfree(grp); 102 } 103 104 static struct vlan_group *vlan_group_alloc(struct net_device *real_dev) 105 { 106 struct vlan_group *grp; 107 108 grp = kzalloc(sizeof(struct vlan_group), GFP_KERNEL); 109 if (!grp) 110 return NULL; 111 112 grp->real_dev = real_dev; 113 hlist_add_head_rcu(&grp->hlist, 114 &vlan_group_hash[vlan_grp_hashfn(real_dev->ifindex)]); 115 return grp; 116 } 117 118 static int vlan_group_prealloc_vid(struct vlan_group *vg, u16 vlan_id) 119 { 120 struct net_device **array; 121 unsigned int size; 122 123 ASSERT_RTNL(); 124 125 array = vg->vlan_devices_arrays[vlan_id / VLAN_GROUP_ARRAY_PART_LEN]; 126 if (array != NULL) 127 return 0; 128 129 size = sizeof(struct net_device *) * VLAN_GROUP_ARRAY_PART_LEN; 130 array = kzalloc(size, GFP_KERNEL); 131 if (array == NULL) 132 return -ENOBUFS; 133 134 vg->vlan_devices_arrays[vlan_id / VLAN_GROUP_ARRAY_PART_LEN] = array; 135 return 0; 136 } 137 138 static void vlan_rcu_free(struct rcu_head *rcu) 139 { 140 vlan_group_free(container_of(rcu, struct vlan_group, rcu)); 141 } 142 143 void unregister_vlan_dev(struct net_device *dev, struct list_head *head) 144 { 145 struct vlan_dev_info *vlan = vlan_dev_info(dev); 146 struct net_device *real_dev = vlan->real_dev; 147 const struct net_device_ops *ops = real_dev->netdev_ops; 148 struct vlan_group *grp; 149 u16 vlan_id = vlan->vlan_id; 150 151 ASSERT_RTNL(); 152 153 grp = __vlan_find_group(real_dev); 154 BUG_ON(!grp); 155 156 /* Take it out of our own structures, but be sure to interlock with 157 * HW accelerating devices or SW vlan input packet processing. 158 */ 159 if (real_dev->features & NETIF_F_HW_VLAN_FILTER) 160 ops->ndo_vlan_rx_kill_vid(real_dev, vlan_id); 161 162 grp->nr_vlans--; 163 164 vlan_group_set_device(grp, vlan_id, NULL); 165 if (!grp->killall) 166 synchronize_net(); 167 168 unregister_netdevice_queue(dev, head); 169 170 /* If the group is now empty, kill off the group. */ 171 if (grp->nr_vlans == 0) { 172 vlan_gvrp_uninit_applicant(real_dev); 173 174 if (real_dev->features & NETIF_F_HW_VLAN_RX) 175 ops->ndo_vlan_rx_register(real_dev, NULL); 176 177 hlist_del_rcu(&grp->hlist); 178 179 /* Free the group, after all cpu's are done. */ 180 call_rcu(&grp->rcu, vlan_rcu_free); 181 } 182 183 /* Get rid of the vlan's reference to real_dev */ 184 dev_put(real_dev); 185 } 186 187 int vlan_check_real_dev(struct net_device *real_dev, u16 vlan_id) 188 { 189 const char *name = real_dev->name; 190 const struct net_device_ops *ops = real_dev->netdev_ops; 191 192 if (real_dev->features & NETIF_F_VLAN_CHALLENGED) { 193 pr_info("8021q: VLANs not supported on %s\n", name); 194 return -EOPNOTSUPP; 195 } 196 197 if ((real_dev->features & NETIF_F_HW_VLAN_RX) && !ops->ndo_vlan_rx_register) { 198 pr_info("8021q: device %s has buggy VLAN hw accel\n", name); 199 return -EOPNOTSUPP; 200 } 201 202 if ((real_dev->features & NETIF_F_HW_VLAN_FILTER) && 203 (!ops->ndo_vlan_rx_add_vid || !ops->ndo_vlan_rx_kill_vid)) { 204 pr_info("8021q: Device %s has buggy VLAN hw accel\n", name); 205 return -EOPNOTSUPP; 206 } 207 208 if (__find_vlan_dev(real_dev, vlan_id) != NULL) 209 return -EEXIST; 210 211 return 0; 212 } 213 214 int register_vlan_dev(struct net_device *dev) 215 { 216 struct vlan_dev_info *vlan = vlan_dev_info(dev); 217 struct net_device *real_dev = vlan->real_dev; 218 const struct net_device_ops *ops = real_dev->netdev_ops; 219 u16 vlan_id = vlan->vlan_id; 220 struct vlan_group *grp, *ngrp = NULL; 221 int err; 222 223 grp = __vlan_find_group(real_dev); 224 if (!grp) { 225 ngrp = grp = vlan_group_alloc(real_dev); 226 if (!grp) 227 return -ENOBUFS; 228 err = vlan_gvrp_init_applicant(real_dev); 229 if (err < 0) 230 goto out_free_group; 231 } 232 233 err = vlan_group_prealloc_vid(grp, vlan_id); 234 if (err < 0) 235 goto out_uninit_applicant; 236 237 err = register_netdevice(dev); 238 if (err < 0) 239 goto out_uninit_applicant; 240 241 /* Account for reference in struct vlan_dev_info */ 242 dev_hold(real_dev); 243 244 netif_stacked_transfer_operstate(real_dev, dev); 245 linkwatch_fire_event(dev); /* _MUST_ call rfc2863_policy() */ 246 247 /* So, got the sucker initialized, now lets place 248 * it into our local structure. 249 */ 250 vlan_group_set_device(grp, vlan_id, dev); 251 grp->nr_vlans++; 252 253 if (ngrp && real_dev->features & NETIF_F_HW_VLAN_RX) 254 ops->ndo_vlan_rx_register(real_dev, ngrp); 255 if (real_dev->features & NETIF_F_HW_VLAN_FILTER) 256 ops->ndo_vlan_rx_add_vid(real_dev, vlan_id); 257 258 return 0; 259 260 out_uninit_applicant: 261 if (ngrp) 262 vlan_gvrp_uninit_applicant(real_dev); 263 out_free_group: 264 if (ngrp) { 265 hlist_del_rcu(&ngrp->hlist); 266 /* Free the group, after all cpu's are done. */ 267 call_rcu(&ngrp->rcu, vlan_rcu_free); 268 } 269 return err; 270 } 271 272 /* Attach a VLAN device to a mac address (ie Ethernet Card). 273 * Returns 0 if the device was created or a negative error code otherwise. 274 */ 275 static int register_vlan_device(struct net_device *real_dev, u16 vlan_id) 276 { 277 struct net_device *new_dev; 278 struct net *net = dev_net(real_dev); 279 struct vlan_net *vn = net_generic(net, vlan_net_id); 280 char name[IFNAMSIZ]; 281 int err; 282 283 if (vlan_id >= VLAN_VID_MASK) 284 return -ERANGE; 285 286 err = vlan_check_real_dev(real_dev, vlan_id); 287 if (err < 0) 288 return err; 289 290 /* Gotta set up the fields for the device. */ 291 switch (vn->name_type) { 292 case VLAN_NAME_TYPE_RAW_PLUS_VID: 293 /* name will look like: eth1.0005 */ 294 snprintf(name, IFNAMSIZ, "%s.%.4i", real_dev->name, vlan_id); 295 break; 296 case VLAN_NAME_TYPE_PLUS_VID_NO_PAD: 297 /* Put our vlan.VID in the name. 298 * Name will look like: vlan5 299 */ 300 snprintf(name, IFNAMSIZ, "vlan%i", vlan_id); 301 break; 302 case VLAN_NAME_TYPE_RAW_PLUS_VID_NO_PAD: 303 /* Put our vlan.VID in the name. 304 * Name will look like: eth0.5 305 */ 306 snprintf(name, IFNAMSIZ, "%s.%i", real_dev->name, vlan_id); 307 break; 308 case VLAN_NAME_TYPE_PLUS_VID: 309 /* Put our vlan.VID in the name. 310 * Name will look like: vlan0005 311 */ 312 default: 313 snprintf(name, IFNAMSIZ, "vlan%.4i", vlan_id); 314 } 315 316 new_dev = alloc_netdev_mq(sizeof(struct vlan_dev_info), name, 317 vlan_setup, real_dev->num_tx_queues); 318 319 if (new_dev == NULL) 320 return -ENOBUFS; 321 322 new_dev->real_num_tx_queues = real_dev->real_num_tx_queues; 323 dev_net_set(new_dev, net); 324 /* need 4 bytes for extra VLAN header info, 325 * hope the underlying device can handle it. 326 */ 327 new_dev->mtu = real_dev->mtu; 328 329 vlan_dev_info(new_dev)->vlan_id = vlan_id; 330 vlan_dev_info(new_dev)->real_dev = real_dev; 331 vlan_dev_info(new_dev)->dent = NULL; 332 vlan_dev_info(new_dev)->flags = VLAN_FLAG_REORDER_HDR; 333 334 new_dev->rtnl_link_ops = &vlan_link_ops; 335 err = register_vlan_dev(new_dev); 336 if (err < 0) 337 goto out_free_newdev; 338 339 return 0; 340 341 out_free_newdev: 342 free_netdev(new_dev); 343 return err; 344 } 345 346 static void vlan_sync_address(struct net_device *dev, 347 struct net_device *vlandev) 348 { 349 struct vlan_dev_info *vlan = vlan_dev_info(vlandev); 350 351 /* May be called without an actual change */ 352 if (!compare_ether_addr(vlan->real_dev_addr, dev->dev_addr)) 353 return; 354 355 /* vlan address was different from the old address and is equal to 356 * the new address */ 357 if (compare_ether_addr(vlandev->dev_addr, vlan->real_dev_addr) && 358 !compare_ether_addr(vlandev->dev_addr, dev->dev_addr)) 359 dev_unicast_delete(dev, vlandev->dev_addr); 360 361 /* vlan address was equal to the old address and is different from 362 * the new address */ 363 if (!compare_ether_addr(vlandev->dev_addr, vlan->real_dev_addr) && 364 compare_ether_addr(vlandev->dev_addr, dev->dev_addr)) 365 dev_unicast_add(dev, vlandev->dev_addr); 366 367 memcpy(vlan->real_dev_addr, dev->dev_addr, ETH_ALEN); 368 } 369 370 static void vlan_transfer_features(struct net_device *dev, 371 struct net_device *vlandev) 372 { 373 unsigned long old_features = vlandev->features; 374 375 vlandev->features &= ~dev->vlan_features; 376 vlandev->features |= dev->features & dev->vlan_features; 377 vlandev->gso_max_size = dev->gso_max_size; 378 #if defined(CONFIG_FCOE) || defined(CONFIG_FCOE_MODULE) 379 vlandev->fcoe_ddp_xid = dev->fcoe_ddp_xid; 380 #endif 381 382 if (old_features != vlandev->features) 383 netdev_features_change(vlandev); 384 } 385 386 static void __vlan_device_event(struct net_device *dev, unsigned long event) 387 { 388 switch (event) { 389 case NETDEV_CHANGENAME: 390 vlan_proc_rem_dev(dev); 391 if (vlan_proc_add_dev(dev) < 0) 392 pr_warning("8021q: failed to change proc name for %s\n", 393 dev->name); 394 break; 395 case NETDEV_REGISTER: 396 if (vlan_proc_add_dev(dev) < 0) 397 pr_warning("8021q: failed to add proc entry for %s\n", 398 dev->name); 399 break; 400 case NETDEV_UNREGISTER: 401 vlan_proc_rem_dev(dev); 402 break; 403 } 404 } 405 406 static int vlan_device_event(struct notifier_block *unused, unsigned long event, 407 void *ptr) 408 { 409 struct net_device *dev = ptr; 410 struct vlan_group *grp; 411 int i, flgs; 412 struct net_device *vlandev; 413 struct vlan_dev_info *vlan; 414 LIST_HEAD(list); 415 416 if (is_vlan_dev(dev)) 417 __vlan_device_event(dev, event); 418 419 grp = __vlan_find_group(dev); 420 if (!grp) 421 goto out; 422 423 /* It is OK that we do not hold the group lock right now, 424 * as we run under the RTNL lock. 425 */ 426 427 switch (event) { 428 case NETDEV_CHANGE: 429 /* Propagate real device state to vlan devices */ 430 for (i = 0; i < VLAN_GROUP_ARRAY_LEN; i++) { 431 vlandev = vlan_group_get_device(grp, i); 432 if (!vlandev) 433 continue; 434 435 netif_stacked_transfer_operstate(dev, vlandev); 436 } 437 break; 438 439 case NETDEV_CHANGEADDR: 440 /* Adjust unicast filters on underlying device */ 441 for (i = 0; i < VLAN_GROUP_ARRAY_LEN; i++) { 442 vlandev = vlan_group_get_device(grp, i); 443 if (!vlandev) 444 continue; 445 446 flgs = vlandev->flags; 447 if (!(flgs & IFF_UP)) 448 continue; 449 450 vlan_sync_address(dev, vlandev); 451 } 452 break; 453 454 case NETDEV_CHANGEMTU: 455 for (i = 0; i < VLAN_GROUP_ARRAY_LEN; i++) { 456 vlandev = vlan_group_get_device(grp, i); 457 if (!vlandev) 458 continue; 459 460 if (vlandev->mtu <= dev->mtu) 461 continue; 462 463 dev_set_mtu(vlandev, dev->mtu); 464 } 465 break; 466 467 case NETDEV_FEAT_CHANGE: 468 /* Propagate device features to underlying device */ 469 for (i = 0; i < VLAN_GROUP_ARRAY_LEN; i++) { 470 vlandev = vlan_group_get_device(grp, i); 471 if (!vlandev) 472 continue; 473 474 vlan_transfer_features(dev, vlandev); 475 } 476 477 break; 478 479 case NETDEV_DOWN: 480 /* Put all VLANs for this dev in the down state too. */ 481 for (i = 0; i < VLAN_GROUP_ARRAY_LEN; i++) { 482 vlandev = vlan_group_get_device(grp, i); 483 if (!vlandev) 484 continue; 485 486 flgs = vlandev->flags; 487 if (!(flgs & IFF_UP)) 488 continue; 489 490 vlan = vlan_dev_info(vlandev); 491 if (!(vlan->flags & VLAN_FLAG_LOOSE_BINDING)) 492 dev_change_flags(vlandev, flgs & ~IFF_UP); 493 netif_stacked_transfer_operstate(dev, vlandev); 494 } 495 break; 496 497 case NETDEV_UP: 498 /* Put all VLANs for this dev in the up state too. */ 499 for (i = 0; i < VLAN_GROUP_ARRAY_LEN; i++) { 500 vlandev = vlan_group_get_device(grp, i); 501 if (!vlandev) 502 continue; 503 504 flgs = vlandev->flags; 505 if (flgs & IFF_UP) 506 continue; 507 508 vlan = vlan_dev_info(vlandev); 509 if (!(vlan->flags & VLAN_FLAG_LOOSE_BINDING)) 510 dev_change_flags(vlandev, flgs | IFF_UP); 511 netif_stacked_transfer_operstate(dev, vlandev); 512 } 513 break; 514 515 case NETDEV_UNREGISTER: 516 /* Delete all VLANs for this dev. */ 517 grp->killall = 1; 518 519 for (i = 0; i < VLAN_GROUP_ARRAY_LEN; i++) { 520 vlandev = vlan_group_get_device(grp, i); 521 if (!vlandev) 522 continue; 523 524 /* unregistration of last vlan destroys group, abort 525 * afterwards */ 526 if (grp->nr_vlans == 1) 527 i = VLAN_GROUP_ARRAY_LEN; 528 529 unregister_vlan_dev(vlandev, &list); 530 } 531 unregister_netdevice_many(&list); 532 break; 533 } 534 535 out: 536 return NOTIFY_DONE; 537 } 538 539 static struct notifier_block vlan_notifier_block __read_mostly = { 540 .notifier_call = vlan_device_event, 541 }; 542 543 /* 544 * VLAN IOCTL handler. 545 * o execute requested action or pass command to the device driver 546 * arg is really a struct vlan_ioctl_args __user *. 547 */ 548 static int vlan_ioctl_handler(struct net *net, void __user *arg) 549 { 550 int err; 551 struct vlan_ioctl_args args; 552 struct net_device *dev = NULL; 553 554 if (copy_from_user(&args, arg, sizeof(struct vlan_ioctl_args))) 555 return -EFAULT; 556 557 /* Null terminate this sucker, just in case. */ 558 args.device1[23] = 0; 559 args.u.device2[23] = 0; 560 561 rtnl_lock(); 562 563 switch (args.cmd) { 564 case SET_VLAN_INGRESS_PRIORITY_CMD: 565 case SET_VLAN_EGRESS_PRIORITY_CMD: 566 case SET_VLAN_FLAG_CMD: 567 case ADD_VLAN_CMD: 568 case DEL_VLAN_CMD: 569 case GET_VLAN_REALDEV_NAME_CMD: 570 case GET_VLAN_VID_CMD: 571 err = -ENODEV; 572 dev = __dev_get_by_name(net, args.device1); 573 if (!dev) 574 goto out; 575 576 err = -EINVAL; 577 if (args.cmd != ADD_VLAN_CMD && !is_vlan_dev(dev)) 578 goto out; 579 } 580 581 switch (args.cmd) { 582 case SET_VLAN_INGRESS_PRIORITY_CMD: 583 err = -EPERM; 584 if (!capable(CAP_NET_ADMIN)) 585 break; 586 vlan_dev_set_ingress_priority(dev, 587 args.u.skb_priority, 588 args.vlan_qos); 589 err = 0; 590 break; 591 592 case SET_VLAN_EGRESS_PRIORITY_CMD: 593 err = -EPERM; 594 if (!capable(CAP_NET_ADMIN)) 595 break; 596 err = vlan_dev_set_egress_priority(dev, 597 args.u.skb_priority, 598 args.vlan_qos); 599 break; 600 601 case SET_VLAN_FLAG_CMD: 602 err = -EPERM; 603 if (!capable(CAP_NET_ADMIN)) 604 break; 605 err = vlan_dev_change_flags(dev, 606 args.vlan_qos ? args.u.flag : 0, 607 args.u.flag); 608 break; 609 610 case SET_VLAN_NAME_TYPE_CMD: 611 err = -EPERM; 612 if (!capable(CAP_NET_ADMIN)) 613 break; 614 if ((args.u.name_type >= 0) && 615 (args.u.name_type < VLAN_NAME_TYPE_HIGHEST)) { 616 struct vlan_net *vn; 617 618 vn = net_generic(net, vlan_net_id); 619 vn->name_type = args.u.name_type; 620 err = 0; 621 } else { 622 err = -EINVAL; 623 } 624 break; 625 626 case ADD_VLAN_CMD: 627 err = -EPERM; 628 if (!capable(CAP_NET_ADMIN)) 629 break; 630 err = register_vlan_device(dev, args.u.VID); 631 break; 632 633 case DEL_VLAN_CMD: 634 err = -EPERM; 635 if (!capable(CAP_NET_ADMIN)) 636 break; 637 unregister_vlan_dev(dev, NULL); 638 err = 0; 639 break; 640 641 case GET_VLAN_REALDEV_NAME_CMD: 642 err = 0; 643 vlan_dev_get_realdev_name(dev, args.u.device2); 644 if (copy_to_user(arg, &args, 645 sizeof(struct vlan_ioctl_args))) 646 err = -EFAULT; 647 break; 648 649 case GET_VLAN_VID_CMD: 650 err = 0; 651 args.u.VID = vlan_dev_vlan_id(dev); 652 if (copy_to_user(arg, &args, 653 sizeof(struct vlan_ioctl_args))) 654 err = -EFAULT; 655 break; 656 657 default: 658 err = -EOPNOTSUPP; 659 break; 660 } 661 out: 662 rtnl_unlock(); 663 return err; 664 } 665 666 static int vlan_init_net(struct net *net) 667 { 668 struct vlan_net *vn = net_generic(net, vlan_net_id); 669 int err; 670 671 vn->name_type = VLAN_NAME_TYPE_RAW_PLUS_VID_NO_PAD; 672 673 err = vlan_proc_init(net); 674 675 return err; 676 } 677 678 static void vlan_exit_net(struct net *net) 679 { 680 vlan_proc_cleanup(net); 681 } 682 683 static struct pernet_operations vlan_net_ops = { 684 .init = vlan_init_net, 685 .exit = vlan_exit_net, 686 .id = &vlan_net_id, 687 .size = sizeof(struct vlan_net), 688 }; 689 690 static int __init vlan_proto_init(void) 691 { 692 int err; 693 694 pr_info("%s v%s %s\n", vlan_fullname, vlan_version, vlan_copyright); 695 pr_info("All bugs added by %s\n", vlan_buggyright); 696 697 err = register_pernet_subsys(&vlan_net_ops); 698 if (err < 0) 699 goto err0; 700 701 err = register_netdevice_notifier(&vlan_notifier_block); 702 if (err < 0) 703 goto err2; 704 705 err = vlan_gvrp_init(); 706 if (err < 0) 707 goto err3; 708 709 err = vlan_netlink_init(); 710 if (err < 0) 711 goto err4; 712 713 dev_add_pack(&vlan_packet_type); 714 vlan_ioctl_set(vlan_ioctl_handler); 715 return 0; 716 717 err4: 718 vlan_gvrp_uninit(); 719 err3: 720 unregister_netdevice_notifier(&vlan_notifier_block); 721 err2: 722 unregister_pernet_subsys(&vlan_net_ops); 723 err0: 724 return err; 725 } 726 727 static void __exit vlan_cleanup_module(void) 728 { 729 unsigned int i; 730 731 vlan_ioctl_set(NULL); 732 vlan_netlink_fini(); 733 734 unregister_netdevice_notifier(&vlan_notifier_block); 735 736 dev_remove_pack(&vlan_packet_type); 737 738 /* This table must be empty if there are no module references left. */ 739 for (i = 0; i < VLAN_GRP_HASH_SIZE; i++) 740 BUG_ON(!hlist_empty(&vlan_group_hash[i])); 741 742 unregister_pernet_subsys(&vlan_net_ops); 743 rcu_barrier(); /* Wait for completion of call_rcu()'s */ 744 745 vlan_gvrp_uninit(); 746 } 747 748 module_init(vlan_proto_init); 749 module_exit(vlan_cleanup_module); 750 751 MODULE_LICENSE("GPL"); 752 MODULE_VERSION(DRV_VERSION); 753