1 /* 2 * INET 802.1Q VLAN 3 * Ethernet-type device handling. 4 * 5 * Authors: Ben Greear <greearb@candelatech.com> 6 * Please send support related email to: netdev@vger.kernel.org 7 * VLAN Home Page: http://www.candelatech.com/~greear/vlan.html 8 * 9 * Fixes: 10 * Fix for packet capture - Nick Eggleston <nick@dccinc.com>; 11 * Add HW acceleration hooks - David S. Miller <davem@redhat.com>; 12 * Correct all the locking - David S. Miller <davem@redhat.com>; 13 * Use hash table for VLAN groups - David S. Miller <davem@redhat.com> 14 * 15 * This program is free software; you can redistribute it and/or 16 * modify it under the terms of the GNU General Public License 17 * as published by the Free Software Foundation; either version 18 * 2 of the License, or (at your option) any later version. 19 */ 20 21 #define pr_fmt(fmt) KBUILD_MODNAME ": " fmt 22 23 #include <linux/capability.h> 24 #include <linux/module.h> 25 #include <linux/netdevice.h> 26 #include <linux/skbuff.h> 27 #include <linux/slab.h> 28 #include <linux/init.h> 29 #include <linux/rculist.h> 30 #include <net/p8022.h> 31 #include <net/arp.h> 32 #include <linux/rtnetlink.h> 33 #include <linux/notifier.h> 34 #include <net/rtnetlink.h> 35 #include <net/net_namespace.h> 36 #include <net/netns/generic.h> 37 #include <linux/uaccess.h> 38 39 #include <linux/if_vlan.h> 40 #include "vlan.h" 41 #include "vlanproc.h" 42 43 #define DRV_VERSION "1.8" 44 45 /* Global VLAN variables */ 46 47 unsigned int vlan_net_id __read_mostly; 48 49 const char vlan_fullname[] = "802.1Q VLAN Support"; 50 const char vlan_version[] = DRV_VERSION; 51 52 /* End of global variables definitions. */ 53 54 static int vlan_group_prealloc_vid(struct vlan_group *vg, 55 __be16 vlan_proto, u16 vlan_id) 56 { 57 struct net_device **array; 58 unsigned int pidx, vidx; 59 unsigned int size; 60 61 ASSERT_RTNL(); 62 63 pidx = vlan_proto_idx(vlan_proto); 64 vidx = vlan_id / VLAN_GROUP_ARRAY_PART_LEN; 65 array = vg->vlan_devices_arrays[pidx][vidx]; 66 if (array != NULL) 67 return 0; 68 69 size = sizeof(struct net_device *) * VLAN_GROUP_ARRAY_PART_LEN; 70 array = kzalloc(size, GFP_KERNEL); 71 if (array == NULL) 72 return -ENOBUFS; 73 74 vg->vlan_devices_arrays[pidx][vidx] = array; 75 return 0; 76 } 77 78 void unregister_vlan_dev(struct net_device *dev, struct list_head *head) 79 { 80 struct vlan_dev_priv *vlan = vlan_dev_priv(dev); 81 struct net_device *real_dev = vlan->real_dev; 82 struct vlan_info *vlan_info; 83 struct vlan_group *grp; 84 u16 vlan_id = vlan->vlan_id; 85 86 ASSERT_RTNL(); 87 88 vlan_info = rtnl_dereference(real_dev->vlan_info); 89 BUG_ON(!vlan_info); 90 91 grp = &vlan_info->grp; 92 93 grp->nr_vlan_devs--; 94 95 if (vlan->flags & VLAN_FLAG_MVRP) 96 vlan_mvrp_request_leave(dev); 97 if (vlan->flags & VLAN_FLAG_GVRP) 98 vlan_gvrp_request_leave(dev); 99 100 vlan_group_set_device(grp, vlan->vlan_proto, vlan_id, NULL); 101 102 netdev_upper_dev_unlink(real_dev, dev); 103 /* Because unregister_netdevice_queue() makes sure at least one rcu 104 * grace period is respected before device freeing, 105 * we dont need to call synchronize_net() here. 106 */ 107 unregister_netdevice_queue(dev, head); 108 109 if (grp->nr_vlan_devs == 0) { 110 vlan_mvrp_uninit_applicant(real_dev); 111 vlan_gvrp_uninit_applicant(real_dev); 112 } 113 114 /* Take it out of our own structures, but be sure to interlock with 115 * HW accelerating devices or SW vlan input packet processing if 116 * VLAN is not 0 (leave it there for 802.1p). 117 */ 118 if (vlan_id) 119 vlan_vid_del(real_dev, vlan->vlan_proto, vlan_id); 120 121 /* Get rid of the vlan's reference to real_dev */ 122 dev_put(real_dev); 123 } 124 125 int vlan_check_real_dev(struct net_device *real_dev, 126 __be16 protocol, u16 vlan_id) 127 { 128 const char *name = real_dev->name; 129 130 if (real_dev->features & NETIF_F_VLAN_CHALLENGED) { 131 pr_info("VLANs not supported on %s\n", name); 132 return -EOPNOTSUPP; 133 } 134 135 if (vlan_find_dev(real_dev, protocol, vlan_id) != NULL) 136 return -EEXIST; 137 138 return 0; 139 } 140 141 int register_vlan_dev(struct net_device *dev, struct netlink_ext_ack *extack) 142 { 143 struct vlan_dev_priv *vlan = vlan_dev_priv(dev); 144 struct net_device *real_dev = vlan->real_dev; 145 u16 vlan_id = vlan->vlan_id; 146 struct vlan_info *vlan_info; 147 struct vlan_group *grp; 148 int err; 149 150 err = vlan_vid_add(real_dev, vlan->vlan_proto, vlan_id); 151 if (err) 152 return err; 153 154 vlan_info = rtnl_dereference(real_dev->vlan_info); 155 /* vlan_info should be there now. vlan_vid_add took care of it */ 156 BUG_ON(!vlan_info); 157 158 grp = &vlan_info->grp; 159 if (grp->nr_vlan_devs == 0) { 160 err = vlan_gvrp_init_applicant(real_dev); 161 if (err < 0) 162 goto out_vid_del; 163 err = vlan_mvrp_init_applicant(real_dev); 164 if (err < 0) 165 goto out_uninit_gvrp; 166 } 167 168 err = vlan_group_prealloc_vid(grp, vlan->vlan_proto, vlan_id); 169 if (err < 0) 170 goto out_uninit_mvrp; 171 172 vlan->nest_level = dev_get_nest_level(real_dev) + 1; 173 err = register_netdevice(dev); 174 if (err < 0) 175 goto out_uninit_mvrp; 176 177 err = netdev_upper_dev_link(real_dev, dev, extack); 178 if (err) 179 goto out_unregister_netdev; 180 181 /* Account for reference in struct vlan_dev_priv */ 182 dev_hold(real_dev); 183 184 netif_stacked_transfer_operstate(real_dev, dev); 185 linkwatch_fire_event(dev); /* _MUST_ call rfc2863_policy() */ 186 187 /* So, got the sucker initialized, now lets place 188 * it into our local structure. 189 */ 190 vlan_group_set_device(grp, vlan->vlan_proto, vlan_id, dev); 191 grp->nr_vlan_devs++; 192 193 return 0; 194 195 out_unregister_netdev: 196 unregister_netdevice(dev); 197 out_uninit_mvrp: 198 if (grp->nr_vlan_devs == 0) 199 vlan_mvrp_uninit_applicant(real_dev); 200 out_uninit_gvrp: 201 if (grp->nr_vlan_devs == 0) 202 vlan_gvrp_uninit_applicant(real_dev); 203 out_vid_del: 204 vlan_vid_del(real_dev, vlan->vlan_proto, vlan_id); 205 return err; 206 } 207 208 /* Attach a VLAN device to a mac address (ie Ethernet Card). 209 * Returns 0 if the device was created or a negative error code otherwise. 210 */ 211 static int register_vlan_device(struct net_device *real_dev, u16 vlan_id) 212 { 213 struct net_device *new_dev; 214 struct vlan_dev_priv *vlan; 215 struct net *net = dev_net(real_dev); 216 struct vlan_net *vn = net_generic(net, vlan_net_id); 217 char name[IFNAMSIZ]; 218 int err; 219 220 if (vlan_id >= VLAN_VID_MASK) 221 return -ERANGE; 222 223 err = vlan_check_real_dev(real_dev, htons(ETH_P_8021Q), vlan_id); 224 if (err < 0) 225 return err; 226 227 /* Gotta set up the fields for the device. */ 228 switch (vn->name_type) { 229 case VLAN_NAME_TYPE_RAW_PLUS_VID: 230 /* name will look like: eth1.0005 */ 231 snprintf(name, IFNAMSIZ, "%s.%.4i", real_dev->name, vlan_id); 232 break; 233 case VLAN_NAME_TYPE_PLUS_VID_NO_PAD: 234 /* Put our vlan.VID in the name. 235 * Name will look like: vlan5 236 */ 237 snprintf(name, IFNAMSIZ, "vlan%i", vlan_id); 238 break; 239 case VLAN_NAME_TYPE_RAW_PLUS_VID_NO_PAD: 240 /* Put our vlan.VID in the name. 241 * Name will look like: eth0.5 242 */ 243 snprintf(name, IFNAMSIZ, "%s.%i", real_dev->name, vlan_id); 244 break; 245 case VLAN_NAME_TYPE_PLUS_VID: 246 /* Put our vlan.VID in the name. 247 * Name will look like: vlan0005 248 */ 249 default: 250 snprintf(name, IFNAMSIZ, "vlan%.4i", vlan_id); 251 } 252 253 new_dev = alloc_netdev(sizeof(struct vlan_dev_priv), name, 254 NET_NAME_UNKNOWN, vlan_setup); 255 256 if (new_dev == NULL) 257 return -ENOBUFS; 258 259 dev_net_set(new_dev, net); 260 /* need 4 bytes for extra VLAN header info, 261 * hope the underlying device can handle it. 262 */ 263 new_dev->mtu = real_dev->mtu; 264 265 vlan = vlan_dev_priv(new_dev); 266 vlan->vlan_proto = htons(ETH_P_8021Q); 267 vlan->vlan_id = vlan_id; 268 vlan->real_dev = real_dev; 269 vlan->dent = NULL; 270 vlan->flags = VLAN_FLAG_REORDER_HDR; 271 272 new_dev->rtnl_link_ops = &vlan_link_ops; 273 err = register_vlan_dev(new_dev, NULL); 274 if (err < 0) 275 goto out_free_newdev; 276 277 return 0; 278 279 out_free_newdev: 280 if (new_dev->reg_state == NETREG_UNINITIALIZED) 281 free_netdev(new_dev); 282 return err; 283 } 284 285 static void vlan_sync_address(struct net_device *dev, 286 struct net_device *vlandev) 287 { 288 struct vlan_dev_priv *vlan = vlan_dev_priv(vlandev); 289 290 /* May be called without an actual change */ 291 if (ether_addr_equal(vlan->real_dev_addr, dev->dev_addr)) 292 return; 293 294 /* vlan continues to inherit address of lower device */ 295 if (vlan_dev_inherit_address(vlandev, dev)) 296 goto out; 297 298 /* vlan address was different from the old address and is equal to 299 * the new address */ 300 if (!ether_addr_equal(vlandev->dev_addr, vlan->real_dev_addr) && 301 ether_addr_equal(vlandev->dev_addr, dev->dev_addr)) 302 dev_uc_del(dev, vlandev->dev_addr); 303 304 /* vlan address was equal to the old address and is different from 305 * the new address */ 306 if (ether_addr_equal(vlandev->dev_addr, vlan->real_dev_addr) && 307 !ether_addr_equal(vlandev->dev_addr, dev->dev_addr)) 308 dev_uc_add(dev, vlandev->dev_addr); 309 310 out: 311 ether_addr_copy(vlan->real_dev_addr, dev->dev_addr); 312 } 313 314 static void vlan_transfer_features(struct net_device *dev, 315 struct net_device *vlandev) 316 { 317 struct vlan_dev_priv *vlan = vlan_dev_priv(vlandev); 318 319 vlandev->gso_max_size = dev->gso_max_size; 320 vlandev->gso_max_segs = dev->gso_max_segs; 321 322 if (vlan_hw_offload_capable(dev->features, vlan->vlan_proto)) 323 vlandev->hard_header_len = dev->hard_header_len; 324 else 325 vlandev->hard_header_len = dev->hard_header_len + VLAN_HLEN; 326 327 #if IS_ENABLED(CONFIG_FCOE) 328 vlandev->fcoe_ddp_xid = dev->fcoe_ddp_xid; 329 #endif 330 331 vlandev->priv_flags &= ~IFF_XMIT_DST_RELEASE; 332 vlandev->priv_flags |= (vlan->real_dev->priv_flags & IFF_XMIT_DST_RELEASE); 333 334 netdev_update_features(vlandev); 335 } 336 337 static int __vlan_device_event(struct net_device *dev, unsigned long event) 338 { 339 int err = 0; 340 341 switch (event) { 342 case NETDEV_CHANGENAME: 343 vlan_proc_rem_dev(dev); 344 err = vlan_proc_add_dev(dev); 345 break; 346 case NETDEV_REGISTER: 347 err = vlan_proc_add_dev(dev); 348 break; 349 case NETDEV_UNREGISTER: 350 vlan_proc_rem_dev(dev); 351 break; 352 } 353 354 return err; 355 } 356 357 static int vlan_device_event(struct notifier_block *unused, unsigned long event, 358 void *ptr) 359 { 360 struct net_device *dev = netdev_notifier_info_to_dev(ptr); 361 struct vlan_group *grp; 362 struct vlan_info *vlan_info; 363 int i, flgs; 364 struct net_device *vlandev; 365 struct vlan_dev_priv *vlan; 366 bool last = false; 367 LIST_HEAD(list); 368 369 if (is_vlan_dev(dev)) { 370 int err = __vlan_device_event(dev, event); 371 372 if (err) 373 return notifier_from_errno(err); 374 } 375 376 if ((event == NETDEV_UP) && 377 (dev->features & NETIF_F_HW_VLAN_CTAG_FILTER)) { 378 pr_info("adding VLAN 0 to HW filter on device %s\n", 379 dev->name); 380 vlan_vid_add(dev, htons(ETH_P_8021Q), 0); 381 } 382 if (event == NETDEV_DOWN && 383 (dev->features & NETIF_F_HW_VLAN_CTAG_FILTER)) 384 vlan_vid_del(dev, htons(ETH_P_8021Q), 0); 385 386 vlan_info = rtnl_dereference(dev->vlan_info); 387 if (!vlan_info) 388 goto out; 389 grp = &vlan_info->grp; 390 391 /* It is OK that we do not hold the group lock right now, 392 * as we run under the RTNL lock. 393 */ 394 395 switch (event) { 396 case NETDEV_CHANGE: 397 /* Propagate real device state to vlan devices */ 398 vlan_group_for_each_dev(grp, i, vlandev) 399 netif_stacked_transfer_operstate(dev, vlandev); 400 break; 401 402 case NETDEV_CHANGEADDR: 403 /* Adjust unicast filters on underlying device */ 404 vlan_group_for_each_dev(grp, i, vlandev) { 405 flgs = vlandev->flags; 406 if (!(flgs & IFF_UP)) 407 continue; 408 409 vlan_sync_address(dev, vlandev); 410 } 411 break; 412 413 case NETDEV_CHANGEMTU: 414 vlan_group_for_each_dev(grp, i, vlandev) { 415 if (vlandev->mtu <= dev->mtu) 416 continue; 417 418 dev_set_mtu(vlandev, dev->mtu); 419 } 420 break; 421 422 case NETDEV_FEAT_CHANGE: 423 /* Propagate device features to underlying device */ 424 vlan_group_for_each_dev(grp, i, vlandev) 425 vlan_transfer_features(dev, vlandev); 426 break; 427 428 case NETDEV_DOWN: { 429 struct net_device *tmp; 430 LIST_HEAD(close_list); 431 432 /* Put all VLANs for this dev in the down state too. */ 433 vlan_group_for_each_dev(grp, i, vlandev) { 434 flgs = vlandev->flags; 435 if (!(flgs & IFF_UP)) 436 continue; 437 438 vlan = vlan_dev_priv(vlandev); 439 if (!(vlan->flags & VLAN_FLAG_LOOSE_BINDING)) 440 list_add(&vlandev->close_list, &close_list); 441 } 442 443 dev_close_many(&close_list, false); 444 445 list_for_each_entry_safe(vlandev, tmp, &close_list, close_list) { 446 netif_stacked_transfer_operstate(dev, vlandev); 447 list_del_init(&vlandev->close_list); 448 } 449 list_del(&close_list); 450 break; 451 } 452 case NETDEV_UP: 453 /* Put all VLANs for this dev in the up state too. */ 454 vlan_group_for_each_dev(grp, i, vlandev) { 455 flgs = dev_get_flags(vlandev); 456 if (flgs & IFF_UP) 457 continue; 458 459 vlan = vlan_dev_priv(vlandev); 460 if (!(vlan->flags & VLAN_FLAG_LOOSE_BINDING)) 461 dev_change_flags(vlandev, flgs | IFF_UP); 462 netif_stacked_transfer_operstate(dev, vlandev); 463 } 464 break; 465 466 case NETDEV_UNREGISTER: 467 /* twiddle thumbs on netns device moves */ 468 if (dev->reg_state != NETREG_UNREGISTERING) 469 break; 470 471 vlan_group_for_each_dev(grp, i, vlandev) { 472 /* removal of last vid destroys vlan_info, abort 473 * afterwards */ 474 if (vlan_info->nr_vids == 1) 475 last = true; 476 477 unregister_vlan_dev(vlandev, &list); 478 if (last) 479 break; 480 } 481 unregister_netdevice_many(&list); 482 break; 483 484 case NETDEV_PRE_TYPE_CHANGE: 485 /* Forbid underlaying device to change its type. */ 486 if (vlan_uses_dev(dev)) 487 return NOTIFY_BAD; 488 break; 489 490 case NETDEV_NOTIFY_PEERS: 491 case NETDEV_BONDING_FAILOVER: 492 case NETDEV_RESEND_IGMP: 493 /* Propagate to vlan devices */ 494 vlan_group_for_each_dev(grp, i, vlandev) 495 call_netdevice_notifiers(event, vlandev); 496 break; 497 } 498 499 out: 500 return NOTIFY_DONE; 501 } 502 503 static struct notifier_block vlan_notifier_block __read_mostly = { 504 .notifier_call = vlan_device_event, 505 }; 506 507 /* 508 * VLAN IOCTL handler. 509 * o execute requested action or pass command to the device driver 510 * arg is really a struct vlan_ioctl_args __user *. 511 */ 512 static int vlan_ioctl_handler(struct net *net, void __user *arg) 513 { 514 int err; 515 struct vlan_ioctl_args args; 516 struct net_device *dev = NULL; 517 518 if (copy_from_user(&args, arg, sizeof(struct vlan_ioctl_args))) 519 return -EFAULT; 520 521 /* Null terminate this sucker, just in case. */ 522 args.device1[sizeof(args.device1) - 1] = 0; 523 args.u.device2[sizeof(args.u.device2) - 1] = 0; 524 525 rtnl_lock(); 526 527 switch (args.cmd) { 528 case SET_VLAN_INGRESS_PRIORITY_CMD: 529 case SET_VLAN_EGRESS_PRIORITY_CMD: 530 case SET_VLAN_FLAG_CMD: 531 case ADD_VLAN_CMD: 532 case DEL_VLAN_CMD: 533 case GET_VLAN_REALDEV_NAME_CMD: 534 case GET_VLAN_VID_CMD: 535 err = -ENODEV; 536 dev = __dev_get_by_name(net, args.device1); 537 if (!dev) 538 goto out; 539 540 err = -EINVAL; 541 if (args.cmd != ADD_VLAN_CMD && !is_vlan_dev(dev)) 542 goto out; 543 } 544 545 switch (args.cmd) { 546 case SET_VLAN_INGRESS_PRIORITY_CMD: 547 err = -EPERM; 548 if (!ns_capable(net->user_ns, CAP_NET_ADMIN)) 549 break; 550 vlan_dev_set_ingress_priority(dev, 551 args.u.skb_priority, 552 args.vlan_qos); 553 err = 0; 554 break; 555 556 case SET_VLAN_EGRESS_PRIORITY_CMD: 557 err = -EPERM; 558 if (!ns_capable(net->user_ns, CAP_NET_ADMIN)) 559 break; 560 err = vlan_dev_set_egress_priority(dev, 561 args.u.skb_priority, 562 args.vlan_qos); 563 break; 564 565 case SET_VLAN_FLAG_CMD: 566 err = -EPERM; 567 if (!ns_capable(net->user_ns, CAP_NET_ADMIN)) 568 break; 569 err = vlan_dev_change_flags(dev, 570 args.vlan_qos ? args.u.flag : 0, 571 args.u.flag); 572 break; 573 574 case SET_VLAN_NAME_TYPE_CMD: 575 err = -EPERM; 576 if (!ns_capable(net->user_ns, CAP_NET_ADMIN)) 577 break; 578 if (args.u.name_type < VLAN_NAME_TYPE_HIGHEST) { 579 struct vlan_net *vn; 580 581 vn = net_generic(net, vlan_net_id); 582 vn->name_type = args.u.name_type; 583 err = 0; 584 } else { 585 err = -EINVAL; 586 } 587 break; 588 589 case ADD_VLAN_CMD: 590 err = -EPERM; 591 if (!ns_capable(net->user_ns, CAP_NET_ADMIN)) 592 break; 593 err = register_vlan_device(dev, args.u.VID); 594 break; 595 596 case DEL_VLAN_CMD: 597 err = -EPERM; 598 if (!ns_capable(net->user_ns, CAP_NET_ADMIN)) 599 break; 600 unregister_vlan_dev(dev, NULL); 601 err = 0; 602 break; 603 604 case GET_VLAN_REALDEV_NAME_CMD: 605 err = 0; 606 vlan_dev_get_realdev_name(dev, args.u.device2); 607 if (copy_to_user(arg, &args, 608 sizeof(struct vlan_ioctl_args))) 609 err = -EFAULT; 610 break; 611 612 case GET_VLAN_VID_CMD: 613 err = 0; 614 args.u.VID = vlan_dev_vlan_id(dev); 615 if (copy_to_user(arg, &args, 616 sizeof(struct vlan_ioctl_args))) 617 err = -EFAULT; 618 break; 619 620 default: 621 err = -EOPNOTSUPP; 622 break; 623 } 624 out: 625 rtnl_unlock(); 626 return err; 627 } 628 629 static struct sk_buff **vlan_gro_receive(struct sk_buff **head, 630 struct sk_buff *skb) 631 { 632 struct sk_buff *p, **pp = NULL; 633 struct vlan_hdr *vhdr; 634 unsigned int hlen, off_vlan; 635 const struct packet_offload *ptype; 636 __be16 type; 637 int flush = 1; 638 639 off_vlan = skb_gro_offset(skb); 640 hlen = off_vlan + sizeof(*vhdr); 641 vhdr = skb_gro_header_fast(skb, off_vlan); 642 if (skb_gro_header_hard(skb, hlen)) { 643 vhdr = skb_gro_header_slow(skb, hlen, off_vlan); 644 if (unlikely(!vhdr)) 645 goto out; 646 } 647 648 type = vhdr->h_vlan_encapsulated_proto; 649 650 rcu_read_lock(); 651 ptype = gro_find_receive_by_type(type); 652 if (!ptype) 653 goto out_unlock; 654 655 flush = 0; 656 657 for (p = *head; p; p = p->next) { 658 struct vlan_hdr *vhdr2; 659 660 if (!NAPI_GRO_CB(p)->same_flow) 661 continue; 662 663 vhdr2 = (struct vlan_hdr *)(p->data + off_vlan); 664 if (compare_vlan_header(vhdr, vhdr2)) 665 NAPI_GRO_CB(p)->same_flow = 0; 666 } 667 668 skb_gro_pull(skb, sizeof(*vhdr)); 669 skb_gro_postpull_rcsum(skb, vhdr, sizeof(*vhdr)); 670 pp = call_gro_receive(ptype->callbacks.gro_receive, head, skb); 671 672 out_unlock: 673 rcu_read_unlock(); 674 out: 675 NAPI_GRO_CB(skb)->flush |= flush; 676 677 return pp; 678 } 679 680 static int vlan_gro_complete(struct sk_buff *skb, int nhoff) 681 { 682 struct vlan_hdr *vhdr = (struct vlan_hdr *)(skb->data + nhoff); 683 __be16 type = vhdr->h_vlan_encapsulated_proto; 684 struct packet_offload *ptype; 685 int err = -ENOENT; 686 687 rcu_read_lock(); 688 ptype = gro_find_complete_by_type(type); 689 if (ptype) 690 err = ptype->callbacks.gro_complete(skb, nhoff + sizeof(*vhdr)); 691 692 rcu_read_unlock(); 693 return err; 694 } 695 696 static struct packet_offload vlan_packet_offloads[] __read_mostly = { 697 { 698 .type = cpu_to_be16(ETH_P_8021Q), 699 .priority = 10, 700 .callbacks = { 701 .gro_receive = vlan_gro_receive, 702 .gro_complete = vlan_gro_complete, 703 }, 704 }, 705 { 706 .type = cpu_to_be16(ETH_P_8021AD), 707 .priority = 10, 708 .callbacks = { 709 .gro_receive = vlan_gro_receive, 710 .gro_complete = vlan_gro_complete, 711 }, 712 }, 713 }; 714 715 static int __net_init vlan_init_net(struct net *net) 716 { 717 struct vlan_net *vn = net_generic(net, vlan_net_id); 718 int err; 719 720 vn->name_type = VLAN_NAME_TYPE_RAW_PLUS_VID_NO_PAD; 721 722 err = vlan_proc_init(net); 723 724 return err; 725 } 726 727 static void __net_exit vlan_exit_net(struct net *net) 728 { 729 vlan_proc_cleanup(net); 730 } 731 732 static struct pernet_operations vlan_net_ops = { 733 .init = vlan_init_net, 734 .exit = vlan_exit_net, 735 .id = &vlan_net_id, 736 .size = sizeof(struct vlan_net), 737 }; 738 739 static int __init vlan_proto_init(void) 740 { 741 int err; 742 unsigned int i; 743 744 pr_info("%s v%s\n", vlan_fullname, vlan_version); 745 746 err = register_pernet_subsys(&vlan_net_ops); 747 if (err < 0) 748 goto err0; 749 750 err = register_netdevice_notifier(&vlan_notifier_block); 751 if (err < 0) 752 goto err2; 753 754 err = vlan_gvrp_init(); 755 if (err < 0) 756 goto err3; 757 758 err = vlan_mvrp_init(); 759 if (err < 0) 760 goto err4; 761 762 err = vlan_netlink_init(); 763 if (err < 0) 764 goto err5; 765 766 for (i = 0; i < ARRAY_SIZE(vlan_packet_offloads); i++) 767 dev_add_offload(&vlan_packet_offloads[i]); 768 769 vlan_ioctl_set(vlan_ioctl_handler); 770 return 0; 771 772 err5: 773 vlan_mvrp_uninit(); 774 err4: 775 vlan_gvrp_uninit(); 776 err3: 777 unregister_netdevice_notifier(&vlan_notifier_block); 778 err2: 779 unregister_pernet_subsys(&vlan_net_ops); 780 err0: 781 return err; 782 } 783 784 static void __exit vlan_cleanup_module(void) 785 { 786 unsigned int i; 787 788 vlan_ioctl_set(NULL); 789 790 for (i = 0; i < ARRAY_SIZE(vlan_packet_offloads); i++) 791 dev_remove_offload(&vlan_packet_offloads[i]); 792 793 vlan_netlink_fini(); 794 795 unregister_netdevice_notifier(&vlan_notifier_block); 796 797 unregister_pernet_subsys(&vlan_net_ops); 798 rcu_barrier(); /* Wait for completion of call_rcu()'s */ 799 800 vlan_mvrp_uninit(); 801 vlan_gvrp_uninit(); 802 } 803 804 module_init(vlan_proto_init); 805 module_exit(vlan_cleanup_module); 806 807 MODULE_LICENSE("GPL"); 808 MODULE_VERSION(DRV_VERSION); 809