1 /* 2 * Copyright (C) 2001 - 2007 Jeff Dike (jdike@{addtoit,linux.intel}.com) 3 * Copyright (C) 2001 Lennert Buytenhek (buytenh@gnu.org) and 4 * James Leu (jleu@mindspring.net). 5 * Copyright (C) 2001 by various other people who didn't put their name here. 6 * Licensed under the GPL. 7 */ 8 9 #include <linux/bootmem.h> 10 #include <linux/etherdevice.h> 11 #include <linux/ethtool.h> 12 #include <linux/inetdevice.h> 13 #include <linux/init.h> 14 #include <linux/list.h> 15 #include <linux/netdevice.h> 16 #include <linux/platform_device.h> 17 #include <linux/rtnetlink.h> 18 #include <linux/skbuff.h> 19 #include <linux/spinlock.h> 20 #include "init.h" 21 #include "irq_kern.h" 22 #include "irq_user.h" 23 #include "mconsole_kern.h" 24 #include "net_kern.h" 25 #include "net_user.h" 26 27 static inline void set_ether_mac(struct net_device *dev, unsigned char *addr) 28 { 29 memcpy(dev->dev_addr, addr, ETH_ALEN); 30 } 31 32 #define DRIVER_NAME "uml-netdev" 33 34 static DEFINE_SPINLOCK(opened_lock); 35 static LIST_HEAD(opened); 36 37 /* 38 * The drop_skb is used when we can't allocate an skb. The 39 * packet is read into drop_skb in order to get the data off the 40 * connection to the host. 41 * It is reallocated whenever a maximum packet size is seen which is 42 * larger than any seen before. update_drop_skb is called from 43 * eth_configure when a new interface is added. 44 */ 45 static DEFINE_SPINLOCK(drop_lock); 46 static struct sk_buff *drop_skb; 47 static int drop_max; 48 49 static int update_drop_skb(int max) 50 { 51 struct sk_buff *new; 52 unsigned long flags; 53 int err = 0; 54 55 spin_lock_irqsave(&drop_lock, flags); 56 57 if (max <= drop_max) 58 goto out; 59 60 err = -ENOMEM; 61 new = dev_alloc_skb(max); 62 if (new == NULL) 63 goto out; 64 65 skb_put(new, max); 66 67 kfree_skb(drop_skb); 68 drop_skb = new; 69 drop_max = max; 70 err = 0; 71 out: 72 spin_unlock_irqrestore(&drop_lock, flags); 73 74 return err; 75 } 76 77 static int uml_net_rx(struct net_device *dev) 78 { 79 struct uml_net_private *lp = netdev_priv(dev); 80 int pkt_len; 81 struct sk_buff *skb; 82 83 /* If we can't allocate memory, try again next round. */ 84 skb = dev_alloc_skb(lp->max_packet); 85 if (skb == NULL) { 86 drop_skb->dev = dev; 87 /* Read a packet into drop_skb and don't do anything with it. */ 88 (*lp->read)(lp->fd, drop_skb, lp); 89 dev->stats.rx_dropped++; 90 return 0; 91 } 92 93 skb->dev = dev; 94 skb_put(skb, lp->max_packet); 95 skb_reset_mac_header(skb); 96 pkt_len = (*lp->read)(lp->fd, skb, lp); 97 98 if (pkt_len > 0) { 99 skb_trim(skb, pkt_len); 100 skb->protocol = (*lp->protocol)(skb); 101 102 dev->stats.rx_bytes += skb->len; 103 dev->stats.rx_packets++; 104 netif_rx(skb); 105 return pkt_len; 106 } 107 108 kfree_skb(skb); 109 return pkt_len; 110 } 111 112 static void uml_dev_close(struct work_struct *work) 113 { 114 struct uml_net_private *lp = 115 container_of(work, struct uml_net_private, work); 116 dev_close(lp->dev); 117 } 118 119 static irqreturn_t uml_net_interrupt(int irq, void *dev_id) 120 { 121 struct net_device *dev = dev_id; 122 struct uml_net_private *lp = netdev_priv(dev); 123 int err; 124 125 if (!netif_running(dev)) 126 return IRQ_NONE; 127 128 spin_lock(&lp->lock); 129 while ((err = uml_net_rx(dev)) > 0) ; 130 if (err < 0) { 131 printk(KERN_ERR 132 "Device '%s' read returned %d, shutting it down\n", 133 dev->name, err); 134 /* dev_close can't be called in interrupt context, and takes 135 * again lp->lock. 136 * And dev_close() can be safely called multiple times on the 137 * same device, since it tests for (dev->flags & IFF_UP). So 138 * there's no harm in delaying the device shutdown. 139 * Furthermore, the workqueue will not re-enqueue an already 140 * enqueued work item. */ 141 schedule_work(&lp->work); 142 goto out; 143 } 144 reactivate_fd(lp->fd, UM_ETH_IRQ); 145 146 out: 147 spin_unlock(&lp->lock); 148 return IRQ_HANDLED; 149 } 150 151 static int uml_net_open(struct net_device *dev) 152 { 153 struct uml_net_private *lp = netdev_priv(dev); 154 int err; 155 156 if (lp->fd >= 0) { 157 err = -ENXIO; 158 goto out; 159 } 160 161 lp->fd = (*lp->open)(&lp->user); 162 if (lp->fd < 0) { 163 err = lp->fd; 164 goto out; 165 } 166 167 err = um_request_irq(dev->irq, lp->fd, IRQ_READ, uml_net_interrupt, 168 IRQF_DISABLED | IRQF_SHARED, dev->name, dev); 169 if (err != 0) { 170 printk(KERN_ERR "uml_net_open: failed to get irq(%d)\n", err); 171 err = -ENETUNREACH; 172 goto out_close; 173 } 174 175 lp->tl.data = (unsigned long) &lp->user; 176 netif_start_queue(dev); 177 178 /* clear buffer - it can happen that the host side of the interface 179 * is full when we get here. In this case, new data is never queued, 180 * SIGIOs never arrive, and the net never works. 181 */ 182 while ((err = uml_net_rx(dev)) > 0) ; 183 184 spin_lock(&opened_lock); 185 list_add(&lp->list, &opened); 186 spin_unlock(&opened_lock); 187 188 return 0; 189 out_close: 190 if (lp->close != NULL) (*lp->close)(lp->fd, &lp->user); 191 lp->fd = -1; 192 out: 193 return err; 194 } 195 196 static int uml_net_close(struct net_device *dev) 197 { 198 struct uml_net_private *lp = netdev_priv(dev); 199 200 netif_stop_queue(dev); 201 202 free_irq(dev->irq, dev); 203 if (lp->close != NULL) 204 (*lp->close)(lp->fd, &lp->user); 205 lp->fd = -1; 206 207 spin_lock(&opened_lock); 208 list_del(&lp->list); 209 spin_unlock(&opened_lock); 210 211 return 0; 212 } 213 214 static int uml_net_start_xmit(struct sk_buff *skb, struct net_device *dev) 215 { 216 struct uml_net_private *lp = netdev_priv(dev); 217 unsigned long flags; 218 int len; 219 220 netif_stop_queue(dev); 221 222 spin_lock_irqsave(&lp->lock, flags); 223 224 len = (*lp->write)(lp->fd, skb, lp); 225 226 if (len == skb->len) { 227 dev->stats.tx_packets++; 228 dev->stats.tx_bytes += skb->len; 229 dev->trans_start = jiffies; 230 netif_start_queue(dev); 231 232 /* this is normally done in the interrupt when tx finishes */ 233 netif_wake_queue(dev); 234 } 235 else if (len == 0) { 236 netif_start_queue(dev); 237 dev->stats.tx_dropped++; 238 } 239 else { 240 netif_start_queue(dev); 241 printk(KERN_ERR "uml_net_start_xmit: failed(%d)\n", len); 242 } 243 244 spin_unlock_irqrestore(&lp->lock, flags); 245 246 dev_kfree_skb(skb); 247 248 return 0; 249 } 250 251 static void uml_net_set_multicast_list(struct net_device *dev) 252 { 253 return; 254 } 255 256 static void uml_net_tx_timeout(struct net_device *dev) 257 { 258 dev->trans_start = jiffies; 259 netif_wake_queue(dev); 260 } 261 262 static int uml_net_set_mac(struct net_device *dev, void *addr) 263 { 264 struct uml_net_private *lp = netdev_priv(dev); 265 struct sockaddr *hwaddr = addr; 266 267 spin_lock_irq(&lp->lock); 268 set_ether_mac(dev, hwaddr->sa_data); 269 spin_unlock_irq(&lp->lock); 270 271 return 0; 272 } 273 274 static int uml_net_change_mtu(struct net_device *dev, int new_mtu) 275 { 276 dev->mtu = new_mtu; 277 278 return 0; 279 } 280 281 static void uml_net_get_drvinfo(struct net_device *dev, 282 struct ethtool_drvinfo *info) 283 { 284 strcpy(info->driver, DRIVER_NAME); 285 strcpy(info->version, "42"); 286 } 287 288 static struct ethtool_ops uml_net_ethtool_ops = { 289 .get_drvinfo = uml_net_get_drvinfo, 290 .get_link = ethtool_op_get_link, 291 }; 292 293 static void uml_net_user_timer_expire(unsigned long _conn) 294 { 295 #ifdef undef 296 struct connection *conn = (struct connection *)_conn; 297 298 dprintk(KERN_INFO "uml_net_user_timer_expire [%p]\n", conn); 299 do_connect(conn); 300 #endif 301 } 302 303 static void setup_etheraddr(char *str, unsigned char *addr, char *name) 304 { 305 char *end; 306 int i; 307 308 if (str == NULL) 309 goto random; 310 311 for (i = 0; i < 6; i++) { 312 addr[i] = simple_strtoul(str, &end, 16); 313 if ((end == str) || 314 ((*end != ':') && (*end != ',') && (*end != '\0'))) { 315 printk(KERN_ERR 316 "setup_etheraddr: failed to parse '%s' " 317 "as an ethernet address\n", str); 318 goto random; 319 } 320 str = end + 1; 321 } 322 if (is_multicast_ether_addr(addr)) { 323 printk(KERN_ERR 324 "Attempt to assign a multicast ethernet address to a " 325 "device disallowed\n"); 326 goto random; 327 } 328 if (!is_valid_ether_addr(addr)) { 329 printk(KERN_ERR 330 "Attempt to assign an invalid ethernet address to a " 331 "device disallowed\n"); 332 goto random; 333 } 334 if (!is_local_ether_addr(addr)) { 335 printk(KERN_WARNING 336 "Warning: Assigning a globally valid ethernet " 337 "address to a device\n"); 338 printk(KERN_WARNING "You should set the 2nd rightmost bit in " 339 "the first byte of the MAC,\n"); 340 printk(KERN_WARNING "i.e. %02x:%02x:%02x:%02x:%02x:%02x\n", 341 addr[0] | 0x02, addr[1], addr[2], addr[3], addr[4], 342 addr[5]); 343 } 344 return; 345 346 random: 347 printk(KERN_INFO 348 "Choosing a random ethernet address for device %s\n", name); 349 random_ether_addr(addr); 350 } 351 352 static DEFINE_SPINLOCK(devices_lock); 353 static LIST_HEAD(devices); 354 355 static struct platform_driver uml_net_driver = { 356 .driver = { 357 .name = DRIVER_NAME, 358 }, 359 }; 360 361 static void net_device_release(struct device *dev) 362 { 363 struct uml_net *device = dev->driver_data; 364 struct net_device *netdev = device->dev; 365 struct uml_net_private *lp = netdev_priv(netdev); 366 367 if (lp->remove != NULL) 368 (*lp->remove)(&lp->user); 369 list_del(&device->list); 370 kfree(device); 371 free_netdev(netdev); 372 } 373 374 static const struct net_device_ops uml_netdev_ops = { 375 .ndo_open = uml_net_open, 376 .ndo_stop = uml_net_close, 377 .ndo_start_xmit = uml_net_start_xmit, 378 .ndo_set_multicast_list = uml_net_set_multicast_list, 379 .ndo_tx_timeout = uml_net_tx_timeout, 380 .ndo_set_mac_address = uml_net_set_mac, 381 .ndo_change_mtu = uml_net_change_mtu, 382 .ndo_set_mac_address = eth_mac_addr, 383 .ndo_validate_addr = eth_validate_addr, 384 }; 385 386 /* 387 * Ensures that platform_driver_register is called only once by 388 * eth_configure. Will be set in an initcall. 389 */ 390 static int driver_registered; 391 392 static void eth_configure(int n, void *init, char *mac, 393 struct transport *transport) 394 { 395 struct uml_net *device; 396 struct net_device *dev; 397 struct uml_net_private *lp; 398 int err, size; 399 400 size = transport->private_size + sizeof(struct uml_net_private); 401 402 device = kzalloc(sizeof(*device), GFP_KERNEL); 403 if (device == NULL) { 404 printk(KERN_ERR "eth_configure failed to allocate struct " 405 "uml_net\n"); 406 return; 407 } 408 409 dev = alloc_etherdev(size); 410 if (dev == NULL) { 411 printk(KERN_ERR "eth_configure: failed to allocate struct " 412 "net_device for eth%d\n", n); 413 goto out_free_device; 414 } 415 416 INIT_LIST_HEAD(&device->list); 417 device->index = n; 418 419 /* If this name ends up conflicting with an existing registered 420 * netdevice, that is OK, register_netdev{,ice}() will notice this 421 * and fail. 422 */ 423 snprintf(dev->name, sizeof(dev->name), "eth%d", n); 424 425 setup_etheraddr(mac, device->mac, dev->name); 426 427 printk(KERN_INFO "Netdevice %d (%pM) : ", n, device->mac); 428 429 lp = netdev_priv(dev); 430 /* This points to the transport private data. It's still clear, but we 431 * must memset it to 0 *now*. Let's help the drivers. */ 432 memset(lp, 0, size); 433 INIT_WORK(&lp->work, uml_dev_close); 434 435 /* sysfs register */ 436 if (!driver_registered) { 437 platform_driver_register(¨_net_driver); 438 driver_registered = 1; 439 } 440 device->pdev.id = n; 441 device->pdev.name = DRIVER_NAME; 442 device->pdev.dev.release = net_device_release; 443 device->pdev.dev.driver_data = device; 444 if (platform_device_register(&device->pdev)) 445 goto out_free_netdev; 446 SET_NETDEV_DEV(dev,&device->pdev.dev); 447 448 device->dev = dev; 449 450 /* 451 * These just fill in a data structure, so there's no failure 452 * to be worried about. 453 */ 454 (*transport->kern->init)(dev, init); 455 456 *lp = ((struct uml_net_private) 457 { .list = LIST_HEAD_INIT(lp->list), 458 .dev = dev, 459 .fd = -1, 460 .mac = { 0xfe, 0xfd, 0x0, 0x0, 0x0, 0x0}, 461 .max_packet = transport->user->max_packet, 462 .protocol = transport->kern->protocol, 463 .open = transport->user->open, 464 .close = transport->user->close, 465 .remove = transport->user->remove, 466 .read = transport->kern->read, 467 .write = transport->kern->write, 468 .add_address = transport->user->add_address, 469 .delete_address = transport->user->delete_address }); 470 471 init_timer(&lp->tl); 472 spin_lock_init(&lp->lock); 473 lp->tl.function = uml_net_user_timer_expire; 474 memcpy(lp->mac, device->mac, sizeof(lp->mac)); 475 476 if ((transport->user->init != NULL) && 477 ((*transport->user->init)(&lp->user, dev) != 0)) 478 goto out_unregister; 479 480 set_ether_mac(dev, device->mac); 481 dev->mtu = transport->user->mtu; 482 dev->netdev_ops = ¨_netdev_ops; 483 dev->ethtool_ops = ¨_net_ethtool_ops; 484 dev->watchdog_timeo = (HZ >> 1); 485 dev->irq = UM_ETH_IRQ; 486 487 err = update_drop_skb(lp->max_packet); 488 if (err) 489 goto out_undo_user_init; 490 491 rtnl_lock(); 492 err = register_netdevice(dev); 493 rtnl_unlock(); 494 if (err) 495 goto out_undo_user_init; 496 497 spin_lock(&devices_lock); 498 list_add(&device->list, &devices); 499 spin_unlock(&devices_lock); 500 501 return; 502 503 out_undo_user_init: 504 if (transport->user->remove != NULL) 505 (*transport->user->remove)(&lp->user); 506 out_unregister: 507 platform_device_unregister(&device->pdev); 508 return; /* platform_device_unregister frees dev and device */ 509 out_free_netdev: 510 free_netdev(dev); 511 out_free_device: 512 kfree(device); 513 } 514 515 static struct uml_net *find_device(int n) 516 { 517 struct uml_net *device; 518 struct list_head *ele; 519 520 spin_lock(&devices_lock); 521 list_for_each(ele, &devices) { 522 device = list_entry(ele, struct uml_net, list); 523 if (device->index == n) 524 goto out; 525 } 526 device = NULL; 527 out: 528 spin_unlock(&devices_lock); 529 return device; 530 } 531 532 static int eth_parse(char *str, int *index_out, char **str_out, 533 char **error_out) 534 { 535 char *end; 536 int n, err = -EINVAL;; 537 538 n = simple_strtoul(str, &end, 0); 539 if (end == str) { 540 *error_out = "Bad device number"; 541 return err; 542 } 543 544 str = end; 545 if (*str != '=') { 546 *error_out = "Expected '=' after device number"; 547 return err; 548 } 549 550 str++; 551 if (find_device(n)) { 552 *error_out = "Device already configured"; 553 return err; 554 } 555 556 *index_out = n; 557 *str_out = str; 558 return 0; 559 } 560 561 struct eth_init { 562 struct list_head list; 563 char *init; 564 int index; 565 }; 566 567 static DEFINE_SPINLOCK(transports_lock); 568 static LIST_HEAD(transports); 569 570 /* Filled in during early boot */ 571 static LIST_HEAD(eth_cmd_line); 572 573 static int check_transport(struct transport *transport, char *eth, int n, 574 void **init_out, char **mac_out) 575 { 576 int len; 577 578 len = strlen(transport->name); 579 if (strncmp(eth, transport->name, len)) 580 return 0; 581 582 eth += len; 583 if (*eth == ',') 584 eth++; 585 else if (*eth != '\0') 586 return 0; 587 588 *init_out = kmalloc(transport->setup_size, GFP_KERNEL); 589 if (*init_out == NULL) 590 return 1; 591 592 if (!transport->setup(eth, mac_out, *init_out)) { 593 kfree(*init_out); 594 *init_out = NULL; 595 } 596 return 1; 597 } 598 599 void register_transport(struct transport *new) 600 { 601 struct list_head *ele, *next; 602 struct eth_init *eth; 603 void *init; 604 char *mac = NULL; 605 int match; 606 607 spin_lock(&transports_lock); 608 BUG_ON(!list_empty(&new->list)); 609 list_add(&new->list, &transports); 610 spin_unlock(&transports_lock); 611 612 list_for_each_safe(ele, next, ð_cmd_line) { 613 eth = list_entry(ele, struct eth_init, list); 614 match = check_transport(new, eth->init, eth->index, &init, 615 &mac); 616 if (!match) 617 continue; 618 else if (init != NULL) { 619 eth_configure(eth->index, init, mac, new); 620 kfree(init); 621 } 622 list_del(ð->list); 623 } 624 } 625 626 static int eth_setup_common(char *str, int index) 627 { 628 struct list_head *ele; 629 struct transport *transport; 630 void *init; 631 char *mac = NULL; 632 int found = 0; 633 634 spin_lock(&transports_lock); 635 list_for_each(ele, &transports) { 636 transport = list_entry(ele, struct transport, list); 637 if (!check_transport(transport, str, index, &init, &mac)) 638 continue; 639 if (init != NULL) { 640 eth_configure(index, init, mac, transport); 641 kfree(init); 642 } 643 found = 1; 644 break; 645 } 646 647 spin_unlock(&transports_lock); 648 return found; 649 } 650 651 static int __init eth_setup(char *str) 652 { 653 struct eth_init *new; 654 char *error; 655 int n, err; 656 657 err = eth_parse(str, &n, &str, &error); 658 if (err) { 659 printk(KERN_ERR "eth_setup - Couldn't parse '%s' : %s\n", 660 str, error); 661 return 1; 662 } 663 664 new = alloc_bootmem(sizeof(*new)); 665 if (new == NULL) { 666 printk(KERN_ERR "eth_init : alloc_bootmem failed\n"); 667 return 1; 668 } 669 670 INIT_LIST_HEAD(&new->list); 671 new->index = n; 672 new->init = str; 673 674 list_add_tail(&new->list, ð_cmd_line); 675 return 1; 676 } 677 678 __setup("eth", eth_setup); 679 __uml_help(eth_setup, 680 "eth[0-9]+=<transport>,<options>\n" 681 " Configure a network device.\n\n" 682 ); 683 684 static int net_config(char *str, char **error_out) 685 { 686 int n, err; 687 688 err = eth_parse(str, &n, &str, error_out); 689 if (err) 690 return err; 691 692 /* This string is broken up and the pieces used by the underlying 693 * driver. So, it is freed only if eth_setup_common fails. 694 */ 695 str = kstrdup(str, GFP_KERNEL); 696 if (str == NULL) { 697 *error_out = "net_config failed to strdup string"; 698 return -ENOMEM; 699 } 700 err = !eth_setup_common(str, n); 701 if (err) 702 kfree(str); 703 return err; 704 } 705 706 static int net_id(char **str, int *start_out, int *end_out) 707 { 708 char *end; 709 int n; 710 711 n = simple_strtoul(*str, &end, 0); 712 if ((*end != '\0') || (end == *str)) 713 return -1; 714 715 *start_out = n; 716 *end_out = n; 717 *str = end; 718 return n; 719 } 720 721 static int net_remove(int n, char **error_out) 722 { 723 struct uml_net *device; 724 struct net_device *dev; 725 struct uml_net_private *lp; 726 727 device = find_device(n); 728 if (device == NULL) 729 return -ENODEV; 730 731 dev = device->dev; 732 lp = netdev_priv(dev); 733 if (lp->fd > 0) 734 return -EBUSY; 735 unregister_netdev(dev); 736 platform_device_unregister(&device->pdev); 737 738 return 0; 739 } 740 741 static struct mc_device net_mc = { 742 .list = LIST_HEAD_INIT(net_mc.list), 743 .name = "eth", 744 .config = net_config, 745 .get_config = NULL, 746 .id = net_id, 747 .remove = net_remove, 748 }; 749 750 #ifdef CONFIG_INET 751 static int uml_inetaddr_event(struct notifier_block *this, unsigned long event, 752 void *ptr) 753 { 754 struct in_ifaddr *ifa = ptr; 755 struct net_device *dev = ifa->ifa_dev->dev; 756 struct uml_net_private *lp; 757 void (*proc)(unsigned char *, unsigned char *, void *); 758 unsigned char addr_buf[4], netmask_buf[4]; 759 760 if (dev->netdev_ops->ndo_open != uml_net_open) 761 return NOTIFY_DONE; 762 763 lp = netdev_priv(dev); 764 765 proc = NULL; 766 switch (event) { 767 case NETDEV_UP: 768 proc = lp->add_address; 769 break; 770 case NETDEV_DOWN: 771 proc = lp->delete_address; 772 break; 773 } 774 if (proc != NULL) { 775 memcpy(addr_buf, &ifa->ifa_address, sizeof(addr_buf)); 776 memcpy(netmask_buf, &ifa->ifa_mask, sizeof(netmask_buf)); 777 (*proc)(addr_buf, netmask_buf, &lp->user); 778 } 779 return NOTIFY_DONE; 780 } 781 782 /* uml_net_init shouldn't be called twice on two CPUs at the same time */ 783 static struct notifier_block uml_inetaddr_notifier = { 784 .notifier_call = uml_inetaddr_event, 785 }; 786 787 static void inet_register(void) 788 { 789 struct list_head *ele; 790 struct uml_net_private *lp; 791 struct in_device *ip; 792 struct in_ifaddr *in; 793 794 register_inetaddr_notifier(¨_inetaddr_notifier); 795 796 /* Devices may have been opened already, so the uml_inetaddr_notifier 797 * didn't get a chance to run for them. This fakes it so that 798 * addresses which have already been set up get handled properly. 799 */ 800 spin_lock(&opened_lock); 801 list_for_each(ele, &opened) { 802 lp = list_entry(ele, struct uml_net_private, list); 803 ip = lp->dev->ip_ptr; 804 if (ip == NULL) 805 continue; 806 in = ip->ifa_list; 807 while (in != NULL) { 808 uml_inetaddr_event(NULL, NETDEV_UP, in); 809 in = in->ifa_next; 810 } 811 } 812 spin_unlock(&opened_lock); 813 } 814 #else 815 static inline void inet_register(void) 816 { 817 } 818 #endif 819 820 static int uml_net_init(void) 821 { 822 mconsole_register_dev(&net_mc); 823 inet_register(); 824 return 0; 825 } 826 827 __initcall(uml_net_init); 828 829 static void close_devices(void) 830 { 831 struct list_head *ele; 832 struct uml_net_private *lp; 833 834 spin_lock(&opened_lock); 835 list_for_each(ele, &opened) { 836 lp = list_entry(ele, struct uml_net_private, list); 837 free_irq(lp->dev->irq, lp->dev); 838 if ((lp->close != NULL) && (lp->fd >= 0)) 839 (*lp->close)(lp->fd, &lp->user); 840 if (lp->remove != NULL) 841 (*lp->remove)(&lp->user); 842 } 843 spin_unlock(&opened_lock); 844 } 845 846 __uml_exitcall(close_devices); 847 848 void iter_addresses(void *d, void (*cb)(unsigned char *, unsigned char *, 849 void *), 850 void *arg) 851 { 852 struct net_device *dev = d; 853 struct in_device *ip = dev->ip_ptr; 854 struct in_ifaddr *in; 855 unsigned char address[4], netmask[4]; 856 857 if (ip == NULL) return; 858 in = ip->ifa_list; 859 while (in != NULL) { 860 memcpy(address, &in->ifa_address, sizeof(address)); 861 memcpy(netmask, &in->ifa_mask, sizeof(netmask)); 862 (*cb)(address, netmask, arg); 863 in = in->ifa_next; 864 } 865 } 866 867 int dev_netmask(void *d, void *m) 868 { 869 struct net_device *dev = d; 870 struct in_device *ip = dev->ip_ptr; 871 struct in_ifaddr *in; 872 __be32 *mask_out = m; 873 874 if (ip == NULL) 875 return 1; 876 877 in = ip->ifa_list; 878 if (in == NULL) 879 return 1; 880 881 *mask_out = in->ifa_mask; 882 return 0; 883 } 884 885 void *get_output_buffer(int *len_out) 886 { 887 void *ret; 888 889 ret = (void *) __get_free_pages(GFP_KERNEL, 0); 890 if (ret) *len_out = PAGE_SIZE; 891 else *len_out = 0; 892 return ret; 893 } 894 895 void free_output_buffer(void *buffer) 896 { 897 free_pages((unsigned long) buffer, 0); 898 } 899 900 int tap_setup_common(char *str, char *type, char **dev_name, char **mac_out, 901 char **gate_addr) 902 { 903 char *remain; 904 905 remain = split_if_spec(str, dev_name, mac_out, gate_addr, NULL); 906 if (remain != NULL) { 907 printk(KERN_ERR "tap_setup_common - Extra garbage on " 908 "specification : '%s'\n", remain); 909 return 1; 910 } 911 912 return 0; 913 } 914 915 unsigned short eth_protocol(struct sk_buff *skb) 916 { 917 return eth_type_trans(skb, skb->dev); 918 } 919