1 /* 2 * NET3 IP device support routines. 3 * 4 * This program is free software; you can redistribute it and/or 5 * modify it under the terms of the GNU General Public License 6 * as published by the Free Software Foundation; either version 7 * 2 of the License, or (at your option) any later version. 8 * 9 * Derived from the IP parts of dev.c 1.0.19 10 * Authors: Ross Biro 11 * Fred N. van Kempen, <waltje@uWalt.NL.Mugnet.ORG> 12 * Mark Evans, <evansmp@uhura.aston.ac.uk> 13 * 14 * Additional Authors: 15 * Alan Cox, <gw4pts@gw4pts.ampr.org> 16 * Alexey Kuznetsov, <kuznet@ms2.inr.ac.ru> 17 * 18 * Changes: 19 * Alexey Kuznetsov: pa_* fields are replaced with ifaddr 20 * lists. 21 * Cyrus Durgin: updated for kmod 22 * Matthias Andree: in devinet_ioctl, compare label and 23 * address (4.4BSD alias style support), 24 * fall back to comparing just the label 25 * if no match found. 26 */ 27 28 29 #include <asm/uaccess.h> 30 #include <linux/bitops.h> 31 #include <linux/capability.h> 32 #include <linux/module.h> 33 #include <linux/types.h> 34 #include <linux/kernel.h> 35 #include <linux/string.h> 36 #include <linux/mm.h> 37 #include <linux/socket.h> 38 #include <linux/sockios.h> 39 #include <linux/in.h> 40 #include <linux/errno.h> 41 #include <linux/interrupt.h> 42 #include <linux/if_addr.h> 43 #include <linux/if_ether.h> 44 #include <linux/inet.h> 45 #include <linux/netdevice.h> 46 #include <linux/etherdevice.h> 47 #include <linux/skbuff.h> 48 #include <linux/init.h> 49 #include <linux/notifier.h> 50 #include <linux/inetdevice.h> 51 #include <linux/igmp.h> 52 #include <linux/slab.h> 53 #include <linux/hash.h> 54 #ifdef CONFIG_SYSCTL 55 #include <linux/sysctl.h> 56 #endif 57 #include <linux/kmod.h> 58 #include <linux/netconf.h> 59 60 #include <net/arp.h> 61 #include <net/ip.h> 62 #include <net/route.h> 63 #include <net/ip_fib.h> 64 #include <net/rtnetlink.h> 65 #include <net/net_namespace.h> 66 #include <net/addrconf.h> 67 68 #include "fib_lookup.h" 69 70 static struct ipv4_devconf ipv4_devconf = { 71 .data = { 72 [IPV4_DEVCONF_ACCEPT_REDIRECTS - 1] = 1, 73 [IPV4_DEVCONF_SEND_REDIRECTS - 1] = 1, 74 [IPV4_DEVCONF_SECURE_REDIRECTS - 1] = 1, 75 [IPV4_DEVCONF_SHARED_MEDIA - 1] = 1, 76 [IPV4_DEVCONF_IGMPV2_UNSOLICITED_REPORT_INTERVAL - 1] = 10000 /*ms*/, 77 [IPV4_DEVCONF_IGMPV3_UNSOLICITED_REPORT_INTERVAL - 1] = 1000 /*ms*/, 78 }, 79 }; 80 81 static struct ipv4_devconf ipv4_devconf_dflt = { 82 .data = { 83 [IPV4_DEVCONF_ACCEPT_REDIRECTS - 1] = 1, 84 [IPV4_DEVCONF_SEND_REDIRECTS - 1] = 1, 85 [IPV4_DEVCONF_SECURE_REDIRECTS - 1] = 1, 86 [IPV4_DEVCONF_SHARED_MEDIA - 1] = 1, 87 [IPV4_DEVCONF_ACCEPT_SOURCE_ROUTE - 1] = 1, 88 [IPV4_DEVCONF_IGMPV2_UNSOLICITED_REPORT_INTERVAL - 1] = 10000 /*ms*/, 89 [IPV4_DEVCONF_IGMPV3_UNSOLICITED_REPORT_INTERVAL - 1] = 1000 /*ms*/, 90 }, 91 }; 92 93 #define IPV4_DEVCONF_DFLT(net, attr) \ 94 IPV4_DEVCONF((*net->ipv4.devconf_dflt), attr) 95 96 static const struct nla_policy ifa_ipv4_policy[IFA_MAX+1] = { 97 [IFA_LOCAL] = { .type = NLA_U32 }, 98 [IFA_ADDRESS] = { .type = NLA_U32 }, 99 [IFA_BROADCAST] = { .type = NLA_U32 }, 100 [IFA_LABEL] = { .type = NLA_STRING, .len = IFNAMSIZ - 1 }, 101 [IFA_CACHEINFO] = { .len = sizeof(struct ifa_cacheinfo) }, 102 [IFA_FLAGS] = { .type = NLA_U32 }, 103 }; 104 105 #define IN4_ADDR_HSIZE_SHIFT 8 106 #define IN4_ADDR_HSIZE (1U << IN4_ADDR_HSIZE_SHIFT) 107 108 static struct hlist_head inet_addr_lst[IN4_ADDR_HSIZE]; 109 110 static u32 inet_addr_hash(const struct net *net, __be32 addr) 111 { 112 u32 val = (__force u32) addr ^ net_hash_mix(net); 113 114 return hash_32(val, IN4_ADDR_HSIZE_SHIFT); 115 } 116 117 static void inet_hash_insert(struct net *net, struct in_ifaddr *ifa) 118 { 119 u32 hash = inet_addr_hash(net, ifa->ifa_local); 120 121 ASSERT_RTNL(); 122 hlist_add_head_rcu(&ifa->hash, &inet_addr_lst[hash]); 123 } 124 125 static void inet_hash_remove(struct in_ifaddr *ifa) 126 { 127 ASSERT_RTNL(); 128 hlist_del_init_rcu(&ifa->hash); 129 } 130 131 /** 132 * __ip_dev_find - find the first device with a given source address. 133 * @net: the net namespace 134 * @addr: the source address 135 * @devref: if true, take a reference on the found device 136 * 137 * If a caller uses devref=false, it should be protected by RCU, or RTNL 138 */ 139 struct net_device *__ip_dev_find(struct net *net, __be32 addr, bool devref) 140 { 141 u32 hash = inet_addr_hash(net, addr); 142 struct net_device *result = NULL; 143 struct in_ifaddr *ifa; 144 145 rcu_read_lock(); 146 hlist_for_each_entry_rcu(ifa, &inet_addr_lst[hash], hash) { 147 if (ifa->ifa_local == addr) { 148 struct net_device *dev = ifa->ifa_dev->dev; 149 150 if (!net_eq(dev_net(dev), net)) 151 continue; 152 result = dev; 153 break; 154 } 155 } 156 if (!result) { 157 struct flowi4 fl4 = { .daddr = addr }; 158 struct fib_result res = { 0 }; 159 struct fib_table *local; 160 161 /* Fallback to FIB local table so that communication 162 * over loopback subnets work. 163 */ 164 local = fib_get_table(net, RT_TABLE_LOCAL); 165 if (local && 166 !fib_table_lookup(local, &fl4, &res, FIB_LOOKUP_NOREF) && 167 res.type == RTN_LOCAL) 168 result = FIB_RES_DEV(res); 169 } 170 if (result && devref) 171 dev_hold(result); 172 rcu_read_unlock(); 173 return result; 174 } 175 EXPORT_SYMBOL(__ip_dev_find); 176 177 static void rtmsg_ifa(int event, struct in_ifaddr *, struct nlmsghdr *, u32); 178 179 static BLOCKING_NOTIFIER_HEAD(inetaddr_chain); 180 static void inet_del_ifa(struct in_device *in_dev, struct in_ifaddr **ifap, 181 int destroy); 182 #ifdef CONFIG_SYSCTL 183 static int devinet_sysctl_register(struct in_device *idev); 184 static void devinet_sysctl_unregister(struct in_device *idev); 185 #else 186 static int devinet_sysctl_register(struct in_device *idev) 187 { 188 return 0; 189 } 190 static void devinet_sysctl_unregister(struct in_device *idev) 191 { 192 } 193 #endif 194 195 /* Locks all the inet devices. */ 196 197 static struct in_ifaddr *inet_alloc_ifa(void) 198 { 199 return kzalloc(sizeof(struct in_ifaddr), GFP_KERNEL); 200 } 201 202 static void inet_rcu_free_ifa(struct rcu_head *head) 203 { 204 struct in_ifaddr *ifa = container_of(head, struct in_ifaddr, rcu_head); 205 if (ifa->ifa_dev) 206 in_dev_put(ifa->ifa_dev); 207 kfree(ifa); 208 } 209 210 static void inet_free_ifa(struct in_ifaddr *ifa) 211 { 212 call_rcu(&ifa->rcu_head, inet_rcu_free_ifa); 213 } 214 215 void in_dev_finish_destroy(struct in_device *idev) 216 { 217 struct net_device *dev = idev->dev; 218 219 WARN_ON(idev->ifa_list); 220 WARN_ON(idev->mc_list); 221 kfree(rcu_dereference_protected(idev->mc_hash, 1)); 222 #ifdef NET_REFCNT_DEBUG 223 pr_debug("%s: %p=%s\n", __func__, idev, dev ? dev->name : "NIL"); 224 #endif 225 dev_put(dev); 226 if (!idev->dead) 227 pr_err("Freeing alive in_device %p\n", idev); 228 else 229 kfree(idev); 230 } 231 EXPORT_SYMBOL(in_dev_finish_destroy); 232 233 static struct in_device *inetdev_init(struct net_device *dev) 234 { 235 struct in_device *in_dev; 236 int err = -ENOMEM; 237 238 ASSERT_RTNL(); 239 240 in_dev = kzalloc(sizeof(*in_dev), GFP_KERNEL); 241 if (!in_dev) 242 goto out; 243 memcpy(&in_dev->cnf, dev_net(dev)->ipv4.devconf_dflt, 244 sizeof(in_dev->cnf)); 245 in_dev->cnf.sysctl = NULL; 246 in_dev->dev = dev; 247 in_dev->arp_parms = neigh_parms_alloc(dev, &arp_tbl); 248 if (!in_dev->arp_parms) 249 goto out_kfree; 250 if (IPV4_DEVCONF(in_dev->cnf, FORWARDING)) 251 dev_disable_lro(dev); 252 /* Reference in_dev->dev */ 253 dev_hold(dev); 254 /* Account for reference dev->ip_ptr (below) */ 255 in_dev_hold(in_dev); 256 257 err = devinet_sysctl_register(in_dev); 258 if (err) { 259 in_dev->dead = 1; 260 in_dev_put(in_dev); 261 in_dev = NULL; 262 goto out; 263 } 264 ip_mc_init_dev(in_dev); 265 if (dev->flags & IFF_UP) 266 ip_mc_up(in_dev); 267 268 /* we can receive as soon as ip_ptr is set -- do this last */ 269 rcu_assign_pointer(dev->ip_ptr, in_dev); 270 out: 271 return in_dev ?: ERR_PTR(err); 272 out_kfree: 273 kfree(in_dev); 274 in_dev = NULL; 275 goto out; 276 } 277 278 static void in_dev_rcu_put(struct rcu_head *head) 279 { 280 struct in_device *idev = container_of(head, struct in_device, rcu_head); 281 in_dev_put(idev); 282 } 283 284 static void inetdev_destroy(struct in_device *in_dev) 285 { 286 struct in_ifaddr *ifa; 287 struct net_device *dev; 288 289 ASSERT_RTNL(); 290 291 dev = in_dev->dev; 292 293 in_dev->dead = 1; 294 295 ip_mc_destroy_dev(in_dev); 296 297 while ((ifa = in_dev->ifa_list) != NULL) { 298 inet_del_ifa(in_dev, &in_dev->ifa_list, 0); 299 inet_free_ifa(ifa); 300 } 301 302 RCU_INIT_POINTER(dev->ip_ptr, NULL); 303 304 devinet_sysctl_unregister(in_dev); 305 neigh_parms_release(&arp_tbl, in_dev->arp_parms); 306 arp_ifdown(dev); 307 308 call_rcu(&in_dev->rcu_head, in_dev_rcu_put); 309 } 310 311 int inet_addr_onlink(struct in_device *in_dev, __be32 a, __be32 b) 312 { 313 rcu_read_lock(); 314 for_primary_ifa(in_dev) { 315 if (inet_ifa_match(a, ifa)) { 316 if (!b || inet_ifa_match(b, ifa)) { 317 rcu_read_unlock(); 318 return 1; 319 } 320 } 321 } endfor_ifa(in_dev); 322 rcu_read_unlock(); 323 return 0; 324 } 325 326 static void __inet_del_ifa(struct in_device *in_dev, struct in_ifaddr **ifap, 327 int destroy, struct nlmsghdr *nlh, u32 portid) 328 { 329 struct in_ifaddr *promote = NULL; 330 struct in_ifaddr *ifa, *ifa1 = *ifap; 331 struct in_ifaddr *last_prim = in_dev->ifa_list; 332 struct in_ifaddr *prev_prom = NULL; 333 int do_promote = IN_DEV_PROMOTE_SECONDARIES(in_dev); 334 335 ASSERT_RTNL(); 336 337 /* 1. Deleting primary ifaddr forces deletion all secondaries 338 * unless alias promotion is set 339 **/ 340 341 if (!(ifa1->ifa_flags & IFA_F_SECONDARY)) { 342 struct in_ifaddr **ifap1 = &ifa1->ifa_next; 343 344 while ((ifa = *ifap1) != NULL) { 345 if (!(ifa->ifa_flags & IFA_F_SECONDARY) && 346 ifa1->ifa_scope <= ifa->ifa_scope) 347 last_prim = ifa; 348 349 if (!(ifa->ifa_flags & IFA_F_SECONDARY) || 350 ifa1->ifa_mask != ifa->ifa_mask || 351 !inet_ifa_match(ifa1->ifa_address, ifa)) { 352 ifap1 = &ifa->ifa_next; 353 prev_prom = ifa; 354 continue; 355 } 356 357 if (!do_promote) { 358 inet_hash_remove(ifa); 359 *ifap1 = ifa->ifa_next; 360 361 rtmsg_ifa(RTM_DELADDR, ifa, nlh, portid); 362 blocking_notifier_call_chain(&inetaddr_chain, 363 NETDEV_DOWN, ifa); 364 inet_free_ifa(ifa); 365 } else { 366 promote = ifa; 367 break; 368 } 369 } 370 } 371 372 /* On promotion all secondaries from subnet are changing 373 * the primary IP, we must remove all their routes silently 374 * and later to add them back with new prefsrc. Do this 375 * while all addresses are on the device list. 376 */ 377 for (ifa = promote; ifa; ifa = ifa->ifa_next) { 378 if (ifa1->ifa_mask == ifa->ifa_mask && 379 inet_ifa_match(ifa1->ifa_address, ifa)) 380 fib_del_ifaddr(ifa, ifa1); 381 } 382 383 /* 2. Unlink it */ 384 385 *ifap = ifa1->ifa_next; 386 inet_hash_remove(ifa1); 387 388 /* 3. Announce address deletion */ 389 390 /* Send message first, then call notifier. 391 At first sight, FIB update triggered by notifier 392 will refer to already deleted ifaddr, that could confuse 393 netlink listeners. It is not true: look, gated sees 394 that route deleted and if it still thinks that ifaddr 395 is valid, it will try to restore deleted routes... Grr. 396 So that, this order is correct. 397 */ 398 rtmsg_ifa(RTM_DELADDR, ifa1, nlh, portid); 399 blocking_notifier_call_chain(&inetaddr_chain, NETDEV_DOWN, ifa1); 400 401 if (promote) { 402 struct in_ifaddr *next_sec = promote->ifa_next; 403 404 if (prev_prom) { 405 prev_prom->ifa_next = promote->ifa_next; 406 promote->ifa_next = last_prim->ifa_next; 407 last_prim->ifa_next = promote; 408 } 409 410 promote->ifa_flags &= ~IFA_F_SECONDARY; 411 rtmsg_ifa(RTM_NEWADDR, promote, nlh, portid); 412 blocking_notifier_call_chain(&inetaddr_chain, 413 NETDEV_UP, promote); 414 for (ifa = next_sec; ifa; ifa = ifa->ifa_next) { 415 if (ifa1->ifa_mask != ifa->ifa_mask || 416 !inet_ifa_match(ifa1->ifa_address, ifa)) 417 continue; 418 fib_add_ifaddr(ifa); 419 } 420 421 } 422 if (destroy) 423 inet_free_ifa(ifa1); 424 } 425 426 static void inet_del_ifa(struct in_device *in_dev, struct in_ifaddr **ifap, 427 int destroy) 428 { 429 __inet_del_ifa(in_dev, ifap, destroy, NULL, 0); 430 } 431 432 static void check_lifetime(struct work_struct *work); 433 434 static DECLARE_DELAYED_WORK(check_lifetime_work, check_lifetime); 435 436 static int __inet_insert_ifa(struct in_ifaddr *ifa, struct nlmsghdr *nlh, 437 u32 portid) 438 { 439 struct in_device *in_dev = ifa->ifa_dev; 440 struct in_ifaddr *ifa1, **ifap, **last_primary; 441 442 ASSERT_RTNL(); 443 444 if (!ifa->ifa_local) { 445 inet_free_ifa(ifa); 446 return 0; 447 } 448 449 ifa->ifa_flags &= ~IFA_F_SECONDARY; 450 last_primary = &in_dev->ifa_list; 451 452 for (ifap = &in_dev->ifa_list; (ifa1 = *ifap) != NULL; 453 ifap = &ifa1->ifa_next) { 454 if (!(ifa1->ifa_flags & IFA_F_SECONDARY) && 455 ifa->ifa_scope <= ifa1->ifa_scope) 456 last_primary = &ifa1->ifa_next; 457 if (ifa1->ifa_mask == ifa->ifa_mask && 458 inet_ifa_match(ifa1->ifa_address, ifa)) { 459 if (ifa1->ifa_local == ifa->ifa_local) { 460 inet_free_ifa(ifa); 461 return -EEXIST; 462 } 463 if (ifa1->ifa_scope != ifa->ifa_scope) { 464 inet_free_ifa(ifa); 465 return -EINVAL; 466 } 467 ifa->ifa_flags |= IFA_F_SECONDARY; 468 } 469 } 470 471 if (!(ifa->ifa_flags & IFA_F_SECONDARY)) { 472 prandom_seed((__force u32) ifa->ifa_local); 473 ifap = last_primary; 474 } 475 476 ifa->ifa_next = *ifap; 477 *ifap = ifa; 478 479 inet_hash_insert(dev_net(in_dev->dev), ifa); 480 481 cancel_delayed_work(&check_lifetime_work); 482 queue_delayed_work(system_power_efficient_wq, &check_lifetime_work, 0); 483 484 /* Send message first, then call notifier. 485 Notifier will trigger FIB update, so that 486 listeners of netlink will know about new ifaddr */ 487 rtmsg_ifa(RTM_NEWADDR, ifa, nlh, portid); 488 blocking_notifier_call_chain(&inetaddr_chain, NETDEV_UP, ifa); 489 490 return 0; 491 } 492 493 static int inet_insert_ifa(struct in_ifaddr *ifa) 494 { 495 return __inet_insert_ifa(ifa, NULL, 0); 496 } 497 498 static int inet_set_ifa(struct net_device *dev, struct in_ifaddr *ifa) 499 { 500 struct in_device *in_dev = __in_dev_get_rtnl(dev); 501 502 ASSERT_RTNL(); 503 504 if (!in_dev) { 505 inet_free_ifa(ifa); 506 return -ENOBUFS; 507 } 508 ipv4_devconf_setall(in_dev); 509 neigh_parms_data_state_setall(in_dev->arp_parms); 510 if (ifa->ifa_dev != in_dev) { 511 WARN_ON(ifa->ifa_dev); 512 in_dev_hold(in_dev); 513 ifa->ifa_dev = in_dev; 514 } 515 if (ipv4_is_loopback(ifa->ifa_local)) 516 ifa->ifa_scope = RT_SCOPE_HOST; 517 return inet_insert_ifa(ifa); 518 } 519 520 /* Caller must hold RCU or RTNL : 521 * We dont take a reference on found in_device 522 */ 523 struct in_device *inetdev_by_index(struct net *net, int ifindex) 524 { 525 struct net_device *dev; 526 struct in_device *in_dev = NULL; 527 528 rcu_read_lock(); 529 dev = dev_get_by_index_rcu(net, ifindex); 530 if (dev) 531 in_dev = rcu_dereference_rtnl(dev->ip_ptr); 532 rcu_read_unlock(); 533 return in_dev; 534 } 535 EXPORT_SYMBOL(inetdev_by_index); 536 537 /* Called only from RTNL semaphored context. No locks. */ 538 539 struct in_ifaddr *inet_ifa_byprefix(struct in_device *in_dev, __be32 prefix, 540 __be32 mask) 541 { 542 ASSERT_RTNL(); 543 544 for_primary_ifa(in_dev) { 545 if (ifa->ifa_mask == mask && inet_ifa_match(prefix, ifa)) 546 return ifa; 547 } endfor_ifa(in_dev); 548 return NULL; 549 } 550 551 static int ip_mc_config(struct sock *sk, bool join, const struct in_ifaddr *ifa) 552 { 553 struct ip_mreqn mreq = { 554 .imr_multiaddr.s_addr = ifa->ifa_address, 555 .imr_ifindex = ifa->ifa_dev->dev->ifindex, 556 }; 557 int ret; 558 559 ASSERT_RTNL(); 560 561 lock_sock(sk); 562 if (join) 563 ret = ip_mc_join_group(sk, &mreq); 564 else 565 ret = ip_mc_leave_group(sk, &mreq); 566 release_sock(sk); 567 568 return ret; 569 } 570 571 static int inet_rtm_deladdr(struct sk_buff *skb, struct nlmsghdr *nlh) 572 { 573 struct net *net = sock_net(skb->sk); 574 struct nlattr *tb[IFA_MAX+1]; 575 struct in_device *in_dev; 576 struct ifaddrmsg *ifm; 577 struct in_ifaddr *ifa, **ifap; 578 int err = -EINVAL; 579 580 ASSERT_RTNL(); 581 582 err = nlmsg_parse(nlh, sizeof(*ifm), tb, IFA_MAX, ifa_ipv4_policy); 583 if (err < 0) 584 goto errout; 585 586 ifm = nlmsg_data(nlh); 587 in_dev = inetdev_by_index(net, ifm->ifa_index); 588 if (!in_dev) { 589 err = -ENODEV; 590 goto errout; 591 } 592 593 for (ifap = &in_dev->ifa_list; (ifa = *ifap) != NULL; 594 ifap = &ifa->ifa_next) { 595 if (tb[IFA_LOCAL] && 596 ifa->ifa_local != nla_get_in_addr(tb[IFA_LOCAL])) 597 continue; 598 599 if (tb[IFA_LABEL] && nla_strcmp(tb[IFA_LABEL], ifa->ifa_label)) 600 continue; 601 602 if (tb[IFA_ADDRESS] && 603 (ifm->ifa_prefixlen != ifa->ifa_prefixlen || 604 !inet_ifa_match(nla_get_in_addr(tb[IFA_ADDRESS]), ifa))) 605 continue; 606 607 if (ipv4_is_multicast(ifa->ifa_address)) 608 ip_mc_config(net->ipv4.mc_autojoin_sk, false, ifa); 609 __inet_del_ifa(in_dev, ifap, 1, nlh, NETLINK_CB(skb).portid); 610 return 0; 611 } 612 613 err = -EADDRNOTAVAIL; 614 errout: 615 return err; 616 } 617 618 #define INFINITY_LIFE_TIME 0xFFFFFFFF 619 620 static void check_lifetime(struct work_struct *work) 621 { 622 unsigned long now, next, next_sec, next_sched; 623 struct in_ifaddr *ifa; 624 struct hlist_node *n; 625 int i; 626 627 now = jiffies; 628 next = round_jiffies_up(now + ADDR_CHECK_FREQUENCY); 629 630 for (i = 0; i < IN4_ADDR_HSIZE; i++) { 631 bool change_needed = false; 632 633 rcu_read_lock(); 634 hlist_for_each_entry_rcu(ifa, &inet_addr_lst[i], hash) { 635 unsigned long age; 636 637 if (ifa->ifa_flags & IFA_F_PERMANENT) 638 continue; 639 640 /* We try to batch several events at once. */ 641 age = (now - ifa->ifa_tstamp + 642 ADDRCONF_TIMER_FUZZ_MINUS) / HZ; 643 644 if (ifa->ifa_valid_lft != INFINITY_LIFE_TIME && 645 age >= ifa->ifa_valid_lft) { 646 change_needed = true; 647 } else if (ifa->ifa_preferred_lft == 648 INFINITY_LIFE_TIME) { 649 continue; 650 } else if (age >= ifa->ifa_preferred_lft) { 651 if (time_before(ifa->ifa_tstamp + 652 ifa->ifa_valid_lft * HZ, next)) 653 next = ifa->ifa_tstamp + 654 ifa->ifa_valid_lft * HZ; 655 656 if (!(ifa->ifa_flags & IFA_F_DEPRECATED)) 657 change_needed = true; 658 } else if (time_before(ifa->ifa_tstamp + 659 ifa->ifa_preferred_lft * HZ, 660 next)) { 661 next = ifa->ifa_tstamp + 662 ifa->ifa_preferred_lft * HZ; 663 } 664 } 665 rcu_read_unlock(); 666 if (!change_needed) 667 continue; 668 rtnl_lock(); 669 hlist_for_each_entry_safe(ifa, n, &inet_addr_lst[i], hash) { 670 unsigned long age; 671 672 if (ifa->ifa_flags & IFA_F_PERMANENT) 673 continue; 674 675 /* We try to batch several events at once. */ 676 age = (now - ifa->ifa_tstamp + 677 ADDRCONF_TIMER_FUZZ_MINUS) / HZ; 678 679 if (ifa->ifa_valid_lft != INFINITY_LIFE_TIME && 680 age >= ifa->ifa_valid_lft) { 681 struct in_ifaddr **ifap; 682 683 for (ifap = &ifa->ifa_dev->ifa_list; 684 *ifap != NULL; ifap = &(*ifap)->ifa_next) { 685 if (*ifap == ifa) { 686 inet_del_ifa(ifa->ifa_dev, 687 ifap, 1); 688 break; 689 } 690 } 691 } else if (ifa->ifa_preferred_lft != 692 INFINITY_LIFE_TIME && 693 age >= ifa->ifa_preferred_lft && 694 !(ifa->ifa_flags & IFA_F_DEPRECATED)) { 695 ifa->ifa_flags |= IFA_F_DEPRECATED; 696 rtmsg_ifa(RTM_NEWADDR, ifa, NULL, 0); 697 } 698 } 699 rtnl_unlock(); 700 } 701 702 next_sec = round_jiffies_up(next); 703 next_sched = next; 704 705 /* If rounded timeout is accurate enough, accept it. */ 706 if (time_before(next_sec, next + ADDRCONF_TIMER_FUZZ)) 707 next_sched = next_sec; 708 709 now = jiffies; 710 /* And minimum interval is ADDRCONF_TIMER_FUZZ_MAX. */ 711 if (time_before(next_sched, now + ADDRCONF_TIMER_FUZZ_MAX)) 712 next_sched = now + ADDRCONF_TIMER_FUZZ_MAX; 713 714 queue_delayed_work(system_power_efficient_wq, &check_lifetime_work, 715 next_sched - now); 716 } 717 718 static void set_ifa_lifetime(struct in_ifaddr *ifa, __u32 valid_lft, 719 __u32 prefered_lft) 720 { 721 unsigned long timeout; 722 723 ifa->ifa_flags &= ~(IFA_F_PERMANENT | IFA_F_DEPRECATED); 724 725 timeout = addrconf_timeout_fixup(valid_lft, HZ); 726 if (addrconf_finite_timeout(timeout)) 727 ifa->ifa_valid_lft = timeout; 728 else 729 ifa->ifa_flags |= IFA_F_PERMANENT; 730 731 timeout = addrconf_timeout_fixup(prefered_lft, HZ); 732 if (addrconf_finite_timeout(timeout)) { 733 if (timeout == 0) 734 ifa->ifa_flags |= IFA_F_DEPRECATED; 735 ifa->ifa_preferred_lft = timeout; 736 } 737 ifa->ifa_tstamp = jiffies; 738 if (!ifa->ifa_cstamp) 739 ifa->ifa_cstamp = ifa->ifa_tstamp; 740 } 741 742 static struct in_ifaddr *rtm_to_ifaddr(struct net *net, struct nlmsghdr *nlh, 743 __u32 *pvalid_lft, __u32 *pprefered_lft) 744 { 745 struct nlattr *tb[IFA_MAX+1]; 746 struct in_ifaddr *ifa; 747 struct ifaddrmsg *ifm; 748 struct net_device *dev; 749 struct in_device *in_dev; 750 int err; 751 752 err = nlmsg_parse(nlh, sizeof(*ifm), tb, IFA_MAX, ifa_ipv4_policy); 753 if (err < 0) 754 goto errout; 755 756 ifm = nlmsg_data(nlh); 757 err = -EINVAL; 758 if (ifm->ifa_prefixlen > 32 || !tb[IFA_LOCAL]) 759 goto errout; 760 761 dev = __dev_get_by_index(net, ifm->ifa_index); 762 err = -ENODEV; 763 if (!dev) 764 goto errout; 765 766 in_dev = __in_dev_get_rtnl(dev); 767 err = -ENOBUFS; 768 if (!in_dev) 769 goto errout; 770 771 ifa = inet_alloc_ifa(); 772 if (!ifa) 773 /* 774 * A potential indev allocation can be left alive, it stays 775 * assigned to its device and is destroy with it. 776 */ 777 goto errout; 778 779 ipv4_devconf_setall(in_dev); 780 neigh_parms_data_state_setall(in_dev->arp_parms); 781 in_dev_hold(in_dev); 782 783 if (!tb[IFA_ADDRESS]) 784 tb[IFA_ADDRESS] = tb[IFA_LOCAL]; 785 786 INIT_HLIST_NODE(&ifa->hash); 787 ifa->ifa_prefixlen = ifm->ifa_prefixlen; 788 ifa->ifa_mask = inet_make_mask(ifm->ifa_prefixlen); 789 ifa->ifa_flags = tb[IFA_FLAGS] ? nla_get_u32(tb[IFA_FLAGS]) : 790 ifm->ifa_flags; 791 ifa->ifa_scope = ifm->ifa_scope; 792 ifa->ifa_dev = in_dev; 793 794 ifa->ifa_local = nla_get_in_addr(tb[IFA_LOCAL]); 795 ifa->ifa_address = nla_get_in_addr(tb[IFA_ADDRESS]); 796 797 if (tb[IFA_BROADCAST]) 798 ifa->ifa_broadcast = nla_get_in_addr(tb[IFA_BROADCAST]); 799 800 if (tb[IFA_LABEL]) 801 nla_strlcpy(ifa->ifa_label, tb[IFA_LABEL], IFNAMSIZ); 802 else 803 memcpy(ifa->ifa_label, dev->name, IFNAMSIZ); 804 805 if (tb[IFA_CACHEINFO]) { 806 struct ifa_cacheinfo *ci; 807 808 ci = nla_data(tb[IFA_CACHEINFO]); 809 if (!ci->ifa_valid || ci->ifa_prefered > ci->ifa_valid) { 810 err = -EINVAL; 811 goto errout_free; 812 } 813 *pvalid_lft = ci->ifa_valid; 814 *pprefered_lft = ci->ifa_prefered; 815 } 816 817 return ifa; 818 819 errout_free: 820 inet_free_ifa(ifa); 821 errout: 822 return ERR_PTR(err); 823 } 824 825 static struct in_ifaddr *find_matching_ifa(struct in_ifaddr *ifa) 826 { 827 struct in_device *in_dev = ifa->ifa_dev; 828 struct in_ifaddr *ifa1, **ifap; 829 830 if (!ifa->ifa_local) 831 return NULL; 832 833 for (ifap = &in_dev->ifa_list; (ifa1 = *ifap) != NULL; 834 ifap = &ifa1->ifa_next) { 835 if (ifa1->ifa_mask == ifa->ifa_mask && 836 inet_ifa_match(ifa1->ifa_address, ifa) && 837 ifa1->ifa_local == ifa->ifa_local) 838 return ifa1; 839 } 840 return NULL; 841 } 842 843 static int inet_rtm_newaddr(struct sk_buff *skb, struct nlmsghdr *nlh) 844 { 845 struct net *net = sock_net(skb->sk); 846 struct in_ifaddr *ifa; 847 struct in_ifaddr *ifa_existing; 848 __u32 valid_lft = INFINITY_LIFE_TIME; 849 __u32 prefered_lft = INFINITY_LIFE_TIME; 850 851 ASSERT_RTNL(); 852 853 ifa = rtm_to_ifaddr(net, nlh, &valid_lft, &prefered_lft); 854 if (IS_ERR(ifa)) 855 return PTR_ERR(ifa); 856 857 ifa_existing = find_matching_ifa(ifa); 858 if (!ifa_existing) { 859 /* It would be best to check for !NLM_F_CREATE here but 860 * userspace already relies on not having to provide this. 861 */ 862 set_ifa_lifetime(ifa, valid_lft, prefered_lft); 863 if (ifa->ifa_flags & IFA_F_MCAUTOJOIN) { 864 int ret = ip_mc_config(net->ipv4.mc_autojoin_sk, 865 true, ifa); 866 867 if (ret < 0) { 868 inet_free_ifa(ifa); 869 return ret; 870 } 871 } 872 return __inet_insert_ifa(ifa, nlh, NETLINK_CB(skb).portid); 873 } else { 874 inet_free_ifa(ifa); 875 876 if (nlh->nlmsg_flags & NLM_F_EXCL || 877 !(nlh->nlmsg_flags & NLM_F_REPLACE)) 878 return -EEXIST; 879 ifa = ifa_existing; 880 set_ifa_lifetime(ifa, valid_lft, prefered_lft); 881 cancel_delayed_work(&check_lifetime_work); 882 queue_delayed_work(system_power_efficient_wq, 883 &check_lifetime_work, 0); 884 rtmsg_ifa(RTM_NEWADDR, ifa, nlh, NETLINK_CB(skb).portid); 885 blocking_notifier_call_chain(&inetaddr_chain, NETDEV_UP, ifa); 886 } 887 return 0; 888 } 889 890 /* 891 * Determine a default network mask, based on the IP address. 892 */ 893 894 static int inet_abc_len(__be32 addr) 895 { 896 int rc = -1; /* Something else, probably a multicast. */ 897 898 if (ipv4_is_zeronet(addr)) 899 rc = 0; 900 else { 901 __u32 haddr = ntohl(addr); 902 903 if (IN_CLASSA(haddr)) 904 rc = 8; 905 else if (IN_CLASSB(haddr)) 906 rc = 16; 907 else if (IN_CLASSC(haddr)) 908 rc = 24; 909 } 910 911 return rc; 912 } 913 914 915 int devinet_ioctl(struct net *net, unsigned int cmd, void __user *arg) 916 { 917 struct ifreq ifr; 918 struct sockaddr_in sin_orig; 919 struct sockaddr_in *sin = (struct sockaddr_in *)&ifr.ifr_addr; 920 struct in_device *in_dev; 921 struct in_ifaddr **ifap = NULL; 922 struct in_ifaddr *ifa = NULL; 923 struct net_device *dev; 924 char *colon; 925 int ret = -EFAULT; 926 int tryaddrmatch = 0; 927 928 /* 929 * Fetch the caller's info block into kernel space 930 */ 931 932 if (copy_from_user(&ifr, arg, sizeof(struct ifreq))) 933 goto out; 934 ifr.ifr_name[IFNAMSIZ - 1] = 0; 935 936 /* save original address for comparison */ 937 memcpy(&sin_orig, sin, sizeof(*sin)); 938 939 colon = strchr(ifr.ifr_name, ':'); 940 if (colon) 941 *colon = 0; 942 943 dev_load(net, ifr.ifr_name); 944 945 switch (cmd) { 946 case SIOCGIFADDR: /* Get interface address */ 947 case SIOCGIFBRDADDR: /* Get the broadcast address */ 948 case SIOCGIFDSTADDR: /* Get the destination address */ 949 case SIOCGIFNETMASK: /* Get the netmask for the interface */ 950 /* Note that these ioctls will not sleep, 951 so that we do not impose a lock. 952 One day we will be forced to put shlock here (I mean SMP) 953 */ 954 tryaddrmatch = (sin_orig.sin_family == AF_INET); 955 memset(sin, 0, sizeof(*sin)); 956 sin->sin_family = AF_INET; 957 break; 958 959 case SIOCSIFFLAGS: 960 ret = -EPERM; 961 if (!ns_capable(net->user_ns, CAP_NET_ADMIN)) 962 goto out; 963 break; 964 case SIOCSIFADDR: /* Set interface address (and family) */ 965 case SIOCSIFBRDADDR: /* Set the broadcast address */ 966 case SIOCSIFDSTADDR: /* Set the destination address */ 967 case SIOCSIFNETMASK: /* Set the netmask for the interface */ 968 ret = -EPERM; 969 if (!ns_capable(net->user_ns, CAP_NET_ADMIN)) 970 goto out; 971 ret = -EINVAL; 972 if (sin->sin_family != AF_INET) 973 goto out; 974 break; 975 default: 976 ret = -EINVAL; 977 goto out; 978 } 979 980 rtnl_lock(); 981 982 ret = -ENODEV; 983 dev = __dev_get_by_name(net, ifr.ifr_name); 984 if (!dev) 985 goto done; 986 987 if (colon) 988 *colon = ':'; 989 990 in_dev = __in_dev_get_rtnl(dev); 991 if (in_dev) { 992 if (tryaddrmatch) { 993 /* Matthias Andree */ 994 /* compare label and address (4.4BSD style) */ 995 /* note: we only do this for a limited set of ioctls 996 and only if the original address family was AF_INET. 997 This is checked above. */ 998 for (ifap = &in_dev->ifa_list; (ifa = *ifap) != NULL; 999 ifap = &ifa->ifa_next) { 1000 if (!strcmp(ifr.ifr_name, ifa->ifa_label) && 1001 sin_orig.sin_addr.s_addr == 1002 ifa->ifa_local) { 1003 break; /* found */ 1004 } 1005 } 1006 } 1007 /* we didn't get a match, maybe the application is 1008 4.3BSD-style and passed in junk so we fall back to 1009 comparing just the label */ 1010 if (!ifa) { 1011 for (ifap = &in_dev->ifa_list; (ifa = *ifap) != NULL; 1012 ifap = &ifa->ifa_next) 1013 if (!strcmp(ifr.ifr_name, ifa->ifa_label)) 1014 break; 1015 } 1016 } 1017 1018 ret = -EADDRNOTAVAIL; 1019 if (!ifa && cmd != SIOCSIFADDR && cmd != SIOCSIFFLAGS) 1020 goto done; 1021 1022 switch (cmd) { 1023 case SIOCGIFADDR: /* Get interface address */ 1024 sin->sin_addr.s_addr = ifa->ifa_local; 1025 goto rarok; 1026 1027 case SIOCGIFBRDADDR: /* Get the broadcast address */ 1028 sin->sin_addr.s_addr = ifa->ifa_broadcast; 1029 goto rarok; 1030 1031 case SIOCGIFDSTADDR: /* Get the destination address */ 1032 sin->sin_addr.s_addr = ifa->ifa_address; 1033 goto rarok; 1034 1035 case SIOCGIFNETMASK: /* Get the netmask for the interface */ 1036 sin->sin_addr.s_addr = ifa->ifa_mask; 1037 goto rarok; 1038 1039 case SIOCSIFFLAGS: 1040 if (colon) { 1041 ret = -EADDRNOTAVAIL; 1042 if (!ifa) 1043 break; 1044 ret = 0; 1045 if (!(ifr.ifr_flags & IFF_UP)) 1046 inet_del_ifa(in_dev, ifap, 1); 1047 break; 1048 } 1049 ret = dev_change_flags(dev, ifr.ifr_flags); 1050 break; 1051 1052 case SIOCSIFADDR: /* Set interface address (and family) */ 1053 ret = -EINVAL; 1054 if (inet_abc_len(sin->sin_addr.s_addr) < 0) 1055 break; 1056 1057 if (!ifa) { 1058 ret = -ENOBUFS; 1059 ifa = inet_alloc_ifa(); 1060 if (!ifa) 1061 break; 1062 INIT_HLIST_NODE(&ifa->hash); 1063 if (colon) 1064 memcpy(ifa->ifa_label, ifr.ifr_name, IFNAMSIZ); 1065 else 1066 memcpy(ifa->ifa_label, dev->name, IFNAMSIZ); 1067 } else { 1068 ret = 0; 1069 if (ifa->ifa_local == sin->sin_addr.s_addr) 1070 break; 1071 inet_del_ifa(in_dev, ifap, 0); 1072 ifa->ifa_broadcast = 0; 1073 ifa->ifa_scope = 0; 1074 } 1075 1076 ifa->ifa_address = ifa->ifa_local = sin->sin_addr.s_addr; 1077 1078 if (!(dev->flags & IFF_POINTOPOINT)) { 1079 ifa->ifa_prefixlen = inet_abc_len(ifa->ifa_address); 1080 ifa->ifa_mask = inet_make_mask(ifa->ifa_prefixlen); 1081 if ((dev->flags & IFF_BROADCAST) && 1082 ifa->ifa_prefixlen < 31) 1083 ifa->ifa_broadcast = ifa->ifa_address | 1084 ~ifa->ifa_mask; 1085 } else { 1086 ifa->ifa_prefixlen = 32; 1087 ifa->ifa_mask = inet_make_mask(32); 1088 } 1089 set_ifa_lifetime(ifa, INFINITY_LIFE_TIME, INFINITY_LIFE_TIME); 1090 ret = inet_set_ifa(dev, ifa); 1091 break; 1092 1093 case SIOCSIFBRDADDR: /* Set the broadcast address */ 1094 ret = 0; 1095 if (ifa->ifa_broadcast != sin->sin_addr.s_addr) { 1096 inet_del_ifa(in_dev, ifap, 0); 1097 ifa->ifa_broadcast = sin->sin_addr.s_addr; 1098 inet_insert_ifa(ifa); 1099 } 1100 break; 1101 1102 case SIOCSIFDSTADDR: /* Set the destination address */ 1103 ret = 0; 1104 if (ifa->ifa_address == sin->sin_addr.s_addr) 1105 break; 1106 ret = -EINVAL; 1107 if (inet_abc_len(sin->sin_addr.s_addr) < 0) 1108 break; 1109 ret = 0; 1110 inet_del_ifa(in_dev, ifap, 0); 1111 ifa->ifa_address = sin->sin_addr.s_addr; 1112 inet_insert_ifa(ifa); 1113 break; 1114 1115 case SIOCSIFNETMASK: /* Set the netmask for the interface */ 1116 1117 /* 1118 * The mask we set must be legal. 1119 */ 1120 ret = -EINVAL; 1121 if (bad_mask(sin->sin_addr.s_addr, 0)) 1122 break; 1123 ret = 0; 1124 if (ifa->ifa_mask != sin->sin_addr.s_addr) { 1125 __be32 old_mask = ifa->ifa_mask; 1126 inet_del_ifa(in_dev, ifap, 0); 1127 ifa->ifa_mask = sin->sin_addr.s_addr; 1128 ifa->ifa_prefixlen = inet_mask_len(ifa->ifa_mask); 1129 1130 /* See if current broadcast address matches 1131 * with current netmask, then recalculate 1132 * the broadcast address. Otherwise it's a 1133 * funny address, so don't touch it since 1134 * the user seems to know what (s)he's doing... 1135 */ 1136 if ((dev->flags & IFF_BROADCAST) && 1137 (ifa->ifa_prefixlen < 31) && 1138 (ifa->ifa_broadcast == 1139 (ifa->ifa_local|~old_mask))) { 1140 ifa->ifa_broadcast = (ifa->ifa_local | 1141 ~sin->sin_addr.s_addr); 1142 } 1143 inet_insert_ifa(ifa); 1144 } 1145 break; 1146 } 1147 done: 1148 rtnl_unlock(); 1149 out: 1150 return ret; 1151 rarok: 1152 rtnl_unlock(); 1153 ret = copy_to_user(arg, &ifr, sizeof(struct ifreq)) ? -EFAULT : 0; 1154 goto out; 1155 } 1156 1157 static int inet_gifconf(struct net_device *dev, char __user *buf, int len) 1158 { 1159 struct in_device *in_dev = __in_dev_get_rtnl(dev); 1160 struct in_ifaddr *ifa; 1161 struct ifreq ifr; 1162 int done = 0; 1163 1164 if (!in_dev) 1165 goto out; 1166 1167 for (ifa = in_dev->ifa_list; ifa; ifa = ifa->ifa_next) { 1168 if (!buf) { 1169 done += sizeof(ifr); 1170 continue; 1171 } 1172 if (len < (int) sizeof(ifr)) 1173 break; 1174 memset(&ifr, 0, sizeof(struct ifreq)); 1175 strcpy(ifr.ifr_name, ifa->ifa_label); 1176 1177 (*(struct sockaddr_in *)&ifr.ifr_addr).sin_family = AF_INET; 1178 (*(struct sockaddr_in *)&ifr.ifr_addr).sin_addr.s_addr = 1179 ifa->ifa_local; 1180 1181 if (copy_to_user(buf, &ifr, sizeof(struct ifreq))) { 1182 done = -EFAULT; 1183 break; 1184 } 1185 buf += sizeof(struct ifreq); 1186 len -= sizeof(struct ifreq); 1187 done += sizeof(struct ifreq); 1188 } 1189 out: 1190 return done; 1191 } 1192 1193 __be32 inet_select_addr(const struct net_device *dev, __be32 dst, int scope) 1194 { 1195 __be32 addr = 0; 1196 struct in_device *in_dev; 1197 struct net *net = dev_net(dev); 1198 1199 rcu_read_lock(); 1200 in_dev = __in_dev_get_rcu(dev); 1201 if (!in_dev) 1202 goto no_in_dev; 1203 1204 for_primary_ifa(in_dev) { 1205 if (ifa->ifa_scope > scope) 1206 continue; 1207 if (!dst || inet_ifa_match(dst, ifa)) { 1208 addr = ifa->ifa_local; 1209 break; 1210 } 1211 if (!addr) 1212 addr = ifa->ifa_local; 1213 } endfor_ifa(in_dev); 1214 1215 if (addr) 1216 goto out_unlock; 1217 no_in_dev: 1218 1219 /* Not loopback addresses on loopback should be preferred 1220 in this case. It is important that lo is the first interface 1221 in dev_base list. 1222 */ 1223 for_each_netdev_rcu(net, dev) { 1224 in_dev = __in_dev_get_rcu(dev); 1225 if (!in_dev) 1226 continue; 1227 1228 for_primary_ifa(in_dev) { 1229 if (ifa->ifa_scope != RT_SCOPE_LINK && 1230 ifa->ifa_scope <= scope) { 1231 addr = ifa->ifa_local; 1232 goto out_unlock; 1233 } 1234 } endfor_ifa(in_dev); 1235 } 1236 out_unlock: 1237 rcu_read_unlock(); 1238 return addr; 1239 } 1240 EXPORT_SYMBOL(inet_select_addr); 1241 1242 static __be32 confirm_addr_indev(struct in_device *in_dev, __be32 dst, 1243 __be32 local, int scope) 1244 { 1245 int same = 0; 1246 __be32 addr = 0; 1247 1248 for_ifa(in_dev) { 1249 if (!addr && 1250 (local == ifa->ifa_local || !local) && 1251 ifa->ifa_scope <= scope) { 1252 addr = ifa->ifa_local; 1253 if (same) 1254 break; 1255 } 1256 if (!same) { 1257 same = (!local || inet_ifa_match(local, ifa)) && 1258 (!dst || inet_ifa_match(dst, ifa)); 1259 if (same && addr) { 1260 if (local || !dst) 1261 break; 1262 /* Is the selected addr into dst subnet? */ 1263 if (inet_ifa_match(addr, ifa)) 1264 break; 1265 /* No, then can we use new local src? */ 1266 if (ifa->ifa_scope <= scope) { 1267 addr = ifa->ifa_local; 1268 break; 1269 } 1270 /* search for large dst subnet for addr */ 1271 same = 0; 1272 } 1273 } 1274 } endfor_ifa(in_dev); 1275 1276 return same ? addr : 0; 1277 } 1278 1279 /* 1280 * Confirm that local IP address exists using wildcards: 1281 * - net: netns to check, cannot be NULL 1282 * - in_dev: only on this interface, NULL=any interface 1283 * - dst: only in the same subnet as dst, 0=any dst 1284 * - local: address, 0=autoselect the local address 1285 * - scope: maximum allowed scope value for the local address 1286 */ 1287 __be32 inet_confirm_addr(struct net *net, struct in_device *in_dev, 1288 __be32 dst, __be32 local, int scope) 1289 { 1290 __be32 addr = 0; 1291 struct net_device *dev; 1292 1293 if (in_dev) 1294 return confirm_addr_indev(in_dev, dst, local, scope); 1295 1296 rcu_read_lock(); 1297 for_each_netdev_rcu(net, dev) { 1298 in_dev = __in_dev_get_rcu(dev); 1299 if (in_dev) { 1300 addr = confirm_addr_indev(in_dev, dst, local, scope); 1301 if (addr) 1302 break; 1303 } 1304 } 1305 rcu_read_unlock(); 1306 1307 return addr; 1308 } 1309 EXPORT_SYMBOL(inet_confirm_addr); 1310 1311 /* 1312 * Device notifier 1313 */ 1314 1315 int register_inetaddr_notifier(struct notifier_block *nb) 1316 { 1317 return blocking_notifier_chain_register(&inetaddr_chain, nb); 1318 } 1319 EXPORT_SYMBOL(register_inetaddr_notifier); 1320 1321 int unregister_inetaddr_notifier(struct notifier_block *nb) 1322 { 1323 return blocking_notifier_chain_unregister(&inetaddr_chain, nb); 1324 } 1325 EXPORT_SYMBOL(unregister_inetaddr_notifier); 1326 1327 /* Rename ifa_labels for a device name change. Make some effort to preserve 1328 * existing alias numbering and to create unique labels if possible. 1329 */ 1330 static void inetdev_changename(struct net_device *dev, struct in_device *in_dev) 1331 { 1332 struct in_ifaddr *ifa; 1333 int named = 0; 1334 1335 for (ifa = in_dev->ifa_list; ifa; ifa = ifa->ifa_next) { 1336 char old[IFNAMSIZ], *dot; 1337 1338 memcpy(old, ifa->ifa_label, IFNAMSIZ); 1339 memcpy(ifa->ifa_label, dev->name, IFNAMSIZ); 1340 if (named++ == 0) 1341 goto skip; 1342 dot = strchr(old, ':'); 1343 if (!dot) { 1344 sprintf(old, ":%d", named); 1345 dot = old; 1346 } 1347 if (strlen(dot) + strlen(dev->name) < IFNAMSIZ) 1348 strcat(ifa->ifa_label, dot); 1349 else 1350 strcpy(ifa->ifa_label + (IFNAMSIZ - strlen(dot) - 1), dot); 1351 skip: 1352 rtmsg_ifa(RTM_NEWADDR, ifa, NULL, 0); 1353 } 1354 } 1355 1356 static bool inetdev_valid_mtu(unsigned int mtu) 1357 { 1358 return mtu >= 68; 1359 } 1360 1361 static void inetdev_send_gratuitous_arp(struct net_device *dev, 1362 struct in_device *in_dev) 1363 1364 { 1365 struct in_ifaddr *ifa; 1366 1367 for (ifa = in_dev->ifa_list; ifa; 1368 ifa = ifa->ifa_next) { 1369 arp_send(ARPOP_REQUEST, ETH_P_ARP, 1370 ifa->ifa_local, dev, 1371 ifa->ifa_local, NULL, 1372 dev->dev_addr, NULL); 1373 } 1374 } 1375 1376 /* Called only under RTNL semaphore */ 1377 1378 static int inetdev_event(struct notifier_block *this, unsigned long event, 1379 void *ptr) 1380 { 1381 struct net_device *dev = netdev_notifier_info_to_dev(ptr); 1382 struct in_device *in_dev = __in_dev_get_rtnl(dev); 1383 1384 ASSERT_RTNL(); 1385 1386 if (!in_dev) { 1387 if (event == NETDEV_REGISTER) { 1388 in_dev = inetdev_init(dev); 1389 if (IS_ERR(in_dev)) 1390 return notifier_from_errno(PTR_ERR(in_dev)); 1391 if (dev->flags & IFF_LOOPBACK) { 1392 IN_DEV_CONF_SET(in_dev, NOXFRM, 1); 1393 IN_DEV_CONF_SET(in_dev, NOPOLICY, 1); 1394 } 1395 } else if (event == NETDEV_CHANGEMTU) { 1396 /* Re-enabling IP */ 1397 if (inetdev_valid_mtu(dev->mtu)) 1398 in_dev = inetdev_init(dev); 1399 } 1400 goto out; 1401 } 1402 1403 switch (event) { 1404 case NETDEV_REGISTER: 1405 pr_debug("%s: bug\n", __func__); 1406 RCU_INIT_POINTER(dev->ip_ptr, NULL); 1407 break; 1408 case NETDEV_UP: 1409 if (!inetdev_valid_mtu(dev->mtu)) 1410 break; 1411 if (dev->flags & IFF_LOOPBACK) { 1412 struct in_ifaddr *ifa = inet_alloc_ifa(); 1413 1414 if (ifa) { 1415 INIT_HLIST_NODE(&ifa->hash); 1416 ifa->ifa_local = 1417 ifa->ifa_address = htonl(INADDR_LOOPBACK); 1418 ifa->ifa_prefixlen = 8; 1419 ifa->ifa_mask = inet_make_mask(8); 1420 in_dev_hold(in_dev); 1421 ifa->ifa_dev = in_dev; 1422 ifa->ifa_scope = RT_SCOPE_HOST; 1423 memcpy(ifa->ifa_label, dev->name, IFNAMSIZ); 1424 set_ifa_lifetime(ifa, INFINITY_LIFE_TIME, 1425 INFINITY_LIFE_TIME); 1426 ipv4_devconf_setall(in_dev); 1427 neigh_parms_data_state_setall(in_dev->arp_parms); 1428 inet_insert_ifa(ifa); 1429 } 1430 } 1431 ip_mc_up(in_dev); 1432 /* fall through */ 1433 case NETDEV_CHANGEADDR: 1434 if (!IN_DEV_ARP_NOTIFY(in_dev)) 1435 break; 1436 /* fall through */ 1437 case NETDEV_NOTIFY_PEERS: 1438 /* Send gratuitous ARP to notify of link change */ 1439 inetdev_send_gratuitous_arp(dev, in_dev); 1440 break; 1441 case NETDEV_DOWN: 1442 ip_mc_down(in_dev); 1443 break; 1444 case NETDEV_PRE_TYPE_CHANGE: 1445 ip_mc_unmap(in_dev); 1446 break; 1447 case NETDEV_POST_TYPE_CHANGE: 1448 ip_mc_remap(in_dev); 1449 break; 1450 case NETDEV_CHANGEMTU: 1451 if (inetdev_valid_mtu(dev->mtu)) 1452 break; 1453 /* disable IP when MTU is not enough */ 1454 case NETDEV_UNREGISTER: 1455 inetdev_destroy(in_dev); 1456 break; 1457 case NETDEV_CHANGENAME: 1458 /* Do not notify about label change, this event is 1459 * not interesting to applications using netlink. 1460 */ 1461 inetdev_changename(dev, in_dev); 1462 1463 devinet_sysctl_unregister(in_dev); 1464 devinet_sysctl_register(in_dev); 1465 break; 1466 } 1467 out: 1468 return NOTIFY_DONE; 1469 } 1470 1471 static struct notifier_block ip_netdev_notifier = { 1472 .notifier_call = inetdev_event, 1473 }; 1474 1475 static size_t inet_nlmsg_size(void) 1476 { 1477 return NLMSG_ALIGN(sizeof(struct ifaddrmsg)) 1478 + nla_total_size(4) /* IFA_ADDRESS */ 1479 + nla_total_size(4) /* IFA_LOCAL */ 1480 + nla_total_size(4) /* IFA_BROADCAST */ 1481 + nla_total_size(IFNAMSIZ) /* IFA_LABEL */ 1482 + nla_total_size(4) /* IFA_FLAGS */ 1483 + nla_total_size(sizeof(struct ifa_cacheinfo)); /* IFA_CACHEINFO */ 1484 } 1485 1486 static inline u32 cstamp_delta(unsigned long cstamp) 1487 { 1488 return (cstamp - INITIAL_JIFFIES) * 100UL / HZ; 1489 } 1490 1491 static int put_cacheinfo(struct sk_buff *skb, unsigned long cstamp, 1492 unsigned long tstamp, u32 preferred, u32 valid) 1493 { 1494 struct ifa_cacheinfo ci; 1495 1496 ci.cstamp = cstamp_delta(cstamp); 1497 ci.tstamp = cstamp_delta(tstamp); 1498 ci.ifa_prefered = preferred; 1499 ci.ifa_valid = valid; 1500 1501 return nla_put(skb, IFA_CACHEINFO, sizeof(ci), &ci); 1502 } 1503 1504 static int inet_fill_ifaddr(struct sk_buff *skb, struct in_ifaddr *ifa, 1505 u32 portid, u32 seq, int event, unsigned int flags) 1506 { 1507 struct ifaddrmsg *ifm; 1508 struct nlmsghdr *nlh; 1509 u32 preferred, valid; 1510 1511 nlh = nlmsg_put(skb, portid, seq, event, sizeof(*ifm), flags); 1512 if (!nlh) 1513 return -EMSGSIZE; 1514 1515 ifm = nlmsg_data(nlh); 1516 ifm->ifa_family = AF_INET; 1517 ifm->ifa_prefixlen = ifa->ifa_prefixlen; 1518 ifm->ifa_flags = ifa->ifa_flags; 1519 ifm->ifa_scope = ifa->ifa_scope; 1520 ifm->ifa_index = ifa->ifa_dev->dev->ifindex; 1521 1522 if (!(ifm->ifa_flags & IFA_F_PERMANENT)) { 1523 preferred = ifa->ifa_preferred_lft; 1524 valid = ifa->ifa_valid_lft; 1525 if (preferred != INFINITY_LIFE_TIME) { 1526 long tval = (jiffies - ifa->ifa_tstamp) / HZ; 1527 1528 if (preferred > tval) 1529 preferred -= tval; 1530 else 1531 preferred = 0; 1532 if (valid != INFINITY_LIFE_TIME) { 1533 if (valid > tval) 1534 valid -= tval; 1535 else 1536 valid = 0; 1537 } 1538 } 1539 } else { 1540 preferred = INFINITY_LIFE_TIME; 1541 valid = INFINITY_LIFE_TIME; 1542 } 1543 if ((ifa->ifa_address && 1544 nla_put_in_addr(skb, IFA_ADDRESS, ifa->ifa_address)) || 1545 (ifa->ifa_local && 1546 nla_put_in_addr(skb, IFA_LOCAL, ifa->ifa_local)) || 1547 (ifa->ifa_broadcast && 1548 nla_put_in_addr(skb, IFA_BROADCAST, ifa->ifa_broadcast)) || 1549 (ifa->ifa_label[0] && 1550 nla_put_string(skb, IFA_LABEL, ifa->ifa_label)) || 1551 nla_put_u32(skb, IFA_FLAGS, ifa->ifa_flags) || 1552 put_cacheinfo(skb, ifa->ifa_cstamp, ifa->ifa_tstamp, 1553 preferred, valid)) 1554 goto nla_put_failure; 1555 1556 nlmsg_end(skb, nlh); 1557 return 0; 1558 1559 nla_put_failure: 1560 nlmsg_cancel(skb, nlh); 1561 return -EMSGSIZE; 1562 } 1563 1564 static int inet_dump_ifaddr(struct sk_buff *skb, struct netlink_callback *cb) 1565 { 1566 struct net *net = sock_net(skb->sk); 1567 int h, s_h; 1568 int idx, s_idx; 1569 int ip_idx, s_ip_idx; 1570 struct net_device *dev; 1571 struct in_device *in_dev; 1572 struct in_ifaddr *ifa; 1573 struct hlist_head *head; 1574 1575 s_h = cb->args[0]; 1576 s_idx = idx = cb->args[1]; 1577 s_ip_idx = ip_idx = cb->args[2]; 1578 1579 for (h = s_h; h < NETDEV_HASHENTRIES; h++, s_idx = 0) { 1580 idx = 0; 1581 head = &net->dev_index_head[h]; 1582 rcu_read_lock(); 1583 cb->seq = atomic_read(&net->ipv4.dev_addr_genid) ^ 1584 net->dev_base_seq; 1585 hlist_for_each_entry_rcu(dev, head, index_hlist) { 1586 if (idx < s_idx) 1587 goto cont; 1588 if (h > s_h || idx > s_idx) 1589 s_ip_idx = 0; 1590 in_dev = __in_dev_get_rcu(dev); 1591 if (!in_dev) 1592 goto cont; 1593 1594 for (ifa = in_dev->ifa_list, ip_idx = 0; ifa; 1595 ifa = ifa->ifa_next, ip_idx++) { 1596 if (ip_idx < s_ip_idx) 1597 continue; 1598 if (inet_fill_ifaddr(skb, ifa, 1599 NETLINK_CB(cb->skb).portid, 1600 cb->nlh->nlmsg_seq, 1601 RTM_NEWADDR, NLM_F_MULTI) < 0) { 1602 rcu_read_unlock(); 1603 goto done; 1604 } 1605 nl_dump_check_consistent(cb, nlmsg_hdr(skb)); 1606 } 1607 cont: 1608 idx++; 1609 } 1610 rcu_read_unlock(); 1611 } 1612 1613 done: 1614 cb->args[0] = h; 1615 cb->args[1] = idx; 1616 cb->args[2] = ip_idx; 1617 1618 return skb->len; 1619 } 1620 1621 static void rtmsg_ifa(int event, struct in_ifaddr *ifa, struct nlmsghdr *nlh, 1622 u32 portid) 1623 { 1624 struct sk_buff *skb; 1625 u32 seq = nlh ? nlh->nlmsg_seq : 0; 1626 int err = -ENOBUFS; 1627 struct net *net; 1628 1629 net = dev_net(ifa->ifa_dev->dev); 1630 skb = nlmsg_new(inet_nlmsg_size(), GFP_KERNEL); 1631 if (!skb) 1632 goto errout; 1633 1634 err = inet_fill_ifaddr(skb, ifa, portid, seq, event, 0); 1635 if (err < 0) { 1636 /* -EMSGSIZE implies BUG in inet_nlmsg_size() */ 1637 WARN_ON(err == -EMSGSIZE); 1638 kfree_skb(skb); 1639 goto errout; 1640 } 1641 rtnl_notify(skb, net, portid, RTNLGRP_IPV4_IFADDR, nlh, GFP_KERNEL); 1642 return; 1643 errout: 1644 if (err < 0) 1645 rtnl_set_sk_err(net, RTNLGRP_IPV4_IFADDR, err); 1646 } 1647 1648 static size_t inet_get_link_af_size(const struct net_device *dev) 1649 { 1650 struct in_device *in_dev = rcu_dereference_rtnl(dev->ip_ptr); 1651 1652 if (!in_dev) 1653 return 0; 1654 1655 return nla_total_size(IPV4_DEVCONF_MAX * 4); /* IFLA_INET_CONF */ 1656 } 1657 1658 static int inet_fill_link_af(struct sk_buff *skb, const struct net_device *dev) 1659 { 1660 struct in_device *in_dev = rcu_dereference_rtnl(dev->ip_ptr); 1661 struct nlattr *nla; 1662 int i; 1663 1664 if (!in_dev) 1665 return -ENODATA; 1666 1667 nla = nla_reserve(skb, IFLA_INET_CONF, IPV4_DEVCONF_MAX * 4); 1668 if (!nla) 1669 return -EMSGSIZE; 1670 1671 for (i = 0; i < IPV4_DEVCONF_MAX; i++) 1672 ((u32 *) nla_data(nla))[i] = in_dev->cnf.data[i]; 1673 1674 return 0; 1675 } 1676 1677 static const struct nla_policy inet_af_policy[IFLA_INET_MAX+1] = { 1678 [IFLA_INET_CONF] = { .type = NLA_NESTED }, 1679 }; 1680 1681 static int inet_validate_link_af(const struct net_device *dev, 1682 const struct nlattr *nla) 1683 { 1684 struct nlattr *a, *tb[IFLA_INET_MAX+1]; 1685 int err, rem; 1686 1687 if (dev && !__in_dev_get_rtnl(dev)) 1688 return -EAFNOSUPPORT; 1689 1690 err = nla_parse_nested(tb, IFLA_INET_MAX, nla, inet_af_policy); 1691 if (err < 0) 1692 return err; 1693 1694 if (tb[IFLA_INET_CONF]) { 1695 nla_for_each_nested(a, tb[IFLA_INET_CONF], rem) { 1696 int cfgid = nla_type(a); 1697 1698 if (nla_len(a) < 4) 1699 return -EINVAL; 1700 1701 if (cfgid <= 0 || cfgid > IPV4_DEVCONF_MAX) 1702 return -EINVAL; 1703 } 1704 } 1705 1706 return 0; 1707 } 1708 1709 static int inet_set_link_af(struct net_device *dev, const struct nlattr *nla) 1710 { 1711 struct in_device *in_dev = __in_dev_get_rtnl(dev); 1712 struct nlattr *a, *tb[IFLA_INET_MAX+1]; 1713 int rem; 1714 1715 if (!in_dev) 1716 return -EAFNOSUPPORT; 1717 1718 if (nla_parse_nested(tb, IFLA_INET_MAX, nla, NULL) < 0) 1719 BUG(); 1720 1721 if (tb[IFLA_INET_CONF]) { 1722 nla_for_each_nested(a, tb[IFLA_INET_CONF], rem) 1723 ipv4_devconf_set(in_dev, nla_type(a), nla_get_u32(a)); 1724 } 1725 1726 return 0; 1727 } 1728 1729 static int inet_netconf_msgsize_devconf(int type) 1730 { 1731 int size = NLMSG_ALIGN(sizeof(struct netconfmsg)) 1732 + nla_total_size(4); /* NETCONFA_IFINDEX */ 1733 1734 /* type -1 is used for ALL */ 1735 if (type == -1 || type == NETCONFA_FORWARDING) 1736 size += nla_total_size(4); 1737 if (type == -1 || type == NETCONFA_RP_FILTER) 1738 size += nla_total_size(4); 1739 if (type == -1 || type == NETCONFA_MC_FORWARDING) 1740 size += nla_total_size(4); 1741 if (type == -1 || type == NETCONFA_PROXY_NEIGH) 1742 size += nla_total_size(4); 1743 1744 return size; 1745 } 1746 1747 static int inet_netconf_fill_devconf(struct sk_buff *skb, int ifindex, 1748 struct ipv4_devconf *devconf, u32 portid, 1749 u32 seq, int event, unsigned int flags, 1750 int type) 1751 { 1752 struct nlmsghdr *nlh; 1753 struct netconfmsg *ncm; 1754 1755 nlh = nlmsg_put(skb, portid, seq, event, sizeof(struct netconfmsg), 1756 flags); 1757 if (!nlh) 1758 return -EMSGSIZE; 1759 1760 ncm = nlmsg_data(nlh); 1761 ncm->ncm_family = AF_INET; 1762 1763 if (nla_put_s32(skb, NETCONFA_IFINDEX, ifindex) < 0) 1764 goto nla_put_failure; 1765 1766 /* type -1 is used for ALL */ 1767 if ((type == -1 || type == NETCONFA_FORWARDING) && 1768 nla_put_s32(skb, NETCONFA_FORWARDING, 1769 IPV4_DEVCONF(*devconf, FORWARDING)) < 0) 1770 goto nla_put_failure; 1771 if ((type == -1 || type == NETCONFA_RP_FILTER) && 1772 nla_put_s32(skb, NETCONFA_RP_FILTER, 1773 IPV4_DEVCONF(*devconf, RP_FILTER)) < 0) 1774 goto nla_put_failure; 1775 if ((type == -1 || type == NETCONFA_MC_FORWARDING) && 1776 nla_put_s32(skb, NETCONFA_MC_FORWARDING, 1777 IPV4_DEVCONF(*devconf, MC_FORWARDING)) < 0) 1778 goto nla_put_failure; 1779 if ((type == -1 || type == NETCONFA_PROXY_NEIGH) && 1780 nla_put_s32(skb, NETCONFA_PROXY_NEIGH, 1781 IPV4_DEVCONF(*devconf, PROXY_ARP)) < 0) 1782 goto nla_put_failure; 1783 1784 nlmsg_end(skb, nlh); 1785 return 0; 1786 1787 nla_put_failure: 1788 nlmsg_cancel(skb, nlh); 1789 return -EMSGSIZE; 1790 } 1791 1792 void inet_netconf_notify_devconf(struct net *net, int type, int ifindex, 1793 struct ipv4_devconf *devconf) 1794 { 1795 struct sk_buff *skb; 1796 int err = -ENOBUFS; 1797 1798 skb = nlmsg_new(inet_netconf_msgsize_devconf(type), GFP_ATOMIC); 1799 if (!skb) 1800 goto errout; 1801 1802 err = inet_netconf_fill_devconf(skb, ifindex, devconf, 0, 0, 1803 RTM_NEWNETCONF, 0, type); 1804 if (err < 0) { 1805 /* -EMSGSIZE implies BUG in inet_netconf_msgsize_devconf() */ 1806 WARN_ON(err == -EMSGSIZE); 1807 kfree_skb(skb); 1808 goto errout; 1809 } 1810 rtnl_notify(skb, net, 0, RTNLGRP_IPV4_NETCONF, NULL, GFP_ATOMIC); 1811 return; 1812 errout: 1813 if (err < 0) 1814 rtnl_set_sk_err(net, RTNLGRP_IPV4_NETCONF, err); 1815 } 1816 1817 static const struct nla_policy devconf_ipv4_policy[NETCONFA_MAX+1] = { 1818 [NETCONFA_IFINDEX] = { .len = sizeof(int) }, 1819 [NETCONFA_FORWARDING] = { .len = sizeof(int) }, 1820 [NETCONFA_RP_FILTER] = { .len = sizeof(int) }, 1821 [NETCONFA_PROXY_NEIGH] = { .len = sizeof(int) }, 1822 }; 1823 1824 static int inet_netconf_get_devconf(struct sk_buff *in_skb, 1825 struct nlmsghdr *nlh) 1826 { 1827 struct net *net = sock_net(in_skb->sk); 1828 struct nlattr *tb[NETCONFA_MAX+1]; 1829 struct netconfmsg *ncm; 1830 struct sk_buff *skb; 1831 struct ipv4_devconf *devconf; 1832 struct in_device *in_dev; 1833 struct net_device *dev; 1834 int ifindex; 1835 int err; 1836 1837 err = nlmsg_parse(nlh, sizeof(*ncm), tb, NETCONFA_MAX, 1838 devconf_ipv4_policy); 1839 if (err < 0) 1840 goto errout; 1841 1842 err = EINVAL; 1843 if (!tb[NETCONFA_IFINDEX]) 1844 goto errout; 1845 1846 ifindex = nla_get_s32(tb[NETCONFA_IFINDEX]); 1847 switch (ifindex) { 1848 case NETCONFA_IFINDEX_ALL: 1849 devconf = net->ipv4.devconf_all; 1850 break; 1851 case NETCONFA_IFINDEX_DEFAULT: 1852 devconf = net->ipv4.devconf_dflt; 1853 break; 1854 default: 1855 dev = __dev_get_by_index(net, ifindex); 1856 if (!dev) 1857 goto errout; 1858 in_dev = __in_dev_get_rtnl(dev); 1859 if (!in_dev) 1860 goto errout; 1861 devconf = &in_dev->cnf; 1862 break; 1863 } 1864 1865 err = -ENOBUFS; 1866 skb = nlmsg_new(inet_netconf_msgsize_devconf(-1), GFP_ATOMIC); 1867 if (!skb) 1868 goto errout; 1869 1870 err = inet_netconf_fill_devconf(skb, ifindex, devconf, 1871 NETLINK_CB(in_skb).portid, 1872 nlh->nlmsg_seq, RTM_NEWNETCONF, 0, 1873 -1); 1874 if (err < 0) { 1875 /* -EMSGSIZE implies BUG in inet_netconf_msgsize_devconf() */ 1876 WARN_ON(err == -EMSGSIZE); 1877 kfree_skb(skb); 1878 goto errout; 1879 } 1880 err = rtnl_unicast(skb, net, NETLINK_CB(in_skb).portid); 1881 errout: 1882 return err; 1883 } 1884 1885 static int inet_netconf_dump_devconf(struct sk_buff *skb, 1886 struct netlink_callback *cb) 1887 { 1888 struct net *net = sock_net(skb->sk); 1889 int h, s_h; 1890 int idx, s_idx; 1891 struct net_device *dev; 1892 struct in_device *in_dev; 1893 struct hlist_head *head; 1894 1895 s_h = cb->args[0]; 1896 s_idx = idx = cb->args[1]; 1897 1898 for (h = s_h; h < NETDEV_HASHENTRIES; h++, s_idx = 0) { 1899 idx = 0; 1900 head = &net->dev_index_head[h]; 1901 rcu_read_lock(); 1902 cb->seq = atomic_read(&net->ipv4.dev_addr_genid) ^ 1903 net->dev_base_seq; 1904 hlist_for_each_entry_rcu(dev, head, index_hlist) { 1905 if (idx < s_idx) 1906 goto cont; 1907 in_dev = __in_dev_get_rcu(dev); 1908 if (!in_dev) 1909 goto cont; 1910 1911 if (inet_netconf_fill_devconf(skb, dev->ifindex, 1912 &in_dev->cnf, 1913 NETLINK_CB(cb->skb).portid, 1914 cb->nlh->nlmsg_seq, 1915 RTM_NEWNETCONF, 1916 NLM_F_MULTI, 1917 -1) < 0) { 1918 rcu_read_unlock(); 1919 goto done; 1920 } 1921 nl_dump_check_consistent(cb, nlmsg_hdr(skb)); 1922 cont: 1923 idx++; 1924 } 1925 rcu_read_unlock(); 1926 } 1927 if (h == NETDEV_HASHENTRIES) { 1928 if (inet_netconf_fill_devconf(skb, NETCONFA_IFINDEX_ALL, 1929 net->ipv4.devconf_all, 1930 NETLINK_CB(cb->skb).portid, 1931 cb->nlh->nlmsg_seq, 1932 RTM_NEWNETCONF, NLM_F_MULTI, 1933 -1) < 0) 1934 goto done; 1935 else 1936 h++; 1937 } 1938 if (h == NETDEV_HASHENTRIES + 1) { 1939 if (inet_netconf_fill_devconf(skb, NETCONFA_IFINDEX_DEFAULT, 1940 net->ipv4.devconf_dflt, 1941 NETLINK_CB(cb->skb).portid, 1942 cb->nlh->nlmsg_seq, 1943 RTM_NEWNETCONF, NLM_F_MULTI, 1944 -1) < 0) 1945 goto done; 1946 else 1947 h++; 1948 } 1949 done: 1950 cb->args[0] = h; 1951 cb->args[1] = idx; 1952 1953 return skb->len; 1954 } 1955 1956 #ifdef CONFIG_SYSCTL 1957 1958 static void devinet_copy_dflt_conf(struct net *net, int i) 1959 { 1960 struct net_device *dev; 1961 1962 rcu_read_lock(); 1963 for_each_netdev_rcu(net, dev) { 1964 struct in_device *in_dev; 1965 1966 in_dev = __in_dev_get_rcu(dev); 1967 if (in_dev && !test_bit(i, in_dev->cnf.state)) 1968 in_dev->cnf.data[i] = net->ipv4.devconf_dflt->data[i]; 1969 } 1970 rcu_read_unlock(); 1971 } 1972 1973 /* called with RTNL locked */ 1974 static void inet_forward_change(struct net *net) 1975 { 1976 struct net_device *dev; 1977 int on = IPV4_DEVCONF_ALL(net, FORWARDING); 1978 1979 IPV4_DEVCONF_ALL(net, ACCEPT_REDIRECTS) = !on; 1980 IPV4_DEVCONF_DFLT(net, FORWARDING) = on; 1981 inet_netconf_notify_devconf(net, NETCONFA_FORWARDING, 1982 NETCONFA_IFINDEX_ALL, 1983 net->ipv4.devconf_all); 1984 inet_netconf_notify_devconf(net, NETCONFA_FORWARDING, 1985 NETCONFA_IFINDEX_DEFAULT, 1986 net->ipv4.devconf_dflt); 1987 1988 for_each_netdev(net, dev) { 1989 struct in_device *in_dev; 1990 if (on) 1991 dev_disable_lro(dev); 1992 rcu_read_lock(); 1993 in_dev = __in_dev_get_rcu(dev); 1994 if (in_dev) { 1995 IN_DEV_CONF_SET(in_dev, FORWARDING, on); 1996 inet_netconf_notify_devconf(net, NETCONFA_FORWARDING, 1997 dev->ifindex, &in_dev->cnf); 1998 } 1999 rcu_read_unlock(); 2000 } 2001 } 2002 2003 static int devinet_conf_ifindex(struct net *net, struct ipv4_devconf *cnf) 2004 { 2005 if (cnf == net->ipv4.devconf_dflt) 2006 return NETCONFA_IFINDEX_DEFAULT; 2007 else if (cnf == net->ipv4.devconf_all) 2008 return NETCONFA_IFINDEX_ALL; 2009 else { 2010 struct in_device *idev 2011 = container_of(cnf, struct in_device, cnf); 2012 return idev->dev->ifindex; 2013 } 2014 } 2015 2016 static int devinet_conf_proc(struct ctl_table *ctl, int write, 2017 void __user *buffer, 2018 size_t *lenp, loff_t *ppos) 2019 { 2020 int old_value = *(int *)ctl->data; 2021 int ret = proc_dointvec(ctl, write, buffer, lenp, ppos); 2022 int new_value = *(int *)ctl->data; 2023 2024 if (write) { 2025 struct ipv4_devconf *cnf = ctl->extra1; 2026 struct net *net = ctl->extra2; 2027 int i = (int *)ctl->data - cnf->data; 2028 int ifindex; 2029 2030 set_bit(i, cnf->state); 2031 2032 if (cnf == net->ipv4.devconf_dflt) 2033 devinet_copy_dflt_conf(net, i); 2034 if (i == IPV4_DEVCONF_ACCEPT_LOCAL - 1 || 2035 i == IPV4_DEVCONF_ROUTE_LOCALNET - 1) 2036 if ((new_value == 0) && (old_value != 0)) 2037 rt_cache_flush(net); 2038 2039 if (i == IPV4_DEVCONF_RP_FILTER - 1 && 2040 new_value != old_value) { 2041 ifindex = devinet_conf_ifindex(net, cnf); 2042 inet_netconf_notify_devconf(net, NETCONFA_RP_FILTER, 2043 ifindex, cnf); 2044 } 2045 if (i == IPV4_DEVCONF_PROXY_ARP - 1 && 2046 new_value != old_value) { 2047 ifindex = devinet_conf_ifindex(net, cnf); 2048 inet_netconf_notify_devconf(net, NETCONFA_PROXY_NEIGH, 2049 ifindex, cnf); 2050 } 2051 } 2052 2053 return ret; 2054 } 2055 2056 static int devinet_sysctl_forward(struct ctl_table *ctl, int write, 2057 void __user *buffer, 2058 size_t *lenp, loff_t *ppos) 2059 { 2060 int *valp = ctl->data; 2061 int val = *valp; 2062 loff_t pos = *ppos; 2063 int ret = proc_dointvec(ctl, write, buffer, lenp, ppos); 2064 2065 if (write && *valp != val) { 2066 struct net *net = ctl->extra2; 2067 2068 if (valp != &IPV4_DEVCONF_DFLT(net, FORWARDING)) { 2069 if (!rtnl_trylock()) { 2070 /* Restore the original values before restarting */ 2071 *valp = val; 2072 *ppos = pos; 2073 return restart_syscall(); 2074 } 2075 if (valp == &IPV4_DEVCONF_ALL(net, FORWARDING)) { 2076 inet_forward_change(net); 2077 } else { 2078 struct ipv4_devconf *cnf = ctl->extra1; 2079 struct in_device *idev = 2080 container_of(cnf, struct in_device, cnf); 2081 if (*valp) 2082 dev_disable_lro(idev->dev); 2083 inet_netconf_notify_devconf(net, 2084 NETCONFA_FORWARDING, 2085 idev->dev->ifindex, 2086 cnf); 2087 } 2088 rtnl_unlock(); 2089 rt_cache_flush(net); 2090 } else 2091 inet_netconf_notify_devconf(net, NETCONFA_FORWARDING, 2092 NETCONFA_IFINDEX_DEFAULT, 2093 net->ipv4.devconf_dflt); 2094 } 2095 2096 return ret; 2097 } 2098 2099 static int ipv4_doint_and_flush(struct ctl_table *ctl, int write, 2100 void __user *buffer, 2101 size_t *lenp, loff_t *ppos) 2102 { 2103 int *valp = ctl->data; 2104 int val = *valp; 2105 int ret = proc_dointvec(ctl, write, buffer, lenp, ppos); 2106 struct net *net = ctl->extra2; 2107 2108 if (write && *valp != val) 2109 rt_cache_flush(net); 2110 2111 return ret; 2112 } 2113 2114 #define DEVINET_SYSCTL_ENTRY(attr, name, mval, proc) \ 2115 { \ 2116 .procname = name, \ 2117 .data = ipv4_devconf.data + \ 2118 IPV4_DEVCONF_ ## attr - 1, \ 2119 .maxlen = sizeof(int), \ 2120 .mode = mval, \ 2121 .proc_handler = proc, \ 2122 .extra1 = &ipv4_devconf, \ 2123 } 2124 2125 #define DEVINET_SYSCTL_RW_ENTRY(attr, name) \ 2126 DEVINET_SYSCTL_ENTRY(attr, name, 0644, devinet_conf_proc) 2127 2128 #define DEVINET_SYSCTL_RO_ENTRY(attr, name) \ 2129 DEVINET_SYSCTL_ENTRY(attr, name, 0444, devinet_conf_proc) 2130 2131 #define DEVINET_SYSCTL_COMPLEX_ENTRY(attr, name, proc) \ 2132 DEVINET_SYSCTL_ENTRY(attr, name, 0644, proc) 2133 2134 #define DEVINET_SYSCTL_FLUSHING_ENTRY(attr, name) \ 2135 DEVINET_SYSCTL_COMPLEX_ENTRY(attr, name, ipv4_doint_and_flush) 2136 2137 static struct devinet_sysctl_table { 2138 struct ctl_table_header *sysctl_header; 2139 struct ctl_table devinet_vars[__IPV4_DEVCONF_MAX]; 2140 } devinet_sysctl = { 2141 .devinet_vars = { 2142 DEVINET_SYSCTL_COMPLEX_ENTRY(FORWARDING, "forwarding", 2143 devinet_sysctl_forward), 2144 DEVINET_SYSCTL_RO_ENTRY(MC_FORWARDING, "mc_forwarding"), 2145 2146 DEVINET_SYSCTL_RW_ENTRY(ACCEPT_REDIRECTS, "accept_redirects"), 2147 DEVINET_SYSCTL_RW_ENTRY(SECURE_REDIRECTS, "secure_redirects"), 2148 DEVINET_SYSCTL_RW_ENTRY(SHARED_MEDIA, "shared_media"), 2149 DEVINET_SYSCTL_RW_ENTRY(RP_FILTER, "rp_filter"), 2150 DEVINET_SYSCTL_RW_ENTRY(SEND_REDIRECTS, "send_redirects"), 2151 DEVINET_SYSCTL_RW_ENTRY(ACCEPT_SOURCE_ROUTE, 2152 "accept_source_route"), 2153 DEVINET_SYSCTL_RW_ENTRY(ACCEPT_LOCAL, "accept_local"), 2154 DEVINET_SYSCTL_RW_ENTRY(SRC_VMARK, "src_valid_mark"), 2155 DEVINET_SYSCTL_RW_ENTRY(PROXY_ARP, "proxy_arp"), 2156 DEVINET_SYSCTL_RW_ENTRY(MEDIUM_ID, "medium_id"), 2157 DEVINET_SYSCTL_RW_ENTRY(BOOTP_RELAY, "bootp_relay"), 2158 DEVINET_SYSCTL_RW_ENTRY(LOG_MARTIANS, "log_martians"), 2159 DEVINET_SYSCTL_RW_ENTRY(TAG, "tag"), 2160 DEVINET_SYSCTL_RW_ENTRY(ARPFILTER, "arp_filter"), 2161 DEVINET_SYSCTL_RW_ENTRY(ARP_ANNOUNCE, "arp_announce"), 2162 DEVINET_SYSCTL_RW_ENTRY(ARP_IGNORE, "arp_ignore"), 2163 DEVINET_SYSCTL_RW_ENTRY(ARP_ACCEPT, "arp_accept"), 2164 DEVINET_SYSCTL_RW_ENTRY(ARP_NOTIFY, "arp_notify"), 2165 DEVINET_SYSCTL_RW_ENTRY(PROXY_ARP_PVLAN, "proxy_arp_pvlan"), 2166 DEVINET_SYSCTL_RW_ENTRY(FORCE_IGMP_VERSION, 2167 "force_igmp_version"), 2168 DEVINET_SYSCTL_RW_ENTRY(IGMPV2_UNSOLICITED_REPORT_INTERVAL, 2169 "igmpv2_unsolicited_report_interval"), 2170 DEVINET_SYSCTL_RW_ENTRY(IGMPV3_UNSOLICITED_REPORT_INTERVAL, 2171 "igmpv3_unsolicited_report_interval"), 2172 2173 DEVINET_SYSCTL_FLUSHING_ENTRY(NOXFRM, "disable_xfrm"), 2174 DEVINET_SYSCTL_FLUSHING_ENTRY(NOPOLICY, "disable_policy"), 2175 DEVINET_SYSCTL_FLUSHING_ENTRY(PROMOTE_SECONDARIES, 2176 "promote_secondaries"), 2177 DEVINET_SYSCTL_FLUSHING_ENTRY(ROUTE_LOCALNET, 2178 "route_localnet"), 2179 }, 2180 }; 2181 2182 static int __devinet_sysctl_register(struct net *net, char *dev_name, 2183 struct ipv4_devconf *p) 2184 { 2185 int i; 2186 struct devinet_sysctl_table *t; 2187 char path[sizeof("net/ipv4/conf/") + IFNAMSIZ]; 2188 2189 t = kmemdup(&devinet_sysctl, sizeof(*t), GFP_KERNEL); 2190 if (!t) 2191 goto out; 2192 2193 for (i = 0; i < ARRAY_SIZE(t->devinet_vars) - 1; i++) { 2194 t->devinet_vars[i].data += (char *)p - (char *)&ipv4_devconf; 2195 t->devinet_vars[i].extra1 = p; 2196 t->devinet_vars[i].extra2 = net; 2197 } 2198 2199 snprintf(path, sizeof(path), "net/ipv4/conf/%s", dev_name); 2200 2201 t->sysctl_header = register_net_sysctl(net, path, t->devinet_vars); 2202 if (!t->sysctl_header) 2203 goto free; 2204 2205 p->sysctl = t; 2206 return 0; 2207 2208 free: 2209 kfree(t); 2210 out: 2211 return -ENOBUFS; 2212 } 2213 2214 static void __devinet_sysctl_unregister(struct ipv4_devconf *cnf) 2215 { 2216 struct devinet_sysctl_table *t = cnf->sysctl; 2217 2218 if (!t) 2219 return; 2220 2221 cnf->sysctl = NULL; 2222 unregister_net_sysctl_table(t->sysctl_header); 2223 kfree(t); 2224 } 2225 2226 static int devinet_sysctl_register(struct in_device *idev) 2227 { 2228 int err; 2229 2230 if (!sysctl_dev_name_is_allowed(idev->dev->name)) 2231 return -EINVAL; 2232 2233 err = neigh_sysctl_register(idev->dev, idev->arp_parms, NULL); 2234 if (err) 2235 return err; 2236 err = __devinet_sysctl_register(dev_net(idev->dev), idev->dev->name, 2237 &idev->cnf); 2238 if (err) 2239 neigh_sysctl_unregister(idev->arp_parms); 2240 return err; 2241 } 2242 2243 static void devinet_sysctl_unregister(struct in_device *idev) 2244 { 2245 __devinet_sysctl_unregister(&idev->cnf); 2246 neigh_sysctl_unregister(idev->arp_parms); 2247 } 2248 2249 static struct ctl_table ctl_forward_entry[] = { 2250 { 2251 .procname = "ip_forward", 2252 .data = &ipv4_devconf.data[ 2253 IPV4_DEVCONF_FORWARDING - 1], 2254 .maxlen = sizeof(int), 2255 .mode = 0644, 2256 .proc_handler = devinet_sysctl_forward, 2257 .extra1 = &ipv4_devconf, 2258 .extra2 = &init_net, 2259 }, 2260 { }, 2261 }; 2262 #endif 2263 2264 static __net_init int devinet_init_net(struct net *net) 2265 { 2266 int err; 2267 struct ipv4_devconf *all, *dflt; 2268 #ifdef CONFIG_SYSCTL 2269 struct ctl_table *tbl = ctl_forward_entry; 2270 struct ctl_table_header *forw_hdr; 2271 #endif 2272 2273 err = -ENOMEM; 2274 all = &ipv4_devconf; 2275 dflt = &ipv4_devconf_dflt; 2276 2277 if (!net_eq(net, &init_net)) { 2278 all = kmemdup(all, sizeof(ipv4_devconf), GFP_KERNEL); 2279 if (!all) 2280 goto err_alloc_all; 2281 2282 dflt = kmemdup(dflt, sizeof(ipv4_devconf_dflt), GFP_KERNEL); 2283 if (!dflt) 2284 goto err_alloc_dflt; 2285 2286 #ifdef CONFIG_SYSCTL 2287 tbl = kmemdup(tbl, sizeof(ctl_forward_entry), GFP_KERNEL); 2288 if (!tbl) 2289 goto err_alloc_ctl; 2290 2291 tbl[0].data = &all->data[IPV4_DEVCONF_FORWARDING - 1]; 2292 tbl[0].extra1 = all; 2293 tbl[0].extra2 = net; 2294 #endif 2295 } 2296 2297 #ifdef CONFIG_SYSCTL 2298 err = __devinet_sysctl_register(net, "all", all); 2299 if (err < 0) 2300 goto err_reg_all; 2301 2302 err = __devinet_sysctl_register(net, "default", dflt); 2303 if (err < 0) 2304 goto err_reg_dflt; 2305 2306 err = -ENOMEM; 2307 forw_hdr = register_net_sysctl(net, "net/ipv4", tbl); 2308 if (!forw_hdr) 2309 goto err_reg_ctl; 2310 net->ipv4.forw_hdr = forw_hdr; 2311 #endif 2312 2313 net->ipv4.devconf_all = all; 2314 net->ipv4.devconf_dflt = dflt; 2315 return 0; 2316 2317 #ifdef CONFIG_SYSCTL 2318 err_reg_ctl: 2319 __devinet_sysctl_unregister(dflt); 2320 err_reg_dflt: 2321 __devinet_sysctl_unregister(all); 2322 err_reg_all: 2323 if (tbl != ctl_forward_entry) 2324 kfree(tbl); 2325 err_alloc_ctl: 2326 #endif 2327 if (dflt != &ipv4_devconf_dflt) 2328 kfree(dflt); 2329 err_alloc_dflt: 2330 if (all != &ipv4_devconf) 2331 kfree(all); 2332 err_alloc_all: 2333 return err; 2334 } 2335 2336 static __net_exit void devinet_exit_net(struct net *net) 2337 { 2338 #ifdef CONFIG_SYSCTL 2339 struct ctl_table *tbl; 2340 2341 tbl = net->ipv4.forw_hdr->ctl_table_arg; 2342 unregister_net_sysctl_table(net->ipv4.forw_hdr); 2343 __devinet_sysctl_unregister(net->ipv4.devconf_dflt); 2344 __devinet_sysctl_unregister(net->ipv4.devconf_all); 2345 kfree(tbl); 2346 #endif 2347 kfree(net->ipv4.devconf_dflt); 2348 kfree(net->ipv4.devconf_all); 2349 } 2350 2351 static __net_initdata struct pernet_operations devinet_ops = { 2352 .init = devinet_init_net, 2353 .exit = devinet_exit_net, 2354 }; 2355 2356 static struct rtnl_af_ops inet_af_ops __read_mostly = { 2357 .family = AF_INET, 2358 .fill_link_af = inet_fill_link_af, 2359 .get_link_af_size = inet_get_link_af_size, 2360 .validate_link_af = inet_validate_link_af, 2361 .set_link_af = inet_set_link_af, 2362 }; 2363 2364 void __init devinet_init(void) 2365 { 2366 int i; 2367 2368 for (i = 0; i < IN4_ADDR_HSIZE; i++) 2369 INIT_HLIST_HEAD(&inet_addr_lst[i]); 2370 2371 register_pernet_subsys(&devinet_ops); 2372 2373 register_gifconf(PF_INET, inet_gifconf); 2374 register_netdevice_notifier(&ip_netdev_notifier); 2375 2376 queue_delayed_work(system_power_efficient_wq, &check_lifetime_work, 0); 2377 2378 rtnl_af_register(&inet_af_ops); 2379 2380 rtnl_register(PF_INET, RTM_NEWADDR, inet_rtm_newaddr, NULL, NULL); 2381 rtnl_register(PF_INET, RTM_DELADDR, inet_rtm_deladdr, NULL, NULL); 2382 rtnl_register(PF_INET, RTM_GETADDR, NULL, inet_dump_ifaddr, NULL); 2383 rtnl_register(PF_INET, RTM_GETNETCONF, inet_netconf_get_devconf, 2384 inet_netconf_dump_devconf, NULL); 2385 } 2386 2387