1 /* 2 * Bond several ethernet interfaces into a Cisco, running 'Etherchannel'. 3 * 4 * Portions are (c) Copyright 1995 Simon "Guru Aleph-Null" Janes 5 * NCM: Network and Communications Management, Inc. 6 * 7 * BUT, I'm the one who modified it for ethernet, so: 8 * (c) Copyright 1999, Thomas Davis, tadavis@lbl.gov 9 * 10 * This software may be used and distributed according to the terms 11 * of the GNU Public License, incorporated herein by reference. 12 * 13 */ 14 15 #ifndef _NET_BONDING_H 16 #define _NET_BONDING_H 17 18 #include <linux/timer.h> 19 #include <linux/proc_fs.h> 20 #include <linux/if_bonding.h> 21 #include <linux/cpumask.h> 22 #include <linux/in6.h> 23 #include <linux/netpoll.h> 24 #include <linux/inetdevice.h> 25 #include <linux/etherdevice.h> 26 #include <linux/reciprocal_div.h> 27 #include <linux/if_link.h> 28 29 #include <net/bond_3ad.h> 30 #include <net/bond_alb.h> 31 #include <net/bond_options.h> 32 33 #define BOND_MAX_ARP_TARGETS 16 34 35 #define BOND_DEFAULT_MIIMON 100 36 37 #ifndef __long_aligned 38 #define __long_aligned __attribute__((aligned((sizeof(long))))) 39 #endif 40 41 #define BOND_MODE(bond) ((bond)->params.mode) 42 43 /* slave list primitives */ 44 #define bond_slave_list(bond) (&(bond)->dev->adj_list.lower) 45 46 #define bond_has_slaves(bond) !list_empty(bond_slave_list(bond)) 47 48 /* IMPORTANT: bond_first/last_slave can return NULL in case of an empty list */ 49 #define bond_first_slave(bond) \ 50 (bond_has_slaves(bond) ? \ 51 netdev_adjacent_get_private(bond_slave_list(bond)->next) : \ 52 NULL) 53 #define bond_last_slave(bond) \ 54 (bond_has_slaves(bond) ? \ 55 netdev_adjacent_get_private(bond_slave_list(bond)->prev) : \ 56 NULL) 57 58 /* Caller must have rcu_read_lock */ 59 #define bond_first_slave_rcu(bond) \ 60 netdev_lower_get_first_private_rcu(bond->dev) 61 62 #define bond_is_first_slave(bond, pos) (pos == bond_first_slave(bond)) 63 #define bond_is_last_slave(bond, pos) (pos == bond_last_slave(bond)) 64 65 /** 66 * bond_for_each_slave - iterate over all slaves 67 * @bond: the bond holding this list 68 * @pos: current slave 69 * @iter: list_head * iterator 70 * 71 * Caller must hold RTNL 72 */ 73 #define bond_for_each_slave(bond, pos, iter) \ 74 netdev_for_each_lower_private((bond)->dev, pos, iter) 75 76 /* Caller must have rcu_read_lock */ 77 #define bond_for_each_slave_rcu(bond, pos, iter) \ 78 netdev_for_each_lower_private_rcu((bond)->dev, pos, iter) 79 80 #ifdef CONFIG_NET_POLL_CONTROLLER 81 extern atomic_t netpoll_block_tx; 82 83 static inline void block_netpoll_tx(void) 84 { 85 atomic_inc(&netpoll_block_tx); 86 } 87 88 static inline void unblock_netpoll_tx(void) 89 { 90 atomic_dec(&netpoll_block_tx); 91 } 92 93 static inline int is_netpoll_tx_blocked(struct net_device *dev) 94 { 95 if (unlikely(netpoll_tx_running(dev))) 96 return atomic_read(&netpoll_block_tx); 97 return 0; 98 } 99 #else 100 #define block_netpoll_tx() 101 #define unblock_netpoll_tx() 102 #define is_netpoll_tx_blocked(dev) (0) 103 #endif 104 105 struct bond_params { 106 int mode; 107 int xmit_policy; 108 int miimon; 109 u8 num_peer_notif; 110 int arp_interval; 111 int arp_validate; 112 int arp_all_targets; 113 int use_carrier; 114 int fail_over_mac; 115 int updelay; 116 int downdelay; 117 int lacp_fast; 118 unsigned int min_links; 119 int ad_select; 120 char primary[IFNAMSIZ]; 121 int primary_reselect; 122 __be32 arp_targets[BOND_MAX_ARP_TARGETS]; 123 int tx_queues; 124 int all_slaves_active; 125 int resend_igmp; 126 int lp_interval; 127 int packets_per_slave; 128 int tlb_dynamic_lb; 129 struct reciprocal_value reciprocal_packets_per_slave; 130 u16 ad_actor_sys_prio; 131 u16 ad_user_port_key; 132 133 /* 2 bytes of padding : see ether_addr_equal_64bits() */ 134 u8 ad_actor_system[ETH_ALEN + 2]; 135 }; 136 137 struct bond_parm_tbl { 138 char *modename; 139 int mode; 140 }; 141 142 struct slave { 143 struct net_device *dev; /* first - useful for panic debug */ 144 struct bonding *bond; /* our master */ 145 int delay; 146 /* all three in jiffies */ 147 unsigned long last_link_up; 148 unsigned long last_rx; 149 unsigned long target_last_arp_rx[BOND_MAX_ARP_TARGETS]; 150 s8 link; /* one of BOND_LINK_XXXX */ 151 s8 link_new_state; /* one of BOND_LINK_XXXX */ 152 s8 new_link; 153 u8 backup:1, /* indicates backup slave. Value corresponds with 154 BOND_STATE_ACTIVE and BOND_STATE_BACKUP */ 155 inactive:1, /* indicates inactive slave */ 156 should_notify:1, /* indicates whether the state changed */ 157 should_notify_link:1; /* indicates whether the link changed */ 158 u8 duplex; 159 u32 original_mtu; 160 u32 link_failure_count; 161 u32 speed; 162 u16 queue_id; 163 u8 perm_hwaddr[MAX_ADDR_LEN]; 164 struct ad_slave_info *ad_info; 165 struct tlb_slave_info tlb_info; 166 #ifdef CONFIG_NET_POLL_CONTROLLER 167 struct netpoll *np; 168 #endif 169 struct delayed_work notify_work; 170 struct kobject kobj; 171 struct rtnl_link_stats64 slave_stats; 172 }; 173 174 struct bond_up_slave { 175 unsigned int count; 176 struct rcu_head rcu; 177 struct slave *arr[0]; 178 }; 179 180 /* 181 * Link pseudo-state only used internally by monitors 182 */ 183 #define BOND_LINK_NOCHANGE -1 184 185 /* 186 * Here are the locking policies for the two bonding locks: 187 * Get rcu_read_lock when reading or RTNL when writing slave list. 188 */ 189 struct bonding { 190 struct net_device *dev; /* first - useful for panic debug */ 191 struct slave __rcu *curr_active_slave; 192 struct slave __rcu *current_arp_slave; 193 struct slave __rcu *primary_slave; 194 struct bond_up_slave __rcu *slave_arr; /* Array of usable slaves */ 195 bool force_primary; 196 u32 nest_level; 197 s32 slave_cnt; /* never change this value outside the attach/detach wrappers */ 198 int (*recv_probe)(const struct sk_buff *, struct bonding *, 199 struct slave *); 200 /* mode_lock is used for mode-specific locking needs, currently used by: 201 * 3ad mode (4) - protect against running bond_3ad_unbind_slave() and 202 * bond_3ad_state_machine_handler() concurrently and also 203 * the access to the state machine shared variables. 204 * TLB mode (5) - to sync the use and modifications of its hash table 205 * ALB mode (6) - to sync the use and modifications of its hash table 206 */ 207 spinlock_t mode_lock; 208 spinlock_t stats_lock; 209 u8 send_peer_notif; 210 u8 igmp_retrans; 211 #ifdef CONFIG_PROC_FS 212 struct proc_dir_entry *proc_entry; 213 char proc_file_name[IFNAMSIZ]; 214 #endif /* CONFIG_PROC_FS */ 215 struct list_head bond_list; 216 u32 rr_tx_counter; 217 struct ad_bond_info ad_info; 218 struct alb_bond_info alb_info; 219 struct bond_params params; 220 struct workqueue_struct *wq; 221 struct delayed_work mii_work; 222 struct delayed_work arp_work; 223 struct delayed_work alb_work; 224 struct delayed_work ad_work; 225 struct delayed_work mcast_work; 226 struct delayed_work slave_arr_work; 227 #ifdef CONFIG_DEBUG_FS 228 /* debugging support via debugfs */ 229 struct dentry *debug_dir; 230 #endif /* CONFIG_DEBUG_FS */ 231 struct rtnl_link_stats64 bond_stats; 232 }; 233 234 #define bond_slave_get_rcu(dev) \ 235 ((struct slave *) rcu_dereference(dev->rx_handler_data)) 236 237 #define bond_slave_get_rtnl(dev) \ 238 ((struct slave *) rtnl_dereference(dev->rx_handler_data)) 239 240 void bond_queue_slave_event(struct slave *slave); 241 void bond_lower_state_changed(struct slave *slave); 242 243 struct bond_vlan_tag { 244 __be16 vlan_proto; 245 unsigned short vlan_id; 246 }; 247 248 /** 249 * Returns NULL if the net_device does not belong to any of the bond's slaves 250 * 251 * Caller must hold bond lock for read 252 */ 253 static inline struct slave *bond_get_slave_by_dev(struct bonding *bond, 254 struct net_device *slave_dev) 255 { 256 return netdev_lower_dev_get_private(bond->dev, slave_dev); 257 } 258 259 static inline struct bonding *bond_get_bond_by_slave(struct slave *slave) 260 { 261 return slave->bond; 262 } 263 264 static inline bool bond_should_override_tx_queue(struct bonding *bond) 265 { 266 return BOND_MODE(bond) == BOND_MODE_ACTIVEBACKUP || 267 BOND_MODE(bond) == BOND_MODE_ROUNDROBIN; 268 } 269 270 static inline bool bond_is_lb(const struct bonding *bond) 271 { 272 return BOND_MODE(bond) == BOND_MODE_TLB || 273 BOND_MODE(bond) == BOND_MODE_ALB; 274 } 275 276 static inline bool bond_needs_speed_duplex(const struct bonding *bond) 277 { 278 return BOND_MODE(bond) == BOND_MODE_8023AD || bond_is_lb(bond); 279 } 280 281 static inline bool bond_is_nondyn_tlb(const struct bonding *bond) 282 { 283 return (bond_is_lb(bond) && bond->params.tlb_dynamic_lb == 0); 284 } 285 286 static inline bool bond_mode_can_use_xmit_hash(const struct bonding *bond) 287 { 288 return (BOND_MODE(bond) == BOND_MODE_8023AD || 289 BOND_MODE(bond) == BOND_MODE_XOR || 290 BOND_MODE(bond) == BOND_MODE_TLB || 291 BOND_MODE(bond) == BOND_MODE_ALB); 292 } 293 294 static inline bool bond_mode_uses_xmit_hash(const struct bonding *bond) 295 { 296 return (BOND_MODE(bond) == BOND_MODE_8023AD || 297 BOND_MODE(bond) == BOND_MODE_XOR || 298 bond_is_nondyn_tlb(bond)); 299 } 300 301 static inline bool bond_mode_uses_arp(int mode) 302 { 303 return mode != BOND_MODE_8023AD && mode != BOND_MODE_TLB && 304 mode != BOND_MODE_ALB; 305 } 306 307 static inline bool bond_mode_uses_primary(int mode) 308 { 309 return mode == BOND_MODE_ACTIVEBACKUP || mode == BOND_MODE_TLB || 310 mode == BOND_MODE_ALB; 311 } 312 313 static inline bool bond_uses_primary(struct bonding *bond) 314 { 315 return bond_mode_uses_primary(BOND_MODE(bond)); 316 } 317 318 static inline struct net_device *bond_option_active_slave_get_rcu(struct bonding *bond) 319 { 320 struct slave *slave = rcu_dereference(bond->curr_active_slave); 321 322 return bond_uses_primary(bond) && slave ? slave->dev : NULL; 323 } 324 325 static inline bool bond_slave_is_up(struct slave *slave) 326 { 327 return netif_running(slave->dev) && netif_carrier_ok(slave->dev); 328 } 329 330 static inline void bond_set_active_slave(struct slave *slave) 331 { 332 if (slave->backup) { 333 slave->backup = 0; 334 bond_queue_slave_event(slave); 335 bond_lower_state_changed(slave); 336 } 337 } 338 339 static inline void bond_set_backup_slave(struct slave *slave) 340 { 341 if (!slave->backup) { 342 slave->backup = 1; 343 bond_queue_slave_event(slave); 344 bond_lower_state_changed(slave); 345 } 346 } 347 348 static inline void bond_set_slave_state(struct slave *slave, 349 int slave_state, bool notify) 350 { 351 if (slave->backup == slave_state) 352 return; 353 354 slave->backup = slave_state; 355 if (notify) { 356 bond_lower_state_changed(slave); 357 bond_queue_slave_event(slave); 358 slave->should_notify = 0; 359 } else { 360 if (slave->should_notify) 361 slave->should_notify = 0; 362 else 363 slave->should_notify = 1; 364 } 365 } 366 367 static inline void bond_slave_state_change(struct bonding *bond) 368 { 369 struct list_head *iter; 370 struct slave *tmp; 371 372 bond_for_each_slave(bond, tmp, iter) { 373 if (tmp->link == BOND_LINK_UP) 374 bond_set_active_slave(tmp); 375 else if (tmp->link == BOND_LINK_DOWN) 376 bond_set_backup_slave(tmp); 377 } 378 } 379 380 static inline void bond_slave_state_notify(struct bonding *bond) 381 { 382 struct list_head *iter; 383 struct slave *tmp; 384 385 bond_for_each_slave(bond, tmp, iter) { 386 if (tmp->should_notify) { 387 bond_lower_state_changed(tmp); 388 tmp->should_notify = 0; 389 } 390 } 391 } 392 393 static inline int bond_slave_state(struct slave *slave) 394 { 395 return slave->backup; 396 } 397 398 static inline bool bond_is_active_slave(struct slave *slave) 399 { 400 return !bond_slave_state(slave); 401 } 402 403 static inline bool bond_slave_can_tx(struct slave *slave) 404 { 405 return bond_slave_is_up(slave) && slave->link == BOND_LINK_UP && 406 bond_is_active_slave(slave); 407 } 408 409 static inline bool bond_is_active_slave_dev(const struct net_device *slave_dev) 410 { 411 struct slave *slave; 412 bool active; 413 414 rcu_read_lock(); 415 slave = bond_slave_get_rcu(slave_dev); 416 active = bond_is_active_slave(slave); 417 rcu_read_unlock(); 418 419 return active; 420 } 421 422 static inline void bond_hw_addr_copy(u8 *dst, const u8 *src, unsigned int len) 423 { 424 if (len == ETH_ALEN) { 425 ether_addr_copy(dst, src); 426 return; 427 } 428 429 memcpy(dst, src, len); 430 } 431 432 #define BOND_PRI_RESELECT_ALWAYS 0 433 #define BOND_PRI_RESELECT_BETTER 1 434 #define BOND_PRI_RESELECT_FAILURE 2 435 436 #define BOND_FOM_NONE 0 437 #define BOND_FOM_ACTIVE 1 438 #define BOND_FOM_FOLLOW 2 439 440 #define BOND_ARP_TARGETS_ANY 0 441 #define BOND_ARP_TARGETS_ALL 1 442 443 #define BOND_ARP_VALIDATE_NONE 0 444 #define BOND_ARP_VALIDATE_ACTIVE (1 << BOND_STATE_ACTIVE) 445 #define BOND_ARP_VALIDATE_BACKUP (1 << BOND_STATE_BACKUP) 446 #define BOND_ARP_VALIDATE_ALL (BOND_ARP_VALIDATE_ACTIVE | \ 447 BOND_ARP_VALIDATE_BACKUP) 448 #define BOND_ARP_FILTER (BOND_ARP_VALIDATE_ALL + 1) 449 #define BOND_ARP_FILTER_ACTIVE (BOND_ARP_VALIDATE_ACTIVE | \ 450 BOND_ARP_FILTER) 451 #define BOND_ARP_FILTER_BACKUP (BOND_ARP_VALIDATE_BACKUP | \ 452 BOND_ARP_FILTER) 453 454 #define BOND_SLAVE_NOTIFY_NOW true 455 #define BOND_SLAVE_NOTIFY_LATER false 456 457 static inline int slave_do_arp_validate(struct bonding *bond, 458 struct slave *slave) 459 { 460 return bond->params.arp_validate & (1 << bond_slave_state(slave)); 461 } 462 463 static inline int slave_do_arp_validate_only(struct bonding *bond) 464 { 465 return bond->params.arp_validate & BOND_ARP_FILTER; 466 } 467 468 static inline int bond_is_ip_target_ok(__be32 addr) 469 { 470 return !ipv4_is_lbcast(addr) && !ipv4_is_zeronet(addr); 471 } 472 473 /* Get the oldest arp which we've received on this slave for bond's 474 * arp_targets. 475 */ 476 static inline unsigned long slave_oldest_target_arp_rx(struct bonding *bond, 477 struct slave *slave) 478 { 479 int i = 1; 480 unsigned long ret = slave->target_last_arp_rx[0]; 481 482 for (; (i < BOND_MAX_ARP_TARGETS) && bond->params.arp_targets[i]; i++) 483 if (time_before(slave->target_last_arp_rx[i], ret)) 484 ret = slave->target_last_arp_rx[i]; 485 486 return ret; 487 } 488 489 static inline unsigned long slave_last_rx(struct bonding *bond, 490 struct slave *slave) 491 { 492 if (bond->params.arp_all_targets == BOND_ARP_TARGETS_ALL) 493 return slave_oldest_target_arp_rx(bond, slave); 494 495 return slave->last_rx; 496 } 497 498 #ifdef CONFIG_NET_POLL_CONTROLLER 499 static inline void bond_netpoll_send_skb(const struct slave *slave, 500 struct sk_buff *skb) 501 { 502 struct netpoll *np = slave->np; 503 504 if (np) 505 netpoll_send_skb(np, skb); 506 } 507 #else 508 static inline void bond_netpoll_send_skb(const struct slave *slave, 509 struct sk_buff *skb) 510 { 511 } 512 #endif 513 514 static inline void bond_set_slave_inactive_flags(struct slave *slave, 515 bool notify) 516 { 517 if (!bond_is_lb(slave->bond)) 518 bond_set_slave_state(slave, BOND_STATE_BACKUP, notify); 519 if (!slave->bond->params.all_slaves_active) 520 slave->inactive = 1; 521 } 522 523 static inline void bond_set_slave_active_flags(struct slave *slave, 524 bool notify) 525 { 526 bond_set_slave_state(slave, BOND_STATE_ACTIVE, notify); 527 slave->inactive = 0; 528 } 529 530 static inline bool bond_is_slave_inactive(struct slave *slave) 531 { 532 return slave->inactive; 533 } 534 535 static inline void bond_propose_link_state(struct slave *slave, int state) 536 { 537 slave->link_new_state = state; 538 } 539 540 static inline void bond_commit_link_state(struct slave *slave, bool notify) 541 { 542 if (slave->link == slave->link_new_state) 543 return; 544 545 slave->link = slave->link_new_state; 546 if (notify) { 547 bond_queue_slave_event(slave); 548 bond_lower_state_changed(slave); 549 slave->should_notify_link = 0; 550 } else { 551 if (slave->should_notify_link) 552 slave->should_notify_link = 0; 553 else 554 slave->should_notify_link = 1; 555 } 556 } 557 558 static inline void bond_set_slave_link_state(struct slave *slave, int state, 559 bool notify) 560 { 561 bond_propose_link_state(slave, state); 562 bond_commit_link_state(slave, notify); 563 } 564 565 static inline void bond_slave_link_notify(struct bonding *bond) 566 { 567 struct list_head *iter; 568 struct slave *tmp; 569 570 bond_for_each_slave(bond, tmp, iter) { 571 if (tmp->should_notify_link) { 572 bond_queue_slave_event(tmp); 573 bond_lower_state_changed(tmp); 574 tmp->should_notify_link = 0; 575 } 576 } 577 } 578 579 static inline __be32 bond_confirm_addr(struct net_device *dev, __be32 dst, __be32 local) 580 { 581 struct in_device *in_dev; 582 __be32 addr = 0; 583 584 rcu_read_lock(); 585 in_dev = __in_dev_get_rcu(dev); 586 587 if (in_dev) 588 addr = inet_confirm_addr(dev_net(dev), in_dev, dst, local, 589 RT_SCOPE_HOST); 590 rcu_read_unlock(); 591 return addr; 592 } 593 594 struct bond_net { 595 struct net *net; /* Associated network namespace */ 596 struct list_head dev_list; 597 #ifdef CONFIG_PROC_FS 598 struct proc_dir_entry *proc_dir; 599 #endif 600 struct class_attribute class_attr_bonding_masters; 601 }; 602 603 int bond_arp_rcv(const struct sk_buff *skb, struct bonding *bond, struct slave *slave); 604 void bond_dev_queue_xmit(struct bonding *bond, struct sk_buff *skb, struct net_device *slave_dev); 605 int bond_create(struct net *net, const char *name); 606 int bond_create_sysfs(struct bond_net *net); 607 void bond_destroy_sysfs(struct bond_net *net); 608 void bond_prepare_sysfs_group(struct bonding *bond); 609 int bond_sysfs_slave_add(struct slave *slave); 610 void bond_sysfs_slave_del(struct slave *slave); 611 int bond_enslave(struct net_device *bond_dev, struct net_device *slave_dev, 612 struct netlink_ext_ack *extack); 613 int bond_release(struct net_device *bond_dev, struct net_device *slave_dev); 614 u32 bond_xmit_hash(struct bonding *bond, struct sk_buff *skb); 615 int bond_set_carrier(struct bonding *bond); 616 void bond_select_active_slave(struct bonding *bond); 617 void bond_change_active_slave(struct bonding *bond, struct slave *new_active); 618 void bond_create_debugfs(void); 619 void bond_destroy_debugfs(void); 620 void bond_debug_register(struct bonding *bond); 621 void bond_debug_unregister(struct bonding *bond); 622 void bond_debug_reregister(struct bonding *bond); 623 const char *bond_mode_name(int mode); 624 void bond_setup(struct net_device *bond_dev); 625 unsigned int bond_get_num_tx_queues(void); 626 int bond_netlink_init(void); 627 void bond_netlink_fini(void); 628 struct net_device *bond_option_active_slave_get_rcu(struct bonding *bond); 629 const char *bond_slave_link_status(s8 link); 630 struct bond_vlan_tag *bond_verify_device_path(struct net_device *start_dev, 631 struct net_device *end_dev, 632 int level); 633 int bond_update_slave_arr(struct bonding *bond, struct slave *skipslave); 634 void bond_slave_arr_work_rearm(struct bonding *bond, unsigned long delay); 635 void bond_work_init_all(struct bonding *bond); 636 637 #ifdef CONFIG_PROC_FS 638 void bond_create_proc_entry(struct bonding *bond); 639 void bond_remove_proc_entry(struct bonding *bond); 640 void bond_create_proc_dir(struct bond_net *bn); 641 void bond_destroy_proc_dir(struct bond_net *bn); 642 #else 643 static inline void bond_create_proc_entry(struct bonding *bond) 644 { 645 } 646 647 static inline void bond_remove_proc_entry(struct bonding *bond) 648 { 649 } 650 651 static inline void bond_create_proc_dir(struct bond_net *bn) 652 { 653 } 654 655 static inline void bond_destroy_proc_dir(struct bond_net *bn) 656 { 657 } 658 #endif 659 660 static inline struct slave *bond_slave_has_mac(struct bonding *bond, 661 const u8 *mac) 662 { 663 struct list_head *iter; 664 struct slave *tmp; 665 666 bond_for_each_slave(bond, tmp, iter) 667 if (ether_addr_equal_64bits(mac, tmp->dev->dev_addr)) 668 return tmp; 669 670 return NULL; 671 } 672 673 /* Caller must hold rcu_read_lock() for read */ 674 static inline struct slave *bond_slave_has_mac_rcu(struct bonding *bond, 675 const u8 *mac) 676 { 677 struct list_head *iter; 678 struct slave *tmp; 679 680 bond_for_each_slave_rcu(bond, tmp, iter) 681 if (ether_addr_equal_64bits(mac, tmp->dev->dev_addr)) 682 return tmp; 683 684 return NULL; 685 } 686 687 /* Caller must hold rcu_read_lock() for read */ 688 static inline bool bond_slave_has_mac_rx(struct bonding *bond, const u8 *mac) 689 { 690 struct list_head *iter; 691 struct slave *tmp; 692 struct netdev_hw_addr *ha; 693 694 bond_for_each_slave_rcu(bond, tmp, iter) 695 if (ether_addr_equal_64bits(mac, tmp->dev->dev_addr)) 696 return true; 697 698 if (netdev_uc_empty(bond->dev)) 699 return false; 700 701 netdev_for_each_uc_addr(ha, bond->dev) 702 if (ether_addr_equal_64bits(mac, ha->addr)) 703 return true; 704 705 return false; 706 } 707 708 /* Check if the ip is present in arp ip list, or first free slot if ip == 0 709 * Returns -1 if not found, index if found 710 */ 711 static inline int bond_get_targets_ip(__be32 *targets, __be32 ip) 712 { 713 int i; 714 715 for (i = 0; i < BOND_MAX_ARP_TARGETS; i++) 716 if (targets[i] == ip) 717 return i; 718 else if (targets[i] == 0) 719 break; 720 721 return -1; 722 } 723 724 /* exported from bond_main.c */ 725 extern unsigned int bond_net_id; 726 extern const struct bond_parm_tbl bond_lacp_tbl[]; 727 extern const struct bond_parm_tbl xmit_hashtype_tbl[]; 728 extern const struct bond_parm_tbl arp_validate_tbl[]; 729 extern const struct bond_parm_tbl arp_all_targets_tbl[]; 730 extern const struct bond_parm_tbl fail_over_mac_tbl[]; 731 extern const struct bond_parm_tbl pri_reselect_tbl[]; 732 extern struct bond_parm_tbl ad_select_tbl[]; 733 734 /* exported from bond_netlink.c */ 735 extern struct rtnl_link_ops bond_link_ops; 736 737 static inline void bond_tx_drop(struct net_device *dev, struct sk_buff *skb) 738 { 739 atomic_long_inc(&dev->tx_dropped); 740 dev_kfree_skb_any(skb); 741 } 742 743 #endif /* _NET_BONDING_H */ 744