1 // SPDX-License-Identifier: GPL-2.0 2 /* Copyright (C) B.A.T.M.A.N. contributors: 3 * 4 * Linus Lüssing 5 */ 6 7 #include "multicast.h" 8 #include "main.h" 9 10 #include <linux/atomic.h> 11 #include <linux/bitops.h> 12 #include <linux/bug.h> 13 #include <linux/byteorder/generic.h> 14 #include <linux/errno.h> 15 #include <linux/etherdevice.h> 16 #include <linux/gfp.h> 17 #include <linux/icmpv6.h> 18 #include <linux/if_bridge.h> 19 #include <linux/if_ether.h> 20 #include <linux/igmp.h> 21 #include <linux/in.h> 22 #include <linux/in6.h> 23 #include <linux/inetdevice.h> 24 #include <linux/ip.h> 25 #include <linux/ipv6.h> 26 #include <linux/jiffies.h> 27 #include <linux/kernel.h> 28 #include <linux/kref.h> 29 #include <linux/list.h> 30 #include <linux/lockdep.h> 31 #include <linux/netdevice.h> 32 #include <linux/netlink.h> 33 #include <linux/printk.h> 34 #include <linux/rculist.h> 35 #include <linux/rcupdate.h> 36 #include <linux/skbuff.h> 37 #include <linux/slab.h> 38 #include <linux/spinlock.h> 39 #include <linux/stddef.h> 40 #include <linux/string.h> 41 #include <linux/types.h> 42 #include <linux/workqueue.h> 43 #include <net/addrconf.h> 44 #include <net/genetlink.h> 45 #include <net/if_inet6.h> 46 #include <net/ip.h> 47 #include <net/ipv6.h> 48 #include <net/netlink.h> 49 #include <net/sock.h> 50 #include <uapi/linux/batadv_packet.h> 51 #include <uapi/linux/batman_adv.h> 52 53 #include "bridge_loop_avoidance.h" 54 #include "hard-interface.h" 55 #include "hash.h" 56 #include "log.h" 57 #include "netlink.h" 58 #include "send.h" 59 #include "soft-interface.h" 60 #include "translation-table.h" 61 #include "tvlv.h" 62 63 static void batadv_mcast_mla_update(struct work_struct *work); 64 65 /** 66 * batadv_mcast_start_timer() - schedule the multicast periodic worker 67 * @bat_priv: the bat priv with all the soft interface information 68 */ 69 static void batadv_mcast_start_timer(struct batadv_priv *bat_priv) 70 { 71 queue_delayed_work(batadv_event_workqueue, &bat_priv->mcast.work, 72 msecs_to_jiffies(BATADV_MCAST_WORK_PERIOD)); 73 } 74 75 /** 76 * batadv_mcast_get_bridge() - get the bridge on top of the softif if it exists 77 * @soft_iface: netdev struct of the mesh interface 78 * 79 * If the given soft interface has a bridge on top then the refcount 80 * of the according net device is increased. 81 * 82 * Return: NULL if no such bridge exists. Otherwise the net device of the 83 * bridge. 84 */ 85 static struct net_device *batadv_mcast_get_bridge(struct net_device *soft_iface) 86 { 87 struct net_device *upper = soft_iface; 88 89 rcu_read_lock(); 90 do { 91 upper = netdev_master_upper_dev_get_rcu(upper); 92 } while (upper && !netif_is_bridge_master(upper)); 93 94 dev_hold(upper); 95 rcu_read_unlock(); 96 97 return upper; 98 } 99 100 /** 101 * batadv_mcast_mla_rtr_flags_softif_get_ipv4() - get mcast router flags from 102 * node for IPv4 103 * @dev: the interface to check 104 * 105 * Checks the presence of an IPv4 multicast router on this node. 106 * 107 * Caller needs to hold rcu read lock. 108 * 109 * Return: BATADV_NO_FLAGS if present, BATADV_MCAST_WANT_NO_RTR4 otherwise. 110 */ 111 static u8 batadv_mcast_mla_rtr_flags_softif_get_ipv4(struct net_device *dev) 112 { 113 struct in_device *in_dev = __in_dev_get_rcu(dev); 114 115 if (in_dev && IN_DEV_MFORWARD(in_dev)) 116 return BATADV_NO_FLAGS; 117 else 118 return BATADV_MCAST_WANT_NO_RTR4; 119 } 120 121 /** 122 * batadv_mcast_mla_rtr_flags_softif_get_ipv6() - get mcast router flags from 123 * node for IPv6 124 * @dev: the interface to check 125 * 126 * Checks the presence of an IPv6 multicast router on this node. 127 * 128 * Caller needs to hold rcu read lock. 129 * 130 * Return: BATADV_NO_FLAGS if present, BATADV_MCAST_WANT_NO_RTR6 otherwise. 131 */ 132 #if IS_ENABLED(CONFIG_IPV6_MROUTE) 133 static u8 batadv_mcast_mla_rtr_flags_softif_get_ipv6(struct net_device *dev) 134 { 135 struct inet6_dev *in6_dev = __in6_dev_get(dev); 136 137 if (in6_dev && in6_dev->cnf.mc_forwarding) 138 return BATADV_NO_FLAGS; 139 else 140 return BATADV_MCAST_WANT_NO_RTR6; 141 } 142 #else 143 static inline u8 144 batadv_mcast_mla_rtr_flags_softif_get_ipv6(struct net_device *dev) 145 { 146 return BATADV_MCAST_WANT_NO_RTR6; 147 } 148 #endif 149 150 /** 151 * batadv_mcast_mla_rtr_flags_softif_get() - get mcast router flags from node 152 * @bat_priv: the bat priv with all the soft interface information 153 * @bridge: bridge interface on top of the soft_iface if present, 154 * otherwise pass NULL 155 * 156 * Checks the presence of IPv4 and IPv6 multicast routers on this 157 * node. 158 * 159 * Return: 160 * BATADV_NO_FLAGS: Both an IPv4 and IPv6 multicast router is present 161 * BATADV_MCAST_WANT_NO_RTR4: No IPv4 multicast router is present 162 * BATADV_MCAST_WANT_NO_RTR6: No IPv6 multicast router is present 163 * The former two OR'd: no multicast router is present 164 */ 165 static u8 batadv_mcast_mla_rtr_flags_softif_get(struct batadv_priv *bat_priv, 166 struct net_device *bridge) 167 { 168 struct net_device *dev = bridge ? bridge : bat_priv->soft_iface; 169 u8 flags = BATADV_NO_FLAGS; 170 171 rcu_read_lock(); 172 173 flags |= batadv_mcast_mla_rtr_flags_softif_get_ipv4(dev); 174 flags |= batadv_mcast_mla_rtr_flags_softif_get_ipv6(dev); 175 176 rcu_read_unlock(); 177 178 return flags; 179 } 180 181 /** 182 * batadv_mcast_mla_rtr_flags_bridge_get() - get mcast router flags from bridge 183 * @bat_priv: the bat priv with all the soft interface information 184 * @bridge: bridge interface on top of the soft_iface if present, 185 * otherwise pass NULL 186 * 187 * Checks the presence of IPv4 and IPv6 multicast routers behind a bridge. 188 * 189 * Return: 190 * BATADV_NO_FLAGS: Both an IPv4 and IPv6 multicast router is present 191 * BATADV_MCAST_WANT_NO_RTR4: No IPv4 multicast router is present 192 * BATADV_MCAST_WANT_NO_RTR6: No IPv6 multicast router is present 193 * The former two OR'd: no multicast router is present 194 */ 195 static u8 batadv_mcast_mla_rtr_flags_bridge_get(struct batadv_priv *bat_priv, 196 struct net_device *bridge) 197 { 198 struct net_device *dev = bat_priv->soft_iface; 199 u8 flags = BATADV_NO_FLAGS; 200 201 if (!bridge) 202 return BATADV_MCAST_WANT_NO_RTR4 | BATADV_MCAST_WANT_NO_RTR6; 203 204 if (!br_multicast_has_router_adjacent(dev, ETH_P_IP)) 205 flags |= BATADV_MCAST_WANT_NO_RTR4; 206 if (!br_multicast_has_router_adjacent(dev, ETH_P_IPV6)) 207 flags |= BATADV_MCAST_WANT_NO_RTR6; 208 209 return flags; 210 } 211 212 /** 213 * batadv_mcast_mla_rtr_flags_get() - get multicast router flags 214 * @bat_priv: the bat priv with all the soft interface information 215 * @bridge: bridge interface on top of the soft_iface if present, 216 * otherwise pass NULL 217 * 218 * Checks the presence of IPv4 and IPv6 multicast routers on this 219 * node or behind its bridge. 220 * 221 * Return: 222 * BATADV_NO_FLAGS: Both an IPv4 and IPv6 multicast router is present 223 * BATADV_MCAST_WANT_NO_RTR4: No IPv4 multicast router is present 224 * BATADV_MCAST_WANT_NO_RTR6: No IPv6 multicast router is present 225 * The former two OR'd: no multicast router is present 226 */ 227 static u8 batadv_mcast_mla_rtr_flags_get(struct batadv_priv *bat_priv, 228 struct net_device *bridge) 229 { 230 u8 flags = BATADV_MCAST_WANT_NO_RTR4 | BATADV_MCAST_WANT_NO_RTR6; 231 232 flags &= batadv_mcast_mla_rtr_flags_softif_get(bat_priv, bridge); 233 flags &= batadv_mcast_mla_rtr_flags_bridge_get(bat_priv, bridge); 234 235 return flags; 236 } 237 238 /** 239 * batadv_mcast_mla_flags_get() - get the new multicast flags 240 * @bat_priv: the bat priv with all the soft interface information 241 * 242 * Return: A set of flags for the current/next TVLV, querier and 243 * bridge state. 244 */ 245 static struct batadv_mcast_mla_flags 246 batadv_mcast_mla_flags_get(struct batadv_priv *bat_priv) 247 { 248 struct net_device *dev = bat_priv->soft_iface; 249 struct batadv_mcast_querier_state *qr4, *qr6; 250 struct batadv_mcast_mla_flags mla_flags; 251 struct net_device *bridge; 252 253 bridge = batadv_mcast_get_bridge(dev); 254 255 memset(&mla_flags, 0, sizeof(mla_flags)); 256 mla_flags.enabled = 1; 257 mla_flags.tvlv_flags |= batadv_mcast_mla_rtr_flags_get(bat_priv, 258 bridge); 259 260 if (!bridge) 261 return mla_flags; 262 263 dev_put(bridge); 264 265 mla_flags.bridged = 1; 266 qr4 = &mla_flags.querier_ipv4; 267 qr6 = &mla_flags.querier_ipv6; 268 269 if (!IS_ENABLED(CONFIG_BRIDGE_IGMP_SNOOPING)) 270 pr_warn_once("No bridge IGMP snooping compiled - multicast optimizations disabled\n"); 271 272 qr4->exists = br_multicast_has_querier_anywhere(dev, ETH_P_IP); 273 qr4->shadowing = br_multicast_has_querier_adjacent(dev, ETH_P_IP); 274 275 qr6->exists = br_multicast_has_querier_anywhere(dev, ETH_P_IPV6); 276 qr6->shadowing = br_multicast_has_querier_adjacent(dev, ETH_P_IPV6); 277 278 mla_flags.tvlv_flags |= BATADV_MCAST_WANT_ALL_UNSNOOPABLES; 279 280 /* 1) If no querier exists at all, then multicast listeners on 281 * our local TT clients behind the bridge will keep silent. 282 * 2) If the selected querier is on one of our local TT clients, 283 * behind the bridge, then this querier might shadow multicast 284 * listeners on our local TT clients, behind this bridge. 285 * 286 * In both cases, we will signalize other batman nodes that 287 * we need all multicast traffic of the according protocol. 288 */ 289 if (!qr4->exists || qr4->shadowing) { 290 mla_flags.tvlv_flags |= BATADV_MCAST_WANT_ALL_IPV4; 291 mla_flags.tvlv_flags &= ~BATADV_MCAST_WANT_NO_RTR4; 292 } 293 294 if (!qr6->exists || qr6->shadowing) { 295 mla_flags.tvlv_flags |= BATADV_MCAST_WANT_ALL_IPV6; 296 mla_flags.tvlv_flags &= ~BATADV_MCAST_WANT_NO_RTR6; 297 } 298 299 return mla_flags; 300 } 301 302 /** 303 * batadv_mcast_mla_is_duplicate() - check whether an address is in a list 304 * @mcast_addr: the multicast address to check 305 * @mcast_list: the list with multicast addresses to search in 306 * 307 * Return: true if the given address is already in the given list. 308 * Otherwise returns false. 309 */ 310 static bool batadv_mcast_mla_is_duplicate(u8 *mcast_addr, 311 struct hlist_head *mcast_list) 312 { 313 struct batadv_hw_addr *mcast_entry; 314 315 hlist_for_each_entry(mcast_entry, mcast_list, list) 316 if (batadv_compare_eth(mcast_entry->addr, mcast_addr)) 317 return true; 318 319 return false; 320 } 321 322 /** 323 * batadv_mcast_mla_softif_get_ipv4() - get softif IPv4 multicast listeners 324 * @dev: the device to collect multicast addresses from 325 * @mcast_list: a list to put found addresses into 326 * @flags: flags indicating the new multicast state 327 * 328 * Collects multicast addresses of IPv4 multicast listeners residing 329 * on this kernel on the given soft interface, dev, in 330 * the given mcast_list. In general, multicast listeners provided by 331 * your multicast receiving applications run directly on this node. 332 * 333 * Return: -ENOMEM on memory allocation error or the number of 334 * items added to the mcast_list otherwise. 335 */ 336 static int 337 batadv_mcast_mla_softif_get_ipv4(struct net_device *dev, 338 struct hlist_head *mcast_list, 339 struct batadv_mcast_mla_flags *flags) 340 { 341 struct batadv_hw_addr *new; 342 struct in_device *in_dev; 343 u8 mcast_addr[ETH_ALEN]; 344 struct ip_mc_list *pmc; 345 int ret = 0; 346 347 if (flags->tvlv_flags & BATADV_MCAST_WANT_ALL_IPV4) 348 return 0; 349 350 rcu_read_lock(); 351 352 in_dev = __in_dev_get_rcu(dev); 353 if (!in_dev) { 354 rcu_read_unlock(); 355 return 0; 356 } 357 358 for (pmc = rcu_dereference(in_dev->mc_list); pmc; 359 pmc = rcu_dereference(pmc->next_rcu)) { 360 if (flags->tvlv_flags & BATADV_MCAST_WANT_ALL_UNSNOOPABLES && 361 ipv4_is_local_multicast(pmc->multiaddr)) 362 continue; 363 364 if (!(flags->tvlv_flags & BATADV_MCAST_WANT_NO_RTR4) && 365 !ipv4_is_local_multicast(pmc->multiaddr)) 366 continue; 367 368 ip_eth_mc_map(pmc->multiaddr, mcast_addr); 369 370 if (batadv_mcast_mla_is_duplicate(mcast_addr, mcast_list)) 371 continue; 372 373 new = kmalloc(sizeof(*new), GFP_ATOMIC); 374 if (!new) { 375 ret = -ENOMEM; 376 break; 377 } 378 379 ether_addr_copy(new->addr, mcast_addr); 380 hlist_add_head(&new->list, mcast_list); 381 ret++; 382 } 383 rcu_read_unlock(); 384 385 return ret; 386 } 387 388 /** 389 * batadv_mcast_mla_softif_get_ipv6() - get softif IPv6 multicast listeners 390 * @dev: the device to collect multicast addresses from 391 * @mcast_list: a list to put found addresses into 392 * @flags: flags indicating the new multicast state 393 * 394 * Collects multicast addresses of IPv6 multicast listeners residing 395 * on this kernel on the given soft interface, dev, in 396 * the given mcast_list. In general, multicast listeners provided by 397 * your multicast receiving applications run directly on this node. 398 * 399 * Return: -ENOMEM on memory allocation error or the number of 400 * items added to the mcast_list otherwise. 401 */ 402 #if IS_ENABLED(CONFIG_IPV6) 403 static int 404 batadv_mcast_mla_softif_get_ipv6(struct net_device *dev, 405 struct hlist_head *mcast_list, 406 struct batadv_mcast_mla_flags *flags) 407 { 408 struct batadv_hw_addr *new; 409 struct inet6_dev *in6_dev; 410 u8 mcast_addr[ETH_ALEN]; 411 struct ifmcaddr6 *pmc6; 412 int ret = 0; 413 414 if (flags->tvlv_flags & BATADV_MCAST_WANT_ALL_IPV6) 415 return 0; 416 417 rcu_read_lock(); 418 419 in6_dev = __in6_dev_get(dev); 420 if (!in6_dev) { 421 rcu_read_unlock(); 422 return 0; 423 } 424 425 for (pmc6 = rcu_dereference(in6_dev->mc_list); 426 pmc6; 427 pmc6 = rcu_dereference(pmc6->next)) { 428 if (IPV6_ADDR_MC_SCOPE(&pmc6->mca_addr) < 429 IPV6_ADDR_SCOPE_LINKLOCAL) 430 continue; 431 432 if (flags->tvlv_flags & BATADV_MCAST_WANT_ALL_UNSNOOPABLES && 433 ipv6_addr_is_ll_all_nodes(&pmc6->mca_addr)) 434 continue; 435 436 if (!(flags->tvlv_flags & BATADV_MCAST_WANT_NO_RTR6) && 437 IPV6_ADDR_MC_SCOPE(&pmc6->mca_addr) > 438 IPV6_ADDR_SCOPE_LINKLOCAL) 439 continue; 440 441 ipv6_eth_mc_map(&pmc6->mca_addr, mcast_addr); 442 443 if (batadv_mcast_mla_is_duplicate(mcast_addr, mcast_list)) 444 continue; 445 446 new = kmalloc(sizeof(*new), GFP_ATOMIC); 447 if (!new) { 448 ret = -ENOMEM; 449 break; 450 } 451 452 ether_addr_copy(new->addr, mcast_addr); 453 hlist_add_head(&new->list, mcast_list); 454 ret++; 455 } 456 rcu_read_unlock(); 457 458 return ret; 459 } 460 #else 461 static inline int 462 batadv_mcast_mla_softif_get_ipv6(struct net_device *dev, 463 struct hlist_head *mcast_list, 464 struct batadv_mcast_mla_flags *flags) 465 { 466 return 0; 467 } 468 #endif 469 470 /** 471 * batadv_mcast_mla_softif_get() - get softif multicast listeners 472 * @dev: the device to collect multicast addresses from 473 * @mcast_list: a list to put found addresses into 474 * @flags: flags indicating the new multicast state 475 * 476 * Collects multicast addresses of multicast listeners residing 477 * on this kernel on the given soft interface, dev, in 478 * the given mcast_list. In general, multicast listeners provided by 479 * your multicast receiving applications run directly on this node. 480 * 481 * If there is a bridge interface on top of dev, collect from that one 482 * instead. Just like with IP addresses and routes, multicast listeners 483 * will(/should) register to the bridge interface instead of an 484 * enslaved bat0. 485 * 486 * Return: -ENOMEM on memory allocation error or the number of 487 * items added to the mcast_list otherwise. 488 */ 489 static int 490 batadv_mcast_mla_softif_get(struct net_device *dev, 491 struct hlist_head *mcast_list, 492 struct batadv_mcast_mla_flags *flags) 493 { 494 struct net_device *bridge = batadv_mcast_get_bridge(dev); 495 int ret4, ret6 = 0; 496 497 if (bridge) 498 dev = bridge; 499 500 ret4 = batadv_mcast_mla_softif_get_ipv4(dev, mcast_list, flags); 501 if (ret4 < 0) 502 goto out; 503 504 ret6 = batadv_mcast_mla_softif_get_ipv6(dev, mcast_list, flags); 505 if (ret6 < 0) { 506 ret4 = 0; 507 goto out; 508 } 509 510 out: 511 dev_put(bridge); 512 513 return ret4 + ret6; 514 } 515 516 /** 517 * batadv_mcast_mla_br_addr_cpy() - copy a bridge multicast address 518 * @dst: destination to write to - a multicast MAC address 519 * @src: source to read from - a multicast IP address 520 * 521 * Converts a given multicast IPv4/IPv6 address from a bridge 522 * to its matching multicast MAC address and copies it into the given 523 * destination buffer. 524 * 525 * Caller needs to make sure the destination buffer can hold 526 * at least ETH_ALEN bytes. 527 */ 528 static void batadv_mcast_mla_br_addr_cpy(char *dst, const struct br_ip *src) 529 { 530 if (src->proto == htons(ETH_P_IP)) 531 ip_eth_mc_map(src->dst.ip4, dst); 532 #if IS_ENABLED(CONFIG_IPV6) 533 else if (src->proto == htons(ETH_P_IPV6)) 534 ipv6_eth_mc_map(&src->dst.ip6, dst); 535 #endif 536 else 537 eth_zero_addr(dst); 538 } 539 540 /** 541 * batadv_mcast_mla_bridge_get() - get bridged-in multicast listeners 542 * @dev: a bridge slave whose bridge to collect multicast addresses from 543 * @mcast_list: a list to put found addresses into 544 * @flags: flags indicating the new multicast state 545 * 546 * Collects multicast addresses of multicast listeners residing 547 * on foreign, non-mesh devices which we gave access to our mesh via 548 * a bridge on top of the given soft interface, dev, in the given 549 * mcast_list. 550 * 551 * Return: -ENOMEM on memory allocation error or the number of 552 * items added to the mcast_list otherwise. 553 */ 554 static int batadv_mcast_mla_bridge_get(struct net_device *dev, 555 struct hlist_head *mcast_list, 556 struct batadv_mcast_mla_flags *flags) 557 { 558 struct list_head bridge_mcast_list = LIST_HEAD_INIT(bridge_mcast_list); 559 struct br_ip_list *br_ip_entry, *tmp; 560 u8 tvlv_flags = flags->tvlv_flags; 561 struct batadv_hw_addr *new; 562 u8 mcast_addr[ETH_ALEN]; 563 int ret; 564 565 /* we don't need to detect these devices/listeners, the IGMP/MLD 566 * snooping code of the Linux bridge already does that for us 567 */ 568 ret = br_multicast_list_adjacent(dev, &bridge_mcast_list); 569 if (ret < 0) 570 goto out; 571 572 list_for_each_entry(br_ip_entry, &bridge_mcast_list, list) { 573 if (br_ip_entry->addr.proto == htons(ETH_P_IP)) { 574 if (tvlv_flags & BATADV_MCAST_WANT_ALL_IPV4) 575 continue; 576 577 if (tvlv_flags & BATADV_MCAST_WANT_ALL_UNSNOOPABLES && 578 ipv4_is_local_multicast(br_ip_entry->addr.dst.ip4)) 579 continue; 580 581 if (!(tvlv_flags & BATADV_MCAST_WANT_NO_RTR4) && 582 !ipv4_is_local_multicast(br_ip_entry->addr.dst.ip4)) 583 continue; 584 } 585 586 #if IS_ENABLED(CONFIG_IPV6) 587 if (br_ip_entry->addr.proto == htons(ETH_P_IPV6)) { 588 if (tvlv_flags & BATADV_MCAST_WANT_ALL_IPV6) 589 continue; 590 591 if (tvlv_flags & BATADV_MCAST_WANT_ALL_UNSNOOPABLES && 592 ipv6_addr_is_ll_all_nodes(&br_ip_entry->addr.dst.ip6)) 593 continue; 594 595 if (!(tvlv_flags & BATADV_MCAST_WANT_NO_RTR6) && 596 IPV6_ADDR_MC_SCOPE(&br_ip_entry->addr.dst.ip6) > 597 IPV6_ADDR_SCOPE_LINKLOCAL) 598 continue; 599 } 600 #endif 601 602 batadv_mcast_mla_br_addr_cpy(mcast_addr, &br_ip_entry->addr); 603 if (batadv_mcast_mla_is_duplicate(mcast_addr, mcast_list)) 604 continue; 605 606 new = kmalloc(sizeof(*new), GFP_ATOMIC); 607 if (!new) { 608 ret = -ENOMEM; 609 break; 610 } 611 612 ether_addr_copy(new->addr, mcast_addr); 613 hlist_add_head(&new->list, mcast_list); 614 } 615 616 out: 617 list_for_each_entry_safe(br_ip_entry, tmp, &bridge_mcast_list, list) { 618 list_del(&br_ip_entry->list); 619 kfree(br_ip_entry); 620 } 621 622 return ret; 623 } 624 625 /** 626 * batadv_mcast_mla_list_free() - free a list of multicast addresses 627 * @mcast_list: the list to free 628 * 629 * Removes and frees all items in the given mcast_list. 630 */ 631 static void batadv_mcast_mla_list_free(struct hlist_head *mcast_list) 632 { 633 struct batadv_hw_addr *mcast_entry; 634 struct hlist_node *tmp; 635 636 hlist_for_each_entry_safe(mcast_entry, tmp, mcast_list, list) { 637 hlist_del(&mcast_entry->list); 638 kfree(mcast_entry); 639 } 640 } 641 642 /** 643 * batadv_mcast_mla_tt_retract() - clean up multicast listener announcements 644 * @bat_priv: the bat priv with all the soft interface information 645 * @mcast_list: a list of addresses which should _not_ be removed 646 * 647 * Retracts the announcement of any multicast listener from the 648 * translation table except the ones listed in the given mcast_list. 649 * 650 * If mcast_list is NULL then all are retracted. 651 */ 652 static void batadv_mcast_mla_tt_retract(struct batadv_priv *bat_priv, 653 struct hlist_head *mcast_list) 654 { 655 struct batadv_hw_addr *mcast_entry; 656 struct hlist_node *tmp; 657 658 hlist_for_each_entry_safe(mcast_entry, tmp, &bat_priv->mcast.mla_list, 659 list) { 660 if (mcast_list && 661 batadv_mcast_mla_is_duplicate(mcast_entry->addr, 662 mcast_list)) 663 continue; 664 665 batadv_tt_local_remove(bat_priv, mcast_entry->addr, 666 BATADV_NO_FLAGS, 667 "mcast TT outdated", false); 668 669 hlist_del(&mcast_entry->list); 670 kfree(mcast_entry); 671 } 672 } 673 674 /** 675 * batadv_mcast_mla_tt_add() - add multicast listener announcements 676 * @bat_priv: the bat priv with all the soft interface information 677 * @mcast_list: a list of addresses which are going to get added 678 * 679 * Adds multicast listener announcements from the given mcast_list to the 680 * translation table if they have not been added yet. 681 */ 682 static void batadv_mcast_mla_tt_add(struct batadv_priv *bat_priv, 683 struct hlist_head *mcast_list) 684 { 685 struct batadv_hw_addr *mcast_entry; 686 struct hlist_node *tmp; 687 688 if (!mcast_list) 689 return; 690 691 hlist_for_each_entry_safe(mcast_entry, tmp, mcast_list, list) { 692 if (batadv_mcast_mla_is_duplicate(mcast_entry->addr, 693 &bat_priv->mcast.mla_list)) 694 continue; 695 696 if (!batadv_tt_local_add(bat_priv->soft_iface, 697 mcast_entry->addr, BATADV_NO_FLAGS, 698 BATADV_NULL_IFINDEX, BATADV_NO_MARK)) 699 continue; 700 701 hlist_del(&mcast_entry->list); 702 hlist_add_head(&mcast_entry->list, &bat_priv->mcast.mla_list); 703 } 704 } 705 706 /** 707 * batadv_mcast_querier_log() - debug output regarding the querier status on 708 * link 709 * @bat_priv: the bat priv with all the soft interface information 710 * @str_proto: a string for the querier protocol (e.g. "IGMP" or "MLD") 711 * @old_state: the previous querier state on our link 712 * @new_state: the new querier state on our link 713 * 714 * Outputs debug messages to the logging facility with log level 'mcast' 715 * regarding changes to the querier status on the link which are relevant 716 * to our multicast optimizations. 717 * 718 * Usually this is about whether a querier appeared or vanished in 719 * our mesh or whether the querier is in the suboptimal position of being 720 * behind our local bridge segment: Snooping switches will directly 721 * forward listener reports to the querier, therefore batman-adv and 722 * the bridge will potentially not see these listeners - the querier is 723 * potentially shadowing listeners from us then. 724 * 725 * This is only interesting for nodes with a bridge on top of their 726 * soft interface. 727 */ 728 static void 729 batadv_mcast_querier_log(struct batadv_priv *bat_priv, char *str_proto, 730 struct batadv_mcast_querier_state *old_state, 731 struct batadv_mcast_querier_state *new_state) 732 { 733 if (!old_state->exists && new_state->exists) 734 batadv_info(bat_priv->soft_iface, "%s Querier appeared\n", 735 str_proto); 736 else if (old_state->exists && !new_state->exists) 737 batadv_info(bat_priv->soft_iface, 738 "%s Querier disappeared - multicast optimizations disabled\n", 739 str_proto); 740 else if (!bat_priv->mcast.mla_flags.bridged && !new_state->exists) 741 batadv_info(bat_priv->soft_iface, 742 "No %s Querier present - multicast optimizations disabled\n", 743 str_proto); 744 745 if (new_state->exists) { 746 if ((!old_state->shadowing && new_state->shadowing) || 747 (!old_state->exists && new_state->shadowing)) 748 batadv_dbg(BATADV_DBG_MCAST, bat_priv, 749 "%s Querier is behind our bridged segment: Might shadow listeners\n", 750 str_proto); 751 else if (old_state->shadowing && !new_state->shadowing) 752 batadv_dbg(BATADV_DBG_MCAST, bat_priv, 753 "%s Querier is not behind our bridged segment\n", 754 str_proto); 755 } 756 } 757 758 /** 759 * batadv_mcast_bridge_log() - debug output for topology changes in bridged 760 * setups 761 * @bat_priv: the bat priv with all the soft interface information 762 * @new_flags: flags indicating the new multicast state 763 * 764 * If no bridges are ever used on this node, then this function does nothing. 765 * 766 * Otherwise this function outputs debug information to the 'mcast' log level 767 * which might be relevant to our multicast optimizations. 768 * 769 * More precisely, it outputs information when a bridge interface is added or 770 * removed from a soft interface. And when a bridge is present, it further 771 * outputs information about the querier state which is relevant for the 772 * multicast flags this node is going to set. 773 */ 774 static void 775 batadv_mcast_bridge_log(struct batadv_priv *bat_priv, 776 struct batadv_mcast_mla_flags *new_flags) 777 { 778 struct batadv_mcast_mla_flags *old_flags = &bat_priv->mcast.mla_flags; 779 780 if (!old_flags->bridged && new_flags->bridged) 781 batadv_dbg(BATADV_DBG_MCAST, bat_priv, 782 "Bridge added: Setting Unsnoopables(U)-flag\n"); 783 else if (old_flags->bridged && !new_flags->bridged) 784 batadv_dbg(BATADV_DBG_MCAST, bat_priv, 785 "Bridge removed: Unsetting Unsnoopables(U)-flag\n"); 786 787 if (new_flags->bridged) { 788 batadv_mcast_querier_log(bat_priv, "IGMP", 789 &old_flags->querier_ipv4, 790 &new_flags->querier_ipv4); 791 batadv_mcast_querier_log(bat_priv, "MLD", 792 &old_flags->querier_ipv6, 793 &new_flags->querier_ipv6); 794 } 795 } 796 797 /** 798 * batadv_mcast_flags_log() - output debug information about mcast flag changes 799 * @bat_priv: the bat priv with all the soft interface information 800 * @flags: TVLV flags indicating the new multicast state 801 * 802 * Whenever the multicast TVLV flags this node announces change, this function 803 * should be used to notify userspace about the change. 804 */ 805 static void batadv_mcast_flags_log(struct batadv_priv *bat_priv, u8 flags) 806 { 807 bool old_enabled = bat_priv->mcast.mla_flags.enabled; 808 u8 old_flags = bat_priv->mcast.mla_flags.tvlv_flags; 809 char str_old_flags[] = "[.... . ]"; 810 811 sprintf(str_old_flags, "[%c%c%c%s%s]", 812 (old_flags & BATADV_MCAST_WANT_ALL_UNSNOOPABLES) ? 'U' : '.', 813 (old_flags & BATADV_MCAST_WANT_ALL_IPV4) ? '4' : '.', 814 (old_flags & BATADV_MCAST_WANT_ALL_IPV6) ? '6' : '.', 815 !(old_flags & BATADV_MCAST_WANT_NO_RTR4) ? "R4" : ". ", 816 !(old_flags & BATADV_MCAST_WANT_NO_RTR6) ? "R6" : ". "); 817 818 batadv_dbg(BATADV_DBG_MCAST, bat_priv, 819 "Changing multicast flags from '%s' to '[%c%c%c%s%s]'\n", 820 old_enabled ? str_old_flags : "<undefined>", 821 (flags & BATADV_MCAST_WANT_ALL_UNSNOOPABLES) ? 'U' : '.', 822 (flags & BATADV_MCAST_WANT_ALL_IPV4) ? '4' : '.', 823 (flags & BATADV_MCAST_WANT_ALL_IPV6) ? '6' : '.', 824 !(flags & BATADV_MCAST_WANT_NO_RTR4) ? "R4" : ". ", 825 !(flags & BATADV_MCAST_WANT_NO_RTR6) ? "R6" : ". "); 826 } 827 828 /** 829 * batadv_mcast_mla_flags_update() - update multicast flags 830 * @bat_priv: the bat priv with all the soft interface information 831 * @flags: flags indicating the new multicast state 832 * 833 * Updates the own multicast tvlv with our current multicast related settings, 834 * capabilities and inabilities. 835 */ 836 static void 837 batadv_mcast_mla_flags_update(struct batadv_priv *bat_priv, 838 struct batadv_mcast_mla_flags *flags) 839 { 840 struct batadv_tvlv_mcast_data mcast_data; 841 842 if (!memcmp(flags, &bat_priv->mcast.mla_flags, sizeof(*flags))) 843 return; 844 845 batadv_mcast_bridge_log(bat_priv, flags); 846 batadv_mcast_flags_log(bat_priv, flags->tvlv_flags); 847 848 mcast_data.flags = flags->tvlv_flags; 849 memset(mcast_data.reserved, 0, sizeof(mcast_data.reserved)); 850 851 batadv_tvlv_container_register(bat_priv, BATADV_TVLV_MCAST, 2, 852 &mcast_data, sizeof(mcast_data)); 853 854 bat_priv->mcast.mla_flags = *flags; 855 } 856 857 /** 858 * __batadv_mcast_mla_update() - update the own MLAs 859 * @bat_priv: the bat priv with all the soft interface information 860 * 861 * Updates the own multicast listener announcements in the translation 862 * table as well as the own, announced multicast tvlv container. 863 * 864 * Note that non-conflicting reads and writes to bat_priv->mcast.mla_list 865 * in batadv_mcast_mla_tt_retract() and batadv_mcast_mla_tt_add() are 866 * ensured by the non-parallel execution of the worker this function 867 * belongs to. 868 */ 869 static void __batadv_mcast_mla_update(struct batadv_priv *bat_priv) 870 { 871 struct net_device *soft_iface = bat_priv->soft_iface; 872 struct hlist_head mcast_list = HLIST_HEAD_INIT; 873 struct batadv_mcast_mla_flags flags; 874 int ret; 875 876 flags = batadv_mcast_mla_flags_get(bat_priv); 877 878 ret = batadv_mcast_mla_softif_get(soft_iface, &mcast_list, &flags); 879 if (ret < 0) 880 goto out; 881 882 ret = batadv_mcast_mla_bridge_get(soft_iface, &mcast_list, &flags); 883 if (ret < 0) 884 goto out; 885 886 spin_lock(&bat_priv->mcast.mla_lock); 887 batadv_mcast_mla_tt_retract(bat_priv, &mcast_list); 888 batadv_mcast_mla_tt_add(bat_priv, &mcast_list); 889 batadv_mcast_mla_flags_update(bat_priv, &flags); 890 spin_unlock(&bat_priv->mcast.mla_lock); 891 892 out: 893 batadv_mcast_mla_list_free(&mcast_list); 894 } 895 896 /** 897 * batadv_mcast_mla_update() - update the own MLAs 898 * @work: kernel work struct 899 * 900 * Updates the own multicast listener announcements in the translation 901 * table as well as the own, announced multicast tvlv container. 902 * 903 * In the end, reschedules the work timer. 904 */ 905 static void batadv_mcast_mla_update(struct work_struct *work) 906 { 907 struct delayed_work *delayed_work; 908 struct batadv_priv_mcast *priv_mcast; 909 struct batadv_priv *bat_priv; 910 911 delayed_work = to_delayed_work(work); 912 priv_mcast = container_of(delayed_work, struct batadv_priv_mcast, work); 913 bat_priv = container_of(priv_mcast, struct batadv_priv, mcast); 914 915 __batadv_mcast_mla_update(bat_priv); 916 batadv_mcast_start_timer(bat_priv); 917 } 918 919 /** 920 * batadv_mcast_is_report_ipv4() - check for IGMP reports 921 * @skb: the ethernet frame destined for the mesh 922 * 923 * This call might reallocate skb data. 924 * 925 * Checks whether the given frame is a valid IGMP report. 926 * 927 * Return: If so then true, otherwise false. 928 */ 929 static bool batadv_mcast_is_report_ipv4(struct sk_buff *skb) 930 { 931 if (ip_mc_check_igmp(skb) < 0) 932 return false; 933 934 switch (igmp_hdr(skb)->type) { 935 case IGMP_HOST_MEMBERSHIP_REPORT: 936 case IGMPV2_HOST_MEMBERSHIP_REPORT: 937 case IGMPV3_HOST_MEMBERSHIP_REPORT: 938 return true; 939 } 940 941 return false; 942 } 943 944 /** 945 * batadv_mcast_forw_mode_check_ipv4() - check for optimized forwarding 946 * potential 947 * @bat_priv: the bat priv with all the soft interface information 948 * @skb: the IPv4 packet to check 949 * @is_unsnoopable: stores whether the destination is snoopable 950 * @is_routable: stores whether the destination is routable 951 * 952 * Checks whether the given IPv4 packet has the potential to be forwarded with a 953 * mode more optimal than classic flooding. 954 * 955 * Return: If so then 0. Otherwise -EINVAL or -ENOMEM in case of memory 956 * allocation failure. 957 */ 958 static int batadv_mcast_forw_mode_check_ipv4(struct batadv_priv *bat_priv, 959 struct sk_buff *skb, 960 bool *is_unsnoopable, 961 int *is_routable) 962 { 963 struct iphdr *iphdr; 964 965 /* We might fail due to out-of-memory -> drop it */ 966 if (!pskb_may_pull(skb, sizeof(struct ethhdr) + sizeof(*iphdr))) 967 return -ENOMEM; 968 969 if (batadv_mcast_is_report_ipv4(skb)) 970 return -EINVAL; 971 972 iphdr = ip_hdr(skb); 973 974 /* link-local multicast listeners behind a bridge are 975 * not snoopable (see RFC4541, section 2.1.2.2) 976 */ 977 if (ipv4_is_local_multicast(iphdr->daddr)) 978 *is_unsnoopable = true; 979 else 980 *is_routable = ETH_P_IP; 981 982 return 0; 983 } 984 985 /** 986 * batadv_mcast_is_report_ipv6() - check for MLD reports 987 * @skb: the ethernet frame destined for the mesh 988 * 989 * This call might reallocate skb data. 990 * 991 * Checks whether the given frame is a valid MLD report. 992 * 993 * Return: If so then true, otherwise false. 994 */ 995 static bool batadv_mcast_is_report_ipv6(struct sk_buff *skb) 996 { 997 if (ipv6_mc_check_mld(skb) < 0) 998 return false; 999 1000 switch (icmp6_hdr(skb)->icmp6_type) { 1001 case ICMPV6_MGM_REPORT: 1002 case ICMPV6_MLD2_REPORT: 1003 return true; 1004 } 1005 1006 return false; 1007 } 1008 1009 /** 1010 * batadv_mcast_forw_mode_check_ipv6() - check for optimized forwarding 1011 * potential 1012 * @bat_priv: the bat priv with all the soft interface information 1013 * @skb: the IPv6 packet to check 1014 * @is_unsnoopable: stores whether the destination is snoopable 1015 * @is_routable: stores whether the destination is routable 1016 * 1017 * Checks whether the given IPv6 packet has the potential to be forwarded with a 1018 * mode more optimal than classic flooding. 1019 * 1020 * Return: If so then 0. Otherwise -EINVAL is or -ENOMEM if we are out of memory 1021 */ 1022 static int batadv_mcast_forw_mode_check_ipv6(struct batadv_priv *bat_priv, 1023 struct sk_buff *skb, 1024 bool *is_unsnoopable, 1025 int *is_routable) 1026 { 1027 struct ipv6hdr *ip6hdr; 1028 1029 /* We might fail due to out-of-memory -> drop it */ 1030 if (!pskb_may_pull(skb, sizeof(struct ethhdr) + sizeof(*ip6hdr))) 1031 return -ENOMEM; 1032 1033 if (batadv_mcast_is_report_ipv6(skb)) 1034 return -EINVAL; 1035 1036 ip6hdr = ipv6_hdr(skb); 1037 1038 if (IPV6_ADDR_MC_SCOPE(&ip6hdr->daddr) < IPV6_ADDR_SCOPE_LINKLOCAL) 1039 return -EINVAL; 1040 1041 /* link-local-all-nodes multicast listeners behind a bridge are 1042 * not snoopable (see RFC4541, section 3, paragraph 3) 1043 */ 1044 if (ipv6_addr_is_ll_all_nodes(&ip6hdr->daddr)) 1045 *is_unsnoopable = true; 1046 else if (IPV6_ADDR_MC_SCOPE(&ip6hdr->daddr) > IPV6_ADDR_SCOPE_LINKLOCAL) 1047 *is_routable = ETH_P_IPV6; 1048 1049 return 0; 1050 } 1051 1052 /** 1053 * batadv_mcast_forw_mode_check() - check for optimized forwarding potential 1054 * @bat_priv: the bat priv with all the soft interface information 1055 * @skb: the multicast frame to check 1056 * @is_unsnoopable: stores whether the destination is snoopable 1057 * @is_routable: stores whether the destination is routable 1058 * 1059 * Checks whether the given multicast ethernet frame has the potential to be 1060 * forwarded with a mode more optimal than classic flooding. 1061 * 1062 * Return: If so then 0. Otherwise -EINVAL is or -ENOMEM if we are out of memory 1063 */ 1064 static int batadv_mcast_forw_mode_check(struct batadv_priv *bat_priv, 1065 struct sk_buff *skb, 1066 bool *is_unsnoopable, 1067 int *is_routable) 1068 { 1069 struct ethhdr *ethhdr = eth_hdr(skb); 1070 1071 if (!atomic_read(&bat_priv->multicast_mode)) 1072 return -EINVAL; 1073 1074 switch (ntohs(ethhdr->h_proto)) { 1075 case ETH_P_IP: 1076 return batadv_mcast_forw_mode_check_ipv4(bat_priv, skb, 1077 is_unsnoopable, 1078 is_routable); 1079 case ETH_P_IPV6: 1080 if (!IS_ENABLED(CONFIG_IPV6)) 1081 return -EINVAL; 1082 1083 return batadv_mcast_forw_mode_check_ipv6(bat_priv, skb, 1084 is_unsnoopable, 1085 is_routable); 1086 default: 1087 return -EINVAL; 1088 } 1089 } 1090 1091 /** 1092 * batadv_mcast_forw_want_all_ip_count() - count nodes with unspecific mcast 1093 * interest 1094 * @bat_priv: the bat priv with all the soft interface information 1095 * @ethhdr: ethernet header of a packet 1096 * 1097 * Return: the number of nodes which want all IPv4 multicast traffic if the 1098 * given ethhdr is from an IPv4 packet or the number of nodes which want all 1099 * IPv6 traffic if it matches an IPv6 packet. 1100 */ 1101 static int batadv_mcast_forw_want_all_ip_count(struct batadv_priv *bat_priv, 1102 struct ethhdr *ethhdr) 1103 { 1104 switch (ntohs(ethhdr->h_proto)) { 1105 case ETH_P_IP: 1106 return atomic_read(&bat_priv->mcast.num_want_all_ipv4); 1107 case ETH_P_IPV6: 1108 return atomic_read(&bat_priv->mcast.num_want_all_ipv6); 1109 default: 1110 /* we shouldn't be here... */ 1111 return 0; 1112 } 1113 } 1114 1115 /** 1116 * batadv_mcast_forw_rtr_count() - count nodes with a multicast router 1117 * @bat_priv: the bat priv with all the soft interface information 1118 * @protocol: the ethernet protocol type to count multicast routers for 1119 * 1120 * Return: the number of nodes which want all routable IPv4 multicast traffic 1121 * if the protocol is ETH_P_IP or the number of nodes which want all routable 1122 * IPv6 traffic if the protocol is ETH_P_IPV6. Otherwise returns 0. 1123 */ 1124 1125 static int batadv_mcast_forw_rtr_count(struct batadv_priv *bat_priv, 1126 int protocol) 1127 { 1128 switch (protocol) { 1129 case ETH_P_IP: 1130 return atomic_read(&bat_priv->mcast.num_want_all_rtr4); 1131 case ETH_P_IPV6: 1132 return atomic_read(&bat_priv->mcast.num_want_all_rtr6); 1133 default: 1134 return 0; 1135 } 1136 } 1137 1138 /** 1139 * batadv_mcast_forw_tt_node_get() - get a multicast tt node 1140 * @bat_priv: the bat priv with all the soft interface information 1141 * @ethhdr: the ether header containing the multicast destination 1142 * 1143 * Return: an orig_node matching the multicast address provided by ethhdr 1144 * via a translation table lookup. This increases the returned nodes refcount. 1145 */ 1146 static struct batadv_orig_node * 1147 batadv_mcast_forw_tt_node_get(struct batadv_priv *bat_priv, 1148 struct ethhdr *ethhdr) 1149 { 1150 return batadv_transtable_search(bat_priv, NULL, ethhdr->h_dest, 1151 BATADV_NO_FLAGS); 1152 } 1153 1154 /** 1155 * batadv_mcast_forw_ipv4_node_get() - get a node with an ipv4 flag 1156 * @bat_priv: the bat priv with all the soft interface information 1157 * 1158 * Return: an orig_node which has the BATADV_MCAST_WANT_ALL_IPV4 flag set and 1159 * increases its refcount. 1160 */ 1161 static struct batadv_orig_node * 1162 batadv_mcast_forw_ipv4_node_get(struct batadv_priv *bat_priv) 1163 { 1164 struct batadv_orig_node *tmp_orig_node, *orig_node = NULL; 1165 1166 rcu_read_lock(); 1167 hlist_for_each_entry_rcu(tmp_orig_node, 1168 &bat_priv->mcast.want_all_ipv4_list, 1169 mcast_want_all_ipv4_node) { 1170 if (!kref_get_unless_zero(&tmp_orig_node->refcount)) 1171 continue; 1172 1173 orig_node = tmp_orig_node; 1174 break; 1175 } 1176 rcu_read_unlock(); 1177 1178 return orig_node; 1179 } 1180 1181 /** 1182 * batadv_mcast_forw_ipv6_node_get() - get a node with an ipv6 flag 1183 * @bat_priv: the bat priv with all the soft interface information 1184 * 1185 * Return: an orig_node which has the BATADV_MCAST_WANT_ALL_IPV6 flag set 1186 * and increases its refcount. 1187 */ 1188 static struct batadv_orig_node * 1189 batadv_mcast_forw_ipv6_node_get(struct batadv_priv *bat_priv) 1190 { 1191 struct batadv_orig_node *tmp_orig_node, *orig_node = NULL; 1192 1193 rcu_read_lock(); 1194 hlist_for_each_entry_rcu(tmp_orig_node, 1195 &bat_priv->mcast.want_all_ipv6_list, 1196 mcast_want_all_ipv6_node) { 1197 if (!kref_get_unless_zero(&tmp_orig_node->refcount)) 1198 continue; 1199 1200 orig_node = tmp_orig_node; 1201 break; 1202 } 1203 rcu_read_unlock(); 1204 1205 return orig_node; 1206 } 1207 1208 /** 1209 * batadv_mcast_forw_ip_node_get() - get a node with an ipv4/ipv6 flag 1210 * @bat_priv: the bat priv with all the soft interface information 1211 * @ethhdr: an ethernet header to determine the protocol family from 1212 * 1213 * Return: an orig_node which has the BATADV_MCAST_WANT_ALL_IPV4 or 1214 * BATADV_MCAST_WANT_ALL_IPV6 flag, depending on the provided ethhdr, sets and 1215 * increases its refcount. 1216 */ 1217 static struct batadv_orig_node * 1218 batadv_mcast_forw_ip_node_get(struct batadv_priv *bat_priv, 1219 struct ethhdr *ethhdr) 1220 { 1221 switch (ntohs(ethhdr->h_proto)) { 1222 case ETH_P_IP: 1223 return batadv_mcast_forw_ipv4_node_get(bat_priv); 1224 case ETH_P_IPV6: 1225 return batadv_mcast_forw_ipv6_node_get(bat_priv); 1226 default: 1227 /* we shouldn't be here... */ 1228 return NULL; 1229 } 1230 } 1231 1232 /** 1233 * batadv_mcast_forw_unsnoop_node_get() - get a node with an unsnoopable flag 1234 * @bat_priv: the bat priv with all the soft interface information 1235 * 1236 * Return: an orig_node which has the BATADV_MCAST_WANT_ALL_UNSNOOPABLES flag 1237 * set and increases its refcount. 1238 */ 1239 static struct batadv_orig_node * 1240 batadv_mcast_forw_unsnoop_node_get(struct batadv_priv *bat_priv) 1241 { 1242 struct batadv_orig_node *tmp_orig_node, *orig_node = NULL; 1243 1244 rcu_read_lock(); 1245 hlist_for_each_entry_rcu(tmp_orig_node, 1246 &bat_priv->mcast.want_all_unsnoopables_list, 1247 mcast_want_all_unsnoopables_node) { 1248 if (!kref_get_unless_zero(&tmp_orig_node->refcount)) 1249 continue; 1250 1251 orig_node = tmp_orig_node; 1252 break; 1253 } 1254 rcu_read_unlock(); 1255 1256 return orig_node; 1257 } 1258 1259 /** 1260 * batadv_mcast_forw_rtr4_node_get() - get a node with an ipv4 mcast router flag 1261 * @bat_priv: the bat priv with all the soft interface information 1262 * 1263 * Return: an orig_node which has the BATADV_MCAST_WANT_NO_RTR4 flag unset and 1264 * increases its refcount. 1265 */ 1266 static struct batadv_orig_node * 1267 batadv_mcast_forw_rtr4_node_get(struct batadv_priv *bat_priv) 1268 { 1269 struct batadv_orig_node *tmp_orig_node, *orig_node = NULL; 1270 1271 rcu_read_lock(); 1272 hlist_for_each_entry_rcu(tmp_orig_node, 1273 &bat_priv->mcast.want_all_rtr4_list, 1274 mcast_want_all_rtr4_node) { 1275 if (!kref_get_unless_zero(&tmp_orig_node->refcount)) 1276 continue; 1277 1278 orig_node = tmp_orig_node; 1279 break; 1280 } 1281 rcu_read_unlock(); 1282 1283 return orig_node; 1284 } 1285 1286 /** 1287 * batadv_mcast_forw_rtr6_node_get() - get a node with an ipv6 mcast router flag 1288 * @bat_priv: the bat priv with all the soft interface information 1289 * 1290 * Return: an orig_node which has the BATADV_MCAST_WANT_NO_RTR6 flag unset 1291 * and increases its refcount. 1292 */ 1293 static struct batadv_orig_node * 1294 batadv_mcast_forw_rtr6_node_get(struct batadv_priv *bat_priv) 1295 { 1296 struct batadv_orig_node *tmp_orig_node, *orig_node = NULL; 1297 1298 rcu_read_lock(); 1299 hlist_for_each_entry_rcu(tmp_orig_node, 1300 &bat_priv->mcast.want_all_rtr6_list, 1301 mcast_want_all_rtr6_node) { 1302 if (!kref_get_unless_zero(&tmp_orig_node->refcount)) 1303 continue; 1304 1305 orig_node = tmp_orig_node; 1306 break; 1307 } 1308 rcu_read_unlock(); 1309 1310 return orig_node; 1311 } 1312 1313 /** 1314 * batadv_mcast_forw_rtr_node_get() - get a node with an ipv4/ipv6 router flag 1315 * @bat_priv: the bat priv with all the soft interface information 1316 * @ethhdr: an ethernet header to determine the protocol family from 1317 * 1318 * Return: an orig_node which has no BATADV_MCAST_WANT_NO_RTR4 or 1319 * BATADV_MCAST_WANT_NO_RTR6 flag, depending on the provided ethhdr, set and 1320 * increases its refcount. 1321 */ 1322 static struct batadv_orig_node * 1323 batadv_mcast_forw_rtr_node_get(struct batadv_priv *bat_priv, 1324 struct ethhdr *ethhdr) 1325 { 1326 switch (ntohs(ethhdr->h_proto)) { 1327 case ETH_P_IP: 1328 return batadv_mcast_forw_rtr4_node_get(bat_priv); 1329 case ETH_P_IPV6: 1330 return batadv_mcast_forw_rtr6_node_get(bat_priv); 1331 default: 1332 /* we shouldn't be here... */ 1333 return NULL; 1334 } 1335 } 1336 1337 /** 1338 * batadv_mcast_forw_mode() - check on how to forward a multicast packet 1339 * @bat_priv: the bat priv with all the soft interface information 1340 * @skb: The multicast packet to check 1341 * @orig: an originator to be set to forward the skb to 1342 * @is_routable: stores whether the destination is routable 1343 * 1344 * Return: the forwarding mode as enum batadv_forw_mode and in case of 1345 * BATADV_FORW_SINGLE set the orig to the single originator the skb 1346 * should be forwarded to. 1347 */ 1348 enum batadv_forw_mode 1349 batadv_mcast_forw_mode(struct batadv_priv *bat_priv, struct sk_buff *skb, 1350 struct batadv_orig_node **orig, int *is_routable) 1351 { 1352 int ret, tt_count, ip_count, unsnoop_count, total_count; 1353 bool is_unsnoopable = false; 1354 unsigned int mcast_fanout; 1355 struct ethhdr *ethhdr; 1356 int rtr_count = 0; 1357 1358 ret = batadv_mcast_forw_mode_check(bat_priv, skb, &is_unsnoopable, 1359 is_routable); 1360 if (ret == -ENOMEM) 1361 return BATADV_FORW_NONE; 1362 else if (ret < 0) 1363 return BATADV_FORW_ALL; 1364 1365 ethhdr = eth_hdr(skb); 1366 1367 tt_count = batadv_tt_global_hash_count(bat_priv, ethhdr->h_dest, 1368 BATADV_NO_FLAGS); 1369 ip_count = batadv_mcast_forw_want_all_ip_count(bat_priv, ethhdr); 1370 unsnoop_count = !is_unsnoopable ? 0 : 1371 atomic_read(&bat_priv->mcast.num_want_all_unsnoopables); 1372 rtr_count = batadv_mcast_forw_rtr_count(bat_priv, *is_routable); 1373 1374 total_count = tt_count + ip_count + unsnoop_count + rtr_count; 1375 1376 switch (total_count) { 1377 case 1: 1378 if (tt_count) 1379 *orig = batadv_mcast_forw_tt_node_get(bat_priv, ethhdr); 1380 else if (ip_count) 1381 *orig = batadv_mcast_forw_ip_node_get(bat_priv, ethhdr); 1382 else if (unsnoop_count) 1383 *orig = batadv_mcast_forw_unsnoop_node_get(bat_priv); 1384 else if (rtr_count) 1385 *orig = batadv_mcast_forw_rtr_node_get(bat_priv, 1386 ethhdr); 1387 1388 if (*orig) 1389 return BATADV_FORW_SINGLE; 1390 1391 fallthrough; 1392 case 0: 1393 return BATADV_FORW_NONE; 1394 default: 1395 mcast_fanout = atomic_read(&bat_priv->multicast_fanout); 1396 1397 if (!unsnoop_count && total_count <= mcast_fanout) 1398 return BATADV_FORW_SOME; 1399 } 1400 1401 return BATADV_FORW_ALL; 1402 } 1403 1404 /** 1405 * batadv_mcast_forw_send_orig() - send a multicast packet to an originator 1406 * @bat_priv: the bat priv with all the soft interface information 1407 * @skb: the multicast packet to send 1408 * @vid: the vlan identifier 1409 * @orig_node: the originator to send the packet to 1410 * 1411 * Return: NET_XMIT_DROP in case of error or NET_XMIT_SUCCESS otherwise. 1412 */ 1413 int batadv_mcast_forw_send_orig(struct batadv_priv *bat_priv, 1414 struct sk_buff *skb, 1415 unsigned short vid, 1416 struct batadv_orig_node *orig_node) 1417 { 1418 /* Avoid sending multicast-in-unicast packets to other BLA 1419 * gateways - they already got the frame from the LAN side 1420 * we share with them. 1421 * TODO: Refactor to take BLA into account earlier, to avoid 1422 * reducing the mcast_fanout count. 1423 */ 1424 if (batadv_bla_is_backbone_gw_orig(bat_priv, orig_node->orig, vid)) { 1425 dev_kfree_skb(skb); 1426 return NET_XMIT_SUCCESS; 1427 } 1428 1429 return batadv_send_skb_unicast(bat_priv, skb, BATADV_UNICAST, 0, 1430 orig_node, vid); 1431 } 1432 1433 /** 1434 * batadv_mcast_forw_tt() - forwards a packet to multicast listeners 1435 * @bat_priv: the bat priv with all the soft interface information 1436 * @skb: the multicast packet to transmit 1437 * @vid: the vlan identifier 1438 * 1439 * Sends copies of a frame with multicast destination to any multicast 1440 * listener registered in the translation table. A transmission is performed 1441 * via a batman-adv unicast packet for each such destination node. 1442 * 1443 * Return: NET_XMIT_DROP on memory allocation failure, NET_XMIT_SUCCESS 1444 * otherwise. 1445 */ 1446 static int 1447 batadv_mcast_forw_tt(struct batadv_priv *bat_priv, struct sk_buff *skb, 1448 unsigned short vid) 1449 { 1450 int ret = NET_XMIT_SUCCESS; 1451 struct sk_buff *newskb; 1452 1453 struct batadv_tt_orig_list_entry *orig_entry; 1454 1455 struct batadv_tt_global_entry *tt_global; 1456 const u8 *addr = eth_hdr(skb)->h_dest; 1457 1458 tt_global = batadv_tt_global_hash_find(bat_priv, addr, vid); 1459 if (!tt_global) 1460 goto out; 1461 1462 rcu_read_lock(); 1463 hlist_for_each_entry_rcu(orig_entry, &tt_global->orig_list, list) { 1464 newskb = skb_copy(skb, GFP_ATOMIC); 1465 if (!newskb) { 1466 ret = NET_XMIT_DROP; 1467 break; 1468 } 1469 1470 batadv_mcast_forw_send_orig(bat_priv, newskb, vid, 1471 orig_entry->orig_node); 1472 } 1473 rcu_read_unlock(); 1474 1475 batadv_tt_global_entry_put(tt_global); 1476 1477 out: 1478 return ret; 1479 } 1480 1481 /** 1482 * batadv_mcast_forw_want_all_ipv4() - forward to nodes with want-all-ipv4 1483 * @bat_priv: the bat priv with all the soft interface information 1484 * @skb: the multicast packet to transmit 1485 * @vid: the vlan identifier 1486 * 1487 * Sends copies of a frame with multicast destination to any node with a 1488 * BATADV_MCAST_WANT_ALL_IPV4 flag set. A transmission is performed via a 1489 * batman-adv unicast packet for each such destination node. 1490 * 1491 * Return: NET_XMIT_DROP on memory allocation failure, NET_XMIT_SUCCESS 1492 * otherwise. 1493 */ 1494 static int 1495 batadv_mcast_forw_want_all_ipv4(struct batadv_priv *bat_priv, 1496 struct sk_buff *skb, unsigned short vid) 1497 { 1498 struct batadv_orig_node *orig_node; 1499 int ret = NET_XMIT_SUCCESS; 1500 struct sk_buff *newskb; 1501 1502 rcu_read_lock(); 1503 hlist_for_each_entry_rcu(orig_node, 1504 &bat_priv->mcast.want_all_ipv4_list, 1505 mcast_want_all_ipv4_node) { 1506 newskb = skb_copy(skb, GFP_ATOMIC); 1507 if (!newskb) { 1508 ret = NET_XMIT_DROP; 1509 break; 1510 } 1511 1512 batadv_mcast_forw_send_orig(bat_priv, newskb, vid, orig_node); 1513 } 1514 rcu_read_unlock(); 1515 return ret; 1516 } 1517 1518 /** 1519 * batadv_mcast_forw_want_all_ipv6() - forward to nodes with want-all-ipv6 1520 * @bat_priv: the bat priv with all the soft interface information 1521 * @skb: The multicast packet to transmit 1522 * @vid: the vlan identifier 1523 * 1524 * Sends copies of a frame with multicast destination to any node with a 1525 * BATADV_MCAST_WANT_ALL_IPV6 flag set. A transmission is performed via a 1526 * batman-adv unicast packet for each such destination node. 1527 * 1528 * Return: NET_XMIT_DROP on memory allocation failure, NET_XMIT_SUCCESS 1529 * otherwise. 1530 */ 1531 static int 1532 batadv_mcast_forw_want_all_ipv6(struct batadv_priv *bat_priv, 1533 struct sk_buff *skb, unsigned short vid) 1534 { 1535 struct batadv_orig_node *orig_node; 1536 int ret = NET_XMIT_SUCCESS; 1537 struct sk_buff *newskb; 1538 1539 rcu_read_lock(); 1540 hlist_for_each_entry_rcu(orig_node, 1541 &bat_priv->mcast.want_all_ipv6_list, 1542 mcast_want_all_ipv6_node) { 1543 newskb = skb_copy(skb, GFP_ATOMIC); 1544 if (!newskb) { 1545 ret = NET_XMIT_DROP; 1546 break; 1547 } 1548 1549 batadv_mcast_forw_send_orig(bat_priv, newskb, vid, orig_node); 1550 } 1551 rcu_read_unlock(); 1552 return ret; 1553 } 1554 1555 /** 1556 * batadv_mcast_forw_want_all() - forward packet to nodes in a want-all list 1557 * @bat_priv: the bat priv with all the soft interface information 1558 * @skb: the multicast packet to transmit 1559 * @vid: the vlan identifier 1560 * 1561 * Sends copies of a frame with multicast destination to any node with a 1562 * BATADV_MCAST_WANT_ALL_IPV4 or BATADV_MCAST_WANT_ALL_IPV6 flag set. A 1563 * transmission is performed via a batman-adv unicast packet for each such 1564 * destination node. 1565 * 1566 * Return: NET_XMIT_DROP on memory allocation failure or if the protocol family 1567 * is neither IPv4 nor IPv6. NET_XMIT_SUCCESS otherwise. 1568 */ 1569 static int 1570 batadv_mcast_forw_want_all(struct batadv_priv *bat_priv, 1571 struct sk_buff *skb, unsigned short vid) 1572 { 1573 switch (ntohs(eth_hdr(skb)->h_proto)) { 1574 case ETH_P_IP: 1575 return batadv_mcast_forw_want_all_ipv4(bat_priv, skb, vid); 1576 case ETH_P_IPV6: 1577 return batadv_mcast_forw_want_all_ipv6(bat_priv, skb, vid); 1578 default: 1579 /* we shouldn't be here... */ 1580 return NET_XMIT_DROP; 1581 } 1582 } 1583 1584 /** 1585 * batadv_mcast_forw_want_all_rtr4() - forward to nodes with want-all-rtr4 1586 * @bat_priv: the bat priv with all the soft interface information 1587 * @skb: the multicast packet to transmit 1588 * @vid: the vlan identifier 1589 * 1590 * Sends copies of a frame with multicast destination to any node with a 1591 * BATADV_MCAST_WANT_NO_RTR4 flag unset. A transmission is performed via a 1592 * batman-adv unicast packet for each such destination node. 1593 * 1594 * Return: NET_XMIT_DROP on memory allocation failure, NET_XMIT_SUCCESS 1595 * otherwise. 1596 */ 1597 static int 1598 batadv_mcast_forw_want_all_rtr4(struct batadv_priv *bat_priv, 1599 struct sk_buff *skb, unsigned short vid) 1600 { 1601 struct batadv_orig_node *orig_node; 1602 int ret = NET_XMIT_SUCCESS; 1603 struct sk_buff *newskb; 1604 1605 rcu_read_lock(); 1606 hlist_for_each_entry_rcu(orig_node, 1607 &bat_priv->mcast.want_all_rtr4_list, 1608 mcast_want_all_rtr4_node) { 1609 newskb = skb_copy(skb, GFP_ATOMIC); 1610 if (!newskb) { 1611 ret = NET_XMIT_DROP; 1612 break; 1613 } 1614 1615 batadv_mcast_forw_send_orig(bat_priv, newskb, vid, orig_node); 1616 } 1617 rcu_read_unlock(); 1618 return ret; 1619 } 1620 1621 /** 1622 * batadv_mcast_forw_want_all_rtr6() - forward to nodes with want-all-rtr6 1623 * @bat_priv: the bat priv with all the soft interface information 1624 * @skb: The multicast packet to transmit 1625 * @vid: the vlan identifier 1626 * 1627 * Sends copies of a frame with multicast destination to any node with a 1628 * BATADV_MCAST_WANT_NO_RTR6 flag unset. A transmission is performed via a 1629 * batman-adv unicast packet for each such destination node. 1630 * 1631 * Return: NET_XMIT_DROP on memory allocation failure, NET_XMIT_SUCCESS 1632 * otherwise. 1633 */ 1634 static int 1635 batadv_mcast_forw_want_all_rtr6(struct batadv_priv *bat_priv, 1636 struct sk_buff *skb, unsigned short vid) 1637 { 1638 struct batadv_orig_node *orig_node; 1639 int ret = NET_XMIT_SUCCESS; 1640 struct sk_buff *newskb; 1641 1642 rcu_read_lock(); 1643 hlist_for_each_entry_rcu(orig_node, 1644 &bat_priv->mcast.want_all_rtr6_list, 1645 mcast_want_all_rtr6_node) { 1646 newskb = skb_copy(skb, GFP_ATOMIC); 1647 if (!newskb) { 1648 ret = NET_XMIT_DROP; 1649 break; 1650 } 1651 1652 batadv_mcast_forw_send_orig(bat_priv, newskb, vid, orig_node); 1653 } 1654 rcu_read_unlock(); 1655 return ret; 1656 } 1657 1658 /** 1659 * batadv_mcast_forw_want_rtr() - forward packet to nodes in a want-all-rtr list 1660 * @bat_priv: the bat priv with all the soft interface information 1661 * @skb: the multicast packet to transmit 1662 * @vid: the vlan identifier 1663 * 1664 * Sends copies of a frame with multicast destination to any node with a 1665 * BATADV_MCAST_WANT_NO_RTR4 or BATADV_MCAST_WANT_NO_RTR6 flag unset. A 1666 * transmission is performed via a batman-adv unicast packet for each such 1667 * destination node. 1668 * 1669 * Return: NET_XMIT_DROP on memory allocation failure or if the protocol family 1670 * is neither IPv4 nor IPv6. NET_XMIT_SUCCESS otherwise. 1671 */ 1672 static int 1673 batadv_mcast_forw_want_rtr(struct batadv_priv *bat_priv, 1674 struct sk_buff *skb, unsigned short vid) 1675 { 1676 switch (ntohs(eth_hdr(skb)->h_proto)) { 1677 case ETH_P_IP: 1678 return batadv_mcast_forw_want_all_rtr4(bat_priv, skb, vid); 1679 case ETH_P_IPV6: 1680 return batadv_mcast_forw_want_all_rtr6(bat_priv, skb, vid); 1681 default: 1682 /* we shouldn't be here... */ 1683 return NET_XMIT_DROP; 1684 } 1685 } 1686 1687 /** 1688 * batadv_mcast_forw_send() - send packet to any detected multicast recipient 1689 * @bat_priv: the bat priv with all the soft interface information 1690 * @skb: the multicast packet to transmit 1691 * @vid: the vlan identifier 1692 * @is_routable: stores whether the destination is routable 1693 * 1694 * Sends copies of a frame with multicast destination to any node that signaled 1695 * interest in it, that is either via the translation table or the according 1696 * want-all flags. A transmission is performed via a batman-adv unicast packet 1697 * for each such destination node. 1698 * 1699 * The given skb is consumed/freed. 1700 * 1701 * Return: NET_XMIT_DROP on memory allocation failure or if the protocol family 1702 * is neither IPv4 nor IPv6. NET_XMIT_SUCCESS otherwise. 1703 */ 1704 int batadv_mcast_forw_send(struct batadv_priv *bat_priv, struct sk_buff *skb, 1705 unsigned short vid, int is_routable) 1706 { 1707 int ret; 1708 1709 ret = batadv_mcast_forw_tt(bat_priv, skb, vid); 1710 if (ret != NET_XMIT_SUCCESS) { 1711 kfree_skb(skb); 1712 return ret; 1713 } 1714 1715 ret = batadv_mcast_forw_want_all(bat_priv, skb, vid); 1716 if (ret != NET_XMIT_SUCCESS) { 1717 kfree_skb(skb); 1718 return ret; 1719 } 1720 1721 if (!is_routable) 1722 goto skip_mc_router; 1723 1724 ret = batadv_mcast_forw_want_rtr(bat_priv, skb, vid); 1725 if (ret != NET_XMIT_SUCCESS) { 1726 kfree_skb(skb); 1727 return ret; 1728 } 1729 1730 skip_mc_router: 1731 consume_skb(skb); 1732 return ret; 1733 } 1734 1735 /** 1736 * batadv_mcast_want_unsnoop_update() - update unsnoop counter and list 1737 * @bat_priv: the bat priv with all the soft interface information 1738 * @orig: the orig_node which multicast state might have changed of 1739 * @mcast_flags: flags indicating the new multicast state 1740 * 1741 * If the BATADV_MCAST_WANT_ALL_UNSNOOPABLES flag of this originator, 1742 * orig, has toggled then this method updates the counter and the list 1743 * accordingly. 1744 * 1745 * Caller needs to hold orig->mcast_handler_lock. 1746 */ 1747 static void batadv_mcast_want_unsnoop_update(struct batadv_priv *bat_priv, 1748 struct batadv_orig_node *orig, 1749 u8 mcast_flags) 1750 { 1751 struct hlist_node *node = &orig->mcast_want_all_unsnoopables_node; 1752 struct hlist_head *head = &bat_priv->mcast.want_all_unsnoopables_list; 1753 1754 lockdep_assert_held(&orig->mcast_handler_lock); 1755 1756 /* switched from flag unset to set */ 1757 if (mcast_flags & BATADV_MCAST_WANT_ALL_UNSNOOPABLES && 1758 !(orig->mcast_flags & BATADV_MCAST_WANT_ALL_UNSNOOPABLES)) { 1759 atomic_inc(&bat_priv->mcast.num_want_all_unsnoopables); 1760 1761 spin_lock_bh(&bat_priv->mcast.want_lists_lock); 1762 /* flag checks above + mcast_handler_lock prevents this */ 1763 WARN_ON(!hlist_unhashed(node)); 1764 1765 hlist_add_head_rcu(node, head); 1766 spin_unlock_bh(&bat_priv->mcast.want_lists_lock); 1767 /* switched from flag set to unset */ 1768 } else if (!(mcast_flags & BATADV_MCAST_WANT_ALL_UNSNOOPABLES) && 1769 orig->mcast_flags & BATADV_MCAST_WANT_ALL_UNSNOOPABLES) { 1770 atomic_dec(&bat_priv->mcast.num_want_all_unsnoopables); 1771 1772 spin_lock_bh(&bat_priv->mcast.want_lists_lock); 1773 /* flag checks above + mcast_handler_lock prevents this */ 1774 WARN_ON(hlist_unhashed(node)); 1775 1776 hlist_del_init_rcu(node); 1777 spin_unlock_bh(&bat_priv->mcast.want_lists_lock); 1778 } 1779 } 1780 1781 /** 1782 * batadv_mcast_want_ipv4_update() - update want-all-ipv4 counter and list 1783 * @bat_priv: the bat priv with all the soft interface information 1784 * @orig: the orig_node which multicast state might have changed of 1785 * @mcast_flags: flags indicating the new multicast state 1786 * 1787 * If the BATADV_MCAST_WANT_ALL_IPV4 flag of this originator, orig, has 1788 * toggled then this method updates the counter and the list accordingly. 1789 * 1790 * Caller needs to hold orig->mcast_handler_lock. 1791 */ 1792 static void batadv_mcast_want_ipv4_update(struct batadv_priv *bat_priv, 1793 struct batadv_orig_node *orig, 1794 u8 mcast_flags) 1795 { 1796 struct hlist_node *node = &orig->mcast_want_all_ipv4_node; 1797 struct hlist_head *head = &bat_priv->mcast.want_all_ipv4_list; 1798 1799 lockdep_assert_held(&orig->mcast_handler_lock); 1800 1801 /* switched from flag unset to set */ 1802 if (mcast_flags & BATADV_MCAST_WANT_ALL_IPV4 && 1803 !(orig->mcast_flags & BATADV_MCAST_WANT_ALL_IPV4)) { 1804 atomic_inc(&bat_priv->mcast.num_want_all_ipv4); 1805 1806 spin_lock_bh(&bat_priv->mcast.want_lists_lock); 1807 /* flag checks above + mcast_handler_lock prevents this */ 1808 WARN_ON(!hlist_unhashed(node)); 1809 1810 hlist_add_head_rcu(node, head); 1811 spin_unlock_bh(&bat_priv->mcast.want_lists_lock); 1812 /* switched from flag set to unset */ 1813 } else if (!(mcast_flags & BATADV_MCAST_WANT_ALL_IPV4) && 1814 orig->mcast_flags & BATADV_MCAST_WANT_ALL_IPV4) { 1815 atomic_dec(&bat_priv->mcast.num_want_all_ipv4); 1816 1817 spin_lock_bh(&bat_priv->mcast.want_lists_lock); 1818 /* flag checks above + mcast_handler_lock prevents this */ 1819 WARN_ON(hlist_unhashed(node)); 1820 1821 hlist_del_init_rcu(node); 1822 spin_unlock_bh(&bat_priv->mcast.want_lists_lock); 1823 } 1824 } 1825 1826 /** 1827 * batadv_mcast_want_ipv6_update() - update want-all-ipv6 counter and list 1828 * @bat_priv: the bat priv with all the soft interface information 1829 * @orig: the orig_node which multicast state might have changed of 1830 * @mcast_flags: flags indicating the new multicast state 1831 * 1832 * If the BATADV_MCAST_WANT_ALL_IPV6 flag of this originator, orig, has 1833 * toggled then this method updates the counter and the list accordingly. 1834 * 1835 * Caller needs to hold orig->mcast_handler_lock. 1836 */ 1837 static void batadv_mcast_want_ipv6_update(struct batadv_priv *bat_priv, 1838 struct batadv_orig_node *orig, 1839 u8 mcast_flags) 1840 { 1841 struct hlist_node *node = &orig->mcast_want_all_ipv6_node; 1842 struct hlist_head *head = &bat_priv->mcast.want_all_ipv6_list; 1843 1844 lockdep_assert_held(&orig->mcast_handler_lock); 1845 1846 /* switched from flag unset to set */ 1847 if (mcast_flags & BATADV_MCAST_WANT_ALL_IPV6 && 1848 !(orig->mcast_flags & BATADV_MCAST_WANT_ALL_IPV6)) { 1849 atomic_inc(&bat_priv->mcast.num_want_all_ipv6); 1850 1851 spin_lock_bh(&bat_priv->mcast.want_lists_lock); 1852 /* flag checks above + mcast_handler_lock prevents this */ 1853 WARN_ON(!hlist_unhashed(node)); 1854 1855 hlist_add_head_rcu(node, head); 1856 spin_unlock_bh(&bat_priv->mcast.want_lists_lock); 1857 /* switched from flag set to unset */ 1858 } else if (!(mcast_flags & BATADV_MCAST_WANT_ALL_IPV6) && 1859 orig->mcast_flags & BATADV_MCAST_WANT_ALL_IPV6) { 1860 atomic_dec(&bat_priv->mcast.num_want_all_ipv6); 1861 1862 spin_lock_bh(&bat_priv->mcast.want_lists_lock); 1863 /* flag checks above + mcast_handler_lock prevents this */ 1864 WARN_ON(hlist_unhashed(node)); 1865 1866 hlist_del_init_rcu(node); 1867 spin_unlock_bh(&bat_priv->mcast.want_lists_lock); 1868 } 1869 } 1870 1871 /** 1872 * batadv_mcast_want_rtr4_update() - update want-all-rtr4 counter and list 1873 * @bat_priv: the bat priv with all the soft interface information 1874 * @orig: the orig_node which multicast state might have changed of 1875 * @mcast_flags: flags indicating the new multicast state 1876 * 1877 * If the BATADV_MCAST_WANT_NO_RTR4 flag of this originator, orig, has 1878 * toggled then this method updates the counter and the list accordingly. 1879 * 1880 * Caller needs to hold orig->mcast_handler_lock. 1881 */ 1882 static void batadv_mcast_want_rtr4_update(struct batadv_priv *bat_priv, 1883 struct batadv_orig_node *orig, 1884 u8 mcast_flags) 1885 { 1886 struct hlist_node *node = &orig->mcast_want_all_rtr4_node; 1887 struct hlist_head *head = &bat_priv->mcast.want_all_rtr4_list; 1888 1889 lockdep_assert_held(&orig->mcast_handler_lock); 1890 1891 /* switched from flag set to unset */ 1892 if (!(mcast_flags & BATADV_MCAST_WANT_NO_RTR4) && 1893 orig->mcast_flags & BATADV_MCAST_WANT_NO_RTR4) { 1894 atomic_inc(&bat_priv->mcast.num_want_all_rtr4); 1895 1896 spin_lock_bh(&bat_priv->mcast.want_lists_lock); 1897 /* flag checks above + mcast_handler_lock prevents this */ 1898 WARN_ON(!hlist_unhashed(node)); 1899 1900 hlist_add_head_rcu(node, head); 1901 spin_unlock_bh(&bat_priv->mcast.want_lists_lock); 1902 /* switched from flag unset to set */ 1903 } else if (mcast_flags & BATADV_MCAST_WANT_NO_RTR4 && 1904 !(orig->mcast_flags & BATADV_MCAST_WANT_NO_RTR4)) { 1905 atomic_dec(&bat_priv->mcast.num_want_all_rtr4); 1906 1907 spin_lock_bh(&bat_priv->mcast.want_lists_lock); 1908 /* flag checks above + mcast_handler_lock prevents this */ 1909 WARN_ON(hlist_unhashed(node)); 1910 1911 hlist_del_init_rcu(node); 1912 spin_unlock_bh(&bat_priv->mcast.want_lists_lock); 1913 } 1914 } 1915 1916 /** 1917 * batadv_mcast_want_rtr6_update() - update want-all-rtr6 counter and list 1918 * @bat_priv: the bat priv with all the soft interface information 1919 * @orig: the orig_node which multicast state might have changed of 1920 * @mcast_flags: flags indicating the new multicast state 1921 * 1922 * If the BATADV_MCAST_WANT_NO_RTR6 flag of this originator, orig, has 1923 * toggled then this method updates the counter and the list accordingly. 1924 * 1925 * Caller needs to hold orig->mcast_handler_lock. 1926 */ 1927 static void batadv_mcast_want_rtr6_update(struct batadv_priv *bat_priv, 1928 struct batadv_orig_node *orig, 1929 u8 mcast_flags) 1930 { 1931 struct hlist_node *node = &orig->mcast_want_all_rtr6_node; 1932 struct hlist_head *head = &bat_priv->mcast.want_all_rtr6_list; 1933 1934 lockdep_assert_held(&orig->mcast_handler_lock); 1935 1936 /* switched from flag set to unset */ 1937 if (!(mcast_flags & BATADV_MCAST_WANT_NO_RTR6) && 1938 orig->mcast_flags & BATADV_MCAST_WANT_NO_RTR6) { 1939 atomic_inc(&bat_priv->mcast.num_want_all_rtr6); 1940 1941 spin_lock_bh(&bat_priv->mcast.want_lists_lock); 1942 /* flag checks above + mcast_handler_lock prevents this */ 1943 WARN_ON(!hlist_unhashed(node)); 1944 1945 hlist_add_head_rcu(node, head); 1946 spin_unlock_bh(&bat_priv->mcast.want_lists_lock); 1947 /* switched from flag unset to set */ 1948 } else if (mcast_flags & BATADV_MCAST_WANT_NO_RTR6 && 1949 !(orig->mcast_flags & BATADV_MCAST_WANT_NO_RTR6)) { 1950 atomic_dec(&bat_priv->mcast.num_want_all_rtr6); 1951 1952 spin_lock_bh(&bat_priv->mcast.want_lists_lock); 1953 /* flag checks above + mcast_handler_lock prevents this */ 1954 WARN_ON(hlist_unhashed(node)); 1955 1956 hlist_del_init_rcu(node); 1957 spin_unlock_bh(&bat_priv->mcast.want_lists_lock); 1958 } 1959 } 1960 1961 /** 1962 * batadv_mcast_tvlv_flags_get() - get multicast flags from an OGM TVLV 1963 * @enabled: whether the originator has multicast TVLV support enabled 1964 * @tvlv_value: tvlv buffer containing the multicast flags 1965 * @tvlv_value_len: tvlv buffer length 1966 * 1967 * Return: multicast flags for the given tvlv buffer 1968 */ 1969 static u8 1970 batadv_mcast_tvlv_flags_get(bool enabled, void *tvlv_value, u16 tvlv_value_len) 1971 { 1972 u8 mcast_flags = BATADV_NO_FLAGS; 1973 1974 if (enabled && tvlv_value && tvlv_value_len >= sizeof(mcast_flags)) 1975 mcast_flags = *(u8 *)tvlv_value; 1976 1977 if (!enabled) { 1978 mcast_flags |= BATADV_MCAST_WANT_ALL_IPV4; 1979 mcast_flags |= BATADV_MCAST_WANT_ALL_IPV6; 1980 } 1981 1982 /* remove redundant flags to avoid sending duplicate packets later */ 1983 if (mcast_flags & BATADV_MCAST_WANT_ALL_IPV4) 1984 mcast_flags |= BATADV_MCAST_WANT_NO_RTR4; 1985 1986 if (mcast_flags & BATADV_MCAST_WANT_ALL_IPV6) 1987 mcast_flags |= BATADV_MCAST_WANT_NO_RTR6; 1988 1989 return mcast_flags; 1990 } 1991 1992 /** 1993 * batadv_mcast_tvlv_ogm_handler() - process incoming multicast tvlv container 1994 * @bat_priv: the bat priv with all the soft interface information 1995 * @orig: the orig_node of the ogm 1996 * @flags: flags indicating the tvlv state (see batadv_tvlv_handler_flags) 1997 * @tvlv_value: tvlv buffer containing the multicast data 1998 * @tvlv_value_len: tvlv buffer length 1999 */ 2000 static void batadv_mcast_tvlv_ogm_handler(struct batadv_priv *bat_priv, 2001 struct batadv_orig_node *orig, 2002 u8 flags, 2003 void *tvlv_value, 2004 u16 tvlv_value_len) 2005 { 2006 bool orig_mcast_enabled = !(flags & BATADV_TVLV_HANDLER_OGM_CIFNOTFND); 2007 u8 mcast_flags; 2008 2009 mcast_flags = batadv_mcast_tvlv_flags_get(orig_mcast_enabled, 2010 tvlv_value, tvlv_value_len); 2011 2012 spin_lock_bh(&orig->mcast_handler_lock); 2013 2014 if (orig_mcast_enabled && 2015 !test_bit(BATADV_ORIG_CAPA_HAS_MCAST, &orig->capabilities)) { 2016 set_bit(BATADV_ORIG_CAPA_HAS_MCAST, &orig->capabilities); 2017 } else if (!orig_mcast_enabled && 2018 test_bit(BATADV_ORIG_CAPA_HAS_MCAST, &orig->capabilities)) { 2019 clear_bit(BATADV_ORIG_CAPA_HAS_MCAST, &orig->capabilities); 2020 } 2021 2022 set_bit(BATADV_ORIG_CAPA_HAS_MCAST, &orig->capa_initialized); 2023 2024 batadv_mcast_want_unsnoop_update(bat_priv, orig, mcast_flags); 2025 batadv_mcast_want_ipv4_update(bat_priv, orig, mcast_flags); 2026 batadv_mcast_want_ipv6_update(bat_priv, orig, mcast_flags); 2027 batadv_mcast_want_rtr4_update(bat_priv, orig, mcast_flags); 2028 batadv_mcast_want_rtr6_update(bat_priv, orig, mcast_flags); 2029 2030 orig->mcast_flags = mcast_flags; 2031 spin_unlock_bh(&orig->mcast_handler_lock); 2032 } 2033 2034 /** 2035 * batadv_mcast_init() - initialize the multicast optimizations structures 2036 * @bat_priv: the bat priv with all the soft interface information 2037 */ 2038 void batadv_mcast_init(struct batadv_priv *bat_priv) 2039 { 2040 batadv_tvlv_handler_register(bat_priv, batadv_mcast_tvlv_ogm_handler, 2041 NULL, BATADV_TVLV_MCAST, 2, 2042 BATADV_TVLV_HANDLER_OGM_CIFNOTFND); 2043 2044 INIT_DELAYED_WORK(&bat_priv->mcast.work, batadv_mcast_mla_update); 2045 batadv_mcast_start_timer(bat_priv); 2046 } 2047 2048 /** 2049 * batadv_mcast_mesh_info_put() - put multicast info into a netlink message 2050 * @msg: buffer for the message 2051 * @bat_priv: the bat priv with all the soft interface information 2052 * 2053 * Return: 0 or error code. 2054 */ 2055 int batadv_mcast_mesh_info_put(struct sk_buff *msg, 2056 struct batadv_priv *bat_priv) 2057 { 2058 u32 flags = bat_priv->mcast.mla_flags.tvlv_flags; 2059 u32 flags_priv = BATADV_NO_FLAGS; 2060 2061 if (bat_priv->mcast.mla_flags.bridged) { 2062 flags_priv |= BATADV_MCAST_FLAGS_BRIDGED; 2063 2064 if (bat_priv->mcast.mla_flags.querier_ipv4.exists) 2065 flags_priv |= BATADV_MCAST_FLAGS_QUERIER_IPV4_EXISTS; 2066 if (bat_priv->mcast.mla_flags.querier_ipv6.exists) 2067 flags_priv |= BATADV_MCAST_FLAGS_QUERIER_IPV6_EXISTS; 2068 if (bat_priv->mcast.mla_flags.querier_ipv4.shadowing) 2069 flags_priv |= BATADV_MCAST_FLAGS_QUERIER_IPV4_SHADOWING; 2070 if (bat_priv->mcast.mla_flags.querier_ipv6.shadowing) 2071 flags_priv |= BATADV_MCAST_FLAGS_QUERIER_IPV6_SHADOWING; 2072 } 2073 2074 if (nla_put_u32(msg, BATADV_ATTR_MCAST_FLAGS, flags) || 2075 nla_put_u32(msg, BATADV_ATTR_MCAST_FLAGS_PRIV, flags_priv)) 2076 return -EMSGSIZE; 2077 2078 return 0; 2079 } 2080 2081 /** 2082 * batadv_mcast_flags_dump_entry() - dump one entry of the multicast flags table 2083 * to a netlink socket 2084 * @msg: buffer for the message 2085 * @portid: netlink port 2086 * @cb: Control block containing additional options 2087 * @orig_node: originator to dump the multicast flags of 2088 * 2089 * Return: 0 or error code. 2090 */ 2091 static int 2092 batadv_mcast_flags_dump_entry(struct sk_buff *msg, u32 portid, 2093 struct netlink_callback *cb, 2094 struct batadv_orig_node *orig_node) 2095 { 2096 void *hdr; 2097 2098 hdr = genlmsg_put(msg, portid, cb->nlh->nlmsg_seq, 2099 &batadv_netlink_family, NLM_F_MULTI, 2100 BATADV_CMD_GET_MCAST_FLAGS); 2101 if (!hdr) 2102 return -ENOBUFS; 2103 2104 genl_dump_check_consistent(cb, hdr); 2105 2106 if (nla_put(msg, BATADV_ATTR_ORIG_ADDRESS, ETH_ALEN, 2107 orig_node->orig)) { 2108 genlmsg_cancel(msg, hdr); 2109 return -EMSGSIZE; 2110 } 2111 2112 if (test_bit(BATADV_ORIG_CAPA_HAS_MCAST, 2113 &orig_node->capabilities)) { 2114 if (nla_put_u32(msg, BATADV_ATTR_MCAST_FLAGS, 2115 orig_node->mcast_flags)) { 2116 genlmsg_cancel(msg, hdr); 2117 return -EMSGSIZE; 2118 } 2119 } 2120 2121 genlmsg_end(msg, hdr); 2122 return 0; 2123 } 2124 2125 /** 2126 * batadv_mcast_flags_dump_bucket() - dump one bucket of the multicast flags 2127 * table to a netlink socket 2128 * @msg: buffer for the message 2129 * @portid: netlink port 2130 * @cb: Control block containing additional options 2131 * @hash: hash to dump 2132 * @bucket: bucket index to dump 2133 * @idx_skip: How many entries to skip 2134 * 2135 * Return: 0 or error code. 2136 */ 2137 static int 2138 batadv_mcast_flags_dump_bucket(struct sk_buff *msg, u32 portid, 2139 struct netlink_callback *cb, 2140 struct batadv_hashtable *hash, 2141 unsigned int bucket, long *idx_skip) 2142 { 2143 struct batadv_orig_node *orig_node; 2144 long idx = 0; 2145 2146 spin_lock_bh(&hash->list_locks[bucket]); 2147 cb->seq = atomic_read(&hash->generation) << 1 | 1; 2148 2149 hlist_for_each_entry(orig_node, &hash->table[bucket], hash_entry) { 2150 if (!test_bit(BATADV_ORIG_CAPA_HAS_MCAST, 2151 &orig_node->capa_initialized)) 2152 continue; 2153 2154 if (idx < *idx_skip) 2155 goto skip; 2156 2157 if (batadv_mcast_flags_dump_entry(msg, portid, cb, orig_node)) { 2158 spin_unlock_bh(&hash->list_locks[bucket]); 2159 *idx_skip = idx; 2160 2161 return -EMSGSIZE; 2162 } 2163 2164 skip: 2165 idx++; 2166 } 2167 spin_unlock_bh(&hash->list_locks[bucket]); 2168 2169 return 0; 2170 } 2171 2172 /** 2173 * __batadv_mcast_flags_dump() - dump multicast flags table to a netlink socket 2174 * @msg: buffer for the message 2175 * @portid: netlink port 2176 * @cb: Control block containing additional options 2177 * @bat_priv: the bat priv with all the soft interface information 2178 * @bucket: current bucket to dump 2179 * @idx: index in current bucket to the next entry to dump 2180 * 2181 * Return: 0 or error code. 2182 */ 2183 static int 2184 __batadv_mcast_flags_dump(struct sk_buff *msg, u32 portid, 2185 struct netlink_callback *cb, 2186 struct batadv_priv *bat_priv, long *bucket, long *idx) 2187 { 2188 struct batadv_hashtable *hash = bat_priv->orig_hash; 2189 long bucket_tmp = *bucket; 2190 long idx_tmp = *idx; 2191 2192 while (bucket_tmp < hash->size) { 2193 if (batadv_mcast_flags_dump_bucket(msg, portid, cb, hash, 2194 bucket_tmp, &idx_tmp)) 2195 break; 2196 2197 bucket_tmp++; 2198 idx_tmp = 0; 2199 } 2200 2201 *bucket = bucket_tmp; 2202 *idx = idx_tmp; 2203 2204 return msg->len; 2205 } 2206 2207 /** 2208 * batadv_mcast_netlink_get_primary() - get primary interface from netlink 2209 * callback 2210 * @cb: netlink callback structure 2211 * @primary_if: the primary interface pointer to return the result in 2212 * 2213 * Return: 0 or error code. 2214 */ 2215 static int 2216 batadv_mcast_netlink_get_primary(struct netlink_callback *cb, 2217 struct batadv_hard_iface **primary_if) 2218 { 2219 struct batadv_hard_iface *hard_iface = NULL; 2220 struct net *net = sock_net(cb->skb->sk); 2221 struct net_device *soft_iface; 2222 struct batadv_priv *bat_priv; 2223 int ifindex; 2224 int ret = 0; 2225 2226 ifindex = batadv_netlink_get_ifindex(cb->nlh, BATADV_ATTR_MESH_IFINDEX); 2227 if (!ifindex) 2228 return -EINVAL; 2229 2230 soft_iface = dev_get_by_index(net, ifindex); 2231 if (!soft_iface || !batadv_softif_is_valid(soft_iface)) { 2232 ret = -ENODEV; 2233 goto out; 2234 } 2235 2236 bat_priv = netdev_priv(soft_iface); 2237 2238 hard_iface = batadv_primary_if_get_selected(bat_priv); 2239 if (!hard_iface || hard_iface->if_status != BATADV_IF_ACTIVE) { 2240 ret = -ENOENT; 2241 goto out; 2242 } 2243 2244 out: 2245 dev_put(soft_iface); 2246 2247 if (!ret && primary_if) 2248 *primary_if = hard_iface; 2249 else 2250 batadv_hardif_put(hard_iface); 2251 2252 return ret; 2253 } 2254 2255 /** 2256 * batadv_mcast_flags_dump() - dump multicast flags table to a netlink socket 2257 * @msg: buffer for the message 2258 * @cb: callback structure containing arguments 2259 * 2260 * Return: message length. 2261 */ 2262 int batadv_mcast_flags_dump(struct sk_buff *msg, struct netlink_callback *cb) 2263 { 2264 struct batadv_hard_iface *primary_if = NULL; 2265 int portid = NETLINK_CB(cb->skb).portid; 2266 struct batadv_priv *bat_priv; 2267 long *bucket = &cb->args[0]; 2268 long *idx = &cb->args[1]; 2269 int ret; 2270 2271 ret = batadv_mcast_netlink_get_primary(cb, &primary_if); 2272 if (ret) 2273 return ret; 2274 2275 bat_priv = netdev_priv(primary_if->soft_iface); 2276 ret = __batadv_mcast_flags_dump(msg, portid, cb, bat_priv, bucket, idx); 2277 2278 batadv_hardif_put(primary_if); 2279 return ret; 2280 } 2281 2282 /** 2283 * batadv_mcast_free() - free the multicast optimizations structures 2284 * @bat_priv: the bat priv with all the soft interface information 2285 */ 2286 void batadv_mcast_free(struct batadv_priv *bat_priv) 2287 { 2288 cancel_delayed_work_sync(&bat_priv->mcast.work); 2289 2290 batadv_tvlv_container_unregister(bat_priv, BATADV_TVLV_MCAST, 2); 2291 batadv_tvlv_handler_unregister(bat_priv, BATADV_TVLV_MCAST, 2); 2292 2293 /* safely calling outside of worker, as worker was canceled above */ 2294 batadv_mcast_mla_tt_retract(bat_priv, NULL); 2295 } 2296 2297 /** 2298 * batadv_mcast_purge_orig() - reset originator global mcast state modifications 2299 * @orig: the originator which is going to get purged 2300 */ 2301 void batadv_mcast_purge_orig(struct batadv_orig_node *orig) 2302 { 2303 struct batadv_priv *bat_priv = orig->bat_priv; 2304 2305 spin_lock_bh(&orig->mcast_handler_lock); 2306 2307 batadv_mcast_want_unsnoop_update(bat_priv, orig, BATADV_NO_FLAGS); 2308 batadv_mcast_want_ipv4_update(bat_priv, orig, BATADV_NO_FLAGS); 2309 batadv_mcast_want_ipv6_update(bat_priv, orig, BATADV_NO_FLAGS); 2310 batadv_mcast_want_rtr4_update(bat_priv, orig, 2311 BATADV_MCAST_WANT_NO_RTR4); 2312 batadv_mcast_want_rtr6_update(bat_priv, orig, 2313 BATADV_MCAST_WANT_NO_RTR6); 2314 2315 spin_unlock_bh(&orig->mcast_handler_lock); 2316 } 2317