1 // SPDX-License-Identifier: GPL-2.0 2 /* Copyright (C) 2011-2020 B.A.T.M.A.N. contributors: 3 * 4 * Simon Wunderlich 5 */ 6 7 #include "bridge_loop_avoidance.h" 8 #include "main.h" 9 10 #include <linux/atomic.h> 11 #include <linux/byteorder/generic.h> 12 #include <linux/compiler.h> 13 #include <linux/crc16.h> 14 #include <linux/errno.h> 15 #include <linux/etherdevice.h> 16 #include <linux/gfp.h> 17 #include <linux/if_arp.h> 18 #include <linux/if_ether.h> 19 #include <linux/if_vlan.h> 20 #include <linux/jhash.h> 21 #include <linux/jiffies.h> 22 #include <linux/kernel.h> 23 #include <linux/kref.h> 24 #include <linux/list.h> 25 #include <linux/lockdep.h> 26 #include <linux/netdevice.h> 27 #include <linux/netlink.h> 28 #include <linux/rculist.h> 29 #include <linux/rcupdate.h> 30 #include <linux/seq_file.h> 31 #include <linux/skbuff.h> 32 #include <linux/slab.h> 33 #include <linux/spinlock.h> 34 #include <linux/stddef.h> 35 #include <linux/string.h> 36 #include <linux/workqueue.h> 37 #include <net/arp.h> 38 #include <net/genetlink.h> 39 #include <net/netlink.h> 40 #include <net/sock.h> 41 #include <uapi/linux/batadv_packet.h> 42 #include <uapi/linux/batman_adv.h> 43 44 #include "hard-interface.h" 45 #include "hash.h" 46 #include "log.h" 47 #include "netlink.h" 48 #include "originator.h" 49 #include "soft-interface.h" 50 #include "translation-table.h" 51 52 static const u8 batadv_announce_mac[4] = {0x43, 0x05, 0x43, 0x05}; 53 54 static void batadv_bla_periodic_work(struct work_struct *work); 55 static void 56 batadv_bla_send_announce(struct batadv_priv *bat_priv, 57 struct batadv_bla_backbone_gw *backbone_gw); 58 59 /** 60 * batadv_choose_claim() - choose the right bucket for a claim. 61 * @data: data to hash 62 * @size: size of the hash table 63 * 64 * Return: the hash index of the claim 65 */ 66 static inline u32 batadv_choose_claim(const void *data, u32 size) 67 { 68 struct batadv_bla_claim *claim = (struct batadv_bla_claim *)data; 69 u32 hash = 0; 70 71 hash = jhash(&claim->addr, sizeof(claim->addr), hash); 72 hash = jhash(&claim->vid, sizeof(claim->vid), hash); 73 74 return hash % size; 75 } 76 77 /** 78 * batadv_choose_backbone_gw() - choose the right bucket for a backbone gateway. 79 * @data: data to hash 80 * @size: size of the hash table 81 * 82 * Return: the hash index of the backbone gateway 83 */ 84 static inline u32 batadv_choose_backbone_gw(const void *data, u32 size) 85 { 86 const struct batadv_bla_claim *claim = (struct batadv_bla_claim *)data; 87 u32 hash = 0; 88 89 hash = jhash(&claim->addr, sizeof(claim->addr), hash); 90 hash = jhash(&claim->vid, sizeof(claim->vid), hash); 91 92 return hash % size; 93 } 94 95 /** 96 * batadv_compare_backbone_gw() - compare address and vid of two backbone gws 97 * @node: list node of the first entry to compare 98 * @data2: pointer to the second backbone gateway 99 * 100 * Return: true if the backbones have the same data, false otherwise 101 */ 102 static bool batadv_compare_backbone_gw(const struct hlist_node *node, 103 const void *data2) 104 { 105 const void *data1 = container_of(node, struct batadv_bla_backbone_gw, 106 hash_entry); 107 const struct batadv_bla_backbone_gw *gw1 = data1; 108 const struct batadv_bla_backbone_gw *gw2 = data2; 109 110 if (!batadv_compare_eth(gw1->orig, gw2->orig)) 111 return false; 112 113 if (gw1->vid != gw2->vid) 114 return false; 115 116 return true; 117 } 118 119 /** 120 * batadv_compare_claim() - compare address and vid of two claims 121 * @node: list node of the first entry to compare 122 * @data2: pointer to the second claims 123 * 124 * Return: true if the claim have the same data, 0 otherwise 125 */ 126 static bool batadv_compare_claim(const struct hlist_node *node, 127 const void *data2) 128 { 129 const void *data1 = container_of(node, struct batadv_bla_claim, 130 hash_entry); 131 const struct batadv_bla_claim *cl1 = data1; 132 const struct batadv_bla_claim *cl2 = data2; 133 134 if (!batadv_compare_eth(cl1->addr, cl2->addr)) 135 return false; 136 137 if (cl1->vid != cl2->vid) 138 return false; 139 140 return true; 141 } 142 143 /** 144 * batadv_backbone_gw_release() - release backbone gw from lists and queue for 145 * free after rcu grace period 146 * @ref: kref pointer of the backbone gw 147 */ 148 static void batadv_backbone_gw_release(struct kref *ref) 149 { 150 struct batadv_bla_backbone_gw *backbone_gw; 151 152 backbone_gw = container_of(ref, struct batadv_bla_backbone_gw, 153 refcount); 154 155 kfree_rcu(backbone_gw, rcu); 156 } 157 158 /** 159 * batadv_backbone_gw_put() - decrement the backbone gw refcounter and possibly 160 * release it 161 * @backbone_gw: backbone gateway to be free'd 162 */ 163 static void batadv_backbone_gw_put(struct batadv_bla_backbone_gw *backbone_gw) 164 { 165 kref_put(&backbone_gw->refcount, batadv_backbone_gw_release); 166 } 167 168 /** 169 * batadv_claim_release() - release claim from lists and queue for free after 170 * rcu grace period 171 * @ref: kref pointer of the claim 172 */ 173 static void batadv_claim_release(struct kref *ref) 174 { 175 struct batadv_bla_claim *claim; 176 struct batadv_bla_backbone_gw *old_backbone_gw; 177 178 claim = container_of(ref, struct batadv_bla_claim, refcount); 179 180 spin_lock_bh(&claim->backbone_lock); 181 old_backbone_gw = claim->backbone_gw; 182 claim->backbone_gw = NULL; 183 spin_unlock_bh(&claim->backbone_lock); 184 185 spin_lock_bh(&old_backbone_gw->crc_lock); 186 old_backbone_gw->crc ^= crc16(0, claim->addr, ETH_ALEN); 187 spin_unlock_bh(&old_backbone_gw->crc_lock); 188 189 batadv_backbone_gw_put(old_backbone_gw); 190 191 kfree_rcu(claim, rcu); 192 } 193 194 /** 195 * batadv_claim_put() - decrement the claim refcounter and possibly release it 196 * @claim: claim to be free'd 197 */ 198 static void batadv_claim_put(struct batadv_bla_claim *claim) 199 { 200 kref_put(&claim->refcount, batadv_claim_release); 201 } 202 203 /** 204 * batadv_claim_hash_find() - looks for a claim in the claim hash 205 * @bat_priv: the bat priv with all the soft interface information 206 * @data: search data (may be local/static data) 207 * 208 * Return: claim if found or NULL otherwise. 209 */ 210 static struct batadv_bla_claim * 211 batadv_claim_hash_find(struct batadv_priv *bat_priv, 212 struct batadv_bla_claim *data) 213 { 214 struct batadv_hashtable *hash = bat_priv->bla.claim_hash; 215 struct hlist_head *head; 216 struct batadv_bla_claim *claim; 217 struct batadv_bla_claim *claim_tmp = NULL; 218 int index; 219 220 if (!hash) 221 return NULL; 222 223 index = batadv_choose_claim(data, hash->size); 224 head = &hash->table[index]; 225 226 rcu_read_lock(); 227 hlist_for_each_entry_rcu(claim, head, hash_entry) { 228 if (!batadv_compare_claim(&claim->hash_entry, data)) 229 continue; 230 231 if (!kref_get_unless_zero(&claim->refcount)) 232 continue; 233 234 claim_tmp = claim; 235 break; 236 } 237 rcu_read_unlock(); 238 239 return claim_tmp; 240 } 241 242 /** 243 * batadv_backbone_hash_find() - looks for a backbone gateway in the hash 244 * @bat_priv: the bat priv with all the soft interface information 245 * @addr: the address of the originator 246 * @vid: the VLAN ID 247 * 248 * Return: backbone gateway if found or NULL otherwise 249 */ 250 static struct batadv_bla_backbone_gw * 251 batadv_backbone_hash_find(struct batadv_priv *bat_priv, u8 *addr, 252 unsigned short vid) 253 { 254 struct batadv_hashtable *hash = bat_priv->bla.backbone_hash; 255 struct hlist_head *head; 256 struct batadv_bla_backbone_gw search_entry, *backbone_gw; 257 struct batadv_bla_backbone_gw *backbone_gw_tmp = NULL; 258 int index; 259 260 if (!hash) 261 return NULL; 262 263 ether_addr_copy(search_entry.orig, addr); 264 search_entry.vid = vid; 265 266 index = batadv_choose_backbone_gw(&search_entry, hash->size); 267 head = &hash->table[index]; 268 269 rcu_read_lock(); 270 hlist_for_each_entry_rcu(backbone_gw, head, hash_entry) { 271 if (!batadv_compare_backbone_gw(&backbone_gw->hash_entry, 272 &search_entry)) 273 continue; 274 275 if (!kref_get_unless_zero(&backbone_gw->refcount)) 276 continue; 277 278 backbone_gw_tmp = backbone_gw; 279 break; 280 } 281 rcu_read_unlock(); 282 283 return backbone_gw_tmp; 284 } 285 286 /** 287 * batadv_bla_del_backbone_claims() - delete all claims for a backbone 288 * @backbone_gw: backbone gateway where the claims should be removed 289 */ 290 static void 291 batadv_bla_del_backbone_claims(struct batadv_bla_backbone_gw *backbone_gw) 292 { 293 struct batadv_hashtable *hash; 294 struct hlist_node *node_tmp; 295 struct hlist_head *head; 296 struct batadv_bla_claim *claim; 297 int i; 298 spinlock_t *list_lock; /* protects write access to the hash lists */ 299 300 hash = backbone_gw->bat_priv->bla.claim_hash; 301 if (!hash) 302 return; 303 304 for (i = 0; i < hash->size; i++) { 305 head = &hash->table[i]; 306 list_lock = &hash->list_locks[i]; 307 308 spin_lock_bh(list_lock); 309 hlist_for_each_entry_safe(claim, node_tmp, 310 head, hash_entry) { 311 if (claim->backbone_gw != backbone_gw) 312 continue; 313 314 batadv_claim_put(claim); 315 hlist_del_rcu(&claim->hash_entry); 316 } 317 spin_unlock_bh(list_lock); 318 } 319 320 /* all claims gone, initialize CRC */ 321 spin_lock_bh(&backbone_gw->crc_lock); 322 backbone_gw->crc = BATADV_BLA_CRC_INIT; 323 spin_unlock_bh(&backbone_gw->crc_lock); 324 } 325 326 /** 327 * batadv_bla_send_claim() - sends a claim frame according to the provided info 328 * @bat_priv: the bat priv with all the soft interface information 329 * @mac: the mac address to be announced within the claim 330 * @vid: the VLAN ID 331 * @claimtype: the type of the claim (CLAIM, UNCLAIM, ANNOUNCE, ...) 332 */ 333 static void batadv_bla_send_claim(struct batadv_priv *bat_priv, u8 *mac, 334 unsigned short vid, int claimtype) 335 { 336 struct sk_buff *skb; 337 struct ethhdr *ethhdr; 338 struct batadv_hard_iface *primary_if; 339 struct net_device *soft_iface; 340 u8 *hw_src; 341 struct batadv_bla_claim_dst local_claim_dest; 342 __be32 zeroip = 0; 343 344 primary_if = batadv_primary_if_get_selected(bat_priv); 345 if (!primary_if) 346 return; 347 348 memcpy(&local_claim_dest, &bat_priv->bla.claim_dest, 349 sizeof(local_claim_dest)); 350 local_claim_dest.type = claimtype; 351 352 soft_iface = primary_if->soft_iface; 353 354 skb = arp_create(ARPOP_REPLY, ETH_P_ARP, 355 /* IP DST: 0.0.0.0 */ 356 zeroip, 357 primary_if->soft_iface, 358 /* IP SRC: 0.0.0.0 */ 359 zeroip, 360 /* Ethernet DST: Broadcast */ 361 NULL, 362 /* Ethernet SRC/HW SRC: originator mac */ 363 primary_if->net_dev->dev_addr, 364 /* HW DST: FF:43:05:XX:YY:YY 365 * with XX = claim type 366 * and YY:YY = group id 367 */ 368 (u8 *)&local_claim_dest); 369 370 if (!skb) 371 goto out; 372 373 ethhdr = (struct ethhdr *)skb->data; 374 hw_src = (u8 *)ethhdr + ETH_HLEN + sizeof(struct arphdr); 375 376 /* now we pretend that the client would have sent this ... */ 377 switch (claimtype) { 378 case BATADV_CLAIM_TYPE_CLAIM: 379 /* normal claim frame 380 * set Ethernet SRC to the clients mac 381 */ 382 ether_addr_copy(ethhdr->h_source, mac); 383 batadv_dbg(BATADV_DBG_BLA, bat_priv, 384 "%s(): CLAIM %pM on vid %d\n", __func__, mac, 385 batadv_print_vid(vid)); 386 break; 387 case BATADV_CLAIM_TYPE_UNCLAIM: 388 /* unclaim frame 389 * set HW SRC to the clients mac 390 */ 391 ether_addr_copy(hw_src, mac); 392 batadv_dbg(BATADV_DBG_BLA, bat_priv, 393 "%s(): UNCLAIM %pM on vid %d\n", __func__, mac, 394 batadv_print_vid(vid)); 395 break; 396 case BATADV_CLAIM_TYPE_ANNOUNCE: 397 /* announcement frame 398 * set HW SRC to the special mac containg the crc 399 */ 400 ether_addr_copy(hw_src, mac); 401 batadv_dbg(BATADV_DBG_BLA, bat_priv, 402 "%s(): ANNOUNCE of %pM on vid %d\n", __func__, 403 ethhdr->h_source, batadv_print_vid(vid)); 404 break; 405 case BATADV_CLAIM_TYPE_REQUEST: 406 /* request frame 407 * set HW SRC and header destination to the receiving backbone 408 * gws mac 409 */ 410 ether_addr_copy(hw_src, mac); 411 ether_addr_copy(ethhdr->h_dest, mac); 412 batadv_dbg(BATADV_DBG_BLA, bat_priv, 413 "%s(): REQUEST of %pM to %pM on vid %d\n", __func__, 414 ethhdr->h_source, ethhdr->h_dest, 415 batadv_print_vid(vid)); 416 break; 417 case BATADV_CLAIM_TYPE_LOOPDETECT: 418 ether_addr_copy(ethhdr->h_source, mac); 419 batadv_dbg(BATADV_DBG_BLA, bat_priv, 420 "%s(): LOOPDETECT of %pM to %pM on vid %d\n", 421 __func__, ethhdr->h_source, ethhdr->h_dest, 422 batadv_print_vid(vid)); 423 424 break; 425 } 426 427 if (vid & BATADV_VLAN_HAS_TAG) { 428 skb = vlan_insert_tag(skb, htons(ETH_P_8021Q), 429 vid & VLAN_VID_MASK); 430 if (!skb) 431 goto out; 432 } 433 434 skb_reset_mac_header(skb); 435 skb->protocol = eth_type_trans(skb, soft_iface); 436 batadv_inc_counter(bat_priv, BATADV_CNT_RX); 437 batadv_add_counter(bat_priv, BATADV_CNT_RX_BYTES, 438 skb->len + ETH_HLEN); 439 440 netif_rx(skb); 441 out: 442 if (primary_if) 443 batadv_hardif_put(primary_if); 444 } 445 446 /** 447 * batadv_bla_loopdetect_report() - worker for reporting the loop 448 * @work: work queue item 449 * 450 * Throws an uevent, as the loopdetect check function can't do that itself 451 * since the kernel may sleep while throwing uevents. 452 */ 453 static void batadv_bla_loopdetect_report(struct work_struct *work) 454 { 455 struct batadv_bla_backbone_gw *backbone_gw; 456 struct batadv_priv *bat_priv; 457 char vid_str[6] = { '\0' }; 458 459 backbone_gw = container_of(work, struct batadv_bla_backbone_gw, 460 report_work); 461 bat_priv = backbone_gw->bat_priv; 462 463 batadv_info(bat_priv->soft_iface, 464 "Possible loop on VLAN %d detected which can't be handled by BLA - please check your network setup!\n", 465 batadv_print_vid(backbone_gw->vid)); 466 snprintf(vid_str, sizeof(vid_str), "%d", 467 batadv_print_vid(backbone_gw->vid)); 468 vid_str[sizeof(vid_str) - 1] = 0; 469 470 batadv_throw_uevent(bat_priv, BATADV_UEV_BLA, BATADV_UEV_LOOPDETECT, 471 vid_str); 472 473 batadv_backbone_gw_put(backbone_gw); 474 } 475 476 /** 477 * batadv_bla_get_backbone_gw() - finds or creates a backbone gateway 478 * @bat_priv: the bat priv with all the soft interface information 479 * @orig: the mac address of the originator 480 * @vid: the VLAN ID 481 * @own_backbone: set if the requested backbone is local 482 * 483 * Return: the (possibly created) backbone gateway or NULL on error 484 */ 485 static struct batadv_bla_backbone_gw * 486 batadv_bla_get_backbone_gw(struct batadv_priv *bat_priv, u8 *orig, 487 unsigned short vid, bool own_backbone) 488 { 489 struct batadv_bla_backbone_gw *entry; 490 struct batadv_orig_node *orig_node; 491 int hash_added; 492 493 entry = batadv_backbone_hash_find(bat_priv, orig, vid); 494 495 if (entry) 496 return entry; 497 498 batadv_dbg(BATADV_DBG_BLA, bat_priv, 499 "%s(): not found (%pM, %d), creating new entry\n", __func__, 500 orig, batadv_print_vid(vid)); 501 502 entry = kzalloc(sizeof(*entry), GFP_ATOMIC); 503 if (!entry) 504 return NULL; 505 506 entry->vid = vid; 507 entry->lasttime = jiffies; 508 entry->crc = BATADV_BLA_CRC_INIT; 509 entry->bat_priv = bat_priv; 510 spin_lock_init(&entry->crc_lock); 511 atomic_set(&entry->request_sent, 0); 512 atomic_set(&entry->wait_periods, 0); 513 ether_addr_copy(entry->orig, orig); 514 INIT_WORK(&entry->report_work, batadv_bla_loopdetect_report); 515 kref_init(&entry->refcount); 516 517 kref_get(&entry->refcount); 518 hash_added = batadv_hash_add(bat_priv->bla.backbone_hash, 519 batadv_compare_backbone_gw, 520 batadv_choose_backbone_gw, entry, 521 &entry->hash_entry); 522 523 if (unlikely(hash_added != 0)) { 524 /* hash failed, free the structure */ 525 kfree(entry); 526 return NULL; 527 } 528 529 /* this is a gateway now, remove any TT entry on this VLAN */ 530 orig_node = batadv_orig_hash_find(bat_priv, orig); 531 if (orig_node) { 532 batadv_tt_global_del_orig(bat_priv, orig_node, vid, 533 "became a backbone gateway"); 534 batadv_orig_node_put(orig_node); 535 } 536 537 if (own_backbone) { 538 batadv_bla_send_announce(bat_priv, entry); 539 540 /* this will be decreased in the worker thread */ 541 atomic_inc(&entry->request_sent); 542 atomic_set(&entry->wait_periods, BATADV_BLA_WAIT_PERIODS); 543 atomic_inc(&bat_priv->bla.num_requests); 544 } 545 546 return entry; 547 } 548 549 /** 550 * batadv_bla_update_own_backbone_gw() - updates the own backbone gw for a VLAN 551 * @bat_priv: the bat priv with all the soft interface information 552 * @primary_if: the selected primary interface 553 * @vid: VLAN identifier 554 * 555 * update or add the own backbone gw to make sure we announce 556 * where we receive other backbone gws 557 */ 558 static void 559 batadv_bla_update_own_backbone_gw(struct batadv_priv *bat_priv, 560 struct batadv_hard_iface *primary_if, 561 unsigned short vid) 562 { 563 struct batadv_bla_backbone_gw *backbone_gw; 564 565 backbone_gw = batadv_bla_get_backbone_gw(bat_priv, 566 primary_if->net_dev->dev_addr, 567 vid, true); 568 if (unlikely(!backbone_gw)) 569 return; 570 571 backbone_gw->lasttime = jiffies; 572 batadv_backbone_gw_put(backbone_gw); 573 } 574 575 /** 576 * batadv_bla_answer_request() - answer a bla request by sending own claims 577 * @bat_priv: the bat priv with all the soft interface information 578 * @primary_if: interface where the request came on 579 * @vid: the vid where the request came on 580 * 581 * Repeat all of our own claims, and finally send an ANNOUNCE frame 582 * to allow the requester another check if the CRC is correct now. 583 */ 584 static void batadv_bla_answer_request(struct batadv_priv *bat_priv, 585 struct batadv_hard_iface *primary_if, 586 unsigned short vid) 587 { 588 struct hlist_head *head; 589 struct batadv_hashtable *hash; 590 struct batadv_bla_claim *claim; 591 struct batadv_bla_backbone_gw *backbone_gw; 592 int i; 593 594 batadv_dbg(BATADV_DBG_BLA, bat_priv, 595 "%s(): received a claim request, send all of our own claims again\n", 596 __func__); 597 598 backbone_gw = batadv_backbone_hash_find(bat_priv, 599 primary_if->net_dev->dev_addr, 600 vid); 601 if (!backbone_gw) 602 return; 603 604 hash = bat_priv->bla.claim_hash; 605 for (i = 0; i < hash->size; i++) { 606 head = &hash->table[i]; 607 608 rcu_read_lock(); 609 hlist_for_each_entry_rcu(claim, head, hash_entry) { 610 /* only own claims are interesting */ 611 if (claim->backbone_gw != backbone_gw) 612 continue; 613 614 batadv_bla_send_claim(bat_priv, claim->addr, claim->vid, 615 BATADV_CLAIM_TYPE_CLAIM); 616 } 617 rcu_read_unlock(); 618 } 619 620 /* finally, send an announcement frame */ 621 batadv_bla_send_announce(bat_priv, backbone_gw); 622 batadv_backbone_gw_put(backbone_gw); 623 } 624 625 /** 626 * batadv_bla_send_request() - send a request to repeat claims 627 * @backbone_gw: the backbone gateway from whom we are out of sync 628 * 629 * When the crc is wrong, ask the backbone gateway for a full table update. 630 * After the request, it will repeat all of his own claims and finally 631 * send an announcement claim with which we can check again. 632 */ 633 static void batadv_bla_send_request(struct batadv_bla_backbone_gw *backbone_gw) 634 { 635 /* first, remove all old entries */ 636 batadv_bla_del_backbone_claims(backbone_gw); 637 638 batadv_dbg(BATADV_DBG_BLA, backbone_gw->bat_priv, 639 "Sending REQUEST to %pM\n", backbone_gw->orig); 640 641 /* send request */ 642 batadv_bla_send_claim(backbone_gw->bat_priv, backbone_gw->orig, 643 backbone_gw->vid, BATADV_CLAIM_TYPE_REQUEST); 644 645 /* no local broadcasts should be sent or received, for now. */ 646 if (!atomic_read(&backbone_gw->request_sent)) { 647 atomic_inc(&backbone_gw->bat_priv->bla.num_requests); 648 atomic_set(&backbone_gw->request_sent, 1); 649 } 650 } 651 652 /** 653 * batadv_bla_send_announce() - Send an announcement frame 654 * @bat_priv: the bat priv with all the soft interface information 655 * @backbone_gw: our backbone gateway which should be announced 656 */ 657 static void batadv_bla_send_announce(struct batadv_priv *bat_priv, 658 struct batadv_bla_backbone_gw *backbone_gw) 659 { 660 u8 mac[ETH_ALEN]; 661 __be16 crc; 662 663 memcpy(mac, batadv_announce_mac, 4); 664 spin_lock_bh(&backbone_gw->crc_lock); 665 crc = htons(backbone_gw->crc); 666 spin_unlock_bh(&backbone_gw->crc_lock); 667 memcpy(&mac[4], &crc, 2); 668 669 batadv_bla_send_claim(bat_priv, mac, backbone_gw->vid, 670 BATADV_CLAIM_TYPE_ANNOUNCE); 671 } 672 673 /** 674 * batadv_bla_add_claim() - Adds a claim in the claim hash 675 * @bat_priv: the bat priv with all the soft interface information 676 * @mac: the mac address of the claim 677 * @vid: the VLAN ID of the frame 678 * @backbone_gw: the backbone gateway which claims it 679 */ 680 static void batadv_bla_add_claim(struct batadv_priv *bat_priv, 681 const u8 *mac, const unsigned short vid, 682 struct batadv_bla_backbone_gw *backbone_gw) 683 { 684 struct batadv_bla_backbone_gw *old_backbone_gw; 685 struct batadv_bla_claim *claim; 686 struct batadv_bla_claim search_claim; 687 bool remove_crc = false; 688 int hash_added; 689 690 ether_addr_copy(search_claim.addr, mac); 691 search_claim.vid = vid; 692 claim = batadv_claim_hash_find(bat_priv, &search_claim); 693 694 /* create a new claim entry if it does not exist yet. */ 695 if (!claim) { 696 claim = kzalloc(sizeof(*claim), GFP_ATOMIC); 697 if (!claim) 698 return; 699 700 ether_addr_copy(claim->addr, mac); 701 spin_lock_init(&claim->backbone_lock); 702 claim->vid = vid; 703 claim->lasttime = jiffies; 704 kref_get(&backbone_gw->refcount); 705 claim->backbone_gw = backbone_gw; 706 kref_init(&claim->refcount); 707 708 batadv_dbg(BATADV_DBG_BLA, bat_priv, 709 "%s(): adding new entry %pM, vid %d to hash ...\n", 710 __func__, mac, batadv_print_vid(vid)); 711 712 kref_get(&claim->refcount); 713 hash_added = batadv_hash_add(bat_priv->bla.claim_hash, 714 batadv_compare_claim, 715 batadv_choose_claim, claim, 716 &claim->hash_entry); 717 718 if (unlikely(hash_added != 0)) { 719 /* only local changes happened. */ 720 kfree(claim); 721 return; 722 } 723 } else { 724 claim->lasttime = jiffies; 725 if (claim->backbone_gw == backbone_gw) 726 /* no need to register a new backbone */ 727 goto claim_free_ref; 728 729 batadv_dbg(BATADV_DBG_BLA, bat_priv, 730 "%s(): changing ownership for %pM, vid %d to gw %pM\n", 731 __func__, mac, batadv_print_vid(vid), 732 backbone_gw->orig); 733 734 remove_crc = true; 735 } 736 737 /* replace backbone_gw atomically and adjust reference counters */ 738 spin_lock_bh(&claim->backbone_lock); 739 old_backbone_gw = claim->backbone_gw; 740 kref_get(&backbone_gw->refcount); 741 claim->backbone_gw = backbone_gw; 742 spin_unlock_bh(&claim->backbone_lock); 743 744 if (remove_crc) { 745 /* remove claim address from old backbone_gw */ 746 spin_lock_bh(&old_backbone_gw->crc_lock); 747 old_backbone_gw->crc ^= crc16(0, claim->addr, ETH_ALEN); 748 spin_unlock_bh(&old_backbone_gw->crc_lock); 749 } 750 751 batadv_backbone_gw_put(old_backbone_gw); 752 753 /* add claim address to new backbone_gw */ 754 spin_lock_bh(&backbone_gw->crc_lock); 755 backbone_gw->crc ^= crc16(0, claim->addr, ETH_ALEN); 756 spin_unlock_bh(&backbone_gw->crc_lock); 757 backbone_gw->lasttime = jiffies; 758 759 claim_free_ref: 760 batadv_claim_put(claim); 761 } 762 763 /** 764 * batadv_bla_claim_get_backbone_gw() - Get valid reference for backbone_gw of 765 * claim 766 * @claim: claim whose backbone_gw should be returned 767 * 768 * Return: valid reference to claim::backbone_gw 769 */ 770 static struct batadv_bla_backbone_gw * 771 batadv_bla_claim_get_backbone_gw(struct batadv_bla_claim *claim) 772 { 773 struct batadv_bla_backbone_gw *backbone_gw; 774 775 spin_lock_bh(&claim->backbone_lock); 776 backbone_gw = claim->backbone_gw; 777 kref_get(&backbone_gw->refcount); 778 spin_unlock_bh(&claim->backbone_lock); 779 780 return backbone_gw; 781 } 782 783 /** 784 * batadv_bla_del_claim() - delete a claim from the claim hash 785 * @bat_priv: the bat priv with all the soft interface information 786 * @mac: mac address of the claim to be removed 787 * @vid: VLAN id for the claim to be removed 788 */ 789 static void batadv_bla_del_claim(struct batadv_priv *bat_priv, 790 const u8 *mac, const unsigned short vid) 791 { 792 struct batadv_bla_claim search_claim, *claim; 793 struct batadv_bla_claim *claim_removed_entry; 794 struct hlist_node *claim_removed_node; 795 796 ether_addr_copy(search_claim.addr, mac); 797 search_claim.vid = vid; 798 claim = batadv_claim_hash_find(bat_priv, &search_claim); 799 if (!claim) 800 return; 801 802 batadv_dbg(BATADV_DBG_BLA, bat_priv, "%s(): %pM, vid %d\n", __func__, 803 mac, batadv_print_vid(vid)); 804 805 claim_removed_node = batadv_hash_remove(bat_priv->bla.claim_hash, 806 batadv_compare_claim, 807 batadv_choose_claim, claim); 808 if (!claim_removed_node) 809 goto free_claim; 810 811 /* reference from the hash is gone */ 812 claim_removed_entry = hlist_entry(claim_removed_node, 813 struct batadv_bla_claim, hash_entry); 814 batadv_claim_put(claim_removed_entry); 815 816 free_claim: 817 /* don't need the reference from hash_find() anymore */ 818 batadv_claim_put(claim); 819 } 820 821 /** 822 * batadv_handle_announce() - check for ANNOUNCE frame 823 * @bat_priv: the bat priv with all the soft interface information 824 * @an_addr: announcement mac address (ARP Sender HW address) 825 * @backbone_addr: originator address of the sender (Ethernet source MAC) 826 * @vid: the VLAN ID of the frame 827 * 828 * Return: true if handled 829 */ 830 static bool batadv_handle_announce(struct batadv_priv *bat_priv, u8 *an_addr, 831 u8 *backbone_addr, unsigned short vid) 832 { 833 struct batadv_bla_backbone_gw *backbone_gw; 834 u16 backbone_crc, crc; 835 836 if (memcmp(an_addr, batadv_announce_mac, 4) != 0) 837 return false; 838 839 backbone_gw = batadv_bla_get_backbone_gw(bat_priv, backbone_addr, vid, 840 false); 841 842 if (unlikely(!backbone_gw)) 843 return true; 844 845 /* handle as ANNOUNCE frame */ 846 backbone_gw->lasttime = jiffies; 847 crc = ntohs(*((__force __be16 *)(&an_addr[4]))); 848 849 batadv_dbg(BATADV_DBG_BLA, bat_priv, 850 "%s(): ANNOUNCE vid %d (sent by %pM)... CRC = %#.4x\n", 851 __func__, batadv_print_vid(vid), backbone_gw->orig, crc); 852 853 spin_lock_bh(&backbone_gw->crc_lock); 854 backbone_crc = backbone_gw->crc; 855 spin_unlock_bh(&backbone_gw->crc_lock); 856 857 if (backbone_crc != crc) { 858 batadv_dbg(BATADV_DBG_BLA, backbone_gw->bat_priv, 859 "%s(): CRC FAILED for %pM/%d (my = %#.4x, sent = %#.4x)\n", 860 __func__, backbone_gw->orig, 861 batadv_print_vid(backbone_gw->vid), 862 backbone_crc, crc); 863 864 batadv_bla_send_request(backbone_gw); 865 } else { 866 /* if we have sent a request and the crc was OK, 867 * we can allow traffic again. 868 */ 869 if (atomic_read(&backbone_gw->request_sent)) { 870 atomic_dec(&backbone_gw->bat_priv->bla.num_requests); 871 atomic_set(&backbone_gw->request_sent, 0); 872 } 873 } 874 875 batadv_backbone_gw_put(backbone_gw); 876 return true; 877 } 878 879 /** 880 * batadv_handle_request() - check for REQUEST frame 881 * @bat_priv: the bat priv with all the soft interface information 882 * @primary_if: the primary hard interface of this batman soft interface 883 * @backbone_addr: backbone address to be requested (ARP sender HW MAC) 884 * @ethhdr: ethernet header of a packet 885 * @vid: the VLAN ID of the frame 886 * 887 * Return: true if handled 888 */ 889 static bool batadv_handle_request(struct batadv_priv *bat_priv, 890 struct batadv_hard_iface *primary_if, 891 u8 *backbone_addr, struct ethhdr *ethhdr, 892 unsigned short vid) 893 { 894 /* check for REQUEST frame */ 895 if (!batadv_compare_eth(backbone_addr, ethhdr->h_dest)) 896 return false; 897 898 /* sanity check, this should not happen on a normal switch, 899 * we ignore it in this case. 900 */ 901 if (!batadv_compare_eth(ethhdr->h_dest, primary_if->net_dev->dev_addr)) 902 return true; 903 904 batadv_dbg(BATADV_DBG_BLA, bat_priv, 905 "%s(): REQUEST vid %d (sent by %pM)...\n", 906 __func__, batadv_print_vid(vid), ethhdr->h_source); 907 908 batadv_bla_answer_request(bat_priv, primary_if, vid); 909 return true; 910 } 911 912 /** 913 * batadv_handle_unclaim() - check for UNCLAIM frame 914 * @bat_priv: the bat priv with all the soft interface information 915 * @primary_if: the primary hard interface of this batman soft interface 916 * @backbone_addr: originator address of the backbone (Ethernet source) 917 * @claim_addr: Client to be unclaimed (ARP sender HW MAC) 918 * @vid: the VLAN ID of the frame 919 * 920 * Return: true if handled 921 */ 922 static bool batadv_handle_unclaim(struct batadv_priv *bat_priv, 923 struct batadv_hard_iface *primary_if, 924 u8 *backbone_addr, u8 *claim_addr, 925 unsigned short vid) 926 { 927 struct batadv_bla_backbone_gw *backbone_gw; 928 929 /* unclaim in any case if it is our own */ 930 if (primary_if && batadv_compare_eth(backbone_addr, 931 primary_if->net_dev->dev_addr)) 932 batadv_bla_send_claim(bat_priv, claim_addr, vid, 933 BATADV_CLAIM_TYPE_UNCLAIM); 934 935 backbone_gw = batadv_backbone_hash_find(bat_priv, backbone_addr, vid); 936 937 if (!backbone_gw) 938 return true; 939 940 /* this must be an UNCLAIM frame */ 941 batadv_dbg(BATADV_DBG_BLA, bat_priv, 942 "%s(): UNCLAIM %pM on vid %d (sent by %pM)...\n", __func__, 943 claim_addr, batadv_print_vid(vid), backbone_gw->orig); 944 945 batadv_bla_del_claim(bat_priv, claim_addr, vid); 946 batadv_backbone_gw_put(backbone_gw); 947 return true; 948 } 949 950 /** 951 * batadv_handle_claim() - check for CLAIM frame 952 * @bat_priv: the bat priv with all the soft interface information 953 * @primary_if: the primary hard interface of this batman soft interface 954 * @backbone_addr: originator address of the backbone (Ethernet Source) 955 * @claim_addr: client mac address to be claimed (ARP sender HW MAC) 956 * @vid: the VLAN ID of the frame 957 * 958 * Return: true if handled 959 */ 960 static bool batadv_handle_claim(struct batadv_priv *bat_priv, 961 struct batadv_hard_iface *primary_if, 962 u8 *backbone_addr, u8 *claim_addr, 963 unsigned short vid) 964 { 965 struct batadv_bla_backbone_gw *backbone_gw; 966 967 /* register the gateway if not yet available, and add the claim. */ 968 969 backbone_gw = batadv_bla_get_backbone_gw(bat_priv, backbone_addr, vid, 970 false); 971 972 if (unlikely(!backbone_gw)) 973 return true; 974 975 /* this must be a CLAIM frame */ 976 batadv_bla_add_claim(bat_priv, claim_addr, vid, backbone_gw); 977 if (batadv_compare_eth(backbone_addr, primary_if->net_dev->dev_addr)) 978 batadv_bla_send_claim(bat_priv, claim_addr, vid, 979 BATADV_CLAIM_TYPE_CLAIM); 980 981 /* TODO: we could call something like tt_local_del() here. */ 982 983 batadv_backbone_gw_put(backbone_gw); 984 return true; 985 } 986 987 /** 988 * batadv_check_claim_group() - check for claim group membership 989 * @bat_priv: the bat priv with all the soft interface information 990 * @primary_if: the primary interface of this batman interface 991 * @hw_src: the Hardware source in the ARP Header 992 * @hw_dst: the Hardware destination in the ARP Header 993 * @ethhdr: pointer to the Ethernet header of the claim frame 994 * 995 * checks if it is a claim packet and if its on the same group. 996 * This function also applies the group ID of the sender 997 * if it is in the same mesh. 998 * 999 * Return: 1000 * 2 - if it is a claim packet and on the same group 1001 * 1 - if is a claim packet from another group 1002 * 0 - if it is not a claim packet 1003 */ 1004 static int batadv_check_claim_group(struct batadv_priv *bat_priv, 1005 struct batadv_hard_iface *primary_if, 1006 u8 *hw_src, u8 *hw_dst, 1007 struct ethhdr *ethhdr) 1008 { 1009 u8 *backbone_addr; 1010 struct batadv_orig_node *orig_node; 1011 struct batadv_bla_claim_dst *bla_dst, *bla_dst_own; 1012 1013 bla_dst = (struct batadv_bla_claim_dst *)hw_dst; 1014 bla_dst_own = &bat_priv->bla.claim_dest; 1015 1016 /* if announcement packet, use the source, 1017 * otherwise assume it is in the hw_src 1018 */ 1019 switch (bla_dst->type) { 1020 case BATADV_CLAIM_TYPE_CLAIM: 1021 backbone_addr = hw_src; 1022 break; 1023 case BATADV_CLAIM_TYPE_REQUEST: 1024 case BATADV_CLAIM_TYPE_ANNOUNCE: 1025 case BATADV_CLAIM_TYPE_UNCLAIM: 1026 backbone_addr = ethhdr->h_source; 1027 break; 1028 default: 1029 return 0; 1030 } 1031 1032 /* don't accept claim frames from ourselves */ 1033 if (batadv_compare_eth(backbone_addr, primary_if->net_dev->dev_addr)) 1034 return 0; 1035 1036 /* if its already the same group, it is fine. */ 1037 if (bla_dst->group == bla_dst_own->group) 1038 return 2; 1039 1040 /* lets see if this originator is in our mesh */ 1041 orig_node = batadv_orig_hash_find(bat_priv, backbone_addr); 1042 1043 /* dont accept claims from gateways which are not in 1044 * the same mesh or group. 1045 */ 1046 if (!orig_node) 1047 return 1; 1048 1049 /* if our mesh friends mac is bigger, use it for ourselves. */ 1050 if (ntohs(bla_dst->group) > ntohs(bla_dst_own->group)) { 1051 batadv_dbg(BATADV_DBG_BLA, bat_priv, 1052 "taking other backbones claim group: %#.4x\n", 1053 ntohs(bla_dst->group)); 1054 bla_dst_own->group = bla_dst->group; 1055 } 1056 1057 batadv_orig_node_put(orig_node); 1058 1059 return 2; 1060 } 1061 1062 /** 1063 * batadv_bla_process_claim() - Check if this is a claim frame, and process it 1064 * @bat_priv: the bat priv with all the soft interface information 1065 * @primary_if: the primary hard interface of this batman soft interface 1066 * @skb: the frame to be checked 1067 * 1068 * Return: true if it was a claim frame, otherwise return false to 1069 * tell the callee that it can use the frame on its own. 1070 */ 1071 static bool batadv_bla_process_claim(struct batadv_priv *bat_priv, 1072 struct batadv_hard_iface *primary_if, 1073 struct sk_buff *skb) 1074 { 1075 struct batadv_bla_claim_dst *bla_dst, *bla_dst_own; 1076 u8 *hw_src, *hw_dst; 1077 struct vlan_hdr *vhdr, vhdr_buf; 1078 struct ethhdr *ethhdr; 1079 struct arphdr *arphdr; 1080 unsigned short vid; 1081 int vlan_depth = 0; 1082 __be16 proto; 1083 int headlen; 1084 int ret; 1085 1086 vid = batadv_get_vid(skb, 0); 1087 ethhdr = eth_hdr(skb); 1088 1089 proto = ethhdr->h_proto; 1090 headlen = ETH_HLEN; 1091 if (vid & BATADV_VLAN_HAS_TAG) { 1092 /* Traverse the VLAN/Ethertypes. 1093 * 1094 * At this point it is known that the first protocol is a VLAN 1095 * header, so start checking at the encapsulated protocol. 1096 * 1097 * The depth of the VLAN headers is recorded to drop BLA claim 1098 * frames encapsulated into multiple VLAN headers (QinQ). 1099 */ 1100 do { 1101 vhdr = skb_header_pointer(skb, headlen, VLAN_HLEN, 1102 &vhdr_buf); 1103 if (!vhdr) 1104 return false; 1105 1106 proto = vhdr->h_vlan_encapsulated_proto; 1107 headlen += VLAN_HLEN; 1108 vlan_depth++; 1109 } while (proto == htons(ETH_P_8021Q)); 1110 } 1111 1112 if (proto != htons(ETH_P_ARP)) 1113 return false; /* not a claim frame */ 1114 1115 /* this must be a ARP frame. check if it is a claim. */ 1116 1117 if (unlikely(!pskb_may_pull(skb, headlen + arp_hdr_len(skb->dev)))) 1118 return false; 1119 1120 /* pskb_may_pull() may have modified the pointers, get ethhdr again */ 1121 ethhdr = eth_hdr(skb); 1122 arphdr = (struct arphdr *)((u8 *)ethhdr + headlen); 1123 1124 /* Check whether the ARP frame carries a valid 1125 * IP information 1126 */ 1127 if (arphdr->ar_hrd != htons(ARPHRD_ETHER)) 1128 return false; 1129 if (arphdr->ar_pro != htons(ETH_P_IP)) 1130 return false; 1131 if (arphdr->ar_hln != ETH_ALEN) 1132 return false; 1133 if (arphdr->ar_pln != 4) 1134 return false; 1135 1136 hw_src = (u8 *)arphdr + sizeof(struct arphdr); 1137 hw_dst = hw_src + ETH_ALEN + 4; 1138 bla_dst = (struct batadv_bla_claim_dst *)hw_dst; 1139 bla_dst_own = &bat_priv->bla.claim_dest; 1140 1141 /* check if it is a claim frame in general */ 1142 if (memcmp(bla_dst->magic, bla_dst_own->magic, 1143 sizeof(bla_dst->magic)) != 0) 1144 return false; 1145 1146 /* check if there is a claim frame encapsulated deeper in (QinQ) and 1147 * drop that, as this is not supported by BLA but should also not be 1148 * sent via the mesh. 1149 */ 1150 if (vlan_depth > 1) 1151 return true; 1152 1153 /* Let the loopdetect frames on the mesh in any case. */ 1154 if (bla_dst->type == BATADV_CLAIM_TYPE_LOOPDETECT) 1155 return false; 1156 1157 /* check if it is a claim frame. */ 1158 ret = batadv_check_claim_group(bat_priv, primary_if, hw_src, hw_dst, 1159 ethhdr); 1160 if (ret == 1) 1161 batadv_dbg(BATADV_DBG_BLA, bat_priv, 1162 "%s(): received a claim frame from another group. From: %pM on vid %d ...(hw_src %pM, hw_dst %pM)\n", 1163 __func__, ethhdr->h_source, batadv_print_vid(vid), 1164 hw_src, hw_dst); 1165 1166 if (ret < 2) 1167 return !!ret; 1168 1169 /* become a backbone gw ourselves on this vlan if not happened yet */ 1170 batadv_bla_update_own_backbone_gw(bat_priv, primary_if, vid); 1171 1172 /* check for the different types of claim frames ... */ 1173 switch (bla_dst->type) { 1174 case BATADV_CLAIM_TYPE_CLAIM: 1175 if (batadv_handle_claim(bat_priv, primary_if, hw_src, 1176 ethhdr->h_source, vid)) 1177 return true; 1178 break; 1179 case BATADV_CLAIM_TYPE_UNCLAIM: 1180 if (batadv_handle_unclaim(bat_priv, primary_if, 1181 ethhdr->h_source, hw_src, vid)) 1182 return true; 1183 break; 1184 1185 case BATADV_CLAIM_TYPE_ANNOUNCE: 1186 if (batadv_handle_announce(bat_priv, hw_src, ethhdr->h_source, 1187 vid)) 1188 return true; 1189 break; 1190 case BATADV_CLAIM_TYPE_REQUEST: 1191 if (batadv_handle_request(bat_priv, primary_if, hw_src, ethhdr, 1192 vid)) 1193 return true; 1194 break; 1195 } 1196 1197 batadv_dbg(BATADV_DBG_BLA, bat_priv, 1198 "%s(): ERROR - this looks like a claim frame, but is useless. eth src %pM on vid %d ...(hw_src %pM, hw_dst %pM)\n", 1199 __func__, ethhdr->h_source, batadv_print_vid(vid), hw_src, 1200 hw_dst); 1201 return true; 1202 } 1203 1204 /** 1205 * batadv_bla_purge_backbone_gw() - Remove backbone gateways after a timeout or 1206 * immediately 1207 * @bat_priv: the bat priv with all the soft interface information 1208 * @now: whether the whole hash shall be wiped now 1209 * 1210 * Check when we last heard from other nodes, and remove them in case of 1211 * a time out, or clean all backbone gws if now is set. 1212 */ 1213 static void batadv_bla_purge_backbone_gw(struct batadv_priv *bat_priv, int now) 1214 { 1215 struct batadv_bla_backbone_gw *backbone_gw; 1216 struct hlist_node *node_tmp; 1217 struct hlist_head *head; 1218 struct batadv_hashtable *hash; 1219 spinlock_t *list_lock; /* protects write access to the hash lists */ 1220 int i; 1221 1222 hash = bat_priv->bla.backbone_hash; 1223 if (!hash) 1224 return; 1225 1226 for (i = 0; i < hash->size; i++) { 1227 head = &hash->table[i]; 1228 list_lock = &hash->list_locks[i]; 1229 1230 spin_lock_bh(list_lock); 1231 hlist_for_each_entry_safe(backbone_gw, node_tmp, 1232 head, hash_entry) { 1233 if (now) 1234 goto purge_now; 1235 if (!batadv_has_timed_out(backbone_gw->lasttime, 1236 BATADV_BLA_BACKBONE_TIMEOUT)) 1237 continue; 1238 1239 batadv_dbg(BATADV_DBG_BLA, backbone_gw->bat_priv, 1240 "%s(): backbone gw %pM timed out\n", 1241 __func__, backbone_gw->orig); 1242 1243 purge_now: 1244 /* don't wait for the pending request anymore */ 1245 if (atomic_read(&backbone_gw->request_sent)) 1246 atomic_dec(&bat_priv->bla.num_requests); 1247 1248 batadv_bla_del_backbone_claims(backbone_gw); 1249 1250 hlist_del_rcu(&backbone_gw->hash_entry); 1251 batadv_backbone_gw_put(backbone_gw); 1252 } 1253 spin_unlock_bh(list_lock); 1254 } 1255 } 1256 1257 /** 1258 * batadv_bla_purge_claims() - Remove claims after a timeout or immediately 1259 * @bat_priv: the bat priv with all the soft interface information 1260 * @primary_if: the selected primary interface, may be NULL if now is set 1261 * @now: whether the whole hash shall be wiped now 1262 * 1263 * Check when we heard last time from our own claims, and remove them in case of 1264 * a time out, or clean all claims if now is set 1265 */ 1266 static void batadv_bla_purge_claims(struct batadv_priv *bat_priv, 1267 struct batadv_hard_iface *primary_if, 1268 int now) 1269 { 1270 struct batadv_bla_backbone_gw *backbone_gw; 1271 struct batadv_bla_claim *claim; 1272 struct hlist_head *head; 1273 struct batadv_hashtable *hash; 1274 int i; 1275 1276 hash = bat_priv->bla.claim_hash; 1277 if (!hash) 1278 return; 1279 1280 for (i = 0; i < hash->size; i++) { 1281 head = &hash->table[i]; 1282 1283 rcu_read_lock(); 1284 hlist_for_each_entry_rcu(claim, head, hash_entry) { 1285 backbone_gw = batadv_bla_claim_get_backbone_gw(claim); 1286 if (now) 1287 goto purge_now; 1288 1289 if (!batadv_compare_eth(backbone_gw->orig, 1290 primary_if->net_dev->dev_addr)) 1291 goto skip; 1292 1293 if (!batadv_has_timed_out(claim->lasttime, 1294 BATADV_BLA_CLAIM_TIMEOUT)) 1295 goto skip; 1296 1297 batadv_dbg(BATADV_DBG_BLA, bat_priv, 1298 "%s(): timed out.\n", __func__); 1299 1300 purge_now: 1301 batadv_dbg(BATADV_DBG_BLA, bat_priv, 1302 "%s(): %pM, vid %d\n", __func__, 1303 claim->addr, claim->vid); 1304 1305 batadv_handle_unclaim(bat_priv, primary_if, 1306 backbone_gw->orig, 1307 claim->addr, claim->vid); 1308 skip: 1309 batadv_backbone_gw_put(backbone_gw); 1310 } 1311 rcu_read_unlock(); 1312 } 1313 } 1314 1315 /** 1316 * batadv_bla_update_orig_address() - Update the backbone gateways when the own 1317 * originator address changes 1318 * @bat_priv: the bat priv with all the soft interface information 1319 * @primary_if: the new selected primary_if 1320 * @oldif: the old primary interface, may be NULL 1321 */ 1322 void batadv_bla_update_orig_address(struct batadv_priv *bat_priv, 1323 struct batadv_hard_iface *primary_if, 1324 struct batadv_hard_iface *oldif) 1325 { 1326 struct batadv_bla_backbone_gw *backbone_gw; 1327 struct hlist_head *head; 1328 struct batadv_hashtable *hash; 1329 __be16 group; 1330 int i; 1331 1332 /* reset bridge loop avoidance group id */ 1333 group = htons(crc16(0, primary_if->net_dev->dev_addr, ETH_ALEN)); 1334 bat_priv->bla.claim_dest.group = group; 1335 1336 /* purge everything when bridge loop avoidance is turned off */ 1337 if (!atomic_read(&bat_priv->bridge_loop_avoidance)) 1338 oldif = NULL; 1339 1340 if (!oldif) { 1341 batadv_bla_purge_claims(bat_priv, NULL, 1); 1342 batadv_bla_purge_backbone_gw(bat_priv, 1); 1343 return; 1344 } 1345 1346 hash = bat_priv->bla.backbone_hash; 1347 if (!hash) 1348 return; 1349 1350 for (i = 0; i < hash->size; i++) { 1351 head = &hash->table[i]; 1352 1353 rcu_read_lock(); 1354 hlist_for_each_entry_rcu(backbone_gw, head, hash_entry) { 1355 /* own orig still holds the old value. */ 1356 if (!batadv_compare_eth(backbone_gw->orig, 1357 oldif->net_dev->dev_addr)) 1358 continue; 1359 1360 ether_addr_copy(backbone_gw->orig, 1361 primary_if->net_dev->dev_addr); 1362 /* send an announce frame so others will ask for our 1363 * claims and update their tables. 1364 */ 1365 batadv_bla_send_announce(bat_priv, backbone_gw); 1366 } 1367 rcu_read_unlock(); 1368 } 1369 } 1370 1371 /** 1372 * batadv_bla_send_loopdetect() - send a loopdetect frame 1373 * @bat_priv: the bat priv with all the soft interface information 1374 * @backbone_gw: the backbone gateway for which a loop should be detected 1375 * 1376 * To detect loops that the bridge loop avoidance can't handle, send a loop 1377 * detection packet on the backbone. Unlike other BLA frames, this frame will 1378 * be allowed on the mesh by other nodes. If it is received on the mesh, this 1379 * indicates that there is a loop. 1380 */ 1381 static void 1382 batadv_bla_send_loopdetect(struct batadv_priv *bat_priv, 1383 struct batadv_bla_backbone_gw *backbone_gw) 1384 { 1385 batadv_dbg(BATADV_DBG_BLA, bat_priv, "Send loopdetect frame for vid %d\n", 1386 backbone_gw->vid); 1387 batadv_bla_send_claim(bat_priv, bat_priv->bla.loopdetect_addr, 1388 backbone_gw->vid, BATADV_CLAIM_TYPE_LOOPDETECT); 1389 } 1390 1391 /** 1392 * batadv_bla_status_update() - purge bla interfaces if necessary 1393 * @net_dev: the soft interface net device 1394 */ 1395 void batadv_bla_status_update(struct net_device *net_dev) 1396 { 1397 struct batadv_priv *bat_priv = netdev_priv(net_dev); 1398 struct batadv_hard_iface *primary_if; 1399 1400 primary_if = batadv_primary_if_get_selected(bat_priv); 1401 if (!primary_if) 1402 return; 1403 1404 /* this function already purges everything when bla is disabled, 1405 * so just call that one. 1406 */ 1407 batadv_bla_update_orig_address(bat_priv, primary_if, primary_if); 1408 batadv_hardif_put(primary_if); 1409 } 1410 1411 /** 1412 * batadv_bla_periodic_work() - performs periodic bla work 1413 * @work: kernel work struct 1414 * 1415 * periodic work to do: 1416 * * purge structures when they are too old 1417 * * send announcements 1418 */ 1419 static void batadv_bla_periodic_work(struct work_struct *work) 1420 { 1421 struct delayed_work *delayed_work; 1422 struct batadv_priv *bat_priv; 1423 struct batadv_priv_bla *priv_bla; 1424 struct hlist_head *head; 1425 struct batadv_bla_backbone_gw *backbone_gw; 1426 struct batadv_hashtable *hash; 1427 struct batadv_hard_iface *primary_if; 1428 bool send_loopdetect = false; 1429 int i; 1430 1431 delayed_work = to_delayed_work(work); 1432 priv_bla = container_of(delayed_work, struct batadv_priv_bla, work); 1433 bat_priv = container_of(priv_bla, struct batadv_priv, bla); 1434 primary_if = batadv_primary_if_get_selected(bat_priv); 1435 if (!primary_if) 1436 goto out; 1437 1438 batadv_bla_purge_claims(bat_priv, primary_if, 0); 1439 batadv_bla_purge_backbone_gw(bat_priv, 0); 1440 1441 if (!atomic_read(&bat_priv->bridge_loop_avoidance)) 1442 goto out; 1443 1444 if (atomic_dec_and_test(&bat_priv->bla.loopdetect_next)) { 1445 /* set a new random mac address for the next bridge loop 1446 * detection frames. Set the locally administered bit to avoid 1447 * collisions with users mac addresses. 1448 */ 1449 eth_random_addr(bat_priv->bla.loopdetect_addr); 1450 bat_priv->bla.loopdetect_addr[0] = 0xba; 1451 bat_priv->bla.loopdetect_addr[1] = 0xbe; 1452 bat_priv->bla.loopdetect_lasttime = jiffies; 1453 atomic_set(&bat_priv->bla.loopdetect_next, 1454 BATADV_BLA_LOOPDETECT_PERIODS); 1455 1456 /* mark for sending loop detect on all VLANs */ 1457 send_loopdetect = true; 1458 } 1459 1460 hash = bat_priv->bla.backbone_hash; 1461 if (!hash) 1462 goto out; 1463 1464 for (i = 0; i < hash->size; i++) { 1465 head = &hash->table[i]; 1466 1467 rcu_read_lock(); 1468 hlist_for_each_entry_rcu(backbone_gw, head, hash_entry) { 1469 if (!batadv_compare_eth(backbone_gw->orig, 1470 primary_if->net_dev->dev_addr)) 1471 continue; 1472 1473 backbone_gw->lasttime = jiffies; 1474 1475 batadv_bla_send_announce(bat_priv, backbone_gw); 1476 if (send_loopdetect) 1477 batadv_bla_send_loopdetect(bat_priv, 1478 backbone_gw); 1479 1480 /* request_sent is only set after creation to avoid 1481 * problems when we are not yet known as backbone gw 1482 * in the backbone. 1483 * 1484 * We can reset this now after we waited some periods 1485 * to give bridge forward delays and bla group forming 1486 * some grace time. 1487 */ 1488 1489 if (atomic_read(&backbone_gw->request_sent) == 0) 1490 continue; 1491 1492 if (!atomic_dec_and_test(&backbone_gw->wait_periods)) 1493 continue; 1494 1495 atomic_dec(&backbone_gw->bat_priv->bla.num_requests); 1496 atomic_set(&backbone_gw->request_sent, 0); 1497 } 1498 rcu_read_unlock(); 1499 } 1500 out: 1501 if (primary_if) 1502 batadv_hardif_put(primary_if); 1503 1504 queue_delayed_work(batadv_event_workqueue, &bat_priv->bla.work, 1505 msecs_to_jiffies(BATADV_BLA_PERIOD_LENGTH)); 1506 } 1507 1508 /* The hash for claim and backbone hash receive the same key because they 1509 * are getting initialized by hash_new with the same key. Reinitializing 1510 * them with to different keys to allow nested locking without generating 1511 * lockdep warnings 1512 */ 1513 static struct lock_class_key batadv_claim_hash_lock_class_key; 1514 static struct lock_class_key batadv_backbone_hash_lock_class_key; 1515 1516 /** 1517 * batadv_bla_init() - initialize all bla structures 1518 * @bat_priv: the bat priv with all the soft interface information 1519 * 1520 * Return: 0 on success, < 0 on error. 1521 */ 1522 int batadv_bla_init(struct batadv_priv *bat_priv) 1523 { 1524 int i; 1525 u8 claim_dest[ETH_ALEN] = {0xff, 0x43, 0x05, 0x00, 0x00, 0x00}; 1526 struct batadv_hard_iface *primary_if; 1527 u16 crc; 1528 unsigned long entrytime; 1529 1530 spin_lock_init(&bat_priv->bla.bcast_duplist_lock); 1531 1532 batadv_dbg(BATADV_DBG_BLA, bat_priv, "bla hash registering\n"); 1533 1534 /* setting claim destination address */ 1535 memcpy(&bat_priv->bla.claim_dest.magic, claim_dest, 3); 1536 bat_priv->bla.claim_dest.type = 0; 1537 primary_if = batadv_primary_if_get_selected(bat_priv); 1538 if (primary_if) { 1539 crc = crc16(0, primary_if->net_dev->dev_addr, ETH_ALEN); 1540 bat_priv->bla.claim_dest.group = htons(crc); 1541 batadv_hardif_put(primary_if); 1542 } else { 1543 bat_priv->bla.claim_dest.group = 0; /* will be set later */ 1544 } 1545 1546 /* initialize the duplicate list */ 1547 entrytime = jiffies - msecs_to_jiffies(BATADV_DUPLIST_TIMEOUT); 1548 for (i = 0; i < BATADV_DUPLIST_SIZE; i++) 1549 bat_priv->bla.bcast_duplist[i].entrytime = entrytime; 1550 bat_priv->bla.bcast_duplist_curr = 0; 1551 1552 atomic_set(&bat_priv->bla.loopdetect_next, 1553 BATADV_BLA_LOOPDETECT_PERIODS); 1554 1555 if (bat_priv->bla.claim_hash) 1556 return 0; 1557 1558 bat_priv->bla.claim_hash = batadv_hash_new(128); 1559 bat_priv->bla.backbone_hash = batadv_hash_new(32); 1560 1561 if (!bat_priv->bla.claim_hash || !bat_priv->bla.backbone_hash) 1562 return -ENOMEM; 1563 1564 batadv_hash_set_lock_class(bat_priv->bla.claim_hash, 1565 &batadv_claim_hash_lock_class_key); 1566 batadv_hash_set_lock_class(bat_priv->bla.backbone_hash, 1567 &batadv_backbone_hash_lock_class_key); 1568 1569 batadv_dbg(BATADV_DBG_BLA, bat_priv, "bla hashes initialized\n"); 1570 1571 INIT_DELAYED_WORK(&bat_priv->bla.work, batadv_bla_periodic_work); 1572 1573 queue_delayed_work(batadv_event_workqueue, &bat_priv->bla.work, 1574 msecs_to_jiffies(BATADV_BLA_PERIOD_LENGTH)); 1575 return 0; 1576 } 1577 1578 /** 1579 * batadv_bla_check_bcast_duplist() - Check if a frame is in the broadcast dup. 1580 * @bat_priv: the bat priv with all the soft interface information 1581 * @skb: contains the bcast_packet to be checked 1582 * 1583 * check if it is on our broadcast list. Another gateway might 1584 * have sent the same packet because it is connected to the same backbone, 1585 * so we have to remove this duplicate. 1586 * 1587 * This is performed by checking the CRC, which will tell us 1588 * with a good chance that it is the same packet. If it is furthermore 1589 * sent by another host, drop it. We allow equal packets from 1590 * the same host however as this might be intended. 1591 * 1592 * Return: true if a packet is in the duplicate list, false otherwise. 1593 */ 1594 bool batadv_bla_check_bcast_duplist(struct batadv_priv *bat_priv, 1595 struct sk_buff *skb) 1596 { 1597 int i, curr; 1598 __be32 crc; 1599 struct batadv_bcast_packet *bcast_packet; 1600 struct batadv_bcast_duplist_entry *entry; 1601 bool ret = false; 1602 1603 bcast_packet = (struct batadv_bcast_packet *)skb->data; 1604 1605 /* calculate the crc ... */ 1606 crc = batadv_skb_crc32(skb, (u8 *)(bcast_packet + 1)); 1607 1608 spin_lock_bh(&bat_priv->bla.bcast_duplist_lock); 1609 1610 for (i = 0; i < BATADV_DUPLIST_SIZE; i++) { 1611 curr = (bat_priv->bla.bcast_duplist_curr + i); 1612 curr %= BATADV_DUPLIST_SIZE; 1613 entry = &bat_priv->bla.bcast_duplist[curr]; 1614 1615 /* we can stop searching if the entry is too old ; 1616 * later entries will be even older 1617 */ 1618 if (batadv_has_timed_out(entry->entrytime, 1619 BATADV_DUPLIST_TIMEOUT)) 1620 break; 1621 1622 if (entry->crc != crc) 1623 continue; 1624 1625 if (batadv_compare_eth(entry->orig, bcast_packet->orig)) 1626 continue; 1627 1628 /* this entry seems to match: same crc, not too old, 1629 * and from another gw. therefore return true to forbid it. 1630 */ 1631 ret = true; 1632 goto out; 1633 } 1634 /* not found, add a new entry (overwrite the oldest entry) 1635 * and allow it, its the first occurrence. 1636 */ 1637 curr = (bat_priv->bla.bcast_duplist_curr + BATADV_DUPLIST_SIZE - 1); 1638 curr %= BATADV_DUPLIST_SIZE; 1639 entry = &bat_priv->bla.bcast_duplist[curr]; 1640 entry->crc = crc; 1641 entry->entrytime = jiffies; 1642 ether_addr_copy(entry->orig, bcast_packet->orig); 1643 bat_priv->bla.bcast_duplist_curr = curr; 1644 1645 out: 1646 spin_unlock_bh(&bat_priv->bla.bcast_duplist_lock); 1647 1648 return ret; 1649 } 1650 1651 /** 1652 * batadv_bla_is_backbone_gw_orig() - Check if the originator is a gateway for 1653 * the VLAN identified by vid. 1654 * @bat_priv: the bat priv with all the soft interface information 1655 * @orig: originator mac address 1656 * @vid: VLAN identifier 1657 * 1658 * Return: true if orig is a backbone for this vid, false otherwise. 1659 */ 1660 bool batadv_bla_is_backbone_gw_orig(struct batadv_priv *bat_priv, u8 *orig, 1661 unsigned short vid) 1662 { 1663 struct batadv_hashtable *hash = bat_priv->bla.backbone_hash; 1664 struct hlist_head *head; 1665 struct batadv_bla_backbone_gw *backbone_gw; 1666 int i; 1667 1668 if (!atomic_read(&bat_priv->bridge_loop_avoidance)) 1669 return false; 1670 1671 if (!hash) 1672 return false; 1673 1674 for (i = 0; i < hash->size; i++) { 1675 head = &hash->table[i]; 1676 1677 rcu_read_lock(); 1678 hlist_for_each_entry_rcu(backbone_gw, head, hash_entry) { 1679 if (batadv_compare_eth(backbone_gw->orig, orig) && 1680 backbone_gw->vid == vid) { 1681 rcu_read_unlock(); 1682 return true; 1683 } 1684 } 1685 rcu_read_unlock(); 1686 } 1687 1688 return false; 1689 } 1690 1691 /** 1692 * batadv_bla_is_backbone_gw() - check if originator is a backbone gw for a VLAN 1693 * @skb: the frame to be checked 1694 * @orig_node: the orig_node of the frame 1695 * @hdr_size: maximum length of the frame 1696 * 1697 * Return: true if the orig_node is also a gateway on the soft interface, 1698 * otherwise it returns false. 1699 */ 1700 bool batadv_bla_is_backbone_gw(struct sk_buff *skb, 1701 struct batadv_orig_node *orig_node, int hdr_size) 1702 { 1703 struct batadv_bla_backbone_gw *backbone_gw; 1704 unsigned short vid; 1705 1706 if (!atomic_read(&orig_node->bat_priv->bridge_loop_avoidance)) 1707 return false; 1708 1709 /* first, find out the vid. */ 1710 if (!pskb_may_pull(skb, hdr_size + ETH_HLEN)) 1711 return false; 1712 1713 vid = batadv_get_vid(skb, hdr_size); 1714 1715 /* see if this originator is a backbone gw for this VLAN */ 1716 backbone_gw = batadv_backbone_hash_find(orig_node->bat_priv, 1717 orig_node->orig, vid); 1718 if (!backbone_gw) 1719 return false; 1720 1721 batadv_backbone_gw_put(backbone_gw); 1722 return true; 1723 } 1724 1725 /** 1726 * batadv_bla_free() - free all bla structures 1727 * @bat_priv: the bat priv with all the soft interface information 1728 * 1729 * for softinterface free or module unload 1730 */ 1731 void batadv_bla_free(struct batadv_priv *bat_priv) 1732 { 1733 struct batadv_hard_iface *primary_if; 1734 1735 cancel_delayed_work_sync(&bat_priv->bla.work); 1736 primary_if = batadv_primary_if_get_selected(bat_priv); 1737 1738 if (bat_priv->bla.claim_hash) { 1739 batadv_bla_purge_claims(bat_priv, primary_if, 1); 1740 batadv_hash_destroy(bat_priv->bla.claim_hash); 1741 bat_priv->bla.claim_hash = NULL; 1742 } 1743 if (bat_priv->bla.backbone_hash) { 1744 batadv_bla_purge_backbone_gw(bat_priv, 1); 1745 batadv_hash_destroy(bat_priv->bla.backbone_hash); 1746 bat_priv->bla.backbone_hash = NULL; 1747 } 1748 if (primary_if) 1749 batadv_hardif_put(primary_if); 1750 } 1751 1752 /** 1753 * batadv_bla_loopdetect_check() - check and handle a detected loop 1754 * @bat_priv: the bat priv with all the soft interface information 1755 * @skb: the packet to check 1756 * @primary_if: interface where the request came on 1757 * @vid: the VLAN ID of the frame 1758 * 1759 * Checks if this packet is a loop detect frame which has been sent by us, 1760 * throw an uevent and log the event if that is the case. 1761 * 1762 * Return: true if it is a loop detect frame which is to be dropped, false 1763 * otherwise. 1764 */ 1765 static bool 1766 batadv_bla_loopdetect_check(struct batadv_priv *bat_priv, struct sk_buff *skb, 1767 struct batadv_hard_iface *primary_if, 1768 unsigned short vid) 1769 { 1770 struct batadv_bla_backbone_gw *backbone_gw; 1771 struct ethhdr *ethhdr; 1772 bool ret; 1773 1774 ethhdr = eth_hdr(skb); 1775 1776 /* Only check for the MAC address and skip more checks here for 1777 * performance reasons - this function is on the hotpath, after all. 1778 */ 1779 if (!batadv_compare_eth(ethhdr->h_source, 1780 bat_priv->bla.loopdetect_addr)) 1781 return false; 1782 1783 /* If the packet came too late, don't forward it on the mesh 1784 * but don't consider that as loop. It might be a coincidence. 1785 */ 1786 if (batadv_has_timed_out(bat_priv->bla.loopdetect_lasttime, 1787 BATADV_BLA_LOOPDETECT_TIMEOUT)) 1788 return true; 1789 1790 backbone_gw = batadv_bla_get_backbone_gw(bat_priv, 1791 primary_if->net_dev->dev_addr, 1792 vid, true); 1793 if (unlikely(!backbone_gw)) 1794 return true; 1795 1796 ret = queue_work(batadv_event_workqueue, &backbone_gw->report_work); 1797 1798 /* backbone_gw is unreferenced in the report work function function 1799 * if queue_work() call was successful 1800 */ 1801 if (!ret) 1802 batadv_backbone_gw_put(backbone_gw); 1803 1804 return true; 1805 } 1806 1807 /** 1808 * batadv_bla_rx() - check packets coming from the mesh. 1809 * @bat_priv: the bat priv with all the soft interface information 1810 * @skb: the frame to be checked 1811 * @vid: the VLAN ID of the frame 1812 * @is_bcast: the packet came in a broadcast packet type. 1813 * 1814 * batadv_bla_rx avoidance checks if: 1815 * * we have to race for a claim 1816 * * if the frame is allowed on the LAN 1817 * 1818 * in these cases, the skb is further handled by this function 1819 * 1820 * Return: true if handled, otherwise it returns false and the caller shall 1821 * further process the skb. 1822 */ 1823 bool batadv_bla_rx(struct batadv_priv *bat_priv, struct sk_buff *skb, 1824 unsigned short vid, bool is_bcast) 1825 { 1826 struct batadv_bla_backbone_gw *backbone_gw; 1827 struct ethhdr *ethhdr; 1828 struct batadv_bla_claim search_claim, *claim = NULL; 1829 struct batadv_hard_iface *primary_if; 1830 bool own_claim; 1831 bool ret; 1832 1833 ethhdr = eth_hdr(skb); 1834 1835 primary_if = batadv_primary_if_get_selected(bat_priv); 1836 if (!primary_if) 1837 goto handled; 1838 1839 if (!atomic_read(&bat_priv->bridge_loop_avoidance)) 1840 goto allow; 1841 1842 if (batadv_bla_loopdetect_check(bat_priv, skb, primary_if, vid)) 1843 goto handled; 1844 1845 if (unlikely(atomic_read(&bat_priv->bla.num_requests))) 1846 /* don't allow broadcasts while requests are in flight */ 1847 if (is_multicast_ether_addr(ethhdr->h_dest) && is_bcast) 1848 goto handled; 1849 1850 ether_addr_copy(search_claim.addr, ethhdr->h_source); 1851 search_claim.vid = vid; 1852 claim = batadv_claim_hash_find(bat_priv, &search_claim); 1853 1854 if (!claim) { 1855 /* possible optimization: race for a claim */ 1856 /* No claim exists yet, claim it for us! 1857 */ 1858 1859 batadv_dbg(BATADV_DBG_BLA, bat_priv, 1860 "%s(): Unclaimed MAC %pM found. Claim it. Local: %s\n", 1861 __func__, ethhdr->h_source, 1862 batadv_is_my_client(bat_priv, 1863 ethhdr->h_source, vid) ? 1864 "yes" : "no"); 1865 batadv_handle_claim(bat_priv, primary_if, 1866 primary_if->net_dev->dev_addr, 1867 ethhdr->h_source, vid); 1868 goto allow; 1869 } 1870 1871 /* if it is our own claim ... */ 1872 backbone_gw = batadv_bla_claim_get_backbone_gw(claim); 1873 own_claim = batadv_compare_eth(backbone_gw->orig, 1874 primary_if->net_dev->dev_addr); 1875 batadv_backbone_gw_put(backbone_gw); 1876 1877 if (own_claim) { 1878 /* ... allow it in any case */ 1879 claim->lasttime = jiffies; 1880 goto allow; 1881 } 1882 1883 /* if it is a broadcast ... */ 1884 if (is_multicast_ether_addr(ethhdr->h_dest) && is_bcast) { 1885 /* ... drop it. the responsible gateway is in charge. 1886 * 1887 * We need to check is_bcast because with the gateway 1888 * feature, broadcasts (like DHCP requests) may be sent 1889 * using a unicast packet type. 1890 */ 1891 goto handled; 1892 } else { 1893 /* seems the client considers us as its best gateway. 1894 * send a claim and update the claim table 1895 * immediately. 1896 */ 1897 batadv_handle_claim(bat_priv, primary_if, 1898 primary_if->net_dev->dev_addr, 1899 ethhdr->h_source, vid); 1900 goto allow; 1901 } 1902 allow: 1903 batadv_bla_update_own_backbone_gw(bat_priv, primary_if, vid); 1904 ret = false; 1905 goto out; 1906 1907 handled: 1908 kfree_skb(skb); 1909 ret = true; 1910 1911 out: 1912 if (primary_if) 1913 batadv_hardif_put(primary_if); 1914 if (claim) 1915 batadv_claim_put(claim); 1916 return ret; 1917 } 1918 1919 /** 1920 * batadv_bla_tx() - check packets going into the mesh 1921 * @bat_priv: the bat priv with all the soft interface information 1922 * @skb: the frame to be checked 1923 * @vid: the VLAN ID of the frame 1924 * 1925 * batadv_bla_tx checks if: 1926 * * a claim was received which has to be processed 1927 * * the frame is allowed on the mesh 1928 * 1929 * in these cases, the skb is further handled by this function. 1930 * 1931 * This call might reallocate skb data. 1932 * 1933 * Return: true if handled, otherwise it returns false and the caller shall 1934 * further process the skb. 1935 */ 1936 bool batadv_bla_tx(struct batadv_priv *bat_priv, struct sk_buff *skb, 1937 unsigned short vid) 1938 { 1939 struct ethhdr *ethhdr; 1940 struct batadv_bla_claim search_claim, *claim = NULL; 1941 struct batadv_bla_backbone_gw *backbone_gw; 1942 struct batadv_hard_iface *primary_if; 1943 bool client_roamed; 1944 bool ret = false; 1945 1946 primary_if = batadv_primary_if_get_selected(bat_priv); 1947 if (!primary_if) 1948 goto out; 1949 1950 if (!atomic_read(&bat_priv->bridge_loop_avoidance)) 1951 goto allow; 1952 1953 if (batadv_bla_process_claim(bat_priv, primary_if, skb)) 1954 goto handled; 1955 1956 ethhdr = eth_hdr(skb); 1957 1958 if (unlikely(atomic_read(&bat_priv->bla.num_requests))) 1959 /* don't allow broadcasts while requests are in flight */ 1960 if (is_multicast_ether_addr(ethhdr->h_dest)) 1961 goto handled; 1962 1963 ether_addr_copy(search_claim.addr, ethhdr->h_source); 1964 search_claim.vid = vid; 1965 1966 claim = batadv_claim_hash_find(bat_priv, &search_claim); 1967 1968 /* if no claim exists, allow it. */ 1969 if (!claim) 1970 goto allow; 1971 1972 /* check if we are responsible. */ 1973 backbone_gw = batadv_bla_claim_get_backbone_gw(claim); 1974 client_roamed = batadv_compare_eth(backbone_gw->orig, 1975 primary_if->net_dev->dev_addr); 1976 batadv_backbone_gw_put(backbone_gw); 1977 1978 if (client_roamed) { 1979 /* if yes, the client has roamed and we have 1980 * to unclaim it. 1981 */ 1982 if (batadv_has_timed_out(claim->lasttime, 100)) { 1983 /* only unclaim if the last claim entry is 1984 * older than 100 ms to make sure we really 1985 * have a roaming client here. 1986 */ 1987 batadv_dbg(BATADV_DBG_BLA, bat_priv, "%s(): Roaming client %pM detected. Unclaim it.\n", 1988 __func__, ethhdr->h_source); 1989 batadv_handle_unclaim(bat_priv, primary_if, 1990 primary_if->net_dev->dev_addr, 1991 ethhdr->h_source, vid); 1992 goto allow; 1993 } else { 1994 batadv_dbg(BATADV_DBG_BLA, bat_priv, "%s(): Race for claim %pM detected. Drop packet.\n", 1995 __func__, ethhdr->h_source); 1996 goto handled; 1997 } 1998 } 1999 2000 /* check if it is a multicast/broadcast frame */ 2001 if (is_multicast_ether_addr(ethhdr->h_dest)) { 2002 /* drop it. the responsible gateway has forwarded it into 2003 * the backbone network. 2004 */ 2005 goto handled; 2006 } else { 2007 /* we must allow it. at least if we are 2008 * responsible for the DESTINATION. 2009 */ 2010 goto allow; 2011 } 2012 allow: 2013 batadv_bla_update_own_backbone_gw(bat_priv, primary_if, vid); 2014 ret = false; 2015 goto out; 2016 handled: 2017 ret = true; 2018 out: 2019 if (primary_if) 2020 batadv_hardif_put(primary_if); 2021 if (claim) 2022 batadv_claim_put(claim); 2023 return ret; 2024 } 2025 2026 #ifdef CONFIG_BATMAN_ADV_DEBUGFS 2027 /** 2028 * batadv_bla_claim_table_seq_print_text() - print the claim table in a seq file 2029 * @seq: seq file to print on 2030 * @offset: not used 2031 * 2032 * Return: always 0 2033 */ 2034 int batadv_bla_claim_table_seq_print_text(struct seq_file *seq, void *offset) 2035 { 2036 struct net_device *net_dev = (struct net_device *)seq->private; 2037 struct batadv_priv *bat_priv = netdev_priv(net_dev); 2038 struct batadv_hashtable *hash = bat_priv->bla.claim_hash; 2039 struct batadv_bla_backbone_gw *backbone_gw; 2040 struct batadv_bla_claim *claim; 2041 struct batadv_hard_iface *primary_if; 2042 struct hlist_head *head; 2043 u16 backbone_crc; 2044 u32 i; 2045 bool is_own; 2046 u8 *primary_addr; 2047 2048 primary_if = batadv_seq_print_text_primary_if_get(seq); 2049 if (!primary_if) 2050 goto out; 2051 2052 primary_addr = primary_if->net_dev->dev_addr; 2053 seq_printf(seq, 2054 "Claims announced for the mesh %s (orig %pM, group id %#.4x)\n", 2055 net_dev->name, primary_addr, 2056 ntohs(bat_priv->bla.claim_dest.group)); 2057 seq_puts(seq, 2058 " Client VID Originator [o] (CRC )\n"); 2059 for (i = 0; i < hash->size; i++) { 2060 head = &hash->table[i]; 2061 2062 rcu_read_lock(); 2063 hlist_for_each_entry_rcu(claim, head, hash_entry) { 2064 backbone_gw = batadv_bla_claim_get_backbone_gw(claim); 2065 2066 is_own = batadv_compare_eth(backbone_gw->orig, 2067 primary_addr); 2068 2069 spin_lock_bh(&backbone_gw->crc_lock); 2070 backbone_crc = backbone_gw->crc; 2071 spin_unlock_bh(&backbone_gw->crc_lock); 2072 seq_printf(seq, " * %pM on %5d by %pM [%c] (%#.4x)\n", 2073 claim->addr, batadv_print_vid(claim->vid), 2074 backbone_gw->orig, 2075 (is_own ? 'x' : ' '), 2076 backbone_crc); 2077 2078 batadv_backbone_gw_put(backbone_gw); 2079 } 2080 rcu_read_unlock(); 2081 } 2082 out: 2083 if (primary_if) 2084 batadv_hardif_put(primary_if); 2085 return 0; 2086 } 2087 #endif 2088 2089 /** 2090 * batadv_bla_claim_dump_entry() - dump one entry of the claim table 2091 * to a netlink socket 2092 * @msg: buffer for the message 2093 * @portid: netlink port 2094 * @cb: Control block containing additional options 2095 * @primary_if: primary interface 2096 * @claim: entry to dump 2097 * 2098 * Return: 0 or error code. 2099 */ 2100 static int 2101 batadv_bla_claim_dump_entry(struct sk_buff *msg, u32 portid, 2102 struct netlink_callback *cb, 2103 struct batadv_hard_iface *primary_if, 2104 struct batadv_bla_claim *claim) 2105 { 2106 u8 *primary_addr = primary_if->net_dev->dev_addr; 2107 u16 backbone_crc; 2108 bool is_own; 2109 void *hdr; 2110 int ret = -EINVAL; 2111 2112 hdr = genlmsg_put(msg, portid, cb->nlh->nlmsg_seq, 2113 &batadv_netlink_family, NLM_F_MULTI, 2114 BATADV_CMD_GET_BLA_CLAIM); 2115 if (!hdr) { 2116 ret = -ENOBUFS; 2117 goto out; 2118 } 2119 2120 genl_dump_check_consistent(cb, hdr); 2121 2122 is_own = batadv_compare_eth(claim->backbone_gw->orig, 2123 primary_addr); 2124 2125 spin_lock_bh(&claim->backbone_gw->crc_lock); 2126 backbone_crc = claim->backbone_gw->crc; 2127 spin_unlock_bh(&claim->backbone_gw->crc_lock); 2128 2129 if (is_own) 2130 if (nla_put_flag(msg, BATADV_ATTR_BLA_OWN)) { 2131 genlmsg_cancel(msg, hdr); 2132 goto out; 2133 } 2134 2135 if (nla_put(msg, BATADV_ATTR_BLA_ADDRESS, ETH_ALEN, claim->addr) || 2136 nla_put_u16(msg, BATADV_ATTR_BLA_VID, claim->vid) || 2137 nla_put(msg, BATADV_ATTR_BLA_BACKBONE, ETH_ALEN, 2138 claim->backbone_gw->orig) || 2139 nla_put_u16(msg, BATADV_ATTR_BLA_CRC, 2140 backbone_crc)) { 2141 genlmsg_cancel(msg, hdr); 2142 goto out; 2143 } 2144 2145 genlmsg_end(msg, hdr); 2146 ret = 0; 2147 2148 out: 2149 return ret; 2150 } 2151 2152 /** 2153 * batadv_bla_claim_dump_bucket() - dump one bucket of the claim table 2154 * to a netlink socket 2155 * @msg: buffer for the message 2156 * @portid: netlink port 2157 * @cb: Control block containing additional options 2158 * @primary_if: primary interface 2159 * @hash: hash to dump 2160 * @bucket: bucket index to dump 2161 * @idx_skip: How many entries to skip 2162 * 2163 * Return: always 0. 2164 */ 2165 static int 2166 batadv_bla_claim_dump_bucket(struct sk_buff *msg, u32 portid, 2167 struct netlink_callback *cb, 2168 struct batadv_hard_iface *primary_if, 2169 struct batadv_hashtable *hash, unsigned int bucket, 2170 int *idx_skip) 2171 { 2172 struct batadv_bla_claim *claim; 2173 int idx = 0; 2174 int ret = 0; 2175 2176 spin_lock_bh(&hash->list_locks[bucket]); 2177 cb->seq = atomic_read(&hash->generation) << 1 | 1; 2178 2179 hlist_for_each_entry(claim, &hash->table[bucket], hash_entry) { 2180 if (idx++ < *idx_skip) 2181 continue; 2182 2183 ret = batadv_bla_claim_dump_entry(msg, portid, cb, 2184 primary_if, claim); 2185 if (ret) { 2186 *idx_skip = idx - 1; 2187 goto unlock; 2188 } 2189 } 2190 2191 *idx_skip = 0; 2192 unlock: 2193 spin_unlock_bh(&hash->list_locks[bucket]); 2194 return ret; 2195 } 2196 2197 /** 2198 * batadv_bla_claim_dump() - dump claim table to a netlink socket 2199 * @msg: buffer for the message 2200 * @cb: callback structure containing arguments 2201 * 2202 * Return: message length. 2203 */ 2204 int batadv_bla_claim_dump(struct sk_buff *msg, struct netlink_callback *cb) 2205 { 2206 struct batadv_hard_iface *primary_if = NULL; 2207 int portid = NETLINK_CB(cb->skb).portid; 2208 struct net *net = sock_net(cb->skb->sk); 2209 struct net_device *soft_iface; 2210 struct batadv_hashtable *hash; 2211 struct batadv_priv *bat_priv; 2212 int bucket = cb->args[0]; 2213 int idx = cb->args[1]; 2214 int ifindex; 2215 int ret = 0; 2216 2217 ifindex = batadv_netlink_get_ifindex(cb->nlh, 2218 BATADV_ATTR_MESH_IFINDEX); 2219 if (!ifindex) 2220 return -EINVAL; 2221 2222 soft_iface = dev_get_by_index(net, ifindex); 2223 if (!soft_iface || !batadv_softif_is_valid(soft_iface)) { 2224 ret = -ENODEV; 2225 goto out; 2226 } 2227 2228 bat_priv = netdev_priv(soft_iface); 2229 hash = bat_priv->bla.claim_hash; 2230 2231 primary_if = batadv_primary_if_get_selected(bat_priv); 2232 if (!primary_if || primary_if->if_status != BATADV_IF_ACTIVE) { 2233 ret = -ENOENT; 2234 goto out; 2235 } 2236 2237 while (bucket < hash->size) { 2238 if (batadv_bla_claim_dump_bucket(msg, portid, cb, primary_if, 2239 hash, bucket, &idx)) 2240 break; 2241 bucket++; 2242 } 2243 2244 cb->args[0] = bucket; 2245 cb->args[1] = idx; 2246 2247 ret = msg->len; 2248 2249 out: 2250 if (primary_if) 2251 batadv_hardif_put(primary_if); 2252 2253 if (soft_iface) 2254 dev_put(soft_iface); 2255 2256 return ret; 2257 } 2258 2259 #ifdef CONFIG_BATMAN_ADV_DEBUGFS 2260 /** 2261 * batadv_bla_backbone_table_seq_print_text() - print the backbone table in a 2262 * seq file 2263 * @seq: seq file to print on 2264 * @offset: not used 2265 * 2266 * Return: always 0 2267 */ 2268 int batadv_bla_backbone_table_seq_print_text(struct seq_file *seq, void *offset) 2269 { 2270 struct net_device *net_dev = (struct net_device *)seq->private; 2271 struct batadv_priv *bat_priv = netdev_priv(net_dev); 2272 struct batadv_hashtable *hash = bat_priv->bla.backbone_hash; 2273 struct batadv_bla_backbone_gw *backbone_gw; 2274 struct batadv_hard_iface *primary_if; 2275 struct hlist_head *head; 2276 int secs, msecs; 2277 u16 backbone_crc; 2278 u32 i; 2279 bool is_own; 2280 u8 *primary_addr; 2281 2282 primary_if = batadv_seq_print_text_primary_if_get(seq); 2283 if (!primary_if) 2284 goto out; 2285 2286 primary_addr = primary_if->net_dev->dev_addr; 2287 seq_printf(seq, 2288 "Backbones announced for the mesh %s (orig %pM, group id %#.4x)\n", 2289 net_dev->name, primary_addr, 2290 ntohs(bat_priv->bla.claim_dest.group)); 2291 seq_puts(seq, " Originator VID last seen (CRC )\n"); 2292 for (i = 0; i < hash->size; i++) { 2293 head = &hash->table[i]; 2294 2295 rcu_read_lock(); 2296 hlist_for_each_entry_rcu(backbone_gw, head, hash_entry) { 2297 msecs = jiffies_to_msecs(jiffies - 2298 backbone_gw->lasttime); 2299 secs = msecs / 1000; 2300 msecs = msecs % 1000; 2301 2302 is_own = batadv_compare_eth(backbone_gw->orig, 2303 primary_addr); 2304 if (is_own) 2305 continue; 2306 2307 spin_lock_bh(&backbone_gw->crc_lock); 2308 backbone_crc = backbone_gw->crc; 2309 spin_unlock_bh(&backbone_gw->crc_lock); 2310 2311 seq_printf(seq, " * %pM on %5d %4i.%03is (%#.4x)\n", 2312 backbone_gw->orig, 2313 batadv_print_vid(backbone_gw->vid), secs, 2314 msecs, backbone_crc); 2315 } 2316 rcu_read_unlock(); 2317 } 2318 out: 2319 if (primary_if) 2320 batadv_hardif_put(primary_if); 2321 return 0; 2322 } 2323 #endif 2324 2325 /** 2326 * batadv_bla_backbone_dump_entry() - dump one entry of the backbone table to a 2327 * netlink socket 2328 * @msg: buffer for the message 2329 * @portid: netlink port 2330 * @cb: Control block containing additional options 2331 * @primary_if: primary interface 2332 * @backbone_gw: entry to dump 2333 * 2334 * Return: 0 or error code. 2335 */ 2336 static int 2337 batadv_bla_backbone_dump_entry(struct sk_buff *msg, u32 portid, 2338 struct netlink_callback *cb, 2339 struct batadv_hard_iface *primary_if, 2340 struct batadv_bla_backbone_gw *backbone_gw) 2341 { 2342 u8 *primary_addr = primary_if->net_dev->dev_addr; 2343 u16 backbone_crc; 2344 bool is_own; 2345 int msecs; 2346 void *hdr; 2347 int ret = -EINVAL; 2348 2349 hdr = genlmsg_put(msg, portid, cb->nlh->nlmsg_seq, 2350 &batadv_netlink_family, NLM_F_MULTI, 2351 BATADV_CMD_GET_BLA_BACKBONE); 2352 if (!hdr) { 2353 ret = -ENOBUFS; 2354 goto out; 2355 } 2356 2357 genl_dump_check_consistent(cb, hdr); 2358 2359 is_own = batadv_compare_eth(backbone_gw->orig, primary_addr); 2360 2361 spin_lock_bh(&backbone_gw->crc_lock); 2362 backbone_crc = backbone_gw->crc; 2363 spin_unlock_bh(&backbone_gw->crc_lock); 2364 2365 msecs = jiffies_to_msecs(jiffies - backbone_gw->lasttime); 2366 2367 if (is_own) 2368 if (nla_put_flag(msg, BATADV_ATTR_BLA_OWN)) { 2369 genlmsg_cancel(msg, hdr); 2370 goto out; 2371 } 2372 2373 if (nla_put(msg, BATADV_ATTR_BLA_BACKBONE, ETH_ALEN, 2374 backbone_gw->orig) || 2375 nla_put_u16(msg, BATADV_ATTR_BLA_VID, backbone_gw->vid) || 2376 nla_put_u16(msg, BATADV_ATTR_BLA_CRC, 2377 backbone_crc) || 2378 nla_put_u32(msg, BATADV_ATTR_LAST_SEEN_MSECS, msecs)) { 2379 genlmsg_cancel(msg, hdr); 2380 goto out; 2381 } 2382 2383 genlmsg_end(msg, hdr); 2384 ret = 0; 2385 2386 out: 2387 return ret; 2388 } 2389 2390 /** 2391 * batadv_bla_backbone_dump_bucket() - dump one bucket of the backbone table to 2392 * a netlink socket 2393 * @msg: buffer for the message 2394 * @portid: netlink port 2395 * @cb: Control block containing additional options 2396 * @primary_if: primary interface 2397 * @hash: hash to dump 2398 * @bucket: bucket index to dump 2399 * @idx_skip: How many entries to skip 2400 * 2401 * Return: always 0. 2402 */ 2403 static int 2404 batadv_bla_backbone_dump_bucket(struct sk_buff *msg, u32 portid, 2405 struct netlink_callback *cb, 2406 struct batadv_hard_iface *primary_if, 2407 struct batadv_hashtable *hash, 2408 unsigned int bucket, int *idx_skip) 2409 { 2410 struct batadv_bla_backbone_gw *backbone_gw; 2411 int idx = 0; 2412 int ret = 0; 2413 2414 spin_lock_bh(&hash->list_locks[bucket]); 2415 cb->seq = atomic_read(&hash->generation) << 1 | 1; 2416 2417 hlist_for_each_entry(backbone_gw, &hash->table[bucket], hash_entry) { 2418 if (idx++ < *idx_skip) 2419 continue; 2420 2421 ret = batadv_bla_backbone_dump_entry(msg, portid, cb, 2422 primary_if, backbone_gw); 2423 if (ret) { 2424 *idx_skip = idx - 1; 2425 goto unlock; 2426 } 2427 } 2428 2429 *idx_skip = 0; 2430 unlock: 2431 spin_unlock_bh(&hash->list_locks[bucket]); 2432 return ret; 2433 } 2434 2435 /** 2436 * batadv_bla_backbone_dump() - dump backbone table to a netlink socket 2437 * @msg: buffer for the message 2438 * @cb: callback structure containing arguments 2439 * 2440 * Return: message length. 2441 */ 2442 int batadv_bla_backbone_dump(struct sk_buff *msg, struct netlink_callback *cb) 2443 { 2444 struct batadv_hard_iface *primary_if = NULL; 2445 int portid = NETLINK_CB(cb->skb).portid; 2446 struct net *net = sock_net(cb->skb->sk); 2447 struct net_device *soft_iface; 2448 struct batadv_hashtable *hash; 2449 struct batadv_priv *bat_priv; 2450 int bucket = cb->args[0]; 2451 int idx = cb->args[1]; 2452 int ifindex; 2453 int ret = 0; 2454 2455 ifindex = batadv_netlink_get_ifindex(cb->nlh, 2456 BATADV_ATTR_MESH_IFINDEX); 2457 if (!ifindex) 2458 return -EINVAL; 2459 2460 soft_iface = dev_get_by_index(net, ifindex); 2461 if (!soft_iface || !batadv_softif_is_valid(soft_iface)) { 2462 ret = -ENODEV; 2463 goto out; 2464 } 2465 2466 bat_priv = netdev_priv(soft_iface); 2467 hash = bat_priv->bla.backbone_hash; 2468 2469 primary_if = batadv_primary_if_get_selected(bat_priv); 2470 if (!primary_if || primary_if->if_status != BATADV_IF_ACTIVE) { 2471 ret = -ENOENT; 2472 goto out; 2473 } 2474 2475 while (bucket < hash->size) { 2476 if (batadv_bla_backbone_dump_bucket(msg, portid, cb, primary_if, 2477 hash, bucket, &idx)) 2478 break; 2479 bucket++; 2480 } 2481 2482 cb->args[0] = bucket; 2483 cb->args[1] = idx; 2484 2485 ret = msg->len; 2486 2487 out: 2488 if (primary_if) 2489 batadv_hardif_put(primary_if); 2490 2491 if (soft_iface) 2492 dev_put(soft_iface); 2493 2494 return ret; 2495 } 2496 2497 #ifdef CONFIG_BATMAN_ADV_DAT 2498 /** 2499 * batadv_bla_check_claim() - check if address is claimed 2500 * 2501 * @bat_priv: the bat priv with all the soft interface information 2502 * @addr: mac address of which the claim status is checked 2503 * @vid: the VLAN ID 2504 * 2505 * addr is checked if this address is claimed by the local device itself. 2506 * 2507 * Return: true if bla is disabled or the mac is claimed by the device, 2508 * false if the device addr is already claimed by another gateway 2509 */ 2510 bool batadv_bla_check_claim(struct batadv_priv *bat_priv, 2511 u8 *addr, unsigned short vid) 2512 { 2513 struct batadv_bla_claim search_claim; 2514 struct batadv_bla_claim *claim = NULL; 2515 struct batadv_hard_iface *primary_if = NULL; 2516 bool ret = true; 2517 2518 if (!atomic_read(&bat_priv->bridge_loop_avoidance)) 2519 return ret; 2520 2521 primary_if = batadv_primary_if_get_selected(bat_priv); 2522 if (!primary_if) 2523 return ret; 2524 2525 /* First look if the mac address is claimed */ 2526 ether_addr_copy(search_claim.addr, addr); 2527 search_claim.vid = vid; 2528 2529 claim = batadv_claim_hash_find(bat_priv, &search_claim); 2530 2531 /* If there is a claim and we are not owner of the claim, 2532 * return false. 2533 */ 2534 if (claim) { 2535 if (!batadv_compare_eth(claim->backbone_gw->orig, 2536 primary_if->net_dev->dev_addr)) 2537 ret = false; 2538 batadv_claim_put(claim); 2539 } 2540 2541 batadv_hardif_put(primary_if); 2542 return ret; 2543 } 2544 #endif 2545