1 /* 2 * Bridge netlink control interface 3 * 4 * Authors: 5 * Stephen Hemminger <shemminger@osdl.org> 6 * 7 * This program is free software; you can redistribute it and/or 8 * modify it under the terms of the GNU General Public License 9 * as published by the Free Software Foundation; either version 10 * 2 of the License, or (at your option) any later version. 11 */ 12 13 #include <linux/kernel.h> 14 #include <linux/slab.h> 15 #include <linux/etherdevice.h> 16 #include <net/rtnetlink.h> 17 #include <net/net_namespace.h> 18 #include <net/sock.h> 19 #include <uapi/linux/if_bridge.h> 20 21 #include "br_private.h" 22 #include "br_private_stp.h" 23 #include "br_private_tunnel.h" 24 25 static int __get_num_vlan_infos(struct net_bridge_vlan_group *vg, 26 u32 filter_mask) 27 { 28 struct net_bridge_vlan *v; 29 u16 vid_range_start = 0, vid_range_end = 0, vid_range_flags = 0; 30 u16 flags, pvid; 31 int num_vlans = 0; 32 33 if (!(filter_mask & RTEXT_FILTER_BRVLAN_COMPRESSED)) 34 return 0; 35 36 pvid = br_get_pvid(vg); 37 /* Count number of vlan infos */ 38 list_for_each_entry_rcu(v, &vg->vlan_list, vlist) { 39 flags = 0; 40 /* only a context, bridge vlan not activated */ 41 if (!br_vlan_should_use(v)) 42 continue; 43 if (v->vid == pvid) 44 flags |= BRIDGE_VLAN_INFO_PVID; 45 46 if (v->flags & BRIDGE_VLAN_INFO_UNTAGGED) 47 flags |= BRIDGE_VLAN_INFO_UNTAGGED; 48 49 if (vid_range_start == 0) { 50 goto initvars; 51 } else if ((v->vid - vid_range_end) == 1 && 52 flags == vid_range_flags) { 53 vid_range_end = v->vid; 54 continue; 55 } else { 56 if ((vid_range_end - vid_range_start) > 0) 57 num_vlans += 2; 58 else 59 num_vlans += 1; 60 } 61 initvars: 62 vid_range_start = v->vid; 63 vid_range_end = v->vid; 64 vid_range_flags = flags; 65 } 66 67 if (vid_range_start != 0) { 68 if ((vid_range_end - vid_range_start) > 0) 69 num_vlans += 2; 70 else 71 num_vlans += 1; 72 } 73 74 return num_vlans; 75 } 76 77 static int br_get_num_vlan_infos(struct net_bridge_vlan_group *vg, 78 u32 filter_mask) 79 { 80 int num_vlans; 81 82 if (!vg) 83 return 0; 84 85 if (filter_mask & RTEXT_FILTER_BRVLAN) 86 return vg->num_vlans; 87 88 rcu_read_lock(); 89 num_vlans = __get_num_vlan_infos(vg, filter_mask); 90 rcu_read_unlock(); 91 92 return num_vlans; 93 } 94 95 static size_t br_get_link_af_size_filtered(const struct net_device *dev, 96 u32 filter_mask) 97 { 98 struct net_bridge_vlan_group *vg = NULL; 99 struct net_bridge_port *p = NULL; 100 struct net_bridge *br; 101 int num_vlan_infos; 102 size_t vinfo_sz = 0; 103 104 rcu_read_lock(); 105 if (br_port_exists(dev)) { 106 p = br_port_get_rcu(dev); 107 vg = nbp_vlan_group_rcu(p); 108 } else if (dev->priv_flags & IFF_EBRIDGE) { 109 br = netdev_priv(dev); 110 vg = br_vlan_group_rcu(br); 111 } 112 num_vlan_infos = br_get_num_vlan_infos(vg, filter_mask); 113 rcu_read_unlock(); 114 115 if (p && (p->flags & BR_VLAN_TUNNEL)) 116 vinfo_sz += br_get_vlan_tunnel_info_size(vg); 117 118 /* Each VLAN is returned in bridge_vlan_info along with flags */ 119 vinfo_sz += num_vlan_infos * nla_total_size(sizeof(struct bridge_vlan_info)); 120 121 return vinfo_sz; 122 } 123 124 static inline size_t br_port_info_size(void) 125 { 126 return nla_total_size(1) /* IFLA_BRPORT_STATE */ 127 + nla_total_size(2) /* IFLA_BRPORT_PRIORITY */ 128 + nla_total_size(4) /* IFLA_BRPORT_COST */ 129 + nla_total_size(1) /* IFLA_BRPORT_MODE */ 130 + nla_total_size(1) /* IFLA_BRPORT_GUARD */ 131 + nla_total_size(1) /* IFLA_BRPORT_PROTECT */ 132 + nla_total_size(1) /* IFLA_BRPORT_FAST_LEAVE */ 133 + nla_total_size(1) /* IFLA_BRPORT_MCAST_TO_UCAST */ 134 + nla_total_size(1) /* IFLA_BRPORT_LEARNING */ 135 + nla_total_size(1) /* IFLA_BRPORT_UNICAST_FLOOD */ 136 + nla_total_size(1) /* IFLA_BRPORT_MCAST_FLOOD */ 137 + nla_total_size(1) /* IFLA_BRPORT_BCAST_FLOOD */ 138 + nla_total_size(1) /* IFLA_BRPORT_PROXYARP */ 139 + nla_total_size(1) /* IFLA_BRPORT_PROXYARP_WIFI */ 140 + nla_total_size(1) /* IFLA_BRPORT_VLAN_TUNNEL */ 141 + nla_total_size(1) /* IFLA_BRPORT_NEIGH_SUPPRESS */ 142 + nla_total_size(sizeof(struct ifla_bridge_id)) /* IFLA_BRPORT_ROOT_ID */ 143 + nla_total_size(sizeof(struct ifla_bridge_id)) /* IFLA_BRPORT_BRIDGE_ID */ 144 + nla_total_size(sizeof(u16)) /* IFLA_BRPORT_DESIGNATED_PORT */ 145 + nla_total_size(sizeof(u16)) /* IFLA_BRPORT_DESIGNATED_COST */ 146 + nla_total_size(sizeof(u16)) /* IFLA_BRPORT_ID */ 147 + nla_total_size(sizeof(u16)) /* IFLA_BRPORT_NO */ 148 + nla_total_size(sizeof(u8)) /* IFLA_BRPORT_TOPOLOGY_CHANGE_ACK */ 149 + nla_total_size(sizeof(u8)) /* IFLA_BRPORT_CONFIG_PENDING */ 150 + nla_total_size_64bit(sizeof(u64)) /* IFLA_BRPORT_MESSAGE_AGE_TIMER */ 151 + nla_total_size_64bit(sizeof(u64)) /* IFLA_BRPORT_FORWARD_DELAY_TIMER */ 152 + nla_total_size_64bit(sizeof(u64)) /* IFLA_BRPORT_HOLD_TIMER */ 153 #ifdef CONFIG_BRIDGE_IGMP_SNOOPING 154 + nla_total_size(sizeof(u8)) /* IFLA_BRPORT_MULTICAST_ROUTER */ 155 #endif 156 + nla_total_size(sizeof(u16)) /* IFLA_BRPORT_GROUP_FWD_MASK */ 157 + 0; 158 } 159 160 static inline size_t br_nlmsg_size(struct net_device *dev, u32 filter_mask) 161 { 162 return NLMSG_ALIGN(sizeof(struct ifinfomsg)) 163 + nla_total_size(IFNAMSIZ) /* IFLA_IFNAME */ 164 + nla_total_size(MAX_ADDR_LEN) /* IFLA_ADDRESS */ 165 + nla_total_size(4) /* IFLA_MASTER */ 166 + nla_total_size(4) /* IFLA_MTU */ 167 + nla_total_size(4) /* IFLA_LINK */ 168 + nla_total_size(1) /* IFLA_OPERSTATE */ 169 + nla_total_size(br_port_info_size()) /* IFLA_PROTINFO */ 170 + nla_total_size(br_get_link_af_size_filtered(dev, 171 filter_mask)); /* IFLA_AF_SPEC */ 172 } 173 174 static int br_port_fill_attrs(struct sk_buff *skb, 175 const struct net_bridge_port *p) 176 { 177 u8 mode = !!(p->flags & BR_HAIRPIN_MODE); 178 u64 timerval; 179 180 if (nla_put_u8(skb, IFLA_BRPORT_STATE, p->state) || 181 nla_put_u16(skb, IFLA_BRPORT_PRIORITY, p->priority) || 182 nla_put_u32(skb, IFLA_BRPORT_COST, p->path_cost) || 183 nla_put_u8(skb, IFLA_BRPORT_MODE, mode) || 184 nla_put_u8(skb, IFLA_BRPORT_GUARD, !!(p->flags & BR_BPDU_GUARD)) || 185 nla_put_u8(skb, IFLA_BRPORT_PROTECT, 186 !!(p->flags & BR_ROOT_BLOCK)) || 187 nla_put_u8(skb, IFLA_BRPORT_FAST_LEAVE, 188 !!(p->flags & BR_MULTICAST_FAST_LEAVE)) || 189 nla_put_u8(skb, IFLA_BRPORT_MCAST_TO_UCAST, 190 !!(p->flags & BR_MULTICAST_TO_UNICAST)) || 191 nla_put_u8(skb, IFLA_BRPORT_LEARNING, !!(p->flags & BR_LEARNING)) || 192 nla_put_u8(skb, IFLA_BRPORT_UNICAST_FLOOD, 193 !!(p->flags & BR_FLOOD)) || 194 nla_put_u8(skb, IFLA_BRPORT_MCAST_FLOOD, 195 !!(p->flags & BR_MCAST_FLOOD)) || 196 nla_put_u8(skb, IFLA_BRPORT_BCAST_FLOOD, 197 !!(p->flags & BR_BCAST_FLOOD)) || 198 nla_put_u8(skb, IFLA_BRPORT_PROXYARP, !!(p->flags & BR_PROXYARP)) || 199 nla_put_u8(skb, IFLA_BRPORT_PROXYARP_WIFI, 200 !!(p->flags & BR_PROXYARP_WIFI)) || 201 nla_put(skb, IFLA_BRPORT_ROOT_ID, sizeof(struct ifla_bridge_id), 202 &p->designated_root) || 203 nla_put(skb, IFLA_BRPORT_BRIDGE_ID, sizeof(struct ifla_bridge_id), 204 &p->designated_bridge) || 205 nla_put_u16(skb, IFLA_BRPORT_DESIGNATED_PORT, p->designated_port) || 206 nla_put_u16(skb, IFLA_BRPORT_DESIGNATED_COST, p->designated_cost) || 207 nla_put_u16(skb, IFLA_BRPORT_ID, p->port_id) || 208 nla_put_u16(skb, IFLA_BRPORT_NO, p->port_no) || 209 nla_put_u8(skb, IFLA_BRPORT_TOPOLOGY_CHANGE_ACK, 210 p->topology_change_ack) || 211 nla_put_u8(skb, IFLA_BRPORT_CONFIG_PENDING, p->config_pending) || 212 nla_put_u8(skb, IFLA_BRPORT_VLAN_TUNNEL, !!(p->flags & 213 BR_VLAN_TUNNEL)) || 214 nla_put_u16(skb, IFLA_BRPORT_GROUP_FWD_MASK, p->group_fwd_mask) || 215 nla_put_u8(skb, IFLA_BRPORT_NEIGH_SUPPRESS, 216 !!(p->flags & BR_NEIGH_SUPPRESS))) 217 return -EMSGSIZE; 218 219 timerval = br_timer_value(&p->message_age_timer); 220 if (nla_put_u64_64bit(skb, IFLA_BRPORT_MESSAGE_AGE_TIMER, timerval, 221 IFLA_BRPORT_PAD)) 222 return -EMSGSIZE; 223 timerval = br_timer_value(&p->forward_delay_timer); 224 if (nla_put_u64_64bit(skb, IFLA_BRPORT_FORWARD_DELAY_TIMER, timerval, 225 IFLA_BRPORT_PAD)) 226 return -EMSGSIZE; 227 timerval = br_timer_value(&p->hold_timer); 228 if (nla_put_u64_64bit(skb, IFLA_BRPORT_HOLD_TIMER, timerval, 229 IFLA_BRPORT_PAD)) 230 return -EMSGSIZE; 231 232 #ifdef CONFIG_BRIDGE_IGMP_SNOOPING 233 if (nla_put_u8(skb, IFLA_BRPORT_MULTICAST_ROUTER, 234 p->multicast_router)) 235 return -EMSGSIZE; 236 #endif 237 238 return 0; 239 } 240 241 static int br_fill_ifvlaninfo_range(struct sk_buff *skb, u16 vid_start, 242 u16 vid_end, u16 flags) 243 { 244 struct bridge_vlan_info vinfo; 245 246 if ((vid_end - vid_start) > 0) { 247 /* add range to skb */ 248 vinfo.vid = vid_start; 249 vinfo.flags = flags | BRIDGE_VLAN_INFO_RANGE_BEGIN; 250 if (nla_put(skb, IFLA_BRIDGE_VLAN_INFO, 251 sizeof(vinfo), &vinfo)) 252 goto nla_put_failure; 253 254 vinfo.vid = vid_end; 255 vinfo.flags = flags | BRIDGE_VLAN_INFO_RANGE_END; 256 if (nla_put(skb, IFLA_BRIDGE_VLAN_INFO, 257 sizeof(vinfo), &vinfo)) 258 goto nla_put_failure; 259 } else { 260 vinfo.vid = vid_start; 261 vinfo.flags = flags; 262 if (nla_put(skb, IFLA_BRIDGE_VLAN_INFO, 263 sizeof(vinfo), &vinfo)) 264 goto nla_put_failure; 265 } 266 267 return 0; 268 269 nla_put_failure: 270 return -EMSGSIZE; 271 } 272 273 static int br_fill_ifvlaninfo_compressed(struct sk_buff *skb, 274 struct net_bridge_vlan_group *vg) 275 { 276 struct net_bridge_vlan *v; 277 u16 vid_range_start = 0, vid_range_end = 0, vid_range_flags = 0; 278 u16 flags, pvid; 279 int err = 0; 280 281 /* Pack IFLA_BRIDGE_VLAN_INFO's for every vlan 282 * and mark vlan info with begin and end flags 283 * if vlaninfo represents a range 284 */ 285 pvid = br_get_pvid(vg); 286 list_for_each_entry_rcu(v, &vg->vlan_list, vlist) { 287 flags = 0; 288 if (!br_vlan_should_use(v)) 289 continue; 290 if (v->vid == pvid) 291 flags |= BRIDGE_VLAN_INFO_PVID; 292 293 if (v->flags & BRIDGE_VLAN_INFO_UNTAGGED) 294 flags |= BRIDGE_VLAN_INFO_UNTAGGED; 295 296 if (vid_range_start == 0) { 297 goto initvars; 298 } else if ((v->vid - vid_range_end) == 1 && 299 flags == vid_range_flags) { 300 vid_range_end = v->vid; 301 continue; 302 } else { 303 err = br_fill_ifvlaninfo_range(skb, vid_range_start, 304 vid_range_end, 305 vid_range_flags); 306 if (err) 307 return err; 308 } 309 310 initvars: 311 vid_range_start = v->vid; 312 vid_range_end = v->vid; 313 vid_range_flags = flags; 314 } 315 316 if (vid_range_start != 0) { 317 /* Call it once more to send any left over vlans */ 318 err = br_fill_ifvlaninfo_range(skb, vid_range_start, 319 vid_range_end, 320 vid_range_flags); 321 if (err) 322 return err; 323 } 324 325 return 0; 326 } 327 328 static int br_fill_ifvlaninfo(struct sk_buff *skb, 329 struct net_bridge_vlan_group *vg) 330 { 331 struct bridge_vlan_info vinfo; 332 struct net_bridge_vlan *v; 333 u16 pvid; 334 335 pvid = br_get_pvid(vg); 336 list_for_each_entry_rcu(v, &vg->vlan_list, vlist) { 337 if (!br_vlan_should_use(v)) 338 continue; 339 340 vinfo.vid = v->vid; 341 vinfo.flags = 0; 342 if (v->vid == pvid) 343 vinfo.flags |= BRIDGE_VLAN_INFO_PVID; 344 345 if (v->flags & BRIDGE_VLAN_INFO_UNTAGGED) 346 vinfo.flags |= BRIDGE_VLAN_INFO_UNTAGGED; 347 348 if (nla_put(skb, IFLA_BRIDGE_VLAN_INFO, 349 sizeof(vinfo), &vinfo)) 350 goto nla_put_failure; 351 } 352 353 return 0; 354 355 nla_put_failure: 356 return -EMSGSIZE; 357 } 358 359 /* 360 * Create one netlink message for one interface 361 * Contains port and master info as well as carrier and bridge state. 362 */ 363 static int br_fill_ifinfo(struct sk_buff *skb, 364 struct net_bridge_port *port, 365 u32 pid, u32 seq, int event, unsigned int flags, 366 u32 filter_mask, const struct net_device *dev) 367 { 368 struct net_bridge *br; 369 struct ifinfomsg *hdr; 370 struct nlmsghdr *nlh; 371 u8 operstate = netif_running(dev) ? dev->operstate : IF_OPER_DOWN; 372 373 if (port) 374 br = port->br; 375 else 376 br = netdev_priv(dev); 377 378 br_debug(br, "br_fill_info event %d port %s master %s\n", 379 event, dev->name, br->dev->name); 380 381 nlh = nlmsg_put(skb, pid, seq, event, sizeof(*hdr), flags); 382 if (nlh == NULL) 383 return -EMSGSIZE; 384 385 hdr = nlmsg_data(nlh); 386 hdr->ifi_family = AF_BRIDGE; 387 hdr->__ifi_pad = 0; 388 hdr->ifi_type = dev->type; 389 hdr->ifi_index = dev->ifindex; 390 hdr->ifi_flags = dev_get_flags(dev); 391 hdr->ifi_change = 0; 392 393 if (nla_put_string(skb, IFLA_IFNAME, dev->name) || 394 nla_put_u32(skb, IFLA_MASTER, br->dev->ifindex) || 395 nla_put_u32(skb, IFLA_MTU, dev->mtu) || 396 nla_put_u8(skb, IFLA_OPERSTATE, operstate) || 397 (dev->addr_len && 398 nla_put(skb, IFLA_ADDRESS, dev->addr_len, dev->dev_addr)) || 399 (dev->ifindex != dev_get_iflink(dev) && 400 nla_put_u32(skb, IFLA_LINK, dev_get_iflink(dev)))) 401 goto nla_put_failure; 402 403 if (event == RTM_NEWLINK && port) { 404 struct nlattr *nest 405 = nla_nest_start(skb, IFLA_PROTINFO | NLA_F_NESTED); 406 407 if (nest == NULL || br_port_fill_attrs(skb, port) < 0) 408 goto nla_put_failure; 409 nla_nest_end(skb, nest); 410 } 411 412 /* Check if the VID information is requested */ 413 if ((filter_mask & RTEXT_FILTER_BRVLAN) || 414 (filter_mask & RTEXT_FILTER_BRVLAN_COMPRESSED)) { 415 struct net_bridge_vlan_group *vg; 416 struct nlattr *af; 417 int err; 418 419 /* RCU needed because of the VLAN locking rules (rcu || rtnl) */ 420 rcu_read_lock(); 421 if (port) 422 vg = nbp_vlan_group_rcu(port); 423 else 424 vg = br_vlan_group_rcu(br); 425 426 if (!vg || !vg->num_vlans) { 427 rcu_read_unlock(); 428 goto done; 429 } 430 af = nla_nest_start(skb, IFLA_AF_SPEC); 431 if (!af) { 432 rcu_read_unlock(); 433 goto nla_put_failure; 434 } 435 if (filter_mask & RTEXT_FILTER_BRVLAN_COMPRESSED) 436 err = br_fill_ifvlaninfo_compressed(skb, vg); 437 else 438 err = br_fill_ifvlaninfo(skb, vg); 439 440 if (port && (port->flags & BR_VLAN_TUNNEL)) 441 err = br_fill_vlan_tunnel_info(skb, vg); 442 rcu_read_unlock(); 443 if (err) 444 goto nla_put_failure; 445 nla_nest_end(skb, af); 446 } 447 448 done: 449 nlmsg_end(skb, nlh); 450 return 0; 451 452 nla_put_failure: 453 nlmsg_cancel(skb, nlh); 454 return -EMSGSIZE; 455 } 456 457 /* 458 * Notify listeners of a change in port information 459 */ 460 void br_ifinfo_notify(int event, struct net_bridge_port *port) 461 { 462 struct net *net; 463 struct sk_buff *skb; 464 int err = -ENOBUFS; 465 u32 filter = RTEXT_FILTER_BRVLAN_COMPRESSED; 466 467 if (!port) 468 return; 469 470 net = dev_net(port->dev); 471 br_debug(port->br, "port %u(%s) event %d\n", 472 (unsigned int)port->port_no, port->dev->name, event); 473 474 skb = nlmsg_new(br_nlmsg_size(port->dev, filter), GFP_ATOMIC); 475 if (skb == NULL) 476 goto errout; 477 478 err = br_fill_ifinfo(skb, port, 0, 0, event, 0, filter, port->dev); 479 if (err < 0) { 480 /* -EMSGSIZE implies BUG in br_nlmsg_size() */ 481 WARN_ON(err == -EMSGSIZE); 482 kfree_skb(skb); 483 goto errout; 484 } 485 rtnl_notify(skb, net, 0, RTNLGRP_LINK, NULL, GFP_ATOMIC); 486 return; 487 errout: 488 rtnl_set_sk_err(net, RTNLGRP_LINK, err); 489 } 490 491 492 /* 493 * Dump information about all ports, in response to GETLINK 494 */ 495 int br_getlink(struct sk_buff *skb, u32 pid, u32 seq, 496 struct net_device *dev, u32 filter_mask, int nlflags) 497 { 498 struct net_bridge_port *port = br_port_get_rtnl(dev); 499 500 if (!port && !(filter_mask & RTEXT_FILTER_BRVLAN) && 501 !(filter_mask & RTEXT_FILTER_BRVLAN_COMPRESSED)) 502 return 0; 503 504 return br_fill_ifinfo(skb, port, pid, seq, RTM_NEWLINK, nlflags, 505 filter_mask, dev); 506 } 507 508 static int br_vlan_info(struct net_bridge *br, struct net_bridge_port *p, 509 int cmd, struct bridge_vlan_info *vinfo, bool *changed) 510 { 511 bool curr_change; 512 int err = 0; 513 514 switch (cmd) { 515 case RTM_SETLINK: 516 if (p) { 517 /* if the MASTER flag is set this will act on the global 518 * per-VLAN entry as well 519 */ 520 err = nbp_vlan_add(p, vinfo->vid, vinfo->flags, 521 &curr_change); 522 } else { 523 vinfo->flags |= BRIDGE_VLAN_INFO_BRENTRY; 524 err = br_vlan_add(br, vinfo->vid, vinfo->flags, 525 &curr_change); 526 } 527 if (curr_change) 528 *changed = true; 529 break; 530 531 case RTM_DELLINK: 532 if (p) { 533 if (!nbp_vlan_delete(p, vinfo->vid)) 534 *changed = true; 535 536 if ((vinfo->flags & BRIDGE_VLAN_INFO_MASTER) && 537 !br_vlan_delete(p->br, vinfo->vid)) 538 *changed = true; 539 } else if (!br_vlan_delete(br, vinfo->vid)) { 540 *changed = true; 541 } 542 break; 543 } 544 545 return err; 546 } 547 548 static int br_process_vlan_info(struct net_bridge *br, 549 struct net_bridge_port *p, int cmd, 550 struct bridge_vlan_info *vinfo_curr, 551 struct bridge_vlan_info **vinfo_last, 552 bool *changed) 553 { 554 if (!vinfo_curr->vid || vinfo_curr->vid >= VLAN_VID_MASK) 555 return -EINVAL; 556 557 if (vinfo_curr->flags & BRIDGE_VLAN_INFO_RANGE_BEGIN) { 558 /* check if we are already processing a range */ 559 if (*vinfo_last) 560 return -EINVAL; 561 *vinfo_last = vinfo_curr; 562 /* don't allow range of pvids */ 563 if ((*vinfo_last)->flags & BRIDGE_VLAN_INFO_PVID) 564 return -EINVAL; 565 return 0; 566 } 567 568 if (*vinfo_last) { 569 struct bridge_vlan_info tmp_vinfo; 570 int v, err; 571 572 if (!(vinfo_curr->flags & BRIDGE_VLAN_INFO_RANGE_END)) 573 return -EINVAL; 574 575 if (vinfo_curr->vid <= (*vinfo_last)->vid) 576 return -EINVAL; 577 578 memcpy(&tmp_vinfo, *vinfo_last, 579 sizeof(struct bridge_vlan_info)); 580 for (v = (*vinfo_last)->vid; v <= vinfo_curr->vid; v++) { 581 tmp_vinfo.vid = v; 582 err = br_vlan_info(br, p, cmd, &tmp_vinfo, changed); 583 if (err) 584 break; 585 } 586 *vinfo_last = NULL; 587 588 return err; 589 } 590 591 return br_vlan_info(br, p, cmd, vinfo_curr, changed); 592 } 593 594 static int br_afspec(struct net_bridge *br, 595 struct net_bridge_port *p, 596 struct nlattr *af_spec, 597 int cmd, bool *changed) 598 { 599 struct bridge_vlan_info *vinfo_curr = NULL; 600 struct bridge_vlan_info *vinfo_last = NULL; 601 struct nlattr *attr; 602 struct vtunnel_info tinfo_last = {}; 603 struct vtunnel_info tinfo_curr = {}; 604 int err = 0, rem; 605 606 nla_for_each_nested(attr, af_spec, rem) { 607 err = 0; 608 switch (nla_type(attr)) { 609 case IFLA_BRIDGE_VLAN_TUNNEL_INFO: 610 if (!p || !(p->flags & BR_VLAN_TUNNEL)) 611 return -EINVAL; 612 err = br_parse_vlan_tunnel_info(attr, &tinfo_curr); 613 if (err) 614 return err; 615 err = br_process_vlan_tunnel_info(br, p, cmd, 616 &tinfo_curr, 617 &tinfo_last, 618 changed); 619 if (err) 620 return err; 621 break; 622 case IFLA_BRIDGE_VLAN_INFO: 623 if (nla_len(attr) != sizeof(struct bridge_vlan_info)) 624 return -EINVAL; 625 vinfo_curr = nla_data(attr); 626 err = br_process_vlan_info(br, p, cmd, vinfo_curr, 627 &vinfo_last, changed); 628 if (err) 629 return err; 630 break; 631 } 632 } 633 634 return err; 635 } 636 637 static const struct nla_policy br_port_policy[IFLA_BRPORT_MAX + 1] = { 638 [IFLA_BRPORT_STATE] = { .type = NLA_U8 }, 639 [IFLA_BRPORT_COST] = { .type = NLA_U32 }, 640 [IFLA_BRPORT_PRIORITY] = { .type = NLA_U16 }, 641 [IFLA_BRPORT_MODE] = { .type = NLA_U8 }, 642 [IFLA_BRPORT_GUARD] = { .type = NLA_U8 }, 643 [IFLA_BRPORT_PROTECT] = { .type = NLA_U8 }, 644 [IFLA_BRPORT_FAST_LEAVE]= { .type = NLA_U8 }, 645 [IFLA_BRPORT_LEARNING] = { .type = NLA_U8 }, 646 [IFLA_BRPORT_UNICAST_FLOOD] = { .type = NLA_U8 }, 647 [IFLA_BRPORT_PROXYARP] = { .type = NLA_U8 }, 648 [IFLA_BRPORT_PROXYARP_WIFI] = { .type = NLA_U8 }, 649 [IFLA_BRPORT_MULTICAST_ROUTER] = { .type = NLA_U8 }, 650 [IFLA_BRPORT_MCAST_TO_UCAST] = { .type = NLA_U8 }, 651 [IFLA_BRPORT_MCAST_FLOOD] = { .type = NLA_U8 }, 652 [IFLA_BRPORT_BCAST_FLOOD] = { .type = NLA_U8 }, 653 [IFLA_BRPORT_GROUP_FWD_MASK] = { .type = NLA_U16 }, 654 [IFLA_BRPORT_NEIGH_SUPPRESS] = { .type = NLA_U8 }, 655 }; 656 657 /* Change the state of the port and notify spanning tree */ 658 static int br_set_port_state(struct net_bridge_port *p, u8 state) 659 { 660 if (state > BR_STATE_BLOCKING) 661 return -EINVAL; 662 663 /* if kernel STP is running, don't allow changes */ 664 if (p->br->stp_enabled == BR_KERNEL_STP) 665 return -EBUSY; 666 667 /* if device is not up, change is not allowed 668 * if link is not present, only allowable state is disabled 669 */ 670 if (!netif_running(p->dev) || 671 (!netif_oper_up(p->dev) && state != BR_STATE_DISABLED)) 672 return -ENETDOWN; 673 674 br_set_state(p, state); 675 br_port_state_selection(p->br); 676 return 0; 677 } 678 679 /* Set/clear or port flags based on attribute */ 680 static int br_set_port_flag(struct net_bridge_port *p, struct nlattr *tb[], 681 int attrtype, unsigned long mask) 682 { 683 unsigned long flags; 684 int err; 685 686 if (!tb[attrtype]) 687 return 0; 688 689 if (nla_get_u8(tb[attrtype])) 690 flags = p->flags | mask; 691 else 692 flags = p->flags & ~mask; 693 694 err = br_switchdev_set_port_flag(p, flags, mask); 695 if (err) 696 return err; 697 698 p->flags = flags; 699 return 0; 700 } 701 702 /* Process bridge protocol info on port */ 703 static int br_setport(struct net_bridge_port *p, struct nlattr *tb[]) 704 { 705 unsigned long old_flags = p->flags; 706 bool br_vlan_tunnel_old = false; 707 int err; 708 709 err = br_set_port_flag(p, tb, IFLA_BRPORT_MODE, BR_HAIRPIN_MODE); 710 if (err) 711 return err; 712 713 err = br_set_port_flag(p, tb, IFLA_BRPORT_GUARD, BR_BPDU_GUARD); 714 if (err) 715 return err; 716 717 err = br_set_port_flag(p, tb, IFLA_BRPORT_FAST_LEAVE, BR_MULTICAST_FAST_LEAVE); 718 if (err) 719 return err; 720 721 err = br_set_port_flag(p, tb, IFLA_BRPORT_PROTECT, BR_ROOT_BLOCK); 722 if (err) 723 return err; 724 725 err = br_set_port_flag(p, tb, IFLA_BRPORT_LEARNING, BR_LEARNING); 726 if (err) 727 return err; 728 729 err = br_set_port_flag(p, tb, IFLA_BRPORT_UNICAST_FLOOD, BR_FLOOD); 730 if (err) 731 return err; 732 733 err = br_set_port_flag(p, tb, IFLA_BRPORT_MCAST_FLOOD, BR_MCAST_FLOOD); 734 if (err) 735 return err; 736 737 err = br_set_port_flag(p, tb, IFLA_BRPORT_MCAST_TO_UCAST, BR_MULTICAST_TO_UNICAST); 738 if (err) 739 return err; 740 741 err = br_set_port_flag(p, tb, IFLA_BRPORT_BCAST_FLOOD, BR_BCAST_FLOOD); 742 if (err) 743 return err; 744 745 err = br_set_port_flag(p, tb, IFLA_BRPORT_PROXYARP, BR_PROXYARP); 746 if (err) 747 return err; 748 749 err = br_set_port_flag(p, tb, IFLA_BRPORT_PROXYARP_WIFI, BR_PROXYARP_WIFI); 750 if (err) 751 return err; 752 753 br_vlan_tunnel_old = (p->flags & BR_VLAN_TUNNEL) ? true : false; 754 err = br_set_port_flag(p, tb, IFLA_BRPORT_VLAN_TUNNEL, BR_VLAN_TUNNEL); 755 if (err) 756 return err; 757 758 if (br_vlan_tunnel_old && !(p->flags & BR_VLAN_TUNNEL)) 759 nbp_vlan_tunnel_info_flush(p); 760 761 if (tb[IFLA_BRPORT_COST]) { 762 err = br_stp_set_path_cost(p, nla_get_u32(tb[IFLA_BRPORT_COST])); 763 if (err) 764 return err; 765 } 766 767 if (tb[IFLA_BRPORT_PRIORITY]) { 768 err = br_stp_set_port_priority(p, nla_get_u16(tb[IFLA_BRPORT_PRIORITY])); 769 if (err) 770 return err; 771 } 772 773 if (tb[IFLA_BRPORT_STATE]) { 774 err = br_set_port_state(p, nla_get_u8(tb[IFLA_BRPORT_STATE])); 775 if (err) 776 return err; 777 } 778 779 if (tb[IFLA_BRPORT_FLUSH]) 780 br_fdb_delete_by_port(p->br, p, 0, 0); 781 782 #ifdef CONFIG_BRIDGE_IGMP_SNOOPING 783 if (tb[IFLA_BRPORT_MULTICAST_ROUTER]) { 784 u8 mcast_router = nla_get_u8(tb[IFLA_BRPORT_MULTICAST_ROUTER]); 785 786 err = br_multicast_set_port_router(p, mcast_router); 787 if (err) 788 return err; 789 } 790 #endif 791 792 if (tb[IFLA_BRPORT_GROUP_FWD_MASK]) { 793 u16 fwd_mask = nla_get_u16(tb[IFLA_BRPORT_GROUP_FWD_MASK]); 794 795 if (fwd_mask & BR_GROUPFWD_MACPAUSE) 796 return -EINVAL; 797 p->group_fwd_mask = fwd_mask; 798 } 799 800 err = br_set_port_flag(p, tb, IFLA_BRPORT_NEIGH_SUPPRESS, 801 BR_NEIGH_SUPPRESS); 802 if (err) 803 return err; 804 805 br_port_flags_change(p, old_flags ^ p->flags); 806 return 0; 807 } 808 809 /* Change state and parameters on port. */ 810 int br_setlink(struct net_device *dev, struct nlmsghdr *nlh, u16 flags) 811 { 812 struct nlattr *protinfo; 813 struct nlattr *afspec; 814 struct net_bridge_port *p; 815 struct nlattr *tb[IFLA_BRPORT_MAX + 1]; 816 bool changed = false; 817 int err = 0; 818 819 protinfo = nlmsg_find_attr(nlh, sizeof(struct ifinfomsg), IFLA_PROTINFO); 820 afspec = nlmsg_find_attr(nlh, sizeof(struct ifinfomsg), IFLA_AF_SPEC); 821 if (!protinfo && !afspec) 822 return 0; 823 824 p = br_port_get_rtnl(dev); 825 /* We want to accept dev as bridge itself if the AF_SPEC 826 * is set to see if someone is setting vlan info on the bridge 827 */ 828 if (!p && !afspec) 829 return -EINVAL; 830 831 if (p && protinfo) { 832 if (protinfo->nla_type & NLA_F_NESTED) { 833 err = nla_parse_nested(tb, IFLA_BRPORT_MAX, protinfo, 834 br_port_policy, NULL); 835 if (err) 836 return err; 837 838 spin_lock_bh(&p->br->lock); 839 err = br_setport(p, tb); 840 spin_unlock_bh(&p->br->lock); 841 } else { 842 /* Binary compatibility with old RSTP */ 843 if (nla_len(protinfo) < sizeof(u8)) 844 return -EINVAL; 845 846 spin_lock_bh(&p->br->lock); 847 err = br_set_port_state(p, nla_get_u8(protinfo)); 848 spin_unlock_bh(&p->br->lock); 849 } 850 if (err) 851 goto out; 852 changed = true; 853 } 854 855 if (afspec) { 856 err = br_afspec((struct net_bridge *)netdev_priv(dev), p, 857 afspec, RTM_SETLINK, &changed); 858 } 859 860 if (changed) 861 br_ifinfo_notify(RTM_NEWLINK, p); 862 out: 863 return err; 864 } 865 866 /* Delete port information */ 867 int br_dellink(struct net_device *dev, struct nlmsghdr *nlh, u16 flags) 868 { 869 struct nlattr *afspec; 870 struct net_bridge_port *p; 871 bool changed = false; 872 int err = 0; 873 874 afspec = nlmsg_find_attr(nlh, sizeof(struct ifinfomsg), IFLA_AF_SPEC); 875 if (!afspec) 876 return 0; 877 878 p = br_port_get_rtnl(dev); 879 /* We want to accept dev as bridge itself as well */ 880 if (!p && !(dev->priv_flags & IFF_EBRIDGE)) 881 return -EINVAL; 882 883 err = br_afspec((struct net_bridge *)netdev_priv(dev), p, 884 afspec, RTM_DELLINK, &changed); 885 if (changed) 886 /* Send RTM_NEWLINK because userspace 887 * expects RTM_NEWLINK for vlan dels 888 */ 889 br_ifinfo_notify(RTM_NEWLINK, p); 890 891 return err; 892 } 893 894 static int br_validate(struct nlattr *tb[], struct nlattr *data[], 895 struct netlink_ext_ack *extack) 896 { 897 if (tb[IFLA_ADDRESS]) { 898 if (nla_len(tb[IFLA_ADDRESS]) != ETH_ALEN) 899 return -EINVAL; 900 if (!is_valid_ether_addr(nla_data(tb[IFLA_ADDRESS]))) 901 return -EADDRNOTAVAIL; 902 } 903 904 if (!data) 905 return 0; 906 907 #ifdef CONFIG_BRIDGE_VLAN_FILTERING 908 if (data[IFLA_BR_VLAN_PROTOCOL]) { 909 switch (nla_get_be16(data[IFLA_BR_VLAN_PROTOCOL])) { 910 case htons(ETH_P_8021Q): 911 case htons(ETH_P_8021AD): 912 break; 913 default: 914 return -EPROTONOSUPPORT; 915 } 916 } 917 918 if (data[IFLA_BR_VLAN_DEFAULT_PVID]) { 919 __u16 defpvid = nla_get_u16(data[IFLA_BR_VLAN_DEFAULT_PVID]); 920 921 if (defpvid >= VLAN_VID_MASK) 922 return -EINVAL; 923 } 924 #endif 925 926 return 0; 927 } 928 929 static int br_port_slave_changelink(struct net_device *brdev, 930 struct net_device *dev, 931 struct nlattr *tb[], 932 struct nlattr *data[], 933 struct netlink_ext_ack *extack) 934 { 935 struct net_bridge *br = netdev_priv(brdev); 936 int ret; 937 938 if (!data) 939 return 0; 940 941 spin_lock_bh(&br->lock); 942 ret = br_setport(br_port_get_rtnl(dev), data); 943 spin_unlock_bh(&br->lock); 944 945 return ret; 946 } 947 948 static int br_port_fill_slave_info(struct sk_buff *skb, 949 const struct net_device *brdev, 950 const struct net_device *dev) 951 { 952 return br_port_fill_attrs(skb, br_port_get_rtnl(dev)); 953 } 954 955 static size_t br_port_get_slave_size(const struct net_device *brdev, 956 const struct net_device *dev) 957 { 958 return br_port_info_size(); 959 } 960 961 static const struct nla_policy br_policy[IFLA_BR_MAX + 1] = { 962 [IFLA_BR_FORWARD_DELAY] = { .type = NLA_U32 }, 963 [IFLA_BR_HELLO_TIME] = { .type = NLA_U32 }, 964 [IFLA_BR_MAX_AGE] = { .type = NLA_U32 }, 965 [IFLA_BR_AGEING_TIME] = { .type = NLA_U32 }, 966 [IFLA_BR_STP_STATE] = { .type = NLA_U32 }, 967 [IFLA_BR_PRIORITY] = { .type = NLA_U16 }, 968 [IFLA_BR_VLAN_FILTERING] = { .type = NLA_U8 }, 969 [IFLA_BR_VLAN_PROTOCOL] = { .type = NLA_U16 }, 970 [IFLA_BR_GROUP_FWD_MASK] = { .type = NLA_U16 }, 971 [IFLA_BR_GROUP_ADDR] = { .type = NLA_BINARY, 972 .len = ETH_ALEN }, 973 [IFLA_BR_MCAST_ROUTER] = { .type = NLA_U8 }, 974 [IFLA_BR_MCAST_SNOOPING] = { .type = NLA_U8 }, 975 [IFLA_BR_MCAST_QUERY_USE_IFADDR] = { .type = NLA_U8 }, 976 [IFLA_BR_MCAST_QUERIER] = { .type = NLA_U8 }, 977 [IFLA_BR_MCAST_HASH_ELASTICITY] = { .type = NLA_U32 }, 978 [IFLA_BR_MCAST_HASH_MAX] = { .type = NLA_U32 }, 979 [IFLA_BR_MCAST_LAST_MEMBER_CNT] = { .type = NLA_U32 }, 980 [IFLA_BR_MCAST_STARTUP_QUERY_CNT] = { .type = NLA_U32 }, 981 [IFLA_BR_MCAST_LAST_MEMBER_INTVL] = { .type = NLA_U64 }, 982 [IFLA_BR_MCAST_MEMBERSHIP_INTVL] = { .type = NLA_U64 }, 983 [IFLA_BR_MCAST_QUERIER_INTVL] = { .type = NLA_U64 }, 984 [IFLA_BR_MCAST_QUERY_INTVL] = { .type = NLA_U64 }, 985 [IFLA_BR_MCAST_QUERY_RESPONSE_INTVL] = { .type = NLA_U64 }, 986 [IFLA_BR_MCAST_STARTUP_QUERY_INTVL] = { .type = NLA_U64 }, 987 [IFLA_BR_NF_CALL_IPTABLES] = { .type = NLA_U8 }, 988 [IFLA_BR_NF_CALL_IP6TABLES] = { .type = NLA_U8 }, 989 [IFLA_BR_NF_CALL_ARPTABLES] = { .type = NLA_U8 }, 990 [IFLA_BR_VLAN_DEFAULT_PVID] = { .type = NLA_U16 }, 991 [IFLA_BR_VLAN_STATS_ENABLED] = { .type = NLA_U8 }, 992 [IFLA_BR_MCAST_STATS_ENABLED] = { .type = NLA_U8 }, 993 [IFLA_BR_MCAST_IGMP_VERSION] = { .type = NLA_U8 }, 994 [IFLA_BR_MCAST_MLD_VERSION] = { .type = NLA_U8 }, 995 }; 996 997 static int br_changelink(struct net_device *brdev, struct nlattr *tb[], 998 struct nlattr *data[], 999 struct netlink_ext_ack *extack) 1000 { 1001 struct net_bridge *br = netdev_priv(brdev); 1002 int err; 1003 1004 if (!data) 1005 return 0; 1006 1007 if (data[IFLA_BR_FORWARD_DELAY]) { 1008 err = br_set_forward_delay(br, nla_get_u32(data[IFLA_BR_FORWARD_DELAY])); 1009 if (err) 1010 return err; 1011 } 1012 1013 if (data[IFLA_BR_HELLO_TIME]) { 1014 err = br_set_hello_time(br, nla_get_u32(data[IFLA_BR_HELLO_TIME])); 1015 if (err) 1016 return err; 1017 } 1018 1019 if (data[IFLA_BR_MAX_AGE]) { 1020 err = br_set_max_age(br, nla_get_u32(data[IFLA_BR_MAX_AGE])); 1021 if (err) 1022 return err; 1023 } 1024 1025 if (data[IFLA_BR_AGEING_TIME]) { 1026 err = br_set_ageing_time(br, nla_get_u32(data[IFLA_BR_AGEING_TIME])); 1027 if (err) 1028 return err; 1029 } 1030 1031 if (data[IFLA_BR_STP_STATE]) { 1032 u32 stp_enabled = nla_get_u32(data[IFLA_BR_STP_STATE]); 1033 1034 br_stp_set_enabled(br, stp_enabled); 1035 } 1036 1037 if (data[IFLA_BR_PRIORITY]) { 1038 u32 priority = nla_get_u16(data[IFLA_BR_PRIORITY]); 1039 1040 br_stp_set_bridge_priority(br, priority); 1041 } 1042 1043 if (data[IFLA_BR_VLAN_FILTERING]) { 1044 u8 vlan_filter = nla_get_u8(data[IFLA_BR_VLAN_FILTERING]); 1045 1046 err = __br_vlan_filter_toggle(br, vlan_filter); 1047 if (err) 1048 return err; 1049 } 1050 1051 #ifdef CONFIG_BRIDGE_VLAN_FILTERING 1052 if (data[IFLA_BR_VLAN_PROTOCOL]) { 1053 __be16 vlan_proto = nla_get_be16(data[IFLA_BR_VLAN_PROTOCOL]); 1054 1055 err = __br_vlan_set_proto(br, vlan_proto); 1056 if (err) 1057 return err; 1058 } 1059 1060 if (data[IFLA_BR_VLAN_DEFAULT_PVID]) { 1061 __u16 defpvid = nla_get_u16(data[IFLA_BR_VLAN_DEFAULT_PVID]); 1062 1063 err = __br_vlan_set_default_pvid(br, defpvid); 1064 if (err) 1065 return err; 1066 } 1067 1068 if (data[IFLA_BR_VLAN_STATS_ENABLED]) { 1069 __u8 vlan_stats = nla_get_u8(data[IFLA_BR_VLAN_STATS_ENABLED]); 1070 1071 err = br_vlan_set_stats(br, vlan_stats); 1072 if (err) 1073 return err; 1074 } 1075 #endif 1076 1077 if (data[IFLA_BR_GROUP_FWD_MASK]) { 1078 u16 fwd_mask = nla_get_u16(data[IFLA_BR_GROUP_FWD_MASK]); 1079 1080 if (fwd_mask & BR_GROUPFWD_RESTRICTED) 1081 return -EINVAL; 1082 br->group_fwd_mask = fwd_mask; 1083 } 1084 1085 if (data[IFLA_BR_GROUP_ADDR]) { 1086 u8 new_addr[ETH_ALEN]; 1087 1088 if (nla_len(data[IFLA_BR_GROUP_ADDR]) != ETH_ALEN) 1089 return -EINVAL; 1090 memcpy(new_addr, nla_data(data[IFLA_BR_GROUP_ADDR]), ETH_ALEN); 1091 if (!is_link_local_ether_addr(new_addr)) 1092 return -EINVAL; 1093 if (new_addr[5] == 1 || /* 802.3x Pause address */ 1094 new_addr[5] == 2 || /* 802.3ad Slow protocols */ 1095 new_addr[5] == 3) /* 802.1X PAE address */ 1096 return -EINVAL; 1097 spin_lock_bh(&br->lock); 1098 memcpy(br->group_addr, new_addr, sizeof(br->group_addr)); 1099 spin_unlock_bh(&br->lock); 1100 br->group_addr_set = true; 1101 br_recalculate_fwd_mask(br); 1102 } 1103 1104 if (data[IFLA_BR_FDB_FLUSH]) 1105 br_fdb_flush(br); 1106 1107 #ifdef CONFIG_BRIDGE_IGMP_SNOOPING 1108 if (data[IFLA_BR_MCAST_ROUTER]) { 1109 u8 multicast_router = nla_get_u8(data[IFLA_BR_MCAST_ROUTER]); 1110 1111 err = br_multicast_set_router(br, multicast_router); 1112 if (err) 1113 return err; 1114 } 1115 1116 if (data[IFLA_BR_MCAST_SNOOPING]) { 1117 u8 mcast_snooping = nla_get_u8(data[IFLA_BR_MCAST_SNOOPING]); 1118 1119 err = br_multicast_toggle(br, mcast_snooping); 1120 if (err) 1121 return err; 1122 } 1123 1124 if (data[IFLA_BR_MCAST_QUERY_USE_IFADDR]) { 1125 u8 val; 1126 1127 val = nla_get_u8(data[IFLA_BR_MCAST_QUERY_USE_IFADDR]); 1128 br->multicast_query_use_ifaddr = !!val; 1129 } 1130 1131 if (data[IFLA_BR_MCAST_QUERIER]) { 1132 u8 mcast_querier = nla_get_u8(data[IFLA_BR_MCAST_QUERIER]); 1133 1134 err = br_multicast_set_querier(br, mcast_querier); 1135 if (err) 1136 return err; 1137 } 1138 1139 if (data[IFLA_BR_MCAST_HASH_ELASTICITY]) { 1140 u32 val = nla_get_u32(data[IFLA_BR_MCAST_HASH_ELASTICITY]); 1141 1142 br->hash_elasticity = val; 1143 } 1144 1145 if (data[IFLA_BR_MCAST_HASH_MAX]) { 1146 u32 hash_max = nla_get_u32(data[IFLA_BR_MCAST_HASH_MAX]); 1147 1148 err = br_multicast_set_hash_max(br, hash_max); 1149 if (err) 1150 return err; 1151 } 1152 1153 if (data[IFLA_BR_MCAST_LAST_MEMBER_CNT]) { 1154 u32 val = nla_get_u32(data[IFLA_BR_MCAST_LAST_MEMBER_CNT]); 1155 1156 br->multicast_last_member_count = val; 1157 } 1158 1159 if (data[IFLA_BR_MCAST_STARTUP_QUERY_CNT]) { 1160 u32 val = nla_get_u32(data[IFLA_BR_MCAST_STARTUP_QUERY_CNT]); 1161 1162 br->multicast_startup_query_count = val; 1163 } 1164 1165 if (data[IFLA_BR_MCAST_LAST_MEMBER_INTVL]) { 1166 u64 val = nla_get_u64(data[IFLA_BR_MCAST_LAST_MEMBER_INTVL]); 1167 1168 br->multicast_last_member_interval = clock_t_to_jiffies(val); 1169 } 1170 1171 if (data[IFLA_BR_MCAST_MEMBERSHIP_INTVL]) { 1172 u64 val = nla_get_u64(data[IFLA_BR_MCAST_MEMBERSHIP_INTVL]); 1173 1174 br->multicast_membership_interval = clock_t_to_jiffies(val); 1175 } 1176 1177 if (data[IFLA_BR_MCAST_QUERIER_INTVL]) { 1178 u64 val = nla_get_u64(data[IFLA_BR_MCAST_QUERIER_INTVL]); 1179 1180 br->multicast_querier_interval = clock_t_to_jiffies(val); 1181 } 1182 1183 if (data[IFLA_BR_MCAST_QUERY_INTVL]) { 1184 u64 val = nla_get_u64(data[IFLA_BR_MCAST_QUERY_INTVL]); 1185 1186 br->multicast_query_interval = clock_t_to_jiffies(val); 1187 } 1188 1189 if (data[IFLA_BR_MCAST_QUERY_RESPONSE_INTVL]) { 1190 u64 val = nla_get_u64(data[IFLA_BR_MCAST_QUERY_RESPONSE_INTVL]); 1191 1192 br->multicast_query_response_interval = clock_t_to_jiffies(val); 1193 } 1194 1195 if (data[IFLA_BR_MCAST_STARTUP_QUERY_INTVL]) { 1196 u64 val = nla_get_u64(data[IFLA_BR_MCAST_STARTUP_QUERY_INTVL]); 1197 1198 br->multicast_startup_query_interval = clock_t_to_jiffies(val); 1199 } 1200 1201 if (data[IFLA_BR_MCAST_STATS_ENABLED]) { 1202 __u8 mcast_stats; 1203 1204 mcast_stats = nla_get_u8(data[IFLA_BR_MCAST_STATS_ENABLED]); 1205 br->multicast_stats_enabled = !!mcast_stats; 1206 } 1207 1208 if (data[IFLA_BR_MCAST_IGMP_VERSION]) { 1209 __u8 igmp_version; 1210 1211 igmp_version = nla_get_u8(data[IFLA_BR_MCAST_IGMP_VERSION]); 1212 err = br_multicast_set_igmp_version(br, igmp_version); 1213 if (err) 1214 return err; 1215 } 1216 1217 #if IS_ENABLED(CONFIG_IPV6) 1218 if (data[IFLA_BR_MCAST_MLD_VERSION]) { 1219 __u8 mld_version; 1220 1221 mld_version = nla_get_u8(data[IFLA_BR_MCAST_MLD_VERSION]); 1222 err = br_multicast_set_mld_version(br, mld_version); 1223 if (err) 1224 return err; 1225 } 1226 #endif 1227 #endif 1228 #if IS_ENABLED(CONFIG_BRIDGE_NETFILTER) 1229 if (data[IFLA_BR_NF_CALL_IPTABLES]) { 1230 u8 val = nla_get_u8(data[IFLA_BR_NF_CALL_IPTABLES]); 1231 1232 br->nf_call_iptables = val ? true : false; 1233 } 1234 1235 if (data[IFLA_BR_NF_CALL_IP6TABLES]) { 1236 u8 val = nla_get_u8(data[IFLA_BR_NF_CALL_IP6TABLES]); 1237 1238 br->nf_call_ip6tables = val ? true : false; 1239 } 1240 1241 if (data[IFLA_BR_NF_CALL_ARPTABLES]) { 1242 u8 val = nla_get_u8(data[IFLA_BR_NF_CALL_ARPTABLES]); 1243 1244 br->nf_call_arptables = val ? true : false; 1245 } 1246 #endif 1247 1248 return 0; 1249 } 1250 1251 static int br_dev_newlink(struct net *src_net, struct net_device *dev, 1252 struct nlattr *tb[], struct nlattr *data[], 1253 struct netlink_ext_ack *extack) 1254 { 1255 struct net_bridge *br = netdev_priv(dev); 1256 int err; 1257 1258 if (tb[IFLA_ADDRESS]) { 1259 spin_lock_bh(&br->lock); 1260 br_stp_change_bridge_id(br, nla_data(tb[IFLA_ADDRESS])); 1261 spin_unlock_bh(&br->lock); 1262 } 1263 1264 err = register_netdevice(dev); 1265 if (err) 1266 return err; 1267 1268 err = br_changelink(dev, tb, data, extack); 1269 if (err) 1270 unregister_netdevice(dev); 1271 return err; 1272 } 1273 1274 static size_t br_get_size(const struct net_device *brdev) 1275 { 1276 return nla_total_size(sizeof(u32)) + /* IFLA_BR_FORWARD_DELAY */ 1277 nla_total_size(sizeof(u32)) + /* IFLA_BR_HELLO_TIME */ 1278 nla_total_size(sizeof(u32)) + /* IFLA_BR_MAX_AGE */ 1279 nla_total_size(sizeof(u32)) + /* IFLA_BR_AGEING_TIME */ 1280 nla_total_size(sizeof(u32)) + /* IFLA_BR_STP_STATE */ 1281 nla_total_size(sizeof(u16)) + /* IFLA_BR_PRIORITY */ 1282 nla_total_size(sizeof(u8)) + /* IFLA_BR_VLAN_FILTERING */ 1283 #ifdef CONFIG_BRIDGE_VLAN_FILTERING 1284 nla_total_size(sizeof(__be16)) + /* IFLA_BR_VLAN_PROTOCOL */ 1285 nla_total_size(sizeof(u16)) + /* IFLA_BR_VLAN_DEFAULT_PVID */ 1286 nla_total_size(sizeof(u8)) + /* IFLA_BR_VLAN_STATS_ENABLED */ 1287 #endif 1288 nla_total_size(sizeof(u16)) + /* IFLA_BR_GROUP_FWD_MASK */ 1289 nla_total_size(sizeof(struct ifla_bridge_id)) + /* IFLA_BR_ROOT_ID */ 1290 nla_total_size(sizeof(struct ifla_bridge_id)) + /* IFLA_BR_BRIDGE_ID */ 1291 nla_total_size(sizeof(u16)) + /* IFLA_BR_ROOT_PORT */ 1292 nla_total_size(sizeof(u32)) + /* IFLA_BR_ROOT_PATH_COST */ 1293 nla_total_size(sizeof(u8)) + /* IFLA_BR_TOPOLOGY_CHANGE */ 1294 nla_total_size(sizeof(u8)) + /* IFLA_BR_TOPOLOGY_CHANGE_DETECTED */ 1295 nla_total_size_64bit(sizeof(u64)) + /* IFLA_BR_HELLO_TIMER */ 1296 nla_total_size_64bit(sizeof(u64)) + /* IFLA_BR_TCN_TIMER */ 1297 nla_total_size_64bit(sizeof(u64)) + /* IFLA_BR_TOPOLOGY_CHANGE_TIMER */ 1298 nla_total_size_64bit(sizeof(u64)) + /* IFLA_BR_GC_TIMER */ 1299 nla_total_size(ETH_ALEN) + /* IFLA_BR_GROUP_ADDR */ 1300 #ifdef CONFIG_BRIDGE_IGMP_SNOOPING 1301 nla_total_size(sizeof(u8)) + /* IFLA_BR_MCAST_ROUTER */ 1302 nla_total_size(sizeof(u8)) + /* IFLA_BR_MCAST_SNOOPING */ 1303 nla_total_size(sizeof(u8)) + /* IFLA_BR_MCAST_QUERY_USE_IFADDR */ 1304 nla_total_size(sizeof(u8)) + /* IFLA_BR_MCAST_QUERIER */ 1305 nla_total_size(sizeof(u8)) + /* IFLA_BR_MCAST_STATS_ENABLED */ 1306 nla_total_size(sizeof(u32)) + /* IFLA_BR_MCAST_HASH_ELASTICITY */ 1307 nla_total_size(sizeof(u32)) + /* IFLA_BR_MCAST_HASH_MAX */ 1308 nla_total_size(sizeof(u32)) + /* IFLA_BR_MCAST_LAST_MEMBER_CNT */ 1309 nla_total_size(sizeof(u32)) + /* IFLA_BR_MCAST_STARTUP_QUERY_CNT */ 1310 nla_total_size_64bit(sizeof(u64)) + /* IFLA_BR_MCAST_LAST_MEMBER_INTVL */ 1311 nla_total_size_64bit(sizeof(u64)) + /* IFLA_BR_MCAST_MEMBERSHIP_INTVL */ 1312 nla_total_size_64bit(sizeof(u64)) + /* IFLA_BR_MCAST_QUERIER_INTVL */ 1313 nla_total_size_64bit(sizeof(u64)) + /* IFLA_BR_MCAST_QUERY_INTVL */ 1314 nla_total_size_64bit(sizeof(u64)) + /* IFLA_BR_MCAST_QUERY_RESPONSE_INTVL */ 1315 nla_total_size_64bit(sizeof(u64)) + /* IFLA_BR_MCAST_STARTUP_QUERY_INTVL */ 1316 nla_total_size(sizeof(u8)) + /* IFLA_BR_MCAST_IGMP_VERSION */ 1317 nla_total_size(sizeof(u8)) + /* IFLA_BR_MCAST_MLD_VERSION */ 1318 #endif 1319 #if IS_ENABLED(CONFIG_BRIDGE_NETFILTER) 1320 nla_total_size(sizeof(u8)) + /* IFLA_BR_NF_CALL_IPTABLES */ 1321 nla_total_size(sizeof(u8)) + /* IFLA_BR_NF_CALL_IP6TABLES */ 1322 nla_total_size(sizeof(u8)) + /* IFLA_BR_NF_CALL_ARPTABLES */ 1323 #endif 1324 0; 1325 } 1326 1327 static int br_fill_info(struct sk_buff *skb, const struct net_device *brdev) 1328 { 1329 struct net_bridge *br = netdev_priv(brdev); 1330 u32 forward_delay = jiffies_to_clock_t(br->forward_delay); 1331 u32 hello_time = jiffies_to_clock_t(br->hello_time); 1332 u32 age_time = jiffies_to_clock_t(br->max_age); 1333 u32 ageing_time = jiffies_to_clock_t(br->ageing_time); 1334 u32 stp_enabled = br->stp_enabled; 1335 u16 priority = (br->bridge_id.prio[0] << 8) | br->bridge_id.prio[1]; 1336 u8 vlan_enabled = br_vlan_enabled(br->dev); 1337 u64 clockval; 1338 1339 clockval = br_timer_value(&br->hello_timer); 1340 if (nla_put_u64_64bit(skb, IFLA_BR_HELLO_TIMER, clockval, IFLA_BR_PAD)) 1341 return -EMSGSIZE; 1342 clockval = br_timer_value(&br->tcn_timer); 1343 if (nla_put_u64_64bit(skb, IFLA_BR_TCN_TIMER, clockval, IFLA_BR_PAD)) 1344 return -EMSGSIZE; 1345 clockval = br_timer_value(&br->topology_change_timer); 1346 if (nla_put_u64_64bit(skb, IFLA_BR_TOPOLOGY_CHANGE_TIMER, clockval, 1347 IFLA_BR_PAD)) 1348 return -EMSGSIZE; 1349 clockval = br_timer_value(&br->gc_work.timer); 1350 if (nla_put_u64_64bit(skb, IFLA_BR_GC_TIMER, clockval, IFLA_BR_PAD)) 1351 return -EMSGSIZE; 1352 1353 if (nla_put_u32(skb, IFLA_BR_FORWARD_DELAY, forward_delay) || 1354 nla_put_u32(skb, IFLA_BR_HELLO_TIME, hello_time) || 1355 nla_put_u32(skb, IFLA_BR_MAX_AGE, age_time) || 1356 nla_put_u32(skb, IFLA_BR_AGEING_TIME, ageing_time) || 1357 nla_put_u32(skb, IFLA_BR_STP_STATE, stp_enabled) || 1358 nla_put_u16(skb, IFLA_BR_PRIORITY, priority) || 1359 nla_put_u8(skb, IFLA_BR_VLAN_FILTERING, vlan_enabled) || 1360 nla_put_u16(skb, IFLA_BR_GROUP_FWD_MASK, br->group_fwd_mask) || 1361 nla_put(skb, IFLA_BR_BRIDGE_ID, sizeof(struct ifla_bridge_id), 1362 &br->bridge_id) || 1363 nla_put(skb, IFLA_BR_ROOT_ID, sizeof(struct ifla_bridge_id), 1364 &br->designated_root) || 1365 nla_put_u16(skb, IFLA_BR_ROOT_PORT, br->root_port) || 1366 nla_put_u32(skb, IFLA_BR_ROOT_PATH_COST, br->root_path_cost) || 1367 nla_put_u8(skb, IFLA_BR_TOPOLOGY_CHANGE, br->topology_change) || 1368 nla_put_u8(skb, IFLA_BR_TOPOLOGY_CHANGE_DETECTED, 1369 br->topology_change_detected) || 1370 nla_put(skb, IFLA_BR_GROUP_ADDR, ETH_ALEN, br->group_addr)) 1371 return -EMSGSIZE; 1372 1373 #ifdef CONFIG_BRIDGE_VLAN_FILTERING 1374 if (nla_put_be16(skb, IFLA_BR_VLAN_PROTOCOL, br->vlan_proto) || 1375 nla_put_u16(skb, IFLA_BR_VLAN_DEFAULT_PVID, br->default_pvid) || 1376 nla_put_u8(skb, IFLA_BR_VLAN_STATS_ENABLED, br->vlan_stats_enabled)) 1377 return -EMSGSIZE; 1378 #endif 1379 #ifdef CONFIG_BRIDGE_IGMP_SNOOPING 1380 if (nla_put_u8(skb, IFLA_BR_MCAST_ROUTER, br->multicast_router) || 1381 nla_put_u8(skb, IFLA_BR_MCAST_SNOOPING, !br->multicast_disabled) || 1382 nla_put_u8(skb, IFLA_BR_MCAST_QUERY_USE_IFADDR, 1383 br->multicast_query_use_ifaddr) || 1384 nla_put_u8(skb, IFLA_BR_MCAST_QUERIER, br->multicast_querier) || 1385 nla_put_u8(skb, IFLA_BR_MCAST_STATS_ENABLED, 1386 br->multicast_stats_enabled) || 1387 nla_put_u32(skb, IFLA_BR_MCAST_HASH_ELASTICITY, 1388 br->hash_elasticity) || 1389 nla_put_u32(skb, IFLA_BR_MCAST_HASH_MAX, br->hash_max) || 1390 nla_put_u32(skb, IFLA_BR_MCAST_LAST_MEMBER_CNT, 1391 br->multicast_last_member_count) || 1392 nla_put_u32(skb, IFLA_BR_MCAST_STARTUP_QUERY_CNT, 1393 br->multicast_startup_query_count) || 1394 nla_put_u8(skb, IFLA_BR_MCAST_IGMP_VERSION, 1395 br->multicast_igmp_version)) 1396 return -EMSGSIZE; 1397 #if IS_ENABLED(CONFIG_IPV6) 1398 if (nla_put_u8(skb, IFLA_BR_MCAST_MLD_VERSION, 1399 br->multicast_mld_version)) 1400 return -EMSGSIZE; 1401 #endif 1402 clockval = jiffies_to_clock_t(br->multicast_last_member_interval); 1403 if (nla_put_u64_64bit(skb, IFLA_BR_MCAST_LAST_MEMBER_INTVL, clockval, 1404 IFLA_BR_PAD)) 1405 return -EMSGSIZE; 1406 clockval = jiffies_to_clock_t(br->multicast_membership_interval); 1407 if (nla_put_u64_64bit(skb, IFLA_BR_MCAST_MEMBERSHIP_INTVL, clockval, 1408 IFLA_BR_PAD)) 1409 return -EMSGSIZE; 1410 clockval = jiffies_to_clock_t(br->multicast_querier_interval); 1411 if (nla_put_u64_64bit(skb, IFLA_BR_MCAST_QUERIER_INTVL, clockval, 1412 IFLA_BR_PAD)) 1413 return -EMSGSIZE; 1414 clockval = jiffies_to_clock_t(br->multicast_query_interval); 1415 if (nla_put_u64_64bit(skb, IFLA_BR_MCAST_QUERY_INTVL, clockval, 1416 IFLA_BR_PAD)) 1417 return -EMSGSIZE; 1418 clockval = jiffies_to_clock_t(br->multicast_query_response_interval); 1419 if (nla_put_u64_64bit(skb, IFLA_BR_MCAST_QUERY_RESPONSE_INTVL, clockval, 1420 IFLA_BR_PAD)) 1421 return -EMSGSIZE; 1422 clockval = jiffies_to_clock_t(br->multicast_startup_query_interval); 1423 if (nla_put_u64_64bit(skb, IFLA_BR_MCAST_STARTUP_QUERY_INTVL, clockval, 1424 IFLA_BR_PAD)) 1425 return -EMSGSIZE; 1426 #endif 1427 #if IS_ENABLED(CONFIG_BRIDGE_NETFILTER) 1428 if (nla_put_u8(skb, IFLA_BR_NF_CALL_IPTABLES, 1429 br->nf_call_iptables ? 1 : 0) || 1430 nla_put_u8(skb, IFLA_BR_NF_CALL_IP6TABLES, 1431 br->nf_call_ip6tables ? 1 : 0) || 1432 nla_put_u8(skb, IFLA_BR_NF_CALL_ARPTABLES, 1433 br->nf_call_arptables ? 1 : 0)) 1434 return -EMSGSIZE; 1435 #endif 1436 1437 return 0; 1438 } 1439 1440 static size_t br_get_linkxstats_size(const struct net_device *dev, int attr) 1441 { 1442 struct net_bridge_port *p = NULL; 1443 struct net_bridge_vlan_group *vg; 1444 struct net_bridge_vlan *v; 1445 struct net_bridge *br; 1446 int numvls = 0; 1447 1448 switch (attr) { 1449 case IFLA_STATS_LINK_XSTATS: 1450 br = netdev_priv(dev); 1451 vg = br_vlan_group(br); 1452 break; 1453 case IFLA_STATS_LINK_XSTATS_SLAVE: 1454 p = br_port_get_rtnl(dev); 1455 if (!p) 1456 return 0; 1457 br = p->br; 1458 vg = nbp_vlan_group(p); 1459 break; 1460 default: 1461 return 0; 1462 } 1463 1464 if (vg) { 1465 /* we need to count all, even placeholder entries */ 1466 list_for_each_entry(v, &vg->vlan_list, vlist) 1467 numvls++; 1468 } 1469 1470 return numvls * nla_total_size(sizeof(struct bridge_vlan_xstats)) + 1471 nla_total_size(sizeof(struct br_mcast_stats)) + 1472 nla_total_size(0); 1473 } 1474 1475 static int br_fill_linkxstats(struct sk_buff *skb, 1476 const struct net_device *dev, 1477 int *prividx, int attr) 1478 { 1479 struct nlattr *nla __maybe_unused; 1480 struct net_bridge_port *p = NULL; 1481 struct net_bridge_vlan_group *vg; 1482 struct net_bridge_vlan *v; 1483 struct net_bridge *br; 1484 struct nlattr *nest; 1485 int vl_idx = 0; 1486 1487 switch (attr) { 1488 case IFLA_STATS_LINK_XSTATS: 1489 br = netdev_priv(dev); 1490 vg = br_vlan_group(br); 1491 break; 1492 case IFLA_STATS_LINK_XSTATS_SLAVE: 1493 p = br_port_get_rtnl(dev); 1494 if (!p) 1495 return 0; 1496 br = p->br; 1497 vg = nbp_vlan_group(p); 1498 break; 1499 default: 1500 return -EINVAL; 1501 } 1502 1503 nest = nla_nest_start(skb, LINK_XSTATS_TYPE_BRIDGE); 1504 if (!nest) 1505 return -EMSGSIZE; 1506 1507 if (vg) { 1508 u16 pvid; 1509 1510 pvid = br_get_pvid(vg); 1511 list_for_each_entry(v, &vg->vlan_list, vlist) { 1512 struct bridge_vlan_xstats vxi; 1513 struct br_vlan_stats stats; 1514 1515 if (++vl_idx < *prividx) 1516 continue; 1517 memset(&vxi, 0, sizeof(vxi)); 1518 vxi.vid = v->vid; 1519 vxi.flags = v->flags; 1520 if (v->vid == pvid) 1521 vxi.flags |= BRIDGE_VLAN_INFO_PVID; 1522 br_vlan_get_stats(v, &stats); 1523 vxi.rx_bytes = stats.rx_bytes; 1524 vxi.rx_packets = stats.rx_packets; 1525 vxi.tx_bytes = stats.tx_bytes; 1526 vxi.tx_packets = stats.tx_packets; 1527 1528 if (nla_put(skb, BRIDGE_XSTATS_VLAN, sizeof(vxi), &vxi)) 1529 goto nla_put_failure; 1530 } 1531 } 1532 1533 #ifdef CONFIG_BRIDGE_IGMP_SNOOPING 1534 if (++vl_idx >= *prividx) { 1535 nla = nla_reserve_64bit(skb, BRIDGE_XSTATS_MCAST, 1536 sizeof(struct br_mcast_stats), 1537 BRIDGE_XSTATS_PAD); 1538 if (!nla) 1539 goto nla_put_failure; 1540 br_multicast_get_stats(br, p, nla_data(nla)); 1541 } 1542 #endif 1543 nla_nest_end(skb, nest); 1544 *prividx = 0; 1545 1546 return 0; 1547 1548 nla_put_failure: 1549 nla_nest_end(skb, nest); 1550 *prividx = vl_idx; 1551 1552 return -EMSGSIZE; 1553 } 1554 1555 static struct rtnl_af_ops br_af_ops __read_mostly = { 1556 .family = AF_BRIDGE, 1557 .get_link_af_size = br_get_link_af_size_filtered, 1558 }; 1559 1560 struct rtnl_link_ops br_link_ops __read_mostly = { 1561 .kind = "bridge", 1562 .priv_size = sizeof(struct net_bridge), 1563 .setup = br_dev_setup, 1564 .maxtype = IFLA_BR_MAX, 1565 .policy = br_policy, 1566 .validate = br_validate, 1567 .newlink = br_dev_newlink, 1568 .changelink = br_changelink, 1569 .dellink = br_dev_delete, 1570 .get_size = br_get_size, 1571 .fill_info = br_fill_info, 1572 .fill_linkxstats = br_fill_linkxstats, 1573 .get_linkxstats_size = br_get_linkxstats_size, 1574 1575 .slave_maxtype = IFLA_BRPORT_MAX, 1576 .slave_policy = br_port_policy, 1577 .slave_changelink = br_port_slave_changelink, 1578 .get_slave_size = br_port_get_slave_size, 1579 .fill_slave_info = br_port_fill_slave_info, 1580 }; 1581 1582 int __init br_netlink_init(void) 1583 { 1584 int err; 1585 1586 br_mdb_init(); 1587 rtnl_af_register(&br_af_ops); 1588 1589 err = rtnl_link_register(&br_link_ops); 1590 if (err) 1591 goto out_af; 1592 1593 return 0; 1594 1595 out_af: 1596 rtnl_af_unregister(&br_af_ops); 1597 br_mdb_uninit(); 1598 return err; 1599 } 1600 1601 void br_netlink_fini(void) 1602 { 1603 br_mdb_uninit(); 1604 rtnl_af_unregister(&br_af_ops); 1605 rtnl_link_unregister(&br_link_ops); 1606 } 1607