1 // SPDX-License-Identifier: GPL-2.0-or-later 2 /* 3 * Copyright Gavin Shan, IBM Corporation 2016. 4 */ 5 6 #include <linux/module.h> 7 #include <linux/kernel.h> 8 #include <linux/init.h> 9 #include <linux/netdevice.h> 10 #include <linux/skbuff.h> 11 #include <linux/of.h> 12 #include <linux/platform_device.h> 13 14 #include <net/ncsi.h> 15 #include <net/net_namespace.h> 16 #include <net/sock.h> 17 #include <net/addrconf.h> 18 #include <net/ipv6.h> 19 #include <net/genetlink.h> 20 21 #include "internal.h" 22 #include "ncsi-pkt.h" 23 #include "ncsi-netlink.h" 24 25 LIST_HEAD(ncsi_dev_list); 26 DEFINE_SPINLOCK(ncsi_dev_lock); 27 28 bool ncsi_channel_has_link(struct ncsi_channel *channel) 29 { 30 return !!(channel->modes[NCSI_MODE_LINK].data[2] & 0x1); 31 } 32 33 bool ncsi_channel_is_last(struct ncsi_dev_priv *ndp, 34 struct ncsi_channel *channel) 35 { 36 struct ncsi_package *np; 37 struct ncsi_channel *nc; 38 39 NCSI_FOR_EACH_PACKAGE(ndp, np) 40 NCSI_FOR_EACH_CHANNEL(np, nc) { 41 if (nc == channel) 42 continue; 43 if (nc->state == NCSI_CHANNEL_ACTIVE && 44 ncsi_channel_has_link(nc)) 45 return false; 46 } 47 48 return true; 49 } 50 51 static void ncsi_report_link(struct ncsi_dev_priv *ndp, bool force_down) 52 { 53 struct ncsi_dev *nd = &ndp->ndev; 54 struct ncsi_package *np; 55 struct ncsi_channel *nc; 56 unsigned long flags; 57 58 nd->state = ncsi_dev_state_functional; 59 if (force_down) { 60 nd->link_up = 0; 61 goto report; 62 } 63 64 nd->link_up = 0; 65 NCSI_FOR_EACH_PACKAGE(ndp, np) { 66 NCSI_FOR_EACH_CHANNEL(np, nc) { 67 spin_lock_irqsave(&nc->lock, flags); 68 69 if (!list_empty(&nc->link) || 70 nc->state != NCSI_CHANNEL_ACTIVE) { 71 spin_unlock_irqrestore(&nc->lock, flags); 72 continue; 73 } 74 75 if (ncsi_channel_has_link(nc)) { 76 spin_unlock_irqrestore(&nc->lock, flags); 77 nd->link_up = 1; 78 goto report; 79 } 80 81 spin_unlock_irqrestore(&nc->lock, flags); 82 } 83 } 84 85 report: 86 nd->handler(nd); 87 } 88 89 static void ncsi_channel_monitor(struct timer_list *t) 90 { 91 struct ncsi_channel *nc = from_timer(nc, t, monitor.timer); 92 struct ncsi_package *np = nc->package; 93 struct ncsi_dev_priv *ndp = np->ndp; 94 struct ncsi_channel_mode *ncm; 95 struct ncsi_cmd_arg nca; 96 bool enabled, chained; 97 unsigned int monitor_state; 98 unsigned long flags; 99 int state, ret; 100 101 spin_lock_irqsave(&nc->lock, flags); 102 state = nc->state; 103 chained = !list_empty(&nc->link); 104 enabled = nc->monitor.enabled; 105 monitor_state = nc->monitor.state; 106 spin_unlock_irqrestore(&nc->lock, flags); 107 108 if (!enabled || chained) { 109 ncsi_stop_channel_monitor(nc); 110 return; 111 } 112 if (state != NCSI_CHANNEL_INACTIVE && 113 state != NCSI_CHANNEL_ACTIVE) { 114 ncsi_stop_channel_monitor(nc); 115 return; 116 } 117 118 switch (monitor_state) { 119 case NCSI_CHANNEL_MONITOR_START: 120 case NCSI_CHANNEL_MONITOR_RETRY: 121 nca.ndp = ndp; 122 nca.package = np->id; 123 nca.channel = nc->id; 124 nca.type = NCSI_PKT_CMD_GLS; 125 nca.req_flags = 0; 126 ret = ncsi_xmit_cmd(&nca); 127 if (ret) 128 netdev_err(ndp->ndev.dev, "Error %d sending GLS\n", 129 ret); 130 break; 131 case NCSI_CHANNEL_MONITOR_WAIT ... NCSI_CHANNEL_MONITOR_WAIT_MAX: 132 break; 133 default: 134 netdev_err(ndp->ndev.dev, "NCSI Channel %d timed out!\n", 135 nc->id); 136 ncsi_report_link(ndp, true); 137 ndp->flags |= NCSI_DEV_RESHUFFLE; 138 139 ncsi_stop_channel_monitor(nc); 140 141 ncm = &nc->modes[NCSI_MODE_LINK]; 142 spin_lock_irqsave(&nc->lock, flags); 143 nc->state = NCSI_CHANNEL_INVISIBLE; 144 ncm->data[2] &= ~0x1; 145 spin_unlock_irqrestore(&nc->lock, flags); 146 147 spin_lock_irqsave(&ndp->lock, flags); 148 nc->state = NCSI_CHANNEL_ACTIVE; 149 list_add_tail_rcu(&nc->link, &ndp->channel_queue); 150 spin_unlock_irqrestore(&ndp->lock, flags); 151 ncsi_process_next_channel(ndp); 152 return; 153 } 154 155 spin_lock_irqsave(&nc->lock, flags); 156 nc->monitor.state++; 157 spin_unlock_irqrestore(&nc->lock, flags); 158 mod_timer(&nc->monitor.timer, jiffies + HZ); 159 } 160 161 void ncsi_start_channel_monitor(struct ncsi_channel *nc) 162 { 163 unsigned long flags; 164 165 spin_lock_irqsave(&nc->lock, flags); 166 WARN_ON_ONCE(nc->monitor.enabled); 167 nc->monitor.enabled = true; 168 nc->monitor.state = NCSI_CHANNEL_MONITOR_START; 169 spin_unlock_irqrestore(&nc->lock, flags); 170 171 mod_timer(&nc->monitor.timer, jiffies + HZ); 172 } 173 174 void ncsi_stop_channel_monitor(struct ncsi_channel *nc) 175 { 176 unsigned long flags; 177 178 spin_lock_irqsave(&nc->lock, flags); 179 if (!nc->monitor.enabled) { 180 spin_unlock_irqrestore(&nc->lock, flags); 181 return; 182 } 183 nc->monitor.enabled = false; 184 spin_unlock_irqrestore(&nc->lock, flags); 185 186 del_timer_sync(&nc->monitor.timer); 187 } 188 189 struct ncsi_channel *ncsi_find_channel(struct ncsi_package *np, 190 unsigned char id) 191 { 192 struct ncsi_channel *nc; 193 194 NCSI_FOR_EACH_CHANNEL(np, nc) { 195 if (nc->id == id) 196 return nc; 197 } 198 199 return NULL; 200 } 201 202 struct ncsi_channel *ncsi_add_channel(struct ncsi_package *np, unsigned char id) 203 { 204 struct ncsi_channel *nc, *tmp; 205 int index; 206 unsigned long flags; 207 208 nc = kzalloc(sizeof(*nc), GFP_ATOMIC); 209 if (!nc) 210 return NULL; 211 212 nc->id = id; 213 nc->package = np; 214 nc->state = NCSI_CHANNEL_INACTIVE; 215 nc->monitor.enabled = false; 216 timer_setup(&nc->monitor.timer, ncsi_channel_monitor, 0); 217 spin_lock_init(&nc->lock); 218 INIT_LIST_HEAD(&nc->link); 219 for (index = 0; index < NCSI_CAP_MAX; index++) 220 nc->caps[index].index = index; 221 for (index = 0; index < NCSI_MODE_MAX; index++) 222 nc->modes[index].index = index; 223 224 spin_lock_irqsave(&np->lock, flags); 225 tmp = ncsi_find_channel(np, id); 226 if (tmp) { 227 spin_unlock_irqrestore(&np->lock, flags); 228 kfree(nc); 229 return tmp; 230 } 231 232 list_add_tail_rcu(&nc->node, &np->channels); 233 np->channel_num++; 234 spin_unlock_irqrestore(&np->lock, flags); 235 236 return nc; 237 } 238 239 static void ncsi_remove_channel(struct ncsi_channel *nc) 240 { 241 struct ncsi_package *np = nc->package; 242 unsigned long flags; 243 244 spin_lock_irqsave(&nc->lock, flags); 245 246 /* Release filters */ 247 kfree(nc->mac_filter.addrs); 248 kfree(nc->vlan_filter.vids); 249 250 nc->state = NCSI_CHANNEL_INACTIVE; 251 spin_unlock_irqrestore(&nc->lock, flags); 252 ncsi_stop_channel_monitor(nc); 253 254 /* Remove and free channel */ 255 spin_lock_irqsave(&np->lock, flags); 256 list_del_rcu(&nc->node); 257 np->channel_num--; 258 spin_unlock_irqrestore(&np->lock, flags); 259 260 kfree(nc); 261 } 262 263 struct ncsi_package *ncsi_find_package(struct ncsi_dev_priv *ndp, 264 unsigned char id) 265 { 266 struct ncsi_package *np; 267 268 NCSI_FOR_EACH_PACKAGE(ndp, np) { 269 if (np->id == id) 270 return np; 271 } 272 273 return NULL; 274 } 275 276 struct ncsi_package *ncsi_add_package(struct ncsi_dev_priv *ndp, 277 unsigned char id) 278 { 279 struct ncsi_package *np, *tmp; 280 unsigned long flags; 281 282 np = kzalloc(sizeof(*np), GFP_ATOMIC); 283 if (!np) 284 return NULL; 285 286 np->id = id; 287 np->ndp = ndp; 288 spin_lock_init(&np->lock); 289 INIT_LIST_HEAD(&np->channels); 290 np->channel_whitelist = UINT_MAX; 291 292 spin_lock_irqsave(&ndp->lock, flags); 293 tmp = ncsi_find_package(ndp, id); 294 if (tmp) { 295 spin_unlock_irqrestore(&ndp->lock, flags); 296 kfree(np); 297 return tmp; 298 } 299 300 list_add_tail_rcu(&np->node, &ndp->packages); 301 ndp->package_num++; 302 spin_unlock_irqrestore(&ndp->lock, flags); 303 304 return np; 305 } 306 307 void ncsi_remove_package(struct ncsi_package *np) 308 { 309 struct ncsi_dev_priv *ndp = np->ndp; 310 struct ncsi_channel *nc, *tmp; 311 unsigned long flags; 312 313 /* Release all child channels */ 314 list_for_each_entry_safe(nc, tmp, &np->channels, node) 315 ncsi_remove_channel(nc); 316 317 /* Remove and free package */ 318 spin_lock_irqsave(&ndp->lock, flags); 319 list_del_rcu(&np->node); 320 ndp->package_num--; 321 spin_unlock_irqrestore(&ndp->lock, flags); 322 323 kfree(np); 324 } 325 326 void ncsi_find_package_and_channel(struct ncsi_dev_priv *ndp, 327 unsigned char id, 328 struct ncsi_package **np, 329 struct ncsi_channel **nc) 330 { 331 struct ncsi_package *p; 332 struct ncsi_channel *c; 333 334 p = ncsi_find_package(ndp, NCSI_PACKAGE_INDEX(id)); 335 c = p ? ncsi_find_channel(p, NCSI_CHANNEL_INDEX(id)) : NULL; 336 337 if (np) 338 *np = p; 339 if (nc) 340 *nc = c; 341 } 342 343 /* For two consecutive NCSI commands, the packet IDs shouldn't 344 * be same. Otherwise, the bogus response might be replied. So 345 * the available IDs are allocated in round-robin fashion. 346 */ 347 struct ncsi_request *ncsi_alloc_request(struct ncsi_dev_priv *ndp, 348 unsigned int req_flags) 349 { 350 struct ncsi_request *nr = NULL; 351 int i, limit = ARRAY_SIZE(ndp->requests); 352 unsigned long flags; 353 354 /* Check if there is one available request until the ceiling */ 355 spin_lock_irqsave(&ndp->lock, flags); 356 for (i = ndp->request_id; i < limit; i++) { 357 if (ndp->requests[i].used) 358 continue; 359 360 nr = &ndp->requests[i]; 361 nr->used = true; 362 nr->flags = req_flags; 363 ndp->request_id = i + 1; 364 goto found; 365 } 366 367 /* Fail back to check from the starting cursor */ 368 for (i = NCSI_REQ_START_IDX; i < ndp->request_id; i++) { 369 if (ndp->requests[i].used) 370 continue; 371 372 nr = &ndp->requests[i]; 373 nr->used = true; 374 nr->flags = req_flags; 375 ndp->request_id = i + 1; 376 goto found; 377 } 378 379 found: 380 spin_unlock_irqrestore(&ndp->lock, flags); 381 return nr; 382 } 383 384 void ncsi_free_request(struct ncsi_request *nr) 385 { 386 struct ncsi_dev_priv *ndp = nr->ndp; 387 struct sk_buff *cmd, *rsp; 388 unsigned long flags; 389 bool driven; 390 391 if (nr->enabled) { 392 nr->enabled = false; 393 del_timer_sync(&nr->timer); 394 } 395 396 spin_lock_irqsave(&ndp->lock, flags); 397 cmd = nr->cmd; 398 rsp = nr->rsp; 399 nr->cmd = NULL; 400 nr->rsp = NULL; 401 nr->used = false; 402 driven = !!(nr->flags & NCSI_REQ_FLAG_EVENT_DRIVEN); 403 spin_unlock_irqrestore(&ndp->lock, flags); 404 405 if (driven && cmd && --ndp->pending_req_num == 0) 406 schedule_work(&ndp->work); 407 408 /* Release command and response */ 409 consume_skb(cmd); 410 consume_skb(rsp); 411 } 412 413 struct ncsi_dev *ncsi_find_dev(struct net_device *dev) 414 { 415 struct ncsi_dev_priv *ndp; 416 417 NCSI_FOR_EACH_DEV(ndp) { 418 if (ndp->ndev.dev == dev) 419 return &ndp->ndev; 420 } 421 422 return NULL; 423 } 424 425 static void ncsi_request_timeout(struct timer_list *t) 426 { 427 struct ncsi_request *nr = from_timer(nr, t, timer); 428 struct ncsi_dev_priv *ndp = nr->ndp; 429 struct ncsi_cmd_pkt *cmd; 430 struct ncsi_package *np; 431 struct ncsi_channel *nc; 432 unsigned long flags; 433 434 /* If the request already had associated response, 435 * let the response handler to release it. 436 */ 437 spin_lock_irqsave(&ndp->lock, flags); 438 nr->enabled = false; 439 if (nr->rsp || !nr->cmd) { 440 spin_unlock_irqrestore(&ndp->lock, flags); 441 return; 442 } 443 spin_unlock_irqrestore(&ndp->lock, flags); 444 445 if (nr->flags == NCSI_REQ_FLAG_NETLINK_DRIVEN) { 446 if (nr->cmd) { 447 /* Find the package */ 448 cmd = (struct ncsi_cmd_pkt *) 449 skb_network_header(nr->cmd); 450 ncsi_find_package_and_channel(ndp, 451 cmd->cmd.common.channel, 452 &np, &nc); 453 ncsi_send_netlink_timeout(nr, np, nc); 454 } 455 } 456 457 /* Release the request */ 458 ncsi_free_request(nr); 459 } 460 461 static void ncsi_suspend_channel(struct ncsi_dev_priv *ndp) 462 { 463 struct ncsi_dev *nd = &ndp->ndev; 464 struct ncsi_package *np; 465 struct ncsi_channel *nc, *tmp; 466 struct ncsi_cmd_arg nca; 467 unsigned long flags; 468 int ret; 469 470 np = ndp->active_package; 471 nc = ndp->active_channel; 472 nca.ndp = ndp; 473 nca.req_flags = NCSI_REQ_FLAG_EVENT_DRIVEN; 474 switch (nd->state) { 475 case ncsi_dev_state_suspend: 476 nd->state = ncsi_dev_state_suspend_select; 477 fallthrough; 478 case ncsi_dev_state_suspend_select: 479 ndp->pending_req_num = 1; 480 481 nca.type = NCSI_PKT_CMD_SP; 482 nca.package = np->id; 483 nca.channel = NCSI_RESERVED_CHANNEL; 484 if (ndp->flags & NCSI_DEV_HWA) 485 nca.bytes[0] = 0; 486 else 487 nca.bytes[0] = 1; 488 489 /* To retrieve the last link states of channels in current 490 * package when current active channel needs fail over to 491 * another one. It means we will possibly select another 492 * channel as next active one. The link states of channels 493 * are most important factor of the selection. So we need 494 * accurate link states. Unfortunately, the link states on 495 * inactive channels can't be updated with LSC AEN in time. 496 */ 497 if (ndp->flags & NCSI_DEV_RESHUFFLE) 498 nd->state = ncsi_dev_state_suspend_gls; 499 else 500 nd->state = ncsi_dev_state_suspend_dcnt; 501 ret = ncsi_xmit_cmd(&nca); 502 if (ret) 503 goto error; 504 505 break; 506 case ncsi_dev_state_suspend_gls: 507 ndp->pending_req_num = np->channel_num; 508 509 nca.type = NCSI_PKT_CMD_GLS; 510 nca.package = np->id; 511 512 nd->state = ncsi_dev_state_suspend_dcnt; 513 NCSI_FOR_EACH_CHANNEL(np, nc) { 514 nca.channel = nc->id; 515 ret = ncsi_xmit_cmd(&nca); 516 if (ret) 517 goto error; 518 } 519 520 break; 521 case ncsi_dev_state_suspend_dcnt: 522 ndp->pending_req_num = 1; 523 524 nca.type = NCSI_PKT_CMD_DCNT; 525 nca.package = np->id; 526 nca.channel = nc->id; 527 528 nd->state = ncsi_dev_state_suspend_dc; 529 ret = ncsi_xmit_cmd(&nca); 530 if (ret) 531 goto error; 532 533 break; 534 case ncsi_dev_state_suspend_dc: 535 ndp->pending_req_num = 1; 536 537 nca.type = NCSI_PKT_CMD_DC; 538 nca.package = np->id; 539 nca.channel = nc->id; 540 nca.bytes[0] = 1; 541 542 nd->state = ncsi_dev_state_suspend_deselect; 543 ret = ncsi_xmit_cmd(&nca); 544 if (ret) 545 goto error; 546 547 NCSI_FOR_EACH_CHANNEL(np, tmp) { 548 /* If there is another channel active on this package 549 * do not deselect the package. 550 */ 551 if (tmp != nc && tmp->state == NCSI_CHANNEL_ACTIVE) { 552 nd->state = ncsi_dev_state_suspend_done; 553 break; 554 } 555 } 556 break; 557 case ncsi_dev_state_suspend_deselect: 558 ndp->pending_req_num = 1; 559 560 nca.type = NCSI_PKT_CMD_DP; 561 nca.package = np->id; 562 nca.channel = NCSI_RESERVED_CHANNEL; 563 564 nd->state = ncsi_dev_state_suspend_done; 565 ret = ncsi_xmit_cmd(&nca); 566 if (ret) 567 goto error; 568 569 break; 570 case ncsi_dev_state_suspend_done: 571 spin_lock_irqsave(&nc->lock, flags); 572 nc->state = NCSI_CHANNEL_INACTIVE; 573 spin_unlock_irqrestore(&nc->lock, flags); 574 if (ndp->flags & NCSI_DEV_RESET) 575 ncsi_reset_dev(nd); 576 else 577 ncsi_process_next_channel(ndp); 578 break; 579 default: 580 netdev_warn(nd->dev, "Wrong NCSI state 0x%x in suspend\n", 581 nd->state); 582 } 583 584 return; 585 error: 586 nd->state = ncsi_dev_state_functional; 587 } 588 589 /* Check the VLAN filter bitmap for a set filter, and construct a 590 * "Set VLAN Filter - Disable" packet if found. 591 */ 592 static int clear_one_vid(struct ncsi_dev_priv *ndp, struct ncsi_channel *nc, 593 struct ncsi_cmd_arg *nca) 594 { 595 struct ncsi_channel_vlan_filter *ncf; 596 unsigned long flags; 597 void *bitmap; 598 int index; 599 u16 vid; 600 601 ncf = &nc->vlan_filter; 602 bitmap = &ncf->bitmap; 603 604 spin_lock_irqsave(&nc->lock, flags); 605 index = find_next_bit(bitmap, ncf->n_vids, 0); 606 if (index >= ncf->n_vids) { 607 spin_unlock_irqrestore(&nc->lock, flags); 608 return -1; 609 } 610 vid = ncf->vids[index]; 611 612 clear_bit(index, bitmap); 613 ncf->vids[index] = 0; 614 spin_unlock_irqrestore(&nc->lock, flags); 615 616 nca->type = NCSI_PKT_CMD_SVF; 617 nca->words[1] = vid; 618 /* HW filter index starts at 1 */ 619 nca->bytes[6] = index + 1; 620 nca->bytes[7] = 0x00; 621 return 0; 622 } 623 624 /* Find an outstanding VLAN tag and constuct a "Set VLAN Filter - Enable" 625 * packet. 626 */ 627 static int set_one_vid(struct ncsi_dev_priv *ndp, struct ncsi_channel *nc, 628 struct ncsi_cmd_arg *nca) 629 { 630 struct ncsi_channel_vlan_filter *ncf; 631 struct vlan_vid *vlan = NULL; 632 unsigned long flags; 633 int i, index; 634 void *bitmap; 635 u16 vid; 636 637 if (list_empty(&ndp->vlan_vids)) 638 return -1; 639 640 ncf = &nc->vlan_filter; 641 bitmap = &ncf->bitmap; 642 643 spin_lock_irqsave(&nc->lock, flags); 644 645 rcu_read_lock(); 646 list_for_each_entry_rcu(vlan, &ndp->vlan_vids, list) { 647 vid = vlan->vid; 648 for (i = 0; i < ncf->n_vids; i++) 649 if (ncf->vids[i] == vid) { 650 vid = 0; 651 break; 652 } 653 if (vid) 654 break; 655 } 656 rcu_read_unlock(); 657 658 if (!vid) { 659 /* No VLAN ID is not set */ 660 spin_unlock_irqrestore(&nc->lock, flags); 661 return -1; 662 } 663 664 index = find_next_zero_bit(bitmap, ncf->n_vids, 0); 665 if (index < 0 || index >= ncf->n_vids) { 666 netdev_err(ndp->ndev.dev, 667 "Channel %u already has all VLAN filters set\n", 668 nc->id); 669 spin_unlock_irqrestore(&nc->lock, flags); 670 return -1; 671 } 672 673 ncf->vids[index] = vid; 674 set_bit(index, bitmap); 675 spin_unlock_irqrestore(&nc->lock, flags); 676 677 nca->type = NCSI_PKT_CMD_SVF; 678 nca->words[1] = vid; 679 /* HW filter index starts at 1 */ 680 nca->bytes[6] = index + 1; 681 nca->bytes[7] = 0x01; 682 683 return 0; 684 } 685 686 #if IS_ENABLED(CONFIG_NCSI_OEM_CMD_GET_MAC) 687 688 /* NCSI OEM Command APIs */ 689 static int ncsi_oem_gma_handler_bcm(struct ncsi_cmd_arg *nca) 690 { 691 unsigned char data[NCSI_OEM_BCM_CMD_GMA_LEN]; 692 int ret = 0; 693 694 nca->payload = NCSI_OEM_BCM_CMD_GMA_LEN; 695 696 memset(data, 0, NCSI_OEM_BCM_CMD_GMA_LEN); 697 *(unsigned int *)data = ntohl(NCSI_OEM_MFR_BCM_ID); 698 data[5] = NCSI_OEM_BCM_CMD_GMA; 699 700 nca->data = data; 701 702 ret = ncsi_xmit_cmd(nca); 703 if (ret) 704 netdev_err(nca->ndp->ndev.dev, 705 "NCSI: Failed to transmit cmd 0x%x during configure\n", 706 nca->type); 707 return ret; 708 } 709 710 static int ncsi_oem_gma_handler_mlx(struct ncsi_cmd_arg *nca) 711 { 712 union { 713 u8 data_u8[NCSI_OEM_MLX_CMD_GMA_LEN]; 714 u32 data_u32[NCSI_OEM_MLX_CMD_GMA_LEN / sizeof(u32)]; 715 } u; 716 int ret = 0; 717 718 nca->payload = NCSI_OEM_MLX_CMD_GMA_LEN; 719 720 memset(&u, 0, sizeof(u)); 721 u.data_u32[0] = ntohl(NCSI_OEM_MFR_MLX_ID); 722 u.data_u8[5] = NCSI_OEM_MLX_CMD_GMA; 723 u.data_u8[6] = NCSI_OEM_MLX_CMD_GMA_PARAM; 724 725 nca->data = u.data_u8; 726 727 ret = ncsi_xmit_cmd(nca); 728 if (ret) 729 netdev_err(nca->ndp->ndev.dev, 730 "NCSI: Failed to transmit cmd 0x%x during configure\n", 731 nca->type); 732 return ret; 733 } 734 735 static int ncsi_oem_smaf_mlx(struct ncsi_cmd_arg *nca) 736 { 737 union { 738 u8 data_u8[NCSI_OEM_MLX_CMD_SMAF_LEN]; 739 u32 data_u32[NCSI_OEM_MLX_CMD_SMAF_LEN / sizeof(u32)]; 740 } u; 741 int ret = 0; 742 743 memset(&u, 0, sizeof(u)); 744 u.data_u32[0] = ntohl(NCSI_OEM_MFR_MLX_ID); 745 u.data_u8[5] = NCSI_OEM_MLX_CMD_SMAF; 746 u.data_u8[6] = NCSI_OEM_MLX_CMD_SMAF_PARAM; 747 memcpy(&u.data_u8[MLX_SMAF_MAC_ADDR_OFFSET], 748 nca->ndp->ndev.dev->dev_addr, ETH_ALEN); 749 u.data_u8[MLX_SMAF_MED_SUPPORT_OFFSET] = 750 (MLX_MC_RBT_AVL | MLX_MC_RBT_SUPPORT); 751 752 nca->payload = NCSI_OEM_MLX_CMD_SMAF_LEN; 753 nca->data = u.data_u8; 754 755 ret = ncsi_xmit_cmd(nca); 756 if (ret) 757 netdev_err(nca->ndp->ndev.dev, 758 "NCSI: Failed to transmit cmd 0x%x during probe\n", 759 nca->type); 760 return ret; 761 } 762 763 /* OEM Command handlers initialization */ 764 static struct ncsi_oem_gma_handler { 765 unsigned int mfr_id; 766 int (*handler)(struct ncsi_cmd_arg *nca); 767 } ncsi_oem_gma_handlers[] = { 768 { NCSI_OEM_MFR_BCM_ID, ncsi_oem_gma_handler_bcm }, 769 { NCSI_OEM_MFR_MLX_ID, ncsi_oem_gma_handler_mlx } 770 }; 771 772 static int ncsi_gma_handler(struct ncsi_cmd_arg *nca, unsigned int mf_id) 773 { 774 struct ncsi_oem_gma_handler *nch = NULL; 775 int i; 776 777 /* This function should only be called once, return if flag set */ 778 if (nca->ndp->gma_flag == 1) 779 return -1; 780 781 /* Find gma handler for given manufacturer id */ 782 for (i = 0; i < ARRAY_SIZE(ncsi_oem_gma_handlers); i++) { 783 if (ncsi_oem_gma_handlers[i].mfr_id == mf_id) { 784 if (ncsi_oem_gma_handlers[i].handler) 785 nch = &ncsi_oem_gma_handlers[i]; 786 break; 787 } 788 } 789 790 if (!nch) { 791 netdev_err(nca->ndp->ndev.dev, 792 "NCSI: No GMA handler available for MFR-ID (0x%x)\n", 793 mf_id); 794 return -1; 795 } 796 797 /* Get Mac address from NCSI device */ 798 return nch->handler(nca); 799 } 800 801 #endif /* CONFIG_NCSI_OEM_CMD_GET_MAC */ 802 803 /* Determine if a given channel from the channel_queue should be used for Tx */ 804 static bool ncsi_channel_is_tx(struct ncsi_dev_priv *ndp, 805 struct ncsi_channel *nc) 806 { 807 struct ncsi_channel_mode *ncm; 808 struct ncsi_channel *channel; 809 struct ncsi_package *np; 810 811 /* Check if any other channel has Tx enabled; a channel may have already 812 * been configured and removed from the channel queue. 813 */ 814 NCSI_FOR_EACH_PACKAGE(ndp, np) { 815 if (!ndp->multi_package && np != nc->package) 816 continue; 817 NCSI_FOR_EACH_CHANNEL(np, channel) { 818 ncm = &channel->modes[NCSI_MODE_TX_ENABLE]; 819 if (ncm->enable) 820 return false; 821 } 822 } 823 824 /* This channel is the preferred channel and has link */ 825 list_for_each_entry_rcu(channel, &ndp->channel_queue, link) { 826 np = channel->package; 827 if (np->preferred_channel && 828 ncsi_channel_has_link(np->preferred_channel)) { 829 return np->preferred_channel == nc; 830 } 831 } 832 833 /* This channel has link */ 834 if (ncsi_channel_has_link(nc)) 835 return true; 836 837 list_for_each_entry_rcu(channel, &ndp->channel_queue, link) 838 if (ncsi_channel_has_link(channel)) 839 return false; 840 841 /* No other channel has link; default to this one */ 842 return true; 843 } 844 845 /* Change the active Tx channel in a multi-channel setup */ 846 int ncsi_update_tx_channel(struct ncsi_dev_priv *ndp, 847 struct ncsi_package *package, 848 struct ncsi_channel *disable, 849 struct ncsi_channel *enable) 850 { 851 struct ncsi_cmd_arg nca; 852 struct ncsi_channel *nc; 853 struct ncsi_package *np; 854 int ret = 0; 855 856 if (!package->multi_channel && !ndp->multi_package) 857 netdev_warn(ndp->ndev.dev, 858 "NCSI: Trying to update Tx channel in single-channel mode\n"); 859 nca.ndp = ndp; 860 nca.req_flags = 0; 861 862 /* Find current channel with Tx enabled */ 863 NCSI_FOR_EACH_PACKAGE(ndp, np) { 864 if (disable) 865 break; 866 if (!ndp->multi_package && np != package) 867 continue; 868 869 NCSI_FOR_EACH_CHANNEL(np, nc) 870 if (nc->modes[NCSI_MODE_TX_ENABLE].enable) { 871 disable = nc; 872 break; 873 } 874 } 875 876 /* Find a suitable channel for Tx */ 877 NCSI_FOR_EACH_PACKAGE(ndp, np) { 878 if (enable) 879 break; 880 if (!ndp->multi_package && np != package) 881 continue; 882 if (!(ndp->package_whitelist & (0x1 << np->id))) 883 continue; 884 885 if (np->preferred_channel && 886 ncsi_channel_has_link(np->preferred_channel)) { 887 enable = np->preferred_channel; 888 break; 889 } 890 891 NCSI_FOR_EACH_CHANNEL(np, nc) { 892 if (!(np->channel_whitelist & 0x1 << nc->id)) 893 continue; 894 if (nc->state != NCSI_CHANNEL_ACTIVE) 895 continue; 896 if (ncsi_channel_has_link(nc)) { 897 enable = nc; 898 break; 899 } 900 } 901 } 902 903 if (disable == enable) 904 return -1; 905 906 if (!enable) 907 return -1; 908 909 if (disable) { 910 nca.channel = disable->id; 911 nca.package = disable->package->id; 912 nca.type = NCSI_PKT_CMD_DCNT; 913 ret = ncsi_xmit_cmd(&nca); 914 if (ret) 915 netdev_err(ndp->ndev.dev, 916 "Error %d sending DCNT\n", 917 ret); 918 } 919 920 netdev_info(ndp->ndev.dev, "NCSI: channel %u enables Tx\n", enable->id); 921 922 nca.channel = enable->id; 923 nca.package = enable->package->id; 924 nca.type = NCSI_PKT_CMD_ECNT; 925 ret = ncsi_xmit_cmd(&nca); 926 if (ret) 927 netdev_err(ndp->ndev.dev, 928 "Error %d sending ECNT\n", 929 ret); 930 931 return ret; 932 } 933 934 static void ncsi_configure_channel(struct ncsi_dev_priv *ndp) 935 { 936 struct ncsi_package *np = ndp->active_package; 937 struct ncsi_channel *nc = ndp->active_channel; 938 struct ncsi_channel *hot_nc = NULL; 939 struct ncsi_dev *nd = &ndp->ndev; 940 struct net_device *dev = nd->dev; 941 struct ncsi_cmd_arg nca; 942 unsigned char index; 943 unsigned long flags; 944 int ret; 945 946 nca.ndp = ndp; 947 nca.req_flags = NCSI_REQ_FLAG_EVENT_DRIVEN; 948 switch (nd->state) { 949 case ncsi_dev_state_config: 950 case ncsi_dev_state_config_sp: 951 ndp->pending_req_num = 1; 952 953 /* Select the specific package */ 954 nca.type = NCSI_PKT_CMD_SP; 955 if (ndp->flags & NCSI_DEV_HWA) 956 nca.bytes[0] = 0; 957 else 958 nca.bytes[0] = 1; 959 nca.package = np->id; 960 nca.channel = NCSI_RESERVED_CHANNEL; 961 ret = ncsi_xmit_cmd(&nca); 962 if (ret) { 963 netdev_err(ndp->ndev.dev, 964 "NCSI: Failed to transmit CMD_SP\n"); 965 goto error; 966 } 967 968 nd->state = ncsi_dev_state_config_cis; 969 break; 970 case ncsi_dev_state_config_cis: 971 ndp->pending_req_num = 1; 972 973 /* Clear initial state */ 974 nca.type = NCSI_PKT_CMD_CIS; 975 nca.package = np->id; 976 nca.channel = nc->id; 977 ret = ncsi_xmit_cmd(&nca); 978 if (ret) { 979 netdev_err(ndp->ndev.dev, 980 "NCSI: Failed to transmit CMD_CIS\n"); 981 goto error; 982 } 983 984 nd->state = ncsi_dev_state_config_oem_gma; 985 break; 986 case ncsi_dev_state_config_oem_gma: 987 nd->state = ncsi_dev_state_config_clear_vids; 988 ret = -1; 989 990 #if IS_ENABLED(CONFIG_NCSI_OEM_CMD_GET_MAC) 991 nca.type = NCSI_PKT_CMD_OEM; 992 nca.package = np->id; 993 nca.channel = nc->id; 994 ndp->pending_req_num = 1; 995 ret = ncsi_gma_handler(&nca, nc->version.mf_id); 996 #endif /* CONFIG_NCSI_OEM_CMD_GET_MAC */ 997 998 if (ret < 0) 999 schedule_work(&ndp->work); 1000 1001 break; 1002 case ncsi_dev_state_config_clear_vids: 1003 case ncsi_dev_state_config_svf: 1004 case ncsi_dev_state_config_ev: 1005 case ncsi_dev_state_config_sma: 1006 case ncsi_dev_state_config_ebf: 1007 case ncsi_dev_state_config_dgmf: 1008 case ncsi_dev_state_config_ecnt: 1009 case ncsi_dev_state_config_ec: 1010 case ncsi_dev_state_config_ae: 1011 case ncsi_dev_state_config_gls: 1012 ndp->pending_req_num = 1; 1013 1014 nca.package = np->id; 1015 nca.channel = nc->id; 1016 1017 /* Clear any active filters on the channel before setting */ 1018 if (nd->state == ncsi_dev_state_config_clear_vids) { 1019 ret = clear_one_vid(ndp, nc, &nca); 1020 if (ret) { 1021 nd->state = ncsi_dev_state_config_svf; 1022 schedule_work(&ndp->work); 1023 break; 1024 } 1025 /* Repeat */ 1026 nd->state = ncsi_dev_state_config_clear_vids; 1027 /* Add known VLAN tags to the filter */ 1028 } else if (nd->state == ncsi_dev_state_config_svf) { 1029 ret = set_one_vid(ndp, nc, &nca); 1030 if (ret) { 1031 nd->state = ncsi_dev_state_config_ev; 1032 schedule_work(&ndp->work); 1033 break; 1034 } 1035 /* Repeat */ 1036 nd->state = ncsi_dev_state_config_svf; 1037 /* Enable/Disable the VLAN filter */ 1038 } else if (nd->state == ncsi_dev_state_config_ev) { 1039 if (list_empty(&ndp->vlan_vids)) { 1040 nca.type = NCSI_PKT_CMD_DV; 1041 } else { 1042 nca.type = NCSI_PKT_CMD_EV; 1043 nca.bytes[3] = NCSI_CAP_VLAN_NO; 1044 } 1045 nd->state = ncsi_dev_state_config_sma; 1046 } else if (nd->state == ncsi_dev_state_config_sma) { 1047 /* Use first entry in unicast filter table. Note that 1048 * the MAC filter table starts from entry 1 instead of 1049 * 0. 1050 */ 1051 nca.type = NCSI_PKT_CMD_SMA; 1052 for (index = 0; index < 6; index++) 1053 nca.bytes[index] = dev->dev_addr[index]; 1054 nca.bytes[6] = 0x1; 1055 nca.bytes[7] = 0x1; 1056 nd->state = ncsi_dev_state_config_ebf; 1057 } else if (nd->state == ncsi_dev_state_config_ebf) { 1058 nca.type = NCSI_PKT_CMD_EBF; 1059 nca.dwords[0] = nc->caps[NCSI_CAP_BC].cap; 1060 /* if multicast global filtering is supported then 1061 * disable it so that all multicast packet will be 1062 * forwarded to management controller 1063 */ 1064 if (nc->caps[NCSI_CAP_GENERIC].cap & 1065 NCSI_CAP_GENERIC_MC) 1066 nd->state = ncsi_dev_state_config_dgmf; 1067 else if (ncsi_channel_is_tx(ndp, nc)) 1068 nd->state = ncsi_dev_state_config_ecnt; 1069 else 1070 nd->state = ncsi_dev_state_config_ec; 1071 } else if (nd->state == ncsi_dev_state_config_dgmf) { 1072 nca.type = NCSI_PKT_CMD_DGMF; 1073 if (ncsi_channel_is_tx(ndp, nc)) 1074 nd->state = ncsi_dev_state_config_ecnt; 1075 else 1076 nd->state = ncsi_dev_state_config_ec; 1077 } else if (nd->state == ncsi_dev_state_config_ecnt) { 1078 if (np->preferred_channel && 1079 nc != np->preferred_channel) 1080 netdev_info(ndp->ndev.dev, 1081 "NCSI: Tx failed over to channel %u\n", 1082 nc->id); 1083 nca.type = NCSI_PKT_CMD_ECNT; 1084 nd->state = ncsi_dev_state_config_ec; 1085 } else if (nd->state == ncsi_dev_state_config_ec) { 1086 /* Enable AEN if it's supported */ 1087 nca.type = NCSI_PKT_CMD_EC; 1088 nd->state = ncsi_dev_state_config_ae; 1089 if (!(nc->caps[NCSI_CAP_AEN].cap & NCSI_CAP_AEN_MASK)) 1090 nd->state = ncsi_dev_state_config_gls; 1091 } else if (nd->state == ncsi_dev_state_config_ae) { 1092 nca.type = NCSI_PKT_CMD_AE; 1093 nca.bytes[0] = 0; 1094 nca.dwords[1] = nc->caps[NCSI_CAP_AEN].cap; 1095 nd->state = ncsi_dev_state_config_gls; 1096 } else if (nd->state == ncsi_dev_state_config_gls) { 1097 nca.type = NCSI_PKT_CMD_GLS; 1098 nd->state = ncsi_dev_state_config_done; 1099 } 1100 1101 ret = ncsi_xmit_cmd(&nca); 1102 if (ret) { 1103 netdev_err(ndp->ndev.dev, 1104 "NCSI: Failed to transmit CMD %x\n", 1105 nca.type); 1106 goto error; 1107 } 1108 break; 1109 case ncsi_dev_state_config_done: 1110 netdev_dbg(ndp->ndev.dev, "NCSI: channel %u config done\n", 1111 nc->id); 1112 spin_lock_irqsave(&nc->lock, flags); 1113 nc->state = NCSI_CHANNEL_ACTIVE; 1114 1115 if (ndp->flags & NCSI_DEV_RESET) { 1116 /* A reset event happened during config, start it now */ 1117 nc->reconfigure_needed = false; 1118 spin_unlock_irqrestore(&nc->lock, flags); 1119 ncsi_reset_dev(nd); 1120 break; 1121 } 1122 1123 if (nc->reconfigure_needed) { 1124 /* This channel's configuration has been updated 1125 * part-way during the config state - start the 1126 * channel configuration over 1127 */ 1128 nc->reconfigure_needed = false; 1129 nc->state = NCSI_CHANNEL_INACTIVE; 1130 spin_unlock_irqrestore(&nc->lock, flags); 1131 1132 spin_lock_irqsave(&ndp->lock, flags); 1133 list_add_tail_rcu(&nc->link, &ndp->channel_queue); 1134 spin_unlock_irqrestore(&ndp->lock, flags); 1135 1136 netdev_dbg(dev, "Dirty NCSI channel state reset\n"); 1137 ncsi_process_next_channel(ndp); 1138 break; 1139 } 1140 1141 if (nc->modes[NCSI_MODE_LINK].data[2] & 0x1) { 1142 hot_nc = nc; 1143 } else { 1144 hot_nc = NULL; 1145 netdev_dbg(ndp->ndev.dev, 1146 "NCSI: channel %u link down after config\n", 1147 nc->id); 1148 } 1149 spin_unlock_irqrestore(&nc->lock, flags); 1150 1151 /* Update the hot channel */ 1152 spin_lock_irqsave(&ndp->lock, flags); 1153 ndp->hot_channel = hot_nc; 1154 spin_unlock_irqrestore(&ndp->lock, flags); 1155 1156 ncsi_start_channel_monitor(nc); 1157 ncsi_process_next_channel(ndp); 1158 break; 1159 default: 1160 netdev_alert(dev, "Wrong NCSI state 0x%x in config\n", 1161 nd->state); 1162 } 1163 1164 return; 1165 1166 error: 1167 ncsi_report_link(ndp, true); 1168 } 1169 1170 static int ncsi_choose_active_channel(struct ncsi_dev_priv *ndp) 1171 { 1172 struct ncsi_channel *nc, *found, *hot_nc; 1173 struct ncsi_channel_mode *ncm; 1174 unsigned long flags, cflags; 1175 struct ncsi_package *np; 1176 bool with_link; 1177 1178 spin_lock_irqsave(&ndp->lock, flags); 1179 hot_nc = ndp->hot_channel; 1180 spin_unlock_irqrestore(&ndp->lock, flags); 1181 1182 /* By default the search is done once an inactive channel with up 1183 * link is found, unless a preferred channel is set. 1184 * If multi_package or multi_channel are configured all channels in the 1185 * whitelist are added to the channel queue. 1186 */ 1187 found = NULL; 1188 with_link = false; 1189 NCSI_FOR_EACH_PACKAGE(ndp, np) { 1190 if (!(ndp->package_whitelist & (0x1 << np->id))) 1191 continue; 1192 NCSI_FOR_EACH_CHANNEL(np, nc) { 1193 if (!(np->channel_whitelist & (0x1 << nc->id))) 1194 continue; 1195 1196 spin_lock_irqsave(&nc->lock, cflags); 1197 1198 if (!list_empty(&nc->link) || 1199 nc->state != NCSI_CHANNEL_INACTIVE) { 1200 spin_unlock_irqrestore(&nc->lock, cflags); 1201 continue; 1202 } 1203 1204 if (!found) 1205 found = nc; 1206 1207 if (nc == hot_nc) 1208 found = nc; 1209 1210 ncm = &nc->modes[NCSI_MODE_LINK]; 1211 if (ncm->data[2] & 0x1) { 1212 found = nc; 1213 with_link = true; 1214 } 1215 1216 /* If multi_channel is enabled configure all valid 1217 * channels whether or not they currently have link 1218 * so they will have AENs enabled. 1219 */ 1220 if (with_link || np->multi_channel) { 1221 spin_lock_irqsave(&ndp->lock, flags); 1222 list_add_tail_rcu(&nc->link, 1223 &ndp->channel_queue); 1224 spin_unlock_irqrestore(&ndp->lock, flags); 1225 1226 netdev_dbg(ndp->ndev.dev, 1227 "NCSI: Channel %u added to queue (link %s)\n", 1228 nc->id, 1229 ncm->data[2] & 0x1 ? "up" : "down"); 1230 } 1231 1232 spin_unlock_irqrestore(&nc->lock, cflags); 1233 1234 if (with_link && !np->multi_channel) 1235 break; 1236 } 1237 if (with_link && !ndp->multi_package) 1238 break; 1239 } 1240 1241 if (list_empty(&ndp->channel_queue) && found) { 1242 netdev_info(ndp->ndev.dev, 1243 "NCSI: No channel with link found, configuring channel %u\n", 1244 found->id); 1245 spin_lock_irqsave(&ndp->lock, flags); 1246 list_add_tail_rcu(&found->link, &ndp->channel_queue); 1247 spin_unlock_irqrestore(&ndp->lock, flags); 1248 } else if (!found) { 1249 netdev_warn(ndp->ndev.dev, 1250 "NCSI: No channel found to configure!\n"); 1251 ncsi_report_link(ndp, true); 1252 return -ENODEV; 1253 } 1254 1255 return ncsi_process_next_channel(ndp); 1256 } 1257 1258 static bool ncsi_check_hwa(struct ncsi_dev_priv *ndp) 1259 { 1260 struct ncsi_package *np; 1261 struct ncsi_channel *nc; 1262 unsigned int cap; 1263 bool has_channel = false; 1264 1265 /* The hardware arbitration is disabled if any one channel 1266 * doesn't support explicitly. 1267 */ 1268 NCSI_FOR_EACH_PACKAGE(ndp, np) { 1269 NCSI_FOR_EACH_CHANNEL(np, nc) { 1270 has_channel = true; 1271 1272 cap = nc->caps[NCSI_CAP_GENERIC].cap; 1273 if (!(cap & NCSI_CAP_GENERIC_HWA) || 1274 (cap & NCSI_CAP_GENERIC_HWA_MASK) != 1275 NCSI_CAP_GENERIC_HWA_SUPPORT) { 1276 ndp->flags &= ~NCSI_DEV_HWA; 1277 return false; 1278 } 1279 } 1280 } 1281 1282 if (has_channel) { 1283 ndp->flags |= NCSI_DEV_HWA; 1284 return true; 1285 } 1286 1287 ndp->flags &= ~NCSI_DEV_HWA; 1288 return false; 1289 } 1290 1291 static void ncsi_probe_channel(struct ncsi_dev_priv *ndp) 1292 { 1293 struct ncsi_dev *nd = &ndp->ndev; 1294 struct ncsi_package *np; 1295 struct ncsi_channel *nc; 1296 struct ncsi_cmd_arg nca; 1297 unsigned char index; 1298 int ret; 1299 1300 nca.ndp = ndp; 1301 nca.req_flags = NCSI_REQ_FLAG_EVENT_DRIVEN; 1302 switch (nd->state) { 1303 case ncsi_dev_state_probe: 1304 nd->state = ncsi_dev_state_probe_deselect; 1305 fallthrough; 1306 case ncsi_dev_state_probe_deselect: 1307 ndp->pending_req_num = 8; 1308 1309 /* Deselect all possible packages */ 1310 nca.type = NCSI_PKT_CMD_DP; 1311 nca.channel = NCSI_RESERVED_CHANNEL; 1312 for (index = 0; index < 8; index++) { 1313 nca.package = index; 1314 ret = ncsi_xmit_cmd(&nca); 1315 if (ret) 1316 goto error; 1317 } 1318 1319 nd->state = ncsi_dev_state_probe_package; 1320 break; 1321 case ncsi_dev_state_probe_package: 1322 ndp->pending_req_num = 1; 1323 1324 nca.type = NCSI_PKT_CMD_SP; 1325 nca.bytes[0] = 1; 1326 nca.package = ndp->package_probe_id; 1327 nca.channel = NCSI_RESERVED_CHANNEL; 1328 ret = ncsi_xmit_cmd(&nca); 1329 if (ret) 1330 goto error; 1331 nd->state = ncsi_dev_state_probe_channel; 1332 break; 1333 case ncsi_dev_state_probe_channel: 1334 ndp->active_package = ncsi_find_package(ndp, 1335 ndp->package_probe_id); 1336 if (!ndp->active_package) { 1337 /* No response */ 1338 nd->state = ncsi_dev_state_probe_dp; 1339 schedule_work(&ndp->work); 1340 break; 1341 } 1342 nd->state = ncsi_dev_state_probe_cis; 1343 if (IS_ENABLED(CONFIG_NCSI_OEM_CMD_GET_MAC) && 1344 ndp->mlx_multi_host) 1345 nd->state = ncsi_dev_state_probe_mlx_gma; 1346 1347 schedule_work(&ndp->work); 1348 break; 1349 #if IS_ENABLED(CONFIG_NCSI_OEM_CMD_GET_MAC) 1350 case ncsi_dev_state_probe_mlx_gma: 1351 ndp->pending_req_num = 1; 1352 1353 nca.type = NCSI_PKT_CMD_OEM; 1354 nca.package = ndp->active_package->id; 1355 nca.channel = 0; 1356 ret = ncsi_oem_gma_handler_mlx(&nca); 1357 if (ret) 1358 goto error; 1359 1360 nd->state = ncsi_dev_state_probe_mlx_smaf; 1361 break; 1362 case ncsi_dev_state_probe_mlx_smaf: 1363 ndp->pending_req_num = 1; 1364 1365 nca.type = NCSI_PKT_CMD_OEM; 1366 nca.package = ndp->active_package->id; 1367 nca.channel = 0; 1368 ret = ncsi_oem_smaf_mlx(&nca); 1369 if (ret) 1370 goto error; 1371 1372 nd->state = ncsi_dev_state_probe_cis; 1373 break; 1374 #endif /* CONFIG_NCSI_OEM_CMD_GET_MAC */ 1375 case ncsi_dev_state_probe_cis: 1376 ndp->pending_req_num = NCSI_RESERVED_CHANNEL; 1377 1378 /* Clear initial state */ 1379 nca.type = NCSI_PKT_CMD_CIS; 1380 nca.package = ndp->active_package->id; 1381 for (index = 0; index < NCSI_RESERVED_CHANNEL; index++) { 1382 nca.channel = index; 1383 ret = ncsi_xmit_cmd(&nca); 1384 if (ret) 1385 goto error; 1386 } 1387 1388 nd->state = ncsi_dev_state_probe_gvi; 1389 break; 1390 case ncsi_dev_state_probe_gvi: 1391 case ncsi_dev_state_probe_gc: 1392 case ncsi_dev_state_probe_gls: 1393 np = ndp->active_package; 1394 ndp->pending_req_num = np->channel_num; 1395 1396 /* Retrieve version, capability or link status */ 1397 if (nd->state == ncsi_dev_state_probe_gvi) 1398 nca.type = NCSI_PKT_CMD_GVI; 1399 else if (nd->state == ncsi_dev_state_probe_gc) 1400 nca.type = NCSI_PKT_CMD_GC; 1401 else 1402 nca.type = NCSI_PKT_CMD_GLS; 1403 1404 nca.package = np->id; 1405 NCSI_FOR_EACH_CHANNEL(np, nc) { 1406 nca.channel = nc->id; 1407 ret = ncsi_xmit_cmd(&nca); 1408 if (ret) 1409 goto error; 1410 } 1411 1412 if (nd->state == ncsi_dev_state_probe_gvi) 1413 nd->state = ncsi_dev_state_probe_gc; 1414 else if (nd->state == ncsi_dev_state_probe_gc) 1415 nd->state = ncsi_dev_state_probe_gls; 1416 else 1417 nd->state = ncsi_dev_state_probe_dp; 1418 break; 1419 case ncsi_dev_state_probe_dp: 1420 ndp->pending_req_num = 1; 1421 1422 /* Deselect the current package */ 1423 nca.type = NCSI_PKT_CMD_DP; 1424 nca.package = ndp->package_probe_id; 1425 nca.channel = NCSI_RESERVED_CHANNEL; 1426 ret = ncsi_xmit_cmd(&nca); 1427 if (ret) 1428 goto error; 1429 1430 /* Probe next package */ 1431 ndp->package_probe_id++; 1432 if (ndp->package_probe_id >= 8) { 1433 /* Probe finished */ 1434 ndp->flags |= NCSI_DEV_PROBED; 1435 break; 1436 } 1437 nd->state = ncsi_dev_state_probe_package; 1438 ndp->active_package = NULL; 1439 break; 1440 default: 1441 netdev_warn(nd->dev, "Wrong NCSI state 0x%0x in enumeration\n", 1442 nd->state); 1443 } 1444 1445 if (ndp->flags & NCSI_DEV_PROBED) { 1446 /* Check if all packages have HWA support */ 1447 ncsi_check_hwa(ndp); 1448 ncsi_choose_active_channel(ndp); 1449 } 1450 1451 return; 1452 error: 1453 netdev_err(ndp->ndev.dev, 1454 "NCSI: Failed to transmit cmd 0x%x during probe\n", 1455 nca.type); 1456 ncsi_report_link(ndp, true); 1457 } 1458 1459 static void ncsi_dev_work(struct work_struct *work) 1460 { 1461 struct ncsi_dev_priv *ndp = container_of(work, 1462 struct ncsi_dev_priv, work); 1463 struct ncsi_dev *nd = &ndp->ndev; 1464 1465 switch (nd->state & ncsi_dev_state_major) { 1466 case ncsi_dev_state_probe: 1467 ncsi_probe_channel(ndp); 1468 break; 1469 case ncsi_dev_state_suspend: 1470 ncsi_suspend_channel(ndp); 1471 break; 1472 case ncsi_dev_state_config: 1473 ncsi_configure_channel(ndp); 1474 break; 1475 default: 1476 netdev_warn(nd->dev, "Wrong NCSI state 0x%x in workqueue\n", 1477 nd->state); 1478 } 1479 } 1480 1481 int ncsi_process_next_channel(struct ncsi_dev_priv *ndp) 1482 { 1483 struct ncsi_channel *nc; 1484 int old_state; 1485 unsigned long flags; 1486 1487 spin_lock_irqsave(&ndp->lock, flags); 1488 nc = list_first_or_null_rcu(&ndp->channel_queue, 1489 struct ncsi_channel, link); 1490 if (!nc) { 1491 spin_unlock_irqrestore(&ndp->lock, flags); 1492 goto out; 1493 } 1494 1495 list_del_init(&nc->link); 1496 spin_unlock_irqrestore(&ndp->lock, flags); 1497 1498 spin_lock_irqsave(&nc->lock, flags); 1499 old_state = nc->state; 1500 nc->state = NCSI_CHANNEL_INVISIBLE; 1501 spin_unlock_irqrestore(&nc->lock, flags); 1502 1503 ndp->active_channel = nc; 1504 ndp->active_package = nc->package; 1505 1506 switch (old_state) { 1507 case NCSI_CHANNEL_INACTIVE: 1508 ndp->ndev.state = ncsi_dev_state_config; 1509 netdev_dbg(ndp->ndev.dev, "NCSI: configuring channel %u\n", 1510 nc->id); 1511 ncsi_configure_channel(ndp); 1512 break; 1513 case NCSI_CHANNEL_ACTIVE: 1514 ndp->ndev.state = ncsi_dev_state_suspend; 1515 netdev_dbg(ndp->ndev.dev, "NCSI: suspending channel %u\n", 1516 nc->id); 1517 ncsi_suspend_channel(ndp); 1518 break; 1519 default: 1520 netdev_err(ndp->ndev.dev, "Invalid state 0x%x on %d:%d\n", 1521 old_state, nc->package->id, nc->id); 1522 ncsi_report_link(ndp, false); 1523 return -EINVAL; 1524 } 1525 1526 return 0; 1527 1528 out: 1529 ndp->active_channel = NULL; 1530 ndp->active_package = NULL; 1531 if (ndp->flags & NCSI_DEV_RESHUFFLE) { 1532 ndp->flags &= ~NCSI_DEV_RESHUFFLE; 1533 return ncsi_choose_active_channel(ndp); 1534 } 1535 1536 ncsi_report_link(ndp, false); 1537 return -ENODEV; 1538 } 1539 1540 static int ncsi_kick_channels(struct ncsi_dev_priv *ndp) 1541 { 1542 struct ncsi_dev *nd = &ndp->ndev; 1543 struct ncsi_channel *nc; 1544 struct ncsi_package *np; 1545 unsigned long flags; 1546 unsigned int n = 0; 1547 1548 NCSI_FOR_EACH_PACKAGE(ndp, np) { 1549 NCSI_FOR_EACH_CHANNEL(np, nc) { 1550 spin_lock_irqsave(&nc->lock, flags); 1551 1552 /* Channels may be busy, mark dirty instead of 1553 * kicking if; 1554 * a) not ACTIVE (configured) 1555 * b) in the channel_queue (to be configured) 1556 * c) it's ndev is in the config state 1557 */ 1558 if (nc->state != NCSI_CHANNEL_ACTIVE) { 1559 if ((ndp->ndev.state & 0xff00) == 1560 ncsi_dev_state_config || 1561 !list_empty(&nc->link)) { 1562 netdev_dbg(nd->dev, 1563 "NCSI: channel %p marked dirty\n", 1564 nc); 1565 nc->reconfigure_needed = true; 1566 } 1567 spin_unlock_irqrestore(&nc->lock, flags); 1568 continue; 1569 } 1570 1571 spin_unlock_irqrestore(&nc->lock, flags); 1572 1573 ncsi_stop_channel_monitor(nc); 1574 spin_lock_irqsave(&nc->lock, flags); 1575 nc->state = NCSI_CHANNEL_INACTIVE; 1576 spin_unlock_irqrestore(&nc->lock, flags); 1577 1578 spin_lock_irqsave(&ndp->lock, flags); 1579 list_add_tail_rcu(&nc->link, &ndp->channel_queue); 1580 spin_unlock_irqrestore(&ndp->lock, flags); 1581 1582 netdev_dbg(nd->dev, "NCSI: kicked channel %p\n", nc); 1583 n++; 1584 } 1585 } 1586 1587 return n; 1588 } 1589 1590 int ncsi_vlan_rx_add_vid(struct net_device *dev, __be16 proto, u16 vid) 1591 { 1592 struct ncsi_dev_priv *ndp; 1593 unsigned int n_vids = 0; 1594 struct vlan_vid *vlan; 1595 struct ncsi_dev *nd; 1596 bool found = false; 1597 1598 if (vid == 0) 1599 return 0; 1600 1601 nd = ncsi_find_dev(dev); 1602 if (!nd) { 1603 netdev_warn(dev, "NCSI: No net_device?\n"); 1604 return 0; 1605 } 1606 1607 ndp = TO_NCSI_DEV_PRIV(nd); 1608 1609 /* Add the VLAN id to our internal list */ 1610 list_for_each_entry_rcu(vlan, &ndp->vlan_vids, list) { 1611 n_vids++; 1612 if (vlan->vid == vid) { 1613 netdev_dbg(dev, "NCSI: vid %u already registered\n", 1614 vid); 1615 return 0; 1616 } 1617 } 1618 if (n_vids >= NCSI_MAX_VLAN_VIDS) { 1619 netdev_warn(dev, 1620 "tried to add vlan id %u but NCSI max already registered (%u)\n", 1621 vid, NCSI_MAX_VLAN_VIDS); 1622 return -ENOSPC; 1623 } 1624 1625 vlan = kzalloc(sizeof(*vlan), GFP_KERNEL); 1626 if (!vlan) 1627 return -ENOMEM; 1628 1629 vlan->proto = proto; 1630 vlan->vid = vid; 1631 list_add_rcu(&vlan->list, &ndp->vlan_vids); 1632 1633 netdev_dbg(dev, "NCSI: Added new vid %u\n", vid); 1634 1635 found = ncsi_kick_channels(ndp) != 0; 1636 1637 return found ? ncsi_process_next_channel(ndp) : 0; 1638 } 1639 EXPORT_SYMBOL_GPL(ncsi_vlan_rx_add_vid); 1640 1641 int ncsi_vlan_rx_kill_vid(struct net_device *dev, __be16 proto, u16 vid) 1642 { 1643 struct vlan_vid *vlan, *tmp; 1644 struct ncsi_dev_priv *ndp; 1645 struct ncsi_dev *nd; 1646 bool found = false; 1647 1648 if (vid == 0) 1649 return 0; 1650 1651 nd = ncsi_find_dev(dev); 1652 if (!nd) { 1653 netdev_warn(dev, "NCSI: no net_device?\n"); 1654 return 0; 1655 } 1656 1657 ndp = TO_NCSI_DEV_PRIV(nd); 1658 1659 /* Remove the VLAN id from our internal list */ 1660 list_for_each_entry_safe(vlan, tmp, &ndp->vlan_vids, list) 1661 if (vlan->vid == vid) { 1662 netdev_dbg(dev, "NCSI: vid %u found, removing\n", vid); 1663 list_del_rcu(&vlan->list); 1664 found = true; 1665 kfree(vlan); 1666 } 1667 1668 if (!found) { 1669 netdev_err(dev, "NCSI: vid %u wasn't registered!\n", vid); 1670 return -EINVAL; 1671 } 1672 1673 found = ncsi_kick_channels(ndp) != 0; 1674 1675 return found ? ncsi_process_next_channel(ndp) : 0; 1676 } 1677 EXPORT_SYMBOL_GPL(ncsi_vlan_rx_kill_vid); 1678 1679 struct ncsi_dev *ncsi_register_dev(struct net_device *dev, 1680 void (*handler)(struct ncsi_dev *ndev)) 1681 { 1682 struct ncsi_dev_priv *ndp; 1683 struct ncsi_dev *nd; 1684 struct platform_device *pdev; 1685 struct device_node *np; 1686 unsigned long flags; 1687 int i; 1688 1689 /* Check if the device has been registered or not */ 1690 nd = ncsi_find_dev(dev); 1691 if (nd) 1692 return nd; 1693 1694 /* Create NCSI device */ 1695 ndp = kzalloc(sizeof(*ndp), GFP_ATOMIC); 1696 if (!ndp) 1697 return NULL; 1698 1699 nd = &ndp->ndev; 1700 nd->state = ncsi_dev_state_registered; 1701 nd->dev = dev; 1702 nd->handler = handler; 1703 ndp->pending_req_num = 0; 1704 INIT_LIST_HEAD(&ndp->channel_queue); 1705 INIT_LIST_HEAD(&ndp->vlan_vids); 1706 INIT_WORK(&ndp->work, ncsi_dev_work); 1707 ndp->package_whitelist = UINT_MAX; 1708 1709 /* Initialize private NCSI device */ 1710 spin_lock_init(&ndp->lock); 1711 INIT_LIST_HEAD(&ndp->packages); 1712 ndp->request_id = NCSI_REQ_START_IDX; 1713 for (i = 0; i < ARRAY_SIZE(ndp->requests); i++) { 1714 ndp->requests[i].id = i; 1715 ndp->requests[i].ndp = ndp; 1716 timer_setup(&ndp->requests[i].timer, ncsi_request_timeout, 0); 1717 } 1718 1719 spin_lock_irqsave(&ncsi_dev_lock, flags); 1720 list_add_tail_rcu(&ndp->node, &ncsi_dev_list); 1721 spin_unlock_irqrestore(&ncsi_dev_lock, flags); 1722 1723 /* Register NCSI packet Rx handler */ 1724 ndp->ptype.type = cpu_to_be16(ETH_P_NCSI); 1725 ndp->ptype.func = ncsi_rcv_rsp; 1726 ndp->ptype.dev = dev; 1727 dev_add_pack(&ndp->ptype); 1728 1729 pdev = to_platform_device(dev->dev.parent); 1730 if (pdev) { 1731 np = pdev->dev.of_node; 1732 if (np && of_get_property(np, "mlx,multi-host", NULL)) 1733 ndp->mlx_multi_host = true; 1734 } 1735 1736 return nd; 1737 } 1738 EXPORT_SYMBOL_GPL(ncsi_register_dev); 1739 1740 int ncsi_start_dev(struct ncsi_dev *nd) 1741 { 1742 struct ncsi_dev_priv *ndp = TO_NCSI_DEV_PRIV(nd); 1743 1744 if (nd->state != ncsi_dev_state_registered && 1745 nd->state != ncsi_dev_state_functional) 1746 return -ENOTTY; 1747 1748 if (!(ndp->flags & NCSI_DEV_PROBED)) { 1749 ndp->package_probe_id = 0; 1750 nd->state = ncsi_dev_state_probe; 1751 schedule_work(&ndp->work); 1752 return 0; 1753 } 1754 1755 return ncsi_reset_dev(nd); 1756 } 1757 EXPORT_SYMBOL_GPL(ncsi_start_dev); 1758 1759 void ncsi_stop_dev(struct ncsi_dev *nd) 1760 { 1761 struct ncsi_dev_priv *ndp = TO_NCSI_DEV_PRIV(nd); 1762 struct ncsi_package *np; 1763 struct ncsi_channel *nc; 1764 bool chained; 1765 int old_state; 1766 unsigned long flags; 1767 1768 /* Stop the channel monitor on any active channels. Don't reset the 1769 * channel state so we know which were active when ncsi_start_dev() 1770 * is next called. 1771 */ 1772 NCSI_FOR_EACH_PACKAGE(ndp, np) { 1773 NCSI_FOR_EACH_CHANNEL(np, nc) { 1774 ncsi_stop_channel_monitor(nc); 1775 1776 spin_lock_irqsave(&nc->lock, flags); 1777 chained = !list_empty(&nc->link); 1778 old_state = nc->state; 1779 spin_unlock_irqrestore(&nc->lock, flags); 1780 1781 WARN_ON_ONCE(chained || 1782 old_state == NCSI_CHANNEL_INVISIBLE); 1783 } 1784 } 1785 1786 netdev_dbg(ndp->ndev.dev, "NCSI: Stopping device\n"); 1787 ncsi_report_link(ndp, true); 1788 } 1789 EXPORT_SYMBOL_GPL(ncsi_stop_dev); 1790 1791 int ncsi_reset_dev(struct ncsi_dev *nd) 1792 { 1793 struct ncsi_dev_priv *ndp = TO_NCSI_DEV_PRIV(nd); 1794 struct ncsi_channel *nc, *active, *tmp; 1795 struct ncsi_package *np; 1796 unsigned long flags; 1797 1798 spin_lock_irqsave(&ndp->lock, flags); 1799 1800 if (!(ndp->flags & NCSI_DEV_RESET)) { 1801 /* Haven't been called yet, check states */ 1802 switch (nd->state & ncsi_dev_state_major) { 1803 case ncsi_dev_state_registered: 1804 case ncsi_dev_state_probe: 1805 /* Not even probed yet - do nothing */ 1806 spin_unlock_irqrestore(&ndp->lock, flags); 1807 return 0; 1808 case ncsi_dev_state_suspend: 1809 case ncsi_dev_state_config: 1810 /* Wait for the channel to finish its suspend/config 1811 * operation; once it finishes it will check for 1812 * NCSI_DEV_RESET and reset the state. 1813 */ 1814 ndp->flags |= NCSI_DEV_RESET; 1815 spin_unlock_irqrestore(&ndp->lock, flags); 1816 return 0; 1817 } 1818 } else { 1819 switch (nd->state) { 1820 case ncsi_dev_state_suspend_done: 1821 case ncsi_dev_state_config_done: 1822 case ncsi_dev_state_functional: 1823 /* Ok */ 1824 break; 1825 default: 1826 /* Current reset operation happening */ 1827 spin_unlock_irqrestore(&ndp->lock, flags); 1828 return 0; 1829 } 1830 } 1831 1832 if (!list_empty(&ndp->channel_queue)) { 1833 /* Clear any channel queue we may have interrupted */ 1834 list_for_each_entry_safe(nc, tmp, &ndp->channel_queue, link) 1835 list_del_init(&nc->link); 1836 } 1837 spin_unlock_irqrestore(&ndp->lock, flags); 1838 1839 active = NULL; 1840 NCSI_FOR_EACH_PACKAGE(ndp, np) { 1841 NCSI_FOR_EACH_CHANNEL(np, nc) { 1842 spin_lock_irqsave(&nc->lock, flags); 1843 1844 if (nc->state == NCSI_CHANNEL_ACTIVE) { 1845 active = nc; 1846 nc->state = NCSI_CHANNEL_INVISIBLE; 1847 spin_unlock_irqrestore(&nc->lock, flags); 1848 ncsi_stop_channel_monitor(nc); 1849 break; 1850 } 1851 1852 spin_unlock_irqrestore(&nc->lock, flags); 1853 } 1854 if (active) 1855 break; 1856 } 1857 1858 if (!active) { 1859 /* Done */ 1860 spin_lock_irqsave(&ndp->lock, flags); 1861 ndp->flags &= ~NCSI_DEV_RESET; 1862 spin_unlock_irqrestore(&ndp->lock, flags); 1863 return ncsi_choose_active_channel(ndp); 1864 } 1865 1866 spin_lock_irqsave(&ndp->lock, flags); 1867 ndp->flags |= NCSI_DEV_RESET; 1868 ndp->active_channel = active; 1869 ndp->active_package = active->package; 1870 spin_unlock_irqrestore(&ndp->lock, flags); 1871 1872 nd->state = ncsi_dev_state_suspend; 1873 schedule_work(&ndp->work); 1874 return 0; 1875 } 1876 1877 void ncsi_unregister_dev(struct ncsi_dev *nd) 1878 { 1879 struct ncsi_dev_priv *ndp = TO_NCSI_DEV_PRIV(nd); 1880 struct ncsi_package *np, *tmp; 1881 unsigned long flags; 1882 1883 dev_remove_pack(&ndp->ptype); 1884 1885 list_for_each_entry_safe(np, tmp, &ndp->packages, node) 1886 ncsi_remove_package(np); 1887 1888 spin_lock_irqsave(&ncsi_dev_lock, flags); 1889 list_del_rcu(&ndp->node); 1890 spin_unlock_irqrestore(&ncsi_dev_lock, flags); 1891 1892 kfree(ndp); 1893 } 1894 EXPORT_SYMBOL_GPL(ncsi_unregister_dev); 1895