1 // SPDX-License-Identifier: GPL-2.0 OR Linux-OpenIB 2 /* 3 * Copyright 2015-2020 Amazon.com, Inc. or its affiliates. All rights reserved. 4 */ 5 6 #include <linux/ethtool.h> 7 #include <linux/pci.h> 8 9 #include "ena_netdev.h" 10 11 struct ena_stats { 12 char name[ETH_GSTRING_LEN]; 13 int stat_offset; 14 }; 15 16 #define ENA_STAT_ENA_COM_ENTRY(stat) { \ 17 .name = #stat, \ 18 .stat_offset = offsetof(struct ena_com_stats_admin, stat) / sizeof(u64) \ 19 } 20 21 #define ENA_STAT_ENTRY(stat, stat_type) { \ 22 .name = #stat, \ 23 .stat_offset = offsetof(struct ena_stats_##stat_type, stat) / sizeof(u64) \ 24 } 25 26 #define ENA_STAT_HW_ENTRY(stat, stat_type) { \ 27 .name = #stat, \ 28 .stat_offset = offsetof(struct ena_admin_##stat_type, stat) / sizeof(u64) \ 29 } 30 31 #define ENA_STAT_RX_ENTRY(stat) \ 32 ENA_STAT_ENTRY(stat, rx) 33 34 #define ENA_STAT_TX_ENTRY(stat) \ 35 ENA_STAT_ENTRY(stat, tx) 36 37 #define ENA_STAT_GLOBAL_ENTRY(stat) \ 38 ENA_STAT_ENTRY(stat, dev) 39 40 #define ENA_STAT_ENI_ENTRY(stat) \ 41 ENA_STAT_HW_ENTRY(stat, eni_stats) 42 43 static const struct ena_stats ena_stats_global_strings[] = { 44 ENA_STAT_GLOBAL_ENTRY(tx_timeout), 45 ENA_STAT_GLOBAL_ENTRY(suspend), 46 ENA_STAT_GLOBAL_ENTRY(resume), 47 ENA_STAT_GLOBAL_ENTRY(wd_expired), 48 ENA_STAT_GLOBAL_ENTRY(interface_up), 49 ENA_STAT_GLOBAL_ENTRY(interface_down), 50 ENA_STAT_GLOBAL_ENTRY(admin_q_pause), 51 }; 52 53 static const struct ena_stats ena_stats_eni_strings[] = { 54 ENA_STAT_ENI_ENTRY(bw_in_allowance_exceeded), 55 ENA_STAT_ENI_ENTRY(bw_out_allowance_exceeded), 56 ENA_STAT_ENI_ENTRY(pps_allowance_exceeded), 57 ENA_STAT_ENI_ENTRY(conntrack_allowance_exceeded), 58 ENA_STAT_ENI_ENTRY(linklocal_allowance_exceeded), 59 }; 60 61 static const struct ena_stats ena_stats_tx_strings[] = { 62 ENA_STAT_TX_ENTRY(cnt), 63 ENA_STAT_TX_ENTRY(bytes), 64 ENA_STAT_TX_ENTRY(queue_stop), 65 ENA_STAT_TX_ENTRY(queue_wakeup), 66 ENA_STAT_TX_ENTRY(dma_mapping_err), 67 ENA_STAT_TX_ENTRY(linearize), 68 ENA_STAT_TX_ENTRY(linearize_failed), 69 ENA_STAT_TX_ENTRY(napi_comp), 70 ENA_STAT_TX_ENTRY(tx_poll), 71 ENA_STAT_TX_ENTRY(doorbells), 72 ENA_STAT_TX_ENTRY(prepare_ctx_err), 73 ENA_STAT_TX_ENTRY(bad_req_id), 74 ENA_STAT_TX_ENTRY(llq_buffer_copy), 75 ENA_STAT_TX_ENTRY(missed_tx), 76 ENA_STAT_TX_ENTRY(unmask_interrupt), 77 }; 78 79 static const struct ena_stats ena_stats_rx_strings[] = { 80 ENA_STAT_RX_ENTRY(cnt), 81 ENA_STAT_RX_ENTRY(bytes), 82 ENA_STAT_RX_ENTRY(rx_copybreak_pkt), 83 ENA_STAT_RX_ENTRY(csum_good), 84 ENA_STAT_RX_ENTRY(refil_partial), 85 ENA_STAT_RX_ENTRY(bad_csum), 86 ENA_STAT_RX_ENTRY(page_alloc_fail), 87 ENA_STAT_RX_ENTRY(skb_alloc_fail), 88 ENA_STAT_RX_ENTRY(dma_mapping_err), 89 ENA_STAT_RX_ENTRY(bad_desc_num), 90 ENA_STAT_RX_ENTRY(bad_req_id), 91 ENA_STAT_RX_ENTRY(empty_rx_ring), 92 ENA_STAT_RX_ENTRY(csum_unchecked), 93 ENA_STAT_RX_ENTRY(xdp_aborted), 94 ENA_STAT_RX_ENTRY(xdp_drop), 95 ENA_STAT_RX_ENTRY(xdp_pass), 96 ENA_STAT_RX_ENTRY(xdp_tx), 97 ENA_STAT_RX_ENTRY(xdp_invalid), 98 ENA_STAT_RX_ENTRY(xdp_redirect), 99 }; 100 101 static const struct ena_stats ena_stats_ena_com_strings[] = { 102 ENA_STAT_ENA_COM_ENTRY(aborted_cmd), 103 ENA_STAT_ENA_COM_ENTRY(submitted_cmd), 104 ENA_STAT_ENA_COM_ENTRY(completed_cmd), 105 ENA_STAT_ENA_COM_ENTRY(out_of_space), 106 ENA_STAT_ENA_COM_ENTRY(no_completion), 107 }; 108 109 #define ENA_STATS_ARRAY_GLOBAL ARRAY_SIZE(ena_stats_global_strings) 110 #define ENA_STATS_ARRAY_TX ARRAY_SIZE(ena_stats_tx_strings) 111 #define ENA_STATS_ARRAY_RX ARRAY_SIZE(ena_stats_rx_strings) 112 #define ENA_STATS_ARRAY_ENA_COM ARRAY_SIZE(ena_stats_ena_com_strings) 113 #define ENA_STATS_ARRAY_ENI(adapter) \ 114 (ARRAY_SIZE(ena_stats_eni_strings) * (adapter)->eni_stats_supported) 115 116 static void ena_safe_update_stat(u64 *src, u64 *dst, 117 struct u64_stats_sync *syncp) 118 { 119 unsigned int start; 120 121 do { 122 start = u64_stats_fetch_begin_irq(syncp); 123 *(dst) = *src; 124 } while (u64_stats_fetch_retry_irq(syncp, start)); 125 } 126 127 static void ena_queue_stats(struct ena_adapter *adapter, u64 **data) 128 { 129 const struct ena_stats *ena_stats; 130 struct ena_ring *ring; 131 132 u64 *ptr; 133 int i, j; 134 135 for (i = 0; i < adapter->num_io_queues + adapter->xdp_num_queues; i++) { 136 /* Tx stats */ 137 ring = &adapter->tx_ring[i]; 138 139 for (j = 0; j < ENA_STATS_ARRAY_TX; j++) { 140 ena_stats = &ena_stats_tx_strings[j]; 141 142 ptr = (u64 *)&ring->tx_stats + ena_stats->stat_offset; 143 144 ena_safe_update_stat(ptr, (*data)++, &ring->syncp); 145 } 146 /* XDP TX queues don't have a RX queue counterpart */ 147 if (!ENA_IS_XDP_INDEX(adapter, i)) { 148 /* Rx stats */ 149 ring = &adapter->rx_ring[i]; 150 151 for (j = 0; j < ENA_STATS_ARRAY_RX; j++) { 152 ena_stats = &ena_stats_rx_strings[j]; 153 154 ptr = (u64 *)&ring->rx_stats + 155 ena_stats->stat_offset; 156 157 ena_safe_update_stat(ptr, (*data)++, &ring->syncp); 158 } 159 } 160 } 161 } 162 163 static void ena_dev_admin_queue_stats(struct ena_adapter *adapter, u64 **data) 164 { 165 const struct ena_stats *ena_stats; 166 u64 *ptr; 167 int i; 168 169 for (i = 0; i < ENA_STATS_ARRAY_ENA_COM; i++) { 170 ena_stats = &ena_stats_ena_com_strings[i]; 171 172 ptr = (u64 *)&adapter->ena_dev->admin_queue.stats + 173 ena_stats->stat_offset; 174 175 *(*data)++ = *ptr; 176 } 177 } 178 179 static void ena_get_stats(struct ena_adapter *adapter, 180 u64 *data, 181 bool eni_stats_needed) 182 { 183 const struct ena_stats *ena_stats; 184 u64 *ptr; 185 int i; 186 187 for (i = 0; i < ENA_STATS_ARRAY_GLOBAL; i++) { 188 ena_stats = &ena_stats_global_strings[i]; 189 190 ptr = (u64 *)&adapter->dev_stats + ena_stats->stat_offset; 191 192 ena_safe_update_stat(ptr, data++, &adapter->syncp); 193 } 194 195 if (eni_stats_needed) { 196 ena_update_hw_stats(adapter); 197 for (i = 0; i < ENA_STATS_ARRAY_ENI(adapter); i++) { 198 ena_stats = &ena_stats_eni_strings[i]; 199 200 ptr = (u64 *)&adapter->eni_stats + 201 ena_stats->stat_offset; 202 203 ena_safe_update_stat(ptr, data++, &adapter->syncp); 204 } 205 } 206 207 ena_queue_stats(adapter, &data); 208 ena_dev_admin_queue_stats(adapter, &data); 209 } 210 211 static void ena_get_ethtool_stats(struct net_device *netdev, 212 struct ethtool_stats *stats, 213 u64 *data) 214 { 215 struct ena_adapter *adapter = netdev_priv(netdev); 216 217 ena_get_stats(adapter, data, adapter->eni_stats_supported); 218 } 219 220 static int ena_get_sw_stats_count(struct ena_adapter *adapter) 221 { 222 return adapter->num_io_queues * (ENA_STATS_ARRAY_TX + ENA_STATS_ARRAY_RX) 223 + adapter->xdp_num_queues * ENA_STATS_ARRAY_TX 224 + ENA_STATS_ARRAY_GLOBAL + ENA_STATS_ARRAY_ENA_COM; 225 } 226 227 static int ena_get_hw_stats_count(struct ena_adapter *adapter) 228 { 229 return ENA_STATS_ARRAY_ENI(adapter); 230 } 231 232 int ena_get_sset_count(struct net_device *netdev, int sset) 233 { 234 struct ena_adapter *adapter = netdev_priv(netdev); 235 236 switch (sset) { 237 case ETH_SS_STATS: 238 return ena_get_sw_stats_count(adapter) + 239 ena_get_hw_stats_count(adapter); 240 } 241 242 return -EOPNOTSUPP; 243 } 244 245 static void ena_queue_strings(struct ena_adapter *adapter, u8 **data) 246 { 247 const struct ena_stats *ena_stats; 248 bool is_xdp; 249 int i, j; 250 251 for (i = 0; i < adapter->num_io_queues + adapter->xdp_num_queues; i++) { 252 is_xdp = ENA_IS_XDP_INDEX(adapter, i); 253 /* Tx stats */ 254 for (j = 0; j < ENA_STATS_ARRAY_TX; j++) { 255 ena_stats = &ena_stats_tx_strings[j]; 256 257 ethtool_sprintf(data, 258 "queue_%u_%s_%s", i, 259 is_xdp ? "xdp_tx" : "tx", 260 ena_stats->name); 261 } 262 263 if (!is_xdp) { 264 /* RX stats, in XDP there isn't a RX queue 265 * counterpart 266 */ 267 for (j = 0; j < ENA_STATS_ARRAY_RX; j++) { 268 ena_stats = &ena_stats_rx_strings[j]; 269 270 ethtool_sprintf(data, 271 "queue_%u_rx_%s", i, 272 ena_stats->name); 273 } 274 } 275 } 276 } 277 278 static void ena_com_dev_strings(u8 **data) 279 { 280 const struct ena_stats *ena_stats; 281 int i; 282 283 for (i = 0; i < ENA_STATS_ARRAY_ENA_COM; i++) { 284 ena_stats = &ena_stats_ena_com_strings[i]; 285 286 ethtool_sprintf(data, 287 "ena_admin_q_%s", ena_stats->name); 288 } 289 } 290 291 static void ena_get_strings(struct ena_adapter *adapter, 292 u8 *data, 293 bool eni_stats_needed) 294 { 295 const struct ena_stats *ena_stats; 296 int i; 297 298 for (i = 0; i < ENA_STATS_ARRAY_GLOBAL; i++) { 299 ena_stats = &ena_stats_global_strings[i]; 300 ethtool_sprintf(&data, ena_stats->name); 301 } 302 303 if (eni_stats_needed) { 304 for (i = 0; i < ENA_STATS_ARRAY_ENI(adapter); i++) { 305 ena_stats = &ena_stats_eni_strings[i]; 306 ethtool_sprintf(&data, ena_stats->name); 307 } 308 } 309 310 ena_queue_strings(adapter, &data); 311 ena_com_dev_strings(&data); 312 } 313 314 static void ena_get_ethtool_strings(struct net_device *netdev, 315 u32 sset, 316 u8 *data) 317 { 318 struct ena_adapter *adapter = netdev_priv(netdev); 319 320 switch (sset) { 321 case ETH_SS_STATS: 322 ena_get_strings(adapter, data, adapter->eni_stats_supported); 323 break; 324 } 325 } 326 327 static int ena_get_link_ksettings(struct net_device *netdev, 328 struct ethtool_link_ksettings *link_ksettings) 329 { 330 struct ena_adapter *adapter = netdev_priv(netdev); 331 struct ena_com_dev *ena_dev = adapter->ena_dev; 332 struct ena_admin_get_feature_link_desc *link; 333 struct ena_admin_get_feat_resp feat_resp; 334 int rc; 335 336 rc = ena_com_get_link_params(ena_dev, &feat_resp); 337 if (rc) 338 return rc; 339 340 link = &feat_resp.u.link; 341 link_ksettings->base.speed = link->speed; 342 343 if (link->flags & ENA_ADMIN_GET_FEATURE_LINK_DESC_AUTONEG_MASK) { 344 ethtool_link_ksettings_add_link_mode(link_ksettings, 345 supported, Autoneg); 346 ethtool_link_ksettings_add_link_mode(link_ksettings, 347 supported, Autoneg); 348 } 349 350 link_ksettings->base.autoneg = 351 (link->flags & ENA_ADMIN_GET_FEATURE_LINK_DESC_AUTONEG_MASK) ? 352 AUTONEG_ENABLE : AUTONEG_DISABLE; 353 354 link_ksettings->base.duplex = DUPLEX_FULL; 355 356 return 0; 357 } 358 359 static int ena_get_coalesce(struct net_device *net_dev, 360 struct ethtool_coalesce *coalesce, 361 struct kernel_ethtool_coalesce *kernel_coal, 362 struct netlink_ext_ack *extack) 363 { 364 struct ena_adapter *adapter = netdev_priv(net_dev); 365 struct ena_com_dev *ena_dev = adapter->ena_dev; 366 367 if (!ena_com_interrupt_moderation_supported(ena_dev)) 368 return -EOPNOTSUPP; 369 370 coalesce->tx_coalesce_usecs = 371 ena_com_get_nonadaptive_moderation_interval_tx(ena_dev) * 372 ena_dev->intr_delay_resolution; 373 374 coalesce->rx_coalesce_usecs = 375 ena_com_get_nonadaptive_moderation_interval_rx(ena_dev) 376 * ena_dev->intr_delay_resolution; 377 378 coalesce->use_adaptive_rx_coalesce = 379 ena_com_get_adaptive_moderation_enabled(ena_dev); 380 381 return 0; 382 } 383 384 static void ena_update_tx_rings_nonadaptive_intr_moderation(struct ena_adapter *adapter) 385 { 386 unsigned int val; 387 int i; 388 389 val = ena_com_get_nonadaptive_moderation_interval_tx(adapter->ena_dev); 390 391 for (i = 0; i < adapter->num_io_queues; i++) 392 adapter->tx_ring[i].smoothed_interval = val; 393 } 394 395 static void ena_update_rx_rings_nonadaptive_intr_moderation(struct ena_adapter *adapter) 396 { 397 unsigned int val; 398 int i; 399 400 val = ena_com_get_nonadaptive_moderation_interval_rx(adapter->ena_dev); 401 402 for (i = 0; i < adapter->num_io_queues; i++) 403 adapter->rx_ring[i].smoothed_interval = val; 404 } 405 406 static int ena_set_coalesce(struct net_device *net_dev, 407 struct ethtool_coalesce *coalesce, 408 struct kernel_ethtool_coalesce *kernel_coal, 409 struct netlink_ext_ack *extack) 410 { 411 struct ena_adapter *adapter = netdev_priv(net_dev); 412 struct ena_com_dev *ena_dev = adapter->ena_dev; 413 int rc; 414 415 if (!ena_com_interrupt_moderation_supported(ena_dev)) 416 return -EOPNOTSUPP; 417 418 rc = ena_com_update_nonadaptive_moderation_interval_tx(ena_dev, 419 coalesce->tx_coalesce_usecs); 420 if (rc) 421 return rc; 422 423 ena_update_tx_rings_nonadaptive_intr_moderation(adapter); 424 425 rc = ena_com_update_nonadaptive_moderation_interval_rx(ena_dev, 426 coalesce->rx_coalesce_usecs); 427 if (rc) 428 return rc; 429 430 ena_update_rx_rings_nonadaptive_intr_moderation(adapter); 431 432 if (coalesce->use_adaptive_rx_coalesce && 433 !ena_com_get_adaptive_moderation_enabled(ena_dev)) 434 ena_com_enable_adaptive_moderation(ena_dev); 435 436 if (!coalesce->use_adaptive_rx_coalesce && 437 ena_com_get_adaptive_moderation_enabled(ena_dev)) 438 ena_com_disable_adaptive_moderation(ena_dev); 439 440 return 0; 441 } 442 443 static u32 ena_get_msglevel(struct net_device *netdev) 444 { 445 struct ena_adapter *adapter = netdev_priv(netdev); 446 447 return adapter->msg_enable; 448 } 449 450 static void ena_set_msglevel(struct net_device *netdev, u32 value) 451 { 452 struct ena_adapter *adapter = netdev_priv(netdev); 453 454 adapter->msg_enable = value; 455 } 456 457 static void ena_get_drvinfo(struct net_device *dev, 458 struct ethtool_drvinfo *info) 459 { 460 struct ena_adapter *adapter = netdev_priv(dev); 461 462 strlcpy(info->driver, DRV_MODULE_NAME, sizeof(info->driver)); 463 strlcpy(info->bus_info, pci_name(adapter->pdev), 464 sizeof(info->bus_info)); 465 } 466 467 static void ena_get_ringparam(struct net_device *netdev, 468 struct ethtool_ringparam *ring) 469 { 470 struct ena_adapter *adapter = netdev_priv(netdev); 471 472 ring->tx_max_pending = adapter->max_tx_ring_size; 473 ring->rx_max_pending = adapter->max_rx_ring_size; 474 ring->tx_pending = adapter->tx_ring[0].ring_size; 475 ring->rx_pending = adapter->rx_ring[0].ring_size; 476 } 477 478 static int ena_set_ringparam(struct net_device *netdev, 479 struct ethtool_ringparam *ring) 480 { 481 struct ena_adapter *adapter = netdev_priv(netdev); 482 u32 new_tx_size, new_rx_size; 483 484 new_tx_size = ring->tx_pending < ENA_MIN_RING_SIZE ? 485 ENA_MIN_RING_SIZE : ring->tx_pending; 486 new_tx_size = rounddown_pow_of_two(new_tx_size); 487 488 new_rx_size = ring->rx_pending < ENA_MIN_RING_SIZE ? 489 ENA_MIN_RING_SIZE : ring->rx_pending; 490 new_rx_size = rounddown_pow_of_two(new_rx_size); 491 492 if (new_tx_size == adapter->requested_tx_ring_size && 493 new_rx_size == adapter->requested_rx_ring_size) 494 return 0; 495 496 return ena_update_queue_sizes(adapter, new_tx_size, new_rx_size); 497 } 498 499 static u32 ena_flow_hash_to_flow_type(u16 hash_fields) 500 { 501 u32 data = 0; 502 503 if (hash_fields & ENA_ADMIN_RSS_L2_DA) 504 data |= RXH_L2DA; 505 506 if (hash_fields & ENA_ADMIN_RSS_L3_DA) 507 data |= RXH_IP_DST; 508 509 if (hash_fields & ENA_ADMIN_RSS_L3_SA) 510 data |= RXH_IP_SRC; 511 512 if (hash_fields & ENA_ADMIN_RSS_L4_DP) 513 data |= RXH_L4_B_2_3; 514 515 if (hash_fields & ENA_ADMIN_RSS_L4_SP) 516 data |= RXH_L4_B_0_1; 517 518 return data; 519 } 520 521 static u16 ena_flow_data_to_flow_hash(u32 hash_fields) 522 { 523 u16 data = 0; 524 525 if (hash_fields & RXH_L2DA) 526 data |= ENA_ADMIN_RSS_L2_DA; 527 528 if (hash_fields & RXH_IP_DST) 529 data |= ENA_ADMIN_RSS_L3_DA; 530 531 if (hash_fields & RXH_IP_SRC) 532 data |= ENA_ADMIN_RSS_L3_SA; 533 534 if (hash_fields & RXH_L4_B_2_3) 535 data |= ENA_ADMIN_RSS_L4_DP; 536 537 if (hash_fields & RXH_L4_B_0_1) 538 data |= ENA_ADMIN_RSS_L4_SP; 539 540 return data; 541 } 542 543 static int ena_get_rss_hash(struct ena_com_dev *ena_dev, 544 struct ethtool_rxnfc *cmd) 545 { 546 enum ena_admin_flow_hash_proto proto; 547 u16 hash_fields; 548 int rc; 549 550 cmd->data = 0; 551 552 switch (cmd->flow_type) { 553 case TCP_V4_FLOW: 554 proto = ENA_ADMIN_RSS_TCP4; 555 break; 556 case UDP_V4_FLOW: 557 proto = ENA_ADMIN_RSS_UDP4; 558 break; 559 case TCP_V6_FLOW: 560 proto = ENA_ADMIN_RSS_TCP6; 561 break; 562 case UDP_V6_FLOW: 563 proto = ENA_ADMIN_RSS_UDP6; 564 break; 565 case IPV4_FLOW: 566 proto = ENA_ADMIN_RSS_IP4; 567 break; 568 case IPV6_FLOW: 569 proto = ENA_ADMIN_RSS_IP6; 570 break; 571 case ETHER_FLOW: 572 proto = ENA_ADMIN_RSS_NOT_IP; 573 break; 574 case AH_V4_FLOW: 575 case ESP_V4_FLOW: 576 case AH_V6_FLOW: 577 case ESP_V6_FLOW: 578 case SCTP_V4_FLOW: 579 case AH_ESP_V4_FLOW: 580 return -EOPNOTSUPP; 581 default: 582 return -EINVAL; 583 } 584 585 rc = ena_com_get_hash_ctrl(ena_dev, proto, &hash_fields); 586 if (rc) 587 return rc; 588 589 cmd->data = ena_flow_hash_to_flow_type(hash_fields); 590 591 return 0; 592 } 593 594 static int ena_set_rss_hash(struct ena_com_dev *ena_dev, 595 struct ethtool_rxnfc *cmd) 596 { 597 enum ena_admin_flow_hash_proto proto; 598 u16 hash_fields; 599 600 switch (cmd->flow_type) { 601 case TCP_V4_FLOW: 602 proto = ENA_ADMIN_RSS_TCP4; 603 break; 604 case UDP_V4_FLOW: 605 proto = ENA_ADMIN_RSS_UDP4; 606 break; 607 case TCP_V6_FLOW: 608 proto = ENA_ADMIN_RSS_TCP6; 609 break; 610 case UDP_V6_FLOW: 611 proto = ENA_ADMIN_RSS_UDP6; 612 break; 613 case IPV4_FLOW: 614 proto = ENA_ADMIN_RSS_IP4; 615 break; 616 case IPV6_FLOW: 617 proto = ENA_ADMIN_RSS_IP6; 618 break; 619 case ETHER_FLOW: 620 proto = ENA_ADMIN_RSS_NOT_IP; 621 break; 622 case AH_V4_FLOW: 623 case ESP_V4_FLOW: 624 case AH_V6_FLOW: 625 case ESP_V6_FLOW: 626 case SCTP_V4_FLOW: 627 case AH_ESP_V4_FLOW: 628 return -EOPNOTSUPP; 629 default: 630 return -EINVAL; 631 } 632 633 hash_fields = ena_flow_data_to_flow_hash(cmd->data); 634 635 return ena_com_fill_hash_ctrl(ena_dev, proto, hash_fields); 636 } 637 638 static int ena_set_rxnfc(struct net_device *netdev, struct ethtool_rxnfc *info) 639 { 640 struct ena_adapter *adapter = netdev_priv(netdev); 641 int rc = 0; 642 643 switch (info->cmd) { 644 case ETHTOOL_SRXFH: 645 rc = ena_set_rss_hash(adapter->ena_dev, info); 646 break; 647 case ETHTOOL_SRXCLSRLDEL: 648 case ETHTOOL_SRXCLSRLINS: 649 default: 650 netif_err(adapter, drv, netdev, 651 "Command parameter %d is not supported\n", info->cmd); 652 rc = -EOPNOTSUPP; 653 } 654 655 return rc; 656 } 657 658 static int ena_get_rxnfc(struct net_device *netdev, struct ethtool_rxnfc *info, 659 u32 *rules) 660 { 661 struct ena_adapter *adapter = netdev_priv(netdev); 662 int rc = 0; 663 664 switch (info->cmd) { 665 case ETHTOOL_GRXRINGS: 666 info->data = adapter->num_io_queues; 667 rc = 0; 668 break; 669 case ETHTOOL_GRXFH: 670 rc = ena_get_rss_hash(adapter->ena_dev, info); 671 break; 672 case ETHTOOL_GRXCLSRLCNT: 673 case ETHTOOL_GRXCLSRULE: 674 case ETHTOOL_GRXCLSRLALL: 675 default: 676 netif_err(adapter, drv, netdev, 677 "Command parameter %d is not supported\n", info->cmd); 678 rc = -EOPNOTSUPP; 679 } 680 681 return rc; 682 } 683 684 static u32 ena_get_rxfh_indir_size(struct net_device *netdev) 685 { 686 return ENA_RX_RSS_TABLE_SIZE; 687 } 688 689 static u32 ena_get_rxfh_key_size(struct net_device *netdev) 690 { 691 return ENA_HASH_KEY_SIZE; 692 } 693 694 static int ena_indirection_table_set(struct ena_adapter *adapter, 695 const u32 *indir) 696 { 697 struct ena_com_dev *ena_dev = adapter->ena_dev; 698 int i, rc; 699 700 for (i = 0; i < ENA_RX_RSS_TABLE_SIZE; i++) { 701 rc = ena_com_indirect_table_fill_entry(ena_dev, 702 i, 703 ENA_IO_RXQ_IDX(indir[i])); 704 if (unlikely(rc)) { 705 netif_err(adapter, drv, adapter->netdev, 706 "Cannot fill indirect table (index is too large)\n"); 707 return rc; 708 } 709 } 710 711 rc = ena_com_indirect_table_set(ena_dev); 712 if (rc) { 713 netif_err(adapter, drv, adapter->netdev, 714 "Cannot set indirect table\n"); 715 return rc == -EPERM ? -EOPNOTSUPP : rc; 716 } 717 return rc; 718 } 719 720 static int ena_indirection_table_get(struct ena_adapter *adapter, u32 *indir) 721 { 722 struct ena_com_dev *ena_dev = adapter->ena_dev; 723 int i, rc; 724 725 if (!indir) 726 return 0; 727 728 rc = ena_com_indirect_table_get(ena_dev, indir); 729 if (rc) 730 return rc; 731 732 /* Our internal representation of the indices is: even indices 733 * for Tx and uneven indices for Rx. We need to convert the Rx 734 * indices to be consecutive 735 */ 736 for (i = 0; i < ENA_RX_RSS_TABLE_SIZE; i++) 737 indir[i] = ENA_IO_RXQ_IDX_TO_COMBINED_IDX(indir[i]); 738 739 return rc; 740 } 741 742 static int ena_get_rxfh(struct net_device *netdev, u32 *indir, u8 *key, 743 u8 *hfunc) 744 { 745 struct ena_adapter *adapter = netdev_priv(netdev); 746 enum ena_admin_hash_functions ena_func; 747 u8 func; 748 int rc; 749 750 rc = ena_indirection_table_get(adapter, indir); 751 if (rc) 752 return rc; 753 754 /* We call this function in order to check if the device 755 * supports getting/setting the hash function. 756 */ 757 rc = ena_com_get_hash_function(adapter->ena_dev, &ena_func); 758 if (rc) { 759 if (rc == -EOPNOTSUPP) 760 rc = 0; 761 762 return rc; 763 } 764 765 rc = ena_com_get_hash_key(adapter->ena_dev, key); 766 if (rc) 767 return rc; 768 769 switch (ena_func) { 770 case ENA_ADMIN_TOEPLITZ: 771 func = ETH_RSS_HASH_TOP; 772 break; 773 case ENA_ADMIN_CRC32: 774 func = ETH_RSS_HASH_CRC32; 775 break; 776 default: 777 netif_err(adapter, drv, netdev, 778 "Command parameter is not supported\n"); 779 return -EOPNOTSUPP; 780 } 781 782 if (hfunc) 783 *hfunc = func; 784 785 return 0; 786 } 787 788 static int ena_set_rxfh(struct net_device *netdev, const u32 *indir, 789 const u8 *key, const u8 hfunc) 790 { 791 struct ena_adapter *adapter = netdev_priv(netdev); 792 struct ena_com_dev *ena_dev = adapter->ena_dev; 793 enum ena_admin_hash_functions func = 0; 794 int rc; 795 796 if (indir) { 797 rc = ena_indirection_table_set(adapter, indir); 798 if (rc) 799 return rc; 800 } 801 802 switch (hfunc) { 803 case ETH_RSS_HASH_NO_CHANGE: 804 func = ena_com_get_current_hash_function(ena_dev); 805 break; 806 case ETH_RSS_HASH_TOP: 807 func = ENA_ADMIN_TOEPLITZ; 808 break; 809 case ETH_RSS_HASH_CRC32: 810 func = ENA_ADMIN_CRC32; 811 break; 812 default: 813 netif_err(adapter, drv, netdev, "Unsupported hfunc %d\n", 814 hfunc); 815 return -EOPNOTSUPP; 816 } 817 818 if (key || func) { 819 rc = ena_com_fill_hash_function(ena_dev, func, key, 820 ENA_HASH_KEY_SIZE, 821 0xFFFFFFFF); 822 if (unlikely(rc)) { 823 netif_err(adapter, drv, netdev, "Cannot fill key\n"); 824 return rc == -EPERM ? -EOPNOTSUPP : rc; 825 } 826 } 827 828 return 0; 829 } 830 831 static void ena_get_channels(struct net_device *netdev, 832 struct ethtool_channels *channels) 833 { 834 struct ena_adapter *adapter = netdev_priv(netdev); 835 836 channels->max_combined = adapter->max_num_io_queues; 837 channels->combined_count = adapter->num_io_queues; 838 } 839 840 static int ena_set_channels(struct net_device *netdev, 841 struct ethtool_channels *channels) 842 { 843 struct ena_adapter *adapter = netdev_priv(netdev); 844 u32 count = channels->combined_count; 845 /* The check for max value is already done in ethtool */ 846 if (count < ENA_MIN_NUM_IO_QUEUES || 847 (ena_xdp_present(adapter) && 848 !ena_xdp_legal_queue_count(adapter, count))) 849 return -EINVAL; 850 851 return ena_update_queue_count(adapter, count); 852 } 853 854 static int ena_get_tunable(struct net_device *netdev, 855 const struct ethtool_tunable *tuna, void *data) 856 { 857 struct ena_adapter *adapter = netdev_priv(netdev); 858 int ret = 0; 859 860 switch (tuna->id) { 861 case ETHTOOL_RX_COPYBREAK: 862 *(u32 *)data = adapter->rx_copybreak; 863 break; 864 default: 865 ret = -EINVAL; 866 break; 867 } 868 869 return ret; 870 } 871 872 static int ena_set_tunable(struct net_device *netdev, 873 const struct ethtool_tunable *tuna, 874 const void *data) 875 { 876 struct ena_adapter *adapter = netdev_priv(netdev); 877 int ret = 0; 878 u32 len; 879 880 switch (tuna->id) { 881 case ETHTOOL_RX_COPYBREAK: 882 len = *(u32 *)data; 883 if (len > adapter->netdev->mtu) { 884 ret = -EINVAL; 885 break; 886 } 887 adapter->rx_copybreak = len; 888 break; 889 default: 890 ret = -EINVAL; 891 break; 892 } 893 894 return ret; 895 } 896 897 static const struct ethtool_ops ena_ethtool_ops = { 898 .supported_coalesce_params = ETHTOOL_COALESCE_USECS | 899 ETHTOOL_COALESCE_USE_ADAPTIVE_RX, 900 .get_link_ksettings = ena_get_link_ksettings, 901 .get_drvinfo = ena_get_drvinfo, 902 .get_msglevel = ena_get_msglevel, 903 .set_msglevel = ena_set_msglevel, 904 .get_link = ethtool_op_get_link, 905 .get_coalesce = ena_get_coalesce, 906 .set_coalesce = ena_set_coalesce, 907 .get_ringparam = ena_get_ringparam, 908 .set_ringparam = ena_set_ringparam, 909 .get_sset_count = ena_get_sset_count, 910 .get_strings = ena_get_ethtool_strings, 911 .get_ethtool_stats = ena_get_ethtool_stats, 912 .get_rxnfc = ena_get_rxnfc, 913 .set_rxnfc = ena_set_rxnfc, 914 .get_rxfh_indir_size = ena_get_rxfh_indir_size, 915 .get_rxfh_key_size = ena_get_rxfh_key_size, 916 .get_rxfh = ena_get_rxfh, 917 .set_rxfh = ena_set_rxfh, 918 .get_channels = ena_get_channels, 919 .set_channels = ena_set_channels, 920 .get_tunable = ena_get_tunable, 921 .set_tunable = ena_set_tunable, 922 .get_ts_info = ethtool_op_get_ts_info, 923 }; 924 925 void ena_set_ethtool_ops(struct net_device *netdev) 926 { 927 netdev->ethtool_ops = &ena_ethtool_ops; 928 } 929 930 static void ena_dump_stats_ex(struct ena_adapter *adapter, u8 *buf) 931 { 932 struct net_device *netdev = adapter->netdev; 933 u8 *strings_buf; 934 u64 *data_buf; 935 int strings_num; 936 int i, rc; 937 938 strings_num = ena_get_sw_stats_count(adapter); 939 if (strings_num <= 0) { 940 netif_err(adapter, drv, netdev, "Can't get stats num\n"); 941 return; 942 } 943 944 strings_buf = devm_kcalloc(&adapter->pdev->dev, 945 ETH_GSTRING_LEN, strings_num, 946 GFP_ATOMIC); 947 if (!strings_buf) { 948 netif_err(adapter, drv, netdev, 949 "Failed to allocate strings_buf\n"); 950 return; 951 } 952 953 data_buf = devm_kcalloc(&adapter->pdev->dev, 954 strings_num, sizeof(u64), 955 GFP_ATOMIC); 956 if (!data_buf) { 957 netif_err(adapter, drv, netdev, 958 "Failed to allocate data buf\n"); 959 devm_kfree(&adapter->pdev->dev, strings_buf); 960 return; 961 } 962 963 ena_get_strings(adapter, strings_buf, false); 964 ena_get_stats(adapter, data_buf, false); 965 966 /* If there is a buffer, dump stats, otherwise print them to dmesg */ 967 if (buf) 968 for (i = 0; i < strings_num; i++) { 969 rc = snprintf(buf, ETH_GSTRING_LEN + sizeof(u64), 970 "%s %llu\n", 971 strings_buf + i * ETH_GSTRING_LEN, 972 data_buf[i]); 973 buf += rc; 974 } 975 else 976 for (i = 0; i < strings_num; i++) 977 netif_err(adapter, drv, netdev, "%s: %llu\n", 978 strings_buf + i * ETH_GSTRING_LEN, 979 data_buf[i]); 980 981 devm_kfree(&adapter->pdev->dev, strings_buf); 982 devm_kfree(&adapter->pdev->dev, data_buf); 983 } 984 985 void ena_dump_stats_to_buf(struct ena_adapter *adapter, u8 *buf) 986 { 987 if (!buf) 988 return; 989 990 ena_dump_stats_ex(adapter, buf); 991 } 992 993 void ena_dump_stats_to_dmesg(struct ena_adapter *adapter) 994 { 995 ena_dump_stats_ex(adapter, NULL); 996 } 997