1 /* 2 * Copyright 2015 Amazon.com, Inc. or its affiliates. 3 * 4 * This software is available to you under a choice of one of two 5 * licenses. You may choose to be licensed under the terms of the GNU 6 * General Public License (GPL) Version 2, available from the file 7 * COPYING in the main directory of this source tree, or the 8 * BSD license below: 9 * 10 * Redistribution and use in source and binary forms, with or 11 * without modification, are permitted provided that the following 12 * conditions are met: 13 * 14 * - Redistributions of source code must retain the above 15 * copyright notice, this list of conditions and the following 16 * disclaimer. 17 * 18 * - Redistributions in binary form must reproduce the above 19 * copyright notice, this list of conditions and the following 20 * disclaimer in the documentation and/or other materials 21 * provided with the distribution. 22 * 23 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, 24 * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF 25 * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND 26 * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS 27 * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN 28 * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN 29 * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE 30 * SOFTWARE. 31 */ 32 33 #include <linux/pci.h> 34 35 #include "ena_netdev.h" 36 37 struct ena_stats { 38 char name[ETH_GSTRING_LEN]; 39 int stat_offset; 40 }; 41 42 #define ENA_STAT_ENA_COM_ENTRY(stat) { \ 43 .name = #stat, \ 44 .stat_offset = offsetof(struct ena_com_stats_admin, stat) \ 45 } 46 47 #define ENA_STAT_ENTRY(stat, stat_type) { \ 48 .name = #stat, \ 49 .stat_offset = offsetof(struct ena_stats_##stat_type, stat) \ 50 } 51 52 #define ENA_STAT_RX_ENTRY(stat) \ 53 ENA_STAT_ENTRY(stat, rx) 54 55 #define ENA_STAT_TX_ENTRY(stat) \ 56 ENA_STAT_ENTRY(stat, tx) 57 58 #define ENA_STAT_GLOBAL_ENTRY(stat) \ 59 ENA_STAT_ENTRY(stat, dev) 60 61 static const struct ena_stats ena_stats_global_strings[] = { 62 ENA_STAT_GLOBAL_ENTRY(tx_timeout), 63 ENA_STAT_GLOBAL_ENTRY(suspend), 64 ENA_STAT_GLOBAL_ENTRY(resume), 65 ENA_STAT_GLOBAL_ENTRY(wd_expired), 66 ENA_STAT_GLOBAL_ENTRY(interface_up), 67 ENA_STAT_GLOBAL_ENTRY(interface_down), 68 ENA_STAT_GLOBAL_ENTRY(admin_q_pause), 69 }; 70 71 static const struct ena_stats ena_stats_tx_strings[] = { 72 ENA_STAT_TX_ENTRY(cnt), 73 ENA_STAT_TX_ENTRY(bytes), 74 ENA_STAT_TX_ENTRY(queue_stop), 75 ENA_STAT_TX_ENTRY(queue_wakeup), 76 ENA_STAT_TX_ENTRY(dma_mapping_err), 77 ENA_STAT_TX_ENTRY(linearize), 78 ENA_STAT_TX_ENTRY(linearize_failed), 79 ENA_STAT_TX_ENTRY(napi_comp), 80 ENA_STAT_TX_ENTRY(tx_poll), 81 ENA_STAT_TX_ENTRY(doorbells), 82 ENA_STAT_TX_ENTRY(prepare_ctx_err), 83 ENA_STAT_TX_ENTRY(bad_req_id), 84 ENA_STAT_TX_ENTRY(llq_buffer_copy), 85 ENA_STAT_TX_ENTRY(missed_tx), 86 }; 87 88 static const struct ena_stats ena_stats_rx_strings[] = { 89 ENA_STAT_RX_ENTRY(cnt), 90 ENA_STAT_RX_ENTRY(bytes), 91 ENA_STAT_RX_ENTRY(rx_copybreak_pkt), 92 ENA_STAT_RX_ENTRY(csum_good), 93 ENA_STAT_RX_ENTRY(refil_partial), 94 ENA_STAT_RX_ENTRY(bad_csum), 95 ENA_STAT_RX_ENTRY(page_alloc_fail), 96 ENA_STAT_RX_ENTRY(skb_alloc_fail), 97 ENA_STAT_RX_ENTRY(dma_mapping_err), 98 ENA_STAT_RX_ENTRY(bad_desc_num), 99 ENA_STAT_RX_ENTRY(bad_req_id), 100 ENA_STAT_RX_ENTRY(empty_rx_ring), 101 ENA_STAT_RX_ENTRY(csum_unchecked), 102 }; 103 104 static const struct ena_stats ena_stats_ena_com_strings[] = { 105 ENA_STAT_ENA_COM_ENTRY(aborted_cmd), 106 ENA_STAT_ENA_COM_ENTRY(submitted_cmd), 107 ENA_STAT_ENA_COM_ENTRY(completed_cmd), 108 ENA_STAT_ENA_COM_ENTRY(out_of_space), 109 ENA_STAT_ENA_COM_ENTRY(no_completion), 110 }; 111 112 #define ENA_STATS_ARRAY_GLOBAL ARRAY_SIZE(ena_stats_global_strings) 113 #define ENA_STATS_ARRAY_TX ARRAY_SIZE(ena_stats_tx_strings) 114 #define ENA_STATS_ARRAY_RX ARRAY_SIZE(ena_stats_rx_strings) 115 #define ENA_STATS_ARRAY_ENA_COM ARRAY_SIZE(ena_stats_ena_com_strings) 116 117 static void ena_safe_update_stat(u64 *src, u64 *dst, 118 struct u64_stats_sync *syncp) 119 { 120 unsigned int start; 121 122 do { 123 start = u64_stats_fetch_begin_irq(syncp); 124 *(dst) = *src; 125 } while (u64_stats_fetch_retry_irq(syncp, start)); 126 } 127 128 static void ena_queue_stats(struct ena_adapter *adapter, u64 **data) 129 { 130 const struct ena_stats *ena_stats; 131 struct ena_ring *ring; 132 133 u64 *ptr; 134 int i, j; 135 136 for (i = 0; i < adapter->num_io_queues; i++) { 137 /* Tx stats */ 138 ring = &adapter->tx_ring[i]; 139 140 for (j = 0; j < ENA_STATS_ARRAY_TX; j++) { 141 ena_stats = &ena_stats_tx_strings[j]; 142 143 ptr = (u64 *)((uintptr_t)&ring->tx_stats + 144 (uintptr_t)ena_stats->stat_offset); 145 146 ena_safe_update_stat(ptr, (*data)++, &ring->syncp); 147 } 148 149 /* Rx stats */ 150 ring = &adapter->rx_ring[i]; 151 152 for (j = 0; j < ENA_STATS_ARRAY_RX; j++) { 153 ena_stats = &ena_stats_rx_strings[j]; 154 155 ptr = (u64 *)((uintptr_t)&ring->rx_stats + 156 (uintptr_t)ena_stats->stat_offset); 157 158 ena_safe_update_stat(ptr, (*data)++, &ring->syncp); 159 } 160 } 161 } 162 163 static void ena_dev_admin_queue_stats(struct ena_adapter *adapter, u64 **data) 164 { 165 const struct ena_stats *ena_stats; 166 u32 *ptr; 167 int i; 168 169 for (i = 0; i < ENA_STATS_ARRAY_ENA_COM; i++) { 170 ena_stats = &ena_stats_ena_com_strings[i]; 171 172 ptr = (u32 *)((uintptr_t)&adapter->ena_dev->admin_queue.stats + 173 (uintptr_t)ena_stats->stat_offset); 174 175 *(*data)++ = *ptr; 176 } 177 } 178 179 static void ena_get_ethtool_stats(struct net_device *netdev, 180 struct ethtool_stats *stats, 181 u64 *data) 182 { 183 struct ena_adapter *adapter = netdev_priv(netdev); 184 const struct ena_stats *ena_stats; 185 u64 *ptr; 186 int i; 187 188 for (i = 0; i < ENA_STATS_ARRAY_GLOBAL; i++) { 189 ena_stats = &ena_stats_global_strings[i]; 190 191 ptr = (u64 *)((uintptr_t)&adapter->dev_stats + 192 (uintptr_t)ena_stats->stat_offset); 193 194 ena_safe_update_stat(ptr, data++, &adapter->syncp); 195 } 196 197 ena_queue_stats(adapter, &data); 198 ena_dev_admin_queue_stats(adapter, &data); 199 } 200 201 int ena_get_sset_count(struct net_device *netdev, int sset) 202 { 203 struct ena_adapter *adapter = netdev_priv(netdev); 204 205 if (sset != ETH_SS_STATS) 206 return -EOPNOTSUPP; 207 208 return adapter->num_io_queues * (ENA_STATS_ARRAY_TX + ENA_STATS_ARRAY_RX) 209 + ENA_STATS_ARRAY_GLOBAL + ENA_STATS_ARRAY_ENA_COM; 210 } 211 212 static void ena_queue_strings(struct ena_adapter *adapter, u8 **data) 213 { 214 const struct ena_stats *ena_stats; 215 int i, j; 216 217 for (i = 0; i < adapter->num_io_queues; i++) { 218 /* Tx stats */ 219 for (j = 0; j < ENA_STATS_ARRAY_TX; j++) { 220 ena_stats = &ena_stats_tx_strings[j]; 221 222 snprintf(*data, ETH_GSTRING_LEN, 223 "queue_%u_tx_%s", i, ena_stats->name); 224 (*data) += ETH_GSTRING_LEN; 225 } 226 /* Rx stats */ 227 for (j = 0; j < ENA_STATS_ARRAY_RX; j++) { 228 ena_stats = &ena_stats_rx_strings[j]; 229 230 snprintf(*data, ETH_GSTRING_LEN, 231 "queue_%u_rx_%s", i, ena_stats->name); 232 (*data) += ETH_GSTRING_LEN; 233 } 234 } 235 } 236 237 static void ena_com_dev_strings(u8 **data) 238 { 239 const struct ena_stats *ena_stats; 240 int i; 241 242 for (i = 0; i < ENA_STATS_ARRAY_ENA_COM; i++) { 243 ena_stats = &ena_stats_ena_com_strings[i]; 244 245 snprintf(*data, ETH_GSTRING_LEN, 246 "ena_admin_q_%s", ena_stats->name); 247 (*data) += ETH_GSTRING_LEN; 248 } 249 } 250 251 static void ena_get_strings(struct net_device *netdev, u32 sset, u8 *data) 252 { 253 struct ena_adapter *adapter = netdev_priv(netdev); 254 const struct ena_stats *ena_stats; 255 int i; 256 257 if (sset != ETH_SS_STATS) 258 return; 259 260 for (i = 0; i < ENA_STATS_ARRAY_GLOBAL; i++) { 261 ena_stats = &ena_stats_global_strings[i]; 262 263 memcpy(data, ena_stats->name, ETH_GSTRING_LEN); 264 data += ETH_GSTRING_LEN; 265 } 266 267 ena_queue_strings(adapter, &data); 268 ena_com_dev_strings(&data); 269 } 270 271 static int ena_get_link_ksettings(struct net_device *netdev, 272 struct ethtool_link_ksettings *link_ksettings) 273 { 274 struct ena_adapter *adapter = netdev_priv(netdev); 275 struct ena_com_dev *ena_dev = adapter->ena_dev; 276 struct ena_admin_get_feature_link_desc *link; 277 struct ena_admin_get_feat_resp feat_resp; 278 int rc; 279 280 rc = ena_com_get_link_params(ena_dev, &feat_resp); 281 if (rc) 282 return rc; 283 284 link = &feat_resp.u.link; 285 link_ksettings->base.speed = link->speed; 286 287 if (link->flags & ENA_ADMIN_GET_FEATURE_LINK_DESC_AUTONEG_MASK) { 288 ethtool_link_ksettings_add_link_mode(link_ksettings, 289 supported, Autoneg); 290 ethtool_link_ksettings_add_link_mode(link_ksettings, 291 supported, Autoneg); 292 } 293 294 link_ksettings->base.autoneg = 295 (link->flags & ENA_ADMIN_GET_FEATURE_LINK_DESC_AUTONEG_MASK) ? 296 AUTONEG_ENABLE : AUTONEG_DISABLE; 297 298 link_ksettings->base.duplex = DUPLEX_FULL; 299 300 return 0; 301 } 302 303 static int ena_get_coalesce(struct net_device *net_dev, 304 struct ethtool_coalesce *coalesce) 305 { 306 struct ena_adapter *adapter = netdev_priv(net_dev); 307 struct ena_com_dev *ena_dev = adapter->ena_dev; 308 309 if (!ena_com_interrupt_moderation_supported(ena_dev)) { 310 /* the devie doesn't support interrupt moderation */ 311 return -EOPNOTSUPP; 312 } 313 314 coalesce->tx_coalesce_usecs = 315 ena_com_get_nonadaptive_moderation_interval_tx(ena_dev) * 316 ena_dev->intr_delay_resolution; 317 318 coalesce->rx_coalesce_usecs = 319 ena_com_get_nonadaptive_moderation_interval_rx(ena_dev) 320 * ena_dev->intr_delay_resolution; 321 322 coalesce->use_adaptive_rx_coalesce = 323 ena_com_get_adaptive_moderation_enabled(ena_dev); 324 325 return 0; 326 } 327 328 static void ena_update_tx_rings_intr_moderation(struct ena_adapter *adapter) 329 { 330 unsigned int val; 331 int i; 332 333 val = ena_com_get_nonadaptive_moderation_interval_tx(adapter->ena_dev); 334 335 for (i = 0; i < adapter->num_io_queues; i++) 336 adapter->tx_ring[i].smoothed_interval = val; 337 } 338 339 static void ena_update_rx_rings_intr_moderation(struct ena_adapter *adapter) 340 { 341 unsigned int val; 342 int i; 343 344 val = ena_com_get_nonadaptive_moderation_interval_rx(adapter->ena_dev); 345 346 for (i = 0; i < adapter->num_io_queues; i++) 347 adapter->rx_ring[i].smoothed_interval = val; 348 } 349 350 static int ena_set_coalesce(struct net_device *net_dev, 351 struct ethtool_coalesce *coalesce) 352 { 353 struct ena_adapter *adapter = netdev_priv(net_dev); 354 struct ena_com_dev *ena_dev = adapter->ena_dev; 355 int rc; 356 357 if (!ena_com_interrupt_moderation_supported(ena_dev)) { 358 /* the devie doesn't support interrupt moderation */ 359 return -EOPNOTSUPP; 360 } 361 362 rc = ena_com_update_nonadaptive_moderation_interval_tx(ena_dev, 363 coalesce->tx_coalesce_usecs); 364 if (rc) 365 return rc; 366 367 ena_update_tx_rings_intr_moderation(adapter); 368 369 rc = ena_com_update_nonadaptive_moderation_interval_rx(ena_dev, 370 coalesce->rx_coalesce_usecs); 371 if (rc) 372 return rc; 373 374 ena_update_rx_rings_intr_moderation(adapter); 375 376 if (coalesce->use_adaptive_rx_coalesce && 377 !ena_com_get_adaptive_moderation_enabled(ena_dev)) 378 ena_com_enable_adaptive_moderation(ena_dev); 379 380 if (!coalesce->use_adaptive_rx_coalesce && 381 ena_com_get_adaptive_moderation_enabled(ena_dev)) 382 ena_com_disable_adaptive_moderation(ena_dev); 383 384 return 0; 385 } 386 387 static u32 ena_get_msglevel(struct net_device *netdev) 388 { 389 struct ena_adapter *adapter = netdev_priv(netdev); 390 391 return adapter->msg_enable; 392 } 393 394 static void ena_set_msglevel(struct net_device *netdev, u32 value) 395 { 396 struct ena_adapter *adapter = netdev_priv(netdev); 397 398 adapter->msg_enable = value; 399 } 400 401 static void ena_get_drvinfo(struct net_device *dev, 402 struct ethtool_drvinfo *info) 403 { 404 struct ena_adapter *adapter = netdev_priv(dev); 405 406 strlcpy(info->driver, DRV_MODULE_NAME, sizeof(info->driver)); 407 strlcpy(info->bus_info, pci_name(adapter->pdev), 408 sizeof(info->bus_info)); 409 } 410 411 static void ena_get_ringparam(struct net_device *netdev, 412 struct ethtool_ringparam *ring) 413 { 414 struct ena_adapter *adapter = netdev_priv(netdev); 415 416 ring->tx_max_pending = adapter->max_tx_ring_size; 417 ring->rx_max_pending = adapter->max_rx_ring_size; 418 ring->tx_pending = adapter->tx_ring[0].ring_size; 419 ring->rx_pending = adapter->rx_ring[0].ring_size; 420 } 421 422 static int ena_set_ringparam(struct net_device *netdev, 423 struct ethtool_ringparam *ring) 424 { 425 struct ena_adapter *adapter = netdev_priv(netdev); 426 u32 new_tx_size, new_rx_size; 427 428 new_tx_size = ring->tx_pending < ENA_MIN_RING_SIZE ? 429 ENA_MIN_RING_SIZE : ring->tx_pending; 430 new_tx_size = rounddown_pow_of_two(new_tx_size); 431 432 new_rx_size = ring->rx_pending < ENA_MIN_RING_SIZE ? 433 ENA_MIN_RING_SIZE : ring->rx_pending; 434 new_rx_size = rounddown_pow_of_two(new_rx_size); 435 436 if (new_tx_size == adapter->requested_tx_ring_size && 437 new_rx_size == adapter->requested_rx_ring_size) 438 return 0; 439 440 return ena_update_queue_sizes(adapter, new_tx_size, new_rx_size); 441 } 442 443 static u32 ena_flow_hash_to_flow_type(u16 hash_fields) 444 { 445 u32 data = 0; 446 447 if (hash_fields & ENA_ADMIN_RSS_L2_DA) 448 data |= RXH_L2DA; 449 450 if (hash_fields & ENA_ADMIN_RSS_L3_DA) 451 data |= RXH_IP_DST; 452 453 if (hash_fields & ENA_ADMIN_RSS_L3_SA) 454 data |= RXH_IP_SRC; 455 456 if (hash_fields & ENA_ADMIN_RSS_L4_DP) 457 data |= RXH_L4_B_2_3; 458 459 if (hash_fields & ENA_ADMIN_RSS_L4_SP) 460 data |= RXH_L4_B_0_1; 461 462 return data; 463 } 464 465 static u16 ena_flow_data_to_flow_hash(u32 hash_fields) 466 { 467 u16 data = 0; 468 469 if (hash_fields & RXH_L2DA) 470 data |= ENA_ADMIN_RSS_L2_DA; 471 472 if (hash_fields & RXH_IP_DST) 473 data |= ENA_ADMIN_RSS_L3_DA; 474 475 if (hash_fields & RXH_IP_SRC) 476 data |= ENA_ADMIN_RSS_L3_SA; 477 478 if (hash_fields & RXH_L4_B_2_3) 479 data |= ENA_ADMIN_RSS_L4_DP; 480 481 if (hash_fields & RXH_L4_B_0_1) 482 data |= ENA_ADMIN_RSS_L4_SP; 483 484 return data; 485 } 486 487 static int ena_get_rss_hash(struct ena_com_dev *ena_dev, 488 struct ethtool_rxnfc *cmd) 489 { 490 enum ena_admin_flow_hash_proto proto; 491 u16 hash_fields; 492 int rc; 493 494 cmd->data = 0; 495 496 switch (cmd->flow_type) { 497 case TCP_V4_FLOW: 498 proto = ENA_ADMIN_RSS_TCP4; 499 break; 500 case UDP_V4_FLOW: 501 proto = ENA_ADMIN_RSS_UDP4; 502 break; 503 case TCP_V6_FLOW: 504 proto = ENA_ADMIN_RSS_TCP6; 505 break; 506 case UDP_V6_FLOW: 507 proto = ENA_ADMIN_RSS_UDP6; 508 break; 509 case IPV4_FLOW: 510 proto = ENA_ADMIN_RSS_IP4; 511 break; 512 case IPV6_FLOW: 513 proto = ENA_ADMIN_RSS_IP6; 514 break; 515 case ETHER_FLOW: 516 proto = ENA_ADMIN_RSS_NOT_IP; 517 break; 518 case AH_V4_FLOW: 519 case ESP_V4_FLOW: 520 case AH_V6_FLOW: 521 case ESP_V6_FLOW: 522 case SCTP_V4_FLOW: 523 case AH_ESP_V4_FLOW: 524 return -EOPNOTSUPP; 525 default: 526 return -EINVAL; 527 } 528 529 rc = ena_com_get_hash_ctrl(ena_dev, proto, &hash_fields); 530 if (rc) 531 return rc; 532 533 cmd->data = ena_flow_hash_to_flow_type(hash_fields); 534 535 return 0; 536 } 537 538 static int ena_set_rss_hash(struct ena_com_dev *ena_dev, 539 struct ethtool_rxnfc *cmd) 540 { 541 enum ena_admin_flow_hash_proto proto; 542 u16 hash_fields; 543 544 switch (cmd->flow_type) { 545 case TCP_V4_FLOW: 546 proto = ENA_ADMIN_RSS_TCP4; 547 break; 548 case UDP_V4_FLOW: 549 proto = ENA_ADMIN_RSS_UDP4; 550 break; 551 case TCP_V6_FLOW: 552 proto = ENA_ADMIN_RSS_TCP6; 553 break; 554 case UDP_V6_FLOW: 555 proto = ENA_ADMIN_RSS_UDP6; 556 break; 557 case IPV4_FLOW: 558 proto = ENA_ADMIN_RSS_IP4; 559 break; 560 case IPV6_FLOW: 561 proto = ENA_ADMIN_RSS_IP6; 562 break; 563 case ETHER_FLOW: 564 proto = ENA_ADMIN_RSS_NOT_IP; 565 break; 566 case AH_V4_FLOW: 567 case ESP_V4_FLOW: 568 case AH_V6_FLOW: 569 case ESP_V6_FLOW: 570 case SCTP_V4_FLOW: 571 case AH_ESP_V4_FLOW: 572 return -EOPNOTSUPP; 573 default: 574 return -EINVAL; 575 } 576 577 hash_fields = ena_flow_data_to_flow_hash(cmd->data); 578 579 return ena_com_fill_hash_ctrl(ena_dev, proto, hash_fields); 580 } 581 582 static int ena_set_rxnfc(struct net_device *netdev, struct ethtool_rxnfc *info) 583 { 584 struct ena_adapter *adapter = netdev_priv(netdev); 585 int rc = 0; 586 587 switch (info->cmd) { 588 case ETHTOOL_SRXFH: 589 rc = ena_set_rss_hash(adapter->ena_dev, info); 590 break; 591 case ETHTOOL_SRXCLSRLDEL: 592 case ETHTOOL_SRXCLSRLINS: 593 default: 594 netif_err(adapter, drv, netdev, 595 "Command parameter %d is not supported\n", info->cmd); 596 rc = -EOPNOTSUPP; 597 } 598 599 return rc; 600 } 601 602 static int ena_get_rxnfc(struct net_device *netdev, struct ethtool_rxnfc *info, 603 u32 *rules) 604 { 605 struct ena_adapter *adapter = netdev_priv(netdev); 606 int rc = 0; 607 608 switch (info->cmd) { 609 case ETHTOOL_GRXRINGS: 610 info->data = adapter->num_io_queues; 611 rc = 0; 612 break; 613 case ETHTOOL_GRXFH: 614 rc = ena_get_rss_hash(adapter->ena_dev, info); 615 break; 616 case ETHTOOL_GRXCLSRLCNT: 617 case ETHTOOL_GRXCLSRULE: 618 case ETHTOOL_GRXCLSRLALL: 619 default: 620 netif_err(adapter, drv, netdev, 621 "Command parameter %d is not supported\n", info->cmd); 622 rc = -EOPNOTSUPP; 623 } 624 625 return rc; 626 } 627 628 static u32 ena_get_rxfh_indir_size(struct net_device *netdev) 629 { 630 return ENA_RX_RSS_TABLE_SIZE; 631 } 632 633 static u32 ena_get_rxfh_key_size(struct net_device *netdev) 634 { 635 return ENA_HASH_KEY_SIZE; 636 } 637 638 static int ena_indirection_table_get(struct ena_adapter *adapter, u32 *indir) 639 { 640 struct ena_com_dev *ena_dev = adapter->ena_dev; 641 int i, rc; 642 643 if (!indir) 644 return 0; 645 646 rc = ena_com_indirect_table_get(ena_dev, indir); 647 if (rc) 648 return rc; 649 650 /* Our internal representation of the indices is: even indices 651 * for Tx and uneven indices for Rx. We need to convert the Rx 652 * indices to be consecutive 653 */ 654 for (i = 0; i < ENA_RX_RSS_TABLE_SIZE; i++) 655 indir[i] = ENA_IO_RXQ_IDX_TO_COMBINED_IDX(indir[i]); 656 657 return rc; 658 } 659 660 static int ena_get_rxfh(struct net_device *netdev, u32 *indir, u8 *key, 661 u8 *hfunc) 662 { 663 struct ena_adapter *adapter = netdev_priv(netdev); 664 enum ena_admin_hash_functions ena_func; 665 u8 func; 666 int rc; 667 668 rc = ena_indirection_table_get(adapter, indir); 669 if (rc) 670 return rc; 671 672 /* We call this function in order to check if the device 673 * supports getting/setting the hash function. 674 */ 675 rc = ena_com_get_hash_function(adapter->ena_dev, &ena_func, key); 676 if (rc) { 677 if (rc == -EOPNOTSUPP) { 678 key = NULL; 679 hfunc = NULL; 680 rc = 0; 681 } 682 683 return rc; 684 } 685 686 switch (ena_func) { 687 case ENA_ADMIN_TOEPLITZ: 688 func = ETH_RSS_HASH_TOP; 689 break; 690 case ENA_ADMIN_CRC32: 691 func = ETH_RSS_HASH_CRC32; 692 break; 693 default: 694 netif_err(adapter, drv, netdev, 695 "Command parameter is not supported\n"); 696 return -EOPNOTSUPP; 697 } 698 699 if (hfunc) 700 *hfunc = func; 701 702 return rc; 703 } 704 705 static int ena_set_rxfh(struct net_device *netdev, const u32 *indir, 706 const u8 *key, const u8 hfunc) 707 { 708 struct ena_adapter *adapter = netdev_priv(netdev); 709 struct ena_com_dev *ena_dev = adapter->ena_dev; 710 enum ena_admin_hash_functions func; 711 int rc, i; 712 713 if (indir) { 714 for (i = 0; i < ENA_RX_RSS_TABLE_SIZE; i++) { 715 rc = ena_com_indirect_table_fill_entry(ena_dev, 716 i, 717 ENA_IO_RXQ_IDX(indir[i])); 718 if (unlikely(rc)) { 719 netif_err(adapter, drv, netdev, 720 "Cannot fill indirect table (index is too large)\n"); 721 return rc; 722 } 723 } 724 725 rc = ena_com_indirect_table_set(ena_dev); 726 if (rc) { 727 netif_err(adapter, drv, netdev, 728 "Cannot set indirect table\n"); 729 return rc == -EPERM ? -EOPNOTSUPP : rc; 730 } 731 } 732 733 switch (hfunc) { 734 case ETH_RSS_HASH_NO_CHANGE: 735 func = ena_com_get_current_hash_function(ena_dev); 736 break; 737 case ETH_RSS_HASH_TOP: 738 func = ENA_ADMIN_TOEPLITZ; 739 break; 740 case ETH_RSS_HASH_CRC32: 741 func = ENA_ADMIN_CRC32; 742 break; 743 default: 744 netif_err(adapter, drv, netdev, "Unsupported hfunc %d\n", 745 hfunc); 746 return -EOPNOTSUPP; 747 } 748 749 if (key) { 750 rc = ena_com_fill_hash_function(ena_dev, func, key, 751 ENA_HASH_KEY_SIZE, 752 0xFFFFFFFF); 753 if (unlikely(rc)) { 754 netif_err(adapter, drv, netdev, "Cannot fill key\n"); 755 return rc == -EPERM ? -EOPNOTSUPP : rc; 756 } 757 } 758 759 return 0; 760 } 761 762 static void ena_get_channels(struct net_device *netdev, 763 struct ethtool_channels *channels) 764 { 765 struct ena_adapter *adapter = netdev_priv(netdev); 766 767 channels->max_combined = adapter->max_num_io_queues; 768 channels->combined_count = adapter->num_io_queues; 769 } 770 771 static int ena_set_channels(struct net_device *netdev, 772 struct ethtool_channels *channels) 773 { 774 struct ena_adapter *adapter = netdev_priv(netdev); 775 u32 count = channels->combined_count; 776 /* The check for max value is already done in ethtool */ 777 if (count < ENA_MIN_NUM_IO_QUEUES || 778 (ena_xdp_present(adapter) && 779 !ena_xdp_legal_queue_count(adapter, channels->combined_count))) 780 return -EINVAL; 781 782 return ena_update_queue_count(adapter, count); 783 } 784 785 static int ena_get_tunable(struct net_device *netdev, 786 const struct ethtool_tunable *tuna, void *data) 787 { 788 struct ena_adapter *adapter = netdev_priv(netdev); 789 int ret = 0; 790 791 switch (tuna->id) { 792 case ETHTOOL_RX_COPYBREAK: 793 *(u32 *)data = adapter->rx_copybreak; 794 break; 795 default: 796 ret = -EINVAL; 797 break; 798 } 799 800 return ret; 801 } 802 803 static int ena_set_tunable(struct net_device *netdev, 804 const struct ethtool_tunable *tuna, 805 const void *data) 806 { 807 struct ena_adapter *adapter = netdev_priv(netdev); 808 int ret = 0; 809 u32 len; 810 811 switch (tuna->id) { 812 case ETHTOOL_RX_COPYBREAK: 813 len = *(u32 *)data; 814 if (len > adapter->netdev->mtu) { 815 ret = -EINVAL; 816 break; 817 } 818 adapter->rx_copybreak = len; 819 break; 820 default: 821 ret = -EINVAL; 822 break; 823 } 824 825 return ret; 826 } 827 828 static const struct ethtool_ops ena_ethtool_ops = { 829 .supported_coalesce_params = ETHTOOL_COALESCE_USECS | 830 ETHTOOL_COALESCE_USE_ADAPTIVE_RX, 831 .get_link_ksettings = ena_get_link_ksettings, 832 .get_drvinfo = ena_get_drvinfo, 833 .get_msglevel = ena_get_msglevel, 834 .set_msglevel = ena_set_msglevel, 835 .get_link = ethtool_op_get_link, 836 .get_coalesce = ena_get_coalesce, 837 .set_coalesce = ena_set_coalesce, 838 .get_ringparam = ena_get_ringparam, 839 .set_ringparam = ena_set_ringparam, 840 .get_sset_count = ena_get_sset_count, 841 .get_strings = ena_get_strings, 842 .get_ethtool_stats = ena_get_ethtool_stats, 843 .get_rxnfc = ena_get_rxnfc, 844 .set_rxnfc = ena_set_rxnfc, 845 .get_rxfh_indir_size = ena_get_rxfh_indir_size, 846 .get_rxfh_key_size = ena_get_rxfh_key_size, 847 .get_rxfh = ena_get_rxfh, 848 .set_rxfh = ena_set_rxfh, 849 .get_channels = ena_get_channels, 850 .set_channels = ena_set_channels, 851 .get_tunable = ena_get_tunable, 852 .set_tunable = ena_set_tunable, 853 .get_ts_info = ethtool_op_get_ts_info, 854 }; 855 856 void ena_set_ethtool_ops(struct net_device *netdev) 857 { 858 netdev->ethtool_ops = &ena_ethtool_ops; 859 } 860 861 static void ena_dump_stats_ex(struct ena_adapter *adapter, u8 *buf) 862 { 863 struct net_device *netdev = adapter->netdev; 864 u8 *strings_buf; 865 u64 *data_buf; 866 int strings_num; 867 int i, rc; 868 869 strings_num = ena_get_sset_count(netdev, ETH_SS_STATS); 870 if (strings_num <= 0) { 871 netif_err(adapter, drv, netdev, "Can't get stats num\n"); 872 return; 873 } 874 875 strings_buf = devm_kcalloc(&adapter->pdev->dev, 876 ETH_GSTRING_LEN, strings_num, 877 GFP_ATOMIC); 878 if (!strings_buf) { 879 netif_err(adapter, drv, netdev, 880 "failed to alloc strings_buf\n"); 881 return; 882 } 883 884 data_buf = devm_kcalloc(&adapter->pdev->dev, 885 strings_num, sizeof(u64), 886 GFP_ATOMIC); 887 if (!data_buf) { 888 netif_err(adapter, drv, netdev, 889 "failed to allocate data buf\n"); 890 devm_kfree(&adapter->pdev->dev, strings_buf); 891 return; 892 } 893 894 ena_get_strings(netdev, ETH_SS_STATS, strings_buf); 895 ena_get_ethtool_stats(netdev, NULL, data_buf); 896 897 /* If there is a buffer, dump stats, otherwise print them to dmesg */ 898 if (buf) 899 for (i = 0; i < strings_num; i++) { 900 rc = snprintf(buf, ETH_GSTRING_LEN + sizeof(u64), 901 "%s %llu\n", 902 strings_buf + i * ETH_GSTRING_LEN, 903 data_buf[i]); 904 buf += rc; 905 } 906 else 907 for (i = 0; i < strings_num; i++) 908 netif_err(adapter, drv, netdev, "%s: %llu\n", 909 strings_buf + i * ETH_GSTRING_LEN, 910 data_buf[i]); 911 912 devm_kfree(&adapter->pdev->dev, strings_buf); 913 devm_kfree(&adapter->pdev->dev, data_buf); 914 } 915 916 void ena_dump_stats_to_buf(struct ena_adapter *adapter, u8 *buf) 917 { 918 if (!buf) 919 return; 920 921 ena_dump_stats_ex(adapter, buf); 922 } 923 924 void ena_dump_stats_to_dmesg(struct ena_adapter *adapter) 925 { 926 ena_dump_stats_ex(adapter, NULL); 927 } 928