1 /* Copyright 2008-2016 Freescale Semiconductor, Inc. 2 * 3 * Redistribution and use in source and binary forms, with or without 4 * modification, are permitted provided that the following conditions are met: 5 * * Redistributions of source code must retain the above copyright 6 * notice, this list of conditions and the following disclaimer. 7 * * Redistributions in binary form must reproduce the above copyright 8 * notice, this list of conditions and the following disclaimer in the 9 * documentation and/or other materials provided with the distribution. 10 * * Neither the name of Freescale Semiconductor nor the 11 * names of its contributors may be used to endorse or promote products 12 * derived from this software without specific prior written permission. 13 * 14 * 15 * ALTERNATIVELY, this software may be distributed under the terms of the 16 * GNU General Public License ("GPL") as published by the Free Software 17 * Foundation, either version 2 of that License or (at your option) any 18 * later version. 19 * 20 * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY 21 * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED 22 * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE 23 * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY 24 * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES 25 * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; 26 * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND 27 * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT 28 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS 29 * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. 30 */ 31 32 #define pr_fmt(fmt) KBUILD_MODNAME ": " fmt 33 34 #include <linux/string.h> 35 #include <linux/of_platform.h> 36 #include <linux/net_tstamp.h> 37 #include <linux/fsl/ptp_qoriq.h> 38 39 #include "dpaa_eth.h" 40 #include "mac.h" 41 42 static const char dpaa_stats_percpu[][ETH_GSTRING_LEN] = { 43 "interrupts", 44 "rx packets", 45 "tx packets", 46 "tx confirm", 47 "tx S/G", 48 "tx error", 49 "rx error", 50 }; 51 52 static char dpaa_stats_global[][ETH_GSTRING_LEN] = { 53 /* dpa rx errors */ 54 "rx dma error", 55 "rx frame physical error", 56 "rx frame size error", 57 "rx header error", 58 59 /* demultiplexing errors */ 60 "qman cg_tdrop", 61 "qman wred", 62 "qman error cond", 63 "qman early window", 64 "qman late window", 65 "qman fq tdrop", 66 "qman fq retired", 67 "qman orp disabled", 68 69 /* congestion related stats */ 70 "congestion time (ms)", 71 "entered congestion", 72 "congested (0/1)" 73 }; 74 75 #define DPAA_STATS_PERCPU_LEN ARRAY_SIZE(dpaa_stats_percpu) 76 #define DPAA_STATS_GLOBAL_LEN ARRAY_SIZE(dpaa_stats_global) 77 78 static int dpaa_get_link_ksettings(struct net_device *net_dev, 79 struct ethtool_link_ksettings *cmd) 80 { 81 if (!net_dev->phydev) { 82 netdev_dbg(net_dev, "phy device not initialized\n"); 83 return 0; 84 } 85 86 phy_ethtool_ksettings_get(net_dev->phydev, cmd); 87 88 return 0; 89 } 90 91 static int dpaa_set_link_ksettings(struct net_device *net_dev, 92 const struct ethtool_link_ksettings *cmd) 93 { 94 int err; 95 96 if (!net_dev->phydev) { 97 netdev_err(net_dev, "phy device not initialized\n"); 98 return -ENODEV; 99 } 100 101 err = phy_ethtool_ksettings_set(net_dev->phydev, cmd); 102 if (err < 0) 103 netdev_err(net_dev, "phy_ethtool_ksettings_set() = %d\n", err); 104 105 return err; 106 } 107 108 static void dpaa_get_drvinfo(struct net_device *net_dev, 109 struct ethtool_drvinfo *drvinfo) 110 { 111 int len; 112 113 strlcpy(drvinfo->driver, KBUILD_MODNAME, 114 sizeof(drvinfo->driver)); 115 len = snprintf(drvinfo->version, sizeof(drvinfo->version), 116 "%X", 0); 117 len = snprintf(drvinfo->fw_version, sizeof(drvinfo->fw_version), 118 "%X", 0); 119 120 if (len >= sizeof(drvinfo->fw_version)) { 121 /* Truncated output */ 122 netdev_notice(net_dev, "snprintf() = %d\n", len); 123 } 124 strlcpy(drvinfo->bus_info, dev_name(net_dev->dev.parent->parent), 125 sizeof(drvinfo->bus_info)); 126 } 127 128 static u32 dpaa_get_msglevel(struct net_device *net_dev) 129 { 130 return ((struct dpaa_priv *)netdev_priv(net_dev))->msg_enable; 131 } 132 133 static void dpaa_set_msglevel(struct net_device *net_dev, 134 u32 msg_enable) 135 { 136 ((struct dpaa_priv *)netdev_priv(net_dev))->msg_enable = msg_enable; 137 } 138 139 static int dpaa_nway_reset(struct net_device *net_dev) 140 { 141 int err; 142 143 if (!net_dev->phydev) { 144 netdev_err(net_dev, "phy device not initialized\n"); 145 return -ENODEV; 146 } 147 148 err = 0; 149 if (net_dev->phydev->autoneg) { 150 err = phy_start_aneg(net_dev->phydev); 151 if (err < 0) 152 netdev_err(net_dev, "phy_start_aneg() = %d\n", 153 err); 154 } 155 156 return err; 157 } 158 159 static void dpaa_get_pauseparam(struct net_device *net_dev, 160 struct ethtool_pauseparam *epause) 161 { 162 struct mac_device *mac_dev; 163 struct dpaa_priv *priv; 164 165 priv = netdev_priv(net_dev); 166 mac_dev = priv->mac_dev; 167 168 if (!net_dev->phydev) { 169 netdev_err(net_dev, "phy device not initialized\n"); 170 return; 171 } 172 173 epause->autoneg = mac_dev->autoneg_pause; 174 epause->rx_pause = mac_dev->rx_pause_active; 175 epause->tx_pause = mac_dev->tx_pause_active; 176 } 177 178 static int dpaa_set_pauseparam(struct net_device *net_dev, 179 struct ethtool_pauseparam *epause) 180 { 181 struct mac_device *mac_dev; 182 struct phy_device *phydev; 183 bool rx_pause, tx_pause; 184 struct dpaa_priv *priv; 185 u32 newadv, oldadv; 186 int err; 187 188 priv = netdev_priv(net_dev); 189 mac_dev = priv->mac_dev; 190 191 phydev = net_dev->phydev; 192 if (!phydev) { 193 netdev_err(net_dev, "phy device not initialized\n"); 194 return -ENODEV; 195 } 196 197 if (!phy_validate_pause(phydev, epause)) 198 return -EINVAL; 199 200 /* The MAC should know how to handle PAUSE frame autonegotiation before 201 * adjust_link is triggered by a forced renegotiation of sym/asym PAUSE 202 * settings. 203 */ 204 mac_dev->autoneg_pause = !!epause->autoneg; 205 mac_dev->rx_pause_req = !!epause->rx_pause; 206 mac_dev->tx_pause_req = !!epause->tx_pause; 207 208 /* Determine the sym/asym advertised PAUSE capabilities from the desired 209 * rx/tx pause settings. 210 */ 211 212 phy_set_asym_pause(phydev, epause->rx_pause, epause->tx_pause); 213 214 fman_get_pause_cfg(mac_dev, &rx_pause, &tx_pause); 215 err = fman_set_mac_active_pause(mac_dev, rx_pause, tx_pause); 216 if (err < 0) 217 netdev_err(net_dev, "set_mac_active_pause() = %d\n", err); 218 219 return err; 220 } 221 222 static int dpaa_get_sset_count(struct net_device *net_dev, int type) 223 { 224 unsigned int total_stats, num_stats; 225 226 num_stats = num_online_cpus() + 1; 227 total_stats = num_stats * (DPAA_STATS_PERCPU_LEN + DPAA_BPS_NUM) + 228 DPAA_STATS_GLOBAL_LEN; 229 230 switch (type) { 231 case ETH_SS_STATS: 232 return total_stats; 233 default: 234 return -EOPNOTSUPP; 235 } 236 } 237 238 static void copy_stats(struct dpaa_percpu_priv *percpu_priv, int num_cpus, 239 int crr_cpu, u64 *bp_count, u64 *data) 240 { 241 int num_values = num_cpus + 1; 242 int crr = 0, j; 243 244 /* update current CPU's stats and also add them to the total values */ 245 data[crr * num_values + crr_cpu] = percpu_priv->in_interrupt; 246 data[crr++ * num_values + num_cpus] += percpu_priv->in_interrupt; 247 248 data[crr * num_values + crr_cpu] = percpu_priv->stats.rx_packets; 249 data[crr++ * num_values + num_cpus] += percpu_priv->stats.rx_packets; 250 251 data[crr * num_values + crr_cpu] = percpu_priv->stats.tx_packets; 252 data[crr++ * num_values + num_cpus] += percpu_priv->stats.tx_packets; 253 254 data[crr * num_values + crr_cpu] = percpu_priv->tx_confirm; 255 data[crr++ * num_values + num_cpus] += percpu_priv->tx_confirm; 256 257 data[crr * num_values + crr_cpu] = percpu_priv->tx_frag_skbuffs; 258 data[crr++ * num_values + num_cpus] += percpu_priv->tx_frag_skbuffs; 259 260 data[crr * num_values + crr_cpu] = percpu_priv->stats.tx_errors; 261 data[crr++ * num_values + num_cpus] += percpu_priv->stats.tx_errors; 262 263 data[crr * num_values + crr_cpu] = percpu_priv->stats.rx_errors; 264 data[crr++ * num_values + num_cpus] += percpu_priv->stats.rx_errors; 265 266 for (j = 0; j < DPAA_BPS_NUM; j++) { 267 data[crr * num_values + crr_cpu] = bp_count[j]; 268 data[crr++ * num_values + num_cpus] += bp_count[j]; 269 } 270 } 271 272 static void dpaa_get_ethtool_stats(struct net_device *net_dev, 273 struct ethtool_stats *stats, u64 *data) 274 { 275 u64 bp_count[DPAA_BPS_NUM], cg_time, cg_num; 276 struct dpaa_percpu_priv *percpu_priv; 277 struct dpaa_rx_errors rx_errors; 278 unsigned int num_cpus, offset; 279 struct dpaa_ern_cnt ern_cnt; 280 struct dpaa_bp *dpaa_bp; 281 struct dpaa_priv *priv; 282 int total_stats, i, j; 283 bool cg_status; 284 285 total_stats = dpaa_get_sset_count(net_dev, ETH_SS_STATS); 286 priv = netdev_priv(net_dev); 287 num_cpus = num_online_cpus(); 288 289 memset(&bp_count, 0, sizeof(bp_count)); 290 memset(&rx_errors, 0, sizeof(struct dpaa_rx_errors)); 291 memset(&ern_cnt, 0, sizeof(struct dpaa_ern_cnt)); 292 memset(data, 0, total_stats * sizeof(u64)); 293 294 for_each_online_cpu(i) { 295 percpu_priv = per_cpu_ptr(priv->percpu_priv, i); 296 for (j = 0; j < DPAA_BPS_NUM; j++) { 297 dpaa_bp = priv->dpaa_bps[j]; 298 if (!dpaa_bp->percpu_count) 299 continue; 300 bp_count[j] = *(per_cpu_ptr(dpaa_bp->percpu_count, i)); 301 } 302 rx_errors.dme += percpu_priv->rx_errors.dme; 303 rx_errors.fpe += percpu_priv->rx_errors.fpe; 304 rx_errors.fse += percpu_priv->rx_errors.fse; 305 rx_errors.phe += percpu_priv->rx_errors.phe; 306 307 ern_cnt.cg_tdrop += percpu_priv->ern_cnt.cg_tdrop; 308 ern_cnt.wred += percpu_priv->ern_cnt.wred; 309 ern_cnt.err_cond += percpu_priv->ern_cnt.err_cond; 310 ern_cnt.early_window += percpu_priv->ern_cnt.early_window; 311 ern_cnt.late_window += percpu_priv->ern_cnt.late_window; 312 ern_cnt.fq_tdrop += percpu_priv->ern_cnt.fq_tdrop; 313 ern_cnt.fq_retired += percpu_priv->ern_cnt.fq_retired; 314 ern_cnt.orp_zero += percpu_priv->ern_cnt.orp_zero; 315 316 copy_stats(percpu_priv, num_cpus, i, bp_count, data); 317 } 318 319 offset = (num_cpus + 1) * (DPAA_STATS_PERCPU_LEN + DPAA_BPS_NUM); 320 memcpy(data + offset, &rx_errors, sizeof(struct dpaa_rx_errors)); 321 322 offset += sizeof(struct dpaa_rx_errors) / sizeof(u64); 323 memcpy(data + offset, &ern_cnt, sizeof(struct dpaa_ern_cnt)); 324 325 /* gather congestion related counters */ 326 cg_num = 0; 327 cg_status = false; 328 cg_time = jiffies_to_msecs(priv->cgr_data.congested_jiffies); 329 if (qman_query_cgr_congested(&priv->cgr_data.cgr, &cg_status) == 0) { 330 cg_num = priv->cgr_data.cgr_congested_count; 331 332 /* reset congestion stats (like QMan API does */ 333 priv->cgr_data.congested_jiffies = 0; 334 priv->cgr_data.cgr_congested_count = 0; 335 } 336 337 offset += sizeof(struct dpaa_ern_cnt) / sizeof(u64); 338 data[offset++] = cg_time; 339 data[offset++] = cg_num; 340 data[offset++] = cg_status; 341 } 342 343 static void dpaa_get_strings(struct net_device *net_dev, u32 stringset, 344 u8 *data) 345 { 346 unsigned int i, j, num_cpus, size; 347 char string_cpu[ETH_GSTRING_LEN]; 348 u8 *strings; 349 350 memset(string_cpu, 0, sizeof(string_cpu)); 351 strings = data; 352 num_cpus = num_online_cpus(); 353 size = DPAA_STATS_GLOBAL_LEN * ETH_GSTRING_LEN; 354 355 for (i = 0; i < DPAA_STATS_PERCPU_LEN; i++) { 356 for (j = 0; j < num_cpus; j++) { 357 snprintf(string_cpu, ETH_GSTRING_LEN, "%s [CPU %d]", 358 dpaa_stats_percpu[i], j); 359 memcpy(strings, string_cpu, ETH_GSTRING_LEN); 360 strings += ETH_GSTRING_LEN; 361 } 362 snprintf(string_cpu, ETH_GSTRING_LEN, "%s [TOTAL]", 363 dpaa_stats_percpu[i]); 364 memcpy(strings, string_cpu, ETH_GSTRING_LEN); 365 strings += ETH_GSTRING_LEN; 366 } 367 for (i = 0; i < DPAA_BPS_NUM; i++) { 368 for (j = 0; j < num_cpus; j++) { 369 snprintf(string_cpu, ETH_GSTRING_LEN, 370 "bpool %c [CPU %d]", 'a' + i, j); 371 memcpy(strings, string_cpu, ETH_GSTRING_LEN); 372 strings += ETH_GSTRING_LEN; 373 } 374 snprintf(string_cpu, ETH_GSTRING_LEN, "bpool %c [TOTAL]", 375 'a' + i); 376 memcpy(strings, string_cpu, ETH_GSTRING_LEN); 377 strings += ETH_GSTRING_LEN; 378 } 379 memcpy(strings, dpaa_stats_global, size); 380 } 381 382 static int dpaa_get_hash_opts(struct net_device *dev, 383 struct ethtool_rxnfc *cmd) 384 { 385 struct dpaa_priv *priv = netdev_priv(dev); 386 387 cmd->data = 0; 388 389 switch (cmd->flow_type) { 390 case TCP_V4_FLOW: 391 case TCP_V6_FLOW: 392 case UDP_V4_FLOW: 393 case UDP_V6_FLOW: 394 if (priv->keygen_in_use) 395 cmd->data |= RXH_L4_B_0_1 | RXH_L4_B_2_3; 396 /* Fall through */ 397 case IPV4_FLOW: 398 case IPV6_FLOW: 399 case SCTP_V4_FLOW: 400 case SCTP_V6_FLOW: 401 case AH_ESP_V4_FLOW: 402 case AH_ESP_V6_FLOW: 403 case AH_V4_FLOW: 404 case AH_V6_FLOW: 405 case ESP_V4_FLOW: 406 case ESP_V6_FLOW: 407 if (priv->keygen_in_use) 408 cmd->data |= RXH_IP_SRC | RXH_IP_DST; 409 break; 410 default: 411 cmd->data = 0; 412 break; 413 } 414 415 return 0; 416 } 417 418 static int dpaa_get_rxnfc(struct net_device *dev, struct ethtool_rxnfc *cmd, 419 u32 *unused) 420 { 421 int ret = -EOPNOTSUPP; 422 423 switch (cmd->cmd) { 424 case ETHTOOL_GRXFH: 425 ret = dpaa_get_hash_opts(dev, cmd); 426 break; 427 default: 428 break; 429 } 430 431 return ret; 432 } 433 434 static void dpaa_set_hash(struct net_device *net_dev, bool enable) 435 { 436 struct mac_device *mac_dev; 437 struct fman_port *rxport; 438 struct dpaa_priv *priv; 439 440 priv = netdev_priv(net_dev); 441 mac_dev = priv->mac_dev; 442 rxport = mac_dev->port[0]; 443 444 fman_port_use_kg_hash(rxport, enable); 445 priv->keygen_in_use = enable; 446 } 447 448 static int dpaa_set_hash_opts(struct net_device *dev, 449 struct ethtool_rxnfc *nfc) 450 { 451 int ret = -EINVAL; 452 453 /* we support hashing on IPv4/v6 src/dest IP and L4 src/dest port */ 454 if (nfc->data & 455 ~(RXH_IP_SRC | RXH_IP_DST | RXH_L4_B_0_1 | RXH_L4_B_2_3)) 456 return -EINVAL; 457 458 switch (nfc->flow_type) { 459 case TCP_V4_FLOW: 460 case TCP_V6_FLOW: 461 case UDP_V4_FLOW: 462 case UDP_V6_FLOW: 463 case IPV4_FLOW: 464 case IPV6_FLOW: 465 case SCTP_V4_FLOW: 466 case SCTP_V6_FLOW: 467 case AH_ESP_V4_FLOW: 468 case AH_ESP_V6_FLOW: 469 case AH_V4_FLOW: 470 case AH_V6_FLOW: 471 case ESP_V4_FLOW: 472 case ESP_V6_FLOW: 473 dpaa_set_hash(dev, !!nfc->data); 474 ret = 0; 475 break; 476 default: 477 break; 478 } 479 480 return ret; 481 } 482 483 static int dpaa_set_rxnfc(struct net_device *dev, struct ethtool_rxnfc *cmd) 484 { 485 int ret = -EOPNOTSUPP; 486 487 switch (cmd->cmd) { 488 case ETHTOOL_SRXFH: 489 ret = dpaa_set_hash_opts(dev, cmd); 490 break; 491 default: 492 break; 493 } 494 495 return ret; 496 } 497 498 static int dpaa_get_ts_info(struct net_device *net_dev, 499 struct ethtool_ts_info *info) 500 { 501 struct device *dev = net_dev->dev.parent; 502 struct device_node *mac_node = dev->of_node; 503 struct device_node *fman_node = NULL, *ptp_node = NULL; 504 struct platform_device *ptp_dev = NULL; 505 struct qoriq_ptp *ptp = NULL; 506 507 info->phc_index = -1; 508 509 fman_node = of_get_parent(mac_node); 510 if (fman_node) 511 ptp_node = of_parse_phandle(fman_node, "ptimer-handle", 0); 512 513 if (ptp_node) 514 ptp_dev = of_find_device_by_node(ptp_node); 515 516 if (ptp_dev) 517 ptp = platform_get_drvdata(ptp_dev); 518 519 if (ptp) 520 info->phc_index = ptp->phc_index; 521 522 info->so_timestamping = SOF_TIMESTAMPING_TX_HARDWARE | 523 SOF_TIMESTAMPING_RX_HARDWARE | 524 SOF_TIMESTAMPING_RAW_HARDWARE; 525 info->tx_types = (1 << HWTSTAMP_TX_OFF) | 526 (1 << HWTSTAMP_TX_ON); 527 info->rx_filters = (1 << HWTSTAMP_FILTER_NONE) | 528 (1 << HWTSTAMP_FILTER_ALL); 529 530 return 0; 531 } 532 533 const struct ethtool_ops dpaa_ethtool_ops = { 534 .get_drvinfo = dpaa_get_drvinfo, 535 .get_msglevel = dpaa_get_msglevel, 536 .set_msglevel = dpaa_set_msglevel, 537 .nway_reset = dpaa_nway_reset, 538 .get_pauseparam = dpaa_get_pauseparam, 539 .set_pauseparam = dpaa_set_pauseparam, 540 .get_link = ethtool_op_get_link, 541 .get_sset_count = dpaa_get_sset_count, 542 .get_ethtool_stats = dpaa_get_ethtool_stats, 543 .get_strings = dpaa_get_strings, 544 .get_link_ksettings = dpaa_get_link_ksettings, 545 .set_link_ksettings = dpaa_set_link_ksettings, 546 .get_rxnfc = dpaa_get_rxnfc, 547 .set_rxnfc = dpaa_set_rxnfc, 548 .get_ts_info = dpaa_get_ts_info, 549 }; 550