1 // SPDX-License-Identifier: GPL-2.0-only 2 /* 10G controller driver for Samsung SoCs 3 * 4 * Copyright (C) 2013 Samsung Electronics Co., Ltd. 5 * http://www.samsung.com 6 * 7 * Author: Siva Reddy Kallam <siva.kallam@samsung.com> 8 */ 9 10 #define pr_fmt(fmt) KBUILD_MODNAME ": " fmt 11 12 #include <linux/clk.h> 13 #include <linux/interrupt.h> 14 #include <linux/kernel.h> 15 #include <linux/netdevice.h> 16 #include <linux/net_tstamp.h> 17 #include <linux/phy.h> 18 #include <linux/ptp_clock_kernel.h> 19 20 #include "sxgbe_common.h" 21 #include "sxgbe_reg.h" 22 #include "sxgbe_dma.h" 23 24 struct sxgbe_stats { 25 char stat_string[ETH_GSTRING_LEN]; 26 int sizeof_stat; 27 int stat_offset; 28 }; 29 30 #define SXGBE_STAT(m) \ 31 { \ 32 #m, \ 33 sizeof_field(struct sxgbe_extra_stats, m), \ 34 offsetof(struct sxgbe_priv_data, xstats.m) \ 35 } 36 37 static const struct sxgbe_stats sxgbe_gstrings_stats[] = { 38 /* TX/RX IRQ events */ 39 SXGBE_STAT(tx_process_stopped_irq), 40 SXGBE_STAT(tx_ctxt_desc_err), 41 SXGBE_STAT(tx_threshold), 42 SXGBE_STAT(rx_threshold), 43 SXGBE_STAT(tx_pkt_n), 44 SXGBE_STAT(rx_pkt_n), 45 SXGBE_STAT(normal_irq_n), 46 SXGBE_STAT(tx_normal_irq_n), 47 SXGBE_STAT(rx_normal_irq_n), 48 SXGBE_STAT(napi_poll), 49 SXGBE_STAT(tx_clean), 50 SXGBE_STAT(tx_reset_ic_bit), 51 SXGBE_STAT(rx_process_stopped_irq), 52 SXGBE_STAT(rx_underflow_irq), 53 54 /* Bus access errors */ 55 SXGBE_STAT(fatal_bus_error_irq), 56 SXGBE_STAT(tx_read_transfer_err), 57 SXGBE_STAT(tx_write_transfer_err), 58 SXGBE_STAT(tx_desc_access_err), 59 SXGBE_STAT(tx_buffer_access_err), 60 SXGBE_STAT(tx_data_transfer_err), 61 SXGBE_STAT(rx_read_transfer_err), 62 SXGBE_STAT(rx_write_transfer_err), 63 SXGBE_STAT(rx_desc_access_err), 64 SXGBE_STAT(rx_buffer_access_err), 65 SXGBE_STAT(rx_data_transfer_err), 66 67 /* EEE-LPI stats */ 68 SXGBE_STAT(tx_lpi_entry_n), 69 SXGBE_STAT(tx_lpi_exit_n), 70 SXGBE_STAT(rx_lpi_entry_n), 71 SXGBE_STAT(rx_lpi_exit_n), 72 SXGBE_STAT(eee_wakeup_error_n), 73 74 /* RX specific */ 75 /* L2 error */ 76 SXGBE_STAT(rx_code_gmii_err), 77 SXGBE_STAT(rx_watchdog_err), 78 SXGBE_STAT(rx_crc_err), 79 SXGBE_STAT(rx_gaint_pkt_err), 80 SXGBE_STAT(ip_hdr_err), 81 SXGBE_STAT(ip_payload_err), 82 SXGBE_STAT(overflow_error), 83 84 /* L2 Pkt type */ 85 SXGBE_STAT(len_pkt), 86 SXGBE_STAT(mac_ctl_pkt), 87 SXGBE_STAT(dcb_ctl_pkt), 88 SXGBE_STAT(arp_pkt), 89 SXGBE_STAT(oam_pkt), 90 SXGBE_STAT(untag_okt), 91 SXGBE_STAT(other_pkt), 92 SXGBE_STAT(svlan_tag_pkt), 93 SXGBE_STAT(cvlan_tag_pkt), 94 SXGBE_STAT(dvlan_ocvlan_icvlan_pkt), 95 SXGBE_STAT(dvlan_osvlan_isvlan_pkt), 96 SXGBE_STAT(dvlan_osvlan_icvlan_pkt), 97 SXGBE_STAT(dvan_ocvlan_icvlan_pkt), 98 99 /* L3/L4 Pkt type */ 100 SXGBE_STAT(not_ip_pkt), 101 SXGBE_STAT(ip4_tcp_pkt), 102 SXGBE_STAT(ip4_udp_pkt), 103 SXGBE_STAT(ip4_icmp_pkt), 104 SXGBE_STAT(ip4_unknown_pkt), 105 SXGBE_STAT(ip6_tcp_pkt), 106 SXGBE_STAT(ip6_udp_pkt), 107 SXGBE_STAT(ip6_icmp_pkt), 108 SXGBE_STAT(ip6_unknown_pkt), 109 110 /* Filter specific */ 111 SXGBE_STAT(vlan_filter_match), 112 SXGBE_STAT(sa_filter_fail), 113 SXGBE_STAT(da_filter_fail), 114 SXGBE_STAT(hash_filter_pass), 115 SXGBE_STAT(l3_filter_match), 116 SXGBE_STAT(l4_filter_match), 117 118 /* RX context specific */ 119 SXGBE_STAT(timestamp_dropped), 120 SXGBE_STAT(rx_msg_type_no_ptp), 121 SXGBE_STAT(rx_ptp_type_sync), 122 SXGBE_STAT(rx_ptp_type_follow_up), 123 SXGBE_STAT(rx_ptp_type_delay_req), 124 SXGBE_STAT(rx_ptp_type_delay_resp), 125 SXGBE_STAT(rx_ptp_type_pdelay_req), 126 SXGBE_STAT(rx_ptp_type_pdelay_resp), 127 SXGBE_STAT(rx_ptp_type_pdelay_follow_up), 128 SXGBE_STAT(rx_ptp_announce), 129 SXGBE_STAT(rx_ptp_mgmt), 130 SXGBE_STAT(rx_ptp_signal), 131 SXGBE_STAT(rx_ptp_resv_msg_type), 132 }; 133 #define SXGBE_STATS_LEN ARRAY_SIZE(sxgbe_gstrings_stats) 134 135 static int sxgbe_get_eee(struct net_device *dev, 136 struct ethtool_eee *edata) 137 { 138 struct sxgbe_priv_data *priv = netdev_priv(dev); 139 140 if (!priv->hw_cap.eee) 141 return -EOPNOTSUPP; 142 143 edata->eee_enabled = priv->eee_enabled; 144 edata->eee_active = priv->eee_active; 145 edata->tx_lpi_timer = priv->tx_lpi_timer; 146 147 return phy_ethtool_get_eee(dev->phydev, edata); 148 } 149 150 static int sxgbe_set_eee(struct net_device *dev, 151 struct ethtool_eee *edata) 152 { 153 struct sxgbe_priv_data *priv = netdev_priv(dev); 154 155 priv->eee_enabled = edata->eee_enabled; 156 157 if (!priv->eee_enabled) { 158 sxgbe_disable_eee_mode(priv); 159 } else { 160 /* We are asking for enabling the EEE but it is safe 161 * to verify all by invoking the eee_init function. 162 * In case of failure it will return an error. 163 */ 164 priv->eee_enabled = sxgbe_eee_init(priv); 165 if (!priv->eee_enabled) 166 return -EOPNOTSUPP; 167 168 /* Do not change tx_lpi_timer in case of failure */ 169 priv->tx_lpi_timer = edata->tx_lpi_timer; 170 } 171 172 return phy_ethtool_set_eee(dev->phydev, edata); 173 } 174 175 static void sxgbe_getdrvinfo(struct net_device *dev, 176 struct ethtool_drvinfo *info) 177 { 178 strlcpy(info->driver, KBUILD_MODNAME, sizeof(info->driver)); 179 strlcpy(info->version, DRV_VERSION, sizeof(info->version)); 180 } 181 182 static u32 sxgbe_getmsglevel(struct net_device *dev) 183 { 184 struct sxgbe_priv_data *priv = netdev_priv(dev); 185 return priv->msg_enable; 186 } 187 188 static void sxgbe_setmsglevel(struct net_device *dev, u32 level) 189 { 190 struct sxgbe_priv_data *priv = netdev_priv(dev); 191 priv->msg_enable = level; 192 } 193 194 static void sxgbe_get_strings(struct net_device *dev, u32 stringset, u8 *data) 195 { 196 int i; 197 u8 *p = data; 198 199 switch (stringset) { 200 case ETH_SS_STATS: 201 for (i = 0; i < SXGBE_STATS_LEN; i++) { 202 memcpy(p, sxgbe_gstrings_stats[i].stat_string, 203 ETH_GSTRING_LEN); 204 p += ETH_GSTRING_LEN; 205 } 206 break; 207 default: 208 WARN_ON(1); 209 break; 210 } 211 } 212 213 static int sxgbe_get_sset_count(struct net_device *netdev, int sset) 214 { 215 int len; 216 217 switch (sset) { 218 case ETH_SS_STATS: 219 len = SXGBE_STATS_LEN; 220 return len; 221 default: 222 return -EINVAL; 223 } 224 } 225 226 static void sxgbe_get_ethtool_stats(struct net_device *dev, 227 struct ethtool_stats *dummy, u64 *data) 228 { 229 struct sxgbe_priv_data *priv = netdev_priv(dev); 230 int i; 231 char *p; 232 233 if (priv->eee_enabled) { 234 int val = phy_get_eee_err(dev->phydev); 235 236 if (val) 237 priv->xstats.eee_wakeup_error_n = val; 238 } 239 240 for (i = 0; i < SXGBE_STATS_LEN; i++) { 241 p = (char *)priv + sxgbe_gstrings_stats[i].stat_offset; 242 data[i] = (sxgbe_gstrings_stats[i].sizeof_stat == sizeof(u64)) 243 ? (*(u64 *)p) : (*(u32 *)p); 244 } 245 } 246 247 static void sxgbe_get_channels(struct net_device *dev, 248 struct ethtool_channels *channel) 249 { 250 channel->max_rx = SXGBE_MAX_RX_CHANNELS; 251 channel->max_tx = SXGBE_MAX_TX_CHANNELS; 252 channel->rx_count = SXGBE_RX_QUEUES; 253 channel->tx_count = SXGBE_TX_QUEUES; 254 } 255 256 static u32 sxgbe_riwt2usec(u32 riwt, struct sxgbe_priv_data *priv) 257 { 258 unsigned long clk = clk_get_rate(priv->sxgbe_clk); 259 260 if (!clk) 261 return 0; 262 263 return (riwt * 256) / (clk / 1000000); 264 } 265 266 static u32 sxgbe_usec2riwt(u32 usec, struct sxgbe_priv_data *priv) 267 { 268 unsigned long clk = clk_get_rate(priv->sxgbe_clk); 269 270 if (!clk) 271 return 0; 272 273 return (usec * (clk / 1000000)) / 256; 274 } 275 276 static int sxgbe_get_coalesce(struct net_device *dev, 277 struct ethtool_coalesce *ec) 278 { 279 struct sxgbe_priv_data *priv = netdev_priv(dev); 280 281 if (priv->use_riwt) 282 ec->rx_coalesce_usecs = sxgbe_riwt2usec(priv->rx_riwt, priv); 283 284 return 0; 285 } 286 287 static int sxgbe_set_coalesce(struct net_device *dev, 288 struct ethtool_coalesce *ec) 289 { 290 struct sxgbe_priv_data *priv = netdev_priv(dev); 291 unsigned int rx_riwt; 292 293 if (!ec->rx_coalesce_usecs) 294 return -EINVAL; 295 296 rx_riwt = sxgbe_usec2riwt(ec->rx_coalesce_usecs, priv); 297 298 if ((rx_riwt > SXGBE_MAX_DMA_RIWT) || (rx_riwt < SXGBE_MIN_DMA_RIWT)) 299 return -EINVAL; 300 else if (!priv->use_riwt) 301 return -EOPNOTSUPP; 302 303 priv->rx_riwt = rx_riwt; 304 priv->hw->dma->rx_watchdog(priv->ioaddr, priv->rx_riwt); 305 306 return 0; 307 } 308 309 static int sxgbe_get_rss_hash_opts(struct sxgbe_priv_data *priv, 310 struct ethtool_rxnfc *cmd) 311 { 312 cmd->data = 0; 313 314 /* Report default options for RSS on sxgbe */ 315 switch (cmd->flow_type) { 316 case TCP_V4_FLOW: 317 case UDP_V4_FLOW: 318 cmd->data |= RXH_L4_B_0_1 | RXH_L4_B_2_3; 319 fallthrough; 320 case SCTP_V4_FLOW: 321 case AH_ESP_V4_FLOW: 322 case AH_V4_FLOW: 323 case ESP_V4_FLOW: 324 case IPV4_FLOW: 325 cmd->data |= RXH_IP_SRC | RXH_IP_DST; 326 break; 327 case TCP_V6_FLOW: 328 case UDP_V6_FLOW: 329 cmd->data |= RXH_L4_B_0_1 | RXH_L4_B_2_3; 330 fallthrough; 331 case SCTP_V6_FLOW: 332 case AH_ESP_V6_FLOW: 333 case AH_V6_FLOW: 334 case ESP_V6_FLOW: 335 case IPV6_FLOW: 336 cmd->data |= RXH_IP_SRC | RXH_IP_DST; 337 break; 338 default: 339 return -EINVAL; 340 } 341 342 return 0; 343 } 344 345 static int sxgbe_get_rxnfc(struct net_device *dev, struct ethtool_rxnfc *cmd, 346 u32 *rule_locs) 347 { 348 struct sxgbe_priv_data *priv = netdev_priv(dev); 349 int ret = -EOPNOTSUPP; 350 351 switch (cmd->cmd) { 352 case ETHTOOL_GRXFH: 353 ret = sxgbe_get_rss_hash_opts(priv, cmd); 354 break; 355 default: 356 break; 357 } 358 359 return ret; 360 } 361 362 static int sxgbe_set_rss_hash_opt(struct sxgbe_priv_data *priv, 363 struct ethtool_rxnfc *cmd) 364 { 365 u32 reg_val = 0; 366 367 /* RSS does not support anything other than hashing 368 * to queues on src and dst IPs and ports 369 */ 370 if (cmd->data & ~(RXH_IP_SRC | RXH_IP_DST | 371 RXH_L4_B_0_1 | RXH_L4_B_2_3)) 372 return -EINVAL; 373 374 switch (cmd->flow_type) { 375 case TCP_V4_FLOW: 376 case TCP_V6_FLOW: 377 if (!(cmd->data & RXH_IP_SRC) || 378 !(cmd->data & RXH_IP_DST) || 379 !(cmd->data & RXH_L4_B_0_1) || 380 !(cmd->data & RXH_L4_B_2_3)) 381 return -EINVAL; 382 reg_val = SXGBE_CORE_RSS_CTL_TCP4TE; 383 break; 384 case UDP_V4_FLOW: 385 case UDP_V6_FLOW: 386 if (!(cmd->data & RXH_IP_SRC) || 387 !(cmd->data & RXH_IP_DST) || 388 !(cmd->data & RXH_L4_B_0_1) || 389 !(cmd->data & RXH_L4_B_2_3)) 390 return -EINVAL; 391 reg_val = SXGBE_CORE_RSS_CTL_UDP4TE; 392 break; 393 case SCTP_V4_FLOW: 394 case AH_ESP_V4_FLOW: 395 case AH_V4_FLOW: 396 case ESP_V4_FLOW: 397 case AH_ESP_V6_FLOW: 398 case AH_V6_FLOW: 399 case ESP_V6_FLOW: 400 case SCTP_V6_FLOW: 401 case IPV4_FLOW: 402 case IPV6_FLOW: 403 if (!(cmd->data & RXH_IP_SRC) || 404 !(cmd->data & RXH_IP_DST) || 405 (cmd->data & RXH_L4_B_0_1) || 406 (cmd->data & RXH_L4_B_2_3)) 407 return -EINVAL; 408 reg_val = SXGBE_CORE_RSS_CTL_IP2TE; 409 break; 410 default: 411 return -EINVAL; 412 } 413 414 /* Read SXGBE RSS control register and update */ 415 reg_val |= readl(priv->ioaddr + SXGBE_CORE_RSS_CTL_REG); 416 writel(reg_val, priv->ioaddr + SXGBE_CORE_RSS_CTL_REG); 417 readl(priv->ioaddr + SXGBE_CORE_RSS_CTL_REG); 418 419 return 0; 420 } 421 422 static int sxgbe_set_rxnfc(struct net_device *dev, struct ethtool_rxnfc *cmd) 423 { 424 struct sxgbe_priv_data *priv = netdev_priv(dev); 425 int ret = -EOPNOTSUPP; 426 427 switch (cmd->cmd) { 428 case ETHTOOL_SRXFH: 429 ret = sxgbe_set_rss_hash_opt(priv, cmd); 430 break; 431 default: 432 break; 433 } 434 435 return ret; 436 } 437 438 static void sxgbe_get_regs(struct net_device *dev, 439 struct ethtool_regs *regs, void *space) 440 { 441 struct sxgbe_priv_data *priv = netdev_priv(dev); 442 u32 *reg_space = (u32 *)space; 443 int reg_offset; 444 int reg_ix = 0; 445 void __iomem *ioaddr = priv->ioaddr; 446 447 memset(reg_space, 0x0, REG_SPACE_SIZE); 448 449 /* MAC registers */ 450 for (reg_offset = START_MAC_REG_OFFSET; 451 reg_offset <= MAX_MAC_REG_OFFSET; reg_offset += 4) { 452 reg_space[reg_ix] = readl(ioaddr + reg_offset); 453 reg_ix++; 454 } 455 456 /* MTL registers */ 457 for (reg_offset = START_MTL_REG_OFFSET; 458 reg_offset <= MAX_MTL_REG_OFFSET; reg_offset += 4) { 459 reg_space[reg_ix] = readl(ioaddr + reg_offset); 460 reg_ix++; 461 } 462 463 /* DMA registers */ 464 for (reg_offset = START_DMA_REG_OFFSET; 465 reg_offset <= MAX_DMA_REG_OFFSET; reg_offset += 4) { 466 reg_space[reg_ix] = readl(ioaddr + reg_offset); 467 reg_ix++; 468 } 469 470 BUG_ON(reg_ix * 4 > REG_SPACE_SIZE); 471 } 472 473 static int sxgbe_get_regs_len(struct net_device *dev) 474 { 475 return REG_SPACE_SIZE; 476 } 477 478 static const struct ethtool_ops sxgbe_ethtool_ops = { 479 .supported_coalesce_params = ETHTOOL_COALESCE_RX_USECS, 480 .get_drvinfo = sxgbe_getdrvinfo, 481 .get_msglevel = sxgbe_getmsglevel, 482 .set_msglevel = sxgbe_setmsglevel, 483 .get_link = ethtool_op_get_link, 484 .get_strings = sxgbe_get_strings, 485 .get_ethtool_stats = sxgbe_get_ethtool_stats, 486 .get_sset_count = sxgbe_get_sset_count, 487 .get_channels = sxgbe_get_channels, 488 .get_coalesce = sxgbe_get_coalesce, 489 .set_coalesce = sxgbe_set_coalesce, 490 .get_rxnfc = sxgbe_get_rxnfc, 491 .set_rxnfc = sxgbe_set_rxnfc, 492 .get_regs = sxgbe_get_regs, 493 .get_regs_len = sxgbe_get_regs_len, 494 .get_eee = sxgbe_get_eee, 495 .set_eee = sxgbe_set_eee, 496 .get_link_ksettings = phy_ethtool_get_link_ksettings, 497 .set_link_ksettings = phy_ethtool_set_link_ksettings, 498 }; 499 500 void sxgbe_set_ethtool_ops(struct net_device *netdev) 501 { 502 netdev->ethtool_ops = &sxgbe_ethtool_ops; 503 } 504