1 /* 2 * drivers/net/ethernet/nxp/lpc_eth.c 3 * 4 * Author: Kevin Wells <kevin.wells@nxp.com> 5 * 6 * Copyright (C) 2010 NXP Semiconductors 7 * Copyright (C) 2012 Roland Stigge <stigge@antcom.de> 8 * 9 * This program is free software; you can redistribute it and/or modify 10 * it under the terms of the GNU General Public License as published by 11 * the Free Software Foundation; either version 2 of the License, or 12 * (at your option) any later version. 13 * 14 * This program is distributed in the hope that it will be useful, 15 * but WITHOUT ANY WARRANTY; without even the implied warranty of 16 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the 17 * GNU General Public License for more details. 18 */ 19 20 #define pr_fmt(fmt) KBUILD_MODNAME ": " fmt 21 22 #include <linux/clk.h> 23 #include <linux/crc32.h> 24 #include <linux/etherdevice.h> 25 #include <linux/module.h> 26 #include <linux/of_net.h> 27 #include <linux/phy.h> 28 #include <linux/platform_device.h> 29 #include <linux/spinlock.h> 30 31 #include <mach/board.h> 32 #include <mach/hardware.h> 33 #include <mach/platform.h> 34 35 #define MODNAME "lpc-eth" 36 #define DRV_VERSION "1.00" 37 38 #define ENET_MAXF_SIZE 1536 39 #define ENET_RX_DESC 48 40 #define ENET_TX_DESC 16 41 42 #define NAPI_WEIGHT 16 43 44 /* 45 * Ethernet MAC controller Register offsets 46 */ 47 #define LPC_ENET_MAC1(x) (x + 0x000) 48 #define LPC_ENET_MAC2(x) (x + 0x004) 49 #define LPC_ENET_IPGT(x) (x + 0x008) 50 #define LPC_ENET_IPGR(x) (x + 0x00C) 51 #define LPC_ENET_CLRT(x) (x + 0x010) 52 #define LPC_ENET_MAXF(x) (x + 0x014) 53 #define LPC_ENET_SUPP(x) (x + 0x018) 54 #define LPC_ENET_TEST(x) (x + 0x01C) 55 #define LPC_ENET_MCFG(x) (x + 0x020) 56 #define LPC_ENET_MCMD(x) (x + 0x024) 57 #define LPC_ENET_MADR(x) (x + 0x028) 58 #define LPC_ENET_MWTD(x) (x + 0x02C) 59 #define LPC_ENET_MRDD(x) (x + 0x030) 60 #define LPC_ENET_MIND(x) (x + 0x034) 61 #define LPC_ENET_SA0(x) (x + 0x040) 62 #define LPC_ENET_SA1(x) (x + 0x044) 63 #define LPC_ENET_SA2(x) (x + 0x048) 64 #define LPC_ENET_COMMAND(x) (x + 0x100) 65 #define LPC_ENET_STATUS(x) (x + 0x104) 66 #define LPC_ENET_RXDESCRIPTOR(x) (x + 0x108) 67 #define LPC_ENET_RXSTATUS(x) (x + 0x10C) 68 #define LPC_ENET_RXDESCRIPTORNUMBER(x) (x + 0x110) 69 #define LPC_ENET_RXPRODUCEINDEX(x) (x + 0x114) 70 #define LPC_ENET_RXCONSUMEINDEX(x) (x + 0x118) 71 #define LPC_ENET_TXDESCRIPTOR(x) (x + 0x11C) 72 #define LPC_ENET_TXSTATUS(x) (x + 0x120) 73 #define LPC_ENET_TXDESCRIPTORNUMBER(x) (x + 0x124) 74 #define LPC_ENET_TXPRODUCEINDEX(x) (x + 0x128) 75 #define LPC_ENET_TXCONSUMEINDEX(x) (x + 0x12C) 76 #define LPC_ENET_TSV0(x) (x + 0x158) 77 #define LPC_ENET_TSV1(x) (x + 0x15C) 78 #define LPC_ENET_RSV(x) (x + 0x160) 79 #define LPC_ENET_FLOWCONTROLCOUNTER(x) (x + 0x170) 80 #define LPC_ENET_FLOWCONTROLSTATUS(x) (x + 0x174) 81 #define LPC_ENET_RXFILTER_CTRL(x) (x + 0x200) 82 #define LPC_ENET_RXFILTERWOLSTATUS(x) (x + 0x204) 83 #define LPC_ENET_RXFILTERWOLCLEAR(x) (x + 0x208) 84 #define LPC_ENET_HASHFILTERL(x) (x + 0x210) 85 #define LPC_ENET_HASHFILTERH(x) (x + 0x214) 86 #define LPC_ENET_INTSTATUS(x) (x + 0xFE0) 87 #define LPC_ENET_INTENABLE(x) (x + 0xFE4) 88 #define LPC_ENET_INTCLEAR(x) (x + 0xFE8) 89 #define LPC_ENET_INTSET(x) (x + 0xFEC) 90 #define LPC_ENET_POWERDOWN(x) (x + 0xFF4) 91 92 /* 93 * mac1 register definitions 94 */ 95 #define LPC_MAC1_RECV_ENABLE (1 << 0) 96 #define LPC_MAC1_PASS_ALL_RX_FRAMES (1 << 1) 97 #define LPC_MAC1_RX_FLOW_CONTROL (1 << 2) 98 #define LPC_MAC1_TX_FLOW_CONTROL (1 << 3) 99 #define LPC_MAC1_LOOPBACK (1 << 4) 100 #define LPC_MAC1_RESET_TX (1 << 8) 101 #define LPC_MAC1_RESET_MCS_TX (1 << 9) 102 #define LPC_MAC1_RESET_RX (1 << 10) 103 #define LPC_MAC1_RESET_MCS_RX (1 << 11) 104 #define LPC_MAC1_SIMULATION_RESET (1 << 14) 105 #define LPC_MAC1_SOFT_RESET (1 << 15) 106 107 /* 108 * mac2 register definitions 109 */ 110 #define LPC_MAC2_FULL_DUPLEX (1 << 0) 111 #define LPC_MAC2_FRAME_LENGTH_CHECKING (1 << 1) 112 #define LPC_MAC2_HUGH_LENGTH_CHECKING (1 << 2) 113 #define LPC_MAC2_DELAYED_CRC (1 << 3) 114 #define LPC_MAC2_CRC_ENABLE (1 << 4) 115 #define LPC_MAC2_PAD_CRC_ENABLE (1 << 5) 116 #define LPC_MAC2_VLAN_PAD_ENABLE (1 << 6) 117 #define LPC_MAC2_AUTO_DETECT_PAD_ENABLE (1 << 7) 118 #define LPC_MAC2_PURE_PREAMBLE_ENFORCEMENT (1 << 8) 119 #define LPC_MAC2_LONG_PREAMBLE_ENFORCEMENT (1 << 9) 120 #define LPC_MAC2_NO_BACKOFF (1 << 12) 121 #define LPC_MAC2_BACK_PRESSURE (1 << 13) 122 #define LPC_MAC2_EXCESS_DEFER (1 << 14) 123 124 /* 125 * ipgt register definitions 126 */ 127 #define LPC_IPGT_LOAD(n) ((n) & 0x7F) 128 129 /* 130 * ipgr register definitions 131 */ 132 #define LPC_IPGR_LOAD_PART2(n) ((n) & 0x7F) 133 #define LPC_IPGR_LOAD_PART1(n) (((n) & 0x7F) << 8) 134 135 /* 136 * clrt register definitions 137 */ 138 #define LPC_CLRT_LOAD_RETRY_MAX(n) ((n) & 0xF) 139 #define LPC_CLRT_LOAD_COLLISION_WINDOW(n) (((n) & 0x3F) << 8) 140 141 /* 142 * maxf register definitions 143 */ 144 #define LPC_MAXF_LOAD_MAX_FRAME_LEN(n) ((n) & 0xFFFF) 145 146 /* 147 * supp register definitions 148 */ 149 #define LPC_SUPP_SPEED (1 << 8) 150 #define LPC_SUPP_RESET_RMII (1 << 11) 151 152 /* 153 * test register definitions 154 */ 155 #define LPC_TEST_SHORTCUT_PAUSE_QUANTA (1 << 0) 156 #define LPC_TEST_PAUSE (1 << 1) 157 #define LPC_TEST_BACKPRESSURE (1 << 2) 158 159 /* 160 * mcfg register definitions 161 */ 162 #define LPC_MCFG_SCAN_INCREMENT (1 << 0) 163 #define LPC_MCFG_SUPPRESS_PREAMBLE (1 << 1) 164 #define LPC_MCFG_CLOCK_SELECT(n) (((n) & 0x7) << 2) 165 #define LPC_MCFG_CLOCK_HOST_DIV_4 0 166 #define LPC_MCFG_CLOCK_HOST_DIV_6 2 167 #define LPC_MCFG_CLOCK_HOST_DIV_8 3 168 #define LPC_MCFG_CLOCK_HOST_DIV_10 4 169 #define LPC_MCFG_CLOCK_HOST_DIV_14 5 170 #define LPC_MCFG_CLOCK_HOST_DIV_20 6 171 #define LPC_MCFG_CLOCK_HOST_DIV_28 7 172 #define LPC_MCFG_RESET_MII_MGMT (1 << 15) 173 174 /* 175 * mcmd register definitions 176 */ 177 #define LPC_MCMD_READ (1 << 0) 178 #define LPC_MCMD_SCAN (1 << 1) 179 180 /* 181 * madr register definitions 182 */ 183 #define LPC_MADR_REGISTER_ADDRESS(n) ((n) & 0x1F) 184 #define LPC_MADR_PHY_0ADDRESS(n) (((n) & 0x1F) << 8) 185 186 /* 187 * mwtd register definitions 188 */ 189 #define LPC_MWDT_WRITE(n) ((n) & 0xFFFF) 190 191 /* 192 * mrdd register definitions 193 */ 194 #define LPC_MRDD_READ_MASK 0xFFFF 195 196 /* 197 * mind register definitions 198 */ 199 #define LPC_MIND_BUSY (1 << 0) 200 #define LPC_MIND_SCANNING (1 << 1) 201 #define LPC_MIND_NOT_VALID (1 << 2) 202 #define LPC_MIND_MII_LINK_FAIL (1 << 3) 203 204 /* 205 * command register definitions 206 */ 207 #define LPC_COMMAND_RXENABLE (1 << 0) 208 #define LPC_COMMAND_TXENABLE (1 << 1) 209 #define LPC_COMMAND_REG_RESET (1 << 3) 210 #define LPC_COMMAND_TXRESET (1 << 4) 211 #define LPC_COMMAND_RXRESET (1 << 5) 212 #define LPC_COMMAND_PASSRUNTFRAME (1 << 6) 213 #define LPC_COMMAND_PASSRXFILTER (1 << 7) 214 #define LPC_COMMAND_TXFLOWCONTROL (1 << 8) 215 #define LPC_COMMAND_RMII (1 << 9) 216 #define LPC_COMMAND_FULLDUPLEX (1 << 10) 217 218 /* 219 * status register definitions 220 */ 221 #define LPC_STATUS_RXACTIVE (1 << 0) 222 #define LPC_STATUS_TXACTIVE (1 << 1) 223 224 /* 225 * tsv0 register definitions 226 */ 227 #define LPC_TSV0_CRC_ERROR (1 << 0) 228 #define LPC_TSV0_LENGTH_CHECK_ERROR (1 << 1) 229 #define LPC_TSV0_LENGTH_OUT_OF_RANGE (1 << 2) 230 #define LPC_TSV0_DONE (1 << 3) 231 #define LPC_TSV0_MULTICAST (1 << 4) 232 #define LPC_TSV0_BROADCAST (1 << 5) 233 #define LPC_TSV0_PACKET_DEFER (1 << 6) 234 #define LPC_TSV0_ESCESSIVE_DEFER (1 << 7) 235 #define LPC_TSV0_ESCESSIVE_COLLISION (1 << 8) 236 #define LPC_TSV0_LATE_COLLISION (1 << 9) 237 #define LPC_TSV0_GIANT (1 << 10) 238 #define LPC_TSV0_UNDERRUN (1 << 11) 239 #define LPC_TSV0_TOTAL_BYTES(n) (((n) >> 12) & 0xFFFF) 240 #define LPC_TSV0_CONTROL_FRAME (1 << 28) 241 #define LPC_TSV0_PAUSE (1 << 29) 242 #define LPC_TSV0_BACKPRESSURE (1 << 30) 243 #define LPC_TSV0_VLAN (1 << 31) 244 245 /* 246 * tsv1 register definitions 247 */ 248 #define LPC_TSV1_TRANSMIT_BYTE_COUNT(n) ((n) & 0xFFFF) 249 #define LPC_TSV1_COLLISION_COUNT(n) (((n) >> 16) & 0xF) 250 251 /* 252 * rsv register definitions 253 */ 254 #define LPC_RSV_RECEIVED_BYTE_COUNT(n) ((n) & 0xFFFF) 255 #define LPC_RSV_RXDV_EVENT_IGNORED (1 << 16) 256 #define LPC_RSV_RXDV_EVENT_PREVIOUSLY_SEEN (1 << 17) 257 #define LPC_RSV_CARRIER_EVNT_PREVIOUS_SEEN (1 << 18) 258 #define LPC_RSV_RECEIVE_CODE_VIOLATION (1 << 19) 259 #define LPC_RSV_CRC_ERROR (1 << 20) 260 #define LPC_RSV_LENGTH_CHECK_ERROR (1 << 21) 261 #define LPC_RSV_LENGTH_OUT_OF_RANGE (1 << 22) 262 #define LPC_RSV_RECEIVE_OK (1 << 23) 263 #define LPC_RSV_MULTICAST (1 << 24) 264 #define LPC_RSV_BROADCAST (1 << 25) 265 #define LPC_RSV_DRIBBLE_NIBBLE (1 << 26) 266 #define LPC_RSV_CONTROL_FRAME (1 << 27) 267 #define LPC_RSV_PAUSE (1 << 28) 268 #define LPC_RSV_UNSUPPORTED_OPCODE (1 << 29) 269 #define LPC_RSV_VLAN (1 << 30) 270 271 /* 272 * flowcontrolcounter register definitions 273 */ 274 #define LPC_FCCR_MIRRORCOUNTER(n) ((n) & 0xFFFF) 275 #define LPC_FCCR_PAUSETIMER(n) (((n) >> 16) & 0xFFFF) 276 277 /* 278 * flowcontrolstatus register definitions 279 */ 280 #define LPC_FCCR_MIRRORCOUNTERCURRENT(n) ((n) & 0xFFFF) 281 282 /* 283 * rxfliterctrl, rxfilterwolstatus, and rxfilterwolclear shared 284 * register definitions 285 */ 286 #define LPC_RXFLTRW_ACCEPTUNICAST (1 << 0) 287 #define LPC_RXFLTRW_ACCEPTUBROADCAST (1 << 1) 288 #define LPC_RXFLTRW_ACCEPTUMULTICAST (1 << 2) 289 #define LPC_RXFLTRW_ACCEPTUNICASTHASH (1 << 3) 290 #define LPC_RXFLTRW_ACCEPTUMULTICASTHASH (1 << 4) 291 #define LPC_RXFLTRW_ACCEPTPERFECT (1 << 5) 292 293 /* 294 * rxfliterctrl register definitions 295 */ 296 #define LPC_RXFLTRWSTS_MAGICPACKETENWOL (1 << 12) 297 #define LPC_RXFLTRWSTS_RXFILTERENWOL (1 << 13) 298 299 /* 300 * rxfilterwolstatus/rxfilterwolclear register definitions 301 */ 302 #define LPC_RXFLTRWSTS_RXFILTERWOL (1 << 7) 303 #define LPC_RXFLTRWSTS_MAGICPACKETWOL (1 << 8) 304 305 /* 306 * intstatus, intenable, intclear, and Intset shared register 307 * definitions 308 */ 309 #define LPC_MACINT_RXOVERRUNINTEN (1 << 0) 310 #define LPC_MACINT_RXERRORONINT (1 << 1) 311 #define LPC_MACINT_RXFINISHEDINTEN (1 << 2) 312 #define LPC_MACINT_RXDONEINTEN (1 << 3) 313 #define LPC_MACINT_TXUNDERRUNINTEN (1 << 4) 314 #define LPC_MACINT_TXERRORINTEN (1 << 5) 315 #define LPC_MACINT_TXFINISHEDINTEN (1 << 6) 316 #define LPC_MACINT_TXDONEINTEN (1 << 7) 317 #define LPC_MACINT_SOFTINTEN (1 << 12) 318 #define LPC_MACINT_WAKEUPINTEN (1 << 13) 319 320 /* 321 * powerdown register definitions 322 */ 323 #define LPC_POWERDOWN_MACAHB (1 << 31) 324 325 static phy_interface_t lpc_phy_interface_mode(struct device *dev) 326 { 327 if (dev && dev->of_node) { 328 const char *mode = of_get_property(dev->of_node, 329 "phy-mode", NULL); 330 if (mode && !strcmp(mode, "mii")) 331 return PHY_INTERFACE_MODE_MII; 332 } 333 return PHY_INTERFACE_MODE_RMII; 334 } 335 336 static bool use_iram_for_net(struct device *dev) 337 { 338 if (dev && dev->of_node) 339 return of_property_read_bool(dev->of_node, "use-iram"); 340 return false; 341 } 342 343 /* Receive Status information word */ 344 #define RXSTATUS_SIZE 0x000007FF 345 #define RXSTATUS_CONTROL (1 << 18) 346 #define RXSTATUS_VLAN (1 << 19) 347 #define RXSTATUS_FILTER (1 << 20) 348 #define RXSTATUS_MULTICAST (1 << 21) 349 #define RXSTATUS_BROADCAST (1 << 22) 350 #define RXSTATUS_CRC (1 << 23) 351 #define RXSTATUS_SYMBOL (1 << 24) 352 #define RXSTATUS_LENGTH (1 << 25) 353 #define RXSTATUS_RANGE (1 << 26) 354 #define RXSTATUS_ALIGN (1 << 27) 355 #define RXSTATUS_OVERRUN (1 << 28) 356 #define RXSTATUS_NODESC (1 << 29) 357 #define RXSTATUS_LAST (1 << 30) 358 #define RXSTATUS_ERROR (1 << 31) 359 360 #define RXSTATUS_STATUS_ERROR \ 361 (RXSTATUS_NODESC | RXSTATUS_OVERRUN | RXSTATUS_ALIGN | \ 362 RXSTATUS_RANGE | RXSTATUS_LENGTH | RXSTATUS_SYMBOL | RXSTATUS_CRC) 363 364 /* Receive Descriptor control word */ 365 #define RXDESC_CONTROL_SIZE 0x000007FF 366 #define RXDESC_CONTROL_INT (1 << 31) 367 368 /* Transmit Status information word */ 369 #define TXSTATUS_COLLISIONS_GET(x) (((x) >> 21) & 0xF) 370 #define TXSTATUS_DEFER (1 << 25) 371 #define TXSTATUS_EXCESSDEFER (1 << 26) 372 #define TXSTATUS_EXCESSCOLL (1 << 27) 373 #define TXSTATUS_LATECOLL (1 << 28) 374 #define TXSTATUS_UNDERRUN (1 << 29) 375 #define TXSTATUS_NODESC (1 << 30) 376 #define TXSTATUS_ERROR (1 << 31) 377 378 /* Transmit Descriptor control word */ 379 #define TXDESC_CONTROL_SIZE 0x000007FF 380 #define TXDESC_CONTROL_OVERRIDE (1 << 26) 381 #define TXDESC_CONTROL_HUGE (1 << 27) 382 #define TXDESC_CONTROL_PAD (1 << 28) 383 #define TXDESC_CONTROL_CRC (1 << 29) 384 #define TXDESC_CONTROL_LAST (1 << 30) 385 #define TXDESC_CONTROL_INT (1 << 31) 386 387 /* 388 * Structure of a TX/RX descriptors and RX status 389 */ 390 struct txrx_desc_t { 391 __le32 packet; 392 __le32 control; 393 }; 394 struct rx_status_t { 395 __le32 statusinfo; 396 __le32 statushashcrc; 397 }; 398 399 /* 400 * Device driver data structure 401 */ 402 struct netdata_local { 403 struct platform_device *pdev; 404 struct net_device *ndev; 405 spinlock_t lock; 406 void __iomem *net_base; 407 u32 msg_enable; 408 unsigned int skblen[ENET_TX_DESC]; 409 unsigned int last_tx_idx; 410 unsigned int num_used_tx_buffs; 411 struct mii_bus *mii_bus; 412 struct clk *clk; 413 dma_addr_t dma_buff_base_p; 414 void *dma_buff_base_v; 415 size_t dma_buff_size; 416 struct txrx_desc_t *tx_desc_v; 417 u32 *tx_stat_v; 418 void *tx_buff_v; 419 struct txrx_desc_t *rx_desc_v; 420 struct rx_status_t *rx_stat_v; 421 void *rx_buff_v; 422 int link; 423 int speed; 424 int duplex; 425 struct napi_struct napi; 426 }; 427 428 /* 429 * MAC support functions 430 */ 431 static void __lpc_set_mac(struct netdata_local *pldat, u8 *mac) 432 { 433 u32 tmp; 434 435 /* Set station address */ 436 tmp = mac[0] | ((u32)mac[1] << 8); 437 writel(tmp, LPC_ENET_SA2(pldat->net_base)); 438 tmp = mac[2] | ((u32)mac[3] << 8); 439 writel(tmp, LPC_ENET_SA1(pldat->net_base)); 440 tmp = mac[4] | ((u32)mac[5] << 8); 441 writel(tmp, LPC_ENET_SA0(pldat->net_base)); 442 443 netdev_dbg(pldat->ndev, "Ethernet MAC address %pM\n", mac); 444 } 445 446 static void __lpc_get_mac(struct netdata_local *pldat, u8 *mac) 447 { 448 u32 tmp; 449 450 /* Get station address */ 451 tmp = readl(LPC_ENET_SA2(pldat->net_base)); 452 mac[0] = tmp & 0xFF; 453 mac[1] = tmp >> 8; 454 tmp = readl(LPC_ENET_SA1(pldat->net_base)); 455 mac[2] = tmp & 0xFF; 456 mac[3] = tmp >> 8; 457 tmp = readl(LPC_ENET_SA0(pldat->net_base)); 458 mac[4] = tmp & 0xFF; 459 mac[5] = tmp >> 8; 460 } 461 462 static void __lpc_params_setup(struct netdata_local *pldat) 463 { 464 u32 tmp; 465 466 if (pldat->duplex == DUPLEX_FULL) { 467 tmp = readl(LPC_ENET_MAC2(pldat->net_base)); 468 tmp |= LPC_MAC2_FULL_DUPLEX; 469 writel(tmp, LPC_ENET_MAC2(pldat->net_base)); 470 tmp = readl(LPC_ENET_COMMAND(pldat->net_base)); 471 tmp |= LPC_COMMAND_FULLDUPLEX; 472 writel(tmp, LPC_ENET_COMMAND(pldat->net_base)); 473 writel(LPC_IPGT_LOAD(0x15), LPC_ENET_IPGT(pldat->net_base)); 474 } else { 475 tmp = readl(LPC_ENET_MAC2(pldat->net_base)); 476 tmp &= ~LPC_MAC2_FULL_DUPLEX; 477 writel(tmp, LPC_ENET_MAC2(pldat->net_base)); 478 tmp = readl(LPC_ENET_COMMAND(pldat->net_base)); 479 tmp &= ~LPC_COMMAND_FULLDUPLEX; 480 writel(tmp, LPC_ENET_COMMAND(pldat->net_base)); 481 writel(LPC_IPGT_LOAD(0x12), LPC_ENET_IPGT(pldat->net_base)); 482 } 483 484 if (pldat->speed == SPEED_100) 485 writel(LPC_SUPP_SPEED, LPC_ENET_SUPP(pldat->net_base)); 486 else 487 writel(0, LPC_ENET_SUPP(pldat->net_base)); 488 } 489 490 static void __lpc_eth_reset(struct netdata_local *pldat) 491 { 492 /* Reset all MAC logic */ 493 writel((LPC_MAC1_RESET_TX | LPC_MAC1_RESET_MCS_TX | LPC_MAC1_RESET_RX | 494 LPC_MAC1_RESET_MCS_RX | LPC_MAC1_SIMULATION_RESET | 495 LPC_MAC1_SOFT_RESET), LPC_ENET_MAC1(pldat->net_base)); 496 writel((LPC_COMMAND_REG_RESET | LPC_COMMAND_TXRESET | 497 LPC_COMMAND_RXRESET), LPC_ENET_COMMAND(pldat->net_base)); 498 } 499 500 static int __lpc_mii_mngt_reset(struct netdata_local *pldat) 501 { 502 /* Reset MII management hardware */ 503 writel(LPC_MCFG_RESET_MII_MGMT, LPC_ENET_MCFG(pldat->net_base)); 504 505 /* Setup MII clock to slowest rate with a /28 divider */ 506 writel(LPC_MCFG_CLOCK_SELECT(LPC_MCFG_CLOCK_HOST_DIV_28), 507 LPC_ENET_MCFG(pldat->net_base)); 508 509 return 0; 510 } 511 512 static inline phys_addr_t __va_to_pa(void *addr, struct netdata_local *pldat) 513 { 514 phys_addr_t phaddr; 515 516 phaddr = addr - pldat->dma_buff_base_v; 517 phaddr += pldat->dma_buff_base_p; 518 519 return phaddr; 520 } 521 522 static void lpc_eth_enable_int(void __iomem *regbase) 523 { 524 writel((LPC_MACINT_RXDONEINTEN | LPC_MACINT_TXDONEINTEN), 525 LPC_ENET_INTENABLE(regbase)); 526 } 527 528 static void lpc_eth_disable_int(void __iomem *regbase) 529 { 530 writel(0, LPC_ENET_INTENABLE(regbase)); 531 } 532 533 /* Setup TX/RX descriptors */ 534 static void __lpc_txrx_desc_setup(struct netdata_local *pldat) 535 { 536 u32 *ptxstat; 537 void *tbuff; 538 int i; 539 struct txrx_desc_t *ptxrxdesc; 540 struct rx_status_t *prxstat; 541 542 tbuff = PTR_ALIGN(pldat->dma_buff_base_v, 16); 543 544 /* Setup TX descriptors, status, and buffers */ 545 pldat->tx_desc_v = tbuff; 546 tbuff += sizeof(struct txrx_desc_t) * ENET_TX_DESC; 547 548 pldat->tx_stat_v = tbuff; 549 tbuff += sizeof(u32) * ENET_TX_DESC; 550 551 tbuff = PTR_ALIGN(tbuff, 16); 552 pldat->tx_buff_v = tbuff; 553 tbuff += ENET_MAXF_SIZE * ENET_TX_DESC; 554 555 /* Setup RX descriptors, status, and buffers */ 556 pldat->rx_desc_v = tbuff; 557 tbuff += sizeof(struct txrx_desc_t) * ENET_RX_DESC; 558 559 tbuff = PTR_ALIGN(tbuff, 16); 560 pldat->rx_stat_v = tbuff; 561 tbuff += sizeof(struct rx_status_t) * ENET_RX_DESC; 562 563 tbuff = PTR_ALIGN(tbuff, 16); 564 pldat->rx_buff_v = tbuff; 565 tbuff += ENET_MAXF_SIZE * ENET_RX_DESC; 566 567 /* Map the TX descriptors to the TX buffers in hardware */ 568 for (i = 0; i < ENET_TX_DESC; i++) { 569 ptxstat = &pldat->tx_stat_v[i]; 570 ptxrxdesc = &pldat->tx_desc_v[i]; 571 572 ptxrxdesc->packet = __va_to_pa( 573 pldat->tx_buff_v + i * ENET_MAXF_SIZE, pldat); 574 ptxrxdesc->control = 0; 575 *ptxstat = 0; 576 } 577 578 /* Map the RX descriptors to the RX buffers in hardware */ 579 for (i = 0; i < ENET_RX_DESC; i++) { 580 prxstat = &pldat->rx_stat_v[i]; 581 ptxrxdesc = &pldat->rx_desc_v[i]; 582 583 ptxrxdesc->packet = __va_to_pa( 584 pldat->rx_buff_v + i * ENET_MAXF_SIZE, pldat); 585 ptxrxdesc->control = RXDESC_CONTROL_INT | (ENET_MAXF_SIZE - 1); 586 prxstat->statusinfo = 0; 587 prxstat->statushashcrc = 0; 588 } 589 590 /* Setup base addresses in hardware to point to buffers and 591 * descriptors 592 */ 593 writel((ENET_TX_DESC - 1), 594 LPC_ENET_TXDESCRIPTORNUMBER(pldat->net_base)); 595 writel(__va_to_pa(pldat->tx_desc_v, pldat), 596 LPC_ENET_TXDESCRIPTOR(pldat->net_base)); 597 writel(__va_to_pa(pldat->tx_stat_v, pldat), 598 LPC_ENET_TXSTATUS(pldat->net_base)); 599 writel((ENET_RX_DESC - 1), 600 LPC_ENET_RXDESCRIPTORNUMBER(pldat->net_base)); 601 writel(__va_to_pa(pldat->rx_desc_v, pldat), 602 LPC_ENET_RXDESCRIPTOR(pldat->net_base)); 603 writel(__va_to_pa(pldat->rx_stat_v, pldat), 604 LPC_ENET_RXSTATUS(pldat->net_base)); 605 } 606 607 static void __lpc_eth_init(struct netdata_local *pldat) 608 { 609 u32 tmp; 610 611 /* Disable controller and reset */ 612 tmp = readl(LPC_ENET_COMMAND(pldat->net_base)); 613 tmp &= ~LPC_COMMAND_RXENABLE | LPC_COMMAND_TXENABLE; 614 writel(tmp, LPC_ENET_COMMAND(pldat->net_base)); 615 tmp = readl(LPC_ENET_MAC1(pldat->net_base)); 616 tmp &= ~LPC_MAC1_RECV_ENABLE; 617 writel(tmp, LPC_ENET_MAC1(pldat->net_base)); 618 619 /* Initial MAC setup */ 620 writel(LPC_MAC1_PASS_ALL_RX_FRAMES, LPC_ENET_MAC1(pldat->net_base)); 621 writel((LPC_MAC2_PAD_CRC_ENABLE | LPC_MAC2_CRC_ENABLE), 622 LPC_ENET_MAC2(pldat->net_base)); 623 writel(ENET_MAXF_SIZE, LPC_ENET_MAXF(pldat->net_base)); 624 625 /* Collision window, gap */ 626 writel((LPC_CLRT_LOAD_RETRY_MAX(0xF) | 627 LPC_CLRT_LOAD_COLLISION_WINDOW(0x37)), 628 LPC_ENET_CLRT(pldat->net_base)); 629 writel(LPC_IPGR_LOAD_PART2(0x12), LPC_ENET_IPGR(pldat->net_base)); 630 631 if (lpc_phy_interface_mode(&pldat->pdev->dev) == PHY_INTERFACE_MODE_MII) 632 writel(LPC_COMMAND_PASSRUNTFRAME, 633 LPC_ENET_COMMAND(pldat->net_base)); 634 else { 635 writel((LPC_COMMAND_PASSRUNTFRAME | LPC_COMMAND_RMII), 636 LPC_ENET_COMMAND(pldat->net_base)); 637 writel(LPC_SUPP_RESET_RMII, LPC_ENET_SUPP(pldat->net_base)); 638 } 639 640 __lpc_params_setup(pldat); 641 642 /* Setup TX and RX descriptors */ 643 __lpc_txrx_desc_setup(pldat); 644 645 /* Setup packet filtering */ 646 writel((LPC_RXFLTRW_ACCEPTUBROADCAST | LPC_RXFLTRW_ACCEPTPERFECT), 647 LPC_ENET_RXFILTER_CTRL(pldat->net_base)); 648 649 /* Get the next TX buffer output index */ 650 pldat->num_used_tx_buffs = 0; 651 pldat->last_tx_idx = 652 readl(LPC_ENET_TXCONSUMEINDEX(pldat->net_base)); 653 654 /* Clear and enable interrupts */ 655 writel(0xFFFF, LPC_ENET_INTCLEAR(pldat->net_base)); 656 smp_wmb(); 657 lpc_eth_enable_int(pldat->net_base); 658 659 /* Enable controller */ 660 tmp = readl(LPC_ENET_COMMAND(pldat->net_base)); 661 tmp |= LPC_COMMAND_RXENABLE | LPC_COMMAND_TXENABLE; 662 writel(tmp, LPC_ENET_COMMAND(pldat->net_base)); 663 tmp = readl(LPC_ENET_MAC1(pldat->net_base)); 664 tmp |= LPC_MAC1_RECV_ENABLE; 665 writel(tmp, LPC_ENET_MAC1(pldat->net_base)); 666 } 667 668 static void __lpc_eth_shutdown(struct netdata_local *pldat) 669 { 670 /* Reset ethernet and power down PHY */ 671 __lpc_eth_reset(pldat); 672 writel(0, LPC_ENET_MAC1(pldat->net_base)); 673 writel(0, LPC_ENET_MAC2(pldat->net_base)); 674 } 675 676 /* 677 * MAC<--->PHY support functions 678 */ 679 static int lpc_mdio_read(struct mii_bus *bus, int phy_id, int phyreg) 680 { 681 struct netdata_local *pldat = bus->priv; 682 unsigned long timeout = jiffies + msecs_to_jiffies(100); 683 int lps; 684 685 writel(((phy_id << 8) | phyreg), LPC_ENET_MADR(pldat->net_base)); 686 writel(LPC_MCMD_READ, LPC_ENET_MCMD(pldat->net_base)); 687 688 /* Wait for unbusy status */ 689 while (readl(LPC_ENET_MIND(pldat->net_base)) & LPC_MIND_BUSY) { 690 if (time_after(jiffies, timeout)) 691 return -EIO; 692 cpu_relax(); 693 } 694 695 lps = readl(LPC_ENET_MRDD(pldat->net_base)); 696 writel(0, LPC_ENET_MCMD(pldat->net_base)); 697 698 return lps; 699 } 700 701 static int lpc_mdio_write(struct mii_bus *bus, int phy_id, int phyreg, 702 u16 phydata) 703 { 704 struct netdata_local *pldat = bus->priv; 705 unsigned long timeout = jiffies + msecs_to_jiffies(100); 706 707 writel(((phy_id << 8) | phyreg), LPC_ENET_MADR(pldat->net_base)); 708 writel(phydata, LPC_ENET_MWTD(pldat->net_base)); 709 710 /* Wait for completion */ 711 while (readl(LPC_ENET_MIND(pldat->net_base)) & LPC_MIND_BUSY) { 712 if (time_after(jiffies, timeout)) 713 return -EIO; 714 cpu_relax(); 715 } 716 717 return 0; 718 } 719 720 static int lpc_mdio_reset(struct mii_bus *bus) 721 { 722 return __lpc_mii_mngt_reset((struct netdata_local *)bus->priv); 723 } 724 725 static void lpc_handle_link_change(struct net_device *ndev) 726 { 727 struct netdata_local *pldat = netdev_priv(ndev); 728 struct phy_device *phydev = ndev->phydev; 729 unsigned long flags; 730 731 bool status_change = false; 732 733 spin_lock_irqsave(&pldat->lock, flags); 734 735 if (phydev->link) { 736 if ((pldat->speed != phydev->speed) || 737 (pldat->duplex != phydev->duplex)) { 738 pldat->speed = phydev->speed; 739 pldat->duplex = phydev->duplex; 740 status_change = true; 741 } 742 } 743 744 if (phydev->link != pldat->link) { 745 if (!phydev->link) { 746 pldat->speed = 0; 747 pldat->duplex = -1; 748 } 749 pldat->link = phydev->link; 750 751 status_change = true; 752 } 753 754 spin_unlock_irqrestore(&pldat->lock, flags); 755 756 if (status_change) 757 __lpc_params_setup(pldat); 758 } 759 760 static int lpc_mii_probe(struct net_device *ndev) 761 { 762 struct netdata_local *pldat = netdev_priv(ndev); 763 struct phy_device *phydev = phy_find_first(pldat->mii_bus); 764 765 if (!phydev) { 766 netdev_err(ndev, "no PHY found\n"); 767 return -ENODEV; 768 } 769 770 /* Attach to the PHY */ 771 if (lpc_phy_interface_mode(&pldat->pdev->dev) == PHY_INTERFACE_MODE_MII) 772 netdev_info(ndev, "using MII interface\n"); 773 else 774 netdev_info(ndev, "using RMII interface\n"); 775 phydev = phy_connect(ndev, phydev_name(phydev), 776 &lpc_handle_link_change, 777 lpc_phy_interface_mode(&pldat->pdev->dev)); 778 779 if (IS_ERR(phydev)) { 780 netdev_err(ndev, "Could not attach to PHY\n"); 781 return PTR_ERR(phydev); 782 } 783 784 phy_set_max_speed(phydev, SPEED_100); 785 786 phydev->advertising = phydev->supported; 787 788 pldat->link = 0; 789 pldat->speed = 0; 790 pldat->duplex = -1; 791 792 phy_attached_info(phydev); 793 794 return 0; 795 } 796 797 static int lpc_mii_init(struct netdata_local *pldat) 798 { 799 int err = -ENXIO; 800 801 pldat->mii_bus = mdiobus_alloc(); 802 if (!pldat->mii_bus) { 803 err = -ENOMEM; 804 goto err_out; 805 } 806 807 /* Setup MII mode */ 808 if (lpc_phy_interface_mode(&pldat->pdev->dev) == PHY_INTERFACE_MODE_MII) 809 writel(LPC_COMMAND_PASSRUNTFRAME, 810 LPC_ENET_COMMAND(pldat->net_base)); 811 else { 812 writel((LPC_COMMAND_PASSRUNTFRAME | LPC_COMMAND_RMII), 813 LPC_ENET_COMMAND(pldat->net_base)); 814 writel(LPC_SUPP_RESET_RMII, LPC_ENET_SUPP(pldat->net_base)); 815 } 816 817 pldat->mii_bus->name = "lpc_mii_bus"; 818 pldat->mii_bus->read = &lpc_mdio_read; 819 pldat->mii_bus->write = &lpc_mdio_write; 820 pldat->mii_bus->reset = &lpc_mdio_reset; 821 snprintf(pldat->mii_bus->id, MII_BUS_ID_SIZE, "%s-%x", 822 pldat->pdev->name, pldat->pdev->id); 823 pldat->mii_bus->priv = pldat; 824 pldat->mii_bus->parent = &pldat->pdev->dev; 825 826 platform_set_drvdata(pldat->pdev, pldat->mii_bus); 827 828 if (mdiobus_register(pldat->mii_bus)) 829 goto err_out_unregister_bus; 830 831 if (lpc_mii_probe(pldat->ndev) != 0) 832 goto err_out_unregister_bus; 833 834 return 0; 835 836 err_out_unregister_bus: 837 mdiobus_unregister(pldat->mii_bus); 838 mdiobus_free(pldat->mii_bus); 839 err_out: 840 return err; 841 } 842 843 static void __lpc_handle_xmit(struct net_device *ndev) 844 { 845 struct netdata_local *pldat = netdev_priv(ndev); 846 u32 txcidx, *ptxstat, txstat; 847 848 txcidx = readl(LPC_ENET_TXCONSUMEINDEX(pldat->net_base)); 849 while (pldat->last_tx_idx != txcidx) { 850 unsigned int skblen = pldat->skblen[pldat->last_tx_idx]; 851 852 /* A buffer is available, get buffer status */ 853 ptxstat = &pldat->tx_stat_v[pldat->last_tx_idx]; 854 txstat = *ptxstat; 855 856 /* Next buffer and decrement used buffer counter */ 857 pldat->num_used_tx_buffs--; 858 pldat->last_tx_idx++; 859 if (pldat->last_tx_idx >= ENET_TX_DESC) 860 pldat->last_tx_idx = 0; 861 862 /* Update collision counter */ 863 ndev->stats.collisions += TXSTATUS_COLLISIONS_GET(txstat); 864 865 /* Any errors occurred? */ 866 if (txstat & TXSTATUS_ERROR) { 867 if (txstat & TXSTATUS_UNDERRUN) { 868 /* FIFO underrun */ 869 ndev->stats.tx_fifo_errors++; 870 } 871 if (txstat & TXSTATUS_LATECOLL) { 872 /* Late collision */ 873 ndev->stats.tx_aborted_errors++; 874 } 875 if (txstat & TXSTATUS_EXCESSCOLL) { 876 /* Excessive collision */ 877 ndev->stats.tx_aborted_errors++; 878 } 879 if (txstat & TXSTATUS_EXCESSDEFER) { 880 /* Defer limit */ 881 ndev->stats.tx_aborted_errors++; 882 } 883 ndev->stats.tx_errors++; 884 } else { 885 /* Update stats */ 886 ndev->stats.tx_packets++; 887 ndev->stats.tx_bytes += skblen; 888 } 889 890 txcidx = readl(LPC_ENET_TXCONSUMEINDEX(pldat->net_base)); 891 } 892 893 if (pldat->num_used_tx_buffs <= ENET_TX_DESC/2) { 894 if (netif_queue_stopped(ndev)) 895 netif_wake_queue(ndev); 896 } 897 } 898 899 static int __lpc_handle_recv(struct net_device *ndev, int budget) 900 { 901 struct netdata_local *pldat = netdev_priv(ndev); 902 struct sk_buff *skb; 903 u32 rxconsidx, len, ethst; 904 struct rx_status_t *prxstat; 905 int rx_done = 0; 906 907 /* Get the current RX buffer indexes */ 908 rxconsidx = readl(LPC_ENET_RXCONSUMEINDEX(pldat->net_base)); 909 while (rx_done < budget && rxconsidx != 910 readl(LPC_ENET_RXPRODUCEINDEX(pldat->net_base))) { 911 /* Get pointer to receive status */ 912 prxstat = &pldat->rx_stat_v[rxconsidx]; 913 len = (prxstat->statusinfo & RXSTATUS_SIZE) + 1; 914 915 /* Status error? */ 916 ethst = prxstat->statusinfo; 917 if ((ethst & (RXSTATUS_ERROR | RXSTATUS_STATUS_ERROR)) == 918 (RXSTATUS_ERROR | RXSTATUS_RANGE)) 919 ethst &= ~RXSTATUS_ERROR; 920 921 if (ethst & RXSTATUS_ERROR) { 922 int si = prxstat->statusinfo; 923 /* Check statuses */ 924 if (si & RXSTATUS_OVERRUN) { 925 /* Overrun error */ 926 ndev->stats.rx_fifo_errors++; 927 } else if (si & RXSTATUS_CRC) { 928 /* CRC error */ 929 ndev->stats.rx_crc_errors++; 930 } else if (si & RXSTATUS_LENGTH) { 931 /* Length error */ 932 ndev->stats.rx_length_errors++; 933 } else if (si & RXSTATUS_ERROR) { 934 /* Other error */ 935 ndev->stats.rx_length_errors++; 936 } 937 ndev->stats.rx_errors++; 938 } else { 939 /* Packet is good */ 940 skb = dev_alloc_skb(len); 941 if (!skb) { 942 ndev->stats.rx_dropped++; 943 } else { 944 /* Copy packet from buffer */ 945 skb_put_data(skb, 946 pldat->rx_buff_v + rxconsidx * ENET_MAXF_SIZE, 947 len); 948 949 /* Pass to upper layer */ 950 skb->protocol = eth_type_trans(skb, ndev); 951 netif_receive_skb(skb); 952 ndev->stats.rx_packets++; 953 ndev->stats.rx_bytes += len; 954 } 955 } 956 957 /* Increment consume index */ 958 rxconsidx = rxconsidx + 1; 959 if (rxconsidx >= ENET_RX_DESC) 960 rxconsidx = 0; 961 writel(rxconsidx, 962 LPC_ENET_RXCONSUMEINDEX(pldat->net_base)); 963 rx_done++; 964 } 965 966 return rx_done; 967 } 968 969 static int lpc_eth_poll(struct napi_struct *napi, int budget) 970 { 971 struct netdata_local *pldat = container_of(napi, 972 struct netdata_local, napi); 973 struct net_device *ndev = pldat->ndev; 974 int rx_done = 0; 975 struct netdev_queue *txq = netdev_get_tx_queue(ndev, 0); 976 977 __netif_tx_lock(txq, smp_processor_id()); 978 __lpc_handle_xmit(ndev); 979 __netif_tx_unlock(txq); 980 rx_done = __lpc_handle_recv(ndev, budget); 981 982 if (rx_done < budget) { 983 napi_complete_done(napi, rx_done); 984 lpc_eth_enable_int(pldat->net_base); 985 } 986 987 return rx_done; 988 } 989 990 static irqreturn_t __lpc_eth_interrupt(int irq, void *dev_id) 991 { 992 struct net_device *ndev = dev_id; 993 struct netdata_local *pldat = netdev_priv(ndev); 994 u32 tmp; 995 996 spin_lock(&pldat->lock); 997 998 tmp = readl(LPC_ENET_INTSTATUS(pldat->net_base)); 999 /* Clear interrupts */ 1000 writel(tmp, LPC_ENET_INTCLEAR(pldat->net_base)); 1001 1002 lpc_eth_disable_int(pldat->net_base); 1003 if (likely(napi_schedule_prep(&pldat->napi))) 1004 __napi_schedule(&pldat->napi); 1005 1006 spin_unlock(&pldat->lock); 1007 1008 return IRQ_HANDLED; 1009 } 1010 1011 static int lpc_eth_close(struct net_device *ndev) 1012 { 1013 unsigned long flags; 1014 struct netdata_local *pldat = netdev_priv(ndev); 1015 1016 if (netif_msg_ifdown(pldat)) 1017 dev_dbg(&pldat->pdev->dev, "shutting down %s\n", ndev->name); 1018 1019 napi_disable(&pldat->napi); 1020 netif_stop_queue(ndev); 1021 1022 if (ndev->phydev) 1023 phy_stop(ndev->phydev); 1024 1025 spin_lock_irqsave(&pldat->lock, flags); 1026 __lpc_eth_reset(pldat); 1027 netif_carrier_off(ndev); 1028 writel(0, LPC_ENET_MAC1(pldat->net_base)); 1029 writel(0, LPC_ENET_MAC2(pldat->net_base)); 1030 spin_unlock_irqrestore(&pldat->lock, flags); 1031 1032 clk_disable_unprepare(pldat->clk); 1033 1034 return 0; 1035 } 1036 1037 static int lpc_eth_hard_start_xmit(struct sk_buff *skb, struct net_device *ndev) 1038 { 1039 struct netdata_local *pldat = netdev_priv(ndev); 1040 u32 len, txidx; 1041 u32 *ptxstat; 1042 struct txrx_desc_t *ptxrxdesc; 1043 1044 len = skb->len; 1045 1046 spin_lock_irq(&pldat->lock); 1047 1048 if (pldat->num_used_tx_buffs >= (ENET_TX_DESC - 1)) { 1049 /* This function should never be called when there are no 1050 buffers */ 1051 netif_stop_queue(ndev); 1052 spin_unlock_irq(&pldat->lock); 1053 WARN(1, "BUG! TX request when no free TX buffers!\n"); 1054 return NETDEV_TX_BUSY; 1055 } 1056 1057 /* Get the next TX descriptor index */ 1058 txidx = readl(LPC_ENET_TXPRODUCEINDEX(pldat->net_base)); 1059 1060 /* Setup control for the transfer */ 1061 ptxstat = &pldat->tx_stat_v[txidx]; 1062 *ptxstat = 0; 1063 ptxrxdesc = &pldat->tx_desc_v[txidx]; 1064 ptxrxdesc->control = 1065 (len - 1) | TXDESC_CONTROL_LAST | TXDESC_CONTROL_INT; 1066 1067 /* Copy data to the DMA buffer */ 1068 memcpy(pldat->tx_buff_v + txidx * ENET_MAXF_SIZE, skb->data, len); 1069 1070 /* Save the buffer and increment the buffer counter */ 1071 pldat->skblen[txidx] = len; 1072 pldat->num_used_tx_buffs++; 1073 1074 /* Start transmit */ 1075 txidx++; 1076 if (txidx >= ENET_TX_DESC) 1077 txidx = 0; 1078 writel(txidx, LPC_ENET_TXPRODUCEINDEX(pldat->net_base)); 1079 1080 /* Stop queue if no more TX buffers */ 1081 if (pldat->num_used_tx_buffs >= (ENET_TX_DESC - 1)) 1082 netif_stop_queue(ndev); 1083 1084 spin_unlock_irq(&pldat->lock); 1085 1086 dev_kfree_skb(skb); 1087 return NETDEV_TX_OK; 1088 } 1089 1090 static int lpc_set_mac_address(struct net_device *ndev, void *p) 1091 { 1092 struct sockaddr *addr = p; 1093 struct netdata_local *pldat = netdev_priv(ndev); 1094 unsigned long flags; 1095 1096 if (!is_valid_ether_addr(addr->sa_data)) 1097 return -EADDRNOTAVAIL; 1098 memcpy(ndev->dev_addr, addr->sa_data, ETH_ALEN); 1099 1100 spin_lock_irqsave(&pldat->lock, flags); 1101 1102 /* Set station address */ 1103 __lpc_set_mac(pldat, ndev->dev_addr); 1104 1105 spin_unlock_irqrestore(&pldat->lock, flags); 1106 1107 return 0; 1108 } 1109 1110 static void lpc_eth_set_multicast_list(struct net_device *ndev) 1111 { 1112 struct netdata_local *pldat = netdev_priv(ndev); 1113 struct netdev_hw_addr_list *mcptr = &ndev->mc; 1114 struct netdev_hw_addr *ha; 1115 u32 tmp32, hash_val, hashlo, hashhi; 1116 unsigned long flags; 1117 1118 spin_lock_irqsave(&pldat->lock, flags); 1119 1120 /* Set station address */ 1121 __lpc_set_mac(pldat, ndev->dev_addr); 1122 1123 tmp32 = LPC_RXFLTRW_ACCEPTUBROADCAST | LPC_RXFLTRW_ACCEPTPERFECT; 1124 1125 if (ndev->flags & IFF_PROMISC) 1126 tmp32 |= LPC_RXFLTRW_ACCEPTUNICAST | 1127 LPC_RXFLTRW_ACCEPTUMULTICAST; 1128 if (ndev->flags & IFF_ALLMULTI) 1129 tmp32 |= LPC_RXFLTRW_ACCEPTUMULTICAST; 1130 1131 if (netdev_hw_addr_list_count(mcptr)) 1132 tmp32 |= LPC_RXFLTRW_ACCEPTUMULTICASTHASH; 1133 1134 writel(tmp32, LPC_ENET_RXFILTER_CTRL(pldat->net_base)); 1135 1136 1137 /* Set initial hash table */ 1138 hashlo = 0x0; 1139 hashhi = 0x0; 1140 1141 /* 64 bits : multicast address in hash table */ 1142 netdev_hw_addr_list_for_each(ha, mcptr) { 1143 hash_val = (ether_crc(6, ha->addr) >> 23) & 0x3F; 1144 1145 if (hash_val >= 32) 1146 hashhi |= 1 << (hash_val - 32); 1147 else 1148 hashlo |= 1 << hash_val; 1149 } 1150 1151 writel(hashlo, LPC_ENET_HASHFILTERL(pldat->net_base)); 1152 writel(hashhi, LPC_ENET_HASHFILTERH(pldat->net_base)); 1153 1154 spin_unlock_irqrestore(&pldat->lock, flags); 1155 } 1156 1157 static int lpc_eth_ioctl(struct net_device *ndev, struct ifreq *req, int cmd) 1158 { 1159 struct phy_device *phydev = ndev->phydev; 1160 1161 if (!netif_running(ndev)) 1162 return -EINVAL; 1163 1164 if (!phydev) 1165 return -ENODEV; 1166 1167 return phy_mii_ioctl(phydev, req, cmd); 1168 } 1169 1170 static int lpc_eth_open(struct net_device *ndev) 1171 { 1172 struct netdata_local *pldat = netdev_priv(ndev); 1173 int ret; 1174 1175 if (netif_msg_ifup(pldat)) 1176 dev_dbg(&pldat->pdev->dev, "enabling %s\n", ndev->name); 1177 1178 ret = clk_prepare_enable(pldat->clk); 1179 if (ret) 1180 return ret; 1181 1182 /* Suspended PHY makes LPC ethernet core block, so resume now */ 1183 phy_resume(ndev->phydev); 1184 1185 /* Reset and initialize */ 1186 __lpc_eth_reset(pldat); 1187 __lpc_eth_init(pldat); 1188 1189 /* schedule a link state check */ 1190 phy_start(ndev->phydev); 1191 netif_start_queue(ndev); 1192 napi_enable(&pldat->napi); 1193 1194 return 0; 1195 } 1196 1197 /* 1198 * Ethtool ops 1199 */ 1200 static void lpc_eth_ethtool_getdrvinfo(struct net_device *ndev, 1201 struct ethtool_drvinfo *info) 1202 { 1203 strlcpy(info->driver, MODNAME, sizeof(info->driver)); 1204 strlcpy(info->version, DRV_VERSION, sizeof(info->version)); 1205 strlcpy(info->bus_info, dev_name(ndev->dev.parent), 1206 sizeof(info->bus_info)); 1207 } 1208 1209 static u32 lpc_eth_ethtool_getmsglevel(struct net_device *ndev) 1210 { 1211 struct netdata_local *pldat = netdev_priv(ndev); 1212 1213 return pldat->msg_enable; 1214 } 1215 1216 static void lpc_eth_ethtool_setmsglevel(struct net_device *ndev, u32 level) 1217 { 1218 struct netdata_local *pldat = netdev_priv(ndev); 1219 1220 pldat->msg_enable = level; 1221 } 1222 1223 static const struct ethtool_ops lpc_eth_ethtool_ops = { 1224 .get_drvinfo = lpc_eth_ethtool_getdrvinfo, 1225 .get_msglevel = lpc_eth_ethtool_getmsglevel, 1226 .set_msglevel = lpc_eth_ethtool_setmsglevel, 1227 .get_link = ethtool_op_get_link, 1228 .get_link_ksettings = phy_ethtool_get_link_ksettings, 1229 .set_link_ksettings = phy_ethtool_set_link_ksettings, 1230 }; 1231 1232 static const struct net_device_ops lpc_netdev_ops = { 1233 .ndo_open = lpc_eth_open, 1234 .ndo_stop = lpc_eth_close, 1235 .ndo_start_xmit = lpc_eth_hard_start_xmit, 1236 .ndo_set_rx_mode = lpc_eth_set_multicast_list, 1237 .ndo_do_ioctl = lpc_eth_ioctl, 1238 .ndo_set_mac_address = lpc_set_mac_address, 1239 .ndo_validate_addr = eth_validate_addr, 1240 }; 1241 1242 static int lpc_eth_drv_probe(struct platform_device *pdev) 1243 { 1244 struct device *dev = &pdev->dev; 1245 struct device_node *np = dev->of_node; 1246 struct netdata_local *pldat; 1247 struct net_device *ndev; 1248 dma_addr_t dma_handle; 1249 struct resource *res; 1250 int irq, ret; 1251 u32 tmp; 1252 1253 /* Setup network interface for RMII or MII mode */ 1254 tmp = __raw_readl(LPC32XX_CLKPWR_MACCLK_CTRL); 1255 tmp &= ~LPC32XX_CLKPWR_MACCTRL_PINS_MSK; 1256 if (lpc_phy_interface_mode(dev) == PHY_INTERFACE_MODE_MII) 1257 tmp |= LPC32XX_CLKPWR_MACCTRL_USE_MII_PINS; 1258 else 1259 tmp |= LPC32XX_CLKPWR_MACCTRL_USE_RMII_PINS; 1260 __raw_writel(tmp, LPC32XX_CLKPWR_MACCLK_CTRL); 1261 1262 /* Get platform resources */ 1263 res = platform_get_resource(pdev, IORESOURCE_MEM, 0); 1264 irq = platform_get_irq(pdev, 0); 1265 if (!res || irq < 0) { 1266 dev_err(dev, "error getting resources.\n"); 1267 ret = -ENXIO; 1268 goto err_exit; 1269 } 1270 1271 /* Allocate net driver data structure */ 1272 ndev = alloc_etherdev(sizeof(struct netdata_local)); 1273 if (!ndev) { 1274 dev_err(dev, "could not allocate device.\n"); 1275 ret = -ENOMEM; 1276 goto err_exit; 1277 } 1278 1279 SET_NETDEV_DEV(ndev, dev); 1280 1281 pldat = netdev_priv(ndev); 1282 pldat->pdev = pdev; 1283 pldat->ndev = ndev; 1284 1285 spin_lock_init(&pldat->lock); 1286 1287 /* Save resources */ 1288 ndev->irq = irq; 1289 1290 /* Get clock for the device */ 1291 pldat->clk = clk_get(dev, NULL); 1292 if (IS_ERR(pldat->clk)) { 1293 dev_err(dev, "error getting clock.\n"); 1294 ret = PTR_ERR(pldat->clk); 1295 goto err_out_free_dev; 1296 } 1297 1298 /* Enable network clock */ 1299 ret = clk_prepare_enable(pldat->clk); 1300 if (ret) 1301 goto err_out_clk_put; 1302 1303 /* Map IO space */ 1304 pldat->net_base = ioremap(res->start, resource_size(res)); 1305 if (!pldat->net_base) { 1306 dev_err(dev, "failed to map registers\n"); 1307 ret = -ENOMEM; 1308 goto err_out_disable_clocks; 1309 } 1310 ret = request_irq(ndev->irq, __lpc_eth_interrupt, 0, 1311 ndev->name, ndev); 1312 if (ret) { 1313 dev_err(dev, "error requesting interrupt.\n"); 1314 goto err_out_iounmap; 1315 } 1316 1317 /* Setup driver functions */ 1318 ndev->netdev_ops = &lpc_netdev_ops; 1319 ndev->ethtool_ops = &lpc_eth_ethtool_ops; 1320 ndev->watchdog_timeo = msecs_to_jiffies(2500); 1321 1322 /* Get size of DMA buffers/descriptors region */ 1323 pldat->dma_buff_size = (ENET_TX_DESC + ENET_RX_DESC) * (ENET_MAXF_SIZE + 1324 sizeof(struct txrx_desc_t) + sizeof(struct rx_status_t)); 1325 pldat->dma_buff_base_v = 0; 1326 1327 if (use_iram_for_net(dev)) { 1328 dma_handle = LPC32XX_IRAM_BASE; 1329 if (pldat->dma_buff_size <= lpc32xx_return_iram_size()) 1330 pldat->dma_buff_base_v = 1331 io_p2v(LPC32XX_IRAM_BASE); 1332 else 1333 netdev_err(ndev, 1334 "IRAM not big enough for net buffers, using SDRAM instead.\n"); 1335 } 1336 1337 if (pldat->dma_buff_base_v == 0) { 1338 ret = dma_coerce_mask_and_coherent(dev, DMA_BIT_MASK(32)); 1339 if (ret) 1340 goto err_out_free_irq; 1341 1342 pldat->dma_buff_size = PAGE_ALIGN(pldat->dma_buff_size); 1343 1344 /* Allocate a chunk of memory for the DMA ethernet buffers 1345 and descriptors */ 1346 pldat->dma_buff_base_v = 1347 dma_alloc_coherent(dev, 1348 pldat->dma_buff_size, &dma_handle, 1349 GFP_KERNEL); 1350 if (pldat->dma_buff_base_v == NULL) { 1351 ret = -ENOMEM; 1352 goto err_out_free_irq; 1353 } 1354 } 1355 pldat->dma_buff_base_p = dma_handle; 1356 1357 netdev_dbg(ndev, "IO address space :%pR\n", res); 1358 netdev_dbg(ndev, "IO address size :%d\n", resource_size(res)); 1359 netdev_dbg(ndev, "IO address (mapped) :0x%p\n", 1360 pldat->net_base); 1361 netdev_dbg(ndev, "IRQ number :%d\n", ndev->irq); 1362 netdev_dbg(ndev, "DMA buffer size :%d\n", pldat->dma_buff_size); 1363 netdev_dbg(ndev, "DMA buffer P address :0x%08x\n", 1364 pldat->dma_buff_base_p); 1365 netdev_dbg(ndev, "DMA buffer V address :0x%p\n", 1366 pldat->dma_buff_base_v); 1367 1368 /* Get MAC address from current HW setting (POR state is all zeros) */ 1369 __lpc_get_mac(pldat, ndev->dev_addr); 1370 1371 if (!is_valid_ether_addr(ndev->dev_addr)) { 1372 const char *macaddr = of_get_mac_address(np); 1373 if (macaddr) 1374 memcpy(ndev->dev_addr, macaddr, ETH_ALEN); 1375 } 1376 if (!is_valid_ether_addr(ndev->dev_addr)) 1377 eth_hw_addr_random(ndev); 1378 1379 /* Reset the ethernet controller */ 1380 __lpc_eth_reset(pldat); 1381 1382 /* then shut everything down to save power */ 1383 __lpc_eth_shutdown(pldat); 1384 1385 /* Set default parameters */ 1386 pldat->msg_enable = NETIF_MSG_LINK; 1387 1388 /* Force an MII interface reset and clock setup */ 1389 __lpc_mii_mngt_reset(pldat); 1390 1391 /* Force default PHY interface setup in chip, this will probably be 1392 changed by the PHY driver */ 1393 pldat->link = 0; 1394 pldat->speed = 100; 1395 pldat->duplex = DUPLEX_FULL; 1396 __lpc_params_setup(pldat); 1397 1398 netif_napi_add(ndev, &pldat->napi, lpc_eth_poll, NAPI_WEIGHT); 1399 1400 ret = register_netdev(ndev); 1401 if (ret) { 1402 dev_err(dev, "Cannot register net device, aborting.\n"); 1403 goto err_out_dma_unmap; 1404 } 1405 platform_set_drvdata(pdev, ndev); 1406 1407 ret = lpc_mii_init(pldat); 1408 if (ret) 1409 goto err_out_unregister_netdev; 1410 1411 netdev_info(ndev, "LPC mac at 0x%08x irq %d\n", 1412 res->start, ndev->irq); 1413 1414 device_init_wakeup(dev, 1); 1415 device_set_wakeup_enable(dev, 0); 1416 1417 return 0; 1418 1419 err_out_unregister_netdev: 1420 unregister_netdev(ndev); 1421 err_out_dma_unmap: 1422 if (!use_iram_for_net(dev) || 1423 pldat->dma_buff_size > lpc32xx_return_iram_size()) 1424 dma_free_coherent(dev, pldat->dma_buff_size, 1425 pldat->dma_buff_base_v, 1426 pldat->dma_buff_base_p); 1427 err_out_free_irq: 1428 free_irq(ndev->irq, ndev); 1429 err_out_iounmap: 1430 iounmap(pldat->net_base); 1431 err_out_disable_clocks: 1432 clk_disable_unprepare(pldat->clk); 1433 err_out_clk_put: 1434 clk_put(pldat->clk); 1435 err_out_free_dev: 1436 free_netdev(ndev); 1437 err_exit: 1438 pr_err("%s: not found (%d).\n", MODNAME, ret); 1439 return ret; 1440 } 1441 1442 static int lpc_eth_drv_remove(struct platform_device *pdev) 1443 { 1444 struct net_device *ndev = platform_get_drvdata(pdev); 1445 struct netdata_local *pldat = netdev_priv(ndev); 1446 1447 unregister_netdev(ndev); 1448 1449 if (!use_iram_for_net(&pldat->pdev->dev) || 1450 pldat->dma_buff_size > lpc32xx_return_iram_size()) 1451 dma_free_coherent(&pldat->pdev->dev, pldat->dma_buff_size, 1452 pldat->dma_buff_base_v, 1453 pldat->dma_buff_base_p); 1454 free_irq(ndev->irq, ndev); 1455 iounmap(pldat->net_base); 1456 mdiobus_unregister(pldat->mii_bus); 1457 mdiobus_free(pldat->mii_bus); 1458 clk_disable_unprepare(pldat->clk); 1459 clk_put(pldat->clk); 1460 free_netdev(ndev); 1461 1462 return 0; 1463 } 1464 1465 #ifdef CONFIG_PM 1466 static int lpc_eth_drv_suspend(struct platform_device *pdev, 1467 pm_message_t state) 1468 { 1469 struct net_device *ndev = platform_get_drvdata(pdev); 1470 struct netdata_local *pldat = netdev_priv(ndev); 1471 1472 if (device_may_wakeup(&pdev->dev)) 1473 enable_irq_wake(ndev->irq); 1474 1475 if (ndev) { 1476 if (netif_running(ndev)) { 1477 netif_device_detach(ndev); 1478 __lpc_eth_shutdown(pldat); 1479 clk_disable_unprepare(pldat->clk); 1480 1481 /* 1482 * Reset again now clock is disable to be sure 1483 * EMC_MDC is down 1484 */ 1485 __lpc_eth_reset(pldat); 1486 } 1487 } 1488 1489 return 0; 1490 } 1491 1492 static int lpc_eth_drv_resume(struct platform_device *pdev) 1493 { 1494 struct net_device *ndev = platform_get_drvdata(pdev); 1495 struct netdata_local *pldat; 1496 1497 if (device_may_wakeup(&pdev->dev)) 1498 disable_irq_wake(ndev->irq); 1499 1500 if (ndev) { 1501 if (netif_running(ndev)) { 1502 pldat = netdev_priv(ndev); 1503 1504 /* Enable interface clock */ 1505 clk_enable(pldat->clk); 1506 1507 /* Reset and initialize */ 1508 __lpc_eth_reset(pldat); 1509 __lpc_eth_init(pldat); 1510 1511 netif_device_attach(ndev); 1512 } 1513 } 1514 1515 return 0; 1516 } 1517 #endif 1518 1519 static const struct of_device_id lpc_eth_match[] = { 1520 { .compatible = "nxp,lpc-eth" }, 1521 { } 1522 }; 1523 MODULE_DEVICE_TABLE(of, lpc_eth_match); 1524 1525 static struct platform_driver lpc_eth_driver = { 1526 .probe = lpc_eth_drv_probe, 1527 .remove = lpc_eth_drv_remove, 1528 #ifdef CONFIG_PM 1529 .suspend = lpc_eth_drv_suspend, 1530 .resume = lpc_eth_drv_resume, 1531 #endif 1532 .driver = { 1533 .name = MODNAME, 1534 .of_match_table = lpc_eth_match, 1535 }, 1536 }; 1537 1538 module_platform_driver(lpc_eth_driver); 1539 1540 MODULE_AUTHOR("Kevin Wells <kevin.wells@nxp.com>"); 1541 MODULE_AUTHOR("Roland Stigge <stigge@antcom.de>"); 1542 MODULE_DESCRIPTION("LPC Ethernet Driver"); 1543 MODULE_LICENSE("GPL"); 1544