1 // SPDX-License-Identifier: GPL-2.0-only 2 /* 3 * ks8842.c timberdale KS8842 ethernet driver 4 * Copyright (c) 2009 Intel Corporation 5 */ 6 7 /* Supports: 8 * The Micrel KS8842 behind the timberdale FPGA 9 * The genuine Micrel KS8841/42 device with ISA 16/32bit bus interface 10 */ 11 12 #define pr_fmt(fmt) KBUILD_MODNAME ": " fmt 13 14 #include <linux/interrupt.h> 15 #include <linux/kernel.h> 16 #include <linux/module.h> 17 #include <linux/platform_device.h> 18 #include <linux/netdevice.h> 19 #include <linux/etherdevice.h> 20 #include <linux/ethtool.h> 21 #include <linux/ks8842.h> 22 #include <linux/dmaengine.h> 23 #include <linux/dma-mapping.h> 24 #include <linux/scatterlist.h> 25 26 #define DRV_NAME "ks8842" 27 28 /* Timberdale specific Registers */ 29 #define REG_TIMB_RST 0x1c 30 #define REG_TIMB_FIFO 0x20 31 #define REG_TIMB_ISR 0x24 32 #define REG_TIMB_IER 0x28 33 #define REG_TIMB_IAR 0x2C 34 #define REQ_TIMB_DMA_RESUME 0x30 35 36 /* KS8842 registers */ 37 38 #define REG_SELECT_BANK 0x0e 39 40 /* bank 0 registers */ 41 #define REG_QRFCR 0x04 42 43 /* bank 2 registers */ 44 #define REG_MARL 0x00 45 #define REG_MARM 0x02 46 #define REG_MARH 0x04 47 48 /* bank 3 registers */ 49 #define REG_GRR 0x06 50 51 /* bank 16 registers */ 52 #define REG_TXCR 0x00 53 #define REG_TXSR 0x02 54 #define REG_RXCR 0x04 55 #define REG_TXMIR 0x08 56 #define REG_RXMIR 0x0A 57 58 /* bank 17 registers */ 59 #define REG_TXQCR 0x00 60 #define REG_RXQCR 0x02 61 #define REG_TXFDPR 0x04 62 #define REG_RXFDPR 0x06 63 #define REG_QMU_DATA_LO 0x08 64 #define REG_QMU_DATA_HI 0x0A 65 66 /* bank 18 registers */ 67 #define REG_IER 0x00 68 #define IRQ_LINK_CHANGE 0x8000 69 #define IRQ_TX 0x4000 70 #define IRQ_RX 0x2000 71 #define IRQ_RX_OVERRUN 0x0800 72 #define IRQ_TX_STOPPED 0x0200 73 #define IRQ_RX_STOPPED 0x0100 74 #define IRQ_RX_ERROR 0x0080 75 #define ENABLED_IRQS (IRQ_LINK_CHANGE | IRQ_TX | IRQ_RX | IRQ_RX_STOPPED | \ 76 IRQ_TX_STOPPED | IRQ_RX_OVERRUN | IRQ_RX_ERROR) 77 /* When running via timberdale in DMA mode, the RX interrupt should be 78 enabled in the KS8842, but not in the FPGA IP, since the IP handles 79 RX DMA internally. 80 TX interrupts are not needed it is handled by the FPGA the driver is 81 notified via DMA callbacks. 82 */ 83 #define ENABLED_IRQS_DMA_IP (IRQ_LINK_CHANGE | IRQ_RX_STOPPED | \ 84 IRQ_TX_STOPPED | IRQ_RX_OVERRUN | IRQ_RX_ERROR) 85 #define ENABLED_IRQS_DMA (ENABLED_IRQS_DMA_IP | IRQ_RX) 86 #define REG_ISR 0x02 87 #define REG_RXSR 0x04 88 #define RXSR_VALID 0x8000 89 #define RXSR_BROADCAST 0x80 90 #define RXSR_MULTICAST 0x40 91 #define RXSR_UNICAST 0x20 92 #define RXSR_FRAMETYPE 0x08 93 #define RXSR_TOO_LONG 0x04 94 #define RXSR_RUNT 0x02 95 #define RXSR_CRC_ERROR 0x01 96 #define RXSR_ERROR (RXSR_TOO_LONG | RXSR_RUNT | RXSR_CRC_ERROR) 97 98 /* bank 32 registers */ 99 #define REG_SW_ID_AND_ENABLE 0x00 100 #define REG_SGCR1 0x02 101 #define REG_SGCR2 0x04 102 #define REG_SGCR3 0x06 103 104 /* bank 39 registers */ 105 #define REG_MACAR1 0x00 106 #define REG_MACAR2 0x02 107 #define REG_MACAR3 0x04 108 109 /* bank 45 registers */ 110 #define REG_P1MBCR 0x00 111 #define REG_P1MBSR 0x02 112 113 /* bank 46 registers */ 114 #define REG_P2MBCR 0x00 115 #define REG_P2MBSR 0x02 116 117 /* bank 48 registers */ 118 #define REG_P1CR2 0x02 119 120 /* bank 49 registers */ 121 #define REG_P1CR4 0x02 122 #define REG_P1SR 0x04 123 124 /* flags passed by platform_device for configuration */ 125 #define MICREL_KS884X 0x01 /* 0=Timeberdale(FPGA), 1=Micrel */ 126 #define KS884X_16BIT 0x02 /* 1=16bit, 0=32bit */ 127 128 #define DMA_BUFFER_SIZE 2048 129 130 struct ks8842_tx_dma_ctl { 131 struct dma_chan *chan; 132 struct dma_async_tx_descriptor *adesc; 133 void *buf; 134 struct scatterlist sg; 135 int channel; 136 }; 137 138 struct ks8842_rx_dma_ctl { 139 struct dma_chan *chan; 140 struct dma_async_tx_descriptor *adesc; 141 struct sk_buff *skb; 142 struct scatterlist sg; 143 struct tasklet_struct tasklet; 144 int channel; 145 }; 146 147 #define KS8842_USE_DMA(adapter) (((adapter)->dma_tx.channel != -1) && \ 148 ((adapter)->dma_rx.channel != -1)) 149 150 struct ks8842_adapter { 151 void __iomem *hw_addr; 152 int irq; 153 unsigned long conf_flags; /* copy of platform_device config */ 154 struct tasklet_struct tasklet; 155 spinlock_t lock; /* spinlock to be interrupt safe */ 156 struct work_struct timeout_work; 157 struct net_device *netdev; 158 struct device *dev; 159 struct ks8842_tx_dma_ctl dma_tx; 160 struct ks8842_rx_dma_ctl dma_rx; 161 }; 162 163 static void ks8842_dma_rx_cb(void *data); 164 static void ks8842_dma_tx_cb(void *data); 165 166 static inline void ks8842_resume_dma(struct ks8842_adapter *adapter) 167 { 168 iowrite32(1, adapter->hw_addr + REQ_TIMB_DMA_RESUME); 169 } 170 171 static inline void ks8842_select_bank(struct ks8842_adapter *adapter, u16 bank) 172 { 173 iowrite16(bank, adapter->hw_addr + REG_SELECT_BANK); 174 } 175 176 static inline void ks8842_write8(struct ks8842_adapter *adapter, u16 bank, 177 u8 value, int offset) 178 { 179 ks8842_select_bank(adapter, bank); 180 iowrite8(value, adapter->hw_addr + offset); 181 } 182 183 static inline void ks8842_write16(struct ks8842_adapter *adapter, u16 bank, 184 u16 value, int offset) 185 { 186 ks8842_select_bank(adapter, bank); 187 iowrite16(value, adapter->hw_addr + offset); 188 } 189 190 static inline void ks8842_enable_bits(struct ks8842_adapter *adapter, u16 bank, 191 u16 bits, int offset) 192 { 193 u16 reg; 194 ks8842_select_bank(adapter, bank); 195 reg = ioread16(adapter->hw_addr + offset); 196 reg |= bits; 197 iowrite16(reg, adapter->hw_addr + offset); 198 } 199 200 static inline void ks8842_clear_bits(struct ks8842_adapter *adapter, u16 bank, 201 u16 bits, int offset) 202 { 203 u16 reg; 204 ks8842_select_bank(adapter, bank); 205 reg = ioread16(adapter->hw_addr + offset); 206 reg &= ~bits; 207 iowrite16(reg, adapter->hw_addr + offset); 208 } 209 210 static inline void ks8842_write32(struct ks8842_adapter *adapter, u16 bank, 211 u32 value, int offset) 212 { 213 ks8842_select_bank(adapter, bank); 214 iowrite32(value, adapter->hw_addr + offset); 215 } 216 217 static inline u8 ks8842_read8(struct ks8842_adapter *adapter, u16 bank, 218 int offset) 219 { 220 ks8842_select_bank(adapter, bank); 221 return ioread8(adapter->hw_addr + offset); 222 } 223 224 static inline u16 ks8842_read16(struct ks8842_adapter *adapter, u16 bank, 225 int offset) 226 { 227 ks8842_select_bank(adapter, bank); 228 return ioread16(adapter->hw_addr + offset); 229 } 230 231 static inline u32 ks8842_read32(struct ks8842_adapter *adapter, u16 bank, 232 int offset) 233 { 234 ks8842_select_bank(adapter, bank); 235 return ioread32(adapter->hw_addr + offset); 236 } 237 238 static void ks8842_reset(struct ks8842_adapter *adapter) 239 { 240 if (adapter->conf_flags & MICREL_KS884X) { 241 ks8842_write16(adapter, 3, 1, REG_GRR); 242 msleep(10); 243 iowrite16(0, adapter->hw_addr + REG_GRR); 244 } else { 245 /* The KS8842 goes haywire when doing softare reset 246 * a work around in the timberdale IP is implemented to 247 * do a hardware reset instead 248 ks8842_write16(adapter, 3, 1, REG_GRR); 249 msleep(10); 250 iowrite16(0, adapter->hw_addr + REG_GRR); 251 */ 252 iowrite32(0x1, adapter->hw_addr + REG_TIMB_RST); 253 msleep(20); 254 } 255 } 256 257 static void ks8842_update_link_status(struct net_device *netdev, 258 struct ks8842_adapter *adapter) 259 { 260 /* check the status of the link */ 261 if (ks8842_read16(adapter, 45, REG_P1MBSR) & 0x4) { 262 netif_carrier_on(netdev); 263 netif_wake_queue(netdev); 264 } else { 265 netif_stop_queue(netdev); 266 netif_carrier_off(netdev); 267 } 268 } 269 270 static void ks8842_enable_tx(struct ks8842_adapter *adapter) 271 { 272 ks8842_enable_bits(adapter, 16, 0x01, REG_TXCR); 273 } 274 275 static void ks8842_disable_tx(struct ks8842_adapter *adapter) 276 { 277 ks8842_clear_bits(adapter, 16, 0x01, REG_TXCR); 278 } 279 280 static void ks8842_enable_rx(struct ks8842_adapter *adapter) 281 { 282 ks8842_enable_bits(adapter, 16, 0x01, REG_RXCR); 283 } 284 285 static void ks8842_disable_rx(struct ks8842_adapter *adapter) 286 { 287 ks8842_clear_bits(adapter, 16, 0x01, REG_RXCR); 288 } 289 290 static void ks8842_reset_hw(struct ks8842_adapter *adapter) 291 { 292 /* reset the HW */ 293 ks8842_reset(adapter); 294 295 /* Enable QMU Transmit flow control / transmit padding / Transmit CRC */ 296 ks8842_write16(adapter, 16, 0x000E, REG_TXCR); 297 298 /* enable the receiver, uni + multi + broadcast + flow ctrl 299 + crc strip */ 300 ks8842_write16(adapter, 16, 0x8 | 0x20 | 0x40 | 0x80 | 0x400, 301 REG_RXCR); 302 303 /* TX frame pointer autoincrement */ 304 ks8842_write16(adapter, 17, 0x4000, REG_TXFDPR); 305 306 /* RX frame pointer autoincrement */ 307 ks8842_write16(adapter, 17, 0x4000, REG_RXFDPR); 308 309 /* RX 2 kb high watermark */ 310 ks8842_write16(adapter, 0, 0x1000, REG_QRFCR); 311 312 /* aggressive back off in half duplex */ 313 ks8842_enable_bits(adapter, 32, 1 << 8, REG_SGCR1); 314 315 /* enable no excessive collison drop */ 316 ks8842_enable_bits(adapter, 32, 1 << 3, REG_SGCR2); 317 318 /* Enable port 1 force flow control / back pressure / transmit / recv */ 319 ks8842_write16(adapter, 48, 0x1E07, REG_P1CR2); 320 321 /* restart port auto-negotiation */ 322 ks8842_enable_bits(adapter, 49, 1 << 13, REG_P1CR4); 323 324 /* Enable the transmitter */ 325 ks8842_enable_tx(adapter); 326 327 /* Enable the receiver */ 328 ks8842_enable_rx(adapter); 329 330 /* clear all interrupts */ 331 ks8842_write16(adapter, 18, 0xffff, REG_ISR); 332 333 /* enable interrupts */ 334 if (KS8842_USE_DMA(adapter)) { 335 /* When running in DMA Mode the RX interrupt is not enabled in 336 timberdale because RX data is received by DMA callbacks 337 it must still be enabled in the KS8842 because it indicates 338 to timberdale when there is RX data for it's DMA FIFOs */ 339 iowrite16(ENABLED_IRQS_DMA_IP, adapter->hw_addr + REG_TIMB_IER); 340 ks8842_write16(adapter, 18, ENABLED_IRQS_DMA, REG_IER); 341 } else { 342 if (!(adapter->conf_flags & MICREL_KS884X)) 343 iowrite16(ENABLED_IRQS, 344 adapter->hw_addr + REG_TIMB_IER); 345 ks8842_write16(adapter, 18, ENABLED_IRQS, REG_IER); 346 } 347 /* enable the switch */ 348 ks8842_write16(adapter, 32, 0x1, REG_SW_ID_AND_ENABLE); 349 } 350 351 static void ks8842_read_mac_addr(struct ks8842_adapter *adapter, u8 *dest) 352 { 353 int i; 354 u16 mac; 355 356 for (i = 0; i < ETH_ALEN; i++) 357 dest[ETH_ALEN - i - 1] = ks8842_read8(adapter, 2, REG_MARL + i); 358 359 if (adapter->conf_flags & MICREL_KS884X) { 360 /* 361 the sequence of saving mac addr between MAC and Switch is 362 different. 363 */ 364 365 mac = ks8842_read16(adapter, 2, REG_MARL); 366 ks8842_write16(adapter, 39, mac, REG_MACAR3); 367 mac = ks8842_read16(adapter, 2, REG_MARM); 368 ks8842_write16(adapter, 39, mac, REG_MACAR2); 369 mac = ks8842_read16(adapter, 2, REG_MARH); 370 ks8842_write16(adapter, 39, mac, REG_MACAR1); 371 } else { 372 373 /* make sure the switch port uses the same MAC as the QMU */ 374 mac = ks8842_read16(adapter, 2, REG_MARL); 375 ks8842_write16(adapter, 39, mac, REG_MACAR1); 376 mac = ks8842_read16(adapter, 2, REG_MARM); 377 ks8842_write16(adapter, 39, mac, REG_MACAR2); 378 mac = ks8842_read16(adapter, 2, REG_MARH); 379 ks8842_write16(adapter, 39, mac, REG_MACAR3); 380 } 381 } 382 383 static void ks8842_write_mac_addr(struct ks8842_adapter *adapter, u8 *mac) 384 { 385 unsigned long flags; 386 unsigned i; 387 388 spin_lock_irqsave(&adapter->lock, flags); 389 for (i = 0; i < ETH_ALEN; i++) { 390 ks8842_write8(adapter, 2, mac[ETH_ALEN - i - 1], REG_MARL + i); 391 if (!(adapter->conf_flags & MICREL_KS884X)) 392 ks8842_write8(adapter, 39, mac[ETH_ALEN - i - 1], 393 REG_MACAR1 + i); 394 } 395 396 if (adapter->conf_flags & MICREL_KS884X) { 397 /* 398 the sequence of saving mac addr between MAC and Switch is 399 different. 400 */ 401 402 u16 mac; 403 404 mac = ks8842_read16(adapter, 2, REG_MARL); 405 ks8842_write16(adapter, 39, mac, REG_MACAR3); 406 mac = ks8842_read16(adapter, 2, REG_MARM); 407 ks8842_write16(adapter, 39, mac, REG_MACAR2); 408 mac = ks8842_read16(adapter, 2, REG_MARH); 409 ks8842_write16(adapter, 39, mac, REG_MACAR1); 410 } 411 spin_unlock_irqrestore(&adapter->lock, flags); 412 } 413 414 static inline u16 ks8842_tx_fifo_space(struct ks8842_adapter *adapter) 415 { 416 return ks8842_read16(adapter, 16, REG_TXMIR) & 0x1fff; 417 } 418 419 static int ks8842_tx_frame_dma(struct sk_buff *skb, struct net_device *netdev) 420 { 421 struct ks8842_adapter *adapter = netdev_priv(netdev); 422 struct ks8842_tx_dma_ctl *ctl = &adapter->dma_tx; 423 u8 *buf = ctl->buf; 424 425 if (ctl->adesc) { 426 netdev_dbg(netdev, "%s: TX ongoing\n", __func__); 427 /* transfer ongoing */ 428 return NETDEV_TX_BUSY; 429 } 430 431 sg_dma_len(&ctl->sg) = skb->len + sizeof(u32); 432 433 /* copy data to the TX buffer */ 434 /* the control word, enable IRQ, port 1 and the length */ 435 *buf++ = 0x00; 436 *buf++ = 0x01; /* Port 1 */ 437 *buf++ = skb->len & 0xff; 438 *buf++ = (skb->len >> 8) & 0xff; 439 skb_copy_from_linear_data(skb, buf, skb->len); 440 441 dma_sync_single_range_for_device(adapter->dev, 442 sg_dma_address(&ctl->sg), 0, sg_dma_len(&ctl->sg), 443 DMA_TO_DEVICE); 444 445 /* make sure the length is a multiple of 4 */ 446 if (sg_dma_len(&ctl->sg) % 4) 447 sg_dma_len(&ctl->sg) += 4 - sg_dma_len(&ctl->sg) % 4; 448 449 ctl->adesc = dmaengine_prep_slave_sg(ctl->chan, 450 &ctl->sg, 1, DMA_MEM_TO_DEV, DMA_PREP_INTERRUPT); 451 if (!ctl->adesc) 452 return NETDEV_TX_BUSY; 453 454 ctl->adesc->callback_param = netdev; 455 ctl->adesc->callback = ks8842_dma_tx_cb; 456 ctl->adesc->tx_submit(ctl->adesc); 457 458 netdev->stats.tx_bytes += skb->len; 459 460 dev_kfree_skb(skb); 461 462 return NETDEV_TX_OK; 463 } 464 465 static int ks8842_tx_frame(struct sk_buff *skb, struct net_device *netdev) 466 { 467 struct ks8842_adapter *adapter = netdev_priv(netdev); 468 int len = skb->len; 469 470 netdev_dbg(netdev, "%s: len %u head %p data %p tail %p end %p\n", 471 __func__, skb->len, skb->head, skb->data, 472 skb_tail_pointer(skb), skb_end_pointer(skb)); 473 474 /* check FIFO buffer space, we need space for CRC and command bits */ 475 if (ks8842_tx_fifo_space(adapter) < len + 8) 476 return NETDEV_TX_BUSY; 477 478 if (adapter->conf_flags & KS884X_16BIT) { 479 u16 *ptr16 = (u16 *)skb->data; 480 ks8842_write16(adapter, 17, 0x8000 | 0x100, REG_QMU_DATA_LO); 481 ks8842_write16(adapter, 17, (u16)len, REG_QMU_DATA_HI); 482 netdev->stats.tx_bytes += len; 483 484 /* copy buffer */ 485 while (len > 0) { 486 iowrite16(*ptr16++, adapter->hw_addr + REG_QMU_DATA_LO); 487 iowrite16(*ptr16++, adapter->hw_addr + REG_QMU_DATA_HI); 488 len -= sizeof(u32); 489 } 490 } else { 491 492 u32 *ptr = (u32 *)skb->data; 493 u32 ctrl; 494 /* the control word, enable IRQ, port 1 and the length */ 495 ctrl = 0x8000 | 0x100 | (len << 16); 496 ks8842_write32(adapter, 17, ctrl, REG_QMU_DATA_LO); 497 498 netdev->stats.tx_bytes += len; 499 500 /* copy buffer */ 501 while (len > 0) { 502 iowrite32(*ptr, adapter->hw_addr + REG_QMU_DATA_LO); 503 len -= sizeof(u32); 504 ptr++; 505 } 506 } 507 508 /* enqueue packet */ 509 ks8842_write16(adapter, 17, 1, REG_TXQCR); 510 511 dev_kfree_skb(skb); 512 513 return NETDEV_TX_OK; 514 } 515 516 static void ks8842_update_rx_err_counters(struct net_device *netdev, u32 status) 517 { 518 netdev_dbg(netdev, "RX error, status: %x\n", status); 519 520 netdev->stats.rx_errors++; 521 if (status & RXSR_TOO_LONG) 522 netdev->stats.rx_length_errors++; 523 if (status & RXSR_CRC_ERROR) 524 netdev->stats.rx_crc_errors++; 525 if (status & RXSR_RUNT) 526 netdev->stats.rx_frame_errors++; 527 } 528 529 static void ks8842_update_rx_counters(struct net_device *netdev, u32 status, 530 int len) 531 { 532 netdev_dbg(netdev, "RX packet, len: %d\n", len); 533 534 netdev->stats.rx_packets++; 535 netdev->stats.rx_bytes += len; 536 if (status & RXSR_MULTICAST) 537 netdev->stats.multicast++; 538 } 539 540 static int __ks8842_start_new_rx_dma(struct net_device *netdev) 541 { 542 struct ks8842_adapter *adapter = netdev_priv(netdev); 543 struct ks8842_rx_dma_ctl *ctl = &adapter->dma_rx; 544 struct scatterlist *sg = &ctl->sg; 545 int err; 546 547 ctl->skb = netdev_alloc_skb(netdev, DMA_BUFFER_SIZE); 548 if (ctl->skb) { 549 sg_init_table(sg, 1); 550 sg_dma_address(sg) = dma_map_single(adapter->dev, 551 ctl->skb->data, DMA_BUFFER_SIZE, DMA_FROM_DEVICE); 552 if (dma_mapping_error(adapter->dev, sg_dma_address(sg))) { 553 err = -ENOMEM; 554 sg_dma_address(sg) = 0; 555 goto out; 556 } 557 558 sg_dma_len(sg) = DMA_BUFFER_SIZE; 559 560 ctl->adesc = dmaengine_prep_slave_sg(ctl->chan, 561 sg, 1, DMA_DEV_TO_MEM, DMA_PREP_INTERRUPT); 562 563 if (!ctl->adesc) { 564 err = -ENOMEM; 565 goto out; 566 } 567 568 ctl->adesc->callback_param = netdev; 569 ctl->adesc->callback = ks8842_dma_rx_cb; 570 ctl->adesc->tx_submit(ctl->adesc); 571 } else { 572 err = -ENOMEM; 573 sg_dma_address(sg) = 0; 574 goto out; 575 } 576 577 return 0; 578 out: 579 if (sg_dma_address(sg)) 580 dma_unmap_single(adapter->dev, sg_dma_address(sg), 581 DMA_BUFFER_SIZE, DMA_FROM_DEVICE); 582 sg_dma_address(sg) = 0; 583 dev_kfree_skb(ctl->skb); 584 ctl->skb = NULL; 585 586 printk(KERN_ERR DRV_NAME": Failed to start RX DMA: %d\n", err); 587 return err; 588 } 589 590 static void ks8842_rx_frame_dma_tasklet(unsigned long arg) 591 { 592 struct net_device *netdev = (struct net_device *)arg; 593 struct ks8842_adapter *adapter = netdev_priv(netdev); 594 struct ks8842_rx_dma_ctl *ctl = &adapter->dma_rx; 595 struct sk_buff *skb = ctl->skb; 596 dma_addr_t addr = sg_dma_address(&ctl->sg); 597 u32 status; 598 599 ctl->adesc = NULL; 600 601 /* kick next transfer going */ 602 __ks8842_start_new_rx_dma(netdev); 603 604 /* now handle the data we got */ 605 dma_unmap_single(adapter->dev, addr, DMA_BUFFER_SIZE, DMA_FROM_DEVICE); 606 607 status = *((u32 *)skb->data); 608 609 netdev_dbg(netdev, "%s - rx_data: status: %x\n", 610 __func__, status & 0xffff); 611 612 /* check the status */ 613 if ((status & RXSR_VALID) && !(status & RXSR_ERROR)) { 614 int len = (status >> 16) & 0x7ff; 615 616 ks8842_update_rx_counters(netdev, status, len); 617 618 /* reserve 4 bytes which is the status word */ 619 skb_reserve(skb, 4); 620 skb_put(skb, len); 621 622 skb->protocol = eth_type_trans(skb, netdev); 623 netif_rx(skb); 624 } else { 625 ks8842_update_rx_err_counters(netdev, status); 626 dev_kfree_skb(skb); 627 } 628 } 629 630 static void ks8842_rx_frame(struct net_device *netdev, 631 struct ks8842_adapter *adapter) 632 { 633 u32 status; 634 int len; 635 636 if (adapter->conf_flags & KS884X_16BIT) { 637 status = ks8842_read16(adapter, 17, REG_QMU_DATA_LO); 638 len = ks8842_read16(adapter, 17, REG_QMU_DATA_HI); 639 netdev_dbg(netdev, "%s - rx_data: status: %x\n", 640 __func__, status); 641 } else { 642 status = ks8842_read32(adapter, 17, REG_QMU_DATA_LO); 643 len = (status >> 16) & 0x7ff; 644 status &= 0xffff; 645 netdev_dbg(netdev, "%s - rx_data: status: %x\n", 646 __func__, status); 647 } 648 649 /* check the status */ 650 if ((status & RXSR_VALID) && !(status & RXSR_ERROR)) { 651 struct sk_buff *skb = netdev_alloc_skb_ip_align(netdev, len + 3); 652 653 if (skb) { 654 655 ks8842_update_rx_counters(netdev, status, len); 656 657 if (adapter->conf_flags & KS884X_16BIT) { 658 u16 *data16 = skb_put(skb, len); 659 ks8842_select_bank(adapter, 17); 660 while (len > 0) { 661 *data16++ = ioread16(adapter->hw_addr + 662 REG_QMU_DATA_LO); 663 *data16++ = ioread16(adapter->hw_addr + 664 REG_QMU_DATA_HI); 665 len -= sizeof(u32); 666 } 667 } else { 668 u32 *data = skb_put(skb, len); 669 670 ks8842_select_bank(adapter, 17); 671 while (len > 0) { 672 *data++ = ioread32(adapter->hw_addr + 673 REG_QMU_DATA_LO); 674 len -= sizeof(u32); 675 } 676 } 677 skb->protocol = eth_type_trans(skb, netdev); 678 netif_rx(skb); 679 } else 680 netdev->stats.rx_dropped++; 681 } else 682 ks8842_update_rx_err_counters(netdev, status); 683 684 /* set high watermark to 3K */ 685 ks8842_clear_bits(adapter, 0, 1 << 12, REG_QRFCR); 686 687 /* release the frame */ 688 ks8842_write16(adapter, 17, 0x01, REG_RXQCR); 689 690 /* set high watermark to 2K */ 691 ks8842_enable_bits(adapter, 0, 1 << 12, REG_QRFCR); 692 } 693 694 static void ks8842_handle_rx(struct net_device *netdev, 695 struct ks8842_adapter *adapter) 696 { 697 u16 rx_data = ks8842_read16(adapter, 16, REG_RXMIR) & 0x1fff; 698 netdev_dbg(netdev, "%s Entry - rx_data: %d\n", __func__, rx_data); 699 while (rx_data) { 700 ks8842_rx_frame(netdev, adapter); 701 rx_data = ks8842_read16(adapter, 16, REG_RXMIR) & 0x1fff; 702 } 703 } 704 705 static void ks8842_handle_tx(struct net_device *netdev, 706 struct ks8842_adapter *adapter) 707 { 708 u16 sr = ks8842_read16(adapter, 16, REG_TXSR); 709 netdev_dbg(netdev, "%s - entry, sr: %x\n", __func__, sr); 710 netdev->stats.tx_packets++; 711 if (netif_queue_stopped(netdev)) 712 netif_wake_queue(netdev); 713 } 714 715 static void ks8842_handle_rx_overrun(struct net_device *netdev, 716 struct ks8842_adapter *adapter) 717 { 718 netdev_dbg(netdev, "%s: entry\n", __func__); 719 netdev->stats.rx_errors++; 720 netdev->stats.rx_fifo_errors++; 721 } 722 723 static void ks8842_tasklet(unsigned long arg) 724 { 725 struct net_device *netdev = (struct net_device *)arg; 726 struct ks8842_adapter *adapter = netdev_priv(netdev); 727 u16 isr; 728 unsigned long flags; 729 u16 entry_bank; 730 731 /* read current bank to be able to set it back */ 732 spin_lock_irqsave(&adapter->lock, flags); 733 entry_bank = ioread16(adapter->hw_addr + REG_SELECT_BANK); 734 spin_unlock_irqrestore(&adapter->lock, flags); 735 736 isr = ks8842_read16(adapter, 18, REG_ISR); 737 netdev_dbg(netdev, "%s - ISR: 0x%x\n", __func__, isr); 738 739 /* when running in DMA mode, do not ack RX interrupts, it is handled 740 internally by timberdale, otherwise it's DMA FIFO:s would stop 741 */ 742 if (KS8842_USE_DMA(adapter)) 743 isr &= ~IRQ_RX; 744 745 /* Ack */ 746 ks8842_write16(adapter, 18, isr, REG_ISR); 747 748 if (!(adapter->conf_flags & MICREL_KS884X)) 749 /* Ack in the timberdale IP as well */ 750 iowrite32(0x1, adapter->hw_addr + REG_TIMB_IAR); 751 752 if (!netif_running(netdev)) 753 return; 754 755 if (isr & IRQ_LINK_CHANGE) 756 ks8842_update_link_status(netdev, adapter); 757 758 /* should not get IRQ_RX when running DMA mode */ 759 if (isr & (IRQ_RX | IRQ_RX_ERROR) && !KS8842_USE_DMA(adapter)) 760 ks8842_handle_rx(netdev, adapter); 761 762 /* should only happen when in PIO mode */ 763 if (isr & IRQ_TX) 764 ks8842_handle_tx(netdev, adapter); 765 766 if (isr & IRQ_RX_OVERRUN) 767 ks8842_handle_rx_overrun(netdev, adapter); 768 769 if (isr & IRQ_TX_STOPPED) { 770 ks8842_disable_tx(adapter); 771 ks8842_enable_tx(adapter); 772 } 773 774 if (isr & IRQ_RX_STOPPED) { 775 ks8842_disable_rx(adapter); 776 ks8842_enable_rx(adapter); 777 } 778 779 /* re-enable interrupts, put back the bank selection register */ 780 spin_lock_irqsave(&adapter->lock, flags); 781 if (KS8842_USE_DMA(adapter)) 782 ks8842_write16(adapter, 18, ENABLED_IRQS_DMA, REG_IER); 783 else 784 ks8842_write16(adapter, 18, ENABLED_IRQS, REG_IER); 785 iowrite16(entry_bank, adapter->hw_addr + REG_SELECT_BANK); 786 787 /* Make sure timberdale continues DMA operations, they are stopped while 788 we are handling the ks8842 because we might change bank */ 789 if (KS8842_USE_DMA(adapter)) 790 ks8842_resume_dma(adapter); 791 792 spin_unlock_irqrestore(&adapter->lock, flags); 793 } 794 795 static irqreturn_t ks8842_irq(int irq, void *devid) 796 { 797 struct net_device *netdev = devid; 798 struct ks8842_adapter *adapter = netdev_priv(netdev); 799 u16 isr; 800 u16 entry_bank = ioread16(adapter->hw_addr + REG_SELECT_BANK); 801 irqreturn_t ret = IRQ_NONE; 802 803 isr = ks8842_read16(adapter, 18, REG_ISR); 804 netdev_dbg(netdev, "%s - ISR: 0x%x\n", __func__, isr); 805 806 if (isr) { 807 if (KS8842_USE_DMA(adapter)) 808 /* disable all but RX IRQ, since the FPGA relies on it*/ 809 ks8842_write16(adapter, 18, IRQ_RX, REG_IER); 810 else 811 /* disable IRQ */ 812 ks8842_write16(adapter, 18, 0x00, REG_IER); 813 814 /* schedule tasklet */ 815 tasklet_schedule(&adapter->tasklet); 816 817 ret = IRQ_HANDLED; 818 } 819 820 iowrite16(entry_bank, adapter->hw_addr + REG_SELECT_BANK); 821 822 /* After an interrupt, tell timberdale to continue DMA operations. 823 DMA is disabled while we are handling the ks8842 because we might 824 change bank */ 825 ks8842_resume_dma(adapter); 826 827 return ret; 828 } 829 830 static void ks8842_dma_rx_cb(void *data) 831 { 832 struct net_device *netdev = data; 833 struct ks8842_adapter *adapter = netdev_priv(netdev); 834 835 netdev_dbg(netdev, "RX DMA finished\n"); 836 /* schedule tasklet */ 837 if (adapter->dma_rx.adesc) 838 tasklet_schedule(&adapter->dma_rx.tasklet); 839 } 840 841 static void ks8842_dma_tx_cb(void *data) 842 { 843 struct net_device *netdev = data; 844 struct ks8842_adapter *adapter = netdev_priv(netdev); 845 struct ks8842_tx_dma_ctl *ctl = &adapter->dma_tx; 846 847 netdev_dbg(netdev, "TX DMA finished\n"); 848 849 if (!ctl->adesc) 850 return; 851 852 netdev->stats.tx_packets++; 853 ctl->adesc = NULL; 854 855 if (netif_queue_stopped(netdev)) 856 netif_wake_queue(netdev); 857 } 858 859 static void ks8842_stop_dma(struct ks8842_adapter *adapter) 860 { 861 struct ks8842_tx_dma_ctl *tx_ctl = &adapter->dma_tx; 862 struct ks8842_rx_dma_ctl *rx_ctl = &adapter->dma_rx; 863 864 tx_ctl->adesc = NULL; 865 if (tx_ctl->chan) 866 dmaengine_terminate_all(tx_ctl->chan); 867 868 rx_ctl->adesc = NULL; 869 if (rx_ctl->chan) 870 dmaengine_terminate_all(rx_ctl->chan); 871 872 if (sg_dma_address(&rx_ctl->sg)) 873 dma_unmap_single(adapter->dev, sg_dma_address(&rx_ctl->sg), 874 DMA_BUFFER_SIZE, DMA_FROM_DEVICE); 875 sg_dma_address(&rx_ctl->sg) = 0; 876 877 dev_kfree_skb(rx_ctl->skb); 878 rx_ctl->skb = NULL; 879 } 880 881 static void ks8842_dealloc_dma_bufs(struct ks8842_adapter *adapter) 882 { 883 struct ks8842_tx_dma_ctl *tx_ctl = &adapter->dma_tx; 884 struct ks8842_rx_dma_ctl *rx_ctl = &adapter->dma_rx; 885 886 ks8842_stop_dma(adapter); 887 888 if (tx_ctl->chan) 889 dma_release_channel(tx_ctl->chan); 890 tx_ctl->chan = NULL; 891 892 if (rx_ctl->chan) 893 dma_release_channel(rx_ctl->chan); 894 rx_ctl->chan = NULL; 895 896 tasklet_kill(&rx_ctl->tasklet); 897 898 if (sg_dma_address(&tx_ctl->sg)) 899 dma_unmap_single(adapter->dev, sg_dma_address(&tx_ctl->sg), 900 DMA_BUFFER_SIZE, DMA_TO_DEVICE); 901 sg_dma_address(&tx_ctl->sg) = 0; 902 903 kfree(tx_ctl->buf); 904 tx_ctl->buf = NULL; 905 } 906 907 static bool ks8842_dma_filter_fn(struct dma_chan *chan, void *filter_param) 908 { 909 return chan->chan_id == (long)filter_param; 910 } 911 912 static int ks8842_alloc_dma_bufs(struct net_device *netdev) 913 { 914 struct ks8842_adapter *adapter = netdev_priv(netdev); 915 struct ks8842_tx_dma_ctl *tx_ctl = &adapter->dma_tx; 916 struct ks8842_rx_dma_ctl *rx_ctl = &adapter->dma_rx; 917 int err; 918 919 dma_cap_mask_t mask; 920 921 dma_cap_zero(mask); 922 dma_cap_set(DMA_SLAVE, mask); 923 dma_cap_set(DMA_PRIVATE, mask); 924 925 sg_init_table(&tx_ctl->sg, 1); 926 927 tx_ctl->chan = dma_request_channel(mask, ks8842_dma_filter_fn, 928 (void *)(long)tx_ctl->channel); 929 if (!tx_ctl->chan) { 930 err = -ENODEV; 931 goto err; 932 } 933 934 /* allocate DMA buffer */ 935 tx_ctl->buf = kmalloc(DMA_BUFFER_SIZE, GFP_KERNEL); 936 if (!tx_ctl->buf) { 937 err = -ENOMEM; 938 goto err; 939 } 940 941 sg_dma_address(&tx_ctl->sg) = dma_map_single(adapter->dev, 942 tx_ctl->buf, DMA_BUFFER_SIZE, DMA_TO_DEVICE); 943 if (dma_mapping_error(adapter->dev, sg_dma_address(&tx_ctl->sg))) { 944 err = -ENOMEM; 945 sg_dma_address(&tx_ctl->sg) = 0; 946 goto err; 947 } 948 949 rx_ctl->chan = dma_request_channel(mask, ks8842_dma_filter_fn, 950 (void *)(long)rx_ctl->channel); 951 if (!rx_ctl->chan) { 952 err = -ENODEV; 953 goto err; 954 } 955 956 tasklet_init(&rx_ctl->tasklet, ks8842_rx_frame_dma_tasklet, 957 (unsigned long)netdev); 958 959 return 0; 960 err: 961 ks8842_dealloc_dma_bufs(adapter); 962 return err; 963 } 964 965 /* Netdevice operations */ 966 967 static int ks8842_open(struct net_device *netdev) 968 { 969 struct ks8842_adapter *adapter = netdev_priv(netdev); 970 int err; 971 972 netdev_dbg(netdev, "%s - entry\n", __func__); 973 974 if (KS8842_USE_DMA(adapter)) { 975 err = ks8842_alloc_dma_bufs(netdev); 976 977 if (!err) { 978 /* start RX dma */ 979 err = __ks8842_start_new_rx_dma(netdev); 980 if (err) 981 ks8842_dealloc_dma_bufs(adapter); 982 } 983 984 if (err) { 985 printk(KERN_WARNING DRV_NAME 986 ": Failed to initiate DMA, running PIO\n"); 987 ks8842_dealloc_dma_bufs(adapter); 988 adapter->dma_rx.channel = -1; 989 adapter->dma_tx.channel = -1; 990 } 991 } 992 993 /* reset the HW */ 994 ks8842_reset_hw(adapter); 995 996 ks8842_write_mac_addr(adapter, netdev->dev_addr); 997 998 ks8842_update_link_status(netdev, adapter); 999 1000 err = request_irq(adapter->irq, ks8842_irq, IRQF_SHARED, DRV_NAME, 1001 netdev); 1002 if (err) { 1003 pr_err("Failed to request IRQ: %d: %d\n", adapter->irq, err); 1004 return err; 1005 } 1006 1007 return 0; 1008 } 1009 1010 static int ks8842_close(struct net_device *netdev) 1011 { 1012 struct ks8842_adapter *adapter = netdev_priv(netdev); 1013 1014 netdev_dbg(netdev, "%s - entry\n", __func__); 1015 1016 cancel_work_sync(&adapter->timeout_work); 1017 1018 if (KS8842_USE_DMA(adapter)) 1019 ks8842_dealloc_dma_bufs(adapter); 1020 1021 /* free the irq */ 1022 free_irq(adapter->irq, netdev); 1023 1024 /* disable the switch */ 1025 ks8842_write16(adapter, 32, 0x0, REG_SW_ID_AND_ENABLE); 1026 1027 return 0; 1028 } 1029 1030 static netdev_tx_t ks8842_xmit_frame(struct sk_buff *skb, 1031 struct net_device *netdev) 1032 { 1033 int ret; 1034 struct ks8842_adapter *adapter = netdev_priv(netdev); 1035 1036 netdev_dbg(netdev, "%s: entry\n", __func__); 1037 1038 if (KS8842_USE_DMA(adapter)) { 1039 unsigned long flags; 1040 ret = ks8842_tx_frame_dma(skb, netdev); 1041 /* for now only allow one transfer at the time */ 1042 spin_lock_irqsave(&adapter->lock, flags); 1043 if (adapter->dma_tx.adesc) 1044 netif_stop_queue(netdev); 1045 spin_unlock_irqrestore(&adapter->lock, flags); 1046 return ret; 1047 } 1048 1049 ret = ks8842_tx_frame(skb, netdev); 1050 1051 if (ks8842_tx_fifo_space(adapter) < netdev->mtu + 8) 1052 netif_stop_queue(netdev); 1053 1054 return ret; 1055 } 1056 1057 static int ks8842_set_mac(struct net_device *netdev, void *p) 1058 { 1059 struct ks8842_adapter *adapter = netdev_priv(netdev); 1060 struct sockaddr *addr = p; 1061 char *mac = (u8 *)addr->sa_data; 1062 1063 netdev_dbg(netdev, "%s: entry\n", __func__); 1064 1065 if (!is_valid_ether_addr(addr->sa_data)) 1066 return -EADDRNOTAVAIL; 1067 1068 memcpy(netdev->dev_addr, mac, netdev->addr_len); 1069 1070 ks8842_write_mac_addr(adapter, mac); 1071 return 0; 1072 } 1073 1074 static void ks8842_tx_timeout_work(struct work_struct *work) 1075 { 1076 struct ks8842_adapter *adapter = 1077 container_of(work, struct ks8842_adapter, timeout_work); 1078 struct net_device *netdev = adapter->netdev; 1079 unsigned long flags; 1080 1081 netdev_dbg(netdev, "%s: entry\n", __func__); 1082 1083 spin_lock_irqsave(&adapter->lock, flags); 1084 1085 if (KS8842_USE_DMA(adapter)) 1086 ks8842_stop_dma(adapter); 1087 1088 /* disable interrupts */ 1089 ks8842_write16(adapter, 18, 0, REG_IER); 1090 ks8842_write16(adapter, 18, 0xFFFF, REG_ISR); 1091 1092 netif_stop_queue(netdev); 1093 1094 spin_unlock_irqrestore(&adapter->lock, flags); 1095 1096 ks8842_reset_hw(adapter); 1097 1098 ks8842_write_mac_addr(adapter, netdev->dev_addr); 1099 1100 ks8842_update_link_status(netdev, adapter); 1101 1102 if (KS8842_USE_DMA(adapter)) 1103 __ks8842_start_new_rx_dma(netdev); 1104 } 1105 1106 static void ks8842_tx_timeout(struct net_device *netdev, unsigned int txqueue) 1107 { 1108 struct ks8842_adapter *adapter = netdev_priv(netdev); 1109 1110 netdev_dbg(netdev, "%s: entry\n", __func__); 1111 1112 schedule_work(&adapter->timeout_work); 1113 } 1114 1115 static const struct net_device_ops ks8842_netdev_ops = { 1116 .ndo_open = ks8842_open, 1117 .ndo_stop = ks8842_close, 1118 .ndo_start_xmit = ks8842_xmit_frame, 1119 .ndo_set_mac_address = ks8842_set_mac, 1120 .ndo_tx_timeout = ks8842_tx_timeout, 1121 .ndo_validate_addr = eth_validate_addr 1122 }; 1123 1124 static const struct ethtool_ops ks8842_ethtool_ops = { 1125 .get_link = ethtool_op_get_link, 1126 }; 1127 1128 static int ks8842_probe(struct platform_device *pdev) 1129 { 1130 int err = -ENOMEM; 1131 struct resource *iomem; 1132 struct net_device *netdev; 1133 struct ks8842_adapter *adapter; 1134 struct ks8842_platform_data *pdata = dev_get_platdata(&pdev->dev); 1135 u16 id; 1136 unsigned i; 1137 1138 iomem = platform_get_resource(pdev, IORESOURCE_MEM, 0); 1139 if (!request_mem_region(iomem->start, resource_size(iomem), DRV_NAME)) 1140 goto err_mem_region; 1141 1142 netdev = alloc_etherdev(sizeof(struct ks8842_adapter)); 1143 if (!netdev) 1144 goto err_alloc_etherdev; 1145 1146 SET_NETDEV_DEV(netdev, &pdev->dev); 1147 1148 adapter = netdev_priv(netdev); 1149 adapter->netdev = netdev; 1150 INIT_WORK(&adapter->timeout_work, ks8842_tx_timeout_work); 1151 adapter->hw_addr = ioremap(iomem->start, resource_size(iomem)); 1152 adapter->conf_flags = iomem->flags; 1153 1154 if (!adapter->hw_addr) 1155 goto err_ioremap; 1156 1157 adapter->irq = platform_get_irq(pdev, 0); 1158 if (adapter->irq < 0) { 1159 err = adapter->irq; 1160 goto err_get_irq; 1161 } 1162 1163 adapter->dev = (pdev->dev.parent) ? pdev->dev.parent : &pdev->dev; 1164 1165 /* DMA is only supported when accessed via timberdale */ 1166 if (!(adapter->conf_flags & MICREL_KS884X) && pdata && 1167 (pdata->tx_dma_channel != -1) && 1168 (pdata->rx_dma_channel != -1)) { 1169 adapter->dma_rx.channel = pdata->rx_dma_channel; 1170 adapter->dma_tx.channel = pdata->tx_dma_channel; 1171 } else { 1172 adapter->dma_rx.channel = -1; 1173 adapter->dma_tx.channel = -1; 1174 } 1175 1176 tasklet_init(&adapter->tasklet, ks8842_tasklet, (unsigned long)netdev); 1177 spin_lock_init(&adapter->lock); 1178 1179 netdev->netdev_ops = &ks8842_netdev_ops; 1180 netdev->ethtool_ops = &ks8842_ethtool_ops; 1181 1182 /* Check if a mac address was given */ 1183 i = netdev->addr_len; 1184 if (pdata) { 1185 for (i = 0; i < netdev->addr_len; i++) 1186 if (pdata->macaddr[i] != 0) 1187 break; 1188 1189 if (i < netdev->addr_len) 1190 /* an address was passed, use it */ 1191 memcpy(netdev->dev_addr, pdata->macaddr, 1192 netdev->addr_len); 1193 } 1194 1195 if (i == netdev->addr_len) { 1196 ks8842_read_mac_addr(adapter, netdev->dev_addr); 1197 1198 if (!is_valid_ether_addr(netdev->dev_addr)) 1199 eth_hw_addr_random(netdev); 1200 } 1201 1202 id = ks8842_read16(adapter, 32, REG_SW_ID_AND_ENABLE); 1203 1204 strcpy(netdev->name, "eth%d"); 1205 err = register_netdev(netdev); 1206 if (err) 1207 goto err_register; 1208 1209 platform_set_drvdata(pdev, netdev); 1210 1211 pr_info("Found chip, family: 0x%x, id: 0x%x, rev: 0x%x\n", 1212 (id >> 8) & 0xff, (id >> 4) & 0xf, (id >> 1) & 0x7); 1213 1214 return 0; 1215 1216 err_register: 1217 err_get_irq: 1218 iounmap(adapter->hw_addr); 1219 err_ioremap: 1220 free_netdev(netdev); 1221 err_alloc_etherdev: 1222 release_mem_region(iomem->start, resource_size(iomem)); 1223 err_mem_region: 1224 return err; 1225 } 1226 1227 static int ks8842_remove(struct platform_device *pdev) 1228 { 1229 struct net_device *netdev = platform_get_drvdata(pdev); 1230 struct ks8842_adapter *adapter = netdev_priv(netdev); 1231 struct resource *iomem = platform_get_resource(pdev, IORESOURCE_MEM, 0); 1232 1233 unregister_netdev(netdev); 1234 tasklet_kill(&adapter->tasklet); 1235 iounmap(adapter->hw_addr); 1236 free_netdev(netdev); 1237 release_mem_region(iomem->start, resource_size(iomem)); 1238 return 0; 1239 } 1240 1241 1242 static struct platform_driver ks8842_platform_driver = { 1243 .driver = { 1244 .name = DRV_NAME, 1245 }, 1246 .probe = ks8842_probe, 1247 .remove = ks8842_remove, 1248 }; 1249 1250 module_platform_driver(ks8842_platform_driver); 1251 1252 MODULE_DESCRIPTION("Timberdale KS8842 ethernet driver"); 1253 MODULE_AUTHOR("Mocean Laboratories <info@mocean-labs.com>"); 1254 MODULE_LICENSE("GPL v2"); 1255 MODULE_ALIAS("platform:ks8842"); 1256 1257