1caab277bSThomas Gleixner // SPDX-License-Identifier: GPL-2.0-only 2ae7668d0SJeff Kirsher /* 3ae7668d0SJeff Kirsher * 4ae7668d0SJeff Kirsher * Copyright (C) 2011 John Crispin <blogic@openwrt.org> 5ae7668d0SJeff Kirsher */ 6ae7668d0SJeff Kirsher 7ae7668d0SJeff Kirsher #include <linux/kernel.h> 8ae7668d0SJeff Kirsher #include <linux/slab.h> 9ae7668d0SJeff Kirsher #include <linux/errno.h> 10ae7668d0SJeff Kirsher #include <linux/types.h> 11ae7668d0SJeff Kirsher #include <linux/interrupt.h> 12ae7668d0SJeff Kirsher #include <linux/uaccess.h> 13ae7668d0SJeff Kirsher #include <linux/in.h> 14ae7668d0SJeff Kirsher #include <linux/netdevice.h> 15ae7668d0SJeff Kirsher #include <linux/etherdevice.h> 16ae7668d0SJeff Kirsher #include <linux/phy.h> 17ae7668d0SJeff Kirsher #include <linux/ip.h> 18ae7668d0SJeff Kirsher #include <linux/tcp.h> 19ae7668d0SJeff Kirsher #include <linux/skbuff.h> 20ae7668d0SJeff Kirsher #include <linux/mm.h> 21ae7668d0SJeff Kirsher #include <linux/platform_device.h> 22ae7668d0SJeff Kirsher #include <linux/ethtool.h> 23ae7668d0SJeff Kirsher #include <linux/init.h> 24ae7668d0SJeff Kirsher #include <linux/delay.h> 25ae7668d0SJeff Kirsher #include <linux/io.h> 26a32fd63dSJohn Crispin #include <linux/dma-mapping.h> 27a32fd63dSJohn Crispin #include <linux/module.h> 28ae7668d0SJeff Kirsher 29ae7668d0SJeff Kirsher #include <asm/checksum.h> 30ae7668d0SJeff Kirsher 31ae7668d0SJeff Kirsher #include <lantiq_soc.h> 32ae7668d0SJeff Kirsher #include <xway_dma.h> 33ae7668d0SJeff Kirsher #include <lantiq_platform.h> 34ae7668d0SJeff Kirsher 35ae7668d0SJeff Kirsher #define LTQ_ETOP_MDIO 0x11804 36ae7668d0SJeff Kirsher #define MDIO_REQUEST 0x80000000 37ae7668d0SJeff Kirsher #define MDIO_READ 0x40000000 38ae7668d0SJeff Kirsher #define MDIO_ADDR_MASK 0x1f 39ae7668d0SJeff Kirsher #define MDIO_ADDR_OFFSET 0x15 40ae7668d0SJeff Kirsher #define MDIO_REG_MASK 0x1f 41ae7668d0SJeff Kirsher #define MDIO_REG_OFFSET 0x10 42ae7668d0SJeff Kirsher #define MDIO_VAL_MASK 0xffff 43ae7668d0SJeff Kirsher 44ae7668d0SJeff Kirsher #define PPE32_CGEN 0x800 45ae7668d0SJeff Kirsher #define LQ_PPE32_ENET_MAC_CFG 0x1840 46ae7668d0SJeff Kirsher 47ae7668d0SJeff Kirsher #define LTQ_ETOP_ENETS0 0x11850 48ae7668d0SJeff Kirsher #define LTQ_ETOP_MAC_DA0 0x1186C 49ae7668d0SJeff Kirsher #define LTQ_ETOP_MAC_DA1 0x11870 50ae7668d0SJeff Kirsher #define LTQ_ETOP_CFG 0x16020 51ae7668d0SJeff Kirsher #define LTQ_ETOP_IGPLEN 0x16080 52ae7668d0SJeff Kirsher 53ae7668d0SJeff Kirsher #define MAX_DMA_CHAN 0x8 54ae7668d0SJeff Kirsher #define MAX_DMA_CRC_LEN 0x4 55ae7668d0SJeff Kirsher #define MAX_DMA_DATA_LEN 0x600 56ae7668d0SJeff Kirsher 57ae7668d0SJeff Kirsher #define ETOP_FTCU BIT(28) 58ae7668d0SJeff Kirsher #define ETOP_MII_MASK 0xf 59ae7668d0SJeff Kirsher #define ETOP_MII_NORMAL 0xd 60ae7668d0SJeff Kirsher #define ETOP_MII_REVERSE 0xe 61ae7668d0SJeff Kirsher #define ETOP_PLEN_UNDER 0x40 62ae7668d0SJeff Kirsher #define ETOP_CGEN 0x800 63ae7668d0SJeff Kirsher 64ae7668d0SJeff Kirsher /* use 2 static channels for TX/RX */ 65ae7668d0SJeff Kirsher #define LTQ_ETOP_TX_CHANNEL 1 66ae7668d0SJeff Kirsher #define LTQ_ETOP_RX_CHANNEL 6 67ae7668d0SJeff Kirsher #define IS_TX(x) (x == LTQ_ETOP_TX_CHANNEL) 68ae7668d0SJeff Kirsher #define IS_RX(x) (x == LTQ_ETOP_RX_CHANNEL) 69ae7668d0SJeff Kirsher 70ae7668d0SJeff Kirsher #define ltq_etop_r32(x) ltq_r32(ltq_etop_membase + (x)) 71ae7668d0SJeff Kirsher #define ltq_etop_w32(x, y) ltq_w32(x, ltq_etop_membase + (y)) 72ae7668d0SJeff Kirsher #define ltq_etop_w32_mask(x, y, z) \ 73ae7668d0SJeff Kirsher ltq_w32_mask(x, y, ltq_etop_membase + (z)) 74ae7668d0SJeff Kirsher 75ae7668d0SJeff Kirsher #define DRV_VERSION "1.0" 76ae7668d0SJeff Kirsher 77ae7668d0SJeff Kirsher static void __iomem *ltq_etop_membase; 78ae7668d0SJeff Kirsher 79ae7668d0SJeff Kirsher struct ltq_etop_chan { 80ae7668d0SJeff Kirsher int idx; 81ae7668d0SJeff Kirsher int tx_free; 82ae7668d0SJeff Kirsher struct net_device *netdev; 83ae7668d0SJeff Kirsher struct napi_struct napi; 84ae7668d0SJeff Kirsher struct ltq_dma_channel dma; 85ae7668d0SJeff Kirsher struct sk_buff *skb[LTQ_DESC_NUM]; 86ae7668d0SJeff Kirsher }; 87ae7668d0SJeff Kirsher 88ae7668d0SJeff Kirsher struct ltq_etop_priv { 89ae7668d0SJeff Kirsher struct net_device *netdev; 90d1b86507SFlorian Fainelli struct platform_device *pdev; 91ae7668d0SJeff Kirsher struct ltq_eth_data *pldata; 92ae7668d0SJeff Kirsher struct resource *res; 93ae7668d0SJeff Kirsher 94ae7668d0SJeff Kirsher struct mii_bus *mii_bus; 95ae7668d0SJeff Kirsher 96ae7668d0SJeff Kirsher struct ltq_etop_chan ch[MAX_DMA_CHAN]; 97ae7668d0SJeff Kirsher int tx_free[MAX_DMA_CHAN >> 1]; 98ae7668d0SJeff Kirsher 99*14d4e308SAleksander Jan Bajkowski int tx_burst_len; 100*14d4e308SAleksander Jan Bajkowski int rx_burst_len; 101*14d4e308SAleksander Jan Bajkowski 102ae7668d0SJeff Kirsher spinlock_t lock; 103ae7668d0SJeff Kirsher }; 104ae7668d0SJeff Kirsher 105ae7668d0SJeff Kirsher static int 106ae7668d0SJeff Kirsher ltq_etop_alloc_skb(struct ltq_etop_chan *ch) 107ae7668d0SJeff Kirsher { 10874e0deb8SChristoph Hellwig struct ltq_etop_priv *priv = netdev_priv(ch->netdev); 10974e0deb8SChristoph Hellwig 110c056b734SPradeep A Dalvi ch->skb[ch->dma.desc] = netdev_alloc_skb(ch->netdev, MAX_DMA_DATA_LEN); 111ae7668d0SJeff Kirsher if (!ch->skb[ch->dma.desc]) 112ae7668d0SJeff Kirsher return -ENOMEM; 11374e0deb8SChristoph Hellwig ch->dma.desc_base[ch->dma.desc].addr = dma_map_single(&priv->pdev->dev, 114ae7668d0SJeff Kirsher ch->skb[ch->dma.desc]->data, MAX_DMA_DATA_LEN, 115ae7668d0SJeff Kirsher DMA_FROM_DEVICE); 116ae7668d0SJeff Kirsher ch->dma.desc_base[ch->dma.desc].addr = 117ae7668d0SJeff Kirsher CPHYSADDR(ch->skb[ch->dma.desc]->data); 118ae7668d0SJeff Kirsher ch->dma.desc_base[ch->dma.desc].ctl = 119ae7668d0SJeff Kirsher LTQ_DMA_OWN | LTQ_DMA_RX_OFFSET(NET_IP_ALIGN) | 120ae7668d0SJeff Kirsher MAX_DMA_DATA_LEN; 121ae7668d0SJeff Kirsher skb_reserve(ch->skb[ch->dma.desc], NET_IP_ALIGN); 122ae7668d0SJeff Kirsher return 0; 123ae7668d0SJeff Kirsher } 124ae7668d0SJeff Kirsher 125ae7668d0SJeff Kirsher static void 126ae7668d0SJeff Kirsher ltq_etop_hw_receive(struct ltq_etop_chan *ch) 127ae7668d0SJeff Kirsher { 128ae7668d0SJeff Kirsher struct ltq_etop_priv *priv = netdev_priv(ch->netdev); 129ae7668d0SJeff Kirsher struct ltq_dma_desc *desc = &ch->dma.desc_base[ch->dma.desc]; 130ae7668d0SJeff Kirsher struct sk_buff *skb = ch->skb[ch->dma.desc]; 131ae7668d0SJeff Kirsher int len = (desc->ctl & LTQ_DMA_SIZE_MASK) - MAX_DMA_CRC_LEN; 132ae7668d0SJeff Kirsher unsigned long flags; 133ae7668d0SJeff Kirsher 134ae7668d0SJeff Kirsher spin_lock_irqsave(&priv->lock, flags); 135ae7668d0SJeff Kirsher if (ltq_etop_alloc_skb(ch)) { 136ae7668d0SJeff Kirsher netdev_err(ch->netdev, 137ae7668d0SJeff Kirsher "failed to allocate new rx buffer, stopping DMA\n"); 138ae7668d0SJeff Kirsher ltq_dma_close(&ch->dma); 139ae7668d0SJeff Kirsher } 140ae7668d0SJeff Kirsher ch->dma.desc++; 141ae7668d0SJeff Kirsher ch->dma.desc %= LTQ_DESC_NUM; 142ae7668d0SJeff Kirsher spin_unlock_irqrestore(&priv->lock, flags); 143ae7668d0SJeff Kirsher 144ae7668d0SJeff Kirsher skb_put(skb, len); 145ae7668d0SJeff Kirsher skb->protocol = eth_type_trans(skb, ch->netdev); 146ae7668d0SJeff Kirsher netif_receive_skb(skb); 147ae7668d0SJeff Kirsher } 148ae7668d0SJeff Kirsher 149ae7668d0SJeff Kirsher static int 150ae7668d0SJeff Kirsher ltq_etop_poll_rx(struct napi_struct *napi, int budget) 151ae7668d0SJeff Kirsher { 152ae7668d0SJeff Kirsher struct ltq_etop_chan *ch = container_of(napi, 153ae7668d0SJeff Kirsher struct ltq_etop_chan, napi); 1546ad20165SEric Dumazet int work_done = 0; 155ae7668d0SJeff Kirsher 1566ad20165SEric Dumazet while (work_done < budget) { 157ae7668d0SJeff Kirsher struct ltq_dma_desc *desc = &ch->dma.desc_base[ch->dma.desc]; 158ae7668d0SJeff Kirsher 1596ad20165SEric Dumazet if ((desc->ctl & (LTQ_DMA_OWN | LTQ_DMA_C)) != LTQ_DMA_C) 1606ad20165SEric Dumazet break; 161ae7668d0SJeff Kirsher ltq_etop_hw_receive(ch); 1626ad20165SEric Dumazet work_done++; 163ae7668d0SJeff Kirsher } 1646ad20165SEric Dumazet if (work_done < budget) { 1656ad20165SEric Dumazet napi_complete_done(&ch->napi, work_done); 166ae7668d0SJeff Kirsher ltq_dma_ack_irq(&ch->dma); 167ae7668d0SJeff Kirsher } 1686ad20165SEric Dumazet return work_done; 169ae7668d0SJeff Kirsher } 170ae7668d0SJeff Kirsher 171ae7668d0SJeff Kirsher static int 172ae7668d0SJeff Kirsher ltq_etop_poll_tx(struct napi_struct *napi, int budget) 173ae7668d0SJeff Kirsher { 174ae7668d0SJeff Kirsher struct ltq_etop_chan *ch = 175ae7668d0SJeff Kirsher container_of(napi, struct ltq_etop_chan, napi); 176ae7668d0SJeff Kirsher struct ltq_etop_priv *priv = netdev_priv(ch->netdev); 177ae7668d0SJeff Kirsher struct netdev_queue *txq = 178ae7668d0SJeff Kirsher netdev_get_tx_queue(ch->netdev, ch->idx >> 1); 179ae7668d0SJeff Kirsher unsigned long flags; 180ae7668d0SJeff Kirsher 181ae7668d0SJeff Kirsher spin_lock_irqsave(&priv->lock, flags); 182ae7668d0SJeff Kirsher while ((ch->dma.desc_base[ch->tx_free].ctl & 183ae7668d0SJeff Kirsher (LTQ_DMA_OWN | LTQ_DMA_C)) == LTQ_DMA_C) { 184ae7668d0SJeff Kirsher dev_kfree_skb_any(ch->skb[ch->tx_free]); 185ae7668d0SJeff Kirsher ch->skb[ch->tx_free] = NULL; 186ae7668d0SJeff Kirsher memset(&ch->dma.desc_base[ch->tx_free], 0, 187ae7668d0SJeff Kirsher sizeof(struct ltq_dma_desc)); 188ae7668d0SJeff Kirsher ch->tx_free++; 189ae7668d0SJeff Kirsher ch->tx_free %= LTQ_DESC_NUM; 190ae7668d0SJeff Kirsher } 191ae7668d0SJeff Kirsher spin_unlock_irqrestore(&priv->lock, flags); 192ae7668d0SJeff Kirsher 193ae7668d0SJeff Kirsher if (netif_tx_queue_stopped(txq)) 194ae7668d0SJeff Kirsher netif_tx_start_queue(txq); 195ae7668d0SJeff Kirsher napi_complete(&ch->napi); 196ae7668d0SJeff Kirsher ltq_dma_ack_irq(&ch->dma); 197ae7668d0SJeff Kirsher return 1; 198ae7668d0SJeff Kirsher } 199ae7668d0SJeff Kirsher 200ae7668d0SJeff Kirsher static irqreturn_t 201ae7668d0SJeff Kirsher ltq_etop_dma_irq(int irq, void *_priv) 202ae7668d0SJeff Kirsher { 203ae7668d0SJeff Kirsher struct ltq_etop_priv *priv = _priv; 204ae7668d0SJeff Kirsher int ch = irq - LTQ_DMA_CH0_INT; 205ae7668d0SJeff Kirsher 206ae7668d0SJeff Kirsher napi_schedule(&priv->ch[ch].napi); 207ae7668d0SJeff Kirsher return IRQ_HANDLED; 208ae7668d0SJeff Kirsher } 209ae7668d0SJeff Kirsher 210ae7668d0SJeff Kirsher static void 211ae7668d0SJeff Kirsher ltq_etop_free_channel(struct net_device *dev, struct ltq_etop_chan *ch) 212ae7668d0SJeff Kirsher { 213ae7668d0SJeff Kirsher struct ltq_etop_priv *priv = netdev_priv(dev); 214ae7668d0SJeff Kirsher 215ae7668d0SJeff Kirsher ltq_dma_free(&ch->dma); 216ae7668d0SJeff Kirsher if (ch->dma.irq) 217ae7668d0SJeff Kirsher free_irq(ch->dma.irq, priv); 218ae7668d0SJeff Kirsher if (IS_RX(ch->idx)) { 219ae7668d0SJeff Kirsher int desc; 220ae7668d0SJeff Kirsher for (desc = 0; desc < LTQ_DESC_NUM; desc++) 221ae7668d0SJeff Kirsher dev_kfree_skb_any(ch->skb[ch->dma.desc]); 222ae7668d0SJeff Kirsher } 223ae7668d0SJeff Kirsher } 224ae7668d0SJeff Kirsher 225ae7668d0SJeff Kirsher static void 226ae7668d0SJeff Kirsher ltq_etop_hw_exit(struct net_device *dev) 227ae7668d0SJeff Kirsher { 228ae7668d0SJeff Kirsher struct ltq_etop_priv *priv = netdev_priv(dev); 229ae7668d0SJeff Kirsher int i; 230ae7668d0SJeff Kirsher 231ae7668d0SJeff Kirsher ltq_pmu_disable(PMU_PPE); 232ae7668d0SJeff Kirsher for (i = 0; i < MAX_DMA_CHAN; i++) 233ae7668d0SJeff Kirsher if (IS_TX(i) || IS_RX(i)) 234ae7668d0SJeff Kirsher ltq_etop_free_channel(dev, &priv->ch[i]); 235ae7668d0SJeff Kirsher } 236ae7668d0SJeff Kirsher 237ae7668d0SJeff Kirsher static int 238ae7668d0SJeff Kirsher ltq_etop_hw_init(struct net_device *dev) 239ae7668d0SJeff Kirsher { 240ae7668d0SJeff Kirsher struct ltq_etop_priv *priv = netdev_priv(dev); 241ae7668d0SJeff Kirsher int i; 242ae7668d0SJeff Kirsher 243ae7668d0SJeff Kirsher ltq_pmu_enable(PMU_PPE); 244ae7668d0SJeff Kirsher 245ae7668d0SJeff Kirsher switch (priv->pldata->mii_mode) { 246ae7668d0SJeff Kirsher case PHY_INTERFACE_MODE_RMII: 247ae7668d0SJeff Kirsher ltq_etop_w32_mask(ETOP_MII_MASK, 248ae7668d0SJeff Kirsher ETOP_MII_REVERSE, LTQ_ETOP_CFG); 249ae7668d0SJeff Kirsher break; 250ae7668d0SJeff Kirsher 251ae7668d0SJeff Kirsher case PHY_INTERFACE_MODE_MII: 252ae7668d0SJeff Kirsher ltq_etop_w32_mask(ETOP_MII_MASK, 253ae7668d0SJeff Kirsher ETOP_MII_NORMAL, LTQ_ETOP_CFG); 254ae7668d0SJeff Kirsher break; 255ae7668d0SJeff Kirsher 256ae7668d0SJeff Kirsher default: 257ae7668d0SJeff Kirsher netdev_err(dev, "unknown mii mode %d\n", 258ae7668d0SJeff Kirsher priv->pldata->mii_mode); 259ae7668d0SJeff Kirsher return -ENOTSUPP; 260ae7668d0SJeff Kirsher } 261ae7668d0SJeff Kirsher 262ae7668d0SJeff Kirsher /* enable crc generation */ 263ae7668d0SJeff Kirsher ltq_etop_w32(PPE32_CGEN, LQ_PPE32_ENET_MAC_CFG); 264ae7668d0SJeff Kirsher 265*14d4e308SAleksander Jan Bajkowski ltq_dma_init_port(DMA_PORT_ETOP, priv->tx_burst_len, rx_burst_len); 266ae7668d0SJeff Kirsher 267ae7668d0SJeff Kirsher for (i = 0; i < MAX_DMA_CHAN; i++) { 268ae7668d0SJeff Kirsher int irq = LTQ_DMA_CH0_INT + i; 269ae7668d0SJeff Kirsher struct ltq_etop_chan *ch = &priv->ch[i]; 270ae7668d0SJeff Kirsher 271ae7668d0SJeff Kirsher ch->idx = ch->dma.nr = i; 2722d946e5bSHauke Mehrtens ch->dma.dev = &priv->pdev->dev; 273ae7668d0SJeff Kirsher 274ae7668d0SJeff Kirsher if (IS_TX(i)) { 275ae7668d0SJeff Kirsher ltq_dma_alloc_tx(&ch->dma); 276dddb29e4SMichael Opdenacker request_irq(irq, ltq_etop_dma_irq, 0, "etop_tx", priv); 277ae7668d0SJeff Kirsher } else if (IS_RX(i)) { 278ae7668d0SJeff Kirsher ltq_dma_alloc_rx(&ch->dma); 279ae7668d0SJeff Kirsher for (ch->dma.desc = 0; ch->dma.desc < LTQ_DESC_NUM; 280ae7668d0SJeff Kirsher ch->dma.desc++) 281ae7668d0SJeff Kirsher if (ltq_etop_alloc_skb(ch)) 282ae7668d0SJeff Kirsher return -ENOMEM; 283ae7668d0SJeff Kirsher ch->dma.desc = 0; 284dddb29e4SMichael Opdenacker request_irq(irq, ltq_etop_dma_irq, 0, "etop_rx", priv); 285ae7668d0SJeff Kirsher } 286ae7668d0SJeff Kirsher ch->dma.irq = irq; 287ae7668d0SJeff Kirsher } 288ae7668d0SJeff Kirsher return 0; 289ae7668d0SJeff Kirsher } 290ae7668d0SJeff Kirsher 291ae7668d0SJeff Kirsher static void 292ae7668d0SJeff Kirsher ltq_etop_get_drvinfo(struct net_device *dev, struct ethtool_drvinfo *info) 293ae7668d0SJeff Kirsher { 2947826d43fSJiri Pirko strlcpy(info->driver, "Lantiq ETOP", sizeof(info->driver)); 2957826d43fSJiri Pirko strlcpy(info->bus_info, "internal", sizeof(info->bus_info)); 2967826d43fSJiri Pirko strlcpy(info->version, DRV_VERSION, sizeof(info->version)); 297ae7668d0SJeff Kirsher } 298ae7668d0SJeff Kirsher 299ae7668d0SJeff Kirsher static const struct ethtool_ops ltq_etop_ethtool_ops = { 300ae7668d0SJeff Kirsher .get_drvinfo = ltq_etop_get_drvinfo, 301e3979ce9SFlorian Fainelli .nway_reset = phy_ethtool_nway_reset, 3025376d95fSPhilippe Reynes .get_link_ksettings = phy_ethtool_get_link_ksettings, 3035376d95fSPhilippe Reynes .set_link_ksettings = phy_ethtool_set_link_ksettings, 304ae7668d0SJeff Kirsher }; 305ae7668d0SJeff Kirsher 306ae7668d0SJeff Kirsher static int 307ae7668d0SJeff Kirsher ltq_etop_mdio_wr(struct mii_bus *bus, int phy_addr, int phy_reg, u16 phy_data) 308ae7668d0SJeff Kirsher { 309ae7668d0SJeff Kirsher u32 val = MDIO_REQUEST | 310ae7668d0SJeff Kirsher ((phy_addr & MDIO_ADDR_MASK) << MDIO_ADDR_OFFSET) | 311ae7668d0SJeff Kirsher ((phy_reg & MDIO_REG_MASK) << MDIO_REG_OFFSET) | 312ae7668d0SJeff Kirsher phy_data; 313ae7668d0SJeff Kirsher 314ae7668d0SJeff Kirsher while (ltq_etop_r32(LTQ_ETOP_MDIO) & MDIO_REQUEST) 315ae7668d0SJeff Kirsher ; 316ae7668d0SJeff Kirsher ltq_etop_w32(val, LTQ_ETOP_MDIO); 317ae7668d0SJeff Kirsher return 0; 318ae7668d0SJeff Kirsher } 319ae7668d0SJeff Kirsher 320ae7668d0SJeff Kirsher static int 321ae7668d0SJeff Kirsher ltq_etop_mdio_rd(struct mii_bus *bus, int phy_addr, int phy_reg) 322ae7668d0SJeff Kirsher { 323ae7668d0SJeff Kirsher u32 val = MDIO_REQUEST | MDIO_READ | 324ae7668d0SJeff Kirsher ((phy_addr & MDIO_ADDR_MASK) << MDIO_ADDR_OFFSET) | 325ae7668d0SJeff Kirsher ((phy_reg & MDIO_REG_MASK) << MDIO_REG_OFFSET); 326ae7668d0SJeff Kirsher 327ae7668d0SJeff Kirsher while (ltq_etop_r32(LTQ_ETOP_MDIO) & MDIO_REQUEST) 328ae7668d0SJeff Kirsher ; 329ae7668d0SJeff Kirsher ltq_etop_w32(val, LTQ_ETOP_MDIO); 330ae7668d0SJeff Kirsher while (ltq_etop_r32(LTQ_ETOP_MDIO) & MDIO_REQUEST) 331ae7668d0SJeff Kirsher ; 332ae7668d0SJeff Kirsher val = ltq_etop_r32(LTQ_ETOP_MDIO) & MDIO_VAL_MASK; 333ae7668d0SJeff Kirsher return val; 334ae7668d0SJeff Kirsher } 335ae7668d0SJeff Kirsher 336ae7668d0SJeff Kirsher static void 337ae7668d0SJeff Kirsher ltq_etop_mdio_link(struct net_device *dev) 338ae7668d0SJeff Kirsher { 339ae7668d0SJeff Kirsher /* nothing to do */ 340ae7668d0SJeff Kirsher } 341ae7668d0SJeff Kirsher 342ae7668d0SJeff Kirsher static int 343ae7668d0SJeff Kirsher ltq_etop_mdio_probe(struct net_device *dev) 344ae7668d0SJeff Kirsher { 345ae7668d0SJeff Kirsher struct ltq_etop_priv *priv = netdev_priv(dev); 3462a4fc4eaSAndrew Lunn struct phy_device *phydev; 347ae7668d0SJeff Kirsher 3482a4fc4eaSAndrew Lunn phydev = phy_find_first(priv->mii_bus); 349ae7668d0SJeff Kirsher 350ae7668d0SJeff Kirsher if (!phydev) { 351ae7668d0SJeff Kirsher netdev_err(dev, "no PHY found\n"); 352ae7668d0SJeff Kirsher return -ENODEV; 353ae7668d0SJeff Kirsher } 354ae7668d0SJeff Kirsher 35584eff6d1SAndrew Lunn phydev = phy_connect(dev, phydev_name(phydev), 356f9a8f83bSFlorian Fainelli <q_etop_mdio_link, priv->pldata->mii_mode); 357ae7668d0SJeff Kirsher 358ae7668d0SJeff Kirsher if (IS_ERR(phydev)) { 359ae7668d0SJeff Kirsher netdev_err(dev, "Could not attach to PHY\n"); 360ae7668d0SJeff Kirsher return PTR_ERR(phydev); 361ae7668d0SJeff Kirsher } 362ae7668d0SJeff Kirsher 36358056c1eSAndrew Lunn phy_set_max_speed(phydev, SPEED_100); 364ae7668d0SJeff Kirsher 3652220943aSAndrew Lunn phy_attached_info(phydev); 366ae7668d0SJeff Kirsher 367ae7668d0SJeff Kirsher return 0; 368ae7668d0SJeff Kirsher } 369ae7668d0SJeff Kirsher 370ae7668d0SJeff Kirsher static int 371ae7668d0SJeff Kirsher ltq_etop_mdio_init(struct net_device *dev) 372ae7668d0SJeff Kirsher { 373ae7668d0SJeff Kirsher struct ltq_etop_priv *priv = netdev_priv(dev); 374ae7668d0SJeff Kirsher int err; 375ae7668d0SJeff Kirsher 376ae7668d0SJeff Kirsher priv->mii_bus = mdiobus_alloc(); 377ae7668d0SJeff Kirsher if (!priv->mii_bus) { 378ae7668d0SJeff Kirsher netdev_err(dev, "failed to allocate mii bus\n"); 379ae7668d0SJeff Kirsher err = -ENOMEM; 380ae7668d0SJeff Kirsher goto err_out; 381ae7668d0SJeff Kirsher } 382ae7668d0SJeff Kirsher 383ae7668d0SJeff Kirsher priv->mii_bus->priv = dev; 384ae7668d0SJeff Kirsher priv->mii_bus->read = ltq_etop_mdio_rd; 385ae7668d0SJeff Kirsher priv->mii_bus->write = ltq_etop_mdio_wr; 386ae7668d0SJeff Kirsher priv->mii_bus->name = "ltq_mii"; 387d1b86507SFlorian Fainelli snprintf(priv->mii_bus->id, MII_BUS_ID_SIZE, "%s-%x", 388d1b86507SFlorian Fainelli priv->pdev->name, priv->pdev->id); 389ae7668d0SJeff Kirsher if (mdiobus_register(priv->mii_bus)) { 390ae7668d0SJeff Kirsher err = -ENXIO; 391e7f4dc35SAndrew Lunn goto err_out_free_mdiobus; 392ae7668d0SJeff Kirsher } 393ae7668d0SJeff Kirsher 394ae7668d0SJeff Kirsher if (ltq_etop_mdio_probe(dev)) { 395ae7668d0SJeff Kirsher err = -ENXIO; 396ae7668d0SJeff Kirsher goto err_out_unregister_bus; 397ae7668d0SJeff Kirsher } 398ae7668d0SJeff Kirsher return 0; 399ae7668d0SJeff Kirsher 400ae7668d0SJeff Kirsher err_out_unregister_bus: 401ae7668d0SJeff Kirsher mdiobus_unregister(priv->mii_bus); 402ae7668d0SJeff Kirsher err_out_free_mdiobus: 403ae7668d0SJeff Kirsher mdiobus_free(priv->mii_bus); 404ae7668d0SJeff Kirsher err_out: 405ae7668d0SJeff Kirsher return err; 406ae7668d0SJeff Kirsher } 407ae7668d0SJeff Kirsher 408ae7668d0SJeff Kirsher static void 409ae7668d0SJeff Kirsher ltq_etop_mdio_cleanup(struct net_device *dev) 410ae7668d0SJeff Kirsher { 411ae7668d0SJeff Kirsher struct ltq_etop_priv *priv = netdev_priv(dev); 412ae7668d0SJeff Kirsher 413d1e3a356SPhilippe Reynes phy_disconnect(dev->phydev); 414ae7668d0SJeff Kirsher mdiobus_unregister(priv->mii_bus); 415ae7668d0SJeff Kirsher mdiobus_free(priv->mii_bus); 416ae7668d0SJeff Kirsher } 417ae7668d0SJeff Kirsher 418ae7668d0SJeff Kirsher static int 419ae7668d0SJeff Kirsher ltq_etop_open(struct net_device *dev) 420ae7668d0SJeff Kirsher { 421ae7668d0SJeff Kirsher struct ltq_etop_priv *priv = netdev_priv(dev); 422ae7668d0SJeff Kirsher int i; 423ae7668d0SJeff Kirsher 424ae7668d0SJeff Kirsher for (i = 0; i < MAX_DMA_CHAN; i++) { 425ae7668d0SJeff Kirsher struct ltq_etop_chan *ch = &priv->ch[i]; 426ae7668d0SJeff Kirsher 427ae7668d0SJeff Kirsher if (!IS_TX(i) && (!IS_RX(i))) 428ae7668d0SJeff Kirsher continue; 429ae7668d0SJeff Kirsher ltq_dma_open(&ch->dma); 430cc973aecSHauke Mehrtens ltq_dma_enable_irq(&ch->dma); 431ae7668d0SJeff Kirsher napi_enable(&ch->napi); 432ae7668d0SJeff Kirsher } 433d1e3a356SPhilippe Reynes phy_start(dev->phydev); 434ae7668d0SJeff Kirsher netif_tx_start_all_queues(dev); 435ae7668d0SJeff Kirsher return 0; 436ae7668d0SJeff Kirsher } 437ae7668d0SJeff Kirsher 438ae7668d0SJeff Kirsher static int 439ae7668d0SJeff Kirsher ltq_etop_stop(struct net_device *dev) 440ae7668d0SJeff Kirsher { 441ae7668d0SJeff Kirsher struct ltq_etop_priv *priv = netdev_priv(dev); 442ae7668d0SJeff Kirsher int i; 443ae7668d0SJeff Kirsher 444ae7668d0SJeff Kirsher netif_tx_stop_all_queues(dev); 445d1e3a356SPhilippe Reynes phy_stop(dev->phydev); 446ae7668d0SJeff Kirsher for (i = 0; i < MAX_DMA_CHAN; i++) { 447ae7668d0SJeff Kirsher struct ltq_etop_chan *ch = &priv->ch[i]; 448ae7668d0SJeff Kirsher 449ae7668d0SJeff Kirsher if (!IS_RX(i) && !IS_TX(i)) 450ae7668d0SJeff Kirsher continue; 451ae7668d0SJeff Kirsher napi_disable(&ch->napi); 452ae7668d0SJeff Kirsher ltq_dma_close(&ch->dma); 453ae7668d0SJeff Kirsher } 454ae7668d0SJeff Kirsher return 0; 455ae7668d0SJeff Kirsher } 456ae7668d0SJeff Kirsher 457ae7668d0SJeff Kirsher static int 458ae7668d0SJeff Kirsher ltq_etop_tx(struct sk_buff *skb, struct net_device *dev) 459ae7668d0SJeff Kirsher { 460ae7668d0SJeff Kirsher int queue = skb_get_queue_mapping(skb); 461ae7668d0SJeff Kirsher struct netdev_queue *txq = netdev_get_tx_queue(dev, queue); 462ae7668d0SJeff Kirsher struct ltq_etop_priv *priv = netdev_priv(dev); 463ae7668d0SJeff Kirsher struct ltq_etop_chan *ch = &priv->ch[(queue << 1) | 1]; 464ae7668d0SJeff Kirsher struct ltq_dma_desc *desc = &ch->dma.desc_base[ch->dma.desc]; 465ae7668d0SJeff Kirsher int len; 466ae7668d0SJeff Kirsher unsigned long flags; 467ae7668d0SJeff Kirsher u32 byte_offset; 468ae7668d0SJeff Kirsher 469ae7668d0SJeff Kirsher len = skb->len < ETH_ZLEN ? ETH_ZLEN : skb->len; 470ae7668d0SJeff Kirsher 471ae7668d0SJeff Kirsher if ((desc->ctl & (LTQ_DMA_OWN | LTQ_DMA_C)) || ch->skb[ch->dma.desc]) { 472ae7668d0SJeff Kirsher dev_kfree_skb_any(skb); 473ae7668d0SJeff Kirsher netdev_err(dev, "tx ring full\n"); 474ae7668d0SJeff Kirsher netif_tx_stop_queue(txq); 475ae7668d0SJeff Kirsher return NETDEV_TX_BUSY; 476ae7668d0SJeff Kirsher } 477ae7668d0SJeff Kirsher 478*14d4e308SAleksander Jan Bajkowski /* dma needs to start on a burst length value aligned address */ 479*14d4e308SAleksander Jan Bajkowski byte_offset = CPHYSADDR(skb->data) % (priv->tx_burst_len * 4); 480ae7668d0SJeff Kirsher ch->skb[ch->dma.desc] = skb; 481ae7668d0SJeff Kirsher 482860e9538SFlorian Westphal netif_trans_update(dev); 483ae7668d0SJeff Kirsher 484ae7668d0SJeff Kirsher spin_lock_irqsave(&priv->lock, flags); 48574e0deb8SChristoph Hellwig desc->addr = ((unsigned int) dma_map_single(&priv->pdev->dev, skb->data, len, 486ae7668d0SJeff Kirsher DMA_TO_DEVICE)) - byte_offset; 487ae7668d0SJeff Kirsher wmb(); 488ae7668d0SJeff Kirsher desc->ctl = LTQ_DMA_OWN | LTQ_DMA_SOP | LTQ_DMA_EOP | 489ae7668d0SJeff Kirsher LTQ_DMA_TX_OFFSET(byte_offset) | (len & LTQ_DMA_SIZE_MASK); 490ae7668d0SJeff Kirsher ch->dma.desc++; 491ae7668d0SJeff Kirsher ch->dma.desc %= LTQ_DESC_NUM; 492ae7668d0SJeff Kirsher spin_unlock_irqrestore(&priv->lock, flags); 493ae7668d0SJeff Kirsher 494ae7668d0SJeff Kirsher if (ch->dma.desc_base[ch->dma.desc].ctl & LTQ_DMA_OWN) 495ae7668d0SJeff Kirsher netif_tx_stop_queue(txq); 496ae7668d0SJeff Kirsher 497ae7668d0SJeff Kirsher return NETDEV_TX_OK; 498ae7668d0SJeff Kirsher } 499ae7668d0SJeff Kirsher 500ae7668d0SJeff Kirsher static int 501ae7668d0SJeff Kirsher ltq_etop_change_mtu(struct net_device *dev, int new_mtu) 502ae7668d0SJeff Kirsher { 503ae7668d0SJeff Kirsher struct ltq_etop_priv *priv = netdev_priv(dev); 504ae7668d0SJeff Kirsher unsigned long flags; 505ae7668d0SJeff Kirsher 506a52ad514SJarod Wilson dev->mtu = new_mtu; 507a52ad514SJarod Wilson 508ae7668d0SJeff Kirsher spin_lock_irqsave(&priv->lock, flags); 509a52ad514SJarod Wilson ltq_etop_w32((ETOP_PLEN_UNDER << 16) | new_mtu, LTQ_ETOP_IGPLEN); 510ae7668d0SJeff Kirsher spin_unlock_irqrestore(&priv->lock, flags); 511a52ad514SJarod Wilson 512a52ad514SJarod Wilson return 0; 513ae7668d0SJeff Kirsher } 514ae7668d0SJeff Kirsher 515ae7668d0SJeff Kirsher static int 516ae7668d0SJeff Kirsher ltq_etop_set_mac_address(struct net_device *dev, void *p) 517ae7668d0SJeff Kirsher { 518ae7668d0SJeff Kirsher int ret = eth_mac_addr(dev, p); 519ae7668d0SJeff Kirsher 520ae7668d0SJeff Kirsher if (!ret) { 521ae7668d0SJeff Kirsher struct ltq_etop_priv *priv = netdev_priv(dev); 522ae7668d0SJeff Kirsher unsigned long flags; 523ae7668d0SJeff Kirsher 524ae7668d0SJeff Kirsher /* store the mac for the unicast filter */ 525ae7668d0SJeff Kirsher spin_lock_irqsave(&priv->lock, flags); 526ae7668d0SJeff Kirsher ltq_etop_w32(*((u32 *)dev->dev_addr), LTQ_ETOP_MAC_DA0); 527ae7668d0SJeff Kirsher ltq_etop_w32(*((u16 *)&dev->dev_addr[4]) << 16, 528ae7668d0SJeff Kirsher LTQ_ETOP_MAC_DA1); 529ae7668d0SJeff Kirsher spin_unlock_irqrestore(&priv->lock, flags); 530ae7668d0SJeff Kirsher } 531ae7668d0SJeff Kirsher return ret; 532ae7668d0SJeff Kirsher } 533ae7668d0SJeff Kirsher 534ae7668d0SJeff Kirsher static void 535ae7668d0SJeff Kirsher ltq_etop_set_multicast_list(struct net_device *dev) 536ae7668d0SJeff Kirsher { 537ae7668d0SJeff Kirsher struct ltq_etop_priv *priv = netdev_priv(dev); 538ae7668d0SJeff Kirsher unsigned long flags; 539ae7668d0SJeff Kirsher 540ae7668d0SJeff Kirsher /* ensure that the unicast filter is not enabled in promiscious mode */ 541ae7668d0SJeff Kirsher spin_lock_irqsave(&priv->lock, flags); 542ae7668d0SJeff Kirsher if ((dev->flags & IFF_PROMISC) || (dev->flags & IFF_ALLMULTI)) 543ae7668d0SJeff Kirsher ltq_etop_w32_mask(ETOP_FTCU, 0, LTQ_ETOP_ENETS0); 544ae7668d0SJeff Kirsher else 545ae7668d0SJeff Kirsher ltq_etop_w32_mask(0, ETOP_FTCU, LTQ_ETOP_ENETS0); 546ae7668d0SJeff Kirsher spin_unlock_irqrestore(&priv->lock, flags); 547ae7668d0SJeff Kirsher } 548ae7668d0SJeff Kirsher 549ae7668d0SJeff Kirsher static int 550ae7668d0SJeff Kirsher ltq_etop_init(struct net_device *dev) 551ae7668d0SJeff Kirsher { 552ae7668d0SJeff Kirsher struct ltq_etop_priv *priv = netdev_priv(dev); 553ae7668d0SJeff Kirsher struct sockaddr mac; 554ae7668d0SJeff Kirsher int err; 55543aabec5SDanny Kukawka bool random_mac = false; 556ae7668d0SJeff Kirsher 557ae7668d0SJeff Kirsher dev->watchdog_timeo = 10 * HZ; 558ae7668d0SJeff Kirsher err = ltq_etop_hw_init(dev); 559ae7668d0SJeff Kirsher if (err) 560ae7668d0SJeff Kirsher goto err_hw; 561ae7668d0SJeff Kirsher ltq_etop_change_mtu(dev, 1500); 562ae7668d0SJeff Kirsher 563ae7668d0SJeff Kirsher memcpy(&mac, &priv->pldata->mac, sizeof(struct sockaddr)); 564ae7668d0SJeff Kirsher if (!is_valid_ether_addr(mac.sa_data)) { 565ae7668d0SJeff Kirsher pr_warn("etop: invalid MAC, using random\n"); 5667efd26d0SJoe Perches eth_random_addr(mac.sa_data); 56743aabec5SDanny Kukawka random_mac = true; 568ae7668d0SJeff Kirsher } 569ae7668d0SJeff Kirsher 570ae7668d0SJeff Kirsher err = ltq_etop_set_mac_address(dev, &mac); 571ae7668d0SJeff Kirsher if (err) 572ae7668d0SJeff Kirsher goto err_netdev; 57343aabec5SDanny Kukawka 57443aabec5SDanny Kukawka /* Set addr_assign_type here, ltq_etop_set_mac_address would reset it. */ 57543aabec5SDanny Kukawka if (random_mac) 576e41b2d7fSJiri Pirko dev->addr_assign_type = NET_ADDR_RANDOM; 57743aabec5SDanny Kukawka 578ae7668d0SJeff Kirsher ltq_etop_set_multicast_list(dev); 579ae7668d0SJeff Kirsher err = ltq_etop_mdio_init(dev); 580ae7668d0SJeff Kirsher if (err) 581ae7668d0SJeff Kirsher goto err_netdev; 582ae7668d0SJeff Kirsher return 0; 583ae7668d0SJeff Kirsher 584ae7668d0SJeff Kirsher err_netdev: 585ae7668d0SJeff Kirsher unregister_netdev(dev); 586ae7668d0SJeff Kirsher free_netdev(dev); 587ae7668d0SJeff Kirsher err_hw: 588ae7668d0SJeff Kirsher ltq_etop_hw_exit(dev); 589ae7668d0SJeff Kirsher return err; 590ae7668d0SJeff Kirsher } 591ae7668d0SJeff Kirsher 592ae7668d0SJeff Kirsher static void 5930290bd29SMichael S. Tsirkin ltq_etop_tx_timeout(struct net_device *dev, unsigned int txqueue) 594ae7668d0SJeff Kirsher { 595ae7668d0SJeff Kirsher int err; 596ae7668d0SJeff Kirsher 597ae7668d0SJeff Kirsher ltq_etop_hw_exit(dev); 598ae7668d0SJeff Kirsher err = ltq_etop_hw_init(dev); 599ae7668d0SJeff Kirsher if (err) 600ae7668d0SJeff Kirsher goto err_hw; 601860e9538SFlorian Westphal netif_trans_update(dev); 602ae7668d0SJeff Kirsher netif_wake_queue(dev); 603ae7668d0SJeff Kirsher return; 604ae7668d0SJeff Kirsher 605ae7668d0SJeff Kirsher err_hw: 606ae7668d0SJeff Kirsher ltq_etop_hw_exit(dev); 607ae7668d0SJeff Kirsher netdev_err(dev, "failed to restart etop after TX timeout\n"); 608ae7668d0SJeff Kirsher } 609ae7668d0SJeff Kirsher 610ae7668d0SJeff Kirsher static const struct net_device_ops ltq_eth_netdev_ops = { 611ae7668d0SJeff Kirsher .ndo_open = ltq_etop_open, 612ae7668d0SJeff Kirsher .ndo_stop = ltq_etop_stop, 613ae7668d0SJeff Kirsher .ndo_start_xmit = ltq_etop_tx, 614ae7668d0SJeff Kirsher .ndo_change_mtu = ltq_etop_change_mtu, 615a7605370SArnd Bergmann .ndo_eth_ioctl = phy_do_ioctl, 616ae7668d0SJeff Kirsher .ndo_set_mac_address = ltq_etop_set_mac_address, 617ae7668d0SJeff Kirsher .ndo_validate_addr = eth_validate_addr, 618afc4b13dSJiri Pirko .ndo_set_rx_mode = ltq_etop_set_multicast_list, 619a4ea8a3dSAlexander Duyck .ndo_select_queue = dev_pick_tx_zero, 620ae7668d0SJeff Kirsher .ndo_init = ltq_etop_init, 621ae7668d0SJeff Kirsher .ndo_tx_timeout = ltq_etop_tx_timeout, 622ae7668d0SJeff Kirsher }; 623ae7668d0SJeff Kirsher 624ae7668d0SJeff Kirsher static int __init 625ae7668d0SJeff Kirsher ltq_etop_probe(struct platform_device *pdev) 626ae7668d0SJeff Kirsher { 627ae7668d0SJeff Kirsher struct net_device *dev; 628ae7668d0SJeff Kirsher struct ltq_etop_priv *priv; 629ae7668d0SJeff Kirsher struct resource *res; 630ae7668d0SJeff Kirsher int err; 631ae7668d0SJeff Kirsher int i; 632ae7668d0SJeff Kirsher 633ae7668d0SJeff Kirsher res = platform_get_resource(pdev, IORESOURCE_MEM, 0); 634ae7668d0SJeff Kirsher if (!res) { 635ae7668d0SJeff Kirsher dev_err(&pdev->dev, "failed to get etop resource\n"); 636ae7668d0SJeff Kirsher err = -ENOENT; 637ae7668d0SJeff Kirsher goto err_out; 638ae7668d0SJeff Kirsher } 639ae7668d0SJeff Kirsher 640ae7668d0SJeff Kirsher res = devm_request_mem_region(&pdev->dev, res->start, 641ae7668d0SJeff Kirsher resource_size(res), dev_name(&pdev->dev)); 642ae7668d0SJeff Kirsher if (!res) { 643ae7668d0SJeff Kirsher dev_err(&pdev->dev, "failed to request etop resource\n"); 644ae7668d0SJeff Kirsher err = -EBUSY; 645ae7668d0SJeff Kirsher goto err_out; 646ae7668d0SJeff Kirsher } 647ae7668d0SJeff Kirsher 6484bdc0d67SChristoph Hellwig ltq_etop_membase = devm_ioremap(&pdev->dev, 649ae7668d0SJeff Kirsher res->start, resource_size(res)); 650ae7668d0SJeff Kirsher if (!ltq_etop_membase) { 651ae7668d0SJeff Kirsher dev_err(&pdev->dev, "failed to remap etop engine %d\n", 652ae7668d0SJeff Kirsher pdev->id); 653ae7668d0SJeff Kirsher err = -ENOMEM; 654ae7668d0SJeff Kirsher goto err_out; 655ae7668d0SJeff Kirsher } 656ae7668d0SJeff Kirsher 657ae7668d0SJeff Kirsher dev = alloc_etherdev_mq(sizeof(struct ltq_etop_priv), 4); 65841de8d4cSJoe Perches if (!dev) { 65941de8d4cSJoe Perches err = -ENOMEM; 66041de8d4cSJoe Perches goto err_out; 66141de8d4cSJoe Perches } 662ae7668d0SJeff Kirsher strcpy(dev->name, "eth%d"); 663ae7668d0SJeff Kirsher dev->netdev_ops = <q_eth_netdev_ops; 664ae7668d0SJeff Kirsher dev->ethtool_ops = <q_etop_ethtool_ops; 665ae7668d0SJeff Kirsher priv = netdev_priv(dev); 666ae7668d0SJeff Kirsher priv->res = res; 667d1b86507SFlorian Fainelli priv->pdev = pdev; 668ae7668d0SJeff Kirsher priv->pldata = dev_get_platdata(&pdev->dev); 669ae7668d0SJeff Kirsher priv->netdev = dev; 670ae7668d0SJeff Kirsher spin_lock_init(&priv->lock); 6719cecb138SFlorian Fainelli SET_NETDEV_DEV(dev, &pdev->dev); 672ae7668d0SJeff Kirsher 673*14d4e308SAleksander Jan Bajkowski err = device_property_read_u32(&pdev->dev, "lantiq,tx-burst-length", &priv->tx_burst_len); 674*14d4e308SAleksander Jan Bajkowski if (err < 0) { 675*14d4e308SAleksander Jan Bajkowski dev_err(&pdev->dev, "unable to read tx-burst-length property\n"); 676*14d4e308SAleksander Jan Bajkowski return err; 677*14d4e308SAleksander Jan Bajkowski } 678*14d4e308SAleksander Jan Bajkowski 679*14d4e308SAleksander Jan Bajkowski err = device_property_read_u32(&pdev->dev, "lantiq,rx-burst-length", &priv->rx_burst_len); 680*14d4e308SAleksander Jan Bajkowski if (err < 0) { 681*14d4e308SAleksander Jan Bajkowski dev_err(&pdev->dev, "unable to read rx-burst-length property\n"); 682*14d4e308SAleksander Jan Bajkowski return err; 683*14d4e308SAleksander Jan Bajkowski } 684*14d4e308SAleksander Jan Bajkowski 685ae7668d0SJeff Kirsher for (i = 0; i < MAX_DMA_CHAN; i++) { 686ae7668d0SJeff Kirsher if (IS_TX(i)) 687ae7668d0SJeff Kirsher netif_napi_add(dev, &priv->ch[i].napi, 688ae7668d0SJeff Kirsher ltq_etop_poll_tx, 8); 689ae7668d0SJeff Kirsher else if (IS_RX(i)) 690ae7668d0SJeff Kirsher netif_napi_add(dev, &priv->ch[i].napi, 691ae7668d0SJeff Kirsher ltq_etop_poll_rx, 32); 692ae7668d0SJeff Kirsher priv->ch[i].netdev = dev; 693ae7668d0SJeff Kirsher } 694ae7668d0SJeff Kirsher 695ae7668d0SJeff Kirsher err = register_netdev(dev); 696ae7668d0SJeff Kirsher if (err) 697ae7668d0SJeff Kirsher goto err_free; 698ae7668d0SJeff Kirsher 699ae7668d0SJeff Kirsher platform_set_drvdata(pdev, dev); 700ae7668d0SJeff Kirsher return 0; 701ae7668d0SJeff Kirsher 702ae7668d0SJeff Kirsher err_free: 703cb0e51d8SWei Yongjun free_netdev(dev); 704ae7668d0SJeff Kirsher err_out: 705ae7668d0SJeff Kirsher return err; 706ae7668d0SJeff Kirsher } 707ae7668d0SJeff Kirsher 708a0a4efedSBill Pemberton static int 709ae7668d0SJeff Kirsher ltq_etop_remove(struct platform_device *pdev) 710ae7668d0SJeff Kirsher { 711ae7668d0SJeff Kirsher struct net_device *dev = platform_get_drvdata(pdev); 712ae7668d0SJeff Kirsher 713ae7668d0SJeff Kirsher if (dev) { 714ae7668d0SJeff Kirsher netif_tx_stop_all_queues(dev); 715ae7668d0SJeff Kirsher ltq_etop_hw_exit(dev); 716ae7668d0SJeff Kirsher ltq_etop_mdio_cleanup(dev); 717ae7668d0SJeff Kirsher unregister_netdev(dev); 718ae7668d0SJeff Kirsher } 719ae7668d0SJeff Kirsher return 0; 720ae7668d0SJeff Kirsher } 721ae7668d0SJeff Kirsher 722ae7668d0SJeff Kirsher static struct platform_driver ltq_mii_driver = { 723a0a4efedSBill Pemberton .remove = ltq_etop_remove, 724ae7668d0SJeff Kirsher .driver = { 725ae7668d0SJeff Kirsher .name = "ltq_etop", 726ae7668d0SJeff Kirsher }, 727ae7668d0SJeff Kirsher }; 728ae7668d0SJeff Kirsher 729ae7668d0SJeff Kirsher int __init 730ae7668d0SJeff Kirsher init_ltq_etop(void) 731ae7668d0SJeff Kirsher { 732ae7668d0SJeff Kirsher int ret = platform_driver_probe(<q_mii_driver, ltq_etop_probe); 733ae7668d0SJeff Kirsher 734ae7668d0SJeff Kirsher if (ret) 735772301b6SMasanari Iida pr_err("ltq_etop: Error registering platform driver!"); 736ae7668d0SJeff Kirsher return ret; 737ae7668d0SJeff Kirsher } 738ae7668d0SJeff Kirsher 739ae7668d0SJeff Kirsher static void __exit 740ae7668d0SJeff Kirsher exit_ltq_etop(void) 741ae7668d0SJeff Kirsher { 742ae7668d0SJeff Kirsher platform_driver_unregister(<q_mii_driver); 743ae7668d0SJeff Kirsher } 744ae7668d0SJeff Kirsher 745ae7668d0SJeff Kirsher module_init(init_ltq_etop); 746ae7668d0SJeff Kirsher module_exit(exit_ltq_etop); 747ae7668d0SJeff Kirsher 748ae7668d0SJeff Kirsher MODULE_AUTHOR("John Crispin <blogic@openwrt.org>"); 749ae7668d0SJeff Kirsher MODULE_DESCRIPTION("Lantiq SoC ETOP"); 750ae7668d0SJeff Kirsher MODULE_LICENSE("GPL"); 751