1adfc5217SJeff Kirsher /* 2adfc5217SJeff Kirsher * Driver for BCM963xx builtin Ethernet mac 3adfc5217SJeff Kirsher * 4adfc5217SJeff Kirsher * Copyright (C) 2008 Maxime Bizon <mbizon@freebox.fr> 5adfc5217SJeff Kirsher * 6adfc5217SJeff Kirsher * This program is free software; you can redistribute it and/or modify 7adfc5217SJeff Kirsher * it under the terms of the GNU General Public License as published by 8adfc5217SJeff Kirsher * the Free Software Foundation; either version 2 of the License, or 9adfc5217SJeff Kirsher * (at your option) any later version. 10adfc5217SJeff Kirsher * 11adfc5217SJeff Kirsher * This program is distributed in the hope that it will be useful, 12adfc5217SJeff Kirsher * but WITHOUT ANY WARRANTY; without even the implied warranty of 13adfc5217SJeff Kirsher * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the 14adfc5217SJeff Kirsher * GNU General Public License for more details. 15adfc5217SJeff Kirsher * 16adfc5217SJeff Kirsher * You should have received a copy of the GNU General Public License 17adfc5217SJeff Kirsher * along with this program; if not, write to the Free Software 18adfc5217SJeff Kirsher * Foundation, Inc., 675 Mass Ave, Cambridge, MA 02139, USA. 19adfc5217SJeff Kirsher */ 20adfc5217SJeff Kirsher #include <linux/init.h> 21adfc5217SJeff Kirsher #include <linux/interrupt.h> 22adfc5217SJeff Kirsher #include <linux/module.h> 23adfc5217SJeff Kirsher #include <linux/clk.h> 24adfc5217SJeff Kirsher #include <linux/etherdevice.h> 25adfc5217SJeff Kirsher #include <linux/slab.h> 26adfc5217SJeff Kirsher #include <linux/delay.h> 27adfc5217SJeff Kirsher #include <linux/ethtool.h> 28adfc5217SJeff Kirsher #include <linux/crc32.h> 29adfc5217SJeff Kirsher #include <linux/err.h> 30adfc5217SJeff Kirsher #include <linux/dma-mapping.h> 31adfc5217SJeff Kirsher #include <linux/platform_device.h> 32adfc5217SJeff Kirsher #include <linux/if_vlan.h> 33adfc5217SJeff Kirsher 34adfc5217SJeff Kirsher #include <bcm63xx_dev_enet.h> 35adfc5217SJeff Kirsher #include "bcm63xx_enet.h" 36adfc5217SJeff Kirsher 37adfc5217SJeff Kirsher static char bcm_enet_driver_name[] = "bcm63xx_enet"; 38adfc5217SJeff Kirsher static char bcm_enet_driver_version[] = "1.0"; 39adfc5217SJeff Kirsher 40adfc5217SJeff Kirsher static int copybreak __read_mostly = 128; 41adfc5217SJeff Kirsher module_param(copybreak, int, 0); 42adfc5217SJeff Kirsher MODULE_PARM_DESC(copybreak, "Receive copy threshold"); 43adfc5217SJeff Kirsher 440ae99b5fSMaxime Bizon /* io registers memory shared between all devices */ 450ae99b5fSMaxime Bizon static void __iomem *bcm_enet_shared_base[3]; 46adfc5217SJeff Kirsher 47adfc5217SJeff Kirsher /* 48adfc5217SJeff Kirsher * io helpers to access mac registers 49adfc5217SJeff Kirsher */ 50adfc5217SJeff Kirsher static inline u32 enet_readl(struct bcm_enet_priv *priv, u32 off) 51adfc5217SJeff Kirsher { 52adfc5217SJeff Kirsher return bcm_readl(priv->base + off); 53adfc5217SJeff Kirsher } 54adfc5217SJeff Kirsher 55adfc5217SJeff Kirsher static inline void enet_writel(struct bcm_enet_priv *priv, 56adfc5217SJeff Kirsher u32 val, u32 off) 57adfc5217SJeff Kirsher { 58adfc5217SJeff Kirsher bcm_writel(val, priv->base + off); 59adfc5217SJeff Kirsher } 60adfc5217SJeff Kirsher 61adfc5217SJeff Kirsher /* 626f00a022SMaxime Bizon * io helpers to access switch registers 63adfc5217SJeff Kirsher */ 646f00a022SMaxime Bizon static inline u32 enetsw_readl(struct bcm_enet_priv *priv, u32 off) 656f00a022SMaxime Bizon { 666f00a022SMaxime Bizon return bcm_readl(priv->base + off); 676f00a022SMaxime Bizon } 686f00a022SMaxime Bizon 696f00a022SMaxime Bizon static inline void enetsw_writel(struct bcm_enet_priv *priv, 706f00a022SMaxime Bizon u32 val, u32 off) 716f00a022SMaxime Bizon { 726f00a022SMaxime Bizon bcm_writel(val, priv->base + off); 736f00a022SMaxime Bizon } 746f00a022SMaxime Bizon 756f00a022SMaxime Bizon static inline u16 enetsw_readw(struct bcm_enet_priv *priv, u32 off) 766f00a022SMaxime Bizon { 776f00a022SMaxime Bizon return bcm_readw(priv->base + off); 786f00a022SMaxime Bizon } 796f00a022SMaxime Bizon 806f00a022SMaxime Bizon static inline void enetsw_writew(struct bcm_enet_priv *priv, 816f00a022SMaxime Bizon u16 val, u32 off) 826f00a022SMaxime Bizon { 836f00a022SMaxime Bizon bcm_writew(val, priv->base + off); 846f00a022SMaxime Bizon } 856f00a022SMaxime Bizon 866f00a022SMaxime Bizon static inline u8 enetsw_readb(struct bcm_enet_priv *priv, u32 off) 876f00a022SMaxime Bizon { 886f00a022SMaxime Bizon return bcm_readb(priv->base + off); 896f00a022SMaxime Bizon } 906f00a022SMaxime Bizon 916f00a022SMaxime Bizon static inline void enetsw_writeb(struct bcm_enet_priv *priv, 926f00a022SMaxime Bizon u8 val, u32 off) 936f00a022SMaxime Bizon { 946f00a022SMaxime Bizon bcm_writeb(val, priv->base + off); 956f00a022SMaxime Bizon } 966f00a022SMaxime Bizon 976f00a022SMaxime Bizon 986f00a022SMaxime Bizon /* io helpers to access shared registers */ 99adfc5217SJeff Kirsher static inline u32 enet_dma_readl(struct bcm_enet_priv *priv, u32 off) 100adfc5217SJeff Kirsher { 1010ae99b5fSMaxime Bizon return bcm_readl(bcm_enet_shared_base[0] + off); 102adfc5217SJeff Kirsher } 103adfc5217SJeff Kirsher 104adfc5217SJeff Kirsher static inline void enet_dma_writel(struct bcm_enet_priv *priv, 105adfc5217SJeff Kirsher u32 val, u32 off) 106adfc5217SJeff Kirsher { 1070ae99b5fSMaxime Bizon bcm_writel(val, bcm_enet_shared_base[0] + off); 1080ae99b5fSMaxime Bizon } 1090ae99b5fSMaxime Bizon 1100ae99b5fSMaxime Bizon static inline u32 enet_dmac_readl(struct bcm_enet_priv *priv, u32 off) 1110ae99b5fSMaxime Bizon { 1120ae99b5fSMaxime Bizon return bcm_readl(bcm_enet_shared_base[1] + off); 1130ae99b5fSMaxime Bizon } 1140ae99b5fSMaxime Bizon 1150ae99b5fSMaxime Bizon static inline void enet_dmac_writel(struct bcm_enet_priv *priv, 1160ae99b5fSMaxime Bizon u32 val, u32 off) 1170ae99b5fSMaxime Bizon { 1180ae99b5fSMaxime Bizon bcm_writel(val, bcm_enet_shared_base[1] + off); 1190ae99b5fSMaxime Bizon } 1200ae99b5fSMaxime Bizon 1210ae99b5fSMaxime Bizon static inline u32 enet_dmas_readl(struct bcm_enet_priv *priv, u32 off) 1220ae99b5fSMaxime Bizon { 1230ae99b5fSMaxime Bizon return bcm_readl(bcm_enet_shared_base[2] + off); 1240ae99b5fSMaxime Bizon } 1250ae99b5fSMaxime Bizon 1260ae99b5fSMaxime Bizon static inline void enet_dmas_writel(struct bcm_enet_priv *priv, 1270ae99b5fSMaxime Bizon u32 val, u32 off) 1280ae99b5fSMaxime Bizon { 1290ae99b5fSMaxime Bizon bcm_writel(val, bcm_enet_shared_base[2] + off); 130adfc5217SJeff Kirsher } 131adfc5217SJeff Kirsher 132adfc5217SJeff Kirsher /* 133adfc5217SJeff Kirsher * write given data into mii register and wait for transfer to end 134adfc5217SJeff Kirsher * with timeout (average measured transfer time is 25us) 135adfc5217SJeff Kirsher */ 136adfc5217SJeff Kirsher static int do_mdio_op(struct bcm_enet_priv *priv, unsigned int data) 137adfc5217SJeff Kirsher { 138adfc5217SJeff Kirsher int limit; 139adfc5217SJeff Kirsher 140adfc5217SJeff Kirsher /* make sure mii interrupt status is cleared */ 141adfc5217SJeff Kirsher enet_writel(priv, ENET_IR_MII, ENET_IR_REG); 142adfc5217SJeff Kirsher 143adfc5217SJeff Kirsher enet_writel(priv, data, ENET_MIIDATA_REG); 144adfc5217SJeff Kirsher wmb(); 145adfc5217SJeff Kirsher 146adfc5217SJeff Kirsher /* busy wait on mii interrupt bit, with timeout */ 147adfc5217SJeff Kirsher limit = 1000; 148adfc5217SJeff Kirsher do { 149adfc5217SJeff Kirsher if (enet_readl(priv, ENET_IR_REG) & ENET_IR_MII) 150adfc5217SJeff Kirsher break; 151adfc5217SJeff Kirsher udelay(1); 152adfc5217SJeff Kirsher } while (limit-- > 0); 153adfc5217SJeff Kirsher 154adfc5217SJeff Kirsher return (limit < 0) ? 1 : 0; 155adfc5217SJeff Kirsher } 156adfc5217SJeff Kirsher 157adfc5217SJeff Kirsher /* 158adfc5217SJeff Kirsher * MII internal read callback 159adfc5217SJeff Kirsher */ 160adfc5217SJeff Kirsher static int bcm_enet_mdio_read(struct bcm_enet_priv *priv, int mii_id, 161adfc5217SJeff Kirsher int regnum) 162adfc5217SJeff Kirsher { 163adfc5217SJeff Kirsher u32 tmp, val; 164adfc5217SJeff Kirsher 165adfc5217SJeff Kirsher tmp = regnum << ENET_MIIDATA_REG_SHIFT; 166adfc5217SJeff Kirsher tmp |= 0x2 << ENET_MIIDATA_TA_SHIFT; 167adfc5217SJeff Kirsher tmp |= mii_id << ENET_MIIDATA_PHYID_SHIFT; 168adfc5217SJeff Kirsher tmp |= ENET_MIIDATA_OP_READ_MASK; 169adfc5217SJeff Kirsher 170adfc5217SJeff Kirsher if (do_mdio_op(priv, tmp)) 171adfc5217SJeff Kirsher return -1; 172adfc5217SJeff Kirsher 173adfc5217SJeff Kirsher val = enet_readl(priv, ENET_MIIDATA_REG); 174adfc5217SJeff Kirsher val &= 0xffff; 175adfc5217SJeff Kirsher return val; 176adfc5217SJeff Kirsher } 177adfc5217SJeff Kirsher 178adfc5217SJeff Kirsher /* 179adfc5217SJeff Kirsher * MII internal write callback 180adfc5217SJeff Kirsher */ 181adfc5217SJeff Kirsher static int bcm_enet_mdio_write(struct bcm_enet_priv *priv, int mii_id, 182adfc5217SJeff Kirsher int regnum, u16 value) 183adfc5217SJeff Kirsher { 184adfc5217SJeff Kirsher u32 tmp; 185adfc5217SJeff Kirsher 186adfc5217SJeff Kirsher tmp = (value & 0xffff) << ENET_MIIDATA_DATA_SHIFT; 187adfc5217SJeff Kirsher tmp |= 0x2 << ENET_MIIDATA_TA_SHIFT; 188adfc5217SJeff Kirsher tmp |= regnum << ENET_MIIDATA_REG_SHIFT; 189adfc5217SJeff Kirsher tmp |= mii_id << ENET_MIIDATA_PHYID_SHIFT; 190adfc5217SJeff Kirsher tmp |= ENET_MIIDATA_OP_WRITE_MASK; 191adfc5217SJeff Kirsher 192adfc5217SJeff Kirsher (void)do_mdio_op(priv, tmp); 193adfc5217SJeff Kirsher return 0; 194adfc5217SJeff Kirsher } 195adfc5217SJeff Kirsher 196adfc5217SJeff Kirsher /* 197adfc5217SJeff Kirsher * MII read callback from phylib 198adfc5217SJeff Kirsher */ 199adfc5217SJeff Kirsher static int bcm_enet_mdio_read_phylib(struct mii_bus *bus, int mii_id, 200adfc5217SJeff Kirsher int regnum) 201adfc5217SJeff Kirsher { 202adfc5217SJeff Kirsher return bcm_enet_mdio_read(bus->priv, mii_id, regnum); 203adfc5217SJeff Kirsher } 204adfc5217SJeff Kirsher 205adfc5217SJeff Kirsher /* 206adfc5217SJeff Kirsher * MII write callback from phylib 207adfc5217SJeff Kirsher */ 208adfc5217SJeff Kirsher static int bcm_enet_mdio_write_phylib(struct mii_bus *bus, int mii_id, 209adfc5217SJeff Kirsher int regnum, u16 value) 210adfc5217SJeff Kirsher { 211adfc5217SJeff Kirsher return bcm_enet_mdio_write(bus->priv, mii_id, regnum, value); 212adfc5217SJeff Kirsher } 213adfc5217SJeff Kirsher 214adfc5217SJeff Kirsher /* 215adfc5217SJeff Kirsher * MII read callback from mii core 216adfc5217SJeff Kirsher */ 217adfc5217SJeff Kirsher static int bcm_enet_mdio_read_mii(struct net_device *dev, int mii_id, 218adfc5217SJeff Kirsher int regnum) 219adfc5217SJeff Kirsher { 220adfc5217SJeff Kirsher return bcm_enet_mdio_read(netdev_priv(dev), mii_id, regnum); 221adfc5217SJeff Kirsher } 222adfc5217SJeff Kirsher 223adfc5217SJeff Kirsher /* 224adfc5217SJeff Kirsher * MII write callback from mii core 225adfc5217SJeff Kirsher */ 226adfc5217SJeff Kirsher static void bcm_enet_mdio_write_mii(struct net_device *dev, int mii_id, 227adfc5217SJeff Kirsher int regnum, int value) 228adfc5217SJeff Kirsher { 229adfc5217SJeff Kirsher bcm_enet_mdio_write(netdev_priv(dev), mii_id, regnum, value); 230adfc5217SJeff Kirsher } 231adfc5217SJeff Kirsher 232adfc5217SJeff Kirsher /* 233adfc5217SJeff Kirsher * refill rx queue 234adfc5217SJeff Kirsher */ 235adfc5217SJeff Kirsher static int bcm_enet_refill_rx(struct net_device *dev) 236adfc5217SJeff Kirsher { 237adfc5217SJeff Kirsher struct bcm_enet_priv *priv; 238adfc5217SJeff Kirsher 239adfc5217SJeff Kirsher priv = netdev_priv(dev); 240adfc5217SJeff Kirsher 241adfc5217SJeff Kirsher while (priv->rx_desc_count < priv->rx_ring_size) { 242adfc5217SJeff Kirsher struct bcm_enet_desc *desc; 243adfc5217SJeff Kirsher struct sk_buff *skb; 244adfc5217SJeff Kirsher dma_addr_t p; 245adfc5217SJeff Kirsher int desc_idx; 246adfc5217SJeff Kirsher u32 len_stat; 247adfc5217SJeff Kirsher 248adfc5217SJeff Kirsher desc_idx = priv->rx_dirty_desc; 249adfc5217SJeff Kirsher desc = &priv->rx_desc_cpu[desc_idx]; 250adfc5217SJeff Kirsher 251adfc5217SJeff Kirsher if (!priv->rx_skb[desc_idx]) { 252adfc5217SJeff Kirsher skb = netdev_alloc_skb(dev, priv->rx_skb_size); 253adfc5217SJeff Kirsher if (!skb) 254adfc5217SJeff Kirsher break; 255adfc5217SJeff Kirsher priv->rx_skb[desc_idx] = skb; 256adfc5217SJeff Kirsher p = dma_map_single(&priv->pdev->dev, skb->data, 257adfc5217SJeff Kirsher priv->rx_skb_size, 258adfc5217SJeff Kirsher DMA_FROM_DEVICE); 259adfc5217SJeff Kirsher desc->address = p; 260adfc5217SJeff Kirsher } 261adfc5217SJeff Kirsher 262adfc5217SJeff Kirsher len_stat = priv->rx_skb_size << DMADESC_LENGTH_SHIFT; 263adfc5217SJeff Kirsher len_stat |= DMADESC_OWNER_MASK; 264adfc5217SJeff Kirsher if (priv->rx_dirty_desc == priv->rx_ring_size - 1) { 265adfc5217SJeff Kirsher len_stat |= DMADESC_WRAP_MASK; 266adfc5217SJeff Kirsher priv->rx_dirty_desc = 0; 267adfc5217SJeff Kirsher } else { 268adfc5217SJeff Kirsher priv->rx_dirty_desc++; 269adfc5217SJeff Kirsher } 270adfc5217SJeff Kirsher wmb(); 271adfc5217SJeff Kirsher desc->len_stat = len_stat; 272adfc5217SJeff Kirsher 273adfc5217SJeff Kirsher priv->rx_desc_count++; 274adfc5217SJeff Kirsher 275adfc5217SJeff Kirsher /* tell dma engine we allocated one buffer */ 276adfc5217SJeff Kirsher enet_dma_writel(priv, 1, ENETDMA_BUFALLOC_REG(priv->rx_chan)); 277adfc5217SJeff Kirsher } 278adfc5217SJeff Kirsher 279adfc5217SJeff Kirsher /* If rx ring is still empty, set a timer to try allocating 280adfc5217SJeff Kirsher * again at a later time. */ 281adfc5217SJeff Kirsher if (priv->rx_desc_count == 0 && netif_running(dev)) { 282adfc5217SJeff Kirsher dev_warn(&priv->pdev->dev, "unable to refill rx ring\n"); 283adfc5217SJeff Kirsher priv->rx_timeout.expires = jiffies + HZ; 284adfc5217SJeff Kirsher add_timer(&priv->rx_timeout); 285adfc5217SJeff Kirsher } 286adfc5217SJeff Kirsher 287adfc5217SJeff Kirsher return 0; 288adfc5217SJeff Kirsher } 289adfc5217SJeff Kirsher 290adfc5217SJeff Kirsher /* 291adfc5217SJeff Kirsher * timer callback to defer refill rx queue in case we're OOM 292adfc5217SJeff Kirsher */ 293adfc5217SJeff Kirsher static void bcm_enet_refill_rx_timer(unsigned long data) 294adfc5217SJeff Kirsher { 295adfc5217SJeff Kirsher struct net_device *dev; 296adfc5217SJeff Kirsher struct bcm_enet_priv *priv; 297adfc5217SJeff Kirsher 298adfc5217SJeff Kirsher dev = (struct net_device *)data; 299adfc5217SJeff Kirsher priv = netdev_priv(dev); 300adfc5217SJeff Kirsher 301adfc5217SJeff Kirsher spin_lock(&priv->rx_lock); 302adfc5217SJeff Kirsher bcm_enet_refill_rx((struct net_device *)data); 303adfc5217SJeff Kirsher spin_unlock(&priv->rx_lock); 304adfc5217SJeff Kirsher } 305adfc5217SJeff Kirsher 306adfc5217SJeff Kirsher /* 307adfc5217SJeff Kirsher * extract packet from rx queue 308adfc5217SJeff Kirsher */ 309adfc5217SJeff Kirsher static int bcm_enet_receive_queue(struct net_device *dev, int budget) 310adfc5217SJeff Kirsher { 311adfc5217SJeff Kirsher struct bcm_enet_priv *priv; 312adfc5217SJeff Kirsher struct device *kdev; 313adfc5217SJeff Kirsher int processed; 314adfc5217SJeff Kirsher 315adfc5217SJeff Kirsher priv = netdev_priv(dev); 316adfc5217SJeff Kirsher kdev = &priv->pdev->dev; 317adfc5217SJeff Kirsher processed = 0; 318adfc5217SJeff Kirsher 319adfc5217SJeff Kirsher /* don't scan ring further than number of refilled 320adfc5217SJeff Kirsher * descriptor */ 321adfc5217SJeff Kirsher if (budget > priv->rx_desc_count) 322adfc5217SJeff Kirsher budget = priv->rx_desc_count; 323adfc5217SJeff Kirsher 324adfc5217SJeff Kirsher do { 325adfc5217SJeff Kirsher struct bcm_enet_desc *desc; 326adfc5217SJeff Kirsher struct sk_buff *skb; 327adfc5217SJeff Kirsher int desc_idx; 328adfc5217SJeff Kirsher u32 len_stat; 329adfc5217SJeff Kirsher unsigned int len; 330adfc5217SJeff Kirsher 331adfc5217SJeff Kirsher desc_idx = priv->rx_curr_desc; 332adfc5217SJeff Kirsher desc = &priv->rx_desc_cpu[desc_idx]; 333adfc5217SJeff Kirsher 334adfc5217SJeff Kirsher /* make sure we actually read the descriptor status at 335adfc5217SJeff Kirsher * each loop */ 336adfc5217SJeff Kirsher rmb(); 337adfc5217SJeff Kirsher 338adfc5217SJeff Kirsher len_stat = desc->len_stat; 339adfc5217SJeff Kirsher 340adfc5217SJeff Kirsher /* break if dma ownership belongs to hw */ 341adfc5217SJeff Kirsher if (len_stat & DMADESC_OWNER_MASK) 342adfc5217SJeff Kirsher break; 343adfc5217SJeff Kirsher 344adfc5217SJeff Kirsher processed++; 345adfc5217SJeff Kirsher priv->rx_curr_desc++; 346adfc5217SJeff Kirsher if (priv->rx_curr_desc == priv->rx_ring_size) 347adfc5217SJeff Kirsher priv->rx_curr_desc = 0; 348adfc5217SJeff Kirsher priv->rx_desc_count--; 349adfc5217SJeff Kirsher 350adfc5217SJeff Kirsher /* if the packet does not have start of packet _and_ 351adfc5217SJeff Kirsher * end of packet flag set, then just recycle it */ 352adfc5217SJeff Kirsher if ((len_stat & DMADESC_ESOP_MASK) != DMADESC_ESOP_MASK) { 353adfc5217SJeff Kirsher dev->stats.rx_dropped++; 354adfc5217SJeff Kirsher continue; 355adfc5217SJeff Kirsher } 356adfc5217SJeff Kirsher 357adfc5217SJeff Kirsher /* recycle packet if it's marked as bad */ 3586f00a022SMaxime Bizon if (!priv->enet_is_sw && 3596f00a022SMaxime Bizon unlikely(len_stat & DMADESC_ERR_MASK)) { 360adfc5217SJeff Kirsher dev->stats.rx_errors++; 361adfc5217SJeff Kirsher 362adfc5217SJeff Kirsher if (len_stat & DMADESC_OVSIZE_MASK) 363adfc5217SJeff Kirsher dev->stats.rx_length_errors++; 364adfc5217SJeff Kirsher if (len_stat & DMADESC_CRC_MASK) 365adfc5217SJeff Kirsher dev->stats.rx_crc_errors++; 366adfc5217SJeff Kirsher if (len_stat & DMADESC_UNDER_MASK) 367adfc5217SJeff Kirsher dev->stats.rx_frame_errors++; 368adfc5217SJeff Kirsher if (len_stat & DMADESC_OV_MASK) 369adfc5217SJeff Kirsher dev->stats.rx_fifo_errors++; 370adfc5217SJeff Kirsher continue; 371adfc5217SJeff Kirsher } 372adfc5217SJeff Kirsher 373adfc5217SJeff Kirsher /* valid packet */ 374adfc5217SJeff Kirsher skb = priv->rx_skb[desc_idx]; 375adfc5217SJeff Kirsher len = (len_stat & DMADESC_LENGTH_MASK) >> DMADESC_LENGTH_SHIFT; 376adfc5217SJeff Kirsher /* don't include FCS */ 377adfc5217SJeff Kirsher len -= 4; 378adfc5217SJeff Kirsher 379adfc5217SJeff Kirsher if (len < copybreak) { 380adfc5217SJeff Kirsher struct sk_buff *nskb; 381adfc5217SJeff Kirsher 382adfc5217SJeff Kirsher nskb = netdev_alloc_skb_ip_align(dev, len); 383adfc5217SJeff Kirsher if (!nskb) { 384adfc5217SJeff Kirsher /* forget packet, just rearm desc */ 385adfc5217SJeff Kirsher dev->stats.rx_dropped++; 386adfc5217SJeff Kirsher continue; 387adfc5217SJeff Kirsher } 388adfc5217SJeff Kirsher 389adfc5217SJeff Kirsher dma_sync_single_for_cpu(kdev, desc->address, 390adfc5217SJeff Kirsher len, DMA_FROM_DEVICE); 391adfc5217SJeff Kirsher memcpy(nskb->data, skb->data, len); 392adfc5217SJeff Kirsher dma_sync_single_for_device(kdev, desc->address, 393adfc5217SJeff Kirsher len, DMA_FROM_DEVICE); 394adfc5217SJeff Kirsher skb = nskb; 395adfc5217SJeff Kirsher } else { 396adfc5217SJeff Kirsher dma_unmap_single(&priv->pdev->dev, desc->address, 397adfc5217SJeff Kirsher priv->rx_skb_size, DMA_FROM_DEVICE); 398adfc5217SJeff Kirsher priv->rx_skb[desc_idx] = NULL; 399adfc5217SJeff Kirsher } 400adfc5217SJeff Kirsher 401adfc5217SJeff Kirsher skb_put(skb, len); 402adfc5217SJeff Kirsher skb->protocol = eth_type_trans(skb, dev); 403adfc5217SJeff Kirsher dev->stats.rx_packets++; 404adfc5217SJeff Kirsher dev->stats.rx_bytes += len; 405adfc5217SJeff Kirsher netif_receive_skb(skb); 406adfc5217SJeff Kirsher 407adfc5217SJeff Kirsher } while (--budget > 0); 408adfc5217SJeff Kirsher 409adfc5217SJeff Kirsher if (processed || !priv->rx_desc_count) { 410adfc5217SJeff Kirsher bcm_enet_refill_rx(dev); 411adfc5217SJeff Kirsher 412adfc5217SJeff Kirsher /* kick rx dma */ 4130ae99b5fSMaxime Bizon enet_dmac_writel(priv, ENETDMAC_CHANCFG_EN_MASK, 4140ae99b5fSMaxime Bizon ENETDMAC_CHANCFG_REG(priv->rx_chan)); 415adfc5217SJeff Kirsher } 416adfc5217SJeff Kirsher 417adfc5217SJeff Kirsher return processed; 418adfc5217SJeff Kirsher } 419adfc5217SJeff Kirsher 420adfc5217SJeff Kirsher 421adfc5217SJeff Kirsher /* 422adfc5217SJeff Kirsher * try to or force reclaim of transmitted buffers 423adfc5217SJeff Kirsher */ 424adfc5217SJeff Kirsher static int bcm_enet_tx_reclaim(struct net_device *dev, int force) 425adfc5217SJeff Kirsher { 426adfc5217SJeff Kirsher struct bcm_enet_priv *priv; 427adfc5217SJeff Kirsher int released; 428adfc5217SJeff Kirsher 429adfc5217SJeff Kirsher priv = netdev_priv(dev); 430adfc5217SJeff Kirsher released = 0; 431adfc5217SJeff Kirsher 432adfc5217SJeff Kirsher while (priv->tx_desc_count < priv->tx_ring_size) { 433adfc5217SJeff Kirsher struct bcm_enet_desc *desc; 434adfc5217SJeff Kirsher struct sk_buff *skb; 435adfc5217SJeff Kirsher 436adfc5217SJeff Kirsher /* We run in a bh and fight against start_xmit, which 437adfc5217SJeff Kirsher * is called with bh disabled */ 438adfc5217SJeff Kirsher spin_lock(&priv->tx_lock); 439adfc5217SJeff Kirsher 440adfc5217SJeff Kirsher desc = &priv->tx_desc_cpu[priv->tx_dirty_desc]; 441adfc5217SJeff Kirsher 442adfc5217SJeff Kirsher if (!force && (desc->len_stat & DMADESC_OWNER_MASK)) { 443adfc5217SJeff Kirsher spin_unlock(&priv->tx_lock); 444adfc5217SJeff Kirsher break; 445adfc5217SJeff Kirsher } 446adfc5217SJeff Kirsher 447adfc5217SJeff Kirsher /* ensure other field of the descriptor were not read 448adfc5217SJeff Kirsher * before we checked ownership */ 449adfc5217SJeff Kirsher rmb(); 450adfc5217SJeff Kirsher 451adfc5217SJeff Kirsher skb = priv->tx_skb[priv->tx_dirty_desc]; 452adfc5217SJeff Kirsher priv->tx_skb[priv->tx_dirty_desc] = NULL; 453adfc5217SJeff Kirsher dma_unmap_single(&priv->pdev->dev, desc->address, skb->len, 454adfc5217SJeff Kirsher DMA_TO_DEVICE); 455adfc5217SJeff Kirsher 456adfc5217SJeff Kirsher priv->tx_dirty_desc++; 457adfc5217SJeff Kirsher if (priv->tx_dirty_desc == priv->tx_ring_size) 458adfc5217SJeff Kirsher priv->tx_dirty_desc = 0; 459adfc5217SJeff Kirsher priv->tx_desc_count++; 460adfc5217SJeff Kirsher 461adfc5217SJeff Kirsher spin_unlock(&priv->tx_lock); 462adfc5217SJeff Kirsher 463adfc5217SJeff Kirsher if (desc->len_stat & DMADESC_UNDER_MASK) 464adfc5217SJeff Kirsher dev->stats.tx_errors++; 465adfc5217SJeff Kirsher 466adfc5217SJeff Kirsher dev_kfree_skb(skb); 467adfc5217SJeff Kirsher released++; 468adfc5217SJeff Kirsher } 469adfc5217SJeff Kirsher 470adfc5217SJeff Kirsher if (netif_queue_stopped(dev) && released) 471adfc5217SJeff Kirsher netif_wake_queue(dev); 472adfc5217SJeff Kirsher 473adfc5217SJeff Kirsher return released; 474adfc5217SJeff Kirsher } 475adfc5217SJeff Kirsher 476adfc5217SJeff Kirsher /* 477adfc5217SJeff Kirsher * poll func, called by network core 478adfc5217SJeff Kirsher */ 479adfc5217SJeff Kirsher static int bcm_enet_poll(struct napi_struct *napi, int budget) 480adfc5217SJeff Kirsher { 481adfc5217SJeff Kirsher struct bcm_enet_priv *priv; 482adfc5217SJeff Kirsher struct net_device *dev; 483adfc5217SJeff Kirsher int tx_work_done, rx_work_done; 484adfc5217SJeff Kirsher 485adfc5217SJeff Kirsher priv = container_of(napi, struct bcm_enet_priv, napi); 486adfc5217SJeff Kirsher dev = priv->net_dev; 487adfc5217SJeff Kirsher 488adfc5217SJeff Kirsher /* ack interrupts */ 4890ae99b5fSMaxime Bizon enet_dmac_writel(priv, ENETDMAC_IR_PKTDONE_MASK, 4900ae99b5fSMaxime Bizon ENETDMAC_IR_REG(priv->rx_chan)); 4910ae99b5fSMaxime Bizon enet_dmac_writel(priv, ENETDMAC_IR_PKTDONE_MASK, 4920ae99b5fSMaxime Bizon ENETDMAC_IR_REG(priv->tx_chan)); 493adfc5217SJeff Kirsher 494adfc5217SJeff Kirsher /* reclaim sent skb */ 495adfc5217SJeff Kirsher tx_work_done = bcm_enet_tx_reclaim(dev, 0); 496adfc5217SJeff Kirsher 497adfc5217SJeff Kirsher spin_lock(&priv->rx_lock); 498adfc5217SJeff Kirsher rx_work_done = bcm_enet_receive_queue(dev, budget); 499adfc5217SJeff Kirsher spin_unlock(&priv->rx_lock); 500adfc5217SJeff Kirsher 501adfc5217SJeff Kirsher if (rx_work_done >= budget || tx_work_done > 0) { 502adfc5217SJeff Kirsher /* rx/tx queue is not yet empty/clean */ 503adfc5217SJeff Kirsher return rx_work_done; 504adfc5217SJeff Kirsher } 505adfc5217SJeff Kirsher 506adfc5217SJeff Kirsher /* no more packet in rx/tx queue, remove device from poll 507adfc5217SJeff Kirsher * queue */ 508adfc5217SJeff Kirsher napi_complete(napi); 509adfc5217SJeff Kirsher 510adfc5217SJeff Kirsher /* restore rx/tx interrupt */ 5110ae99b5fSMaxime Bizon enet_dmac_writel(priv, ENETDMAC_IR_PKTDONE_MASK, 5120ae99b5fSMaxime Bizon ENETDMAC_IRMASK_REG(priv->rx_chan)); 5130ae99b5fSMaxime Bizon enet_dmac_writel(priv, ENETDMAC_IR_PKTDONE_MASK, 5140ae99b5fSMaxime Bizon ENETDMAC_IRMASK_REG(priv->tx_chan)); 515adfc5217SJeff Kirsher 516adfc5217SJeff Kirsher return rx_work_done; 517adfc5217SJeff Kirsher } 518adfc5217SJeff Kirsher 519adfc5217SJeff Kirsher /* 520adfc5217SJeff Kirsher * mac interrupt handler 521adfc5217SJeff Kirsher */ 522adfc5217SJeff Kirsher static irqreturn_t bcm_enet_isr_mac(int irq, void *dev_id) 523adfc5217SJeff Kirsher { 524adfc5217SJeff Kirsher struct net_device *dev; 525adfc5217SJeff Kirsher struct bcm_enet_priv *priv; 526adfc5217SJeff Kirsher u32 stat; 527adfc5217SJeff Kirsher 528adfc5217SJeff Kirsher dev = dev_id; 529adfc5217SJeff Kirsher priv = netdev_priv(dev); 530adfc5217SJeff Kirsher 531adfc5217SJeff Kirsher stat = enet_readl(priv, ENET_IR_REG); 532adfc5217SJeff Kirsher if (!(stat & ENET_IR_MIB)) 533adfc5217SJeff Kirsher return IRQ_NONE; 534adfc5217SJeff Kirsher 535adfc5217SJeff Kirsher /* clear & mask interrupt */ 536adfc5217SJeff Kirsher enet_writel(priv, ENET_IR_MIB, ENET_IR_REG); 537adfc5217SJeff Kirsher enet_writel(priv, 0, ENET_IRMASK_REG); 538adfc5217SJeff Kirsher 539adfc5217SJeff Kirsher /* read mib registers in workqueue */ 540adfc5217SJeff Kirsher schedule_work(&priv->mib_update_task); 541adfc5217SJeff Kirsher 542adfc5217SJeff Kirsher return IRQ_HANDLED; 543adfc5217SJeff Kirsher } 544adfc5217SJeff Kirsher 545adfc5217SJeff Kirsher /* 546adfc5217SJeff Kirsher * rx/tx dma interrupt handler 547adfc5217SJeff Kirsher */ 548adfc5217SJeff Kirsher static irqreturn_t bcm_enet_isr_dma(int irq, void *dev_id) 549adfc5217SJeff Kirsher { 550adfc5217SJeff Kirsher struct net_device *dev; 551adfc5217SJeff Kirsher struct bcm_enet_priv *priv; 552adfc5217SJeff Kirsher 553adfc5217SJeff Kirsher dev = dev_id; 554adfc5217SJeff Kirsher priv = netdev_priv(dev); 555adfc5217SJeff Kirsher 556adfc5217SJeff Kirsher /* mask rx/tx interrupts */ 5570ae99b5fSMaxime Bizon enet_dmac_writel(priv, 0, ENETDMAC_IRMASK_REG(priv->rx_chan)); 5580ae99b5fSMaxime Bizon enet_dmac_writel(priv, 0, ENETDMAC_IRMASK_REG(priv->tx_chan)); 559adfc5217SJeff Kirsher 560adfc5217SJeff Kirsher napi_schedule(&priv->napi); 561adfc5217SJeff Kirsher 562adfc5217SJeff Kirsher return IRQ_HANDLED; 563adfc5217SJeff Kirsher } 564adfc5217SJeff Kirsher 565adfc5217SJeff Kirsher /* 566adfc5217SJeff Kirsher * tx request callback 567adfc5217SJeff Kirsher */ 568adfc5217SJeff Kirsher static int bcm_enet_start_xmit(struct sk_buff *skb, struct net_device *dev) 569adfc5217SJeff Kirsher { 570adfc5217SJeff Kirsher struct bcm_enet_priv *priv; 571adfc5217SJeff Kirsher struct bcm_enet_desc *desc; 572adfc5217SJeff Kirsher u32 len_stat; 573adfc5217SJeff Kirsher int ret; 574adfc5217SJeff Kirsher 575adfc5217SJeff Kirsher priv = netdev_priv(dev); 576adfc5217SJeff Kirsher 577adfc5217SJeff Kirsher /* lock against tx reclaim */ 578adfc5217SJeff Kirsher spin_lock(&priv->tx_lock); 579adfc5217SJeff Kirsher 580adfc5217SJeff Kirsher /* make sure the tx hw queue is not full, should not happen 581adfc5217SJeff Kirsher * since we stop queue before it's the case */ 582adfc5217SJeff Kirsher if (unlikely(!priv->tx_desc_count)) { 583adfc5217SJeff Kirsher netif_stop_queue(dev); 584adfc5217SJeff Kirsher dev_err(&priv->pdev->dev, "xmit called with no tx desc " 585adfc5217SJeff Kirsher "available?\n"); 586adfc5217SJeff Kirsher ret = NETDEV_TX_BUSY; 587adfc5217SJeff Kirsher goto out_unlock; 588adfc5217SJeff Kirsher } 589adfc5217SJeff Kirsher 5906f00a022SMaxime Bizon /* pad small packets sent on a switch device */ 5916f00a022SMaxime Bizon if (priv->enet_is_sw && skb->len < 64) { 5926f00a022SMaxime Bizon int needed = 64 - skb->len; 5936f00a022SMaxime Bizon char *data; 5946f00a022SMaxime Bizon 5956f00a022SMaxime Bizon if (unlikely(skb_tailroom(skb) < needed)) { 5966f00a022SMaxime Bizon struct sk_buff *nskb; 5976f00a022SMaxime Bizon 5986f00a022SMaxime Bizon nskb = skb_copy_expand(skb, 0, needed, GFP_ATOMIC); 5996f00a022SMaxime Bizon if (!nskb) { 6006f00a022SMaxime Bizon ret = NETDEV_TX_BUSY; 6016f00a022SMaxime Bizon goto out_unlock; 6026f00a022SMaxime Bizon } 6036f00a022SMaxime Bizon dev_kfree_skb(skb); 6046f00a022SMaxime Bizon skb = nskb; 6056f00a022SMaxime Bizon } 6066f00a022SMaxime Bizon data = skb_put(skb, needed); 6076f00a022SMaxime Bizon memset(data, 0, needed); 6086f00a022SMaxime Bizon } 6096f00a022SMaxime Bizon 610adfc5217SJeff Kirsher /* point to the next available desc */ 611adfc5217SJeff Kirsher desc = &priv->tx_desc_cpu[priv->tx_curr_desc]; 612adfc5217SJeff Kirsher priv->tx_skb[priv->tx_curr_desc] = skb; 613adfc5217SJeff Kirsher 614adfc5217SJeff Kirsher /* fill descriptor */ 615adfc5217SJeff Kirsher desc->address = dma_map_single(&priv->pdev->dev, skb->data, skb->len, 616adfc5217SJeff Kirsher DMA_TO_DEVICE); 617adfc5217SJeff Kirsher 618adfc5217SJeff Kirsher len_stat = (skb->len << DMADESC_LENGTH_SHIFT) & DMADESC_LENGTH_MASK; 619adfc5217SJeff Kirsher len_stat |= DMADESC_ESOP_MASK | 620adfc5217SJeff Kirsher DMADESC_APPEND_CRC | 621adfc5217SJeff Kirsher DMADESC_OWNER_MASK; 622adfc5217SJeff Kirsher 623adfc5217SJeff Kirsher priv->tx_curr_desc++; 624adfc5217SJeff Kirsher if (priv->tx_curr_desc == priv->tx_ring_size) { 625adfc5217SJeff Kirsher priv->tx_curr_desc = 0; 626adfc5217SJeff Kirsher len_stat |= DMADESC_WRAP_MASK; 627adfc5217SJeff Kirsher } 628adfc5217SJeff Kirsher priv->tx_desc_count--; 629adfc5217SJeff Kirsher 630adfc5217SJeff Kirsher /* dma might be already polling, make sure we update desc 631adfc5217SJeff Kirsher * fields in correct order */ 632adfc5217SJeff Kirsher wmb(); 633adfc5217SJeff Kirsher desc->len_stat = len_stat; 634adfc5217SJeff Kirsher wmb(); 635adfc5217SJeff Kirsher 636adfc5217SJeff Kirsher /* kick tx dma */ 6370ae99b5fSMaxime Bizon enet_dmac_writel(priv, ENETDMAC_CHANCFG_EN_MASK, 6380ae99b5fSMaxime Bizon ENETDMAC_CHANCFG_REG(priv->tx_chan)); 639adfc5217SJeff Kirsher 640adfc5217SJeff Kirsher /* stop queue if no more desc available */ 641adfc5217SJeff Kirsher if (!priv->tx_desc_count) 642adfc5217SJeff Kirsher netif_stop_queue(dev); 643adfc5217SJeff Kirsher 644adfc5217SJeff Kirsher dev->stats.tx_bytes += skb->len; 645adfc5217SJeff Kirsher dev->stats.tx_packets++; 646adfc5217SJeff Kirsher ret = NETDEV_TX_OK; 647adfc5217SJeff Kirsher 648adfc5217SJeff Kirsher out_unlock: 649adfc5217SJeff Kirsher spin_unlock(&priv->tx_lock); 650adfc5217SJeff Kirsher return ret; 651adfc5217SJeff Kirsher } 652adfc5217SJeff Kirsher 653adfc5217SJeff Kirsher /* 654adfc5217SJeff Kirsher * Change the interface's mac address. 655adfc5217SJeff Kirsher */ 656adfc5217SJeff Kirsher static int bcm_enet_set_mac_address(struct net_device *dev, void *p) 657adfc5217SJeff Kirsher { 658adfc5217SJeff Kirsher struct bcm_enet_priv *priv; 659adfc5217SJeff Kirsher struct sockaddr *addr = p; 660adfc5217SJeff Kirsher u32 val; 661adfc5217SJeff Kirsher 662adfc5217SJeff Kirsher priv = netdev_priv(dev); 663adfc5217SJeff Kirsher memcpy(dev->dev_addr, addr->sa_data, ETH_ALEN); 664adfc5217SJeff Kirsher 665adfc5217SJeff Kirsher /* use perfect match register 0 to store my mac address */ 666adfc5217SJeff Kirsher val = (dev->dev_addr[2] << 24) | (dev->dev_addr[3] << 16) | 667adfc5217SJeff Kirsher (dev->dev_addr[4] << 8) | dev->dev_addr[5]; 668adfc5217SJeff Kirsher enet_writel(priv, val, ENET_PML_REG(0)); 669adfc5217SJeff Kirsher 670adfc5217SJeff Kirsher val = (dev->dev_addr[0] << 8 | dev->dev_addr[1]); 671adfc5217SJeff Kirsher val |= ENET_PMH_DATAVALID_MASK; 672adfc5217SJeff Kirsher enet_writel(priv, val, ENET_PMH_REG(0)); 673adfc5217SJeff Kirsher 674adfc5217SJeff Kirsher return 0; 675adfc5217SJeff Kirsher } 676adfc5217SJeff Kirsher 677adfc5217SJeff Kirsher /* 678adfc5217SJeff Kirsher * Change rx mode (promiscuous/allmulti) and update multicast list 679adfc5217SJeff Kirsher */ 680adfc5217SJeff Kirsher static void bcm_enet_set_multicast_list(struct net_device *dev) 681adfc5217SJeff Kirsher { 682adfc5217SJeff Kirsher struct bcm_enet_priv *priv; 683adfc5217SJeff Kirsher struct netdev_hw_addr *ha; 684adfc5217SJeff Kirsher u32 val; 685adfc5217SJeff Kirsher int i; 686adfc5217SJeff Kirsher 687adfc5217SJeff Kirsher priv = netdev_priv(dev); 688adfc5217SJeff Kirsher 689adfc5217SJeff Kirsher val = enet_readl(priv, ENET_RXCFG_REG); 690adfc5217SJeff Kirsher 691adfc5217SJeff Kirsher if (dev->flags & IFF_PROMISC) 692adfc5217SJeff Kirsher val |= ENET_RXCFG_PROMISC_MASK; 693adfc5217SJeff Kirsher else 694adfc5217SJeff Kirsher val &= ~ENET_RXCFG_PROMISC_MASK; 695adfc5217SJeff Kirsher 696adfc5217SJeff Kirsher /* only 3 perfect match registers left, first one is used for 697adfc5217SJeff Kirsher * own mac address */ 698adfc5217SJeff Kirsher if ((dev->flags & IFF_ALLMULTI) || netdev_mc_count(dev) > 3) 699adfc5217SJeff Kirsher val |= ENET_RXCFG_ALLMCAST_MASK; 700adfc5217SJeff Kirsher else 701adfc5217SJeff Kirsher val &= ~ENET_RXCFG_ALLMCAST_MASK; 702adfc5217SJeff Kirsher 703adfc5217SJeff Kirsher /* no need to set perfect match registers if we catch all 704adfc5217SJeff Kirsher * multicast */ 705adfc5217SJeff Kirsher if (val & ENET_RXCFG_ALLMCAST_MASK) { 706adfc5217SJeff Kirsher enet_writel(priv, val, ENET_RXCFG_REG); 707adfc5217SJeff Kirsher return; 708adfc5217SJeff Kirsher } 709adfc5217SJeff Kirsher 710adfc5217SJeff Kirsher i = 0; 711adfc5217SJeff Kirsher netdev_for_each_mc_addr(ha, dev) { 712adfc5217SJeff Kirsher u8 *dmi_addr; 713adfc5217SJeff Kirsher u32 tmp; 714adfc5217SJeff Kirsher 715adfc5217SJeff Kirsher if (i == 3) 716adfc5217SJeff Kirsher break; 717adfc5217SJeff Kirsher /* update perfect match registers */ 718adfc5217SJeff Kirsher dmi_addr = ha->addr; 719adfc5217SJeff Kirsher tmp = (dmi_addr[2] << 24) | (dmi_addr[3] << 16) | 720adfc5217SJeff Kirsher (dmi_addr[4] << 8) | dmi_addr[5]; 721adfc5217SJeff Kirsher enet_writel(priv, tmp, ENET_PML_REG(i + 1)); 722adfc5217SJeff Kirsher 723adfc5217SJeff Kirsher tmp = (dmi_addr[0] << 8 | dmi_addr[1]); 724adfc5217SJeff Kirsher tmp |= ENET_PMH_DATAVALID_MASK; 725adfc5217SJeff Kirsher enet_writel(priv, tmp, ENET_PMH_REG(i++ + 1)); 726adfc5217SJeff Kirsher } 727adfc5217SJeff Kirsher 728adfc5217SJeff Kirsher for (; i < 3; i++) { 729adfc5217SJeff Kirsher enet_writel(priv, 0, ENET_PML_REG(i + 1)); 730adfc5217SJeff Kirsher enet_writel(priv, 0, ENET_PMH_REG(i + 1)); 731adfc5217SJeff Kirsher } 732adfc5217SJeff Kirsher 733adfc5217SJeff Kirsher enet_writel(priv, val, ENET_RXCFG_REG); 734adfc5217SJeff Kirsher } 735adfc5217SJeff Kirsher 736adfc5217SJeff Kirsher /* 737adfc5217SJeff Kirsher * set mac duplex parameters 738adfc5217SJeff Kirsher */ 739adfc5217SJeff Kirsher static void bcm_enet_set_duplex(struct bcm_enet_priv *priv, int fullduplex) 740adfc5217SJeff Kirsher { 741adfc5217SJeff Kirsher u32 val; 742adfc5217SJeff Kirsher 743adfc5217SJeff Kirsher val = enet_readl(priv, ENET_TXCTL_REG); 744adfc5217SJeff Kirsher if (fullduplex) 745adfc5217SJeff Kirsher val |= ENET_TXCTL_FD_MASK; 746adfc5217SJeff Kirsher else 747adfc5217SJeff Kirsher val &= ~ENET_TXCTL_FD_MASK; 748adfc5217SJeff Kirsher enet_writel(priv, val, ENET_TXCTL_REG); 749adfc5217SJeff Kirsher } 750adfc5217SJeff Kirsher 751adfc5217SJeff Kirsher /* 752adfc5217SJeff Kirsher * set mac flow control parameters 753adfc5217SJeff Kirsher */ 754adfc5217SJeff Kirsher static void bcm_enet_set_flow(struct bcm_enet_priv *priv, int rx_en, int tx_en) 755adfc5217SJeff Kirsher { 756adfc5217SJeff Kirsher u32 val; 757adfc5217SJeff Kirsher 758adfc5217SJeff Kirsher /* rx flow control (pause frame handling) */ 759adfc5217SJeff Kirsher val = enet_readl(priv, ENET_RXCFG_REG); 760adfc5217SJeff Kirsher if (rx_en) 761adfc5217SJeff Kirsher val |= ENET_RXCFG_ENFLOW_MASK; 762adfc5217SJeff Kirsher else 763adfc5217SJeff Kirsher val &= ~ENET_RXCFG_ENFLOW_MASK; 764adfc5217SJeff Kirsher enet_writel(priv, val, ENET_RXCFG_REG); 765adfc5217SJeff Kirsher 766adfc5217SJeff Kirsher /* tx flow control (pause frame generation) */ 767adfc5217SJeff Kirsher val = enet_dma_readl(priv, ENETDMA_CFG_REG); 768adfc5217SJeff Kirsher if (tx_en) 769adfc5217SJeff Kirsher val |= ENETDMA_CFG_FLOWCH_MASK(priv->rx_chan); 770adfc5217SJeff Kirsher else 771adfc5217SJeff Kirsher val &= ~ENETDMA_CFG_FLOWCH_MASK(priv->rx_chan); 772adfc5217SJeff Kirsher enet_dma_writel(priv, val, ENETDMA_CFG_REG); 773adfc5217SJeff Kirsher } 774adfc5217SJeff Kirsher 775adfc5217SJeff Kirsher /* 776adfc5217SJeff Kirsher * link changed callback (from phylib) 777adfc5217SJeff Kirsher */ 778adfc5217SJeff Kirsher static void bcm_enet_adjust_phy_link(struct net_device *dev) 779adfc5217SJeff Kirsher { 780adfc5217SJeff Kirsher struct bcm_enet_priv *priv; 781adfc5217SJeff Kirsher struct phy_device *phydev; 782adfc5217SJeff Kirsher int status_changed; 783adfc5217SJeff Kirsher 784adfc5217SJeff Kirsher priv = netdev_priv(dev); 785adfc5217SJeff Kirsher phydev = priv->phydev; 786adfc5217SJeff Kirsher status_changed = 0; 787adfc5217SJeff Kirsher 788adfc5217SJeff Kirsher if (priv->old_link != phydev->link) { 789adfc5217SJeff Kirsher status_changed = 1; 790adfc5217SJeff Kirsher priv->old_link = phydev->link; 791adfc5217SJeff Kirsher } 792adfc5217SJeff Kirsher 793adfc5217SJeff Kirsher /* reflect duplex change in mac configuration */ 794adfc5217SJeff Kirsher if (phydev->link && phydev->duplex != priv->old_duplex) { 795adfc5217SJeff Kirsher bcm_enet_set_duplex(priv, 796adfc5217SJeff Kirsher (phydev->duplex == DUPLEX_FULL) ? 1 : 0); 797adfc5217SJeff Kirsher status_changed = 1; 798adfc5217SJeff Kirsher priv->old_duplex = phydev->duplex; 799adfc5217SJeff Kirsher } 800adfc5217SJeff Kirsher 801adfc5217SJeff Kirsher /* enable flow control if remote advertise it (trust phylib to 802adfc5217SJeff Kirsher * check that duplex is full */ 803adfc5217SJeff Kirsher if (phydev->link && phydev->pause != priv->old_pause) { 804adfc5217SJeff Kirsher int rx_pause_en, tx_pause_en; 805adfc5217SJeff Kirsher 806adfc5217SJeff Kirsher if (phydev->pause) { 807adfc5217SJeff Kirsher /* pause was advertised by lpa and us */ 808adfc5217SJeff Kirsher rx_pause_en = 1; 809adfc5217SJeff Kirsher tx_pause_en = 1; 810adfc5217SJeff Kirsher } else if (!priv->pause_auto) { 811adfc5217SJeff Kirsher /* pause setting overrided by user */ 812adfc5217SJeff Kirsher rx_pause_en = priv->pause_rx; 813adfc5217SJeff Kirsher tx_pause_en = priv->pause_tx; 814adfc5217SJeff Kirsher } else { 815adfc5217SJeff Kirsher rx_pause_en = 0; 816adfc5217SJeff Kirsher tx_pause_en = 0; 817adfc5217SJeff Kirsher } 818adfc5217SJeff Kirsher 819adfc5217SJeff Kirsher bcm_enet_set_flow(priv, rx_pause_en, tx_pause_en); 820adfc5217SJeff Kirsher status_changed = 1; 821adfc5217SJeff Kirsher priv->old_pause = phydev->pause; 822adfc5217SJeff Kirsher } 823adfc5217SJeff Kirsher 824adfc5217SJeff Kirsher if (status_changed) { 825adfc5217SJeff Kirsher pr_info("%s: link %s", dev->name, phydev->link ? 826adfc5217SJeff Kirsher "UP" : "DOWN"); 827adfc5217SJeff Kirsher if (phydev->link) 828adfc5217SJeff Kirsher pr_cont(" - %d/%s - flow control %s", phydev->speed, 829adfc5217SJeff Kirsher DUPLEX_FULL == phydev->duplex ? "full" : "half", 830adfc5217SJeff Kirsher phydev->pause == 1 ? "rx&tx" : "off"); 831adfc5217SJeff Kirsher 832adfc5217SJeff Kirsher pr_cont("\n"); 833adfc5217SJeff Kirsher } 834adfc5217SJeff Kirsher } 835adfc5217SJeff Kirsher 836adfc5217SJeff Kirsher /* 837adfc5217SJeff Kirsher * link changed callback (if phylib is not used) 838adfc5217SJeff Kirsher */ 839adfc5217SJeff Kirsher static void bcm_enet_adjust_link(struct net_device *dev) 840adfc5217SJeff Kirsher { 841adfc5217SJeff Kirsher struct bcm_enet_priv *priv; 842adfc5217SJeff Kirsher 843adfc5217SJeff Kirsher priv = netdev_priv(dev); 844adfc5217SJeff Kirsher bcm_enet_set_duplex(priv, priv->force_duplex_full); 845adfc5217SJeff Kirsher bcm_enet_set_flow(priv, priv->pause_rx, priv->pause_tx); 846adfc5217SJeff Kirsher netif_carrier_on(dev); 847adfc5217SJeff Kirsher 848adfc5217SJeff Kirsher pr_info("%s: link forced UP - %d/%s - flow control %s/%s\n", 849adfc5217SJeff Kirsher dev->name, 850adfc5217SJeff Kirsher priv->force_speed_100 ? 100 : 10, 851adfc5217SJeff Kirsher priv->force_duplex_full ? "full" : "half", 852adfc5217SJeff Kirsher priv->pause_rx ? "rx" : "off", 853adfc5217SJeff Kirsher priv->pause_tx ? "tx" : "off"); 854adfc5217SJeff Kirsher } 855adfc5217SJeff Kirsher 856adfc5217SJeff Kirsher /* 857adfc5217SJeff Kirsher * open callback, allocate dma rings & buffers and start rx operation 858adfc5217SJeff Kirsher */ 859adfc5217SJeff Kirsher static int bcm_enet_open(struct net_device *dev) 860adfc5217SJeff Kirsher { 861adfc5217SJeff Kirsher struct bcm_enet_priv *priv; 862adfc5217SJeff Kirsher struct sockaddr addr; 863adfc5217SJeff Kirsher struct device *kdev; 864adfc5217SJeff Kirsher struct phy_device *phydev; 865adfc5217SJeff Kirsher int i, ret; 866adfc5217SJeff Kirsher unsigned int size; 867adfc5217SJeff Kirsher char phy_id[MII_BUS_ID_SIZE + 3]; 868adfc5217SJeff Kirsher void *p; 869adfc5217SJeff Kirsher u32 val; 870adfc5217SJeff Kirsher 871adfc5217SJeff Kirsher priv = netdev_priv(dev); 872adfc5217SJeff Kirsher kdev = &priv->pdev->dev; 873adfc5217SJeff Kirsher 874adfc5217SJeff Kirsher if (priv->has_phy) { 875adfc5217SJeff Kirsher /* connect to PHY */ 876adfc5217SJeff Kirsher snprintf(phy_id, sizeof(phy_id), PHY_ID_FMT, 877c56e9e2aSFlorian Fainelli priv->mii_bus->id, priv->phy_id); 878adfc5217SJeff Kirsher 879f9a8f83bSFlorian Fainelli phydev = phy_connect(dev, phy_id, bcm_enet_adjust_phy_link, 880adfc5217SJeff Kirsher PHY_INTERFACE_MODE_MII); 881adfc5217SJeff Kirsher 882adfc5217SJeff Kirsher if (IS_ERR(phydev)) { 883adfc5217SJeff Kirsher dev_err(kdev, "could not attach to PHY\n"); 884adfc5217SJeff Kirsher return PTR_ERR(phydev); 885adfc5217SJeff Kirsher } 886adfc5217SJeff Kirsher 887adfc5217SJeff Kirsher /* mask with MAC supported features */ 888adfc5217SJeff Kirsher phydev->supported &= (SUPPORTED_10baseT_Half | 889adfc5217SJeff Kirsher SUPPORTED_10baseT_Full | 890adfc5217SJeff Kirsher SUPPORTED_100baseT_Half | 891adfc5217SJeff Kirsher SUPPORTED_100baseT_Full | 892adfc5217SJeff Kirsher SUPPORTED_Autoneg | 893adfc5217SJeff Kirsher SUPPORTED_Pause | 894adfc5217SJeff Kirsher SUPPORTED_MII); 895adfc5217SJeff Kirsher phydev->advertising = phydev->supported; 896adfc5217SJeff Kirsher 897adfc5217SJeff Kirsher if (priv->pause_auto && priv->pause_rx && priv->pause_tx) 898adfc5217SJeff Kirsher phydev->advertising |= SUPPORTED_Pause; 899adfc5217SJeff Kirsher else 900adfc5217SJeff Kirsher phydev->advertising &= ~SUPPORTED_Pause; 901adfc5217SJeff Kirsher 902adfc5217SJeff Kirsher dev_info(kdev, "attached PHY at address %d [%s]\n", 903adfc5217SJeff Kirsher phydev->addr, phydev->drv->name); 904adfc5217SJeff Kirsher 905adfc5217SJeff Kirsher priv->old_link = 0; 906adfc5217SJeff Kirsher priv->old_duplex = -1; 907adfc5217SJeff Kirsher priv->old_pause = -1; 908adfc5217SJeff Kirsher priv->phydev = phydev; 909adfc5217SJeff Kirsher } 910adfc5217SJeff Kirsher 911adfc5217SJeff Kirsher /* mask all interrupts and request them */ 912adfc5217SJeff Kirsher enet_writel(priv, 0, ENET_IRMASK_REG); 9130ae99b5fSMaxime Bizon enet_dmac_writel(priv, 0, ENETDMAC_IRMASK_REG(priv->rx_chan)); 9140ae99b5fSMaxime Bizon enet_dmac_writel(priv, 0, ENETDMAC_IRMASK_REG(priv->tx_chan)); 915adfc5217SJeff Kirsher 916adfc5217SJeff Kirsher ret = request_irq(dev->irq, bcm_enet_isr_mac, 0, dev->name, dev); 917adfc5217SJeff Kirsher if (ret) 918adfc5217SJeff Kirsher goto out_phy_disconnect; 919adfc5217SJeff Kirsher 920adfc5217SJeff Kirsher ret = request_irq(priv->irq_rx, bcm_enet_isr_dma, IRQF_DISABLED, 921adfc5217SJeff Kirsher dev->name, dev); 922adfc5217SJeff Kirsher if (ret) 923adfc5217SJeff Kirsher goto out_freeirq; 924adfc5217SJeff Kirsher 925adfc5217SJeff Kirsher ret = request_irq(priv->irq_tx, bcm_enet_isr_dma, 926adfc5217SJeff Kirsher IRQF_DISABLED, dev->name, dev); 927adfc5217SJeff Kirsher if (ret) 928adfc5217SJeff Kirsher goto out_freeirq_rx; 929adfc5217SJeff Kirsher 930adfc5217SJeff Kirsher /* initialize perfect match registers */ 931adfc5217SJeff Kirsher for (i = 0; i < 4; i++) { 932adfc5217SJeff Kirsher enet_writel(priv, 0, ENET_PML_REG(i)); 933adfc5217SJeff Kirsher enet_writel(priv, 0, ENET_PMH_REG(i)); 934adfc5217SJeff Kirsher } 935adfc5217SJeff Kirsher 936adfc5217SJeff Kirsher /* write device mac address */ 937adfc5217SJeff Kirsher memcpy(addr.sa_data, dev->dev_addr, ETH_ALEN); 938adfc5217SJeff Kirsher bcm_enet_set_mac_address(dev, &addr); 939adfc5217SJeff Kirsher 940adfc5217SJeff Kirsher /* allocate rx dma ring */ 941adfc5217SJeff Kirsher size = priv->rx_ring_size * sizeof(struct bcm_enet_desc); 9421f9061d2SJoe Perches p = dma_alloc_coherent(kdev, size, &priv->rx_desc_dma, 9431f9061d2SJoe Perches GFP_KERNEL | __GFP_ZERO); 944adfc5217SJeff Kirsher if (!p) { 945adfc5217SJeff Kirsher ret = -ENOMEM; 946adfc5217SJeff Kirsher goto out_freeirq_tx; 947adfc5217SJeff Kirsher } 948adfc5217SJeff Kirsher 949adfc5217SJeff Kirsher priv->rx_desc_alloc_size = size; 950adfc5217SJeff Kirsher priv->rx_desc_cpu = p; 951adfc5217SJeff Kirsher 952adfc5217SJeff Kirsher /* allocate tx dma ring */ 953adfc5217SJeff Kirsher size = priv->tx_ring_size * sizeof(struct bcm_enet_desc); 9541f9061d2SJoe Perches p = dma_alloc_coherent(kdev, size, &priv->tx_desc_dma, 9551f9061d2SJoe Perches GFP_KERNEL | __GFP_ZERO); 956adfc5217SJeff Kirsher if (!p) { 957adfc5217SJeff Kirsher ret = -ENOMEM; 958adfc5217SJeff Kirsher goto out_free_rx_ring; 959adfc5217SJeff Kirsher } 960adfc5217SJeff Kirsher 961adfc5217SJeff Kirsher priv->tx_desc_alloc_size = size; 962adfc5217SJeff Kirsher priv->tx_desc_cpu = p; 963adfc5217SJeff Kirsher 964b2adaca9SJoe Perches priv->tx_skb = kcalloc(priv->tx_ring_size, sizeof(struct sk_buff *), 965adfc5217SJeff Kirsher GFP_KERNEL); 966adfc5217SJeff Kirsher if (!priv->tx_skb) { 967adfc5217SJeff Kirsher ret = -ENOMEM; 968adfc5217SJeff Kirsher goto out_free_tx_ring; 969adfc5217SJeff Kirsher } 970adfc5217SJeff Kirsher 971adfc5217SJeff Kirsher priv->tx_desc_count = priv->tx_ring_size; 972adfc5217SJeff Kirsher priv->tx_dirty_desc = 0; 973adfc5217SJeff Kirsher priv->tx_curr_desc = 0; 974adfc5217SJeff Kirsher spin_lock_init(&priv->tx_lock); 975adfc5217SJeff Kirsher 976adfc5217SJeff Kirsher /* init & fill rx ring with skbs */ 977b2adaca9SJoe Perches priv->rx_skb = kcalloc(priv->rx_ring_size, sizeof(struct sk_buff *), 978adfc5217SJeff Kirsher GFP_KERNEL); 979adfc5217SJeff Kirsher if (!priv->rx_skb) { 980adfc5217SJeff Kirsher ret = -ENOMEM; 981adfc5217SJeff Kirsher goto out_free_tx_skb; 982adfc5217SJeff Kirsher } 983adfc5217SJeff Kirsher 984adfc5217SJeff Kirsher priv->rx_desc_count = 0; 985adfc5217SJeff Kirsher priv->rx_dirty_desc = 0; 986adfc5217SJeff Kirsher priv->rx_curr_desc = 0; 987adfc5217SJeff Kirsher 988adfc5217SJeff Kirsher /* initialize flow control buffer allocation */ 989adfc5217SJeff Kirsher enet_dma_writel(priv, ENETDMA_BUFALLOC_FORCE_MASK | 0, 990adfc5217SJeff Kirsher ENETDMA_BUFALLOC_REG(priv->rx_chan)); 991adfc5217SJeff Kirsher 992adfc5217SJeff Kirsher if (bcm_enet_refill_rx(dev)) { 993adfc5217SJeff Kirsher dev_err(kdev, "cannot allocate rx skb queue\n"); 994adfc5217SJeff Kirsher ret = -ENOMEM; 995adfc5217SJeff Kirsher goto out; 996adfc5217SJeff Kirsher } 997adfc5217SJeff Kirsher 998adfc5217SJeff Kirsher /* write rx & tx ring addresses */ 9990ae99b5fSMaxime Bizon enet_dmas_writel(priv, priv->rx_desc_dma, 10000ae99b5fSMaxime Bizon ENETDMAS_RSTART_REG(priv->rx_chan)); 10010ae99b5fSMaxime Bizon enet_dmas_writel(priv, priv->tx_desc_dma, 10020ae99b5fSMaxime Bizon ENETDMAS_RSTART_REG(priv->tx_chan)); 1003adfc5217SJeff Kirsher 1004adfc5217SJeff Kirsher /* clear remaining state ram for rx & tx channel */ 10050ae99b5fSMaxime Bizon enet_dmas_writel(priv, 0, ENETDMAS_SRAM2_REG(priv->rx_chan)); 10060ae99b5fSMaxime Bizon enet_dmas_writel(priv, 0, ENETDMAS_SRAM2_REG(priv->tx_chan)); 10070ae99b5fSMaxime Bizon enet_dmas_writel(priv, 0, ENETDMAS_SRAM3_REG(priv->rx_chan)); 10080ae99b5fSMaxime Bizon enet_dmas_writel(priv, 0, ENETDMAS_SRAM3_REG(priv->tx_chan)); 10090ae99b5fSMaxime Bizon enet_dmas_writel(priv, 0, ENETDMAS_SRAM4_REG(priv->rx_chan)); 10100ae99b5fSMaxime Bizon enet_dmas_writel(priv, 0, ENETDMAS_SRAM4_REG(priv->tx_chan)); 1011adfc5217SJeff Kirsher 1012adfc5217SJeff Kirsher /* set max rx/tx length */ 1013adfc5217SJeff Kirsher enet_writel(priv, priv->hw_mtu, ENET_RXMAXLEN_REG); 1014adfc5217SJeff Kirsher enet_writel(priv, priv->hw_mtu, ENET_TXMAXLEN_REG); 1015adfc5217SJeff Kirsher 1016adfc5217SJeff Kirsher /* set dma maximum burst len */ 10176f00a022SMaxime Bizon enet_dmac_writel(priv, priv->dma_maxburst, 10180ae99b5fSMaxime Bizon ENETDMAC_MAXBURST_REG(priv->rx_chan)); 10196f00a022SMaxime Bizon enet_dmac_writel(priv, priv->dma_maxburst, 10200ae99b5fSMaxime Bizon ENETDMAC_MAXBURST_REG(priv->tx_chan)); 1021adfc5217SJeff Kirsher 1022adfc5217SJeff Kirsher /* set correct transmit fifo watermark */ 1023adfc5217SJeff Kirsher enet_writel(priv, BCMENET_TX_FIFO_TRESH, ENET_TXWMARK_REG); 1024adfc5217SJeff Kirsher 1025adfc5217SJeff Kirsher /* set flow control low/high threshold to 1/3 / 2/3 */ 1026adfc5217SJeff Kirsher val = priv->rx_ring_size / 3; 1027adfc5217SJeff Kirsher enet_dma_writel(priv, val, ENETDMA_FLOWCL_REG(priv->rx_chan)); 1028adfc5217SJeff Kirsher val = (priv->rx_ring_size * 2) / 3; 1029adfc5217SJeff Kirsher enet_dma_writel(priv, val, ENETDMA_FLOWCH_REG(priv->rx_chan)); 1030adfc5217SJeff Kirsher 1031adfc5217SJeff Kirsher /* all set, enable mac and interrupts, start dma engine and 1032adfc5217SJeff Kirsher * kick rx dma channel */ 1033adfc5217SJeff Kirsher wmb(); 1034adfc5217SJeff Kirsher val = enet_readl(priv, ENET_CTL_REG); 1035adfc5217SJeff Kirsher val |= ENET_CTL_ENABLE_MASK; 1036adfc5217SJeff Kirsher enet_writel(priv, val, ENET_CTL_REG); 1037adfc5217SJeff Kirsher enet_dma_writel(priv, ENETDMA_CFG_EN_MASK, ENETDMA_CFG_REG); 10380ae99b5fSMaxime Bizon enet_dmac_writel(priv, ENETDMAC_CHANCFG_EN_MASK, 10390ae99b5fSMaxime Bizon ENETDMAC_CHANCFG_REG(priv->rx_chan)); 1040adfc5217SJeff Kirsher 1041adfc5217SJeff Kirsher /* watch "mib counters about to overflow" interrupt */ 1042adfc5217SJeff Kirsher enet_writel(priv, ENET_IR_MIB, ENET_IR_REG); 1043adfc5217SJeff Kirsher enet_writel(priv, ENET_IR_MIB, ENET_IRMASK_REG); 1044adfc5217SJeff Kirsher 1045adfc5217SJeff Kirsher /* watch "packet transferred" interrupt in rx and tx */ 10460ae99b5fSMaxime Bizon enet_dmac_writel(priv, ENETDMAC_IR_PKTDONE_MASK, 10470ae99b5fSMaxime Bizon ENETDMAC_IR_REG(priv->rx_chan)); 10480ae99b5fSMaxime Bizon enet_dmac_writel(priv, ENETDMAC_IR_PKTDONE_MASK, 10490ae99b5fSMaxime Bizon ENETDMAC_IR_REG(priv->tx_chan)); 1050adfc5217SJeff Kirsher 1051adfc5217SJeff Kirsher /* make sure we enable napi before rx interrupt */ 1052adfc5217SJeff Kirsher napi_enable(&priv->napi); 1053adfc5217SJeff Kirsher 10540ae99b5fSMaxime Bizon enet_dmac_writel(priv, ENETDMAC_IR_PKTDONE_MASK, 10550ae99b5fSMaxime Bizon ENETDMAC_IRMASK_REG(priv->rx_chan)); 10560ae99b5fSMaxime Bizon enet_dmac_writel(priv, ENETDMAC_IR_PKTDONE_MASK, 10570ae99b5fSMaxime Bizon ENETDMAC_IRMASK_REG(priv->tx_chan)); 1058adfc5217SJeff Kirsher 1059adfc5217SJeff Kirsher if (priv->has_phy) 1060adfc5217SJeff Kirsher phy_start(priv->phydev); 1061adfc5217SJeff Kirsher else 1062adfc5217SJeff Kirsher bcm_enet_adjust_link(dev); 1063adfc5217SJeff Kirsher 1064adfc5217SJeff Kirsher netif_start_queue(dev); 1065adfc5217SJeff Kirsher return 0; 1066adfc5217SJeff Kirsher 1067adfc5217SJeff Kirsher out: 1068adfc5217SJeff Kirsher for (i = 0; i < priv->rx_ring_size; i++) { 1069adfc5217SJeff Kirsher struct bcm_enet_desc *desc; 1070adfc5217SJeff Kirsher 1071adfc5217SJeff Kirsher if (!priv->rx_skb[i]) 1072adfc5217SJeff Kirsher continue; 1073adfc5217SJeff Kirsher 1074adfc5217SJeff Kirsher desc = &priv->rx_desc_cpu[i]; 1075adfc5217SJeff Kirsher dma_unmap_single(kdev, desc->address, priv->rx_skb_size, 1076adfc5217SJeff Kirsher DMA_FROM_DEVICE); 1077adfc5217SJeff Kirsher kfree_skb(priv->rx_skb[i]); 1078adfc5217SJeff Kirsher } 1079adfc5217SJeff Kirsher kfree(priv->rx_skb); 1080adfc5217SJeff Kirsher 1081adfc5217SJeff Kirsher out_free_tx_skb: 1082adfc5217SJeff Kirsher kfree(priv->tx_skb); 1083adfc5217SJeff Kirsher 1084adfc5217SJeff Kirsher out_free_tx_ring: 1085adfc5217SJeff Kirsher dma_free_coherent(kdev, priv->tx_desc_alloc_size, 1086adfc5217SJeff Kirsher priv->tx_desc_cpu, priv->tx_desc_dma); 1087adfc5217SJeff Kirsher 1088adfc5217SJeff Kirsher out_free_rx_ring: 1089adfc5217SJeff Kirsher dma_free_coherent(kdev, priv->rx_desc_alloc_size, 1090adfc5217SJeff Kirsher priv->rx_desc_cpu, priv->rx_desc_dma); 1091adfc5217SJeff Kirsher 1092adfc5217SJeff Kirsher out_freeirq_tx: 1093adfc5217SJeff Kirsher free_irq(priv->irq_tx, dev); 1094adfc5217SJeff Kirsher 1095adfc5217SJeff Kirsher out_freeirq_rx: 1096adfc5217SJeff Kirsher free_irq(priv->irq_rx, dev); 1097adfc5217SJeff Kirsher 1098adfc5217SJeff Kirsher out_freeirq: 1099adfc5217SJeff Kirsher free_irq(dev->irq, dev); 1100adfc5217SJeff Kirsher 1101adfc5217SJeff Kirsher out_phy_disconnect: 1102adfc5217SJeff Kirsher phy_disconnect(priv->phydev); 1103adfc5217SJeff Kirsher 1104adfc5217SJeff Kirsher return ret; 1105adfc5217SJeff Kirsher } 1106adfc5217SJeff Kirsher 1107adfc5217SJeff Kirsher /* 1108adfc5217SJeff Kirsher * disable mac 1109adfc5217SJeff Kirsher */ 1110adfc5217SJeff Kirsher static void bcm_enet_disable_mac(struct bcm_enet_priv *priv) 1111adfc5217SJeff Kirsher { 1112adfc5217SJeff Kirsher int limit; 1113adfc5217SJeff Kirsher u32 val; 1114adfc5217SJeff Kirsher 1115adfc5217SJeff Kirsher val = enet_readl(priv, ENET_CTL_REG); 1116adfc5217SJeff Kirsher val |= ENET_CTL_DISABLE_MASK; 1117adfc5217SJeff Kirsher enet_writel(priv, val, ENET_CTL_REG); 1118adfc5217SJeff Kirsher 1119adfc5217SJeff Kirsher limit = 1000; 1120adfc5217SJeff Kirsher do { 1121adfc5217SJeff Kirsher u32 val; 1122adfc5217SJeff Kirsher 1123adfc5217SJeff Kirsher val = enet_readl(priv, ENET_CTL_REG); 1124adfc5217SJeff Kirsher if (!(val & ENET_CTL_DISABLE_MASK)) 1125adfc5217SJeff Kirsher break; 1126adfc5217SJeff Kirsher udelay(1); 1127adfc5217SJeff Kirsher } while (limit--); 1128adfc5217SJeff Kirsher } 1129adfc5217SJeff Kirsher 1130adfc5217SJeff Kirsher /* 1131adfc5217SJeff Kirsher * disable dma in given channel 1132adfc5217SJeff Kirsher */ 1133adfc5217SJeff Kirsher static void bcm_enet_disable_dma(struct bcm_enet_priv *priv, int chan) 1134adfc5217SJeff Kirsher { 1135adfc5217SJeff Kirsher int limit; 1136adfc5217SJeff Kirsher 11370ae99b5fSMaxime Bizon enet_dmac_writel(priv, 0, ENETDMAC_CHANCFG_REG(chan)); 1138adfc5217SJeff Kirsher 1139adfc5217SJeff Kirsher limit = 1000; 1140adfc5217SJeff Kirsher do { 1141adfc5217SJeff Kirsher u32 val; 1142adfc5217SJeff Kirsher 11430ae99b5fSMaxime Bizon val = enet_dmac_readl(priv, ENETDMAC_CHANCFG_REG(chan)); 11440ae99b5fSMaxime Bizon if (!(val & ENETDMAC_CHANCFG_EN_MASK)) 1145adfc5217SJeff Kirsher break; 1146adfc5217SJeff Kirsher udelay(1); 1147adfc5217SJeff Kirsher } while (limit--); 1148adfc5217SJeff Kirsher } 1149adfc5217SJeff Kirsher 1150adfc5217SJeff Kirsher /* 1151adfc5217SJeff Kirsher * stop callback 1152adfc5217SJeff Kirsher */ 1153adfc5217SJeff Kirsher static int bcm_enet_stop(struct net_device *dev) 1154adfc5217SJeff Kirsher { 1155adfc5217SJeff Kirsher struct bcm_enet_priv *priv; 1156adfc5217SJeff Kirsher struct device *kdev; 1157adfc5217SJeff Kirsher int i; 1158adfc5217SJeff Kirsher 1159adfc5217SJeff Kirsher priv = netdev_priv(dev); 1160adfc5217SJeff Kirsher kdev = &priv->pdev->dev; 1161adfc5217SJeff Kirsher 1162adfc5217SJeff Kirsher netif_stop_queue(dev); 1163adfc5217SJeff Kirsher napi_disable(&priv->napi); 1164adfc5217SJeff Kirsher if (priv->has_phy) 1165adfc5217SJeff Kirsher phy_stop(priv->phydev); 1166adfc5217SJeff Kirsher del_timer_sync(&priv->rx_timeout); 1167adfc5217SJeff Kirsher 1168adfc5217SJeff Kirsher /* mask all interrupts */ 1169adfc5217SJeff Kirsher enet_writel(priv, 0, ENET_IRMASK_REG); 11700ae99b5fSMaxime Bizon enet_dmac_writel(priv, 0, ENETDMAC_IRMASK_REG(priv->rx_chan)); 11710ae99b5fSMaxime Bizon enet_dmac_writel(priv, 0, ENETDMAC_IRMASK_REG(priv->tx_chan)); 1172adfc5217SJeff Kirsher 1173adfc5217SJeff Kirsher /* make sure no mib update is scheduled */ 1174adfc5217SJeff Kirsher cancel_work_sync(&priv->mib_update_task); 1175adfc5217SJeff Kirsher 1176adfc5217SJeff Kirsher /* disable dma & mac */ 1177adfc5217SJeff Kirsher bcm_enet_disable_dma(priv, priv->tx_chan); 1178adfc5217SJeff Kirsher bcm_enet_disable_dma(priv, priv->rx_chan); 1179adfc5217SJeff Kirsher bcm_enet_disable_mac(priv); 1180adfc5217SJeff Kirsher 1181adfc5217SJeff Kirsher /* force reclaim of all tx buffers */ 1182adfc5217SJeff Kirsher bcm_enet_tx_reclaim(dev, 1); 1183adfc5217SJeff Kirsher 1184adfc5217SJeff Kirsher /* free the rx skb ring */ 1185adfc5217SJeff Kirsher for (i = 0; i < priv->rx_ring_size; i++) { 1186adfc5217SJeff Kirsher struct bcm_enet_desc *desc; 1187adfc5217SJeff Kirsher 1188adfc5217SJeff Kirsher if (!priv->rx_skb[i]) 1189adfc5217SJeff Kirsher continue; 1190adfc5217SJeff Kirsher 1191adfc5217SJeff Kirsher desc = &priv->rx_desc_cpu[i]; 1192adfc5217SJeff Kirsher dma_unmap_single(kdev, desc->address, priv->rx_skb_size, 1193adfc5217SJeff Kirsher DMA_FROM_DEVICE); 1194adfc5217SJeff Kirsher kfree_skb(priv->rx_skb[i]); 1195adfc5217SJeff Kirsher } 1196adfc5217SJeff Kirsher 1197adfc5217SJeff Kirsher /* free remaining allocated memory */ 1198adfc5217SJeff Kirsher kfree(priv->rx_skb); 1199adfc5217SJeff Kirsher kfree(priv->tx_skb); 1200adfc5217SJeff Kirsher dma_free_coherent(kdev, priv->rx_desc_alloc_size, 1201adfc5217SJeff Kirsher priv->rx_desc_cpu, priv->rx_desc_dma); 1202adfc5217SJeff Kirsher dma_free_coherent(kdev, priv->tx_desc_alloc_size, 1203adfc5217SJeff Kirsher priv->tx_desc_cpu, priv->tx_desc_dma); 1204adfc5217SJeff Kirsher free_irq(priv->irq_tx, dev); 1205adfc5217SJeff Kirsher free_irq(priv->irq_rx, dev); 1206adfc5217SJeff Kirsher free_irq(dev->irq, dev); 1207adfc5217SJeff Kirsher 1208adfc5217SJeff Kirsher /* release phy */ 1209adfc5217SJeff Kirsher if (priv->has_phy) { 1210adfc5217SJeff Kirsher phy_disconnect(priv->phydev); 1211adfc5217SJeff Kirsher priv->phydev = NULL; 1212adfc5217SJeff Kirsher } 1213adfc5217SJeff Kirsher 1214adfc5217SJeff Kirsher return 0; 1215adfc5217SJeff Kirsher } 1216adfc5217SJeff Kirsher 1217adfc5217SJeff Kirsher /* 1218adfc5217SJeff Kirsher * ethtool callbacks 1219adfc5217SJeff Kirsher */ 1220adfc5217SJeff Kirsher struct bcm_enet_stats { 1221adfc5217SJeff Kirsher char stat_string[ETH_GSTRING_LEN]; 1222adfc5217SJeff Kirsher int sizeof_stat; 1223adfc5217SJeff Kirsher int stat_offset; 1224adfc5217SJeff Kirsher int mib_reg; 1225adfc5217SJeff Kirsher }; 1226adfc5217SJeff Kirsher 1227adfc5217SJeff Kirsher #define GEN_STAT(m) sizeof(((struct bcm_enet_priv *)0)->m), \ 1228adfc5217SJeff Kirsher offsetof(struct bcm_enet_priv, m) 1229adfc5217SJeff Kirsher #define DEV_STAT(m) sizeof(((struct net_device_stats *)0)->m), \ 1230adfc5217SJeff Kirsher offsetof(struct net_device_stats, m) 1231adfc5217SJeff Kirsher 1232adfc5217SJeff Kirsher static const struct bcm_enet_stats bcm_enet_gstrings_stats[] = { 1233adfc5217SJeff Kirsher { "rx_packets", DEV_STAT(rx_packets), -1 }, 1234adfc5217SJeff Kirsher { "tx_packets", DEV_STAT(tx_packets), -1 }, 1235adfc5217SJeff Kirsher { "rx_bytes", DEV_STAT(rx_bytes), -1 }, 1236adfc5217SJeff Kirsher { "tx_bytes", DEV_STAT(tx_bytes), -1 }, 1237adfc5217SJeff Kirsher { "rx_errors", DEV_STAT(rx_errors), -1 }, 1238adfc5217SJeff Kirsher { "tx_errors", DEV_STAT(tx_errors), -1 }, 1239adfc5217SJeff Kirsher { "rx_dropped", DEV_STAT(rx_dropped), -1 }, 1240adfc5217SJeff Kirsher { "tx_dropped", DEV_STAT(tx_dropped), -1 }, 1241adfc5217SJeff Kirsher 1242adfc5217SJeff Kirsher { "rx_good_octets", GEN_STAT(mib.rx_gd_octets), ETH_MIB_RX_GD_OCTETS}, 1243adfc5217SJeff Kirsher { "rx_good_pkts", GEN_STAT(mib.rx_gd_pkts), ETH_MIB_RX_GD_PKTS }, 1244adfc5217SJeff Kirsher { "rx_broadcast", GEN_STAT(mib.rx_brdcast), ETH_MIB_RX_BRDCAST }, 1245adfc5217SJeff Kirsher { "rx_multicast", GEN_STAT(mib.rx_mult), ETH_MIB_RX_MULT }, 1246adfc5217SJeff Kirsher { "rx_64_octets", GEN_STAT(mib.rx_64), ETH_MIB_RX_64 }, 1247adfc5217SJeff Kirsher { "rx_65_127_oct", GEN_STAT(mib.rx_65_127), ETH_MIB_RX_65_127 }, 1248adfc5217SJeff Kirsher { "rx_128_255_oct", GEN_STAT(mib.rx_128_255), ETH_MIB_RX_128_255 }, 1249adfc5217SJeff Kirsher { "rx_256_511_oct", GEN_STAT(mib.rx_256_511), ETH_MIB_RX_256_511 }, 1250adfc5217SJeff Kirsher { "rx_512_1023_oct", GEN_STAT(mib.rx_512_1023), ETH_MIB_RX_512_1023 }, 1251adfc5217SJeff Kirsher { "rx_1024_max_oct", GEN_STAT(mib.rx_1024_max), ETH_MIB_RX_1024_MAX }, 1252adfc5217SJeff Kirsher { "rx_jabber", GEN_STAT(mib.rx_jab), ETH_MIB_RX_JAB }, 1253adfc5217SJeff Kirsher { "rx_oversize", GEN_STAT(mib.rx_ovr), ETH_MIB_RX_OVR }, 1254adfc5217SJeff Kirsher { "rx_fragment", GEN_STAT(mib.rx_frag), ETH_MIB_RX_FRAG }, 1255adfc5217SJeff Kirsher { "rx_dropped", GEN_STAT(mib.rx_drop), ETH_MIB_RX_DROP }, 1256adfc5217SJeff Kirsher { "rx_crc_align", GEN_STAT(mib.rx_crc_align), ETH_MIB_RX_CRC_ALIGN }, 1257adfc5217SJeff Kirsher { "rx_undersize", GEN_STAT(mib.rx_und), ETH_MIB_RX_UND }, 1258adfc5217SJeff Kirsher { "rx_crc", GEN_STAT(mib.rx_crc), ETH_MIB_RX_CRC }, 1259adfc5217SJeff Kirsher { "rx_align", GEN_STAT(mib.rx_align), ETH_MIB_RX_ALIGN }, 1260adfc5217SJeff Kirsher { "rx_symbol_error", GEN_STAT(mib.rx_sym), ETH_MIB_RX_SYM }, 1261adfc5217SJeff Kirsher { "rx_pause", GEN_STAT(mib.rx_pause), ETH_MIB_RX_PAUSE }, 1262adfc5217SJeff Kirsher { "rx_control", GEN_STAT(mib.rx_cntrl), ETH_MIB_RX_CNTRL }, 1263adfc5217SJeff Kirsher 1264adfc5217SJeff Kirsher { "tx_good_octets", GEN_STAT(mib.tx_gd_octets), ETH_MIB_TX_GD_OCTETS }, 1265adfc5217SJeff Kirsher { "tx_good_pkts", GEN_STAT(mib.tx_gd_pkts), ETH_MIB_TX_GD_PKTS }, 1266adfc5217SJeff Kirsher { "tx_broadcast", GEN_STAT(mib.tx_brdcast), ETH_MIB_TX_BRDCAST }, 1267adfc5217SJeff Kirsher { "tx_multicast", GEN_STAT(mib.tx_mult), ETH_MIB_TX_MULT }, 1268adfc5217SJeff Kirsher { "tx_64_oct", GEN_STAT(mib.tx_64), ETH_MIB_TX_64 }, 1269adfc5217SJeff Kirsher { "tx_65_127_oct", GEN_STAT(mib.tx_65_127), ETH_MIB_TX_65_127 }, 1270adfc5217SJeff Kirsher { "tx_128_255_oct", GEN_STAT(mib.tx_128_255), ETH_MIB_TX_128_255 }, 1271adfc5217SJeff Kirsher { "tx_256_511_oct", GEN_STAT(mib.tx_256_511), ETH_MIB_TX_256_511 }, 1272adfc5217SJeff Kirsher { "tx_512_1023_oct", GEN_STAT(mib.tx_512_1023), ETH_MIB_TX_512_1023}, 1273adfc5217SJeff Kirsher { "tx_1024_max_oct", GEN_STAT(mib.tx_1024_max), ETH_MIB_TX_1024_MAX }, 1274adfc5217SJeff Kirsher { "tx_jabber", GEN_STAT(mib.tx_jab), ETH_MIB_TX_JAB }, 1275adfc5217SJeff Kirsher { "tx_oversize", GEN_STAT(mib.tx_ovr), ETH_MIB_TX_OVR }, 1276adfc5217SJeff Kirsher { "tx_fragment", GEN_STAT(mib.tx_frag), ETH_MIB_TX_FRAG }, 1277adfc5217SJeff Kirsher { "tx_underrun", GEN_STAT(mib.tx_underrun), ETH_MIB_TX_UNDERRUN }, 1278adfc5217SJeff Kirsher { "tx_collisions", GEN_STAT(mib.tx_col), ETH_MIB_TX_COL }, 1279adfc5217SJeff Kirsher { "tx_single_collision", GEN_STAT(mib.tx_1_col), ETH_MIB_TX_1_COL }, 1280adfc5217SJeff Kirsher { "tx_multiple_collision", GEN_STAT(mib.tx_m_col), ETH_MIB_TX_M_COL }, 1281adfc5217SJeff Kirsher { "tx_excess_collision", GEN_STAT(mib.tx_ex_col), ETH_MIB_TX_EX_COL }, 1282adfc5217SJeff Kirsher { "tx_late_collision", GEN_STAT(mib.tx_late), ETH_MIB_TX_LATE }, 1283adfc5217SJeff Kirsher { "tx_deferred", GEN_STAT(mib.tx_def), ETH_MIB_TX_DEF }, 1284adfc5217SJeff Kirsher { "tx_carrier_sense", GEN_STAT(mib.tx_crs), ETH_MIB_TX_CRS }, 1285adfc5217SJeff Kirsher { "tx_pause", GEN_STAT(mib.tx_pause), ETH_MIB_TX_PAUSE }, 1286adfc5217SJeff Kirsher 1287adfc5217SJeff Kirsher }; 1288adfc5217SJeff Kirsher 1289adfc5217SJeff Kirsher #define BCM_ENET_STATS_LEN \ 1290adfc5217SJeff Kirsher (sizeof(bcm_enet_gstrings_stats) / sizeof(struct bcm_enet_stats)) 1291adfc5217SJeff Kirsher 1292adfc5217SJeff Kirsher static const u32 unused_mib_regs[] = { 1293adfc5217SJeff Kirsher ETH_MIB_TX_ALL_OCTETS, 1294adfc5217SJeff Kirsher ETH_MIB_TX_ALL_PKTS, 1295adfc5217SJeff Kirsher ETH_MIB_RX_ALL_OCTETS, 1296adfc5217SJeff Kirsher ETH_MIB_RX_ALL_PKTS, 1297adfc5217SJeff Kirsher }; 1298adfc5217SJeff Kirsher 1299adfc5217SJeff Kirsher 1300adfc5217SJeff Kirsher static void bcm_enet_get_drvinfo(struct net_device *netdev, 1301adfc5217SJeff Kirsher struct ethtool_drvinfo *drvinfo) 1302adfc5217SJeff Kirsher { 13037826d43fSJiri Pirko strlcpy(drvinfo->driver, bcm_enet_driver_name, sizeof(drvinfo->driver)); 13047826d43fSJiri Pirko strlcpy(drvinfo->version, bcm_enet_driver_version, 13057826d43fSJiri Pirko sizeof(drvinfo->version)); 13067826d43fSJiri Pirko strlcpy(drvinfo->fw_version, "N/A", sizeof(drvinfo->fw_version)); 13077826d43fSJiri Pirko strlcpy(drvinfo->bus_info, "bcm63xx", sizeof(drvinfo->bus_info)); 1308adfc5217SJeff Kirsher drvinfo->n_stats = BCM_ENET_STATS_LEN; 1309adfc5217SJeff Kirsher } 1310adfc5217SJeff Kirsher 1311adfc5217SJeff Kirsher static int bcm_enet_get_sset_count(struct net_device *netdev, 1312adfc5217SJeff Kirsher int string_set) 1313adfc5217SJeff Kirsher { 1314adfc5217SJeff Kirsher switch (string_set) { 1315adfc5217SJeff Kirsher case ETH_SS_STATS: 1316adfc5217SJeff Kirsher return BCM_ENET_STATS_LEN; 1317adfc5217SJeff Kirsher default: 1318adfc5217SJeff Kirsher return -EINVAL; 1319adfc5217SJeff Kirsher } 1320adfc5217SJeff Kirsher } 1321adfc5217SJeff Kirsher 1322adfc5217SJeff Kirsher static void bcm_enet_get_strings(struct net_device *netdev, 1323adfc5217SJeff Kirsher u32 stringset, u8 *data) 1324adfc5217SJeff Kirsher { 1325adfc5217SJeff Kirsher int i; 1326adfc5217SJeff Kirsher 1327adfc5217SJeff Kirsher switch (stringset) { 1328adfc5217SJeff Kirsher case ETH_SS_STATS: 1329adfc5217SJeff Kirsher for (i = 0; i < BCM_ENET_STATS_LEN; i++) { 1330adfc5217SJeff Kirsher memcpy(data + i * ETH_GSTRING_LEN, 1331adfc5217SJeff Kirsher bcm_enet_gstrings_stats[i].stat_string, 1332adfc5217SJeff Kirsher ETH_GSTRING_LEN); 1333adfc5217SJeff Kirsher } 1334adfc5217SJeff Kirsher break; 1335adfc5217SJeff Kirsher } 1336adfc5217SJeff Kirsher } 1337adfc5217SJeff Kirsher 1338adfc5217SJeff Kirsher static void update_mib_counters(struct bcm_enet_priv *priv) 1339adfc5217SJeff Kirsher { 1340adfc5217SJeff Kirsher int i; 1341adfc5217SJeff Kirsher 1342adfc5217SJeff Kirsher for (i = 0; i < BCM_ENET_STATS_LEN; i++) { 1343adfc5217SJeff Kirsher const struct bcm_enet_stats *s; 1344adfc5217SJeff Kirsher u32 val; 1345adfc5217SJeff Kirsher char *p; 1346adfc5217SJeff Kirsher 1347adfc5217SJeff Kirsher s = &bcm_enet_gstrings_stats[i]; 1348adfc5217SJeff Kirsher if (s->mib_reg == -1) 1349adfc5217SJeff Kirsher continue; 1350adfc5217SJeff Kirsher 1351adfc5217SJeff Kirsher val = enet_readl(priv, ENET_MIB_REG(s->mib_reg)); 1352adfc5217SJeff Kirsher p = (char *)priv + s->stat_offset; 1353adfc5217SJeff Kirsher 1354adfc5217SJeff Kirsher if (s->sizeof_stat == sizeof(u64)) 1355adfc5217SJeff Kirsher *(u64 *)p += val; 1356adfc5217SJeff Kirsher else 1357adfc5217SJeff Kirsher *(u32 *)p += val; 1358adfc5217SJeff Kirsher } 1359adfc5217SJeff Kirsher 1360adfc5217SJeff Kirsher /* also empty unused mib counters to make sure mib counter 1361adfc5217SJeff Kirsher * overflow interrupt is cleared */ 1362adfc5217SJeff Kirsher for (i = 0; i < ARRAY_SIZE(unused_mib_regs); i++) 1363adfc5217SJeff Kirsher (void)enet_readl(priv, ENET_MIB_REG(unused_mib_regs[i])); 1364adfc5217SJeff Kirsher } 1365adfc5217SJeff Kirsher 1366adfc5217SJeff Kirsher static void bcm_enet_update_mib_counters_defer(struct work_struct *t) 1367adfc5217SJeff Kirsher { 1368adfc5217SJeff Kirsher struct bcm_enet_priv *priv; 1369adfc5217SJeff Kirsher 1370adfc5217SJeff Kirsher priv = container_of(t, struct bcm_enet_priv, mib_update_task); 1371adfc5217SJeff Kirsher mutex_lock(&priv->mib_update_lock); 1372adfc5217SJeff Kirsher update_mib_counters(priv); 1373adfc5217SJeff Kirsher mutex_unlock(&priv->mib_update_lock); 1374adfc5217SJeff Kirsher 1375adfc5217SJeff Kirsher /* reenable mib interrupt */ 1376adfc5217SJeff Kirsher if (netif_running(priv->net_dev)) 1377adfc5217SJeff Kirsher enet_writel(priv, ENET_IR_MIB, ENET_IRMASK_REG); 1378adfc5217SJeff Kirsher } 1379adfc5217SJeff Kirsher 1380adfc5217SJeff Kirsher static void bcm_enet_get_ethtool_stats(struct net_device *netdev, 1381adfc5217SJeff Kirsher struct ethtool_stats *stats, 1382adfc5217SJeff Kirsher u64 *data) 1383adfc5217SJeff Kirsher { 1384adfc5217SJeff Kirsher struct bcm_enet_priv *priv; 1385adfc5217SJeff Kirsher int i; 1386adfc5217SJeff Kirsher 1387adfc5217SJeff Kirsher priv = netdev_priv(netdev); 1388adfc5217SJeff Kirsher 1389adfc5217SJeff Kirsher mutex_lock(&priv->mib_update_lock); 1390adfc5217SJeff Kirsher update_mib_counters(priv); 1391adfc5217SJeff Kirsher 1392adfc5217SJeff Kirsher for (i = 0; i < BCM_ENET_STATS_LEN; i++) { 1393adfc5217SJeff Kirsher const struct bcm_enet_stats *s; 1394adfc5217SJeff Kirsher char *p; 1395adfc5217SJeff Kirsher 1396adfc5217SJeff Kirsher s = &bcm_enet_gstrings_stats[i]; 1397adfc5217SJeff Kirsher if (s->mib_reg == -1) 1398adfc5217SJeff Kirsher p = (char *)&netdev->stats; 1399adfc5217SJeff Kirsher else 1400adfc5217SJeff Kirsher p = (char *)priv; 1401adfc5217SJeff Kirsher p += s->stat_offset; 1402adfc5217SJeff Kirsher data[i] = (s->sizeof_stat == sizeof(u64)) ? 1403adfc5217SJeff Kirsher *(u64 *)p : *(u32 *)p; 1404adfc5217SJeff Kirsher } 1405adfc5217SJeff Kirsher mutex_unlock(&priv->mib_update_lock); 1406adfc5217SJeff Kirsher } 1407adfc5217SJeff Kirsher 14087260aac9SMaxime Bizon static int bcm_enet_nway_reset(struct net_device *dev) 14097260aac9SMaxime Bizon { 14107260aac9SMaxime Bizon struct bcm_enet_priv *priv; 14117260aac9SMaxime Bizon 14127260aac9SMaxime Bizon priv = netdev_priv(dev); 14137260aac9SMaxime Bizon if (priv->has_phy) { 14147260aac9SMaxime Bizon if (!priv->phydev) 14157260aac9SMaxime Bizon return -ENODEV; 14167260aac9SMaxime Bizon return genphy_restart_aneg(priv->phydev); 14177260aac9SMaxime Bizon } 14187260aac9SMaxime Bizon 14197260aac9SMaxime Bizon return -EOPNOTSUPP; 14207260aac9SMaxime Bizon } 14217260aac9SMaxime Bizon 1422adfc5217SJeff Kirsher static int bcm_enet_get_settings(struct net_device *dev, 1423adfc5217SJeff Kirsher struct ethtool_cmd *cmd) 1424adfc5217SJeff Kirsher { 1425adfc5217SJeff Kirsher struct bcm_enet_priv *priv; 1426adfc5217SJeff Kirsher 1427adfc5217SJeff Kirsher priv = netdev_priv(dev); 1428adfc5217SJeff Kirsher 1429adfc5217SJeff Kirsher cmd->maxrxpkt = 0; 1430adfc5217SJeff Kirsher cmd->maxtxpkt = 0; 1431adfc5217SJeff Kirsher 1432adfc5217SJeff Kirsher if (priv->has_phy) { 1433adfc5217SJeff Kirsher if (!priv->phydev) 1434adfc5217SJeff Kirsher return -ENODEV; 1435adfc5217SJeff Kirsher return phy_ethtool_gset(priv->phydev, cmd); 1436adfc5217SJeff Kirsher } else { 1437adfc5217SJeff Kirsher cmd->autoneg = 0; 1438adfc5217SJeff Kirsher ethtool_cmd_speed_set(cmd, ((priv->force_speed_100) 1439adfc5217SJeff Kirsher ? SPEED_100 : SPEED_10)); 1440adfc5217SJeff Kirsher cmd->duplex = (priv->force_duplex_full) ? 1441adfc5217SJeff Kirsher DUPLEX_FULL : DUPLEX_HALF; 1442adfc5217SJeff Kirsher cmd->supported = ADVERTISED_10baseT_Half | 1443adfc5217SJeff Kirsher ADVERTISED_10baseT_Full | 1444adfc5217SJeff Kirsher ADVERTISED_100baseT_Half | 1445adfc5217SJeff Kirsher ADVERTISED_100baseT_Full; 1446adfc5217SJeff Kirsher cmd->advertising = 0; 1447adfc5217SJeff Kirsher cmd->port = PORT_MII; 1448adfc5217SJeff Kirsher cmd->transceiver = XCVR_EXTERNAL; 1449adfc5217SJeff Kirsher } 1450adfc5217SJeff Kirsher return 0; 1451adfc5217SJeff Kirsher } 1452adfc5217SJeff Kirsher 1453adfc5217SJeff Kirsher static int bcm_enet_set_settings(struct net_device *dev, 1454adfc5217SJeff Kirsher struct ethtool_cmd *cmd) 1455adfc5217SJeff Kirsher { 1456adfc5217SJeff Kirsher struct bcm_enet_priv *priv; 1457adfc5217SJeff Kirsher 1458adfc5217SJeff Kirsher priv = netdev_priv(dev); 1459adfc5217SJeff Kirsher if (priv->has_phy) { 1460adfc5217SJeff Kirsher if (!priv->phydev) 1461adfc5217SJeff Kirsher return -ENODEV; 1462adfc5217SJeff Kirsher return phy_ethtool_sset(priv->phydev, cmd); 1463adfc5217SJeff Kirsher } else { 1464adfc5217SJeff Kirsher 1465adfc5217SJeff Kirsher if (cmd->autoneg || 1466adfc5217SJeff Kirsher (cmd->speed != SPEED_100 && cmd->speed != SPEED_10) || 1467adfc5217SJeff Kirsher cmd->port != PORT_MII) 1468adfc5217SJeff Kirsher return -EINVAL; 1469adfc5217SJeff Kirsher 1470adfc5217SJeff Kirsher priv->force_speed_100 = (cmd->speed == SPEED_100) ? 1 : 0; 1471adfc5217SJeff Kirsher priv->force_duplex_full = (cmd->duplex == DUPLEX_FULL) ? 1 : 0; 1472adfc5217SJeff Kirsher 1473adfc5217SJeff Kirsher if (netif_running(dev)) 1474adfc5217SJeff Kirsher bcm_enet_adjust_link(dev); 1475adfc5217SJeff Kirsher return 0; 1476adfc5217SJeff Kirsher } 1477adfc5217SJeff Kirsher } 1478adfc5217SJeff Kirsher 1479adfc5217SJeff Kirsher static void bcm_enet_get_ringparam(struct net_device *dev, 1480adfc5217SJeff Kirsher struct ethtool_ringparam *ering) 1481adfc5217SJeff Kirsher { 1482adfc5217SJeff Kirsher struct bcm_enet_priv *priv; 1483adfc5217SJeff Kirsher 1484adfc5217SJeff Kirsher priv = netdev_priv(dev); 1485adfc5217SJeff Kirsher 1486adfc5217SJeff Kirsher /* rx/tx ring is actually only limited by memory */ 1487adfc5217SJeff Kirsher ering->rx_max_pending = 8192; 1488adfc5217SJeff Kirsher ering->tx_max_pending = 8192; 1489adfc5217SJeff Kirsher ering->rx_pending = priv->rx_ring_size; 1490adfc5217SJeff Kirsher ering->tx_pending = priv->tx_ring_size; 1491adfc5217SJeff Kirsher } 1492adfc5217SJeff Kirsher 1493adfc5217SJeff Kirsher static int bcm_enet_set_ringparam(struct net_device *dev, 1494adfc5217SJeff Kirsher struct ethtool_ringparam *ering) 1495adfc5217SJeff Kirsher { 1496adfc5217SJeff Kirsher struct bcm_enet_priv *priv; 1497adfc5217SJeff Kirsher int was_running; 1498adfc5217SJeff Kirsher 1499adfc5217SJeff Kirsher priv = netdev_priv(dev); 1500adfc5217SJeff Kirsher 1501adfc5217SJeff Kirsher was_running = 0; 1502adfc5217SJeff Kirsher if (netif_running(dev)) { 1503adfc5217SJeff Kirsher bcm_enet_stop(dev); 1504adfc5217SJeff Kirsher was_running = 1; 1505adfc5217SJeff Kirsher } 1506adfc5217SJeff Kirsher 1507adfc5217SJeff Kirsher priv->rx_ring_size = ering->rx_pending; 1508adfc5217SJeff Kirsher priv->tx_ring_size = ering->tx_pending; 1509adfc5217SJeff Kirsher 1510adfc5217SJeff Kirsher if (was_running) { 1511adfc5217SJeff Kirsher int err; 1512adfc5217SJeff Kirsher 1513adfc5217SJeff Kirsher err = bcm_enet_open(dev); 1514adfc5217SJeff Kirsher if (err) 1515adfc5217SJeff Kirsher dev_close(dev); 1516adfc5217SJeff Kirsher else 1517adfc5217SJeff Kirsher bcm_enet_set_multicast_list(dev); 1518adfc5217SJeff Kirsher } 1519adfc5217SJeff Kirsher return 0; 1520adfc5217SJeff Kirsher } 1521adfc5217SJeff Kirsher 1522adfc5217SJeff Kirsher static void bcm_enet_get_pauseparam(struct net_device *dev, 1523adfc5217SJeff Kirsher struct ethtool_pauseparam *ecmd) 1524adfc5217SJeff Kirsher { 1525adfc5217SJeff Kirsher struct bcm_enet_priv *priv; 1526adfc5217SJeff Kirsher 1527adfc5217SJeff Kirsher priv = netdev_priv(dev); 1528adfc5217SJeff Kirsher ecmd->autoneg = priv->pause_auto; 1529adfc5217SJeff Kirsher ecmd->rx_pause = priv->pause_rx; 1530adfc5217SJeff Kirsher ecmd->tx_pause = priv->pause_tx; 1531adfc5217SJeff Kirsher } 1532adfc5217SJeff Kirsher 1533adfc5217SJeff Kirsher static int bcm_enet_set_pauseparam(struct net_device *dev, 1534adfc5217SJeff Kirsher struct ethtool_pauseparam *ecmd) 1535adfc5217SJeff Kirsher { 1536adfc5217SJeff Kirsher struct bcm_enet_priv *priv; 1537adfc5217SJeff Kirsher 1538adfc5217SJeff Kirsher priv = netdev_priv(dev); 1539adfc5217SJeff Kirsher 1540adfc5217SJeff Kirsher if (priv->has_phy) { 1541adfc5217SJeff Kirsher if (ecmd->autoneg && (ecmd->rx_pause != ecmd->tx_pause)) { 1542adfc5217SJeff Kirsher /* asymetric pause mode not supported, 1543adfc5217SJeff Kirsher * actually possible but integrated PHY has RO 1544adfc5217SJeff Kirsher * asym_pause bit */ 1545adfc5217SJeff Kirsher return -EINVAL; 1546adfc5217SJeff Kirsher } 1547adfc5217SJeff Kirsher } else { 1548adfc5217SJeff Kirsher /* no pause autoneg on direct mii connection */ 1549adfc5217SJeff Kirsher if (ecmd->autoneg) 1550adfc5217SJeff Kirsher return -EINVAL; 1551adfc5217SJeff Kirsher } 1552adfc5217SJeff Kirsher 1553adfc5217SJeff Kirsher priv->pause_auto = ecmd->autoneg; 1554adfc5217SJeff Kirsher priv->pause_rx = ecmd->rx_pause; 1555adfc5217SJeff Kirsher priv->pause_tx = ecmd->tx_pause; 1556adfc5217SJeff Kirsher 1557adfc5217SJeff Kirsher return 0; 1558adfc5217SJeff Kirsher } 1559adfc5217SJeff Kirsher 15601aff0cbeSstephen hemminger static const struct ethtool_ops bcm_enet_ethtool_ops = { 1561adfc5217SJeff Kirsher .get_strings = bcm_enet_get_strings, 1562adfc5217SJeff Kirsher .get_sset_count = bcm_enet_get_sset_count, 1563adfc5217SJeff Kirsher .get_ethtool_stats = bcm_enet_get_ethtool_stats, 15647260aac9SMaxime Bizon .nway_reset = bcm_enet_nway_reset, 1565adfc5217SJeff Kirsher .get_settings = bcm_enet_get_settings, 1566adfc5217SJeff Kirsher .set_settings = bcm_enet_set_settings, 1567adfc5217SJeff Kirsher .get_drvinfo = bcm_enet_get_drvinfo, 1568adfc5217SJeff Kirsher .get_link = ethtool_op_get_link, 1569adfc5217SJeff Kirsher .get_ringparam = bcm_enet_get_ringparam, 1570adfc5217SJeff Kirsher .set_ringparam = bcm_enet_set_ringparam, 1571adfc5217SJeff Kirsher .get_pauseparam = bcm_enet_get_pauseparam, 1572adfc5217SJeff Kirsher .set_pauseparam = bcm_enet_set_pauseparam, 1573adfc5217SJeff Kirsher }; 1574adfc5217SJeff Kirsher 1575adfc5217SJeff Kirsher static int bcm_enet_ioctl(struct net_device *dev, struct ifreq *rq, int cmd) 1576adfc5217SJeff Kirsher { 1577adfc5217SJeff Kirsher struct bcm_enet_priv *priv; 1578adfc5217SJeff Kirsher 1579adfc5217SJeff Kirsher priv = netdev_priv(dev); 1580adfc5217SJeff Kirsher if (priv->has_phy) { 1581adfc5217SJeff Kirsher if (!priv->phydev) 1582adfc5217SJeff Kirsher return -ENODEV; 1583adfc5217SJeff Kirsher return phy_mii_ioctl(priv->phydev, rq, cmd); 1584adfc5217SJeff Kirsher } else { 1585adfc5217SJeff Kirsher struct mii_if_info mii; 1586adfc5217SJeff Kirsher 1587adfc5217SJeff Kirsher mii.dev = dev; 1588adfc5217SJeff Kirsher mii.mdio_read = bcm_enet_mdio_read_mii; 1589adfc5217SJeff Kirsher mii.mdio_write = bcm_enet_mdio_write_mii; 1590adfc5217SJeff Kirsher mii.phy_id = 0; 1591adfc5217SJeff Kirsher mii.phy_id_mask = 0x3f; 1592adfc5217SJeff Kirsher mii.reg_num_mask = 0x1f; 1593adfc5217SJeff Kirsher return generic_mii_ioctl(&mii, if_mii(rq), cmd, NULL); 1594adfc5217SJeff Kirsher } 1595adfc5217SJeff Kirsher } 1596adfc5217SJeff Kirsher 1597adfc5217SJeff Kirsher /* 1598adfc5217SJeff Kirsher * calculate actual hardware mtu 1599adfc5217SJeff Kirsher */ 1600adfc5217SJeff Kirsher static int compute_hw_mtu(struct bcm_enet_priv *priv, int mtu) 1601adfc5217SJeff Kirsher { 1602adfc5217SJeff Kirsher int actual_mtu; 1603adfc5217SJeff Kirsher 1604adfc5217SJeff Kirsher actual_mtu = mtu; 1605adfc5217SJeff Kirsher 1606adfc5217SJeff Kirsher /* add ethernet header + vlan tag size */ 1607adfc5217SJeff Kirsher actual_mtu += VLAN_ETH_HLEN; 1608adfc5217SJeff Kirsher 1609adfc5217SJeff Kirsher if (actual_mtu < 64 || actual_mtu > BCMENET_MAX_MTU) 1610adfc5217SJeff Kirsher return -EINVAL; 1611adfc5217SJeff Kirsher 1612adfc5217SJeff Kirsher /* 1613adfc5217SJeff Kirsher * setup maximum size before we get overflow mark in 1614adfc5217SJeff Kirsher * descriptor, note that this will not prevent reception of 1615adfc5217SJeff Kirsher * big frames, they will be split into multiple buffers 1616adfc5217SJeff Kirsher * anyway 1617adfc5217SJeff Kirsher */ 1618adfc5217SJeff Kirsher priv->hw_mtu = actual_mtu; 1619adfc5217SJeff Kirsher 1620adfc5217SJeff Kirsher /* 1621adfc5217SJeff Kirsher * align rx buffer size to dma burst len, account FCS since 1622adfc5217SJeff Kirsher * it's appended 1623adfc5217SJeff Kirsher */ 1624adfc5217SJeff Kirsher priv->rx_skb_size = ALIGN(actual_mtu + ETH_FCS_LEN, 16256f00a022SMaxime Bizon priv->dma_maxburst * 4); 1626adfc5217SJeff Kirsher return 0; 1627adfc5217SJeff Kirsher } 1628adfc5217SJeff Kirsher 1629adfc5217SJeff Kirsher /* 1630adfc5217SJeff Kirsher * adjust mtu, can't be called while device is running 1631adfc5217SJeff Kirsher */ 1632adfc5217SJeff Kirsher static int bcm_enet_change_mtu(struct net_device *dev, int new_mtu) 1633adfc5217SJeff Kirsher { 1634adfc5217SJeff Kirsher int ret; 1635adfc5217SJeff Kirsher 1636adfc5217SJeff Kirsher if (netif_running(dev)) 1637adfc5217SJeff Kirsher return -EBUSY; 1638adfc5217SJeff Kirsher 1639adfc5217SJeff Kirsher ret = compute_hw_mtu(netdev_priv(dev), new_mtu); 1640adfc5217SJeff Kirsher if (ret) 1641adfc5217SJeff Kirsher return ret; 1642adfc5217SJeff Kirsher dev->mtu = new_mtu; 1643adfc5217SJeff Kirsher return 0; 1644adfc5217SJeff Kirsher } 1645adfc5217SJeff Kirsher 1646adfc5217SJeff Kirsher /* 1647adfc5217SJeff Kirsher * preinit hardware to allow mii operation while device is down 1648adfc5217SJeff Kirsher */ 1649adfc5217SJeff Kirsher static void bcm_enet_hw_preinit(struct bcm_enet_priv *priv) 1650adfc5217SJeff Kirsher { 1651adfc5217SJeff Kirsher u32 val; 1652adfc5217SJeff Kirsher int limit; 1653adfc5217SJeff Kirsher 1654adfc5217SJeff Kirsher /* make sure mac is disabled */ 1655adfc5217SJeff Kirsher bcm_enet_disable_mac(priv); 1656adfc5217SJeff Kirsher 1657adfc5217SJeff Kirsher /* soft reset mac */ 1658adfc5217SJeff Kirsher val = ENET_CTL_SRESET_MASK; 1659adfc5217SJeff Kirsher enet_writel(priv, val, ENET_CTL_REG); 1660adfc5217SJeff Kirsher wmb(); 1661adfc5217SJeff Kirsher 1662adfc5217SJeff Kirsher limit = 1000; 1663adfc5217SJeff Kirsher do { 1664adfc5217SJeff Kirsher val = enet_readl(priv, ENET_CTL_REG); 1665adfc5217SJeff Kirsher if (!(val & ENET_CTL_SRESET_MASK)) 1666adfc5217SJeff Kirsher break; 1667adfc5217SJeff Kirsher udelay(1); 1668adfc5217SJeff Kirsher } while (limit--); 1669adfc5217SJeff Kirsher 1670adfc5217SJeff Kirsher /* select correct mii interface */ 1671adfc5217SJeff Kirsher val = enet_readl(priv, ENET_CTL_REG); 1672adfc5217SJeff Kirsher if (priv->use_external_mii) 1673adfc5217SJeff Kirsher val |= ENET_CTL_EPHYSEL_MASK; 1674adfc5217SJeff Kirsher else 1675adfc5217SJeff Kirsher val &= ~ENET_CTL_EPHYSEL_MASK; 1676adfc5217SJeff Kirsher enet_writel(priv, val, ENET_CTL_REG); 1677adfc5217SJeff Kirsher 1678adfc5217SJeff Kirsher /* turn on mdc clock */ 1679adfc5217SJeff Kirsher enet_writel(priv, (0x1f << ENET_MIISC_MDCFREQDIV_SHIFT) | 1680adfc5217SJeff Kirsher ENET_MIISC_PREAMBLEEN_MASK, ENET_MIISC_REG); 1681adfc5217SJeff Kirsher 1682adfc5217SJeff Kirsher /* set mib counters to self-clear when read */ 1683adfc5217SJeff Kirsher val = enet_readl(priv, ENET_MIBCTL_REG); 1684adfc5217SJeff Kirsher val |= ENET_MIBCTL_RDCLEAR_MASK; 1685adfc5217SJeff Kirsher enet_writel(priv, val, ENET_MIBCTL_REG); 1686adfc5217SJeff Kirsher } 1687adfc5217SJeff Kirsher 1688adfc5217SJeff Kirsher static const struct net_device_ops bcm_enet_ops = { 1689adfc5217SJeff Kirsher .ndo_open = bcm_enet_open, 1690adfc5217SJeff Kirsher .ndo_stop = bcm_enet_stop, 1691adfc5217SJeff Kirsher .ndo_start_xmit = bcm_enet_start_xmit, 1692adfc5217SJeff Kirsher .ndo_set_mac_address = bcm_enet_set_mac_address, 1693afc4b13dSJiri Pirko .ndo_set_rx_mode = bcm_enet_set_multicast_list, 1694adfc5217SJeff Kirsher .ndo_do_ioctl = bcm_enet_ioctl, 1695adfc5217SJeff Kirsher .ndo_change_mtu = bcm_enet_change_mtu, 1696adfc5217SJeff Kirsher #ifdef CONFIG_NET_POLL_CONTROLLER 1697adfc5217SJeff Kirsher .ndo_poll_controller = bcm_enet_netpoll, 1698adfc5217SJeff Kirsher #endif 1699adfc5217SJeff Kirsher }; 1700adfc5217SJeff Kirsher 1701adfc5217SJeff Kirsher /* 1702adfc5217SJeff Kirsher * allocate netdevice, request register memory and register device. 1703adfc5217SJeff Kirsher */ 1704047fc566SBill Pemberton static int bcm_enet_probe(struct platform_device *pdev) 1705adfc5217SJeff Kirsher { 1706adfc5217SJeff Kirsher struct bcm_enet_priv *priv; 1707adfc5217SJeff Kirsher struct net_device *dev; 1708adfc5217SJeff Kirsher struct bcm63xx_enet_platform_data *pd; 1709adfc5217SJeff Kirsher struct resource *res_mem, *res_irq, *res_irq_rx, *res_irq_tx; 1710adfc5217SJeff Kirsher struct mii_bus *bus; 1711adfc5217SJeff Kirsher const char *clk_name; 1712adfc5217SJeff Kirsher int i, ret; 1713adfc5217SJeff Kirsher 1714adfc5217SJeff Kirsher /* stop if shared driver failed, assume driver->probe will be 1715adfc5217SJeff Kirsher * called in the same order we register devices (correct ?) */ 17160ae99b5fSMaxime Bizon if (!bcm_enet_shared_base[0]) 1717adfc5217SJeff Kirsher return -ENODEV; 1718adfc5217SJeff Kirsher 1719adfc5217SJeff Kirsher res_mem = platform_get_resource(pdev, IORESOURCE_MEM, 0); 1720adfc5217SJeff Kirsher res_irq = platform_get_resource(pdev, IORESOURCE_IRQ, 0); 1721adfc5217SJeff Kirsher res_irq_rx = platform_get_resource(pdev, IORESOURCE_IRQ, 1); 1722adfc5217SJeff Kirsher res_irq_tx = platform_get_resource(pdev, IORESOURCE_IRQ, 2); 1723adfc5217SJeff Kirsher if (!res_mem || !res_irq || !res_irq_rx || !res_irq_tx) 1724adfc5217SJeff Kirsher return -ENODEV; 1725adfc5217SJeff Kirsher 1726adfc5217SJeff Kirsher ret = 0; 1727adfc5217SJeff Kirsher dev = alloc_etherdev(sizeof(*priv)); 1728adfc5217SJeff Kirsher if (!dev) 1729adfc5217SJeff Kirsher return -ENOMEM; 1730adfc5217SJeff Kirsher priv = netdev_priv(dev); 1731adfc5217SJeff Kirsher 17326f00a022SMaxime Bizon priv->enet_is_sw = false; 17336f00a022SMaxime Bizon priv->dma_maxburst = BCMENET_DMA_MAXBURST; 17346f00a022SMaxime Bizon 1735adfc5217SJeff Kirsher ret = compute_hw_mtu(priv, dev->mtu); 1736adfc5217SJeff Kirsher if (ret) 1737adfc5217SJeff Kirsher goto out; 1738adfc5217SJeff Kirsher 17391c03da05SJonas Gorski priv->base = devm_request_and_ioremap(&pdev->dev, res_mem); 17401c03da05SJonas Gorski if (priv->base == NULL) { 17411c03da05SJonas Gorski ret = -ENOMEM; 1742adfc5217SJeff Kirsher goto out; 1743adfc5217SJeff Kirsher } 1744adfc5217SJeff Kirsher 1745adfc5217SJeff Kirsher dev->irq = priv->irq = res_irq->start; 1746adfc5217SJeff Kirsher priv->irq_rx = res_irq_rx->start; 1747adfc5217SJeff Kirsher priv->irq_tx = res_irq_tx->start; 1748adfc5217SJeff Kirsher priv->mac_id = pdev->id; 1749adfc5217SJeff Kirsher 1750adfc5217SJeff Kirsher /* get rx & tx dma channel id for this mac */ 1751adfc5217SJeff Kirsher if (priv->mac_id == 0) { 1752adfc5217SJeff Kirsher priv->rx_chan = 0; 1753adfc5217SJeff Kirsher priv->tx_chan = 1; 1754adfc5217SJeff Kirsher clk_name = "enet0"; 1755adfc5217SJeff Kirsher } else { 1756adfc5217SJeff Kirsher priv->rx_chan = 2; 1757adfc5217SJeff Kirsher priv->tx_chan = 3; 1758adfc5217SJeff Kirsher clk_name = "enet1"; 1759adfc5217SJeff Kirsher } 1760adfc5217SJeff Kirsher 1761adfc5217SJeff Kirsher priv->mac_clk = clk_get(&pdev->dev, clk_name); 1762adfc5217SJeff Kirsher if (IS_ERR(priv->mac_clk)) { 1763adfc5217SJeff Kirsher ret = PTR_ERR(priv->mac_clk); 17641c03da05SJonas Gorski goto out; 1765adfc5217SJeff Kirsher } 1766624e2d21SJonas Gorski clk_prepare_enable(priv->mac_clk); 1767adfc5217SJeff Kirsher 1768adfc5217SJeff Kirsher /* initialize default and fetch platform data */ 1769adfc5217SJeff Kirsher priv->rx_ring_size = BCMENET_DEF_RX_DESC; 1770adfc5217SJeff Kirsher priv->tx_ring_size = BCMENET_DEF_TX_DESC; 1771adfc5217SJeff Kirsher 1772adfc5217SJeff Kirsher pd = pdev->dev.platform_data; 1773adfc5217SJeff Kirsher if (pd) { 1774adfc5217SJeff Kirsher memcpy(dev->dev_addr, pd->mac_addr, ETH_ALEN); 1775adfc5217SJeff Kirsher priv->has_phy = pd->has_phy; 1776adfc5217SJeff Kirsher priv->phy_id = pd->phy_id; 1777adfc5217SJeff Kirsher priv->has_phy_interrupt = pd->has_phy_interrupt; 1778adfc5217SJeff Kirsher priv->phy_interrupt = pd->phy_interrupt; 1779adfc5217SJeff Kirsher priv->use_external_mii = !pd->use_internal_phy; 1780adfc5217SJeff Kirsher priv->pause_auto = pd->pause_auto; 1781adfc5217SJeff Kirsher priv->pause_rx = pd->pause_rx; 1782adfc5217SJeff Kirsher priv->pause_tx = pd->pause_tx; 1783adfc5217SJeff Kirsher priv->force_duplex_full = pd->force_duplex_full; 1784adfc5217SJeff Kirsher priv->force_speed_100 = pd->force_speed_100; 1785adfc5217SJeff Kirsher } 1786adfc5217SJeff Kirsher 1787adfc5217SJeff Kirsher if (priv->mac_id == 0 && priv->has_phy && !priv->use_external_mii) { 1788adfc5217SJeff Kirsher /* using internal PHY, enable clock */ 1789adfc5217SJeff Kirsher priv->phy_clk = clk_get(&pdev->dev, "ephy"); 1790adfc5217SJeff Kirsher if (IS_ERR(priv->phy_clk)) { 1791adfc5217SJeff Kirsher ret = PTR_ERR(priv->phy_clk); 1792adfc5217SJeff Kirsher priv->phy_clk = NULL; 1793adfc5217SJeff Kirsher goto out_put_clk_mac; 1794adfc5217SJeff Kirsher } 1795624e2d21SJonas Gorski clk_prepare_enable(priv->phy_clk); 1796adfc5217SJeff Kirsher } 1797adfc5217SJeff Kirsher 1798adfc5217SJeff Kirsher /* do minimal hardware init to be able to probe mii bus */ 1799adfc5217SJeff Kirsher bcm_enet_hw_preinit(priv); 1800adfc5217SJeff Kirsher 1801adfc5217SJeff Kirsher /* MII bus registration */ 1802adfc5217SJeff Kirsher if (priv->has_phy) { 1803adfc5217SJeff Kirsher 1804adfc5217SJeff Kirsher priv->mii_bus = mdiobus_alloc(); 1805adfc5217SJeff Kirsher if (!priv->mii_bus) { 1806adfc5217SJeff Kirsher ret = -ENOMEM; 1807adfc5217SJeff Kirsher goto out_uninit_hw; 1808adfc5217SJeff Kirsher } 1809adfc5217SJeff Kirsher 1810adfc5217SJeff Kirsher bus = priv->mii_bus; 1811adfc5217SJeff Kirsher bus->name = "bcm63xx_enet MII bus"; 1812adfc5217SJeff Kirsher bus->parent = &pdev->dev; 1813adfc5217SJeff Kirsher bus->priv = priv; 1814adfc5217SJeff Kirsher bus->read = bcm_enet_mdio_read_phylib; 1815adfc5217SJeff Kirsher bus->write = bcm_enet_mdio_write_phylib; 18163e617506SFlorian Fainelli sprintf(bus->id, "%s-%d", pdev->name, priv->mac_id); 1817adfc5217SJeff Kirsher 1818adfc5217SJeff Kirsher /* only probe bus where we think the PHY is, because 1819adfc5217SJeff Kirsher * the mdio read operation return 0 instead of 0xffff 1820adfc5217SJeff Kirsher * if a slave is not present on hw */ 1821adfc5217SJeff Kirsher bus->phy_mask = ~(1 << priv->phy_id); 1822adfc5217SJeff Kirsher 18232a80b5e1SJonas Gorski bus->irq = devm_kzalloc(&pdev->dev, sizeof(int) * PHY_MAX_ADDR, 18242a80b5e1SJonas Gorski GFP_KERNEL); 1825adfc5217SJeff Kirsher if (!bus->irq) { 1826adfc5217SJeff Kirsher ret = -ENOMEM; 1827adfc5217SJeff Kirsher goto out_free_mdio; 1828adfc5217SJeff Kirsher } 1829adfc5217SJeff Kirsher 1830adfc5217SJeff Kirsher if (priv->has_phy_interrupt) 1831adfc5217SJeff Kirsher bus->irq[priv->phy_id] = priv->phy_interrupt; 1832adfc5217SJeff Kirsher else 1833adfc5217SJeff Kirsher bus->irq[priv->phy_id] = PHY_POLL; 1834adfc5217SJeff Kirsher 1835adfc5217SJeff Kirsher ret = mdiobus_register(bus); 1836adfc5217SJeff Kirsher if (ret) { 1837adfc5217SJeff Kirsher dev_err(&pdev->dev, "unable to register mdio bus\n"); 1838adfc5217SJeff Kirsher goto out_free_mdio; 1839adfc5217SJeff Kirsher } 1840adfc5217SJeff Kirsher } else { 1841adfc5217SJeff Kirsher 1842adfc5217SJeff Kirsher /* run platform code to initialize PHY device */ 1843adfc5217SJeff Kirsher if (pd->mii_config && 1844adfc5217SJeff Kirsher pd->mii_config(dev, 1, bcm_enet_mdio_read_mii, 1845adfc5217SJeff Kirsher bcm_enet_mdio_write_mii)) { 1846adfc5217SJeff Kirsher dev_err(&pdev->dev, "unable to configure mdio bus\n"); 1847adfc5217SJeff Kirsher goto out_uninit_hw; 1848adfc5217SJeff Kirsher } 1849adfc5217SJeff Kirsher } 1850adfc5217SJeff Kirsher 1851adfc5217SJeff Kirsher spin_lock_init(&priv->rx_lock); 1852adfc5217SJeff Kirsher 1853adfc5217SJeff Kirsher /* init rx timeout (used for oom) */ 1854adfc5217SJeff Kirsher init_timer(&priv->rx_timeout); 1855adfc5217SJeff Kirsher priv->rx_timeout.function = bcm_enet_refill_rx_timer; 1856adfc5217SJeff Kirsher priv->rx_timeout.data = (unsigned long)dev; 1857adfc5217SJeff Kirsher 1858adfc5217SJeff Kirsher /* init the mib update lock&work */ 1859adfc5217SJeff Kirsher mutex_init(&priv->mib_update_lock); 1860adfc5217SJeff Kirsher INIT_WORK(&priv->mib_update_task, bcm_enet_update_mib_counters_defer); 1861adfc5217SJeff Kirsher 1862adfc5217SJeff Kirsher /* zero mib counters */ 1863adfc5217SJeff Kirsher for (i = 0; i < ENET_MIB_REG_COUNT; i++) 1864adfc5217SJeff Kirsher enet_writel(priv, 0, ENET_MIB_REG(i)); 1865adfc5217SJeff Kirsher 1866adfc5217SJeff Kirsher /* register netdevice */ 1867adfc5217SJeff Kirsher dev->netdev_ops = &bcm_enet_ops; 1868adfc5217SJeff Kirsher netif_napi_add(dev, &priv->napi, bcm_enet_poll, 16); 1869adfc5217SJeff Kirsher 1870adfc5217SJeff Kirsher SET_ETHTOOL_OPS(dev, &bcm_enet_ethtool_ops); 1871adfc5217SJeff Kirsher SET_NETDEV_DEV(dev, &pdev->dev); 1872adfc5217SJeff Kirsher 1873adfc5217SJeff Kirsher ret = register_netdev(dev); 1874adfc5217SJeff Kirsher if (ret) 1875adfc5217SJeff Kirsher goto out_unregister_mdio; 1876adfc5217SJeff Kirsher 1877adfc5217SJeff Kirsher netif_carrier_off(dev); 1878adfc5217SJeff Kirsher platform_set_drvdata(pdev, dev); 1879adfc5217SJeff Kirsher priv->pdev = pdev; 1880adfc5217SJeff Kirsher priv->net_dev = dev; 1881adfc5217SJeff Kirsher 1882adfc5217SJeff Kirsher return 0; 1883adfc5217SJeff Kirsher 1884adfc5217SJeff Kirsher out_unregister_mdio: 18852a80b5e1SJonas Gorski if (priv->mii_bus) 1886adfc5217SJeff Kirsher mdiobus_unregister(priv->mii_bus); 1887adfc5217SJeff Kirsher 1888adfc5217SJeff Kirsher out_free_mdio: 1889adfc5217SJeff Kirsher if (priv->mii_bus) 1890adfc5217SJeff Kirsher mdiobus_free(priv->mii_bus); 1891adfc5217SJeff Kirsher 1892adfc5217SJeff Kirsher out_uninit_hw: 1893adfc5217SJeff Kirsher /* turn off mdc clock */ 1894adfc5217SJeff Kirsher enet_writel(priv, 0, ENET_MIISC_REG); 1895adfc5217SJeff Kirsher if (priv->phy_clk) { 1896624e2d21SJonas Gorski clk_disable_unprepare(priv->phy_clk); 1897adfc5217SJeff Kirsher clk_put(priv->phy_clk); 1898adfc5217SJeff Kirsher } 1899adfc5217SJeff Kirsher 1900adfc5217SJeff Kirsher out_put_clk_mac: 1901624e2d21SJonas Gorski clk_disable_unprepare(priv->mac_clk); 1902adfc5217SJeff Kirsher clk_put(priv->mac_clk); 1903adfc5217SJeff Kirsher out: 1904adfc5217SJeff Kirsher free_netdev(dev); 1905adfc5217SJeff Kirsher return ret; 1906adfc5217SJeff Kirsher } 1907adfc5217SJeff Kirsher 1908adfc5217SJeff Kirsher 1909adfc5217SJeff Kirsher /* 1910adfc5217SJeff Kirsher * exit func, stops hardware and unregisters netdevice 1911adfc5217SJeff Kirsher */ 1912047fc566SBill Pemberton static int bcm_enet_remove(struct platform_device *pdev) 1913adfc5217SJeff Kirsher { 1914adfc5217SJeff Kirsher struct bcm_enet_priv *priv; 1915adfc5217SJeff Kirsher struct net_device *dev; 1916adfc5217SJeff Kirsher 1917adfc5217SJeff Kirsher /* stop netdevice */ 1918adfc5217SJeff Kirsher dev = platform_get_drvdata(pdev); 1919adfc5217SJeff Kirsher priv = netdev_priv(dev); 1920adfc5217SJeff Kirsher unregister_netdev(dev); 1921adfc5217SJeff Kirsher 1922adfc5217SJeff Kirsher /* turn off mdc clock */ 1923adfc5217SJeff Kirsher enet_writel(priv, 0, ENET_MIISC_REG); 1924adfc5217SJeff Kirsher 1925adfc5217SJeff Kirsher if (priv->has_phy) { 1926adfc5217SJeff Kirsher mdiobus_unregister(priv->mii_bus); 1927adfc5217SJeff Kirsher mdiobus_free(priv->mii_bus); 1928adfc5217SJeff Kirsher } else { 1929adfc5217SJeff Kirsher struct bcm63xx_enet_platform_data *pd; 1930adfc5217SJeff Kirsher 1931adfc5217SJeff Kirsher pd = pdev->dev.platform_data; 1932adfc5217SJeff Kirsher if (pd && pd->mii_config) 1933adfc5217SJeff Kirsher pd->mii_config(dev, 0, bcm_enet_mdio_read_mii, 1934adfc5217SJeff Kirsher bcm_enet_mdio_write_mii); 1935adfc5217SJeff Kirsher } 1936adfc5217SJeff Kirsher 1937adfc5217SJeff Kirsher /* disable hw block clocks */ 1938adfc5217SJeff Kirsher if (priv->phy_clk) { 1939624e2d21SJonas Gorski clk_disable_unprepare(priv->phy_clk); 1940adfc5217SJeff Kirsher clk_put(priv->phy_clk); 1941adfc5217SJeff Kirsher } 1942624e2d21SJonas Gorski clk_disable_unprepare(priv->mac_clk); 1943adfc5217SJeff Kirsher clk_put(priv->mac_clk); 1944adfc5217SJeff Kirsher 1945adfc5217SJeff Kirsher free_netdev(dev); 1946adfc5217SJeff Kirsher return 0; 1947adfc5217SJeff Kirsher } 1948adfc5217SJeff Kirsher 1949adfc5217SJeff Kirsher struct platform_driver bcm63xx_enet_driver = { 1950adfc5217SJeff Kirsher .probe = bcm_enet_probe, 1951047fc566SBill Pemberton .remove = bcm_enet_remove, 1952adfc5217SJeff Kirsher .driver = { 1953adfc5217SJeff Kirsher .name = "bcm63xx_enet", 1954adfc5217SJeff Kirsher .owner = THIS_MODULE, 1955adfc5217SJeff Kirsher }, 1956adfc5217SJeff Kirsher }; 1957adfc5217SJeff Kirsher 1958adfc5217SJeff Kirsher /* 19596f00a022SMaxime Bizon * switch mii access callbacks 1960adfc5217SJeff Kirsher */ 19616f00a022SMaxime Bizon static int bcmenet_sw_mdio_read(struct bcm_enet_priv *priv, 19626f00a022SMaxime Bizon int ext, int phy_id, int location) 19636f00a022SMaxime Bizon { 19646f00a022SMaxime Bizon u32 reg; 19656f00a022SMaxime Bizon int ret; 19666f00a022SMaxime Bizon 19676f00a022SMaxime Bizon spin_lock_bh(&priv->enetsw_mdio_lock); 19686f00a022SMaxime Bizon enetsw_writel(priv, 0, ENETSW_MDIOC_REG); 19696f00a022SMaxime Bizon 19706f00a022SMaxime Bizon reg = ENETSW_MDIOC_RD_MASK | 19716f00a022SMaxime Bizon (phy_id << ENETSW_MDIOC_PHYID_SHIFT) | 19726f00a022SMaxime Bizon (location << ENETSW_MDIOC_REG_SHIFT); 19736f00a022SMaxime Bizon 19746f00a022SMaxime Bizon if (ext) 19756f00a022SMaxime Bizon reg |= ENETSW_MDIOC_EXT_MASK; 19766f00a022SMaxime Bizon 19776f00a022SMaxime Bizon enetsw_writel(priv, reg, ENETSW_MDIOC_REG); 19786f00a022SMaxime Bizon udelay(50); 19796f00a022SMaxime Bizon ret = enetsw_readw(priv, ENETSW_MDIOD_REG); 19806f00a022SMaxime Bizon spin_unlock_bh(&priv->enetsw_mdio_lock); 19816f00a022SMaxime Bizon return ret; 19826f00a022SMaxime Bizon } 19836f00a022SMaxime Bizon 19846f00a022SMaxime Bizon static void bcmenet_sw_mdio_write(struct bcm_enet_priv *priv, 19856f00a022SMaxime Bizon int ext, int phy_id, int location, 19866f00a022SMaxime Bizon uint16_t data) 19876f00a022SMaxime Bizon { 19886f00a022SMaxime Bizon u32 reg; 19896f00a022SMaxime Bizon 19906f00a022SMaxime Bizon spin_lock_bh(&priv->enetsw_mdio_lock); 19916f00a022SMaxime Bizon enetsw_writel(priv, 0, ENETSW_MDIOC_REG); 19926f00a022SMaxime Bizon 19936f00a022SMaxime Bizon reg = ENETSW_MDIOC_WR_MASK | 19946f00a022SMaxime Bizon (phy_id << ENETSW_MDIOC_PHYID_SHIFT) | 19956f00a022SMaxime Bizon (location << ENETSW_MDIOC_REG_SHIFT); 19966f00a022SMaxime Bizon 19976f00a022SMaxime Bizon if (ext) 19986f00a022SMaxime Bizon reg |= ENETSW_MDIOC_EXT_MASK; 19996f00a022SMaxime Bizon 20006f00a022SMaxime Bizon reg |= data; 20016f00a022SMaxime Bizon 20026f00a022SMaxime Bizon enetsw_writel(priv, reg, ENETSW_MDIOC_REG); 20036f00a022SMaxime Bizon udelay(50); 20046f00a022SMaxime Bizon spin_unlock_bh(&priv->enetsw_mdio_lock); 20056f00a022SMaxime Bizon } 20066f00a022SMaxime Bizon 20076f00a022SMaxime Bizon static inline int bcm_enet_port_is_rgmii(int portid) 20086f00a022SMaxime Bizon { 20096f00a022SMaxime Bizon return portid >= ENETSW_RGMII_PORT0; 20106f00a022SMaxime Bizon } 20116f00a022SMaxime Bizon 20126f00a022SMaxime Bizon /* 20136f00a022SMaxime Bizon * enet sw PHY polling 20146f00a022SMaxime Bizon */ 20156f00a022SMaxime Bizon static void swphy_poll_timer(unsigned long data) 20166f00a022SMaxime Bizon { 20176f00a022SMaxime Bizon struct bcm_enet_priv *priv = (struct bcm_enet_priv *)data; 20186f00a022SMaxime Bizon unsigned int i; 20196f00a022SMaxime Bizon 20206f00a022SMaxime Bizon for (i = 0; i < priv->num_ports; i++) { 20216f00a022SMaxime Bizon struct bcm63xx_enetsw_port *port; 20226f00a022SMaxime Bizon int val, j, up, advertise, lpa, lpa2, speed, duplex, media; 20236f00a022SMaxime Bizon int external_phy = bcm_enet_port_is_rgmii(i); 20246f00a022SMaxime Bizon u8 override; 20256f00a022SMaxime Bizon 20266f00a022SMaxime Bizon port = &priv->used_ports[i]; 20276f00a022SMaxime Bizon if (!port->used) 20286f00a022SMaxime Bizon continue; 20296f00a022SMaxime Bizon 20306f00a022SMaxime Bizon if (port->bypass_link) 20316f00a022SMaxime Bizon continue; 20326f00a022SMaxime Bizon 20336f00a022SMaxime Bizon /* dummy read to clear */ 20346f00a022SMaxime Bizon for (j = 0; j < 2; j++) 20356f00a022SMaxime Bizon val = bcmenet_sw_mdio_read(priv, external_phy, 20366f00a022SMaxime Bizon port->phy_id, MII_BMSR); 20376f00a022SMaxime Bizon 20386f00a022SMaxime Bizon if (val == 0xffff) 20396f00a022SMaxime Bizon continue; 20406f00a022SMaxime Bizon 20416f00a022SMaxime Bizon up = (val & BMSR_LSTATUS) ? 1 : 0; 20426f00a022SMaxime Bizon if (!(up ^ priv->sw_port_link[i])) 20436f00a022SMaxime Bizon continue; 20446f00a022SMaxime Bizon 20456f00a022SMaxime Bizon priv->sw_port_link[i] = up; 20466f00a022SMaxime Bizon 20476f00a022SMaxime Bizon /* link changed */ 20486f00a022SMaxime Bizon if (!up) { 20496f00a022SMaxime Bizon dev_info(&priv->pdev->dev, "link DOWN on %s\n", 20506f00a022SMaxime Bizon port->name); 20516f00a022SMaxime Bizon enetsw_writeb(priv, ENETSW_PORTOV_ENABLE_MASK, 20526f00a022SMaxime Bizon ENETSW_PORTOV_REG(i)); 20536f00a022SMaxime Bizon enetsw_writeb(priv, ENETSW_PTCTRL_RXDIS_MASK | 20546f00a022SMaxime Bizon ENETSW_PTCTRL_TXDIS_MASK, 20556f00a022SMaxime Bizon ENETSW_PTCTRL_REG(i)); 20566f00a022SMaxime Bizon continue; 20576f00a022SMaxime Bizon } 20586f00a022SMaxime Bizon 20596f00a022SMaxime Bizon advertise = bcmenet_sw_mdio_read(priv, external_phy, 20606f00a022SMaxime Bizon port->phy_id, MII_ADVERTISE); 20616f00a022SMaxime Bizon 20626f00a022SMaxime Bizon lpa = bcmenet_sw_mdio_read(priv, external_phy, port->phy_id, 20636f00a022SMaxime Bizon MII_LPA); 20646f00a022SMaxime Bizon 20656f00a022SMaxime Bizon lpa2 = bcmenet_sw_mdio_read(priv, external_phy, port->phy_id, 20666f00a022SMaxime Bizon MII_STAT1000); 20676f00a022SMaxime Bizon 20686f00a022SMaxime Bizon /* figure out media and duplex from advertise and LPA values */ 20696f00a022SMaxime Bizon media = mii_nway_result(lpa & advertise); 20706f00a022SMaxime Bizon duplex = (media & ADVERTISE_FULL) ? 1 : 0; 20716f00a022SMaxime Bizon if (lpa2 & LPA_1000FULL) 20726f00a022SMaxime Bizon duplex = 1; 20736f00a022SMaxime Bizon 20746f00a022SMaxime Bizon if (lpa2 & (LPA_1000FULL | LPA_1000HALF)) 20756f00a022SMaxime Bizon speed = 1000; 20766f00a022SMaxime Bizon else { 20776f00a022SMaxime Bizon if (media & (ADVERTISE_100FULL | ADVERTISE_100HALF)) 20786f00a022SMaxime Bizon speed = 100; 20796f00a022SMaxime Bizon else 20806f00a022SMaxime Bizon speed = 10; 20816f00a022SMaxime Bizon } 20826f00a022SMaxime Bizon 20836f00a022SMaxime Bizon dev_info(&priv->pdev->dev, 20846f00a022SMaxime Bizon "link UP on %s, %dMbps, %s-duplex\n", 20856f00a022SMaxime Bizon port->name, speed, duplex ? "full" : "half"); 20866f00a022SMaxime Bizon 20876f00a022SMaxime Bizon override = ENETSW_PORTOV_ENABLE_MASK | 20886f00a022SMaxime Bizon ENETSW_PORTOV_LINKUP_MASK; 20896f00a022SMaxime Bizon 20906f00a022SMaxime Bizon if (speed == 1000) 20916f00a022SMaxime Bizon override |= ENETSW_IMPOV_1000_MASK; 20926f00a022SMaxime Bizon else if (speed == 100) 20936f00a022SMaxime Bizon override |= ENETSW_IMPOV_100_MASK; 20946f00a022SMaxime Bizon if (duplex) 20956f00a022SMaxime Bizon override |= ENETSW_IMPOV_FDX_MASK; 20966f00a022SMaxime Bizon 20976f00a022SMaxime Bizon enetsw_writeb(priv, override, ENETSW_PORTOV_REG(i)); 20986f00a022SMaxime Bizon enetsw_writeb(priv, 0, ENETSW_PTCTRL_REG(i)); 20996f00a022SMaxime Bizon } 21006f00a022SMaxime Bizon 21016f00a022SMaxime Bizon priv->swphy_poll.expires = jiffies + HZ; 21026f00a022SMaxime Bizon add_timer(&priv->swphy_poll); 21036f00a022SMaxime Bizon } 21046f00a022SMaxime Bizon 21056f00a022SMaxime Bizon /* 21066f00a022SMaxime Bizon * open callback, allocate dma rings & buffers and start rx operation 21076f00a022SMaxime Bizon */ 21086f00a022SMaxime Bizon static int bcm_enetsw_open(struct net_device *dev) 21096f00a022SMaxime Bizon { 21106f00a022SMaxime Bizon struct bcm_enet_priv *priv; 21116f00a022SMaxime Bizon struct device *kdev; 21126f00a022SMaxime Bizon int i, ret; 21136f00a022SMaxime Bizon unsigned int size; 21146f00a022SMaxime Bizon void *p; 21156f00a022SMaxime Bizon u32 val; 21166f00a022SMaxime Bizon 21176f00a022SMaxime Bizon priv = netdev_priv(dev); 21186f00a022SMaxime Bizon kdev = &priv->pdev->dev; 21196f00a022SMaxime Bizon 21206f00a022SMaxime Bizon /* mask all interrupts and request them */ 21216f00a022SMaxime Bizon enet_dmac_writel(priv, 0, ENETDMAC_IRMASK_REG(priv->rx_chan)); 21226f00a022SMaxime Bizon enet_dmac_writel(priv, 0, ENETDMAC_IRMASK_REG(priv->tx_chan)); 21236f00a022SMaxime Bizon 21246f00a022SMaxime Bizon ret = request_irq(priv->irq_rx, bcm_enet_isr_dma, 21256f00a022SMaxime Bizon IRQF_DISABLED, dev->name, dev); 21266f00a022SMaxime Bizon if (ret) 21276f00a022SMaxime Bizon goto out_freeirq; 21286f00a022SMaxime Bizon 21296f00a022SMaxime Bizon if (priv->irq_tx != -1) { 21306f00a022SMaxime Bizon ret = request_irq(priv->irq_tx, bcm_enet_isr_dma, 21316f00a022SMaxime Bizon IRQF_DISABLED, dev->name, dev); 21326f00a022SMaxime Bizon if (ret) 21336f00a022SMaxime Bizon goto out_freeirq_rx; 21346f00a022SMaxime Bizon } 21356f00a022SMaxime Bizon 21366f00a022SMaxime Bizon /* allocate rx dma ring */ 21376f00a022SMaxime Bizon size = priv->rx_ring_size * sizeof(struct bcm_enet_desc); 21386f00a022SMaxime Bizon p = dma_alloc_coherent(kdev, size, &priv->rx_desc_dma, GFP_KERNEL); 21396f00a022SMaxime Bizon if (!p) { 21406f00a022SMaxime Bizon dev_err(kdev, "cannot allocate rx ring %u\n", size); 21416f00a022SMaxime Bizon ret = -ENOMEM; 21426f00a022SMaxime Bizon goto out_freeirq_tx; 21436f00a022SMaxime Bizon } 21446f00a022SMaxime Bizon 21456f00a022SMaxime Bizon memset(p, 0, size); 21466f00a022SMaxime Bizon priv->rx_desc_alloc_size = size; 21476f00a022SMaxime Bizon priv->rx_desc_cpu = p; 21486f00a022SMaxime Bizon 21496f00a022SMaxime Bizon /* allocate tx dma ring */ 21506f00a022SMaxime Bizon size = priv->tx_ring_size * sizeof(struct bcm_enet_desc); 21516f00a022SMaxime Bizon p = dma_alloc_coherent(kdev, size, &priv->tx_desc_dma, GFP_KERNEL); 21526f00a022SMaxime Bizon if (!p) { 21536f00a022SMaxime Bizon dev_err(kdev, "cannot allocate tx ring\n"); 21546f00a022SMaxime Bizon ret = -ENOMEM; 21556f00a022SMaxime Bizon goto out_free_rx_ring; 21566f00a022SMaxime Bizon } 21576f00a022SMaxime Bizon 21586f00a022SMaxime Bizon memset(p, 0, size); 21596f00a022SMaxime Bizon priv->tx_desc_alloc_size = size; 21606f00a022SMaxime Bizon priv->tx_desc_cpu = p; 21616f00a022SMaxime Bizon 21626f00a022SMaxime Bizon priv->tx_skb = kzalloc(sizeof(struct sk_buff *) * priv->tx_ring_size, 21636f00a022SMaxime Bizon GFP_KERNEL); 21646f00a022SMaxime Bizon if (!priv->tx_skb) { 21656f00a022SMaxime Bizon dev_err(kdev, "cannot allocate rx skb queue\n"); 21666f00a022SMaxime Bizon ret = -ENOMEM; 21676f00a022SMaxime Bizon goto out_free_tx_ring; 21686f00a022SMaxime Bizon } 21696f00a022SMaxime Bizon 21706f00a022SMaxime Bizon priv->tx_desc_count = priv->tx_ring_size; 21716f00a022SMaxime Bizon priv->tx_dirty_desc = 0; 21726f00a022SMaxime Bizon priv->tx_curr_desc = 0; 21736f00a022SMaxime Bizon spin_lock_init(&priv->tx_lock); 21746f00a022SMaxime Bizon 21756f00a022SMaxime Bizon /* init & fill rx ring with skbs */ 21766f00a022SMaxime Bizon priv->rx_skb = kzalloc(sizeof(struct sk_buff *) * priv->rx_ring_size, 21776f00a022SMaxime Bizon GFP_KERNEL); 21786f00a022SMaxime Bizon if (!priv->rx_skb) { 21796f00a022SMaxime Bizon dev_err(kdev, "cannot allocate rx skb queue\n"); 21806f00a022SMaxime Bizon ret = -ENOMEM; 21816f00a022SMaxime Bizon goto out_free_tx_skb; 21826f00a022SMaxime Bizon } 21836f00a022SMaxime Bizon 21846f00a022SMaxime Bizon priv->rx_desc_count = 0; 21856f00a022SMaxime Bizon priv->rx_dirty_desc = 0; 21866f00a022SMaxime Bizon priv->rx_curr_desc = 0; 21876f00a022SMaxime Bizon 21886f00a022SMaxime Bizon /* disable all ports */ 21896f00a022SMaxime Bizon for (i = 0; i < priv->num_ports; i++) { 21906f00a022SMaxime Bizon enetsw_writeb(priv, ENETSW_PORTOV_ENABLE_MASK, 21916f00a022SMaxime Bizon ENETSW_PORTOV_REG(i)); 21926f00a022SMaxime Bizon enetsw_writeb(priv, ENETSW_PTCTRL_RXDIS_MASK | 21936f00a022SMaxime Bizon ENETSW_PTCTRL_TXDIS_MASK, 21946f00a022SMaxime Bizon ENETSW_PTCTRL_REG(i)); 21956f00a022SMaxime Bizon 21966f00a022SMaxime Bizon priv->sw_port_link[i] = 0; 21976f00a022SMaxime Bizon } 21986f00a022SMaxime Bizon 21996f00a022SMaxime Bizon /* reset mib */ 22006f00a022SMaxime Bizon val = enetsw_readb(priv, ENETSW_GMCR_REG); 22016f00a022SMaxime Bizon val |= ENETSW_GMCR_RST_MIB_MASK; 22026f00a022SMaxime Bizon enetsw_writeb(priv, val, ENETSW_GMCR_REG); 22036f00a022SMaxime Bizon mdelay(1); 22046f00a022SMaxime Bizon val &= ~ENETSW_GMCR_RST_MIB_MASK; 22056f00a022SMaxime Bizon enetsw_writeb(priv, val, ENETSW_GMCR_REG); 22066f00a022SMaxime Bizon mdelay(1); 22076f00a022SMaxime Bizon 22086f00a022SMaxime Bizon /* force CPU port state */ 22096f00a022SMaxime Bizon val = enetsw_readb(priv, ENETSW_IMPOV_REG); 22106f00a022SMaxime Bizon val |= ENETSW_IMPOV_FORCE_MASK | ENETSW_IMPOV_LINKUP_MASK; 22116f00a022SMaxime Bizon enetsw_writeb(priv, val, ENETSW_IMPOV_REG); 22126f00a022SMaxime Bizon 22136f00a022SMaxime Bizon /* enable switch forward engine */ 22146f00a022SMaxime Bizon val = enetsw_readb(priv, ENETSW_SWMODE_REG); 22156f00a022SMaxime Bizon val |= ENETSW_SWMODE_FWD_EN_MASK; 22166f00a022SMaxime Bizon enetsw_writeb(priv, val, ENETSW_SWMODE_REG); 22176f00a022SMaxime Bizon 22186f00a022SMaxime Bizon /* enable jumbo on all ports */ 22196f00a022SMaxime Bizon enetsw_writel(priv, 0x1ff, ENETSW_JMBCTL_PORT_REG); 22206f00a022SMaxime Bizon enetsw_writew(priv, 9728, ENETSW_JMBCTL_MAXSIZE_REG); 22216f00a022SMaxime Bizon 22226f00a022SMaxime Bizon /* initialize flow control buffer allocation */ 22236f00a022SMaxime Bizon enet_dma_writel(priv, ENETDMA_BUFALLOC_FORCE_MASK | 0, 22246f00a022SMaxime Bizon ENETDMA_BUFALLOC_REG(priv->rx_chan)); 22256f00a022SMaxime Bizon 22266f00a022SMaxime Bizon if (bcm_enet_refill_rx(dev)) { 22276f00a022SMaxime Bizon dev_err(kdev, "cannot allocate rx skb queue\n"); 22286f00a022SMaxime Bizon ret = -ENOMEM; 22296f00a022SMaxime Bizon goto out; 22306f00a022SMaxime Bizon } 22316f00a022SMaxime Bizon 22326f00a022SMaxime Bizon /* write rx & tx ring addresses */ 22336f00a022SMaxime Bizon enet_dmas_writel(priv, priv->rx_desc_dma, 22346f00a022SMaxime Bizon ENETDMAS_RSTART_REG(priv->rx_chan)); 22356f00a022SMaxime Bizon enet_dmas_writel(priv, priv->tx_desc_dma, 22366f00a022SMaxime Bizon ENETDMAS_RSTART_REG(priv->tx_chan)); 22376f00a022SMaxime Bizon 22386f00a022SMaxime Bizon /* clear remaining state ram for rx & tx channel */ 22396f00a022SMaxime Bizon enet_dmas_writel(priv, 0, ENETDMAS_SRAM2_REG(priv->rx_chan)); 22406f00a022SMaxime Bizon enet_dmas_writel(priv, 0, ENETDMAS_SRAM2_REG(priv->tx_chan)); 22416f00a022SMaxime Bizon enet_dmas_writel(priv, 0, ENETDMAS_SRAM3_REG(priv->rx_chan)); 22426f00a022SMaxime Bizon enet_dmas_writel(priv, 0, ENETDMAS_SRAM3_REG(priv->tx_chan)); 22436f00a022SMaxime Bizon enet_dmas_writel(priv, 0, ENETDMAS_SRAM4_REG(priv->rx_chan)); 22446f00a022SMaxime Bizon enet_dmas_writel(priv, 0, ENETDMAS_SRAM4_REG(priv->tx_chan)); 22456f00a022SMaxime Bizon 22466f00a022SMaxime Bizon /* set dma maximum burst len */ 22476f00a022SMaxime Bizon enet_dmac_writel(priv, priv->dma_maxburst, 22486f00a022SMaxime Bizon ENETDMAC_MAXBURST_REG(priv->rx_chan)); 22496f00a022SMaxime Bizon enet_dmac_writel(priv, priv->dma_maxburst, 22506f00a022SMaxime Bizon ENETDMAC_MAXBURST_REG(priv->tx_chan)); 22516f00a022SMaxime Bizon 22526f00a022SMaxime Bizon /* set flow control low/high threshold to 1/3 / 2/3 */ 22536f00a022SMaxime Bizon val = priv->rx_ring_size / 3; 22546f00a022SMaxime Bizon enet_dma_writel(priv, val, ENETDMA_FLOWCL_REG(priv->rx_chan)); 22556f00a022SMaxime Bizon val = (priv->rx_ring_size * 2) / 3; 22566f00a022SMaxime Bizon enet_dma_writel(priv, val, ENETDMA_FLOWCH_REG(priv->rx_chan)); 22576f00a022SMaxime Bizon 22586f00a022SMaxime Bizon /* all set, enable mac and interrupts, start dma engine and 22596f00a022SMaxime Bizon * kick rx dma channel 22606f00a022SMaxime Bizon */ 22616f00a022SMaxime Bizon wmb(); 22626f00a022SMaxime Bizon enet_dma_writel(priv, ENETDMA_CFG_EN_MASK, ENETDMA_CFG_REG); 22636f00a022SMaxime Bizon enet_dmac_writel(priv, ENETDMAC_CHANCFG_EN_MASK, 22646f00a022SMaxime Bizon ENETDMAC_CHANCFG_REG(priv->rx_chan)); 22656f00a022SMaxime Bizon 22666f00a022SMaxime Bizon /* watch "packet transferred" interrupt in rx and tx */ 22676f00a022SMaxime Bizon enet_dmac_writel(priv, ENETDMAC_IR_PKTDONE_MASK, 22686f00a022SMaxime Bizon ENETDMAC_IR_REG(priv->rx_chan)); 22696f00a022SMaxime Bizon enet_dmac_writel(priv, ENETDMAC_IR_PKTDONE_MASK, 22706f00a022SMaxime Bizon ENETDMAC_IR_REG(priv->tx_chan)); 22716f00a022SMaxime Bizon 22726f00a022SMaxime Bizon /* make sure we enable napi before rx interrupt */ 22736f00a022SMaxime Bizon napi_enable(&priv->napi); 22746f00a022SMaxime Bizon 22756f00a022SMaxime Bizon enet_dmac_writel(priv, ENETDMAC_IR_PKTDONE_MASK, 22766f00a022SMaxime Bizon ENETDMAC_IRMASK_REG(priv->rx_chan)); 22776f00a022SMaxime Bizon enet_dmac_writel(priv, ENETDMAC_IR_PKTDONE_MASK, 22786f00a022SMaxime Bizon ENETDMAC_IRMASK_REG(priv->tx_chan)); 22796f00a022SMaxime Bizon 22806f00a022SMaxime Bizon netif_carrier_on(dev); 22816f00a022SMaxime Bizon netif_start_queue(dev); 22826f00a022SMaxime Bizon 22836f00a022SMaxime Bizon /* apply override config for bypass_link ports here. */ 22846f00a022SMaxime Bizon for (i = 0; i < priv->num_ports; i++) { 22856f00a022SMaxime Bizon struct bcm63xx_enetsw_port *port; 22866f00a022SMaxime Bizon u8 override; 22876f00a022SMaxime Bizon port = &priv->used_ports[i]; 22886f00a022SMaxime Bizon if (!port->used) 22896f00a022SMaxime Bizon continue; 22906f00a022SMaxime Bizon 22916f00a022SMaxime Bizon if (!port->bypass_link) 22926f00a022SMaxime Bizon continue; 22936f00a022SMaxime Bizon 22946f00a022SMaxime Bizon override = ENETSW_PORTOV_ENABLE_MASK | 22956f00a022SMaxime Bizon ENETSW_PORTOV_LINKUP_MASK; 22966f00a022SMaxime Bizon 22976f00a022SMaxime Bizon switch (port->force_speed) { 22986f00a022SMaxime Bizon case 1000: 22996f00a022SMaxime Bizon override |= ENETSW_IMPOV_1000_MASK; 23006f00a022SMaxime Bizon break; 23016f00a022SMaxime Bizon case 100: 23026f00a022SMaxime Bizon override |= ENETSW_IMPOV_100_MASK; 23036f00a022SMaxime Bizon break; 23046f00a022SMaxime Bizon case 10: 23056f00a022SMaxime Bizon break; 23066f00a022SMaxime Bizon default: 23076f00a022SMaxime Bizon pr_warn("invalid forced speed on port %s: assume 10\n", 23086f00a022SMaxime Bizon port->name); 23096f00a022SMaxime Bizon break; 23106f00a022SMaxime Bizon } 23116f00a022SMaxime Bizon 23126f00a022SMaxime Bizon if (port->force_duplex_full) 23136f00a022SMaxime Bizon override |= ENETSW_IMPOV_FDX_MASK; 23146f00a022SMaxime Bizon 23156f00a022SMaxime Bizon 23166f00a022SMaxime Bizon enetsw_writeb(priv, override, ENETSW_PORTOV_REG(i)); 23176f00a022SMaxime Bizon enetsw_writeb(priv, 0, ENETSW_PTCTRL_REG(i)); 23186f00a022SMaxime Bizon } 23196f00a022SMaxime Bizon 23206f00a022SMaxime Bizon /* start phy polling timer */ 23216f00a022SMaxime Bizon init_timer(&priv->swphy_poll); 23226f00a022SMaxime Bizon priv->swphy_poll.function = swphy_poll_timer; 23236f00a022SMaxime Bizon priv->swphy_poll.data = (unsigned long)priv; 23246f00a022SMaxime Bizon priv->swphy_poll.expires = jiffies; 23256f00a022SMaxime Bizon add_timer(&priv->swphy_poll); 23266f00a022SMaxime Bizon return 0; 23276f00a022SMaxime Bizon 23286f00a022SMaxime Bizon out: 23296f00a022SMaxime Bizon for (i = 0; i < priv->rx_ring_size; i++) { 23306f00a022SMaxime Bizon struct bcm_enet_desc *desc; 23316f00a022SMaxime Bizon 23326f00a022SMaxime Bizon if (!priv->rx_skb[i]) 23336f00a022SMaxime Bizon continue; 23346f00a022SMaxime Bizon 23356f00a022SMaxime Bizon desc = &priv->rx_desc_cpu[i]; 23366f00a022SMaxime Bizon dma_unmap_single(kdev, desc->address, priv->rx_skb_size, 23376f00a022SMaxime Bizon DMA_FROM_DEVICE); 23386f00a022SMaxime Bizon kfree_skb(priv->rx_skb[i]); 23396f00a022SMaxime Bizon } 23406f00a022SMaxime Bizon kfree(priv->rx_skb); 23416f00a022SMaxime Bizon 23426f00a022SMaxime Bizon out_free_tx_skb: 23436f00a022SMaxime Bizon kfree(priv->tx_skb); 23446f00a022SMaxime Bizon 23456f00a022SMaxime Bizon out_free_tx_ring: 23466f00a022SMaxime Bizon dma_free_coherent(kdev, priv->tx_desc_alloc_size, 23476f00a022SMaxime Bizon priv->tx_desc_cpu, priv->tx_desc_dma); 23486f00a022SMaxime Bizon 23496f00a022SMaxime Bizon out_free_rx_ring: 23506f00a022SMaxime Bizon dma_free_coherent(kdev, priv->rx_desc_alloc_size, 23516f00a022SMaxime Bizon priv->rx_desc_cpu, priv->rx_desc_dma); 23526f00a022SMaxime Bizon 23536f00a022SMaxime Bizon out_freeirq_tx: 23546f00a022SMaxime Bizon if (priv->irq_tx != -1) 23556f00a022SMaxime Bizon free_irq(priv->irq_tx, dev); 23566f00a022SMaxime Bizon 23576f00a022SMaxime Bizon out_freeirq_rx: 23586f00a022SMaxime Bizon free_irq(priv->irq_rx, dev); 23596f00a022SMaxime Bizon 23606f00a022SMaxime Bizon out_freeirq: 23616f00a022SMaxime Bizon return ret; 23626f00a022SMaxime Bizon } 23636f00a022SMaxime Bizon 23646f00a022SMaxime Bizon /* stop callback */ 23656f00a022SMaxime Bizon static int bcm_enetsw_stop(struct net_device *dev) 23666f00a022SMaxime Bizon { 23676f00a022SMaxime Bizon struct bcm_enet_priv *priv; 23686f00a022SMaxime Bizon struct device *kdev; 23696f00a022SMaxime Bizon int i; 23706f00a022SMaxime Bizon 23716f00a022SMaxime Bizon priv = netdev_priv(dev); 23726f00a022SMaxime Bizon kdev = &priv->pdev->dev; 23736f00a022SMaxime Bizon 23746f00a022SMaxime Bizon del_timer_sync(&priv->swphy_poll); 23756f00a022SMaxime Bizon netif_stop_queue(dev); 23766f00a022SMaxime Bizon napi_disable(&priv->napi); 23776f00a022SMaxime Bizon del_timer_sync(&priv->rx_timeout); 23786f00a022SMaxime Bizon 23796f00a022SMaxime Bizon /* mask all interrupts */ 23806f00a022SMaxime Bizon enet_dmac_writel(priv, 0, ENETDMAC_IRMASK_REG(priv->rx_chan)); 23816f00a022SMaxime Bizon enet_dmac_writel(priv, 0, ENETDMAC_IRMASK_REG(priv->tx_chan)); 23826f00a022SMaxime Bizon 23836f00a022SMaxime Bizon /* disable dma & mac */ 23846f00a022SMaxime Bizon bcm_enet_disable_dma(priv, priv->tx_chan); 23856f00a022SMaxime Bizon bcm_enet_disable_dma(priv, priv->rx_chan); 23866f00a022SMaxime Bizon 23876f00a022SMaxime Bizon /* force reclaim of all tx buffers */ 23886f00a022SMaxime Bizon bcm_enet_tx_reclaim(dev, 1); 23896f00a022SMaxime Bizon 23906f00a022SMaxime Bizon /* free the rx skb ring */ 23916f00a022SMaxime Bizon for (i = 0; i < priv->rx_ring_size; i++) { 23926f00a022SMaxime Bizon struct bcm_enet_desc *desc; 23936f00a022SMaxime Bizon 23946f00a022SMaxime Bizon if (!priv->rx_skb[i]) 23956f00a022SMaxime Bizon continue; 23966f00a022SMaxime Bizon 23976f00a022SMaxime Bizon desc = &priv->rx_desc_cpu[i]; 23986f00a022SMaxime Bizon dma_unmap_single(kdev, desc->address, priv->rx_skb_size, 23996f00a022SMaxime Bizon DMA_FROM_DEVICE); 24006f00a022SMaxime Bizon kfree_skb(priv->rx_skb[i]); 24016f00a022SMaxime Bizon } 24026f00a022SMaxime Bizon 24036f00a022SMaxime Bizon /* free remaining allocated memory */ 24046f00a022SMaxime Bizon kfree(priv->rx_skb); 24056f00a022SMaxime Bizon kfree(priv->tx_skb); 24066f00a022SMaxime Bizon dma_free_coherent(kdev, priv->rx_desc_alloc_size, 24076f00a022SMaxime Bizon priv->rx_desc_cpu, priv->rx_desc_dma); 24086f00a022SMaxime Bizon dma_free_coherent(kdev, priv->tx_desc_alloc_size, 24096f00a022SMaxime Bizon priv->tx_desc_cpu, priv->tx_desc_dma); 24106f00a022SMaxime Bizon if (priv->irq_tx != -1) 24116f00a022SMaxime Bizon free_irq(priv->irq_tx, dev); 24126f00a022SMaxime Bizon free_irq(priv->irq_rx, dev); 24136f00a022SMaxime Bizon 24146f00a022SMaxime Bizon return 0; 24156f00a022SMaxime Bizon } 24166f00a022SMaxime Bizon 24176f00a022SMaxime Bizon /* try to sort out phy external status by walking the used_port field 24186f00a022SMaxime Bizon * in the bcm_enet_priv structure. in case the phy address is not 24196f00a022SMaxime Bizon * assigned to any physical port on the switch, assume it is external 24206f00a022SMaxime Bizon * (and yell at the user). 24216f00a022SMaxime Bizon */ 24226f00a022SMaxime Bizon static int bcm_enetsw_phy_is_external(struct bcm_enet_priv *priv, int phy_id) 24236f00a022SMaxime Bizon { 24246f00a022SMaxime Bizon int i; 24256f00a022SMaxime Bizon 24266f00a022SMaxime Bizon for (i = 0; i < priv->num_ports; ++i) { 24276f00a022SMaxime Bizon if (!priv->used_ports[i].used) 24286f00a022SMaxime Bizon continue; 24296f00a022SMaxime Bizon if (priv->used_ports[i].phy_id == phy_id) 24306f00a022SMaxime Bizon return bcm_enet_port_is_rgmii(i); 24316f00a022SMaxime Bizon } 24326f00a022SMaxime Bizon 24336f00a022SMaxime Bizon printk_once(KERN_WARNING "bcm63xx_enet: could not find a used port with phy_id %i, assuming phy is external\n", 24346f00a022SMaxime Bizon phy_id); 24356f00a022SMaxime Bizon return 1; 24366f00a022SMaxime Bizon } 24376f00a022SMaxime Bizon 24386f00a022SMaxime Bizon /* can't use bcmenet_sw_mdio_read directly as we need to sort out 24396f00a022SMaxime Bizon * external/internal status of the given phy_id first. 24406f00a022SMaxime Bizon */ 24416f00a022SMaxime Bizon static int bcm_enetsw_mii_mdio_read(struct net_device *dev, int phy_id, 24426f00a022SMaxime Bizon int location) 24436f00a022SMaxime Bizon { 24446f00a022SMaxime Bizon struct bcm_enet_priv *priv; 24456f00a022SMaxime Bizon 24466f00a022SMaxime Bizon priv = netdev_priv(dev); 24476f00a022SMaxime Bizon return bcmenet_sw_mdio_read(priv, 24486f00a022SMaxime Bizon bcm_enetsw_phy_is_external(priv, phy_id), 24496f00a022SMaxime Bizon phy_id, location); 24506f00a022SMaxime Bizon } 24516f00a022SMaxime Bizon 24526f00a022SMaxime Bizon /* can't use bcmenet_sw_mdio_write directly as we need to sort out 24536f00a022SMaxime Bizon * external/internal status of the given phy_id first. 24546f00a022SMaxime Bizon */ 24556f00a022SMaxime Bizon static void bcm_enetsw_mii_mdio_write(struct net_device *dev, int phy_id, 24566f00a022SMaxime Bizon int location, 24576f00a022SMaxime Bizon int val) 24586f00a022SMaxime Bizon { 24596f00a022SMaxime Bizon struct bcm_enet_priv *priv; 24606f00a022SMaxime Bizon 24616f00a022SMaxime Bizon priv = netdev_priv(dev); 24626f00a022SMaxime Bizon bcmenet_sw_mdio_write(priv, bcm_enetsw_phy_is_external(priv, phy_id), 24636f00a022SMaxime Bizon phy_id, location, val); 24646f00a022SMaxime Bizon } 24656f00a022SMaxime Bizon 24666f00a022SMaxime Bizon static int bcm_enetsw_ioctl(struct net_device *dev, struct ifreq *rq, int cmd) 24676f00a022SMaxime Bizon { 24686f00a022SMaxime Bizon struct mii_if_info mii; 24696f00a022SMaxime Bizon 24706f00a022SMaxime Bizon mii.dev = dev; 24716f00a022SMaxime Bizon mii.mdio_read = bcm_enetsw_mii_mdio_read; 24726f00a022SMaxime Bizon mii.mdio_write = bcm_enetsw_mii_mdio_write; 24736f00a022SMaxime Bizon mii.phy_id = 0; 24746f00a022SMaxime Bizon mii.phy_id_mask = 0x3f; 24756f00a022SMaxime Bizon mii.reg_num_mask = 0x1f; 24766f00a022SMaxime Bizon return generic_mii_ioctl(&mii, if_mii(rq), cmd, NULL); 24776f00a022SMaxime Bizon 24786f00a022SMaxime Bizon } 24796f00a022SMaxime Bizon 24806f00a022SMaxime Bizon static const struct net_device_ops bcm_enetsw_ops = { 24816f00a022SMaxime Bizon .ndo_open = bcm_enetsw_open, 24826f00a022SMaxime Bizon .ndo_stop = bcm_enetsw_stop, 24836f00a022SMaxime Bizon .ndo_start_xmit = bcm_enet_start_xmit, 24846f00a022SMaxime Bizon .ndo_change_mtu = bcm_enet_change_mtu, 24856f00a022SMaxime Bizon .ndo_do_ioctl = bcm_enetsw_ioctl, 24866f00a022SMaxime Bizon }; 24876f00a022SMaxime Bizon 24886f00a022SMaxime Bizon 24896f00a022SMaxime Bizon static const struct bcm_enet_stats bcm_enetsw_gstrings_stats[] = { 24906f00a022SMaxime Bizon { "rx_packets", DEV_STAT(rx_packets), -1 }, 24916f00a022SMaxime Bizon { "tx_packets", DEV_STAT(tx_packets), -1 }, 24926f00a022SMaxime Bizon { "rx_bytes", DEV_STAT(rx_bytes), -1 }, 24936f00a022SMaxime Bizon { "tx_bytes", DEV_STAT(tx_bytes), -1 }, 24946f00a022SMaxime Bizon { "rx_errors", DEV_STAT(rx_errors), -1 }, 24956f00a022SMaxime Bizon { "tx_errors", DEV_STAT(tx_errors), -1 }, 24966f00a022SMaxime Bizon { "rx_dropped", DEV_STAT(rx_dropped), -1 }, 24976f00a022SMaxime Bizon { "tx_dropped", DEV_STAT(tx_dropped), -1 }, 24986f00a022SMaxime Bizon 24996f00a022SMaxime Bizon { "tx_good_octets", GEN_STAT(mib.tx_gd_octets), ETHSW_MIB_RX_GD_OCT }, 25006f00a022SMaxime Bizon { "tx_unicast", GEN_STAT(mib.tx_unicast), ETHSW_MIB_RX_BRDCAST }, 25016f00a022SMaxime Bizon { "tx_broadcast", GEN_STAT(mib.tx_brdcast), ETHSW_MIB_RX_BRDCAST }, 25026f00a022SMaxime Bizon { "tx_multicast", GEN_STAT(mib.tx_mult), ETHSW_MIB_RX_MULT }, 25036f00a022SMaxime Bizon { "tx_64_octets", GEN_STAT(mib.tx_64), ETHSW_MIB_RX_64 }, 25046f00a022SMaxime Bizon { "tx_65_127_oct", GEN_STAT(mib.tx_65_127), ETHSW_MIB_RX_65_127 }, 25056f00a022SMaxime Bizon { "tx_128_255_oct", GEN_STAT(mib.tx_128_255), ETHSW_MIB_RX_128_255 }, 25066f00a022SMaxime Bizon { "tx_256_511_oct", GEN_STAT(mib.tx_256_511), ETHSW_MIB_RX_256_511 }, 25076f00a022SMaxime Bizon { "tx_512_1023_oct", GEN_STAT(mib.tx_512_1023), ETHSW_MIB_RX_512_1023}, 25086f00a022SMaxime Bizon { "tx_1024_1522_oct", GEN_STAT(mib.tx_1024_max), 25096f00a022SMaxime Bizon ETHSW_MIB_RX_1024_1522 }, 25106f00a022SMaxime Bizon { "tx_1523_2047_oct", GEN_STAT(mib.tx_1523_2047), 25116f00a022SMaxime Bizon ETHSW_MIB_RX_1523_2047 }, 25126f00a022SMaxime Bizon { "tx_2048_4095_oct", GEN_STAT(mib.tx_2048_4095), 25136f00a022SMaxime Bizon ETHSW_MIB_RX_2048_4095 }, 25146f00a022SMaxime Bizon { "tx_4096_8191_oct", GEN_STAT(mib.tx_4096_8191), 25156f00a022SMaxime Bizon ETHSW_MIB_RX_4096_8191 }, 25166f00a022SMaxime Bizon { "tx_8192_9728_oct", GEN_STAT(mib.tx_8192_9728), 25176f00a022SMaxime Bizon ETHSW_MIB_RX_8192_9728 }, 25186f00a022SMaxime Bizon { "tx_oversize", GEN_STAT(mib.tx_ovr), ETHSW_MIB_RX_OVR }, 25196f00a022SMaxime Bizon { "tx_oversize_drop", GEN_STAT(mib.tx_ovr), ETHSW_MIB_RX_OVR_DISC }, 25206f00a022SMaxime Bizon { "tx_dropped", GEN_STAT(mib.tx_drop), ETHSW_MIB_RX_DROP }, 25216f00a022SMaxime Bizon { "tx_undersize", GEN_STAT(mib.tx_underrun), ETHSW_MIB_RX_UND }, 25226f00a022SMaxime Bizon { "tx_pause", GEN_STAT(mib.tx_pause), ETHSW_MIB_RX_PAUSE }, 25236f00a022SMaxime Bizon 25246f00a022SMaxime Bizon { "rx_good_octets", GEN_STAT(mib.rx_gd_octets), ETHSW_MIB_TX_ALL_OCT }, 25256f00a022SMaxime Bizon { "rx_broadcast", GEN_STAT(mib.rx_brdcast), ETHSW_MIB_TX_BRDCAST }, 25266f00a022SMaxime Bizon { "rx_multicast", GEN_STAT(mib.rx_mult), ETHSW_MIB_TX_MULT }, 25276f00a022SMaxime Bizon { "rx_unicast", GEN_STAT(mib.rx_unicast), ETHSW_MIB_TX_MULT }, 25286f00a022SMaxime Bizon { "rx_pause", GEN_STAT(mib.rx_pause), ETHSW_MIB_TX_PAUSE }, 25296f00a022SMaxime Bizon { "rx_dropped", GEN_STAT(mib.rx_drop), ETHSW_MIB_TX_DROP_PKTS }, 25306f00a022SMaxime Bizon 25316f00a022SMaxime Bizon }; 25326f00a022SMaxime Bizon 25336f00a022SMaxime Bizon #define BCM_ENETSW_STATS_LEN \ 25346f00a022SMaxime Bizon (sizeof(bcm_enetsw_gstrings_stats) / sizeof(struct bcm_enet_stats)) 25356f00a022SMaxime Bizon 25366f00a022SMaxime Bizon static void bcm_enetsw_get_strings(struct net_device *netdev, 25376f00a022SMaxime Bizon u32 stringset, u8 *data) 25386f00a022SMaxime Bizon { 25396f00a022SMaxime Bizon int i; 25406f00a022SMaxime Bizon 25416f00a022SMaxime Bizon switch (stringset) { 25426f00a022SMaxime Bizon case ETH_SS_STATS: 25436f00a022SMaxime Bizon for (i = 0; i < BCM_ENETSW_STATS_LEN; i++) { 25446f00a022SMaxime Bizon memcpy(data + i * ETH_GSTRING_LEN, 25456f00a022SMaxime Bizon bcm_enetsw_gstrings_stats[i].stat_string, 25466f00a022SMaxime Bizon ETH_GSTRING_LEN); 25476f00a022SMaxime Bizon } 25486f00a022SMaxime Bizon break; 25496f00a022SMaxime Bizon } 25506f00a022SMaxime Bizon } 25516f00a022SMaxime Bizon 25526f00a022SMaxime Bizon static int bcm_enetsw_get_sset_count(struct net_device *netdev, 25536f00a022SMaxime Bizon int string_set) 25546f00a022SMaxime Bizon { 25556f00a022SMaxime Bizon switch (string_set) { 25566f00a022SMaxime Bizon case ETH_SS_STATS: 25576f00a022SMaxime Bizon return BCM_ENETSW_STATS_LEN; 25586f00a022SMaxime Bizon default: 25596f00a022SMaxime Bizon return -EINVAL; 25606f00a022SMaxime Bizon } 25616f00a022SMaxime Bizon } 25626f00a022SMaxime Bizon 25636f00a022SMaxime Bizon static void bcm_enetsw_get_drvinfo(struct net_device *netdev, 25646f00a022SMaxime Bizon struct ethtool_drvinfo *drvinfo) 25656f00a022SMaxime Bizon { 25666f00a022SMaxime Bizon strncpy(drvinfo->driver, bcm_enet_driver_name, 32); 25676f00a022SMaxime Bizon strncpy(drvinfo->version, bcm_enet_driver_version, 32); 25686f00a022SMaxime Bizon strncpy(drvinfo->fw_version, "N/A", 32); 25696f00a022SMaxime Bizon strncpy(drvinfo->bus_info, "bcm63xx", 32); 25706f00a022SMaxime Bizon drvinfo->n_stats = BCM_ENETSW_STATS_LEN; 25716f00a022SMaxime Bizon } 25726f00a022SMaxime Bizon 25736f00a022SMaxime Bizon static void bcm_enetsw_get_ethtool_stats(struct net_device *netdev, 25746f00a022SMaxime Bizon struct ethtool_stats *stats, 25756f00a022SMaxime Bizon u64 *data) 25766f00a022SMaxime Bizon { 25776f00a022SMaxime Bizon struct bcm_enet_priv *priv; 25786f00a022SMaxime Bizon int i; 25796f00a022SMaxime Bizon 25806f00a022SMaxime Bizon priv = netdev_priv(netdev); 25816f00a022SMaxime Bizon 25826f00a022SMaxime Bizon for (i = 0; i < BCM_ENETSW_STATS_LEN; i++) { 25836f00a022SMaxime Bizon const struct bcm_enet_stats *s; 25846f00a022SMaxime Bizon u32 lo, hi; 25856f00a022SMaxime Bizon char *p; 25866f00a022SMaxime Bizon int reg; 25876f00a022SMaxime Bizon 25886f00a022SMaxime Bizon s = &bcm_enetsw_gstrings_stats[i]; 25896f00a022SMaxime Bizon 25906f00a022SMaxime Bizon reg = s->mib_reg; 25916f00a022SMaxime Bizon if (reg == -1) 25926f00a022SMaxime Bizon continue; 25936f00a022SMaxime Bizon 25946f00a022SMaxime Bizon lo = enetsw_readl(priv, ENETSW_MIB_REG(reg)); 25956f00a022SMaxime Bizon p = (char *)priv + s->stat_offset; 25966f00a022SMaxime Bizon 25976f00a022SMaxime Bizon if (s->sizeof_stat == sizeof(u64)) { 25986f00a022SMaxime Bizon hi = enetsw_readl(priv, ENETSW_MIB_REG(reg + 1)); 25996f00a022SMaxime Bizon *(u64 *)p = ((u64)hi << 32 | lo); 26006f00a022SMaxime Bizon } else { 26016f00a022SMaxime Bizon *(u32 *)p = lo; 26026f00a022SMaxime Bizon } 26036f00a022SMaxime Bizon } 26046f00a022SMaxime Bizon 26056f00a022SMaxime Bizon for (i = 0; i < BCM_ENETSW_STATS_LEN; i++) { 26066f00a022SMaxime Bizon const struct bcm_enet_stats *s; 26076f00a022SMaxime Bizon char *p; 26086f00a022SMaxime Bizon 26096f00a022SMaxime Bizon s = &bcm_enetsw_gstrings_stats[i]; 26106f00a022SMaxime Bizon 26116f00a022SMaxime Bizon if (s->mib_reg == -1) 26126f00a022SMaxime Bizon p = (char *)&netdev->stats + s->stat_offset; 26136f00a022SMaxime Bizon else 26146f00a022SMaxime Bizon p = (char *)priv + s->stat_offset; 26156f00a022SMaxime Bizon 26166f00a022SMaxime Bizon data[i] = (s->sizeof_stat == sizeof(u64)) ? 26176f00a022SMaxime Bizon *(u64 *)p : *(u32 *)p; 26186f00a022SMaxime Bizon } 26196f00a022SMaxime Bizon } 26206f00a022SMaxime Bizon 26216f00a022SMaxime Bizon static void bcm_enetsw_get_ringparam(struct net_device *dev, 26226f00a022SMaxime Bizon struct ethtool_ringparam *ering) 26236f00a022SMaxime Bizon { 26246f00a022SMaxime Bizon struct bcm_enet_priv *priv; 26256f00a022SMaxime Bizon 26266f00a022SMaxime Bizon priv = netdev_priv(dev); 26276f00a022SMaxime Bizon 26286f00a022SMaxime Bizon /* rx/tx ring is actually only limited by memory */ 26296f00a022SMaxime Bizon ering->rx_max_pending = 8192; 26306f00a022SMaxime Bizon ering->tx_max_pending = 8192; 26316f00a022SMaxime Bizon ering->rx_mini_max_pending = 0; 26326f00a022SMaxime Bizon ering->rx_jumbo_max_pending = 0; 26336f00a022SMaxime Bizon ering->rx_pending = priv->rx_ring_size; 26346f00a022SMaxime Bizon ering->tx_pending = priv->tx_ring_size; 26356f00a022SMaxime Bizon } 26366f00a022SMaxime Bizon 26376f00a022SMaxime Bizon static int bcm_enetsw_set_ringparam(struct net_device *dev, 26386f00a022SMaxime Bizon struct ethtool_ringparam *ering) 26396f00a022SMaxime Bizon { 26406f00a022SMaxime Bizon struct bcm_enet_priv *priv; 26416f00a022SMaxime Bizon int was_running; 26426f00a022SMaxime Bizon 26436f00a022SMaxime Bizon priv = netdev_priv(dev); 26446f00a022SMaxime Bizon 26456f00a022SMaxime Bizon was_running = 0; 26466f00a022SMaxime Bizon if (netif_running(dev)) { 26476f00a022SMaxime Bizon bcm_enetsw_stop(dev); 26486f00a022SMaxime Bizon was_running = 1; 26496f00a022SMaxime Bizon } 26506f00a022SMaxime Bizon 26516f00a022SMaxime Bizon priv->rx_ring_size = ering->rx_pending; 26526f00a022SMaxime Bizon priv->tx_ring_size = ering->tx_pending; 26536f00a022SMaxime Bizon 26546f00a022SMaxime Bizon if (was_running) { 26556f00a022SMaxime Bizon int err; 26566f00a022SMaxime Bizon 26576f00a022SMaxime Bizon err = bcm_enetsw_open(dev); 26586f00a022SMaxime Bizon if (err) 26596f00a022SMaxime Bizon dev_close(dev); 26606f00a022SMaxime Bizon } 26616f00a022SMaxime Bizon return 0; 26626f00a022SMaxime Bizon } 26636f00a022SMaxime Bizon 26646f00a022SMaxime Bizon static struct ethtool_ops bcm_enetsw_ethtool_ops = { 26656f00a022SMaxime Bizon .get_strings = bcm_enetsw_get_strings, 26666f00a022SMaxime Bizon .get_sset_count = bcm_enetsw_get_sset_count, 26676f00a022SMaxime Bizon .get_ethtool_stats = bcm_enetsw_get_ethtool_stats, 26686f00a022SMaxime Bizon .get_drvinfo = bcm_enetsw_get_drvinfo, 26696f00a022SMaxime Bizon .get_ringparam = bcm_enetsw_get_ringparam, 26706f00a022SMaxime Bizon .set_ringparam = bcm_enetsw_set_ringparam, 26716f00a022SMaxime Bizon }; 26726f00a022SMaxime Bizon 26736f00a022SMaxime Bizon /* allocate netdevice, request register memory and register device. */ 26746f00a022SMaxime Bizon static int bcm_enetsw_probe(struct platform_device *pdev) 26756f00a022SMaxime Bizon { 26766f00a022SMaxime Bizon struct bcm_enet_priv *priv; 26776f00a022SMaxime Bizon struct net_device *dev; 26786f00a022SMaxime Bizon struct bcm63xx_enetsw_platform_data *pd; 26796f00a022SMaxime Bizon struct resource *res_mem; 26806f00a022SMaxime Bizon int ret, irq_rx, irq_tx; 26816f00a022SMaxime Bizon 26826f00a022SMaxime Bizon /* stop if shared driver failed, assume driver->probe will be 26836f00a022SMaxime Bizon * called in the same order we register devices (correct ?) 26846f00a022SMaxime Bizon */ 26856f00a022SMaxime Bizon if (!bcm_enet_shared_base[0]) 26866f00a022SMaxime Bizon return -ENODEV; 26876f00a022SMaxime Bizon 26886f00a022SMaxime Bizon res_mem = platform_get_resource(pdev, IORESOURCE_MEM, 0); 26896f00a022SMaxime Bizon irq_rx = platform_get_irq(pdev, 0); 26906f00a022SMaxime Bizon irq_tx = platform_get_irq(pdev, 1); 26916f00a022SMaxime Bizon if (!res_mem || irq_rx < 0) 26926f00a022SMaxime Bizon return -ENODEV; 26936f00a022SMaxime Bizon 26946f00a022SMaxime Bizon ret = 0; 26956f00a022SMaxime Bizon dev = alloc_etherdev(sizeof(*priv)); 26966f00a022SMaxime Bizon if (!dev) 26976f00a022SMaxime Bizon return -ENOMEM; 26986f00a022SMaxime Bizon priv = netdev_priv(dev); 26996f00a022SMaxime Bizon memset(priv, 0, sizeof(*priv)); 27006f00a022SMaxime Bizon 27016f00a022SMaxime Bizon /* initialize default and fetch platform data */ 27026f00a022SMaxime Bizon priv->enet_is_sw = true; 27036f00a022SMaxime Bizon priv->irq_rx = irq_rx; 27046f00a022SMaxime Bizon priv->irq_tx = irq_tx; 27056f00a022SMaxime Bizon priv->rx_ring_size = BCMENET_DEF_RX_DESC; 27066f00a022SMaxime Bizon priv->tx_ring_size = BCMENET_DEF_TX_DESC; 27076f00a022SMaxime Bizon priv->dma_maxburst = BCMENETSW_DMA_MAXBURST; 27086f00a022SMaxime Bizon 27096f00a022SMaxime Bizon pd = pdev->dev.platform_data; 27106f00a022SMaxime Bizon if (pd) { 27116f00a022SMaxime Bizon memcpy(dev->dev_addr, pd->mac_addr, ETH_ALEN); 27126f00a022SMaxime Bizon memcpy(priv->used_ports, pd->used_ports, 27136f00a022SMaxime Bizon sizeof(pd->used_ports)); 27146f00a022SMaxime Bizon priv->num_ports = pd->num_ports; 27156f00a022SMaxime Bizon } 27166f00a022SMaxime Bizon 27176f00a022SMaxime Bizon ret = compute_hw_mtu(priv, dev->mtu); 27186f00a022SMaxime Bizon if (ret) 27196f00a022SMaxime Bizon goto out; 27206f00a022SMaxime Bizon 27216f00a022SMaxime Bizon if (!request_mem_region(res_mem->start, resource_size(res_mem), 27226f00a022SMaxime Bizon "bcm63xx_enetsw")) { 27236f00a022SMaxime Bizon ret = -EBUSY; 27246f00a022SMaxime Bizon goto out; 27256f00a022SMaxime Bizon } 27266f00a022SMaxime Bizon 27276f00a022SMaxime Bizon priv->base = ioremap(res_mem->start, resource_size(res_mem)); 27286f00a022SMaxime Bizon if (priv->base == NULL) { 27296f00a022SMaxime Bizon ret = -ENOMEM; 27306f00a022SMaxime Bizon goto out_release_mem; 27316f00a022SMaxime Bizon } 27326f00a022SMaxime Bizon 27336f00a022SMaxime Bizon priv->mac_clk = clk_get(&pdev->dev, "enetsw"); 27346f00a022SMaxime Bizon if (IS_ERR(priv->mac_clk)) { 27356f00a022SMaxime Bizon ret = PTR_ERR(priv->mac_clk); 27366f00a022SMaxime Bizon goto out_unmap; 27376f00a022SMaxime Bizon } 27386f00a022SMaxime Bizon clk_enable(priv->mac_clk); 27396f00a022SMaxime Bizon 27406f00a022SMaxime Bizon priv->rx_chan = 0; 27416f00a022SMaxime Bizon priv->tx_chan = 1; 27426f00a022SMaxime Bizon spin_lock_init(&priv->rx_lock); 27436f00a022SMaxime Bizon 27446f00a022SMaxime Bizon /* init rx timeout (used for oom) */ 27456f00a022SMaxime Bizon init_timer(&priv->rx_timeout); 27466f00a022SMaxime Bizon priv->rx_timeout.function = bcm_enet_refill_rx_timer; 27476f00a022SMaxime Bizon priv->rx_timeout.data = (unsigned long)dev; 27486f00a022SMaxime Bizon 27496f00a022SMaxime Bizon /* register netdevice */ 27506f00a022SMaxime Bizon dev->netdev_ops = &bcm_enetsw_ops; 27516f00a022SMaxime Bizon netif_napi_add(dev, &priv->napi, bcm_enet_poll, 16); 27526f00a022SMaxime Bizon SET_ETHTOOL_OPS(dev, &bcm_enetsw_ethtool_ops); 27536f00a022SMaxime Bizon SET_NETDEV_DEV(dev, &pdev->dev); 27546f00a022SMaxime Bizon 27556f00a022SMaxime Bizon spin_lock_init(&priv->enetsw_mdio_lock); 27566f00a022SMaxime Bizon 27576f00a022SMaxime Bizon ret = register_netdev(dev); 27586f00a022SMaxime Bizon if (ret) 27596f00a022SMaxime Bizon goto out_put_clk; 27606f00a022SMaxime Bizon 27616f00a022SMaxime Bizon netif_carrier_off(dev); 27626f00a022SMaxime Bizon platform_set_drvdata(pdev, dev); 27636f00a022SMaxime Bizon priv->pdev = pdev; 27646f00a022SMaxime Bizon priv->net_dev = dev; 27656f00a022SMaxime Bizon 27666f00a022SMaxime Bizon return 0; 27676f00a022SMaxime Bizon 27686f00a022SMaxime Bizon out_put_clk: 27696f00a022SMaxime Bizon clk_put(priv->mac_clk); 27706f00a022SMaxime Bizon 27716f00a022SMaxime Bizon out_unmap: 27726f00a022SMaxime Bizon iounmap(priv->base); 27736f00a022SMaxime Bizon 27746f00a022SMaxime Bizon out_release_mem: 27756f00a022SMaxime Bizon release_mem_region(res_mem->start, resource_size(res_mem)); 27766f00a022SMaxime Bizon out: 27776f00a022SMaxime Bizon free_netdev(dev); 27786f00a022SMaxime Bizon return ret; 27796f00a022SMaxime Bizon } 27806f00a022SMaxime Bizon 27816f00a022SMaxime Bizon 27826f00a022SMaxime Bizon /* exit func, stops hardware and unregisters netdevice */ 27836f00a022SMaxime Bizon static int bcm_enetsw_remove(struct platform_device *pdev) 27846f00a022SMaxime Bizon { 27856f00a022SMaxime Bizon struct bcm_enet_priv *priv; 27866f00a022SMaxime Bizon struct net_device *dev; 27876f00a022SMaxime Bizon struct resource *res; 27886f00a022SMaxime Bizon 27896f00a022SMaxime Bizon /* stop netdevice */ 27906f00a022SMaxime Bizon dev = platform_get_drvdata(pdev); 27916f00a022SMaxime Bizon priv = netdev_priv(dev); 27926f00a022SMaxime Bizon unregister_netdev(dev); 27936f00a022SMaxime Bizon 27946f00a022SMaxime Bizon /* release device resources */ 27956f00a022SMaxime Bizon iounmap(priv->base); 27966f00a022SMaxime Bizon res = platform_get_resource(pdev, IORESOURCE_MEM, 0); 27976f00a022SMaxime Bizon release_mem_region(res->start, resource_size(res)); 27986f00a022SMaxime Bizon 27996f00a022SMaxime Bizon platform_set_drvdata(pdev, NULL); 28006f00a022SMaxime Bizon free_netdev(dev); 28016f00a022SMaxime Bizon return 0; 28026f00a022SMaxime Bizon } 28036f00a022SMaxime Bizon 28046f00a022SMaxime Bizon struct platform_driver bcm63xx_enetsw_driver = { 28056f00a022SMaxime Bizon .probe = bcm_enetsw_probe, 28066f00a022SMaxime Bizon .remove = bcm_enetsw_remove, 28076f00a022SMaxime Bizon .driver = { 28086f00a022SMaxime Bizon .name = "bcm63xx_enetsw", 28096f00a022SMaxime Bizon .owner = THIS_MODULE, 28106f00a022SMaxime Bizon }, 28116f00a022SMaxime Bizon }; 28126f00a022SMaxime Bizon 28136f00a022SMaxime Bizon /* reserve & remap memory space shared between all macs */ 2814047fc566SBill Pemberton static int bcm_enet_shared_probe(struct platform_device *pdev) 2815adfc5217SJeff Kirsher { 2816adfc5217SJeff Kirsher struct resource *res; 28170ae99b5fSMaxime Bizon void __iomem *p[3]; 28180ae99b5fSMaxime Bizon unsigned int i; 2819adfc5217SJeff Kirsher 28200ae99b5fSMaxime Bizon memset(bcm_enet_shared_base, 0, sizeof(bcm_enet_shared_base)); 2821adfc5217SJeff Kirsher 28220ae99b5fSMaxime Bizon for (i = 0; i < 3; i++) { 28230ae99b5fSMaxime Bizon res = platform_get_resource(pdev, IORESOURCE_MEM, i); 28240ae99b5fSMaxime Bizon p[i] = devm_ioremap_resource(&pdev->dev, res); 28250ae99b5fSMaxime Bizon if (!p[i]) 2826adfc5217SJeff Kirsher return -ENOMEM; 28270ae99b5fSMaxime Bizon } 28280ae99b5fSMaxime Bizon 28290ae99b5fSMaxime Bizon memcpy(bcm_enet_shared_base, p, sizeof(bcm_enet_shared_base)); 28301c03da05SJonas Gorski 2831adfc5217SJeff Kirsher return 0; 2832adfc5217SJeff Kirsher } 2833adfc5217SJeff Kirsher 2834047fc566SBill Pemberton static int bcm_enet_shared_remove(struct platform_device *pdev) 2835adfc5217SJeff Kirsher { 2836adfc5217SJeff Kirsher return 0; 2837adfc5217SJeff Kirsher } 2838adfc5217SJeff Kirsher 28396f00a022SMaxime Bizon /* this "shared" driver is needed because both macs share a single 2840adfc5217SJeff Kirsher * address space 2841adfc5217SJeff Kirsher */ 2842adfc5217SJeff Kirsher struct platform_driver bcm63xx_enet_shared_driver = { 2843adfc5217SJeff Kirsher .probe = bcm_enet_shared_probe, 2844047fc566SBill Pemberton .remove = bcm_enet_shared_remove, 2845adfc5217SJeff Kirsher .driver = { 2846adfc5217SJeff Kirsher .name = "bcm63xx_enet_shared", 2847adfc5217SJeff Kirsher .owner = THIS_MODULE, 2848adfc5217SJeff Kirsher }, 2849adfc5217SJeff Kirsher }; 2850adfc5217SJeff Kirsher 28516f00a022SMaxime Bizon /* entry point */ 2852adfc5217SJeff Kirsher static int __init bcm_enet_init(void) 2853adfc5217SJeff Kirsher { 2854adfc5217SJeff Kirsher int ret; 2855adfc5217SJeff Kirsher 2856adfc5217SJeff Kirsher ret = platform_driver_register(&bcm63xx_enet_shared_driver); 2857adfc5217SJeff Kirsher if (ret) 2858adfc5217SJeff Kirsher return ret; 2859adfc5217SJeff Kirsher 2860adfc5217SJeff Kirsher ret = platform_driver_register(&bcm63xx_enet_driver); 2861adfc5217SJeff Kirsher if (ret) 2862adfc5217SJeff Kirsher platform_driver_unregister(&bcm63xx_enet_shared_driver); 2863adfc5217SJeff Kirsher 28646f00a022SMaxime Bizon ret = platform_driver_register(&bcm63xx_enetsw_driver); 28656f00a022SMaxime Bizon if (ret) { 28666f00a022SMaxime Bizon platform_driver_unregister(&bcm63xx_enet_driver); 28676f00a022SMaxime Bizon platform_driver_unregister(&bcm63xx_enet_shared_driver); 28686f00a022SMaxime Bizon } 28696f00a022SMaxime Bizon 2870adfc5217SJeff Kirsher return ret; 2871adfc5217SJeff Kirsher } 2872adfc5217SJeff Kirsher 2873adfc5217SJeff Kirsher static void __exit bcm_enet_exit(void) 2874adfc5217SJeff Kirsher { 2875adfc5217SJeff Kirsher platform_driver_unregister(&bcm63xx_enet_driver); 28766f00a022SMaxime Bizon platform_driver_unregister(&bcm63xx_enetsw_driver); 2877adfc5217SJeff Kirsher platform_driver_unregister(&bcm63xx_enet_shared_driver); 2878adfc5217SJeff Kirsher } 2879adfc5217SJeff Kirsher 2880adfc5217SJeff Kirsher 2881adfc5217SJeff Kirsher module_init(bcm_enet_init); 2882adfc5217SJeff Kirsher module_exit(bcm_enet_exit); 2883adfc5217SJeff Kirsher 2884adfc5217SJeff Kirsher MODULE_DESCRIPTION("BCM63xx internal ethernet mac driver"); 2885adfc5217SJeff Kirsher MODULE_AUTHOR("Maxime Bizon <mbizon@freebox.fr>"); 2886adfc5217SJeff Kirsher MODULE_LICENSE("GPL"); 2887