1*527a6266SJeff Kirsher /* 2*527a6266SJeff Kirsher * New driver for Marvell Yukon 2 chipset. 3*527a6266SJeff Kirsher * Based on earlier sk98lin, and skge driver. 4*527a6266SJeff Kirsher * 5*527a6266SJeff Kirsher * This driver intentionally does not support all the features 6*527a6266SJeff Kirsher * of the original driver such as link fail-over and link management because 7*527a6266SJeff Kirsher * those should be done at higher levels. 8*527a6266SJeff Kirsher * 9*527a6266SJeff Kirsher * Copyright (C) 2005 Stephen Hemminger <shemminger@osdl.org> 10*527a6266SJeff Kirsher * 11*527a6266SJeff Kirsher * This program is free software; you can redistribute it and/or modify 12*527a6266SJeff Kirsher * it under the terms of the GNU General Public License as published by 13*527a6266SJeff Kirsher * the Free Software Foundation; either version 2 of the License. 14*527a6266SJeff Kirsher * 15*527a6266SJeff Kirsher * This program is distributed in the hope that it will be useful, 16*527a6266SJeff Kirsher * but WITHOUT ANY WARRANTY; without even the implied warranty of 17*527a6266SJeff Kirsher * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the 18*527a6266SJeff Kirsher * GNU General Public License for more details. 19*527a6266SJeff Kirsher * 20*527a6266SJeff Kirsher * You should have received a copy of the GNU General Public License 21*527a6266SJeff Kirsher * along with this program; if not, write to the Free Software 22*527a6266SJeff Kirsher * Foundation, Inc., 675 Mass Ave, Cambridge, MA 02139, USA. 23*527a6266SJeff Kirsher */ 24*527a6266SJeff Kirsher 25*527a6266SJeff Kirsher #define pr_fmt(fmt) KBUILD_MODNAME ": " fmt 26*527a6266SJeff Kirsher 27*527a6266SJeff Kirsher #include <linux/crc32.h> 28*527a6266SJeff Kirsher #include <linux/kernel.h> 29*527a6266SJeff Kirsher #include <linux/module.h> 30*527a6266SJeff Kirsher #include <linux/netdevice.h> 31*527a6266SJeff Kirsher #include <linux/dma-mapping.h> 32*527a6266SJeff Kirsher #include <linux/etherdevice.h> 33*527a6266SJeff Kirsher #include <linux/ethtool.h> 34*527a6266SJeff Kirsher #include <linux/pci.h> 35*527a6266SJeff Kirsher #include <linux/interrupt.h> 36*527a6266SJeff Kirsher #include <linux/ip.h> 37*527a6266SJeff Kirsher #include <linux/slab.h> 38*527a6266SJeff Kirsher #include <net/ip.h> 39*527a6266SJeff Kirsher #include <linux/tcp.h> 40*527a6266SJeff Kirsher #include <linux/in.h> 41*527a6266SJeff Kirsher #include <linux/delay.h> 42*527a6266SJeff Kirsher #include <linux/workqueue.h> 43*527a6266SJeff Kirsher #include <linux/if_vlan.h> 44*527a6266SJeff Kirsher #include <linux/prefetch.h> 45*527a6266SJeff Kirsher #include <linux/debugfs.h> 46*527a6266SJeff Kirsher #include <linux/mii.h> 47*527a6266SJeff Kirsher 48*527a6266SJeff Kirsher #include <asm/irq.h> 49*527a6266SJeff Kirsher 50*527a6266SJeff Kirsher #include "sky2.h" 51*527a6266SJeff Kirsher 52*527a6266SJeff Kirsher #define DRV_NAME "sky2" 53*527a6266SJeff Kirsher #define DRV_VERSION "1.29" 54*527a6266SJeff Kirsher 55*527a6266SJeff Kirsher /* 56*527a6266SJeff Kirsher * The Yukon II chipset takes 64 bit command blocks (called list elements) 57*527a6266SJeff Kirsher * that are organized into three (receive, transmit, status) different rings 58*527a6266SJeff Kirsher * similar to Tigon3. 59*527a6266SJeff Kirsher */ 60*527a6266SJeff Kirsher 61*527a6266SJeff Kirsher #define RX_LE_SIZE 1024 62*527a6266SJeff Kirsher #define RX_LE_BYTES (RX_LE_SIZE*sizeof(struct sky2_rx_le)) 63*527a6266SJeff Kirsher #define RX_MAX_PENDING (RX_LE_SIZE/6 - 2) 64*527a6266SJeff Kirsher #define RX_DEF_PENDING RX_MAX_PENDING 65*527a6266SJeff Kirsher 66*527a6266SJeff Kirsher /* This is the worst case number of transmit list elements for a single skb: 67*527a6266SJeff Kirsher VLAN:GSO + CKSUM + Data + skb_frags * DMA */ 68*527a6266SJeff Kirsher #define MAX_SKB_TX_LE (2 + (sizeof(dma_addr_t)/sizeof(u32))*(MAX_SKB_FRAGS+1)) 69*527a6266SJeff Kirsher #define TX_MIN_PENDING (MAX_SKB_TX_LE+1) 70*527a6266SJeff Kirsher #define TX_MAX_PENDING 1024 71*527a6266SJeff Kirsher #define TX_DEF_PENDING 127 72*527a6266SJeff Kirsher 73*527a6266SJeff Kirsher #define TX_WATCHDOG (5 * HZ) 74*527a6266SJeff Kirsher #define NAPI_WEIGHT 64 75*527a6266SJeff Kirsher #define PHY_RETRIES 1000 76*527a6266SJeff Kirsher 77*527a6266SJeff Kirsher #define SKY2_EEPROM_MAGIC 0x9955aabb 78*527a6266SJeff Kirsher 79*527a6266SJeff Kirsher #define RING_NEXT(x, s) (((x)+1) & ((s)-1)) 80*527a6266SJeff Kirsher 81*527a6266SJeff Kirsher static const u32 default_msg = 82*527a6266SJeff Kirsher NETIF_MSG_DRV | NETIF_MSG_PROBE | NETIF_MSG_LINK 83*527a6266SJeff Kirsher | NETIF_MSG_TIMER | NETIF_MSG_TX_ERR | NETIF_MSG_RX_ERR 84*527a6266SJeff Kirsher | NETIF_MSG_IFUP | NETIF_MSG_IFDOWN; 85*527a6266SJeff Kirsher 86*527a6266SJeff Kirsher static int debug = -1; /* defaults above */ 87*527a6266SJeff Kirsher module_param(debug, int, 0); 88*527a6266SJeff Kirsher MODULE_PARM_DESC(debug, "Debug level (0=none,...,16=all)"); 89*527a6266SJeff Kirsher 90*527a6266SJeff Kirsher static int copybreak __read_mostly = 128; 91*527a6266SJeff Kirsher module_param(copybreak, int, 0); 92*527a6266SJeff Kirsher MODULE_PARM_DESC(copybreak, "Receive copy threshold"); 93*527a6266SJeff Kirsher 94*527a6266SJeff Kirsher static int disable_msi = 0; 95*527a6266SJeff Kirsher module_param(disable_msi, int, 0); 96*527a6266SJeff Kirsher MODULE_PARM_DESC(disable_msi, "Disable Message Signaled Interrupt (MSI)"); 97*527a6266SJeff Kirsher 98*527a6266SJeff Kirsher static DEFINE_PCI_DEVICE_TABLE(sky2_id_table) = { 99*527a6266SJeff Kirsher { PCI_DEVICE(PCI_VENDOR_ID_SYSKONNECT, 0x9000) }, /* SK-9Sxx */ 100*527a6266SJeff Kirsher { PCI_DEVICE(PCI_VENDOR_ID_SYSKONNECT, 0x9E00) }, /* SK-9Exx */ 101*527a6266SJeff Kirsher { PCI_DEVICE(PCI_VENDOR_ID_SYSKONNECT, 0x9E01) }, /* SK-9E21M */ 102*527a6266SJeff Kirsher { PCI_DEVICE(PCI_VENDOR_ID_DLINK, 0x4b00) }, /* DGE-560T */ 103*527a6266SJeff Kirsher { PCI_DEVICE(PCI_VENDOR_ID_DLINK, 0x4001) }, /* DGE-550SX */ 104*527a6266SJeff Kirsher { PCI_DEVICE(PCI_VENDOR_ID_DLINK, 0x4B02) }, /* DGE-560SX */ 105*527a6266SJeff Kirsher { PCI_DEVICE(PCI_VENDOR_ID_DLINK, 0x4B03) }, /* DGE-550T */ 106*527a6266SJeff Kirsher { PCI_DEVICE(PCI_VENDOR_ID_MARVELL, 0x4340) }, /* 88E8021 */ 107*527a6266SJeff Kirsher { PCI_DEVICE(PCI_VENDOR_ID_MARVELL, 0x4341) }, /* 88E8022 */ 108*527a6266SJeff Kirsher { PCI_DEVICE(PCI_VENDOR_ID_MARVELL, 0x4342) }, /* 88E8061 */ 109*527a6266SJeff Kirsher { PCI_DEVICE(PCI_VENDOR_ID_MARVELL, 0x4343) }, /* 88E8062 */ 110*527a6266SJeff Kirsher { PCI_DEVICE(PCI_VENDOR_ID_MARVELL, 0x4344) }, /* 88E8021 */ 111*527a6266SJeff Kirsher { PCI_DEVICE(PCI_VENDOR_ID_MARVELL, 0x4345) }, /* 88E8022 */ 112*527a6266SJeff Kirsher { PCI_DEVICE(PCI_VENDOR_ID_MARVELL, 0x4346) }, /* 88E8061 */ 113*527a6266SJeff Kirsher { PCI_DEVICE(PCI_VENDOR_ID_MARVELL, 0x4347) }, /* 88E8062 */ 114*527a6266SJeff Kirsher { PCI_DEVICE(PCI_VENDOR_ID_MARVELL, 0x4350) }, /* 88E8035 */ 115*527a6266SJeff Kirsher { PCI_DEVICE(PCI_VENDOR_ID_MARVELL, 0x4351) }, /* 88E8036 */ 116*527a6266SJeff Kirsher { PCI_DEVICE(PCI_VENDOR_ID_MARVELL, 0x4352) }, /* 88E8038 */ 117*527a6266SJeff Kirsher { PCI_DEVICE(PCI_VENDOR_ID_MARVELL, 0x4353) }, /* 88E8039 */ 118*527a6266SJeff Kirsher { PCI_DEVICE(PCI_VENDOR_ID_MARVELL, 0x4354) }, /* 88E8040 */ 119*527a6266SJeff Kirsher { PCI_DEVICE(PCI_VENDOR_ID_MARVELL, 0x4355) }, /* 88E8040T */ 120*527a6266SJeff Kirsher { PCI_DEVICE(PCI_VENDOR_ID_MARVELL, 0x4356) }, /* 88EC033 */ 121*527a6266SJeff Kirsher { PCI_DEVICE(PCI_VENDOR_ID_MARVELL, 0x4357) }, /* 88E8042 */ 122*527a6266SJeff Kirsher { PCI_DEVICE(PCI_VENDOR_ID_MARVELL, 0x435A) }, /* 88E8048 */ 123*527a6266SJeff Kirsher { PCI_DEVICE(PCI_VENDOR_ID_MARVELL, 0x4360) }, /* 88E8052 */ 124*527a6266SJeff Kirsher { PCI_DEVICE(PCI_VENDOR_ID_MARVELL, 0x4361) }, /* 88E8050 */ 125*527a6266SJeff Kirsher { PCI_DEVICE(PCI_VENDOR_ID_MARVELL, 0x4362) }, /* 88E8053 */ 126*527a6266SJeff Kirsher { PCI_DEVICE(PCI_VENDOR_ID_MARVELL, 0x4363) }, /* 88E8055 */ 127*527a6266SJeff Kirsher { PCI_DEVICE(PCI_VENDOR_ID_MARVELL, 0x4364) }, /* 88E8056 */ 128*527a6266SJeff Kirsher { PCI_DEVICE(PCI_VENDOR_ID_MARVELL, 0x4365) }, /* 88E8070 */ 129*527a6266SJeff Kirsher { PCI_DEVICE(PCI_VENDOR_ID_MARVELL, 0x4366) }, /* 88EC036 */ 130*527a6266SJeff Kirsher { PCI_DEVICE(PCI_VENDOR_ID_MARVELL, 0x4367) }, /* 88EC032 */ 131*527a6266SJeff Kirsher { PCI_DEVICE(PCI_VENDOR_ID_MARVELL, 0x4368) }, /* 88EC034 */ 132*527a6266SJeff Kirsher { PCI_DEVICE(PCI_VENDOR_ID_MARVELL, 0x4369) }, /* 88EC042 */ 133*527a6266SJeff Kirsher { PCI_DEVICE(PCI_VENDOR_ID_MARVELL, 0x436A) }, /* 88E8058 */ 134*527a6266SJeff Kirsher { PCI_DEVICE(PCI_VENDOR_ID_MARVELL, 0x436B) }, /* 88E8071 */ 135*527a6266SJeff Kirsher { PCI_DEVICE(PCI_VENDOR_ID_MARVELL, 0x436C) }, /* 88E8072 */ 136*527a6266SJeff Kirsher { PCI_DEVICE(PCI_VENDOR_ID_MARVELL, 0x436D) }, /* 88E8055 */ 137*527a6266SJeff Kirsher { PCI_DEVICE(PCI_VENDOR_ID_MARVELL, 0x4370) }, /* 88E8075 */ 138*527a6266SJeff Kirsher { PCI_DEVICE(PCI_VENDOR_ID_MARVELL, 0x4380) }, /* 88E8057 */ 139*527a6266SJeff Kirsher { PCI_DEVICE(PCI_VENDOR_ID_MARVELL, 0x4381) }, /* 88E8059 */ 140*527a6266SJeff Kirsher { 0 } 141*527a6266SJeff Kirsher }; 142*527a6266SJeff Kirsher 143*527a6266SJeff Kirsher MODULE_DEVICE_TABLE(pci, sky2_id_table); 144*527a6266SJeff Kirsher 145*527a6266SJeff Kirsher /* Avoid conditionals by using array */ 146*527a6266SJeff Kirsher static const unsigned txqaddr[] = { Q_XA1, Q_XA2 }; 147*527a6266SJeff Kirsher static const unsigned rxqaddr[] = { Q_R1, Q_R2 }; 148*527a6266SJeff Kirsher static const u32 portirq_msk[] = { Y2_IS_PORT_1, Y2_IS_PORT_2 }; 149*527a6266SJeff Kirsher 150*527a6266SJeff Kirsher static void sky2_set_multicast(struct net_device *dev); 151*527a6266SJeff Kirsher 152*527a6266SJeff Kirsher /* Access to PHY via serial interconnect */ 153*527a6266SJeff Kirsher static int gm_phy_write(struct sky2_hw *hw, unsigned port, u16 reg, u16 val) 154*527a6266SJeff Kirsher { 155*527a6266SJeff Kirsher int i; 156*527a6266SJeff Kirsher 157*527a6266SJeff Kirsher gma_write16(hw, port, GM_SMI_DATA, val); 158*527a6266SJeff Kirsher gma_write16(hw, port, GM_SMI_CTRL, 159*527a6266SJeff Kirsher GM_SMI_CT_PHY_AD(PHY_ADDR_MARV) | GM_SMI_CT_REG_AD(reg)); 160*527a6266SJeff Kirsher 161*527a6266SJeff Kirsher for (i = 0; i < PHY_RETRIES; i++) { 162*527a6266SJeff Kirsher u16 ctrl = gma_read16(hw, port, GM_SMI_CTRL); 163*527a6266SJeff Kirsher if (ctrl == 0xffff) 164*527a6266SJeff Kirsher goto io_error; 165*527a6266SJeff Kirsher 166*527a6266SJeff Kirsher if (!(ctrl & GM_SMI_CT_BUSY)) 167*527a6266SJeff Kirsher return 0; 168*527a6266SJeff Kirsher 169*527a6266SJeff Kirsher udelay(10); 170*527a6266SJeff Kirsher } 171*527a6266SJeff Kirsher 172*527a6266SJeff Kirsher dev_warn(&hw->pdev->dev, "%s: phy write timeout\n", hw->dev[port]->name); 173*527a6266SJeff Kirsher return -ETIMEDOUT; 174*527a6266SJeff Kirsher 175*527a6266SJeff Kirsher io_error: 176*527a6266SJeff Kirsher dev_err(&hw->pdev->dev, "%s: phy I/O error\n", hw->dev[port]->name); 177*527a6266SJeff Kirsher return -EIO; 178*527a6266SJeff Kirsher } 179*527a6266SJeff Kirsher 180*527a6266SJeff Kirsher static int __gm_phy_read(struct sky2_hw *hw, unsigned port, u16 reg, u16 *val) 181*527a6266SJeff Kirsher { 182*527a6266SJeff Kirsher int i; 183*527a6266SJeff Kirsher 184*527a6266SJeff Kirsher gma_write16(hw, port, GM_SMI_CTRL, GM_SMI_CT_PHY_AD(PHY_ADDR_MARV) 185*527a6266SJeff Kirsher | GM_SMI_CT_REG_AD(reg) | GM_SMI_CT_OP_RD); 186*527a6266SJeff Kirsher 187*527a6266SJeff Kirsher for (i = 0; i < PHY_RETRIES; i++) { 188*527a6266SJeff Kirsher u16 ctrl = gma_read16(hw, port, GM_SMI_CTRL); 189*527a6266SJeff Kirsher if (ctrl == 0xffff) 190*527a6266SJeff Kirsher goto io_error; 191*527a6266SJeff Kirsher 192*527a6266SJeff Kirsher if (ctrl & GM_SMI_CT_RD_VAL) { 193*527a6266SJeff Kirsher *val = gma_read16(hw, port, GM_SMI_DATA); 194*527a6266SJeff Kirsher return 0; 195*527a6266SJeff Kirsher } 196*527a6266SJeff Kirsher 197*527a6266SJeff Kirsher udelay(10); 198*527a6266SJeff Kirsher } 199*527a6266SJeff Kirsher 200*527a6266SJeff Kirsher dev_warn(&hw->pdev->dev, "%s: phy read timeout\n", hw->dev[port]->name); 201*527a6266SJeff Kirsher return -ETIMEDOUT; 202*527a6266SJeff Kirsher io_error: 203*527a6266SJeff Kirsher dev_err(&hw->pdev->dev, "%s: phy I/O error\n", hw->dev[port]->name); 204*527a6266SJeff Kirsher return -EIO; 205*527a6266SJeff Kirsher } 206*527a6266SJeff Kirsher 207*527a6266SJeff Kirsher static inline u16 gm_phy_read(struct sky2_hw *hw, unsigned port, u16 reg) 208*527a6266SJeff Kirsher { 209*527a6266SJeff Kirsher u16 v; 210*527a6266SJeff Kirsher __gm_phy_read(hw, port, reg, &v); 211*527a6266SJeff Kirsher return v; 212*527a6266SJeff Kirsher } 213*527a6266SJeff Kirsher 214*527a6266SJeff Kirsher 215*527a6266SJeff Kirsher static void sky2_power_on(struct sky2_hw *hw) 216*527a6266SJeff Kirsher { 217*527a6266SJeff Kirsher /* switch power to VCC (WA for VAUX problem) */ 218*527a6266SJeff Kirsher sky2_write8(hw, B0_POWER_CTRL, 219*527a6266SJeff Kirsher PC_VAUX_ENA | PC_VCC_ENA | PC_VAUX_OFF | PC_VCC_ON); 220*527a6266SJeff Kirsher 221*527a6266SJeff Kirsher /* disable Core Clock Division, */ 222*527a6266SJeff Kirsher sky2_write32(hw, B2_Y2_CLK_CTRL, Y2_CLK_DIV_DIS); 223*527a6266SJeff Kirsher 224*527a6266SJeff Kirsher if (hw->chip_id == CHIP_ID_YUKON_XL && hw->chip_rev > CHIP_REV_YU_XL_A1) 225*527a6266SJeff Kirsher /* enable bits are inverted */ 226*527a6266SJeff Kirsher sky2_write8(hw, B2_Y2_CLK_GATE, 227*527a6266SJeff Kirsher Y2_PCI_CLK_LNK1_DIS | Y2_COR_CLK_LNK1_DIS | 228*527a6266SJeff Kirsher Y2_CLK_GAT_LNK1_DIS | Y2_PCI_CLK_LNK2_DIS | 229*527a6266SJeff Kirsher Y2_COR_CLK_LNK2_DIS | Y2_CLK_GAT_LNK2_DIS); 230*527a6266SJeff Kirsher else 231*527a6266SJeff Kirsher sky2_write8(hw, B2_Y2_CLK_GATE, 0); 232*527a6266SJeff Kirsher 233*527a6266SJeff Kirsher if (hw->flags & SKY2_HW_ADV_POWER_CTL) { 234*527a6266SJeff Kirsher u32 reg; 235*527a6266SJeff Kirsher 236*527a6266SJeff Kirsher sky2_pci_write32(hw, PCI_DEV_REG3, 0); 237*527a6266SJeff Kirsher 238*527a6266SJeff Kirsher reg = sky2_pci_read32(hw, PCI_DEV_REG4); 239*527a6266SJeff Kirsher /* set all bits to 0 except bits 15..12 and 8 */ 240*527a6266SJeff Kirsher reg &= P_ASPM_CONTROL_MSK; 241*527a6266SJeff Kirsher sky2_pci_write32(hw, PCI_DEV_REG4, reg); 242*527a6266SJeff Kirsher 243*527a6266SJeff Kirsher reg = sky2_pci_read32(hw, PCI_DEV_REG5); 244*527a6266SJeff Kirsher /* set all bits to 0 except bits 28 & 27 */ 245*527a6266SJeff Kirsher reg &= P_CTL_TIM_VMAIN_AV_MSK; 246*527a6266SJeff Kirsher sky2_pci_write32(hw, PCI_DEV_REG5, reg); 247*527a6266SJeff Kirsher 248*527a6266SJeff Kirsher sky2_pci_write32(hw, PCI_CFG_REG_1, 0); 249*527a6266SJeff Kirsher 250*527a6266SJeff Kirsher sky2_write16(hw, B0_CTST, Y2_HW_WOL_ON); 251*527a6266SJeff Kirsher 252*527a6266SJeff Kirsher /* Enable workaround for dev 4.107 on Yukon-Ultra & Extreme */ 253*527a6266SJeff Kirsher reg = sky2_read32(hw, B2_GP_IO); 254*527a6266SJeff Kirsher reg |= GLB_GPIO_STAT_RACE_DIS; 255*527a6266SJeff Kirsher sky2_write32(hw, B2_GP_IO, reg); 256*527a6266SJeff Kirsher 257*527a6266SJeff Kirsher sky2_read32(hw, B2_GP_IO); 258*527a6266SJeff Kirsher } 259*527a6266SJeff Kirsher 260*527a6266SJeff Kirsher /* Turn on "driver loaded" LED */ 261*527a6266SJeff Kirsher sky2_write16(hw, B0_CTST, Y2_LED_STAT_ON); 262*527a6266SJeff Kirsher } 263*527a6266SJeff Kirsher 264*527a6266SJeff Kirsher static void sky2_power_aux(struct sky2_hw *hw) 265*527a6266SJeff Kirsher { 266*527a6266SJeff Kirsher if (hw->chip_id == CHIP_ID_YUKON_XL && hw->chip_rev > CHIP_REV_YU_XL_A1) 267*527a6266SJeff Kirsher sky2_write8(hw, B2_Y2_CLK_GATE, 0); 268*527a6266SJeff Kirsher else 269*527a6266SJeff Kirsher /* enable bits are inverted */ 270*527a6266SJeff Kirsher sky2_write8(hw, B2_Y2_CLK_GATE, 271*527a6266SJeff Kirsher Y2_PCI_CLK_LNK1_DIS | Y2_COR_CLK_LNK1_DIS | 272*527a6266SJeff Kirsher Y2_CLK_GAT_LNK1_DIS | Y2_PCI_CLK_LNK2_DIS | 273*527a6266SJeff Kirsher Y2_COR_CLK_LNK2_DIS | Y2_CLK_GAT_LNK2_DIS); 274*527a6266SJeff Kirsher 275*527a6266SJeff Kirsher /* switch power to VAUX if supported and PME from D3cold */ 276*527a6266SJeff Kirsher if ( (sky2_read32(hw, B0_CTST) & Y2_VAUX_AVAIL) && 277*527a6266SJeff Kirsher pci_pme_capable(hw->pdev, PCI_D3cold)) 278*527a6266SJeff Kirsher sky2_write8(hw, B0_POWER_CTRL, 279*527a6266SJeff Kirsher (PC_VAUX_ENA | PC_VCC_ENA | 280*527a6266SJeff Kirsher PC_VAUX_ON | PC_VCC_OFF)); 281*527a6266SJeff Kirsher 282*527a6266SJeff Kirsher /* turn off "driver loaded LED" */ 283*527a6266SJeff Kirsher sky2_write16(hw, B0_CTST, Y2_LED_STAT_OFF); 284*527a6266SJeff Kirsher } 285*527a6266SJeff Kirsher 286*527a6266SJeff Kirsher static void sky2_gmac_reset(struct sky2_hw *hw, unsigned port) 287*527a6266SJeff Kirsher { 288*527a6266SJeff Kirsher u16 reg; 289*527a6266SJeff Kirsher 290*527a6266SJeff Kirsher /* disable all GMAC IRQ's */ 291*527a6266SJeff Kirsher sky2_write8(hw, SK_REG(port, GMAC_IRQ_MSK), 0); 292*527a6266SJeff Kirsher 293*527a6266SJeff Kirsher gma_write16(hw, port, GM_MC_ADDR_H1, 0); /* clear MC hash */ 294*527a6266SJeff Kirsher gma_write16(hw, port, GM_MC_ADDR_H2, 0); 295*527a6266SJeff Kirsher gma_write16(hw, port, GM_MC_ADDR_H3, 0); 296*527a6266SJeff Kirsher gma_write16(hw, port, GM_MC_ADDR_H4, 0); 297*527a6266SJeff Kirsher 298*527a6266SJeff Kirsher reg = gma_read16(hw, port, GM_RX_CTRL); 299*527a6266SJeff Kirsher reg |= GM_RXCR_UCF_ENA | GM_RXCR_MCF_ENA; 300*527a6266SJeff Kirsher gma_write16(hw, port, GM_RX_CTRL, reg); 301*527a6266SJeff Kirsher } 302*527a6266SJeff Kirsher 303*527a6266SJeff Kirsher /* flow control to advertise bits */ 304*527a6266SJeff Kirsher static const u16 copper_fc_adv[] = { 305*527a6266SJeff Kirsher [FC_NONE] = 0, 306*527a6266SJeff Kirsher [FC_TX] = PHY_M_AN_ASP, 307*527a6266SJeff Kirsher [FC_RX] = PHY_M_AN_PC, 308*527a6266SJeff Kirsher [FC_BOTH] = PHY_M_AN_PC | PHY_M_AN_ASP, 309*527a6266SJeff Kirsher }; 310*527a6266SJeff Kirsher 311*527a6266SJeff Kirsher /* flow control to advertise bits when using 1000BaseX */ 312*527a6266SJeff Kirsher static const u16 fiber_fc_adv[] = { 313*527a6266SJeff Kirsher [FC_NONE] = PHY_M_P_NO_PAUSE_X, 314*527a6266SJeff Kirsher [FC_TX] = PHY_M_P_ASYM_MD_X, 315*527a6266SJeff Kirsher [FC_RX] = PHY_M_P_SYM_MD_X, 316*527a6266SJeff Kirsher [FC_BOTH] = PHY_M_P_BOTH_MD_X, 317*527a6266SJeff Kirsher }; 318*527a6266SJeff Kirsher 319*527a6266SJeff Kirsher /* flow control to GMA disable bits */ 320*527a6266SJeff Kirsher static const u16 gm_fc_disable[] = { 321*527a6266SJeff Kirsher [FC_NONE] = GM_GPCR_FC_RX_DIS | GM_GPCR_FC_TX_DIS, 322*527a6266SJeff Kirsher [FC_TX] = GM_GPCR_FC_RX_DIS, 323*527a6266SJeff Kirsher [FC_RX] = GM_GPCR_FC_TX_DIS, 324*527a6266SJeff Kirsher [FC_BOTH] = 0, 325*527a6266SJeff Kirsher }; 326*527a6266SJeff Kirsher 327*527a6266SJeff Kirsher 328*527a6266SJeff Kirsher static void sky2_phy_init(struct sky2_hw *hw, unsigned port) 329*527a6266SJeff Kirsher { 330*527a6266SJeff Kirsher struct sky2_port *sky2 = netdev_priv(hw->dev[port]); 331*527a6266SJeff Kirsher u16 ctrl, ct1000, adv, pg, ledctrl, ledover, reg; 332*527a6266SJeff Kirsher 333*527a6266SJeff Kirsher if ( (sky2->flags & SKY2_FLAG_AUTO_SPEED) && 334*527a6266SJeff Kirsher !(hw->flags & SKY2_HW_NEWER_PHY)) { 335*527a6266SJeff Kirsher u16 ectrl = gm_phy_read(hw, port, PHY_MARV_EXT_CTRL); 336*527a6266SJeff Kirsher 337*527a6266SJeff Kirsher ectrl &= ~(PHY_M_EC_M_DSC_MSK | PHY_M_EC_S_DSC_MSK | 338*527a6266SJeff Kirsher PHY_M_EC_MAC_S_MSK); 339*527a6266SJeff Kirsher ectrl |= PHY_M_EC_MAC_S(MAC_TX_CLK_25_MHZ); 340*527a6266SJeff Kirsher 341*527a6266SJeff Kirsher /* on PHY 88E1040 Rev.D0 (and newer) downshift control changed */ 342*527a6266SJeff Kirsher if (hw->chip_id == CHIP_ID_YUKON_EC) 343*527a6266SJeff Kirsher /* set downshift counter to 3x and enable downshift */ 344*527a6266SJeff Kirsher ectrl |= PHY_M_EC_DSC_2(2) | PHY_M_EC_DOWN_S_ENA; 345*527a6266SJeff Kirsher else 346*527a6266SJeff Kirsher /* set master & slave downshift counter to 1x */ 347*527a6266SJeff Kirsher ectrl |= PHY_M_EC_M_DSC(0) | PHY_M_EC_S_DSC(1); 348*527a6266SJeff Kirsher 349*527a6266SJeff Kirsher gm_phy_write(hw, port, PHY_MARV_EXT_CTRL, ectrl); 350*527a6266SJeff Kirsher } 351*527a6266SJeff Kirsher 352*527a6266SJeff Kirsher ctrl = gm_phy_read(hw, port, PHY_MARV_PHY_CTRL); 353*527a6266SJeff Kirsher if (sky2_is_copper(hw)) { 354*527a6266SJeff Kirsher if (!(hw->flags & SKY2_HW_GIGABIT)) { 355*527a6266SJeff Kirsher /* enable automatic crossover */ 356*527a6266SJeff Kirsher ctrl |= PHY_M_PC_MDI_XMODE(PHY_M_PC_ENA_AUTO) >> 1; 357*527a6266SJeff Kirsher 358*527a6266SJeff Kirsher if (hw->chip_id == CHIP_ID_YUKON_FE_P && 359*527a6266SJeff Kirsher hw->chip_rev == CHIP_REV_YU_FE2_A0) { 360*527a6266SJeff Kirsher u16 spec; 361*527a6266SJeff Kirsher 362*527a6266SJeff Kirsher /* Enable Class A driver for FE+ A0 */ 363*527a6266SJeff Kirsher spec = gm_phy_read(hw, port, PHY_MARV_FE_SPEC_2); 364*527a6266SJeff Kirsher spec |= PHY_M_FESC_SEL_CL_A; 365*527a6266SJeff Kirsher gm_phy_write(hw, port, PHY_MARV_FE_SPEC_2, spec); 366*527a6266SJeff Kirsher } 367*527a6266SJeff Kirsher } else { 368*527a6266SJeff Kirsher if (hw->chip_id >= CHIP_ID_YUKON_OPT) { 369*527a6266SJeff Kirsher u16 ctrl2 = gm_phy_read(hw, port, PHY_MARV_EXT_CTRL_2); 370*527a6266SJeff Kirsher 371*527a6266SJeff Kirsher /* enable PHY Reverse Auto-Negotiation */ 372*527a6266SJeff Kirsher ctrl2 |= 1u << 13; 373*527a6266SJeff Kirsher 374*527a6266SJeff Kirsher /* Write PHY changes (SW-reset must follow) */ 375*527a6266SJeff Kirsher gm_phy_write(hw, port, PHY_MARV_EXT_CTRL_2, ctrl2); 376*527a6266SJeff Kirsher } 377*527a6266SJeff Kirsher 378*527a6266SJeff Kirsher 379*527a6266SJeff Kirsher /* disable energy detect */ 380*527a6266SJeff Kirsher ctrl &= ~PHY_M_PC_EN_DET_MSK; 381*527a6266SJeff Kirsher 382*527a6266SJeff Kirsher /* enable automatic crossover */ 383*527a6266SJeff Kirsher ctrl |= PHY_M_PC_MDI_XMODE(PHY_M_PC_ENA_AUTO); 384*527a6266SJeff Kirsher 385*527a6266SJeff Kirsher /* downshift on PHY 88E1112 and 88E1149 is changed */ 386*527a6266SJeff Kirsher if ( (sky2->flags & SKY2_FLAG_AUTO_SPEED) && 387*527a6266SJeff Kirsher (hw->flags & SKY2_HW_NEWER_PHY)) { 388*527a6266SJeff Kirsher /* set downshift counter to 3x and enable downshift */ 389*527a6266SJeff Kirsher ctrl &= ~PHY_M_PC_DSC_MSK; 390*527a6266SJeff Kirsher ctrl |= PHY_M_PC_DSC(2) | PHY_M_PC_DOWN_S_ENA; 391*527a6266SJeff Kirsher } 392*527a6266SJeff Kirsher } 393*527a6266SJeff Kirsher } else { 394*527a6266SJeff Kirsher /* workaround for deviation #4.88 (CRC errors) */ 395*527a6266SJeff Kirsher /* disable Automatic Crossover */ 396*527a6266SJeff Kirsher 397*527a6266SJeff Kirsher ctrl &= ~PHY_M_PC_MDIX_MSK; 398*527a6266SJeff Kirsher } 399*527a6266SJeff Kirsher 400*527a6266SJeff Kirsher gm_phy_write(hw, port, PHY_MARV_PHY_CTRL, ctrl); 401*527a6266SJeff Kirsher 402*527a6266SJeff Kirsher /* special setup for PHY 88E1112 Fiber */ 403*527a6266SJeff Kirsher if (hw->chip_id == CHIP_ID_YUKON_XL && (hw->flags & SKY2_HW_FIBRE_PHY)) { 404*527a6266SJeff Kirsher pg = gm_phy_read(hw, port, PHY_MARV_EXT_ADR); 405*527a6266SJeff Kirsher 406*527a6266SJeff Kirsher /* Fiber: select 1000BASE-X only mode MAC Specific Ctrl Reg. */ 407*527a6266SJeff Kirsher gm_phy_write(hw, port, PHY_MARV_EXT_ADR, 2); 408*527a6266SJeff Kirsher ctrl = gm_phy_read(hw, port, PHY_MARV_PHY_CTRL); 409*527a6266SJeff Kirsher ctrl &= ~PHY_M_MAC_MD_MSK; 410*527a6266SJeff Kirsher ctrl |= PHY_M_MAC_MODE_SEL(PHY_M_MAC_MD_1000BX); 411*527a6266SJeff Kirsher gm_phy_write(hw, port, PHY_MARV_PHY_CTRL, ctrl); 412*527a6266SJeff Kirsher 413*527a6266SJeff Kirsher if (hw->pmd_type == 'P') { 414*527a6266SJeff Kirsher /* select page 1 to access Fiber registers */ 415*527a6266SJeff Kirsher gm_phy_write(hw, port, PHY_MARV_EXT_ADR, 1); 416*527a6266SJeff Kirsher 417*527a6266SJeff Kirsher /* for SFP-module set SIGDET polarity to low */ 418*527a6266SJeff Kirsher ctrl = gm_phy_read(hw, port, PHY_MARV_PHY_CTRL); 419*527a6266SJeff Kirsher ctrl |= PHY_M_FIB_SIGD_POL; 420*527a6266SJeff Kirsher gm_phy_write(hw, port, PHY_MARV_PHY_CTRL, ctrl); 421*527a6266SJeff Kirsher } 422*527a6266SJeff Kirsher 423*527a6266SJeff Kirsher gm_phy_write(hw, port, PHY_MARV_EXT_ADR, pg); 424*527a6266SJeff Kirsher } 425*527a6266SJeff Kirsher 426*527a6266SJeff Kirsher ctrl = PHY_CT_RESET; 427*527a6266SJeff Kirsher ct1000 = 0; 428*527a6266SJeff Kirsher adv = PHY_AN_CSMA; 429*527a6266SJeff Kirsher reg = 0; 430*527a6266SJeff Kirsher 431*527a6266SJeff Kirsher if (sky2->flags & SKY2_FLAG_AUTO_SPEED) { 432*527a6266SJeff Kirsher if (sky2_is_copper(hw)) { 433*527a6266SJeff Kirsher if (sky2->advertising & ADVERTISED_1000baseT_Full) 434*527a6266SJeff Kirsher ct1000 |= PHY_M_1000C_AFD; 435*527a6266SJeff Kirsher if (sky2->advertising & ADVERTISED_1000baseT_Half) 436*527a6266SJeff Kirsher ct1000 |= PHY_M_1000C_AHD; 437*527a6266SJeff Kirsher if (sky2->advertising & ADVERTISED_100baseT_Full) 438*527a6266SJeff Kirsher adv |= PHY_M_AN_100_FD; 439*527a6266SJeff Kirsher if (sky2->advertising & ADVERTISED_100baseT_Half) 440*527a6266SJeff Kirsher adv |= PHY_M_AN_100_HD; 441*527a6266SJeff Kirsher if (sky2->advertising & ADVERTISED_10baseT_Full) 442*527a6266SJeff Kirsher adv |= PHY_M_AN_10_FD; 443*527a6266SJeff Kirsher if (sky2->advertising & ADVERTISED_10baseT_Half) 444*527a6266SJeff Kirsher adv |= PHY_M_AN_10_HD; 445*527a6266SJeff Kirsher 446*527a6266SJeff Kirsher } else { /* special defines for FIBER (88E1040S only) */ 447*527a6266SJeff Kirsher if (sky2->advertising & ADVERTISED_1000baseT_Full) 448*527a6266SJeff Kirsher adv |= PHY_M_AN_1000X_AFD; 449*527a6266SJeff Kirsher if (sky2->advertising & ADVERTISED_1000baseT_Half) 450*527a6266SJeff Kirsher adv |= PHY_M_AN_1000X_AHD; 451*527a6266SJeff Kirsher } 452*527a6266SJeff Kirsher 453*527a6266SJeff Kirsher /* Restart Auto-negotiation */ 454*527a6266SJeff Kirsher ctrl |= PHY_CT_ANE | PHY_CT_RE_CFG; 455*527a6266SJeff Kirsher } else { 456*527a6266SJeff Kirsher /* forced speed/duplex settings */ 457*527a6266SJeff Kirsher ct1000 = PHY_M_1000C_MSE; 458*527a6266SJeff Kirsher 459*527a6266SJeff Kirsher /* Disable auto update for duplex flow control and duplex */ 460*527a6266SJeff Kirsher reg |= GM_GPCR_AU_DUP_DIS | GM_GPCR_AU_SPD_DIS; 461*527a6266SJeff Kirsher 462*527a6266SJeff Kirsher switch (sky2->speed) { 463*527a6266SJeff Kirsher case SPEED_1000: 464*527a6266SJeff Kirsher ctrl |= PHY_CT_SP1000; 465*527a6266SJeff Kirsher reg |= GM_GPCR_SPEED_1000; 466*527a6266SJeff Kirsher break; 467*527a6266SJeff Kirsher case SPEED_100: 468*527a6266SJeff Kirsher ctrl |= PHY_CT_SP100; 469*527a6266SJeff Kirsher reg |= GM_GPCR_SPEED_100; 470*527a6266SJeff Kirsher break; 471*527a6266SJeff Kirsher } 472*527a6266SJeff Kirsher 473*527a6266SJeff Kirsher if (sky2->duplex == DUPLEX_FULL) { 474*527a6266SJeff Kirsher reg |= GM_GPCR_DUP_FULL; 475*527a6266SJeff Kirsher ctrl |= PHY_CT_DUP_MD; 476*527a6266SJeff Kirsher } else if (sky2->speed < SPEED_1000) 477*527a6266SJeff Kirsher sky2->flow_mode = FC_NONE; 478*527a6266SJeff Kirsher } 479*527a6266SJeff Kirsher 480*527a6266SJeff Kirsher if (sky2->flags & SKY2_FLAG_AUTO_PAUSE) { 481*527a6266SJeff Kirsher if (sky2_is_copper(hw)) 482*527a6266SJeff Kirsher adv |= copper_fc_adv[sky2->flow_mode]; 483*527a6266SJeff Kirsher else 484*527a6266SJeff Kirsher adv |= fiber_fc_adv[sky2->flow_mode]; 485*527a6266SJeff Kirsher } else { 486*527a6266SJeff Kirsher reg |= GM_GPCR_AU_FCT_DIS; 487*527a6266SJeff Kirsher reg |= gm_fc_disable[sky2->flow_mode]; 488*527a6266SJeff Kirsher 489*527a6266SJeff Kirsher /* Forward pause packets to GMAC? */ 490*527a6266SJeff Kirsher if (sky2->flow_mode & FC_RX) 491*527a6266SJeff Kirsher sky2_write8(hw, SK_REG(port, GMAC_CTRL), GMC_PAUSE_ON); 492*527a6266SJeff Kirsher else 493*527a6266SJeff Kirsher sky2_write8(hw, SK_REG(port, GMAC_CTRL), GMC_PAUSE_OFF); 494*527a6266SJeff Kirsher } 495*527a6266SJeff Kirsher 496*527a6266SJeff Kirsher gma_write16(hw, port, GM_GP_CTRL, reg); 497*527a6266SJeff Kirsher 498*527a6266SJeff Kirsher if (hw->flags & SKY2_HW_GIGABIT) 499*527a6266SJeff Kirsher gm_phy_write(hw, port, PHY_MARV_1000T_CTRL, ct1000); 500*527a6266SJeff Kirsher 501*527a6266SJeff Kirsher gm_phy_write(hw, port, PHY_MARV_AUNE_ADV, adv); 502*527a6266SJeff Kirsher gm_phy_write(hw, port, PHY_MARV_CTRL, ctrl); 503*527a6266SJeff Kirsher 504*527a6266SJeff Kirsher /* Setup Phy LED's */ 505*527a6266SJeff Kirsher ledctrl = PHY_M_LED_PULS_DUR(PULS_170MS); 506*527a6266SJeff Kirsher ledover = 0; 507*527a6266SJeff Kirsher 508*527a6266SJeff Kirsher switch (hw->chip_id) { 509*527a6266SJeff Kirsher case CHIP_ID_YUKON_FE: 510*527a6266SJeff Kirsher /* on 88E3082 these bits are at 11..9 (shifted left) */ 511*527a6266SJeff Kirsher ledctrl |= PHY_M_LED_BLINK_RT(BLINK_84MS) << 1; 512*527a6266SJeff Kirsher 513*527a6266SJeff Kirsher ctrl = gm_phy_read(hw, port, PHY_MARV_FE_LED_PAR); 514*527a6266SJeff Kirsher 515*527a6266SJeff Kirsher /* delete ACT LED control bits */ 516*527a6266SJeff Kirsher ctrl &= ~PHY_M_FELP_LED1_MSK; 517*527a6266SJeff Kirsher /* change ACT LED control to blink mode */ 518*527a6266SJeff Kirsher ctrl |= PHY_M_FELP_LED1_CTRL(LED_PAR_CTRL_ACT_BL); 519*527a6266SJeff Kirsher gm_phy_write(hw, port, PHY_MARV_FE_LED_PAR, ctrl); 520*527a6266SJeff Kirsher break; 521*527a6266SJeff Kirsher 522*527a6266SJeff Kirsher case CHIP_ID_YUKON_FE_P: 523*527a6266SJeff Kirsher /* Enable Link Partner Next Page */ 524*527a6266SJeff Kirsher ctrl = gm_phy_read(hw, port, PHY_MARV_PHY_CTRL); 525*527a6266SJeff Kirsher ctrl |= PHY_M_PC_ENA_LIP_NP; 526*527a6266SJeff Kirsher 527*527a6266SJeff Kirsher /* disable Energy Detect and enable scrambler */ 528*527a6266SJeff Kirsher ctrl &= ~(PHY_M_PC_ENA_ENE_DT | PHY_M_PC_DIS_SCRAMB); 529*527a6266SJeff Kirsher gm_phy_write(hw, port, PHY_MARV_PHY_CTRL, ctrl); 530*527a6266SJeff Kirsher 531*527a6266SJeff Kirsher /* set LED2 -> ACT, LED1 -> LINK, LED0 -> SPEED */ 532*527a6266SJeff Kirsher ctrl = PHY_M_FELP_LED2_CTRL(LED_PAR_CTRL_ACT_BL) | 533*527a6266SJeff Kirsher PHY_M_FELP_LED1_CTRL(LED_PAR_CTRL_LINK) | 534*527a6266SJeff Kirsher PHY_M_FELP_LED0_CTRL(LED_PAR_CTRL_SPEED); 535*527a6266SJeff Kirsher 536*527a6266SJeff Kirsher gm_phy_write(hw, port, PHY_MARV_FE_LED_PAR, ctrl); 537*527a6266SJeff Kirsher break; 538*527a6266SJeff Kirsher 539*527a6266SJeff Kirsher case CHIP_ID_YUKON_XL: 540*527a6266SJeff Kirsher pg = gm_phy_read(hw, port, PHY_MARV_EXT_ADR); 541*527a6266SJeff Kirsher 542*527a6266SJeff Kirsher /* select page 3 to access LED control register */ 543*527a6266SJeff Kirsher gm_phy_write(hw, port, PHY_MARV_EXT_ADR, 3); 544*527a6266SJeff Kirsher 545*527a6266SJeff Kirsher /* set LED Function Control register */ 546*527a6266SJeff Kirsher gm_phy_write(hw, port, PHY_MARV_PHY_CTRL, 547*527a6266SJeff Kirsher (PHY_M_LEDC_LOS_CTRL(1) | /* LINK/ACT */ 548*527a6266SJeff Kirsher PHY_M_LEDC_INIT_CTRL(7) | /* 10 Mbps */ 549*527a6266SJeff Kirsher PHY_M_LEDC_STA1_CTRL(7) | /* 100 Mbps */ 550*527a6266SJeff Kirsher PHY_M_LEDC_STA0_CTRL(7))); /* 1000 Mbps */ 551*527a6266SJeff Kirsher 552*527a6266SJeff Kirsher /* set Polarity Control register */ 553*527a6266SJeff Kirsher gm_phy_write(hw, port, PHY_MARV_PHY_STAT, 554*527a6266SJeff Kirsher (PHY_M_POLC_LS1_P_MIX(4) | 555*527a6266SJeff Kirsher PHY_M_POLC_IS0_P_MIX(4) | 556*527a6266SJeff Kirsher PHY_M_POLC_LOS_CTRL(2) | 557*527a6266SJeff Kirsher PHY_M_POLC_INIT_CTRL(2) | 558*527a6266SJeff Kirsher PHY_M_POLC_STA1_CTRL(2) | 559*527a6266SJeff Kirsher PHY_M_POLC_STA0_CTRL(2))); 560*527a6266SJeff Kirsher 561*527a6266SJeff Kirsher /* restore page register */ 562*527a6266SJeff Kirsher gm_phy_write(hw, port, PHY_MARV_EXT_ADR, pg); 563*527a6266SJeff Kirsher break; 564*527a6266SJeff Kirsher 565*527a6266SJeff Kirsher case CHIP_ID_YUKON_EC_U: 566*527a6266SJeff Kirsher case CHIP_ID_YUKON_EX: 567*527a6266SJeff Kirsher case CHIP_ID_YUKON_SUPR: 568*527a6266SJeff Kirsher pg = gm_phy_read(hw, port, PHY_MARV_EXT_ADR); 569*527a6266SJeff Kirsher 570*527a6266SJeff Kirsher /* select page 3 to access LED control register */ 571*527a6266SJeff Kirsher gm_phy_write(hw, port, PHY_MARV_EXT_ADR, 3); 572*527a6266SJeff Kirsher 573*527a6266SJeff Kirsher /* set LED Function Control register */ 574*527a6266SJeff Kirsher gm_phy_write(hw, port, PHY_MARV_PHY_CTRL, 575*527a6266SJeff Kirsher (PHY_M_LEDC_LOS_CTRL(1) | /* LINK/ACT */ 576*527a6266SJeff Kirsher PHY_M_LEDC_INIT_CTRL(8) | /* 10 Mbps */ 577*527a6266SJeff Kirsher PHY_M_LEDC_STA1_CTRL(7) | /* 100 Mbps */ 578*527a6266SJeff Kirsher PHY_M_LEDC_STA0_CTRL(7)));/* 1000 Mbps */ 579*527a6266SJeff Kirsher 580*527a6266SJeff Kirsher /* set Blink Rate in LED Timer Control Register */ 581*527a6266SJeff Kirsher gm_phy_write(hw, port, PHY_MARV_INT_MASK, 582*527a6266SJeff Kirsher ledctrl | PHY_M_LED_BLINK_RT(BLINK_84MS)); 583*527a6266SJeff Kirsher /* restore page register */ 584*527a6266SJeff Kirsher gm_phy_write(hw, port, PHY_MARV_EXT_ADR, pg); 585*527a6266SJeff Kirsher break; 586*527a6266SJeff Kirsher 587*527a6266SJeff Kirsher default: 588*527a6266SJeff Kirsher /* set Tx LED (LED_TX) to blink mode on Rx OR Tx activity */ 589*527a6266SJeff Kirsher ledctrl |= PHY_M_LED_BLINK_RT(BLINK_84MS) | PHY_M_LEDC_TX_CTRL; 590*527a6266SJeff Kirsher 591*527a6266SJeff Kirsher /* turn off the Rx LED (LED_RX) */ 592*527a6266SJeff Kirsher ledover |= PHY_M_LED_MO_RX(MO_LED_OFF); 593*527a6266SJeff Kirsher } 594*527a6266SJeff Kirsher 595*527a6266SJeff Kirsher if (hw->chip_id == CHIP_ID_YUKON_EC_U || hw->chip_id == CHIP_ID_YUKON_UL_2) { 596*527a6266SJeff Kirsher /* apply fixes in PHY AFE */ 597*527a6266SJeff Kirsher gm_phy_write(hw, port, PHY_MARV_EXT_ADR, 255); 598*527a6266SJeff Kirsher 599*527a6266SJeff Kirsher /* increase differential signal amplitude in 10BASE-T */ 600*527a6266SJeff Kirsher gm_phy_write(hw, port, 0x18, 0xaa99); 601*527a6266SJeff Kirsher gm_phy_write(hw, port, 0x17, 0x2011); 602*527a6266SJeff Kirsher 603*527a6266SJeff Kirsher if (hw->chip_id == CHIP_ID_YUKON_EC_U) { 604*527a6266SJeff Kirsher /* fix for IEEE A/B Symmetry failure in 1000BASE-T */ 605*527a6266SJeff Kirsher gm_phy_write(hw, port, 0x18, 0xa204); 606*527a6266SJeff Kirsher gm_phy_write(hw, port, 0x17, 0x2002); 607*527a6266SJeff Kirsher } 608*527a6266SJeff Kirsher 609*527a6266SJeff Kirsher /* set page register to 0 */ 610*527a6266SJeff Kirsher gm_phy_write(hw, port, PHY_MARV_EXT_ADR, 0); 611*527a6266SJeff Kirsher } else if (hw->chip_id == CHIP_ID_YUKON_FE_P && 612*527a6266SJeff Kirsher hw->chip_rev == CHIP_REV_YU_FE2_A0) { 613*527a6266SJeff Kirsher /* apply workaround for integrated resistors calibration */ 614*527a6266SJeff Kirsher gm_phy_write(hw, port, PHY_MARV_PAGE_ADDR, 17); 615*527a6266SJeff Kirsher gm_phy_write(hw, port, PHY_MARV_PAGE_DATA, 0x3f60); 616*527a6266SJeff Kirsher } else if (hw->chip_id == CHIP_ID_YUKON_OPT && hw->chip_rev == 0) { 617*527a6266SJeff Kirsher /* apply fixes in PHY AFE */ 618*527a6266SJeff Kirsher gm_phy_write(hw, port, PHY_MARV_EXT_ADR, 0x00ff); 619*527a6266SJeff Kirsher 620*527a6266SJeff Kirsher /* apply RDAC termination workaround */ 621*527a6266SJeff Kirsher gm_phy_write(hw, port, 24, 0x2800); 622*527a6266SJeff Kirsher gm_phy_write(hw, port, 23, 0x2001); 623*527a6266SJeff Kirsher 624*527a6266SJeff Kirsher /* set page register back to 0 */ 625*527a6266SJeff Kirsher gm_phy_write(hw, port, PHY_MARV_EXT_ADR, 0); 626*527a6266SJeff Kirsher } else if (hw->chip_id != CHIP_ID_YUKON_EX && 627*527a6266SJeff Kirsher hw->chip_id < CHIP_ID_YUKON_SUPR) { 628*527a6266SJeff Kirsher /* no effect on Yukon-XL */ 629*527a6266SJeff Kirsher gm_phy_write(hw, port, PHY_MARV_LED_CTRL, ledctrl); 630*527a6266SJeff Kirsher 631*527a6266SJeff Kirsher if (!(sky2->flags & SKY2_FLAG_AUTO_SPEED) || 632*527a6266SJeff Kirsher sky2->speed == SPEED_100) { 633*527a6266SJeff Kirsher /* turn on 100 Mbps LED (LED_LINK100) */ 634*527a6266SJeff Kirsher ledover |= PHY_M_LED_MO_100(MO_LED_ON); 635*527a6266SJeff Kirsher } 636*527a6266SJeff Kirsher 637*527a6266SJeff Kirsher if (ledover) 638*527a6266SJeff Kirsher gm_phy_write(hw, port, PHY_MARV_LED_OVER, ledover); 639*527a6266SJeff Kirsher 640*527a6266SJeff Kirsher } else if (hw->chip_id == CHIP_ID_YUKON_PRM && 641*527a6266SJeff Kirsher (sky2_read8(hw, B2_MAC_CFG) & 0xf) == 0x7) { 642*527a6266SJeff Kirsher int i; 643*527a6266SJeff Kirsher /* This a phy register setup workaround copied from vendor driver. */ 644*527a6266SJeff Kirsher static const struct { 645*527a6266SJeff Kirsher u16 reg, val; 646*527a6266SJeff Kirsher } eee_afe[] = { 647*527a6266SJeff Kirsher { 0x156, 0x58ce }, 648*527a6266SJeff Kirsher { 0x153, 0x99eb }, 649*527a6266SJeff Kirsher { 0x141, 0x8064 }, 650*527a6266SJeff Kirsher /* { 0x155, 0x130b },*/ 651*527a6266SJeff Kirsher { 0x000, 0x0000 }, 652*527a6266SJeff Kirsher { 0x151, 0x8433 }, 653*527a6266SJeff Kirsher { 0x14b, 0x8c44 }, 654*527a6266SJeff Kirsher { 0x14c, 0x0f90 }, 655*527a6266SJeff Kirsher { 0x14f, 0x39aa }, 656*527a6266SJeff Kirsher /* { 0x154, 0x2f39 },*/ 657*527a6266SJeff Kirsher { 0x14d, 0xba33 }, 658*527a6266SJeff Kirsher { 0x144, 0x0048 }, 659*527a6266SJeff Kirsher { 0x152, 0x2010 }, 660*527a6266SJeff Kirsher /* { 0x158, 0x1223 },*/ 661*527a6266SJeff Kirsher { 0x140, 0x4444 }, 662*527a6266SJeff Kirsher { 0x154, 0x2f3b }, 663*527a6266SJeff Kirsher { 0x158, 0xb203 }, 664*527a6266SJeff Kirsher { 0x157, 0x2029 }, 665*527a6266SJeff Kirsher }; 666*527a6266SJeff Kirsher 667*527a6266SJeff Kirsher /* Start Workaround for OptimaEEE Rev.Z0 */ 668*527a6266SJeff Kirsher gm_phy_write(hw, port, PHY_MARV_EXT_ADR, 0x00fb); 669*527a6266SJeff Kirsher 670*527a6266SJeff Kirsher gm_phy_write(hw, port, 1, 0x4099); 671*527a6266SJeff Kirsher gm_phy_write(hw, port, 3, 0x1120); 672*527a6266SJeff Kirsher gm_phy_write(hw, port, 11, 0x113c); 673*527a6266SJeff Kirsher gm_phy_write(hw, port, 14, 0x8100); 674*527a6266SJeff Kirsher gm_phy_write(hw, port, 15, 0x112a); 675*527a6266SJeff Kirsher gm_phy_write(hw, port, 17, 0x1008); 676*527a6266SJeff Kirsher 677*527a6266SJeff Kirsher gm_phy_write(hw, port, PHY_MARV_EXT_ADR, 0x00fc); 678*527a6266SJeff Kirsher gm_phy_write(hw, port, 1, 0x20b0); 679*527a6266SJeff Kirsher 680*527a6266SJeff Kirsher gm_phy_write(hw, port, PHY_MARV_EXT_ADR, 0x00ff); 681*527a6266SJeff Kirsher 682*527a6266SJeff Kirsher for (i = 0; i < ARRAY_SIZE(eee_afe); i++) { 683*527a6266SJeff Kirsher /* apply AFE settings */ 684*527a6266SJeff Kirsher gm_phy_write(hw, port, 17, eee_afe[i].val); 685*527a6266SJeff Kirsher gm_phy_write(hw, port, 16, eee_afe[i].reg | 1u<<13); 686*527a6266SJeff Kirsher } 687*527a6266SJeff Kirsher 688*527a6266SJeff Kirsher /* End Workaround for OptimaEEE */ 689*527a6266SJeff Kirsher gm_phy_write(hw, port, PHY_MARV_EXT_ADR, 0); 690*527a6266SJeff Kirsher 691*527a6266SJeff Kirsher /* Enable 10Base-Te (EEE) */ 692*527a6266SJeff Kirsher if (hw->chip_id >= CHIP_ID_YUKON_PRM) { 693*527a6266SJeff Kirsher reg = gm_phy_read(hw, port, PHY_MARV_EXT_CTRL); 694*527a6266SJeff Kirsher gm_phy_write(hw, port, PHY_MARV_EXT_CTRL, 695*527a6266SJeff Kirsher reg | PHY_M_10B_TE_ENABLE); 696*527a6266SJeff Kirsher } 697*527a6266SJeff Kirsher } 698*527a6266SJeff Kirsher 699*527a6266SJeff Kirsher /* Enable phy interrupt on auto-negotiation complete (or link up) */ 700*527a6266SJeff Kirsher if (sky2->flags & SKY2_FLAG_AUTO_SPEED) 701*527a6266SJeff Kirsher gm_phy_write(hw, port, PHY_MARV_INT_MASK, PHY_M_IS_AN_COMPL); 702*527a6266SJeff Kirsher else 703*527a6266SJeff Kirsher gm_phy_write(hw, port, PHY_MARV_INT_MASK, PHY_M_DEF_MSK); 704*527a6266SJeff Kirsher } 705*527a6266SJeff Kirsher 706*527a6266SJeff Kirsher static const u32 phy_power[] = { PCI_Y2_PHY1_POWD, PCI_Y2_PHY2_POWD }; 707*527a6266SJeff Kirsher static const u32 coma_mode[] = { PCI_Y2_PHY1_COMA, PCI_Y2_PHY2_COMA }; 708*527a6266SJeff Kirsher 709*527a6266SJeff Kirsher static void sky2_phy_power_up(struct sky2_hw *hw, unsigned port) 710*527a6266SJeff Kirsher { 711*527a6266SJeff Kirsher u32 reg1; 712*527a6266SJeff Kirsher 713*527a6266SJeff Kirsher sky2_write8(hw, B2_TST_CTRL1, TST_CFG_WRITE_ON); 714*527a6266SJeff Kirsher reg1 = sky2_pci_read32(hw, PCI_DEV_REG1); 715*527a6266SJeff Kirsher reg1 &= ~phy_power[port]; 716*527a6266SJeff Kirsher 717*527a6266SJeff Kirsher if (hw->chip_id == CHIP_ID_YUKON_XL && hw->chip_rev > CHIP_REV_YU_XL_A1) 718*527a6266SJeff Kirsher reg1 |= coma_mode[port]; 719*527a6266SJeff Kirsher 720*527a6266SJeff Kirsher sky2_pci_write32(hw, PCI_DEV_REG1, reg1); 721*527a6266SJeff Kirsher sky2_write8(hw, B2_TST_CTRL1, TST_CFG_WRITE_OFF); 722*527a6266SJeff Kirsher sky2_pci_read32(hw, PCI_DEV_REG1); 723*527a6266SJeff Kirsher 724*527a6266SJeff Kirsher if (hw->chip_id == CHIP_ID_YUKON_FE) 725*527a6266SJeff Kirsher gm_phy_write(hw, port, PHY_MARV_CTRL, PHY_CT_ANE); 726*527a6266SJeff Kirsher else if (hw->flags & SKY2_HW_ADV_POWER_CTL) 727*527a6266SJeff Kirsher sky2_write8(hw, SK_REG(port, GPHY_CTRL), GPC_RST_CLR); 728*527a6266SJeff Kirsher } 729*527a6266SJeff Kirsher 730*527a6266SJeff Kirsher static void sky2_phy_power_down(struct sky2_hw *hw, unsigned port) 731*527a6266SJeff Kirsher { 732*527a6266SJeff Kirsher u32 reg1; 733*527a6266SJeff Kirsher u16 ctrl; 734*527a6266SJeff Kirsher 735*527a6266SJeff Kirsher /* release GPHY Control reset */ 736*527a6266SJeff Kirsher sky2_write8(hw, SK_REG(port, GPHY_CTRL), GPC_RST_CLR); 737*527a6266SJeff Kirsher 738*527a6266SJeff Kirsher /* release GMAC reset */ 739*527a6266SJeff Kirsher sky2_write8(hw, SK_REG(port, GMAC_CTRL), GMC_RST_CLR); 740*527a6266SJeff Kirsher 741*527a6266SJeff Kirsher if (hw->flags & SKY2_HW_NEWER_PHY) { 742*527a6266SJeff Kirsher /* select page 2 to access MAC control register */ 743*527a6266SJeff Kirsher gm_phy_write(hw, port, PHY_MARV_EXT_ADR, 2); 744*527a6266SJeff Kirsher 745*527a6266SJeff Kirsher ctrl = gm_phy_read(hw, port, PHY_MARV_PHY_CTRL); 746*527a6266SJeff Kirsher /* allow GMII Power Down */ 747*527a6266SJeff Kirsher ctrl &= ~PHY_M_MAC_GMIF_PUP; 748*527a6266SJeff Kirsher gm_phy_write(hw, port, PHY_MARV_PHY_CTRL, ctrl); 749*527a6266SJeff Kirsher 750*527a6266SJeff Kirsher /* set page register back to 0 */ 751*527a6266SJeff Kirsher gm_phy_write(hw, port, PHY_MARV_EXT_ADR, 0); 752*527a6266SJeff Kirsher } 753*527a6266SJeff Kirsher 754*527a6266SJeff Kirsher /* setup General Purpose Control Register */ 755*527a6266SJeff Kirsher gma_write16(hw, port, GM_GP_CTRL, 756*527a6266SJeff Kirsher GM_GPCR_FL_PASS | GM_GPCR_SPEED_100 | 757*527a6266SJeff Kirsher GM_GPCR_AU_DUP_DIS | GM_GPCR_AU_FCT_DIS | 758*527a6266SJeff Kirsher GM_GPCR_AU_SPD_DIS); 759*527a6266SJeff Kirsher 760*527a6266SJeff Kirsher if (hw->chip_id != CHIP_ID_YUKON_EC) { 761*527a6266SJeff Kirsher if (hw->chip_id == CHIP_ID_YUKON_EC_U) { 762*527a6266SJeff Kirsher /* select page 2 to access MAC control register */ 763*527a6266SJeff Kirsher gm_phy_write(hw, port, PHY_MARV_EXT_ADR, 2); 764*527a6266SJeff Kirsher 765*527a6266SJeff Kirsher ctrl = gm_phy_read(hw, port, PHY_MARV_PHY_CTRL); 766*527a6266SJeff Kirsher /* enable Power Down */ 767*527a6266SJeff Kirsher ctrl |= PHY_M_PC_POW_D_ENA; 768*527a6266SJeff Kirsher gm_phy_write(hw, port, PHY_MARV_PHY_CTRL, ctrl); 769*527a6266SJeff Kirsher 770*527a6266SJeff Kirsher /* set page register back to 0 */ 771*527a6266SJeff Kirsher gm_phy_write(hw, port, PHY_MARV_EXT_ADR, 0); 772*527a6266SJeff Kirsher } 773*527a6266SJeff Kirsher 774*527a6266SJeff Kirsher /* set IEEE compatible Power Down Mode (dev. #4.99) */ 775*527a6266SJeff Kirsher gm_phy_write(hw, port, PHY_MARV_CTRL, PHY_CT_PDOWN); 776*527a6266SJeff Kirsher } 777*527a6266SJeff Kirsher 778*527a6266SJeff Kirsher sky2_write8(hw, B2_TST_CTRL1, TST_CFG_WRITE_ON); 779*527a6266SJeff Kirsher reg1 = sky2_pci_read32(hw, PCI_DEV_REG1); 780*527a6266SJeff Kirsher reg1 |= phy_power[port]; /* set PHY to PowerDown/COMA Mode */ 781*527a6266SJeff Kirsher sky2_pci_write32(hw, PCI_DEV_REG1, reg1); 782*527a6266SJeff Kirsher sky2_write8(hw, B2_TST_CTRL1, TST_CFG_WRITE_OFF); 783*527a6266SJeff Kirsher } 784*527a6266SJeff Kirsher 785*527a6266SJeff Kirsher /* configure IPG according to used link speed */ 786*527a6266SJeff Kirsher static void sky2_set_ipg(struct sky2_port *sky2) 787*527a6266SJeff Kirsher { 788*527a6266SJeff Kirsher u16 reg; 789*527a6266SJeff Kirsher 790*527a6266SJeff Kirsher reg = gma_read16(sky2->hw, sky2->port, GM_SERIAL_MODE); 791*527a6266SJeff Kirsher reg &= ~GM_SMOD_IPG_MSK; 792*527a6266SJeff Kirsher if (sky2->speed > SPEED_100) 793*527a6266SJeff Kirsher reg |= IPG_DATA_VAL(IPG_DATA_DEF_1000); 794*527a6266SJeff Kirsher else 795*527a6266SJeff Kirsher reg |= IPG_DATA_VAL(IPG_DATA_DEF_10_100); 796*527a6266SJeff Kirsher gma_write16(sky2->hw, sky2->port, GM_SERIAL_MODE, reg); 797*527a6266SJeff Kirsher } 798*527a6266SJeff Kirsher 799*527a6266SJeff Kirsher /* Enable Rx/Tx */ 800*527a6266SJeff Kirsher static void sky2_enable_rx_tx(struct sky2_port *sky2) 801*527a6266SJeff Kirsher { 802*527a6266SJeff Kirsher struct sky2_hw *hw = sky2->hw; 803*527a6266SJeff Kirsher unsigned port = sky2->port; 804*527a6266SJeff Kirsher u16 reg; 805*527a6266SJeff Kirsher 806*527a6266SJeff Kirsher reg = gma_read16(hw, port, GM_GP_CTRL); 807*527a6266SJeff Kirsher reg |= GM_GPCR_RX_ENA | GM_GPCR_TX_ENA; 808*527a6266SJeff Kirsher gma_write16(hw, port, GM_GP_CTRL, reg); 809*527a6266SJeff Kirsher } 810*527a6266SJeff Kirsher 811*527a6266SJeff Kirsher /* Force a renegotiation */ 812*527a6266SJeff Kirsher static void sky2_phy_reinit(struct sky2_port *sky2) 813*527a6266SJeff Kirsher { 814*527a6266SJeff Kirsher spin_lock_bh(&sky2->phy_lock); 815*527a6266SJeff Kirsher sky2_phy_init(sky2->hw, sky2->port); 816*527a6266SJeff Kirsher sky2_enable_rx_tx(sky2); 817*527a6266SJeff Kirsher spin_unlock_bh(&sky2->phy_lock); 818*527a6266SJeff Kirsher } 819*527a6266SJeff Kirsher 820*527a6266SJeff Kirsher /* Put device in state to listen for Wake On Lan */ 821*527a6266SJeff Kirsher static void sky2_wol_init(struct sky2_port *sky2) 822*527a6266SJeff Kirsher { 823*527a6266SJeff Kirsher struct sky2_hw *hw = sky2->hw; 824*527a6266SJeff Kirsher unsigned port = sky2->port; 825*527a6266SJeff Kirsher enum flow_control save_mode; 826*527a6266SJeff Kirsher u16 ctrl; 827*527a6266SJeff Kirsher 828*527a6266SJeff Kirsher /* Bring hardware out of reset */ 829*527a6266SJeff Kirsher sky2_write16(hw, B0_CTST, CS_RST_CLR); 830*527a6266SJeff Kirsher sky2_write16(hw, SK_REG(port, GMAC_LINK_CTRL), GMLC_RST_CLR); 831*527a6266SJeff Kirsher 832*527a6266SJeff Kirsher sky2_write8(hw, SK_REG(port, GPHY_CTRL), GPC_RST_CLR); 833*527a6266SJeff Kirsher sky2_write8(hw, SK_REG(port, GMAC_CTRL), GMC_RST_CLR); 834*527a6266SJeff Kirsher 835*527a6266SJeff Kirsher /* Force to 10/100 836*527a6266SJeff Kirsher * sky2_reset will re-enable on resume 837*527a6266SJeff Kirsher */ 838*527a6266SJeff Kirsher save_mode = sky2->flow_mode; 839*527a6266SJeff Kirsher ctrl = sky2->advertising; 840*527a6266SJeff Kirsher 841*527a6266SJeff Kirsher sky2->advertising &= ~(ADVERTISED_1000baseT_Half|ADVERTISED_1000baseT_Full); 842*527a6266SJeff Kirsher sky2->flow_mode = FC_NONE; 843*527a6266SJeff Kirsher 844*527a6266SJeff Kirsher spin_lock_bh(&sky2->phy_lock); 845*527a6266SJeff Kirsher sky2_phy_power_up(hw, port); 846*527a6266SJeff Kirsher sky2_phy_init(hw, port); 847*527a6266SJeff Kirsher spin_unlock_bh(&sky2->phy_lock); 848*527a6266SJeff Kirsher 849*527a6266SJeff Kirsher sky2->flow_mode = save_mode; 850*527a6266SJeff Kirsher sky2->advertising = ctrl; 851*527a6266SJeff Kirsher 852*527a6266SJeff Kirsher /* Set GMAC to no flow control and auto update for speed/duplex */ 853*527a6266SJeff Kirsher gma_write16(hw, port, GM_GP_CTRL, 854*527a6266SJeff Kirsher GM_GPCR_FC_TX_DIS|GM_GPCR_TX_ENA|GM_GPCR_RX_ENA| 855*527a6266SJeff Kirsher GM_GPCR_DUP_FULL|GM_GPCR_FC_RX_DIS|GM_GPCR_AU_FCT_DIS); 856*527a6266SJeff Kirsher 857*527a6266SJeff Kirsher /* Set WOL address */ 858*527a6266SJeff Kirsher memcpy_toio(hw->regs + WOL_REGS(port, WOL_MAC_ADDR), 859*527a6266SJeff Kirsher sky2->netdev->dev_addr, ETH_ALEN); 860*527a6266SJeff Kirsher 861*527a6266SJeff Kirsher /* Turn on appropriate WOL control bits */ 862*527a6266SJeff Kirsher sky2_write16(hw, WOL_REGS(port, WOL_CTRL_STAT), WOL_CTL_CLEAR_RESULT); 863*527a6266SJeff Kirsher ctrl = 0; 864*527a6266SJeff Kirsher if (sky2->wol & WAKE_PHY) 865*527a6266SJeff Kirsher ctrl |= WOL_CTL_ENA_PME_ON_LINK_CHG|WOL_CTL_ENA_LINK_CHG_UNIT; 866*527a6266SJeff Kirsher else 867*527a6266SJeff Kirsher ctrl |= WOL_CTL_DIS_PME_ON_LINK_CHG|WOL_CTL_DIS_LINK_CHG_UNIT; 868*527a6266SJeff Kirsher 869*527a6266SJeff Kirsher if (sky2->wol & WAKE_MAGIC) 870*527a6266SJeff Kirsher ctrl |= WOL_CTL_ENA_PME_ON_MAGIC_PKT|WOL_CTL_ENA_MAGIC_PKT_UNIT; 871*527a6266SJeff Kirsher else 872*527a6266SJeff Kirsher ctrl |= WOL_CTL_DIS_PME_ON_MAGIC_PKT|WOL_CTL_DIS_MAGIC_PKT_UNIT; 873*527a6266SJeff Kirsher 874*527a6266SJeff Kirsher ctrl |= WOL_CTL_DIS_PME_ON_PATTERN|WOL_CTL_DIS_PATTERN_UNIT; 875*527a6266SJeff Kirsher sky2_write16(hw, WOL_REGS(port, WOL_CTRL_STAT), ctrl); 876*527a6266SJeff Kirsher 877*527a6266SJeff Kirsher /* Disable PiG firmware */ 878*527a6266SJeff Kirsher sky2_write16(hw, B0_CTST, Y2_HW_WOL_OFF); 879*527a6266SJeff Kirsher 880*527a6266SJeff Kirsher /* block receiver */ 881*527a6266SJeff Kirsher sky2_write8(hw, SK_REG(port, RX_GMF_CTRL_T), GMF_RST_SET); 882*527a6266SJeff Kirsher } 883*527a6266SJeff Kirsher 884*527a6266SJeff Kirsher static void sky2_set_tx_stfwd(struct sky2_hw *hw, unsigned port) 885*527a6266SJeff Kirsher { 886*527a6266SJeff Kirsher struct net_device *dev = hw->dev[port]; 887*527a6266SJeff Kirsher 888*527a6266SJeff Kirsher if ( (hw->chip_id == CHIP_ID_YUKON_EX && 889*527a6266SJeff Kirsher hw->chip_rev != CHIP_REV_YU_EX_A0) || 890*527a6266SJeff Kirsher hw->chip_id >= CHIP_ID_YUKON_FE_P) { 891*527a6266SJeff Kirsher /* Yukon-Extreme B0 and further Extreme devices */ 892*527a6266SJeff Kirsher sky2_write32(hw, SK_REG(port, TX_GMF_CTRL_T), TX_STFW_ENA); 893*527a6266SJeff Kirsher } else if (dev->mtu > ETH_DATA_LEN) { 894*527a6266SJeff Kirsher /* set Tx GMAC FIFO Almost Empty Threshold */ 895*527a6266SJeff Kirsher sky2_write32(hw, SK_REG(port, TX_GMF_AE_THR), 896*527a6266SJeff Kirsher (ECU_JUMBO_WM << 16) | ECU_AE_THR); 897*527a6266SJeff Kirsher 898*527a6266SJeff Kirsher sky2_write32(hw, SK_REG(port, TX_GMF_CTRL_T), TX_STFW_DIS); 899*527a6266SJeff Kirsher } else 900*527a6266SJeff Kirsher sky2_write32(hw, SK_REG(port, TX_GMF_CTRL_T), TX_STFW_ENA); 901*527a6266SJeff Kirsher } 902*527a6266SJeff Kirsher 903*527a6266SJeff Kirsher static void sky2_mac_init(struct sky2_hw *hw, unsigned port) 904*527a6266SJeff Kirsher { 905*527a6266SJeff Kirsher struct sky2_port *sky2 = netdev_priv(hw->dev[port]); 906*527a6266SJeff Kirsher u16 reg; 907*527a6266SJeff Kirsher u32 rx_reg; 908*527a6266SJeff Kirsher int i; 909*527a6266SJeff Kirsher const u8 *addr = hw->dev[port]->dev_addr; 910*527a6266SJeff Kirsher 911*527a6266SJeff Kirsher sky2_write8(hw, SK_REG(port, GPHY_CTRL), GPC_RST_SET); 912*527a6266SJeff Kirsher sky2_write8(hw, SK_REG(port, GPHY_CTRL), GPC_RST_CLR); 913*527a6266SJeff Kirsher 914*527a6266SJeff Kirsher sky2_write8(hw, SK_REG(port, GMAC_CTRL), GMC_RST_CLR); 915*527a6266SJeff Kirsher 916*527a6266SJeff Kirsher if (hw->chip_id == CHIP_ID_YUKON_XL && 917*527a6266SJeff Kirsher hw->chip_rev == CHIP_REV_YU_XL_A0 && 918*527a6266SJeff Kirsher port == 1) { 919*527a6266SJeff Kirsher /* WA DEV_472 -- looks like crossed wires on port 2 */ 920*527a6266SJeff Kirsher /* clear GMAC 1 Control reset */ 921*527a6266SJeff Kirsher sky2_write8(hw, SK_REG(0, GMAC_CTRL), GMC_RST_CLR); 922*527a6266SJeff Kirsher do { 923*527a6266SJeff Kirsher sky2_write8(hw, SK_REG(1, GMAC_CTRL), GMC_RST_SET); 924*527a6266SJeff Kirsher sky2_write8(hw, SK_REG(1, GMAC_CTRL), GMC_RST_CLR); 925*527a6266SJeff Kirsher } while (gm_phy_read(hw, 1, PHY_MARV_ID0) != PHY_MARV_ID0_VAL || 926*527a6266SJeff Kirsher gm_phy_read(hw, 1, PHY_MARV_ID1) != PHY_MARV_ID1_Y2 || 927*527a6266SJeff Kirsher gm_phy_read(hw, 1, PHY_MARV_INT_MASK) != 0); 928*527a6266SJeff Kirsher } 929*527a6266SJeff Kirsher 930*527a6266SJeff Kirsher sky2_read16(hw, SK_REG(port, GMAC_IRQ_SRC)); 931*527a6266SJeff Kirsher 932*527a6266SJeff Kirsher /* Enable Transmit FIFO Underrun */ 933*527a6266SJeff Kirsher sky2_write8(hw, SK_REG(port, GMAC_IRQ_MSK), GMAC_DEF_MSK); 934*527a6266SJeff Kirsher 935*527a6266SJeff Kirsher spin_lock_bh(&sky2->phy_lock); 936*527a6266SJeff Kirsher sky2_phy_power_up(hw, port); 937*527a6266SJeff Kirsher sky2_phy_init(hw, port); 938*527a6266SJeff Kirsher spin_unlock_bh(&sky2->phy_lock); 939*527a6266SJeff Kirsher 940*527a6266SJeff Kirsher /* MIB clear */ 941*527a6266SJeff Kirsher reg = gma_read16(hw, port, GM_PHY_ADDR); 942*527a6266SJeff Kirsher gma_write16(hw, port, GM_PHY_ADDR, reg | GM_PAR_MIB_CLR); 943*527a6266SJeff Kirsher 944*527a6266SJeff Kirsher for (i = GM_MIB_CNT_BASE; i <= GM_MIB_CNT_END; i += 4) 945*527a6266SJeff Kirsher gma_read16(hw, port, i); 946*527a6266SJeff Kirsher gma_write16(hw, port, GM_PHY_ADDR, reg); 947*527a6266SJeff Kirsher 948*527a6266SJeff Kirsher /* transmit control */ 949*527a6266SJeff Kirsher gma_write16(hw, port, GM_TX_CTRL, TX_COL_THR(TX_COL_DEF)); 950*527a6266SJeff Kirsher 951*527a6266SJeff Kirsher /* receive control reg: unicast + multicast + no FCS */ 952*527a6266SJeff Kirsher gma_write16(hw, port, GM_RX_CTRL, 953*527a6266SJeff Kirsher GM_RXCR_UCF_ENA | GM_RXCR_CRC_DIS | GM_RXCR_MCF_ENA); 954*527a6266SJeff Kirsher 955*527a6266SJeff Kirsher /* transmit flow control */ 956*527a6266SJeff Kirsher gma_write16(hw, port, GM_TX_FLOW_CTRL, 0xffff); 957*527a6266SJeff Kirsher 958*527a6266SJeff Kirsher /* transmit parameter */ 959*527a6266SJeff Kirsher gma_write16(hw, port, GM_TX_PARAM, 960*527a6266SJeff Kirsher TX_JAM_LEN_VAL(TX_JAM_LEN_DEF) | 961*527a6266SJeff Kirsher TX_JAM_IPG_VAL(TX_JAM_IPG_DEF) | 962*527a6266SJeff Kirsher TX_IPG_JAM_DATA(TX_IPG_JAM_DEF) | 963*527a6266SJeff Kirsher TX_BACK_OFF_LIM(TX_BOF_LIM_DEF)); 964*527a6266SJeff Kirsher 965*527a6266SJeff Kirsher /* serial mode register */ 966*527a6266SJeff Kirsher reg = DATA_BLIND_VAL(DATA_BLIND_DEF) | 967*527a6266SJeff Kirsher GM_SMOD_VLAN_ENA | IPG_DATA_VAL(IPG_DATA_DEF_1000); 968*527a6266SJeff Kirsher 969*527a6266SJeff Kirsher if (hw->dev[port]->mtu > ETH_DATA_LEN) 970*527a6266SJeff Kirsher reg |= GM_SMOD_JUMBO_ENA; 971*527a6266SJeff Kirsher 972*527a6266SJeff Kirsher if (hw->chip_id == CHIP_ID_YUKON_EC_U && 973*527a6266SJeff Kirsher hw->chip_rev == CHIP_REV_YU_EC_U_B1) 974*527a6266SJeff Kirsher reg |= GM_NEW_FLOW_CTRL; 975*527a6266SJeff Kirsher 976*527a6266SJeff Kirsher gma_write16(hw, port, GM_SERIAL_MODE, reg); 977*527a6266SJeff Kirsher 978*527a6266SJeff Kirsher /* virtual address for data */ 979*527a6266SJeff Kirsher gma_set_addr(hw, port, GM_SRC_ADDR_2L, addr); 980*527a6266SJeff Kirsher 981*527a6266SJeff Kirsher /* physical address: used for pause frames */ 982*527a6266SJeff Kirsher gma_set_addr(hw, port, GM_SRC_ADDR_1L, addr); 983*527a6266SJeff Kirsher 984*527a6266SJeff Kirsher /* ignore counter overflows */ 985*527a6266SJeff Kirsher gma_write16(hw, port, GM_TX_IRQ_MSK, 0); 986*527a6266SJeff Kirsher gma_write16(hw, port, GM_RX_IRQ_MSK, 0); 987*527a6266SJeff Kirsher gma_write16(hw, port, GM_TR_IRQ_MSK, 0); 988*527a6266SJeff Kirsher 989*527a6266SJeff Kirsher /* Configure Rx MAC FIFO */ 990*527a6266SJeff Kirsher sky2_write8(hw, SK_REG(port, RX_GMF_CTRL_T), GMF_RST_CLR); 991*527a6266SJeff Kirsher rx_reg = GMF_OPER_ON | GMF_RX_F_FL_ON; 992*527a6266SJeff Kirsher if (hw->chip_id == CHIP_ID_YUKON_EX || 993*527a6266SJeff Kirsher hw->chip_id == CHIP_ID_YUKON_FE_P) 994*527a6266SJeff Kirsher rx_reg |= GMF_RX_OVER_ON; 995*527a6266SJeff Kirsher 996*527a6266SJeff Kirsher sky2_write32(hw, SK_REG(port, RX_GMF_CTRL_T), rx_reg); 997*527a6266SJeff Kirsher 998*527a6266SJeff Kirsher if (hw->chip_id == CHIP_ID_YUKON_XL) { 999*527a6266SJeff Kirsher /* Hardware errata - clear flush mask */ 1000*527a6266SJeff Kirsher sky2_write16(hw, SK_REG(port, RX_GMF_FL_MSK), 0); 1001*527a6266SJeff Kirsher } else { 1002*527a6266SJeff Kirsher /* Flush Rx MAC FIFO on any flow control or error */ 1003*527a6266SJeff Kirsher sky2_write16(hw, SK_REG(port, RX_GMF_FL_MSK), GMR_FS_ANY_ERR); 1004*527a6266SJeff Kirsher } 1005*527a6266SJeff Kirsher 1006*527a6266SJeff Kirsher /* Set threshold to 0xa (64 bytes) + 1 to workaround pause bug */ 1007*527a6266SJeff Kirsher reg = RX_GMF_FL_THR_DEF + 1; 1008*527a6266SJeff Kirsher /* Another magic mystery workaround from sk98lin */ 1009*527a6266SJeff Kirsher if (hw->chip_id == CHIP_ID_YUKON_FE_P && 1010*527a6266SJeff Kirsher hw->chip_rev == CHIP_REV_YU_FE2_A0) 1011*527a6266SJeff Kirsher reg = 0x178; 1012*527a6266SJeff Kirsher sky2_write16(hw, SK_REG(port, RX_GMF_FL_THR), reg); 1013*527a6266SJeff Kirsher 1014*527a6266SJeff Kirsher /* Configure Tx MAC FIFO */ 1015*527a6266SJeff Kirsher sky2_write8(hw, SK_REG(port, TX_GMF_CTRL_T), GMF_RST_CLR); 1016*527a6266SJeff Kirsher sky2_write16(hw, SK_REG(port, TX_GMF_CTRL_T), GMF_OPER_ON); 1017*527a6266SJeff Kirsher 1018*527a6266SJeff Kirsher /* On chips without ram buffer, pause is controlled by MAC level */ 1019*527a6266SJeff Kirsher if (!(hw->flags & SKY2_HW_RAM_BUFFER)) { 1020*527a6266SJeff Kirsher /* Pause threshold is scaled by 8 in bytes */ 1021*527a6266SJeff Kirsher if (hw->chip_id == CHIP_ID_YUKON_FE_P && 1022*527a6266SJeff Kirsher hw->chip_rev == CHIP_REV_YU_FE2_A0) 1023*527a6266SJeff Kirsher reg = 1568 / 8; 1024*527a6266SJeff Kirsher else 1025*527a6266SJeff Kirsher reg = 1024 / 8; 1026*527a6266SJeff Kirsher sky2_write16(hw, SK_REG(port, RX_GMF_UP_THR), reg); 1027*527a6266SJeff Kirsher sky2_write16(hw, SK_REG(port, RX_GMF_LP_THR), 768 / 8); 1028*527a6266SJeff Kirsher 1029*527a6266SJeff Kirsher sky2_set_tx_stfwd(hw, port); 1030*527a6266SJeff Kirsher } 1031*527a6266SJeff Kirsher 1032*527a6266SJeff Kirsher if (hw->chip_id == CHIP_ID_YUKON_FE_P && 1033*527a6266SJeff Kirsher hw->chip_rev == CHIP_REV_YU_FE2_A0) { 1034*527a6266SJeff Kirsher /* disable dynamic watermark */ 1035*527a6266SJeff Kirsher reg = sky2_read16(hw, SK_REG(port, TX_GMF_EA)); 1036*527a6266SJeff Kirsher reg &= ~TX_DYN_WM_ENA; 1037*527a6266SJeff Kirsher sky2_write16(hw, SK_REG(port, TX_GMF_EA), reg); 1038*527a6266SJeff Kirsher } 1039*527a6266SJeff Kirsher } 1040*527a6266SJeff Kirsher 1041*527a6266SJeff Kirsher /* Assign Ram Buffer allocation to queue */ 1042*527a6266SJeff Kirsher static void sky2_ramset(struct sky2_hw *hw, u16 q, u32 start, u32 space) 1043*527a6266SJeff Kirsher { 1044*527a6266SJeff Kirsher u32 end; 1045*527a6266SJeff Kirsher 1046*527a6266SJeff Kirsher /* convert from K bytes to qwords used for hw register */ 1047*527a6266SJeff Kirsher start *= 1024/8; 1048*527a6266SJeff Kirsher space *= 1024/8; 1049*527a6266SJeff Kirsher end = start + space - 1; 1050*527a6266SJeff Kirsher 1051*527a6266SJeff Kirsher sky2_write8(hw, RB_ADDR(q, RB_CTRL), RB_RST_CLR); 1052*527a6266SJeff Kirsher sky2_write32(hw, RB_ADDR(q, RB_START), start); 1053*527a6266SJeff Kirsher sky2_write32(hw, RB_ADDR(q, RB_END), end); 1054*527a6266SJeff Kirsher sky2_write32(hw, RB_ADDR(q, RB_WP), start); 1055*527a6266SJeff Kirsher sky2_write32(hw, RB_ADDR(q, RB_RP), start); 1056*527a6266SJeff Kirsher 1057*527a6266SJeff Kirsher if (q == Q_R1 || q == Q_R2) { 1058*527a6266SJeff Kirsher u32 tp = space - space/4; 1059*527a6266SJeff Kirsher 1060*527a6266SJeff Kirsher /* On receive queue's set the thresholds 1061*527a6266SJeff Kirsher * give receiver priority when > 3/4 full 1062*527a6266SJeff Kirsher * send pause when down to 2K 1063*527a6266SJeff Kirsher */ 1064*527a6266SJeff Kirsher sky2_write32(hw, RB_ADDR(q, RB_RX_UTHP), tp); 1065*527a6266SJeff Kirsher sky2_write32(hw, RB_ADDR(q, RB_RX_LTHP), space/2); 1066*527a6266SJeff Kirsher 1067*527a6266SJeff Kirsher tp = space - 2048/8; 1068*527a6266SJeff Kirsher sky2_write32(hw, RB_ADDR(q, RB_RX_UTPP), tp); 1069*527a6266SJeff Kirsher sky2_write32(hw, RB_ADDR(q, RB_RX_LTPP), space/4); 1070*527a6266SJeff Kirsher } else { 1071*527a6266SJeff Kirsher /* Enable store & forward on Tx queue's because 1072*527a6266SJeff Kirsher * Tx FIFO is only 1K on Yukon 1073*527a6266SJeff Kirsher */ 1074*527a6266SJeff Kirsher sky2_write8(hw, RB_ADDR(q, RB_CTRL), RB_ENA_STFWD); 1075*527a6266SJeff Kirsher } 1076*527a6266SJeff Kirsher 1077*527a6266SJeff Kirsher sky2_write8(hw, RB_ADDR(q, RB_CTRL), RB_ENA_OP_MD); 1078*527a6266SJeff Kirsher sky2_read8(hw, RB_ADDR(q, RB_CTRL)); 1079*527a6266SJeff Kirsher } 1080*527a6266SJeff Kirsher 1081*527a6266SJeff Kirsher /* Setup Bus Memory Interface */ 1082*527a6266SJeff Kirsher static void sky2_qset(struct sky2_hw *hw, u16 q) 1083*527a6266SJeff Kirsher { 1084*527a6266SJeff Kirsher sky2_write32(hw, Q_ADDR(q, Q_CSR), BMU_CLR_RESET); 1085*527a6266SJeff Kirsher sky2_write32(hw, Q_ADDR(q, Q_CSR), BMU_OPER_INIT); 1086*527a6266SJeff Kirsher sky2_write32(hw, Q_ADDR(q, Q_CSR), BMU_FIFO_OP_ON); 1087*527a6266SJeff Kirsher sky2_write32(hw, Q_ADDR(q, Q_WM), BMU_WM_DEFAULT); 1088*527a6266SJeff Kirsher } 1089*527a6266SJeff Kirsher 1090*527a6266SJeff Kirsher /* Setup prefetch unit registers. This is the interface between 1091*527a6266SJeff Kirsher * hardware and driver list elements 1092*527a6266SJeff Kirsher */ 1093*527a6266SJeff Kirsher static void sky2_prefetch_init(struct sky2_hw *hw, u32 qaddr, 1094*527a6266SJeff Kirsher dma_addr_t addr, u32 last) 1095*527a6266SJeff Kirsher { 1096*527a6266SJeff Kirsher sky2_write32(hw, Y2_QADDR(qaddr, PREF_UNIT_CTRL), PREF_UNIT_RST_SET); 1097*527a6266SJeff Kirsher sky2_write32(hw, Y2_QADDR(qaddr, PREF_UNIT_CTRL), PREF_UNIT_RST_CLR); 1098*527a6266SJeff Kirsher sky2_write32(hw, Y2_QADDR(qaddr, PREF_UNIT_ADDR_HI), upper_32_bits(addr)); 1099*527a6266SJeff Kirsher sky2_write32(hw, Y2_QADDR(qaddr, PREF_UNIT_ADDR_LO), lower_32_bits(addr)); 1100*527a6266SJeff Kirsher sky2_write16(hw, Y2_QADDR(qaddr, PREF_UNIT_LAST_IDX), last); 1101*527a6266SJeff Kirsher sky2_write32(hw, Y2_QADDR(qaddr, PREF_UNIT_CTRL), PREF_UNIT_OP_ON); 1102*527a6266SJeff Kirsher 1103*527a6266SJeff Kirsher sky2_read32(hw, Y2_QADDR(qaddr, PREF_UNIT_CTRL)); 1104*527a6266SJeff Kirsher } 1105*527a6266SJeff Kirsher 1106*527a6266SJeff Kirsher static inline struct sky2_tx_le *get_tx_le(struct sky2_port *sky2, u16 *slot) 1107*527a6266SJeff Kirsher { 1108*527a6266SJeff Kirsher struct sky2_tx_le *le = sky2->tx_le + *slot; 1109*527a6266SJeff Kirsher 1110*527a6266SJeff Kirsher *slot = RING_NEXT(*slot, sky2->tx_ring_size); 1111*527a6266SJeff Kirsher le->ctrl = 0; 1112*527a6266SJeff Kirsher return le; 1113*527a6266SJeff Kirsher } 1114*527a6266SJeff Kirsher 1115*527a6266SJeff Kirsher static void tx_init(struct sky2_port *sky2) 1116*527a6266SJeff Kirsher { 1117*527a6266SJeff Kirsher struct sky2_tx_le *le; 1118*527a6266SJeff Kirsher 1119*527a6266SJeff Kirsher sky2->tx_prod = sky2->tx_cons = 0; 1120*527a6266SJeff Kirsher sky2->tx_tcpsum = 0; 1121*527a6266SJeff Kirsher sky2->tx_last_mss = 0; 1122*527a6266SJeff Kirsher 1123*527a6266SJeff Kirsher le = get_tx_le(sky2, &sky2->tx_prod); 1124*527a6266SJeff Kirsher le->addr = 0; 1125*527a6266SJeff Kirsher le->opcode = OP_ADDR64 | HW_OWNER; 1126*527a6266SJeff Kirsher sky2->tx_last_upper = 0; 1127*527a6266SJeff Kirsher } 1128*527a6266SJeff Kirsher 1129*527a6266SJeff Kirsher /* Update chip's next pointer */ 1130*527a6266SJeff Kirsher static inline void sky2_put_idx(struct sky2_hw *hw, unsigned q, u16 idx) 1131*527a6266SJeff Kirsher { 1132*527a6266SJeff Kirsher /* Make sure write' to descriptors are complete before we tell hardware */ 1133*527a6266SJeff Kirsher wmb(); 1134*527a6266SJeff Kirsher sky2_write16(hw, Y2_QADDR(q, PREF_UNIT_PUT_IDX), idx); 1135*527a6266SJeff Kirsher 1136*527a6266SJeff Kirsher /* Synchronize I/O on since next processor may write to tail */ 1137*527a6266SJeff Kirsher mmiowb(); 1138*527a6266SJeff Kirsher } 1139*527a6266SJeff Kirsher 1140*527a6266SJeff Kirsher 1141*527a6266SJeff Kirsher static inline struct sky2_rx_le *sky2_next_rx(struct sky2_port *sky2) 1142*527a6266SJeff Kirsher { 1143*527a6266SJeff Kirsher struct sky2_rx_le *le = sky2->rx_le + sky2->rx_put; 1144*527a6266SJeff Kirsher sky2->rx_put = RING_NEXT(sky2->rx_put, RX_LE_SIZE); 1145*527a6266SJeff Kirsher le->ctrl = 0; 1146*527a6266SJeff Kirsher return le; 1147*527a6266SJeff Kirsher } 1148*527a6266SJeff Kirsher 1149*527a6266SJeff Kirsher static unsigned sky2_get_rx_threshold(struct sky2_port *sky2) 1150*527a6266SJeff Kirsher { 1151*527a6266SJeff Kirsher unsigned size; 1152*527a6266SJeff Kirsher 1153*527a6266SJeff Kirsher /* Space needed for frame data + headers rounded up */ 1154*527a6266SJeff Kirsher size = roundup(sky2->netdev->mtu + ETH_HLEN + VLAN_HLEN, 8); 1155*527a6266SJeff Kirsher 1156*527a6266SJeff Kirsher /* Stopping point for hardware truncation */ 1157*527a6266SJeff Kirsher return (size - 8) / sizeof(u32); 1158*527a6266SJeff Kirsher } 1159*527a6266SJeff Kirsher 1160*527a6266SJeff Kirsher static unsigned sky2_get_rx_data_size(struct sky2_port *sky2) 1161*527a6266SJeff Kirsher { 1162*527a6266SJeff Kirsher struct rx_ring_info *re; 1163*527a6266SJeff Kirsher unsigned size; 1164*527a6266SJeff Kirsher 1165*527a6266SJeff Kirsher /* Space needed for frame data + headers rounded up */ 1166*527a6266SJeff Kirsher size = roundup(sky2->netdev->mtu + ETH_HLEN + VLAN_HLEN, 8); 1167*527a6266SJeff Kirsher 1168*527a6266SJeff Kirsher sky2->rx_nfrags = size >> PAGE_SHIFT; 1169*527a6266SJeff Kirsher BUG_ON(sky2->rx_nfrags > ARRAY_SIZE(re->frag_addr)); 1170*527a6266SJeff Kirsher 1171*527a6266SJeff Kirsher /* Compute residue after pages */ 1172*527a6266SJeff Kirsher size -= sky2->rx_nfrags << PAGE_SHIFT; 1173*527a6266SJeff Kirsher 1174*527a6266SJeff Kirsher /* Optimize to handle small packets and headers */ 1175*527a6266SJeff Kirsher if (size < copybreak) 1176*527a6266SJeff Kirsher size = copybreak; 1177*527a6266SJeff Kirsher if (size < ETH_HLEN) 1178*527a6266SJeff Kirsher size = ETH_HLEN; 1179*527a6266SJeff Kirsher 1180*527a6266SJeff Kirsher return size; 1181*527a6266SJeff Kirsher } 1182*527a6266SJeff Kirsher 1183*527a6266SJeff Kirsher /* Build description to hardware for one receive segment */ 1184*527a6266SJeff Kirsher static void sky2_rx_add(struct sky2_port *sky2, u8 op, 1185*527a6266SJeff Kirsher dma_addr_t map, unsigned len) 1186*527a6266SJeff Kirsher { 1187*527a6266SJeff Kirsher struct sky2_rx_le *le; 1188*527a6266SJeff Kirsher 1189*527a6266SJeff Kirsher if (sizeof(dma_addr_t) > sizeof(u32)) { 1190*527a6266SJeff Kirsher le = sky2_next_rx(sky2); 1191*527a6266SJeff Kirsher le->addr = cpu_to_le32(upper_32_bits(map)); 1192*527a6266SJeff Kirsher le->opcode = OP_ADDR64 | HW_OWNER; 1193*527a6266SJeff Kirsher } 1194*527a6266SJeff Kirsher 1195*527a6266SJeff Kirsher le = sky2_next_rx(sky2); 1196*527a6266SJeff Kirsher le->addr = cpu_to_le32(lower_32_bits(map)); 1197*527a6266SJeff Kirsher le->length = cpu_to_le16(len); 1198*527a6266SJeff Kirsher le->opcode = op | HW_OWNER; 1199*527a6266SJeff Kirsher } 1200*527a6266SJeff Kirsher 1201*527a6266SJeff Kirsher /* Build description to hardware for one possibly fragmented skb */ 1202*527a6266SJeff Kirsher static void sky2_rx_submit(struct sky2_port *sky2, 1203*527a6266SJeff Kirsher const struct rx_ring_info *re) 1204*527a6266SJeff Kirsher { 1205*527a6266SJeff Kirsher int i; 1206*527a6266SJeff Kirsher 1207*527a6266SJeff Kirsher sky2_rx_add(sky2, OP_PACKET, re->data_addr, sky2->rx_data_size); 1208*527a6266SJeff Kirsher 1209*527a6266SJeff Kirsher for (i = 0; i < skb_shinfo(re->skb)->nr_frags; i++) 1210*527a6266SJeff Kirsher sky2_rx_add(sky2, OP_BUFFER, re->frag_addr[i], PAGE_SIZE); 1211*527a6266SJeff Kirsher } 1212*527a6266SJeff Kirsher 1213*527a6266SJeff Kirsher 1214*527a6266SJeff Kirsher static int sky2_rx_map_skb(struct pci_dev *pdev, struct rx_ring_info *re, 1215*527a6266SJeff Kirsher unsigned size) 1216*527a6266SJeff Kirsher { 1217*527a6266SJeff Kirsher struct sk_buff *skb = re->skb; 1218*527a6266SJeff Kirsher int i; 1219*527a6266SJeff Kirsher 1220*527a6266SJeff Kirsher re->data_addr = pci_map_single(pdev, skb->data, size, PCI_DMA_FROMDEVICE); 1221*527a6266SJeff Kirsher if (pci_dma_mapping_error(pdev, re->data_addr)) 1222*527a6266SJeff Kirsher goto mapping_error; 1223*527a6266SJeff Kirsher 1224*527a6266SJeff Kirsher dma_unmap_len_set(re, data_size, size); 1225*527a6266SJeff Kirsher 1226*527a6266SJeff Kirsher for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { 1227*527a6266SJeff Kirsher skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; 1228*527a6266SJeff Kirsher 1229*527a6266SJeff Kirsher re->frag_addr[i] = pci_map_page(pdev, frag->page, 1230*527a6266SJeff Kirsher frag->page_offset, 1231*527a6266SJeff Kirsher frag->size, 1232*527a6266SJeff Kirsher PCI_DMA_FROMDEVICE); 1233*527a6266SJeff Kirsher 1234*527a6266SJeff Kirsher if (pci_dma_mapping_error(pdev, re->frag_addr[i])) 1235*527a6266SJeff Kirsher goto map_page_error; 1236*527a6266SJeff Kirsher } 1237*527a6266SJeff Kirsher return 0; 1238*527a6266SJeff Kirsher 1239*527a6266SJeff Kirsher map_page_error: 1240*527a6266SJeff Kirsher while (--i >= 0) { 1241*527a6266SJeff Kirsher pci_unmap_page(pdev, re->frag_addr[i], 1242*527a6266SJeff Kirsher skb_shinfo(skb)->frags[i].size, 1243*527a6266SJeff Kirsher PCI_DMA_FROMDEVICE); 1244*527a6266SJeff Kirsher } 1245*527a6266SJeff Kirsher 1246*527a6266SJeff Kirsher pci_unmap_single(pdev, re->data_addr, dma_unmap_len(re, data_size), 1247*527a6266SJeff Kirsher PCI_DMA_FROMDEVICE); 1248*527a6266SJeff Kirsher 1249*527a6266SJeff Kirsher mapping_error: 1250*527a6266SJeff Kirsher if (net_ratelimit()) 1251*527a6266SJeff Kirsher dev_warn(&pdev->dev, "%s: rx mapping error\n", 1252*527a6266SJeff Kirsher skb->dev->name); 1253*527a6266SJeff Kirsher return -EIO; 1254*527a6266SJeff Kirsher } 1255*527a6266SJeff Kirsher 1256*527a6266SJeff Kirsher static void sky2_rx_unmap_skb(struct pci_dev *pdev, struct rx_ring_info *re) 1257*527a6266SJeff Kirsher { 1258*527a6266SJeff Kirsher struct sk_buff *skb = re->skb; 1259*527a6266SJeff Kirsher int i; 1260*527a6266SJeff Kirsher 1261*527a6266SJeff Kirsher pci_unmap_single(pdev, re->data_addr, dma_unmap_len(re, data_size), 1262*527a6266SJeff Kirsher PCI_DMA_FROMDEVICE); 1263*527a6266SJeff Kirsher 1264*527a6266SJeff Kirsher for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) 1265*527a6266SJeff Kirsher pci_unmap_page(pdev, re->frag_addr[i], 1266*527a6266SJeff Kirsher skb_shinfo(skb)->frags[i].size, 1267*527a6266SJeff Kirsher PCI_DMA_FROMDEVICE); 1268*527a6266SJeff Kirsher } 1269*527a6266SJeff Kirsher 1270*527a6266SJeff Kirsher /* Tell chip where to start receive checksum. 1271*527a6266SJeff Kirsher * Actually has two checksums, but set both same to avoid possible byte 1272*527a6266SJeff Kirsher * order problems. 1273*527a6266SJeff Kirsher */ 1274*527a6266SJeff Kirsher static void rx_set_checksum(struct sky2_port *sky2) 1275*527a6266SJeff Kirsher { 1276*527a6266SJeff Kirsher struct sky2_rx_le *le = sky2_next_rx(sky2); 1277*527a6266SJeff Kirsher 1278*527a6266SJeff Kirsher le->addr = cpu_to_le32((ETH_HLEN << 16) | ETH_HLEN); 1279*527a6266SJeff Kirsher le->ctrl = 0; 1280*527a6266SJeff Kirsher le->opcode = OP_TCPSTART | HW_OWNER; 1281*527a6266SJeff Kirsher 1282*527a6266SJeff Kirsher sky2_write32(sky2->hw, 1283*527a6266SJeff Kirsher Q_ADDR(rxqaddr[sky2->port], Q_CSR), 1284*527a6266SJeff Kirsher (sky2->netdev->features & NETIF_F_RXCSUM) 1285*527a6266SJeff Kirsher ? BMU_ENA_RX_CHKSUM : BMU_DIS_RX_CHKSUM); 1286*527a6266SJeff Kirsher } 1287*527a6266SJeff Kirsher 1288*527a6266SJeff Kirsher /* Enable/disable receive hash calculation (RSS) */ 1289*527a6266SJeff Kirsher static void rx_set_rss(struct net_device *dev, u32 features) 1290*527a6266SJeff Kirsher { 1291*527a6266SJeff Kirsher struct sky2_port *sky2 = netdev_priv(dev); 1292*527a6266SJeff Kirsher struct sky2_hw *hw = sky2->hw; 1293*527a6266SJeff Kirsher int i, nkeys = 4; 1294*527a6266SJeff Kirsher 1295*527a6266SJeff Kirsher /* Supports IPv6 and other modes */ 1296*527a6266SJeff Kirsher if (hw->flags & SKY2_HW_NEW_LE) { 1297*527a6266SJeff Kirsher nkeys = 10; 1298*527a6266SJeff Kirsher sky2_write32(hw, SK_REG(sky2->port, RSS_CFG), HASH_ALL); 1299*527a6266SJeff Kirsher } 1300*527a6266SJeff Kirsher 1301*527a6266SJeff Kirsher /* Program RSS initial values */ 1302*527a6266SJeff Kirsher if (features & NETIF_F_RXHASH) { 1303*527a6266SJeff Kirsher u32 key[nkeys]; 1304*527a6266SJeff Kirsher 1305*527a6266SJeff Kirsher get_random_bytes(key, nkeys * sizeof(u32)); 1306*527a6266SJeff Kirsher for (i = 0; i < nkeys; i++) 1307*527a6266SJeff Kirsher sky2_write32(hw, SK_REG(sky2->port, RSS_KEY + i * 4), 1308*527a6266SJeff Kirsher key[i]); 1309*527a6266SJeff Kirsher 1310*527a6266SJeff Kirsher /* Need to turn on (undocumented) flag to make hashing work */ 1311*527a6266SJeff Kirsher sky2_write32(hw, SK_REG(sky2->port, RX_GMF_CTRL_T), 1312*527a6266SJeff Kirsher RX_STFW_ENA); 1313*527a6266SJeff Kirsher 1314*527a6266SJeff Kirsher sky2_write32(hw, Q_ADDR(rxqaddr[sky2->port], Q_CSR), 1315*527a6266SJeff Kirsher BMU_ENA_RX_RSS_HASH); 1316*527a6266SJeff Kirsher } else 1317*527a6266SJeff Kirsher sky2_write32(hw, Q_ADDR(rxqaddr[sky2->port], Q_CSR), 1318*527a6266SJeff Kirsher BMU_DIS_RX_RSS_HASH); 1319*527a6266SJeff Kirsher } 1320*527a6266SJeff Kirsher 1321*527a6266SJeff Kirsher /* 1322*527a6266SJeff Kirsher * The RX Stop command will not work for Yukon-2 if the BMU does not 1323*527a6266SJeff Kirsher * reach the end of packet and since we can't make sure that we have 1324*527a6266SJeff Kirsher * incoming data, we must reset the BMU while it is not doing a DMA 1325*527a6266SJeff Kirsher * transfer. Since it is possible that the RX path is still active, 1326*527a6266SJeff Kirsher * the RX RAM buffer will be stopped first, so any possible incoming 1327*527a6266SJeff Kirsher * data will not trigger a DMA. After the RAM buffer is stopped, the 1328*527a6266SJeff Kirsher * BMU is polled until any DMA in progress is ended and only then it 1329*527a6266SJeff Kirsher * will be reset. 1330*527a6266SJeff Kirsher */ 1331*527a6266SJeff Kirsher static void sky2_rx_stop(struct sky2_port *sky2) 1332*527a6266SJeff Kirsher { 1333*527a6266SJeff Kirsher struct sky2_hw *hw = sky2->hw; 1334*527a6266SJeff Kirsher unsigned rxq = rxqaddr[sky2->port]; 1335*527a6266SJeff Kirsher int i; 1336*527a6266SJeff Kirsher 1337*527a6266SJeff Kirsher /* disable the RAM Buffer receive queue */ 1338*527a6266SJeff Kirsher sky2_write8(hw, RB_ADDR(rxq, RB_CTRL), RB_DIS_OP_MD); 1339*527a6266SJeff Kirsher 1340*527a6266SJeff Kirsher for (i = 0; i < 0xffff; i++) 1341*527a6266SJeff Kirsher if (sky2_read8(hw, RB_ADDR(rxq, Q_RSL)) 1342*527a6266SJeff Kirsher == sky2_read8(hw, RB_ADDR(rxq, Q_RL))) 1343*527a6266SJeff Kirsher goto stopped; 1344*527a6266SJeff Kirsher 1345*527a6266SJeff Kirsher netdev_warn(sky2->netdev, "receiver stop failed\n"); 1346*527a6266SJeff Kirsher stopped: 1347*527a6266SJeff Kirsher sky2_write32(hw, Q_ADDR(rxq, Q_CSR), BMU_RST_SET | BMU_FIFO_RST); 1348*527a6266SJeff Kirsher 1349*527a6266SJeff Kirsher /* reset the Rx prefetch unit */ 1350*527a6266SJeff Kirsher sky2_write32(hw, Y2_QADDR(rxq, PREF_UNIT_CTRL), PREF_UNIT_RST_SET); 1351*527a6266SJeff Kirsher mmiowb(); 1352*527a6266SJeff Kirsher } 1353*527a6266SJeff Kirsher 1354*527a6266SJeff Kirsher /* Clean out receive buffer area, assumes receiver hardware stopped */ 1355*527a6266SJeff Kirsher static void sky2_rx_clean(struct sky2_port *sky2) 1356*527a6266SJeff Kirsher { 1357*527a6266SJeff Kirsher unsigned i; 1358*527a6266SJeff Kirsher 1359*527a6266SJeff Kirsher memset(sky2->rx_le, 0, RX_LE_BYTES); 1360*527a6266SJeff Kirsher for (i = 0; i < sky2->rx_pending; i++) { 1361*527a6266SJeff Kirsher struct rx_ring_info *re = sky2->rx_ring + i; 1362*527a6266SJeff Kirsher 1363*527a6266SJeff Kirsher if (re->skb) { 1364*527a6266SJeff Kirsher sky2_rx_unmap_skb(sky2->hw->pdev, re); 1365*527a6266SJeff Kirsher kfree_skb(re->skb); 1366*527a6266SJeff Kirsher re->skb = NULL; 1367*527a6266SJeff Kirsher } 1368*527a6266SJeff Kirsher } 1369*527a6266SJeff Kirsher } 1370*527a6266SJeff Kirsher 1371*527a6266SJeff Kirsher /* Basic MII support */ 1372*527a6266SJeff Kirsher static int sky2_ioctl(struct net_device *dev, struct ifreq *ifr, int cmd) 1373*527a6266SJeff Kirsher { 1374*527a6266SJeff Kirsher struct mii_ioctl_data *data = if_mii(ifr); 1375*527a6266SJeff Kirsher struct sky2_port *sky2 = netdev_priv(dev); 1376*527a6266SJeff Kirsher struct sky2_hw *hw = sky2->hw; 1377*527a6266SJeff Kirsher int err = -EOPNOTSUPP; 1378*527a6266SJeff Kirsher 1379*527a6266SJeff Kirsher if (!netif_running(dev)) 1380*527a6266SJeff Kirsher return -ENODEV; /* Phy still in reset */ 1381*527a6266SJeff Kirsher 1382*527a6266SJeff Kirsher switch (cmd) { 1383*527a6266SJeff Kirsher case SIOCGMIIPHY: 1384*527a6266SJeff Kirsher data->phy_id = PHY_ADDR_MARV; 1385*527a6266SJeff Kirsher 1386*527a6266SJeff Kirsher /* fallthru */ 1387*527a6266SJeff Kirsher case SIOCGMIIREG: { 1388*527a6266SJeff Kirsher u16 val = 0; 1389*527a6266SJeff Kirsher 1390*527a6266SJeff Kirsher spin_lock_bh(&sky2->phy_lock); 1391*527a6266SJeff Kirsher err = __gm_phy_read(hw, sky2->port, data->reg_num & 0x1f, &val); 1392*527a6266SJeff Kirsher spin_unlock_bh(&sky2->phy_lock); 1393*527a6266SJeff Kirsher 1394*527a6266SJeff Kirsher data->val_out = val; 1395*527a6266SJeff Kirsher break; 1396*527a6266SJeff Kirsher } 1397*527a6266SJeff Kirsher 1398*527a6266SJeff Kirsher case SIOCSMIIREG: 1399*527a6266SJeff Kirsher spin_lock_bh(&sky2->phy_lock); 1400*527a6266SJeff Kirsher err = gm_phy_write(hw, sky2->port, data->reg_num & 0x1f, 1401*527a6266SJeff Kirsher data->val_in); 1402*527a6266SJeff Kirsher spin_unlock_bh(&sky2->phy_lock); 1403*527a6266SJeff Kirsher break; 1404*527a6266SJeff Kirsher } 1405*527a6266SJeff Kirsher return err; 1406*527a6266SJeff Kirsher } 1407*527a6266SJeff Kirsher 1408*527a6266SJeff Kirsher #define SKY2_VLAN_OFFLOADS (NETIF_F_IP_CSUM | NETIF_F_SG | NETIF_F_TSO) 1409*527a6266SJeff Kirsher 1410*527a6266SJeff Kirsher static void sky2_vlan_mode(struct net_device *dev, u32 features) 1411*527a6266SJeff Kirsher { 1412*527a6266SJeff Kirsher struct sky2_port *sky2 = netdev_priv(dev); 1413*527a6266SJeff Kirsher struct sky2_hw *hw = sky2->hw; 1414*527a6266SJeff Kirsher u16 port = sky2->port; 1415*527a6266SJeff Kirsher 1416*527a6266SJeff Kirsher if (features & NETIF_F_HW_VLAN_RX) 1417*527a6266SJeff Kirsher sky2_write32(hw, SK_REG(port, RX_GMF_CTRL_T), 1418*527a6266SJeff Kirsher RX_VLAN_STRIP_ON); 1419*527a6266SJeff Kirsher else 1420*527a6266SJeff Kirsher sky2_write32(hw, SK_REG(port, RX_GMF_CTRL_T), 1421*527a6266SJeff Kirsher RX_VLAN_STRIP_OFF); 1422*527a6266SJeff Kirsher 1423*527a6266SJeff Kirsher if (features & NETIF_F_HW_VLAN_TX) { 1424*527a6266SJeff Kirsher sky2_write32(hw, SK_REG(port, TX_GMF_CTRL_T), 1425*527a6266SJeff Kirsher TX_VLAN_TAG_ON); 1426*527a6266SJeff Kirsher 1427*527a6266SJeff Kirsher dev->vlan_features |= SKY2_VLAN_OFFLOADS; 1428*527a6266SJeff Kirsher } else { 1429*527a6266SJeff Kirsher sky2_write32(hw, SK_REG(port, TX_GMF_CTRL_T), 1430*527a6266SJeff Kirsher TX_VLAN_TAG_OFF); 1431*527a6266SJeff Kirsher 1432*527a6266SJeff Kirsher /* Can't do transmit offload of vlan without hw vlan */ 1433*527a6266SJeff Kirsher dev->vlan_features &= ~SKY2_VLAN_OFFLOADS; 1434*527a6266SJeff Kirsher } 1435*527a6266SJeff Kirsher } 1436*527a6266SJeff Kirsher 1437*527a6266SJeff Kirsher /* Amount of required worst case padding in rx buffer */ 1438*527a6266SJeff Kirsher static inline unsigned sky2_rx_pad(const struct sky2_hw *hw) 1439*527a6266SJeff Kirsher { 1440*527a6266SJeff Kirsher return (hw->flags & SKY2_HW_RAM_BUFFER) ? 8 : 2; 1441*527a6266SJeff Kirsher } 1442*527a6266SJeff Kirsher 1443*527a6266SJeff Kirsher /* 1444*527a6266SJeff Kirsher * Allocate an skb for receiving. If the MTU is large enough 1445*527a6266SJeff Kirsher * make the skb non-linear with a fragment list of pages. 1446*527a6266SJeff Kirsher */ 1447*527a6266SJeff Kirsher static struct sk_buff *sky2_rx_alloc(struct sky2_port *sky2, gfp_t gfp) 1448*527a6266SJeff Kirsher { 1449*527a6266SJeff Kirsher struct sk_buff *skb; 1450*527a6266SJeff Kirsher int i; 1451*527a6266SJeff Kirsher 1452*527a6266SJeff Kirsher skb = __netdev_alloc_skb(sky2->netdev, 1453*527a6266SJeff Kirsher sky2->rx_data_size + sky2_rx_pad(sky2->hw), 1454*527a6266SJeff Kirsher gfp); 1455*527a6266SJeff Kirsher if (!skb) 1456*527a6266SJeff Kirsher goto nomem; 1457*527a6266SJeff Kirsher 1458*527a6266SJeff Kirsher if (sky2->hw->flags & SKY2_HW_RAM_BUFFER) { 1459*527a6266SJeff Kirsher unsigned char *start; 1460*527a6266SJeff Kirsher /* 1461*527a6266SJeff Kirsher * Workaround for a bug in FIFO that cause hang 1462*527a6266SJeff Kirsher * if the FIFO if the receive buffer is not 64 byte aligned. 1463*527a6266SJeff Kirsher * The buffer returned from netdev_alloc_skb is 1464*527a6266SJeff Kirsher * aligned except if slab debugging is enabled. 1465*527a6266SJeff Kirsher */ 1466*527a6266SJeff Kirsher start = PTR_ALIGN(skb->data, 8); 1467*527a6266SJeff Kirsher skb_reserve(skb, start - skb->data); 1468*527a6266SJeff Kirsher } else 1469*527a6266SJeff Kirsher skb_reserve(skb, NET_IP_ALIGN); 1470*527a6266SJeff Kirsher 1471*527a6266SJeff Kirsher for (i = 0; i < sky2->rx_nfrags; i++) { 1472*527a6266SJeff Kirsher struct page *page = alloc_page(gfp); 1473*527a6266SJeff Kirsher 1474*527a6266SJeff Kirsher if (!page) 1475*527a6266SJeff Kirsher goto free_partial; 1476*527a6266SJeff Kirsher skb_fill_page_desc(skb, i, page, 0, PAGE_SIZE); 1477*527a6266SJeff Kirsher } 1478*527a6266SJeff Kirsher 1479*527a6266SJeff Kirsher return skb; 1480*527a6266SJeff Kirsher free_partial: 1481*527a6266SJeff Kirsher kfree_skb(skb); 1482*527a6266SJeff Kirsher nomem: 1483*527a6266SJeff Kirsher return NULL; 1484*527a6266SJeff Kirsher } 1485*527a6266SJeff Kirsher 1486*527a6266SJeff Kirsher static inline void sky2_rx_update(struct sky2_port *sky2, unsigned rxq) 1487*527a6266SJeff Kirsher { 1488*527a6266SJeff Kirsher sky2_put_idx(sky2->hw, rxq, sky2->rx_put); 1489*527a6266SJeff Kirsher } 1490*527a6266SJeff Kirsher 1491*527a6266SJeff Kirsher static int sky2_alloc_rx_skbs(struct sky2_port *sky2) 1492*527a6266SJeff Kirsher { 1493*527a6266SJeff Kirsher struct sky2_hw *hw = sky2->hw; 1494*527a6266SJeff Kirsher unsigned i; 1495*527a6266SJeff Kirsher 1496*527a6266SJeff Kirsher sky2->rx_data_size = sky2_get_rx_data_size(sky2); 1497*527a6266SJeff Kirsher 1498*527a6266SJeff Kirsher /* Fill Rx ring */ 1499*527a6266SJeff Kirsher for (i = 0; i < sky2->rx_pending; i++) { 1500*527a6266SJeff Kirsher struct rx_ring_info *re = sky2->rx_ring + i; 1501*527a6266SJeff Kirsher 1502*527a6266SJeff Kirsher re->skb = sky2_rx_alloc(sky2, GFP_KERNEL); 1503*527a6266SJeff Kirsher if (!re->skb) 1504*527a6266SJeff Kirsher return -ENOMEM; 1505*527a6266SJeff Kirsher 1506*527a6266SJeff Kirsher if (sky2_rx_map_skb(hw->pdev, re, sky2->rx_data_size)) { 1507*527a6266SJeff Kirsher dev_kfree_skb(re->skb); 1508*527a6266SJeff Kirsher re->skb = NULL; 1509*527a6266SJeff Kirsher return -ENOMEM; 1510*527a6266SJeff Kirsher } 1511*527a6266SJeff Kirsher } 1512*527a6266SJeff Kirsher return 0; 1513*527a6266SJeff Kirsher } 1514*527a6266SJeff Kirsher 1515*527a6266SJeff Kirsher /* 1516*527a6266SJeff Kirsher * Setup receiver buffer pool. 1517*527a6266SJeff Kirsher * Normal case this ends up creating one list element for skb 1518*527a6266SJeff Kirsher * in the receive ring. Worst case if using large MTU and each 1519*527a6266SJeff Kirsher * allocation falls on a different 64 bit region, that results 1520*527a6266SJeff Kirsher * in 6 list elements per ring entry. 1521*527a6266SJeff Kirsher * One element is used for checksum enable/disable, and one 1522*527a6266SJeff Kirsher * extra to avoid wrap. 1523*527a6266SJeff Kirsher */ 1524*527a6266SJeff Kirsher static void sky2_rx_start(struct sky2_port *sky2) 1525*527a6266SJeff Kirsher { 1526*527a6266SJeff Kirsher struct sky2_hw *hw = sky2->hw; 1527*527a6266SJeff Kirsher struct rx_ring_info *re; 1528*527a6266SJeff Kirsher unsigned rxq = rxqaddr[sky2->port]; 1529*527a6266SJeff Kirsher unsigned i, thresh; 1530*527a6266SJeff Kirsher 1531*527a6266SJeff Kirsher sky2->rx_put = sky2->rx_next = 0; 1532*527a6266SJeff Kirsher sky2_qset(hw, rxq); 1533*527a6266SJeff Kirsher 1534*527a6266SJeff Kirsher /* On PCI express lowering the watermark gives better performance */ 1535*527a6266SJeff Kirsher if (pci_is_pcie(hw->pdev)) 1536*527a6266SJeff Kirsher sky2_write32(hw, Q_ADDR(rxq, Q_WM), BMU_WM_PEX); 1537*527a6266SJeff Kirsher 1538*527a6266SJeff Kirsher /* These chips have no ram buffer? 1539*527a6266SJeff Kirsher * MAC Rx RAM Read is controlled by hardware */ 1540*527a6266SJeff Kirsher if (hw->chip_id == CHIP_ID_YUKON_EC_U && 1541*527a6266SJeff Kirsher hw->chip_rev > CHIP_REV_YU_EC_U_A0) 1542*527a6266SJeff Kirsher sky2_write32(hw, Q_ADDR(rxq, Q_TEST), F_M_RX_RAM_DIS); 1543*527a6266SJeff Kirsher 1544*527a6266SJeff Kirsher sky2_prefetch_init(hw, rxq, sky2->rx_le_map, RX_LE_SIZE - 1); 1545*527a6266SJeff Kirsher 1546*527a6266SJeff Kirsher if (!(hw->flags & SKY2_HW_NEW_LE)) 1547*527a6266SJeff Kirsher rx_set_checksum(sky2); 1548*527a6266SJeff Kirsher 1549*527a6266SJeff Kirsher if (!(hw->flags & SKY2_HW_RSS_BROKEN)) 1550*527a6266SJeff Kirsher rx_set_rss(sky2->netdev, sky2->netdev->features); 1551*527a6266SJeff Kirsher 1552*527a6266SJeff Kirsher /* submit Rx ring */ 1553*527a6266SJeff Kirsher for (i = 0; i < sky2->rx_pending; i++) { 1554*527a6266SJeff Kirsher re = sky2->rx_ring + i; 1555*527a6266SJeff Kirsher sky2_rx_submit(sky2, re); 1556*527a6266SJeff Kirsher } 1557*527a6266SJeff Kirsher 1558*527a6266SJeff Kirsher /* 1559*527a6266SJeff Kirsher * The receiver hangs if it receives frames larger than the 1560*527a6266SJeff Kirsher * packet buffer. As a workaround, truncate oversize frames, but 1561*527a6266SJeff Kirsher * the register is limited to 9 bits, so if you do frames > 2052 1562*527a6266SJeff Kirsher * you better get the MTU right! 1563*527a6266SJeff Kirsher */ 1564*527a6266SJeff Kirsher thresh = sky2_get_rx_threshold(sky2); 1565*527a6266SJeff Kirsher if (thresh > 0x1ff) 1566*527a6266SJeff Kirsher sky2_write32(hw, SK_REG(sky2->port, RX_GMF_CTRL_T), RX_TRUNC_OFF); 1567*527a6266SJeff Kirsher else { 1568*527a6266SJeff Kirsher sky2_write16(hw, SK_REG(sky2->port, RX_GMF_TR_THR), thresh); 1569*527a6266SJeff Kirsher sky2_write32(hw, SK_REG(sky2->port, RX_GMF_CTRL_T), RX_TRUNC_ON); 1570*527a6266SJeff Kirsher } 1571*527a6266SJeff Kirsher 1572*527a6266SJeff Kirsher /* Tell chip about available buffers */ 1573*527a6266SJeff Kirsher sky2_rx_update(sky2, rxq); 1574*527a6266SJeff Kirsher 1575*527a6266SJeff Kirsher if (hw->chip_id == CHIP_ID_YUKON_EX || 1576*527a6266SJeff Kirsher hw->chip_id == CHIP_ID_YUKON_SUPR) { 1577*527a6266SJeff Kirsher /* 1578*527a6266SJeff Kirsher * Disable flushing of non ASF packets; 1579*527a6266SJeff Kirsher * must be done after initializing the BMUs; 1580*527a6266SJeff Kirsher * drivers without ASF support should do this too, otherwise 1581*527a6266SJeff Kirsher * it may happen that they cannot run on ASF devices; 1582*527a6266SJeff Kirsher * remember that the MAC FIFO isn't reset during initialization. 1583*527a6266SJeff Kirsher */ 1584*527a6266SJeff Kirsher sky2_write32(hw, SK_REG(sky2->port, RX_GMF_CTRL_T), RX_MACSEC_FLUSH_OFF); 1585*527a6266SJeff Kirsher } 1586*527a6266SJeff Kirsher 1587*527a6266SJeff Kirsher if (hw->chip_id >= CHIP_ID_YUKON_SUPR) { 1588*527a6266SJeff Kirsher /* Enable RX Home Address & Routing Header checksum fix */ 1589*527a6266SJeff Kirsher sky2_write16(hw, SK_REG(sky2->port, RX_GMF_FL_CTRL), 1590*527a6266SJeff Kirsher RX_IPV6_SA_MOB_ENA | RX_IPV6_DA_MOB_ENA); 1591*527a6266SJeff Kirsher 1592*527a6266SJeff Kirsher /* Enable TX Home Address & Routing Header checksum fix */ 1593*527a6266SJeff Kirsher sky2_write32(hw, Q_ADDR(txqaddr[sky2->port], Q_TEST), 1594*527a6266SJeff Kirsher TBMU_TEST_HOME_ADD_FIX_EN | TBMU_TEST_ROUTING_ADD_FIX_EN); 1595*527a6266SJeff Kirsher } 1596*527a6266SJeff Kirsher } 1597*527a6266SJeff Kirsher 1598*527a6266SJeff Kirsher static int sky2_alloc_buffers(struct sky2_port *sky2) 1599*527a6266SJeff Kirsher { 1600*527a6266SJeff Kirsher struct sky2_hw *hw = sky2->hw; 1601*527a6266SJeff Kirsher 1602*527a6266SJeff Kirsher /* must be power of 2 */ 1603*527a6266SJeff Kirsher sky2->tx_le = pci_alloc_consistent(hw->pdev, 1604*527a6266SJeff Kirsher sky2->tx_ring_size * 1605*527a6266SJeff Kirsher sizeof(struct sky2_tx_le), 1606*527a6266SJeff Kirsher &sky2->tx_le_map); 1607*527a6266SJeff Kirsher if (!sky2->tx_le) 1608*527a6266SJeff Kirsher goto nomem; 1609*527a6266SJeff Kirsher 1610*527a6266SJeff Kirsher sky2->tx_ring = kcalloc(sky2->tx_ring_size, sizeof(struct tx_ring_info), 1611*527a6266SJeff Kirsher GFP_KERNEL); 1612*527a6266SJeff Kirsher if (!sky2->tx_ring) 1613*527a6266SJeff Kirsher goto nomem; 1614*527a6266SJeff Kirsher 1615*527a6266SJeff Kirsher sky2->rx_le = pci_alloc_consistent(hw->pdev, RX_LE_BYTES, 1616*527a6266SJeff Kirsher &sky2->rx_le_map); 1617*527a6266SJeff Kirsher if (!sky2->rx_le) 1618*527a6266SJeff Kirsher goto nomem; 1619*527a6266SJeff Kirsher memset(sky2->rx_le, 0, RX_LE_BYTES); 1620*527a6266SJeff Kirsher 1621*527a6266SJeff Kirsher sky2->rx_ring = kcalloc(sky2->rx_pending, sizeof(struct rx_ring_info), 1622*527a6266SJeff Kirsher GFP_KERNEL); 1623*527a6266SJeff Kirsher if (!sky2->rx_ring) 1624*527a6266SJeff Kirsher goto nomem; 1625*527a6266SJeff Kirsher 1626*527a6266SJeff Kirsher return sky2_alloc_rx_skbs(sky2); 1627*527a6266SJeff Kirsher nomem: 1628*527a6266SJeff Kirsher return -ENOMEM; 1629*527a6266SJeff Kirsher } 1630*527a6266SJeff Kirsher 1631*527a6266SJeff Kirsher static void sky2_free_buffers(struct sky2_port *sky2) 1632*527a6266SJeff Kirsher { 1633*527a6266SJeff Kirsher struct sky2_hw *hw = sky2->hw; 1634*527a6266SJeff Kirsher 1635*527a6266SJeff Kirsher sky2_rx_clean(sky2); 1636*527a6266SJeff Kirsher 1637*527a6266SJeff Kirsher if (sky2->rx_le) { 1638*527a6266SJeff Kirsher pci_free_consistent(hw->pdev, RX_LE_BYTES, 1639*527a6266SJeff Kirsher sky2->rx_le, sky2->rx_le_map); 1640*527a6266SJeff Kirsher sky2->rx_le = NULL; 1641*527a6266SJeff Kirsher } 1642*527a6266SJeff Kirsher if (sky2->tx_le) { 1643*527a6266SJeff Kirsher pci_free_consistent(hw->pdev, 1644*527a6266SJeff Kirsher sky2->tx_ring_size * sizeof(struct sky2_tx_le), 1645*527a6266SJeff Kirsher sky2->tx_le, sky2->tx_le_map); 1646*527a6266SJeff Kirsher sky2->tx_le = NULL; 1647*527a6266SJeff Kirsher } 1648*527a6266SJeff Kirsher kfree(sky2->tx_ring); 1649*527a6266SJeff Kirsher kfree(sky2->rx_ring); 1650*527a6266SJeff Kirsher 1651*527a6266SJeff Kirsher sky2->tx_ring = NULL; 1652*527a6266SJeff Kirsher sky2->rx_ring = NULL; 1653*527a6266SJeff Kirsher } 1654*527a6266SJeff Kirsher 1655*527a6266SJeff Kirsher static void sky2_hw_up(struct sky2_port *sky2) 1656*527a6266SJeff Kirsher { 1657*527a6266SJeff Kirsher struct sky2_hw *hw = sky2->hw; 1658*527a6266SJeff Kirsher unsigned port = sky2->port; 1659*527a6266SJeff Kirsher u32 ramsize; 1660*527a6266SJeff Kirsher int cap; 1661*527a6266SJeff Kirsher struct net_device *otherdev = hw->dev[sky2->port^1]; 1662*527a6266SJeff Kirsher 1663*527a6266SJeff Kirsher tx_init(sky2); 1664*527a6266SJeff Kirsher 1665*527a6266SJeff Kirsher /* 1666*527a6266SJeff Kirsher * On dual port PCI-X card, there is an problem where status 1667*527a6266SJeff Kirsher * can be received out of order due to split transactions 1668*527a6266SJeff Kirsher */ 1669*527a6266SJeff Kirsher if (otherdev && netif_running(otherdev) && 1670*527a6266SJeff Kirsher (cap = pci_find_capability(hw->pdev, PCI_CAP_ID_PCIX))) { 1671*527a6266SJeff Kirsher u16 cmd; 1672*527a6266SJeff Kirsher 1673*527a6266SJeff Kirsher cmd = sky2_pci_read16(hw, cap + PCI_X_CMD); 1674*527a6266SJeff Kirsher cmd &= ~PCI_X_CMD_MAX_SPLIT; 1675*527a6266SJeff Kirsher sky2_pci_write16(hw, cap + PCI_X_CMD, cmd); 1676*527a6266SJeff Kirsher } 1677*527a6266SJeff Kirsher 1678*527a6266SJeff Kirsher sky2_mac_init(hw, port); 1679*527a6266SJeff Kirsher 1680*527a6266SJeff Kirsher /* Register is number of 4K blocks on internal RAM buffer. */ 1681*527a6266SJeff Kirsher ramsize = sky2_read8(hw, B2_E_0) * 4; 1682*527a6266SJeff Kirsher if (ramsize > 0) { 1683*527a6266SJeff Kirsher u32 rxspace; 1684*527a6266SJeff Kirsher 1685*527a6266SJeff Kirsher netdev_dbg(sky2->netdev, "ram buffer %dK\n", ramsize); 1686*527a6266SJeff Kirsher if (ramsize < 16) 1687*527a6266SJeff Kirsher rxspace = ramsize / 2; 1688*527a6266SJeff Kirsher else 1689*527a6266SJeff Kirsher rxspace = 8 + (2*(ramsize - 16))/3; 1690*527a6266SJeff Kirsher 1691*527a6266SJeff Kirsher sky2_ramset(hw, rxqaddr[port], 0, rxspace); 1692*527a6266SJeff Kirsher sky2_ramset(hw, txqaddr[port], rxspace, ramsize - rxspace); 1693*527a6266SJeff Kirsher 1694*527a6266SJeff Kirsher /* Make sure SyncQ is disabled */ 1695*527a6266SJeff Kirsher sky2_write8(hw, RB_ADDR(port == 0 ? Q_XS1 : Q_XS2, RB_CTRL), 1696*527a6266SJeff Kirsher RB_RST_SET); 1697*527a6266SJeff Kirsher } 1698*527a6266SJeff Kirsher 1699*527a6266SJeff Kirsher sky2_qset(hw, txqaddr[port]); 1700*527a6266SJeff Kirsher 1701*527a6266SJeff Kirsher /* This is copied from sk98lin 10.0.5.3; no one tells me about erratta's */ 1702*527a6266SJeff Kirsher if (hw->chip_id == CHIP_ID_YUKON_EX && hw->chip_rev == CHIP_REV_YU_EX_B0) 1703*527a6266SJeff Kirsher sky2_write32(hw, Q_ADDR(txqaddr[port], Q_TEST), F_TX_CHK_AUTO_OFF); 1704*527a6266SJeff Kirsher 1705*527a6266SJeff Kirsher /* Set almost empty threshold */ 1706*527a6266SJeff Kirsher if (hw->chip_id == CHIP_ID_YUKON_EC_U && 1707*527a6266SJeff Kirsher hw->chip_rev == CHIP_REV_YU_EC_U_A0) 1708*527a6266SJeff Kirsher sky2_write16(hw, Q_ADDR(txqaddr[port], Q_AL), ECU_TXFF_LEV); 1709*527a6266SJeff Kirsher 1710*527a6266SJeff Kirsher sky2_prefetch_init(hw, txqaddr[port], sky2->tx_le_map, 1711*527a6266SJeff Kirsher sky2->tx_ring_size - 1); 1712*527a6266SJeff Kirsher 1713*527a6266SJeff Kirsher sky2_vlan_mode(sky2->netdev, sky2->netdev->features); 1714*527a6266SJeff Kirsher netdev_update_features(sky2->netdev); 1715*527a6266SJeff Kirsher 1716*527a6266SJeff Kirsher sky2_rx_start(sky2); 1717*527a6266SJeff Kirsher } 1718*527a6266SJeff Kirsher 1719*527a6266SJeff Kirsher /* Bring up network interface. */ 1720*527a6266SJeff Kirsher static int sky2_up(struct net_device *dev) 1721*527a6266SJeff Kirsher { 1722*527a6266SJeff Kirsher struct sky2_port *sky2 = netdev_priv(dev); 1723*527a6266SJeff Kirsher struct sky2_hw *hw = sky2->hw; 1724*527a6266SJeff Kirsher unsigned port = sky2->port; 1725*527a6266SJeff Kirsher u32 imask; 1726*527a6266SJeff Kirsher int err; 1727*527a6266SJeff Kirsher 1728*527a6266SJeff Kirsher netif_carrier_off(dev); 1729*527a6266SJeff Kirsher 1730*527a6266SJeff Kirsher err = sky2_alloc_buffers(sky2); 1731*527a6266SJeff Kirsher if (err) 1732*527a6266SJeff Kirsher goto err_out; 1733*527a6266SJeff Kirsher 1734*527a6266SJeff Kirsher sky2_hw_up(sky2); 1735*527a6266SJeff Kirsher 1736*527a6266SJeff Kirsher /* Enable interrupts from phy/mac for port */ 1737*527a6266SJeff Kirsher imask = sky2_read32(hw, B0_IMSK); 1738*527a6266SJeff Kirsher imask |= portirq_msk[port]; 1739*527a6266SJeff Kirsher sky2_write32(hw, B0_IMSK, imask); 1740*527a6266SJeff Kirsher sky2_read32(hw, B0_IMSK); 1741*527a6266SJeff Kirsher 1742*527a6266SJeff Kirsher netif_info(sky2, ifup, dev, "enabling interface\n"); 1743*527a6266SJeff Kirsher 1744*527a6266SJeff Kirsher return 0; 1745*527a6266SJeff Kirsher 1746*527a6266SJeff Kirsher err_out: 1747*527a6266SJeff Kirsher sky2_free_buffers(sky2); 1748*527a6266SJeff Kirsher return err; 1749*527a6266SJeff Kirsher } 1750*527a6266SJeff Kirsher 1751*527a6266SJeff Kirsher /* Modular subtraction in ring */ 1752*527a6266SJeff Kirsher static inline int tx_inuse(const struct sky2_port *sky2) 1753*527a6266SJeff Kirsher { 1754*527a6266SJeff Kirsher return (sky2->tx_prod - sky2->tx_cons) & (sky2->tx_ring_size - 1); 1755*527a6266SJeff Kirsher } 1756*527a6266SJeff Kirsher 1757*527a6266SJeff Kirsher /* Number of list elements available for next tx */ 1758*527a6266SJeff Kirsher static inline int tx_avail(const struct sky2_port *sky2) 1759*527a6266SJeff Kirsher { 1760*527a6266SJeff Kirsher return sky2->tx_pending - tx_inuse(sky2); 1761*527a6266SJeff Kirsher } 1762*527a6266SJeff Kirsher 1763*527a6266SJeff Kirsher /* Estimate of number of transmit list elements required */ 1764*527a6266SJeff Kirsher static unsigned tx_le_req(const struct sk_buff *skb) 1765*527a6266SJeff Kirsher { 1766*527a6266SJeff Kirsher unsigned count; 1767*527a6266SJeff Kirsher 1768*527a6266SJeff Kirsher count = (skb_shinfo(skb)->nr_frags + 1) 1769*527a6266SJeff Kirsher * (sizeof(dma_addr_t) / sizeof(u32)); 1770*527a6266SJeff Kirsher 1771*527a6266SJeff Kirsher if (skb_is_gso(skb)) 1772*527a6266SJeff Kirsher ++count; 1773*527a6266SJeff Kirsher else if (sizeof(dma_addr_t) == sizeof(u32)) 1774*527a6266SJeff Kirsher ++count; /* possible vlan */ 1775*527a6266SJeff Kirsher 1776*527a6266SJeff Kirsher if (skb->ip_summed == CHECKSUM_PARTIAL) 1777*527a6266SJeff Kirsher ++count; 1778*527a6266SJeff Kirsher 1779*527a6266SJeff Kirsher return count; 1780*527a6266SJeff Kirsher } 1781*527a6266SJeff Kirsher 1782*527a6266SJeff Kirsher static void sky2_tx_unmap(struct pci_dev *pdev, struct tx_ring_info *re) 1783*527a6266SJeff Kirsher { 1784*527a6266SJeff Kirsher if (re->flags & TX_MAP_SINGLE) 1785*527a6266SJeff Kirsher pci_unmap_single(pdev, dma_unmap_addr(re, mapaddr), 1786*527a6266SJeff Kirsher dma_unmap_len(re, maplen), 1787*527a6266SJeff Kirsher PCI_DMA_TODEVICE); 1788*527a6266SJeff Kirsher else if (re->flags & TX_MAP_PAGE) 1789*527a6266SJeff Kirsher pci_unmap_page(pdev, dma_unmap_addr(re, mapaddr), 1790*527a6266SJeff Kirsher dma_unmap_len(re, maplen), 1791*527a6266SJeff Kirsher PCI_DMA_TODEVICE); 1792*527a6266SJeff Kirsher re->flags = 0; 1793*527a6266SJeff Kirsher } 1794*527a6266SJeff Kirsher 1795*527a6266SJeff Kirsher /* 1796*527a6266SJeff Kirsher * Put one packet in ring for transmit. 1797*527a6266SJeff Kirsher * A single packet can generate multiple list elements, and 1798*527a6266SJeff Kirsher * the number of ring elements will probably be less than the number 1799*527a6266SJeff Kirsher * of list elements used. 1800*527a6266SJeff Kirsher */ 1801*527a6266SJeff Kirsher static netdev_tx_t sky2_xmit_frame(struct sk_buff *skb, 1802*527a6266SJeff Kirsher struct net_device *dev) 1803*527a6266SJeff Kirsher { 1804*527a6266SJeff Kirsher struct sky2_port *sky2 = netdev_priv(dev); 1805*527a6266SJeff Kirsher struct sky2_hw *hw = sky2->hw; 1806*527a6266SJeff Kirsher struct sky2_tx_le *le = NULL; 1807*527a6266SJeff Kirsher struct tx_ring_info *re; 1808*527a6266SJeff Kirsher unsigned i, len; 1809*527a6266SJeff Kirsher dma_addr_t mapping; 1810*527a6266SJeff Kirsher u32 upper; 1811*527a6266SJeff Kirsher u16 slot; 1812*527a6266SJeff Kirsher u16 mss; 1813*527a6266SJeff Kirsher u8 ctrl; 1814*527a6266SJeff Kirsher 1815*527a6266SJeff Kirsher if (unlikely(tx_avail(sky2) < tx_le_req(skb))) 1816*527a6266SJeff Kirsher return NETDEV_TX_BUSY; 1817*527a6266SJeff Kirsher 1818*527a6266SJeff Kirsher len = skb_headlen(skb); 1819*527a6266SJeff Kirsher mapping = pci_map_single(hw->pdev, skb->data, len, PCI_DMA_TODEVICE); 1820*527a6266SJeff Kirsher 1821*527a6266SJeff Kirsher if (pci_dma_mapping_error(hw->pdev, mapping)) 1822*527a6266SJeff Kirsher goto mapping_error; 1823*527a6266SJeff Kirsher 1824*527a6266SJeff Kirsher slot = sky2->tx_prod; 1825*527a6266SJeff Kirsher netif_printk(sky2, tx_queued, KERN_DEBUG, dev, 1826*527a6266SJeff Kirsher "tx queued, slot %u, len %d\n", slot, skb->len); 1827*527a6266SJeff Kirsher 1828*527a6266SJeff Kirsher /* Send high bits if needed */ 1829*527a6266SJeff Kirsher upper = upper_32_bits(mapping); 1830*527a6266SJeff Kirsher if (upper != sky2->tx_last_upper) { 1831*527a6266SJeff Kirsher le = get_tx_le(sky2, &slot); 1832*527a6266SJeff Kirsher le->addr = cpu_to_le32(upper); 1833*527a6266SJeff Kirsher sky2->tx_last_upper = upper; 1834*527a6266SJeff Kirsher le->opcode = OP_ADDR64 | HW_OWNER; 1835*527a6266SJeff Kirsher } 1836*527a6266SJeff Kirsher 1837*527a6266SJeff Kirsher /* Check for TCP Segmentation Offload */ 1838*527a6266SJeff Kirsher mss = skb_shinfo(skb)->gso_size; 1839*527a6266SJeff Kirsher if (mss != 0) { 1840*527a6266SJeff Kirsher 1841*527a6266SJeff Kirsher if (!(hw->flags & SKY2_HW_NEW_LE)) 1842*527a6266SJeff Kirsher mss += ETH_HLEN + ip_hdrlen(skb) + tcp_hdrlen(skb); 1843*527a6266SJeff Kirsher 1844*527a6266SJeff Kirsher if (mss != sky2->tx_last_mss) { 1845*527a6266SJeff Kirsher le = get_tx_le(sky2, &slot); 1846*527a6266SJeff Kirsher le->addr = cpu_to_le32(mss); 1847*527a6266SJeff Kirsher 1848*527a6266SJeff Kirsher if (hw->flags & SKY2_HW_NEW_LE) 1849*527a6266SJeff Kirsher le->opcode = OP_MSS | HW_OWNER; 1850*527a6266SJeff Kirsher else 1851*527a6266SJeff Kirsher le->opcode = OP_LRGLEN | HW_OWNER; 1852*527a6266SJeff Kirsher sky2->tx_last_mss = mss; 1853*527a6266SJeff Kirsher } 1854*527a6266SJeff Kirsher } 1855*527a6266SJeff Kirsher 1856*527a6266SJeff Kirsher ctrl = 0; 1857*527a6266SJeff Kirsher 1858*527a6266SJeff Kirsher /* Add VLAN tag, can piggyback on LRGLEN or ADDR64 */ 1859*527a6266SJeff Kirsher if (vlan_tx_tag_present(skb)) { 1860*527a6266SJeff Kirsher if (!le) { 1861*527a6266SJeff Kirsher le = get_tx_le(sky2, &slot); 1862*527a6266SJeff Kirsher le->addr = 0; 1863*527a6266SJeff Kirsher le->opcode = OP_VLAN|HW_OWNER; 1864*527a6266SJeff Kirsher } else 1865*527a6266SJeff Kirsher le->opcode |= OP_VLAN; 1866*527a6266SJeff Kirsher le->length = cpu_to_be16(vlan_tx_tag_get(skb)); 1867*527a6266SJeff Kirsher ctrl |= INS_VLAN; 1868*527a6266SJeff Kirsher } 1869*527a6266SJeff Kirsher 1870*527a6266SJeff Kirsher /* Handle TCP checksum offload */ 1871*527a6266SJeff Kirsher if (skb->ip_summed == CHECKSUM_PARTIAL) { 1872*527a6266SJeff Kirsher /* On Yukon EX (some versions) encoding change. */ 1873*527a6266SJeff Kirsher if (hw->flags & SKY2_HW_AUTO_TX_SUM) 1874*527a6266SJeff Kirsher ctrl |= CALSUM; /* auto checksum */ 1875*527a6266SJeff Kirsher else { 1876*527a6266SJeff Kirsher const unsigned offset = skb_transport_offset(skb); 1877*527a6266SJeff Kirsher u32 tcpsum; 1878*527a6266SJeff Kirsher 1879*527a6266SJeff Kirsher tcpsum = offset << 16; /* sum start */ 1880*527a6266SJeff Kirsher tcpsum |= offset + skb->csum_offset; /* sum write */ 1881*527a6266SJeff Kirsher 1882*527a6266SJeff Kirsher ctrl |= CALSUM | WR_SUM | INIT_SUM | LOCK_SUM; 1883*527a6266SJeff Kirsher if (ip_hdr(skb)->protocol == IPPROTO_UDP) 1884*527a6266SJeff Kirsher ctrl |= UDPTCP; 1885*527a6266SJeff Kirsher 1886*527a6266SJeff Kirsher if (tcpsum != sky2->tx_tcpsum) { 1887*527a6266SJeff Kirsher sky2->tx_tcpsum = tcpsum; 1888*527a6266SJeff Kirsher 1889*527a6266SJeff Kirsher le = get_tx_le(sky2, &slot); 1890*527a6266SJeff Kirsher le->addr = cpu_to_le32(tcpsum); 1891*527a6266SJeff Kirsher le->length = 0; /* initial checksum value */ 1892*527a6266SJeff Kirsher le->ctrl = 1; /* one packet */ 1893*527a6266SJeff Kirsher le->opcode = OP_TCPLISW | HW_OWNER; 1894*527a6266SJeff Kirsher } 1895*527a6266SJeff Kirsher } 1896*527a6266SJeff Kirsher } 1897*527a6266SJeff Kirsher 1898*527a6266SJeff Kirsher re = sky2->tx_ring + slot; 1899*527a6266SJeff Kirsher re->flags = TX_MAP_SINGLE; 1900*527a6266SJeff Kirsher dma_unmap_addr_set(re, mapaddr, mapping); 1901*527a6266SJeff Kirsher dma_unmap_len_set(re, maplen, len); 1902*527a6266SJeff Kirsher 1903*527a6266SJeff Kirsher le = get_tx_le(sky2, &slot); 1904*527a6266SJeff Kirsher le->addr = cpu_to_le32(lower_32_bits(mapping)); 1905*527a6266SJeff Kirsher le->length = cpu_to_le16(len); 1906*527a6266SJeff Kirsher le->ctrl = ctrl; 1907*527a6266SJeff Kirsher le->opcode = mss ? (OP_LARGESEND | HW_OWNER) : (OP_PACKET | HW_OWNER); 1908*527a6266SJeff Kirsher 1909*527a6266SJeff Kirsher 1910*527a6266SJeff Kirsher for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { 1911*527a6266SJeff Kirsher const skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; 1912*527a6266SJeff Kirsher 1913*527a6266SJeff Kirsher mapping = pci_map_page(hw->pdev, frag->page, frag->page_offset, 1914*527a6266SJeff Kirsher frag->size, PCI_DMA_TODEVICE); 1915*527a6266SJeff Kirsher 1916*527a6266SJeff Kirsher if (pci_dma_mapping_error(hw->pdev, mapping)) 1917*527a6266SJeff Kirsher goto mapping_unwind; 1918*527a6266SJeff Kirsher 1919*527a6266SJeff Kirsher upper = upper_32_bits(mapping); 1920*527a6266SJeff Kirsher if (upper != sky2->tx_last_upper) { 1921*527a6266SJeff Kirsher le = get_tx_le(sky2, &slot); 1922*527a6266SJeff Kirsher le->addr = cpu_to_le32(upper); 1923*527a6266SJeff Kirsher sky2->tx_last_upper = upper; 1924*527a6266SJeff Kirsher le->opcode = OP_ADDR64 | HW_OWNER; 1925*527a6266SJeff Kirsher } 1926*527a6266SJeff Kirsher 1927*527a6266SJeff Kirsher re = sky2->tx_ring + slot; 1928*527a6266SJeff Kirsher re->flags = TX_MAP_PAGE; 1929*527a6266SJeff Kirsher dma_unmap_addr_set(re, mapaddr, mapping); 1930*527a6266SJeff Kirsher dma_unmap_len_set(re, maplen, frag->size); 1931*527a6266SJeff Kirsher 1932*527a6266SJeff Kirsher le = get_tx_le(sky2, &slot); 1933*527a6266SJeff Kirsher le->addr = cpu_to_le32(lower_32_bits(mapping)); 1934*527a6266SJeff Kirsher le->length = cpu_to_le16(frag->size); 1935*527a6266SJeff Kirsher le->ctrl = ctrl; 1936*527a6266SJeff Kirsher le->opcode = OP_BUFFER | HW_OWNER; 1937*527a6266SJeff Kirsher } 1938*527a6266SJeff Kirsher 1939*527a6266SJeff Kirsher re->skb = skb; 1940*527a6266SJeff Kirsher le->ctrl |= EOP; 1941*527a6266SJeff Kirsher 1942*527a6266SJeff Kirsher sky2->tx_prod = slot; 1943*527a6266SJeff Kirsher 1944*527a6266SJeff Kirsher if (tx_avail(sky2) <= MAX_SKB_TX_LE) 1945*527a6266SJeff Kirsher netif_stop_queue(dev); 1946*527a6266SJeff Kirsher 1947*527a6266SJeff Kirsher sky2_put_idx(hw, txqaddr[sky2->port], sky2->tx_prod); 1948*527a6266SJeff Kirsher 1949*527a6266SJeff Kirsher return NETDEV_TX_OK; 1950*527a6266SJeff Kirsher 1951*527a6266SJeff Kirsher mapping_unwind: 1952*527a6266SJeff Kirsher for (i = sky2->tx_prod; i != slot; i = RING_NEXT(i, sky2->tx_ring_size)) { 1953*527a6266SJeff Kirsher re = sky2->tx_ring + i; 1954*527a6266SJeff Kirsher 1955*527a6266SJeff Kirsher sky2_tx_unmap(hw->pdev, re); 1956*527a6266SJeff Kirsher } 1957*527a6266SJeff Kirsher 1958*527a6266SJeff Kirsher mapping_error: 1959*527a6266SJeff Kirsher if (net_ratelimit()) 1960*527a6266SJeff Kirsher dev_warn(&hw->pdev->dev, "%s: tx mapping error\n", dev->name); 1961*527a6266SJeff Kirsher dev_kfree_skb(skb); 1962*527a6266SJeff Kirsher return NETDEV_TX_OK; 1963*527a6266SJeff Kirsher } 1964*527a6266SJeff Kirsher 1965*527a6266SJeff Kirsher /* 1966*527a6266SJeff Kirsher * Free ring elements from starting at tx_cons until "done" 1967*527a6266SJeff Kirsher * 1968*527a6266SJeff Kirsher * NB: 1969*527a6266SJeff Kirsher * 1. The hardware will tell us about partial completion of multi-part 1970*527a6266SJeff Kirsher * buffers so make sure not to free skb to early. 1971*527a6266SJeff Kirsher * 2. This may run in parallel start_xmit because the it only 1972*527a6266SJeff Kirsher * looks at the tail of the queue of FIFO (tx_cons), not 1973*527a6266SJeff Kirsher * the head (tx_prod) 1974*527a6266SJeff Kirsher */ 1975*527a6266SJeff Kirsher static void sky2_tx_complete(struct sky2_port *sky2, u16 done) 1976*527a6266SJeff Kirsher { 1977*527a6266SJeff Kirsher struct net_device *dev = sky2->netdev; 1978*527a6266SJeff Kirsher unsigned idx; 1979*527a6266SJeff Kirsher 1980*527a6266SJeff Kirsher BUG_ON(done >= sky2->tx_ring_size); 1981*527a6266SJeff Kirsher 1982*527a6266SJeff Kirsher for (idx = sky2->tx_cons; idx != done; 1983*527a6266SJeff Kirsher idx = RING_NEXT(idx, sky2->tx_ring_size)) { 1984*527a6266SJeff Kirsher struct tx_ring_info *re = sky2->tx_ring + idx; 1985*527a6266SJeff Kirsher struct sk_buff *skb = re->skb; 1986*527a6266SJeff Kirsher 1987*527a6266SJeff Kirsher sky2_tx_unmap(sky2->hw->pdev, re); 1988*527a6266SJeff Kirsher 1989*527a6266SJeff Kirsher if (skb) { 1990*527a6266SJeff Kirsher netif_printk(sky2, tx_done, KERN_DEBUG, dev, 1991*527a6266SJeff Kirsher "tx done %u\n", idx); 1992*527a6266SJeff Kirsher 1993*527a6266SJeff Kirsher u64_stats_update_begin(&sky2->tx_stats.syncp); 1994*527a6266SJeff Kirsher ++sky2->tx_stats.packets; 1995*527a6266SJeff Kirsher sky2->tx_stats.bytes += skb->len; 1996*527a6266SJeff Kirsher u64_stats_update_end(&sky2->tx_stats.syncp); 1997*527a6266SJeff Kirsher 1998*527a6266SJeff Kirsher re->skb = NULL; 1999*527a6266SJeff Kirsher dev_kfree_skb_any(skb); 2000*527a6266SJeff Kirsher 2001*527a6266SJeff Kirsher sky2->tx_next = RING_NEXT(idx, sky2->tx_ring_size); 2002*527a6266SJeff Kirsher } 2003*527a6266SJeff Kirsher } 2004*527a6266SJeff Kirsher 2005*527a6266SJeff Kirsher sky2->tx_cons = idx; 2006*527a6266SJeff Kirsher smp_mb(); 2007*527a6266SJeff Kirsher } 2008*527a6266SJeff Kirsher 2009*527a6266SJeff Kirsher static void sky2_tx_reset(struct sky2_hw *hw, unsigned port) 2010*527a6266SJeff Kirsher { 2011*527a6266SJeff Kirsher /* Disable Force Sync bit and Enable Alloc bit */ 2012*527a6266SJeff Kirsher sky2_write8(hw, SK_REG(port, TXA_CTRL), 2013*527a6266SJeff Kirsher TXA_DIS_FSYNC | TXA_DIS_ALLOC | TXA_STOP_RC); 2014*527a6266SJeff Kirsher 2015*527a6266SJeff Kirsher /* Stop Interval Timer and Limit Counter of Tx Arbiter */ 2016*527a6266SJeff Kirsher sky2_write32(hw, SK_REG(port, TXA_ITI_INI), 0L); 2017*527a6266SJeff Kirsher sky2_write32(hw, SK_REG(port, TXA_LIM_INI), 0L); 2018*527a6266SJeff Kirsher 2019*527a6266SJeff Kirsher /* Reset the PCI FIFO of the async Tx queue */ 2020*527a6266SJeff Kirsher sky2_write32(hw, Q_ADDR(txqaddr[port], Q_CSR), 2021*527a6266SJeff Kirsher BMU_RST_SET | BMU_FIFO_RST); 2022*527a6266SJeff Kirsher 2023*527a6266SJeff Kirsher /* Reset the Tx prefetch units */ 2024*527a6266SJeff Kirsher sky2_write32(hw, Y2_QADDR(txqaddr[port], PREF_UNIT_CTRL), 2025*527a6266SJeff Kirsher PREF_UNIT_RST_SET); 2026*527a6266SJeff Kirsher 2027*527a6266SJeff Kirsher sky2_write32(hw, RB_ADDR(txqaddr[port], RB_CTRL), RB_RST_SET); 2028*527a6266SJeff Kirsher sky2_write8(hw, SK_REG(port, TX_GMF_CTRL_T), GMF_RST_SET); 2029*527a6266SJeff Kirsher } 2030*527a6266SJeff Kirsher 2031*527a6266SJeff Kirsher static void sky2_hw_down(struct sky2_port *sky2) 2032*527a6266SJeff Kirsher { 2033*527a6266SJeff Kirsher struct sky2_hw *hw = sky2->hw; 2034*527a6266SJeff Kirsher unsigned port = sky2->port; 2035*527a6266SJeff Kirsher u16 ctrl; 2036*527a6266SJeff Kirsher 2037*527a6266SJeff Kirsher /* Force flow control off */ 2038*527a6266SJeff Kirsher sky2_write8(hw, SK_REG(port, GMAC_CTRL), GMC_PAUSE_OFF); 2039*527a6266SJeff Kirsher 2040*527a6266SJeff Kirsher /* Stop transmitter */ 2041*527a6266SJeff Kirsher sky2_write32(hw, Q_ADDR(txqaddr[port], Q_CSR), BMU_STOP); 2042*527a6266SJeff Kirsher sky2_read32(hw, Q_ADDR(txqaddr[port], Q_CSR)); 2043*527a6266SJeff Kirsher 2044*527a6266SJeff Kirsher sky2_write32(hw, RB_ADDR(txqaddr[port], RB_CTRL), 2045*527a6266SJeff Kirsher RB_RST_SET | RB_DIS_OP_MD); 2046*527a6266SJeff Kirsher 2047*527a6266SJeff Kirsher ctrl = gma_read16(hw, port, GM_GP_CTRL); 2048*527a6266SJeff Kirsher ctrl &= ~(GM_GPCR_TX_ENA | GM_GPCR_RX_ENA); 2049*527a6266SJeff Kirsher gma_write16(hw, port, GM_GP_CTRL, ctrl); 2050*527a6266SJeff Kirsher 2051*527a6266SJeff Kirsher sky2_write8(hw, SK_REG(port, GPHY_CTRL), GPC_RST_SET); 2052*527a6266SJeff Kirsher 2053*527a6266SJeff Kirsher /* Workaround shared GMAC reset */ 2054*527a6266SJeff Kirsher if (!(hw->chip_id == CHIP_ID_YUKON_XL && hw->chip_rev == 0 && 2055*527a6266SJeff Kirsher port == 0 && hw->dev[1] && netif_running(hw->dev[1]))) 2056*527a6266SJeff Kirsher sky2_write8(hw, SK_REG(port, GMAC_CTRL), GMC_RST_SET); 2057*527a6266SJeff Kirsher 2058*527a6266SJeff Kirsher sky2_write8(hw, SK_REG(port, RX_GMF_CTRL_T), GMF_RST_SET); 2059*527a6266SJeff Kirsher 2060*527a6266SJeff Kirsher /* Force any delayed status interrrupt and NAPI */ 2061*527a6266SJeff Kirsher sky2_write32(hw, STAT_LEV_TIMER_CNT, 0); 2062*527a6266SJeff Kirsher sky2_write32(hw, STAT_TX_TIMER_CNT, 0); 2063*527a6266SJeff Kirsher sky2_write32(hw, STAT_ISR_TIMER_CNT, 0); 2064*527a6266SJeff Kirsher sky2_read8(hw, STAT_ISR_TIMER_CTRL); 2065*527a6266SJeff Kirsher 2066*527a6266SJeff Kirsher sky2_rx_stop(sky2); 2067*527a6266SJeff Kirsher 2068*527a6266SJeff Kirsher spin_lock_bh(&sky2->phy_lock); 2069*527a6266SJeff Kirsher sky2_phy_power_down(hw, port); 2070*527a6266SJeff Kirsher spin_unlock_bh(&sky2->phy_lock); 2071*527a6266SJeff Kirsher 2072*527a6266SJeff Kirsher sky2_tx_reset(hw, port); 2073*527a6266SJeff Kirsher 2074*527a6266SJeff Kirsher /* Free any pending frames stuck in HW queue */ 2075*527a6266SJeff Kirsher sky2_tx_complete(sky2, sky2->tx_prod); 2076*527a6266SJeff Kirsher } 2077*527a6266SJeff Kirsher 2078*527a6266SJeff Kirsher /* Network shutdown */ 2079*527a6266SJeff Kirsher static int sky2_down(struct net_device *dev) 2080*527a6266SJeff Kirsher { 2081*527a6266SJeff Kirsher struct sky2_port *sky2 = netdev_priv(dev); 2082*527a6266SJeff Kirsher struct sky2_hw *hw = sky2->hw; 2083*527a6266SJeff Kirsher 2084*527a6266SJeff Kirsher /* Never really got started! */ 2085*527a6266SJeff Kirsher if (!sky2->tx_le) 2086*527a6266SJeff Kirsher return 0; 2087*527a6266SJeff Kirsher 2088*527a6266SJeff Kirsher netif_info(sky2, ifdown, dev, "disabling interface\n"); 2089*527a6266SJeff Kirsher 2090*527a6266SJeff Kirsher /* Disable port IRQ */ 2091*527a6266SJeff Kirsher sky2_write32(hw, B0_IMSK, 2092*527a6266SJeff Kirsher sky2_read32(hw, B0_IMSK) & ~portirq_msk[sky2->port]); 2093*527a6266SJeff Kirsher sky2_read32(hw, B0_IMSK); 2094*527a6266SJeff Kirsher 2095*527a6266SJeff Kirsher synchronize_irq(hw->pdev->irq); 2096*527a6266SJeff Kirsher napi_synchronize(&hw->napi); 2097*527a6266SJeff Kirsher 2098*527a6266SJeff Kirsher sky2_hw_down(sky2); 2099*527a6266SJeff Kirsher 2100*527a6266SJeff Kirsher sky2_free_buffers(sky2); 2101*527a6266SJeff Kirsher 2102*527a6266SJeff Kirsher return 0; 2103*527a6266SJeff Kirsher } 2104*527a6266SJeff Kirsher 2105*527a6266SJeff Kirsher static u16 sky2_phy_speed(const struct sky2_hw *hw, u16 aux) 2106*527a6266SJeff Kirsher { 2107*527a6266SJeff Kirsher if (hw->flags & SKY2_HW_FIBRE_PHY) 2108*527a6266SJeff Kirsher return SPEED_1000; 2109*527a6266SJeff Kirsher 2110*527a6266SJeff Kirsher if (!(hw->flags & SKY2_HW_GIGABIT)) { 2111*527a6266SJeff Kirsher if (aux & PHY_M_PS_SPEED_100) 2112*527a6266SJeff Kirsher return SPEED_100; 2113*527a6266SJeff Kirsher else 2114*527a6266SJeff Kirsher return SPEED_10; 2115*527a6266SJeff Kirsher } 2116*527a6266SJeff Kirsher 2117*527a6266SJeff Kirsher switch (aux & PHY_M_PS_SPEED_MSK) { 2118*527a6266SJeff Kirsher case PHY_M_PS_SPEED_1000: 2119*527a6266SJeff Kirsher return SPEED_1000; 2120*527a6266SJeff Kirsher case PHY_M_PS_SPEED_100: 2121*527a6266SJeff Kirsher return SPEED_100; 2122*527a6266SJeff Kirsher default: 2123*527a6266SJeff Kirsher return SPEED_10; 2124*527a6266SJeff Kirsher } 2125*527a6266SJeff Kirsher } 2126*527a6266SJeff Kirsher 2127*527a6266SJeff Kirsher static void sky2_link_up(struct sky2_port *sky2) 2128*527a6266SJeff Kirsher { 2129*527a6266SJeff Kirsher struct sky2_hw *hw = sky2->hw; 2130*527a6266SJeff Kirsher unsigned port = sky2->port; 2131*527a6266SJeff Kirsher static const char *fc_name[] = { 2132*527a6266SJeff Kirsher [FC_NONE] = "none", 2133*527a6266SJeff Kirsher [FC_TX] = "tx", 2134*527a6266SJeff Kirsher [FC_RX] = "rx", 2135*527a6266SJeff Kirsher [FC_BOTH] = "both", 2136*527a6266SJeff Kirsher }; 2137*527a6266SJeff Kirsher 2138*527a6266SJeff Kirsher sky2_set_ipg(sky2); 2139*527a6266SJeff Kirsher 2140*527a6266SJeff Kirsher sky2_enable_rx_tx(sky2); 2141*527a6266SJeff Kirsher 2142*527a6266SJeff Kirsher gm_phy_write(hw, port, PHY_MARV_INT_MASK, PHY_M_DEF_MSK); 2143*527a6266SJeff Kirsher 2144*527a6266SJeff Kirsher netif_carrier_on(sky2->netdev); 2145*527a6266SJeff Kirsher 2146*527a6266SJeff Kirsher mod_timer(&hw->watchdog_timer, jiffies + 1); 2147*527a6266SJeff Kirsher 2148*527a6266SJeff Kirsher /* Turn on link LED */ 2149*527a6266SJeff Kirsher sky2_write8(hw, SK_REG(port, LNK_LED_REG), 2150*527a6266SJeff Kirsher LINKLED_ON | LINKLED_BLINK_OFF | LINKLED_LINKSYNC_OFF); 2151*527a6266SJeff Kirsher 2152*527a6266SJeff Kirsher netif_info(sky2, link, sky2->netdev, 2153*527a6266SJeff Kirsher "Link is up at %d Mbps, %s duplex, flow control %s\n", 2154*527a6266SJeff Kirsher sky2->speed, 2155*527a6266SJeff Kirsher sky2->duplex == DUPLEX_FULL ? "full" : "half", 2156*527a6266SJeff Kirsher fc_name[sky2->flow_status]); 2157*527a6266SJeff Kirsher } 2158*527a6266SJeff Kirsher 2159*527a6266SJeff Kirsher static void sky2_link_down(struct sky2_port *sky2) 2160*527a6266SJeff Kirsher { 2161*527a6266SJeff Kirsher struct sky2_hw *hw = sky2->hw; 2162*527a6266SJeff Kirsher unsigned port = sky2->port; 2163*527a6266SJeff Kirsher u16 reg; 2164*527a6266SJeff Kirsher 2165*527a6266SJeff Kirsher gm_phy_write(hw, port, PHY_MARV_INT_MASK, 0); 2166*527a6266SJeff Kirsher 2167*527a6266SJeff Kirsher reg = gma_read16(hw, port, GM_GP_CTRL); 2168*527a6266SJeff Kirsher reg &= ~(GM_GPCR_RX_ENA | GM_GPCR_TX_ENA); 2169*527a6266SJeff Kirsher gma_write16(hw, port, GM_GP_CTRL, reg); 2170*527a6266SJeff Kirsher 2171*527a6266SJeff Kirsher netif_carrier_off(sky2->netdev); 2172*527a6266SJeff Kirsher 2173*527a6266SJeff Kirsher /* Turn off link LED */ 2174*527a6266SJeff Kirsher sky2_write8(hw, SK_REG(port, LNK_LED_REG), LINKLED_OFF); 2175*527a6266SJeff Kirsher 2176*527a6266SJeff Kirsher netif_info(sky2, link, sky2->netdev, "Link is down\n"); 2177*527a6266SJeff Kirsher 2178*527a6266SJeff Kirsher sky2_phy_init(hw, port); 2179*527a6266SJeff Kirsher } 2180*527a6266SJeff Kirsher 2181*527a6266SJeff Kirsher static enum flow_control sky2_flow(int rx, int tx) 2182*527a6266SJeff Kirsher { 2183*527a6266SJeff Kirsher if (rx) 2184*527a6266SJeff Kirsher return tx ? FC_BOTH : FC_RX; 2185*527a6266SJeff Kirsher else 2186*527a6266SJeff Kirsher return tx ? FC_TX : FC_NONE; 2187*527a6266SJeff Kirsher } 2188*527a6266SJeff Kirsher 2189*527a6266SJeff Kirsher static int sky2_autoneg_done(struct sky2_port *sky2, u16 aux) 2190*527a6266SJeff Kirsher { 2191*527a6266SJeff Kirsher struct sky2_hw *hw = sky2->hw; 2192*527a6266SJeff Kirsher unsigned port = sky2->port; 2193*527a6266SJeff Kirsher u16 advert, lpa; 2194*527a6266SJeff Kirsher 2195*527a6266SJeff Kirsher advert = gm_phy_read(hw, port, PHY_MARV_AUNE_ADV); 2196*527a6266SJeff Kirsher lpa = gm_phy_read(hw, port, PHY_MARV_AUNE_LP); 2197*527a6266SJeff Kirsher if (lpa & PHY_M_AN_RF) { 2198*527a6266SJeff Kirsher netdev_err(sky2->netdev, "remote fault\n"); 2199*527a6266SJeff Kirsher return -1; 2200*527a6266SJeff Kirsher } 2201*527a6266SJeff Kirsher 2202*527a6266SJeff Kirsher if (!(aux & PHY_M_PS_SPDUP_RES)) { 2203*527a6266SJeff Kirsher netdev_err(sky2->netdev, "speed/duplex mismatch\n"); 2204*527a6266SJeff Kirsher return -1; 2205*527a6266SJeff Kirsher } 2206*527a6266SJeff Kirsher 2207*527a6266SJeff Kirsher sky2->speed = sky2_phy_speed(hw, aux); 2208*527a6266SJeff Kirsher sky2->duplex = (aux & PHY_M_PS_FULL_DUP) ? DUPLEX_FULL : DUPLEX_HALF; 2209*527a6266SJeff Kirsher 2210*527a6266SJeff Kirsher /* Since the pause result bits seem to in different positions on 2211*527a6266SJeff Kirsher * different chips. look at registers. 2212*527a6266SJeff Kirsher */ 2213*527a6266SJeff Kirsher if (hw->flags & SKY2_HW_FIBRE_PHY) { 2214*527a6266SJeff Kirsher /* Shift for bits in fiber PHY */ 2215*527a6266SJeff Kirsher advert &= ~(ADVERTISE_PAUSE_CAP|ADVERTISE_PAUSE_ASYM); 2216*527a6266SJeff Kirsher lpa &= ~(LPA_PAUSE_CAP|LPA_PAUSE_ASYM); 2217*527a6266SJeff Kirsher 2218*527a6266SJeff Kirsher if (advert & ADVERTISE_1000XPAUSE) 2219*527a6266SJeff Kirsher advert |= ADVERTISE_PAUSE_CAP; 2220*527a6266SJeff Kirsher if (advert & ADVERTISE_1000XPSE_ASYM) 2221*527a6266SJeff Kirsher advert |= ADVERTISE_PAUSE_ASYM; 2222*527a6266SJeff Kirsher if (lpa & LPA_1000XPAUSE) 2223*527a6266SJeff Kirsher lpa |= LPA_PAUSE_CAP; 2224*527a6266SJeff Kirsher if (lpa & LPA_1000XPAUSE_ASYM) 2225*527a6266SJeff Kirsher lpa |= LPA_PAUSE_ASYM; 2226*527a6266SJeff Kirsher } 2227*527a6266SJeff Kirsher 2228*527a6266SJeff Kirsher sky2->flow_status = FC_NONE; 2229*527a6266SJeff Kirsher if (advert & ADVERTISE_PAUSE_CAP) { 2230*527a6266SJeff Kirsher if (lpa & LPA_PAUSE_CAP) 2231*527a6266SJeff Kirsher sky2->flow_status = FC_BOTH; 2232*527a6266SJeff Kirsher else if (advert & ADVERTISE_PAUSE_ASYM) 2233*527a6266SJeff Kirsher sky2->flow_status = FC_RX; 2234*527a6266SJeff Kirsher } else if (advert & ADVERTISE_PAUSE_ASYM) { 2235*527a6266SJeff Kirsher if ((lpa & LPA_PAUSE_CAP) && (lpa & LPA_PAUSE_ASYM)) 2236*527a6266SJeff Kirsher sky2->flow_status = FC_TX; 2237*527a6266SJeff Kirsher } 2238*527a6266SJeff Kirsher 2239*527a6266SJeff Kirsher if (sky2->duplex == DUPLEX_HALF && sky2->speed < SPEED_1000 && 2240*527a6266SJeff Kirsher !(hw->chip_id == CHIP_ID_YUKON_EC_U || hw->chip_id == CHIP_ID_YUKON_EX)) 2241*527a6266SJeff Kirsher sky2->flow_status = FC_NONE; 2242*527a6266SJeff Kirsher 2243*527a6266SJeff Kirsher if (sky2->flow_status & FC_TX) 2244*527a6266SJeff Kirsher sky2_write8(hw, SK_REG(port, GMAC_CTRL), GMC_PAUSE_ON); 2245*527a6266SJeff Kirsher else 2246*527a6266SJeff Kirsher sky2_write8(hw, SK_REG(port, GMAC_CTRL), GMC_PAUSE_OFF); 2247*527a6266SJeff Kirsher 2248*527a6266SJeff Kirsher return 0; 2249*527a6266SJeff Kirsher } 2250*527a6266SJeff Kirsher 2251*527a6266SJeff Kirsher /* Interrupt from PHY */ 2252*527a6266SJeff Kirsher static void sky2_phy_intr(struct sky2_hw *hw, unsigned port) 2253*527a6266SJeff Kirsher { 2254*527a6266SJeff Kirsher struct net_device *dev = hw->dev[port]; 2255*527a6266SJeff Kirsher struct sky2_port *sky2 = netdev_priv(dev); 2256*527a6266SJeff Kirsher u16 istatus, phystat; 2257*527a6266SJeff Kirsher 2258*527a6266SJeff Kirsher if (!netif_running(dev)) 2259*527a6266SJeff Kirsher return; 2260*527a6266SJeff Kirsher 2261*527a6266SJeff Kirsher spin_lock(&sky2->phy_lock); 2262*527a6266SJeff Kirsher istatus = gm_phy_read(hw, port, PHY_MARV_INT_STAT); 2263*527a6266SJeff Kirsher phystat = gm_phy_read(hw, port, PHY_MARV_PHY_STAT); 2264*527a6266SJeff Kirsher 2265*527a6266SJeff Kirsher netif_info(sky2, intr, sky2->netdev, "phy interrupt status 0x%x 0x%x\n", 2266*527a6266SJeff Kirsher istatus, phystat); 2267*527a6266SJeff Kirsher 2268*527a6266SJeff Kirsher if (istatus & PHY_M_IS_AN_COMPL) { 2269*527a6266SJeff Kirsher if (sky2_autoneg_done(sky2, phystat) == 0 && 2270*527a6266SJeff Kirsher !netif_carrier_ok(dev)) 2271*527a6266SJeff Kirsher sky2_link_up(sky2); 2272*527a6266SJeff Kirsher goto out; 2273*527a6266SJeff Kirsher } 2274*527a6266SJeff Kirsher 2275*527a6266SJeff Kirsher if (istatus & PHY_M_IS_LSP_CHANGE) 2276*527a6266SJeff Kirsher sky2->speed = sky2_phy_speed(hw, phystat); 2277*527a6266SJeff Kirsher 2278*527a6266SJeff Kirsher if (istatus & PHY_M_IS_DUP_CHANGE) 2279*527a6266SJeff Kirsher sky2->duplex = 2280*527a6266SJeff Kirsher (phystat & PHY_M_PS_FULL_DUP) ? DUPLEX_FULL : DUPLEX_HALF; 2281*527a6266SJeff Kirsher 2282*527a6266SJeff Kirsher if (istatus & PHY_M_IS_LST_CHANGE) { 2283*527a6266SJeff Kirsher if (phystat & PHY_M_PS_LINK_UP) 2284*527a6266SJeff Kirsher sky2_link_up(sky2); 2285*527a6266SJeff Kirsher else 2286*527a6266SJeff Kirsher sky2_link_down(sky2); 2287*527a6266SJeff Kirsher } 2288*527a6266SJeff Kirsher out: 2289*527a6266SJeff Kirsher spin_unlock(&sky2->phy_lock); 2290*527a6266SJeff Kirsher } 2291*527a6266SJeff Kirsher 2292*527a6266SJeff Kirsher /* Special quick link interrupt (Yukon-2 Optima only) */ 2293*527a6266SJeff Kirsher static void sky2_qlink_intr(struct sky2_hw *hw) 2294*527a6266SJeff Kirsher { 2295*527a6266SJeff Kirsher struct sky2_port *sky2 = netdev_priv(hw->dev[0]); 2296*527a6266SJeff Kirsher u32 imask; 2297*527a6266SJeff Kirsher u16 phy; 2298*527a6266SJeff Kirsher 2299*527a6266SJeff Kirsher /* disable irq */ 2300*527a6266SJeff Kirsher imask = sky2_read32(hw, B0_IMSK); 2301*527a6266SJeff Kirsher imask &= ~Y2_IS_PHY_QLNK; 2302*527a6266SJeff Kirsher sky2_write32(hw, B0_IMSK, imask); 2303*527a6266SJeff Kirsher 2304*527a6266SJeff Kirsher /* reset PHY Link Detect */ 2305*527a6266SJeff Kirsher phy = sky2_pci_read16(hw, PSM_CONFIG_REG4); 2306*527a6266SJeff Kirsher sky2_write8(hw, B2_TST_CTRL1, TST_CFG_WRITE_ON); 2307*527a6266SJeff Kirsher sky2_pci_write16(hw, PSM_CONFIG_REG4, phy | 1); 2308*527a6266SJeff Kirsher sky2_write8(hw, B2_TST_CTRL1, TST_CFG_WRITE_OFF); 2309*527a6266SJeff Kirsher 2310*527a6266SJeff Kirsher sky2_link_up(sky2); 2311*527a6266SJeff Kirsher } 2312*527a6266SJeff Kirsher 2313*527a6266SJeff Kirsher /* Transmit timeout is only called if we are running, carrier is up 2314*527a6266SJeff Kirsher * and tx queue is full (stopped). 2315*527a6266SJeff Kirsher */ 2316*527a6266SJeff Kirsher static void sky2_tx_timeout(struct net_device *dev) 2317*527a6266SJeff Kirsher { 2318*527a6266SJeff Kirsher struct sky2_port *sky2 = netdev_priv(dev); 2319*527a6266SJeff Kirsher struct sky2_hw *hw = sky2->hw; 2320*527a6266SJeff Kirsher 2321*527a6266SJeff Kirsher netif_err(sky2, timer, dev, "tx timeout\n"); 2322*527a6266SJeff Kirsher 2323*527a6266SJeff Kirsher netdev_printk(KERN_DEBUG, dev, "transmit ring %u .. %u report=%u done=%u\n", 2324*527a6266SJeff Kirsher sky2->tx_cons, sky2->tx_prod, 2325*527a6266SJeff Kirsher sky2_read16(hw, sky2->port == 0 ? STAT_TXA1_RIDX : STAT_TXA2_RIDX), 2326*527a6266SJeff Kirsher sky2_read16(hw, Q_ADDR(txqaddr[sky2->port], Q_DONE))); 2327*527a6266SJeff Kirsher 2328*527a6266SJeff Kirsher /* can't restart safely under softirq */ 2329*527a6266SJeff Kirsher schedule_work(&hw->restart_work); 2330*527a6266SJeff Kirsher } 2331*527a6266SJeff Kirsher 2332*527a6266SJeff Kirsher static int sky2_change_mtu(struct net_device *dev, int new_mtu) 2333*527a6266SJeff Kirsher { 2334*527a6266SJeff Kirsher struct sky2_port *sky2 = netdev_priv(dev); 2335*527a6266SJeff Kirsher struct sky2_hw *hw = sky2->hw; 2336*527a6266SJeff Kirsher unsigned port = sky2->port; 2337*527a6266SJeff Kirsher int err; 2338*527a6266SJeff Kirsher u16 ctl, mode; 2339*527a6266SJeff Kirsher u32 imask; 2340*527a6266SJeff Kirsher 2341*527a6266SJeff Kirsher /* MTU size outside the spec */ 2342*527a6266SJeff Kirsher if (new_mtu < ETH_ZLEN || new_mtu > ETH_JUMBO_MTU) 2343*527a6266SJeff Kirsher return -EINVAL; 2344*527a6266SJeff Kirsher 2345*527a6266SJeff Kirsher /* MTU > 1500 on yukon FE and FE+ not allowed */ 2346*527a6266SJeff Kirsher if (new_mtu > ETH_DATA_LEN && 2347*527a6266SJeff Kirsher (hw->chip_id == CHIP_ID_YUKON_FE || 2348*527a6266SJeff Kirsher hw->chip_id == CHIP_ID_YUKON_FE_P)) 2349*527a6266SJeff Kirsher return -EINVAL; 2350*527a6266SJeff Kirsher 2351*527a6266SJeff Kirsher if (!netif_running(dev)) { 2352*527a6266SJeff Kirsher dev->mtu = new_mtu; 2353*527a6266SJeff Kirsher netdev_update_features(dev); 2354*527a6266SJeff Kirsher return 0; 2355*527a6266SJeff Kirsher } 2356*527a6266SJeff Kirsher 2357*527a6266SJeff Kirsher imask = sky2_read32(hw, B0_IMSK); 2358*527a6266SJeff Kirsher sky2_write32(hw, B0_IMSK, 0); 2359*527a6266SJeff Kirsher 2360*527a6266SJeff Kirsher dev->trans_start = jiffies; /* prevent tx timeout */ 2361*527a6266SJeff Kirsher napi_disable(&hw->napi); 2362*527a6266SJeff Kirsher netif_tx_disable(dev); 2363*527a6266SJeff Kirsher 2364*527a6266SJeff Kirsher synchronize_irq(hw->pdev->irq); 2365*527a6266SJeff Kirsher 2366*527a6266SJeff Kirsher if (!(hw->flags & SKY2_HW_RAM_BUFFER)) 2367*527a6266SJeff Kirsher sky2_set_tx_stfwd(hw, port); 2368*527a6266SJeff Kirsher 2369*527a6266SJeff Kirsher ctl = gma_read16(hw, port, GM_GP_CTRL); 2370*527a6266SJeff Kirsher gma_write16(hw, port, GM_GP_CTRL, ctl & ~GM_GPCR_RX_ENA); 2371*527a6266SJeff Kirsher sky2_rx_stop(sky2); 2372*527a6266SJeff Kirsher sky2_rx_clean(sky2); 2373*527a6266SJeff Kirsher 2374*527a6266SJeff Kirsher dev->mtu = new_mtu; 2375*527a6266SJeff Kirsher netdev_update_features(dev); 2376*527a6266SJeff Kirsher 2377*527a6266SJeff Kirsher mode = DATA_BLIND_VAL(DATA_BLIND_DEF) | GM_SMOD_VLAN_ENA; 2378*527a6266SJeff Kirsher if (sky2->speed > SPEED_100) 2379*527a6266SJeff Kirsher mode |= IPG_DATA_VAL(IPG_DATA_DEF_1000); 2380*527a6266SJeff Kirsher else 2381*527a6266SJeff Kirsher mode |= IPG_DATA_VAL(IPG_DATA_DEF_10_100); 2382*527a6266SJeff Kirsher 2383*527a6266SJeff Kirsher if (dev->mtu > ETH_DATA_LEN) 2384*527a6266SJeff Kirsher mode |= GM_SMOD_JUMBO_ENA; 2385*527a6266SJeff Kirsher 2386*527a6266SJeff Kirsher gma_write16(hw, port, GM_SERIAL_MODE, mode); 2387*527a6266SJeff Kirsher 2388*527a6266SJeff Kirsher sky2_write8(hw, RB_ADDR(rxqaddr[port], RB_CTRL), RB_ENA_OP_MD); 2389*527a6266SJeff Kirsher 2390*527a6266SJeff Kirsher err = sky2_alloc_rx_skbs(sky2); 2391*527a6266SJeff Kirsher if (!err) 2392*527a6266SJeff Kirsher sky2_rx_start(sky2); 2393*527a6266SJeff Kirsher else 2394*527a6266SJeff Kirsher sky2_rx_clean(sky2); 2395*527a6266SJeff Kirsher sky2_write32(hw, B0_IMSK, imask); 2396*527a6266SJeff Kirsher 2397*527a6266SJeff Kirsher sky2_read32(hw, B0_Y2_SP_LISR); 2398*527a6266SJeff Kirsher napi_enable(&hw->napi); 2399*527a6266SJeff Kirsher 2400*527a6266SJeff Kirsher if (err) 2401*527a6266SJeff Kirsher dev_close(dev); 2402*527a6266SJeff Kirsher else { 2403*527a6266SJeff Kirsher gma_write16(hw, port, GM_GP_CTRL, ctl); 2404*527a6266SJeff Kirsher 2405*527a6266SJeff Kirsher netif_wake_queue(dev); 2406*527a6266SJeff Kirsher } 2407*527a6266SJeff Kirsher 2408*527a6266SJeff Kirsher return err; 2409*527a6266SJeff Kirsher } 2410*527a6266SJeff Kirsher 2411*527a6266SJeff Kirsher /* For small just reuse existing skb for next receive */ 2412*527a6266SJeff Kirsher static struct sk_buff *receive_copy(struct sky2_port *sky2, 2413*527a6266SJeff Kirsher const struct rx_ring_info *re, 2414*527a6266SJeff Kirsher unsigned length) 2415*527a6266SJeff Kirsher { 2416*527a6266SJeff Kirsher struct sk_buff *skb; 2417*527a6266SJeff Kirsher 2418*527a6266SJeff Kirsher skb = netdev_alloc_skb_ip_align(sky2->netdev, length); 2419*527a6266SJeff Kirsher if (likely(skb)) { 2420*527a6266SJeff Kirsher pci_dma_sync_single_for_cpu(sky2->hw->pdev, re->data_addr, 2421*527a6266SJeff Kirsher length, PCI_DMA_FROMDEVICE); 2422*527a6266SJeff Kirsher skb_copy_from_linear_data(re->skb, skb->data, length); 2423*527a6266SJeff Kirsher skb->ip_summed = re->skb->ip_summed; 2424*527a6266SJeff Kirsher skb->csum = re->skb->csum; 2425*527a6266SJeff Kirsher pci_dma_sync_single_for_device(sky2->hw->pdev, re->data_addr, 2426*527a6266SJeff Kirsher length, PCI_DMA_FROMDEVICE); 2427*527a6266SJeff Kirsher re->skb->ip_summed = CHECKSUM_NONE; 2428*527a6266SJeff Kirsher skb_put(skb, length); 2429*527a6266SJeff Kirsher } 2430*527a6266SJeff Kirsher return skb; 2431*527a6266SJeff Kirsher } 2432*527a6266SJeff Kirsher 2433*527a6266SJeff Kirsher /* Adjust length of skb with fragments to match received data */ 2434*527a6266SJeff Kirsher static void skb_put_frags(struct sk_buff *skb, unsigned int hdr_space, 2435*527a6266SJeff Kirsher unsigned int length) 2436*527a6266SJeff Kirsher { 2437*527a6266SJeff Kirsher int i, num_frags; 2438*527a6266SJeff Kirsher unsigned int size; 2439*527a6266SJeff Kirsher 2440*527a6266SJeff Kirsher /* put header into skb */ 2441*527a6266SJeff Kirsher size = min(length, hdr_space); 2442*527a6266SJeff Kirsher skb->tail += size; 2443*527a6266SJeff Kirsher skb->len += size; 2444*527a6266SJeff Kirsher length -= size; 2445*527a6266SJeff Kirsher 2446*527a6266SJeff Kirsher num_frags = skb_shinfo(skb)->nr_frags; 2447*527a6266SJeff Kirsher for (i = 0; i < num_frags; i++) { 2448*527a6266SJeff Kirsher skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; 2449*527a6266SJeff Kirsher 2450*527a6266SJeff Kirsher if (length == 0) { 2451*527a6266SJeff Kirsher /* don't need this page */ 2452*527a6266SJeff Kirsher __free_page(frag->page); 2453*527a6266SJeff Kirsher --skb_shinfo(skb)->nr_frags; 2454*527a6266SJeff Kirsher } else { 2455*527a6266SJeff Kirsher size = min(length, (unsigned) PAGE_SIZE); 2456*527a6266SJeff Kirsher 2457*527a6266SJeff Kirsher frag->size = size; 2458*527a6266SJeff Kirsher skb->data_len += size; 2459*527a6266SJeff Kirsher skb->truesize += size; 2460*527a6266SJeff Kirsher skb->len += size; 2461*527a6266SJeff Kirsher length -= size; 2462*527a6266SJeff Kirsher } 2463*527a6266SJeff Kirsher } 2464*527a6266SJeff Kirsher } 2465*527a6266SJeff Kirsher 2466*527a6266SJeff Kirsher /* Normal packet - take skb from ring element and put in a new one */ 2467*527a6266SJeff Kirsher static struct sk_buff *receive_new(struct sky2_port *sky2, 2468*527a6266SJeff Kirsher struct rx_ring_info *re, 2469*527a6266SJeff Kirsher unsigned int length) 2470*527a6266SJeff Kirsher { 2471*527a6266SJeff Kirsher struct sk_buff *skb; 2472*527a6266SJeff Kirsher struct rx_ring_info nre; 2473*527a6266SJeff Kirsher unsigned hdr_space = sky2->rx_data_size; 2474*527a6266SJeff Kirsher 2475*527a6266SJeff Kirsher nre.skb = sky2_rx_alloc(sky2, GFP_ATOMIC); 2476*527a6266SJeff Kirsher if (unlikely(!nre.skb)) 2477*527a6266SJeff Kirsher goto nobuf; 2478*527a6266SJeff Kirsher 2479*527a6266SJeff Kirsher if (sky2_rx_map_skb(sky2->hw->pdev, &nre, hdr_space)) 2480*527a6266SJeff Kirsher goto nomap; 2481*527a6266SJeff Kirsher 2482*527a6266SJeff Kirsher skb = re->skb; 2483*527a6266SJeff Kirsher sky2_rx_unmap_skb(sky2->hw->pdev, re); 2484*527a6266SJeff Kirsher prefetch(skb->data); 2485*527a6266SJeff Kirsher *re = nre; 2486*527a6266SJeff Kirsher 2487*527a6266SJeff Kirsher if (skb_shinfo(skb)->nr_frags) 2488*527a6266SJeff Kirsher skb_put_frags(skb, hdr_space, length); 2489*527a6266SJeff Kirsher else 2490*527a6266SJeff Kirsher skb_put(skb, length); 2491*527a6266SJeff Kirsher return skb; 2492*527a6266SJeff Kirsher 2493*527a6266SJeff Kirsher nomap: 2494*527a6266SJeff Kirsher dev_kfree_skb(nre.skb); 2495*527a6266SJeff Kirsher nobuf: 2496*527a6266SJeff Kirsher return NULL; 2497*527a6266SJeff Kirsher } 2498*527a6266SJeff Kirsher 2499*527a6266SJeff Kirsher /* 2500*527a6266SJeff Kirsher * Receive one packet. 2501*527a6266SJeff Kirsher * For larger packets, get new buffer. 2502*527a6266SJeff Kirsher */ 2503*527a6266SJeff Kirsher static struct sk_buff *sky2_receive(struct net_device *dev, 2504*527a6266SJeff Kirsher u16 length, u32 status) 2505*527a6266SJeff Kirsher { 2506*527a6266SJeff Kirsher struct sky2_port *sky2 = netdev_priv(dev); 2507*527a6266SJeff Kirsher struct rx_ring_info *re = sky2->rx_ring + sky2->rx_next; 2508*527a6266SJeff Kirsher struct sk_buff *skb = NULL; 2509*527a6266SJeff Kirsher u16 count = (status & GMR_FS_LEN) >> 16; 2510*527a6266SJeff Kirsher 2511*527a6266SJeff Kirsher if (status & GMR_FS_VLAN) 2512*527a6266SJeff Kirsher count -= VLAN_HLEN; /* Account for vlan tag */ 2513*527a6266SJeff Kirsher 2514*527a6266SJeff Kirsher netif_printk(sky2, rx_status, KERN_DEBUG, dev, 2515*527a6266SJeff Kirsher "rx slot %u status 0x%x len %d\n", 2516*527a6266SJeff Kirsher sky2->rx_next, status, length); 2517*527a6266SJeff Kirsher 2518*527a6266SJeff Kirsher sky2->rx_next = (sky2->rx_next + 1) % sky2->rx_pending; 2519*527a6266SJeff Kirsher prefetch(sky2->rx_ring + sky2->rx_next); 2520*527a6266SJeff Kirsher 2521*527a6266SJeff Kirsher /* This chip has hardware problems that generates bogus status. 2522*527a6266SJeff Kirsher * So do only marginal checking and expect higher level protocols 2523*527a6266SJeff Kirsher * to handle crap frames. 2524*527a6266SJeff Kirsher */ 2525*527a6266SJeff Kirsher if (sky2->hw->chip_id == CHIP_ID_YUKON_FE_P && 2526*527a6266SJeff Kirsher sky2->hw->chip_rev == CHIP_REV_YU_FE2_A0 && 2527*527a6266SJeff Kirsher length != count) 2528*527a6266SJeff Kirsher goto okay; 2529*527a6266SJeff Kirsher 2530*527a6266SJeff Kirsher if (status & GMR_FS_ANY_ERR) 2531*527a6266SJeff Kirsher goto error; 2532*527a6266SJeff Kirsher 2533*527a6266SJeff Kirsher if (!(status & GMR_FS_RX_OK)) 2534*527a6266SJeff Kirsher goto resubmit; 2535*527a6266SJeff Kirsher 2536*527a6266SJeff Kirsher /* if length reported by DMA does not match PHY, packet was truncated */ 2537*527a6266SJeff Kirsher if (length != count) 2538*527a6266SJeff Kirsher goto error; 2539*527a6266SJeff Kirsher 2540*527a6266SJeff Kirsher okay: 2541*527a6266SJeff Kirsher if (length < copybreak) 2542*527a6266SJeff Kirsher skb = receive_copy(sky2, re, length); 2543*527a6266SJeff Kirsher else 2544*527a6266SJeff Kirsher skb = receive_new(sky2, re, length); 2545*527a6266SJeff Kirsher 2546*527a6266SJeff Kirsher dev->stats.rx_dropped += (skb == NULL); 2547*527a6266SJeff Kirsher 2548*527a6266SJeff Kirsher resubmit: 2549*527a6266SJeff Kirsher sky2_rx_submit(sky2, re); 2550*527a6266SJeff Kirsher 2551*527a6266SJeff Kirsher return skb; 2552*527a6266SJeff Kirsher 2553*527a6266SJeff Kirsher error: 2554*527a6266SJeff Kirsher ++dev->stats.rx_errors; 2555*527a6266SJeff Kirsher 2556*527a6266SJeff Kirsher if (net_ratelimit()) 2557*527a6266SJeff Kirsher netif_info(sky2, rx_err, dev, 2558*527a6266SJeff Kirsher "rx error, status 0x%x length %d\n", status, length); 2559*527a6266SJeff Kirsher 2560*527a6266SJeff Kirsher goto resubmit; 2561*527a6266SJeff Kirsher } 2562*527a6266SJeff Kirsher 2563*527a6266SJeff Kirsher /* Transmit complete */ 2564*527a6266SJeff Kirsher static inline void sky2_tx_done(struct net_device *dev, u16 last) 2565*527a6266SJeff Kirsher { 2566*527a6266SJeff Kirsher struct sky2_port *sky2 = netdev_priv(dev); 2567*527a6266SJeff Kirsher 2568*527a6266SJeff Kirsher if (netif_running(dev)) { 2569*527a6266SJeff Kirsher sky2_tx_complete(sky2, last); 2570*527a6266SJeff Kirsher 2571*527a6266SJeff Kirsher /* Wake unless it's detached, and called e.g. from sky2_down() */ 2572*527a6266SJeff Kirsher if (tx_avail(sky2) > MAX_SKB_TX_LE + 4) 2573*527a6266SJeff Kirsher netif_wake_queue(dev); 2574*527a6266SJeff Kirsher } 2575*527a6266SJeff Kirsher } 2576*527a6266SJeff Kirsher 2577*527a6266SJeff Kirsher static inline void sky2_skb_rx(const struct sky2_port *sky2, 2578*527a6266SJeff Kirsher u32 status, struct sk_buff *skb) 2579*527a6266SJeff Kirsher { 2580*527a6266SJeff Kirsher if (status & GMR_FS_VLAN) 2581*527a6266SJeff Kirsher __vlan_hwaccel_put_tag(skb, be16_to_cpu(sky2->rx_tag)); 2582*527a6266SJeff Kirsher 2583*527a6266SJeff Kirsher if (skb->ip_summed == CHECKSUM_NONE) 2584*527a6266SJeff Kirsher netif_receive_skb(skb); 2585*527a6266SJeff Kirsher else 2586*527a6266SJeff Kirsher napi_gro_receive(&sky2->hw->napi, skb); 2587*527a6266SJeff Kirsher } 2588*527a6266SJeff Kirsher 2589*527a6266SJeff Kirsher static inline void sky2_rx_done(struct sky2_hw *hw, unsigned port, 2590*527a6266SJeff Kirsher unsigned packets, unsigned bytes) 2591*527a6266SJeff Kirsher { 2592*527a6266SJeff Kirsher struct net_device *dev = hw->dev[port]; 2593*527a6266SJeff Kirsher struct sky2_port *sky2 = netdev_priv(dev); 2594*527a6266SJeff Kirsher 2595*527a6266SJeff Kirsher if (packets == 0) 2596*527a6266SJeff Kirsher return; 2597*527a6266SJeff Kirsher 2598*527a6266SJeff Kirsher u64_stats_update_begin(&sky2->rx_stats.syncp); 2599*527a6266SJeff Kirsher sky2->rx_stats.packets += packets; 2600*527a6266SJeff Kirsher sky2->rx_stats.bytes += bytes; 2601*527a6266SJeff Kirsher u64_stats_update_end(&sky2->rx_stats.syncp); 2602*527a6266SJeff Kirsher 2603*527a6266SJeff Kirsher dev->last_rx = jiffies; 2604*527a6266SJeff Kirsher sky2_rx_update(netdev_priv(dev), rxqaddr[port]); 2605*527a6266SJeff Kirsher } 2606*527a6266SJeff Kirsher 2607*527a6266SJeff Kirsher static void sky2_rx_checksum(struct sky2_port *sky2, u32 status) 2608*527a6266SJeff Kirsher { 2609*527a6266SJeff Kirsher /* If this happens then driver assuming wrong format for chip type */ 2610*527a6266SJeff Kirsher BUG_ON(sky2->hw->flags & SKY2_HW_NEW_LE); 2611*527a6266SJeff Kirsher 2612*527a6266SJeff Kirsher /* Both checksum counters are programmed to start at 2613*527a6266SJeff Kirsher * the same offset, so unless there is a problem they 2614*527a6266SJeff Kirsher * should match. This failure is an early indication that 2615*527a6266SJeff Kirsher * hardware receive checksumming won't work. 2616*527a6266SJeff Kirsher */ 2617*527a6266SJeff Kirsher if (likely((u16)(status >> 16) == (u16)status)) { 2618*527a6266SJeff Kirsher struct sk_buff *skb = sky2->rx_ring[sky2->rx_next].skb; 2619*527a6266SJeff Kirsher skb->ip_summed = CHECKSUM_COMPLETE; 2620*527a6266SJeff Kirsher skb->csum = le16_to_cpu(status); 2621*527a6266SJeff Kirsher } else { 2622*527a6266SJeff Kirsher dev_notice(&sky2->hw->pdev->dev, 2623*527a6266SJeff Kirsher "%s: receive checksum problem (status = %#x)\n", 2624*527a6266SJeff Kirsher sky2->netdev->name, status); 2625*527a6266SJeff Kirsher 2626*527a6266SJeff Kirsher /* Disable checksum offload 2627*527a6266SJeff Kirsher * It will be reenabled on next ndo_set_features, but if it's 2628*527a6266SJeff Kirsher * really broken, will get disabled again 2629*527a6266SJeff Kirsher */ 2630*527a6266SJeff Kirsher sky2->netdev->features &= ~NETIF_F_RXCSUM; 2631*527a6266SJeff Kirsher sky2_write32(sky2->hw, Q_ADDR(rxqaddr[sky2->port], Q_CSR), 2632*527a6266SJeff Kirsher BMU_DIS_RX_CHKSUM); 2633*527a6266SJeff Kirsher } 2634*527a6266SJeff Kirsher } 2635*527a6266SJeff Kirsher 2636*527a6266SJeff Kirsher static void sky2_rx_hash(struct sky2_port *sky2, u32 status) 2637*527a6266SJeff Kirsher { 2638*527a6266SJeff Kirsher struct sk_buff *skb; 2639*527a6266SJeff Kirsher 2640*527a6266SJeff Kirsher skb = sky2->rx_ring[sky2->rx_next].skb; 2641*527a6266SJeff Kirsher skb->rxhash = le32_to_cpu(status); 2642*527a6266SJeff Kirsher } 2643*527a6266SJeff Kirsher 2644*527a6266SJeff Kirsher /* Process status response ring */ 2645*527a6266SJeff Kirsher static int sky2_status_intr(struct sky2_hw *hw, int to_do, u16 idx) 2646*527a6266SJeff Kirsher { 2647*527a6266SJeff Kirsher int work_done = 0; 2648*527a6266SJeff Kirsher unsigned int total_bytes[2] = { 0 }; 2649*527a6266SJeff Kirsher unsigned int total_packets[2] = { 0 }; 2650*527a6266SJeff Kirsher 2651*527a6266SJeff Kirsher rmb(); 2652*527a6266SJeff Kirsher do { 2653*527a6266SJeff Kirsher struct sky2_port *sky2; 2654*527a6266SJeff Kirsher struct sky2_status_le *le = hw->st_le + hw->st_idx; 2655*527a6266SJeff Kirsher unsigned port; 2656*527a6266SJeff Kirsher struct net_device *dev; 2657*527a6266SJeff Kirsher struct sk_buff *skb; 2658*527a6266SJeff Kirsher u32 status; 2659*527a6266SJeff Kirsher u16 length; 2660*527a6266SJeff Kirsher u8 opcode = le->opcode; 2661*527a6266SJeff Kirsher 2662*527a6266SJeff Kirsher if (!(opcode & HW_OWNER)) 2663*527a6266SJeff Kirsher break; 2664*527a6266SJeff Kirsher 2665*527a6266SJeff Kirsher hw->st_idx = RING_NEXT(hw->st_idx, hw->st_size); 2666*527a6266SJeff Kirsher 2667*527a6266SJeff Kirsher port = le->css & CSS_LINK_BIT; 2668*527a6266SJeff Kirsher dev = hw->dev[port]; 2669*527a6266SJeff Kirsher sky2 = netdev_priv(dev); 2670*527a6266SJeff Kirsher length = le16_to_cpu(le->length); 2671*527a6266SJeff Kirsher status = le32_to_cpu(le->status); 2672*527a6266SJeff Kirsher 2673*527a6266SJeff Kirsher le->opcode = 0; 2674*527a6266SJeff Kirsher switch (opcode & ~HW_OWNER) { 2675*527a6266SJeff Kirsher case OP_RXSTAT: 2676*527a6266SJeff Kirsher total_packets[port]++; 2677*527a6266SJeff Kirsher total_bytes[port] += length; 2678*527a6266SJeff Kirsher 2679*527a6266SJeff Kirsher skb = sky2_receive(dev, length, status); 2680*527a6266SJeff Kirsher if (!skb) 2681*527a6266SJeff Kirsher break; 2682*527a6266SJeff Kirsher 2683*527a6266SJeff Kirsher /* This chip reports checksum status differently */ 2684*527a6266SJeff Kirsher if (hw->flags & SKY2_HW_NEW_LE) { 2685*527a6266SJeff Kirsher if ((dev->features & NETIF_F_RXCSUM) && 2686*527a6266SJeff Kirsher (le->css & (CSS_ISIPV4 | CSS_ISIPV6)) && 2687*527a6266SJeff Kirsher (le->css & CSS_TCPUDPCSOK)) 2688*527a6266SJeff Kirsher skb->ip_summed = CHECKSUM_UNNECESSARY; 2689*527a6266SJeff Kirsher else 2690*527a6266SJeff Kirsher skb->ip_summed = CHECKSUM_NONE; 2691*527a6266SJeff Kirsher } 2692*527a6266SJeff Kirsher 2693*527a6266SJeff Kirsher skb->protocol = eth_type_trans(skb, dev); 2694*527a6266SJeff Kirsher 2695*527a6266SJeff Kirsher sky2_skb_rx(sky2, status, skb); 2696*527a6266SJeff Kirsher 2697*527a6266SJeff Kirsher /* Stop after net poll weight */ 2698*527a6266SJeff Kirsher if (++work_done >= to_do) 2699*527a6266SJeff Kirsher goto exit_loop; 2700*527a6266SJeff Kirsher break; 2701*527a6266SJeff Kirsher 2702*527a6266SJeff Kirsher case OP_RXVLAN: 2703*527a6266SJeff Kirsher sky2->rx_tag = length; 2704*527a6266SJeff Kirsher break; 2705*527a6266SJeff Kirsher 2706*527a6266SJeff Kirsher case OP_RXCHKSVLAN: 2707*527a6266SJeff Kirsher sky2->rx_tag = length; 2708*527a6266SJeff Kirsher /* fall through */ 2709*527a6266SJeff Kirsher case OP_RXCHKS: 2710*527a6266SJeff Kirsher if (likely(dev->features & NETIF_F_RXCSUM)) 2711*527a6266SJeff Kirsher sky2_rx_checksum(sky2, status); 2712*527a6266SJeff Kirsher break; 2713*527a6266SJeff Kirsher 2714*527a6266SJeff Kirsher case OP_RSS_HASH: 2715*527a6266SJeff Kirsher sky2_rx_hash(sky2, status); 2716*527a6266SJeff Kirsher break; 2717*527a6266SJeff Kirsher 2718*527a6266SJeff Kirsher case OP_TXINDEXLE: 2719*527a6266SJeff Kirsher /* TX index reports status for both ports */ 2720*527a6266SJeff Kirsher sky2_tx_done(hw->dev[0], status & 0xfff); 2721*527a6266SJeff Kirsher if (hw->dev[1]) 2722*527a6266SJeff Kirsher sky2_tx_done(hw->dev[1], 2723*527a6266SJeff Kirsher ((status >> 24) & 0xff) 2724*527a6266SJeff Kirsher | (u16)(length & 0xf) << 8); 2725*527a6266SJeff Kirsher break; 2726*527a6266SJeff Kirsher 2727*527a6266SJeff Kirsher default: 2728*527a6266SJeff Kirsher if (net_ratelimit()) 2729*527a6266SJeff Kirsher pr_warning("unknown status opcode 0x%x\n", opcode); 2730*527a6266SJeff Kirsher } 2731*527a6266SJeff Kirsher } while (hw->st_idx != idx); 2732*527a6266SJeff Kirsher 2733*527a6266SJeff Kirsher /* Fully processed status ring so clear irq */ 2734*527a6266SJeff Kirsher sky2_write32(hw, STAT_CTRL, SC_STAT_CLR_IRQ); 2735*527a6266SJeff Kirsher 2736*527a6266SJeff Kirsher exit_loop: 2737*527a6266SJeff Kirsher sky2_rx_done(hw, 0, total_packets[0], total_bytes[0]); 2738*527a6266SJeff Kirsher sky2_rx_done(hw, 1, total_packets[1], total_bytes[1]); 2739*527a6266SJeff Kirsher 2740*527a6266SJeff Kirsher return work_done; 2741*527a6266SJeff Kirsher } 2742*527a6266SJeff Kirsher 2743*527a6266SJeff Kirsher static void sky2_hw_error(struct sky2_hw *hw, unsigned port, u32 status) 2744*527a6266SJeff Kirsher { 2745*527a6266SJeff Kirsher struct net_device *dev = hw->dev[port]; 2746*527a6266SJeff Kirsher 2747*527a6266SJeff Kirsher if (net_ratelimit()) 2748*527a6266SJeff Kirsher netdev_info(dev, "hw error interrupt status 0x%x\n", status); 2749*527a6266SJeff Kirsher 2750*527a6266SJeff Kirsher if (status & Y2_IS_PAR_RD1) { 2751*527a6266SJeff Kirsher if (net_ratelimit()) 2752*527a6266SJeff Kirsher netdev_err(dev, "ram data read parity error\n"); 2753*527a6266SJeff Kirsher /* Clear IRQ */ 2754*527a6266SJeff Kirsher sky2_write16(hw, RAM_BUFFER(port, B3_RI_CTRL), RI_CLR_RD_PERR); 2755*527a6266SJeff Kirsher } 2756*527a6266SJeff Kirsher 2757*527a6266SJeff Kirsher if (status & Y2_IS_PAR_WR1) { 2758*527a6266SJeff Kirsher if (net_ratelimit()) 2759*527a6266SJeff Kirsher netdev_err(dev, "ram data write parity error\n"); 2760*527a6266SJeff Kirsher 2761*527a6266SJeff Kirsher sky2_write16(hw, RAM_BUFFER(port, B3_RI_CTRL), RI_CLR_WR_PERR); 2762*527a6266SJeff Kirsher } 2763*527a6266SJeff Kirsher 2764*527a6266SJeff Kirsher if (status & Y2_IS_PAR_MAC1) { 2765*527a6266SJeff Kirsher if (net_ratelimit()) 2766*527a6266SJeff Kirsher netdev_err(dev, "MAC parity error\n"); 2767*527a6266SJeff Kirsher sky2_write8(hw, SK_REG(port, TX_GMF_CTRL_T), GMF_CLI_TX_PE); 2768*527a6266SJeff Kirsher } 2769*527a6266SJeff Kirsher 2770*527a6266SJeff Kirsher if (status & Y2_IS_PAR_RX1) { 2771*527a6266SJeff Kirsher if (net_ratelimit()) 2772*527a6266SJeff Kirsher netdev_err(dev, "RX parity error\n"); 2773*527a6266SJeff Kirsher sky2_write32(hw, Q_ADDR(rxqaddr[port], Q_CSR), BMU_CLR_IRQ_PAR); 2774*527a6266SJeff Kirsher } 2775*527a6266SJeff Kirsher 2776*527a6266SJeff Kirsher if (status & Y2_IS_TCP_TXA1) { 2777*527a6266SJeff Kirsher if (net_ratelimit()) 2778*527a6266SJeff Kirsher netdev_err(dev, "TCP segmentation error\n"); 2779*527a6266SJeff Kirsher sky2_write32(hw, Q_ADDR(txqaddr[port], Q_CSR), BMU_CLR_IRQ_TCP); 2780*527a6266SJeff Kirsher } 2781*527a6266SJeff Kirsher } 2782*527a6266SJeff Kirsher 2783*527a6266SJeff Kirsher static void sky2_hw_intr(struct sky2_hw *hw) 2784*527a6266SJeff Kirsher { 2785*527a6266SJeff Kirsher struct pci_dev *pdev = hw->pdev; 2786*527a6266SJeff Kirsher u32 status = sky2_read32(hw, B0_HWE_ISRC); 2787*527a6266SJeff Kirsher u32 hwmsk = sky2_read32(hw, B0_HWE_IMSK); 2788*527a6266SJeff Kirsher 2789*527a6266SJeff Kirsher status &= hwmsk; 2790*527a6266SJeff Kirsher 2791*527a6266SJeff Kirsher if (status & Y2_IS_TIST_OV) 2792*527a6266SJeff Kirsher sky2_write8(hw, GMAC_TI_ST_CTRL, GMT_ST_CLR_IRQ); 2793*527a6266SJeff Kirsher 2794*527a6266SJeff Kirsher if (status & (Y2_IS_MST_ERR | Y2_IS_IRQ_STAT)) { 2795*527a6266SJeff Kirsher u16 pci_err; 2796*527a6266SJeff Kirsher 2797*527a6266SJeff Kirsher sky2_write8(hw, B2_TST_CTRL1, TST_CFG_WRITE_ON); 2798*527a6266SJeff Kirsher pci_err = sky2_pci_read16(hw, PCI_STATUS); 2799*527a6266SJeff Kirsher if (net_ratelimit()) 2800*527a6266SJeff Kirsher dev_err(&pdev->dev, "PCI hardware error (0x%x)\n", 2801*527a6266SJeff Kirsher pci_err); 2802*527a6266SJeff Kirsher 2803*527a6266SJeff Kirsher sky2_pci_write16(hw, PCI_STATUS, 2804*527a6266SJeff Kirsher pci_err | PCI_STATUS_ERROR_BITS); 2805*527a6266SJeff Kirsher sky2_write8(hw, B2_TST_CTRL1, TST_CFG_WRITE_OFF); 2806*527a6266SJeff Kirsher } 2807*527a6266SJeff Kirsher 2808*527a6266SJeff Kirsher if (status & Y2_IS_PCI_EXP) { 2809*527a6266SJeff Kirsher /* PCI-Express uncorrectable Error occurred */ 2810*527a6266SJeff Kirsher u32 err; 2811*527a6266SJeff Kirsher 2812*527a6266SJeff Kirsher sky2_write8(hw, B2_TST_CTRL1, TST_CFG_WRITE_ON); 2813*527a6266SJeff Kirsher err = sky2_read32(hw, Y2_CFG_AER + PCI_ERR_UNCOR_STATUS); 2814*527a6266SJeff Kirsher sky2_write32(hw, Y2_CFG_AER + PCI_ERR_UNCOR_STATUS, 2815*527a6266SJeff Kirsher 0xfffffffful); 2816*527a6266SJeff Kirsher if (net_ratelimit()) 2817*527a6266SJeff Kirsher dev_err(&pdev->dev, "PCI Express error (0x%x)\n", err); 2818*527a6266SJeff Kirsher 2819*527a6266SJeff Kirsher sky2_read32(hw, Y2_CFG_AER + PCI_ERR_UNCOR_STATUS); 2820*527a6266SJeff Kirsher sky2_write8(hw, B2_TST_CTRL1, TST_CFG_WRITE_OFF); 2821*527a6266SJeff Kirsher } 2822*527a6266SJeff Kirsher 2823*527a6266SJeff Kirsher if (status & Y2_HWE_L1_MASK) 2824*527a6266SJeff Kirsher sky2_hw_error(hw, 0, status); 2825*527a6266SJeff Kirsher status >>= 8; 2826*527a6266SJeff Kirsher if (status & Y2_HWE_L1_MASK) 2827*527a6266SJeff Kirsher sky2_hw_error(hw, 1, status); 2828*527a6266SJeff Kirsher } 2829*527a6266SJeff Kirsher 2830*527a6266SJeff Kirsher static void sky2_mac_intr(struct sky2_hw *hw, unsigned port) 2831*527a6266SJeff Kirsher { 2832*527a6266SJeff Kirsher struct net_device *dev = hw->dev[port]; 2833*527a6266SJeff Kirsher struct sky2_port *sky2 = netdev_priv(dev); 2834*527a6266SJeff Kirsher u8 status = sky2_read8(hw, SK_REG(port, GMAC_IRQ_SRC)); 2835*527a6266SJeff Kirsher 2836*527a6266SJeff Kirsher netif_info(sky2, intr, dev, "mac interrupt status 0x%x\n", status); 2837*527a6266SJeff Kirsher 2838*527a6266SJeff Kirsher if (status & GM_IS_RX_CO_OV) 2839*527a6266SJeff Kirsher gma_read16(hw, port, GM_RX_IRQ_SRC); 2840*527a6266SJeff Kirsher 2841*527a6266SJeff Kirsher if (status & GM_IS_TX_CO_OV) 2842*527a6266SJeff Kirsher gma_read16(hw, port, GM_TX_IRQ_SRC); 2843*527a6266SJeff Kirsher 2844*527a6266SJeff Kirsher if (status & GM_IS_RX_FF_OR) { 2845*527a6266SJeff Kirsher ++dev->stats.rx_fifo_errors; 2846*527a6266SJeff Kirsher sky2_write8(hw, SK_REG(port, RX_GMF_CTRL_T), GMF_CLI_RX_FO); 2847*527a6266SJeff Kirsher } 2848*527a6266SJeff Kirsher 2849*527a6266SJeff Kirsher if (status & GM_IS_TX_FF_UR) { 2850*527a6266SJeff Kirsher ++dev->stats.tx_fifo_errors; 2851*527a6266SJeff Kirsher sky2_write8(hw, SK_REG(port, TX_GMF_CTRL_T), GMF_CLI_TX_FU); 2852*527a6266SJeff Kirsher } 2853*527a6266SJeff Kirsher } 2854*527a6266SJeff Kirsher 2855*527a6266SJeff Kirsher /* This should never happen it is a bug. */ 2856*527a6266SJeff Kirsher static void sky2_le_error(struct sky2_hw *hw, unsigned port, u16 q) 2857*527a6266SJeff Kirsher { 2858*527a6266SJeff Kirsher struct net_device *dev = hw->dev[port]; 2859*527a6266SJeff Kirsher u16 idx = sky2_read16(hw, Y2_QADDR(q, PREF_UNIT_GET_IDX)); 2860*527a6266SJeff Kirsher 2861*527a6266SJeff Kirsher dev_err(&hw->pdev->dev, "%s: descriptor error q=%#x get=%u put=%u\n", 2862*527a6266SJeff Kirsher dev->name, (unsigned) q, (unsigned) idx, 2863*527a6266SJeff Kirsher (unsigned) sky2_read16(hw, Y2_QADDR(q, PREF_UNIT_PUT_IDX))); 2864*527a6266SJeff Kirsher 2865*527a6266SJeff Kirsher sky2_write32(hw, Q_ADDR(q, Q_CSR), BMU_CLR_IRQ_CHK); 2866*527a6266SJeff Kirsher } 2867*527a6266SJeff Kirsher 2868*527a6266SJeff Kirsher static int sky2_rx_hung(struct net_device *dev) 2869*527a6266SJeff Kirsher { 2870*527a6266SJeff Kirsher struct sky2_port *sky2 = netdev_priv(dev); 2871*527a6266SJeff Kirsher struct sky2_hw *hw = sky2->hw; 2872*527a6266SJeff Kirsher unsigned port = sky2->port; 2873*527a6266SJeff Kirsher unsigned rxq = rxqaddr[port]; 2874*527a6266SJeff Kirsher u32 mac_rp = sky2_read32(hw, SK_REG(port, RX_GMF_RP)); 2875*527a6266SJeff Kirsher u8 mac_lev = sky2_read8(hw, SK_REG(port, RX_GMF_RLEV)); 2876*527a6266SJeff Kirsher u8 fifo_rp = sky2_read8(hw, Q_ADDR(rxq, Q_RP)); 2877*527a6266SJeff Kirsher u8 fifo_lev = sky2_read8(hw, Q_ADDR(rxq, Q_RL)); 2878*527a6266SJeff Kirsher 2879*527a6266SJeff Kirsher /* If idle and MAC or PCI is stuck */ 2880*527a6266SJeff Kirsher if (sky2->check.last == dev->last_rx && 2881*527a6266SJeff Kirsher ((mac_rp == sky2->check.mac_rp && 2882*527a6266SJeff Kirsher mac_lev != 0 && mac_lev >= sky2->check.mac_lev) || 2883*527a6266SJeff Kirsher /* Check if the PCI RX hang */ 2884*527a6266SJeff Kirsher (fifo_rp == sky2->check.fifo_rp && 2885*527a6266SJeff Kirsher fifo_lev != 0 && fifo_lev >= sky2->check.fifo_lev))) { 2886*527a6266SJeff Kirsher netdev_printk(KERN_DEBUG, dev, 2887*527a6266SJeff Kirsher "hung mac %d:%d fifo %d (%d:%d)\n", 2888*527a6266SJeff Kirsher mac_lev, mac_rp, fifo_lev, 2889*527a6266SJeff Kirsher fifo_rp, sky2_read8(hw, Q_ADDR(rxq, Q_WP))); 2890*527a6266SJeff Kirsher return 1; 2891*527a6266SJeff Kirsher } else { 2892*527a6266SJeff Kirsher sky2->check.last = dev->last_rx; 2893*527a6266SJeff Kirsher sky2->check.mac_rp = mac_rp; 2894*527a6266SJeff Kirsher sky2->check.mac_lev = mac_lev; 2895*527a6266SJeff Kirsher sky2->check.fifo_rp = fifo_rp; 2896*527a6266SJeff Kirsher sky2->check.fifo_lev = fifo_lev; 2897*527a6266SJeff Kirsher return 0; 2898*527a6266SJeff Kirsher } 2899*527a6266SJeff Kirsher } 2900*527a6266SJeff Kirsher 2901*527a6266SJeff Kirsher static void sky2_watchdog(unsigned long arg) 2902*527a6266SJeff Kirsher { 2903*527a6266SJeff Kirsher struct sky2_hw *hw = (struct sky2_hw *) arg; 2904*527a6266SJeff Kirsher 2905*527a6266SJeff Kirsher /* Check for lost IRQ once a second */ 2906*527a6266SJeff Kirsher if (sky2_read32(hw, B0_ISRC)) { 2907*527a6266SJeff Kirsher napi_schedule(&hw->napi); 2908*527a6266SJeff Kirsher } else { 2909*527a6266SJeff Kirsher int i, active = 0; 2910*527a6266SJeff Kirsher 2911*527a6266SJeff Kirsher for (i = 0; i < hw->ports; i++) { 2912*527a6266SJeff Kirsher struct net_device *dev = hw->dev[i]; 2913*527a6266SJeff Kirsher if (!netif_running(dev)) 2914*527a6266SJeff Kirsher continue; 2915*527a6266SJeff Kirsher ++active; 2916*527a6266SJeff Kirsher 2917*527a6266SJeff Kirsher /* For chips with Rx FIFO, check if stuck */ 2918*527a6266SJeff Kirsher if ((hw->flags & SKY2_HW_RAM_BUFFER) && 2919*527a6266SJeff Kirsher sky2_rx_hung(dev)) { 2920*527a6266SJeff Kirsher netdev_info(dev, "receiver hang detected\n"); 2921*527a6266SJeff Kirsher schedule_work(&hw->restart_work); 2922*527a6266SJeff Kirsher return; 2923*527a6266SJeff Kirsher } 2924*527a6266SJeff Kirsher } 2925*527a6266SJeff Kirsher 2926*527a6266SJeff Kirsher if (active == 0) 2927*527a6266SJeff Kirsher return; 2928*527a6266SJeff Kirsher } 2929*527a6266SJeff Kirsher 2930*527a6266SJeff Kirsher mod_timer(&hw->watchdog_timer, round_jiffies(jiffies + HZ)); 2931*527a6266SJeff Kirsher } 2932*527a6266SJeff Kirsher 2933*527a6266SJeff Kirsher /* Hardware/software error handling */ 2934*527a6266SJeff Kirsher static void sky2_err_intr(struct sky2_hw *hw, u32 status) 2935*527a6266SJeff Kirsher { 2936*527a6266SJeff Kirsher if (net_ratelimit()) 2937*527a6266SJeff Kirsher dev_warn(&hw->pdev->dev, "error interrupt status=%#x\n", status); 2938*527a6266SJeff Kirsher 2939*527a6266SJeff Kirsher if (status & Y2_IS_HW_ERR) 2940*527a6266SJeff Kirsher sky2_hw_intr(hw); 2941*527a6266SJeff Kirsher 2942*527a6266SJeff Kirsher if (status & Y2_IS_IRQ_MAC1) 2943*527a6266SJeff Kirsher sky2_mac_intr(hw, 0); 2944*527a6266SJeff Kirsher 2945*527a6266SJeff Kirsher if (status & Y2_IS_IRQ_MAC2) 2946*527a6266SJeff Kirsher sky2_mac_intr(hw, 1); 2947*527a6266SJeff Kirsher 2948*527a6266SJeff Kirsher if (status & Y2_IS_CHK_RX1) 2949*527a6266SJeff Kirsher sky2_le_error(hw, 0, Q_R1); 2950*527a6266SJeff Kirsher 2951*527a6266SJeff Kirsher if (status & Y2_IS_CHK_RX2) 2952*527a6266SJeff Kirsher sky2_le_error(hw, 1, Q_R2); 2953*527a6266SJeff Kirsher 2954*527a6266SJeff Kirsher if (status & Y2_IS_CHK_TXA1) 2955*527a6266SJeff Kirsher sky2_le_error(hw, 0, Q_XA1); 2956*527a6266SJeff Kirsher 2957*527a6266SJeff Kirsher if (status & Y2_IS_CHK_TXA2) 2958*527a6266SJeff Kirsher sky2_le_error(hw, 1, Q_XA2); 2959*527a6266SJeff Kirsher } 2960*527a6266SJeff Kirsher 2961*527a6266SJeff Kirsher static int sky2_poll(struct napi_struct *napi, int work_limit) 2962*527a6266SJeff Kirsher { 2963*527a6266SJeff Kirsher struct sky2_hw *hw = container_of(napi, struct sky2_hw, napi); 2964*527a6266SJeff Kirsher u32 status = sky2_read32(hw, B0_Y2_SP_EISR); 2965*527a6266SJeff Kirsher int work_done = 0; 2966*527a6266SJeff Kirsher u16 idx; 2967*527a6266SJeff Kirsher 2968*527a6266SJeff Kirsher if (unlikely(status & Y2_IS_ERROR)) 2969*527a6266SJeff Kirsher sky2_err_intr(hw, status); 2970*527a6266SJeff Kirsher 2971*527a6266SJeff Kirsher if (status & Y2_IS_IRQ_PHY1) 2972*527a6266SJeff Kirsher sky2_phy_intr(hw, 0); 2973*527a6266SJeff Kirsher 2974*527a6266SJeff Kirsher if (status & Y2_IS_IRQ_PHY2) 2975*527a6266SJeff Kirsher sky2_phy_intr(hw, 1); 2976*527a6266SJeff Kirsher 2977*527a6266SJeff Kirsher if (status & Y2_IS_PHY_QLNK) 2978*527a6266SJeff Kirsher sky2_qlink_intr(hw); 2979*527a6266SJeff Kirsher 2980*527a6266SJeff Kirsher while ((idx = sky2_read16(hw, STAT_PUT_IDX)) != hw->st_idx) { 2981*527a6266SJeff Kirsher work_done += sky2_status_intr(hw, work_limit - work_done, idx); 2982*527a6266SJeff Kirsher 2983*527a6266SJeff Kirsher if (work_done >= work_limit) 2984*527a6266SJeff Kirsher goto done; 2985*527a6266SJeff Kirsher } 2986*527a6266SJeff Kirsher 2987*527a6266SJeff Kirsher napi_complete(napi); 2988*527a6266SJeff Kirsher sky2_read32(hw, B0_Y2_SP_LISR); 2989*527a6266SJeff Kirsher done: 2990*527a6266SJeff Kirsher 2991*527a6266SJeff Kirsher return work_done; 2992*527a6266SJeff Kirsher } 2993*527a6266SJeff Kirsher 2994*527a6266SJeff Kirsher static irqreturn_t sky2_intr(int irq, void *dev_id) 2995*527a6266SJeff Kirsher { 2996*527a6266SJeff Kirsher struct sky2_hw *hw = dev_id; 2997*527a6266SJeff Kirsher u32 status; 2998*527a6266SJeff Kirsher 2999*527a6266SJeff Kirsher /* Reading this mask interrupts as side effect */ 3000*527a6266SJeff Kirsher status = sky2_read32(hw, B0_Y2_SP_ISRC2); 3001*527a6266SJeff Kirsher if (status == 0 || status == ~0) 3002*527a6266SJeff Kirsher return IRQ_NONE; 3003*527a6266SJeff Kirsher 3004*527a6266SJeff Kirsher prefetch(&hw->st_le[hw->st_idx]); 3005*527a6266SJeff Kirsher 3006*527a6266SJeff Kirsher napi_schedule(&hw->napi); 3007*527a6266SJeff Kirsher 3008*527a6266SJeff Kirsher return IRQ_HANDLED; 3009*527a6266SJeff Kirsher } 3010*527a6266SJeff Kirsher 3011*527a6266SJeff Kirsher #ifdef CONFIG_NET_POLL_CONTROLLER 3012*527a6266SJeff Kirsher static void sky2_netpoll(struct net_device *dev) 3013*527a6266SJeff Kirsher { 3014*527a6266SJeff Kirsher struct sky2_port *sky2 = netdev_priv(dev); 3015*527a6266SJeff Kirsher 3016*527a6266SJeff Kirsher napi_schedule(&sky2->hw->napi); 3017*527a6266SJeff Kirsher } 3018*527a6266SJeff Kirsher #endif 3019*527a6266SJeff Kirsher 3020*527a6266SJeff Kirsher /* Chip internal frequency for clock calculations */ 3021*527a6266SJeff Kirsher static u32 sky2_mhz(const struct sky2_hw *hw) 3022*527a6266SJeff Kirsher { 3023*527a6266SJeff Kirsher switch (hw->chip_id) { 3024*527a6266SJeff Kirsher case CHIP_ID_YUKON_EC: 3025*527a6266SJeff Kirsher case CHIP_ID_YUKON_EC_U: 3026*527a6266SJeff Kirsher case CHIP_ID_YUKON_EX: 3027*527a6266SJeff Kirsher case CHIP_ID_YUKON_SUPR: 3028*527a6266SJeff Kirsher case CHIP_ID_YUKON_UL_2: 3029*527a6266SJeff Kirsher case CHIP_ID_YUKON_OPT: 3030*527a6266SJeff Kirsher case CHIP_ID_YUKON_PRM: 3031*527a6266SJeff Kirsher case CHIP_ID_YUKON_OP_2: 3032*527a6266SJeff Kirsher return 125; 3033*527a6266SJeff Kirsher 3034*527a6266SJeff Kirsher case CHIP_ID_YUKON_FE: 3035*527a6266SJeff Kirsher return 100; 3036*527a6266SJeff Kirsher 3037*527a6266SJeff Kirsher case CHIP_ID_YUKON_FE_P: 3038*527a6266SJeff Kirsher return 50; 3039*527a6266SJeff Kirsher 3040*527a6266SJeff Kirsher case CHIP_ID_YUKON_XL: 3041*527a6266SJeff Kirsher return 156; 3042*527a6266SJeff Kirsher 3043*527a6266SJeff Kirsher default: 3044*527a6266SJeff Kirsher BUG(); 3045*527a6266SJeff Kirsher } 3046*527a6266SJeff Kirsher } 3047*527a6266SJeff Kirsher 3048*527a6266SJeff Kirsher static inline u32 sky2_us2clk(const struct sky2_hw *hw, u32 us) 3049*527a6266SJeff Kirsher { 3050*527a6266SJeff Kirsher return sky2_mhz(hw) * us; 3051*527a6266SJeff Kirsher } 3052*527a6266SJeff Kirsher 3053*527a6266SJeff Kirsher static inline u32 sky2_clk2us(const struct sky2_hw *hw, u32 clk) 3054*527a6266SJeff Kirsher { 3055*527a6266SJeff Kirsher return clk / sky2_mhz(hw); 3056*527a6266SJeff Kirsher } 3057*527a6266SJeff Kirsher 3058*527a6266SJeff Kirsher 3059*527a6266SJeff Kirsher static int __devinit sky2_init(struct sky2_hw *hw) 3060*527a6266SJeff Kirsher { 3061*527a6266SJeff Kirsher u8 t8; 3062*527a6266SJeff Kirsher 3063*527a6266SJeff Kirsher /* Enable all clocks and check for bad PCI access */ 3064*527a6266SJeff Kirsher sky2_pci_write32(hw, PCI_DEV_REG3, 0); 3065*527a6266SJeff Kirsher 3066*527a6266SJeff Kirsher sky2_write8(hw, B0_CTST, CS_RST_CLR); 3067*527a6266SJeff Kirsher 3068*527a6266SJeff Kirsher hw->chip_id = sky2_read8(hw, B2_CHIP_ID); 3069*527a6266SJeff Kirsher hw->chip_rev = (sky2_read8(hw, B2_MAC_CFG) & CFG_CHIP_R_MSK) >> 4; 3070*527a6266SJeff Kirsher 3071*527a6266SJeff Kirsher switch (hw->chip_id) { 3072*527a6266SJeff Kirsher case CHIP_ID_YUKON_XL: 3073*527a6266SJeff Kirsher hw->flags = SKY2_HW_GIGABIT | SKY2_HW_NEWER_PHY; 3074*527a6266SJeff Kirsher if (hw->chip_rev < CHIP_REV_YU_XL_A2) 3075*527a6266SJeff Kirsher hw->flags |= SKY2_HW_RSS_BROKEN; 3076*527a6266SJeff Kirsher break; 3077*527a6266SJeff Kirsher 3078*527a6266SJeff Kirsher case CHIP_ID_YUKON_EC_U: 3079*527a6266SJeff Kirsher hw->flags = SKY2_HW_GIGABIT 3080*527a6266SJeff Kirsher | SKY2_HW_NEWER_PHY 3081*527a6266SJeff Kirsher | SKY2_HW_ADV_POWER_CTL; 3082*527a6266SJeff Kirsher break; 3083*527a6266SJeff Kirsher 3084*527a6266SJeff Kirsher case CHIP_ID_YUKON_EX: 3085*527a6266SJeff Kirsher hw->flags = SKY2_HW_GIGABIT 3086*527a6266SJeff Kirsher | SKY2_HW_NEWER_PHY 3087*527a6266SJeff Kirsher | SKY2_HW_NEW_LE 3088*527a6266SJeff Kirsher | SKY2_HW_ADV_POWER_CTL 3089*527a6266SJeff Kirsher | SKY2_HW_RSS_CHKSUM; 3090*527a6266SJeff Kirsher 3091*527a6266SJeff Kirsher /* New transmit checksum */ 3092*527a6266SJeff Kirsher if (hw->chip_rev != CHIP_REV_YU_EX_B0) 3093*527a6266SJeff Kirsher hw->flags |= SKY2_HW_AUTO_TX_SUM; 3094*527a6266SJeff Kirsher break; 3095*527a6266SJeff Kirsher 3096*527a6266SJeff Kirsher case CHIP_ID_YUKON_EC: 3097*527a6266SJeff Kirsher /* This rev is really old, and requires untested workarounds */ 3098*527a6266SJeff Kirsher if (hw->chip_rev == CHIP_REV_YU_EC_A1) { 3099*527a6266SJeff Kirsher dev_err(&hw->pdev->dev, "unsupported revision Yukon-EC rev A1\n"); 3100*527a6266SJeff Kirsher return -EOPNOTSUPP; 3101*527a6266SJeff Kirsher } 3102*527a6266SJeff Kirsher hw->flags = SKY2_HW_GIGABIT | SKY2_HW_RSS_BROKEN; 3103*527a6266SJeff Kirsher break; 3104*527a6266SJeff Kirsher 3105*527a6266SJeff Kirsher case CHIP_ID_YUKON_FE: 3106*527a6266SJeff Kirsher hw->flags = SKY2_HW_RSS_BROKEN; 3107*527a6266SJeff Kirsher break; 3108*527a6266SJeff Kirsher 3109*527a6266SJeff Kirsher case CHIP_ID_YUKON_FE_P: 3110*527a6266SJeff Kirsher hw->flags = SKY2_HW_NEWER_PHY 3111*527a6266SJeff Kirsher | SKY2_HW_NEW_LE 3112*527a6266SJeff Kirsher | SKY2_HW_AUTO_TX_SUM 3113*527a6266SJeff Kirsher | SKY2_HW_ADV_POWER_CTL; 3114*527a6266SJeff Kirsher 3115*527a6266SJeff Kirsher /* The workaround for status conflicts VLAN tag detection. */ 3116*527a6266SJeff Kirsher if (hw->chip_rev == CHIP_REV_YU_FE2_A0) 3117*527a6266SJeff Kirsher hw->flags |= SKY2_HW_VLAN_BROKEN | SKY2_HW_RSS_CHKSUM; 3118*527a6266SJeff Kirsher break; 3119*527a6266SJeff Kirsher 3120*527a6266SJeff Kirsher case CHIP_ID_YUKON_SUPR: 3121*527a6266SJeff Kirsher hw->flags = SKY2_HW_GIGABIT 3122*527a6266SJeff Kirsher | SKY2_HW_NEWER_PHY 3123*527a6266SJeff Kirsher | SKY2_HW_NEW_LE 3124*527a6266SJeff Kirsher | SKY2_HW_AUTO_TX_SUM 3125*527a6266SJeff Kirsher | SKY2_HW_ADV_POWER_CTL; 3126*527a6266SJeff Kirsher 3127*527a6266SJeff Kirsher if (hw->chip_rev == CHIP_REV_YU_SU_A0) 3128*527a6266SJeff Kirsher hw->flags |= SKY2_HW_RSS_CHKSUM; 3129*527a6266SJeff Kirsher break; 3130*527a6266SJeff Kirsher 3131*527a6266SJeff Kirsher case CHIP_ID_YUKON_UL_2: 3132*527a6266SJeff Kirsher hw->flags = SKY2_HW_GIGABIT 3133*527a6266SJeff Kirsher | SKY2_HW_ADV_POWER_CTL; 3134*527a6266SJeff Kirsher break; 3135*527a6266SJeff Kirsher 3136*527a6266SJeff Kirsher case CHIP_ID_YUKON_OPT: 3137*527a6266SJeff Kirsher case CHIP_ID_YUKON_PRM: 3138*527a6266SJeff Kirsher case CHIP_ID_YUKON_OP_2: 3139*527a6266SJeff Kirsher hw->flags = SKY2_HW_GIGABIT 3140*527a6266SJeff Kirsher | SKY2_HW_NEW_LE 3141*527a6266SJeff Kirsher | SKY2_HW_ADV_POWER_CTL; 3142*527a6266SJeff Kirsher break; 3143*527a6266SJeff Kirsher 3144*527a6266SJeff Kirsher default: 3145*527a6266SJeff Kirsher dev_err(&hw->pdev->dev, "unsupported chip type 0x%x\n", 3146*527a6266SJeff Kirsher hw->chip_id); 3147*527a6266SJeff Kirsher return -EOPNOTSUPP; 3148*527a6266SJeff Kirsher } 3149*527a6266SJeff Kirsher 3150*527a6266SJeff Kirsher hw->pmd_type = sky2_read8(hw, B2_PMD_TYP); 3151*527a6266SJeff Kirsher if (hw->pmd_type == 'L' || hw->pmd_type == 'S' || hw->pmd_type == 'P') 3152*527a6266SJeff Kirsher hw->flags |= SKY2_HW_FIBRE_PHY; 3153*527a6266SJeff Kirsher 3154*527a6266SJeff Kirsher hw->ports = 1; 3155*527a6266SJeff Kirsher t8 = sky2_read8(hw, B2_Y2_HW_RES); 3156*527a6266SJeff Kirsher if ((t8 & CFG_DUAL_MAC_MSK) == CFG_DUAL_MAC_MSK) { 3157*527a6266SJeff Kirsher if (!(sky2_read8(hw, B2_Y2_CLK_GATE) & Y2_STATUS_LNK2_INAC)) 3158*527a6266SJeff Kirsher ++hw->ports; 3159*527a6266SJeff Kirsher } 3160*527a6266SJeff Kirsher 3161*527a6266SJeff Kirsher if (sky2_read8(hw, B2_E_0)) 3162*527a6266SJeff Kirsher hw->flags |= SKY2_HW_RAM_BUFFER; 3163*527a6266SJeff Kirsher 3164*527a6266SJeff Kirsher return 0; 3165*527a6266SJeff Kirsher } 3166*527a6266SJeff Kirsher 3167*527a6266SJeff Kirsher static void sky2_reset(struct sky2_hw *hw) 3168*527a6266SJeff Kirsher { 3169*527a6266SJeff Kirsher struct pci_dev *pdev = hw->pdev; 3170*527a6266SJeff Kirsher u16 status; 3171*527a6266SJeff Kirsher int i; 3172*527a6266SJeff Kirsher u32 hwe_mask = Y2_HWE_ALL_MASK; 3173*527a6266SJeff Kirsher 3174*527a6266SJeff Kirsher /* disable ASF */ 3175*527a6266SJeff Kirsher if (hw->chip_id == CHIP_ID_YUKON_EX 3176*527a6266SJeff Kirsher || hw->chip_id == CHIP_ID_YUKON_SUPR) { 3177*527a6266SJeff Kirsher sky2_write32(hw, CPU_WDOG, 0); 3178*527a6266SJeff Kirsher status = sky2_read16(hw, HCU_CCSR); 3179*527a6266SJeff Kirsher status &= ~(HCU_CCSR_AHB_RST | HCU_CCSR_CPU_RST_MODE | 3180*527a6266SJeff Kirsher HCU_CCSR_UC_STATE_MSK); 3181*527a6266SJeff Kirsher /* 3182*527a6266SJeff Kirsher * CPU clock divider shouldn't be used because 3183*527a6266SJeff Kirsher * - ASF firmware may malfunction 3184*527a6266SJeff Kirsher * - Yukon-Supreme: Parallel FLASH doesn't support divided clocks 3185*527a6266SJeff Kirsher */ 3186*527a6266SJeff Kirsher status &= ~HCU_CCSR_CPU_CLK_DIVIDE_MSK; 3187*527a6266SJeff Kirsher sky2_write16(hw, HCU_CCSR, status); 3188*527a6266SJeff Kirsher sky2_write32(hw, CPU_WDOG, 0); 3189*527a6266SJeff Kirsher } else 3190*527a6266SJeff Kirsher sky2_write8(hw, B28_Y2_ASF_STAT_CMD, Y2_ASF_RESET); 3191*527a6266SJeff Kirsher sky2_write16(hw, B0_CTST, Y2_ASF_DISABLE); 3192*527a6266SJeff Kirsher 3193*527a6266SJeff Kirsher /* do a SW reset */ 3194*527a6266SJeff Kirsher sky2_write8(hw, B0_CTST, CS_RST_SET); 3195*527a6266SJeff Kirsher sky2_write8(hw, B0_CTST, CS_RST_CLR); 3196*527a6266SJeff Kirsher 3197*527a6266SJeff Kirsher /* allow writes to PCI config */ 3198*527a6266SJeff Kirsher sky2_write8(hw, B2_TST_CTRL1, TST_CFG_WRITE_ON); 3199*527a6266SJeff Kirsher 3200*527a6266SJeff Kirsher /* clear PCI errors, if any */ 3201*527a6266SJeff Kirsher status = sky2_pci_read16(hw, PCI_STATUS); 3202*527a6266SJeff Kirsher status |= PCI_STATUS_ERROR_BITS; 3203*527a6266SJeff Kirsher sky2_pci_write16(hw, PCI_STATUS, status); 3204*527a6266SJeff Kirsher 3205*527a6266SJeff Kirsher sky2_write8(hw, B0_CTST, CS_MRST_CLR); 3206*527a6266SJeff Kirsher 3207*527a6266SJeff Kirsher if (pci_is_pcie(pdev)) { 3208*527a6266SJeff Kirsher sky2_write32(hw, Y2_CFG_AER + PCI_ERR_UNCOR_STATUS, 3209*527a6266SJeff Kirsher 0xfffffffful); 3210*527a6266SJeff Kirsher 3211*527a6266SJeff Kirsher /* If error bit is stuck on ignore it */ 3212*527a6266SJeff Kirsher if (sky2_read32(hw, B0_HWE_ISRC) & Y2_IS_PCI_EXP) 3213*527a6266SJeff Kirsher dev_info(&pdev->dev, "ignoring stuck error report bit\n"); 3214*527a6266SJeff Kirsher else 3215*527a6266SJeff Kirsher hwe_mask |= Y2_IS_PCI_EXP; 3216*527a6266SJeff Kirsher } 3217*527a6266SJeff Kirsher 3218*527a6266SJeff Kirsher sky2_power_on(hw); 3219*527a6266SJeff Kirsher sky2_write8(hw, B2_TST_CTRL1, TST_CFG_WRITE_OFF); 3220*527a6266SJeff Kirsher 3221*527a6266SJeff Kirsher for (i = 0; i < hw->ports; i++) { 3222*527a6266SJeff Kirsher sky2_write8(hw, SK_REG(i, GMAC_LINK_CTRL), GMLC_RST_SET); 3223*527a6266SJeff Kirsher sky2_write8(hw, SK_REG(i, GMAC_LINK_CTRL), GMLC_RST_CLR); 3224*527a6266SJeff Kirsher 3225*527a6266SJeff Kirsher if (hw->chip_id == CHIP_ID_YUKON_EX || 3226*527a6266SJeff Kirsher hw->chip_id == CHIP_ID_YUKON_SUPR) 3227*527a6266SJeff Kirsher sky2_write16(hw, SK_REG(i, GMAC_CTRL), 3228*527a6266SJeff Kirsher GMC_BYP_MACSECRX_ON | GMC_BYP_MACSECTX_ON 3229*527a6266SJeff Kirsher | GMC_BYP_RETR_ON); 3230*527a6266SJeff Kirsher 3231*527a6266SJeff Kirsher } 3232*527a6266SJeff Kirsher 3233*527a6266SJeff Kirsher if (hw->chip_id == CHIP_ID_YUKON_SUPR && hw->chip_rev > CHIP_REV_YU_SU_B0) { 3234*527a6266SJeff Kirsher /* enable MACSec clock gating */ 3235*527a6266SJeff Kirsher sky2_pci_write32(hw, PCI_DEV_REG3, P_CLK_MACSEC_DIS); 3236*527a6266SJeff Kirsher } 3237*527a6266SJeff Kirsher 3238*527a6266SJeff Kirsher if (hw->chip_id == CHIP_ID_YUKON_OPT || 3239*527a6266SJeff Kirsher hw->chip_id == CHIP_ID_YUKON_PRM || 3240*527a6266SJeff Kirsher hw->chip_id == CHIP_ID_YUKON_OP_2) { 3241*527a6266SJeff Kirsher u16 reg; 3242*527a6266SJeff Kirsher u32 msk; 3243*527a6266SJeff Kirsher 3244*527a6266SJeff Kirsher if (hw->chip_id == CHIP_ID_YUKON_OPT && hw->chip_rev == 0) { 3245*527a6266SJeff Kirsher /* disable PCI-E PHY power down (set PHY reg 0x80, bit 7 */ 3246*527a6266SJeff Kirsher sky2_write32(hw, Y2_PEX_PHY_DATA, (0x80UL << 16) | (1 << 7)); 3247*527a6266SJeff Kirsher 3248*527a6266SJeff Kirsher /* set PHY Link Detect Timer to 1.1 second (11x 100ms) */ 3249*527a6266SJeff Kirsher reg = 10; 3250*527a6266SJeff Kirsher 3251*527a6266SJeff Kirsher /* re-enable PEX PM in PEX PHY debug reg. 8 (clear bit 12) */ 3252*527a6266SJeff Kirsher sky2_write32(hw, Y2_PEX_PHY_DATA, PEX_DB_ACCESS | (0x08UL << 16)); 3253*527a6266SJeff Kirsher } else { 3254*527a6266SJeff Kirsher /* set PHY Link Detect Timer to 0.4 second (4x 100ms) */ 3255*527a6266SJeff Kirsher reg = 3; 3256*527a6266SJeff Kirsher } 3257*527a6266SJeff Kirsher 3258*527a6266SJeff Kirsher reg <<= PSM_CONFIG_REG4_TIMER_PHY_LINK_DETECT_BASE; 3259*527a6266SJeff Kirsher reg |= PSM_CONFIG_REG4_RST_PHY_LINK_DETECT; 3260*527a6266SJeff Kirsher 3261*527a6266SJeff Kirsher /* reset PHY Link Detect */ 3262*527a6266SJeff Kirsher sky2_write8(hw, B2_TST_CTRL1, TST_CFG_WRITE_ON); 3263*527a6266SJeff Kirsher sky2_pci_write16(hw, PSM_CONFIG_REG4, reg); 3264*527a6266SJeff Kirsher 3265*527a6266SJeff Kirsher /* enable PHY Quick Link */ 3266*527a6266SJeff Kirsher msk = sky2_read32(hw, B0_IMSK); 3267*527a6266SJeff Kirsher msk |= Y2_IS_PHY_QLNK; 3268*527a6266SJeff Kirsher sky2_write32(hw, B0_IMSK, msk); 3269*527a6266SJeff Kirsher 3270*527a6266SJeff Kirsher /* check if PSMv2 was running before */ 3271*527a6266SJeff Kirsher reg = sky2_pci_read16(hw, PSM_CONFIG_REG3); 3272*527a6266SJeff Kirsher if (reg & PCI_EXP_LNKCTL_ASPMC) 3273*527a6266SJeff Kirsher /* restore the PCIe Link Control register */ 3274*527a6266SJeff Kirsher sky2_pci_write16(hw, pdev->pcie_cap + PCI_EXP_LNKCTL, 3275*527a6266SJeff Kirsher reg); 3276*527a6266SJeff Kirsher 3277*527a6266SJeff Kirsher sky2_write8(hw, B2_TST_CTRL1, TST_CFG_WRITE_OFF); 3278*527a6266SJeff Kirsher 3279*527a6266SJeff Kirsher /* re-enable PEX PM in PEX PHY debug reg. 8 (clear bit 12) */ 3280*527a6266SJeff Kirsher sky2_write32(hw, Y2_PEX_PHY_DATA, PEX_DB_ACCESS | (0x08UL << 16)); 3281*527a6266SJeff Kirsher } 3282*527a6266SJeff Kirsher 3283*527a6266SJeff Kirsher /* Clear I2C IRQ noise */ 3284*527a6266SJeff Kirsher sky2_write32(hw, B2_I2C_IRQ, 1); 3285*527a6266SJeff Kirsher 3286*527a6266SJeff Kirsher /* turn off hardware timer (unused) */ 3287*527a6266SJeff Kirsher sky2_write8(hw, B2_TI_CTRL, TIM_STOP); 3288*527a6266SJeff Kirsher sky2_write8(hw, B2_TI_CTRL, TIM_CLR_IRQ); 3289*527a6266SJeff Kirsher 3290*527a6266SJeff Kirsher /* Turn off descriptor polling */ 3291*527a6266SJeff Kirsher sky2_write32(hw, B28_DPT_CTRL, DPT_STOP); 3292*527a6266SJeff Kirsher 3293*527a6266SJeff Kirsher /* Turn off receive timestamp */ 3294*527a6266SJeff Kirsher sky2_write8(hw, GMAC_TI_ST_CTRL, GMT_ST_STOP); 3295*527a6266SJeff Kirsher sky2_write8(hw, GMAC_TI_ST_CTRL, GMT_ST_CLR_IRQ); 3296*527a6266SJeff Kirsher 3297*527a6266SJeff Kirsher /* enable the Tx Arbiters */ 3298*527a6266SJeff Kirsher for (i = 0; i < hw->ports; i++) 3299*527a6266SJeff Kirsher sky2_write8(hw, SK_REG(i, TXA_CTRL), TXA_ENA_ARB); 3300*527a6266SJeff Kirsher 3301*527a6266SJeff Kirsher /* Initialize ram interface */ 3302*527a6266SJeff Kirsher for (i = 0; i < hw->ports; i++) { 3303*527a6266SJeff Kirsher sky2_write8(hw, RAM_BUFFER(i, B3_RI_CTRL), RI_RST_CLR); 3304*527a6266SJeff Kirsher 3305*527a6266SJeff Kirsher sky2_write8(hw, RAM_BUFFER(i, B3_RI_WTO_R1), SK_RI_TO_53); 3306*527a6266SJeff Kirsher sky2_write8(hw, RAM_BUFFER(i, B3_RI_WTO_XA1), SK_RI_TO_53); 3307*527a6266SJeff Kirsher sky2_write8(hw, RAM_BUFFER(i, B3_RI_WTO_XS1), SK_RI_TO_53); 3308*527a6266SJeff Kirsher sky2_write8(hw, RAM_BUFFER(i, B3_RI_RTO_R1), SK_RI_TO_53); 3309*527a6266SJeff Kirsher sky2_write8(hw, RAM_BUFFER(i, B3_RI_RTO_XA1), SK_RI_TO_53); 3310*527a6266SJeff Kirsher sky2_write8(hw, RAM_BUFFER(i, B3_RI_RTO_XS1), SK_RI_TO_53); 3311*527a6266SJeff Kirsher sky2_write8(hw, RAM_BUFFER(i, B3_RI_WTO_R2), SK_RI_TO_53); 3312*527a6266SJeff Kirsher sky2_write8(hw, RAM_BUFFER(i, B3_RI_WTO_XA2), SK_RI_TO_53); 3313*527a6266SJeff Kirsher sky2_write8(hw, RAM_BUFFER(i, B3_RI_WTO_XS2), SK_RI_TO_53); 3314*527a6266SJeff Kirsher sky2_write8(hw, RAM_BUFFER(i, B3_RI_RTO_R2), SK_RI_TO_53); 3315*527a6266SJeff Kirsher sky2_write8(hw, RAM_BUFFER(i, B3_RI_RTO_XA2), SK_RI_TO_53); 3316*527a6266SJeff Kirsher sky2_write8(hw, RAM_BUFFER(i, B3_RI_RTO_XS2), SK_RI_TO_53); 3317*527a6266SJeff Kirsher } 3318*527a6266SJeff Kirsher 3319*527a6266SJeff Kirsher sky2_write32(hw, B0_HWE_IMSK, hwe_mask); 3320*527a6266SJeff Kirsher 3321*527a6266SJeff Kirsher for (i = 0; i < hw->ports; i++) 3322*527a6266SJeff Kirsher sky2_gmac_reset(hw, i); 3323*527a6266SJeff Kirsher 3324*527a6266SJeff Kirsher memset(hw->st_le, 0, hw->st_size * sizeof(struct sky2_status_le)); 3325*527a6266SJeff Kirsher hw->st_idx = 0; 3326*527a6266SJeff Kirsher 3327*527a6266SJeff Kirsher sky2_write32(hw, STAT_CTRL, SC_STAT_RST_SET); 3328*527a6266SJeff Kirsher sky2_write32(hw, STAT_CTRL, SC_STAT_RST_CLR); 3329*527a6266SJeff Kirsher 3330*527a6266SJeff Kirsher sky2_write32(hw, STAT_LIST_ADDR_LO, hw->st_dma); 3331*527a6266SJeff Kirsher sky2_write32(hw, STAT_LIST_ADDR_HI, (u64) hw->st_dma >> 32); 3332*527a6266SJeff Kirsher 3333*527a6266SJeff Kirsher /* Set the list last index */ 3334*527a6266SJeff Kirsher sky2_write16(hw, STAT_LAST_IDX, hw->st_size - 1); 3335*527a6266SJeff Kirsher 3336*527a6266SJeff Kirsher sky2_write16(hw, STAT_TX_IDX_TH, 10); 3337*527a6266SJeff Kirsher sky2_write8(hw, STAT_FIFO_WM, 16); 3338*527a6266SJeff Kirsher 3339*527a6266SJeff Kirsher /* set Status-FIFO ISR watermark */ 3340*527a6266SJeff Kirsher if (hw->chip_id == CHIP_ID_YUKON_XL && hw->chip_rev == 0) 3341*527a6266SJeff Kirsher sky2_write8(hw, STAT_FIFO_ISR_WM, 4); 3342*527a6266SJeff Kirsher else 3343*527a6266SJeff Kirsher sky2_write8(hw, STAT_FIFO_ISR_WM, 16); 3344*527a6266SJeff Kirsher 3345*527a6266SJeff Kirsher sky2_write32(hw, STAT_TX_TIMER_INI, sky2_us2clk(hw, 1000)); 3346*527a6266SJeff Kirsher sky2_write32(hw, STAT_ISR_TIMER_INI, sky2_us2clk(hw, 20)); 3347*527a6266SJeff Kirsher sky2_write32(hw, STAT_LEV_TIMER_INI, sky2_us2clk(hw, 100)); 3348*527a6266SJeff Kirsher 3349*527a6266SJeff Kirsher /* enable status unit */ 3350*527a6266SJeff Kirsher sky2_write32(hw, STAT_CTRL, SC_STAT_OP_ON); 3351*527a6266SJeff Kirsher 3352*527a6266SJeff Kirsher sky2_write8(hw, STAT_TX_TIMER_CTRL, TIM_START); 3353*527a6266SJeff Kirsher sky2_write8(hw, STAT_LEV_TIMER_CTRL, TIM_START); 3354*527a6266SJeff Kirsher sky2_write8(hw, STAT_ISR_TIMER_CTRL, TIM_START); 3355*527a6266SJeff Kirsher } 3356*527a6266SJeff Kirsher 3357*527a6266SJeff Kirsher /* Take device down (offline). 3358*527a6266SJeff Kirsher * Equivalent to doing dev_stop() but this does not 3359*527a6266SJeff Kirsher * inform upper layers of the transition. 3360*527a6266SJeff Kirsher */ 3361*527a6266SJeff Kirsher static void sky2_detach(struct net_device *dev) 3362*527a6266SJeff Kirsher { 3363*527a6266SJeff Kirsher if (netif_running(dev)) { 3364*527a6266SJeff Kirsher netif_tx_lock(dev); 3365*527a6266SJeff Kirsher netif_device_detach(dev); /* stop txq */ 3366*527a6266SJeff Kirsher netif_tx_unlock(dev); 3367*527a6266SJeff Kirsher sky2_down(dev); 3368*527a6266SJeff Kirsher } 3369*527a6266SJeff Kirsher } 3370*527a6266SJeff Kirsher 3371*527a6266SJeff Kirsher /* Bring device back after doing sky2_detach */ 3372*527a6266SJeff Kirsher static int sky2_reattach(struct net_device *dev) 3373*527a6266SJeff Kirsher { 3374*527a6266SJeff Kirsher int err = 0; 3375*527a6266SJeff Kirsher 3376*527a6266SJeff Kirsher if (netif_running(dev)) { 3377*527a6266SJeff Kirsher err = sky2_up(dev); 3378*527a6266SJeff Kirsher if (err) { 3379*527a6266SJeff Kirsher netdev_info(dev, "could not restart %d\n", err); 3380*527a6266SJeff Kirsher dev_close(dev); 3381*527a6266SJeff Kirsher } else { 3382*527a6266SJeff Kirsher netif_device_attach(dev); 3383*527a6266SJeff Kirsher sky2_set_multicast(dev); 3384*527a6266SJeff Kirsher } 3385*527a6266SJeff Kirsher } 3386*527a6266SJeff Kirsher 3387*527a6266SJeff Kirsher return err; 3388*527a6266SJeff Kirsher } 3389*527a6266SJeff Kirsher 3390*527a6266SJeff Kirsher static void sky2_all_down(struct sky2_hw *hw) 3391*527a6266SJeff Kirsher { 3392*527a6266SJeff Kirsher int i; 3393*527a6266SJeff Kirsher 3394*527a6266SJeff Kirsher sky2_read32(hw, B0_IMSK); 3395*527a6266SJeff Kirsher sky2_write32(hw, B0_IMSK, 0); 3396*527a6266SJeff Kirsher synchronize_irq(hw->pdev->irq); 3397*527a6266SJeff Kirsher napi_disable(&hw->napi); 3398*527a6266SJeff Kirsher 3399*527a6266SJeff Kirsher for (i = 0; i < hw->ports; i++) { 3400*527a6266SJeff Kirsher struct net_device *dev = hw->dev[i]; 3401*527a6266SJeff Kirsher struct sky2_port *sky2 = netdev_priv(dev); 3402*527a6266SJeff Kirsher 3403*527a6266SJeff Kirsher if (!netif_running(dev)) 3404*527a6266SJeff Kirsher continue; 3405*527a6266SJeff Kirsher 3406*527a6266SJeff Kirsher netif_carrier_off(dev); 3407*527a6266SJeff Kirsher netif_tx_disable(dev); 3408*527a6266SJeff Kirsher sky2_hw_down(sky2); 3409*527a6266SJeff Kirsher } 3410*527a6266SJeff Kirsher } 3411*527a6266SJeff Kirsher 3412*527a6266SJeff Kirsher static void sky2_all_up(struct sky2_hw *hw) 3413*527a6266SJeff Kirsher { 3414*527a6266SJeff Kirsher u32 imask = Y2_IS_BASE; 3415*527a6266SJeff Kirsher int i; 3416*527a6266SJeff Kirsher 3417*527a6266SJeff Kirsher for (i = 0; i < hw->ports; i++) { 3418*527a6266SJeff Kirsher struct net_device *dev = hw->dev[i]; 3419*527a6266SJeff Kirsher struct sky2_port *sky2 = netdev_priv(dev); 3420*527a6266SJeff Kirsher 3421*527a6266SJeff Kirsher if (!netif_running(dev)) 3422*527a6266SJeff Kirsher continue; 3423*527a6266SJeff Kirsher 3424*527a6266SJeff Kirsher sky2_hw_up(sky2); 3425*527a6266SJeff Kirsher sky2_set_multicast(dev); 3426*527a6266SJeff Kirsher imask |= portirq_msk[i]; 3427*527a6266SJeff Kirsher netif_wake_queue(dev); 3428*527a6266SJeff Kirsher } 3429*527a6266SJeff Kirsher 3430*527a6266SJeff Kirsher sky2_write32(hw, B0_IMSK, imask); 3431*527a6266SJeff Kirsher sky2_read32(hw, B0_IMSK); 3432*527a6266SJeff Kirsher 3433*527a6266SJeff Kirsher sky2_read32(hw, B0_Y2_SP_LISR); 3434*527a6266SJeff Kirsher napi_enable(&hw->napi); 3435*527a6266SJeff Kirsher } 3436*527a6266SJeff Kirsher 3437*527a6266SJeff Kirsher static void sky2_restart(struct work_struct *work) 3438*527a6266SJeff Kirsher { 3439*527a6266SJeff Kirsher struct sky2_hw *hw = container_of(work, struct sky2_hw, restart_work); 3440*527a6266SJeff Kirsher 3441*527a6266SJeff Kirsher rtnl_lock(); 3442*527a6266SJeff Kirsher 3443*527a6266SJeff Kirsher sky2_all_down(hw); 3444*527a6266SJeff Kirsher sky2_reset(hw); 3445*527a6266SJeff Kirsher sky2_all_up(hw); 3446*527a6266SJeff Kirsher 3447*527a6266SJeff Kirsher rtnl_unlock(); 3448*527a6266SJeff Kirsher } 3449*527a6266SJeff Kirsher 3450*527a6266SJeff Kirsher static inline u8 sky2_wol_supported(const struct sky2_hw *hw) 3451*527a6266SJeff Kirsher { 3452*527a6266SJeff Kirsher return sky2_is_copper(hw) ? (WAKE_PHY | WAKE_MAGIC) : 0; 3453*527a6266SJeff Kirsher } 3454*527a6266SJeff Kirsher 3455*527a6266SJeff Kirsher static void sky2_get_wol(struct net_device *dev, struct ethtool_wolinfo *wol) 3456*527a6266SJeff Kirsher { 3457*527a6266SJeff Kirsher const struct sky2_port *sky2 = netdev_priv(dev); 3458*527a6266SJeff Kirsher 3459*527a6266SJeff Kirsher wol->supported = sky2_wol_supported(sky2->hw); 3460*527a6266SJeff Kirsher wol->wolopts = sky2->wol; 3461*527a6266SJeff Kirsher } 3462*527a6266SJeff Kirsher 3463*527a6266SJeff Kirsher static int sky2_set_wol(struct net_device *dev, struct ethtool_wolinfo *wol) 3464*527a6266SJeff Kirsher { 3465*527a6266SJeff Kirsher struct sky2_port *sky2 = netdev_priv(dev); 3466*527a6266SJeff Kirsher struct sky2_hw *hw = sky2->hw; 3467*527a6266SJeff Kirsher bool enable_wakeup = false; 3468*527a6266SJeff Kirsher int i; 3469*527a6266SJeff Kirsher 3470*527a6266SJeff Kirsher if ((wol->wolopts & ~sky2_wol_supported(sky2->hw)) || 3471*527a6266SJeff Kirsher !device_can_wakeup(&hw->pdev->dev)) 3472*527a6266SJeff Kirsher return -EOPNOTSUPP; 3473*527a6266SJeff Kirsher 3474*527a6266SJeff Kirsher sky2->wol = wol->wolopts; 3475*527a6266SJeff Kirsher 3476*527a6266SJeff Kirsher for (i = 0; i < hw->ports; i++) { 3477*527a6266SJeff Kirsher struct net_device *dev = hw->dev[i]; 3478*527a6266SJeff Kirsher struct sky2_port *sky2 = netdev_priv(dev); 3479*527a6266SJeff Kirsher 3480*527a6266SJeff Kirsher if (sky2->wol) 3481*527a6266SJeff Kirsher enable_wakeup = true; 3482*527a6266SJeff Kirsher } 3483*527a6266SJeff Kirsher device_set_wakeup_enable(&hw->pdev->dev, enable_wakeup); 3484*527a6266SJeff Kirsher 3485*527a6266SJeff Kirsher return 0; 3486*527a6266SJeff Kirsher } 3487*527a6266SJeff Kirsher 3488*527a6266SJeff Kirsher static u32 sky2_supported_modes(const struct sky2_hw *hw) 3489*527a6266SJeff Kirsher { 3490*527a6266SJeff Kirsher if (sky2_is_copper(hw)) { 3491*527a6266SJeff Kirsher u32 modes = SUPPORTED_10baseT_Half 3492*527a6266SJeff Kirsher | SUPPORTED_10baseT_Full 3493*527a6266SJeff Kirsher | SUPPORTED_100baseT_Half 3494*527a6266SJeff Kirsher | SUPPORTED_100baseT_Full; 3495*527a6266SJeff Kirsher 3496*527a6266SJeff Kirsher if (hw->flags & SKY2_HW_GIGABIT) 3497*527a6266SJeff Kirsher modes |= SUPPORTED_1000baseT_Half 3498*527a6266SJeff Kirsher | SUPPORTED_1000baseT_Full; 3499*527a6266SJeff Kirsher return modes; 3500*527a6266SJeff Kirsher } else 3501*527a6266SJeff Kirsher return SUPPORTED_1000baseT_Half 3502*527a6266SJeff Kirsher | SUPPORTED_1000baseT_Full; 3503*527a6266SJeff Kirsher } 3504*527a6266SJeff Kirsher 3505*527a6266SJeff Kirsher static int sky2_get_settings(struct net_device *dev, struct ethtool_cmd *ecmd) 3506*527a6266SJeff Kirsher { 3507*527a6266SJeff Kirsher struct sky2_port *sky2 = netdev_priv(dev); 3508*527a6266SJeff Kirsher struct sky2_hw *hw = sky2->hw; 3509*527a6266SJeff Kirsher 3510*527a6266SJeff Kirsher ecmd->transceiver = XCVR_INTERNAL; 3511*527a6266SJeff Kirsher ecmd->supported = sky2_supported_modes(hw); 3512*527a6266SJeff Kirsher ecmd->phy_address = PHY_ADDR_MARV; 3513*527a6266SJeff Kirsher if (sky2_is_copper(hw)) { 3514*527a6266SJeff Kirsher ecmd->port = PORT_TP; 3515*527a6266SJeff Kirsher ethtool_cmd_speed_set(ecmd, sky2->speed); 3516*527a6266SJeff Kirsher ecmd->supported |= SUPPORTED_Autoneg | SUPPORTED_TP; 3517*527a6266SJeff Kirsher } else { 3518*527a6266SJeff Kirsher ethtool_cmd_speed_set(ecmd, SPEED_1000); 3519*527a6266SJeff Kirsher ecmd->port = PORT_FIBRE; 3520*527a6266SJeff Kirsher ecmd->supported |= SUPPORTED_Autoneg | SUPPORTED_FIBRE; 3521*527a6266SJeff Kirsher } 3522*527a6266SJeff Kirsher 3523*527a6266SJeff Kirsher ecmd->advertising = sky2->advertising; 3524*527a6266SJeff Kirsher ecmd->autoneg = (sky2->flags & SKY2_FLAG_AUTO_SPEED) 3525*527a6266SJeff Kirsher ? AUTONEG_ENABLE : AUTONEG_DISABLE; 3526*527a6266SJeff Kirsher ecmd->duplex = sky2->duplex; 3527*527a6266SJeff Kirsher return 0; 3528*527a6266SJeff Kirsher } 3529*527a6266SJeff Kirsher 3530*527a6266SJeff Kirsher static int sky2_set_settings(struct net_device *dev, struct ethtool_cmd *ecmd) 3531*527a6266SJeff Kirsher { 3532*527a6266SJeff Kirsher struct sky2_port *sky2 = netdev_priv(dev); 3533*527a6266SJeff Kirsher const struct sky2_hw *hw = sky2->hw; 3534*527a6266SJeff Kirsher u32 supported = sky2_supported_modes(hw); 3535*527a6266SJeff Kirsher 3536*527a6266SJeff Kirsher if (ecmd->autoneg == AUTONEG_ENABLE) { 3537*527a6266SJeff Kirsher if (ecmd->advertising & ~supported) 3538*527a6266SJeff Kirsher return -EINVAL; 3539*527a6266SJeff Kirsher 3540*527a6266SJeff Kirsher if (sky2_is_copper(hw)) 3541*527a6266SJeff Kirsher sky2->advertising = ecmd->advertising | 3542*527a6266SJeff Kirsher ADVERTISED_TP | 3543*527a6266SJeff Kirsher ADVERTISED_Autoneg; 3544*527a6266SJeff Kirsher else 3545*527a6266SJeff Kirsher sky2->advertising = ecmd->advertising | 3546*527a6266SJeff Kirsher ADVERTISED_FIBRE | 3547*527a6266SJeff Kirsher ADVERTISED_Autoneg; 3548*527a6266SJeff Kirsher 3549*527a6266SJeff Kirsher sky2->flags |= SKY2_FLAG_AUTO_SPEED; 3550*527a6266SJeff Kirsher sky2->duplex = -1; 3551*527a6266SJeff Kirsher sky2->speed = -1; 3552*527a6266SJeff Kirsher } else { 3553*527a6266SJeff Kirsher u32 setting; 3554*527a6266SJeff Kirsher u32 speed = ethtool_cmd_speed(ecmd); 3555*527a6266SJeff Kirsher 3556*527a6266SJeff Kirsher switch (speed) { 3557*527a6266SJeff Kirsher case SPEED_1000: 3558*527a6266SJeff Kirsher if (ecmd->duplex == DUPLEX_FULL) 3559*527a6266SJeff Kirsher setting = SUPPORTED_1000baseT_Full; 3560*527a6266SJeff Kirsher else if (ecmd->duplex == DUPLEX_HALF) 3561*527a6266SJeff Kirsher setting = SUPPORTED_1000baseT_Half; 3562*527a6266SJeff Kirsher else 3563*527a6266SJeff Kirsher return -EINVAL; 3564*527a6266SJeff Kirsher break; 3565*527a6266SJeff Kirsher case SPEED_100: 3566*527a6266SJeff Kirsher if (ecmd->duplex == DUPLEX_FULL) 3567*527a6266SJeff Kirsher setting = SUPPORTED_100baseT_Full; 3568*527a6266SJeff Kirsher else if (ecmd->duplex == DUPLEX_HALF) 3569*527a6266SJeff Kirsher setting = SUPPORTED_100baseT_Half; 3570*527a6266SJeff Kirsher else 3571*527a6266SJeff Kirsher return -EINVAL; 3572*527a6266SJeff Kirsher break; 3573*527a6266SJeff Kirsher 3574*527a6266SJeff Kirsher case SPEED_10: 3575*527a6266SJeff Kirsher if (ecmd->duplex == DUPLEX_FULL) 3576*527a6266SJeff Kirsher setting = SUPPORTED_10baseT_Full; 3577*527a6266SJeff Kirsher else if (ecmd->duplex == DUPLEX_HALF) 3578*527a6266SJeff Kirsher setting = SUPPORTED_10baseT_Half; 3579*527a6266SJeff Kirsher else 3580*527a6266SJeff Kirsher return -EINVAL; 3581*527a6266SJeff Kirsher break; 3582*527a6266SJeff Kirsher default: 3583*527a6266SJeff Kirsher return -EINVAL; 3584*527a6266SJeff Kirsher } 3585*527a6266SJeff Kirsher 3586*527a6266SJeff Kirsher if ((setting & supported) == 0) 3587*527a6266SJeff Kirsher return -EINVAL; 3588*527a6266SJeff Kirsher 3589*527a6266SJeff Kirsher sky2->speed = speed; 3590*527a6266SJeff Kirsher sky2->duplex = ecmd->duplex; 3591*527a6266SJeff Kirsher sky2->flags &= ~SKY2_FLAG_AUTO_SPEED; 3592*527a6266SJeff Kirsher } 3593*527a6266SJeff Kirsher 3594*527a6266SJeff Kirsher if (netif_running(dev)) { 3595*527a6266SJeff Kirsher sky2_phy_reinit(sky2); 3596*527a6266SJeff Kirsher sky2_set_multicast(dev); 3597*527a6266SJeff Kirsher } 3598*527a6266SJeff Kirsher 3599*527a6266SJeff Kirsher return 0; 3600*527a6266SJeff Kirsher } 3601*527a6266SJeff Kirsher 3602*527a6266SJeff Kirsher static void sky2_get_drvinfo(struct net_device *dev, 3603*527a6266SJeff Kirsher struct ethtool_drvinfo *info) 3604*527a6266SJeff Kirsher { 3605*527a6266SJeff Kirsher struct sky2_port *sky2 = netdev_priv(dev); 3606*527a6266SJeff Kirsher 3607*527a6266SJeff Kirsher strcpy(info->driver, DRV_NAME); 3608*527a6266SJeff Kirsher strcpy(info->version, DRV_VERSION); 3609*527a6266SJeff Kirsher strcpy(info->fw_version, "N/A"); 3610*527a6266SJeff Kirsher strcpy(info->bus_info, pci_name(sky2->hw->pdev)); 3611*527a6266SJeff Kirsher } 3612*527a6266SJeff Kirsher 3613*527a6266SJeff Kirsher static const struct sky2_stat { 3614*527a6266SJeff Kirsher char name[ETH_GSTRING_LEN]; 3615*527a6266SJeff Kirsher u16 offset; 3616*527a6266SJeff Kirsher } sky2_stats[] = { 3617*527a6266SJeff Kirsher { "tx_bytes", GM_TXO_OK_HI }, 3618*527a6266SJeff Kirsher { "rx_bytes", GM_RXO_OK_HI }, 3619*527a6266SJeff Kirsher { "tx_broadcast", GM_TXF_BC_OK }, 3620*527a6266SJeff Kirsher { "rx_broadcast", GM_RXF_BC_OK }, 3621*527a6266SJeff Kirsher { "tx_multicast", GM_TXF_MC_OK }, 3622*527a6266SJeff Kirsher { "rx_multicast", GM_RXF_MC_OK }, 3623*527a6266SJeff Kirsher { "tx_unicast", GM_TXF_UC_OK }, 3624*527a6266SJeff Kirsher { "rx_unicast", GM_RXF_UC_OK }, 3625*527a6266SJeff Kirsher { "tx_mac_pause", GM_TXF_MPAUSE }, 3626*527a6266SJeff Kirsher { "rx_mac_pause", GM_RXF_MPAUSE }, 3627*527a6266SJeff Kirsher { "collisions", GM_TXF_COL }, 3628*527a6266SJeff Kirsher { "late_collision",GM_TXF_LAT_COL }, 3629*527a6266SJeff Kirsher { "aborted", GM_TXF_ABO_COL }, 3630*527a6266SJeff Kirsher { "single_collisions", GM_TXF_SNG_COL }, 3631*527a6266SJeff Kirsher { "multi_collisions", GM_TXF_MUL_COL }, 3632*527a6266SJeff Kirsher 3633*527a6266SJeff Kirsher { "rx_short", GM_RXF_SHT }, 3634*527a6266SJeff Kirsher { "rx_runt", GM_RXE_FRAG }, 3635*527a6266SJeff Kirsher { "rx_64_byte_packets", GM_RXF_64B }, 3636*527a6266SJeff Kirsher { "rx_65_to_127_byte_packets", GM_RXF_127B }, 3637*527a6266SJeff Kirsher { "rx_128_to_255_byte_packets", GM_RXF_255B }, 3638*527a6266SJeff Kirsher { "rx_256_to_511_byte_packets", GM_RXF_511B }, 3639*527a6266SJeff Kirsher { "rx_512_to_1023_byte_packets", GM_RXF_1023B }, 3640*527a6266SJeff Kirsher { "rx_1024_to_1518_byte_packets", GM_RXF_1518B }, 3641*527a6266SJeff Kirsher { "rx_1518_to_max_byte_packets", GM_RXF_MAX_SZ }, 3642*527a6266SJeff Kirsher { "rx_too_long", GM_RXF_LNG_ERR }, 3643*527a6266SJeff Kirsher { "rx_fifo_overflow", GM_RXE_FIFO_OV }, 3644*527a6266SJeff Kirsher { "rx_jabber", GM_RXF_JAB_PKT }, 3645*527a6266SJeff Kirsher { "rx_fcs_error", GM_RXF_FCS_ERR }, 3646*527a6266SJeff Kirsher 3647*527a6266SJeff Kirsher { "tx_64_byte_packets", GM_TXF_64B }, 3648*527a6266SJeff Kirsher { "tx_65_to_127_byte_packets", GM_TXF_127B }, 3649*527a6266SJeff Kirsher { "tx_128_to_255_byte_packets", GM_TXF_255B }, 3650*527a6266SJeff Kirsher { "tx_256_to_511_byte_packets", GM_TXF_511B }, 3651*527a6266SJeff Kirsher { "tx_512_to_1023_byte_packets", GM_TXF_1023B }, 3652*527a6266SJeff Kirsher { "tx_1024_to_1518_byte_packets", GM_TXF_1518B }, 3653*527a6266SJeff Kirsher { "tx_1519_to_max_byte_packets", GM_TXF_MAX_SZ }, 3654*527a6266SJeff Kirsher { "tx_fifo_underrun", GM_TXE_FIFO_UR }, 3655*527a6266SJeff Kirsher }; 3656*527a6266SJeff Kirsher 3657*527a6266SJeff Kirsher static u32 sky2_get_msglevel(struct net_device *netdev) 3658*527a6266SJeff Kirsher { 3659*527a6266SJeff Kirsher struct sky2_port *sky2 = netdev_priv(netdev); 3660*527a6266SJeff Kirsher return sky2->msg_enable; 3661*527a6266SJeff Kirsher } 3662*527a6266SJeff Kirsher 3663*527a6266SJeff Kirsher static int sky2_nway_reset(struct net_device *dev) 3664*527a6266SJeff Kirsher { 3665*527a6266SJeff Kirsher struct sky2_port *sky2 = netdev_priv(dev); 3666*527a6266SJeff Kirsher 3667*527a6266SJeff Kirsher if (!netif_running(dev) || !(sky2->flags & SKY2_FLAG_AUTO_SPEED)) 3668*527a6266SJeff Kirsher return -EINVAL; 3669*527a6266SJeff Kirsher 3670*527a6266SJeff Kirsher sky2_phy_reinit(sky2); 3671*527a6266SJeff Kirsher sky2_set_multicast(dev); 3672*527a6266SJeff Kirsher 3673*527a6266SJeff Kirsher return 0; 3674*527a6266SJeff Kirsher } 3675*527a6266SJeff Kirsher 3676*527a6266SJeff Kirsher static void sky2_phy_stats(struct sky2_port *sky2, u64 * data, unsigned count) 3677*527a6266SJeff Kirsher { 3678*527a6266SJeff Kirsher struct sky2_hw *hw = sky2->hw; 3679*527a6266SJeff Kirsher unsigned port = sky2->port; 3680*527a6266SJeff Kirsher int i; 3681*527a6266SJeff Kirsher 3682*527a6266SJeff Kirsher data[0] = get_stats64(hw, port, GM_TXO_OK_LO); 3683*527a6266SJeff Kirsher data[1] = get_stats64(hw, port, GM_RXO_OK_LO); 3684*527a6266SJeff Kirsher 3685*527a6266SJeff Kirsher for (i = 2; i < count; i++) 3686*527a6266SJeff Kirsher data[i] = get_stats32(hw, port, sky2_stats[i].offset); 3687*527a6266SJeff Kirsher } 3688*527a6266SJeff Kirsher 3689*527a6266SJeff Kirsher static void sky2_set_msglevel(struct net_device *netdev, u32 value) 3690*527a6266SJeff Kirsher { 3691*527a6266SJeff Kirsher struct sky2_port *sky2 = netdev_priv(netdev); 3692*527a6266SJeff Kirsher sky2->msg_enable = value; 3693*527a6266SJeff Kirsher } 3694*527a6266SJeff Kirsher 3695*527a6266SJeff Kirsher static int sky2_get_sset_count(struct net_device *dev, int sset) 3696*527a6266SJeff Kirsher { 3697*527a6266SJeff Kirsher switch (sset) { 3698*527a6266SJeff Kirsher case ETH_SS_STATS: 3699*527a6266SJeff Kirsher return ARRAY_SIZE(sky2_stats); 3700*527a6266SJeff Kirsher default: 3701*527a6266SJeff Kirsher return -EOPNOTSUPP; 3702*527a6266SJeff Kirsher } 3703*527a6266SJeff Kirsher } 3704*527a6266SJeff Kirsher 3705*527a6266SJeff Kirsher static void sky2_get_ethtool_stats(struct net_device *dev, 3706*527a6266SJeff Kirsher struct ethtool_stats *stats, u64 * data) 3707*527a6266SJeff Kirsher { 3708*527a6266SJeff Kirsher struct sky2_port *sky2 = netdev_priv(dev); 3709*527a6266SJeff Kirsher 3710*527a6266SJeff Kirsher sky2_phy_stats(sky2, data, ARRAY_SIZE(sky2_stats)); 3711*527a6266SJeff Kirsher } 3712*527a6266SJeff Kirsher 3713*527a6266SJeff Kirsher static void sky2_get_strings(struct net_device *dev, u32 stringset, u8 * data) 3714*527a6266SJeff Kirsher { 3715*527a6266SJeff Kirsher int i; 3716*527a6266SJeff Kirsher 3717*527a6266SJeff Kirsher switch (stringset) { 3718*527a6266SJeff Kirsher case ETH_SS_STATS: 3719*527a6266SJeff Kirsher for (i = 0; i < ARRAY_SIZE(sky2_stats); i++) 3720*527a6266SJeff Kirsher memcpy(data + i * ETH_GSTRING_LEN, 3721*527a6266SJeff Kirsher sky2_stats[i].name, ETH_GSTRING_LEN); 3722*527a6266SJeff Kirsher break; 3723*527a6266SJeff Kirsher } 3724*527a6266SJeff Kirsher } 3725*527a6266SJeff Kirsher 3726*527a6266SJeff Kirsher static int sky2_set_mac_address(struct net_device *dev, void *p) 3727*527a6266SJeff Kirsher { 3728*527a6266SJeff Kirsher struct sky2_port *sky2 = netdev_priv(dev); 3729*527a6266SJeff Kirsher struct sky2_hw *hw = sky2->hw; 3730*527a6266SJeff Kirsher unsigned port = sky2->port; 3731*527a6266SJeff Kirsher const struct sockaddr *addr = p; 3732*527a6266SJeff Kirsher 3733*527a6266SJeff Kirsher if (!is_valid_ether_addr(addr->sa_data)) 3734*527a6266SJeff Kirsher return -EADDRNOTAVAIL; 3735*527a6266SJeff Kirsher 3736*527a6266SJeff Kirsher memcpy(dev->dev_addr, addr->sa_data, ETH_ALEN); 3737*527a6266SJeff Kirsher memcpy_toio(hw->regs + B2_MAC_1 + port * 8, 3738*527a6266SJeff Kirsher dev->dev_addr, ETH_ALEN); 3739*527a6266SJeff Kirsher memcpy_toio(hw->regs + B2_MAC_2 + port * 8, 3740*527a6266SJeff Kirsher dev->dev_addr, ETH_ALEN); 3741*527a6266SJeff Kirsher 3742*527a6266SJeff Kirsher /* virtual address for data */ 3743*527a6266SJeff Kirsher gma_set_addr(hw, port, GM_SRC_ADDR_2L, dev->dev_addr); 3744*527a6266SJeff Kirsher 3745*527a6266SJeff Kirsher /* physical address: used for pause frames */ 3746*527a6266SJeff Kirsher gma_set_addr(hw, port, GM_SRC_ADDR_1L, dev->dev_addr); 3747*527a6266SJeff Kirsher 3748*527a6266SJeff Kirsher return 0; 3749*527a6266SJeff Kirsher } 3750*527a6266SJeff Kirsher 3751*527a6266SJeff Kirsher static inline void sky2_add_filter(u8 filter[8], const u8 *addr) 3752*527a6266SJeff Kirsher { 3753*527a6266SJeff Kirsher u32 bit; 3754*527a6266SJeff Kirsher 3755*527a6266SJeff Kirsher bit = ether_crc(ETH_ALEN, addr) & 63; 3756*527a6266SJeff Kirsher filter[bit >> 3] |= 1 << (bit & 7); 3757*527a6266SJeff Kirsher } 3758*527a6266SJeff Kirsher 3759*527a6266SJeff Kirsher static void sky2_set_multicast(struct net_device *dev) 3760*527a6266SJeff Kirsher { 3761*527a6266SJeff Kirsher struct sky2_port *sky2 = netdev_priv(dev); 3762*527a6266SJeff Kirsher struct sky2_hw *hw = sky2->hw; 3763*527a6266SJeff Kirsher unsigned port = sky2->port; 3764*527a6266SJeff Kirsher struct netdev_hw_addr *ha; 3765*527a6266SJeff Kirsher u16 reg; 3766*527a6266SJeff Kirsher u8 filter[8]; 3767*527a6266SJeff Kirsher int rx_pause; 3768*527a6266SJeff Kirsher static const u8 pause_mc_addr[ETH_ALEN] = { 0x1, 0x80, 0xc2, 0x0, 0x0, 0x1 }; 3769*527a6266SJeff Kirsher 3770*527a6266SJeff Kirsher rx_pause = (sky2->flow_status == FC_RX || sky2->flow_status == FC_BOTH); 3771*527a6266SJeff Kirsher memset(filter, 0, sizeof(filter)); 3772*527a6266SJeff Kirsher 3773*527a6266SJeff Kirsher reg = gma_read16(hw, port, GM_RX_CTRL); 3774*527a6266SJeff Kirsher reg |= GM_RXCR_UCF_ENA; 3775*527a6266SJeff Kirsher 3776*527a6266SJeff Kirsher if (dev->flags & IFF_PROMISC) /* promiscuous */ 3777*527a6266SJeff Kirsher reg &= ~(GM_RXCR_UCF_ENA | GM_RXCR_MCF_ENA); 3778*527a6266SJeff Kirsher else if (dev->flags & IFF_ALLMULTI) 3779*527a6266SJeff Kirsher memset(filter, 0xff, sizeof(filter)); 3780*527a6266SJeff Kirsher else if (netdev_mc_empty(dev) && !rx_pause) 3781*527a6266SJeff Kirsher reg &= ~GM_RXCR_MCF_ENA; 3782*527a6266SJeff Kirsher else { 3783*527a6266SJeff Kirsher reg |= GM_RXCR_MCF_ENA; 3784*527a6266SJeff Kirsher 3785*527a6266SJeff Kirsher if (rx_pause) 3786*527a6266SJeff Kirsher sky2_add_filter(filter, pause_mc_addr); 3787*527a6266SJeff Kirsher 3788*527a6266SJeff Kirsher netdev_for_each_mc_addr(ha, dev) 3789*527a6266SJeff Kirsher sky2_add_filter(filter, ha->addr); 3790*527a6266SJeff Kirsher } 3791*527a6266SJeff Kirsher 3792*527a6266SJeff Kirsher gma_write16(hw, port, GM_MC_ADDR_H1, 3793*527a6266SJeff Kirsher (u16) filter[0] | ((u16) filter[1] << 8)); 3794*527a6266SJeff Kirsher gma_write16(hw, port, GM_MC_ADDR_H2, 3795*527a6266SJeff Kirsher (u16) filter[2] | ((u16) filter[3] << 8)); 3796*527a6266SJeff Kirsher gma_write16(hw, port, GM_MC_ADDR_H3, 3797*527a6266SJeff Kirsher (u16) filter[4] | ((u16) filter[5] << 8)); 3798*527a6266SJeff Kirsher gma_write16(hw, port, GM_MC_ADDR_H4, 3799*527a6266SJeff Kirsher (u16) filter[6] | ((u16) filter[7] << 8)); 3800*527a6266SJeff Kirsher 3801*527a6266SJeff Kirsher gma_write16(hw, port, GM_RX_CTRL, reg); 3802*527a6266SJeff Kirsher } 3803*527a6266SJeff Kirsher 3804*527a6266SJeff Kirsher static struct rtnl_link_stats64 *sky2_get_stats(struct net_device *dev, 3805*527a6266SJeff Kirsher struct rtnl_link_stats64 *stats) 3806*527a6266SJeff Kirsher { 3807*527a6266SJeff Kirsher struct sky2_port *sky2 = netdev_priv(dev); 3808*527a6266SJeff Kirsher struct sky2_hw *hw = sky2->hw; 3809*527a6266SJeff Kirsher unsigned port = sky2->port; 3810*527a6266SJeff Kirsher unsigned int start; 3811*527a6266SJeff Kirsher u64 _bytes, _packets; 3812*527a6266SJeff Kirsher 3813*527a6266SJeff Kirsher do { 3814*527a6266SJeff Kirsher start = u64_stats_fetch_begin_bh(&sky2->rx_stats.syncp); 3815*527a6266SJeff Kirsher _bytes = sky2->rx_stats.bytes; 3816*527a6266SJeff Kirsher _packets = sky2->rx_stats.packets; 3817*527a6266SJeff Kirsher } while (u64_stats_fetch_retry_bh(&sky2->rx_stats.syncp, start)); 3818*527a6266SJeff Kirsher 3819*527a6266SJeff Kirsher stats->rx_packets = _packets; 3820*527a6266SJeff Kirsher stats->rx_bytes = _bytes; 3821*527a6266SJeff Kirsher 3822*527a6266SJeff Kirsher do { 3823*527a6266SJeff Kirsher start = u64_stats_fetch_begin_bh(&sky2->tx_stats.syncp); 3824*527a6266SJeff Kirsher _bytes = sky2->tx_stats.bytes; 3825*527a6266SJeff Kirsher _packets = sky2->tx_stats.packets; 3826*527a6266SJeff Kirsher } while (u64_stats_fetch_retry_bh(&sky2->tx_stats.syncp, start)); 3827*527a6266SJeff Kirsher 3828*527a6266SJeff Kirsher stats->tx_packets = _packets; 3829*527a6266SJeff Kirsher stats->tx_bytes = _bytes; 3830*527a6266SJeff Kirsher 3831*527a6266SJeff Kirsher stats->multicast = get_stats32(hw, port, GM_RXF_MC_OK) 3832*527a6266SJeff Kirsher + get_stats32(hw, port, GM_RXF_BC_OK); 3833*527a6266SJeff Kirsher 3834*527a6266SJeff Kirsher stats->collisions = get_stats32(hw, port, GM_TXF_COL); 3835*527a6266SJeff Kirsher 3836*527a6266SJeff Kirsher stats->rx_length_errors = get_stats32(hw, port, GM_RXF_LNG_ERR); 3837*527a6266SJeff Kirsher stats->rx_crc_errors = get_stats32(hw, port, GM_RXF_FCS_ERR); 3838*527a6266SJeff Kirsher stats->rx_frame_errors = get_stats32(hw, port, GM_RXF_SHT) 3839*527a6266SJeff Kirsher + get_stats32(hw, port, GM_RXE_FRAG); 3840*527a6266SJeff Kirsher stats->rx_over_errors = get_stats32(hw, port, GM_RXE_FIFO_OV); 3841*527a6266SJeff Kirsher 3842*527a6266SJeff Kirsher stats->rx_dropped = dev->stats.rx_dropped; 3843*527a6266SJeff Kirsher stats->rx_fifo_errors = dev->stats.rx_fifo_errors; 3844*527a6266SJeff Kirsher stats->tx_fifo_errors = dev->stats.tx_fifo_errors; 3845*527a6266SJeff Kirsher 3846*527a6266SJeff Kirsher return stats; 3847*527a6266SJeff Kirsher } 3848*527a6266SJeff Kirsher 3849*527a6266SJeff Kirsher /* Can have one global because blinking is controlled by 3850*527a6266SJeff Kirsher * ethtool and that is always under RTNL mutex 3851*527a6266SJeff Kirsher */ 3852*527a6266SJeff Kirsher static void sky2_led(struct sky2_port *sky2, enum led_mode mode) 3853*527a6266SJeff Kirsher { 3854*527a6266SJeff Kirsher struct sky2_hw *hw = sky2->hw; 3855*527a6266SJeff Kirsher unsigned port = sky2->port; 3856*527a6266SJeff Kirsher 3857*527a6266SJeff Kirsher spin_lock_bh(&sky2->phy_lock); 3858*527a6266SJeff Kirsher if (hw->chip_id == CHIP_ID_YUKON_EC_U || 3859*527a6266SJeff Kirsher hw->chip_id == CHIP_ID_YUKON_EX || 3860*527a6266SJeff Kirsher hw->chip_id == CHIP_ID_YUKON_SUPR) { 3861*527a6266SJeff Kirsher u16 pg; 3862*527a6266SJeff Kirsher pg = gm_phy_read(hw, port, PHY_MARV_EXT_ADR); 3863*527a6266SJeff Kirsher gm_phy_write(hw, port, PHY_MARV_EXT_ADR, 3); 3864*527a6266SJeff Kirsher 3865*527a6266SJeff Kirsher switch (mode) { 3866*527a6266SJeff Kirsher case MO_LED_OFF: 3867*527a6266SJeff Kirsher gm_phy_write(hw, port, PHY_MARV_PHY_CTRL, 3868*527a6266SJeff Kirsher PHY_M_LEDC_LOS_CTRL(8) | 3869*527a6266SJeff Kirsher PHY_M_LEDC_INIT_CTRL(8) | 3870*527a6266SJeff Kirsher PHY_M_LEDC_STA1_CTRL(8) | 3871*527a6266SJeff Kirsher PHY_M_LEDC_STA0_CTRL(8)); 3872*527a6266SJeff Kirsher break; 3873*527a6266SJeff Kirsher case MO_LED_ON: 3874*527a6266SJeff Kirsher gm_phy_write(hw, port, PHY_MARV_PHY_CTRL, 3875*527a6266SJeff Kirsher PHY_M_LEDC_LOS_CTRL(9) | 3876*527a6266SJeff Kirsher PHY_M_LEDC_INIT_CTRL(9) | 3877*527a6266SJeff Kirsher PHY_M_LEDC_STA1_CTRL(9) | 3878*527a6266SJeff Kirsher PHY_M_LEDC_STA0_CTRL(9)); 3879*527a6266SJeff Kirsher break; 3880*527a6266SJeff Kirsher case MO_LED_BLINK: 3881*527a6266SJeff Kirsher gm_phy_write(hw, port, PHY_MARV_PHY_CTRL, 3882*527a6266SJeff Kirsher PHY_M_LEDC_LOS_CTRL(0xa) | 3883*527a6266SJeff Kirsher PHY_M_LEDC_INIT_CTRL(0xa) | 3884*527a6266SJeff Kirsher PHY_M_LEDC_STA1_CTRL(0xa) | 3885*527a6266SJeff Kirsher PHY_M_LEDC_STA0_CTRL(0xa)); 3886*527a6266SJeff Kirsher break; 3887*527a6266SJeff Kirsher case MO_LED_NORM: 3888*527a6266SJeff Kirsher gm_phy_write(hw, port, PHY_MARV_PHY_CTRL, 3889*527a6266SJeff Kirsher PHY_M_LEDC_LOS_CTRL(1) | 3890*527a6266SJeff Kirsher PHY_M_LEDC_INIT_CTRL(8) | 3891*527a6266SJeff Kirsher PHY_M_LEDC_STA1_CTRL(7) | 3892*527a6266SJeff Kirsher PHY_M_LEDC_STA0_CTRL(7)); 3893*527a6266SJeff Kirsher } 3894*527a6266SJeff Kirsher 3895*527a6266SJeff Kirsher gm_phy_write(hw, port, PHY_MARV_EXT_ADR, pg); 3896*527a6266SJeff Kirsher } else 3897*527a6266SJeff Kirsher gm_phy_write(hw, port, PHY_MARV_LED_OVER, 3898*527a6266SJeff Kirsher PHY_M_LED_MO_DUP(mode) | 3899*527a6266SJeff Kirsher PHY_M_LED_MO_10(mode) | 3900*527a6266SJeff Kirsher PHY_M_LED_MO_100(mode) | 3901*527a6266SJeff Kirsher PHY_M_LED_MO_1000(mode) | 3902*527a6266SJeff Kirsher PHY_M_LED_MO_RX(mode) | 3903*527a6266SJeff Kirsher PHY_M_LED_MO_TX(mode)); 3904*527a6266SJeff Kirsher 3905*527a6266SJeff Kirsher spin_unlock_bh(&sky2->phy_lock); 3906*527a6266SJeff Kirsher } 3907*527a6266SJeff Kirsher 3908*527a6266SJeff Kirsher /* blink LED's for finding board */ 3909*527a6266SJeff Kirsher static int sky2_set_phys_id(struct net_device *dev, 3910*527a6266SJeff Kirsher enum ethtool_phys_id_state state) 3911*527a6266SJeff Kirsher { 3912*527a6266SJeff Kirsher struct sky2_port *sky2 = netdev_priv(dev); 3913*527a6266SJeff Kirsher 3914*527a6266SJeff Kirsher switch (state) { 3915*527a6266SJeff Kirsher case ETHTOOL_ID_ACTIVE: 3916*527a6266SJeff Kirsher return 1; /* cycle on/off once per second */ 3917*527a6266SJeff Kirsher case ETHTOOL_ID_INACTIVE: 3918*527a6266SJeff Kirsher sky2_led(sky2, MO_LED_NORM); 3919*527a6266SJeff Kirsher break; 3920*527a6266SJeff Kirsher case ETHTOOL_ID_ON: 3921*527a6266SJeff Kirsher sky2_led(sky2, MO_LED_ON); 3922*527a6266SJeff Kirsher break; 3923*527a6266SJeff Kirsher case ETHTOOL_ID_OFF: 3924*527a6266SJeff Kirsher sky2_led(sky2, MO_LED_OFF); 3925*527a6266SJeff Kirsher break; 3926*527a6266SJeff Kirsher } 3927*527a6266SJeff Kirsher 3928*527a6266SJeff Kirsher return 0; 3929*527a6266SJeff Kirsher } 3930*527a6266SJeff Kirsher 3931*527a6266SJeff Kirsher static void sky2_get_pauseparam(struct net_device *dev, 3932*527a6266SJeff Kirsher struct ethtool_pauseparam *ecmd) 3933*527a6266SJeff Kirsher { 3934*527a6266SJeff Kirsher struct sky2_port *sky2 = netdev_priv(dev); 3935*527a6266SJeff Kirsher 3936*527a6266SJeff Kirsher switch (sky2->flow_mode) { 3937*527a6266SJeff Kirsher case FC_NONE: 3938*527a6266SJeff Kirsher ecmd->tx_pause = ecmd->rx_pause = 0; 3939*527a6266SJeff Kirsher break; 3940*527a6266SJeff Kirsher case FC_TX: 3941*527a6266SJeff Kirsher ecmd->tx_pause = 1, ecmd->rx_pause = 0; 3942*527a6266SJeff Kirsher break; 3943*527a6266SJeff Kirsher case FC_RX: 3944*527a6266SJeff Kirsher ecmd->tx_pause = 0, ecmd->rx_pause = 1; 3945*527a6266SJeff Kirsher break; 3946*527a6266SJeff Kirsher case FC_BOTH: 3947*527a6266SJeff Kirsher ecmd->tx_pause = ecmd->rx_pause = 1; 3948*527a6266SJeff Kirsher } 3949*527a6266SJeff Kirsher 3950*527a6266SJeff Kirsher ecmd->autoneg = (sky2->flags & SKY2_FLAG_AUTO_PAUSE) 3951*527a6266SJeff Kirsher ? AUTONEG_ENABLE : AUTONEG_DISABLE; 3952*527a6266SJeff Kirsher } 3953*527a6266SJeff Kirsher 3954*527a6266SJeff Kirsher static int sky2_set_pauseparam(struct net_device *dev, 3955*527a6266SJeff Kirsher struct ethtool_pauseparam *ecmd) 3956*527a6266SJeff Kirsher { 3957*527a6266SJeff Kirsher struct sky2_port *sky2 = netdev_priv(dev); 3958*527a6266SJeff Kirsher 3959*527a6266SJeff Kirsher if (ecmd->autoneg == AUTONEG_ENABLE) 3960*527a6266SJeff Kirsher sky2->flags |= SKY2_FLAG_AUTO_PAUSE; 3961*527a6266SJeff Kirsher else 3962*527a6266SJeff Kirsher sky2->flags &= ~SKY2_FLAG_AUTO_PAUSE; 3963*527a6266SJeff Kirsher 3964*527a6266SJeff Kirsher sky2->flow_mode = sky2_flow(ecmd->rx_pause, ecmd->tx_pause); 3965*527a6266SJeff Kirsher 3966*527a6266SJeff Kirsher if (netif_running(dev)) 3967*527a6266SJeff Kirsher sky2_phy_reinit(sky2); 3968*527a6266SJeff Kirsher 3969*527a6266SJeff Kirsher return 0; 3970*527a6266SJeff Kirsher } 3971*527a6266SJeff Kirsher 3972*527a6266SJeff Kirsher static int sky2_get_coalesce(struct net_device *dev, 3973*527a6266SJeff Kirsher struct ethtool_coalesce *ecmd) 3974*527a6266SJeff Kirsher { 3975*527a6266SJeff Kirsher struct sky2_port *sky2 = netdev_priv(dev); 3976*527a6266SJeff Kirsher struct sky2_hw *hw = sky2->hw; 3977*527a6266SJeff Kirsher 3978*527a6266SJeff Kirsher if (sky2_read8(hw, STAT_TX_TIMER_CTRL) == TIM_STOP) 3979*527a6266SJeff Kirsher ecmd->tx_coalesce_usecs = 0; 3980*527a6266SJeff Kirsher else { 3981*527a6266SJeff Kirsher u32 clks = sky2_read32(hw, STAT_TX_TIMER_INI); 3982*527a6266SJeff Kirsher ecmd->tx_coalesce_usecs = sky2_clk2us(hw, clks); 3983*527a6266SJeff Kirsher } 3984*527a6266SJeff Kirsher ecmd->tx_max_coalesced_frames = sky2_read16(hw, STAT_TX_IDX_TH); 3985*527a6266SJeff Kirsher 3986*527a6266SJeff Kirsher if (sky2_read8(hw, STAT_LEV_TIMER_CTRL) == TIM_STOP) 3987*527a6266SJeff Kirsher ecmd->rx_coalesce_usecs = 0; 3988*527a6266SJeff Kirsher else { 3989*527a6266SJeff Kirsher u32 clks = sky2_read32(hw, STAT_LEV_TIMER_INI); 3990*527a6266SJeff Kirsher ecmd->rx_coalesce_usecs = sky2_clk2us(hw, clks); 3991*527a6266SJeff Kirsher } 3992*527a6266SJeff Kirsher ecmd->rx_max_coalesced_frames = sky2_read8(hw, STAT_FIFO_WM); 3993*527a6266SJeff Kirsher 3994*527a6266SJeff Kirsher if (sky2_read8(hw, STAT_ISR_TIMER_CTRL) == TIM_STOP) 3995*527a6266SJeff Kirsher ecmd->rx_coalesce_usecs_irq = 0; 3996*527a6266SJeff Kirsher else { 3997*527a6266SJeff Kirsher u32 clks = sky2_read32(hw, STAT_ISR_TIMER_INI); 3998*527a6266SJeff Kirsher ecmd->rx_coalesce_usecs_irq = sky2_clk2us(hw, clks); 3999*527a6266SJeff Kirsher } 4000*527a6266SJeff Kirsher 4001*527a6266SJeff Kirsher ecmd->rx_max_coalesced_frames_irq = sky2_read8(hw, STAT_FIFO_ISR_WM); 4002*527a6266SJeff Kirsher 4003*527a6266SJeff Kirsher return 0; 4004*527a6266SJeff Kirsher } 4005*527a6266SJeff Kirsher 4006*527a6266SJeff Kirsher /* Note: this affect both ports */ 4007*527a6266SJeff Kirsher static int sky2_set_coalesce(struct net_device *dev, 4008*527a6266SJeff Kirsher struct ethtool_coalesce *ecmd) 4009*527a6266SJeff Kirsher { 4010*527a6266SJeff Kirsher struct sky2_port *sky2 = netdev_priv(dev); 4011*527a6266SJeff Kirsher struct sky2_hw *hw = sky2->hw; 4012*527a6266SJeff Kirsher const u32 tmax = sky2_clk2us(hw, 0x0ffffff); 4013*527a6266SJeff Kirsher 4014*527a6266SJeff Kirsher if (ecmd->tx_coalesce_usecs > tmax || 4015*527a6266SJeff Kirsher ecmd->rx_coalesce_usecs > tmax || 4016*527a6266SJeff Kirsher ecmd->rx_coalesce_usecs_irq > tmax) 4017*527a6266SJeff Kirsher return -EINVAL; 4018*527a6266SJeff Kirsher 4019*527a6266SJeff Kirsher if (ecmd->tx_max_coalesced_frames >= sky2->tx_ring_size-1) 4020*527a6266SJeff Kirsher return -EINVAL; 4021*527a6266SJeff Kirsher if (ecmd->rx_max_coalesced_frames > RX_MAX_PENDING) 4022*527a6266SJeff Kirsher return -EINVAL; 4023*527a6266SJeff Kirsher if (ecmd->rx_max_coalesced_frames_irq > RX_MAX_PENDING) 4024*527a6266SJeff Kirsher return -EINVAL; 4025*527a6266SJeff Kirsher 4026*527a6266SJeff Kirsher if (ecmd->tx_coalesce_usecs == 0) 4027*527a6266SJeff Kirsher sky2_write8(hw, STAT_TX_TIMER_CTRL, TIM_STOP); 4028*527a6266SJeff Kirsher else { 4029*527a6266SJeff Kirsher sky2_write32(hw, STAT_TX_TIMER_INI, 4030*527a6266SJeff Kirsher sky2_us2clk(hw, ecmd->tx_coalesce_usecs)); 4031*527a6266SJeff Kirsher sky2_write8(hw, STAT_TX_TIMER_CTRL, TIM_START); 4032*527a6266SJeff Kirsher } 4033*527a6266SJeff Kirsher sky2_write16(hw, STAT_TX_IDX_TH, ecmd->tx_max_coalesced_frames); 4034*527a6266SJeff Kirsher 4035*527a6266SJeff Kirsher if (ecmd->rx_coalesce_usecs == 0) 4036*527a6266SJeff Kirsher sky2_write8(hw, STAT_LEV_TIMER_CTRL, TIM_STOP); 4037*527a6266SJeff Kirsher else { 4038*527a6266SJeff Kirsher sky2_write32(hw, STAT_LEV_TIMER_INI, 4039*527a6266SJeff Kirsher sky2_us2clk(hw, ecmd->rx_coalesce_usecs)); 4040*527a6266SJeff Kirsher sky2_write8(hw, STAT_LEV_TIMER_CTRL, TIM_START); 4041*527a6266SJeff Kirsher } 4042*527a6266SJeff Kirsher sky2_write8(hw, STAT_FIFO_WM, ecmd->rx_max_coalesced_frames); 4043*527a6266SJeff Kirsher 4044*527a6266SJeff Kirsher if (ecmd->rx_coalesce_usecs_irq == 0) 4045*527a6266SJeff Kirsher sky2_write8(hw, STAT_ISR_TIMER_CTRL, TIM_STOP); 4046*527a6266SJeff Kirsher else { 4047*527a6266SJeff Kirsher sky2_write32(hw, STAT_ISR_TIMER_INI, 4048*527a6266SJeff Kirsher sky2_us2clk(hw, ecmd->rx_coalesce_usecs_irq)); 4049*527a6266SJeff Kirsher sky2_write8(hw, STAT_ISR_TIMER_CTRL, TIM_START); 4050*527a6266SJeff Kirsher } 4051*527a6266SJeff Kirsher sky2_write8(hw, STAT_FIFO_ISR_WM, ecmd->rx_max_coalesced_frames_irq); 4052*527a6266SJeff Kirsher return 0; 4053*527a6266SJeff Kirsher } 4054*527a6266SJeff Kirsher 4055*527a6266SJeff Kirsher static void sky2_get_ringparam(struct net_device *dev, 4056*527a6266SJeff Kirsher struct ethtool_ringparam *ering) 4057*527a6266SJeff Kirsher { 4058*527a6266SJeff Kirsher struct sky2_port *sky2 = netdev_priv(dev); 4059*527a6266SJeff Kirsher 4060*527a6266SJeff Kirsher ering->rx_max_pending = RX_MAX_PENDING; 4061*527a6266SJeff Kirsher ering->rx_mini_max_pending = 0; 4062*527a6266SJeff Kirsher ering->rx_jumbo_max_pending = 0; 4063*527a6266SJeff Kirsher ering->tx_max_pending = TX_MAX_PENDING; 4064*527a6266SJeff Kirsher 4065*527a6266SJeff Kirsher ering->rx_pending = sky2->rx_pending; 4066*527a6266SJeff Kirsher ering->rx_mini_pending = 0; 4067*527a6266SJeff Kirsher ering->rx_jumbo_pending = 0; 4068*527a6266SJeff Kirsher ering->tx_pending = sky2->tx_pending; 4069*527a6266SJeff Kirsher } 4070*527a6266SJeff Kirsher 4071*527a6266SJeff Kirsher static int sky2_set_ringparam(struct net_device *dev, 4072*527a6266SJeff Kirsher struct ethtool_ringparam *ering) 4073*527a6266SJeff Kirsher { 4074*527a6266SJeff Kirsher struct sky2_port *sky2 = netdev_priv(dev); 4075*527a6266SJeff Kirsher 4076*527a6266SJeff Kirsher if (ering->rx_pending > RX_MAX_PENDING || 4077*527a6266SJeff Kirsher ering->rx_pending < 8 || 4078*527a6266SJeff Kirsher ering->tx_pending < TX_MIN_PENDING || 4079*527a6266SJeff Kirsher ering->tx_pending > TX_MAX_PENDING) 4080*527a6266SJeff Kirsher return -EINVAL; 4081*527a6266SJeff Kirsher 4082*527a6266SJeff Kirsher sky2_detach(dev); 4083*527a6266SJeff Kirsher 4084*527a6266SJeff Kirsher sky2->rx_pending = ering->rx_pending; 4085*527a6266SJeff Kirsher sky2->tx_pending = ering->tx_pending; 4086*527a6266SJeff Kirsher sky2->tx_ring_size = roundup_pow_of_two(sky2->tx_pending+1); 4087*527a6266SJeff Kirsher 4088*527a6266SJeff Kirsher return sky2_reattach(dev); 4089*527a6266SJeff Kirsher } 4090*527a6266SJeff Kirsher 4091*527a6266SJeff Kirsher static int sky2_get_regs_len(struct net_device *dev) 4092*527a6266SJeff Kirsher { 4093*527a6266SJeff Kirsher return 0x4000; 4094*527a6266SJeff Kirsher } 4095*527a6266SJeff Kirsher 4096*527a6266SJeff Kirsher static int sky2_reg_access_ok(struct sky2_hw *hw, unsigned int b) 4097*527a6266SJeff Kirsher { 4098*527a6266SJeff Kirsher /* This complicated switch statement is to make sure and 4099*527a6266SJeff Kirsher * only access regions that are unreserved. 4100*527a6266SJeff Kirsher * Some blocks are only valid on dual port cards. 4101*527a6266SJeff Kirsher */ 4102*527a6266SJeff Kirsher switch (b) { 4103*527a6266SJeff Kirsher /* second port */ 4104*527a6266SJeff Kirsher case 5: /* Tx Arbiter 2 */ 4105*527a6266SJeff Kirsher case 9: /* RX2 */ 4106*527a6266SJeff Kirsher case 14 ... 15: /* TX2 */ 4107*527a6266SJeff Kirsher case 17: case 19: /* Ram Buffer 2 */ 4108*527a6266SJeff Kirsher case 22 ... 23: /* Tx Ram Buffer 2 */ 4109*527a6266SJeff Kirsher case 25: /* Rx MAC Fifo 1 */ 4110*527a6266SJeff Kirsher case 27: /* Tx MAC Fifo 2 */ 4111*527a6266SJeff Kirsher case 31: /* GPHY 2 */ 4112*527a6266SJeff Kirsher case 40 ... 47: /* Pattern Ram 2 */ 4113*527a6266SJeff Kirsher case 52: case 54: /* TCP Segmentation 2 */ 4114*527a6266SJeff Kirsher case 112 ... 116: /* GMAC 2 */ 4115*527a6266SJeff Kirsher return hw->ports > 1; 4116*527a6266SJeff Kirsher 4117*527a6266SJeff Kirsher case 0: /* Control */ 4118*527a6266SJeff Kirsher case 2: /* Mac address */ 4119*527a6266SJeff Kirsher case 4: /* Tx Arbiter 1 */ 4120*527a6266SJeff Kirsher case 7: /* PCI express reg */ 4121*527a6266SJeff Kirsher case 8: /* RX1 */ 4122*527a6266SJeff Kirsher case 12 ... 13: /* TX1 */ 4123*527a6266SJeff Kirsher case 16: case 18:/* Rx Ram Buffer 1 */ 4124*527a6266SJeff Kirsher case 20 ... 21: /* Tx Ram Buffer 1 */ 4125*527a6266SJeff Kirsher case 24: /* Rx MAC Fifo 1 */ 4126*527a6266SJeff Kirsher case 26: /* Tx MAC Fifo 1 */ 4127*527a6266SJeff Kirsher case 28 ... 29: /* Descriptor and status unit */ 4128*527a6266SJeff Kirsher case 30: /* GPHY 1*/ 4129*527a6266SJeff Kirsher case 32 ... 39: /* Pattern Ram 1 */ 4130*527a6266SJeff Kirsher case 48: case 50: /* TCP Segmentation 1 */ 4131*527a6266SJeff Kirsher case 56 ... 60: /* PCI space */ 4132*527a6266SJeff Kirsher case 80 ... 84: /* GMAC 1 */ 4133*527a6266SJeff Kirsher return 1; 4134*527a6266SJeff Kirsher 4135*527a6266SJeff Kirsher default: 4136*527a6266SJeff Kirsher return 0; 4137*527a6266SJeff Kirsher } 4138*527a6266SJeff Kirsher } 4139*527a6266SJeff Kirsher 4140*527a6266SJeff Kirsher /* 4141*527a6266SJeff Kirsher * Returns copy of control register region 4142*527a6266SJeff Kirsher * Note: ethtool_get_regs always provides full size (16k) buffer 4143*527a6266SJeff Kirsher */ 4144*527a6266SJeff Kirsher static void sky2_get_regs(struct net_device *dev, struct ethtool_regs *regs, 4145*527a6266SJeff Kirsher void *p) 4146*527a6266SJeff Kirsher { 4147*527a6266SJeff Kirsher const struct sky2_port *sky2 = netdev_priv(dev); 4148*527a6266SJeff Kirsher const void __iomem *io = sky2->hw->regs; 4149*527a6266SJeff Kirsher unsigned int b; 4150*527a6266SJeff Kirsher 4151*527a6266SJeff Kirsher regs->version = 1; 4152*527a6266SJeff Kirsher 4153*527a6266SJeff Kirsher for (b = 0; b < 128; b++) { 4154*527a6266SJeff Kirsher /* skip poisonous diagnostic ram region in block 3 */ 4155*527a6266SJeff Kirsher if (b == 3) 4156*527a6266SJeff Kirsher memcpy_fromio(p + 0x10, io + 0x10, 128 - 0x10); 4157*527a6266SJeff Kirsher else if (sky2_reg_access_ok(sky2->hw, b)) 4158*527a6266SJeff Kirsher memcpy_fromio(p, io, 128); 4159*527a6266SJeff Kirsher else 4160*527a6266SJeff Kirsher memset(p, 0, 128); 4161*527a6266SJeff Kirsher 4162*527a6266SJeff Kirsher p += 128; 4163*527a6266SJeff Kirsher io += 128; 4164*527a6266SJeff Kirsher } 4165*527a6266SJeff Kirsher } 4166*527a6266SJeff Kirsher 4167*527a6266SJeff Kirsher static int sky2_get_eeprom_len(struct net_device *dev) 4168*527a6266SJeff Kirsher { 4169*527a6266SJeff Kirsher struct sky2_port *sky2 = netdev_priv(dev); 4170*527a6266SJeff Kirsher struct sky2_hw *hw = sky2->hw; 4171*527a6266SJeff Kirsher u16 reg2; 4172*527a6266SJeff Kirsher 4173*527a6266SJeff Kirsher reg2 = sky2_pci_read16(hw, PCI_DEV_REG2); 4174*527a6266SJeff Kirsher return 1 << ( ((reg2 & PCI_VPD_ROM_SZ) >> 14) + 8); 4175*527a6266SJeff Kirsher } 4176*527a6266SJeff Kirsher 4177*527a6266SJeff Kirsher static int sky2_vpd_wait(const struct sky2_hw *hw, int cap, u16 busy) 4178*527a6266SJeff Kirsher { 4179*527a6266SJeff Kirsher unsigned long start = jiffies; 4180*527a6266SJeff Kirsher 4181*527a6266SJeff Kirsher while ( (sky2_pci_read16(hw, cap + PCI_VPD_ADDR) & PCI_VPD_ADDR_F) == busy) { 4182*527a6266SJeff Kirsher /* Can take up to 10.6 ms for write */ 4183*527a6266SJeff Kirsher if (time_after(jiffies, start + HZ/4)) { 4184*527a6266SJeff Kirsher dev_err(&hw->pdev->dev, "VPD cycle timed out\n"); 4185*527a6266SJeff Kirsher return -ETIMEDOUT; 4186*527a6266SJeff Kirsher } 4187*527a6266SJeff Kirsher mdelay(1); 4188*527a6266SJeff Kirsher } 4189*527a6266SJeff Kirsher 4190*527a6266SJeff Kirsher return 0; 4191*527a6266SJeff Kirsher } 4192*527a6266SJeff Kirsher 4193*527a6266SJeff Kirsher static int sky2_vpd_read(struct sky2_hw *hw, int cap, void *data, 4194*527a6266SJeff Kirsher u16 offset, size_t length) 4195*527a6266SJeff Kirsher { 4196*527a6266SJeff Kirsher int rc = 0; 4197*527a6266SJeff Kirsher 4198*527a6266SJeff Kirsher while (length > 0) { 4199*527a6266SJeff Kirsher u32 val; 4200*527a6266SJeff Kirsher 4201*527a6266SJeff Kirsher sky2_pci_write16(hw, cap + PCI_VPD_ADDR, offset); 4202*527a6266SJeff Kirsher rc = sky2_vpd_wait(hw, cap, 0); 4203*527a6266SJeff Kirsher if (rc) 4204*527a6266SJeff Kirsher break; 4205*527a6266SJeff Kirsher 4206*527a6266SJeff Kirsher val = sky2_pci_read32(hw, cap + PCI_VPD_DATA); 4207*527a6266SJeff Kirsher 4208*527a6266SJeff Kirsher memcpy(data, &val, min(sizeof(val), length)); 4209*527a6266SJeff Kirsher offset += sizeof(u32); 4210*527a6266SJeff Kirsher data += sizeof(u32); 4211*527a6266SJeff Kirsher length -= sizeof(u32); 4212*527a6266SJeff Kirsher } 4213*527a6266SJeff Kirsher 4214*527a6266SJeff Kirsher return rc; 4215*527a6266SJeff Kirsher } 4216*527a6266SJeff Kirsher 4217*527a6266SJeff Kirsher static int sky2_vpd_write(struct sky2_hw *hw, int cap, const void *data, 4218*527a6266SJeff Kirsher u16 offset, unsigned int length) 4219*527a6266SJeff Kirsher { 4220*527a6266SJeff Kirsher unsigned int i; 4221*527a6266SJeff Kirsher int rc = 0; 4222*527a6266SJeff Kirsher 4223*527a6266SJeff Kirsher for (i = 0; i < length; i += sizeof(u32)) { 4224*527a6266SJeff Kirsher u32 val = *(u32 *)(data + i); 4225*527a6266SJeff Kirsher 4226*527a6266SJeff Kirsher sky2_pci_write32(hw, cap + PCI_VPD_DATA, val); 4227*527a6266SJeff Kirsher sky2_pci_write32(hw, cap + PCI_VPD_ADDR, offset | PCI_VPD_ADDR_F); 4228*527a6266SJeff Kirsher 4229*527a6266SJeff Kirsher rc = sky2_vpd_wait(hw, cap, PCI_VPD_ADDR_F); 4230*527a6266SJeff Kirsher if (rc) 4231*527a6266SJeff Kirsher break; 4232*527a6266SJeff Kirsher } 4233*527a6266SJeff Kirsher return rc; 4234*527a6266SJeff Kirsher } 4235*527a6266SJeff Kirsher 4236*527a6266SJeff Kirsher static int sky2_get_eeprom(struct net_device *dev, struct ethtool_eeprom *eeprom, 4237*527a6266SJeff Kirsher u8 *data) 4238*527a6266SJeff Kirsher { 4239*527a6266SJeff Kirsher struct sky2_port *sky2 = netdev_priv(dev); 4240*527a6266SJeff Kirsher int cap = pci_find_capability(sky2->hw->pdev, PCI_CAP_ID_VPD); 4241*527a6266SJeff Kirsher 4242*527a6266SJeff Kirsher if (!cap) 4243*527a6266SJeff Kirsher return -EINVAL; 4244*527a6266SJeff Kirsher 4245*527a6266SJeff Kirsher eeprom->magic = SKY2_EEPROM_MAGIC; 4246*527a6266SJeff Kirsher 4247*527a6266SJeff Kirsher return sky2_vpd_read(sky2->hw, cap, data, eeprom->offset, eeprom->len); 4248*527a6266SJeff Kirsher } 4249*527a6266SJeff Kirsher 4250*527a6266SJeff Kirsher static int sky2_set_eeprom(struct net_device *dev, struct ethtool_eeprom *eeprom, 4251*527a6266SJeff Kirsher u8 *data) 4252*527a6266SJeff Kirsher { 4253*527a6266SJeff Kirsher struct sky2_port *sky2 = netdev_priv(dev); 4254*527a6266SJeff Kirsher int cap = pci_find_capability(sky2->hw->pdev, PCI_CAP_ID_VPD); 4255*527a6266SJeff Kirsher 4256*527a6266SJeff Kirsher if (!cap) 4257*527a6266SJeff Kirsher return -EINVAL; 4258*527a6266SJeff Kirsher 4259*527a6266SJeff Kirsher if (eeprom->magic != SKY2_EEPROM_MAGIC) 4260*527a6266SJeff Kirsher return -EINVAL; 4261*527a6266SJeff Kirsher 4262*527a6266SJeff Kirsher /* Partial writes not supported */ 4263*527a6266SJeff Kirsher if ((eeprom->offset & 3) || (eeprom->len & 3)) 4264*527a6266SJeff Kirsher return -EINVAL; 4265*527a6266SJeff Kirsher 4266*527a6266SJeff Kirsher return sky2_vpd_write(sky2->hw, cap, data, eeprom->offset, eeprom->len); 4267*527a6266SJeff Kirsher } 4268*527a6266SJeff Kirsher 4269*527a6266SJeff Kirsher static u32 sky2_fix_features(struct net_device *dev, u32 features) 4270*527a6266SJeff Kirsher { 4271*527a6266SJeff Kirsher const struct sky2_port *sky2 = netdev_priv(dev); 4272*527a6266SJeff Kirsher const struct sky2_hw *hw = sky2->hw; 4273*527a6266SJeff Kirsher 4274*527a6266SJeff Kirsher /* In order to do Jumbo packets on these chips, need to turn off the 4275*527a6266SJeff Kirsher * transmit store/forward. Therefore checksum offload won't work. 4276*527a6266SJeff Kirsher */ 4277*527a6266SJeff Kirsher if (dev->mtu > ETH_DATA_LEN && hw->chip_id == CHIP_ID_YUKON_EC_U) { 4278*527a6266SJeff Kirsher netdev_info(dev, "checksum offload not possible with jumbo frames\n"); 4279*527a6266SJeff Kirsher features &= ~(NETIF_F_TSO|NETIF_F_SG|NETIF_F_ALL_CSUM); 4280*527a6266SJeff Kirsher } 4281*527a6266SJeff Kirsher 4282*527a6266SJeff Kirsher /* Some hardware requires receive checksum for RSS to work. */ 4283*527a6266SJeff Kirsher if ( (features & NETIF_F_RXHASH) && 4284*527a6266SJeff Kirsher !(features & NETIF_F_RXCSUM) && 4285*527a6266SJeff Kirsher (sky2->hw->flags & SKY2_HW_RSS_CHKSUM)) { 4286*527a6266SJeff Kirsher netdev_info(dev, "receive hashing forces receive checksum\n"); 4287*527a6266SJeff Kirsher features |= NETIF_F_RXCSUM; 4288*527a6266SJeff Kirsher } 4289*527a6266SJeff Kirsher 4290*527a6266SJeff Kirsher return features; 4291*527a6266SJeff Kirsher } 4292*527a6266SJeff Kirsher 4293*527a6266SJeff Kirsher static int sky2_set_features(struct net_device *dev, u32 features) 4294*527a6266SJeff Kirsher { 4295*527a6266SJeff Kirsher struct sky2_port *sky2 = netdev_priv(dev); 4296*527a6266SJeff Kirsher u32 changed = dev->features ^ features; 4297*527a6266SJeff Kirsher 4298*527a6266SJeff Kirsher if (changed & NETIF_F_RXCSUM) { 4299*527a6266SJeff Kirsher u32 on = features & NETIF_F_RXCSUM; 4300*527a6266SJeff Kirsher sky2_write32(sky2->hw, Q_ADDR(rxqaddr[sky2->port], Q_CSR), 4301*527a6266SJeff Kirsher on ? BMU_ENA_RX_CHKSUM : BMU_DIS_RX_CHKSUM); 4302*527a6266SJeff Kirsher } 4303*527a6266SJeff Kirsher 4304*527a6266SJeff Kirsher if (changed & NETIF_F_RXHASH) 4305*527a6266SJeff Kirsher rx_set_rss(dev, features); 4306*527a6266SJeff Kirsher 4307*527a6266SJeff Kirsher if (changed & (NETIF_F_HW_VLAN_TX|NETIF_F_HW_VLAN_RX)) 4308*527a6266SJeff Kirsher sky2_vlan_mode(dev, features); 4309*527a6266SJeff Kirsher 4310*527a6266SJeff Kirsher return 0; 4311*527a6266SJeff Kirsher } 4312*527a6266SJeff Kirsher 4313*527a6266SJeff Kirsher static const struct ethtool_ops sky2_ethtool_ops = { 4314*527a6266SJeff Kirsher .get_settings = sky2_get_settings, 4315*527a6266SJeff Kirsher .set_settings = sky2_set_settings, 4316*527a6266SJeff Kirsher .get_drvinfo = sky2_get_drvinfo, 4317*527a6266SJeff Kirsher .get_wol = sky2_get_wol, 4318*527a6266SJeff Kirsher .set_wol = sky2_set_wol, 4319*527a6266SJeff Kirsher .get_msglevel = sky2_get_msglevel, 4320*527a6266SJeff Kirsher .set_msglevel = sky2_set_msglevel, 4321*527a6266SJeff Kirsher .nway_reset = sky2_nway_reset, 4322*527a6266SJeff Kirsher .get_regs_len = sky2_get_regs_len, 4323*527a6266SJeff Kirsher .get_regs = sky2_get_regs, 4324*527a6266SJeff Kirsher .get_link = ethtool_op_get_link, 4325*527a6266SJeff Kirsher .get_eeprom_len = sky2_get_eeprom_len, 4326*527a6266SJeff Kirsher .get_eeprom = sky2_get_eeprom, 4327*527a6266SJeff Kirsher .set_eeprom = sky2_set_eeprom, 4328*527a6266SJeff Kirsher .get_strings = sky2_get_strings, 4329*527a6266SJeff Kirsher .get_coalesce = sky2_get_coalesce, 4330*527a6266SJeff Kirsher .set_coalesce = sky2_set_coalesce, 4331*527a6266SJeff Kirsher .get_ringparam = sky2_get_ringparam, 4332*527a6266SJeff Kirsher .set_ringparam = sky2_set_ringparam, 4333*527a6266SJeff Kirsher .get_pauseparam = sky2_get_pauseparam, 4334*527a6266SJeff Kirsher .set_pauseparam = sky2_set_pauseparam, 4335*527a6266SJeff Kirsher .set_phys_id = sky2_set_phys_id, 4336*527a6266SJeff Kirsher .get_sset_count = sky2_get_sset_count, 4337*527a6266SJeff Kirsher .get_ethtool_stats = sky2_get_ethtool_stats, 4338*527a6266SJeff Kirsher }; 4339*527a6266SJeff Kirsher 4340*527a6266SJeff Kirsher #ifdef CONFIG_SKY2_DEBUG 4341*527a6266SJeff Kirsher 4342*527a6266SJeff Kirsher static struct dentry *sky2_debug; 4343*527a6266SJeff Kirsher 4344*527a6266SJeff Kirsher 4345*527a6266SJeff Kirsher /* 4346*527a6266SJeff Kirsher * Read and parse the first part of Vital Product Data 4347*527a6266SJeff Kirsher */ 4348*527a6266SJeff Kirsher #define VPD_SIZE 128 4349*527a6266SJeff Kirsher #define VPD_MAGIC 0x82 4350*527a6266SJeff Kirsher 4351*527a6266SJeff Kirsher static const struct vpd_tag { 4352*527a6266SJeff Kirsher char tag[2]; 4353*527a6266SJeff Kirsher char *label; 4354*527a6266SJeff Kirsher } vpd_tags[] = { 4355*527a6266SJeff Kirsher { "PN", "Part Number" }, 4356*527a6266SJeff Kirsher { "EC", "Engineering Level" }, 4357*527a6266SJeff Kirsher { "MN", "Manufacturer" }, 4358*527a6266SJeff Kirsher { "SN", "Serial Number" }, 4359*527a6266SJeff Kirsher { "YA", "Asset Tag" }, 4360*527a6266SJeff Kirsher { "VL", "First Error Log Message" }, 4361*527a6266SJeff Kirsher { "VF", "Second Error Log Message" }, 4362*527a6266SJeff Kirsher { "VB", "Boot Agent ROM Configuration" }, 4363*527a6266SJeff Kirsher { "VE", "EFI UNDI Configuration" }, 4364*527a6266SJeff Kirsher }; 4365*527a6266SJeff Kirsher 4366*527a6266SJeff Kirsher static void sky2_show_vpd(struct seq_file *seq, struct sky2_hw *hw) 4367*527a6266SJeff Kirsher { 4368*527a6266SJeff Kirsher size_t vpd_size; 4369*527a6266SJeff Kirsher loff_t offs; 4370*527a6266SJeff Kirsher u8 len; 4371*527a6266SJeff Kirsher unsigned char *buf; 4372*527a6266SJeff Kirsher u16 reg2; 4373*527a6266SJeff Kirsher 4374*527a6266SJeff Kirsher reg2 = sky2_pci_read16(hw, PCI_DEV_REG2); 4375*527a6266SJeff Kirsher vpd_size = 1 << ( ((reg2 & PCI_VPD_ROM_SZ) >> 14) + 8); 4376*527a6266SJeff Kirsher 4377*527a6266SJeff Kirsher seq_printf(seq, "%s Product Data\n", pci_name(hw->pdev)); 4378*527a6266SJeff Kirsher buf = kmalloc(vpd_size, GFP_KERNEL); 4379*527a6266SJeff Kirsher if (!buf) { 4380*527a6266SJeff Kirsher seq_puts(seq, "no memory!\n"); 4381*527a6266SJeff Kirsher return; 4382*527a6266SJeff Kirsher } 4383*527a6266SJeff Kirsher 4384*527a6266SJeff Kirsher if (pci_read_vpd(hw->pdev, 0, vpd_size, buf) < 0) { 4385*527a6266SJeff Kirsher seq_puts(seq, "VPD read failed\n"); 4386*527a6266SJeff Kirsher goto out; 4387*527a6266SJeff Kirsher } 4388*527a6266SJeff Kirsher 4389*527a6266SJeff Kirsher if (buf[0] != VPD_MAGIC) { 4390*527a6266SJeff Kirsher seq_printf(seq, "VPD tag mismatch: %#x\n", buf[0]); 4391*527a6266SJeff Kirsher goto out; 4392*527a6266SJeff Kirsher } 4393*527a6266SJeff Kirsher len = buf[1]; 4394*527a6266SJeff Kirsher if (len == 0 || len > vpd_size - 4) { 4395*527a6266SJeff Kirsher seq_printf(seq, "Invalid id length: %d\n", len); 4396*527a6266SJeff Kirsher goto out; 4397*527a6266SJeff Kirsher } 4398*527a6266SJeff Kirsher 4399*527a6266SJeff Kirsher seq_printf(seq, "%.*s\n", len, buf + 3); 4400*527a6266SJeff Kirsher offs = len + 3; 4401*527a6266SJeff Kirsher 4402*527a6266SJeff Kirsher while (offs < vpd_size - 4) { 4403*527a6266SJeff Kirsher int i; 4404*527a6266SJeff Kirsher 4405*527a6266SJeff Kirsher if (!memcmp("RW", buf + offs, 2)) /* end marker */ 4406*527a6266SJeff Kirsher break; 4407*527a6266SJeff Kirsher len = buf[offs + 2]; 4408*527a6266SJeff Kirsher if (offs + len + 3 >= vpd_size) 4409*527a6266SJeff Kirsher break; 4410*527a6266SJeff Kirsher 4411*527a6266SJeff Kirsher for (i = 0; i < ARRAY_SIZE(vpd_tags); i++) { 4412*527a6266SJeff Kirsher if (!memcmp(vpd_tags[i].tag, buf + offs, 2)) { 4413*527a6266SJeff Kirsher seq_printf(seq, " %s: %.*s\n", 4414*527a6266SJeff Kirsher vpd_tags[i].label, len, buf + offs + 3); 4415*527a6266SJeff Kirsher break; 4416*527a6266SJeff Kirsher } 4417*527a6266SJeff Kirsher } 4418*527a6266SJeff Kirsher offs += len + 3; 4419*527a6266SJeff Kirsher } 4420*527a6266SJeff Kirsher out: 4421*527a6266SJeff Kirsher kfree(buf); 4422*527a6266SJeff Kirsher } 4423*527a6266SJeff Kirsher 4424*527a6266SJeff Kirsher static int sky2_debug_show(struct seq_file *seq, void *v) 4425*527a6266SJeff Kirsher { 4426*527a6266SJeff Kirsher struct net_device *dev = seq->private; 4427*527a6266SJeff Kirsher const struct sky2_port *sky2 = netdev_priv(dev); 4428*527a6266SJeff Kirsher struct sky2_hw *hw = sky2->hw; 4429*527a6266SJeff Kirsher unsigned port = sky2->port; 4430*527a6266SJeff Kirsher unsigned idx, last; 4431*527a6266SJeff Kirsher int sop; 4432*527a6266SJeff Kirsher 4433*527a6266SJeff Kirsher sky2_show_vpd(seq, hw); 4434*527a6266SJeff Kirsher 4435*527a6266SJeff Kirsher seq_printf(seq, "\nIRQ src=%x mask=%x control=%x\n", 4436*527a6266SJeff Kirsher sky2_read32(hw, B0_ISRC), 4437*527a6266SJeff Kirsher sky2_read32(hw, B0_IMSK), 4438*527a6266SJeff Kirsher sky2_read32(hw, B0_Y2_SP_ICR)); 4439*527a6266SJeff Kirsher 4440*527a6266SJeff Kirsher if (!netif_running(dev)) { 4441*527a6266SJeff Kirsher seq_printf(seq, "network not running\n"); 4442*527a6266SJeff Kirsher return 0; 4443*527a6266SJeff Kirsher } 4444*527a6266SJeff Kirsher 4445*527a6266SJeff Kirsher napi_disable(&hw->napi); 4446*527a6266SJeff Kirsher last = sky2_read16(hw, STAT_PUT_IDX); 4447*527a6266SJeff Kirsher 4448*527a6266SJeff Kirsher seq_printf(seq, "Status ring %u\n", hw->st_size); 4449*527a6266SJeff Kirsher if (hw->st_idx == last) 4450*527a6266SJeff Kirsher seq_puts(seq, "Status ring (empty)\n"); 4451*527a6266SJeff Kirsher else { 4452*527a6266SJeff Kirsher seq_puts(seq, "Status ring\n"); 4453*527a6266SJeff Kirsher for (idx = hw->st_idx; idx != last && idx < hw->st_size; 4454*527a6266SJeff Kirsher idx = RING_NEXT(idx, hw->st_size)) { 4455*527a6266SJeff Kirsher const struct sky2_status_le *le = hw->st_le + idx; 4456*527a6266SJeff Kirsher seq_printf(seq, "[%d] %#x %d %#x\n", 4457*527a6266SJeff Kirsher idx, le->opcode, le->length, le->status); 4458*527a6266SJeff Kirsher } 4459*527a6266SJeff Kirsher seq_puts(seq, "\n"); 4460*527a6266SJeff Kirsher } 4461*527a6266SJeff Kirsher 4462*527a6266SJeff Kirsher seq_printf(seq, "Tx ring pending=%u...%u report=%d done=%d\n", 4463*527a6266SJeff Kirsher sky2->tx_cons, sky2->tx_prod, 4464*527a6266SJeff Kirsher sky2_read16(hw, port == 0 ? STAT_TXA1_RIDX : STAT_TXA2_RIDX), 4465*527a6266SJeff Kirsher sky2_read16(hw, Q_ADDR(txqaddr[port], Q_DONE))); 4466*527a6266SJeff Kirsher 4467*527a6266SJeff Kirsher /* Dump contents of tx ring */ 4468*527a6266SJeff Kirsher sop = 1; 4469*527a6266SJeff Kirsher for (idx = sky2->tx_next; idx != sky2->tx_prod && idx < sky2->tx_ring_size; 4470*527a6266SJeff Kirsher idx = RING_NEXT(idx, sky2->tx_ring_size)) { 4471*527a6266SJeff Kirsher const struct sky2_tx_le *le = sky2->tx_le + idx; 4472*527a6266SJeff Kirsher u32 a = le32_to_cpu(le->addr); 4473*527a6266SJeff Kirsher 4474*527a6266SJeff Kirsher if (sop) 4475*527a6266SJeff Kirsher seq_printf(seq, "%u:", idx); 4476*527a6266SJeff Kirsher sop = 0; 4477*527a6266SJeff Kirsher 4478*527a6266SJeff Kirsher switch (le->opcode & ~HW_OWNER) { 4479*527a6266SJeff Kirsher case OP_ADDR64: 4480*527a6266SJeff Kirsher seq_printf(seq, " %#x:", a); 4481*527a6266SJeff Kirsher break; 4482*527a6266SJeff Kirsher case OP_LRGLEN: 4483*527a6266SJeff Kirsher seq_printf(seq, " mtu=%d", a); 4484*527a6266SJeff Kirsher break; 4485*527a6266SJeff Kirsher case OP_VLAN: 4486*527a6266SJeff Kirsher seq_printf(seq, " vlan=%d", be16_to_cpu(le->length)); 4487*527a6266SJeff Kirsher break; 4488*527a6266SJeff Kirsher case OP_TCPLISW: 4489*527a6266SJeff Kirsher seq_printf(seq, " csum=%#x", a); 4490*527a6266SJeff Kirsher break; 4491*527a6266SJeff Kirsher case OP_LARGESEND: 4492*527a6266SJeff Kirsher seq_printf(seq, " tso=%#x(%d)", a, le16_to_cpu(le->length)); 4493*527a6266SJeff Kirsher break; 4494*527a6266SJeff Kirsher case OP_PACKET: 4495*527a6266SJeff Kirsher seq_printf(seq, " %#x(%d)", a, le16_to_cpu(le->length)); 4496*527a6266SJeff Kirsher break; 4497*527a6266SJeff Kirsher case OP_BUFFER: 4498*527a6266SJeff Kirsher seq_printf(seq, " frag=%#x(%d)", a, le16_to_cpu(le->length)); 4499*527a6266SJeff Kirsher break; 4500*527a6266SJeff Kirsher default: 4501*527a6266SJeff Kirsher seq_printf(seq, " op=%#x,%#x(%d)", le->opcode, 4502*527a6266SJeff Kirsher a, le16_to_cpu(le->length)); 4503*527a6266SJeff Kirsher } 4504*527a6266SJeff Kirsher 4505*527a6266SJeff Kirsher if (le->ctrl & EOP) { 4506*527a6266SJeff Kirsher seq_putc(seq, '\n'); 4507*527a6266SJeff Kirsher sop = 1; 4508*527a6266SJeff Kirsher } 4509*527a6266SJeff Kirsher } 4510*527a6266SJeff Kirsher 4511*527a6266SJeff Kirsher seq_printf(seq, "\nRx ring hw get=%d put=%d last=%d\n", 4512*527a6266SJeff Kirsher sky2_read16(hw, Y2_QADDR(rxqaddr[port], PREF_UNIT_GET_IDX)), 4513*527a6266SJeff Kirsher sky2_read16(hw, Y2_QADDR(rxqaddr[port], PREF_UNIT_PUT_IDX)), 4514*527a6266SJeff Kirsher sky2_read16(hw, Y2_QADDR(rxqaddr[port], PREF_UNIT_LAST_IDX))); 4515*527a6266SJeff Kirsher 4516*527a6266SJeff Kirsher sky2_read32(hw, B0_Y2_SP_LISR); 4517*527a6266SJeff Kirsher napi_enable(&hw->napi); 4518*527a6266SJeff Kirsher return 0; 4519*527a6266SJeff Kirsher } 4520*527a6266SJeff Kirsher 4521*527a6266SJeff Kirsher static int sky2_debug_open(struct inode *inode, struct file *file) 4522*527a6266SJeff Kirsher { 4523*527a6266SJeff Kirsher return single_open(file, sky2_debug_show, inode->i_private); 4524*527a6266SJeff Kirsher } 4525*527a6266SJeff Kirsher 4526*527a6266SJeff Kirsher static const struct file_operations sky2_debug_fops = { 4527*527a6266SJeff Kirsher .owner = THIS_MODULE, 4528*527a6266SJeff Kirsher .open = sky2_debug_open, 4529*527a6266SJeff Kirsher .read = seq_read, 4530*527a6266SJeff Kirsher .llseek = seq_lseek, 4531*527a6266SJeff Kirsher .release = single_release, 4532*527a6266SJeff Kirsher }; 4533*527a6266SJeff Kirsher 4534*527a6266SJeff Kirsher /* 4535*527a6266SJeff Kirsher * Use network device events to create/remove/rename 4536*527a6266SJeff Kirsher * debugfs file entries 4537*527a6266SJeff Kirsher */ 4538*527a6266SJeff Kirsher static int sky2_device_event(struct notifier_block *unused, 4539*527a6266SJeff Kirsher unsigned long event, void *ptr) 4540*527a6266SJeff Kirsher { 4541*527a6266SJeff Kirsher struct net_device *dev = ptr; 4542*527a6266SJeff Kirsher struct sky2_port *sky2 = netdev_priv(dev); 4543*527a6266SJeff Kirsher 4544*527a6266SJeff Kirsher if (dev->netdev_ops->ndo_open != sky2_up || !sky2_debug) 4545*527a6266SJeff Kirsher return NOTIFY_DONE; 4546*527a6266SJeff Kirsher 4547*527a6266SJeff Kirsher switch (event) { 4548*527a6266SJeff Kirsher case NETDEV_CHANGENAME: 4549*527a6266SJeff Kirsher if (sky2->debugfs) { 4550*527a6266SJeff Kirsher sky2->debugfs = debugfs_rename(sky2_debug, sky2->debugfs, 4551*527a6266SJeff Kirsher sky2_debug, dev->name); 4552*527a6266SJeff Kirsher } 4553*527a6266SJeff Kirsher break; 4554*527a6266SJeff Kirsher 4555*527a6266SJeff Kirsher case NETDEV_GOING_DOWN: 4556*527a6266SJeff Kirsher if (sky2->debugfs) { 4557*527a6266SJeff Kirsher netdev_printk(KERN_DEBUG, dev, "remove debugfs\n"); 4558*527a6266SJeff Kirsher debugfs_remove(sky2->debugfs); 4559*527a6266SJeff Kirsher sky2->debugfs = NULL; 4560*527a6266SJeff Kirsher } 4561*527a6266SJeff Kirsher break; 4562*527a6266SJeff Kirsher 4563*527a6266SJeff Kirsher case NETDEV_UP: 4564*527a6266SJeff Kirsher sky2->debugfs = debugfs_create_file(dev->name, S_IRUGO, 4565*527a6266SJeff Kirsher sky2_debug, dev, 4566*527a6266SJeff Kirsher &sky2_debug_fops); 4567*527a6266SJeff Kirsher if (IS_ERR(sky2->debugfs)) 4568*527a6266SJeff Kirsher sky2->debugfs = NULL; 4569*527a6266SJeff Kirsher } 4570*527a6266SJeff Kirsher 4571*527a6266SJeff Kirsher return NOTIFY_DONE; 4572*527a6266SJeff Kirsher } 4573*527a6266SJeff Kirsher 4574*527a6266SJeff Kirsher static struct notifier_block sky2_notifier = { 4575*527a6266SJeff Kirsher .notifier_call = sky2_device_event, 4576*527a6266SJeff Kirsher }; 4577*527a6266SJeff Kirsher 4578*527a6266SJeff Kirsher 4579*527a6266SJeff Kirsher static __init void sky2_debug_init(void) 4580*527a6266SJeff Kirsher { 4581*527a6266SJeff Kirsher struct dentry *ent; 4582*527a6266SJeff Kirsher 4583*527a6266SJeff Kirsher ent = debugfs_create_dir("sky2", NULL); 4584*527a6266SJeff Kirsher if (!ent || IS_ERR(ent)) 4585*527a6266SJeff Kirsher return; 4586*527a6266SJeff Kirsher 4587*527a6266SJeff Kirsher sky2_debug = ent; 4588*527a6266SJeff Kirsher register_netdevice_notifier(&sky2_notifier); 4589*527a6266SJeff Kirsher } 4590*527a6266SJeff Kirsher 4591*527a6266SJeff Kirsher static __exit void sky2_debug_cleanup(void) 4592*527a6266SJeff Kirsher { 4593*527a6266SJeff Kirsher if (sky2_debug) { 4594*527a6266SJeff Kirsher unregister_netdevice_notifier(&sky2_notifier); 4595*527a6266SJeff Kirsher debugfs_remove(sky2_debug); 4596*527a6266SJeff Kirsher sky2_debug = NULL; 4597*527a6266SJeff Kirsher } 4598*527a6266SJeff Kirsher } 4599*527a6266SJeff Kirsher 4600*527a6266SJeff Kirsher #else 4601*527a6266SJeff Kirsher #define sky2_debug_init() 4602*527a6266SJeff Kirsher #define sky2_debug_cleanup() 4603*527a6266SJeff Kirsher #endif 4604*527a6266SJeff Kirsher 4605*527a6266SJeff Kirsher /* Two copies of network device operations to handle special case of 4606*527a6266SJeff Kirsher not allowing netpoll on second port */ 4607*527a6266SJeff Kirsher static const struct net_device_ops sky2_netdev_ops[2] = { 4608*527a6266SJeff Kirsher { 4609*527a6266SJeff Kirsher .ndo_open = sky2_up, 4610*527a6266SJeff Kirsher .ndo_stop = sky2_down, 4611*527a6266SJeff Kirsher .ndo_start_xmit = sky2_xmit_frame, 4612*527a6266SJeff Kirsher .ndo_do_ioctl = sky2_ioctl, 4613*527a6266SJeff Kirsher .ndo_validate_addr = eth_validate_addr, 4614*527a6266SJeff Kirsher .ndo_set_mac_address = sky2_set_mac_address, 4615*527a6266SJeff Kirsher .ndo_set_multicast_list = sky2_set_multicast, 4616*527a6266SJeff Kirsher .ndo_change_mtu = sky2_change_mtu, 4617*527a6266SJeff Kirsher .ndo_fix_features = sky2_fix_features, 4618*527a6266SJeff Kirsher .ndo_set_features = sky2_set_features, 4619*527a6266SJeff Kirsher .ndo_tx_timeout = sky2_tx_timeout, 4620*527a6266SJeff Kirsher .ndo_get_stats64 = sky2_get_stats, 4621*527a6266SJeff Kirsher #ifdef CONFIG_NET_POLL_CONTROLLER 4622*527a6266SJeff Kirsher .ndo_poll_controller = sky2_netpoll, 4623*527a6266SJeff Kirsher #endif 4624*527a6266SJeff Kirsher }, 4625*527a6266SJeff Kirsher { 4626*527a6266SJeff Kirsher .ndo_open = sky2_up, 4627*527a6266SJeff Kirsher .ndo_stop = sky2_down, 4628*527a6266SJeff Kirsher .ndo_start_xmit = sky2_xmit_frame, 4629*527a6266SJeff Kirsher .ndo_do_ioctl = sky2_ioctl, 4630*527a6266SJeff Kirsher .ndo_validate_addr = eth_validate_addr, 4631*527a6266SJeff Kirsher .ndo_set_mac_address = sky2_set_mac_address, 4632*527a6266SJeff Kirsher .ndo_set_multicast_list = sky2_set_multicast, 4633*527a6266SJeff Kirsher .ndo_change_mtu = sky2_change_mtu, 4634*527a6266SJeff Kirsher .ndo_fix_features = sky2_fix_features, 4635*527a6266SJeff Kirsher .ndo_set_features = sky2_set_features, 4636*527a6266SJeff Kirsher .ndo_tx_timeout = sky2_tx_timeout, 4637*527a6266SJeff Kirsher .ndo_get_stats64 = sky2_get_stats, 4638*527a6266SJeff Kirsher }, 4639*527a6266SJeff Kirsher }; 4640*527a6266SJeff Kirsher 4641*527a6266SJeff Kirsher /* Initialize network device */ 4642*527a6266SJeff Kirsher static __devinit struct net_device *sky2_init_netdev(struct sky2_hw *hw, 4643*527a6266SJeff Kirsher unsigned port, 4644*527a6266SJeff Kirsher int highmem, int wol) 4645*527a6266SJeff Kirsher { 4646*527a6266SJeff Kirsher struct sky2_port *sky2; 4647*527a6266SJeff Kirsher struct net_device *dev = alloc_etherdev(sizeof(*sky2)); 4648*527a6266SJeff Kirsher 4649*527a6266SJeff Kirsher if (!dev) { 4650*527a6266SJeff Kirsher dev_err(&hw->pdev->dev, "etherdev alloc failed\n"); 4651*527a6266SJeff Kirsher return NULL; 4652*527a6266SJeff Kirsher } 4653*527a6266SJeff Kirsher 4654*527a6266SJeff Kirsher SET_NETDEV_DEV(dev, &hw->pdev->dev); 4655*527a6266SJeff Kirsher dev->irq = hw->pdev->irq; 4656*527a6266SJeff Kirsher SET_ETHTOOL_OPS(dev, &sky2_ethtool_ops); 4657*527a6266SJeff Kirsher dev->watchdog_timeo = TX_WATCHDOG; 4658*527a6266SJeff Kirsher dev->netdev_ops = &sky2_netdev_ops[port]; 4659*527a6266SJeff Kirsher 4660*527a6266SJeff Kirsher sky2 = netdev_priv(dev); 4661*527a6266SJeff Kirsher sky2->netdev = dev; 4662*527a6266SJeff Kirsher sky2->hw = hw; 4663*527a6266SJeff Kirsher sky2->msg_enable = netif_msg_init(debug, default_msg); 4664*527a6266SJeff Kirsher 4665*527a6266SJeff Kirsher /* Auto speed and flow control */ 4666*527a6266SJeff Kirsher sky2->flags = SKY2_FLAG_AUTO_SPEED | SKY2_FLAG_AUTO_PAUSE; 4667*527a6266SJeff Kirsher if (hw->chip_id != CHIP_ID_YUKON_XL) 4668*527a6266SJeff Kirsher dev->hw_features |= NETIF_F_RXCSUM; 4669*527a6266SJeff Kirsher 4670*527a6266SJeff Kirsher sky2->flow_mode = FC_BOTH; 4671*527a6266SJeff Kirsher 4672*527a6266SJeff Kirsher sky2->duplex = -1; 4673*527a6266SJeff Kirsher sky2->speed = -1; 4674*527a6266SJeff Kirsher sky2->advertising = sky2_supported_modes(hw); 4675*527a6266SJeff Kirsher sky2->wol = wol; 4676*527a6266SJeff Kirsher 4677*527a6266SJeff Kirsher spin_lock_init(&sky2->phy_lock); 4678*527a6266SJeff Kirsher 4679*527a6266SJeff Kirsher sky2->tx_pending = TX_DEF_PENDING; 4680*527a6266SJeff Kirsher sky2->tx_ring_size = roundup_pow_of_two(TX_DEF_PENDING+1); 4681*527a6266SJeff Kirsher sky2->rx_pending = RX_DEF_PENDING; 4682*527a6266SJeff Kirsher 4683*527a6266SJeff Kirsher hw->dev[port] = dev; 4684*527a6266SJeff Kirsher 4685*527a6266SJeff Kirsher sky2->port = port; 4686*527a6266SJeff Kirsher 4687*527a6266SJeff Kirsher dev->hw_features |= NETIF_F_IP_CSUM | NETIF_F_SG | NETIF_F_TSO; 4688*527a6266SJeff Kirsher 4689*527a6266SJeff Kirsher if (highmem) 4690*527a6266SJeff Kirsher dev->features |= NETIF_F_HIGHDMA; 4691*527a6266SJeff Kirsher 4692*527a6266SJeff Kirsher /* Enable receive hashing unless hardware is known broken */ 4693*527a6266SJeff Kirsher if (!(hw->flags & SKY2_HW_RSS_BROKEN)) 4694*527a6266SJeff Kirsher dev->hw_features |= NETIF_F_RXHASH; 4695*527a6266SJeff Kirsher 4696*527a6266SJeff Kirsher if (!(hw->flags & SKY2_HW_VLAN_BROKEN)) { 4697*527a6266SJeff Kirsher dev->hw_features |= NETIF_F_HW_VLAN_TX | NETIF_F_HW_VLAN_RX; 4698*527a6266SJeff Kirsher dev->vlan_features |= SKY2_VLAN_OFFLOADS; 4699*527a6266SJeff Kirsher } 4700*527a6266SJeff Kirsher 4701*527a6266SJeff Kirsher dev->features |= dev->hw_features; 4702*527a6266SJeff Kirsher 4703*527a6266SJeff Kirsher /* read the mac address */ 4704*527a6266SJeff Kirsher memcpy_fromio(dev->dev_addr, hw->regs + B2_MAC_1 + port * 8, ETH_ALEN); 4705*527a6266SJeff Kirsher memcpy(dev->perm_addr, dev->dev_addr, dev->addr_len); 4706*527a6266SJeff Kirsher 4707*527a6266SJeff Kirsher return dev; 4708*527a6266SJeff Kirsher } 4709*527a6266SJeff Kirsher 4710*527a6266SJeff Kirsher static void __devinit sky2_show_addr(struct net_device *dev) 4711*527a6266SJeff Kirsher { 4712*527a6266SJeff Kirsher const struct sky2_port *sky2 = netdev_priv(dev); 4713*527a6266SJeff Kirsher 4714*527a6266SJeff Kirsher netif_info(sky2, probe, dev, "addr %pM\n", dev->dev_addr); 4715*527a6266SJeff Kirsher } 4716*527a6266SJeff Kirsher 4717*527a6266SJeff Kirsher /* Handle software interrupt used during MSI test */ 4718*527a6266SJeff Kirsher static irqreturn_t __devinit sky2_test_intr(int irq, void *dev_id) 4719*527a6266SJeff Kirsher { 4720*527a6266SJeff Kirsher struct sky2_hw *hw = dev_id; 4721*527a6266SJeff Kirsher u32 status = sky2_read32(hw, B0_Y2_SP_ISRC2); 4722*527a6266SJeff Kirsher 4723*527a6266SJeff Kirsher if (status == 0) 4724*527a6266SJeff Kirsher return IRQ_NONE; 4725*527a6266SJeff Kirsher 4726*527a6266SJeff Kirsher if (status & Y2_IS_IRQ_SW) { 4727*527a6266SJeff Kirsher hw->flags |= SKY2_HW_USE_MSI; 4728*527a6266SJeff Kirsher wake_up(&hw->msi_wait); 4729*527a6266SJeff Kirsher sky2_write8(hw, B0_CTST, CS_CL_SW_IRQ); 4730*527a6266SJeff Kirsher } 4731*527a6266SJeff Kirsher sky2_write32(hw, B0_Y2_SP_ICR, 2); 4732*527a6266SJeff Kirsher 4733*527a6266SJeff Kirsher return IRQ_HANDLED; 4734*527a6266SJeff Kirsher } 4735*527a6266SJeff Kirsher 4736*527a6266SJeff Kirsher /* Test interrupt path by forcing a a software IRQ */ 4737*527a6266SJeff Kirsher static int __devinit sky2_test_msi(struct sky2_hw *hw) 4738*527a6266SJeff Kirsher { 4739*527a6266SJeff Kirsher struct pci_dev *pdev = hw->pdev; 4740*527a6266SJeff Kirsher int err; 4741*527a6266SJeff Kirsher 4742*527a6266SJeff Kirsher init_waitqueue_head(&hw->msi_wait); 4743*527a6266SJeff Kirsher 4744*527a6266SJeff Kirsher sky2_write32(hw, B0_IMSK, Y2_IS_IRQ_SW); 4745*527a6266SJeff Kirsher 4746*527a6266SJeff Kirsher err = request_irq(pdev->irq, sky2_test_intr, 0, DRV_NAME, hw); 4747*527a6266SJeff Kirsher if (err) { 4748*527a6266SJeff Kirsher dev_err(&pdev->dev, "cannot assign irq %d\n", pdev->irq); 4749*527a6266SJeff Kirsher return err; 4750*527a6266SJeff Kirsher } 4751*527a6266SJeff Kirsher 4752*527a6266SJeff Kirsher sky2_write8(hw, B0_CTST, CS_ST_SW_IRQ); 4753*527a6266SJeff Kirsher sky2_read8(hw, B0_CTST); 4754*527a6266SJeff Kirsher 4755*527a6266SJeff Kirsher wait_event_timeout(hw->msi_wait, (hw->flags & SKY2_HW_USE_MSI), HZ/10); 4756*527a6266SJeff Kirsher 4757*527a6266SJeff Kirsher if (!(hw->flags & SKY2_HW_USE_MSI)) { 4758*527a6266SJeff Kirsher /* MSI test failed, go back to INTx mode */ 4759*527a6266SJeff Kirsher dev_info(&pdev->dev, "No interrupt generated using MSI, " 4760*527a6266SJeff Kirsher "switching to INTx mode.\n"); 4761*527a6266SJeff Kirsher 4762*527a6266SJeff Kirsher err = -EOPNOTSUPP; 4763*527a6266SJeff Kirsher sky2_write8(hw, B0_CTST, CS_CL_SW_IRQ); 4764*527a6266SJeff Kirsher } 4765*527a6266SJeff Kirsher 4766*527a6266SJeff Kirsher sky2_write32(hw, B0_IMSK, 0); 4767*527a6266SJeff Kirsher sky2_read32(hw, B0_IMSK); 4768*527a6266SJeff Kirsher 4769*527a6266SJeff Kirsher free_irq(pdev->irq, hw); 4770*527a6266SJeff Kirsher 4771*527a6266SJeff Kirsher return err; 4772*527a6266SJeff Kirsher } 4773*527a6266SJeff Kirsher 4774*527a6266SJeff Kirsher /* This driver supports yukon2 chipset only */ 4775*527a6266SJeff Kirsher static const char *sky2_name(u8 chipid, char *buf, int sz) 4776*527a6266SJeff Kirsher { 4777*527a6266SJeff Kirsher const char *name[] = { 4778*527a6266SJeff Kirsher "XL", /* 0xb3 */ 4779*527a6266SJeff Kirsher "EC Ultra", /* 0xb4 */ 4780*527a6266SJeff Kirsher "Extreme", /* 0xb5 */ 4781*527a6266SJeff Kirsher "EC", /* 0xb6 */ 4782*527a6266SJeff Kirsher "FE", /* 0xb7 */ 4783*527a6266SJeff Kirsher "FE+", /* 0xb8 */ 4784*527a6266SJeff Kirsher "Supreme", /* 0xb9 */ 4785*527a6266SJeff Kirsher "UL 2", /* 0xba */ 4786*527a6266SJeff Kirsher "Unknown", /* 0xbb */ 4787*527a6266SJeff Kirsher "Optima", /* 0xbc */ 4788*527a6266SJeff Kirsher "Optima Prime", /* 0xbd */ 4789*527a6266SJeff Kirsher "Optima 2", /* 0xbe */ 4790*527a6266SJeff Kirsher }; 4791*527a6266SJeff Kirsher 4792*527a6266SJeff Kirsher if (chipid >= CHIP_ID_YUKON_XL && chipid <= CHIP_ID_YUKON_OP_2) 4793*527a6266SJeff Kirsher strncpy(buf, name[chipid - CHIP_ID_YUKON_XL], sz); 4794*527a6266SJeff Kirsher else 4795*527a6266SJeff Kirsher snprintf(buf, sz, "(chip %#x)", chipid); 4796*527a6266SJeff Kirsher return buf; 4797*527a6266SJeff Kirsher } 4798*527a6266SJeff Kirsher 4799*527a6266SJeff Kirsher static int __devinit sky2_probe(struct pci_dev *pdev, 4800*527a6266SJeff Kirsher const struct pci_device_id *ent) 4801*527a6266SJeff Kirsher { 4802*527a6266SJeff Kirsher struct net_device *dev; 4803*527a6266SJeff Kirsher struct sky2_hw *hw; 4804*527a6266SJeff Kirsher int err, using_dac = 0, wol_default; 4805*527a6266SJeff Kirsher u32 reg; 4806*527a6266SJeff Kirsher char buf1[16]; 4807*527a6266SJeff Kirsher 4808*527a6266SJeff Kirsher err = pci_enable_device(pdev); 4809*527a6266SJeff Kirsher if (err) { 4810*527a6266SJeff Kirsher dev_err(&pdev->dev, "cannot enable PCI device\n"); 4811*527a6266SJeff Kirsher goto err_out; 4812*527a6266SJeff Kirsher } 4813*527a6266SJeff Kirsher 4814*527a6266SJeff Kirsher /* Get configuration information 4815*527a6266SJeff Kirsher * Note: only regular PCI config access once to test for HW issues 4816*527a6266SJeff Kirsher * other PCI access through shared memory for speed and to 4817*527a6266SJeff Kirsher * avoid MMCONFIG problems. 4818*527a6266SJeff Kirsher */ 4819*527a6266SJeff Kirsher err = pci_read_config_dword(pdev, PCI_DEV_REG2, ®); 4820*527a6266SJeff Kirsher if (err) { 4821*527a6266SJeff Kirsher dev_err(&pdev->dev, "PCI read config failed\n"); 4822*527a6266SJeff Kirsher goto err_out; 4823*527a6266SJeff Kirsher } 4824*527a6266SJeff Kirsher 4825*527a6266SJeff Kirsher if (~reg == 0) { 4826*527a6266SJeff Kirsher dev_err(&pdev->dev, "PCI configuration read error\n"); 4827*527a6266SJeff Kirsher goto err_out; 4828*527a6266SJeff Kirsher } 4829*527a6266SJeff Kirsher 4830*527a6266SJeff Kirsher err = pci_request_regions(pdev, DRV_NAME); 4831*527a6266SJeff Kirsher if (err) { 4832*527a6266SJeff Kirsher dev_err(&pdev->dev, "cannot obtain PCI resources\n"); 4833*527a6266SJeff Kirsher goto err_out_disable; 4834*527a6266SJeff Kirsher } 4835*527a6266SJeff Kirsher 4836*527a6266SJeff Kirsher pci_set_master(pdev); 4837*527a6266SJeff Kirsher 4838*527a6266SJeff Kirsher if (sizeof(dma_addr_t) > sizeof(u32) && 4839*527a6266SJeff Kirsher !(err = pci_set_dma_mask(pdev, DMA_BIT_MASK(64)))) { 4840*527a6266SJeff Kirsher using_dac = 1; 4841*527a6266SJeff Kirsher err = pci_set_consistent_dma_mask(pdev, DMA_BIT_MASK(64)); 4842*527a6266SJeff Kirsher if (err < 0) { 4843*527a6266SJeff Kirsher dev_err(&pdev->dev, "unable to obtain 64 bit DMA " 4844*527a6266SJeff Kirsher "for consistent allocations\n"); 4845*527a6266SJeff Kirsher goto err_out_free_regions; 4846*527a6266SJeff Kirsher } 4847*527a6266SJeff Kirsher } else { 4848*527a6266SJeff Kirsher err = pci_set_dma_mask(pdev, DMA_BIT_MASK(32)); 4849*527a6266SJeff Kirsher if (err) { 4850*527a6266SJeff Kirsher dev_err(&pdev->dev, "no usable DMA configuration\n"); 4851*527a6266SJeff Kirsher goto err_out_free_regions; 4852*527a6266SJeff Kirsher } 4853*527a6266SJeff Kirsher } 4854*527a6266SJeff Kirsher 4855*527a6266SJeff Kirsher 4856*527a6266SJeff Kirsher #ifdef __BIG_ENDIAN 4857*527a6266SJeff Kirsher /* The sk98lin vendor driver uses hardware byte swapping but 4858*527a6266SJeff Kirsher * this driver uses software swapping. 4859*527a6266SJeff Kirsher */ 4860*527a6266SJeff Kirsher reg &= ~PCI_REV_DESC; 4861*527a6266SJeff Kirsher err = pci_write_config_dword(pdev, PCI_DEV_REG2, reg); 4862*527a6266SJeff Kirsher if (err) { 4863*527a6266SJeff Kirsher dev_err(&pdev->dev, "PCI write config failed\n"); 4864*527a6266SJeff Kirsher goto err_out_free_regions; 4865*527a6266SJeff Kirsher } 4866*527a6266SJeff Kirsher #endif 4867*527a6266SJeff Kirsher 4868*527a6266SJeff Kirsher wol_default = device_may_wakeup(&pdev->dev) ? WAKE_MAGIC : 0; 4869*527a6266SJeff Kirsher 4870*527a6266SJeff Kirsher err = -ENOMEM; 4871*527a6266SJeff Kirsher 4872*527a6266SJeff Kirsher hw = kzalloc(sizeof(*hw) + strlen(DRV_NAME "@pci:") 4873*527a6266SJeff Kirsher + strlen(pci_name(pdev)) + 1, GFP_KERNEL); 4874*527a6266SJeff Kirsher if (!hw) { 4875*527a6266SJeff Kirsher dev_err(&pdev->dev, "cannot allocate hardware struct\n"); 4876*527a6266SJeff Kirsher goto err_out_free_regions; 4877*527a6266SJeff Kirsher } 4878*527a6266SJeff Kirsher 4879*527a6266SJeff Kirsher hw->pdev = pdev; 4880*527a6266SJeff Kirsher sprintf(hw->irq_name, DRV_NAME "@pci:%s", pci_name(pdev)); 4881*527a6266SJeff Kirsher 4882*527a6266SJeff Kirsher hw->regs = ioremap_nocache(pci_resource_start(pdev, 0), 0x4000); 4883*527a6266SJeff Kirsher if (!hw->regs) { 4884*527a6266SJeff Kirsher dev_err(&pdev->dev, "cannot map device registers\n"); 4885*527a6266SJeff Kirsher goto err_out_free_hw; 4886*527a6266SJeff Kirsher } 4887*527a6266SJeff Kirsher 4888*527a6266SJeff Kirsher err = sky2_init(hw); 4889*527a6266SJeff Kirsher if (err) 4890*527a6266SJeff Kirsher goto err_out_iounmap; 4891*527a6266SJeff Kirsher 4892*527a6266SJeff Kirsher /* ring for status responses */ 4893*527a6266SJeff Kirsher hw->st_size = hw->ports * roundup_pow_of_two(3*RX_MAX_PENDING + TX_MAX_PENDING); 4894*527a6266SJeff Kirsher hw->st_le = pci_alloc_consistent(pdev, hw->st_size * sizeof(struct sky2_status_le), 4895*527a6266SJeff Kirsher &hw->st_dma); 4896*527a6266SJeff Kirsher if (!hw->st_le) 4897*527a6266SJeff Kirsher goto err_out_reset; 4898*527a6266SJeff Kirsher 4899*527a6266SJeff Kirsher dev_info(&pdev->dev, "Yukon-2 %s chip revision %d\n", 4900*527a6266SJeff Kirsher sky2_name(hw->chip_id, buf1, sizeof(buf1)), hw->chip_rev); 4901*527a6266SJeff Kirsher 4902*527a6266SJeff Kirsher sky2_reset(hw); 4903*527a6266SJeff Kirsher 4904*527a6266SJeff Kirsher dev = sky2_init_netdev(hw, 0, using_dac, wol_default); 4905*527a6266SJeff Kirsher if (!dev) { 4906*527a6266SJeff Kirsher err = -ENOMEM; 4907*527a6266SJeff Kirsher goto err_out_free_pci; 4908*527a6266SJeff Kirsher } 4909*527a6266SJeff Kirsher 4910*527a6266SJeff Kirsher if (!disable_msi && pci_enable_msi(pdev) == 0) { 4911*527a6266SJeff Kirsher err = sky2_test_msi(hw); 4912*527a6266SJeff Kirsher if (err == -EOPNOTSUPP) 4913*527a6266SJeff Kirsher pci_disable_msi(pdev); 4914*527a6266SJeff Kirsher else if (err) 4915*527a6266SJeff Kirsher goto err_out_free_netdev; 4916*527a6266SJeff Kirsher } 4917*527a6266SJeff Kirsher 4918*527a6266SJeff Kirsher err = register_netdev(dev); 4919*527a6266SJeff Kirsher if (err) { 4920*527a6266SJeff Kirsher dev_err(&pdev->dev, "cannot register net device\n"); 4921*527a6266SJeff Kirsher goto err_out_free_netdev; 4922*527a6266SJeff Kirsher } 4923*527a6266SJeff Kirsher 4924*527a6266SJeff Kirsher netif_carrier_off(dev); 4925*527a6266SJeff Kirsher 4926*527a6266SJeff Kirsher netif_napi_add(dev, &hw->napi, sky2_poll, NAPI_WEIGHT); 4927*527a6266SJeff Kirsher 4928*527a6266SJeff Kirsher err = request_irq(pdev->irq, sky2_intr, 4929*527a6266SJeff Kirsher (hw->flags & SKY2_HW_USE_MSI) ? 0 : IRQF_SHARED, 4930*527a6266SJeff Kirsher hw->irq_name, hw); 4931*527a6266SJeff Kirsher if (err) { 4932*527a6266SJeff Kirsher dev_err(&pdev->dev, "cannot assign irq %d\n", pdev->irq); 4933*527a6266SJeff Kirsher goto err_out_unregister; 4934*527a6266SJeff Kirsher } 4935*527a6266SJeff Kirsher sky2_write32(hw, B0_IMSK, Y2_IS_BASE); 4936*527a6266SJeff Kirsher napi_enable(&hw->napi); 4937*527a6266SJeff Kirsher 4938*527a6266SJeff Kirsher sky2_show_addr(dev); 4939*527a6266SJeff Kirsher 4940*527a6266SJeff Kirsher if (hw->ports > 1) { 4941*527a6266SJeff Kirsher struct net_device *dev1; 4942*527a6266SJeff Kirsher 4943*527a6266SJeff Kirsher err = -ENOMEM; 4944*527a6266SJeff Kirsher dev1 = sky2_init_netdev(hw, 1, using_dac, wol_default); 4945*527a6266SJeff Kirsher if (dev1 && (err = register_netdev(dev1)) == 0) 4946*527a6266SJeff Kirsher sky2_show_addr(dev1); 4947*527a6266SJeff Kirsher else { 4948*527a6266SJeff Kirsher dev_warn(&pdev->dev, 4949*527a6266SJeff Kirsher "register of second port failed (%d)\n", err); 4950*527a6266SJeff Kirsher hw->dev[1] = NULL; 4951*527a6266SJeff Kirsher hw->ports = 1; 4952*527a6266SJeff Kirsher if (dev1) 4953*527a6266SJeff Kirsher free_netdev(dev1); 4954*527a6266SJeff Kirsher } 4955*527a6266SJeff Kirsher } 4956*527a6266SJeff Kirsher 4957*527a6266SJeff Kirsher setup_timer(&hw->watchdog_timer, sky2_watchdog, (unsigned long) hw); 4958*527a6266SJeff Kirsher INIT_WORK(&hw->restart_work, sky2_restart); 4959*527a6266SJeff Kirsher 4960*527a6266SJeff Kirsher pci_set_drvdata(pdev, hw); 4961*527a6266SJeff Kirsher pdev->d3_delay = 150; 4962*527a6266SJeff Kirsher 4963*527a6266SJeff Kirsher return 0; 4964*527a6266SJeff Kirsher 4965*527a6266SJeff Kirsher err_out_unregister: 4966*527a6266SJeff Kirsher if (hw->flags & SKY2_HW_USE_MSI) 4967*527a6266SJeff Kirsher pci_disable_msi(pdev); 4968*527a6266SJeff Kirsher unregister_netdev(dev); 4969*527a6266SJeff Kirsher err_out_free_netdev: 4970*527a6266SJeff Kirsher free_netdev(dev); 4971*527a6266SJeff Kirsher err_out_free_pci: 4972*527a6266SJeff Kirsher pci_free_consistent(pdev, hw->st_size * sizeof(struct sky2_status_le), 4973*527a6266SJeff Kirsher hw->st_le, hw->st_dma); 4974*527a6266SJeff Kirsher err_out_reset: 4975*527a6266SJeff Kirsher sky2_write8(hw, B0_CTST, CS_RST_SET); 4976*527a6266SJeff Kirsher err_out_iounmap: 4977*527a6266SJeff Kirsher iounmap(hw->regs); 4978*527a6266SJeff Kirsher err_out_free_hw: 4979*527a6266SJeff Kirsher kfree(hw); 4980*527a6266SJeff Kirsher err_out_free_regions: 4981*527a6266SJeff Kirsher pci_release_regions(pdev); 4982*527a6266SJeff Kirsher err_out_disable: 4983*527a6266SJeff Kirsher pci_disable_device(pdev); 4984*527a6266SJeff Kirsher err_out: 4985*527a6266SJeff Kirsher pci_set_drvdata(pdev, NULL); 4986*527a6266SJeff Kirsher return err; 4987*527a6266SJeff Kirsher } 4988*527a6266SJeff Kirsher 4989*527a6266SJeff Kirsher static void __devexit sky2_remove(struct pci_dev *pdev) 4990*527a6266SJeff Kirsher { 4991*527a6266SJeff Kirsher struct sky2_hw *hw = pci_get_drvdata(pdev); 4992*527a6266SJeff Kirsher int i; 4993*527a6266SJeff Kirsher 4994*527a6266SJeff Kirsher if (!hw) 4995*527a6266SJeff Kirsher return; 4996*527a6266SJeff Kirsher 4997*527a6266SJeff Kirsher del_timer_sync(&hw->watchdog_timer); 4998*527a6266SJeff Kirsher cancel_work_sync(&hw->restart_work); 4999*527a6266SJeff Kirsher 5000*527a6266SJeff Kirsher for (i = hw->ports-1; i >= 0; --i) 5001*527a6266SJeff Kirsher unregister_netdev(hw->dev[i]); 5002*527a6266SJeff Kirsher 5003*527a6266SJeff Kirsher sky2_write32(hw, B0_IMSK, 0); 5004*527a6266SJeff Kirsher 5005*527a6266SJeff Kirsher sky2_power_aux(hw); 5006*527a6266SJeff Kirsher 5007*527a6266SJeff Kirsher sky2_write8(hw, B0_CTST, CS_RST_SET); 5008*527a6266SJeff Kirsher sky2_read8(hw, B0_CTST); 5009*527a6266SJeff Kirsher 5010*527a6266SJeff Kirsher free_irq(pdev->irq, hw); 5011*527a6266SJeff Kirsher if (hw->flags & SKY2_HW_USE_MSI) 5012*527a6266SJeff Kirsher pci_disable_msi(pdev); 5013*527a6266SJeff Kirsher pci_free_consistent(pdev, hw->st_size * sizeof(struct sky2_status_le), 5014*527a6266SJeff Kirsher hw->st_le, hw->st_dma); 5015*527a6266SJeff Kirsher pci_release_regions(pdev); 5016*527a6266SJeff Kirsher pci_disable_device(pdev); 5017*527a6266SJeff Kirsher 5018*527a6266SJeff Kirsher for (i = hw->ports-1; i >= 0; --i) 5019*527a6266SJeff Kirsher free_netdev(hw->dev[i]); 5020*527a6266SJeff Kirsher 5021*527a6266SJeff Kirsher iounmap(hw->regs); 5022*527a6266SJeff Kirsher kfree(hw); 5023*527a6266SJeff Kirsher 5024*527a6266SJeff Kirsher pci_set_drvdata(pdev, NULL); 5025*527a6266SJeff Kirsher } 5026*527a6266SJeff Kirsher 5027*527a6266SJeff Kirsher static int sky2_suspend(struct device *dev) 5028*527a6266SJeff Kirsher { 5029*527a6266SJeff Kirsher struct pci_dev *pdev = to_pci_dev(dev); 5030*527a6266SJeff Kirsher struct sky2_hw *hw = pci_get_drvdata(pdev); 5031*527a6266SJeff Kirsher int i; 5032*527a6266SJeff Kirsher 5033*527a6266SJeff Kirsher if (!hw) 5034*527a6266SJeff Kirsher return 0; 5035*527a6266SJeff Kirsher 5036*527a6266SJeff Kirsher del_timer_sync(&hw->watchdog_timer); 5037*527a6266SJeff Kirsher cancel_work_sync(&hw->restart_work); 5038*527a6266SJeff Kirsher 5039*527a6266SJeff Kirsher rtnl_lock(); 5040*527a6266SJeff Kirsher 5041*527a6266SJeff Kirsher sky2_all_down(hw); 5042*527a6266SJeff Kirsher for (i = 0; i < hw->ports; i++) { 5043*527a6266SJeff Kirsher struct net_device *dev = hw->dev[i]; 5044*527a6266SJeff Kirsher struct sky2_port *sky2 = netdev_priv(dev); 5045*527a6266SJeff Kirsher 5046*527a6266SJeff Kirsher if (sky2->wol) 5047*527a6266SJeff Kirsher sky2_wol_init(sky2); 5048*527a6266SJeff Kirsher } 5049*527a6266SJeff Kirsher 5050*527a6266SJeff Kirsher sky2_power_aux(hw); 5051*527a6266SJeff Kirsher rtnl_unlock(); 5052*527a6266SJeff Kirsher 5053*527a6266SJeff Kirsher return 0; 5054*527a6266SJeff Kirsher } 5055*527a6266SJeff Kirsher 5056*527a6266SJeff Kirsher #ifdef CONFIG_PM_SLEEP 5057*527a6266SJeff Kirsher static int sky2_resume(struct device *dev) 5058*527a6266SJeff Kirsher { 5059*527a6266SJeff Kirsher struct pci_dev *pdev = to_pci_dev(dev); 5060*527a6266SJeff Kirsher struct sky2_hw *hw = pci_get_drvdata(pdev); 5061*527a6266SJeff Kirsher int err; 5062*527a6266SJeff Kirsher 5063*527a6266SJeff Kirsher if (!hw) 5064*527a6266SJeff Kirsher return 0; 5065*527a6266SJeff Kirsher 5066*527a6266SJeff Kirsher /* Re-enable all clocks */ 5067*527a6266SJeff Kirsher err = pci_write_config_dword(pdev, PCI_DEV_REG3, 0); 5068*527a6266SJeff Kirsher if (err) { 5069*527a6266SJeff Kirsher dev_err(&pdev->dev, "PCI write config failed\n"); 5070*527a6266SJeff Kirsher goto out; 5071*527a6266SJeff Kirsher } 5072*527a6266SJeff Kirsher 5073*527a6266SJeff Kirsher rtnl_lock(); 5074*527a6266SJeff Kirsher sky2_reset(hw); 5075*527a6266SJeff Kirsher sky2_all_up(hw); 5076*527a6266SJeff Kirsher rtnl_unlock(); 5077*527a6266SJeff Kirsher 5078*527a6266SJeff Kirsher return 0; 5079*527a6266SJeff Kirsher out: 5080*527a6266SJeff Kirsher 5081*527a6266SJeff Kirsher dev_err(&pdev->dev, "resume failed (%d)\n", err); 5082*527a6266SJeff Kirsher pci_disable_device(pdev); 5083*527a6266SJeff Kirsher return err; 5084*527a6266SJeff Kirsher } 5085*527a6266SJeff Kirsher 5086*527a6266SJeff Kirsher static SIMPLE_DEV_PM_OPS(sky2_pm_ops, sky2_suspend, sky2_resume); 5087*527a6266SJeff Kirsher #define SKY2_PM_OPS (&sky2_pm_ops) 5088*527a6266SJeff Kirsher 5089*527a6266SJeff Kirsher #else 5090*527a6266SJeff Kirsher 5091*527a6266SJeff Kirsher #define SKY2_PM_OPS NULL 5092*527a6266SJeff Kirsher #endif 5093*527a6266SJeff Kirsher 5094*527a6266SJeff Kirsher static void sky2_shutdown(struct pci_dev *pdev) 5095*527a6266SJeff Kirsher { 5096*527a6266SJeff Kirsher sky2_suspend(&pdev->dev); 5097*527a6266SJeff Kirsher pci_wake_from_d3(pdev, device_may_wakeup(&pdev->dev)); 5098*527a6266SJeff Kirsher pci_set_power_state(pdev, PCI_D3hot); 5099*527a6266SJeff Kirsher } 5100*527a6266SJeff Kirsher 5101*527a6266SJeff Kirsher static struct pci_driver sky2_driver = { 5102*527a6266SJeff Kirsher .name = DRV_NAME, 5103*527a6266SJeff Kirsher .id_table = sky2_id_table, 5104*527a6266SJeff Kirsher .probe = sky2_probe, 5105*527a6266SJeff Kirsher .remove = __devexit_p(sky2_remove), 5106*527a6266SJeff Kirsher .shutdown = sky2_shutdown, 5107*527a6266SJeff Kirsher .driver.pm = SKY2_PM_OPS, 5108*527a6266SJeff Kirsher }; 5109*527a6266SJeff Kirsher 5110*527a6266SJeff Kirsher static int __init sky2_init_module(void) 5111*527a6266SJeff Kirsher { 5112*527a6266SJeff Kirsher pr_info("driver version " DRV_VERSION "\n"); 5113*527a6266SJeff Kirsher 5114*527a6266SJeff Kirsher sky2_debug_init(); 5115*527a6266SJeff Kirsher return pci_register_driver(&sky2_driver); 5116*527a6266SJeff Kirsher } 5117*527a6266SJeff Kirsher 5118*527a6266SJeff Kirsher static void __exit sky2_cleanup_module(void) 5119*527a6266SJeff Kirsher { 5120*527a6266SJeff Kirsher pci_unregister_driver(&sky2_driver); 5121*527a6266SJeff Kirsher sky2_debug_cleanup(); 5122*527a6266SJeff Kirsher } 5123*527a6266SJeff Kirsher 5124*527a6266SJeff Kirsher module_init(sky2_init_module); 5125*527a6266SJeff Kirsher module_exit(sky2_cleanup_module); 5126*527a6266SJeff Kirsher 5127*527a6266SJeff Kirsher MODULE_DESCRIPTION("Marvell Yukon 2 Gigabit Ethernet driver"); 5128*527a6266SJeff Kirsher MODULE_AUTHOR("Stephen Hemminger <shemminger@linux-foundation.org>"); 5129*527a6266SJeff Kirsher MODULE_LICENSE("GPL"); 5130*527a6266SJeff Kirsher MODULE_VERSION(DRV_VERSION); 5131