1*df8bc08cSHitoshi Mitake /* 2*df8bc08cSHitoshi Mitake * Intel X38 Memory Controller kernel module 3*df8bc08cSHitoshi Mitake * Copyright (C) 2008 Cluster Computing, Inc. 4*df8bc08cSHitoshi Mitake * 5*df8bc08cSHitoshi Mitake * This file may be distributed under the terms of the 6*df8bc08cSHitoshi Mitake * GNU General Public License. 7*df8bc08cSHitoshi Mitake * 8*df8bc08cSHitoshi Mitake * This file is based on i3200_edac.c 9*df8bc08cSHitoshi Mitake * 10*df8bc08cSHitoshi Mitake */ 11*df8bc08cSHitoshi Mitake 12*df8bc08cSHitoshi Mitake #include <linux/module.h> 13*df8bc08cSHitoshi Mitake #include <linux/init.h> 14*df8bc08cSHitoshi Mitake #include <linux/pci.h> 15*df8bc08cSHitoshi Mitake #include <linux/pci_ids.h> 16*df8bc08cSHitoshi Mitake #include <linux/slab.h> 17*df8bc08cSHitoshi Mitake #include <linux/edac.h> 18*df8bc08cSHitoshi Mitake #include "edac_core.h" 19*df8bc08cSHitoshi Mitake 20*df8bc08cSHitoshi Mitake #define X38_REVISION "1.1" 21*df8bc08cSHitoshi Mitake 22*df8bc08cSHitoshi Mitake #define EDAC_MOD_STR "x38_edac" 23*df8bc08cSHitoshi Mitake 24*df8bc08cSHitoshi Mitake #define PCI_DEVICE_ID_INTEL_X38_HB 0x29e0 25*df8bc08cSHitoshi Mitake 26*df8bc08cSHitoshi Mitake #define X38_RANKS 8 27*df8bc08cSHitoshi Mitake #define X38_RANKS_PER_CHANNEL 4 28*df8bc08cSHitoshi Mitake #define X38_CHANNELS 2 29*df8bc08cSHitoshi Mitake 30*df8bc08cSHitoshi Mitake /* Intel X38 register addresses - device 0 function 0 - DRAM Controller */ 31*df8bc08cSHitoshi Mitake 32*df8bc08cSHitoshi Mitake #define X38_MCHBAR_LOW 0x48 /* MCH Memory Mapped Register BAR */ 33*df8bc08cSHitoshi Mitake #define X38_MCHBAR_HIGH 0x4b 34*df8bc08cSHitoshi Mitake #define X38_MCHBAR_MASK 0xfffffc000ULL /* bits 35:14 */ 35*df8bc08cSHitoshi Mitake #define X38_MMR_WINDOW_SIZE 16384 36*df8bc08cSHitoshi Mitake 37*df8bc08cSHitoshi Mitake #define X38_TOM 0xa0 /* Top of Memory (16b) 38*df8bc08cSHitoshi Mitake * 39*df8bc08cSHitoshi Mitake * 15:10 reserved 40*df8bc08cSHitoshi Mitake * 9:0 total populated physical memory 41*df8bc08cSHitoshi Mitake */ 42*df8bc08cSHitoshi Mitake #define X38_TOM_MASK 0x3ff /* bits 9:0 */ 43*df8bc08cSHitoshi Mitake #define X38_TOM_SHIFT 26 /* 64MiB grain */ 44*df8bc08cSHitoshi Mitake 45*df8bc08cSHitoshi Mitake #define X38_ERRSTS 0xc8 /* Error Status Register (16b) 46*df8bc08cSHitoshi Mitake * 47*df8bc08cSHitoshi Mitake * 15 reserved 48*df8bc08cSHitoshi Mitake * 14 Isochronous TBWRR Run Behind FIFO Full 49*df8bc08cSHitoshi Mitake * (ITCV) 50*df8bc08cSHitoshi Mitake * 13 Isochronous TBWRR Run Behind FIFO Put 51*df8bc08cSHitoshi Mitake * (ITSTV) 52*df8bc08cSHitoshi Mitake * 12 reserved 53*df8bc08cSHitoshi Mitake * 11 MCH Thermal Sensor Event 54*df8bc08cSHitoshi Mitake * for SMI/SCI/SERR (GTSE) 55*df8bc08cSHitoshi Mitake * 10 reserved 56*df8bc08cSHitoshi Mitake * 9 LOCK to non-DRAM Memory Flag (LCKF) 57*df8bc08cSHitoshi Mitake * 8 reserved 58*df8bc08cSHitoshi Mitake * 7 DRAM Throttle Flag (DTF) 59*df8bc08cSHitoshi Mitake * 6:2 reserved 60*df8bc08cSHitoshi Mitake * 1 Multi-bit DRAM ECC Error Flag (DMERR) 61*df8bc08cSHitoshi Mitake * 0 Single-bit DRAM ECC Error Flag (DSERR) 62*df8bc08cSHitoshi Mitake */ 63*df8bc08cSHitoshi Mitake #define X38_ERRSTS_UE 0x0002 64*df8bc08cSHitoshi Mitake #define X38_ERRSTS_CE 0x0001 65*df8bc08cSHitoshi Mitake #define X38_ERRSTS_BITS (X38_ERRSTS_UE | X38_ERRSTS_CE) 66*df8bc08cSHitoshi Mitake 67*df8bc08cSHitoshi Mitake 68*df8bc08cSHitoshi Mitake /* Intel MMIO register space - device 0 function 0 - MMR space */ 69*df8bc08cSHitoshi Mitake 70*df8bc08cSHitoshi Mitake #define X38_C0DRB 0x200 /* Channel 0 DRAM Rank Boundary (16b x 4) 71*df8bc08cSHitoshi Mitake * 72*df8bc08cSHitoshi Mitake * 15:10 reserved 73*df8bc08cSHitoshi Mitake * 9:0 Channel 0 DRAM Rank Boundary Address 74*df8bc08cSHitoshi Mitake */ 75*df8bc08cSHitoshi Mitake #define X38_C1DRB 0x600 /* Channel 1 DRAM Rank Boundary (16b x 4) */ 76*df8bc08cSHitoshi Mitake #define X38_DRB_MASK 0x3ff /* bits 9:0 */ 77*df8bc08cSHitoshi Mitake #define X38_DRB_SHIFT 26 /* 64MiB grain */ 78*df8bc08cSHitoshi Mitake 79*df8bc08cSHitoshi Mitake #define X38_C0ECCERRLOG 0x280 /* Channel 0 ECC Error Log (64b) 80*df8bc08cSHitoshi Mitake * 81*df8bc08cSHitoshi Mitake * 63:48 Error Column Address (ERRCOL) 82*df8bc08cSHitoshi Mitake * 47:32 Error Row Address (ERRROW) 83*df8bc08cSHitoshi Mitake * 31:29 Error Bank Address (ERRBANK) 84*df8bc08cSHitoshi Mitake * 28:27 Error Rank Address (ERRRANK) 85*df8bc08cSHitoshi Mitake * 26:24 reserved 86*df8bc08cSHitoshi Mitake * 23:16 Error Syndrome (ERRSYND) 87*df8bc08cSHitoshi Mitake * 15: 2 reserved 88*df8bc08cSHitoshi Mitake * 1 Multiple Bit Error Status (MERRSTS) 89*df8bc08cSHitoshi Mitake * 0 Correctable Error Status (CERRSTS) 90*df8bc08cSHitoshi Mitake */ 91*df8bc08cSHitoshi Mitake #define X38_C1ECCERRLOG 0x680 /* Channel 1 ECC Error Log (64b) */ 92*df8bc08cSHitoshi Mitake #define X38_ECCERRLOG_CE 0x1 93*df8bc08cSHitoshi Mitake #define X38_ECCERRLOG_UE 0x2 94*df8bc08cSHitoshi Mitake #define X38_ECCERRLOG_RANK_BITS 0x18000000 95*df8bc08cSHitoshi Mitake #define X38_ECCERRLOG_SYNDROME_BITS 0xff0000 96*df8bc08cSHitoshi Mitake 97*df8bc08cSHitoshi Mitake #define X38_CAPID0 0xe0 /* see P.94 of spec for details */ 98*df8bc08cSHitoshi Mitake 99*df8bc08cSHitoshi Mitake static int x38_channel_num; 100*df8bc08cSHitoshi Mitake 101*df8bc08cSHitoshi Mitake static int how_many_channel(struct pci_dev *pdev) 102*df8bc08cSHitoshi Mitake { 103*df8bc08cSHitoshi Mitake unsigned char capid0_8b; /* 8th byte of CAPID0 */ 104*df8bc08cSHitoshi Mitake 105*df8bc08cSHitoshi Mitake pci_read_config_byte(pdev, X38_CAPID0 + 8, &capid0_8b); 106*df8bc08cSHitoshi Mitake if (capid0_8b & 0x20) { /* check DCD: Dual Channel Disable */ 107*df8bc08cSHitoshi Mitake debugf0("In single channel mode.\n"); 108*df8bc08cSHitoshi Mitake x38_channel_num = 1; 109*df8bc08cSHitoshi Mitake } else { 110*df8bc08cSHitoshi Mitake debugf0("In dual channel mode.\n"); 111*df8bc08cSHitoshi Mitake x38_channel_num = 2; 112*df8bc08cSHitoshi Mitake } 113*df8bc08cSHitoshi Mitake 114*df8bc08cSHitoshi Mitake return x38_channel_num; 115*df8bc08cSHitoshi Mitake } 116*df8bc08cSHitoshi Mitake 117*df8bc08cSHitoshi Mitake static unsigned long eccerrlog_syndrome(u64 log) 118*df8bc08cSHitoshi Mitake { 119*df8bc08cSHitoshi Mitake return (log & X38_ECCERRLOG_SYNDROME_BITS) >> 16; 120*df8bc08cSHitoshi Mitake } 121*df8bc08cSHitoshi Mitake 122*df8bc08cSHitoshi Mitake static int eccerrlog_row(int channel, u64 log) 123*df8bc08cSHitoshi Mitake { 124*df8bc08cSHitoshi Mitake return ((log & X38_ECCERRLOG_RANK_BITS) >> 27) | 125*df8bc08cSHitoshi Mitake (channel * X38_RANKS_PER_CHANNEL); 126*df8bc08cSHitoshi Mitake } 127*df8bc08cSHitoshi Mitake 128*df8bc08cSHitoshi Mitake enum x38_chips { 129*df8bc08cSHitoshi Mitake X38 = 0, 130*df8bc08cSHitoshi Mitake }; 131*df8bc08cSHitoshi Mitake 132*df8bc08cSHitoshi Mitake struct x38_dev_info { 133*df8bc08cSHitoshi Mitake const char *ctl_name; 134*df8bc08cSHitoshi Mitake }; 135*df8bc08cSHitoshi Mitake 136*df8bc08cSHitoshi Mitake struct x38_error_info { 137*df8bc08cSHitoshi Mitake u16 errsts; 138*df8bc08cSHitoshi Mitake u16 errsts2; 139*df8bc08cSHitoshi Mitake u64 eccerrlog[X38_CHANNELS]; 140*df8bc08cSHitoshi Mitake }; 141*df8bc08cSHitoshi Mitake 142*df8bc08cSHitoshi Mitake static const struct x38_dev_info x38_devs[] = { 143*df8bc08cSHitoshi Mitake [X38] = { 144*df8bc08cSHitoshi Mitake .ctl_name = "x38"}, 145*df8bc08cSHitoshi Mitake }; 146*df8bc08cSHitoshi Mitake 147*df8bc08cSHitoshi Mitake static struct pci_dev *mci_pdev; 148*df8bc08cSHitoshi Mitake static int x38_registered = 1; 149*df8bc08cSHitoshi Mitake 150*df8bc08cSHitoshi Mitake 151*df8bc08cSHitoshi Mitake static void x38_clear_error_info(struct mem_ctl_info *mci) 152*df8bc08cSHitoshi Mitake { 153*df8bc08cSHitoshi Mitake struct pci_dev *pdev; 154*df8bc08cSHitoshi Mitake 155*df8bc08cSHitoshi Mitake pdev = to_pci_dev(mci->dev); 156*df8bc08cSHitoshi Mitake 157*df8bc08cSHitoshi Mitake /* 158*df8bc08cSHitoshi Mitake * Clear any error bits. 159*df8bc08cSHitoshi Mitake * (Yes, we really clear bits by writing 1 to them.) 160*df8bc08cSHitoshi Mitake */ 161*df8bc08cSHitoshi Mitake pci_write_bits16(pdev, X38_ERRSTS, X38_ERRSTS_BITS, 162*df8bc08cSHitoshi Mitake X38_ERRSTS_BITS); 163*df8bc08cSHitoshi Mitake } 164*df8bc08cSHitoshi Mitake 165*df8bc08cSHitoshi Mitake static u64 x38_readq(const void __iomem *addr) 166*df8bc08cSHitoshi Mitake { 167*df8bc08cSHitoshi Mitake return readl(addr) | (((u64)readl(addr + 4)) << 32); 168*df8bc08cSHitoshi Mitake } 169*df8bc08cSHitoshi Mitake 170*df8bc08cSHitoshi Mitake static void x38_get_and_clear_error_info(struct mem_ctl_info *mci, 171*df8bc08cSHitoshi Mitake struct x38_error_info *info) 172*df8bc08cSHitoshi Mitake { 173*df8bc08cSHitoshi Mitake struct pci_dev *pdev; 174*df8bc08cSHitoshi Mitake void __iomem *window = mci->pvt_info; 175*df8bc08cSHitoshi Mitake 176*df8bc08cSHitoshi Mitake pdev = to_pci_dev(mci->dev); 177*df8bc08cSHitoshi Mitake 178*df8bc08cSHitoshi Mitake /* 179*df8bc08cSHitoshi Mitake * This is a mess because there is no atomic way to read all the 180*df8bc08cSHitoshi Mitake * registers at once and the registers can transition from CE being 181*df8bc08cSHitoshi Mitake * overwritten by UE. 182*df8bc08cSHitoshi Mitake */ 183*df8bc08cSHitoshi Mitake pci_read_config_word(pdev, X38_ERRSTS, &info->errsts); 184*df8bc08cSHitoshi Mitake if (!(info->errsts & X38_ERRSTS_BITS)) 185*df8bc08cSHitoshi Mitake return; 186*df8bc08cSHitoshi Mitake 187*df8bc08cSHitoshi Mitake info->eccerrlog[0] = x38_readq(window + X38_C0ECCERRLOG); 188*df8bc08cSHitoshi Mitake if (x38_channel_num == 2) 189*df8bc08cSHitoshi Mitake info->eccerrlog[1] = x38_readq(window + X38_C1ECCERRLOG); 190*df8bc08cSHitoshi Mitake 191*df8bc08cSHitoshi Mitake pci_read_config_word(pdev, X38_ERRSTS, &info->errsts2); 192*df8bc08cSHitoshi Mitake 193*df8bc08cSHitoshi Mitake /* 194*df8bc08cSHitoshi Mitake * If the error is the same for both reads then the first set 195*df8bc08cSHitoshi Mitake * of reads is valid. If there is a change then there is a CE 196*df8bc08cSHitoshi Mitake * with no info and the second set of reads is valid and 197*df8bc08cSHitoshi Mitake * should be UE info. 198*df8bc08cSHitoshi Mitake */ 199*df8bc08cSHitoshi Mitake if ((info->errsts ^ info->errsts2) & X38_ERRSTS_BITS) { 200*df8bc08cSHitoshi Mitake info->eccerrlog[0] = x38_readq(window + X38_C0ECCERRLOG); 201*df8bc08cSHitoshi Mitake if (x38_channel_num == 2) 202*df8bc08cSHitoshi Mitake info->eccerrlog[1] = 203*df8bc08cSHitoshi Mitake x38_readq(window + X38_C1ECCERRLOG); 204*df8bc08cSHitoshi Mitake } 205*df8bc08cSHitoshi Mitake 206*df8bc08cSHitoshi Mitake x38_clear_error_info(mci); 207*df8bc08cSHitoshi Mitake } 208*df8bc08cSHitoshi Mitake 209*df8bc08cSHitoshi Mitake static void x38_process_error_info(struct mem_ctl_info *mci, 210*df8bc08cSHitoshi Mitake struct x38_error_info *info) 211*df8bc08cSHitoshi Mitake { 212*df8bc08cSHitoshi Mitake int channel; 213*df8bc08cSHitoshi Mitake u64 log; 214*df8bc08cSHitoshi Mitake 215*df8bc08cSHitoshi Mitake if (!(info->errsts & X38_ERRSTS_BITS)) 216*df8bc08cSHitoshi Mitake return; 217*df8bc08cSHitoshi Mitake 218*df8bc08cSHitoshi Mitake if ((info->errsts ^ info->errsts2) & X38_ERRSTS_BITS) { 219*df8bc08cSHitoshi Mitake edac_mc_handle_ce_no_info(mci, "UE overwrote CE"); 220*df8bc08cSHitoshi Mitake info->errsts = info->errsts2; 221*df8bc08cSHitoshi Mitake } 222*df8bc08cSHitoshi Mitake 223*df8bc08cSHitoshi Mitake for (channel = 0; channel < x38_channel_num; channel++) { 224*df8bc08cSHitoshi Mitake log = info->eccerrlog[channel]; 225*df8bc08cSHitoshi Mitake if (log & X38_ECCERRLOG_UE) { 226*df8bc08cSHitoshi Mitake edac_mc_handle_ue(mci, 0, 0, 227*df8bc08cSHitoshi Mitake eccerrlog_row(channel, log), "x38 UE"); 228*df8bc08cSHitoshi Mitake } else if (log & X38_ECCERRLOG_CE) { 229*df8bc08cSHitoshi Mitake edac_mc_handle_ce(mci, 0, 0, 230*df8bc08cSHitoshi Mitake eccerrlog_syndrome(log), 231*df8bc08cSHitoshi Mitake eccerrlog_row(channel, log), 0, "x38 CE"); 232*df8bc08cSHitoshi Mitake } 233*df8bc08cSHitoshi Mitake } 234*df8bc08cSHitoshi Mitake } 235*df8bc08cSHitoshi Mitake 236*df8bc08cSHitoshi Mitake static void x38_check(struct mem_ctl_info *mci) 237*df8bc08cSHitoshi Mitake { 238*df8bc08cSHitoshi Mitake struct x38_error_info info; 239*df8bc08cSHitoshi Mitake 240*df8bc08cSHitoshi Mitake debugf1("MC%d: %s()\n", mci->mc_idx, __func__); 241*df8bc08cSHitoshi Mitake x38_get_and_clear_error_info(mci, &info); 242*df8bc08cSHitoshi Mitake x38_process_error_info(mci, &info); 243*df8bc08cSHitoshi Mitake } 244*df8bc08cSHitoshi Mitake 245*df8bc08cSHitoshi Mitake 246*df8bc08cSHitoshi Mitake void __iomem *x38_map_mchbar(struct pci_dev *pdev) 247*df8bc08cSHitoshi Mitake { 248*df8bc08cSHitoshi Mitake union { 249*df8bc08cSHitoshi Mitake u64 mchbar; 250*df8bc08cSHitoshi Mitake struct { 251*df8bc08cSHitoshi Mitake u32 mchbar_low; 252*df8bc08cSHitoshi Mitake u32 mchbar_high; 253*df8bc08cSHitoshi Mitake }; 254*df8bc08cSHitoshi Mitake } u; 255*df8bc08cSHitoshi Mitake void __iomem *window; 256*df8bc08cSHitoshi Mitake 257*df8bc08cSHitoshi Mitake pci_read_config_dword(pdev, X38_MCHBAR_LOW, &u.mchbar_low); 258*df8bc08cSHitoshi Mitake pci_write_config_dword(pdev, X38_MCHBAR_LOW, u.mchbar_low | 0x1); 259*df8bc08cSHitoshi Mitake pci_read_config_dword(pdev, X38_MCHBAR_HIGH, &u.mchbar_high); 260*df8bc08cSHitoshi Mitake u.mchbar &= X38_MCHBAR_MASK; 261*df8bc08cSHitoshi Mitake 262*df8bc08cSHitoshi Mitake if (u.mchbar != (resource_size_t)u.mchbar) { 263*df8bc08cSHitoshi Mitake printk(KERN_ERR 264*df8bc08cSHitoshi Mitake "x38: mmio space beyond accessible range (0x%llx)\n", 265*df8bc08cSHitoshi Mitake (unsigned long long)u.mchbar); 266*df8bc08cSHitoshi Mitake return NULL; 267*df8bc08cSHitoshi Mitake } 268*df8bc08cSHitoshi Mitake 269*df8bc08cSHitoshi Mitake window = ioremap_nocache(u.mchbar, X38_MMR_WINDOW_SIZE); 270*df8bc08cSHitoshi Mitake if (!window) 271*df8bc08cSHitoshi Mitake printk(KERN_ERR "x38: cannot map mmio space at 0x%llx\n", 272*df8bc08cSHitoshi Mitake (unsigned long long)u.mchbar); 273*df8bc08cSHitoshi Mitake 274*df8bc08cSHitoshi Mitake return window; 275*df8bc08cSHitoshi Mitake } 276*df8bc08cSHitoshi Mitake 277*df8bc08cSHitoshi Mitake 278*df8bc08cSHitoshi Mitake static void x38_get_drbs(void __iomem *window, 279*df8bc08cSHitoshi Mitake u16 drbs[X38_CHANNELS][X38_RANKS_PER_CHANNEL]) 280*df8bc08cSHitoshi Mitake { 281*df8bc08cSHitoshi Mitake int i; 282*df8bc08cSHitoshi Mitake 283*df8bc08cSHitoshi Mitake for (i = 0; i < X38_RANKS_PER_CHANNEL; i++) { 284*df8bc08cSHitoshi Mitake drbs[0][i] = readw(window + X38_C0DRB + 2*i) & X38_DRB_MASK; 285*df8bc08cSHitoshi Mitake drbs[1][i] = readw(window + X38_C1DRB + 2*i) & X38_DRB_MASK; 286*df8bc08cSHitoshi Mitake } 287*df8bc08cSHitoshi Mitake } 288*df8bc08cSHitoshi Mitake 289*df8bc08cSHitoshi Mitake static bool x38_is_stacked(struct pci_dev *pdev, 290*df8bc08cSHitoshi Mitake u16 drbs[X38_CHANNELS][X38_RANKS_PER_CHANNEL]) 291*df8bc08cSHitoshi Mitake { 292*df8bc08cSHitoshi Mitake u16 tom; 293*df8bc08cSHitoshi Mitake 294*df8bc08cSHitoshi Mitake pci_read_config_word(pdev, X38_TOM, &tom); 295*df8bc08cSHitoshi Mitake tom &= X38_TOM_MASK; 296*df8bc08cSHitoshi Mitake 297*df8bc08cSHitoshi Mitake return drbs[X38_CHANNELS - 1][X38_RANKS_PER_CHANNEL - 1] == tom; 298*df8bc08cSHitoshi Mitake } 299*df8bc08cSHitoshi Mitake 300*df8bc08cSHitoshi Mitake static unsigned long drb_to_nr_pages( 301*df8bc08cSHitoshi Mitake u16 drbs[X38_CHANNELS][X38_RANKS_PER_CHANNEL], 302*df8bc08cSHitoshi Mitake bool stacked, int channel, int rank) 303*df8bc08cSHitoshi Mitake { 304*df8bc08cSHitoshi Mitake int n; 305*df8bc08cSHitoshi Mitake 306*df8bc08cSHitoshi Mitake n = drbs[channel][rank]; 307*df8bc08cSHitoshi Mitake if (rank > 0) 308*df8bc08cSHitoshi Mitake n -= drbs[channel][rank - 1]; 309*df8bc08cSHitoshi Mitake if (stacked && (channel == 1) && drbs[channel][rank] == 310*df8bc08cSHitoshi Mitake drbs[channel][X38_RANKS_PER_CHANNEL - 1]) { 311*df8bc08cSHitoshi Mitake n -= drbs[0][X38_RANKS_PER_CHANNEL - 1]; 312*df8bc08cSHitoshi Mitake } 313*df8bc08cSHitoshi Mitake 314*df8bc08cSHitoshi Mitake n <<= (X38_DRB_SHIFT - PAGE_SHIFT); 315*df8bc08cSHitoshi Mitake return n; 316*df8bc08cSHitoshi Mitake } 317*df8bc08cSHitoshi Mitake 318*df8bc08cSHitoshi Mitake static int x38_probe1(struct pci_dev *pdev, int dev_idx) 319*df8bc08cSHitoshi Mitake { 320*df8bc08cSHitoshi Mitake int rc; 321*df8bc08cSHitoshi Mitake int i; 322*df8bc08cSHitoshi Mitake struct mem_ctl_info *mci = NULL; 323*df8bc08cSHitoshi Mitake unsigned long last_page; 324*df8bc08cSHitoshi Mitake u16 drbs[X38_CHANNELS][X38_RANKS_PER_CHANNEL]; 325*df8bc08cSHitoshi Mitake bool stacked; 326*df8bc08cSHitoshi Mitake void __iomem *window; 327*df8bc08cSHitoshi Mitake 328*df8bc08cSHitoshi Mitake debugf0("MC: %s()\n", __func__); 329*df8bc08cSHitoshi Mitake 330*df8bc08cSHitoshi Mitake window = x38_map_mchbar(pdev); 331*df8bc08cSHitoshi Mitake if (!window) 332*df8bc08cSHitoshi Mitake return -ENODEV; 333*df8bc08cSHitoshi Mitake 334*df8bc08cSHitoshi Mitake x38_get_drbs(window, drbs); 335*df8bc08cSHitoshi Mitake 336*df8bc08cSHitoshi Mitake how_many_channel(pdev); 337*df8bc08cSHitoshi Mitake 338*df8bc08cSHitoshi Mitake /* FIXME: unconventional pvt_info usage */ 339*df8bc08cSHitoshi Mitake mci = edac_mc_alloc(0, X38_RANKS, x38_channel_num, 0); 340*df8bc08cSHitoshi Mitake if (!mci) 341*df8bc08cSHitoshi Mitake return -ENOMEM; 342*df8bc08cSHitoshi Mitake 343*df8bc08cSHitoshi Mitake debugf3("MC: %s(): init mci\n", __func__); 344*df8bc08cSHitoshi Mitake 345*df8bc08cSHitoshi Mitake mci->dev = &pdev->dev; 346*df8bc08cSHitoshi Mitake mci->mtype_cap = MEM_FLAG_DDR2; 347*df8bc08cSHitoshi Mitake 348*df8bc08cSHitoshi Mitake mci->edac_ctl_cap = EDAC_FLAG_SECDED; 349*df8bc08cSHitoshi Mitake mci->edac_cap = EDAC_FLAG_SECDED; 350*df8bc08cSHitoshi Mitake 351*df8bc08cSHitoshi Mitake mci->mod_name = EDAC_MOD_STR; 352*df8bc08cSHitoshi Mitake mci->mod_ver = X38_REVISION; 353*df8bc08cSHitoshi Mitake mci->ctl_name = x38_devs[dev_idx].ctl_name; 354*df8bc08cSHitoshi Mitake mci->dev_name = pci_name(pdev); 355*df8bc08cSHitoshi Mitake mci->edac_check = x38_check; 356*df8bc08cSHitoshi Mitake mci->ctl_page_to_phys = NULL; 357*df8bc08cSHitoshi Mitake mci->pvt_info = window; 358*df8bc08cSHitoshi Mitake 359*df8bc08cSHitoshi Mitake stacked = x38_is_stacked(pdev, drbs); 360*df8bc08cSHitoshi Mitake 361*df8bc08cSHitoshi Mitake /* 362*df8bc08cSHitoshi Mitake * The dram rank boundary (DRB) reg values are boundary addresses 363*df8bc08cSHitoshi Mitake * for each DRAM rank with a granularity of 64MB. DRB regs are 364*df8bc08cSHitoshi Mitake * cumulative; the last one will contain the total memory 365*df8bc08cSHitoshi Mitake * contained in all ranks. 366*df8bc08cSHitoshi Mitake */ 367*df8bc08cSHitoshi Mitake last_page = -1UL; 368*df8bc08cSHitoshi Mitake for (i = 0; i < mci->nr_csrows; i++) { 369*df8bc08cSHitoshi Mitake unsigned long nr_pages; 370*df8bc08cSHitoshi Mitake struct csrow_info *csrow = &mci->csrows[i]; 371*df8bc08cSHitoshi Mitake 372*df8bc08cSHitoshi Mitake nr_pages = drb_to_nr_pages(drbs, stacked, 373*df8bc08cSHitoshi Mitake i / X38_RANKS_PER_CHANNEL, 374*df8bc08cSHitoshi Mitake i % X38_RANKS_PER_CHANNEL); 375*df8bc08cSHitoshi Mitake 376*df8bc08cSHitoshi Mitake if (nr_pages == 0) { 377*df8bc08cSHitoshi Mitake csrow->mtype = MEM_EMPTY; 378*df8bc08cSHitoshi Mitake continue; 379*df8bc08cSHitoshi Mitake } 380*df8bc08cSHitoshi Mitake 381*df8bc08cSHitoshi Mitake csrow->first_page = last_page + 1; 382*df8bc08cSHitoshi Mitake last_page += nr_pages; 383*df8bc08cSHitoshi Mitake csrow->last_page = last_page; 384*df8bc08cSHitoshi Mitake csrow->nr_pages = nr_pages; 385*df8bc08cSHitoshi Mitake 386*df8bc08cSHitoshi Mitake csrow->grain = nr_pages << PAGE_SHIFT; 387*df8bc08cSHitoshi Mitake csrow->mtype = MEM_DDR2; 388*df8bc08cSHitoshi Mitake csrow->dtype = DEV_UNKNOWN; 389*df8bc08cSHitoshi Mitake csrow->edac_mode = EDAC_UNKNOWN; 390*df8bc08cSHitoshi Mitake } 391*df8bc08cSHitoshi Mitake 392*df8bc08cSHitoshi Mitake x38_clear_error_info(mci); 393*df8bc08cSHitoshi Mitake 394*df8bc08cSHitoshi Mitake rc = -ENODEV; 395*df8bc08cSHitoshi Mitake if (edac_mc_add_mc(mci)) { 396*df8bc08cSHitoshi Mitake debugf3("MC: %s(): failed edac_mc_add_mc()\n", __func__); 397*df8bc08cSHitoshi Mitake goto fail; 398*df8bc08cSHitoshi Mitake } 399*df8bc08cSHitoshi Mitake 400*df8bc08cSHitoshi Mitake /* get this far and it's successful */ 401*df8bc08cSHitoshi Mitake debugf3("MC: %s(): success\n", __func__); 402*df8bc08cSHitoshi Mitake return 0; 403*df8bc08cSHitoshi Mitake 404*df8bc08cSHitoshi Mitake fail: 405*df8bc08cSHitoshi Mitake iounmap(window); 406*df8bc08cSHitoshi Mitake if (mci) 407*df8bc08cSHitoshi Mitake edac_mc_free(mci); 408*df8bc08cSHitoshi Mitake 409*df8bc08cSHitoshi Mitake return rc; 410*df8bc08cSHitoshi Mitake } 411*df8bc08cSHitoshi Mitake 412*df8bc08cSHitoshi Mitake static int __devinit x38_init_one(struct pci_dev *pdev, 413*df8bc08cSHitoshi Mitake const struct pci_device_id *ent) 414*df8bc08cSHitoshi Mitake { 415*df8bc08cSHitoshi Mitake int rc; 416*df8bc08cSHitoshi Mitake 417*df8bc08cSHitoshi Mitake debugf0("MC: %s()\n", __func__); 418*df8bc08cSHitoshi Mitake 419*df8bc08cSHitoshi Mitake if (pci_enable_device(pdev) < 0) 420*df8bc08cSHitoshi Mitake return -EIO; 421*df8bc08cSHitoshi Mitake 422*df8bc08cSHitoshi Mitake rc = x38_probe1(pdev, ent->driver_data); 423*df8bc08cSHitoshi Mitake if (!mci_pdev) 424*df8bc08cSHitoshi Mitake mci_pdev = pci_dev_get(pdev); 425*df8bc08cSHitoshi Mitake 426*df8bc08cSHitoshi Mitake return rc; 427*df8bc08cSHitoshi Mitake } 428*df8bc08cSHitoshi Mitake 429*df8bc08cSHitoshi Mitake static void __devexit x38_remove_one(struct pci_dev *pdev) 430*df8bc08cSHitoshi Mitake { 431*df8bc08cSHitoshi Mitake struct mem_ctl_info *mci; 432*df8bc08cSHitoshi Mitake 433*df8bc08cSHitoshi Mitake debugf0("%s()\n", __func__); 434*df8bc08cSHitoshi Mitake 435*df8bc08cSHitoshi Mitake mci = edac_mc_del_mc(&pdev->dev); 436*df8bc08cSHitoshi Mitake if (!mci) 437*df8bc08cSHitoshi Mitake return; 438*df8bc08cSHitoshi Mitake 439*df8bc08cSHitoshi Mitake iounmap(mci->pvt_info); 440*df8bc08cSHitoshi Mitake 441*df8bc08cSHitoshi Mitake edac_mc_free(mci); 442*df8bc08cSHitoshi Mitake } 443*df8bc08cSHitoshi Mitake 444*df8bc08cSHitoshi Mitake static const struct pci_device_id x38_pci_tbl[] __devinitdata = { 445*df8bc08cSHitoshi Mitake { 446*df8bc08cSHitoshi Mitake PCI_VEND_DEV(INTEL, X38_HB), PCI_ANY_ID, PCI_ANY_ID, 0, 0, 447*df8bc08cSHitoshi Mitake X38}, 448*df8bc08cSHitoshi Mitake { 449*df8bc08cSHitoshi Mitake 0, 450*df8bc08cSHitoshi Mitake } /* 0 terminated list. */ 451*df8bc08cSHitoshi Mitake }; 452*df8bc08cSHitoshi Mitake 453*df8bc08cSHitoshi Mitake MODULE_DEVICE_TABLE(pci, x38_pci_tbl); 454*df8bc08cSHitoshi Mitake 455*df8bc08cSHitoshi Mitake static struct pci_driver x38_driver = { 456*df8bc08cSHitoshi Mitake .name = EDAC_MOD_STR, 457*df8bc08cSHitoshi Mitake .probe = x38_init_one, 458*df8bc08cSHitoshi Mitake .remove = __devexit_p(x38_remove_one), 459*df8bc08cSHitoshi Mitake .id_table = x38_pci_tbl, 460*df8bc08cSHitoshi Mitake }; 461*df8bc08cSHitoshi Mitake 462*df8bc08cSHitoshi Mitake static int __init x38_init(void) 463*df8bc08cSHitoshi Mitake { 464*df8bc08cSHitoshi Mitake int pci_rc; 465*df8bc08cSHitoshi Mitake 466*df8bc08cSHitoshi Mitake debugf3("MC: %s()\n", __func__); 467*df8bc08cSHitoshi Mitake 468*df8bc08cSHitoshi Mitake /* Ensure that the OPSTATE is set correctly for POLL or NMI */ 469*df8bc08cSHitoshi Mitake opstate_init(); 470*df8bc08cSHitoshi Mitake 471*df8bc08cSHitoshi Mitake pci_rc = pci_register_driver(&x38_driver); 472*df8bc08cSHitoshi Mitake if (pci_rc < 0) 473*df8bc08cSHitoshi Mitake goto fail0; 474*df8bc08cSHitoshi Mitake 475*df8bc08cSHitoshi Mitake if (!mci_pdev) { 476*df8bc08cSHitoshi Mitake x38_registered = 0; 477*df8bc08cSHitoshi Mitake mci_pdev = pci_get_device(PCI_VENDOR_ID_INTEL, 478*df8bc08cSHitoshi Mitake PCI_DEVICE_ID_INTEL_X38_HB, NULL); 479*df8bc08cSHitoshi Mitake if (!mci_pdev) { 480*df8bc08cSHitoshi Mitake debugf0("x38 pci_get_device fail\n"); 481*df8bc08cSHitoshi Mitake pci_rc = -ENODEV; 482*df8bc08cSHitoshi Mitake goto fail1; 483*df8bc08cSHitoshi Mitake } 484*df8bc08cSHitoshi Mitake 485*df8bc08cSHitoshi Mitake pci_rc = x38_init_one(mci_pdev, x38_pci_tbl); 486*df8bc08cSHitoshi Mitake if (pci_rc < 0) { 487*df8bc08cSHitoshi Mitake debugf0("x38 init fail\n"); 488*df8bc08cSHitoshi Mitake pci_rc = -ENODEV; 489*df8bc08cSHitoshi Mitake goto fail1; 490*df8bc08cSHitoshi Mitake } 491*df8bc08cSHitoshi Mitake } 492*df8bc08cSHitoshi Mitake 493*df8bc08cSHitoshi Mitake return 0; 494*df8bc08cSHitoshi Mitake 495*df8bc08cSHitoshi Mitake fail1: 496*df8bc08cSHitoshi Mitake pci_unregister_driver(&x38_driver); 497*df8bc08cSHitoshi Mitake 498*df8bc08cSHitoshi Mitake fail0: 499*df8bc08cSHitoshi Mitake if (mci_pdev) 500*df8bc08cSHitoshi Mitake pci_dev_put(mci_pdev); 501*df8bc08cSHitoshi Mitake 502*df8bc08cSHitoshi Mitake return pci_rc; 503*df8bc08cSHitoshi Mitake } 504*df8bc08cSHitoshi Mitake 505*df8bc08cSHitoshi Mitake static void __exit x38_exit(void) 506*df8bc08cSHitoshi Mitake { 507*df8bc08cSHitoshi Mitake debugf3("MC: %s()\n", __func__); 508*df8bc08cSHitoshi Mitake 509*df8bc08cSHitoshi Mitake pci_unregister_driver(&x38_driver); 510*df8bc08cSHitoshi Mitake if (!x38_registered) { 511*df8bc08cSHitoshi Mitake x38_remove_one(mci_pdev); 512*df8bc08cSHitoshi Mitake pci_dev_put(mci_pdev); 513*df8bc08cSHitoshi Mitake } 514*df8bc08cSHitoshi Mitake } 515*df8bc08cSHitoshi Mitake 516*df8bc08cSHitoshi Mitake module_init(x38_init); 517*df8bc08cSHitoshi Mitake module_exit(x38_exit); 518*df8bc08cSHitoshi Mitake 519*df8bc08cSHitoshi Mitake MODULE_LICENSE("GPL"); 520*df8bc08cSHitoshi Mitake MODULE_AUTHOR("Cluster Computing, Inc. Hitoshi Mitake"); 521*df8bc08cSHitoshi Mitake MODULE_DESCRIPTION("MC support for Intel X38 memory hub controllers"); 522*df8bc08cSHitoshi Mitake 523*df8bc08cSHitoshi Mitake module_param(edac_op_state, int, 0444); 524*df8bc08cSHitoshi Mitake MODULE_PARM_DESC(edac_op_state, "EDAC Error Reporting state: 0=Poll,1=NMI"); 525