xref: /openbmc/linux/drivers/edac/e752x_edac.c (revision 8730046c)
1 /*
2  * Intel e752x Memory Controller kernel module
3  * (C) 2004 Linux Networx (http://lnxi.com)
4  * This file may be distributed under the terms of the
5  * GNU General Public License.
6  *
7  * Implement support for the e7520, E7525, e7320 and i3100 memory controllers.
8  *
9  * Datasheets:
10  *	http://www.intel.in/content/www/in/en/chipsets/e7525-memory-controller-hub-datasheet.html
11  *	ftp://download.intel.com/design/intarch/datashts/31345803.pdf
12  *
13  * Written by Tom Zimmerman
14  *
15  * Contributors:
16  * 	Thayne Harbaugh at realmsys.com (?)
17  * 	Wang Zhenyu at intel.com
18  * 	Dave Jiang at mvista.com
19  *
20  */
21 
22 #include <linux/module.h>
23 #include <linux/init.h>
24 #include <linux/pci.h>
25 #include <linux/pci_ids.h>
26 #include <linux/edac.h>
27 #include "edac_module.h"
28 
29 #define E752X_REVISION	" Ver: 2.0.2"
30 #define EDAC_MOD_STR	"e752x_edac"
31 
32 static int report_non_memory_errors;
33 static int force_function_unhide;
34 static int sysbus_parity = -1;
35 
36 static struct edac_pci_ctl_info *e752x_pci;
37 
38 #define e752x_printk(level, fmt, arg...) \
39 	edac_printk(level, "e752x", fmt, ##arg)
40 
41 #define e752x_mc_printk(mci, level, fmt, arg...) \
42 	edac_mc_chipset_printk(mci, level, "e752x", fmt, ##arg)
43 
44 #ifndef PCI_DEVICE_ID_INTEL_7520_0
45 #define PCI_DEVICE_ID_INTEL_7520_0      0x3590
46 #endif				/* PCI_DEVICE_ID_INTEL_7520_0      */
47 
48 #ifndef PCI_DEVICE_ID_INTEL_7520_1_ERR
49 #define PCI_DEVICE_ID_INTEL_7520_1_ERR  0x3591
50 #endif				/* PCI_DEVICE_ID_INTEL_7520_1_ERR  */
51 
52 #ifndef PCI_DEVICE_ID_INTEL_7525_0
53 #define PCI_DEVICE_ID_INTEL_7525_0      0x359E
54 #endif				/* PCI_DEVICE_ID_INTEL_7525_0      */
55 
56 #ifndef PCI_DEVICE_ID_INTEL_7525_1_ERR
57 #define PCI_DEVICE_ID_INTEL_7525_1_ERR  0x3593
58 #endif				/* PCI_DEVICE_ID_INTEL_7525_1_ERR  */
59 
60 #ifndef PCI_DEVICE_ID_INTEL_7320_0
61 #define PCI_DEVICE_ID_INTEL_7320_0	0x3592
62 #endif				/* PCI_DEVICE_ID_INTEL_7320_0 */
63 
64 #ifndef PCI_DEVICE_ID_INTEL_7320_1_ERR
65 #define PCI_DEVICE_ID_INTEL_7320_1_ERR	0x3593
66 #endif				/* PCI_DEVICE_ID_INTEL_7320_1_ERR */
67 
68 #ifndef PCI_DEVICE_ID_INTEL_3100_0
69 #define PCI_DEVICE_ID_INTEL_3100_0	0x35B0
70 #endif				/* PCI_DEVICE_ID_INTEL_3100_0 */
71 
72 #ifndef PCI_DEVICE_ID_INTEL_3100_1_ERR
73 #define PCI_DEVICE_ID_INTEL_3100_1_ERR	0x35B1
74 #endif				/* PCI_DEVICE_ID_INTEL_3100_1_ERR */
75 
76 #define E752X_NR_CSROWS		8	/* number of csrows */
77 
78 /* E752X register addresses - device 0 function 0 */
79 #define E752X_MCHSCRB		0x52	/* Memory Scrub register (16b) */
80 					/*
81 					 * 6:5     Scrub Completion Count
82 					 * 3:2     Scrub Rate (i3100 only)
83 					 *      01=fast 10=normal
84 					 * 1:0     Scrub Mode enable
85 					 *      00=off 10=on
86 					 */
87 #define E752X_DRB		0x60	/* DRAM row boundary register (8b) */
88 #define E752X_DRA		0x70	/* DRAM row attribute register (8b) */
89 					/*
90 					 * 31:30   Device width row 7
91 					 *      01=x8 10=x4 11=x8 DDR2
92 					 * 27:26   Device width row 6
93 					 * 23:22   Device width row 5
94 					 * 19:20   Device width row 4
95 					 * 15:14   Device width row 3
96 					 * 11:10   Device width row 2
97 					 *  7:6    Device width row 1
98 					 *  3:2    Device width row 0
99 					 */
100 #define E752X_DRC		0x7C	/* DRAM controller mode reg (32b) */
101 					/* FIXME:IS THIS RIGHT? */
102 					/*
103 					 * 22    Number channels 0=1,1=2
104 					 * 19:18 DRB Granularity 32/64MB
105 					 */
106 #define E752X_DRM		0x80	/* Dimm mapping register */
107 #define E752X_DDRCSR		0x9A	/* DDR control and status reg (16b) */
108 					/*
109 					 * 14:12 1 single A, 2 single B, 3 dual
110 					 */
111 #define E752X_TOLM		0xC4	/* DRAM top of low memory reg (16b) */
112 #define E752X_REMAPBASE		0xC6	/* DRAM remap base address reg (16b) */
113 #define E752X_REMAPLIMIT	0xC8	/* DRAM remap limit address reg (16b) */
114 #define E752X_REMAPOFFSET	0xCA	/* DRAM remap limit offset reg (16b) */
115 
116 /* E752X register addresses - device 0 function 1 */
117 #define E752X_FERR_GLOBAL	0x40	/* Global first error register (32b) */
118 #define E752X_NERR_GLOBAL	0x44	/* Global next error register (32b) */
119 #define E752X_HI_FERR		0x50	/* Hub interface first error reg (8b) */
120 #define E752X_HI_NERR		0x52	/* Hub interface next error reg (8b) */
121 #define E752X_HI_ERRMASK	0x54	/* Hub interface error mask reg (8b) */
122 #define E752X_HI_SMICMD		0x5A	/* Hub interface SMI command reg (8b) */
123 #define E752X_SYSBUS_FERR	0x60	/* System buss first error reg (16b) */
124 #define E752X_SYSBUS_NERR	0x62	/* System buss next error reg (16b) */
125 #define E752X_SYSBUS_ERRMASK	0x64	/* System buss error mask reg (16b) */
126 #define E752X_SYSBUS_SMICMD	0x6A	/* System buss SMI command reg (16b) */
127 #define E752X_BUF_FERR		0x70	/* Memory buffer first error reg (8b) */
128 #define E752X_BUF_NERR		0x72	/* Memory buffer next error reg (8b) */
129 #define E752X_BUF_ERRMASK	0x74	/* Memory buffer error mask reg (8b) */
130 #define E752X_BUF_SMICMD	0x7A	/* Memory buffer SMI cmd reg (8b) */
131 #define E752X_DRAM_FERR		0x80	/* DRAM first error register (16b) */
132 #define E752X_DRAM_NERR		0x82	/* DRAM next error register (16b) */
133 #define E752X_DRAM_ERRMASK	0x84	/* DRAM error mask register (8b) */
134 #define E752X_DRAM_SMICMD	0x8A	/* DRAM SMI command register (8b) */
135 #define E752X_DRAM_RETR_ADD	0xAC	/* DRAM Retry address register (32b) */
136 #define E752X_DRAM_SEC1_ADD	0xA0	/* DRAM first correctable memory */
137 					/*     error address register (32b) */
138 					/*
139 					 * 31    Reserved
140 					 * 30:2  CE address (64 byte block 34:6
141 					 * 1     Reserved
142 					 * 0     HiLoCS
143 					 */
144 #define E752X_DRAM_SEC2_ADD	0xC8	/* DRAM first correctable memory */
145 					/*     error address register (32b) */
146 					/*
147 					 * 31    Reserved
148 					 * 30:2  CE address (64 byte block 34:6)
149 					 * 1     Reserved
150 					 * 0     HiLoCS
151 					 */
152 #define E752X_DRAM_DED_ADD	0xA4	/* DRAM first uncorrectable memory */
153 					/*     error address register (32b) */
154 					/*
155 					 * 31    Reserved
156 					 * 30:2  CE address (64 byte block 34:6)
157 					 * 1     Reserved
158 					 * 0     HiLoCS
159 					 */
160 #define E752X_DRAM_SCRB_ADD	0xA8	/* DRAM 1st uncorrectable scrub mem */
161 					/*     error address register (32b) */
162 					/*
163 					 * 31    Reserved
164 					 * 30:2  CE address (64 byte block 34:6
165 					 * 1     Reserved
166 					 * 0     HiLoCS
167 					 */
168 #define E752X_DRAM_SEC1_SYNDROME 0xC4	/* DRAM first correctable memory */
169 					/*     error syndrome register (16b) */
170 #define E752X_DRAM_SEC2_SYNDROME 0xC6	/* DRAM second correctable memory */
171 					/*     error syndrome register (16b) */
172 #define E752X_DEVPRES1		0xF4	/* Device Present 1 register (8b) */
173 
174 /* 3100 IMCH specific register addresses - device 0 function 1 */
175 #define I3100_NSI_FERR		0x48	/* NSI first error reg (32b) */
176 #define I3100_NSI_NERR		0x4C	/* NSI next error reg (32b) */
177 #define I3100_NSI_SMICMD	0x54	/* NSI SMI command register (32b) */
178 #define I3100_NSI_EMASK		0x90	/* NSI error mask register (32b) */
179 
180 /* ICH5R register addresses - device 30 function 0 */
181 #define ICH5R_PCI_STAT		0x06	/* PCI status register (16b) */
182 #define ICH5R_PCI_2ND_STAT	0x1E	/* PCI status secondary reg (16b) */
183 #define ICH5R_PCI_BRIDGE_CTL	0x3E	/* PCI bridge control register (16b) */
184 
185 enum e752x_chips {
186 	E7520 = 0,
187 	E7525 = 1,
188 	E7320 = 2,
189 	I3100 = 3
190 };
191 
192 /*
193  * Those chips Support single-rank and dual-rank memories only.
194  *
195  * On e752x chips, the odd rows are present only on dual-rank memories.
196  * Dividing the rank by two will provide the dimm#
197  *
198  * i3100 MC has a different mapping: it supports only 4 ranks.
199  *
200  * The mapping is (from 1 to n):
201  *	slot	   single-ranked	double-ranked
202  *	dimm #1 -> rank #4		NA
203  *	dimm #2 -> rank #3		NA
204  *	dimm #3 -> rank #2		Ranks 2 and 3
205  *	dimm #4 -> rank $1		Ranks 1 and 4
206  *
207  * FIXME: The current mapping for i3100 considers that it supports up to 8
208  *	  ranks/chanel, but datasheet says that the MC supports only 4 ranks.
209  */
210 
211 struct e752x_pvt {
212 	struct pci_dev *dev_d0f0;
213 	struct pci_dev *dev_d0f1;
214 	u32 tolm;
215 	u32 remapbase;
216 	u32 remaplimit;
217 	int mc_symmetric;
218 	u8 map[8];
219 	int map_type;
220 	const struct e752x_dev_info *dev_info;
221 };
222 
223 struct e752x_dev_info {
224 	u16 err_dev;
225 	u16 ctl_dev;
226 	const char *ctl_name;
227 };
228 
229 struct e752x_error_info {
230 	u32 ferr_global;
231 	u32 nerr_global;
232 	u32 nsi_ferr;	/* 3100 only */
233 	u32 nsi_nerr;	/* 3100 only */
234 	u8 hi_ferr;	/* all but 3100 */
235 	u8 hi_nerr;	/* all but 3100 */
236 	u16 sysbus_ferr;
237 	u16 sysbus_nerr;
238 	u8 buf_ferr;
239 	u8 buf_nerr;
240 	u16 dram_ferr;
241 	u16 dram_nerr;
242 	u32 dram_sec1_add;
243 	u32 dram_sec2_add;
244 	u16 dram_sec1_syndrome;
245 	u16 dram_sec2_syndrome;
246 	u32 dram_ded_add;
247 	u32 dram_scrb_add;
248 	u32 dram_retr_add;
249 };
250 
251 static const struct e752x_dev_info e752x_devs[] = {
252 	[E7520] = {
253 		.err_dev = PCI_DEVICE_ID_INTEL_7520_1_ERR,
254 		.ctl_dev = PCI_DEVICE_ID_INTEL_7520_0,
255 		.ctl_name = "E7520"},
256 	[E7525] = {
257 		.err_dev = PCI_DEVICE_ID_INTEL_7525_1_ERR,
258 		.ctl_dev = PCI_DEVICE_ID_INTEL_7525_0,
259 		.ctl_name = "E7525"},
260 	[E7320] = {
261 		.err_dev = PCI_DEVICE_ID_INTEL_7320_1_ERR,
262 		.ctl_dev = PCI_DEVICE_ID_INTEL_7320_0,
263 		.ctl_name = "E7320"},
264 	[I3100] = {
265 		.err_dev = PCI_DEVICE_ID_INTEL_3100_1_ERR,
266 		.ctl_dev = PCI_DEVICE_ID_INTEL_3100_0,
267 		.ctl_name = "3100"},
268 };
269 
270 /* Valid scrub rates for the e752x/3100 hardware memory scrubber. We
271  * map the scrubbing bandwidth to a hardware register value. The 'set'
272  * operation finds the 'matching or higher value'.  Note that scrubbing
273  * on the e752x can only be enabled/disabled.  The 3100 supports
274  * a normal and fast mode.
275  */
276 
277 #define SDRATE_EOT 0xFFFFFFFF
278 
279 struct scrubrate {
280 	u32 bandwidth;	/* bandwidth consumed by scrubbing in bytes/sec */
281 	u16 scrubval;	/* register value for scrub rate */
282 };
283 
284 /* Rate below assumes same performance as i3100 using PC3200 DDR2 in
285  * normal mode.  e752x bridges don't support choosing normal or fast mode,
286  * so the scrubbing bandwidth value isn't all that important - scrubbing is
287  * either on or off.
288  */
289 static const struct scrubrate scrubrates_e752x[] = {
290 	{0,		0x00},	/* Scrubbing Off */
291 	{500000,	0x02},	/* Scrubbing On */
292 	{SDRATE_EOT,	0x00}	/* End of Table */
293 };
294 
295 /* Fast mode: 2 GByte PC3200 DDR2 scrubbed in 33s = 63161283 bytes/s
296  * Normal mode: 125 (32000 / 256) times slower than fast mode.
297  */
298 static const struct scrubrate scrubrates_i3100[] = {
299 	{0,		0x00},	/* Scrubbing Off */
300 	{500000,	0x0a},	/* Normal mode - 32k clocks */
301 	{62500000,	0x06},	/* Fast mode - 256 clocks */
302 	{SDRATE_EOT,	0x00}	/* End of Table */
303 };
304 
305 static unsigned long ctl_page_to_phys(struct mem_ctl_info *mci,
306 				unsigned long page)
307 {
308 	u32 remap;
309 	struct e752x_pvt *pvt = (struct e752x_pvt *)mci->pvt_info;
310 
311 	edac_dbg(3, "\n");
312 
313 	if (page < pvt->tolm)
314 		return page;
315 
316 	if ((page >= 0x100000) && (page < pvt->remapbase))
317 		return page;
318 
319 	remap = (page - pvt->tolm) + pvt->remapbase;
320 
321 	if (remap < pvt->remaplimit)
322 		return remap;
323 
324 	e752x_printk(KERN_ERR, "Invalid page %lx - out of range\n", page);
325 	return pvt->tolm - 1;
326 }
327 
328 static void do_process_ce(struct mem_ctl_info *mci, u16 error_one,
329 			u32 sec1_add, u16 sec1_syndrome)
330 {
331 	u32 page;
332 	int row;
333 	int channel;
334 	int i;
335 	struct e752x_pvt *pvt = (struct e752x_pvt *)mci->pvt_info;
336 
337 	edac_dbg(3, "\n");
338 
339 	/* convert the addr to 4k page */
340 	page = sec1_add >> (PAGE_SHIFT - 4);
341 
342 	/* FIXME - check for -1 */
343 	if (pvt->mc_symmetric) {
344 		/* chip select are bits 14 & 13 */
345 		row = ((page >> 1) & 3);
346 		e752x_printk(KERN_WARNING,
347 			"Test row %d Table %d %d %d %d %d %d %d %d\n", row,
348 			pvt->map[0], pvt->map[1], pvt->map[2], pvt->map[3],
349 			pvt->map[4], pvt->map[5], pvt->map[6],
350 			pvt->map[7]);
351 
352 		/* test for channel remapping */
353 		for (i = 0; i < 8; i++) {
354 			if (pvt->map[i] == row)
355 				break;
356 		}
357 
358 		e752x_printk(KERN_WARNING, "Test computed row %d\n", i);
359 
360 		if (i < 8)
361 			row = i;
362 		else
363 			e752x_mc_printk(mci, KERN_WARNING,
364 					"row %d not found in remap table\n",
365 					row);
366 	} else
367 		row = edac_mc_find_csrow_by_page(mci, page);
368 
369 	/* 0 = channel A, 1 = channel B */
370 	channel = !(error_one & 1);
371 
372 	/* e752x mc reads 34:6 of the DRAM linear address */
373 	edac_mc_handle_error(HW_EVENT_ERR_CORRECTED, mci, 1,
374 			     page, offset_in_page(sec1_add << 4), sec1_syndrome,
375 			     row, channel, -1,
376 			     "e752x CE", "");
377 }
378 
379 static inline void process_ce(struct mem_ctl_info *mci, u16 error_one,
380 			u32 sec1_add, u16 sec1_syndrome, int *error_found,
381 			int handle_error)
382 {
383 	*error_found = 1;
384 
385 	if (handle_error)
386 		do_process_ce(mci, error_one, sec1_add, sec1_syndrome);
387 }
388 
389 static void do_process_ue(struct mem_ctl_info *mci, u16 error_one,
390 			u32 ded_add, u32 scrb_add)
391 {
392 	u32 error_2b, block_page;
393 	int row;
394 	struct e752x_pvt *pvt = (struct e752x_pvt *)mci->pvt_info;
395 
396 	edac_dbg(3, "\n");
397 
398 	if (error_one & 0x0202) {
399 		error_2b = ded_add;
400 
401 		/* convert to 4k address */
402 		block_page = error_2b >> (PAGE_SHIFT - 4);
403 
404 		row = pvt->mc_symmetric ?
405 		/* chip select are bits 14 & 13 */
406 			((block_page >> 1) & 3) :
407 			edac_mc_find_csrow_by_page(mci, block_page);
408 
409 		/* e752x mc reads 34:6 of the DRAM linear address */
410 		edac_mc_handle_error(HW_EVENT_ERR_UNCORRECTED, mci, 1,
411 					block_page,
412 					offset_in_page(error_2b << 4), 0,
413 					 row, -1, -1,
414 					"e752x UE from Read", "");
415 
416 	}
417 	if (error_one & 0x0404) {
418 		error_2b = scrb_add;
419 
420 		/* convert to 4k address */
421 		block_page = error_2b >> (PAGE_SHIFT - 4);
422 
423 		row = pvt->mc_symmetric ?
424 		/* chip select are bits 14 & 13 */
425 			((block_page >> 1) & 3) :
426 			edac_mc_find_csrow_by_page(mci, block_page);
427 
428 		/* e752x mc reads 34:6 of the DRAM linear address */
429 		edac_mc_handle_error(HW_EVENT_ERR_UNCORRECTED, mci, 1,
430 					block_page,
431 					offset_in_page(error_2b << 4), 0,
432 					row, -1, -1,
433 					"e752x UE from Scruber", "");
434 	}
435 }
436 
437 static inline void process_ue(struct mem_ctl_info *mci, u16 error_one,
438 			u32 ded_add, u32 scrb_add, int *error_found,
439 			int handle_error)
440 {
441 	*error_found = 1;
442 
443 	if (handle_error)
444 		do_process_ue(mci, error_one, ded_add, scrb_add);
445 }
446 
447 static inline void process_ue_no_info_wr(struct mem_ctl_info *mci,
448 					 int *error_found, int handle_error)
449 {
450 	*error_found = 1;
451 
452 	if (!handle_error)
453 		return;
454 
455 	edac_dbg(3, "\n");
456 	edac_mc_handle_error(HW_EVENT_ERR_UNCORRECTED, mci, 1, 0, 0, 0,
457 			     -1, -1, -1,
458 			     "e752x UE log memory write", "");
459 }
460 
461 static void do_process_ded_retry(struct mem_ctl_info *mci, u16 error,
462 				 u32 retry_add)
463 {
464 	u32 error_1b, page;
465 	int row;
466 	struct e752x_pvt *pvt = (struct e752x_pvt *)mci->pvt_info;
467 
468 	error_1b = retry_add;
469 	page = error_1b >> (PAGE_SHIFT - 4);  /* convert the addr to 4k page */
470 
471 	/* chip select are bits 14 & 13 */
472 	row = pvt->mc_symmetric ? ((page >> 1) & 3) :
473 		edac_mc_find_csrow_by_page(mci, page);
474 
475 	e752x_mc_printk(mci, KERN_WARNING,
476 			"CE page 0x%lx, row %d : Memory read retry\n",
477 			(long unsigned int)page, row);
478 }
479 
480 static inline void process_ded_retry(struct mem_ctl_info *mci, u16 error,
481 				u32 retry_add, int *error_found,
482 				int handle_error)
483 {
484 	*error_found = 1;
485 
486 	if (handle_error)
487 		do_process_ded_retry(mci, error, retry_add);
488 }
489 
490 static inline void process_threshold_ce(struct mem_ctl_info *mci, u16 error,
491 					int *error_found, int handle_error)
492 {
493 	*error_found = 1;
494 
495 	if (handle_error)
496 		e752x_mc_printk(mci, KERN_WARNING, "Memory threshold CE\n");
497 }
498 
499 static char *global_message[11] = {
500 	"PCI Express C1",
501 	"PCI Express C",
502 	"PCI Express B1",
503 	"PCI Express B",
504 	"PCI Express A1",
505 	"PCI Express A",
506 	"DMA Controller",
507 	"HUB or NS Interface",
508 	"System Bus",
509 	"DRAM Controller",  /* 9th entry */
510 	"Internal Buffer"
511 };
512 
513 #define DRAM_ENTRY	9
514 
515 static char *fatal_message[2] = { "Non-Fatal ", "Fatal " };
516 
517 static void do_global_error(int fatal, u32 errors)
518 {
519 	int i;
520 
521 	for (i = 0; i < 11; i++) {
522 		if (errors & (1 << i)) {
523 			/* If the error is from DRAM Controller OR
524 			 * we are to report ALL errors, then
525 			 * report the error
526 			 */
527 			if ((i == DRAM_ENTRY) || report_non_memory_errors)
528 				e752x_printk(KERN_WARNING, "%sError %s\n",
529 					fatal_message[fatal],
530 					global_message[i]);
531 		}
532 	}
533 }
534 
535 static inline void global_error(int fatal, u32 errors, int *error_found,
536 				int handle_error)
537 {
538 	*error_found = 1;
539 
540 	if (handle_error)
541 		do_global_error(fatal, errors);
542 }
543 
544 static char *hub_message[7] = {
545 	"HI Address or Command Parity", "HI Illegal Access",
546 	"HI Internal Parity", "Out of Range Access",
547 	"HI Data Parity", "Enhanced Config Access",
548 	"Hub Interface Target Abort"
549 };
550 
551 static void do_hub_error(int fatal, u8 errors)
552 {
553 	int i;
554 
555 	for (i = 0; i < 7; i++) {
556 		if (errors & (1 << i))
557 			e752x_printk(KERN_WARNING, "%sError %s\n",
558 				fatal_message[fatal], hub_message[i]);
559 	}
560 }
561 
562 static inline void hub_error(int fatal, u8 errors, int *error_found,
563 			int handle_error)
564 {
565 	*error_found = 1;
566 
567 	if (handle_error)
568 		do_hub_error(fatal, errors);
569 }
570 
571 #define NSI_FATAL_MASK		0x0c080081
572 #define NSI_NON_FATAL_MASK	0x23a0ba64
573 #define NSI_ERR_MASK		(NSI_FATAL_MASK | NSI_NON_FATAL_MASK)
574 
575 static char *nsi_message[30] = {
576 	"NSI Link Down",	/* NSI_FERR/NSI_NERR bit 0, fatal error */
577 	"",						/* reserved */
578 	"NSI Parity Error",				/* bit 2, non-fatal */
579 	"",						/* reserved */
580 	"",						/* reserved */
581 	"Correctable Error Message",			/* bit 5, non-fatal */
582 	"Non-Fatal Error Message",			/* bit 6, non-fatal */
583 	"Fatal Error Message",				/* bit 7, fatal */
584 	"",						/* reserved */
585 	"Receiver Error",				/* bit 9, non-fatal */
586 	"",						/* reserved */
587 	"Bad TLP",					/* bit 11, non-fatal */
588 	"Bad DLLP",					/* bit 12, non-fatal */
589 	"REPLAY_NUM Rollover",				/* bit 13, non-fatal */
590 	"",						/* reserved */
591 	"Replay Timer Timeout",				/* bit 15, non-fatal */
592 	"",						/* reserved */
593 	"",						/* reserved */
594 	"",						/* reserved */
595 	"Data Link Protocol Error",			/* bit 19, fatal */
596 	"",						/* reserved */
597 	"Poisoned TLP",					/* bit 21, non-fatal */
598 	"",						/* reserved */
599 	"Completion Timeout",				/* bit 23, non-fatal */
600 	"Completer Abort",				/* bit 24, non-fatal */
601 	"Unexpected Completion",			/* bit 25, non-fatal */
602 	"Receiver Overflow",				/* bit 26, fatal */
603 	"Malformed TLP",				/* bit 27, fatal */
604 	"",						/* reserved */
605 	"Unsupported Request"				/* bit 29, non-fatal */
606 };
607 
608 static void do_nsi_error(int fatal, u32 errors)
609 {
610 	int i;
611 
612 	for (i = 0; i < 30; i++) {
613 		if (errors & (1 << i))
614 			printk(KERN_WARNING "%sError %s\n",
615 			       fatal_message[fatal], nsi_message[i]);
616 	}
617 }
618 
619 static inline void nsi_error(int fatal, u32 errors, int *error_found,
620 		int handle_error)
621 {
622 	*error_found = 1;
623 
624 	if (handle_error)
625 		do_nsi_error(fatal, errors);
626 }
627 
628 static char *membuf_message[4] = {
629 	"Internal PMWB to DRAM parity",
630 	"Internal PMWB to System Bus Parity",
631 	"Internal System Bus or IO to PMWB Parity",
632 	"Internal DRAM to PMWB Parity"
633 };
634 
635 static void do_membuf_error(u8 errors)
636 {
637 	int i;
638 
639 	for (i = 0; i < 4; i++) {
640 		if (errors & (1 << i))
641 			e752x_printk(KERN_WARNING, "Non-Fatal Error %s\n",
642 				membuf_message[i]);
643 	}
644 }
645 
646 static inline void membuf_error(u8 errors, int *error_found, int handle_error)
647 {
648 	*error_found = 1;
649 
650 	if (handle_error)
651 		do_membuf_error(errors);
652 }
653 
654 static char *sysbus_message[10] = {
655 	"Addr or Request Parity",
656 	"Data Strobe Glitch",
657 	"Addr Strobe Glitch",
658 	"Data Parity",
659 	"Addr Above TOM",
660 	"Non DRAM Lock Error",
661 	"MCERR", "BINIT",
662 	"Memory Parity",
663 	"IO Subsystem Parity"
664 };
665 
666 static void do_sysbus_error(int fatal, u32 errors)
667 {
668 	int i;
669 
670 	for (i = 0; i < 10; i++) {
671 		if (errors & (1 << i))
672 			e752x_printk(KERN_WARNING, "%sError System Bus %s\n",
673 				fatal_message[fatal], sysbus_message[i]);
674 	}
675 }
676 
677 static inline void sysbus_error(int fatal, u32 errors, int *error_found,
678 				int handle_error)
679 {
680 	*error_found = 1;
681 
682 	if (handle_error)
683 		do_sysbus_error(fatal, errors);
684 }
685 
686 static void e752x_check_hub_interface(struct e752x_error_info *info,
687 				int *error_found, int handle_error)
688 {
689 	u8 stat8;
690 
691 	//pci_read_config_byte(dev,E752X_HI_FERR,&stat8);
692 
693 	stat8 = info->hi_ferr;
694 
695 	if (stat8 & 0x7f) {	/* Error, so process */
696 		stat8 &= 0x7f;
697 
698 		if (stat8 & 0x2b)
699 			hub_error(1, stat8 & 0x2b, error_found, handle_error);
700 
701 		if (stat8 & 0x54)
702 			hub_error(0, stat8 & 0x54, error_found, handle_error);
703 	}
704 	//pci_read_config_byte(dev,E752X_HI_NERR,&stat8);
705 
706 	stat8 = info->hi_nerr;
707 
708 	if (stat8 & 0x7f) {	/* Error, so process */
709 		stat8 &= 0x7f;
710 
711 		if (stat8 & 0x2b)
712 			hub_error(1, stat8 & 0x2b, error_found, handle_error);
713 
714 		if (stat8 & 0x54)
715 			hub_error(0, stat8 & 0x54, error_found, handle_error);
716 	}
717 }
718 
719 static void e752x_check_ns_interface(struct e752x_error_info *info,
720 				int *error_found, int handle_error)
721 {
722 	u32 stat32;
723 
724 	stat32 = info->nsi_ferr;
725 	if (stat32 & NSI_ERR_MASK) { /* Error, so process */
726 		if (stat32 & NSI_FATAL_MASK)	/* check for fatal errors */
727 			nsi_error(1, stat32 & NSI_FATAL_MASK, error_found,
728 				  handle_error);
729 		if (stat32 & NSI_NON_FATAL_MASK) /* check for non-fatal ones */
730 			nsi_error(0, stat32 & NSI_NON_FATAL_MASK, error_found,
731 				  handle_error);
732 	}
733 	stat32 = info->nsi_nerr;
734 	if (stat32 & NSI_ERR_MASK) {
735 		if (stat32 & NSI_FATAL_MASK)
736 			nsi_error(1, stat32 & NSI_FATAL_MASK, error_found,
737 				  handle_error);
738 		if (stat32 & NSI_NON_FATAL_MASK)
739 			nsi_error(0, stat32 & NSI_NON_FATAL_MASK, error_found,
740 				  handle_error);
741 	}
742 }
743 
744 static void e752x_check_sysbus(struct e752x_error_info *info,
745 			int *error_found, int handle_error)
746 {
747 	u32 stat32, error32;
748 
749 	//pci_read_config_dword(dev,E752X_SYSBUS_FERR,&stat32);
750 	stat32 = info->sysbus_ferr + (info->sysbus_nerr << 16);
751 
752 	if (stat32 == 0)
753 		return;		/* no errors */
754 
755 	error32 = (stat32 >> 16) & 0x3ff;
756 	stat32 = stat32 & 0x3ff;
757 
758 	if (stat32 & 0x087)
759 		sysbus_error(1, stat32 & 0x087, error_found, handle_error);
760 
761 	if (stat32 & 0x378)
762 		sysbus_error(0, stat32 & 0x378, error_found, handle_error);
763 
764 	if (error32 & 0x087)
765 		sysbus_error(1, error32 & 0x087, error_found, handle_error);
766 
767 	if (error32 & 0x378)
768 		sysbus_error(0, error32 & 0x378, error_found, handle_error);
769 }
770 
771 static void e752x_check_membuf(struct e752x_error_info *info,
772 			int *error_found, int handle_error)
773 {
774 	u8 stat8;
775 
776 	stat8 = info->buf_ferr;
777 
778 	if (stat8 & 0x0f) {	/* Error, so process */
779 		stat8 &= 0x0f;
780 		membuf_error(stat8, error_found, handle_error);
781 	}
782 
783 	stat8 = info->buf_nerr;
784 
785 	if (stat8 & 0x0f) {	/* Error, so process */
786 		stat8 &= 0x0f;
787 		membuf_error(stat8, error_found, handle_error);
788 	}
789 }
790 
791 static void e752x_check_dram(struct mem_ctl_info *mci,
792 			struct e752x_error_info *info, int *error_found,
793 			int handle_error)
794 {
795 	u16 error_one, error_next;
796 
797 	error_one = info->dram_ferr;
798 	error_next = info->dram_nerr;
799 
800 	/* decode and report errors */
801 	if (error_one & 0x0101)	/* check first error correctable */
802 		process_ce(mci, error_one, info->dram_sec1_add,
803 			info->dram_sec1_syndrome, error_found, handle_error);
804 
805 	if (error_next & 0x0101)	/* check next error correctable */
806 		process_ce(mci, error_next, info->dram_sec2_add,
807 			info->dram_sec2_syndrome, error_found, handle_error);
808 
809 	if (error_one & 0x4040)
810 		process_ue_no_info_wr(mci, error_found, handle_error);
811 
812 	if (error_next & 0x4040)
813 		process_ue_no_info_wr(mci, error_found, handle_error);
814 
815 	if (error_one & 0x2020)
816 		process_ded_retry(mci, error_one, info->dram_retr_add,
817 				error_found, handle_error);
818 
819 	if (error_next & 0x2020)
820 		process_ded_retry(mci, error_next, info->dram_retr_add,
821 				error_found, handle_error);
822 
823 	if (error_one & 0x0808)
824 		process_threshold_ce(mci, error_one, error_found, handle_error);
825 
826 	if (error_next & 0x0808)
827 		process_threshold_ce(mci, error_next, error_found,
828 				handle_error);
829 
830 	if (error_one & 0x0606)
831 		process_ue(mci, error_one, info->dram_ded_add,
832 			info->dram_scrb_add, error_found, handle_error);
833 
834 	if (error_next & 0x0606)
835 		process_ue(mci, error_next, info->dram_ded_add,
836 			info->dram_scrb_add, error_found, handle_error);
837 }
838 
839 static void e752x_get_error_info(struct mem_ctl_info *mci,
840 				 struct e752x_error_info *info)
841 {
842 	struct pci_dev *dev;
843 	struct e752x_pvt *pvt;
844 
845 	memset(info, 0, sizeof(*info));
846 	pvt = (struct e752x_pvt *)mci->pvt_info;
847 	dev = pvt->dev_d0f1;
848 	pci_read_config_dword(dev, E752X_FERR_GLOBAL, &info->ferr_global);
849 
850 	if (info->ferr_global) {
851 		if (pvt->dev_info->err_dev == PCI_DEVICE_ID_INTEL_3100_1_ERR) {
852 			pci_read_config_dword(dev, I3100_NSI_FERR,
853 					     &info->nsi_ferr);
854 			info->hi_ferr = 0;
855 		} else {
856 			pci_read_config_byte(dev, E752X_HI_FERR,
857 					     &info->hi_ferr);
858 			info->nsi_ferr = 0;
859 		}
860 		pci_read_config_word(dev, E752X_SYSBUS_FERR,
861 				&info->sysbus_ferr);
862 		pci_read_config_byte(dev, E752X_BUF_FERR, &info->buf_ferr);
863 		pci_read_config_word(dev, E752X_DRAM_FERR, &info->dram_ferr);
864 		pci_read_config_dword(dev, E752X_DRAM_SEC1_ADD,
865 				&info->dram_sec1_add);
866 		pci_read_config_word(dev, E752X_DRAM_SEC1_SYNDROME,
867 				&info->dram_sec1_syndrome);
868 		pci_read_config_dword(dev, E752X_DRAM_DED_ADD,
869 				&info->dram_ded_add);
870 		pci_read_config_dword(dev, E752X_DRAM_SCRB_ADD,
871 				&info->dram_scrb_add);
872 		pci_read_config_dword(dev, E752X_DRAM_RETR_ADD,
873 				&info->dram_retr_add);
874 
875 		/* ignore the reserved bits just in case */
876 		if (info->hi_ferr & 0x7f)
877 			pci_write_config_byte(dev, E752X_HI_FERR,
878 					info->hi_ferr);
879 
880 		if (info->nsi_ferr & NSI_ERR_MASK)
881 			pci_write_config_dword(dev, I3100_NSI_FERR,
882 					info->nsi_ferr);
883 
884 		if (info->sysbus_ferr)
885 			pci_write_config_word(dev, E752X_SYSBUS_FERR,
886 					info->sysbus_ferr);
887 
888 		if (info->buf_ferr & 0x0f)
889 			pci_write_config_byte(dev, E752X_BUF_FERR,
890 					info->buf_ferr);
891 
892 		if (info->dram_ferr)
893 			pci_write_bits16(pvt->dev_d0f1, E752X_DRAM_FERR,
894 					 info->dram_ferr, info->dram_ferr);
895 
896 		pci_write_config_dword(dev, E752X_FERR_GLOBAL,
897 				info->ferr_global);
898 	}
899 
900 	pci_read_config_dword(dev, E752X_NERR_GLOBAL, &info->nerr_global);
901 
902 	if (info->nerr_global) {
903 		if (pvt->dev_info->err_dev == PCI_DEVICE_ID_INTEL_3100_1_ERR) {
904 			pci_read_config_dword(dev, I3100_NSI_NERR,
905 					     &info->nsi_nerr);
906 			info->hi_nerr = 0;
907 		} else {
908 			pci_read_config_byte(dev, E752X_HI_NERR,
909 					     &info->hi_nerr);
910 			info->nsi_nerr = 0;
911 		}
912 		pci_read_config_word(dev, E752X_SYSBUS_NERR,
913 				&info->sysbus_nerr);
914 		pci_read_config_byte(dev, E752X_BUF_NERR, &info->buf_nerr);
915 		pci_read_config_word(dev, E752X_DRAM_NERR, &info->dram_nerr);
916 		pci_read_config_dword(dev, E752X_DRAM_SEC2_ADD,
917 				&info->dram_sec2_add);
918 		pci_read_config_word(dev, E752X_DRAM_SEC2_SYNDROME,
919 				&info->dram_sec2_syndrome);
920 
921 		if (info->hi_nerr & 0x7f)
922 			pci_write_config_byte(dev, E752X_HI_NERR,
923 					info->hi_nerr);
924 
925 		if (info->nsi_nerr & NSI_ERR_MASK)
926 			pci_write_config_dword(dev, I3100_NSI_NERR,
927 					info->nsi_nerr);
928 
929 		if (info->sysbus_nerr)
930 			pci_write_config_word(dev, E752X_SYSBUS_NERR,
931 					info->sysbus_nerr);
932 
933 		if (info->buf_nerr & 0x0f)
934 			pci_write_config_byte(dev, E752X_BUF_NERR,
935 					info->buf_nerr);
936 
937 		if (info->dram_nerr)
938 			pci_write_bits16(pvt->dev_d0f1, E752X_DRAM_NERR,
939 					 info->dram_nerr, info->dram_nerr);
940 
941 		pci_write_config_dword(dev, E752X_NERR_GLOBAL,
942 				info->nerr_global);
943 	}
944 }
945 
946 static int e752x_process_error_info(struct mem_ctl_info *mci,
947 				struct e752x_error_info *info,
948 				int handle_errors)
949 {
950 	u32 error32, stat32;
951 	int error_found;
952 
953 	error_found = 0;
954 	error32 = (info->ferr_global >> 18) & 0x3ff;
955 	stat32 = (info->ferr_global >> 4) & 0x7ff;
956 
957 	if (error32)
958 		global_error(1, error32, &error_found, handle_errors);
959 
960 	if (stat32)
961 		global_error(0, stat32, &error_found, handle_errors);
962 
963 	error32 = (info->nerr_global >> 18) & 0x3ff;
964 	stat32 = (info->nerr_global >> 4) & 0x7ff;
965 
966 	if (error32)
967 		global_error(1, error32, &error_found, handle_errors);
968 
969 	if (stat32)
970 		global_error(0, stat32, &error_found, handle_errors);
971 
972 	e752x_check_hub_interface(info, &error_found, handle_errors);
973 	e752x_check_ns_interface(info, &error_found, handle_errors);
974 	e752x_check_sysbus(info, &error_found, handle_errors);
975 	e752x_check_membuf(info, &error_found, handle_errors);
976 	e752x_check_dram(mci, info, &error_found, handle_errors);
977 	return error_found;
978 }
979 
980 static void e752x_check(struct mem_ctl_info *mci)
981 {
982 	struct e752x_error_info info;
983 
984 	edac_dbg(3, "\n");
985 	e752x_get_error_info(mci, &info);
986 	e752x_process_error_info(mci, &info, 1);
987 }
988 
989 /* Program byte/sec bandwidth scrub rate to hardware */
990 static int set_sdram_scrub_rate(struct mem_ctl_info *mci, u32 new_bw)
991 {
992 	const struct scrubrate *scrubrates;
993 	struct e752x_pvt *pvt = (struct e752x_pvt *) mci->pvt_info;
994 	struct pci_dev *pdev = pvt->dev_d0f0;
995 	int i;
996 
997 	if (pvt->dev_info->ctl_dev == PCI_DEVICE_ID_INTEL_3100_0)
998 		scrubrates = scrubrates_i3100;
999 	else
1000 		scrubrates = scrubrates_e752x;
1001 
1002 	/* Translate the desired scrub rate to a e752x/3100 register value.
1003 	 * Search for the bandwidth that is equal or greater than the
1004 	 * desired rate and program the cooresponding register value.
1005 	 */
1006 	for (i = 0; scrubrates[i].bandwidth != SDRATE_EOT; i++)
1007 		if (scrubrates[i].bandwidth >= new_bw)
1008 			break;
1009 
1010 	if (scrubrates[i].bandwidth == SDRATE_EOT)
1011 		return -1;
1012 
1013 	pci_write_config_word(pdev, E752X_MCHSCRB, scrubrates[i].scrubval);
1014 
1015 	return scrubrates[i].bandwidth;
1016 }
1017 
1018 /* Convert current scrub rate value into byte/sec bandwidth */
1019 static int get_sdram_scrub_rate(struct mem_ctl_info *mci)
1020 {
1021 	const struct scrubrate *scrubrates;
1022 	struct e752x_pvt *pvt = (struct e752x_pvt *) mci->pvt_info;
1023 	struct pci_dev *pdev = pvt->dev_d0f0;
1024 	u16 scrubval;
1025 	int i;
1026 
1027 	if (pvt->dev_info->ctl_dev == PCI_DEVICE_ID_INTEL_3100_0)
1028 		scrubrates = scrubrates_i3100;
1029 	else
1030 		scrubrates = scrubrates_e752x;
1031 
1032 	/* Find the bandwidth matching the memory scrubber configuration */
1033 	pci_read_config_word(pdev, E752X_MCHSCRB, &scrubval);
1034 	scrubval = scrubval & 0x0f;
1035 
1036 	for (i = 0; scrubrates[i].bandwidth != SDRATE_EOT; i++)
1037 		if (scrubrates[i].scrubval == scrubval)
1038 			break;
1039 
1040 	if (scrubrates[i].bandwidth == SDRATE_EOT) {
1041 		e752x_printk(KERN_WARNING,
1042 			"Invalid sdram scrub control value: 0x%x\n", scrubval);
1043 		return -1;
1044 	}
1045 	return scrubrates[i].bandwidth;
1046 
1047 }
1048 
1049 /* Return 1 if dual channel mode is active.  Else return 0. */
1050 static inline int dual_channel_active(u16 ddrcsr)
1051 {
1052 	return (((ddrcsr >> 12) & 3) == 3);
1053 }
1054 
1055 /* Remap csrow index numbers if map_type is "reverse"
1056  */
1057 static inline int remap_csrow_index(struct mem_ctl_info *mci, int index)
1058 {
1059 	struct e752x_pvt *pvt = mci->pvt_info;
1060 
1061 	if (!pvt->map_type)
1062 		return (7 - index);
1063 
1064 	return (index);
1065 }
1066 
1067 static void e752x_init_csrows(struct mem_ctl_info *mci, struct pci_dev *pdev,
1068 			u16 ddrcsr)
1069 {
1070 	struct csrow_info *csrow;
1071 	enum edac_type edac_mode;
1072 	unsigned long last_cumul_size;
1073 	int index, mem_dev, drc_chan;
1074 	int drc_drbg;		/* DRB granularity 0=64mb, 1=128mb */
1075 	int drc_ddim;		/* DRAM Data Integrity Mode 0=none, 2=edac */
1076 	u8 value;
1077 	u32 dra, drc, cumul_size, i, nr_pages;
1078 
1079 	dra = 0;
1080 	for (index = 0; index < 4; index++) {
1081 		u8 dra_reg;
1082 		pci_read_config_byte(pdev, E752X_DRA + index, &dra_reg);
1083 		dra |= dra_reg << (index * 8);
1084 	}
1085 	pci_read_config_dword(pdev, E752X_DRC, &drc);
1086 	drc_chan = dual_channel_active(ddrcsr) ? 1 : 0;
1087 	drc_drbg = drc_chan + 1;	/* 128 in dual mode, 64 in single */
1088 	drc_ddim = (drc >> 20) & 0x3;
1089 
1090 	/* The dram row boundary (DRB) reg values are boundary address for
1091 	 * each DRAM row with a granularity of 64 or 128MB (single/dual
1092 	 * channel operation).  DRB regs are cumulative; therefore DRB7 will
1093 	 * contain the total memory contained in all eight rows.
1094 	 */
1095 	for (last_cumul_size = index = 0; index < mci->nr_csrows; index++) {
1096 		/* mem_dev 0=x8, 1=x4 */
1097 		mem_dev = (dra >> (index * 4 + 2)) & 0x3;
1098 		csrow = mci->csrows[remap_csrow_index(mci, index)];
1099 
1100 		mem_dev = (mem_dev == 2);
1101 		pci_read_config_byte(pdev, E752X_DRB + index, &value);
1102 		/* convert a 128 or 64 MiB DRB to a page size. */
1103 		cumul_size = value << (25 + drc_drbg - PAGE_SHIFT);
1104 		edac_dbg(3, "(%d) cumul_size 0x%x\n", index, cumul_size);
1105 		if (cumul_size == last_cumul_size)
1106 			continue;	/* not populated */
1107 
1108 		csrow->first_page = last_cumul_size;
1109 		csrow->last_page = cumul_size - 1;
1110 		nr_pages = cumul_size - last_cumul_size;
1111 		last_cumul_size = cumul_size;
1112 
1113 		/*
1114 		* if single channel or x8 devices then SECDED
1115 		* if dual channel and x4 then S4ECD4ED
1116 		*/
1117 		if (drc_ddim) {
1118 			if (drc_chan && mem_dev) {
1119 				edac_mode = EDAC_S4ECD4ED;
1120 				mci->edac_cap |= EDAC_FLAG_S4ECD4ED;
1121 			} else {
1122 				edac_mode = EDAC_SECDED;
1123 				mci->edac_cap |= EDAC_FLAG_SECDED;
1124 			}
1125 		} else
1126 			edac_mode = EDAC_NONE;
1127 		for (i = 0; i < csrow->nr_channels; i++) {
1128 			struct dimm_info *dimm = csrow->channels[i]->dimm;
1129 
1130 			edac_dbg(3, "Initializing rank at (%i,%i)\n", index, i);
1131 			dimm->nr_pages = nr_pages / csrow->nr_channels;
1132 			dimm->grain = 1 << 12;	/* 4KiB - resolution of CELOG */
1133 			dimm->mtype = MEM_RDDR;	/* only one type supported */
1134 			dimm->dtype = mem_dev ? DEV_X4 : DEV_X8;
1135 			dimm->edac_mode = edac_mode;
1136 		}
1137 	}
1138 }
1139 
1140 static void e752x_init_mem_map_table(struct pci_dev *pdev,
1141 				struct e752x_pvt *pvt)
1142 {
1143 	int index;
1144 	u8 value, last, row;
1145 
1146 	last = 0;
1147 	row = 0;
1148 
1149 	for (index = 0; index < 8; index += 2) {
1150 		pci_read_config_byte(pdev, E752X_DRB + index, &value);
1151 		/* test if there is a dimm in this slot */
1152 		if (value == last) {
1153 			/* no dimm in the slot, so flag it as empty */
1154 			pvt->map[index] = 0xff;
1155 			pvt->map[index + 1] = 0xff;
1156 		} else {	/* there is a dimm in the slot */
1157 			pvt->map[index] = row;
1158 			row++;
1159 			last = value;
1160 			/* test the next value to see if the dimm is double
1161 			 * sided
1162 			 */
1163 			pci_read_config_byte(pdev, E752X_DRB + index + 1,
1164 					&value);
1165 
1166 			/* the dimm is single sided, so flag as empty */
1167 			/* this is a double sided dimm to save the next row #*/
1168 			pvt->map[index + 1] = (value == last) ? 0xff :	row;
1169 			row++;
1170 			last = value;
1171 		}
1172 	}
1173 }
1174 
1175 /* Return 0 on success or 1 on failure. */
1176 static int e752x_get_devs(struct pci_dev *pdev, int dev_idx,
1177 			struct e752x_pvt *pvt)
1178 {
1179 	pvt->dev_d0f1 = pci_get_device(PCI_VENDOR_ID_INTEL,
1180 				pvt->dev_info->err_dev, NULL);
1181 
1182 	if (pvt->dev_d0f1 == NULL) {
1183 		pvt->dev_d0f1 = pci_scan_single_device(pdev->bus,
1184 							PCI_DEVFN(0, 1));
1185 		pci_dev_get(pvt->dev_d0f1);
1186 	}
1187 
1188 	if (pvt->dev_d0f1 == NULL) {
1189 		e752x_printk(KERN_ERR, "error reporting device not found:"
1190 			"vendor %x device 0x%x (broken BIOS?)\n",
1191 			PCI_VENDOR_ID_INTEL, e752x_devs[dev_idx].err_dev);
1192 		return 1;
1193 	}
1194 
1195 	pvt->dev_d0f0 = pci_get_device(PCI_VENDOR_ID_INTEL,
1196 				e752x_devs[dev_idx].ctl_dev,
1197 				NULL);
1198 
1199 	if (pvt->dev_d0f0 == NULL)
1200 		goto fail;
1201 
1202 	return 0;
1203 
1204 fail:
1205 	pci_dev_put(pvt->dev_d0f1);
1206 	return 1;
1207 }
1208 
1209 /* Setup system bus parity mask register.
1210  * Sysbus parity supported on:
1211  * e7320/e7520/e7525 + Xeon
1212  */
1213 static void e752x_init_sysbus_parity_mask(struct e752x_pvt *pvt)
1214 {
1215 	char *cpu_id = cpu_data(0).x86_model_id;
1216 	struct pci_dev *dev = pvt->dev_d0f1;
1217 	int enable = 1;
1218 
1219 	/* Allow module parameter override, else see if CPU supports parity */
1220 	if (sysbus_parity != -1) {
1221 		enable = sysbus_parity;
1222 	} else if (cpu_id[0] && !strstr(cpu_id, "Xeon")) {
1223 		e752x_printk(KERN_INFO, "System Bus Parity not "
1224 			     "supported by CPU, disabling\n");
1225 		enable = 0;
1226 	}
1227 
1228 	if (enable)
1229 		pci_write_config_word(dev, E752X_SYSBUS_ERRMASK, 0x0000);
1230 	else
1231 		pci_write_config_word(dev, E752X_SYSBUS_ERRMASK, 0x0309);
1232 }
1233 
1234 static void e752x_init_error_reporting_regs(struct e752x_pvt *pvt)
1235 {
1236 	struct pci_dev *dev;
1237 
1238 	dev = pvt->dev_d0f1;
1239 	/* Turn off error disable & SMI in case the BIOS turned it on */
1240 	if (pvt->dev_info->err_dev == PCI_DEVICE_ID_INTEL_3100_1_ERR) {
1241 		pci_write_config_dword(dev, I3100_NSI_EMASK, 0);
1242 		pci_write_config_dword(dev, I3100_NSI_SMICMD, 0);
1243 	} else {
1244 		pci_write_config_byte(dev, E752X_HI_ERRMASK, 0x00);
1245 		pci_write_config_byte(dev, E752X_HI_SMICMD, 0x00);
1246 	}
1247 
1248 	e752x_init_sysbus_parity_mask(pvt);
1249 
1250 	pci_write_config_word(dev, E752X_SYSBUS_SMICMD, 0x00);
1251 	pci_write_config_byte(dev, E752X_BUF_ERRMASK, 0x00);
1252 	pci_write_config_byte(dev, E752X_BUF_SMICMD, 0x00);
1253 	pci_write_config_byte(dev, E752X_DRAM_ERRMASK, 0x00);
1254 	pci_write_config_byte(dev, E752X_DRAM_SMICMD, 0x00);
1255 }
1256 
1257 static int e752x_probe1(struct pci_dev *pdev, int dev_idx)
1258 {
1259 	u16 pci_data;
1260 	u8 stat8;
1261 	struct mem_ctl_info *mci;
1262 	struct edac_mc_layer layers[2];
1263 	struct e752x_pvt *pvt;
1264 	u16 ddrcsr;
1265 	int drc_chan;		/* Number of channels 0=1chan,1=2chan */
1266 	struct e752x_error_info discard;
1267 
1268 	edac_dbg(0, "mci\n");
1269 	edac_dbg(0, "Starting Probe1\n");
1270 
1271 	/* check to see if device 0 function 1 is enabled; if it isn't, we
1272 	 * assume the BIOS has reserved it for a reason and is expecting
1273 	 * exclusive access, we take care not to violate that assumption and
1274 	 * fail the probe. */
1275 	pci_read_config_byte(pdev, E752X_DEVPRES1, &stat8);
1276 	if (!force_function_unhide && !(stat8 & (1 << 5))) {
1277 		printk(KERN_INFO "Contact your BIOS vendor to see if the "
1278 			"E752x error registers can be safely un-hidden\n");
1279 		return -ENODEV;
1280 	}
1281 	stat8 |= (1 << 5);
1282 	pci_write_config_byte(pdev, E752X_DEVPRES1, stat8);
1283 
1284 	pci_read_config_word(pdev, E752X_DDRCSR, &ddrcsr);
1285 	/* FIXME: should check >>12 or 0xf, true for all? */
1286 	/* Dual channel = 1, Single channel = 0 */
1287 	drc_chan = dual_channel_active(ddrcsr);
1288 
1289 	layers[0].type = EDAC_MC_LAYER_CHIP_SELECT;
1290 	layers[0].size = E752X_NR_CSROWS;
1291 	layers[0].is_virt_csrow = true;
1292 	layers[1].type = EDAC_MC_LAYER_CHANNEL;
1293 	layers[1].size = drc_chan + 1;
1294 	layers[1].is_virt_csrow = false;
1295 	mci = edac_mc_alloc(0, ARRAY_SIZE(layers), layers, sizeof(*pvt));
1296 	if (mci == NULL)
1297 		return -ENOMEM;
1298 
1299 	edac_dbg(3, "init mci\n");
1300 	mci->mtype_cap = MEM_FLAG_RDDR;
1301 	/* 3100 IMCH supports SECDEC only */
1302 	mci->edac_ctl_cap = (dev_idx == I3100) ? EDAC_FLAG_SECDED :
1303 		(EDAC_FLAG_NONE | EDAC_FLAG_SECDED | EDAC_FLAG_S4ECD4ED);
1304 	/* FIXME - what if different memory types are in different csrows? */
1305 	mci->mod_name = EDAC_MOD_STR;
1306 	mci->mod_ver = E752X_REVISION;
1307 	mci->pdev = &pdev->dev;
1308 
1309 	edac_dbg(3, "init pvt\n");
1310 	pvt = (struct e752x_pvt *)mci->pvt_info;
1311 	pvt->dev_info = &e752x_devs[dev_idx];
1312 	pvt->mc_symmetric = ((ddrcsr & 0x10) != 0);
1313 
1314 	if (e752x_get_devs(pdev, dev_idx, pvt)) {
1315 		edac_mc_free(mci);
1316 		return -ENODEV;
1317 	}
1318 
1319 	edac_dbg(3, "more mci init\n");
1320 	mci->ctl_name = pvt->dev_info->ctl_name;
1321 	mci->dev_name = pci_name(pdev);
1322 	mci->edac_check = e752x_check;
1323 	mci->ctl_page_to_phys = ctl_page_to_phys;
1324 	mci->set_sdram_scrub_rate = set_sdram_scrub_rate;
1325 	mci->get_sdram_scrub_rate = get_sdram_scrub_rate;
1326 
1327 	/* set the map type.  1 = normal, 0 = reversed
1328 	 * Must be set before e752x_init_csrows in case csrow mapping
1329 	 * is reversed.
1330 	 */
1331 	pci_read_config_byte(pdev, E752X_DRM, &stat8);
1332 	pvt->map_type = ((stat8 & 0x0f) > ((stat8 >> 4) & 0x0f));
1333 
1334 	e752x_init_csrows(mci, pdev, ddrcsr);
1335 	e752x_init_mem_map_table(pdev, pvt);
1336 
1337 	if (dev_idx == I3100)
1338 		mci->edac_cap = EDAC_FLAG_SECDED; /* the only mode supported */
1339 	else
1340 		mci->edac_cap |= EDAC_FLAG_NONE;
1341 	edac_dbg(3, "tolm, remapbase, remaplimit\n");
1342 
1343 	/* load the top of low memory, remap base, and remap limit vars */
1344 	pci_read_config_word(pdev, E752X_TOLM, &pci_data);
1345 	pvt->tolm = ((u32) pci_data) << 4;
1346 	pci_read_config_word(pdev, E752X_REMAPBASE, &pci_data);
1347 	pvt->remapbase = ((u32) pci_data) << 14;
1348 	pci_read_config_word(pdev, E752X_REMAPLIMIT, &pci_data);
1349 	pvt->remaplimit = ((u32) pci_data) << 14;
1350 	e752x_printk(KERN_INFO,
1351 			"tolm = %x, remapbase = %x, remaplimit = %x\n",
1352 			pvt->tolm, pvt->remapbase, pvt->remaplimit);
1353 
1354 	/* Here we assume that we will never see multiple instances of this
1355 	 * type of memory controller.  The ID is therefore hardcoded to 0.
1356 	 */
1357 	if (edac_mc_add_mc(mci)) {
1358 		edac_dbg(3, "failed edac_mc_add_mc()\n");
1359 		goto fail;
1360 	}
1361 
1362 	e752x_init_error_reporting_regs(pvt);
1363 	e752x_get_error_info(mci, &discard);	/* clear other MCH errors */
1364 
1365 	/* allocating generic PCI control info */
1366 	e752x_pci = edac_pci_create_generic_ctl(&pdev->dev, EDAC_MOD_STR);
1367 	if (!e752x_pci) {
1368 		printk(KERN_WARNING
1369 			"%s(): Unable to create PCI control\n", __func__);
1370 		printk(KERN_WARNING
1371 			"%s(): PCI error report via EDAC not setup\n",
1372 			__func__);
1373 	}
1374 
1375 	/* get this far and it's successful */
1376 	edac_dbg(3, "success\n");
1377 	return 0;
1378 
1379 fail:
1380 	pci_dev_put(pvt->dev_d0f0);
1381 	pci_dev_put(pvt->dev_d0f1);
1382 	edac_mc_free(mci);
1383 
1384 	return -ENODEV;
1385 }
1386 
1387 /* returns count (>= 0), or negative on error */
1388 static int e752x_init_one(struct pci_dev *pdev, const struct pci_device_id *ent)
1389 {
1390 	edac_dbg(0, "\n");
1391 
1392 	/* wake up and enable device */
1393 	if (pci_enable_device(pdev) < 0)
1394 		return -EIO;
1395 
1396 	return e752x_probe1(pdev, ent->driver_data);
1397 }
1398 
1399 static void e752x_remove_one(struct pci_dev *pdev)
1400 {
1401 	struct mem_ctl_info *mci;
1402 	struct e752x_pvt *pvt;
1403 
1404 	edac_dbg(0, "\n");
1405 
1406 	if (e752x_pci)
1407 		edac_pci_release_generic_ctl(e752x_pci);
1408 
1409 	if ((mci = edac_mc_del_mc(&pdev->dev)) == NULL)
1410 		return;
1411 
1412 	pvt = (struct e752x_pvt *)mci->pvt_info;
1413 	pci_dev_put(pvt->dev_d0f0);
1414 	pci_dev_put(pvt->dev_d0f1);
1415 	edac_mc_free(mci);
1416 }
1417 
1418 static const struct pci_device_id e752x_pci_tbl[] = {
1419 	{
1420 	 PCI_VEND_DEV(INTEL, 7520_0), PCI_ANY_ID, PCI_ANY_ID, 0, 0,
1421 	 E7520},
1422 	{
1423 	 PCI_VEND_DEV(INTEL, 7525_0), PCI_ANY_ID, PCI_ANY_ID, 0, 0,
1424 	 E7525},
1425 	{
1426 	 PCI_VEND_DEV(INTEL, 7320_0), PCI_ANY_ID, PCI_ANY_ID, 0, 0,
1427 	 E7320},
1428 	{
1429 	 PCI_VEND_DEV(INTEL, 3100_0), PCI_ANY_ID, PCI_ANY_ID, 0, 0,
1430 	 I3100},
1431 	{
1432 	 0,
1433 	 }			/* 0 terminated list. */
1434 };
1435 
1436 MODULE_DEVICE_TABLE(pci, e752x_pci_tbl);
1437 
1438 static struct pci_driver e752x_driver = {
1439 	.name = EDAC_MOD_STR,
1440 	.probe = e752x_init_one,
1441 	.remove = e752x_remove_one,
1442 	.id_table = e752x_pci_tbl,
1443 };
1444 
1445 static int __init e752x_init(void)
1446 {
1447 	int pci_rc;
1448 
1449 	edac_dbg(3, "\n");
1450 
1451        /* Ensure that the OPSTATE is set correctly for POLL or NMI */
1452        opstate_init();
1453 
1454 	pci_rc = pci_register_driver(&e752x_driver);
1455 	return (pci_rc < 0) ? pci_rc : 0;
1456 }
1457 
1458 static void __exit e752x_exit(void)
1459 {
1460 	edac_dbg(3, "\n");
1461 	pci_unregister_driver(&e752x_driver);
1462 }
1463 
1464 module_init(e752x_init);
1465 module_exit(e752x_exit);
1466 
1467 MODULE_LICENSE("GPL");
1468 MODULE_AUTHOR("Linux Networx (http://lnxi.com) Tom Zimmerman\n");
1469 MODULE_DESCRIPTION("MC support for Intel e752x/3100 memory controllers");
1470 
1471 module_param(force_function_unhide, int, 0444);
1472 MODULE_PARM_DESC(force_function_unhide, "if BIOS sets Dev0:Fun1 up as hidden:"
1473 		 " 1=force unhide and hope BIOS doesn't fight driver for "
1474 		"Dev0:Fun1 access");
1475 
1476 module_param(edac_op_state, int, 0444);
1477 MODULE_PARM_DESC(edac_op_state, "EDAC Error Reporting state: 0=Poll,1=NMI");
1478 
1479 module_param(sysbus_parity, int, 0444);
1480 MODULE_PARM_DESC(sysbus_parity, "0=disable system bus parity checking,"
1481 		" 1=enable system bus parity checking, default=auto-detect");
1482 module_param(report_non_memory_errors, int, 0644);
1483 MODULE_PARM_DESC(report_non_memory_errors, "0=disable non-memory error "
1484 		"reporting, 1=enable non-memory error reporting");
1485