1 // SPDX-License-Identifier: GPL-2.0-only
2 /*
3 * Driver for Pondicherry2 memory controller.
4 *
5 * Copyright (c) 2016, Intel Corporation.
6 *
7 * [Derived from sb_edac.c]
8 *
9 * Translation of system physical addresses to DIMM addresses
10 * is a two stage process:
11 *
12 * First the Pondicherry 2 memory controller handles slice and channel interleaving
13 * in "sys2pmi()". This is (almost) completley common between platforms.
14 *
15 * Then a platform specific dunit (DIMM unit) completes the process to provide DIMM,
16 * rank, bank, row and column using the appropriate "dunit_ops" functions/parameters.
17 */
18
19 #include <linux/module.h>
20 #include <linux/init.h>
21 #include <linux/pci.h>
22 #include <linux/pci_ids.h>
23 #include <linux/slab.h>
24 #include <linux/delay.h>
25 #include <linux/edac.h>
26 #include <linux/mmzone.h>
27 #include <linux/smp.h>
28 #include <linux/bitmap.h>
29 #include <linux/math64.h>
30 #include <linux/mod_devicetable.h>
31 #include <linux/platform_data/x86/p2sb.h>
32
33 #include <asm/cpu_device_id.h>
34 #include <asm/intel-family.h>
35 #include <asm/processor.h>
36 #include <asm/mce.h>
37
38 #include "edac_mc.h"
39 #include "edac_module.h"
40 #include "pnd2_edac.h"
41
42 #define EDAC_MOD_STR "pnd2_edac"
43
44 #define APL_NUM_CHANNELS 4
45 #define DNV_NUM_CHANNELS 2
46 #define DNV_MAX_DIMMS 2 /* Max DIMMs per channel */
47
48 enum type {
49 APL,
50 DNV, /* All requests go to PMI CH0 on each slice (CH1 disabled) */
51 };
52
53 struct dram_addr {
54 int chan;
55 int dimm;
56 int rank;
57 int bank;
58 int row;
59 int col;
60 };
61
62 struct pnd2_pvt {
63 int dimm_geom[APL_NUM_CHANNELS];
64 u64 tolm, tohm;
65 };
66
67 /*
68 * System address space is divided into multiple regions with
69 * different interleave rules in each. The as0/as1 regions
70 * have no interleaving at all. The as2 region is interleaved
71 * between two channels. The mot region is magic and may overlap
72 * other regions, with its interleave rules taking precedence.
73 * Addresses not in any of these regions are interleaved across
74 * all four channels.
75 */
76 static struct region {
77 u64 base;
78 u64 limit;
79 u8 enabled;
80 } mot, as0, as1, as2;
81
82 static struct dunit_ops {
83 char *name;
84 enum type type;
85 int pmiaddr_shift;
86 int pmiidx_shift;
87 int channels;
88 int dimms_per_channel;
89 int (*rd_reg)(int port, int off, int op, void *data, size_t sz, char *name);
90 int (*get_registers)(void);
91 int (*check_ecc)(void);
92 void (*mk_region)(char *name, struct region *rp, void *asym);
93 void (*get_dimm_config)(struct mem_ctl_info *mci);
94 int (*pmi2mem)(struct mem_ctl_info *mci, u64 pmiaddr, u32 pmiidx,
95 struct dram_addr *daddr, char *msg);
96 } *ops;
97
98 static struct mem_ctl_info *pnd2_mci;
99
100 #define PND2_MSG_SIZE 256
101
102 /* Debug macros */
103 #define pnd2_printk(level, fmt, arg...) \
104 edac_printk(level, "pnd2", fmt, ##arg)
105
106 #define pnd2_mc_printk(mci, level, fmt, arg...) \
107 edac_mc_chipset_printk(mci, level, "pnd2", fmt, ##arg)
108
109 #define MOT_CHAN_INTLV_BIT_1SLC_2CH 12
110 #define MOT_CHAN_INTLV_BIT_2SLC_2CH 13
111 #define SELECTOR_DISABLED (-1)
112 #define _4GB (1ul << 32)
113
114 #define PMI_ADDRESS_WIDTH 31
115 #define PND_MAX_PHYS_BIT 39
116
117 #define APL_ASYMSHIFT 28
118 #define DNV_ASYMSHIFT 31
119 #define CH_HASH_MASK_LSB 6
120 #define SLICE_HASH_MASK_LSB 6
121 #define MOT_SLC_INTLV_BIT 12
122 #define LOG2_PMI_ADDR_GRANULARITY 5
123 #define MOT_SHIFT 24
124
125 #define GET_BITFIELD(v, lo, hi) (((v) & GENMASK_ULL(hi, lo)) >> (lo))
126 #define U64_LSHIFT(val, s) ((u64)(val) << (s))
127
128 /*
129 * On Apollo Lake we access memory controller registers via a
130 * side-band mailbox style interface in a hidden PCI device
131 * configuration space.
132 */
133 static struct pci_bus *p2sb_bus;
134 #define P2SB_DEVFN PCI_DEVFN(0xd, 0)
135 #define P2SB_ADDR_OFF 0xd0
136 #define P2SB_DATA_OFF 0xd4
137 #define P2SB_STAT_OFF 0xd8
138 #define P2SB_ROUT_OFF 0xda
139 #define P2SB_EADD_OFF 0xdc
140 #define P2SB_HIDE_OFF 0xe1
141
142 #define P2SB_BUSY 1
143
144 #define P2SB_READ(size, off, ptr) \
145 pci_bus_read_config_##size(p2sb_bus, P2SB_DEVFN, off, ptr)
146 #define P2SB_WRITE(size, off, val) \
147 pci_bus_write_config_##size(p2sb_bus, P2SB_DEVFN, off, val)
148
p2sb_is_busy(u16 * status)149 static bool p2sb_is_busy(u16 *status)
150 {
151 P2SB_READ(word, P2SB_STAT_OFF, status);
152
153 return !!(*status & P2SB_BUSY);
154 }
155
_apl_rd_reg(int port,int off,int op,u32 * data)156 static int _apl_rd_reg(int port, int off, int op, u32 *data)
157 {
158 int retries = 0xff, ret;
159 u16 status;
160 u8 hidden;
161
162 /* Unhide the P2SB device, if it's hidden */
163 P2SB_READ(byte, P2SB_HIDE_OFF, &hidden);
164 if (hidden)
165 P2SB_WRITE(byte, P2SB_HIDE_OFF, 0);
166
167 if (p2sb_is_busy(&status)) {
168 ret = -EAGAIN;
169 goto out;
170 }
171
172 P2SB_WRITE(dword, P2SB_ADDR_OFF, (port << 24) | off);
173 P2SB_WRITE(dword, P2SB_DATA_OFF, 0);
174 P2SB_WRITE(dword, P2SB_EADD_OFF, 0);
175 P2SB_WRITE(word, P2SB_ROUT_OFF, 0);
176 P2SB_WRITE(word, P2SB_STAT_OFF, (op << 8) | P2SB_BUSY);
177
178 while (p2sb_is_busy(&status)) {
179 if (retries-- == 0) {
180 ret = -EBUSY;
181 goto out;
182 }
183 }
184
185 P2SB_READ(dword, P2SB_DATA_OFF, data);
186 ret = (status >> 1) & 0x3;
187 out:
188 /* Hide the P2SB device, if it was hidden before */
189 if (hidden)
190 P2SB_WRITE(byte, P2SB_HIDE_OFF, hidden);
191
192 return ret;
193 }
194
apl_rd_reg(int port,int off,int op,void * data,size_t sz,char * name)195 static int apl_rd_reg(int port, int off, int op, void *data, size_t sz, char *name)
196 {
197 int ret = 0;
198
199 edac_dbg(2, "Read %s port=%x off=%x op=%x\n", name, port, off, op);
200 switch (sz) {
201 case 8:
202 ret = _apl_rd_reg(port, off + 4, op, (u32 *)(data + 4));
203 fallthrough;
204 case 4:
205 ret |= _apl_rd_reg(port, off, op, (u32 *)data);
206 pnd2_printk(KERN_DEBUG, "%s=%x%08x ret=%d\n", name,
207 sz == 8 ? *((u32 *)(data + 4)) : 0, *((u32 *)data), ret);
208 break;
209 }
210
211 return ret;
212 }
213
get_mem_ctrl_hub_base_addr(void)214 static u64 get_mem_ctrl_hub_base_addr(void)
215 {
216 struct b_cr_mchbar_lo_pci lo;
217 struct b_cr_mchbar_hi_pci hi;
218 struct pci_dev *pdev;
219
220 pdev = pci_get_device(PCI_VENDOR_ID_INTEL, 0x1980, NULL);
221 if (pdev) {
222 pci_read_config_dword(pdev, 0x48, (u32 *)&lo);
223 pci_read_config_dword(pdev, 0x4c, (u32 *)&hi);
224 pci_dev_put(pdev);
225 } else {
226 return 0;
227 }
228
229 if (!lo.enable) {
230 edac_dbg(2, "MMIO via memory controller hub base address is disabled!\n");
231 return 0;
232 }
233
234 return U64_LSHIFT(hi.base, 32) | U64_LSHIFT(lo.base, 15);
235 }
236
237 #define DNV_MCHBAR_SIZE 0x8000
238 #define DNV_SB_PORT_SIZE 0x10000
dnv_rd_reg(int port,int off,int op,void * data,size_t sz,char * name)239 static int dnv_rd_reg(int port, int off, int op, void *data, size_t sz, char *name)
240 {
241 struct pci_dev *pdev;
242 void __iomem *base;
243 struct resource r;
244 int ret;
245
246 if (op == 4) {
247 pdev = pci_get_device(PCI_VENDOR_ID_INTEL, 0x1980, NULL);
248 if (!pdev)
249 return -ENODEV;
250
251 pci_read_config_dword(pdev, off, data);
252 pci_dev_put(pdev);
253 } else {
254 /* MMIO via memory controller hub base address */
255 if (op == 0 && port == 0x4c) {
256 memset(&r, 0, sizeof(r));
257
258 r.start = get_mem_ctrl_hub_base_addr();
259 if (!r.start)
260 return -ENODEV;
261 r.end = r.start + DNV_MCHBAR_SIZE - 1;
262 } else {
263 /* MMIO via sideband register base address */
264 ret = p2sb_bar(NULL, 0, &r);
265 if (ret)
266 return ret;
267
268 r.start += (port << 16);
269 r.end = r.start + DNV_SB_PORT_SIZE - 1;
270 }
271
272 base = ioremap(r.start, resource_size(&r));
273 if (!base)
274 return -ENODEV;
275
276 if (sz == 8)
277 *(u64 *)data = readq(base + off);
278 else
279 *(u32 *)data = readl(base + off);
280
281 iounmap(base);
282 }
283
284 edac_dbg(2, "Read %s=%.8x_%.8x\n", name,
285 (sz == 8) ? *(u32 *)(data + 4) : 0, *(u32 *)data);
286
287 return 0;
288 }
289
290 #define RD_REGP(regp, regname, port) \
291 ops->rd_reg(port, \
292 regname##_offset, \
293 regname##_r_opcode, \
294 regp, sizeof(struct regname), \
295 #regname)
296
297 #define RD_REG(regp, regname) \
298 ops->rd_reg(regname ## _port, \
299 regname##_offset, \
300 regname##_r_opcode, \
301 regp, sizeof(struct regname), \
302 #regname)
303
304 static u64 top_lm, top_hm;
305 static bool two_slices;
306 static bool two_channels; /* Both PMI channels in one slice enabled */
307
308 static u8 sym_chan_mask;
309 static u8 asym_chan_mask;
310 static u8 chan_mask;
311
312 static int slice_selector = -1;
313 static int chan_selector = -1;
314 static u64 slice_hash_mask;
315 static u64 chan_hash_mask;
316
mk_region(char * name,struct region * rp,u64 base,u64 limit)317 static void mk_region(char *name, struct region *rp, u64 base, u64 limit)
318 {
319 rp->enabled = 1;
320 rp->base = base;
321 rp->limit = limit;
322 edac_dbg(2, "Region:%s [%llx, %llx]\n", name, base, limit);
323 }
324
mk_region_mask(char * name,struct region * rp,u64 base,u64 mask)325 static void mk_region_mask(char *name, struct region *rp, u64 base, u64 mask)
326 {
327 if (mask == 0) {
328 pr_info(FW_BUG "MOT mask cannot be zero\n");
329 return;
330 }
331 if (mask != GENMASK_ULL(PND_MAX_PHYS_BIT, __ffs(mask))) {
332 pr_info(FW_BUG "MOT mask not power of two\n");
333 return;
334 }
335 if (base & ~mask) {
336 pr_info(FW_BUG "MOT region base/mask alignment error\n");
337 return;
338 }
339 rp->base = base;
340 rp->limit = (base | ~mask) & GENMASK_ULL(PND_MAX_PHYS_BIT, 0);
341 rp->enabled = 1;
342 edac_dbg(2, "Region:%s [%llx, %llx]\n", name, base, rp->limit);
343 }
344
in_region(struct region * rp,u64 addr)345 static bool in_region(struct region *rp, u64 addr)
346 {
347 if (!rp->enabled)
348 return false;
349
350 return rp->base <= addr && addr <= rp->limit;
351 }
352
gen_sym_mask(struct b_cr_slice_channel_hash * p)353 static int gen_sym_mask(struct b_cr_slice_channel_hash *p)
354 {
355 int mask = 0;
356
357 if (!p->slice_0_mem_disabled)
358 mask |= p->sym_slice0_channel_enabled;
359
360 if (!p->slice_1_disabled)
361 mask |= p->sym_slice1_channel_enabled << 2;
362
363 if (p->ch_1_disabled || p->enable_pmi_dual_data_mode)
364 mask &= 0x5;
365
366 return mask;
367 }
368
gen_asym_mask(struct b_cr_slice_channel_hash * p,struct b_cr_asym_mem_region0_mchbar * as0,struct b_cr_asym_mem_region1_mchbar * as1,struct b_cr_asym_2way_mem_region_mchbar * as2way)369 static int gen_asym_mask(struct b_cr_slice_channel_hash *p,
370 struct b_cr_asym_mem_region0_mchbar *as0,
371 struct b_cr_asym_mem_region1_mchbar *as1,
372 struct b_cr_asym_2way_mem_region_mchbar *as2way)
373 {
374 const int intlv[] = { 0x5, 0xA, 0x3, 0xC };
375 int mask = 0;
376
377 if (as2way->asym_2way_interleave_enable)
378 mask = intlv[as2way->asym_2way_intlv_mode];
379 if (as0->slice0_asym_enable)
380 mask |= (1 << as0->slice0_asym_channel_select);
381 if (as1->slice1_asym_enable)
382 mask |= (4 << as1->slice1_asym_channel_select);
383 if (p->slice_0_mem_disabled)
384 mask &= 0xc;
385 if (p->slice_1_disabled)
386 mask &= 0x3;
387 if (p->ch_1_disabled || p->enable_pmi_dual_data_mode)
388 mask &= 0x5;
389
390 return mask;
391 }
392
393 static struct b_cr_tolud_pci tolud;
394 static struct b_cr_touud_lo_pci touud_lo;
395 static struct b_cr_touud_hi_pci touud_hi;
396 static struct b_cr_asym_mem_region0_mchbar asym0;
397 static struct b_cr_asym_mem_region1_mchbar asym1;
398 static struct b_cr_asym_2way_mem_region_mchbar asym_2way;
399 static struct b_cr_mot_out_base_mchbar mot_base;
400 static struct b_cr_mot_out_mask_mchbar mot_mask;
401 static struct b_cr_slice_channel_hash chash;
402
403 /* Apollo Lake dunit */
404 /*
405 * Validated on board with just two DIMMs in the [0] and [2] positions
406 * in this array. Other port number matches documentation, but caution
407 * advised.
408 */
409 static const int apl_dports[APL_NUM_CHANNELS] = { 0x18, 0x10, 0x11, 0x19 };
410 static struct d_cr_drp0 drp0[APL_NUM_CHANNELS];
411
412 /* Denverton dunit */
413 static const int dnv_dports[DNV_NUM_CHANNELS] = { 0x10, 0x12 };
414 static struct d_cr_dsch dsch;
415 static struct d_cr_ecc_ctrl ecc_ctrl[DNV_NUM_CHANNELS];
416 static struct d_cr_drp drp[DNV_NUM_CHANNELS];
417 static struct d_cr_dmap dmap[DNV_NUM_CHANNELS];
418 static struct d_cr_dmap1 dmap1[DNV_NUM_CHANNELS];
419 static struct d_cr_dmap2 dmap2[DNV_NUM_CHANNELS];
420 static struct d_cr_dmap3 dmap3[DNV_NUM_CHANNELS];
421 static struct d_cr_dmap4 dmap4[DNV_NUM_CHANNELS];
422 static struct d_cr_dmap5 dmap5[DNV_NUM_CHANNELS];
423
apl_mk_region(char * name,struct region * rp,void * asym)424 static void apl_mk_region(char *name, struct region *rp, void *asym)
425 {
426 struct b_cr_asym_mem_region0_mchbar *a = asym;
427
428 mk_region(name, rp,
429 U64_LSHIFT(a->slice0_asym_base, APL_ASYMSHIFT),
430 U64_LSHIFT(a->slice0_asym_limit, APL_ASYMSHIFT) +
431 GENMASK_ULL(APL_ASYMSHIFT - 1, 0));
432 }
433
dnv_mk_region(char * name,struct region * rp,void * asym)434 static void dnv_mk_region(char *name, struct region *rp, void *asym)
435 {
436 struct b_cr_asym_mem_region_denverton *a = asym;
437
438 mk_region(name, rp,
439 U64_LSHIFT(a->slice_asym_base, DNV_ASYMSHIFT),
440 U64_LSHIFT(a->slice_asym_limit, DNV_ASYMSHIFT) +
441 GENMASK_ULL(DNV_ASYMSHIFT - 1, 0));
442 }
443
apl_get_registers(void)444 static int apl_get_registers(void)
445 {
446 int ret = -ENODEV;
447 int i;
448
449 if (RD_REG(&asym_2way, b_cr_asym_2way_mem_region_mchbar))
450 return -ENODEV;
451
452 /*
453 * RD_REGP() will fail for unpopulated or non-existent
454 * DIMM slots. Return success if we find at least one DIMM.
455 */
456 for (i = 0; i < APL_NUM_CHANNELS; i++)
457 if (!RD_REGP(&drp0[i], d_cr_drp0, apl_dports[i]))
458 ret = 0;
459
460 return ret;
461 }
462
dnv_get_registers(void)463 static int dnv_get_registers(void)
464 {
465 int i;
466
467 if (RD_REG(&dsch, d_cr_dsch))
468 return -ENODEV;
469
470 for (i = 0; i < DNV_NUM_CHANNELS; i++)
471 if (RD_REGP(&ecc_ctrl[i], d_cr_ecc_ctrl, dnv_dports[i]) ||
472 RD_REGP(&drp[i], d_cr_drp, dnv_dports[i]) ||
473 RD_REGP(&dmap[i], d_cr_dmap, dnv_dports[i]) ||
474 RD_REGP(&dmap1[i], d_cr_dmap1, dnv_dports[i]) ||
475 RD_REGP(&dmap2[i], d_cr_dmap2, dnv_dports[i]) ||
476 RD_REGP(&dmap3[i], d_cr_dmap3, dnv_dports[i]) ||
477 RD_REGP(&dmap4[i], d_cr_dmap4, dnv_dports[i]) ||
478 RD_REGP(&dmap5[i], d_cr_dmap5, dnv_dports[i]))
479 return -ENODEV;
480
481 return 0;
482 }
483
484 /*
485 * Read all the h/w config registers once here (they don't
486 * change at run time. Figure out which address ranges have
487 * which interleave characteristics.
488 */
get_registers(void)489 static int get_registers(void)
490 {
491 const int intlv[] = { 10, 11, 12, 12 };
492
493 if (RD_REG(&tolud, b_cr_tolud_pci) ||
494 RD_REG(&touud_lo, b_cr_touud_lo_pci) ||
495 RD_REG(&touud_hi, b_cr_touud_hi_pci) ||
496 RD_REG(&asym0, b_cr_asym_mem_region0_mchbar) ||
497 RD_REG(&asym1, b_cr_asym_mem_region1_mchbar) ||
498 RD_REG(&mot_base, b_cr_mot_out_base_mchbar) ||
499 RD_REG(&mot_mask, b_cr_mot_out_mask_mchbar) ||
500 RD_REG(&chash, b_cr_slice_channel_hash))
501 return -ENODEV;
502
503 if (ops->get_registers())
504 return -ENODEV;
505
506 if (ops->type == DNV) {
507 /* PMI channel idx (always 0) for asymmetric region */
508 asym0.slice0_asym_channel_select = 0;
509 asym1.slice1_asym_channel_select = 0;
510 /* PMI channel bitmap (always 1) for symmetric region */
511 chash.sym_slice0_channel_enabled = 0x1;
512 chash.sym_slice1_channel_enabled = 0x1;
513 }
514
515 if (asym0.slice0_asym_enable)
516 ops->mk_region("as0", &as0, &asym0);
517
518 if (asym1.slice1_asym_enable)
519 ops->mk_region("as1", &as1, &asym1);
520
521 if (asym_2way.asym_2way_interleave_enable) {
522 mk_region("as2way", &as2,
523 U64_LSHIFT(asym_2way.asym_2way_base, APL_ASYMSHIFT),
524 U64_LSHIFT(asym_2way.asym_2way_limit, APL_ASYMSHIFT) +
525 GENMASK_ULL(APL_ASYMSHIFT - 1, 0));
526 }
527
528 if (mot_base.imr_en) {
529 mk_region_mask("mot", &mot,
530 U64_LSHIFT(mot_base.mot_out_base, MOT_SHIFT),
531 U64_LSHIFT(mot_mask.mot_out_mask, MOT_SHIFT));
532 }
533
534 top_lm = U64_LSHIFT(tolud.tolud, 20);
535 top_hm = U64_LSHIFT(touud_hi.touud, 32) | U64_LSHIFT(touud_lo.touud, 20);
536
537 two_slices = !chash.slice_1_disabled &&
538 !chash.slice_0_mem_disabled &&
539 (chash.sym_slice0_channel_enabled != 0) &&
540 (chash.sym_slice1_channel_enabled != 0);
541 two_channels = !chash.ch_1_disabled &&
542 !chash.enable_pmi_dual_data_mode &&
543 ((chash.sym_slice0_channel_enabled == 3) ||
544 (chash.sym_slice1_channel_enabled == 3));
545
546 sym_chan_mask = gen_sym_mask(&chash);
547 asym_chan_mask = gen_asym_mask(&chash, &asym0, &asym1, &asym_2way);
548 chan_mask = sym_chan_mask | asym_chan_mask;
549
550 if (two_slices && !two_channels) {
551 if (chash.hvm_mode)
552 slice_selector = 29;
553 else
554 slice_selector = intlv[chash.interleave_mode];
555 } else if (!two_slices && two_channels) {
556 if (chash.hvm_mode)
557 chan_selector = 29;
558 else
559 chan_selector = intlv[chash.interleave_mode];
560 } else if (two_slices && two_channels) {
561 if (chash.hvm_mode) {
562 slice_selector = 29;
563 chan_selector = 30;
564 } else {
565 slice_selector = intlv[chash.interleave_mode];
566 chan_selector = intlv[chash.interleave_mode] + 1;
567 }
568 }
569
570 if (two_slices) {
571 if (!chash.hvm_mode)
572 slice_hash_mask = chash.slice_hash_mask << SLICE_HASH_MASK_LSB;
573 if (!two_channels)
574 slice_hash_mask |= BIT_ULL(slice_selector);
575 }
576
577 if (two_channels) {
578 if (!chash.hvm_mode)
579 chan_hash_mask = chash.ch_hash_mask << CH_HASH_MASK_LSB;
580 if (!two_slices)
581 chan_hash_mask |= BIT_ULL(chan_selector);
582 }
583
584 return 0;
585 }
586
587 /* Get a contiguous memory address (remove the MMIO gap) */
remove_mmio_gap(u64 sys)588 static u64 remove_mmio_gap(u64 sys)
589 {
590 return (sys < _4GB) ? sys : sys - (_4GB - top_lm);
591 }
592
593 /* Squeeze out one address bit, shift upper part down to fill gap */
remove_addr_bit(u64 * addr,int bitidx)594 static void remove_addr_bit(u64 *addr, int bitidx)
595 {
596 u64 mask;
597
598 if (bitidx == -1)
599 return;
600
601 mask = (1ull << bitidx) - 1;
602 *addr = ((*addr >> 1) & ~mask) | (*addr & mask);
603 }
604
605 /* XOR all the bits from addr specified in mask */
hash_by_mask(u64 addr,u64 mask)606 static int hash_by_mask(u64 addr, u64 mask)
607 {
608 u64 result = addr & mask;
609
610 result = (result >> 32) ^ result;
611 result = (result >> 16) ^ result;
612 result = (result >> 8) ^ result;
613 result = (result >> 4) ^ result;
614 result = (result >> 2) ^ result;
615 result = (result >> 1) ^ result;
616
617 return (int)result & 1;
618 }
619
620 /*
621 * First stage decode. Take the system address and figure out which
622 * second stage will deal with it based on interleave modes.
623 */
sys2pmi(const u64 addr,u32 * pmiidx,u64 * pmiaddr,char * msg)624 static int sys2pmi(const u64 addr, u32 *pmiidx, u64 *pmiaddr, char *msg)
625 {
626 u64 contig_addr, contig_base, contig_offset, contig_base_adj;
627 int mot_intlv_bit = two_slices ? MOT_CHAN_INTLV_BIT_2SLC_2CH :
628 MOT_CHAN_INTLV_BIT_1SLC_2CH;
629 int slice_intlv_bit_rm = SELECTOR_DISABLED;
630 int chan_intlv_bit_rm = SELECTOR_DISABLED;
631 /* Determine if address is in the MOT region. */
632 bool mot_hit = in_region(&mot, addr);
633 /* Calculate the number of symmetric regions enabled. */
634 int sym_channels = hweight8(sym_chan_mask);
635
636 /*
637 * The amount we need to shift the asym base can be determined by the
638 * number of enabled symmetric channels.
639 * NOTE: This can only work because symmetric memory is not supposed
640 * to do a 3-way interleave.
641 */
642 int sym_chan_shift = sym_channels >> 1;
643
644 /* Give up if address is out of range, or in MMIO gap */
645 if (addr >= (1ul << PND_MAX_PHYS_BIT) ||
646 (addr >= top_lm && addr < _4GB) || addr >= top_hm) {
647 snprintf(msg, PND2_MSG_SIZE, "Error address 0x%llx is not DRAM", addr);
648 return -EINVAL;
649 }
650
651 /* Get a contiguous memory address (remove the MMIO gap) */
652 contig_addr = remove_mmio_gap(addr);
653
654 if (in_region(&as0, addr)) {
655 *pmiidx = asym0.slice0_asym_channel_select;
656
657 contig_base = remove_mmio_gap(as0.base);
658 contig_offset = contig_addr - contig_base;
659 contig_base_adj = (contig_base >> sym_chan_shift) *
660 ((chash.sym_slice0_channel_enabled >> (*pmiidx & 1)) & 1);
661 contig_addr = contig_offset + ((sym_channels > 0) ? contig_base_adj : 0ull);
662 } else if (in_region(&as1, addr)) {
663 *pmiidx = 2u + asym1.slice1_asym_channel_select;
664
665 contig_base = remove_mmio_gap(as1.base);
666 contig_offset = contig_addr - contig_base;
667 contig_base_adj = (contig_base >> sym_chan_shift) *
668 ((chash.sym_slice1_channel_enabled >> (*pmiidx & 1)) & 1);
669 contig_addr = contig_offset + ((sym_channels > 0) ? contig_base_adj : 0ull);
670 } else if (in_region(&as2, addr) && (asym_2way.asym_2way_intlv_mode == 0x3ul)) {
671 bool channel1;
672
673 mot_intlv_bit = MOT_CHAN_INTLV_BIT_1SLC_2CH;
674 *pmiidx = (asym_2way.asym_2way_intlv_mode & 1) << 1;
675 channel1 = mot_hit ? ((bool)((addr >> mot_intlv_bit) & 1)) :
676 hash_by_mask(contig_addr, chan_hash_mask);
677 *pmiidx |= (u32)channel1;
678
679 contig_base = remove_mmio_gap(as2.base);
680 chan_intlv_bit_rm = mot_hit ? mot_intlv_bit : chan_selector;
681 contig_offset = contig_addr - contig_base;
682 remove_addr_bit(&contig_offset, chan_intlv_bit_rm);
683 contig_addr = (contig_base >> sym_chan_shift) + contig_offset;
684 } else {
685 /* Otherwise we're in normal, boring symmetric mode. */
686 *pmiidx = 0u;
687
688 if (two_slices) {
689 bool slice1;
690
691 if (mot_hit) {
692 slice_intlv_bit_rm = MOT_SLC_INTLV_BIT;
693 slice1 = (addr >> MOT_SLC_INTLV_BIT) & 1;
694 } else {
695 slice_intlv_bit_rm = slice_selector;
696 slice1 = hash_by_mask(addr, slice_hash_mask);
697 }
698
699 *pmiidx = (u32)slice1 << 1;
700 }
701
702 if (two_channels) {
703 bool channel1;
704
705 mot_intlv_bit = two_slices ? MOT_CHAN_INTLV_BIT_2SLC_2CH :
706 MOT_CHAN_INTLV_BIT_1SLC_2CH;
707
708 if (mot_hit) {
709 chan_intlv_bit_rm = mot_intlv_bit;
710 channel1 = (addr >> mot_intlv_bit) & 1;
711 } else {
712 chan_intlv_bit_rm = chan_selector;
713 channel1 = hash_by_mask(contig_addr, chan_hash_mask);
714 }
715
716 *pmiidx |= (u32)channel1;
717 }
718 }
719
720 /* Remove the chan_selector bit first */
721 remove_addr_bit(&contig_addr, chan_intlv_bit_rm);
722 /* Remove the slice bit (we remove it second because it must be lower */
723 remove_addr_bit(&contig_addr, slice_intlv_bit_rm);
724 *pmiaddr = contig_addr;
725
726 return 0;
727 }
728
729 /* Translate PMI address to memory (rank, row, bank, column) */
730 #define C(n) (0x10 | (n)) /* column */
731 #define B(n) (0x20 | (n)) /* bank */
732 #define R(n) (0x40 | (n)) /* row */
733 #define RS (0x80) /* rank */
734
735 /* addrdec values */
736 #define AMAP_1KB 0
737 #define AMAP_2KB 1
738 #define AMAP_4KB 2
739 #define AMAP_RSVD 3
740
741 /* dden values */
742 #define DEN_4Gb 0
743 #define DEN_8Gb 2
744
745 /* dwid values */
746 #define X8 0
747 #define X16 1
748
749 static struct dimm_geometry {
750 u8 addrdec;
751 u8 dden;
752 u8 dwid;
753 u8 rowbits, colbits;
754 u16 bits[PMI_ADDRESS_WIDTH];
755 } dimms[] = {
756 {
757 .addrdec = AMAP_1KB, .dden = DEN_4Gb, .dwid = X16,
758 .rowbits = 15, .colbits = 10,
759 .bits = {
760 C(2), C(3), C(4), C(5), C(6), B(0), B(1), B(2), R(0),
761 R(1), R(2), R(3), R(4), R(5), R(6), R(7), R(8), R(9),
762 R(10), C(7), C(8), C(9), R(11), RS, R(12), R(13), R(14),
763 0, 0, 0, 0
764 }
765 },
766 {
767 .addrdec = AMAP_1KB, .dden = DEN_4Gb, .dwid = X8,
768 .rowbits = 16, .colbits = 10,
769 .bits = {
770 C(2), C(3), C(4), C(5), C(6), B(0), B(1), B(2), R(0),
771 R(1), R(2), R(3), R(4), R(5), R(6), R(7), R(8), R(9),
772 R(10), C(7), C(8), C(9), R(11), RS, R(12), R(13), R(14),
773 R(15), 0, 0, 0
774 }
775 },
776 {
777 .addrdec = AMAP_1KB, .dden = DEN_8Gb, .dwid = X16,
778 .rowbits = 16, .colbits = 10,
779 .bits = {
780 C(2), C(3), C(4), C(5), C(6), B(0), B(1), B(2), R(0),
781 R(1), R(2), R(3), R(4), R(5), R(6), R(7), R(8), R(9),
782 R(10), C(7), C(8), C(9), R(11), RS, R(12), R(13), R(14),
783 R(15), 0, 0, 0
784 }
785 },
786 {
787 .addrdec = AMAP_1KB, .dden = DEN_8Gb, .dwid = X8,
788 .rowbits = 16, .colbits = 11,
789 .bits = {
790 C(2), C(3), C(4), C(5), C(6), B(0), B(1), B(2), R(0),
791 R(1), R(2), R(3), R(4), R(5), R(6), R(7), R(8), R(9),
792 R(10), C(7), C(8), C(9), R(11), RS, C(11), R(12), R(13),
793 R(14), R(15), 0, 0
794 }
795 },
796 {
797 .addrdec = AMAP_2KB, .dden = DEN_4Gb, .dwid = X16,
798 .rowbits = 15, .colbits = 10,
799 .bits = {
800 C(2), C(3), C(4), C(5), C(6), C(7), B(0), B(1), B(2),
801 R(0), R(1), R(2), R(3), R(4), R(5), R(6), R(7), R(8),
802 R(9), R(10), C(8), C(9), R(11), RS, R(12), R(13), R(14),
803 0, 0, 0, 0
804 }
805 },
806 {
807 .addrdec = AMAP_2KB, .dden = DEN_4Gb, .dwid = X8,
808 .rowbits = 16, .colbits = 10,
809 .bits = {
810 C(2), C(3), C(4), C(5), C(6), C(7), B(0), B(1), B(2),
811 R(0), R(1), R(2), R(3), R(4), R(5), R(6), R(7), R(8),
812 R(9), R(10), C(8), C(9), R(11), RS, R(12), R(13), R(14),
813 R(15), 0, 0, 0
814 }
815 },
816 {
817 .addrdec = AMAP_2KB, .dden = DEN_8Gb, .dwid = X16,
818 .rowbits = 16, .colbits = 10,
819 .bits = {
820 C(2), C(3), C(4), C(5), C(6), C(7), B(0), B(1), B(2),
821 R(0), R(1), R(2), R(3), R(4), R(5), R(6), R(7), R(8),
822 R(9), R(10), C(8), C(9), R(11), RS, R(12), R(13), R(14),
823 R(15), 0, 0, 0
824 }
825 },
826 {
827 .addrdec = AMAP_2KB, .dden = DEN_8Gb, .dwid = X8,
828 .rowbits = 16, .colbits = 11,
829 .bits = {
830 C(2), C(3), C(4), C(5), C(6), C(7), B(0), B(1), B(2),
831 R(0), R(1), R(2), R(3), R(4), R(5), R(6), R(7), R(8),
832 R(9), R(10), C(8), C(9), R(11), RS, C(11), R(12), R(13),
833 R(14), R(15), 0, 0
834 }
835 },
836 {
837 .addrdec = AMAP_4KB, .dden = DEN_4Gb, .dwid = X16,
838 .rowbits = 15, .colbits = 10,
839 .bits = {
840 C(2), C(3), C(4), C(5), C(6), C(7), C(8), B(0), B(1),
841 B(2), R(0), R(1), R(2), R(3), R(4), R(5), R(6), R(7),
842 R(8), R(9), R(10), C(9), R(11), RS, R(12), R(13), R(14),
843 0, 0, 0, 0
844 }
845 },
846 {
847 .addrdec = AMAP_4KB, .dden = DEN_4Gb, .dwid = X8,
848 .rowbits = 16, .colbits = 10,
849 .bits = {
850 C(2), C(3), C(4), C(5), C(6), C(7), C(8), B(0), B(1),
851 B(2), R(0), R(1), R(2), R(3), R(4), R(5), R(6), R(7),
852 R(8), R(9), R(10), C(9), R(11), RS, R(12), R(13), R(14),
853 R(15), 0, 0, 0
854 }
855 },
856 {
857 .addrdec = AMAP_4KB, .dden = DEN_8Gb, .dwid = X16,
858 .rowbits = 16, .colbits = 10,
859 .bits = {
860 C(2), C(3), C(4), C(5), C(6), C(7), C(8), B(0), B(1),
861 B(2), R(0), R(1), R(2), R(3), R(4), R(5), R(6), R(7),
862 R(8), R(9), R(10), C(9), R(11), RS, R(12), R(13), R(14),
863 R(15), 0, 0, 0
864 }
865 },
866 {
867 .addrdec = AMAP_4KB, .dden = DEN_8Gb, .dwid = X8,
868 .rowbits = 16, .colbits = 11,
869 .bits = {
870 C(2), C(3), C(4), C(5), C(6), C(7), C(8), B(0), B(1),
871 B(2), R(0), R(1), R(2), R(3), R(4), R(5), R(6), R(7),
872 R(8), R(9), R(10), C(9), R(11), RS, C(11), R(12), R(13),
873 R(14), R(15), 0, 0
874 }
875 }
876 };
877
bank_hash(u64 pmiaddr,int idx,int shft)878 static int bank_hash(u64 pmiaddr, int idx, int shft)
879 {
880 int bhash = 0;
881
882 switch (idx) {
883 case 0:
884 bhash ^= ((pmiaddr >> (12 + shft)) ^ (pmiaddr >> (9 + shft))) & 1;
885 break;
886 case 1:
887 bhash ^= (((pmiaddr >> (10 + shft)) ^ (pmiaddr >> (8 + shft))) & 1) << 1;
888 bhash ^= ((pmiaddr >> 22) & 1) << 1;
889 break;
890 case 2:
891 bhash ^= (((pmiaddr >> (13 + shft)) ^ (pmiaddr >> (11 + shft))) & 1) << 2;
892 break;
893 }
894
895 return bhash;
896 }
897
rank_hash(u64 pmiaddr)898 static int rank_hash(u64 pmiaddr)
899 {
900 return ((pmiaddr >> 16) ^ (pmiaddr >> 10)) & 1;
901 }
902
903 /* Second stage decode. Compute rank, bank, row & column. */
apl_pmi2mem(struct mem_ctl_info * mci,u64 pmiaddr,u32 pmiidx,struct dram_addr * daddr,char * msg)904 static int apl_pmi2mem(struct mem_ctl_info *mci, u64 pmiaddr, u32 pmiidx,
905 struct dram_addr *daddr, char *msg)
906 {
907 struct d_cr_drp0 *cr_drp0 = &drp0[pmiidx];
908 struct pnd2_pvt *pvt = mci->pvt_info;
909 int g = pvt->dimm_geom[pmiidx];
910 struct dimm_geometry *d = &dimms[g];
911 int column = 0, bank = 0, row = 0, rank = 0;
912 int i, idx, type, skiprs = 0;
913
914 for (i = 0; i < PMI_ADDRESS_WIDTH; i++) {
915 int bit = (pmiaddr >> i) & 1;
916
917 if (i + skiprs >= PMI_ADDRESS_WIDTH) {
918 snprintf(msg, PND2_MSG_SIZE, "Bad dimm_geometry[] table\n");
919 return -EINVAL;
920 }
921
922 type = d->bits[i + skiprs] & ~0xf;
923 idx = d->bits[i + skiprs] & 0xf;
924
925 /*
926 * On single rank DIMMs ignore the rank select bit
927 * and shift remainder of "bits[]" down one place.
928 */
929 if (type == RS && (cr_drp0->rken0 + cr_drp0->rken1) == 1) {
930 skiprs = 1;
931 type = d->bits[i + skiprs] & ~0xf;
932 idx = d->bits[i + skiprs] & 0xf;
933 }
934
935 switch (type) {
936 case C(0):
937 column |= (bit << idx);
938 break;
939 case B(0):
940 bank |= (bit << idx);
941 if (cr_drp0->bahen)
942 bank ^= bank_hash(pmiaddr, idx, d->addrdec);
943 break;
944 case R(0):
945 row |= (bit << idx);
946 break;
947 case RS:
948 rank = bit;
949 if (cr_drp0->rsien)
950 rank ^= rank_hash(pmiaddr);
951 break;
952 default:
953 if (bit) {
954 snprintf(msg, PND2_MSG_SIZE, "Bad translation\n");
955 return -EINVAL;
956 }
957 goto done;
958 }
959 }
960
961 done:
962 daddr->col = column;
963 daddr->bank = bank;
964 daddr->row = row;
965 daddr->rank = rank;
966 daddr->dimm = 0;
967
968 return 0;
969 }
970
971 /* Pluck bit "in" from pmiaddr and return value shifted to bit "out" */
972 #define dnv_get_bit(pmi, in, out) ((int)(((pmi) >> (in)) & 1u) << (out))
973
dnv_pmi2mem(struct mem_ctl_info * mci,u64 pmiaddr,u32 pmiidx,struct dram_addr * daddr,char * msg)974 static int dnv_pmi2mem(struct mem_ctl_info *mci, u64 pmiaddr, u32 pmiidx,
975 struct dram_addr *daddr, char *msg)
976 {
977 /* Rank 0 or 1 */
978 daddr->rank = dnv_get_bit(pmiaddr, dmap[pmiidx].rs0 + 13, 0);
979 /* Rank 2 or 3 */
980 daddr->rank |= dnv_get_bit(pmiaddr, dmap[pmiidx].rs1 + 13, 1);
981
982 /*
983 * Normally ranks 0,1 are DIMM0, and 2,3 are DIMM1, but we
984 * flip them if DIMM1 is larger than DIMM0.
985 */
986 daddr->dimm = (daddr->rank >= 2) ^ drp[pmiidx].dimmflip;
987
988 daddr->bank = dnv_get_bit(pmiaddr, dmap[pmiidx].ba0 + 6, 0);
989 daddr->bank |= dnv_get_bit(pmiaddr, dmap[pmiidx].ba1 + 6, 1);
990 daddr->bank |= dnv_get_bit(pmiaddr, dmap[pmiidx].bg0 + 6, 2);
991 if (dsch.ddr4en)
992 daddr->bank |= dnv_get_bit(pmiaddr, dmap[pmiidx].bg1 + 6, 3);
993 if (dmap1[pmiidx].bxor) {
994 if (dsch.ddr4en) {
995 daddr->bank ^= dnv_get_bit(pmiaddr, dmap3[pmiidx].row6 + 6, 0);
996 daddr->bank ^= dnv_get_bit(pmiaddr, dmap3[pmiidx].row7 + 6, 1);
997 if (dsch.chan_width == 0)
998 /* 64/72 bit dram channel width */
999 daddr->bank ^= dnv_get_bit(pmiaddr, dmap5[pmiidx].ca3 + 6, 2);
1000 else
1001 /* 32/40 bit dram channel width */
1002 daddr->bank ^= dnv_get_bit(pmiaddr, dmap5[pmiidx].ca4 + 6, 2);
1003 daddr->bank ^= dnv_get_bit(pmiaddr, dmap2[pmiidx].row2 + 6, 3);
1004 } else {
1005 daddr->bank ^= dnv_get_bit(pmiaddr, dmap2[pmiidx].row2 + 6, 0);
1006 daddr->bank ^= dnv_get_bit(pmiaddr, dmap3[pmiidx].row6 + 6, 1);
1007 if (dsch.chan_width == 0)
1008 daddr->bank ^= dnv_get_bit(pmiaddr, dmap5[pmiidx].ca3 + 6, 2);
1009 else
1010 daddr->bank ^= dnv_get_bit(pmiaddr, dmap5[pmiidx].ca4 + 6, 2);
1011 }
1012 }
1013
1014 daddr->row = dnv_get_bit(pmiaddr, dmap2[pmiidx].row0 + 6, 0);
1015 daddr->row |= dnv_get_bit(pmiaddr, dmap2[pmiidx].row1 + 6, 1);
1016 daddr->row |= dnv_get_bit(pmiaddr, dmap2[pmiidx].row2 + 6, 2);
1017 daddr->row |= dnv_get_bit(pmiaddr, dmap2[pmiidx].row3 + 6, 3);
1018 daddr->row |= dnv_get_bit(pmiaddr, dmap2[pmiidx].row4 + 6, 4);
1019 daddr->row |= dnv_get_bit(pmiaddr, dmap2[pmiidx].row5 + 6, 5);
1020 daddr->row |= dnv_get_bit(pmiaddr, dmap3[pmiidx].row6 + 6, 6);
1021 daddr->row |= dnv_get_bit(pmiaddr, dmap3[pmiidx].row7 + 6, 7);
1022 daddr->row |= dnv_get_bit(pmiaddr, dmap3[pmiidx].row8 + 6, 8);
1023 daddr->row |= dnv_get_bit(pmiaddr, dmap3[pmiidx].row9 + 6, 9);
1024 daddr->row |= dnv_get_bit(pmiaddr, dmap3[pmiidx].row10 + 6, 10);
1025 daddr->row |= dnv_get_bit(pmiaddr, dmap3[pmiidx].row11 + 6, 11);
1026 daddr->row |= dnv_get_bit(pmiaddr, dmap4[pmiidx].row12 + 6, 12);
1027 daddr->row |= dnv_get_bit(pmiaddr, dmap4[pmiidx].row13 + 6, 13);
1028 if (dmap4[pmiidx].row14 != 31)
1029 daddr->row |= dnv_get_bit(pmiaddr, dmap4[pmiidx].row14 + 6, 14);
1030 if (dmap4[pmiidx].row15 != 31)
1031 daddr->row |= dnv_get_bit(pmiaddr, dmap4[pmiidx].row15 + 6, 15);
1032 if (dmap4[pmiidx].row16 != 31)
1033 daddr->row |= dnv_get_bit(pmiaddr, dmap4[pmiidx].row16 + 6, 16);
1034 if (dmap4[pmiidx].row17 != 31)
1035 daddr->row |= dnv_get_bit(pmiaddr, dmap4[pmiidx].row17 + 6, 17);
1036
1037 daddr->col = dnv_get_bit(pmiaddr, dmap5[pmiidx].ca3 + 6, 3);
1038 daddr->col |= dnv_get_bit(pmiaddr, dmap5[pmiidx].ca4 + 6, 4);
1039 daddr->col |= dnv_get_bit(pmiaddr, dmap5[pmiidx].ca5 + 6, 5);
1040 daddr->col |= dnv_get_bit(pmiaddr, dmap5[pmiidx].ca6 + 6, 6);
1041 daddr->col |= dnv_get_bit(pmiaddr, dmap5[pmiidx].ca7 + 6, 7);
1042 daddr->col |= dnv_get_bit(pmiaddr, dmap5[pmiidx].ca8 + 6, 8);
1043 daddr->col |= dnv_get_bit(pmiaddr, dmap5[pmiidx].ca9 + 6, 9);
1044 if (!dsch.ddr4en && dmap1[pmiidx].ca11 != 0x3f)
1045 daddr->col |= dnv_get_bit(pmiaddr, dmap1[pmiidx].ca11 + 13, 11);
1046
1047 return 0;
1048 }
1049
check_channel(int ch)1050 static int check_channel(int ch)
1051 {
1052 if (drp0[ch].dramtype != 0) {
1053 pnd2_printk(KERN_INFO, "Unsupported DIMM in channel %d\n", ch);
1054 return 1;
1055 } else if (drp0[ch].eccen == 0) {
1056 pnd2_printk(KERN_INFO, "ECC disabled on channel %d\n", ch);
1057 return 1;
1058 }
1059 return 0;
1060 }
1061
apl_check_ecc_active(void)1062 static int apl_check_ecc_active(void)
1063 {
1064 int i, ret = 0;
1065
1066 /* Check dramtype and ECC mode for each present DIMM */
1067 for (i = 0; i < APL_NUM_CHANNELS; i++)
1068 if (chan_mask & BIT(i))
1069 ret += check_channel(i);
1070 return ret ? -EINVAL : 0;
1071 }
1072
1073 #define DIMMS_PRESENT(d) ((d)->rken0 + (d)->rken1 + (d)->rken2 + (d)->rken3)
1074
check_unit(int ch)1075 static int check_unit(int ch)
1076 {
1077 struct d_cr_drp *d = &drp[ch];
1078
1079 if (DIMMS_PRESENT(d) && !ecc_ctrl[ch].eccen) {
1080 pnd2_printk(KERN_INFO, "ECC disabled on channel %d\n", ch);
1081 return 1;
1082 }
1083 return 0;
1084 }
1085
dnv_check_ecc_active(void)1086 static int dnv_check_ecc_active(void)
1087 {
1088 int i, ret = 0;
1089
1090 for (i = 0; i < DNV_NUM_CHANNELS; i++)
1091 ret += check_unit(i);
1092 return ret ? -EINVAL : 0;
1093 }
1094
get_memory_error_data(struct mem_ctl_info * mci,u64 addr,struct dram_addr * daddr,char * msg)1095 static int get_memory_error_data(struct mem_ctl_info *mci, u64 addr,
1096 struct dram_addr *daddr, char *msg)
1097 {
1098 u64 pmiaddr;
1099 u32 pmiidx;
1100 int ret;
1101
1102 ret = sys2pmi(addr, &pmiidx, &pmiaddr, msg);
1103 if (ret)
1104 return ret;
1105
1106 pmiaddr >>= ops->pmiaddr_shift;
1107 /* pmi channel idx to dimm channel idx */
1108 pmiidx >>= ops->pmiidx_shift;
1109 daddr->chan = pmiidx;
1110
1111 ret = ops->pmi2mem(mci, pmiaddr, pmiidx, daddr, msg);
1112 if (ret)
1113 return ret;
1114
1115 edac_dbg(0, "SysAddr=%llx PmiAddr=%llx Channel=%d DIMM=%d Rank=%d Bank=%d Row=%d Column=%d\n",
1116 addr, pmiaddr, daddr->chan, daddr->dimm, daddr->rank, daddr->bank, daddr->row, daddr->col);
1117
1118 return 0;
1119 }
1120
pnd2_mce_output_error(struct mem_ctl_info * mci,const struct mce * m,struct dram_addr * daddr)1121 static void pnd2_mce_output_error(struct mem_ctl_info *mci, const struct mce *m,
1122 struct dram_addr *daddr)
1123 {
1124 enum hw_event_mc_err_type tp_event;
1125 char *optype, msg[PND2_MSG_SIZE];
1126 bool ripv = m->mcgstatus & MCG_STATUS_RIPV;
1127 bool overflow = m->status & MCI_STATUS_OVER;
1128 bool uc_err = m->status & MCI_STATUS_UC;
1129 bool recov = m->status & MCI_STATUS_S;
1130 u32 core_err_cnt = GET_BITFIELD(m->status, 38, 52);
1131 u32 mscod = GET_BITFIELD(m->status, 16, 31);
1132 u32 errcode = GET_BITFIELD(m->status, 0, 15);
1133 u32 optypenum = GET_BITFIELD(m->status, 4, 6);
1134 int rc;
1135
1136 tp_event = uc_err ? (ripv ? HW_EVENT_ERR_UNCORRECTED : HW_EVENT_ERR_FATAL) :
1137 HW_EVENT_ERR_CORRECTED;
1138
1139 /*
1140 * According with Table 15-9 of the Intel Architecture spec vol 3A,
1141 * memory errors should fit in this mask:
1142 * 000f 0000 1mmm cccc (binary)
1143 * where:
1144 * f = Correction Report Filtering Bit. If 1, subsequent errors
1145 * won't be shown
1146 * mmm = error type
1147 * cccc = channel
1148 * If the mask doesn't match, report an error to the parsing logic
1149 */
1150 if (!((errcode & 0xef80) == 0x80)) {
1151 optype = "Can't parse: it is not a mem";
1152 } else {
1153 switch (optypenum) {
1154 case 0:
1155 optype = "generic undef request error";
1156 break;
1157 case 1:
1158 optype = "memory read error";
1159 break;
1160 case 2:
1161 optype = "memory write error";
1162 break;
1163 case 3:
1164 optype = "addr/cmd error";
1165 break;
1166 case 4:
1167 optype = "memory scrubbing error";
1168 break;
1169 default:
1170 optype = "reserved";
1171 break;
1172 }
1173 }
1174
1175 /* Only decode errors with an valid address (ADDRV) */
1176 if (!(m->status & MCI_STATUS_ADDRV))
1177 return;
1178
1179 rc = get_memory_error_data(mci, m->addr, daddr, msg);
1180 if (rc)
1181 goto address_error;
1182
1183 snprintf(msg, sizeof(msg),
1184 "%s%s err_code:%04x:%04x channel:%d DIMM:%d rank:%d row:%d bank:%d col:%d",
1185 overflow ? " OVERFLOW" : "", (uc_err && recov) ? " recoverable" : "", mscod,
1186 errcode, daddr->chan, daddr->dimm, daddr->rank, daddr->row, daddr->bank, daddr->col);
1187
1188 edac_dbg(0, "%s\n", msg);
1189
1190 /* Call the helper to output message */
1191 edac_mc_handle_error(tp_event, mci, core_err_cnt, m->addr >> PAGE_SHIFT,
1192 m->addr & ~PAGE_MASK, 0, daddr->chan, daddr->dimm, -1, optype, msg);
1193
1194 return;
1195
1196 address_error:
1197 edac_mc_handle_error(tp_event, mci, core_err_cnt, 0, 0, 0, -1, -1, -1, msg, "");
1198 }
1199
apl_get_dimm_config(struct mem_ctl_info * mci)1200 static void apl_get_dimm_config(struct mem_ctl_info *mci)
1201 {
1202 struct pnd2_pvt *pvt = mci->pvt_info;
1203 struct dimm_info *dimm;
1204 struct d_cr_drp0 *d;
1205 u64 capacity;
1206 int i, g;
1207
1208 for (i = 0; i < APL_NUM_CHANNELS; i++) {
1209 if (!(chan_mask & BIT(i)))
1210 continue;
1211
1212 dimm = edac_get_dimm(mci, i, 0, 0);
1213 if (!dimm) {
1214 edac_dbg(0, "No allocated DIMM for channel %d\n", i);
1215 continue;
1216 }
1217
1218 d = &drp0[i];
1219 for (g = 0; g < ARRAY_SIZE(dimms); g++)
1220 if (dimms[g].addrdec == d->addrdec &&
1221 dimms[g].dden == d->dden &&
1222 dimms[g].dwid == d->dwid)
1223 break;
1224
1225 if (g == ARRAY_SIZE(dimms)) {
1226 edac_dbg(0, "Channel %d: unrecognized DIMM\n", i);
1227 continue;
1228 }
1229
1230 pvt->dimm_geom[i] = g;
1231 capacity = (d->rken0 + d->rken1) * 8 * (1ul << dimms[g].rowbits) *
1232 (1ul << dimms[g].colbits);
1233 edac_dbg(0, "Channel %d: %lld MByte DIMM\n", i, capacity >> (20 - 3));
1234 dimm->nr_pages = MiB_TO_PAGES(capacity >> (20 - 3));
1235 dimm->grain = 32;
1236 dimm->dtype = (d->dwid == 0) ? DEV_X8 : DEV_X16;
1237 dimm->mtype = MEM_DDR3;
1238 dimm->edac_mode = EDAC_SECDED;
1239 snprintf(dimm->label, sizeof(dimm->label), "Slice#%d_Chan#%d", i / 2, i % 2);
1240 }
1241 }
1242
1243 static const int dnv_dtypes[] = {
1244 DEV_X8, DEV_X4, DEV_X16, DEV_UNKNOWN
1245 };
1246
dnv_get_dimm_config(struct mem_ctl_info * mci)1247 static void dnv_get_dimm_config(struct mem_ctl_info *mci)
1248 {
1249 int i, j, ranks_of_dimm[DNV_MAX_DIMMS], banks, rowbits, colbits, memtype;
1250 struct dimm_info *dimm;
1251 struct d_cr_drp *d;
1252 u64 capacity;
1253
1254 if (dsch.ddr4en) {
1255 memtype = MEM_DDR4;
1256 banks = 16;
1257 colbits = 10;
1258 } else {
1259 memtype = MEM_DDR3;
1260 banks = 8;
1261 }
1262
1263 for (i = 0; i < DNV_NUM_CHANNELS; i++) {
1264 if (dmap4[i].row14 == 31)
1265 rowbits = 14;
1266 else if (dmap4[i].row15 == 31)
1267 rowbits = 15;
1268 else if (dmap4[i].row16 == 31)
1269 rowbits = 16;
1270 else if (dmap4[i].row17 == 31)
1271 rowbits = 17;
1272 else
1273 rowbits = 18;
1274
1275 if (memtype == MEM_DDR3) {
1276 if (dmap1[i].ca11 != 0x3f)
1277 colbits = 12;
1278 else
1279 colbits = 10;
1280 }
1281
1282 d = &drp[i];
1283 /* DIMM0 is present if rank0 and/or rank1 is enabled */
1284 ranks_of_dimm[0] = d->rken0 + d->rken1;
1285 /* DIMM1 is present if rank2 and/or rank3 is enabled */
1286 ranks_of_dimm[1] = d->rken2 + d->rken3;
1287
1288 for (j = 0; j < DNV_MAX_DIMMS; j++) {
1289 if (!ranks_of_dimm[j])
1290 continue;
1291
1292 dimm = edac_get_dimm(mci, i, j, 0);
1293 if (!dimm) {
1294 edac_dbg(0, "No allocated DIMM for channel %d DIMM %d\n", i, j);
1295 continue;
1296 }
1297
1298 capacity = ranks_of_dimm[j] * banks * (1ul << rowbits) * (1ul << colbits);
1299 edac_dbg(0, "Channel %d DIMM %d: %lld MByte DIMM\n", i, j, capacity >> (20 - 3));
1300 dimm->nr_pages = MiB_TO_PAGES(capacity >> (20 - 3));
1301 dimm->grain = 32;
1302 dimm->dtype = dnv_dtypes[j ? d->dimmdwid0 : d->dimmdwid1];
1303 dimm->mtype = memtype;
1304 dimm->edac_mode = EDAC_SECDED;
1305 snprintf(dimm->label, sizeof(dimm->label), "Chan#%d_DIMM#%d", i, j);
1306 }
1307 }
1308 }
1309
pnd2_register_mci(struct mem_ctl_info ** ppmci)1310 static int pnd2_register_mci(struct mem_ctl_info **ppmci)
1311 {
1312 struct edac_mc_layer layers[2];
1313 struct mem_ctl_info *mci;
1314 struct pnd2_pvt *pvt;
1315 int rc;
1316
1317 rc = ops->check_ecc();
1318 if (rc < 0)
1319 return rc;
1320
1321 /* Allocate a new MC control structure */
1322 layers[0].type = EDAC_MC_LAYER_CHANNEL;
1323 layers[0].size = ops->channels;
1324 layers[0].is_virt_csrow = false;
1325 layers[1].type = EDAC_MC_LAYER_SLOT;
1326 layers[1].size = ops->dimms_per_channel;
1327 layers[1].is_virt_csrow = true;
1328 mci = edac_mc_alloc(0, ARRAY_SIZE(layers), layers, sizeof(*pvt));
1329 if (!mci)
1330 return -ENOMEM;
1331
1332 pvt = mci->pvt_info;
1333 memset(pvt, 0, sizeof(*pvt));
1334
1335 mci->mod_name = EDAC_MOD_STR;
1336 mci->dev_name = ops->name;
1337 mci->ctl_name = "Pondicherry2";
1338
1339 /* Get dimm basic config and the memory layout */
1340 ops->get_dimm_config(mci);
1341
1342 if (edac_mc_add_mc(mci)) {
1343 edac_dbg(0, "MC: failed edac_mc_add_mc()\n");
1344 edac_mc_free(mci);
1345 return -EINVAL;
1346 }
1347
1348 *ppmci = mci;
1349
1350 return 0;
1351 }
1352
pnd2_unregister_mci(struct mem_ctl_info * mci)1353 static void pnd2_unregister_mci(struct mem_ctl_info *mci)
1354 {
1355 if (unlikely(!mci || !mci->pvt_info)) {
1356 pnd2_printk(KERN_ERR, "Couldn't find mci handler\n");
1357 return;
1358 }
1359
1360 /* Remove MC sysfs nodes */
1361 edac_mc_del_mc(NULL);
1362 edac_dbg(1, "%s: free mci struct\n", mci->ctl_name);
1363 edac_mc_free(mci);
1364 }
1365
1366 /*
1367 * Callback function registered with core kernel mce code.
1368 * Called once for each logged error.
1369 */
pnd2_mce_check_error(struct notifier_block * nb,unsigned long val,void * data)1370 static int pnd2_mce_check_error(struct notifier_block *nb, unsigned long val, void *data)
1371 {
1372 struct mce *mce = (struct mce *)data;
1373 struct mem_ctl_info *mci;
1374 struct dram_addr daddr;
1375 char *type;
1376
1377 mci = pnd2_mci;
1378 if (!mci || (mce->kflags & MCE_HANDLED_CEC))
1379 return NOTIFY_DONE;
1380
1381 /*
1382 * Just let mcelog handle it if the error is
1383 * outside the memory controller. A memory error
1384 * is indicated by bit 7 = 1 and bits = 8-11,13-15 = 0.
1385 * bit 12 has an special meaning.
1386 */
1387 if ((mce->status & 0xefff) >> 7 != 1)
1388 return NOTIFY_DONE;
1389
1390 if (mce->mcgstatus & MCG_STATUS_MCIP)
1391 type = "Exception";
1392 else
1393 type = "Event";
1394
1395 pnd2_mc_printk(mci, KERN_INFO, "HANDLING MCE MEMORY ERROR\n");
1396 pnd2_mc_printk(mci, KERN_INFO, "CPU %u: Machine Check %s: %llx Bank %u: %llx\n",
1397 mce->extcpu, type, mce->mcgstatus, mce->bank, mce->status);
1398 pnd2_mc_printk(mci, KERN_INFO, "TSC %llx ", mce->tsc);
1399 pnd2_mc_printk(mci, KERN_INFO, "ADDR %llx ", mce->addr);
1400 pnd2_mc_printk(mci, KERN_INFO, "MISC %llx ", mce->misc);
1401 pnd2_mc_printk(mci, KERN_INFO, "PROCESSOR %u:%x TIME %llu SOCKET %u APIC %x\n",
1402 mce->cpuvendor, mce->cpuid, mce->time, mce->socketid, mce->apicid);
1403
1404 pnd2_mce_output_error(mci, mce, &daddr);
1405
1406 /* Advice mcelog that the error were handled */
1407 mce->kflags |= MCE_HANDLED_EDAC;
1408 return NOTIFY_OK;
1409 }
1410
1411 static struct notifier_block pnd2_mce_dec = {
1412 .notifier_call = pnd2_mce_check_error,
1413 .priority = MCE_PRIO_EDAC,
1414 };
1415
1416 #ifdef CONFIG_EDAC_DEBUG
1417 /*
1418 * Write an address to this file to exercise the address decode
1419 * logic in this driver.
1420 */
1421 static u64 pnd2_fake_addr;
1422 #define PND2_BLOB_SIZE 1024
1423 static char pnd2_result[PND2_BLOB_SIZE];
1424 static struct dentry *pnd2_test;
1425 static struct debugfs_blob_wrapper pnd2_blob = {
1426 .data = pnd2_result,
1427 .size = 0
1428 };
1429
debugfs_u64_set(void * data,u64 val)1430 static int debugfs_u64_set(void *data, u64 val)
1431 {
1432 struct dram_addr daddr;
1433 struct mce m;
1434
1435 *(u64 *)data = val;
1436 m.mcgstatus = 0;
1437 /* ADDRV + MemRd + Unknown channel */
1438 m.status = MCI_STATUS_ADDRV + 0x9f;
1439 m.addr = val;
1440 pnd2_mce_output_error(pnd2_mci, &m, &daddr);
1441 snprintf(pnd2_blob.data, PND2_BLOB_SIZE,
1442 "SysAddr=%llx Channel=%d DIMM=%d Rank=%d Bank=%d Row=%d Column=%d\n",
1443 m.addr, daddr.chan, daddr.dimm, daddr.rank, daddr.bank, daddr.row, daddr.col);
1444 pnd2_blob.size = strlen(pnd2_blob.data);
1445
1446 return 0;
1447 }
1448 DEFINE_DEBUGFS_ATTRIBUTE(fops_u64_wo, NULL, debugfs_u64_set, "%llu\n");
1449
setup_pnd2_debug(void)1450 static void setup_pnd2_debug(void)
1451 {
1452 pnd2_test = edac_debugfs_create_dir("pnd2_test");
1453 edac_debugfs_create_file("pnd2_debug_addr", 0200, pnd2_test,
1454 &pnd2_fake_addr, &fops_u64_wo);
1455 debugfs_create_blob("pnd2_debug_results", 0400, pnd2_test, &pnd2_blob);
1456 }
1457
teardown_pnd2_debug(void)1458 static void teardown_pnd2_debug(void)
1459 {
1460 debugfs_remove_recursive(pnd2_test);
1461 }
1462 #else
setup_pnd2_debug(void)1463 static void setup_pnd2_debug(void) {}
teardown_pnd2_debug(void)1464 static void teardown_pnd2_debug(void) {}
1465 #endif /* CONFIG_EDAC_DEBUG */
1466
1467
pnd2_probe(void)1468 static int pnd2_probe(void)
1469 {
1470 int rc;
1471
1472 edac_dbg(2, "\n");
1473 rc = get_registers();
1474 if (rc)
1475 return rc;
1476
1477 return pnd2_register_mci(&pnd2_mci);
1478 }
1479
pnd2_remove(void)1480 static void pnd2_remove(void)
1481 {
1482 edac_dbg(0, "\n");
1483 pnd2_unregister_mci(pnd2_mci);
1484 }
1485
1486 static struct dunit_ops apl_ops = {
1487 .name = "pnd2/apl",
1488 .type = APL,
1489 .pmiaddr_shift = LOG2_PMI_ADDR_GRANULARITY,
1490 .pmiidx_shift = 0,
1491 .channels = APL_NUM_CHANNELS,
1492 .dimms_per_channel = 1,
1493 .rd_reg = apl_rd_reg,
1494 .get_registers = apl_get_registers,
1495 .check_ecc = apl_check_ecc_active,
1496 .mk_region = apl_mk_region,
1497 .get_dimm_config = apl_get_dimm_config,
1498 .pmi2mem = apl_pmi2mem,
1499 };
1500
1501 static struct dunit_ops dnv_ops = {
1502 .name = "pnd2/dnv",
1503 .type = DNV,
1504 .pmiaddr_shift = 0,
1505 .pmiidx_shift = 1,
1506 .channels = DNV_NUM_CHANNELS,
1507 .dimms_per_channel = 2,
1508 .rd_reg = dnv_rd_reg,
1509 .get_registers = dnv_get_registers,
1510 .check_ecc = dnv_check_ecc_active,
1511 .mk_region = dnv_mk_region,
1512 .get_dimm_config = dnv_get_dimm_config,
1513 .pmi2mem = dnv_pmi2mem,
1514 };
1515
1516 static const struct x86_cpu_id pnd2_cpuids[] = {
1517 X86_MATCH_INTEL_FAM6_MODEL(ATOM_GOLDMONT, &apl_ops),
1518 X86_MATCH_INTEL_FAM6_MODEL(ATOM_GOLDMONT_D, &dnv_ops),
1519 { }
1520 };
1521 MODULE_DEVICE_TABLE(x86cpu, pnd2_cpuids);
1522
pnd2_init(void)1523 static int __init pnd2_init(void)
1524 {
1525 const struct x86_cpu_id *id;
1526 const char *owner;
1527 int rc;
1528
1529 edac_dbg(2, "\n");
1530
1531 if (ghes_get_devices())
1532 return -EBUSY;
1533
1534 owner = edac_get_owner();
1535 if (owner && strncmp(owner, EDAC_MOD_STR, sizeof(EDAC_MOD_STR)))
1536 return -EBUSY;
1537
1538 if (cpu_feature_enabled(X86_FEATURE_HYPERVISOR))
1539 return -ENODEV;
1540
1541 id = x86_match_cpu(pnd2_cpuids);
1542 if (!id)
1543 return -ENODEV;
1544
1545 ops = (struct dunit_ops *)id->driver_data;
1546
1547 if (ops->type == APL) {
1548 p2sb_bus = pci_find_bus(0, 0);
1549 if (!p2sb_bus)
1550 return -ENODEV;
1551 }
1552
1553 /* Ensure that the OPSTATE is set correctly for POLL or NMI */
1554 opstate_init();
1555
1556 rc = pnd2_probe();
1557 if (rc < 0) {
1558 pnd2_printk(KERN_ERR, "Failed to register device with error %d.\n", rc);
1559 return rc;
1560 }
1561
1562 if (!pnd2_mci)
1563 return -ENODEV;
1564
1565 mce_register_decode_chain(&pnd2_mce_dec);
1566 setup_pnd2_debug();
1567
1568 return 0;
1569 }
1570
pnd2_exit(void)1571 static void __exit pnd2_exit(void)
1572 {
1573 edac_dbg(2, "\n");
1574 teardown_pnd2_debug();
1575 mce_unregister_decode_chain(&pnd2_mce_dec);
1576 pnd2_remove();
1577 }
1578
1579 module_init(pnd2_init);
1580 module_exit(pnd2_exit);
1581
1582 module_param(edac_op_state, int, 0444);
1583 MODULE_PARM_DESC(edac_op_state, "EDAC Error Reporting state: 0=Poll,1=NMI");
1584
1585 MODULE_LICENSE("GPL v2");
1586 MODULE_AUTHOR("Tony Luck");
1587 MODULE_DESCRIPTION("MC Driver for Intel SoC using Pondicherry memory controller");
1588