1 // SPDX-License-Identifier: GPL-2.0-only 2 #include "amd64_edac.h" 3 #include <asm/amd_nb.h> 4 5 static struct edac_pci_ctl_info *pci_ctl; 6 7 /* 8 * Set by command line parameter. If BIOS has enabled the ECC, this override is 9 * cleared to prevent re-enabling the hardware by this driver. 10 */ 11 static int ecc_enable_override; 12 module_param(ecc_enable_override, int, 0644); 13 14 static struct msr __percpu *msrs; 15 16 static struct amd64_family_type *fam_type; 17 18 static inline u32 get_umc_reg(u32 reg) 19 { 20 if (!fam_type->flags.zn_regs_v2) 21 return reg; 22 23 switch (reg) { 24 case UMCCH_ADDR_CFG: return UMCCH_ADDR_CFG_DDR5; 25 case UMCCH_ADDR_MASK_SEC: return UMCCH_ADDR_MASK_SEC_DDR5; 26 case UMCCH_DIMM_CFG: return UMCCH_DIMM_CFG_DDR5; 27 } 28 29 WARN_ONCE(1, "%s: unknown register 0x%x", __func__, reg); 30 return 0; 31 } 32 33 /* Per-node stuff */ 34 static struct ecc_settings **ecc_stngs; 35 36 /* Device for the PCI component */ 37 static struct device *pci_ctl_dev; 38 39 /* 40 * Valid scrub rates for the K8 hardware memory scrubber. We map the scrubbing 41 * bandwidth to a valid bit pattern. The 'set' operation finds the 'matching- 42 * or higher value'. 43 * 44 *FIXME: Produce a better mapping/linearisation. 45 */ 46 static const struct scrubrate { 47 u32 scrubval; /* bit pattern for scrub rate */ 48 u32 bandwidth; /* bandwidth consumed (bytes/sec) */ 49 } scrubrates[] = { 50 { 0x01, 1600000000UL}, 51 { 0x02, 800000000UL}, 52 { 0x03, 400000000UL}, 53 { 0x04, 200000000UL}, 54 { 0x05, 100000000UL}, 55 { 0x06, 50000000UL}, 56 { 0x07, 25000000UL}, 57 { 0x08, 12284069UL}, 58 { 0x09, 6274509UL}, 59 { 0x0A, 3121951UL}, 60 { 0x0B, 1560975UL}, 61 { 0x0C, 781440UL}, 62 { 0x0D, 390720UL}, 63 { 0x0E, 195300UL}, 64 { 0x0F, 97650UL}, 65 { 0x10, 48854UL}, 66 { 0x11, 24427UL}, 67 { 0x12, 12213UL}, 68 { 0x13, 6101UL}, 69 { 0x14, 3051UL}, 70 { 0x15, 1523UL}, 71 { 0x16, 761UL}, 72 { 0x00, 0UL}, /* scrubbing off */ 73 }; 74 75 int __amd64_read_pci_cfg_dword(struct pci_dev *pdev, int offset, 76 u32 *val, const char *func) 77 { 78 int err = 0; 79 80 err = pci_read_config_dword(pdev, offset, val); 81 if (err) 82 amd64_warn("%s: error reading F%dx%03x.\n", 83 func, PCI_FUNC(pdev->devfn), offset); 84 85 return err; 86 } 87 88 int __amd64_write_pci_cfg_dword(struct pci_dev *pdev, int offset, 89 u32 val, const char *func) 90 { 91 int err = 0; 92 93 err = pci_write_config_dword(pdev, offset, val); 94 if (err) 95 amd64_warn("%s: error writing to F%dx%03x.\n", 96 func, PCI_FUNC(pdev->devfn), offset); 97 98 return err; 99 } 100 101 /* 102 * Select DCT to which PCI cfg accesses are routed 103 */ 104 static void f15h_select_dct(struct amd64_pvt *pvt, u8 dct) 105 { 106 u32 reg = 0; 107 108 amd64_read_pci_cfg(pvt->F1, DCT_CFG_SEL, ®); 109 reg &= (pvt->model == 0x30) ? ~3 : ~1; 110 reg |= dct; 111 amd64_write_pci_cfg(pvt->F1, DCT_CFG_SEL, reg); 112 } 113 114 /* 115 * 116 * Depending on the family, F2 DCT reads need special handling: 117 * 118 * K8: has a single DCT only and no address offsets >= 0x100 119 * 120 * F10h: each DCT has its own set of regs 121 * DCT0 -> F2x040.. 122 * DCT1 -> F2x140.. 123 * 124 * F16h: has only 1 DCT 125 * 126 * F15h: we select which DCT we access using F1x10C[DctCfgSel] 127 */ 128 static inline int amd64_read_dct_pci_cfg(struct amd64_pvt *pvt, u8 dct, 129 int offset, u32 *val) 130 { 131 switch (pvt->fam) { 132 case 0xf: 133 if (dct || offset >= 0x100) 134 return -EINVAL; 135 break; 136 137 case 0x10: 138 if (dct) { 139 /* 140 * Note: If ganging is enabled, barring the regs 141 * F2x[1,0]98 and F2x[1,0]9C; reads reads to F2x1xx 142 * return 0. (cf. Section 2.8.1 F10h BKDG) 143 */ 144 if (dct_ganging_enabled(pvt)) 145 return 0; 146 147 offset += 0x100; 148 } 149 break; 150 151 case 0x15: 152 /* 153 * F15h: F2x1xx addresses do not map explicitly to DCT1. 154 * We should select which DCT we access using F1x10C[DctCfgSel] 155 */ 156 dct = (dct && pvt->model == 0x30) ? 3 : dct; 157 f15h_select_dct(pvt, dct); 158 break; 159 160 case 0x16: 161 if (dct) 162 return -EINVAL; 163 break; 164 165 default: 166 break; 167 } 168 return amd64_read_pci_cfg(pvt->F2, offset, val); 169 } 170 171 /* 172 * Memory scrubber control interface. For K8, memory scrubbing is handled by 173 * hardware and can involve L2 cache, dcache as well as the main memory. With 174 * F10, this is extended to L3 cache scrubbing on CPU models sporting that 175 * functionality. 176 * 177 * This causes the "units" for the scrubbing speed to vary from 64 byte blocks 178 * (dram) over to cache lines. This is nasty, so we will use bandwidth in 179 * bytes/sec for the setting. 180 * 181 * Currently, we only do dram scrubbing. If the scrubbing is done in software on 182 * other archs, we might not have access to the caches directly. 183 */ 184 185 static inline void __f17h_set_scrubval(struct amd64_pvt *pvt, u32 scrubval) 186 { 187 /* 188 * Fam17h supports scrub values between 0x5 and 0x14. Also, the values 189 * are shifted down by 0x5, so scrubval 0x5 is written to the register 190 * as 0x0, scrubval 0x6 as 0x1, etc. 191 */ 192 if (scrubval >= 0x5 && scrubval <= 0x14) { 193 scrubval -= 0x5; 194 pci_write_bits32(pvt->F6, F17H_SCR_LIMIT_ADDR, scrubval, 0xF); 195 pci_write_bits32(pvt->F6, F17H_SCR_BASE_ADDR, 1, 0x1); 196 } else { 197 pci_write_bits32(pvt->F6, F17H_SCR_BASE_ADDR, 0, 0x1); 198 } 199 } 200 /* 201 * Scan the scrub rate mapping table for a close or matching bandwidth value to 202 * issue. If requested is too big, then use last maximum value found. 203 */ 204 static int __set_scrub_rate(struct amd64_pvt *pvt, u32 new_bw, u32 min_rate) 205 { 206 u32 scrubval; 207 int i; 208 209 /* 210 * map the configured rate (new_bw) to a value specific to the AMD64 211 * memory controller and apply to register. Search for the first 212 * bandwidth entry that is greater or equal than the setting requested 213 * and program that. If at last entry, turn off DRAM scrubbing. 214 * 215 * If no suitable bandwidth is found, turn off DRAM scrubbing entirely 216 * by falling back to the last element in scrubrates[]. 217 */ 218 for (i = 0; i < ARRAY_SIZE(scrubrates) - 1; i++) { 219 /* 220 * skip scrub rates which aren't recommended 221 * (see F10 BKDG, F3x58) 222 */ 223 if (scrubrates[i].scrubval < min_rate) 224 continue; 225 226 if (scrubrates[i].bandwidth <= new_bw) 227 break; 228 } 229 230 scrubval = scrubrates[i].scrubval; 231 232 if (pvt->umc) { 233 __f17h_set_scrubval(pvt, scrubval); 234 } else if (pvt->fam == 0x15 && pvt->model == 0x60) { 235 f15h_select_dct(pvt, 0); 236 pci_write_bits32(pvt->F2, F15H_M60H_SCRCTRL, scrubval, 0x001F); 237 f15h_select_dct(pvt, 1); 238 pci_write_bits32(pvt->F2, F15H_M60H_SCRCTRL, scrubval, 0x001F); 239 } else { 240 pci_write_bits32(pvt->F3, SCRCTRL, scrubval, 0x001F); 241 } 242 243 if (scrubval) 244 return scrubrates[i].bandwidth; 245 246 return 0; 247 } 248 249 static int set_scrub_rate(struct mem_ctl_info *mci, u32 bw) 250 { 251 struct amd64_pvt *pvt = mci->pvt_info; 252 u32 min_scrubrate = 0x5; 253 254 if (pvt->fam == 0xf) 255 min_scrubrate = 0x0; 256 257 if (pvt->fam == 0x15) { 258 /* Erratum #505 */ 259 if (pvt->model < 0x10) 260 f15h_select_dct(pvt, 0); 261 262 if (pvt->model == 0x60) 263 min_scrubrate = 0x6; 264 } 265 return __set_scrub_rate(pvt, bw, min_scrubrate); 266 } 267 268 static int get_scrub_rate(struct mem_ctl_info *mci) 269 { 270 struct amd64_pvt *pvt = mci->pvt_info; 271 int i, retval = -EINVAL; 272 u32 scrubval = 0; 273 274 if (pvt->umc) { 275 amd64_read_pci_cfg(pvt->F6, F17H_SCR_BASE_ADDR, &scrubval); 276 if (scrubval & BIT(0)) { 277 amd64_read_pci_cfg(pvt->F6, F17H_SCR_LIMIT_ADDR, &scrubval); 278 scrubval &= 0xF; 279 scrubval += 0x5; 280 } else { 281 scrubval = 0; 282 } 283 } else if (pvt->fam == 0x15) { 284 /* Erratum #505 */ 285 if (pvt->model < 0x10) 286 f15h_select_dct(pvt, 0); 287 288 if (pvt->model == 0x60) 289 amd64_read_pci_cfg(pvt->F2, F15H_M60H_SCRCTRL, &scrubval); 290 else 291 amd64_read_pci_cfg(pvt->F3, SCRCTRL, &scrubval); 292 } else { 293 amd64_read_pci_cfg(pvt->F3, SCRCTRL, &scrubval); 294 } 295 296 scrubval = scrubval & 0x001F; 297 298 for (i = 0; i < ARRAY_SIZE(scrubrates); i++) { 299 if (scrubrates[i].scrubval == scrubval) { 300 retval = scrubrates[i].bandwidth; 301 break; 302 } 303 } 304 return retval; 305 } 306 307 /* 308 * returns true if the SysAddr given by sys_addr matches the 309 * DRAM base/limit associated with node_id 310 */ 311 static bool base_limit_match(struct amd64_pvt *pvt, u64 sys_addr, u8 nid) 312 { 313 u64 addr; 314 315 /* The K8 treats this as a 40-bit value. However, bits 63-40 will be 316 * all ones if the most significant implemented address bit is 1. 317 * Here we discard bits 63-40. See section 3.4.2 of AMD publication 318 * 24592: AMD x86-64 Architecture Programmer's Manual Volume 1 319 * Application Programming. 320 */ 321 addr = sys_addr & 0x000000ffffffffffull; 322 323 return ((addr >= get_dram_base(pvt, nid)) && 324 (addr <= get_dram_limit(pvt, nid))); 325 } 326 327 /* 328 * Attempt to map a SysAddr to a node. On success, return a pointer to the 329 * mem_ctl_info structure for the node that the SysAddr maps to. 330 * 331 * On failure, return NULL. 332 */ 333 static struct mem_ctl_info *find_mc_by_sys_addr(struct mem_ctl_info *mci, 334 u64 sys_addr) 335 { 336 struct amd64_pvt *pvt; 337 u8 node_id; 338 u32 intlv_en, bits; 339 340 /* 341 * Here we use the DRAM Base (section 3.4.4.1) and DRAM Limit (section 342 * 3.4.4.2) registers to map the SysAddr to a node ID. 343 */ 344 pvt = mci->pvt_info; 345 346 /* 347 * The value of this field should be the same for all DRAM Base 348 * registers. Therefore we arbitrarily choose to read it from the 349 * register for node 0. 350 */ 351 intlv_en = dram_intlv_en(pvt, 0); 352 353 if (intlv_en == 0) { 354 for (node_id = 0; node_id < DRAM_RANGES; node_id++) { 355 if (base_limit_match(pvt, sys_addr, node_id)) 356 goto found; 357 } 358 goto err_no_match; 359 } 360 361 if (unlikely((intlv_en != 0x01) && 362 (intlv_en != 0x03) && 363 (intlv_en != 0x07))) { 364 amd64_warn("DRAM Base[IntlvEn] junk value: 0x%x, BIOS bug?\n", intlv_en); 365 return NULL; 366 } 367 368 bits = (((u32) sys_addr) >> 12) & intlv_en; 369 370 for (node_id = 0; ; ) { 371 if ((dram_intlv_sel(pvt, node_id) & intlv_en) == bits) 372 break; /* intlv_sel field matches */ 373 374 if (++node_id >= DRAM_RANGES) 375 goto err_no_match; 376 } 377 378 /* sanity test for sys_addr */ 379 if (unlikely(!base_limit_match(pvt, sys_addr, node_id))) { 380 amd64_warn("%s: sys_addr 0x%llx falls outside base/limit address" 381 "range for node %d with node interleaving enabled.\n", 382 __func__, sys_addr, node_id); 383 return NULL; 384 } 385 386 found: 387 return edac_mc_find((int)node_id); 388 389 err_no_match: 390 edac_dbg(2, "sys_addr 0x%lx doesn't match any node\n", 391 (unsigned long)sys_addr); 392 393 return NULL; 394 } 395 396 /* 397 * compute the CS base address of the @csrow on the DRAM controller @dct. 398 * For details see F2x[5C:40] in the processor's BKDG 399 */ 400 static void get_cs_base_and_mask(struct amd64_pvt *pvt, int csrow, u8 dct, 401 u64 *base, u64 *mask) 402 { 403 u64 csbase, csmask, base_bits, mask_bits; 404 u8 addr_shift; 405 406 if (pvt->fam == 0xf && pvt->ext_model < K8_REV_F) { 407 csbase = pvt->csels[dct].csbases[csrow]; 408 csmask = pvt->csels[dct].csmasks[csrow]; 409 base_bits = GENMASK_ULL(31, 21) | GENMASK_ULL(15, 9); 410 mask_bits = GENMASK_ULL(29, 21) | GENMASK_ULL(15, 9); 411 addr_shift = 4; 412 413 /* 414 * F16h and F15h, models 30h and later need two addr_shift values: 415 * 8 for high and 6 for low (cf. F16h BKDG). 416 */ 417 } else if (pvt->fam == 0x16 || 418 (pvt->fam == 0x15 && pvt->model >= 0x30)) { 419 csbase = pvt->csels[dct].csbases[csrow]; 420 csmask = pvt->csels[dct].csmasks[csrow >> 1]; 421 422 *base = (csbase & GENMASK_ULL(15, 5)) << 6; 423 *base |= (csbase & GENMASK_ULL(30, 19)) << 8; 424 425 *mask = ~0ULL; 426 /* poke holes for the csmask */ 427 *mask &= ~((GENMASK_ULL(15, 5) << 6) | 428 (GENMASK_ULL(30, 19) << 8)); 429 430 *mask |= (csmask & GENMASK_ULL(15, 5)) << 6; 431 *mask |= (csmask & GENMASK_ULL(30, 19)) << 8; 432 433 return; 434 } else { 435 csbase = pvt->csels[dct].csbases[csrow]; 436 csmask = pvt->csels[dct].csmasks[csrow >> 1]; 437 addr_shift = 8; 438 439 if (pvt->fam == 0x15) 440 base_bits = mask_bits = 441 GENMASK_ULL(30,19) | GENMASK_ULL(13,5); 442 else 443 base_bits = mask_bits = 444 GENMASK_ULL(28,19) | GENMASK_ULL(13,5); 445 } 446 447 *base = (csbase & base_bits) << addr_shift; 448 449 *mask = ~0ULL; 450 /* poke holes for the csmask */ 451 *mask &= ~(mask_bits << addr_shift); 452 /* OR them in */ 453 *mask |= (csmask & mask_bits) << addr_shift; 454 } 455 456 #define for_each_chip_select(i, dct, pvt) \ 457 for (i = 0; i < pvt->csels[dct].b_cnt; i++) 458 459 #define chip_select_base(i, dct, pvt) \ 460 pvt->csels[dct].csbases[i] 461 462 #define for_each_chip_select_mask(i, dct, pvt) \ 463 for (i = 0; i < pvt->csels[dct].m_cnt; i++) 464 465 #define for_each_umc(i) \ 466 for (i = 0; i < fam_type->max_mcs; i++) 467 468 /* 469 * @input_addr is an InputAddr associated with the node given by mci. Return the 470 * csrow that input_addr maps to, or -1 on failure (no csrow claims input_addr). 471 */ 472 static int input_addr_to_csrow(struct mem_ctl_info *mci, u64 input_addr) 473 { 474 struct amd64_pvt *pvt; 475 int csrow; 476 u64 base, mask; 477 478 pvt = mci->pvt_info; 479 480 for_each_chip_select(csrow, 0, pvt) { 481 if (!csrow_enabled(csrow, 0, pvt)) 482 continue; 483 484 get_cs_base_and_mask(pvt, csrow, 0, &base, &mask); 485 486 mask = ~mask; 487 488 if ((input_addr & mask) == (base & mask)) { 489 edac_dbg(2, "InputAddr 0x%lx matches csrow %d (node %d)\n", 490 (unsigned long)input_addr, csrow, 491 pvt->mc_node_id); 492 493 return csrow; 494 } 495 } 496 edac_dbg(2, "no matching csrow for InputAddr 0x%lx (MC node %d)\n", 497 (unsigned long)input_addr, pvt->mc_node_id); 498 499 return -1; 500 } 501 502 /* 503 * Obtain info from the DRAM Hole Address Register (section 3.4.8, pub #26094) 504 * for the node represented by mci. Info is passed back in *hole_base, 505 * *hole_offset, and *hole_size. Function returns 0 if info is valid or 1 if 506 * info is invalid. Info may be invalid for either of the following reasons: 507 * 508 * - The revision of the node is not E or greater. In this case, the DRAM Hole 509 * Address Register does not exist. 510 * 511 * - The DramHoleValid bit is cleared in the DRAM Hole Address Register, 512 * indicating that its contents are not valid. 513 * 514 * The values passed back in *hole_base, *hole_offset, and *hole_size are 515 * complete 32-bit values despite the fact that the bitfields in the DHAR 516 * only represent bits 31-24 of the base and offset values. 517 */ 518 static int get_dram_hole_info(struct mem_ctl_info *mci, u64 *hole_base, 519 u64 *hole_offset, u64 *hole_size) 520 { 521 struct amd64_pvt *pvt = mci->pvt_info; 522 523 /* only revE and later have the DRAM Hole Address Register */ 524 if (pvt->fam == 0xf && pvt->ext_model < K8_REV_E) { 525 edac_dbg(1, " revision %d for node %d does not support DHAR\n", 526 pvt->ext_model, pvt->mc_node_id); 527 return 1; 528 } 529 530 /* valid for Fam10h and above */ 531 if (pvt->fam >= 0x10 && !dhar_mem_hoist_valid(pvt)) { 532 edac_dbg(1, " Dram Memory Hoisting is DISABLED on this system\n"); 533 return 1; 534 } 535 536 if (!dhar_valid(pvt)) { 537 edac_dbg(1, " Dram Memory Hoisting is DISABLED on this node %d\n", 538 pvt->mc_node_id); 539 return 1; 540 } 541 542 /* This node has Memory Hoisting */ 543 544 /* +------------------+--------------------+--------------------+----- 545 * | memory | DRAM hole | relocated | 546 * | [0, (x - 1)] | [x, 0xffffffff] | addresses from | 547 * | | | DRAM hole | 548 * | | | [0x100000000, | 549 * | | | (0x100000000+ | 550 * | | | (0xffffffff-x))] | 551 * +------------------+--------------------+--------------------+----- 552 * 553 * Above is a diagram of physical memory showing the DRAM hole and the 554 * relocated addresses from the DRAM hole. As shown, the DRAM hole 555 * starts at address x (the base address) and extends through address 556 * 0xffffffff. The DRAM Hole Address Register (DHAR) relocates the 557 * addresses in the hole so that they start at 0x100000000. 558 */ 559 560 *hole_base = dhar_base(pvt); 561 *hole_size = (1ULL << 32) - *hole_base; 562 563 *hole_offset = (pvt->fam > 0xf) ? f10_dhar_offset(pvt) 564 : k8_dhar_offset(pvt); 565 566 edac_dbg(1, " DHAR info for node %d base 0x%lx offset 0x%lx size 0x%lx\n", 567 pvt->mc_node_id, (unsigned long)*hole_base, 568 (unsigned long)*hole_offset, (unsigned long)*hole_size); 569 570 return 0; 571 } 572 573 #ifdef CONFIG_EDAC_DEBUG 574 #define EDAC_DCT_ATTR_SHOW(reg) \ 575 static ssize_t reg##_show(struct device *dev, \ 576 struct device_attribute *mattr, char *data) \ 577 { \ 578 struct mem_ctl_info *mci = to_mci(dev); \ 579 struct amd64_pvt *pvt = mci->pvt_info; \ 580 \ 581 return sprintf(data, "0x%016llx\n", (u64)pvt->reg); \ 582 } 583 584 EDAC_DCT_ATTR_SHOW(dhar); 585 EDAC_DCT_ATTR_SHOW(dbam0); 586 EDAC_DCT_ATTR_SHOW(top_mem); 587 EDAC_DCT_ATTR_SHOW(top_mem2); 588 589 static ssize_t dram_hole_show(struct device *dev, struct device_attribute *mattr, 590 char *data) 591 { 592 struct mem_ctl_info *mci = to_mci(dev); 593 594 u64 hole_base = 0; 595 u64 hole_offset = 0; 596 u64 hole_size = 0; 597 598 get_dram_hole_info(mci, &hole_base, &hole_offset, &hole_size); 599 600 return sprintf(data, "%llx %llx %llx\n", hole_base, hole_offset, 601 hole_size); 602 } 603 604 /* 605 * update NUM_DBG_ATTRS in case you add new members 606 */ 607 static DEVICE_ATTR(dhar, S_IRUGO, dhar_show, NULL); 608 static DEVICE_ATTR(dbam, S_IRUGO, dbam0_show, NULL); 609 static DEVICE_ATTR(topmem, S_IRUGO, top_mem_show, NULL); 610 static DEVICE_ATTR(topmem2, S_IRUGO, top_mem2_show, NULL); 611 static DEVICE_ATTR_RO(dram_hole); 612 613 static struct attribute *dbg_attrs[] = { 614 &dev_attr_dhar.attr, 615 &dev_attr_dbam.attr, 616 &dev_attr_topmem.attr, 617 &dev_attr_topmem2.attr, 618 &dev_attr_dram_hole.attr, 619 NULL 620 }; 621 622 static const struct attribute_group dbg_group = { 623 .attrs = dbg_attrs, 624 }; 625 626 static ssize_t inject_section_show(struct device *dev, 627 struct device_attribute *mattr, char *buf) 628 { 629 struct mem_ctl_info *mci = to_mci(dev); 630 struct amd64_pvt *pvt = mci->pvt_info; 631 return sprintf(buf, "0x%x\n", pvt->injection.section); 632 } 633 634 /* 635 * store error injection section value which refers to one of 4 16-byte sections 636 * within a 64-byte cacheline 637 * 638 * range: 0..3 639 */ 640 static ssize_t inject_section_store(struct device *dev, 641 struct device_attribute *mattr, 642 const char *data, size_t count) 643 { 644 struct mem_ctl_info *mci = to_mci(dev); 645 struct amd64_pvt *pvt = mci->pvt_info; 646 unsigned long value; 647 int ret; 648 649 ret = kstrtoul(data, 10, &value); 650 if (ret < 0) 651 return ret; 652 653 if (value > 3) { 654 amd64_warn("%s: invalid section 0x%lx\n", __func__, value); 655 return -EINVAL; 656 } 657 658 pvt->injection.section = (u32) value; 659 return count; 660 } 661 662 static ssize_t inject_word_show(struct device *dev, 663 struct device_attribute *mattr, char *buf) 664 { 665 struct mem_ctl_info *mci = to_mci(dev); 666 struct amd64_pvt *pvt = mci->pvt_info; 667 return sprintf(buf, "0x%x\n", pvt->injection.word); 668 } 669 670 /* 671 * store error injection word value which refers to one of 9 16-bit word of the 672 * 16-byte (128-bit + ECC bits) section 673 * 674 * range: 0..8 675 */ 676 static ssize_t inject_word_store(struct device *dev, 677 struct device_attribute *mattr, 678 const char *data, size_t count) 679 { 680 struct mem_ctl_info *mci = to_mci(dev); 681 struct amd64_pvt *pvt = mci->pvt_info; 682 unsigned long value; 683 int ret; 684 685 ret = kstrtoul(data, 10, &value); 686 if (ret < 0) 687 return ret; 688 689 if (value > 8) { 690 amd64_warn("%s: invalid word 0x%lx\n", __func__, value); 691 return -EINVAL; 692 } 693 694 pvt->injection.word = (u32) value; 695 return count; 696 } 697 698 static ssize_t inject_ecc_vector_show(struct device *dev, 699 struct device_attribute *mattr, 700 char *buf) 701 { 702 struct mem_ctl_info *mci = to_mci(dev); 703 struct amd64_pvt *pvt = mci->pvt_info; 704 return sprintf(buf, "0x%x\n", pvt->injection.bit_map); 705 } 706 707 /* 708 * store 16 bit error injection vector which enables injecting errors to the 709 * corresponding bit within the error injection word above. When used during a 710 * DRAM ECC read, it holds the contents of the of the DRAM ECC bits. 711 */ 712 static ssize_t inject_ecc_vector_store(struct device *dev, 713 struct device_attribute *mattr, 714 const char *data, size_t count) 715 { 716 struct mem_ctl_info *mci = to_mci(dev); 717 struct amd64_pvt *pvt = mci->pvt_info; 718 unsigned long value; 719 int ret; 720 721 ret = kstrtoul(data, 16, &value); 722 if (ret < 0) 723 return ret; 724 725 if (value & 0xFFFF0000) { 726 amd64_warn("%s: invalid EccVector: 0x%lx\n", __func__, value); 727 return -EINVAL; 728 } 729 730 pvt->injection.bit_map = (u32) value; 731 return count; 732 } 733 734 /* 735 * Do a DRAM ECC read. Assemble staged values in the pvt area, format into 736 * fields needed by the injection registers and read the NB Array Data Port. 737 */ 738 static ssize_t inject_read_store(struct device *dev, 739 struct device_attribute *mattr, 740 const char *data, size_t count) 741 { 742 struct mem_ctl_info *mci = to_mci(dev); 743 struct amd64_pvt *pvt = mci->pvt_info; 744 unsigned long value; 745 u32 section, word_bits; 746 int ret; 747 748 ret = kstrtoul(data, 10, &value); 749 if (ret < 0) 750 return ret; 751 752 /* Form value to choose 16-byte section of cacheline */ 753 section = F10_NB_ARRAY_DRAM | SET_NB_ARRAY_ADDR(pvt->injection.section); 754 755 amd64_write_pci_cfg(pvt->F3, F10_NB_ARRAY_ADDR, section); 756 757 word_bits = SET_NB_DRAM_INJECTION_READ(pvt->injection); 758 759 /* Issue 'word' and 'bit' along with the READ request */ 760 amd64_write_pci_cfg(pvt->F3, F10_NB_ARRAY_DATA, word_bits); 761 762 edac_dbg(0, "section=0x%x word_bits=0x%x\n", section, word_bits); 763 764 return count; 765 } 766 767 /* 768 * Do a DRAM ECC write. Assemble staged values in the pvt area and format into 769 * fields needed by the injection registers. 770 */ 771 static ssize_t inject_write_store(struct device *dev, 772 struct device_attribute *mattr, 773 const char *data, size_t count) 774 { 775 struct mem_ctl_info *mci = to_mci(dev); 776 struct amd64_pvt *pvt = mci->pvt_info; 777 u32 section, word_bits, tmp; 778 unsigned long value; 779 int ret; 780 781 ret = kstrtoul(data, 10, &value); 782 if (ret < 0) 783 return ret; 784 785 /* Form value to choose 16-byte section of cacheline */ 786 section = F10_NB_ARRAY_DRAM | SET_NB_ARRAY_ADDR(pvt->injection.section); 787 788 amd64_write_pci_cfg(pvt->F3, F10_NB_ARRAY_ADDR, section); 789 790 word_bits = SET_NB_DRAM_INJECTION_WRITE(pvt->injection); 791 792 pr_notice_once("Don't forget to decrease MCE polling interval in\n" 793 "/sys/bus/machinecheck/devices/machinecheck<CPUNUM>/check_interval\n" 794 "so that you can get the error report faster.\n"); 795 796 on_each_cpu(disable_caches, NULL, 1); 797 798 /* Issue 'word' and 'bit' along with the READ request */ 799 amd64_write_pci_cfg(pvt->F3, F10_NB_ARRAY_DATA, word_bits); 800 801 retry: 802 /* wait until injection happens */ 803 amd64_read_pci_cfg(pvt->F3, F10_NB_ARRAY_DATA, &tmp); 804 if (tmp & F10_NB_ARR_ECC_WR_REQ) { 805 cpu_relax(); 806 goto retry; 807 } 808 809 on_each_cpu(enable_caches, NULL, 1); 810 811 edac_dbg(0, "section=0x%x word_bits=0x%x\n", section, word_bits); 812 813 return count; 814 } 815 816 /* 817 * update NUM_INJ_ATTRS in case you add new members 818 */ 819 820 static DEVICE_ATTR_RW(inject_section); 821 static DEVICE_ATTR_RW(inject_word); 822 static DEVICE_ATTR_RW(inject_ecc_vector); 823 static DEVICE_ATTR_WO(inject_write); 824 static DEVICE_ATTR_WO(inject_read); 825 826 static struct attribute *inj_attrs[] = { 827 &dev_attr_inject_section.attr, 828 &dev_attr_inject_word.attr, 829 &dev_attr_inject_ecc_vector.attr, 830 &dev_attr_inject_write.attr, 831 &dev_attr_inject_read.attr, 832 NULL 833 }; 834 835 static umode_t inj_is_visible(struct kobject *kobj, struct attribute *attr, int idx) 836 { 837 struct device *dev = kobj_to_dev(kobj); 838 struct mem_ctl_info *mci = container_of(dev, struct mem_ctl_info, dev); 839 struct amd64_pvt *pvt = mci->pvt_info; 840 841 /* Families which have that injection hw */ 842 if (pvt->fam >= 0x10 && pvt->fam <= 0x16) 843 return attr->mode; 844 845 return 0; 846 } 847 848 static const struct attribute_group inj_group = { 849 .attrs = inj_attrs, 850 .is_visible = inj_is_visible, 851 }; 852 #endif /* CONFIG_EDAC_DEBUG */ 853 854 /* 855 * Return the DramAddr that the SysAddr given by @sys_addr maps to. It is 856 * assumed that sys_addr maps to the node given by mci. 857 * 858 * The first part of section 3.4.4 (p. 70) shows how the DRAM Base (section 859 * 3.4.4.1) and DRAM Limit (section 3.4.4.2) registers are used to translate a 860 * SysAddr to a DramAddr. If the DRAM Hole Address Register (DHAR) is enabled, 861 * then it is also involved in translating a SysAddr to a DramAddr. Sections 862 * 3.4.8 and 3.5.8.2 describe the DHAR and how it is used for memory hoisting. 863 * These parts of the documentation are unclear. I interpret them as follows: 864 * 865 * When node n receives a SysAddr, it processes the SysAddr as follows: 866 * 867 * 1. It extracts the DRAMBase and DRAMLimit values from the DRAM Base and DRAM 868 * Limit registers for node n. If the SysAddr is not within the range 869 * specified by the base and limit values, then node n ignores the Sysaddr 870 * (since it does not map to node n). Otherwise continue to step 2 below. 871 * 872 * 2. If the DramHoleValid bit of the DHAR for node n is clear, the DHAR is 873 * disabled so skip to step 3 below. Otherwise see if the SysAddr is within 874 * the range of relocated addresses (starting at 0x100000000) from the DRAM 875 * hole. If not, skip to step 3 below. Else get the value of the 876 * DramHoleOffset field from the DHAR. To obtain the DramAddr, subtract the 877 * offset defined by this value from the SysAddr. 878 * 879 * 3. Obtain the base address for node n from the DRAMBase field of the DRAM 880 * Base register for node n. To obtain the DramAddr, subtract the base 881 * address from the SysAddr, as shown near the start of section 3.4.4 (p.70). 882 */ 883 static u64 sys_addr_to_dram_addr(struct mem_ctl_info *mci, u64 sys_addr) 884 { 885 struct amd64_pvt *pvt = mci->pvt_info; 886 u64 dram_base, hole_base, hole_offset, hole_size, dram_addr; 887 int ret; 888 889 dram_base = get_dram_base(pvt, pvt->mc_node_id); 890 891 ret = get_dram_hole_info(mci, &hole_base, &hole_offset, &hole_size); 892 if (!ret) { 893 if ((sys_addr >= (1ULL << 32)) && 894 (sys_addr < ((1ULL << 32) + hole_size))) { 895 /* use DHAR to translate SysAddr to DramAddr */ 896 dram_addr = sys_addr - hole_offset; 897 898 edac_dbg(2, "using DHAR to translate SysAddr 0x%lx to DramAddr 0x%lx\n", 899 (unsigned long)sys_addr, 900 (unsigned long)dram_addr); 901 902 return dram_addr; 903 } 904 } 905 906 /* 907 * Translate the SysAddr to a DramAddr as shown near the start of 908 * section 3.4.4 (p. 70). Although sys_addr is a 64-bit value, the k8 909 * only deals with 40-bit values. Therefore we discard bits 63-40 of 910 * sys_addr below. If bit 39 of sys_addr is 1 then the bits we 911 * discard are all 1s. Otherwise the bits we discard are all 0s. See 912 * section 3.4.2 of AMD publication 24592: AMD x86-64 Architecture 913 * Programmer's Manual Volume 1 Application Programming. 914 */ 915 dram_addr = (sys_addr & GENMASK_ULL(39, 0)) - dram_base; 916 917 edac_dbg(2, "using DRAM Base register to translate SysAddr 0x%lx to DramAddr 0x%lx\n", 918 (unsigned long)sys_addr, (unsigned long)dram_addr); 919 return dram_addr; 920 } 921 922 /* 923 * @intlv_en is the value of the IntlvEn field from a DRAM Base register 924 * (section 3.4.4.1). Return the number of bits from a SysAddr that are used 925 * for node interleaving. 926 */ 927 static int num_node_interleave_bits(unsigned intlv_en) 928 { 929 static const int intlv_shift_table[] = { 0, 1, 0, 2, 0, 0, 0, 3 }; 930 int n; 931 932 BUG_ON(intlv_en > 7); 933 n = intlv_shift_table[intlv_en]; 934 return n; 935 } 936 937 /* Translate the DramAddr given by @dram_addr to an InputAddr. */ 938 static u64 dram_addr_to_input_addr(struct mem_ctl_info *mci, u64 dram_addr) 939 { 940 struct amd64_pvt *pvt; 941 int intlv_shift; 942 u64 input_addr; 943 944 pvt = mci->pvt_info; 945 946 /* 947 * See the start of section 3.4.4 (p. 70, BKDG #26094, K8, revA-E) 948 * concerning translating a DramAddr to an InputAddr. 949 */ 950 intlv_shift = num_node_interleave_bits(dram_intlv_en(pvt, 0)); 951 input_addr = ((dram_addr >> intlv_shift) & GENMASK_ULL(35, 12)) + 952 (dram_addr & 0xfff); 953 954 edac_dbg(2, " Intlv Shift=%d DramAddr=0x%lx maps to InputAddr=0x%lx\n", 955 intlv_shift, (unsigned long)dram_addr, 956 (unsigned long)input_addr); 957 958 return input_addr; 959 } 960 961 /* 962 * Translate the SysAddr represented by @sys_addr to an InputAddr. It is 963 * assumed that @sys_addr maps to the node given by mci. 964 */ 965 static u64 sys_addr_to_input_addr(struct mem_ctl_info *mci, u64 sys_addr) 966 { 967 u64 input_addr; 968 969 input_addr = 970 dram_addr_to_input_addr(mci, sys_addr_to_dram_addr(mci, sys_addr)); 971 972 edac_dbg(2, "SysAddr 0x%lx translates to InputAddr 0x%lx\n", 973 (unsigned long)sys_addr, (unsigned long)input_addr); 974 975 return input_addr; 976 } 977 978 /* Map the Error address to a PAGE and PAGE OFFSET. */ 979 static inline void error_address_to_page_and_offset(u64 error_address, 980 struct err_info *err) 981 { 982 err->page = (u32) (error_address >> PAGE_SHIFT); 983 err->offset = ((u32) error_address) & ~PAGE_MASK; 984 } 985 986 /* 987 * @sys_addr is an error address (a SysAddr) extracted from the MCA NB Address 988 * Low (section 3.6.4.5) and MCA NB Address High (section 3.6.4.6) registers 989 * of a node that detected an ECC memory error. mci represents the node that 990 * the error address maps to (possibly different from the node that detected 991 * the error). Return the number of the csrow that sys_addr maps to, or -1 on 992 * error. 993 */ 994 static int sys_addr_to_csrow(struct mem_ctl_info *mci, u64 sys_addr) 995 { 996 int csrow; 997 998 csrow = input_addr_to_csrow(mci, sys_addr_to_input_addr(mci, sys_addr)); 999 1000 if (csrow == -1) 1001 amd64_mc_err(mci, "Failed to translate InputAddr to csrow for " 1002 "address 0x%lx\n", (unsigned long)sys_addr); 1003 return csrow; 1004 } 1005 1006 /* Protect the PCI config register pairs used for DF indirect access. */ 1007 static DEFINE_MUTEX(df_indirect_mutex); 1008 1009 /* 1010 * Data Fabric Indirect Access uses FICAA/FICAD. 1011 * 1012 * Fabric Indirect Configuration Access Address (FICAA): Constructed based 1013 * on the device's Instance Id and the PCI function and register offset of 1014 * the desired register. 1015 * 1016 * Fabric Indirect Configuration Access Data (FICAD): There are FICAD LO 1017 * and FICAD HI registers but so far we only need the LO register. 1018 * 1019 * Use Instance Id 0xFF to indicate a broadcast read. 1020 */ 1021 #define DF_BROADCAST 0xFF 1022 static int __df_indirect_read(u16 node, u8 func, u16 reg, u8 instance_id, u32 *lo) 1023 { 1024 struct pci_dev *F4; 1025 u32 ficaa; 1026 int err = -ENODEV; 1027 1028 if (node >= amd_nb_num()) 1029 goto out; 1030 1031 F4 = node_to_amd_nb(node)->link; 1032 if (!F4) 1033 goto out; 1034 1035 ficaa = (instance_id == DF_BROADCAST) ? 0 : 1; 1036 ficaa |= reg & 0x3FC; 1037 ficaa |= (func & 0x7) << 11; 1038 ficaa |= instance_id << 16; 1039 1040 mutex_lock(&df_indirect_mutex); 1041 1042 err = pci_write_config_dword(F4, 0x5C, ficaa); 1043 if (err) { 1044 pr_warn("Error writing DF Indirect FICAA, FICAA=0x%x\n", ficaa); 1045 goto out_unlock; 1046 } 1047 1048 err = pci_read_config_dword(F4, 0x98, lo); 1049 if (err) 1050 pr_warn("Error reading DF Indirect FICAD LO, FICAA=0x%x.\n", ficaa); 1051 1052 out_unlock: 1053 mutex_unlock(&df_indirect_mutex); 1054 1055 out: 1056 return err; 1057 } 1058 1059 static int df_indirect_read_instance(u16 node, u8 func, u16 reg, u8 instance_id, u32 *lo) 1060 { 1061 return __df_indirect_read(node, func, reg, instance_id, lo); 1062 } 1063 1064 static int df_indirect_read_broadcast(u16 node, u8 func, u16 reg, u32 *lo) 1065 { 1066 return __df_indirect_read(node, func, reg, DF_BROADCAST, lo); 1067 } 1068 1069 struct addr_ctx { 1070 u64 ret_addr; 1071 u32 tmp; 1072 u16 nid; 1073 u8 inst_id; 1074 }; 1075 1076 static int umc_normaddr_to_sysaddr(u64 norm_addr, u16 nid, u8 umc, u64 *sys_addr) 1077 { 1078 u64 dram_base_addr, dram_limit_addr, dram_hole_base; 1079 1080 u8 die_id_shift, die_id_mask, socket_id_shift, socket_id_mask; 1081 u8 intlv_num_dies, intlv_num_chan, intlv_num_sockets; 1082 u8 intlv_addr_sel, intlv_addr_bit; 1083 u8 num_intlv_bits, hashed_bit; 1084 u8 lgcy_mmio_hole_en, base = 0; 1085 u8 cs_mask, cs_id = 0; 1086 bool hash_enabled = false; 1087 1088 struct addr_ctx ctx; 1089 1090 memset(&ctx, 0, sizeof(ctx)); 1091 1092 /* Start from the normalized address */ 1093 ctx.ret_addr = norm_addr; 1094 1095 ctx.nid = nid; 1096 ctx.inst_id = umc; 1097 1098 /* Read D18F0x1B4 (DramOffset), check if base 1 is used. */ 1099 if (df_indirect_read_instance(nid, 0, 0x1B4, umc, &ctx.tmp)) 1100 goto out_err; 1101 1102 /* Remove HiAddrOffset from normalized address, if enabled: */ 1103 if (ctx.tmp & BIT(0)) { 1104 u64 hi_addr_offset = (ctx.tmp & GENMASK_ULL(31, 20)) << 8; 1105 1106 if (norm_addr >= hi_addr_offset) { 1107 ctx.ret_addr -= hi_addr_offset; 1108 base = 1; 1109 } 1110 } 1111 1112 /* Read D18F0x110 (DramBaseAddress). */ 1113 if (df_indirect_read_instance(nid, 0, 0x110 + (8 * base), umc, &ctx.tmp)) 1114 goto out_err; 1115 1116 /* Check if address range is valid. */ 1117 if (!(ctx.tmp & BIT(0))) { 1118 pr_err("%s: Invalid DramBaseAddress range: 0x%x.\n", 1119 __func__, ctx.tmp); 1120 goto out_err; 1121 } 1122 1123 lgcy_mmio_hole_en = ctx.tmp & BIT(1); 1124 intlv_num_chan = (ctx.tmp >> 4) & 0xF; 1125 intlv_addr_sel = (ctx.tmp >> 8) & 0x7; 1126 dram_base_addr = (ctx.tmp & GENMASK_ULL(31, 12)) << 16; 1127 1128 /* {0, 1, 2, 3} map to address bits {8, 9, 10, 11} respectively */ 1129 if (intlv_addr_sel > 3) { 1130 pr_err("%s: Invalid interleave address select %d.\n", 1131 __func__, intlv_addr_sel); 1132 goto out_err; 1133 } 1134 1135 /* Read D18F0x114 (DramLimitAddress). */ 1136 if (df_indirect_read_instance(nid, 0, 0x114 + (8 * base), umc, &ctx.tmp)) 1137 goto out_err; 1138 1139 intlv_num_sockets = (ctx.tmp >> 8) & 0x1; 1140 intlv_num_dies = (ctx.tmp >> 10) & 0x3; 1141 dram_limit_addr = ((ctx.tmp & GENMASK_ULL(31, 12)) << 16) | GENMASK_ULL(27, 0); 1142 1143 intlv_addr_bit = intlv_addr_sel + 8; 1144 1145 /* Re-use intlv_num_chan by setting it equal to log2(#channels) */ 1146 switch (intlv_num_chan) { 1147 case 0: intlv_num_chan = 0; break; 1148 case 1: intlv_num_chan = 1; break; 1149 case 3: intlv_num_chan = 2; break; 1150 case 5: intlv_num_chan = 3; break; 1151 case 7: intlv_num_chan = 4; break; 1152 1153 case 8: intlv_num_chan = 1; 1154 hash_enabled = true; 1155 break; 1156 default: 1157 pr_err("%s: Invalid number of interleaved channels %d.\n", 1158 __func__, intlv_num_chan); 1159 goto out_err; 1160 } 1161 1162 num_intlv_bits = intlv_num_chan; 1163 1164 if (intlv_num_dies > 2) { 1165 pr_err("%s: Invalid number of interleaved nodes/dies %d.\n", 1166 __func__, intlv_num_dies); 1167 goto out_err; 1168 } 1169 1170 num_intlv_bits += intlv_num_dies; 1171 1172 /* Add a bit if sockets are interleaved. */ 1173 num_intlv_bits += intlv_num_sockets; 1174 1175 /* Assert num_intlv_bits <= 4 */ 1176 if (num_intlv_bits > 4) { 1177 pr_err("%s: Invalid interleave bits %d.\n", 1178 __func__, num_intlv_bits); 1179 goto out_err; 1180 } 1181 1182 if (num_intlv_bits > 0) { 1183 u64 temp_addr_x, temp_addr_i, temp_addr_y; 1184 u8 die_id_bit, sock_id_bit, cs_fabric_id; 1185 1186 /* 1187 * Read FabricBlockInstanceInformation3_CS[BlockFabricID]. 1188 * This is the fabric id for this coherent slave. Use 1189 * umc/channel# as instance id of the coherent slave 1190 * for FICAA. 1191 */ 1192 if (df_indirect_read_instance(nid, 0, 0x50, umc, &ctx.tmp)) 1193 goto out_err; 1194 1195 cs_fabric_id = (ctx.tmp >> 8) & 0xFF; 1196 die_id_bit = 0; 1197 1198 /* If interleaved over more than 1 channel: */ 1199 if (intlv_num_chan) { 1200 die_id_bit = intlv_num_chan; 1201 cs_mask = (1 << die_id_bit) - 1; 1202 cs_id = cs_fabric_id & cs_mask; 1203 } 1204 1205 sock_id_bit = die_id_bit; 1206 1207 /* Read D18F1x208 (SystemFabricIdMask). */ 1208 if (intlv_num_dies || intlv_num_sockets) 1209 if (df_indirect_read_broadcast(nid, 1, 0x208, &ctx.tmp)) 1210 goto out_err; 1211 1212 /* If interleaved over more than 1 die. */ 1213 if (intlv_num_dies) { 1214 sock_id_bit = die_id_bit + intlv_num_dies; 1215 die_id_shift = (ctx.tmp >> 24) & 0xF; 1216 die_id_mask = (ctx.tmp >> 8) & 0xFF; 1217 1218 cs_id |= ((cs_fabric_id & die_id_mask) >> die_id_shift) << die_id_bit; 1219 } 1220 1221 /* If interleaved over more than 1 socket. */ 1222 if (intlv_num_sockets) { 1223 socket_id_shift = (ctx.tmp >> 28) & 0xF; 1224 socket_id_mask = (ctx.tmp >> 16) & 0xFF; 1225 1226 cs_id |= ((cs_fabric_id & socket_id_mask) >> socket_id_shift) << sock_id_bit; 1227 } 1228 1229 /* 1230 * The pre-interleaved address consists of XXXXXXIIIYYYYY 1231 * where III is the ID for this CS, and XXXXXXYYYYY are the 1232 * address bits from the post-interleaved address. 1233 * "num_intlv_bits" has been calculated to tell us how many "I" 1234 * bits there are. "intlv_addr_bit" tells us how many "Y" bits 1235 * there are (where "I" starts). 1236 */ 1237 temp_addr_y = ctx.ret_addr & GENMASK_ULL(intlv_addr_bit - 1, 0); 1238 temp_addr_i = (cs_id << intlv_addr_bit); 1239 temp_addr_x = (ctx.ret_addr & GENMASK_ULL(63, intlv_addr_bit)) << num_intlv_bits; 1240 ctx.ret_addr = temp_addr_x | temp_addr_i | temp_addr_y; 1241 } 1242 1243 /* Add dram base address */ 1244 ctx.ret_addr += dram_base_addr; 1245 1246 /* If legacy MMIO hole enabled */ 1247 if (lgcy_mmio_hole_en) { 1248 if (df_indirect_read_broadcast(nid, 0, 0x104, &ctx.tmp)) 1249 goto out_err; 1250 1251 dram_hole_base = ctx.tmp & GENMASK(31, 24); 1252 if (ctx.ret_addr >= dram_hole_base) 1253 ctx.ret_addr += (BIT_ULL(32) - dram_hole_base); 1254 } 1255 1256 if (hash_enabled) { 1257 /* Save some parentheses and grab ls-bit at the end. */ 1258 hashed_bit = (ctx.ret_addr >> 12) ^ 1259 (ctx.ret_addr >> 18) ^ 1260 (ctx.ret_addr >> 21) ^ 1261 (ctx.ret_addr >> 30) ^ 1262 cs_id; 1263 1264 hashed_bit &= BIT(0); 1265 1266 if (hashed_bit != ((ctx.ret_addr >> intlv_addr_bit) & BIT(0))) 1267 ctx.ret_addr ^= BIT(intlv_addr_bit); 1268 } 1269 1270 /* Is calculated system address is above DRAM limit address? */ 1271 if (ctx.ret_addr > dram_limit_addr) 1272 goto out_err; 1273 1274 *sys_addr = ctx.ret_addr; 1275 return 0; 1276 1277 out_err: 1278 return -EINVAL; 1279 } 1280 1281 static int get_channel_from_ecc_syndrome(struct mem_ctl_info *, u16); 1282 1283 /* 1284 * Determine if the DIMMs have ECC enabled. ECC is enabled ONLY if all the DIMMs 1285 * are ECC capable. 1286 */ 1287 static unsigned long determine_edac_cap(struct amd64_pvt *pvt) 1288 { 1289 unsigned long edac_cap = EDAC_FLAG_NONE; 1290 u8 bit; 1291 1292 if (pvt->umc) { 1293 u8 i, umc_en_mask = 0, dimm_ecc_en_mask = 0; 1294 1295 for_each_umc(i) { 1296 if (!(pvt->umc[i].sdp_ctrl & UMC_SDP_INIT)) 1297 continue; 1298 1299 umc_en_mask |= BIT(i); 1300 1301 /* UMC Configuration bit 12 (DimmEccEn) */ 1302 if (pvt->umc[i].umc_cfg & BIT(12)) 1303 dimm_ecc_en_mask |= BIT(i); 1304 } 1305 1306 if (umc_en_mask == dimm_ecc_en_mask) 1307 edac_cap = EDAC_FLAG_SECDED; 1308 } else { 1309 bit = (pvt->fam > 0xf || pvt->ext_model >= K8_REV_F) 1310 ? 19 1311 : 17; 1312 1313 if (pvt->dclr0 & BIT(bit)) 1314 edac_cap = EDAC_FLAG_SECDED; 1315 } 1316 1317 return edac_cap; 1318 } 1319 1320 static void debug_display_dimm_sizes(struct amd64_pvt *, u8); 1321 1322 static void debug_dump_dramcfg_low(struct amd64_pvt *pvt, u32 dclr, int chan) 1323 { 1324 edac_dbg(1, "F2x%d90 (DRAM Cfg Low): 0x%08x\n", chan, dclr); 1325 1326 if (pvt->dram_type == MEM_LRDDR3) { 1327 u32 dcsm = pvt->csels[chan].csmasks[0]; 1328 /* 1329 * It's assumed all LRDIMMs in a DCT are going to be of 1330 * same 'type' until proven otherwise. So, use a cs 1331 * value of '0' here to get dcsm value. 1332 */ 1333 edac_dbg(1, " LRDIMM %dx rank multiply\n", (dcsm & 0x3)); 1334 } 1335 1336 edac_dbg(1, "All DIMMs support ECC:%s\n", 1337 (dclr & BIT(19)) ? "yes" : "no"); 1338 1339 1340 edac_dbg(1, " PAR/ERR parity: %s\n", 1341 (dclr & BIT(8)) ? "enabled" : "disabled"); 1342 1343 if (pvt->fam == 0x10) 1344 edac_dbg(1, " DCT 128bit mode width: %s\n", 1345 (dclr & BIT(11)) ? "128b" : "64b"); 1346 1347 edac_dbg(1, " x4 logical DIMMs present: L0: %s L1: %s L2: %s L3: %s\n", 1348 (dclr & BIT(12)) ? "yes" : "no", 1349 (dclr & BIT(13)) ? "yes" : "no", 1350 (dclr & BIT(14)) ? "yes" : "no", 1351 (dclr & BIT(15)) ? "yes" : "no"); 1352 } 1353 1354 #define CS_EVEN_PRIMARY BIT(0) 1355 #define CS_ODD_PRIMARY BIT(1) 1356 #define CS_EVEN_SECONDARY BIT(2) 1357 #define CS_ODD_SECONDARY BIT(3) 1358 #define CS_3R_INTERLEAVE BIT(4) 1359 1360 #define CS_EVEN (CS_EVEN_PRIMARY | CS_EVEN_SECONDARY) 1361 #define CS_ODD (CS_ODD_PRIMARY | CS_ODD_SECONDARY) 1362 1363 static int f17_get_cs_mode(int dimm, u8 ctrl, struct amd64_pvt *pvt) 1364 { 1365 u8 base, count = 0; 1366 int cs_mode = 0; 1367 1368 if (csrow_enabled(2 * dimm, ctrl, pvt)) 1369 cs_mode |= CS_EVEN_PRIMARY; 1370 1371 if (csrow_enabled(2 * dimm + 1, ctrl, pvt)) 1372 cs_mode |= CS_ODD_PRIMARY; 1373 1374 /* Asymmetric dual-rank DIMM support. */ 1375 if (csrow_sec_enabled(2 * dimm + 1, ctrl, pvt)) 1376 cs_mode |= CS_ODD_SECONDARY; 1377 1378 /* 1379 * 3 Rank inteleaving support. 1380 * There should be only three bases enabled and their two masks should 1381 * be equal. 1382 */ 1383 for_each_chip_select(base, ctrl, pvt) 1384 count += csrow_enabled(base, ctrl, pvt); 1385 1386 if (count == 3 && 1387 pvt->csels[ctrl].csmasks[0] == pvt->csels[ctrl].csmasks[1]) { 1388 edac_dbg(1, "3R interleaving in use.\n"); 1389 cs_mode |= CS_3R_INTERLEAVE; 1390 } 1391 1392 return cs_mode; 1393 } 1394 1395 static void debug_display_dimm_sizes_df(struct amd64_pvt *pvt, u8 ctrl) 1396 { 1397 int dimm, size0, size1, cs0, cs1, cs_mode; 1398 1399 edac_printk(KERN_DEBUG, EDAC_MC, "UMC%d chip selects:\n", ctrl); 1400 1401 for (dimm = 0; dimm < 2; dimm++) { 1402 cs0 = dimm * 2; 1403 cs1 = dimm * 2 + 1; 1404 1405 cs_mode = f17_get_cs_mode(dimm, ctrl, pvt); 1406 1407 size0 = pvt->ops->dbam_to_cs(pvt, ctrl, cs_mode, cs0); 1408 size1 = pvt->ops->dbam_to_cs(pvt, ctrl, cs_mode, cs1); 1409 1410 amd64_info(EDAC_MC ": %d: %5dMB %d: %5dMB\n", 1411 cs0, size0, 1412 cs1, size1); 1413 } 1414 } 1415 1416 static void __dump_misc_regs_df(struct amd64_pvt *pvt) 1417 { 1418 struct amd64_umc *umc; 1419 u32 i, tmp, umc_base; 1420 1421 for_each_umc(i) { 1422 umc_base = get_umc_base(i); 1423 umc = &pvt->umc[i]; 1424 1425 edac_dbg(1, "UMC%d DIMM cfg: 0x%x\n", i, umc->dimm_cfg); 1426 edac_dbg(1, "UMC%d UMC cfg: 0x%x\n", i, umc->umc_cfg); 1427 edac_dbg(1, "UMC%d SDP ctrl: 0x%x\n", i, umc->sdp_ctrl); 1428 edac_dbg(1, "UMC%d ECC ctrl: 0x%x\n", i, umc->ecc_ctrl); 1429 1430 amd_smn_read(pvt->mc_node_id, umc_base + UMCCH_ECC_BAD_SYMBOL, &tmp); 1431 edac_dbg(1, "UMC%d ECC bad symbol: 0x%x\n", i, tmp); 1432 1433 amd_smn_read(pvt->mc_node_id, umc_base + UMCCH_UMC_CAP, &tmp); 1434 edac_dbg(1, "UMC%d UMC cap: 0x%x\n", i, tmp); 1435 edac_dbg(1, "UMC%d UMC cap high: 0x%x\n", i, umc->umc_cap_hi); 1436 1437 edac_dbg(1, "UMC%d ECC capable: %s, ChipKill ECC capable: %s\n", 1438 i, (umc->umc_cap_hi & BIT(30)) ? "yes" : "no", 1439 (umc->umc_cap_hi & BIT(31)) ? "yes" : "no"); 1440 edac_dbg(1, "UMC%d All DIMMs support ECC: %s\n", 1441 i, (umc->umc_cfg & BIT(12)) ? "yes" : "no"); 1442 edac_dbg(1, "UMC%d x4 DIMMs present: %s\n", 1443 i, (umc->dimm_cfg & BIT(6)) ? "yes" : "no"); 1444 edac_dbg(1, "UMC%d x16 DIMMs present: %s\n", 1445 i, (umc->dimm_cfg & BIT(7)) ? "yes" : "no"); 1446 1447 if (umc->dram_type == MEM_LRDDR4 || umc->dram_type == MEM_LRDDR5) { 1448 amd_smn_read(pvt->mc_node_id, 1449 umc_base + get_umc_reg(UMCCH_ADDR_CFG), 1450 &tmp); 1451 edac_dbg(1, "UMC%d LRDIMM %dx rank multiply\n", 1452 i, 1 << ((tmp >> 4) & 0x3)); 1453 } 1454 1455 debug_display_dimm_sizes_df(pvt, i); 1456 } 1457 1458 edac_dbg(1, "F0x104 (DRAM Hole Address): 0x%08x, base: 0x%08x\n", 1459 pvt->dhar, dhar_base(pvt)); 1460 } 1461 1462 /* Display and decode various NB registers for debug purposes. */ 1463 static void __dump_misc_regs(struct amd64_pvt *pvt) 1464 { 1465 edac_dbg(1, "F3xE8 (NB Cap): 0x%08x\n", pvt->nbcap); 1466 1467 edac_dbg(1, " NB two channel DRAM capable: %s\n", 1468 (pvt->nbcap & NBCAP_DCT_DUAL) ? "yes" : "no"); 1469 1470 edac_dbg(1, " ECC capable: %s, ChipKill ECC capable: %s\n", 1471 (pvt->nbcap & NBCAP_SECDED) ? "yes" : "no", 1472 (pvt->nbcap & NBCAP_CHIPKILL) ? "yes" : "no"); 1473 1474 debug_dump_dramcfg_low(pvt, pvt->dclr0, 0); 1475 1476 edac_dbg(1, "F3xB0 (Online Spare): 0x%08x\n", pvt->online_spare); 1477 1478 edac_dbg(1, "F1xF0 (DRAM Hole Address): 0x%08x, base: 0x%08x, offset: 0x%08x\n", 1479 pvt->dhar, dhar_base(pvt), 1480 (pvt->fam == 0xf) ? k8_dhar_offset(pvt) 1481 : f10_dhar_offset(pvt)); 1482 1483 debug_display_dimm_sizes(pvt, 0); 1484 1485 /* everything below this point is Fam10h and above */ 1486 if (pvt->fam == 0xf) 1487 return; 1488 1489 debug_display_dimm_sizes(pvt, 1); 1490 1491 /* Only if NOT ganged does dclr1 have valid info */ 1492 if (!dct_ganging_enabled(pvt)) 1493 debug_dump_dramcfg_low(pvt, pvt->dclr1, 1); 1494 } 1495 1496 /* Display and decode various NB registers for debug purposes. */ 1497 static void dump_misc_regs(struct amd64_pvt *pvt) 1498 { 1499 if (pvt->umc) 1500 __dump_misc_regs_df(pvt); 1501 else 1502 __dump_misc_regs(pvt); 1503 1504 edac_dbg(1, " DramHoleValid: %s\n", dhar_valid(pvt) ? "yes" : "no"); 1505 1506 amd64_info("using x%u syndromes.\n", pvt->ecc_sym_sz); 1507 } 1508 1509 /* 1510 * See BKDG, F2x[1,0][5C:40], F2[1,0][6C:60] 1511 */ 1512 static void prep_chip_selects(struct amd64_pvt *pvt) 1513 { 1514 if (pvt->fam == 0xf && pvt->ext_model < K8_REV_F) { 1515 pvt->csels[0].b_cnt = pvt->csels[1].b_cnt = 8; 1516 pvt->csels[0].m_cnt = pvt->csels[1].m_cnt = 8; 1517 } else if (pvt->fam == 0x15 && pvt->model == 0x30) { 1518 pvt->csels[0].b_cnt = pvt->csels[1].b_cnt = 4; 1519 pvt->csels[0].m_cnt = pvt->csels[1].m_cnt = 2; 1520 } else if (pvt->fam >= 0x17) { 1521 int umc; 1522 1523 for_each_umc(umc) { 1524 pvt->csels[umc].b_cnt = 4; 1525 pvt->csels[umc].m_cnt = fam_type->flags.zn_regs_v2 ? 4 : 2; 1526 } 1527 1528 } else { 1529 pvt->csels[0].b_cnt = pvt->csels[1].b_cnt = 8; 1530 pvt->csels[0].m_cnt = pvt->csels[1].m_cnt = 4; 1531 } 1532 } 1533 1534 static void read_umc_base_mask(struct amd64_pvt *pvt) 1535 { 1536 u32 umc_base_reg, umc_base_reg_sec; 1537 u32 umc_mask_reg, umc_mask_reg_sec; 1538 u32 base_reg, base_reg_sec; 1539 u32 mask_reg, mask_reg_sec; 1540 u32 *base, *base_sec; 1541 u32 *mask, *mask_sec; 1542 int cs, umc; 1543 1544 for_each_umc(umc) { 1545 umc_base_reg = get_umc_base(umc) + UMCCH_BASE_ADDR; 1546 umc_base_reg_sec = get_umc_base(umc) + UMCCH_BASE_ADDR_SEC; 1547 1548 for_each_chip_select(cs, umc, pvt) { 1549 base = &pvt->csels[umc].csbases[cs]; 1550 base_sec = &pvt->csels[umc].csbases_sec[cs]; 1551 1552 base_reg = umc_base_reg + (cs * 4); 1553 base_reg_sec = umc_base_reg_sec + (cs * 4); 1554 1555 if (!amd_smn_read(pvt->mc_node_id, base_reg, base)) 1556 edac_dbg(0, " DCSB%d[%d]=0x%08x reg: 0x%x\n", 1557 umc, cs, *base, base_reg); 1558 1559 if (!amd_smn_read(pvt->mc_node_id, base_reg_sec, base_sec)) 1560 edac_dbg(0, " DCSB_SEC%d[%d]=0x%08x reg: 0x%x\n", 1561 umc, cs, *base_sec, base_reg_sec); 1562 } 1563 1564 umc_mask_reg = get_umc_base(umc) + UMCCH_ADDR_MASK; 1565 umc_mask_reg_sec = get_umc_base(umc) + get_umc_reg(UMCCH_ADDR_MASK_SEC); 1566 1567 for_each_chip_select_mask(cs, umc, pvt) { 1568 mask = &pvt->csels[umc].csmasks[cs]; 1569 mask_sec = &pvt->csels[umc].csmasks_sec[cs]; 1570 1571 mask_reg = umc_mask_reg + (cs * 4); 1572 mask_reg_sec = umc_mask_reg_sec + (cs * 4); 1573 1574 if (!amd_smn_read(pvt->mc_node_id, mask_reg, mask)) 1575 edac_dbg(0, " DCSM%d[%d]=0x%08x reg: 0x%x\n", 1576 umc, cs, *mask, mask_reg); 1577 1578 if (!amd_smn_read(pvt->mc_node_id, mask_reg_sec, mask_sec)) 1579 edac_dbg(0, " DCSM_SEC%d[%d]=0x%08x reg: 0x%x\n", 1580 umc, cs, *mask_sec, mask_reg_sec); 1581 } 1582 } 1583 } 1584 1585 /* 1586 * Function 2 Offset F10_DCSB0; read in the DCS Base and DCS Mask registers 1587 */ 1588 static void read_dct_base_mask(struct amd64_pvt *pvt) 1589 { 1590 int cs; 1591 1592 prep_chip_selects(pvt); 1593 1594 if (pvt->umc) 1595 return read_umc_base_mask(pvt); 1596 1597 for_each_chip_select(cs, 0, pvt) { 1598 int reg0 = DCSB0 + (cs * 4); 1599 int reg1 = DCSB1 + (cs * 4); 1600 u32 *base0 = &pvt->csels[0].csbases[cs]; 1601 u32 *base1 = &pvt->csels[1].csbases[cs]; 1602 1603 if (!amd64_read_dct_pci_cfg(pvt, 0, reg0, base0)) 1604 edac_dbg(0, " DCSB0[%d]=0x%08x reg: F2x%x\n", 1605 cs, *base0, reg0); 1606 1607 if (pvt->fam == 0xf) 1608 continue; 1609 1610 if (!amd64_read_dct_pci_cfg(pvt, 1, reg0, base1)) 1611 edac_dbg(0, " DCSB1[%d]=0x%08x reg: F2x%x\n", 1612 cs, *base1, (pvt->fam == 0x10) ? reg1 1613 : reg0); 1614 } 1615 1616 for_each_chip_select_mask(cs, 0, pvt) { 1617 int reg0 = DCSM0 + (cs * 4); 1618 int reg1 = DCSM1 + (cs * 4); 1619 u32 *mask0 = &pvt->csels[0].csmasks[cs]; 1620 u32 *mask1 = &pvt->csels[1].csmasks[cs]; 1621 1622 if (!amd64_read_dct_pci_cfg(pvt, 0, reg0, mask0)) 1623 edac_dbg(0, " DCSM0[%d]=0x%08x reg: F2x%x\n", 1624 cs, *mask0, reg0); 1625 1626 if (pvt->fam == 0xf) 1627 continue; 1628 1629 if (!amd64_read_dct_pci_cfg(pvt, 1, reg0, mask1)) 1630 edac_dbg(0, " DCSM1[%d]=0x%08x reg: F2x%x\n", 1631 cs, *mask1, (pvt->fam == 0x10) ? reg1 1632 : reg0); 1633 } 1634 } 1635 1636 static void determine_memory_type_df(struct amd64_pvt *pvt) 1637 { 1638 struct amd64_umc *umc; 1639 u32 i; 1640 1641 for_each_umc(i) { 1642 umc = &pvt->umc[i]; 1643 1644 if (!(umc->sdp_ctrl & UMC_SDP_INIT)) { 1645 umc->dram_type = MEM_EMPTY; 1646 continue; 1647 } 1648 1649 /* 1650 * Check if the system supports the "DDR Type" field in UMC Config 1651 * and has DDR5 DIMMs in use. 1652 */ 1653 if (fam_type->flags.zn_regs_v2 && ((umc->umc_cfg & GENMASK(2, 0)) == 0x1)) { 1654 if (umc->dimm_cfg & BIT(5)) 1655 umc->dram_type = MEM_LRDDR5; 1656 else if (umc->dimm_cfg & BIT(4)) 1657 umc->dram_type = MEM_RDDR5; 1658 else 1659 umc->dram_type = MEM_DDR5; 1660 } else { 1661 if (umc->dimm_cfg & BIT(5)) 1662 umc->dram_type = MEM_LRDDR4; 1663 else if (umc->dimm_cfg & BIT(4)) 1664 umc->dram_type = MEM_RDDR4; 1665 else 1666 umc->dram_type = MEM_DDR4; 1667 } 1668 1669 edac_dbg(1, " UMC%d DIMM type: %s\n", i, edac_mem_types[umc->dram_type]); 1670 } 1671 } 1672 1673 static void determine_memory_type(struct amd64_pvt *pvt) 1674 { 1675 u32 dram_ctrl, dcsm; 1676 1677 if (pvt->umc) 1678 return determine_memory_type_df(pvt); 1679 1680 switch (pvt->fam) { 1681 case 0xf: 1682 if (pvt->ext_model >= K8_REV_F) 1683 goto ddr3; 1684 1685 pvt->dram_type = (pvt->dclr0 & BIT(18)) ? MEM_DDR : MEM_RDDR; 1686 return; 1687 1688 case 0x10: 1689 if (pvt->dchr0 & DDR3_MODE) 1690 goto ddr3; 1691 1692 pvt->dram_type = (pvt->dclr0 & BIT(16)) ? MEM_DDR2 : MEM_RDDR2; 1693 return; 1694 1695 case 0x15: 1696 if (pvt->model < 0x60) 1697 goto ddr3; 1698 1699 /* 1700 * Model 0x60h needs special handling: 1701 * 1702 * We use a Chip Select value of '0' to obtain dcsm. 1703 * Theoretically, it is possible to populate LRDIMMs of different 1704 * 'Rank' value on a DCT. But this is not the common case. So, 1705 * it's reasonable to assume all DIMMs are going to be of same 1706 * 'type' until proven otherwise. 1707 */ 1708 amd64_read_dct_pci_cfg(pvt, 0, DRAM_CONTROL, &dram_ctrl); 1709 dcsm = pvt->csels[0].csmasks[0]; 1710 1711 if (((dram_ctrl >> 8) & 0x7) == 0x2) 1712 pvt->dram_type = MEM_DDR4; 1713 else if (pvt->dclr0 & BIT(16)) 1714 pvt->dram_type = MEM_DDR3; 1715 else if (dcsm & 0x3) 1716 pvt->dram_type = MEM_LRDDR3; 1717 else 1718 pvt->dram_type = MEM_RDDR3; 1719 1720 return; 1721 1722 case 0x16: 1723 goto ddr3; 1724 1725 default: 1726 WARN(1, KERN_ERR "%s: Family??? 0x%x\n", __func__, pvt->fam); 1727 pvt->dram_type = MEM_EMPTY; 1728 } 1729 return; 1730 1731 ddr3: 1732 pvt->dram_type = (pvt->dclr0 & BIT(16)) ? MEM_DDR3 : MEM_RDDR3; 1733 } 1734 1735 /* Get the number of DCT channels the memory controller is using. */ 1736 static int k8_early_channel_count(struct amd64_pvt *pvt) 1737 { 1738 int flag; 1739 1740 if (pvt->ext_model >= K8_REV_F) 1741 /* RevF (NPT) and later */ 1742 flag = pvt->dclr0 & WIDTH_128; 1743 else 1744 /* RevE and earlier */ 1745 flag = pvt->dclr0 & REVE_WIDTH_128; 1746 1747 /* not used */ 1748 pvt->dclr1 = 0; 1749 1750 return (flag) ? 2 : 1; 1751 } 1752 1753 /* On F10h and later ErrAddr is MC4_ADDR[47:1] */ 1754 static u64 get_error_address(struct amd64_pvt *pvt, struct mce *m) 1755 { 1756 u16 mce_nid = topology_die_id(m->extcpu); 1757 struct mem_ctl_info *mci; 1758 u8 start_bit = 1; 1759 u8 end_bit = 47; 1760 u64 addr; 1761 1762 mci = edac_mc_find(mce_nid); 1763 if (!mci) 1764 return 0; 1765 1766 pvt = mci->pvt_info; 1767 1768 if (pvt->fam == 0xf) { 1769 start_bit = 3; 1770 end_bit = 39; 1771 } 1772 1773 addr = m->addr & GENMASK_ULL(end_bit, start_bit); 1774 1775 /* 1776 * Erratum 637 workaround 1777 */ 1778 if (pvt->fam == 0x15) { 1779 u64 cc6_base, tmp_addr; 1780 u32 tmp; 1781 u8 intlv_en; 1782 1783 if ((addr & GENMASK_ULL(47, 24)) >> 24 != 0x00fdf7) 1784 return addr; 1785 1786 1787 amd64_read_pci_cfg(pvt->F1, DRAM_LOCAL_NODE_LIM, &tmp); 1788 intlv_en = tmp >> 21 & 0x7; 1789 1790 /* add [47:27] + 3 trailing bits */ 1791 cc6_base = (tmp & GENMASK_ULL(20, 0)) << 3; 1792 1793 /* reverse and add DramIntlvEn */ 1794 cc6_base |= intlv_en ^ 0x7; 1795 1796 /* pin at [47:24] */ 1797 cc6_base <<= 24; 1798 1799 if (!intlv_en) 1800 return cc6_base | (addr & GENMASK_ULL(23, 0)); 1801 1802 amd64_read_pci_cfg(pvt->F1, DRAM_LOCAL_NODE_BASE, &tmp); 1803 1804 /* faster log2 */ 1805 tmp_addr = (addr & GENMASK_ULL(23, 12)) << __fls(intlv_en + 1); 1806 1807 /* OR DramIntlvSel into bits [14:12] */ 1808 tmp_addr |= (tmp & GENMASK_ULL(23, 21)) >> 9; 1809 1810 /* add remaining [11:0] bits from original MC4_ADDR */ 1811 tmp_addr |= addr & GENMASK_ULL(11, 0); 1812 1813 return cc6_base | tmp_addr; 1814 } 1815 1816 return addr; 1817 } 1818 1819 static struct pci_dev *pci_get_related_function(unsigned int vendor, 1820 unsigned int device, 1821 struct pci_dev *related) 1822 { 1823 struct pci_dev *dev = NULL; 1824 1825 while ((dev = pci_get_device(vendor, device, dev))) { 1826 if (pci_domain_nr(dev->bus) == pci_domain_nr(related->bus) && 1827 (dev->bus->number == related->bus->number) && 1828 (PCI_SLOT(dev->devfn) == PCI_SLOT(related->devfn))) 1829 break; 1830 } 1831 1832 return dev; 1833 } 1834 1835 static void read_dram_base_limit_regs(struct amd64_pvt *pvt, unsigned range) 1836 { 1837 struct amd_northbridge *nb; 1838 struct pci_dev *f1 = NULL; 1839 unsigned int pci_func; 1840 int off = range << 3; 1841 u32 llim; 1842 1843 amd64_read_pci_cfg(pvt->F1, DRAM_BASE_LO + off, &pvt->ranges[range].base.lo); 1844 amd64_read_pci_cfg(pvt->F1, DRAM_LIMIT_LO + off, &pvt->ranges[range].lim.lo); 1845 1846 if (pvt->fam == 0xf) 1847 return; 1848 1849 if (!dram_rw(pvt, range)) 1850 return; 1851 1852 amd64_read_pci_cfg(pvt->F1, DRAM_BASE_HI + off, &pvt->ranges[range].base.hi); 1853 amd64_read_pci_cfg(pvt->F1, DRAM_LIMIT_HI + off, &pvt->ranges[range].lim.hi); 1854 1855 /* F15h: factor in CC6 save area by reading dst node's limit reg */ 1856 if (pvt->fam != 0x15) 1857 return; 1858 1859 nb = node_to_amd_nb(dram_dst_node(pvt, range)); 1860 if (WARN_ON(!nb)) 1861 return; 1862 1863 if (pvt->model == 0x60) 1864 pci_func = PCI_DEVICE_ID_AMD_15H_M60H_NB_F1; 1865 else if (pvt->model == 0x30) 1866 pci_func = PCI_DEVICE_ID_AMD_15H_M30H_NB_F1; 1867 else 1868 pci_func = PCI_DEVICE_ID_AMD_15H_NB_F1; 1869 1870 f1 = pci_get_related_function(nb->misc->vendor, pci_func, nb->misc); 1871 if (WARN_ON(!f1)) 1872 return; 1873 1874 amd64_read_pci_cfg(f1, DRAM_LOCAL_NODE_LIM, &llim); 1875 1876 pvt->ranges[range].lim.lo &= GENMASK_ULL(15, 0); 1877 1878 /* {[39:27],111b} */ 1879 pvt->ranges[range].lim.lo |= ((llim & 0x1fff) << 3 | 0x7) << 16; 1880 1881 pvt->ranges[range].lim.hi &= GENMASK_ULL(7, 0); 1882 1883 /* [47:40] */ 1884 pvt->ranges[range].lim.hi |= llim >> 13; 1885 1886 pci_dev_put(f1); 1887 } 1888 1889 static void k8_map_sysaddr_to_csrow(struct mem_ctl_info *mci, u64 sys_addr, 1890 struct err_info *err) 1891 { 1892 struct amd64_pvt *pvt = mci->pvt_info; 1893 1894 error_address_to_page_and_offset(sys_addr, err); 1895 1896 /* 1897 * Find out which node the error address belongs to. This may be 1898 * different from the node that detected the error. 1899 */ 1900 err->src_mci = find_mc_by_sys_addr(mci, sys_addr); 1901 if (!err->src_mci) { 1902 amd64_mc_err(mci, "failed to map error addr 0x%lx to a node\n", 1903 (unsigned long)sys_addr); 1904 err->err_code = ERR_NODE; 1905 return; 1906 } 1907 1908 /* Now map the sys_addr to a CSROW */ 1909 err->csrow = sys_addr_to_csrow(err->src_mci, sys_addr); 1910 if (err->csrow < 0) { 1911 err->err_code = ERR_CSROW; 1912 return; 1913 } 1914 1915 /* CHIPKILL enabled */ 1916 if (pvt->nbcfg & NBCFG_CHIPKILL) { 1917 err->channel = get_channel_from_ecc_syndrome(mci, err->syndrome); 1918 if (err->channel < 0) { 1919 /* 1920 * Syndrome didn't map, so we don't know which of the 1921 * 2 DIMMs is in error. So we need to ID 'both' of them 1922 * as suspect. 1923 */ 1924 amd64_mc_warn(err->src_mci, "unknown syndrome 0x%04x - " 1925 "possible error reporting race\n", 1926 err->syndrome); 1927 err->err_code = ERR_CHANNEL; 1928 return; 1929 } 1930 } else { 1931 /* 1932 * non-chipkill ecc mode 1933 * 1934 * The k8 documentation is unclear about how to determine the 1935 * channel number when using non-chipkill memory. This method 1936 * was obtained from email communication with someone at AMD. 1937 * (Wish the email was placed in this comment - norsk) 1938 */ 1939 err->channel = ((sys_addr & BIT(3)) != 0); 1940 } 1941 } 1942 1943 static int ddr2_cs_size(unsigned i, bool dct_width) 1944 { 1945 unsigned shift = 0; 1946 1947 if (i <= 2) 1948 shift = i; 1949 else if (!(i & 0x1)) 1950 shift = i >> 1; 1951 else 1952 shift = (i + 1) >> 1; 1953 1954 return 128 << (shift + !!dct_width); 1955 } 1956 1957 static int k8_dbam_to_chip_select(struct amd64_pvt *pvt, u8 dct, 1958 unsigned cs_mode, int cs_mask_nr) 1959 { 1960 u32 dclr = dct ? pvt->dclr1 : pvt->dclr0; 1961 1962 if (pvt->ext_model >= K8_REV_F) { 1963 WARN_ON(cs_mode > 11); 1964 return ddr2_cs_size(cs_mode, dclr & WIDTH_128); 1965 } 1966 else if (pvt->ext_model >= K8_REV_D) { 1967 unsigned diff; 1968 WARN_ON(cs_mode > 10); 1969 1970 /* 1971 * the below calculation, besides trying to win an obfuscated C 1972 * contest, maps cs_mode values to DIMM chip select sizes. The 1973 * mappings are: 1974 * 1975 * cs_mode CS size (mb) 1976 * ======= ============ 1977 * 0 32 1978 * 1 64 1979 * 2 128 1980 * 3 128 1981 * 4 256 1982 * 5 512 1983 * 6 256 1984 * 7 512 1985 * 8 1024 1986 * 9 1024 1987 * 10 2048 1988 * 1989 * Basically, it calculates a value with which to shift the 1990 * smallest CS size of 32MB. 1991 * 1992 * ddr[23]_cs_size have a similar purpose. 1993 */ 1994 diff = cs_mode/3 + (unsigned)(cs_mode > 5); 1995 1996 return 32 << (cs_mode - diff); 1997 } 1998 else { 1999 WARN_ON(cs_mode > 6); 2000 return 32 << cs_mode; 2001 } 2002 } 2003 2004 /* 2005 * Get the number of DCT channels in use. 2006 * 2007 * Return: 2008 * number of Memory Channels in operation 2009 * Pass back: 2010 * contents of the DCL0_LOW register 2011 */ 2012 static int f1x_early_channel_count(struct amd64_pvt *pvt) 2013 { 2014 int i, j, channels = 0; 2015 2016 /* On F10h, if we are in 128 bit mode, then we are using 2 channels */ 2017 if (pvt->fam == 0x10 && (pvt->dclr0 & WIDTH_128)) 2018 return 2; 2019 2020 /* 2021 * Need to check if in unganged mode: In such, there are 2 channels, 2022 * but they are not in 128 bit mode and thus the above 'dclr0' status 2023 * bit will be OFF. 2024 * 2025 * Need to check DCT0[0] and DCT1[0] to see if only one of them has 2026 * their CSEnable bit on. If so, then SINGLE DIMM case. 2027 */ 2028 edac_dbg(0, "Data width is not 128 bits - need more decoding\n"); 2029 2030 /* 2031 * Check DRAM Bank Address Mapping values for each DIMM to see if there 2032 * is more than just one DIMM present in unganged mode. Need to check 2033 * both controllers since DIMMs can be placed in either one. 2034 */ 2035 for (i = 0; i < 2; i++) { 2036 u32 dbam = (i ? pvt->dbam1 : pvt->dbam0); 2037 2038 for (j = 0; j < 4; j++) { 2039 if (DBAM_DIMM(j, dbam) > 0) { 2040 channels++; 2041 break; 2042 } 2043 } 2044 } 2045 2046 if (channels > 2) 2047 channels = 2; 2048 2049 amd64_info("MCT channel count: %d\n", channels); 2050 2051 return channels; 2052 } 2053 2054 static int f17_early_channel_count(struct amd64_pvt *pvt) 2055 { 2056 int i, channels = 0; 2057 2058 /* SDP Control bit 31 (SdpInit) is clear for unused UMC channels */ 2059 for_each_umc(i) 2060 channels += !!(pvt->umc[i].sdp_ctrl & UMC_SDP_INIT); 2061 2062 amd64_info("MCT channel count: %d\n", channels); 2063 2064 return channels; 2065 } 2066 2067 static int ddr3_cs_size(unsigned i, bool dct_width) 2068 { 2069 unsigned shift = 0; 2070 int cs_size = 0; 2071 2072 if (i == 0 || i == 3 || i == 4) 2073 cs_size = -1; 2074 else if (i <= 2) 2075 shift = i; 2076 else if (i == 12) 2077 shift = 7; 2078 else if (!(i & 0x1)) 2079 shift = i >> 1; 2080 else 2081 shift = (i + 1) >> 1; 2082 2083 if (cs_size != -1) 2084 cs_size = (128 * (1 << !!dct_width)) << shift; 2085 2086 return cs_size; 2087 } 2088 2089 static int ddr3_lrdimm_cs_size(unsigned i, unsigned rank_multiply) 2090 { 2091 unsigned shift = 0; 2092 int cs_size = 0; 2093 2094 if (i < 4 || i == 6) 2095 cs_size = -1; 2096 else if (i == 12) 2097 shift = 7; 2098 else if (!(i & 0x1)) 2099 shift = i >> 1; 2100 else 2101 shift = (i + 1) >> 1; 2102 2103 if (cs_size != -1) 2104 cs_size = rank_multiply * (128 << shift); 2105 2106 return cs_size; 2107 } 2108 2109 static int ddr4_cs_size(unsigned i) 2110 { 2111 int cs_size = 0; 2112 2113 if (i == 0) 2114 cs_size = -1; 2115 else if (i == 1) 2116 cs_size = 1024; 2117 else 2118 /* Min cs_size = 1G */ 2119 cs_size = 1024 * (1 << (i >> 1)); 2120 2121 return cs_size; 2122 } 2123 2124 static int f10_dbam_to_chip_select(struct amd64_pvt *pvt, u8 dct, 2125 unsigned cs_mode, int cs_mask_nr) 2126 { 2127 u32 dclr = dct ? pvt->dclr1 : pvt->dclr0; 2128 2129 WARN_ON(cs_mode > 11); 2130 2131 if (pvt->dchr0 & DDR3_MODE || pvt->dchr1 & DDR3_MODE) 2132 return ddr3_cs_size(cs_mode, dclr & WIDTH_128); 2133 else 2134 return ddr2_cs_size(cs_mode, dclr & WIDTH_128); 2135 } 2136 2137 /* 2138 * F15h supports only 64bit DCT interfaces 2139 */ 2140 static int f15_dbam_to_chip_select(struct amd64_pvt *pvt, u8 dct, 2141 unsigned cs_mode, int cs_mask_nr) 2142 { 2143 WARN_ON(cs_mode > 12); 2144 2145 return ddr3_cs_size(cs_mode, false); 2146 } 2147 2148 /* F15h M60h supports DDR4 mapping as well.. */ 2149 static int f15_m60h_dbam_to_chip_select(struct amd64_pvt *pvt, u8 dct, 2150 unsigned cs_mode, int cs_mask_nr) 2151 { 2152 int cs_size; 2153 u32 dcsm = pvt->csels[dct].csmasks[cs_mask_nr]; 2154 2155 WARN_ON(cs_mode > 12); 2156 2157 if (pvt->dram_type == MEM_DDR4) { 2158 if (cs_mode > 9) 2159 return -1; 2160 2161 cs_size = ddr4_cs_size(cs_mode); 2162 } else if (pvt->dram_type == MEM_LRDDR3) { 2163 unsigned rank_multiply = dcsm & 0xf; 2164 2165 if (rank_multiply == 3) 2166 rank_multiply = 4; 2167 cs_size = ddr3_lrdimm_cs_size(cs_mode, rank_multiply); 2168 } else { 2169 /* Minimum cs size is 512mb for F15hM60h*/ 2170 if (cs_mode == 0x1) 2171 return -1; 2172 2173 cs_size = ddr3_cs_size(cs_mode, false); 2174 } 2175 2176 return cs_size; 2177 } 2178 2179 /* 2180 * F16h and F15h model 30h have only limited cs_modes. 2181 */ 2182 static int f16_dbam_to_chip_select(struct amd64_pvt *pvt, u8 dct, 2183 unsigned cs_mode, int cs_mask_nr) 2184 { 2185 WARN_ON(cs_mode > 12); 2186 2187 if (cs_mode == 6 || cs_mode == 8 || 2188 cs_mode == 9 || cs_mode == 12) 2189 return -1; 2190 else 2191 return ddr3_cs_size(cs_mode, false); 2192 } 2193 2194 static int f17_addr_mask_to_cs_size(struct amd64_pvt *pvt, u8 umc, 2195 unsigned int cs_mode, int csrow_nr) 2196 { 2197 u32 addr_mask_orig, addr_mask_deinterleaved; 2198 u32 msb, weight, num_zero_bits; 2199 int cs_mask_nr = csrow_nr; 2200 int dimm, size = 0; 2201 2202 /* No Chip Selects are enabled. */ 2203 if (!cs_mode) 2204 return size; 2205 2206 /* Requested size of an even CS but none are enabled. */ 2207 if (!(cs_mode & CS_EVEN) && !(csrow_nr & 1)) 2208 return size; 2209 2210 /* Requested size of an odd CS but none are enabled. */ 2211 if (!(cs_mode & CS_ODD) && (csrow_nr & 1)) 2212 return size; 2213 2214 /* 2215 * Family 17h introduced systems with one mask per DIMM, 2216 * and two Chip Selects per DIMM. 2217 * 2218 * CS0 and CS1 -> MASK0 / DIMM0 2219 * CS2 and CS3 -> MASK1 / DIMM1 2220 * 2221 * Family 19h Model 10h introduced systems with one mask per Chip Select, 2222 * and two Chip Selects per DIMM. 2223 * 2224 * CS0 -> MASK0 -> DIMM0 2225 * CS1 -> MASK1 -> DIMM0 2226 * CS2 -> MASK2 -> DIMM1 2227 * CS3 -> MASK3 -> DIMM1 2228 * 2229 * Keep the mask number equal to the Chip Select number for newer systems, 2230 * and shift the mask number for older systems. 2231 */ 2232 dimm = csrow_nr >> 1; 2233 2234 if (!fam_type->flags.zn_regs_v2) 2235 cs_mask_nr >>= 1; 2236 2237 /* Asymmetric dual-rank DIMM support. */ 2238 if ((csrow_nr & 1) && (cs_mode & CS_ODD_SECONDARY)) 2239 addr_mask_orig = pvt->csels[umc].csmasks_sec[cs_mask_nr]; 2240 else 2241 addr_mask_orig = pvt->csels[umc].csmasks[cs_mask_nr]; 2242 2243 /* 2244 * The number of zero bits in the mask is equal to the number of bits 2245 * in a full mask minus the number of bits in the current mask. 2246 * 2247 * The MSB is the number of bits in the full mask because BIT[0] is 2248 * always 0. 2249 * 2250 * In the special 3 Rank interleaving case, a single bit is flipped 2251 * without swapping with the most significant bit. This can be handled 2252 * by keeping the MSB where it is and ignoring the single zero bit. 2253 */ 2254 msb = fls(addr_mask_orig) - 1; 2255 weight = hweight_long(addr_mask_orig); 2256 num_zero_bits = msb - weight - !!(cs_mode & CS_3R_INTERLEAVE); 2257 2258 /* Take the number of zero bits off from the top of the mask. */ 2259 addr_mask_deinterleaved = GENMASK_ULL(msb - num_zero_bits, 1); 2260 2261 edac_dbg(1, "CS%d DIMM%d AddrMasks:\n", csrow_nr, dimm); 2262 edac_dbg(1, " Original AddrMask: 0x%x\n", addr_mask_orig); 2263 edac_dbg(1, " Deinterleaved AddrMask: 0x%x\n", addr_mask_deinterleaved); 2264 2265 /* Register [31:1] = Address [39:9]. Size is in kBs here. */ 2266 size = (addr_mask_deinterleaved >> 2) + 1; 2267 2268 /* Return size in MBs. */ 2269 return size >> 10; 2270 } 2271 2272 static void read_dram_ctl_register(struct amd64_pvt *pvt) 2273 { 2274 2275 if (pvt->fam == 0xf) 2276 return; 2277 2278 if (!amd64_read_pci_cfg(pvt->F2, DCT_SEL_LO, &pvt->dct_sel_lo)) { 2279 edac_dbg(0, "F2x110 (DCTSelLow): 0x%08x, High range addrs at: 0x%x\n", 2280 pvt->dct_sel_lo, dct_sel_baseaddr(pvt)); 2281 2282 edac_dbg(0, " DCTs operate in %s mode\n", 2283 (dct_ganging_enabled(pvt) ? "ganged" : "unganged")); 2284 2285 if (!dct_ganging_enabled(pvt)) 2286 edac_dbg(0, " Address range split per DCT: %s\n", 2287 (dct_high_range_enabled(pvt) ? "yes" : "no")); 2288 2289 edac_dbg(0, " data interleave for ECC: %s, DRAM cleared since last warm reset: %s\n", 2290 (dct_data_intlv_enabled(pvt) ? "enabled" : "disabled"), 2291 (dct_memory_cleared(pvt) ? "yes" : "no")); 2292 2293 edac_dbg(0, " channel interleave: %s, " 2294 "interleave bits selector: 0x%x\n", 2295 (dct_interleave_enabled(pvt) ? "enabled" : "disabled"), 2296 dct_sel_interleave_addr(pvt)); 2297 } 2298 2299 amd64_read_pci_cfg(pvt->F2, DCT_SEL_HI, &pvt->dct_sel_hi); 2300 } 2301 2302 /* 2303 * Determine channel (DCT) based on the interleaving mode (see F15h M30h BKDG, 2304 * 2.10.12 Memory Interleaving Modes). 2305 */ 2306 static u8 f15_m30h_determine_channel(struct amd64_pvt *pvt, u64 sys_addr, 2307 u8 intlv_en, int num_dcts_intlv, 2308 u32 dct_sel) 2309 { 2310 u8 channel = 0; 2311 u8 select; 2312 2313 if (!(intlv_en)) 2314 return (u8)(dct_sel); 2315 2316 if (num_dcts_intlv == 2) { 2317 select = (sys_addr >> 8) & 0x3; 2318 channel = select ? 0x3 : 0; 2319 } else if (num_dcts_intlv == 4) { 2320 u8 intlv_addr = dct_sel_interleave_addr(pvt); 2321 switch (intlv_addr) { 2322 case 0x4: 2323 channel = (sys_addr >> 8) & 0x3; 2324 break; 2325 case 0x5: 2326 channel = (sys_addr >> 9) & 0x3; 2327 break; 2328 } 2329 } 2330 return channel; 2331 } 2332 2333 /* 2334 * Determine channel (DCT) based on the interleaving mode: F10h BKDG, 2.8.9 Memory 2335 * Interleaving Modes. 2336 */ 2337 static u8 f1x_determine_channel(struct amd64_pvt *pvt, u64 sys_addr, 2338 bool hi_range_sel, u8 intlv_en) 2339 { 2340 u8 dct_sel_high = (pvt->dct_sel_lo >> 1) & 1; 2341 2342 if (dct_ganging_enabled(pvt)) 2343 return 0; 2344 2345 if (hi_range_sel) 2346 return dct_sel_high; 2347 2348 /* 2349 * see F2x110[DctSelIntLvAddr] - channel interleave mode 2350 */ 2351 if (dct_interleave_enabled(pvt)) { 2352 u8 intlv_addr = dct_sel_interleave_addr(pvt); 2353 2354 /* return DCT select function: 0=DCT0, 1=DCT1 */ 2355 if (!intlv_addr) 2356 return sys_addr >> 6 & 1; 2357 2358 if (intlv_addr & 0x2) { 2359 u8 shift = intlv_addr & 0x1 ? 9 : 6; 2360 u32 temp = hweight_long((u32) ((sys_addr >> 16) & 0x1F)) & 1; 2361 2362 return ((sys_addr >> shift) & 1) ^ temp; 2363 } 2364 2365 if (intlv_addr & 0x4) { 2366 u8 shift = intlv_addr & 0x1 ? 9 : 8; 2367 2368 return (sys_addr >> shift) & 1; 2369 } 2370 2371 return (sys_addr >> (12 + hweight8(intlv_en))) & 1; 2372 } 2373 2374 if (dct_high_range_enabled(pvt)) 2375 return ~dct_sel_high & 1; 2376 2377 return 0; 2378 } 2379 2380 /* Convert the sys_addr to the normalized DCT address */ 2381 static u64 f1x_get_norm_dct_addr(struct amd64_pvt *pvt, u8 range, 2382 u64 sys_addr, bool hi_rng, 2383 u32 dct_sel_base_addr) 2384 { 2385 u64 chan_off; 2386 u64 dram_base = get_dram_base(pvt, range); 2387 u64 hole_off = f10_dhar_offset(pvt); 2388 u64 dct_sel_base_off = (u64)(pvt->dct_sel_hi & 0xFFFFFC00) << 16; 2389 2390 if (hi_rng) { 2391 /* 2392 * if 2393 * base address of high range is below 4Gb 2394 * (bits [47:27] at [31:11]) 2395 * DRAM address space on this DCT is hoisted above 4Gb && 2396 * sys_addr > 4Gb 2397 * 2398 * remove hole offset from sys_addr 2399 * else 2400 * remove high range offset from sys_addr 2401 */ 2402 if ((!(dct_sel_base_addr >> 16) || 2403 dct_sel_base_addr < dhar_base(pvt)) && 2404 dhar_valid(pvt) && 2405 (sys_addr >= BIT_64(32))) 2406 chan_off = hole_off; 2407 else 2408 chan_off = dct_sel_base_off; 2409 } else { 2410 /* 2411 * if 2412 * we have a valid hole && 2413 * sys_addr > 4Gb 2414 * 2415 * remove hole 2416 * else 2417 * remove dram base to normalize to DCT address 2418 */ 2419 if (dhar_valid(pvt) && (sys_addr >= BIT_64(32))) 2420 chan_off = hole_off; 2421 else 2422 chan_off = dram_base; 2423 } 2424 2425 return (sys_addr & GENMASK_ULL(47,6)) - (chan_off & GENMASK_ULL(47,23)); 2426 } 2427 2428 /* 2429 * checks if the csrow passed in is marked as SPARED, if so returns the new 2430 * spare row 2431 */ 2432 static int f10_process_possible_spare(struct amd64_pvt *pvt, u8 dct, int csrow) 2433 { 2434 int tmp_cs; 2435 2436 if (online_spare_swap_done(pvt, dct) && 2437 csrow == online_spare_bad_dramcs(pvt, dct)) { 2438 2439 for_each_chip_select(tmp_cs, dct, pvt) { 2440 if (chip_select_base(tmp_cs, dct, pvt) & 0x2) { 2441 csrow = tmp_cs; 2442 break; 2443 } 2444 } 2445 } 2446 return csrow; 2447 } 2448 2449 /* 2450 * Iterate over the DRAM DCT "base" and "mask" registers looking for a 2451 * SystemAddr match on the specified 'ChannelSelect' and 'NodeID' 2452 * 2453 * Return: 2454 * -EINVAL: NOT FOUND 2455 * 0..csrow = Chip-Select Row 2456 */ 2457 static int f1x_lookup_addr_in_dct(u64 in_addr, u8 nid, u8 dct) 2458 { 2459 struct mem_ctl_info *mci; 2460 struct amd64_pvt *pvt; 2461 u64 cs_base, cs_mask; 2462 int cs_found = -EINVAL; 2463 int csrow; 2464 2465 mci = edac_mc_find(nid); 2466 if (!mci) 2467 return cs_found; 2468 2469 pvt = mci->pvt_info; 2470 2471 edac_dbg(1, "input addr: 0x%llx, DCT: %d\n", in_addr, dct); 2472 2473 for_each_chip_select(csrow, dct, pvt) { 2474 if (!csrow_enabled(csrow, dct, pvt)) 2475 continue; 2476 2477 get_cs_base_and_mask(pvt, csrow, dct, &cs_base, &cs_mask); 2478 2479 edac_dbg(1, " CSROW=%d CSBase=0x%llx CSMask=0x%llx\n", 2480 csrow, cs_base, cs_mask); 2481 2482 cs_mask = ~cs_mask; 2483 2484 edac_dbg(1, " (InputAddr & ~CSMask)=0x%llx (CSBase & ~CSMask)=0x%llx\n", 2485 (in_addr & cs_mask), (cs_base & cs_mask)); 2486 2487 if ((in_addr & cs_mask) == (cs_base & cs_mask)) { 2488 if (pvt->fam == 0x15 && pvt->model >= 0x30) { 2489 cs_found = csrow; 2490 break; 2491 } 2492 cs_found = f10_process_possible_spare(pvt, dct, csrow); 2493 2494 edac_dbg(1, " MATCH csrow=%d\n", cs_found); 2495 break; 2496 } 2497 } 2498 return cs_found; 2499 } 2500 2501 /* 2502 * See F2x10C. Non-interleaved graphics framebuffer memory under the 16G is 2503 * swapped with a region located at the bottom of memory so that the GPU can use 2504 * the interleaved region and thus two channels. 2505 */ 2506 static u64 f1x_swap_interleaved_region(struct amd64_pvt *pvt, u64 sys_addr) 2507 { 2508 u32 swap_reg, swap_base, swap_limit, rgn_size, tmp_addr; 2509 2510 if (pvt->fam == 0x10) { 2511 /* only revC3 and revE have that feature */ 2512 if (pvt->model < 4 || (pvt->model < 0xa && pvt->stepping < 3)) 2513 return sys_addr; 2514 } 2515 2516 amd64_read_pci_cfg(pvt->F2, SWAP_INTLV_REG, &swap_reg); 2517 2518 if (!(swap_reg & 0x1)) 2519 return sys_addr; 2520 2521 swap_base = (swap_reg >> 3) & 0x7f; 2522 swap_limit = (swap_reg >> 11) & 0x7f; 2523 rgn_size = (swap_reg >> 20) & 0x7f; 2524 tmp_addr = sys_addr >> 27; 2525 2526 if (!(sys_addr >> 34) && 2527 (((tmp_addr >= swap_base) && 2528 (tmp_addr <= swap_limit)) || 2529 (tmp_addr < rgn_size))) 2530 return sys_addr ^ (u64)swap_base << 27; 2531 2532 return sys_addr; 2533 } 2534 2535 /* For a given @dram_range, check if @sys_addr falls within it. */ 2536 static int f1x_match_to_this_node(struct amd64_pvt *pvt, unsigned range, 2537 u64 sys_addr, int *chan_sel) 2538 { 2539 int cs_found = -EINVAL; 2540 u64 chan_addr; 2541 u32 dct_sel_base; 2542 u8 channel; 2543 bool high_range = false; 2544 2545 u8 node_id = dram_dst_node(pvt, range); 2546 u8 intlv_en = dram_intlv_en(pvt, range); 2547 u32 intlv_sel = dram_intlv_sel(pvt, range); 2548 2549 edac_dbg(1, "(range %d) SystemAddr= 0x%llx Limit=0x%llx\n", 2550 range, sys_addr, get_dram_limit(pvt, range)); 2551 2552 if (dhar_valid(pvt) && 2553 dhar_base(pvt) <= sys_addr && 2554 sys_addr < BIT_64(32)) { 2555 amd64_warn("Huh? Address is in the MMIO hole: 0x%016llx\n", 2556 sys_addr); 2557 return -EINVAL; 2558 } 2559 2560 if (intlv_en && (intlv_sel != ((sys_addr >> 12) & intlv_en))) 2561 return -EINVAL; 2562 2563 sys_addr = f1x_swap_interleaved_region(pvt, sys_addr); 2564 2565 dct_sel_base = dct_sel_baseaddr(pvt); 2566 2567 /* 2568 * check whether addresses >= DctSelBaseAddr[47:27] are to be used to 2569 * select between DCT0 and DCT1. 2570 */ 2571 if (dct_high_range_enabled(pvt) && 2572 !dct_ganging_enabled(pvt) && 2573 ((sys_addr >> 27) >= (dct_sel_base >> 11))) 2574 high_range = true; 2575 2576 channel = f1x_determine_channel(pvt, sys_addr, high_range, intlv_en); 2577 2578 chan_addr = f1x_get_norm_dct_addr(pvt, range, sys_addr, 2579 high_range, dct_sel_base); 2580 2581 /* Remove node interleaving, see F1x120 */ 2582 if (intlv_en) 2583 chan_addr = ((chan_addr >> (12 + hweight8(intlv_en))) << 12) | 2584 (chan_addr & 0xfff); 2585 2586 /* remove channel interleave */ 2587 if (dct_interleave_enabled(pvt) && 2588 !dct_high_range_enabled(pvt) && 2589 !dct_ganging_enabled(pvt)) { 2590 2591 if (dct_sel_interleave_addr(pvt) != 1) { 2592 if (dct_sel_interleave_addr(pvt) == 0x3) 2593 /* hash 9 */ 2594 chan_addr = ((chan_addr >> 10) << 9) | 2595 (chan_addr & 0x1ff); 2596 else 2597 /* A[6] or hash 6 */ 2598 chan_addr = ((chan_addr >> 7) << 6) | 2599 (chan_addr & 0x3f); 2600 } else 2601 /* A[12] */ 2602 chan_addr = ((chan_addr >> 13) << 12) | 2603 (chan_addr & 0xfff); 2604 } 2605 2606 edac_dbg(1, " Normalized DCT addr: 0x%llx\n", chan_addr); 2607 2608 cs_found = f1x_lookup_addr_in_dct(chan_addr, node_id, channel); 2609 2610 if (cs_found >= 0) 2611 *chan_sel = channel; 2612 2613 return cs_found; 2614 } 2615 2616 static int f15_m30h_match_to_this_node(struct amd64_pvt *pvt, unsigned range, 2617 u64 sys_addr, int *chan_sel) 2618 { 2619 int cs_found = -EINVAL; 2620 int num_dcts_intlv = 0; 2621 u64 chan_addr, chan_offset; 2622 u64 dct_base, dct_limit; 2623 u32 dct_cont_base_reg, dct_cont_limit_reg, tmp; 2624 u8 channel, alias_channel, leg_mmio_hole, dct_sel, dct_offset_en; 2625 2626 u64 dhar_offset = f10_dhar_offset(pvt); 2627 u8 intlv_addr = dct_sel_interleave_addr(pvt); 2628 u8 node_id = dram_dst_node(pvt, range); 2629 u8 intlv_en = dram_intlv_en(pvt, range); 2630 2631 amd64_read_pci_cfg(pvt->F1, DRAM_CONT_BASE, &dct_cont_base_reg); 2632 amd64_read_pci_cfg(pvt->F1, DRAM_CONT_LIMIT, &dct_cont_limit_reg); 2633 2634 dct_offset_en = (u8) ((dct_cont_base_reg >> 3) & BIT(0)); 2635 dct_sel = (u8) ((dct_cont_base_reg >> 4) & 0x7); 2636 2637 edac_dbg(1, "(range %d) SystemAddr= 0x%llx Limit=0x%llx\n", 2638 range, sys_addr, get_dram_limit(pvt, range)); 2639 2640 if (!(get_dram_base(pvt, range) <= sys_addr) && 2641 !(get_dram_limit(pvt, range) >= sys_addr)) 2642 return -EINVAL; 2643 2644 if (dhar_valid(pvt) && 2645 dhar_base(pvt) <= sys_addr && 2646 sys_addr < BIT_64(32)) { 2647 amd64_warn("Huh? Address is in the MMIO hole: 0x%016llx\n", 2648 sys_addr); 2649 return -EINVAL; 2650 } 2651 2652 /* Verify sys_addr is within DCT Range. */ 2653 dct_base = (u64) dct_sel_baseaddr(pvt); 2654 dct_limit = (dct_cont_limit_reg >> 11) & 0x1FFF; 2655 2656 if (!(dct_cont_base_reg & BIT(0)) && 2657 !(dct_base <= (sys_addr >> 27) && 2658 dct_limit >= (sys_addr >> 27))) 2659 return -EINVAL; 2660 2661 /* Verify number of dct's that participate in channel interleaving. */ 2662 num_dcts_intlv = (int) hweight8(intlv_en); 2663 2664 if (!(num_dcts_intlv % 2 == 0) || (num_dcts_intlv > 4)) 2665 return -EINVAL; 2666 2667 if (pvt->model >= 0x60) 2668 channel = f1x_determine_channel(pvt, sys_addr, false, intlv_en); 2669 else 2670 channel = f15_m30h_determine_channel(pvt, sys_addr, intlv_en, 2671 num_dcts_intlv, dct_sel); 2672 2673 /* Verify we stay within the MAX number of channels allowed */ 2674 if (channel > 3) 2675 return -EINVAL; 2676 2677 leg_mmio_hole = (u8) (dct_cont_base_reg >> 1 & BIT(0)); 2678 2679 /* Get normalized DCT addr */ 2680 if (leg_mmio_hole && (sys_addr >= BIT_64(32))) 2681 chan_offset = dhar_offset; 2682 else 2683 chan_offset = dct_base << 27; 2684 2685 chan_addr = sys_addr - chan_offset; 2686 2687 /* remove channel interleave */ 2688 if (num_dcts_intlv == 2) { 2689 if (intlv_addr == 0x4) 2690 chan_addr = ((chan_addr >> 9) << 8) | 2691 (chan_addr & 0xff); 2692 else if (intlv_addr == 0x5) 2693 chan_addr = ((chan_addr >> 10) << 9) | 2694 (chan_addr & 0x1ff); 2695 else 2696 return -EINVAL; 2697 2698 } else if (num_dcts_intlv == 4) { 2699 if (intlv_addr == 0x4) 2700 chan_addr = ((chan_addr >> 10) << 8) | 2701 (chan_addr & 0xff); 2702 else if (intlv_addr == 0x5) 2703 chan_addr = ((chan_addr >> 11) << 9) | 2704 (chan_addr & 0x1ff); 2705 else 2706 return -EINVAL; 2707 } 2708 2709 if (dct_offset_en) { 2710 amd64_read_pci_cfg(pvt->F1, 2711 DRAM_CONT_HIGH_OFF + (int) channel * 4, 2712 &tmp); 2713 chan_addr += (u64) ((tmp >> 11) & 0xfff) << 27; 2714 } 2715 2716 f15h_select_dct(pvt, channel); 2717 2718 edac_dbg(1, " Normalized DCT addr: 0x%llx\n", chan_addr); 2719 2720 /* 2721 * Find Chip select: 2722 * if channel = 3, then alias it to 1. This is because, in F15 M30h, 2723 * there is support for 4 DCT's, but only 2 are currently functional. 2724 * They are DCT0 and DCT3. But we have read all registers of DCT3 into 2725 * pvt->csels[1]. So we need to use '1' here to get correct info. 2726 * Refer F15 M30h BKDG Section 2.10 and 2.10.3 for clarifications. 2727 */ 2728 alias_channel = (channel == 3) ? 1 : channel; 2729 2730 cs_found = f1x_lookup_addr_in_dct(chan_addr, node_id, alias_channel); 2731 2732 if (cs_found >= 0) 2733 *chan_sel = alias_channel; 2734 2735 return cs_found; 2736 } 2737 2738 static int f1x_translate_sysaddr_to_cs(struct amd64_pvt *pvt, 2739 u64 sys_addr, 2740 int *chan_sel) 2741 { 2742 int cs_found = -EINVAL; 2743 unsigned range; 2744 2745 for (range = 0; range < DRAM_RANGES; range++) { 2746 if (!dram_rw(pvt, range)) 2747 continue; 2748 2749 if (pvt->fam == 0x15 && pvt->model >= 0x30) 2750 cs_found = f15_m30h_match_to_this_node(pvt, range, 2751 sys_addr, 2752 chan_sel); 2753 2754 else if ((get_dram_base(pvt, range) <= sys_addr) && 2755 (get_dram_limit(pvt, range) >= sys_addr)) { 2756 cs_found = f1x_match_to_this_node(pvt, range, 2757 sys_addr, chan_sel); 2758 if (cs_found >= 0) 2759 break; 2760 } 2761 } 2762 return cs_found; 2763 } 2764 2765 /* 2766 * For reference see "2.8.5 Routing DRAM Requests" in F10 BKDG. This code maps 2767 * a @sys_addr to NodeID, DCT (channel) and chip select (CSROW). 2768 * 2769 * The @sys_addr is usually an error address received from the hardware 2770 * (MCX_ADDR). 2771 */ 2772 static void f1x_map_sysaddr_to_csrow(struct mem_ctl_info *mci, u64 sys_addr, 2773 struct err_info *err) 2774 { 2775 struct amd64_pvt *pvt = mci->pvt_info; 2776 2777 error_address_to_page_and_offset(sys_addr, err); 2778 2779 err->csrow = f1x_translate_sysaddr_to_cs(pvt, sys_addr, &err->channel); 2780 if (err->csrow < 0) { 2781 err->err_code = ERR_CSROW; 2782 return; 2783 } 2784 2785 /* 2786 * We need the syndromes for channel detection only when we're 2787 * ganged. Otherwise @chan should already contain the channel at 2788 * this point. 2789 */ 2790 if (dct_ganging_enabled(pvt)) 2791 err->channel = get_channel_from_ecc_syndrome(mci, err->syndrome); 2792 } 2793 2794 /* 2795 * debug routine to display the memory sizes of all logical DIMMs and its 2796 * CSROWs 2797 */ 2798 static void debug_display_dimm_sizes(struct amd64_pvt *pvt, u8 ctrl) 2799 { 2800 int dimm, size0, size1; 2801 u32 *dcsb = ctrl ? pvt->csels[1].csbases : pvt->csels[0].csbases; 2802 u32 dbam = ctrl ? pvt->dbam1 : pvt->dbam0; 2803 2804 if (pvt->fam == 0xf) { 2805 /* K8 families < revF not supported yet */ 2806 if (pvt->ext_model < K8_REV_F) 2807 return; 2808 else 2809 WARN_ON(ctrl != 0); 2810 } 2811 2812 if (pvt->fam == 0x10) { 2813 dbam = (ctrl && !dct_ganging_enabled(pvt)) ? pvt->dbam1 2814 : pvt->dbam0; 2815 dcsb = (ctrl && !dct_ganging_enabled(pvt)) ? 2816 pvt->csels[1].csbases : 2817 pvt->csels[0].csbases; 2818 } else if (ctrl) { 2819 dbam = pvt->dbam0; 2820 dcsb = pvt->csels[1].csbases; 2821 } 2822 edac_dbg(1, "F2x%d80 (DRAM Bank Address Mapping): 0x%08x\n", 2823 ctrl, dbam); 2824 2825 edac_printk(KERN_DEBUG, EDAC_MC, "DCT%d chip selects:\n", ctrl); 2826 2827 /* Dump memory sizes for DIMM and its CSROWs */ 2828 for (dimm = 0; dimm < 4; dimm++) { 2829 2830 size0 = 0; 2831 if (dcsb[dimm*2] & DCSB_CS_ENABLE) 2832 /* 2833 * For F15m60h, we need multiplier for LRDIMM cs_size 2834 * calculation. We pass dimm value to the dbam_to_cs 2835 * mapper so we can find the multiplier from the 2836 * corresponding DCSM. 2837 */ 2838 size0 = pvt->ops->dbam_to_cs(pvt, ctrl, 2839 DBAM_DIMM(dimm, dbam), 2840 dimm); 2841 2842 size1 = 0; 2843 if (dcsb[dimm*2 + 1] & DCSB_CS_ENABLE) 2844 size1 = pvt->ops->dbam_to_cs(pvt, ctrl, 2845 DBAM_DIMM(dimm, dbam), 2846 dimm); 2847 2848 amd64_info(EDAC_MC ": %d: %5dMB %d: %5dMB\n", 2849 dimm * 2, size0, 2850 dimm * 2 + 1, size1); 2851 } 2852 } 2853 2854 static struct amd64_family_type family_types[] = { 2855 [K8_CPUS] = { 2856 .ctl_name = "K8", 2857 .f1_id = PCI_DEVICE_ID_AMD_K8_NB_ADDRMAP, 2858 .f2_id = PCI_DEVICE_ID_AMD_K8_NB_MEMCTL, 2859 .max_mcs = 2, 2860 .ops = { 2861 .early_channel_count = k8_early_channel_count, 2862 .map_sysaddr_to_csrow = k8_map_sysaddr_to_csrow, 2863 .dbam_to_cs = k8_dbam_to_chip_select, 2864 } 2865 }, 2866 [F10_CPUS] = { 2867 .ctl_name = "F10h", 2868 .f1_id = PCI_DEVICE_ID_AMD_10H_NB_MAP, 2869 .f2_id = PCI_DEVICE_ID_AMD_10H_NB_DRAM, 2870 .max_mcs = 2, 2871 .ops = { 2872 .early_channel_count = f1x_early_channel_count, 2873 .map_sysaddr_to_csrow = f1x_map_sysaddr_to_csrow, 2874 .dbam_to_cs = f10_dbam_to_chip_select, 2875 } 2876 }, 2877 [F15_CPUS] = { 2878 .ctl_name = "F15h", 2879 .f1_id = PCI_DEVICE_ID_AMD_15H_NB_F1, 2880 .f2_id = PCI_DEVICE_ID_AMD_15H_NB_F2, 2881 .max_mcs = 2, 2882 .ops = { 2883 .early_channel_count = f1x_early_channel_count, 2884 .map_sysaddr_to_csrow = f1x_map_sysaddr_to_csrow, 2885 .dbam_to_cs = f15_dbam_to_chip_select, 2886 } 2887 }, 2888 [F15_M30H_CPUS] = { 2889 .ctl_name = "F15h_M30h", 2890 .f1_id = PCI_DEVICE_ID_AMD_15H_M30H_NB_F1, 2891 .f2_id = PCI_DEVICE_ID_AMD_15H_M30H_NB_F2, 2892 .max_mcs = 2, 2893 .ops = { 2894 .early_channel_count = f1x_early_channel_count, 2895 .map_sysaddr_to_csrow = f1x_map_sysaddr_to_csrow, 2896 .dbam_to_cs = f16_dbam_to_chip_select, 2897 } 2898 }, 2899 [F15_M60H_CPUS] = { 2900 .ctl_name = "F15h_M60h", 2901 .f1_id = PCI_DEVICE_ID_AMD_15H_M60H_NB_F1, 2902 .f2_id = PCI_DEVICE_ID_AMD_15H_M60H_NB_F2, 2903 .max_mcs = 2, 2904 .ops = { 2905 .early_channel_count = f1x_early_channel_count, 2906 .map_sysaddr_to_csrow = f1x_map_sysaddr_to_csrow, 2907 .dbam_to_cs = f15_m60h_dbam_to_chip_select, 2908 } 2909 }, 2910 [F16_CPUS] = { 2911 .ctl_name = "F16h", 2912 .f1_id = PCI_DEVICE_ID_AMD_16H_NB_F1, 2913 .f2_id = PCI_DEVICE_ID_AMD_16H_NB_F2, 2914 .max_mcs = 2, 2915 .ops = { 2916 .early_channel_count = f1x_early_channel_count, 2917 .map_sysaddr_to_csrow = f1x_map_sysaddr_to_csrow, 2918 .dbam_to_cs = f16_dbam_to_chip_select, 2919 } 2920 }, 2921 [F16_M30H_CPUS] = { 2922 .ctl_name = "F16h_M30h", 2923 .f1_id = PCI_DEVICE_ID_AMD_16H_M30H_NB_F1, 2924 .f2_id = PCI_DEVICE_ID_AMD_16H_M30H_NB_F2, 2925 .max_mcs = 2, 2926 .ops = { 2927 .early_channel_count = f1x_early_channel_count, 2928 .map_sysaddr_to_csrow = f1x_map_sysaddr_to_csrow, 2929 .dbam_to_cs = f16_dbam_to_chip_select, 2930 } 2931 }, 2932 [F17_CPUS] = { 2933 .ctl_name = "F17h", 2934 .f0_id = PCI_DEVICE_ID_AMD_17H_DF_F0, 2935 .f6_id = PCI_DEVICE_ID_AMD_17H_DF_F6, 2936 .max_mcs = 2, 2937 .ops = { 2938 .early_channel_count = f17_early_channel_count, 2939 .dbam_to_cs = f17_addr_mask_to_cs_size, 2940 } 2941 }, 2942 [F17_M10H_CPUS] = { 2943 .ctl_name = "F17h_M10h", 2944 .f0_id = PCI_DEVICE_ID_AMD_17H_M10H_DF_F0, 2945 .f6_id = PCI_DEVICE_ID_AMD_17H_M10H_DF_F6, 2946 .max_mcs = 2, 2947 .ops = { 2948 .early_channel_count = f17_early_channel_count, 2949 .dbam_to_cs = f17_addr_mask_to_cs_size, 2950 } 2951 }, 2952 [F17_M30H_CPUS] = { 2953 .ctl_name = "F17h_M30h", 2954 .f0_id = PCI_DEVICE_ID_AMD_17H_M30H_DF_F0, 2955 .f6_id = PCI_DEVICE_ID_AMD_17H_M30H_DF_F6, 2956 .max_mcs = 8, 2957 .ops = { 2958 .early_channel_count = f17_early_channel_count, 2959 .dbam_to_cs = f17_addr_mask_to_cs_size, 2960 } 2961 }, 2962 [F17_M60H_CPUS] = { 2963 .ctl_name = "F17h_M60h", 2964 .f0_id = PCI_DEVICE_ID_AMD_17H_M60H_DF_F0, 2965 .f6_id = PCI_DEVICE_ID_AMD_17H_M60H_DF_F6, 2966 .max_mcs = 2, 2967 .ops = { 2968 .early_channel_count = f17_early_channel_count, 2969 .dbam_to_cs = f17_addr_mask_to_cs_size, 2970 } 2971 }, 2972 [F17_M70H_CPUS] = { 2973 .ctl_name = "F17h_M70h", 2974 .f0_id = PCI_DEVICE_ID_AMD_17H_M70H_DF_F0, 2975 .f6_id = PCI_DEVICE_ID_AMD_17H_M70H_DF_F6, 2976 .max_mcs = 2, 2977 .ops = { 2978 .early_channel_count = f17_early_channel_count, 2979 .dbam_to_cs = f17_addr_mask_to_cs_size, 2980 } 2981 }, 2982 [F19_CPUS] = { 2983 .ctl_name = "F19h", 2984 .f0_id = PCI_DEVICE_ID_AMD_19H_DF_F0, 2985 .f6_id = PCI_DEVICE_ID_AMD_19H_DF_F6, 2986 .max_mcs = 8, 2987 .ops = { 2988 .early_channel_count = f17_early_channel_count, 2989 .dbam_to_cs = f17_addr_mask_to_cs_size, 2990 } 2991 }, 2992 [F19_M10H_CPUS] = { 2993 .ctl_name = "F19h_M10h", 2994 .f0_id = PCI_DEVICE_ID_AMD_19H_M10H_DF_F0, 2995 .f6_id = PCI_DEVICE_ID_AMD_19H_M10H_DF_F6, 2996 .max_mcs = 12, 2997 .flags.zn_regs_v2 = 1, 2998 .ops = { 2999 .early_channel_count = f17_early_channel_count, 3000 .dbam_to_cs = f17_addr_mask_to_cs_size, 3001 } 3002 }, 3003 [F19_M50H_CPUS] = { 3004 .ctl_name = "F19h_M50h", 3005 .f0_id = PCI_DEVICE_ID_AMD_19H_M50H_DF_F0, 3006 .f6_id = PCI_DEVICE_ID_AMD_19H_M50H_DF_F6, 3007 .max_mcs = 2, 3008 .ops = { 3009 .early_channel_count = f17_early_channel_count, 3010 .dbam_to_cs = f17_addr_mask_to_cs_size, 3011 } 3012 }, 3013 }; 3014 3015 /* 3016 * These are tables of eigenvectors (one per line) which can be used for the 3017 * construction of the syndrome tables. The modified syndrome search algorithm 3018 * uses those to find the symbol in error and thus the DIMM. 3019 * 3020 * Algorithm courtesy of Ross LaFetra from AMD. 3021 */ 3022 static const u16 x4_vectors[] = { 3023 0x2f57, 0x1afe, 0x66cc, 0xdd88, 3024 0x11eb, 0x3396, 0x7f4c, 0xeac8, 3025 0x0001, 0x0002, 0x0004, 0x0008, 3026 0x1013, 0x3032, 0x4044, 0x8088, 3027 0x106b, 0x30d6, 0x70fc, 0xe0a8, 3028 0x4857, 0xc4fe, 0x13cc, 0x3288, 3029 0x1ac5, 0x2f4a, 0x5394, 0xa1e8, 3030 0x1f39, 0x251e, 0xbd6c, 0x6bd8, 3031 0x15c1, 0x2a42, 0x89ac, 0x4758, 3032 0x2b03, 0x1602, 0x4f0c, 0xca08, 3033 0x1f07, 0x3a0e, 0x6b04, 0xbd08, 3034 0x8ba7, 0x465e, 0x244c, 0x1cc8, 3035 0x2b87, 0x164e, 0x642c, 0xdc18, 3036 0x40b9, 0x80de, 0x1094, 0x20e8, 3037 0x27db, 0x1eb6, 0x9dac, 0x7b58, 3038 0x11c1, 0x2242, 0x84ac, 0x4c58, 3039 0x1be5, 0x2d7a, 0x5e34, 0xa718, 3040 0x4b39, 0x8d1e, 0x14b4, 0x28d8, 3041 0x4c97, 0xc87e, 0x11fc, 0x33a8, 3042 0x8e97, 0x497e, 0x2ffc, 0x1aa8, 3043 0x16b3, 0x3d62, 0x4f34, 0x8518, 3044 0x1e2f, 0x391a, 0x5cac, 0xf858, 3045 0x1d9f, 0x3b7a, 0x572c, 0xfe18, 3046 0x15f5, 0x2a5a, 0x5264, 0xa3b8, 3047 0x1dbb, 0x3b66, 0x715c, 0xe3f8, 3048 0x4397, 0xc27e, 0x17fc, 0x3ea8, 3049 0x1617, 0x3d3e, 0x6464, 0xb8b8, 3050 0x23ff, 0x12aa, 0xab6c, 0x56d8, 3051 0x2dfb, 0x1ba6, 0x913c, 0x7328, 3052 0x185d, 0x2ca6, 0x7914, 0x9e28, 3053 0x171b, 0x3e36, 0x7d7c, 0xebe8, 3054 0x4199, 0x82ee, 0x19f4, 0x2e58, 3055 0x4807, 0xc40e, 0x130c, 0x3208, 3056 0x1905, 0x2e0a, 0x5804, 0xac08, 3057 0x213f, 0x132a, 0xadfc, 0x5ba8, 3058 0x19a9, 0x2efe, 0xb5cc, 0x6f88, 3059 }; 3060 3061 static const u16 x8_vectors[] = { 3062 0x0145, 0x028a, 0x2374, 0x43c8, 0xa1f0, 0x0520, 0x0a40, 0x1480, 3063 0x0211, 0x0422, 0x0844, 0x1088, 0x01b0, 0x44e0, 0x23c0, 0xed80, 3064 0x1011, 0x0116, 0x022c, 0x0458, 0x08b0, 0x8c60, 0x2740, 0x4e80, 3065 0x0411, 0x0822, 0x1044, 0x0158, 0x02b0, 0x2360, 0x46c0, 0xab80, 3066 0x0811, 0x1022, 0x012c, 0x0258, 0x04b0, 0x4660, 0x8cc0, 0x2780, 3067 0x2071, 0x40e2, 0xa0c4, 0x0108, 0x0210, 0x0420, 0x0840, 0x1080, 3068 0x4071, 0x80e2, 0x0104, 0x0208, 0x0410, 0x0820, 0x1040, 0x2080, 3069 0x8071, 0x0102, 0x0204, 0x0408, 0x0810, 0x1020, 0x2040, 0x4080, 3070 0x019d, 0x03d6, 0x136c, 0x2198, 0x50b0, 0xb2e0, 0x0740, 0x0e80, 3071 0x0189, 0x03ea, 0x072c, 0x0e58, 0x1cb0, 0x56e0, 0x37c0, 0xf580, 3072 0x01fd, 0x0376, 0x06ec, 0x0bb8, 0x1110, 0x2220, 0x4440, 0x8880, 3073 0x0163, 0x02c6, 0x1104, 0x0758, 0x0eb0, 0x2be0, 0x6140, 0xc280, 3074 0x02fd, 0x01c6, 0x0b5c, 0x1108, 0x07b0, 0x25a0, 0x8840, 0x6180, 3075 0x0801, 0x012e, 0x025c, 0x04b8, 0x1370, 0x26e0, 0x57c0, 0xb580, 3076 0x0401, 0x0802, 0x015c, 0x02b8, 0x22b0, 0x13e0, 0x7140, 0xe280, 3077 0x0201, 0x0402, 0x0804, 0x01b8, 0x11b0, 0x31a0, 0x8040, 0x7180, 3078 0x0101, 0x0202, 0x0404, 0x0808, 0x1010, 0x2020, 0x4040, 0x8080, 3079 0x0001, 0x0002, 0x0004, 0x0008, 0x0010, 0x0020, 0x0040, 0x0080, 3080 0x0100, 0x0200, 0x0400, 0x0800, 0x1000, 0x2000, 0x4000, 0x8000, 3081 }; 3082 3083 static int decode_syndrome(u16 syndrome, const u16 *vectors, unsigned num_vecs, 3084 unsigned v_dim) 3085 { 3086 unsigned int i, err_sym; 3087 3088 for (err_sym = 0; err_sym < num_vecs / v_dim; err_sym++) { 3089 u16 s = syndrome; 3090 unsigned v_idx = err_sym * v_dim; 3091 unsigned v_end = (err_sym + 1) * v_dim; 3092 3093 /* walk over all 16 bits of the syndrome */ 3094 for (i = 1; i < (1U << 16); i <<= 1) { 3095 3096 /* if bit is set in that eigenvector... */ 3097 if (v_idx < v_end && vectors[v_idx] & i) { 3098 u16 ev_comp = vectors[v_idx++]; 3099 3100 /* ... and bit set in the modified syndrome, */ 3101 if (s & i) { 3102 /* remove it. */ 3103 s ^= ev_comp; 3104 3105 if (!s) 3106 return err_sym; 3107 } 3108 3109 } else if (s & i) 3110 /* can't get to zero, move to next symbol */ 3111 break; 3112 } 3113 } 3114 3115 edac_dbg(0, "syndrome(%x) not found\n", syndrome); 3116 return -1; 3117 } 3118 3119 static int map_err_sym_to_channel(int err_sym, int sym_size) 3120 { 3121 if (sym_size == 4) 3122 switch (err_sym) { 3123 case 0x20: 3124 case 0x21: 3125 return 0; 3126 case 0x22: 3127 case 0x23: 3128 return 1; 3129 default: 3130 return err_sym >> 4; 3131 } 3132 /* x8 symbols */ 3133 else 3134 switch (err_sym) { 3135 /* imaginary bits not in a DIMM */ 3136 case 0x10: 3137 WARN(1, KERN_ERR "Invalid error symbol: 0x%x\n", 3138 err_sym); 3139 return -1; 3140 case 0x11: 3141 return 0; 3142 case 0x12: 3143 return 1; 3144 default: 3145 return err_sym >> 3; 3146 } 3147 return -1; 3148 } 3149 3150 static int get_channel_from_ecc_syndrome(struct mem_ctl_info *mci, u16 syndrome) 3151 { 3152 struct amd64_pvt *pvt = mci->pvt_info; 3153 int err_sym = -1; 3154 3155 if (pvt->ecc_sym_sz == 8) 3156 err_sym = decode_syndrome(syndrome, x8_vectors, 3157 ARRAY_SIZE(x8_vectors), 3158 pvt->ecc_sym_sz); 3159 else if (pvt->ecc_sym_sz == 4) 3160 err_sym = decode_syndrome(syndrome, x4_vectors, 3161 ARRAY_SIZE(x4_vectors), 3162 pvt->ecc_sym_sz); 3163 else { 3164 amd64_warn("Illegal syndrome type: %u\n", pvt->ecc_sym_sz); 3165 return err_sym; 3166 } 3167 3168 return map_err_sym_to_channel(err_sym, pvt->ecc_sym_sz); 3169 } 3170 3171 static void __log_ecc_error(struct mem_ctl_info *mci, struct err_info *err, 3172 u8 ecc_type) 3173 { 3174 enum hw_event_mc_err_type err_type; 3175 const char *string; 3176 3177 if (ecc_type == 2) 3178 err_type = HW_EVENT_ERR_CORRECTED; 3179 else if (ecc_type == 1) 3180 err_type = HW_EVENT_ERR_UNCORRECTED; 3181 else if (ecc_type == 3) 3182 err_type = HW_EVENT_ERR_DEFERRED; 3183 else { 3184 WARN(1, "Something is rotten in the state of Denmark.\n"); 3185 return; 3186 } 3187 3188 switch (err->err_code) { 3189 case DECODE_OK: 3190 string = ""; 3191 break; 3192 case ERR_NODE: 3193 string = "Failed to map error addr to a node"; 3194 break; 3195 case ERR_CSROW: 3196 string = "Failed to map error addr to a csrow"; 3197 break; 3198 case ERR_CHANNEL: 3199 string = "Unknown syndrome - possible error reporting race"; 3200 break; 3201 case ERR_SYND: 3202 string = "MCA_SYND not valid - unknown syndrome and csrow"; 3203 break; 3204 case ERR_NORM_ADDR: 3205 string = "Cannot decode normalized address"; 3206 break; 3207 default: 3208 string = "WTF error"; 3209 break; 3210 } 3211 3212 edac_mc_handle_error(err_type, mci, 1, 3213 err->page, err->offset, err->syndrome, 3214 err->csrow, err->channel, -1, 3215 string, ""); 3216 } 3217 3218 static inline void decode_bus_error(int node_id, struct mce *m) 3219 { 3220 struct mem_ctl_info *mci; 3221 struct amd64_pvt *pvt; 3222 u8 ecc_type = (m->status >> 45) & 0x3; 3223 u8 xec = XEC(m->status, 0x1f); 3224 u16 ec = EC(m->status); 3225 u64 sys_addr; 3226 struct err_info err; 3227 3228 mci = edac_mc_find(node_id); 3229 if (!mci) 3230 return; 3231 3232 pvt = mci->pvt_info; 3233 3234 /* Bail out early if this was an 'observed' error */ 3235 if (PP(ec) == NBSL_PP_OBS) 3236 return; 3237 3238 /* Do only ECC errors */ 3239 if (xec && xec != F10_NBSL_EXT_ERR_ECC) 3240 return; 3241 3242 memset(&err, 0, sizeof(err)); 3243 3244 sys_addr = get_error_address(pvt, m); 3245 3246 if (ecc_type == 2) 3247 err.syndrome = extract_syndrome(m->status); 3248 3249 pvt->ops->map_sysaddr_to_csrow(mci, sys_addr, &err); 3250 3251 __log_ecc_error(mci, &err, ecc_type); 3252 } 3253 3254 /* 3255 * To find the UMC channel represented by this bank we need to match on its 3256 * instance_id. The instance_id of a bank is held in the lower 32 bits of its 3257 * IPID. 3258 * 3259 * Currently, we can derive the channel number by looking at the 6th nibble in 3260 * the instance_id. For example, instance_id=0xYXXXXX where Y is the channel 3261 * number. 3262 */ 3263 static int find_umc_channel(struct mce *m) 3264 { 3265 return (m->ipid & GENMASK(31, 0)) >> 20; 3266 } 3267 3268 static void decode_umc_error(int node_id, struct mce *m) 3269 { 3270 u8 ecc_type = (m->status >> 45) & 0x3; 3271 struct mem_ctl_info *mci; 3272 struct amd64_pvt *pvt; 3273 struct err_info err; 3274 u64 sys_addr; 3275 3276 mci = edac_mc_find(node_id); 3277 if (!mci) 3278 return; 3279 3280 pvt = mci->pvt_info; 3281 3282 memset(&err, 0, sizeof(err)); 3283 3284 if (m->status & MCI_STATUS_DEFERRED) 3285 ecc_type = 3; 3286 3287 err.channel = find_umc_channel(m); 3288 3289 if (!(m->status & MCI_STATUS_SYNDV)) { 3290 err.err_code = ERR_SYND; 3291 goto log_error; 3292 } 3293 3294 if (ecc_type == 2) { 3295 u8 length = (m->synd >> 18) & 0x3f; 3296 3297 if (length) 3298 err.syndrome = (m->synd >> 32) & GENMASK(length - 1, 0); 3299 else 3300 err.err_code = ERR_CHANNEL; 3301 } 3302 3303 err.csrow = m->synd & 0x7; 3304 3305 if (umc_normaddr_to_sysaddr(m->addr, pvt->mc_node_id, err.channel, &sys_addr)) { 3306 err.err_code = ERR_NORM_ADDR; 3307 goto log_error; 3308 } 3309 3310 error_address_to_page_and_offset(sys_addr, &err); 3311 3312 log_error: 3313 __log_ecc_error(mci, &err, ecc_type); 3314 } 3315 3316 /* 3317 * Use pvt->F3 which contains the F3 CPU PCI device to get the related 3318 * F1 (AddrMap) and F2 (Dct) devices. Return negative value on error. 3319 * Reserve F0 and F6 on systems with a UMC. 3320 */ 3321 static int 3322 reserve_mc_sibling_devs(struct amd64_pvt *pvt, u16 pci_id1, u16 pci_id2) 3323 { 3324 if (pvt->umc) { 3325 pvt->F0 = pci_get_related_function(pvt->F3->vendor, pci_id1, pvt->F3); 3326 if (!pvt->F0) { 3327 edac_dbg(1, "F0 not found, device 0x%x\n", pci_id1); 3328 return -ENODEV; 3329 } 3330 3331 pvt->F6 = pci_get_related_function(pvt->F3->vendor, pci_id2, pvt->F3); 3332 if (!pvt->F6) { 3333 pci_dev_put(pvt->F0); 3334 pvt->F0 = NULL; 3335 3336 edac_dbg(1, "F6 not found: device 0x%x\n", pci_id2); 3337 return -ENODEV; 3338 } 3339 3340 if (!pci_ctl_dev) 3341 pci_ctl_dev = &pvt->F0->dev; 3342 3343 edac_dbg(1, "F0: %s\n", pci_name(pvt->F0)); 3344 edac_dbg(1, "F3: %s\n", pci_name(pvt->F3)); 3345 edac_dbg(1, "F6: %s\n", pci_name(pvt->F6)); 3346 3347 return 0; 3348 } 3349 3350 /* Reserve the ADDRESS MAP Device */ 3351 pvt->F1 = pci_get_related_function(pvt->F3->vendor, pci_id1, pvt->F3); 3352 if (!pvt->F1) { 3353 edac_dbg(1, "F1 not found: device 0x%x\n", pci_id1); 3354 return -ENODEV; 3355 } 3356 3357 /* Reserve the DCT Device */ 3358 pvt->F2 = pci_get_related_function(pvt->F3->vendor, pci_id2, pvt->F3); 3359 if (!pvt->F2) { 3360 pci_dev_put(pvt->F1); 3361 pvt->F1 = NULL; 3362 3363 edac_dbg(1, "F2 not found: device 0x%x\n", pci_id2); 3364 return -ENODEV; 3365 } 3366 3367 if (!pci_ctl_dev) 3368 pci_ctl_dev = &pvt->F2->dev; 3369 3370 edac_dbg(1, "F1: %s\n", pci_name(pvt->F1)); 3371 edac_dbg(1, "F2: %s\n", pci_name(pvt->F2)); 3372 edac_dbg(1, "F3: %s\n", pci_name(pvt->F3)); 3373 3374 return 0; 3375 } 3376 3377 static void free_mc_sibling_devs(struct amd64_pvt *pvt) 3378 { 3379 if (pvt->umc) { 3380 pci_dev_put(pvt->F0); 3381 pci_dev_put(pvt->F6); 3382 } else { 3383 pci_dev_put(pvt->F1); 3384 pci_dev_put(pvt->F2); 3385 } 3386 } 3387 3388 static void determine_ecc_sym_sz(struct amd64_pvt *pvt) 3389 { 3390 pvt->ecc_sym_sz = 4; 3391 3392 if (pvt->umc) { 3393 u8 i; 3394 3395 for_each_umc(i) { 3396 /* Check enabled channels only: */ 3397 if (pvt->umc[i].sdp_ctrl & UMC_SDP_INIT) { 3398 if (pvt->umc[i].ecc_ctrl & BIT(9)) { 3399 pvt->ecc_sym_sz = 16; 3400 return; 3401 } else if (pvt->umc[i].ecc_ctrl & BIT(7)) { 3402 pvt->ecc_sym_sz = 8; 3403 return; 3404 } 3405 } 3406 } 3407 } else if (pvt->fam >= 0x10) { 3408 u32 tmp; 3409 3410 amd64_read_pci_cfg(pvt->F3, EXT_NB_MCA_CFG, &tmp); 3411 /* F16h has only DCT0, so no need to read dbam1. */ 3412 if (pvt->fam != 0x16) 3413 amd64_read_dct_pci_cfg(pvt, 1, DBAM0, &pvt->dbam1); 3414 3415 /* F10h, revD and later can do x8 ECC too. */ 3416 if ((pvt->fam > 0x10 || pvt->model > 7) && tmp & BIT(25)) 3417 pvt->ecc_sym_sz = 8; 3418 } 3419 } 3420 3421 /* 3422 * Retrieve the hardware registers of the memory controller. 3423 */ 3424 static void __read_mc_regs_df(struct amd64_pvt *pvt) 3425 { 3426 u8 nid = pvt->mc_node_id; 3427 struct amd64_umc *umc; 3428 u32 i, umc_base; 3429 3430 /* Read registers from each UMC */ 3431 for_each_umc(i) { 3432 3433 umc_base = get_umc_base(i); 3434 umc = &pvt->umc[i]; 3435 3436 amd_smn_read(nid, umc_base + get_umc_reg(UMCCH_DIMM_CFG), &umc->dimm_cfg); 3437 amd_smn_read(nid, umc_base + UMCCH_UMC_CFG, &umc->umc_cfg); 3438 amd_smn_read(nid, umc_base + UMCCH_SDP_CTRL, &umc->sdp_ctrl); 3439 amd_smn_read(nid, umc_base + UMCCH_ECC_CTRL, &umc->ecc_ctrl); 3440 amd_smn_read(nid, umc_base + UMCCH_UMC_CAP_HI, &umc->umc_cap_hi); 3441 } 3442 } 3443 3444 /* 3445 * Retrieve the hardware registers of the memory controller (this includes the 3446 * 'Address Map' and 'Misc' device regs) 3447 */ 3448 static void read_mc_regs(struct amd64_pvt *pvt) 3449 { 3450 unsigned int range; 3451 u64 msr_val; 3452 3453 /* 3454 * Retrieve TOP_MEM and TOP_MEM2; no masking off of reserved bits since 3455 * those are Read-As-Zero. 3456 */ 3457 rdmsrl(MSR_K8_TOP_MEM1, pvt->top_mem); 3458 edac_dbg(0, " TOP_MEM: 0x%016llx\n", pvt->top_mem); 3459 3460 /* Check first whether TOP_MEM2 is enabled: */ 3461 rdmsrl(MSR_AMD64_SYSCFG, msr_val); 3462 if (msr_val & BIT(21)) { 3463 rdmsrl(MSR_K8_TOP_MEM2, pvt->top_mem2); 3464 edac_dbg(0, " TOP_MEM2: 0x%016llx\n", pvt->top_mem2); 3465 } else { 3466 edac_dbg(0, " TOP_MEM2 disabled\n"); 3467 } 3468 3469 if (pvt->umc) { 3470 __read_mc_regs_df(pvt); 3471 amd64_read_pci_cfg(pvt->F0, DF_DHAR, &pvt->dhar); 3472 3473 goto skip; 3474 } 3475 3476 amd64_read_pci_cfg(pvt->F3, NBCAP, &pvt->nbcap); 3477 3478 read_dram_ctl_register(pvt); 3479 3480 for (range = 0; range < DRAM_RANGES; range++) { 3481 u8 rw; 3482 3483 /* read settings for this DRAM range */ 3484 read_dram_base_limit_regs(pvt, range); 3485 3486 rw = dram_rw(pvt, range); 3487 if (!rw) 3488 continue; 3489 3490 edac_dbg(1, " DRAM range[%d], base: 0x%016llx; limit: 0x%016llx\n", 3491 range, 3492 get_dram_base(pvt, range), 3493 get_dram_limit(pvt, range)); 3494 3495 edac_dbg(1, " IntlvEn=%s; Range access: %s%s IntlvSel=%d DstNode=%d\n", 3496 dram_intlv_en(pvt, range) ? "Enabled" : "Disabled", 3497 (rw & 0x1) ? "R" : "-", 3498 (rw & 0x2) ? "W" : "-", 3499 dram_intlv_sel(pvt, range), 3500 dram_dst_node(pvt, range)); 3501 } 3502 3503 amd64_read_pci_cfg(pvt->F1, DHAR, &pvt->dhar); 3504 amd64_read_dct_pci_cfg(pvt, 0, DBAM0, &pvt->dbam0); 3505 3506 amd64_read_pci_cfg(pvt->F3, F10_ONLINE_SPARE, &pvt->online_spare); 3507 3508 amd64_read_dct_pci_cfg(pvt, 0, DCLR0, &pvt->dclr0); 3509 amd64_read_dct_pci_cfg(pvt, 0, DCHR0, &pvt->dchr0); 3510 3511 if (!dct_ganging_enabled(pvt)) { 3512 amd64_read_dct_pci_cfg(pvt, 1, DCLR0, &pvt->dclr1); 3513 amd64_read_dct_pci_cfg(pvt, 1, DCHR0, &pvt->dchr1); 3514 } 3515 3516 skip: 3517 read_dct_base_mask(pvt); 3518 3519 determine_memory_type(pvt); 3520 3521 if (!pvt->umc) 3522 edac_dbg(1, " DIMM type: %s\n", edac_mem_types[pvt->dram_type]); 3523 3524 determine_ecc_sym_sz(pvt); 3525 } 3526 3527 /* 3528 * NOTE: CPU Revision Dependent code 3529 * 3530 * Input: 3531 * @csrow_nr ChipSelect Row Number (0..NUM_CHIPSELECTS-1) 3532 * k8 private pointer to --> 3533 * DRAM Bank Address mapping register 3534 * node_id 3535 * DCL register where dual_channel_active is 3536 * 3537 * The DBAM register consists of 4 sets of 4 bits each definitions: 3538 * 3539 * Bits: CSROWs 3540 * 0-3 CSROWs 0 and 1 3541 * 4-7 CSROWs 2 and 3 3542 * 8-11 CSROWs 4 and 5 3543 * 12-15 CSROWs 6 and 7 3544 * 3545 * Values range from: 0 to 15 3546 * The meaning of the values depends on CPU revision and dual-channel state, 3547 * see relevant BKDG more info. 3548 * 3549 * The memory controller provides for total of only 8 CSROWs in its current 3550 * architecture. Each "pair" of CSROWs normally represents just one DIMM in 3551 * single channel or two (2) DIMMs in dual channel mode. 3552 * 3553 * The following code logic collapses the various tables for CSROW based on CPU 3554 * revision. 3555 * 3556 * Returns: 3557 * The number of PAGE_SIZE pages on the specified CSROW number it 3558 * encompasses 3559 * 3560 */ 3561 static u32 get_csrow_nr_pages(struct amd64_pvt *pvt, u8 dct, int csrow_nr_orig) 3562 { 3563 u32 dbam = dct ? pvt->dbam1 : pvt->dbam0; 3564 int csrow_nr = csrow_nr_orig; 3565 u32 cs_mode, nr_pages; 3566 3567 if (!pvt->umc) { 3568 csrow_nr >>= 1; 3569 cs_mode = DBAM_DIMM(csrow_nr, dbam); 3570 } else { 3571 cs_mode = f17_get_cs_mode(csrow_nr >> 1, dct, pvt); 3572 } 3573 3574 nr_pages = pvt->ops->dbam_to_cs(pvt, dct, cs_mode, csrow_nr); 3575 nr_pages <<= 20 - PAGE_SHIFT; 3576 3577 edac_dbg(0, "csrow: %d, channel: %d, DBAM idx: %d\n", 3578 csrow_nr_orig, dct, cs_mode); 3579 edac_dbg(0, "nr_pages/channel: %u\n", nr_pages); 3580 3581 return nr_pages; 3582 } 3583 3584 static int init_csrows_df(struct mem_ctl_info *mci) 3585 { 3586 struct amd64_pvt *pvt = mci->pvt_info; 3587 enum edac_type edac_mode = EDAC_NONE; 3588 enum dev_type dev_type = DEV_UNKNOWN; 3589 struct dimm_info *dimm; 3590 int empty = 1; 3591 u8 umc, cs; 3592 3593 if (mci->edac_ctl_cap & EDAC_FLAG_S16ECD16ED) { 3594 edac_mode = EDAC_S16ECD16ED; 3595 dev_type = DEV_X16; 3596 } else if (mci->edac_ctl_cap & EDAC_FLAG_S8ECD8ED) { 3597 edac_mode = EDAC_S8ECD8ED; 3598 dev_type = DEV_X8; 3599 } else if (mci->edac_ctl_cap & EDAC_FLAG_S4ECD4ED) { 3600 edac_mode = EDAC_S4ECD4ED; 3601 dev_type = DEV_X4; 3602 } else if (mci->edac_ctl_cap & EDAC_FLAG_SECDED) { 3603 edac_mode = EDAC_SECDED; 3604 } 3605 3606 for_each_umc(umc) { 3607 for_each_chip_select(cs, umc, pvt) { 3608 if (!csrow_enabled(cs, umc, pvt)) 3609 continue; 3610 3611 empty = 0; 3612 dimm = mci->csrows[cs]->channels[umc]->dimm; 3613 3614 edac_dbg(1, "MC node: %d, csrow: %d\n", 3615 pvt->mc_node_id, cs); 3616 3617 dimm->nr_pages = get_csrow_nr_pages(pvt, umc, cs); 3618 dimm->mtype = pvt->umc[umc].dram_type; 3619 dimm->edac_mode = edac_mode; 3620 dimm->dtype = dev_type; 3621 dimm->grain = 64; 3622 } 3623 } 3624 3625 return empty; 3626 } 3627 3628 /* 3629 * Initialize the array of csrow attribute instances, based on the values 3630 * from pci config hardware registers. 3631 */ 3632 static int init_csrows(struct mem_ctl_info *mci) 3633 { 3634 struct amd64_pvt *pvt = mci->pvt_info; 3635 enum edac_type edac_mode = EDAC_NONE; 3636 struct csrow_info *csrow; 3637 struct dimm_info *dimm; 3638 int i, j, empty = 1; 3639 int nr_pages = 0; 3640 u32 val; 3641 3642 if (pvt->umc) 3643 return init_csrows_df(mci); 3644 3645 amd64_read_pci_cfg(pvt->F3, NBCFG, &val); 3646 3647 pvt->nbcfg = val; 3648 3649 edac_dbg(0, "node %d, NBCFG=0x%08x[ChipKillEccCap: %d|DramEccEn: %d]\n", 3650 pvt->mc_node_id, val, 3651 !!(val & NBCFG_CHIPKILL), !!(val & NBCFG_ECC_ENABLE)); 3652 3653 /* 3654 * We iterate over DCT0 here but we look at DCT1 in parallel, if needed. 3655 */ 3656 for_each_chip_select(i, 0, pvt) { 3657 bool row_dct0 = !!csrow_enabled(i, 0, pvt); 3658 bool row_dct1 = false; 3659 3660 if (pvt->fam != 0xf) 3661 row_dct1 = !!csrow_enabled(i, 1, pvt); 3662 3663 if (!row_dct0 && !row_dct1) 3664 continue; 3665 3666 csrow = mci->csrows[i]; 3667 empty = 0; 3668 3669 edac_dbg(1, "MC node: %d, csrow: %d\n", 3670 pvt->mc_node_id, i); 3671 3672 if (row_dct0) { 3673 nr_pages = get_csrow_nr_pages(pvt, 0, i); 3674 csrow->channels[0]->dimm->nr_pages = nr_pages; 3675 } 3676 3677 /* K8 has only one DCT */ 3678 if (pvt->fam != 0xf && row_dct1) { 3679 int row_dct1_pages = get_csrow_nr_pages(pvt, 1, i); 3680 3681 csrow->channels[1]->dimm->nr_pages = row_dct1_pages; 3682 nr_pages += row_dct1_pages; 3683 } 3684 3685 edac_dbg(1, "Total csrow%d pages: %u\n", i, nr_pages); 3686 3687 /* Determine DIMM ECC mode: */ 3688 if (pvt->nbcfg & NBCFG_ECC_ENABLE) { 3689 edac_mode = (pvt->nbcfg & NBCFG_CHIPKILL) 3690 ? EDAC_S4ECD4ED 3691 : EDAC_SECDED; 3692 } 3693 3694 for (j = 0; j < pvt->channel_count; j++) { 3695 dimm = csrow->channels[j]->dimm; 3696 dimm->mtype = pvt->dram_type; 3697 dimm->edac_mode = edac_mode; 3698 dimm->grain = 64; 3699 } 3700 } 3701 3702 return empty; 3703 } 3704 3705 /* get all cores on this DCT */ 3706 static void get_cpus_on_this_dct_cpumask(struct cpumask *mask, u16 nid) 3707 { 3708 int cpu; 3709 3710 for_each_online_cpu(cpu) 3711 if (topology_die_id(cpu) == nid) 3712 cpumask_set_cpu(cpu, mask); 3713 } 3714 3715 /* check MCG_CTL on all the cpus on this node */ 3716 static bool nb_mce_bank_enabled_on_node(u16 nid) 3717 { 3718 cpumask_var_t mask; 3719 int cpu, nbe; 3720 bool ret = false; 3721 3722 if (!zalloc_cpumask_var(&mask, GFP_KERNEL)) { 3723 amd64_warn("%s: Error allocating mask\n", __func__); 3724 return false; 3725 } 3726 3727 get_cpus_on_this_dct_cpumask(mask, nid); 3728 3729 rdmsr_on_cpus(mask, MSR_IA32_MCG_CTL, msrs); 3730 3731 for_each_cpu(cpu, mask) { 3732 struct msr *reg = per_cpu_ptr(msrs, cpu); 3733 nbe = reg->l & MSR_MCGCTL_NBE; 3734 3735 edac_dbg(0, "core: %u, MCG_CTL: 0x%llx, NB MSR is %s\n", 3736 cpu, reg->q, 3737 (nbe ? "enabled" : "disabled")); 3738 3739 if (!nbe) 3740 goto out; 3741 } 3742 ret = true; 3743 3744 out: 3745 free_cpumask_var(mask); 3746 return ret; 3747 } 3748 3749 static int toggle_ecc_err_reporting(struct ecc_settings *s, u16 nid, bool on) 3750 { 3751 cpumask_var_t cmask; 3752 int cpu; 3753 3754 if (!zalloc_cpumask_var(&cmask, GFP_KERNEL)) { 3755 amd64_warn("%s: error allocating mask\n", __func__); 3756 return -ENOMEM; 3757 } 3758 3759 get_cpus_on_this_dct_cpumask(cmask, nid); 3760 3761 rdmsr_on_cpus(cmask, MSR_IA32_MCG_CTL, msrs); 3762 3763 for_each_cpu(cpu, cmask) { 3764 3765 struct msr *reg = per_cpu_ptr(msrs, cpu); 3766 3767 if (on) { 3768 if (reg->l & MSR_MCGCTL_NBE) 3769 s->flags.nb_mce_enable = 1; 3770 3771 reg->l |= MSR_MCGCTL_NBE; 3772 } else { 3773 /* 3774 * Turn off NB MCE reporting only when it was off before 3775 */ 3776 if (!s->flags.nb_mce_enable) 3777 reg->l &= ~MSR_MCGCTL_NBE; 3778 } 3779 } 3780 wrmsr_on_cpus(cmask, MSR_IA32_MCG_CTL, msrs); 3781 3782 free_cpumask_var(cmask); 3783 3784 return 0; 3785 } 3786 3787 static bool enable_ecc_error_reporting(struct ecc_settings *s, u16 nid, 3788 struct pci_dev *F3) 3789 { 3790 bool ret = true; 3791 u32 value, mask = 0x3; /* UECC/CECC enable */ 3792 3793 if (toggle_ecc_err_reporting(s, nid, ON)) { 3794 amd64_warn("Error enabling ECC reporting over MCGCTL!\n"); 3795 return false; 3796 } 3797 3798 amd64_read_pci_cfg(F3, NBCTL, &value); 3799 3800 s->old_nbctl = value & mask; 3801 s->nbctl_valid = true; 3802 3803 value |= mask; 3804 amd64_write_pci_cfg(F3, NBCTL, value); 3805 3806 amd64_read_pci_cfg(F3, NBCFG, &value); 3807 3808 edac_dbg(0, "1: node %d, NBCFG=0x%08x[DramEccEn: %d]\n", 3809 nid, value, !!(value & NBCFG_ECC_ENABLE)); 3810 3811 if (!(value & NBCFG_ECC_ENABLE)) { 3812 amd64_warn("DRAM ECC disabled on this node, enabling...\n"); 3813 3814 s->flags.nb_ecc_prev = 0; 3815 3816 /* Attempt to turn on DRAM ECC Enable */ 3817 value |= NBCFG_ECC_ENABLE; 3818 amd64_write_pci_cfg(F3, NBCFG, value); 3819 3820 amd64_read_pci_cfg(F3, NBCFG, &value); 3821 3822 if (!(value & NBCFG_ECC_ENABLE)) { 3823 amd64_warn("Hardware rejected DRAM ECC enable," 3824 "check memory DIMM configuration.\n"); 3825 ret = false; 3826 } else { 3827 amd64_info("Hardware accepted DRAM ECC Enable\n"); 3828 } 3829 } else { 3830 s->flags.nb_ecc_prev = 1; 3831 } 3832 3833 edac_dbg(0, "2: node %d, NBCFG=0x%08x[DramEccEn: %d]\n", 3834 nid, value, !!(value & NBCFG_ECC_ENABLE)); 3835 3836 return ret; 3837 } 3838 3839 static void restore_ecc_error_reporting(struct ecc_settings *s, u16 nid, 3840 struct pci_dev *F3) 3841 { 3842 u32 value, mask = 0x3; /* UECC/CECC enable */ 3843 3844 if (!s->nbctl_valid) 3845 return; 3846 3847 amd64_read_pci_cfg(F3, NBCTL, &value); 3848 value &= ~mask; 3849 value |= s->old_nbctl; 3850 3851 amd64_write_pci_cfg(F3, NBCTL, value); 3852 3853 /* restore previous BIOS DRAM ECC "off" setting we force-enabled */ 3854 if (!s->flags.nb_ecc_prev) { 3855 amd64_read_pci_cfg(F3, NBCFG, &value); 3856 value &= ~NBCFG_ECC_ENABLE; 3857 amd64_write_pci_cfg(F3, NBCFG, value); 3858 } 3859 3860 /* restore the NB Enable MCGCTL bit */ 3861 if (toggle_ecc_err_reporting(s, nid, OFF)) 3862 amd64_warn("Error restoring NB MCGCTL settings!\n"); 3863 } 3864 3865 static bool ecc_enabled(struct amd64_pvt *pvt) 3866 { 3867 u16 nid = pvt->mc_node_id; 3868 bool nb_mce_en = false; 3869 u8 ecc_en = 0, i; 3870 u32 value; 3871 3872 if (boot_cpu_data.x86 >= 0x17) { 3873 u8 umc_en_mask = 0, ecc_en_mask = 0; 3874 struct amd64_umc *umc; 3875 3876 for_each_umc(i) { 3877 umc = &pvt->umc[i]; 3878 3879 /* Only check enabled UMCs. */ 3880 if (!(umc->sdp_ctrl & UMC_SDP_INIT)) 3881 continue; 3882 3883 umc_en_mask |= BIT(i); 3884 3885 if (umc->umc_cap_hi & UMC_ECC_ENABLED) 3886 ecc_en_mask |= BIT(i); 3887 } 3888 3889 /* Check whether at least one UMC is enabled: */ 3890 if (umc_en_mask) 3891 ecc_en = umc_en_mask == ecc_en_mask; 3892 else 3893 edac_dbg(0, "Node %d: No enabled UMCs.\n", nid); 3894 3895 /* Assume UMC MCA banks are enabled. */ 3896 nb_mce_en = true; 3897 } else { 3898 amd64_read_pci_cfg(pvt->F3, NBCFG, &value); 3899 3900 ecc_en = !!(value & NBCFG_ECC_ENABLE); 3901 3902 nb_mce_en = nb_mce_bank_enabled_on_node(nid); 3903 if (!nb_mce_en) 3904 edac_dbg(0, "NB MCE bank disabled, set MSR 0x%08x[4] on node %d to enable.\n", 3905 MSR_IA32_MCG_CTL, nid); 3906 } 3907 3908 edac_dbg(3, "Node %d: DRAM ECC %s.\n", nid, (ecc_en ? "enabled" : "disabled")); 3909 3910 if (!ecc_en || !nb_mce_en) 3911 return false; 3912 else 3913 return true; 3914 } 3915 3916 static inline void 3917 f17h_determine_edac_ctl_cap(struct mem_ctl_info *mci, struct amd64_pvt *pvt) 3918 { 3919 u8 i, ecc_en = 1, cpk_en = 1, dev_x4 = 1, dev_x16 = 1; 3920 3921 for_each_umc(i) { 3922 if (pvt->umc[i].sdp_ctrl & UMC_SDP_INIT) { 3923 ecc_en &= !!(pvt->umc[i].umc_cap_hi & UMC_ECC_ENABLED); 3924 cpk_en &= !!(pvt->umc[i].umc_cap_hi & UMC_ECC_CHIPKILL_CAP); 3925 3926 dev_x4 &= !!(pvt->umc[i].dimm_cfg & BIT(6)); 3927 dev_x16 &= !!(pvt->umc[i].dimm_cfg & BIT(7)); 3928 } 3929 } 3930 3931 /* Set chipkill only if ECC is enabled: */ 3932 if (ecc_en) { 3933 mci->edac_ctl_cap |= EDAC_FLAG_SECDED; 3934 3935 if (!cpk_en) 3936 return; 3937 3938 if (dev_x4) 3939 mci->edac_ctl_cap |= EDAC_FLAG_S4ECD4ED; 3940 else if (dev_x16) 3941 mci->edac_ctl_cap |= EDAC_FLAG_S16ECD16ED; 3942 else 3943 mci->edac_ctl_cap |= EDAC_FLAG_S8ECD8ED; 3944 } 3945 } 3946 3947 static void setup_mci_misc_attrs(struct mem_ctl_info *mci) 3948 { 3949 struct amd64_pvt *pvt = mci->pvt_info; 3950 3951 mci->mtype_cap = MEM_FLAG_DDR2 | MEM_FLAG_RDDR2; 3952 mci->edac_ctl_cap = EDAC_FLAG_NONE; 3953 3954 if (pvt->umc) { 3955 f17h_determine_edac_ctl_cap(mci, pvt); 3956 } else { 3957 if (pvt->nbcap & NBCAP_SECDED) 3958 mci->edac_ctl_cap |= EDAC_FLAG_SECDED; 3959 3960 if (pvt->nbcap & NBCAP_CHIPKILL) 3961 mci->edac_ctl_cap |= EDAC_FLAG_S4ECD4ED; 3962 } 3963 3964 mci->edac_cap = determine_edac_cap(pvt); 3965 mci->mod_name = EDAC_MOD_STR; 3966 mci->ctl_name = fam_type->ctl_name; 3967 mci->dev_name = pci_name(pvt->F3); 3968 mci->ctl_page_to_phys = NULL; 3969 3970 /* memory scrubber interface */ 3971 mci->set_sdram_scrub_rate = set_scrub_rate; 3972 mci->get_sdram_scrub_rate = get_scrub_rate; 3973 } 3974 3975 /* 3976 * returns a pointer to the family descriptor on success, NULL otherwise. 3977 */ 3978 static struct amd64_family_type *per_family_init(struct amd64_pvt *pvt) 3979 { 3980 pvt->ext_model = boot_cpu_data.x86_model >> 4; 3981 pvt->stepping = boot_cpu_data.x86_stepping; 3982 pvt->model = boot_cpu_data.x86_model; 3983 pvt->fam = boot_cpu_data.x86; 3984 3985 switch (pvt->fam) { 3986 case 0xf: 3987 fam_type = &family_types[K8_CPUS]; 3988 pvt->ops = &family_types[K8_CPUS].ops; 3989 break; 3990 3991 case 0x10: 3992 fam_type = &family_types[F10_CPUS]; 3993 pvt->ops = &family_types[F10_CPUS].ops; 3994 break; 3995 3996 case 0x15: 3997 if (pvt->model == 0x30) { 3998 fam_type = &family_types[F15_M30H_CPUS]; 3999 pvt->ops = &family_types[F15_M30H_CPUS].ops; 4000 break; 4001 } else if (pvt->model == 0x60) { 4002 fam_type = &family_types[F15_M60H_CPUS]; 4003 pvt->ops = &family_types[F15_M60H_CPUS].ops; 4004 break; 4005 /* Richland is only client */ 4006 } else if (pvt->model == 0x13) { 4007 return NULL; 4008 } else { 4009 fam_type = &family_types[F15_CPUS]; 4010 pvt->ops = &family_types[F15_CPUS].ops; 4011 } 4012 break; 4013 4014 case 0x16: 4015 if (pvt->model == 0x30) { 4016 fam_type = &family_types[F16_M30H_CPUS]; 4017 pvt->ops = &family_types[F16_M30H_CPUS].ops; 4018 break; 4019 } 4020 fam_type = &family_types[F16_CPUS]; 4021 pvt->ops = &family_types[F16_CPUS].ops; 4022 break; 4023 4024 case 0x17: 4025 if (pvt->model >= 0x10 && pvt->model <= 0x2f) { 4026 fam_type = &family_types[F17_M10H_CPUS]; 4027 pvt->ops = &family_types[F17_M10H_CPUS].ops; 4028 break; 4029 } else if (pvt->model >= 0x30 && pvt->model <= 0x3f) { 4030 fam_type = &family_types[F17_M30H_CPUS]; 4031 pvt->ops = &family_types[F17_M30H_CPUS].ops; 4032 break; 4033 } else if (pvt->model >= 0x60 && pvt->model <= 0x6f) { 4034 fam_type = &family_types[F17_M60H_CPUS]; 4035 pvt->ops = &family_types[F17_M60H_CPUS].ops; 4036 break; 4037 } else if (pvt->model >= 0x70 && pvt->model <= 0x7f) { 4038 fam_type = &family_types[F17_M70H_CPUS]; 4039 pvt->ops = &family_types[F17_M70H_CPUS].ops; 4040 break; 4041 } 4042 fallthrough; 4043 case 0x18: 4044 fam_type = &family_types[F17_CPUS]; 4045 pvt->ops = &family_types[F17_CPUS].ops; 4046 4047 if (pvt->fam == 0x18) 4048 family_types[F17_CPUS].ctl_name = "F18h"; 4049 break; 4050 4051 case 0x19: 4052 if (pvt->model >= 0x10 && pvt->model <= 0x1f) { 4053 fam_type = &family_types[F19_M10H_CPUS]; 4054 pvt->ops = &family_types[F19_M10H_CPUS].ops; 4055 break; 4056 } else if (pvt->model >= 0x20 && pvt->model <= 0x2f) { 4057 fam_type = &family_types[F17_M70H_CPUS]; 4058 pvt->ops = &family_types[F17_M70H_CPUS].ops; 4059 fam_type->ctl_name = "F19h_M20h"; 4060 break; 4061 } else if (pvt->model >= 0x50 && pvt->model <= 0x5f) { 4062 fam_type = &family_types[F19_M50H_CPUS]; 4063 pvt->ops = &family_types[F19_M50H_CPUS].ops; 4064 fam_type->ctl_name = "F19h_M50h"; 4065 break; 4066 } else if (pvt->model >= 0xa0 && pvt->model <= 0xaf) { 4067 fam_type = &family_types[F19_M10H_CPUS]; 4068 pvt->ops = &family_types[F19_M10H_CPUS].ops; 4069 fam_type->ctl_name = "F19h_MA0h"; 4070 break; 4071 } 4072 fam_type = &family_types[F19_CPUS]; 4073 pvt->ops = &family_types[F19_CPUS].ops; 4074 family_types[F19_CPUS].ctl_name = "F19h"; 4075 break; 4076 4077 default: 4078 amd64_err("Unsupported family!\n"); 4079 return NULL; 4080 } 4081 4082 return fam_type; 4083 } 4084 4085 static const struct attribute_group *amd64_edac_attr_groups[] = { 4086 #ifdef CONFIG_EDAC_DEBUG 4087 &dbg_group, 4088 &inj_group, 4089 #endif 4090 NULL 4091 }; 4092 4093 static int hw_info_get(struct amd64_pvt *pvt) 4094 { 4095 u16 pci_id1, pci_id2; 4096 int ret; 4097 4098 if (pvt->fam >= 0x17) { 4099 pvt->umc = kcalloc(fam_type->max_mcs, sizeof(struct amd64_umc), GFP_KERNEL); 4100 if (!pvt->umc) 4101 return -ENOMEM; 4102 4103 pci_id1 = fam_type->f0_id; 4104 pci_id2 = fam_type->f6_id; 4105 } else { 4106 pci_id1 = fam_type->f1_id; 4107 pci_id2 = fam_type->f2_id; 4108 } 4109 4110 ret = reserve_mc_sibling_devs(pvt, pci_id1, pci_id2); 4111 if (ret) 4112 return ret; 4113 4114 read_mc_regs(pvt); 4115 4116 return 0; 4117 } 4118 4119 static void hw_info_put(struct amd64_pvt *pvt) 4120 { 4121 if (pvt->F0 || pvt->F1) 4122 free_mc_sibling_devs(pvt); 4123 4124 kfree(pvt->umc); 4125 } 4126 4127 static int init_one_instance(struct amd64_pvt *pvt) 4128 { 4129 struct mem_ctl_info *mci = NULL; 4130 struct edac_mc_layer layers[2]; 4131 int ret = -EINVAL; 4132 4133 /* 4134 * We need to determine how many memory channels there are. Then use 4135 * that information for calculating the size of the dynamic instance 4136 * tables in the 'mci' structure. 4137 */ 4138 pvt->channel_count = pvt->ops->early_channel_count(pvt); 4139 if (pvt->channel_count < 0) 4140 return ret; 4141 4142 ret = -ENOMEM; 4143 layers[0].type = EDAC_MC_LAYER_CHIP_SELECT; 4144 layers[0].size = pvt->csels[0].b_cnt; 4145 layers[0].is_virt_csrow = true; 4146 layers[1].type = EDAC_MC_LAYER_CHANNEL; 4147 4148 /* 4149 * Always allocate two channels since we can have setups with DIMMs on 4150 * only one channel. Also, this simplifies handling later for the price 4151 * of a couple of KBs tops. 4152 */ 4153 layers[1].size = fam_type->max_mcs; 4154 layers[1].is_virt_csrow = false; 4155 4156 mci = edac_mc_alloc(pvt->mc_node_id, ARRAY_SIZE(layers), layers, 0); 4157 if (!mci) 4158 return ret; 4159 4160 mci->pvt_info = pvt; 4161 mci->pdev = &pvt->F3->dev; 4162 4163 setup_mci_misc_attrs(mci); 4164 4165 if (init_csrows(mci)) 4166 mci->edac_cap = EDAC_FLAG_NONE; 4167 4168 ret = -ENODEV; 4169 if (edac_mc_add_mc_with_groups(mci, amd64_edac_attr_groups)) { 4170 edac_dbg(1, "failed edac_mc_add_mc()\n"); 4171 edac_mc_free(mci); 4172 return ret; 4173 } 4174 4175 return 0; 4176 } 4177 4178 static bool instance_has_memory(struct amd64_pvt *pvt) 4179 { 4180 bool cs_enabled = false; 4181 int cs = 0, dct = 0; 4182 4183 for (dct = 0; dct < fam_type->max_mcs; dct++) { 4184 for_each_chip_select(cs, dct, pvt) 4185 cs_enabled |= csrow_enabled(cs, dct, pvt); 4186 } 4187 4188 return cs_enabled; 4189 } 4190 4191 static int probe_one_instance(unsigned int nid) 4192 { 4193 struct pci_dev *F3 = node_to_amd_nb(nid)->misc; 4194 struct amd64_pvt *pvt = NULL; 4195 struct ecc_settings *s; 4196 int ret; 4197 4198 ret = -ENOMEM; 4199 s = kzalloc(sizeof(struct ecc_settings), GFP_KERNEL); 4200 if (!s) 4201 goto err_out; 4202 4203 ecc_stngs[nid] = s; 4204 4205 pvt = kzalloc(sizeof(struct amd64_pvt), GFP_KERNEL); 4206 if (!pvt) 4207 goto err_settings; 4208 4209 pvt->mc_node_id = nid; 4210 pvt->F3 = F3; 4211 4212 ret = -ENODEV; 4213 fam_type = per_family_init(pvt); 4214 if (!fam_type) 4215 goto err_enable; 4216 4217 ret = hw_info_get(pvt); 4218 if (ret < 0) 4219 goto err_enable; 4220 4221 ret = 0; 4222 if (!instance_has_memory(pvt)) { 4223 amd64_info("Node %d: No DIMMs detected.\n", nid); 4224 goto err_enable; 4225 } 4226 4227 if (!ecc_enabled(pvt)) { 4228 ret = -ENODEV; 4229 4230 if (!ecc_enable_override) 4231 goto err_enable; 4232 4233 if (boot_cpu_data.x86 >= 0x17) { 4234 amd64_warn("Forcing ECC on is not recommended on newer systems. Please enable ECC in BIOS."); 4235 goto err_enable; 4236 } else 4237 amd64_warn("Forcing ECC on!\n"); 4238 4239 if (!enable_ecc_error_reporting(s, nid, F3)) 4240 goto err_enable; 4241 } 4242 4243 ret = init_one_instance(pvt); 4244 if (ret < 0) { 4245 amd64_err("Error probing instance: %d\n", nid); 4246 4247 if (boot_cpu_data.x86 < 0x17) 4248 restore_ecc_error_reporting(s, nid, F3); 4249 4250 goto err_enable; 4251 } 4252 4253 amd64_info("%s %sdetected (node %d).\n", fam_type->ctl_name, 4254 (pvt->fam == 0xf ? 4255 (pvt->ext_model >= K8_REV_F ? "revF or later " 4256 : "revE or earlier ") 4257 : ""), pvt->mc_node_id); 4258 4259 dump_misc_regs(pvt); 4260 4261 return ret; 4262 4263 err_enable: 4264 hw_info_put(pvt); 4265 kfree(pvt); 4266 4267 err_settings: 4268 kfree(s); 4269 ecc_stngs[nid] = NULL; 4270 4271 err_out: 4272 return ret; 4273 } 4274 4275 static void remove_one_instance(unsigned int nid) 4276 { 4277 struct pci_dev *F3 = node_to_amd_nb(nid)->misc; 4278 struct ecc_settings *s = ecc_stngs[nid]; 4279 struct mem_ctl_info *mci; 4280 struct amd64_pvt *pvt; 4281 4282 /* Remove from EDAC CORE tracking list */ 4283 mci = edac_mc_del_mc(&F3->dev); 4284 if (!mci) 4285 return; 4286 4287 pvt = mci->pvt_info; 4288 4289 restore_ecc_error_reporting(s, nid, F3); 4290 4291 kfree(ecc_stngs[nid]); 4292 ecc_stngs[nid] = NULL; 4293 4294 /* Free the EDAC CORE resources */ 4295 mci->pvt_info = NULL; 4296 4297 hw_info_put(pvt); 4298 kfree(pvt); 4299 edac_mc_free(mci); 4300 } 4301 4302 static void setup_pci_device(void) 4303 { 4304 if (pci_ctl) 4305 return; 4306 4307 pci_ctl = edac_pci_create_generic_ctl(pci_ctl_dev, EDAC_MOD_STR); 4308 if (!pci_ctl) { 4309 pr_warn("%s(): Unable to create PCI control\n", __func__); 4310 pr_warn("%s(): PCI error report via EDAC not set\n", __func__); 4311 } 4312 } 4313 4314 static const struct x86_cpu_id amd64_cpuids[] = { 4315 X86_MATCH_VENDOR_FAM(AMD, 0x0F, NULL), 4316 X86_MATCH_VENDOR_FAM(AMD, 0x10, NULL), 4317 X86_MATCH_VENDOR_FAM(AMD, 0x15, NULL), 4318 X86_MATCH_VENDOR_FAM(AMD, 0x16, NULL), 4319 X86_MATCH_VENDOR_FAM(AMD, 0x17, NULL), 4320 X86_MATCH_VENDOR_FAM(HYGON, 0x18, NULL), 4321 X86_MATCH_VENDOR_FAM(AMD, 0x19, NULL), 4322 { } 4323 }; 4324 MODULE_DEVICE_TABLE(x86cpu, amd64_cpuids); 4325 4326 static int __init amd64_edac_init(void) 4327 { 4328 const char *owner; 4329 int err = -ENODEV; 4330 int i; 4331 4332 if (ghes_get_devices()) 4333 return -EBUSY; 4334 4335 owner = edac_get_owner(); 4336 if (owner && strncmp(owner, EDAC_MOD_STR, sizeof(EDAC_MOD_STR))) 4337 return -EBUSY; 4338 4339 if (!x86_match_cpu(amd64_cpuids)) 4340 return -ENODEV; 4341 4342 if (!amd_nb_num()) 4343 return -ENODEV; 4344 4345 opstate_init(); 4346 4347 err = -ENOMEM; 4348 ecc_stngs = kcalloc(amd_nb_num(), sizeof(ecc_stngs[0]), GFP_KERNEL); 4349 if (!ecc_stngs) 4350 goto err_free; 4351 4352 msrs = msrs_alloc(); 4353 if (!msrs) 4354 goto err_free; 4355 4356 for (i = 0; i < amd_nb_num(); i++) { 4357 err = probe_one_instance(i); 4358 if (err) { 4359 /* unwind properly */ 4360 while (--i >= 0) 4361 remove_one_instance(i); 4362 4363 goto err_pci; 4364 } 4365 } 4366 4367 if (!edac_has_mcs()) { 4368 err = -ENODEV; 4369 goto err_pci; 4370 } 4371 4372 /* register stuff with EDAC MCE */ 4373 if (boot_cpu_data.x86 >= 0x17) 4374 amd_register_ecc_decoder(decode_umc_error); 4375 else 4376 amd_register_ecc_decoder(decode_bus_error); 4377 4378 setup_pci_device(); 4379 4380 #ifdef CONFIG_X86_32 4381 amd64_err("%s on 32-bit is unsupported. USE AT YOUR OWN RISK!\n", EDAC_MOD_STR); 4382 #endif 4383 4384 printk(KERN_INFO "AMD64 EDAC driver v%s\n", EDAC_AMD64_VERSION); 4385 4386 return 0; 4387 4388 err_pci: 4389 pci_ctl_dev = NULL; 4390 4391 msrs_free(msrs); 4392 msrs = NULL; 4393 4394 err_free: 4395 kfree(ecc_stngs); 4396 ecc_stngs = NULL; 4397 4398 return err; 4399 } 4400 4401 static void __exit amd64_edac_exit(void) 4402 { 4403 int i; 4404 4405 if (pci_ctl) 4406 edac_pci_release_generic_ctl(pci_ctl); 4407 4408 /* unregister from EDAC MCE */ 4409 if (boot_cpu_data.x86 >= 0x17) 4410 amd_unregister_ecc_decoder(decode_umc_error); 4411 else 4412 amd_unregister_ecc_decoder(decode_bus_error); 4413 4414 for (i = 0; i < amd_nb_num(); i++) 4415 remove_one_instance(i); 4416 4417 kfree(ecc_stngs); 4418 ecc_stngs = NULL; 4419 4420 pci_ctl_dev = NULL; 4421 4422 msrs_free(msrs); 4423 msrs = NULL; 4424 } 4425 4426 module_init(amd64_edac_init); 4427 module_exit(amd64_edac_exit); 4428 4429 MODULE_LICENSE("GPL"); 4430 MODULE_AUTHOR("SoftwareBitMaker: Doug Thompson, " 4431 "Dave Peterson, Thayne Harbaugh"); 4432 MODULE_DESCRIPTION("MC support for AMD64 memory controllers - " 4433 EDAC_AMD64_VERSION); 4434 4435 module_param(edac_op_state, int, 0444); 4436 MODULE_PARM_DESC(edac_op_state, "EDAC Error Reporting state: 0=Poll,1=NMI"); 4437