1 // SPDX-License-Identifier: GPL-2.0 2 /* 3 * Copyright (C) 2017 Pengutronix, Jan Luebbe <kernel@pengutronix.de> 4 */ 5 6 #include <linux/kernel.h> 7 #include <linux/edac.h> 8 #include <linux/of_platform.h> 9 10 #include <asm/hardware/cache-l2x0.h> 11 #include <asm/hardware/cache-aurora-l2.h> 12 13 #include "edac_mc.h" 14 #include "edac_device.h" 15 #include "edac_module.h" 16 17 /************************ EDAC MC (DDR RAM) ********************************/ 18 19 #define SDRAM_NUM_CS 4 20 21 #define SDRAM_CONFIG_REG 0x0 22 #define SDRAM_CONFIG_ECC_MASK BIT(18) 23 #define SDRAM_CONFIG_REGISTERED_MASK BIT(17) 24 #define SDRAM_CONFIG_BUS_WIDTH_MASK BIT(15) 25 26 #define SDRAM_ADDR_CTRL_REG 0x10 27 #define SDRAM_ADDR_CTRL_SIZE_HIGH_OFFSET(cs) (20+cs) 28 #define SDRAM_ADDR_CTRL_SIZE_HIGH_MASK(cs) (0x1 << SDRAM_ADDR_CTRL_SIZE_HIGH_OFFSET(cs)) 29 #define SDRAM_ADDR_CTRL_ADDR_SEL_MASK(cs) BIT(16+cs) 30 #define SDRAM_ADDR_CTRL_SIZE_LOW_OFFSET(cs) (cs*4+2) 31 #define SDRAM_ADDR_CTRL_SIZE_LOW_MASK(cs) (0x3 << SDRAM_ADDR_CTRL_SIZE_LOW_OFFSET(cs)) 32 #define SDRAM_ADDR_CTRL_STRUCT_OFFSET(cs) (cs*4) 33 #define SDRAM_ADDR_CTRL_STRUCT_MASK(cs) (0x3 << SDRAM_ADDR_CTRL_STRUCT_OFFSET(cs)) 34 35 #define SDRAM_ERR_DATA_H_REG 0x40 36 #define SDRAM_ERR_DATA_L_REG 0x44 37 38 #define SDRAM_ERR_RECV_ECC_REG 0x48 39 #define SDRAM_ERR_RECV_ECC_VALUE_MASK 0xff 40 41 #define SDRAM_ERR_CALC_ECC_REG 0x4c 42 #define SDRAM_ERR_CALC_ECC_ROW_OFFSET 8 43 #define SDRAM_ERR_CALC_ECC_ROW_MASK (0xffff << SDRAM_ERR_CALC_ECC_ROW_OFFSET) 44 #define SDRAM_ERR_CALC_ECC_VALUE_MASK 0xff 45 46 #define SDRAM_ERR_ADDR_REG 0x50 47 #define SDRAM_ERR_ADDR_BANK_OFFSET 23 48 #define SDRAM_ERR_ADDR_BANK_MASK (0x7 << SDRAM_ERR_ADDR_BANK_OFFSET) 49 #define SDRAM_ERR_ADDR_COL_OFFSET 8 50 #define SDRAM_ERR_ADDR_COL_MASK (0x7fff << SDRAM_ERR_ADDR_COL_OFFSET) 51 #define SDRAM_ERR_ADDR_CS_OFFSET 1 52 #define SDRAM_ERR_ADDR_CS_MASK (0x3 << SDRAM_ERR_ADDR_CS_OFFSET) 53 #define SDRAM_ERR_ADDR_TYPE_MASK BIT(0) 54 55 #define SDRAM_ERR_CTRL_REG 0x54 56 #define SDRAM_ERR_CTRL_THR_OFFSET 16 57 #define SDRAM_ERR_CTRL_THR_MASK (0xff << SDRAM_ERR_CTRL_THR_OFFSET) 58 #define SDRAM_ERR_CTRL_PROP_MASK BIT(9) 59 60 #define SDRAM_ERR_SBE_COUNT_REG 0x58 61 #define SDRAM_ERR_DBE_COUNT_REG 0x5c 62 63 #define SDRAM_ERR_CAUSE_ERR_REG 0xd0 64 #define SDRAM_ERR_CAUSE_MSG_REG 0xd8 65 #define SDRAM_ERR_CAUSE_DBE_MASK BIT(1) 66 #define SDRAM_ERR_CAUSE_SBE_MASK BIT(0) 67 68 #define SDRAM_RANK_CTRL_REG 0x1e0 69 #define SDRAM_RANK_CTRL_EXIST_MASK(cs) BIT(cs) 70 71 struct axp_mc_drvdata { 72 void __iomem *base; 73 /* width in bytes */ 74 unsigned int width; 75 /* bank interleaving */ 76 bool cs_addr_sel[SDRAM_NUM_CS]; 77 78 char msg[128]; 79 }; 80 81 /* derived from "DRAM Address Multiplexing" in the ARAMDA XP Functional Spec */ 82 static uint32_t axp_mc_calc_address(struct axp_mc_drvdata *drvdata, 83 uint8_t cs, uint8_t bank, uint16_t row, 84 uint16_t col) 85 { 86 if (drvdata->width == 8) { 87 /* 64 bit */ 88 if (drvdata->cs_addr_sel[cs]) 89 /* bank interleaved */ 90 return (((row & 0xfff8) << 16) | 91 ((bank & 0x7) << 16) | 92 ((row & 0x7) << 13) | 93 ((col & 0x3ff) << 3)); 94 else 95 return (((row & 0xffff << 16) | 96 ((bank & 0x7) << 13) | 97 ((col & 0x3ff)) << 3)); 98 } else if (drvdata->width == 4) { 99 /* 32 bit */ 100 if (drvdata->cs_addr_sel[cs]) 101 /* bank interleaved */ 102 return (((row & 0xfff0) << 15) | 103 ((bank & 0x7) << 16) | 104 ((row & 0xf) << 12) | 105 ((col & 0x3ff) << 2)); 106 else 107 return (((row & 0xffff << 15) | 108 ((bank & 0x7) << 12) | 109 ((col & 0x3ff)) << 2)); 110 } else { 111 /* 16 bit */ 112 if (drvdata->cs_addr_sel[cs]) 113 /* bank interleaved */ 114 return (((row & 0xffe0) << 14) | 115 ((bank & 0x7) << 16) | 116 ((row & 0x1f) << 11) | 117 ((col & 0x3ff) << 1)); 118 else 119 return (((row & 0xffff << 14) | 120 ((bank & 0x7) << 11) | 121 ((col & 0x3ff)) << 1)); 122 } 123 } 124 125 static void axp_mc_check(struct mem_ctl_info *mci) 126 { 127 struct axp_mc_drvdata *drvdata = mci->pvt_info; 128 uint32_t data_h, data_l, recv_ecc, calc_ecc, addr; 129 uint32_t cnt_sbe, cnt_dbe, cause_err, cause_msg; 130 uint32_t row_val, col_val, bank_val, addr_val; 131 uint8_t syndrome_val, cs_val; 132 char *msg = drvdata->msg; 133 134 data_h = readl(drvdata->base + SDRAM_ERR_DATA_H_REG); 135 data_l = readl(drvdata->base + SDRAM_ERR_DATA_L_REG); 136 recv_ecc = readl(drvdata->base + SDRAM_ERR_RECV_ECC_REG); 137 calc_ecc = readl(drvdata->base + SDRAM_ERR_CALC_ECC_REG); 138 addr = readl(drvdata->base + SDRAM_ERR_ADDR_REG); 139 cnt_sbe = readl(drvdata->base + SDRAM_ERR_SBE_COUNT_REG); 140 cnt_dbe = readl(drvdata->base + SDRAM_ERR_DBE_COUNT_REG); 141 cause_err = readl(drvdata->base + SDRAM_ERR_CAUSE_ERR_REG); 142 cause_msg = readl(drvdata->base + SDRAM_ERR_CAUSE_MSG_REG); 143 144 /* clear cause registers */ 145 writel(~(SDRAM_ERR_CAUSE_DBE_MASK | SDRAM_ERR_CAUSE_SBE_MASK), 146 drvdata->base + SDRAM_ERR_CAUSE_ERR_REG); 147 writel(~(SDRAM_ERR_CAUSE_DBE_MASK | SDRAM_ERR_CAUSE_SBE_MASK), 148 drvdata->base + SDRAM_ERR_CAUSE_MSG_REG); 149 150 /* clear error counter registers */ 151 if (cnt_sbe) 152 writel(0, drvdata->base + SDRAM_ERR_SBE_COUNT_REG); 153 if (cnt_dbe) 154 writel(0, drvdata->base + SDRAM_ERR_DBE_COUNT_REG); 155 156 if (!cnt_sbe && !cnt_dbe) 157 return; 158 159 if (!(addr & SDRAM_ERR_ADDR_TYPE_MASK)) { 160 if (cnt_sbe) 161 cnt_sbe--; 162 else 163 dev_warn(mci->pdev, "inconsistent SBE count detected"); 164 } else { 165 if (cnt_dbe) 166 cnt_dbe--; 167 else 168 dev_warn(mci->pdev, "inconsistent DBE count detected"); 169 } 170 171 /* report earlier errors */ 172 if (cnt_sbe) 173 edac_mc_handle_error(HW_EVENT_ERR_CORRECTED, mci, 174 cnt_sbe, /* error count */ 175 0, 0, 0, /* pfn, offset, syndrome */ 176 -1, -1, -1, /* top, mid, low layer */ 177 mci->ctl_name, 178 "details unavailable (multiple errors)"); 179 if (cnt_dbe) 180 edac_mc_handle_error(HW_EVENT_ERR_UNCORRECTED, mci, 181 cnt_sbe, /* error count */ 182 0, 0, 0, /* pfn, offset, syndrome */ 183 -1, -1, -1, /* top, mid, low layer */ 184 mci->ctl_name, 185 "details unavailable (multiple errors)"); 186 187 /* report details for most recent error */ 188 cs_val = (addr & SDRAM_ERR_ADDR_CS_MASK) >> SDRAM_ERR_ADDR_CS_OFFSET; 189 bank_val = (addr & SDRAM_ERR_ADDR_BANK_MASK) >> SDRAM_ERR_ADDR_BANK_OFFSET; 190 row_val = (calc_ecc & SDRAM_ERR_CALC_ECC_ROW_MASK) >> SDRAM_ERR_CALC_ECC_ROW_OFFSET; 191 col_val = (addr & SDRAM_ERR_ADDR_COL_MASK) >> SDRAM_ERR_ADDR_COL_OFFSET; 192 syndrome_val = (recv_ecc ^ calc_ecc) & 0xff; 193 addr_val = axp_mc_calc_address(drvdata, cs_val, bank_val, row_val, 194 col_val); 195 msg += sprintf(msg, "row=0x%04x ", row_val); /* 11 chars */ 196 msg += sprintf(msg, "bank=0x%x ", bank_val); /* 9 chars */ 197 msg += sprintf(msg, "col=0x%04x ", col_val); /* 11 chars */ 198 msg += sprintf(msg, "cs=%d", cs_val); /* 4 chars */ 199 200 if (!(addr & SDRAM_ERR_ADDR_TYPE_MASK)) { 201 edac_mc_handle_error(HW_EVENT_ERR_CORRECTED, mci, 202 1, /* error count */ 203 addr_val >> PAGE_SHIFT, 204 addr_val & ~PAGE_MASK, 205 syndrome_val, 206 cs_val, -1, -1, /* top, mid, low layer */ 207 mci->ctl_name, drvdata->msg); 208 } else { 209 edac_mc_handle_error(HW_EVENT_ERR_UNCORRECTED, mci, 210 1, /* error count */ 211 addr_val >> PAGE_SHIFT, 212 addr_val & ~PAGE_MASK, 213 syndrome_val, 214 cs_val, -1, -1, /* top, mid, low layer */ 215 mci->ctl_name, drvdata->msg); 216 } 217 } 218 219 static void axp_mc_read_config(struct mem_ctl_info *mci) 220 { 221 struct axp_mc_drvdata *drvdata = mci->pvt_info; 222 uint32_t config, addr_ctrl, rank_ctrl; 223 unsigned int i, cs_struct, cs_size; 224 struct dimm_info *dimm; 225 226 config = readl(drvdata->base + SDRAM_CONFIG_REG); 227 if (config & SDRAM_CONFIG_BUS_WIDTH_MASK) 228 /* 64 bit */ 229 drvdata->width = 8; 230 else 231 /* 32 bit */ 232 drvdata->width = 4; 233 234 addr_ctrl = readl(drvdata->base + SDRAM_ADDR_CTRL_REG); 235 rank_ctrl = readl(drvdata->base + SDRAM_RANK_CTRL_REG); 236 for (i = 0; i < SDRAM_NUM_CS; i++) { 237 dimm = mci->dimms[i]; 238 239 if (!(rank_ctrl & SDRAM_RANK_CTRL_EXIST_MASK(i))) 240 continue; 241 242 drvdata->cs_addr_sel[i] = 243 !!(addr_ctrl & SDRAM_ADDR_CTRL_ADDR_SEL_MASK(i)); 244 245 cs_struct = (addr_ctrl & SDRAM_ADDR_CTRL_STRUCT_MASK(i)) >> SDRAM_ADDR_CTRL_STRUCT_OFFSET(i); 246 cs_size = ((addr_ctrl & SDRAM_ADDR_CTRL_SIZE_HIGH_MASK(i)) >> (SDRAM_ADDR_CTRL_SIZE_HIGH_OFFSET(i) - 2) | 247 ((addr_ctrl & SDRAM_ADDR_CTRL_SIZE_LOW_MASK(i)) >> SDRAM_ADDR_CTRL_SIZE_LOW_OFFSET(i))); 248 249 switch (cs_size) { 250 case 0: /* 2GBit */ 251 dimm->nr_pages = 524288; 252 break; 253 case 1: /* 256MBit */ 254 dimm->nr_pages = 65536; 255 break; 256 case 2: /* 512MBit */ 257 dimm->nr_pages = 131072; 258 break; 259 case 3: /* 1GBit */ 260 dimm->nr_pages = 262144; 261 break; 262 case 4: /* 4GBit */ 263 dimm->nr_pages = 1048576; 264 break; 265 case 5: /* 8GBit */ 266 dimm->nr_pages = 2097152; 267 break; 268 } 269 dimm->grain = 8; 270 dimm->dtype = cs_struct ? DEV_X16 : DEV_X8; 271 dimm->mtype = (config & SDRAM_CONFIG_REGISTERED_MASK) ? 272 MEM_RDDR3 : MEM_DDR3; 273 dimm->edac_mode = EDAC_SECDED; 274 } 275 } 276 277 static const struct of_device_id axp_mc_of_match[] = { 278 {.compatible = "marvell,armada-xp-sdram-controller",}, 279 {}, 280 }; 281 MODULE_DEVICE_TABLE(of, axp_mc_of_match); 282 283 static int axp_mc_probe(struct platform_device *pdev) 284 { 285 struct axp_mc_drvdata *drvdata; 286 struct edac_mc_layer layers[1]; 287 const struct of_device_id *id; 288 struct mem_ctl_info *mci; 289 struct resource *r; 290 void __iomem *base; 291 uint32_t config; 292 293 r = platform_get_resource(pdev, IORESOURCE_MEM, 0); 294 if (!r) { 295 dev_err(&pdev->dev, "Unable to get mem resource\n"); 296 return -ENODEV; 297 } 298 299 base = devm_ioremap_resource(&pdev->dev, r); 300 if (IS_ERR(base)) { 301 dev_err(&pdev->dev, "Unable to map regs\n"); 302 return PTR_ERR(base); 303 } 304 305 config = readl(base + SDRAM_CONFIG_REG); 306 if (!(config & SDRAM_CONFIG_ECC_MASK)) { 307 dev_warn(&pdev->dev, "SDRAM ECC is not enabled"); 308 return -EINVAL; 309 } 310 311 layers[0].type = EDAC_MC_LAYER_CHIP_SELECT; 312 layers[0].size = SDRAM_NUM_CS; 313 layers[0].is_virt_csrow = true; 314 315 mci = edac_mc_alloc(0, ARRAY_SIZE(layers), layers, sizeof(*drvdata)); 316 if (!mci) 317 return -ENOMEM; 318 319 drvdata = mci->pvt_info; 320 drvdata->base = base; 321 mci->pdev = &pdev->dev; 322 platform_set_drvdata(pdev, mci); 323 324 id = of_match_device(axp_mc_of_match, &pdev->dev); 325 mci->edac_check = axp_mc_check; 326 mci->mtype_cap = MEM_FLAG_DDR3; 327 mci->edac_cap = EDAC_FLAG_SECDED; 328 mci->mod_name = pdev->dev.driver->name; 329 mci->ctl_name = id ? id->compatible : "unknown"; 330 mci->dev_name = dev_name(&pdev->dev); 331 mci->scrub_mode = SCRUB_NONE; 332 333 axp_mc_read_config(mci); 334 335 /* These SoCs have a reduced width bus */ 336 if (of_machine_is_compatible("marvell,armada380") || 337 of_machine_is_compatible("marvell,armadaxp-98dx3236")) 338 drvdata->width /= 2; 339 340 /* configure SBE threshold */ 341 /* it seems that SBEs are not captured otherwise */ 342 writel(1 << SDRAM_ERR_CTRL_THR_OFFSET, drvdata->base + SDRAM_ERR_CTRL_REG); 343 344 /* clear cause registers */ 345 writel(~(SDRAM_ERR_CAUSE_DBE_MASK | SDRAM_ERR_CAUSE_SBE_MASK), drvdata->base + SDRAM_ERR_CAUSE_ERR_REG); 346 writel(~(SDRAM_ERR_CAUSE_DBE_MASK | SDRAM_ERR_CAUSE_SBE_MASK), drvdata->base + SDRAM_ERR_CAUSE_MSG_REG); 347 348 /* clear counter registers */ 349 writel(0, drvdata->base + SDRAM_ERR_SBE_COUNT_REG); 350 writel(0, drvdata->base + SDRAM_ERR_DBE_COUNT_REG); 351 352 if (edac_mc_add_mc(mci)) { 353 edac_mc_free(mci); 354 return -EINVAL; 355 } 356 edac_op_state = EDAC_OPSTATE_POLL; 357 358 return 0; 359 } 360 361 static int axp_mc_remove(struct platform_device *pdev) 362 { 363 struct mem_ctl_info *mci = platform_get_drvdata(pdev); 364 365 edac_mc_del_mc(&pdev->dev); 366 edac_mc_free(mci); 367 platform_set_drvdata(pdev, NULL); 368 369 return 0; 370 } 371 372 static struct platform_driver axp_mc_driver = { 373 .probe = axp_mc_probe, 374 .remove = axp_mc_remove, 375 .driver = { 376 .name = "armada_xp_mc_edac", 377 .of_match_table = of_match_ptr(axp_mc_of_match), 378 }, 379 }; 380 381 /************************ EDAC Device (L2 Cache) ***************************/ 382 383 struct aurora_l2_drvdata { 384 void __iomem *base; 385 386 char msg[128]; 387 388 /* error injection via debugfs */ 389 uint32_t inject_addr; 390 uint32_t inject_mask; 391 uint8_t inject_ctl; 392 393 struct dentry *debugfs; 394 }; 395 396 #ifdef CONFIG_EDAC_DEBUG 397 static void aurora_l2_inject(struct aurora_l2_drvdata *drvdata) 398 { 399 drvdata->inject_addr &= AURORA_ERR_INJECT_CTL_ADDR_MASK; 400 drvdata->inject_ctl &= AURORA_ERR_INJECT_CTL_EN_MASK; 401 writel(0, drvdata->base + AURORA_ERR_INJECT_CTL_REG); 402 writel(drvdata->inject_mask, drvdata->base + AURORA_ERR_INJECT_MASK_REG); 403 writel(drvdata->inject_addr | drvdata->inject_ctl, drvdata->base + AURORA_ERR_INJECT_CTL_REG); 404 } 405 #endif 406 407 static void aurora_l2_check(struct edac_device_ctl_info *dci) 408 { 409 struct aurora_l2_drvdata *drvdata = dci->pvt_info; 410 uint32_t cnt, src, txn, err, attr_cap, addr_cap, way_cap; 411 unsigned int cnt_ce, cnt_ue; 412 char *msg = drvdata->msg; 413 size_t size = sizeof(drvdata->msg); 414 size_t len = 0; 415 416 cnt = readl(drvdata->base + AURORA_ERR_CNT_REG); 417 attr_cap = readl(drvdata->base + AURORA_ERR_ATTR_CAP_REG); 418 addr_cap = readl(drvdata->base + AURORA_ERR_ADDR_CAP_REG); 419 way_cap = readl(drvdata->base + AURORA_ERR_WAY_CAP_REG); 420 421 cnt_ce = (cnt & AURORA_ERR_CNT_CE_MASK) >> AURORA_ERR_CNT_CE_OFFSET; 422 cnt_ue = (cnt & AURORA_ERR_CNT_UE_MASK) >> AURORA_ERR_CNT_UE_OFFSET; 423 /* clear error counter registers */ 424 if (cnt_ce || cnt_ue) 425 writel(AURORA_ERR_CNT_CLR, drvdata->base + AURORA_ERR_CNT_REG); 426 427 if (!(attr_cap & AURORA_ERR_ATTR_CAP_VALID)) 428 goto clear_remaining; 429 430 src = (attr_cap & AURORA_ERR_ATTR_SRC_MSK) >> AURORA_ERR_ATTR_SRC_OFF; 431 if (src <= 3) 432 len += snprintf(msg+len, size-len, "src=CPU%d ", src); 433 else 434 len += snprintf(msg+len, size-len, "src=IO "); 435 436 txn = (attr_cap & AURORA_ERR_ATTR_TXN_MSK) >> AURORA_ERR_ATTR_TXN_OFF; 437 switch (txn) { 438 case 0: 439 len += snprintf(msg+len, size-len, "txn=Data-Read "); 440 break; 441 case 1: 442 len += snprintf(msg+len, size-len, "txn=Isn-Read "); 443 break; 444 case 2: 445 len += snprintf(msg+len, size-len, "txn=Clean-Flush "); 446 break; 447 case 3: 448 len += snprintf(msg+len, size-len, "txn=Eviction "); 449 break; 450 case 4: 451 len += snprintf(msg+len, size-len, 452 "txn=Read-Modify-Write "); 453 break; 454 } 455 456 err = (attr_cap & AURORA_ERR_ATTR_ERR_MSK) >> AURORA_ERR_ATTR_ERR_OFF; 457 switch (err) { 458 case 0: 459 len += snprintf(msg+len, size-len, "err=CorrECC "); 460 break; 461 case 1: 462 len += snprintf(msg+len, size-len, "err=UnCorrECC "); 463 break; 464 case 2: 465 len += snprintf(msg+len, size-len, "err=TagParity "); 466 break; 467 } 468 469 len += snprintf(msg+len, size-len, "addr=0x%x ", addr_cap & AURORA_ERR_ADDR_CAP_ADDR_MASK); 470 len += snprintf(msg+len, size-len, "index=0x%x ", (way_cap & AURORA_ERR_WAY_IDX_MSK) >> AURORA_ERR_WAY_IDX_OFF); 471 len += snprintf(msg+len, size-len, "way=0x%x", (way_cap & AURORA_ERR_WAY_CAP_WAY_MASK) >> AURORA_ERR_WAY_CAP_WAY_OFFSET); 472 473 /* clear error capture registers */ 474 writel(AURORA_ERR_ATTR_CAP_VALID, drvdata->base + AURORA_ERR_ATTR_CAP_REG); 475 if (err) { 476 /* UnCorrECC or TagParity */ 477 if (cnt_ue) 478 cnt_ue--; 479 edac_device_handle_ue(dci, 0, 0, drvdata->msg); 480 } else { 481 if (cnt_ce) 482 cnt_ce--; 483 edac_device_handle_ce(dci, 0, 0, drvdata->msg); 484 } 485 486 clear_remaining: 487 /* report remaining errors */ 488 while (cnt_ue--) 489 edac_device_handle_ue(dci, 0, 0, "details unavailable (multiple errors)"); 490 while (cnt_ce--) 491 edac_device_handle_ue(dci, 0, 0, "details unavailable (multiple errors)"); 492 } 493 494 static void aurora_l2_poll(struct edac_device_ctl_info *dci) 495 { 496 #ifdef CONFIG_EDAC_DEBUG 497 struct aurora_l2_drvdata *drvdata = dci->pvt_info; 498 #endif 499 500 aurora_l2_check(dci); 501 #ifdef CONFIG_EDAC_DEBUG 502 aurora_l2_inject(drvdata); 503 #endif 504 } 505 506 static const struct of_device_id aurora_l2_of_match[] = { 507 {.compatible = "marvell,aurora-system-cache",}, 508 {}, 509 }; 510 MODULE_DEVICE_TABLE(of, aurora_l2_of_match); 511 512 static int aurora_l2_probe(struct platform_device *pdev) 513 { 514 struct aurora_l2_drvdata *drvdata; 515 struct edac_device_ctl_info *dci; 516 const struct of_device_id *id; 517 uint32_t l2x0_aux_ctrl; 518 void __iomem *base; 519 struct resource *r; 520 521 r = platform_get_resource(pdev, IORESOURCE_MEM, 0); 522 if (!r) { 523 dev_err(&pdev->dev, "Unable to get mem resource\n"); 524 return -ENODEV; 525 } 526 527 base = devm_ioremap_resource(&pdev->dev, r); 528 if (IS_ERR(base)) { 529 dev_err(&pdev->dev, "Unable to map regs\n"); 530 return PTR_ERR(base); 531 } 532 533 l2x0_aux_ctrl = readl(base + L2X0_AUX_CTRL); 534 if (!(l2x0_aux_ctrl & AURORA_ACR_PARITY_EN)) 535 dev_warn(&pdev->dev, "tag parity is not enabled"); 536 if (!(l2x0_aux_ctrl & AURORA_ACR_ECC_EN)) 537 dev_warn(&pdev->dev, "data ECC is not enabled"); 538 539 dci = edac_device_alloc_ctl_info(sizeof(*drvdata), 540 "cpu", 1, "L", 1, 2, NULL, 0, 0); 541 if (!dci) 542 return -ENOMEM; 543 544 drvdata = dci->pvt_info; 545 drvdata->base = base; 546 dci->dev = &pdev->dev; 547 platform_set_drvdata(pdev, dci); 548 549 id = of_match_device(aurora_l2_of_match, &pdev->dev); 550 dci->edac_check = aurora_l2_poll; 551 dci->mod_name = pdev->dev.driver->name; 552 dci->ctl_name = id ? id->compatible : "unknown"; 553 dci->dev_name = dev_name(&pdev->dev); 554 555 /* clear registers */ 556 writel(AURORA_ERR_CNT_CLR, drvdata->base + AURORA_ERR_CNT_REG); 557 writel(AURORA_ERR_ATTR_CAP_VALID, drvdata->base + AURORA_ERR_ATTR_CAP_REG); 558 559 if (edac_device_add_device(dci)) { 560 edac_device_free_ctl_info(dci); 561 return -EINVAL; 562 } 563 564 #ifdef CONFIG_EDAC_DEBUG 565 drvdata->debugfs = edac_debugfs_create_dir(dev_name(&pdev->dev)); 566 if (drvdata->debugfs) { 567 edac_debugfs_create_x32("inject_addr", 0644, 568 drvdata->debugfs, 569 &drvdata->inject_addr); 570 edac_debugfs_create_x32("inject_mask", 0644, 571 drvdata->debugfs, 572 &drvdata->inject_mask); 573 edac_debugfs_create_x8("inject_ctl", 0644, 574 drvdata->debugfs, &drvdata->inject_ctl); 575 } 576 #endif 577 578 return 0; 579 } 580 581 static int aurora_l2_remove(struct platform_device *pdev) 582 { 583 struct edac_device_ctl_info *dci = platform_get_drvdata(pdev); 584 #ifdef CONFIG_EDAC_DEBUG 585 struct aurora_l2_drvdata *drvdata = dci->pvt_info; 586 587 edac_debugfs_remove_recursive(drvdata->debugfs); 588 #endif 589 edac_device_del_device(&pdev->dev); 590 edac_device_free_ctl_info(dci); 591 platform_set_drvdata(pdev, NULL); 592 593 return 0; 594 } 595 596 static struct platform_driver aurora_l2_driver = { 597 .probe = aurora_l2_probe, 598 .remove = aurora_l2_remove, 599 .driver = { 600 .name = "aurora_l2_edac", 601 .of_match_table = of_match_ptr(aurora_l2_of_match), 602 }, 603 }; 604 605 /************************ Driver registration ******************************/ 606 607 static struct platform_driver * const drivers[] = { 608 &axp_mc_driver, 609 &aurora_l2_driver, 610 }; 611 612 static int __init armada_xp_edac_init(void) 613 { 614 int res; 615 616 /* only polling is supported */ 617 edac_op_state = EDAC_OPSTATE_POLL; 618 619 res = platform_register_drivers(drivers, ARRAY_SIZE(drivers)); 620 if (res) 621 pr_warn("Aramda XP EDAC drivers fail to register\n"); 622 623 return 0; 624 } 625 module_init(armada_xp_edac_init); 626 627 static void __exit armada_xp_edac_exit(void) 628 { 629 platform_unregister_drivers(drivers, ARRAY_SIZE(drivers)); 630 } 631 module_exit(armada_xp_edac_exit); 632 633 MODULE_LICENSE("GPL v2"); 634 MODULE_AUTHOR("Pengutronix"); 635 MODULE_DESCRIPTION("EDAC Drivers for Marvell Armada XP SDRAM and L2 Cache Controller"); 636