1 // SPDX-License-Identifier: GPL-2.0 2 /* 3 * Page Deallocation Table (PDT) support 4 * 5 * The Page Deallocation Table (PDT) is maintained by firmware and holds a 6 * list of memory addresses in which memory errors were detected. 7 * The list contains both single-bit (correctable) and double-bit 8 * (uncorrectable) errors. 9 * 10 * Copyright 2017 by Helge Deller <deller@gmx.de> 11 * 12 * possible future enhancements: 13 * - add userspace interface via procfs or sysfs to clear PDT 14 */ 15 16 #include <linux/memblock.h> 17 #include <linux/seq_file.h> 18 #include <linux/kthread.h> 19 #include <linux/initrd.h> 20 #include <linux/pgtable.h> 21 #include <linux/mm.h> 22 23 #include <asm/pdc.h> 24 #include <asm/pdcpat.h> 25 #include <asm/sections.h> 26 #include <asm/pgtable.h> 27 28 enum pdt_access_type { 29 PDT_NONE, 30 PDT_PDC, 31 PDT_PAT_NEW, 32 PDT_PAT_CELL 33 }; 34 35 static enum pdt_access_type pdt_type; 36 37 /* PDT poll interval: 1 minute if errors, 5 minutes if everything OK. */ 38 #define PDT_POLL_INTERVAL_DEFAULT (5*60*HZ) 39 #define PDT_POLL_INTERVAL_SHORT (1*60*HZ) 40 static unsigned long pdt_poll_interval = PDT_POLL_INTERVAL_DEFAULT; 41 42 /* global PDT status information */ 43 static struct pdc_mem_retinfo pdt_status; 44 45 #define MAX_PDT_TABLE_SIZE PAGE_SIZE 46 #define MAX_PDT_ENTRIES (MAX_PDT_TABLE_SIZE / sizeof(unsigned long)) 47 static unsigned long pdt_entry[MAX_PDT_ENTRIES] __page_aligned_bss; 48 49 /* 50 * Constants for the pdt_entry format: 51 * A pdt_entry holds the physical address in bits 0-57, bits 58-61 are 52 * reserved, bit 62 is the perm bit and bit 63 is the error_type bit. 53 * The perm bit indicates whether the error have been verified as a permanent 54 * error (value of 1) or has not been verified, and may be transient (value 55 * of 0). The error_type bit indicates whether the error is a single bit error 56 * (value of 1) or a multiple bit error. 57 * On non-PAT machines phys_addr is encoded in bits 0-59 and error_type in bit 58 * 63. Those machines don't provide the perm bit. 59 */ 60 61 #define PDT_ADDR_PHYS_MASK (pdt_type != PDT_PDC ? ~0x3f : ~0x0f) 62 #define PDT_ADDR_PERM_ERR (pdt_type != PDT_PDC ? 2UL : 0UL) 63 #define PDT_ADDR_SINGLE_ERR 1UL 64 65 /* report PDT entries via /proc/meminfo */ 66 void arch_report_meminfo(struct seq_file *m) 67 { 68 if (pdt_type == PDT_NONE) 69 return; 70 71 seq_printf(m, "PDT_max_entries: %7lu\n", 72 pdt_status.pdt_size); 73 seq_printf(m, "PDT_cur_entries: %7lu\n", 74 pdt_status.pdt_entries); 75 } 76 77 static int get_info_pat_new(void) 78 { 79 struct pdc_pat_mem_retinfo pat_rinfo; 80 int ret; 81 82 /* newer PAT machines like C8000 report info for all cells */ 83 if (is_pdc_pat()) 84 ret = pdc_pat_mem_pdt_info(&pat_rinfo); 85 else 86 return PDC_BAD_PROC; 87 88 pdt_status.pdt_size = pat_rinfo.max_pdt_entries; 89 pdt_status.pdt_entries = pat_rinfo.current_pdt_entries; 90 pdt_status.pdt_status = 0; 91 pdt_status.first_dbe_loc = pat_rinfo.first_dbe_loc; 92 pdt_status.good_mem = pat_rinfo.good_mem; 93 94 return ret; 95 } 96 97 static int get_info_pat_cell(void) 98 { 99 struct pdc_pat_mem_cell_pdt_retinfo cell_rinfo; 100 int ret; 101 102 /* older PAT machines like rp5470 report cell info only */ 103 if (is_pdc_pat()) 104 ret = pdc_pat_mem_pdt_cell_info(&cell_rinfo, parisc_cell_num); 105 else 106 return PDC_BAD_PROC; 107 108 pdt_status.pdt_size = cell_rinfo.max_pdt_entries; 109 pdt_status.pdt_entries = cell_rinfo.current_pdt_entries; 110 pdt_status.pdt_status = 0; 111 pdt_status.first_dbe_loc = cell_rinfo.first_dbe_loc; 112 pdt_status.good_mem = cell_rinfo.good_mem; 113 114 return ret; 115 } 116 117 static void report_mem_err(unsigned long pde) 118 { 119 struct pdc_pat_mem_phys_mem_location loc; 120 unsigned long addr; 121 char dimm_txt[32]; 122 123 addr = pde & PDT_ADDR_PHYS_MASK; 124 125 /* show DIMM slot description on PAT machines */ 126 if (is_pdc_pat()) { 127 pdc_pat_mem_get_dimm_phys_location(&loc, addr); 128 sprintf(dimm_txt, "DIMM slot %02x, ", loc.dimm_slot); 129 } else 130 dimm_txt[0] = 0; 131 132 pr_warn("PDT: BAD MEMORY at 0x%08lx, %s%s%s-bit error.\n", 133 addr, dimm_txt, 134 pde & PDT_ADDR_PERM_ERR ? "permanent ":"", 135 pde & PDT_ADDR_SINGLE_ERR ? "single":"multi"); 136 } 137 138 139 /* 140 * pdc_pdt_init() 141 * 142 * Initialize kernel PDT structures, read initial PDT table from firmware, 143 * report all current PDT entries and mark bad memory with memblock_reserve() 144 * to avoid that the kernel will use broken memory areas. 145 * 146 */ 147 void __init pdc_pdt_init(void) 148 { 149 int ret, i; 150 unsigned long entries; 151 struct pdc_mem_read_pdt pdt_read_ret; 152 153 pdt_type = PDT_PAT_NEW; 154 ret = get_info_pat_new(); 155 156 if (ret != PDC_OK) { 157 pdt_type = PDT_PAT_CELL; 158 ret = get_info_pat_cell(); 159 } 160 161 if (ret != PDC_OK) { 162 pdt_type = PDT_PDC; 163 /* non-PAT machines provide the standard PDC call */ 164 ret = pdc_mem_pdt_info(&pdt_status); 165 } 166 167 if (ret != PDC_OK) { 168 pdt_type = PDT_NONE; 169 pr_info("PDT: Firmware does not provide any page deallocation" 170 " information.\n"); 171 return; 172 } 173 174 entries = pdt_status.pdt_entries; 175 if (WARN_ON(entries > MAX_PDT_ENTRIES)) 176 entries = pdt_status.pdt_entries = MAX_PDT_ENTRIES; 177 178 pr_info("PDT: type %s, size %lu, entries %lu, status %lu, dbe_loc 0x%lx," 179 " good_mem %lu MB\n", 180 pdt_type == PDT_PDC ? __stringify(PDT_PDC) : 181 pdt_type == PDT_PAT_CELL ? __stringify(PDT_PAT_CELL) 182 : __stringify(PDT_PAT_NEW), 183 pdt_status.pdt_size, pdt_status.pdt_entries, 184 pdt_status.pdt_status, pdt_status.first_dbe_loc, 185 pdt_status.good_mem / 1024 / 1024); 186 187 if (entries == 0) { 188 pr_info("PDT: Firmware reports all memory OK.\n"); 189 return; 190 } 191 192 if (pdt_status.first_dbe_loc && 193 pdt_status.first_dbe_loc <= __pa((unsigned long)&_end)) 194 pr_crit("CRITICAL: Bad memory inside kernel image memory area!\n"); 195 196 pr_warn("PDT: Firmware reports %lu entries of faulty memory:\n", 197 entries); 198 199 if (pdt_type == PDT_PDC) 200 ret = pdc_mem_pdt_read_entries(&pdt_read_ret, pdt_entry); 201 else { 202 #ifdef CONFIG_64BIT 203 struct pdc_pat_mem_read_pd_retinfo pat_pret; 204 205 if (pdt_type == PDT_PAT_CELL) 206 ret = pdc_pat_mem_read_cell_pdt(&pat_pret, pdt_entry, 207 MAX_PDT_ENTRIES); 208 else 209 ret = pdc_pat_mem_read_pd_pdt(&pat_pret, pdt_entry, 210 MAX_PDT_TABLE_SIZE, 0); 211 #else 212 ret = PDC_BAD_PROC; 213 #endif 214 } 215 216 if (ret != PDC_OK) { 217 pdt_type = PDT_NONE; 218 pr_warn("PDT: Get PDT entries failed with %d\n", ret); 219 return; 220 } 221 222 for (i = 0; i < pdt_status.pdt_entries; i++) { 223 unsigned long addr; 224 225 report_mem_err(pdt_entry[i]); 226 227 addr = pdt_entry[i] & PDT_ADDR_PHYS_MASK; 228 if (IS_ENABLED(CONFIG_BLK_DEV_INITRD) && 229 addr >= initrd_start && addr < initrd_end) 230 pr_crit("CRITICAL: initrd possibly broken " 231 "due to bad memory!\n"); 232 233 /* mark memory page bad */ 234 memblock_reserve(pdt_entry[i] & PAGE_MASK, PAGE_SIZE); 235 num_poisoned_pages_inc(addr >> PAGE_SHIFT); 236 } 237 } 238 239 240 /* 241 * This is the PDT kernel thread main loop. 242 */ 243 244 static int pdt_mainloop(void *unused) 245 { 246 struct pdc_mem_read_pdt pdt_read_ret; 247 struct pdc_pat_mem_read_pd_retinfo pat_pret __maybe_unused; 248 unsigned long old_num_entries; 249 unsigned long *bad_mem_ptr; 250 int num, ret; 251 252 for (;;) { 253 set_current_state(TASK_INTERRUPTIBLE); 254 255 old_num_entries = pdt_status.pdt_entries; 256 257 schedule_timeout(pdt_poll_interval); 258 if (kthread_should_stop()) 259 break; 260 261 /* Do we have new PDT entries? */ 262 switch (pdt_type) { 263 case PDT_PAT_NEW: 264 ret = get_info_pat_new(); 265 break; 266 case PDT_PAT_CELL: 267 ret = get_info_pat_cell(); 268 break; 269 default: 270 ret = pdc_mem_pdt_info(&pdt_status); 271 break; 272 } 273 274 if (ret != PDC_OK) { 275 pr_warn("PDT: unexpected failure %d\n", ret); 276 return -EINVAL; 277 } 278 279 /* if no new PDT entries, just wait again */ 280 num = pdt_status.pdt_entries - old_num_entries; 281 if (num <= 0) 282 continue; 283 284 /* decrease poll interval in case we found memory errors */ 285 if (pdt_status.pdt_entries && 286 pdt_poll_interval == PDT_POLL_INTERVAL_DEFAULT) 287 pdt_poll_interval = PDT_POLL_INTERVAL_SHORT; 288 289 /* limit entries to get */ 290 if (num > MAX_PDT_ENTRIES) { 291 num = MAX_PDT_ENTRIES; 292 pdt_status.pdt_entries = old_num_entries + num; 293 } 294 295 /* get new entries */ 296 switch (pdt_type) { 297 #ifdef CONFIG_64BIT 298 case PDT_PAT_CELL: 299 if (pdt_status.pdt_entries > MAX_PDT_ENTRIES) { 300 pr_crit("PDT: too many entries.\n"); 301 return -ENOMEM; 302 } 303 ret = pdc_pat_mem_read_cell_pdt(&pat_pret, pdt_entry, 304 MAX_PDT_ENTRIES); 305 bad_mem_ptr = &pdt_entry[old_num_entries]; 306 break; 307 case PDT_PAT_NEW: 308 ret = pdc_pat_mem_read_pd_pdt(&pat_pret, 309 pdt_entry, 310 num * sizeof(unsigned long), 311 old_num_entries * sizeof(unsigned long)); 312 bad_mem_ptr = &pdt_entry[0]; 313 break; 314 #endif 315 default: 316 ret = pdc_mem_pdt_read_entries(&pdt_read_ret, 317 pdt_entry); 318 bad_mem_ptr = &pdt_entry[old_num_entries]; 319 break; 320 } 321 322 /* report and mark memory broken */ 323 while (num--) { 324 unsigned long pde = *bad_mem_ptr++; 325 326 report_mem_err(pde); 327 328 #ifdef CONFIG_MEMORY_FAILURE 329 if ((pde & PDT_ADDR_PERM_ERR) || 330 ((pde & PDT_ADDR_SINGLE_ERR) == 0)) 331 memory_failure(pde >> PAGE_SHIFT, 0); 332 else 333 soft_offline_page(pde >> PAGE_SHIFT, 0); 334 #else 335 pr_crit("PDT: memory error at 0x%lx ignored.\n" 336 "Rebuild kernel with CONFIG_MEMORY_FAILURE=y " 337 "for real handling.\n", 338 pde & PDT_ADDR_PHYS_MASK); 339 #endif 340 341 } 342 } 343 344 return 0; 345 } 346 347 348 static int __init pdt_initcall(void) 349 { 350 struct task_struct *kpdtd_task; 351 352 if (pdt_type == PDT_NONE) 353 return -ENODEV; 354 355 kpdtd_task = kthread_run(pdt_mainloop, NULL, "kpdtd"); 356 if (IS_ERR(kpdtd_task)) 357 return PTR_ERR(kpdtd_task); 358 359 return 0; 360 } 361 362 late_initcall(pdt_initcall); 363