xref: /openbmc/linux/arch/parisc/kernel/pdt.c (revision b755c25f)
1 // SPDX-License-Identifier: GPL-2.0
2 /*
3  *    Page Deallocation Table (PDT) support
4  *
5  *    The Page Deallocation Table (PDT) is maintained by firmware and holds a
6  *    list of memory addresses in which memory errors were detected.
7  *    The list contains both single-bit (correctable) and double-bit
8  *    (uncorrectable) errors.
9  *
10  *    Copyright 2017 by Helge Deller <deller@gmx.de>
11  *
12  *    possible future enhancements:
13  *    - add userspace interface via procfs or sysfs to clear PDT
14  */
15 
16 #include <linux/memblock.h>
17 #include <linux/seq_file.h>
18 #include <linux/kthread.h>
19 #include <linux/proc_fs.h>
20 #include <linux/initrd.h>
21 #include <linux/pgtable.h>
22 #include <linux/mm.h>
23 
24 #include <asm/pdc.h>
25 #include <asm/pdcpat.h>
26 #include <asm/sections.h>
27 #include <asm/pgtable.h>
28 
29 enum pdt_access_type {
30 	PDT_NONE,
31 	PDT_PDC,
32 	PDT_PAT_NEW,
33 	PDT_PAT_CELL
34 };
35 
36 static enum pdt_access_type pdt_type;
37 
38 /* PDT poll interval: 1 minute if errors, 5 minutes if everything OK. */
39 #define PDT_POLL_INTERVAL_DEFAULT	(5*60*HZ)
40 #define PDT_POLL_INTERVAL_SHORT		(1*60*HZ)
41 static unsigned long pdt_poll_interval = PDT_POLL_INTERVAL_DEFAULT;
42 
43 /* global PDT status information */
44 static struct pdc_mem_retinfo pdt_status;
45 
46 #define MAX_PDT_TABLE_SIZE	PAGE_SIZE
47 #define MAX_PDT_ENTRIES		(MAX_PDT_TABLE_SIZE / sizeof(unsigned long))
48 static unsigned long pdt_entry[MAX_PDT_ENTRIES] __page_aligned_bss;
49 
50 /*
51  * Constants for the pdt_entry format:
52  * A pdt_entry holds the physical address in bits 0-57, bits 58-61 are
53  * reserved, bit 62 is the perm bit and bit 63 is the error_type bit.
54  * The perm bit indicates whether the error have been verified as a permanent
55  * error (value of 1) or has not been verified, and may be transient (value
56  * of 0). The error_type bit indicates whether the error is a single bit error
57  * (value of 1) or a multiple bit error.
58  * On non-PAT machines phys_addr is encoded in bits 0-59 and error_type in bit
59  * 63. Those machines don't provide the perm bit.
60  */
61 
62 #define PDT_ADDR_PHYS_MASK	(pdt_type != PDT_PDC ? ~0x3f : ~0x0f)
63 #define PDT_ADDR_PERM_ERR	(pdt_type != PDT_PDC ? 2UL : 0UL)
64 #define PDT_ADDR_SINGLE_ERR	1UL
65 
66 /* report PDT entries via /proc/meminfo */
67 void arch_report_meminfo(struct seq_file *m)
68 {
69 	if (pdt_type == PDT_NONE)
70 		return;
71 
72 	seq_printf(m, "PDT_max_entries: %7lu\n",
73 			pdt_status.pdt_size);
74 	seq_printf(m, "PDT_cur_entries: %7lu\n",
75 			pdt_status.pdt_entries);
76 }
77 
78 static int get_info_pat_new(void)
79 {
80 	struct pdc_pat_mem_retinfo pat_rinfo;
81 	int ret;
82 
83 	/* newer PAT machines like C8000 report info for all cells */
84 	if (is_pdc_pat())
85 		ret = pdc_pat_mem_pdt_info(&pat_rinfo);
86 	else
87 		return PDC_BAD_PROC;
88 
89 	pdt_status.pdt_size = pat_rinfo.max_pdt_entries;
90 	pdt_status.pdt_entries = pat_rinfo.current_pdt_entries;
91 	pdt_status.pdt_status = 0;
92 	pdt_status.first_dbe_loc = pat_rinfo.first_dbe_loc;
93 	pdt_status.good_mem = pat_rinfo.good_mem;
94 
95 	return ret;
96 }
97 
98 static int get_info_pat_cell(void)
99 {
100 	struct pdc_pat_mem_cell_pdt_retinfo cell_rinfo;
101 	int ret;
102 
103 	/* older PAT machines like rp5470 report cell info only */
104 	if (is_pdc_pat())
105 		ret = pdc_pat_mem_pdt_cell_info(&cell_rinfo, parisc_cell_num);
106 	else
107 		return PDC_BAD_PROC;
108 
109 	pdt_status.pdt_size = cell_rinfo.max_pdt_entries;
110 	pdt_status.pdt_entries = cell_rinfo.current_pdt_entries;
111 	pdt_status.pdt_status = 0;
112 	pdt_status.first_dbe_loc = cell_rinfo.first_dbe_loc;
113 	pdt_status.good_mem = cell_rinfo.good_mem;
114 
115 	return ret;
116 }
117 
118 static void report_mem_err(unsigned long pde)
119 {
120 	struct pdc_pat_mem_phys_mem_location loc;
121 	unsigned long addr;
122 	char dimm_txt[32];
123 
124 	addr = pde & PDT_ADDR_PHYS_MASK;
125 
126 	/* show DIMM slot description on PAT machines */
127 	if (is_pdc_pat()) {
128 		pdc_pat_mem_get_dimm_phys_location(&loc, addr);
129 		sprintf(dimm_txt, "DIMM slot %02x, ", loc.dimm_slot);
130 	} else
131 		dimm_txt[0] = 0;
132 
133 	pr_warn("PDT: BAD MEMORY at 0x%08lx, %s%s%s-bit error.\n",
134 		addr, dimm_txt,
135 		pde & PDT_ADDR_PERM_ERR ? "permanent ":"",
136 		pde & PDT_ADDR_SINGLE_ERR ? "single":"multi");
137 }
138 
139 
140 /*
141  * pdc_pdt_init()
142  *
143  * Initialize kernel PDT structures, read initial PDT table from firmware,
144  * report all current PDT entries and mark bad memory with memblock_reserve()
145  * to avoid that the kernel will use broken memory areas.
146  *
147  */
148 void __init pdc_pdt_init(void)
149 {
150 	int ret, i;
151 	unsigned long entries;
152 	struct pdc_mem_read_pdt pdt_read_ret;
153 
154 	pdt_type = PDT_PAT_NEW;
155 	ret = get_info_pat_new();
156 
157 	if (ret != PDC_OK) {
158 		pdt_type = PDT_PAT_CELL;
159 		ret = get_info_pat_cell();
160 	}
161 
162 	if (ret != PDC_OK) {
163 		pdt_type = PDT_PDC;
164 		/* non-PAT machines provide the standard PDC call */
165 		ret = pdc_mem_pdt_info(&pdt_status);
166 	}
167 
168 	if (ret != PDC_OK) {
169 		pdt_type = PDT_NONE;
170 		pr_info("PDT: Firmware does not provide any page deallocation"
171 			" information.\n");
172 		return;
173 	}
174 
175 	entries = pdt_status.pdt_entries;
176 	if (WARN_ON(entries > MAX_PDT_ENTRIES))
177 		entries = pdt_status.pdt_entries = MAX_PDT_ENTRIES;
178 
179 	pr_info("PDT: type %s, size %lu, entries %lu, status %lu, dbe_loc 0x%lx,"
180 		" good_mem %lu MB\n",
181 			pdt_type == PDT_PDC ? __stringify(PDT_PDC) :
182 			pdt_type == PDT_PAT_CELL ? __stringify(PDT_PAT_CELL)
183 						 : __stringify(PDT_PAT_NEW),
184 			pdt_status.pdt_size, pdt_status.pdt_entries,
185 			pdt_status.pdt_status, pdt_status.first_dbe_loc,
186 			pdt_status.good_mem / 1024 / 1024);
187 
188 	if (entries == 0) {
189 		pr_info("PDT: Firmware reports all memory OK.\n");
190 		return;
191 	}
192 
193 	if (pdt_status.first_dbe_loc &&
194 		pdt_status.first_dbe_loc <= __pa((unsigned long)&_end))
195 		pr_crit("CRITICAL: Bad memory inside kernel image memory area!\n");
196 
197 	pr_warn("PDT: Firmware reports %lu entries of faulty memory:\n",
198 		entries);
199 
200 	if (pdt_type == PDT_PDC)
201 		ret = pdc_mem_pdt_read_entries(&pdt_read_ret, pdt_entry);
202 	else {
203 #ifdef CONFIG_64BIT
204 		struct pdc_pat_mem_read_pd_retinfo pat_pret;
205 
206 		if (pdt_type == PDT_PAT_CELL)
207 			ret = pdc_pat_mem_read_cell_pdt(&pat_pret, pdt_entry,
208 				MAX_PDT_ENTRIES);
209 		else
210 			ret = pdc_pat_mem_read_pd_pdt(&pat_pret, pdt_entry,
211 				MAX_PDT_TABLE_SIZE, 0);
212 #else
213 		ret = PDC_BAD_PROC;
214 #endif
215 	}
216 
217 	if (ret != PDC_OK) {
218 		pdt_type = PDT_NONE;
219 		pr_warn("PDT: Get PDT entries failed with %d\n", ret);
220 		return;
221 	}
222 
223 	for (i = 0; i < pdt_status.pdt_entries; i++) {
224 		unsigned long addr;
225 
226 		report_mem_err(pdt_entry[i]);
227 
228 		addr = pdt_entry[i] & PDT_ADDR_PHYS_MASK;
229 		if (IS_ENABLED(CONFIG_BLK_DEV_INITRD) &&
230 			addr >= initrd_start && addr < initrd_end)
231 			pr_crit("CRITICAL: initrd possibly broken "
232 				"due to bad memory!\n");
233 
234 		/* mark memory page bad */
235 		memblock_reserve(pdt_entry[i] & PAGE_MASK, PAGE_SIZE);
236 		num_poisoned_pages_inc(addr >> PAGE_SHIFT);
237 	}
238 }
239 
240 
241 /*
242  * This is the PDT kernel thread main loop.
243  */
244 
245 static int pdt_mainloop(void *unused)
246 {
247 	struct pdc_mem_read_pdt pdt_read_ret;
248 	struct pdc_pat_mem_read_pd_retinfo pat_pret __maybe_unused;
249 	unsigned long old_num_entries;
250 	unsigned long *bad_mem_ptr;
251 	int num, ret;
252 
253 	for (;;) {
254 		set_current_state(TASK_INTERRUPTIBLE);
255 
256 		old_num_entries = pdt_status.pdt_entries;
257 
258 		schedule_timeout(pdt_poll_interval);
259 		if (kthread_should_stop())
260 			break;
261 
262 		/* Do we have new PDT entries? */
263 		switch (pdt_type) {
264 		case PDT_PAT_NEW:
265 			ret = get_info_pat_new();
266 			break;
267 		case PDT_PAT_CELL:
268 			ret = get_info_pat_cell();
269 			break;
270 		default:
271 			ret = pdc_mem_pdt_info(&pdt_status);
272 			break;
273 		}
274 
275 		if (ret != PDC_OK) {
276 			pr_warn("PDT: unexpected failure %d\n", ret);
277 			return -EINVAL;
278 		}
279 
280 		/* if no new PDT entries, just wait again */
281 		num = pdt_status.pdt_entries - old_num_entries;
282 		if (num <= 0)
283 			continue;
284 
285 		/* decrease poll interval in case we found memory errors */
286 		if (pdt_status.pdt_entries &&
287 			pdt_poll_interval == PDT_POLL_INTERVAL_DEFAULT)
288 			pdt_poll_interval = PDT_POLL_INTERVAL_SHORT;
289 
290 		/* limit entries to get */
291 		if (num > MAX_PDT_ENTRIES) {
292 			num = MAX_PDT_ENTRIES;
293 			pdt_status.pdt_entries = old_num_entries + num;
294 		}
295 
296 		/* get new entries */
297 		switch (pdt_type) {
298 #ifdef CONFIG_64BIT
299 		case PDT_PAT_CELL:
300 			if (pdt_status.pdt_entries > MAX_PDT_ENTRIES) {
301 				pr_crit("PDT: too many entries.\n");
302 				return -ENOMEM;
303 			}
304 			ret = pdc_pat_mem_read_cell_pdt(&pat_pret, pdt_entry,
305 				MAX_PDT_ENTRIES);
306 			bad_mem_ptr = &pdt_entry[old_num_entries];
307 			break;
308 		case PDT_PAT_NEW:
309 			ret = pdc_pat_mem_read_pd_pdt(&pat_pret,
310 				pdt_entry,
311 				num * sizeof(unsigned long),
312 				old_num_entries * sizeof(unsigned long));
313 			bad_mem_ptr = &pdt_entry[0];
314 			break;
315 #endif
316 		default:
317 			ret = pdc_mem_pdt_read_entries(&pdt_read_ret,
318 				pdt_entry);
319 			bad_mem_ptr = &pdt_entry[old_num_entries];
320 			break;
321 		}
322 
323 		/* report and mark memory broken */
324 		while (num--) {
325 			unsigned long pde = *bad_mem_ptr++;
326 
327 			report_mem_err(pde);
328 
329 #ifdef CONFIG_MEMORY_FAILURE
330 			if ((pde & PDT_ADDR_PERM_ERR) ||
331 			    ((pde & PDT_ADDR_SINGLE_ERR) == 0))
332 				memory_failure(pde >> PAGE_SHIFT, 0);
333 			else
334 				soft_offline_page(pde >> PAGE_SHIFT, 0);
335 #else
336 			pr_crit("PDT: memory error at 0x%lx ignored.\n"
337 				"Rebuild kernel with CONFIG_MEMORY_FAILURE=y "
338 				"for real handling.\n",
339 				pde & PDT_ADDR_PHYS_MASK);
340 #endif
341 
342 		}
343 	}
344 
345 	return 0;
346 }
347 
348 
349 static int __init pdt_initcall(void)
350 {
351 	struct task_struct *kpdtd_task;
352 
353 	if (pdt_type == PDT_NONE)
354 		return -ENODEV;
355 
356 	kpdtd_task = kthread_run(pdt_mainloop, NULL, "kpdtd");
357 	if (IS_ERR(kpdtd_task))
358 		return PTR_ERR(kpdtd_task);
359 
360 	return 0;
361 }
362 
363 late_initcall(pdt_initcall);
364