xref: /openbmc/linux/drivers/acpi/sysfs.c (revision 55b37d9c)
1 // SPDX-License-Identifier: GPL-2.0
2 /*
3  * sysfs.c - ACPI sysfs interface to userspace.
4  */
5 
6 #define pr_fmt(fmt) "ACPI: " fmt
7 
8 #include <linux/acpi.h>
9 #include <linux/bitmap.h>
10 #include <linux/init.h>
11 #include <linux/kernel.h>
12 #include <linux/kstrtox.h>
13 #include <linux/moduleparam.h>
14 
15 #include "internal.h"
16 
17 #ifdef CONFIG_ACPI_DEBUG
18 /*
19  * ACPI debug sysfs I/F, including:
20  * /sys/modules/acpi/parameters/debug_layer
21  * /sys/modules/acpi/parameters/debug_level
22  * /sys/modules/acpi/parameters/trace_method_name
23  * /sys/modules/acpi/parameters/trace_state
24  * /sys/modules/acpi/parameters/trace_debug_layer
25  * /sys/modules/acpi/parameters/trace_debug_level
26  */
27 
28 struct acpi_dlayer {
29 	const char *name;
30 	unsigned long value;
31 };
32 struct acpi_dlevel {
33 	const char *name;
34 	unsigned long value;
35 };
36 #define ACPI_DEBUG_INIT(v)	{ .name = #v, .value = v }
37 
38 static const struct acpi_dlayer acpi_debug_layers[] = {
39 	ACPI_DEBUG_INIT(ACPI_UTILITIES),
40 	ACPI_DEBUG_INIT(ACPI_HARDWARE),
41 	ACPI_DEBUG_INIT(ACPI_EVENTS),
42 	ACPI_DEBUG_INIT(ACPI_TABLES),
43 	ACPI_DEBUG_INIT(ACPI_NAMESPACE),
44 	ACPI_DEBUG_INIT(ACPI_PARSER),
45 	ACPI_DEBUG_INIT(ACPI_DISPATCHER),
46 	ACPI_DEBUG_INIT(ACPI_EXECUTER),
47 	ACPI_DEBUG_INIT(ACPI_RESOURCES),
48 	ACPI_DEBUG_INIT(ACPI_CA_DEBUGGER),
49 	ACPI_DEBUG_INIT(ACPI_OS_SERVICES),
50 	ACPI_DEBUG_INIT(ACPI_CA_DISASSEMBLER),
51 	ACPI_DEBUG_INIT(ACPI_COMPILER),
52 	ACPI_DEBUG_INIT(ACPI_TOOLS),
53 };
54 
55 static const struct acpi_dlevel acpi_debug_levels[] = {
56 	ACPI_DEBUG_INIT(ACPI_LV_INIT),
57 	ACPI_DEBUG_INIT(ACPI_LV_DEBUG_OBJECT),
58 	ACPI_DEBUG_INIT(ACPI_LV_INFO),
59 	ACPI_DEBUG_INIT(ACPI_LV_REPAIR),
60 	ACPI_DEBUG_INIT(ACPI_LV_TRACE_POINT),
61 
62 	ACPI_DEBUG_INIT(ACPI_LV_INIT_NAMES),
63 	ACPI_DEBUG_INIT(ACPI_LV_PARSE),
64 	ACPI_DEBUG_INIT(ACPI_LV_LOAD),
65 	ACPI_DEBUG_INIT(ACPI_LV_DISPATCH),
66 	ACPI_DEBUG_INIT(ACPI_LV_EXEC),
67 	ACPI_DEBUG_INIT(ACPI_LV_NAMES),
68 	ACPI_DEBUG_INIT(ACPI_LV_OPREGION),
69 	ACPI_DEBUG_INIT(ACPI_LV_BFIELD),
70 	ACPI_DEBUG_INIT(ACPI_LV_TABLES),
71 	ACPI_DEBUG_INIT(ACPI_LV_VALUES),
72 	ACPI_DEBUG_INIT(ACPI_LV_OBJECTS),
73 	ACPI_DEBUG_INIT(ACPI_LV_RESOURCES),
74 	ACPI_DEBUG_INIT(ACPI_LV_USER_REQUESTS),
75 	ACPI_DEBUG_INIT(ACPI_LV_PACKAGE),
76 
77 	ACPI_DEBUG_INIT(ACPI_LV_ALLOCATIONS),
78 	ACPI_DEBUG_INIT(ACPI_LV_FUNCTIONS),
79 	ACPI_DEBUG_INIT(ACPI_LV_OPTIMIZATIONS),
80 
81 	ACPI_DEBUG_INIT(ACPI_LV_MUTEX),
82 	ACPI_DEBUG_INIT(ACPI_LV_THREADS),
83 	ACPI_DEBUG_INIT(ACPI_LV_IO),
84 	ACPI_DEBUG_INIT(ACPI_LV_INTERRUPTS),
85 
86 	ACPI_DEBUG_INIT(ACPI_LV_AML_DISASSEMBLE),
87 	ACPI_DEBUG_INIT(ACPI_LV_VERBOSE_INFO),
88 	ACPI_DEBUG_INIT(ACPI_LV_FULL_TABLES),
89 	ACPI_DEBUG_INIT(ACPI_LV_EVENTS),
90 };
91 
92 static int param_get_debug_layer(char *buffer, const struct kernel_param *kp)
93 {
94 	int result = 0;
95 	int i;
96 
97 	result = sprintf(buffer, "%-25s\tHex        SET\n", "Description");
98 
99 	for (i = 0; i < ARRAY_SIZE(acpi_debug_layers); i++) {
100 		result += sprintf(buffer + result, "%-25s\t0x%08lX [%c]\n",
101 				  acpi_debug_layers[i].name,
102 				  acpi_debug_layers[i].value,
103 				  (acpi_dbg_layer & acpi_debug_layers[i].value)
104 				  ? '*' : ' ');
105 	}
106 	result +=
107 	    sprintf(buffer + result, "%-25s\t0x%08X [%c]\n", "ACPI_ALL_DRIVERS",
108 		    ACPI_ALL_DRIVERS,
109 		    (acpi_dbg_layer & ACPI_ALL_DRIVERS) ==
110 		    ACPI_ALL_DRIVERS ? '*' : (acpi_dbg_layer & ACPI_ALL_DRIVERS)
111 		    == 0 ? ' ' : '-');
112 	result +=
113 	    sprintf(buffer + result,
114 		    "--\ndebug_layer = 0x%08X ( * = enabled)\n",
115 		    acpi_dbg_layer);
116 
117 	return result;
118 }
119 
120 static int param_get_debug_level(char *buffer, const struct kernel_param *kp)
121 {
122 	int result = 0;
123 	int i;
124 
125 	result = sprintf(buffer, "%-25s\tHex        SET\n", "Description");
126 
127 	for (i = 0; i < ARRAY_SIZE(acpi_debug_levels); i++) {
128 		result += sprintf(buffer + result, "%-25s\t0x%08lX [%c]\n",
129 				  acpi_debug_levels[i].name,
130 				  acpi_debug_levels[i].value,
131 				  (acpi_dbg_level & acpi_debug_levels[i].value)
132 				  ? '*' : ' ');
133 	}
134 	result +=
135 	    sprintf(buffer + result, "--\ndebug_level = 0x%08X (* = enabled)\n",
136 		    acpi_dbg_level);
137 
138 	return result;
139 }
140 
141 static const struct kernel_param_ops param_ops_debug_layer = {
142 	.set = param_set_uint,
143 	.get = param_get_debug_layer,
144 };
145 
146 static const struct kernel_param_ops param_ops_debug_level = {
147 	.set = param_set_uint,
148 	.get = param_get_debug_level,
149 };
150 
151 module_param_cb(debug_layer, &param_ops_debug_layer, &acpi_dbg_layer, 0644);
152 module_param_cb(debug_level, &param_ops_debug_level, &acpi_dbg_level, 0644);
153 
154 static char trace_method_name[1024];
155 
156 static int param_set_trace_method_name(const char *val,
157 				       const struct kernel_param *kp)
158 {
159 	u32 saved_flags = 0;
160 	bool is_abs_path = true;
161 
162 	if (*val != '\\')
163 		is_abs_path = false;
164 
165 	if ((is_abs_path && strlen(val) > 1023) ||
166 	    (!is_abs_path && strlen(val) > 1022)) {
167 		pr_err("%s: string parameter too long\n", kp->name);
168 		return -ENOSPC;
169 	}
170 
171 	/*
172 	 * It's not safe to update acpi_gbl_trace_method_name without
173 	 * having the tracer stopped, so we save the original tracer
174 	 * state and disable it.
175 	 */
176 	saved_flags = acpi_gbl_trace_flags;
177 	(void)acpi_debug_trace(NULL,
178 			       acpi_gbl_trace_dbg_level,
179 			       acpi_gbl_trace_dbg_layer,
180 			       0);
181 
182 	/* This is a hack.  We can't kmalloc in early boot. */
183 	if (is_abs_path)
184 		strcpy(trace_method_name, val);
185 	else {
186 		trace_method_name[0] = '\\';
187 		strcpy(trace_method_name+1, val);
188 	}
189 
190 	/* Restore the original tracer state */
191 	(void)acpi_debug_trace(trace_method_name,
192 			       acpi_gbl_trace_dbg_level,
193 			       acpi_gbl_trace_dbg_layer,
194 			       saved_flags);
195 
196 	return 0;
197 }
198 
199 static int param_get_trace_method_name(char *buffer, const struct kernel_param *kp)
200 {
201 	return sysfs_emit(buffer, "%s\n", acpi_gbl_trace_method_name);
202 }
203 
204 static const struct kernel_param_ops param_ops_trace_method = {
205 	.set = param_set_trace_method_name,
206 	.get = param_get_trace_method_name,
207 };
208 
209 static const struct kernel_param_ops param_ops_trace_attrib = {
210 	.set = param_set_uint,
211 	.get = param_get_uint,
212 };
213 
214 module_param_cb(trace_method_name, &param_ops_trace_method, &trace_method_name, 0644);
215 module_param_cb(trace_debug_layer, &param_ops_trace_attrib, &acpi_gbl_trace_dbg_layer, 0644);
216 module_param_cb(trace_debug_level, &param_ops_trace_attrib, &acpi_gbl_trace_dbg_level, 0644);
217 
218 static int param_set_trace_state(const char *val,
219 				 const struct kernel_param *kp)
220 {
221 	acpi_status status;
222 	const char *method = trace_method_name;
223 	u32 flags = 0;
224 
225 /* So "xxx-once" comparison should go prior than "xxx" comparison */
226 #define acpi_compare_param(val, key)	\
227 	strncmp((val), (key), sizeof(key) - 1)
228 
229 	if (!acpi_compare_param(val, "enable")) {
230 		method = NULL;
231 		flags = ACPI_TRACE_ENABLED;
232 	} else if (!acpi_compare_param(val, "disable"))
233 		method = NULL;
234 	else if (!acpi_compare_param(val, "method-once"))
235 		flags = ACPI_TRACE_ENABLED | ACPI_TRACE_ONESHOT;
236 	else if (!acpi_compare_param(val, "method"))
237 		flags = ACPI_TRACE_ENABLED;
238 	else if (!acpi_compare_param(val, "opcode-once"))
239 		flags = ACPI_TRACE_ENABLED | ACPI_TRACE_ONESHOT | ACPI_TRACE_OPCODE;
240 	else if (!acpi_compare_param(val, "opcode"))
241 		flags = ACPI_TRACE_ENABLED | ACPI_TRACE_OPCODE;
242 	else
243 		return -EINVAL;
244 
245 	status = acpi_debug_trace(method,
246 				  acpi_gbl_trace_dbg_level,
247 				  acpi_gbl_trace_dbg_layer,
248 				  flags);
249 	if (ACPI_FAILURE(status))
250 		return -EBUSY;
251 
252 	return 0;
253 }
254 
255 static int param_get_trace_state(char *buffer, const struct kernel_param *kp)
256 {
257 	if (!(acpi_gbl_trace_flags & ACPI_TRACE_ENABLED))
258 		return sprintf(buffer, "disable\n");
259 	if (!acpi_gbl_trace_method_name)
260 		return sprintf(buffer, "enable\n");
261 	if (acpi_gbl_trace_flags & ACPI_TRACE_ONESHOT)
262 		return sprintf(buffer, "method-once\n");
263 	else
264 		return sprintf(buffer, "method\n");
265 }
266 
267 module_param_call(trace_state, param_set_trace_state, param_get_trace_state,
268 		  NULL, 0644);
269 #endif /* CONFIG_ACPI_DEBUG */
270 
271 
272 /* /sys/modules/acpi/parameters/aml_debug_output */
273 
274 module_param_named(aml_debug_output, acpi_gbl_enable_aml_debug_object,
275 		   byte, 0644);
276 MODULE_PARM_DESC(aml_debug_output,
277 		 "To enable/disable the ACPI Debug Object output.");
278 
279 /* /sys/module/acpi/parameters/acpica_version */
280 static int param_get_acpica_version(char *buffer,
281 				    const struct kernel_param *kp)
282 {
283 	int result;
284 
285 	result = sprintf(buffer, "%x\n", ACPI_CA_VERSION);
286 
287 	return result;
288 }
289 
290 module_param_call(acpica_version, NULL, param_get_acpica_version, NULL, 0444);
291 
292 /*
293  * ACPI table sysfs I/F:
294  * /sys/firmware/acpi/tables/
295  * /sys/firmware/acpi/tables/data/
296  * /sys/firmware/acpi/tables/dynamic/
297  */
298 
299 static LIST_HEAD(acpi_table_attr_list);
300 static struct kobject *tables_kobj;
301 static struct kobject *tables_data_kobj;
302 static struct kobject *dynamic_tables_kobj;
303 static struct kobject *hotplug_kobj;
304 
305 #define ACPI_MAX_TABLE_INSTANCES	999
306 #define ACPI_INST_SIZE			4 /* including trailing 0 */
307 
308 struct acpi_table_attr {
309 	struct bin_attribute attr;
310 	char name[ACPI_NAMESEG_SIZE];
311 	int instance;
312 	char filename[ACPI_NAMESEG_SIZE+ACPI_INST_SIZE];
313 	struct list_head node;
314 };
315 
316 struct acpi_data_attr {
317 	struct bin_attribute attr;
318 	u64	addr;
319 };
320 
321 static ssize_t acpi_table_show(struct file *filp, struct kobject *kobj,
322 			       struct bin_attribute *bin_attr, char *buf,
323 			       loff_t offset, size_t count)
324 {
325 	struct acpi_table_attr *table_attr =
326 	    container_of(bin_attr, struct acpi_table_attr, attr);
327 	struct acpi_table_header *table_header = NULL;
328 	acpi_status status;
329 	ssize_t rc;
330 
331 	status = acpi_get_table(table_attr->name, table_attr->instance,
332 				&table_header);
333 	if (ACPI_FAILURE(status))
334 		return -ENODEV;
335 
336 	rc = memory_read_from_buffer(buf, count, &offset, table_header,
337 			table_header->length);
338 	acpi_put_table(table_header);
339 	return rc;
340 }
341 
342 static int acpi_table_attr_init(struct kobject *tables_obj,
343 				struct acpi_table_attr *table_attr,
344 				struct acpi_table_header *table_header)
345 {
346 	struct acpi_table_header *header = NULL;
347 	struct acpi_table_attr *attr = NULL;
348 	char instance_str[ACPI_INST_SIZE];
349 
350 	sysfs_attr_init(&table_attr->attr.attr);
351 	ACPI_COPY_NAMESEG(table_attr->name, table_header->signature);
352 
353 	list_for_each_entry(attr, &acpi_table_attr_list, node) {
354 		if (ACPI_COMPARE_NAMESEG(table_attr->name, attr->name))
355 			if (table_attr->instance < attr->instance)
356 				table_attr->instance = attr->instance;
357 	}
358 	table_attr->instance++;
359 	if (table_attr->instance > ACPI_MAX_TABLE_INSTANCES) {
360 		pr_warn("%4.4s: too many table instances\n", table_attr->name);
361 		return -ERANGE;
362 	}
363 
364 	ACPI_COPY_NAMESEG(table_attr->filename, table_header->signature);
365 	table_attr->filename[ACPI_NAMESEG_SIZE] = '\0';
366 	if (table_attr->instance > 1 || (table_attr->instance == 1 &&
367 					 !acpi_get_table
368 					 (table_header->signature, 2, &header))) {
369 		snprintf(instance_str, sizeof(instance_str), "%u",
370 			 table_attr->instance);
371 		strcat(table_attr->filename, instance_str);
372 	}
373 
374 	table_attr->attr.size = table_header->length;
375 	table_attr->attr.read = acpi_table_show;
376 	table_attr->attr.attr.name = table_attr->filename;
377 	table_attr->attr.attr.mode = 0400;
378 
379 	return sysfs_create_bin_file(tables_obj, &table_attr->attr);
380 }
381 
382 acpi_status acpi_sysfs_table_handler(u32 event, void *table, void *context)
383 {
384 	struct acpi_table_attr *table_attr;
385 
386 	switch (event) {
387 	case ACPI_TABLE_EVENT_INSTALL:
388 		table_attr = kzalloc(sizeof(*table_attr), GFP_KERNEL);
389 		if (!table_attr)
390 			return AE_NO_MEMORY;
391 
392 		if (acpi_table_attr_init(dynamic_tables_kobj,
393 					 table_attr, table)) {
394 			kfree(table_attr);
395 			return AE_ERROR;
396 		}
397 		list_add_tail(&table_attr->node, &acpi_table_attr_list);
398 		break;
399 	case ACPI_TABLE_EVENT_LOAD:
400 	case ACPI_TABLE_EVENT_UNLOAD:
401 	case ACPI_TABLE_EVENT_UNINSTALL:
402 		/*
403 		 * we do not need to do anything right now
404 		 * because the table is not deleted from the
405 		 * global table list when unloading it.
406 		 */
407 		break;
408 	default:
409 		return AE_BAD_PARAMETER;
410 	}
411 	return AE_OK;
412 }
413 
414 static ssize_t acpi_data_show(struct file *filp, struct kobject *kobj,
415 			      struct bin_attribute *bin_attr, char *buf,
416 			      loff_t offset, size_t count)
417 {
418 	struct acpi_data_attr *data_attr;
419 	void __iomem *base;
420 	ssize_t size;
421 
422 	data_attr = container_of(bin_attr, struct acpi_data_attr, attr);
423 	size = data_attr->attr.size;
424 
425 	if (offset < 0)
426 		return -EINVAL;
427 
428 	if (offset >= size)
429 		return 0;
430 
431 	if (count > size - offset)
432 		count = size - offset;
433 
434 	base = acpi_os_map_iomem(data_attr->addr, size);
435 	if (!base)
436 		return -ENOMEM;
437 
438 	memcpy_fromio(buf, base + offset, count);
439 
440 	acpi_os_unmap_iomem(base, size);
441 
442 	return count;
443 }
444 
445 static int acpi_bert_data_init(void *th, struct acpi_data_attr *data_attr)
446 {
447 	struct acpi_table_bert *bert = th;
448 
449 	if (bert->header.length < sizeof(struct acpi_table_bert) ||
450 	    bert->region_length < sizeof(struct acpi_hest_generic_status)) {
451 		kfree(data_attr);
452 		return -EINVAL;
453 	}
454 	data_attr->addr = bert->address;
455 	data_attr->attr.size = bert->region_length;
456 	data_attr->attr.attr.name = "BERT";
457 
458 	return sysfs_create_bin_file(tables_data_kobj, &data_attr->attr);
459 }
460 
461 static int acpi_ccel_data_init(void *th, struct acpi_data_attr *data_attr)
462 {
463 	struct acpi_table_ccel *ccel = th;
464 
465 	if (ccel->header.length < sizeof(struct acpi_table_ccel) ||
466 	    !ccel->log_area_start_address || !ccel->log_area_minimum_length) {
467 		kfree(data_attr);
468 		return -EINVAL;
469 	}
470 	data_attr->addr = ccel->log_area_start_address;
471 	data_attr->attr.size = ccel->log_area_minimum_length;
472 	data_attr->attr.attr.name = "CCEL";
473 
474 	return sysfs_create_bin_file(tables_data_kobj, &data_attr->attr);
475 }
476 
477 static struct acpi_data_obj {
478 	char *name;
479 	int (*fn)(void *, struct acpi_data_attr *);
480 } acpi_data_objs[] = {
481 	{ ACPI_SIG_BERT, acpi_bert_data_init },
482 	{ ACPI_SIG_CCEL, acpi_ccel_data_init },
483 };
484 
485 #define NUM_ACPI_DATA_OBJS ARRAY_SIZE(acpi_data_objs)
486 
487 static int acpi_table_data_init(struct acpi_table_header *th)
488 {
489 	struct acpi_data_attr *data_attr;
490 	int i;
491 
492 	for (i = 0; i < NUM_ACPI_DATA_OBJS; i++) {
493 		if (ACPI_COMPARE_NAMESEG(th->signature, acpi_data_objs[i].name)) {
494 			data_attr = kzalloc(sizeof(*data_attr), GFP_KERNEL);
495 			if (!data_attr)
496 				return -ENOMEM;
497 			sysfs_attr_init(&data_attr->attr.attr);
498 			data_attr->attr.read = acpi_data_show;
499 			data_attr->attr.attr.mode = 0400;
500 			return acpi_data_objs[i].fn(th, data_attr);
501 		}
502 	}
503 	return 0;
504 }
505 
506 static int acpi_tables_sysfs_init(void)
507 {
508 	struct acpi_table_attr *table_attr;
509 	struct acpi_table_header *table_header = NULL;
510 	int table_index;
511 	acpi_status status;
512 	int ret;
513 
514 	tables_kobj = kobject_create_and_add("tables", acpi_kobj);
515 	if (!tables_kobj)
516 		goto err;
517 
518 	tables_data_kobj = kobject_create_and_add("data", tables_kobj);
519 	if (!tables_data_kobj)
520 		goto err_tables_data;
521 
522 	dynamic_tables_kobj = kobject_create_and_add("dynamic", tables_kobj);
523 	if (!dynamic_tables_kobj)
524 		goto err_dynamic_tables;
525 
526 	for (table_index = 0;; table_index++) {
527 		status = acpi_get_table_by_index(table_index, &table_header);
528 
529 		if (status == AE_BAD_PARAMETER)
530 			break;
531 
532 		if (ACPI_FAILURE(status))
533 			continue;
534 
535 		table_attr = kzalloc(sizeof(*table_attr), GFP_KERNEL);
536 		if (!table_attr)
537 			return -ENOMEM;
538 
539 		ret = acpi_table_attr_init(tables_kobj,
540 					   table_attr, table_header);
541 		if (ret) {
542 			kfree(table_attr);
543 			return ret;
544 		}
545 		list_add_tail(&table_attr->node, &acpi_table_attr_list);
546 		acpi_table_data_init(table_header);
547 	}
548 
549 	kobject_uevent(tables_kobj, KOBJ_ADD);
550 	kobject_uevent(tables_data_kobj, KOBJ_ADD);
551 	kobject_uevent(dynamic_tables_kobj, KOBJ_ADD);
552 
553 	return 0;
554 err_dynamic_tables:
555 	kobject_put(tables_data_kobj);
556 err_tables_data:
557 	kobject_put(tables_kobj);
558 err:
559 	return -ENOMEM;
560 }
561 
562 /*
563  * Detailed ACPI IRQ counters:
564  * /sys/firmware/acpi/interrupts/
565  */
566 
567 u32 acpi_irq_handled;
568 u32 acpi_irq_not_handled;
569 
570 #define COUNT_GPE 0
571 #define COUNT_SCI 1		/* acpi_irq_handled */
572 #define COUNT_SCI_NOT 2		/* acpi_irq_not_handled */
573 #define COUNT_ERROR 3		/* other */
574 #define NUM_COUNTERS_EXTRA 4
575 
576 struct event_counter {
577 	u32 count;
578 	u32 flags;
579 };
580 
581 static struct event_counter *all_counters;
582 static u32 num_gpes;
583 static u32 num_counters;
584 static struct attribute **all_attrs;
585 static u32 acpi_gpe_count;
586 
587 static struct attribute_group interrupt_stats_attr_group = {
588 	.name = "interrupts",
589 };
590 
591 static struct kobj_attribute *counter_attrs;
592 
593 static void delete_gpe_attr_array(void)
594 {
595 	struct event_counter *tmp = all_counters;
596 
597 	all_counters = NULL;
598 	kfree(tmp);
599 
600 	if (counter_attrs) {
601 		int i;
602 
603 		for (i = 0; i < num_gpes; i++)
604 			kfree(counter_attrs[i].attr.name);
605 
606 		kfree(counter_attrs);
607 	}
608 	kfree(all_attrs);
609 }
610 
611 static void gpe_count(u32 gpe_number)
612 {
613 	acpi_gpe_count++;
614 
615 	if (!all_counters)
616 		return;
617 
618 	if (gpe_number < num_gpes)
619 		all_counters[gpe_number].count++;
620 	else
621 		all_counters[num_gpes + ACPI_NUM_FIXED_EVENTS +
622 			     COUNT_ERROR].count++;
623 }
624 
625 static void fixed_event_count(u32 event_number)
626 {
627 	if (!all_counters)
628 		return;
629 
630 	if (event_number < ACPI_NUM_FIXED_EVENTS)
631 		all_counters[num_gpes + event_number].count++;
632 	else
633 		all_counters[num_gpes + ACPI_NUM_FIXED_EVENTS +
634 			     COUNT_ERROR].count++;
635 }
636 
637 static void acpi_global_event_handler(u32 event_type, acpi_handle device,
638 	u32 event_number, void *context)
639 {
640 	if (event_type == ACPI_EVENT_TYPE_GPE) {
641 		gpe_count(event_number);
642 		pr_debug("GPE event 0x%02x\n", event_number);
643 	} else if (event_type == ACPI_EVENT_TYPE_FIXED) {
644 		fixed_event_count(event_number);
645 		pr_debug("Fixed event 0x%02x\n", event_number);
646 	} else {
647 		pr_debug("Other event 0x%02x\n", event_number);
648 	}
649 }
650 
651 static int get_status(u32 index, acpi_event_status *ret,
652 		      acpi_handle *handle)
653 {
654 	acpi_status status;
655 
656 	if (index >= num_gpes + ACPI_NUM_FIXED_EVENTS)
657 		return -EINVAL;
658 
659 	if (index < num_gpes) {
660 		status = acpi_get_gpe_device(index, handle);
661 		if (ACPI_FAILURE(status)) {
662 			pr_warn("Invalid GPE 0x%x", index);
663 			return -ENXIO;
664 		}
665 		status = acpi_get_gpe_status(*handle, index, ret);
666 	} else {
667 		status = acpi_get_event_status(index - num_gpes, ret);
668 	}
669 	if (ACPI_FAILURE(status))
670 		return -EIO;
671 
672 	return 0;
673 }
674 
675 static ssize_t counter_show(struct kobject *kobj,
676 			    struct kobj_attribute *attr, char *buf)
677 {
678 	int index = attr - counter_attrs;
679 	int size;
680 	acpi_handle handle;
681 	acpi_event_status status;
682 	int result = 0;
683 
684 	all_counters[num_gpes + ACPI_NUM_FIXED_EVENTS + COUNT_SCI].count =
685 	    acpi_irq_handled;
686 	all_counters[num_gpes + ACPI_NUM_FIXED_EVENTS + COUNT_SCI_NOT].count =
687 	    acpi_irq_not_handled;
688 	all_counters[num_gpes + ACPI_NUM_FIXED_EVENTS + COUNT_GPE].count =
689 	    acpi_gpe_count;
690 	size = sprintf(buf, "%8u", all_counters[index].count);
691 
692 	/* "gpe_all" or "sci" */
693 	if (index >= num_gpes + ACPI_NUM_FIXED_EVENTS)
694 		goto end;
695 
696 	result = get_status(index, &status, &handle);
697 	if (result)
698 		goto end;
699 
700 	if (status & ACPI_EVENT_FLAG_ENABLE_SET)
701 		size += sprintf(buf + size, "  EN");
702 	else
703 		size += sprintf(buf + size, "    ");
704 	if (status & ACPI_EVENT_FLAG_STATUS_SET)
705 		size += sprintf(buf + size, " STS");
706 	else
707 		size += sprintf(buf + size, "    ");
708 
709 	if (!(status & ACPI_EVENT_FLAG_HAS_HANDLER))
710 		size += sprintf(buf + size, " invalid     ");
711 	else if (status & ACPI_EVENT_FLAG_ENABLED)
712 		size += sprintf(buf + size, " enabled     ");
713 	else if (status & ACPI_EVENT_FLAG_WAKE_ENABLED)
714 		size += sprintf(buf + size, " wake_enabled");
715 	else
716 		size += sprintf(buf + size, " disabled    ");
717 	if (status & ACPI_EVENT_FLAG_MASKED)
718 		size += sprintf(buf + size, " masked  ");
719 	else
720 		size += sprintf(buf + size, " unmasked");
721 
722 end:
723 	size += sprintf(buf + size, "\n");
724 	return result ? result : size;
725 }
726 
727 /*
728  * counter_set() sets the specified counter.
729  * setting the total "sci" file to any value clears all counters.
730  * enable/disable/clear a gpe/fixed event in user space.
731  */
732 static ssize_t counter_set(struct kobject *kobj,
733 			   struct kobj_attribute *attr, const char *buf,
734 			   size_t size)
735 {
736 	int index = attr - counter_attrs;
737 	acpi_event_status status;
738 	acpi_handle handle;
739 	int result = 0;
740 	unsigned long tmp;
741 
742 	if (index == num_gpes + ACPI_NUM_FIXED_EVENTS + COUNT_SCI) {
743 		int i;
744 		for (i = 0; i < num_counters; ++i)
745 			all_counters[i].count = 0;
746 		acpi_gpe_count = 0;
747 		acpi_irq_handled = 0;
748 		acpi_irq_not_handled = 0;
749 		goto end;
750 	}
751 
752 	/* show the event status for both GPEs and Fixed Events */
753 	result = get_status(index, &status, &handle);
754 	if (result)
755 		goto end;
756 
757 	if (!(status & ACPI_EVENT_FLAG_HAS_HANDLER)) {
758 		pr_warn("Can not change Invalid GPE/Fixed Event status\n");
759 		return -EINVAL;
760 	}
761 
762 	if (index < num_gpes) {
763 		if (!strcmp(buf, "disable\n") &&
764 		    (status & ACPI_EVENT_FLAG_ENABLED))
765 			result = acpi_disable_gpe(handle, index);
766 		else if (!strcmp(buf, "enable\n") &&
767 			 !(status & ACPI_EVENT_FLAG_ENABLED))
768 			result = acpi_enable_gpe(handle, index);
769 		else if (!strcmp(buf, "clear\n") &&
770 			 (status & ACPI_EVENT_FLAG_STATUS_SET))
771 			result = acpi_clear_gpe(handle, index);
772 		else if (!strcmp(buf, "mask\n"))
773 			result = acpi_mask_gpe(handle, index, TRUE);
774 		else if (!strcmp(buf, "unmask\n"))
775 			result = acpi_mask_gpe(handle, index, FALSE);
776 		else if (!kstrtoul(buf, 0, &tmp))
777 			all_counters[index].count = tmp;
778 		else
779 			result = -EINVAL;
780 	} else if (index < num_gpes + ACPI_NUM_FIXED_EVENTS) {
781 		int event = index - num_gpes;
782 		if (!strcmp(buf, "disable\n") &&
783 		    (status & ACPI_EVENT_FLAG_ENABLE_SET))
784 			result = acpi_disable_event(event, ACPI_NOT_ISR);
785 		else if (!strcmp(buf, "enable\n") &&
786 			 !(status & ACPI_EVENT_FLAG_ENABLE_SET))
787 			result = acpi_enable_event(event, ACPI_NOT_ISR);
788 		else if (!strcmp(buf, "clear\n") &&
789 			 (status & ACPI_EVENT_FLAG_STATUS_SET))
790 			result = acpi_clear_event(event);
791 		else if (!kstrtoul(buf, 0, &tmp))
792 			all_counters[index].count = tmp;
793 		else
794 			result = -EINVAL;
795 	} else
796 		all_counters[index].count = strtoul(buf, NULL, 0);
797 
798 	if (ACPI_FAILURE(result))
799 		result = -EINVAL;
800 end:
801 	return result ? result : size;
802 }
803 
804 /*
805  * A Quirk Mechanism for GPE Flooding Prevention:
806  *
807  * Quirks may be needed to prevent GPE flooding on a specific GPE. The
808  * flooding typically cannot be detected and automatically prevented by
809  * ACPI_GPE_DISPATCH_NONE check because there is a _Lxx/_Exx prepared in
810  * the AML tables. This normally indicates a feature gap in Linux, thus
811  * instead of providing endless quirk tables, we provide a boot parameter
812  * for those who want this quirk. For example, if the users want to prevent
813  * the GPE flooding for GPE 00, they need to specify the following boot
814  * parameter:
815  *   acpi_mask_gpe=0x00
816  * Note, the parameter can be a list (see bitmap_parselist() for the details).
817  * The masking status can be modified by the following runtime controlling
818  * interface:
819  *   echo unmask > /sys/firmware/acpi/interrupts/gpe00
820  */
821 #define ACPI_MASKABLE_GPE_MAX	0x100
822 static DECLARE_BITMAP(acpi_masked_gpes_map, ACPI_MASKABLE_GPE_MAX) __initdata;
823 
824 static int __init acpi_gpe_set_masked_gpes(char *val)
825 {
826 	int ret;
827 	u8 gpe;
828 
829 	ret = kstrtou8(val, 0, &gpe);
830 	if (ret) {
831 		ret = bitmap_parselist(val, acpi_masked_gpes_map, ACPI_MASKABLE_GPE_MAX);
832 		if (ret)
833 			return ret;
834 	} else
835 		set_bit(gpe, acpi_masked_gpes_map);
836 
837 	return 1;
838 }
839 __setup("acpi_mask_gpe=", acpi_gpe_set_masked_gpes);
840 
841 void __init acpi_gpe_apply_masked_gpes(void)
842 {
843 	acpi_handle handle;
844 	acpi_status status;
845 	u16 gpe;
846 
847 	for_each_set_bit(gpe, acpi_masked_gpes_map, ACPI_MASKABLE_GPE_MAX) {
848 		status = acpi_get_gpe_device(gpe, &handle);
849 		if (ACPI_SUCCESS(status)) {
850 			pr_info("Masking GPE 0x%x.\n", gpe);
851 			(void)acpi_mask_gpe(handle, gpe, TRUE);
852 		}
853 	}
854 }
855 
856 void acpi_irq_stats_init(void)
857 {
858 	acpi_status status;
859 	int i;
860 
861 	if (all_counters)
862 		return;
863 
864 	num_gpes = acpi_current_gpe_count;
865 	num_counters = num_gpes + ACPI_NUM_FIXED_EVENTS + NUM_COUNTERS_EXTRA;
866 
867 	all_attrs = kcalloc(num_counters + 1, sizeof(*all_attrs), GFP_KERNEL);
868 	if (all_attrs == NULL)
869 		return;
870 
871 	all_counters = kcalloc(num_counters, sizeof(*all_counters), GFP_KERNEL);
872 	if (all_counters == NULL)
873 		goto fail;
874 
875 	status = acpi_install_global_event_handler(acpi_global_event_handler, NULL);
876 	if (ACPI_FAILURE(status))
877 		goto fail;
878 
879 	counter_attrs = kcalloc(num_counters, sizeof(*counter_attrs), GFP_KERNEL);
880 	if (counter_attrs == NULL)
881 		goto fail;
882 
883 	for (i = 0; i < num_counters; ++i) {
884 		char buffer[12];
885 		char *name;
886 
887 		if (i < num_gpes)
888 			sprintf(buffer, "gpe%02X", i);
889 		else if (i == num_gpes + ACPI_EVENT_PMTIMER)
890 			sprintf(buffer, "ff_pmtimer");
891 		else if (i == num_gpes + ACPI_EVENT_GLOBAL)
892 			sprintf(buffer, "ff_gbl_lock");
893 		else if (i == num_gpes + ACPI_EVENT_POWER_BUTTON)
894 			sprintf(buffer, "ff_pwr_btn");
895 		else if (i == num_gpes + ACPI_EVENT_SLEEP_BUTTON)
896 			sprintf(buffer, "ff_slp_btn");
897 		else if (i == num_gpes + ACPI_EVENT_RTC)
898 			sprintf(buffer, "ff_rt_clk");
899 		else if (i == num_gpes + ACPI_NUM_FIXED_EVENTS + COUNT_GPE)
900 			sprintf(buffer, "gpe_all");
901 		else if (i == num_gpes + ACPI_NUM_FIXED_EVENTS + COUNT_SCI)
902 			sprintf(buffer, "sci");
903 		else if (i == num_gpes + ACPI_NUM_FIXED_EVENTS + COUNT_SCI_NOT)
904 			sprintf(buffer, "sci_not");
905 		else if (i == num_gpes + ACPI_NUM_FIXED_EVENTS + COUNT_ERROR)
906 			sprintf(buffer, "error");
907 		else
908 			sprintf(buffer, "bug%02X", i);
909 
910 		name = kstrdup(buffer, GFP_KERNEL);
911 		if (name == NULL)
912 			goto fail;
913 
914 		sysfs_attr_init(&counter_attrs[i].attr);
915 		counter_attrs[i].attr.name = name;
916 		counter_attrs[i].attr.mode = 0644;
917 		counter_attrs[i].show = counter_show;
918 		counter_attrs[i].store = counter_set;
919 
920 		all_attrs[i] = &counter_attrs[i].attr;
921 	}
922 
923 	interrupt_stats_attr_group.attrs = all_attrs;
924 	if (!sysfs_create_group(acpi_kobj, &interrupt_stats_attr_group))
925 		return;
926 
927 fail:
928 	delete_gpe_attr_array();
929 }
930 
931 static void __exit interrupt_stats_exit(void)
932 {
933 	sysfs_remove_group(acpi_kobj, &interrupt_stats_attr_group);
934 
935 	delete_gpe_attr_array();
936 }
937 
938 static ssize_t pm_profile_show(struct kobject *kobj, struct kobj_attribute *attr, char *buf)
939 {
940 	return sprintf(buf, "%d\n", acpi_gbl_FADT.preferred_profile);
941 }
942 
943 static const struct kobj_attribute pm_profile_attr = __ATTR_RO(pm_profile);
944 
945 static ssize_t enabled_show(struct kobject *kobj, struct kobj_attribute *attr, char *buf)
946 {
947 	struct acpi_hotplug_profile *hotplug = to_acpi_hotplug_profile(kobj);
948 
949 	return sprintf(buf, "%d\n", hotplug->enabled);
950 }
951 
952 static ssize_t enabled_store(struct kobject *kobj, struct kobj_attribute *attr,
953 			     const char *buf, size_t size)
954 {
955 	struct acpi_hotplug_profile *hotplug = to_acpi_hotplug_profile(kobj);
956 	unsigned int val;
957 
958 	if (kstrtouint(buf, 10, &val) || val > 1)
959 		return -EINVAL;
960 
961 	acpi_scan_hotplug_enabled(hotplug, val);
962 	return size;
963 }
964 
965 static struct kobj_attribute hotplug_enabled_attr = __ATTR_RW(enabled);
966 
967 static struct attribute *hotplug_profile_attrs[] = {
968 	&hotplug_enabled_attr.attr,
969 	NULL
970 };
971 ATTRIBUTE_GROUPS(hotplug_profile);
972 
973 static const struct kobj_type acpi_hotplug_profile_ktype = {
974 	.sysfs_ops = &kobj_sysfs_ops,
975 	.default_groups = hotplug_profile_groups,
976 };
977 
978 void acpi_sysfs_add_hotplug_profile(struct acpi_hotplug_profile *hotplug,
979 				    const char *name)
980 {
981 	int error;
982 
983 	if (!hotplug_kobj)
984 		goto err_out;
985 
986 	error = kobject_init_and_add(&hotplug->kobj,
987 		&acpi_hotplug_profile_ktype, hotplug_kobj, "%s", name);
988 	if (error) {
989 		kobject_put(&hotplug->kobj);
990 		goto err_out;
991 	}
992 
993 	kobject_uevent(&hotplug->kobj, KOBJ_ADD);
994 	return;
995 
996  err_out:
997 	pr_err("Unable to add hotplug profile '%s'\n", name);
998 }
999 
1000 static ssize_t force_remove_show(struct kobject *kobj,
1001 				 struct kobj_attribute *attr, char *buf)
1002 {
1003 	return sprintf(buf, "%d\n", 0);
1004 }
1005 
1006 static ssize_t force_remove_store(struct kobject *kobj,
1007 				  struct kobj_attribute *attr,
1008 				  const char *buf, size_t size)
1009 {
1010 	bool val;
1011 	int ret;
1012 
1013 	ret = kstrtobool(buf, &val);
1014 	if (ret < 0)
1015 		return ret;
1016 
1017 	if (val) {
1018 		pr_err("Enabling force_remove is not supported anymore. Please report to linux-acpi@vger.kernel.org if you depend on this functionality\n");
1019 		return -EINVAL;
1020 	}
1021 	return size;
1022 }
1023 
1024 static const struct kobj_attribute force_remove_attr = __ATTR_RW(force_remove);
1025 
1026 int __init acpi_sysfs_init(void)
1027 {
1028 	int result;
1029 
1030 	result = acpi_tables_sysfs_init();
1031 	if (result)
1032 		return result;
1033 
1034 	hotplug_kobj = kobject_create_and_add("hotplug", acpi_kobj);
1035 	if (!hotplug_kobj)
1036 		return -ENOMEM;
1037 
1038 	result = sysfs_create_file(hotplug_kobj, &force_remove_attr.attr);
1039 	if (result)
1040 		return result;
1041 
1042 	result = sysfs_create_file(acpi_kobj, &pm_profile_attr.attr);
1043 	return result;
1044 }
1045