xref: /openbmc/linux/drivers/acpi/sysfs.c (revision 31368ce8)
1 /*
2  * sysfs.c - ACPI sysfs interface to userspace.
3  */
4 
5 #include <linux/init.h>
6 #include <linux/kernel.h>
7 #include <linux/moduleparam.h>
8 #include <linux/acpi.h>
9 
10 #include "internal.h"
11 
12 #define _COMPONENT		ACPI_SYSTEM_COMPONENT
13 ACPI_MODULE_NAME("sysfs");
14 
15 #ifdef CONFIG_ACPI_DEBUG
16 /*
17  * ACPI debug sysfs I/F, including:
18  * /sys/modules/acpi/parameters/debug_layer
19  * /sys/modules/acpi/parameters/debug_level
20  * /sys/modules/acpi/parameters/trace_method_name
21  * /sys/modules/acpi/parameters/trace_state
22  * /sys/modules/acpi/parameters/trace_debug_layer
23  * /sys/modules/acpi/parameters/trace_debug_level
24  */
25 
26 struct acpi_dlayer {
27 	const char *name;
28 	unsigned long value;
29 };
30 struct acpi_dlevel {
31 	const char *name;
32 	unsigned long value;
33 };
34 #define ACPI_DEBUG_INIT(v)	{ .name = #v, .value = v }
35 
36 static const struct acpi_dlayer acpi_debug_layers[] = {
37 	ACPI_DEBUG_INIT(ACPI_UTILITIES),
38 	ACPI_DEBUG_INIT(ACPI_HARDWARE),
39 	ACPI_DEBUG_INIT(ACPI_EVENTS),
40 	ACPI_DEBUG_INIT(ACPI_TABLES),
41 	ACPI_DEBUG_INIT(ACPI_NAMESPACE),
42 	ACPI_DEBUG_INIT(ACPI_PARSER),
43 	ACPI_DEBUG_INIT(ACPI_DISPATCHER),
44 	ACPI_DEBUG_INIT(ACPI_EXECUTER),
45 	ACPI_DEBUG_INIT(ACPI_RESOURCES),
46 	ACPI_DEBUG_INIT(ACPI_CA_DEBUGGER),
47 	ACPI_DEBUG_INIT(ACPI_OS_SERVICES),
48 	ACPI_DEBUG_INIT(ACPI_CA_DISASSEMBLER),
49 	ACPI_DEBUG_INIT(ACPI_COMPILER),
50 	ACPI_DEBUG_INIT(ACPI_TOOLS),
51 
52 	ACPI_DEBUG_INIT(ACPI_BUS_COMPONENT),
53 	ACPI_DEBUG_INIT(ACPI_AC_COMPONENT),
54 	ACPI_DEBUG_INIT(ACPI_BATTERY_COMPONENT),
55 	ACPI_DEBUG_INIT(ACPI_BUTTON_COMPONENT),
56 	ACPI_DEBUG_INIT(ACPI_SBS_COMPONENT),
57 	ACPI_DEBUG_INIT(ACPI_FAN_COMPONENT),
58 	ACPI_DEBUG_INIT(ACPI_PCI_COMPONENT),
59 	ACPI_DEBUG_INIT(ACPI_POWER_COMPONENT),
60 	ACPI_DEBUG_INIT(ACPI_CONTAINER_COMPONENT),
61 	ACPI_DEBUG_INIT(ACPI_SYSTEM_COMPONENT),
62 	ACPI_DEBUG_INIT(ACPI_THERMAL_COMPONENT),
63 	ACPI_DEBUG_INIT(ACPI_MEMORY_DEVICE_COMPONENT),
64 	ACPI_DEBUG_INIT(ACPI_VIDEO_COMPONENT),
65 	ACPI_DEBUG_INIT(ACPI_PROCESSOR_COMPONENT),
66 };
67 
68 static const struct acpi_dlevel acpi_debug_levels[] = {
69 	ACPI_DEBUG_INIT(ACPI_LV_INIT),
70 	ACPI_DEBUG_INIT(ACPI_LV_DEBUG_OBJECT),
71 	ACPI_DEBUG_INIT(ACPI_LV_INFO),
72 	ACPI_DEBUG_INIT(ACPI_LV_REPAIR),
73 	ACPI_DEBUG_INIT(ACPI_LV_TRACE_POINT),
74 
75 	ACPI_DEBUG_INIT(ACPI_LV_INIT_NAMES),
76 	ACPI_DEBUG_INIT(ACPI_LV_PARSE),
77 	ACPI_DEBUG_INIT(ACPI_LV_LOAD),
78 	ACPI_DEBUG_INIT(ACPI_LV_DISPATCH),
79 	ACPI_DEBUG_INIT(ACPI_LV_EXEC),
80 	ACPI_DEBUG_INIT(ACPI_LV_NAMES),
81 	ACPI_DEBUG_INIT(ACPI_LV_OPREGION),
82 	ACPI_DEBUG_INIT(ACPI_LV_BFIELD),
83 	ACPI_DEBUG_INIT(ACPI_LV_TABLES),
84 	ACPI_DEBUG_INIT(ACPI_LV_VALUES),
85 	ACPI_DEBUG_INIT(ACPI_LV_OBJECTS),
86 	ACPI_DEBUG_INIT(ACPI_LV_RESOURCES),
87 	ACPI_DEBUG_INIT(ACPI_LV_USER_REQUESTS),
88 	ACPI_DEBUG_INIT(ACPI_LV_PACKAGE),
89 
90 	ACPI_DEBUG_INIT(ACPI_LV_ALLOCATIONS),
91 	ACPI_DEBUG_INIT(ACPI_LV_FUNCTIONS),
92 	ACPI_DEBUG_INIT(ACPI_LV_OPTIMIZATIONS),
93 
94 	ACPI_DEBUG_INIT(ACPI_LV_MUTEX),
95 	ACPI_DEBUG_INIT(ACPI_LV_THREADS),
96 	ACPI_DEBUG_INIT(ACPI_LV_IO),
97 	ACPI_DEBUG_INIT(ACPI_LV_INTERRUPTS),
98 
99 	ACPI_DEBUG_INIT(ACPI_LV_AML_DISASSEMBLE),
100 	ACPI_DEBUG_INIT(ACPI_LV_VERBOSE_INFO),
101 	ACPI_DEBUG_INIT(ACPI_LV_FULL_TABLES),
102 	ACPI_DEBUG_INIT(ACPI_LV_EVENTS),
103 };
104 
105 static int param_get_debug_layer(char *buffer, const struct kernel_param *kp)
106 {
107 	int result = 0;
108 	int i;
109 
110 	result = sprintf(buffer, "%-25s\tHex        SET\n", "Description");
111 
112 	for (i = 0; i < ARRAY_SIZE(acpi_debug_layers); i++) {
113 		result += sprintf(buffer + result, "%-25s\t0x%08lX [%c]\n",
114 				  acpi_debug_layers[i].name,
115 				  acpi_debug_layers[i].value,
116 				  (acpi_dbg_layer & acpi_debug_layers[i].value)
117 				  ? '*' : ' ');
118 	}
119 	result +=
120 	    sprintf(buffer + result, "%-25s\t0x%08X [%c]\n", "ACPI_ALL_DRIVERS",
121 		    ACPI_ALL_DRIVERS,
122 		    (acpi_dbg_layer & ACPI_ALL_DRIVERS) ==
123 		    ACPI_ALL_DRIVERS ? '*' : (acpi_dbg_layer & ACPI_ALL_DRIVERS)
124 		    == 0 ? ' ' : '-');
125 	result +=
126 	    sprintf(buffer + result,
127 		    "--\ndebug_layer = 0x%08X ( * = enabled)\n",
128 		    acpi_dbg_layer);
129 
130 	return result;
131 }
132 
133 static int param_get_debug_level(char *buffer, const struct kernel_param *kp)
134 {
135 	int result = 0;
136 	int i;
137 
138 	result = sprintf(buffer, "%-25s\tHex        SET\n", "Description");
139 
140 	for (i = 0; i < ARRAY_SIZE(acpi_debug_levels); i++) {
141 		result += sprintf(buffer + result, "%-25s\t0x%08lX [%c]\n",
142 				  acpi_debug_levels[i].name,
143 				  acpi_debug_levels[i].value,
144 				  (acpi_dbg_level & acpi_debug_levels[i].value)
145 				  ? '*' : ' ');
146 	}
147 	result +=
148 	    sprintf(buffer + result, "--\ndebug_level = 0x%08X (* = enabled)\n",
149 		    acpi_dbg_level);
150 
151 	return result;
152 }
153 
154 static const struct kernel_param_ops param_ops_debug_layer = {
155 	.set = param_set_uint,
156 	.get = param_get_debug_layer,
157 };
158 
159 static const struct kernel_param_ops param_ops_debug_level = {
160 	.set = param_set_uint,
161 	.get = param_get_debug_level,
162 };
163 
164 module_param_cb(debug_layer, &param_ops_debug_layer, &acpi_dbg_layer, 0644);
165 module_param_cb(debug_level, &param_ops_debug_level, &acpi_dbg_level, 0644);
166 
167 static char trace_method_name[1024];
168 
169 int param_set_trace_method_name(const char *val, const struct kernel_param *kp)
170 {
171 	u32 saved_flags = 0;
172 	bool is_abs_path = true;
173 
174 	if (*val != '\\')
175 		is_abs_path = false;
176 
177 	if ((is_abs_path && strlen(val) > 1023) ||
178 	    (!is_abs_path && strlen(val) > 1022)) {
179 		pr_err("%s: string parameter too long\n", kp->name);
180 		return -ENOSPC;
181 	}
182 
183 	/*
184 	 * It's not safe to update acpi_gbl_trace_method_name without
185 	 * having the tracer stopped, so we save the original tracer
186 	 * state and disable it.
187 	 */
188 	saved_flags = acpi_gbl_trace_flags;
189 	(void)acpi_debug_trace(NULL,
190 			       acpi_gbl_trace_dbg_level,
191 			       acpi_gbl_trace_dbg_layer,
192 			       0);
193 
194 	/* This is a hack.  We can't kmalloc in early boot. */
195 	if (is_abs_path)
196 		strcpy(trace_method_name, val);
197 	else {
198 		trace_method_name[0] = '\\';
199 		strcpy(trace_method_name+1, val);
200 	}
201 
202 	/* Restore the original tracer state */
203 	(void)acpi_debug_trace(trace_method_name,
204 			       acpi_gbl_trace_dbg_level,
205 			       acpi_gbl_trace_dbg_layer,
206 			       saved_flags);
207 
208 	return 0;
209 }
210 
211 static int param_get_trace_method_name(char *buffer, const struct kernel_param *kp)
212 {
213 	return scnprintf(buffer, PAGE_SIZE, "%s", acpi_gbl_trace_method_name);
214 }
215 
216 static const struct kernel_param_ops param_ops_trace_method = {
217 	.set = param_set_trace_method_name,
218 	.get = param_get_trace_method_name,
219 };
220 
221 static const struct kernel_param_ops param_ops_trace_attrib = {
222 	.set = param_set_uint,
223 	.get = param_get_uint,
224 };
225 
226 module_param_cb(trace_method_name, &param_ops_trace_method, &trace_method_name, 0644);
227 module_param_cb(trace_debug_layer, &param_ops_trace_attrib, &acpi_gbl_trace_dbg_layer, 0644);
228 module_param_cb(trace_debug_level, &param_ops_trace_attrib, &acpi_gbl_trace_dbg_level, 0644);
229 
230 static int param_set_trace_state(const char *val, struct kernel_param *kp)
231 {
232 	acpi_status status;
233 	const char *method = trace_method_name;
234 	u32 flags = 0;
235 
236 /* So "xxx-once" comparison should go prior than "xxx" comparison */
237 #define acpi_compare_param(val, key)	\
238 	strncmp((val), (key), sizeof(key) - 1)
239 
240 	if (!acpi_compare_param(val, "enable")) {
241 		method = NULL;
242 		flags = ACPI_TRACE_ENABLED;
243 	} else if (!acpi_compare_param(val, "disable"))
244 		method = NULL;
245 	else if (!acpi_compare_param(val, "method-once"))
246 		flags = ACPI_TRACE_ENABLED | ACPI_TRACE_ONESHOT;
247 	else if (!acpi_compare_param(val, "method"))
248 		flags = ACPI_TRACE_ENABLED;
249 	else if (!acpi_compare_param(val, "opcode-once"))
250 		flags = ACPI_TRACE_ENABLED | ACPI_TRACE_ONESHOT | ACPI_TRACE_OPCODE;
251 	else if (!acpi_compare_param(val, "opcode"))
252 		flags = ACPI_TRACE_ENABLED | ACPI_TRACE_OPCODE;
253 	else
254 		return -EINVAL;
255 
256 	status = acpi_debug_trace(method,
257 				  acpi_gbl_trace_dbg_level,
258 				  acpi_gbl_trace_dbg_layer,
259 				  flags);
260 	if (ACPI_FAILURE(status))
261 		return -EBUSY;
262 
263 	return 0;
264 }
265 
266 static int param_get_trace_state(char *buffer, struct kernel_param *kp)
267 {
268 	if (!(acpi_gbl_trace_flags & ACPI_TRACE_ENABLED))
269 		return sprintf(buffer, "disable");
270 	else {
271 		if (acpi_gbl_trace_method_name) {
272 			if (acpi_gbl_trace_flags & ACPI_TRACE_ONESHOT)
273 				return sprintf(buffer, "method-once");
274 			else
275 				return sprintf(buffer, "method");
276 		} else
277 			return sprintf(buffer, "enable");
278 	}
279 	return 0;
280 }
281 
282 module_param_call(trace_state, param_set_trace_state, param_get_trace_state,
283 		  NULL, 0644);
284 #endif /* CONFIG_ACPI_DEBUG */
285 
286 
287 /* /sys/modules/acpi/parameters/aml_debug_output */
288 
289 module_param_named(aml_debug_output, acpi_gbl_enable_aml_debug_object,
290 		   byte, 0644);
291 MODULE_PARM_DESC(aml_debug_output,
292 		 "To enable/disable the ACPI Debug Object output.");
293 
294 /* /sys/module/acpi/parameters/acpica_version */
295 static int param_get_acpica_version(char *buffer, struct kernel_param *kp)
296 {
297 	int result;
298 
299 	result = sprintf(buffer, "%x", ACPI_CA_VERSION);
300 
301 	return result;
302 }
303 
304 module_param_call(acpica_version, NULL, param_get_acpica_version, NULL, 0444);
305 
306 /*
307  * ACPI table sysfs I/F:
308  * /sys/firmware/acpi/tables/
309  * /sys/firmware/acpi/tables/dynamic/
310  */
311 
312 static LIST_HEAD(acpi_table_attr_list);
313 static struct kobject *tables_kobj;
314 static struct kobject *dynamic_tables_kobj;
315 static struct kobject *hotplug_kobj;
316 
317 #define ACPI_MAX_TABLE_INSTANCES	999
318 #define ACPI_INST_SIZE			4 /* including trailing 0 */
319 
320 struct acpi_table_attr {
321 	struct bin_attribute attr;
322 	char name[ACPI_NAME_SIZE];
323 	int instance;
324 	char filename[ACPI_NAME_SIZE+ACPI_INST_SIZE];
325 	struct list_head node;
326 };
327 
328 static ssize_t acpi_table_show(struct file *filp, struct kobject *kobj,
329 			       struct bin_attribute *bin_attr, char *buf,
330 			       loff_t offset, size_t count)
331 {
332 	struct acpi_table_attr *table_attr =
333 	    container_of(bin_attr, struct acpi_table_attr, attr);
334 	struct acpi_table_header *table_header = NULL;
335 	acpi_status status;
336 	ssize_t rc;
337 
338 	status = acpi_get_table(table_attr->name, table_attr->instance,
339 				&table_header);
340 	if (ACPI_FAILURE(status))
341 		return -ENODEV;
342 
343 	rc = memory_read_from_buffer(buf, count, &offset, table_header,
344 			table_header->length);
345 	acpi_put_table(table_header);
346 	return rc;
347 }
348 
349 static int acpi_table_attr_init(struct kobject *tables_obj,
350 				struct acpi_table_attr *table_attr,
351 				struct acpi_table_header *table_header)
352 {
353 	struct acpi_table_header *header = NULL;
354 	struct acpi_table_attr *attr = NULL;
355 	char instance_str[ACPI_INST_SIZE];
356 
357 	sysfs_attr_init(&table_attr->attr.attr);
358 	ACPI_MOVE_NAME(table_attr->name, table_header->signature);
359 
360 	list_for_each_entry(attr, &acpi_table_attr_list, node) {
361 		if (ACPI_COMPARE_NAME(table_attr->name, attr->name))
362 			if (table_attr->instance < attr->instance)
363 				table_attr->instance = attr->instance;
364 	}
365 	table_attr->instance++;
366 	if (table_attr->instance > ACPI_MAX_TABLE_INSTANCES) {
367 		pr_warn("%4.4s: too many table instances\n",
368 			table_attr->name);
369 		return -ERANGE;
370 	}
371 
372 	ACPI_MOVE_NAME(table_attr->filename, table_header->signature);
373 	table_attr->filename[ACPI_NAME_SIZE] = '\0';
374 	if (table_attr->instance > 1 || (table_attr->instance == 1 &&
375 					 !acpi_get_table
376 					 (table_header->signature, 2, &header))) {
377 		snprintf(instance_str, sizeof(instance_str), "%u",
378 			 table_attr->instance);
379 		strcat(table_attr->filename, instance_str);
380 	}
381 
382 	table_attr->attr.size = table_header->length;
383 	table_attr->attr.read = acpi_table_show;
384 	table_attr->attr.attr.name = table_attr->filename;
385 	table_attr->attr.attr.mode = 0400;
386 
387 	return sysfs_create_bin_file(tables_obj, &table_attr->attr);
388 }
389 
390 acpi_status acpi_sysfs_table_handler(u32 event, void *table, void *context)
391 {
392 	struct acpi_table_attr *table_attr;
393 
394 	switch (event) {
395 	case ACPI_TABLE_EVENT_INSTALL:
396 		table_attr =
397 		    kzalloc(sizeof(struct acpi_table_attr), GFP_KERNEL);
398 		if (!table_attr)
399 			return AE_NO_MEMORY;
400 
401 		if (acpi_table_attr_init(dynamic_tables_kobj,
402 					 table_attr, table)) {
403 			kfree(table_attr);
404 			return AE_ERROR;
405 		}
406 		list_add_tail(&table_attr->node, &acpi_table_attr_list);
407 		break;
408 	case ACPI_TABLE_EVENT_LOAD:
409 	case ACPI_TABLE_EVENT_UNLOAD:
410 	case ACPI_TABLE_EVENT_UNINSTALL:
411 		/*
412 		 * we do not need to do anything right now
413 		 * because the table is not deleted from the
414 		 * global table list when unloading it.
415 		 */
416 		break;
417 	default:
418 		return AE_BAD_PARAMETER;
419 	}
420 	return AE_OK;
421 }
422 
423 static int acpi_tables_sysfs_init(void)
424 {
425 	struct acpi_table_attr *table_attr;
426 	struct acpi_table_header *table_header = NULL;
427 	int table_index;
428 	acpi_status status;
429 	int ret;
430 
431 	tables_kobj = kobject_create_and_add("tables", acpi_kobj);
432 	if (!tables_kobj)
433 		goto err;
434 
435 	dynamic_tables_kobj = kobject_create_and_add("dynamic", tables_kobj);
436 	if (!dynamic_tables_kobj)
437 		goto err_dynamic_tables;
438 
439 	for (table_index = 0;; table_index++) {
440 		status = acpi_get_table_by_index(table_index, &table_header);
441 
442 		if (status == AE_BAD_PARAMETER)
443 			break;
444 
445 		if (ACPI_FAILURE(status))
446 			continue;
447 
448 		table_attr = kzalloc(sizeof(*table_attr), GFP_KERNEL);
449 		if (!table_attr)
450 			return -ENOMEM;
451 
452 		ret = acpi_table_attr_init(tables_kobj,
453 					   table_attr, table_header);
454 		if (ret) {
455 			kfree(table_attr);
456 			return ret;
457 		}
458 		list_add_tail(&table_attr->node, &acpi_table_attr_list);
459 	}
460 
461 	kobject_uevent(tables_kobj, KOBJ_ADD);
462 	kobject_uevent(dynamic_tables_kobj, KOBJ_ADD);
463 
464 	return 0;
465 err_dynamic_tables:
466 	kobject_put(tables_kobj);
467 err:
468 	return -ENOMEM;
469 }
470 
471 /*
472  * Detailed ACPI IRQ counters:
473  * /sys/firmware/acpi/interrupts/
474  */
475 
476 u32 acpi_irq_handled;
477 u32 acpi_irq_not_handled;
478 
479 #define COUNT_GPE 0
480 #define COUNT_SCI 1		/* acpi_irq_handled */
481 #define COUNT_SCI_NOT 2		/* acpi_irq_not_handled */
482 #define COUNT_ERROR 3		/* other */
483 #define NUM_COUNTERS_EXTRA 4
484 
485 struct event_counter {
486 	u32 count;
487 	u32 flags;
488 };
489 
490 static struct event_counter *all_counters;
491 static u32 num_gpes;
492 static u32 num_counters;
493 static struct attribute **all_attrs;
494 static u32 acpi_gpe_count;
495 
496 static struct attribute_group interrupt_stats_attr_group = {
497 	.name = "interrupts",
498 };
499 
500 static struct kobj_attribute *counter_attrs;
501 
502 static void delete_gpe_attr_array(void)
503 {
504 	struct event_counter *tmp = all_counters;
505 
506 	all_counters = NULL;
507 	kfree(tmp);
508 
509 	if (counter_attrs) {
510 		int i;
511 
512 		for (i = 0; i < num_gpes; i++)
513 			kfree(counter_attrs[i].attr.name);
514 
515 		kfree(counter_attrs);
516 	}
517 	kfree(all_attrs);
518 
519 	return;
520 }
521 
522 static void gpe_count(u32 gpe_number)
523 {
524 	acpi_gpe_count++;
525 
526 	if (!all_counters)
527 		return;
528 
529 	if (gpe_number < num_gpes)
530 		all_counters[gpe_number].count++;
531 	else
532 		all_counters[num_gpes + ACPI_NUM_FIXED_EVENTS +
533 			     COUNT_ERROR].count++;
534 
535 	return;
536 }
537 
538 static void fixed_event_count(u32 event_number)
539 {
540 	if (!all_counters)
541 		return;
542 
543 	if (event_number < ACPI_NUM_FIXED_EVENTS)
544 		all_counters[num_gpes + event_number].count++;
545 	else
546 		all_counters[num_gpes + ACPI_NUM_FIXED_EVENTS +
547 			     COUNT_ERROR].count++;
548 
549 	return;
550 }
551 
552 static void acpi_global_event_handler(u32 event_type, acpi_handle device,
553 	u32 event_number, void *context)
554 {
555 	if (event_type == ACPI_EVENT_TYPE_GPE)
556 		gpe_count(event_number);
557 
558 	if (event_type == ACPI_EVENT_TYPE_FIXED)
559 		fixed_event_count(event_number);
560 }
561 
562 static int get_status(u32 index, acpi_event_status *status,
563 		      acpi_handle *handle)
564 {
565 	int result;
566 
567 	if (index >= num_gpes + ACPI_NUM_FIXED_EVENTS)
568 		return -EINVAL;
569 
570 	if (index < num_gpes) {
571 		result = acpi_get_gpe_device(index, handle);
572 		if (result) {
573 			ACPI_EXCEPTION((AE_INFO, AE_NOT_FOUND,
574 					"Invalid GPE 0x%x", index));
575 			return result;
576 		}
577 		result = acpi_get_gpe_status(*handle, index, status);
578 	} else if (index < (num_gpes + ACPI_NUM_FIXED_EVENTS))
579 		result = acpi_get_event_status(index - num_gpes, status);
580 
581 	return result;
582 }
583 
584 static ssize_t counter_show(struct kobject *kobj,
585 			    struct kobj_attribute *attr, char *buf)
586 {
587 	int index = attr - counter_attrs;
588 	int size;
589 	acpi_handle handle;
590 	acpi_event_status status;
591 	int result = 0;
592 
593 	all_counters[num_gpes + ACPI_NUM_FIXED_EVENTS + COUNT_SCI].count =
594 	    acpi_irq_handled;
595 	all_counters[num_gpes + ACPI_NUM_FIXED_EVENTS + COUNT_SCI_NOT].count =
596 	    acpi_irq_not_handled;
597 	all_counters[num_gpes + ACPI_NUM_FIXED_EVENTS + COUNT_GPE].count =
598 	    acpi_gpe_count;
599 	size = sprintf(buf, "%8u", all_counters[index].count);
600 
601 	/* "gpe_all" or "sci" */
602 	if (index >= num_gpes + ACPI_NUM_FIXED_EVENTS)
603 		goto end;
604 
605 	result = get_status(index, &status, &handle);
606 	if (result)
607 		goto end;
608 
609 	if (status & ACPI_EVENT_FLAG_ENABLE_SET)
610 		size += sprintf(buf + size, "  EN");
611 	else
612 		size += sprintf(buf + size, "    ");
613 	if (status & ACPI_EVENT_FLAG_STATUS_SET)
614 		size += sprintf(buf + size, " STS");
615 	else
616 		size += sprintf(buf + size, "    ");
617 
618 	if (!(status & ACPI_EVENT_FLAG_HAS_HANDLER))
619 		size += sprintf(buf + size, " invalid     ");
620 	else if (status & ACPI_EVENT_FLAG_ENABLED)
621 		size += sprintf(buf + size, " enabled     ");
622 	else if (status & ACPI_EVENT_FLAG_WAKE_ENABLED)
623 		size += sprintf(buf + size, " wake_enabled");
624 	else
625 		size += sprintf(buf + size, " disabled    ");
626 	if (status & ACPI_EVENT_FLAG_MASKED)
627 		size += sprintf(buf + size, " masked  ");
628 	else
629 		size += sprintf(buf + size, " unmasked");
630 
631 end:
632 	size += sprintf(buf + size, "\n");
633 	return result ? result : size;
634 }
635 
636 /*
637  * counter_set() sets the specified counter.
638  * setting the total "sci" file to any value clears all counters.
639  * enable/disable/clear a gpe/fixed event in user space.
640  */
641 static ssize_t counter_set(struct kobject *kobj,
642 			   struct kobj_attribute *attr, const char *buf,
643 			   size_t size)
644 {
645 	int index = attr - counter_attrs;
646 	acpi_event_status status;
647 	acpi_handle handle;
648 	int result = 0;
649 	unsigned long tmp;
650 
651 	if (index == num_gpes + ACPI_NUM_FIXED_EVENTS + COUNT_SCI) {
652 		int i;
653 		for (i = 0; i < num_counters; ++i)
654 			all_counters[i].count = 0;
655 		acpi_gpe_count = 0;
656 		acpi_irq_handled = 0;
657 		acpi_irq_not_handled = 0;
658 		goto end;
659 	}
660 
661 	/* show the event status for both GPEs and Fixed Events */
662 	result = get_status(index, &status, &handle);
663 	if (result)
664 		goto end;
665 
666 	if (!(status & ACPI_EVENT_FLAG_HAS_HANDLER)) {
667 		printk(KERN_WARNING PREFIX
668 		       "Can not change Invalid GPE/Fixed Event status\n");
669 		return -EINVAL;
670 	}
671 
672 	if (index < num_gpes) {
673 		if (!strcmp(buf, "disable\n") &&
674 		    (status & ACPI_EVENT_FLAG_ENABLED))
675 			result = acpi_disable_gpe(handle, index);
676 		else if (!strcmp(buf, "enable\n") &&
677 			 !(status & ACPI_EVENT_FLAG_ENABLED))
678 			result = acpi_enable_gpe(handle, index);
679 		else if (!strcmp(buf, "clear\n") &&
680 			 (status & ACPI_EVENT_FLAG_STATUS_SET))
681 			result = acpi_clear_gpe(handle, index);
682 		else if (!strcmp(buf, "mask\n"))
683 			result = acpi_mask_gpe(handle, index, TRUE);
684 		else if (!strcmp(buf, "unmask\n"))
685 			result = acpi_mask_gpe(handle, index, FALSE);
686 		else if (!kstrtoul(buf, 0, &tmp))
687 			all_counters[index].count = tmp;
688 		else
689 			result = -EINVAL;
690 	} else if (index < num_gpes + ACPI_NUM_FIXED_EVENTS) {
691 		int event = index - num_gpes;
692 		if (!strcmp(buf, "disable\n") &&
693 		    (status & ACPI_EVENT_FLAG_ENABLE_SET))
694 			result = acpi_disable_event(event, ACPI_NOT_ISR);
695 		else if (!strcmp(buf, "enable\n") &&
696 			 !(status & ACPI_EVENT_FLAG_ENABLE_SET))
697 			result = acpi_enable_event(event, ACPI_NOT_ISR);
698 		else if (!strcmp(buf, "clear\n") &&
699 			 (status & ACPI_EVENT_FLAG_STATUS_SET))
700 			result = acpi_clear_event(event);
701 		else if (!kstrtoul(buf, 0, &tmp))
702 			all_counters[index].count = tmp;
703 		else
704 			result = -EINVAL;
705 	} else
706 		all_counters[index].count = strtoul(buf, NULL, 0);
707 
708 	if (ACPI_FAILURE(result))
709 		result = -EINVAL;
710 end:
711 	return result ? result : size;
712 }
713 
714 /*
715  * A Quirk Mechanism for GPE Flooding Prevention:
716  *
717  * Quirks may be needed to prevent GPE flooding on a specific GPE. The
718  * flooding typically cannot be detected and automatically prevented by
719  * ACPI_GPE_DISPATCH_NONE check because there is a _Lxx/_Exx prepared in
720  * the AML tables. This normally indicates a feature gap in Linux, thus
721  * instead of providing endless quirk tables, we provide a boot parameter
722  * for those who want this quirk. For example, if the users want to prevent
723  * the GPE flooding for GPE 00, they need to specify the following boot
724  * parameter:
725  *   acpi_mask_gpe=0x00
726  * The masking status can be modified by the following runtime controlling
727  * interface:
728  *   echo unmask > /sys/firmware/acpi/interrupts/gpe00
729  */
730 
731 /*
732  * Currently, the GPE flooding prevention only supports to mask the GPEs
733  * numbered from 00 to 7f.
734  */
735 #define ACPI_MASKABLE_GPE_MAX	0x80
736 
737 static u64 __initdata acpi_masked_gpes;
738 
739 static int __init acpi_gpe_set_masked_gpes(char *val)
740 {
741 	u8 gpe;
742 
743 	if (kstrtou8(val, 0, &gpe) || gpe > ACPI_MASKABLE_GPE_MAX)
744 		return -EINVAL;
745 	acpi_masked_gpes |= ((u64)1<<gpe);
746 
747 	return 1;
748 }
749 __setup("acpi_mask_gpe=", acpi_gpe_set_masked_gpes);
750 
751 void __init acpi_gpe_apply_masked_gpes(void)
752 {
753 	acpi_handle handle;
754 	acpi_status status;
755 	u8 gpe;
756 
757 	for (gpe = 0;
758 	     gpe < min_t(u8, ACPI_MASKABLE_GPE_MAX, acpi_current_gpe_count);
759 	     gpe++) {
760 		if (acpi_masked_gpes & ((u64)1<<gpe)) {
761 			status = acpi_get_gpe_device(gpe, &handle);
762 			if (ACPI_SUCCESS(status)) {
763 				pr_info("Masking GPE 0x%x.\n", gpe);
764 				(void)acpi_mask_gpe(handle, gpe, TRUE);
765 			}
766 		}
767 	}
768 }
769 
770 void acpi_irq_stats_init(void)
771 {
772 	acpi_status status;
773 	int i;
774 
775 	if (all_counters)
776 		return;
777 
778 	num_gpes = acpi_current_gpe_count;
779 	num_counters = num_gpes + ACPI_NUM_FIXED_EVENTS + NUM_COUNTERS_EXTRA;
780 
781 	all_attrs = kzalloc(sizeof(struct attribute *) * (num_counters + 1),
782 			    GFP_KERNEL);
783 	if (all_attrs == NULL)
784 		return;
785 
786 	all_counters = kzalloc(sizeof(struct event_counter) * (num_counters),
787 			       GFP_KERNEL);
788 	if (all_counters == NULL)
789 		goto fail;
790 
791 	status = acpi_install_global_event_handler(acpi_global_event_handler, NULL);
792 	if (ACPI_FAILURE(status))
793 		goto fail;
794 
795 	counter_attrs = kzalloc(sizeof(struct kobj_attribute) * (num_counters),
796 				GFP_KERNEL);
797 	if (counter_attrs == NULL)
798 		goto fail;
799 
800 	for (i = 0; i < num_counters; ++i) {
801 		char buffer[12];
802 		char *name;
803 
804 		if (i < num_gpes)
805 			sprintf(buffer, "gpe%02X", i);
806 		else if (i == num_gpes + ACPI_EVENT_PMTIMER)
807 			sprintf(buffer, "ff_pmtimer");
808 		else if (i == num_gpes + ACPI_EVENT_GLOBAL)
809 			sprintf(buffer, "ff_gbl_lock");
810 		else if (i == num_gpes + ACPI_EVENT_POWER_BUTTON)
811 			sprintf(buffer, "ff_pwr_btn");
812 		else if (i == num_gpes + ACPI_EVENT_SLEEP_BUTTON)
813 			sprintf(buffer, "ff_slp_btn");
814 		else if (i == num_gpes + ACPI_EVENT_RTC)
815 			sprintf(buffer, "ff_rt_clk");
816 		else if (i == num_gpes + ACPI_NUM_FIXED_EVENTS + COUNT_GPE)
817 			sprintf(buffer, "gpe_all");
818 		else if (i == num_gpes + ACPI_NUM_FIXED_EVENTS + COUNT_SCI)
819 			sprintf(buffer, "sci");
820 		else if (i == num_gpes + ACPI_NUM_FIXED_EVENTS + COUNT_SCI_NOT)
821 			sprintf(buffer, "sci_not");
822 		else if (i == num_gpes + ACPI_NUM_FIXED_EVENTS + COUNT_ERROR)
823 			sprintf(buffer, "error");
824 		else
825 			sprintf(buffer, "bug%02X", i);
826 
827 		name = kstrdup(buffer, GFP_KERNEL);
828 		if (name == NULL)
829 			goto fail;
830 
831 		sysfs_attr_init(&counter_attrs[i].attr);
832 		counter_attrs[i].attr.name = name;
833 		counter_attrs[i].attr.mode = 0644;
834 		counter_attrs[i].show = counter_show;
835 		counter_attrs[i].store = counter_set;
836 
837 		all_attrs[i] = &counter_attrs[i].attr;
838 	}
839 
840 	interrupt_stats_attr_group.attrs = all_attrs;
841 	if (!sysfs_create_group(acpi_kobj, &interrupt_stats_attr_group))
842 		return;
843 
844 fail:
845 	delete_gpe_attr_array();
846 	return;
847 }
848 
849 static void __exit interrupt_stats_exit(void)
850 {
851 	sysfs_remove_group(acpi_kobj, &interrupt_stats_attr_group);
852 
853 	delete_gpe_attr_array();
854 
855 	return;
856 }
857 
858 static ssize_t
859 acpi_show_profile(struct device *dev, struct device_attribute *attr,
860 		  char *buf)
861 {
862 	return sprintf(buf, "%d\n", acpi_gbl_FADT.preferred_profile);
863 }
864 
865 static const struct device_attribute pm_profile_attr =
866 	__ATTR(pm_profile, S_IRUGO, acpi_show_profile, NULL);
867 
868 static ssize_t hotplug_enabled_show(struct kobject *kobj,
869 				    struct kobj_attribute *attr, char *buf)
870 {
871 	struct acpi_hotplug_profile *hotplug = to_acpi_hotplug_profile(kobj);
872 
873 	return sprintf(buf, "%d\n", hotplug->enabled);
874 }
875 
876 static ssize_t hotplug_enabled_store(struct kobject *kobj,
877 				     struct kobj_attribute *attr,
878 				     const char *buf, size_t size)
879 {
880 	struct acpi_hotplug_profile *hotplug = to_acpi_hotplug_profile(kobj);
881 	unsigned int val;
882 
883 	if (kstrtouint(buf, 10, &val) || val > 1)
884 		return -EINVAL;
885 
886 	acpi_scan_hotplug_enabled(hotplug, val);
887 	return size;
888 }
889 
890 static struct kobj_attribute hotplug_enabled_attr =
891 	__ATTR(enabled, S_IRUGO | S_IWUSR, hotplug_enabled_show,
892 		hotplug_enabled_store);
893 
894 static struct attribute *hotplug_profile_attrs[] = {
895 	&hotplug_enabled_attr.attr,
896 	NULL
897 };
898 
899 static struct kobj_type acpi_hotplug_profile_ktype = {
900 	.sysfs_ops = &kobj_sysfs_ops,
901 	.default_attrs = hotplug_profile_attrs,
902 };
903 
904 void acpi_sysfs_add_hotplug_profile(struct acpi_hotplug_profile *hotplug,
905 				    const char *name)
906 {
907 	int error;
908 
909 	if (!hotplug_kobj)
910 		goto err_out;
911 
912 	error = kobject_init_and_add(&hotplug->kobj,
913 		&acpi_hotplug_profile_ktype, hotplug_kobj, "%s", name);
914 	if (error)
915 		goto err_out;
916 
917 	kobject_uevent(&hotplug->kobj, KOBJ_ADD);
918 	return;
919 
920  err_out:
921 	pr_err(PREFIX "Unable to add hotplug profile '%s'\n", name);
922 }
923 
924 static ssize_t force_remove_show(struct kobject *kobj,
925 				 struct kobj_attribute *attr, char *buf)
926 {
927 	return sprintf(buf, "%d\n", 0);
928 }
929 
930 static ssize_t force_remove_store(struct kobject *kobj,
931 				  struct kobj_attribute *attr,
932 				  const char *buf, size_t size)
933 {
934 	bool val;
935 	int ret;
936 
937 	ret = strtobool(buf, &val);
938 	if (ret < 0)
939 		return ret;
940 
941 	if (val) {
942 		pr_err("Enabling force_remove is not supported anymore. Please report to linux-acpi@vger.kernel.org if you depend on this functionality\n");
943 		return -EINVAL;
944 	}
945 	return size;
946 }
947 
948 static const struct kobj_attribute force_remove_attr =
949 	__ATTR(force_remove, S_IRUGO | S_IWUSR, force_remove_show,
950 	       force_remove_store);
951 
952 int __init acpi_sysfs_init(void)
953 {
954 	int result;
955 
956 	result = acpi_tables_sysfs_init();
957 	if (result)
958 		return result;
959 
960 	hotplug_kobj = kobject_create_and_add("hotplug", acpi_kobj);
961 	if (!hotplug_kobj)
962 		return -ENOMEM;
963 
964 	result = sysfs_create_file(hotplug_kobj, &force_remove_attr.attr);
965 	if (result)
966 		return result;
967 
968 	result = sysfs_create_file(acpi_kobj, &pm_profile_attr.attr);
969 	return result;
970 }
971