1 // SPDX-License-Identifier: GPL-2.0 2 /* 3 * sysfs.c - ACPI sysfs interface to userspace. 4 */ 5 6 #define pr_fmt(fmt) "ACPI: " fmt 7 8 #include <linux/init.h> 9 #include <linux/kernel.h> 10 #include <linux/moduleparam.h> 11 #include <linux/acpi.h> 12 13 #include "internal.h" 14 15 #define _COMPONENT ACPI_SYSTEM_COMPONENT 16 ACPI_MODULE_NAME("sysfs"); 17 18 #ifdef CONFIG_ACPI_DEBUG 19 /* 20 * ACPI debug sysfs I/F, including: 21 * /sys/modules/acpi/parameters/debug_layer 22 * /sys/modules/acpi/parameters/debug_level 23 * /sys/modules/acpi/parameters/trace_method_name 24 * /sys/modules/acpi/parameters/trace_state 25 * /sys/modules/acpi/parameters/trace_debug_layer 26 * /sys/modules/acpi/parameters/trace_debug_level 27 */ 28 29 struct acpi_dlayer { 30 const char *name; 31 unsigned long value; 32 }; 33 struct acpi_dlevel { 34 const char *name; 35 unsigned long value; 36 }; 37 #define ACPI_DEBUG_INIT(v) { .name = #v, .value = v } 38 39 static const struct acpi_dlayer acpi_debug_layers[] = { 40 ACPI_DEBUG_INIT(ACPI_UTILITIES), 41 ACPI_DEBUG_INIT(ACPI_HARDWARE), 42 ACPI_DEBUG_INIT(ACPI_EVENTS), 43 ACPI_DEBUG_INIT(ACPI_TABLES), 44 ACPI_DEBUG_INIT(ACPI_NAMESPACE), 45 ACPI_DEBUG_INIT(ACPI_PARSER), 46 ACPI_DEBUG_INIT(ACPI_DISPATCHER), 47 ACPI_DEBUG_INIT(ACPI_EXECUTER), 48 ACPI_DEBUG_INIT(ACPI_RESOURCES), 49 ACPI_DEBUG_INIT(ACPI_CA_DEBUGGER), 50 ACPI_DEBUG_INIT(ACPI_OS_SERVICES), 51 ACPI_DEBUG_INIT(ACPI_CA_DISASSEMBLER), 52 ACPI_DEBUG_INIT(ACPI_COMPILER), 53 ACPI_DEBUG_INIT(ACPI_TOOLS), 54 55 ACPI_DEBUG_INIT(ACPI_BUS_COMPONENT), 56 ACPI_DEBUG_INIT(ACPI_AC_COMPONENT), 57 ACPI_DEBUG_INIT(ACPI_BATTERY_COMPONENT), 58 ACPI_DEBUG_INIT(ACPI_BUTTON_COMPONENT), 59 ACPI_DEBUG_INIT(ACPI_SBS_COMPONENT), 60 ACPI_DEBUG_INIT(ACPI_FAN_COMPONENT), 61 ACPI_DEBUG_INIT(ACPI_PCI_COMPONENT), 62 ACPI_DEBUG_INIT(ACPI_POWER_COMPONENT), 63 ACPI_DEBUG_INIT(ACPI_CONTAINER_COMPONENT), 64 ACPI_DEBUG_INIT(ACPI_SYSTEM_COMPONENT), 65 ACPI_DEBUG_INIT(ACPI_THERMAL_COMPONENT), 66 ACPI_DEBUG_INIT(ACPI_MEMORY_DEVICE_COMPONENT), 67 ACPI_DEBUG_INIT(ACPI_VIDEO_COMPONENT), 68 ACPI_DEBUG_INIT(ACPI_PROCESSOR_COMPONENT), 69 }; 70 71 static const struct acpi_dlevel acpi_debug_levels[] = { 72 ACPI_DEBUG_INIT(ACPI_LV_INIT), 73 ACPI_DEBUG_INIT(ACPI_LV_DEBUG_OBJECT), 74 ACPI_DEBUG_INIT(ACPI_LV_INFO), 75 ACPI_DEBUG_INIT(ACPI_LV_REPAIR), 76 ACPI_DEBUG_INIT(ACPI_LV_TRACE_POINT), 77 78 ACPI_DEBUG_INIT(ACPI_LV_INIT_NAMES), 79 ACPI_DEBUG_INIT(ACPI_LV_PARSE), 80 ACPI_DEBUG_INIT(ACPI_LV_LOAD), 81 ACPI_DEBUG_INIT(ACPI_LV_DISPATCH), 82 ACPI_DEBUG_INIT(ACPI_LV_EXEC), 83 ACPI_DEBUG_INIT(ACPI_LV_NAMES), 84 ACPI_DEBUG_INIT(ACPI_LV_OPREGION), 85 ACPI_DEBUG_INIT(ACPI_LV_BFIELD), 86 ACPI_DEBUG_INIT(ACPI_LV_TABLES), 87 ACPI_DEBUG_INIT(ACPI_LV_VALUES), 88 ACPI_DEBUG_INIT(ACPI_LV_OBJECTS), 89 ACPI_DEBUG_INIT(ACPI_LV_RESOURCES), 90 ACPI_DEBUG_INIT(ACPI_LV_USER_REQUESTS), 91 ACPI_DEBUG_INIT(ACPI_LV_PACKAGE), 92 93 ACPI_DEBUG_INIT(ACPI_LV_ALLOCATIONS), 94 ACPI_DEBUG_INIT(ACPI_LV_FUNCTIONS), 95 ACPI_DEBUG_INIT(ACPI_LV_OPTIMIZATIONS), 96 97 ACPI_DEBUG_INIT(ACPI_LV_MUTEX), 98 ACPI_DEBUG_INIT(ACPI_LV_THREADS), 99 ACPI_DEBUG_INIT(ACPI_LV_IO), 100 ACPI_DEBUG_INIT(ACPI_LV_INTERRUPTS), 101 102 ACPI_DEBUG_INIT(ACPI_LV_AML_DISASSEMBLE), 103 ACPI_DEBUG_INIT(ACPI_LV_VERBOSE_INFO), 104 ACPI_DEBUG_INIT(ACPI_LV_FULL_TABLES), 105 ACPI_DEBUG_INIT(ACPI_LV_EVENTS), 106 }; 107 108 static int param_get_debug_layer(char *buffer, const struct kernel_param *kp) 109 { 110 int result = 0; 111 int i; 112 113 result = sprintf(buffer, "%-25s\tHex SET\n", "Description"); 114 115 for (i = 0; i < ARRAY_SIZE(acpi_debug_layers); i++) { 116 result += sprintf(buffer + result, "%-25s\t0x%08lX [%c]\n", 117 acpi_debug_layers[i].name, 118 acpi_debug_layers[i].value, 119 (acpi_dbg_layer & acpi_debug_layers[i].value) 120 ? '*' : ' '); 121 } 122 result += 123 sprintf(buffer + result, "%-25s\t0x%08X [%c]\n", "ACPI_ALL_DRIVERS", 124 ACPI_ALL_DRIVERS, 125 (acpi_dbg_layer & ACPI_ALL_DRIVERS) == 126 ACPI_ALL_DRIVERS ? '*' : (acpi_dbg_layer & ACPI_ALL_DRIVERS) 127 == 0 ? ' ' : '-'); 128 result += 129 sprintf(buffer + result, 130 "--\ndebug_layer = 0x%08X ( * = enabled)\n", 131 acpi_dbg_layer); 132 133 return result; 134 } 135 136 static int param_get_debug_level(char *buffer, const struct kernel_param *kp) 137 { 138 int result = 0; 139 int i; 140 141 result = sprintf(buffer, "%-25s\tHex SET\n", "Description"); 142 143 for (i = 0; i < ARRAY_SIZE(acpi_debug_levels); i++) { 144 result += sprintf(buffer + result, "%-25s\t0x%08lX [%c]\n", 145 acpi_debug_levels[i].name, 146 acpi_debug_levels[i].value, 147 (acpi_dbg_level & acpi_debug_levels[i].value) 148 ? '*' : ' '); 149 } 150 result += 151 sprintf(buffer + result, "--\ndebug_level = 0x%08X (* = enabled)\n", 152 acpi_dbg_level); 153 154 return result; 155 } 156 157 static const struct kernel_param_ops param_ops_debug_layer = { 158 .set = param_set_uint, 159 .get = param_get_debug_layer, 160 }; 161 162 static const struct kernel_param_ops param_ops_debug_level = { 163 .set = param_set_uint, 164 .get = param_get_debug_level, 165 }; 166 167 module_param_cb(debug_layer, ¶m_ops_debug_layer, &acpi_dbg_layer, 0644); 168 module_param_cb(debug_level, ¶m_ops_debug_level, &acpi_dbg_level, 0644); 169 170 static char trace_method_name[1024]; 171 172 static int param_set_trace_method_name(const char *val, 173 const struct kernel_param *kp) 174 { 175 u32 saved_flags = 0; 176 bool is_abs_path = true; 177 178 if (*val != '\\') 179 is_abs_path = false; 180 181 if ((is_abs_path && strlen(val) > 1023) || 182 (!is_abs_path && strlen(val) > 1022)) { 183 pr_err("%s: string parameter too long\n", kp->name); 184 return -ENOSPC; 185 } 186 187 /* 188 * It's not safe to update acpi_gbl_trace_method_name without 189 * having the tracer stopped, so we save the original tracer 190 * state and disable it. 191 */ 192 saved_flags = acpi_gbl_trace_flags; 193 (void)acpi_debug_trace(NULL, 194 acpi_gbl_trace_dbg_level, 195 acpi_gbl_trace_dbg_layer, 196 0); 197 198 /* This is a hack. We can't kmalloc in early boot. */ 199 if (is_abs_path) 200 strcpy(trace_method_name, val); 201 else { 202 trace_method_name[0] = '\\'; 203 strcpy(trace_method_name+1, val); 204 } 205 206 /* Restore the original tracer state */ 207 (void)acpi_debug_trace(trace_method_name, 208 acpi_gbl_trace_dbg_level, 209 acpi_gbl_trace_dbg_layer, 210 saved_flags); 211 212 return 0; 213 } 214 215 static int param_get_trace_method_name(char *buffer, const struct kernel_param *kp) 216 { 217 return scnprintf(buffer, PAGE_SIZE, "%s", acpi_gbl_trace_method_name); 218 } 219 220 static const struct kernel_param_ops param_ops_trace_method = { 221 .set = param_set_trace_method_name, 222 .get = param_get_trace_method_name, 223 }; 224 225 static const struct kernel_param_ops param_ops_trace_attrib = { 226 .set = param_set_uint, 227 .get = param_get_uint, 228 }; 229 230 module_param_cb(trace_method_name, ¶m_ops_trace_method, &trace_method_name, 0644); 231 module_param_cb(trace_debug_layer, ¶m_ops_trace_attrib, &acpi_gbl_trace_dbg_layer, 0644); 232 module_param_cb(trace_debug_level, ¶m_ops_trace_attrib, &acpi_gbl_trace_dbg_level, 0644); 233 234 static int param_set_trace_state(const char *val, 235 const struct kernel_param *kp) 236 { 237 acpi_status status; 238 const char *method = trace_method_name; 239 u32 flags = 0; 240 241 /* So "xxx-once" comparison should go prior than "xxx" comparison */ 242 #define acpi_compare_param(val, key) \ 243 strncmp((val), (key), sizeof(key) - 1) 244 245 if (!acpi_compare_param(val, "enable")) { 246 method = NULL; 247 flags = ACPI_TRACE_ENABLED; 248 } else if (!acpi_compare_param(val, "disable")) 249 method = NULL; 250 else if (!acpi_compare_param(val, "method-once")) 251 flags = ACPI_TRACE_ENABLED | ACPI_TRACE_ONESHOT; 252 else if (!acpi_compare_param(val, "method")) 253 flags = ACPI_TRACE_ENABLED; 254 else if (!acpi_compare_param(val, "opcode-once")) 255 flags = ACPI_TRACE_ENABLED | ACPI_TRACE_ONESHOT | ACPI_TRACE_OPCODE; 256 else if (!acpi_compare_param(val, "opcode")) 257 flags = ACPI_TRACE_ENABLED | ACPI_TRACE_OPCODE; 258 else 259 return -EINVAL; 260 261 status = acpi_debug_trace(method, 262 acpi_gbl_trace_dbg_level, 263 acpi_gbl_trace_dbg_layer, 264 flags); 265 if (ACPI_FAILURE(status)) 266 return -EBUSY; 267 268 return 0; 269 } 270 271 static int param_get_trace_state(char *buffer, const struct kernel_param *kp) 272 { 273 if (!(acpi_gbl_trace_flags & ACPI_TRACE_ENABLED)) 274 return sprintf(buffer, "disable"); 275 else { 276 if (acpi_gbl_trace_method_name) { 277 if (acpi_gbl_trace_flags & ACPI_TRACE_ONESHOT) 278 return sprintf(buffer, "method-once"); 279 else 280 return sprintf(buffer, "method"); 281 } else 282 return sprintf(buffer, "enable"); 283 } 284 return 0; 285 } 286 287 module_param_call(trace_state, param_set_trace_state, param_get_trace_state, 288 NULL, 0644); 289 #endif /* CONFIG_ACPI_DEBUG */ 290 291 292 /* /sys/modules/acpi/parameters/aml_debug_output */ 293 294 module_param_named(aml_debug_output, acpi_gbl_enable_aml_debug_object, 295 byte, 0644); 296 MODULE_PARM_DESC(aml_debug_output, 297 "To enable/disable the ACPI Debug Object output."); 298 299 /* /sys/module/acpi/parameters/acpica_version */ 300 static int param_get_acpica_version(char *buffer, 301 const struct kernel_param *kp) 302 { 303 int result; 304 305 result = sprintf(buffer, "%x", ACPI_CA_VERSION); 306 307 return result; 308 } 309 310 module_param_call(acpica_version, NULL, param_get_acpica_version, NULL, 0444); 311 312 /* 313 * ACPI table sysfs I/F: 314 * /sys/firmware/acpi/tables/ 315 * /sys/firmware/acpi/tables/data/ 316 * /sys/firmware/acpi/tables/dynamic/ 317 */ 318 319 static LIST_HEAD(acpi_table_attr_list); 320 static struct kobject *tables_kobj; 321 static struct kobject *tables_data_kobj; 322 static struct kobject *dynamic_tables_kobj; 323 static struct kobject *hotplug_kobj; 324 325 #define ACPI_MAX_TABLE_INSTANCES 999 326 #define ACPI_INST_SIZE 4 /* including trailing 0 */ 327 328 struct acpi_table_attr { 329 struct bin_attribute attr; 330 char name[ACPI_NAME_SIZE]; 331 int instance; 332 char filename[ACPI_NAME_SIZE+ACPI_INST_SIZE]; 333 struct list_head node; 334 }; 335 336 struct acpi_data_attr { 337 struct bin_attribute attr; 338 u64 addr; 339 }; 340 341 static ssize_t acpi_table_show(struct file *filp, struct kobject *kobj, 342 struct bin_attribute *bin_attr, char *buf, 343 loff_t offset, size_t count) 344 { 345 struct acpi_table_attr *table_attr = 346 container_of(bin_attr, struct acpi_table_attr, attr); 347 struct acpi_table_header *table_header = NULL; 348 acpi_status status; 349 ssize_t rc; 350 351 status = acpi_get_table(table_attr->name, table_attr->instance, 352 &table_header); 353 if (ACPI_FAILURE(status)) 354 return -ENODEV; 355 356 rc = memory_read_from_buffer(buf, count, &offset, table_header, 357 table_header->length); 358 acpi_put_table(table_header); 359 return rc; 360 } 361 362 static int acpi_table_attr_init(struct kobject *tables_obj, 363 struct acpi_table_attr *table_attr, 364 struct acpi_table_header *table_header) 365 { 366 struct acpi_table_header *header = NULL; 367 struct acpi_table_attr *attr = NULL; 368 char instance_str[ACPI_INST_SIZE]; 369 370 sysfs_attr_init(&table_attr->attr.attr); 371 ACPI_MOVE_NAME(table_attr->name, table_header->signature); 372 373 list_for_each_entry(attr, &acpi_table_attr_list, node) { 374 if (ACPI_COMPARE_NAME(table_attr->name, attr->name)) 375 if (table_attr->instance < attr->instance) 376 table_attr->instance = attr->instance; 377 } 378 table_attr->instance++; 379 if (table_attr->instance > ACPI_MAX_TABLE_INSTANCES) { 380 pr_warn("%4.4s: too many table instances\n", 381 table_attr->name); 382 return -ERANGE; 383 } 384 385 ACPI_MOVE_NAME(table_attr->filename, table_header->signature); 386 table_attr->filename[ACPI_NAME_SIZE] = '\0'; 387 if (table_attr->instance > 1 || (table_attr->instance == 1 && 388 !acpi_get_table 389 (table_header->signature, 2, &header))) { 390 snprintf(instance_str, sizeof(instance_str), "%u", 391 table_attr->instance); 392 strcat(table_attr->filename, instance_str); 393 } 394 395 table_attr->attr.size = table_header->length; 396 table_attr->attr.read = acpi_table_show; 397 table_attr->attr.attr.name = table_attr->filename; 398 table_attr->attr.attr.mode = 0400; 399 400 return sysfs_create_bin_file(tables_obj, &table_attr->attr); 401 } 402 403 acpi_status acpi_sysfs_table_handler(u32 event, void *table, void *context) 404 { 405 struct acpi_table_attr *table_attr; 406 407 switch (event) { 408 case ACPI_TABLE_EVENT_INSTALL: 409 table_attr = 410 kzalloc(sizeof(struct acpi_table_attr), GFP_KERNEL); 411 if (!table_attr) 412 return AE_NO_MEMORY; 413 414 if (acpi_table_attr_init(dynamic_tables_kobj, 415 table_attr, table)) { 416 kfree(table_attr); 417 return AE_ERROR; 418 } 419 list_add_tail(&table_attr->node, &acpi_table_attr_list); 420 break; 421 case ACPI_TABLE_EVENT_LOAD: 422 case ACPI_TABLE_EVENT_UNLOAD: 423 case ACPI_TABLE_EVENT_UNINSTALL: 424 /* 425 * we do not need to do anything right now 426 * because the table is not deleted from the 427 * global table list when unloading it. 428 */ 429 break; 430 default: 431 return AE_BAD_PARAMETER; 432 } 433 return AE_OK; 434 } 435 436 static ssize_t acpi_data_show(struct file *filp, struct kobject *kobj, 437 struct bin_attribute *bin_attr, char *buf, 438 loff_t offset, size_t count) 439 { 440 struct acpi_data_attr *data_attr; 441 void __iomem *base; 442 ssize_t rc; 443 444 data_attr = container_of(bin_attr, struct acpi_data_attr, attr); 445 446 base = acpi_os_map_memory(data_attr->addr, data_attr->attr.size); 447 if (!base) 448 return -ENOMEM; 449 rc = memory_read_from_buffer(buf, count, &offset, base, 450 data_attr->attr.size); 451 acpi_os_unmap_memory(base, data_attr->attr.size); 452 453 return rc; 454 } 455 456 static int acpi_bert_data_init(void *th, struct acpi_data_attr *data_attr) 457 { 458 struct acpi_table_bert *bert = th; 459 460 if (bert->header.length < sizeof(struct acpi_table_bert) || 461 bert->region_length < sizeof(struct acpi_hest_generic_status)) { 462 kfree(data_attr); 463 return -EINVAL; 464 } 465 data_attr->addr = bert->address; 466 data_attr->attr.size = bert->region_length; 467 data_attr->attr.attr.name = "BERT"; 468 469 return sysfs_create_bin_file(tables_data_kobj, &data_attr->attr); 470 } 471 472 static struct acpi_data_obj { 473 char *name; 474 int (*fn)(void *, struct acpi_data_attr *); 475 } acpi_data_objs[] = { 476 { ACPI_SIG_BERT, acpi_bert_data_init }, 477 }; 478 479 #define NUM_ACPI_DATA_OBJS ARRAY_SIZE(acpi_data_objs) 480 481 static int acpi_table_data_init(struct acpi_table_header *th) 482 { 483 struct acpi_data_attr *data_attr; 484 int i; 485 486 for (i = 0; i < NUM_ACPI_DATA_OBJS; i++) { 487 if (ACPI_COMPARE_NAME(th->signature, acpi_data_objs[i].name)) { 488 data_attr = kzalloc(sizeof(*data_attr), GFP_KERNEL); 489 if (!data_attr) 490 return -ENOMEM; 491 sysfs_attr_init(&data_attr->attr.attr); 492 data_attr->attr.read = acpi_data_show; 493 data_attr->attr.attr.mode = 0400; 494 return acpi_data_objs[i].fn(th, data_attr); 495 } 496 } 497 return 0; 498 } 499 500 static int acpi_tables_sysfs_init(void) 501 { 502 struct acpi_table_attr *table_attr; 503 struct acpi_table_header *table_header = NULL; 504 int table_index; 505 acpi_status status; 506 int ret; 507 508 tables_kobj = kobject_create_and_add("tables", acpi_kobj); 509 if (!tables_kobj) 510 goto err; 511 512 tables_data_kobj = kobject_create_and_add("data", tables_kobj); 513 if (!tables_data_kobj) 514 goto err_tables_data; 515 516 dynamic_tables_kobj = kobject_create_and_add("dynamic", tables_kobj); 517 if (!dynamic_tables_kobj) 518 goto err_dynamic_tables; 519 520 for (table_index = 0;; table_index++) { 521 status = acpi_get_table_by_index(table_index, &table_header); 522 523 if (status == AE_BAD_PARAMETER) 524 break; 525 526 if (ACPI_FAILURE(status)) 527 continue; 528 529 table_attr = kzalloc(sizeof(*table_attr), GFP_KERNEL); 530 if (!table_attr) 531 return -ENOMEM; 532 533 ret = acpi_table_attr_init(tables_kobj, 534 table_attr, table_header); 535 if (ret) { 536 kfree(table_attr); 537 return ret; 538 } 539 list_add_tail(&table_attr->node, &acpi_table_attr_list); 540 acpi_table_data_init(table_header); 541 } 542 543 kobject_uevent(tables_kobj, KOBJ_ADD); 544 kobject_uevent(tables_data_kobj, KOBJ_ADD); 545 kobject_uevent(dynamic_tables_kobj, KOBJ_ADD); 546 547 return 0; 548 err_dynamic_tables: 549 kobject_put(tables_data_kobj); 550 err_tables_data: 551 kobject_put(tables_kobj); 552 err: 553 return -ENOMEM; 554 } 555 556 /* 557 * Detailed ACPI IRQ counters: 558 * /sys/firmware/acpi/interrupts/ 559 */ 560 561 u32 acpi_irq_handled; 562 u32 acpi_irq_not_handled; 563 564 #define COUNT_GPE 0 565 #define COUNT_SCI 1 /* acpi_irq_handled */ 566 #define COUNT_SCI_NOT 2 /* acpi_irq_not_handled */ 567 #define COUNT_ERROR 3 /* other */ 568 #define NUM_COUNTERS_EXTRA 4 569 570 struct event_counter { 571 u32 count; 572 u32 flags; 573 }; 574 575 static struct event_counter *all_counters; 576 static u32 num_gpes; 577 static u32 num_counters; 578 static struct attribute **all_attrs; 579 static u32 acpi_gpe_count; 580 581 static struct attribute_group interrupt_stats_attr_group = { 582 .name = "interrupts", 583 }; 584 585 static struct kobj_attribute *counter_attrs; 586 587 static void delete_gpe_attr_array(void) 588 { 589 struct event_counter *tmp = all_counters; 590 591 all_counters = NULL; 592 kfree(tmp); 593 594 if (counter_attrs) { 595 int i; 596 597 for (i = 0; i < num_gpes; i++) 598 kfree(counter_attrs[i].attr.name); 599 600 kfree(counter_attrs); 601 } 602 kfree(all_attrs); 603 604 return; 605 } 606 607 static void gpe_count(u32 gpe_number) 608 { 609 acpi_gpe_count++; 610 611 if (!all_counters) 612 return; 613 614 if (gpe_number < num_gpes) 615 all_counters[gpe_number].count++; 616 else 617 all_counters[num_gpes + ACPI_NUM_FIXED_EVENTS + 618 COUNT_ERROR].count++; 619 620 return; 621 } 622 623 static void fixed_event_count(u32 event_number) 624 { 625 if (!all_counters) 626 return; 627 628 if (event_number < ACPI_NUM_FIXED_EVENTS) 629 all_counters[num_gpes + event_number].count++; 630 else 631 all_counters[num_gpes + ACPI_NUM_FIXED_EVENTS + 632 COUNT_ERROR].count++; 633 634 return; 635 } 636 637 static void acpi_global_event_handler(u32 event_type, acpi_handle device, 638 u32 event_number, void *context) 639 { 640 if (event_type == ACPI_EVENT_TYPE_GPE) { 641 gpe_count(event_number); 642 pr_debug("GPE event 0x%02x\n", event_number); 643 } else if (event_type == ACPI_EVENT_TYPE_FIXED) { 644 fixed_event_count(event_number); 645 pr_debug("Fixed event 0x%02x\n", event_number); 646 } else { 647 pr_debug("Other event 0x%02x\n", event_number); 648 } 649 } 650 651 static int get_status(u32 index, acpi_event_status *status, 652 acpi_handle *handle) 653 { 654 int result; 655 656 if (index >= num_gpes + ACPI_NUM_FIXED_EVENTS) 657 return -EINVAL; 658 659 if (index < num_gpes) { 660 result = acpi_get_gpe_device(index, handle); 661 if (result) { 662 ACPI_EXCEPTION((AE_INFO, AE_NOT_FOUND, 663 "Invalid GPE 0x%x", index)); 664 return result; 665 } 666 result = acpi_get_gpe_status(*handle, index, status); 667 } else if (index < (num_gpes + ACPI_NUM_FIXED_EVENTS)) 668 result = acpi_get_event_status(index - num_gpes, status); 669 670 return result; 671 } 672 673 static ssize_t counter_show(struct kobject *kobj, 674 struct kobj_attribute *attr, char *buf) 675 { 676 int index = attr - counter_attrs; 677 int size; 678 acpi_handle handle; 679 acpi_event_status status; 680 int result = 0; 681 682 all_counters[num_gpes + ACPI_NUM_FIXED_EVENTS + COUNT_SCI].count = 683 acpi_irq_handled; 684 all_counters[num_gpes + ACPI_NUM_FIXED_EVENTS + COUNT_SCI_NOT].count = 685 acpi_irq_not_handled; 686 all_counters[num_gpes + ACPI_NUM_FIXED_EVENTS + COUNT_GPE].count = 687 acpi_gpe_count; 688 size = sprintf(buf, "%8u", all_counters[index].count); 689 690 /* "gpe_all" or "sci" */ 691 if (index >= num_gpes + ACPI_NUM_FIXED_EVENTS) 692 goto end; 693 694 result = get_status(index, &status, &handle); 695 if (result) 696 goto end; 697 698 if (status & ACPI_EVENT_FLAG_ENABLE_SET) 699 size += sprintf(buf + size, " EN"); 700 else 701 size += sprintf(buf + size, " "); 702 if (status & ACPI_EVENT_FLAG_STATUS_SET) 703 size += sprintf(buf + size, " STS"); 704 else 705 size += sprintf(buf + size, " "); 706 707 if (!(status & ACPI_EVENT_FLAG_HAS_HANDLER)) 708 size += sprintf(buf + size, " invalid "); 709 else if (status & ACPI_EVENT_FLAG_ENABLED) 710 size += sprintf(buf + size, " enabled "); 711 else if (status & ACPI_EVENT_FLAG_WAKE_ENABLED) 712 size += sprintf(buf + size, " wake_enabled"); 713 else 714 size += sprintf(buf + size, " disabled "); 715 if (status & ACPI_EVENT_FLAG_MASKED) 716 size += sprintf(buf + size, " masked "); 717 else 718 size += sprintf(buf + size, " unmasked"); 719 720 end: 721 size += sprintf(buf + size, "\n"); 722 return result ? result : size; 723 } 724 725 /* 726 * counter_set() sets the specified counter. 727 * setting the total "sci" file to any value clears all counters. 728 * enable/disable/clear a gpe/fixed event in user space. 729 */ 730 static ssize_t counter_set(struct kobject *kobj, 731 struct kobj_attribute *attr, const char *buf, 732 size_t size) 733 { 734 int index = attr - counter_attrs; 735 acpi_event_status status; 736 acpi_handle handle; 737 int result = 0; 738 unsigned long tmp; 739 740 if (index == num_gpes + ACPI_NUM_FIXED_EVENTS + COUNT_SCI) { 741 int i; 742 for (i = 0; i < num_counters; ++i) 743 all_counters[i].count = 0; 744 acpi_gpe_count = 0; 745 acpi_irq_handled = 0; 746 acpi_irq_not_handled = 0; 747 goto end; 748 } 749 750 /* show the event status for both GPEs and Fixed Events */ 751 result = get_status(index, &status, &handle); 752 if (result) 753 goto end; 754 755 if (!(status & ACPI_EVENT_FLAG_HAS_HANDLER)) { 756 printk(KERN_WARNING PREFIX 757 "Can not change Invalid GPE/Fixed Event status\n"); 758 return -EINVAL; 759 } 760 761 if (index < num_gpes) { 762 if (!strcmp(buf, "disable\n") && 763 (status & ACPI_EVENT_FLAG_ENABLED)) 764 result = acpi_disable_gpe(handle, index); 765 else if (!strcmp(buf, "enable\n") && 766 !(status & ACPI_EVENT_FLAG_ENABLED)) 767 result = acpi_enable_gpe(handle, index); 768 else if (!strcmp(buf, "clear\n") && 769 (status & ACPI_EVENT_FLAG_STATUS_SET)) 770 result = acpi_clear_gpe(handle, index); 771 else if (!strcmp(buf, "mask\n")) 772 result = acpi_mask_gpe(handle, index, TRUE); 773 else if (!strcmp(buf, "unmask\n")) 774 result = acpi_mask_gpe(handle, index, FALSE); 775 else if (!kstrtoul(buf, 0, &tmp)) 776 all_counters[index].count = tmp; 777 else 778 result = -EINVAL; 779 } else if (index < num_gpes + ACPI_NUM_FIXED_EVENTS) { 780 int event = index - num_gpes; 781 if (!strcmp(buf, "disable\n") && 782 (status & ACPI_EVENT_FLAG_ENABLE_SET)) 783 result = acpi_disable_event(event, ACPI_NOT_ISR); 784 else if (!strcmp(buf, "enable\n") && 785 !(status & ACPI_EVENT_FLAG_ENABLE_SET)) 786 result = acpi_enable_event(event, ACPI_NOT_ISR); 787 else if (!strcmp(buf, "clear\n") && 788 (status & ACPI_EVENT_FLAG_STATUS_SET)) 789 result = acpi_clear_event(event); 790 else if (!kstrtoul(buf, 0, &tmp)) 791 all_counters[index].count = tmp; 792 else 793 result = -EINVAL; 794 } else 795 all_counters[index].count = strtoul(buf, NULL, 0); 796 797 if (ACPI_FAILURE(result)) 798 result = -EINVAL; 799 end: 800 return result ? result : size; 801 } 802 803 /* 804 * A Quirk Mechanism for GPE Flooding Prevention: 805 * 806 * Quirks may be needed to prevent GPE flooding on a specific GPE. The 807 * flooding typically cannot be detected and automatically prevented by 808 * ACPI_GPE_DISPATCH_NONE check because there is a _Lxx/_Exx prepared in 809 * the AML tables. This normally indicates a feature gap in Linux, thus 810 * instead of providing endless quirk tables, we provide a boot parameter 811 * for those who want this quirk. For example, if the users want to prevent 812 * the GPE flooding for GPE 00, they need to specify the following boot 813 * parameter: 814 * acpi_mask_gpe=0x00 815 * The masking status can be modified by the following runtime controlling 816 * interface: 817 * echo unmask > /sys/firmware/acpi/interrupts/gpe00 818 */ 819 820 /* 821 * Currently, the GPE flooding prevention only supports to mask the GPEs 822 * numbered from 00 to 7f. 823 */ 824 #define ACPI_MASKABLE_GPE_MAX 0x80 825 826 static u64 __initdata acpi_masked_gpes; 827 828 static int __init acpi_gpe_set_masked_gpes(char *val) 829 { 830 u8 gpe; 831 832 if (kstrtou8(val, 0, &gpe) || gpe > ACPI_MASKABLE_GPE_MAX) 833 return -EINVAL; 834 acpi_masked_gpes |= ((u64)1<<gpe); 835 836 return 1; 837 } 838 __setup("acpi_mask_gpe=", acpi_gpe_set_masked_gpes); 839 840 void __init acpi_gpe_apply_masked_gpes(void) 841 { 842 acpi_handle handle; 843 acpi_status status; 844 u8 gpe; 845 846 for (gpe = 0; 847 gpe < min_t(u8, ACPI_MASKABLE_GPE_MAX, acpi_current_gpe_count); 848 gpe++) { 849 if (acpi_masked_gpes & ((u64)1<<gpe)) { 850 status = acpi_get_gpe_device(gpe, &handle); 851 if (ACPI_SUCCESS(status)) { 852 pr_info("Masking GPE 0x%x.\n", gpe); 853 (void)acpi_mask_gpe(handle, gpe, TRUE); 854 } 855 } 856 } 857 } 858 859 void acpi_irq_stats_init(void) 860 { 861 acpi_status status; 862 int i; 863 864 if (all_counters) 865 return; 866 867 num_gpes = acpi_current_gpe_count; 868 num_counters = num_gpes + ACPI_NUM_FIXED_EVENTS + NUM_COUNTERS_EXTRA; 869 870 all_attrs = kzalloc(sizeof(struct attribute *) * (num_counters + 1), 871 GFP_KERNEL); 872 if (all_attrs == NULL) 873 return; 874 875 all_counters = kzalloc(sizeof(struct event_counter) * (num_counters), 876 GFP_KERNEL); 877 if (all_counters == NULL) 878 goto fail; 879 880 status = acpi_install_global_event_handler(acpi_global_event_handler, NULL); 881 if (ACPI_FAILURE(status)) 882 goto fail; 883 884 counter_attrs = kzalloc(sizeof(struct kobj_attribute) * (num_counters), 885 GFP_KERNEL); 886 if (counter_attrs == NULL) 887 goto fail; 888 889 for (i = 0; i < num_counters; ++i) { 890 char buffer[12]; 891 char *name; 892 893 if (i < num_gpes) 894 sprintf(buffer, "gpe%02X", i); 895 else if (i == num_gpes + ACPI_EVENT_PMTIMER) 896 sprintf(buffer, "ff_pmtimer"); 897 else if (i == num_gpes + ACPI_EVENT_GLOBAL) 898 sprintf(buffer, "ff_gbl_lock"); 899 else if (i == num_gpes + ACPI_EVENT_POWER_BUTTON) 900 sprintf(buffer, "ff_pwr_btn"); 901 else if (i == num_gpes + ACPI_EVENT_SLEEP_BUTTON) 902 sprintf(buffer, "ff_slp_btn"); 903 else if (i == num_gpes + ACPI_EVENT_RTC) 904 sprintf(buffer, "ff_rt_clk"); 905 else if (i == num_gpes + ACPI_NUM_FIXED_EVENTS + COUNT_GPE) 906 sprintf(buffer, "gpe_all"); 907 else if (i == num_gpes + ACPI_NUM_FIXED_EVENTS + COUNT_SCI) 908 sprintf(buffer, "sci"); 909 else if (i == num_gpes + ACPI_NUM_FIXED_EVENTS + COUNT_SCI_NOT) 910 sprintf(buffer, "sci_not"); 911 else if (i == num_gpes + ACPI_NUM_FIXED_EVENTS + COUNT_ERROR) 912 sprintf(buffer, "error"); 913 else 914 sprintf(buffer, "bug%02X", i); 915 916 name = kstrdup(buffer, GFP_KERNEL); 917 if (name == NULL) 918 goto fail; 919 920 sysfs_attr_init(&counter_attrs[i].attr); 921 counter_attrs[i].attr.name = name; 922 counter_attrs[i].attr.mode = 0644; 923 counter_attrs[i].show = counter_show; 924 counter_attrs[i].store = counter_set; 925 926 all_attrs[i] = &counter_attrs[i].attr; 927 } 928 929 interrupt_stats_attr_group.attrs = all_attrs; 930 if (!sysfs_create_group(acpi_kobj, &interrupt_stats_attr_group)) 931 return; 932 933 fail: 934 delete_gpe_attr_array(); 935 return; 936 } 937 938 static void __exit interrupt_stats_exit(void) 939 { 940 sysfs_remove_group(acpi_kobj, &interrupt_stats_attr_group); 941 942 delete_gpe_attr_array(); 943 944 return; 945 } 946 947 static ssize_t 948 acpi_show_profile(struct device *dev, struct device_attribute *attr, 949 char *buf) 950 { 951 return sprintf(buf, "%d\n", acpi_gbl_FADT.preferred_profile); 952 } 953 954 static const struct device_attribute pm_profile_attr = 955 __ATTR(pm_profile, S_IRUGO, acpi_show_profile, NULL); 956 957 static ssize_t hotplug_enabled_show(struct kobject *kobj, 958 struct kobj_attribute *attr, char *buf) 959 { 960 struct acpi_hotplug_profile *hotplug = to_acpi_hotplug_profile(kobj); 961 962 return sprintf(buf, "%d\n", hotplug->enabled); 963 } 964 965 static ssize_t hotplug_enabled_store(struct kobject *kobj, 966 struct kobj_attribute *attr, 967 const char *buf, size_t size) 968 { 969 struct acpi_hotplug_profile *hotplug = to_acpi_hotplug_profile(kobj); 970 unsigned int val; 971 972 if (kstrtouint(buf, 10, &val) || val > 1) 973 return -EINVAL; 974 975 acpi_scan_hotplug_enabled(hotplug, val); 976 return size; 977 } 978 979 static struct kobj_attribute hotplug_enabled_attr = 980 __ATTR(enabled, S_IRUGO | S_IWUSR, hotplug_enabled_show, 981 hotplug_enabled_store); 982 983 static struct attribute *hotplug_profile_attrs[] = { 984 &hotplug_enabled_attr.attr, 985 NULL 986 }; 987 988 static struct kobj_type acpi_hotplug_profile_ktype = { 989 .sysfs_ops = &kobj_sysfs_ops, 990 .default_attrs = hotplug_profile_attrs, 991 }; 992 993 void acpi_sysfs_add_hotplug_profile(struct acpi_hotplug_profile *hotplug, 994 const char *name) 995 { 996 int error; 997 998 if (!hotplug_kobj) 999 goto err_out; 1000 1001 error = kobject_init_and_add(&hotplug->kobj, 1002 &acpi_hotplug_profile_ktype, hotplug_kobj, "%s", name); 1003 if (error) 1004 goto err_out; 1005 1006 kobject_uevent(&hotplug->kobj, KOBJ_ADD); 1007 return; 1008 1009 err_out: 1010 pr_err(PREFIX "Unable to add hotplug profile '%s'\n", name); 1011 } 1012 1013 static ssize_t force_remove_show(struct kobject *kobj, 1014 struct kobj_attribute *attr, char *buf) 1015 { 1016 return sprintf(buf, "%d\n", 0); 1017 } 1018 1019 static ssize_t force_remove_store(struct kobject *kobj, 1020 struct kobj_attribute *attr, 1021 const char *buf, size_t size) 1022 { 1023 bool val; 1024 int ret; 1025 1026 ret = strtobool(buf, &val); 1027 if (ret < 0) 1028 return ret; 1029 1030 if (val) { 1031 pr_err("Enabling force_remove is not supported anymore. Please report to linux-acpi@vger.kernel.org if you depend on this functionality\n"); 1032 return -EINVAL; 1033 } 1034 return size; 1035 } 1036 1037 static const struct kobj_attribute force_remove_attr = 1038 __ATTR(force_remove, S_IRUGO | S_IWUSR, force_remove_show, 1039 force_remove_store); 1040 1041 int __init acpi_sysfs_init(void) 1042 { 1043 int result; 1044 1045 result = acpi_tables_sysfs_init(); 1046 if (result) 1047 return result; 1048 1049 hotplug_kobj = kobject_create_and_add("hotplug", acpi_kobj); 1050 if (!hotplug_kobj) 1051 return -ENOMEM; 1052 1053 result = sysfs_create_file(hotplug_kobj, &force_remove_attr.attr); 1054 if (result) 1055 return result; 1056 1057 result = sysfs_create_file(acpi_kobj, &pm_profile_attr.attr); 1058 return result; 1059 } 1060