1 // SPDX-License-Identifier: GPL-2.0 2 /* 3 * Copyright (c) 2014 Red Hat, Inc. 4 * All Rights Reserved. 5 */ 6 7 #include "xfs.h" 8 #include "xfs_shared.h" 9 #include "xfs_format.h" 10 #include "xfs_log_format.h" 11 #include "xfs_trans_resv.h" 12 #include "xfs_sysfs.h" 13 #include "xfs_log.h" 14 #include "xfs_log_priv.h" 15 #include "xfs_mount.h" 16 17 struct xfs_sysfs_attr { 18 struct attribute attr; 19 ssize_t (*show)(struct kobject *kobject, char *buf); 20 ssize_t (*store)(struct kobject *kobject, const char *buf, 21 size_t count); 22 }; 23 24 static inline struct xfs_sysfs_attr * 25 to_attr(struct attribute *attr) 26 { 27 return container_of(attr, struct xfs_sysfs_attr, attr); 28 } 29 30 #define XFS_SYSFS_ATTR_RW(name) \ 31 static struct xfs_sysfs_attr xfs_sysfs_attr_##name = __ATTR_RW(name) 32 #define XFS_SYSFS_ATTR_RO(name) \ 33 static struct xfs_sysfs_attr xfs_sysfs_attr_##name = __ATTR_RO(name) 34 #define XFS_SYSFS_ATTR_WO(name) \ 35 static struct xfs_sysfs_attr xfs_sysfs_attr_##name = __ATTR_WO(name) 36 37 #define ATTR_LIST(name) &xfs_sysfs_attr_##name.attr 38 39 STATIC ssize_t 40 xfs_sysfs_object_show( 41 struct kobject *kobject, 42 struct attribute *attr, 43 char *buf) 44 { 45 struct xfs_sysfs_attr *xfs_attr = to_attr(attr); 46 47 return xfs_attr->show ? xfs_attr->show(kobject, buf) : 0; 48 } 49 50 STATIC ssize_t 51 xfs_sysfs_object_store( 52 struct kobject *kobject, 53 struct attribute *attr, 54 const char *buf, 55 size_t count) 56 { 57 struct xfs_sysfs_attr *xfs_attr = to_attr(attr); 58 59 return xfs_attr->store ? xfs_attr->store(kobject, buf, count) : 0; 60 } 61 62 static const struct sysfs_ops xfs_sysfs_ops = { 63 .show = xfs_sysfs_object_show, 64 .store = xfs_sysfs_object_store, 65 }; 66 67 static struct attribute *xfs_mp_attrs[] = { 68 NULL, 69 }; 70 71 struct kobj_type xfs_mp_ktype = { 72 .release = xfs_sysfs_release, 73 .sysfs_ops = &xfs_sysfs_ops, 74 .default_attrs = xfs_mp_attrs, 75 }; 76 77 #ifdef DEBUG 78 /* debug */ 79 80 STATIC ssize_t 81 bug_on_assert_store( 82 struct kobject *kobject, 83 const char *buf, 84 size_t count) 85 { 86 int ret; 87 int val; 88 89 ret = kstrtoint(buf, 0, &val); 90 if (ret) 91 return ret; 92 93 if (val == 1) 94 xfs_globals.bug_on_assert = true; 95 else if (val == 0) 96 xfs_globals.bug_on_assert = false; 97 else 98 return -EINVAL; 99 100 return count; 101 } 102 103 STATIC ssize_t 104 bug_on_assert_show( 105 struct kobject *kobject, 106 char *buf) 107 { 108 return sysfs_emit(buf, "%d\n", xfs_globals.bug_on_assert); 109 } 110 XFS_SYSFS_ATTR_RW(bug_on_assert); 111 112 STATIC ssize_t 113 log_recovery_delay_store( 114 struct kobject *kobject, 115 const char *buf, 116 size_t count) 117 { 118 int ret; 119 int val; 120 121 ret = kstrtoint(buf, 0, &val); 122 if (ret) 123 return ret; 124 125 if (val < 0 || val > 60) 126 return -EINVAL; 127 128 xfs_globals.log_recovery_delay = val; 129 130 return count; 131 } 132 133 STATIC ssize_t 134 log_recovery_delay_show( 135 struct kobject *kobject, 136 char *buf) 137 { 138 return sysfs_emit(buf, "%d\n", xfs_globals.log_recovery_delay); 139 } 140 XFS_SYSFS_ATTR_RW(log_recovery_delay); 141 142 STATIC ssize_t 143 mount_delay_store( 144 struct kobject *kobject, 145 const char *buf, 146 size_t count) 147 { 148 int ret; 149 int val; 150 151 ret = kstrtoint(buf, 0, &val); 152 if (ret) 153 return ret; 154 155 if (val < 0 || val > 60) 156 return -EINVAL; 157 158 xfs_globals.mount_delay = val; 159 160 return count; 161 } 162 163 STATIC ssize_t 164 mount_delay_show( 165 struct kobject *kobject, 166 char *buf) 167 { 168 return sysfs_emit(buf, "%d\n", xfs_globals.mount_delay); 169 } 170 XFS_SYSFS_ATTR_RW(mount_delay); 171 172 static ssize_t 173 always_cow_store( 174 struct kobject *kobject, 175 const char *buf, 176 size_t count) 177 { 178 ssize_t ret; 179 180 ret = kstrtobool(buf, &xfs_globals.always_cow); 181 if (ret < 0) 182 return ret; 183 return count; 184 } 185 186 static ssize_t 187 always_cow_show( 188 struct kobject *kobject, 189 char *buf) 190 { 191 return sysfs_emit(buf, "%d\n", xfs_globals.always_cow); 192 } 193 XFS_SYSFS_ATTR_RW(always_cow); 194 195 #ifdef DEBUG 196 /* 197 * Override how many threads the parallel work queue is allowed to create. 198 * This has to be a debug-only global (instead of an errortag) because one of 199 * the main users of parallel workqueues is mount time quotacheck. 200 */ 201 STATIC ssize_t 202 pwork_threads_store( 203 struct kobject *kobject, 204 const char *buf, 205 size_t count) 206 { 207 int ret; 208 int val; 209 210 ret = kstrtoint(buf, 0, &val); 211 if (ret) 212 return ret; 213 214 if (val < -1 || val > num_possible_cpus()) 215 return -EINVAL; 216 217 xfs_globals.pwork_threads = val; 218 219 return count; 220 } 221 222 STATIC ssize_t 223 pwork_threads_show( 224 struct kobject *kobject, 225 char *buf) 226 { 227 return sysfs_emit(buf, "%d\n", xfs_globals.pwork_threads); 228 } 229 XFS_SYSFS_ATTR_RW(pwork_threads); 230 #endif /* DEBUG */ 231 232 static struct attribute *xfs_dbg_attrs[] = { 233 ATTR_LIST(bug_on_assert), 234 ATTR_LIST(log_recovery_delay), 235 ATTR_LIST(mount_delay), 236 ATTR_LIST(always_cow), 237 #ifdef DEBUG 238 ATTR_LIST(pwork_threads), 239 #endif 240 NULL, 241 }; 242 243 struct kobj_type xfs_dbg_ktype = { 244 .release = xfs_sysfs_release, 245 .sysfs_ops = &xfs_sysfs_ops, 246 .default_attrs = xfs_dbg_attrs, 247 }; 248 249 #endif /* DEBUG */ 250 251 /* stats */ 252 253 static inline struct xstats * 254 to_xstats(struct kobject *kobject) 255 { 256 struct xfs_kobj *kobj = to_kobj(kobject); 257 258 return container_of(kobj, struct xstats, xs_kobj); 259 } 260 261 STATIC ssize_t 262 stats_show( 263 struct kobject *kobject, 264 char *buf) 265 { 266 struct xstats *stats = to_xstats(kobject); 267 268 return xfs_stats_format(stats->xs_stats, buf); 269 } 270 XFS_SYSFS_ATTR_RO(stats); 271 272 STATIC ssize_t 273 stats_clear_store( 274 struct kobject *kobject, 275 const char *buf, 276 size_t count) 277 { 278 int ret; 279 int val; 280 struct xstats *stats = to_xstats(kobject); 281 282 ret = kstrtoint(buf, 0, &val); 283 if (ret) 284 return ret; 285 286 if (val != 1) 287 return -EINVAL; 288 289 xfs_stats_clearall(stats->xs_stats); 290 return count; 291 } 292 XFS_SYSFS_ATTR_WO(stats_clear); 293 294 static struct attribute *xfs_stats_attrs[] = { 295 ATTR_LIST(stats), 296 ATTR_LIST(stats_clear), 297 NULL, 298 }; 299 300 struct kobj_type xfs_stats_ktype = { 301 .release = xfs_sysfs_release, 302 .sysfs_ops = &xfs_sysfs_ops, 303 .default_attrs = xfs_stats_attrs, 304 }; 305 306 /* xlog */ 307 308 static inline struct xlog * 309 to_xlog(struct kobject *kobject) 310 { 311 struct xfs_kobj *kobj = to_kobj(kobject); 312 313 return container_of(kobj, struct xlog, l_kobj); 314 } 315 316 STATIC ssize_t 317 log_head_lsn_show( 318 struct kobject *kobject, 319 char *buf) 320 { 321 int cycle; 322 int block; 323 struct xlog *log = to_xlog(kobject); 324 325 spin_lock(&log->l_icloglock); 326 cycle = log->l_curr_cycle; 327 block = log->l_curr_block; 328 spin_unlock(&log->l_icloglock); 329 330 return sysfs_emit(buf, "%d:%d\n", cycle, block); 331 } 332 XFS_SYSFS_ATTR_RO(log_head_lsn); 333 334 STATIC ssize_t 335 log_tail_lsn_show( 336 struct kobject *kobject, 337 char *buf) 338 { 339 int cycle; 340 int block; 341 struct xlog *log = to_xlog(kobject); 342 343 xlog_crack_atomic_lsn(&log->l_tail_lsn, &cycle, &block); 344 return sysfs_emit(buf, "%d:%d\n", cycle, block); 345 } 346 XFS_SYSFS_ATTR_RO(log_tail_lsn); 347 348 STATIC ssize_t 349 reserve_grant_head_show( 350 struct kobject *kobject, 351 char *buf) 352 353 { 354 int cycle; 355 int bytes; 356 struct xlog *log = to_xlog(kobject); 357 358 xlog_crack_grant_head(&log->l_reserve_head.grant, &cycle, &bytes); 359 return sysfs_emit(buf, "%d:%d\n", cycle, bytes); 360 } 361 XFS_SYSFS_ATTR_RO(reserve_grant_head); 362 363 STATIC ssize_t 364 write_grant_head_show( 365 struct kobject *kobject, 366 char *buf) 367 { 368 int cycle; 369 int bytes; 370 struct xlog *log = to_xlog(kobject); 371 372 xlog_crack_grant_head(&log->l_write_head.grant, &cycle, &bytes); 373 return sysfs_emit(buf, "%d:%d\n", cycle, bytes); 374 } 375 XFS_SYSFS_ATTR_RO(write_grant_head); 376 377 static struct attribute *xfs_log_attrs[] = { 378 ATTR_LIST(log_head_lsn), 379 ATTR_LIST(log_tail_lsn), 380 ATTR_LIST(reserve_grant_head), 381 ATTR_LIST(write_grant_head), 382 NULL, 383 }; 384 385 struct kobj_type xfs_log_ktype = { 386 .release = xfs_sysfs_release, 387 .sysfs_ops = &xfs_sysfs_ops, 388 .default_attrs = xfs_log_attrs, 389 }; 390 391 /* 392 * Metadata IO error configuration 393 * 394 * The sysfs structure here is: 395 * ...xfs/<dev>/error/<class>/<errno>/<error_attrs> 396 * 397 * where <class> allows us to discriminate between data IO and metadata IO, 398 * and any other future type of IO (e.g. special inode or directory error 399 * handling) we care to support. 400 */ 401 static inline struct xfs_error_cfg * 402 to_error_cfg(struct kobject *kobject) 403 { 404 struct xfs_kobj *kobj = to_kobj(kobject); 405 return container_of(kobj, struct xfs_error_cfg, kobj); 406 } 407 408 static inline struct xfs_mount * 409 err_to_mp(struct kobject *kobject) 410 { 411 struct xfs_kobj *kobj = to_kobj(kobject); 412 return container_of(kobj, struct xfs_mount, m_error_kobj); 413 } 414 415 static ssize_t 416 max_retries_show( 417 struct kobject *kobject, 418 char *buf) 419 { 420 int retries; 421 struct xfs_error_cfg *cfg = to_error_cfg(kobject); 422 423 if (cfg->max_retries == XFS_ERR_RETRY_FOREVER) 424 retries = -1; 425 else 426 retries = cfg->max_retries; 427 428 return sysfs_emit(buf, "%d\n", retries); 429 } 430 431 static ssize_t 432 max_retries_store( 433 struct kobject *kobject, 434 const char *buf, 435 size_t count) 436 { 437 struct xfs_error_cfg *cfg = to_error_cfg(kobject); 438 int ret; 439 int val; 440 441 ret = kstrtoint(buf, 0, &val); 442 if (ret) 443 return ret; 444 445 if (val < -1) 446 return -EINVAL; 447 448 if (val == -1) 449 cfg->max_retries = XFS_ERR_RETRY_FOREVER; 450 else 451 cfg->max_retries = val; 452 return count; 453 } 454 XFS_SYSFS_ATTR_RW(max_retries); 455 456 static ssize_t 457 retry_timeout_seconds_show( 458 struct kobject *kobject, 459 char *buf) 460 { 461 int timeout; 462 struct xfs_error_cfg *cfg = to_error_cfg(kobject); 463 464 if (cfg->retry_timeout == XFS_ERR_RETRY_FOREVER) 465 timeout = -1; 466 else 467 timeout = jiffies_to_msecs(cfg->retry_timeout) / MSEC_PER_SEC; 468 469 return sysfs_emit(buf, "%d\n", timeout); 470 } 471 472 static ssize_t 473 retry_timeout_seconds_store( 474 struct kobject *kobject, 475 const char *buf, 476 size_t count) 477 { 478 struct xfs_error_cfg *cfg = to_error_cfg(kobject); 479 int ret; 480 int val; 481 482 ret = kstrtoint(buf, 0, &val); 483 if (ret) 484 return ret; 485 486 /* 1 day timeout maximum, -1 means infinite */ 487 if (val < -1 || val > 86400) 488 return -EINVAL; 489 490 if (val == -1) 491 cfg->retry_timeout = XFS_ERR_RETRY_FOREVER; 492 else { 493 cfg->retry_timeout = msecs_to_jiffies(val * MSEC_PER_SEC); 494 ASSERT(msecs_to_jiffies(val * MSEC_PER_SEC) < LONG_MAX); 495 } 496 return count; 497 } 498 XFS_SYSFS_ATTR_RW(retry_timeout_seconds); 499 500 static ssize_t 501 fail_at_unmount_show( 502 struct kobject *kobject, 503 char *buf) 504 { 505 struct xfs_mount *mp = err_to_mp(kobject); 506 507 return sysfs_emit(buf, "%d\n", mp->m_fail_unmount); 508 } 509 510 static ssize_t 511 fail_at_unmount_store( 512 struct kobject *kobject, 513 const char *buf, 514 size_t count) 515 { 516 struct xfs_mount *mp = err_to_mp(kobject); 517 int ret; 518 int val; 519 520 ret = kstrtoint(buf, 0, &val); 521 if (ret) 522 return ret; 523 524 if (val < 0 || val > 1) 525 return -EINVAL; 526 527 mp->m_fail_unmount = val; 528 return count; 529 } 530 XFS_SYSFS_ATTR_RW(fail_at_unmount); 531 532 static struct attribute *xfs_error_attrs[] = { 533 ATTR_LIST(max_retries), 534 ATTR_LIST(retry_timeout_seconds), 535 NULL, 536 }; 537 538 539 static struct kobj_type xfs_error_cfg_ktype = { 540 .release = xfs_sysfs_release, 541 .sysfs_ops = &xfs_sysfs_ops, 542 .default_attrs = xfs_error_attrs, 543 }; 544 545 static struct kobj_type xfs_error_ktype = { 546 .release = xfs_sysfs_release, 547 .sysfs_ops = &xfs_sysfs_ops, 548 }; 549 550 /* 551 * Error initialization tables. These need to be ordered in the same 552 * order as the enums used to index the array. All class init tables need to 553 * define a "default" behaviour as the first entry, all other entries can be 554 * empty. 555 */ 556 struct xfs_error_init { 557 char *name; 558 int max_retries; 559 int retry_timeout; /* in seconds */ 560 }; 561 562 static const struct xfs_error_init xfs_error_meta_init[XFS_ERR_ERRNO_MAX] = { 563 { .name = "default", 564 .max_retries = XFS_ERR_RETRY_FOREVER, 565 .retry_timeout = XFS_ERR_RETRY_FOREVER, 566 }, 567 { .name = "EIO", 568 .max_retries = XFS_ERR_RETRY_FOREVER, 569 .retry_timeout = XFS_ERR_RETRY_FOREVER, 570 }, 571 { .name = "ENOSPC", 572 .max_retries = XFS_ERR_RETRY_FOREVER, 573 .retry_timeout = XFS_ERR_RETRY_FOREVER, 574 }, 575 { .name = "ENODEV", 576 .max_retries = 0, /* We can't recover from devices disappearing */ 577 .retry_timeout = 0, 578 }, 579 }; 580 581 static int 582 xfs_error_sysfs_init_class( 583 struct xfs_mount *mp, 584 int class, 585 const char *parent_name, 586 struct xfs_kobj *parent_kobj, 587 const struct xfs_error_init init[]) 588 { 589 struct xfs_error_cfg *cfg; 590 int error; 591 int i; 592 593 ASSERT(class < XFS_ERR_CLASS_MAX); 594 595 error = xfs_sysfs_init(parent_kobj, &xfs_error_ktype, 596 &mp->m_error_kobj, parent_name); 597 if (error) 598 return error; 599 600 for (i = 0; i < XFS_ERR_ERRNO_MAX; i++) { 601 cfg = &mp->m_error_cfg[class][i]; 602 error = xfs_sysfs_init(&cfg->kobj, &xfs_error_cfg_ktype, 603 parent_kobj, init[i].name); 604 if (error) 605 goto out_error; 606 607 cfg->max_retries = init[i].max_retries; 608 if (init[i].retry_timeout == XFS_ERR_RETRY_FOREVER) 609 cfg->retry_timeout = XFS_ERR_RETRY_FOREVER; 610 else 611 cfg->retry_timeout = msecs_to_jiffies( 612 init[i].retry_timeout * MSEC_PER_SEC); 613 } 614 return 0; 615 616 out_error: 617 /* unwind the entries that succeeded */ 618 for (i--; i >= 0; i--) { 619 cfg = &mp->m_error_cfg[class][i]; 620 xfs_sysfs_del(&cfg->kobj); 621 } 622 xfs_sysfs_del(parent_kobj); 623 return error; 624 } 625 626 int 627 xfs_error_sysfs_init( 628 struct xfs_mount *mp) 629 { 630 int error; 631 632 /* .../xfs/<dev>/error/ */ 633 error = xfs_sysfs_init(&mp->m_error_kobj, &xfs_error_ktype, 634 &mp->m_kobj, "error"); 635 if (error) 636 return error; 637 638 error = sysfs_create_file(&mp->m_error_kobj.kobject, 639 ATTR_LIST(fail_at_unmount)); 640 641 if (error) 642 goto out_error; 643 644 /* .../xfs/<dev>/error/metadata/ */ 645 error = xfs_error_sysfs_init_class(mp, XFS_ERR_METADATA, 646 "metadata", &mp->m_error_meta_kobj, 647 xfs_error_meta_init); 648 if (error) 649 goto out_error; 650 651 return 0; 652 653 out_error: 654 xfs_sysfs_del(&mp->m_error_kobj); 655 return error; 656 } 657 658 void 659 xfs_error_sysfs_del( 660 struct xfs_mount *mp) 661 { 662 struct xfs_error_cfg *cfg; 663 int i, j; 664 665 for (i = 0; i < XFS_ERR_CLASS_MAX; i++) { 666 for (j = 0; j < XFS_ERR_ERRNO_MAX; j++) { 667 cfg = &mp->m_error_cfg[i][j]; 668 669 xfs_sysfs_del(&cfg->kobj); 670 } 671 } 672 xfs_sysfs_del(&mp->m_error_meta_kobj); 673 xfs_sysfs_del(&mp->m_error_kobj); 674 } 675 676 struct xfs_error_cfg * 677 xfs_error_get_cfg( 678 struct xfs_mount *mp, 679 int error_class, 680 int error) 681 { 682 struct xfs_error_cfg *cfg; 683 684 if (error < 0) 685 error = -error; 686 687 switch (error) { 688 case EIO: 689 cfg = &mp->m_error_cfg[error_class][XFS_ERR_EIO]; 690 break; 691 case ENOSPC: 692 cfg = &mp->m_error_cfg[error_class][XFS_ERR_ENOSPC]; 693 break; 694 case ENODEV: 695 cfg = &mp->m_error_cfg[error_class][XFS_ERR_ENODEV]; 696 break; 697 default: 698 cfg = &mp->m_error_cfg[error_class][XFS_ERR_DEFAULT]; 699 break; 700 } 701 702 return cfg; 703 } 704