/openbmc/linux/drivers/md/ |
H A D | md-multipath.c | 141 conf->raid_disks - mddev->degraded); in multipath_status() 158 if (conf->raid_disks - mddev->degraded <= 1) { in multipath_error() 174 mddev->degraded++; in multipath_error() 182 conf->raid_disks - mddev->degraded); in multipath_error() 195 pr_debug(" --- wd:%d rd:%d\n", conf->raid_disks - conf->mddev->degraded, in print_multipath_conf() 230 mddev->degraded--; in multipath_add_disk() 397 mddev->degraded = conf->raid_disks - working_disks; in multipath_run() 410 mdname(mddev), conf->raid_disks - mddev->degraded, in multipath_run()
|
H A D | raid0.c | 645 if (mddev->degraded != 1) { in raid0_takeover_raid45() 648 mddev->degraded); in raid0_takeover_raid45() 698 if (mddev->degraded != (mddev->raid_disks>>1)) { in raid0_takeover_raid10() 710 mddev->degraded = 0; in raid0_takeover_raid10() 727 if ((mddev->raid_disks - 1) != mddev->degraded) { in raid0_takeover_raid1()
|
H A D | raid1.c | 390 if (r1_bio->mddev->degraded == conf->raid_disks || in raid1_end_read_request() 391 (r1_bio->mddev->degraded == conf->raid_disks-1 && in raid1_end_read_request() 1543 conf->raid_disks - mddev->degraded > 1) in raid1_write_request() 1602 conf->raid_disks - mddev->degraded); in raid1_status() 1637 (conf->raid_disks - mddev->degraded) == 1) { in raid1_error() 1648 mddev->degraded++; in raid1_error() 1660 mdname(mddev), conf->raid_disks - mddev->degraded); in raid1_error() 1672 pr_debug(" --- wd:%d rd:%d\n", conf->raid_disks - conf->mddev->degraded, in print_conf() 1744 mddev->degraded -= count; in raid1_spare_active() 1842 mddev->degraded < conf->raid_disks) { in raid1_remove_disk() [all …]
|
H A D | raid10.c | 1938 conf->geo.raid_disks - mddev->degraded); in raid10_status() 2031 mddev->degraded++; in raid10_error() 2042 mdname(mddev), conf->geo.raid_disks - mddev->degraded); in raid10_error() 2055 pr_debug(" --- wd:%d rd:%d\n", conf->geo.raid_disks - conf->mddev->degraded, in print_conf() 2119 mddev->degraded -= count; in raid10_spare_active() 3694 &sync_blocks, mddev->degraded) && in raid10_sync_request() 4222 mddev->degraded = 0; in raid10_run() 4240 mddev->degraded++; in raid10_run() 4259 mdname(mddev), conf->geo.raid_disks - mddev->degraded, in raid10_run() 4373 if (mddev->degraded > 0) { in raid10_takeover_raid0() [all …]
|
H A D | md-bitmap.h | 268 int md_bitmap_start_sync(struct bitmap *bitmap, sector_t offset, sector_t *blocks, int degraded);
|
H A D | raid5.c | 691 int degraded, degraded2; in raid5_calc_degraded() local 695 degraded = 0; in raid5_calc_degraded() 701 degraded++; in raid5_calc_degraded() 715 degraded++; in raid5_calc_degraded() 719 return degraded; in raid5_calc_degraded() 740 if (degraded2 > degraded) in raid5_calc_degraded() 742 return degraded; in raid5_calc_degraded() 747 int degraded = conf->mddev->degraded; in has_failed() local 753 degraded = raid5_calc_degraded(conf); in has_failed() 755 return degraded > conf->max_degraded; in has_failed() [all …]
|
H A D | md-bitmap.c | 1539 if (!bitmap->mddev->degraded) { in md_bitmap_endwrite() 1568 int degraded) in __bitmap_start_sync() argument 1585 if (!degraded) { /* don't set/clear bits if degraded */ in __bitmap_start_sync() 1596 int degraded) in md_bitmap_start_sync() argument 1611 &blocks1, degraded); in md_bitmap_start_sync() 2016 if (mddev->degraded == 0 in md_bitmap_load() 2682 if (mddev->degraded) in can_clear_store()
|
H A D | md.h | 436 int degraded; /* whether md should consider member
|
H A D | dm-raid.c | 1720 if (rs->md.degraded) { in rs_check_takeover() 1790 mddev->degraded = 1; in rs_check_takeover() 2019 else if (mddev->degraded) in rs_check_reshape() 2786 mddev->degraded = rs->delta_disks < 0 ? -rs->delta_disks : rs->delta_disks; in rs_prepare_reshape()
|
H A D | md.c | 2738 if (mddev->degraded) in md_update_sb() 2872 if (mddev->degraded) in add_bound_rdev() 4045 mddev->degraded = 0; in level_store() 5027 return sprintf(page, "%d\n", mddev->degraded); in degraded_show() 5029 static struct md_sysfs_entry md_degraded = __ATTR_RO(degraded); 6073 if (mddev->degraded) in md_run() 6112 if (mddev->degraded && md_is_rdwr(mddev)) in md_run() 6280 mddev->degraded = 0; in md_clean() 8091 if (mddev->degraded && !test_bit(MD_BROKEN, &mddev->flags)) in md_error() 9578 mddev->degraded != mddev->raid_disks) { in md_reap_sync_thread() [all …]
|
/openbmc/linux/Documentation/driver-api/md/ |
H A D | raid5-ppl.rst | 8 in degraded state, there is no way to recalculate parity, because one of the 10 array or using it is as degraded - data calculated from parity for array blocks 13 this, md by default does not allow starting a dirty degraded array.
|
H A D | raid5-cache.rst | 29 unclean shutdown. We call an array degraded if it has inconsistent data. MD 37 shutdown even the array is degraded. Thus the cache can close the 'write hole'.
|
/openbmc/dbus-sensors/src/mctp/ |
H A D | MCTPEndpoint.hpp | 115 virtual void subscribe(Event&& degraded, Event&& available, 216 void subscribe(Event&& degraded, Event&& available,
|
H A D | MCTPEndpoint.cpp | 221 void MCTPDEndpoint::subscribe(Event&& degraded, Event&& available, in subscribe() argument 228 this->notifyDegraded = std::move(degraded); in subscribe()
|
/openbmc/dbus-sensors/src/fan/ |
H A D | TachSensor.hpp | 27 constexpr const char* degraded = "Degraded"; variable
|
H A D | TachSensor.cpp | 260 newState = redundancy::degraded; in update()
|
/openbmc/linux/Documentation/admin-guide/ |
H A D | md.rst | 89 Boot time assembly of degraded/dirty arrays 92 If a raid5 or raid6 array is both dirty and degraded, it could have 95 is degraded means that some datablocks are missing and cannot reliably 107 when set to 1, bypassed the checks and will allows dirty degraded 110 So, to boot with a root filesystem of a dirty degraded raid 5 or 6, use:: 399 degraded array, or if the array becomes degraded during a write. 401 once the array becomes non-degraded, and this fact has been 663 degraded 665 arrays is degraded. So an optimal array will show ``0``. A
|
/openbmc/phosphor-logging/extensions/openpower-pels/ |
H A D | README.md | 725 The degraded mode reporting functionality (i.e. nag) implemented by IBM creates 726 periodic degraded mode reports when a system is running in degraded mode. This 729 includes the PELs created for the fails leading to the degraded mode. These PELs 734 FRU is no longer considered degraded the PELs that led to its replacement no 735 longer need be picked up in the degraded mode report.
|
/openbmc/dbus-sensors/src/tests/ |
H A D | test_MCTPReactor.cpp | 37 (Event && degraded, Event&& available, Event&& removed),
|
/openbmc/linux/arch/arm/boot/dts/nxp/imx/ |
H A D | imx7d-flex-concentrator.dts | 55 label = "hmi-b:red:heartbeat-degraded";
|
/openbmc/linux/Documentation/networking/ |
H A D | ipvlan.rst | 131 puts the NIC in promiscuous mode and degraded performance is a concern.
|
/openbmc/docs/designs/ |
H A D | target-fail-monitoring.md | 101 - Changes BMC state (CurrentBMCState) to indicate a degraded mode of the BMC
|
H A D | bmc-service-failure-debug-and-recovery.md | 17 the BMC is still operational in a degraded mode. 456 - Changes BMC state (CurrentBMCState) to indicate a degraded mode of the BMC
|
/openbmc/linux/Documentation/PCI/ |
H A D | pcieaer-howto.rst | 104 of those errors, which may result in degraded performance or function
|
/openbmc/linux/Documentation/scsi/ |
H A D | ChangeLog.megaraid_sas | 105 1. Fix reglockFlags for degraded raid5/6 for MR 9360/9380. 130 1. Fix FastPath I/O to work with degraded RAID 1.
|