Searched refs:sector_nr (Results 1 – 6 of 6) sorted by relevance
592 int sector_nr) in scrub_stripe_get_page_offset() argument655 for (int i = sector_nr + 1; i < sector_nr + sectors_per_tree; i++) { in scrub_verify_one_metadata()700 ASSERT(sector_nr >= 0 && sector_nr < stripe->nr_sectors); in scrub_verify_one_sector()756 int sector_nr; in scrub_verify_one_stripe() local761 sector_nr += sectors_per_tree - 1; in scrub_verify_one_stripe()793 ASSERT(sector_nr < stripe->nr_sectors); in scrub_repair_read_endio()879 int sector_nr; in scrub_stripe_report_errors() local1200 int sector_nr; in scrub_write_sectors() local1211 if (bbio && sector_nr && !test_bit(sector_nr - 1, &write_bitmap)) { in scrub_write_sectors()1603 int sector_nr; in scrub_find_fill_first_stripe() local[all …]
609 sector_nr)]; in rbio_stripe_sector()879 ASSERT(sector_nr >= 0 && sector_nr < rbio->stripe_nsectors); in sector_in_rbio()1039 unsigned int sector_nr, in rbio_add_io_sector() argument1055 ASSERT(sector_nr >= 0 && sector_nr < rbio->stripe_nsectors); in rbio_add_io_sector()1994 int sector_nr; in set_rbio_raid6_extra_error() local2003 for (sector_nr = 0; sector_nr < rbio->stripe_nsectors; sector_nr++) { in set_rbio_raid6_extra_error()2563 int sector_nr; in recover_scrub_rbio() local2579 for (sector_nr = 0; sector_nr < rbio->stripe_nsectors; sector_nr++) { in recover_scrub_rbio()2691 int sector_nr; in scrub_rbio() local2715 for (sector_nr = 0; sector_nr < rbio->stripe_nsectors; sector_nr++) { in scrub_rbio()[all …]
871 int idx = sector_to_idx(sector_nr); in raise_barrier()921 int idx = sector_to_idx(sector_nr); in lower_barrier()1001 int idx = sector_to_idx(sector_nr); in wait_read_barrier()1044 int idx = sector_to_idx(sector_nr); in wait_barrier()2684 if (sector_nr >= max_sector) { in raid1_sync_request()2711 return max_sector - sector_nr; in raid1_sync_request()2754 r1_bio->sector = sector_nr; in raid1_sync_request()2780 if (first_bad > sector_nr) in raid1_sync_request()2789 if (sector_nr < first_bad) { in raid1_sync_request()2873 rv = max_sector - sector_nr; in raid1_sync_request()[all …]
3333 if (sector_nr >= max_sector) { in raid10_sync_request()3734 r10_bio->sector = sector_nr; in raid10_sync_request()3737 r10_bio->sectors = (sector_nr | chunk_mask) - sector_nr + 1; in raid10_sync_request()3843 sector_nr += len>>9; in raid10_sync_request()3925 sector_nr = max_sector; in raid10_sync_request()4825 if (sector_nr == 0) { in reshape_request()4834 if (sector_nr) { in reshape_request()4838 return sector_nr; in reshape_request()4923 r10_bio->sector = sector_nr; in reshape_request()5020 sector_nr += len >> 9; in reshape_request()[all …]
6286 if (sector_nr == 0) { in reshape_request()6290 sector_nr = raid5_size(mddev, 0, 0) in reshape_request()6295 sector_nr = MaxSector; in reshape_request()6298 sector_nr = conf->reshape_progress; in reshape_request()6300 if (sector_nr) { in reshape_request()6304 retn = sector_nr; in reshape_request()6361 stripe_addr = sector_nr; in reshape_request()6409 rdev->recovery_offset = sector_nr; in reshape_request()6497 sector_nr += reshape_sectors; in reshape_request()6518 rdev->recovery_offset = sector_nr; in reshape_request()[all …]
639 sector_t (*sync_request)(struct mddev *mddev, sector_t sector_nr, int *skipped);