Home
last modified time | relevance | path

Searched refs:sector_nr (Results 1 – 6 of 6) sorted by relevance

/openbmc/linux/fs/btrfs/
H A Dscrub.c592 int sector_nr) in scrub_stripe_get_page_offset() argument
655 for (int i = sector_nr + 1; i < sector_nr + sectors_per_tree; i++) { in scrub_verify_one_metadata()
700 ASSERT(sector_nr >= 0 && sector_nr < stripe->nr_sectors); in scrub_verify_one_sector()
756 int sector_nr; in scrub_verify_one_stripe() local
761 sector_nr += sectors_per_tree - 1; in scrub_verify_one_stripe()
793 ASSERT(sector_nr < stripe->nr_sectors); in scrub_repair_read_endio()
879 int sector_nr; in scrub_stripe_report_errors() local
1200 int sector_nr; in scrub_write_sectors() local
1211 if (bbio && sector_nr && !test_bit(sector_nr - 1, &write_bitmap)) { in scrub_write_sectors()
1603 int sector_nr; in scrub_find_fill_first_stripe() local
[all …]
H A Draid56.c609 sector_nr)]; in rbio_stripe_sector()
879 ASSERT(sector_nr >= 0 && sector_nr < rbio->stripe_nsectors); in sector_in_rbio()
1039 unsigned int sector_nr, in rbio_add_io_sector() argument
1055 ASSERT(sector_nr >= 0 && sector_nr < rbio->stripe_nsectors); in rbio_add_io_sector()
1994 int sector_nr; in set_rbio_raid6_extra_error() local
2003 for (sector_nr = 0; sector_nr < rbio->stripe_nsectors; sector_nr++) { in set_rbio_raid6_extra_error()
2563 int sector_nr; in recover_scrub_rbio() local
2579 for (sector_nr = 0; sector_nr < rbio->stripe_nsectors; sector_nr++) { in recover_scrub_rbio()
2691 int sector_nr; in scrub_rbio() local
2715 for (sector_nr = 0; sector_nr < rbio->stripe_nsectors; sector_nr++) { in scrub_rbio()
[all …]
/openbmc/linux/drivers/md/
H A Draid1.c871 int idx = sector_to_idx(sector_nr); in raise_barrier()
921 int idx = sector_to_idx(sector_nr); in lower_barrier()
1001 int idx = sector_to_idx(sector_nr); in wait_read_barrier()
1044 int idx = sector_to_idx(sector_nr); in wait_barrier()
2684 if (sector_nr >= max_sector) { in raid1_sync_request()
2711 return max_sector - sector_nr; in raid1_sync_request()
2754 r1_bio->sector = sector_nr; in raid1_sync_request()
2780 if (first_bad > sector_nr) in raid1_sync_request()
2789 if (sector_nr < first_bad) { in raid1_sync_request()
2873 rv = max_sector - sector_nr; in raid1_sync_request()
[all …]
H A Draid10.c3333 if (sector_nr >= max_sector) { in raid10_sync_request()
3734 r10_bio->sector = sector_nr; in raid10_sync_request()
3737 r10_bio->sectors = (sector_nr | chunk_mask) - sector_nr + 1; in raid10_sync_request()
3843 sector_nr += len>>9; in raid10_sync_request()
3925 sector_nr = max_sector; in raid10_sync_request()
4825 if (sector_nr == 0) { in reshape_request()
4834 if (sector_nr) { in reshape_request()
4838 return sector_nr; in reshape_request()
4923 r10_bio->sector = sector_nr; in reshape_request()
5020 sector_nr += len >> 9; in reshape_request()
[all …]
H A Draid5.c6286 if (sector_nr == 0) { in reshape_request()
6290 sector_nr = raid5_size(mddev, 0, 0) in reshape_request()
6295 sector_nr = MaxSector; in reshape_request()
6298 sector_nr = conf->reshape_progress; in reshape_request()
6300 if (sector_nr) { in reshape_request()
6304 retn = sector_nr; in reshape_request()
6361 stripe_addr = sector_nr; in reshape_request()
6409 rdev->recovery_offset = sector_nr; in reshape_request()
6497 sector_nr += reshape_sectors; in reshape_request()
6518 rdev->recovery_offset = sector_nr; in reshape_request()
[all …]
H A Dmd.h639 sector_t (*sync_request)(struct mddev *mddev, sector_t sector_nr, int *skipped);