/openbmc/qemu/hw/block/ |
H A D | nand.c | 127 # define PAGE_SECTORS 1 macro 132 # define PAGE_SECTORS 1 macro 137 # define PAGE_SECTORS 4 macro 658 uint8_t iobuf[(PAGE_SECTORS + 2) * 0x200]; 670 PAGE_SECTORS << BDRV_SECTOR_BITS, iobuf, 0) < 0) { 683 PAGE_SECTORS << BDRV_SECTOR_BITS, iobuf, 0) < 0) { 691 (PAGE_SECTORS + 2) << BDRV_SECTOR_BITS, iobuf, 0) < 0) { 699 (PAGE_SECTORS + 2) << BDRV_SECTOR_BITS, iobuf, 0) < 0) { 776 PAGE_SECTORS << BDRV_SECTOR_BITS, s->io, 0) < 0) { 786 (PAGE_SECTORS + 2) << BDRV_SECTOR_BITS, s->io, 0) [all …]
|
/openbmc/linux/fs/squashfs/ |
H A D | block.c | 119 bio_trim(bio, start_idx * PAGE_SECTORS, in squashfs_bio_read_cached() 120 (end_idx - start_idx) * PAGE_SECTORS); in squashfs_bio_read_cached() 134 bio_trim(bio, start_idx * PAGE_SECTORS, in squashfs_bio_read_cached() 135 (end_idx - start_idx) * PAGE_SECTORS); in squashfs_bio_read_cached()
|
/openbmc/linux/drivers/block/ |
H A D | brd.c | 126 unsigned int offset = (sector & (PAGE_SECTORS-1)) << SECTOR_SHIFT; in copy_to_brd_setup() 149 unsigned int offset = (sector & (PAGE_SECTORS-1)) << SECTOR_SHIFT; in copy_to_brd() 181 unsigned int offset = (sector & (PAGE_SECTORS-1)) << SECTOR_SHIFT; in copy_from_brd()
|
/openbmc/linux/drivers/md/bcache/ |
H A D | movinggc.c | 83 DIV_ROUND_UP(KEY_SIZE(&io->w->key), PAGE_SECTORS), 0); in moving_init() 149 DIV_ROUND_UP(KEY_SIZE(&w->key), PAGE_SECTORS)), in read_moving()
|
H A D | bcache.h | 763 ((unsigned int) ((PAGE_SECTORS * (c)->btree_pages) >> (c)->block_bits)) 773 __rounddown_pow_of_two(USHRT_MAX) / PAGE_SECTORS, in meta_bucket_pages() 776 n = sb->bucket_size / PAGE_SECTORS; in meta_bucket_pages()
|
H A D | writeback.c | 335 DIV_ROUND_UP(KEY_SIZE(&w->key), PAGE_SECTORS), 0); in dirty_init() 540 DIV_ROUND_UP(KEY_SIZE(&w->key), PAGE_SECTORS)), in read_dirty()
|
H A D | journal.c | 54 len = min_t(unsigned int, left, PAGE_SECTORS << JSET_BITS); in journal_read_bucket() 871 PAGE_SECTORS << JSET_BITS)) in journal_wait_for_write()
|
H A D | super.c | 121 if (sb->block_size > PAGE_SECTORS) in read_super_common() 129 if (sb->bucket_size < PAGE_SECTORS) in read_super_common() 512 size = meta_bucket_pages(&ca->sb) * PAGE_SECTORS; in __uuid_write() 1917 ((meta_bucket_pages(sb) * PAGE_SECTORS) / sb->block_size) * in bch_cache_set_alloc()
|
H A D | request.c | 895 size_limit = min_t(unsigned int, BIO_MAX_VECS * PAGE_SECTORS, in cached_dev_cache_miss() 916 DIV_ROUND_UP(s->insert_bio_sectors, PAGE_SECTORS), in cached_dev_cache_miss()
|
H A D | btree.c | 545 return ilog2(KEY_SIZE(k) / PAGE_SECTORS ?: 1); in btree_order() 1131 SET_KEY_SIZE(&k.key, c->btree_pages * PAGE_SECTORS); in __bch_btree_node_alloc()
|
/openbmc/linux/include/linux/ |
H A D | blk_types.h | 37 #define PAGE_SECTORS (1 << PAGE_SECTORS_SHIFT) macro 38 #define SECTOR_MASK (PAGE_SECTORS - 1)
|
/openbmc/linux/drivers/md/ |
H A D | dm-stripe.c | 308 stripe_map_sector(sc, *pgoff * PAGE_SECTORS, &stripe, &dev_sector); in stripe_dax_pgoff()
|
H A D | dm.c | 1224 sector_t sector = pgoff * PAGE_SECTORS; in dm_dax_direct_access() 1235 len = max_io_len(ti, sector) / PAGE_SECTORS; in dm_dax_direct_access() 1251 sector_t sector = pgoff * PAGE_SECTORS; in dm_dax_zero_page_range() 1278 sector_t sector = pgoff * PAGE_SECTORS; in dm_dax_recovery_write()
|
H A D | dm-log-writes.c | 308 sector += bio_pages * PAGE_SECTORS; in write_inline_data()
|
/openbmc/linux/drivers/block/null_blk/ |
H A D | main.c | 1065 for (i = 0; i < PAGE_SECTORS; in null_flush_cache_page()
|