Home
last modified time | relevance | path

Searched refs:bdev_get_queue (Results 1 – 25 of 31) sorted by relevance

12

/openbmc/linux/include/linux/
H A Dblkdev.h871 static inline struct request_queue *bdev_get_queue(struct block_device *bdev) in bdev_get_queue() function
1149 return queue_max_zone_append_sectors(bdev_get_queue(bdev)); in bdev_max_zone_append_sectors()
1154 return queue_max_segments(bdev_get_queue(bdev)); in bdev_max_segments()
1169 return queue_logical_block_size(bdev_get_queue(bdev)); in bdev_logical_block_size()
1179 return queue_physical_block_size(bdev_get_queue(bdev)); in bdev_physical_block_size()
1189 return queue_io_min(bdev_get_queue(bdev)); in bdev_io_min()
1199 return queue_io_opt(bdev_get_queue(bdev)); in bdev_io_opt()
1211 return queue_zone_write_granularity(bdev_get_queue(bdev)); in bdev_zone_write_granularity()
1219 return bdev_get_queue(bdev)->limits.max_discard_sectors; in bdev_max_discard_sectors()
1224 return bdev_get_queue(bdev)->limits.discard_granularity; in bdev_discard_granularity()
[all …]
/openbmc/linux/block/
H A Dblk-crypto.c363 return __blk_crypto_cfg_supported(bdev_get_queue(bdev)->crypto_profile, in blk_crypto_config_supported_natively()
420 struct request_queue *q = bdev_get_queue(bdev); in blk_crypto_evict_key()
H A Dblk-core.c648 struct request_queue *q = bdev_get_queue(bio->bi_bdev); in __submit_bio_noacct()
666 if (q == bdev_get_queue(bio->bi_bdev)) in __submit_bio_noacct()
735 struct request_queue *q = bdev_get_queue(bdev); in submit_bio_noacct()
881 q = bdev_get_queue(bdev); in bio_poll()
H A Dblk-settings.c712 if (blk_stack_limits(&t->limits, &bdev_get_queue(bdev)->limits, in disk_stack_limits()
971 struct request_queue *q = bdev_get_queue(bdev); in bdev_alignment_offset()
984 struct request_queue *q = bdev_get_queue(bdev); in bdev_discard_alignment()
H A Dblk-rq-qos.h141 struct request_queue *q = bdev_get_queue(bio->bi_bdev); in rq_qos_done_bio()
H A Dbio.c1045 struct request_queue *q = bdev_get_queue(bio->bi_bdev); in bio_add_zone_append_page()
1176 struct request_queue *q = bdev_get_queue(bio->bi_bdev); in bio_iov_bvec_set()
1212 struct request_queue *q = bdev_get_queue(bio->bi_bdev); in bio_iov_add_zone_append_page()
1585 trace_block_bio_complete(bdev_get_queue(bio->bi_bdev), bio); in bio_endio()
H A Dbio-integrity.c126 struct request_queue *q = bdev_get_queue(bio->bi_bdev); in bio_integrity_add_page()
H A Dblk-cgroup.c962 spin_unlock_irq(&bdev_get_queue(ctx->bdev)->queue_lock); in blkg_conf_exit()
2094 blkg_get(bdev_get_queue(bio->bi_bdev)->root_blkg); in bio_associate_blkg_from_css()
2095 bio->bi_blkg = bdev_get_queue(bio->bi_bdev)->root_blkg; in bio_associate_blkg_from_css()
H A Dblk-zoned.c253 struct request_queue *q = bdev_get_queue(bdev); in blkdev_zone_mgmt()
H A Dblk.h66 struct request_queue *q = bdev_get_queue(bio->bi_bdev); in bio_queue_enter()
H A Dgenhd.c944 struct request_queue *q = bdev_get_queue(bdev); in part_stat_show()
993 struct request_queue *q = bdev_get_queue(bdev); in part_inflight_show()
H A Dioctl.c542 queue_max_sectors(bdev_get_queue(bdev))); in blkdev_common_ioctl()
H A Dblk-merge.c411 const struct queue_limits *lim = &bdev_get_queue(bio->bi_bdev)->limits; in bio_split_to_limits()
/openbmc/linux/drivers/md/
H A Ddm-table.c410 struct request_queue *q = bdev_get_queue(bdev); in dm_set_device_limits()
864 struct request_queue *q = bdev_get_queue(bdev); in device_is_rq_stackable()
1265 bdev_get_queue(dev->bdev)->crypto_profile; in device_intersect_crypto_capabilities()
1498 struct request_queue *q = bdev_get_queue(dev->bdev); in device_not_poll_capable()
1589 struct request_queue *q = bdev_get_queue(dev->bdev); in device_not_zoned_model()
1782 struct request_queue *q = bdev_get_queue(dev->bdev); in device_flush_capable()
1835 struct request_queue *q = bdev_get_queue(dev->bdev); in device_is_not_random()
1843 struct request_queue *q = bdev_get_queue(dev->bdev); in device_not_write_zeroes_capable()
H A Ddm-mpath.c539 q = bdev_get_queue(bdev); in multipath_clone_and_map()
885 struct request_queue *q = bdev_get_queue(bdev); in setup_scsi_dh()
961 q = bdev_get_queue(p->path.dev->bdev); in parse_path()
1626 struct request_queue *q = bdev_get_queue(pgpath->path.dev->bdev); in activate_or_offline_path()
2100 struct request_queue *q = bdev_get_queue(pgpath->path.dev->bdev); in pgpath_busy()
H A Ddm-io.c316 struct request_queue *q = bdev_get_queue(where->bdev); in do_region()
H A Ddm-clone-target.c2027 struct queue_limits *dest_limits = &bdev_get_queue(dest_dev)->limits; in disable_passdown_if_not_supported()
2048 struct queue_limits *dest_limits = &bdev_get_queue(dest_bdev)->limits; in set_discard_limits()
H A Ddm-zoned-target.c589 if (blk_queue_dying(bdev_get_queue(dmz_dev->bdev))) { in dmz_bdev_is_dying()
H A Ddm-cache-target.c3373 struct queue_limits *origin_limits = &bdev_get_queue(origin_bdev)->limits; in disable_passdown_if_not_supported()
3395 struct queue_limits *origin_limits = &bdev_get_queue(origin_bdev)->limits; in set_discard_limits()
/openbmc/linux/kernel/trace/
H A Dblktrace.c732 struct request_queue *q = bdev_get_queue(bdev); in blk_trace_ioctl()
1768 struct request_queue *q = bdev_get_queue(bdev); in sysfs_blk_trace_attr_show()
1802 struct request_queue *q = bdev_get_queue(bdev); in sysfs_blk_trace_attr_store()
/openbmc/linux/drivers/block/rnbd/
H A Drnbd-srv.c546 cpu_to_le32(queue_max_hw_sectors(bdev_get_queue(bdev))); in rnbd_srv_fill_msg_open_rsp()
/openbmc/linux/drivers/target/
H A Dtarget_core_iblock.c126 q = bdev_get_queue(bd); in iblock_configure_device()
/openbmc/linux/drivers/block/
H A Dpktcdvd.c717 struct request_queue *q = bdev_get_queue(pd->bdev); in pkt_generic_packet()
2185 q = bdev_get_queue(pd->bdev); in pkt_open_dev()
H A Dloop.c767 struct request_queue *backingq = bdev_get_queue(I_BDEV(inode)); in loop_config_discard()
/openbmc/linux/drivers/md/bcache/
H A Dsuper.c1018 q = bdev_get_queue(dc->bdev); in cached_dev_status_update()
1400 struct request_queue *q = bdev_get_queue(dc->bdev); in cached_dev_init()

12