/openbmc/linux/drivers/nvme/target/ |
H A D | zns.c | 541 const blk_opf_t opf = REQ_OP_ZONE_APPEND | REQ_SYNC | REQ_IDLE; in nvmet_bdev_execute_zone_append() local 572 ARRAY_SIZE(req->inline_bvec), opf); in nvmet_bdev_execute_zone_append() 574 bio = bio_alloc(req->ns->bdev, req->sg_cnt, opf, GFP_KERNEL); in nvmet_bdev_execute_zone_append()
|
/openbmc/linux/fs/iomap/ |
H A D | direct-io.c | 56 struct iomap_dio *dio, unsigned short nr_vecs, blk_opf_t opf) in iomap_dio_alloc_bio() argument 59 return bio_alloc_bioset(iter->iomap.bdev, nr_vecs, opf, in iomap_dio_alloc_bio() 61 return bio_alloc(iter->iomap.bdev, nr_vecs, opf, GFP_KERNEL); in iomap_dio_alloc_bio()
|
/openbmc/linux/block/ |
H A D | kyber-iosched.c | 196 static unsigned int kyber_sched_domain(blk_opf_t opf) in kyber_sched_domain() argument 198 switch (opf & REQ_OP_MASK) { in kyber_sched_domain() 554 static void kyber_limit_depth(blk_opf_t opf, struct blk_mq_alloc_data *data) in kyber_limit_depth() argument 560 if (!op_is_sync(opf)) { in kyber_limit_depth()
|
H A D | blk-mq.c | 520 blk_opf_t opf, in blk_mq_rq_cache_fill() argument 526 .cmd_flags = opf, in blk_mq_rq_cache_fill() 544 blk_opf_t opf, in blk_mq_alloc_cached_request() argument 556 rq = blk_mq_rq_cache_fill(q, plug, opf, flags); in blk_mq_alloc_cached_request() 564 if (blk_mq_get_hctx_type(opf) != rq->mq_hctx->type) in blk_mq_alloc_cached_request() 566 if (op_is_flush(rq->cmd_flags) != op_is_flush(opf)) in blk_mq_alloc_cached_request() 573 rq->cmd_flags = opf; in blk_mq_alloc_cached_request() 583 rq = blk_mq_alloc_cached_request(q, opf, flags); in blk_mq_alloc_request() 588 .cmd_flags = opf, in blk_mq_alloc_request() 612 blk_opf_t opf, blk_mq_req_flags_t flags, unsigned int hctx_idx) in blk_mq_alloc_request_hctx() argument [all …]
|
H A D | mq-deadline.c | 642 static void dd_limit_depth(blk_opf_t opf, struct blk_mq_alloc_data *data) in dd_limit_depth() argument 647 if (op_is_sync(opf) && !op_is_write(opf)) in dd_limit_depth()
|
H A D | bfq-iosched.c | 690 static void bfq_limit_depth(blk_opf_t opf, struct blk_mq_alloc_data *data) in bfq_limit_depth() argument 699 if (op_is_sync(opf) && !op_is_write(opf)) { in bfq_limit_depth() 702 depth = bfqd->word_depths[!!bfqd->wr_busy_queues][op_is_sync(opf)]; in bfq_limit_depth() 708 bic_to_bfqq(bic, op_is_sync(opf), act_idx); in bfq_limit_depth() 722 __func__, bfqd->wr_busy_queues, op_is_sync(opf), depth); in bfq_limit_depth()
|
/openbmc/linux/fs/gfs2/ |
H A D | meta_io.c | 220 static void gfs2_submit_bhs(blk_opf_t opf, struct buffer_head *bhs[], int num) in gfs2_submit_bhs() argument 226 bio = bio_alloc(bh->b_bdev, num, opf, GFP_NOIO); in gfs2_submit_bhs()
|
H A D | lops.c | 241 void gfs2_log_submit_bio(struct bio **biop, blk_opf_t opf) in gfs2_log_submit_bio() argument 247 bio->bi_opf = opf; in gfs2_log_submit_bio()
|
/openbmc/linux/drivers/scsi/device_handler/ |
H A D | scsi_dh_alua.c | 130 blk_opf_t opf = REQ_OP_DRV_IN | REQ_FAILFAST_DEV | in submit_rtpg() local 145 return scsi_execute_cmd(sdev, cdb, opf, buff, bufflen, in submit_rtpg() 163 blk_opf_t opf = REQ_OP_DRV_OUT | REQ_FAILFAST_DEV | in submit_stpg() local 180 return scsi_execute_cmd(sdev, cdb, opf, stpg_data, in submit_stpg()
|
H A D | scsi_dh_rdac.c | 539 blk_opf_t opf = REQ_OP_DRV_OUT | REQ_FAILFAST_DEV | in send_mode_select() local 561 if (scsi_execute_cmd(sdev, cdb, opf, &h->ctlr->mode_select, data_size, in send_mode_select()
|
/openbmc/linux/fs/ |
H A D | buffer.c | 57 static void submit_bh_wbc(blk_opf_t opf, struct buffer_head *bh, 2786 static void submit_bh_wbc(blk_opf_t opf, struct buffer_head *bh, in submit_bh_wbc() argument 2789 const enum req_op op = opf & REQ_OP_MASK; in submit_bh_wbc() 2805 opf |= REQ_META; in submit_bh_wbc() 2807 opf |= REQ_PRIO; in submit_bh_wbc() 2809 bio = bio_alloc(bh->b_bdev, 1, opf, GFP_NOIO); in submit_bh_wbc() 2831 void submit_bh(blk_opf_t opf, struct buffer_head *bh) in submit_bh() argument 2833 submit_bh_wbc(opf, bh, NULL); in submit_bh()
|
/openbmc/linux/include/scsi/ |
H A D | scsi_cmnd.h | 395 struct request *scsi_alloc_request(struct request_queue *q, blk_opf_t opf,
|
H A D | scsi_device.h | 501 blk_opf_t opf, void *buffer, unsigned int bufflen,
|
/openbmc/linux/include/linux/ |
H A D | blk-mq.h | 726 struct request *blk_mq_alloc_request(struct request_queue *q, blk_opf_t opf, 729 blk_opf_t opf, blk_mq_req_flags_t flags,
|
/openbmc/linux/drivers/md/bcache/ |
H A D | super.c | 417 static void uuid_io(struct cache_set *c, blk_opf_t opf, struct bkey *k, in uuid_io() argument 432 bio->bi_opf = opf | REQ_SYNC | REQ_META; in uuid_io() 441 if ((opf & REQ_OP_MASK) != REQ_OP_WRITE) in uuid_io() 446 pr_debug("%s UUIDs at %s\n", (opf & REQ_OP_MASK) == REQ_OP_WRITE ? in uuid_io() 590 static void prio_io(struct cache *ca, uint64_t bucket, blk_opf_t opf) in prio_io() argument 603 bio->bi_opf = opf | REQ_SYNC | REQ_META; in prio_io()
|
/openbmc/linux/fs/btrfs/ |
H A D | bio.c | 62 struct btrfs_bio *btrfs_bio_alloc(unsigned int nr_vecs, blk_opf_t opf, in btrfs_bio_alloc() argument 69 bio = bio_alloc_bioset(NULL, nr_vecs, opf, GFP_NOFS, &btrfs_bioset); in btrfs_bio_alloc()
|
H A D | extent_io.c | 103 blk_opf_t opf; member 740 bbio = btrfs_bio_alloc(BIO_MAX_VECS, bio_ctrl->opf, fs_info, in alloc_new_bio() 1120 struct btrfs_bio_ctrl bio_ctrl = { .opf = REQ_OP_READ }; in btrfs_read_folio() 2159 .opf = REQ_OP_WRITE | wbc_to_write_flags(wbc), in extent_write_locked_range() 2163 bio_ctrl.opf |= REQ_BTRFS_CGROUP_PUNT; in extent_write_locked_range() 2211 .opf = REQ_OP_WRITE | wbc_to_write_flags(wbc), in extent_writepages() 2227 struct btrfs_bio_ctrl bio_ctrl = { .opf = REQ_OP_READ | REQ_RAHEAD }; in extent_readahead()
|
/openbmc/linux/drivers/scsi/ |
H A D | scsi_lib.c | 202 blk_opf_t opf, void *buffer, unsigned int bufflen, in scsi_execute_cmd() argument 217 req = scsi_alloc_request(sdev->request_queue, opf, args->req_flags); in scsi_execute_cmd() 1131 struct request *scsi_alloc_request(struct request_queue *q, blk_opf_t opf, in scsi_alloc_request() argument 1136 rq = blk_mq_alloc_request(q, opf, flags); in scsi_alloc_request()
|
H A D | scsi_transport_spi.c | 113 blk_opf_t opf = op | REQ_FAILFAST_DEV | REQ_FAILFAST_TRANSPORT | in spi_execute() local 127 result = scsi_execute_cmd(sdev, cmd, opf, buffer, bufflen, in spi_execute()
|
/openbmc/linux/fs/hfsplus/ |
H A D | hfsplus_fs.h | 530 void **data, blk_opf_t opf);
|
/openbmc/linux/kernel/power/ |
H A D | swap.c | 272 static int hib_submit_io(blk_opf_t opf, pgoff_t page_off, void *addr, in hib_submit_io() argument 279 bio = bio_alloc(hib_resume_bdev, 1, opf, GFP_NOIO | __GFP_HIGH); in hib_submit_io()
|
/openbmc/linux/drivers/md/ |
H A D | md.h | 787 struct page *page, blk_opf_t opf, bool metadata_op);
|
H A D | dm-integrity.c | 543 static int sync_rw_sb(struct dm_integrity_c *ic, blk_opf_t opf) in sync_rw_sb() argument 547 const enum req_op op = opf & REQ_OP_MASK; in sync_rw_sb() 550 io_req.bi_opf = opf; in sync_rw_sb() 1050 static void rw_journal_sectors(struct dm_integrity_c *ic, blk_opf_t opf, in rw_journal_sectors() argument 1068 io_req.bi_opf = opf; in rw_journal_sectors() 1088 dm_integrity_io_error(ic, (opf & REQ_OP_MASK) == REQ_OP_READ ? in rw_journal_sectors() 1097 static void rw_journal(struct dm_integrity_c *ic, blk_opf_t opf, in rw_journal() argument 1106 rw_journal_sectors(ic, opf, sector, n_sectors, comp); in rw_journal()
|
/openbmc/linux/drivers/block/drbd/ |
H A D | drbd_int.h | 326 blk_opf_t opf; member 340 ((peer_req)->opf & REQ_OP_MASK)
|
/openbmc/linux/include/trace/events/ |
H A D | btrfs.h | 2455 __field( u8, opf ) 2466 __entry->opf = bio_op(bio); 2489 __entry->offset, __entry->opf, __entry->physical, __entry->len)
|