/openbmc/linux/fs/btrfs/ |
H A D | bio.c | 336 else if (!(bio->bi_opf & REQ_RAHEAD)) in btrfs_log_dev_io_error() 338 if (bio->bi_opf & REQ_PREFLUSH) in btrfs_log_dev_io_error() 345 if (bio->bi_opf & REQ_META) in btrfs_end_io_wq() 462 __func__, bio_op(bio), bio->bi_opf, bio->bi_iter.bi_sector, in btrfs_submit_dev_bio() 468 if (bio->bi_opf & REQ_BTRFS_CGROUP_PUNT) in btrfs_submit_dev_bio() 528 if (bbio->bio.bi_opf & REQ_META) in btrfs_bio_csum() 589 bio->bi_opf |= REQ_BTRFS_CGROUP_PUNT; in run_one_async_done() 608 if (op_is_sync(bbio->bio.bi_opf)) in should_async_write() 612 if ((bbio->bio.bi_opf & REQ_META) && btrfs_is_zoned(bbio->fs_info)) in should_async_write() 689 bio->bi_opf &= ~REQ_OP_WRITE; in btrfs_submit_chunk() [all …]
|
/openbmc/linux/block/ |
H A D | blk-merge.c | 324 if (bio->bi_opf & REQ_NOWAIT) { in bio_split_rw() 387 split->bi_opf |= REQ_NOMERGE; in __bio_split_to_limits() 755 WARN_ON_ONCE((bio->bi_opf & REQ_FAILFAST_MASK) && in blk_rq_set_mixed_merge() 756 (bio->bi_opf & REQ_FAILFAST_MASK) != ff); in blk_rq_set_mixed_merge() 757 bio->bi_opf |= ff; in blk_rq_set_mixed_merge() 764 if (bio->bi_opf & REQ_RAHEAD) in bio_failfast() 767 return bio->bi_opf & REQ_FAILFAST_MASK; in bio_failfast() 779 if (bio->bi_opf & REQ_RAHEAD) in blk_update_mixed_merge() 780 bio->bi_opf |= REQ_FAILFAST_MASK; in blk_update_mixed_merge() 784 req->cmd_flags |= bio->bi_opf & REQ_FAILFAST_MASK; in blk_update_mixed_merge()
|
H A D | blk-core.c | 344 if (bio->bi_opf & REQ_NOWAIT) { in __bio_queue_enter() 505 if (op_is_flush(bio->bi_opf) && !bio_sectors(bio)) in bio_check_ro() 544 current->comm, bio->bi_bdev, bio->bi_opf, in bio_check_eod() 599 bio->bi_opf |= REQ_NOMERGE; in blk_check_zone_append() 744 if ((bio->bi_opf & REQ_NOWAIT) && !bdev_nowait(bdev)) in submit_bio_noacct() 761 if (op_is_flush(bio->bi_opf)) { in submit_bio_noacct() 766 bio->bi_opf &= ~(REQ_PREFLUSH | REQ_FUA); in submit_bio_noacct() 1211 sizeof_field(struct bio, bi_opf)); in blk_dev_init()
|
H A D | blk-throttle.h | 191 blkg_rwstat_add(&tg->stat_bytes, bio->bi_opf, in blk_should_throtl() 194 blkg_rwstat_add(&tg->stat_ios, bio->bi_opf, 1); in blk_should_throtl()
|
H A D | bio.c | 251 bio->bi_opf = opf; in bio_init() 308 bio->bi_opf = opf; in bio_reset() 773 if ((bio->bi_opf & REQ_POLLED) && !WARN_ON_ONCE(in_interrupt())) { in bio_put_percpu_cache() 805 if (bio->bi_opf & REQ_ALLOC_CACHE) in bio_put() 850 bio = bio_alloc_bioset(bdev, 0, bio_src->bi_opf, gfp, bs); in bio_alloc_clone() 881 bio_init(bio, bdev, bio_src->bi_io_vec, 0, bio_src->bi_opf); in bio_init_clone() 1374 bio->bi_opf |= REQ_SYNC; in submit_bio_wait()
|
H A D | blk-mq-sched.h | 33 return !(bio->bi_opf & REQ_NOMERGE_FLAGS); in bio_mergeable()
|
H A D | fops.c | 87 bio.bi_opf |= REQ_NOWAIT; in __blkdev_direct_IO_simple() 232 bio->bi_opf |= REQ_NOWAIT; in __blkdev_direct_IO() 354 bio->bi_opf |= REQ_NOWAIT; in __blkdev_direct_IO_async() 357 bio->bi_opf |= REQ_POLLED; in __blkdev_direct_IO_async()
|
/openbmc/linux/drivers/md/bcache/ |
H A D | writeback.h | 123 return (op_is_sync(bio->bi_opf) || in should_writeback() 124 bio->bi_opf & (REQ_META|REQ_PRIO) || in should_writeback()
|
H A D | request.c | 202 bio->bi_opf &= ~(REQ_PREFLUSH|REQ_FUA); in bch_data_insert_start() 247 n->bi_opf = REQ_OP_WRITE; in bch_data_insert_start() 391 if ((bio->bi_opf & (REQ_RAHEAD|REQ_BACKGROUND))) { in check_should_bypass() 392 if (!(bio->bi_opf & (REQ_META|REQ_PRIO)) && in check_should_bypass() 653 bio->bi_opf & REQ_PREFLUSH)) { in backing_request_endio() 744 s->iop.flush_journal = op_is_flush(bio->bi_opf); in search_alloc() 1019 if (bio->bi_opf & REQ_PREFLUSH) { in cached_dev_write()
|
/openbmc/linux/drivers/md/ |
H A D | md-multipath.c | 86 else if (!(bio->bi_opf & REQ_RAHEAD)) { in multipath_end_request() 106 if (unlikely(bio->bi_opf & REQ_PREFLUSH) in multipath_make_request() 127 mp_bh->bio.bi_opf |= REQ_FAILFAST_TRANSPORT; in multipath_make_request() 320 bio->bi_opf |= REQ_FAILFAST_TRANSPORT; in multipathd()
|
H A D | dm-raid1.c | 268 .bi_opf = REQ_OP_WRITE | REQ_PREFLUSH | REQ_SYNC, in mirror_flush() 547 .bi_opf = REQ_OP_READ, in read_async_bio() 659 blk_opf_t op_flags = bio->bi_opf & (REQ_FUA | REQ_PREFLUSH); in do_write() 661 .bi_opf = REQ_OP_WRITE | op_flags, in do_write() 670 io_req.bi_opf = REQ_OP_DISCARD | op_flags; in do_write() 708 if ((bio->bi_opf & REQ_PREFLUSH) || in do_writes() 1222 if (bio->bi_opf & REQ_RAHEAD) in mirror_map() 1259 if (!(bio->bi_opf & REQ_PREFLUSH) && in mirror_end_io() 1268 if (bio->bi_opf & REQ_RAHEAD) in mirror_end_io()
|
H A D | dm-zone.c | 135 return !op_is_flush(bio->bi_opf) && bio_sectors(bio); in dm_is_zone_write() 401 clone->bi_opf = REQ_OP_WRITE | REQ_NOMERGE | in dm_zone_map_bio_begin() 402 (clone->bi_opf & (~REQ_OP_MASK)); in dm_zone_map_bio_begin() 495 if (op_is_flush(bio->bi_opf) && !bio_sectors(bio)) in dm_need_zone_wp_tracking()
|
H A D | raid1.c | 467 (bio->bi_opf & MD_FAILFAST) && in raid1_end_write_request() 1218 const blk_opf_t do_sync = bio->bi_opf & REQ_SYNC; in raid1_read_request() 1248 bio->bi_opf & REQ_NOWAIT)) { in raid1_read_request() 1318 read_bio->bi_opf = op | do_sync; in raid1_read_request() 1321 read_bio->bi_opf |= MD_FAILFAST; in raid1_read_request() 1349 if (bio->bi_opf & REQ_NOWAIT) { in raid1_write_request() 1371 bio->bi_opf & REQ_NOWAIT)) { in raid1_write_request() 1461 if (bio->bi_opf & REQ_NOWAIT) { in raid1_write_request() 1540 mbio->bi_opf = bio_op(bio) | (bio->bi_opf & (REQ_SYNC | REQ_FUA)); in raid1_write_request() 1544 mbio->bi_opf |= MD_FAILFAST; in raid1_write_request() [all …]
|
H A D | dm-zero.c | 42 if (bio->bi_opf & REQ_RAHEAD) { in zero_map()
|
H A D | dm.c | 492 return ((bio->bi_opf & REQ_PREFLUSH) && bio->bi_iter.bi_size); in bio_is_flush_with_data() 631 clone->bi_opf &= ~REQ_DM_POLL_LIST; in alloc_tio() 887 (bio->bi_opf & REQ_POLLED)); in dm_handle_requeue() 894 if (bio->bi_opf & REQ_POLLED) { in dm_handle_requeue() 968 bio->bi_opf &= ~REQ_PREFLUSH; in __dm_io_complete() 1089 return unlikely((bio->bi_opf & REQ_SWAP) != 0) && unlikely(ti->limit_swap_bios); in swap_bios_limit() 1673 if (!(bio->bi_opf & REQ_DM_POLL_LIST)) { in dm_queue_poll_io() 1674 bio->bi_opf |= REQ_DM_POLL_LIST; in dm_queue_poll_io() 1710 if (unlikely((ci->bio->bi_opf & REQ_NOWAIT) != 0) && in __split_and_process_bio() 1721 ci->submit_as_polled = !!(ci->bio->bi_opf & REQ_POLLED); in __split_and_process_bio() [all …]
|
H A D | raid10.c | 487 (bio->bi_opf & MD_FAILFAST)) { in raid10_end_write_request() 1144 if (!wait_barrier(conf, bio->bi_opf & REQ_NOWAIT)) { in regular_request_wait() 1152 if (bio->bi_opf & REQ_NOWAIT) { in regular_request_wait() 1172 const blk_opf_t do_sync = bio->bi_opf & REQ_SYNC; in raid10_read_request() 1250 read_bio->bi_opf = op | do_sync; in raid10_read_request() 1253 read_bio->bi_opf |= MD_FAILFAST; in raid10_read_request() 1268 const blk_opf_t do_sync = bio->bi_opf & REQ_SYNC; in raid10_write_one_disk() 1269 const blk_opf_t do_fua = bio->bi_opf & REQ_FUA; in raid10_write_one_disk() 1295 mbio->bi_opf = op | do_sync | do_fua; in raid10_write_one_disk() 1299 mbio->bi_opf |= MD_FAILFAST; in raid10_write_one_disk() [all …]
|
H A D | dm-io.c | 507 if ((io_req->bi_opf & REQ_OP_MASK) == REQ_OP_READ) { in dp_init() 538 io_req->bi_opf, &dp, sync_error_bits, ioprio); in dm_io() 541 io_req->bi_opf, &dp, io_req->notify.fn, in dm_io()
|
H A D | dm-flakey.c | 22 (((bio)->bi_opf & (fc)->corrupt_bio_flags) == (fc)->corrupt_bio_flags) 359 (bio_data_dir(bio) == WRITE) ? 'w' : 'r', bio->bi_opf, in corrupt_bio_common() 429 bio_init(clone, fc->dev->bdev, bio->bi_inline_vecs, nr_iovecs, bio->bi_opf); in clone_bio()
|
/openbmc/linux/include/trace/events/ |
H A D | bcache.h | 31 blk_fill_rwbs(__entry->rwbs, bio->bi_opf); 105 blk_fill_rwbs(__entry->rwbs, bio->bi_opf); 140 blk_fill_rwbs(__entry->rwbs, bio->bi_opf); 171 blk_fill_rwbs(__entry->rwbs, bio->bi_opf); 241 blk_fill_rwbs(__entry->rwbs, bio->bi_opf);
|
H A D | block.h | 303 blk_fill_rwbs(__entry->rwbs, bio->bi_opf); 330 blk_fill_rwbs(__entry->rwbs, bio->bi_opf); 487 blk_fill_rwbs(__entry->rwbs, bio->bi_opf); 528 blk_fill_rwbs(__entry->rwbs, bio->bi_opf);
|
/openbmc/linux/include/linux/ |
H A D | dm-io.h | 62 blk_opf_t bi_opf; /* Request type and flags */ member
|
H A D | bio.h | 356 if (bio->bi_opf & REQ_INTEGRITY) in bio_integrity() 806 bio->bi_opf |= REQ_POLLED; in bio_set_polled() 808 bio->bi_opf |= REQ_NOWAIT; in bio_set_polled() 813 bio->bi_opf &= ~REQ_POLLED; in bio_clear_polled()
|
H A D | blk_types.h | 268 blk_opf_t bi_opf; /* bottom bits REQ_OP, top bits member 477 return bio->bi_opf & REQ_OP_MASK; in bio_op()
|
/openbmc/linux/fs/xfs/ |
H A D | xfs_bio_io.c | 42 prev->bi_opf, GFP_KERNEL); in xfs_rw_bdev()
|
/openbmc/linux/drivers/block/ |
H A D | brd.c | 259 bio->bi_opf, sector); in brd_submit_bio() 261 if (err == -ENOMEM && bio->bi_opf & REQ_NOWAIT) { in brd_submit_bio()
|