Home
last modified time | relevance | path

Searched refs:mq_hctx (Results 1 – 15 of 15) sorted by relevance

/openbmc/linux/block/
H A Dblk-mq-sched.c53 return rqa->mq_hctx > rqb->mq_hctx; in sched_rq_cmp()
59 list_first_entry(rq_list, struct request, queuelist)->mq_hctx; in blk_mq_dispatch_hctx_list()
65 if (rq->mq_hctx != hctx) { in blk_mq_dispatch_hctx_list()
141 if (rq->mq_hctx != hctx) in __blk_mq_do_dispatch_sched()
264 } while (blk_mq_dispatch_rq_list(rq->mq_hctx, &rq_list, 1)); in blk_mq_do_dispatch_ctx()
H A Dblk-mq.c359 rq->mq_hctx = hctx; in blk_mq_rq_ctx_init()
565 if (blk_mq_get_hctx_type(opf) != rq->mq_hctx->type) in blk_mq_alloc_cached_request()
705 struct blk_mq_hw_ctx *hctx = rq->mq_hctx; in __blk_mq_free_request()
710 rq->mq_hctx = NULL; in __blk_mq_free_request()
1116 if (nr_tags == TAG_COMP_BATCH || cur_hctx != rq->mq_hctx) { in blk_mq_end_request_batch()
1120 cur_hctx = rq->mq_hctx; in blk_mq_end_request_batch()
1210 if ((rq->mq_hctx->nr_ctx == 1 && in blk_mq_complete_request_remote()
1273 WRITE_ONCE(rq->bio->bi_cookie, rq->mq_hctx->queue_num); in blk_mq_start_request()
1330 struct blk_mq_hw_ctx *hctx = rq->mq_hctx; in blk_execute_rq_nowait()
1368 if (!rq->mq_hctx) in blk_rq_is_poll()
[all …]
H A Dblk-flush.c325 flush_rq->mq_hctx = first_rq->mq_hctx; in blk_kick_flush()
363 struct blk_mq_hw_ctx *hctx = rq->mq_hctx; in mq_flush_data_end_io()
H A Dblk-mq.h332 __blk_mq_put_driver_tag(rq->mq_hctx, rq); in blk_mq_put_driver_tag()
339 struct blk_mq_hw_ctx *hctx = rq->mq_hctx; in blk_mq_get_driver_tag()
H A Dblk-mq-tag.c292 if (rq->q == q && (!hctx || rq->mq_hctx == hctx)) in bt_iter()
681 return (rq->mq_hctx->queue_num << BLK_MQ_UNIQUE_TAG_BITS) | in blk_mq_unique_tag()
H A Dmq-deadline.c957 if (dd_has_write_work(rq->mq_hctx)) in dd_finish_request()
958 blk_mq_sched_mark_restart_hctx(rq->mq_hctx); in dd_finish_request()
H A Dblk-mq-debugfs.c358 if (rq->mq_hctx == params->hctx) in hctx_show_busy_rq()
/openbmc/linux/drivers/block/
H A Dvirtio_blk.c339 struct virtio_blk *vblk = req->mq_hctx->queue->queuedata; in virtblk_request_done()
470 struct virtio_blk *vblk = req->mq_hctx->queue->queuedata; in virtblk_prep_rq_batch()
473 req->mq_hctx->tags->rqs[req->tag] = req; in virtblk_prep_rq_batch()
475 return virtblk_prep_rq(req->mq_hctx, vblk, req, vbr) == BLK_STS_OK; in virtblk_prep_rq_batch()
515 struct virtio_blk_vq *this_vq = get_virtio_blk_vq(req->mq_hctx); in virtio_queue_rqs()
H A Dublk_drv.c1056 struct ublk_queue *ubq = req->mq_hctx->driver_data; in __ublk_complete_rq()
1171 struct ublk_queue *ubq = req->mq_hctx->driver_data; in __ublk_rq_task_work()
1309 struct ublk_queue *ubq = rq->mq_hctx->driver_data; in ublk_timeout()
1930 if (!req->mq_hctx || !req->mq_hctx->driver_data) in ublk_check_and_get_req()
1955 ubq = req->mq_hctx->driver_data; in ublk_ch_read_iter()
1973 ubq = req->mq_hctx->driver_data; in ublk_ch_write_iter()
/openbmc/linux/drivers/nvme/host/
H A Dpci.c511 struct nvme_queue *nvmeq = req->mq_hctx->driver_data; in nvme_pci_use_sgls()
778 struct nvme_queue *nvmeq = req->mq_hctx->driver_data; in nvme_map_data()
931 req->mq_hctx->tags->rqs[req->tag] = req; in nvme_prep_rq_batch()
944 if (nvmeq && nvmeq != req->mq_hctx->driver_data) in nvme_queue_rqs()
946 nvmeq = req->mq_hctx->driver_data; in nvme_queue_rqs()
961 struct nvme_queue *nvmeq = req->mq_hctx->driver_data; in nvme_pci_unmap_rq()
1218 struct nvme_queue *nvmeq = req->mq_hctx->driver_data; in abort_endio()
1281 struct nvme_queue *nvmeq = req->mq_hctx->driver_data; in nvme_timeout()
1309 nvme_poll(req->mq_hctx, NULL); in nvme_timeout()
H A Dnvme.h221 return req->mq_hctx->queue_num + 1; in nvme_req_qid()
H A Dcore.c666 if (req->mq_hctx->type == HCTX_TYPE_POLL) in nvme_init_request()
/openbmc/linux/drivers/ufs/core/
H A Dufs-mcq.c114 struct blk_mq_hw_ctx *hctx = READ_ONCE(req->mq_hctx); in ufshcd_mcq_req_to_hwq()
/openbmc/linux/include/linux/
H A Dblk-mq.h83 struct blk_mq_hw_ctx *mq_hctx; member
/openbmc/linux/drivers/block/null_blk/
H A Dmain.c1690 struct blk_mq_hw_ctx *hctx = rq->mq_hctx; in null_timeout_rq()