Lines Matching refs:blkif

173 	struct xen_blkif *blkif = ring->blkif;  in add_persistent_gnt()  local
176 if (!blkif->vbd.overflow_max_grants) in add_persistent_gnt()
177 blkif->vbd.overflow_max_grants = 1; in add_persistent_gnt()
347 !ring->blkif->vbd.overflow_max_grants)) { in purge_persistent_gnt()
400 ring->blkif->vbd.overflow_max_grants = 0; in purge_persistent_gnt()
449 static int xen_vbd_translate(struct phys_req *req, struct xen_blkif *blkif, in xen_vbd_translate() argument
452 struct xen_vbd *vbd = &blkif->vbd; in xen_vbd_translate()
475 static void xen_vbd_resize(struct xen_blkif *blkif) in xen_vbd_resize() argument
477 struct xen_vbd *vbd = &blkif->vbd; in xen_vbd_resize()
480 struct xenbus_device *dev = xen_blkbk_xenbus(blkif->be); in xen_vbd_resize()
484 blkif->domid, MAJOR(vbd->pdevice), MINOR(vbd->pdevice)); in xen_vbd_resize()
557 struct xen_blkif *blkif = ring->blkif; in xen_blkif_schedule() local
558 struct xen_vbd *vbd = &blkif->vbd; in xen_blkif_schedule()
569 xen_vbd_resize(blkif); in xen_blkif_schedule()
605 if (blkif->vbd.feature_gnt_persistent && in xen_blkif_schedule()
612 if (time_before(jiffies, blkif->buffer_squeeze_end)) in xen_blkif_schedule()
675 struct xen_blkif *blkif = ring->blkif; in xen_blkbk_unmap_and_respond_callback() local
697 if (atomic_dec_and_test(&ring->inflight) && atomic_read(&blkif->drain)) { in xen_blkbk_unmap_and_respond_callback()
698 complete(&blkif->drain_complete); in xen_blkbk_unmap_and_respond_callback()
700 xen_blkif_put(blkif); in xen_blkbk_unmap_and_respond_callback()
769 struct xen_blkif *blkif = ring->blkif; in xen_blkbk_map() local
771 use_persistent_gnts = (blkif->vbd.feature_gnt_persistent); in xen_blkbk_map()
812 blkif->domid); in xen_blkbk_map()
874 if (use_persistent_gnts && !blkif->vbd.overflow_max_grants) { in xen_blkbk_map()
875 blkif->vbd.overflow_max_grants = 1; in xen_blkbk_map()
877 blkif->domid, blkif->vbd.handle); in xen_blkbk_map()
971 struct xen_blkif *blkif = ring->blkif; in dispatch_discard_io() local
972 struct block_device *bdev = blkif->vbd.bdev; in dispatch_discard_io()
975 xen_blkif_get(blkif); in dispatch_discard_io()
980 err = xen_vbd_translate(&preq, blkif, REQ_OP_WRITE); in dispatch_discard_io()
984 preq.sector_number + preq.nr_sects, blkif->vbd.pdevice); in dispatch_discard_io()
989 if (blkif->vbd.discard_secure && in dispatch_discard_io()
1006 xen_blkif_put(blkif); in dispatch_discard_io()
1022 struct xen_blkif *blkif = ring->blkif; in xen_blk_drain_io() local
1024 atomic_set(&blkif->drain, 1); in xen_blk_drain_io()
1029 &blkif->drain_complete, HZ); in xen_blk_drain_io()
1031 if (!atomic_read(&blkif->drain)) in xen_blk_drain_io()
1034 atomic_set(&blkif->drain, 0); in xen_blk_drain_io()
1044 xen_blkbk_flush_diskcache(XBT_NIL, pending_req->ring->blkif->be, 0); in __end_block_io_op()
1049 xen_blkbk_barrier(XBT_NIL, pending_req->ring->blkif->be, 0); in __end_block_io_op()
1202 rp, rc, rp - rc, ring->blkif->vbd.pdevice); in __do_block_io_op()
1225 switch (ring->blkif->blk_protocol) { in __do_block_io_op()
1384 if (xen_vbd_translate(&preq, ring->blkif, operation) != 0) { in dispatch_rw_block_io()
1389 ring->blkif->vbd.pdevice); in dispatch_rw_block_io()
1401 ring->blkif->domid); in dispatch_rw_block_io()
1425 xen_blkif_get(ring->blkif); in dispatch_rw_block_io()
1500 switch (ring->blkif->blk_protocol) { in make_response()