xref: /openbmc/qemu/block/block-backend.c (revision 9c4218e9)
1 /*
2  * QEMU Block backends
3  *
4  * Copyright (C) 2014 Red Hat, Inc.
5  *
6  * Authors:
7  *  Markus Armbruster <armbru@redhat.com>,
8  *
9  * This work is licensed under the terms of the GNU LGPL, version 2.1
10  * or later.  See the COPYING.LIB file in the top-level directory.
11  */
12 
13 #include "qemu/osdep.h"
14 #include "sysemu/block-backend.h"
15 #include "block/block_int.h"
16 #include "block/blockjob.h"
17 #include "block/throttle-groups.h"
18 #include "sysemu/blockdev.h"
19 #include "sysemu/sysemu.h"
20 #include "qapi-event.h"
21 
22 /* Number of coroutines to reserve per attached device model */
23 #define COROUTINE_POOL_RESERVATION 64
24 
25 static AioContext *blk_aiocb_get_aio_context(BlockAIOCB *acb);
26 
27 struct BlockBackend {
28     char *name;
29     int refcnt;
30     BlockDriverState *bs;
31     DriveInfo *legacy_dinfo;    /* null unless created by drive_new() */
32     QTAILQ_ENTRY(BlockBackend) link; /* for blk_backends */
33 
34     void *dev;                  /* attached device model, if any */
35     /* TODO change to DeviceState when all users are qdevified */
36     const BlockDevOps *dev_ops;
37     void *dev_opaque;
38 
39     /* the block size for which the guest device expects atomicity */
40     int guest_block_size;
41 
42     /* If the BDS tree is removed, some of its options are stored here (which
43      * can be used to restore those options in the new BDS on insert) */
44     BlockBackendRootState root_state;
45 
46     /* I/O stats (display with "info blockstats"). */
47     BlockAcctStats stats;
48 
49     BlockdevOnError on_read_error, on_write_error;
50     bool iostatus_enabled;
51     BlockDeviceIoStatus iostatus;
52 
53     NotifierList remove_bs_notifiers, insert_bs_notifiers;
54 };
55 
56 typedef struct BlockBackendAIOCB {
57     BlockAIOCB common;
58     QEMUBH *bh;
59     BlockBackend *blk;
60     int ret;
61 } BlockBackendAIOCB;
62 
63 static const AIOCBInfo block_backend_aiocb_info = {
64     .get_aio_context = blk_aiocb_get_aio_context,
65     .aiocb_size = sizeof(BlockBackendAIOCB),
66 };
67 
68 static void drive_info_del(DriveInfo *dinfo);
69 
70 /* All the BlockBackends (except for hidden ones) */
71 static QTAILQ_HEAD(, BlockBackend) blk_backends =
72     QTAILQ_HEAD_INITIALIZER(blk_backends);
73 
74 /*
75  * Create a new BlockBackend with @name, with a reference count of one.
76  * @name must not be null or empty.
77  * Fail if a BlockBackend with this name already exists.
78  * Store an error through @errp on failure, unless it's null.
79  * Return the new BlockBackend on success, null on failure.
80  */
81 BlockBackend *blk_new(const char *name, Error **errp)
82 {
83     BlockBackend *blk;
84 
85     assert(name && name[0]);
86     if (!id_wellformed(name)) {
87         error_setg(errp, "Invalid device name");
88         return NULL;
89     }
90     if (blk_by_name(name)) {
91         error_setg(errp, "Device with id '%s' already exists", name);
92         return NULL;
93     }
94     if (bdrv_find_node(name)) {
95         error_setg(errp,
96                    "Device name '%s' conflicts with an existing node name",
97                    name);
98         return NULL;
99     }
100 
101     blk = g_new0(BlockBackend, 1);
102     blk->name = g_strdup(name);
103     blk->refcnt = 1;
104     notifier_list_init(&blk->remove_bs_notifiers);
105     notifier_list_init(&blk->insert_bs_notifiers);
106     QTAILQ_INSERT_TAIL(&blk_backends, blk, link);
107     return blk;
108 }
109 
110 /*
111  * Create a new BlockBackend with a new BlockDriverState attached.
112  * Otherwise just like blk_new(), which see.
113  */
114 BlockBackend *blk_new_with_bs(const char *name, Error **errp)
115 {
116     BlockBackend *blk;
117     BlockDriverState *bs;
118 
119     blk = blk_new(name, errp);
120     if (!blk) {
121         return NULL;
122     }
123 
124     bs = bdrv_new_root();
125     blk->bs = bs;
126     bs->blk = blk;
127     return blk;
128 }
129 
130 /*
131  * Calls blk_new_with_bs() and then calls bdrv_open() on the BlockDriverState.
132  *
133  * Just as with bdrv_open(), after having called this function the reference to
134  * @options belongs to the block layer (even on failure).
135  *
136  * TODO: Remove @filename and @flags; it should be possible to specify a whole
137  * BDS tree just by specifying the @options QDict (or @reference,
138  * alternatively). At the time of adding this function, this is not possible,
139  * though, so callers of this function have to be able to specify @filename and
140  * @flags.
141  */
142 BlockBackend *blk_new_open(const char *name, const char *filename,
143                            const char *reference, QDict *options, int flags,
144                            Error **errp)
145 {
146     BlockBackend *blk;
147     int ret;
148 
149     blk = blk_new_with_bs(name, errp);
150     if (!blk) {
151         QDECREF(options);
152         return NULL;
153     }
154 
155     ret = bdrv_open(&blk->bs, filename, reference, options, flags, errp);
156     if (ret < 0) {
157         blk_unref(blk);
158         return NULL;
159     }
160 
161     return blk;
162 }
163 
164 static void blk_delete(BlockBackend *blk)
165 {
166     assert(!blk->refcnt);
167     assert(!blk->dev);
168     if (blk->bs) {
169         blk_remove_bs(blk);
170     }
171     assert(QLIST_EMPTY(&blk->remove_bs_notifiers.notifiers));
172     assert(QLIST_EMPTY(&blk->insert_bs_notifiers.notifiers));
173     if (blk->root_state.throttle_state) {
174         g_free(blk->root_state.throttle_group);
175         throttle_group_unref(blk->root_state.throttle_state);
176     }
177     /* Avoid double-remove after blk_hide_on_behalf_of_hmp_drive_del() */
178     if (blk->name[0]) {
179         QTAILQ_REMOVE(&blk_backends, blk, link);
180     }
181     g_free(blk->name);
182     drive_info_del(blk->legacy_dinfo);
183     block_acct_cleanup(&blk->stats);
184     g_free(blk);
185 }
186 
187 static void drive_info_del(DriveInfo *dinfo)
188 {
189     if (!dinfo) {
190         return;
191     }
192     qemu_opts_del(dinfo->opts);
193     g_free(dinfo->serial);
194     g_free(dinfo);
195 }
196 
197 int blk_get_refcnt(BlockBackend *blk)
198 {
199     return blk ? blk->refcnt : 0;
200 }
201 
202 /*
203  * Increment @blk's reference count.
204  * @blk must not be null.
205  */
206 void blk_ref(BlockBackend *blk)
207 {
208     blk->refcnt++;
209 }
210 
211 /*
212  * Decrement @blk's reference count.
213  * If this drops it to zero, destroy @blk.
214  * For convenience, do nothing if @blk is null.
215  */
216 void blk_unref(BlockBackend *blk)
217 {
218     if (blk) {
219         assert(blk->refcnt > 0);
220         if (!--blk->refcnt) {
221             blk_delete(blk);
222         }
223     }
224 }
225 
226 /*
227  * Return the BlockBackend after @blk.
228  * If @blk is null, return the first one.
229  * Else, return @blk's next sibling, which may be null.
230  *
231  * To iterate over all BlockBackends, do
232  * for (blk = blk_next(NULL); blk; blk = blk_next(blk)) {
233  *     ...
234  * }
235  */
236 BlockBackend *blk_next(BlockBackend *blk)
237 {
238     return blk ? QTAILQ_NEXT(blk, link) : QTAILQ_FIRST(&blk_backends);
239 }
240 
241 /*
242  * Return @blk's name, a non-null string.
243  * Wart: the name is empty iff @blk has been hidden with
244  * blk_hide_on_behalf_of_hmp_drive_del().
245  */
246 const char *blk_name(BlockBackend *blk)
247 {
248     return blk->name;
249 }
250 
251 /*
252  * Return the BlockBackend with name @name if it exists, else null.
253  * @name must not be null.
254  */
255 BlockBackend *blk_by_name(const char *name)
256 {
257     BlockBackend *blk;
258 
259     assert(name);
260     QTAILQ_FOREACH(blk, &blk_backends, link) {
261         if (!strcmp(name, blk->name)) {
262             return blk;
263         }
264     }
265     return NULL;
266 }
267 
268 /*
269  * Return the BlockDriverState attached to @blk if any, else null.
270  */
271 BlockDriverState *blk_bs(BlockBackend *blk)
272 {
273     return blk->bs;
274 }
275 
276 /*
277  * Changes the BlockDriverState attached to @blk
278  */
279 void blk_set_bs(BlockBackend *blk, BlockDriverState *bs)
280 {
281     bdrv_ref(bs);
282 
283     if (blk->bs) {
284         blk->bs->blk = NULL;
285         bdrv_unref(blk->bs);
286     }
287     assert(bs->blk == NULL);
288 
289     blk->bs = bs;
290     bs->blk = blk;
291 }
292 
293 /*
294  * Return @blk's DriveInfo if any, else null.
295  */
296 DriveInfo *blk_legacy_dinfo(BlockBackend *blk)
297 {
298     return blk->legacy_dinfo;
299 }
300 
301 /*
302  * Set @blk's DriveInfo to @dinfo, and return it.
303  * @blk must not have a DriveInfo set already.
304  * No other BlockBackend may have the same DriveInfo set.
305  */
306 DriveInfo *blk_set_legacy_dinfo(BlockBackend *blk, DriveInfo *dinfo)
307 {
308     assert(!blk->legacy_dinfo);
309     return blk->legacy_dinfo = dinfo;
310 }
311 
312 /*
313  * Return the BlockBackend with DriveInfo @dinfo.
314  * It must exist.
315  */
316 BlockBackend *blk_by_legacy_dinfo(DriveInfo *dinfo)
317 {
318     BlockBackend *blk;
319 
320     QTAILQ_FOREACH(blk, &blk_backends, link) {
321         if (blk->legacy_dinfo == dinfo) {
322             return blk;
323         }
324     }
325     abort();
326 }
327 
328 /*
329  * Hide @blk.
330  * @blk must not have been hidden already.
331  * Make attached BlockDriverState, if any, anonymous.
332  * Once hidden, @blk is invisible to all functions that don't receive
333  * it as argument.  For example, blk_by_name() won't return it.
334  * Strictly for use by do_drive_del().
335  * TODO get rid of it!
336  */
337 void blk_hide_on_behalf_of_hmp_drive_del(BlockBackend *blk)
338 {
339     QTAILQ_REMOVE(&blk_backends, blk, link);
340     blk->name[0] = 0;
341     if (blk->bs) {
342         bdrv_make_anon(blk->bs);
343     }
344 }
345 
346 /*
347  * Disassociates the currently associated BlockDriverState from @blk.
348  */
349 void blk_remove_bs(BlockBackend *blk)
350 {
351     assert(blk->bs->blk == blk);
352 
353     notifier_list_notify(&blk->remove_bs_notifiers, blk);
354 
355     blk_update_root_state(blk);
356 
357     blk->bs->blk = NULL;
358     bdrv_unref(blk->bs);
359     blk->bs = NULL;
360 }
361 
362 /*
363  * Associates a new BlockDriverState with @blk.
364  */
365 void blk_insert_bs(BlockBackend *blk, BlockDriverState *bs)
366 {
367     assert(!blk->bs && !bs->blk);
368     bdrv_ref(bs);
369     blk->bs = bs;
370     bs->blk = blk;
371 
372     notifier_list_notify(&blk->insert_bs_notifiers, blk);
373 }
374 
375 /*
376  * Attach device model @dev to @blk.
377  * Return 0 on success, -EBUSY when a device model is attached already.
378  */
379 int blk_attach_dev(BlockBackend *blk, void *dev)
380 /* TODO change to DeviceState *dev when all users are qdevified */
381 {
382     if (blk->dev) {
383         return -EBUSY;
384     }
385     blk_ref(blk);
386     blk->dev = dev;
387     blk_iostatus_reset(blk);
388     return 0;
389 }
390 
391 /*
392  * Attach device model @dev to @blk.
393  * @blk must not have a device model attached already.
394  * TODO qdevified devices don't use this, remove when devices are qdevified
395  */
396 void blk_attach_dev_nofail(BlockBackend *blk, void *dev)
397 {
398     if (blk_attach_dev(blk, dev) < 0) {
399         abort();
400     }
401 }
402 
403 /*
404  * Detach device model @dev from @blk.
405  * @dev must be currently attached to @blk.
406  */
407 void blk_detach_dev(BlockBackend *blk, void *dev)
408 /* TODO change to DeviceState *dev when all users are qdevified */
409 {
410     assert(blk->dev == dev);
411     blk->dev = NULL;
412     blk->dev_ops = NULL;
413     blk->dev_opaque = NULL;
414     blk->guest_block_size = 512;
415     blk_unref(blk);
416 }
417 
418 /*
419  * Return the device model attached to @blk if any, else null.
420  */
421 void *blk_get_attached_dev(BlockBackend *blk)
422 /* TODO change to return DeviceState * when all users are qdevified */
423 {
424     return blk->dev;
425 }
426 
427 /*
428  * Set @blk's device model callbacks to @ops.
429  * @opaque is the opaque argument to pass to the callbacks.
430  * This is for use by device models.
431  */
432 void blk_set_dev_ops(BlockBackend *blk, const BlockDevOps *ops,
433                      void *opaque)
434 {
435     blk->dev_ops = ops;
436     blk->dev_opaque = opaque;
437 }
438 
439 /*
440  * Notify @blk's attached device model of media change.
441  * If @load is true, notify of media load.
442  * Else, notify of media eject.
443  * Also send DEVICE_TRAY_MOVED events as appropriate.
444  */
445 void blk_dev_change_media_cb(BlockBackend *blk, bool load)
446 {
447     if (blk->dev_ops && blk->dev_ops->change_media_cb) {
448         bool tray_was_open, tray_is_open;
449 
450         tray_was_open = blk_dev_is_tray_open(blk);
451         blk->dev_ops->change_media_cb(blk->dev_opaque, load);
452         tray_is_open = blk_dev_is_tray_open(blk);
453 
454         if (tray_was_open != tray_is_open) {
455             qapi_event_send_device_tray_moved(blk_name(blk), tray_is_open,
456                                               &error_abort);
457         }
458     }
459 }
460 
461 /*
462  * Does @blk's attached device model have removable media?
463  * %true if no device model is attached.
464  */
465 bool blk_dev_has_removable_media(BlockBackend *blk)
466 {
467     return !blk->dev || (blk->dev_ops && blk->dev_ops->change_media_cb);
468 }
469 
470 /*
471  * Does @blk's attached device model have a tray?
472  */
473 bool blk_dev_has_tray(BlockBackend *blk)
474 {
475     return blk->dev_ops && blk->dev_ops->is_tray_open;
476 }
477 
478 /*
479  * Notify @blk's attached device model of a media eject request.
480  * If @force is true, the medium is about to be yanked out forcefully.
481  */
482 void blk_dev_eject_request(BlockBackend *blk, bool force)
483 {
484     if (blk->dev_ops && blk->dev_ops->eject_request_cb) {
485         blk->dev_ops->eject_request_cb(blk->dev_opaque, force);
486     }
487 }
488 
489 /*
490  * Does @blk's attached device model have a tray, and is it open?
491  */
492 bool blk_dev_is_tray_open(BlockBackend *blk)
493 {
494     if (blk_dev_has_tray(blk)) {
495         return blk->dev_ops->is_tray_open(blk->dev_opaque);
496     }
497     return false;
498 }
499 
500 /*
501  * Does @blk's attached device model have the medium locked?
502  * %false if the device model has no such lock.
503  */
504 bool blk_dev_is_medium_locked(BlockBackend *blk)
505 {
506     if (blk->dev_ops && blk->dev_ops->is_medium_locked) {
507         return blk->dev_ops->is_medium_locked(blk->dev_opaque);
508     }
509     return false;
510 }
511 
512 /*
513  * Notify @blk's attached device model of a backend size change.
514  */
515 void blk_dev_resize_cb(BlockBackend *blk)
516 {
517     if (blk->dev_ops && blk->dev_ops->resize_cb) {
518         blk->dev_ops->resize_cb(blk->dev_opaque);
519     }
520 }
521 
522 void blk_iostatus_enable(BlockBackend *blk)
523 {
524     blk->iostatus_enabled = true;
525     blk->iostatus = BLOCK_DEVICE_IO_STATUS_OK;
526 }
527 
528 /* The I/O status is only enabled if the drive explicitly
529  * enables it _and_ the VM is configured to stop on errors */
530 bool blk_iostatus_is_enabled(const BlockBackend *blk)
531 {
532     return (blk->iostatus_enabled &&
533            (blk->on_write_error == BLOCKDEV_ON_ERROR_ENOSPC ||
534             blk->on_write_error == BLOCKDEV_ON_ERROR_STOP   ||
535             blk->on_read_error == BLOCKDEV_ON_ERROR_STOP));
536 }
537 
538 BlockDeviceIoStatus blk_iostatus(const BlockBackend *blk)
539 {
540     return blk->iostatus;
541 }
542 
543 void blk_iostatus_disable(BlockBackend *blk)
544 {
545     blk->iostatus_enabled = false;
546 }
547 
548 void blk_iostatus_reset(BlockBackend *blk)
549 {
550     if (blk_iostatus_is_enabled(blk)) {
551         blk->iostatus = BLOCK_DEVICE_IO_STATUS_OK;
552         if (blk->bs && blk->bs->job) {
553             block_job_iostatus_reset(blk->bs->job);
554         }
555     }
556 }
557 
558 void blk_iostatus_set_err(BlockBackend *blk, int error)
559 {
560     assert(blk_iostatus_is_enabled(blk));
561     if (blk->iostatus == BLOCK_DEVICE_IO_STATUS_OK) {
562         blk->iostatus = error == ENOSPC ? BLOCK_DEVICE_IO_STATUS_NOSPACE :
563                                           BLOCK_DEVICE_IO_STATUS_FAILED;
564     }
565 }
566 
567 static int blk_check_byte_request(BlockBackend *blk, int64_t offset,
568                                   size_t size)
569 {
570     int64_t len;
571 
572     if (size > INT_MAX) {
573         return -EIO;
574     }
575 
576     if (!blk_is_available(blk)) {
577         return -ENOMEDIUM;
578     }
579 
580     len = blk_getlength(blk);
581     if (len < 0) {
582         return len;
583     }
584 
585     if (offset < 0) {
586         return -EIO;
587     }
588 
589     if (offset > len || len - offset < size) {
590         return -EIO;
591     }
592 
593     return 0;
594 }
595 
596 static int blk_check_request(BlockBackend *blk, int64_t sector_num,
597                              int nb_sectors)
598 {
599     if (sector_num < 0 || sector_num > INT64_MAX / BDRV_SECTOR_SIZE) {
600         return -EIO;
601     }
602 
603     if (nb_sectors < 0 || nb_sectors > INT_MAX / BDRV_SECTOR_SIZE) {
604         return -EIO;
605     }
606 
607     return blk_check_byte_request(blk, sector_num * BDRV_SECTOR_SIZE,
608                                   nb_sectors * BDRV_SECTOR_SIZE);
609 }
610 
611 int blk_read(BlockBackend *blk, int64_t sector_num, uint8_t *buf,
612              int nb_sectors)
613 {
614     int ret = blk_check_request(blk, sector_num, nb_sectors);
615     if (ret < 0) {
616         return ret;
617     }
618 
619     return bdrv_read(blk->bs, sector_num, buf, nb_sectors);
620 }
621 
622 int blk_read_unthrottled(BlockBackend *blk, int64_t sector_num, uint8_t *buf,
623                          int nb_sectors)
624 {
625     int ret = blk_check_request(blk, sector_num, nb_sectors);
626     if (ret < 0) {
627         return ret;
628     }
629 
630     return bdrv_read_unthrottled(blk->bs, sector_num, buf, nb_sectors);
631 }
632 
633 int blk_write(BlockBackend *blk, int64_t sector_num, const uint8_t *buf,
634               int nb_sectors)
635 {
636     int ret = blk_check_request(blk, sector_num, nb_sectors);
637     if (ret < 0) {
638         return ret;
639     }
640 
641     return bdrv_write(blk->bs, sector_num, buf, nb_sectors);
642 }
643 
644 int blk_write_zeroes(BlockBackend *blk, int64_t sector_num,
645                      int nb_sectors, BdrvRequestFlags flags)
646 {
647     int ret = blk_check_request(blk, sector_num, nb_sectors);
648     if (ret < 0) {
649         return ret;
650     }
651 
652     return bdrv_write_zeroes(blk->bs, sector_num, nb_sectors, flags);
653 }
654 
655 static void error_callback_bh(void *opaque)
656 {
657     struct BlockBackendAIOCB *acb = opaque;
658     qemu_bh_delete(acb->bh);
659     acb->common.cb(acb->common.opaque, acb->ret);
660     qemu_aio_unref(acb);
661 }
662 
663 BlockAIOCB *blk_abort_aio_request(BlockBackend *blk,
664                                   BlockCompletionFunc *cb,
665                                   void *opaque, int ret)
666 {
667     struct BlockBackendAIOCB *acb;
668     QEMUBH *bh;
669 
670     acb = blk_aio_get(&block_backend_aiocb_info, blk, cb, opaque);
671     acb->blk = blk;
672     acb->ret = ret;
673 
674     bh = aio_bh_new(blk_get_aio_context(blk), error_callback_bh, acb);
675     acb->bh = bh;
676     qemu_bh_schedule(bh);
677 
678     return &acb->common;
679 }
680 
681 BlockAIOCB *blk_aio_write_zeroes(BlockBackend *blk, int64_t sector_num,
682                                  int nb_sectors, BdrvRequestFlags flags,
683                                  BlockCompletionFunc *cb, void *opaque)
684 {
685     int ret = blk_check_request(blk, sector_num, nb_sectors);
686     if (ret < 0) {
687         return blk_abort_aio_request(blk, cb, opaque, ret);
688     }
689 
690     return bdrv_aio_write_zeroes(blk->bs, sector_num, nb_sectors, flags,
691                                  cb, opaque);
692 }
693 
694 int blk_pread(BlockBackend *blk, int64_t offset, void *buf, int count)
695 {
696     int ret = blk_check_byte_request(blk, offset, count);
697     if (ret < 0) {
698         return ret;
699     }
700 
701     return bdrv_pread(blk->bs, offset, buf, count);
702 }
703 
704 int blk_pwrite(BlockBackend *blk, int64_t offset, const void *buf, int count)
705 {
706     int ret = blk_check_byte_request(blk, offset, count);
707     if (ret < 0) {
708         return ret;
709     }
710 
711     return bdrv_pwrite(blk->bs, offset, buf, count);
712 }
713 
714 int64_t blk_getlength(BlockBackend *blk)
715 {
716     if (!blk_is_available(blk)) {
717         return -ENOMEDIUM;
718     }
719 
720     return bdrv_getlength(blk->bs);
721 }
722 
723 void blk_get_geometry(BlockBackend *blk, uint64_t *nb_sectors_ptr)
724 {
725     if (!blk->bs) {
726         *nb_sectors_ptr = 0;
727     } else {
728         bdrv_get_geometry(blk->bs, nb_sectors_ptr);
729     }
730 }
731 
732 int64_t blk_nb_sectors(BlockBackend *blk)
733 {
734     if (!blk_is_available(blk)) {
735         return -ENOMEDIUM;
736     }
737 
738     return bdrv_nb_sectors(blk->bs);
739 }
740 
741 BlockAIOCB *blk_aio_readv(BlockBackend *blk, int64_t sector_num,
742                           QEMUIOVector *iov, int nb_sectors,
743                           BlockCompletionFunc *cb, void *opaque)
744 {
745     int ret = blk_check_request(blk, sector_num, nb_sectors);
746     if (ret < 0) {
747         return blk_abort_aio_request(blk, cb, opaque, ret);
748     }
749 
750     return bdrv_aio_readv(blk->bs, sector_num, iov, nb_sectors, cb, opaque);
751 }
752 
753 BlockAIOCB *blk_aio_writev(BlockBackend *blk, int64_t sector_num,
754                            QEMUIOVector *iov, int nb_sectors,
755                            BlockCompletionFunc *cb, void *opaque)
756 {
757     int ret = blk_check_request(blk, sector_num, nb_sectors);
758     if (ret < 0) {
759         return blk_abort_aio_request(blk, cb, opaque, ret);
760     }
761 
762     return bdrv_aio_writev(blk->bs, sector_num, iov, nb_sectors, cb, opaque);
763 }
764 
765 BlockAIOCB *blk_aio_flush(BlockBackend *blk,
766                           BlockCompletionFunc *cb, void *opaque)
767 {
768     if (!blk_is_available(blk)) {
769         return blk_abort_aio_request(blk, cb, opaque, -ENOMEDIUM);
770     }
771 
772     return bdrv_aio_flush(blk->bs, cb, opaque);
773 }
774 
775 BlockAIOCB *blk_aio_discard(BlockBackend *blk,
776                             int64_t sector_num, int nb_sectors,
777                             BlockCompletionFunc *cb, void *opaque)
778 {
779     int ret = blk_check_request(blk, sector_num, nb_sectors);
780     if (ret < 0) {
781         return blk_abort_aio_request(blk, cb, opaque, ret);
782     }
783 
784     return bdrv_aio_discard(blk->bs, sector_num, nb_sectors, cb, opaque);
785 }
786 
787 void blk_aio_cancel(BlockAIOCB *acb)
788 {
789     bdrv_aio_cancel(acb);
790 }
791 
792 void blk_aio_cancel_async(BlockAIOCB *acb)
793 {
794     bdrv_aio_cancel_async(acb);
795 }
796 
797 int blk_aio_multiwrite(BlockBackend *blk, BlockRequest *reqs, int num_reqs)
798 {
799     int i, ret;
800 
801     for (i = 0; i < num_reqs; i++) {
802         ret = blk_check_request(blk, reqs[i].sector, reqs[i].nb_sectors);
803         if (ret < 0) {
804             return ret;
805         }
806     }
807 
808     return bdrv_aio_multiwrite(blk->bs, reqs, num_reqs);
809 }
810 
811 int blk_ioctl(BlockBackend *blk, unsigned long int req, void *buf)
812 {
813     if (!blk_is_available(blk)) {
814         return -ENOMEDIUM;
815     }
816 
817     return bdrv_ioctl(blk->bs, req, buf);
818 }
819 
820 BlockAIOCB *blk_aio_ioctl(BlockBackend *blk, unsigned long int req, void *buf,
821                           BlockCompletionFunc *cb, void *opaque)
822 {
823     if (!blk_is_available(blk)) {
824         return blk_abort_aio_request(blk, cb, opaque, -ENOMEDIUM);
825     }
826 
827     return bdrv_aio_ioctl(blk->bs, req, buf, cb, opaque);
828 }
829 
830 int blk_co_discard(BlockBackend *blk, int64_t sector_num, int nb_sectors)
831 {
832     int ret = blk_check_request(blk, sector_num, nb_sectors);
833     if (ret < 0) {
834         return ret;
835     }
836 
837     return bdrv_co_discard(blk->bs, sector_num, nb_sectors);
838 }
839 
840 int blk_co_flush(BlockBackend *blk)
841 {
842     if (!blk_is_available(blk)) {
843         return -ENOMEDIUM;
844     }
845 
846     return bdrv_co_flush(blk->bs);
847 }
848 
849 int blk_flush(BlockBackend *blk)
850 {
851     if (!blk_is_available(blk)) {
852         return -ENOMEDIUM;
853     }
854 
855     return bdrv_flush(blk->bs);
856 }
857 
858 int blk_flush_all(void)
859 {
860     return bdrv_flush_all();
861 }
862 
863 void blk_drain(BlockBackend *blk)
864 {
865     if (blk->bs) {
866         bdrv_drain(blk->bs);
867     }
868 }
869 
870 void blk_drain_all(void)
871 {
872     bdrv_drain_all();
873 }
874 
875 void blk_set_on_error(BlockBackend *blk, BlockdevOnError on_read_error,
876                       BlockdevOnError on_write_error)
877 {
878     blk->on_read_error = on_read_error;
879     blk->on_write_error = on_write_error;
880 }
881 
882 BlockdevOnError blk_get_on_error(BlockBackend *blk, bool is_read)
883 {
884     return is_read ? blk->on_read_error : blk->on_write_error;
885 }
886 
887 BlockErrorAction blk_get_error_action(BlockBackend *blk, bool is_read,
888                                       int error)
889 {
890     BlockdevOnError on_err = blk_get_on_error(blk, is_read);
891 
892     switch (on_err) {
893     case BLOCKDEV_ON_ERROR_ENOSPC:
894         return (error == ENOSPC) ?
895                BLOCK_ERROR_ACTION_STOP : BLOCK_ERROR_ACTION_REPORT;
896     case BLOCKDEV_ON_ERROR_STOP:
897         return BLOCK_ERROR_ACTION_STOP;
898     case BLOCKDEV_ON_ERROR_REPORT:
899         return BLOCK_ERROR_ACTION_REPORT;
900     case BLOCKDEV_ON_ERROR_IGNORE:
901         return BLOCK_ERROR_ACTION_IGNORE;
902     default:
903         abort();
904     }
905 }
906 
907 static void send_qmp_error_event(BlockBackend *blk,
908                                  BlockErrorAction action,
909                                  bool is_read, int error)
910 {
911     IoOperationType optype;
912 
913     optype = is_read ? IO_OPERATION_TYPE_READ : IO_OPERATION_TYPE_WRITE;
914     qapi_event_send_block_io_error(blk_name(blk), optype, action,
915                                    blk_iostatus_is_enabled(blk),
916                                    error == ENOSPC, strerror(error),
917                                    &error_abort);
918 }
919 
920 /* This is done by device models because, while the block layer knows
921  * about the error, it does not know whether an operation comes from
922  * the device or the block layer (from a job, for example).
923  */
924 void blk_error_action(BlockBackend *blk, BlockErrorAction action,
925                       bool is_read, int error)
926 {
927     assert(error >= 0);
928 
929     if (action == BLOCK_ERROR_ACTION_STOP) {
930         /* First set the iostatus, so that "info block" returns an iostatus
931          * that matches the events raised so far (an additional error iostatus
932          * is fine, but not a lost one).
933          */
934         blk_iostatus_set_err(blk, error);
935 
936         /* Then raise the request to stop the VM and the event.
937          * qemu_system_vmstop_request_prepare has two effects.  First,
938          * it ensures that the STOP event always comes after the
939          * BLOCK_IO_ERROR event.  Second, it ensures that even if management
940          * can observe the STOP event and do a "cont" before the STOP
941          * event is issued, the VM will not stop.  In this case, vm_start()
942          * also ensures that the STOP/RESUME pair of events is emitted.
943          */
944         qemu_system_vmstop_request_prepare();
945         send_qmp_error_event(blk, action, is_read, error);
946         qemu_system_vmstop_request(RUN_STATE_IO_ERROR);
947     } else {
948         send_qmp_error_event(blk, action, is_read, error);
949     }
950 }
951 
952 int blk_is_read_only(BlockBackend *blk)
953 {
954     if (blk->bs) {
955         return bdrv_is_read_only(blk->bs);
956     } else {
957         return blk->root_state.read_only;
958     }
959 }
960 
961 int blk_is_sg(BlockBackend *blk)
962 {
963     if (!blk->bs) {
964         return 0;
965     }
966 
967     return bdrv_is_sg(blk->bs);
968 }
969 
970 int blk_enable_write_cache(BlockBackend *blk)
971 {
972     if (blk->bs) {
973         return bdrv_enable_write_cache(blk->bs);
974     } else {
975         return !!(blk->root_state.open_flags & BDRV_O_CACHE_WB);
976     }
977 }
978 
979 void blk_set_enable_write_cache(BlockBackend *blk, bool wce)
980 {
981     if (blk->bs) {
982         bdrv_set_enable_write_cache(blk->bs, wce);
983     } else {
984         if (wce) {
985             blk->root_state.open_flags |= BDRV_O_CACHE_WB;
986         } else {
987             blk->root_state.open_flags &= ~BDRV_O_CACHE_WB;
988         }
989     }
990 }
991 
992 void blk_invalidate_cache(BlockBackend *blk, Error **errp)
993 {
994     if (!blk->bs) {
995         error_setg(errp, "Device '%s' has no medium", blk->name);
996         return;
997     }
998 
999     bdrv_invalidate_cache(blk->bs, errp);
1000 }
1001 
1002 bool blk_is_inserted(BlockBackend *blk)
1003 {
1004     return blk->bs && bdrv_is_inserted(blk->bs);
1005 }
1006 
1007 bool blk_is_available(BlockBackend *blk)
1008 {
1009     return blk_is_inserted(blk) && !blk_dev_is_tray_open(blk);
1010 }
1011 
1012 void blk_lock_medium(BlockBackend *blk, bool locked)
1013 {
1014     if (blk->bs) {
1015         bdrv_lock_medium(blk->bs, locked);
1016     }
1017 }
1018 
1019 void blk_eject(BlockBackend *blk, bool eject_flag)
1020 {
1021     if (blk->bs) {
1022         bdrv_eject(blk->bs, eject_flag);
1023     }
1024 }
1025 
1026 int blk_get_flags(BlockBackend *blk)
1027 {
1028     if (blk->bs) {
1029         return bdrv_get_flags(blk->bs);
1030     } else {
1031         return blk->root_state.open_flags;
1032     }
1033 }
1034 
1035 int blk_get_max_transfer_length(BlockBackend *blk)
1036 {
1037     if (blk->bs) {
1038         return blk->bs->bl.max_transfer_length;
1039     } else {
1040         return 0;
1041     }
1042 }
1043 
1044 int blk_get_max_iov(BlockBackend *blk)
1045 {
1046     return blk->bs->bl.max_iov;
1047 }
1048 
1049 void blk_set_guest_block_size(BlockBackend *blk, int align)
1050 {
1051     blk->guest_block_size = align;
1052 }
1053 
1054 void *blk_try_blockalign(BlockBackend *blk, size_t size)
1055 {
1056     return qemu_try_blockalign(blk ? blk->bs : NULL, size);
1057 }
1058 
1059 void *blk_blockalign(BlockBackend *blk, size_t size)
1060 {
1061     return qemu_blockalign(blk ? blk->bs : NULL, size);
1062 }
1063 
1064 bool blk_op_is_blocked(BlockBackend *blk, BlockOpType op, Error **errp)
1065 {
1066     if (!blk->bs) {
1067         return false;
1068     }
1069 
1070     return bdrv_op_is_blocked(blk->bs, op, errp);
1071 }
1072 
1073 void blk_op_unblock(BlockBackend *blk, BlockOpType op, Error *reason)
1074 {
1075     if (blk->bs) {
1076         bdrv_op_unblock(blk->bs, op, reason);
1077     }
1078 }
1079 
1080 void blk_op_block_all(BlockBackend *blk, Error *reason)
1081 {
1082     if (blk->bs) {
1083         bdrv_op_block_all(blk->bs, reason);
1084     }
1085 }
1086 
1087 void blk_op_unblock_all(BlockBackend *blk, Error *reason)
1088 {
1089     if (blk->bs) {
1090         bdrv_op_unblock_all(blk->bs, reason);
1091     }
1092 }
1093 
1094 AioContext *blk_get_aio_context(BlockBackend *blk)
1095 {
1096     if (blk->bs) {
1097         return bdrv_get_aio_context(blk->bs);
1098     } else {
1099         return qemu_get_aio_context();
1100     }
1101 }
1102 
1103 static AioContext *blk_aiocb_get_aio_context(BlockAIOCB *acb)
1104 {
1105     BlockBackendAIOCB *blk_acb = DO_UPCAST(BlockBackendAIOCB, common, acb);
1106     return blk_get_aio_context(blk_acb->blk);
1107 }
1108 
1109 void blk_set_aio_context(BlockBackend *blk, AioContext *new_context)
1110 {
1111     if (blk->bs) {
1112         bdrv_set_aio_context(blk->bs, new_context);
1113     }
1114 }
1115 
1116 void blk_add_aio_context_notifier(BlockBackend *blk,
1117         void (*attached_aio_context)(AioContext *new_context, void *opaque),
1118         void (*detach_aio_context)(void *opaque), void *opaque)
1119 {
1120     if (blk->bs) {
1121         bdrv_add_aio_context_notifier(blk->bs, attached_aio_context,
1122                                       detach_aio_context, opaque);
1123     }
1124 }
1125 
1126 void blk_remove_aio_context_notifier(BlockBackend *blk,
1127                                      void (*attached_aio_context)(AioContext *,
1128                                                                   void *),
1129                                      void (*detach_aio_context)(void *),
1130                                      void *opaque)
1131 {
1132     if (blk->bs) {
1133         bdrv_remove_aio_context_notifier(blk->bs, attached_aio_context,
1134                                          detach_aio_context, opaque);
1135     }
1136 }
1137 
1138 void blk_add_remove_bs_notifier(BlockBackend *blk, Notifier *notify)
1139 {
1140     notifier_list_add(&blk->remove_bs_notifiers, notify);
1141 }
1142 
1143 void blk_add_insert_bs_notifier(BlockBackend *blk, Notifier *notify)
1144 {
1145     notifier_list_add(&blk->insert_bs_notifiers, notify);
1146 }
1147 
1148 void blk_io_plug(BlockBackend *blk)
1149 {
1150     if (blk->bs) {
1151         bdrv_io_plug(blk->bs);
1152     }
1153 }
1154 
1155 void blk_io_unplug(BlockBackend *blk)
1156 {
1157     if (blk->bs) {
1158         bdrv_io_unplug(blk->bs);
1159     }
1160 }
1161 
1162 BlockAcctStats *blk_get_stats(BlockBackend *blk)
1163 {
1164     return &blk->stats;
1165 }
1166 
1167 void *blk_aio_get(const AIOCBInfo *aiocb_info, BlockBackend *blk,
1168                   BlockCompletionFunc *cb, void *opaque)
1169 {
1170     return qemu_aio_get(aiocb_info, blk_bs(blk), cb, opaque);
1171 }
1172 
1173 int coroutine_fn blk_co_write_zeroes(BlockBackend *blk, int64_t sector_num,
1174                                      int nb_sectors, BdrvRequestFlags flags)
1175 {
1176     int ret = blk_check_request(blk, sector_num, nb_sectors);
1177     if (ret < 0) {
1178         return ret;
1179     }
1180 
1181     return bdrv_co_write_zeroes(blk->bs, sector_num, nb_sectors, flags);
1182 }
1183 
1184 int blk_write_compressed(BlockBackend *blk, int64_t sector_num,
1185                          const uint8_t *buf, int nb_sectors)
1186 {
1187     int ret = blk_check_request(blk, sector_num, nb_sectors);
1188     if (ret < 0) {
1189         return ret;
1190     }
1191 
1192     return bdrv_write_compressed(blk->bs, sector_num, buf, nb_sectors);
1193 }
1194 
1195 int blk_truncate(BlockBackend *blk, int64_t offset)
1196 {
1197     if (!blk_is_available(blk)) {
1198         return -ENOMEDIUM;
1199     }
1200 
1201     return bdrv_truncate(blk->bs, offset);
1202 }
1203 
1204 int blk_discard(BlockBackend *blk, int64_t sector_num, int nb_sectors)
1205 {
1206     int ret = blk_check_request(blk, sector_num, nb_sectors);
1207     if (ret < 0) {
1208         return ret;
1209     }
1210 
1211     return bdrv_discard(blk->bs, sector_num, nb_sectors);
1212 }
1213 
1214 int blk_save_vmstate(BlockBackend *blk, const uint8_t *buf,
1215                      int64_t pos, int size)
1216 {
1217     if (!blk_is_available(blk)) {
1218         return -ENOMEDIUM;
1219     }
1220 
1221     return bdrv_save_vmstate(blk->bs, buf, pos, size);
1222 }
1223 
1224 int blk_load_vmstate(BlockBackend *blk, uint8_t *buf, int64_t pos, int size)
1225 {
1226     if (!blk_is_available(blk)) {
1227         return -ENOMEDIUM;
1228     }
1229 
1230     return bdrv_load_vmstate(blk->bs, buf, pos, size);
1231 }
1232 
1233 int blk_probe_blocksizes(BlockBackend *blk, BlockSizes *bsz)
1234 {
1235     if (!blk_is_available(blk)) {
1236         return -ENOMEDIUM;
1237     }
1238 
1239     return bdrv_probe_blocksizes(blk->bs, bsz);
1240 }
1241 
1242 int blk_probe_geometry(BlockBackend *blk, HDGeometry *geo)
1243 {
1244     if (!blk_is_available(blk)) {
1245         return -ENOMEDIUM;
1246     }
1247 
1248     return bdrv_probe_geometry(blk->bs, geo);
1249 }
1250 
1251 /*
1252  * Updates the BlockBackendRootState object with data from the currently
1253  * attached BlockDriverState.
1254  */
1255 void blk_update_root_state(BlockBackend *blk)
1256 {
1257     assert(blk->bs);
1258 
1259     blk->root_state.open_flags    = blk->bs->open_flags;
1260     blk->root_state.read_only     = blk->bs->read_only;
1261     blk->root_state.detect_zeroes = blk->bs->detect_zeroes;
1262 
1263     if (blk->root_state.throttle_group) {
1264         g_free(blk->root_state.throttle_group);
1265         throttle_group_unref(blk->root_state.throttle_state);
1266     }
1267     if (blk->bs->throttle_state) {
1268         const char *name = throttle_group_get_name(blk->bs);
1269         blk->root_state.throttle_group = g_strdup(name);
1270         blk->root_state.throttle_state = throttle_group_incref(name);
1271     } else {
1272         blk->root_state.throttle_group = NULL;
1273         blk->root_state.throttle_state = NULL;
1274     }
1275 }
1276 
1277 /*
1278  * Applies the information in the root state to the given BlockDriverState. This
1279  * does not include the flags which have to be specified for bdrv_open(), use
1280  * blk_get_open_flags_from_root_state() to inquire them.
1281  */
1282 void blk_apply_root_state(BlockBackend *blk, BlockDriverState *bs)
1283 {
1284     bs->detect_zeroes = blk->root_state.detect_zeroes;
1285     if (blk->root_state.throttle_group) {
1286         bdrv_io_limits_enable(bs, blk->root_state.throttle_group);
1287     }
1288 }
1289 
1290 /*
1291  * Returns the flags to be used for bdrv_open() of a BlockDriverState which is
1292  * supposed to inherit the root state.
1293  */
1294 int blk_get_open_flags_from_root_state(BlockBackend *blk)
1295 {
1296     int bs_flags;
1297 
1298     bs_flags = blk->root_state.read_only ? 0 : BDRV_O_RDWR;
1299     bs_flags |= blk->root_state.open_flags & ~BDRV_O_RDWR;
1300 
1301     return bs_flags;
1302 }
1303 
1304 BlockBackendRootState *blk_get_root_state(BlockBackend *blk)
1305 {
1306     return &blk->root_state;
1307 }
1308