1 /* 2 * QEMU System Emulator block driver 3 * 4 * Copyright (c) 2003 Fabrice Bellard 5 * 6 * Permission is hereby granted, free of charge, to any person obtaining a copy 7 * of this software and associated documentation files (the "Software"), to deal 8 * in the Software without restriction, including without limitation the rights 9 * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell 10 * copies of the Software, and to permit persons to whom the Software is 11 * furnished to do so, subject to the following conditions: 12 * 13 * The above copyright notice and this permission notice shall be included in 14 * all copies or substantial portions of the Software. 15 * 16 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR 17 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, 18 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL 19 * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER 20 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, 21 * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN 22 * THE SOFTWARE. 23 */ 24 #ifndef BLOCK_INT_H 25 #define BLOCK_INT_H 26 27 #include "block/accounting.h" 28 #include "block/block.h" 29 #include "qemu/option.h" 30 #include "qemu/queue.h" 31 #include "qemu/coroutine.h" 32 #include "qemu/timer.h" 33 #include "qapi-types.h" 34 #include "qemu/hbitmap.h" 35 #include "block/snapshot.h" 36 #include "qemu/main-loop.h" 37 #include "qemu/throttle.h" 38 39 #define BLOCK_FLAG_ENCRYPT 1 40 #define BLOCK_FLAG_COMPAT6 4 41 #define BLOCK_FLAG_LAZY_REFCOUNTS 8 42 43 #define BLOCK_OPT_SIZE "size" 44 #define BLOCK_OPT_ENCRYPT "encryption" 45 #define BLOCK_OPT_COMPAT6 "compat6" 46 #define BLOCK_OPT_BACKING_FILE "backing_file" 47 #define BLOCK_OPT_BACKING_FMT "backing_fmt" 48 #define BLOCK_OPT_CLUSTER_SIZE "cluster_size" 49 #define BLOCK_OPT_TABLE_SIZE "table_size" 50 #define BLOCK_OPT_PREALLOC "preallocation" 51 #define BLOCK_OPT_SUBFMT "subformat" 52 #define BLOCK_OPT_COMPAT_LEVEL "compat" 53 #define BLOCK_OPT_LAZY_REFCOUNTS "lazy_refcounts" 54 #define BLOCK_OPT_ADAPTER_TYPE "adapter_type" 55 #define BLOCK_OPT_REDUNDANCY "redundancy" 56 #define BLOCK_OPT_NOCOW "nocow" 57 #define BLOCK_OPT_OBJECT_SIZE "object_size" 58 #define BLOCK_OPT_REFCOUNT_BITS "refcount_bits" 59 60 #define BLOCK_PROBE_BUF_SIZE 512 61 62 typedef struct BdrvTrackedRequest { 63 BlockDriverState *bs; 64 int64_t offset; 65 unsigned int bytes; 66 bool is_write; 67 68 bool serialising; 69 int64_t overlap_offset; 70 unsigned int overlap_bytes; 71 72 QLIST_ENTRY(BdrvTrackedRequest) list; 73 Coroutine *co; /* owner, used for deadlock detection */ 74 CoQueue wait_queue; /* coroutines blocked on this request */ 75 76 struct BdrvTrackedRequest *waiting_for; 77 } BdrvTrackedRequest; 78 79 struct BlockDriver { 80 const char *format_name; 81 int instance_size; 82 83 /* set to true if the BlockDriver is a block filter */ 84 bool is_filter; 85 /* for snapshots block filter like Quorum can implement the 86 * following recursive callback. 87 * It's purpose is to recurse on the filter children while calling 88 * bdrv_recurse_is_first_non_filter on them. 89 * For a sample implementation look in the future Quorum block filter. 90 */ 91 bool (*bdrv_recurse_is_first_non_filter)(BlockDriverState *bs, 92 BlockDriverState *candidate); 93 94 int (*bdrv_probe)(const uint8_t *buf, int buf_size, const char *filename); 95 int (*bdrv_probe_device)(const char *filename); 96 97 /* Any driver implementing this callback is expected to be able to handle 98 * NULL file names in its .bdrv_open() implementation */ 99 void (*bdrv_parse_filename)(const char *filename, QDict *options, Error **errp); 100 /* Drivers not implementing bdrv_parse_filename nor bdrv_open should have 101 * this field set to true, except ones that are defined only by their 102 * child's bs. 103 * An example of the last type will be the quorum block driver. 104 */ 105 bool bdrv_needs_filename; 106 107 /* Set if a driver can support backing files */ 108 bool supports_backing; 109 110 /* For handling image reopen for split or non-split files */ 111 int (*bdrv_reopen_prepare)(BDRVReopenState *reopen_state, 112 BlockReopenQueue *queue, Error **errp); 113 void (*bdrv_reopen_commit)(BDRVReopenState *reopen_state); 114 void (*bdrv_reopen_abort)(BDRVReopenState *reopen_state); 115 116 int (*bdrv_open)(BlockDriverState *bs, QDict *options, int flags, 117 Error **errp); 118 int (*bdrv_file_open)(BlockDriverState *bs, QDict *options, int flags, 119 Error **errp); 120 int (*bdrv_read)(BlockDriverState *bs, int64_t sector_num, 121 uint8_t *buf, int nb_sectors); 122 int (*bdrv_write)(BlockDriverState *bs, int64_t sector_num, 123 const uint8_t *buf, int nb_sectors); 124 void (*bdrv_close)(BlockDriverState *bs); 125 int (*bdrv_create)(const char *filename, QemuOpts *opts, Error **errp); 126 int (*bdrv_set_key)(BlockDriverState *bs, const char *key); 127 int (*bdrv_make_empty)(BlockDriverState *bs); 128 129 void (*bdrv_refresh_filename)(BlockDriverState *bs); 130 131 /* aio */ 132 BlockAIOCB *(*bdrv_aio_readv)(BlockDriverState *bs, 133 int64_t sector_num, QEMUIOVector *qiov, int nb_sectors, 134 BlockCompletionFunc *cb, void *opaque); 135 BlockAIOCB *(*bdrv_aio_writev)(BlockDriverState *bs, 136 int64_t sector_num, QEMUIOVector *qiov, int nb_sectors, 137 BlockCompletionFunc *cb, void *opaque); 138 BlockAIOCB *(*bdrv_aio_flush)(BlockDriverState *bs, 139 BlockCompletionFunc *cb, void *opaque); 140 BlockAIOCB *(*bdrv_aio_discard)(BlockDriverState *bs, 141 int64_t sector_num, int nb_sectors, 142 BlockCompletionFunc *cb, void *opaque); 143 144 int coroutine_fn (*bdrv_co_readv)(BlockDriverState *bs, 145 int64_t sector_num, int nb_sectors, QEMUIOVector *qiov); 146 int coroutine_fn (*bdrv_co_writev)(BlockDriverState *bs, 147 int64_t sector_num, int nb_sectors, QEMUIOVector *qiov); 148 /* 149 * Efficiently zero a region of the disk image. Typically an image format 150 * would use a compact metadata representation to implement this. This 151 * function pointer may be NULL and .bdrv_co_writev() will be called 152 * instead. 153 */ 154 int coroutine_fn (*bdrv_co_write_zeroes)(BlockDriverState *bs, 155 int64_t sector_num, int nb_sectors, BdrvRequestFlags flags); 156 int coroutine_fn (*bdrv_co_discard)(BlockDriverState *bs, 157 int64_t sector_num, int nb_sectors); 158 int64_t coroutine_fn (*bdrv_co_get_block_status)(BlockDriverState *bs, 159 int64_t sector_num, int nb_sectors, int *pnum); 160 161 /* 162 * Invalidate any cached meta-data. 163 */ 164 void (*bdrv_invalidate_cache)(BlockDriverState *bs, Error **errp); 165 166 /* 167 * Flushes all data that was already written to the OS all the way down to 168 * the disk (for example raw-posix calls fsync()). 169 */ 170 int coroutine_fn (*bdrv_co_flush_to_disk)(BlockDriverState *bs); 171 172 /* 173 * Flushes all internal caches to the OS. The data may still sit in a 174 * writeback cache of the host OS, but it will survive a crash of the qemu 175 * process. 176 */ 177 int coroutine_fn (*bdrv_co_flush_to_os)(BlockDriverState *bs); 178 179 const char *protocol_name; 180 int (*bdrv_truncate)(BlockDriverState *bs, int64_t offset); 181 182 int64_t (*bdrv_getlength)(BlockDriverState *bs); 183 bool has_variable_length; 184 int64_t (*bdrv_get_allocated_file_size)(BlockDriverState *bs); 185 186 int (*bdrv_write_compressed)(BlockDriverState *bs, int64_t sector_num, 187 const uint8_t *buf, int nb_sectors); 188 189 int (*bdrv_snapshot_create)(BlockDriverState *bs, 190 QEMUSnapshotInfo *sn_info); 191 int (*bdrv_snapshot_goto)(BlockDriverState *bs, 192 const char *snapshot_id); 193 int (*bdrv_snapshot_delete)(BlockDriverState *bs, 194 const char *snapshot_id, 195 const char *name, 196 Error **errp); 197 int (*bdrv_snapshot_list)(BlockDriverState *bs, 198 QEMUSnapshotInfo **psn_info); 199 int (*bdrv_snapshot_load_tmp)(BlockDriverState *bs, 200 const char *snapshot_id, 201 const char *name, 202 Error **errp); 203 int (*bdrv_get_info)(BlockDriverState *bs, BlockDriverInfo *bdi); 204 ImageInfoSpecific *(*bdrv_get_specific_info)(BlockDriverState *bs); 205 206 int (*bdrv_save_vmstate)(BlockDriverState *bs, QEMUIOVector *qiov, 207 int64_t pos); 208 int (*bdrv_load_vmstate)(BlockDriverState *bs, uint8_t *buf, 209 int64_t pos, int size); 210 211 int (*bdrv_change_backing_file)(BlockDriverState *bs, 212 const char *backing_file, const char *backing_fmt); 213 214 /* removable device specific */ 215 int (*bdrv_is_inserted)(BlockDriverState *bs); 216 int (*bdrv_media_changed)(BlockDriverState *bs); 217 void (*bdrv_eject)(BlockDriverState *bs, bool eject_flag); 218 void (*bdrv_lock_medium)(BlockDriverState *bs, bool locked); 219 220 /* to control generic scsi devices */ 221 int (*bdrv_ioctl)(BlockDriverState *bs, unsigned long int req, void *buf); 222 BlockAIOCB *(*bdrv_aio_ioctl)(BlockDriverState *bs, 223 unsigned long int req, void *buf, 224 BlockCompletionFunc *cb, void *opaque); 225 226 /* List of options for creating images, terminated by name == NULL */ 227 QemuOptsList *create_opts; 228 229 /* 230 * Returns 0 for completed check, -errno for internal errors. 231 * The check results are stored in result. 232 */ 233 int (*bdrv_check)(BlockDriverState* bs, BdrvCheckResult *result, 234 BdrvCheckMode fix); 235 236 int (*bdrv_amend_options)(BlockDriverState *bs, QemuOpts *opts, 237 BlockDriverAmendStatusCB *status_cb); 238 239 void (*bdrv_debug_event)(BlockDriverState *bs, BlkDebugEvent event); 240 241 /* TODO Better pass a option string/QDict/QemuOpts to add any rule? */ 242 int (*bdrv_debug_breakpoint)(BlockDriverState *bs, const char *event, 243 const char *tag); 244 int (*bdrv_debug_remove_breakpoint)(BlockDriverState *bs, 245 const char *tag); 246 int (*bdrv_debug_resume)(BlockDriverState *bs, const char *tag); 247 bool (*bdrv_debug_is_suspended)(BlockDriverState *bs, const char *tag); 248 249 void (*bdrv_refresh_limits)(BlockDriverState *bs, Error **errp); 250 251 /* 252 * Returns 1 if newly created images are guaranteed to contain only 253 * zeros, 0 otherwise. 254 */ 255 int (*bdrv_has_zero_init)(BlockDriverState *bs); 256 257 /* Remove fd handlers, timers, and other event loop callbacks so the event 258 * loop is no longer in use. Called with no in-flight requests and in 259 * depth-first traversal order with parents before child nodes. 260 */ 261 void (*bdrv_detach_aio_context)(BlockDriverState *bs); 262 263 /* Add fd handlers, timers, and other event loop callbacks so I/O requests 264 * can be processed again. Called with no in-flight requests and in 265 * depth-first traversal order with child nodes before parent nodes. 266 */ 267 void (*bdrv_attach_aio_context)(BlockDriverState *bs, 268 AioContext *new_context); 269 270 /* io queue for linux-aio */ 271 void (*bdrv_io_plug)(BlockDriverState *bs); 272 void (*bdrv_io_unplug)(BlockDriverState *bs); 273 void (*bdrv_flush_io_queue)(BlockDriverState *bs); 274 275 /** 276 * Try to get @bs's logical and physical block size. 277 * On success, store them in @bsz and return zero. 278 * On failure, return negative errno. 279 */ 280 int (*bdrv_probe_blocksizes)(BlockDriverState *bs, BlockSizes *bsz); 281 /** 282 * Try to get @bs's geometry (cyls, heads, sectors) 283 * On success, store them in @geo and return 0. 284 * On failure return -errno. 285 * Only drivers that want to override guest geometry implement this 286 * callback; see hd_geometry_guess(). 287 */ 288 int (*bdrv_probe_geometry)(BlockDriverState *bs, HDGeometry *geo); 289 290 QLIST_ENTRY(BlockDriver) list; 291 }; 292 293 typedef struct BlockLimits { 294 /* maximum number of sectors that can be discarded at once */ 295 int max_discard; 296 297 /* optimal alignment for discard requests in sectors */ 298 int64_t discard_alignment; 299 300 /* maximum number of sectors that can zeroized at once */ 301 int max_write_zeroes; 302 303 /* optimal alignment for write zeroes requests in sectors */ 304 int64_t write_zeroes_alignment; 305 306 /* optimal transfer length in sectors */ 307 int opt_transfer_length; 308 309 /* maximal transfer length in sectors */ 310 int max_transfer_length; 311 312 /* memory alignment so that no bounce buffer is needed */ 313 size_t min_mem_alignment; 314 315 /* memory alignment for bounce buffer */ 316 size_t opt_mem_alignment; 317 } BlockLimits; 318 319 typedef struct BdrvOpBlocker BdrvOpBlocker; 320 321 typedef struct BdrvAioNotifier { 322 void (*attached_aio_context)(AioContext *new_context, void *opaque); 323 void (*detach_aio_context)(void *opaque); 324 325 void *opaque; 326 327 QLIST_ENTRY(BdrvAioNotifier) list; 328 } BdrvAioNotifier; 329 330 struct BdrvChildRole { 331 int (*inherit_flags)(int parent_flags); 332 }; 333 334 extern const BdrvChildRole child_file; 335 extern const BdrvChildRole child_format; 336 337 struct BdrvChild { 338 BlockDriverState *bs; 339 const BdrvChildRole *role; 340 QLIST_ENTRY(BdrvChild) next; 341 QLIST_ENTRY(BdrvChild) next_parent; 342 }; 343 344 /* 345 * Note: the function bdrv_append() copies and swaps contents of 346 * BlockDriverStates, so if you add new fields to this struct, please 347 * inspect bdrv_append() to determine if the new fields need to be 348 * copied as well. 349 */ 350 struct BlockDriverState { 351 int64_t total_sectors; /* if we are reading a disk image, give its 352 size in sectors */ 353 int read_only; /* if true, the media is read only */ 354 int open_flags; /* flags used to open the file, re-used for re-open */ 355 int encrypted; /* if true, the media is encrypted */ 356 int valid_key; /* if true, a valid encryption key has been set */ 357 int sg; /* if true, the device is a /dev/sg* */ 358 int copy_on_read; /* if true, copy read backing sectors into image 359 note this is a reference count */ 360 bool probed; 361 362 BlockDriver *drv; /* NULL means no media */ 363 void *opaque; 364 365 BlockBackend *blk; /* owning backend, if any */ 366 367 AioContext *aio_context; /* event loop used for fd handlers, timers, etc */ 368 /* long-running tasks intended to always use the same AioContext as this 369 * BDS may register themselves in this list to be notified of changes 370 * regarding this BDS's context */ 371 QLIST_HEAD(, BdrvAioNotifier) aio_notifiers; 372 373 char filename[PATH_MAX]; 374 char backing_file[PATH_MAX]; /* if non zero, the image is a diff of 375 this file image */ 376 char backing_format[16]; /* if non-zero and backing_file exists */ 377 378 QDict *full_open_options; 379 char exact_filename[PATH_MAX]; 380 381 BdrvChild *backing; 382 BdrvChild *file; 383 384 NotifierList close_notifiers; 385 386 /* Callback before write request is processed */ 387 NotifierWithReturnList before_write_notifiers; 388 389 /* number of in-flight serialising requests */ 390 unsigned int serialising_in_flight; 391 392 /* I/O throttling */ 393 CoQueue throttled_reqs[2]; 394 bool io_limits_enabled; 395 /* The following fields are protected by the ThrottleGroup lock. 396 * See the ThrottleGroup documentation for details. */ 397 ThrottleState *throttle_state; 398 ThrottleTimers throttle_timers; 399 unsigned pending_reqs[2]; 400 QLIST_ENTRY(BlockDriverState) round_robin; 401 402 /* I/O stats (display with "info blockstats"). */ 403 BlockAcctStats stats; 404 405 /* I/O Limits */ 406 BlockLimits bl; 407 408 /* Whether produces zeros when read beyond eof */ 409 bool zero_beyond_eof; 410 411 /* Alignment requirement for offset/length of I/O requests */ 412 unsigned int request_alignment; 413 414 /* the block size for which the guest device expects atomicity */ 415 int guest_block_size; 416 417 /* do we need to tell the quest if we have a volatile write cache? */ 418 int enable_write_cache; 419 420 /* NOTE: the following infos are only hints for real hardware 421 drivers. They are not used by the block driver */ 422 BlockdevOnError on_read_error, on_write_error; 423 bool iostatus_enabled; 424 BlockDeviceIoStatus iostatus; 425 426 /* the following member gives a name to every node on the bs graph. */ 427 char node_name[32]; 428 /* element of the list of named nodes building the graph */ 429 QTAILQ_ENTRY(BlockDriverState) node_list; 430 /* element of the list of "drives" the guest sees */ 431 QTAILQ_ENTRY(BlockDriverState) device_list; 432 QLIST_HEAD(, BdrvDirtyBitmap) dirty_bitmaps; 433 int refcnt; 434 435 QLIST_HEAD(, BdrvTrackedRequest) tracked_requests; 436 437 /* operation blockers */ 438 QLIST_HEAD(, BdrvOpBlocker) op_blockers[BLOCK_OP_TYPE_MAX]; 439 440 /* long-running background operation */ 441 BlockJob *job; 442 443 /* The node that this node inherited default options from (and a reopen on 444 * which can affect this node by changing these defaults). This is always a 445 * parent node of this node. */ 446 BlockDriverState *inherits_from; 447 QLIST_HEAD(, BdrvChild) children; 448 QLIST_HEAD(, BdrvChild) parents; 449 450 QDict *options; 451 BlockdevDetectZeroesOptions detect_zeroes; 452 453 /* The error object in use for blocking operations on backing_hd */ 454 Error *backing_blocker; 455 456 /* threshold limit for writes, in bytes. "High water mark". */ 457 uint64_t write_threshold_offset; 458 NotifierWithReturn write_threshold_notifier; 459 }; 460 461 static inline BlockDriverState *backing_bs(BlockDriverState *bs) 462 { 463 return bs->backing ? bs->backing->bs : NULL; 464 } 465 466 467 /* Essential block drivers which must always be statically linked into qemu, and 468 * which therefore can be accessed without using bdrv_find_format() */ 469 extern BlockDriver bdrv_file; 470 extern BlockDriver bdrv_raw; 471 extern BlockDriver bdrv_qcow2; 472 473 /** 474 * bdrv_setup_io_funcs: 475 * 476 * Prepare a #BlockDriver for I/O request processing by populating 477 * unimplemented coroutine and AIO interfaces with generic wrapper functions 478 * that fall back to implemented interfaces. 479 */ 480 void bdrv_setup_io_funcs(BlockDriver *bdrv); 481 482 int get_tmp_filename(char *filename, int size); 483 BlockDriver *bdrv_probe_all(const uint8_t *buf, int buf_size, 484 const char *filename); 485 486 void bdrv_set_io_limits(BlockDriverState *bs, 487 ThrottleConfig *cfg); 488 489 490 /** 491 * bdrv_add_before_write_notifier: 492 * 493 * Register a callback that is invoked before write requests are processed but 494 * after any throttling or waiting for overlapping requests. 495 */ 496 void bdrv_add_before_write_notifier(BlockDriverState *bs, 497 NotifierWithReturn *notifier); 498 499 /** 500 * bdrv_detach_aio_context: 501 * 502 * May be called from .bdrv_detach_aio_context() to detach children from the 503 * current #AioContext. This is only needed by block drivers that manage their 504 * own children. Both ->file and ->backing are automatically handled and 505 * block drivers should not call this function on them explicitly. 506 */ 507 void bdrv_detach_aio_context(BlockDriverState *bs); 508 509 /** 510 * bdrv_attach_aio_context: 511 * 512 * May be called from .bdrv_attach_aio_context() to attach children to the new 513 * #AioContext. This is only needed by block drivers that manage their own 514 * children. Both ->file and ->backing are automatically handled and block 515 * drivers should not call this function on them explicitly. 516 */ 517 void bdrv_attach_aio_context(BlockDriverState *bs, 518 AioContext *new_context); 519 520 /** 521 * bdrv_add_aio_context_notifier: 522 * 523 * If a long-running job intends to be always run in the same AioContext as a 524 * certain BDS, it may use this function to be notified of changes regarding the 525 * association of the BDS to an AioContext. 526 * 527 * attached_aio_context() is called after the target BDS has been attached to a 528 * new AioContext; detach_aio_context() is called before the target BDS is being 529 * detached from its old AioContext. 530 */ 531 void bdrv_add_aio_context_notifier(BlockDriverState *bs, 532 void (*attached_aio_context)(AioContext *new_context, void *opaque), 533 void (*detach_aio_context)(void *opaque), void *opaque); 534 535 /** 536 * bdrv_remove_aio_context_notifier: 537 * 538 * Unsubscribe of change notifications regarding the BDS's AioContext. The 539 * parameters given here have to be the same as those given to 540 * bdrv_add_aio_context_notifier(). 541 */ 542 void bdrv_remove_aio_context_notifier(BlockDriverState *bs, 543 void (*aio_context_attached)(AioContext *, 544 void *), 545 void (*aio_context_detached)(void *), 546 void *opaque); 547 548 #ifdef _WIN32 549 int is_windows_drive(const char *filename); 550 #endif 551 552 /** 553 * stream_start: 554 * @bs: Block device to operate on. 555 * @base: Block device that will become the new base, or %NULL to 556 * flatten the whole backing file chain onto @bs. 557 * @base_id: The file name that will be written to @bs as the new 558 * backing file if the job completes. Ignored if @base is %NULL. 559 * @speed: The maximum speed, in bytes per second, or 0 for unlimited. 560 * @on_error: The action to take upon error. 561 * @cb: Completion function for the job. 562 * @opaque: Opaque pointer value passed to @cb. 563 * @errp: Error object. 564 * 565 * Start a streaming operation on @bs. Clusters that are unallocated 566 * in @bs, but allocated in any image between @base and @bs (both 567 * exclusive) will be written to @bs. At the end of a successful 568 * streaming job, the backing file of @bs will be changed to 569 * @base_id in the written image and to @base in the live BlockDriverState. 570 */ 571 void stream_start(BlockDriverState *bs, BlockDriverState *base, 572 const char *base_id, int64_t speed, BlockdevOnError on_error, 573 BlockCompletionFunc *cb, 574 void *opaque, Error **errp); 575 576 /** 577 * commit_start: 578 * @bs: Active block device. 579 * @top: Top block device to be committed. 580 * @base: Block device that will be written into, and become the new top. 581 * @speed: The maximum speed, in bytes per second, or 0 for unlimited. 582 * @on_error: The action to take upon error. 583 * @cb: Completion function for the job. 584 * @opaque: Opaque pointer value passed to @cb. 585 * @backing_file_str: String to use as the backing file in @top's overlay 586 * @errp: Error object. 587 * 588 */ 589 void commit_start(BlockDriverState *bs, BlockDriverState *base, 590 BlockDriverState *top, int64_t speed, 591 BlockdevOnError on_error, BlockCompletionFunc *cb, 592 void *opaque, const char *backing_file_str, Error **errp); 593 /** 594 * commit_active_start: 595 * @bs: Active block device to be committed. 596 * @base: Block device that will be written into, and become the new top. 597 * @speed: The maximum speed, in bytes per second, or 0 for unlimited. 598 * @on_error: The action to take upon error. 599 * @cb: Completion function for the job. 600 * @opaque: Opaque pointer value passed to @cb. 601 * @errp: Error object. 602 * 603 */ 604 void commit_active_start(BlockDriverState *bs, BlockDriverState *base, 605 int64_t speed, 606 BlockdevOnError on_error, 607 BlockCompletionFunc *cb, 608 void *opaque, Error **errp); 609 /* 610 * mirror_start: 611 * @bs: Block device to operate on. 612 * @target: Block device to write to. 613 * @replaces: Block graph node name to replace once the mirror is done. Can 614 * only be used when full mirroring is selected. 615 * @speed: The maximum speed, in bytes per second, or 0 for unlimited. 616 * @granularity: The chosen granularity for the dirty bitmap. 617 * @buf_size: The amount of data that can be in flight at one time. 618 * @mode: Whether to collapse all images in the chain to the target. 619 * @on_source_error: The action to take upon error reading from the source. 620 * @on_target_error: The action to take upon error writing to the target. 621 * @unmap: Whether to unmap target where source sectors only contain zeroes. 622 * @cb: Completion function for the job. 623 * @opaque: Opaque pointer value passed to @cb. 624 * @errp: Error object. 625 * 626 * Start a mirroring operation on @bs. Clusters that are allocated 627 * in @bs will be written to @bs until the job is cancelled or 628 * manually completed. At the end of a successful mirroring job, 629 * @bs will be switched to read from @target. 630 */ 631 void mirror_start(BlockDriverState *bs, BlockDriverState *target, 632 const char *replaces, 633 int64_t speed, uint32_t granularity, int64_t buf_size, 634 MirrorSyncMode mode, BlockdevOnError on_source_error, 635 BlockdevOnError on_target_error, 636 bool unmap, 637 BlockCompletionFunc *cb, 638 void *opaque, Error **errp); 639 640 /* 641 * backup_start: 642 * @bs: Block device to operate on. 643 * @target: Block device to write to. 644 * @speed: The maximum speed, in bytes per second, or 0 for unlimited. 645 * @sync_mode: What parts of the disk image should be copied to the destination. 646 * @sync_bitmap: The dirty bitmap if sync_mode is MIRROR_SYNC_MODE_INCREMENTAL. 647 * @on_source_error: The action to take upon error reading from the source. 648 * @on_target_error: The action to take upon error writing to the target. 649 * @cb: Completion function for the job. 650 * @opaque: Opaque pointer value passed to @cb. 651 * 652 * Start a backup operation on @bs. Clusters in @bs are written to @target 653 * until the job is cancelled or manually completed. 654 */ 655 void backup_start(BlockDriverState *bs, BlockDriverState *target, 656 int64_t speed, MirrorSyncMode sync_mode, 657 BdrvDirtyBitmap *sync_bitmap, 658 BlockdevOnError on_source_error, 659 BlockdevOnError on_target_error, 660 BlockCompletionFunc *cb, void *opaque, 661 Error **errp); 662 663 void blk_set_bs(BlockBackend *blk, BlockDriverState *bs); 664 665 void blk_dev_change_media_cb(BlockBackend *blk, bool load); 666 bool blk_dev_has_removable_media(BlockBackend *blk); 667 void blk_dev_eject_request(BlockBackend *blk, bool force); 668 bool blk_dev_is_tray_open(BlockBackend *blk); 669 bool blk_dev_is_medium_locked(BlockBackend *blk); 670 void blk_dev_resize_cb(BlockBackend *blk); 671 672 void bdrv_set_dirty(BlockDriverState *bs, int64_t cur_sector, int nr_sectors); 673 bool bdrv_requests_pending(BlockDriverState *bs); 674 675 #endif /* BLOCK_INT_H */ 676