1 /* 2 * vhost-user-blk host device 3 * 4 * Copyright(C) 2017 Intel Corporation. 5 * 6 * Authors: 7 * Changpeng Liu <changpeng.liu@intel.com> 8 * 9 * Largely based on the "vhost-user-scsi.c" and "vhost-scsi.c" implemented by: 10 * Felipe Franciosi <felipe@nutanix.com> 11 * Stefan Hajnoczi <stefanha@linux.vnet.ibm.com> 12 * Nicholas Bellinger <nab@risingtidesystems.com> 13 * 14 * This work is licensed under the terms of the GNU LGPL, version 2 or later. 15 * See the COPYING.LIB file in the top-level directory. 16 * 17 */ 18 19 #include "qemu/osdep.h" 20 #include "qapi/error.h" 21 #include "qemu/error-report.h" 22 #include "qemu/cutils.h" 23 #include "hw/qdev-core.h" 24 #include "hw/qdev-properties.h" 25 #include "hw/qdev-properties-system.h" 26 #include "hw/virtio/virtio-blk-common.h" 27 #include "hw/virtio/vhost.h" 28 #include "hw/virtio/vhost-user-blk.h" 29 #include "hw/virtio/virtio.h" 30 #include "hw/virtio/virtio-bus.h" 31 #include "hw/virtio/virtio-access.h" 32 #include "sysemu/sysemu.h" 33 #include "sysemu/runstate.h" 34 35 #define REALIZE_CONNECTION_RETRIES 3 36 37 static const int user_feature_bits[] = { 38 VIRTIO_BLK_F_SIZE_MAX, 39 VIRTIO_BLK_F_SEG_MAX, 40 VIRTIO_BLK_F_GEOMETRY, 41 VIRTIO_BLK_F_BLK_SIZE, 42 VIRTIO_BLK_F_TOPOLOGY, 43 VIRTIO_BLK_F_MQ, 44 VIRTIO_BLK_F_RO, 45 VIRTIO_BLK_F_FLUSH, 46 VIRTIO_BLK_F_CONFIG_WCE, 47 VIRTIO_BLK_F_DISCARD, 48 VIRTIO_BLK_F_WRITE_ZEROES, 49 VIRTIO_F_VERSION_1, 50 VIRTIO_RING_F_INDIRECT_DESC, 51 VIRTIO_RING_F_EVENT_IDX, 52 VIRTIO_F_NOTIFY_ON_EMPTY, 53 VIRTIO_F_RING_PACKED, 54 VIRTIO_F_IOMMU_PLATFORM, 55 VHOST_INVALID_FEATURE_BIT 56 }; 57 58 static void vhost_user_blk_event(void *opaque, QEMUChrEvent event); 59 60 static void vhost_user_blk_update_config(VirtIODevice *vdev, uint8_t *config) 61 { 62 VHostUserBlk *s = VHOST_USER_BLK(vdev); 63 64 /* Our num_queues overrides the device backend */ 65 virtio_stw_p(vdev, &s->blkcfg.num_queues, s->num_queues); 66 67 memcpy(config, &s->blkcfg, vdev->config_len); 68 } 69 70 static void vhost_user_blk_set_config(VirtIODevice *vdev, const uint8_t *config) 71 { 72 VHostUserBlk *s = VHOST_USER_BLK(vdev); 73 struct virtio_blk_config *blkcfg = (struct virtio_blk_config *)config; 74 int ret; 75 76 if (blkcfg->wce == s->blkcfg.wce) { 77 return; 78 } 79 80 ret = vhost_dev_set_config(&s->dev, &blkcfg->wce, 81 offsetof(struct virtio_blk_config, wce), 82 sizeof(blkcfg->wce), 83 VHOST_SET_CONFIG_TYPE_MASTER); 84 if (ret) { 85 error_report("set device config space failed"); 86 return; 87 } 88 89 s->blkcfg.wce = blkcfg->wce; 90 } 91 92 static int vhost_user_blk_handle_config_change(struct vhost_dev *dev) 93 { 94 int ret; 95 struct virtio_blk_config blkcfg; 96 VirtIODevice *vdev = dev->vdev; 97 VHostUserBlk *s = VHOST_USER_BLK(dev->vdev); 98 Error *local_err = NULL; 99 100 if (!dev->started) { 101 return 0; 102 } 103 104 ret = vhost_dev_get_config(dev, (uint8_t *)&blkcfg, 105 vdev->config_len, &local_err); 106 if (ret < 0) { 107 error_report_err(local_err); 108 return ret; 109 } 110 111 /* valid for resize only */ 112 if (blkcfg.capacity != s->blkcfg.capacity) { 113 s->blkcfg.capacity = blkcfg.capacity; 114 memcpy(dev->vdev->config, &s->blkcfg, vdev->config_len); 115 virtio_notify_config(dev->vdev); 116 } 117 118 return 0; 119 } 120 121 const VhostDevConfigOps blk_ops = { 122 .vhost_dev_config_notifier = vhost_user_blk_handle_config_change, 123 }; 124 125 static int vhost_user_blk_start(VirtIODevice *vdev, Error **errp) 126 { 127 VHostUserBlk *s = VHOST_USER_BLK(vdev); 128 BusState *qbus = BUS(qdev_get_parent_bus(DEVICE(vdev))); 129 VirtioBusClass *k = VIRTIO_BUS_GET_CLASS(qbus); 130 int i, ret; 131 132 if (!k->set_guest_notifiers) { 133 error_setg(errp, "binding does not support guest notifiers"); 134 return -ENOSYS; 135 } 136 137 ret = vhost_dev_enable_notifiers(&s->dev, vdev); 138 if (ret < 0) { 139 error_setg_errno(errp, -ret, "Error enabling host notifiers"); 140 return ret; 141 } 142 143 ret = k->set_guest_notifiers(qbus->parent, s->dev.nvqs, true); 144 if (ret < 0) { 145 error_setg_errno(errp, -ret, "Error binding guest notifier"); 146 goto err_host_notifiers; 147 } 148 149 s->dev.acked_features = vdev->guest_features; 150 151 ret = vhost_dev_prepare_inflight(&s->dev, vdev); 152 if (ret < 0) { 153 error_setg_errno(errp, -ret, "Error setting inflight format"); 154 goto err_guest_notifiers; 155 } 156 157 if (!s->inflight->addr) { 158 ret = vhost_dev_get_inflight(&s->dev, s->queue_size, s->inflight); 159 if (ret < 0) { 160 error_setg_errno(errp, -ret, "Error getting inflight"); 161 goto err_guest_notifiers; 162 } 163 } 164 165 ret = vhost_dev_set_inflight(&s->dev, s->inflight); 166 if (ret < 0) { 167 error_setg_errno(errp, -ret, "Error setting inflight"); 168 goto err_guest_notifiers; 169 } 170 171 ret = vhost_dev_start(&s->dev, vdev); 172 if (ret < 0) { 173 error_setg_errno(errp, -ret, "Error starting vhost"); 174 goto err_guest_notifiers; 175 } 176 s->started_vu = true; 177 178 /* guest_notifier_mask/pending not used yet, so just unmask 179 * everything here. virtio-pci will do the right thing by 180 * enabling/disabling irqfd. 181 */ 182 for (i = 0; i < s->dev.nvqs; i++) { 183 vhost_virtqueue_mask(&s->dev, vdev, i, false); 184 } 185 186 return ret; 187 188 err_guest_notifiers: 189 k->set_guest_notifiers(qbus->parent, s->dev.nvqs, false); 190 err_host_notifiers: 191 vhost_dev_disable_notifiers(&s->dev, vdev); 192 return ret; 193 } 194 195 static void vhost_user_blk_stop(VirtIODevice *vdev) 196 { 197 VHostUserBlk *s = VHOST_USER_BLK(vdev); 198 BusState *qbus = BUS(qdev_get_parent_bus(DEVICE(vdev))); 199 VirtioBusClass *k = VIRTIO_BUS_GET_CLASS(qbus); 200 int ret; 201 202 if (!s->started_vu) { 203 return; 204 } 205 s->started_vu = false; 206 207 if (!k->set_guest_notifiers) { 208 return; 209 } 210 211 vhost_dev_stop(&s->dev, vdev); 212 213 ret = k->set_guest_notifiers(qbus->parent, s->dev.nvqs, false); 214 if (ret < 0) { 215 error_report("vhost guest notifier cleanup failed: %d", ret); 216 return; 217 } 218 219 vhost_dev_disable_notifiers(&s->dev, vdev); 220 } 221 222 static void vhost_user_blk_set_status(VirtIODevice *vdev, uint8_t status) 223 { 224 VHostUserBlk *s = VHOST_USER_BLK(vdev); 225 bool should_start = virtio_device_started(vdev, status); 226 Error *local_err = NULL; 227 int ret; 228 229 if (!vdev->vm_running) { 230 should_start = false; 231 } 232 233 if (!s->connected) { 234 return; 235 } 236 237 if (vhost_dev_is_started(&s->dev) == should_start) { 238 return; 239 } 240 241 if (should_start) { 242 ret = vhost_user_blk_start(vdev, &local_err); 243 if (ret < 0) { 244 error_reportf_err(local_err, "vhost-user-blk: vhost start failed: "); 245 qemu_chr_fe_disconnect(&s->chardev); 246 } 247 } else { 248 vhost_user_blk_stop(vdev); 249 } 250 251 } 252 253 static uint64_t vhost_user_blk_get_features(VirtIODevice *vdev, 254 uint64_t features, 255 Error **errp) 256 { 257 VHostUserBlk *s = VHOST_USER_BLK(vdev); 258 259 /* Turn on pre-defined features */ 260 virtio_add_feature(&features, VIRTIO_BLK_F_SIZE_MAX); 261 virtio_add_feature(&features, VIRTIO_BLK_F_SEG_MAX); 262 virtio_add_feature(&features, VIRTIO_BLK_F_GEOMETRY); 263 virtio_add_feature(&features, VIRTIO_BLK_F_TOPOLOGY); 264 virtio_add_feature(&features, VIRTIO_BLK_F_BLK_SIZE); 265 virtio_add_feature(&features, VIRTIO_BLK_F_FLUSH); 266 virtio_add_feature(&features, VIRTIO_BLK_F_RO); 267 268 if (s->num_queues > 1) { 269 virtio_add_feature(&features, VIRTIO_BLK_F_MQ); 270 } 271 272 return vhost_get_features(&s->dev, user_feature_bits, features); 273 } 274 275 static void vhost_user_blk_handle_output(VirtIODevice *vdev, VirtQueue *vq) 276 { 277 VHostUserBlk *s = VHOST_USER_BLK(vdev); 278 Error *local_err = NULL; 279 int i, ret; 280 281 if (!vdev->start_on_kick) { 282 return; 283 } 284 285 if (!s->connected) { 286 return; 287 } 288 289 if (vhost_dev_is_started(&s->dev)) { 290 return; 291 } 292 293 /* Some guests kick before setting VIRTIO_CONFIG_S_DRIVER_OK so start 294 * vhost here instead of waiting for .set_status(). 295 */ 296 ret = vhost_user_blk_start(vdev, &local_err); 297 if (ret < 0) { 298 error_reportf_err(local_err, "vhost-user-blk: vhost start failed: "); 299 qemu_chr_fe_disconnect(&s->chardev); 300 return; 301 } 302 303 /* Kick right away to begin processing requests already in vring */ 304 for (i = 0; i < s->dev.nvqs; i++) { 305 VirtQueue *kick_vq = virtio_get_queue(vdev, i); 306 307 if (!virtio_queue_get_desc_addr(vdev, i)) { 308 continue; 309 } 310 event_notifier_set(virtio_queue_get_host_notifier(kick_vq)); 311 } 312 } 313 314 static void vhost_user_blk_reset(VirtIODevice *vdev) 315 { 316 VHostUserBlk *s = VHOST_USER_BLK(vdev); 317 318 vhost_dev_free_inflight(s->inflight); 319 } 320 321 static int vhost_user_blk_connect(DeviceState *dev, Error **errp) 322 { 323 VirtIODevice *vdev = VIRTIO_DEVICE(dev); 324 VHostUserBlk *s = VHOST_USER_BLK(vdev); 325 int ret = 0; 326 327 if (s->connected) { 328 return 0; 329 } 330 s->connected = true; 331 332 s->dev.num_queues = s->num_queues; 333 s->dev.nvqs = s->num_queues; 334 s->dev.vqs = s->vhost_vqs; 335 s->dev.vq_index = 0; 336 s->dev.backend_features = 0; 337 338 vhost_dev_set_config_notifier(&s->dev, &blk_ops); 339 340 s->vhost_user.supports_config = true; 341 ret = vhost_dev_init(&s->dev, &s->vhost_user, VHOST_BACKEND_TYPE_USER, 0, 342 errp); 343 if (ret < 0) { 344 return ret; 345 } 346 347 /* restore vhost state */ 348 if (virtio_device_started(vdev, vdev->status)) { 349 ret = vhost_user_blk_start(vdev, errp); 350 if (ret < 0) { 351 return ret; 352 } 353 } 354 355 return 0; 356 } 357 358 static void vhost_user_blk_disconnect(DeviceState *dev) 359 { 360 VirtIODevice *vdev = VIRTIO_DEVICE(dev); 361 VHostUserBlk *s = VHOST_USER_BLK(vdev); 362 363 if (!s->connected) { 364 return; 365 } 366 s->connected = false; 367 368 vhost_user_blk_stop(vdev); 369 370 vhost_dev_cleanup(&s->dev); 371 } 372 373 static void vhost_user_blk_chr_closed_bh(void *opaque) 374 { 375 DeviceState *dev = opaque; 376 VirtIODevice *vdev = VIRTIO_DEVICE(dev); 377 VHostUserBlk *s = VHOST_USER_BLK(vdev); 378 379 vhost_user_blk_disconnect(dev); 380 qemu_chr_fe_set_handlers(&s->chardev, NULL, NULL, vhost_user_blk_event, 381 NULL, opaque, NULL, true); 382 } 383 384 static void vhost_user_blk_event(void *opaque, QEMUChrEvent event) 385 { 386 DeviceState *dev = opaque; 387 VirtIODevice *vdev = VIRTIO_DEVICE(dev); 388 VHostUserBlk *s = VHOST_USER_BLK(vdev); 389 Error *local_err = NULL; 390 391 switch (event) { 392 case CHR_EVENT_OPENED: 393 if (vhost_user_blk_connect(dev, &local_err) < 0) { 394 error_report_err(local_err); 395 qemu_chr_fe_disconnect(&s->chardev); 396 return; 397 } 398 break; 399 case CHR_EVENT_CLOSED: 400 if (!runstate_check(RUN_STATE_SHUTDOWN)) { 401 /* 402 * A close event may happen during a read/write, but vhost 403 * code assumes the vhost_dev remains setup, so delay the 404 * stop & clear. 405 */ 406 AioContext *ctx = qemu_get_current_aio_context(); 407 408 qemu_chr_fe_set_handlers(&s->chardev, NULL, NULL, NULL, NULL, 409 NULL, NULL, false); 410 aio_bh_schedule_oneshot(ctx, vhost_user_blk_chr_closed_bh, opaque); 411 412 /* 413 * Move vhost device to the stopped state. The vhost-user device 414 * will be clean up and disconnected in BH. This can be useful in 415 * the vhost migration code. If disconnect was caught there is an 416 * option for the general vhost code to get the dev state without 417 * knowing its type (in this case vhost-user). 418 * 419 * FIXME: this is sketchy to be reaching into vhost_dev 420 * now because we are forcing something that implies we 421 * have executed vhost_dev_stop() but that won't happen 422 * until vhost_user_blk_stop() gets called from the bh. 423 * Really this state check should be tracked locally. 424 */ 425 s->dev.started = false; 426 } 427 break; 428 case CHR_EVENT_BREAK: 429 case CHR_EVENT_MUX_IN: 430 case CHR_EVENT_MUX_OUT: 431 /* Ignore */ 432 break; 433 } 434 } 435 436 static int vhost_user_blk_realize_connect(VHostUserBlk *s, Error **errp) 437 { 438 DeviceState *dev = &s->parent_obj.parent_obj; 439 int ret; 440 441 s->connected = false; 442 443 ret = qemu_chr_fe_wait_connected(&s->chardev, errp); 444 if (ret < 0) { 445 return ret; 446 } 447 448 ret = vhost_user_blk_connect(dev, errp); 449 if (ret < 0) { 450 qemu_chr_fe_disconnect(&s->chardev); 451 return ret; 452 } 453 assert(s->connected); 454 455 ret = vhost_dev_get_config(&s->dev, (uint8_t *)&s->blkcfg, 456 s->parent_obj.config_len, errp); 457 if (ret < 0) { 458 qemu_chr_fe_disconnect(&s->chardev); 459 vhost_dev_cleanup(&s->dev); 460 return ret; 461 } 462 463 return 0; 464 } 465 466 static void vhost_user_blk_device_realize(DeviceState *dev, Error **errp) 467 { 468 ERRP_GUARD(); 469 VirtIODevice *vdev = VIRTIO_DEVICE(dev); 470 VHostUserBlk *s = VHOST_USER_BLK(vdev); 471 size_t config_size; 472 int retries; 473 int i, ret; 474 475 if (!s->chardev.chr) { 476 error_setg(errp, "chardev is mandatory"); 477 return; 478 } 479 480 if (s->num_queues == VHOST_USER_BLK_AUTO_NUM_QUEUES) { 481 s->num_queues = 1; 482 } 483 if (!s->num_queues || s->num_queues > VIRTIO_QUEUE_MAX) { 484 error_setg(errp, "invalid number of IO queues"); 485 return; 486 } 487 488 if (!s->queue_size) { 489 error_setg(errp, "queue size must be non-zero"); 490 return; 491 } 492 if (s->queue_size > VIRTQUEUE_MAX_SIZE) { 493 error_setg(errp, "queue size must not exceed %d", 494 VIRTQUEUE_MAX_SIZE); 495 return; 496 } 497 498 if (!vhost_user_init(&s->vhost_user, &s->chardev, errp)) { 499 return; 500 } 501 502 config_size = virtio_get_config_size(&virtio_blk_cfg_size_params, 503 vdev->host_features); 504 virtio_init(vdev, VIRTIO_ID_BLOCK, config_size); 505 506 s->virtqs = g_new(VirtQueue *, s->num_queues); 507 for (i = 0; i < s->num_queues; i++) { 508 s->virtqs[i] = virtio_add_queue(vdev, s->queue_size, 509 vhost_user_blk_handle_output); 510 } 511 512 s->inflight = g_new0(struct vhost_inflight, 1); 513 s->vhost_vqs = g_new0(struct vhost_virtqueue, s->num_queues); 514 515 retries = REALIZE_CONNECTION_RETRIES; 516 assert(!*errp); 517 do { 518 if (*errp) { 519 error_prepend(errp, "Reconnecting after error: "); 520 error_report_err(*errp); 521 *errp = NULL; 522 } 523 ret = vhost_user_blk_realize_connect(s, errp); 524 } while (ret < 0 && retries--); 525 526 if (ret < 0) { 527 goto virtio_err; 528 } 529 530 /* we're fully initialized, now we can operate, so add the handler */ 531 qemu_chr_fe_set_handlers(&s->chardev, NULL, NULL, 532 vhost_user_blk_event, NULL, (void *)dev, 533 NULL, true); 534 return; 535 536 virtio_err: 537 g_free(s->vhost_vqs); 538 s->vhost_vqs = NULL; 539 g_free(s->inflight); 540 s->inflight = NULL; 541 for (i = 0; i < s->num_queues; i++) { 542 virtio_delete_queue(s->virtqs[i]); 543 } 544 g_free(s->virtqs); 545 virtio_cleanup(vdev); 546 vhost_user_cleanup(&s->vhost_user); 547 } 548 549 static void vhost_user_blk_device_unrealize(DeviceState *dev) 550 { 551 VirtIODevice *vdev = VIRTIO_DEVICE(dev); 552 VHostUserBlk *s = VHOST_USER_BLK(dev); 553 int i; 554 555 virtio_set_status(vdev, 0); 556 qemu_chr_fe_set_handlers(&s->chardev, NULL, NULL, NULL, 557 NULL, NULL, NULL, false); 558 vhost_dev_cleanup(&s->dev); 559 vhost_dev_free_inflight(s->inflight); 560 g_free(s->vhost_vqs); 561 s->vhost_vqs = NULL; 562 g_free(s->inflight); 563 s->inflight = NULL; 564 565 for (i = 0; i < s->num_queues; i++) { 566 virtio_delete_queue(s->virtqs[i]); 567 } 568 g_free(s->virtqs); 569 virtio_cleanup(vdev); 570 vhost_user_cleanup(&s->vhost_user); 571 } 572 573 static void vhost_user_blk_instance_init(Object *obj) 574 { 575 VHostUserBlk *s = VHOST_USER_BLK(obj); 576 577 device_add_bootindex_property(obj, &s->bootindex, "bootindex", 578 "/disk@0,0", DEVICE(obj)); 579 } 580 581 static struct vhost_dev *vhost_user_blk_get_vhost(VirtIODevice *vdev) 582 { 583 VHostUserBlk *s = VHOST_USER_BLK(vdev); 584 return &s->dev; 585 } 586 587 static const VMStateDescription vmstate_vhost_user_blk = { 588 .name = "vhost-user-blk", 589 .minimum_version_id = 1, 590 .version_id = 1, 591 .fields = (VMStateField[]) { 592 VMSTATE_VIRTIO_DEVICE, 593 VMSTATE_END_OF_LIST() 594 }, 595 }; 596 597 static Property vhost_user_blk_properties[] = { 598 DEFINE_PROP_CHR("chardev", VHostUserBlk, chardev), 599 DEFINE_PROP_UINT16("num-queues", VHostUserBlk, num_queues, 600 VHOST_USER_BLK_AUTO_NUM_QUEUES), 601 DEFINE_PROP_UINT32("queue-size", VHostUserBlk, queue_size, 128), 602 DEFINE_PROP_BIT64("config-wce", VHostUserBlk, parent_obj.host_features, 603 VIRTIO_BLK_F_CONFIG_WCE, true), 604 DEFINE_PROP_BIT64("discard", VHostUserBlk, parent_obj.host_features, 605 VIRTIO_BLK_F_DISCARD, true), 606 DEFINE_PROP_BIT64("write-zeroes", VHostUserBlk, parent_obj.host_features, 607 VIRTIO_BLK_F_WRITE_ZEROES, true), 608 DEFINE_PROP_END_OF_LIST(), 609 }; 610 611 static void vhost_user_blk_class_init(ObjectClass *klass, void *data) 612 { 613 DeviceClass *dc = DEVICE_CLASS(klass); 614 VirtioDeviceClass *vdc = VIRTIO_DEVICE_CLASS(klass); 615 616 device_class_set_props(dc, vhost_user_blk_properties); 617 dc->vmsd = &vmstate_vhost_user_blk; 618 set_bit(DEVICE_CATEGORY_STORAGE, dc->categories); 619 vdc->realize = vhost_user_blk_device_realize; 620 vdc->unrealize = vhost_user_blk_device_unrealize; 621 vdc->get_config = vhost_user_blk_update_config; 622 vdc->set_config = vhost_user_blk_set_config; 623 vdc->get_features = vhost_user_blk_get_features; 624 vdc->set_status = vhost_user_blk_set_status; 625 vdc->reset = vhost_user_blk_reset; 626 vdc->get_vhost = vhost_user_blk_get_vhost; 627 } 628 629 static const TypeInfo vhost_user_blk_info = { 630 .name = TYPE_VHOST_USER_BLK, 631 .parent = TYPE_VIRTIO_DEVICE, 632 .instance_size = sizeof(VHostUserBlk), 633 .instance_init = vhost_user_blk_instance_init, 634 .class_init = vhost_user_blk_class_init, 635 }; 636 637 static void virtio_register_types(void) 638 { 639 type_register_static(&vhost_user_blk_info); 640 } 641 642 type_init(virtio_register_types) 643