1 /* 2 * Virtio PCI Bindings 3 * 4 * Copyright IBM, Corp. 2007 5 * Copyright (c) 2009 CodeSourcery 6 * 7 * Authors: 8 * Anthony Liguori <aliguori@us.ibm.com> 9 * Paul Brook <paul@codesourcery.com> 10 * 11 * This work is licensed under the terms of the GNU GPL, version 2. See 12 * the COPYING file in the top-level directory. 13 * 14 * Contributions after 2012-01-13 are licensed under the terms of the 15 * GNU GPL, version 2 or (at your option) any later version. 16 */ 17 18 #include <inttypes.h> 19 20 #include "hw/virtio/virtio.h" 21 #include "hw/virtio/virtio-blk.h" 22 #include "hw/virtio/virtio-net.h" 23 #include "hw/virtio/virtio-serial.h" 24 #include "hw/virtio/virtio-scsi.h" 25 #include "hw/virtio/virtio-balloon.h" 26 #include "hw/pci/pci.h" 27 #include "qemu/error-report.h" 28 #include "hw/pci/msi.h" 29 #include "hw/pci/msix.h" 30 #include "hw/loader.h" 31 #include "sysemu/kvm.h" 32 #include "sysemu/blockdev.h" 33 #include "virtio-pci.h" 34 #include "qemu/range.h" 35 #include "hw/virtio/virtio-bus.h" 36 #include "qapi/visitor.h" 37 38 /* from Linux's linux/virtio_pci.h */ 39 40 /* A 32-bit r/o bitmask of the features supported by the host */ 41 #define VIRTIO_PCI_HOST_FEATURES 0 42 43 /* A 32-bit r/w bitmask of features activated by the guest */ 44 #define VIRTIO_PCI_GUEST_FEATURES 4 45 46 /* A 32-bit r/w PFN for the currently selected queue */ 47 #define VIRTIO_PCI_QUEUE_PFN 8 48 49 /* A 16-bit r/o queue size for the currently selected queue */ 50 #define VIRTIO_PCI_QUEUE_NUM 12 51 52 /* A 16-bit r/w queue selector */ 53 #define VIRTIO_PCI_QUEUE_SEL 14 54 55 /* A 16-bit r/w queue notifier */ 56 #define VIRTIO_PCI_QUEUE_NOTIFY 16 57 58 /* An 8-bit device status register. */ 59 #define VIRTIO_PCI_STATUS 18 60 61 /* An 8-bit r/o interrupt status register. Reading the value will return the 62 * current contents of the ISR and will also clear it. This is effectively 63 * a read-and-acknowledge. */ 64 #define VIRTIO_PCI_ISR 19 65 66 /* MSI-X registers: only enabled if MSI-X is enabled. */ 67 /* A 16-bit vector for configuration changes. */ 68 #define VIRTIO_MSI_CONFIG_VECTOR 20 69 /* A 16-bit vector for selected queue notifications. */ 70 #define VIRTIO_MSI_QUEUE_VECTOR 22 71 72 /* Config space size */ 73 #define VIRTIO_PCI_CONFIG_NOMSI 20 74 #define VIRTIO_PCI_CONFIG_MSI 24 75 #define VIRTIO_PCI_REGION_SIZE(dev) (msix_present(dev) ? \ 76 VIRTIO_PCI_CONFIG_MSI : \ 77 VIRTIO_PCI_CONFIG_NOMSI) 78 79 /* The remaining space is defined by each driver as the per-driver 80 * configuration space */ 81 #define VIRTIO_PCI_CONFIG(dev) (msix_enabled(dev) ? \ 82 VIRTIO_PCI_CONFIG_MSI : \ 83 VIRTIO_PCI_CONFIG_NOMSI) 84 85 /* How many bits to shift physical queue address written to QUEUE_PFN. 86 * 12 is historical, and due to x86 page size. */ 87 #define VIRTIO_PCI_QUEUE_ADDR_SHIFT 12 88 89 /* Flags track per-device state like workarounds for quirks in older guests. */ 90 #define VIRTIO_PCI_FLAG_BUS_MASTER_BUG (1 << 0) 91 92 /* QEMU doesn't strictly need write barriers since everything runs in 93 * lock-step. We'll leave the calls to wmb() in though to make it obvious for 94 * KVM or if kqemu gets SMP support. 95 */ 96 #define wmb() do { } while (0) 97 98 /* HACK for virtio to determine if it's running a big endian guest */ 99 bool virtio_is_big_endian(void); 100 101 /* virtio device */ 102 /* DeviceState to VirtIOPCIProxy. For use off data-path. TODO: use QOM. */ 103 static inline VirtIOPCIProxy *to_virtio_pci_proxy(DeviceState *d) 104 { 105 return container_of(d, VirtIOPCIProxy, pci_dev.qdev); 106 } 107 108 /* DeviceState to VirtIOPCIProxy. Note: used on datapath, 109 * be careful and test performance if you change this. 110 */ 111 static inline VirtIOPCIProxy *to_virtio_pci_proxy_fast(DeviceState *d) 112 { 113 return container_of(d, VirtIOPCIProxy, pci_dev.qdev); 114 } 115 116 static void virtio_pci_notify(DeviceState *d, uint16_t vector) 117 { 118 VirtIOPCIProxy *proxy = to_virtio_pci_proxy_fast(d); 119 if (msix_enabled(&proxy->pci_dev)) 120 msix_notify(&proxy->pci_dev, vector); 121 else 122 qemu_set_irq(proxy->pci_dev.irq[0], proxy->vdev->isr & 1); 123 } 124 125 static void virtio_pci_save_config(DeviceState *d, QEMUFile *f) 126 { 127 VirtIOPCIProxy *proxy = to_virtio_pci_proxy(d); 128 pci_device_save(&proxy->pci_dev, f); 129 msix_save(&proxy->pci_dev, f); 130 if (msix_present(&proxy->pci_dev)) 131 qemu_put_be16(f, proxy->vdev->config_vector); 132 } 133 134 static void virtio_pci_save_queue(DeviceState *d, int n, QEMUFile *f) 135 { 136 VirtIOPCIProxy *proxy = to_virtio_pci_proxy(d); 137 if (msix_present(&proxy->pci_dev)) 138 qemu_put_be16(f, virtio_queue_vector(proxy->vdev, n)); 139 } 140 141 static int virtio_pci_load_config(DeviceState *d, QEMUFile *f) 142 { 143 VirtIOPCIProxy *proxy = to_virtio_pci_proxy(d); 144 int ret; 145 ret = pci_device_load(&proxy->pci_dev, f); 146 if (ret) { 147 return ret; 148 } 149 msix_unuse_all_vectors(&proxy->pci_dev); 150 msix_load(&proxy->pci_dev, f); 151 if (msix_present(&proxy->pci_dev)) { 152 qemu_get_be16s(f, &proxy->vdev->config_vector); 153 } else { 154 proxy->vdev->config_vector = VIRTIO_NO_VECTOR; 155 } 156 if (proxy->vdev->config_vector != VIRTIO_NO_VECTOR) { 157 return msix_vector_use(&proxy->pci_dev, proxy->vdev->config_vector); 158 } 159 return 0; 160 } 161 162 static int virtio_pci_load_queue(DeviceState *d, int n, QEMUFile *f) 163 { 164 VirtIOPCIProxy *proxy = to_virtio_pci_proxy(d); 165 uint16_t vector; 166 if (msix_present(&proxy->pci_dev)) { 167 qemu_get_be16s(f, &vector); 168 } else { 169 vector = VIRTIO_NO_VECTOR; 170 } 171 virtio_queue_set_vector(proxy->vdev, n, vector); 172 if (vector != VIRTIO_NO_VECTOR) { 173 return msix_vector_use(&proxy->pci_dev, vector); 174 } 175 return 0; 176 } 177 178 static int virtio_pci_set_host_notifier_internal(VirtIOPCIProxy *proxy, 179 int n, bool assign, bool set_handler) 180 { 181 VirtQueue *vq = virtio_get_queue(proxy->vdev, n); 182 EventNotifier *notifier = virtio_queue_get_host_notifier(vq); 183 int r = 0; 184 185 if (assign) { 186 r = event_notifier_init(notifier, 1); 187 if (r < 0) { 188 error_report("%s: unable to init event notifier: %d", 189 __func__, r); 190 return r; 191 } 192 virtio_queue_set_host_notifier_fd_handler(vq, true, set_handler); 193 memory_region_add_eventfd(&proxy->bar, VIRTIO_PCI_QUEUE_NOTIFY, 2, 194 true, n, notifier); 195 } else { 196 memory_region_del_eventfd(&proxy->bar, VIRTIO_PCI_QUEUE_NOTIFY, 2, 197 true, n, notifier); 198 virtio_queue_set_host_notifier_fd_handler(vq, false, false); 199 event_notifier_cleanup(notifier); 200 } 201 return r; 202 } 203 204 static void virtio_pci_start_ioeventfd(VirtIOPCIProxy *proxy) 205 { 206 int n, r; 207 208 if (!(proxy->flags & VIRTIO_PCI_FLAG_USE_IOEVENTFD) || 209 proxy->ioeventfd_disabled || 210 proxy->ioeventfd_started) { 211 return; 212 } 213 214 for (n = 0; n < VIRTIO_PCI_QUEUE_MAX; n++) { 215 if (!virtio_queue_get_num(proxy->vdev, n)) { 216 continue; 217 } 218 219 r = virtio_pci_set_host_notifier_internal(proxy, n, true, true); 220 if (r < 0) { 221 goto assign_error; 222 } 223 } 224 proxy->ioeventfd_started = true; 225 return; 226 227 assign_error: 228 while (--n >= 0) { 229 if (!virtio_queue_get_num(proxy->vdev, n)) { 230 continue; 231 } 232 233 r = virtio_pci_set_host_notifier_internal(proxy, n, false, false); 234 assert(r >= 0); 235 } 236 proxy->ioeventfd_started = false; 237 error_report("%s: failed. Fallback to a userspace (slower).", __func__); 238 } 239 240 static void virtio_pci_stop_ioeventfd(VirtIOPCIProxy *proxy) 241 { 242 int r; 243 int n; 244 245 if (!proxy->ioeventfd_started) { 246 return; 247 } 248 249 for (n = 0; n < VIRTIO_PCI_QUEUE_MAX; n++) { 250 if (!virtio_queue_get_num(proxy->vdev, n)) { 251 continue; 252 } 253 254 r = virtio_pci_set_host_notifier_internal(proxy, n, false, false); 255 assert(r >= 0); 256 } 257 proxy->ioeventfd_started = false; 258 } 259 260 static void virtio_pci_reset(DeviceState *d) 261 { 262 VirtIOPCIProxy *proxy = to_virtio_pci_proxy(d); 263 virtio_pci_stop_ioeventfd(proxy); 264 virtio_reset(proxy->vdev); 265 msix_unuse_all_vectors(&proxy->pci_dev); 266 proxy->flags &= ~VIRTIO_PCI_FLAG_BUS_MASTER_BUG; 267 } 268 269 static void virtio_ioport_write(void *opaque, uint32_t addr, uint32_t val) 270 { 271 VirtIOPCIProxy *proxy = opaque; 272 VirtIODevice *vdev = proxy->vdev; 273 hwaddr pa; 274 275 switch (addr) { 276 case VIRTIO_PCI_GUEST_FEATURES: 277 /* Guest does not negotiate properly? We have to assume nothing. */ 278 if (val & (1 << VIRTIO_F_BAD_FEATURE)) { 279 val = vdev->bad_features ? vdev->bad_features(vdev) : 0; 280 } 281 virtio_set_features(vdev, val); 282 break; 283 case VIRTIO_PCI_QUEUE_PFN: 284 pa = (hwaddr)val << VIRTIO_PCI_QUEUE_ADDR_SHIFT; 285 if (pa == 0) { 286 virtio_pci_stop_ioeventfd(proxy); 287 virtio_reset(proxy->vdev); 288 msix_unuse_all_vectors(&proxy->pci_dev); 289 } 290 else 291 virtio_queue_set_addr(vdev, vdev->queue_sel, pa); 292 break; 293 case VIRTIO_PCI_QUEUE_SEL: 294 if (val < VIRTIO_PCI_QUEUE_MAX) 295 vdev->queue_sel = val; 296 break; 297 case VIRTIO_PCI_QUEUE_NOTIFY: 298 if (val < VIRTIO_PCI_QUEUE_MAX) { 299 virtio_queue_notify(vdev, val); 300 } 301 break; 302 case VIRTIO_PCI_STATUS: 303 if (!(val & VIRTIO_CONFIG_S_DRIVER_OK)) { 304 virtio_pci_stop_ioeventfd(proxy); 305 } 306 307 virtio_set_status(vdev, val & 0xFF); 308 309 if (val & VIRTIO_CONFIG_S_DRIVER_OK) { 310 virtio_pci_start_ioeventfd(proxy); 311 } 312 313 if (vdev->status == 0) { 314 virtio_reset(proxy->vdev); 315 msix_unuse_all_vectors(&proxy->pci_dev); 316 } 317 318 /* Linux before 2.6.34 sets the device as OK without enabling 319 the PCI device bus master bit. In this case we need to disable 320 some safety checks. */ 321 if ((val & VIRTIO_CONFIG_S_DRIVER_OK) && 322 !(proxy->pci_dev.config[PCI_COMMAND] & PCI_COMMAND_MASTER)) { 323 proxy->flags |= VIRTIO_PCI_FLAG_BUS_MASTER_BUG; 324 } 325 break; 326 case VIRTIO_MSI_CONFIG_VECTOR: 327 msix_vector_unuse(&proxy->pci_dev, vdev->config_vector); 328 /* Make it possible for guest to discover an error took place. */ 329 if (msix_vector_use(&proxy->pci_dev, val) < 0) 330 val = VIRTIO_NO_VECTOR; 331 vdev->config_vector = val; 332 break; 333 case VIRTIO_MSI_QUEUE_VECTOR: 334 msix_vector_unuse(&proxy->pci_dev, 335 virtio_queue_vector(vdev, vdev->queue_sel)); 336 /* Make it possible for guest to discover an error took place. */ 337 if (msix_vector_use(&proxy->pci_dev, val) < 0) 338 val = VIRTIO_NO_VECTOR; 339 virtio_queue_set_vector(vdev, vdev->queue_sel, val); 340 break; 341 default: 342 error_report("%s: unexpected address 0x%x value 0x%x", 343 __func__, addr, val); 344 break; 345 } 346 } 347 348 static uint32_t virtio_ioport_read(VirtIOPCIProxy *proxy, uint32_t addr) 349 { 350 VirtIODevice *vdev = proxy->vdev; 351 uint32_t ret = 0xFFFFFFFF; 352 353 switch (addr) { 354 case VIRTIO_PCI_HOST_FEATURES: 355 ret = proxy->host_features; 356 break; 357 case VIRTIO_PCI_GUEST_FEATURES: 358 ret = vdev->guest_features; 359 break; 360 case VIRTIO_PCI_QUEUE_PFN: 361 ret = virtio_queue_get_addr(vdev, vdev->queue_sel) 362 >> VIRTIO_PCI_QUEUE_ADDR_SHIFT; 363 break; 364 case VIRTIO_PCI_QUEUE_NUM: 365 ret = virtio_queue_get_num(vdev, vdev->queue_sel); 366 break; 367 case VIRTIO_PCI_QUEUE_SEL: 368 ret = vdev->queue_sel; 369 break; 370 case VIRTIO_PCI_STATUS: 371 ret = vdev->status; 372 break; 373 case VIRTIO_PCI_ISR: 374 /* reading from the ISR also clears it. */ 375 ret = vdev->isr; 376 vdev->isr = 0; 377 qemu_set_irq(proxy->pci_dev.irq[0], 0); 378 break; 379 case VIRTIO_MSI_CONFIG_VECTOR: 380 ret = vdev->config_vector; 381 break; 382 case VIRTIO_MSI_QUEUE_VECTOR: 383 ret = virtio_queue_vector(vdev, vdev->queue_sel); 384 break; 385 default: 386 break; 387 } 388 389 return ret; 390 } 391 392 static uint64_t virtio_pci_config_read(void *opaque, hwaddr addr, 393 unsigned size) 394 { 395 VirtIOPCIProxy *proxy = opaque; 396 uint32_t config = VIRTIO_PCI_CONFIG(&proxy->pci_dev); 397 uint64_t val = 0; 398 if (addr < config) { 399 return virtio_ioport_read(proxy, addr); 400 } 401 addr -= config; 402 403 switch (size) { 404 case 1: 405 val = virtio_config_readb(proxy->vdev, addr); 406 break; 407 case 2: 408 val = virtio_config_readw(proxy->vdev, addr); 409 if (virtio_is_big_endian()) { 410 val = bswap16(val); 411 } 412 break; 413 case 4: 414 val = virtio_config_readl(proxy->vdev, addr); 415 if (virtio_is_big_endian()) { 416 val = bswap32(val); 417 } 418 break; 419 } 420 return val; 421 } 422 423 static void virtio_pci_config_write(void *opaque, hwaddr addr, 424 uint64_t val, unsigned size) 425 { 426 VirtIOPCIProxy *proxy = opaque; 427 uint32_t config = VIRTIO_PCI_CONFIG(&proxy->pci_dev); 428 if (addr < config) { 429 virtio_ioport_write(proxy, addr, val); 430 return; 431 } 432 addr -= config; 433 /* 434 * Virtio-PCI is odd. Ioports are LE but config space is target native 435 * endian. 436 */ 437 switch (size) { 438 case 1: 439 virtio_config_writeb(proxy->vdev, addr, val); 440 break; 441 case 2: 442 if (virtio_is_big_endian()) { 443 val = bswap16(val); 444 } 445 virtio_config_writew(proxy->vdev, addr, val); 446 break; 447 case 4: 448 if (virtio_is_big_endian()) { 449 val = bswap32(val); 450 } 451 virtio_config_writel(proxy->vdev, addr, val); 452 break; 453 } 454 } 455 456 static const MemoryRegionOps virtio_pci_config_ops = { 457 .read = virtio_pci_config_read, 458 .write = virtio_pci_config_write, 459 .impl = { 460 .min_access_size = 1, 461 .max_access_size = 4, 462 }, 463 .endianness = DEVICE_LITTLE_ENDIAN, 464 }; 465 466 static void virtio_write_config(PCIDevice *pci_dev, uint32_t address, 467 uint32_t val, int len) 468 { 469 VirtIOPCIProxy *proxy = DO_UPCAST(VirtIOPCIProxy, pci_dev, pci_dev); 470 471 pci_default_write_config(pci_dev, address, val, len); 472 473 if (range_covers_byte(address, len, PCI_COMMAND) && 474 !(pci_dev->config[PCI_COMMAND] & PCI_COMMAND_MASTER) && 475 !(proxy->flags & VIRTIO_PCI_FLAG_BUS_MASTER_BUG)) { 476 virtio_pci_stop_ioeventfd(proxy); 477 virtio_set_status(proxy->vdev, 478 proxy->vdev->status & ~VIRTIO_CONFIG_S_DRIVER_OK); 479 } 480 } 481 482 static unsigned virtio_pci_get_features(DeviceState *d) 483 { 484 VirtIOPCIProxy *proxy = to_virtio_pci_proxy(d); 485 return proxy->host_features; 486 } 487 488 static int kvm_virtio_pci_vq_vector_use(VirtIOPCIProxy *proxy, 489 unsigned int queue_no, 490 unsigned int vector, 491 MSIMessage msg) 492 { 493 VirtIOIRQFD *irqfd = &proxy->vector_irqfd[vector]; 494 int ret; 495 496 if (irqfd->users == 0) { 497 ret = kvm_irqchip_add_msi_route(kvm_state, msg); 498 if (ret < 0) { 499 return ret; 500 } 501 irqfd->virq = ret; 502 } 503 irqfd->users++; 504 return 0; 505 } 506 507 static void kvm_virtio_pci_vq_vector_release(VirtIOPCIProxy *proxy, 508 unsigned int vector) 509 { 510 VirtIOIRQFD *irqfd = &proxy->vector_irqfd[vector]; 511 if (--irqfd->users == 0) { 512 kvm_irqchip_release_virq(kvm_state, irqfd->virq); 513 } 514 } 515 516 static int kvm_virtio_pci_irqfd_use(VirtIOPCIProxy *proxy, 517 unsigned int queue_no, 518 unsigned int vector) 519 { 520 VirtIOIRQFD *irqfd = &proxy->vector_irqfd[vector]; 521 VirtQueue *vq = virtio_get_queue(proxy->vdev, queue_no); 522 EventNotifier *n = virtio_queue_get_guest_notifier(vq); 523 int ret; 524 ret = kvm_irqchip_add_irqfd_notifier(kvm_state, n, irqfd->virq); 525 return ret; 526 } 527 528 static void kvm_virtio_pci_irqfd_release(VirtIOPCIProxy *proxy, 529 unsigned int queue_no, 530 unsigned int vector) 531 { 532 VirtQueue *vq = virtio_get_queue(proxy->vdev, queue_no); 533 EventNotifier *n = virtio_queue_get_guest_notifier(vq); 534 VirtIOIRQFD *irqfd = &proxy->vector_irqfd[vector]; 535 int ret; 536 537 ret = kvm_irqchip_remove_irqfd_notifier(kvm_state, n, irqfd->virq); 538 assert(ret == 0); 539 } 540 541 static int kvm_virtio_pci_vector_use(VirtIOPCIProxy *proxy, int nvqs) 542 { 543 PCIDevice *dev = &proxy->pci_dev; 544 VirtIODevice *vdev = proxy->vdev; 545 unsigned int vector; 546 int ret, queue_no; 547 MSIMessage msg; 548 549 for (queue_no = 0; queue_no < nvqs; queue_no++) { 550 if (!virtio_queue_get_num(vdev, queue_no)) { 551 break; 552 } 553 vector = virtio_queue_vector(vdev, queue_no); 554 if (vector >= msix_nr_vectors_allocated(dev)) { 555 continue; 556 } 557 msg = msix_get_message(dev, vector); 558 ret = kvm_virtio_pci_vq_vector_use(proxy, queue_no, vector, msg); 559 if (ret < 0) { 560 goto undo; 561 } 562 /* If guest supports masking, set up irqfd now. 563 * Otherwise, delay until unmasked in the frontend. 564 */ 565 if (proxy->vdev->guest_notifier_mask) { 566 ret = kvm_virtio_pci_irqfd_use(proxy, queue_no, vector); 567 if (ret < 0) { 568 kvm_virtio_pci_vq_vector_release(proxy, vector); 569 goto undo; 570 } 571 } 572 } 573 return 0; 574 575 undo: 576 while (--queue_no >= 0) { 577 vector = virtio_queue_vector(vdev, queue_no); 578 if (vector >= msix_nr_vectors_allocated(dev)) { 579 continue; 580 } 581 if (proxy->vdev->guest_notifier_mask) { 582 kvm_virtio_pci_irqfd_release(proxy, queue_no, vector); 583 } 584 kvm_virtio_pci_vq_vector_release(proxy, vector); 585 } 586 return ret; 587 } 588 589 static void kvm_virtio_pci_vector_release(VirtIOPCIProxy *proxy, int nvqs) 590 { 591 PCIDevice *dev = &proxy->pci_dev; 592 VirtIODevice *vdev = proxy->vdev; 593 unsigned int vector; 594 int queue_no; 595 596 for (queue_no = 0; queue_no < nvqs; queue_no++) { 597 if (!virtio_queue_get_num(vdev, queue_no)) { 598 break; 599 } 600 vector = virtio_queue_vector(vdev, queue_no); 601 if (vector >= msix_nr_vectors_allocated(dev)) { 602 continue; 603 } 604 /* If guest supports masking, clean up irqfd now. 605 * Otherwise, it was cleaned when masked in the frontend. 606 */ 607 if (proxy->vdev->guest_notifier_mask) { 608 kvm_virtio_pci_irqfd_release(proxy, queue_no, vector); 609 } 610 kvm_virtio_pci_vq_vector_release(proxy, vector); 611 } 612 } 613 614 static int virtio_pci_vq_vector_unmask(VirtIOPCIProxy *proxy, 615 unsigned int queue_no, 616 unsigned int vector, 617 MSIMessage msg) 618 { 619 VirtQueue *vq = virtio_get_queue(proxy->vdev, queue_no); 620 EventNotifier *n = virtio_queue_get_guest_notifier(vq); 621 VirtIOIRQFD *irqfd; 622 int ret = 0; 623 624 if (proxy->vector_irqfd) { 625 irqfd = &proxy->vector_irqfd[vector]; 626 if (irqfd->msg.data != msg.data || irqfd->msg.address != msg.address) { 627 ret = kvm_irqchip_update_msi_route(kvm_state, irqfd->virq, msg); 628 if (ret < 0) { 629 return ret; 630 } 631 } 632 } 633 634 /* If guest supports masking, irqfd is already setup, unmask it. 635 * Otherwise, set it up now. 636 */ 637 if (proxy->vdev->guest_notifier_mask) { 638 proxy->vdev->guest_notifier_mask(proxy->vdev, queue_no, false); 639 /* Test after unmasking to avoid losing events. */ 640 if (proxy->vdev->guest_notifier_pending && 641 proxy->vdev->guest_notifier_pending(proxy->vdev, queue_no)) { 642 event_notifier_set(n); 643 } 644 } else { 645 ret = kvm_virtio_pci_irqfd_use(proxy, queue_no, vector); 646 } 647 return ret; 648 } 649 650 static void virtio_pci_vq_vector_mask(VirtIOPCIProxy *proxy, 651 unsigned int queue_no, 652 unsigned int vector) 653 { 654 /* If guest supports masking, keep irqfd but mask it. 655 * Otherwise, clean it up now. 656 */ 657 if (proxy->vdev->guest_notifier_mask) { 658 proxy->vdev->guest_notifier_mask(proxy->vdev, queue_no, true); 659 } else { 660 kvm_virtio_pci_irqfd_release(proxy, queue_no, vector); 661 } 662 } 663 664 static int virtio_pci_vector_unmask(PCIDevice *dev, unsigned vector, 665 MSIMessage msg) 666 { 667 VirtIOPCIProxy *proxy = container_of(dev, VirtIOPCIProxy, pci_dev); 668 VirtIODevice *vdev = proxy->vdev; 669 int ret, queue_no; 670 671 for (queue_no = 0; queue_no < proxy->nvqs_with_notifiers; queue_no++) { 672 if (!virtio_queue_get_num(vdev, queue_no)) { 673 break; 674 } 675 if (virtio_queue_vector(vdev, queue_no) != vector) { 676 continue; 677 } 678 ret = virtio_pci_vq_vector_unmask(proxy, queue_no, vector, msg); 679 if (ret < 0) { 680 goto undo; 681 } 682 } 683 return 0; 684 685 undo: 686 while (--queue_no >= 0) { 687 if (virtio_queue_vector(vdev, queue_no) != vector) { 688 continue; 689 } 690 virtio_pci_vq_vector_mask(proxy, queue_no, vector); 691 } 692 return ret; 693 } 694 695 static void virtio_pci_vector_mask(PCIDevice *dev, unsigned vector) 696 { 697 VirtIOPCIProxy *proxy = container_of(dev, VirtIOPCIProxy, pci_dev); 698 VirtIODevice *vdev = proxy->vdev; 699 int queue_no; 700 701 for (queue_no = 0; queue_no < proxy->nvqs_with_notifiers; queue_no++) { 702 if (!virtio_queue_get_num(vdev, queue_no)) { 703 break; 704 } 705 if (virtio_queue_vector(vdev, queue_no) != vector) { 706 continue; 707 } 708 virtio_pci_vq_vector_mask(proxy, queue_no, vector); 709 } 710 } 711 712 static void virtio_pci_vector_poll(PCIDevice *dev, 713 unsigned int vector_start, 714 unsigned int vector_end) 715 { 716 VirtIOPCIProxy *proxy = container_of(dev, VirtIOPCIProxy, pci_dev); 717 VirtIODevice *vdev = proxy->vdev; 718 int queue_no; 719 unsigned int vector; 720 EventNotifier *notifier; 721 VirtQueue *vq; 722 723 for (queue_no = 0; queue_no < proxy->nvqs_with_notifiers; queue_no++) { 724 if (!virtio_queue_get_num(vdev, queue_no)) { 725 break; 726 } 727 vector = virtio_queue_vector(vdev, queue_no); 728 if (vector < vector_start || vector >= vector_end || 729 !msix_is_masked(dev, vector)) { 730 continue; 731 } 732 vq = virtio_get_queue(vdev, queue_no); 733 notifier = virtio_queue_get_guest_notifier(vq); 734 if (vdev->guest_notifier_pending) { 735 if (vdev->guest_notifier_pending(vdev, queue_no)) { 736 msix_set_pending(dev, vector); 737 } 738 } else if (event_notifier_test_and_clear(notifier)) { 739 msix_set_pending(dev, vector); 740 } 741 } 742 } 743 744 static int virtio_pci_set_guest_notifier(DeviceState *d, int n, bool assign, 745 bool with_irqfd) 746 { 747 VirtIOPCIProxy *proxy = to_virtio_pci_proxy(d); 748 VirtQueue *vq = virtio_get_queue(proxy->vdev, n); 749 EventNotifier *notifier = virtio_queue_get_guest_notifier(vq); 750 751 if (assign) { 752 int r = event_notifier_init(notifier, 0); 753 if (r < 0) { 754 return r; 755 } 756 virtio_queue_set_guest_notifier_fd_handler(vq, true, with_irqfd); 757 } else { 758 virtio_queue_set_guest_notifier_fd_handler(vq, false, with_irqfd); 759 event_notifier_cleanup(notifier); 760 } 761 762 return 0; 763 } 764 765 static bool virtio_pci_query_guest_notifiers(DeviceState *d) 766 { 767 VirtIOPCIProxy *proxy = to_virtio_pci_proxy(d); 768 return msix_enabled(&proxy->pci_dev); 769 } 770 771 static int virtio_pci_set_guest_notifiers(DeviceState *d, int nvqs, bool assign) 772 { 773 VirtIOPCIProxy *proxy = to_virtio_pci_proxy(d); 774 VirtIODevice *vdev = proxy->vdev; 775 int r, n; 776 bool with_irqfd = msix_enabled(&proxy->pci_dev) && 777 kvm_msi_via_irqfd_enabled(); 778 779 nvqs = MIN(nvqs, VIRTIO_PCI_QUEUE_MAX); 780 781 /* When deassigning, pass a consistent nvqs value 782 * to avoid leaking notifiers. 783 */ 784 assert(assign || nvqs == proxy->nvqs_with_notifiers); 785 786 proxy->nvqs_with_notifiers = nvqs; 787 788 /* Must unset vector notifier while guest notifier is still assigned */ 789 if ((proxy->vector_irqfd || vdev->guest_notifier_mask) && !assign) { 790 msix_unset_vector_notifiers(&proxy->pci_dev); 791 if (proxy->vector_irqfd) { 792 kvm_virtio_pci_vector_release(proxy, nvqs); 793 g_free(proxy->vector_irqfd); 794 proxy->vector_irqfd = NULL; 795 } 796 } 797 798 for (n = 0; n < nvqs; n++) { 799 if (!virtio_queue_get_num(vdev, n)) { 800 break; 801 } 802 803 r = virtio_pci_set_guest_notifier(d, n, assign, 804 kvm_msi_via_irqfd_enabled()); 805 if (r < 0) { 806 goto assign_error; 807 } 808 } 809 810 /* Must set vector notifier after guest notifier has been assigned */ 811 if ((with_irqfd || vdev->guest_notifier_mask) && assign) { 812 if (with_irqfd) { 813 proxy->vector_irqfd = 814 g_malloc0(sizeof(*proxy->vector_irqfd) * 815 msix_nr_vectors_allocated(&proxy->pci_dev)); 816 r = kvm_virtio_pci_vector_use(proxy, nvqs); 817 if (r < 0) { 818 goto assign_error; 819 } 820 } 821 r = msix_set_vector_notifiers(&proxy->pci_dev, 822 virtio_pci_vector_unmask, 823 virtio_pci_vector_mask, 824 virtio_pci_vector_poll); 825 if (r < 0) { 826 goto notifiers_error; 827 } 828 } 829 830 return 0; 831 832 notifiers_error: 833 if (with_irqfd) { 834 assert(assign); 835 kvm_virtio_pci_vector_release(proxy, nvqs); 836 } 837 838 assign_error: 839 /* We get here on assignment failure. Recover by undoing for VQs 0 .. n. */ 840 assert(assign); 841 while (--n >= 0) { 842 virtio_pci_set_guest_notifier(d, n, !assign, with_irqfd); 843 } 844 return r; 845 } 846 847 static int virtio_pci_set_host_notifier(DeviceState *d, int n, bool assign) 848 { 849 VirtIOPCIProxy *proxy = to_virtio_pci_proxy(d); 850 851 /* Stop using ioeventfd for virtqueue kick if the device starts using host 852 * notifiers. This makes it easy to avoid stepping on each others' toes. 853 */ 854 proxy->ioeventfd_disabled = assign; 855 if (assign) { 856 virtio_pci_stop_ioeventfd(proxy); 857 } 858 /* We don't need to start here: it's not needed because backend 859 * currently only stops on status change away from ok, 860 * reset, vmstop and such. If we do add code to start here, 861 * need to check vmstate, device state etc. */ 862 return virtio_pci_set_host_notifier_internal(proxy, n, assign, false); 863 } 864 865 static void virtio_pci_vmstate_change(DeviceState *d, bool running) 866 { 867 VirtIOPCIProxy *proxy = to_virtio_pci_proxy(d); 868 869 if (running) { 870 /* Try to find out if the guest has bus master disabled, but is 871 in ready state. Then we have a buggy guest OS. */ 872 if ((proxy->vdev->status & VIRTIO_CONFIG_S_DRIVER_OK) && 873 !(proxy->pci_dev.config[PCI_COMMAND] & PCI_COMMAND_MASTER)) { 874 proxy->flags |= VIRTIO_PCI_FLAG_BUS_MASTER_BUG; 875 } 876 virtio_pci_start_ioeventfd(proxy); 877 } else { 878 virtio_pci_stop_ioeventfd(proxy); 879 } 880 } 881 882 static const VirtIOBindings virtio_pci_bindings = { 883 .notify = virtio_pci_notify, 884 .save_config = virtio_pci_save_config, 885 .load_config = virtio_pci_load_config, 886 .save_queue = virtio_pci_save_queue, 887 .load_queue = virtio_pci_load_queue, 888 .get_features = virtio_pci_get_features, 889 .query_guest_notifiers = virtio_pci_query_guest_notifiers, 890 .set_host_notifier = virtio_pci_set_host_notifier, 891 .set_guest_notifiers = virtio_pci_set_guest_notifiers, 892 .vmstate_change = virtio_pci_vmstate_change, 893 }; 894 895 void virtio_init_pci(VirtIOPCIProxy *proxy, VirtIODevice *vdev) 896 { 897 uint8_t *config; 898 uint32_t size; 899 900 proxy->vdev = vdev; 901 902 config = proxy->pci_dev.config; 903 904 if (proxy->class_code) { 905 pci_config_set_class(config, proxy->class_code); 906 } 907 pci_set_word(config + PCI_SUBSYSTEM_VENDOR_ID, 908 pci_get_word(config + PCI_VENDOR_ID)); 909 pci_set_word(config + PCI_SUBSYSTEM_ID, vdev->device_id); 910 config[PCI_INTERRUPT_PIN] = 1; 911 912 if (vdev->nvectors && 913 msix_init_exclusive_bar(&proxy->pci_dev, vdev->nvectors, 1)) { 914 vdev->nvectors = 0; 915 } 916 917 proxy->pci_dev.config_write = virtio_write_config; 918 919 size = VIRTIO_PCI_REGION_SIZE(&proxy->pci_dev) + vdev->config_len; 920 if (size & (size-1)) 921 size = 1 << qemu_fls(size); 922 923 memory_region_init_io(&proxy->bar, &virtio_pci_config_ops, proxy, 924 "virtio-pci", size); 925 pci_register_bar(&proxy->pci_dev, 0, PCI_BASE_ADDRESS_SPACE_IO, 926 &proxy->bar); 927 928 if (!kvm_has_many_ioeventfds()) { 929 proxy->flags &= ~VIRTIO_PCI_FLAG_USE_IOEVENTFD; 930 } 931 932 virtio_bind_device(vdev, &virtio_pci_bindings, DEVICE(proxy)); 933 proxy->host_features |= 0x1 << VIRTIO_F_NOTIFY_ON_EMPTY; 934 proxy->host_features |= 0x1 << VIRTIO_F_BAD_FEATURE; 935 proxy->host_features = vdev->get_features(vdev, proxy->host_features); 936 } 937 938 static void virtio_exit_pci(PCIDevice *pci_dev) 939 { 940 VirtIOPCIProxy *proxy = DO_UPCAST(VirtIOPCIProxy, pci_dev, pci_dev); 941 942 memory_region_destroy(&proxy->bar); 943 msix_uninit_exclusive_bar(pci_dev); 944 } 945 946 static int virtio_net_init_pci(PCIDevice *pci_dev) 947 { 948 VirtIOPCIProxy *proxy = DO_UPCAST(VirtIOPCIProxy, pci_dev, pci_dev); 949 VirtIODevice *vdev; 950 951 vdev = virtio_net_init(&pci_dev->qdev, &proxy->nic, &proxy->net, 952 proxy->host_features); 953 954 vdev->nvectors = proxy->nvectors; 955 virtio_init_pci(proxy, vdev); 956 957 /* make the actual value visible */ 958 proxy->nvectors = vdev->nvectors; 959 return 0; 960 } 961 962 static void virtio_net_exit_pci(PCIDevice *pci_dev) 963 { 964 VirtIOPCIProxy *proxy = DO_UPCAST(VirtIOPCIProxy, pci_dev, pci_dev); 965 966 virtio_pci_stop_ioeventfd(proxy); 967 virtio_net_exit(proxy->vdev); 968 virtio_exit_pci(pci_dev); 969 } 970 971 static int virtio_rng_init_pci(PCIDevice *pci_dev) 972 { 973 VirtIOPCIProxy *proxy = DO_UPCAST(VirtIOPCIProxy, pci_dev, pci_dev); 974 VirtIODevice *vdev; 975 976 if (proxy->rng.rng == NULL) { 977 proxy->rng.default_backend = RNG_RANDOM(object_new(TYPE_RNG_RANDOM)); 978 979 object_property_add_child(OBJECT(pci_dev), 980 "default-backend", 981 OBJECT(proxy->rng.default_backend), 982 NULL); 983 984 object_property_set_link(OBJECT(pci_dev), 985 OBJECT(proxy->rng.default_backend), 986 "rng", NULL); 987 } 988 989 vdev = virtio_rng_init(&pci_dev->qdev, &proxy->rng); 990 if (!vdev) { 991 return -1; 992 } 993 virtio_init_pci(proxy, vdev); 994 return 0; 995 } 996 997 static void virtio_rng_exit_pci(PCIDevice *pci_dev) 998 { 999 VirtIOPCIProxy *proxy = DO_UPCAST(VirtIOPCIProxy, pci_dev, pci_dev); 1000 1001 virtio_pci_stop_ioeventfd(proxy); 1002 virtio_rng_exit(proxy->vdev); 1003 virtio_exit_pci(pci_dev); 1004 } 1005 1006 static Property virtio_net_properties[] = { 1007 DEFINE_PROP_BIT("ioeventfd", VirtIOPCIProxy, flags, VIRTIO_PCI_FLAG_USE_IOEVENTFD_BIT, false), 1008 DEFINE_PROP_UINT32("vectors", VirtIOPCIProxy, nvectors, 3), 1009 DEFINE_VIRTIO_NET_FEATURES(VirtIOPCIProxy, host_features), 1010 DEFINE_NIC_PROPERTIES(VirtIOPCIProxy, nic), 1011 DEFINE_PROP_UINT32("x-txtimer", VirtIOPCIProxy, net.txtimer, TX_TIMER_INTERVAL), 1012 DEFINE_PROP_INT32("x-txburst", VirtIOPCIProxy, net.txburst, TX_BURST), 1013 DEFINE_PROP_STRING("tx", VirtIOPCIProxy, net.tx), 1014 DEFINE_PROP_END_OF_LIST(), 1015 }; 1016 1017 static void virtio_net_class_init(ObjectClass *klass, void *data) 1018 { 1019 DeviceClass *dc = DEVICE_CLASS(klass); 1020 PCIDeviceClass *k = PCI_DEVICE_CLASS(klass); 1021 1022 k->init = virtio_net_init_pci; 1023 k->exit = virtio_net_exit_pci; 1024 k->romfile = "efi-virtio.rom"; 1025 k->vendor_id = PCI_VENDOR_ID_REDHAT_QUMRANET; 1026 k->device_id = PCI_DEVICE_ID_VIRTIO_NET; 1027 k->revision = VIRTIO_PCI_ABI_VERSION; 1028 k->class_id = PCI_CLASS_NETWORK_ETHERNET; 1029 dc->reset = virtio_pci_reset; 1030 dc->props = virtio_net_properties; 1031 } 1032 1033 static const TypeInfo virtio_net_info = { 1034 .name = "virtio-net-pci", 1035 .parent = TYPE_PCI_DEVICE, 1036 .instance_size = sizeof(VirtIOPCIProxy), 1037 .class_init = virtio_net_class_init, 1038 }; 1039 1040 static void virtio_rng_initfn(Object *obj) 1041 { 1042 PCIDevice *pci_dev = PCI_DEVICE(obj); 1043 VirtIOPCIProxy *proxy = DO_UPCAST(VirtIOPCIProxy, pci_dev, pci_dev); 1044 1045 object_property_add_link(obj, "rng", TYPE_RNG_BACKEND, 1046 (Object **)&proxy->rng.rng, NULL); 1047 } 1048 1049 static Property virtio_rng_properties[] = { 1050 DEFINE_VIRTIO_COMMON_FEATURES(VirtIOPCIProxy, host_features), 1051 /* Set a default rate limit of 2^47 bytes per minute or roughly 2TB/s. If 1052 you have an entropy source capable of generating more entropy than this 1053 and you can pass it through via virtio-rng, then hats off to you. Until 1054 then, this is unlimited for all practical purposes. 1055 */ 1056 DEFINE_PROP_UINT64("max-bytes", VirtIOPCIProxy, rng.max_bytes, INT64_MAX), 1057 DEFINE_PROP_UINT32("period", VirtIOPCIProxy, rng.period_ms, 1 << 16), 1058 DEFINE_PROP_END_OF_LIST(), 1059 }; 1060 1061 static void virtio_rng_class_init(ObjectClass *klass, void *data) 1062 { 1063 DeviceClass *dc = DEVICE_CLASS(klass); 1064 PCIDeviceClass *k = PCI_DEVICE_CLASS(klass); 1065 1066 k->init = virtio_rng_init_pci; 1067 k->exit = virtio_rng_exit_pci; 1068 k->vendor_id = PCI_VENDOR_ID_REDHAT_QUMRANET; 1069 k->device_id = PCI_DEVICE_ID_VIRTIO_RNG; 1070 k->revision = VIRTIO_PCI_ABI_VERSION; 1071 k->class_id = PCI_CLASS_OTHERS; 1072 dc->reset = virtio_pci_reset; 1073 dc->props = virtio_rng_properties; 1074 } 1075 1076 static const TypeInfo virtio_rng_info = { 1077 .name = "virtio-rng-pci", 1078 .parent = TYPE_PCI_DEVICE, 1079 .instance_size = sizeof(VirtIOPCIProxy), 1080 .instance_init = virtio_rng_initfn, 1081 .class_init = virtio_rng_class_init, 1082 }; 1083 1084 #ifdef CONFIG_VIRTFS 1085 static int virtio_9p_init_pci(PCIDevice *pci_dev) 1086 { 1087 VirtIOPCIProxy *proxy = DO_UPCAST(VirtIOPCIProxy, pci_dev, pci_dev); 1088 VirtIODevice *vdev; 1089 1090 vdev = virtio_9p_init(&pci_dev->qdev, &proxy->fsconf); 1091 vdev->nvectors = proxy->nvectors; 1092 virtio_init_pci(proxy, vdev); 1093 /* make the actual value visible */ 1094 proxy->nvectors = vdev->nvectors; 1095 return 0; 1096 } 1097 1098 static Property virtio_9p_properties[] = { 1099 DEFINE_PROP_BIT("ioeventfd", VirtIOPCIProxy, flags, VIRTIO_PCI_FLAG_USE_IOEVENTFD_BIT, true), 1100 DEFINE_PROP_UINT32("vectors", VirtIOPCIProxy, nvectors, 2), 1101 DEFINE_VIRTIO_COMMON_FEATURES(VirtIOPCIProxy, host_features), 1102 DEFINE_PROP_STRING("mount_tag", VirtIOPCIProxy, fsconf.tag), 1103 DEFINE_PROP_STRING("fsdev", VirtIOPCIProxy, fsconf.fsdev_id), 1104 DEFINE_PROP_END_OF_LIST(), 1105 }; 1106 1107 static void virtio_9p_class_init(ObjectClass *klass, void *data) 1108 { 1109 DeviceClass *dc = DEVICE_CLASS(klass); 1110 PCIDeviceClass *k = PCI_DEVICE_CLASS(klass); 1111 1112 k->init = virtio_9p_init_pci; 1113 k->vendor_id = PCI_VENDOR_ID_REDHAT_QUMRANET; 1114 k->device_id = PCI_DEVICE_ID_VIRTIO_9P; 1115 k->revision = VIRTIO_PCI_ABI_VERSION; 1116 k->class_id = 0x2; 1117 dc->props = virtio_9p_properties; 1118 dc->reset = virtio_pci_reset; 1119 } 1120 1121 static const TypeInfo virtio_9p_info = { 1122 .name = "virtio-9p-pci", 1123 .parent = TYPE_PCI_DEVICE, 1124 .instance_size = sizeof(VirtIOPCIProxy), 1125 .class_init = virtio_9p_class_init, 1126 }; 1127 #endif 1128 1129 /* 1130 * virtio-pci: This is the PCIDevice which has a virtio-pci-bus. 1131 */ 1132 1133 /* This is called by virtio-bus just after the device is plugged. */ 1134 static void virtio_pci_device_plugged(DeviceState *d) 1135 { 1136 VirtIOPCIProxy *proxy = VIRTIO_PCI(d); 1137 VirtioBusState *bus = &proxy->bus; 1138 uint8_t *config; 1139 uint32_t size; 1140 1141 proxy->vdev = bus->vdev; 1142 1143 config = proxy->pci_dev.config; 1144 if (proxy->class_code) { 1145 pci_config_set_class(config, proxy->class_code); 1146 } 1147 pci_set_word(config + PCI_SUBSYSTEM_VENDOR_ID, 1148 pci_get_word(config + PCI_VENDOR_ID)); 1149 pci_set_word(config + PCI_SUBSYSTEM_ID, virtio_bus_get_vdev_id(bus)); 1150 config[PCI_INTERRUPT_PIN] = 1; 1151 1152 if (proxy->nvectors && 1153 msix_init_exclusive_bar(&proxy->pci_dev, proxy->nvectors, 1)) { 1154 proxy->nvectors = 0; 1155 } 1156 1157 proxy->pci_dev.config_write = virtio_write_config; 1158 1159 size = VIRTIO_PCI_REGION_SIZE(&proxy->pci_dev) 1160 + virtio_bus_get_vdev_config_len(bus); 1161 if (size & (size - 1)) { 1162 size = 1 << qemu_fls(size); 1163 } 1164 1165 memory_region_init_io(&proxy->bar, &virtio_pci_config_ops, proxy, 1166 "virtio-pci", size); 1167 pci_register_bar(&proxy->pci_dev, 0, PCI_BASE_ADDRESS_SPACE_IO, 1168 &proxy->bar); 1169 1170 if (!kvm_has_many_ioeventfds()) { 1171 proxy->flags &= ~VIRTIO_PCI_FLAG_USE_IOEVENTFD; 1172 } 1173 1174 proxy->host_features |= 0x1 << VIRTIO_F_NOTIFY_ON_EMPTY; 1175 proxy->host_features |= 0x1 << VIRTIO_F_BAD_FEATURE; 1176 proxy->host_features = virtio_bus_get_vdev_features(bus, 1177 proxy->host_features); 1178 } 1179 1180 static int virtio_pci_init(PCIDevice *pci_dev) 1181 { 1182 VirtIOPCIProxy *dev = VIRTIO_PCI(pci_dev); 1183 VirtioPCIClass *k = VIRTIO_PCI_GET_CLASS(pci_dev); 1184 virtio_pci_bus_new(&dev->bus, dev); 1185 if (k->init != NULL) { 1186 return k->init(dev); 1187 } 1188 return 0; 1189 } 1190 1191 static void virtio_pci_exit(PCIDevice *pci_dev) 1192 { 1193 VirtIOPCIProxy *proxy = VIRTIO_PCI(pci_dev); 1194 virtio_pci_stop_ioeventfd(proxy); 1195 virtio_exit_pci(pci_dev); 1196 } 1197 1198 /* 1199 * This will be renamed virtio_pci_reset at the end of the series. 1200 * virtio_pci_reset is still in use at this moment. 1201 */ 1202 static void virtio_pci_rst(DeviceState *qdev) 1203 { 1204 VirtIOPCIProxy *proxy = VIRTIO_PCI(qdev); 1205 VirtioBusState *bus = VIRTIO_BUS(&proxy->bus); 1206 virtio_pci_stop_ioeventfd(proxy); 1207 virtio_bus_reset(bus); 1208 msix_unuse_all_vectors(&proxy->pci_dev); 1209 proxy->flags &= ~VIRTIO_PCI_FLAG_BUS_MASTER_BUG; 1210 } 1211 1212 static void virtio_pci_class_init(ObjectClass *klass, void *data) 1213 { 1214 DeviceClass *dc = DEVICE_CLASS(klass); 1215 PCIDeviceClass *k = PCI_DEVICE_CLASS(klass); 1216 1217 k->init = virtio_pci_init; 1218 k->exit = virtio_pci_exit; 1219 k->vendor_id = PCI_VENDOR_ID_REDHAT_QUMRANET; 1220 k->revision = VIRTIO_PCI_ABI_VERSION; 1221 k->class_id = PCI_CLASS_OTHERS; 1222 dc->reset = virtio_pci_rst; 1223 } 1224 1225 static const TypeInfo virtio_pci_info = { 1226 .name = TYPE_VIRTIO_PCI, 1227 .parent = TYPE_PCI_DEVICE, 1228 .instance_size = sizeof(VirtIOPCIProxy), 1229 .class_init = virtio_pci_class_init, 1230 .class_size = sizeof(VirtioPCIClass), 1231 .abstract = true, 1232 }; 1233 1234 /* virtio-blk-pci */ 1235 1236 static Property virtio_blk_pci_properties[] = { 1237 DEFINE_PROP_HEX32("class", VirtIOPCIProxy, class_code, 0), 1238 DEFINE_PROP_BIT("ioeventfd", VirtIOPCIProxy, flags, 1239 VIRTIO_PCI_FLAG_USE_IOEVENTFD_BIT, true), 1240 DEFINE_PROP_UINT32("vectors", VirtIOPCIProxy, nvectors, 2), 1241 #ifdef CONFIG_VIRTIO_BLK_DATA_PLANE 1242 DEFINE_PROP_BIT("x-data-plane", VirtIOBlkPCI, blk.data_plane, 0, false), 1243 #endif 1244 DEFINE_VIRTIO_BLK_FEATURES(VirtIOPCIProxy, host_features), 1245 DEFINE_VIRTIO_BLK_PROPERTIES(VirtIOBlkPCI, blk), 1246 DEFINE_PROP_END_OF_LIST(), 1247 }; 1248 1249 static int virtio_blk_pci_init(VirtIOPCIProxy *vpci_dev) 1250 { 1251 VirtIOBlkPCI *dev = VIRTIO_BLK_PCI(vpci_dev); 1252 DeviceState *vdev = DEVICE(&dev->vdev); 1253 virtio_blk_set_conf(vdev, &(dev->blk)); 1254 qdev_set_parent_bus(vdev, BUS(&vpci_dev->bus)); 1255 if (qdev_init(vdev) < 0) { 1256 return -1; 1257 } 1258 return 0; 1259 } 1260 1261 static void virtio_blk_pci_class_init(ObjectClass *klass, void *data) 1262 { 1263 DeviceClass *dc = DEVICE_CLASS(klass); 1264 VirtioPCIClass *k = VIRTIO_PCI_CLASS(klass); 1265 PCIDeviceClass *pcidev_k = PCI_DEVICE_CLASS(klass); 1266 1267 dc->props = virtio_blk_pci_properties; 1268 k->init = virtio_blk_pci_init; 1269 pcidev_k->vendor_id = PCI_VENDOR_ID_REDHAT_QUMRANET; 1270 pcidev_k->device_id = PCI_DEVICE_ID_VIRTIO_BLOCK; 1271 pcidev_k->revision = VIRTIO_PCI_ABI_VERSION; 1272 pcidev_k->class_id = PCI_CLASS_STORAGE_SCSI; 1273 } 1274 1275 static void virtio_blk_pci_instance_init(Object *obj) 1276 { 1277 VirtIOBlkPCI *dev = VIRTIO_BLK_PCI(obj); 1278 object_initialize(OBJECT(&dev->vdev), TYPE_VIRTIO_BLK); 1279 object_property_add_child(obj, "virtio-backend", OBJECT(&dev->vdev), NULL); 1280 } 1281 1282 static const TypeInfo virtio_blk_pci_info = { 1283 .name = TYPE_VIRTIO_BLK_PCI, 1284 .parent = TYPE_VIRTIO_PCI, 1285 .instance_size = sizeof(VirtIOBlkPCI), 1286 .instance_init = virtio_blk_pci_instance_init, 1287 .class_init = virtio_blk_pci_class_init, 1288 }; 1289 1290 /* virtio-scsi-pci */ 1291 1292 static Property virtio_scsi_pci_properties[] = { 1293 DEFINE_PROP_BIT("ioeventfd", VirtIOPCIProxy, flags, 1294 VIRTIO_PCI_FLAG_USE_IOEVENTFD_BIT, true), 1295 DEFINE_PROP_UINT32("vectors", VirtIOPCIProxy, nvectors, 1296 DEV_NVECTORS_UNSPECIFIED), 1297 DEFINE_VIRTIO_SCSI_FEATURES(VirtIOPCIProxy, host_features), 1298 DEFINE_VIRTIO_SCSI_PROPERTIES(VirtIOSCSIPCI, vdev.conf), 1299 DEFINE_PROP_END_OF_LIST(), 1300 }; 1301 1302 static int virtio_scsi_pci_init_pci(VirtIOPCIProxy *vpci_dev) 1303 { 1304 VirtIOSCSIPCI *dev = VIRTIO_SCSI_PCI(vpci_dev); 1305 DeviceState *vdev = DEVICE(&dev->vdev); 1306 1307 if (vpci_dev->nvectors == DEV_NVECTORS_UNSPECIFIED) { 1308 vpci_dev->nvectors = dev->vdev.conf.num_queues + 3; 1309 } 1310 1311 qdev_set_parent_bus(vdev, BUS(&vpci_dev->bus)); 1312 if (qdev_init(vdev) < 0) { 1313 return -1; 1314 } 1315 return 0; 1316 } 1317 1318 static void virtio_scsi_pci_class_init(ObjectClass *klass, void *data) 1319 { 1320 DeviceClass *dc = DEVICE_CLASS(klass); 1321 VirtioPCIClass *k = VIRTIO_PCI_CLASS(klass); 1322 PCIDeviceClass *pcidev_k = PCI_DEVICE_CLASS(klass); 1323 k->init = virtio_scsi_pci_init_pci; 1324 dc->props = virtio_scsi_pci_properties; 1325 pcidev_k->vendor_id = PCI_VENDOR_ID_REDHAT_QUMRANET; 1326 pcidev_k->device_id = PCI_DEVICE_ID_VIRTIO_SCSI; 1327 pcidev_k->revision = 0x00; 1328 pcidev_k->class_id = PCI_CLASS_STORAGE_SCSI; 1329 } 1330 1331 static void virtio_scsi_pci_instance_init(Object *obj) 1332 { 1333 VirtIOSCSIPCI *dev = VIRTIO_SCSI_PCI(obj); 1334 object_initialize(OBJECT(&dev->vdev), TYPE_VIRTIO_SCSI); 1335 object_property_add_child(obj, "virtio-backend", OBJECT(&dev->vdev), NULL); 1336 } 1337 1338 static const TypeInfo virtio_scsi_pci_info = { 1339 .name = TYPE_VIRTIO_SCSI_PCI, 1340 .parent = TYPE_VIRTIO_PCI, 1341 .instance_size = sizeof(VirtIOSCSIPCI), 1342 .instance_init = virtio_scsi_pci_instance_init, 1343 .class_init = virtio_scsi_pci_class_init, 1344 }; 1345 1346 /* virtio-balloon-pci */ 1347 1348 static void balloon_pci_stats_get_all(Object *obj, struct Visitor *v, 1349 void *opaque, const char *name, 1350 Error **errp) 1351 { 1352 VirtIOBalloonPCI *dev = opaque; 1353 object_property_get(OBJECT(&dev->vdev), v, "guest-stats", errp); 1354 } 1355 1356 static void balloon_pci_stats_get_poll_interval(Object *obj, struct Visitor *v, 1357 void *opaque, const char *name, 1358 Error **errp) 1359 { 1360 VirtIOBalloonPCI *dev = opaque; 1361 object_property_get(OBJECT(&dev->vdev), v, "guest-stats-polling-interval", 1362 errp); 1363 } 1364 1365 static void balloon_pci_stats_set_poll_interval(Object *obj, struct Visitor *v, 1366 void *opaque, const char *name, 1367 Error **errp) 1368 { 1369 VirtIOBalloonPCI *dev = opaque; 1370 object_property_set(OBJECT(&dev->vdev), v, "guest-stats-polling-interval", 1371 errp); 1372 } 1373 1374 static Property virtio_balloon_pci_properties[] = { 1375 DEFINE_VIRTIO_COMMON_FEATURES(VirtIOPCIProxy, host_features), 1376 DEFINE_PROP_HEX32("class", VirtIOPCIProxy, class_code, 0), 1377 DEFINE_PROP_END_OF_LIST(), 1378 }; 1379 1380 static int virtio_balloon_pci_init(VirtIOPCIProxy *vpci_dev) 1381 { 1382 VirtIOBalloonPCI *dev = VIRTIO_BALLOON_PCI(vpci_dev); 1383 DeviceState *vdev = DEVICE(&dev->vdev); 1384 1385 if (vpci_dev->class_code != PCI_CLASS_OTHERS && 1386 vpci_dev->class_code != PCI_CLASS_MEMORY_RAM) { /* qemu < 1.1 */ 1387 vpci_dev->class_code = PCI_CLASS_OTHERS; 1388 } 1389 1390 qdev_set_parent_bus(vdev, BUS(&vpci_dev->bus)); 1391 if (qdev_init(vdev) < 0) { 1392 return -1; 1393 } 1394 return 0; 1395 } 1396 1397 static void virtio_balloon_pci_class_init(ObjectClass *klass, void *data) 1398 { 1399 DeviceClass *dc = DEVICE_CLASS(klass); 1400 VirtioPCIClass *k = VIRTIO_PCI_CLASS(klass); 1401 PCIDeviceClass *pcidev_k = PCI_DEVICE_CLASS(klass); 1402 k->init = virtio_balloon_pci_init; 1403 dc->props = virtio_balloon_pci_properties; 1404 pcidev_k->vendor_id = PCI_VENDOR_ID_REDHAT_QUMRANET; 1405 pcidev_k->device_id = PCI_DEVICE_ID_VIRTIO_BALLOON; 1406 pcidev_k->revision = VIRTIO_PCI_ABI_VERSION; 1407 pcidev_k->class_id = PCI_CLASS_OTHERS; 1408 } 1409 1410 static void virtio_balloon_pci_instance_init(Object *obj) 1411 { 1412 VirtIOBalloonPCI *dev = VIRTIO_BALLOON_PCI(obj); 1413 object_initialize(OBJECT(&dev->vdev), TYPE_VIRTIO_BALLOON); 1414 object_property_add_child(obj, "virtio-backend", OBJECT(&dev->vdev), NULL); 1415 1416 object_property_add(obj, "guest-stats", "guest statistics", 1417 balloon_pci_stats_get_all, NULL, NULL, dev, 1418 NULL); 1419 1420 object_property_add(obj, "guest-stats-polling-interval", "int", 1421 balloon_pci_stats_get_poll_interval, 1422 balloon_pci_stats_set_poll_interval, 1423 NULL, dev, NULL); 1424 } 1425 1426 static const TypeInfo virtio_balloon_pci_info = { 1427 .name = TYPE_VIRTIO_BALLOON_PCI, 1428 .parent = TYPE_VIRTIO_PCI, 1429 .instance_size = sizeof(VirtIOBalloonPCI), 1430 .instance_init = virtio_balloon_pci_instance_init, 1431 .class_init = virtio_balloon_pci_class_init, 1432 }; 1433 1434 /* virtio-serial-pci */ 1435 1436 static int virtio_serial_pci_init(VirtIOPCIProxy *vpci_dev) 1437 { 1438 VirtIOSerialPCI *dev = VIRTIO_SERIAL_PCI(vpci_dev); 1439 DeviceState *vdev = DEVICE(&dev->vdev); 1440 1441 if (vpci_dev->class_code != PCI_CLASS_COMMUNICATION_OTHER && 1442 vpci_dev->class_code != PCI_CLASS_DISPLAY_OTHER && /* qemu 0.10 */ 1443 vpci_dev->class_code != PCI_CLASS_OTHERS) { /* qemu-kvm */ 1444 vpci_dev->class_code = PCI_CLASS_COMMUNICATION_OTHER; 1445 } 1446 1447 /* backwards-compatibility with machines that were created with 1448 DEV_NVECTORS_UNSPECIFIED */ 1449 if (vpci_dev->nvectors == DEV_NVECTORS_UNSPECIFIED) { 1450 vpci_dev->nvectors = dev->vdev.serial.max_virtserial_ports + 1; 1451 } 1452 1453 qdev_set_parent_bus(vdev, BUS(&vpci_dev->bus)); 1454 if (qdev_init(vdev) < 0) { 1455 return -1; 1456 } 1457 return 0; 1458 } 1459 1460 static Property virtio_serial_pci_properties[] = { 1461 DEFINE_PROP_BIT("ioeventfd", VirtIOPCIProxy, flags, 1462 VIRTIO_PCI_FLAG_USE_IOEVENTFD_BIT, true), 1463 DEFINE_PROP_UINT32("vectors", VirtIOPCIProxy, nvectors, 2), 1464 DEFINE_PROP_HEX32("class", VirtIOPCIProxy, class_code, 0), 1465 DEFINE_VIRTIO_COMMON_FEATURES(VirtIOPCIProxy, host_features), 1466 DEFINE_VIRTIO_SERIAL_PROPERTIES(VirtIOSerialPCI, vdev.serial), 1467 DEFINE_PROP_END_OF_LIST(), 1468 }; 1469 1470 static void virtio_serial_pci_class_init(ObjectClass *klass, void *data) 1471 { 1472 DeviceClass *dc = DEVICE_CLASS(klass); 1473 VirtioPCIClass *k = VIRTIO_PCI_CLASS(klass); 1474 PCIDeviceClass *pcidev_k = PCI_DEVICE_CLASS(klass); 1475 k->init = virtio_serial_pci_init; 1476 dc->props = virtio_serial_pci_properties; 1477 pcidev_k->vendor_id = PCI_VENDOR_ID_REDHAT_QUMRANET; 1478 pcidev_k->device_id = PCI_DEVICE_ID_VIRTIO_CONSOLE; 1479 pcidev_k->revision = VIRTIO_PCI_ABI_VERSION; 1480 pcidev_k->class_id = PCI_CLASS_COMMUNICATION_OTHER; 1481 } 1482 1483 static void virtio_serial_pci_instance_init(Object *obj) 1484 { 1485 VirtIOSerialPCI *dev = VIRTIO_SERIAL_PCI(obj); 1486 object_initialize(OBJECT(&dev->vdev), TYPE_VIRTIO_SERIAL); 1487 object_property_add_child(obj, "virtio-backend", OBJECT(&dev->vdev), NULL); 1488 } 1489 1490 static const TypeInfo virtio_serial_pci_info = { 1491 .name = TYPE_VIRTIO_SERIAL_PCI, 1492 .parent = TYPE_VIRTIO_PCI, 1493 .instance_size = sizeof(VirtIOSerialPCI), 1494 .instance_init = virtio_serial_pci_instance_init, 1495 .class_init = virtio_serial_pci_class_init, 1496 }; 1497 1498 /* virtio-pci-bus */ 1499 1500 void virtio_pci_bus_new(VirtioBusState *bus, VirtIOPCIProxy *dev) 1501 { 1502 DeviceState *qdev = DEVICE(dev); 1503 BusState *qbus; 1504 qbus_create_inplace((BusState *)bus, TYPE_VIRTIO_PCI_BUS, qdev, NULL); 1505 qbus = BUS(bus); 1506 qbus->allow_hotplug = 1; 1507 } 1508 1509 static void virtio_pci_bus_class_init(ObjectClass *klass, void *data) 1510 { 1511 BusClass *bus_class = BUS_CLASS(klass); 1512 VirtioBusClass *k = VIRTIO_BUS_CLASS(klass); 1513 bus_class->max_dev = 1; 1514 k->notify = virtio_pci_notify; 1515 k->save_config = virtio_pci_save_config; 1516 k->load_config = virtio_pci_load_config; 1517 k->save_queue = virtio_pci_save_queue; 1518 k->load_queue = virtio_pci_load_queue; 1519 k->get_features = virtio_pci_get_features; 1520 k->query_guest_notifiers = virtio_pci_query_guest_notifiers; 1521 k->set_host_notifier = virtio_pci_set_host_notifier; 1522 k->set_guest_notifiers = virtio_pci_set_guest_notifiers; 1523 k->vmstate_change = virtio_pci_vmstate_change; 1524 k->device_plugged = virtio_pci_device_plugged; 1525 } 1526 1527 static const TypeInfo virtio_pci_bus_info = { 1528 .name = TYPE_VIRTIO_PCI_BUS, 1529 .parent = TYPE_VIRTIO_BUS, 1530 .instance_size = sizeof(VirtioPCIBusState), 1531 .class_init = virtio_pci_bus_class_init, 1532 }; 1533 1534 static void virtio_pci_register_types(void) 1535 { 1536 type_register_static(&virtio_net_info); 1537 type_register_static(&virtio_rng_info); 1538 type_register_static(&virtio_pci_bus_info); 1539 type_register_static(&virtio_pci_info); 1540 #ifdef CONFIG_VIRTFS 1541 type_register_static(&virtio_9p_info); 1542 #endif 1543 type_register_static(&virtio_blk_pci_info); 1544 type_register_static(&virtio_scsi_pci_info); 1545 type_register_static(&virtio_balloon_pci_info); 1546 type_register_static(&virtio_serial_pci_info); 1547 } 1548 1549 type_init(virtio_pci_register_types) 1550