1 #ifndef VHOST_H 2 #define VHOST_H 3 4 #include "net/vhost_net.h" 5 #include "hw/virtio/vhost-backend.h" 6 #include "hw/virtio/virtio.h" 7 #include "system/memory.h" 8 9 #define VHOST_F_DEVICE_IOTLB 63 10 #define VHOST_USER_F_PROTOCOL_FEATURES 30 11 12 #define VU_REALIZE_CONN_RETRIES 3 13 14 /* Generic structures common for any vhost based device. */ 15 16 struct vhost_inflight { 17 int fd; 18 void *addr; 19 uint64_t size; 20 uint64_t offset; 21 uint16_t queue_size; 22 }; 23 24 struct vhost_virtqueue { 25 int kick; 26 int call; 27 void *desc; 28 void *avail; 29 void *used; 30 int num; 31 unsigned long long desc_phys; 32 unsigned desc_size; 33 unsigned long long avail_phys; 34 unsigned avail_size; 35 unsigned long long used_phys; 36 unsigned used_size; 37 EventNotifier masked_notifier; 38 EventNotifier error_notifier; 39 EventNotifier masked_config_notifier; 40 struct vhost_dev *dev; 41 }; 42 43 typedef unsigned long vhost_log_chunk_t; 44 #define VHOST_LOG_PAGE 0x1000 45 #define VHOST_LOG_BITS (8 * sizeof(vhost_log_chunk_t)) 46 #define VHOST_LOG_CHUNK (VHOST_LOG_PAGE * VHOST_LOG_BITS) 47 #define VHOST_INVALID_FEATURE_BIT (0xff) 48 #define VHOST_QUEUE_NUM_CONFIG_INR 0 49 50 struct vhost_log { 51 unsigned long long size; 52 int refcnt; 53 int fd; 54 vhost_log_chunk_t *log; 55 }; 56 57 struct vhost_dev; 58 struct vhost_iommu { 59 struct vhost_dev *hdev; 60 MemoryRegion *mr; 61 hwaddr iommu_offset; 62 IOMMUNotifier n; 63 QLIST_ENTRY(vhost_iommu) iommu_next; 64 }; 65 66 typedef struct VhostDevConfigOps { 67 /* Vhost device config space changed callback 68 */ 69 int (*vhost_dev_config_notifier)(struct vhost_dev *dev); 70 } VhostDevConfigOps; 71 72 struct vhost_memory; 73 74 /** 75 * struct vhost_dev - common vhost_dev structure 76 * @vhost_ops: backend specific ops 77 * @config_ops: ops for config changes (see @vhost_dev_set_config_notifier) 78 */ 79 struct vhost_dev { 80 VirtIODevice *vdev; 81 MemoryListener memory_listener; 82 MemoryListener iommu_listener; 83 struct vhost_memory *mem; 84 int n_mem_sections; 85 MemoryRegionSection *mem_sections; 86 int n_tmp_sections; 87 MemoryRegionSection *tmp_sections; 88 struct vhost_virtqueue *vqs; 89 unsigned int nvqs; 90 /* the first virtqueue which would be used by this vhost dev */ 91 int vq_index; 92 /* one past the last vq index for the virtio device (not vhost) */ 93 int vq_index_end; 94 /* if non-zero, minimum required value for max_queues */ 95 int num_queues; 96 /** 97 * vhost feature handling requires matching the feature set 98 * offered by a backend which may be a subset of the total 99 * features eventually offered to the guest. 100 * 101 * @features: available features provided by the backend 102 * @acked_features: final negotiated features with front-end driver 103 * 104 * @backend_features: this is used in a couple of places to either 105 * store VHOST_USER_F_PROTOCOL_FEATURES to apply to 106 * VHOST_USER_SET_FEATURES or VHOST_NET_F_VIRTIO_NET_HDR. Its 107 * future use should be discouraged and the variable retired as 108 * its easy to confuse with the VirtIO backend_features. 109 */ 110 uint64_t features; 111 uint64_t acked_features; 112 uint64_t backend_features; 113 114 /** 115 * @protocol_features: is the vhost-user only feature set by 116 * VHOST_USER_SET_PROTOCOL_FEATURES. Protocol features are only 117 * negotiated if VHOST_USER_F_PROTOCOL_FEATURES has been offered 118 * by the backend (see @features). 119 */ 120 uint64_t protocol_features; 121 122 uint64_t max_queues; 123 uint64_t backend_cap; 124 /* @started: is the vhost device started? */ 125 bool started; 126 bool log_enabled; 127 uint64_t log_size; 128 Error *migration_blocker; 129 const VhostOps *vhost_ops; 130 void *opaque; 131 struct vhost_log *log; 132 QLIST_ENTRY(vhost_dev) entry; 133 QLIST_ENTRY(vhost_dev) logdev_entry; 134 QLIST_HEAD(, vhost_iommu) iommu_list; 135 IOMMUNotifier n; 136 const VhostDevConfigOps *config_ops; 137 }; 138 139 extern const VhostOps kernel_ops; 140 extern const VhostOps user_ops; 141 extern const VhostOps vdpa_ops; 142 143 struct vhost_net { 144 struct vhost_dev dev; 145 struct vhost_virtqueue vqs[2]; 146 int backend; 147 const int *feature_bits; 148 int max_tx_queue_size; 149 SaveAcketFeatures *save_acked_features; 150 bool is_vhost_user; 151 NetClientState *nc; 152 }; 153 154 /** 155 * vhost_dev_init() - initialise the vhost interface 156 * @hdev: the common vhost_dev structure 157 * @opaque: opaque ptr passed to backend (vhost/vhost-user/vdpa) 158 * @backend_type: type of backend 159 * @busyloop_timeout: timeout for polling virtqueue 160 * @errp: error handle 161 * 162 * The initialisation of the vhost device will trigger the 163 * initialisation of the backend and potentially capability 164 * negotiation of backend interface. Configuration of the VirtIO 165 * itself won't happen until the interface is started. 166 * 167 * Return: 0 on success, non-zero on error while setting errp. 168 */ 169 int vhost_dev_init(struct vhost_dev *hdev, void *opaque, 170 VhostBackendType backend_type, 171 uint32_t busyloop_timeout, Error **errp); 172 173 /** 174 * vhost_dev_cleanup() - tear down and cleanup vhost interface 175 * @hdev: the common vhost_dev structure 176 */ 177 void vhost_dev_cleanup(struct vhost_dev *hdev); 178 179 void vhost_dev_disable_notifiers_nvqs(struct vhost_dev *hdev, 180 VirtIODevice *vdev, 181 unsigned int nvqs); 182 183 /** 184 * vhost_dev_enable_notifiers() - enable event notifiers 185 * @hdev: common vhost_dev structure 186 * @vdev: the VirtIODevice structure 187 * 188 * Enable notifications directly to the vhost device rather than being 189 * triggered by QEMU itself. Notifications should be enabled before 190 * the vhost device is started via @vhost_dev_start. 191 * 192 * Return: 0 on success, < 0 on error. 193 */ 194 int vhost_dev_enable_notifiers(struct vhost_dev *hdev, VirtIODevice *vdev); 195 196 /** 197 * vhost_dev_disable_notifiers - disable event notifications 198 * @hdev: common vhost_dev structure 199 * @vdev: the VirtIODevice structure 200 * 201 * Disable direct notifications to vhost device. 202 */ 203 void vhost_dev_disable_notifiers(struct vhost_dev *hdev, VirtIODevice *vdev); 204 bool vhost_config_pending(struct vhost_dev *hdev); 205 void vhost_config_mask(struct vhost_dev *hdev, VirtIODevice *vdev, bool mask); 206 207 /** 208 * vhost_dev_is_started() - report status of vhost device 209 * @hdev: common vhost_dev structure 210 * 211 * Return the started status of the vhost device 212 */ 213 static inline bool vhost_dev_is_started(struct vhost_dev *hdev) 214 { 215 return hdev->started; 216 } 217 218 /** 219 * vhost_dev_start() - start the vhost device 220 * @hdev: common vhost_dev structure 221 * @vdev: the VirtIODevice structure 222 * @vrings: true to have vrings enabled in this call 223 * 224 * Starts the vhost device. From this point VirtIO feature negotiation 225 * can start and the device can start processing VirtIO transactions. 226 * 227 * Return: 0 on success, < 0 on error. 228 */ 229 int vhost_dev_start(struct vhost_dev *hdev, VirtIODevice *vdev, bool vrings); 230 231 /** 232 * vhost_dev_stop() - stop the vhost device 233 * @hdev: common vhost_dev structure 234 * @vdev: the VirtIODevice structure 235 * @vrings: true to have vrings disabled in this call 236 * 237 * Stop the vhost device. After the device is stopped the notifiers 238 * can be disabled (@vhost_dev_disable_notifiers) and the device can 239 * be torn down (@vhost_dev_cleanup). 240 * 241 * Return: 0 on success, != 0 on error when stopping dev. 242 */ 243 int vhost_dev_stop(struct vhost_dev *hdev, VirtIODevice *vdev, bool vrings); 244 245 /** 246 * vhost_dev_force_stop() - force stop the vhost device 247 * @hdev: common vhost_dev structure 248 * @vdev: the VirtIODevice structure 249 * @vrings: true to have vrings disabled in this call 250 * 251 * Force stop the vhost device. After the device is stopped the notifiers 252 * can be disabled (@vhost_dev_disable_notifiers) and the device can 253 * be torn down (@vhost_dev_cleanup). Unlike @vhost_dev_stop, this doesn't 254 * attempt to flush in-flight backend requests by skipping GET_VRING_BASE 255 * entirely. 256 */ 257 int vhost_dev_force_stop(struct vhost_dev *hdev, VirtIODevice *vdev, 258 bool vrings); 259 260 /** 261 * DOC: vhost device configuration handling 262 * 263 * The VirtIO device configuration space is used for rarely changing 264 * or initialisation time parameters. The configuration can be updated 265 * by either the guest driver or the device itself. If the device can 266 * change the configuration over time the vhost handler should 267 * register a @VhostDevConfigOps structure with 268 * @vhost_dev_set_config_notifier so the guest can be notified. Some 269 * devices register a handler anyway and will signal an error if an 270 * unexpected config change happens. 271 */ 272 273 /** 274 * vhost_dev_get_config() - fetch device configuration 275 * @hdev: common vhost_dev_structure 276 * @config: pointer to device appropriate config structure 277 * @config_len: size of device appropriate config structure 278 * 279 * Return: 0 on success, < 0 on error while setting errp 280 */ 281 int vhost_dev_get_config(struct vhost_dev *hdev, uint8_t *config, 282 uint32_t config_len, Error **errp); 283 284 /** 285 * vhost_dev_set_config() - set device configuration 286 * @hdev: common vhost_dev_structure 287 * @data: pointer to data to set 288 * @offset: offset into configuration space 289 * @size: length of set 290 * @flags: @VhostSetConfigType flags 291 * 292 * By use of @offset/@size a subset of the configuration space can be 293 * written to. The @flags are used to indicate if it is a normal 294 * transaction or related to migration. 295 * 296 * Return: 0 on success, non-zero on error 297 */ 298 int vhost_dev_set_config(struct vhost_dev *dev, const uint8_t *data, 299 uint32_t offset, uint32_t size, uint32_t flags); 300 301 /** 302 * vhost_dev_set_config_notifier() - register VhostDevConfigOps 303 * @hdev: common vhost_dev_structure 304 * @ops: notifier ops 305 * 306 * If the device is expected to change configuration a notifier can be 307 * setup to handle the case. 308 */ 309 void vhost_dev_set_config_notifier(struct vhost_dev *dev, 310 const VhostDevConfigOps *ops); 311 312 313 /* Test and clear masked event pending status. 314 * Should be called after unmask to avoid losing events. 315 */ 316 bool vhost_virtqueue_pending(struct vhost_dev *hdev, int n); 317 318 /* Mask/unmask events from this vq. 319 */ 320 void vhost_virtqueue_mask(struct vhost_dev *hdev, VirtIODevice *vdev, int n, 321 bool mask); 322 323 /** 324 * vhost_get_features() - return a sanitised set of feature bits 325 * @hdev: common vhost_dev structure 326 * @feature_bits: pointer to terminated table of feature bits 327 * @features: original feature set 328 * 329 * This returns a set of features bits that is an intersection of what 330 * is supported by the vhost backend (hdev->features), the supported 331 * feature_bits and the requested feature set. 332 */ 333 uint64_t vhost_get_features(struct vhost_dev *hdev, const int *feature_bits, 334 uint64_t features); 335 336 /** 337 * vhost_ack_features() - set vhost acked_features 338 * @hdev: common vhost_dev structure 339 * @feature_bits: pointer to terminated table of feature bits 340 * @features: requested feature set 341 * 342 * This sets the internal hdev->acked_features to the intersection of 343 * the backends advertised features and the supported feature_bits. 344 */ 345 void vhost_ack_features(struct vhost_dev *hdev, const int *feature_bits, 346 uint64_t features); 347 unsigned int vhost_get_max_memslots(void); 348 unsigned int vhost_get_free_memslots(void); 349 350 int vhost_net_set_backend(struct vhost_dev *hdev, 351 struct vhost_vring_file *file); 352 353 void vhost_toggle_device_iotlb(VirtIODevice *vdev); 354 int vhost_device_iotlb_miss(struct vhost_dev *dev, uint64_t iova, int write); 355 356 int vhost_virtqueue_start(struct vhost_dev *dev, struct VirtIODevice *vdev, 357 struct vhost_virtqueue *vq, unsigned idx); 358 int vhost_virtqueue_stop(struct vhost_dev *dev, struct VirtIODevice *vdev, 359 struct vhost_virtqueue *vq, unsigned idx); 360 361 void vhost_dev_reset_inflight(struct vhost_inflight *inflight); 362 void vhost_dev_free_inflight(struct vhost_inflight *inflight); 363 int vhost_dev_prepare_inflight(struct vhost_dev *hdev, VirtIODevice *vdev); 364 int vhost_dev_set_inflight(struct vhost_dev *dev, 365 struct vhost_inflight *inflight); 366 int vhost_dev_get_inflight(struct vhost_dev *dev, uint16_t queue_size, 367 struct vhost_inflight *inflight); 368 bool vhost_dev_has_iommu(struct vhost_dev *dev); 369 370 #ifdef CONFIG_VHOST 371 int vhost_reset_device(struct vhost_dev *hdev); 372 #else 373 static inline int vhost_reset_device(struct vhost_dev *hdev) 374 { 375 return -ENOSYS; 376 } 377 #endif /* CONFIG_VHOST */ 378 379 /** 380 * vhost_supports_device_state(): Checks whether the back-end supports 381 * transferring internal device state for the purpose of migration. 382 * Support for this feature is required for vhost_set_device_state_fd() 383 * and vhost_check_device_state(). 384 * 385 * @dev: The vhost device 386 * 387 * Returns true if the device supports these commands, and false if it 388 * does not. 389 */ 390 #ifdef CONFIG_VHOST 391 bool vhost_supports_device_state(struct vhost_dev *dev); 392 #else 393 static inline bool vhost_supports_device_state(struct vhost_dev *dev) 394 { 395 return false; 396 } 397 #endif 398 399 /** 400 * vhost_set_device_state_fd(): Begin transfer of internal state from/to 401 * the back-end for the purpose of migration. Data is to be transferred 402 * over a pipe according to @direction and @phase. The sending end must 403 * only write to the pipe, and the receiving end must only read from it. 404 * Once the sending end is done, it closes its FD. The receiving end 405 * must take this as the end-of-transfer signal and close its FD, too. 406 * 407 * @fd is the back-end's end of the pipe: The write FD for SAVE, and the 408 * read FD for LOAD. This function transfers ownership of @fd to the 409 * back-end, i.e. closes it in the front-end. 410 * 411 * The back-end may optionally reply with an FD of its own, if this 412 * improves efficiency on its end. In this case, the returned FD is 413 * stored in *reply_fd. The back-end will discard the FD sent to it, 414 * and the front-end must use *reply_fd for transferring state to/from 415 * the back-end. 416 * 417 * @dev: The vhost device 418 * @direction: The direction in which the state is to be transferred. 419 * For outgoing migrations, this is SAVE, and data is read 420 * from the back-end and stored by the front-end in the 421 * migration stream. 422 * For incoming migrations, this is LOAD, and data is read 423 * by the front-end from the migration stream and sent to 424 * the back-end to restore the saved state. 425 * @phase: Which migration phase we are in. Currently, there is only 426 * STOPPED (device and all vrings are stopped), in the future, 427 * more phases such as PRE_COPY or POST_COPY may be added. 428 * @fd: Back-end's end of the pipe through which to transfer state; note 429 * that ownership is transferred to the back-end, so this function 430 * closes @fd in the front-end. 431 * @reply_fd: If the back-end wishes to use a different pipe for state 432 * transfer, this will contain an FD for the front-end to 433 * use. Otherwise, -1 is stored here. 434 * @errp: Potential error description 435 * 436 * Returns 0 on success, and -errno on failure. 437 */ 438 int vhost_set_device_state_fd(struct vhost_dev *dev, 439 VhostDeviceStateDirection direction, 440 VhostDeviceStatePhase phase, 441 int fd, 442 int *reply_fd, 443 Error **errp); 444 445 /** 446 * vhost_set_device_state_fd(): After transferring state from/to the 447 * back-end via vhost_set_device_state_fd(), i.e. once the sending end 448 * has closed the pipe, inquire the back-end to report any potential 449 * errors that have occurred on its side. This allows to sense errors 450 * like: 451 * - During outgoing migration, when the source side had already started 452 * to produce its state, something went wrong and it failed to finish 453 * - During incoming migration, when the received state is somehow 454 * invalid and cannot be processed by the back-end 455 * 456 * @dev: The vhost device 457 * @errp: Potential error description 458 * 459 * Returns 0 when the back-end reports successful state transfer and 460 * processing, and -errno when an error occurred somewhere. 461 */ 462 int vhost_check_device_state(struct vhost_dev *dev, Error **errp); 463 464 /** 465 * vhost_save_backend_state(): High-level function to receive a vhost 466 * back-end's state, and save it in @f. Uses 467 * `vhost_set_device_state_fd()` to get the data from the back-end, and 468 * stores it in consecutive chunks that are each prefixed by their 469 * respective length (be32). The end is marked by a 0-length chunk. 470 * 471 * Must only be called while the device and all its vrings are stopped 472 * (`VHOST_TRANSFER_STATE_PHASE_STOPPED`). 473 * 474 * @dev: The vhost device from which to save the state 475 * @f: Migration stream in which to save the state 476 * @errp: Potential error message 477 * 478 * Returns 0 on success, and -errno otherwise. 479 */ 480 #ifdef CONFIG_VHOST 481 int vhost_save_backend_state(struct vhost_dev *dev, QEMUFile *f, Error **errp); 482 #else 483 static inline int vhost_save_backend_state(struct vhost_dev *dev, QEMUFile *f, 484 Error **errp) 485 { 486 return -ENOSYS; 487 } 488 #endif 489 490 /** 491 * vhost_load_backend_state(): High-level function to load a vhost 492 * back-end's state from @f, and send it over to the back-end. Reads 493 * the data from @f in the format used by `vhost_save_state()`, and uses 494 * `vhost_set_device_state_fd()` to transfer it to the back-end. 495 * 496 * Must only be called while the device and all its vrings are stopped 497 * (`VHOST_TRANSFER_STATE_PHASE_STOPPED`). 498 * 499 * @dev: The vhost device to which to send the state 500 * @f: Migration stream from which to load the state 501 * @errp: Potential error message 502 * 503 * Returns 0 on success, and -errno otherwise. 504 */ 505 #ifdef CONFIG_VHOST 506 int vhost_load_backend_state(struct vhost_dev *dev, QEMUFile *f, Error **errp); 507 #else 508 static inline int vhost_load_backend_state(struct vhost_dev *dev, QEMUFile *f, 509 Error **errp) 510 { 511 return -ENOSYS; 512 } 513 #endif 514 515 #endif 516