1 #ifndef VHOST_H 2 #define VHOST_H 3 4 #include "net/vhost_net.h" 5 #include "hw/virtio/vhost-backend.h" 6 #include "hw/virtio/virtio.h" 7 #include "system/memory.h" 8 9 #define VHOST_F_DEVICE_IOTLB 63 10 #define VHOST_USER_F_PROTOCOL_FEATURES 30 11 12 #define VU_REALIZE_CONN_RETRIES 3 13 14 /* Generic structures common for any vhost based device. */ 15 16 struct vhost_inflight { 17 int fd; 18 void *addr; 19 uint64_t size; 20 uint64_t offset; 21 uint16_t queue_size; 22 }; 23 24 struct vhost_virtqueue { 25 int kick; 26 int call; 27 void *desc; 28 void *avail; 29 void *used; 30 int num; 31 unsigned long long desc_phys; 32 unsigned desc_size; 33 unsigned long long avail_phys; 34 unsigned avail_size; 35 unsigned long long used_phys; 36 unsigned used_size; 37 EventNotifier masked_notifier; 38 EventNotifier error_notifier; 39 EventNotifier masked_config_notifier; 40 struct vhost_dev *dev; 41 }; 42 43 typedef unsigned long vhost_log_chunk_t; 44 #define VHOST_LOG_PAGE 0x1000 45 #define VHOST_LOG_BITS (8 * sizeof(vhost_log_chunk_t)) 46 #define VHOST_LOG_CHUNK (VHOST_LOG_PAGE * VHOST_LOG_BITS) 47 #define VHOST_INVALID_FEATURE_BIT (0xff) 48 #define VHOST_QUEUE_NUM_CONFIG_INR 0 49 50 struct vhost_log { 51 unsigned long long size; 52 int refcnt; 53 int fd; 54 vhost_log_chunk_t *log; 55 }; 56 57 struct vhost_dev; 58 struct vhost_iommu { 59 struct vhost_dev *hdev; 60 MemoryRegion *mr; 61 hwaddr iommu_offset; 62 IOMMUNotifier n; 63 QLIST_ENTRY(vhost_iommu) iommu_next; 64 }; 65 66 typedef struct VhostDevConfigOps { 67 /* Vhost device config space changed callback 68 */ 69 int (*vhost_dev_config_notifier)(struct vhost_dev *dev); 70 } VhostDevConfigOps; 71 72 struct vhost_memory; 73 74 /** 75 * struct vhost_dev - common vhost_dev structure 76 * @vhost_ops: backend specific ops 77 * @config_ops: ops for config changes (see @vhost_dev_set_config_notifier) 78 */ 79 struct vhost_dev { 80 VirtIODevice *vdev; 81 MemoryListener memory_listener; 82 MemoryListener iommu_listener; 83 struct vhost_memory *mem; 84 int n_mem_sections; 85 MemoryRegionSection *mem_sections; 86 int n_tmp_sections; 87 MemoryRegionSection *tmp_sections; 88 struct vhost_virtqueue *vqs; 89 unsigned int nvqs; 90 /* the first virtqueue which would be used by this vhost dev */ 91 int vq_index; 92 /* one past the last vq index for the virtio device (not vhost) */ 93 int vq_index_end; 94 /* if non-zero, minimum required value for max_queues */ 95 int num_queues; 96 /** 97 * vhost feature handling requires matching the feature set 98 * offered by a backend which may be a subset of the total 99 * features eventually offered to the guest. 100 * 101 * @features: available features provided by the backend 102 * @acked_features: final negotiated features with front-end driver 103 * 104 * @backend_features: this is used in a couple of places to either 105 * store VHOST_USER_F_PROTOCOL_FEATURES to apply to 106 * VHOST_USER_SET_FEATURES or VHOST_NET_F_VIRTIO_NET_HDR. Its 107 * future use should be discouraged and the variable retired as 108 * its easy to confuse with the VirtIO backend_features. 109 */ 110 uint64_t features; 111 uint64_t acked_features; 112 uint64_t backend_features; 113 114 /** 115 * @protocol_features: is the vhost-user only feature set by 116 * VHOST_USER_SET_PROTOCOL_FEATURES. Protocol features are only 117 * negotiated if VHOST_USER_F_PROTOCOL_FEATURES has been offered 118 * by the backend (see @features). 119 */ 120 uint64_t protocol_features; 121 122 uint64_t max_queues; 123 uint64_t backend_cap; 124 /* @started: is the vhost device started? */ 125 bool started; 126 bool log_enabled; 127 uint64_t log_size; 128 Error *migration_blocker; 129 const VhostOps *vhost_ops; 130 void *opaque; 131 struct vhost_log *log; 132 QLIST_ENTRY(vhost_dev) entry; 133 QLIST_ENTRY(vhost_dev) logdev_entry; 134 QLIST_HEAD(, vhost_iommu) iommu_list; 135 IOMMUNotifier n; 136 const VhostDevConfigOps *config_ops; 137 }; 138 139 extern const VhostOps kernel_ops; 140 extern const VhostOps user_ops; 141 extern const VhostOps vdpa_ops; 142 143 struct vhost_net { 144 struct vhost_dev dev; 145 struct vhost_virtqueue vqs[2]; 146 int backend; 147 const int *feature_bits; 148 SaveAcketFeatures *save_acked_features; 149 NetClientState *nc; 150 }; 151 152 /** 153 * vhost_dev_init() - initialise the vhost interface 154 * @hdev: the common vhost_dev structure 155 * @opaque: opaque ptr passed to backend (vhost/vhost-user/vdpa) 156 * @backend_type: type of backend 157 * @busyloop_timeout: timeout for polling virtqueue 158 * @errp: error handle 159 * 160 * The initialisation of the vhost device will trigger the 161 * initialisation of the backend and potentially capability 162 * negotiation of backend interface. Configuration of the VirtIO 163 * itself won't happen until the interface is started. 164 * 165 * Return: 0 on success, non-zero on error while setting errp. 166 */ 167 int vhost_dev_init(struct vhost_dev *hdev, void *opaque, 168 VhostBackendType backend_type, 169 uint32_t busyloop_timeout, Error **errp); 170 171 /** 172 * vhost_dev_cleanup() - tear down and cleanup vhost interface 173 * @hdev: the common vhost_dev structure 174 */ 175 void vhost_dev_cleanup(struct vhost_dev *hdev); 176 177 void vhost_dev_disable_notifiers_nvqs(struct vhost_dev *hdev, 178 VirtIODevice *vdev, 179 unsigned int nvqs); 180 181 /** 182 * vhost_dev_enable_notifiers() - enable event notifiers 183 * @hdev: common vhost_dev structure 184 * @vdev: the VirtIODevice structure 185 * 186 * Enable notifications directly to the vhost device rather than being 187 * triggered by QEMU itself. Notifications should be enabled before 188 * the vhost device is started via @vhost_dev_start. 189 * 190 * Return: 0 on success, < 0 on error. 191 */ 192 int vhost_dev_enable_notifiers(struct vhost_dev *hdev, VirtIODevice *vdev); 193 194 /** 195 * vhost_dev_disable_notifiers - disable event notifications 196 * @hdev: common vhost_dev structure 197 * @vdev: the VirtIODevice structure 198 * 199 * Disable direct notifications to vhost device. 200 */ 201 void vhost_dev_disable_notifiers(struct vhost_dev *hdev, VirtIODevice *vdev); 202 bool vhost_config_pending(struct vhost_dev *hdev); 203 void vhost_config_mask(struct vhost_dev *hdev, VirtIODevice *vdev, bool mask); 204 205 /** 206 * vhost_dev_is_started() - report status of vhost device 207 * @hdev: common vhost_dev structure 208 * 209 * Return the started status of the vhost device 210 */ 211 static inline bool vhost_dev_is_started(struct vhost_dev *hdev) 212 { 213 return hdev->started; 214 } 215 216 /** 217 * vhost_dev_start() - start the vhost device 218 * @hdev: common vhost_dev structure 219 * @vdev: the VirtIODevice structure 220 * @vrings: true to have vrings enabled in this call 221 * 222 * Starts the vhost device. From this point VirtIO feature negotiation 223 * can start and the device can start processing VirtIO transactions. 224 * 225 * Return: 0 on success, < 0 on error. 226 */ 227 int vhost_dev_start(struct vhost_dev *hdev, VirtIODevice *vdev, bool vrings); 228 229 /** 230 * vhost_dev_stop() - stop the vhost device 231 * @hdev: common vhost_dev structure 232 * @vdev: the VirtIODevice structure 233 * @vrings: true to have vrings disabled in this call 234 * 235 * Stop the vhost device. After the device is stopped the notifiers 236 * can be disabled (@vhost_dev_disable_notifiers) and the device can 237 * be torn down (@vhost_dev_cleanup). 238 * 239 * Return: 0 on success, != 0 on error when stopping dev. 240 */ 241 int vhost_dev_stop(struct vhost_dev *hdev, VirtIODevice *vdev, bool vrings); 242 243 /** 244 * DOC: vhost device configuration handling 245 * 246 * The VirtIO device configuration space is used for rarely changing 247 * or initialisation time parameters. The configuration can be updated 248 * by either the guest driver or the device itself. If the device can 249 * change the configuration over time the vhost handler should 250 * register a @VhostDevConfigOps structure with 251 * @vhost_dev_set_config_notifier so the guest can be notified. Some 252 * devices register a handler anyway and will signal an error if an 253 * unexpected config change happens. 254 */ 255 256 /** 257 * vhost_dev_get_config() - fetch device configuration 258 * @hdev: common vhost_dev_structure 259 * @config: pointer to device appropriate config structure 260 * @config_len: size of device appropriate config structure 261 * 262 * Return: 0 on success, < 0 on error while setting errp 263 */ 264 int vhost_dev_get_config(struct vhost_dev *hdev, uint8_t *config, 265 uint32_t config_len, Error **errp); 266 267 /** 268 * vhost_dev_set_config() - set device configuration 269 * @hdev: common vhost_dev_structure 270 * @data: pointer to data to set 271 * @offset: offset into configuration space 272 * @size: length of set 273 * @flags: @VhostSetConfigType flags 274 * 275 * By use of @offset/@size a subset of the configuration space can be 276 * written to. The @flags are used to indicate if it is a normal 277 * transaction or related to migration. 278 * 279 * Return: 0 on success, non-zero on error 280 */ 281 int vhost_dev_set_config(struct vhost_dev *dev, const uint8_t *data, 282 uint32_t offset, uint32_t size, uint32_t flags); 283 284 /** 285 * vhost_dev_set_config_notifier() - register VhostDevConfigOps 286 * @hdev: common vhost_dev_structure 287 * @ops: notifier ops 288 * 289 * If the device is expected to change configuration a notifier can be 290 * setup to handle the case. 291 */ 292 void vhost_dev_set_config_notifier(struct vhost_dev *dev, 293 const VhostDevConfigOps *ops); 294 295 296 /* Test and clear masked event pending status. 297 * Should be called after unmask to avoid losing events. 298 */ 299 bool vhost_virtqueue_pending(struct vhost_dev *hdev, int n); 300 301 /* Mask/unmask events from this vq. 302 */ 303 void vhost_virtqueue_mask(struct vhost_dev *hdev, VirtIODevice *vdev, int n, 304 bool mask); 305 306 /** 307 * vhost_get_features() - return a sanitised set of feature bits 308 * @hdev: common vhost_dev structure 309 * @feature_bits: pointer to terminated table of feature bits 310 * @features: original feature set 311 * 312 * This returns a set of features bits that is an intersection of what 313 * is supported by the vhost backend (hdev->features), the supported 314 * feature_bits and the requested feature set. 315 */ 316 uint64_t vhost_get_features(struct vhost_dev *hdev, const int *feature_bits, 317 uint64_t features); 318 319 /** 320 * vhost_ack_features() - set vhost acked_features 321 * @hdev: common vhost_dev structure 322 * @feature_bits: pointer to terminated table of feature bits 323 * @features: requested feature set 324 * 325 * This sets the internal hdev->acked_features to the intersection of 326 * the backends advertised features and the supported feature_bits. 327 */ 328 void vhost_ack_features(struct vhost_dev *hdev, const int *feature_bits, 329 uint64_t features); 330 unsigned int vhost_get_max_memslots(void); 331 unsigned int vhost_get_free_memslots(void); 332 333 int vhost_net_set_backend(struct vhost_dev *hdev, 334 struct vhost_vring_file *file); 335 336 void vhost_toggle_device_iotlb(VirtIODevice *vdev); 337 int vhost_device_iotlb_miss(struct vhost_dev *dev, uint64_t iova, int write); 338 339 int vhost_virtqueue_start(struct vhost_dev *dev, struct VirtIODevice *vdev, 340 struct vhost_virtqueue *vq, unsigned idx); 341 int vhost_virtqueue_stop(struct vhost_dev *dev, struct VirtIODevice *vdev, 342 struct vhost_virtqueue *vq, unsigned idx); 343 344 void vhost_dev_reset_inflight(struct vhost_inflight *inflight); 345 void vhost_dev_free_inflight(struct vhost_inflight *inflight); 346 int vhost_dev_prepare_inflight(struct vhost_dev *hdev, VirtIODevice *vdev); 347 int vhost_dev_set_inflight(struct vhost_dev *dev, 348 struct vhost_inflight *inflight); 349 int vhost_dev_get_inflight(struct vhost_dev *dev, uint16_t queue_size, 350 struct vhost_inflight *inflight); 351 bool vhost_dev_has_iommu(struct vhost_dev *dev); 352 353 #ifdef CONFIG_VHOST 354 int vhost_reset_device(struct vhost_dev *hdev); 355 #else 356 static inline int vhost_reset_device(struct vhost_dev *hdev) 357 { 358 return -ENOSYS; 359 } 360 #endif /* CONFIG_VHOST */ 361 362 /** 363 * vhost_supports_device_state(): Checks whether the back-end supports 364 * transferring internal device state for the purpose of migration. 365 * Support for this feature is required for vhost_set_device_state_fd() 366 * and vhost_check_device_state(). 367 * 368 * @dev: The vhost device 369 * 370 * Returns true if the device supports these commands, and false if it 371 * does not. 372 */ 373 #ifdef CONFIG_VHOST 374 bool vhost_supports_device_state(struct vhost_dev *dev); 375 #else 376 static inline bool vhost_supports_device_state(struct vhost_dev *dev) 377 { 378 return false; 379 } 380 #endif 381 382 /** 383 * vhost_set_device_state_fd(): Begin transfer of internal state from/to 384 * the back-end for the purpose of migration. Data is to be transferred 385 * over a pipe according to @direction and @phase. The sending end must 386 * only write to the pipe, and the receiving end must only read from it. 387 * Once the sending end is done, it closes its FD. The receiving end 388 * must take this as the end-of-transfer signal and close its FD, too. 389 * 390 * @fd is the back-end's end of the pipe: The write FD for SAVE, and the 391 * read FD for LOAD. This function transfers ownership of @fd to the 392 * back-end, i.e. closes it in the front-end. 393 * 394 * The back-end may optionally reply with an FD of its own, if this 395 * improves efficiency on its end. In this case, the returned FD is 396 * stored in *reply_fd. The back-end will discard the FD sent to it, 397 * and the front-end must use *reply_fd for transferring state to/from 398 * the back-end. 399 * 400 * @dev: The vhost device 401 * @direction: The direction in which the state is to be transferred. 402 * For outgoing migrations, this is SAVE, and data is read 403 * from the back-end and stored by the front-end in the 404 * migration stream. 405 * For incoming migrations, this is LOAD, and data is read 406 * by the front-end from the migration stream and sent to 407 * the back-end to restore the saved state. 408 * @phase: Which migration phase we are in. Currently, there is only 409 * STOPPED (device and all vrings are stopped), in the future, 410 * more phases such as PRE_COPY or POST_COPY may be added. 411 * @fd: Back-end's end of the pipe through which to transfer state; note 412 * that ownership is transferred to the back-end, so this function 413 * closes @fd in the front-end. 414 * @reply_fd: If the back-end wishes to use a different pipe for state 415 * transfer, this will contain an FD for the front-end to 416 * use. Otherwise, -1 is stored here. 417 * @errp: Potential error description 418 * 419 * Returns 0 on success, and -errno on failure. 420 */ 421 int vhost_set_device_state_fd(struct vhost_dev *dev, 422 VhostDeviceStateDirection direction, 423 VhostDeviceStatePhase phase, 424 int fd, 425 int *reply_fd, 426 Error **errp); 427 428 /** 429 * vhost_set_device_state_fd(): After transferring state from/to the 430 * back-end via vhost_set_device_state_fd(), i.e. once the sending end 431 * has closed the pipe, inquire the back-end to report any potential 432 * errors that have occurred on its side. This allows to sense errors 433 * like: 434 * - During outgoing migration, when the source side had already started 435 * to produce its state, something went wrong and it failed to finish 436 * - During incoming migration, when the received state is somehow 437 * invalid and cannot be processed by the back-end 438 * 439 * @dev: The vhost device 440 * @errp: Potential error description 441 * 442 * Returns 0 when the back-end reports successful state transfer and 443 * processing, and -errno when an error occurred somewhere. 444 */ 445 int vhost_check_device_state(struct vhost_dev *dev, Error **errp); 446 447 /** 448 * vhost_save_backend_state(): High-level function to receive a vhost 449 * back-end's state, and save it in @f. Uses 450 * `vhost_set_device_state_fd()` to get the data from the back-end, and 451 * stores it in consecutive chunks that are each prefixed by their 452 * respective length (be32). The end is marked by a 0-length chunk. 453 * 454 * Must only be called while the device and all its vrings are stopped 455 * (`VHOST_TRANSFER_STATE_PHASE_STOPPED`). 456 * 457 * @dev: The vhost device from which to save the state 458 * @f: Migration stream in which to save the state 459 * @errp: Potential error message 460 * 461 * Returns 0 on success, and -errno otherwise. 462 */ 463 #ifdef CONFIG_VHOST 464 int vhost_save_backend_state(struct vhost_dev *dev, QEMUFile *f, Error **errp); 465 #else 466 static inline int vhost_save_backend_state(struct vhost_dev *dev, QEMUFile *f, 467 Error **errp) 468 { 469 return -ENOSYS; 470 } 471 #endif 472 473 /** 474 * vhost_load_backend_state(): High-level function to load a vhost 475 * back-end's state from @f, and send it over to the back-end. Reads 476 * the data from @f in the format used by `vhost_save_state()`, and uses 477 * `vhost_set_device_state_fd()` to transfer it to the back-end. 478 * 479 * Must only be called while the device and all its vrings are stopped 480 * (`VHOST_TRANSFER_STATE_PHASE_STOPPED`). 481 * 482 * @dev: The vhost device to which to send the state 483 * @f: Migration stream from which to load the state 484 * @errp: Potential error message 485 * 486 * Returns 0 on success, and -errno otherwise. 487 */ 488 #ifdef CONFIG_VHOST 489 int vhost_load_backend_state(struct vhost_dev *dev, QEMUFile *f, Error **errp); 490 #else 491 static inline int vhost_load_backend_state(struct vhost_dev *dev, QEMUFile *f, 492 Error **errp) 493 { 494 return -ENOSYS; 495 } 496 #endif 497 498 #endif 499