1 #ifndef VHOST_H
2 #define VHOST_H
3
4 #include "net/vhost_net.h"
5 #include "hw/virtio/vhost-backend.h"
6 #include "hw/virtio/virtio.h"
7 #include "system/memory.h"
8
9 #define VHOST_F_DEVICE_IOTLB 63
10 #define VHOST_USER_F_PROTOCOL_FEATURES 30
11
12 #define VU_REALIZE_CONN_RETRIES 3
13
14 /* Generic structures common for any vhost based device. */
15
16 struct vhost_inflight {
17 int fd;
18 void *addr;
19 uint64_t size;
20 uint64_t offset;
21 uint16_t queue_size;
22 };
23
24 struct vhost_virtqueue {
25 int kick;
26 int call;
27 void *desc;
28 void *avail;
29 void *used;
30 int num;
31 unsigned long long desc_phys;
32 unsigned desc_size;
33 unsigned long long avail_phys;
34 unsigned avail_size;
35 unsigned long long used_phys;
36 unsigned used_size;
37 EventNotifier masked_notifier;
38 EventNotifier error_notifier;
39 EventNotifier masked_config_notifier;
40 struct vhost_dev *dev;
41 };
42
43 typedef unsigned long vhost_log_chunk_t;
44 #define VHOST_LOG_PAGE 0x1000
45 #define VHOST_LOG_BITS (8 * sizeof(vhost_log_chunk_t))
46 #define VHOST_LOG_CHUNK (VHOST_LOG_PAGE * VHOST_LOG_BITS)
47 #define VHOST_INVALID_FEATURE_BIT (0xff)
48 #define VHOST_QUEUE_NUM_CONFIG_INR 0
49
50 struct vhost_log {
51 unsigned long long size;
52 int refcnt;
53 int fd;
54 vhost_log_chunk_t *log;
55 };
56
57 struct vhost_dev;
58 struct vhost_iommu {
59 struct vhost_dev *hdev;
60 MemoryRegion *mr;
61 hwaddr iommu_offset;
62 IOMMUNotifier n;
63 QLIST_ENTRY(vhost_iommu) iommu_next;
64 };
65
66 typedef struct VhostDevConfigOps {
67 /* Vhost device config space changed callback
68 */
69 int (*vhost_dev_config_notifier)(struct vhost_dev *dev);
70 } VhostDevConfigOps;
71
72 struct vhost_memory;
73
74 /**
75 * struct vhost_dev - common vhost_dev structure
76 * @vhost_ops: backend specific ops
77 * @config_ops: ops for config changes (see @vhost_dev_set_config_notifier)
78 */
79 struct vhost_dev {
80 VirtIODevice *vdev;
81 MemoryListener memory_listener;
82 MemoryListener iommu_listener;
83 struct vhost_memory *mem;
84 int n_mem_sections;
85 MemoryRegionSection *mem_sections;
86 int n_tmp_sections;
87 MemoryRegionSection *tmp_sections;
88 struct vhost_virtqueue *vqs;
89 unsigned int nvqs;
90 /* the first virtqueue which would be used by this vhost dev */
91 int vq_index;
92 /* one past the last vq index for the virtio device (not vhost) */
93 int vq_index_end;
94 /* if non-zero, minimum required value for max_queues */
95 int num_queues;
96 /**
97 * vhost feature handling requires matching the feature set
98 * offered by a backend which may be a subset of the total
99 * features eventually offered to the guest.
100 *
101 * @features: available features provided by the backend
102 * @acked_features: final negotiated features with front-end driver
103 *
104 * @backend_features: this is used in a couple of places to either
105 * store VHOST_USER_F_PROTOCOL_FEATURES to apply to
106 * VHOST_USER_SET_FEATURES or VHOST_NET_F_VIRTIO_NET_HDR. Its
107 * future use should be discouraged and the variable retired as
108 * its easy to confuse with the VirtIO backend_features.
109 */
110 uint64_t features;
111 uint64_t acked_features;
112 uint64_t backend_features;
113
114 /**
115 * @protocol_features: is the vhost-user only feature set by
116 * VHOST_USER_SET_PROTOCOL_FEATURES. Protocol features are only
117 * negotiated if VHOST_USER_F_PROTOCOL_FEATURES has been offered
118 * by the backend (see @features).
119 */
120 uint64_t protocol_features;
121
122 uint64_t max_queues;
123 uint64_t backend_cap;
124 /* @started: is the vhost device started? */
125 bool started;
126 bool log_enabled;
127 uint64_t log_size;
128 Error *migration_blocker;
129 const VhostOps *vhost_ops;
130 void *opaque;
131 struct vhost_log *log;
132 QLIST_ENTRY(vhost_dev) entry;
133 QLIST_ENTRY(vhost_dev) logdev_entry;
134 QLIST_HEAD(, vhost_iommu) iommu_list;
135 IOMMUNotifier n;
136 const VhostDevConfigOps *config_ops;
137 };
138
139 extern const VhostOps kernel_ops;
140 extern const VhostOps user_ops;
141 extern const VhostOps vdpa_ops;
142
143 struct vhost_net {
144 struct vhost_dev dev;
145 struct vhost_virtqueue vqs[2];
146 int backend;
147 const int *feature_bits;
148 int max_tx_queue_size;
149 SaveAcketFeatures *save_acked_features;
150 bool is_vhost_user;
151 NetClientState *nc;
152 };
153
154 /**
155 * vhost_dev_init() - initialise the vhost interface
156 * @hdev: the common vhost_dev structure
157 * @opaque: opaque ptr passed to backend (vhost/vhost-user/vdpa)
158 * @backend_type: type of backend
159 * @busyloop_timeout: timeout for polling virtqueue
160 * @errp: error handle
161 *
162 * The initialisation of the vhost device will trigger the
163 * initialisation of the backend and potentially capability
164 * negotiation of backend interface. Configuration of the VirtIO
165 * itself won't happen until the interface is started.
166 *
167 * Return: 0 on success, non-zero on error while setting errp.
168 */
169 int vhost_dev_init(struct vhost_dev *hdev, void *opaque,
170 VhostBackendType backend_type,
171 uint32_t busyloop_timeout, Error **errp);
172
173 /**
174 * vhost_dev_cleanup() - tear down and cleanup vhost interface
175 * @hdev: the common vhost_dev structure
176 */
177 void vhost_dev_cleanup(struct vhost_dev *hdev);
178
179 void vhost_dev_disable_notifiers_nvqs(struct vhost_dev *hdev,
180 VirtIODevice *vdev,
181 unsigned int nvqs);
182
183 /**
184 * vhost_dev_enable_notifiers() - enable event notifiers
185 * @hdev: common vhost_dev structure
186 * @vdev: the VirtIODevice structure
187 *
188 * Enable notifications directly to the vhost device rather than being
189 * triggered by QEMU itself. Notifications should be enabled before
190 * the vhost device is started via @vhost_dev_start.
191 *
192 * Return: 0 on success, < 0 on error.
193 */
194 int vhost_dev_enable_notifiers(struct vhost_dev *hdev, VirtIODevice *vdev);
195
196 /**
197 * vhost_dev_disable_notifiers - disable event notifications
198 * @hdev: common vhost_dev structure
199 * @vdev: the VirtIODevice structure
200 *
201 * Disable direct notifications to vhost device.
202 */
203 void vhost_dev_disable_notifiers(struct vhost_dev *hdev, VirtIODevice *vdev);
204 bool vhost_config_pending(struct vhost_dev *hdev);
205 void vhost_config_mask(struct vhost_dev *hdev, VirtIODevice *vdev, bool mask);
206
207 /**
208 * vhost_dev_is_started() - report status of vhost device
209 * @hdev: common vhost_dev structure
210 *
211 * Return the started status of the vhost device
212 */
vhost_dev_is_started(struct vhost_dev * hdev)213 static inline bool vhost_dev_is_started(struct vhost_dev *hdev)
214 {
215 return hdev->started;
216 }
217
218 /**
219 * vhost_dev_start() - start the vhost device
220 * @hdev: common vhost_dev structure
221 * @vdev: the VirtIODevice structure
222 * @vrings: true to have vrings enabled in this call
223 *
224 * Starts the vhost device. From this point VirtIO feature negotiation
225 * can start and the device can start processing VirtIO transactions.
226 *
227 * Return: 0 on success, < 0 on error.
228 */
229 int vhost_dev_start(struct vhost_dev *hdev, VirtIODevice *vdev, bool vrings);
230
231 /**
232 * vhost_dev_stop() - stop the vhost device
233 * @hdev: common vhost_dev structure
234 * @vdev: the VirtIODevice structure
235 * @vrings: true to have vrings disabled in this call
236 *
237 * Stop the vhost device. After the device is stopped the notifiers
238 * can be disabled (@vhost_dev_disable_notifiers) and the device can
239 * be torn down (@vhost_dev_cleanup).
240 *
241 * Return: 0 on success, != 0 on error when stopping dev.
242 */
243 int vhost_dev_stop(struct vhost_dev *hdev, VirtIODevice *vdev, bool vrings);
244
245 /**
246 * vhost_dev_force_stop() - force stop the vhost device
247 * @hdev: common vhost_dev structure
248 * @vdev: the VirtIODevice structure
249 * @vrings: true to have vrings disabled in this call
250 *
251 * Force stop the vhost device. After the device is stopped the notifiers
252 * can be disabled (@vhost_dev_disable_notifiers) and the device can
253 * be torn down (@vhost_dev_cleanup). Unlike @vhost_dev_stop, this doesn't
254 * attempt to flush in-flight backend requests by skipping GET_VRING_BASE
255 * entirely.
256 */
257 int vhost_dev_force_stop(struct vhost_dev *hdev, VirtIODevice *vdev,
258 bool vrings);
259
260 /**
261 * DOC: vhost device configuration handling
262 *
263 * The VirtIO device configuration space is used for rarely changing
264 * or initialisation time parameters. The configuration can be updated
265 * by either the guest driver or the device itself. If the device can
266 * change the configuration over time the vhost handler should
267 * register a @VhostDevConfigOps structure with
268 * @vhost_dev_set_config_notifier so the guest can be notified. Some
269 * devices register a handler anyway and will signal an error if an
270 * unexpected config change happens.
271 */
272
273 /**
274 * vhost_dev_get_config() - fetch device configuration
275 * @hdev: common vhost_dev_structure
276 * @config: pointer to device appropriate config structure
277 * @config_len: size of device appropriate config structure
278 *
279 * Return: 0 on success, < 0 on error while setting errp
280 */
281 int vhost_dev_get_config(struct vhost_dev *hdev, uint8_t *config,
282 uint32_t config_len, Error **errp);
283
284 /**
285 * vhost_dev_set_config() - set device configuration
286 * @hdev: common vhost_dev_structure
287 * @data: pointer to data to set
288 * @offset: offset into configuration space
289 * @size: length of set
290 * @flags: @VhostSetConfigType flags
291 *
292 * By use of @offset/@size a subset of the configuration space can be
293 * written to. The @flags are used to indicate if it is a normal
294 * transaction or related to migration.
295 *
296 * Return: 0 on success, non-zero on error
297 */
298 int vhost_dev_set_config(struct vhost_dev *dev, const uint8_t *data,
299 uint32_t offset, uint32_t size, uint32_t flags);
300
301 /**
302 * vhost_dev_set_config_notifier() - register VhostDevConfigOps
303 * @hdev: common vhost_dev_structure
304 * @ops: notifier ops
305 *
306 * If the device is expected to change configuration a notifier can be
307 * setup to handle the case.
308 */
309 void vhost_dev_set_config_notifier(struct vhost_dev *dev,
310 const VhostDevConfigOps *ops);
311
312
313 /* Test and clear masked event pending status.
314 * Should be called after unmask to avoid losing events.
315 */
316 bool vhost_virtqueue_pending(struct vhost_dev *hdev, int n);
317
318 /* Mask/unmask events from this vq.
319 */
320 void vhost_virtqueue_mask(struct vhost_dev *hdev, VirtIODevice *vdev, int n,
321 bool mask);
322
323 /**
324 * vhost_get_features() - return a sanitised set of feature bits
325 * @hdev: common vhost_dev structure
326 * @feature_bits: pointer to terminated table of feature bits
327 * @features: original feature set
328 *
329 * This returns a set of features bits that is an intersection of what
330 * is supported by the vhost backend (hdev->features), the supported
331 * feature_bits and the requested feature set.
332 */
333 uint64_t vhost_get_features(struct vhost_dev *hdev, const int *feature_bits,
334 uint64_t features);
335
336 /**
337 * vhost_ack_features() - set vhost acked_features
338 * @hdev: common vhost_dev structure
339 * @feature_bits: pointer to terminated table of feature bits
340 * @features: requested feature set
341 *
342 * This sets the internal hdev->acked_features to the intersection of
343 * the backends advertised features and the supported feature_bits.
344 */
345 void vhost_ack_features(struct vhost_dev *hdev, const int *feature_bits,
346 uint64_t features);
347 unsigned int vhost_get_max_memslots(void);
348 unsigned int vhost_get_free_memslots(void);
349
350 int vhost_net_set_backend(struct vhost_dev *hdev,
351 struct vhost_vring_file *file);
352
353 void vhost_toggle_device_iotlb(VirtIODevice *vdev);
354 int vhost_device_iotlb_miss(struct vhost_dev *dev, uint64_t iova, int write);
355
356 int vhost_virtqueue_start(struct vhost_dev *dev, struct VirtIODevice *vdev,
357 struct vhost_virtqueue *vq, unsigned idx);
358 int vhost_virtqueue_stop(struct vhost_dev *dev, struct VirtIODevice *vdev,
359 struct vhost_virtqueue *vq, unsigned idx);
360
361 void vhost_dev_reset_inflight(struct vhost_inflight *inflight);
362 void vhost_dev_free_inflight(struct vhost_inflight *inflight);
363 int vhost_dev_prepare_inflight(struct vhost_dev *hdev, VirtIODevice *vdev);
364 int vhost_dev_set_inflight(struct vhost_dev *dev,
365 struct vhost_inflight *inflight);
366 int vhost_dev_get_inflight(struct vhost_dev *dev, uint16_t queue_size,
367 struct vhost_inflight *inflight);
368 bool vhost_dev_has_iommu(struct vhost_dev *dev);
369
370 #ifdef CONFIG_VHOST
371 int vhost_reset_device(struct vhost_dev *hdev);
372 #else
vhost_reset_device(struct vhost_dev * hdev)373 static inline int vhost_reset_device(struct vhost_dev *hdev)
374 {
375 return -ENOSYS;
376 }
377 #endif /* CONFIG_VHOST */
378
379 /**
380 * vhost_supports_device_state(): Checks whether the back-end supports
381 * transferring internal device state for the purpose of migration.
382 * Support for this feature is required for vhost_set_device_state_fd()
383 * and vhost_check_device_state().
384 *
385 * @dev: The vhost device
386 *
387 * Returns true if the device supports these commands, and false if it
388 * does not.
389 */
390 #ifdef CONFIG_VHOST
391 bool vhost_supports_device_state(struct vhost_dev *dev);
392 #else
vhost_supports_device_state(struct vhost_dev * dev)393 static inline bool vhost_supports_device_state(struct vhost_dev *dev)
394 {
395 return false;
396 }
397 #endif
398
399 /**
400 * vhost_set_device_state_fd(): Begin transfer of internal state from/to
401 * the back-end for the purpose of migration. Data is to be transferred
402 * over a pipe according to @direction and @phase. The sending end must
403 * only write to the pipe, and the receiving end must only read from it.
404 * Once the sending end is done, it closes its FD. The receiving end
405 * must take this as the end-of-transfer signal and close its FD, too.
406 *
407 * @fd is the back-end's end of the pipe: The write FD for SAVE, and the
408 * read FD for LOAD. This function transfers ownership of @fd to the
409 * back-end, i.e. closes it in the front-end.
410 *
411 * The back-end may optionally reply with an FD of its own, if this
412 * improves efficiency on its end. In this case, the returned FD is
413 * stored in *reply_fd. The back-end will discard the FD sent to it,
414 * and the front-end must use *reply_fd for transferring state to/from
415 * the back-end.
416 *
417 * @dev: The vhost device
418 * @direction: The direction in which the state is to be transferred.
419 * For outgoing migrations, this is SAVE, and data is read
420 * from the back-end and stored by the front-end in the
421 * migration stream.
422 * For incoming migrations, this is LOAD, and data is read
423 * by the front-end from the migration stream and sent to
424 * the back-end to restore the saved state.
425 * @phase: Which migration phase we are in. Currently, there is only
426 * STOPPED (device and all vrings are stopped), in the future,
427 * more phases such as PRE_COPY or POST_COPY may be added.
428 * @fd: Back-end's end of the pipe through which to transfer state; note
429 * that ownership is transferred to the back-end, so this function
430 * closes @fd in the front-end.
431 * @reply_fd: If the back-end wishes to use a different pipe for state
432 * transfer, this will contain an FD for the front-end to
433 * use. Otherwise, -1 is stored here.
434 * @errp: Potential error description
435 *
436 * Returns 0 on success, and -errno on failure.
437 */
438 int vhost_set_device_state_fd(struct vhost_dev *dev,
439 VhostDeviceStateDirection direction,
440 VhostDeviceStatePhase phase,
441 int fd,
442 int *reply_fd,
443 Error **errp);
444
445 /**
446 * vhost_set_device_state_fd(): After transferring state from/to the
447 * back-end via vhost_set_device_state_fd(), i.e. once the sending end
448 * has closed the pipe, inquire the back-end to report any potential
449 * errors that have occurred on its side. This allows to sense errors
450 * like:
451 * - During outgoing migration, when the source side had already started
452 * to produce its state, something went wrong and it failed to finish
453 * - During incoming migration, when the received state is somehow
454 * invalid and cannot be processed by the back-end
455 *
456 * @dev: The vhost device
457 * @errp: Potential error description
458 *
459 * Returns 0 when the back-end reports successful state transfer and
460 * processing, and -errno when an error occurred somewhere.
461 */
462 int vhost_check_device_state(struct vhost_dev *dev, Error **errp);
463
464 /**
465 * vhost_save_backend_state(): High-level function to receive a vhost
466 * back-end's state, and save it in @f. Uses
467 * `vhost_set_device_state_fd()` to get the data from the back-end, and
468 * stores it in consecutive chunks that are each prefixed by their
469 * respective length (be32). The end is marked by a 0-length chunk.
470 *
471 * Must only be called while the device and all its vrings are stopped
472 * (`VHOST_TRANSFER_STATE_PHASE_STOPPED`).
473 *
474 * @dev: The vhost device from which to save the state
475 * @f: Migration stream in which to save the state
476 * @errp: Potential error message
477 *
478 * Returns 0 on success, and -errno otherwise.
479 */
480 #ifdef CONFIG_VHOST
481 int vhost_save_backend_state(struct vhost_dev *dev, QEMUFile *f, Error **errp);
482 #else
vhost_save_backend_state(struct vhost_dev * dev,QEMUFile * f,Error ** errp)483 static inline int vhost_save_backend_state(struct vhost_dev *dev, QEMUFile *f,
484 Error **errp)
485 {
486 return -ENOSYS;
487 }
488 #endif
489
490 /**
491 * vhost_load_backend_state(): High-level function to load a vhost
492 * back-end's state from @f, and send it over to the back-end. Reads
493 * the data from @f in the format used by `vhost_save_state()`, and uses
494 * `vhost_set_device_state_fd()` to transfer it to the back-end.
495 *
496 * Must only be called while the device and all its vrings are stopped
497 * (`VHOST_TRANSFER_STATE_PHASE_STOPPED`).
498 *
499 * @dev: The vhost device to which to send the state
500 * @f: Migration stream from which to load the state
501 * @errp: Potential error message
502 *
503 * Returns 0 on success, and -errno otherwise.
504 */
505 #ifdef CONFIG_VHOST
506 int vhost_load_backend_state(struct vhost_dev *dev, QEMUFile *f, Error **errp);
507 #else
vhost_load_backend_state(struct vhost_dev * dev,QEMUFile * f,Error ** errp)508 static inline int vhost_load_backend_state(struct vhost_dev *dev, QEMUFile *f,
509 Error **errp)
510 {
511 return -ENOSYS;
512 }
513 #endif
514
515 #endif
516