1 /* SPDX-License-Identifier: GPL-2.0-only */
2 /*
3 * Copyright (C) 2012 Red Hat, Inc. All rights reserved.
4 * Author: Alex Williamson <alex.williamson@redhat.com>
5 */
6 #ifndef __VFIO_VFIO_H__
7 #define __VFIO_VFIO_H__
8
9 #include <linux/file.h>
10 #include <linux/device.h>
11 #include <linux/cdev.h>
12 #include <linux/module.h>
13 #include <linux/vfio.h>
14
15 struct iommufd_ctx;
16 struct iommu_group;
17 struct vfio_container;
18
19 struct vfio_device_file {
20 struct vfio_device *device;
21 struct vfio_group *group;
22
23 u8 access_granted;
24 u32 devid; /* only valid when iommufd is valid */
25 spinlock_t kvm_ref_lock; /* protect kvm field */
26 struct kvm *kvm;
27 struct iommufd_ctx *iommufd; /* protected by struct vfio_device_set::lock */
28 };
29
30 void vfio_device_put_registration(struct vfio_device *device);
31 bool vfio_device_try_get_registration(struct vfio_device *device);
32 int vfio_df_open(struct vfio_device_file *df);
33 void vfio_df_close(struct vfio_device_file *df);
34 struct vfio_device_file *
35 vfio_allocate_device_file(struct vfio_device *device);
36
37 extern const struct file_operations vfio_device_fops;
38
39 #ifdef CONFIG_VFIO_NOIOMMU
40 extern bool vfio_noiommu __read_mostly;
41 #else
42 enum { vfio_noiommu = false };
43 #endif
44
45 enum vfio_group_type {
46 /*
47 * Physical device with IOMMU backing.
48 */
49 VFIO_IOMMU,
50
51 /*
52 * Virtual device without IOMMU backing. The VFIO core fakes up an
53 * iommu_group as the iommu_group sysfs interface is part of the
54 * userspace ABI. The user of these devices must not be able to
55 * directly trigger unmediated DMA.
56 */
57 VFIO_EMULATED_IOMMU,
58
59 /*
60 * Physical device without IOMMU backing. The VFIO core fakes up an
61 * iommu_group as the iommu_group sysfs interface is part of the
62 * userspace ABI. Users can trigger unmediated DMA by the device,
63 * usage is highly dangerous, requires an explicit opt-in and will
64 * taint the kernel.
65 */
66 VFIO_NO_IOMMU,
67 };
68
69 #if IS_ENABLED(CONFIG_VFIO_GROUP)
70 struct vfio_group {
71 struct device dev;
72 struct cdev cdev;
73 /*
74 * When drivers is non-zero a driver is attached to the struct device
75 * that provided the iommu_group and thus the iommu_group is a valid
76 * pointer. When drivers is 0 the driver is being detached. Once users
77 * reaches 0 then the iommu_group is invalid.
78 */
79 refcount_t drivers;
80 unsigned int container_users;
81 struct iommu_group *iommu_group;
82 struct vfio_container *container;
83 struct list_head device_list;
84 struct mutex device_lock;
85 struct list_head vfio_next;
86 #if IS_ENABLED(CONFIG_VFIO_CONTAINER)
87 struct list_head container_next;
88 #endif
89 enum vfio_group_type type;
90 struct mutex group_lock;
91 struct kvm *kvm;
92 struct file *opened_file;
93 struct blocking_notifier_head notifier;
94 struct iommufd_ctx *iommufd;
95 spinlock_t kvm_ref_lock;
96 unsigned int cdev_device_open_cnt;
97 };
98
99 int vfio_device_block_group(struct vfio_device *device);
100 void vfio_device_unblock_group(struct vfio_device *device);
101 int vfio_device_set_group(struct vfio_device *device,
102 enum vfio_group_type type);
103 void vfio_device_remove_group(struct vfio_device *device);
104 void vfio_device_group_register(struct vfio_device *device);
105 void vfio_device_group_unregister(struct vfio_device *device);
106 int vfio_device_group_use_iommu(struct vfio_device *device);
107 void vfio_device_group_unuse_iommu(struct vfio_device *device);
108 void vfio_df_group_close(struct vfio_device_file *df);
109 struct vfio_group *vfio_group_from_file(struct file *file);
110 bool vfio_group_enforced_coherent(struct vfio_group *group);
111 void vfio_group_set_kvm(struct vfio_group *group, struct kvm *kvm);
112 bool vfio_device_has_container(struct vfio_device *device);
113 int __init vfio_group_init(void);
114 void vfio_group_cleanup(void);
115
vfio_device_is_noiommu(struct vfio_device * vdev)116 static inline bool vfio_device_is_noiommu(struct vfio_device *vdev)
117 {
118 return IS_ENABLED(CONFIG_VFIO_NOIOMMU) &&
119 vdev->group->type == VFIO_NO_IOMMU;
120 }
121 #else
122 struct vfio_group;
123
vfio_device_block_group(struct vfio_device * device)124 static inline int vfio_device_block_group(struct vfio_device *device)
125 {
126 return 0;
127 }
128
vfio_device_unblock_group(struct vfio_device * device)129 static inline void vfio_device_unblock_group(struct vfio_device *device)
130 {
131 }
132
vfio_device_set_group(struct vfio_device * device,enum vfio_group_type type)133 static inline int vfio_device_set_group(struct vfio_device *device,
134 enum vfio_group_type type)
135 {
136 return 0;
137 }
138
vfio_device_remove_group(struct vfio_device * device)139 static inline void vfio_device_remove_group(struct vfio_device *device)
140 {
141 }
142
vfio_device_group_register(struct vfio_device * device)143 static inline void vfio_device_group_register(struct vfio_device *device)
144 {
145 }
146
vfio_device_group_unregister(struct vfio_device * device)147 static inline void vfio_device_group_unregister(struct vfio_device *device)
148 {
149 }
150
vfio_device_group_use_iommu(struct vfio_device * device)151 static inline int vfio_device_group_use_iommu(struct vfio_device *device)
152 {
153 return -EOPNOTSUPP;
154 }
155
vfio_device_group_unuse_iommu(struct vfio_device * device)156 static inline void vfio_device_group_unuse_iommu(struct vfio_device *device)
157 {
158 }
159
vfio_df_group_close(struct vfio_device_file * df)160 static inline void vfio_df_group_close(struct vfio_device_file *df)
161 {
162 }
163
vfio_group_from_file(struct file * file)164 static inline struct vfio_group *vfio_group_from_file(struct file *file)
165 {
166 return NULL;
167 }
168
vfio_group_enforced_coherent(struct vfio_group * group)169 static inline bool vfio_group_enforced_coherent(struct vfio_group *group)
170 {
171 return true;
172 }
173
vfio_group_set_kvm(struct vfio_group * group,struct kvm * kvm)174 static inline void vfio_group_set_kvm(struct vfio_group *group, struct kvm *kvm)
175 {
176 }
177
vfio_device_has_container(struct vfio_device * device)178 static inline bool vfio_device_has_container(struct vfio_device *device)
179 {
180 return false;
181 }
182
vfio_group_init(void)183 static inline int __init vfio_group_init(void)
184 {
185 return 0;
186 }
187
vfio_group_cleanup(void)188 static inline void vfio_group_cleanup(void)
189 {
190 }
191
vfio_device_is_noiommu(struct vfio_device * vdev)192 static inline bool vfio_device_is_noiommu(struct vfio_device *vdev)
193 {
194 return false;
195 }
196 #endif /* CONFIG_VFIO_GROUP */
197
198 #if IS_ENABLED(CONFIG_VFIO_CONTAINER)
199 /**
200 * struct vfio_iommu_driver_ops - VFIO IOMMU driver callbacks
201 */
202 struct vfio_iommu_driver_ops {
203 char *name;
204 struct module *owner;
205 void *(*open)(unsigned long arg);
206 void (*release)(void *iommu_data);
207 long (*ioctl)(void *iommu_data, unsigned int cmd,
208 unsigned long arg);
209 int (*attach_group)(void *iommu_data,
210 struct iommu_group *group,
211 enum vfio_group_type);
212 void (*detach_group)(void *iommu_data,
213 struct iommu_group *group);
214 int (*pin_pages)(void *iommu_data,
215 struct iommu_group *group,
216 dma_addr_t user_iova,
217 int npage, int prot,
218 struct page **pages);
219 void (*unpin_pages)(void *iommu_data,
220 dma_addr_t user_iova, int npage);
221 void (*register_device)(void *iommu_data,
222 struct vfio_device *vdev);
223 void (*unregister_device)(void *iommu_data,
224 struct vfio_device *vdev);
225 int (*dma_rw)(void *iommu_data, dma_addr_t user_iova,
226 void *data, size_t count, bool write);
227 struct iommu_domain *(*group_iommu_domain)(void *iommu_data,
228 struct iommu_group *group);
229 };
230
231 struct vfio_iommu_driver {
232 const struct vfio_iommu_driver_ops *ops;
233 struct list_head vfio_next;
234 };
235
236 int vfio_register_iommu_driver(const struct vfio_iommu_driver_ops *ops);
237 void vfio_unregister_iommu_driver(const struct vfio_iommu_driver_ops *ops);
238
239 struct vfio_container *vfio_container_from_file(struct file *filep);
240 int vfio_group_use_container(struct vfio_group *group);
241 void vfio_group_unuse_container(struct vfio_group *group);
242 int vfio_container_attach_group(struct vfio_container *container,
243 struct vfio_group *group);
244 void vfio_group_detach_container(struct vfio_group *group);
245 void vfio_device_container_register(struct vfio_device *device);
246 void vfio_device_container_unregister(struct vfio_device *device);
247 int vfio_device_container_pin_pages(struct vfio_device *device,
248 dma_addr_t iova, int npage,
249 int prot, struct page **pages);
250 void vfio_device_container_unpin_pages(struct vfio_device *device,
251 dma_addr_t iova, int npage);
252 int vfio_device_container_dma_rw(struct vfio_device *device,
253 dma_addr_t iova, void *data,
254 size_t len, bool write);
255
256 int __init vfio_container_init(void);
257 void vfio_container_cleanup(void);
258 #else
259 static inline struct vfio_container *
vfio_container_from_file(struct file * filep)260 vfio_container_from_file(struct file *filep)
261 {
262 return NULL;
263 }
264
vfio_group_use_container(struct vfio_group * group)265 static inline int vfio_group_use_container(struct vfio_group *group)
266 {
267 return -EOPNOTSUPP;
268 }
269
vfio_group_unuse_container(struct vfio_group * group)270 static inline void vfio_group_unuse_container(struct vfio_group *group)
271 {
272 }
273
vfio_container_attach_group(struct vfio_container * container,struct vfio_group * group)274 static inline int vfio_container_attach_group(struct vfio_container *container,
275 struct vfio_group *group)
276 {
277 return -EOPNOTSUPP;
278 }
279
vfio_group_detach_container(struct vfio_group * group)280 static inline void vfio_group_detach_container(struct vfio_group *group)
281 {
282 }
283
vfio_device_container_register(struct vfio_device * device)284 static inline void vfio_device_container_register(struct vfio_device *device)
285 {
286 }
287
vfio_device_container_unregister(struct vfio_device * device)288 static inline void vfio_device_container_unregister(struct vfio_device *device)
289 {
290 }
291
vfio_device_container_pin_pages(struct vfio_device * device,dma_addr_t iova,int npage,int prot,struct page ** pages)292 static inline int vfio_device_container_pin_pages(struct vfio_device *device,
293 dma_addr_t iova, int npage,
294 int prot, struct page **pages)
295 {
296 return -EOPNOTSUPP;
297 }
298
vfio_device_container_unpin_pages(struct vfio_device * device,dma_addr_t iova,int npage)299 static inline void vfio_device_container_unpin_pages(struct vfio_device *device,
300 dma_addr_t iova, int npage)
301 {
302 }
303
vfio_device_container_dma_rw(struct vfio_device * device,dma_addr_t iova,void * data,size_t len,bool write)304 static inline int vfio_device_container_dma_rw(struct vfio_device *device,
305 dma_addr_t iova, void *data,
306 size_t len, bool write)
307 {
308 return -EOPNOTSUPP;
309 }
310
vfio_container_init(void)311 static inline int vfio_container_init(void)
312 {
313 return 0;
314 }
vfio_container_cleanup(void)315 static inline void vfio_container_cleanup(void)
316 {
317 }
318 #endif
319
320 #if IS_ENABLED(CONFIG_IOMMUFD)
321 bool vfio_iommufd_device_has_compat_ioas(struct vfio_device *vdev,
322 struct iommufd_ctx *ictx);
323 int vfio_df_iommufd_bind(struct vfio_device_file *df);
324 void vfio_df_iommufd_unbind(struct vfio_device_file *df);
325 int vfio_iommufd_compat_attach_ioas(struct vfio_device *device,
326 struct iommufd_ctx *ictx);
327 #else
328 static inline bool
vfio_iommufd_device_has_compat_ioas(struct vfio_device * vdev,struct iommufd_ctx * ictx)329 vfio_iommufd_device_has_compat_ioas(struct vfio_device *vdev,
330 struct iommufd_ctx *ictx)
331 {
332 return false;
333 }
334
vfio_df_iommufd_bind(struct vfio_device_file * fd)335 static inline int vfio_df_iommufd_bind(struct vfio_device_file *fd)
336 {
337 return -EOPNOTSUPP;
338 }
339
vfio_df_iommufd_unbind(struct vfio_device_file * df)340 static inline void vfio_df_iommufd_unbind(struct vfio_device_file *df)
341 {
342 }
343
344 static inline int
vfio_iommufd_compat_attach_ioas(struct vfio_device * device,struct iommufd_ctx * ictx)345 vfio_iommufd_compat_attach_ioas(struct vfio_device *device,
346 struct iommufd_ctx *ictx)
347 {
348 return -EOPNOTSUPP;
349 }
350 #endif
351
352 int vfio_df_ioctl_attach_pt(struct vfio_device_file *df,
353 struct vfio_device_attach_iommufd_pt __user *arg);
354 int vfio_df_ioctl_detach_pt(struct vfio_device_file *df,
355 struct vfio_device_detach_iommufd_pt __user *arg);
356
357 #if IS_ENABLED(CONFIG_VFIO_DEVICE_CDEV)
358 void vfio_init_device_cdev(struct vfio_device *device);
359
vfio_device_add(struct vfio_device * device)360 static inline int vfio_device_add(struct vfio_device *device)
361 {
362 /* cdev does not support noiommu device */
363 if (vfio_device_is_noiommu(device))
364 return device_add(&device->device);
365 vfio_init_device_cdev(device);
366 return cdev_device_add(&device->cdev, &device->device);
367 }
368
vfio_device_del(struct vfio_device * device)369 static inline void vfio_device_del(struct vfio_device *device)
370 {
371 if (vfio_device_is_noiommu(device))
372 device_del(&device->device);
373 else
374 cdev_device_del(&device->cdev, &device->device);
375 }
376
377 int vfio_device_fops_cdev_open(struct inode *inode, struct file *filep);
378 long vfio_df_ioctl_bind_iommufd(struct vfio_device_file *df,
379 struct vfio_device_bind_iommufd __user *arg);
380 void vfio_df_unbind_iommufd(struct vfio_device_file *df);
381 int vfio_cdev_init(struct class *device_class);
382 void vfio_cdev_cleanup(void);
383 #else
vfio_init_device_cdev(struct vfio_device * device)384 static inline void vfio_init_device_cdev(struct vfio_device *device)
385 {
386 }
387
vfio_device_add(struct vfio_device * device)388 static inline int vfio_device_add(struct vfio_device *device)
389 {
390 return device_add(&device->device);
391 }
392
vfio_device_del(struct vfio_device * device)393 static inline void vfio_device_del(struct vfio_device *device)
394 {
395 device_del(&device->device);
396 }
397
vfio_device_fops_cdev_open(struct inode * inode,struct file * filep)398 static inline int vfio_device_fops_cdev_open(struct inode *inode,
399 struct file *filep)
400 {
401 return 0;
402 }
403
vfio_df_ioctl_bind_iommufd(struct vfio_device_file * df,struct vfio_device_bind_iommufd __user * arg)404 static inline long vfio_df_ioctl_bind_iommufd(struct vfio_device_file *df,
405 struct vfio_device_bind_iommufd __user *arg)
406 {
407 return -ENOTTY;
408 }
409
vfio_df_unbind_iommufd(struct vfio_device_file * df)410 static inline void vfio_df_unbind_iommufd(struct vfio_device_file *df)
411 {
412 }
413
vfio_cdev_init(struct class * device_class)414 static inline int vfio_cdev_init(struct class *device_class)
415 {
416 return 0;
417 }
418
vfio_cdev_cleanup(void)419 static inline void vfio_cdev_cleanup(void)
420 {
421 }
422 #endif /* CONFIG_VFIO_DEVICE_CDEV */
423
424 #if IS_ENABLED(CONFIG_VFIO_VIRQFD)
425 int __init vfio_virqfd_init(void);
426 void vfio_virqfd_exit(void);
427 #else
vfio_virqfd_init(void)428 static inline int __init vfio_virqfd_init(void)
429 {
430 return 0;
431 }
vfio_virqfd_exit(void)432 static inline void vfio_virqfd_exit(void)
433 {
434 }
435 #endif
436
437 #ifdef CONFIG_HAVE_KVM
438 void vfio_device_get_kvm_safe(struct vfio_device *device, struct kvm *kvm);
439 void vfio_device_put_kvm(struct vfio_device *device);
440 #else
vfio_device_get_kvm_safe(struct vfio_device * device,struct kvm * kvm)441 static inline void vfio_device_get_kvm_safe(struct vfio_device *device,
442 struct kvm *kvm)
443 {
444 }
445
vfio_device_put_kvm(struct vfio_device * device)446 static inline void vfio_device_put_kvm(struct vfio_device *device)
447 {
448 }
449 #endif
450
451 #endif
452