xref: /openbmc/qemu/hw/display/vhost-user-gpu.c (revision de799beb)
1 /*
2  * vhost-user GPU Device
3  *
4  * Copyright Red Hat, Inc. 2018
5  *
6  * Authors:
7  *     Marc-André Lureau <marcandre.lureau@redhat.com>
8  *
9  * This work is licensed under the terms of the GNU GPL, version 2 or later.
10  * See the COPYING file in the top-level directory.
11  */
12 
13 #include "qemu/osdep.h"
14 #include "hw/qdev-properties.h"
15 #include "hw/virtio/virtio-gpu.h"
16 #include "chardev/char-fe.h"
17 #include "qapi/error.h"
18 #include "migration/blocker.h"
19 
20 typedef enum VhostUserGpuRequest {
21     VHOST_USER_GPU_NONE = 0,
22     VHOST_USER_GPU_GET_PROTOCOL_FEATURES,
23     VHOST_USER_GPU_SET_PROTOCOL_FEATURES,
24     VHOST_USER_GPU_GET_DISPLAY_INFO,
25     VHOST_USER_GPU_CURSOR_POS,
26     VHOST_USER_GPU_CURSOR_POS_HIDE,
27     VHOST_USER_GPU_CURSOR_UPDATE,
28     VHOST_USER_GPU_SCANOUT,
29     VHOST_USER_GPU_UPDATE,
30     VHOST_USER_GPU_DMABUF_SCANOUT,
31     VHOST_USER_GPU_DMABUF_UPDATE,
32 } VhostUserGpuRequest;
33 
34 typedef struct VhostUserGpuDisplayInfoReply {
35     struct virtio_gpu_resp_display_info info;
36 } VhostUserGpuDisplayInfoReply;
37 
38 typedef struct VhostUserGpuCursorPos {
39     uint32_t scanout_id;
40     uint32_t x;
41     uint32_t y;
42 } QEMU_PACKED VhostUserGpuCursorPos;
43 
44 typedef struct VhostUserGpuCursorUpdate {
45     VhostUserGpuCursorPos pos;
46     uint32_t hot_x;
47     uint32_t hot_y;
48     uint32_t data[64 * 64];
49 } QEMU_PACKED VhostUserGpuCursorUpdate;
50 
51 typedef struct VhostUserGpuScanout {
52     uint32_t scanout_id;
53     uint32_t width;
54     uint32_t height;
55 } QEMU_PACKED VhostUserGpuScanout;
56 
57 typedef struct VhostUserGpuUpdate {
58     uint32_t scanout_id;
59     uint32_t x;
60     uint32_t y;
61     uint32_t width;
62     uint32_t height;
63     uint8_t data[];
64 } QEMU_PACKED VhostUserGpuUpdate;
65 
66 typedef struct VhostUserGpuDMABUFScanout {
67     uint32_t scanout_id;
68     uint32_t x;
69     uint32_t y;
70     uint32_t width;
71     uint32_t height;
72     uint32_t fd_width;
73     uint32_t fd_height;
74     uint32_t fd_stride;
75     uint32_t fd_flags;
76     int fd_drm_fourcc;
77 } QEMU_PACKED VhostUserGpuDMABUFScanout;
78 
79 typedef struct VhostUserGpuMsg {
80     uint32_t request; /* VhostUserGpuRequest */
81     uint32_t flags;
82     uint32_t size; /* the following payload size */
83     union {
84         VhostUserGpuCursorPos cursor_pos;
85         VhostUserGpuCursorUpdate cursor_update;
86         VhostUserGpuScanout scanout;
87         VhostUserGpuUpdate update;
88         VhostUserGpuDMABUFScanout dmabuf_scanout;
89         struct virtio_gpu_resp_display_info display_info;
90         uint64_t u64;
91     } payload;
92 } QEMU_PACKED VhostUserGpuMsg;
93 
94 static VhostUserGpuMsg m __attribute__ ((unused));
95 #define VHOST_USER_GPU_HDR_SIZE \
96     (sizeof(m.request) + sizeof(m.size) + sizeof(m.flags))
97 
98 #define VHOST_USER_GPU_MSG_FLAG_REPLY 0x4
99 
100 static void vhost_user_gpu_update_blocked(VhostUserGPU *g, bool blocked);
101 
102 static void
103 vhost_user_gpu_handle_cursor(VhostUserGPU *g, VhostUserGpuMsg *msg)
104 {
105     VhostUserGpuCursorPos *pos = &msg->payload.cursor_pos;
106     struct virtio_gpu_scanout *s;
107 
108     if (pos->scanout_id >= g->parent_obj.conf.max_outputs) {
109         return;
110     }
111     s = &g->parent_obj.scanout[pos->scanout_id];
112 
113     if (msg->request == VHOST_USER_GPU_CURSOR_UPDATE) {
114         VhostUserGpuCursorUpdate *up = &msg->payload.cursor_update;
115         if (!s->current_cursor) {
116             s->current_cursor = cursor_alloc(64, 64);
117         }
118 
119         s->current_cursor->hot_x = up->hot_x;
120         s->current_cursor->hot_y = up->hot_y;
121 
122         memcpy(s->current_cursor->data, up->data,
123                64 * 64 * sizeof(uint32_t));
124 
125         dpy_cursor_define(s->con, s->current_cursor);
126     }
127 
128     dpy_mouse_set(s->con, pos->x, pos->y,
129                   msg->request != VHOST_USER_GPU_CURSOR_POS_HIDE);
130 }
131 
132 static void
133 vhost_user_gpu_send_msg(VhostUserGPU *g, const VhostUserGpuMsg *msg)
134 {
135     qemu_chr_fe_write(&g->vhost_chr, (uint8_t *)msg,
136                       VHOST_USER_GPU_HDR_SIZE + msg->size);
137 }
138 
139 static void
140 vhost_user_gpu_unblock(VhostUserGPU *g)
141 {
142     VhostUserGpuMsg msg = {
143         .request = VHOST_USER_GPU_DMABUF_UPDATE,
144         .flags = VHOST_USER_GPU_MSG_FLAG_REPLY,
145     };
146 
147     vhost_user_gpu_send_msg(g, &msg);
148 }
149 
150 static void
151 vhost_user_gpu_handle_display(VhostUserGPU *g, VhostUserGpuMsg *msg)
152 {
153     QemuConsole *con = NULL;
154     struct virtio_gpu_scanout *s;
155 
156     switch (msg->request) {
157     case VHOST_USER_GPU_GET_PROTOCOL_FEATURES: {
158         VhostUserGpuMsg reply = {
159             .request = msg->request,
160             .flags = VHOST_USER_GPU_MSG_FLAG_REPLY,
161             .size = sizeof(uint64_t),
162         };
163 
164         vhost_user_gpu_send_msg(g, &reply);
165         break;
166     }
167     case VHOST_USER_GPU_SET_PROTOCOL_FEATURES: {
168         break;
169     }
170     case VHOST_USER_GPU_GET_DISPLAY_INFO: {
171         struct virtio_gpu_resp_display_info display_info = { {} };
172         VhostUserGpuMsg reply = {
173             .request = msg->request,
174             .flags = VHOST_USER_GPU_MSG_FLAG_REPLY,
175             .size = sizeof(struct virtio_gpu_resp_display_info),
176         };
177 
178         display_info.hdr.type = VIRTIO_GPU_RESP_OK_DISPLAY_INFO;
179         virtio_gpu_base_fill_display_info(VIRTIO_GPU_BASE(g), &display_info);
180         memcpy(&reply.payload.display_info, &display_info,
181                sizeof(display_info));
182         vhost_user_gpu_send_msg(g, &reply);
183         break;
184     }
185     case VHOST_USER_GPU_SCANOUT: {
186         VhostUserGpuScanout *m = &msg->payload.scanout;
187 
188         if (m->scanout_id >= g->parent_obj.conf.max_outputs) {
189             return;
190         }
191 
192         g->parent_obj.enable = 1;
193         s = &g->parent_obj.scanout[m->scanout_id];
194         con = s->con;
195 
196         if (m->width == 0) {
197             dpy_gfx_replace_surface(con, NULL);
198         } else {
199             s->ds = qemu_create_displaysurface(m->width, m->height);
200             /* replace surface on next update */
201         }
202 
203         break;
204     }
205     case VHOST_USER_GPU_DMABUF_SCANOUT: {
206         VhostUserGpuDMABUFScanout *m = &msg->payload.dmabuf_scanout;
207         int fd = qemu_chr_fe_get_msgfd(&g->vhost_chr);
208         QemuDmaBuf *dmabuf;
209 
210         if (m->scanout_id >= g->parent_obj.conf.max_outputs) {
211             error_report("invalid scanout: %d", m->scanout_id);
212             if (fd >= 0) {
213                 close(fd);
214             }
215             break;
216         }
217 
218         g->parent_obj.enable = 1;
219         con = g->parent_obj.scanout[m->scanout_id].con;
220         dmabuf = &g->dmabuf[m->scanout_id];
221         if (dmabuf->fd >= 0) {
222             close(dmabuf->fd);
223             dmabuf->fd = -1;
224         }
225         dpy_gl_release_dmabuf(con, dmabuf);
226         if (fd == -1) {
227             dpy_gl_scanout_disable(con);
228             break;
229         }
230         *dmabuf = (QemuDmaBuf) {
231             .fd = fd,
232             .width = m->fd_width,
233             .height = m->fd_height,
234             .stride = m->fd_stride,
235             .fourcc = m->fd_drm_fourcc,
236             .y0_top = m->fd_flags & VIRTIO_GPU_RESOURCE_FLAG_Y_0_TOP,
237         };
238         dpy_gl_scanout_dmabuf(con, dmabuf);
239         break;
240     }
241     case VHOST_USER_GPU_DMABUF_UPDATE: {
242         VhostUserGpuUpdate *m = &msg->payload.update;
243 
244         if (m->scanout_id >= g->parent_obj.conf.max_outputs ||
245             !g->parent_obj.scanout[m->scanout_id].con) {
246             error_report("invalid scanout update: %d", m->scanout_id);
247             vhost_user_gpu_unblock(g);
248             break;
249         }
250 
251         con = g->parent_obj.scanout[m->scanout_id].con;
252         if (!console_has_gl(con)) {
253             error_report("console doesn't support GL!");
254             vhost_user_gpu_unblock(g);
255             break;
256         }
257         g->backend_blocked = true;
258         dpy_gl_update(con, m->x, m->y, m->width, m->height);
259         break;
260     }
261     case VHOST_USER_GPU_UPDATE: {
262         VhostUserGpuUpdate *m = &msg->payload.update;
263 
264         if (m->scanout_id >= g->parent_obj.conf.max_outputs) {
265             break;
266         }
267         s = &g->parent_obj.scanout[m->scanout_id];
268         con = s->con;
269         pixman_image_t *image =
270             pixman_image_create_bits(PIXMAN_x8r8g8b8,
271                                      m->width,
272                                      m->height,
273                                      (uint32_t *)m->data,
274                                      m->width * 4);
275 
276         pixman_image_composite(PIXMAN_OP_SRC,
277                                image, NULL, s->ds->image,
278                                0, 0, 0, 0, m->x, m->y, m->width, m->height);
279 
280         pixman_image_unref(image);
281         if (qemu_console_surface(con) != s->ds) {
282             dpy_gfx_replace_surface(con, s->ds);
283         } else {
284             dpy_gfx_update(con, m->x, m->y, m->width, m->height);
285         }
286         break;
287     }
288     default:
289         g_warning("unhandled message %d %d", msg->request, msg->size);
290     }
291 
292     if (con && qemu_console_is_gl_blocked(con)) {
293         vhost_user_gpu_update_blocked(g, true);
294     }
295 }
296 
297 static void
298 vhost_user_gpu_chr_read(void *opaque)
299 {
300     VhostUserGPU *g = opaque;
301     VhostUserGpuMsg *msg = NULL;
302     VhostUserGpuRequest request;
303     uint32_t size, flags;
304     int r;
305 
306     r = qemu_chr_fe_read_all(&g->vhost_chr,
307                              (uint8_t *)&request, sizeof(uint32_t));
308     if (r != sizeof(uint32_t)) {
309         error_report("failed to read msg header: %d, %d", r, errno);
310         goto end;
311     }
312 
313     r = qemu_chr_fe_read_all(&g->vhost_chr,
314                              (uint8_t *)&flags, sizeof(uint32_t));
315     if (r != sizeof(uint32_t)) {
316         error_report("failed to read msg flags");
317         goto end;
318     }
319 
320     r = qemu_chr_fe_read_all(&g->vhost_chr,
321                              (uint8_t *)&size, sizeof(uint32_t));
322     if (r != sizeof(uint32_t)) {
323         error_report("failed to read msg size");
324         goto end;
325     }
326 
327     msg = g_malloc(VHOST_USER_GPU_HDR_SIZE + size);
328 
329     r = qemu_chr_fe_read_all(&g->vhost_chr,
330                              (uint8_t *)&msg->payload, size);
331     if (r != size) {
332         error_report("failed to read msg payload %d != %d", r, size);
333         goto end;
334     }
335 
336     msg->request = request;
337     msg->flags = size;
338     msg->size = size;
339 
340     if (request == VHOST_USER_GPU_CURSOR_UPDATE ||
341         request == VHOST_USER_GPU_CURSOR_POS ||
342         request == VHOST_USER_GPU_CURSOR_POS_HIDE) {
343         vhost_user_gpu_handle_cursor(g, msg);
344     } else {
345         vhost_user_gpu_handle_display(g, msg);
346     }
347 
348 end:
349     g_free(msg);
350 }
351 
352 static void
353 vhost_user_gpu_update_blocked(VhostUserGPU *g, bool blocked)
354 {
355     qemu_set_fd_handler(g->vhost_gpu_fd,
356                         blocked ? NULL : vhost_user_gpu_chr_read, NULL, g);
357 }
358 
359 static void
360 vhost_user_gpu_gl_flushed(VirtIOGPUBase *b)
361 {
362     VhostUserGPU *g = VHOST_USER_GPU(b);
363 
364     if (g->backend_blocked) {
365         vhost_user_gpu_unblock(VHOST_USER_GPU(g));
366         g->backend_blocked = false;
367     }
368 
369     vhost_user_gpu_update_blocked(VHOST_USER_GPU(g), false);
370 }
371 
372 static bool
373 vhost_user_gpu_do_set_socket(VhostUserGPU *g, Error **errp)
374 {
375     Chardev *chr;
376     int sv[2];
377 
378     if (socketpair(PF_UNIX, SOCK_STREAM, 0, sv) == -1) {
379         error_setg_errno(errp, errno, "socketpair() failed");
380         return false;
381     }
382 
383     chr = CHARDEV(object_new(TYPE_CHARDEV_SOCKET));
384     if (!chr || qemu_chr_add_client(chr, sv[0]) == -1) {
385         error_setg(errp, "Failed to make socket chardev");
386         goto err;
387     }
388     if (!qemu_chr_fe_init(&g->vhost_chr, chr, errp)) {
389         goto err;
390     }
391     if (vhost_user_gpu_set_socket(&g->vhost->dev, sv[1]) < 0) {
392         error_setg(errp, "Failed to set vhost-user-gpu socket");
393         qemu_chr_fe_deinit(&g->vhost_chr, false);
394         goto err;
395     }
396 
397     g->vhost_gpu_fd = sv[0];
398     vhost_user_gpu_update_blocked(g, false);
399     close(sv[1]);
400     return true;
401 
402 err:
403     close(sv[0]);
404     close(sv[1]);
405     if (chr) {
406         object_unref(OBJECT(chr));
407     }
408     return false;
409 }
410 
411 static void
412 vhost_user_gpu_get_config(VirtIODevice *vdev, uint8_t *config_data)
413 {
414     VhostUserGPU *g = VHOST_USER_GPU(vdev);
415     VirtIOGPUBase *b = VIRTIO_GPU_BASE(vdev);
416     struct virtio_gpu_config *vgconfig =
417         (struct virtio_gpu_config *)config_data;
418     Error *local_err = NULL;
419     int ret;
420 
421     memset(config_data, 0, sizeof(struct virtio_gpu_config));
422 
423     ret = vhost_dev_get_config(&g->vhost->dev,
424                                config_data, sizeof(struct virtio_gpu_config),
425                                &local_err);
426     if (ret) {
427         error_report_err(local_err);
428         return;
429     }
430 
431     /* those fields are managed by qemu */
432     vgconfig->num_scanouts = b->virtio_config.num_scanouts;
433     vgconfig->events_read = b->virtio_config.events_read;
434     vgconfig->events_clear = b->virtio_config.events_clear;
435 }
436 
437 static void
438 vhost_user_gpu_set_config(VirtIODevice *vdev,
439                           const uint8_t *config_data)
440 {
441     VhostUserGPU *g = VHOST_USER_GPU(vdev);
442     VirtIOGPUBase *b = VIRTIO_GPU_BASE(vdev);
443     const struct virtio_gpu_config *vgconfig =
444         (const struct virtio_gpu_config *)config_data;
445     int ret;
446 
447     if (vgconfig->events_clear) {
448         b->virtio_config.events_read &= ~vgconfig->events_clear;
449     }
450 
451     ret = vhost_dev_set_config(&g->vhost->dev, config_data,
452                                0, sizeof(struct virtio_gpu_config),
453                                VHOST_SET_CONFIG_TYPE_MASTER);
454     if (ret) {
455         error_report("vhost-user-gpu: set device config space failed");
456         return;
457     }
458 }
459 
460 static void
461 vhost_user_gpu_set_status(VirtIODevice *vdev, uint8_t val)
462 {
463     VhostUserGPU *g = VHOST_USER_GPU(vdev);
464     Error *err = NULL;
465 
466     if (val & VIRTIO_CONFIG_S_DRIVER_OK && vdev->vm_running) {
467         if (!vhost_user_gpu_do_set_socket(g, &err)) {
468             error_report_err(err);
469             return;
470         }
471         vhost_user_backend_start(g->vhost);
472     } else {
473         /* unblock any wait and stop processing */
474         if (g->vhost_gpu_fd != -1) {
475             vhost_user_gpu_update_blocked(g, true);
476             qemu_chr_fe_deinit(&g->vhost_chr, true);
477             g->vhost_gpu_fd = -1;
478         }
479         vhost_user_backend_stop(g->vhost);
480     }
481 }
482 
483 static bool
484 vhost_user_gpu_guest_notifier_pending(VirtIODevice *vdev, int idx)
485 {
486     VhostUserGPU *g = VHOST_USER_GPU(vdev);
487 
488     return vhost_virtqueue_pending(&g->vhost->dev, idx);
489 }
490 
491 static void
492 vhost_user_gpu_guest_notifier_mask(VirtIODevice *vdev, int idx, bool mask)
493 {
494     VhostUserGPU *g = VHOST_USER_GPU(vdev);
495 
496     vhost_virtqueue_mask(&g->vhost->dev, vdev, idx, mask);
497 }
498 
499 static void
500 vhost_user_gpu_instance_init(Object *obj)
501 {
502     VhostUserGPU *g = VHOST_USER_GPU(obj);
503 
504     g->vhost = VHOST_USER_BACKEND(object_new(TYPE_VHOST_USER_BACKEND));
505     object_property_add_alias(obj, "chardev",
506                               OBJECT(g->vhost), "chardev");
507 }
508 
509 static void
510 vhost_user_gpu_instance_finalize(Object *obj)
511 {
512     VhostUserGPU *g = VHOST_USER_GPU(obj);
513 
514     object_unref(OBJECT(g->vhost));
515 }
516 
517 static void
518 vhost_user_gpu_reset(VirtIODevice *vdev)
519 {
520     VhostUserGPU *g = VHOST_USER_GPU(vdev);
521 
522     virtio_gpu_base_reset(VIRTIO_GPU_BASE(vdev));
523 
524     vhost_user_backend_stop(g->vhost);
525 }
526 
527 static int
528 vhost_user_gpu_config_change(struct vhost_dev *dev)
529 {
530     error_report("vhost-user-gpu: unhandled backend config change");
531     return -1;
532 }
533 
534 static const VhostDevConfigOps config_ops = {
535     .vhost_dev_config_notifier = vhost_user_gpu_config_change,
536 };
537 
538 static void
539 vhost_user_gpu_device_realize(DeviceState *qdev, Error **errp)
540 {
541     VhostUserGPU *g = VHOST_USER_GPU(qdev);
542     VirtIODevice *vdev = VIRTIO_DEVICE(g);
543 
544     vhost_dev_set_config_notifier(&g->vhost->dev, &config_ops);
545     if (vhost_user_backend_dev_init(g->vhost, vdev, 2, errp) < 0) {
546         return;
547     }
548 
549     /* existing backend may send DMABUF, so let's add that requirement */
550     g->parent_obj.conf.flags |= 1 << VIRTIO_GPU_FLAG_DMABUF_ENABLED;
551     if (virtio_has_feature(g->vhost->dev.features, VIRTIO_GPU_F_VIRGL)) {
552         g->parent_obj.conf.flags |= 1 << VIRTIO_GPU_FLAG_VIRGL_ENABLED;
553     }
554     if (virtio_has_feature(g->vhost->dev.features, VIRTIO_GPU_F_EDID)) {
555         g->parent_obj.conf.flags |= 1 << VIRTIO_GPU_FLAG_EDID_ENABLED;
556     } else {
557         error_report("EDID requested but the backend doesn't support it.");
558         g->parent_obj.conf.flags &= ~(1 << VIRTIO_GPU_FLAG_EDID_ENABLED);
559     }
560 
561     if (!virtio_gpu_base_device_realize(qdev, NULL, NULL, errp)) {
562         return;
563     }
564 
565     g->vhost_gpu_fd = -1;
566 }
567 
568 static struct vhost_dev *vhost_user_gpu_get_vhost(VirtIODevice *vdev)
569 {
570     VhostUserGPU *g = VHOST_USER_GPU(vdev);
571     return &g->vhost->dev;
572 }
573 
574 static Property vhost_user_gpu_properties[] = {
575     VIRTIO_GPU_BASE_PROPERTIES(VhostUserGPU, parent_obj.conf),
576     DEFINE_PROP_END_OF_LIST(),
577 };
578 
579 static void
580 vhost_user_gpu_class_init(ObjectClass *klass, void *data)
581 {
582     DeviceClass *dc = DEVICE_CLASS(klass);
583     VirtioDeviceClass *vdc = VIRTIO_DEVICE_CLASS(klass);
584     VirtIOGPUBaseClass *vgc = VIRTIO_GPU_BASE_CLASS(klass);
585 
586     vgc->gl_flushed = vhost_user_gpu_gl_flushed;
587 
588     vdc->realize = vhost_user_gpu_device_realize;
589     vdc->reset = vhost_user_gpu_reset;
590     vdc->set_status   = vhost_user_gpu_set_status;
591     vdc->guest_notifier_mask = vhost_user_gpu_guest_notifier_mask;
592     vdc->guest_notifier_pending = vhost_user_gpu_guest_notifier_pending;
593     vdc->get_config = vhost_user_gpu_get_config;
594     vdc->set_config = vhost_user_gpu_set_config;
595     vdc->get_vhost = vhost_user_gpu_get_vhost;
596 
597     device_class_set_props(dc, vhost_user_gpu_properties);
598 }
599 
600 static const TypeInfo vhost_user_gpu_info = {
601     .name = TYPE_VHOST_USER_GPU,
602     .parent = TYPE_VIRTIO_GPU_BASE,
603     .instance_size = sizeof(VhostUserGPU),
604     .instance_init = vhost_user_gpu_instance_init,
605     .instance_finalize = vhost_user_gpu_instance_finalize,
606     .class_init = vhost_user_gpu_class_init,
607 };
608 module_obj(TYPE_VHOST_USER_GPU);
609 module_kconfig(VHOST_USER_GPU);
610 
611 static void vhost_user_gpu_register_types(void)
612 {
613     type_register_static(&vhost_user_gpu_info);
614 }
615 
616 type_init(vhost_user_gpu_register_types)
617