xref: /openbmc/qemu/target/arm/kvm.c (revision f3635813)
1 /*
2  * ARM implementation of KVM hooks
3  *
4  * Copyright Christoffer Dall 2009-2010
5  *
6  * This work is licensed under the terms of the GNU GPL, version 2 or later.
7  * See the COPYING file in the top-level directory.
8  *
9  */
10 
11 #include "qemu/osdep.h"
12 #include <sys/ioctl.h>
13 
14 #include <linux/kvm.h>
15 
16 #include "qemu-common.h"
17 #include "qemu/timer.h"
18 #include "qemu/error-report.h"
19 #include "qemu/main-loop.h"
20 #include "sysemu/sysemu.h"
21 #include "sysemu/kvm.h"
22 #include "sysemu/kvm_int.h"
23 #include "kvm_arm.h"
24 #include "cpu.h"
25 #include "trace.h"
26 #include "internals.h"
27 #include "hw/pci/pci.h"
28 #include "exec/memattrs.h"
29 #include "exec/address-spaces.h"
30 #include "hw/boards.h"
31 #include "hw/irq.h"
32 #include "qemu/log.h"
33 
34 const KVMCapabilityInfo kvm_arch_required_capabilities[] = {
35     KVM_CAP_LAST_INFO
36 };
37 
38 static bool cap_has_mp_state;
39 static bool cap_has_inject_serror_esr;
40 
41 static ARMHostCPUFeatures arm_host_cpu_features;
42 
43 int kvm_arm_vcpu_init(CPUState *cs)
44 {
45     ARMCPU *cpu = ARM_CPU(cs);
46     struct kvm_vcpu_init init;
47 
48     init.target = cpu->kvm_target;
49     memcpy(init.features, cpu->kvm_init_features, sizeof(init.features));
50 
51     return kvm_vcpu_ioctl(cs, KVM_ARM_VCPU_INIT, &init);
52 }
53 
54 int kvm_arm_vcpu_finalize(CPUState *cs, int feature)
55 {
56     return kvm_vcpu_ioctl(cs, KVM_ARM_VCPU_FINALIZE, &feature);
57 }
58 
59 void kvm_arm_init_serror_injection(CPUState *cs)
60 {
61     cap_has_inject_serror_esr = kvm_check_extension(cs->kvm_state,
62                                     KVM_CAP_ARM_INJECT_SERROR_ESR);
63 }
64 
65 bool kvm_arm_create_scratch_host_vcpu(const uint32_t *cpus_to_try,
66                                       int *fdarray,
67                                       struct kvm_vcpu_init *init)
68 {
69     int ret = 0, kvmfd = -1, vmfd = -1, cpufd = -1;
70 
71     kvmfd = qemu_open("/dev/kvm", O_RDWR);
72     if (kvmfd < 0) {
73         goto err;
74     }
75     vmfd = ioctl(kvmfd, KVM_CREATE_VM, 0);
76     if (vmfd < 0) {
77         goto err;
78     }
79     cpufd = ioctl(vmfd, KVM_CREATE_VCPU, 0);
80     if (cpufd < 0) {
81         goto err;
82     }
83 
84     if (!init) {
85         /* Caller doesn't want the VCPU to be initialized, so skip it */
86         goto finish;
87     }
88 
89     if (init->target == -1) {
90         struct kvm_vcpu_init preferred;
91 
92         ret = ioctl(vmfd, KVM_ARM_PREFERRED_TARGET, &preferred);
93         if (!ret) {
94             init->target = preferred.target;
95         }
96     }
97     if (ret >= 0) {
98         ret = ioctl(cpufd, KVM_ARM_VCPU_INIT, init);
99         if (ret < 0) {
100             goto err;
101         }
102     } else if (cpus_to_try) {
103         /* Old kernel which doesn't know about the
104          * PREFERRED_TARGET ioctl: we know it will only support
105          * creating one kind of guest CPU which is its preferred
106          * CPU type.
107          */
108         struct kvm_vcpu_init try;
109 
110         while (*cpus_to_try != QEMU_KVM_ARM_TARGET_NONE) {
111             try.target = *cpus_to_try++;
112             memcpy(try.features, init->features, sizeof(init->features));
113             ret = ioctl(cpufd, KVM_ARM_VCPU_INIT, &try);
114             if (ret >= 0) {
115                 break;
116             }
117         }
118         if (ret < 0) {
119             goto err;
120         }
121         init->target = try.target;
122     } else {
123         /* Treat a NULL cpus_to_try argument the same as an empty
124          * list, which means we will fail the call since this must
125          * be an old kernel which doesn't support PREFERRED_TARGET.
126          */
127         goto err;
128     }
129 
130 finish:
131     fdarray[0] = kvmfd;
132     fdarray[1] = vmfd;
133     fdarray[2] = cpufd;
134 
135     return true;
136 
137 err:
138     if (cpufd >= 0) {
139         close(cpufd);
140     }
141     if (vmfd >= 0) {
142         close(vmfd);
143     }
144     if (kvmfd >= 0) {
145         close(kvmfd);
146     }
147 
148     return false;
149 }
150 
151 void kvm_arm_destroy_scratch_host_vcpu(int *fdarray)
152 {
153     int i;
154 
155     for (i = 2; i >= 0; i--) {
156         close(fdarray[i]);
157     }
158 }
159 
160 void kvm_arm_set_cpu_features_from_host(ARMCPU *cpu)
161 {
162     CPUARMState *env = &cpu->env;
163 
164     if (!arm_host_cpu_features.dtb_compatible) {
165         if (!kvm_enabled() ||
166             !kvm_arm_get_host_cpu_features(&arm_host_cpu_features)) {
167             /* We can't report this error yet, so flag that we need to
168              * in arm_cpu_realizefn().
169              */
170             cpu->kvm_target = QEMU_KVM_ARM_TARGET_NONE;
171             cpu->host_cpu_probe_failed = true;
172             return;
173         }
174     }
175 
176     cpu->kvm_target = arm_host_cpu_features.target;
177     cpu->dtb_compatible = arm_host_cpu_features.dtb_compatible;
178     cpu->isar = arm_host_cpu_features.isar;
179     env->features = arm_host_cpu_features.features;
180 }
181 
182 bool kvm_arm_pmu_supported(CPUState *cpu)
183 {
184     KVMState *s = KVM_STATE(current_machine->accelerator);
185 
186     return kvm_check_extension(s, KVM_CAP_ARM_PMU_V3);
187 }
188 
189 int kvm_arm_get_max_vm_ipa_size(MachineState *ms)
190 {
191     KVMState *s = KVM_STATE(ms->accelerator);
192     int ret;
193 
194     ret = kvm_check_extension(s, KVM_CAP_ARM_VM_IPA_SIZE);
195     return ret > 0 ? ret : 40;
196 }
197 
198 int kvm_arch_init(MachineState *ms, KVMState *s)
199 {
200     int ret = 0;
201     /* For ARM interrupt delivery is always asynchronous,
202      * whether we are using an in-kernel VGIC or not.
203      */
204     kvm_async_interrupts_allowed = true;
205 
206     /*
207      * PSCI wakes up secondary cores, so we always need to
208      * have vCPUs waiting in kernel space
209      */
210     kvm_halt_in_kernel_allowed = true;
211 
212     cap_has_mp_state = kvm_check_extension(s, KVM_CAP_MP_STATE);
213 
214     if (ms->smp.cpus > 256 &&
215         !kvm_check_extension(s, KVM_CAP_ARM_IRQ_LINE_LAYOUT_2)) {
216         error_report("Using more than 256 vcpus requires a host kernel "
217                      "with KVM_CAP_ARM_IRQ_LINE_LAYOUT_2");
218         ret = -EINVAL;
219     }
220 
221     return ret;
222 }
223 
224 unsigned long kvm_arch_vcpu_id(CPUState *cpu)
225 {
226     return cpu->cpu_index;
227 }
228 
229 /* We track all the KVM devices which need their memory addresses
230  * passing to the kernel in a list of these structures.
231  * When board init is complete we run through the list and
232  * tell the kernel the base addresses of the memory regions.
233  * We use a MemoryListener to track mapping and unmapping of
234  * the regions during board creation, so the board models don't
235  * need to do anything special for the KVM case.
236  *
237  * Sometimes the address must be OR'ed with some other fields
238  * (for example for KVM_VGIC_V3_ADDR_TYPE_REDIST_REGION).
239  * @kda_addr_ormask aims at storing the value of those fields.
240  */
241 typedef struct KVMDevice {
242     struct kvm_arm_device_addr kda;
243     struct kvm_device_attr kdattr;
244     uint64_t kda_addr_ormask;
245     MemoryRegion *mr;
246     QSLIST_ENTRY(KVMDevice) entries;
247     int dev_fd;
248 } KVMDevice;
249 
250 static QSLIST_HEAD(, KVMDevice) kvm_devices_head;
251 
252 static void kvm_arm_devlistener_add(MemoryListener *listener,
253                                     MemoryRegionSection *section)
254 {
255     KVMDevice *kd;
256 
257     QSLIST_FOREACH(kd, &kvm_devices_head, entries) {
258         if (section->mr == kd->mr) {
259             kd->kda.addr = section->offset_within_address_space;
260         }
261     }
262 }
263 
264 static void kvm_arm_devlistener_del(MemoryListener *listener,
265                                     MemoryRegionSection *section)
266 {
267     KVMDevice *kd;
268 
269     QSLIST_FOREACH(kd, &kvm_devices_head, entries) {
270         if (section->mr == kd->mr) {
271             kd->kda.addr = -1;
272         }
273     }
274 }
275 
276 static MemoryListener devlistener = {
277     .region_add = kvm_arm_devlistener_add,
278     .region_del = kvm_arm_devlistener_del,
279 };
280 
281 static void kvm_arm_set_device_addr(KVMDevice *kd)
282 {
283     struct kvm_device_attr *attr = &kd->kdattr;
284     int ret;
285 
286     /* If the device control API is available and we have a device fd on the
287      * KVMDevice struct, let's use the newer API
288      */
289     if (kd->dev_fd >= 0) {
290         uint64_t addr = kd->kda.addr;
291 
292         addr |= kd->kda_addr_ormask;
293         attr->addr = (uintptr_t)&addr;
294         ret = kvm_device_ioctl(kd->dev_fd, KVM_SET_DEVICE_ATTR, attr);
295     } else {
296         ret = kvm_vm_ioctl(kvm_state, KVM_ARM_SET_DEVICE_ADDR, &kd->kda);
297     }
298 
299     if (ret < 0) {
300         fprintf(stderr, "Failed to set device address: %s\n",
301                 strerror(-ret));
302         abort();
303     }
304 }
305 
306 static void kvm_arm_machine_init_done(Notifier *notifier, void *data)
307 {
308     KVMDevice *kd, *tkd;
309 
310     QSLIST_FOREACH_SAFE(kd, &kvm_devices_head, entries, tkd) {
311         if (kd->kda.addr != -1) {
312             kvm_arm_set_device_addr(kd);
313         }
314         memory_region_unref(kd->mr);
315         QSLIST_REMOVE_HEAD(&kvm_devices_head, entries);
316         g_free(kd);
317     }
318     memory_listener_unregister(&devlistener);
319 }
320 
321 static Notifier notify = {
322     .notify = kvm_arm_machine_init_done,
323 };
324 
325 void kvm_arm_register_device(MemoryRegion *mr, uint64_t devid, uint64_t group,
326                              uint64_t attr, int dev_fd, uint64_t addr_ormask)
327 {
328     KVMDevice *kd;
329 
330     if (!kvm_irqchip_in_kernel()) {
331         return;
332     }
333 
334     if (QSLIST_EMPTY(&kvm_devices_head)) {
335         memory_listener_register(&devlistener, &address_space_memory);
336         qemu_add_machine_init_done_notifier(&notify);
337     }
338     kd = g_new0(KVMDevice, 1);
339     kd->mr = mr;
340     kd->kda.id = devid;
341     kd->kda.addr = -1;
342     kd->kdattr.flags = 0;
343     kd->kdattr.group = group;
344     kd->kdattr.attr = attr;
345     kd->dev_fd = dev_fd;
346     kd->kda_addr_ormask = addr_ormask;
347     QSLIST_INSERT_HEAD(&kvm_devices_head, kd, entries);
348     memory_region_ref(kd->mr);
349 }
350 
351 static int compare_u64(const void *a, const void *b)
352 {
353     if (*(uint64_t *)a > *(uint64_t *)b) {
354         return 1;
355     }
356     if (*(uint64_t *)a < *(uint64_t *)b) {
357         return -1;
358     }
359     return 0;
360 }
361 
362 /* Initialize the ARMCPU cpreg list according to the kernel's
363  * definition of what CPU registers it knows about (and throw away
364  * the previous TCG-created cpreg list).
365  */
366 int kvm_arm_init_cpreg_list(ARMCPU *cpu)
367 {
368     struct kvm_reg_list rl;
369     struct kvm_reg_list *rlp;
370     int i, ret, arraylen;
371     CPUState *cs = CPU(cpu);
372 
373     rl.n = 0;
374     ret = kvm_vcpu_ioctl(cs, KVM_GET_REG_LIST, &rl);
375     if (ret != -E2BIG) {
376         return ret;
377     }
378     rlp = g_malloc(sizeof(struct kvm_reg_list) + rl.n * sizeof(uint64_t));
379     rlp->n = rl.n;
380     ret = kvm_vcpu_ioctl(cs, KVM_GET_REG_LIST, rlp);
381     if (ret) {
382         goto out;
383     }
384     /* Sort the list we get back from the kernel, since cpreg_tuples
385      * must be in strictly ascending order.
386      */
387     qsort(&rlp->reg, rlp->n, sizeof(rlp->reg[0]), compare_u64);
388 
389     for (i = 0, arraylen = 0; i < rlp->n; i++) {
390         if (!kvm_arm_reg_syncs_via_cpreg_list(rlp->reg[i])) {
391             continue;
392         }
393         switch (rlp->reg[i] & KVM_REG_SIZE_MASK) {
394         case KVM_REG_SIZE_U32:
395         case KVM_REG_SIZE_U64:
396             break;
397         default:
398             fprintf(stderr, "Can't handle size of register in kernel list\n");
399             ret = -EINVAL;
400             goto out;
401         }
402 
403         arraylen++;
404     }
405 
406     cpu->cpreg_indexes = g_renew(uint64_t, cpu->cpreg_indexes, arraylen);
407     cpu->cpreg_values = g_renew(uint64_t, cpu->cpreg_values, arraylen);
408     cpu->cpreg_vmstate_indexes = g_renew(uint64_t, cpu->cpreg_vmstate_indexes,
409                                          arraylen);
410     cpu->cpreg_vmstate_values = g_renew(uint64_t, cpu->cpreg_vmstate_values,
411                                         arraylen);
412     cpu->cpreg_array_len = arraylen;
413     cpu->cpreg_vmstate_array_len = arraylen;
414 
415     for (i = 0, arraylen = 0; i < rlp->n; i++) {
416         uint64_t regidx = rlp->reg[i];
417         if (!kvm_arm_reg_syncs_via_cpreg_list(regidx)) {
418             continue;
419         }
420         cpu->cpreg_indexes[arraylen] = regidx;
421         arraylen++;
422     }
423     assert(cpu->cpreg_array_len == arraylen);
424 
425     if (!write_kvmstate_to_list(cpu)) {
426         /* Shouldn't happen unless kernel is inconsistent about
427          * what registers exist.
428          */
429         fprintf(stderr, "Initial read of kernel register state failed\n");
430         ret = -EINVAL;
431         goto out;
432     }
433 
434 out:
435     g_free(rlp);
436     return ret;
437 }
438 
439 bool write_kvmstate_to_list(ARMCPU *cpu)
440 {
441     CPUState *cs = CPU(cpu);
442     int i;
443     bool ok = true;
444 
445     for (i = 0; i < cpu->cpreg_array_len; i++) {
446         struct kvm_one_reg r;
447         uint64_t regidx = cpu->cpreg_indexes[i];
448         uint32_t v32;
449         int ret;
450 
451         r.id = regidx;
452 
453         switch (regidx & KVM_REG_SIZE_MASK) {
454         case KVM_REG_SIZE_U32:
455             r.addr = (uintptr_t)&v32;
456             ret = kvm_vcpu_ioctl(cs, KVM_GET_ONE_REG, &r);
457             if (!ret) {
458                 cpu->cpreg_values[i] = v32;
459             }
460             break;
461         case KVM_REG_SIZE_U64:
462             r.addr = (uintptr_t)(cpu->cpreg_values + i);
463             ret = kvm_vcpu_ioctl(cs, KVM_GET_ONE_REG, &r);
464             break;
465         default:
466             abort();
467         }
468         if (ret) {
469             ok = false;
470         }
471     }
472     return ok;
473 }
474 
475 bool write_list_to_kvmstate(ARMCPU *cpu, int level)
476 {
477     CPUState *cs = CPU(cpu);
478     int i;
479     bool ok = true;
480 
481     for (i = 0; i < cpu->cpreg_array_len; i++) {
482         struct kvm_one_reg r;
483         uint64_t regidx = cpu->cpreg_indexes[i];
484         uint32_t v32;
485         int ret;
486 
487         if (kvm_arm_cpreg_level(regidx) > level) {
488             continue;
489         }
490 
491         r.id = regidx;
492         switch (regidx & KVM_REG_SIZE_MASK) {
493         case KVM_REG_SIZE_U32:
494             v32 = cpu->cpreg_values[i];
495             r.addr = (uintptr_t)&v32;
496             break;
497         case KVM_REG_SIZE_U64:
498             r.addr = (uintptr_t)(cpu->cpreg_values + i);
499             break;
500         default:
501             abort();
502         }
503         ret = kvm_vcpu_ioctl(cs, KVM_SET_ONE_REG, &r);
504         if (ret) {
505             /* We might fail for "unknown register" and also for
506              * "you tried to set a register which is constant with
507              * a different value from what it actually contains".
508              */
509             ok = false;
510         }
511     }
512     return ok;
513 }
514 
515 void kvm_arm_reset_vcpu(ARMCPU *cpu)
516 {
517     int ret;
518 
519     /* Re-init VCPU so that all registers are set to
520      * their respective reset values.
521      */
522     ret = kvm_arm_vcpu_init(CPU(cpu));
523     if (ret < 0) {
524         fprintf(stderr, "kvm_arm_vcpu_init failed: %s\n", strerror(-ret));
525         abort();
526     }
527     if (!write_kvmstate_to_list(cpu)) {
528         fprintf(stderr, "write_kvmstate_to_list failed\n");
529         abort();
530     }
531     /*
532      * Sync the reset values also into the CPUState. This is necessary
533      * because the next thing we do will be a kvm_arch_put_registers()
534      * which will update the list values from the CPUState before copying
535      * the list values back to KVM. It's OK to ignore failure returns here
536      * for the same reason we do so in kvm_arch_get_registers().
537      */
538     write_list_to_cpustate(cpu);
539 }
540 
541 /*
542  * Update KVM's MP_STATE based on what QEMU thinks it is
543  */
544 int kvm_arm_sync_mpstate_to_kvm(ARMCPU *cpu)
545 {
546     if (cap_has_mp_state) {
547         struct kvm_mp_state mp_state = {
548             .mp_state = (cpu->power_state == PSCI_OFF) ?
549             KVM_MP_STATE_STOPPED : KVM_MP_STATE_RUNNABLE
550         };
551         int ret = kvm_vcpu_ioctl(CPU(cpu), KVM_SET_MP_STATE, &mp_state);
552         if (ret) {
553             fprintf(stderr, "%s: failed to set MP_STATE %d/%s\n",
554                     __func__, ret, strerror(-ret));
555             return -1;
556         }
557     }
558 
559     return 0;
560 }
561 
562 /*
563  * Sync the KVM MP_STATE into QEMU
564  */
565 int kvm_arm_sync_mpstate_to_qemu(ARMCPU *cpu)
566 {
567     if (cap_has_mp_state) {
568         struct kvm_mp_state mp_state;
569         int ret = kvm_vcpu_ioctl(CPU(cpu), KVM_GET_MP_STATE, &mp_state);
570         if (ret) {
571             fprintf(stderr, "%s: failed to get MP_STATE %d/%s\n",
572                     __func__, ret, strerror(-ret));
573             abort();
574         }
575         cpu->power_state = (mp_state.mp_state == KVM_MP_STATE_STOPPED) ?
576             PSCI_OFF : PSCI_ON;
577     }
578 
579     return 0;
580 }
581 
582 int kvm_put_vcpu_events(ARMCPU *cpu)
583 {
584     CPUARMState *env = &cpu->env;
585     struct kvm_vcpu_events events;
586     int ret;
587 
588     if (!kvm_has_vcpu_events()) {
589         return 0;
590     }
591 
592     memset(&events, 0, sizeof(events));
593     events.exception.serror_pending = env->serror.pending;
594 
595     /* Inject SError to guest with specified syndrome if host kernel
596      * supports it, otherwise inject SError without syndrome.
597      */
598     if (cap_has_inject_serror_esr) {
599         events.exception.serror_has_esr = env->serror.has_esr;
600         events.exception.serror_esr = env->serror.esr;
601     }
602 
603     ret = kvm_vcpu_ioctl(CPU(cpu), KVM_SET_VCPU_EVENTS, &events);
604     if (ret) {
605         error_report("failed to put vcpu events");
606     }
607 
608     return ret;
609 }
610 
611 int kvm_get_vcpu_events(ARMCPU *cpu)
612 {
613     CPUARMState *env = &cpu->env;
614     struct kvm_vcpu_events events;
615     int ret;
616 
617     if (!kvm_has_vcpu_events()) {
618         return 0;
619     }
620 
621     memset(&events, 0, sizeof(events));
622     ret = kvm_vcpu_ioctl(CPU(cpu), KVM_GET_VCPU_EVENTS, &events);
623     if (ret) {
624         error_report("failed to get vcpu events");
625         return ret;
626     }
627 
628     env->serror.pending = events.exception.serror_pending;
629     env->serror.has_esr = events.exception.serror_has_esr;
630     env->serror.esr = events.exception.serror_esr;
631 
632     return 0;
633 }
634 
635 void kvm_arch_pre_run(CPUState *cs, struct kvm_run *run)
636 {
637 }
638 
639 MemTxAttrs kvm_arch_post_run(CPUState *cs, struct kvm_run *run)
640 {
641     ARMCPU *cpu;
642     uint32_t switched_level;
643 
644     if (kvm_irqchip_in_kernel()) {
645         /*
646          * We only need to sync timer states with user-space interrupt
647          * controllers, so return early and save cycles if we don't.
648          */
649         return MEMTXATTRS_UNSPECIFIED;
650     }
651 
652     cpu = ARM_CPU(cs);
653 
654     /* Synchronize our shadowed in-kernel device irq lines with the kvm ones */
655     if (run->s.regs.device_irq_level != cpu->device_irq_level) {
656         switched_level = cpu->device_irq_level ^ run->s.regs.device_irq_level;
657 
658         qemu_mutex_lock_iothread();
659 
660         if (switched_level & KVM_ARM_DEV_EL1_VTIMER) {
661             qemu_set_irq(cpu->gt_timer_outputs[GTIMER_VIRT],
662                          !!(run->s.regs.device_irq_level &
663                             KVM_ARM_DEV_EL1_VTIMER));
664             switched_level &= ~KVM_ARM_DEV_EL1_VTIMER;
665         }
666 
667         if (switched_level & KVM_ARM_DEV_EL1_PTIMER) {
668             qemu_set_irq(cpu->gt_timer_outputs[GTIMER_PHYS],
669                          !!(run->s.regs.device_irq_level &
670                             KVM_ARM_DEV_EL1_PTIMER));
671             switched_level &= ~KVM_ARM_DEV_EL1_PTIMER;
672         }
673 
674         if (switched_level & KVM_ARM_DEV_PMU) {
675             qemu_set_irq(cpu->pmu_interrupt,
676                          !!(run->s.regs.device_irq_level & KVM_ARM_DEV_PMU));
677             switched_level &= ~KVM_ARM_DEV_PMU;
678         }
679 
680         if (switched_level) {
681             qemu_log_mask(LOG_UNIMP, "%s: unhandled in-kernel device IRQ %x\n",
682                           __func__, switched_level);
683         }
684 
685         /* We also mark unknown levels as processed to not waste cycles */
686         cpu->device_irq_level = run->s.regs.device_irq_level;
687         qemu_mutex_unlock_iothread();
688     }
689 
690     return MEMTXATTRS_UNSPECIFIED;
691 }
692 
693 
694 int kvm_arch_handle_exit(CPUState *cs, struct kvm_run *run)
695 {
696     int ret = 0;
697 
698     switch (run->exit_reason) {
699     case KVM_EXIT_DEBUG:
700         if (kvm_arm_handle_debug(cs, &run->debug.arch)) {
701             ret = EXCP_DEBUG;
702         } /* otherwise return to guest */
703         break;
704     default:
705         qemu_log_mask(LOG_UNIMP, "%s: un-handled exit reason %d\n",
706                       __func__, run->exit_reason);
707         break;
708     }
709     return ret;
710 }
711 
712 bool kvm_arch_stop_on_emulation_error(CPUState *cs)
713 {
714     return true;
715 }
716 
717 int kvm_arch_process_async_events(CPUState *cs)
718 {
719     return 0;
720 }
721 
722 /* The #ifdef protections are until 32bit headers are imported and can
723  * be removed once both 32 and 64 bit reach feature parity.
724  */
725 void kvm_arch_update_guest_debug(CPUState *cs, struct kvm_guest_debug *dbg)
726 {
727 #ifdef KVM_GUESTDBG_USE_SW_BP
728     if (kvm_sw_breakpoints_active(cs)) {
729         dbg->control |= KVM_GUESTDBG_ENABLE | KVM_GUESTDBG_USE_SW_BP;
730     }
731 #endif
732 #ifdef KVM_GUESTDBG_USE_HW
733     if (kvm_arm_hw_debug_active(cs)) {
734         dbg->control |= KVM_GUESTDBG_ENABLE | KVM_GUESTDBG_USE_HW;
735         kvm_arm_copy_hw_debug_data(&dbg->arch);
736     }
737 #endif
738 }
739 
740 void kvm_arch_init_irq_routing(KVMState *s)
741 {
742 }
743 
744 int kvm_arch_irqchip_create(MachineState *ms, KVMState *s)
745 {
746      if (machine_kernel_irqchip_split(ms)) {
747          perror("-machine kernel_irqchip=split is not supported on ARM.");
748          exit(1);
749     }
750 
751     /* If we can create the VGIC using the newer device control API, we
752      * let the device do this when it initializes itself, otherwise we
753      * fall back to the old API */
754     return kvm_check_extension(s, KVM_CAP_DEVICE_CTRL);
755 }
756 
757 int kvm_arm_vgic_probe(void)
758 {
759     if (kvm_create_device(kvm_state,
760                           KVM_DEV_TYPE_ARM_VGIC_V3, true) == 0) {
761         return 3;
762     } else if (kvm_create_device(kvm_state,
763                                  KVM_DEV_TYPE_ARM_VGIC_V2, true) == 0) {
764         return 2;
765     } else {
766         return 0;
767     }
768 }
769 
770 int kvm_arm_set_irq(int cpu, int irqtype, int irq, int level)
771 {
772     int kvm_irq = (irqtype << KVM_ARM_IRQ_TYPE_SHIFT) | irq;
773     int cpu_idx1 = cpu % 256;
774     int cpu_idx2 = cpu / 256;
775 
776     kvm_irq |= (cpu_idx1 << KVM_ARM_IRQ_VCPU_SHIFT) |
777                (cpu_idx2 << KVM_ARM_IRQ_VCPU2_SHIFT);
778 
779     return kvm_set_irq(kvm_state, kvm_irq, !!level);
780 }
781 
782 int kvm_arch_fixup_msi_route(struct kvm_irq_routing_entry *route,
783                              uint64_t address, uint32_t data, PCIDevice *dev)
784 {
785     AddressSpace *as = pci_device_iommu_address_space(dev);
786     hwaddr xlat, len, doorbell_gpa;
787     MemoryRegionSection mrs;
788     MemoryRegion *mr;
789     int ret = 1;
790 
791     if (as == &address_space_memory) {
792         return 0;
793     }
794 
795     /* MSI doorbell address is translated by an IOMMU */
796 
797     rcu_read_lock();
798     mr = address_space_translate(as, address, &xlat, &len, true,
799                                  MEMTXATTRS_UNSPECIFIED);
800     if (!mr) {
801         goto unlock;
802     }
803     mrs = memory_region_find(mr, xlat, 1);
804     if (!mrs.mr) {
805         goto unlock;
806     }
807 
808     doorbell_gpa = mrs.offset_within_address_space;
809     memory_region_unref(mrs.mr);
810 
811     route->u.msi.address_lo = doorbell_gpa;
812     route->u.msi.address_hi = doorbell_gpa >> 32;
813 
814     trace_kvm_arm_fixup_msi_route(address, doorbell_gpa);
815 
816     ret = 0;
817 
818 unlock:
819     rcu_read_unlock();
820     return ret;
821 }
822 
823 int kvm_arch_add_msi_route_post(struct kvm_irq_routing_entry *route,
824                                 int vector, PCIDevice *dev)
825 {
826     return 0;
827 }
828 
829 int kvm_arch_release_virq_post(int virq)
830 {
831     return 0;
832 }
833 
834 int kvm_arch_msi_data_to_gsi(uint32_t data)
835 {
836     return (data - 32) & 0xffff;
837 }
838