169e0a03cSPaolo Bonzini /* Copyright 2008 IBM Corporation 269e0a03cSPaolo Bonzini * 2008 Red Hat, Inc. 369e0a03cSPaolo Bonzini * Copyright 2011 Intel Corporation 469e0a03cSPaolo Bonzini * Copyright 2016 Veertu, Inc. 569e0a03cSPaolo Bonzini * Copyright 2017 The Android Open Source Project 669e0a03cSPaolo Bonzini * 769e0a03cSPaolo Bonzini * QEMU Hypervisor.framework support 869e0a03cSPaolo Bonzini * 969e0a03cSPaolo Bonzini * This program is free software; you can redistribute it and/or 1069e0a03cSPaolo Bonzini * modify it under the terms of version 2 of the GNU General Public 1169e0a03cSPaolo Bonzini * License as published by the Free Software Foundation. 1269e0a03cSPaolo Bonzini * 1369e0a03cSPaolo Bonzini * This program is distributed in the hope that it will be useful, 1469e0a03cSPaolo Bonzini * but WITHOUT ANY WARRANTY; without even the implied warranty of 1569e0a03cSPaolo Bonzini * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU 16e361a772SThomas Huth * General Public License for more details. 1769e0a03cSPaolo Bonzini * 18e361a772SThomas Huth * You should have received a copy of the GNU General Public License 19e361a772SThomas Huth * along with this program; if not, see <http://www.gnu.org/licenses/>. 20d781e24dSIzik Eidus * 21d781e24dSIzik Eidus * This file contain code under public domain from the hvdos project: 22d781e24dSIzik Eidus * https://github.com/mist64/hvdos 234d98a8e5SPaolo Bonzini * 244d98a8e5SPaolo Bonzini * Parts Copyright (c) 2011 NetApp, Inc. 254d98a8e5SPaolo Bonzini * All rights reserved. 264d98a8e5SPaolo Bonzini * 274d98a8e5SPaolo Bonzini * Redistribution and use in source and binary forms, with or without 284d98a8e5SPaolo Bonzini * modification, are permitted provided that the following conditions 294d98a8e5SPaolo Bonzini * are met: 304d98a8e5SPaolo Bonzini * 1. Redistributions of source code must retain the above copyright 314d98a8e5SPaolo Bonzini * notice, this list of conditions and the following disclaimer. 324d98a8e5SPaolo Bonzini * 2. Redistributions in binary form must reproduce the above copyright 334d98a8e5SPaolo Bonzini * notice, this list of conditions and the following disclaimer in the 344d98a8e5SPaolo Bonzini * documentation and/or other materials provided with the distribution. 354d98a8e5SPaolo Bonzini * 364d98a8e5SPaolo Bonzini * THIS SOFTWARE IS PROVIDED BY NETAPP, INC ``AS IS'' AND 374d98a8e5SPaolo Bonzini * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE 384d98a8e5SPaolo Bonzini * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE 394d98a8e5SPaolo Bonzini * ARE DISCLAIMED. IN NO EVENT SHALL NETAPP, INC OR CONTRIBUTORS BE LIABLE 404d98a8e5SPaolo Bonzini * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL 414d98a8e5SPaolo Bonzini * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS 424d98a8e5SPaolo Bonzini * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) 434d98a8e5SPaolo Bonzini * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT 444d98a8e5SPaolo Bonzini * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY 454d98a8e5SPaolo Bonzini * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF 464d98a8e5SPaolo Bonzini * SUCH DAMAGE. 4769e0a03cSPaolo Bonzini */ 4854d31236SMarkus Armbruster 4969e0a03cSPaolo Bonzini #include "qemu/osdep.h" 5069e0a03cSPaolo Bonzini #include "qemu-common.h" 5169e0a03cSPaolo Bonzini #include "qemu/error-report.h" 5269e0a03cSPaolo Bonzini 5369e0a03cSPaolo Bonzini #include "sysemu/hvf.h" 5454d31236SMarkus Armbruster #include "sysemu/runstate.h" 5569e0a03cSPaolo Bonzini #include "hvf-i386.h" 5669e0a03cSPaolo Bonzini #include "vmcs.h" 5769e0a03cSPaolo Bonzini #include "vmx.h" 5869e0a03cSPaolo Bonzini #include "x86.h" 5969e0a03cSPaolo Bonzini #include "x86_descr.h" 6069e0a03cSPaolo Bonzini #include "x86_mmu.h" 6169e0a03cSPaolo Bonzini #include "x86_decode.h" 6269e0a03cSPaolo Bonzini #include "x86_emu.h" 6369e0a03cSPaolo Bonzini #include "x86_task.h" 6469e0a03cSPaolo Bonzini #include "x86hvf.h" 6569e0a03cSPaolo Bonzini 6669e0a03cSPaolo Bonzini #include <Hypervisor/hv.h> 6769e0a03cSPaolo Bonzini #include <Hypervisor/hv_vmx.h> 6869e0a03cSPaolo Bonzini 6969e0a03cSPaolo Bonzini #include "exec/address-spaces.h" 7069e0a03cSPaolo Bonzini #include "hw/i386/apic_internal.h" 7169e0a03cSPaolo Bonzini #include "qemu/main-loop.h" 7269e0a03cSPaolo Bonzini #include "sysemu/accel.h" 7369e0a03cSPaolo Bonzini #include "target/i386/cpu.h" 7469e0a03cSPaolo Bonzini 75*b52bcba7SClaudio Fontana #include "hvf-cpus.h" 76*b52bcba7SClaudio Fontana 7769e0a03cSPaolo Bonzini HVFState *hvf_state; 7869e0a03cSPaolo Bonzini 7969e0a03cSPaolo Bonzini static void assert_hvf_ok(hv_return_t ret) 8069e0a03cSPaolo Bonzini { 8169e0a03cSPaolo Bonzini if (ret == HV_SUCCESS) { 8269e0a03cSPaolo Bonzini return; 8369e0a03cSPaolo Bonzini } 8469e0a03cSPaolo Bonzini 8569e0a03cSPaolo Bonzini switch (ret) { 8669e0a03cSPaolo Bonzini case HV_ERROR: 872d9178d9SLaurent Vivier error_report("Error: HV_ERROR"); 8869e0a03cSPaolo Bonzini break; 8969e0a03cSPaolo Bonzini case HV_BUSY: 902d9178d9SLaurent Vivier error_report("Error: HV_BUSY"); 9169e0a03cSPaolo Bonzini break; 9269e0a03cSPaolo Bonzini case HV_BAD_ARGUMENT: 932d9178d9SLaurent Vivier error_report("Error: HV_BAD_ARGUMENT"); 9469e0a03cSPaolo Bonzini break; 9569e0a03cSPaolo Bonzini case HV_NO_RESOURCES: 962d9178d9SLaurent Vivier error_report("Error: HV_NO_RESOURCES"); 9769e0a03cSPaolo Bonzini break; 9869e0a03cSPaolo Bonzini case HV_NO_DEVICE: 992d9178d9SLaurent Vivier error_report("Error: HV_NO_DEVICE"); 10069e0a03cSPaolo Bonzini break; 10169e0a03cSPaolo Bonzini case HV_UNSUPPORTED: 1022d9178d9SLaurent Vivier error_report("Error: HV_UNSUPPORTED"); 10369e0a03cSPaolo Bonzini break; 10469e0a03cSPaolo Bonzini default: 1052d9178d9SLaurent Vivier error_report("Unknown Error"); 10669e0a03cSPaolo Bonzini } 10769e0a03cSPaolo Bonzini 10869e0a03cSPaolo Bonzini abort(); 10969e0a03cSPaolo Bonzini } 11069e0a03cSPaolo Bonzini 11169e0a03cSPaolo Bonzini /* Memory slots */ 112fbafbb6dSCameron Esfahani hvf_slot *hvf_find_overlap_slot(uint64_t start, uint64_t size) 11369e0a03cSPaolo Bonzini { 11469e0a03cSPaolo Bonzini hvf_slot *slot; 11569e0a03cSPaolo Bonzini int x; 11669e0a03cSPaolo Bonzini for (x = 0; x < hvf_state->num_slots; ++x) { 11769e0a03cSPaolo Bonzini slot = &hvf_state->slots[x]; 11869e0a03cSPaolo Bonzini if (slot->size && start < (slot->start + slot->size) && 119fbafbb6dSCameron Esfahani (start + size) > slot->start) { 12069e0a03cSPaolo Bonzini return slot; 12169e0a03cSPaolo Bonzini } 12269e0a03cSPaolo Bonzini } 12369e0a03cSPaolo Bonzini return NULL; 12469e0a03cSPaolo Bonzini } 12569e0a03cSPaolo Bonzini 12669e0a03cSPaolo Bonzini struct mac_slot { 12769e0a03cSPaolo Bonzini int present; 12869e0a03cSPaolo Bonzini uint64_t size; 12969e0a03cSPaolo Bonzini uint64_t gpa_start; 13069e0a03cSPaolo Bonzini uint64_t gva; 13169e0a03cSPaolo Bonzini }; 13269e0a03cSPaolo Bonzini 13369e0a03cSPaolo Bonzini struct mac_slot mac_slots[32]; 13469e0a03cSPaolo Bonzini 135fbafbb6dSCameron Esfahani static int do_hvf_set_memory(hvf_slot *slot, hv_memory_flags_t flags) 13669e0a03cSPaolo Bonzini { 13769e0a03cSPaolo Bonzini struct mac_slot *macslot; 13869e0a03cSPaolo Bonzini hv_return_t ret; 13969e0a03cSPaolo Bonzini 14069e0a03cSPaolo Bonzini macslot = &mac_slots[slot->slot_id]; 14169e0a03cSPaolo Bonzini 14269e0a03cSPaolo Bonzini if (macslot->present) { 14369e0a03cSPaolo Bonzini if (macslot->size != slot->size) { 14469e0a03cSPaolo Bonzini macslot->present = 0; 14569e0a03cSPaolo Bonzini ret = hv_vm_unmap(macslot->gpa_start, macslot->size); 14669e0a03cSPaolo Bonzini assert_hvf_ok(ret); 14769e0a03cSPaolo Bonzini } 14869e0a03cSPaolo Bonzini } 14969e0a03cSPaolo Bonzini 15069e0a03cSPaolo Bonzini if (!slot->size) { 15169e0a03cSPaolo Bonzini return 0; 15269e0a03cSPaolo Bonzini } 15369e0a03cSPaolo Bonzini 15469e0a03cSPaolo Bonzini macslot->present = 1; 15569e0a03cSPaolo Bonzini macslot->gpa_start = slot->start; 15669e0a03cSPaolo Bonzini macslot->size = slot->size; 15769e0a03cSPaolo Bonzini ret = hv_vm_map((hv_uvaddr_t)slot->mem, slot->start, slot->size, flags); 15869e0a03cSPaolo Bonzini assert_hvf_ok(ret); 15969e0a03cSPaolo Bonzini return 0; 16069e0a03cSPaolo Bonzini } 16169e0a03cSPaolo Bonzini 16269e0a03cSPaolo Bonzini void hvf_set_phys_mem(MemoryRegionSection *section, bool add) 16369e0a03cSPaolo Bonzini { 16469e0a03cSPaolo Bonzini hvf_slot *mem; 16569e0a03cSPaolo Bonzini MemoryRegion *area = section->mr; 166fbafbb6dSCameron Esfahani bool writeable = !area->readonly && !area->rom_device; 167fbafbb6dSCameron Esfahani hv_memory_flags_t flags; 16869e0a03cSPaolo Bonzini 16969e0a03cSPaolo Bonzini if (!memory_region_is_ram(area)) { 170fbafbb6dSCameron Esfahani if (writeable) { 17169e0a03cSPaolo Bonzini return; 172fbafbb6dSCameron Esfahani } else if (!memory_region_is_romd(area)) { 173fbafbb6dSCameron Esfahani /* 174fbafbb6dSCameron Esfahani * If the memory device is not in romd_mode, then we actually want 175fbafbb6dSCameron Esfahani * to remove the hvf memory slot so all accesses will trap. 176fbafbb6dSCameron Esfahani */ 177fbafbb6dSCameron Esfahani add = false; 178fbafbb6dSCameron Esfahani } 17969e0a03cSPaolo Bonzini } 18069e0a03cSPaolo Bonzini 18169e0a03cSPaolo Bonzini mem = hvf_find_overlap_slot( 18269e0a03cSPaolo Bonzini section->offset_within_address_space, 183fbafbb6dSCameron Esfahani int128_get64(section->size)); 18469e0a03cSPaolo Bonzini 18569e0a03cSPaolo Bonzini if (mem && add) { 18669e0a03cSPaolo Bonzini if (mem->size == int128_get64(section->size) && 18769e0a03cSPaolo Bonzini mem->start == section->offset_within_address_space && 18869e0a03cSPaolo Bonzini mem->mem == (memory_region_get_ram_ptr(area) + 18969e0a03cSPaolo Bonzini section->offset_within_region)) { 19069e0a03cSPaolo Bonzini return; /* Same region was attempted to register, go away. */ 19169e0a03cSPaolo Bonzini } 19269e0a03cSPaolo Bonzini } 19369e0a03cSPaolo Bonzini 19469e0a03cSPaolo Bonzini /* Region needs to be reset. set the size to 0 and remap it. */ 19569e0a03cSPaolo Bonzini if (mem) { 19669e0a03cSPaolo Bonzini mem->size = 0; 197fbafbb6dSCameron Esfahani if (do_hvf_set_memory(mem, 0)) { 1982d9178d9SLaurent Vivier error_report("Failed to reset overlapping slot"); 19969e0a03cSPaolo Bonzini abort(); 20069e0a03cSPaolo Bonzini } 20169e0a03cSPaolo Bonzini } 20269e0a03cSPaolo Bonzini 20369e0a03cSPaolo Bonzini if (!add) { 20469e0a03cSPaolo Bonzini return; 20569e0a03cSPaolo Bonzini } 20669e0a03cSPaolo Bonzini 207fbafbb6dSCameron Esfahani if (area->readonly || 208fbafbb6dSCameron Esfahani (!memory_region_is_ram(area) && memory_region_is_romd(area))) { 209fbafbb6dSCameron Esfahani flags = HV_MEMORY_READ | HV_MEMORY_EXEC; 210fbafbb6dSCameron Esfahani } else { 211fbafbb6dSCameron Esfahani flags = HV_MEMORY_READ | HV_MEMORY_WRITE | HV_MEMORY_EXEC; 212fbafbb6dSCameron Esfahani } 213fbafbb6dSCameron Esfahani 21469e0a03cSPaolo Bonzini /* Now make a new slot. */ 21569e0a03cSPaolo Bonzini int x; 21669e0a03cSPaolo Bonzini 21769e0a03cSPaolo Bonzini for (x = 0; x < hvf_state->num_slots; ++x) { 21869e0a03cSPaolo Bonzini mem = &hvf_state->slots[x]; 21969e0a03cSPaolo Bonzini if (!mem->size) { 22069e0a03cSPaolo Bonzini break; 22169e0a03cSPaolo Bonzini } 22269e0a03cSPaolo Bonzini } 22369e0a03cSPaolo Bonzini 22469e0a03cSPaolo Bonzini if (x == hvf_state->num_slots) { 2252d9178d9SLaurent Vivier error_report("No free slots"); 22669e0a03cSPaolo Bonzini abort(); 22769e0a03cSPaolo Bonzini } 22869e0a03cSPaolo Bonzini 22969e0a03cSPaolo Bonzini mem->size = int128_get64(section->size); 23069e0a03cSPaolo Bonzini mem->mem = memory_region_get_ram_ptr(area) + section->offset_within_region; 23169e0a03cSPaolo Bonzini mem->start = section->offset_within_address_space; 23269e0a03cSPaolo Bonzini mem->region = area; 23369e0a03cSPaolo Bonzini 234fbafbb6dSCameron Esfahani if (do_hvf_set_memory(mem, flags)) { 2352d9178d9SLaurent Vivier error_report("Error registering new memory slot"); 23669e0a03cSPaolo Bonzini abort(); 23769e0a03cSPaolo Bonzini } 23869e0a03cSPaolo Bonzini } 23969e0a03cSPaolo Bonzini 24069e0a03cSPaolo Bonzini void vmx_update_tpr(CPUState *cpu) 24169e0a03cSPaolo Bonzini { 24269e0a03cSPaolo Bonzini /* TODO: need integrate APIC handling */ 24369e0a03cSPaolo Bonzini X86CPU *x86_cpu = X86_CPU(cpu); 24469e0a03cSPaolo Bonzini int tpr = cpu_get_apic_tpr(x86_cpu->apic_state) << 4; 24569e0a03cSPaolo Bonzini int irr = apic_get_highest_priority_irr(x86_cpu->apic_state); 24669e0a03cSPaolo Bonzini 24769e0a03cSPaolo Bonzini wreg(cpu->hvf_fd, HV_X86_TPR, tpr); 24869e0a03cSPaolo Bonzini if (irr == -1) { 24969e0a03cSPaolo Bonzini wvmcs(cpu->hvf_fd, VMCS_TPR_THRESHOLD, 0); 25069e0a03cSPaolo Bonzini } else { 25169e0a03cSPaolo Bonzini wvmcs(cpu->hvf_fd, VMCS_TPR_THRESHOLD, (irr > tpr) ? tpr >> 4 : 25269e0a03cSPaolo Bonzini irr >> 4); 25369e0a03cSPaolo Bonzini } 25469e0a03cSPaolo Bonzini } 25569e0a03cSPaolo Bonzini 256583ae161SRoman Bolshakov static void update_apic_tpr(CPUState *cpu) 25769e0a03cSPaolo Bonzini { 25869e0a03cSPaolo Bonzini X86CPU *x86_cpu = X86_CPU(cpu); 25969e0a03cSPaolo Bonzini int tpr = rreg(cpu->hvf_fd, HV_X86_TPR) >> 4; 26069e0a03cSPaolo Bonzini cpu_set_apic_tpr(x86_cpu->apic_state, tpr); 26169e0a03cSPaolo Bonzini } 26269e0a03cSPaolo Bonzini 26369e0a03cSPaolo Bonzini #define VECTORING_INFO_VECTOR_MASK 0xff 26469e0a03cSPaolo Bonzini 26569e0a03cSPaolo Bonzini static void hvf_handle_interrupt(CPUState * cpu, int mask) 26669e0a03cSPaolo Bonzini { 26769e0a03cSPaolo Bonzini cpu->interrupt_request |= mask; 26869e0a03cSPaolo Bonzini if (!qemu_cpu_is_self(cpu)) { 26969e0a03cSPaolo Bonzini qemu_cpu_kick(cpu); 27069e0a03cSPaolo Bonzini } 27169e0a03cSPaolo Bonzini } 27269e0a03cSPaolo Bonzini 27369e0a03cSPaolo Bonzini void hvf_handle_io(CPUArchState *env, uint16_t port, void *buffer, 27469e0a03cSPaolo Bonzini int direction, int size, int count) 27569e0a03cSPaolo Bonzini { 27669e0a03cSPaolo Bonzini int i; 27769e0a03cSPaolo Bonzini uint8_t *ptr = buffer; 27869e0a03cSPaolo Bonzini 27969e0a03cSPaolo Bonzini for (i = 0; i < count; i++) { 28069e0a03cSPaolo Bonzini address_space_rw(&address_space_io, port, MEMTXATTRS_UNSPECIFIED, 28169e0a03cSPaolo Bonzini ptr, size, 28269e0a03cSPaolo Bonzini direction); 28369e0a03cSPaolo Bonzini ptr += size; 28469e0a03cSPaolo Bonzini } 28569e0a03cSPaolo Bonzini } 28669e0a03cSPaolo Bonzini 28769e0a03cSPaolo Bonzini static void do_hvf_cpu_synchronize_state(CPUState *cpu, run_on_cpu_data arg) 28869e0a03cSPaolo Bonzini { 289eae009deSRoman Bolshakov if (!cpu->vcpu_dirty) { 290eae009deSRoman Bolshakov hvf_get_registers(cpu); 291eae009deSRoman Bolshakov cpu->vcpu_dirty = true; 292eae009deSRoman Bolshakov } 29369e0a03cSPaolo Bonzini } 29469e0a03cSPaolo Bonzini 295eae009deSRoman Bolshakov void hvf_cpu_synchronize_state(CPUState *cpu) 29669e0a03cSPaolo Bonzini { 297eae009deSRoman Bolshakov if (!cpu->vcpu_dirty) { 298eae009deSRoman Bolshakov run_on_cpu(cpu, do_hvf_cpu_synchronize_state, RUN_ON_CPU_NULL); 29969e0a03cSPaolo Bonzini } 30069e0a03cSPaolo Bonzini } 30169e0a03cSPaolo Bonzini 302eae009deSRoman Bolshakov static void do_hvf_cpu_synchronize_post_reset(CPUState *cpu, 303eae009deSRoman Bolshakov run_on_cpu_data arg) 30469e0a03cSPaolo Bonzini { 305eae009deSRoman Bolshakov hvf_put_registers(cpu); 306eae009deSRoman Bolshakov cpu->vcpu_dirty = false; 30769e0a03cSPaolo Bonzini } 30869e0a03cSPaolo Bonzini 309eae009deSRoman Bolshakov void hvf_cpu_synchronize_post_reset(CPUState *cpu) 31069e0a03cSPaolo Bonzini { 311eae009deSRoman Bolshakov run_on_cpu(cpu, do_hvf_cpu_synchronize_post_reset, RUN_ON_CPU_NULL); 31269e0a03cSPaolo Bonzini } 31369e0a03cSPaolo Bonzini 314583ae161SRoman Bolshakov static void do_hvf_cpu_synchronize_post_init(CPUState *cpu, 315583ae161SRoman Bolshakov run_on_cpu_data arg) 31669e0a03cSPaolo Bonzini { 317eae009deSRoman Bolshakov hvf_put_registers(cpu); 318eae009deSRoman Bolshakov cpu->vcpu_dirty = false; 31969e0a03cSPaolo Bonzini } 32069e0a03cSPaolo Bonzini 321eae009deSRoman Bolshakov void hvf_cpu_synchronize_post_init(CPUState *cpu) 32269e0a03cSPaolo Bonzini { 323eae009deSRoman Bolshakov run_on_cpu(cpu, do_hvf_cpu_synchronize_post_init, RUN_ON_CPU_NULL); 32469e0a03cSPaolo Bonzini } 32569e0a03cSPaolo Bonzini 3265536c98eSRoman Bolshakov static void do_hvf_cpu_synchronize_pre_loadvm(CPUState *cpu, 3275536c98eSRoman Bolshakov run_on_cpu_data arg) 3285536c98eSRoman Bolshakov { 3295536c98eSRoman Bolshakov cpu->vcpu_dirty = true; 3305536c98eSRoman Bolshakov } 3315536c98eSRoman Bolshakov 3325536c98eSRoman Bolshakov void hvf_cpu_synchronize_pre_loadvm(CPUState *cpu) 3335536c98eSRoman Bolshakov { 3345536c98eSRoman Bolshakov run_on_cpu(cpu, do_hvf_cpu_synchronize_pre_loadvm, RUN_ON_CPU_NULL); 3355536c98eSRoman Bolshakov } 3365536c98eSRoman Bolshakov 337ff2de166SPaolo Bonzini static bool ept_emulation_fault(hvf_slot *slot, uint64_t gpa, uint64_t ept_qual) 33869e0a03cSPaolo Bonzini { 33969e0a03cSPaolo Bonzini int read, write; 34069e0a03cSPaolo Bonzini 34169e0a03cSPaolo Bonzini /* EPT fault on an instruction fetch doesn't make sense here */ 34269e0a03cSPaolo Bonzini if (ept_qual & EPT_VIOLATION_INST_FETCH) { 34369e0a03cSPaolo Bonzini return false; 34469e0a03cSPaolo Bonzini } 34569e0a03cSPaolo Bonzini 34669e0a03cSPaolo Bonzini /* EPT fault must be a read fault or a write fault */ 34769e0a03cSPaolo Bonzini read = ept_qual & EPT_VIOLATION_DATA_READ ? 1 : 0; 34869e0a03cSPaolo Bonzini write = ept_qual & EPT_VIOLATION_DATA_WRITE ? 1 : 0; 34969e0a03cSPaolo Bonzini if ((read | write) == 0) { 35069e0a03cSPaolo Bonzini return false; 35169e0a03cSPaolo Bonzini } 35269e0a03cSPaolo Bonzini 35369e0a03cSPaolo Bonzini if (write && slot) { 35469e0a03cSPaolo Bonzini if (slot->flags & HVF_SLOT_LOG) { 35569e0a03cSPaolo Bonzini memory_region_set_dirty(slot->region, gpa - slot->start, 1); 35669e0a03cSPaolo Bonzini hv_vm_protect((hv_gpaddr_t)slot->start, (size_t)slot->size, 35769e0a03cSPaolo Bonzini HV_MEMORY_READ | HV_MEMORY_WRITE); 35869e0a03cSPaolo Bonzini } 35969e0a03cSPaolo Bonzini } 36069e0a03cSPaolo Bonzini 36169e0a03cSPaolo Bonzini /* 36269e0a03cSPaolo Bonzini * The EPT violation must have been caused by accessing a 36369e0a03cSPaolo Bonzini * guest-physical address that is a translation of a guest-linear 36469e0a03cSPaolo Bonzini * address. 36569e0a03cSPaolo Bonzini */ 36669e0a03cSPaolo Bonzini if ((ept_qual & EPT_VIOLATION_GLA_VALID) == 0 || 36769e0a03cSPaolo Bonzini (ept_qual & EPT_VIOLATION_XLAT_VALID) == 0) { 36869e0a03cSPaolo Bonzini return false; 36969e0a03cSPaolo Bonzini } 37069e0a03cSPaolo Bonzini 371fbafbb6dSCameron Esfahani if (!slot) { 372fbafbb6dSCameron Esfahani return true; 373fbafbb6dSCameron Esfahani } 374fbafbb6dSCameron Esfahani if (!memory_region_is_ram(slot->region) && 375fbafbb6dSCameron Esfahani !(read && memory_region_is_romd(slot->region))) { 376fbafbb6dSCameron Esfahani return true; 377fbafbb6dSCameron Esfahani } 378fbafbb6dSCameron Esfahani return false; 37969e0a03cSPaolo Bonzini } 38069e0a03cSPaolo Bonzini 38169e0a03cSPaolo Bonzini static void hvf_set_dirty_tracking(MemoryRegionSection *section, bool on) 38269e0a03cSPaolo Bonzini { 38369e0a03cSPaolo Bonzini hvf_slot *slot; 38469e0a03cSPaolo Bonzini 38569e0a03cSPaolo Bonzini slot = hvf_find_overlap_slot( 38669e0a03cSPaolo Bonzini section->offset_within_address_space, 387fbafbb6dSCameron Esfahani int128_get64(section->size)); 38869e0a03cSPaolo Bonzini 38969e0a03cSPaolo Bonzini /* protect region against writes; begin tracking it */ 39069e0a03cSPaolo Bonzini if (on) { 39169e0a03cSPaolo Bonzini slot->flags |= HVF_SLOT_LOG; 39269e0a03cSPaolo Bonzini hv_vm_protect((hv_gpaddr_t)slot->start, (size_t)slot->size, 39369e0a03cSPaolo Bonzini HV_MEMORY_READ); 39469e0a03cSPaolo Bonzini /* stop tracking region*/ 39569e0a03cSPaolo Bonzini } else { 39669e0a03cSPaolo Bonzini slot->flags &= ~HVF_SLOT_LOG; 39769e0a03cSPaolo Bonzini hv_vm_protect((hv_gpaddr_t)slot->start, (size_t)slot->size, 39869e0a03cSPaolo Bonzini HV_MEMORY_READ | HV_MEMORY_WRITE); 39969e0a03cSPaolo Bonzini } 40069e0a03cSPaolo Bonzini } 40169e0a03cSPaolo Bonzini 40269e0a03cSPaolo Bonzini static void hvf_log_start(MemoryListener *listener, 40369e0a03cSPaolo Bonzini MemoryRegionSection *section, int old, int new) 40469e0a03cSPaolo Bonzini { 40569e0a03cSPaolo Bonzini if (old != 0) { 40669e0a03cSPaolo Bonzini return; 40769e0a03cSPaolo Bonzini } 40869e0a03cSPaolo Bonzini 40969e0a03cSPaolo Bonzini hvf_set_dirty_tracking(section, 1); 41069e0a03cSPaolo Bonzini } 41169e0a03cSPaolo Bonzini 41269e0a03cSPaolo Bonzini static void hvf_log_stop(MemoryListener *listener, 41369e0a03cSPaolo Bonzini MemoryRegionSection *section, int old, int new) 41469e0a03cSPaolo Bonzini { 41569e0a03cSPaolo Bonzini if (new != 0) { 41669e0a03cSPaolo Bonzini return; 41769e0a03cSPaolo Bonzini } 41869e0a03cSPaolo Bonzini 41969e0a03cSPaolo Bonzini hvf_set_dirty_tracking(section, 0); 42069e0a03cSPaolo Bonzini } 42169e0a03cSPaolo Bonzini 42269e0a03cSPaolo Bonzini static void hvf_log_sync(MemoryListener *listener, 42369e0a03cSPaolo Bonzini MemoryRegionSection *section) 42469e0a03cSPaolo Bonzini { 42569e0a03cSPaolo Bonzini /* 42669e0a03cSPaolo Bonzini * sync of dirty pages is handled elsewhere; just make sure we keep 42769e0a03cSPaolo Bonzini * tracking the region. 42869e0a03cSPaolo Bonzini */ 42969e0a03cSPaolo Bonzini hvf_set_dirty_tracking(section, 1); 43069e0a03cSPaolo Bonzini } 43169e0a03cSPaolo Bonzini 43269e0a03cSPaolo Bonzini static void hvf_region_add(MemoryListener *listener, 43369e0a03cSPaolo Bonzini MemoryRegionSection *section) 43469e0a03cSPaolo Bonzini { 43569e0a03cSPaolo Bonzini hvf_set_phys_mem(section, true); 43669e0a03cSPaolo Bonzini } 43769e0a03cSPaolo Bonzini 43869e0a03cSPaolo Bonzini static void hvf_region_del(MemoryListener *listener, 43969e0a03cSPaolo Bonzini MemoryRegionSection *section) 44069e0a03cSPaolo Bonzini { 44169e0a03cSPaolo Bonzini hvf_set_phys_mem(section, false); 44269e0a03cSPaolo Bonzini } 44369e0a03cSPaolo Bonzini 44469e0a03cSPaolo Bonzini static MemoryListener hvf_memory_listener = { 44569e0a03cSPaolo Bonzini .priority = 10, 44669e0a03cSPaolo Bonzini .region_add = hvf_region_add, 44769e0a03cSPaolo Bonzini .region_del = hvf_region_del, 44869e0a03cSPaolo Bonzini .log_start = hvf_log_start, 44969e0a03cSPaolo Bonzini .log_stop = hvf_log_stop, 45069e0a03cSPaolo Bonzini .log_sync = hvf_log_sync, 45169e0a03cSPaolo Bonzini }; 45269e0a03cSPaolo Bonzini 45369e0a03cSPaolo Bonzini void hvf_vcpu_destroy(CPUState *cpu) 45469e0a03cSPaolo Bonzini { 455fe76b09cSRoman Bolshakov X86CPU *x86_cpu = X86_CPU(cpu); 456fe76b09cSRoman Bolshakov CPUX86State *env = &x86_cpu->env; 457fe76b09cSRoman Bolshakov 45869e0a03cSPaolo Bonzini hv_return_t ret = hv_vcpu_destroy((hv_vcpuid_t)cpu->hvf_fd); 459fe76b09cSRoman Bolshakov g_free(env->hvf_mmio_buf); 46069e0a03cSPaolo Bonzini assert_hvf_ok(ret); 46169e0a03cSPaolo Bonzini } 46269e0a03cSPaolo Bonzini 46369e0a03cSPaolo Bonzini static void dummy_signal(int sig) 46469e0a03cSPaolo Bonzini { 46569e0a03cSPaolo Bonzini } 46669e0a03cSPaolo Bonzini 46769e0a03cSPaolo Bonzini int hvf_init_vcpu(CPUState *cpu) 46869e0a03cSPaolo Bonzini { 46969e0a03cSPaolo Bonzini 47069e0a03cSPaolo Bonzini X86CPU *x86cpu = X86_CPU(cpu); 47169e0a03cSPaolo Bonzini CPUX86State *env = &x86cpu->env; 47269e0a03cSPaolo Bonzini int r; 47369e0a03cSPaolo Bonzini 47469e0a03cSPaolo Bonzini /* init cpu signals */ 47569e0a03cSPaolo Bonzini sigset_t set; 47669e0a03cSPaolo Bonzini struct sigaction sigact; 47769e0a03cSPaolo Bonzini 47869e0a03cSPaolo Bonzini memset(&sigact, 0, sizeof(sigact)); 47969e0a03cSPaolo Bonzini sigact.sa_handler = dummy_signal; 48069e0a03cSPaolo Bonzini sigaction(SIG_IPI, &sigact, NULL); 48169e0a03cSPaolo Bonzini 48269e0a03cSPaolo Bonzini pthread_sigmask(SIG_BLOCK, NULL, &set); 48369e0a03cSPaolo Bonzini sigdelset(&set, SIG_IPI); 48469e0a03cSPaolo Bonzini 48569e0a03cSPaolo Bonzini init_emu(); 48669e0a03cSPaolo Bonzini init_decoder(); 48769e0a03cSPaolo Bonzini 48869e0a03cSPaolo Bonzini hvf_state->hvf_caps = g_new0(struct hvf_vcpu_caps, 1); 489fe76b09cSRoman Bolshakov env->hvf_mmio_buf = g_new(char, 4096); 49069e0a03cSPaolo Bonzini 49169e0a03cSPaolo Bonzini r = hv_vcpu_create((hv_vcpuid_t *)&cpu->hvf_fd, HV_VCPU_DEFAULT); 49269e0a03cSPaolo Bonzini cpu->vcpu_dirty = 1; 49369e0a03cSPaolo Bonzini assert_hvf_ok(r); 49469e0a03cSPaolo Bonzini 49569e0a03cSPaolo Bonzini if (hv_vmx_read_capability(HV_VMX_CAP_PINBASED, 49669e0a03cSPaolo Bonzini &hvf_state->hvf_caps->vmx_cap_pinbased)) { 49769e0a03cSPaolo Bonzini abort(); 49869e0a03cSPaolo Bonzini } 49969e0a03cSPaolo Bonzini if (hv_vmx_read_capability(HV_VMX_CAP_PROCBASED, 50069e0a03cSPaolo Bonzini &hvf_state->hvf_caps->vmx_cap_procbased)) { 50169e0a03cSPaolo Bonzini abort(); 50269e0a03cSPaolo Bonzini } 50369e0a03cSPaolo Bonzini if (hv_vmx_read_capability(HV_VMX_CAP_PROCBASED2, 50469e0a03cSPaolo Bonzini &hvf_state->hvf_caps->vmx_cap_procbased2)) { 50569e0a03cSPaolo Bonzini abort(); 50669e0a03cSPaolo Bonzini } 50769e0a03cSPaolo Bonzini if (hv_vmx_read_capability(HV_VMX_CAP_ENTRY, 50869e0a03cSPaolo Bonzini &hvf_state->hvf_caps->vmx_cap_entry)) { 50969e0a03cSPaolo Bonzini abort(); 51069e0a03cSPaolo Bonzini } 51169e0a03cSPaolo Bonzini 51269e0a03cSPaolo Bonzini /* set VMCS control fields */ 51369e0a03cSPaolo Bonzini wvmcs(cpu->hvf_fd, VMCS_PIN_BASED_CTLS, 51469e0a03cSPaolo Bonzini cap2ctrl(hvf_state->hvf_caps->vmx_cap_pinbased, 51569e0a03cSPaolo Bonzini VMCS_PIN_BASED_CTLS_EXTINT | 51669e0a03cSPaolo Bonzini VMCS_PIN_BASED_CTLS_NMI | 51769e0a03cSPaolo Bonzini VMCS_PIN_BASED_CTLS_VNMI)); 51869e0a03cSPaolo Bonzini wvmcs(cpu->hvf_fd, VMCS_PRI_PROC_BASED_CTLS, 51969e0a03cSPaolo Bonzini cap2ctrl(hvf_state->hvf_caps->vmx_cap_procbased, 52069e0a03cSPaolo Bonzini VMCS_PRI_PROC_BASED_CTLS_HLT | 52169e0a03cSPaolo Bonzini VMCS_PRI_PROC_BASED_CTLS_MWAIT | 52269e0a03cSPaolo Bonzini VMCS_PRI_PROC_BASED_CTLS_TSC_OFFSET | 52369e0a03cSPaolo Bonzini VMCS_PRI_PROC_BASED_CTLS_TPR_SHADOW) | 52469e0a03cSPaolo Bonzini VMCS_PRI_PROC_BASED_CTLS_SEC_CONTROL); 52569e0a03cSPaolo Bonzini wvmcs(cpu->hvf_fd, VMCS_SEC_PROC_BASED_CTLS, 52669e0a03cSPaolo Bonzini cap2ctrl(hvf_state->hvf_caps->vmx_cap_procbased2, 52769e0a03cSPaolo Bonzini VMCS_PRI_PROC_BASED2_CTLS_APIC_ACCESSES)); 52869e0a03cSPaolo Bonzini 52969e0a03cSPaolo Bonzini wvmcs(cpu->hvf_fd, VMCS_ENTRY_CTLS, cap2ctrl(hvf_state->hvf_caps->vmx_cap_entry, 53069e0a03cSPaolo Bonzini 0)); 53169e0a03cSPaolo Bonzini wvmcs(cpu->hvf_fd, VMCS_EXCEPTION_BITMAP, 0); /* Double fault */ 53269e0a03cSPaolo Bonzini 53369e0a03cSPaolo Bonzini wvmcs(cpu->hvf_fd, VMCS_TPR_THRESHOLD, 0); 53469e0a03cSPaolo Bonzini 53569e0a03cSPaolo Bonzini x86cpu = X86_CPU(cpu); 5365b8063c4SLiran Alon x86cpu->env.xsave_buf = qemu_memalign(4096, 4096); 53769e0a03cSPaolo Bonzini 53869e0a03cSPaolo Bonzini hv_vcpu_enable_native_msr(cpu->hvf_fd, MSR_STAR, 1); 53969e0a03cSPaolo Bonzini hv_vcpu_enable_native_msr(cpu->hvf_fd, MSR_LSTAR, 1); 54069e0a03cSPaolo Bonzini hv_vcpu_enable_native_msr(cpu->hvf_fd, MSR_CSTAR, 1); 54169e0a03cSPaolo Bonzini hv_vcpu_enable_native_msr(cpu->hvf_fd, MSR_FMASK, 1); 54269e0a03cSPaolo Bonzini hv_vcpu_enable_native_msr(cpu->hvf_fd, MSR_FSBASE, 1); 54369e0a03cSPaolo Bonzini hv_vcpu_enable_native_msr(cpu->hvf_fd, MSR_GSBASE, 1); 54469e0a03cSPaolo Bonzini hv_vcpu_enable_native_msr(cpu->hvf_fd, MSR_KERNELGSBASE, 1); 54569e0a03cSPaolo Bonzini hv_vcpu_enable_native_msr(cpu->hvf_fd, MSR_TSC_AUX, 1); 5469fedbbeeSCameron Esfahani hv_vcpu_enable_native_msr(cpu->hvf_fd, MSR_IA32_TSC, 1); 54769e0a03cSPaolo Bonzini hv_vcpu_enable_native_msr(cpu->hvf_fd, MSR_IA32_SYSENTER_CS, 1); 54869e0a03cSPaolo Bonzini hv_vcpu_enable_native_msr(cpu->hvf_fd, MSR_IA32_SYSENTER_EIP, 1); 54969e0a03cSPaolo Bonzini hv_vcpu_enable_native_msr(cpu->hvf_fd, MSR_IA32_SYSENTER_ESP, 1); 55069e0a03cSPaolo Bonzini 55169e0a03cSPaolo Bonzini return 0; 55269e0a03cSPaolo Bonzini } 55369e0a03cSPaolo Bonzini 55469e0a03cSPaolo Bonzini static void hvf_store_events(CPUState *cpu, uint32_t ins_len, uint64_t idtvec_info) 55569e0a03cSPaolo Bonzini { 55669e0a03cSPaolo Bonzini X86CPU *x86_cpu = X86_CPU(cpu); 55769e0a03cSPaolo Bonzini CPUX86State *env = &x86_cpu->env; 55869e0a03cSPaolo Bonzini 559fd13f23bSLiran Alon env->exception_nr = -1; 560fd13f23bSLiran Alon env->exception_pending = 0; 561fd13f23bSLiran Alon env->exception_injected = 0; 56269e0a03cSPaolo Bonzini env->interrupt_injected = -1; 56369e0a03cSPaolo Bonzini env->nmi_injected = false; 56464bef038SCameron Esfahani env->ins_len = 0; 56564bef038SCameron Esfahani env->has_error_code = false; 56669e0a03cSPaolo Bonzini if (idtvec_info & VMCS_IDT_VEC_VALID) { 56769e0a03cSPaolo Bonzini switch (idtvec_info & VMCS_IDT_VEC_TYPE) { 56869e0a03cSPaolo Bonzini case VMCS_IDT_VEC_HWINTR: 56969e0a03cSPaolo Bonzini case VMCS_IDT_VEC_SWINTR: 57069e0a03cSPaolo Bonzini env->interrupt_injected = idtvec_info & VMCS_IDT_VEC_VECNUM; 57169e0a03cSPaolo Bonzini break; 57269e0a03cSPaolo Bonzini case VMCS_IDT_VEC_NMI: 57369e0a03cSPaolo Bonzini env->nmi_injected = true; 57469e0a03cSPaolo Bonzini break; 57569e0a03cSPaolo Bonzini case VMCS_IDT_VEC_HWEXCEPTION: 57669e0a03cSPaolo Bonzini case VMCS_IDT_VEC_SWEXCEPTION: 577fd13f23bSLiran Alon env->exception_nr = idtvec_info & VMCS_IDT_VEC_VECNUM; 578fd13f23bSLiran Alon env->exception_injected = 1; 57969e0a03cSPaolo Bonzini break; 58069e0a03cSPaolo Bonzini case VMCS_IDT_VEC_PRIV_SWEXCEPTION: 58169e0a03cSPaolo Bonzini default: 58269e0a03cSPaolo Bonzini abort(); 58369e0a03cSPaolo Bonzini } 58469e0a03cSPaolo Bonzini if ((idtvec_info & VMCS_IDT_VEC_TYPE) == VMCS_IDT_VEC_SWEXCEPTION || 58569e0a03cSPaolo Bonzini (idtvec_info & VMCS_IDT_VEC_TYPE) == VMCS_IDT_VEC_SWINTR) { 58669e0a03cSPaolo Bonzini env->ins_len = ins_len; 58769e0a03cSPaolo Bonzini } 58864bef038SCameron Esfahani if (idtvec_info & VMCS_IDT_VEC_ERRCODE_VALID) { 58969e0a03cSPaolo Bonzini env->has_error_code = true; 59069e0a03cSPaolo Bonzini env->error_code = rvmcs(cpu->hvf_fd, VMCS_IDT_VECTORING_ERROR); 59169e0a03cSPaolo Bonzini } 59269e0a03cSPaolo Bonzini } 59369e0a03cSPaolo Bonzini if ((rvmcs(cpu->hvf_fd, VMCS_GUEST_INTERRUPTIBILITY) & 59469e0a03cSPaolo Bonzini VMCS_INTERRUPTIBILITY_NMI_BLOCKING)) { 59569e0a03cSPaolo Bonzini env->hflags2 |= HF2_NMI_MASK; 59669e0a03cSPaolo Bonzini } else { 59769e0a03cSPaolo Bonzini env->hflags2 &= ~HF2_NMI_MASK; 59869e0a03cSPaolo Bonzini } 59969e0a03cSPaolo Bonzini if (rvmcs(cpu->hvf_fd, VMCS_GUEST_INTERRUPTIBILITY) & 60069e0a03cSPaolo Bonzini (VMCS_INTERRUPTIBILITY_STI_BLOCKING | 60169e0a03cSPaolo Bonzini VMCS_INTERRUPTIBILITY_MOVSS_BLOCKING)) { 60269e0a03cSPaolo Bonzini env->hflags |= HF_INHIBIT_IRQ_MASK; 60369e0a03cSPaolo Bonzini } else { 60469e0a03cSPaolo Bonzini env->hflags &= ~HF_INHIBIT_IRQ_MASK; 60569e0a03cSPaolo Bonzini } 60669e0a03cSPaolo Bonzini } 60769e0a03cSPaolo Bonzini 60869e0a03cSPaolo Bonzini int hvf_vcpu_exec(CPUState *cpu) 60969e0a03cSPaolo Bonzini { 61069e0a03cSPaolo Bonzini X86CPU *x86_cpu = X86_CPU(cpu); 61169e0a03cSPaolo Bonzini CPUX86State *env = &x86_cpu->env; 61269e0a03cSPaolo Bonzini int ret = 0; 61369e0a03cSPaolo Bonzini uint64_t rip = 0; 61469e0a03cSPaolo Bonzini 61569e0a03cSPaolo Bonzini if (hvf_process_events(cpu)) { 61669e0a03cSPaolo Bonzini return EXCP_HLT; 61769e0a03cSPaolo Bonzini } 61869e0a03cSPaolo Bonzini 61969e0a03cSPaolo Bonzini do { 62069e0a03cSPaolo Bonzini if (cpu->vcpu_dirty) { 62169e0a03cSPaolo Bonzini hvf_put_registers(cpu); 62269e0a03cSPaolo Bonzini cpu->vcpu_dirty = false; 62369e0a03cSPaolo Bonzini } 62469e0a03cSPaolo Bonzini 62569e0a03cSPaolo Bonzini if (hvf_inject_interrupts(cpu)) { 62669e0a03cSPaolo Bonzini return EXCP_INTERRUPT; 62769e0a03cSPaolo Bonzini } 62869e0a03cSPaolo Bonzini vmx_update_tpr(cpu); 62969e0a03cSPaolo Bonzini 63069e0a03cSPaolo Bonzini qemu_mutex_unlock_iothread(); 63169e0a03cSPaolo Bonzini if (!cpu_is_bsp(X86_CPU(cpu)) && cpu->halted) { 63269e0a03cSPaolo Bonzini qemu_mutex_lock_iothread(); 63369e0a03cSPaolo Bonzini return EXCP_HLT; 63469e0a03cSPaolo Bonzini } 63569e0a03cSPaolo Bonzini 63669e0a03cSPaolo Bonzini hv_return_t r = hv_vcpu_run(cpu->hvf_fd); 63769e0a03cSPaolo Bonzini assert_hvf_ok(r); 63869e0a03cSPaolo Bonzini 63969e0a03cSPaolo Bonzini /* handle VMEXIT */ 64069e0a03cSPaolo Bonzini uint64_t exit_reason = rvmcs(cpu->hvf_fd, VMCS_EXIT_REASON); 64169e0a03cSPaolo Bonzini uint64_t exit_qual = rvmcs(cpu->hvf_fd, VMCS_EXIT_QUALIFICATION); 64269e0a03cSPaolo Bonzini uint32_t ins_len = (uint32_t)rvmcs(cpu->hvf_fd, 64369e0a03cSPaolo Bonzini VMCS_EXIT_INSTRUCTION_LENGTH); 64469e0a03cSPaolo Bonzini 64569e0a03cSPaolo Bonzini uint64_t idtvec_info = rvmcs(cpu->hvf_fd, VMCS_IDT_VECTORING_INFO); 64669e0a03cSPaolo Bonzini 64769e0a03cSPaolo Bonzini hvf_store_events(cpu, ins_len, idtvec_info); 64869e0a03cSPaolo Bonzini rip = rreg(cpu->hvf_fd, HV_X86_RIP); 649967f4da2SRoman Bolshakov env->eflags = rreg(cpu->hvf_fd, HV_X86_RFLAGS); 65069e0a03cSPaolo Bonzini 65169e0a03cSPaolo Bonzini qemu_mutex_lock_iothread(); 65269e0a03cSPaolo Bonzini 65369e0a03cSPaolo Bonzini update_apic_tpr(cpu); 65469e0a03cSPaolo Bonzini current_cpu = cpu; 65569e0a03cSPaolo Bonzini 65669e0a03cSPaolo Bonzini ret = 0; 65769e0a03cSPaolo Bonzini switch (exit_reason) { 65869e0a03cSPaolo Bonzini case EXIT_REASON_HLT: { 65969e0a03cSPaolo Bonzini macvm_set_rip(cpu, rip + ins_len); 66069e0a03cSPaolo Bonzini if (!((cpu->interrupt_request & CPU_INTERRUPT_HARD) && 661967f4da2SRoman Bolshakov (env->eflags & IF_MASK)) 66269e0a03cSPaolo Bonzini && !(cpu->interrupt_request & CPU_INTERRUPT_NMI) && 66369e0a03cSPaolo Bonzini !(idtvec_info & VMCS_IDT_VEC_VALID)) { 66469e0a03cSPaolo Bonzini cpu->halted = 1; 66569e0a03cSPaolo Bonzini ret = EXCP_HLT; 6663b9c59daSChen Zhang break; 66769e0a03cSPaolo Bonzini } 66869e0a03cSPaolo Bonzini ret = EXCP_INTERRUPT; 66969e0a03cSPaolo Bonzini break; 67069e0a03cSPaolo Bonzini } 67169e0a03cSPaolo Bonzini case EXIT_REASON_MWAIT: { 67269e0a03cSPaolo Bonzini ret = EXCP_INTERRUPT; 67369e0a03cSPaolo Bonzini break; 67469e0a03cSPaolo Bonzini } 675fbafbb6dSCameron Esfahani /* Need to check if MMIO or unmapped fault */ 67669e0a03cSPaolo Bonzini case EXIT_REASON_EPT_FAULT: 67769e0a03cSPaolo Bonzini { 67869e0a03cSPaolo Bonzini hvf_slot *slot; 679ff2de166SPaolo Bonzini uint64_t gpa = rvmcs(cpu->hvf_fd, VMCS_GUEST_PHYSICAL_ADDRESS); 68069e0a03cSPaolo Bonzini 68169e0a03cSPaolo Bonzini if (((idtvec_info & VMCS_IDT_VEC_VALID) == 0) && 68269e0a03cSPaolo Bonzini ((exit_qual & EXIT_QUAL_NMIUDTI) != 0)) { 68369e0a03cSPaolo Bonzini vmx_set_nmi_blocking(cpu); 68469e0a03cSPaolo Bonzini } 68569e0a03cSPaolo Bonzini 686fbafbb6dSCameron Esfahani slot = hvf_find_overlap_slot(gpa, 1); 68769e0a03cSPaolo Bonzini /* mmio */ 68869e0a03cSPaolo Bonzini if (ept_emulation_fault(slot, gpa, exit_qual)) { 68969e0a03cSPaolo Bonzini struct x86_decode decode; 69069e0a03cSPaolo Bonzini 69169e0a03cSPaolo Bonzini load_regs(cpu); 69269e0a03cSPaolo Bonzini decode_instruction(env, &decode); 69369e0a03cSPaolo Bonzini exec_instruction(env, &decode); 69469e0a03cSPaolo Bonzini store_regs(cpu); 69569e0a03cSPaolo Bonzini break; 69669e0a03cSPaolo Bonzini } 69769e0a03cSPaolo Bonzini break; 69869e0a03cSPaolo Bonzini } 69969e0a03cSPaolo Bonzini case EXIT_REASON_INOUT: 70069e0a03cSPaolo Bonzini { 70169e0a03cSPaolo Bonzini uint32_t in = (exit_qual & 8) != 0; 70269e0a03cSPaolo Bonzini uint32_t size = (exit_qual & 7) + 1; 70369e0a03cSPaolo Bonzini uint32_t string = (exit_qual & 16) != 0; 70469e0a03cSPaolo Bonzini uint32_t port = exit_qual >> 16; 70569e0a03cSPaolo Bonzini /*uint32_t rep = (exit_qual & 0x20) != 0;*/ 70669e0a03cSPaolo Bonzini 70769e0a03cSPaolo Bonzini if (!string && in) { 70869e0a03cSPaolo Bonzini uint64_t val = 0; 70969e0a03cSPaolo Bonzini load_regs(cpu); 71069e0a03cSPaolo Bonzini hvf_handle_io(env, port, &val, 0, size, 1); 71169e0a03cSPaolo Bonzini if (size == 1) { 71269e0a03cSPaolo Bonzini AL(env) = val; 71369e0a03cSPaolo Bonzini } else if (size == 2) { 71469e0a03cSPaolo Bonzini AX(env) = val; 71569e0a03cSPaolo Bonzini } else if (size == 4) { 71669e0a03cSPaolo Bonzini RAX(env) = (uint32_t)val; 71769e0a03cSPaolo Bonzini } else { 718da20f5cdSPaolo Bonzini RAX(env) = (uint64_t)val; 71969e0a03cSPaolo Bonzini } 7205d32173fSRoman Bolshakov env->eip += ins_len; 72169e0a03cSPaolo Bonzini store_regs(cpu); 72269e0a03cSPaolo Bonzini break; 72369e0a03cSPaolo Bonzini } else if (!string && !in) { 72469e0a03cSPaolo Bonzini RAX(env) = rreg(cpu->hvf_fd, HV_X86_RAX); 72569e0a03cSPaolo Bonzini hvf_handle_io(env, port, &RAX(env), 1, size, 1); 72669e0a03cSPaolo Bonzini macvm_set_rip(cpu, rip + ins_len); 72769e0a03cSPaolo Bonzini break; 72869e0a03cSPaolo Bonzini } 72969e0a03cSPaolo Bonzini struct x86_decode decode; 73069e0a03cSPaolo Bonzini 73169e0a03cSPaolo Bonzini load_regs(cpu); 73269e0a03cSPaolo Bonzini decode_instruction(env, &decode); 733e62963bfSPaolo Bonzini assert(ins_len == decode.len); 73469e0a03cSPaolo Bonzini exec_instruction(env, &decode); 73569e0a03cSPaolo Bonzini store_regs(cpu); 73669e0a03cSPaolo Bonzini 73769e0a03cSPaolo Bonzini break; 73869e0a03cSPaolo Bonzini } 73969e0a03cSPaolo Bonzini case EXIT_REASON_CPUID: { 74069e0a03cSPaolo Bonzini uint32_t rax = (uint32_t)rreg(cpu->hvf_fd, HV_X86_RAX); 74169e0a03cSPaolo Bonzini uint32_t rbx = (uint32_t)rreg(cpu->hvf_fd, HV_X86_RBX); 74269e0a03cSPaolo Bonzini uint32_t rcx = (uint32_t)rreg(cpu->hvf_fd, HV_X86_RCX); 74369e0a03cSPaolo Bonzini uint32_t rdx = (uint32_t)rreg(cpu->hvf_fd, HV_X86_RDX); 74469e0a03cSPaolo Bonzini 74569e0a03cSPaolo Bonzini cpu_x86_cpuid(env, rax, rcx, &rax, &rbx, &rcx, &rdx); 74669e0a03cSPaolo Bonzini 74769e0a03cSPaolo Bonzini wreg(cpu->hvf_fd, HV_X86_RAX, rax); 74869e0a03cSPaolo Bonzini wreg(cpu->hvf_fd, HV_X86_RBX, rbx); 74969e0a03cSPaolo Bonzini wreg(cpu->hvf_fd, HV_X86_RCX, rcx); 75069e0a03cSPaolo Bonzini wreg(cpu->hvf_fd, HV_X86_RDX, rdx); 75169e0a03cSPaolo Bonzini 75269e0a03cSPaolo Bonzini macvm_set_rip(cpu, rip + ins_len); 75369e0a03cSPaolo Bonzini break; 75469e0a03cSPaolo Bonzini } 75569e0a03cSPaolo Bonzini case EXIT_REASON_XSETBV: { 75669e0a03cSPaolo Bonzini X86CPU *x86_cpu = X86_CPU(cpu); 75769e0a03cSPaolo Bonzini CPUX86State *env = &x86_cpu->env; 75869e0a03cSPaolo Bonzini uint32_t eax = (uint32_t)rreg(cpu->hvf_fd, HV_X86_RAX); 75969e0a03cSPaolo Bonzini uint32_t ecx = (uint32_t)rreg(cpu->hvf_fd, HV_X86_RCX); 76069e0a03cSPaolo Bonzini uint32_t edx = (uint32_t)rreg(cpu->hvf_fd, HV_X86_RDX); 76169e0a03cSPaolo Bonzini 76269e0a03cSPaolo Bonzini if (ecx) { 76369e0a03cSPaolo Bonzini macvm_set_rip(cpu, rip + ins_len); 76469e0a03cSPaolo Bonzini break; 76569e0a03cSPaolo Bonzini } 76669e0a03cSPaolo Bonzini env->xcr0 = ((uint64_t)edx << 32) | eax; 76769e0a03cSPaolo Bonzini wreg(cpu->hvf_fd, HV_X86_XCR0, env->xcr0 | 1); 76869e0a03cSPaolo Bonzini macvm_set_rip(cpu, rip + ins_len); 76969e0a03cSPaolo Bonzini break; 77069e0a03cSPaolo Bonzini } 77169e0a03cSPaolo Bonzini case EXIT_REASON_INTR_WINDOW: 77269e0a03cSPaolo Bonzini vmx_clear_int_window_exiting(cpu); 77369e0a03cSPaolo Bonzini ret = EXCP_INTERRUPT; 77469e0a03cSPaolo Bonzini break; 77569e0a03cSPaolo Bonzini case EXIT_REASON_NMI_WINDOW: 77669e0a03cSPaolo Bonzini vmx_clear_nmi_window_exiting(cpu); 77769e0a03cSPaolo Bonzini ret = EXCP_INTERRUPT; 77869e0a03cSPaolo Bonzini break; 77969e0a03cSPaolo Bonzini case EXIT_REASON_EXT_INTR: 78069e0a03cSPaolo Bonzini /* force exit and allow io handling */ 78169e0a03cSPaolo Bonzini ret = EXCP_INTERRUPT; 78269e0a03cSPaolo Bonzini break; 78369e0a03cSPaolo Bonzini case EXIT_REASON_RDMSR: 78469e0a03cSPaolo Bonzini case EXIT_REASON_WRMSR: 78569e0a03cSPaolo Bonzini { 78669e0a03cSPaolo Bonzini load_regs(cpu); 78769e0a03cSPaolo Bonzini if (exit_reason == EXIT_REASON_RDMSR) { 78869e0a03cSPaolo Bonzini simulate_rdmsr(cpu); 78969e0a03cSPaolo Bonzini } else { 79069e0a03cSPaolo Bonzini simulate_wrmsr(cpu); 79169e0a03cSPaolo Bonzini } 7925d32173fSRoman Bolshakov env->eip += ins_len; 79369e0a03cSPaolo Bonzini store_regs(cpu); 79469e0a03cSPaolo Bonzini break; 79569e0a03cSPaolo Bonzini } 79669e0a03cSPaolo Bonzini case EXIT_REASON_CR_ACCESS: { 79769e0a03cSPaolo Bonzini int cr; 79869e0a03cSPaolo Bonzini int reg; 79969e0a03cSPaolo Bonzini 80069e0a03cSPaolo Bonzini load_regs(cpu); 80169e0a03cSPaolo Bonzini cr = exit_qual & 15; 80269e0a03cSPaolo Bonzini reg = (exit_qual >> 8) & 15; 80369e0a03cSPaolo Bonzini 80469e0a03cSPaolo Bonzini switch (cr) { 80569e0a03cSPaolo Bonzini case 0x0: { 80669e0a03cSPaolo Bonzini macvm_set_cr0(cpu->hvf_fd, RRX(env, reg)); 80769e0a03cSPaolo Bonzini break; 80869e0a03cSPaolo Bonzini } 80969e0a03cSPaolo Bonzini case 4: { 81069e0a03cSPaolo Bonzini macvm_set_cr4(cpu->hvf_fd, RRX(env, reg)); 81169e0a03cSPaolo Bonzini break; 81269e0a03cSPaolo Bonzini } 81369e0a03cSPaolo Bonzini case 8: { 81469e0a03cSPaolo Bonzini X86CPU *x86_cpu = X86_CPU(cpu); 81569e0a03cSPaolo Bonzini if (exit_qual & 0x10) { 81669e0a03cSPaolo Bonzini RRX(env, reg) = cpu_get_apic_tpr(x86_cpu->apic_state); 81769e0a03cSPaolo Bonzini } else { 81869e0a03cSPaolo Bonzini int tpr = RRX(env, reg); 81969e0a03cSPaolo Bonzini cpu_set_apic_tpr(x86_cpu->apic_state, tpr); 82069e0a03cSPaolo Bonzini ret = EXCP_INTERRUPT; 82169e0a03cSPaolo Bonzini } 82269e0a03cSPaolo Bonzini break; 82369e0a03cSPaolo Bonzini } 82469e0a03cSPaolo Bonzini default: 8252d9178d9SLaurent Vivier error_report("Unrecognized CR %d", cr); 82669e0a03cSPaolo Bonzini abort(); 82769e0a03cSPaolo Bonzini } 8285d32173fSRoman Bolshakov env->eip += ins_len; 82969e0a03cSPaolo Bonzini store_regs(cpu); 83069e0a03cSPaolo Bonzini break; 83169e0a03cSPaolo Bonzini } 83269e0a03cSPaolo Bonzini case EXIT_REASON_APIC_ACCESS: { /* TODO */ 83369e0a03cSPaolo Bonzini struct x86_decode decode; 83469e0a03cSPaolo Bonzini 83569e0a03cSPaolo Bonzini load_regs(cpu); 83669e0a03cSPaolo Bonzini decode_instruction(env, &decode); 83769e0a03cSPaolo Bonzini exec_instruction(env, &decode); 83869e0a03cSPaolo Bonzini store_regs(cpu); 83969e0a03cSPaolo Bonzini break; 84069e0a03cSPaolo Bonzini } 84169e0a03cSPaolo Bonzini case EXIT_REASON_TPR: { 84269e0a03cSPaolo Bonzini ret = 1; 84369e0a03cSPaolo Bonzini break; 84469e0a03cSPaolo Bonzini } 84569e0a03cSPaolo Bonzini case EXIT_REASON_TASK_SWITCH: { 84669e0a03cSPaolo Bonzini uint64_t vinfo = rvmcs(cpu->hvf_fd, VMCS_IDT_VECTORING_INFO); 84769e0a03cSPaolo Bonzini x68_segment_selector sel = {.sel = exit_qual & 0xffff}; 84869e0a03cSPaolo Bonzini vmx_handle_task_switch(cpu, sel, (exit_qual >> 30) & 0x3, 84969e0a03cSPaolo Bonzini vinfo & VMCS_INTR_VALID, vinfo & VECTORING_INFO_VECTOR_MASK, vinfo 85069e0a03cSPaolo Bonzini & VMCS_INTR_T_MASK); 85169e0a03cSPaolo Bonzini break; 85269e0a03cSPaolo Bonzini } 85369e0a03cSPaolo Bonzini case EXIT_REASON_TRIPLE_FAULT: { 85469e0a03cSPaolo Bonzini qemu_system_reset_request(SHUTDOWN_CAUSE_GUEST_RESET); 85569e0a03cSPaolo Bonzini ret = EXCP_INTERRUPT; 85669e0a03cSPaolo Bonzini break; 85769e0a03cSPaolo Bonzini } 85869e0a03cSPaolo Bonzini case EXIT_REASON_RDPMC: 85969e0a03cSPaolo Bonzini wreg(cpu->hvf_fd, HV_X86_RAX, 0); 86069e0a03cSPaolo Bonzini wreg(cpu->hvf_fd, HV_X86_RDX, 0); 86169e0a03cSPaolo Bonzini macvm_set_rip(cpu, rip + ins_len); 86269e0a03cSPaolo Bonzini break; 86369e0a03cSPaolo Bonzini case VMX_REASON_VMCALL: 864fd13f23bSLiran Alon env->exception_nr = EXCP0D_GPF; 865fd13f23bSLiran Alon env->exception_injected = 1; 86669e0a03cSPaolo Bonzini env->has_error_code = true; 86769e0a03cSPaolo Bonzini env->error_code = 0; 86869e0a03cSPaolo Bonzini break; 86969e0a03cSPaolo Bonzini default: 8702d9178d9SLaurent Vivier error_report("%llx: unhandled exit %llx", rip, exit_reason); 87169e0a03cSPaolo Bonzini } 87269e0a03cSPaolo Bonzini } while (ret == 0); 87369e0a03cSPaolo Bonzini 87469e0a03cSPaolo Bonzini return ret; 87569e0a03cSPaolo Bonzini } 87669e0a03cSPaolo Bonzini 87792cc3aaaSRoman Bolshakov bool hvf_allowed; 87869e0a03cSPaolo Bonzini 87969e0a03cSPaolo Bonzini static int hvf_accel_init(MachineState *ms) 88069e0a03cSPaolo Bonzini { 88169e0a03cSPaolo Bonzini int x; 88269e0a03cSPaolo Bonzini hv_return_t ret; 88369e0a03cSPaolo Bonzini HVFState *s; 88469e0a03cSPaolo Bonzini 88569e0a03cSPaolo Bonzini ret = hv_vm_create(HV_VM_DEFAULT); 88669e0a03cSPaolo Bonzini assert_hvf_ok(ret); 88769e0a03cSPaolo Bonzini 88869e0a03cSPaolo Bonzini s = g_new0(HVFState, 1); 88969e0a03cSPaolo Bonzini 89069e0a03cSPaolo Bonzini s->num_slots = 32; 89169e0a03cSPaolo Bonzini for (x = 0; x < s->num_slots; ++x) { 89269e0a03cSPaolo Bonzini s->slots[x].size = 0; 89369e0a03cSPaolo Bonzini s->slots[x].slot_id = x; 89469e0a03cSPaolo Bonzini } 89569e0a03cSPaolo Bonzini 89669e0a03cSPaolo Bonzini hvf_state = s; 89769e0a03cSPaolo Bonzini cpu_interrupt_handler = hvf_handle_interrupt; 89869e0a03cSPaolo Bonzini memory_listener_register(&hvf_memory_listener, &address_space_memory); 899*b52bcba7SClaudio Fontana cpus_register_accel(&hvf_cpus); 90069e0a03cSPaolo Bonzini return 0; 90169e0a03cSPaolo Bonzini } 90269e0a03cSPaolo Bonzini 90369e0a03cSPaolo Bonzini static void hvf_accel_class_init(ObjectClass *oc, void *data) 90469e0a03cSPaolo Bonzini { 90569e0a03cSPaolo Bonzini AccelClass *ac = ACCEL_CLASS(oc); 90669e0a03cSPaolo Bonzini ac->name = "HVF"; 90769e0a03cSPaolo Bonzini ac->init_machine = hvf_accel_init; 90869e0a03cSPaolo Bonzini ac->allowed = &hvf_allowed; 90969e0a03cSPaolo Bonzini } 91069e0a03cSPaolo Bonzini 91169e0a03cSPaolo Bonzini static const TypeInfo hvf_accel_type = { 91269e0a03cSPaolo Bonzini .name = TYPE_HVF_ACCEL, 91369e0a03cSPaolo Bonzini .parent = TYPE_ACCEL, 91469e0a03cSPaolo Bonzini .class_init = hvf_accel_class_init, 91569e0a03cSPaolo Bonzini }; 91669e0a03cSPaolo Bonzini 91769e0a03cSPaolo Bonzini static void hvf_type_init(void) 91869e0a03cSPaolo Bonzini { 91969e0a03cSPaolo Bonzini type_register_static(&hvf_accel_type); 92069e0a03cSPaolo Bonzini } 92169e0a03cSPaolo Bonzini 92269e0a03cSPaolo Bonzini type_init(hvf_type_init); 923