169e0a03cSPaolo Bonzini /* Copyright 2008 IBM Corporation 269e0a03cSPaolo Bonzini * 2008 Red Hat, Inc. 369e0a03cSPaolo Bonzini * Copyright 2011 Intel Corporation 469e0a03cSPaolo Bonzini * Copyright 2016 Veertu, Inc. 569e0a03cSPaolo Bonzini * Copyright 2017 The Android Open Source Project 669e0a03cSPaolo Bonzini * 769e0a03cSPaolo Bonzini * QEMU Hypervisor.framework support 869e0a03cSPaolo Bonzini * 969e0a03cSPaolo Bonzini * This program is free software; you can redistribute it and/or 1069e0a03cSPaolo Bonzini * modify it under the terms of version 2 of the GNU General Public 1169e0a03cSPaolo Bonzini * License as published by the Free Software Foundation. 1269e0a03cSPaolo Bonzini * 1369e0a03cSPaolo Bonzini * This program is distributed in the hope that it will be useful, 1469e0a03cSPaolo Bonzini * but WITHOUT ANY WARRANTY; without even the implied warranty of 1569e0a03cSPaolo Bonzini * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU 16e361a772SThomas Huth * General Public License for more details. 1769e0a03cSPaolo Bonzini * 18e361a772SThomas Huth * You should have received a copy of the GNU General Public License 19e361a772SThomas Huth * along with this program; if not, see <http://www.gnu.org/licenses/>. 20d781e24dSIzik Eidus * 21d781e24dSIzik Eidus * This file contain code under public domain from the hvdos project: 22d781e24dSIzik Eidus * https://github.com/mist64/hvdos 234d98a8e5SPaolo Bonzini * 244d98a8e5SPaolo Bonzini * Parts Copyright (c) 2011 NetApp, Inc. 254d98a8e5SPaolo Bonzini * All rights reserved. 264d98a8e5SPaolo Bonzini * 274d98a8e5SPaolo Bonzini * Redistribution and use in source and binary forms, with or without 284d98a8e5SPaolo Bonzini * modification, are permitted provided that the following conditions 294d98a8e5SPaolo Bonzini * are met: 304d98a8e5SPaolo Bonzini * 1. Redistributions of source code must retain the above copyright 314d98a8e5SPaolo Bonzini * notice, this list of conditions and the following disclaimer. 324d98a8e5SPaolo Bonzini * 2. Redistributions in binary form must reproduce the above copyright 334d98a8e5SPaolo Bonzini * notice, this list of conditions and the following disclaimer in the 344d98a8e5SPaolo Bonzini * documentation and/or other materials provided with the distribution. 354d98a8e5SPaolo Bonzini * 364d98a8e5SPaolo Bonzini * THIS SOFTWARE IS PROVIDED BY NETAPP, INC ``AS IS'' AND 374d98a8e5SPaolo Bonzini * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE 384d98a8e5SPaolo Bonzini * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE 394d98a8e5SPaolo Bonzini * ARE DISCLAIMED. IN NO EVENT SHALL NETAPP, INC OR CONTRIBUTORS BE LIABLE 404d98a8e5SPaolo Bonzini * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL 414d98a8e5SPaolo Bonzini * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS 424d98a8e5SPaolo Bonzini * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) 434d98a8e5SPaolo Bonzini * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT 444d98a8e5SPaolo Bonzini * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY 454d98a8e5SPaolo Bonzini * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF 464d98a8e5SPaolo Bonzini * SUCH DAMAGE. 4769e0a03cSPaolo Bonzini */ 4854d31236SMarkus Armbruster 4969e0a03cSPaolo Bonzini #include "qemu/osdep.h" 5069e0a03cSPaolo Bonzini #include "qemu-common.h" 5169e0a03cSPaolo Bonzini #include "qemu/error-report.h" 5269e0a03cSPaolo Bonzini 5369e0a03cSPaolo Bonzini #include "sysemu/hvf.h" 5454d31236SMarkus Armbruster #include "sysemu/runstate.h" 5569e0a03cSPaolo Bonzini #include "hvf-i386.h" 5669e0a03cSPaolo Bonzini #include "vmcs.h" 5769e0a03cSPaolo Bonzini #include "vmx.h" 5869e0a03cSPaolo Bonzini #include "x86.h" 5969e0a03cSPaolo Bonzini #include "x86_descr.h" 6069e0a03cSPaolo Bonzini #include "x86_mmu.h" 6169e0a03cSPaolo Bonzini #include "x86_decode.h" 6269e0a03cSPaolo Bonzini #include "x86_emu.h" 6369e0a03cSPaolo Bonzini #include "x86_task.h" 6469e0a03cSPaolo Bonzini #include "x86hvf.h" 6569e0a03cSPaolo Bonzini 6669e0a03cSPaolo Bonzini #include <Hypervisor/hv.h> 6769e0a03cSPaolo Bonzini #include <Hypervisor/hv_vmx.h> 6869e0a03cSPaolo Bonzini 6969e0a03cSPaolo Bonzini #include "exec/address-spaces.h" 7069e0a03cSPaolo Bonzini #include "hw/i386/apic_internal.h" 7169e0a03cSPaolo Bonzini #include "qemu/main-loop.h" 7269e0a03cSPaolo Bonzini #include "sysemu/accel.h" 7369e0a03cSPaolo Bonzini #include "target/i386/cpu.h" 7469e0a03cSPaolo Bonzini 7569e0a03cSPaolo Bonzini HVFState *hvf_state; 7669e0a03cSPaolo Bonzini 7769e0a03cSPaolo Bonzini static void assert_hvf_ok(hv_return_t ret) 7869e0a03cSPaolo Bonzini { 7969e0a03cSPaolo Bonzini if (ret == HV_SUCCESS) { 8069e0a03cSPaolo Bonzini return; 8169e0a03cSPaolo Bonzini } 8269e0a03cSPaolo Bonzini 8369e0a03cSPaolo Bonzini switch (ret) { 8469e0a03cSPaolo Bonzini case HV_ERROR: 852d9178d9SLaurent Vivier error_report("Error: HV_ERROR"); 8669e0a03cSPaolo Bonzini break; 8769e0a03cSPaolo Bonzini case HV_BUSY: 882d9178d9SLaurent Vivier error_report("Error: HV_BUSY"); 8969e0a03cSPaolo Bonzini break; 9069e0a03cSPaolo Bonzini case HV_BAD_ARGUMENT: 912d9178d9SLaurent Vivier error_report("Error: HV_BAD_ARGUMENT"); 9269e0a03cSPaolo Bonzini break; 9369e0a03cSPaolo Bonzini case HV_NO_RESOURCES: 942d9178d9SLaurent Vivier error_report("Error: HV_NO_RESOURCES"); 9569e0a03cSPaolo Bonzini break; 9669e0a03cSPaolo Bonzini case HV_NO_DEVICE: 972d9178d9SLaurent Vivier error_report("Error: HV_NO_DEVICE"); 9869e0a03cSPaolo Bonzini break; 9969e0a03cSPaolo Bonzini case HV_UNSUPPORTED: 1002d9178d9SLaurent Vivier error_report("Error: HV_UNSUPPORTED"); 10169e0a03cSPaolo Bonzini break; 10269e0a03cSPaolo Bonzini default: 1032d9178d9SLaurent Vivier error_report("Unknown Error"); 10469e0a03cSPaolo Bonzini } 10569e0a03cSPaolo Bonzini 10669e0a03cSPaolo Bonzini abort(); 10769e0a03cSPaolo Bonzini } 10869e0a03cSPaolo Bonzini 10969e0a03cSPaolo Bonzini /* Memory slots */ 110fbafbb6dSCameron Esfahani hvf_slot *hvf_find_overlap_slot(uint64_t start, uint64_t size) 11169e0a03cSPaolo Bonzini { 11269e0a03cSPaolo Bonzini hvf_slot *slot; 11369e0a03cSPaolo Bonzini int x; 11469e0a03cSPaolo Bonzini for (x = 0; x < hvf_state->num_slots; ++x) { 11569e0a03cSPaolo Bonzini slot = &hvf_state->slots[x]; 11669e0a03cSPaolo Bonzini if (slot->size && start < (slot->start + slot->size) && 117fbafbb6dSCameron Esfahani (start + size) > slot->start) { 11869e0a03cSPaolo Bonzini return slot; 11969e0a03cSPaolo Bonzini } 12069e0a03cSPaolo Bonzini } 12169e0a03cSPaolo Bonzini return NULL; 12269e0a03cSPaolo Bonzini } 12369e0a03cSPaolo Bonzini 12469e0a03cSPaolo Bonzini struct mac_slot { 12569e0a03cSPaolo Bonzini int present; 12669e0a03cSPaolo Bonzini uint64_t size; 12769e0a03cSPaolo Bonzini uint64_t gpa_start; 12869e0a03cSPaolo Bonzini uint64_t gva; 12969e0a03cSPaolo Bonzini }; 13069e0a03cSPaolo Bonzini 13169e0a03cSPaolo Bonzini struct mac_slot mac_slots[32]; 13269e0a03cSPaolo Bonzini 133fbafbb6dSCameron Esfahani static int do_hvf_set_memory(hvf_slot *slot, hv_memory_flags_t flags) 13469e0a03cSPaolo Bonzini { 13569e0a03cSPaolo Bonzini struct mac_slot *macslot; 13669e0a03cSPaolo Bonzini hv_return_t ret; 13769e0a03cSPaolo Bonzini 13869e0a03cSPaolo Bonzini macslot = &mac_slots[slot->slot_id]; 13969e0a03cSPaolo Bonzini 14069e0a03cSPaolo Bonzini if (macslot->present) { 14169e0a03cSPaolo Bonzini if (macslot->size != slot->size) { 14269e0a03cSPaolo Bonzini macslot->present = 0; 14369e0a03cSPaolo Bonzini ret = hv_vm_unmap(macslot->gpa_start, macslot->size); 14469e0a03cSPaolo Bonzini assert_hvf_ok(ret); 14569e0a03cSPaolo Bonzini } 14669e0a03cSPaolo Bonzini } 14769e0a03cSPaolo Bonzini 14869e0a03cSPaolo Bonzini if (!slot->size) { 14969e0a03cSPaolo Bonzini return 0; 15069e0a03cSPaolo Bonzini } 15169e0a03cSPaolo Bonzini 15269e0a03cSPaolo Bonzini macslot->present = 1; 15369e0a03cSPaolo Bonzini macslot->gpa_start = slot->start; 15469e0a03cSPaolo Bonzini macslot->size = slot->size; 15569e0a03cSPaolo Bonzini ret = hv_vm_map((hv_uvaddr_t)slot->mem, slot->start, slot->size, flags); 15669e0a03cSPaolo Bonzini assert_hvf_ok(ret); 15769e0a03cSPaolo Bonzini return 0; 15869e0a03cSPaolo Bonzini } 15969e0a03cSPaolo Bonzini 16069e0a03cSPaolo Bonzini void hvf_set_phys_mem(MemoryRegionSection *section, bool add) 16169e0a03cSPaolo Bonzini { 16269e0a03cSPaolo Bonzini hvf_slot *mem; 16369e0a03cSPaolo Bonzini MemoryRegion *area = section->mr; 164fbafbb6dSCameron Esfahani bool writeable = !area->readonly && !area->rom_device; 165fbafbb6dSCameron Esfahani hv_memory_flags_t flags; 16669e0a03cSPaolo Bonzini 16769e0a03cSPaolo Bonzini if (!memory_region_is_ram(area)) { 168fbafbb6dSCameron Esfahani if (writeable) { 16969e0a03cSPaolo Bonzini return; 170fbafbb6dSCameron Esfahani } else if (!memory_region_is_romd(area)) { 171fbafbb6dSCameron Esfahani /* 172fbafbb6dSCameron Esfahani * If the memory device is not in romd_mode, then we actually want 173fbafbb6dSCameron Esfahani * to remove the hvf memory slot so all accesses will trap. 174fbafbb6dSCameron Esfahani */ 175fbafbb6dSCameron Esfahani add = false; 176fbafbb6dSCameron Esfahani } 17769e0a03cSPaolo Bonzini } 17869e0a03cSPaolo Bonzini 17969e0a03cSPaolo Bonzini mem = hvf_find_overlap_slot( 18069e0a03cSPaolo Bonzini section->offset_within_address_space, 181fbafbb6dSCameron Esfahani int128_get64(section->size)); 18269e0a03cSPaolo Bonzini 18369e0a03cSPaolo Bonzini if (mem && add) { 18469e0a03cSPaolo Bonzini if (mem->size == int128_get64(section->size) && 18569e0a03cSPaolo Bonzini mem->start == section->offset_within_address_space && 18669e0a03cSPaolo Bonzini mem->mem == (memory_region_get_ram_ptr(area) + 18769e0a03cSPaolo Bonzini section->offset_within_region)) { 18869e0a03cSPaolo Bonzini return; /* Same region was attempted to register, go away. */ 18969e0a03cSPaolo Bonzini } 19069e0a03cSPaolo Bonzini } 19169e0a03cSPaolo Bonzini 19269e0a03cSPaolo Bonzini /* Region needs to be reset. set the size to 0 and remap it. */ 19369e0a03cSPaolo Bonzini if (mem) { 19469e0a03cSPaolo Bonzini mem->size = 0; 195fbafbb6dSCameron Esfahani if (do_hvf_set_memory(mem, 0)) { 1962d9178d9SLaurent Vivier error_report("Failed to reset overlapping slot"); 19769e0a03cSPaolo Bonzini abort(); 19869e0a03cSPaolo Bonzini } 19969e0a03cSPaolo Bonzini } 20069e0a03cSPaolo Bonzini 20169e0a03cSPaolo Bonzini if (!add) { 20269e0a03cSPaolo Bonzini return; 20369e0a03cSPaolo Bonzini } 20469e0a03cSPaolo Bonzini 205fbafbb6dSCameron Esfahani if (area->readonly || 206fbafbb6dSCameron Esfahani (!memory_region_is_ram(area) && memory_region_is_romd(area))) { 207fbafbb6dSCameron Esfahani flags = HV_MEMORY_READ | HV_MEMORY_EXEC; 208fbafbb6dSCameron Esfahani } else { 209fbafbb6dSCameron Esfahani flags = HV_MEMORY_READ | HV_MEMORY_WRITE | HV_MEMORY_EXEC; 210fbafbb6dSCameron Esfahani } 211fbafbb6dSCameron Esfahani 21269e0a03cSPaolo Bonzini /* Now make a new slot. */ 21369e0a03cSPaolo Bonzini int x; 21469e0a03cSPaolo Bonzini 21569e0a03cSPaolo Bonzini for (x = 0; x < hvf_state->num_slots; ++x) { 21669e0a03cSPaolo Bonzini mem = &hvf_state->slots[x]; 21769e0a03cSPaolo Bonzini if (!mem->size) { 21869e0a03cSPaolo Bonzini break; 21969e0a03cSPaolo Bonzini } 22069e0a03cSPaolo Bonzini } 22169e0a03cSPaolo Bonzini 22269e0a03cSPaolo Bonzini if (x == hvf_state->num_slots) { 2232d9178d9SLaurent Vivier error_report("No free slots"); 22469e0a03cSPaolo Bonzini abort(); 22569e0a03cSPaolo Bonzini } 22669e0a03cSPaolo Bonzini 22769e0a03cSPaolo Bonzini mem->size = int128_get64(section->size); 22869e0a03cSPaolo Bonzini mem->mem = memory_region_get_ram_ptr(area) + section->offset_within_region; 22969e0a03cSPaolo Bonzini mem->start = section->offset_within_address_space; 23069e0a03cSPaolo Bonzini mem->region = area; 23169e0a03cSPaolo Bonzini 232fbafbb6dSCameron Esfahani if (do_hvf_set_memory(mem, flags)) { 2332d9178d9SLaurent Vivier error_report("Error registering new memory slot"); 23469e0a03cSPaolo Bonzini abort(); 23569e0a03cSPaolo Bonzini } 23669e0a03cSPaolo Bonzini } 23769e0a03cSPaolo Bonzini 23869e0a03cSPaolo Bonzini void vmx_update_tpr(CPUState *cpu) 23969e0a03cSPaolo Bonzini { 24069e0a03cSPaolo Bonzini /* TODO: need integrate APIC handling */ 24169e0a03cSPaolo Bonzini X86CPU *x86_cpu = X86_CPU(cpu); 24269e0a03cSPaolo Bonzini int tpr = cpu_get_apic_tpr(x86_cpu->apic_state) << 4; 24369e0a03cSPaolo Bonzini int irr = apic_get_highest_priority_irr(x86_cpu->apic_state); 24469e0a03cSPaolo Bonzini 24569e0a03cSPaolo Bonzini wreg(cpu->hvf_fd, HV_X86_TPR, tpr); 24669e0a03cSPaolo Bonzini if (irr == -1) { 24769e0a03cSPaolo Bonzini wvmcs(cpu->hvf_fd, VMCS_TPR_THRESHOLD, 0); 24869e0a03cSPaolo Bonzini } else { 24969e0a03cSPaolo Bonzini wvmcs(cpu->hvf_fd, VMCS_TPR_THRESHOLD, (irr > tpr) ? tpr >> 4 : 25069e0a03cSPaolo Bonzini irr >> 4); 25169e0a03cSPaolo Bonzini } 25269e0a03cSPaolo Bonzini } 25369e0a03cSPaolo Bonzini 254583ae161SRoman Bolshakov static void update_apic_tpr(CPUState *cpu) 25569e0a03cSPaolo Bonzini { 25669e0a03cSPaolo Bonzini X86CPU *x86_cpu = X86_CPU(cpu); 25769e0a03cSPaolo Bonzini int tpr = rreg(cpu->hvf_fd, HV_X86_TPR) >> 4; 25869e0a03cSPaolo Bonzini cpu_set_apic_tpr(x86_cpu->apic_state, tpr); 25969e0a03cSPaolo Bonzini } 26069e0a03cSPaolo Bonzini 26169e0a03cSPaolo Bonzini #define VECTORING_INFO_VECTOR_MASK 0xff 26269e0a03cSPaolo Bonzini 26369e0a03cSPaolo Bonzini static void hvf_handle_interrupt(CPUState * cpu, int mask) 26469e0a03cSPaolo Bonzini { 26569e0a03cSPaolo Bonzini cpu->interrupt_request |= mask; 26669e0a03cSPaolo Bonzini if (!qemu_cpu_is_self(cpu)) { 26769e0a03cSPaolo Bonzini qemu_cpu_kick(cpu); 26869e0a03cSPaolo Bonzini } 26969e0a03cSPaolo Bonzini } 27069e0a03cSPaolo Bonzini 27169e0a03cSPaolo Bonzini void hvf_handle_io(CPUArchState *env, uint16_t port, void *buffer, 27269e0a03cSPaolo Bonzini int direction, int size, int count) 27369e0a03cSPaolo Bonzini { 27469e0a03cSPaolo Bonzini int i; 27569e0a03cSPaolo Bonzini uint8_t *ptr = buffer; 27669e0a03cSPaolo Bonzini 27769e0a03cSPaolo Bonzini for (i = 0; i < count; i++) { 27869e0a03cSPaolo Bonzini address_space_rw(&address_space_io, port, MEMTXATTRS_UNSPECIFIED, 27969e0a03cSPaolo Bonzini ptr, size, 28069e0a03cSPaolo Bonzini direction); 28169e0a03cSPaolo Bonzini ptr += size; 28269e0a03cSPaolo Bonzini } 28369e0a03cSPaolo Bonzini } 28469e0a03cSPaolo Bonzini 28569e0a03cSPaolo Bonzini /* TODO: synchronize vcpu state */ 28669e0a03cSPaolo Bonzini static void do_hvf_cpu_synchronize_state(CPUState *cpu, run_on_cpu_data arg) 28769e0a03cSPaolo Bonzini { 28869e0a03cSPaolo Bonzini CPUState *cpu_state = cpu; 28969e0a03cSPaolo Bonzini if (cpu_state->vcpu_dirty == 0) { 29069e0a03cSPaolo Bonzini hvf_get_registers(cpu_state); 29169e0a03cSPaolo Bonzini } 29269e0a03cSPaolo Bonzini 29369e0a03cSPaolo Bonzini cpu_state->vcpu_dirty = 1; 29469e0a03cSPaolo Bonzini } 29569e0a03cSPaolo Bonzini 29669e0a03cSPaolo Bonzini void hvf_cpu_synchronize_state(CPUState *cpu_state) 29769e0a03cSPaolo Bonzini { 29869e0a03cSPaolo Bonzini if (cpu_state->vcpu_dirty == 0) { 29969e0a03cSPaolo Bonzini run_on_cpu(cpu_state, do_hvf_cpu_synchronize_state, RUN_ON_CPU_NULL); 30069e0a03cSPaolo Bonzini } 30169e0a03cSPaolo Bonzini } 30269e0a03cSPaolo Bonzini 30369e0a03cSPaolo Bonzini static void do_hvf_cpu_synchronize_post_reset(CPUState *cpu, run_on_cpu_data arg) 30469e0a03cSPaolo Bonzini { 30569e0a03cSPaolo Bonzini CPUState *cpu_state = cpu; 30669e0a03cSPaolo Bonzini hvf_put_registers(cpu_state); 30769e0a03cSPaolo Bonzini cpu_state->vcpu_dirty = false; 30869e0a03cSPaolo Bonzini } 30969e0a03cSPaolo Bonzini 31069e0a03cSPaolo Bonzini void hvf_cpu_synchronize_post_reset(CPUState *cpu_state) 31169e0a03cSPaolo Bonzini { 31269e0a03cSPaolo Bonzini run_on_cpu(cpu_state, do_hvf_cpu_synchronize_post_reset, RUN_ON_CPU_NULL); 31369e0a03cSPaolo Bonzini } 31469e0a03cSPaolo Bonzini 315583ae161SRoman Bolshakov static void do_hvf_cpu_synchronize_post_init(CPUState *cpu, 316583ae161SRoman Bolshakov run_on_cpu_data arg) 31769e0a03cSPaolo Bonzini { 31869e0a03cSPaolo Bonzini CPUState *cpu_state = cpu; 31969e0a03cSPaolo Bonzini hvf_put_registers(cpu_state); 32069e0a03cSPaolo Bonzini cpu_state->vcpu_dirty = false; 32169e0a03cSPaolo Bonzini } 32269e0a03cSPaolo Bonzini 32369e0a03cSPaolo Bonzini void hvf_cpu_synchronize_post_init(CPUState *cpu_state) 32469e0a03cSPaolo Bonzini { 325583ae161SRoman Bolshakov run_on_cpu(cpu_state, do_hvf_cpu_synchronize_post_init, RUN_ON_CPU_NULL); 32669e0a03cSPaolo Bonzini } 32769e0a03cSPaolo Bonzini 328*5536c98eSRoman Bolshakov static void do_hvf_cpu_synchronize_pre_loadvm(CPUState *cpu, 329*5536c98eSRoman Bolshakov run_on_cpu_data arg) 330*5536c98eSRoman Bolshakov { 331*5536c98eSRoman Bolshakov cpu->vcpu_dirty = true; 332*5536c98eSRoman Bolshakov } 333*5536c98eSRoman Bolshakov 334*5536c98eSRoman Bolshakov void hvf_cpu_synchronize_pre_loadvm(CPUState *cpu) 335*5536c98eSRoman Bolshakov { 336*5536c98eSRoman Bolshakov run_on_cpu(cpu, do_hvf_cpu_synchronize_pre_loadvm, RUN_ON_CPU_NULL); 337*5536c98eSRoman Bolshakov } 338*5536c98eSRoman Bolshakov 339ff2de166SPaolo Bonzini static bool ept_emulation_fault(hvf_slot *slot, uint64_t gpa, uint64_t ept_qual) 34069e0a03cSPaolo Bonzini { 34169e0a03cSPaolo Bonzini int read, write; 34269e0a03cSPaolo Bonzini 34369e0a03cSPaolo Bonzini /* EPT fault on an instruction fetch doesn't make sense here */ 34469e0a03cSPaolo Bonzini if (ept_qual & EPT_VIOLATION_INST_FETCH) { 34569e0a03cSPaolo Bonzini return false; 34669e0a03cSPaolo Bonzini } 34769e0a03cSPaolo Bonzini 34869e0a03cSPaolo Bonzini /* EPT fault must be a read fault or a write fault */ 34969e0a03cSPaolo Bonzini read = ept_qual & EPT_VIOLATION_DATA_READ ? 1 : 0; 35069e0a03cSPaolo Bonzini write = ept_qual & EPT_VIOLATION_DATA_WRITE ? 1 : 0; 35169e0a03cSPaolo Bonzini if ((read | write) == 0) { 35269e0a03cSPaolo Bonzini return false; 35369e0a03cSPaolo Bonzini } 35469e0a03cSPaolo Bonzini 35569e0a03cSPaolo Bonzini if (write && slot) { 35669e0a03cSPaolo Bonzini if (slot->flags & HVF_SLOT_LOG) { 35769e0a03cSPaolo Bonzini memory_region_set_dirty(slot->region, gpa - slot->start, 1); 35869e0a03cSPaolo Bonzini hv_vm_protect((hv_gpaddr_t)slot->start, (size_t)slot->size, 35969e0a03cSPaolo Bonzini HV_MEMORY_READ | HV_MEMORY_WRITE); 36069e0a03cSPaolo Bonzini } 36169e0a03cSPaolo Bonzini } 36269e0a03cSPaolo Bonzini 36369e0a03cSPaolo Bonzini /* 36469e0a03cSPaolo Bonzini * The EPT violation must have been caused by accessing a 36569e0a03cSPaolo Bonzini * guest-physical address that is a translation of a guest-linear 36669e0a03cSPaolo Bonzini * address. 36769e0a03cSPaolo Bonzini */ 36869e0a03cSPaolo Bonzini if ((ept_qual & EPT_VIOLATION_GLA_VALID) == 0 || 36969e0a03cSPaolo Bonzini (ept_qual & EPT_VIOLATION_XLAT_VALID) == 0) { 37069e0a03cSPaolo Bonzini return false; 37169e0a03cSPaolo Bonzini } 37269e0a03cSPaolo Bonzini 373fbafbb6dSCameron Esfahani if (!slot) { 374fbafbb6dSCameron Esfahani return true; 375fbafbb6dSCameron Esfahani } 376fbafbb6dSCameron Esfahani if (!memory_region_is_ram(slot->region) && 377fbafbb6dSCameron Esfahani !(read && memory_region_is_romd(slot->region))) { 378fbafbb6dSCameron Esfahani return true; 379fbafbb6dSCameron Esfahani } 380fbafbb6dSCameron Esfahani return false; 38169e0a03cSPaolo Bonzini } 38269e0a03cSPaolo Bonzini 38369e0a03cSPaolo Bonzini static void hvf_set_dirty_tracking(MemoryRegionSection *section, bool on) 38469e0a03cSPaolo Bonzini { 38569e0a03cSPaolo Bonzini hvf_slot *slot; 38669e0a03cSPaolo Bonzini 38769e0a03cSPaolo Bonzini slot = hvf_find_overlap_slot( 38869e0a03cSPaolo Bonzini section->offset_within_address_space, 389fbafbb6dSCameron Esfahani int128_get64(section->size)); 39069e0a03cSPaolo Bonzini 39169e0a03cSPaolo Bonzini /* protect region against writes; begin tracking it */ 39269e0a03cSPaolo Bonzini if (on) { 39369e0a03cSPaolo Bonzini slot->flags |= HVF_SLOT_LOG; 39469e0a03cSPaolo Bonzini hv_vm_protect((hv_gpaddr_t)slot->start, (size_t)slot->size, 39569e0a03cSPaolo Bonzini HV_MEMORY_READ); 39669e0a03cSPaolo Bonzini /* stop tracking region*/ 39769e0a03cSPaolo Bonzini } else { 39869e0a03cSPaolo Bonzini slot->flags &= ~HVF_SLOT_LOG; 39969e0a03cSPaolo Bonzini hv_vm_protect((hv_gpaddr_t)slot->start, (size_t)slot->size, 40069e0a03cSPaolo Bonzini HV_MEMORY_READ | HV_MEMORY_WRITE); 40169e0a03cSPaolo Bonzini } 40269e0a03cSPaolo Bonzini } 40369e0a03cSPaolo Bonzini 40469e0a03cSPaolo Bonzini static void hvf_log_start(MemoryListener *listener, 40569e0a03cSPaolo Bonzini MemoryRegionSection *section, int old, int new) 40669e0a03cSPaolo Bonzini { 40769e0a03cSPaolo Bonzini if (old != 0) { 40869e0a03cSPaolo Bonzini return; 40969e0a03cSPaolo Bonzini } 41069e0a03cSPaolo Bonzini 41169e0a03cSPaolo Bonzini hvf_set_dirty_tracking(section, 1); 41269e0a03cSPaolo Bonzini } 41369e0a03cSPaolo Bonzini 41469e0a03cSPaolo Bonzini static void hvf_log_stop(MemoryListener *listener, 41569e0a03cSPaolo Bonzini MemoryRegionSection *section, int old, int new) 41669e0a03cSPaolo Bonzini { 41769e0a03cSPaolo Bonzini if (new != 0) { 41869e0a03cSPaolo Bonzini return; 41969e0a03cSPaolo Bonzini } 42069e0a03cSPaolo Bonzini 42169e0a03cSPaolo Bonzini hvf_set_dirty_tracking(section, 0); 42269e0a03cSPaolo Bonzini } 42369e0a03cSPaolo Bonzini 42469e0a03cSPaolo Bonzini static void hvf_log_sync(MemoryListener *listener, 42569e0a03cSPaolo Bonzini MemoryRegionSection *section) 42669e0a03cSPaolo Bonzini { 42769e0a03cSPaolo Bonzini /* 42869e0a03cSPaolo Bonzini * sync of dirty pages is handled elsewhere; just make sure we keep 42969e0a03cSPaolo Bonzini * tracking the region. 43069e0a03cSPaolo Bonzini */ 43169e0a03cSPaolo Bonzini hvf_set_dirty_tracking(section, 1); 43269e0a03cSPaolo Bonzini } 43369e0a03cSPaolo Bonzini 43469e0a03cSPaolo Bonzini static void hvf_region_add(MemoryListener *listener, 43569e0a03cSPaolo Bonzini MemoryRegionSection *section) 43669e0a03cSPaolo Bonzini { 43769e0a03cSPaolo Bonzini hvf_set_phys_mem(section, true); 43869e0a03cSPaolo Bonzini } 43969e0a03cSPaolo Bonzini 44069e0a03cSPaolo Bonzini static void hvf_region_del(MemoryListener *listener, 44169e0a03cSPaolo Bonzini MemoryRegionSection *section) 44269e0a03cSPaolo Bonzini { 44369e0a03cSPaolo Bonzini hvf_set_phys_mem(section, false); 44469e0a03cSPaolo Bonzini } 44569e0a03cSPaolo Bonzini 44669e0a03cSPaolo Bonzini static MemoryListener hvf_memory_listener = { 44769e0a03cSPaolo Bonzini .priority = 10, 44869e0a03cSPaolo Bonzini .region_add = hvf_region_add, 44969e0a03cSPaolo Bonzini .region_del = hvf_region_del, 45069e0a03cSPaolo Bonzini .log_start = hvf_log_start, 45169e0a03cSPaolo Bonzini .log_stop = hvf_log_stop, 45269e0a03cSPaolo Bonzini .log_sync = hvf_log_sync, 45369e0a03cSPaolo Bonzini }; 45469e0a03cSPaolo Bonzini 45569e0a03cSPaolo Bonzini void hvf_reset_vcpu(CPUState *cpu) { 456e37aa8b0SCameron Esfahani uint64_t pdpte[4] = {0, 0, 0, 0}; 457e37aa8b0SCameron Esfahani int i; 45869e0a03cSPaolo Bonzini 45969e0a03cSPaolo Bonzini /* TODO: this shouldn't be needed; there is already a call to 46069e0a03cSPaolo Bonzini * cpu_synchronize_all_post_reset in vl.c 46169e0a03cSPaolo Bonzini */ 46269e0a03cSPaolo Bonzini wvmcs(cpu->hvf_fd, VMCS_ENTRY_CTLS, 0); 46369e0a03cSPaolo Bonzini wvmcs(cpu->hvf_fd, VMCS_GUEST_IA32_EFER, 0); 464e37aa8b0SCameron Esfahani 465e37aa8b0SCameron Esfahani /* Initialize PDPTE */ 466e37aa8b0SCameron Esfahani for (i = 0; i < 4; i++) { 467e37aa8b0SCameron Esfahani wvmcs(cpu->hvf_fd, VMCS_GUEST_PDPTE0 + i * 2, pdpte[i]); 468e37aa8b0SCameron Esfahani } 469e37aa8b0SCameron Esfahani 47069e0a03cSPaolo Bonzini macvm_set_cr0(cpu->hvf_fd, 0x60000010); 47169e0a03cSPaolo Bonzini 47269e0a03cSPaolo Bonzini wvmcs(cpu->hvf_fd, VMCS_CR4_MASK, CR4_VMXE_MASK); 47369e0a03cSPaolo Bonzini wvmcs(cpu->hvf_fd, VMCS_CR4_SHADOW, 0x0); 47469e0a03cSPaolo Bonzini wvmcs(cpu->hvf_fd, VMCS_GUEST_CR4, CR4_VMXE_MASK); 47569e0a03cSPaolo Bonzini 47669e0a03cSPaolo Bonzini /* set VMCS guest state fields */ 47769e0a03cSPaolo Bonzini wvmcs(cpu->hvf_fd, VMCS_GUEST_CS_SELECTOR, 0xf000); 47869e0a03cSPaolo Bonzini wvmcs(cpu->hvf_fd, VMCS_GUEST_CS_LIMIT, 0xffff); 47969e0a03cSPaolo Bonzini wvmcs(cpu->hvf_fd, VMCS_GUEST_CS_ACCESS_RIGHTS, 0x9b); 48069e0a03cSPaolo Bonzini wvmcs(cpu->hvf_fd, VMCS_GUEST_CS_BASE, 0xffff0000); 48169e0a03cSPaolo Bonzini 48269e0a03cSPaolo Bonzini wvmcs(cpu->hvf_fd, VMCS_GUEST_DS_SELECTOR, 0); 48369e0a03cSPaolo Bonzini wvmcs(cpu->hvf_fd, VMCS_GUEST_DS_LIMIT, 0xffff); 48469e0a03cSPaolo Bonzini wvmcs(cpu->hvf_fd, VMCS_GUEST_DS_ACCESS_RIGHTS, 0x93); 48569e0a03cSPaolo Bonzini wvmcs(cpu->hvf_fd, VMCS_GUEST_DS_BASE, 0); 48669e0a03cSPaolo Bonzini 48769e0a03cSPaolo Bonzini wvmcs(cpu->hvf_fd, VMCS_GUEST_ES_SELECTOR, 0); 48869e0a03cSPaolo Bonzini wvmcs(cpu->hvf_fd, VMCS_GUEST_ES_LIMIT, 0xffff); 48969e0a03cSPaolo Bonzini wvmcs(cpu->hvf_fd, VMCS_GUEST_ES_ACCESS_RIGHTS, 0x93); 49069e0a03cSPaolo Bonzini wvmcs(cpu->hvf_fd, VMCS_GUEST_ES_BASE, 0); 49169e0a03cSPaolo Bonzini 49269e0a03cSPaolo Bonzini wvmcs(cpu->hvf_fd, VMCS_GUEST_FS_SELECTOR, 0); 49369e0a03cSPaolo Bonzini wvmcs(cpu->hvf_fd, VMCS_GUEST_FS_LIMIT, 0xffff); 49469e0a03cSPaolo Bonzini wvmcs(cpu->hvf_fd, VMCS_GUEST_FS_ACCESS_RIGHTS, 0x93); 49569e0a03cSPaolo Bonzini wvmcs(cpu->hvf_fd, VMCS_GUEST_FS_BASE, 0); 49669e0a03cSPaolo Bonzini 49769e0a03cSPaolo Bonzini wvmcs(cpu->hvf_fd, VMCS_GUEST_GS_SELECTOR, 0); 49869e0a03cSPaolo Bonzini wvmcs(cpu->hvf_fd, VMCS_GUEST_GS_LIMIT, 0xffff); 49969e0a03cSPaolo Bonzini wvmcs(cpu->hvf_fd, VMCS_GUEST_GS_ACCESS_RIGHTS, 0x93); 50069e0a03cSPaolo Bonzini wvmcs(cpu->hvf_fd, VMCS_GUEST_GS_BASE, 0); 50169e0a03cSPaolo Bonzini 50269e0a03cSPaolo Bonzini wvmcs(cpu->hvf_fd, VMCS_GUEST_SS_SELECTOR, 0); 50369e0a03cSPaolo Bonzini wvmcs(cpu->hvf_fd, VMCS_GUEST_SS_LIMIT, 0xffff); 50469e0a03cSPaolo Bonzini wvmcs(cpu->hvf_fd, VMCS_GUEST_SS_ACCESS_RIGHTS, 0x93); 50569e0a03cSPaolo Bonzini wvmcs(cpu->hvf_fd, VMCS_GUEST_SS_BASE, 0); 50669e0a03cSPaolo Bonzini 50769e0a03cSPaolo Bonzini wvmcs(cpu->hvf_fd, VMCS_GUEST_LDTR_SELECTOR, 0); 50869e0a03cSPaolo Bonzini wvmcs(cpu->hvf_fd, VMCS_GUEST_LDTR_LIMIT, 0); 50969e0a03cSPaolo Bonzini wvmcs(cpu->hvf_fd, VMCS_GUEST_LDTR_ACCESS_RIGHTS, 0x10000); 51069e0a03cSPaolo Bonzini wvmcs(cpu->hvf_fd, VMCS_GUEST_LDTR_BASE, 0); 51169e0a03cSPaolo Bonzini 51269e0a03cSPaolo Bonzini wvmcs(cpu->hvf_fd, VMCS_GUEST_TR_SELECTOR, 0); 51369e0a03cSPaolo Bonzini wvmcs(cpu->hvf_fd, VMCS_GUEST_TR_LIMIT, 0); 51469e0a03cSPaolo Bonzini wvmcs(cpu->hvf_fd, VMCS_GUEST_TR_ACCESS_RIGHTS, 0x83); 51569e0a03cSPaolo Bonzini wvmcs(cpu->hvf_fd, VMCS_GUEST_TR_BASE, 0); 51669e0a03cSPaolo Bonzini 51769e0a03cSPaolo Bonzini wvmcs(cpu->hvf_fd, VMCS_GUEST_GDTR_LIMIT, 0); 51869e0a03cSPaolo Bonzini wvmcs(cpu->hvf_fd, VMCS_GUEST_GDTR_BASE, 0); 51969e0a03cSPaolo Bonzini 52069e0a03cSPaolo Bonzini wvmcs(cpu->hvf_fd, VMCS_GUEST_IDTR_LIMIT, 0); 52169e0a03cSPaolo Bonzini wvmcs(cpu->hvf_fd, VMCS_GUEST_IDTR_BASE, 0); 52269e0a03cSPaolo Bonzini 52369e0a03cSPaolo Bonzini /*wvmcs(cpu->hvf_fd, VMCS_GUEST_CR2, 0x0);*/ 52469e0a03cSPaolo Bonzini wvmcs(cpu->hvf_fd, VMCS_GUEST_CR3, 0x0); 52569e0a03cSPaolo Bonzini 52669e0a03cSPaolo Bonzini wreg(cpu->hvf_fd, HV_X86_RIP, 0xfff0); 52769e0a03cSPaolo Bonzini wreg(cpu->hvf_fd, HV_X86_RDX, 0x623); 52869e0a03cSPaolo Bonzini wreg(cpu->hvf_fd, HV_X86_RFLAGS, 0x2); 52969e0a03cSPaolo Bonzini wreg(cpu->hvf_fd, HV_X86_RSP, 0x0); 53069e0a03cSPaolo Bonzini wreg(cpu->hvf_fd, HV_X86_RAX, 0x0); 53169e0a03cSPaolo Bonzini wreg(cpu->hvf_fd, HV_X86_RBX, 0x0); 53269e0a03cSPaolo Bonzini wreg(cpu->hvf_fd, HV_X86_RCX, 0x0); 53369e0a03cSPaolo Bonzini wreg(cpu->hvf_fd, HV_X86_RSI, 0x0); 53469e0a03cSPaolo Bonzini wreg(cpu->hvf_fd, HV_X86_RDI, 0x0); 53569e0a03cSPaolo Bonzini wreg(cpu->hvf_fd, HV_X86_RBP, 0x0); 53669e0a03cSPaolo Bonzini 53769e0a03cSPaolo Bonzini for (int i = 0; i < 8; i++) { 53869e0a03cSPaolo Bonzini wreg(cpu->hvf_fd, HV_X86_R8 + i, 0x0); 53969e0a03cSPaolo Bonzini } 54069e0a03cSPaolo Bonzini 54169e0a03cSPaolo Bonzini hv_vcpu_invalidate_tlb(cpu->hvf_fd); 54269e0a03cSPaolo Bonzini hv_vcpu_flush(cpu->hvf_fd); 54369e0a03cSPaolo Bonzini } 54469e0a03cSPaolo Bonzini 54569e0a03cSPaolo Bonzini void hvf_vcpu_destroy(CPUState *cpu) 54669e0a03cSPaolo Bonzini { 547fe76b09cSRoman Bolshakov X86CPU *x86_cpu = X86_CPU(cpu); 548fe76b09cSRoman Bolshakov CPUX86State *env = &x86_cpu->env; 549fe76b09cSRoman Bolshakov 55069e0a03cSPaolo Bonzini hv_return_t ret = hv_vcpu_destroy((hv_vcpuid_t)cpu->hvf_fd); 551fe76b09cSRoman Bolshakov g_free(env->hvf_mmio_buf); 55269e0a03cSPaolo Bonzini assert_hvf_ok(ret); 55369e0a03cSPaolo Bonzini } 55469e0a03cSPaolo Bonzini 55569e0a03cSPaolo Bonzini static void dummy_signal(int sig) 55669e0a03cSPaolo Bonzini { 55769e0a03cSPaolo Bonzini } 55869e0a03cSPaolo Bonzini 55969e0a03cSPaolo Bonzini int hvf_init_vcpu(CPUState *cpu) 56069e0a03cSPaolo Bonzini { 56169e0a03cSPaolo Bonzini 56269e0a03cSPaolo Bonzini X86CPU *x86cpu = X86_CPU(cpu); 56369e0a03cSPaolo Bonzini CPUX86State *env = &x86cpu->env; 56469e0a03cSPaolo Bonzini int r; 56569e0a03cSPaolo Bonzini 56669e0a03cSPaolo Bonzini /* init cpu signals */ 56769e0a03cSPaolo Bonzini sigset_t set; 56869e0a03cSPaolo Bonzini struct sigaction sigact; 56969e0a03cSPaolo Bonzini 57069e0a03cSPaolo Bonzini memset(&sigact, 0, sizeof(sigact)); 57169e0a03cSPaolo Bonzini sigact.sa_handler = dummy_signal; 57269e0a03cSPaolo Bonzini sigaction(SIG_IPI, &sigact, NULL); 57369e0a03cSPaolo Bonzini 57469e0a03cSPaolo Bonzini pthread_sigmask(SIG_BLOCK, NULL, &set); 57569e0a03cSPaolo Bonzini sigdelset(&set, SIG_IPI); 57669e0a03cSPaolo Bonzini 57769e0a03cSPaolo Bonzini init_emu(); 57869e0a03cSPaolo Bonzini init_decoder(); 57969e0a03cSPaolo Bonzini 58069e0a03cSPaolo Bonzini hvf_state->hvf_caps = g_new0(struct hvf_vcpu_caps, 1); 581fe76b09cSRoman Bolshakov env->hvf_mmio_buf = g_new(char, 4096); 58269e0a03cSPaolo Bonzini 58369e0a03cSPaolo Bonzini r = hv_vcpu_create((hv_vcpuid_t *)&cpu->hvf_fd, HV_VCPU_DEFAULT); 58469e0a03cSPaolo Bonzini cpu->vcpu_dirty = 1; 58569e0a03cSPaolo Bonzini assert_hvf_ok(r); 58669e0a03cSPaolo Bonzini 58769e0a03cSPaolo Bonzini if (hv_vmx_read_capability(HV_VMX_CAP_PINBASED, 58869e0a03cSPaolo Bonzini &hvf_state->hvf_caps->vmx_cap_pinbased)) { 58969e0a03cSPaolo Bonzini abort(); 59069e0a03cSPaolo Bonzini } 59169e0a03cSPaolo Bonzini if (hv_vmx_read_capability(HV_VMX_CAP_PROCBASED, 59269e0a03cSPaolo Bonzini &hvf_state->hvf_caps->vmx_cap_procbased)) { 59369e0a03cSPaolo Bonzini abort(); 59469e0a03cSPaolo Bonzini } 59569e0a03cSPaolo Bonzini if (hv_vmx_read_capability(HV_VMX_CAP_PROCBASED2, 59669e0a03cSPaolo Bonzini &hvf_state->hvf_caps->vmx_cap_procbased2)) { 59769e0a03cSPaolo Bonzini abort(); 59869e0a03cSPaolo Bonzini } 59969e0a03cSPaolo Bonzini if (hv_vmx_read_capability(HV_VMX_CAP_ENTRY, 60069e0a03cSPaolo Bonzini &hvf_state->hvf_caps->vmx_cap_entry)) { 60169e0a03cSPaolo Bonzini abort(); 60269e0a03cSPaolo Bonzini } 60369e0a03cSPaolo Bonzini 60469e0a03cSPaolo Bonzini /* set VMCS control fields */ 60569e0a03cSPaolo Bonzini wvmcs(cpu->hvf_fd, VMCS_PIN_BASED_CTLS, 60669e0a03cSPaolo Bonzini cap2ctrl(hvf_state->hvf_caps->vmx_cap_pinbased, 60769e0a03cSPaolo Bonzini VMCS_PIN_BASED_CTLS_EXTINT | 60869e0a03cSPaolo Bonzini VMCS_PIN_BASED_CTLS_NMI | 60969e0a03cSPaolo Bonzini VMCS_PIN_BASED_CTLS_VNMI)); 61069e0a03cSPaolo Bonzini wvmcs(cpu->hvf_fd, VMCS_PRI_PROC_BASED_CTLS, 61169e0a03cSPaolo Bonzini cap2ctrl(hvf_state->hvf_caps->vmx_cap_procbased, 61269e0a03cSPaolo Bonzini VMCS_PRI_PROC_BASED_CTLS_HLT | 61369e0a03cSPaolo Bonzini VMCS_PRI_PROC_BASED_CTLS_MWAIT | 61469e0a03cSPaolo Bonzini VMCS_PRI_PROC_BASED_CTLS_TSC_OFFSET | 61569e0a03cSPaolo Bonzini VMCS_PRI_PROC_BASED_CTLS_TPR_SHADOW) | 61669e0a03cSPaolo Bonzini VMCS_PRI_PROC_BASED_CTLS_SEC_CONTROL); 61769e0a03cSPaolo Bonzini wvmcs(cpu->hvf_fd, VMCS_SEC_PROC_BASED_CTLS, 61869e0a03cSPaolo Bonzini cap2ctrl(hvf_state->hvf_caps->vmx_cap_procbased2, 61969e0a03cSPaolo Bonzini VMCS_PRI_PROC_BASED2_CTLS_APIC_ACCESSES)); 62069e0a03cSPaolo Bonzini 62169e0a03cSPaolo Bonzini wvmcs(cpu->hvf_fd, VMCS_ENTRY_CTLS, cap2ctrl(hvf_state->hvf_caps->vmx_cap_entry, 62269e0a03cSPaolo Bonzini 0)); 62369e0a03cSPaolo Bonzini wvmcs(cpu->hvf_fd, VMCS_EXCEPTION_BITMAP, 0); /* Double fault */ 62469e0a03cSPaolo Bonzini 62569e0a03cSPaolo Bonzini wvmcs(cpu->hvf_fd, VMCS_TPR_THRESHOLD, 0); 62669e0a03cSPaolo Bonzini 62769e0a03cSPaolo Bonzini x86cpu = X86_CPU(cpu); 6285b8063c4SLiran Alon x86cpu->env.xsave_buf = qemu_memalign(4096, 4096); 62969e0a03cSPaolo Bonzini 63069e0a03cSPaolo Bonzini hv_vcpu_enable_native_msr(cpu->hvf_fd, MSR_STAR, 1); 63169e0a03cSPaolo Bonzini hv_vcpu_enable_native_msr(cpu->hvf_fd, MSR_LSTAR, 1); 63269e0a03cSPaolo Bonzini hv_vcpu_enable_native_msr(cpu->hvf_fd, MSR_CSTAR, 1); 63369e0a03cSPaolo Bonzini hv_vcpu_enable_native_msr(cpu->hvf_fd, MSR_FMASK, 1); 63469e0a03cSPaolo Bonzini hv_vcpu_enable_native_msr(cpu->hvf_fd, MSR_FSBASE, 1); 63569e0a03cSPaolo Bonzini hv_vcpu_enable_native_msr(cpu->hvf_fd, MSR_GSBASE, 1); 63669e0a03cSPaolo Bonzini hv_vcpu_enable_native_msr(cpu->hvf_fd, MSR_KERNELGSBASE, 1); 63769e0a03cSPaolo Bonzini hv_vcpu_enable_native_msr(cpu->hvf_fd, MSR_TSC_AUX, 1); 6389fedbbeeSCameron Esfahani hv_vcpu_enable_native_msr(cpu->hvf_fd, MSR_IA32_TSC, 1); 63969e0a03cSPaolo Bonzini hv_vcpu_enable_native_msr(cpu->hvf_fd, MSR_IA32_SYSENTER_CS, 1); 64069e0a03cSPaolo Bonzini hv_vcpu_enable_native_msr(cpu->hvf_fd, MSR_IA32_SYSENTER_EIP, 1); 64169e0a03cSPaolo Bonzini hv_vcpu_enable_native_msr(cpu->hvf_fd, MSR_IA32_SYSENTER_ESP, 1); 64269e0a03cSPaolo Bonzini 64369e0a03cSPaolo Bonzini return 0; 64469e0a03cSPaolo Bonzini } 64569e0a03cSPaolo Bonzini 64669e0a03cSPaolo Bonzini static void hvf_store_events(CPUState *cpu, uint32_t ins_len, uint64_t idtvec_info) 64769e0a03cSPaolo Bonzini { 64869e0a03cSPaolo Bonzini X86CPU *x86_cpu = X86_CPU(cpu); 64969e0a03cSPaolo Bonzini CPUX86State *env = &x86_cpu->env; 65069e0a03cSPaolo Bonzini 651fd13f23bSLiran Alon env->exception_nr = -1; 652fd13f23bSLiran Alon env->exception_pending = 0; 653fd13f23bSLiran Alon env->exception_injected = 0; 65469e0a03cSPaolo Bonzini env->interrupt_injected = -1; 65569e0a03cSPaolo Bonzini env->nmi_injected = false; 65664bef038SCameron Esfahani env->ins_len = 0; 65764bef038SCameron Esfahani env->has_error_code = false; 65869e0a03cSPaolo Bonzini if (idtvec_info & VMCS_IDT_VEC_VALID) { 65969e0a03cSPaolo Bonzini switch (idtvec_info & VMCS_IDT_VEC_TYPE) { 66069e0a03cSPaolo Bonzini case VMCS_IDT_VEC_HWINTR: 66169e0a03cSPaolo Bonzini case VMCS_IDT_VEC_SWINTR: 66269e0a03cSPaolo Bonzini env->interrupt_injected = idtvec_info & VMCS_IDT_VEC_VECNUM; 66369e0a03cSPaolo Bonzini break; 66469e0a03cSPaolo Bonzini case VMCS_IDT_VEC_NMI: 66569e0a03cSPaolo Bonzini env->nmi_injected = true; 66669e0a03cSPaolo Bonzini break; 66769e0a03cSPaolo Bonzini case VMCS_IDT_VEC_HWEXCEPTION: 66869e0a03cSPaolo Bonzini case VMCS_IDT_VEC_SWEXCEPTION: 669fd13f23bSLiran Alon env->exception_nr = idtvec_info & VMCS_IDT_VEC_VECNUM; 670fd13f23bSLiran Alon env->exception_injected = 1; 67169e0a03cSPaolo Bonzini break; 67269e0a03cSPaolo Bonzini case VMCS_IDT_VEC_PRIV_SWEXCEPTION: 67369e0a03cSPaolo Bonzini default: 67469e0a03cSPaolo Bonzini abort(); 67569e0a03cSPaolo Bonzini } 67669e0a03cSPaolo Bonzini if ((idtvec_info & VMCS_IDT_VEC_TYPE) == VMCS_IDT_VEC_SWEXCEPTION || 67769e0a03cSPaolo Bonzini (idtvec_info & VMCS_IDT_VEC_TYPE) == VMCS_IDT_VEC_SWINTR) { 67869e0a03cSPaolo Bonzini env->ins_len = ins_len; 67969e0a03cSPaolo Bonzini } 68064bef038SCameron Esfahani if (idtvec_info & VMCS_IDT_VEC_ERRCODE_VALID) { 68169e0a03cSPaolo Bonzini env->has_error_code = true; 68269e0a03cSPaolo Bonzini env->error_code = rvmcs(cpu->hvf_fd, VMCS_IDT_VECTORING_ERROR); 68369e0a03cSPaolo Bonzini } 68469e0a03cSPaolo Bonzini } 68569e0a03cSPaolo Bonzini if ((rvmcs(cpu->hvf_fd, VMCS_GUEST_INTERRUPTIBILITY) & 68669e0a03cSPaolo Bonzini VMCS_INTERRUPTIBILITY_NMI_BLOCKING)) { 68769e0a03cSPaolo Bonzini env->hflags2 |= HF2_NMI_MASK; 68869e0a03cSPaolo Bonzini } else { 68969e0a03cSPaolo Bonzini env->hflags2 &= ~HF2_NMI_MASK; 69069e0a03cSPaolo Bonzini } 69169e0a03cSPaolo Bonzini if (rvmcs(cpu->hvf_fd, VMCS_GUEST_INTERRUPTIBILITY) & 69269e0a03cSPaolo Bonzini (VMCS_INTERRUPTIBILITY_STI_BLOCKING | 69369e0a03cSPaolo Bonzini VMCS_INTERRUPTIBILITY_MOVSS_BLOCKING)) { 69469e0a03cSPaolo Bonzini env->hflags |= HF_INHIBIT_IRQ_MASK; 69569e0a03cSPaolo Bonzini } else { 69669e0a03cSPaolo Bonzini env->hflags &= ~HF_INHIBIT_IRQ_MASK; 69769e0a03cSPaolo Bonzini } 69869e0a03cSPaolo Bonzini } 69969e0a03cSPaolo Bonzini 70069e0a03cSPaolo Bonzini int hvf_vcpu_exec(CPUState *cpu) 70169e0a03cSPaolo Bonzini { 70269e0a03cSPaolo Bonzini X86CPU *x86_cpu = X86_CPU(cpu); 70369e0a03cSPaolo Bonzini CPUX86State *env = &x86_cpu->env; 70469e0a03cSPaolo Bonzini int ret = 0; 70569e0a03cSPaolo Bonzini uint64_t rip = 0; 70669e0a03cSPaolo Bonzini 70769e0a03cSPaolo Bonzini if (hvf_process_events(cpu)) { 70869e0a03cSPaolo Bonzini return EXCP_HLT; 70969e0a03cSPaolo Bonzini } 71069e0a03cSPaolo Bonzini 71169e0a03cSPaolo Bonzini do { 71269e0a03cSPaolo Bonzini if (cpu->vcpu_dirty) { 71369e0a03cSPaolo Bonzini hvf_put_registers(cpu); 71469e0a03cSPaolo Bonzini cpu->vcpu_dirty = false; 71569e0a03cSPaolo Bonzini } 71669e0a03cSPaolo Bonzini 71769e0a03cSPaolo Bonzini if (hvf_inject_interrupts(cpu)) { 71869e0a03cSPaolo Bonzini return EXCP_INTERRUPT; 71969e0a03cSPaolo Bonzini } 72069e0a03cSPaolo Bonzini vmx_update_tpr(cpu); 72169e0a03cSPaolo Bonzini 72269e0a03cSPaolo Bonzini qemu_mutex_unlock_iothread(); 72369e0a03cSPaolo Bonzini if (!cpu_is_bsp(X86_CPU(cpu)) && cpu->halted) { 72469e0a03cSPaolo Bonzini qemu_mutex_lock_iothread(); 72569e0a03cSPaolo Bonzini return EXCP_HLT; 72669e0a03cSPaolo Bonzini } 72769e0a03cSPaolo Bonzini 72869e0a03cSPaolo Bonzini hv_return_t r = hv_vcpu_run(cpu->hvf_fd); 72969e0a03cSPaolo Bonzini assert_hvf_ok(r); 73069e0a03cSPaolo Bonzini 73169e0a03cSPaolo Bonzini /* handle VMEXIT */ 73269e0a03cSPaolo Bonzini uint64_t exit_reason = rvmcs(cpu->hvf_fd, VMCS_EXIT_REASON); 73369e0a03cSPaolo Bonzini uint64_t exit_qual = rvmcs(cpu->hvf_fd, VMCS_EXIT_QUALIFICATION); 73469e0a03cSPaolo Bonzini uint32_t ins_len = (uint32_t)rvmcs(cpu->hvf_fd, 73569e0a03cSPaolo Bonzini VMCS_EXIT_INSTRUCTION_LENGTH); 73669e0a03cSPaolo Bonzini 73769e0a03cSPaolo Bonzini uint64_t idtvec_info = rvmcs(cpu->hvf_fd, VMCS_IDT_VECTORING_INFO); 73869e0a03cSPaolo Bonzini 73969e0a03cSPaolo Bonzini hvf_store_events(cpu, ins_len, idtvec_info); 74069e0a03cSPaolo Bonzini rip = rreg(cpu->hvf_fd, HV_X86_RIP); 741967f4da2SRoman Bolshakov env->eflags = rreg(cpu->hvf_fd, HV_X86_RFLAGS); 74269e0a03cSPaolo Bonzini 74369e0a03cSPaolo Bonzini qemu_mutex_lock_iothread(); 74469e0a03cSPaolo Bonzini 74569e0a03cSPaolo Bonzini update_apic_tpr(cpu); 74669e0a03cSPaolo Bonzini current_cpu = cpu; 74769e0a03cSPaolo Bonzini 74869e0a03cSPaolo Bonzini ret = 0; 74969e0a03cSPaolo Bonzini switch (exit_reason) { 75069e0a03cSPaolo Bonzini case EXIT_REASON_HLT: { 75169e0a03cSPaolo Bonzini macvm_set_rip(cpu, rip + ins_len); 75269e0a03cSPaolo Bonzini if (!((cpu->interrupt_request & CPU_INTERRUPT_HARD) && 753967f4da2SRoman Bolshakov (env->eflags & IF_MASK)) 75469e0a03cSPaolo Bonzini && !(cpu->interrupt_request & CPU_INTERRUPT_NMI) && 75569e0a03cSPaolo Bonzini !(idtvec_info & VMCS_IDT_VEC_VALID)) { 75669e0a03cSPaolo Bonzini cpu->halted = 1; 75769e0a03cSPaolo Bonzini ret = EXCP_HLT; 7583b9c59daSChen Zhang break; 75969e0a03cSPaolo Bonzini } 76069e0a03cSPaolo Bonzini ret = EXCP_INTERRUPT; 76169e0a03cSPaolo Bonzini break; 76269e0a03cSPaolo Bonzini } 76369e0a03cSPaolo Bonzini case EXIT_REASON_MWAIT: { 76469e0a03cSPaolo Bonzini ret = EXCP_INTERRUPT; 76569e0a03cSPaolo Bonzini break; 76669e0a03cSPaolo Bonzini } 767fbafbb6dSCameron Esfahani /* Need to check if MMIO or unmapped fault */ 76869e0a03cSPaolo Bonzini case EXIT_REASON_EPT_FAULT: 76969e0a03cSPaolo Bonzini { 77069e0a03cSPaolo Bonzini hvf_slot *slot; 771ff2de166SPaolo Bonzini uint64_t gpa = rvmcs(cpu->hvf_fd, VMCS_GUEST_PHYSICAL_ADDRESS); 77269e0a03cSPaolo Bonzini 77369e0a03cSPaolo Bonzini if (((idtvec_info & VMCS_IDT_VEC_VALID) == 0) && 77469e0a03cSPaolo Bonzini ((exit_qual & EXIT_QUAL_NMIUDTI) != 0)) { 77569e0a03cSPaolo Bonzini vmx_set_nmi_blocking(cpu); 77669e0a03cSPaolo Bonzini } 77769e0a03cSPaolo Bonzini 778fbafbb6dSCameron Esfahani slot = hvf_find_overlap_slot(gpa, 1); 77969e0a03cSPaolo Bonzini /* mmio */ 78069e0a03cSPaolo Bonzini if (ept_emulation_fault(slot, gpa, exit_qual)) { 78169e0a03cSPaolo Bonzini struct x86_decode decode; 78269e0a03cSPaolo Bonzini 78369e0a03cSPaolo Bonzini load_regs(cpu); 78469e0a03cSPaolo Bonzini decode_instruction(env, &decode); 78569e0a03cSPaolo Bonzini exec_instruction(env, &decode); 78669e0a03cSPaolo Bonzini store_regs(cpu); 78769e0a03cSPaolo Bonzini break; 78869e0a03cSPaolo Bonzini } 78969e0a03cSPaolo Bonzini break; 79069e0a03cSPaolo Bonzini } 79169e0a03cSPaolo Bonzini case EXIT_REASON_INOUT: 79269e0a03cSPaolo Bonzini { 79369e0a03cSPaolo Bonzini uint32_t in = (exit_qual & 8) != 0; 79469e0a03cSPaolo Bonzini uint32_t size = (exit_qual & 7) + 1; 79569e0a03cSPaolo Bonzini uint32_t string = (exit_qual & 16) != 0; 79669e0a03cSPaolo Bonzini uint32_t port = exit_qual >> 16; 79769e0a03cSPaolo Bonzini /*uint32_t rep = (exit_qual & 0x20) != 0;*/ 79869e0a03cSPaolo Bonzini 79969e0a03cSPaolo Bonzini if (!string && in) { 80069e0a03cSPaolo Bonzini uint64_t val = 0; 80169e0a03cSPaolo Bonzini load_regs(cpu); 80269e0a03cSPaolo Bonzini hvf_handle_io(env, port, &val, 0, size, 1); 80369e0a03cSPaolo Bonzini if (size == 1) { 80469e0a03cSPaolo Bonzini AL(env) = val; 80569e0a03cSPaolo Bonzini } else if (size == 2) { 80669e0a03cSPaolo Bonzini AX(env) = val; 80769e0a03cSPaolo Bonzini } else if (size == 4) { 80869e0a03cSPaolo Bonzini RAX(env) = (uint32_t)val; 80969e0a03cSPaolo Bonzini } else { 810da20f5cdSPaolo Bonzini RAX(env) = (uint64_t)val; 81169e0a03cSPaolo Bonzini } 8125d32173fSRoman Bolshakov env->eip += ins_len; 81369e0a03cSPaolo Bonzini store_regs(cpu); 81469e0a03cSPaolo Bonzini break; 81569e0a03cSPaolo Bonzini } else if (!string && !in) { 81669e0a03cSPaolo Bonzini RAX(env) = rreg(cpu->hvf_fd, HV_X86_RAX); 81769e0a03cSPaolo Bonzini hvf_handle_io(env, port, &RAX(env), 1, size, 1); 81869e0a03cSPaolo Bonzini macvm_set_rip(cpu, rip + ins_len); 81969e0a03cSPaolo Bonzini break; 82069e0a03cSPaolo Bonzini } 82169e0a03cSPaolo Bonzini struct x86_decode decode; 82269e0a03cSPaolo Bonzini 82369e0a03cSPaolo Bonzini load_regs(cpu); 82469e0a03cSPaolo Bonzini decode_instruction(env, &decode); 825e62963bfSPaolo Bonzini assert(ins_len == decode.len); 82669e0a03cSPaolo Bonzini exec_instruction(env, &decode); 82769e0a03cSPaolo Bonzini store_regs(cpu); 82869e0a03cSPaolo Bonzini 82969e0a03cSPaolo Bonzini break; 83069e0a03cSPaolo Bonzini } 83169e0a03cSPaolo Bonzini case EXIT_REASON_CPUID: { 83269e0a03cSPaolo Bonzini uint32_t rax = (uint32_t)rreg(cpu->hvf_fd, HV_X86_RAX); 83369e0a03cSPaolo Bonzini uint32_t rbx = (uint32_t)rreg(cpu->hvf_fd, HV_X86_RBX); 83469e0a03cSPaolo Bonzini uint32_t rcx = (uint32_t)rreg(cpu->hvf_fd, HV_X86_RCX); 83569e0a03cSPaolo Bonzini uint32_t rdx = (uint32_t)rreg(cpu->hvf_fd, HV_X86_RDX); 83669e0a03cSPaolo Bonzini 83769e0a03cSPaolo Bonzini cpu_x86_cpuid(env, rax, rcx, &rax, &rbx, &rcx, &rdx); 83869e0a03cSPaolo Bonzini 83969e0a03cSPaolo Bonzini wreg(cpu->hvf_fd, HV_X86_RAX, rax); 84069e0a03cSPaolo Bonzini wreg(cpu->hvf_fd, HV_X86_RBX, rbx); 84169e0a03cSPaolo Bonzini wreg(cpu->hvf_fd, HV_X86_RCX, rcx); 84269e0a03cSPaolo Bonzini wreg(cpu->hvf_fd, HV_X86_RDX, rdx); 84369e0a03cSPaolo Bonzini 84469e0a03cSPaolo Bonzini macvm_set_rip(cpu, rip + ins_len); 84569e0a03cSPaolo Bonzini break; 84669e0a03cSPaolo Bonzini } 84769e0a03cSPaolo Bonzini case EXIT_REASON_XSETBV: { 84869e0a03cSPaolo Bonzini X86CPU *x86_cpu = X86_CPU(cpu); 84969e0a03cSPaolo Bonzini CPUX86State *env = &x86_cpu->env; 85069e0a03cSPaolo Bonzini uint32_t eax = (uint32_t)rreg(cpu->hvf_fd, HV_X86_RAX); 85169e0a03cSPaolo Bonzini uint32_t ecx = (uint32_t)rreg(cpu->hvf_fd, HV_X86_RCX); 85269e0a03cSPaolo Bonzini uint32_t edx = (uint32_t)rreg(cpu->hvf_fd, HV_X86_RDX); 85369e0a03cSPaolo Bonzini 85469e0a03cSPaolo Bonzini if (ecx) { 85569e0a03cSPaolo Bonzini macvm_set_rip(cpu, rip + ins_len); 85669e0a03cSPaolo Bonzini break; 85769e0a03cSPaolo Bonzini } 85869e0a03cSPaolo Bonzini env->xcr0 = ((uint64_t)edx << 32) | eax; 85969e0a03cSPaolo Bonzini wreg(cpu->hvf_fd, HV_X86_XCR0, env->xcr0 | 1); 86069e0a03cSPaolo Bonzini macvm_set_rip(cpu, rip + ins_len); 86169e0a03cSPaolo Bonzini break; 86269e0a03cSPaolo Bonzini } 86369e0a03cSPaolo Bonzini case EXIT_REASON_INTR_WINDOW: 86469e0a03cSPaolo Bonzini vmx_clear_int_window_exiting(cpu); 86569e0a03cSPaolo Bonzini ret = EXCP_INTERRUPT; 86669e0a03cSPaolo Bonzini break; 86769e0a03cSPaolo Bonzini case EXIT_REASON_NMI_WINDOW: 86869e0a03cSPaolo Bonzini vmx_clear_nmi_window_exiting(cpu); 86969e0a03cSPaolo Bonzini ret = EXCP_INTERRUPT; 87069e0a03cSPaolo Bonzini break; 87169e0a03cSPaolo Bonzini case EXIT_REASON_EXT_INTR: 87269e0a03cSPaolo Bonzini /* force exit and allow io handling */ 87369e0a03cSPaolo Bonzini ret = EXCP_INTERRUPT; 87469e0a03cSPaolo Bonzini break; 87569e0a03cSPaolo Bonzini case EXIT_REASON_RDMSR: 87669e0a03cSPaolo Bonzini case EXIT_REASON_WRMSR: 87769e0a03cSPaolo Bonzini { 87869e0a03cSPaolo Bonzini load_regs(cpu); 87969e0a03cSPaolo Bonzini if (exit_reason == EXIT_REASON_RDMSR) { 88069e0a03cSPaolo Bonzini simulate_rdmsr(cpu); 88169e0a03cSPaolo Bonzini } else { 88269e0a03cSPaolo Bonzini simulate_wrmsr(cpu); 88369e0a03cSPaolo Bonzini } 8845d32173fSRoman Bolshakov env->eip += ins_len; 88569e0a03cSPaolo Bonzini store_regs(cpu); 88669e0a03cSPaolo Bonzini break; 88769e0a03cSPaolo Bonzini } 88869e0a03cSPaolo Bonzini case EXIT_REASON_CR_ACCESS: { 88969e0a03cSPaolo Bonzini int cr; 89069e0a03cSPaolo Bonzini int reg; 89169e0a03cSPaolo Bonzini 89269e0a03cSPaolo Bonzini load_regs(cpu); 89369e0a03cSPaolo Bonzini cr = exit_qual & 15; 89469e0a03cSPaolo Bonzini reg = (exit_qual >> 8) & 15; 89569e0a03cSPaolo Bonzini 89669e0a03cSPaolo Bonzini switch (cr) { 89769e0a03cSPaolo Bonzini case 0x0: { 89869e0a03cSPaolo Bonzini macvm_set_cr0(cpu->hvf_fd, RRX(env, reg)); 89969e0a03cSPaolo Bonzini break; 90069e0a03cSPaolo Bonzini } 90169e0a03cSPaolo Bonzini case 4: { 90269e0a03cSPaolo Bonzini macvm_set_cr4(cpu->hvf_fd, RRX(env, reg)); 90369e0a03cSPaolo Bonzini break; 90469e0a03cSPaolo Bonzini } 90569e0a03cSPaolo Bonzini case 8: { 90669e0a03cSPaolo Bonzini X86CPU *x86_cpu = X86_CPU(cpu); 90769e0a03cSPaolo Bonzini if (exit_qual & 0x10) { 90869e0a03cSPaolo Bonzini RRX(env, reg) = cpu_get_apic_tpr(x86_cpu->apic_state); 90969e0a03cSPaolo Bonzini } else { 91069e0a03cSPaolo Bonzini int tpr = RRX(env, reg); 91169e0a03cSPaolo Bonzini cpu_set_apic_tpr(x86_cpu->apic_state, tpr); 91269e0a03cSPaolo Bonzini ret = EXCP_INTERRUPT; 91369e0a03cSPaolo Bonzini } 91469e0a03cSPaolo Bonzini break; 91569e0a03cSPaolo Bonzini } 91669e0a03cSPaolo Bonzini default: 9172d9178d9SLaurent Vivier error_report("Unrecognized CR %d", cr); 91869e0a03cSPaolo Bonzini abort(); 91969e0a03cSPaolo Bonzini } 9205d32173fSRoman Bolshakov env->eip += ins_len; 92169e0a03cSPaolo Bonzini store_regs(cpu); 92269e0a03cSPaolo Bonzini break; 92369e0a03cSPaolo Bonzini } 92469e0a03cSPaolo Bonzini case EXIT_REASON_APIC_ACCESS: { /* TODO */ 92569e0a03cSPaolo Bonzini struct x86_decode decode; 92669e0a03cSPaolo Bonzini 92769e0a03cSPaolo Bonzini load_regs(cpu); 92869e0a03cSPaolo Bonzini decode_instruction(env, &decode); 92969e0a03cSPaolo Bonzini exec_instruction(env, &decode); 93069e0a03cSPaolo Bonzini store_regs(cpu); 93169e0a03cSPaolo Bonzini break; 93269e0a03cSPaolo Bonzini } 93369e0a03cSPaolo Bonzini case EXIT_REASON_TPR: { 93469e0a03cSPaolo Bonzini ret = 1; 93569e0a03cSPaolo Bonzini break; 93669e0a03cSPaolo Bonzini } 93769e0a03cSPaolo Bonzini case EXIT_REASON_TASK_SWITCH: { 93869e0a03cSPaolo Bonzini uint64_t vinfo = rvmcs(cpu->hvf_fd, VMCS_IDT_VECTORING_INFO); 93969e0a03cSPaolo Bonzini x68_segment_selector sel = {.sel = exit_qual & 0xffff}; 94069e0a03cSPaolo Bonzini vmx_handle_task_switch(cpu, sel, (exit_qual >> 30) & 0x3, 94169e0a03cSPaolo Bonzini vinfo & VMCS_INTR_VALID, vinfo & VECTORING_INFO_VECTOR_MASK, vinfo 94269e0a03cSPaolo Bonzini & VMCS_INTR_T_MASK); 94369e0a03cSPaolo Bonzini break; 94469e0a03cSPaolo Bonzini } 94569e0a03cSPaolo Bonzini case EXIT_REASON_TRIPLE_FAULT: { 94669e0a03cSPaolo Bonzini qemu_system_reset_request(SHUTDOWN_CAUSE_GUEST_RESET); 94769e0a03cSPaolo Bonzini ret = EXCP_INTERRUPT; 94869e0a03cSPaolo Bonzini break; 94969e0a03cSPaolo Bonzini } 95069e0a03cSPaolo Bonzini case EXIT_REASON_RDPMC: 95169e0a03cSPaolo Bonzini wreg(cpu->hvf_fd, HV_X86_RAX, 0); 95269e0a03cSPaolo Bonzini wreg(cpu->hvf_fd, HV_X86_RDX, 0); 95369e0a03cSPaolo Bonzini macvm_set_rip(cpu, rip + ins_len); 95469e0a03cSPaolo Bonzini break; 95569e0a03cSPaolo Bonzini case VMX_REASON_VMCALL: 956fd13f23bSLiran Alon env->exception_nr = EXCP0D_GPF; 957fd13f23bSLiran Alon env->exception_injected = 1; 95869e0a03cSPaolo Bonzini env->has_error_code = true; 95969e0a03cSPaolo Bonzini env->error_code = 0; 96069e0a03cSPaolo Bonzini break; 96169e0a03cSPaolo Bonzini default: 9622d9178d9SLaurent Vivier error_report("%llx: unhandled exit %llx", rip, exit_reason); 96369e0a03cSPaolo Bonzini } 96469e0a03cSPaolo Bonzini } while (ret == 0); 96569e0a03cSPaolo Bonzini 96669e0a03cSPaolo Bonzini return ret; 96769e0a03cSPaolo Bonzini } 96869e0a03cSPaolo Bonzini 96992cc3aaaSRoman Bolshakov bool hvf_allowed; 97069e0a03cSPaolo Bonzini 97169e0a03cSPaolo Bonzini static int hvf_accel_init(MachineState *ms) 97269e0a03cSPaolo Bonzini { 97369e0a03cSPaolo Bonzini int x; 97469e0a03cSPaolo Bonzini hv_return_t ret; 97569e0a03cSPaolo Bonzini HVFState *s; 97669e0a03cSPaolo Bonzini 97769e0a03cSPaolo Bonzini ret = hv_vm_create(HV_VM_DEFAULT); 97869e0a03cSPaolo Bonzini assert_hvf_ok(ret); 97969e0a03cSPaolo Bonzini 98069e0a03cSPaolo Bonzini s = g_new0(HVFState, 1); 98169e0a03cSPaolo Bonzini 98269e0a03cSPaolo Bonzini s->num_slots = 32; 98369e0a03cSPaolo Bonzini for (x = 0; x < s->num_slots; ++x) { 98469e0a03cSPaolo Bonzini s->slots[x].size = 0; 98569e0a03cSPaolo Bonzini s->slots[x].slot_id = x; 98669e0a03cSPaolo Bonzini } 98769e0a03cSPaolo Bonzini 98869e0a03cSPaolo Bonzini hvf_state = s; 98969e0a03cSPaolo Bonzini cpu_interrupt_handler = hvf_handle_interrupt; 99069e0a03cSPaolo Bonzini memory_listener_register(&hvf_memory_listener, &address_space_memory); 99169e0a03cSPaolo Bonzini return 0; 99269e0a03cSPaolo Bonzini } 99369e0a03cSPaolo Bonzini 99469e0a03cSPaolo Bonzini static void hvf_accel_class_init(ObjectClass *oc, void *data) 99569e0a03cSPaolo Bonzini { 99669e0a03cSPaolo Bonzini AccelClass *ac = ACCEL_CLASS(oc); 99769e0a03cSPaolo Bonzini ac->name = "HVF"; 99869e0a03cSPaolo Bonzini ac->init_machine = hvf_accel_init; 99969e0a03cSPaolo Bonzini ac->allowed = &hvf_allowed; 100069e0a03cSPaolo Bonzini } 100169e0a03cSPaolo Bonzini 100269e0a03cSPaolo Bonzini static const TypeInfo hvf_accel_type = { 100369e0a03cSPaolo Bonzini .name = TYPE_HVF_ACCEL, 100469e0a03cSPaolo Bonzini .parent = TYPE_ACCEL, 100569e0a03cSPaolo Bonzini .class_init = hvf_accel_class_init, 100669e0a03cSPaolo Bonzini }; 100769e0a03cSPaolo Bonzini 100869e0a03cSPaolo Bonzini static void hvf_type_init(void) 100969e0a03cSPaolo Bonzini { 101069e0a03cSPaolo Bonzini type_register_static(&hvf_accel_type); 101169e0a03cSPaolo Bonzini } 101269e0a03cSPaolo Bonzini 101369e0a03cSPaolo Bonzini type_init(hvf_type_init); 1014