Lines Matching refs:arch

86 	vcpu->arch.mmu = &vcpu->arch.guest_mmu;  in nested_svm_init_mmu_context()
96 vcpu->arch.mmu->get_guest_pgd = nested_svm_get_tdp_cr3; in nested_svm_init_mmu_context()
97 vcpu->arch.mmu->get_pdptr = nested_svm_get_tdp_pdptr; in nested_svm_init_mmu_context()
98 vcpu->arch.mmu->inject_page_fault = nested_svm_inject_npf_exit; in nested_svm_init_mmu_context()
99 vcpu->arch.walk_mmu = &vcpu->arch.nested_mmu; in nested_svm_init_mmu_context()
104 vcpu->arch.mmu = &vcpu->arch.root_mmu; in nested_svm_uninit_mmu_context()
105 vcpu->arch.walk_mmu = &vcpu->arch.root_mmu; in nested_svm_uninit_mmu_context()
447 if (vcpu->arch.exception.injected) { in nested_save_pending_event_to_vmcb12()
448 nr = vcpu->arch.exception.vector; in nested_save_pending_event_to_vmcb12()
451 if (vcpu->arch.exception.has_error_code) { in nested_save_pending_event_to_vmcb12()
454 vcpu->arch.exception.error_code; in nested_save_pending_event_to_vmcb12()
457 } else if (vcpu->arch.nmi_injected) { in nested_save_pending_event_to_vmcb12()
460 } else if (vcpu->arch.interrupt.injected) { in nested_save_pending_event_to_vmcb12()
461 nr = vcpu->arch.interrupt.nr; in nested_save_pending_event_to_vmcb12()
464 if (vcpu->arch.interrupt.soft) in nested_save_pending_event_to_vmcb12()
515 vcpu->arch.cr3 = cr3; in nested_svm_load_cr3()
573 svm->vcpu.arch.cr2 = vmcb12->save.cr2; in nested_vmcb02_prepare_save()
587 svm->vcpu.arch.dr6 = svm->nested.save.dr6 | DR6_ACTIVE_LOW; in nested_vmcb02_prepare_save()
656 svm->vcpu.arch.nmi_pending++; in nested_vmcb02_prepare_control()
679 vcpu->arch.tsc_offset = kvm_calc_nested_tsc_offset( in nested_vmcb02_prepare_control()
680 vcpu->arch.l1_tsc_offset, in nested_vmcb02_prepare_control()
684 vmcb02->control.tsc_offset = vcpu->arch.tsc_offset; in nested_vmcb02_prepare_control()
893 vmcb01->save.efer = vcpu->arch.efer; in nested_svm_vmrun()
895 vmcb01->save.cr4 = vcpu->arch.cr4; in nested_svm_vmrun()
991 svm->vcpu.arch.mp_state = KVM_MP_STATE_RUNNABLE; in nested_svm_vmexit()
1001 vmcb12->save.efer = svm->vcpu.arch.efer; in nested_svm_vmexit()
1005 vmcb12->save.cr4 = svm->vcpu.arch.cr4; in nested_svm_vmexit()
1011 vmcb12->save.dr6 = svm->vcpu.arch.dr6; in nested_svm_vmexit()
1077 if (vcpu->arch.nmi_pending) { in nested_svm_vmexit()
1078 vcpu->arch.nmi_pending--; in nested_svm_vmexit()
1092 svm->vcpu.arch.tsc_offset = svm->vcpu.arch.l1_tsc_offset; in nested_svm_vmexit()
1093 if (vmcb01->control.tsc_offset != svm->vcpu.arch.tsc_offset) { in nested_svm_vmexit()
1094 vmcb01->control.tsc_offset = svm->vcpu.arch.tsc_offset; in nested_svm_vmexit()
1099 vcpu->arch.tsc_scaling_ratio != vcpu->arch.l1_tsc_scaling_ratio) { in nested_svm_vmexit()
1100 vcpu->arch.tsc_scaling_ratio = vcpu->arch.l1_tsc_scaling_ratio; in nested_svm_vmexit()
1117 svm->vcpu.arch.dr7 = DR7_FIXED_1; in nested_svm_vmexit()
1141 svm->vcpu.arch.nmi_injected = false; in nested_svm_vmexit()
1257 msr = svm->vcpu.arch.regs[VCPU_REGS_RCX]; in nested_svm_exit_handled_msr()
1357 if (!(vcpu->arch.efer & EFER_SVME) || !is_paging(vcpu)) { in nested_svm_check_permissions()
1380 struct kvm_queued_exception *ex = &vcpu->arch.exception_vmexit; in nested_svm_inject_exception_vmexit()
1398 vmcb->control.exit_info_2 = vcpu->arch.cr2; in nested_svm_inject_exception_vmexit()
1403 if (vcpu->arch.dr7 & DR7_GD) { in nested_svm_inject_exception_vmexit()
1404 vcpu->arch.dr7 &= ~DR7_GD; in nested_svm_inject_exception_vmexit()
1421 struct kvm_lapic *apic = vcpu->arch.apic; in svm_check_nested_events()
1448 if (vcpu->arch.exception_vmexit.pending) { in svm_check_nested_events()
1455 if (vcpu->arch.exception.pending) { in svm_check_nested_events()
1462 if (vcpu->arch.smi_pending && !svm_smi_blocked(vcpu)) { in svm_check_nested_events()
1472 if (vcpu->arch.nmi_pending && !svm_nmi_blocked(vcpu)) { in svm_check_nested_events()
1511 svm->vcpu.arch.apf.host_apf_flags) in nested_svm_exit_special()
1534 vcpu->arch.tsc_scaling_ratio = in nested_svm_update_tsc_ratio_msr()
1535 kvm_calc_nested_tsc_multiplier(vcpu->arch.l1_tsc_scaling_ratio, in nested_svm_update_tsc_ratio_msr()
1672 if (!(vcpu->arch.efer & EFER_SVME)) { in svm_set_nested_state()
1762 ret = nested_svm_load_cr3(&svm->vcpu, vcpu->arch.cr3, in svm_set_nested_state()
1785 if (!vcpu->arch.pdptrs_from_userspace && in svm_get_nested_state_pages()
1792 if (CC(!load_pdptrs(vcpu, vcpu->arch.cr3))) in svm_get_nested_state_pages()