Searched refs:HF_SMM_MASK (Results 1 – 10 of 10) sorted by relevance
151 return vcpu->arch.hflags & HF_SMM_MASK; in is_smm()
117 vcpu->arch.hflags |= HF_SMM_MASK; in kvm_smm_changed()119 vcpu->arch.hflags &= ~(HF_SMM_MASK | HF_SMM_INSIDE_NMI_MASK); in kvm_smm_changed()
5317 if (!!(vcpu->arch.hflags & HF_SMM_MASK) != events->smi.smm) { in kvm_vcpu_ioctl_x86_set_vcpu_events()
47 env->hflags |= HF_SMM_MASK; in do_smm_enter()315 env->hflags &= ~HF_SMM_MASK; in helper_rsm()
192 #define HF_SMM_MASK (1 << HF_SMM_SHIFT) macro2570 return ((MemTxAttrs) { .secure = (env->hflags & HF_SMM_MASK) != 0 }); in cpu_get_mem_attrs()2575 if (env->hflags & HF_SMM_MASK) { in x86_get_a20_mask()2611 ((env->cr[4] & CR4_VMXE_MASK) || (env->hflags & HF_SMM_MASK)); in cpu_vmx_maybe_enabled()
8288 !(env->hflags & HF_SMM_MASK)) { in x86_cpu_pending_interrupt()
2125 #define HF_SMM_MASK (1 << 1) macro2130 # define kvm_arch_vcpu_memslots_id(vcpu) ((vcpu)->arch.hflags & HF_SMM_MASK ? 1 : 0)
5038 events.smi.smm = !!(env->hflags & HF_SMM_MASK); in kvm_put_vcpu_events()5110 env->hflags |= HF_SMM_MASK; in kvm_get_vcpu_events()5112 env->hflags &= ~HF_SMM_MASK; in kvm_get_vcpu_events()5468 !(env->hflags & HF_SMM_MASK)) { in kvm_arch_pre_run()5531 env->hflags |= HF_SMM_MASK; in kvm_arch_post_run()5533 env->hflags &= ~HF_SMM_MASK; in kvm_arch_post_run()5599 !(env->hflags & HF_SMM_MASK)) { in kvm_arch_process_async_events()
1493 !(env->hflags & HF_SMM_MASK)) { in whpx_vcpu_pre_run()1609 !(env->hflags & HF_SMM_MASK)) { in whpx_vcpu_process_async_events()
2809 if ((decode.e.check & X86_CHECK_smm) && !(s->flags & HF_SMM_MASK)) {