Searched refs:HF_SMM_MASK (Results 1 – 10 of 10) sorted by relevance
151 return vcpu->arch.hflags & HF_SMM_MASK; in is_smm()
117 vcpu->arch.hflags |= HF_SMM_MASK; in kvm_smm_changed()119 vcpu->arch.hflags &= ~(HF_SMM_MASK | HF_SMM_INSIDE_NMI_MASK); in kvm_smm_changed()
5317 if (!!(vcpu->arch.hflags & HF_SMM_MASK) != events->smi.smm) { in kvm_vcpu_ioctl_x86_set_vcpu_events()
47 env->hflags |= HF_SMM_MASK; in do_smm_enter()315 env->hflags &= ~HF_SMM_MASK; in helper_rsm()
193 #define HF_SMM_MASK (1 << HF_SMM_SHIFT) macro2366 return ((MemTxAttrs) { .secure = (env->hflags & HF_SMM_MASK) != 0 }); in cpu_get_mem_attrs()2371 if (env->hflags & HF_SMM_MASK) { in x86_get_a20_mask()2407 ((env->cr[4] & CR4_VMXE_MASK) || (env->hflags & HF_SMM_MASK)); in cpu_vmx_maybe_enabled()
7693 !(env->hflags & HF_SMM_MASK)) { in x86_cpu_pending_interrupt()
4337 events.smi.smm = !!(env->hflags & HF_SMM_MASK); in kvm_put_vcpu_events()4415 env->hflags |= HF_SMM_MASK; in kvm_get_vcpu_events()4417 env->hflags &= ~HF_SMM_MASK; in kvm_get_vcpu_events()4750 !(env->hflags & HF_SMM_MASK)) { in kvm_arch_pre_run()4813 env->hflags |= HF_SMM_MASK; in kvm_arch_post_run()4815 env->hflags &= ~HF_SMM_MASK; in kvm_arch_post_run()4881 !(env->hflags & HF_SMM_MASK)) { in kvm_arch_process_async_events()
2123 #define HF_SMM_MASK (1 << 1) macro2128 # define kvm_arch_vcpu_memslots_id(vcpu) ((vcpu)->arch.hflags & HF_SMM_MASK ? 1 : 0)
1498 !(env->hflags & HF_SMM_MASK)) { in whpx_vcpu_pre_run()1614 !(env->hflags & HF_SMM_MASK)) { in whpx_vcpu_process_async_events()
6755 if (!(s->flags & HF_SMM_MASK)) in disas_insn()