Home
last modified time | relevance | path

Searched refs:kern_hyp_va (Results 1 – 19 of 19) sorted by relevance

/openbmc/linux/arch/arm64/kvm/hyp/nvhe/
H A Dhyp-main.c79 host_vcpu = kern_hyp_va(host_vcpu); in handle___kvm_vcpu_run()
85 host_kvm = kern_hyp_va(host_vcpu->kvm); in handle___kvm_vcpu_run()
112 __kvm_adjust_pc(kern_hyp_va(vcpu)); in handle___kvm_adjust_pc()
126 __kvm_tlb_flush_vmid_ipa(kern_hyp_va(mmu), ipa, level); in handle___kvm_tlb_flush_vmid_ipa()
152 __kvm_tlb_flush_vmid(kern_hyp_va(mmu)); in handle___kvm_tlb_flush_vmid()
159 __kvm_flush_cpu_context(kern_hyp_va(mmu)); in handle___kvm_flush_cpu_context()
205 __vgic_v3_save_aprs(kern_hyp_va(cpu_if)); in handle___vgic_v3_save_aprs()
212 __vgic_v3_restore_aprs(kern_hyp_va(cpu_if)); in handle___vgic_v3_restore_aprs()
287 __pkvm_vcpu_init_traps(kern_hyp_va(vcpu)); in handle___pkvm_vcpu_init_traps()
296 host_kvm = kern_hyp_va(host_kvm); in handle___pkvm_init_vm()
[all …]
H A Dswitch.c210 if (unlikely(kvm_vm_is_protected(kern_hyp_va(vcpu->kvm)))) in kvm_get_exit_handler_array()
229 struct kvm *kvm = kern_hyp_va(vcpu->kvm); in early_exit_filter()
305 mmu = kern_hyp_va(vcpu->arch.hw_mmu); in __kvm_vcpu_run()
306 __load_stage2(mmu, kern_hyp_va(mmu->arch)); in __kvm_vcpu_run()
H A Dhyp-smp.c37 this_cpu_base = kern_hyp_va(cpu_base_array[cpu]); in __hyp_per_cpu_offset()
H A Dtimer-sr.c51 !kern_hyp_va(vcpu->kvm)->arch.timer_data.poffset) in __timer_enable_traps()
H A Dtlb.c65 __load_stage2(mmu, kern_hyp_va(mmu->arch)); in __tlb_switch_to_guest()
H A Dsetup.c117 start = (void *)kern_hyp_va(per_cpu_base[i]); in recreate_hyp_mappings()
H A Dsys_regs.c99 const struct kvm *kvm = (const struct kvm *)kern_hyp_va(vcpu->kvm); in get_pvm_id_aa64pfr1()
H A Dpkvm.c405 void *va = (void *)kern_hyp_va(host_va); in map_donated_memory_noclear()
/openbmc/linux/arch/arm64/kvm/hyp/include/hyp/
H A Ddebug-sr.h141 guest_dbg = kern_hyp_va(vcpu->arch.debug_ptr); in __debug_switch_to_guest_common()
160 guest_dbg = kern_hyp_va(vcpu->arch.debug_ptr); in __debug_switch_to_host_common()
H A Dswitch.h492 val -= *kern_hyp_va(ctxt->offset.vm_offset); in kvm_hyp_handle_cntpct()
494 val -= *kern_hyp_va(ctxt->offset.vcpu_offset); in kvm_hyp_handle_cntpct()
H A Dsysreg-sr.h37 return kvm_has_mte(kern_hyp_va(vcpu->kvm)); in ctxt_has_mte()
/openbmc/linux/arch/arm64/include/asm/
H A Dkvm_mmu.h65 .macro kern_hyp_va reg
145 #define kern_hyp_va(v) ((typeof(v))(__kern_hyp_va((unsigned long)(v)))) macro
H A Dkvm_host.h743 #define vcpu_sve_pffr(vcpu) (kern_hyp_va((vcpu)->arch.sve_state) + \
/openbmc/linux/arch/arm64/kvm/hyp/
H A Dvgic-v2-cpuif-proxy.c39 struct kvm *kvm = kern_hyp_va(vcpu->kvm); in __vgic_v2_perform_cpuif_access()
H A Dexception.c134 if (kvm_has_mte(kern_hyp_va(vcpu->kvm))) in enter_exception64()
/openbmc/linux/arch/arm64/kvm/
H A Dfpsimd.c52 vcpu->arch.host_fpsimd_state = kern_hyp_va(fpsimd); in kvm_arch_vcpu_run_map_fp()
H A Dmmu.c578 unsigned long start = kern_hyp_va((unsigned long)from); in create_hyp_mappings()
579 unsigned long end = kern_hyp_va((unsigned long)to); in create_hyp_mappings()
1922 kern_hyp_va(PAGE_OFFSET), in kvm_mmu_init()
1923 kern_hyp_va((unsigned long)high_memory - 1)); in kvm_mmu_init()
1925 if (hyp_idmap_start >= kern_hyp_va(PAGE_OFFSET) && in kvm_mmu_init()
1926 hyp_idmap_start < kern_hyp_va((unsigned long)high_memory - 1) && in kvm_mmu_init()
H A Darm.c1706 base = kern_hyp_va(kvm_ksym_ref(__kvm_hyp_vector)); in kvm_init_vector_slots()
1709 base = kern_hyp_va(kvm_ksym_ref(__bp_harden_hyp_vecs)); in kvm_init_vector_slots()
2084 num_possible_cpus(), kern_hyp_va(per_cpu_base), in do_pkvm_init()
/openbmc/linux/Documentation/arch/arm64/
H A Dmemory.rst95 random) offset from the linear mapping. See the kern_hyp_va macro and