Lines Matching refs:arch

35 	for (i = 0; i < vcpu->arch.slb_nr; i++) {  in kvmppc_mmu_book3s_64_find_slbe()
38 if (!vcpu->arch.slb[i].valid) in kvmppc_mmu_book3s_64_find_slbe()
41 if (vcpu->arch.slb[i].tb) in kvmppc_mmu_book3s_64_find_slbe()
44 if (vcpu->arch.slb[i].esid == cmp_esid) in kvmppc_mmu_book3s_64_find_slbe()
45 return &vcpu->arch.slb[i]; in kvmppc_mmu_book3s_64_find_slbe()
50 for (i = 0; i < vcpu->arch.slb_nr; i++) { in kvmppc_mmu_book3s_64_find_slbe()
51 if (vcpu->arch.slb[i].vsid) in kvmppc_mmu_book3s_64_find_slbe()
53 vcpu->arch.slb[i].valid ? 'v' : ' ', in kvmppc_mmu_book3s_64_find_slbe()
54 vcpu->arch.slb[i].large ? 'l' : ' ', in kvmppc_mmu_book3s_64_find_slbe()
55 vcpu->arch.slb[i].tb ? 't' : ' ', in kvmppc_mmu_book3s_64_find_slbe()
56 vcpu->arch.slb[i].esid, in kvmppc_mmu_book3s_64_find_slbe()
57 vcpu->arch.slb[i].vsid); in kvmppc_mmu_book3s_64_find_slbe()
145 if (vcpu->arch.papr_enabled) in kvmppc_mmu_book3s_64_get_pteg()
207 ulong mp_ea = vcpu->arch.magic_page_ea; in kvmppc_mmu_book3s_64_xlate()
215 gpte->raddr = vcpu->arch.magic_page_pa | (gpte->raddr & 0xfff); in kvmppc_mmu_book3s_64_xlate()
244 mutex_lock(&vcpu->kvm->arch.hpt_mutex); in kvmppc_mmu_book3s_64_xlate()
270 (vcpu->arch.hflags & BOOK3S_HFLAG_MULTI_PGSIZE)) { in kvmppc_mmu_book3s_64_xlate()
300 if (unlikely(vcpu->arch.disable_kernel_nx) && in kvmppc_mmu_book3s_64_xlate()
348 mutex_unlock(&vcpu->kvm->arch.hpt_mutex); in kvmppc_mmu_book3s_64_xlate()
355 mutex_unlock(&vcpu->kvm->arch.hpt_mutex); in kvmppc_mmu_book3s_64_xlate()
375 if (slb_nr > vcpu->arch.slb_nr) in kvmppc_mmu_book3s_64_slbmte()
378 slbe = &vcpu->arch.slb[slb_nr]; in kvmppc_mmu_book3s_64_slbmte()
392 if (vcpu->arch.hflags & BOOK3S_HFLAG_MULTI_PGSIZE) { in kvmppc_mmu_book3s_64_slbmte()
429 if (slb_nr > vcpu->arch.slb_nr) in kvmppc_mmu_book3s_64_slbmfee()
432 slbe = &vcpu->arch.slb[slb_nr]; in kvmppc_mmu_book3s_64_slbmfee()
441 if (slb_nr > vcpu->arch.slb_nr) in kvmppc_mmu_book3s_64_slbmfev()
444 slbe = &vcpu->arch.slb[slb_nr]; in kvmppc_mmu_book3s_64_slbmfev()
477 for (i = 1; i < vcpu->arch.slb_nr; i++) { in kvmppc_mmu_book3s_64_slbia()
478 vcpu->arch.slb[i].valid = false; in kvmppc_mmu_book3s_64_slbia()
479 vcpu->arch.slb[i].orige = 0; in kvmppc_mmu_book3s_64_slbia()
480 vcpu->arch.slb[i].origv = 0; in kvmppc_mmu_book3s_64_slbia()
544 if (vcpu->arch.hflags & BOOK3S_HFLAG_NEW_TLBIE) { in kvmppc_mmu_book3s_64_tlbie()
565 ulong mp_ea = vcpu->arch.magic_page_ea; in segment_contains_magic_page()
578 ulong mp_ea = vcpu->arch.magic_page_ea; in kvmppc_mmu_book3s_64_esid_to_vsid()
653 struct kvmppc_mmu *mmu = &vcpu->arch.mmu; in kvmppc_mmu_book3s_64_init()
669 vcpu->arch.hflags |= BOOK3S_HFLAG_SLB; in kvmppc_mmu_book3s_64_init()