/openbmc/linux/drivers/staging/media/ipu3/ |
H A D | ipu3-mmu.c | 78 static void imgu_mmu_tlb_invalidate(struct imgu_mmu *mmu) in imgu_mmu_tlb_invalidate() argument 80 writel(TLB_INVALIDATE, mmu->base + REG_TLB_INVALIDATE); in imgu_mmu_tlb_invalidate() 83 static void call_if_imgu_is_powered(struct imgu_mmu *mmu, in call_if_imgu_is_powered() argument 84 void (*func)(struct imgu_mmu *mmu)) in call_if_imgu_is_powered() argument 86 if (!pm_runtime_get_if_in_use(mmu->dev)) in call_if_imgu_is_powered() 89 func(mmu); in call_if_imgu_is_powered() 90 pm_runtime_put(mmu->dev); in call_if_imgu_is_powered() 101 static void imgu_mmu_set_halt(struct imgu_mmu *mmu, bool halt) in imgu_mmu_set_halt() argument 106 writel(halt, mmu->base + REG_GP_HALT); in imgu_mmu_set_halt() 107 ret = readl_poll_timeout(mmu->base + REG_GP_HALTED, in imgu_mmu_set_halt() [all …]
|
/openbmc/linux/drivers/gpu/drm/nouveau/nvkm/subdev/mmu/ |
H A D | base.c | 42 nvkm_mmu_ptp_put(struct nvkm_mmu *mmu, bool force, struct nvkm_mmu_pt *pt) in nvkm_mmu_ptp_put() argument 51 list_add(&ptp->head, &mmu->ptp.list); in nvkm_mmu_ptp_put() 56 nvkm_mmu_ptc_put(mmu, force, &ptp->pt); in nvkm_mmu_ptp_put() 65 nvkm_mmu_ptp_get(struct nvkm_mmu *mmu, u32 size, bool zero) in nvkm_mmu_ptp_get() argument 74 ptp = list_first_entry_or_null(&mmu->ptp.list, typeof(*ptp), head); in nvkm_mmu_ptp_get() 82 ptp->pt = nvkm_mmu_ptc_get(mmu, 0x1000, 0x1000, false); in nvkm_mmu_ptp_get() 93 list_add(&ptp->head, &mmu->ptp.list); in nvkm_mmu_ptp_get() 120 nvkm_mmu_ptc_find(struct nvkm_mmu *mmu, u32 size) in nvkm_mmu_ptc_find() argument 124 list_for_each_entry(ptc, &mmu->ptc.list, head) { in nvkm_mmu_ptc_find() 134 list_add(&ptc->head, &mmu->ptc.list); in nvkm_mmu_ptc_find() [all …]
|
H A D | Kbuild | 2 nvkm-y += nvkm/subdev/mmu/base.o 3 nvkm-y += nvkm/subdev/mmu/nv04.o 4 nvkm-y += nvkm/subdev/mmu/nv41.o 5 nvkm-y += nvkm/subdev/mmu/nv44.o 6 nvkm-y += nvkm/subdev/mmu/nv50.o 7 nvkm-y += nvkm/subdev/mmu/g84.o 8 nvkm-y += nvkm/subdev/mmu/mcp77.o 9 nvkm-y += nvkm/subdev/mmu/gf100.o 10 nvkm-y += nvkm/subdev/mmu/gk104.o 11 nvkm-y += nvkm/subdev/mmu/gk20a.o [all …]
|
H A D | ummu.c | 35 struct nvkm_mmu *mmu = nvkm_ummu(object)->mmu; in nvkm_ummu_sclass() local 37 if (mmu->func->mem.user.oclass) { in nvkm_ummu_sclass() 39 oclass->base = mmu->func->mem.user; in nvkm_ummu_sclass() 45 if (mmu->func->vmm.user.oclass) { in nvkm_ummu_sclass() 47 oclass->base = mmu->func->vmm.user; in nvkm_ummu_sclass() 59 struct nvkm_mmu *mmu = ummu->mmu; in nvkm_ummu_heap() local 67 if ((index = args->v0.index) >= mmu->heap_nr) in nvkm_ummu_heap() 69 args->v0.size = mmu->heap[index].size; in nvkm_ummu_heap() 79 struct nvkm_mmu *mmu = ummu->mmu; in nvkm_ummu_type() local 87 if ((index = args->v0.index) >= mmu->type_nr) in nvkm_ummu_type() [all …]
|
H A D | umem.c | 72 struct nvkm_device *device = umem->mmu->subdev.device; in nvkm_umem_unmap() 90 struct nvkm_mmu *mmu = umem->mmu; in nvkm_umem_map() local 109 int ret = mmu->func->mem.umap(mmu, umem->memory, argv, argc, in nvkm_umem_map() 145 struct nvkm_mmu *mmu = nvkm_ummu(oclass->parent)->mmu; in nvkm_umem_new() local 161 if (type >= mmu->type_nr) in nvkm_umem_new() 167 umem->mmu = mmu; in nvkm_umem_new() 168 umem->type = mmu->type[type].type; in nvkm_umem_new() 172 if (mmu->type[type].type & NVKM_MEM_MAPPABLE) { in nvkm_umem_new() 177 ret = nvkm_mem_new_type(mmu, type, page, size, argv, argc, in nvkm_umem_new()
|
/openbmc/linux/drivers/gpu/drm/nouveau/nvif/ |
H A D | mmu.c | 28 nvif_mmu_dtor(struct nvif_mmu *mmu) in nvif_mmu_dtor() argument 30 if (!nvif_object_constructed(&mmu->object)) in nvif_mmu_dtor() 33 kfree(mmu->kind); in nvif_mmu_dtor() 34 kfree(mmu->type); in nvif_mmu_dtor() 35 kfree(mmu->heap); in nvif_mmu_dtor() 36 nvif_object_dtor(&mmu->object); in nvif_mmu_dtor() 41 struct nvif_mmu *mmu) in nvif_mmu_ctor() argument 53 mmu->heap = NULL; in nvif_mmu_ctor() 54 mmu->type = NULL; in nvif_mmu_ctor() 55 mmu->kind = NULL; in nvif_mmu_ctor() [all …]
|
H A D | mem.c | 28 nvif_mem_ctor_map(struct nvif_mmu *mmu, const char *name, u8 type, u64 size, in nvif_mem_ctor_map() argument 31 int ret = nvif_mem_ctor(mmu, name, mmu->mem, NVIF_MEM_MAPPABLE | type, in nvif_mem_ctor_map() 48 nvif_mem_ctor_type(struct nvif_mmu *mmu, const char *name, s32 oclass, in nvif_mem_ctor_type() argument 72 ret = nvif_object_ctor(&mmu->object, name ? name : "nvifMem", 0, oclass, in nvif_mem_ctor_type() 75 mem->type = mmu->type[type].type; in nvif_mem_ctor_type() 88 nvif_mem_ctor(struct nvif_mmu *mmu, const char *name, s32 oclass, u8 type, in nvif_mem_ctor() argument 95 for (i = 0; ret && i < mmu->type_nr; i++) { in nvif_mem_ctor() 96 if ((mmu->type[i].type & type) == type) { in nvif_mem_ctor() 97 ret = nvif_mem_ctor_type(mmu, name, oclass, i, page, in nvif_mem_ctor()
|
/openbmc/linux/drivers/staging/media/atomisp/pci/mmu/ |
H A D | isp_mmu.c | 57 static void free_mmu_map(struct isp_mmu *mmu, unsigned int start_isp_virt, 79 static phys_addr_t isp_pte_to_pgaddr(struct isp_mmu *mmu, in isp_pte_to_pgaddr() argument 82 return mmu->driver->pte_to_phys(mmu, pte); in isp_pte_to_pgaddr() 85 static unsigned int isp_pgaddr_to_pte_valid(struct isp_mmu *mmu, in isp_pgaddr_to_pte_valid() argument 88 unsigned int pte = mmu->driver->phys_to_pte(mmu, phys); in isp_pgaddr_to_pte_valid() 90 return (unsigned int)(pte | ISP_PTE_VALID_MASK(mmu)); in isp_pgaddr_to_pte_valid() 97 static phys_addr_t alloc_page_table(struct isp_mmu *mmu) in alloc_page_table() argument 119 atomisp_set_pte(page, i, mmu->driver->null_pte); in alloc_page_table() 125 static void free_page_table(struct isp_mmu *mmu, phys_addr_t page) in free_page_table() argument 142 static void mmu_remap_error(struct isp_mmu *mmu, in mmu_remap_error() argument [all …]
|
/openbmc/linux/drivers/iommu/ |
H A D | ipmmu-vmsa.c | 72 struct ipmmu_vmsa_device *mmu; member 150 static bool ipmmu_is_root(struct ipmmu_vmsa_device *mmu) in ipmmu_is_root() argument 152 return mmu->root == mmu; in ipmmu_is_root() 157 struct ipmmu_vmsa_device *mmu = dev_get_drvdata(dev); in __ipmmu_check_device() local 160 if (ipmmu_is_root(mmu)) in __ipmmu_check_device() 161 *rootp = mmu; in __ipmmu_check_device() 178 static u32 ipmmu_read(struct ipmmu_vmsa_device *mmu, unsigned int offset) in ipmmu_read() argument 180 return ioread32(mmu->base + offset); in ipmmu_read() 183 static void ipmmu_write(struct ipmmu_vmsa_device *mmu, unsigned int offset, in ipmmu_write() argument 186 iowrite32(data, mmu->base + offset); in ipmmu_write() [all …]
|
/openbmc/linux/arch/x86/kernel/ |
H A D | paravirt.c | 232 .mmu.flush_tlb_user = native_flush_tlb_local, 233 .mmu.flush_tlb_kernel = native_flush_tlb_global, 234 .mmu.flush_tlb_one_user = native_flush_tlb_one_user, 235 .mmu.flush_tlb_multi = native_flush_tlb_multi, 236 .mmu.tlb_remove_table = native_tlb_remove_table, 238 .mmu.exit_mmap = paravirt_nop, 239 .mmu.notify_page_enc_status_changed = paravirt_nop, 242 .mmu.read_cr2 = __PV_IS_CALLEE_SAVE(pv_native_read_cr2), 243 .mmu.write_cr2 = pv_native_write_cr2, 244 .mmu.read_cr3 = __native_read_cr3, [all …]
|
/openbmc/qemu/target/microblaze/ |
H A D | mmu.c | 40 MicroBlazeMMU *mmu = &env->mmu; in mmu_flush_idx() local 44 t = mmu->rams[RAM_TAG][idx]; in mmu_flush_idx() 60 MicroBlazeMMU *mmu = &env->mmu; in mmu_change_pid() local 67 for (i = 0; i < ARRAY_SIZE(mmu->rams[RAM_TAG]); i++) { in mmu_change_pid() 69 t = mmu->rams[RAM_TAG][i]; in mmu_change_pid() 71 if (mmu->tids[i] && ((mmu->regs[MMU_R_PID] & 0xff) == mmu->tids[i])) in mmu_change_pid() 81 MicroBlazeMMU *mmu = &cpu->env.mmu; in mmu_translate() local 88 for (i = 0; i < ARRAY_SIZE(mmu->rams[RAM_TAG]); i++) { in mmu_translate() 92 t = mmu->rams[RAM_TAG][i]; in mmu_translate() 105 if (mmu->tids[i] in mmu_translate() [all …]
|
/openbmc/linux/drivers/staging/media/atomisp/include/mmu/ |
H A D | isp_mmu.h | 88 unsigned int (*get_pd_base)(struct isp_mmu *mmu, phys_addr_t pd_base); 100 void (*tlb_flush_range)(struct isp_mmu *mmu, 102 void (*tlb_flush_all)(struct isp_mmu *mmu); 103 unsigned int (*phys_to_pte)(struct isp_mmu *mmu, 105 phys_addr_t (*pte_to_phys)(struct isp_mmu *mmu, 120 #define ISP_PTE_VALID_MASK(mmu) \ argument 121 ((mmu)->driver->pte_valid_mask) 123 #define ISP_PTE_VALID(mmu, pte) \ argument 124 ((pte) & ISP_PTE_VALID_MASK(mmu)) 132 int isp_mmu_init(struct isp_mmu *mmu, struct isp_mmu_client *driver); [all …]
|
/openbmc/linux/drivers/gpu/drm/panfrost/ |
H A D | panfrost_mmu.c | 113 struct panfrost_mmu *mmu, in mmu_hw_do_operation() argument 119 ret = mmu_hw_do_operation_locked(pfdev, mmu->as, iova, size, op); in mmu_hw_do_operation() 124 static void panfrost_mmu_enable(struct panfrost_device *pfdev, struct panfrost_mmu *mmu) in panfrost_mmu_enable() argument 126 int as_nr = mmu->as; in panfrost_mmu_enable() 127 struct io_pgtable_cfg *cfg = &mmu->pgtbl_cfg; in panfrost_mmu_enable() 158 u32 panfrost_mmu_as_get(struct panfrost_device *pfdev, struct panfrost_mmu *mmu) in panfrost_mmu_as_get() argument 164 as = mmu->as; in panfrost_mmu_as_get() 166 int en = atomic_inc_return(&mmu->as_count); in panfrost_mmu_as_get() 175 list_move(&mmu->list, &pfdev->as_lru_list); in panfrost_mmu_as_get() 185 panfrost_mmu_enable(pfdev, mmu); in panfrost_mmu_as_get() [all …]
|
/openbmc/linux/drivers/gpu/drm/msm/ |
H A D | msm_mmu.h | 13 void (*detach)(struct msm_mmu *mmu); 14 int (*map)(struct msm_mmu *mmu, uint64_t iova, struct sg_table *sgt, 16 int (*unmap)(struct msm_mmu *mmu, uint64_t iova, size_t len); 17 void (*destroy)(struct msm_mmu *mmu); 18 void (*resume_translation)(struct msm_mmu *mmu); 35 static inline void msm_mmu_init(struct msm_mmu *mmu, struct device *dev, in msm_mmu_init() argument 38 mmu->dev = dev; in msm_mmu_init() 39 mmu->funcs = funcs; in msm_mmu_init() 40 mmu->type = type; in msm_mmu_init() 47 static inline void msm_mmu_set_fault_handler(struct msm_mmu *mmu, void *arg, in msm_mmu_set_fault_handler() argument [all …]
|
H A D | msm_iommu.c | 30 static struct msm_iommu_pagetable *to_pagetable(struct msm_mmu *mmu) in to_pagetable() argument 32 return container_of(mmu, struct msm_iommu_pagetable, base); in to_pagetable() 91 static int msm_iommu_pagetable_unmap(struct msm_mmu *mmu, u64 iova, in msm_iommu_pagetable_unmap() argument 94 struct msm_iommu_pagetable *pagetable = to_pagetable(mmu); in msm_iommu_pagetable_unmap() 115 static int msm_iommu_pagetable_map(struct msm_mmu *mmu, u64 iova, in msm_iommu_pagetable_map() argument 118 struct msm_iommu_pagetable *pagetable = to_pagetable(mmu); in msm_iommu_pagetable_map() 145 msm_iommu_pagetable_unmap(mmu, iova, addr - iova); in msm_iommu_pagetable_map() 154 static void msm_iommu_pagetable_destroy(struct msm_mmu *mmu) in msm_iommu_pagetable_destroy() argument 156 struct msm_iommu_pagetable *pagetable = to_pagetable(mmu); in msm_iommu_pagetable_destroy() 172 int msm_iommu_pagetable_params(struct msm_mmu *mmu, in msm_iommu_pagetable_params() argument [all …]
|
H A D | msm_gem_vma.c | 19 if (aspace->mmu) in msm_gem_address_space_destroy() 20 aspace->mmu->funcs->destroy(aspace->mmu); in msm_gem_address_space_destroy() 51 aspace->mmu->funcs->unmap(aspace->mmu, vma->iova, size); in msm_gem_vma_purge() 84 ret = aspace->mmu->funcs->map(aspace->mmu, vma->iova, sgt, size, prot); in msm_gem_vma_map() 154 msm_gem_address_space_create(struct msm_mmu *mmu, const char *name, in msm_gem_address_space_create() argument 159 if (IS_ERR(mmu)) in msm_gem_address_space_create() 160 return ERR_CAST(mmu); in msm_gem_address_space_create() 168 aspace->mmu = mmu; in msm_gem_address_space_create()
|
H A D | msm_gpummu.c | 24 static void msm_gpummu_detach(struct msm_mmu *mmu) in msm_gpummu_detach() argument 28 static int msm_gpummu_map(struct msm_mmu *mmu, uint64_t iova, in msm_gpummu_map() argument 31 struct msm_gpummu *gpummu = to_msm_gpummu(mmu); in msm_gpummu_map() 56 static int msm_gpummu_unmap(struct msm_mmu *mmu, uint64_t iova, size_t len) in msm_gpummu_unmap() argument 58 struct msm_gpummu *gpummu = to_msm_gpummu(mmu); in msm_gpummu_unmap() 71 static void msm_gpummu_resume_translation(struct msm_mmu *mmu) in msm_gpummu_resume_translation() argument 75 static void msm_gpummu_destroy(struct msm_mmu *mmu) in msm_gpummu_destroy() argument 77 struct msm_gpummu *gpummu = to_msm_gpummu(mmu); in msm_gpummu_destroy() 79 dma_free_attrs(mmu->dev, TABLE_SIZE, gpummu->table, gpummu->pt_base, in msm_gpummu_destroy() 114 void msm_gpummu_params(struct msm_mmu *mmu, dma_addr_t *pt_base, in msm_gpummu_params() argument [all …]
|
/openbmc/linux/drivers/accel/ivpu/ |
H A D | ivpu_mmu.c | 279 struct ivpu_mmu_info *mmu = vdev->mmu; in ivpu_mmu_cdtab_alloc() local 280 struct ivpu_mmu_cdtab *cdtab = &mmu->cdtab; in ivpu_mmu_cdtab_alloc() 294 struct ivpu_mmu_info *mmu = vdev->mmu; in ivpu_mmu_strtab_alloc() local 295 struct ivpu_mmu_strtab *strtab = &mmu->strtab; in ivpu_mmu_strtab_alloc() 314 struct ivpu_mmu_info *mmu = vdev->mmu; in ivpu_mmu_cmdq_alloc() local 315 struct ivpu_mmu_queue *q = &mmu->cmdq; in ivpu_mmu_cmdq_alloc() 333 struct ivpu_mmu_info *mmu = vdev->mmu; in ivpu_mmu_evtq_alloc() local 334 struct ivpu_mmu_queue *q = &mmu->evtq; in ivpu_mmu_evtq_alloc() 408 struct ivpu_mmu_queue *cmdq = &vdev->mmu->cmdq; in ivpu_mmu_cmdq_wait_for_cons() 416 struct ivpu_mmu_queue *q = &vdev->mmu->cmdq; in ivpu_mmu_cmdq_cmd_write() [all …]
|
/openbmc/linux/arch/arc/mm/ |
H A D | tlb.c | 136 struct cpuinfo_arc_mmu *mmu = &mmuinfo; in local_flush_tlb_all() local 139 int num_tlb = mmu->sets * mmu->ways; in local_flush_tlb_all() 571 struct cpuinfo_arc_mmu *mmu = &mmuinfo; in arc_mmu_mumbojumbo() local 579 mmu->ver = (bcr >> 24); in arc_mmu_mumbojumbo() 581 if (is_isa_arcompact() && mmu->ver == 3) { in arc_mmu_mumbojumbo() 583 mmu->pg_sz_k = 1 << (mmu3->pg_sz - 1); in arc_mmu_mumbojumbo() 584 mmu->sets = 1 << mmu3->sets; in arc_mmu_mumbojumbo() 585 mmu->ways = 1 << mmu3->ways; in arc_mmu_mumbojumbo() 591 mmu->pg_sz_k = 1 << (mmu4->sz0 - 1); in arc_mmu_mumbojumbo() 592 mmu->s_pg_sz_m = 1 << (mmu4->sz1 - 11); in arc_mmu_mumbojumbo() [all …]
|
/openbmc/linux/arch/x86/include/asm/ |
H A D | paravirt.h | 71 PVOP_VCALL0(mmu.flush_tlb_user); in __flush_tlb_local() 76 PVOP_VCALL0(mmu.flush_tlb_kernel); in __flush_tlb_global() 81 PVOP_VCALL1(mmu.flush_tlb_one_user, addr); in __flush_tlb_one_user() 87 PVOP_VCALL2(mmu.flush_tlb_multi, cpumask, info); in __flush_tlb_multi() 92 PVOP_VCALL2(mmu.tlb_remove_table, tlb, table); in paravirt_tlb_remove_table() 97 PVOP_VCALL1(mmu.exit_mmap, mm); in paravirt_arch_exit_mmap() 103 PVOP_VCALL3(mmu.notify_page_enc_status_changed, pfn, npages, enc); in notify_page_enc_status_changed() 144 return PVOP_ALT_CALLEE0(unsigned long, mmu.read_cr2, in read_cr2() 151 PVOP_VCALL1(mmu.write_cr2, x); in write_cr2() 156 return PVOP_ALT_CALL0(unsigned long, mmu.read_cr3, in __read_cr3() [all …]
|
/openbmc/linux/arch/arm64/kvm/hyp/nvhe/ |
H A D | tlb.c | 17 static void __tlb_switch_to_guest(struct kvm_s2_mmu *mmu, in __tlb_switch_to_guest() argument 65 __load_stage2(mmu, kern_hyp_va(mmu->arch)); in __tlb_switch_to_guest() 81 void __kvm_tlb_flush_vmid_ipa(struct kvm_s2_mmu *mmu, in __kvm_tlb_flush_vmid_ipa() argument 87 __tlb_switch_to_guest(mmu, &cxt, false); in __kvm_tlb_flush_vmid_ipa() 133 void __kvm_tlb_flush_vmid_ipa_nsh(struct kvm_s2_mmu *mmu, in __kvm_tlb_flush_vmid_ipa_nsh() argument 139 __tlb_switch_to_guest(mmu, &cxt, true); in __kvm_tlb_flush_vmid_ipa_nsh() 185 void __kvm_tlb_flush_vmid_range(struct kvm_s2_mmu *mmu, in __kvm_tlb_flush_vmid_range() argument 199 __tlb_switch_to_guest(mmu, &cxt, false); in __kvm_tlb_flush_vmid_range() 215 void __kvm_tlb_flush_vmid(struct kvm_s2_mmu *mmu) in __kvm_tlb_flush_vmid() argument 220 __tlb_switch_to_guest(mmu, &cxt, false); in __kvm_tlb_flush_vmid() [all …]
|
/openbmc/linux/arch/arm64/kvm/hyp/vhe/ |
H A D | tlb.c | 19 static void __tlb_switch_to_guest(struct kvm_s2_mmu *mmu, in __tlb_switch_to_guest() argument 56 __load_stage2(mmu, mmu->arch); in __tlb_switch_to_guest() 82 void __kvm_tlb_flush_vmid_ipa(struct kvm_s2_mmu *mmu, in __kvm_tlb_flush_vmid_ipa() argument 90 __tlb_switch_to_guest(mmu, &cxt); in __kvm_tlb_flush_vmid_ipa() 114 void __kvm_tlb_flush_vmid_ipa_nsh(struct kvm_s2_mmu *mmu, in __kvm_tlb_flush_vmid_ipa_nsh() argument 122 __tlb_switch_to_guest(mmu, &cxt); in __kvm_tlb_flush_vmid_ipa_nsh() 146 void __kvm_tlb_flush_vmid_range(struct kvm_s2_mmu *mmu, in __kvm_tlb_flush_vmid_range() argument 162 __tlb_switch_to_guest(mmu, &cxt); in __kvm_tlb_flush_vmid_range() 174 void __kvm_tlb_flush_vmid(struct kvm_s2_mmu *mmu) in __kvm_tlb_flush_vmid() argument 181 __tlb_switch_to_guest(mmu, &cxt); in __kvm_tlb_flush_vmid() [all …]
|
/openbmc/linux/drivers/xen/ |
H A D | grant-dma-iommu.c | 36 struct grant_dma_iommu_device *mmu; in grant_dma_iommu_probe() local 39 mmu = devm_kzalloc(&pdev->dev, sizeof(*mmu), GFP_KERNEL); in grant_dma_iommu_probe() 40 if (!mmu) in grant_dma_iommu_probe() 43 mmu->dev = &pdev->dev; in grant_dma_iommu_probe() 45 ret = iommu_device_register(&mmu->iommu, &grant_dma_iommu_ops, &pdev->dev); in grant_dma_iommu_probe() 49 platform_set_drvdata(pdev, mmu); in grant_dma_iommu_probe() 56 struct grant_dma_iommu_device *mmu = platform_get_drvdata(pdev); in grant_dma_iommu_remove() local 59 iommu_device_unregister(&mmu->iommu); in grant_dma_iommu_remove()
|
/openbmc/linux/arch/x86/kvm/mmu/ |
H A D | paging_tmpl.h | 31 #define PT_HAVE_ACCESSED_DIRTY(mmu) true argument 45 #define PT_HAVE_ACCESSED_DIRTY(mmu) true argument 58 #define PT_HAVE_ACCESSED_DIRTY(mmu) (!(mmu)->cpu_role.base.ad_disabled) argument 109 static inline void FNAME(protect_clean_gpte)(struct kvm_mmu *mmu, unsigned *access, in FNAME() 115 if (!PT_HAVE_ACCESSED_DIRTY(mmu)) in FNAME() 145 static bool FNAME(is_rsvd_bits_set)(struct kvm_mmu *mmu, u64 gpte, int level) in FNAME() 147 return __is_rsvd_bits_set(&mmu->guest_rsvd_check, gpte, level) || in FNAME() 148 FNAME(is_bad_mt_xwr)(&mmu->guest_rsvd_check, gpte); in FNAME() 159 if (PT_HAVE_ACCESSED_DIRTY(vcpu->arch.mmu) && in FNAME() 163 if (FNAME(is_rsvd_bits_set)(vcpu->arch.mmu, gpte, PG_LEVEL_4K)) in FNAME() [all …]
|
/openbmc/qemu/target/m68k/ |
H A D | helper.c | 202 env->mmu.tcr = val; in HELPER() 209 env->mmu.mmusr = val; in HELPER() 217 env->mmu.srp = val; in HELPER() 225 env->mmu.urp = val; in HELPER() 254 env->mmu.ttr[M68K_ITTR0] = val; in HELPER() 261 env->mmu.ttr[M68K_ITTR1] = val; in HELPER() 268 env->mmu.ttr[M68K_DTTR0] = val; in HELPER() 275 env->mmu.ttr[M68K_DTTR1] = val; in HELPER() 318 return env->mmu.tcr; in HELPER() 324 return env->mmu.mmusr; in HELPER() [all …]
|