/openbmc/linux/include/linux/ |
H A D | hugetlb.h | 113 struct vm_area_struct *vma; member 143 struct vm_area_struct *vma, 161 struct vm_area_struct *vma, 257 if (is_vm_hugetlb_page(vma)) in hugetlb_zap_begin() 264 if (is_vm_hugetlb_page(vma)) in hugetlb_zap_end() 307 struct vm_area_struct *vma, in huge_pmd_unshare() argument 314 struct vm_area_struct *vma, in adjust_range_if_pmd_sharing_possible() argument 320 struct vm_area_struct *vma, in hugetlb_zap_begin() argument 326 struct vm_area_struct *vma, in hugetlb_zap_end() argument 1261 #define flush_hugetlb_tlb_range(vma, addr, end) flush_tlb_range(vma, addr, end) argument [all …]
|
H A D | mm.h | 810 memset(vma, 0, sizeof(*vma)); in vma_init() 865 vm_flags_init(vma, (vma->vm_flags | set) & ~clear); in __vm_flags_mod() 895 return vma->vm_start <= vma->vm_mm->brk && in vma_is_initial_heap() 896 vma->vm_end >= vma->vm_mm->start_brk; in vma_is_initial_heap() 911 vma->vm_end >= vma->vm_mm->start_stack; in vma_is_initial_stack() 2345 zap_page_range_single(vma, vma->vm_start, in zap_vma_pages() 2346 vma->vm_end - vma->vm_start, NULL); in zap_vma_pages() 3204 vma; vma = vma_interval_tree_iter_next(vma, start, last)) 3428 if (vma && (vma->vm_start != vm_start || vma->vm_end != vm_end)) in find_exact_vma() 3437 return (vma && vma->vm_start <= start && end <= vma->vm_end); in range_in_vma() [all …]
|
/openbmc/linux/fs/proc/ |
H A D | task_mmu.c | 303 if (vma->vm_ops && vma->vm_ops->name) { in show_map_vma() 304 name = vma->vm_ops->name(vma); in show_map_vma() 500 struct vm_area_struct *vma = walk->vma; in smaps_pte_hole() local 526 struct vm_area_struct *vma = walk->vma; in smaps_pte_entry() local 573 struct vm_area_struct *vma = walk->vma; in smaps_pmd_entry() local 612 struct vm_area_struct *vma = walk->vma; in smaps_pte_range() local 725 struct vm_area_struct *vma = walk->vma; in smaps_hugetlb_range() local 1166 struct vm_area_struct *vma = walk->vma; in clear_refs_pte_range() local 1226 struct vm_area_struct *vma = walk->vma; in clear_refs_test_walk() local 1464 struct vm_area_struct *vma = walk->vma; in pagemap_pmd_range() local [all …]
|
H A D | task_nommu.c | 29 for_each_vma(vmi, vma) { in task_mem() 30 bytes += kobjsize(vma); in task_mem() 32 region = vma->vm_region; in task_mem() 37 size = vma->vm_end - vma->vm_start; in task_mem() 88 for_each_vma(vmi, vma) in task_vsize() 89 vsize += vma->vm_end - vma->vm_start; in task_vsize() 137 file = vma->vm_file; in nommu_vma_show() 149 vma->vm_start, in nommu_vma_show() 150 vma->vm_end, in nommu_vma_show() 183 if (vma) { in proc_get_vma() [all …]
|
/openbmc/linux/drivers/char/ |
H A D | mspec.c | 88 mspec_open(struct vm_area_struct *vma) in mspec_open() argument 92 vdata = vma->vm_private_data; in mspec_open() 109 vdata = vma->vm_private_data; in mspec_close() 187 if (vma->vm_pgoff != 0) in mspec_mmap() 193 if ((vma->vm_flags & VM_WRITE) == 0) in mspec_mmap() 196 pages = vma_pages(vma); in mspec_mmap() 202 vdata->vm_start = vma->vm_start; in mspec_mmap() 203 vdata->vm_end = vma->vm_end; in mspec_mmap() 207 vma->vm_private_data = vdata; in mspec_mmap() 211 vma->vm_page_prot = pgprot_noncached(vma->vm_page_prot); in mspec_mmap() [all …]
|
/openbmc/linux/drivers/gpu/drm/ttm/ |
H A D | ttm_bo_vm.c | 62 mmap_read_unlock(vmf->vma->vm_mm); in ttm_bo_vm_fault_idle() 185 struct vm_area_struct *vma = vmf->vma; in ttm_bo_vm_fault_reserved() local 212 page_last = vma_pages(vma) + vma->vm_pgoff - in ttm_bo_vm_fault_reserved() 293 struct vm_area_struct *vma = vmf->vma; in ttm_bo_vm_dummy_page() local 313 for (address = vma->vm_start; address < vma->vm_end; in ttm_bo_vm_dummy_page() 323 struct vm_area_struct *vma = vmf->vma; in ttm_bo_vm_fault() local 334 prot = vma->vm_page_prot; in ttm_bo_vm_fault() 365 vma->vm_private_data = NULL; in ttm_bo_vm_close() 471 if (!vma->vm_ops) in ttm_bo_mmap_obj() 472 vma->vm_ops = &ttm_bo_vm_ops; in ttm_bo_mmap_obj() [all …]
|
/openbmc/linux/mm/ |
H A D | msync.c | 36 struct vm_area_struct *vma; in SYSCALL_DEFINE3() local 63 vma = find_vma(mm, start); in SYSCALL_DEFINE3() 70 if (!vma) in SYSCALL_DEFINE3() 73 if (start < vma->vm_start) { in SYSCALL_DEFINE3() 76 start = vma->vm_start; in SYSCALL_DEFINE3() 83 (vma->vm_flags & VM_LOCKED)) { in SYSCALL_DEFINE3() 87 file = vma->vm_file; in SYSCALL_DEFINE3() 91 start = vma->vm_end; in SYSCALL_DEFINE3() 93 (vma->vm_flags & VM_SHARED)) { in SYSCALL_DEFINE3() 101 vma = find_vma(mm, start); in SYSCALL_DEFINE3() [all …]
|
H A D | pgtable-generic.c | 68 int ptep_set_access_flags(struct vm_area_struct *vma, in ptep_set_access_flags() argument 74 set_pte_at(vma->vm_mm, address, ptep, entry); in ptep_set_access_flags() 75 flush_tlb_fix_spurious_fault(vma, address, ptep); in ptep_set_access_flags() 82 int ptep_clear_flush_young(struct vm_area_struct *vma, in ptep_clear_flush_young() argument 88 flush_tlb_page(vma, address); in ptep_clear_flush_young() 97 struct mm_struct *mm = (vma)->vm_mm; in ptep_clear_flush() 101 flush_tlb_page(vma, address); in ptep_clear_flush() 109 int pmdp_set_access_flags(struct vm_area_struct *vma, in pmdp_set_access_flags() argument 116 set_pmd_at(vma->vm_mm, address, pmdp, entry); in pmdp_set_access_flags() 124 int pmdp_clear_flush_young(struct vm_area_struct *vma, in pmdp_clear_flush_young() argument [all …]
|
H A D | hugetlb.c | 356 struct vm_area_struct *vma = vma_lock->vma; in __hugetlb_vma_unlock_write_put() local 388 if (!vma || !__vma_shareable_lock(vma)) in hugetlb_vma_lock_free() 404 if (!vma || !(vma->vm_flags & VM_MAYSHARE)) in hugetlb_vma_lock_alloc() 429 vma_lock->vma = vma; in hugetlb_vma_lock_alloc() 1014 if (vma->vm_ops && vma->vm_ops->pagesize) in vma_kernel_pagesize() 1015 return vma->vm_ops->pagesize(vma); in vma_kernel_pagesize() 1217 if (vma_lock && vma_lock->vma != vma) in hugetlb_dup_vma_private() 4860 if (vma_lock->vma != vma) { in hugetlb_vm_op_open() 5804 .vma = vma, in hugetlb_handle_userfault() 6204 .vma = vma, in hugetlb_fault() [all …]
|
H A D | mincore.c | 78 if (vma->vm_file) { in __mincore_unmapped_range() 96 walk->vma, walk->private); in mincore_unmapped_range() 104 struct vm_area_struct *vma = walk->vma; in mincore_pte_range() local 109 ptl = pmd_trans_huge_lock(pmd, vma); in mincore_pte_range() 127 vma, vec); in mincore_pte_range() 160 if (vma_is_anonymous(vma)) in can_do_mincore() 162 if (!vma->vm_file) in can_do_mincore() 189 struct vm_area_struct *vma; in do_mincore() local 193 vma = vma_lookup(current->mm, addr); in do_mincore() 194 if (!vma) in do_mincore() [all …]
|
/openbmc/linux/fs/ |
H A D | userfaultfd.c | 283 struct vm_area_struct *vma = vmf->vma; in userfaultfd_huge_must_wait() local 414 struct vm_area_struct *vma = vmf->vma; in handle_userfault() local 926 uffd_wp_range(vma, vma->vm_start, in userfaultfd_release() 927 vma->vm_end - vma->vm_start, false); in userfaultfd_release() 931 vma->vm_file, vma->vm_pgoff, in userfaultfd_release() 1378 if (!vma) in userfaultfd_register() 1397 cur = vma; in userfaultfd_register() 1487 vma->anon_vma, vma->vm_file, pgoff, in userfaultfd_register() 1590 if (!vma) in userfaultfd_unregister() 1638 BUG_ON(!vma_can_userfault(vma, vma->vm_flags)); in userfaultfd_unregister() [all …]
|
/openbmc/linux/arch/arc/include/asm/ |
H A D | tlbflush.h | 13 void local_flush_tlb_page(struct vm_area_struct *vma, unsigned long page); 15 void local_flush_tlb_range(struct vm_area_struct *vma, 18 void local_flush_pmd_tlb_range(struct vm_area_struct *vma, unsigned long start, 23 #define flush_tlb_range(vma, s, e) local_flush_tlb_range(vma, s, e) argument 24 #define flush_tlb_page(vma, page) local_flush_tlb_page(vma, page) argument 29 #define flush_pmd_tlb_range(vma, s, e) local_flush_pmd_tlb_range(vma, s, e) argument 32 extern void flush_tlb_range(struct vm_area_struct *vma, unsigned long start, 34 extern void flush_tlb_page(struct vm_area_struct *vma, unsigned long page); 39 extern void flush_pmd_tlb_range(struct vm_area_struct *vma, unsigned long start, unsigned long end);
|
/openbmc/linux/drivers/gpu/drm/i915/gem/ |
H A D | i915_gem_mman.c | 33 if (vma->vm_file != filp) in __vma_matches() 37 (vma->vm_end - vma->vm_start) == PAGE_ALIGN(size); in __vma_matches() 114 if (vma && __vma_matches(vma, obj->base.filp, addr, args->size)) in i915_gem_mmap_ioctl() 115 vma->vm_page_prot = in i915_gem_mmap_ioctl() 382 if (IS_ERR(vma) && vma != ERR_PTR(-EDEADLK)) { in vm_fault_gtt() 398 if (IS_ERR(vma) && vma != ERR_PTR(-EDEADLK)) { in vm_fault_gtt() 420 if (IS_ERR(vma)) { in vm_fault_gtt() 421 ret = PTR_ERR(vma); in vm_fault_gtt() 464 vma->mmo = mmo; in vm_fault_gtt() 1038 vma->vm_page_prot = vm_get_page_prot(vma->vm_flags); in i915_gem_object_mmap() [all …]
|
/openbmc/linux/scripts/coccinelle/api/ |
H A D | vma_pages.cocci | 3 /// Use vma_pages function on vma object instead of explicit computation. 6 // Keywords: vma_pages vma 19 struct vm_area_struct *vma; 22 * (vma->vm_end - vma->vm_start) >> PAGE_SHIFT 29 struct vm_area_struct *vma; 32 - ((vma->vm_end - vma->vm_start) >> PAGE_SHIFT) 33 + vma_pages(vma) 40 struct vm_area_struct *vma; 44 (vma->vm_end@p - vma->vm_start) >> PAGE_SHIFT 48 x << r_org.vma; [all …]
|
/openbmc/linux/drivers/gpu/drm/i915/gt/ |
H A D | intel_ggtt_fencing.c | 210 if (vma) { in fence_update() 224 GEM_BUG_ON(vma->fence_size > i915_vma_size(vma)); in fence_update() 268 GEM_BUG_ON(vma); in fence_update() 272 WRITE_ONCE(fence->vma, vma); in fence_update() 275 if (vma) { in fence_update() 300 GEM_BUG_ON(fence->vma != vma); in i915_vma_revoke_fence() 326 return fence->vma && i915_vma_is_active(fence->vma); in fence_is_active() 335 GEM_BUG_ON(fence->vma && fence->vma->fence != fence); in fence_find() 372 if (vma->fence) { in __i915_vma_pin_fence() 374 GEM_BUG_ON(fence->vma != vma); in __i915_vma_pin_fence() [all …]
|
/openbmc/linux/Documentation/translations/zh_CN/core-api/ |
H A D | cachetlb.rst | 62 “vma”是用于该区域的备份存储。主要是用于munmap()类型的操作。 71 踪进程的mmap区域的支持结构体,地址空间可以通过vma->vm_mm获得。另 77 中不会有虚拟地址‘addr’的‘vma->vm_mm’的页表项。 85 软件页表中,在地址空间“vma->vm_mm”的虚拟地址“地址”处,现在存在 99 2) flush_cache_range(vma, start, end); 101 flush_tlb_range(vma, start, end); 103 3) flush_cache_page(vma, addr, pfn); 105 flush_tlb_page(vma, addr); 144 “vma”是被用于该区域的备份存储。主要是用于munmap()类型的操作。 152 这一次我们需要从缓存中删除一个PAGE_SIZE大小的区域。“vma”是 [all …]
|
/openbmc/linux/drivers/gpu/drm/nouveau/nvkm/subdev/mmu/ |
H A D | uvmm.c | 127 if (ret = -ENOENT, !vma || vma->addr != addr) { in nvkm_uvmm_mthd_unmap() 129 addr, vma ? vma->addr : ~0ULL); in nvkm_uvmm_mthd_unmap() 191 if (ret = -EINVAL, vma->mapped && !vma->memory) { in nvkm_uvmm_mthd_map() 197 if (addr + size > vma->addr + vma->size || vma->memory || in nvkm_uvmm_mthd_map() 201 !!vma->memory, vma->refd, vma->mapref, in nvkm_uvmm_mthd_map() 202 addr, size, vma->addr, (u64)vma->size); in nvkm_uvmm_mthd_map() 206 vma = nvkm_vmm_node_split(vmm, vma, addr, size); in nvkm_uvmm_mthd_map() 207 if (!vma) { in nvkm_uvmm_mthd_map() 249 if (ret = -ENOENT, !vma || vma->addr != addr || vma->part) { in nvkm_uvmm_mthd_put() 251 vma ? vma->addr : ~0ULL, vma ? vma->part : 0); in nvkm_uvmm_mthd_put() [all …]
|
/openbmc/linux/drivers/gpu/drm/i915/gem/selftests/ |
H A D | huge_pages.c | 415 vma->page_sizes.sg, vma->resource->page_sizes_gtt); in igt_check_page_sizes() 428 struct i915_vma *vma; in igt_mock_exhaust_device_supported_pages() local 462 if (IS_ERR(vma)) { in igt_mock_exhaust_device_supported_pages() 530 if (IS_ERR(vma)) { in igt_mock_memory_region_huge_pages() 568 i915_vma_unpin(vma); in igt_mock_memory_region_huge_pages() 622 if (IS_ERR(vma)) { in igt_mock_ppgtt_misaligned_dma() 772 if (IS_ERR(vma)) { in igt_ppgtt_huge_fill() 1156 if (IS_ERR(vma)) in __igt_write_huge() 1673 if (IS_ERR(vma)) { in igt_ppgtt_mixed() 1807 if (IS_ERR(vma)) { in igt_tmpfs_fallback() [all …]
|
H A D | igt_gem_utils.c | 42 igt_emit_store_dw(struct i915_vma *vma, in igt_emit_store_dw() argument 48 const int ver = GRAPHICS_VER(vma->vm->i915); in igt_emit_store_dw() 66 offset += i915_vma_offset(vma); in igt_emit_store_dw() 92 intel_gt_chipset_flush(vma->vm->gt); in igt_emit_store_dw() 94 vma = i915_vma_instance(obj, vma->vm, NULL); in igt_emit_store_dw() 95 if (IS_ERR(vma)) { in igt_emit_store_dw() 96 err = PTR_ERR(vma); in igt_emit_store_dw() 100 err = i915_vma_pin(vma, 0, 0, PIN_USER); in igt_emit_store_dw() 104 return vma; in igt_emit_store_dw() 112 struct i915_vma *vma, u64 offset, in igt_gpu_fill_dw() argument [all …]
|
/openbmc/linux/drivers/misc/ocxl/ |
H A D | context.c | 141 struct vm_area_struct *vma = vmf->vma; in ocxl_mmap_fault() local 162 struct vm_area_struct *vma) in check_mmap_afu_irq() argument 167 if (vma_pages(vma) != 1) in check_mmap_afu_irq() 180 if ((vma->vm_flags & VM_READ) || (vma->vm_flags & VM_EXEC) || in check_mmap_afu_irq() 181 !(vma->vm_flags & VM_WRITE)) in check_mmap_afu_irq() 188 struct vm_area_struct *vma) in check_mmap_mmio() argument 190 if ((vma_pages(vma) + vma->vm_pgoff) > in check_mmap_mmio() 201 rc = check_mmap_mmio(ctx, vma); in ocxl_context_mmap() 203 rc = check_mmap_afu_irq(ctx, vma); in ocxl_context_mmap() 208 vma->vm_page_prot = pgprot_noncached(vma->vm_page_prot); in ocxl_context_mmap() [all …]
|
/openbmc/linux/arch/x86/entry/vdso/ |
H A D | vma.c | 114 struct vm_area_struct *vma; in vdso_join_timens() local 118 for_each_vma(vmi, vma) { in vdso_join_timens() 120 zap_vma_pages(vma); in vdso_join_timens() 226 struct vm_area_struct *vma; in map_vdso() local 245 vma = _install_special_mapping(mm, in map_vdso() 252 if (IS_ERR(vma)) { in map_vdso() 253 ret = PTR_ERR(vma); in map_vdso() 264 if (IS_ERR(vma)) { in map_vdso() 265 ret = PTR_ERR(vma); in map_vdso() 333 struct vm_area_struct *vma; in map_vdso_once() local [all …]
|
/openbmc/linux/arch/powerpc/include/asm/book3s/64/ |
H A D | tlbflush.h | 47 static inline void flush_pmd_tlb_range(struct vm_area_struct *vma, in flush_pmd_tlb_range() argument 51 radix__flush_pmd_tlb_range(vma, start, end); in flush_pmd_tlb_range() 55 static inline void flush_pud_tlb_range(struct vm_area_struct *vma, in flush_pud_tlb_range() argument 59 radix__flush_pud_tlb_range(vma, start, end); in flush_pud_tlb_range() 68 radix__flush_hugetlb_tlb_range(vma, start, end); in flush_hugetlb_tlb_range() 71 static inline void flush_tlb_range(struct vm_area_struct *vma, in flush_tlb_range() argument 75 radix__flush_tlb_range(vma, start, end); in flush_tlb_range() 95 radix__local_flush_tlb_page(vma, vmaddr); in local_flush_tlb_page() 120 static inline void flush_tlb_page(struct vm_area_struct *vma, in flush_tlb_page() argument 124 radix__flush_tlb_page(vma, vmaddr); in flush_tlb_page() [all …]
|
/openbmc/linux/drivers/gpu/drm/i915/display/ |
H A D | intel_plane_initial.c | 52 struct i915_vma *vma; in initial_plane_vma() local 142 if (IS_ERR(vma)) in initial_plane_vma() 155 return vma; in initial_plane_vma() 170 struct i915_vma *vma; in intel_alloc_initial_plane_obj() local 186 if (!vma) in intel_alloc_initial_plane_obj() 202 plane_config->vma = vma; in intel_alloc_initial_plane_obj() 206 i915_vma_put(vma); in intel_alloc_initial_plane_obj() 221 struct i915_vma *vma; in intel_find_initial_plane_obj() local 233 vma = plane_config->vma; in intel_find_initial_plane_obj() 260 __i915_vma_pin(vma); in intel_find_initial_plane_obj() [all …]
|
/openbmc/linux/tools/testing/selftests/bpf/progs/ |
H A D | bpf_iter_task_vma.c | 28 struct vm_area_struct *vma = ctx->vma; in proc_maps() local 34 if (task == (void *)0 || vma == (void *)0) in proc_maps() 37 file = vma->vm_file; in proc_maps() 43 perm_str[0] = (vma->vm_flags & VM_READ) ? 'r' : '-'; in proc_maps() 44 perm_str[1] = (vma->vm_flags & VM_WRITE) ? 'w' : '-'; in proc_maps() 45 perm_str[2] = (vma->vm_flags & VM_EXEC) ? 'x' : '-'; in proc_maps() 46 perm_str[3] = (vma->vm_flags & VM_MAYSHARE) ? 's' : 'p'; in proc_maps() 47 BPF_SEQ_PRINTF(seq, "%08llx-%08llx %s ", vma->vm_start, vma->vm_end, perm_str); in proc_maps() 54 BPF_SEQ_PRINTF(seq, "%08llx ", vma->vm_pgoff << 12); in proc_maps()
|
/openbmc/linux/drivers/gpu/drm/i915/ |
H A D | i915_gem.c | 143 while (!ret && (vma = list_first_entry_or_null(&obj->vma.list, in i915_gem_object_unbind() 341 vma = NULL; in i915_gem_gtt_prepare() 401 if (IS_ERR(vma)) { in i915_gem_gtt_pread() 579 if (IS_ERR(vma)) { in i915_gem_gtt_pwrite_fast() 885 if (!reg->vma) in i915_gem_runtime_suspend() 899 rb_erase(&vma->obj_node, &obj->vma.tree); in discard_ggtt_vma() 954 if (IS_ERR(vma)) in i915_gem_object_ggtt_pin_ww() 955 return vma; in i915_gem_object_ggtt_pin_ww() 959 if (i915_vma_is_pinned(vma) || i915_vma_is_active(vma)) in i915_gem_object_ggtt_pin_ww() 978 if (i915_vma_is_pinned(vma) || i915_vma_is_active(vma)) { in i915_gem_object_ggtt_pin_ww() [all …]
|