/openbmc/linux/arch/powerpc/mm/ |
H A D | fault.c | 88 struct vm_area_struct *vma) in bad_access_pkey() argument 110 pkey = vma_pkey(vma); in bad_access_pkey() 231 struct vm_area_struct *vma) in access_pkey_error() argument 257 return !(vma->vm_flags & VM_EXEC) && in access_error() 273 if (unlikely(!vma_is_accessible(vma))) in access_error() 408 struct vm_area_struct * vma; in ___do_page_fault() local 476 if (!vma) in ___do_page_fault() 481 vma_end_read(vma); in ___do_page_fault() 486 vma_end_read(vma); in ___do_page_fault() 492 vma_end_read(vma); in ___do_page_fault() [all …]
|
/openbmc/linux/mm/ |
H A D | migrate_device.c | 44 if (!vma_is_anonymous(walk->vma)) in migrate_vma_collect_hole() 63 struct vm_area_struct *vma = walk->vma; in migrate_vma_collect_pmd() local 64 struct mm_struct *mm = vma->vm_mm; in migrate_vma_collect_pmd() 85 split_huge_pmd(vma, pmdp, addr); in migrate_vma_collect_pmd() 117 if (vma_is_anonymous(vma)) { in migrate_vma_collect_pmd() 520 if (!args->vma || is_vm_hugetlb_page(args->vma) || in migrate_vma_setup() 521 (args->vma->vm_flags & VM_SPECIAL) || vma_is_dax(args->vma)) in migrate_vma_setup() 528 if (args->end <= args->vma->vm_start || args->end > args->vma->vm_end) in migrate_vma_setup() 567 struct vm_area_struct *vma = migrate->vma; in migrate_vma_insert_page() local 580 if (!vma_is_anonymous(vma)) in migrate_vma_insert_page() [all …]
|
H A D | ksm.c | 562 if (!vma || !(vma->vm_flags & VM_MERGEABLE) || !vma->anon_vma) in find_mergeable_vma() 564 return vma; in find_mergeable_vma() 581 if (vma) in break_cow() 595 if (!vma) in get_mergeable_page() 1046 vma->vm_start, vma->vm_end, false); in unmerge_and_remove_all_rmap_items() 1362 if (!vma) in try_to_merge_with_ksm_page() 2194 if (vma) { in cmp_and_merge_page() 2553 err = unmerge_ksm_pages(vma, vma->vm_start, vma->vm_end, true); in __ksm_del_vma() 2871 vma = vmac->vma; in rmap_walk_ksm() 2876 if (addr < vma->vm_start || addr >= vma->vm_end) in rmap_walk_ksm() [all …]
|
/openbmc/linux/include/linux/ |
H A D | pgtable.h | 357 int ptep_clear_flush_young(struct vm_area_struct *vma, 737 set_pmd_at(vma->vm_mm, address, pmdp, pmd); in generic_pmdp_establish() 889 struct vm_area_struct *vma, in arch_do_swap_page() argument 907 struct vm_area_struct *vma, in arch_unmap_one() argument 956 #define flush_tlb_fix_spurious_fault(vma, address, ptep) flush_tlb_page(vma, address) argument 1065 return ptep_get_and_clear(vma->vm_mm, addr, ptep); in __ptep_modify_prot_start() 1076 set_pte_at(vma->vm_mm, addr, ptep, pte); in __ptep_modify_prot_commit() 1098 return __ptep_modify_prot_start(vma, addr, ptep); in ptep_modify_prot_start() 1109 __ptep_modify_prot_commit(vma, addr, ptep, pte); in ptep_modify_prot_commit() 1519 #define flush_pmd_tlb_range(vma, addr, end) flush_tlb_range(vma, addr, end) argument [all …]
|
H A D | rmap.h | 83 struct vm_area_struct *vma; member 156 if (likely(vma->anon_vma)) in anon_vma_prepare() 159 return __anon_vma_prepare(vma); in anon_vma_prepare() 165 VM_BUG_ON_VMA(vma->anon_vma != next->anon_vma, vma); in anon_vma_merge() 246 struct vm_area_struct *vma) in page_try_dup_anon_rmap() argument 265 unlikely(page_needs_cow_for_dma(vma, page))) in page_try_dup_anon_rmap() 387 struct vm_area_struct *vma; member 400 .vma = _vma, \ 410 .vma = _vma, \ 440 struct vm_area_struct *vma); [all …]
|
/openbmc/linux/drivers/gpu/drm/i915/gem/ |
H A D | i915_gem_execbuffer.c | 441 struct i915_vma *vma = ev->vma; in eb_pin_vma() local 569 ev->vma = vma; in eb_add_vma() 659 struct i915_vma *vma = ev->vma; in eb_reserve_vma() local 899 vma = i915_vma_tryget(vma); in eb_lookup_vma() 976 eb->vma[i + 1].vma = NULL; in eb_lookup_vmas() 990 eb->vma[i].vma = NULL; in eb_lookup_vmas() 1001 struct i915_vma *vma = ev->vma; in eb_lock_vmas() local 1025 struct i915_vma *vma = ev->vma; in eb_validate_vmas() local 1089 struct i915_vma *vma = ev->vma; in eb_release_vmas() local 2009 struct i915_vma *vma = ev->vma; in eb_capture_stage() local [all …]
|
/openbmc/linux/arch/ia64/include/asm/ |
H A D | fb.h | 13 static inline void fb_pgprotect(struct file *file, struct vm_area_struct *vma, in fb_pgprotect() argument 16 if (efi_range_is_wc(vma->vm_start, vma->vm_end - vma->vm_start)) in fb_pgprotect() 17 vma->vm_page_prot = pgprot_writecombine(vma->vm_page_prot); in fb_pgprotect() 19 vma->vm_page_prot = pgprot_noncached(vma->vm_page_prot); in fb_pgprotect()
|
/openbmc/linux/arch/loongarch/kernel/ |
H A D | vdso.c | 51 struct vm_area_struct *vma, struct vm_fault *vmf) in vvar_fault() argument 54 struct page *timens_page = find_timens_vvar_page(vma); in vvar_fault() 84 return vmf_insert_pfn(vma, vmf->address, pfn); in vvar_fault() 136 struct vm_area_struct *vma; in vdso_join_timens() local 141 for_each_vma(vmi, vma) { in vdso_join_timens() 143 zap_vma_pages(vma); in vdso_join_timens() 168 struct vm_area_struct *vma; in arch_setup_additional_pages() local 189 if (IS_ERR(vma)) { in arch_setup_additional_pages() 190 ret = PTR_ERR(vma); in arch_setup_additional_pages() 198 if (IS_ERR(vma)) { in arch_setup_additional_pages() [all …]
|
/openbmc/linux/drivers/infiniband/sw/rxe/ |
H A D | rxe_mmap.c | 37 static void rxe_vma_open(struct vm_area_struct *vma) in rxe_vma_open() argument 39 struct rxe_mmap_info *ip = vma->vm_private_data; in rxe_vma_open() 44 static void rxe_vma_close(struct vm_area_struct *vma) in rxe_vma_close() argument 46 struct rxe_mmap_info *ip = vma->vm_private_data; in rxe_vma_close() 62 int rxe_mmap(struct ib_ucontext *context, struct vm_area_struct *vma) in rxe_mmap() argument 65 unsigned long offset = vma->vm_pgoff << PAGE_SHIFT; in rxe_mmap() 66 unsigned long size = vma->vm_end - vma->vm_start; in rxe_mmap() 99 ret = remap_vmalloc_range(vma, ip->obj, 0); in rxe_mmap() 105 vma->vm_ops = &rxe_vm_ops; in rxe_mmap() 106 vma->vm_private_data = ip; in rxe_mmap() [all …]
|
/openbmc/linux/drivers/gpu/drm/i915/gem/selftests/ |
H A D | i915_gem_coherency.c | 88 struct i915_vma *vma; in gtt_set() local 99 if (IS_ERR(vma)) in gtt_set() 100 return PTR_ERR(vma); in gtt_set() 105 i915_vma_unpin(vma); in gtt_set() 121 struct i915_vma *vma; in gtt_get() local 132 if (IS_ERR(vma)) in gtt_get() 133 return PTR_ERR(vma); in gtt_get() 138 i915_vma_unpin(vma); in gtt_get() 199 struct i915_vma *vma; in gpu_set() local 204 if (IS_ERR(vma)) in gpu_set() [all …]
|
/openbmc/linux/drivers/char/ |
H A D | uv_mmtimer.c | 43 static int uv_mmtimer_mmap(struct file *file, struct vm_area_struct *vma); 147 static int uv_mmtimer_mmap(struct file *file, struct vm_area_struct *vma) in uv_mmtimer_mmap() argument 151 if (vma->vm_end - vma->vm_start != PAGE_SIZE) in uv_mmtimer_mmap() 154 if (vma->vm_flags & VM_WRITE) in uv_mmtimer_mmap() 160 vma->vm_page_prot = pgprot_noncached(vma->vm_page_prot); in uv_mmtimer_mmap() 166 if (remap_pfn_range(vma, vma->vm_start, uv_mmtimer_addr >> PAGE_SHIFT, in uv_mmtimer_mmap() 167 PAGE_SIZE, vma->vm_page_prot)) { in uv_mmtimer_mmap()
|
/openbmc/linux/arch/sh/include/asm/ |
H A D | tlbflush.h | 16 extern void local_flush_tlb_range(struct vm_area_struct *vma, 19 extern void local_flush_tlb_page(struct vm_area_struct *vma, 31 extern void flush_tlb_range(struct vm_area_struct *vma, unsigned long start, 33 extern void flush_tlb_page(struct vm_area_struct *vma, unsigned long page); 41 #define flush_tlb_page(vma, page) local_flush_tlb_page(vma, page) argument 44 #define flush_tlb_range(vma, start, end) \ argument 45 local_flush_tlb_range(vma, start, end)
|
/openbmc/linux/drivers/gpu/drm/i915/gt/ |
H A D | selftest_ring_submission.c | 12 struct i915_vma *vma; in create_wally() local 20 vma = i915_vma_instance(obj, engine->gt->vm, NULL); in create_wally() 21 if (IS_ERR(vma)) { in create_wally() 23 return vma; in create_wally() 32 err = i915_vma_sync(vma); in create_wally() 53 *cs++ = i915_vma_offset(vma) + 4000; in create_wally() 62 if (IS_ERR(vma->private)) { in create_wally() 63 vma = ERR_CAST(vma->private); in create_wally() 67 return vma; in create_wally() 223 engine->wa_ctx.vma = bb; in __live_ctx_switch_wa() [all …]
|
/openbmc/linux/arch/sparc/include/asm/ |
H A D | cacheflush_64.h | 24 #define flush_cache_range(vma, start, end) \ argument 25 flush_cache_mm((vma)->vm_mm) 26 #define flush_cache_page(vma, page, pfn) \ argument 27 flush_cache_mm((vma)->vm_mm) 60 #define copy_to_user_page(vma, page, vaddr, dst, src, len) \ argument 62 flush_cache_page(vma, vaddr, page_to_pfn(page)); \ 64 flush_ptrace_access(vma, page, vaddr, src, len, 0); \ 67 #define copy_from_user_page(vma, page, vaddr, dst, src, len) \ argument 69 flush_cache_page(vma, vaddr, page_to_pfn(page)); \ 71 flush_ptrace_access(vma, page, vaddr, dst, len, 1); \
|
/openbmc/linux/drivers/gpu/drm/i915/ |
H A D | i915_debugfs.c | 134 if (i915_vma_is_ggtt(vma)) in stringify_vma_type() 137 if (i915_vma_is_dpt(vma)) in stringify_vma_type() 191 struct i915_vma *vma; in i915_debugfs_describe_obj() local 208 spin_lock(&obj->vma.lock); in i915_debugfs_describe_obj() 209 list_for_each_entry(vma, &obj->vma.list, obj_link) { in i915_debugfs_describe_obj() 213 spin_unlock(&obj->vma.lock); in i915_debugfs_describe_obj() 220 i915_vma_offset(vma), i915_vma_size(vma), in i915_debugfs_describe_obj() 223 if (i915_vma_is_ggtt(vma) || i915_vma_is_dpt(vma)) { in i915_debugfs_describe_obj() 268 if (vma->fence) in i915_debugfs_describe_obj() 272 spin_lock(&obj->vma.lock); in i915_debugfs_describe_obj() [all …]
|
H A D | i915_mm.c | 94 int remap_io_mapping(struct vm_area_struct *vma, in remap_io_mapping() argument 101 GEM_BUG_ON((vma->vm_flags & EXPECTED_FLAGS) != EXPECTED_FLAGS); in remap_io_mapping() 104 r.mm = vma->vm_mm; in remap_io_mapping() 107 (pgprot_val(vma->vm_page_prot) & ~_PAGE_CACHE_MASK)); in remap_io_mapping() 111 zap_vma_ptes(vma, addr, (r.pfn - pfn) << PAGE_SHIFT); in remap_io_mapping() 129 int remap_io_sg(struct vm_area_struct *vma, in remap_io_sg() argument 134 .mm = vma->vm_mm, in remap_io_sg() 135 .prot = vma->vm_page_prot, in remap_io_sg() 142 GEM_BUG_ON((vma->vm_flags & EXPECTED_FLAGS) != EXPECTED_FLAGS); in remap_io_sg() 145 flush_cache_range(vma, addr, size); in remap_io_sg() [all …]
|
/openbmc/linux/drivers/dax/ |
H A D | device.c | 35 if (vma->vm_start & mask || vma->vm_end & mask) { in check_vma() 38 current->comm, func, vma->vm_start, vma->vm_end, in check_vma() 43 if (!vma_is_dax(vma)) { in check_vma() 88 pgoff = linear_page_index(vmf->vma, in dax_set_mapping() 161 if (pmd_addr < vmf->vma->vm_start || in __dev_dax_pmd_fault() 206 if (pud_addr < vmf->vma->vm_start || in __dev_dax_pud_fault() 240 vmf->vma->vm_start, vmf->vma->vm_end, order); in dev_dax_huge_fault() 264 struct file *filp = vma->vm_file; in dev_dax_may_split() 274 struct file *filp = vma->vm_file; in dev_dax_pagesize() 304 vma->vm_ops = &dax_vm_ops; in dax_mmap() [all …]
|
/openbmc/linux/drivers/accel/habanalabs/common/ |
H A D | memory_mgr.c | 202 (struct hl_mmap_mem_buf *)vma->vm_private_data; in hl_mmap_mem_buf_vm_close() 205 new_mmap_size = buf->real_mapped_size - (vma->vm_end - vma->vm_start); in hl_mmap_mem_buf_vm_close() 214 vma->vm_private_data = NULL; in hl_mmap_mem_buf_vm_close() 241 handle = vma->vm_pgoff << PAGE_SHIFT; 242 vma->vm_pgoff = 0; 253 user_mem_size = vma->vm_end - vma->vm_start; 270 buf->behavior->topic, vma->vm_start); 284 vma->vm_ops = &hl_mmap_mem_buf_vm_ops; 288 vma->vm_private_data = buf; 290 rc = buf->behavior->mmap(buf, vma, args); [all …]
|
/openbmc/linux/drivers/gpu/drm/virtio/ |
H A D | virtgpu_vram.c | 33 struct vm_area_struct *vma) in virtio_gpu_vram_mmap() argument 39 unsigned long vm_size = vma->vm_end - vma->vm_start; in virtio_gpu_vram_mmap() 48 vma->vm_pgoff -= drm_vma_node_start(&obj->vma_node); in virtio_gpu_vram_mmap() 49 vm_flags_set(vma, VM_MIXEDMAP | VM_DONTEXPAND); in virtio_gpu_vram_mmap() 50 vma->vm_page_prot = vm_get_page_prot(vma->vm_flags); in virtio_gpu_vram_mmap() 51 vma->vm_page_prot = pgprot_decrypted(vma->vm_page_prot); in virtio_gpu_vram_mmap() 52 vma->vm_ops = &virtio_gpu_vram_vm_ops; in virtio_gpu_vram_mmap() 55 vma->vm_page_prot = pgprot_writecombine(vma->vm_page_prot); in virtio_gpu_vram_mmap() 57 vma->vm_page_prot = pgprot_noncached(vma->vm_page_prot); in virtio_gpu_vram_mmap() 63 ret = io_remap_pfn_range(vma, vma->vm_start, in virtio_gpu_vram_mmap() [all …]
|
/openbmc/linux/drivers/infiniband/sw/rdmavt/ |
H A D | mmap.c | 42 static void rvt_vma_open(struct vm_area_struct *vma) in rvt_vma_open() argument 44 struct rvt_mmap_info *ip = vma->vm_private_data; in rvt_vma_open() 49 static void rvt_vma_close(struct vm_area_struct *vma) in rvt_vma_close() argument 51 struct rvt_mmap_info *ip = vma->vm_private_data; in rvt_vma_close() 68 int rvt_mmap(struct ib_ucontext *context, struct vm_area_struct *vma) in rvt_mmap() argument 71 unsigned long offset = vma->vm_pgoff << PAGE_SHIFT; in rvt_mmap() 72 unsigned long size = vma->vm_end - vma->vm_start; in rvt_mmap() 94 ret = remap_vmalloc_range(vma, ip->obj, 0); in rvt_mmap() 97 vma->vm_ops = &rvt_vm_ops; in rvt_mmap() 98 vma->vm_private_data = ip; in rvt_mmap() [all …]
|
/openbmc/linux/arch/mips/include/asm/ |
H A D | tlbflush.h | 17 extern void local_flush_tlb_range(struct vm_area_struct *vma, 21 extern void local_flush_tlb_page(struct vm_area_struct *vma, 31 extern void flush_tlb_range(struct vm_area_struct *vma, unsigned long, 41 #define flush_tlb_range(vma, vmaddr, end) local_flush_tlb_range(vma, vmaddr, end) argument 44 #define flush_tlb_page(vma, page) local_flush_tlb_page(vma, page) argument
|
/openbmc/linux/arch/alpha/include/asm/ |
H A D | tlbflush.h | 39 struct vm_area_struct *vma, in ev4_flush_tlb_current_page() argument 43 if (vma->vm_flags & VM_EXEC) { in ev4_flush_tlb_current_page() 52 struct vm_area_struct *vma, in ev5_flush_tlb_current_page() argument 55 if (vma->vm_flags & VM_EXEC) in ev5_flush_tlb_current_page() 117 flush_tlb_page(struct vm_area_struct *vma, unsigned long addr) in flush_tlb_page() argument 119 struct mm_struct *mm = vma->vm_mm; in flush_tlb_page() 122 flush_tlb_current_page(mm, vma, addr); in flush_tlb_page() 130 flush_tlb_range(struct vm_area_struct *vma, unsigned long start, in flush_tlb_range() argument 133 flush_tlb_mm(vma->vm_mm); in flush_tlb_range()
|
/openbmc/linux/arch/arm64/include/asm/ |
H A D | hugetlb.h | 33 extern int huge_ptep_set_access_flags(struct vm_area_struct *vma, 43 extern pte_t huge_ptep_clear_flush(struct vm_area_struct *vma, 54 extern pte_t huge_ptep_modify_prot_start(struct vm_area_struct *vma, 58 extern void huge_ptep_modify_prot_commit(struct vm_area_struct *vma, 65 static inline void flush_hugetlb_tlb_range(struct vm_area_struct *vma, in flush_hugetlb_tlb_range() argument 69 unsigned long stride = huge_page_size(hstate_vma(vma)); in flush_hugetlb_tlb_range() 72 __flush_tlb_range(vma, start, end, stride, false, 2); in flush_hugetlb_tlb_range() 74 __flush_tlb_range(vma, start, end, stride, false, 1); in flush_hugetlb_tlb_range() 76 __flush_tlb_range(vma, start, end, PAGE_SIZE, false, 0); in flush_hugetlb_tlb_range()
|
/openbmc/linux/kernel/dma/ |
H A D | ops_helpers.c | 34 int dma_common_mmap(struct device *dev, struct vm_area_struct *vma, in dma_common_mmap() argument 39 unsigned long user_count = vma_pages(vma); in dma_common_mmap() 41 unsigned long off = vma->vm_pgoff; in dma_common_mmap() 45 vma->vm_page_prot = dma_pgprot(dev, vma->vm_page_prot, attrs); in dma_common_mmap() 47 if (dma_mmap_from_dev_coherent(dev, vma, cpu_addr, size, &ret)) in dma_common_mmap() 53 return remap_pfn_range(vma, vma->vm_start, in dma_common_mmap() 54 page_to_pfn(page) + vma->vm_pgoff, in dma_common_mmap() 55 user_count << PAGE_SHIFT, vma->vm_page_prot); in dma_common_mmap()
|
/openbmc/linux/tools/testing/selftests/bpf/progs/ |
H A D | bpf_iter_vma_offset.c | 18 struct vm_area_struct *vma = ctx->vma; in get_vma_offset() local 22 if (task == NULL || vma == NULL) in get_vma_offset() 32 if (vma->vm_start <= address && vma->vm_end > address) { in get_vma_offset() 33 offset = address - vma->vm_start + (vma->vm_pgoff << page_shift); in get_vma_offset()
|