Home
last modified time | relevance | path

Searched refs:vma (Results 76 – 100 of 1042) sorted by relevance

12345678910>>...42

/openbmc/linux/arch/powerpc/mm/
H A Dfault.c88 struct vm_area_struct *vma) in bad_access_pkey() argument
110 pkey = vma_pkey(vma); in bad_access_pkey()
231 struct vm_area_struct *vma) in access_pkey_error() argument
257 return !(vma->vm_flags & VM_EXEC) && in access_error()
273 if (unlikely(!vma_is_accessible(vma))) in access_error()
408 struct vm_area_struct * vma; in ___do_page_fault() local
476 if (!vma) in ___do_page_fault()
481 vma_end_read(vma); in ___do_page_fault()
486 vma_end_read(vma); in ___do_page_fault()
492 vma_end_read(vma); in ___do_page_fault()
[all …]
/openbmc/linux/mm/
H A Dmigrate_device.c44 if (!vma_is_anonymous(walk->vma)) in migrate_vma_collect_hole()
63 struct vm_area_struct *vma = walk->vma; in migrate_vma_collect_pmd() local
64 struct mm_struct *mm = vma->vm_mm; in migrate_vma_collect_pmd()
85 split_huge_pmd(vma, pmdp, addr); in migrate_vma_collect_pmd()
117 if (vma_is_anonymous(vma)) { in migrate_vma_collect_pmd()
520 if (!args->vma || is_vm_hugetlb_page(args->vma) || in migrate_vma_setup()
521 (args->vma->vm_flags & VM_SPECIAL) || vma_is_dax(args->vma)) in migrate_vma_setup()
528 if (args->end <= args->vma->vm_start || args->end > args->vma->vm_end) in migrate_vma_setup()
567 struct vm_area_struct *vma = migrate->vma; in migrate_vma_insert_page() local
580 if (!vma_is_anonymous(vma)) in migrate_vma_insert_page()
[all …]
H A Dksm.c562 if (!vma || !(vma->vm_flags & VM_MERGEABLE) || !vma->anon_vma) in find_mergeable_vma()
564 return vma; in find_mergeable_vma()
581 if (vma) in break_cow()
595 if (!vma) in get_mergeable_page()
1046 vma->vm_start, vma->vm_end, false); in unmerge_and_remove_all_rmap_items()
1362 if (!vma) in try_to_merge_with_ksm_page()
2194 if (vma) { in cmp_and_merge_page()
2553 err = unmerge_ksm_pages(vma, vma->vm_start, vma->vm_end, true); in __ksm_del_vma()
2871 vma = vmac->vma; in rmap_walk_ksm()
2876 if (addr < vma->vm_start || addr >= vma->vm_end) in rmap_walk_ksm()
[all …]
/openbmc/linux/include/linux/
H A Dpgtable.h357 int ptep_clear_flush_young(struct vm_area_struct *vma,
737 set_pmd_at(vma->vm_mm, address, pmdp, pmd); in generic_pmdp_establish()
889 struct vm_area_struct *vma, in arch_do_swap_page() argument
907 struct vm_area_struct *vma, in arch_unmap_one() argument
956 #define flush_tlb_fix_spurious_fault(vma, address, ptep) flush_tlb_page(vma, address) argument
1065 return ptep_get_and_clear(vma->vm_mm, addr, ptep); in __ptep_modify_prot_start()
1076 set_pte_at(vma->vm_mm, addr, ptep, pte); in __ptep_modify_prot_commit()
1098 return __ptep_modify_prot_start(vma, addr, ptep); in ptep_modify_prot_start()
1109 __ptep_modify_prot_commit(vma, addr, ptep, pte); in ptep_modify_prot_commit()
1519 #define flush_pmd_tlb_range(vma, addr, end) flush_tlb_range(vma, addr, end) argument
[all …]
H A Drmap.h83 struct vm_area_struct *vma; member
156 if (likely(vma->anon_vma)) in anon_vma_prepare()
159 return __anon_vma_prepare(vma); in anon_vma_prepare()
165 VM_BUG_ON_VMA(vma->anon_vma != next->anon_vma, vma); in anon_vma_merge()
246 struct vm_area_struct *vma) in page_try_dup_anon_rmap() argument
265 unlikely(page_needs_cow_for_dma(vma, page))) in page_try_dup_anon_rmap()
387 struct vm_area_struct *vma; member
400 .vma = _vma, \
410 .vma = _vma, \
440 struct vm_area_struct *vma);
[all …]
/openbmc/linux/drivers/gpu/drm/i915/gem/
H A Di915_gem_execbuffer.c441 struct i915_vma *vma = ev->vma; in eb_pin_vma() local
569 ev->vma = vma; in eb_add_vma()
659 struct i915_vma *vma = ev->vma; in eb_reserve_vma() local
899 vma = i915_vma_tryget(vma); in eb_lookup_vma()
976 eb->vma[i + 1].vma = NULL; in eb_lookup_vmas()
990 eb->vma[i].vma = NULL; in eb_lookup_vmas()
1001 struct i915_vma *vma = ev->vma; in eb_lock_vmas() local
1025 struct i915_vma *vma = ev->vma; in eb_validate_vmas() local
1089 struct i915_vma *vma = ev->vma; in eb_release_vmas() local
2009 struct i915_vma *vma = ev->vma; in eb_capture_stage() local
[all …]
/openbmc/linux/arch/ia64/include/asm/
H A Dfb.h13 static inline void fb_pgprotect(struct file *file, struct vm_area_struct *vma, in fb_pgprotect() argument
16 if (efi_range_is_wc(vma->vm_start, vma->vm_end - vma->vm_start)) in fb_pgprotect()
17 vma->vm_page_prot = pgprot_writecombine(vma->vm_page_prot); in fb_pgprotect()
19 vma->vm_page_prot = pgprot_noncached(vma->vm_page_prot); in fb_pgprotect()
/openbmc/linux/arch/loongarch/kernel/
H A Dvdso.c51 struct vm_area_struct *vma, struct vm_fault *vmf) in vvar_fault() argument
54 struct page *timens_page = find_timens_vvar_page(vma); in vvar_fault()
84 return vmf_insert_pfn(vma, vmf->address, pfn); in vvar_fault()
136 struct vm_area_struct *vma; in vdso_join_timens() local
141 for_each_vma(vmi, vma) { in vdso_join_timens()
143 zap_vma_pages(vma); in vdso_join_timens()
168 struct vm_area_struct *vma; in arch_setup_additional_pages() local
189 if (IS_ERR(vma)) { in arch_setup_additional_pages()
190 ret = PTR_ERR(vma); in arch_setup_additional_pages()
198 if (IS_ERR(vma)) { in arch_setup_additional_pages()
[all …]
/openbmc/linux/drivers/infiniband/sw/rxe/
H A Drxe_mmap.c37 static void rxe_vma_open(struct vm_area_struct *vma) in rxe_vma_open() argument
39 struct rxe_mmap_info *ip = vma->vm_private_data; in rxe_vma_open()
44 static void rxe_vma_close(struct vm_area_struct *vma) in rxe_vma_close() argument
46 struct rxe_mmap_info *ip = vma->vm_private_data; in rxe_vma_close()
62 int rxe_mmap(struct ib_ucontext *context, struct vm_area_struct *vma) in rxe_mmap() argument
65 unsigned long offset = vma->vm_pgoff << PAGE_SHIFT; in rxe_mmap()
66 unsigned long size = vma->vm_end - vma->vm_start; in rxe_mmap()
99 ret = remap_vmalloc_range(vma, ip->obj, 0); in rxe_mmap()
105 vma->vm_ops = &rxe_vm_ops; in rxe_mmap()
106 vma->vm_private_data = ip; in rxe_mmap()
[all …]
/openbmc/linux/drivers/gpu/drm/i915/gem/selftests/
H A Di915_gem_coherency.c88 struct i915_vma *vma; in gtt_set() local
99 if (IS_ERR(vma)) in gtt_set()
100 return PTR_ERR(vma); in gtt_set()
105 i915_vma_unpin(vma); in gtt_set()
121 struct i915_vma *vma; in gtt_get() local
132 if (IS_ERR(vma)) in gtt_get()
133 return PTR_ERR(vma); in gtt_get()
138 i915_vma_unpin(vma); in gtt_get()
199 struct i915_vma *vma; in gpu_set() local
204 if (IS_ERR(vma)) in gpu_set()
[all …]
/openbmc/linux/drivers/char/
H A Duv_mmtimer.c43 static int uv_mmtimer_mmap(struct file *file, struct vm_area_struct *vma);
147 static int uv_mmtimer_mmap(struct file *file, struct vm_area_struct *vma) in uv_mmtimer_mmap() argument
151 if (vma->vm_end - vma->vm_start != PAGE_SIZE) in uv_mmtimer_mmap()
154 if (vma->vm_flags & VM_WRITE) in uv_mmtimer_mmap()
160 vma->vm_page_prot = pgprot_noncached(vma->vm_page_prot); in uv_mmtimer_mmap()
166 if (remap_pfn_range(vma, vma->vm_start, uv_mmtimer_addr >> PAGE_SHIFT, in uv_mmtimer_mmap()
167 PAGE_SIZE, vma->vm_page_prot)) { in uv_mmtimer_mmap()
/openbmc/linux/arch/sh/include/asm/
H A Dtlbflush.h16 extern void local_flush_tlb_range(struct vm_area_struct *vma,
19 extern void local_flush_tlb_page(struct vm_area_struct *vma,
31 extern void flush_tlb_range(struct vm_area_struct *vma, unsigned long start,
33 extern void flush_tlb_page(struct vm_area_struct *vma, unsigned long page);
41 #define flush_tlb_page(vma, page) local_flush_tlb_page(vma, page) argument
44 #define flush_tlb_range(vma, start, end) \ argument
45 local_flush_tlb_range(vma, start, end)
/openbmc/linux/drivers/gpu/drm/i915/gt/
H A Dselftest_ring_submission.c12 struct i915_vma *vma; in create_wally() local
20 vma = i915_vma_instance(obj, engine->gt->vm, NULL); in create_wally()
21 if (IS_ERR(vma)) { in create_wally()
23 return vma; in create_wally()
32 err = i915_vma_sync(vma); in create_wally()
53 *cs++ = i915_vma_offset(vma) + 4000; in create_wally()
62 if (IS_ERR(vma->private)) { in create_wally()
63 vma = ERR_CAST(vma->private); in create_wally()
67 return vma; in create_wally()
223 engine->wa_ctx.vma = bb; in __live_ctx_switch_wa()
[all …]
/openbmc/linux/arch/sparc/include/asm/
H A Dcacheflush_64.h24 #define flush_cache_range(vma, start, end) \ argument
25 flush_cache_mm((vma)->vm_mm)
26 #define flush_cache_page(vma, page, pfn) \ argument
27 flush_cache_mm((vma)->vm_mm)
60 #define copy_to_user_page(vma, page, vaddr, dst, src, len) \ argument
62 flush_cache_page(vma, vaddr, page_to_pfn(page)); \
64 flush_ptrace_access(vma, page, vaddr, src, len, 0); \
67 #define copy_from_user_page(vma, page, vaddr, dst, src, len) \ argument
69 flush_cache_page(vma, vaddr, page_to_pfn(page)); \
71 flush_ptrace_access(vma, page, vaddr, dst, len, 1); \
/openbmc/linux/drivers/gpu/drm/i915/
H A Di915_debugfs.c134 if (i915_vma_is_ggtt(vma)) in stringify_vma_type()
137 if (i915_vma_is_dpt(vma)) in stringify_vma_type()
191 struct i915_vma *vma; in i915_debugfs_describe_obj() local
208 spin_lock(&obj->vma.lock); in i915_debugfs_describe_obj()
209 list_for_each_entry(vma, &obj->vma.list, obj_link) { in i915_debugfs_describe_obj()
213 spin_unlock(&obj->vma.lock); in i915_debugfs_describe_obj()
220 i915_vma_offset(vma), i915_vma_size(vma), in i915_debugfs_describe_obj()
223 if (i915_vma_is_ggtt(vma) || i915_vma_is_dpt(vma)) { in i915_debugfs_describe_obj()
268 if (vma->fence) in i915_debugfs_describe_obj()
272 spin_lock(&obj->vma.lock); in i915_debugfs_describe_obj()
[all …]
H A Di915_mm.c94 int remap_io_mapping(struct vm_area_struct *vma, in remap_io_mapping() argument
101 GEM_BUG_ON((vma->vm_flags & EXPECTED_FLAGS) != EXPECTED_FLAGS); in remap_io_mapping()
104 r.mm = vma->vm_mm; in remap_io_mapping()
107 (pgprot_val(vma->vm_page_prot) & ~_PAGE_CACHE_MASK)); in remap_io_mapping()
111 zap_vma_ptes(vma, addr, (r.pfn - pfn) << PAGE_SHIFT); in remap_io_mapping()
129 int remap_io_sg(struct vm_area_struct *vma, in remap_io_sg() argument
134 .mm = vma->vm_mm, in remap_io_sg()
135 .prot = vma->vm_page_prot, in remap_io_sg()
142 GEM_BUG_ON((vma->vm_flags & EXPECTED_FLAGS) != EXPECTED_FLAGS); in remap_io_sg()
145 flush_cache_range(vma, addr, size); in remap_io_sg()
[all …]
/openbmc/linux/drivers/dax/
H A Ddevice.c35 if (vma->vm_start & mask || vma->vm_end & mask) { in check_vma()
38 current->comm, func, vma->vm_start, vma->vm_end, in check_vma()
43 if (!vma_is_dax(vma)) { in check_vma()
88 pgoff = linear_page_index(vmf->vma, in dax_set_mapping()
161 if (pmd_addr < vmf->vma->vm_start || in __dev_dax_pmd_fault()
206 if (pud_addr < vmf->vma->vm_start || in __dev_dax_pud_fault()
240 vmf->vma->vm_start, vmf->vma->vm_end, order); in dev_dax_huge_fault()
264 struct file *filp = vma->vm_file; in dev_dax_may_split()
274 struct file *filp = vma->vm_file; in dev_dax_pagesize()
304 vma->vm_ops = &dax_vm_ops; in dax_mmap()
[all …]
/openbmc/linux/drivers/accel/habanalabs/common/
H A Dmemory_mgr.c202 (struct hl_mmap_mem_buf *)vma->vm_private_data; in hl_mmap_mem_buf_vm_close()
205 new_mmap_size = buf->real_mapped_size - (vma->vm_end - vma->vm_start); in hl_mmap_mem_buf_vm_close()
214 vma->vm_private_data = NULL; in hl_mmap_mem_buf_vm_close()
241 handle = vma->vm_pgoff << PAGE_SHIFT;
242 vma->vm_pgoff = 0;
253 user_mem_size = vma->vm_end - vma->vm_start;
270 buf->behavior->topic, vma->vm_start);
284 vma->vm_ops = &hl_mmap_mem_buf_vm_ops;
288 vma->vm_private_data = buf;
290 rc = buf->behavior->mmap(buf, vma, args);
[all …]
/openbmc/linux/drivers/gpu/drm/virtio/
H A Dvirtgpu_vram.c33 struct vm_area_struct *vma) in virtio_gpu_vram_mmap() argument
39 unsigned long vm_size = vma->vm_end - vma->vm_start; in virtio_gpu_vram_mmap()
48 vma->vm_pgoff -= drm_vma_node_start(&obj->vma_node); in virtio_gpu_vram_mmap()
49 vm_flags_set(vma, VM_MIXEDMAP | VM_DONTEXPAND); in virtio_gpu_vram_mmap()
50 vma->vm_page_prot = vm_get_page_prot(vma->vm_flags); in virtio_gpu_vram_mmap()
51 vma->vm_page_prot = pgprot_decrypted(vma->vm_page_prot); in virtio_gpu_vram_mmap()
52 vma->vm_ops = &virtio_gpu_vram_vm_ops; in virtio_gpu_vram_mmap()
55 vma->vm_page_prot = pgprot_writecombine(vma->vm_page_prot); in virtio_gpu_vram_mmap()
57 vma->vm_page_prot = pgprot_noncached(vma->vm_page_prot); in virtio_gpu_vram_mmap()
63 ret = io_remap_pfn_range(vma, vma->vm_start, in virtio_gpu_vram_mmap()
[all …]
/openbmc/linux/drivers/infiniband/sw/rdmavt/
H A Dmmap.c42 static void rvt_vma_open(struct vm_area_struct *vma) in rvt_vma_open() argument
44 struct rvt_mmap_info *ip = vma->vm_private_data; in rvt_vma_open()
49 static void rvt_vma_close(struct vm_area_struct *vma) in rvt_vma_close() argument
51 struct rvt_mmap_info *ip = vma->vm_private_data; in rvt_vma_close()
68 int rvt_mmap(struct ib_ucontext *context, struct vm_area_struct *vma) in rvt_mmap() argument
71 unsigned long offset = vma->vm_pgoff << PAGE_SHIFT; in rvt_mmap()
72 unsigned long size = vma->vm_end - vma->vm_start; in rvt_mmap()
94 ret = remap_vmalloc_range(vma, ip->obj, 0); in rvt_mmap()
97 vma->vm_ops = &rvt_vm_ops; in rvt_mmap()
98 vma->vm_private_data = ip; in rvt_mmap()
[all …]
/openbmc/linux/arch/mips/include/asm/
H A Dtlbflush.h17 extern void local_flush_tlb_range(struct vm_area_struct *vma,
21 extern void local_flush_tlb_page(struct vm_area_struct *vma,
31 extern void flush_tlb_range(struct vm_area_struct *vma, unsigned long,
41 #define flush_tlb_range(vma, vmaddr, end) local_flush_tlb_range(vma, vmaddr, end) argument
44 #define flush_tlb_page(vma, page) local_flush_tlb_page(vma, page) argument
/openbmc/linux/arch/alpha/include/asm/
H A Dtlbflush.h39 struct vm_area_struct *vma, in ev4_flush_tlb_current_page() argument
43 if (vma->vm_flags & VM_EXEC) { in ev4_flush_tlb_current_page()
52 struct vm_area_struct *vma, in ev5_flush_tlb_current_page() argument
55 if (vma->vm_flags & VM_EXEC) in ev5_flush_tlb_current_page()
117 flush_tlb_page(struct vm_area_struct *vma, unsigned long addr) in flush_tlb_page() argument
119 struct mm_struct *mm = vma->vm_mm; in flush_tlb_page()
122 flush_tlb_current_page(mm, vma, addr); in flush_tlb_page()
130 flush_tlb_range(struct vm_area_struct *vma, unsigned long start, in flush_tlb_range() argument
133 flush_tlb_mm(vma->vm_mm); in flush_tlb_range()
/openbmc/linux/arch/arm64/include/asm/
H A Dhugetlb.h33 extern int huge_ptep_set_access_flags(struct vm_area_struct *vma,
43 extern pte_t huge_ptep_clear_flush(struct vm_area_struct *vma,
54 extern pte_t huge_ptep_modify_prot_start(struct vm_area_struct *vma,
58 extern void huge_ptep_modify_prot_commit(struct vm_area_struct *vma,
65 static inline void flush_hugetlb_tlb_range(struct vm_area_struct *vma, in flush_hugetlb_tlb_range() argument
69 unsigned long stride = huge_page_size(hstate_vma(vma)); in flush_hugetlb_tlb_range()
72 __flush_tlb_range(vma, start, end, stride, false, 2); in flush_hugetlb_tlb_range()
74 __flush_tlb_range(vma, start, end, stride, false, 1); in flush_hugetlb_tlb_range()
76 __flush_tlb_range(vma, start, end, PAGE_SIZE, false, 0); in flush_hugetlb_tlb_range()
/openbmc/linux/kernel/dma/
H A Dops_helpers.c34 int dma_common_mmap(struct device *dev, struct vm_area_struct *vma, in dma_common_mmap() argument
39 unsigned long user_count = vma_pages(vma); in dma_common_mmap()
41 unsigned long off = vma->vm_pgoff; in dma_common_mmap()
45 vma->vm_page_prot = dma_pgprot(dev, vma->vm_page_prot, attrs); in dma_common_mmap()
47 if (dma_mmap_from_dev_coherent(dev, vma, cpu_addr, size, &ret)) in dma_common_mmap()
53 return remap_pfn_range(vma, vma->vm_start, in dma_common_mmap()
54 page_to_pfn(page) + vma->vm_pgoff, in dma_common_mmap()
55 user_count << PAGE_SHIFT, vma->vm_page_prot); in dma_common_mmap()
/openbmc/linux/tools/testing/selftests/bpf/progs/
H A Dbpf_iter_vma_offset.c18 struct vm_area_struct *vma = ctx->vma; in get_vma_offset() local
22 if (task == NULL || vma == NULL) in get_vma_offset()
32 if (vma->vm_start <= address && vma->vm_end > address) { in get_vma_offset()
33 offset = address - vma->vm_start + (vma->vm_pgoff << page_shift); in get_vma_offset()

12345678910>>...42