Home
last modified time | relevance | path

Searched refs:vma (Results 251 – 275 of 1042) sorted by relevance

1...<<11121314151617181920>>...42

/openbmc/linux/arch/parisc/include/asm/
H A Dtlbflush.h19 #define flush_tlb_range(vma, start, end) \ argument
20 __flush_tlb_range((vma)->vm_mm->context.space_id, start, end)
64 static inline void flush_tlb_page(struct vm_area_struct *vma, in flush_tlb_page() argument
67 purge_tlb_entries(vma->vm_mm, addr); in flush_tlb_page()
/openbmc/linux/arch/ia64/include/asm/
H A Dtlbflush.h95 extern void flush_tlb_range (struct vm_area_struct *vma, unsigned long start, unsigned long end);
101 flush_tlb_page (struct vm_area_struct *vma, unsigned long addr) in flush_tlb_page() argument
104 flush_tlb_range(vma, (addr & PAGE_MASK), (addr & PAGE_MASK) + PAGE_SIZE); in flush_tlb_page()
106 if (vma->vm_mm == current->active_mm) in flush_tlb_page()
109 vma->vm_mm->context = 0; in flush_tlb_page()
/openbmc/linux/arch/xtensa/include/asm/
H A Dtlbflush.h35 void local_flush_tlb_page(struct vm_area_struct *vma,
37 void local_flush_tlb_range(struct vm_area_struct *vma,
54 #define flush_tlb_page(vma, page) local_flush_tlb_page(vma, page) argument
55 #define flush_tlb_range(vma, vmaddr, end) local_flush_tlb_range(vma, vmaddr, \ argument
H A Dcacheflush.h132 void local_flush_cache_range(struct vm_area_struct *vma,
134 void local_flush_cache_page(struct vm_area_struct *vma,
151 #define flush_cache_page(vma, addr, pfn) do { } while (0) argument
152 #define flush_cache_range(vma, start, end) do { } while (0) argument
177 #define copy_to_user_page(vma, page, vaddr, dst, src, len) \ argument
184 #define copy_from_user_page(vma, page, vaddr, dst, src, len) \ argument
/openbmc/linux/arch/powerpc/mm/
H A Dcopro_fault.c26 struct vm_area_struct *vma; in copro_handle_mm_fault() local
36 vma = lock_mm_and_find_vma(mm, ea, NULL); in copro_handle_mm_fault()
37 if (!vma) in copro_handle_mm_fault()
43 if (!(vma->vm_flags & VM_WRITE)) in copro_handle_mm_fault()
46 if (!(vma->vm_flags & (VM_READ | VM_EXEC))) in copro_handle_mm_fault()
59 *flt = handle_mm_fault(vma, ea, is_write ? FAULT_FLAG_WRITE : 0, NULL); in copro_handle_mm_fault()
H A Dpgtable.c144 static pte_t set_access_flags_filter(pte_t pte, struct vm_area_struct *vma, in set_access_flags_filter() argument
168 if (WARN_ON(!(vma->vm_flags & VM_EXEC))) in set_access_flags_filter()
250 entry = set_access_flags_filter(entry, vma, dirty); in ptep_set_access_flags()
253 assert_pte_locked(vma->vm_mm, address); in ptep_set_access_flags()
254 __ptep_set_access_flags(vma, ptep, entry, in ptep_set_access_flags()
261 int huge_ptep_set_access_flags(struct vm_area_struct *vma, in huge_ptep_set_access_flags() argument
271 ptep_set_access_flags(vma, addr, ptep, pte, dirty); in huge_ptep_set_access_flags()
276 pte = set_access_flags_filter(pte, vma, dirty); in huge_ptep_set_access_flags()
281 struct hstate *h = hstate_vma(vma); in huge_ptep_set_access_flags()
285 assert_spin_locked(huge_pte_lockptr(h, vma->vm_mm, ptep)); in huge_ptep_set_access_flags()
[all …]
/openbmc/linux/arch/nios2/mm/
H A Dfault.c46 struct vm_area_struct *vma = NULL; in do_page_fault() local
90 vma = lock_mm_and_find_vma(mm, address, regs); in do_page_fault()
91 if (!vma) in do_page_fault()
105 if (!(vma->vm_flags & VM_EXEC)) in do_page_fault()
109 if (!(vma->vm_flags & VM_READ)) in do_page_fault()
113 if (!(vma->vm_flags & VM_WRITE)) in do_page_fault()
124 fault = handle_mm_fault(vma, address, flags, regs); in do_page_fault()
/openbmc/linux/arch/csky/mm/
H A Dfault.c171 static inline bool access_error(struct pt_regs *regs, struct vm_area_struct *vma) in access_error() argument
174 if (!(vma->vm_flags & VM_WRITE)) in access_error()
177 if (unlikely(!vma_is_accessible(vma))) in access_error()
190 struct vm_area_struct *vma; in do_page_fault() local
240 vma = lock_mm_and_find_vma(mm, addr, regs); in do_page_fault()
241 if (unlikely(!vma)) { in do_page_fault()
252 if (unlikely(access_error(regs, vma))) { in do_page_fault()
263 fault = handle_mm_fault(vma, addr, flags, regs); in do_page_fault()
/openbmc/linux/arch/loongarch/mm/
H A Dtlb.c56 void local_flush_tlb_range(struct vm_area_struct *vma, unsigned long start, in local_flush_tlb_range() argument
59 struct mm_struct *mm = vma->vm_mm; in local_flush_tlb_range()
112 void local_flush_tlb_page(struct vm_area_struct *vma, unsigned long page) in local_flush_tlb_page() argument
116 if (asid_valid(vma->vm_mm, cpu)) { in local_flush_tlb_page()
119 newpid = cpu_asid(cpu, vma->vm_mm); in local_flush_tlb_page()
123 cpumask_clear_cpu(cpu, mm_cpumask(vma->vm_mm)); in local_flush_tlb_page()
137 static void __update_hugetlb(struct vm_area_struct *vma, unsigned long address, pte_t *ptep) in __update_hugetlb() argument
165 void __update_tlb(struct vm_area_struct *vma, unsigned long address, pte_t *ptep) in __update_tlb() argument
176 if (current->active_mm != vma->vm_mm) in __update_tlb()
180 __update_hugetlb(vma, address, ptep); in __update_tlb()
/openbmc/linux/arch/sh/kernel/cpu/sh4/
H A Dsq.c104 struct vm_struct *vma; in __sq_remap() local
106 vma = __get_vm_area_caller(map->size, VM_IOREMAP, map->sq_addr, in __sq_remap()
108 if (!vma) in __sq_remap()
111 vma->phys_addr = map->addr; in __sq_remap()
113 if (ioremap_page_range((unsigned long)vma->addr, in __sq_remap()
114 (unsigned long)vma->addr + map->size, in __sq_remap()
115 vma->phys_addr, prot)) { in __sq_remap()
116 vunmap(vma->addr); in __sq_remap()
230 struct vm_struct *vma; in sq_unmap() local
232 vma = remove_vm_area((void *)(map->sq_addr & PAGE_MASK)); in sq_unmap()
[all …]
/openbmc/linux/arch/hexagon/mm/
H A Dvm_tlb.c25 void flush_tlb_range(struct vm_area_struct *vma, unsigned long start, in flush_tlb_range() argument
28 struct mm_struct *mm = vma->vm_mm; in flush_tlb_range()
66 void flush_tlb_page(struct vm_area_struct *vma, unsigned long vaddr) in flush_tlb_page() argument
68 struct mm_struct *mm = vma->vm_mm; in flush_tlb_page()
/openbmc/linux/drivers/gpu/drm/i915/gt/
H A Dselftest_rps.c70 struct i915_vma *vma; in create_spin_counter() local
83 if (IS_ERR(vma)) { in create_spin_counter()
84 err = PTR_ERR(vma); in create_spin_counter()
92 i915_vma_lock(vma); in create_spin_counter()
140 return vma; in create_spin_counter()
143 i915_vma_unpin(vma); in create_spin_counter()
145 i915_vma_unlock(vma); in create_spin_counter()
644 if (IS_ERR(vma)) { in live_rps_frequency_cs()
722 i915_vma_put(vma); in live_rps_frequency_cs()
783 if (IS_ERR(vma)) { in live_rps_frequency_srm()
[all …]
H A Dselftest_engine_cs.c80 struct i915_vma *vma; in create_empty_batch() local
98 vma = i915_vma_instance(obj, ce->vm, NULL); in create_empty_batch()
99 if (IS_ERR(vma)) { in create_empty_batch()
100 err = PTR_ERR(vma); in create_empty_batch()
104 err = i915_vma_pin(vma, 0, 0, PIN_USER); in create_empty_batch()
109 return vma; in create_empty_batch()
219 struct i915_vma *vma; in create_nop_batch() local
239 if (IS_ERR(vma)) { in create_nop_batch()
240 err = PTR_ERR(vma); in create_nop_batch()
244 err = i915_vma_pin(vma, 0, 0, PIN_USER); in create_nop_batch()
[all …]
/openbmc/linux/arch/x86/kernel/cpu/sgx/
H A Dencl.c433 struct vm_area_struct *vma = vmf->vma; in sgx_vma_fault() local
439 encl = vma->vm_private_data; in sgx_vma_fault()
458 return sgx_encl_eaug_page(vma, encl, addr); in sgx_vma_fault()
499 if (sgx_encl_mm_add(encl, vma->vm_mm)) in sgx_vma_open()
500 vma->vm_private_data = NULL; in sgx_vma_open()
652 vma->vm_flags); in sgx_vma_access()
1140 struct vm_area_struct *vma; in sgx_encl_test_and_clear_young() local
1143 ret = sgx_encl_find(mm, addr, &vma); in sgx_encl_test_and_clear_young()
1147 if (encl != vma->vm_private_data) in sgx_encl_test_and_clear_young()
1203 struct vm_area_struct *vma; in sgx_zap_enclave_ptes() local
[all …]
/openbmc/linux/fs/kernfs/
H A Dfile.c354 struct file *file = vma->vm_file; in kernfs_vma_open()
364 of->vm_ops->open(vma); in kernfs_vma_open()
414 struct file *file = vma->vm_file; in kernfs_vma_access()
462 return vma->vm_policy; in kernfs_vma_get_policy()
465 return vma->vm_policy; in kernfs_vma_get_policy()
467 pol = vma->vm_policy; in kernfs_vma_get_policy()
511 rc = ops->mmap(of, vma); in kernfs_fop_mmap()
520 if (vma->vm_file != file) in kernfs_fop_mmap()
531 if (vma->vm_ops && vma->vm_ops->close) in kernfs_fop_mmap()
538 of->vm_ops = vma->vm_ops; in kernfs_fop_mmap()
[all …]
/openbmc/linux/arch/riscv/mm/
H A Dtlbflush.c136 void flush_tlb_page(struct vm_area_struct *vma, unsigned long addr) in flush_tlb_page() argument
138 __flush_tlb_range(vma->vm_mm, addr, PAGE_SIZE, PAGE_SIZE); in flush_tlb_page()
141 void flush_tlb_range(struct vm_area_struct *vma, unsigned long start, in flush_tlb_range() argument
144 __flush_tlb_range(vma->vm_mm, start, end - start, PAGE_SIZE); in flush_tlb_range()
153 void flush_pmd_tlb_range(struct vm_area_struct *vma, unsigned long start, in flush_pmd_tlb_range() argument
156 __flush_tlb_range(vma->vm_mm, start, end - start, PMD_SIZE); in flush_pmd_tlb_range()
H A Dhugetlbpage.c31 struct vm_area_struct *vma, in huge_pte_alloc() argument
57 if (want_pmd_share(vma, addr) && pud_none(pudp_get(pud))) in huge_pte_alloc()
58 pte = huge_pmd_share(mm, vma, addr, pud); in huge_pte_alloc()
175 struct vm_area_struct vma = TLB_FLUSH_VMA(mm, 0); in get_clear_contig_flush() local
179 flush_tlb_range(&vma, addr, addr + (PAGE_SIZE * pte_num)); in get_clear_contig_flush()
206 struct vm_area_struct vma = TLB_FLUSH_VMA(mm, 0); in clear_flush() local
212 flush_tlb_range(&vma, saddr, addr); in clear_flush()
263 int huge_ptep_set_access_flags(struct vm_area_struct *vma, in huge_ptep_set_access_flags() argument
269 struct mm_struct *mm = vma->vm_mm; in huge_ptep_set_access_flags()
334 pte_t huge_ptep_clear_flush(struct vm_area_struct *vma, in huge_ptep_clear_flush() argument
[all …]
/openbmc/linux/drivers/xen/
H A Dgntalloc.c448 static void gntalloc_vma_open(struct vm_area_struct *vma) in gntalloc_vma_open() argument
450 struct gntalloc_vma_private_data *priv = vma->vm_private_data; in gntalloc_vma_open()
460 static void gntalloc_vma_close(struct vm_area_struct *vma) in gntalloc_vma_close() argument
496 int count = vma_pages(vma); in gntalloc_mmap()
499 if (!(vma->vm_flags & VM_SHARED)) { in gntalloc_mmap()
511 priv, vm_priv, vma->vm_pgoff, count); in gntalloc_mmap()
513 gref = find_grefs(priv, vma->vm_pgoff << PAGE_SHIFT, count); in gntalloc_mmap()
526 vma->vm_private_data = vm_priv; in gntalloc_mmap()
528 vm_flags_set(vma, VM_DONTEXPAND | VM_DONTDUMP); in gntalloc_mmap()
530 vma->vm_ops = &gntalloc_vmops; in gntalloc_mmap()
[all …]
/openbmc/linux/include/linux/
H A Dmm_inline.h365 extern struct anon_vma_name *anon_vma_name(struct vm_area_struct *vma);
403 static inline void free_anon_vma_name(struct vm_area_struct *vma) in free_anon_vma_name() argument
409 anon_vma_name_put(vma->anon_name); in free_anon_vma_name()
423 static inline struct anon_vma_name *anon_vma_name(struct vm_area_struct *vma) in anon_vma_name() argument
437 static inline void free_anon_vma_name(struct vm_area_struct *vma) {} in free_anon_vma_name() argument
565 pte_install_uffd_wp_if_needed(struct vm_area_struct *vma, unsigned long addr, in pte_install_uffd_wp_if_needed() argument
580 if (vma_is_anonymous(vma) || !userfaultfd_wp(vma)) in pte_install_uffd_wp_if_needed()
595 set_pte_at(vma->vm_mm, addr, pte, in pte_install_uffd_wp_if_needed()
600 static inline bool vma_has_recency(struct vm_area_struct *vma) in vma_has_recency() argument
602 if (vma->vm_flags & (VM_SEQ_READ | VM_RAND_READ)) in vma_has_recency()
[all …]
/openbmc/linux/drivers/gpu/drm/
H A Ddrm_gem_dma_helper.c522 int drm_gem_dma_mmap(struct drm_gem_dma_object *dma_obj, struct vm_area_struct *vma) in drm_gem_dma_mmap() argument
532 vma->vm_pgoff -= drm_vma_node_start(&obj->vma_node); in drm_gem_dma_mmap()
533 vm_flags_mod(vma, VM_DONTEXPAND, VM_PFNMAP); in drm_gem_dma_mmap()
536 vma->vm_page_prot = vm_get_page_prot(vma->vm_flags); in drm_gem_dma_mmap()
539 vma, vma->vm_end - vma->vm_start, in drm_gem_dma_mmap()
542 ret = dma_mmap_wc(dma_obj->base.dev->dev, vma, dma_obj->vaddr, in drm_gem_dma_mmap()
544 vma->vm_end - vma->vm_start); in drm_gem_dma_mmap()
547 drm_gem_vm_close(vma); in drm_gem_dma_mmap()
/openbmc/linux/drivers/gpu/drm/i915/
H A Di915_gpu_error.c517 vma = vma->next; in __find_vma()
619 if (!vma) in intel_gpu_error_print_vma()
827 for (vma = ee->vma; vma; vma = vma->next) in err_print_gt_engines()
1461 if (!vma) in capture_vma()
1494 if (vma) { in add_vma()
1495 vma->next = ee->vma; in add_vma()
1496 ee->vma = vma; in add_vma()
1508 if (!vma) in create_vma_coredump()
1565 vma = capture_vma(vma, ce->ring->vma, "ring", gfp); in engine_coredump_add_context()
1579 if (!vma) in intel_engine_coredump_add_request()
[all …]
/openbmc/linux/drivers/video/fbdev/
H A Dsbuslib.c45 struct vm_area_struct *vma) in sbusfb_mmap_helper() argument
52 if (!(vma->vm_flags & (VM_SHARED | VM_MAYSHARE))) in sbusfb_mmap_helper()
55 size = vma->vm_end - vma->vm_start; in sbusfb_mmap_helper()
56 if (vma->vm_pgoff > (~0UL >> PAGE_SHIFT)) in sbusfb_mmap_helper()
59 off = vma->vm_pgoff << PAGE_SHIFT; in sbusfb_mmap_helper()
63 vma->vm_page_prot = pgprot_noncached(vma->vm_page_prot); in sbusfb_mmap_helper()
85 r = io_remap_pfn_range(vma, in sbusfb_mmap_helper()
86 vma->vm_start + page, in sbusfb_mmap_helper()
90 vma->vm_page_prot); in sbusfb_mmap_helper()
/openbmc/linux/drivers/video/fbdev/core/
H A Dfb_defio.c98 struct fb_info *info = vmf->vma->vm_private_data; in fb_deferred_io_fault()
110 if (vmf->vma->vm_file) in fb_deferred_io_fault()
111 page->mapping = vmf->vma->vm_file->f_mapping; in fb_deferred_io_fault()
202 file_update_time(vmf->vma->vm_file); in fb_deferred_io_page_mkwrite()
210 struct fb_info *info = vmf->vma->vm_private_data; in fb_deferred_io_mkwrite()
224 int fb_deferred_io_mmap(struct fb_info *info, struct vm_area_struct *vma) in fb_deferred_io_mmap() argument
226 vma->vm_ops = &fb_deferred_io_vm_ops; in fb_deferred_io_mmap()
227 vm_flags_set(vma, VM_DONTEXPAND | VM_DONTDUMP); in fb_deferred_io_mmap()
229 vm_flags_set(vma, VM_IO); in fb_deferred_io_mmap()
230 vma->vm_private_data = info; in fb_deferred_io_mmap()
/openbmc/linux/arch/mips/mm/
H A Dc-octeon.c63 static void octeon_flush_icache_all_cores(struct vm_area_struct *vma) in octeon_flush_icache_all_cores() argument
81 if (vma) in octeon_flush_icache_all_cores()
82 mask = *mm_cpumask(vma->vm_mm); in octeon_flush_icache_all_cores()
139 static void octeon_flush_cache_range(struct vm_area_struct *vma, in octeon_flush_cache_range() argument
142 if (vma->vm_flags & VM_EXEC) in octeon_flush_cache_range()
143 octeon_flush_icache_all_cores(vma); in octeon_flush_cache_range()
154 static void octeon_flush_cache_page(struct vm_area_struct *vma, in octeon_flush_cache_page() argument
157 if (vma->vm_flags & VM_EXEC) in octeon_flush_cache_page()
158 octeon_flush_icache_all_cores(vma); in octeon_flush_cache_page()
/openbmc/linux/drivers/gpu/drm/i915/gem/selftests/
H A Di915_gem_context.c445 struct i915_vma *vma; in gpu_fill() local
452 if (IS_ERR(vma)) in gpu_fill()
453 return PTR_ERR(vma); in gpu_fill()
472 i915_vma_unpin(vma); in gpu_fill()
958 if (IS_ERR(vma)) in emit_rpcs_query()
1027 i915_vma_unpin(vma); in emit_rpcs_query()
1547 if (IS_ERR(vma)) { in write_to_scratch()
1548 err = PTR_ERR(vma); in write_to_scratch()
1577 i915_vma_unpin(vma); in write_to_scratch()
1621 if (IS_ERR(vma)) { in read_from_scratch()
[all …]

1...<<11121314151617181920>>...42