/openbmc/linux/arch/parisc/include/asm/ |
H A D | tlbflush.h | 19 #define flush_tlb_range(vma, start, end) \ argument 20 __flush_tlb_range((vma)->vm_mm->context.space_id, start, end) 64 static inline void flush_tlb_page(struct vm_area_struct *vma, in flush_tlb_page() argument 67 purge_tlb_entries(vma->vm_mm, addr); in flush_tlb_page()
|
/openbmc/linux/arch/ia64/include/asm/ |
H A D | tlbflush.h | 95 extern void flush_tlb_range (struct vm_area_struct *vma, unsigned long start, unsigned long end); 101 flush_tlb_page (struct vm_area_struct *vma, unsigned long addr) in flush_tlb_page() argument 104 flush_tlb_range(vma, (addr & PAGE_MASK), (addr & PAGE_MASK) + PAGE_SIZE); in flush_tlb_page() 106 if (vma->vm_mm == current->active_mm) in flush_tlb_page() 109 vma->vm_mm->context = 0; in flush_tlb_page()
|
/openbmc/linux/arch/xtensa/include/asm/ |
H A D | tlbflush.h | 35 void local_flush_tlb_page(struct vm_area_struct *vma, 37 void local_flush_tlb_range(struct vm_area_struct *vma, 54 #define flush_tlb_page(vma, page) local_flush_tlb_page(vma, page) argument 55 #define flush_tlb_range(vma, vmaddr, end) local_flush_tlb_range(vma, vmaddr, \ argument
|
H A D | cacheflush.h | 132 void local_flush_cache_range(struct vm_area_struct *vma, 134 void local_flush_cache_page(struct vm_area_struct *vma, 151 #define flush_cache_page(vma, addr, pfn) do { } while (0) argument 152 #define flush_cache_range(vma, start, end) do { } while (0) argument 177 #define copy_to_user_page(vma, page, vaddr, dst, src, len) \ argument 184 #define copy_from_user_page(vma, page, vaddr, dst, src, len) \ argument
|
/openbmc/linux/arch/powerpc/mm/ |
H A D | copro_fault.c | 26 struct vm_area_struct *vma; in copro_handle_mm_fault() local 36 vma = lock_mm_and_find_vma(mm, ea, NULL); in copro_handle_mm_fault() 37 if (!vma) in copro_handle_mm_fault() 43 if (!(vma->vm_flags & VM_WRITE)) in copro_handle_mm_fault() 46 if (!(vma->vm_flags & (VM_READ | VM_EXEC))) in copro_handle_mm_fault() 59 *flt = handle_mm_fault(vma, ea, is_write ? FAULT_FLAG_WRITE : 0, NULL); in copro_handle_mm_fault()
|
H A D | pgtable.c | 144 static pte_t set_access_flags_filter(pte_t pte, struct vm_area_struct *vma, in set_access_flags_filter() argument 168 if (WARN_ON(!(vma->vm_flags & VM_EXEC))) in set_access_flags_filter() 250 entry = set_access_flags_filter(entry, vma, dirty); in ptep_set_access_flags() 253 assert_pte_locked(vma->vm_mm, address); in ptep_set_access_flags() 254 __ptep_set_access_flags(vma, ptep, entry, in ptep_set_access_flags() 261 int huge_ptep_set_access_flags(struct vm_area_struct *vma, in huge_ptep_set_access_flags() argument 271 ptep_set_access_flags(vma, addr, ptep, pte, dirty); in huge_ptep_set_access_flags() 276 pte = set_access_flags_filter(pte, vma, dirty); in huge_ptep_set_access_flags() 281 struct hstate *h = hstate_vma(vma); in huge_ptep_set_access_flags() 285 assert_spin_locked(huge_pte_lockptr(h, vma->vm_mm, ptep)); in huge_ptep_set_access_flags() [all …]
|
/openbmc/linux/arch/nios2/mm/ |
H A D | fault.c | 46 struct vm_area_struct *vma = NULL; in do_page_fault() local 90 vma = lock_mm_and_find_vma(mm, address, regs); in do_page_fault() 91 if (!vma) in do_page_fault() 105 if (!(vma->vm_flags & VM_EXEC)) in do_page_fault() 109 if (!(vma->vm_flags & VM_READ)) in do_page_fault() 113 if (!(vma->vm_flags & VM_WRITE)) in do_page_fault() 124 fault = handle_mm_fault(vma, address, flags, regs); in do_page_fault()
|
/openbmc/linux/arch/csky/mm/ |
H A D | fault.c | 171 static inline bool access_error(struct pt_regs *regs, struct vm_area_struct *vma) in access_error() argument 174 if (!(vma->vm_flags & VM_WRITE)) in access_error() 177 if (unlikely(!vma_is_accessible(vma))) in access_error() 190 struct vm_area_struct *vma; in do_page_fault() local 240 vma = lock_mm_and_find_vma(mm, addr, regs); in do_page_fault() 241 if (unlikely(!vma)) { in do_page_fault() 252 if (unlikely(access_error(regs, vma))) { in do_page_fault() 263 fault = handle_mm_fault(vma, addr, flags, regs); in do_page_fault()
|
/openbmc/linux/arch/loongarch/mm/ |
H A D | tlb.c | 56 void local_flush_tlb_range(struct vm_area_struct *vma, unsigned long start, in local_flush_tlb_range() argument 59 struct mm_struct *mm = vma->vm_mm; in local_flush_tlb_range() 112 void local_flush_tlb_page(struct vm_area_struct *vma, unsigned long page) in local_flush_tlb_page() argument 116 if (asid_valid(vma->vm_mm, cpu)) { in local_flush_tlb_page() 119 newpid = cpu_asid(cpu, vma->vm_mm); in local_flush_tlb_page() 123 cpumask_clear_cpu(cpu, mm_cpumask(vma->vm_mm)); in local_flush_tlb_page() 137 static void __update_hugetlb(struct vm_area_struct *vma, unsigned long address, pte_t *ptep) in __update_hugetlb() argument 165 void __update_tlb(struct vm_area_struct *vma, unsigned long address, pte_t *ptep) in __update_tlb() argument 176 if (current->active_mm != vma->vm_mm) in __update_tlb() 180 __update_hugetlb(vma, address, ptep); in __update_tlb()
|
/openbmc/linux/arch/sh/kernel/cpu/sh4/ |
H A D | sq.c | 104 struct vm_struct *vma; in __sq_remap() local 106 vma = __get_vm_area_caller(map->size, VM_IOREMAP, map->sq_addr, in __sq_remap() 108 if (!vma) in __sq_remap() 111 vma->phys_addr = map->addr; in __sq_remap() 113 if (ioremap_page_range((unsigned long)vma->addr, in __sq_remap() 114 (unsigned long)vma->addr + map->size, in __sq_remap() 115 vma->phys_addr, prot)) { in __sq_remap() 116 vunmap(vma->addr); in __sq_remap() 230 struct vm_struct *vma; in sq_unmap() local 232 vma = remove_vm_area((void *)(map->sq_addr & PAGE_MASK)); in sq_unmap() [all …]
|
/openbmc/linux/arch/hexagon/mm/ |
H A D | vm_tlb.c | 25 void flush_tlb_range(struct vm_area_struct *vma, unsigned long start, in flush_tlb_range() argument 28 struct mm_struct *mm = vma->vm_mm; in flush_tlb_range() 66 void flush_tlb_page(struct vm_area_struct *vma, unsigned long vaddr) in flush_tlb_page() argument 68 struct mm_struct *mm = vma->vm_mm; in flush_tlb_page()
|
/openbmc/linux/drivers/gpu/drm/i915/gt/ |
H A D | selftest_rps.c | 70 struct i915_vma *vma; in create_spin_counter() local 83 if (IS_ERR(vma)) { in create_spin_counter() 84 err = PTR_ERR(vma); in create_spin_counter() 92 i915_vma_lock(vma); in create_spin_counter() 140 return vma; in create_spin_counter() 143 i915_vma_unpin(vma); in create_spin_counter() 145 i915_vma_unlock(vma); in create_spin_counter() 644 if (IS_ERR(vma)) { in live_rps_frequency_cs() 722 i915_vma_put(vma); in live_rps_frequency_cs() 783 if (IS_ERR(vma)) { in live_rps_frequency_srm() [all …]
|
H A D | selftest_engine_cs.c | 80 struct i915_vma *vma; in create_empty_batch() local 98 vma = i915_vma_instance(obj, ce->vm, NULL); in create_empty_batch() 99 if (IS_ERR(vma)) { in create_empty_batch() 100 err = PTR_ERR(vma); in create_empty_batch() 104 err = i915_vma_pin(vma, 0, 0, PIN_USER); in create_empty_batch() 109 return vma; in create_empty_batch() 219 struct i915_vma *vma; in create_nop_batch() local 239 if (IS_ERR(vma)) { in create_nop_batch() 240 err = PTR_ERR(vma); in create_nop_batch() 244 err = i915_vma_pin(vma, 0, 0, PIN_USER); in create_nop_batch() [all …]
|
/openbmc/linux/arch/x86/kernel/cpu/sgx/ |
H A D | encl.c | 433 struct vm_area_struct *vma = vmf->vma; in sgx_vma_fault() local 439 encl = vma->vm_private_data; in sgx_vma_fault() 458 return sgx_encl_eaug_page(vma, encl, addr); in sgx_vma_fault() 499 if (sgx_encl_mm_add(encl, vma->vm_mm)) in sgx_vma_open() 500 vma->vm_private_data = NULL; in sgx_vma_open() 652 vma->vm_flags); in sgx_vma_access() 1140 struct vm_area_struct *vma; in sgx_encl_test_and_clear_young() local 1143 ret = sgx_encl_find(mm, addr, &vma); in sgx_encl_test_and_clear_young() 1147 if (encl != vma->vm_private_data) in sgx_encl_test_and_clear_young() 1203 struct vm_area_struct *vma; in sgx_zap_enclave_ptes() local [all …]
|
/openbmc/linux/fs/kernfs/ |
H A D | file.c | 354 struct file *file = vma->vm_file; in kernfs_vma_open() 364 of->vm_ops->open(vma); in kernfs_vma_open() 414 struct file *file = vma->vm_file; in kernfs_vma_access() 462 return vma->vm_policy; in kernfs_vma_get_policy() 465 return vma->vm_policy; in kernfs_vma_get_policy() 467 pol = vma->vm_policy; in kernfs_vma_get_policy() 511 rc = ops->mmap(of, vma); in kernfs_fop_mmap() 520 if (vma->vm_file != file) in kernfs_fop_mmap() 531 if (vma->vm_ops && vma->vm_ops->close) in kernfs_fop_mmap() 538 of->vm_ops = vma->vm_ops; in kernfs_fop_mmap() [all …]
|
/openbmc/linux/arch/riscv/mm/ |
H A D | tlbflush.c | 136 void flush_tlb_page(struct vm_area_struct *vma, unsigned long addr) in flush_tlb_page() argument 138 __flush_tlb_range(vma->vm_mm, addr, PAGE_SIZE, PAGE_SIZE); in flush_tlb_page() 141 void flush_tlb_range(struct vm_area_struct *vma, unsigned long start, in flush_tlb_range() argument 144 __flush_tlb_range(vma->vm_mm, start, end - start, PAGE_SIZE); in flush_tlb_range() 153 void flush_pmd_tlb_range(struct vm_area_struct *vma, unsigned long start, in flush_pmd_tlb_range() argument 156 __flush_tlb_range(vma->vm_mm, start, end - start, PMD_SIZE); in flush_pmd_tlb_range()
|
H A D | hugetlbpage.c | 31 struct vm_area_struct *vma, in huge_pte_alloc() argument 57 if (want_pmd_share(vma, addr) && pud_none(pudp_get(pud))) in huge_pte_alloc() 58 pte = huge_pmd_share(mm, vma, addr, pud); in huge_pte_alloc() 175 struct vm_area_struct vma = TLB_FLUSH_VMA(mm, 0); in get_clear_contig_flush() local 179 flush_tlb_range(&vma, addr, addr + (PAGE_SIZE * pte_num)); in get_clear_contig_flush() 206 struct vm_area_struct vma = TLB_FLUSH_VMA(mm, 0); in clear_flush() local 212 flush_tlb_range(&vma, saddr, addr); in clear_flush() 263 int huge_ptep_set_access_flags(struct vm_area_struct *vma, in huge_ptep_set_access_flags() argument 269 struct mm_struct *mm = vma->vm_mm; in huge_ptep_set_access_flags() 334 pte_t huge_ptep_clear_flush(struct vm_area_struct *vma, in huge_ptep_clear_flush() argument [all …]
|
/openbmc/linux/drivers/xen/ |
H A D | gntalloc.c | 448 static void gntalloc_vma_open(struct vm_area_struct *vma) in gntalloc_vma_open() argument 450 struct gntalloc_vma_private_data *priv = vma->vm_private_data; in gntalloc_vma_open() 460 static void gntalloc_vma_close(struct vm_area_struct *vma) in gntalloc_vma_close() argument 496 int count = vma_pages(vma); in gntalloc_mmap() 499 if (!(vma->vm_flags & VM_SHARED)) { in gntalloc_mmap() 511 priv, vm_priv, vma->vm_pgoff, count); in gntalloc_mmap() 513 gref = find_grefs(priv, vma->vm_pgoff << PAGE_SHIFT, count); in gntalloc_mmap() 526 vma->vm_private_data = vm_priv; in gntalloc_mmap() 528 vm_flags_set(vma, VM_DONTEXPAND | VM_DONTDUMP); in gntalloc_mmap() 530 vma->vm_ops = &gntalloc_vmops; in gntalloc_mmap() [all …]
|
/openbmc/linux/include/linux/ |
H A D | mm_inline.h | 365 extern struct anon_vma_name *anon_vma_name(struct vm_area_struct *vma); 403 static inline void free_anon_vma_name(struct vm_area_struct *vma) in free_anon_vma_name() argument 409 anon_vma_name_put(vma->anon_name); in free_anon_vma_name() 423 static inline struct anon_vma_name *anon_vma_name(struct vm_area_struct *vma) in anon_vma_name() argument 437 static inline void free_anon_vma_name(struct vm_area_struct *vma) {} in free_anon_vma_name() argument 565 pte_install_uffd_wp_if_needed(struct vm_area_struct *vma, unsigned long addr, in pte_install_uffd_wp_if_needed() argument 580 if (vma_is_anonymous(vma) || !userfaultfd_wp(vma)) in pte_install_uffd_wp_if_needed() 595 set_pte_at(vma->vm_mm, addr, pte, in pte_install_uffd_wp_if_needed() 600 static inline bool vma_has_recency(struct vm_area_struct *vma) in vma_has_recency() argument 602 if (vma->vm_flags & (VM_SEQ_READ | VM_RAND_READ)) in vma_has_recency() [all …]
|
/openbmc/linux/drivers/gpu/drm/ |
H A D | drm_gem_dma_helper.c | 522 int drm_gem_dma_mmap(struct drm_gem_dma_object *dma_obj, struct vm_area_struct *vma) in drm_gem_dma_mmap() argument 532 vma->vm_pgoff -= drm_vma_node_start(&obj->vma_node); in drm_gem_dma_mmap() 533 vm_flags_mod(vma, VM_DONTEXPAND, VM_PFNMAP); in drm_gem_dma_mmap() 536 vma->vm_page_prot = vm_get_page_prot(vma->vm_flags); in drm_gem_dma_mmap() 539 vma, vma->vm_end - vma->vm_start, in drm_gem_dma_mmap() 542 ret = dma_mmap_wc(dma_obj->base.dev->dev, vma, dma_obj->vaddr, in drm_gem_dma_mmap() 544 vma->vm_end - vma->vm_start); in drm_gem_dma_mmap() 547 drm_gem_vm_close(vma); in drm_gem_dma_mmap()
|
/openbmc/linux/drivers/gpu/drm/i915/ |
H A D | i915_gpu_error.c | 517 vma = vma->next; in __find_vma() 619 if (!vma) in intel_gpu_error_print_vma() 827 for (vma = ee->vma; vma; vma = vma->next) in err_print_gt_engines() 1461 if (!vma) in capture_vma() 1494 if (vma) { in add_vma() 1495 vma->next = ee->vma; in add_vma() 1496 ee->vma = vma; in add_vma() 1508 if (!vma) in create_vma_coredump() 1565 vma = capture_vma(vma, ce->ring->vma, "ring", gfp); in engine_coredump_add_context() 1579 if (!vma) in intel_engine_coredump_add_request() [all …]
|
/openbmc/linux/drivers/video/fbdev/ |
H A D | sbuslib.c | 45 struct vm_area_struct *vma) in sbusfb_mmap_helper() argument 52 if (!(vma->vm_flags & (VM_SHARED | VM_MAYSHARE))) in sbusfb_mmap_helper() 55 size = vma->vm_end - vma->vm_start; in sbusfb_mmap_helper() 56 if (vma->vm_pgoff > (~0UL >> PAGE_SHIFT)) in sbusfb_mmap_helper() 59 off = vma->vm_pgoff << PAGE_SHIFT; in sbusfb_mmap_helper() 63 vma->vm_page_prot = pgprot_noncached(vma->vm_page_prot); in sbusfb_mmap_helper() 85 r = io_remap_pfn_range(vma, in sbusfb_mmap_helper() 86 vma->vm_start + page, in sbusfb_mmap_helper() 90 vma->vm_page_prot); in sbusfb_mmap_helper()
|
/openbmc/linux/drivers/video/fbdev/core/ |
H A D | fb_defio.c | 98 struct fb_info *info = vmf->vma->vm_private_data; in fb_deferred_io_fault() 110 if (vmf->vma->vm_file) in fb_deferred_io_fault() 111 page->mapping = vmf->vma->vm_file->f_mapping; in fb_deferred_io_fault() 202 file_update_time(vmf->vma->vm_file); in fb_deferred_io_page_mkwrite() 210 struct fb_info *info = vmf->vma->vm_private_data; in fb_deferred_io_mkwrite() 224 int fb_deferred_io_mmap(struct fb_info *info, struct vm_area_struct *vma) in fb_deferred_io_mmap() argument 226 vma->vm_ops = &fb_deferred_io_vm_ops; in fb_deferred_io_mmap() 227 vm_flags_set(vma, VM_DONTEXPAND | VM_DONTDUMP); in fb_deferred_io_mmap() 229 vm_flags_set(vma, VM_IO); in fb_deferred_io_mmap() 230 vma->vm_private_data = info; in fb_deferred_io_mmap()
|
/openbmc/linux/arch/mips/mm/ |
H A D | c-octeon.c | 63 static void octeon_flush_icache_all_cores(struct vm_area_struct *vma) in octeon_flush_icache_all_cores() argument 81 if (vma) in octeon_flush_icache_all_cores() 82 mask = *mm_cpumask(vma->vm_mm); in octeon_flush_icache_all_cores() 139 static void octeon_flush_cache_range(struct vm_area_struct *vma, in octeon_flush_cache_range() argument 142 if (vma->vm_flags & VM_EXEC) in octeon_flush_cache_range() 143 octeon_flush_icache_all_cores(vma); in octeon_flush_cache_range() 154 static void octeon_flush_cache_page(struct vm_area_struct *vma, in octeon_flush_cache_page() argument 157 if (vma->vm_flags & VM_EXEC) in octeon_flush_cache_page() 158 octeon_flush_icache_all_cores(vma); in octeon_flush_cache_page()
|
/openbmc/linux/drivers/gpu/drm/i915/gem/selftests/ |
H A D | i915_gem_context.c | 445 struct i915_vma *vma; in gpu_fill() local 452 if (IS_ERR(vma)) in gpu_fill() 453 return PTR_ERR(vma); in gpu_fill() 472 i915_vma_unpin(vma); in gpu_fill() 958 if (IS_ERR(vma)) in emit_rpcs_query() 1027 i915_vma_unpin(vma); in emit_rpcs_query() 1547 if (IS_ERR(vma)) { in write_to_scratch() 1548 err = PTR_ERR(vma); in write_to_scratch() 1577 i915_vma_unpin(vma); in write_to_scratch() 1621 if (IS_ERR(vma)) { in read_from_scratch() [all …]
|