Lines Matching refs:vm_flags

84 static pgprot_t vm_pgprot_modify(pgprot_t oldprot, unsigned long vm_flags)  in vm_pgprot_modify()  argument
86 return pgprot_modify(oldprot, vm_get_page_prot(vm_flags)); in vm_pgprot_modify()
92 unsigned long vm_flags = vma->vm_flags; in vma_set_page_prot() local
95 vm_page_prot = vm_pgprot_modify(vma->vm_page_prot, vm_flags); in vma_set_page_prot()
97 vm_flags &= ~VM_SHARED; in vma_set_page_prot()
98 vm_page_prot = vm_pgprot_modify(vm_page_prot, vm_flags); in vma_set_page_prot()
110 if (vma->vm_flags & VM_SHARED) in __remove_shared_vm_struct()
386 if (vma->vm_flags & VM_SHARED) in __vma_link_file()
720 struct file *file, unsigned long vm_flags, in is_mergeable_vma() argument
732 if ((vma->vm_flags ^ vm_flags) & ~VM_SOFTDIRTY) in is_mergeable_vma()
772 can_vma_merge_before(struct vm_area_struct *vma, unsigned long vm_flags, in can_vma_merge_before() argument
777 if (is_mergeable_vma(vma, file, vm_flags, vm_userfaultfd_ctx, anon_name, true) && in can_vma_merge_before()
795 can_vma_merge_after(struct vm_area_struct *vma, unsigned long vm_flags, in can_vma_merge_after() argument
800 if (is_mergeable_vma(vma, file, vm_flags, vm_userfaultfd_ctx, anon_name, false) && in can_vma_merge_after()
864 unsigned long end, unsigned long vm_flags, in vma_merge() argument
888 if (vm_flags & VM_SPECIAL) in vma_merge()
906 && can_vma_merge_after(prev, vm_flags, anon_vma, file, in vma_merge()
915 can_vma_merge_before(next, vm_flags, anon_vma, file, pgoff+pglen, in vma_merge()
1035 khugepaged_enter_vma(res, vm_flags); in vma_merge()
1066 !((a->vm_flags ^ b->vm_flags) & ~(VM_ACCESS_FLAGS | VM_SOFTDIRTY)) && in anon_vma_compatible()
1212 unsigned long flags, vm_flags_t vm_flags, in do_mmap() argument
1276 vm_flags |= calc_vm_prot_bits(prot, pkey) | calc_vm_flag_bits(file, flags) | in do_mmap()
1283 if (!mlock_future_ok(mm, vm_flags, len)) in do_mmap()
1323 vm_flags |= VM_SHARED | VM_MAYSHARE; in do_mmap()
1325 vm_flags &= ~(VM_MAYWRITE | VM_SHARED); in do_mmap()
1331 if (vm_flags & VM_EXEC) in do_mmap()
1333 vm_flags &= ~VM_MAYEXEC; in do_mmap()
1338 if (vm_flags & (VM_GROWSDOWN|VM_GROWSUP)) in do_mmap()
1348 if (vm_flags & (VM_GROWSDOWN|VM_GROWSUP)) in do_mmap()
1354 vm_flags |= VM_SHARED | VM_MAYSHARE; in do_mmap()
1374 vm_flags |= VM_NORESERVE; in do_mmap()
1378 vm_flags |= VM_NORESERVE; in do_mmap()
1381 addr = mmap_region(file, addr, len, vm_flags, pgoff, uf); in do_mmap()
1383 ((vm_flags & VM_LOCKED) || in do_mmap()
1472 return (vma->vm_flags & (VM_WRITE | VM_SHARED)) == in vma_is_shared_writable()
1479 if (vma->vm_flags & VM_PFNMAP) in vma_fs_can_writeback()
1526 pgprot_val(vm_pgprot_modify(vm_page_prot, vma->vm_flags))) in vma_wants_writenotify()
1548 static inline int accountable_mapping(struct file *file, vm_flags_t vm_flags) in accountable_mapping() argument
1557 return (vm_flags & (VM_NORESERVE | VM_SHARED | VM_WRITE)) == VM_WRITE; in accountable_mapping()
1594 if (tmp && (tmp->vm_flags & VM_STARTGAP_FLAGS)) { /* Avoid prev check if possible */ in unmapped_area()
1646 if (tmp && (tmp->vm_flags & VM_STARTGAP_FLAGS)) { /* Avoid prev check if possible */ in unmapped_area_topdown()
1927 if (!may_expand_vm(mm, vma->vm_flags, grow)) in acct_stack_growth()
1935 if (!mlock_future_ok(mm, vma->vm_flags, grow << PAGE_SHIFT)) in acct_stack_growth()
1939 new_start = (vma->vm_flags & VM_GROWSUP) ? vma->vm_start : in acct_stack_growth()
1967 if (!(vma->vm_flags & VM_GROWSUP)) in expand_upwards()
1985 if (!(next->vm_flags & VM_GROWSUP)) in expand_upwards()
2033 if (vma->vm_flags & VM_LOCKED) in expand_upwards()
2035 vm_stat_account(mm, vma->vm_flags, grow); in expand_upwards()
2048 khugepaged_enter_vma(vma, vma->vm_flags); in expand_upwards()
2066 if (!(vma->vm_flags & VM_GROWSDOWN)) in expand_downwards()
2077 if (!(prev->vm_flags & VM_GROWSDOWN) && in expand_downwards()
2126 if (vma->vm_flags & VM_LOCKED) in expand_downwards()
2128 vm_stat_account(mm, vma->vm_flags, grow); in expand_downwards()
2142 khugepaged_enter_vma(vma, vma->vm_flags); in expand_downwards()
2182 if (prev->vm_flags & VM_LOCKED) in find_extend_vma_locked()
2189 if (unlikely(!(vma->vm_flags & VM_GROWSDOWN))) in expand_stack_locked()
2208 if (vma->vm_flags & VM_LOCKED) in find_extend_vma_locked()
2316 if (vma->vm_flags & VM_ACCOUNT) in remove_mt()
2318 vm_stat_account(mm, vma->vm_flags, -nrpages); in remove_mt()
2514 if (next->vm_flags & VM_LOCKED) in do_vmi_align_munmap()
2670 unsigned long len, vm_flags_t vm_flags, unsigned long pgoff, in __mmap_region() argument
2685 if (!may_expand_vm(mm, vm_flags, len >> PAGE_SHIFT)) { in __mmap_region()
2694 if (!may_expand_vm(mm, vm_flags, in __mmap_region()
2706 if (accountable_mapping(file, vm_flags)) { in __mmap_region()
2710 vm_flags |= VM_ACCOUNT; in __mmap_region()
2715 if (vm_flags & VM_SPECIAL) { in __mmap_region()
2724 can_vma_merge_before(next, vm_flags, NULL, file, pgoff+pglen, in __mmap_region()
2733 (vma ? can_vma_merge_after(prev, vm_flags, vma->anon_vma, file, in __mmap_region()
2735 can_vma_merge_after(prev, vm_flags, NULL, file, pgoff, in __mmap_region()
2747 khugepaged_enter_vma(vma, vm_flags); in __mmap_region()
2769 vm_flags_init(vma, vm_flags); in __mmap_region()
2770 vma->vm_page_prot = vm_get_page_prot(vm_flags); in __mmap_region()
2790 VM_WARN_ON_ONCE(vm_flags != vma->vm_flags && in __mmap_region()
2791 !(vm_flags & VM_MAYWRITE) && in __mmap_region()
2792 (vma->vm_flags & VM_MAYWRITE)); in __mmap_region()
2799 if (unlikely(vm_flags != vma->vm_flags && prev)) { in __mmap_region()
2801 vma->vm_end, vma->vm_flags, NULL, in __mmap_region()
2817 vm_flags = vma->vm_flags; in __mmap_region()
2822 vm_flags = vma->vm_flags; in __mmap_region()
2823 } else if (vm_flags & VM_SHARED) { in __mmap_region()
2833 WARN_ON_ONCE(!arch_validate_flags(vm_flags)); in __mmap_region()
2842 if (vma->vm_flags & VM_SHARED) in __mmap_region()
2855 khugepaged_enter_vma(vma, vma->vm_flags); in __mmap_region()
2863 vm_stat_account(mm, vm_flags, len >> PAGE_SHIFT); in __mmap_region()
2864 if (vm_flags & VM_LOCKED) { in __mmap_region()
2865 if ((vm_flags & VM_SPECIAL) || vma_is_dax(vma) || in __mmap_region()
2908 unsigned long len, vm_flags_t vm_flags, unsigned long pgoff, in mmap_region() argument
2915 if (map_deny_write_exec(vm_flags, vm_flags)) in mmap_region()
2919 if (!arch_validate_flags(vm_flags)) in mmap_region()
2923 if (file && (vm_flags & VM_SHARED)) { in mmap_region()
2931 ret = __mmap_region(file, addr, len, vm_flags, pgoff, uf); in mmap_region()
3005 if (!vma || !(vma->vm_flags & VM_SHARED)) in SYSCALL_DEFINE5()
3020 if (next->vm_flags != vma->vm_flags) in SYSCALL_DEFINE5()
3033 prot |= vma->vm_flags & VM_READ ? PROT_READ : 0; in SYSCALL_DEFINE5()
3034 prot |= vma->vm_flags & VM_WRITE ? PROT_WRITE : 0; in SYSCALL_DEFINE5()
3035 prot |= vma->vm_flags & VM_EXEC ? PROT_EXEC : 0; in SYSCALL_DEFINE5()
3039 if (vma->vm_flags & VM_LOCKED) in SYSCALL_DEFINE5()
3279 if (vma->vm_flags & VM_ACCOUNT) in exit_mmap()
3306 if ((vma->vm_flags & VM_ACCOUNT) && in insert_vm_struct()
3363 new_vma = vma_merge(&vmi, mm, prev, addr, addr + len, vma->vm_flags, in copy_vma()
3551 unsigned long vm_flags, void *priv, in __install_special_mapping() argument
3564 vm_flags_init(vma, (vm_flags | mm->def_flags | in __install_special_mapping()
3566 vma->vm_page_prot = vm_get_page_prot(vma->vm_flags); in __install_special_mapping()
3575 vm_stat_account(mm, vma->vm_flags, len >> PAGE_SHIFT); in __install_special_mapping()
3606 unsigned long vm_flags, const struct vm_special_mapping *spec) in _install_special_mapping() argument
3608 return __install_special_mapping(mm, addr, len, vm_flags, (void *)spec, in _install_special_mapping()
3614 unsigned long vm_flags, struct page **pages) in install_special_mapping() argument
3617 mm, addr, len, vm_flags, (void *)pages, in install_special_mapping()