Home
last modified time | relevance | path

Searched refs:locked_vm (Results 1 – 20 of 20) sorted by relevance

/openbmc/linux/mm/
H A Dutil.c487 unsigned long locked_vm, limit; in __account_locked_vm() local
492 locked_vm = mm->locked_vm; in __account_locked_vm()
496 if (locked_vm + pages > limit) in __account_locked_vm()
500 mm->locked_vm = locked_vm + pages; in __account_locked_vm()
502 WARN_ON_ONCE(pages > locked_vm); in __account_locked_vm()
503 mm->locked_vm = locked_vm - pages; in __account_locked_vm()
508 locked_vm << PAGE_SHIFT, task_rlimit(task, RLIMIT_MEMLOCK), in __account_locked_vm()
H A Ddebug.c201 mm->hiwater_rss, mm->hiwater_vm, mm->total_vm, mm->locked_vm, in dump_mm()
H A Dmmap.c1168 locked_pages += mm->locked_vm; in mlock_future_ok()
2035 mm->locked_vm += grow; in expand_upwards()
2128 mm->locked_vm += grow; in expand_downwards()
2468 unsigned long locked_vm = 0; in do_vmi_align_munmap() local
2516 locked_vm += vma_pages(next); in do_vmi_align_munmap()
2568 mm->locked_vm -= locked_vm; in do_vmi_align_munmap()
2881 mm->locked_vm += (len >> PAGE_SHIFT); in mmap_region()
3151 mm->locked_vm += (len >> PAGE_SHIFT); in do_brk_flags()
H A Dmlock.c458 mm->locked_vm += nr_pages; in mlock_fixup()
596 locked += current->mm->locked_vm; in do_mlock()
H A Dmremap.c733 mm->locked_vm += new_len >> PAGE_SHIFT; in move_vma()
1075 mm->locked_vm += pages; in SYSCALL_DEFINE5()
/openbmc/linux/net/xdp/
H A Dxdp_umem.c35 atomic_long_sub(umem->npgs, &umem->user->locked_vm); in xdp_umem_unaccount_pages()
139 old_npgs = atomic_long_read(&umem->user->locked_vm); in xdp_umem_account_pages()
146 } while (atomic_long_cmpxchg(&umem->user->locked_vm, old_npgs, in xdp_umem_account_pages()
/openbmc/linux/include/linux/sched/
H A Duser.h29 atomic_long_t locked_vm; member
/openbmc/linux/arch/s390/kvm/
H A Dpci.c199 atomic_long_sub(nr_pages, &user->locked_vm); in unaccount_mem()
212 cur_pages = atomic_long_read(&user->locked_vm); in account_mem()
216 } while (atomic_long_cmpxchg(&user->locked_vm, cur_pages, in account_mem()
/openbmc/linux/io_uring/
H A Drsrc.h142 atomic_long_sub(nr_pages, &user->locked_vm); in __io_unaccount_mem()
H A Drsrc.c51 cur_pages = atomic_long_read(&user->locked_vm); in __io_account_mem()
56 } while (!atomic_long_try_cmpxchg(&user->locked_vm, in __io_account_mem()
/openbmc/linux/drivers/iommu/iommufd/
H A Dpages.c813 cur_pages = atomic_long_read(&pages->source_user->locked_vm); in incr_user_locked_vm()
817 } while (atomic_long_cmpxchg(&pages->source_user->locked_vm, cur_pages, in incr_user_locked_vm()
824 if (WARN_ON(atomic_long_read(&pages->source_user->locked_vm) < npages)) in decr_user_locked_vm()
826 atomic_long_sub(npages, &pages->source_user->locked_vm); in decr_user_locked_vm()
/openbmc/linux/include/linux/
H A Dmm_types.h820 unsigned long locked_vm; /* Pages that have PG_mlocked set */ member
/openbmc/linux/drivers/vfio/
H A Dvfio_iommu_type1.c101 size_t locked_vm; member
439 dma->locked_vm += npage; in vfio_lock_acct()
675 mm->locked_vm + lock_acct + 1 > limit) { in vfio_pin_pages_remote()
1529 long npage = dma->locked_vm; in vfio_change_dma_owner()
/openbmc/linux/drivers/infiniband/sw/siw/
H A Dsiw_verbs.c1346 if (num_pages > mem_limit - current->mm->locked_vm) { in siw_reg_user_mr()
1349 current->mm->locked_vm); in siw_reg_user_mr()
/openbmc/linux/net/core/
H A Dskbuff.c1513 old_pg = atomic_long_read(&user->locked_vm); in mm_account_pinned_pages()
1518 } while (!atomic_long_try_cmpxchg(&user->locked_vm, &old_pg, new_pg)); in mm_account_pinned_pages()
1534 atomic_long_sub(mmp->num_pg, &mmp->user->locked_vm); in mm_unaccount_pinned_pages()
/openbmc/linux/fs/proc/
H A Dtask_mmu.c62 SEQ_PUT_DEC(" kB\nVmLck:\t", mm->locked_vm); in task_mem()
/openbmc/linux/Documentation/driver-api/
H A Dvfio.rst636 mm::locked_vm counter to make sure we do not exceed the rlimit.
/openbmc/linux/Documentation/mm/
H A Dunevictable-lru.rst363 VMAs against the task's "locked_vm".
/openbmc/linux/kernel/
H A Dfork.c1273 mm->locked_vm = 0; in mm_init()
/openbmc/linux/kernel/events/
H A Dcore.c6378 atomic_long_sub(rb->aux_nr_pages - rb->aux_mmap_locked, &mmap_user->locked_vm); in perf_mmap_close()
6453 &mmap_user->locked_vm); in perf_mmap_close()
6605 user_locked = atomic_long_read(&user->locked_vm); in perf_mmap()
6667 atomic_long_add(user_extra, &user->locked_vm); in perf_mmap()