/openbmc/linux/mm/ |
H A D | swap.c | 122 void __folio_put(struct folio *folio) in __folio_put() argument 141 struct folio *folio, *next; in put_pages_list() local 168 VM_BUG_ON_FOLIO(folio_test_lru(folio), folio); in lru_add_fn() 210 struct folio *folio = fbatch->folios[i]; in folio_batch_move_lru() local 229 struct folio *folio, move_fn_t move_fn) in folio_batch_add_and_move() argument 352 void folio_activate(struct folio *folio) in folio_activate() argument 371 void folio_activate(struct folio *folio) in folio_activate() argument 501 void folio_add_lru(struct folio *folio) in folio_add_lru() argument 506 folio_test_unevictable(folio), folio); in folio_add_lru() 713 void folio_deactivate(struct folio *folio) in folio_deactivate() argument [all …]
|
H A D | filemap.c | 151 struct folio *folio) in filemap_unaccount_folio() argument 285 struct folio *folio; in page_cache_delete_batch() local 475 struct folio *folio; in filemap_range_has_page() local 636 struct folio *folio; in filemap_range_has_writeback() local 964 struct folio *folio; in filemap_alloc_folio() local 1140 key.folio = folio; in folio_wake_bit() 1646 wait->folio = folio; in __folio_lock_async() 1819 struct folio *folio; in filemap_get_entry() local 1866 struct folio *folio; in __filemap_get_folio() local 1972 struct folio *folio; in find_get_entry() local [all …]
|
H A D | truncate.c | 83 struct folio *folio = fbatch->folios[i]; in truncate_folio_batch_exceptionals() local 178 folio_invalidate(folio, 0, folio_size(folio)); in truncate_cleanup_folio() 227 truncate_inode_folio(folio->mapping, folio); in truncate_inode_partial_folio() 246 truncate_inode_folio(folio->mapping, folio); in truncate_inode_partial_folio() 270 struct folio *folio) in mapping_evict_folio() argument 296 struct folio *folio = page_folio(page); in invalidate_inode_page() local 338 struct folio *folio; in truncate_inode_pages_range() local 414 struct folio *folio = fbatch.folios[i]; in truncate_inode_pages_range() local 510 struct folio *folio = fbatch.folios[i]; in mapping_try_invalidate() local 570 struct folio *folio) in invalidate_complete_folio2() argument [all …]
|
H A D | rmap.c | 747 struct folio *folio = page_folio(page); in page_address_in_vma() local 1016 int folio_mkclean(struct folio *folio) in folio_mkclean() argument 1073 int folio_total_mapcount(struct folio *folio) in folio_total_mapcount() argument 1109 struct folio *folio = page_folio(page); in page_move_anon_rmap() local 1206 struct folio *folio = page_folio(page); in page_add_anon_rmap() local 1369 struct folio *folio = page_folio(page); in page_add_file_rmap() local 1393 struct folio *folio = page_folio(page); in page_remove_rmap() local 1877 !folio_test_pmd_mappable(folio), folio); in try_to_migrate_one() 2342 struct folio *folio = page_folio(pages[i]); in make_device_exclusive_range() local 2412 static void rmap_walk_anon(struct folio *folio, in rmap_walk_anon() argument [all …]
|
H A D | page_idle.c | 37 struct folio *folio; in page_idle_get_folio() local 43 if (!folio_test_lru(folio) || !folio_try_get(folio)) in page_idle_get_folio() 47 folio = NULL; in page_idle_get_folio() 49 return folio; in page_idle_get_folio() 52 static bool page_idle_clear_pte_refs_one(struct folio *folio, in page_idle_clear_pte_refs_one() argument 89 static void page_idle_clear_pte_refs(struct folio *folio) in page_idle_clear_pte_refs() argument 101 if (!folio_mapped(folio) || !folio_raw_mapping(folio)) in page_idle_clear_pte_refs() 119 struct folio *folio; in page_idle_bitmap_read() local 139 if (folio) { in page_idle_bitmap_read() 164 struct folio *folio; in page_idle_bitmap_write() local [all …]
|
H A D | mlock.c | 95 folio->mlock_count = !!folio_test_mlocked(folio); in __mlock_folio() 105 VM_BUG_ON_FOLIO(folio_test_lru(folio), folio); in __mlock_new_folio() 169 static inline struct folio *mlock_lru(struct folio *folio) in mlock_lru() argument 174 static inline struct folio *mlock_new(struct folio *folio) in mlock_new() argument 190 struct folio *folio; in mlock_folio_batch() local 196 folio = (struct folio *)((unsigned long)folio - mlock); in mlock_folio_batch() 243 void mlock_folio(struct folio *folio) in mlock_folio() argument 268 void mlock_new_folio(struct folio *folio) in mlock_new_folio() argument 291 void munlock_folio(struct folio *folio) in munlock_folio() argument 316 struct folio *folio; in mlock_pte_range() local [all …]
|
H A D | swap_state.c | 97 VM_BUG_ON_FOLIO(!folio_test_locked(folio), folio); in add_to_swap_cache() 98 VM_BUG_ON_FOLIO(folio_test_swapcache(folio), folio); in add_to_swap_cache() 139 void __delete_from_swap_cache(struct folio *folio, in __delete_from_swap_cache() argument 150 VM_BUG_ON_FOLIO(!folio_test_locked(folio), folio); in __delete_from_swap_cache() 176 bool add_to_swap(struct folio *folio) in add_to_swap() argument 181 VM_BUG_ON_FOLIO(!folio_test_locked(folio), folio); in add_to_swap() 233 void delete_from_swap_cache(struct folio *folio) in delete_from_swap_cache() argument 243 folio_ref_sub(folio, folio_nr_pages(folio)); in delete_from_swap_cache() 286 struct folio *folio = page_folio(page); in free_swap_cache() local 334 struct folio *folio; in swap_cache_get_folio() local [all …]
|
H A D | migrate.c | 62 struct folio *folio = folio_get_nontail_page(page); in isolate_movable_page() local 149 struct folio *folio; in putback_movable_pages() local 205 !folio_test_pmd_mappable(folio), folio); in remove_migration_pte() 379 struct folio *folio) in folio_expected_refs() argument 401 struct folio *newfolio, struct folio *folio, int extra_count) in folio_migrate_mapping() argument 528 struct folio *dst, struct folio *src) in migrate_huge_page_move_mapping() 557 void folio_migrate_flags(struct folio *newfolio, struct folio *folio) in folio_migrate_flags() argument 648 void folio_migrate_copy(struct folio *newfolio, struct folio *folio) in folio_migrate_copy() argument 902 remove_migration_ptes(folio, folio, false); in writeout() 1528 struct folio *folio, *folio2; in migrate_hugetlbs() local [all …]
|
/openbmc/linux/include/linux/ |
H A D | hugetlb_cgroup.h | 70 VM_BUG_ON_FOLIO(!folio_test_hugetlb(folio), folio); in __hugetlb_cgroup_from_folio() 85 hugetlb_cgroup_from_folio_rsvd(struct folio *folio) in hugetlb_cgroup_from_folio_rsvd() argument 93 VM_BUG_ON_FOLIO(!folio_test_hugetlb(folio), folio); in __set_hugetlb_cgroup() 144 struct folio *folio); 147 struct folio *folio); 149 struct folio *folio); 151 struct folio *folio); 233 struct folio *folio) in hugetlb_cgroup_commit_charge() argument 240 struct folio *folio) in hugetlb_cgroup_commit_charge_rsvd() argument 245 struct folio *folio) in hugetlb_cgroup_uncharge_folio() argument [all …]
|
H A D | pagemap.h | 809 return folio->index + folio_nr_pages(folio); in folio_next_index() 993 struct folio *folio; member 999 struct folio *folio; member 1007 if (wait_page->folio != key->folio) in wake_page_match() 1017 void __folio_lock(struct folio *folio); 1021 void folio_unlock(struct folio *folio); 1090 struct folio *folio; in lock_page() local 1168 void folio_end_writeback(struct folio *folio); 1170 void folio_wait_stable(struct folio *folio); 1197 void folio_end_private_2(struct folio *folio); [all …]
|
H A D | mm_inline.h | 27 static inline int folio_is_file_lru(struct folio *folio) in folio_is_file_lru() argument 67 VM_BUG_ON_FOLIO(!folio_test_lru(folio), folio); in __folio_clear_lru_flags() 72 if (folio_test_active(folio) && folio_test_unevictable(folio)) in __folio_clear_lru_flags() 90 VM_BUG_ON_FOLIO(folio_test_active(folio) && folio_test_unevictable(folio), folio); in folio_lru_list() 143 static inline int folio_lru_refs(struct folio *folio) in folio_lru_refs() argument 157 static inline int folio_lru_gen(struct folio *folio) in folio_lru_gen() argument 249 (folio_test_dirty(folio) || folio_test_writeback(folio)))) in lru_gen_add_folio() 279 VM_WARN_ON_ONCE_FOLIO(folio_test_active(folio), folio); in lru_gen_del_folio() 280 VM_WARN_ON_ONCE_FOLIO(folio_test_unevictable(folio), folio); in lru_gen_del_folio() 318 void lruvec_add_folio(struct lruvec *lruvec, struct folio *folio) in lruvec_add_folio() argument [all …]
|
H A D | page_idle.h | 16 static inline bool folio_test_young(struct folio *folio) in folio_test_young() argument 30 static inline void folio_set_young(struct folio *folio) in folio_set_young() argument 41 static inline bool folio_test_clear_young(struct folio *folio) in folio_test_clear_young() argument 55 static inline bool folio_test_idle(struct folio *folio) in folio_test_idle() argument 69 static inline void folio_set_idle(struct folio *folio) in folio_set_idle() argument 80 static inline void folio_clear_idle(struct folio *folio) in folio_clear_idle() argument 94 static inline bool folio_test_young(struct folio *folio) in folio_test_young() argument 99 static inline void folio_set_young(struct folio *folio) in folio_set_young() argument 108 static inline bool folio_test_idle(struct folio *folio) in folio_test_idle() argument 113 static inline void folio_set_idle(struct folio *folio) in folio_set_idle() argument [all …]
|
H A D | page_ref.h | 87 static inline int folio_ref_count(const struct folio *folio) in folio_ref_count() argument 104 static inline void folio_set_count(struct folio *folio, int v) in folio_set_count() argument 125 static inline void folio_ref_add(struct folio *folio, int nr) in folio_ref_add() argument 137 static inline void folio_ref_sub(struct folio *folio, int nr) in folio_ref_sub() argument 151 static inline int folio_ref_sub_return(struct folio *folio, int nr) in folio_ref_sub_return() argument 163 static inline void folio_ref_inc(struct folio *folio) in folio_ref_inc() argument 175 static inline void folio_ref_dec(struct folio *folio) in folio_ref_dec() argument 203 static inline int folio_ref_inc_return(struct folio *folio) in folio_ref_inc_return() argument 217 static inline int folio_ref_dec_and_test(struct folio *folio) in folio_ref_dec_and_test() argument 231 static inline int folio_ref_dec_return(struct folio *folio) in folio_ref_dec_return() argument [all …]
|
H A D | migrate.h | 10 typedef struct folio *new_folio_t(struct folio *folio, unsigned long private); 11 typedef void free_folio_t(struct folio *folio, unsigned long private); 73 struct folio *alloc_migration_target(struct folio *src, unsigned long private); 77 struct folio *dst, struct folio *src); 80 void folio_migrate_flags(struct folio *newfolio, struct folio *folio); 81 void folio_migrate_copy(struct folio *newfolio, struct folio *folio); 83 struct folio *newfolio, struct folio *folio, int extra_count); 92 static inline struct folio *alloc_migration_target(struct folio *src, in alloc_migration_target() 99 struct folio *dst, struct folio *src) in migrate_huge_page_move_mapping() 121 static inline bool folio_test_movable(struct folio *folio) in folio_test_movable() argument [all …]
|
H A D | memcontrol.h | 359 static inline bool folio_memcg_kmem(struct folio *folio); 387 VM_BUG_ON_FOLIO(folio_test_slab(folio), folio); in __folio_memcg() 408 VM_BUG_ON_FOLIO(folio_test_slab(folio), folio); in __folio_objcg() 463 VM_BUG_ON_FOLIO(folio_test_slab(folio), folio); in folio_memcg_rcu() 549 static inline bool folio_memcg_kmem(struct folio *folio) in folio_memcg_kmem() argument 551 VM_BUG_ON_PGFLAGS(PageTail(&folio->page), &folio->page); in folio_memcg_kmem() 558 static inline bool folio_memcg_kmem(struct folio *folio) in folio_memcg_kmem() argument 693 void __mem_cgroup_uncharge(struct folio *folio); 772 struct lruvec *folio_lruvec_lock(struct folio *folio); 956 void folio_memcg_lock(struct folio *folio); [all …]
|
H A D | swap.h | 338 struct folio *folio = page_folio(page); in page_swap_entry() local 349 void workingset_refault(struct folio *folio, void *shadow); 350 void workingset_activation(struct folio *folio); 395 void folio_deactivate(struct folio *folio); 396 void folio_mark_lazyfree(struct folio *folio); 473 swp_entry_t folio_alloc_swap(struct folio *folio); 474 bool folio_free_swap(struct folio *folio); 475 void put_swap_folio(struct folio *folio, swp_entry_t entry); 590 static inline bool folio_free_swap(struct folio *folio) in folio_free_swap() argument 666 extern bool mem_cgroup_swap_full(struct folio *folio); [all …]
|
/openbmc/linux/fs/afs/ |
H A D | write.c | 51 struct folio *folio) in afs_flush_conflicting_write() argument 73 struct folio *folio; in afs_write_begin() local 157 struct folio *folio = page_folio(subpage); in afs_write_end() local 224 struct folio *folio; in afs_kill_pages() local 262 struct folio *folio; in afs_redirty_pages() local 293 struct folio *folio; in afs_pages_written_back() local 471 struct folio *folio; in afs_extend_writeback() local 583 struct folio *folio, in afs_write_back_from_locked_folio() argument 708 struct folio *folio; in afs_writepages_region() local 912 struct folio *folio = page_folio(vmf->page); in afs_page_mkwrite() local [all …]
|
/openbmc/linux/mm/damon/ |
H A D | paddr.c | 36 struct folio *folio = damon_get_folio(PHYS_PFN(paddr)); in damon_pa_mkold() local 43 if (!folio) in damon_pa_mkold() 46 if (!folio_mapped(folio) || !folio_raw_mapping(folio)) { in damon_pa_mkold() 51 need_lock = !folio_test_anon(folio) || folio_test_ksm(folio); in damon_pa_mkold() 116 struct folio *folio = damon_get_folio(PHYS_PFN(paddr)); in damon_pa_young() local 125 if (!folio) in damon_pa_young() 128 if (!folio_mapped(folio) || !folio_raw_mapping(folio)) { in damon_pa_young() 189 struct folio *folio) in __damos_pa_filter_out() argument 234 struct folio *folio = damon_get_folio(PHYS_PFN(addr)); in damon_pa_pageout() local 236 if (!folio) in damon_pa_pageout() [all …]
|
H A D | ops-common.c | 25 struct folio *folio; in damon_get_folio() local 31 if (!folio_test_lru(folio) || !folio_try_get(folio)) in damon_get_folio() 33 if (unlikely(page_folio(page) != folio || !folio_test_lru(folio))) { in damon_get_folio() 34 folio_put(folio); in damon_get_folio() 35 folio = NULL; in damon_get_folio() 37 return folio; in damon_get_folio() 42 struct folio *folio = damon_get_folio(pte_pfn(ptep_get(pte))); in damon_ptep_mkold() local 44 if (!folio) in damon_ptep_mkold() 51 folio_put(folio); in damon_ptep_mkold() 57 struct folio *folio = damon_get_folio(pmd_pfn(pmdp_get(pmd))); in damon_pmdp_mkold() local [all …]
|
/openbmc/linux/fs/iomap/ |
H A D | buffered-io.c | 295 struct folio *folio) in iomap_read_inline_data() argument 340 struct folio *folio = ctx->cur_folio; in iomap_readpage_iter() local 632 size_t len, struct folio *folio) in __iomap_write_begin() argument 706 struct folio *folio) in __iomap_put_folio() argument 719 struct folio *folio) in iomap_write_begin_inline() argument 732 struct folio *folio; in iomap_write_begin() local 880 struct folio *folio; in iomap_write_iter() local 1077 struct folio *folio; in iomap_write_delalloc_scan() local 1281 struct folio *folio; in iomap_unshare_iter() local 1342 struct folio *folio; in iomap_zero_iter() local [all …]
|
/openbmc/linux/include/trace/events/ |
H A D | pagemap.h | 30 TP_PROTO(struct folio *folio), 32 TP_ARGS(folio), 35 __field(struct folio *, folio ) 42 __entry->folio = folio; 43 __entry->pfn = folio_pfn(folio); 50 __entry->folio, 63 TP_PROTO(struct folio *folio), 65 TP_ARGS(folio), 68 __field(struct folio *, folio ) 73 __entry->folio = folio; [all …]
|
/openbmc/linux/fs/9p/ |
H A D | vfs_addr.c | 115 static bool v9fs_release_folio(struct folio *folio, gfp_t gfp) in v9fs_release_folio() argument 130 static void v9fs_invalidate_folio(struct folio *folio, size_t offset, in v9fs_invalidate_folio() argument 152 static int v9fs_vfs_write_folio_locked(struct folio *folio) in v9fs_vfs_write_folio_locked() argument 203 struct folio *folio = page_folio(page); in v9fs_vfs_writepage() local 219 folio_unlock(folio); in v9fs_vfs_writepage() 223 static int v9fs_launder_folio(struct folio *folio) in v9fs_launder_folio() argument 280 struct folio *folio; in v9fs_write_begin() local 302 struct folio *folio = page_folio(subpage); in v9fs_write_end() local 328 folio_mark_dirty(folio); in v9fs_write_end() 330 folio_unlock(folio); in v9fs_write_end() [all …]
|
/openbmc/linux/fs/gfs2/ |
H A D | aops.c | 40 void gfs2_trans_add_databufs(struct gfs2_inode *ip, struct folio *folio, in gfs2_trans_add_databufs() argument 92 static int gfs2_write_jdata_folio(struct folio *folio, in gfs2_write_jdata_folio() argument 106 i_size < folio_pos(folio) + folio_size(folio)) in gfs2_write_jdata_folio() 107 folio_zero_segment(folio, offset_in_folio(folio, i_size), in gfs2_write_jdata_folio() 124 static int __gfs2_jdata_write_folio(struct folio *folio, in __gfs2_jdata_write_folio() argument 137 gfs2_trans_add_databufs(ip, folio, 0, folio_size(folio)); in __gfs2_jdata_write_folio() 153 struct folio *folio = page_folio(page); in gfs2_jdata_writepage() local 229 struct folio *folio = fbatch->folios[i]; in gfs2_write_jdata_batch() local 491 struct folio *folio; in gfs2_internal_read() local 579 struct folio *folio) in jdata_dirty_folio() argument [all …]
|
/openbmc/linux/fs/netfs/ |
H A D | buffered_read.c | 19 struct folio *folio; in netfs_rreq_unlock_folios() local 55 pg_end = folio_pos(folio) + folio_size(folio) - 1; in netfs_rreq_unlock_folios() 224 int netfs_read_folio(struct file *file, struct folio *folio) in netfs_read_folio() argument 234 folio_file_pos(folio), folio_size(folio), in netfs_read_folio() 347 struct folio *folio; in netfs_write_begin() local 356 if (IS_ERR(folio)) in netfs_write_begin() 366 if (!folio) in netfs_write_begin() 384 folio_file_pos(folio), folio_size(folio), in netfs_write_begin() 409 folio_get(folio); in netfs_write_begin() 422 *_folio = folio; in netfs_write_begin() [all …]
|
/openbmc/linux/fs/ |
H A D | mpage.c | 56 folio_unlock(fi.folio); in mpage_read_end_io() 104 static void map_buffer_to_folio(struct folio *folio, struct buffer_head *bh, in map_buffer_to_folio() argument 141 struct folio *folio; member 161 struct folio *folio = args->folio; in do_mpage_readpage() local 182 VM_BUG_ON_FOLIO(folio_test_large(folio), folio); in do_mpage_readpage() 372 struct folio *folio; in mpage_readahead() local 380 args.folio = folio; in mpage_readahead() 392 int mpage_read_folio(struct folio *folio, get_block_t get_block) in mpage_read_folio() argument 395 .folio = folio, in mpage_read_folio() 599 folio_zero_segment(folio, length, folio_size(folio)); in __mpage_writepage() [all …]
|