Searched refs:NM_I (Results 1 – 9 of 9) sorted by relevance
132 return NM_I(sbi)->nat_cnt[DIRTY_NAT] >= NM_I(sbi)->max_nid * in excess_dirty_nats()133 NM_I(sbi)->dirty_nats_ratio / 100; in excess_dirty_nats()138 return NM_I(sbi)->nat_cnt[TOTAL_NAT] >= DEF_NAT_CACHE_THRESHOLD; in excess_cached_nats()168 struct f2fs_nm_info *nm_i = NM_I(sbi); in next_free_nid()186 struct f2fs_nm_info *nm_i = NM_I(sbi); in get_nat_bitmap()198 struct f2fs_nm_info *nm_i = NM_I(sbi); in current_nat_addr()222 struct f2fs_nm_info *nm_i = NM_I(sbi); in next_nat_addr()
35 if (unlikely(nid < F2FS_ROOT_INO(sbi) || nid >= NM_I(sbi)->max_nid)) { in f2fs_check_nid_range()47 struct f2fs_nm_info *nm_i = NM_I(sbi); in f2fs_available_free_memory()145 struct f2fs_nm_info *nm_i = NM_I(sbi); in get_next_nat_page()382 struct f2fs_nm_info *nm_i = NM_I(sbi); in f2fs_need_dentry_mark()399 struct f2fs_nm_info *nm_i = NM_I(sbi); in f2fs_is_checkpointed_node()413 struct f2fs_nm_info *nm_i = NM_I(sbi); in f2fs_need_inode_block_update()431 struct f2fs_nm_info *nm_i = NM_I(sbi); in cache_nat_entry()459 struct f2fs_nm_info *nm_i = NM_I(sbi); in set_node_addr()517 struct f2fs_nm_info *nm_i = NM_I(sbi); in f2fs_try_to_free_nats()549 struct f2fs_nm_info *nm_i = NM_I(sbi); in f2fs_get_node_info()[all …]
170 si->nats = NM_I(sbi)->nat_cnt[TOTAL_NAT]; in update_general_status()171 si->dirty_nats = NM_I(sbi)->nat_cnt[DIRTY_NAT]; in update_general_status()174 si->free_nids = NM_I(sbi)->nid_cnt[FREE_NID]; in update_general_status()175 si->avail_nids = NM_I(sbi)->available_nids; in update_general_status()176 si->alloc_nids = NM_I(sbi)->nid_cnt[PREALLOC_NID]; in update_general_status()279 si->base_mem += (NM_I(sbi)->nat_bits_blocks << F2FS_BLKSIZE_BITS); in update_mem_info()280 si->base_mem += NM_I(sbi)->nat_blocks * in update_mem_info()282 si->base_mem += NM_I(sbi)->nat_blocks / 8; in update_mem_info()283 si->base_mem += NM_I(sbi)->nat_blocks * sizeof(unsigned short); in update_mem_info()302 si->cache_mem += (NM_I(sbi)->nid_cnt[FREE_NID] + in update_mem_info()[all …]
21 return NM_I(sbi)->nat_cnt[RECLAIMABLE_NAT]; in __count_nat_entries()26 long count = NM_I(sbi)->nid_cnt[FREE_NID] - MAX_FREE_NIDS; in __count_free_nids()
265 NAT_BLOCK_OFFSET(NM_I(sbi)->max_nid))) in f2fs_ra_meta_pages()1165 struct f2fs_nm_info *nm_i = NM_I(sbi); in __prepare_cp_block()1332 NM_I(sbi)->nat_bits_blocks > BLKS_PER_SEG(sbi)) { in update_ckpt_flags()1458 struct f2fs_nm_info *nm_i = NM_I(sbi); in do_checkpoint()1664 if (NM_I(sbi)->nat_cnt[DIRTY_NAT] == 0 && in f2fs_write_checkpoint()
59 if (NM_I(sbi)->max_rf_node_blocks && in f2fs_space_for_roll_forward()61 NM_I(sbi)->max_rf_node_blocks) in f2fs_space_for_roll_forward()
76 return (unsigned char *)NM_I(sbi); in __struct_ptr()
2006 static inline struct f2fs_nm_info *NM_I(struct f2fs_sb_info *sbi) in NM_I() function 4028 else if (blkaddr < NM_I(sbi)->nat_blkaddr) \
38 #define F2FS_COMPRESS_INO(sbi) (NM_I(sbi)->max_nid)