1*39a53e0cSJaegeuk Kim /** 2*39a53e0cSJaegeuk Kim * fs/f2fs/f2fs.h 3*39a53e0cSJaegeuk Kim * 4*39a53e0cSJaegeuk Kim * Copyright (c) 2012 Samsung Electronics Co., Ltd. 5*39a53e0cSJaegeuk Kim * http://www.samsung.com/ 6*39a53e0cSJaegeuk Kim * 7*39a53e0cSJaegeuk Kim * This program is free software; you can redistribute it and/or modify 8*39a53e0cSJaegeuk Kim * it under the terms of the GNU General Public License version 2 as 9*39a53e0cSJaegeuk Kim * published by the Free Software Foundation. 10*39a53e0cSJaegeuk Kim */ 11*39a53e0cSJaegeuk Kim #ifndef _LINUX_F2FS_H 12*39a53e0cSJaegeuk Kim #define _LINUX_F2FS_H 13*39a53e0cSJaegeuk Kim 14*39a53e0cSJaegeuk Kim #include <linux/types.h> 15*39a53e0cSJaegeuk Kim #include <linux/page-flags.h> 16*39a53e0cSJaegeuk Kim #include <linux/buffer_head.h> 17*39a53e0cSJaegeuk Kim #include <linux/version.h> 18*39a53e0cSJaegeuk Kim #include <linux/slab.h> 19*39a53e0cSJaegeuk Kim #include <linux/crc32.h> 20*39a53e0cSJaegeuk Kim #include <linux/magic.h> 21*39a53e0cSJaegeuk Kim 22*39a53e0cSJaegeuk Kim /* 23*39a53e0cSJaegeuk Kim * For mount options 24*39a53e0cSJaegeuk Kim */ 25*39a53e0cSJaegeuk Kim #define F2FS_MOUNT_BG_GC 0x00000001 26*39a53e0cSJaegeuk Kim #define F2FS_MOUNT_DISABLE_ROLL_FORWARD 0x00000002 27*39a53e0cSJaegeuk Kim #define F2FS_MOUNT_DISCARD 0x00000004 28*39a53e0cSJaegeuk Kim #define F2FS_MOUNT_NOHEAP 0x00000008 29*39a53e0cSJaegeuk Kim #define F2FS_MOUNT_XATTR_USER 0x00000010 30*39a53e0cSJaegeuk Kim #define F2FS_MOUNT_POSIX_ACL 0x00000020 31*39a53e0cSJaegeuk Kim #define F2FS_MOUNT_DISABLE_EXT_IDENTIFY 0x00000040 32*39a53e0cSJaegeuk Kim 33*39a53e0cSJaegeuk Kim #define clear_opt(sbi, option) (sbi->mount_opt.opt &= ~F2FS_MOUNT_##option) 34*39a53e0cSJaegeuk Kim #define set_opt(sbi, option) (sbi->mount_opt.opt |= F2FS_MOUNT_##option) 35*39a53e0cSJaegeuk Kim #define test_opt(sbi, option) (sbi->mount_opt.opt & F2FS_MOUNT_##option) 36*39a53e0cSJaegeuk Kim 37*39a53e0cSJaegeuk Kim #define ver_after(a, b) (typecheck(unsigned long long, a) && \ 38*39a53e0cSJaegeuk Kim typecheck(unsigned long long, b) && \ 39*39a53e0cSJaegeuk Kim ((long long)((a) - (b)) > 0)) 40*39a53e0cSJaegeuk Kim 41*39a53e0cSJaegeuk Kim typedef u64 block_t; 42*39a53e0cSJaegeuk Kim typedef u32 nid_t; 43*39a53e0cSJaegeuk Kim 44*39a53e0cSJaegeuk Kim struct f2fs_mount_info { 45*39a53e0cSJaegeuk Kim unsigned int opt; 46*39a53e0cSJaegeuk Kim }; 47*39a53e0cSJaegeuk Kim 48*39a53e0cSJaegeuk Kim static inline __u32 f2fs_crc32(void *buff, size_t len) 49*39a53e0cSJaegeuk Kim { 50*39a53e0cSJaegeuk Kim return crc32_le(F2FS_SUPER_MAGIC, buff, len); 51*39a53e0cSJaegeuk Kim } 52*39a53e0cSJaegeuk Kim 53*39a53e0cSJaegeuk Kim static inline bool f2fs_crc_valid(__u32 blk_crc, void *buff, size_t buff_size) 54*39a53e0cSJaegeuk Kim { 55*39a53e0cSJaegeuk Kim return f2fs_crc32(buff, buff_size) == blk_crc; 56*39a53e0cSJaegeuk Kim } 57*39a53e0cSJaegeuk Kim 58*39a53e0cSJaegeuk Kim /* 59*39a53e0cSJaegeuk Kim * For checkpoint manager 60*39a53e0cSJaegeuk Kim */ 61*39a53e0cSJaegeuk Kim enum { 62*39a53e0cSJaegeuk Kim NAT_BITMAP, 63*39a53e0cSJaegeuk Kim SIT_BITMAP 64*39a53e0cSJaegeuk Kim }; 65*39a53e0cSJaegeuk Kim 66*39a53e0cSJaegeuk Kim /* for the list of orphan inodes */ 67*39a53e0cSJaegeuk Kim struct orphan_inode_entry { 68*39a53e0cSJaegeuk Kim struct list_head list; /* list head */ 69*39a53e0cSJaegeuk Kim nid_t ino; /* inode number */ 70*39a53e0cSJaegeuk Kim }; 71*39a53e0cSJaegeuk Kim 72*39a53e0cSJaegeuk Kim /* for the list of directory inodes */ 73*39a53e0cSJaegeuk Kim struct dir_inode_entry { 74*39a53e0cSJaegeuk Kim struct list_head list; /* list head */ 75*39a53e0cSJaegeuk Kim struct inode *inode; /* vfs inode pointer */ 76*39a53e0cSJaegeuk Kim }; 77*39a53e0cSJaegeuk Kim 78*39a53e0cSJaegeuk Kim /* for the list of fsync inodes, used only during recovery */ 79*39a53e0cSJaegeuk Kim struct fsync_inode_entry { 80*39a53e0cSJaegeuk Kim struct list_head list; /* list head */ 81*39a53e0cSJaegeuk Kim struct inode *inode; /* vfs inode pointer */ 82*39a53e0cSJaegeuk Kim block_t blkaddr; /* block address locating the last inode */ 83*39a53e0cSJaegeuk Kim }; 84*39a53e0cSJaegeuk Kim 85*39a53e0cSJaegeuk Kim #define nats_in_cursum(sum) (le16_to_cpu(sum->n_nats)) 86*39a53e0cSJaegeuk Kim #define sits_in_cursum(sum) (le16_to_cpu(sum->n_sits)) 87*39a53e0cSJaegeuk Kim 88*39a53e0cSJaegeuk Kim #define nat_in_journal(sum, i) (sum->nat_j.entries[i].ne) 89*39a53e0cSJaegeuk Kim #define nid_in_journal(sum, i) (sum->nat_j.entries[i].nid) 90*39a53e0cSJaegeuk Kim #define sit_in_journal(sum, i) (sum->sit_j.entries[i].se) 91*39a53e0cSJaegeuk Kim #define segno_in_journal(sum, i) (sum->sit_j.entries[i].segno) 92*39a53e0cSJaegeuk Kim 93*39a53e0cSJaegeuk Kim static inline int update_nats_in_cursum(struct f2fs_summary_block *rs, int i) 94*39a53e0cSJaegeuk Kim { 95*39a53e0cSJaegeuk Kim int before = nats_in_cursum(rs); 96*39a53e0cSJaegeuk Kim rs->n_nats = cpu_to_le16(before + i); 97*39a53e0cSJaegeuk Kim return before; 98*39a53e0cSJaegeuk Kim } 99*39a53e0cSJaegeuk Kim 100*39a53e0cSJaegeuk Kim static inline int update_sits_in_cursum(struct f2fs_summary_block *rs, int i) 101*39a53e0cSJaegeuk Kim { 102*39a53e0cSJaegeuk Kim int before = sits_in_cursum(rs); 103*39a53e0cSJaegeuk Kim rs->n_sits = cpu_to_le16(before + i); 104*39a53e0cSJaegeuk Kim return before; 105*39a53e0cSJaegeuk Kim } 106*39a53e0cSJaegeuk Kim 107*39a53e0cSJaegeuk Kim /* 108*39a53e0cSJaegeuk Kim * For INODE and NODE manager 109*39a53e0cSJaegeuk Kim */ 110*39a53e0cSJaegeuk Kim #define XATTR_NODE_OFFSET (-1) /* 111*39a53e0cSJaegeuk Kim * store xattrs to one node block per 112*39a53e0cSJaegeuk Kim * file keeping -1 as its node offset to 113*39a53e0cSJaegeuk Kim * distinguish from index node blocks. 114*39a53e0cSJaegeuk Kim */ 115*39a53e0cSJaegeuk Kim #define RDONLY_NODE 1 /* 116*39a53e0cSJaegeuk Kim * specify a read-only mode when getting 117*39a53e0cSJaegeuk Kim * a node block. 0 is read-write mode. 118*39a53e0cSJaegeuk Kim * used by get_dnode_of_data(). 119*39a53e0cSJaegeuk Kim */ 120*39a53e0cSJaegeuk Kim #define F2FS_LINK_MAX 32000 /* maximum link count per file */ 121*39a53e0cSJaegeuk Kim 122*39a53e0cSJaegeuk Kim /* for in-memory extent cache entry */ 123*39a53e0cSJaegeuk Kim struct extent_info { 124*39a53e0cSJaegeuk Kim rwlock_t ext_lock; /* rwlock for consistency */ 125*39a53e0cSJaegeuk Kim unsigned int fofs; /* start offset in a file */ 126*39a53e0cSJaegeuk Kim u32 blk_addr; /* start block address of the extent */ 127*39a53e0cSJaegeuk Kim unsigned int len; /* lenth of the extent */ 128*39a53e0cSJaegeuk Kim }; 129*39a53e0cSJaegeuk Kim 130*39a53e0cSJaegeuk Kim /* 131*39a53e0cSJaegeuk Kim * i_advise uses FADVISE_XXX_BIT. We can add additional hints later. 132*39a53e0cSJaegeuk Kim */ 133*39a53e0cSJaegeuk Kim #define FADVISE_COLD_BIT 0x01 134*39a53e0cSJaegeuk Kim 135*39a53e0cSJaegeuk Kim struct f2fs_inode_info { 136*39a53e0cSJaegeuk Kim struct inode vfs_inode; /* serve a vfs inode */ 137*39a53e0cSJaegeuk Kim unsigned long i_flags; /* keep an inode flags for ioctl */ 138*39a53e0cSJaegeuk Kim unsigned char i_advise; /* use to give file attribute hints */ 139*39a53e0cSJaegeuk Kim unsigned int i_current_depth; /* use only in directory structure */ 140*39a53e0cSJaegeuk Kim umode_t i_acl_mode; /* keep file acl mode temporarily */ 141*39a53e0cSJaegeuk Kim 142*39a53e0cSJaegeuk Kim /* Use below internally in f2fs*/ 143*39a53e0cSJaegeuk Kim unsigned long flags; /* use to pass per-file flags */ 144*39a53e0cSJaegeuk Kim unsigned long long data_version;/* lastes version of data for fsync */ 145*39a53e0cSJaegeuk Kim atomic_t dirty_dents; /* # of dirty dentry pages */ 146*39a53e0cSJaegeuk Kim f2fs_hash_t chash; /* hash value of given file name */ 147*39a53e0cSJaegeuk Kim unsigned int clevel; /* maximum level of given file name */ 148*39a53e0cSJaegeuk Kim nid_t i_xattr_nid; /* node id that contains xattrs */ 149*39a53e0cSJaegeuk Kim struct extent_info ext; /* in-memory extent cache entry */ 150*39a53e0cSJaegeuk Kim }; 151*39a53e0cSJaegeuk Kim 152*39a53e0cSJaegeuk Kim static inline void get_extent_info(struct extent_info *ext, 153*39a53e0cSJaegeuk Kim struct f2fs_extent i_ext) 154*39a53e0cSJaegeuk Kim { 155*39a53e0cSJaegeuk Kim write_lock(&ext->ext_lock); 156*39a53e0cSJaegeuk Kim ext->fofs = le32_to_cpu(i_ext.fofs); 157*39a53e0cSJaegeuk Kim ext->blk_addr = le32_to_cpu(i_ext.blk_addr); 158*39a53e0cSJaegeuk Kim ext->len = le32_to_cpu(i_ext.len); 159*39a53e0cSJaegeuk Kim write_unlock(&ext->ext_lock); 160*39a53e0cSJaegeuk Kim } 161*39a53e0cSJaegeuk Kim 162*39a53e0cSJaegeuk Kim static inline void set_raw_extent(struct extent_info *ext, 163*39a53e0cSJaegeuk Kim struct f2fs_extent *i_ext) 164*39a53e0cSJaegeuk Kim { 165*39a53e0cSJaegeuk Kim read_lock(&ext->ext_lock); 166*39a53e0cSJaegeuk Kim i_ext->fofs = cpu_to_le32(ext->fofs); 167*39a53e0cSJaegeuk Kim i_ext->blk_addr = cpu_to_le32(ext->blk_addr); 168*39a53e0cSJaegeuk Kim i_ext->len = cpu_to_le32(ext->len); 169*39a53e0cSJaegeuk Kim read_unlock(&ext->ext_lock); 170*39a53e0cSJaegeuk Kim } 171*39a53e0cSJaegeuk Kim 172*39a53e0cSJaegeuk Kim struct f2fs_nm_info { 173*39a53e0cSJaegeuk Kim block_t nat_blkaddr; /* base disk address of NAT */ 174*39a53e0cSJaegeuk Kim nid_t max_nid; /* maximum possible node ids */ 175*39a53e0cSJaegeuk Kim nid_t init_scan_nid; /* the first nid to be scanned */ 176*39a53e0cSJaegeuk Kim nid_t next_scan_nid; /* the next nid to be scanned */ 177*39a53e0cSJaegeuk Kim 178*39a53e0cSJaegeuk Kim /* NAT cache management */ 179*39a53e0cSJaegeuk Kim struct radix_tree_root nat_root;/* root of the nat entry cache */ 180*39a53e0cSJaegeuk Kim rwlock_t nat_tree_lock; /* protect nat_tree_lock */ 181*39a53e0cSJaegeuk Kim unsigned int nat_cnt; /* the # of cached nat entries */ 182*39a53e0cSJaegeuk Kim struct list_head nat_entries; /* cached nat entry list (clean) */ 183*39a53e0cSJaegeuk Kim struct list_head dirty_nat_entries; /* cached nat entry list (dirty) */ 184*39a53e0cSJaegeuk Kim 185*39a53e0cSJaegeuk Kim /* free node ids management */ 186*39a53e0cSJaegeuk Kim struct list_head free_nid_list; /* a list for free nids */ 187*39a53e0cSJaegeuk Kim spinlock_t free_nid_list_lock; /* protect free nid list */ 188*39a53e0cSJaegeuk Kim unsigned int fcnt; /* the number of free node id */ 189*39a53e0cSJaegeuk Kim struct mutex build_lock; /* lock for build free nids */ 190*39a53e0cSJaegeuk Kim 191*39a53e0cSJaegeuk Kim /* for checkpoint */ 192*39a53e0cSJaegeuk Kim char *nat_bitmap; /* NAT bitmap pointer */ 193*39a53e0cSJaegeuk Kim int bitmap_size; /* bitmap size */ 194*39a53e0cSJaegeuk Kim }; 195*39a53e0cSJaegeuk Kim 196*39a53e0cSJaegeuk Kim /* 197*39a53e0cSJaegeuk Kim * this structure is used as one of function parameters. 198*39a53e0cSJaegeuk Kim * all the information are dedicated to a given direct node block determined 199*39a53e0cSJaegeuk Kim * by the data offset in a file. 200*39a53e0cSJaegeuk Kim */ 201*39a53e0cSJaegeuk Kim struct dnode_of_data { 202*39a53e0cSJaegeuk Kim struct inode *inode; /* vfs inode pointer */ 203*39a53e0cSJaegeuk Kim struct page *inode_page; /* its inode page, NULL is possible */ 204*39a53e0cSJaegeuk Kim struct page *node_page; /* cached direct node page */ 205*39a53e0cSJaegeuk Kim nid_t nid; /* node id of the direct node block */ 206*39a53e0cSJaegeuk Kim unsigned int ofs_in_node; /* data offset in the node page */ 207*39a53e0cSJaegeuk Kim bool inode_page_locked; /* inode page is locked or not */ 208*39a53e0cSJaegeuk Kim block_t data_blkaddr; /* block address of the node block */ 209*39a53e0cSJaegeuk Kim }; 210*39a53e0cSJaegeuk Kim 211*39a53e0cSJaegeuk Kim static inline void set_new_dnode(struct dnode_of_data *dn, struct inode *inode, 212*39a53e0cSJaegeuk Kim struct page *ipage, struct page *npage, nid_t nid) 213*39a53e0cSJaegeuk Kim { 214*39a53e0cSJaegeuk Kim dn->inode = inode; 215*39a53e0cSJaegeuk Kim dn->inode_page = ipage; 216*39a53e0cSJaegeuk Kim dn->node_page = npage; 217*39a53e0cSJaegeuk Kim dn->nid = nid; 218*39a53e0cSJaegeuk Kim dn->inode_page_locked = 0; 219*39a53e0cSJaegeuk Kim } 220*39a53e0cSJaegeuk Kim 221*39a53e0cSJaegeuk Kim /* 222*39a53e0cSJaegeuk Kim * For SIT manager 223*39a53e0cSJaegeuk Kim * 224*39a53e0cSJaegeuk Kim * By default, there are 6 active log areas across the whole main area. 225*39a53e0cSJaegeuk Kim * When considering hot and cold data separation to reduce cleaning overhead, 226*39a53e0cSJaegeuk Kim * we split 3 for data logs and 3 for node logs as hot, warm, and cold types, 227*39a53e0cSJaegeuk Kim * respectively. 228*39a53e0cSJaegeuk Kim * In the current design, you should not change the numbers intentionally. 229*39a53e0cSJaegeuk Kim * Instead, as a mount option such as active_logs=x, you can use 2, 4, and 6 230*39a53e0cSJaegeuk Kim * logs individually according to the underlying devices. (default: 6) 231*39a53e0cSJaegeuk Kim * Just in case, on-disk layout covers maximum 16 logs that consist of 8 for 232*39a53e0cSJaegeuk Kim * data and 8 for node logs. 233*39a53e0cSJaegeuk Kim */ 234*39a53e0cSJaegeuk Kim #define NR_CURSEG_DATA_TYPE (3) 235*39a53e0cSJaegeuk Kim #define NR_CURSEG_NODE_TYPE (3) 236*39a53e0cSJaegeuk Kim #define NR_CURSEG_TYPE (NR_CURSEG_DATA_TYPE + NR_CURSEG_NODE_TYPE) 237*39a53e0cSJaegeuk Kim 238*39a53e0cSJaegeuk Kim enum { 239*39a53e0cSJaegeuk Kim CURSEG_HOT_DATA = 0, /* directory entry blocks */ 240*39a53e0cSJaegeuk Kim CURSEG_WARM_DATA, /* data blocks */ 241*39a53e0cSJaegeuk Kim CURSEG_COLD_DATA, /* multimedia or GCed data blocks */ 242*39a53e0cSJaegeuk Kim CURSEG_HOT_NODE, /* direct node blocks of directory files */ 243*39a53e0cSJaegeuk Kim CURSEG_WARM_NODE, /* direct node blocks of normal files */ 244*39a53e0cSJaegeuk Kim CURSEG_COLD_NODE, /* indirect node blocks */ 245*39a53e0cSJaegeuk Kim NO_CHECK_TYPE 246*39a53e0cSJaegeuk Kim }; 247*39a53e0cSJaegeuk Kim 248*39a53e0cSJaegeuk Kim struct f2fs_sm_info { 249*39a53e0cSJaegeuk Kim struct sit_info *sit_info; /* whole segment information */ 250*39a53e0cSJaegeuk Kim struct free_segmap_info *free_info; /* free segment information */ 251*39a53e0cSJaegeuk Kim struct dirty_seglist_info *dirty_info; /* dirty segment information */ 252*39a53e0cSJaegeuk Kim struct curseg_info *curseg_array; /* active segment information */ 253*39a53e0cSJaegeuk Kim 254*39a53e0cSJaegeuk Kim struct list_head wblist_head; /* list of under-writeback pages */ 255*39a53e0cSJaegeuk Kim spinlock_t wblist_lock; /* lock for checkpoint */ 256*39a53e0cSJaegeuk Kim 257*39a53e0cSJaegeuk Kim block_t seg0_blkaddr; /* block address of 0'th segment */ 258*39a53e0cSJaegeuk Kim block_t main_blkaddr; /* start block address of main area */ 259*39a53e0cSJaegeuk Kim block_t ssa_blkaddr; /* start block address of SSA area */ 260*39a53e0cSJaegeuk Kim 261*39a53e0cSJaegeuk Kim unsigned int segment_count; /* total # of segments */ 262*39a53e0cSJaegeuk Kim unsigned int main_segments; /* # of segments in main area */ 263*39a53e0cSJaegeuk Kim unsigned int reserved_segments; /* # of reserved segments */ 264*39a53e0cSJaegeuk Kim unsigned int ovp_segments; /* # of overprovision segments */ 265*39a53e0cSJaegeuk Kim }; 266*39a53e0cSJaegeuk Kim 267*39a53e0cSJaegeuk Kim /* 268*39a53e0cSJaegeuk Kim * For directory operation 269*39a53e0cSJaegeuk Kim */ 270*39a53e0cSJaegeuk Kim #define NODE_DIR1_BLOCK (ADDRS_PER_INODE + 1) 271*39a53e0cSJaegeuk Kim #define NODE_DIR2_BLOCK (ADDRS_PER_INODE + 2) 272*39a53e0cSJaegeuk Kim #define NODE_IND1_BLOCK (ADDRS_PER_INODE + 3) 273*39a53e0cSJaegeuk Kim #define NODE_IND2_BLOCK (ADDRS_PER_INODE + 4) 274*39a53e0cSJaegeuk Kim #define NODE_DIND_BLOCK (ADDRS_PER_INODE + 5) 275*39a53e0cSJaegeuk Kim 276*39a53e0cSJaegeuk Kim /* 277*39a53e0cSJaegeuk Kim * For superblock 278*39a53e0cSJaegeuk Kim */ 279*39a53e0cSJaegeuk Kim /* 280*39a53e0cSJaegeuk Kim * COUNT_TYPE for monitoring 281*39a53e0cSJaegeuk Kim * 282*39a53e0cSJaegeuk Kim * f2fs monitors the number of several block types such as on-writeback, 283*39a53e0cSJaegeuk Kim * dirty dentry blocks, dirty node blocks, and dirty meta blocks. 284*39a53e0cSJaegeuk Kim */ 285*39a53e0cSJaegeuk Kim enum count_type { 286*39a53e0cSJaegeuk Kim F2FS_WRITEBACK, 287*39a53e0cSJaegeuk Kim F2FS_DIRTY_DENTS, 288*39a53e0cSJaegeuk Kim F2FS_DIRTY_NODES, 289*39a53e0cSJaegeuk Kim F2FS_DIRTY_META, 290*39a53e0cSJaegeuk Kim NR_COUNT_TYPE, 291*39a53e0cSJaegeuk Kim }; 292*39a53e0cSJaegeuk Kim 293*39a53e0cSJaegeuk Kim /* 294*39a53e0cSJaegeuk Kim * FS_LOCK nesting subclasses for the lock validator: 295*39a53e0cSJaegeuk Kim * 296*39a53e0cSJaegeuk Kim * The locking order between these classes is 297*39a53e0cSJaegeuk Kim * RENAME -> DENTRY_OPS -> DATA_WRITE -> DATA_NEW 298*39a53e0cSJaegeuk Kim * -> DATA_TRUNC -> NODE_WRITE -> NODE_NEW -> NODE_TRUNC 299*39a53e0cSJaegeuk Kim */ 300*39a53e0cSJaegeuk Kim enum lock_type { 301*39a53e0cSJaegeuk Kim RENAME, /* for renaming operations */ 302*39a53e0cSJaegeuk Kim DENTRY_OPS, /* for directory operations */ 303*39a53e0cSJaegeuk Kim DATA_WRITE, /* for data write */ 304*39a53e0cSJaegeuk Kim DATA_NEW, /* for data allocation */ 305*39a53e0cSJaegeuk Kim DATA_TRUNC, /* for data truncate */ 306*39a53e0cSJaegeuk Kim NODE_NEW, /* for node allocation */ 307*39a53e0cSJaegeuk Kim NODE_TRUNC, /* for node truncate */ 308*39a53e0cSJaegeuk Kim NODE_WRITE, /* for node write */ 309*39a53e0cSJaegeuk Kim NR_LOCK_TYPE, 310*39a53e0cSJaegeuk Kim }; 311*39a53e0cSJaegeuk Kim 312*39a53e0cSJaegeuk Kim /* 313*39a53e0cSJaegeuk Kim * The below are the page types of bios used in submti_bio(). 314*39a53e0cSJaegeuk Kim * The available types are: 315*39a53e0cSJaegeuk Kim * DATA User data pages. It operates as async mode. 316*39a53e0cSJaegeuk Kim * NODE Node pages. It operates as async mode. 317*39a53e0cSJaegeuk Kim * META FS metadata pages such as SIT, NAT, CP. 318*39a53e0cSJaegeuk Kim * NR_PAGE_TYPE The number of page types. 319*39a53e0cSJaegeuk Kim * META_FLUSH Make sure the previous pages are written 320*39a53e0cSJaegeuk Kim * with waiting the bio's completion 321*39a53e0cSJaegeuk Kim * ... Only can be used with META. 322*39a53e0cSJaegeuk Kim */ 323*39a53e0cSJaegeuk Kim enum page_type { 324*39a53e0cSJaegeuk Kim DATA, 325*39a53e0cSJaegeuk Kim NODE, 326*39a53e0cSJaegeuk Kim META, 327*39a53e0cSJaegeuk Kim NR_PAGE_TYPE, 328*39a53e0cSJaegeuk Kim META_FLUSH, 329*39a53e0cSJaegeuk Kim }; 330*39a53e0cSJaegeuk Kim 331*39a53e0cSJaegeuk Kim struct f2fs_sb_info { 332*39a53e0cSJaegeuk Kim struct super_block *sb; /* pointer to VFS super block */ 333*39a53e0cSJaegeuk Kim struct buffer_head *raw_super_buf; /* buffer head of raw sb */ 334*39a53e0cSJaegeuk Kim struct f2fs_super_block *raw_super; /* raw super block pointer */ 335*39a53e0cSJaegeuk Kim int s_dirty; /* dirty flag for checkpoint */ 336*39a53e0cSJaegeuk Kim 337*39a53e0cSJaegeuk Kim /* for node-related operations */ 338*39a53e0cSJaegeuk Kim struct f2fs_nm_info *nm_info; /* node manager */ 339*39a53e0cSJaegeuk Kim struct inode *node_inode; /* cache node blocks */ 340*39a53e0cSJaegeuk Kim 341*39a53e0cSJaegeuk Kim /* for segment-related operations */ 342*39a53e0cSJaegeuk Kim struct f2fs_sm_info *sm_info; /* segment manager */ 343*39a53e0cSJaegeuk Kim struct bio *bio[NR_PAGE_TYPE]; /* bios to merge */ 344*39a53e0cSJaegeuk Kim sector_t last_block_in_bio[NR_PAGE_TYPE]; /* last block number */ 345*39a53e0cSJaegeuk Kim struct rw_semaphore bio_sem; /* IO semaphore */ 346*39a53e0cSJaegeuk Kim 347*39a53e0cSJaegeuk Kim /* for checkpoint */ 348*39a53e0cSJaegeuk Kim struct f2fs_checkpoint *ckpt; /* raw checkpoint pointer */ 349*39a53e0cSJaegeuk Kim struct inode *meta_inode; /* cache meta blocks */ 350*39a53e0cSJaegeuk Kim struct mutex cp_mutex; /* for checkpoint procedure */ 351*39a53e0cSJaegeuk Kim struct mutex fs_lock[NR_LOCK_TYPE]; /* for blocking FS operations */ 352*39a53e0cSJaegeuk Kim struct mutex write_inode; /* mutex for write inode */ 353*39a53e0cSJaegeuk Kim struct mutex writepages; /* mutex for writepages() */ 354*39a53e0cSJaegeuk Kim int por_doing; /* recovery is doing or not */ 355*39a53e0cSJaegeuk Kim 356*39a53e0cSJaegeuk Kim /* for orphan inode management */ 357*39a53e0cSJaegeuk Kim struct list_head orphan_inode_list; /* orphan inode list */ 358*39a53e0cSJaegeuk Kim struct mutex orphan_inode_mutex; /* for orphan inode list */ 359*39a53e0cSJaegeuk Kim unsigned int n_orphans; /* # of orphan inodes */ 360*39a53e0cSJaegeuk Kim 361*39a53e0cSJaegeuk Kim /* for directory inode management */ 362*39a53e0cSJaegeuk Kim struct list_head dir_inode_list; /* dir inode list */ 363*39a53e0cSJaegeuk Kim spinlock_t dir_inode_lock; /* for dir inode list lock */ 364*39a53e0cSJaegeuk Kim unsigned int n_dirty_dirs; /* # of dir inodes */ 365*39a53e0cSJaegeuk Kim 366*39a53e0cSJaegeuk Kim /* basic file system units */ 367*39a53e0cSJaegeuk Kim unsigned int log_sectors_per_block; /* log2 sectors per block */ 368*39a53e0cSJaegeuk Kim unsigned int log_blocksize; /* log2 block size */ 369*39a53e0cSJaegeuk Kim unsigned int blocksize; /* block size */ 370*39a53e0cSJaegeuk Kim unsigned int root_ino_num; /* root inode number*/ 371*39a53e0cSJaegeuk Kim unsigned int node_ino_num; /* node inode number*/ 372*39a53e0cSJaegeuk Kim unsigned int meta_ino_num; /* meta inode number*/ 373*39a53e0cSJaegeuk Kim unsigned int log_blocks_per_seg; /* log2 blocks per segment */ 374*39a53e0cSJaegeuk Kim unsigned int blocks_per_seg; /* blocks per segment */ 375*39a53e0cSJaegeuk Kim unsigned int segs_per_sec; /* segments per section */ 376*39a53e0cSJaegeuk Kim unsigned int secs_per_zone; /* sections per zone */ 377*39a53e0cSJaegeuk Kim unsigned int total_sections; /* total section count */ 378*39a53e0cSJaegeuk Kim unsigned int total_node_count; /* total node block count */ 379*39a53e0cSJaegeuk Kim unsigned int total_valid_node_count; /* valid node block count */ 380*39a53e0cSJaegeuk Kim unsigned int total_valid_inode_count; /* valid inode count */ 381*39a53e0cSJaegeuk Kim int active_logs; /* # of active logs */ 382*39a53e0cSJaegeuk Kim 383*39a53e0cSJaegeuk Kim block_t user_block_count; /* # of user blocks */ 384*39a53e0cSJaegeuk Kim block_t total_valid_block_count; /* # of valid blocks */ 385*39a53e0cSJaegeuk Kim block_t alloc_valid_block_count; /* # of allocated blocks */ 386*39a53e0cSJaegeuk Kim block_t last_valid_block_count; /* for recovery */ 387*39a53e0cSJaegeuk Kim u32 s_next_generation; /* for NFS support */ 388*39a53e0cSJaegeuk Kim atomic_t nr_pages[NR_COUNT_TYPE]; /* # of pages, see count_type */ 389*39a53e0cSJaegeuk Kim 390*39a53e0cSJaegeuk Kim struct f2fs_mount_info mount_opt; /* mount options */ 391*39a53e0cSJaegeuk Kim 392*39a53e0cSJaegeuk Kim /* for cleaning operations */ 393*39a53e0cSJaegeuk Kim struct mutex gc_mutex; /* mutex for GC */ 394*39a53e0cSJaegeuk Kim struct f2fs_gc_kthread *gc_thread; /* GC thread */ 395*39a53e0cSJaegeuk Kim 396*39a53e0cSJaegeuk Kim /* 397*39a53e0cSJaegeuk Kim * for stat information. 398*39a53e0cSJaegeuk Kim * one is for the LFS mode, and the other is for the SSR mode. 399*39a53e0cSJaegeuk Kim */ 400*39a53e0cSJaegeuk Kim struct f2fs_stat_info *stat_info; /* FS status information */ 401*39a53e0cSJaegeuk Kim unsigned int segment_count[2]; /* # of allocated segments */ 402*39a53e0cSJaegeuk Kim unsigned int block_count[2]; /* # of allocated blocks */ 403*39a53e0cSJaegeuk Kim unsigned int last_victim[2]; /* last victim segment # */ 404*39a53e0cSJaegeuk Kim int total_hit_ext, read_hit_ext; /* extent cache hit ratio */ 405*39a53e0cSJaegeuk Kim int bg_gc; /* background gc calls */ 406*39a53e0cSJaegeuk Kim spinlock_t stat_lock; /* lock for stat operations */ 407*39a53e0cSJaegeuk Kim }; 408*39a53e0cSJaegeuk Kim 409*39a53e0cSJaegeuk Kim /* 410*39a53e0cSJaegeuk Kim * Inline functions 411*39a53e0cSJaegeuk Kim */ 412*39a53e0cSJaegeuk Kim static inline struct f2fs_inode_info *F2FS_I(struct inode *inode) 413*39a53e0cSJaegeuk Kim { 414*39a53e0cSJaegeuk Kim return container_of(inode, struct f2fs_inode_info, vfs_inode); 415*39a53e0cSJaegeuk Kim } 416*39a53e0cSJaegeuk Kim 417*39a53e0cSJaegeuk Kim static inline struct f2fs_sb_info *F2FS_SB(struct super_block *sb) 418*39a53e0cSJaegeuk Kim { 419*39a53e0cSJaegeuk Kim return sb->s_fs_info; 420*39a53e0cSJaegeuk Kim } 421*39a53e0cSJaegeuk Kim 422*39a53e0cSJaegeuk Kim static inline struct f2fs_super_block *F2FS_RAW_SUPER(struct f2fs_sb_info *sbi) 423*39a53e0cSJaegeuk Kim { 424*39a53e0cSJaegeuk Kim return (struct f2fs_super_block *)(sbi->raw_super); 425*39a53e0cSJaegeuk Kim } 426*39a53e0cSJaegeuk Kim 427*39a53e0cSJaegeuk Kim static inline struct f2fs_checkpoint *F2FS_CKPT(struct f2fs_sb_info *sbi) 428*39a53e0cSJaegeuk Kim { 429*39a53e0cSJaegeuk Kim return (struct f2fs_checkpoint *)(sbi->ckpt); 430*39a53e0cSJaegeuk Kim } 431*39a53e0cSJaegeuk Kim 432*39a53e0cSJaegeuk Kim static inline struct f2fs_nm_info *NM_I(struct f2fs_sb_info *sbi) 433*39a53e0cSJaegeuk Kim { 434*39a53e0cSJaegeuk Kim return (struct f2fs_nm_info *)(sbi->nm_info); 435*39a53e0cSJaegeuk Kim } 436*39a53e0cSJaegeuk Kim 437*39a53e0cSJaegeuk Kim static inline struct f2fs_sm_info *SM_I(struct f2fs_sb_info *sbi) 438*39a53e0cSJaegeuk Kim { 439*39a53e0cSJaegeuk Kim return (struct f2fs_sm_info *)(sbi->sm_info); 440*39a53e0cSJaegeuk Kim } 441*39a53e0cSJaegeuk Kim 442*39a53e0cSJaegeuk Kim static inline struct sit_info *SIT_I(struct f2fs_sb_info *sbi) 443*39a53e0cSJaegeuk Kim { 444*39a53e0cSJaegeuk Kim return (struct sit_info *)(SM_I(sbi)->sit_info); 445*39a53e0cSJaegeuk Kim } 446*39a53e0cSJaegeuk Kim 447*39a53e0cSJaegeuk Kim static inline struct free_segmap_info *FREE_I(struct f2fs_sb_info *sbi) 448*39a53e0cSJaegeuk Kim { 449*39a53e0cSJaegeuk Kim return (struct free_segmap_info *)(SM_I(sbi)->free_info); 450*39a53e0cSJaegeuk Kim } 451*39a53e0cSJaegeuk Kim 452*39a53e0cSJaegeuk Kim static inline struct dirty_seglist_info *DIRTY_I(struct f2fs_sb_info *sbi) 453*39a53e0cSJaegeuk Kim { 454*39a53e0cSJaegeuk Kim return (struct dirty_seglist_info *)(SM_I(sbi)->dirty_info); 455*39a53e0cSJaegeuk Kim } 456*39a53e0cSJaegeuk Kim 457*39a53e0cSJaegeuk Kim static inline void F2FS_SET_SB_DIRT(struct f2fs_sb_info *sbi) 458*39a53e0cSJaegeuk Kim { 459*39a53e0cSJaegeuk Kim sbi->s_dirty = 1; 460*39a53e0cSJaegeuk Kim } 461*39a53e0cSJaegeuk Kim 462*39a53e0cSJaegeuk Kim static inline void F2FS_RESET_SB_DIRT(struct f2fs_sb_info *sbi) 463*39a53e0cSJaegeuk Kim { 464*39a53e0cSJaegeuk Kim sbi->s_dirty = 0; 465*39a53e0cSJaegeuk Kim } 466*39a53e0cSJaegeuk Kim 467*39a53e0cSJaegeuk Kim static inline void mutex_lock_op(struct f2fs_sb_info *sbi, enum lock_type t) 468*39a53e0cSJaegeuk Kim { 469*39a53e0cSJaegeuk Kim mutex_lock_nested(&sbi->fs_lock[t], t); 470*39a53e0cSJaegeuk Kim } 471*39a53e0cSJaegeuk Kim 472*39a53e0cSJaegeuk Kim static inline void mutex_unlock_op(struct f2fs_sb_info *sbi, enum lock_type t) 473*39a53e0cSJaegeuk Kim { 474*39a53e0cSJaegeuk Kim mutex_unlock(&sbi->fs_lock[t]); 475*39a53e0cSJaegeuk Kim } 476*39a53e0cSJaegeuk Kim 477*39a53e0cSJaegeuk Kim /* 478*39a53e0cSJaegeuk Kim * Check whether the given nid is within node id range. 479*39a53e0cSJaegeuk Kim */ 480*39a53e0cSJaegeuk Kim static inline void check_nid_range(struct f2fs_sb_info *sbi, nid_t nid) 481*39a53e0cSJaegeuk Kim { 482*39a53e0cSJaegeuk Kim BUG_ON((nid >= NM_I(sbi)->max_nid)); 483*39a53e0cSJaegeuk Kim } 484*39a53e0cSJaegeuk Kim 485*39a53e0cSJaegeuk Kim #define F2FS_DEFAULT_ALLOCATED_BLOCKS 1 486*39a53e0cSJaegeuk Kim 487*39a53e0cSJaegeuk Kim /* 488*39a53e0cSJaegeuk Kim * Check whether the inode has blocks or not 489*39a53e0cSJaegeuk Kim */ 490*39a53e0cSJaegeuk Kim static inline int F2FS_HAS_BLOCKS(struct inode *inode) 491*39a53e0cSJaegeuk Kim { 492*39a53e0cSJaegeuk Kim if (F2FS_I(inode)->i_xattr_nid) 493*39a53e0cSJaegeuk Kim return (inode->i_blocks > F2FS_DEFAULT_ALLOCATED_BLOCKS + 1); 494*39a53e0cSJaegeuk Kim else 495*39a53e0cSJaegeuk Kim return (inode->i_blocks > F2FS_DEFAULT_ALLOCATED_BLOCKS); 496*39a53e0cSJaegeuk Kim } 497*39a53e0cSJaegeuk Kim 498*39a53e0cSJaegeuk Kim static inline bool inc_valid_block_count(struct f2fs_sb_info *sbi, 499*39a53e0cSJaegeuk Kim struct inode *inode, blkcnt_t count) 500*39a53e0cSJaegeuk Kim { 501*39a53e0cSJaegeuk Kim block_t valid_block_count; 502*39a53e0cSJaegeuk Kim 503*39a53e0cSJaegeuk Kim spin_lock(&sbi->stat_lock); 504*39a53e0cSJaegeuk Kim valid_block_count = 505*39a53e0cSJaegeuk Kim sbi->total_valid_block_count + (block_t)count; 506*39a53e0cSJaegeuk Kim if (valid_block_count > sbi->user_block_count) { 507*39a53e0cSJaegeuk Kim spin_unlock(&sbi->stat_lock); 508*39a53e0cSJaegeuk Kim return false; 509*39a53e0cSJaegeuk Kim } 510*39a53e0cSJaegeuk Kim inode->i_blocks += count; 511*39a53e0cSJaegeuk Kim sbi->total_valid_block_count = valid_block_count; 512*39a53e0cSJaegeuk Kim sbi->alloc_valid_block_count += (block_t)count; 513*39a53e0cSJaegeuk Kim spin_unlock(&sbi->stat_lock); 514*39a53e0cSJaegeuk Kim return true; 515*39a53e0cSJaegeuk Kim } 516*39a53e0cSJaegeuk Kim 517*39a53e0cSJaegeuk Kim static inline int dec_valid_block_count(struct f2fs_sb_info *sbi, 518*39a53e0cSJaegeuk Kim struct inode *inode, 519*39a53e0cSJaegeuk Kim blkcnt_t count) 520*39a53e0cSJaegeuk Kim { 521*39a53e0cSJaegeuk Kim spin_lock(&sbi->stat_lock); 522*39a53e0cSJaegeuk Kim BUG_ON(sbi->total_valid_block_count < (block_t) count); 523*39a53e0cSJaegeuk Kim BUG_ON(inode->i_blocks < count); 524*39a53e0cSJaegeuk Kim inode->i_blocks -= count; 525*39a53e0cSJaegeuk Kim sbi->total_valid_block_count -= (block_t)count; 526*39a53e0cSJaegeuk Kim spin_unlock(&sbi->stat_lock); 527*39a53e0cSJaegeuk Kim return 0; 528*39a53e0cSJaegeuk Kim } 529*39a53e0cSJaegeuk Kim 530*39a53e0cSJaegeuk Kim static inline void inc_page_count(struct f2fs_sb_info *sbi, int count_type) 531*39a53e0cSJaegeuk Kim { 532*39a53e0cSJaegeuk Kim atomic_inc(&sbi->nr_pages[count_type]); 533*39a53e0cSJaegeuk Kim F2FS_SET_SB_DIRT(sbi); 534*39a53e0cSJaegeuk Kim } 535*39a53e0cSJaegeuk Kim 536*39a53e0cSJaegeuk Kim static inline void inode_inc_dirty_dents(struct inode *inode) 537*39a53e0cSJaegeuk Kim { 538*39a53e0cSJaegeuk Kim atomic_inc(&F2FS_I(inode)->dirty_dents); 539*39a53e0cSJaegeuk Kim } 540*39a53e0cSJaegeuk Kim 541*39a53e0cSJaegeuk Kim static inline void dec_page_count(struct f2fs_sb_info *sbi, int count_type) 542*39a53e0cSJaegeuk Kim { 543*39a53e0cSJaegeuk Kim atomic_dec(&sbi->nr_pages[count_type]); 544*39a53e0cSJaegeuk Kim } 545*39a53e0cSJaegeuk Kim 546*39a53e0cSJaegeuk Kim static inline void inode_dec_dirty_dents(struct inode *inode) 547*39a53e0cSJaegeuk Kim { 548*39a53e0cSJaegeuk Kim atomic_dec(&F2FS_I(inode)->dirty_dents); 549*39a53e0cSJaegeuk Kim } 550*39a53e0cSJaegeuk Kim 551*39a53e0cSJaegeuk Kim static inline int get_pages(struct f2fs_sb_info *sbi, int count_type) 552*39a53e0cSJaegeuk Kim { 553*39a53e0cSJaegeuk Kim return atomic_read(&sbi->nr_pages[count_type]); 554*39a53e0cSJaegeuk Kim } 555*39a53e0cSJaegeuk Kim 556*39a53e0cSJaegeuk Kim static inline block_t valid_user_blocks(struct f2fs_sb_info *sbi) 557*39a53e0cSJaegeuk Kim { 558*39a53e0cSJaegeuk Kim block_t ret; 559*39a53e0cSJaegeuk Kim spin_lock(&sbi->stat_lock); 560*39a53e0cSJaegeuk Kim ret = sbi->total_valid_block_count; 561*39a53e0cSJaegeuk Kim spin_unlock(&sbi->stat_lock); 562*39a53e0cSJaegeuk Kim return ret; 563*39a53e0cSJaegeuk Kim } 564*39a53e0cSJaegeuk Kim 565*39a53e0cSJaegeuk Kim static inline unsigned long __bitmap_size(struct f2fs_sb_info *sbi, int flag) 566*39a53e0cSJaegeuk Kim { 567*39a53e0cSJaegeuk Kim struct f2fs_checkpoint *ckpt = F2FS_CKPT(sbi); 568*39a53e0cSJaegeuk Kim 569*39a53e0cSJaegeuk Kim /* return NAT or SIT bitmap */ 570*39a53e0cSJaegeuk Kim if (flag == NAT_BITMAP) 571*39a53e0cSJaegeuk Kim return le32_to_cpu(ckpt->nat_ver_bitmap_bytesize); 572*39a53e0cSJaegeuk Kim else if (flag == SIT_BITMAP) 573*39a53e0cSJaegeuk Kim return le32_to_cpu(ckpt->sit_ver_bitmap_bytesize); 574*39a53e0cSJaegeuk Kim 575*39a53e0cSJaegeuk Kim return 0; 576*39a53e0cSJaegeuk Kim } 577*39a53e0cSJaegeuk Kim 578*39a53e0cSJaegeuk Kim static inline void *__bitmap_ptr(struct f2fs_sb_info *sbi, int flag) 579*39a53e0cSJaegeuk Kim { 580*39a53e0cSJaegeuk Kim struct f2fs_checkpoint *ckpt = F2FS_CKPT(sbi); 581*39a53e0cSJaegeuk Kim int offset = (flag == NAT_BITMAP) ? ckpt->sit_ver_bitmap_bytesize : 0; 582*39a53e0cSJaegeuk Kim return &ckpt->sit_nat_version_bitmap + offset; 583*39a53e0cSJaegeuk Kim } 584*39a53e0cSJaegeuk Kim 585*39a53e0cSJaegeuk Kim static inline block_t __start_cp_addr(struct f2fs_sb_info *sbi) 586*39a53e0cSJaegeuk Kim { 587*39a53e0cSJaegeuk Kim block_t start_addr; 588*39a53e0cSJaegeuk Kim struct f2fs_checkpoint *ckpt = F2FS_CKPT(sbi); 589*39a53e0cSJaegeuk Kim unsigned long long ckpt_version = le64_to_cpu(ckpt->checkpoint_ver); 590*39a53e0cSJaegeuk Kim 591*39a53e0cSJaegeuk Kim start_addr = le64_to_cpu(F2FS_RAW_SUPER(sbi)->cp_blkaddr); 592*39a53e0cSJaegeuk Kim 593*39a53e0cSJaegeuk Kim /* 594*39a53e0cSJaegeuk Kim * odd numbered checkpoint should at cp segment 0 595*39a53e0cSJaegeuk Kim * and even segent must be at cp segment 1 596*39a53e0cSJaegeuk Kim */ 597*39a53e0cSJaegeuk Kim if (!(ckpt_version & 1)) 598*39a53e0cSJaegeuk Kim start_addr += sbi->blocks_per_seg; 599*39a53e0cSJaegeuk Kim 600*39a53e0cSJaegeuk Kim return start_addr; 601*39a53e0cSJaegeuk Kim } 602*39a53e0cSJaegeuk Kim 603*39a53e0cSJaegeuk Kim static inline block_t __start_sum_addr(struct f2fs_sb_info *sbi) 604*39a53e0cSJaegeuk Kim { 605*39a53e0cSJaegeuk Kim return le32_to_cpu(F2FS_CKPT(sbi)->cp_pack_start_sum); 606*39a53e0cSJaegeuk Kim } 607*39a53e0cSJaegeuk Kim 608*39a53e0cSJaegeuk Kim static inline bool inc_valid_node_count(struct f2fs_sb_info *sbi, 609*39a53e0cSJaegeuk Kim struct inode *inode, 610*39a53e0cSJaegeuk Kim unsigned int count) 611*39a53e0cSJaegeuk Kim { 612*39a53e0cSJaegeuk Kim block_t valid_block_count; 613*39a53e0cSJaegeuk Kim unsigned int valid_node_count; 614*39a53e0cSJaegeuk Kim 615*39a53e0cSJaegeuk Kim spin_lock(&sbi->stat_lock); 616*39a53e0cSJaegeuk Kim 617*39a53e0cSJaegeuk Kim valid_block_count = sbi->total_valid_block_count + (block_t)count; 618*39a53e0cSJaegeuk Kim sbi->alloc_valid_block_count += (block_t)count; 619*39a53e0cSJaegeuk Kim valid_node_count = sbi->total_valid_node_count + count; 620*39a53e0cSJaegeuk Kim 621*39a53e0cSJaegeuk Kim if (valid_block_count > sbi->user_block_count) { 622*39a53e0cSJaegeuk Kim spin_unlock(&sbi->stat_lock); 623*39a53e0cSJaegeuk Kim return false; 624*39a53e0cSJaegeuk Kim } 625*39a53e0cSJaegeuk Kim 626*39a53e0cSJaegeuk Kim if (valid_node_count > sbi->total_node_count) { 627*39a53e0cSJaegeuk Kim spin_unlock(&sbi->stat_lock); 628*39a53e0cSJaegeuk Kim return false; 629*39a53e0cSJaegeuk Kim } 630*39a53e0cSJaegeuk Kim 631*39a53e0cSJaegeuk Kim if (inode) 632*39a53e0cSJaegeuk Kim inode->i_blocks += count; 633*39a53e0cSJaegeuk Kim sbi->total_valid_node_count = valid_node_count; 634*39a53e0cSJaegeuk Kim sbi->total_valid_block_count = valid_block_count; 635*39a53e0cSJaegeuk Kim spin_unlock(&sbi->stat_lock); 636*39a53e0cSJaegeuk Kim 637*39a53e0cSJaegeuk Kim return true; 638*39a53e0cSJaegeuk Kim } 639*39a53e0cSJaegeuk Kim 640*39a53e0cSJaegeuk Kim static inline void dec_valid_node_count(struct f2fs_sb_info *sbi, 641*39a53e0cSJaegeuk Kim struct inode *inode, 642*39a53e0cSJaegeuk Kim unsigned int count) 643*39a53e0cSJaegeuk Kim { 644*39a53e0cSJaegeuk Kim spin_lock(&sbi->stat_lock); 645*39a53e0cSJaegeuk Kim 646*39a53e0cSJaegeuk Kim BUG_ON(sbi->total_valid_block_count < count); 647*39a53e0cSJaegeuk Kim BUG_ON(sbi->total_valid_node_count < count); 648*39a53e0cSJaegeuk Kim BUG_ON(inode->i_blocks < count); 649*39a53e0cSJaegeuk Kim 650*39a53e0cSJaegeuk Kim inode->i_blocks -= count; 651*39a53e0cSJaegeuk Kim sbi->total_valid_node_count -= count; 652*39a53e0cSJaegeuk Kim sbi->total_valid_block_count -= (block_t)count; 653*39a53e0cSJaegeuk Kim 654*39a53e0cSJaegeuk Kim spin_unlock(&sbi->stat_lock); 655*39a53e0cSJaegeuk Kim } 656*39a53e0cSJaegeuk Kim 657*39a53e0cSJaegeuk Kim static inline unsigned int valid_node_count(struct f2fs_sb_info *sbi) 658*39a53e0cSJaegeuk Kim { 659*39a53e0cSJaegeuk Kim unsigned int ret; 660*39a53e0cSJaegeuk Kim spin_lock(&sbi->stat_lock); 661*39a53e0cSJaegeuk Kim ret = sbi->total_valid_node_count; 662*39a53e0cSJaegeuk Kim spin_unlock(&sbi->stat_lock); 663*39a53e0cSJaegeuk Kim return ret; 664*39a53e0cSJaegeuk Kim } 665*39a53e0cSJaegeuk Kim 666*39a53e0cSJaegeuk Kim static inline void inc_valid_inode_count(struct f2fs_sb_info *sbi) 667*39a53e0cSJaegeuk Kim { 668*39a53e0cSJaegeuk Kim spin_lock(&sbi->stat_lock); 669*39a53e0cSJaegeuk Kim BUG_ON(sbi->total_valid_inode_count == sbi->total_node_count); 670*39a53e0cSJaegeuk Kim sbi->total_valid_inode_count++; 671*39a53e0cSJaegeuk Kim spin_unlock(&sbi->stat_lock); 672*39a53e0cSJaegeuk Kim } 673*39a53e0cSJaegeuk Kim 674*39a53e0cSJaegeuk Kim static inline int dec_valid_inode_count(struct f2fs_sb_info *sbi) 675*39a53e0cSJaegeuk Kim { 676*39a53e0cSJaegeuk Kim spin_lock(&sbi->stat_lock); 677*39a53e0cSJaegeuk Kim BUG_ON(!sbi->total_valid_inode_count); 678*39a53e0cSJaegeuk Kim sbi->total_valid_inode_count--; 679*39a53e0cSJaegeuk Kim spin_unlock(&sbi->stat_lock); 680*39a53e0cSJaegeuk Kim return 0; 681*39a53e0cSJaegeuk Kim } 682*39a53e0cSJaegeuk Kim 683*39a53e0cSJaegeuk Kim static inline unsigned int valid_inode_count(struct f2fs_sb_info *sbi) 684*39a53e0cSJaegeuk Kim { 685*39a53e0cSJaegeuk Kim unsigned int ret; 686*39a53e0cSJaegeuk Kim spin_lock(&sbi->stat_lock); 687*39a53e0cSJaegeuk Kim ret = sbi->total_valid_inode_count; 688*39a53e0cSJaegeuk Kim spin_unlock(&sbi->stat_lock); 689*39a53e0cSJaegeuk Kim return ret; 690*39a53e0cSJaegeuk Kim } 691*39a53e0cSJaegeuk Kim 692*39a53e0cSJaegeuk Kim static inline void f2fs_put_page(struct page *page, int unlock) 693*39a53e0cSJaegeuk Kim { 694*39a53e0cSJaegeuk Kim if (!page || IS_ERR(page)) 695*39a53e0cSJaegeuk Kim return; 696*39a53e0cSJaegeuk Kim 697*39a53e0cSJaegeuk Kim if (unlock) { 698*39a53e0cSJaegeuk Kim BUG_ON(!PageLocked(page)); 699*39a53e0cSJaegeuk Kim unlock_page(page); 700*39a53e0cSJaegeuk Kim } 701*39a53e0cSJaegeuk Kim page_cache_release(page); 702*39a53e0cSJaegeuk Kim } 703*39a53e0cSJaegeuk Kim 704*39a53e0cSJaegeuk Kim static inline void f2fs_put_dnode(struct dnode_of_data *dn) 705*39a53e0cSJaegeuk Kim { 706*39a53e0cSJaegeuk Kim if (dn->node_page) 707*39a53e0cSJaegeuk Kim f2fs_put_page(dn->node_page, 1); 708*39a53e0cSJaegeuk Kim if (dn->inode_page && dn->node_page != dn->inode_page) 709*39a53e0cSJaegeuk Kim f2fs_put_page(dn->inode_page, 0); 710*39a53e0cSJaegeuk Kim dn->node_page = NULL; 711*39a53e0cSJaegeuk Kim dn->inode_page = NULL; 712*39a53e0cSJaegeuk Kim } 713*39a53e0cSJaegeuk Kim 714*39a53e0cSJaegeuk Kim static inline struct kmem_cache *f2fs_kmem_cache_create(const char *name, 715*39a53e0cSJaegeuk Kim size_t size, void (*ctor)(void *)) 716*39a53e0cSJaegeuk Kim { 717*39a53e0cSJaegeuk Kim return kmem_cache_create(name, size, 0, SLAB_RECLAIM_ACCOUNT, ctor); 718*39a53e0cSJaegeuk Kim } 719*39a53e0cSJaegeuk Kim 720*39a53e0cSJaegeuk Kim #define RAW_IS_INODE(p) ((p)->footer.nid == (p)->footer.ino) 721*39a53e0cSJaegeuk Kim 722*39a53e0cSJaegeuk Kim static inline bool IS_INODE(struct page *page) 723*39a53e0cSJaegeuk Kim { 724*39a53e0cSJaegeuk Kim struct f2fs_node *p = (struct f2fs_node *)page_address(page); 725*39a53e0cSJaegeuk Kim return RAW_IS_INODE(p); 726*39a53e0cSJaegeuk Kim } 727*39a53e0cSJaegeuk Kim 728*39a53e0cSJaegeuk Kim static inline __le32 *blkaddr_in_node(struct f2fs_node *node) 729*39a53e0cSJaegeuk Kim { 730*39a53e0cSJaegeuk Kim return RAW_IS_INODE(node) ? node->i.i_addr : node->dn.addr; 731*39a53e0cSJaegeuk Kim } 732*39a53e0cSJaegeuk Kim 733*39a53e0cSJaegeuk Kim static inline block_t datablock_addr(struct page *node_page, 734*39a53e0cSJaegeuk Kim unsigned int offset) 735*39a53e0cSJaegeuk Kim { 736*39a53e0cSJaegeuk Kim struct f2fs_node *raw_node; 737*39a53e0cSJaegeuk Kim __le32 *addr_array; 738*39a53e0cSJaegeuk Kim raw_node = (struct f2fs_node *)page_address(node_page); 739*39a53e0cSJaegeuk Kim addr_array = blkaddr_in_node(raw_node); 740*39a53e0cSJaegeuk Kim return le32_to_cpu(addr_array[offset]); 741*39a53e0cSJaegeuk Kim } 742*39a53e0cSJaegeuk Kim 743*39a53e0cSJaegeuk Kim static inline int f2fs_test_bit(unsigned int nr, char *addr) 744*39a53e0cSJaegeuk Kim { 745*39a53e0cSJaegeuk Kim int mask; 746*39a53e0cSJaegeuk Kim 747*39a53e0cSJaegeuk Kim addr += (nr >> 3); 748*39a53e0cSJaegeuk Kim mask = 1 << (7 - (nr & 0x07)); 749*39a53e0cSJaegeuk Kim return mask & *addr; 750*39a53e0cSJaegeuk Kim } 751*39a53e0cSJaegeuk Kim 752*39a53e0cSJaegeuk Kim static inline int f2fs_set_bit(unsigned int nr, char *addr) 753*39a53e0cSJaegeuk Kim { 754*39a53e0cSJaegeuk Kim int mask; 755*39a53e0cSJaegeuk Kim int ret; 756*39a53e0cSJaegeuk Kim 757*39a53e0cSJaegeuk Kim addr += (nr >> 3); 758*39a53e0cSJaegeuk Kim mask = 1 << (7 - (nr & 0x07)); 759*39a53e0cSJaegeuk Kim ret = mask & *addr; 760*39a53e0cSJaegeuk Kim *addr |= mask; 761*39a53e0cSJaegeuk Kim return ret; 762*39a53e0cSJaegeuk Kim } 763*39a53e0cSJaegeuk Kim 764*39a53e0cSJaegeuk Kim static inline int f2fs_clear_bit(unsigned int nr, char *addr) 765*39a53e0cSJaegeuk Kim { 766*39a53e0cSJaegeuk Kim int mask; 767*39a53e0cSJaegeuk Kim int ret; 768*39a53e0cSJaegeuk Kim 769*39a53e0cSJaegeuk Kim addr += (nr >> 3); 770*39a53e0cSJaegeuk Kim mask = 1 << (7 - (nr & 0x07)); 771*39a53e0cSJaegeuk Kim ret = mask & *addr; 772*39a53e0cSJaegeuk Kim *addr &= ~mask; 773*39a53e0cSJaegeuk Kim return ret; 774*39a53e0cSJaegeuk Kim } 775*39a53e0cSJaegeuk Kim 776*39a53e0cSJaegeuk Kim /* used for f2fs_inode_info->flags */ 777*39a53e0cSJaegeuk Kim enum { 778*39a53e0cSJaegeuk Kim FI_NEW_INODE, /* indicate newly allocated inode */ 779*39a53e0cSJaegeuk Kim FI_NEED_CP, /* need to do checkpoint during fsync */ 780*39a53e0cSJaegeuk Kim FI_INC_LINK, /* need to increment i_nlink */ 781*39a53e0cSJaegeuk Kim FI_ACL_MODE, /* indicate acl mode */ 782*39a53e0cSJaegeuk Kim FI_NO_ALLOC, /* should not allocate any blocks */ 783*39a53e0cSJaegeuk Kim }; 784*39a53e0cSJaegeuk Kim 785*39a53e0cSJaegeuk Kim static inline void set_inode_flag(struct f2fs_inode_info *fi, int flag) 786*39a53e0cSJaegeuk Kim { 787*39a53e0cSJaegeuk Kim set_bit(flag, &fi->flags); 788*39a53e0cSJaegeuk Kim } 789*39a53e0cSJaegeuk Kim 790*39a53e0cSJaegeuk Kim static inline int is_inode_flag_set(struct f2fs_inode_info *fi, int flag) 791*39a53e0cSJaegeuk Kim { 792*39a53e0cSJaegeuk Kim return test_bit(flag, &fi->flags); 793*39a53e0cSJaegeuk Kim } 794*39a53e0cSJaegeuk Kim 795*39a53e0cSJaegeuk Kim static inline void clear_inode_flag(struct f2fs_inode_info *fi, int flag) 796*39a53e0cSJaegeuk Kim { 797*39a53e0cSJaegeuk Kim clear_bit(flag, &fi->flags); 798*39a53e0cSJaegeuk Kim } 799*39a53e0cSJaegeuk Kim 800*39a53e0cSJaegeuk Kim static inline void set_acl_inode(struct f2fs_inode_info *fi, umode_t mode) 801*39a53e0cSJaegeuk Kim { 802*39a53e0cSJaegeuk Kim fi->i_acl_mode = mode; 803*39a53e0cSJaegeuk Kim set_inode_flag(fi, FI_ACL_MODE); 804*39a53e0cSJaegeuk Kim } 805*39a53e0cSJaegeuk Kim 806*39a53e0cSJaegeuk Kim static inline int cond_clear_inode_flag(struct f2fs_inode_info *fi, int flag) 807*39a53e0cSJaegeuk Kim { 808*39a53e0cSJaegeuk Kim if (is_inode_flag_set(fi, FI_ACL_MODE)) { 809*39a53e0cSJaegeuk Kim clear_inode_flag(fi, FI_ACL_MODE); 810*39a53e0cSJaegeuk Kim return 1; 811*39a53e0cSJaegeuk Kim } 812*39a53e0cSJaegeuk Kim return 0; 813*39a53e0cSJaegeuk Kim } 814*39a53e0cSJaegeuk Kim 815*39a53e0cSJaegeuk Kim /* 816*39a53e0cSJaegeuk Kim * file.c 817*39a53e0cSJaegeuk Kim */ 818*39a53e0cSJaegeuk Kim int f2fs_sync_file(struct file *, loff_t, loff_t, int); 819*39a53e0cSJaegeuk Kim void truncate_data_blocks(struct dnode_of_data *); 820*39a53e0cSJaegeuk Kim void f2fs_truncate(struct inode *); 821*39a53e0cSJaegeuk Kim int f2fs_setattr(struct dentry *, struct iattr *); 822*39a53e0cSJaegeuk Kim int truncate_hole(struct inode *, pgoff_t, pgoff_t); 823*39a53e0cSJaegeuk Kim long f2fs_ioctl(struct file *, unsigned int, unsigned long); 824*39a53e0cSJaegeuk Kim 825*39a53e0cSJaegeuk Kim /* 826*39a53e0cSJaegeuk Kim * inode.c 827*39a53e0cSJaegeuk Kim */ 828*39a53e0cSJaegeuk Kim void f2fs_set_inode_flags(struct inode *); 829*39a53e0cSJaegeuk Kim struct inode *f2fs_iget_nowait(struct super_block *, unsigned long); 830*39a53e0cSJaegeuk Kim struct inode *f2fs_iget(struct super_block *, unsigned long); 831*39a53e0cSJaegeuk Kim void update_inode(struct inode *, struct page *); 832*39a53e0cSJaegeuk Kim int f2fs_write_inode(struct inode *, struct writeback_control *); 833*39a53e0cSJaegeuk Kim void f2fs_evict_inode(struct inode *); 834*39a53e0cSJaegeuk Kim 835*39a53e0cSJaegeuk Kim /* 836*39a53e0cSJaegeuk Kim * namei.c 837*39a53e0cSJaegeuk Kim */ 838*39a53e0cSJaegeuk Kim struct dentry *f2fs_get_parent(struct dentry *child); 839*39a53e0cSJaegeuk Kim 840*39a53e0cSJaegeuk Kim /* 841*39a53e0cSJaegeuk Kim * dir.c 842*39a53e0cSJaegeuk Kim */ 843*39a53e0cSJaegeuk Kim struct f2fs_dir_entry *f2fs_find_entry(struct inode *, struct qstr *, 844*39a53e0cSJaegeuk Kim struct page **); 845*39a53e0cSJaegeuk Kim struct f2fs_dir_entry *f2fs_parent_dir(struct inode *, struct page **); 846*39a53e0cSJaegeuk Kim ino_t f2fs_inode_by_name(struct inode *, struct qstr *); 847*39a53e0cSJaegeuk Kim void f2fs_set_link(struct inode *, struct f2fs_dir_entry *, 848*39a53e0cSJaegeuk Kim struct page *, struct inode *); 849*39a53e0cSJaegeuk Kim void init_dent_inode(struct dentry *, struct page *); 850*39a53e0cSJaegeuk Kim int f2fs_add_link(struct dentry *, struct inode *); 851*39a53e0cSJaegeuk Kim void f2fs_delete_entry(struct f2fs_dir_entry *, struct page *, struct inode *); 852*39a53e0cSJaegeuk Kim int f2fs_make_empty(struct inode *, struct inode *); 853*39a53e0cSJaegeuk Kim bool f2fs_empty_dir(struct inode *); 854*39a53e0cSJaegeuk Kim 855*39a53e0cSJaegeuk Kim /* 856*39a53e0cSJaegeuk Kim * super.c 857*39a53e0cSJaegeuk Kim */ 858*39a53e0cSJaegeuk Kim int f2fs_sync_fs(struct super_block *, int); 859*39a53e0cSJaegeuk Kim 860*39a53e0cSJaegeuk Kim /* 861*39a53e0cSJaegeuk Kim * hash.c 862*39a53e0cSJaegeuk Kim */ 863*39a53e0cSJaegeuk Kim f2fs_hash_t f2fs_dentry_hash(const char *, int); 864*39a53e0cSJaegeuk Kim 865*39a53e0cSJaegeuk Kim /* 866*39a53e0cSJaegeuk Kim * node.c 867*39a53e0cSJaegeuk Kim */ 868*39a53e0cSJaegeuk Kim struct dnode_of_data; 869*39a53e0cSJaegeuk Kim struct node_info; 870*39a53e0cSJaegeuk Kim 871*39a53e0cSJaegeuk Kim int is_checkpointed_node(struct f2fs_sb_info *, nid_t); 872*39a53e0cSJaegeuk Kim void get_node_info(struct f2fs_sb_info *, nid_t, struct node_info *); 873*39a53e0cSJaegeuk Kim int get_dnode_of_data(struct dnode_of_data *, pgoff_t, int); 874*39a53e0cSJaegeuk Kim int truncate_inode_blocks(struct inode *, pgoff_t); 875*39a53e0cSJaegeuk Kim int remove_inode_page(struct inode *); 876*39a53e0cSJaegeuk Kim int new_inode_page(struct inode *, struct dentry *); 877*39a53e0cSJaegeuk Kim struct page *new_node_page(struct dnode_of_data *, unsigned int); 878*39a53e0cSJaegeuk Kim void ra_node_page(struct f2fs_sb_info *, nid_t); 879*39a53e0cSJaegeuk Kim struct page *get_node_page(struct f2fs_sb_info *, pgoff_t); 880*39a53e0cSJaegeuk Kim struct page *get_node_page_ra(struct page *, int); 881*39a53e0cSJaegeuk Kim void sync_inode_page(struct dnode_of_data *); 882*39a53e0cSJaegeuk Kim int sync_node_pages(struct f2fs_sb_info *, nid_t, struct writeback_control *); 883*39a53e0cSJaegeuk Kim bool alloc_nid(struct f2fs_sb_info *, nid_t *); 884*39a53e0cSJaegeuk Kim void alloc_nid_done(struct f2fs_sb_info *, nid_t); 885*39a53e0cSJaegeuk Kim void alloc_nid_failed(struct f2fs_sb_info *, nid_t); 886*39a53e0cSJaegeuk Kim void recover_node_page(struct f2fs_sb_info *, struct page *, 887*39a53e0cSJaegeuk Kim struct f2fs_summary *, struct node_info *, block_t); 888*39a53e0cSJaegeuk Kim int recover_inode_page(struct f2fs_sb_info *, struct page *); 889*39a53e0cSJaegeuk Kim int restore_node_summary(struct f2fs_sb_info *, unsigned int, 890*39a53e0cSJaegeuk Kim struct f2fs_summary_block *); 891*39a53e0cSJaegeuk Kim void flush_nat_entries(struct f2fs_sb_info *); 892*39a53e0cSJaegeuk Kim int build_node_manager(struct f2fs_sb_info *); 893*39a53e0cSJaegeuk Kim void destroy_node_manager(struct f2fs_sb_info *); 894*39a53e0cSJaegeuk Kim int create_node_manager_caches(void); 895*39a53e0cSJaegeuk Kim void destroy_node_manager_caches(void); 896*39a53e0cSJaegeuk Kim 897*39a53e0cSJaegeuk Kim /* 898*39a53e0cSJaegeuk Kim * segment.c 899*39a53e0cSJaegeuk Kim */ 900*39a53e0cSJaegeuk Kim void f2fs_balance_fs(struct f2fs_sb_info *); 901*39a53e0cSJaegeuk Kim void invalidate_blocks(struct f2fs_sb_info *, block_t); 902*39a53e0cSJaegeuk Kim void locate_dirty_segment(struct f2fs_sb_info *, unsigned int); 903*39a53e0cSJaegeuk Kim void clear_prefree_segments(struct f2fs_sb_info *); 904*39a53e0cSJaegeuk Kim int npages_for_summary_flush(struct f2fs_sb_info *); 905*39a53e0cSJaegeuk Kim void allocate_new_segments(struct f2fs_sb_info *); 906*39a53e0cSJaegeuk Kim struct page *get_sum_page(struct f2fs_sb_info *, unsigned int); 907*39a53e0cSJaegeuk Kim struct bio *f2fs_bio_alloc(struct block_device *, sector_t, int, gfp_t); 908*39a53e0cSJaegeuk Kim void f2fs_submit_bio(struct f2fs_sb_info *, enum page_type, bool sync); 909*39a53e0cSJaegeuk Kim int write_meta_page(struct f2fs_sb_info *, struct page *, 910*39a53e0cSJaegeuk Kim struct writeback_control *); 911*39a53e0cSJaegeuk Kim void write_node_page(struct f2fs_sb_info *, struct page *, unsigned int, 912*39a53e0cSJaegeuk Kim block_t, block_t *); 913*39a53e0cSJaegeuk Kim void write_data_page(struct inode *, struct page *, struct dnode_of_data*, 914*39a53e0cSJaegeuk Kim block_t, block_t *); 915*39a53e0cSJaegeuk Kim void rewrite_data_page(struct f2fs_sb_info *, struct page *, block_t); 916*39a53e0cSJaegeuk Kim void recover_data_page(struct f2fs_sb_info *, struct page *, 917*39a53e0cSJaegeuk Kim struct f2fs_summary *, block_t, block_t); 918*39a53e0cSJaegeuk Kim void rewrite_node_page(struct f2fs_sb_info *, struct page *, 919*39a53e0cSJaegeuk Kim struct f2fs_summary *, block_t, block_t); 920*39a53e0cSJaegeuk Kim void write_data_summaries(struct f2fs_sb_info *, block_t); 921*39a53e0cSJaegeuk Kim void write_node_summaries(struct f2fs_sb_info *, block_t); 922*39a53e0cSJaegeuk Kim int lookup_journal_in_cursum(struct f2fs_summary_block *, 923*39a53e0cSJaegeuk Kim int, unsigned int, int); 924*39a53e0cSJaegeuk Kim void flush_sit_entries(struct f2fs_sb_info *); 925*39a53e0cSJaegeuk Kim int build_segment_manager(struct f2fs_sb_info *); 926*39a53e0cSJaegeuk Kim void reset_victim_segmap(struct f2fs_sb_info *); 927*39a53e0cSJaegeuk Kim void destroy_segment_manager(struct f2fs_sb_info *); 928*39a53e0cSJaegeuk Kim 929*39a53e0cSJaegeuk Kim /* 930*39a53e0cSJaegeuk Kim * checkpoint.c 931*39a53e0cSJaegeuk Kim */ 932*39a53e0cSJaegeuk Kim struct page *grab_meta_page(struct f2fs_sb_info *, pgoff_t); 933*39a53e0cSJaegeuk Kim struct page *get_meta_page(struct f2fs_sb_info *, pgoff_t); 934*39a53e0cSJaegeuk Kim long sync_meta_pages(struct f2fs_sb_info *, enum page_type, long); 935*39a53e0cSJaegeuk Kim int check_orphan_space(struct f2fs_sb_info *); 936*39a53e0cSJaegeuk Kim void add_orphan_inode(struct f2fs_sb_info *, nid_t); 937*39a53e0cSJaegeuk Kim void remove_orphan_inode(struct f2fs_sb_info *, nid_t); 938*39a53e0cSJaegeuk Kim int recover_orphan_inodes(struct f2fs_sb_info *); 939*39a53e0cSJaegeuk Kim int get_valid_checkpoint(struct f2fs_sb_info *); 940*39a53e0cSJaegeuk Kim void set_dirty_dir_page(struct inode *, struct page *); 941*39a53e0cSJaegeuk Kim void remove_dirty_dir_inode(struct inode *); 942*39a53e0cSJaegeuk Kim void sync_dirty_dir_inodes(struct f2fs_sb_info *); 943*39a53e0cSJaegeuk Kim void block_operations(struct f2fs_sb_info *); 944*39a53e0cSJaegeuk Kim void write_checkpoint(struct f2fs_sb_info *, bool, bool); 945*39a53e0cSJaegeuk Kim void init_orphan_info(struct f2fs_sb_info *); 946*39a53e0cSJaegeuk Kim int create_checkpoint_caches(void); 947*39a53e0cSJaegeuk Kim void destroy_checkpoint_caches(void); 948*39a53e0cSJaegeuk Kim 949*39a53e0cSJaegeuk Kim /* 950*39a53e0cSJaegeuk Kim * data.c 951*39a53e0cSJaegeuk Kim */ 952*39a53e0cSJaegeuk Kim int reserve_new_block(struct dnode_of_data *); 953*39a53e0cSJaegeuk Kim void update_extent_cache(block_t, struct dnode_of_data *); 954*39a53e0cSJaegeuk Kim struct page *find_data_page(struct inode *, pgoff_t); 955*39a53e0cSJaegeuk Kim struct page *get_lock_data_page(struct inode *, pgoff_t); 956*39a53e0cSJaegeuk Kim struct page *get_new_data_page(struct inode *, pgoff_t, bool); 957*39a53e0cSJaegeuk Kim int f2fs_readpage(struct f2fs_sb_info *, struct page *, block_t, int); 958*39a53e0cSJaegeuk Kim int do_write_data_page(struct page *); 959*39a53e0cSJaegeuk Kim 960*39a53e0cSJaegeuk Kim /* 961*39a53e0cSJaegeuk Kim * gc.c 962*39a53e0cSJaegeuk Kim */ 963*39a53e0cSJaegeuk Kim int start_gc_thread(struct f2fs_sb_info *); 964*39a53e0cSJaegeuk Kim void stop_gc_thread(struct f2fs_sb_info *); 965*39a53e0cSJaegeuk Kim block_t start_bidx_of_node(unsigned int); 966*39a53e0cSJaegeuk Kim int f2fs_gc(struct f2fs_sb_info *, int); 967*39a53e0cSJaegeuk Kim void build_gc_manager(struct f2fs_sb_info *); 968*39a53e0cSJaegeuk Kim int create_gc_caches(void); 969*39a53e0cSJaegeuk Kim void destroy_gc_caches(void); 970*39a53e0cSJaegeuk Kim 971*39a53e0cSJaegeuk Kim /* 972*39a53e0cSJaegeuk Kim * recovery.c 973*39a53e0cSJaegeuk Kim */ 974*39a53e0cSJaegeuk Kim void recover_fsync_data(struct f2fs_sb_info *); 975*39a53e0cSJaegeuk Kim bool space_for_roll_forward(struct f2fs_sb_info *); 976*39a53e0cSJaegeuk Kim 977*39a53e0cSJaegeuk Kim /* 978*39a53e0cSJaegeuk Kim * debug.c 979*39a53e0cSJaegeuk Kim */ 980*39a53e0cSJaegeuk Kim #ifdef CONFIG_F2FS_STAT_FS 981*39a53e0cSJaegeuk Kim struct f2fs_stat_info { 982*39a53e0cSJaegeuk Kim struct list_head stat_list; 983*39a53e0cSJaegeuk Kim struct f2fs_sb_info *sbi; 984*39a53e0cSJaegeuk Kim struct mutex stat_lock; 985*39a53e0cSJaegeuk Kim int all_area_segs, sit_area_segs, nat_area_segs, ssa_area_segs; 986*39a53e0cSJaegeuk Kim int main_area_segs, main_area_sections, main_area_zones; 987*39a53e0cSJaegeuk Kim int hit_ext, total_ext; 988*39a53e0cSJaegeuk Kim int ndirty_node, ndirty_dent, ndirty_dirs, ndirty_meta; 989*39a53e0cSJaegeuk Kim int nats, sits, fnids; 990*39a53e0cSJaegeuk Kim int total_count, utilization; 991*39a53e0cSJaegeuk Kim int bg_gc; 992*39a53e0cSJaegeuk Kim unsigned int valid_count, valid_node_count, valid_inode_count; 993*39a53e0cSJaegeuk Kim unsigned int bimodal, avg_vblocks; 994*39a53e0cSJaegeuk Kim int util_free, util_valid, util_invalid; 995*39a53e0cSJaegeuk Kim int rsvd_segs, overp_segs; 996*39a53e0cSJaegeuk Kim int dirty_count, node_pages, meta_pages; 997*39a53e0cSJaegeuk Kim int prefree_count, call_count; 998*39a53e0cSJaegeuk Kim int tot_segs, node_segs, data_segs, free_segs, free_secs; 999*39a53e0cSJaegeuk Kim int tot_blks, data_blks, node_blks; 1000*39a53e0cSJaegeuk Kim int curseg[NR_CURSEG_TYPE]; 1001*39a53e0cSJaegeuk Kim int cursec[NR_CURSEG_TYPE]; 1002*39a53e0cSJaegeuk Kim int curzone[NR_CURSEG_TYPE]; 1003*39a53e0cSJaegeuk Kim 1004*39a53e0cSJaegeuk Kim unsigned int segment_count[2]; 1005*39a53e0cSJaegeuk Kim unsigned int block_count[2]; 1006*39a53e0cSJaegeuk Kim unsigned base_mem, cache_mem; 1007*39a53e0cSJaegeuk Kim }; 1008*39a53e0cSJaegeuk Kim 1009*39a53e0cSJaegeuk Kim #define stat_inc_call_count(si) ((si)->call_count++) 1010*39a53e0cSJaegeuk Kim 1011*39a53e0cSJaegeuk Kim #define stat_inc_seg_count(sbi, type) \ 1012*39a53e0cSJaegeuk Kim do { \ 1013*39a53e0cSJaegeuk Kim struct f2fs_stat_info *si = sbi->stat_info; \ 1014*39a53e0cSJaegeuk Kim (si)->tot_segs++; \ 1015*39a53e0cSJaegeuk Kim if (type == SUM_TYPE_DATA) \ 1016*39a53e0cSJaegeuk Kim si->data_segs++; \ 1017*39a53e0cSJaegeuk Kim else \ 1018*39a53e0cSJaegeuk Kim si->node_segs++; \ 1019*39a53e0cSJaegeuk Kim } while (0) 1020*39a53e0cSJaegeuk Kim 1021*39a53e0cSJaegeuk Kim #define stat_inc_tot_blk_count(si, blks) \ 1022*39a53e0cSJaegeuk Kim (si->tot_blks += (blks)) 1023*39a53e0cSJaegeuk Kim 1024*39a53e0cSJaegeuk Kim #define stat_inc_data_blk_count(sbi, blks) \ 1025*39a53e0cSJaegeuk Kim do { \ 1026*39a53e0cSJaegeuk Kim struct f2fs_stat_info *si = sbi->stat_info; \ 1027*39a53e0cSJaegeuk Kim stat_inc_tot_blk_count(si, blks); \ 1028*39a53e0cSJaegeuk Kim si->data_blks += (blks); \ 1029*39a53e0cSJaegeuk Kim } while (0) 1030*39a53e0cSJaegeuk Kim 1031*39a53e0cSJaegeuk Kim #define stat_inc_node_blk_count(sbi, blks) \ 1032*39a53e0cSJaegeuk Kim do { \ 1033*39a53e0cSJaegeuk Kim struct f2fs_stat_info *si = sbi->stat_info; \ 1034*39a53e0cSJaegeuk Kim stat_inc_tot_blk_count(si, blks); \ 1035*39a53e0cSJaegeuk Kim si->node_blks += (blks); \ 1036*39a53e0cSJaegeuk Kim } while (0) 1037*39a53e0cSJaegeuk Kim 1038*39a53e0cSJaegeuk Kim int f2fs_build_stats(struct f2fs_sb_info *); 1039*39a53e0cSJaegeuk Kim void f2fs_destroy_stats(struct f2fs_sb_info *); 1040*39a53e0cSJaegeuk Kim void destroy_root_stats(void); 1041*39a53e0cSJaegeuk Kim #else 1042*39a53e0cSJaegeuk Kim #define stat_inc_call_count(si) 1043*39a53e0cSJaegeuk Kim #define stat_inc_seg_count(si, type) 1044*39a53e0cSJaegeuk Kim #define stat_inc_tot_blk_count(si, blks) 1045*39a53e0cSJaegeuk Kim #define stat_inc_data_blk_count(si, blks) 1046*39a53e0cSJaegeuk Kim #define stat_inc_node_blk_count(sbi, blks) 1047*39a53e0cSJaegeuk Kim 1048*39a53e0cSJaegeuk Kim static inline int f2fs_build_stats(struct f2fs_sb_info *sbi) { return 0; } 1049*39a53e0cSJaegeuk Kim static inline void f2fs_destroy_stats(struct f2fs_sb_info *sbi) { } 1050*39a53e0cSJaegeuk Kim static inline void destroy_root_stats(void) { } 1051*39a53e0cSJaegeuk Kim #endif 1052*39a53e0cSJaegeuk Kim 1053*39a53e0cSJaegeuk Kim extern const struct file_operations f2fs_dir_operations; 1054*39a53e0cSJaegeuk Kim extern const struct file_operations f2fs_file_operations; 1055*39a53e0cSJaegeuk Kim extern const struct inode_operations f2fs_file_inode_operations; 1056*39a53e0cSJaegeuk Kim extern const struct address_space_operations f2fs_dblock_aops; 1057*39a53e0cSJaegeuk Kim extern const struct address_space_operations f2fs_node_aops; 1058*39a53e0cSJaegeuk Kim extern const struct address_space_operations f2fs_meta_aops; 1059*39a53e0cSJaegeuk Kim extern const struct inode_operations f2fs_dir_inode_operations; 1060*39a53e0cSJaegeuk Kim extern const struct inode_operations f2fs_symlink_inode_operations; 1061*39a53e0cSJaegeuk Kim extern const struct inode_operations f2fs_special_inode_operations; 1062*39a53e0cSJaegeuk Kim #endif 1063