1b16817b6SDave Chinner /* SPDX-License-Identifier: GPL-2.0 */ 2b16817b6SDave Chinner /* 3b16817b6SDave Chinner * Copyright (c) 2000-2005 Silicon Graphics, Inc. 4b16817b6SDave Chinner * Copyright (c) 2018 Red Hat, Inc. 5b16817b6SDave Chinner * All rights reserved. 6b16817b6SDave Chinner */ 7b16817b6SDave Chinner 8b16817b6SDave Chinner #include "xfs.h" 9b16817b6SDave Chinner #include "xfs_fs.h" 10b16817b6SDave Chinner #include "xfs_shared.h" 11b16817b6SDave Chinner #include "xfs_format.h" 12b16817b6SDave Chinner #include "xfs_trans_resv.h" 13*f327a007SDarrick J. Wong #include "xfs_bit.h" 14b16817b6SDave Chinner #include "xfs_sb.h" 15b16817b6SDave Chinner #include "xfs_mount.h" 16b16817b6SDave Chinner #include "xfs_btree.h" 17b16817b6SDave Chinner #include "xfs_alloc_btree.h" 18b16817b6SDave Chinner #include "xfs_rmap_btree.h" 19b16817b6SDave Chinner #include "xfs_alloc.h" 2049dd56f2SDave Chinner #include "xfs_ialloc.h" 21b16817b6SDave Chinner #include "xfs_rmap.h" 22b16817b6SDave Chinner #include "xfs_ag.h" 237cd5006bSDarrick J. Wong #include "xfs_ag_resv.h" 241302c6a2SDarrick J. Wong #include "xfs_health.h" 25b16817b6SDave Chinner 26b16817b6SDave Chinner static struct xfs_buf * 27b16817b6SDave Chinner xfs_get_aghdr_buf( 28b16817b6SDave Chinner struct xfs_mount *mp, 29b16817b6SDave Chinner xfs_daddr_t blkno, 30b16817b6SDave Chinner size_t numblks, 31b16817b6SDave Chinner int flags, 32b16817b6SDave Chinner const struct xfs_buf_ops *ops) 33b16817b6SDave Chinner { 34b16817b6SDave Chinner struct xfs_buf *bp; 35b16817b6SDave Chinner 36b16817b6SDave Chinner bp = xfs_buf_get_uncached(mp->m_ddev_targp, numblks, flags); 37b16817b6SDave Chinner if (!bp) 38b16817b6SDave Chinner return NULL; 39b16817b6SDave Chinner 40b16817b6SDave Chinner xfs_buf_zero(bp, 0, BBTOB(bp->b_length)); 41b16817b6SDave Chinner bp->b_bn = blkno; 42b16817b6SDave Chinner bp->b_maps[0].bm_bn = blkno; 43b16817b6SDave Chinner bp->b_ops = ops; 44b16817b6SDave Chinner 45b16817b6SDave Chinner return bp; 46b16817b6SDave Chinner } 47b16817b6SDave Chinner 48*f327a007SDarrick J. Wong static inline bool is_log_ag(struct xfs_mount *mp, struct aghdr_init_data *id) 49*f327a007SDarrick J. Wong { 50*f327a007SDarrick J. Wong return mp->m_sb.sb_logstart > 0 && 51*f327a007SDarrick J. Wong id->agno == XFS_FSB_TO_AGNO(mp, mp->m_sb.sb_logstart); 52*f327a007SDarrick J. Wong } 53*f327a007SDarrick J. Wong 54b16817b6SDave Chinner /* 55b16817b6SDave Chinner * Generic btree root block init function 56b16817b6SDave Chinner */ 57b16817b6SDave Chinner static void 58b16817b6SDave Chinner xfs_btroot_init( 59b16817b6SDave Chinner struct xfs_mount *mp, 60b16817b6SDave Chinner struct xfs_buf *bp, 61b16817b6SDave Chinner struct aghdr_init_data *id) 62b16817b6SDave Chinner { 63f5b999c0SEric Sandeen xfs_btree_init_block(mp, bp, id->type, 0, 0, id->agno); 64b16817b6SDave Chinner } 65b16817b6SDave Chinner 668d90857cSDarrick J. Wong /* Finish initializing a free space btree. */ 678d90857cSDarrick J. Wong static void 688d90857cSDarrick J. Wong xfs_freesp_init_recs( 698d90857cSDarrick J. Wong struct xfs_mount *mp, 708d90857cSDarrick J. Wong struct xfs_buf *bp, 718d90857cSDarrick J. Wong struct aghdr_init_data *id) 728d90857cSDarrick J. Wong { 738d90857cSDarrick J. Wong struct xfs_alloc_rec *arec; 74*f327a007SDarrick J. Wong struct xfs_btree_block *block = XFS_BUF_TO_BLOCK(bp); 758d90857cSDarrick J. Wong 768d90857cSDarrick J. Wong arec = XFS_ALLOC_REC_ADDR(mp, XFS_BUF_TO_BLOCK(bp), 1); 778d90857cSDarrick J. Wong arec->ar_startblock = cpu_to_be32(mp->m_ag_prealloc_blocks); 78*f327a007SDarrick J. Wong 79*f327a007SDarrick J. Wong if (is_log_ag(mp, id)) { 80*f327a007SDarrick J. Wong struct xfs_alloc_rec *nrec; 81*f327a007SDarrick J. Wong xfs_agblock_t start = XFS_FSB_TO_AGBNO(mp, 82*f327a007SDarrick J. Wong mp->m_sb.sb_logstart); 83*f327a007SDarrick J. Wong 84*f327a007SDarrick J. Wong ASSERT(start >= mp->m_ag_prealloc_blocks); 85*f327a007SDarrick J. Wong if (start != mp->m_ag_prealloc_blocks) { 86*f327a007SDarrick J. Wong /* 87*f327a007SDarrick J. Wong * Modify first record to pad stripe align of log 88*f327a007SDarrick J. Wong */ 89*f327a007SDarrick J. Wong arec->ar_blockcount = cpu_to_be32(start - 90*f327a007SDarrick J. Wong mp->m_ag_prealloc_blocks); 91*f327a007SDarrick J. Wong nrec = arec + 1; 92*f327a007SDarrick J. Wong 93*f327a007SDarrick J. Wong /* 94*f327a007SDarrick J. Wong * Insert second record at start of internal log 95*f327a007SDarrick J. Wong * which then gets trimmed. 96*f327a007SDarrick J. Wong */ 97*f327a007SDarrick J. Wong nrec->ar_startblock = cpu_to_be32( 98*f327a007SDarrick J. Wong be32_to_cpu(arec->ar_startblock) + 99*f327a007SDarrick J. Wong be32_to_cpu(arec->ar_blockcount)); 100*f327a007SDarrick J. Wong arec = nrec; 101*f327a007SDarrick J. Wong be16_add_cpu(&block->bb_numrecs, 1); 102*f327a007SDarrick J. Wong } 103*f327a007SDarrick J. Wong /* 104*f327a007SDarrick J. Wong * Change record start to after the internal log 105*f327a007SDarrick J. Wong */ 106*f327a007SDarrick J. Wong be32_add_cpu(&arec->ar_startblock, mp->m_sb.sb_logblocks); 107*f327a007SDarrick J. Wong } 108*f327a007SDarrick J. Wong 109*f327a007SDarrick J. Wong /* 110*f327a007SDarrick J. Wong * Calculate the record block count and check for the case where 111*f327a007SDarrick J. Wong * the log might have consumed all available space in the AG. If 112*f327a007SDarrick J. Wong * so, reset the record count to 0 to avoid exposure of an invalid 113*f327a007SDarrick J. Wong * record start block. 114*f327a007SDarrick J. Wong */ 1158d90857cSDarrick J. Wong arec->ar_blockcount = cpu_to_be32(id->agsize - 1168d90857cSDarrick J. Wong be32_to_cpu(arec->ar_startblock)); 117*f327a007SDarrick J. Wong if (!arec->ar_blockcount) 118*f327a007SDarrick J. Wong block->bb_numrecs = 0; 1198d90857cSDarrick J. Wong } 1208d90857cSDarrick J. Wong 121b16817b6SDave Chinner /* 122b16817b6SDave Chinner * Alloc btree root block init functions 123b16817b6SDave Chinner */ 124b16817b6SDave Chinner static void 125b16817b6SDave Chinner xfs_bnoroot_init( 126b16817b6SDave Chinner struct xfs_mount *mp, 127b16817b6SDave Chinner struct xfs_buf *bp, 128b16817b6SDave Chinner struct aghdr_init_data *id) 129b16817b6SDave Chinner { 130f5b999c0SEric Sandeen xfs_btree_init_block(mp, bp, XFS_BTNUM_BNO, 0, 1, id->agno); 1318d90857cSDarrick J. Wong xfs_freesp_init_recs(mp, bp, id); 132b16817b6SDave Chinner } 133b16817b6SDave Chinner 134b16817b6SDave Chinner static void 135b16817b6SDave Chinner xfs_cntroot_init( 136b16817b6SDave Chinner struct xfs_mount *mp, 137b16817b6SDave Chinner struct xfs_buf *bp, 138b16817b6SDave Chinner struct aghdr_init_data *id) 139b16817b6SDave Chinner { 140f5b999c0SEric Sandeen xfs_btree_init_block(mp, bp, XFS_BTNUM_CNT, 0, 1, id->agno); 1418d90857cSDarrick J. Wong xfs_freesp_init_recs(mp, bp, id); 142b16817b6SDave Chinner } 143b16817b6SDave Chinner 144b16817b6SDave Chinner /* 145b16817b6SDave Chinner * Reverse map root block init 146b16817b6SDave Chinner */ 147b16817b6SDave Chinner static void 148b16817b6SDave Chinner xfs_rmaproot_init( 149b16817b6SDave Chinner struct xfs_mount *mp, 150b16817b6SDave Chinner struct xfs_buf *bp, 151b16817b6SDave Chinner struct aghdr_init_data *id) 152b16817b6SDave Chinner { 153b16817b6SDave Chinner struct xfs_btree_block *block = XFS_BUF_TO_BLOCK(bp); 154b16817b6SDave Chinner struct xfs_rmap_rec *rrec; 155b16817b6SDave Chinner 156f5b999c0SEric Sandeen xfs_btree_init_block(mp, bp, XFS_BTNUM_RMAP, 0, 4, id->agno); 157b16817b6SDave Chinner 158b16817b6SDave Chinner /* 159b16817b6SDave Chinner * mark the AG header regions as static metadata The BNO 160b16817b6SDave Chinner * btree block is the first block after the headers, so 161b16817b6SDave Chinner * it's location defines the size of region the static 162b16817b6SDave Chinner * metadata consumes. 163b16817b6SDave Chinner * 164b16817b6SDave Chinner * Note: unlike mkfs, we never have to account for log 165b16817b6SDave Chinner * space when growing the data regions 166b16817b6SDave Chinner */ 167b16817b6SDave Chinner rrec = XFS_RMAP_REC_ADDR(block, 1); 168b16817b6SDave Chinner rrec->rm_startblock = 0; 169b16817b6SDave Chinner rrec->rm_blockcount = cpu_to_be32(XFS_BNO_BLOCK(mp)); 170b16817b6SDave Chinner rrec->rm_owner = cpu_to_be64(XFS_RMAP_OWN_FS); 171b16817b6SDave Chinner rrec->rm_offset = 0; 172b16817b6SDave Chinner 173b16817b6SDave Chinner /* account freespace btree root blocks */ 174b16817b6SDave Chinner rrec = XFS_RMAP_REC_ADDR(block, 2); 175b16817b6SDave Chinner rrec->rm_startblock = cpu_to_be32(XFS_BNO_BLOCK(mp)); 176b16817b6SDave Chinner rrec->rm_blockcount = cpu_to_be32(2); 177b16817b6SDave Chinner rrec->rm_owner = cpu_to_be64(XFS_RMAP_OWN_AG); 178b16817b6SDave Chinner rrec->rm_offset = 0; 179b16817b6SDave Chinner 180b16817b6SDave Chinner /* account inode btree root blocks */ 181b16817b6SDave Chinner rrec = XFS_RMAP_REC_ADDR(block, 3); 182b16817b6SDave Chinner rrec->rm_startblock = cpu_to_be32(XFS_IBT_BLOCK(mp)); 183b16817b6SDave Chinner rrec->rm_blockcount = cpu_to_be32(XFS_RMAP_BLOCK(mp) - 184b16817b6SDave Chinner XFS_IBT_BLOCK(mp)); 185b16817b6SDave Chinner rrec->rm_owner = cpu_to_be64(XFS_RMAP_OWN_INOBT); 186b16817b6SDave Chinner rrec->rm_offset = 0; 187b16817b6SDave Chinner 188b16817b6SDave Chinner /* account for rmap btree root */ 189b16817b6SDave Chinner rrec = XFS_RMAP_REC_ADDR(block, 4); 190b16817b6SDave Chinner rrec->rm_startblock = cpu_to_be32(XFS_RMAP_BLOCK(mp)); 191b16817b6SDave Chinner rrec->rm_blockcount = cpu_to_be32(1); 192b16817b6SDave Chinner rrec->rm_owner = cpu_to_be64(XFS_RMAP_OWN_AG); 193b16817b6SDave Chinner rrec->rm_offset = 0; 194b16817b6SDave Chinner 195b16817b6SDave Chinner /* account for refc btree root */ 196b16817b6SDave Chinner if (xfs_sb_version_hasreflink(&mp->m_sb)) { 197b16817b6SDave Chinner rrec = XFS_RMAP_REC_ADDR(block, 5); 198b16817b6SDave Chinner rrec->rm_startblock = cpu_to_be32(xfs_refc_block(mp)); 199b16817b6SDave Chinner rrec->rm_blockcount = cpu_to_be32(1); 200b16817b6SDave Chinner rrec->rm_owner = cpu_to_be64(XFS_RMAP_OWN_REFC); 201b16817b6SDave Chinner rrec->rm_offset = 0; 202b16817b6SDave Chinner be16_add_cpu(&block->bb_numrecs, 1); 203b16817b6SDave Chinner } 204*f327a007SDarrick J. Wong 205*f327a007SDarrick J. Wong /* account for the log space */ 206*f327a007SDarrick J. Wong if (is_log_ag(mp, id)) { 207*f327a007SDarrick J. Wong rrec = XFS_RMAP_REC_ADDR(block, 208*f327a007SDarrick J. Wong be16_to_cpu(block->bb_numrecs) + 1); 209*f327a007SDarrick J. Wong rrec->rm_startblock = cpu_to_be32( 210*f327a007SDarrick J. Wong XFS_FSB_TO_AGBNO(mp, mp->m_sb.sb_logstart)); 211*f327a007SDarrick J. Wong rrec->rm_blockcount = cpu_to_be32(mp->m_sb.sb_logblocks); 212*f327a007SDarrick J. Wong rrec->rm_owner = cpu_to_be64(XFS_RMAP_OWN_LOG); 213*f327a007SDarrick J. Wong rrec->rm_offset = 0; 214*f327a007SDarrick J. Wong be16_add_cpu(&block->bb_numrecs, 1); 215*f327a007SDarrick J. Wong } 216b16817b6SDave Chinner } 217b16817b6SDave Chinner 218b16817b6SDave Chinner /* 219b16817b6SDave Chinner * Initialise new secondary superblocks with the pre-grow geometry, but mark 220b16817b6SDave Chinner * them as "in progress" so we know they haven't yet been activated. This will 221b16817b6SDave Chinner * get cleared when the update with the new geometry information is done after 222b16817b6SDave Chinner * changes to the primary are committed. This isn't strictly necessary, but we 223b16817b6SDave Chinner * get it for free with the delayed buffer write lists and it means we can tell 224b16817b6SDave Chinner * if a grow operation didn't complete properly after the fact. 225b16817b6SDave Chinner */ 226b16817b6SDave Chinner static void 227b16817b6SDave Chinner xfs_sbblock_init( 228b16817b6SDave Chinner struct xfs_mount *mp, 229b16817b6SDave Chinner struct xfs_buf *bp, 230b16817b6SDave Chinner struct aghdr_init_data *id) 231b16817b6SDave Chinner { 232b16817b6SDave Chinner struct xfs_dsb *dsb = XFS_BUF_TO_SBP(bp); 233b16817b6SDave Chinner 234b16817b6SDave Chinner xfs_sb_to_disk(dsb, &mp->m_sb); 235b16817b6SDave Chinner dsb->sb_inprogress = 1; 236b16817b6SDave Chinner } 237b16817b6SDave Chinner 238b16817b6SDave Chinner static void 239b16817b6SDave Chinner xfs_agfblock_init( 240b16817b6SDave Chinner struct xfs_mount *mp, 241b16817b6SDave Chinner struct xfs_buf *bp, 242b16817b6SDave Chinner struct aghdr_init_data *id) 243b16817b6SDave Chinner { 244b16817b6SDave Chinner struct xfs_agf *agf = XFS_BUF_TO_AGF(bp); 245b16817b6SDave Chinner xfs_extlen_t tmpsize; 246b16817b6SDave Chinner 247b16817b6SDave Chinner agf->agf_magicnum = cpu_to_be32(XFS_AGF_MAGIC); 248b16817b6SDave Chinner agf->agf_versionnum = cpu_to_be32(XFS_AGF_VERSION); 249b16817b6SDave Chinner agf->agf_seqno = cpu_to_be32(id->agno); 250b16817b6SDave Chinner agf->agf_length = cpu_to_be32(id->agsize); 251b16817b6SDave Chinner agf->agf_roots[XFS_BTNUM_BNOi] = cpu_to_be32(XFS_BNO_BLOCK(mp)); 252b16817b6SDave Chinner agf->agf_roots[XFS_BTNUM_CNTi] = cpu_to_be32(XFS_CNT_BLOCK(mp)); 253b16817b6SDave Chinner agf->agf_levels[XFS_BTNUM_BNOi] = cpu_to_be32(1); 254b16817b6SDave Chinner agf->agf_levels[XFS_BTNUM_CNTi] = cpu_to_be32(1); 255b16817b6SDave Chinner if (xfs_sb_version_hasrmapbt(&mp->m_sb)) { 256b16817b6SDave Chinner agf->agf_roots[XFS_BTNUM_RMAPi] = 257b16817b6SDave Chinner cpu_to_be32(XFS_RMAP_BLOCK(mp)); 258b16817b6SDave Chinner agf->agf_levels[XFS_BTNUM_RMAPi] = cpu_to_be32(1); 259b16817b6SDave Chinner agf->agf_rmap_blocks = cpu_to_be32(1); 260b16817b6SDave Chinner } 261b16817b6SDave Chinner 262b16817b6SDave Chinner agf->agf_flfirst = cpu_to_be32(1); 263b16817b6SDave Chinner agf->agf_fllast = 0; 264b16817b6SDave Chinner agf->agf_flcount = 0; 265b16817b6SDave Chinner tmpsize = id->agsize - mp->m_ag_prealloc_blocks; 266b16817b6SDave Chinner agf->agf_freeblks = cpu_to_be32(tmpsize); 267b16817b6SDave Chinner agf->agf_longest = cpu_to_be32(tmpsize); 268b16817b6SDave Chinner if (xfs_sb_version_hascrc(&mp->m_sb)) 269b16817b6SDave Chinner uuid_copy(&agf->agf_uuid, &mp->m_sb.sb_meta_uuid); 270b16817b6SDave Chinner if (xfs_sb_version_hasreflink(&mp->m_sb)) { 271b16817b6SDave Chinner agf->agf_refcount_root = cpu_to_be32( 272b16817b6SDave Chinner xfs_refc_block(mp)); 273b16817b6SDave Chinner agf->agf_refcount_level = cpu_to_be32(1); 274b16817b6SDave Chinner agf->agf_refcount_blocks = cpu_to_be32(1); 275b16817b6SDave Chinner } 276*f327a007SDarrick J. Wong 277*f327a007SDarrick J. Wong if (is_log_ag(mp, id)) { 278*f327a007SDarrick J. Wong int64_t logblocks = mp->m_sb.sb_logblocks; 279*f327a007SDarrick J. Wong 280*f327a007SDarrick J. Wong be32_add_cpu(&agf->agf_freeblks, -logblocks); 281*f327a007SDarrick J. Wong agf->agf_longest = cpu_to_be32(id->agsize - 282*f327a007SDarrick J. Wong XFS_FSB_TO_AGBNO(mp, mp->m_sb.sb_logstart) - logblocks); 283*f327a007SDarrick J. Wong } 284b16817b6SDave Chinner } 285b16817b6SDave Chinner 286b16817b6SDave Chinner static void 287b16817b6SDave Chinner xfs_agflblock_init( 288b16817b6SDave Chinner struct xfs_mount *mp, 289b16817b6SDave Chinner struct xfs_buf *bp, 290b16817b6SDave Chinner struct aghdr_init_data *id) 291b16817b6SDave Chinner { 292b16817b6SDave Chinner struct xfs_agfl *agfl = XFS_BUF_TO_AGFL(bp); 293b16817b6SDave Chinner __be32 *agfl_bno; 294b16817b6SDave Chinner int bucket; 295b16817b6SDave Chinner 296b16817b6SDave Chinner if (xfs_sb_version_hascrc(&mp->m_sb)) { 297b16817b6SDave Chinner agfl->agfl_magicnum = cpu_to_be32(XFS_AGFL_MAGIC); 298b16817b6SDave Chinner agfl->agfl_seqno = cpu_to_be32(id->agno); 299b16817b6SDave Chinner uuid_copy(&agfl->agfl_uuid, &mp->m_sb.sb_meta_uuid); 300b16817b6SDave Chinner } 301b16817b6SDave Chinner 302b16817b6SDave Chinner agfl_bno = XFS_BUF_TO_AGFL_BNO(mp, bp); 303b16817b6SDave Chinner for (bucket = 0; bucket < xfs_agfl_size(mp); bucket++) 304b16817b6SDave Chinner agfl_bno[bucket] = cpu_to_be32(NULLAGBLOCK); 305b16817b6SDave Chinner } 306b16817b6SDave Chinner 307b16817b6SDave Chinner static void 308b16817b6SDave Chinner xfs_agiblock_init( 309b16817b6SDave Chinner struct xfs_mount *mp, 310b16817b6SDave Chinner struct xfs_buf *bp, 311b16817b6SDave Chinner struct aghdr_init_data *id) 312b16817b6SDave Chinner { 313b16817b6SDave Chinner struct xfs_agi *agi = XFS_BUF_TO_AGI(bp); 314b16817b6SDave Chinner int bucket; 315b16817b6SDave Chinner 316b16817b6SDave Chinner agi->agi_magicnum = cpu_to_be32(XFS_AGI_MAGIC); 317b16817b6SDave Chinner agi->agi_versionnum = cpu_to_be32(XFS_AGI_VERSION); 318b16817b6SDave Chinner agi->agi_seqno = cpu_to_be32(id->agno); 319b16817b6SDave Chinner agi->agi_length = cpu_to_be32(id->agsize); 320b16817b6SDave Chinner agi->agi_count = 0; 321b16817b6SDave Chinner agi->agi_root = cpu_to_be32(XFS_IBT_BLOCK(mp)); 322b16817b6SDave Chinner agi->agi_level = cpu_to_be32(1); 323b16817b6SDave Chinner agi->agi_freecount = 0; 324b16817b6SDave Chinner agi->agi_newino = cpu_to_be32(NULLAGINO); 325b16817b6SDave Chinner agi->agi_dirino = cpu_to_be32(NULLAGINO); 326b16817b6SDave Chinner if (xfs_sb_version_hascrc(&mp->m_sb)) 327b16817b6SDave Chinner uuid_copy(&agi->agi_uuid, &mp->m_sb.sb_meta_uuid); 328b16817b6SDave Chinner if (xfs_sb_version_hasfinobt(&mp->m_sb)) { 329b16817b6SDave Chinner agi->agi_free_root = cpu_to_be32(XFS_FIBT_BLOCK(mp)); 330b16817b6SDave Chinner agi->agi_free_level = cpu_to_be32(1); 331b16817b6SDave Chinner } 332b16817b6SDave Chinner for (bucket = 0; bucket < XFS_AGI_UNLINKED_BUCKETS; bucket++) 333b16817b6SDave Chinner agi->agi_unlinked[bucket] = cpu_to_be32(NULLAGINO); 334b16817b6SDave Chinner } 335b16817b6SDave Chinner 336b16817b6SDave Chinner typedef void (*aghdr_init_work_f)(struct xfs_mount *mp, struct xfs_buf *bp, 337b16817b6SDave Chinner struct aghdr_init_data *id); 338b16817b6SDave Chinner static int 339b16817b6SDave Chinner xfs_ag_init_hdr( 340b16817b6SDave Chinner struct xfs_mount *mp, 341b16817b6SDave Chinner struct aghdr_init_data *id, 342b16817b6SDave Chinner aghdr_init_work_f work, 343b16817b6SDave Chinner const struct xfs_buf_ops *ops) 344b16817b6SDave Chinner 345b16817b6SDave Chinner { 346b16817b6SDave Chinner struct xfs_buf *bp; 347b16817b6SDave Chinner 348b16817b6SDave Chinner bp = xfs_get_aghdr_buf(mp, id->daddr, id->numblks, 0, ops); 349b16817b6SDave Chinner if (!bp) 350b16817b6SDave Chinner return -ENOMEM; 351b16817b6SDave Chinner 352b16817b6SDave Chinner (*work)(mp, bp, id); 353b16817b6SDave Chinner 354b16817b6SDave Chinner xfs_buf_delwri_queue(bp, &id->buffer_list); 355b16817b6SDave Chinner xfs_buf_relse(bp); 356b16817b6SDave Chinner return 0; 357b16817b6SDave Chinner } 358b16817b6SDave Chinner 359b16817b6SDave Chinner struct xfs_aghdr_grow_data { 360b16817b6SDave Chinner xfs_daddr_t daddr; 361b16817b6SDave Chinner size_t numblks; 362b16817b6SDave Chinner const struct xfs_buf_ops *ops; 363b16817b6SDave Chinner aghdr_init_work_f work; 364b16817b6SDave Chinner xfs_btnum_t type; 365b16817b6SDave Chinner bool need_init; 366b16817b6SDave Chinner }; 367b16817b6SDave Chinner 368b16817b6SDave Chinner /* 369b16817b6SDave Chinner * Prepare new AG headers to be written to disk. We use uncached buffers here, 370b16817b6SDave Chinner * as it is assumed these new AG headers are currently beyond the currently 371b16817b6SDave Chinner * valid filesystem address space. Using cached buffers would trip over EOFS 372b16817b6SDave Chinner * corruption detection alogrithms in the buffer cache lookup routines. 373b16817b6SDave Chinner * 374b16817b6SDave Chinner * This is a non-transactional function, but the prepared buffers are added to a 375b16817b6SDave Chinner * delayed write buffer list supplied by the caller so they can submit them to 376b16817b6SDave Chinner * disk and wait on them as required. 377b16817b6SDave Chinner */ 378b16817b6SDave Chinner int 379b16817b6SDave Chinner xfs_ag_init_headers( 380b16817b6SDave Chinner struct xfs_mount *mp, 381b16817b6SDave Chinner struct aghdr_init_data *id) 382b16817b6SDave Chinner 383b16817b6SDave Chinner { 384b16817b6SDave Chinner struct xfs_aghdr_grow_data aghdr_data[] = { 385b16817b6SDave Chinner { /* SB */ 386b16817b6SDave Chinner .daddr = XFS_AG_DADDR(mp, id->agno, XFS_SB_DADDR), 387b16817b6SDave Chinner .numblks = XFS_FSS_TO_BB(mp, 1), 388b16817b6SDave Chinner .ops = &xfs_sb_buf_ops, 389b16817b6SDave Chinner .work = &xfs_sbblock_init, 390b16817b6SDave Chinner .need_init = true 391b16817b6SDave Chinner }, 392b16817b6SDave Chinner { /* AGF */ 393b16817b6SDave Chinner .daddr = XFS_AG_DADDR(mp, id->agno, XFS_AGF_DADDR(mp)), 394b16817b6SDave Chinner .numblks = XFS_FSS_TO_BB(mp, 1), 395b16817b6SDave Chinner .ops = &xfs_agf_buf_ops, 396b16817b6SDave Chinner .work = &xfs_agfblock_init, 397b16817b6SDave Chinner .need_init = true 398b16817b6SDave Chinner }, 399b16817b6SDave Chinner { /* AGFL */ 400b16817b6SDave Chinner .daddr = XFS_AG_DADDR(mp, id->agno, XFS_AGFL_DADDR(mp)), 401b16817b6SDave Chinner .numblks = XFS_FSS_TO_BB(mp, 1), 402b16817b6SDave Chinner .ops = &xfs_agfl_buf_ops, 403b16817b6SDave Chinner .work = &xfs_agflblock_init, 404b16817b6SDave Chinner .need_init = true 405b16817b6SDave Chinner }, 406b16817b6SDave Chinner { /* AGI */ 407b16817b6SDave Chinner .daddr = XFS_AG_DADDR(mp, id->agno, XFS_AGI_DADDR(mp)), 408b16817b6SDave Chinner .numblks = XFS_FSS_TO_BB(mp, 1), 409b16817b6SDave Chinner .ops = &xfs_agi_buf_ops, 410b16817b6SDave Chinner .work = &xfs_agiblock_init, 411b16817b6SDave Chinner .need_init = true 412b16817b6SDave Chinner }, 413b16817b6SDave Chinner { /* BNO root block */ 414b16817b6SDave Chinner .daddr = XFS_AGB_TO_DADDR(mp, id->agno, XFS_BNO_BLOCK(mp)), 415b16817b6SDave Chinner .numblks = BTOBB(mp->m_sb.sb_blocksize), 41627df4f50SBrian Foster .ops = &xfs_bnobt_buf_ops, 417b16817b6SDave Chinner .work = &xfs_bnoroot_init, 418b16817b6SDave Chinner .need_init = true 419b16817b6SDave Chinner }, 420b16817b6SDave Chinner { /* CNT root block */ 421b16817b6SDave Chinner .daddr = XFS_AGB_TO_DADDR(mp, id->agno, XFS_CNT_BLOCK(mp)), 422b16817b6SDave Chinner .numblks = BTOBB(mp->m_sb.sb_blocksize), 42327df4f50SBrian Foster .ops = &xfs_cntbt_buf_ops, 424b16817b6SDave Chinner .work = &xfs_cntroot_init, 425b16817b6SDave Chinner .need_init = true 426b16817b6SDave Chinner }, 427b16817b6SDave Chinner { /* INO root block */ 428b16817b6SDave Chinner .daddr = XFS_AGB_TO_DADDR(mp, id->agno, XFS_IBT_BLOCK(mp)), 429b16817b6SDave Chinner .numblks = BTOBB(mp->m_sb.sb_blocksize), 430b16817b6SDave Chinner .ops = &xfs_inobt_buf_ops, 431b16817b6SDave Chinner .work = &xfs_btroot_init, 432b16817b6SDave Chinner .type = XFS_BTNUM_INO, 433b16817b6SDave Chinner .need_init = true 434b16817b6SDave Chinner }, 435b16817b6SDave Chinner { /* FINO root block */ 436b16817b6SDave Chinner .daddr = XFS_AGB_TO_DADDR(mp, id->agno, XFS_FIBT_BLOCK(mp)), 437b16817b6SDave Chinner .numblks = BTOBB(mp->m_sb.sb_blocksize), 43801e68f40SBrian Foster .ops = &xfs_finobt_buf_ops, 439b16817b6SDave Chinner .work = &xfs_btroot_init, 440b16817b6SDave Chinner .type = XFS_BTNUM_FINO, 441b16817b6SDave Chinner .need_init = xfs_sb_version_hasfinobt(&mp->m_sb) 442b16817b6SDave Chinner }, 443b16817b6SDave Chinner { /* RMAP root block */ 444b16817b6SDave Chinner .daddr = XFS_AGB_TO_DADDR(mp, id->agno, XFS_RMAP_BLOCK(mp)), 445b16817b6SDave Chinner .numblks = BTOBB(mp->m_sb.sb_blocksize), 446b16817b6SDave Chinner .ops = &xfs_rmapbt_buf_ops, 447b16817b6SDave Chinner .work = &xfs_rmaproot_init, 448b16817b6SDave Chinner .need_init = xfs_sb_version_hasrmapbt(&mp->m_sb) 449b16817b6SDave Chinner }, 450b16817b6SDave Chinner { /* REFC root block */ 451b16817b6SDave Chinner .daddr = XFS_AGB_TO_DADDR(mp, id->agno, xfs_refc_block(mp)), 452b16817b6SDave Chinner .numblks = BTOBB(mp->m_sb.sb_blocksize), 453b16817b6SDave Chinner .ops = &xfs_refcountbt_buf_ops, 454b16817b6SDave Chinner .work = &xfs_btroot_init, 455b16817b6SDave Chinner .type = XFS_BTNUM_REFC, 456b16817b6SDave Chinner .need_init = xfs_sb_version_hasreflink(&mp->m_sb) 457b16817b6SDave Chinner }, 458b16817b6SDave Chinner { /* NULL terminating block */ 459b16817b6SDave Chinner .daddr = XFS_BUF_DADDR_NULL, 460b16817b6SDave Chinner } 461b16817b6SDave Chinner }; 462b16817b6SDave Chinner struct xfs_aghdr_grow_data *dp; 463b16817b6SDave Chinner int error = 0; 464b16817b6SDave Chinner 465b16817b6SDave Chinner /* Account for AG free space in new AG */ 466b16817b6SDave Chinner id->nfree += id->agsize - mp->m_ag_prealloc_blocks; 467b16817b6SDave Chinner for (dp = &aghdr_data[0]; dp->daddr != XFS_BUF_DADDR_NULL; dp++) { 468b16817b6SDave Chinner if (!dp->need_init) 469b16817b6SDave Chinner continue; 470b16817b6SDave Chinner 471b16817b6SDave Chinner id->daddr = dp->daddr; 472b16817b6SDave Chinner id->numblks = dp->numblks; 473b16817b6SDave Chinner id->type = dp->type; 474b16817b6SDave Chinner error = xfs_ag_init_hdr(mp, id, dp->work, dp->ops); 475b16817b6SDave Chinner if (error) 476b16817b6SDave Chinner break; 477b16817b6SDave Chinner } 478b16817b6SDave Chinner return error; 479b16817b6SDave Chinner } 48049dd56f2SDave Chinner 48149dd56f2SDave Chinner /* 48249dd56f2SDave Chinner * Extent the AG indicated by the @id by the length passed in 48349dd56f2SDave Chinner */ 48449dd56f2SDave Chinner int 48549dd56f2SDave Chinner xfs_ag_extend_space( 48649dd56f2SDave Chinner struct xfs_mount *mp, 48749dd56f2SDave Chinner struct xfs_trans *tp, 48849dd56f2SDave Chinner struct aghdr_init_data *id, 48949dd56f2SDave Chinner xfs_extlen_t len) 49049dd56f2SDave Chinner { 49149dd56f2SDave Chinner struct xfs_buf *bp; 49249dd56f2SDave Chinner struct xfs_agi *agi; 49349dd56f2SDave Chinner struct xfs_agf *agf; 49449dd56f2SDave Chinner int error; 49549dd56f2SDave Chinner 49649dd56f2SDave Chinner /* 49749dd56f2SDave Chinner * Change the agi length. 49849dd56f2SDave Chinner */ 49949dd56f2SDave Chinner error = xfs_ialloc_read_agi(mp, tp, id->agno, &bp); 50049dd56f2SDave Chinner if (error) 50149dd56f2SDave Chinner return error; 50249dd56f2SDave Chinner 50349dd56f2SDave Chinner agi = XFS_BUF_TO_AGI(bp); 50449dd56f2SDave Chinner be32_add_cpu(&agi->agi_length, len); 50549dd56f2SDave Chinner ASSERT(id->agno == mp->m_sb.sb_agcount - 1 || 50649dd56f2SDave Chinner be32_to_cpu(agi->agi_length) == mp->m_sb.sb_agblocks); 50749dd56f2SDave Chinner xfs_ialloc_log_agi(tp, bp, XFS_AGI_LENGTH); 50849dd56f2SDave Chinner 50949dd56f2SDave Chinner /* 51049dd56f2SDave Chinner * Change agf length. 51149dd56f2SDave Chinner */ 51249dd56f2SDave Chinner error = xfs_alloc_read_agf(mp, tp, id->agno, 0, &bp); 51349dd56f2SDave Chinner if (error) 51449dd56f2SDave Chinner return error; 51549dd56f2SDave Chinner 51649dd56f2SDave Chinner agf = XFS_BUF_TO_AGF(bp); 51749dd56f2SDave Chinner be32_add_cpu(&agf->agf_length, len); 51849dd56f2SDave Chinner ASSERT(agf->agf_length == agi->agi_length); 51949dd56f2SDave Chinner xfs_alloc_log_agf(tp, bp, XFS_AGF_LENGTH); 52049dd56f2SDave Chinner 52149dd56f2SDave Chinner /* 52249dd56f2SDave Chinner * Free the new space. 52349dd56f2SDave Chinner * 5247280fedaSDarrick J. Wong * XFS_RMAP_OINFO_SKIP_UPDATE is used here to tell the rmap btree that 52549dd56f2SDave Chinner * this doesn't actually exist in the rmap btree. 52649dd56f2SDave Chinner */ 52749dd56f2SDave Chinner error = xfs_rmap_free(tp, bp, id->agno, 52849dd56f2SDave Chinner be32_to_cpu(agf->agf_length) - len, 5297280fedaSDarrick J. Wong len, &XFS_RMAP_OINFO_SKIP_UPDATE); 53049dd56f2SDave Chinner if (error) 53149dd56f2SDave Chinner return error; 53249dd56f2SDave Chinner 53349dd56f2SDave Chinner return xfs_free_extent(tp, XFS_AGB_TO_FSB(mp, id->agno, 53449dd56f2SDave Chinner be32_to_cpu(agf->agf_length) - len), 5357280fedaSDarrick J. Wong len, &XFS_RMAP_OINFO_SKIP_UPDATE, 5367280fedaSDarrick J. Wong XFS_AG_RESV_NONE); 53749dd56f2SDave Chinner } 5387cd5006bSDarrick J. Wong 5397cd5006bSDarrick J. Wong /* Retrieve AG geometry. */ 5407cd5006bSDarrick J. Wong int 5417cd5006bSDarrick J. Wong xfs_ag_get_geometry( 5427cd5006bSDarrick J. Wong struct xfs_mount *mp, 5437cd5006bSDarrick J. Wong xfs_agnumber_t agno, 5447cd5006bSDarrick J. Wong struct xfs_ag_geometry *ageo) 5457cd5006bSDarrick J. Wong { 5467cd5006bSDarrick J. Wong struct xfs_buf *agi_bp; 5477cd5006bSDarrick J. Wong struct xfs_buf *agf_bp; 5487cd5006bSDarrick J. Wong struct xfs_agi *agi; 5497cd5006bSDarrick J. Wong struct xfs_agf *agf; 5507cd5006bSDarrick J. Wong struct xfs_perag *pag; 5517cd5006bSDarrick J. Wong unsigned int freeblks; 5527cd5006bSDarrick J. Wong int error; 5537cd5006bSDarrick J. Wong 5547cd5006bSDarrick J. Wong if (agno >= mp->m_sb.sb_agcount) 5557cd5006bSDarrick J. Wong return -EINVAL; 5567cd5006bSDarrick J. Wong 5577cd5006bSDarrick J. Wong /* Lock the AG headers. */ 5587cd5006bSDarrick J. Wong error = xfs_ialloc_read_agi(mp, NULL, agno, &agi_bp); 5597cd5006bSDarrick J. Wong if (error) 5607cd5006bSDarrick J. Wong return error; 5617cd5006bSDarrick J. Wong error = xfs_alloc_read_agf(mp, NULL, agno, 0, &agf_bp); 5627cd5006bSDarrick J. Wong if (error) 5637cd5006bSDarrick J. Wong goto out_agi; 5647cd5006bSDarrick J. Wong pag = xfs_perag_get(mp, agno); 5657cd5006bSDarrick J. Wong 5667cd5006bSDarrick J. Wong /* Fill out form. */ 5677cd5006bSDarrick J. Wong memset(ageo, 0, sizeof(*ageo)); 5687cd5006bSDarrick J. Wong ageo->ag_number = agno; 5697cd5006bSDarrick J. Wong 5707cd5006bSDarrick J. Wong agi = XFS_BUF_TO_AGI(agi_bp); 5717cd5006bSDarrick J. Wong ageo->ag_icount = be32_to_cpu(agi->agi_count); 5727cd5006bSDarrick J. Wong ageo->ag_ifree = be32_to_cpu(agi->agi_freecount); 5737cd5006bSDarrick J. Wong 5747cd5006bSDarrick J. Wong agf = XFS_BUF_TO_AGF(agf_bp); 5757cd5006bSDarrick J. Wong ageo->ag_length = be32_to_cpu(agf->agf_length); 5767cd5006bSDarrick J. Wong freeblks = pag->pagf_freeblks + 5777cd5006bSDarrick J. Wong pag->pagf_flcount + 5787cd5006bSDarrick J. Wong pag->pagf_btreeblks - 5797cd5006bSDarrick J. Wong xfs_ag_resv_needed(pag, XFS_AG_RESV_NONE); 5807cd5006bSDarrick J. Wong ageo->ag_freeblks = freeblks; 5811302c6a2SDarrick J. Wong xfs_ag_geom_health(pag, ageo); 5827cd5006bSDarrick J. Wong 5837cd5006bSDarrick J. Wong /* Release resources. */ 5847cd5006bSDarrick J. Wong xfs_perag_put(pag); 5857cd5006bSDarrick J. Wong xfs_buf_relse(agf_bp); 5867cd5006bSDarrick J. Wong out_agi: 5877cd5006bSDarrick J. Wong xfs_buf_relse(agi_bp); 5887cd5006bSDarrick J. Wong return error; 5897cd5006bSDarrick J. Wong } 590