10b61f8a4SDave Chinner // SPDX-License-Identifier: GPL-2.0 23ebe7d2dSDave Chinner /* 33ebe7d2dSDave Chinner * Copyright (c) 2008-2010, 2013 Dave Chinner 43ebe7d2dSDave Chinner * All Rights Reserved. 53ebe7d2dSDave Chinner */ 63ebe7d2dSDave Chinner #include "xfs.h" 73ebe7d2dSDave Chinner #include "xfs_fs.h" 870a9883cSDave Chinner #include "xfs_shared.h" 93ec6efa7SDarrick J. Wong #include "xfs_format.h" 10239880efSDave Chinner #include "xfs_log_format.h" 113ec6efa7SDarrick J. Wong #include "xfs_trans_resv.h" 123ec6efa7SDarrick J. Wong #include "xfs_mount.h" 133ec6efa7SDarrick J. Wong #include "xfs_inode.h" 14239880efSDave Chinner #include "xfs_trans.h" 153ebe7d2dSDave Chinner #include "xfs_trans_priv.h" 163ebe7d2dSDave Chinner #include "xfs_icreate_item.h" 171234351cSChristoph Hellwig #include "xfs_log.h" 1886ffa471SDarrick J. Wong #include "xfs_log_priv.h" 1986ffa471SDarrick J. Wong #include "xfs_log_recover.h" 203ec6efa7SDarrick J. Wong #include "xfs_ialloc.h" 213ec6efa7SDarrick J. Wong #include "xfs_trace.h" 223ebe7d2dSDave Chinner 233ebe7d2dSDave Chinner kmem_zone_t *xfs_icreate_zone; /* inode create item zone */ 243ebe7d2dSDave Chinner 253ebe7d2dSDave Chinner static inline struct xfs_icreate_item *ICR_ITEM(struct xfs_log_item *lip) 263ebe7d2dSDave Chinner { 273ebe7d2dSDave Chinner return container_of(lip, struct xfs_icreate_item, ic_item); 283ebe7d2dSDave Chinner } 293ebe7d2dSDave Chinner 303ebe7d2dSDave Chinner /* 313ebe7d2dSDave Chinner * This returns the number of iovecs needed to log the given inode item. 323ebe7d2dSDave Chinner * 333ebe7d2dSDave Chinner * We only need one iovec for the icreate log structure. 343ebe7d2dSDave Chinner */ 35166d1368SDave Chinner STATIC void 363ebe7d2dSDave Chinner xfs_icreate_item_size( 37166d1368SDave Chinner struct xfs_log_item *lip, 38166d1368SDave Chinner int *nvecs, 39166d1368SDave Chinner int *nbytes) 403ebe7d2dSDave Chinner { 41166d1368SDave Chinner *nvecs += 1; 42166d1368SDave Chinner *nbytes += sizeof(struct xfs_icreate_log); 433ebe7d2dSDave Chinner } 443ebe7d2dSDave Chinner 453ebe7d2dSDave Chinner /* 463ebe7d2dSDave Chinner * This is called to fill in the vector of log iovecs for the 473ebe7d2dSDave Chinner * given inode create log item. 483ebe7d2dSDave Chinner */ 493ebe7d2dSDave Chinner STATIC void 503ebe7d2dSDave Chinner xfs_icreate_item_format( 513ebe7d2dSDave Chinner struct xfs_log_item *lip, 52bde7cff6SChristoph Hellwig struct xfs_log_vec *lv) 533ebe7d2dSDave Chinner { 543ebe7d2dSDave Chinner struct xfs_icreate_item *icp = ICR_ITEM(lip); 55bde7cff6SChristoph Hellwig struct xfs_log_iovec *vecp = NULL; 563ebe7d2dSDave Chinner 57bde7cff6SChristoph Hellwig xlog_copy_iovec(lv, &vecp, XLOG_REG_TYPE_ICREATE, 581234351cSChristoph Hellwig &icp->ic_format, 591234351cSChristoph Hellwig sizeof(struct xfs_icreate_log)); 603ebe7d2dSDave Chinner } 613ebe7d2dSDave Chinner 623ebe7d2dSDave Chinner STATIC void 63ddf92053SChristoph Hellwig xfs_icreate_item_release( 643ebe7d2dSDave Chinner struct xfs_log_item *lip) 653ebe7d2dSDave Chinner { 66377bcd5fSCarlos Maiolino kmem_cache_free(xfs_icreate_zone, ICR_ITEM(lip)); 673ebe7d2dSDave Chinner } 683ebe7d2dSDave Chinner 69bb6e0ebeSBhumika Goyal static const struct xfs_item_ops xfs_icreate_item_ops = { 709ce632a2SChristoph Hellwig .flags = XFS_ITEM_RELEASE_WHEN_COMMITTED, 713ebe7d2dSDave Chinner .iop_size = xfs_icreate_item_size, 723ebe7d2dSDave Chinner .iop_format = xfs_icreate_item_format, 73ddf92053SChristoph Hellwig .iop_release = xfs_icreate_item_release, 743ebe7d2dSDave Chinner }; 753ebe7d2dSDave Chinner 763ebe7d2dSDave Chinner 773ebe7d2dSDave Chinner /* 783ebe7d2dSDave Chinner * Initialize the inode log item for a newly allocated (in-core) inode. 793ebe7d2dSDave Chinner * 803ebe7d2dSDave Chinner * Inode extents can only reside within an AG. Hence specify the starting 813ebe7d2dSDave Chinner * block for the inode chunk by offset within an AG as well as the 823ebe7d2dSDave Chinner * length of the allocated extent. 833ebe7d2dSDave Chinner * 843ebe7d2dSDave Chinner * This joins the item to the transaction and marks it dirty so 853ebe7d2dSDave Chinner * that we don't need a separate call to do this, nor does the 863ebe7d2dSDave Chinner * caller need to know anything about the icreate item. 873ebe7d2dSDave Chinner */ 883ebe7d2dSDave Chinner void 893ebe7d2dSDave Chinner xfs_icreate_log( 903ebe7d2dSDave Chinner struct xfs_trans *tp, 913ebe7d2dSDave Chinner xfs_agnumber_t agno, 923ebe7d2dSDave Chinner xfs_agblock_t agbno, 933ebe7d2dSDave Chinner unsigned int count, 943ebe7d2dSDave Chinner unsigned int inode_size, 953ebe7d2dSDave Chinner xfs_agblock_t length, 963ebe7d2dSDave Chinner unsigned int generation) 973ebe7d2dSDave Chinner { 983ebe7d2dSDave Chinner struct xfs_icreate_item *icp; 993ebe7d2dSDave Chinner 10032a2b11fSCarlos Maiolino icp = kmem_cache_zalloc(xfs_icreate_zone, GFP_KERNEL | __GFP_NOFAIL); 1013ebe7d2dSDave Chinner 1023ebe7d2dSDave Chinner xfs_log_item_init(tp->t_mountp, &icp->ic_item, XFS_LI_ICREATE, 1033ebe7d2dSDave Chinner &xfs_icreate_item_ops); 1043ebe7d2dSDave Chinner 1053ebe7d2dSDave Chinner icp->ic_format.icl_type = XFS_LI_ICREATE; 1063ebe7d2dSDave Chinner icp->ic_format.icl_size = 1; /* single vector */ 1073ebe7d2dSDave Chinner icp->ic_format.icl_ag = cpu_to_be32(agno); 1083ebe7d2dSDave Chinner icp->ic_format.icl_agbno = cpu_to_be32(agbno); 1093ebe7d2dSDave Chinner icp->ic_format.icl_count = cpu_to_be32(count); 1103ebe7d2dSDave Chinner icp->ic_format.icl_isize = cpu_to_be32(inode_size); 1113ebe7d2dSDave Chinner icp->ic_format.icl_length = cpu_to_be32(length); 1123ebe7d2dSDave Chinner icp->ic_format.icl_gen = cpu_to_be32(generation); 1133ebe7d2dSDave Chinner 1143ebe7d2dSDave Chinner xfs_trans_add_item(tp, &icp->ic_item); 1153ebe7d2dSDave Chinner tp->t_flags |= XFS_TRANS_DIRTY; 116e6631f85SDave Chinner set_bit(XFS_LI_DIRTY, &icp->ic_item.li_flags); 1173ebe7d2dSDave Chinner } 11886ffa471SDarrick J. Wong 11986ffa471SDarrick J. Wong static enum xlog_recover_reorder 12086ffa471SDarrick J. Wong xlog_recover_icreate_reorder( 12186ffa471SDarrick J. Wong struct xlog_recover_item *item) 12286ffa471SDarrick J. Wong { 12386ffa471SDarrick J. Wong /* 12486ffa471SDarrick J. Wong * Inode allocation buffers must be replayed before subsequent inode 12586ffa471SDarrick J. Wong * items try to modify those buffers. ICREATE items are the logical 12686ffa471SDarrick J. Wong * equivalent of logging a newly initialized inode buffer, so recover 12786ffa471SDarrick J. Wong * these at the same time that we recover logged buffers. 12886ffa471SDarrick J. Wong */ 12986ffa471SDarrick J. Wong return XLOG_REORDER_BUFFER_LIST; 13086ffa471SDarrick J. Wong } 13186ffa471SDarrick J. Wong 1323ec6efa7SDarrick J. Wong /* 1333ec6efa7SDarrick J. Wong * This routine is called when an inode create format structure is found in a 1343ec6efa7SDarrick J. Wong * committed transaction in the log. It's purpose is to initialise the inodes 1353ec6efa7SDarrick J. Wong * being allocated on disk. This requires us to get inode cluster buffers that 1363ec6efa7SDarrick J. Wong * match the range to be initialised, stamped with inode templates and written 1373ec6efa7SDarrick J. Wong * by delayed write so that subsequent modifications will hit the cached buffer 1383ec6efa7SDarrick J. Wong * and only need writing out at the end of recovery. 1393ec6efa7SDarrick J. Wong */ 1403ec6efa7SDarrick J. Wong STATIC int 1413ec6efa7SDarrick J. Wong xlog_recover_icreate_commit_pass2( 1423ec6efa7SDarrick J. Wong struct xlog *log, 1433ec6efa7SDarrick J. Wong struct list_head *buffer_list, 1443ec6efa7SDarrick J. Wong struct xlog_recover_item *item, 1453ec6efa7SDarrick J. Wong xfs_lsn_t lsn) 1463ec6efa7SDarrick J. Wong { 1473ec6efa7SDarrick J. Wong struct xfs_mount *mp = log->l_mp; 1483ec6efa7SDarrick J. Wong struct xfs_icreate_log *icl; 1493ec6efa7SDarrick J. Wong struct xfs_ino_geometry *igeo = M_IGEO(mp); 1503ec6efa7SDarrick J. Wong xfs_agnumber_t agno; 1513ec6efa7SDarrick J. Wong xfs_agblock_t agbno; 1523ec6efa7SDarrick J. Wong unsigned int count; 1533ec6efa7SDarrick J. Wong unsigned int isize; 1543ec6efa7SDarrick J. Wong xfs_agblock_t length; 1553ec6efa7SDarrick J. Wong int bb_per_cluster; 1563ec6efa7SDarrick J. Wong int cancel_count; 1573ec6efa7SDarrick J. Wong int nbufs; 1583ec6efa7SDarrick J. Wong int i; 1593ec6efa7SDarrick J. Wong 1603ec6efa7SDarrick J. Wong icl = (struct xfs_icreate_log *)item->ri_buf[0].i_addr; 1613ec6efa7SDarrick J. Wong if (icl->icl_type != XFS_LI_ICREATE) { 1623ec6efa7SDarrick J. Wong xfs_warn(log->l_mp, "xlog_recover_do_icreate_trans: bad type"); 1633ec6efa7SDarrick J. Wong return -EINVAL; 1643ec6efa7SDarrick J. Wong } 1653ec6efa7SDarrick J. Wong 1663ec6efa7SDarrick J. Wong if (icl->icl_size != 1) { 1673ec6efa7SDarrick J. Wong xfs_warn(log->l_mp, "xlog_recover_do_icreate_trans: bad icl size"); 1683ec6efa7SDarrick J. Wong return -EINVAL; 1693ec6efa7SDarrick J. Wong } 1703ec6efa7SDarrick J. Wong 1713ec6efa7SDarrick J. Wong agno = be32_to_cpu(icl->icl_ag); 1723ec6efa7SDarrick J. Wong if (agno >= mp->m_sb.sb_agcount) { 1733ec6efa7SDarrick J. Wong xfs_warn(log->l_mp, "xlog_recover_do_icreate_trans: bad agno"); 1743ec6efa7SDarrick J. Wong return -EINVAL; 1753ec6efa7SDarrick J. Wong } 1763ec6efa7SDarrick J. Wong agbno = be32_to_cpu(icl->icl_agbno); 1773ec6efa7SDarrick J. Wong if (!agbno || agbno == NULLAGBLOCK || agbno >= mp->m_sb.sb_agblocks) { 1783ec6efa7SDarrick J. Wong xfs_warn(log->l_mp, "xlog_recover_do_icreate_trans: bad agbno"); 1793ec6efa7SDarrick J. Wong return -EINVAL; 1803ec6efa7SDarrick J. Wong } 1813ec6efa7SDarrick J. Wong isize = be32_to_cpu(icl->icl_isize); 1823ec6efa7SDarrick J. Wong if (isize != mp->m_sb.sb_inodesize) { 1833ec6efa7SDarrick J. Wong xfs_warn(log->l_mp, "xlog_recover_do_icreate_trans: bad isize"); 1843ec6efa7SDarrick J. Wong return -EINVAL; 1853ec6efa7SDarrick J. Wong } 1863ec6efa7SDarrick J. Wong count = be32_to_cpu(icl->icl_count); 1873ec6efa7SDarrick J. Wong if (!count) { 1883ec6efa7SDarrick J. Wong xfs_warn(log->l_mp, "xlog_recover_do_icreate_trans: bad count"); 1893ec6efa7SDarrick J. Wong return -EINVAL; 1903ec6efa7SDarrick J. Wong } 1913ec6efa7SDarrick J. Wong length = be32_to_cpu(icl->icl_length); 1923ec6efa7SDarrick J. Wong if (!length || length >= mp->m_sb.sb_agblocks) { 1933ec6efa7SDarrick J. Wong xfs_warn(log->l_mp, "xlog_recover_do_icreate_trans: bad length"); 1943ec6efa7SDarrick J. Wong return -EINVAL; 1953ec6efa7SDarrick J. Wong } 1963ec6efa7SDarrick J. Wong 1973ec6efa7SDarrick J. Wong /* 1983ec6efa7SDarrick J. Wong * The inode chunk is either full or sparse and we only support 1993ec6efa7SDarrick J. Wong * m_ino_geo.ialloc_min_blks sized sparse allocations at this time. 2003ec6efa7SDarrick J. Wong */ 2013ec6efa7SDarrick J. Wong if (length != igeo->ialloc_blks && 2023ec6efa7SDarrick J. Wong length != igeo->ialloc_min_blks) { 2033ec6efa7SDarrick J. Wong xfs_warn(log->l_mp, 204*edf27485SDwaipayan Ray "%s: unsupported chunk length", __func__); 2053ec6efa7SDarrick J. Wong return -EINVAL; 2063ec6efa7SDarrick J. Wong } 2073ec6efa7SDarrick J. Wong 2083ec6efa7SDarrick J. Wong /* verify inode count is consistent with extent length */ 2093ec6efa7SDarrick J. Wong if ((count >> mp->m_sb.sb_inopblog) != length) { 2103ec6efa7SDarrick J. Wong xfs_warn(log->l_mp, 2113ec6efa7SDarrick J. Wong "%s: inconsistent inode count and chunk length", 212*edf27485SDwaipayan Ray __func__); 2133ec6efa7SDarrick J. Wong return -EINVAL; 2143ec6efa7SDarrick J. Wong } 2153ec6efa7SDarrick J. Wong 2163ec6efa7SDarrick J. Wong /* 2173ec6efa7SDarrick J. Wong * The icreate transaction can cover multiple cluster buffers and these 2183ec6efa7SDarrick J. Wong * buffers could have been freed and reused. Check the individual 2193ec6efa7SDarrick J. Wong * buffers for cancellation so we don't overwrite anything written after 2203ec6efa7SDarrick J. Wong * a cancellation. 2213ec6efa7SDarrick J. Wong */ 2223ec6efa7SDarrick J. Wong bb_per_cluster = XFS_FSB_TO_BB(mp, igeo->blocks_per_cluster); 2233ec6efa7SDarrick J. Wong nbufs = length / igeo->blocks_per_cluster; 2243ec6efa7SDarrick J. Wong for (i = 0, cancel_count = 0; i < nbufs; i++) { 2253ec6efa7SDarrick J. Wong xfs_daddr_t daddr; 2263ec6efa7SDarrick J. Wong 2273ec6efa7SDarrick J. Wong daddr = XFS_AGB_TO_DADDR(mp, agno, 2283ec6efa7SDarrick J. Wong agbno + i * igeo->blocks_per_cluster); 2293ec6efa7SDarrick J. Wong if (xlog_is_buffer_cancelled(log, daddr, bb_per_cluster)) 2303ec6efa7SDarrick J. Wong cancel_count++; 2313ec6efa7SDarrick J. Wong } 2323ec6efa7SDarrick J. Wong 2333ec6efa7SDarrick J. Wong /* 2343ec6efa7SDarrick J. Wong * We currently only use icreate for a single allocation at a time. This 2353ec6efa7SDarrick J. Wong * means we should expect either all or none of the buffers to be 2363ec6efa7SDarrick J. Wong * cancelled. Be conservative and skip replay if at least one buffer is 2373ec6efa7SDarrick J. Wong * cancelled, but warn the user that something is awry if the buffers 2383ec6efa7SDarrick J. Wong * are not consistent. 2393ec6efa7SDarrick J. Wong * 2403ec6efa7SDarrick J. Wong * XXX: This must be refined to only skip cancelled clusters once we use 2413ec6efa7SDarrick J. Wong * icreate for multiple chunk allocations. 2423ec6efa7SDarrick J. Wong */ 2433ec6efa7SDarrick J. Wong ASSERT(!cancel_count || cancel_count == nbufs); 2443ec6efa7SDarrick J. Wong if (cancel_count) { 2453ec6efa7SDarrick J. Wong if (cancel_count != nbufs) 2463ec6efa7SDarrick J. Wong xfs_warn(mp, 2473ec6efa7SDarrick J. Wong "WARNING: partial inode chunk cancellation, skipped icreate."); 2483ec6efa7SDarrick J. Wong trace_xfs_log_recover_icreate_cancel(log, icl); 2493ec6efa7SDarrick J. Wong return 0; 2503ec6efa7SDarrick J. Wong } 2513ec6efa7SDarrick J. Wong 2523ec6efa7SDarrick J. Wong trace_xfs_log_recover_icreate_recover(log, icl); 2533ec6efa7SDarrick J. Wong return xfs_ialloc_inode_init(mp, NULL, buffer_list, count, agno, agbno, 2543ec6efa7SDarrick J. Wong length, be32_to_cpu(icl->icl_gen)); 2553ec6efa7SDarrick J. Wong } 2563ec6efa7SDarrick J. Wong 25786ffa471SDarrick J. Wong const struct xlog_recover_item_ops xlog_icreate_item_ops = { 25886ffa471SDarrick J. Wong .item_type = XFS_LI_ICREATE, 25986ffa471SDarrick J. Wong .reorder = xlog_recover_icreate_reorder, 2603ec6efa7SDarrick J. Wong .commit_pass2 = xlog_recover_icreate_commit_pass2, 26186ffa471SDarrick J. Wong }; 262