xref: /openbmc/linux/fs/xfs/scrub/agheader.c (revision ba61bb17)
1 // SPDX-License-Identifier: GPL-2.0+
2 /*
3  * Copyright (C) 2017 Oracle.  All Rights Reserved.
4  * Author: Darrick J. Wong <darrick.wong@oracle.com>
5  */
6 #include "xfs.h"
7 #include "xfs_fs.h"
8 #include "xfs_shared.h"
9 #include "xfs_format.h"
10 #include "xfs_trans_resv.h"
11 #include "xfs_mount.h"
12 #include "xfs_defer.h"
13 #include "xfs_btree.h"
14 #include "xfs_bit.h"
15 #include "xfs_log_format.h"
16 #include "xfs_trans.h"
17 #include "xfs_sb.h"
18 #include "xfs_inode.h"
19 #include "xfs_alloc.h"
20 #include "xfs_ialloc.h"
21 #include "xfs_rmap.h"
22 #include "scrub/xfs_scrub.h"
23 #include "scrub/scrub.h"
24 #include "scrub/common.h"
25 #include "scrub/trace.h"
26 
27 /* Superblock */
28 
29 /* Cross-reference with the other btrees. */
30 STATIC void
31 xfs_scrub_superblock_xref(
32 	struct xfs_scrub_context	*sc,
33 	struct xfs_buf			*bp)
34 {
35 	struct xfs_owner_info		oinfo;
36 	struct xfs_mount		*mp = sc->mp;
37 	xfs_agnumber_t			agno = sc->sm->sm_agno;
38 	xfs_agblock_t			agbno;
39 	int				error;
40 
41 	if (sc->sm->sm_flags & XFS_SCRUB_OFLAG_CORRUPT)
42 		return;
43 
44 	agbno = XFS_SB_BLOCK(mp);
45 
46 	error = xfs_scrub_ag_init(sc, agno, &sc->sa);
47 	if (!xfs_scrub_xref_process_error(sc, agno, agbno, &error))
48 		return;
49 
50 	xfs_scrub_xref_is_used_space(sc, agbno, 1);
51 	xfs_scrub_xref_is_not_inode_chunk(sc, agbno, 1);
52 	xfs_rmap_ag_owner(&oinfo, XFS_RMAP_OWN_FS);
53 	xfs_scrub_xref_is_owned_by(sc, agbno, 1, &oinfo);
54 	xfs_scrub_xref_is_not_shared(sc, agbno, 1);
55 
56 	/* scrub teardown will take care of sc->sa for us */
57 }
58 
59 /*
60  * Scrub the filesystem superblock.
61  *
62  * Note: We do /not/ attempt to check AG 0's superblock.  Mount is
63  * responsible for validating all the geometry information in sb 0, so
64  * if the filesystem is capable of initiating online scrub, then clearly
65  * sb 0 is ok and we can use its information to check everything else.
66  */
67 int
68 xfs_scrub_superblock(
69 	struct xfs_scrub_context	*sc)
70 {
71 	struct xfs_mount		*mp = sc->mp;
72 	struct xfs_buf			*bp;
73 	struct xfs_dsb			*sb;
74 	xfs_agnumber_t			agno;
75 	uint32_t			v2_ok;
76 	__be32				features_mask;
77 	int				error;
78 	__be16				vernum_mask;
79 
80 	agno = sc->sm->sm_agno;
81 	if (agno == 0)
82 		return 0;
83 
84 	error = xfs_sb_read_secondary(mp, sc->tp, agno, &bp);
85 	/*
86 	 * The superblock verifier can return several different error codes
87 	 * if it thinks the superblock doesn't look right.  For a mount these
88 	 * would all get bounced back to userspace, but if we're here then the
89 	 * fs mounted successfully, which means that this secondary superblock
90 	 * is simply incorrect.  Treat all these codes the same way we treat
91 	 * any corruption.
92 	 */
93 	switch (error) {
94 	case -EINVAL:	/* also -EWRONGFS */
95 	case -ENOSYS:
96 	case -EFBIG:
97 		error = -EFSCORRUPTED;
98 	default:
99 		break;
100 	}
101 	if (!xfs_scrub_process_error(sc, agno, XFS_SB_BLOCK(mp), &error))
102 		return error;
103 
104 	sb = XFS_BUF_TO_SBP(bp);
105 
106 	/*
107 	 * Verify the geometries match.  Fields that are permanently
108 	 * set by mkfs are checked; fields that can be updated later
109 	 * (and are not propagated to backup superblocks) are preen
110 	 * checked.
111 	 */
112 	if (sb->sb_blocksize != cpu_to_be32(mp->m_sb.sb_blocksize))
113 		xfs_scrub_block_set_corrupt(sc, bp);
114 
115 	if (sb->sb_dblocks != cpu_to_be64(mp->m_sb.sb_dblocks))
116 		xfs_scrub_block_set_corrupt(sc, bp);
117 
118 	if (sb->sb_rblocks != cpu_to_be64(mp->m_sb.sb_rblocks))
119 		xfs_scrub_block_set_corrupt(sc, bp);
120 
121 	if (sb->sb_rextents != cpu_to_be64(mp->m_sb.sb_rextents))
122 		xfs_scrub_block_set_corrupt(sc, bp);
123 
124 	if (!uuid_equal(&sb->sb_uuid, &mp->m_sb.sb_uuid))
125 		xfs_scrub_block_set_preen(sc, bp);
126 
127 	if (sb->sb_logstart != cpu_to_be64(mp->m_sb.sb_logstart))
128 		xfs_scrub_block_set_corrupt(sc, bp);
129 
130 	if (sb->sb_rootino != cpu_to_be64(mp->m_sb.sb_rootino))
131 		xfs_scrub_block_set_preen(sc, bp);
132 
133 	if (sb->sb_rbmino != cpu_to_be64(mp->m_sb.sb_rbmino))
134 		xfs_scrub_block_set_preen(sc, bp);
135 
136 	if (sb->sb_rsumino != cpu_to_be64(mp->m_sb.sb_rsumino))
137 		xfs_scrub_block_set_preen(sc, bp);
138 
139 	if (sb->sb_rextsize != cpu_to_be32(mp->m_sb.sb_rextsize))
140 		xfs_scrub_block_set_corrupt(sc, bp);
141 
142 	if (sb->sb_agblocks != cpu_to_be32(mp->m_sb.sb_agblocks))
143 		xfs_scrub_block_set_corrupt(sc, bp);
144 
145 	if (sb->sb_agcount != cpu_to_be32(mp->m_sb.sb_agcount))
146 		xfs_scrub_block_set_corrupt(sc, bp);
147 
148 	if (sb->sb_rbmblocks != cpu_to_be32(mp->m_sb.sb_rbmblocks))
149 		xfs_scrub_block_set_corrupt(sc, bp);
150 
151 	if (sb->sb_logblocks != cpu_to_be32(mp->m_sb.sb_logblocks))
152 		xfs_scrub_block_set_corrupt(sc, bp);
153 
154 	/* Check sb_versionnum bits that are set at mkfs time. */
155 	vernum_mask = cpu_to_be16(~XFS_SB_VERSION_OKBITS |
156 				  XFS_SB_VERSION_NUMBITS |
157 				  XFS_SB_VERSION_ALIGNBIT |
158 				  XFS_SB_VERSION_DALIGNBIT |
159 				  XFS_SB_VERSION_SHAREDBIT |
160 				  XFS_SB_VERSION_LOGV2BIT |
161 				  XFS_SB_VERSION_SECTORBIT |
162 				  XFS_SB_VERSION_EXTFLGBIT |
163 				  XFS_SB_VERSION_DIRV2BIT);
164 	if ((sb->sb_versionnum & vernum_mask) !=
165 	    (cpu_to_be16(mp->m_sb.sb_versionnum) & vernum_mask))
166 		xfs_scrub_block_set_corrupt(sc, bp);
167 
168 	/* Check sb_versionnum bits that can be set after mkfs time. */
169 	vernum_mask = cpu_to_be16(XFS_SB_VERSION_ATTRBIT |
170 				  XFS_SB_VERSION_NLINKBIT |
171 				  XFS_SB_VERSION_QUOTABIT);
172 	if ((sb->sb_versionnum & vernum_mask) !=
173 	    (cpu_to_be16(mp->m_sb.sb_versionnum) & vernum_mask))
174 		xfs_scrub_block_set_preen(sc, bp);
175 
176 	if (sb->sb_sectsize != cpu_to_be16(mp->m_sb.sb_sectsize))
177 		xfs_scrub_block_set_corrupt(sc, bp);
178 
179 	if (sb->sb_inodesize != cpu_to_be16(mp->m_sb.sb_inodesize))
180 		xfs_scrub_block_set_corrupt(sc, bp);
181 
182 	if (sb->sb_inopblock != cpu_to_be16(mp->m_sb.sb_inopblock))
183 		xfs_scrub_block_set_corrupt(sc, bp);
184 
185 	if (memcmp(sb->sb_fname, mp->m_sb.sb_fname, sizeof(sb->sb_fname)))
186 		xfs_scrub_block_set_preen(sc, bp);
187 
188 	if (sb->sb_blocklog != mp->m_sb.sb_blocklog)
189 		xfs_scrub_block_set_corrupt(sc, bp);
190 
191 	if (sb->sb_sectlog != mp->m_sb.sb_sectlog)
192 		xfs_scrub_block_set_corrupt(sc, bp);
193 
194 	if (sb->sb_inodelog != mp->m_sb.sb_inodelog)
195 		xfs_scrub_block_set_corrupt(sc, bp);
196 
197 	if (sb->sb_inopblog != mp->m_sb.sb_inopblog)
198 		xfs_scrub_block_set_corrupt(sc, bp);
199 
200 	if (sb->sb_agblklog != mp->m_sb.sb_agblklog)
201 		xfs_scrub_block_set_corrupt(sc, bp);
202 
203 	if (sb->sb_rextslog != mp->m_sb.sb_rextslog)
204 		xfs_scrub_block_set_corrupt(sc, bp);
205 
206 	if (sb->sb_imax_pct != mp->m_sb.sb_imax_pct)
207 		xfs_scrub_block_set_preen(sc, bp);
208 
209 	/*
210 	 * Skip the summary counters since we track them in memory anyway.
211 	 * sb_icount, sb_ifree, sb_fdblocks, sb_frexents
212 	 */
213 
214 	if (sb->sb_uquotino != cpu_to_be64(mp->m_sb.sb_uquotino))
215 		xfs_scrub_block_set_preen(sc, bp);
216 
217 	if (sb->sb_gquotino != cpu_to_be64(mp->m_sb.sb_gquotino))
218 		xfs_scrub_block_set_preen(sc, bp);
219 
220 	/*
221 	 * Skip the quota flags since repair will force quotacheck.
222 	 * sb_qflags
223 	 */
224 
225 	if (sb->sb_flags != mp->m_sb.sb_flags)
226 		xfs_scrub_block_set_corrupt(sc, bp);
227 
228 	if (sb->sb_shared_vn != mp->m_sb.sb_shared_vn)
229 		xfs_scrub_block_set_corrupt(sc, bp);
230 
231 	if (sb->sb_inoalignmt != cpu_to_be32(mp->m_sb.sb_inoalignmt))
232 		xfs_scrub_block_set_corrupt(sc, bp);
233 
234 	if (sb->sb_unit != cpu_to_be32(mp->m_sb.sb_unit))
235 		xfs_scrub_block_set_preen(sc, bp);
236 
237 	if (sb->sb_width != cpu_to_be32(mp->m_sb.sb_width))
238 		xfs_scrub_block_set_preen(sc, bp);
239 
240 	if (sb->sb_dirblklog != mp->m_sb.sb_dirblklog)
241 		xfs_scrub_block_set_corrupt(sc, bp);
242 
243 	if (sb->sb_logsectlog != mp->m_sb.sb_logsectlog)
244 		xfs_scrub_block_set_corrupt(sc, bp);
245 
246 	if (sb->sb_logsectsize != cpu_to_be16(mp->m_sb.sb_logsectsize))
247 		xfs_scrub_block_set_corrupt(sc, bp);
248 
249 	if (sb->sb_logsunit != cpu_to_be32(mp->m_sb.sb_logsunit))
250 		xfs_scrub_block_set_corrupt(sc, bp);
251 
252 	/* Do we see any invalid bits in sb_features2? */
253 	if (!xfs_sb_version_hasmorebits(&mp->m_sb)) {
254 		if (sb->sb_features2 != 0)
255 			xfs_scrub_block_set_corrupt(sc, bp);
256 	} else {
257 		v2_ok = XFS_SB_VERSION2_OKBITS;
258 		if (XFS_SB_VERSION_NUM(&mp->m_sb) >= XFS_SB_VERSION_5)
259 			v2_ok |= XFS_SB_VERSION2_CRCBIT;
260 
261 		if (!!(sb->sb_features2 & cpu_to_be32(~v2_ok)))
262 			xfs_scrub_block_set_corrupt(sc, bp);
263 
264 		if (sb->sb_features2 != sb->sb_bad_features2)
265 			xfs_scrub_block_set_preen(sc, bp);
266 	}
267 
268 	/* Check sb_features2 flags that are set at mkfs time. */
269 	features_mask = cpu_to_be32(XFS_SB_VERSION2_LAZYSBCOUNTBIT |
270 				    XFS_SB_VERSION2_PROJID32BIT |
271 				    XFS_SB_VERSION2_CRCBIT |
272 				    XFS_SB_VERSION2_FTYPE);
273 	if ((sb->sb_features2 & features_mask) !=
274 	    (cpu_to_be32(mp->m_sb.sb_features2) & features_mask))
275 		xfs_scrub_block_set_corrupt(sc, bp);
276 
277 	/* Check sb_features2 flags that can be set after mkfs time. */
278 	features_mask = cpu_to_be32(XFS_SB_VERSION2_ATTR2BIT);
279 	if ((sb->sb_features2 & features_mask) !=
280 	    (cpu_to_be32(mp->m_sb.sb_features2) & features_mask))
281 		xfs_scrub_block_set_corrupt(sc, bp);
282 
283 	if (!xfs_sb_version_hascrc(&mp->m_sb)) {
284 		/* all v5 fields must be zero */
285 		if (memchr_inv(&sb->sb_features_compat, 0,
286 				sizeof(struct xfs_dsb) -
287 				offsetof(struct xfs_dsb, sb_features_compat)))
288 			xfs_scrub_block_set_corrupt(sc, bp);
289 	} else {
290 		/* Check compat flags; all are set at mkfs time. */
291 		features_mask = cpu_to_be32(XFS_SB_FEAT_COMPAT_UNKNOWN);
292 		if ((sb->sb_features_compat & features_mask) !=
293 		    (cpu_to_be32(mp->m_sb.sb_features_compat) & features_mask))
294 			xfs_scrub_block_set_corrupt(sc, bp);
295 
296 		/* Check ro compat flags; all are set at mkfs time. */
297 		features_mask = cpu_to_be32(XFS_SB_FEAT_RO_COMPAT_UNKNOWN |
298 					    XFS_SB_FEAT_RO_COMPAT_FINOBT |
299 					    XFS_SB_FEAT_RO_COMPAT_RMAPBT |
300 					    XFS_SB_FEAT_RO_COMPAT_REFLINK);
301 		if ((sb->sb_features_ro_compat & features_mask) !=
302 		    (cpu_to_be32(mp->m_sb.sb_features_ro_compat) &
303 		     features_mask))
304 			xfs_scrub_block_set_corrupt(sc, bp);
305 
306 		/* Check incompat flags; all are set at mkfs time. */
307 		features_mask = cpu_to_be32(XFS_SB_FEAT_INCOMPAT_UNKNOWN |
308 					    XFS_SB_FEAT_INCOMPAT_FTYPE |
309 					    XFS_SB_FEAT_INCOMPAT_SPINODES |
310 					    XFS_SB_FEAT_INCOMPAT_META_UUID);
311 		if ((sb->sb_features_incompat & features_mask) !=
312 		    (cpu_to_be32(mp->m_sb.sb_features_incompat) &
313 		     features_mask))
314 			xfs_scrub_block_set_corrupt(sc, bp);
315 
316 		/* Check log incompat flags; all are set at mkfs time. */
317 		features_mask = cpu_to_be32(XFS_SB_FEAT_INCOMPAT_LOG_UNKNOWN);
318 		if ((sb->sb_features_log_incompat & features_mask) !=
319 		    (cpu_to_be32(mp->m_sb.sb_features_log_incompat) &
320 		     features_mask))
321 			xfs_scrub_block_set_corrupt(sc, bp);
322 
323 		/* Don't care about sb_crc */
324 
325 		if (sb->sb_spino_align != cpu_to_be32(mp->m_sb.sb_spino_align))
326 			xfs_scrub_block_set_corrupt(sc, bp);
327 
328 		if (sb->sb_pquotino != cpu_to_be64(mp->m_sb.sb_pquotino))
329 			xfs_scrub_block_set_preen(sc, bp);
330 
331 		/* Don't care about sb_lsn */
332 	}
333 
334 	if (xfs_sb_version_hasmetauuid(&mp->m_sb)) {
335 		/* The metadata UUID must be the same for all supers */
336 		if (!uuid_equal(&sb->sb_meta_uuid, &mp->m_sb.sb_meta_uuid))
337 			xfs_scrub_block_set_corrupt(sc, bp);
338 	}
339 
340 	/* Everything else must be zero. */
341 	if (memchr_inv(sb + 1, 0,
342 			BBTOB(bp->b_length) - sizeof(struct xfs_dsb)))
343 		xfs_scrub_block_set_corrupt(sc, bp);
344 
345 	xfs_scrub_superblock_xref(sc, bp);
346 
347 	return error;
348 }
349 
350 /* AGF */
351 
352 /* Tally freespace record lengths. */
353 STATIC int
354 xfs_scrub_agf_record_bno_lengths(
355 	struct xfs_btree_cur		*cur,
356 	struct xfs_alloc_rec_incore	*rec,
357 	void				*priv)
358 {
359 	xfs_extlen_t			*blocks = priv;
360 
361 	(*blocks) += rec->ar_blockcount;
362 	return 0;
363 }
364 
365 /* Check agf_freeblks */
366 static inline void
367 xfs_scrub_agf_xref_freeblks(
368 	struct xfs_scrub_context	*sc)
369 {
370 	struct xfs_agf			*agf = XFS_BUF_TO_AGF(sc->sa.agf_bp);
371 	xfs_extlen_t			blocks = 0;
372 	int				error;
373 
374 	if (!sc->sa.bno_cur)
375 		return;
376 
377 	error = xfs_alloc_query_all(sc->sa.bno_cur,
378 			xfs_scrub_agf_record_bno_lengths, &blocks);
379 	if (!xfs_scrub_should_check_xref(sc, &error, &sc->sa.bno_cur))
380 		return;
381 	if (blocks != be32_to_cpu(agf->agf_freeblks))
382 		xfs_scrub_block_xref_set_corrupt(sc, sc->sa.agf_bp);
383 }
384 
385 /* Cross reference the AGF with the cntbt (freespace by length btree) */
386 static inline void
387 xfs_scrub_agf_xref_cntbt(
388 	struct xfs_scrub_context	*sc)
389 {
390 	struct xfs_agf			*agf = XFS_BUF_TO_AGF(sc->sa.agf_bp);
391 	xfs_agblock_t			agbno;
392 	xfs_extlen_t			blocks;
393 	int				have;
394 	int				error;
395 
396 	if (!sc->sa.cnt_cur)
397 		return;
398 
399 	/* Any freespace at all? */
400 	error = xfs_alloc_lookup_le(sc->sa.cnt_cur, 0, -1U, &have);
401 	if (!xfs_scrub_should_check_xref(sc, &error, &sc->sa.cnt_cur))
402 		return;
403 	if (!have) {
404 		if (agf->agf_freeblks != be32_to_cpu(0))
405 			xfs_scrub_block_xref_set_corrupt(sc, sc->sa.agf_bp);
406 		return;
407 	}
408 
409 	/* Check agf_longest */
410 	error = xfs_alloc_get_rec(sc->sa.cnt_cur, &agbno, &blocks, &have);
411 	if (!xfs_scrub_should_check_xref(sc, &error, &sc->sa.cnt_cur))
412 		return;
413 	if (!have || blocks != be32_to_cpu(agf->agf_longest))
414 		xfs_scrub_block_xref_set_corrupt(sc, sc->sa.agf_bp);
415 }
416 
417 /* Check the btree block counts in the AGF against the btrees. */
418 STATIC void
419 xfs_scrub_agf_xref_btreeblks(
420 	struct xfs_scrub_context	*sc)
421 {
422 	struct xfs_agf			*agf = XFS_BUF_TO_AGF(sc->sa.agf_bp);
423 	struct xfs_mount		*mp = sc->mp;
424 	xfs_agblock_t			blocks;
425 	xfs_agblock_t			btreeblks;
426 	int				error;
427 
428 	/* Check agf_rmap_blocks; set up for agf_btreeblks check */
429 	if (sc->sa.rmap_cur) {
430 		error = xfs_btree_count_blocks(sc->sa.rmap_cur, &blocks);
431 		if (!xfs_scrub_should_check_xref(sc, &error, &sc->sa.rmap_cur))
432 			return;
433 		btreeblks = blocks - 1;
434 		if (blocks != be32_to_cpu(agf->agf_rmap_blocks))
435 			xfs_scrub_block_xref_set_corrupt(sc, sc->sa.agf_bp);
436 	} else {
437 		btreeblks = 0;
438 	}
439 
440 	/*
441 	 * No rmap cursor; we can't xref if we have the rmapbt feature.
442 	 * We also can't do it if we're missing the free space btree cursors.
443 	 */
444 	if ((xfs_sb_version_hasrmapbt(&mp->m_sb) && !sc->sa.rmap_cur) ||
445 	    !sc->sa.bno_cur || !sc->sa.cnt_cur)
446 		return;
447 
448 	/* Check agf_btreeblks */
449 	error = xfs_btree_count_blocks(sc->sa.bno_cur, &blocks);
450 	if (!xfs_scrub_should_check_xref(sc, &error, &sc->sa.bno_cur))
451 		return;
452 	btreeblks += blocks - 1;
453 
454 	error = xfs_btree_count_blocks(sc->sa.cnt_cur, &blocks);
455 	if (!xfs_scrub_should_check_xref(sc, &error, &sc->sa.cnt_cur))
456 		return;
457 	btreeblks += blocks - 1;
458 
459 	if (btreeblks != be32_to_cpu(agf->agf_btreeblks))
460 		xfs_scrub_block_xref_set_corrupt(sc, sc->sa.agf_bp);
461 }
462 
463 /* Check agf_refcount_blocks against tree size */
464 static inline void
465 xfs_scrub_agf_xref_refcblks(
466 	struct xfs_scrub_context	*sc)
467 {
468 	struct xfs_agf			*agf = XFS_BUF_TO_AGF(sc->sa.agf_bp);
469 	xfs_agblock_t			blocks;
470 	int				error;
471 
472 	if (!sc->sa.refc_cur)
473 		return;
474 
475 	error = xfs_btree_count_blocks(sc->sa.refc_cur, &blocks);
476 	if (!xfs_scrub_should_check_xref(sc, &error, &sc->sa.refc_cur))
477 		return;
478 	if (blocks != be32_to_cpu(agf->agf_refcount_blocks))
479 		xfs_scrub_block_xref_set_corrupt(sc, sc->sa.agf_bp);
480 }
481 
482 /* Cross-reference with the other btrees. */
483 STATIC void
484 xfs_scrub_agf_xref(
485 	struct xfs_scrub_context	*sc)
486 {
487 	struct xfs_owner_info		oinfo;
488 	struct xfs_mount		*mp = sc->mp;
489 	xfs_agblock_t			agbno;
490 	int				error;
491 
492 	if (sc->sm->sm_flags & XFS_SCRUB_OFLAG_CORRUPT)
493 		return;
494 
495 	agbno = XFS_AGF_BLOCK(mp);
496 
497 	error = xfs_scrub_ag_btcur_init(sc, &sc->sa);
498 	if (error)
499 		return;
500 
501 	xfs_scrub_xref_is_used_space(sc, agbno, 1);
502 	xfs_scrub_agf_xref_freeblks(sc);
503 	xfs_scrub_agf_xref_cntbt(sc);
504 	xfs_scrub_xref_is_not_inode_chunk(sc, agbno, 1);
505 	xfs_rmap_ag_owner(&oinfo, XFS_RMAP_OWN_FS);
506 	xfs_scrub_xref_is_owned_by(sc, agbno, 1, &oinfo);
507 	xfs_scrub_agf_xref_btreeblks(sc);
508 	xfs_scrub_xref_is_not_shared(sc, agbno, 1);
509 	xfs_scrub_agf_xref_refcblks(sc);
510 
511 	/* scrub teardown will take care of sc->sa for us */
512 }
513 
514 /* Scrub the AGF. */
515 int
516 xfs_scrub_agf(
517 	struct xfs_scrub_context	*sc)
518 {
519 	struct xfs_mount		*mp = sc->mp;
520 	struct xfs_agf			*agf;
521 	xfs_agnumber_t			agno;
522 	xfs_agblock_t			agbno;
523 	xfs_agblock_t			eoag;
524 	xfs_agblock_t			agfl_first;
525 	xfs_agblock_t			agfl_last;
526 	xfs_agblock_t			agfl_count;
527 	xfs_agblock_t			fl_count;
528 	int				level;
529 	int				error = 0;
530 
531 	agno = sc->sa.agno = sc->sm->sm_agno;
532 	error = xfs_scrub_ag_read_headers(sc, agno, &sc->sa.agi_bp,
533 			&sc->sa.agf_bp, &sc->sa.agfl_bp);
534 	if (!xfs_scrub_process_error(sc, agno, XFS_AGF_BLOCK(sc->mp), &error))
535 		goto out;
536 	xfs_scrub_buffer_recheck(sc, sc->sa.agf_bp);
537 
538 	agf = XFS_BUF_TO_AGF(sc->sa.agf_bp);
539 
540 	/* Check the AG length */
541 	eoag = be32_to_cpu(agf->agf_length);
542 	if (eoag != xfs_ag_block_count(mp, agno))
543 		xfs_scrub_block_set_corrupt(sc, sc->sa.agf_bp);
544 
545 	/* Check the AGF btree roots and levels */
546 	agbno = be32_to_cpu(agf->agf_roots[XFS_BTNUM_BNO]);
547 	if (!xfs_verify_agbno(mp, agno, agbno))
548 		xfs_scrub_block_set_corrupt(sc, sc->sa.agf_bp);
549 
550 	agbno = be32_to_cpu(agf->agf_roots[XFS_BTNUM_CNT]);
551 	if (!xfs_verify_agbno(mp, agno, agbno))
552 		xfs_scrub_block_set_corrupt(sc, sc->sa.agf_bp);
553 
554 	level = be32_to_cpu(agf->agf_levels[XFS_BTNUM_BNO]);
555 	if (level <= 0 || level > XFS_BTREE_MAXLEVELS)
556 		xfs_scrub_block_set_corrupt(sc, sc->sa.agf_bp);
557 
558 	level = be32_to_cpu(agf->agf_levels[XFS_BTNUM_CNT]);
559 	if (level <= 0 || level > XFS_BTREE_MAXLEVELS)
560 		xfs_scrub_block_set_corrupt(sc, sc->sa.agf_bp);
561 
562 	if (xfs_sb_version_hasrmapbt(&mp->m_sb)) {
563 		agbno = be32_to_cpu(agf->agf_roots[XFS_BTNUM_RMAP]);
564 		if (!xfs_verify_agbno(mp, agno, agbno))
565 			xfs_scrub_block_set_corrupt(sc, sc->sa.agf_bp);
566 
567 		level = be32_to_cpu(agf->agf_levels[XFS_BTNUM_RMAP]);
568 		if (level <= 0 || level > XFS_BTREE_MAXLEVELS)
569 			xfs_scrub_block_set_corrupt(sc, sc->sa.agf_bp);
570 	}
571 
572 	if (xfs_sb_version_hasreflink(&mp->m_sb)) {
573 		agbno = be32_to_cpu(agf->agf_refcount_root);
574 		if (!xfs_verify_agbno(mp, agno, agbno))
575 			xfs_scrub_block_set_corrupt(sc, sc->sa.agf_bp);
576 
577 		level = be32_to_cpu(agf->agf_refcount_level);
578 		if (level <= 0 || level > XFS_BTREE_MAXLEVELS)
579 			xfs_scrub_block_set_corrupt(sc, sc->sa.agf_bp);
580 	}
581 
582 	/* Check the AGFL counters */
583 	agfl_first = be32_to_cpu(agf->agf_flfirst);
584 	agfl_last = be32_to_cpu(agf->agf_fllast);
585 	agfl_count = be32_to_cpu(agf->agf_flcount);
586 	if (agfl_last > agfl_first)
587 		fl_count = agfl_last - agfl_first + 1;
588 	else
589 		fl_count = xfs_agfl_size(mp) - agfl_first + agfl_last + 1;
590 	if (agfl_count != 0 && fl_count != agfl_count)
591 		xfs_scrub_block_set_corrupt(sc, sc->sa.agf_bp);
592 
593 	xfs_scrub_agf_xref(sc);
594 out:
595 	return error;
596 }
597 
598 /* AGFL */
599 
600 struct xfs_scrub_agfl_info {
601 	struct xfs_owner_info		oinfo;
602 	unsigned int			sz_entries;
603 	unsigned int			nr_entries;
604 	xfs_agblock_t			*entries;
605 	struct xfs_scrub_context	*sc;
606 };
607 
608 /* Cross-reference with the other btrees. */
609 STATIC void
610 xfs_scrub_agfl_block_xref(
611 	struct xfs_scrub_context	*sc,
612 	xfs_agblock_t			agbno,
613 	struct xfs_owner_info		*oinfo)
614 {
615 	if (sc->sm->sm_flags & XFS_SCRUB_OFLAG_CORRUPT)
616 		return;
617 
618 	xfs_scrub_xref_is_used_space(sc, agbno, 1);
619 	xfs_scrub_xref_is_not_inode_chunk(sc, agbno, 1);
620 	xfs_scrub_xref_is_owned_by(sc, agbno, 1, oinfo);
621 	xfs_scrub_xref_is_not_shared(sc, agbno, 1);
622 }
623 
624 /* Scrub an AGFL block. */
625 STATIC int
626 xfs_scrub_agfl_block(
627 	struct xfs_mount		*mp,
628 	xfs_agblock_t			agbno,
629 	void				*priv)
630 {
631 	struct xfs_scrub_agfl_info	*sai = priv;
632 	struct xfs_scrub_context	*sc = sai->sc;
633 	xfs_agnumber_t			agno = sc->sa.agno;
634 
635 	if (xfs_verify_agbno(mp, agno, agbno) &&
636 	    sai->nr_entries < sai->sz_entries)
637 		sai->entries[sai->nr_entries++] = agbno;
638 	else
639 		xfs_scrub_block_set_corrupt(sc, sc->sa.agfl_bp);
640 
641 	xfs_scrub_agfl_block_xref(sc, agbno, priv);
642 
643 	if (sc->sm->sm_flags & XFS_SCRUB_OFLAG_CORRUPT)
644 		return XFS_BTREE_QUERY_RANGE_ABORT;
645 
646 	return 0;
647 }
648 
649 static int
650 xfs_scrub_agblock_cmp(
651 	const void		*pa,
652 	const void		*pb)
653 {
654 	const xfs_agblock_t	*a = pa;
655 	const xfs_agblock_t	*b = pb;
656 
657 	return (int)*a - (int)*b;
658 }
659 
660 /* Cross-reference with the other btrees. */
661 STATIC void
662 xfs_scrub_agfl_xref(
663 	struct xfs_scrub_context	*sc)
664 {
665 	struct xfs_owner_info		oinfo;
666 	struct xfs_mount		*mp = sc->mp;
667 	xfs_agblock_t			agbno;
668 	int				error;
669 
670 	if (sc->sm->sm_flags & XFS_SCRUB_OFLAG_CORRUPT)
671 		return;
672 
673 	agbno = XFS_AGFL_BLOCK(mp);
674 
675 	error = xfs_scrub_ag_btcur_init(sc, &sc->sa);
676 	if (error)
677 		return;
678 
679 	xfs_scrub_xref_is_used_space(sc, agbno, 1);
680 	xfs_scrub_xref_is_not_inode_chunk(sc, agbno, 1);
681 	xfs_rmap_ag_owner(&oinfo, XFS_RMAP_OWN_FS);
682 	xfs_scrub_xref_is_owned_by(sc, agbno, 1, &oinfo);
683 	xfs_scrub_xref_is_not_shared(sc, agbno, 1);
684 
685 	/*
686 	 * Scrub teardown will take care of sc->sa for us.  Leave sc->sa
687 	 * active so that the agfl block xref can use it too.
688 	 */
689 }
690 
691 /* Scrub the AGFL. */
692 int
693 xfs_scrub_agfl(
694 	struct xfs_scrub_context	*sc)
695 {
696 	struct xfs_scrub_agfl_info	sai;
697 	struct xfs_agf			*agf;
698 	xfs_agnumber_t			agno;
699 	unsigned int			agflcount;
700 	unsigned int			i;
701 	int				error;
702 
703 	agno = sc->sa.agno = sc->sm->sm_agno;
704 	error = xfs_scrub_ag_read_headers(sc, agno, &sc->sa.agi_bp,
705 			&sc->sa.agf_bp, &sc->sa.agfl_bp);
706 	if (!xfs_scrub_process_error(sc, agno, XFS_AGFL_BLOCK(sc->mp), &error))
707 		goto out;
708 	if (!sc->sa.agf_bp)
709 		return -EFSCORRUPTED;
710 	xfs_scrub_buffer_recheck(sc, sc->sa.agfl_bp);
711 
712 	xfs_scrub_agfl_xref(sc);
713 
714 	if (sc->sm->sm_flags & XFS_SCRUB_OFLAG_CORRUPT)
715 		goto out;
716 
717 	/* Allocate buffer to ensure uniqueness of AGFL entries. */
718 	agf = XFS_BUF_TO_AGF(sc->sa.agf_bp);
719 	agflcount = be32_to_cpu(agf->agf_flcount);
720 	if (agflcount > xfs_agfl_size(sc->mp)) {
721 		xfs_scrub_block_set_corrupt(sc, sc->sa.agf_bp);
722 		goto out;
723 	}
724 	memset(&sai, 0, sizeof(sai));
725 	sai.sc = sc;
726 	sai.sz_entries = agflcount;
727 	sai.entries = kmem_zalloc(sizeof(xfs_agblock_t) * agflcount,
728 			KM_MAYFAIL);
729 	if (!sai.entries) {
730 		error = -ENOMEM;
731 		goto out;
732 	}
733 
734 	/* Check the blocks in the AGFL. */
735 	xfs_rmap_ag_owner(&sai.oinfo, XFS_RMAP_OWN_AG);
736 	error = xfs_agfl_walk(sc->mp, XFS_BUF_TO_AGF(sc->sa.agf_bp),
737 			sc->sa.agfl_bp, xfs_scrub_agfl_block, &sai);
738 	if (error == XFS_BTREE_QUERY_RANGE_ABORT) {
739 		error = 0;
740 		goto out_free;
741 	}
742 	if (error)
743 		goto out_free;
744 
745 	if (agflcount != sai.nr_entries) {
746 		xfs_scrub_block_set_corrupt(sc, sc->sa.agf_bp);
747 		goto out_free;
748 	}
749 
750 	/* Sort entries, check for duplicates. */
751 	sort(sai.entries, sai.nr_entries, sizeof(sai.entries[0]),
752 			xfs_scrub_agblock_cmp, NULL);
753 	for (i = 1; i < sai.nr_entries; i++) {
754 		if (sai.entries[i] == sai.entries[i - 1]) {
755 			xfs_scrub_block_set_corrupt(sc, sc->sa.agf_bp);
756 			break;
757 		}
758 	}
759 
760 out_free:
761 	kmem_free(sai.entries);
762 out:
763 	return error;
764 }
765 
766 /* AGI */
767 
768 /* Check agi_count/agi_freecount */
769 static inline void
770 xfs_scrub_agi_xref_icounts(
771 	struct xfs_scrub_context	*sc)
772 {
773 	struct xfs_agi			*agi = XFS_BUF_TO_AGI(sc->sa.agi_bp);
774 	xfs_agino_t			icount;
775 	xfs_agino_t			freecount;
776 	int				error;
777 
778 	if (!sc->sa.ino_cur)
779 		return;
780 
781 	error = xfs_ialloc_count_inodes(sc->sa.ino_cur, &icount, &freecount);
782 	if (!xfs_scrub_should_check_xref(sc, &error, &sc->sa.ino_cur))
783 		return;
784 	if (be32_to_cpu(agi->agi_count) != icount ||
785 	    be32_to_cpu(agi->agi_freecount) != freecount)
786 		xfs_scrub_block_xref_set_corrupt(sc, sc->sa.agi_bp);
787 }
788 
789 /* Cross-reference with the other btrees. */
790 STATIC void
791 xfs_scrub_agi_xref(
792 	struct xfs_scrub_context	*sc)
793 {
794 	struct xfs_owner_info		oinfo;
795 	struct xfs_mount		*mp = sc->mp;
796 	xfs_agblock_t			agbno;
797 	int				error;
798 
799 	if (sc->sm->sm_flags & XFS_SCRUB_OFLAG_CORRUPT)
800 		return;
801 
802 	agbno = XFS_AGI_BLOCK(mp);
803 
804 	error = xfs_scrub_ag_btcur_init(sc, &sc->sa);
805 	if (error)
806 		return;
807 
808 	xfs_scrub_xref_is_used_space(sc, agbno, 1);
809 	xfs_scrub_xref_is_not_inode_chunk(sc, agbno, 1);
810 	xfs_scrub_agi_xref_icounts(sc);
811 	xfs_rmap_ag_owner(&oinfo, XFS_RMAP_OWN_FS);
812 	xfs_scrub_xref_is_owned_by(sc, agbno, 1, &oinfo);
813 	xfs_scrub_xref_is_not_shared(sc, agbno, 1);
814 
815 	/* scrub teardown will take care of sc->sa for us */
816 }
817 
818 /* Scrub the AGI. */
819 int
820 xfs_scrub_agi(
821 	struct xfs_scrub_context	*sc)
822 {
823 	struct xfs_mount		*mp = sc->mp;
824 	struct xfs_agi			*agi;
825 	xfs_agnumber_t			agno;
826 	xfs_agblock_t			agbno;
827 	xfs_agblock_t			eoag;
828 	xfs_agino_t			agino;
829 	xfs_agino_t			first_agino;
830 	xfs_agino_t			last_agino;
831 	xfs_agino_t			icount;
832 	int				i;
833 	int				level;
834 	int				error = 0;
835 
836 	agno = sc->sa.agno = sc->sm->sm_agno;
837 	error = xfs_scrub_ag_read_headers(sc, agno, &sc->sa.agi_bp,
838 			&sc->sa.agf_bp, &sc->sa.agfl_bp);
839 	if (!xfs_scrub_process_error(sc, agno, XFS_AGI_BLOCK(sc->mp), &error))
840 		goto out;
841 	xfs_scrub_buffer_recheck(sc, sc->sa.agi_bp);
842 
843 	agi = XFS_BUF_TO_AGI(sc->sa.agi_bp);
844 
845 	/* Check the AG length */
846 	eoag = be32_to_cpu(agi->agi_length);
847 	if (eoag != xfs_ag_block_count(mp, agno))
848 		xfs_scrub_block_set_corrupt(sc, sc->sa.agi_bp);
849 
850 	/* Check btree roots and levels */
851 	agbno = be32_to_cpu(agi->agi_root);
852 	if (!xfs_verify_agbno(mp, agno, agbno))
853 		xfs_scrub_block_set_corrupt(sc, sc->sa.agi_bp);
854 
855 	level = be32_to_cpu(agi->agi_level);
856 	if (level <= 0 || level > XFS_BTREE_MAXLEVELS)
857 		xfs_scrub_block_set_corrupt(sc, sc->sa.agi_bp);
858 
859 	if (xfs_sb_version_hasfinobt(&mp->m_sb)) {
860 		agbno = be32_to_cpu(agi->agi_free_root);
861 		if (!xfs_verify_agbno(mp, agno, agbno))
862 			xfs_scrub_block_set_corrupt(sc, sc->sa.agi_bp);
863 
864 		level = be32_to_cpu(agi->agi_free_level);
865 		if (level <= 0 || level > XFS_BTREE_MAXLEVELS)
866 			xfs_scrub_block_set_corrupt(sc, sc->sa.agi_bp);
867 	}
868 
869 	/* Check inode counters */
870 	xfs_agino_range(mp, agno, &first_agino, &last_agino);
871 	icount = be32_to_cpu(agi->agi_count);
872 	if (icount > last_agino - first_agino + 1 ||
873 	    icount < be32_to_cpu(agi->agi_freecount))
874 		xfs_scrub_block_set_corrupt(sc, sc->sa.agi_bp);
875 
876 	/* Check inode pointers */
877 	agino = be32_to_cpu(agi->agi_newino);
878 	if (agino != NULLAGINO && !xfs_verify_agino(mp, agno, agino))
879 		xfs_scrub_block_set_corrupt(sc, sc->sa.agi_bp);
880 
881 	agino = be32_to_cpu(agi->agi_dirino);
882 	if (agino != NULLAGINO && !xfs_verify_agino(mp, agno, agino))
883 		xfs_scrub_block_set_corrupt(sc, sc->sa.agi_bp);
884 
885 	/* Check unlinked inode buckets */
886 	for (i = 0; i < XFS_AGI_UNLINKED_BUCKETS; i++) {
887 		agino = be32_to_cpu(agi->agi_unlinked[i]);
888 		if (agino == NULLAGINO)
889 			continue;
890 		if (!xfs_verify_agino(mp, agno, agino))
891 			xfs_scrub_block_set_corrupt(sc, sc->sa.agi_bp);
892 	}
893 
894 	if (agi->agi_pad32 != cpu_to_be32(0))
895 		xfs_scrub_block_set_corrupt(sc, sc->sa.agi_bp);
896 
897 	xfs_scrub_agi_xref(sc);
898 out:
899 	return error;
900 }
901