1 // SPDX-License-Identifier: GPL-2.0 2 /* 3 * Copyright (c) 2000-2006 Silicon Graphics, Inc. 4 * All Rights Reserved. 5 */ 6 #include "xfs.h" 7 #include "xfs_fs.h" 8 #include "xfs_shared.h" 9 #include "xfs_format.h" 10 #include "xfs_log_format.h" 11 #include "xfs_trans_resv.h" 12 #include "xfs_mount.h" 13 #include "xfs_inode.h" 14 #include "xfs_errortag.h" 15 #include "xfs_error.h" 16 #include "xfs_icache.h" 17 #include "xfs_trans.h" 18 #include "xfs_ialloc.h" 19 #include "xfs_dir2.h" 20 21 #include <linux/iversion.h> 22 23 /* 24 * Check that none of the inode's in the buffer have a next 25 * unlinked field of 0. 26 */ 27 #if defined(DEBUG) 28 void 29 xfs_inobp_check( 30 xfs_mount_t *mp, 31 xfs_buf_t *bp) 32 { 33 int i; 34 xfs_dinode_t *dip; 35 36 for (i = 0; i < M_IGEO(mp)->inodes_per_cluster; i++) { 37 dip = xfs_buf_offset(bp, i * mp->m_sb.sb_inodesize); 38 if (!dip->di_next_unlinked) { 39 xfs_alert(mp, 40 "Detected bogus zero next_unlinked field in inode %d buffer 0x%llx.", 41 i, (long long)bp->b_bn); 42 } 43 } 44 } 45 #endif 46 47 bool 48 xfs_dinode_good_version( 49 struct xfs_mount *mp, 50 __u8 version) 51 { 52 if (xfs_sb_version_hascrc(&mp->m_sb)) 53 return version == 3; 54 55 return version == 1 || version == 2; 56 } 57 58 /* 59 * If we are doing readahead on an inode buffer, we might be in log recovery 60 * reading an inode allocation buffer that hasn't yet been replayed, and hence 61 * has not had the inode cores stamped into it. Hence for readahead, the buffer 62 * may be potentially invalid. 63 * 64 * If the readahead buffer is invalid, we need to mark it with an error and 65 * clear the DONE status of the buffer so that a followup read will re-read it 66 * from disk. We don't report the error otherwise to avoid warnings during log 67 * recovery and we don't get unnecssary panics on debug kernels. We use EIO here 68 * because all we want to do is say readahead failed; there is no-one to report 69 * the error to, so this will distinguish it from a non-ra verifier failure. 70 * Changes to this readahead error behavour also need to be reflected in 71 * xfs_dquot_buf_readahead_verify(). 72 */ 73 static void 74 xfs_inode_buf_verify( 75 struct xfs_buf *bp, 76 bool readahead) 77 { 78 struct xfs_mount *mp = bp->b_mount; 79 xfs_agnumber_t agno; 80 int i; 81 int ni; 82 83 /* 84 * Validate the magic number and version of every inode in the buffer 85 */ 86 agno = xfs_daddr_to_agno(mp, XFS_BUF_ADDR(bp)); 87 ni = XFS_BB_TO_FSB(mp, bp->b_length) * mp->m_sb.sb_inopblock; 88 for (i = 0; i < ni; i++) { 89 int di_ok; 90 xfs_dinode_t *dip; 91 xfs_agino_t unlinked_ino; 92 93 dip = xfs_buf_offset(bp, (i << mp->m_sb.sb_inodelog)); 94 unlinked_ino = be32_to_cpu(dip->di_next_unlinked); 95 di_ok = xfs_verify_magic16(bp, dip->di_magic) && 96 xfs_dinode_good_version(mp, dip->di_version) && 97 xfs_verify_agino_or_null(mp, agno, unlinked_ino); 98 if (unlikely(XFS_TEST_ERROR(!di_ok, mp, 99 XFS_ERRTAG_ITOBP_INOTOBP))) { 100 if (readahead) { 101 bp->b_flags &= ~XBF_DONE; 102 xfs_buf_ioerror(bp, -EIO); 103 return; 104 } 105 106 #ifdef DEBUG 107 xfs_alert(mp, 108 "bad inode magic/vsn daddr %lld #%d (magic=%x)", 109 (unsigned long long)bp->b_bn, i, 110 be16_to_cpu(dip->di_magic)); 111 #endif 112 xfs_buf_verifier_error(bp, -EFSCORRUPTED, 113 __func__, dip, sizeof(*dip), 114 NULL); 115 return; 116 } 117 } 118 } 119 120 121 static void 122 xfs_inode_buf_read_verify( 123 struct xfs_buf *bp) 124 { 125 xfs_inode_buf_verify(bp, false); 126 } 127 128 static void 129 xfs_inode_buf_readahead_verify( 130 struct xfs_buf *bp) 131 { 132 xfs_inode_buf_verify(bp, true); 133 } 134 135 static void 136 xfs_inode_buf_write_verify( 137 struct xfs_buf *bp) 138 { 139 xfs_inode_buf_verify(bp, false); 140 } 141 142 const struct xfs_buf_ops xfs_inode_buf_ops = { 143 .name = "xfs_inode", 144 .magic16 = { cpu_to_be16(XFS_DINODE_MAGIC), 145 cpu_to_be16(XFS_DINODE_MAGIC) }, 146 .verify_read = xfs_inode_buf_read_verify, 147 .verify_write = xfs_inode_buf_write_verify, 148 }; 149 150 const struct xfs_buf_ops xfs_inode_buf_ra_ops = { 151 .name = "xfs_inode_ra", 152 .magic16 = { cpu_to_be16(XFS_DINODE_MAGIC), 153 cpu_to_be16(XFS_DINODE_MAGIC) }, 154 .verify_read = xfs_inode_buf_readahead_verify, 155 .verify_write = xfs_inode_buf_write_verify, 156 }; 157 158 159 /* 160 * This routine is called to map an inode to the buffer containing the on-disk 161 * version of the inode. It returns a pointer to the buffer containing the 162 * on-disk inode in the bpp parameter, and in the dipp parameter it returns a 163 * pointer to the on-disk inode within that buffer. 164 * 165 * If a non-zero error is returned, then the contents of bpp and dipp are 166 * undefined. 167 */ 168 int 169 xfs_imap_to_bp( 170 struct xfs_mount *mp, 171 struct xfs_trans *tp, 172 struct xfs_imap *imap, 173 struct xfs_dinode **dipp, 174 struct xfs_buf **bpp, 175 uint buf_flags, 176 uint iget_flags) 177 { 178 struct xfs_buf *bp; 179 int error; 180 181 buf_flags |= XBF_UNMAPPED; 182 error = xfs_trans_read_buf(mp, tp, mp->m_ddev_targp, imap->im_blkno, 183 (int)imap->im_len, buf_flags, &bp, 184 &xfs_inode_buf_ops); 185 if (error) { 186 if (error == -EAGAIN) { 187 ASSERT(buf_flags & XBF_TRYLOCK); 188 return error; 189 } 190 xfs_warn(mp, "%s: xfs_trans_read_buf() returned error %d.", 191 __func__, error); 192 return error; 193 } 194 195 *bpp = bp; 196 *dipp = xfs_buf_offset(bp, imap->im_boffset); 197 return 0; 198 } 199 200 void 201 xfs_inode_from_disk( 202 struct xfs_inode *ip, 203 struct xfs_dinode *from) 204 { 205 struct xfs_icdinode *to = &ip->i_d; 206 struct inode *inode = VFS_I(ip); 207 208 209 /* 210 * Convert v1 inodes immediately to v2 inode format as this is the 211 * minimum inode version format we support in the rest of the code. 212 */ 213 to->di_version = from->di_version; 214 if (to->di_version == 1) { 215 set_nlink(inode, be16_to_cpu(from->di_onlink)); 216 to->di_projid = 0; 217 to->di_version = 2; 218 } else { 219 set_nlink(inode, be32_to_cpu(from->di_nlink)); 220 to->di_projid = (prid_t)be16_to_cpu(from->di_projid_hi) << 16 | 221 be16_to_cpu(from->di_projid_lo); 222 } 223 224 to->di_format = from->di_format; 225 to->di_uid = be32_to_cpu(from->di_uid); 226 to->di_gid = be32_to_cpu(from->di_gid); 227 to->di_flushiter = be16_to_cpu(from->di_flushiter); 228 229 /* 230 * Time is signed, so need to convert to signed 32 bit before 231 * storing in inode timestamp which may be 64 bit. Otherwise 232 * a time before epoch is converted to a time long after epoch 233 * on 64 bit systems. 234 */ 235 inode->i_atime.tv_sec = (int)be32_to_cpu(from->di_atime.t_sec); 236 inode->i_atime.tv_nsec = (int)be32_to_cpu(from->di_atime.t_nsec); 237 inode->i_mtime.tv_sec = (int)be32_to_cpu(from->di_mtime.t_sec); 238 inode->i_mtime.tv_nsec = (int)be32_to_cpu(from->di_mtime.t_nsec); 239 inode->i_ctime.tv_sec = (int)be32_to_cpu(from->di_ctime.t_sec); 240 inode->i_ctime.tv_nsec = (int)be32_to_cpu(from->di_ctime.t_nsec); 241 inode->i_generation = be32_to_cpu(from->di_gen); 242 inode->i_mode = be16_to_cpu(from->di_mode); 243 244 to->di_size = be64_to_cpu(from->di_size); 245 to->di_nblocks = be64_to_cpu(from->di_nblocks); 246 to->di_extsize = be32_to_cpu(from->di_extsize); 247 to->di_nextents = be32_to_cpu(from->di_nextents); 248 to->di_anextents = be16_to_cpu(from->di_anextents); 249 to->di_forkoff = from->di_forkoff; 250 to->di_aformat = from->di_aformat; 251 to->di_dmevmask = be32_to_cpu(from->di_dmevmask); 252 to->di_dmstate = be16_to_cpu(from->di_dmstate); 253 to->di_flags = be16_to_cpu(from->di_flags); 254 255 if (to->di_version == 3) { 256 inode_set_iversion_queried(inode, 257 be64_to_cpu(from->di_changecount)); 258 to->di_crtime.tv_sec = be32_to_cpu(from->di_crtime.t_sec); 259 to->di_crtime.tv_nsec = be32_to_cpu(from->di_crtime.t_nsec); 260 to->di_flags2 = be64_to_cpu(from->di_flags2); 261 to->di_cowextsize = be32_to_cpu(from->di_cowextsize); 262 } 263 } 264 265 void 266 xfs_inode_to_disk( 267 struct xfs_inode *ip, 268 struct xfs_dinode *to, 269 xfs_lsn_t lsn) 270 { 271 struct xfs_icdinode *from = &ip->i_d; 272 struct inode *inode = VFS_I(ip); 273 274 to->di_magic = cpu_to_be16(XFS_DINODE_MAGIC); 275 to->di_onlink = 0; 276 277 to->di_version = from->di_version; 278 to->di_format = from->di_format; 279 to->di_uid = cpu_to_be32(from->di_uid); 280 to->di_gid = cpu_to_be32(from->di_gid); 281 to->di_projid_lo = cpu_to_be16(from->di_projid & 0xffff); 282 to->di_projid_hi = cpu_to_be16(from->di_projid >> 16); 283 284 memset(to->di_pad, 0, sizeof(to->di_pad)); 285 to->di_atime.t_sec = cpu_to_be32(inode->i_atime.tv_sec); 286 to->di_atime.t_nsec = cpu_to_be32(inode->i_atime.tv_nsec); 287 to->di_mtime.t_sec = cpu_to_be32(inode->i_mtime.tv_sec); 288 to->di_mtime.t_nsec = cpu_to_be32(inode->i_mtime.tv_nsec); 289 to->di_ctime.t_sec = cpu_to_be32(inode->i_ctime.tv_sec); 290 to->di_ctime.t_nsec = cpu_to_be32(inode->i_ctime.tv_nsec); 291 to->di_nlink = cpu_to_be32(inode->i_nlink); 292 to->di_gen = cpu_to_be32(inode->i_generation); 293 to->di_mode = cpu_to_be16(inode->i_mode); 294 295 to->di_size = cpu_to_be64(from->di_size); 296 to->di_nblocks = cpu_to_be64(from->di_nblocks); 297 to->di_extsize = cpu_to_be32(from->di_extsize); 298 to->di_nextents = cpu_to_be32(from->di_nextents); 299 to->di_anextents = cpu_to_be16(from->di_anextents); 300 to->di_forkoff = from->di_forkoff; 301 to->di_aformat = from->di_aformat; 302 to->di_dmevmask = cpu_to_be32(from->di_dmevmask); 303 to->di_dmstate = cpu_to_be16(from->di_dmstate); 304 to->di_flags = cpu_to_be16(from->di_flags); 305 306 if (from->di_version == 3) { 307 to->di_changecount = cpu_to_be64(inode_peek_iversion(inode)); 308 to->di_crtime.t_sec = cpu_to_be32(from->di_crtime.tv_sec); 309 to->di_crtime.t_nsec = cpu_to_be32(from->di_crtime.tv_nsec); 310 to->di_flags2 = cpu_to_be64(from->di_flags2); 311 to->di_cowextsize = cpu_to_be32(from->di_cowextsize); 312 to->di_ino = cpu_to_be64(ip->i_ino); 313 to->di_lsn = cpu_to_be64(lsn); 314 memset(to->di_pad2, 0, sizeof(to->di_pad2)); 315 uuid_copy(&to->di_uuid, &ip->i_mount->m_sb.sb_meta_uuid); 316 to->di_flushiter = 0; 317 } else { 318 to->di_flushiter = cpu_to_be16(from->di_flushiter); 319 } 320 } 321 322 void 323 xfs_log_dinode_to_disk( 324 struct xfs_log_dinode *from, 325 struct xfs_dinode *to) 326 { 327 to->di_magic = cpu_to_be16(from->di_magic); 328 to->di_mode = cpu_to_be16(from->di_mode); 329 to->di_version = from->di_version; 330 to->di_format = from->di_format; 331 to->di_onlink = 0; 332 to->di_uid = cpu_to_be32(from->di_uid); 333 to->di_gid = cpu_to_be32(from->di_gid); 334 to->di_nlink = cpu_to_be32(from->di_nlink); 335 to->di_projid_lo = cpu_to_be16(from->di_projid_lo); 336 to->di_projid_hi = cpu_to_be16(from->di_projid_hi); 337 memcpy(to->di_pad, from->di_pad, sizeof(to->di_pad)); 338 339 to->di_atime.t_sec = cpu_to_be32(from->di_atime.t_sec); 340 to->di_atime.t_nsec = cpu_to_be32(from->di_atime.t_nsec); 341 to->di_mtime.t_sec = cpu_to_be32(from->di_mtime.t_sec); 342 to->di_mtime.t_nsec = cpu_to_be32(from->di_mtime.t_nsec); 343 to->di_ctime.t_sec = cpu_to_be32(from->di_ctime.t_sec); 344 to->di_ctime.t_nsec = cpu_to_be32(from->di_ctime.t_nsec); 345 346 to->di_size = cpu_to_be64(from->di_size); 347 to->di_nblocks = cpu_to_be64(from->di_nblocks); 348 to->di_extsize = cpu_to_be32(from->di_extsize); 349 to->di_nextents = cpu_to_be32(from->di_nextents); 350 to->di_anextents = cpu_to_be16(from->di_anextents); 351 to->di_forkoff = from->di_forkoff; 352 to->di_aformat = from->di_aformat; 353 to->di_dmevmask = cpu_to_be32(from->di_dmevmask); 354 to->di_dmstate = cpu_to_be16(from->di_dmstate); 355 to->di_flags = cpu_to_be16(from->di_flags); 356 to->di_gen = cpu_to_be32(from->di_gen); 357 358 if (from->di_version == 3) { 359 to->di_changecount = cpu_to_be64(from->di_changecount); 360 to->di_crtime.t_sec = cpu_to_be32(from->di_crtime.t_sec); 361 to->di_crtime.t_nsec = cpu_to_be32(from->di_crtime.t_nsec); 362 to->di_flags2 = cpu_to_be64(from->di_flags2); 363 to->di_cowextsize = cpu_to_be32(from->di_cowextsize); 364 to->di_ino = cpu_to_be64(from->di_ino); 365 to->di_lsn = cpu_to_be64(from->di_lsn); 366 memcpy(to->di_pad2, from->di_pad2, sizeof(to->di_pad2)); 367 uuid_copy(&to->di_uuid, &from->di_uuid); 368 to->di_flushiter = 0; 369 } else { 370 to->di_flushiter = cpu_to_be16(from->di_flushiter); 371 } 372 } 373 374 static xfs_failaddr_t 375 xfs_dinode_verify_fork( 376 struct xfs_dinode *dip, 377 struct xfs_mount *mp, 378 int whichfork) 379 { 380 uint32_t di_nextents = XFS_DFORK_NEXTENTS(dip, whichfork); 381 382 switch (XFS_DFORK_FORMAT(dip, whichfork)) { 383 case XFS_DINODE_FMT_LOCAL: 384 /* 385 * no local regular files yet 386 */ 387 if (whichfork == XFS_DATA_FORK) { 388 if (S_ISREG(be16_to_cpu(dip->di_mode))) 389 return __this_address; 390 if (be64_to_cpu(dip->di_size) > 391 XFS_DFORK_SIZE(dip, mp, whichfork)) 392 return __this_address; 393 } 394 if (di_nextents) 395 return __this_address; 396 break; 397 case XFS_DINODE_FMT_EXTENTS: 398 if (di_nextents > XFS_DFORK_MAXEXT(dip, mp, whichfork)) 399 return __this_address; 400 break; 401 case XFS_DINODE_FMT_BTREE: 402 if (whichfork == XFS_ATTR_FORK) { 403 if (di_nextents > MAXAEXTNUM) 404 return __this_address; 405 } else if (di_nextents > MAXEXTNUM) { 406 return __this_address; 407 } 408 break; 409 default: 410 return __this_address; 411 } 412 return NULL; 413 } 414 415 static xfs_failaddr_t 416 xfs_dinode_verify_forkoff( 417 struct xfs_dinode *dip, 418 struct xfs_mount *mp) 419 { 420 if (!XFS_DFORK_Q(dip)) 421 return NULL; 422 423 switch (dip->di_format) { 424 case XFS_DINODE_FMT_DEV: 425 if (dip->di_forkoff != (roundup(sizeof(xfs_dev_t), 8) >> 3)) 426 return __this_address; 427 break; 428 case XFS_DINODE_FMT_LOCAL: /* fall through ... */ 429 case XFS_DINODE_FMT_EXTENTS: /* fall through ... */ 430 case XFS_DINODE_FMT_BTREE: 431 if (dip->di_forkoff >= (XFS_LITINO(mp, dip->di_version) >> 3)) 432 return __this_address; 433 break; 434 default: 435 return __this_address; 436 } 437 return NULL; 438 } 439 440 xfs_failaddr_t 441 xfs_dinode_verify( 442 struct xfs_mount *mp, 443 xfs_ino_t ino, 444 struct xfs_dinode *dip) 445 { 446 xfs_failaddr_t fa; 447 uint16_t mode; 448 uint16_t flags; 449 uint64_t flags2; 450 uint64_t di_size; 451 452 if (dip->di_magic != cpu_to_be16(XFS_DINODE_MAGIC)) 453 return __this_address; 454 455 /* Verify v3 integrity information first */ 456 if (dip->di_version >= 3) { 457 if (!xfs_sb_version_hascrc(&mp->m_sb)) 458 return __this_address; 459 if (!xfs_verify_cksum((char *)dip, mp->m_sb.sb_inodesize, 460 XFS_DINODE_CRC_OFF)) 461 return __this_address; 462 if (be64_to_cpu(dip->di_ino) != ino) 463 return __this_address; 464 if (!uuid_equal(&dip->di_uuid, &mp->m_sb.sb_meta_uuid)) 465 return __this_address; 466 } 467 468 /* don't allow invalid i_size */ 469 di_size = be64_to_cpu(dip->di_size); 470 if (di_size & (1ULL << 63)) 471 return __this_address; 472 473 mode = be16_to_cpu(dip->di_mode); 474 if (mode && xfs_mode_to_ftype(mode) == XFS_DIR3_FT_UNKNOWN) 475 return __this_address; 476 477 /* No zero-length symlinks/dirs. */ 478 if ((S_ISLNK(mode) || S_ISDIR(mode)) && di_size == 0) 479 return __this_address; 480 481 /* Fork checks carried over from xfs_iformat_fork */ 482 if (mode && 483 be32_to_cpu(dip->di_nextents) + be16_to_cpu(dip->di_anextents) > 484 be64_to_cpu(dip->di_nblocks)) 485 return __this_address; 486 487 if (mode && XFS_DFORK_BOFF(dip) > mp->m_sb.sb_inodesize) 488 return __this_address; 489 490 flags = be16_to_cpu(dip->di_flags); 491 492 if (mode && (flags & XFS_DIFLAG_REALTIME) && !mp->m_rtdev_targp) 493 return __this_address; 494 495 /* check for illegal values of forkoff */ 496 fa = xfs_dinode_verify_forkoff(dip, mp); 497 if (fa) 498 return fa; 499 500 /* Do we have appropriate data fork formats for the mode? */ 501 switch (mode & S_IFMT) { 502 case S_IFIFO: 503 case S_IFCHR: 504 case S_IFBLK: 505 case S_IFSOCK: 506 if (dip->di_format != XFS_DINODE_FMT_DEV) 507 return __this_address; 508 break; 509 case S_IFREG: 510 case S_IFLNK: 511 case S_IFDIR: 512 fa = xfs_dinode_verify_fork(dip, mp, XFS_DATA_FORK); 513 if (fa) 514 return fa; 515 break; 516 case 0: 517 /* Uninitialized inode ok. */ 518 break; 519 default: 520 return __this_address; 521 } 522 523 if (XFS_DFORK_Q(dip)) { 524 fa = xfs_dinode_verify_fork(dip, mp, XFS_ATTR_FORK); 525 if (fa) 526 return fa; 527 } else { 528 /* 529 * If there is no fork offset, this may be a freshly-made inode 530 * in a new disk cluster, in which case di_aformat is zeroed. 531 * Otherwise, such an inode must be in EXTENTS format; this goes 532 * for freed inodes as well. 533 */ 534 switch (dip->di_aformat) { 535 case 0: 536 case XFS_DINODE_FMT_EXTENTS: 537 break; 538 default: 539 return __this_address; 540 } 541 if (dip->di_anextents) 542 return __this_address; 543 } 544 545 /* extent size hint validation */ 546 fa = xfs_inode_validate_extsize(mp, be32_to_cpu(dip->di_extsize), 547 mode, flags); 548 if (fa) 549 return fa; 550 551 /* only version 3 or greater inodes are extensively verified here */ 552 if (dip->di_version < 3) 553 return NULL; 554 555 flags2 = be64_to_cpu(dip->di_flags2); 556 557 /* don't allow reflink/cowextsize if we don't have reflink */ 558 if ((flags2 & (XFS_DIFLAG2_REFLINK | XFS_DIFLAG2_COWEXTSIZE)) && 559 !xfs_sb_version_hasreflink(&mp->m_sb)) 560 return __this_address; 561 562 /* only regular files get reflink */ 563 if ((flags2 & XFS_DIFLAG2_REFLINK) && (mode & S_IFMT) != S_IFREG) 564 return __this_address; 565 566 /* don't let reflink and realtime mix */ 567 if ((flags2 & XFS_DIFLAG2_REFLINK) && (flags & XFS_DIFLAG_REALTIME)) 568 return __this_address; 569 570 /* don't let reflink and dax mix */ 571 if ((flags2 & XFS_DIFLAG2_REFLINK) && (flags2 & XFS_DIFLAG2_DAX)) 572 return __this_address; 573 574 /* COW extent size hint validation */ 575 fa = xfs_inode_validate_cowextsize(mp, be32_to_cpu(dip->di_cowextsize), 576 mode, flags, flags2); 577 if (fa) 578 return fa; 579 580 return NULL; 581 } 582 583 void 584 xfs_dinode_calc_crc( 585 struct xfs_mount *mp, 586 struct xfs_dinode *dip) 587 { 588 uint32_t crc; 589 590 if (dip->di_version < 3) 591 return; 592 593 ASSERT(xfs_sb_version_hascrc(&mp->m_sb)); 594 crc = xfs_start_cksum_update((char *)dip, mp->m_sb.sb_inodesize, 595 XFS_DINODE_CRC_OFF); 596 dip->di_crc = xfs_end_cksum(crc); 597 } 598 599 /* 600 * Read the disk inode attributes into the in-core inode structure. 601 * 602 * For version 5 superblocks, if we are initialising a new inode and we are not 603 * utilising the XFS_MOUNT_IKEEP inode cluster mode, we can simple build the new 604 * inode core with a random generation number. If we are keeping inodes around, 605 * we need to read the inode cluster to get the existing generation number off 606 * disk. Further, if we are using version 4 superblocks (i.e. v1/v2 inode 607 * format) then log recovery is dependent on the di_flushiter field being 608 * initialised from the current on-disk value and hence we must also read the 609 * inode off disk. 610 */ 611 int 612 xfs_iread( 613 xfs_mount_t *mp, 614 xfs_trans_t *tp, 615 xfs_inode_t *ip, 616 uint iget_flags) 617 { 618 xfs_buf_t *bp; 619 xfs_dinode_t *dip; 620 xfs_failaddr_t fa; 621 int error; 622 623 /* 624 * Fill in the location information in the in-core inode. 625 */ 626 error = xfs_imap(mp, tp, ip->i_ino, &ip->i_imap, iget_flags); 627 if (error) 628 return error; 629 630 /* shortcut IO on inode allocation if possible */ 631 if ((iget_flags & XFS_IGET_CREATE) && 632 xfs_sb_version_hascrc(&mp->m_sb) && 633 !(mp->m_flags & XFS_MOUNT_IKEEP)) { 634 VFS_I(ip)->i_generation = prandom_u32(); 635 ip->i_d.di_version = 3; 636 return 0; 637 } 638 639 /* 640 * Get pointers to the on-disk inode and the buffer containing it. 641 */ 642 error = xfs_imap_to_bp(mp, tp, &ip->i_imap, &dip, &bp, 0, iget_flags); 643 if (error) 644 return error; 645 646 /* even unallocated inodes are verified */ 647 fa = xfs_dinode_verify(mp, ip->i_ino, dip); 648 if (fa) { 649 xfs_inode_verifier_error(ip, -EFSCORRUPTED, "dinode", dip, 650 sizeof(*dip), fa); 651 error = -EFSCORRUPTED; 652 goto out_brelse; 653 } 654 655 /* 656 * If the on-disk inode is already linked to a directory 657 * entry, copy all of the inode into the in-core inode. 658 * xfs_iformat_fork() handles copying in the inode format 659 * specific information. 660 * Otherwise, just get the truly permanent information. 661 */ 662 if (dip->di_mode) { 663 xfs_inode_from_disk(ip, dip); 664 error = xfs_iformat_fork(ip, dip); 665 if (error) { 666 #ifdef DEBUG 667 xfs_alert(mp, "%s: xfs_iformat() returned error %d", 668 __func__, error); 669 #endif /* DEBUG */ 670 goto out_brelse; 671 } 672 } else { 673 /* 674 * Partial initialisation of the in-core inode. Just the bits 675 * that xfs_ialloc won't overwrite or relies on being correct. 676 */ 677 ip->i_d.di_version = dip->di_version; 678 VFS_I(ip)->i_generation = be32_to_cpu(dip->di_gen); 679 ip->i_d.di_flushiter = be16_to_cpu(dip->di_flushiter); 680 681 /* 682 * Make sure to pull in the mode here as well in 683 * case the inode is released without being used. 684 * This ensures that xfs_inactive() will see that 685 * the inode is already free and not try to mess 686 * with the uninitialized part of it. 687 */ 688 VFS_I(ip)->i_mode = 0; 689 } 690 691 ASSERT(ip->i_d.di_version >= 2); 692 ip->i_delayed_blks = 0; 693 694 /* 695 * Mark the buffer containing the inode as something to keep 696 * around for a while. This helps to keep recently accessed 697 * meta-data in-core longer. 698 */ 699 xfs_buf_set_ref(bp, XFS_INO_REF); 700 701 /* 702 * Use xfs_trans_brelse() to release the buffer containing the on-disk 703 * inode, because it was acquired with xfs_trans_read_buf() in 704 * xfs_imap_to_bp() above. If tp is NULL, this is just a normal 705 * brelse(). If we're within a transaction, then xfs_trans_brelse() 706 * will only release the buffer if it is not dirty within the 707 * transaction. It will be OK to release the buffer in this case, 708 * because inodes on disk are never destroyed and we will be locking the 709 * new in-core inode before putting it in the cache where other 710 * processes can find it. Thus we don't have to worry about the inode 711 * being changed just because we released the buffer. 712 */ 713 out_brelse: 714 xfs_trans_brelse(tp, bp); 715 return error; 716 } 717 718 /* 719 * Validate di_extsize hint. 720 * 721 * The rules are documented at xfs_ioctl_setattr_check_extsize(). 722 * These functions must be kept in sync with each other. 723 */ 724 xfs_failaddr_t 725 xfs_inode_validate_extsize( 726 struct xfs_mount *mp, 727 uint32_t extsize, 728 uint16_t mode, 729 uint16_t flags) 730 { 731 bool rt_flag; 732 bool hint_flag; 733 bool inherit_flag; 734 uint32_t extsize_bytes; 735 uint32_t blocksize_bytes; 736 737 rt_flag = (flags & XFS_DIFLAG_REALTIME); 738 hint_flag = (flags & XFS_DIFLAG_EXTSIZE); 739 inherit_flag = (flags & XFS_DIFLAG_EXTSZINHERIT); 740 extsize_bytes = XFS_FSB_TO_B(mp, extsize); 741 742 if (rt_flag) 743 blocksize_bytes = mp->m_sb.sb_rextsize << mp->m_sb.sb_blocklog; 744 else 745 blocksize_bytes = mp->m_sb.sb_blocksize; 746 747 if ((hint_flag || inherit_flag) && !(S_ISDIR(mode) || S_ISREG(mode))) 748 return __this_address; 749 750 if (hint_flag && !S_ISREG(mode)) 751 return __this_address; 752 753 if (inherit_flag && !S_ISDIR(mode)) 754 return __this_address; 755 756 if ((hint_flag || inherit_flag) && extsize == 0) 757 return __this_address; 758 759 /* free inodes get flags set to zero but extsize remains */ 760 if (mode && !(hint_flag || inherit_flag) && extsize != 0) 761 return __this_address; 762 763 if (extsize_bytes % blocksize_bytes) 764 return __this_address; 765 766 if (extsize > MAXEXTLEN) 767 return __this_address; 768 769 if (!rt_flag && extsize > mp->m_sb.sb_agblocks / 2) 770 return __this_address; 771 772 return NULL; 773 } 774 775 /* 776 * Validate di_cowextsize hint. 777 * 778 * The rules are documented at xfs_ioctl_setattr_check_cowextsize(). 779 * These functions must be kept in sync with each other. 780 */ 781 xfs_failaddr_t 782 xfs_inode_validate_cowextsize( 783 struct xfs_mount *mp, 784 uint32_t cowextsize, 785 uint16_t mode, 786 uint16_t flags, 787 uint64_t flags2) 788 { 789 bool rt_flag; 790 bool hint_flag; 791 uint32_t cowextsize_bytes; 792 793 rt_flag = (flags & XFS_DIFLAG_REALTIME); 794 hint_flag = (flags2 & XFS_DIFLAG2_COWEXTSIZE); 795 cowextsize_bytes = XFS_FSB_TO_B(mp, cowextsize); 796 797 if (hint_flag && !xfs_sb_version_hasreflink(&mp->m_sb)) 798 return __this_address; 799 800 if (hint_flag && !(S_ISDIR(mode) || S_ISREG(mode))) 801 return __this_address; 802 803 if (hint_flag && cowextsize == 0) 804 return __this_address; 805 806 /* free inodes get flags set to zero but cowextsize remains */ 807 if (mode && !hint_flag && cowextsize != 0) 808 return __this_address; 809 810 if (hint_flag && rt_flag) 811 return __this_address; 812 813 if (cowextsize_bytes % mp->m_sb.sb_blocksize) 814 return __this_address; 815 816 if (cowextsize > MAXEXTLEN) 817 return __this_address; 818 819 if (cowextsize > mp->m_sb.sb_agblocks / 2) 820 return __this_address; 821 822 return NULL; 823 } 824