1 /* 2 * Ext4 orphan inode handling 3 */ 4 #include <linux/fs.h> 5 #include <linux/quotaops.h> 6 #include <linux/buffer_head.h> 7 8 #include "ext4.h" 9 #include "ext4_jbd2.h" 10 11 static int ext4_orphan_file_add(handle_t *handle, struct inode *inode) 12 { 13 int i, j; 14 struct ext4_orphan_info *oi = &EXT4_SB(inode->i_sb)->s_orphan_info; 15 int ret = 0; 16 __le32 *bdata; 17 int inodes_per_ob = ext4_inodes_per_orphan_block(inode->i_sb); 18 19 spin_lock(&oi->of_lock); 20 for (i = 0; i < oi->of_blocks && !oi->of_binfo[i].ob_free_entries; i++); 21 if (i == oi->of_blocks) { 22 spin_unlock(&oi->of_lock); 23 /* 24 * For now we don't grow or shrink orphan file. We just use 25 * whatever was allocated at mke2fs time. The additional 26 * credits we would have to reserve for each orphan inode 27 * operation just don't seem worth it. 28 */ 29 return -ENOSPC; 30 } 31 oi->of_binfo[i].ob_free_entries--; 32 spin_unlock(&oi->of_lock); 33 34 /* 35 * Get access to orphan block. We have dropped of_lock but since we 36 * have decremented number of free entries we are guaranteed free entry 37 * in our block. 38 */ 39 ret = ext4_journal_get_write_access(handle, inode->i_sb, 40 oi->of_binfo[i].ob_bh, EXT4_JTR_ORPHAN_FILE); 41 if (ret) 42 return ret; 43 44 bdata = (__le32 *)(oi->of_binfo[i].ob_bh->b_data); 45 spin_lock(&oi->of_lock); 46 /* Find empty slot in a block */ 47 for (j = 0; j < inodes_per_ob && bdata[j]; j++); 48 BUG_ON(j == inodes_per_ob); 49 bdata[j] = cpu_to_le32(inode->i_ino); 50 EXT4_I(inode)->i_orphan_idx = i * inodes_per_ob + j; 51 ext4_set_inode_state(inode, EXT4_STATE_ORPHAN_FILE); 52 spin_unlock(&oi->of_lock); 53 54 return ext4_handle_dirty_metadata(handle, NULL, oi->of_binfo[i].ob_bh); 55 } 56 57 /* 58 * ext4_orphan_add() links an unlinked or truncated inode into a list of 59 * such inodes, starting at the superblock, in case we crash before the 60 * file is closed/deleted, or in case the inode truncate spans multiple 61 * transactions and the last transaction is not recovered after a crash. 62 * 63 * At filesystem recovery time, we walk this list deleting unlinked 64 * inodes and truncating linked inodes in ext4_orphan_cleanup(). 65 * 66 * Orphan list manipulation functions must be called under i_mutex unless 67 * we are just creating the inode or deleting it. 68 */ 69 int ext4_orphan_add(handle_t *handle, struct inode *inode) 70 { 71 struct super_block *sb = inode->i_sb; 72 struct ext4_sb_info *sbi = EXT4_SB(sb); 73 struct ext4_iloc iloc; 74 int err = 0, rc; 75 bool dirty = false; 76 77 if (!sbi->s_journal || is_bad_inode(inode)) 78 return 0; 79 80 WARN_ON_ONCE(!(inode->i_state & (I_NEW | I_FREEING)) && 81 !inode_is_locked(inode)); 82 /* 83 * Inode orphaned in orphan file or in orphan list? 84 */ 85 if (ext4_test_inode_state(inode, EXT4_STATE_ORPHAN_FILE) || 86 !list_empty(&EXT4_I(inode)->i_orphan)) 87 return 0; 88 89 /* 90 * Orphan handling is only valid for files with data blocks 91 * being truncated, or files being unlinked. Note that we either 92 * hold i_mutex, or the inode can not be referenced from outside, 93 * so i_nlink should not be bumped due to race 94 */ 95 ASSERT((S_ISREG(inode->i_mode) || S_ISDIR(inode->i_mode) || 96 S_ISLNK(inode->i_mode)) || inode->i_nlink == 0); 97 98 if (sbi->s_orphan_info.of_blocks) { 99 err = ext4_orphan_file_add(handle, inode); 100 /* 101 * Fallback to normal orphan list of orphan file is 102 * out of space 103 */ 104 if (err != -ENOSPC) 105 return err; 106 } 107 108 BUFFER_TRACE(sbi->s_sbh, "get_write_access"); 109 err = ext4_journal_get_write_access(handle, sb, sbi->s_sbh, 110 EXT4_JTR_NONE); 111 if (err) 112 goto out; 113 114 err = ext4_reserve_inode_write(handle, inode, &iloc); 115 if (err) 116 goto out; 117 118 mutex_lock(&sbi->s_orphan_lock); 119 /* 120 * Due to previous errors inode may be already a part of on-disk 121 * orphan list. If so skip on-disk list modification. 122 */ 123 if (!NEXT_ORPHAN(inode) || NEXT_ORPHAN(inode) > 124 (le32_to_cpu(sbi->s_es->s_inodes_count))) { 125 /* Insert this inode at the head of the on-disk orphan list */ 126 NEXT_ORPHAN(inode) = le32_to_cpu(sbi->s_es->s_last_orphan); 127 lock_buffer(sbi->s_sbh); 128 sbi->s_es->s_last_orphan = cpu_to_le32(inode->i_ino); 129 ext4_superblock_csum_set(sb); 130 unlock_buffer(sbi->s_sbh); 131 dirty = true; 132 } 133 list_add(&EXT4_I(inode)->i_orphan, &sbi->s_orphan); 134 mutex_unlock(&sbi->s_orphan_lock); 135 136 if (dirty) { 137 err = ext4_handle_dirty_metadata(handle, NULL, sbi->s_sbh); 138 rc = ext4_mark_iloc_dirty(handle, inode, &iloc); 139 if (!err) 140 err = rc; 141 if (err) { 142 /* 143 * We have to remove inode from in-memory list if 144 * addition to on disk orphan list failed. Stray orphan 145 * list entries can cause panics at unmount time. 146 */ 147 mutex_lock(&sbi->s_orphan_lock); 148 list_del_init(&EXT4_I(inode)->i_orphan); 149 mutex_unlock(&sbi->s_orphan_lock); 150 } 151 } else 152 brelse(iloc.bh); 153 154 jbd_debug(4, "superblock will point to %lu\n", inode->i_ino); 155 jbd_debug(4, "orphan inode %lu will point to %d\n", 156 inode->i_ino, NEXT_ORPHAN(inode)); 157 out: 158 ext4_std_error(sb, err); 159 return err; 160 } 161 162 static int ext4_orphan_file_del(handle_t *handle, struct inode *inode) 163 { 164 struct ext4_orphan_info *oi = &EXT4_SB(inode->i_sb)->s_orphan_info; 165 __le32 *bdata; 166 int blk, off; 167 int inodes_per_ob = ext4_inodes_per_orphan_block(inode->i_sb); 168 int ret = 0; 169 170 if (!handle) 171 goto out; 172 blk = EXT4_I(inode)->i_orphan_idx / inodes_per_ob; 173 off = EXT4_I(inode)->i_orphan_idx % inodes_per_ob; 174 if (WARN_ON_ONCE(blk >= oi->of_blocks)) 175 goto out; 176 177 ret = ext4_journal_get_write_access(handle, inode->i_sb, 178 oi->of_binfo[blk].ob_bh, EXT4_JTR_ORPHAN_FILE); 179 if (ret) 180 goto out; 181 182 bdata = (__le32 *)(oi->of_binfo[blk].ob_bh->b_data); 183 spin_lock(&oi->of_lock); 184 bdata[off] = 0; 185 oi->of_binfo[blk].ob_free_entries++; 186 spin_unlock(&oi->of_lock); 187 ret = ext4_handle_dirty_metadata(handle, NULL, oi->of_binfo[blk].ob_bh); 188 out: 189 ext4_clear_inode_state(inode, EXT4_STATE_ORPHAN_FILE); 190 INIT_LIST_HEAD(&EXT4_I(inode)->i_orphan); 191 192 return ret; 193 } 194 195 /* 196 * ext4_orphan_del() removes an unlinked or truncated inode from the list 197 * of such inodes stored on disk, because it is finally being cleaned up. 198 */ 199 int ext4_orphan_del(handle_t *handle, struct inode *inode) 200 { 201 struct list_head *prev; 202 struct ext4_inode_info *ei = EXT4_I(inode); 203 struct ext4_sb_info *sbi = EXT4_SB(inode->i_sb); 204 __u32 ino_next; 205 struct ext4_iloc iloc; 206 int err = 0; 207 208 if (!sbi->s_journal && !(sbi->s_mount_state & EXT4_ORPHAN_FS)) 209 return 0; 210 211 WARN_ON_ONCE(!(inode->i_state & (I_NEW | I_FREEING)) && 212 !inode_is_locked(inode)); 213 if (ext4_test_inode_state(inode, EXT4_STATE_ORPHAN_FILE)) 214 return ext4_orphan_file_del(handle, inode); 215 216 /* Do this quick check before taking global s_orphan_lock. */ 217 if (list_empty(&ei->i_orphan)) 218 return 0; 219 220 if (handle) { 221 /* Grab inode buffer early before taking global s_orphan_lock */ 222 err = ext4_reserve_inode_write(handle, inode, &iloc); 223 } 224 225 mutex_lock(&sbi->s_orphan_lock); 226 jbd_debug(4, "remove inode %lu from orphan list\n", inode->i_ino); 227 228 prev = ei->i_orphan.prev; 229 list_del_init(&ei->i_orphan); 230 231 /* If we're on an error path, we may not have a valid 232 * transaction handle with which to update the orphan list on 233 * disk, but we still need to remove the inode from the linked 234 * list in memory. */ 235 if (!handle || err) { 236 mutex_unlock(&sbi->s_orphan_lock); 237 goto out_err; 238 } 239 240 ino_next = NEXT_ORPHAN(inode); 241 if (prev == &sbi->s_orphan) { 242 jbd_debug(4, "superblock will point to %u\n", ino_next); 243 BUFFER_TRACE(sbi->s_sbh, "get_write_access"); 244 err = ext4_journal_get_write_access(handle, inode->i_sb, 245 sbi->s_sbh, EXT4_JTR_NONE); 246 if (err) { 247 mutex_unlock(&sbi->s_orphan_lock); 248 goto out_brelse; 249 } 250 lock_buffer(sbi->s_sbh); 251 sbi->s_es->s_last_orphan = cpu_to_le32(ino_next); 252 ext4_superblock_csum_set(inode->i_sb); 253 unlock_buffer(sbi->s_sbh); 254 mutex_unlock(&sbi->s_orphan_lock); 255 err = ext4_handle_dirty_metadata(handle, NULL, sbi->s_sbh); 256 } else { 257 struct ext4_iloc iloc2; 258 struct inode *i_prev = 259 &list_entry(prev, struct ext4_inode_info, i_orphan)->vfs_inode; 260 261 jbd_debug(4, "orphan inode %lu will point to %u\n", 262 i_prev->i_ino, ino_next); 263 err = ext4_reserve_inode_write(handle, i_prev, &iloc2); 264 if (err) { 265 mutex_unlock(&sbi->s_orphan_lock); 266 goto out_brelse; 267 } 268 NEXT_ORPHAN(i_prev) = ino_next; 269 err = ext4_mark_iloc_dirty(handle, i_prev, &iloc2); 270 mutex_unlock(&sbi->s_orphan_lock); 271 } 272 if (err) 273 goto out_brelse; 274 NEXT_ORPHAN(inode) = 0; 275 err = ext4_mark_iloc_dirty(handle, inode, &iloc); 276 out_err: 277 ext4_std_error(inode->i_sb, err); 278 return err; 279 280 out_brelse: 281 brelse(iloc.bh); 282 goto out_err; 283 } 284 285 #ifdef CONFIG_QUOTA 286 static int ext4_quota_on_mount(struct super_block *sb, int type) 287 { 288 return dquot_quota_on_mount(sb, 289 rcu_dereference_protected(EXT4_SB(sb)->s_qf_names[type], 290 lockdep_is_held(&sb->s_umount)), 291 EXT4_SB(sb)->s_jquota_fmt, type); 292 } 293 #endif 294 295 static void ext4_process_orphan(struct inode *inode, 296 int *nr_truncates, int *nr_orphans) 297 { 298 struct super_block *sb = inode->i_sb; 299 int ret; 300 301 dquot_initialize(inode); 302 if (inode->i_nlink) { 303 if (test_opt(sb, DEBUG)) 304 ext4_msg(sb, KERN_DEBUG, 305 "%s: truncating inode %lu to %lld bytes", 306 __func__, inode->i_ino, inode->i_size); 307 jbd_debug(2, "truncating inode %lu to %lld bytes\n", 308 inode->i_ino, inode->i_size); 309 inode_lock(inode); 310 truncate_inode_pages(inode->i_mapping, inode->i_size); 311 ret = ext4_truncate(inode); 312 if (ret) { 313 /* 314 * We need to clean up the in-core orphan list 315 * manually if ext4_truncate() failed to get a 316 * transaction handle. 317 */ 318 ext4_orphan_del(NULL, inode); 319 ext4_std_error(inode->i_sb, ret); 320 } 321 inode_unlock(inode); 322 (*nr_truncates)++; 323 } else { 324 if (test_opt(sb, DEBUG)) 325 ext4_msg(sb, KERN_DEBUG, 326 "%s: deleting unreferenced inode %lu", 327 __func__, inode->i_ino); 328 jbd_debug(2, "deleting unreferenced inode %lu\n", 329 inode->i_ino); 330 (*nr_orphans)++; 331 } 332 iput(inode); /* The delete magic happens here! */ 333 } 334 335 /* ext4_orphan_cleanup() walks a singly-linked list of inodes (starting at 336 * the superblock) which were deleted from all directories, but held open by 337 * a process at the time of a crash. We walk the list and try to delete these 338 * inodes at recovery time (only with a read-write filesystem). 339 * 340 * In order to keep the orphan inode chain consistent during traversal (in 341 * case of crash during recovery), we link each inode into the superblock 342 * orphan list_head and handle it the same way as an inode deletion during 343 * normal operation (which journals the operations for us). 344 * 345 * We only do an iget() and an iput() on each inode, which is very safe if we 346 * accidentally point at an in-use or already deleted inode. The worst that 347 * can happen in this case is that we get a "bit already cleared" message from 348 * ext4_free_inode(). The only reason we would point at a wrong inode is if 349 * e2fsck was run on this filesystem, and it must have already done the orphan 350 * inode cleanup for us, so we can safely abort without any further action. 351 */ 352 void ext4_orphan_cleanup(struct super_block *sb, struct ext4_super_block *es) 353 { 354 unsigned int s_flags = sb->s_flags; 355 int nr_orphans = 0, nr_truncates = 0; 356 struct inode *inode; 357 int i, j; 358 #ifdef CONFIG_QUOTA 359 int quota_update = 0; 360 #endif 361 __le32 *bdata; 362 struct ext4_orphan_info *oi = &EXT4_SB(sb)->s_orphan_info; 363 int inodes_per_ob = ext4_inodes_per_orphan_block(sb); 364 365 if (!es->s_last_orphan && !oi->of_blocks) { 366 jbd_debug(4, "no orphan inodes to clean up\n"); 367 return; 368 } 369 370 if (bdev_read_only(sb->s_bdev)) { 371 ext4_msg(sb, KERN_ERR, "write access " 372 "unavailable, skipping orphan cleanup"); 373 return; 374 } 375 376 /* Check if feature set would not allow a r/w mount */ 377 if (!ext4_feature_set_ok(sb, 0)) { 378 ext4_msg(sb, KERN_INFO, "Skipping orphan cleanup due to " 379 "unknown ROCOMPAT features"); 380 return; 381 } 382 383 if (EXT4_SB(sb)->s_mount_state & EXT4_ERROR_FS) { 384 /* don't clear list on RO mount w/ errors */ 385 if (es->s_last_orphan && !(s_flags & SB_RDONLY)) { 386 ext4_msg(sb, KERN_INFO, "Errors on filesystem, " 387 "clearing orphan list.\n"); 388 es->s_last_orphan = 0; 389 } 390 jbd_debug(1, "Skipping orphan recovery on fs with errors.\n"); 391 return; 392 } 393 394 if (s_flags & SB_RDONLY) { 395 ext4_msg(sb, KERN_INFO, "orphan cleanup on readonly fs"); 396 sb->s_flags &= ~SB_RDONLY; 397 } 398 #ifdef CONFIG_QUOTA 399 /* 400 * Turn on quotas which were not enabled for read-only mounts if 401 * filesystem has quota feature, so that they are updated correctly. 402 */ 403 if (ext4_has_feature_quota(sb) && (s_flags & SB_RDONLY)) { 404 int ret = ext4_enable_quotas(sb); 405 406 if (!ret) 407 quota_update = 1; 408 else 409 ext4_msg(sb, KERN_ERR, 410 "Cannot turn on quotas: error %d", ret); 411 } 412 413 /* Turn on journaled quotas used for old sytle */ 414 for (i = 0; i < EXT4_MAXQUOTAS; i++) { 415 if (EXT4_SB(sb)->s_qf_names[i]) { 416 int ret = ext4_quota_on_mount(sb, i); 417 418 if (!ret) 419 quota_update = 1; 420 else 421 ext4_msg(sb, KERN_ERR, 422 "Cannot turn on journaled " 423 "quota: type %d: error %d", i, ret); 424 } 425 } 426 #endif 427 428 while (es->s_last_orphan) { 429 /* 430 * We may have encountered an error during cleanup; if 431 * so, skip the rest. 432 */ 433 if (EXT4_SB(sb)->s_mount_state & EXT4_ERROR_FS) { 434 jbd_debug(1, "Skipping orphan recovery on fs with errors.\n"); 435 es->s_last_orphan = 0; 436 break; 437 } 438 439 inode = ext4_orphan_get(sb, le32_to_cpu(es->s_last_orphan)); 440 if (IS_ERR(inode)) { 441 es->s_last_orphan = 0; 442 break; 443 } 444 445 list_add(&EXT4_I(inode)->i_orphan, &EXT4_SB(sb)->s_orphan); 446 ext4_process_orphan(inode, &nr_truncates, &nr_orphans); 447 } 448 449 for (i = 0; i < oi->of_blocks; i++) { 450 bdata = (__le32 *)(oi->of_binfo[i].ob_bh->b_data); 451 for (j = 0; j < inodes_per_ob; j++) { 452 if (!bdata[j]) 453 continue; 454 inode = ext4_orphan_get(sb, le32_to_cpu(bdata[j])); 455 if (IS_ERR(inode)) 456 continue; 457 ext4_set_inode_state(inode, EXT4_STATE_ORPHAN_FILE); 458 EXT4_I(inode)->i_orphan_idx = i * inodes_per_ob + j; 459 ext4_process_orphan(inode, &nr_truncates, &nr_orphans); 460 } 461 } 462 463 #define PLURAL(x) (x), ((x) == 1) ? "" : "s" 464 465 if (nr_orphans) 466 ext4_msg(sb, KERN_INFO, "%d orphan inode%s deleted", 467 PLURAL(nr_orphans)); 468 if (nr_truncates) 469 ext4_msg(sb, KERN_INFO, "%d truncate%s cleaned up", 470 PLURAL(nr_truncates)); 471 #ifdef CONFIG_QUOTA 472 /* Turn off quotas if they were enabled for orphan cleanup */ 473 if (quota_update) { 474 for (i = 0; i < EXT4_MAXQUOTAS; i++) { 475 if (sb_dqopt(sb)->files[i]) 476 dquot_quota_off(sb, i); 477 } 478 } 479 #endif 480 sb->s_flags = s_flags; /* Restore SB_RDONLY status */ 481 } 482 483 void ext4_release_orphan_info(struct super_block *sb) 484 { 485 int i; 486 struct ext4_orphan_info *oi = &EXT4_SB(sb)->s_orphan_info; 487 488 if (!oi->of_blocks) 489 return; 490 for (i = 0; i < oi->of_blocks; i++) 491 brelse(oi->of_binfo[i].ob_bh); 492 kfree(oi->of_binfo); 493 } 494 495 static struct ext4_orphan_block_tail *ext4_orphan_block_tail( 496 struct super_block *sb, 497 struct buffer_head *bh) 498 { 499 return (struct ext4_orphan_block_tail *)(bh->b_data + sb->s_blocksize - 500 sizeof(struct ext4_orphan_block_tail)); 501 } 502 503 static int ext4_orphan_file_block_csum_verify(struct super_block *sb, 504 struct buffer_head *bh) 505 { 506 __u32 calculated; 507 int inodes_per_ob = ext4_inodes_per_orphan_block(sb); 508 struct ext4_orphan_info *oi = &EXT4_SB(sb)->s_orphan_info; 509 struct ext4_orphan_block_tail *ot; 510 __le64 dsk_block_nr = cpu_to_le64(bh->b_blocknr); 511 512 if (!ext4_has_metadata_csum(sb)) 513 return 1; 514 515 ot = ext4_orphan_block_tail(sb, bh); 516 calculated = ext4_chksum(EXT4_SB(sb), oi->of_csum_seed, 517 (__u8 *)&dsk_block_nr, sizeof(dsk_block_nr)); 518 calculated = ext4_chksum(EXT4_SB(sb), calculated, (__u8 *)bh->b_data, 519 inodes_per_ob * sizeof(__u32)); 520 return le32_to_cpu(ot->ob_checksum) == calculated; 521 } 522 523 /* This gets called only when checksumming is enabled */ 524 void ext4_orphan_file_block_trigger(struct jbd2_buffer_trigger_type *triggers, 525 struct buffer_head *bh, 526 void *data, size_t size) 527 { 528 struct super_block *sb = EXT4_TRIGGER(triggers)->sb; 529 __u32 csum; 530 int inodes_per_ob = ext4_inodes_per_orphan_block(sb); 531 struct ext4_orphan_info *oi = &EXT4_SB(sb)->s_orphan_info; 532 struct ext4_orphan_block_tail *ot; 533 __le64 dsk_block_nr = cpu_to_le64(bh->b_blocknr); 534 535 csum = ext4_chksum(EXT4_SB(sb), oi->of_csum_seed, 536 (__u8 *)&dsk_block_nr, sizeof(dsk_block_nr)); 537 csum = ext4_chksum(EXT4_SB(sb), csum, (__u8 *)data, 538 inodes_per_ob * sizeof(__u32)); 539 ot = ext4_orphan_block_tail(sb, bh); 540 ot->ob_checksum = cpu_to_le32(csum); 541 } 542 543 int ext4_init_orphan_info(struct super_block *sb) 544 { 545 struct ext4_orphan_info *oi = &EXT4_SB(sb)->s_orphan_info; 546 struct inode *inode; 547 int i, j; 548 int ret; 549 int free; 550 __le32 *bdata; 551 int inodes_per_ob = ext4_inodes_per_orphan_block(sb); 552 struct ext4_orphan_block_tail *ot; 553 ino_t orphan_ino = le32_to_cpu(EXT4_SB(sb)->s_es->s_orphan_file_inum); 554 555 spin_lock_init(&oi->of_lock); 556 557 if (!ext4_has_feature_orphan_file(sb)) 558 return 0; 559 560 inode = ext4_iget(sb, orphan_ino, EXT4_IGET_SPECIAL); 561 if (IS_ERR(inode)) { 562 ext4_msg(sb, KERN_ERR, "get orphan inode failed"); 563 return PTR_ERR(inode); 564 } 565 oi->of_blocks = inode->i_size >> sb->s_blocksize_bits; 566 oi->of_csum_seed = EXT4_I(inode)->i_csum_seed; 567 oi->of_binfo = kmalloc(oi->of_blocks*sizeof(struct ext4_orphan_block), 568 GFP_KERNEL); 569 if (!oi->of_binfo) { 570 ret = -ENOMEM; 571 goto out_put; 572 } 573 for (i = 0; i < oi->of_blocks; i++) { 574 oi->of_binfo[i].ob_bh = ext4_bread(NULL, inode, i, 0); 575 if (IS_ERR(oi->of_binfo[i].ob_bh)) { 576 ret = PTR_ERR(oi->of_binfo[i].ob_bh); 577 goto out_free; 578 } 579 if (!oi->of_binfo[i].ob_bh) { 580 ret = -EIO; 581 goto out_free; 582 } 583 ot = ext4_orphan_block_tail(sb, oi->of_binfo[i].ob_bh); 584 if (le32_to_cpu(ot->ob_magic) != EXT4_ORPHAN_BLOCK_MAGIC) { 585 ext4_error(sb, "orphan file block %d: bad magic", i); 586 ret = -EIO; 587 goto out_free; 588 } 589 if (!ext4_orphan_file_block_csum_verify(sb, 590 oi->of_binfo[i].ob_bh)) { 591 ext4_error(sb, "orphan file block %d: bad checksum", i); 592 ret = -EIO; 593 goto out_free; 594 } 595 bdata = (__le32 *)(oi->of_binfo[i].ob_bh->b_data); 596 free = 0; 597 for (j = 0; j < inodes_per_ob; j++) 598 if (bdata[j] == 0) 599 free++; 600 oi->of_binfo[i].ob_free_entries = free; 601 } 602 iput(inode); 603 return 0; 604 out_free: 605 for (i--; i >= 0; i--) 606 brelse(oi->of_binfo[i].ob_bh); 607 kfree(oi->of_binfo); 608 out_put: 609 iput(inode); 610 return ret; 611 } 612 613 int ext4_orphan_file_empty(struct super_block *sb) 614 { 615 struct ext4_orphan_info *oi = &EXT4_SB(sb)->s_orphan_info; 616 int i; 617 int inodes_per_ob = ext4_inodes_per_orphan_block(sb); 618 619 if (!ext4_has_feature_orphan_file(sb)) 620 return 1; 621 for (i = 0; i < oi->of_blocks; i++) 622 if (oi->of_binfo[i].ob_free_entries != inodes_per_ob) 623 return 0; 624 return 1; 625 } 626