1 /* 2 * linux/fs/hfsplus/super.c 3 * 4 * Copyright (C) 2001 5 * Brad Boyer (flar@allandria.com) 6 * (C) 2003 Ardis Technologies <roman@ardistech.com> 7 * 8 */ 9 10 #include <linux/module.h> 11 #include <linux/init.h> 12 #include <linux/pagemap.h> 13 #include <linux/blkdev.h> 14 #include <linux/backing-dev.h> 15 #include <linux/fs.h> 16 #include <linux/slab.h> 17 #include <linux/vfs.h> 18 #include <linux/nls.h> 19 20 static struct inode *hfsplus_alloc_inode(struct super_block *sb); 21 static void hfsplus_destroy_inode(struct inode *inode); 22 23 #include "hfsplus_fs.h" 24 #include "xattr.h" 25 26 static int hfsplus_system_read_inode(struct inode *inode) 27 { 28 struct hfsplus_vh *vhdr = HFSPLUS_SB(inode->i_sb)->s_vhdr; 29 30 switch (inode->i_ino) { 31 case HFSPLUS_EXT_CNID: 32 hfsplus_inode_read_fork(inode, &vhdr->ext_file); 33 inode->i_mapping->a_ops = &hfsplus_btree_aops; 34 break; 35 case HFSPLUS_CAT_CNID: 36 hfsplus_inode_read_fork(inode, &vhdr->cat_file); 37 inode->i_mapping->a_ops = &hfsplus_btree_aops; 38 break; 39 case HFSPLUS_ALLOC_CNID: 40 hfsplus_inode_read_fork(inode, &vhdr->alloc_file); 41 inode->i_mapping->a_ops = &hfsplus_aops; 42 break; 43 case HFSPLUS_START_CNID: 44 hfsplus_inode_read_fork(inode, &vhdr->start_file); 45 break; 46 case HFSPLUS_ATTR_CNID: 47 hfsplus_inode_read_fork(inode, &vhdr->attr_file); 48 inode->i_mapping->a_ops = &hfsplus_btree_aops; 49 break; 50 default: 51 return -EIO; 52 } 53 54 return 0; 55 } 56 57 struct inode *hfsplus_iget(struct super_block *sb, unsigned long ino) 58 { 59 struct hfs_find_data fd; 60 struct inode *inode; 61 int err; 62 63 inode = iget_locked(sb, ino); 64 if (!inode) 65 return ERR_PTR(-ENOMEM); 66 if (!(inode->i_state & I_NEW)) 67 return inode; 68 69 INIT_LIST_HEAD(&HFSPLUS_I(inode)->open_dir_list); 70 spin_lock_init(&HFSPLUS_I(inode)->open_dir_lock); 71 mutex_init(&HFSPLUS_I(inode)->extents_lock); 72 HFSPLUS_I(inode)->flags = 0; 73 HFSPLUS_I(inode)->extent_state = 0; 74 HFSPLUS_I(inode)->rsrc_inode = NULL; 75 atomic_set(&HFSPLUS_I(inode)->opencnt, 0); 76 77 if (inode->i_ino >= HFSPLUS_FIRSTUSER_CNID || 78 inode->i_ino == HFSPLUS_ROOT_CNID) { 79 err = hfs_find_init(HFSPLUS_SB(inode->i_sb)->cat_tree, &fd); 80 if (!err) { 81 err = hfsplus_find_cat(inode->i_sb, inode->i_ino, &fd); 82 if (!err) 83 err = hfsplus_cat_read_inode(inode, &fd); 84 hfs_find_exit(&fd); 85 } 86 } else { 87 err = hfsplus_system_read_inode(inode); 88 } 89 90 if (err) { 91 iget_failed(inode); 92 return ERR_PTR(err); 93 } 94 95 unlock_new_inode(inode); 96 return inode; 97 } 98 99 static int hfsplus_system_write_inode(struct inode *inode) 100 { 101 struct hfsplus_sb_info *sbi = HFSPLUS_SB(inode->i_sb); 102 struct hfsplus_vh *vhdr = sbi->s_vhdr; 103 struct hfsplus_fork_raw *fork; 104 struct hfs_btree *tree = NULL; 105 106 switch (inode->i_ino) { 107 case HFSPLUS_EXT_CNID: 108 fork = &vhdr->ext_file; 109 tree = sbi->ext_tree; 110 break; 111 case HFSPLUS_CAT_CNID: 112 fork = &vhdr->cat_file; 113 tree = sbi->cat_tree; 114 break; 115 case HFSPLUS_ALLOC_CNID: 116 fork = &vhdr->alloc_file; 117 break; 118 case HFSPLUS_START_CNID: 119 fork = &vhdr->start_file; 120 break; 121 case HFSPLUS_ATTR_CNID: 122 fork = &vhdr->attr_file; 123 tree = sbi->attr_tree; 124 break; 125 default: 126 return -EIO; 127 } 128 129 if (fork->total_size != cpu_to_be64(inode->i_size)) { 130 set_bit(HFSPLUS_SB_WRITEBACKUP, &sbi->flags); 131 hfsplus_mark_mdb_dirty(inode->i_sb); 132 } 133 hfsplus_inode_write_fork(inode, fork); 134 if (tree) { 135 int err = hfs_btree_write(tree); 136 137 if (err) { 138 pr_err("b-tree write err: %d, ino %lu\n", 139 err, inode->i_ino); 140 return err; 141 } 142 } 143 return 0; 144 } 145 146 static int hfsplus_write_inode(struct inode *inode, 147 struct writeback_control *wbc) 148 { 149 int err; 150 151 hfs_dbg(INODE, "hfsplus_write_inode: %lu\n", inode->i_ino); 152 153 err = hfsplus_ext_write_extent(inode); 154 if (err) 155 return err; 156 157 if (inode->i_ino >= HFSPLUS_FIRSTUSER_CNID || 158 inode->i_ino == HFSPLUS_ROOT_CNID) 159 return hfsplus_cat_write_inode(inode); 160 else 161 return hfsplus_system_write_inode(inode); 162 } 163 164 static void hfsplus_evict_inode(struct inode *inode) 165 { 166 hfs_dbg(INODE, "hfsplus_evict_inode: %lu\n", inode->i_ino); 167 truncate_inode_pages_final(&inode->i_data); 168 clear_inode(inode); 169 if (HFSPLUS_IS_RSRC(inode)) { 170 HFSPLUS_I(HFSPLUS_I(inode)->rsrc_inode)->rsrc_inode = NULL; 171 iput(HFSPLUS_I(inode)->rsrc_inode); 172 } 173 } 174 175 static int hfsplus_sync_fs(struct super_block *sb, int wait) 176 { 177 struct hfsplus_sb_info *sbi = HFSPLUS_SB(sb); 178 struct hfsplus_vh *vhdr = sbi->s_vhdr; 179 int write_backup = 0; 180 int error, error2; 181 182 if (!wait) 183 return 0; 184 185 hfs_dbg(SUPER, "hfsplus_sync_fs\n"); 186 187 /* 188 * Explicitly write out the special metadata inodes. 189 * 190 * While these special inodes are marked as hashed and written 191 * out peridocically by the flusher threads we redirty them 192 * during writeout of normal inodes, and thus the life lock 193 * prevents us from getting the latest state to disk. 194 */ 195 error = filemap_write_and_wait(sbi->cat_tree->inode->i_mapping); 196 error2 = filemap_write_and_wait(sbi->ext_tree->inode->i_mapping); 197 if (!error) 198 error = error2; 199 if (sbi->attr_tree) { 200 error2 = 201 filemap_write_and_wait(sbi->attr_tree->inode->i_mapping); 202 if (!error) 203 error = error2; 204 } 205 error2 = filemap_write_and_wait(sbi->alloc_file->i_mapping); 206 if (!error) 207 error = error2; 208 209 mutex_lock(&sbi->vh_mutex); 210 mutex_lock(&sbi->alloc_mutex); 211 vhdr->free_blocks = cpu_to_be32(sbi->free_blocks); 212 vhdr->next_cnid = cpu_to_be32(sbi->next_cnid); 213 vhdr->folder_count = cpu_to_be32(sbi->folder_count); 214 vhdr->file_count = cpu_to_be32(sbi->file_count); 215 216 if (test_and_clear_bit(HFSPLUS_SB_WRITEBACKUP, &sbi->flags)) { 217 memcpy(sbi->s_backup_vhdr, sbi->s_vhdr, sizeof(*sbi->s_vhdr)); 218 write_backup = 1; 219 } 220 221 error2 = hfsplus_submit_bio(sb, 222 sbi->part_start + HFSPLUS_VOLHEAD_SECTOR, 223 sbi->s_vhdr_buf, NULL, WRITE_SYNC); 224 if (!error) 225 error = error2; 226 if (!write_backup) 227 goto out; 228 229 error2 = hfsplus_submit_bio(sb, 230 sbi->part_start + sbi->sect_count - 2, 231 sbi->s_backup_vhdr_buf, NULL, WRITE_SYNC); 232 if (!error) 233 error2 = error; 234 out: 235 mutex_unlock(&sbi->alloc_mutex); 236 mutex_unlock(&sbi->vh_mutex); 237 238 if (!test_bit(HFSPLUS_SB_NOBARRIER, &sbi->flags)) 239 blkdev_issue_flush(sb->s_bdev, GFP_KERNEL, NULL); 240 241 return error; 242 } 243 244 static void delayed_sync_fs(struct work_struct *work) 245 { 246 int err; 247 struct hfsplus_sb_info *sbi; 248 249 sbi = container_of(work, struct hfsplus_sb_info, sync_work.work); 250 251 spin_lock(&sbi->work_lock); 252 sbi->work_queued = 0; 253 spin_unlock(&sbi->work_lock); 254 255 err = hfsplus_sync_fs(sbi->alloc_file->i_sb, 1); 256 if (err) 257 pr_err("delayed sync fs err %d\n", err); 258 } 259 260 void hfsplus_mark_mdb_dirty(struct super_block *sb) 261 { 262 struct hfsplus_sb_info *sbi = HFSPLUS_SB(sb); 263 unsigned long delay; 264 265 if (sb->s_flags & MS_RDONLY) 266 return; 267 268 spin_lock(&sbi->work_lock); 269 if (!sbi->work_queued) { 270 delay = msecs_to_jiffies(dirty_writeback_interval * 10); 271 queue_delayed_work(system_long_wq, &sbi->sync_work, delay); 272 sbi->work_queued = 1; 273 } 274 spin_unlock(&sbi->work_lock); 275 } 276 277 static void hfsplus_put_super(struct super_block *sb) 278 { 279 struct hfsplus_sb_info *sbi = HFSPLUS_SB(sb); 280 281 hfs_dbg(SUPER, "hfsplus_put_super\n"); 282 283 cancel_delayed_work_sync(&sbi->sync_work); 284 285 if (!(sb->s_flags & MS_RDONLY) && sbi->s_vhdr) { 286 struct hfsplus_vh *vhdr = sbi->s_vhdr; 287 288 vhdr->modify_date = hfsp_now2mt(); 289 vhdr->attributes |= cpu_to_be32(HFSPLUS_VOL_UNMNT); 290 vhdr->attributes &= cpu_to_be32(~HFSPLUS_VOL_INCNSTNT); 291 292 hfsplus_sync_fs(sb, 1); 293 } 294 295 hfs_btree_close(sbi->attr_tree); 296 hfs_btree_close(sbi->cat_tree); 297 hfs_btree_close(sbi->ext_tree); 298 iput(sbi->alloc_file); 299 iput(sbi->hidden_dir); 300 kfree(sbi->s_vhdr_buf); 301 kfree(sbi->s_backup_vhdr_buf); 302 unload_nls(sbi->nls); 303 kfree(sb->s_fs_info); 304 sb->s_fs_info = NULL; 305 } 306 307 static int hfsplus_statfs(struct dentry *dentry, struct kstatfs *buf) 308 { 309 struct super_block *sb = dentry->d_sb; 310 struct hfsplus_sb_info *sbi = HFSPLUS_SB(sb); 311 u64 id = huge_encode_dev(sb->s_bdev->bd_dev); 312 313 buf->f_type = HFSPLUS_SUPER_MAGIC; 314 buf->f_bsize = sb->s_blocksize; 315 buf->f_blocks = sbi->total_blocks << sbi->fs_shift; 316 buf->f_bfree = sbi->free_blocks << sbi->fs_shift; 317 buf->f_bavail = buf->f_bfree; 318 buf->f_files = 0xFFFFFFFF; 319 buf->f_ffree = 0xFFFFFFFF - sbi->next_cnid; 320 buf->f_fsid.val[0] = (u32)id; 321 buf->f_fsid.val[1] = (u32)(id >> 32); 322 buf->f_namelen = HFSPLUS_MAX_STRLEN; 323 324 return 0; 325 } 326 327 static int hfsplus_remount(struct super_block *sb, int *flags, char *data) 328 { 329 sync_filesystem(sb); 330 if ((*flags & MS_RDONLY) == (sb->s_flags & MS_RDONLY)) 331 return 0; 332 if (!(*flags & MS_RDONLY)) { 333 struct hfsplus_vh *vhdr = HFSPLUS_SB(sb)->s_vhdr; 334 int force = 0; 335 336 if (!hfsplus_parse_options_remount(data, &force)) 337 return -EINVAL; 338 339 if (!(vhdr->attributes & cpu_to_be32(HFSPLUS_VOL_UNMNT))) { 340 pr_warn("filesystem was not cleanly unmounted, running fsck.hfsplus is recommended. leaving read-only.\n"); 341 sb->s_flags |= MS_RDONLY; 342 *flags |= MS_RDONLY; 343 } else if (force) { 344 /* nothing */ 345 } else if (vhdr->attributes & 346 cpu_to_be32(HFSPLUS_VOL_SOFTLOCK)) { 347 pr_warn("filesystem is marked locked, leaving read-only.\n"); 348 sb->s_flags |= MS_RDONLY; 349 *flags |= MS_RDONLY; 350 } else if (vhdr->attributes & 351 cpu_to_be32(HFSPLUS_VOL_JOURNALED)) { 352 pr_warn("filesystem is marked journaled, leaving read-only.\n"); 353 sb->s_flags |= MS_RDONLY; 354 *flags |= MS_RDONLY; 355 } 356 } 357 return 0; 358 } 359 360 static const struct super_operations hfsplus_sops = { 361 .alloc_inode = hfsplus_alloc_inode, 362 .destroy_inode = hfsplus_destroy_inode, 363 .write_inode = hfsplus_write_inode, 364 .evict_inode = hfsplus_evict_inode, 365 .put_super = hfsplus_put_super, 366 .sync_fs = hfsplus_sync_fs, 367 .statfs = hfsplus_statfs, 368 .remount_fs = hfsplus_remount, 369 .show_options = hfsplus_show_options, 370 }; 371 372 static int hfsplus_fill_super(struct super_block *sb, void *data, int silent) 373 { 374 struct hfsplus_vh *vhdr; 375 struct hfsplus_sb_info *sbi; 376 hfsplus_cat_entry entry; 377 struct hfs_find_data fd; 378 struct inode *root, *inode; 379 struct qstr str; 380 struct nls_table *nls = NULL; 381 u64 last_fs_block, last_fs_page; 382 int err; 383 384 err = -ENOMEM; 385 sbi = kzalloc(sizeof(*sbi), GFP_KERNEL); 386 if (!sbi) 387 goto out; 388 389 sb->s_fs_info = sbi; 390 mutex_init(&sbi->alloc_mutex); 391 mutex_init(&sbi->vh_mutex); 392 spin_lock_init(&sbi->work_lock); 393 INIT_DELAYED_WORK(&sbi->sync_work, delayed_sync_fs); 394 hfsplus_fill_defaults(sbi); 395 396 err = -EINVAL; 397 if (!hfsplus_parse_options(data, sbi)) { 398 pr_err("unable to parse mount options\n"); 399 goto out_unload_nls; 400 } 401 402 /* temporarily use utf8 to correctly find the hidden dir below */ 403 nls = sbi->nls; 404 sbi->nls = load_nls("utf8"); 405 if (!sbi->nls) { 406 pr_err("unable to load nls for utf8\n"); 407 goto out_unload_nls; 408 } 409 410 /* Grab the volume header */ 411 if (hfsplus_read_wrapper(sb)) { 412 if (!silent) 413 pr_warn("unable to find HFS+ superblock\n"); 414 goto out_unload_nls; 415 } 416 vhdr = sbi->s_vhdr; 417 418 /* Copy parts of the volume header into the superblock */ 419 sb->s_magic = HFSPLUS_VOLHEAD_SIG; 420 if (be16_to_cpu(vhdr->version) < HFSPLUS_MIN_VERSION || 421 be16_to_cpu(vhdr->version) > HFSPLUS_CURRENT_VERSION) { 422 pr_err("wrong filesystem version\n"); 423 goto out_free_vhdr; 424 } 425 sbi->total_blocks = be32_to_cpu(vhdr->total_blocks); 426 sbi->free_blocks = be32_to_cpu(vhdr->free_blocks); 427 sbi->next_cnid = be32_to_cpu(vhdr->next_cnid); 428 sbi->file_count = be32_to_cpu(vhdr->file_count); 429 sbi->folder_count = be32_to_cpu(vhdr->folder_count); 430 sbi->data_clump_blocks = 431 be32_to_cpu(vhdr->data_clump_sz) >> sbi->alloc_blksz_shift; 432 if (!sbi->data_clump_blocks) 433 sbi->data_clump_blocks = 1; 434 sbi->rsrc_clump_blocks = 435 be32_to_cpu(vhdr->rsrc_clump_sz) >> sbi->alloc_blksz_shift; 436 if (!sbi->rsrc_clump_blocks) 437 sbi->rsrc_clump_blocks = 1; 438 439 err = -EFBIG; 440 last_fs_block = sbi->total_blocks - 1; 441 last_fs_page = (last_fs_block << sbi->alloc_blksz_shift) >> 442 PAGE_SHIFT; 443 444 if ((last_fs_block > (sector_t)(~0ULL) >> (sbi->alloc_blksz_shift - 9)) || 445 (last_fs_page > (pgoff_t)(~0ULL))) { 446 pr_err("filesystem size too large\n"); 447 goto out_free_vhdr; 448 } 449 450 /* Set up operations so we can load metadata */ 451 sb->s_op = &hfsplus_sops; 452 sb->s_maxbytes = MAX_LFS_FILESIZE; 453 454 if (!(vhdr->attributes & cpu_to_be32(HFSPLUS_VOL_UNMNT))) { 455 pr_warn("Filesystem was not cleanly unmounted, running fsck.hfsplus is recommended. mounting read-only.\n"); 456 sb->s_flags |= MS_RDONLY; 457 } else if (test_and_clear_bit(HFSPLUS_SB_FORCE, &sbi->flags)) { 458 /* nothing */ 459 } else if (vhdr->attributes & cpu_to_be32(HFSPLUS_VOL_SOFTLOCK)) { 460 pr_warn("Filesystem is marked locked, mounting read-only.\n"); 461 sb->s_flags |= MS_RDONLY; 462 } else if ((vhdr->attributes & cpu_to_be32(HFSPLUS_VOL_JOURNALED)) && 463 !(sb->s_flags & MS_RDONLY)) { 464 pr_warn("write access to a journaled filesystem is not supported, use the force option at your own risk, mounting read-only.\n"); 465 sb->s_flags |= MS_RDONLY; 466 } 467 468 err = -EINVAL; 469 470 /* Load metadata objects (B*Trees) */ 471 sbi->ext_tree = hfs_btree_open(sb, HFSPLUS_EXT_CNID); 472 if (!sbi->ext_tree) { 473 pr_err("failed to load extents file\n"); 474 goto out_free_vhdr; 475 } 476 sbi->cat_tree = hfs_btree_open(sb, HFSPLUS_CAT_CNID); 477 if (!sbi->cat_tree) { 478 pr_err("failed to load catalog file\n"); 479 goto out_close_ext_tree; 480 } 481 atomic_set(&sbi->attr_tree_state, HFSPLUS_EMPTY_ATTR_TREE); 482 if (vhdr->attr_file.total_blocks != 0) { 483 sbi->attr_tree = hfs_btree_open(sb, HFSPLUS_ATTR_CNID); 484 if (!sbi->attr_tree) { 485 pr_err("failed to load attributes file\n"); 486 goto out_close_cat_tree; 487 } 488 atomic_set(&sbi->attr_tree_state, HFSPLUS_VALID_ATTR_TREE); 489 } 490 sb->s_xattr = hfsplus_xattr_handlers; 491 492 inode = hfsplus_iget(sb, HFSPLUS_ALLOC_CNID); 493 if (IS_ERR(inode)) { 494 pr_err("failed to load allocation file\n"); 495 err = PTR_ERR(inode); 496 goto out_close_attr_tree; 497 } 498 sbi->alloc_file = inode; 499 500 /* Load the root directory */ 501 root = hfsplus_iget(sb, HFSPLUS_ROOT_CNID); 502 if (IS_ERR(root)) { 503 pr_err("failed to load root directory\n"); 504 err = PTR_ERR(root); 505 goto out_put_alloc_file; 506 } 507 508 sb->s_d_op = &hfsplus_dentry_operations; 509 sb->s_root = d_make_root(root); 510 if (!sb->s_root) { 511 err = -ENOMEM; 512 goto out_put_alloc_file; 513 } 514 515 str.len = sizeof(HFSP_HIDDENDIR_NAME) - 1; 516 str.name = HFSP_HIDDENDIR_NAME; 517 err = hfs_find_init(sbi->cat_tree, &fd); 518 if (err) 519 goto out_put_root; 520 err = hfsplus_cat_build_key(sb, fd.search_key, HFSPLUS_ROOT_CNID, &str); 521 if (unlikely(err < 0)) 522 goto out_put_root; 523 if (!hfs_brec_read(&fd, &entry, sizeof(entry))) { 524 hfs_find_exit(&fd); 525 if (entry.type != cpu_to_be16(HFSPLUS_FOLDER)) 526 goto out_put_root; 527 inode = hfsplus_iget(sb, be32_to_cpu(entry.folder.id)); 528 if (IS_ERR(inode)) { 529 err = PTR_ERR(inode); 530 goto out_put_root; 531 } 532 sbi->hidden_dir = inode; 533 } else 534 hfs_find_exit(&fd); 535 536 if (!(sb->s_flags & MS_RDONLY)) { 537 /* 538 * H+LX == hfsplusutils, H+Lx == this driver, H+lx is unused 539 * all three are registered with Apple for our use 540 */ 541 vhdr->last_mount_vers = cpu_to_be32(HFSP_MOUNT_VERSION); 542 vhdr->modify_date = hfsp_now2mt(); 543 be32_add_cpu(&vhdr->write_count, 1); 544 vhdr->attributes &= cpu_to_be32(~HFSPLUS_VOL_UNMNT); 545 vhdr->attributes |= cpu_to_be32(HFSPLUS_VOL_INCNSTNT); 546 hfsplus_sync_fs(sb, 1); 547 548 if (!sbi->hidden_dir) { 549 mutex_lock(&sbi->vh_mutex); 550 sbi->hidden_dir = hfsplus_new_inode(sb, S_IFDIR); 551 if (!sbi->hidden_dir) { 552 mutex_unlock(&sbi->vh_mutex); 553 err = -ENOMEM; 554 goto out_put_root; 555 } 556 err = hfsplus_create_cat(sbi->hidden_dir->i_ino, root, 557 &str, sbi->hidden_dir); 558 if (err) { 559 mutex_unlock(&sbi->vh_mutex); 560 goto out_put_hidden_dir; 561 } 562 563 err = hfsplus_init_inode_security(sbi->hidden_dir, 564 root, &str); 565 if (err == -EOPNOTSUPP) 566 err = 0; /* Operation is not supported. */ 567 else if (err) { 568 /* 569 * Try to delete anyway without 570 * error analysis. 571 */ 572 hfsplus_delete_cat(sbi->hidden_dir->i_ino, 573 root, &str); 574 mutex_unlock(&sbi->vh_mutex); 575 goto out_put_hidden_dir; 576 } 577 578 mutex_unlock(&sbi->vh_mutex); 579 hfsplus_mark_inode_dirty(sbi->hidden_dir, 580 HFSPLUS_I_CAT_DIRTY); 581 } 582 } 583 584 unload_nls(sbi->nls); 585 sbi->nls = nls; 586 return 0; 587 588 out_put_hidden_dir: 589 iput(sbi->hidden_dir); 590 out_put_root: 591 dput(sb->s_root); 592 sb->s_root = NULL; 593 out_put_alloc_file: 594 iput(sbi->alloc_file); 595 out_close_attr_tree: 596 hfs_btree_close(sbi->attr_tree); 597 out_close_cat_tree: 598 hfs_btree_close(sbi->cat_tree); 599 out_close_ext_tree: 600 hfs_btree_close(sbi->ext_tree); 601 out_free_vhdr: 602 kfree(sbi->s_vhdr_buf); 603 kfree(sbi->s_backup_vhdr_buf); 604 out_unload_nls: 605 unload_nls(sbi->nls); 606 unload_nls(nls); 607 kfree(sbi); 608 out: 609 return err; 610 } 611 612 MODULE_AUTHOR("Brad Boyer"); 613 MODULE_DESCRIPTION("Extended Macintosh Filesystem"); 614 MODULE_LICENSE("GPL"); 615 616 static struct kmem_cache *hfsplus_inode_cachep; 617 618 static struct inode *hfsplus_alloc_inode(struct super_block *sb) 619 { 620 struct hfsplus_inode_info *i; 621 622 i = kmem_cache_alloc(hfsplus_inode_cachep, GFP_KERNEL); 623 return i ? &i->vfs_inode : NULL; 624 } 625 626 static void hfsplus_i_callback(struct rcu_head *head) 627 { 628 struct inode *inode = container_of(head, struct inode, i_rcu); 629 630 kmem_cache_free(hfsplus_inode_cachep, HFSPLUS_I(inode)); 631 } 632 633 static void hfsplus_destroy_inode(struct inode *inode) 634 { 635 call_rcu(&inode->i_rcu, hfsplus_i_callback); 636 } 637 638 #define HFSPLUS_INODE_SIZE sizeof(struct hfsplus_inode_info) 639 640 static struct dentry *hfsplus_mount(struct file_system_type *fs_type, 641 int flags, const char *dev_name, void *data) 642 { 643 return mount_bdev(fs_type, flags, dev_name, data, hfsplus_fill_super); 644 } 645 646 static struct file_system_type hfsplus_fs_type = { 647 .owner = THIS_MODULE, 648 .name = "hfsplus", 649 .mount = hfsplus_mount, 650 .kill_sb = kill_block_super, 651 .fs_flags = FS_REQUIRES_DEV, 652 }; 653 MODULE_ALIAS_FS("hfsplus"); 654 655 static void hfsplus_init_once(void *p) 656 { 657 struct hfsplus_inode_info *i = p; 658 659 inode_init_once(&i->vfs_inode); 660 } 661 662 static int __init init_hfsplus_fs(void) 663 { 664 int err; 665 666 hfsplus_inode_cachep = kmem_cache_create("hfsplus_icache", 667 HFSPLUS_INODE_SIZE, 0, SLAB_HWCACHE_ALIGN|SLAB_ACCOUNT, 668 hfsplus_init_once); 669 if (!hfsplus_inode_cachep) 670 return -ENOMEM; 671 err = hfsplus_create_attr_tree_cache(); 672 if (err) 673 goto destroy_inode_cache; 674 err = register_filesystem(&hfsplus_fs_type); 675 if (err) 676 goto destroy_attr_tree_cache; 677 return 0; 678 679 destroy_attr_tree_cache: 680 hfsplus_destroy_attr_tree_cache(); 681 682 destroy_inode_cache: 683 kmem_cache_destroy(hfsplus_inode_cachep); 684 685 return err; 686 } 687 688 static void __exit exit_hfsplus_fs(void) 689 { 690 unregister_filesystem(&hfsplus_fs_type); 691 692 /* 693 * Make sure all delayed rcu free inodes are flushed before we 694 * destroy cache. 695 */ 696 rcu_barrier(); 697 hfsplus_destroy_attr_tree_cache(); 698 kmem_cache_destroy(hfsplus_inode_cachep); 699 } 700 701 module_init(init_hfsplus_fs) 702 module_exit(exit_hfsplus_fs) 703