1 /* 2 * Copyright (C) Sistina Software, Inc. 1997-2003 All rights reserved. 3 * Copyright (C) 2004-2008 Red Hat, Inc. All rights reserved. 4 * 5 * This copyrighted material is made available to anyone wishing to use, 6 * modify, copy, or redistribute it subject to the terms and conditions 7 * of the GNU General Public License version 2. 8 */ 9 10 #include <linux/sched.h> 11 #include <linux/slab.h> 12 #include <linux/spinlock.h> 13 #include <linux/completion.h> 14 #include <linux/buffer_head.h> 15 #include <linux/posix_acl.h> 16 #include <linux/sort.h> 17 #include <linux/gfs2_ondisk.h> 18 #include <linux/crc32.h> 19 #include <linux/lm_interface.h> 20 #include <linux/security.h> 21 22 #include "gfs2.h" 23 #include "incore.h" 24 #include "acl.h" 25 #include "bmap.h" 26 #include "dir.h" 27 #include "eattr.h" 28 #include "glock.h" 29 #include "glops.h" 30 #include "inode.h" 31 #include "log.h" 32 #include "meta_io.h" 33 #include "ops_address.h" 34 #include "ops_inode.h" 35 #include "quota.h" 36 #include "rgrp.h" 37 #include "trans.h" 38 #include "util.h" 39 40 struct gfs2_inum_range_host { 41 u64 ir_start; 42 u64 ir_length; 43 }; 44 45 static int iget_test(struct inode *inode, void *opaque) 46 { 47 struct gfs2_inode *ip = GFS2_I(inode); 48 u64 *no_addr = opaque; 49 50 if (ip->i_no_addr == *no_addr && 51 inode->i_private != NULL) 52 return 1; 53 54 return 0; 55 } 56 57 static int iget_set(struct inode *inode, void *opaque) 58 { 59 struct gfs2_inode *ip = GFS2_I(inode); 60 u64 *no_addr = opaque; 61 62 inode->i_ino = (unsigned long)*no_addr; 63 ip->i_no_addr = *no_addr; 64 return 0; 65 } 66 67 struct inode *gfs2_ilookup(struct super_block *sb, u64 no_addr) 68 { 69 unsigned long hash = (unsigned long)no_addr; 70 return ilookup5(sb, hash, iget_test, &no_addr); 71 } 72 73 static struct inode *gfs2_iget(struct super_block *sb, u64 no_addr) 74 { 75 unsigned long hash = (unsigned long)no_addr; 76 return iget5_locked(sb, hash, iget_test, iget_set, &no_addr); 77 } 78 79 struct gfs2_skip_data { 80 u64 no_addr; 81 int skipped; 82 }; 83 84 static int iget_skip_test(struct inode *inode, void *opaque) 85 { 86 struct gfs2_inode *ip = GFS2_I(inode); 87 struct gfs2_skip_data *data = opaque; 88 89 if (ip->i_no_addr == data->no_addr && inode->i_private != NULL){ 90 if (inode->i_state & (I_FREEING|I_CLEAR|I_WILL_FREE)){ 91 data->skipped = 1; 92 return 0; 93 } 94 return 1; 95 } 96 return 0; 97 } 98 99 static int iget_skip_set(struct inode *inode, void *opaque) 100 { 101 struct gfs2_inode *ip = GFS2_I(inode); 102 struct gfs2_skip_data *data = opaque; 103 104 if (data->skipped) 105 return 1; 106 inode->i_ino = (unsigned long)(data->no_addr); 107 ip->i_no_addr = data->no_addr; 108 return 0; 109 } 110 111 static struct inode *gfs2_iget_skip(struct super_block *sb, 112 u64 no_addr) 113 { 114 struct gfs2_skip_data data; 115 unsigned long hash = (unsigned long)no_addr; 116 117 data.no_addr = no_addr; 118 data.skipped = 0; 119 return iget5_locked(sb, hash, iget_skip_test, iget_skip_set, &data); 120 } 121 122 /** 123 * GFS2 lookup code fills in vfs inode contents based on info obtained 124 * from directory entry inside gfs2_inode_lookup(). This has caused issues 125 * with NFS code path since its get_dentry routine doesn't have the relevant 126 * directory entry when gfs2_inode_lookup() is invoked. Part of the code 127 * segment inside gfs2_inode_lookup code needs to get moved around. 128 * 129 * Clean up I_LOCK and I_NEW as well. 130 **/ 131 132 void gfs2_set_iop(struct inode *inode) 133 { 134 struct gfs2_sbd *sdp = GFS2_SB(inode); 135 umode_t mode = inode->i_mode; 136 137 if (S_ISREG(mode)) { 138 inode->i_op = &gfs2_file_iops; 139 if (sdp->sd_args.ar_localflocks) 140 inode->i_fop = &gfs2_file_fops_nolock; 141 else 142 inode->i_fop = &gfs2_file_fops; 143 } else if (S_ISDIR(mode)) { 144 inode->i_op = &gfs2_dir_iops; 145 if (sdp->sd_args.ar_localflocks) 146 inode->i_fop = &gfs2_dir_fops_nolock; 147 else 148 inode->i_fop = &gfs2_dir_fops; 149 } else if (S_ISLNK(mode)) { 150 inode->i_op = &gfs2_symlink_iops; 151 } else { 152 inode->i_op = &gfs2_file_iops; 153 } 154 155 unlock_new_inode(inode); 156 } 157 158 /** 159 * gfs2_inode_lookup - Lookup an inode 160 * @sb: The super block 161 * @no_addr: The inode number 162 * @type: The type of the inode 163 * @skip_freeing: set this not return an inode if it is currently being freed. 164 * 165 * Returns: A VFS inode, or an error 166 */ 167 168 struct inode *gfs2_inode_lookup(struct super_block *sb, 169 unsigned int type, 170 u64 no_addr, 171 u64 no_formal_ino, int skip_freeing) 172 { 173 struct inode *inode; 174 struct gfs2_inode *ip; 175 struct gfs2_glock *io_gl; 176 int error; 177 178 if (skip_freeing) 179 inode = gfs2_iget_skip(sb, no_addr); 180 else 181 inode = gfs2_iget(sb, no_addr); 182 ip = GFS2_I(inode); 183 184 if (!inode) 185 return ERR_PTR(-ENOBUFS); 186 187 if (inode->i_state & I_NEW) { 188 struct gfs2_sbd *sdp = GFS2_SB(inode); 189 inode->i_private = ip; 190 ip->i_no_formal_ino = no_formal_ino; 191 192 error = gfs2_glock_get(sdp, no_addr, &gfs2_inode_glops, CREATE, &ip->i_gl); 193 if (unlikely(error)) 194 goto fail; 195 ip->i_gl->gl_object = ip; 196 197 error = gfs2_glock_get(sdp, no_addr, &gfs2_iopen_glops, CREATE, &io_gl); 198 if (unlikely(error)) 199 goto fail_put; 200 201 set_bit(GIF_INVALID, &ip->i_flags); 202 error = gfs2_glock_nq_init(io_gl, LM_ST_SHARED, GL_EXACT, &ip->i_iopen_gh); 203 if (unlikely(error)) 204 goto fail_iopen; 205 ip->i_iopen_gh.gh_gl->gl_object = ip; 206 207 gfs2_glock_put(io_gl); 208 209 if ((type == DT_UNKNOWN) && (no_formal_ino == 0)) 210 goto gfs2_nfsbypass; 211 212 inode->i_mode = DT2IF(type); 213 214 /* 215 * We must read the inode in order to work out its type in 216 * this case. Note that this doesn't happen often as we normally 217 * know the type beforehand. This code path only occurs during 218 * unlinked inode recovery (where it is safe to do this glock, 219 * which is not true in the general case). 220 */ 221 if (type == DT_UNKNOWN) { 222 struct gfs2_holder gh; 223 error = gfs2_glock_nq_init(ip->i_gl, LM_ST_EXCLUSIVE, 0, &gh); 224 if (unlikely(error)) 225 goto fail_glock; 226 /* Inode is now uptodate */ 227 gfs2_glock_dq_uninit(&gh); 228 } 229 230 gfs2_set_iop(inode); 231 } 232 233 gfs2_nfsbypass: 234 return inode; 235 fail_glock: 236 gfs2_glock_dq(&ip->i_iopen_gh); 237 fail_iopen: 238 gfs2_glock_put(io_gl); 239 fail_put: 240 ip->i_gl->gl_object = NULL; 241 gfs2_glock_put(ip->i_gl); 242 fail: 243 iget_failed(inode); 244 return ERR_PTR(error); 245 } 246 247 static int gfs2_dinode_in(struct gfs2_inode *ip, const void *buf) 248 { 249 struct gfs2_dinode_host *di = &ip->i_di; 250 const struct gfs2_dinode *str = buf; 251 u16 height, depth; 252 253 if (unlikely(ip->i_no_addr != be64_to_cpu(str->di_num.no_addr))) 254 goto corrupt; 255 ip->i_no_formal_ino = be64_to_cpu(str->di_num.no_formal_ino); 256 ip->i_inode.i_mode = be32_to_cpu(str->di_mode); 257 ip->i_inode.i_rdev = 0; 258 switch (ip->i_inode.i_mode & S_IFMT) { 259 case S_IFBLK: 260 case S_IFCHR: 261 ip->i_inode.i_rdev = MKDEV(be32_to_cpu(str->di_major), 262 be32_to_cpu(str->di_minor)); 263 break; 264 }; 265 266 ip->i_inode.i_uid = be32_to_cpu(str->di_uid); 267 ip->i_inode.i_gid = be32_to_cpu(str->di_gid); 268 /* 269 * We will need to review setting the nlink count here in the 270 * light of the forthcoming ro bind mount work. This is a reminder 271 * to do that. 272 */ 273 ip->i_inode.i_nlink = be32_to_cpu(str->di_nlink); 274 di->di_size = be64_to_cpu(str->di_size); 275 i_size_write(&ip->i_inode, di->di_size); 276 gfs2_set_inode_blocks(&ip->i_inode, be64_to_cpu(str->di_blocks)); 277 ip->i_inode.i_atime.tv_sec = be64_to_cpu(str->di_atime); 278 ip->i_inode.i_atime.tv_nsec = be32_to_cpu(str->di_atime_nsec); 279 ip->i_inode.i_mtime.tv_sec = be64_to_cpu(str->di_mtime); 280 ip->i_inode.i_mtime.tv_nsec = be32_to_cpu(str->di_mtime_nsec); 281 ip->i_inode.i_ctime.tv_sec = be64_to_cpu(str->di_ctime); 282 ip->i_inode.i_ctime.tv_nsec = be32_to_cpu(str->di_ctime_nsec); 283 284 ip->i_goal = be64_to_cpu(str->di_goal_meta); 285 di->di_generation = be64_to_cpu(str->di_generation); 286 287 di->di_flags = be32_to_cpu(str->di_flags); 288 gfs2_set_inode_flags(&ip->i_inode); 289 height = be16_to_cpu(str->di_height); 290 if (unlikely(height > GFS2_MAX_META_HEIGHT)) 291 goto corrupt; 292 ip->i_height = (u8)height; 293 294 depth = be16_to_cpu(str->di_depth); 295 if (unlikely(depth > GFS2_DIR_MAX_DEPTH)) 296 goto corrupt; 297 ip->i_depth = (u8)depth; 298 di->di_entries = be32_to_cpu(str->di_entries); 299 300 di->di_eattr = be64_to_cpu(str->di_eattr); 301 if (S_ISREG(ip->i_inode.i_mode)) 302 gfs2_set_aops(&ip->i_inode); 303 304 return 0; 305 corrupt: 306 if (gfs2_consist_inode(ip)) 307 gfs2_dinode_print(ip); 308 return -EIO; 309 } 310 311 /** 312 * gfs2_inode_refresh - Refresh the incore copy of the dinode 313 * @ip: The GFS2 inode 314 * 315 * Returns: errno 316 */ 317 318 int gfs2_inode_refresh(struct gfs2_inode *ip) 319 { 320 struct buffer_head *dibh; 321 int error; 322 323 error = gfs2_meta_inode_buffer(ip, &dibh); 324 if (error) 325 return error; 326 327 if (gfs2_metatype_check(GFS2_SB(&ip->i_inode), dibh, GFS2_METATYPE_DI)) { 328 brelse(dibh); 329 return -EIO; 330 } 331 332 error = gfs2_dinode_in(ip, dibh->b_data); 333 brelse(dibh); 334 clear_bit(GIF_INVALID, &ip->i_flags); 335 336 return error; 337 } 338 339 int gfs2_dinode_dealloc(struct gfs2_inode *ip) 340 { 341 struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode); 342 struct gfs2_alloc *al; 343 struct gfs2_rgrpd *rgd; 344 int error; 345 346 if (gfs2_get_inode_blocks(&ip->i_inode) != 1) { 347 if (gfs2_consist_inode(ip)) 348 gfs2_dinode_print(ip); 349 return -EIO; 350 } 351 352 al = gfs2_alloc_get(ip); 353 354 error = gfs2_quota_hold(ip, NO_QUOTA_CHANGE, NO_QUOTA_CHANGE); 355 if (error) 356 goto out; 357 358 error = gfs2_rindex_hold(sdp, &al->al_ri_gh); 359 if (error) 360 goto out_qs; 361 362 rgd = gfs2_blk2rgrpd(sdp, ip->i_no_addr); 363 if (!rgd) { 364 gfs2_consist_inode(ip); 365 error = -EIO; 366 goto out_rindex_relse; 367 } 368 369 error = gfs2_glock_nq_init(rgd->rd_gl, LM_ST_EXCLUSIVE, 0, 370 &al->al_rgd_gh); 371 if (error) 372 goto out_rindex_relse; 373 374 error = gfs2_trans_begin(sdp, RES_RG_BIT + RES_STATFS + RES_QUOTA, 1); 375 if (error) 376 goto out_rg_gunlock; 377 378 set_bit(GLF_DIRTY, &ip->i_gl->gl_flags); 379 set_bit(GLF_LFLUSH, &ip->i_gl->gl_flags); 380 381 gfs2_free_di(rgd, ip); 382 383 gfs2_trans_end(sdp); 384 clear_bit(GLF_STICKY, &ip->i_gl->gl_flags); 385 386 out_rg_gunlock: 387 gfs2_glock_dq_uninit(&al->al_rgd_gh); 388 out_rindex_relse: 389 gfs2_glock_dq_uninit(&al->al_ri_gh); 390 out_qs: 391 gfs2_quota_unhold(ip); 392 out: 393 gfs2_alloc_put(ip); 394 return error; 395 } 396 397 /** 398 * gfs2_change_nlink - Change nlink count on inode 399 * @ip: The GFS2 inode 400 * @diff: The change in the nlink count required 401 * 402 * Returns: errno 403 */ 404 int gfs2_change_nlink(struct gfs2_inode *ip, int diff) 405 { 406 struct buffer_head *dibh; 407 u32 nlink; 408 int error; 409 410 BUG_ON(diff != 1 && diff != -1); 411 nlink = ip->i_inode.i_nlink + diff; 412 413 /* If we are reducing the nlink count, but the new value ends up being 414 bigger than the old one, we must have underflowed. */ 415 if (diff < 0 && nlink > ip->i_inode.i_nlink) { 416 if (gfs2_consist_inode(ip)) 417 gfs2_dinode_print(ip); 418 return -EIO; 419 } 420 421 error = gfs2_meta_inode_buffer(ip, &dibh); 422 if (error) 423 return error; 424 425 if (diff > 0) 426 inc_nlink(&ip->i_inode); 427 else 428 drop_nlink(&ip->i_inode); 429 430 ip->i_inode.i_ctime = CURRENT_TIME; 431 432 gfs2_trans_add_bh(ip->i_gl, dibh, 1); 433 gfs2_dinode_out(ip, dibh->b_data); 434 brelse(dibh); 435 mark_inode_dirty(&ip->i_inode); 436 437 if (ip->i_inode.i_nlink == 0) 438 gfs2_unlink_di(&ip->i_inode); /* mark inode unlinked */ 439 440 return error; 441 } 442 443 struct inode *gfs2_lookup_simple(struct inode *dip, const char *name) 444 { 445 struct qstr qstr; 446 struct inode *inode; 447 gfs2_str2qstr(&qstr, name); 448 inode = gfs2_lookupi(dip, &qstr, 1, NULL); 449 /* gfs2_lookupi has inconsistent callers: vfs 450 * related routines expect NULL for no entry found, 451 * gfs2_lookup_simple callers expect ENOENT 452 * and do not check for NULL. 453 */ 454 if (inode == NULL) 455 return ERR_PTR(-ENOENT); 456 else 457 return inode; 458 } 459 460 461 /** 462 * gfs2_lookupi - Look up a filename in a directory and return its inode 463 * @d_gh: An initialized holder for the directory glock 464 * @name: The name of the inode to look for 465 * @is_root: If 1, ignore the caller's permissions 466 * @i_gh: An uninitialized holder for the new inode glock 467 * 468 * This can be called via the VFS filldir function when NFS is doing 469 * a readdirplus and the inode which its intending to stat isn't 470 * already in cache. In this case we must not take the directory glock 471 * again, since the readdir call will have already taken that lock. 472 * 473 * Returns: errno 474 */ 475 476 struct inode *gfs2_lookupi(struct inode *dir, const struct qstr *name, 477 int is_root, struct nameidata *nd) 478 { 479 struct super_block *sb = dir->i_sb; 480 struct gfs2_inode *dip = GFS2_I(dir); 481 struct gfs2_holder d_gh; 482 int error = 0; 483 struct inode *inode = NULL; 484 int unlock = 0; 485 486 if (!name->len || name->len > GFS2_FNAMESIZE) 487 return ERR_PTR(-ENAMETOOLONG); 488 489 if ((name->len == 1 && memcmp(name->name, ".", 1) == 0) || 490 (name->len == 2 && memcmp(name->name, "..", 2) == 0 && 491 dir == sb->s_root->d_inode)) { 492 igrab(dir); 493 return dir; 494 } 495 496 if (gfs2_glock_is_locked_by_me(dip->i_gl) == NULL) { 497 error = gfs2_glock_nq_init(dip->i_gl, LM_ST_SHARED, 0, &d_gh); 498 if (error) 499 return ERR_PTR(error); 500 unlock = 1; 501 } 502 503 if (!is_root) { 504 error = permission(dir, MAY_EXEC, NULL); 505 if (error) 506 goto out; 507 } 508 509 inode = gfs2_dir_search(dir, name); 510 if (IS_ERR(inode)) 511 error = PTR_ERR(inode); 512 out: 513 if (unlock) 514 gfs2_glock_dq_uninit(&d_gh); 515 if (error == -ENOENT) 516 return NULL; 517 return inode ? inode : ERR_PTR(error); 518 } 519 520 static void gfs2_inum_range_in(struct gfs2_inum_range_host *ir, const void *buf) 521 { 522 const struct gfs2_inum_range *str = buf; 523 524 ir->ir_start = be64_to_cpu(str->ir_start); 525 ir->ir_length = be64_to_cpu(str->ir_length); 526 } 527 528 static void gfs2_inum_range_out(const struct gfs2_inum_range_host *ir, void *buf) 529 { 530 struct gfs2_inum_range *str = buf; 531 532 str->ir_start = cpu_to_be64(ir->ir_start); 533 str->ir_length = cpu_to_be64(ir->ir_length); 534 } 535 536 static int pick_formal_ino_1(struct gfs2_sbd *sdp, u64 *formal_ino) 537 { 538 struct gfs2_inode *ip = GFS2_I(sdp->sd_ir_inode); 539 struct buffer_head *bh; 540 struct gfs2_inum_range_host ir; 541 int error; 542 543 error = gfs2_trans_begin(sdp, RES_DINODE, 0); 544 if (error) 545 return error; 546 mutex_lock(&sdp->sd_inum_mutex); 547 548 error = gfs2_meta_inode_buffer(ip, &bh); 549 if (error) { 550 mutex_unlock(&sdp->sd_inum_mutex); 551 gfs2_trans_end(sdp); 552 return error; 553 } 554 555 gfs2_inum_range_in(&ir, bh->b_data + sizeof(struct gfs2_dinode)); 556 557 if (ir.ir_length) { 558 *formal_ino = ir.ir_start++; 559 ir.ir_length--; 560 gfs2_trans_add_bh(ip->i_gl, bh, 1); 561 gfs2_inum_range_out(&ir, 562 bh->b_data + sizeof(struct gfs2_dinode)); 563 brelse(bh); 564 mutex_unlock(&sdp->sd_inum_mutex); 565 gfs2_trans_end(sdp); 566 return 0; 567 } 568 569 brelse(bh); 570 571 mutex_unlock(&sdp->sd_inum_mutex); 572 gfs2_trans_end(sdp); 573 574 return 1; 575 } 576 577 static int pick_formal_ino_2(struct gfs2_sbd *sdp, u64 *formal_ino) 578 { 579 struct gfs2_inode *ip = GFS2_I(sdp->sd_ir_inode); 580 struct gfs2_inode *m_ip = GFS2_I(sdp->sd_inum_inode); 581 struct gfs2_holder gh; 582 struct buffer_head *bh; 583 struct gfs2_inum_range_host ir; 584 int error; 585 586 error = gfs2_glock_nq_init(m_ip->i_gl, LM_ST_EXCLUSIVE, 0, &gh); 587 if (error) 588 return error; 589 590 error = gfs2_trans_begin(sdp, 2 * RES_DINODE, 0); 591 if (error) 592 goto out; 593 mutex_lock(&sdp->sd_inum_mutex); 594 595 error = gfs2_meta_inode_buffer(ip, &bh); 596 if (error) 597 goto out_end_trans; 598 599 gfs2_inum_range_in(&ir, bh->b_data + sizeof(struct gfs2_dinode)); 600 601 if (!ir.ir_length) { 602 struct buffer_head *m_bh; 603 u64 x, y; 604 __be64 z; 605 606 error = gfs2_meta_inode_buffer(m_ip, &m_bh); 607 if (error) 608 goto out_brelse; 609 610 z = *(__be64 *)(m_bh->b_data + sizeof(struct gfs2_dinode)); 611 x = y = be64_to_cpu(z); 612 ir.ir_start = x; 613 ir.ir_length = GFS2_INUM_QUANTUM; 614 x += GFS2_INUM_QUANTUM; 615 if (x < y) 616 gfs2_consist_inode(m_ip); 617 z = cpu_to_be64(x); 618 gfs2_trans_add_bh(m_ip->i_gl, m_bh, 1); 619 *(__be64 *)(m_bh->b_data + sizeof(struct gfs2_dinode)) = z; 620 621 brelse(m_bh); 622 } 623 624 *formal_ino = ir.ir_start++; 625 ir.ir_length--; 626 627 gfs2_trans_add_bh(ip->i_gl, bh, 1); 628 gfs2_inum_range_out(&ir, bh->b_data + sizeof(struct gfs2_dinode)); 629 630 out_brelse: 631 brelse(bh); 632 out_end_trans: 633 mutex_unlock(&sdp->sd_inum_mutex); 634 gfs2_trans_end(sdp); 635 out: 636 gfs2_glock_dq_uninit(&gh); 637 return error; 638 } 639 640 static int pick_formal_ino(struct gfs2_sbd *sdp, u64 *inum) 641 { 642 int error; 643 644 error = pick_formal_ino_1(sdp, inum); 645 if (error <= 0) 646 return error; 647 648 error = pick_formal_ino_2(sdp, inum); 649 650 return error; 651 } 652 653 /** 654 * create_ok - OK to create a new on-disk inode here? 655 * @dip: Directory in which dinode is to be created 656 * @name: Name of new dinode 657 * @mode: 658 * 659 * Returns: errno 660 */ 661 662 static int create_ok(struct gfs2_inode *dip, const struct qstr *name, 663 unsigned int mode) 664 { 665 int error; 666 667 error = permission(&dip->i_inode, MAY_WRITE | MAY_EXEC, NULL); 668 if (error) 669 return error; 670 671 /* Don't create entries in an unlinked directory */ 672 if (!dip->i_inode.i_nlink) 673 return -EPERM; 674 675 error = gfs2_dir_check(&dip->i_inode, name, NULL); 676 switch (error) { 677 case -ENOENT: 678 error = 0; 679 break; 680 case 0: 681 return -EEXIST; 682 default: 683 return error; 684 } 685 686 if (dip->i_di.di_entries == (u32)-1) 687 return -EFBIG; 688 if (S_ISDIR(mode) && dip->i_inode.i_nlink == (u32)-1) 689 return -EMLINK; 690 691 return 0; 692 } 693 694 static void munge_mode_uid_gid(struct gfs2_inode *dip, unsigned int *mode, 695 unsigned int *uid, unsigned int *gid) 696 { 697 if (GFS2_SB(&dip->i_inode)->sd_args.ar_suiddir && 698 (dip->i_inode.i_mode & S_ISUID) && dip->i_inode.i_uid) { 699 if (S_ISDIR(*mode)) 700 *mode |= S_ISUID; 701 else if (dip->i_inode.i_uid != current->fsuid) 702 *mode &= ~07111; 703 *uid = dip->i_inode.i_uid; 704 } else 705 *uid = current->fsuid; 706 707 if (dip->i_inode.i_mode & S_ISGID) { 708 if (S_ISDIR(*mode)) 709 *mode |= S_ISGID; 710 *gid = dip->i_inode.i_gid; 711 } else 712 *gid = current->fsgid; 713 } 714 715 static int alloc_dinode(struct gfs2_inode *dip, u64 *no_addr, u64 *generation) 716 { 717 struct gfs2_sbd *sdp = GFS2_SB(&dip->i_inode); 718 int error; 719 720 if (gfs2_alloc_get(dip) == NULL) 721 return -ENOMEM; 722 723 dip->i_alloc->al_requested = RES_DINODE; 724 error = gfs2_inplace_reserve(dip); 725 if (error) 726 goto out; 727 728 error = gfs2_trans_begin(sdp, RES_RG_BIT + RES_STATFS, 0); 729 if (error) 730 goto out_ipreserv; 731 732 *no_addr = gfs2_alloc_di(dip, generation); 733 734 gfs2_trans_end(sdp); 735 736 out_ipreserv: 737 gfs2_inplace_release(dip); 738 out: 739 gfs2_alloc_put(dip); 740 return error; 741 } 742 743 /** 744 * init_dinode - Fill in a new dinode structure 745 * @dip: the directory this inode is being created in 746 * @gl: The glock covering the new inode 747 * @inum: the inode number 748 * @mode: the file permissions 749 * @uid: 750 * @gid: 751 * 752 */ 753 754 static void init_dinode(struct gfs2_inode *dip, struct gfs2_glock *gl, 755 const struct gfs2_inum_host *inum, unsigned int mode, 756 unsigned int uid, unsigned int gid, 757 const u64 *generation, dev_t dev, struct buffer_head **bhp) 758 { 759 struct gfs2_sbd *sdp = GFS2_SB(&dip->i_inode); 760 struct gfs2_dinode *di; 761 struct buffer_head *dibh; 762 struct timespec tv = CURRENT_TIME; 763 764 dibh = gfs2_meta_new(gl, inum->no_addr); 765 gfs2_trans_add_bh(gl, dibh, 1); 766 gfs2_metatype_set(dibh, GFS2_METATYPE_DI, GFS2_FORMAT_DI); 767 gfs2_buffer_clear_tail(dibh, sizeof(struct gfs2_dinode)); 768 di = (struct gfs2_dinode *)dibh->b_data; 769 770 di->di_num.no_formal_ino = cpu_to_be64(inum->no_formal_ino); 771 di->di_num.no_addr = cpu_to_be64(inum->no_addr); 772 di->di_mode = cpu_to_be32(mode); 773 di->di_uid = cpu_to_be32(uid); 774 di->di_gid = cpu_to_be32(gid); 775 di->di_nlink = 0; 776 di->di_size = 0; 777 di->di_blocks = cpu_to_be64(1); 778 di->di_atime = di->di_mtime = di->di_ctime = cpu_to_be64(tv.tv_sec); 779 di->di_major = cpu_to_be32(MAJOR(dev)); 780 di->di_minor = cpu_to_be32(MINOR(dev)); 781 di->di_goal_meta = di->di_goal_data = cpu_to_be64(inum->no_addr); 782 di->di_generation = cpu_to_be64(*generation); 783 di->di_flags = 0; 784 785 if (S_ISREG(mode)) { 786 if ((dip->i_di.di_flags & GFS2_DIF_INHERIT_JDATA) || 787 gfs2_tune_get(sdp, gt_new_files_jdata)) 788 di->di_flags |= cpu_to_be32(GFS2_DIF_JDATA); 789 if ((dip->i_di.di_flags & GFS2_DIF_INHERIT_DIRECTIO) || 790 gfs2_tune_get(sdp, gt_new_files_directio)) 791 di->di_flags |= cpu_to_be32(GFS2_DIF_DIRECTIO); 792 } else if (S_ISDIR(mode)) { 793 di->di_flags |= cpu_to_be32(dip->i_di.di_flags & 794 GFS2_DIF_INHERIT_DIRECTIO); 795 di->di_flags |= cpu_to_be32(dip->i_di.di_flags & 796 GFS2_DIF_INHERIT_JDATA); 797 } 798 799 di->__pad1 = 0; 800 di->di_payload_format = cpu_to_be32(S_ISDIR(mode) ? GFS2_FORMAT_DE : 0); 801 di->di_height = 0; 802 di->__pad2 = 0; 803 di->__pad3 = 0; 804 di->di_depth = 0; 805 di->di_entries = 0; 806 memset(&di->__pad4, 0, sizeof(di->__pad4)); 807 di->di_eattr = 0; 808 di->di_atime_nsec = cpu_to_be32(tv.tv_nsec); 809 di->di_mtime_nsec = cpu_to_be32(tv.tv_nsec); 810 di->di_ctime_nsec = cpu_to_be32(tv.tv_nsec); 811 memset(&di->di_reserved, 0, sizeof(di->di_reserved)); 812 813 set_buffer_uptodate(dibh); 814 815 *bhp = dibh; 816 } 817 818 static int make_dinode(struct gfs2_inode *dip, struct gfs2_glock *gl, 819 unsigned int mode, const struct gfs2_inum_host *inum, 820 const u64 *generation, dev_t dev, struct buffer_head **bhp) 821 { 822 struct gfs2_sbd *sdp = GFS2_SB(&dip->i_inode); 823 unsigned int uid, gid; 824 int error; 825 826 munge_mode_uid_gid(dip, &mode, &uid, &gid); 827 gfs2_alloc_get(dip); 828 829 error = gfs2_quota_lock(dip, uid, gid); 830 if (error) 831 goto out; 832 833 error = gfs2_quota_check(dip, uid, gid); 834 if (error) 835 goto out_quota; 836 837 error = gfs2_trans_begin(sdp, RES_DINODE + RES_QUOTA, 0); 838 if (error) 839 goto out_quota; 840 841 init_dinode(dip, gl, inum, mode, uid, gid, generation, dev, bhp); 842 gfs2_quota_change(dip, +1, uid, gid); 843 gfs2_trans_end(sdp); 844 845 out_quota: 846 gfs2_quota_unlock(dip); 847 out: 848 gfs2_alloc_put(dip); 849 return error; 850 } 851 852 static int link_dinode(struct gfs2_inode *dip, const struct qstr *name, 853 struct gfs2_inode *ip) 854 { 855 struct gfs2_sbd *sdp = GFS2_SB(&dip->i_inode); 856 struct gfs2_alloc *al; 857 int alloc_required; 858 struct buffer_head *dibh; 859 int error; 860 861 al = gfs2_alloc_get(dip); 862 863 error = gfs2_quota_lock(dip, NO_QUOTA_CHANGE, NO_QUOTA_CHANGE); 864 if (error) 865 goto fail; 866 867 error = alloc_required = gfs2_diradd_alloc_required(&dip->i_inode, name); 868 if (alloc_required < 0) 869 goto fail_quota_locks; 870 if (alloc_required) { 871 error = gfs2_quota_check(dip, dip->i_inode.i_uid, dip->i_inode.i_gid); 872 if (error) 873 goto fail_quota_locks; 874 875 al->al_requested = sdp->sd_max_dirres; 876 877 error = gfs2_inplace_reserve(dip); 878 if (error) 879 goto fail_quota_locks; 880 881 error = gfs2_trans_begin(sdp, sdp->sd_max_dirres + 882 al->al_rgd->rd_length + 883 2 * RES_DINODE + 884 RES_STATFS + RES_QUOTA, 0); 885 if (error) 886 goto fail_ipreserv; 887 } else { 888 error = gfs2_trans_begin(sdp, RES_LEAF + 2 * RES_DINODE, 0); 889 if (error) 890 goto fail_quota_locks; 891 } 892 893 error = gfs2_dir_add(&dip->i_inode, name, ip, IF2DT(ip->i_inode.i_mode)); 894 if (error) 895 goto fail_end_trans; 896 897 error = gfs2_meta_inode_buffer(ip, &dibh); 898 if (error) 899 goto fail_end_trans; 900 ip->i_inode.i_nlink = 1; 901 gfs2_trans_add_bh(ip->i_gl, dibh, 1); 902 gfs2_dinode_out(ip, dibh->b_data); 903 brelse(dibh); 904 return 0; 905 906 fail_end_trans: 907 gfs2_trans_end(sdp); 908 909 fail_ipreserv: 910 if (dip->i_alloc->al_rgd) 911 gfs2_inplace_release(dip); 912 913 fail_quota_locks: 914 gfs2_quota_unlock(dip); 915 916 fail: 917 gfs2_alloc_put(dip); 918 return error; 919 } 920 921 static int gfs2_security_init(struct gfs2_inode *dip, struct gfs2_inode *ip) 922 { 923 int err; 924 size_t len; 925 void *value; 926 char *name; 927 struct gfs2_ea_request er; 928 929 err = security_inode_init_security(&ip->i_inode, &dip->i_inode, 930 &name, &value, &len); 931 932 if (err) { 933 if (err == -EOPNOTSUPP) 934 return 0; 935 return err; 936 } 937 938 memset(&er, 0, sizeof(struct gfs2_ea_request)); 939 940 er.er_type = GFS2_EATYPE_SECURITY; 941 er.er_name = name; 942 er.er_data = value; 943 er.er_name_len = strlen(name); 944 er.er_data_len = len; 945 946 err = gfs2_ea_set_i(ip, &er); 947 948 kfree(value); 949 kfree(name); 950 951 return err; 952 } 953 954 /** 955 * gfs2_createi - Create a new inode 956 * @ghs: An array of two holders 957 * @name: The name of the new file 958 * @mode: the permissions on the new inode 959 * 960 * @ghs[0] is an initialized holder for the directory 961 * @ghs[1] is the holder for the inode lock 962 * 963 * If the return value is not NULL, the glocks on both the directory and the new 964 * file are held. A transaction has been started and an inplace reservation 965 * is held, as well. 966 * 967 * Returns: An inode 968 */ 969 970 struct inode *gfs2_createi(struct gfs2_holder *ghs, const struct qstr *name, 971 unsigned int mode, dev_t dev) 972 { 973 struct inode *inode = NULL; 974 struct gfs2_inode *dip = ghs->gh_gl->gl_object; 975 struct inode *dir = &dip->i_inode; 976 struct gfs2_sbd *sdp = GFS2_SB(&dip->i_inode); 977 struct gfs2_inum_host inum = { .no_addr = 0, .no_formal_ino = 0 }; 978 int error; 979 u64 generation; 980 struct buffer_head *bh = NULL; 981 982 if (!name->len || name->len > GFS2_FNAMESIZE) 983 return ERR_PTR(-ENAMETOOLONG); 984 985 gfs2_holder_reinit(LM_ST_EXCLUSIVE, 0, ghs); 986 error = gfs2_glock_nq(ghs); 987 if (error) 988 goto fail; 989 990 error = create_ok(dip, name, mode); 991 if (error) 992 goto fail_gunlock; 993 994 error = pick_formal_ino(sdp, &inum.no_formal_ino); 995 if (error) 996 goto fail_gunlock; 997 998 error = alloc_dinode(dip, &inum.no_addr, &generation); 999 if (error) 1000 goto fail_gunlock; 1001 1002 error = gfs2_glock_nq_num(sdp, inum.no_addr, &gfs2_inode_glops, 1003 LM_ST_EXCLUSIVE, GL_SKIP, ghs + 1); 1004 if (error) 1005 goto fail_gunlock; 1006 1007 error = make_dinode(dip, ghs[1].gh_gl, mode, &inum, &generation, dev, &bh); 1008 if (error) 1009 goto fail_gunlock2; 1010 1011 inode = gfs2_inode_lookup(dir->i_sb, IF2DT(mode), 1012 inum.no_addr, 1013 inum.no_formal_ino, 0); 1014 if (IS_ERR(inode)) 1015 goto fail_gunlock2; 1016 1017 error = gfs2_inode_refresh(GFS2_I(inode)); 1018 if (error) 1019 goto fail_gunlock2; 1020 1021 error = gfs2_acl_create(dip, GFS2_I(inode)); 1022 if (error) 1023 goto fail_gunlock2; 1024 1025 error = gfs2_security_init(dip, GFS2_I(inode)); 1026 if (error) 1027 goto fail_gunlock2; 1028 1029 error = link_dinode(dip, name, GFS2_I(inode)); 1030 if (error) 1031 goto fail_gunlock2; 1032 1033 if (bh) 1034 brelse(bh); 1035 if (!inode) 1036 return ERR_PTR(-ENOMEM); 1037 return inode; 1038 1039 fail_gunlock2: 1040 gfs2_glock_dq_uninit(ghs + 1); 1041 if (inode) 1042 iput(inode); 1043 fail_gunlock: 1044 gfs2_glock_dq(ghs); 1045 fail: 1046 if (bh) 1047 brelse(bh); 1048 return ERR_PTR(error); 1049 } 1050 1051 /** 1052 * gfs2_rmdiri - Remove a directory 1053 * @dip: The parent directory of the directory to be removed 1054 * @name: The name of the directory to be removed 1055 * @ip: The GFS2 inode of the directory to be removed 1056 * 1057 * Assumes Glocks on dip and ip are held 1058 * 1059 * Returns: errno 1060 */ 1061 1062 int gfs2_rmdiri(struct gfs2_inode *dip, const struct qstr *name, 1063 struct gfs2_inode *ip) 1064 { 1065 struct qstr dotname; 1066 int error; 1067 1068 if (ip->i_di.di_entries != 2) { 1069 if (gfs2_consist_inode(ip)) 1070 gfs2_dinode_print(ip); 1071 return -EIO; 1072 } 1073 1074 error = gfs2_dir_del(dip, name); 1075 if (error) 1076 return error; 1077 1078 error = gfs2_change_nlink(dip, -1); 1079 if (error) 1080 return error; 1081 1082 gfs2_str2qstr(&dotname, "."); 1083 error = gfs2_dir_del(ip, &dotname); 1084 if (error) 1085 return error; 1086 1087 gfs2_str2qstr(&dotname, ".."); 1088 error = gfs2_dir_del(ip, &dotname); 1089 if (error) 1090 return error; 1091 1092 /* It looks odd, but it really should be done twice */ 1093 error = gfs2_change_nlink(ip, -1); 1094 if (error) 1095 return error; 1096 1097 error = gfs2_change_nlink(ip, -1); 1098 if (error) 1099 return error; 1100 1101 return error; 1102 } 1103 1104 /* 1105 * gfs2_unlink_ok - check to see that a inode is still in a directory 1106 * @dip: the directory 1107 * @name: the name of the file 1108 * @ip: the inode 1109 * 1110 * Assumes that the lock on (at least) @dip is held. 1111 * 1112 * Returns: 0 if the parent/child relationship is correct, errno if it isn't 1113 */ 1114 1115 int gfs2_unlink_ok(struct gfs2_inode *dip, const struct qstr *name, 1116 const struct gfs2_inode *ip) 1117 { 1118 int error; 1119 1120 if (IS_IMMUTABLE(&ip->i_inode) || IS_APPEND(&ip->i_inode)) 1121 return -EPERM; 1122 1123 if ((dip->i_inode.i_mode & S_ISVTX) && 1124 dip->i_inode.i_uid != current->fsuid && 1125 ip->i_inode.i_uid != current->fsuid && !capable(CAP_FOWNER)) 1126 return -EPERM; 1127 1128 if (IS_APPEND(&dip->i_inode)) 1129 return -EPERM; 1130 1131 error = permission(&dip->i_inode, MAY_WRITE | MAY_EXEC, NULL); 1132 if (error) 1133 return error; 1134 1135 error = gfs2_dir_check(&dip->i_inode, name, ip); 1136 if (error) 1137 return error; 1138 1139 return 0; 1140 } 1141 1142 /* 1143 * gfs2_ok_to_move - check if it's ok to move a directory to another directory 1144 * @this: move this 1145 * @to: to here 1146 * 1147 * Follow @to back to the root and make sure we don't encounter @this 1148 * Assumes we already hold the rename lock. 1149 * 1150 * Returns: errno 1151 */ 1152 1153 int gfs2_ok_to_move(struct gfs2_inode *this, struct gfs2_inode *to) 1154 { 1155 struct inode *dir = &to->i_inode; 1156 struct super_block *sb = dir->i_sb; 1157 struct inode *tmp; 1158 struct qstr dotdot; 1159 int error = 0; 1160 1161 gfs2_str2qstr(&dotdot, ".."); 1162 1163 igrab(dir); 1164 1165 for (;;) { 1166 if (dir == &this->i_inode) { 1167 error = -EINVAL; 1168 break; 1169 } 1170 if (dir == sb->s_root->d_inode) { 1171 error = 0; 1172 break; 1173 } 1174 1175 tmp = gfs2_lookupi(dir, &dotdot, 1, NULL); 1176 if (IS_ERR(tmp)) { 1177 error = PTR_ERR(tmp); 1178 break; 1179 } 1180 1181 iput(dir); 1182 dir = tmp; 1183 } 1184 1185 iput(dir); 1186 1187 return error; 1188 } 1189 1190 /** 1191 * gfs2_readlinki - return the contents of a symlink 1192 * @ip: the symlink's inode 1193 * @buf: a pointer to the buffer to be filled 1194 * @len: a pointer to the length of @buf 1195 * 1196 * If @buf is too small, a piece of memory is kmalloc()ed and needs 1197 * to be freed by the caller. 1198 * 1199 * Returns: errno 1200 */ 1201 1202 int gfs2_readlinki(struct gfs2_inode *ip, char **buf, unsigned int *len) 1203 { 1204 struct gfs2_holder i_gh; 1205 struct buffer_head *dibh; 1206 unsigned int x; 1207 int error; 1208 1209 gfs2_holder_init(ip->i_gl, LM_ST_SHARED, GL_ATIME, &i_gh); 1210 error = gfs2_glock_nq_atime(&i_gh); 1211 if (error) { 1212 gfs2_holder_uninit(&i_gh); 1213 return error; 1214 } 1215 1216 if (!ip->i_di.di_size) { 1217 gfs2_consist_inode(ip); 1218 error = -EIO; 1219 goto out; 1220 } 1221 1222 error = gfs2_meta_inode_buffer(ip, &dibh); 1223 if (error) 1224 goto out; 1225 1226 x = ip->i_di.di_size + 1; 1227 if (x > *len) { 1228 *buf = kmalloc(x, GFP_KERNEL); 1229 if (!*buf) { 1230 error = -ENOMEM; 1231 goto out_brelse; 1232 } 1233 } 1234 1235 memcpy(*buf, dibh->b_data + sizeof(struct gfs2_dinode), x); 1236 *len = x; 1237 1238 out_brelse: 1239 brelse(dibh); 1240 out: 1241 gfs2_glock_dq_uninit(&i_gh); 1242 return error; 1243 } 1244 1245 /** 1246 * gfs2_glock_nq_atime - Acquire a hold on an inode's glock, and 1247 * conditionally update the inode's atime 1248 * @gh: the holder to acquire 1249 * 1250 * Tests atime (access time) for gfs2_read, gfs2_readdir and gfs2_mmap 1251 * Update if the difference between the current time and the inode's current 1252 * atime is greater than an interval specified at mount. 1253 * 1254 * Returns: errno 1255 */ 1256 1257 int gfs2_glock_nq_atime(struct gfs2_holder *gh) 1258 { 1259 struct gfs2_glock *gl = gh->gh_gl; 1260 struct gfs2_sbd *sdp = gl->gl_sbd; 1261 struct gfs2_inode *ip = gl->gl_object; 1262 s64 quantum = gfs2_tune_get(sdp, gt_atime_quantum); 1263 unsigned int state; 1264 int flags; 1265 int error; 1266 struct timespec tv = CURRENT_TIME; 1267 1268 if (gfs2_assert_warn(sdp, gh->gh_flags & GL_ATIME) || 1269 gfs2_assert_warn(sdp, !(gh->gh_flags & GL_ASYNC)) || 1270 gfs2_assert_warn(sdp, gl->gl_ops == &gfs2_inode_glops)) 1271 return -EINVAL; 1272 1273 state = gh->gh_state; 1274 flags = gh->gh_flags; 1275 1276 error = gfs2_glock_nq(gh); 1277 if (error) 1278 return error; 1279 1280 if (test_bit(SDF_NOATIME, &sdp->sd_flags) || 1281 (sdp->sd_vfs->s_flags & MS_RDONLY)) 1282 return 0; 1283 1284 if (tv.tv_sec - ip->i_inode.i_atime.tv_sec >= quantum) { 1285 gfs2_glock_dq(gh); 1286 gfs2_holder_reinit(LM_ST_EXCLUSIVE, gh->gh_flags & ~LM_FLAG_ANY, 1287 gh); 1288 error = gfs2_glock_nq(gh); 1289 if (error) 1290 return error; 1291 1292 /* Verify that atime hasn't been updated while we were 1293 trying to get exclusive lock. */ 1294 1295 tv = CURRENT_TIME; 1296 if (tv.tv_sec - ip->i_inode.i_atime.tv_sec >= quantum) { 1297 struct buffer_head *dibh; 1298 struct gfs2_dinode *di; 1299 1300 error = gfs2_trans_begin(sdp, RES_DINODE, 0); 1301 if (error == -EROFS) 1302 return 0; 1303 if (error) 1304 goto fail; 1305 1306 error = gfs2_meta_inode_buffer(ip, &dibh); 1307 if (error) 1308 goto fail_end_trans; 1309 1310 ip->i_inode.i_atime = tv; 1311 1312 gfs2_trans_add_bh(ip->i_gl, dibh, 1); 1313 di = (struct gfs2_dinode *)dibh->b_data; 1314 di->di_atime = cpu_to_be64(ip->i_inode.i_atime.tv_sec); 1315 di->di_atime_nsec = cpu_to_be32(ip->i_inode.i_atime.tv_nsec); 1316 brelse(dibh); 1317 1318 gfs2_trans_end(sdp); 1319 } 1320 1321 /* If someone else has asked for the glock, 1322 unlock and let them have it. Then reacquire 1323 in the original state. */ 1324 if (gfs2_glock_is_blocking(gl)) { 1325 gfs2_glock_dq(gh); 1326 gfs2_holder_reinit(state, flags, gh); 1327 return gfs2_glock_nq(gh); 1328 } 1329 } 1330 1331 return 0; 1332 1333 fail_end_trans: 1334 gfs2_trans_end(sdp); 1335 fail: 1336 gfs2_glock_dq(gh); 1337 return error; 1338 } 1339 1340 static int 1341 __gfs2_setattr_simple(struct gfs2_inode *ip, struct iattr *attr) 1342 { 1343 struct buffer_head *dibh; 1344 int error; 1345 1346 error = gfs2_meta_inode_buffer(ip, &dibh); 1347 if (!error) { 1348 error = inode_setattr(&ip->i_inode, attr); 1349 gfs2_assert_warn(GFS2_SB(&ip->i_inode), !error); 1350 gfs2_trans_add_bh(ip->i_gl, dibh, 1); 1351 gfs2_dinode_out(ip, dibh->b_data); 1352 brelse(dibh); 1353 } 1354 return error; 1355 } 1356 1357 /** 1358 * gfs2_setattr_simple - 1359 * @ip: 1360 * @attr: 1361 * 1362 * Called with a reference on the vnode. 1363 * 1364 * Returns: errno 1365 */ 1366 1367 int gfs2_setattr_simple(struct gfs2_inode *ip, struct iattr *attr) 1368 { 1369 int error; 1370 1371 if (current->journal_info) 1372 return __gfs2_setattr_simple(ip, attr); 1373 1374 error = gfs2_trans_begin(GFS2_SB(&ip->i_inode), RES_DINODE, 0); 1375 if (error) 1376 return error; 1377 1378 error = __gfs2_setattr_simple(ip, attr); 1379 gfs2_trans_end(GFS2_SB(&ip->i_inode)); 1380 return error; 1381 } 1382 1383 void gfs2_dinode_out(const struct gfs2_inode *ip, void *buf) 1384 { 1385 const struct gfs2_dinode_host *di = &ip->i_di; 1386 struct gfs2_dinode *str = buf; 1387 1388 str->di_header.mh_magic = cpu_to_be32(GFS2_MAGIC); 1389 str->di_header.mh_type = cpu_to_be32(GFS2_METATYPE_DI); 1390 str->di_header.__pad0 = 0; 1391 str->di_header.mh_format = cpu_to_be32(GFS2_FORMAT_DI); 1392 str->di_header.__pad1 = 0; 1393 str->di_num.no_addr = cpu_to_be64(ip->i_no_addr); 1394 str->di_num.no_formal_ino = cpu_to_be64(ip->i_no_formal_ino); 1395 str->di_mode = cpu_to_be32(ip->i_inode.i_mode); 1396 str->di_uid = cpu_to_be32(ip->i_inode.i_uid); 1397 str->di_gid = cpu_to_be32(ip->i_inode.i_gid); 1398 str->di_nlink = cpu_to_be32(ip->i_inode.i_nlink); 1399 str->di_size = cpu_to_be64(di->di_size); 1400 str->di_blocks = cpu_to_be64(gfs2_get_inode_blocks(&ip->i_inode)); 1401 str->di_atime = cpu_to_be64(ip->i_inode.i_atime.tv_sec); 1402 str->di_mtime = cpu_to_be64(ip->i_inode.i_mtime.tv_sec); 1403 str->di_ctime = cpu_to_be64(ip->i_inode.i_ctime.tv_sec); 1404 1405 str->di_goal_meta = cpu_to_be64(ip->i_goal); 1406 str->di_goal_data = cpu_to_be64(ip->i_goal); 1407 str->di_generation = cpu_to_be64(di->di_generation); 1408 1409 str->di_flags = cpu_to_be32(di->di_flags); 1410 str->di_height = cpu_to_be16(ip->i_height); 1411 str->di_payload_format = cpu_to_be32(S_ISDIR(ip->i_inode.i_mode) && 1412 !(ip->i_di.di_flags & GFS2_DIF_EXHASH) ? 1413 GFS2_FORMAT_DE : 0); 1414 str->di_depth = cpu_to_be16(ip->i_depth); 1415 str->di_entries = cpu_to_be32(di->di_entries); 1416 1417 str->di_eattr = cpu_to_be64(di->di_eattr); 1418 str->di_atime_nsec = cpu_to_be32(ip->i_inode.i_atime.tv_nsec); 1419 str->di_mtime_nsec = cpu_to_be32(ip->i_inode.i_mtime.tv_nsec); 1420 str->di_ctime_nsec = cpu_to_be32(ip->i_inode.i_ctime.tv_nsec); 1421 } 1422 1423 void gfs2_dinode_print(const struct gfs2_inode *ip) 1424 { 1425 const struct gfs2_dinode_host *di = &ip->i_di; 1426 1427 printk(KERN_INFO " no_formal_ino = %llu\n", 1428 (unsigned long long)ip->i_no_formal_ino); 1429 printk(KERN_INFO " no_addr = %llu\n", 1430 (unsigned long long)ip->i_no_addr); 1431 printk(KERN_INFO " di_size = %llu\n", (unsigned long long)di->di_size); 1432 printk(KERN_INFO " blocks = %llu\n", 1433 (unsigned long long)gfs2_get_inode_blocks(&ip->i_inode)); 1434 printk(KERN_INFO " i_goal = %llu\n", 1435 (unsigned long long)ip->i_goal); 1436 printk(KERN_INFO " di_flags = 0x%.8X\n", di->di_flags); 1437 printk(KERN_INFO " i_height = %u\n", ip->i_height); 1438 printk(KERN_INFO " i_depth = %u\n", ip->i_depth); 1439 printk(KERN_INFO " di_entries = %u\n", di->di_entries); 1440 printk(KERN_INFO " di_eattr = %llu\n", 1441 (unsigned long long)di->di_eattr); 1442 } 1443 1444