1 /* 2 * Copyright (C) Sistina Software, Inc. 1997-2003 All rights reserved. 3 * Copyright (C) 2004-2006 Red Hat, Inc. All rights reserved. 4 * 5 * This copyrighted material is made available to anyone wishing to use, 6 * modify, copy, or redistribute it subject to the terms and conditions 7 * of the GNU General Public License version 2. 8 */ 9 10 /* 11 * Implements Extendible Hashing as described in: 12 * "Extendible Hashing" by Fagin, et al in 13 * __ACM Trans. on Database Systems__, Sept 1979. 14 * 15 * 16 * Here's the layout of dirents which is essentially the same as that of ext2 17 * within a single block. The field de_name_len is the number of bytes 18 * actually required for the name (no null terminator). The field de_rec_len 19 * is the number of bytes allocated to the dirent. The offset of the next 20 * dirent in the block is (dirent + dirent->de_rec_len). When a dirent is 21 * deleted, the preceding dirent inherits its allocated space, ie 22 * prev->de_rec_len += deleted->de_rec_len. Since the next dirent is obtained 23 * by adding de_rec_len to the current dirent, this essentially causes the 24 * deleted dirent to get jumped over when iterating through all the dirents. 25 * 26 * When deleting the first dirent in a block, there is no previous dirent so 27 * the field de_ino is set to zero to designate it as deleted. When allocating 28 * a dirent, gfs2_dirent_alloc iterates through the dirents in a block. If the 29 * first dirent has (de_ino == 0) and de_rec_len is large enough, this first 30 * dirent is allocated. Otherwise it must go through all the 'used' dirents 31 * searching for one in which the amount of total space minus the amount of 32 * used space will provide enough space for the new dirent. 33 * 34 * There are two types of blocks in which dirents reside. In a stuffed dinode, 35 * the dirents begin at offset sizeof(struct gfs2_dinode) from the beginning of 36 * the block. In leaves, they begin at offset sizeof(struct gfs2_leaf) from the 37 * beginning of the leaf block. The dirents reside in leaves when 38 * 39 * dip->i_diskflags & GFS2_DIF_EXHASH is true 40 * 41 * Otherwise, the dirents are "linear", within a single stuffed dinode block. 42 * 43 * When the dirents are in leaves, the actual contents of the directory file are 44 * used as an array of 64-bit block pointers pointing to the leaf blocks. The 45 * dirents are NOT in the directory file itself. There can be more than one 46 * block pointer in the array that points to the same leaf. In fact, when a 47 * directory is first converted from linear to exhash, all of the pointers 48 * point to the same leaf. 49 * 50 * When a leaf is completely full, the size of the hash table can be 51 * doubled unless it is already at the maximum size which is hard coded into 52 * GFS2_DIR_MAX_DEPTH. After that, leaves are chained together in a linked list, 53 * but never before the maximum hash table size has been reached. 54 */ 55 56 #define pr_fmt(fmt) KBUILD_MODNAME ": " fmt 57 58 #include <linux/slab.h> 59 #include <linux/spinlock.h> 60 #include <linux/buffer_head.h> 61 #include <linux/sort.h> 62 #include <linux/gfs2_ondisk.h> 63 #include <linux/crc32.h> 64 #include <linux/vmalloc.h> 65 66 #include "gfs2.h" 67 #include "incore.h" 68 #include "dir.h" 69 #include "glock.h" 70 #include "inode.h" 71 #include "meta_io.h" 72 #include "quota.h" 73 #include "rgrp.h" 74 #include "trans.h" 75 #include "bmap.h" 76 #include "util.h" 77 78 #define IS_LEAF 1 /* Hashed (leaf) directory */ 79 #define IS_DINODE 2 /* Linear (stuffed dinode block) directory */ 80 81 #define MAX_RA_BLOCKS 32 /* max read-ahead blocks */ 82 83 #define gfs2_disk_hash2offset(h) (((u64)(h)) >> 1) 84 #define gfs2_dir_offset2hash(p) ((u32)(((u64)(p)) << 1)) 85 86 struct qstr gfs2_qdot __read_mostly; 87 struct qstr gfs2_qdotdot __read_mostly; 88 89 typedef int (*gfs2_dscan_t)(const struct gfs2_dirent *dent, 90 const struct qstr *name, void *opaque); 91 92 int gfs2_dir_get_new_buffer(struct gfs2_inode *ip, u64 block, 93 struct buffer_head **bhp) 94 { 95 struct buffer_head *bh; 96 97 bh = gfs2_meta_new(ip->i_gl, block); 98 gfs2_trans_add_meta(ip->i_gl, bh); 99 gfs2_metatype_set(bh, GFS2_METATYPE_JD, GFS2_FORMAT_JD); 100 gfs2_buffer_clear_tail(bh, sizeof(struct gfs2_meta_header)); 101 *bhp = bh; 102 return 0; 103 } 104 105 static int gfs2_dir_get_existing_buffer(struct gfs2_inode *ip, u64 block, 106 struct buffer_head **bhp) 107 { 108 struct buffer_head *bh; 109 int error; 110 111 error = gfs2_meta_read(ip->i_gl, block, DIO_WAIT, &bh); 112 if (error) 113 return error; 114 if (gfs2_metatype_check(GFS2_SB(&ip->i_inode), bh, GFS2_METATYPE_JD)) { 115 brelse(bh); 116 return -EIO; 117 } 118 *bhp = bh; 119 return 0; 120 } 121 122 static int gfs2_dir_write_stuffed(struct gfs2_inode *ip, const char *buf, 123 unsigned int offset, unsigned int size) 124 { 125 struct buffer_head *dibh; 126 int error; 127 128 error = gfs2_meta_inode_buffer(ip, &dibh); 129 if (error) 130 return error; 131 132 gfs2_trans_add_meta(ip->i_gl, dibh); 133 memcpy(dibh->b_data + offset + sizeof(struct gfs2_dinode), buf, size); 134 if (ip->i_inode.i_size < offset + size) 135 i_size_write(&ip->i_inode, offset + size); 136 ip->i_inode.i_mtime = ip->i_inode.i_ctime = CURRENT_TIME; 137 gfs2_dinode_out(ip, dibh->b_data); 138 139 brelse(dibh); 140 141 return size; 142 } 143 144 145 146 /** 147 * gfs2_dir_write_data - Write directory information to the inode 148 * @ip: The GFS2 inode 149 * @buf: The buffer containing information to be written 150 * @offset: The file offset to start writing at 151 * @size: The amount of data to write 152 * 153 * Returns: The number of bytes correctly written or error code 154 */ 155 static int gfs2_dir_write_data(struct gfs2_inode *ip, const char *buf, 156 u64 offset, unsigned int size) 157 { 158 struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode); 159 struct buffer_head *dibh; 160 u64 lblock, dblock; 161 u32 extlen = 0; 162 unsigned int o; 163 int copied = 0; 164 int error = 0; 165 int new = 0; 166 167 if (!size) 168 return 0; 169 170 if (gfs2_is_stuffed(ip) && 171 offset + size <= sdp->sd_sb.sb_bsize - sizeof(struct gfs2_dinode)) 172 return gfs2_dir_write_stuffed(ip, buf, (unsigned int)offset, 173 size); 174 175 if (gfs2_assert_warn(sdp, gfs2_is_jdata(ip))) 176 return -EINVAL; 177 178 if (gfs2_is_stuffed(ip)) { 179 error = gfs2_unstuff_dinode(ip, NULL); 180 if (error) 181 return error; 182 } 183 184 lblock = offset; 185 o = do_div(lblock, sdp->sd_jbsize) + sizeof(struct gfs2_meta_header); 186 187 while (copied < size) { 188 unsigned int amount; 189 struct buffer_head *bh; 190 191 amount = size - copied; 192 if (amount > sdp->sd_sb.sb_bsize - o) 193 amount = sdp->sd_sb.sb_bsize - o; 194 195 if (!extlen) { 196 new = 1; 197 error = gfs2_extent_map(&ip->i_inode, lblock, &new, 198 &dblock, &extlen); 199 if (error) 200 goto fail; 201 error = -EIO; 202 if (gfs2_assert_withdraw(sdp, dblock)) 203 goto fail; 204 } 205 206 if (amount == sdp->sd_jbsize || new) 207 error = gfs2_dir_get_new_buffer(ip, dblock, &bh); 208 else 209 error = gfs2_dir_get_existing_buffer(ip, dblock, &bh); 210 211 if (error) 212 goto fail; 213 214 gfs2_trans_add_meta(ip->i_gl, bh); 215 memcpy(bh->b_data + o, buf, amount); 216 brelse(bh); 217 218 buf += amount; 219 copied += amount; 220 lblock++; 221 dblock++; 222 extlen--; 223 224 o = sizeof(struct gfs2_meta_header); 225 } 226 227 out: 228 error = gfs2_meta_inode_buffer(ip, &dibh); 229 if (error) 230 return error; 231 232 if (ip->i_inode.i_size < offset + copied) 233 i_size_write(&ip->i_inode, offset + copied); 234 ip->i_inode.i_mtime = ip->i_inode.i_ctime = CURRENT_TIME; 235 236 gfs2_trans_add_meta(ip->i_gl, dibh); 237 gfs2_dinode_out(ip, dibh->b_data); 238 brelse(dibh); 239 240 return copied; 241 fail: 242 if (copied) 243 goto out; 244 return error; 245 } 246 247 static int gfs2_dir_read_stuffed(struct gfs2_inode *ip, __be64 *buf, 248 unsigned int size) 249 { 250 struct buffer_head *dibh; 251 int error; 252 253 error = gfs2_meta_inode_buffer(ip, &dibh); 254 if (!error) { 255 memcpy(buf, dibh->b_data + sizeof(struct gfs2_dinode), size); 256 brelse(dibh); 257 } 258 259 return (error) ? error : size; 260 } 261 262 263 /** 264 * gfs2_dir_read_data - Read a data from a directory inode 265 * @ip: The GFS2 Inode 266 * @buf: The buffer to place result into 267 * @size: Amount of data to transfer 268 * 269 * Returns: The amount of data actually copied or the error 270 */ 271 static int gfs2_dir_read_data(struct gfs2_inode *ip, __be64 *buf, 272 unsigned int size) 273 { 274 struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode); 275 u64 lblock, dblock; 276 u32 extlen = 0; 277 unsigned int o; 278 int copied = 0; 279 int error = 0; 280 281 if (gfs2_is_stuffed(ip)) 282 return gfs2_dir_read_stuffed(ip, buf, size); 283 284 if (gfs2_assert_warn(sdp, gfs2_is_jdata(ip))) 285 return -EINVAL; 286 287 lblock = 0; 288 o = do_div(lblock, sdp->sd_jbsize) + sizeof(struct gfs2_meta_header); 289 290 while (copied < size) { 291 unsigned int amount; 292 struct buffer_head *bh; 293 int new; 294 295 amount = size - copied; 296 if (amount > sdp->sd_sb.sb_bsize - o) 297 amount = sdp->sd_sb.sb_bsize - o; 298 299 if (!extlen) { 300 new = 0; 301 error = gfs2_extent_map(&ip->i_inode, lblock, &new, 302 &dblock, &extlen); 303 if (error || !dblock) 304 goto fail; 305 BUG_ON(extlen < 1); 306 bh = gfs2_meta_ra(ip->i_gl, dblock, extlen); 307 } else { 308 error = gfs2_meta_read(ip->i_gl, dblock, DIO_WAIT, &bh); 309 if (error) 310 goto fail; 311 } 312 error = gfs2_metatype_check(sdp, bh, GFS2_METATYPE_JD); 313 if (error) { 314 brelse(bh); 315 goto fail; 316 } 317 dblock++; 318 extlen--; 319 memcpy(buf, bh->b_data + o, amount); 320 brelse(bh); 321 buf += (amount/sizeof(__be64)); 322 copied += amount; 323 lblock++; 324 o = sizeof(struct gfs2_meta_header); 325 } 326 327 return copied; 328 fail: 329 return (copied) ? copied : error; 330 } 331 332 /** 333 * gfs2_dir_get_hash_table - Get pointer to the dir hash table 334 * @ip: The inode in question 335 * 336 * Returns: The hash table or an error 337 */ 338 339 static __be64 *gfs2_dir_get_hash_table(struct gfs2_inode *ip) 340 { 341 struct inode *inode = &ip->i_inode; 342 int ret; 343 u32 hsize; 344 __be64 *hc; 345 346 BUG_ON(!(ip->i_diskflags & GFS2_DIF_EXHASH)); 347 348 hc = ip->i_hash_cache; 349 if (hc) 350 return hc; 351 352 hsize = 1 << ip->i_depth; 353 hsize *= sizeof(__be64); 354 if (hsize != i_size_read(&ip->i_inode)) { 355 gfs2_consist_inode(ip); 356 return ERR_PTR(-EIO); 357 } 358 359 hc = kmalloc(hsize, GFP_NOFS | __GFP_NOWARN); 360 if (hc == NULL) 361 hc = __vmalloc(hsize, GFP_NOFS, PAGE_KERNEL); 362 363 if (hc == NULL) 364 return ERR_PTR(-ENOMEM); 365 366 ret = gfs2_dir_read_data(ip, hc, hsize); 367 if (ret < 0) { 368 if (is_vmalloc_addr(hc)) 369 vfree(hc); 370 else 371 kfree(hc); 372 return ERR_PTR(ret); 373 } 374 375 spin_lock(&inode->i_lock); 376 if (ip->i_hash_cache) { 377 if (is_vmalloc_addr(hc)) 378 vfree(hc); 379 else 380 kfree(hc); 381 } else { 382 ip->i_hash_cache = hc; 383 } 384 spin_unlock(&inode->i_lock); 385 386 return ip->i_hash_cache; 387 } 388 389 /** 390 * gfs2_dir_hash_inval - Invalidate dir hash 391 * @ip: The directory inode 392 * 393 * Must be called with an exclusive glock, or during glock invalidation. 394 */ 395 void gfs2_dir_hash_inval(struct gfs2_inode *ip) 396 { 397 __be64 *hc = ip->i_hash_cache; 398 ip->i_hash_cache = NULL; 399 if (is_vmalloc_addr(hc)) 400 vfree(hc); 401 else 402 kfree(hc); 403 } 404 405 static inline int gfs2_dirent_sentinel(const struct gfs2_dirent *dent) 406 { 407 return dent->de_inum.no_addr == 0 || dent->de_inum.no_formal_ino == 0; 408 } 409 410 static inline int __gfs2_dirent_find(const struct gfs2_dirent *dent, 411 const struct qstr *name, int ret) 412 { 413 if (!gfs2_dirent_sentinel(dent) && 414 be32_to_cpu(dent->de_hash) == name->hash && 415 be16_to_cpu(dent->de_name_len) == name->len && 416 memcmp(dent+1, name->name, name->len) == 0) 417 return ret; 418 return 0; 419 } 420 421 static int gfs2_dirent_find(const struct gfs2_dirent *dent, 422 const struct qstr *name, 423 void *opaque) 424 { 425 return __gfs2_dirent_find(dent, name, 1); 426 } 427 428 static int gfs2_dirent_prev(const struct gfs2_dirent *dent, 429 const struct qstr *name, 430 void *opaque) 431 { 432 return __gfs2_dirent_find(dent, name, 2); 433 } 434 435 /* 436 * name->name holds ptr to start of block. 437 * name->len holds size of block. 438 */ 439 static int gfs2_dirent_last(const struct gfs2_dirent *dent, 440 const struct qstr *name, 441 void *opaque) 442 { 443 const char *start = name->name; 444 const char *end = (const char *)dent + be16_to_cpu(dent->de_rec_len); 445 if (name->len == (end - start)) 446 return 1; 447 return 0; 448 } 449 450 static int gfs2_dirent_find_space(const struct gfs2_dirent *dent, 451 const struct qstr *name, 452 void *opaque) 453 { 454 unsigned required = GFS2_DIRENT_SIZE(name->len); 455 unsigned actual = GFS2_DIRENT_SIZE(be16_to_cpu(dent->de_name_len)); 456 unsigned totlen = be16_to_cpu(dent->de_rec_len); 457 458 if (gfs2_dirent_sentinel(dent)) 459 actual = 0; 460 if (totlen - actual >= required) 461 return 1; 462 return 0; 463 } 464 465 struct dirent_gather { 466 const struct gfs2_dirent **pdent; 467 unsigned offset; 468 }; 469 470 static int gfs2_dirent_gather(const struct gfs2_dirent *dent, 471 const struct qstr *name, 472 void *opaque) 473 { 474 struct dirent_gather *g = opaque; 475 if (!gfs2_dirent_sentinel(dent)) { 476 g->pdent[g->offset++] = dent; 477 } 478 return 0; 479 } 480 481 /* 482 * Other possible things to check: 483 * - Inode located within filesystem size (and on valid block) 484 * - Valid directory entry type 485 * Not sure how heavy-weight we want to make this... could also check 486 * hash is correct for example, but that would take a lot of extra time. 487 * For now the most important thing is to check that the various sizes 488 * are correct. 489 */ 490 static int gfs2_check_dirent(struct gfs2_dirent *dent, unsigned int offset, 491 unsigned int size, unsigned int len, int first) 492 { 493 const char *msg = "gfs2_dirent too small"; 494 if (unlikely(size < sizeof(struct gfs2_dirent))) 495 goto error; 496 msg = "gfs2_dirent misaligned"; 497 if (unlikely(offset & 0x7)) 498 goto error; 499 msg = "gfs2_dirent points beyond end of block"; 500 if (unlikely(offset + size > len)) 501 goto error; 502 msg = "zero inode number"; 503 if (unlikely(!first && gfs2_dirent_sentinel(dent))) 504 goto error; 505 msg = "name length is greater than space in dirent"; 506 if (!gfs2_dirent_sentinel(dent) && 507 unlikely(sizeof(struct gfs2_dirent)+be16_to_cpu(dent->de_name_len) > 508 size)) 509 goto error; 510 return 0; 511 error: 512 pr_warn("%s: %s (%s)\n", 513 __func__, msg, first ? "first in block" : "not first in block"); 514 return -EIO; 515 } 516 517 static int gfs2_dirent_offset(const void *buf) 518 { 519 const struct gfs2_meta_header *h = buf; 520 int offset; 521 522 BUG_ON(buf == NULL); 523 524 switch(be32_to_cpu(h->mh_type)) { 525 case GFS2_METATYPE_LF: 526 offset = sizeof(struct gfs2_leaf); 527 break; 528 case GFS2_METATYPE_DI: 529 offset = sizeof(struct gfs2_dinode); 530 break; 531 default: 532 goto wrong_type; 533 } 534 return offset; 535 wrong_type: 536 pr_warn("%s: wrong block type %u\n", __func__, be32_to_cpu(h->mh_type)); 537 return -1; 538 } 539 540 static struct gfs2_dirent *gfs2_dirent_scan(struct inode *inode, void *buf, 541 unsigned int len, gfs2_dscan_t scan, 542 const struct qstr *name, 543 void *opaque) 544 { 545 struct gfs2_dirent *dent, *prev; 546 unsigned offset; 547 unsigned size; 548 int ret = 0; 549 550 ret = gfs2_dirent_offset(buf); 551 if (ret < 0) 552 goto consist_inode; 553 554 offset = ret; 555 prev = NULL; 556 dent = buf + offset; 557 size = be16_to_cpu(dent->de_rec_len); 558 if (gfs2_check_dirent(dent, offset, size, len, 1)) 559 goto consist_inode; 560 do { 561 ret = scan(dent, name, opaque); 562 if (ret) 563 break; 564 offset += size; 565 if (offset == len) 566 break; 567 prev = dent; 568 dent = buf + offset; 569 size = be16_to_cpu(dent->de_rec_len); 570 if (gfs2_check_dirent(dent, offset, size, len, 0)) 571 goto consist_inode; 572 } while(1); 573 574 switch(ret) { 575 case 0: 576 return NULL; 577 case 1: 578 return dent; 579 case 2: 580 return prev ? prev : dent; 581 default: 582 BUG_ON(ret > 0); 583 return ERR_PTR(ret); 584 } 585 586 consist_inode: 587 gfs2_consist_inode(GFS2_I(inode)); 588 return ERR_PTR(-EIO); 589 } 590 591 static int dirent_check_reclen(struct gfs2_inode *dip, 592 const struct gfs2_dirent *d, const void *end_p) 593 { 594 const void *ptr = d; 595 u16 rec_len = be16_to_cpu(d->de_rec_len); 596 597 if (unlikely(rec_len < sizeof(struct gfs2_dirent))) 598 goto broken; 599 ptr += rec_len; 600 if (ptr < end_p) 601 return rec_len; 602 if (ptr == end_p) 603 return -ENOENT; 604 broken: 605 gfs2_consist_inode(dip); 606 return -EIO; 607 } 608 609 /** 610 * dirent_next - Next dirent 611 * @dip: the directory 612 * @bh: The buffer 613 * @dent: Pointer to list of dirents 614 * 615 * Returns: 0 on success, error code otherwise 616 */ 617 618 static int dirent_next(struct gfs2_inode *dip, struct buffer_head *bh, 619 struct gfs2_dirent **dent) 620 { 621 struct gfs2_dirent *cur = *dent, *tmp; 622 char *bh_end = bh->b_data + bh->b_size; 623 int ret; 624 625 ret = dirent_check_reclen(dip, cur, bh_end); 626 if (ret < 0) 627 return ret; 628 629 tmp = (void *)cur + ret; 630 ret = dirent_check_reclen(dip, tmp, bh_end); 631 if (ret == -EIO) 632 return ret; 633 634 /* Only the first dent could ever have de_inum.no_addr == 0 */ 635 if (gfs2_dirent_sentinel(tmp)) { 636 gfs2_consist_inode(dip); 637 return -EIO; 638 } 639 640 *dent = tmp; 641 return 0; 642 } 643 644 /** 645 * dirent_del - Delete a dirent 646 * @dip: The GFS2 inode 647 * @bh: The buffer 648 * @prev: The previous dirent 649 * @cur: The current dirent 650 * 651 */ 652 653 static void dirent_del(struct gfs2_inode *dip, struct buffer_head *bh, 654 struct gfs2_dirent *prev, struct gfs2_dirent *cur) 655 { 656 u16 cur_rec_len, prev_rec_len; 657 658 if (gfs2_dirent_sentinel(cur)) { 659 gfs2_consist_inode(dip); 660 return; 661 } 662 663 gfs2_trans_add_meta(dip->i_gl, bh); 664 665 /* If there is no prev entry, this is the first entry in the block. 666 The de_rec_len is already as big as it needs to be. Just zero 667 out the inode number and return. */ 668 669 if (!prev) { 670 cur->de_inum.no_addr = 0; 671 cur->de_inum.no_formal_ino = 0; 672 return; 673 } 674 675 /* Combine this dentry with the previous one. */ 676 677 prev_rec_len = be16_to_cpu(prev->de_rec_len); 678 cur_rec_len = be16_to_cpu(cur->de_rec_len); 679 680 if ((char *)prev + prev_rec_len != (char *)cur) 681 gfs2_consist_inode(dip); 682 if ((char *)cur + cur_rec_len > bh->b_data + bh->b_size) 683 gfs2_consist_inode(dip); 684 685 prev_rec_len += cur_rec_len; 686 prev->de_rec_len = cpu_to_be16(prev_rec_len); 687 } 688 689 /* 690 * Takes a dent from which to grab space as an argument. Returns the 691 * newly created dent. 692 */ 693 static struct gfs2_dirent *gfs2_init_dirent(struct inode *inode, 694 struct gfs2_dirent *dent, 695 const struct qstr *name, 696 struct buffer_head *bh) 697 { 698 struct gfs2_inode *ip = GFS2_I(inode); 699 struct gfs2_dirent *ndent; 700 unsigned offset = 0, totlen; 701 702 if (!gfs2_dirent_sentinel(dent)) 703 offset = GFS2_DIRENT_SIZE(be16_to_cpu(dent->de_name_len)); 704 totlen = be16_to_cpu(dent->de_rec_len); 705 BUG_ON(offset + name->len > totlen); 706 gfs2_trans_add_meta(ip->i_gl, bh); 707 ndent = (struct gfs2_dirent *)((char *)dent + offset); 708 dent->de_rec_len = cpu_to_be16(offset); 709 gfs2_qstr2dirent(name, totlen - offset, ndent); 710 return ndent; 711 } 712 713 static struct gfs2_dirent *gfs2_dirent_alloc(struct inode *inode, 714 struct buffer_head *bh, 715 const struct qstr *name) 716 { 717 struct gfs2_dirent *dent; 718 dent = gfs2_dirent_scan(inode, bh->b_data, bh->b_size, 719 gfs2_dirent_find_space, name, NULL); 720 if (!dent || IS_ERR(dent)) 721 return dent; 722 return gfs2_init_dirent(inode, dent, name, bh); 723 } 724 725 static int get_leaf(struct gfs2_inode *dip, u64 leaf_no, 726 struct buffer_head **bhp) 727 { 728 int error; 729 730 error = gfs2_meta_read(dip->i_gl, leaf_no, DIO_WAIT, bhp); 731 if (!error && gfs2_metatype_check(GFS2_SB(&dip->i_inode), *bhp, GFS2_METATYPE_LF)) { 732 /* pr_info("block num=%llu\n", leaf_no); */ 733 error = -EIO; 734 } 735 736 return error; 737 } 738 739 /** 740 * get_leaf_nr - Get a leaf number associated with the index 741 * @dip: The GFS2 inode 742 * @index: 743 * @leaf_out: 744 * 745 * Returns: 0 on success, error code otherwise 746 */ 747 748 static int get_leaf_nr(struct gfs2_inode *dip, u32 index, 749 u64 *leaf_out) 750 { 751 __be64 *hash; 752 753 hash = gfs2_dir_get_hash_table(dip); 754 if (IS_ERR(hash)) 755 return PTR_ERR(hash); 756 *leaf_out = be64_to_cpu(*(hash + index)); 757 return 0; 758 } 759 760 static int get_first_leaf(struct gfs2_inode *dip, u32 index, 761 struct buffer_head **bh_out) 762 { 763 u64 leaf_no; 764 int error; 765 766 error = get_leaf_nr(dip, index, &leaf_no); 767 if (!error) 768 error = get_leaf(dip, leaf_no, bh_out); 769 770 return error; 771 } 772 773 static struct gfs2_dirent *gfs2_dirent_search(struct inode *inode, 774 const struct qstr *name, 775 gfs2_dscan_t scan, 776 struct buffer_head **pbh) 777 { 778 struct buffer_head *bh; 779 struct gfs2_dirent *dent; 780 struct gfs2_inode *ip = GFS2_I(inode); 781 int error; 782 783 if (ip->i_diskflags & GFS2_DIF_EXHASH) { 784 struct gfs2_leaf *leaf; 785 unsigned hsize = 1 << ip->i_depth; 786 unsigned index; 787 u64 ln; 788 if (hsize * sizeof(u64) != i_size_read(inode)) { 789 gfs2_consist_inode(ip); 790 return ERR_PTR(-EIO); 791 } 792 793 index = name->hash >> (32 - ip->i_depth); 794 error = get_first_leaf(ip, index, &bh); 795 if (error) 796 return ERR_PTR(error); 797 do { 798 dent = gfs2_dirent_scan(inode, bh->b_data, bh->b_size, 799 scan, name, NULL); 800 if (dent) 801 goto got_dent; 802 leaf = (struct gfs2_leaf *)bh->b_data; 803 ln = be64_to_cpu(leaf->lf_next); 804 brelse(bh); 805 if (!ln) 806 break; 807 808 error = get_leaf(ip, ln, &bh); 809 } while(!error); 810 811 return error ? ERR_PTR(error) : NULL; 812 } 813 814 815 error = gfs2_meta_inode_buffer(ip, &bh); 816 if (error) 817 return ERR_PTR(error); 818 dent = gfs2_dirent_scan(inode, bh->b_data, bh->b_size, scan, name, NULL); 819 got_dent: 820 if (unlikely(dent == NULL || IS_ERR(dent))) { 821 brelse(bh); 822 bh = NULL; 823 } 824 *pbh = bh; 825 return dent; 826 } 827 828 static struct gfs2_leaf *new_leaf(struct inode *inode, struct buffer_head **pbh, u16 depth) 829 { 830 struct gfs2_inode *ip = GFS2_I(inode); 831 unsigned int n = 1; 832 u64 bn; 833 int error; 834 struct buffer_head *bh; 835 struct gfs2_leaf *leaf; 836 struct gfs2_dirent *dent; 837 struct qstr name = { .name = "" }; 838 struct timespec tv = CURRENT_TIME; 839 840 error = gfs2_alloc_blocks(ip, &bn, &n, 0, NULL); 841 if (error) 842 return NULL; 843 bh = gfs2_meta_new(ip->i_gl, bn); 844 if (!bh) 845 return NULL; 846 847 gfs2_trans_add_unrevoke(GFS2_SB(inode), bn, 1); 848 gfs2_trans_add_meta(ip->i_gl, bh); 849 gfs2_metatype_set(bh, GFS2_METATYPE_LF, GFS2_FORMAT_LF); 850 leaf = (struct gfs2_leaf *)bh->b_data; 851 leaf->lf_depth = cpu_to_be16(depth); 852 leaf->lf_entries = 0; 853 leaf->lf_dirent_format = cpu_to_be32(GFS2_FORMAT_DE); 854 leaf->lf_next = 0; 855 leaf->lf_inode = cpu_to_be64(ip->i_no_addr); 856 leaf->lf_dist = cpu_to_be32(1); 857 leaf->lf_nsec = cpu_to_be32(tv.tv_nsec); 858 leaf->lf_sec = cpu_to_be64(tv.tv_sec); 859 memset(leaf->lf_reserved2, 0, sizeof(leaf->lf_reserved2)); 860 dent = (struct gfs2_dirent *)(leaf+1); 861 gfs2_qstr2dirent(&name, bh->b_size - sizeof(struct gfs2_leaf), dent); 862 *pbh = bh; 863 return leaf; 864 } 865 866 /** 867 * dir_make_exhash - Convert a stuffed directory into an ExHash directory 868 * @dip: The GFS2 inode 869 * 870 * Returns: 0 on success, error code otherwise 871 */ 872 873 static int dir_make_exhash(struct inode *inode) 874 { 875 struct gfs2_inode *dip = GFS2_I(inode); 876 struct gfs2_sbd *sdp = GFS2_SB(inode); 877 struct gfs2_dirent *dent; 878 struct qstr args; 879 struct buffer_head *bh, *dibh; 880 struct gfs2_leaf *leaf; 881 int y; 882 u32 x; 883 __be64 *lp; 884 u64 bn; 885 int error; 886 887 error = gfs2_meta_inode_buffer(dip, &dibh); 888 if (error) 889 return error; 890 891 /* Turn over a new leaf */ 892 893 leaf = new_leaf(inode, &bh, 0); 894 if (!leaf) 895 return -ENOSPC; 896 bn = bh->b_blocknr; 897 898 gfs2_assert(sdp, dip->i_entries < (1 << 16)); 899 leaf->lf_entries = cpu_to_be16(dip->i_entries); 900 901 /* Copy dirents */ 902 903 gfs2_buffer_copy_tail(bh, sizeof(struct gfs2_leaf), dibh, 904 sizeof(struct gfs2_dinode)); 905 906 /* Find last entry */ 907 908 x = 0; 909 args.len = bh->b_size - sizeof(struct gfs2_dinode) + 910 sizeof(struct gfs2_leaf); 911 args.name = bh->b_data; 912 dent = gfs2_dirent_scan(&dip->i_inode, bh->b_data, bh->b_size, 913 gfs2_dirent_last, &args, NULL); 914 if (!dent) { 915 brelse(bh); 916 brelse(dibh); 917 return -EIO; 918 } 919 if (IS_ERR(dent)) { 920 brelse(bh); 921 brelse(dibh); 922 return PTR_ERR(dent); 923 } 924 925 /* Adjust the last dirent's record length 926 (Remember that dent still points to the last entry.) */ 927 928 dent->de_rec_len = cpu_to_be16(be16_to_cpu(dent->de_rec_len) + 929 sizeof(struct gfs2_dinode) - 930 sizeof(struct gfs2_leaf)); 931 932 brelse(bh); 933 934 /* We're done with the new leaf block, now setup the new 935 hash table. */ 936 937 gfs2_trans_add_meta(dip->i_gl, dibh); 938 gfs2_buffer_clear_tail(dibh, sizeof(struct gfs2_dinode)); 939 940 lp = (__be64 *)(dibh->b_data + sizeof(struct gfs2_dinode)); 941 942 for (x = sdp->sd_hash_ptrs; x--; lp++) 943 *lp = cpu_to_be64(bn); 944 945 i_size_write(inode, sdp->sd_sb.sb_bsize / 2); 946 gfs2_add_inode_blocks(&dip->i_inode, 1); 947 dip->i_diskflags |= GFS2_DIF_EXHASH; 948 949 for (x = sdp->sd_hash_ptrs, y = -1; x; x >>= 1, y++) ; 950 dip->i_depth = y; 951 952 gfs2_dinode_out(dip, dibh->b_data); 953 954 brelse(dibh); 955 956 return 0; 957 } 958 959 /** 960 * dir_split_leaf - Split a leaf block into two 961 * @dip: The GFS2 inode 962 * @index: 963 * @leaf_no: 964 * 965 * Returns: 0 on success, error code on failure 966 */ 967 968 static int dir_split_leaf(struct inode *inode, const struct qstr *name) 969 { 970 struct gfs2_inode *dip = GFS2_I(inode); 971 struct buffer_head *nbh, *obh, *dibh; 972 struct gfs2_leaf *nleaf, *oleaf; 973 struct gfs2_dirent *dent = NULL, *prev = NULL, *next = NULL, *new; 974 u32 start, len, half_len, divider; 975 u64 bn, leaf_no; 976 __be64 *lp; 977 u32 index; 978 int x, moved = 0; 979 int error; 980 981 index = name->hash >> (32 - dip->i_depth); 982 error = get_leaf_nr(dip, index, &leaf_no); 983 if (error) 984 return error; 985 986 /* Get the old leaf block */ 987 error = get_leaf(dip, leaf_no, &obh); 988 if (error) 989 return error; 990 991 oleaf = (struct gfs2_leaf *)obh->b_data; 992 if (dip->i_depth == be16_to_cpu(oleaf->lf_depth)) { 993 brelse(obh); 994 return 1; /* can't split */ 995 } 996 997 gfs2_trans_add_meta(dip->i_gl, obh); 998 999 nleaf = new_leaf(inode, &nbh, be16_to_cpu(oleaf->lf_depth) + 1); 1000 if (!nleaf) { 1001 brelse(obh); 1002 return -ENOSPC; 1003 } 1004 bn = nbh->b_blocknr; 1005 1006 /* Compute the start and len of leaf pointers in the hash table. */ 1007 len = 1 << (dip->i_depth - be16_to_cpu(oleaf->lf_depth)); 1008 half_len = len >> 1; 1009 if (!half_len) { 1010 pr_warn("i_depth %u lf_depth %u index %u\n", 1011 dip->i_depth, be16_to_cpu(oleaf->lf_depth), index); 1012 gfs2_consist_inode(dip); 1013 error = -EIO; 1014 goto fail_brelse; 1015 } 1016 1017 start = (index & ~(len - 1)); 1018 1019 /* Change the pointers. 1020 Don't bother distinguishing stuffed from non-stuffed. 1021 This code is complicated enough already. */ 1022 lp = kmalloc(half_len * sizeof(__be64), GFP_NOFS); 1023 if (!lp) { 1024 error = -ENOMEM; 1025 goto fail_brelse; 1026 } 1027 1028 /* Change the pointers */ 1029 for (x = 0; x < half_len; x++) 1030 lp[x] = cpu_to_be64(bn); 1031 1032 gfs2_dir_hash_inval(dip); 1033 1034 error = gfs2_dir_write_data(dip, (char *)lp, start * sizeof(u64), 1035 half_len * sizeof(u64)); 1036 if (error != half_len * sizeof(u64)) { 1037 if (error >= 0) 1038 error = -EIO; 1039 goto fail_lpfree; 1040 } 1041 1042 kfree(lp); 1043 1044 /* Compute the divider */ 1045 divider = (start + half_len) << (32 - dip->i_depth); 1046 1047 /* Copy the entries */ 1048 dent = (struct gfs2_dirent *)(obh->b_data + sizeof(struct gfs2_leaf)); 1049 1050 do { 1051 next = dent; 1052 if (dirent_next(dip, obh, &next)) 1053 next = NULL; 1054 1055 if (!gfs2_dirent_sentinel(dent) && 1056 be32_to_cpu(dent->de_hash) < divider) { 1057 struct qstr str; 1058 str.name = (char*)(dent+1); 1059 str.len = be16_to_cpu(dent->de_name_len); 1060 str.hash = be32_to_cpu(dent->de_hash); 1061 new = gfs2_dirent_alloc(inode, nbh, &str); 1062 if (IS_ERR(new)) { 1063 error = PTR_ERR(new); 1064 break; 1065 } 1066 1067 new->de_inum = dent->de_inum; /* No endian worries */ 1068 new->de_type = dent->de_type; /* No endian worries */ 1069 be16_add_cpu(&nleaf->lf_entries, 1); 1070 1071 dirent_del(dip, obh, prev, dent); 1072 1073 if (!oleaf->lf_entries) 1074 gfs2_consist_inode(dip); 1075 be16_add_cpu(&oleaf->lf_entries, -1); 1076 1077 if (!prev) 1078 prev = dent; 1079 1080 moved = 1; 1081 } else { 1082 prev = dent; 1083 } 1084 dent = next; 1085 } while (dent); 1086 1087 oleaf->lf_depth = nleaf->lf_depth; 1088 1089 error = gfs2_meta_inode_buffer(dip, &dibh); 1090 if (!gfs2_assert_withdraw(GFS2_SB(&dip->i_inode), !error)) { 1091 gfs2_trans_add_meta(dip->i_gl, dibh); 1092 gfs2_add_inode_blocks(&dip->i_inode, 1); 1093 gfs2_dinode_out(dip, dibh->b_data); 1094 brelse(dibh); 1095 } 1096 1097 brelse(obh); 1098 brelse(nbh); 1099 1100 return error; 1101 1102 fail_lpfree: 1103 kfree(lp); 1104 1105 fail_brelse: 1106 brelse(obh); 1107 brelse(nbh); 1108 return error; 1109 } 1110 1111 /** 1112 * dir_double_exhash - Double size of ExHash table 1113 * @dip: The GFS2 dinode 1114 * 1115 * Returns: 0 on success, error code on failure 1116 */ 1117 1118 static int dir_double_exhash(struct gfs2_inode *dip) 1119 { 1120 struct buffer_head *dibh; 1121 u32 hsize; 1122 u32 hsize_bytes; 1123 __be64 *hc; 1124 __be64 *hc2, *h; 1125 int x; 1126 int error = 0; 1127 1128 hsize = 1 << dip->i_depth; 1129 hsize_bytes = hsize * sizeof(__be64); 1130 1131 hc = gfs2_dir_get_hash_table(dip); 1132 if (IS_ERR(hc)) 1133 return PTR_ERR(hc); 1134 1135 hc2 = kmalloc(hsize_bytes * 2, GFP_NOFS | __GFP_NOWARN); 1136 if (hc2 == NULL) 1137 hc2 = __vmalloc(hsize_bytes * 2, GFP_NOFS, PAGE_KERNEL); 1138 1139 if (!hc2) 1140 return -ENOMEM; 1141 1142 h = hc2; 1143 error = gfs2_meta_inode_buffer(dip, &dibh); 1144 if (error) 1145 goto out_kfree; 1146 1147 for (x = 0; x < hsize; x++) { 1148 *h++ = *hc; 1149 *h++ = *hc; 1150 hc++; 1151 } 1152 1153 error = gfs2_dir_write_data(dip, (char *)hc2, 0, hsize_bytes * 2); 1154 if (error != (hsize_bytes * 2)) 1155 goto fail; 1156 1157 gfs2_dir_hash_inval(dip); 1158 dip->i_hash_cache = hc2; 1159 dip->i_depth++; 1160 gfs2_dinode_out(dip, dibh->b_data); 1161 brelse(dibh); 1162 return 0; 1163 1164 fail: 1165 /* Replace original hash table & size */ 1166 gfs2_dir_write_data(dip, (char *)hc, 0, hsize_bytes); 1167 i_size_write(&dip->i_inode, hsize_bytes); 1168 gfs2_dinode_out(dip, dibh->b_data); 1169 brelse(dibh); 1170 out_kfree: 1171 if (is_vmalloc_addr(hc2)) 1172 vfree(hc2); 1173 else 1174 kfree(hc2); 1175 return error; 1176 } 1177 1178 /** 1179 * compare_dents - compare directory entries by hash value 1180 * @a: first dent 1181 * @b: second dent 1182 * 1183 * When comparing the hash entries of @a to @b: 1184 * gt: returns 1 1185 * lt: returns -1 1186 * eq: returns 0 1187 */ 1188 1189 static int compare_dents(const void *a, const void *b) 1190 { 1191 const struct gfs2_dirent *dent_a, *dent_b; 1192 u32 hash_a, hash_b; 1193 int ret = 0; 1194 1195 dent_a = *(const struct gfs2_dirent **)a; 1196 hash_a = be32_to_cpu(dent_a->de_hash); 1197 1198 dent_b = *(const struct gfs2_dirent **)b; 1199 hash_b = be32_to_cpu(dent_b->de_hash); 1200 1201 if (hash_a > hash_b) 1202 ret = 1; 1203 else if (hash_a < hash_b) 1204 ret = -1; 1205 else { 1206 unsigned int len_a = be16_to_cpu(dent_a->de_name_len); 1207 unsigned int len_b = be16_to_cpu(dent_b->de_name_len); 1208 1209 if (len_a > len_b) 1210 ret = 1; 1211 else if (len_a < len_b) 1212 ret = -1; 1213 else 1214 ret = memcmp(dent_a + 1, dent_b + 1, len_a); 1215 } 1216 1217 return ret; 1218 } 1219 1220 /** 1221 * do_filldir_main - read out directory entries 1222 * @dip: The GFS2 inode 1223 * @ctx: what to feed the entries to 1224 * @darr: an array of struct gfs2_dirent pointers to read 1225 * @entries: the number of entries in darr 1226 * @copied: pointer to int that's non-zero if a entry has been copied out 1227 * 1228 * Jump through some hoops to make sure that if there are hash collsions, 1229 * they are read out at the beginning of a buffer. We want to minimize 1230 * the possibility that they will fall into different readdir buffers or 1231 * that someone will want to seek to that location. 1232 * 1233 * Returns: errno, >0 if the actor tells you to stop 1234 */ 1235 1236 static int do_filldir_main(struct gfs2_inode *dip, struct dir_context *ctx, 1237 const struct gfs2_dirent **darr, u32 entries, 1238 int *copied) 1239 { 1240 const struct gfs2_dirent *dent, *dent_next; 1241 u64 off, off_next; 1242 unsigned int x, y; 1243 int run = 0; 1244 1245 sort(darr, entries, sizeof(struct gfs2_dirent *), compare_dents, NULL); 1246 1247 dent_next = darr[0]; 1248 off_next = be32_to_cpu(dent_next->de_hash); 1249 off_next = gfs2_disk_hash2offset(off_next); 1250 1251 for (x = 0, y = 1; x < entries; x++, y++) { 1252 dent = dent_next; 1253 off = off_next; 1254 1255 if (y < entries) { 1256 dent_next = darr[y]; 1257 off_next = be32_to_cpu(dent_next->de_hash); 1258 off_next = gfs2_disk_hash2offset(off_next); 1259 1260 if (off < ctx->pos) 1261 continue; 1262 ctx->pos = off; 1263 1264 if (off_next == off) { 1265 if (*copied && !run) 1266 return 1; 1267 run = 1; 1268 } else 1269 run = 0; 1270 } else { 1271 if (off < ctx->pos) 1272 continue; 1273 ctx->pos = off; 1274 } 1275 1276 if (!dir_emit(ctx, (const char *)(dent + 1), 1277 be16_to_cpu(dent->de_name_len), 1278 be64_to_cpu(dent->de_inum.no_addr), 1279 be16_to_cpu(dent->de_type))) 1280 return 1; 1281 1282 *copied = 1; 1283 } 1284 1285 /* Increment the ctx->pos by one, so the next time we come into the 1286 do_filldir fxn, we get the next entry instead of the last one in the 1287 current leaf */ 1288 1289 ctx->pos++; 1290 1291 return 0; 1292 } 1293 1294 static void *gfs2_alloc_sort_buffer(unsigned size) 1295 { 1296 void *ptr = NULL; 1297 1298 if (size < KMALLOC_MAX_SIZE) 1299 ptr = kmalloc(size, GFP_NOFS | __GFP_NOWARN); 1300 if (!ptr) 1301 ptr = __vmalloc(size, GFP_NOFS, PAGE_KERNEL); 1302 return ptr; 1303 } 1304 1305 static void gfs2_free_sort_buffer(void *ptr) 1306 { 1307 if (is_vmalloc_addr(ptr)) 1308 vfree(ptr); 1309 else 1310 kfree(ptr); 1311 } 1312 1313 static int gfs2_dir_read_leaf(struct inode *inode, struct dir_context *ctx, 1314 int *copied, unsigned *depth, 1315 u64 leaf_no) 1316 { 1317 struct gfs2_inode *ip = GFS2_I(inode); 1318 struct gfs2_sbd *sdp = GFS2_SB(inode); 1319 struct buffer_head *bh; 1320 struct gfs2_leaf *lf; 1321 unsigned entries = 0, entries2 = 0; 1322 unsigned leaves = 0; 1323 const struct gfs2_dirent **darr, *dent; 1324 struct dirent_gather g; 1325 struct buffer_head **larr; 1326 int leaf = 0; 1327 int error, i; 1328 u64 lfn = leaf_no; 1329 1330 do { 1331 error = get_leaf(ip, lfn, &bh); 1332 if (error) 1333 goto out; 1334 lf = (struct gfs2_leaf *)bh->b_data; 1335 if (leaves == 0) 1336 *depth = be16_to_cpu(lf->lf_depth); 1337 entries += be16_to_cpu(lf->lf_entries); 1338 leaves++; 1339 lfn = be64_to_cpu(lf->lf_next); 1340 brelse(bh); 1341 } while(lfn); 1342 1343 if (!entries) 1344 return 0; 1345 1346 error = -ENOMEM; 1347 /* 1348 * The extra 99 entries are not normally used, but are a buffer 1349 * zone in case the number of entries in the leaf is corrupt. 1350 * 99 is the maximum number of entries that can fit in a single 1351 * leaf block. 1352 */ 1353 larr = gfs2_alloc_sort_buffer((leaves + entries + 99) * sizeof(void *)); 1354 if (!larr) 1355 goto out; 1356 darr = (const struct gfs2_dirent **)(larr + leaves); 1357 g.pdent = darr; 1358 g.offset = 0; 1359 lfn = leaf_no; 1360 1361 do { 1362 error = get_leaf(ip, lfn, &bh); 1363 if (error) 1364 goto out_free; 1365 lf = (struct gfs2_leaf *)bh->b_data; 1366 lfn = be64_to_cpu(lf->lf_next); 1367 if (lf->lf_entries) { 1368 entries2 += be16_to_cpu(lf->lf_entries); 1369 dent = gfs2_dirent_scan(inode, bh->b_data, bh->b_size, 1370 gfs2_dirent_gather, NULL, &g); 1371 error = PTR_ERR(dent); 1372 if (IS_ERR(dent)) 1373 goto out_free; 1374 if (entries2 != g.offset) { 1375 fs_warn(sdp, "Number of entries corrupt in dir " 1376 "leaf %llu, entries2 (%u) != " 1377 "g.offset (%u)\n", 1378 (unsigned long long)bh->b_blocknr, 1379 entries2, g.offset); 1380 1381 error = -EIO; 1382 goto out_free; 1383 } 1384 error = 0; 1385 larr[leaf++] = bh; 1386 } else { 1387 brelse(bh); 1388 } 1389 } while(lfn); 1390 1391 BUG_ON(entries2 != entries); 1392 error = do_filldir_main(ip, ctx, darr, entries, copied); 1393 out_free: 1394 for(i = 0; i < leaf; i++) 1395 brelse(larr[i]); 1396 gfs2_free_sort_buffer(larr); 1397 out: 1398 return error; 1399 } 1400 1401 /** 1402 * gfs2_dir_readahead - Issue read-ahead requests for leaf blocks. 1403 * 1404 * Note: we can't calculate each index like dir_e_read can because we don't 1405 * have the leaf, and therefore we don't have the depth, and therefore we 1406 * don't have the length. So we have to just read enough ahead to make up 1407 * for the loss of information. 1408 */ 1409 static void gfs2_dir_readahead(struct inode *inode, unsigned hsize, u32 index, 1410 struct file_ra_state *f_ra) 1411 { 1412 struct gfs2_inode *ip = GFS2_I(inode); 1413 struct gfs2_glock *gl = ip->i_gl; 1414 struct buffer_head *bh; 1415 u64 blocknr = 0, last; 1416 unsigned count; 1417 1418 /* First check if we've already read-ahead for the whole range. */ 1419 if (index + MAX_RA_BLOCKS < f_ra->start) 1420 return; 1421 1422 f_ra->start = max((pgoff_t)index, f_ra->start); 1423 for (count = 0; count < MAX_RA_BLOCKS; count++) { 1424 if (f_ra->start >= hsize) /* if exceeded the hash table */ 1425 break; 1426 1427 last = blocknr; 1428 blocknr = be64_to_cpu(ip->i_hash_cache[f_ra->start]); 1429 f_ra->start++; 1430 if (blocknr == last) 1431 continue; 1432 1433 bh = gfs2_getbuf(gl, blocknr, 1); 1434 if (trylock_buffer(bh)) { 1435 if (buffer_uptodate(bh)) { 1436 unlock_buffer(bh); 1437 brelse(bh); 1438 continue; 1439 } 1440 bh->b_end_io = end_buffer_read_sync; 1441 submit_bh(READA | REQ_META, bh); 1442 continue; 1443 } 1444 brelse(bh); 1445 } 1446 } 1447 1448 /** 1449 * dir_e_read - Reads the entries from a directory into a filldir buffer 1450 * @dip: dinode pointer 1451 * @ctx: actor to feed the entries to 1452 * 1453 * Returns: errno 1454 */ 1455 1456 static int dir_e_read(struct inode *inode, struct dir_context *ctx, 1457 struct file_ra_state *f_ra) 1458 { 1459 struct gfs2_inode *dip = GFS2_I(inode); 1460 u32 hsize, len = 0; 1461 u32 hash, index; 1462 __be64 *lp; 1463 int copied = 0; 1464 int error = 0; 1465 unsigned depth = 0; 1466 1467 hsize = 1 << dip->i_depth; 1468 hash = gfs2_dir_offset2hash(ctx->pos); 1469 index = hash >> (32 - dip->i_depth); 1470 1471 if (dip->i_hash_cache == NULL) 1472 f_ra->start = 0; 1473 lp = gfs2_dir_get_hash_table(dip); 1474 if (IS_ERR(lp)) 1475 return PTR_ERR(lp); 1476 1477 gfs2_dir_readahead(inode, hsize, index, f_ra); 1478 1479 while (index < hsize) { 1480 error = gfs2_dir_read_leaf(inode, ctx, 1481 &copied, &depth, 1482 be64_to_cpu(lp[index])); 1483 if (error) 1484 break; 1485 1486 len = 1 << (dip->i_depth - depth); 1487 index = (index & ~(len - 1)) + len; 1488 } 1489 1490 if (error > 0) 1491 error = 0; 1492 return error; 1493 } 1494 1495 int gfs2_dir_read(struct inode *inode, struct dir_context *ctx, 1496 struct file_ra_state *f_ra) 1497 { 1498 struct gfs2_inode *dip = GFS2_I(inode); 1499 struct gfs2_sbd *sdp = GFS2_SB(inode); 1500 struct dirent_gather g; 1501 const struct gfs2_dirent **darr, *dent; 1502 struct buffer_head *dibh; 1503 int copied = 0; 1504 int error; 1505 1506 if (!dip->i_entries) 1507 return 0; 1508 1509 if (dip->i_diskflags & GFS2_DIF_EXHASH) 1510 return dir_e_read(inode, ctx, f_ra); 1511 1512 if (!gfs2_is_stuffed(dip)) { 1513 gfs2_consist_inode(dip); 1514 return -EIO; 1515 } 1516 1517 error = gfs2_meta_inode_buffer(dip, &dibh); 1518 if (error) 1519 return error; 1520 1521 error = -ENOMEM; 1522 /* 96 is max number of dirents which can be stuffed into an inode */ 1523 darr = kmalloc(96 * sizeof(struct gfs2_dirent *), GFP_NOFS); 1524 if (darr) { 1525 g.pdent = darr; 1526 g.offset = 0; 1527 dent = gfs2_dirent_scan(inode, dibh->b_data, dibh->b_size, 1528 gfs2_dirent_gather, NULL, &g); 1529 if (IS_ERR(dent)) { 1530 error = PTR_ERR(dent); 1531 goto out; 1532 } 1533 if (dip->i_entries != g.offset) { 1534 fs_warn(sdp, "Number of entries corrupt in dir %llu, " 1535 "ip->i_entries (%u) != g.offset (%u)\n", 1536 (unsigned long long)dip->i_no_addr, 1537 dip->i_entries, 1538 g.offset); 1539 error = -EIO; 1540 goto out; 1541 } 1542 error = do_filldir_main(dip, ctx, darr, 1543 dip->i_entries, &copied); 1544 out: 1545 kfree(darr); 1546 } 1547 1548 if (error > 0) 1549 error = 0; 1550 1551 brelse(dibh); 1552 1553 return error; 1554 } 1555 1556 /** 1557 * gfs2_dir_search - Search a directory 1558 * @dip: The GFS2 dir inode 1559 * @name: The name we are looking up 1560 * @fail_on_exist: Fail if the name exists rather than looking it up 1561 * 1562 * This routine searches a directory for a file or another directory. 1563 * Assumes a glock is held on dip. 1564 * 1565 * Returns: errno 1566 */ 1567 1568 struct inode *gfs2_dir_search(struct inode *dir, const struct qstr *name, 1569 bool fail_on_exist) 1570 { 1571 struct buffer_head *bh; 1572 struct gfs2_dirent *dent; 1573 u64 addr, formal_ino; 1574 u16 dtype; 1575 1576 dent = gfs2_dirent_search(dir, name, gfs2_dirent_find, &bh); 1577 if (dent) { 1578 if (IS_ERR(dent)) 1579 return ERR_CAST(dent); 1580 dtype = be16_to_cpu(dent->de_type); 1581 addr = be64_to_cpu(dent->de_inum.no_addr); 1582 formal_ino = be64_to_cpu(dent->de_inum.no_formal_ino); 1583 brelse(bh); 1584 if (fail_on_exist) 1585 return ERR_PTR(-EEXIST); 1586 return gfs2_inode_lookup(dir->i_sb, dtype, addr, formal_ino, 0); 1587 } 1588 return ERR_PTR(-ENOENT); 1589 } 1590 1591 int gfs2_dir_check(struct inode *dir, const struct qstr *name, 1592 const struct gfs2_inode *ip) 1593 { 1594 struct buffer_head *bh; 1595 struct gfs2_dirent *dent; 1596 int ret = -ENOENT; 1597 1598 dent = gfs2_dirent_search(dir, name, gfs2_dirent_find, &bh); 1599 if (dent) { 1600 if (IS_ERR(dent)) 1601 return PTR_ERR(dent); 1602 if (ip) { 1603 if (be64_to_cpu(dent->de_inum.no_addr) != ip->i_no_addr) 1604 goto out; 1605 if (be64_to_cpu(dent->de_inum.no_formal_ino) != 1606 ip->i_no_formal_ino) 1607 goto out; 1608 if (unlikely(IF2DT(ip->i_inode.i_mode) != 1609 be16_to_cpu(dent->de_type))) { 1610 gfs2_consist_inode(GFS2_I(dir)); 1611 ret = -EIO; 1612 goto out; 1613 } 1614 } 1615 ret = 0; 1616 out: 1617 brelse(bh); 1618 } 1619 return ret; 1620 } 1621 1622 /** 1623 * dir_new_leaf - Add a new leaf onto hash chain 1624 * @inode: The directory 1625 * @name: The name we are adding 1626 * 1627 * This adds a new dir leaf onto an existing leaf when there is not 1628 * enough space to add a new dir entry. This is a last resort after 1629 * we've expanded the hash table to max size and also split existing 1630 * leaf blocks, so it will only occur for very large directories. 1631 * 1632 * The dist parameter is set to 1 for leaf blocks directly attached 1633 * to the hash table, 2 for one layer of indirection, 3 for two layers 1634 * etc. We are thus able to tell the difference between an old leaf 1635 * with dist set to zero (i.e. "don't know") and a new one where we 1636 * set this information for debug/fsck purposes. 1637 * 1638 * Returns: 0 on success, or -ve on error 1639 */ 1640 1641 static int dir_new_leaf(struct inode *inode, const struct qstr *name) 1642 { 1643 struct buffer_head *bh, *obh; 1644 struct gfs2_inode *ip = GFS2_I(inode); 1645 struct gfs2_leaf *leaf, *oleaf; 1646 u32 dist = 1; 1647 int error; 1648 u32 index; 1649 u64 bn; 1650 1651 index = name->hash >> (32 - ip->i_depth); 1652 error = get_first_leaf(ip, index, &obh); 1653 if (error) 1654 return error; 1655 do { 1656 dist++; 1657 oleaf = (struct gfs2_leaf *)obh->b_data; 1658 bn = be64_to_cpu(oleaf->lf_next); 1659 if (!bn) 1660 break; 1661 brelse(obh); 1662 error = get_leaf(ip, bn, &obh); 1663 if (error) 1664 return error; 1665 } while(1); 1666 1667 gfs2_trans_add_meta(ip->i_gl, obh); 1668 1669 leaf = new_leaf(inode, &bh, be16_to_cpu(oleaf->lf_depth)); 1670 if (!leaf) { 1671 brelse(obh); 1672 return -ENOSPC; 1673 } 1674 leaf->lf_dist = cpu_to_be32(dist); 1675 oleaf->lf_next = cpu_to_be64(bh->b_blocknr); 1676 brelse(bh); 1677 brelse(obh); 1678 1679 error = gfs2_meta_inode_buffer(ip, &bh); 1680 if (error) 1681 return error; 1682 gfs2_trans_add_meta(ip->i_gl, bh); 1683 gfs2_add_inode_blocks(&ip->i_inode, 1); 1684 gfs2_dinode_out(ip, bh->b_data); 1685 brelse(bh); 1686 return 0; 1687 } 1688 1689 static u16 gfs2_inode_ra_len(const struct gfs2_inode *ip) 1690 { 1691 u64 where = ip->i_no_addr + 1; 1692 if (ip->i_eattr == where) 1693 return 1; 1694 return 0; 1695 } 1696 1697 /** 1698 * gfs2_dir_add - Add new filename into directory 1699 * @inode: The directory inode 1700 * @name: The new name 1701 * @nip: The GFS2 inode to be linked in to the directory 1702 * @da: The directory addition info 1703 * 1704 * If the call to gfs2_diradd_alloc_required resulted in there being 1705 * no need to allocate any new directory blocks, then it will contain 1706 * a pointer to the directory entry and the bh in which it resides. We 1707 * can use that without having to repeat the search. If there was no 1708 * free space, then we must now create more space. 1709 * 1710 * Returns: 0 on success, error code on failure 1711 */ 1712 1713 int gfs2_dir_add(struct inode *inode, const struct qstr *name, 1714 const struct gfs2_inode *nip, struct gfs2_diradd *da) 1715 { 1716 struct gfs2_inode *ip = GFS2_I(inode); 1717 struct buffer_head *bh = da->bh; 1718 struct gfs2_dirent *dent = da->dent; 1719 struct timespec tv; 1720 struct gfs2_leaf *leaf; 1721 int error; 1722 1723 while(1) { 1724 if (da->bh == NULL) { 1725 dent = gfs2_dirent_search(inode, name, 1726 gfs2_dirent_find_space, &bh); 1727 } 1728 if (dent) { 1729 if (IS_ERR(dent)) 1730 return PTR_ERR(dent); 1731 dent = gfs2_init_dirent(inode, dent, name, bh); 1732 gfs2_inum_out(nip, dent); 1733 dent->de_type = cpu_to_be16(IF2DT(nip->i_inode.i_mode)); 1734 dent->de_rahead = cpu_to_be16(gfs2_inode_ra_len(nip)); 1735 tv = CURRENT_TIME; 1736 if (ip->i_diskflags & GFS2_DIF_EXHASH) { 1737 leaf = (struct gfs2_leaf *)bh->b_data; 1738 be16_add_cpu(&leaf->lf_entries, 1); 1739 leaf->lf_nsec = cpu_to_be32(tv.tv_nsec); 1740 leaf->lf_sec = cpu_to_be64(tv.tv_sec); 1741 } 1742 da->dent = NULL; 1743 da->bh = NULL; 1744 brelse(bh); 1745 ip->i_entries++; 1746 ip->i_inode.i_mtime = ip->i_inode.i_ctime = tv; 1747 if (S_ISDIR(nip->i_inode.i_mode)) 1748 inc_nlink(&ip->i_inode); 1749 mark_inode_dirty(inode); 1750 error = 0; 1751 break; 1752 } 1753 if (!(ip->i_diskflags & GFS2_DIF_EXHASH)) { 1754 error = dir_make_exhash(inode); 1755 if (error) 1756 break; 1757 continue; 1758 } 1759 error = dir_split_leaf(inode, name); 1760 if (error == 0) 1761 continue; 1762 if (error < 0) 1763 break; 1764 if (ip->i_depth < GFS2_DIR_MAX_DEPTH) { 1765 error = dir_double_exhash(ip); 1766 if (error) 1767 break; 1768 error = dir_split_leaf(inode, name); 1769 if (error < 0) 1770 break; 1771 if (error == 0) 1772 continue; 1773 } 1774 error = dir_new_leaf(inode, name); 1775 if (!error) 1776 continue; 1777 error = -ENOSPC; 1778 break; 1779 } 1780 return error; 1781 } 1782 1783 1784 /** 1785 * gfs2_dir_del - Delete a directory entry 1786 * @dip: The GFS2 inode 1787 * @filename: The filename 1788 * 1789 * Returns: 0 on success, error code on failure 1790 */ 1791 1792 int gfs2_dir_del(struct gfs2_inode *dip, const struct dentry *dentry) 1793 { 1794 const struct qstr *name = &dentry->d_name; 1795 struct gfs2_dirent *dent, *prev = NULL; 1796 struct buffer_head *bh; 1797 struct timespec tv = CURRENT_TIME; 1798 1799 /* Returns _either_ the entry (if its first in block) or the 1800 previous entry otherwise */ 1801 dent = gfs2_dirent_search(&dip->i_inode, name, gfs2_dirent_prev, &bh); 1802 if (!dent) { 1803 gfs2_consist_inode(dip); 1804 return -EIO; 1805 } 1806 if (IS_ERR(dent)) { 1807 gfs2_consist_inode(dip); 1808 return PTR_ERR(dent); 1809 } 1810 /* If not first in block, adjust pointers accordingly */ 1811 if (gfs2_dirent_find(dent, name, NULL) == 0) { 1812 prev = dent; 1813 dent = (struct gfs2_dirent *)((char *)dent + be16_to_cpu(prev->de_rec_len)); 1814 } 1815 1816 dirent_del(dip, bh, prev, dent); 1817 if (dip->i_diskflags & GFS2_DIF_EXHASH) { 1818 struct gfs2_leaf *leaf = (struct gfs2_leaf *)bh->b_data; 1819 u16 entries = be16_to_cpu(leaf->lf_entries); 1820 if (!entries) 1821 gfs2_consist_inode(dip); 1822 leaf->lf_entries = cpu_to_be16(--entries); 1823 leaf->lf_nsec = cpu_to_be32(tv.tv_nsec); 1824 leaf->lf_sec = cpu_to_be64(tv.tv_sec); 1825 } 1826 brelse(bh); 1827 1828 if (!dip->i_entries) 1829 gfs2_consist_inode(dip); 1830 dip->i_entries--; 1831 dip->i_inode.i_mtime = dip->i_inode.i_ctime = tv; 1832 if (S_ISDIR(dentry->d_inode->i_mode)) 1833 drop_nlink(&dip->i_inode); 1834 mark_inode_dirty(&dip->i_inode); 1835 1836 return 0; 1837 } 1838 1839 /** 1840 * gfs2_dir_mvino - Change inode number of directory entry 1841 * @dip: The GFS2 inode 1842 * @filename: 1843 * @new_inode: 1844 * 1845 * This routine changes the inode number of a directory entry. It's used 1846 * by rename to change ".." when a directory is moved. 1847 * Assumes a glock is held on dvp. 1848 * 1849 * Returns: errno 1850 */ 1851 1852 int gfs2_dir_mvino(struct gfs2_inode *dip, const struct qstr *filename, 1853 const struct gfs2_inode *nip, unsigned int new_type) 1854 { 1855 struct buffer_head *bh; 1856 struct gfs2_dirent *dent; 1857 int error; 1858 1859 dent = gfs2_dirent_search(&dip->i_inode, filename, gfs2_dirent_find, &bh); 1860 if (!dent) { 1861 gfs2_consist_inode(dip); 1862 return -EIO; 1863 } 1864 if (IS_ERR(dent)) 1865 return PTR_ERR(dent); 1866 1867 gfs2_trans_add_meta(dip->i_gl, bh); 1868 gfs2_inum_out(nip, dent); 1869 dent->de_type = cpu_to_be16(new_type); 1870 1871 if (dip->i_diskflags & GFS2_DIF_EXHASH) { 1872 brelse(bh); 1873 error = gfs2_meta_inode_buffer(dip, &bh); 1874 if (error) 1875 return error; 1876 gfs2_trans_add_meta(dip->i_gl, bh); 1877 } 1878 1879 dip->i_inode.i_mtime = dip->i_inode.i_ctime = CURRENT_TIME; 1880 gfs2_dinode_out(dip, bh->b_data); 1881 brelse(bh); 1882 return 0; 1883 } 1884 1885 /** 1886 * leaf_dealloc - Deallocate a directory leaf 1887 * @dip: the directory 1888 * @index: the hash table offset in the directory 1889 * @len: the number of pointers to this leaf 1890 * @leaf_no: the leaf number 1891 * @leaf_bh: buffer_head for the starting leaf 1892 * last_dealloc: 1 if this is the final dealloc for the leaf, else 0 1893 * 1894 * Returns: errno 1895 */ 1896 1897 static int leaf_dealloc(struct gfs2_inode *dip, u32 index, u32 len, 1898 u64 leaf_no, struct buffer_head *leaf_bh, 1899 int last_dealloc) 1900 { 1901 struct gfs2_sbd *sdp = GFS2_SB(&dip->i_inode); 1902 struct gfs2_leaf *tmp_leaf; 1903 struct gfs2_rgrp_list rlist; 1904 struct buffer_head *bh, *dibh; 1905 u64 blk, nblk; 1906 unsigned int rg_blocks = 0, l_blocks = 0; 1907 char *ht; 1908 unsigned int x, size = len * sizeof(u64); 1909 int error; 1910 1911 error = gfs2_rindex_update(sdp); 1912 if (error) 1913 return error; 1914 1915 memset(&rlist, 0, sizeof(struct gfs2_rgrp_list)); 1916 1917 ht = kzalloc(size, GFP_NOFS | __GFP_NOWARN); 1918 if (ht == NULL) 1919 ht = vzalloc(size); 1920 if (!ht) 1921 return -ENOMEM; 1922 1923 error = gfs2_quota_hold(dip, NO_UID_QUOTA_CHANGE, NO_GID_QUOTA_CHANGE); 1924 if (error) 1925 goto out; 1926 1927 /* Count the number of leaves */ 1928 bh = leaf_bh; 1929 1930 for (blk = leaf_no; blk; blk = nblk) { 1931 if (blk != leaf_no) { 1932 error = get_leaf(dip, blk, &bh); 1933 if (error) 1934 goto out_rlist; 1935 } 1936 tmp_leaf = (struct gfs2_leaf *)bh->b_data; 1937 nblk = be64_to_cpu(tmp_leaf->lf_next); 1938 if (blk != leaf_no) 1939 brelse(bh); 1940 1941 gfs2_rlist_add(dip, &rlist, blk); 1942 l_blocks++; 1943 } 1944 1945 gfs2_rlist_alloc(&rlist, LM_ST_EXCLUSIVE); 1946 1947 for (x = 0; x < rlist.rl_rgrps; x++) { 1948 struct gfs2_rgrpd *rgd; 1949 rgd = rlist.rl_ghs[x].gh_gl->gl_object; 1950 rg_blocks += rgd->rd_length; 1951 } 1952 1953 error = gfs2_glock_nq_m(rlist.rl_rgrps, rlist.rl_ghs); 1954 if (error) 1955 goto out_rlist; 1956 1957 error = gfs2_trans_begin(sdp, 1958 rg_blocks + (DIV_ROUND_UP(size, sdp->sd_jbsize) + 1) + 1959 RES_DINODE + RES_STATFS + RES_QUOTA, l_blocks); 1960 if (error) 1961 goto out_rg_gunlock; 1962 1963 bh = leaf_bh; 1964 1965 for (blk = leaf_no; blk; blk = nblk) { 1966 if (blk != leaf_no) { 1967 error = get_leaf(dip, blk, &bh); 1968 if (error) 1969 goto out_end_trans; 1970 } 1971 tmp_leaf = (struct gfs2_leaf *)bh->b_data; 1972 nblk = be64_to_cpu(tmp_leaf->lf_next); 1973 if (blk != leaf_no) 1974 brelse(bh); 1975 1976 gfs2_free_meta(dip, blk, 1); 1977 gfs2_add_inode_blocks(&dip->i_inode, -1); 1978 } 1979 1980 error = gfs2_dir_write_data(dip, ht, index * sizeof(u64), size); 1981 if (error != size) { 1982 if (error >= 0) 1983 error = -EIO; 1984 goto out_end_trans; 1985 } 1986 1987 error = gfs2_meta_inode_buffer(dip, &dibh); 1988 if (error) 1989 goto out_end_trans; 1990 1991 gfs2_trans_add_meta(dip->i_gl, dibh); 1992 /* On the last dealloc, make this a regular file in case we crash. 1993 (We don't want to free these blocks a second time.) */ 1994 if (last_dealloc) 1995 dip->i_inode.i_mode = S_IFREG; 1996 gfs2_dinode_out(dip, dibh->b_data); 1997 brelse(dibh); 1998 1999 out_end_trans: 2000 gfs2_trans_end(sdp); 2001 out_rg_gunlock: 2002 gfs2_glock_dq_m(rlist.rl_rgrps, rlist.rl_ghs); 2003 out_rlist: 2004 gfs2_rlist_free(&rlist); 2005 gfs2_quota_unhold(dip); 2006 out: 2007 if (is_vmalloc_addr(ht)) 2008 vfree(ht); 2009 else 2010 kfree(ht); 2011 return error; 2012 } 2013 2014 /** 2015 * gfs2_dir_exhash_dealloc - free all the leaf blocks in a directory 2016 * @dip: the directory 2017 * 2018 * Dealloc all on-disk directory leaves to FREEMETA state 2019 * Change on-disk inode type to "regular file" 2020 * 2021 * Returns: errno 2022 */ 2023 2024 int gfs2_dir_exhash_dealloc(struct gfs2_inode *dip) 2025 { 2026 struct buffer_head *bh; 2027 struct gfs2_leaf *leaf; 2028 u32 hsize, len; 2029 u32 index = 0, next_index; 2030 __be64 *lp; 2031 u64 leaf_no; 2032 int error = 0, last; 2033 2034 hsize = 1 << dip->i_depth; 2035 2036 lp = gfs2_dir_get_hash_table(dip); 2037 if (IS_ERR(lp)) 2038 return PTR_ERR(lp); 2039 2040 while (index < hsize) { 2041 leaf_no = be64_to_cpu(lp[index]); 2042 if (leaf_no) { 2043 error = get_leaf(dip, leaf_no, &bh); 2044 if (error) 2045 goto out; 2046 leaf = (struct gfs2_leaf *)bh->b_data; 2047 len = 1 << (dip->i_depth - be16_to_cpu(leaf->lf_depth)); 2048 2049 next_index = (index & ~(len - 1)) + len; 2050 last = ((next_index >= hsize) ? 1 : 0); 2051 error = leaf_dealloc(dip, index, len, leaf_no, bh, 2052 last); 2053 brelse(bh); 2054 if (error) 2055 goto out; 2056 index = next_index; 2057 } else 2058 index++; 2059 } 2060 2061 if (index != hsize) { 2062 gfs2_consist_inode(dip); 2063 error = -EIO; 2064 } 2065 2066 out: 2067 2068 return error; 2069 } 2070 2071 /** 2072 * gfs2_diradd_alloc_required - find if adding entry will require an allocation 2073 * @ip: the file being written to 2074 * @filname: the filename that's going to be added 2075 * @da: The structure to return dir alloc info 2076 * 2077 * Returns: 0 if ok, -ve on error 2078 */ 2079 2080 int gfs2_diradd_alloc_required(struct inode *inode, const struct qstr *name, 2081 struct gfs2_diradd *da) 2082 { 2083 struct gfs2_inode *ip = GFS2_I(inode); 2084 struct gfs2_sbd *sdp = GFS2_SB(inode); 2085 const unsigned int extra = sizeof(struct gfs2_dinode) - sizeof(struct gfs2_leaf); 2086 struct gfs2_dirent *dent; 2087 struct buffer_head *bh; 2088 2089 da->nr_blocks = 0; 2090 da->bh = NULL; 2091 da->dent = NULL; 2092 2093 dent = gfs2_dirent_search(inode, name, gfs2_dirent_find_space, &bh); 2094 if (!dent) { 2095 da->nr_blocks = sdp->sd_max_dirres; 2096 if (!(ip->i_diskflags & GFS2_DIF_EXHASH) && 2097 (GFS2_DIRENT_SIZE(name->len) < extra)) 2098 da->nr_blocks = 1; 2099 return 0; 2100 } 2101 if (IS_ERR(dent)) 2102 return PTR_ERR(dent); 2103 2104 if (da->save_loc) { 2105 da->bh = bh; 2106 da->dent = dent; 2107 } else { 2108 brelse(bh); 2109 } 2110 return 0; 2111 } 2112 2113