1 /* 2 * Copyright (C) Sistina Software, Inc. 1997-2003 All rights reserved. 3 * Copyright (C) 2004-2006 Red Hat, Inc. All rights reserved. 4 * 5 * This copyrighted material is made available to anyone wishing to use, 6 * modify, copy, or redistribute it subject to the terms and conditions 7 * of the GNU General Public License version 2. 8 */ 9 10 /* 11 * Implements Extendible Hashing as described in: 12 * "Extendible Hashing" by Fagin, et al in 13 * __ACM Trans. on Database Systems__, Sept 1979. 14 * 15 * 16 * Here's the layout of dirents which is essentially the same as that of ext2 17 * within a single block. The field de_name_len is the number of bytes 18 * actually required for the name (no null terminator). The field de_rec_len 19 * is the number of bytes allocated to the dirent. The offset of the next 20 * dirent in the block is (dirent + dirent->de_rec_len). When a dirent is 21 * deleted, the preceding dirent inherits its allocated space, ie 22 * prev->de_rec_len += deleted->de_rec_len. Since the next dirent is obtained 23 * by adding de_rec_len to the current dirent, this essentially causes the 24 * deleted dirent to get jumped over when iterating through all the dirents. 25 * 26 * When deleting the first dirent in a block, there is no previous dirent so 27 * the field de_ino is set to zero to designate it as deleted. When allocating 28 * a dirent, gfs2_dirent_alloc iterates through the dirents in a block. If the 29 * first dirent has (de_ino == 0) and de_rec_len is large enough, this first 30 * dirent is allocated. Otherwise it must go through all the 'used' dirents 31 * searching for one in which the amount of total space minus the amount of 32 * used space will provide enough space for the new dirent. 33 * 34 * There are two types of blocks in which dirents reside. In a stuffed dinode, 35 * the dirents begin at offset sizeof(struct gfs2_dinode) from the beginning of 36 * the block. In leaves, they begin at offset sizeof(struct gfs2_leaf) from the 37 * beginning of the leaf block. The dirents reside in leaves when 38 * 39 * dip->i_diskflags & GFS2_DIF_EXHASH is true 40 * 41 * Otherwise, the dirents are "linear", within a single stuffed dinode block. 42 * 43 * When the dirents are in leaves, the actual contents of the directory file are 44 * used as an array of 64-bit block pointers pointing to the leaf blocks. The 45 * dirents are NOT in the directory file itself. There can be more than one 46 * block pointer in the array that points to the same leaf. In fact, when a 47 * directory is first converted from linear to exhash, all of the pointers 48 * point to the same leaf. 49 * 50 * When a leaf is completely full, the size of the hash table can be 51 * doubled unless it is already at the maximum size which is hard coded into 52 * GFS2_DIR_MAX_DEPTH. After that, leaves are chained together in a linked list, 53 * but never before the maximum hash table size has been reached. 54 */ 55 56 #include <linux/slab.h> 57 #include <linux/spinlock.h> 58 #include <linux/buffer_head.h> 59 #include <linux/sort.h> 60 #include <linux/gfs2_ondisk.h> 61 #include <linux/crc32.h> 62 #include <linux/vmalloc.h> 63 64 #include "gfs2.h" 65 #include "incore.h" 66 #include "dir.h" 67 #include "glock.h" 68 #include "inode.h" 69 #include "meta_io.h" 70 #include "quota.h" 71 #include "rgrp.h" 72 #include "trans.h" 73 #include "bmap.h" 74 #include "util.h" 75 76 #define IS_LEAF 1 /* Hashed (leaf) directory */ 77 #define IS_DINODE 2 /* Linear (stuffed dinode block) directory */ 78 79 #define MAX_RA_BLOCKS 32 /* max read-ahead blocks */ 80 81 #define gfs2_disk_hash2offset(h) (((u64)(h)) >> 1) 82 #define gfs2_dir_offset2hash(p) ((u32)(((u64)(p)) << 1)) 83 84 struct qstr gfs2_qdot __read_mostly; 85 struct qstr gfs2_qdotdot __read_mostly; 86 87 typedef int (*gfs2_dscan_t)(const struct gfs2_dirent *dent, 88 const struct qstr *name, void *opaque); 89 90 int gfs2_dir_get_new_buffer(struct gfs2_inode *ip, u64 block, 91 struct buffer_head **bhp) 92 { 93 struct buffer_head *bh; 94 95 bh = gfs2_meta_new(ip->i_gl, block); 96 gfs2_trans_add_meta(ip->i_gl, bh); 97 gfs2_metatype_set(bh, GFS2_METATYPE_JD, GFS2_FORMAT_JD); 98 gfs2_buffer_clear_tail(bh, sizeof(struct gfs2_meta_header)); 99 *bhp = bh; 100 return 0; 101 } 102 103 static int gfs2_dir_get_existing_buffer(struct gfs2_inode *ip, u64 block, 104 struct buffer_head **bhp) 105 { 106 struct buffer_head *bh; 107 int error; 108 109 error = gfs2_meta_read(ip->i_gl, block, DIO_WAIT, &bh); 110 if (error) 111 return error; 112 if (gfs2_metatype_check(GFS2_SB(&ip->i_inode), bh, GFS2_METATYPE_JD)) { 113 brelse(bh); 114 return -EIO; 115 } 116 *bhp = bh; 117 return 0; 118 } 119 120 static int gfs2_dir_write_stuffed(struct gfs2_inode *ip, const char *buf, 121 unsigned int offset, unsigned int size) 122 { 123 struct buffer_head *dibh; 124 int error; 125 126 error = gfs2_meta_inode_buffer(ip, &dibh); 127 if (error) 128 return error; 129 130 gfs2_trans_add_meta(ip->i_gl, dibh); 131 memcpy(dibh->b_data + offset + sizeof(struct gfs2_dinode), buf, size); 132 if (ip->i_inode.i_size < offset + size) 133 i_size_write(&ip->i_inode, offset + size); 134 ip->i_inode.i_mtime = ip->i_inode.i_ctime = CURRENT_TIME; 135 gfs2_dinode_out(ip, dibh->b_data); 136 137 brelse(dibh); 138 139 return size; 140 } 141 142 143 144 /** 145 * gfs2_dir_write_data - Write directory information to the inode 146 * @ip: The GFS2 inode 147 * @buf: The buffer containing information to be written 148 * @offset: The file offset to start writing at 149 * @size: The amount of data to write 150 * 151 * Returns: The number of bytes correctly written or error code 152 */ 153 static int gfs2_dir_write_data(struct gfs2_inode *ip, const char *buf, 154 u64 offset, unsigned int size) 155 { 156 struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode); 157 struct buffer_head *dibh; 158 u64 lblock, dblock; 159 u32 extlen = 0; 160 unsigned int o; 161 int copied = 0; 162 int error = 0; 163 int new = 0; 164 165 if (!size) 166 return 0; 167 168 if (gfs2_is_stuffed(ip) && 169 offset + size <= sdp->sd_sb.sb_bsize - sizeof(struct gfs2_dinode)) 170 return gfs2_dir_write_stuffed(ip, buf, (unsigned int)offset, 171 size); 172 173 if (gfs2_assert_warn(sdp, gfs2_is_jdata(ip))) 174 return -EINVAL; 175 176 if (gfs2_is_stuffed(ip)) { 177 error = gfs2_unstuff_dinode(ip, NULL); 178 if (error) 179 return error; 180 } 181 182 lblock = offset; 183 o = do_div(lblock, sdp->sd_jbsize) + sizeof(struct gfs2_meta_header); 184 185 while (copied < size) { 186 unsigned int amount; 187 struct buffer_head *bh; 188 189 amount = size - copied; 190 if (amount > sdp->sd_sb.sb_bsize - o) 191 amount = sdp->sd_sb.sb_bsize - o; 192 193 if (!extlen) { 194 new = 1; 195 error = gfs2_extent_map(&ip->i_inode, lblock, &new, 196 &dblock, &extlen); 197 if (error) 198 goto fail; 199 error = -EIO; 200 if (gfs2_assert_withdraw(sdp, dblock)) 201 goto fail; 202 } 203 204 if (amount == sdp->sd_jbsize || new) 205 error = gfs2_dir_get_new_buffer(ip, dblock, &bh); 206 else 207 error = gfs2_dir_get_existing_buffer(ip, dblock, &bh); 208 209 if (error) 210 goto fail; 211 212 gfs2_trans_add_meta(ip->i_gl, bh); 213 memcpy(bh->b_data + o, buf, amount); 214 brelse(bh); 215 216 buf += amount; 217 copied += amount; 218 lblock++; 219 dblock++; 220 extlen--; 221 222 o = sizeof(struct gfs2_meta_header); 223 } 224 225 out: 226 error = gfs2_meta_inode_buffer(ip, &dibh); 227 if (error) 228 return error; 229 230 if (ip->i_inode.i_size < offset + copied) 231 i_size_write(&ip->i_inode, offset + copied); 232 ip->i_inode.i_mtime = ip->i_inode.i_ctime = CURRENT_TIME; 233 234 gfs2_trans_add_meta(ip->i_gl, dibh); 235 gfs2_dinode_out(ip, dibh->b_data); 236 brelse(dibh); 237 238 return copied; 239 fail: 240 if (copied) 241 goto out; 242 return error; 243 } 244 245 static int gfs2_dir_read_stuffed(struct gfs2_inode *ip, __be64 *buf, 246 unsigned int size) 247 { 248 struct buffer_head *dibh; 249 int error; 250 251 error = gfs2_meta_inode_buffer(ip, &dibh); 252 if (!error) { 253 memcpy(buf, dibh->b_data + sizeof(struct gfs2_dinode), size); 254 brelse(dibh); 255 } 256 257 return (error) ? error : size; 258 } 259 260 261 /** 262 * gfs2_dir_read_data - Read a data from a directory inode 263 * @ip: The GFS2 Inode 264 * @buf: The buffer to place result into 265 * @size: Amount of data to transfer 266 * 267 * Returns: The amount of data actually copied or the error 268 */ 269 static int gfs2_dir_read_data(struct gfs2_inode *ip, __be64 *buf, 270 unsigned int size) 271 { 272 struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode); 273 u64 lblock, dblock; 274 u32 extlen = 0; 275 unsigned int o; 276 int copied = 0; 277 int error = 0; 278 279 if (gfs2_is_stuffed(ip)) 280 return gfs2_dir_read_stuffed(ip, buf, size); 281 282 if (gfs2_assert_warn(sdp, gfs2_is_jdata(ip))) 283 return -EINVAL; 284 285 lblock = 0; 286 o = do_div(lblock, sdp->sd_jbsize) + sizeof(struct gfs2_meta_header); 287 288 while (copied < size) { 289 unsigned int amount; 290 struct buffer_head *bh; 291 int new; 292 293 amount = size - copied; 294 if (amount > sdp->sd_sb.sb_bsize - o) 295 amount = sdp->sd_sb.sb_bsize - o; 296 297 if (!extlen) { 298 new = 0; 299 error = gfs2_extent_map(&ip->i_inode, lblock, &new, 300 &dblock, &extlen); 301 if (error || !dblock) 302 goto fail; 303 BUG_ON(extlen < 1); 304 bh = gfs2_meta_ra(ip->i_gl, dblock, extlen); 305 } else { 306 error = gfs2_meta_read(ip->i_gl, dblock, DIO_WAIT, &bh); 307 if (error) 308 goto fail; 309 } 310 error = gfs2_metatype_check(sdp, bh, GFS2_METATYPE_JD); 311 if (error) { 312 brelse(bh); 313 goto fail; 314 } 315 dblock++; 316 extlen--; 317 memcpy(buf, bh->b_data + o, amount); 318 brelse(bh); 319 buf += (amount/sizeof(__be64)); 320 copied += amount; 321 lblock++; 322 o = sizeof(struct gfs2_meta_header); 323 } 324 325 return copied; 326 fail: 327 return (copied) ? copied : error; 328 } 329 330 /** 331 * gfs2_dir_get_hash_table - Get pointer to the dir hash table 332 * @ip: The inode in question 333 * 334 * Returns: The hash table or an error 335 */ 336 337 static __be64 *gfs2_dir_get_hash_table(struct gfs2_inode *ip) 338 { 339 struct inode *inode = &ip->i_inode; 340 int ret; 341 u32 hsize; 342 __be64 *hc; 343 344 BUG_ON(!(ip->i_diskflags & GFS2_DIF_EXHASH)); 345 346 hc = ip->i_hash_cache; 347 if (hc) 348 return hc; 349 350 hsize = 1 << ip->i_depth; 351 hsize *= sizeof(__be64); 352 if (hsize != i_size_read(&ip->i_inode)) { 353 gfs2_consist_inode(ip); 354 return ERR_PTR(-EIO); 355 } 356 357 hc = kmalloc(hsize, GFP_NOFS | __GFP_NOWARN); 358 if (hc == NULL) 359 hc = __vmalloc(hsize, GFP_NOFS, PAGE_KERNEL); 360 361 if (hc == NULL) 362 return ERR_PTR(-ENOMEM); 363 364 ret = gfs2_dir_read_data(ip, hc, hsize); 365 if (ret < 0) { 366 if (is_vmalloc_addr(hc)) 367 vfree(hc); 368 else 369 kfree(hc); 370 return ERR_PTR(ret); 371 } 372 373 spin_lock(&inode->i_lock); 374 if (ip->i_hash_cache) { 375 if (is_vmalloc_addr(hc)) 376 vfree(hc); 377 else 378 kfree(hc); 379 } else { 380 ip->i_hash_cache = hc; 381 } 382 spin_unlock(&inode->i_lock); 383 384 return ip->i_hash_cache; 385 } 386 387 /** 388 * gfs2_dir_hash_inval - Invalidate dir hash 389 * @ip: The directory inode 390 * 391 * Must be called with an exclusive glock, or during glock invalidation. 392 */ 393 void gfs2_dir_hash_inval(struct gfs2_inode *ip) 394 { 395 __be64 *hc = ip->i_hash_cache; 396 ip->i_hash_cache = NULL; 397 if (is_vmalloc_addr(hc)) 398 vfree(hc); 399 else 400 kfree(hc); 401 } 402 403 static inline int gfs2_dirent_sentinel(const struct gfs2_dirent *dent) 404 { 405 return dent->de_inum.no_addr == 0 || dent->de_inum.no_formal_ino == 0; 406 } 407 408 static inline int __gfs2_dirent_find(const struct gfs2_dirent *dent, 409 const struct qstr *name, int ret) 410 { 411 if (!gfs2_dirent_sentinel(dent) && 412 be32_to_cpu(dent->de_hash) == name->hash && 413 be16_to_cpu(dent->de_name_len) == name->len && 414 memcmp(dent+1, name->name, name->len) == 0) 415 return ret; 416 return 0; 417 } 418 419 static int gfs2_dirent_find(const struct gfs2_dirent *dent, 420 const struct qstr *name, 421 void *opaque) 422 { 423 return __gfs2_dirent_find(dent, name, 1); 424 } 425 426 static int gfs2_dirent_prev(const struct gfs2_dirent *dent, 427 const struct qstr *name, 428 void *opaque) 429 { 430 return __gfs2_dirent_find(dent, name, 2); 431 } 432 433 /* 434 * name->name holds ptr to start of block. 435 * name->len holds size of block. 436 */ 437 static int gfs2_dirent_last(const struct gfs2_dirent *dent, 438 const struct qstr *name, 439 void *opaque) 440 { 441 const char *start = name->name; 442 const char *end = (const char *)dent + be16_to_cpu(dent->de_rec_len); 443 if (name->len == (end - start)) 444 return 1; 445 return 0; 446 } 447 448 static int gfs2_dirent_find_space(const struct gfs2_dirent *dent, 449 const struct qstr *name, 450 void *opaque) 451 { 452 unsigned required = GFS2_DIRENT_SIZE(name->len); 453 unsigned actual = GFS2_DIRENT_SIZE(be16_to_cpu(dent->de_name_len)); 454 unsigned totlen = be16_to_cpu(dent->de_rec_len); 455 456 if (gfs2_dirent_sentinel(dent)) 457 actual = 0; 458 if (totlen - actual >= required) 459 return 1; 460 return 0; 461 } 462 463 struct dirent_gather { 464 const struct gfs2_dirent **pdent; 465 unsigned offset; 466 }; 467 468 static int gfs2_dirent_gather(const struct gfs2_dirent *dent, 469 const struct qstr *name, 470 void *opaque) 471 { 472 struct dirent_gather *g = opaque; 473 if (!gfs2_dirent_sentinel(dent)) { 474 g->pdent[g->offset++] = dent; 475 } 476 return 0; 477 } 478 479 /* 480 * Other possible things to check: 481 * - Inode located within filesystem size (and on valid block) 482 * - Valid directory entry type 483 * Not sure how heavy-weight we want to make this... could also check 484 * hash is correct for example, but that would take a lot of extra time. 485 * For now the most important thing is to check that the various sizes 486 * are correct. 487 */ 488 static int gfs2_check_dirent(struct gfs2_dirent *dent, unsigned int offset, 489 unsigned int size, unsigned int len, int first) 490 { 491 const char *msg = "gfs2_dirent too small"; 492 if (unlikely(size < sizeof(struct gfs2_dirent))) 493 goto error; 494 msg = "gfs2_dirent misaligned"; 495 if (unlikely(offset & 0x7)) 496 goto error; 497 msg = "gfs2_dirent points beyond end of block"; 498 if (unlikely(offset + size > len)) 499 goto error; 500 msg = "zero inode number"; 501 if (unlikely(!first && gfs2_dirent_sentinel(dent))) 502 goto error; 503 msg = "name length is greater than space in dirent"; 504 if (!gfs2_dirent_sentinel(dent) && 505 unlikely(sizeof(struct gfs2_dirent)+be16_to_cpu(dent->de_name_len) > 506 size)) 507 goto error; 508 return 0; 509 error: 510 printk(KERN_WARNING "gfs2_check_dirent: %s (%s)\n", msg, 511 first ? "first in block" : "not first in block"); 512 return -EIO; 513 } 514 515 static int gfs2_dirent_offset(const void *buf) 516 { 517 const struct gfs2_meta_header *h = buf; 518 int offset; 519 520 BUG_ON(buf == NULL); 521 522 switch(be32_to_cpu(h->mh_type)) { 523 case GFS2_METATYPE_LF: 524 offset = sizeof(struct gfs2_leaf); 525 break; 526 case GFS2_METATYPE_DI: 527 offset = sizeof(struct gfs2_dinode); 528 break; 529 default: 530 goto wrong_type; 531 } 532 return offset; 533 wrong_type: 534 printk(KERN_WARNING "gfs2_scan_dirent: wrong block type %u\n", 535 be32_to_cpu(h->mh_type)); 536 return -1; 537 } 538 539 static struct gfs2_dirent *gfs2_dirent_scan(struct inode *inode, void *buf, 540 unsigned int len, gfs2_dscan_t scan, 541 const struct qstr *name, 542 void *opaque) 543 { 544 struct gfs2_dirent *dent, *prev; 545 unsigned offset; 546 unsigned size; 547 int ret = 0; 548 549 ret = gfs2_dirent_offset(buf); 550 if (ret < 0) 551 goto consist_inode; 552 553 offset = ret; 554 prev = NULL; 555 dent = buf + offset; 556 size = be16_to_cpu(dent->de_rec_len); 557 if (gfs2_check_dirent(dent, offset, size, len, 1)) 558 goto consist_inode; 559 do { 560 ret = scan(dent, name, opaque); 561 if (ret) 562 break; 563 offset += size; 564 if (offset == len) 565 break; 566 prev = dent; 567 dent = buf + offset; 568 size = be16_to_cpu(dent->de_rec_len); 569 if (gfs2_check_dirent(dent, offset, size, len, 0)) 570 goto consist_inode; 571 } while(1); 572 573 switch(ret) { 574 case 0: 575 return NULL; 576 case 1: 577 return dent; 578 case 2: 579 return prev ? prev : dent; 580 default: 581 BUG_ON(ret > 0); 582 return ERR_PTR(ret); 583 } 584 585 consist_inode: 586 gfs2_consist_inode(GFS2_I(inode)); 587 return ERR_PTR(-EIO); 588 } 589 590 static int dirent_check_reclen(struct gfs2_inode *dip, 591 const struct gfs2_dirent *d, const void *end_p) 592 { 593 const void *ptr = d; 594 u16 rec_len = be16_to_cpu(d->de_rec_len); 595 596 if (unlikely(rec_len < sizeof(struct gfs2_dirent))) 597 goto broken; 598 ptr += rec_len; 599 if (ptr < end_p) 600 return rec_len; 601 if (ptr == end_p) 602 return -ENOENT; 603 broken: 604 gfs2_consist_inode(dip); 605 return -EIO; 606 } 607 608 /** 609 * dirent_next - Next dirent 610 * @dip: the directory 611 * @bh: The buffer 612 * @dent: Pointer to list of dirents 613 * 614 * Returns: 0 on success, error code otherwise 615 */ 616 617 static int dirent_next(struct gfs2_inode *dip, struct buffer_head *bh, 618 struct gfs2_dirent **dent) 619 { 620 struct gfs2_dirent *cur = *dent, *tmp; 621 char *bh_end = bh->b_data + bh->b_size; 622 int ret; 623 624 ret = dirent_check_reclen(dip, cur, bh_end); 625 if (ret < 0) 626 return ret; 627 628 tmp = (void *)cur + ret; 629 ret = dirent_check_reclen(dip, tmp, bh_end); 630 if (ret == -EIO) 631 return ret; 632 633 /* Only the first dent could ever have de_inum.no_addr == 0 */ 634 if (gfs2_dirent_sentinel(tmp)) { 635 gfs2_consist_inode(dip); 636 return -EIO; 637 } 638 639 *dent = tmp; 640 return 0; 641 } 642 643 /** 644 * dirent_del - Delete a dirent 645 * @dip: The GFS2 inode 646 * @bh: The buffer 647 * @prev: The previous dirent 648 * @cur: The current dirent 649 * 650 */ 651 652 static void dirent_del(struct gfs2_inode *dip, struct buffer_head *bh, 653 struct gfs2_dirent *prev, struct gfs2_dirent *cur) 654 { 655 u16 cur_rec_len, prev_rec_len; 656 657 if (gfs2_dirent_sentinel(cur)) { 658 gfs2_consist_inode(dip); 659 return; 660 } 661 662 gfs2_trans_add_meta(dip->i_gl, bh); 663 664 /* If there is no prev entry, this is the first entry in the block. 665 The de_rec_len is already as big as it needs to be. Just zero 666 out the inode number and return. */ 667 668 if (!prev) { 669 cur->de_inum.no_addr = 0; 670 cur->de_inum.no_formal_ino = 0; 671 return; 672 } 673 674 /* Combine this dentry with the previous one. */ 675 676 prev_rec_len = be16_to_cpu(prev->de_rec_len); 677 cur_rec_len = be16_to_cpu(cur->de_rec_len); 678 679 if ((char *)prev + prev_rec_len != (char *)cur) 680 gfs2_consist_inode(dip); 681 if ((char *)cur + cur_rec_len > bh->b_data + bh->b_size) 682 gfs2_consist_inode(dip); 683 684 prev_rec_len += cur_rec_len; 685 prev->de_rec_len = cpu_to_be16(prev_rec_len); 686 } 687 688 /* 689 * Takes a dent from which to grab space as an argument. Returns the 690 * newly created dent. 691 */ 692 static struct gfs2_dirent *gfs2_init_dirent(struct inode *inode, 693 struct gfs2_dirent *dent, 694 const struct qstr *name, 695 struct buffer_head *bh) 696 { 697 struct gfs2_inode *ip = GFS2_I(inode); 698 struct gfs2_dirent *ndent; 699 unsigned offset = 0, totlen; 700 701 if (!gfs2_dirent_sentinel(dent)) 702 offset = GFS2_DIRENT_SIZE(be16_to_cpu(dent->de_name_len)); 703 totlen = be16_to_cpu(dent->de_rec_len); 704 BUG_ON(offset + name->len > totlen); 705 gfs2_trans_add_meta(ip->i_gl, bh); 706 ndent = (struct gfs2_dirent *)((char *)dent + offset); 707 dent->de_rec_len = cpu_to_be16(offset); 708 gfs2_qstr2dirent(name, totlen - offset, ndent); 709 return ndent; 710 } 711 712 static struct gfs2_dirent *gfs2_dirent_alloc(struct inode *inode, 713 struct buffer_head *bh, 714 const struct qstr *name) 715 { 716 struct gfs2_dirent *dent; 717 dent = gfs2_dirent_scan(inode, bh->b_data, bh->b_size, 718 gfs2_dirent_find_space, name, NULL); 719 if (!dent || IS_ERR(dent)) 720 return dent; 721 return gfs2_init_dirent(inode, dent, name, bh); 722 } 723 724 static int get_leaf(struct gfs2_inode *dip, u64 leaf_no, 725 struct buffer_head **bhp) 726 { 727 int error; 728 729 error = gfs2_meta_read(dip->i_gl, leaf_no, DIO_WAIT, bhp); 730 if (!error && gfs2_metatype_check(GFS2_SB(&dip->i_inode), *bhp, GFS2_METATYPE_LF)) { 731 /* printk(KERN_INFO "block num=%llu\n", leaf_no); */ 732 error = -EIO; 733 } 734 735 return error; 736 } 737 738 /** 739 * get_leaf_nr - Get a leaf number associated with the index 740 * @dip: The GFS2 inode 741 * @index: 742 * @leaf_out: 743 * 744 * Returns: 0 on success, error code otherwise 745 */ 746 747 static int get_leaf_nr(struct gfs2_inode *dip, u32 index, 748 u64 *leaf_out) 749 { 750 __be64 *hash; 751 752 hash = gfs2_dir_get_hash_table(dip); 753 if (IS_ERR(hash)) 754 return PTR_ERR(hash); 755 *leaf_out = be64_to_cpu(*(hash + index)); 756 return 0; 757 } 758 759 static int get_first_leaf(struct gfs2_inode *dip, u32 index, 760 struct buffer_head **bh_out) 761 { 762 u64 leaf_no; 763 int error; 764 765 error = get_leaf_nr(dip, index, &leaf_no); 766 if (!error) 767 error = get_leaf(dip, leaf_no, bh_out); 768 769 return error; 770 } 771 772 static struct gfs2_dirent *gfs2_dirent_search(struct inode *inode, 773 const struct qstr *name, 774 gfs2_dscan_t scan, 775 struct buffer_head **pbh) 776 { 777 struct buffer_head *bh; 778 struct gfs2_dirent *dent; 779 struct gfs2_inode *ip = GFS2_I(inode); 780 int error; 781 782 if (ip->i_diskflags & GFS2_DIF_EXHASH) { 783 struct gfs2_leaf *leaf; 784 unsigned hsize = 1 << ip->i_depth; 785 unsigned index; 786 u64 ln; 787 if (hsize * sizeof(u64) != i_size_read(inode)) { 788 gfs2_consist_inode(ip); 789 return ERR_PTR(-EIO); 790 } 791 792 index = name->hash >> (32 - ip->i_depth); 793 error = get_first_leaf(ip, index, &bh); 794 if (error) 795 return ERR_PTR(error); 796 do { 797 dent = gfs2_dirent_scan(inode, bh->b_data, bh->b_size, 798 scan, name, NULL); 799 if (dent) 800 goto got_dent; 801 leaf = (struct gfs2_leaf *)bh->b_data; 802 ln = be64_to_cpu(leaf->lf_next); 803 brelse(bh); 804 if (!ln) 805 break; 806 807 error = get_leaf(ip, ln, &bh); 808 } while(!error); 809 810 return error ? ERR_PTR(error) : NULL; 811 } 812 813 814 error = gfs2_meta_inode_buffer(ip, &bh); 815 if (error) 816 return ERR_PTR(error); 817 dent = gfs2_dirent_scan(inode, bh->b_data, bh->b_size, scan, name, NULL); 818 got_dent: 819 if (unlikely(dent == NULL || IS_ERR(dent))) { 820 brelse(bh); 821 bh = NULL; 822 } 823 *pbh = bh; 824 return dent; 825 } 826 827 static struct gfs2_leaf *new_leaf(struct inode *inode, struct buffer_head **pbh, u16 depth) 828 { 829 struct gfs2_inode *ip = GFS2_I(inode); 830 unsigned int n = 1; 831 u64 bn; 832 int error; 833 struct buffer_head *bh; 834 struct gfs2_leaf *leaf; 835 struct gfs2_dirent *dent; 836 struct qstr name = { .name = "" }; 837 struct timespec tv = CURRENT_TIME; 838 839 error = gfs2_alloc_blocks(ip, &bn, &n, 0, NULL); 840 if (error) 841 return NULL; 842 bh = gfs2_meta_new(ip->i_gl, bn); 843 if (!bh) 844 return NULL; 845 846 gfs2_trans_add_unrevoke(GFS2_SB(inode), bn, 1); 847 gfs2_trans_add_meta(ip->i_gl, bh); 848 gfs2_metatype_set(bh, GFS2_METATYPE_LF, GFS2_FORMAT_LF); 849 leaf = (struct gfs2_leaf *)bh->b_data; 850 leaf->lf_depth = cpu_to_be16(depth); 851 leaf->lf_entries = 0; 852 leaf->lf_dirent_format = cpu_to_be32(GFS2_FORMAT_DE); 853 leaf->lf_next = 0; 854 leaf->lf_inode = cpu_to_be64(ip->i_no_addr); 855 leaf->lf_dist = cpu_to_be32(1); 856 leaf->lf_nsec = cpu_to_be32(tv.tv_nsec); 857 leaf->lf_sec = cpu_to_be64(tv.tv_sec); 858 memset(leaf->lf_reserved2, 0, sizeof(leaf->lf_reserved2)); 859 dent = (struct gfs2_dirent *)(leaf+1); 860 gfs2_qstr2dirent(&name, bh->b_size - sizeof(struct gfs2_leaf), dent); 861 *pbh = bh; 862 return leaf; 863 } 864 865 /** 866 * dir_make_exhash - Convert a stuffed directory into an ExHash directory 867 * @dip: The GFS2 inode 868 * 869 * Returns: 0 on success, error code otherwise 870 */ 871 872 static int dir_make_exhash(struct inode *inode) 873 { 874 struct gfs2_inode *dip = GFS2_I(inode); 875 struct gfs2_sbd *sdp = GFS2_SB(inode); 876 struct gfs2_dirent *dent; 877 struct qstr args; 878 struct buffer_head *bh, *dibh; 879 struct gfs2_leaf *leaf; 880 int y; 881 u32 x; 882 __be64 *lp; 883 u64 bn; 884 int error; 885 886 error = gfs2_meta_inode_buffer(dip, &dibh); 887 if (error) 888 return error; 889 890 /* Turn over a new leaf */ 891 892 leaf = new_leaf(inode, &bh, 0); 893 if (!leaf) 894 return -ENOSPC; 895 bn = bh->b_blocknr; 896 897 gfs2_assert(sdp, dip->i_entries < (1 << 16)); 898 leaf->lf_entries = cpu_to_be16(dip->i_entries); 899 900 /* Copy dirents */ 901 902 gfs2_buffer_copy_tail(bh, sizeof(struct gfs2_leaf), dibh, 903 sizeof(struct gfs2_dinode)); 904 905 /* Find last entry */ 906 907 x = 0; 908 args.len = bh->b_size - sizeof(struct gfs2_dinode) + 909 sizeof(struct gfs2_leaf); 910 args.name = bh->b_data; 911 dent = gfs2_dirent_scan(&dip->i_inode, bh->b_data, bh->b_size, 912 gfs2_dirent_last, &args, NULL); 913 if (!dent) { 914 brelse(bh); 915 brelse(dibh); 916 return -EIO; 917 } 918 if (IS_ERR(dent)) { 919 brelse(bh); 920 brelse(dibh); 921 return PTR_ERR(dent); 922 } 923 924 /* Adjust the last dirent's record length 925 (Remember that dent still points to the last entry.) */ 926 927 dent->de_rec_len = cpu_to_be16(be16_to_cpu(dent->de_rec_len) + 928 sizeof(struct gfs2_dinode) - 929 sizeof(struct gfs2_leaf)); 930 931 brelse(bh); 932 933 /* We're done with the new leaf block, now setup the new 934 hash table. */ 935 936 gfs2_trans_add_meta(dip->i_gl, dibh); 937 gfs2_buffer_clear_tail(dibh, sizeof(struct gfs2_dinode)); 938 939 lp = (__be64 *)(dibh->b_data + sizeof(struct gfs2_dinode)); 940 941 for (x = sdp->sd_hash_ptrs; x--; lp++) 942 *lp = cpu_to_be64(bn); 943 944 i_size_write(inode, sdp->sd_sb.sb_bsize / 2); 945 gfs2_add_inode_blocks(&dip->i_inode, 1); 946 dip->i_diskflags |= GFS2_DIF_EXHASH; 947 948 for (x = sdp->sd_hash_ptrs, y = -1; x; x >>= 1, y++) ; 949 dip->i_depth = y; 950 951 gfs2_dinode_out(dip, dibh->b_data); 952 953 brelse(dibh); 954 955 return 0; 956 } 957 958 /** 959 * dir_split_leaf - Split a leaf block into two 960 * @dip: The GFS2 inode 961 * @index: 962 * @leaf_no: 963 * 964 * Returns: 0 on success, error code on failure 965 */ 966 967 static int dir_split_leaf(struct inode *inode, const struct qstr *name) 968 { 969 struct gfs2_inode *dip = GFS2_I(inode); 970 struct buffer_head *nbh, *obh, *dibh; 971 struct gfs2_leaf *nleaf, *oleaf; 972 struct gfs2_dirent *dent = NULL, *prev = NULL, *next = NULL, *new; 973 u32 start, len, half_len, divider; 974 u64 bn, leaf_no; 975 __be64 *lp; 976 u32 index; 977 int x, moved = 0; 978 int error; 979 980 index = name->hash >> (32 - dip->i_depth); 981 error = get_leaf_nr(dip, index, &leaf_no); 982 if (error) 983 return error; 984 985 /* Get the old leaf block */ 986 error = get_leaf(dip, leaf_no, &obh); 987 if (error) 988 return error; 989 990 oleaf = (struct gfs2_leaf *)obh->b_data; 991 if (dip->i_depth == be16_to_cpu(oleaf->lf_depth)) { 992 brelse(obh); 993 return 1; /* can't split */ 994 } 995 996 gfs2_trans_add_meta(dip->i_gl, obh); 997 998 nleaf = new_leaf(inode, &nbh, be16_to_cpu(oleaf->lf_depth) + 1); 999 if (!nleaf) { 1000 brelse(obh); 1001 return -ENOSPC; 1002 } 1003 bn = nbh->b_blocknr; 1004 1005 /* Compute the start and len of leaf pointers in the hash table. */ 1006 len = 1 << (dip->i_depth - be16_to_cpu(oleaf->lf_depth)); 1007 half_len = len >> 1; 1008 if (!half_len) { 1009 printk(KERN_WARNING "i_depth %u lf_depth %u index %u\n", dip->i_depth, be16_to_cpu(oleaf->lf_depth), index); 1010 gfs2_consist_inode(dip); 1011 error = -EIO; 1012 goto fail_brelse; 1013 } 1014 1015 start = (index & ~(len - 1)); 1016 1017 /* Change the pointers. 1018 Don't bother distinguishing stuffed from non-stuffed. 1019 This code is complicated enough already. */ 1020 lp = kmalloc(half_len * sizeof(__be64), GFP_NOFS); 1021 if (!lp) { 1022 error = -ENOMEM; 1023 goto fail_brelse; 1024 } 1025 1026 /* Change the pointers */ 1027 for (x = 0; x < half_len; x++) 1028 lp[x] = cpu_to_be64(bn); 1029 1030 gfs2_dir_hash_inval(dip); 1031 1032 error = gfs2_dir_write_data(dip, (char *)lp, start * sizeof(u64), 1033 half_len * sizeof(u64)); 1034 if (error != half_len * sizeof(u64)) { 1035 if (error >= 0) 1036 error = -EIO; 1037 goto fail_lpfree; 1038 } 1039 1040 kfree(lp); 1041 1042 /* Compute the divider */ 1043 divider = (start + half_len) << (32 - dip->i_depth); 1044 1045 /* Copy the entries */ 1046 dent = (struct gfs2_dirent *)(obh->b_data + sizeof(struct gfs2_leaf)); 1047 1048 do { 1049 next = dent; 1050 if (dirent_next(dip, obh, &next)) 1051 next = NULL; 1052 1053 if (!gfs2_dirent_sentinel(dent) && 1054 be32_to_cpu(dent->de_hash) < divider) { 1055 struct qstr str; 1056 str.name = (char*)(dent+1); 1057 str.len = be16_to_cpu(dent->de_name_len); 1058 str.hash = be32_to_cpu(dent->de_hash); 1059 new = gfs2_dirent_alloc(inode, nbh, &str); 1060 if (IS_ERR(new)) { 1061 error = PTR_ERR(new); 1062 break; 1063 } 1064 1065 new->de_inum = dent->de_inum; /* No endian worries */ 1066 new->de_type = dent->de_type; /* No endian worries */ 1067 be16_add_cpu(&nleaf->lf_entries, 1); 1068 1069 dirent_del(dip, obh, prev, dent); 1070 1071 if (!oleaf->lf_entries) 1072 gfs2_consist_inode(dip); 1073 be16_add_cpu(&oleaf->lf_entries, -1); 1074 1075 if (!prev) 1076 prev = dent; 1077 1078 moved = 1; 1079 } else { 1080 prev = dent; 1081 } 1082 dent = next; 1083 } while (dent); 1084 1085 oleaf->lf_depth = nleaf->lf_depth; 1086 1087 error = gfs2_meta_inode_buffer(dip, &dibh); 1088 if (!gfs2_assert_withdraw(GFS2_SB(&dip->i_inode), !error)) { 1089 gfs2_trans_add_meta(dip->i_gl, dibh); 1090 gfs2_add_inode_blocks(&dip->i_inode, 1); 1091 gfs2_dinode_out(dip, dibh->b_data); 1092 brelse(dibh); 1093 } 1094 1095 brelse(obh); 1096 brelse(nbh); 1097 1098 return error; 1099 1100 fail_lpfree: 1101 kfree(lp); 1102 1103 fail_brelse: 1104 brelse(obh); 1105 brelse(nbh); 1106 return error; 1107 } 1108 1109 /** 1110 * dir_double_exhash - Double size of ExHash table 1111 * @dip: The GFS2 dinode 1112 * 1113 * Returns: 0 on success, error code on failure 1114 */ 1115 1116 static int dir_double_exhash(struct gfs2_inode *dip) 1117 { 1118 struct buffer_head *dibh; 1119 u32 hsize; 1120 u32 hsize_bytes; 1121 __be64 *hc; 1122 __be64 *hc2, *h; 1123 int x; 1124 int error = 0; 1125 1126 hsize = 1 << dip->i_depth; 1127 hsize_bytes = hsize * sizeof(__be64); 1128 1129 hc = gfs2_dir_get_hash_table(dip); 1130 if (IS_ERR(hc)) 1131 return PTR_ERR(hc); 1132 1133 hc2 = kmalloc(hsize_bytes * 2, GFP_NOFS | __GFP_NOWARN); 1134 if (hc2 == NULL) 1135 hc2 = __vmalloc(hsize_bytes * 2, GFP_NOFS, PAGE_KERNEL); 1136 1137 if (!hc2) 1138 return -ENOMEM; 1139 1140 h = hc2; 1141 error = gfs2_meta_inode_buffer(dip, &dibh); 1142 if (error) 1143 goto out_kfree; 1144 1145 for (x = 0; x < hsize; x++) { 1146 *h++ = *hc; 1147 *h++ = *hc; 1148 hc++; 1149 } 1150 1151 error = gfs2_dir_write_data(dip, (char *)hc2, 0, hsize_bytes * 2); 1152 if (error != (hsize_bytes * 2)) 1153 goto fail; 1154 1155 gfs2_dir_hash_inval(dip); 1156 dip->i_hash_cache = hc2; 1157 dip->i_depth++; 1158 gfs2_dinode_out(dip, dibh->b_data); 1159 brelse(dibh); 1160 return 0; 1161 1162 fail: 1163 /* Replace original hash table & size */ 1164 gfs2_dir_write_data(dip, (char *)hc, 0, hsize_bytes); 1165 i_size_write(&dip->i_inode, hsize_bytes); 1166 gfs2_dinode_out(dip, dibh->b_data); 1167 brelse(dibh); 1168 out_kfree: 1169 if (is_vmalloc_addr(hc2)) 1170 vfree(hc2); 1171 else 1172 kfree(hc2); 1173 return error; 1174 } 1175 1176 /** 1177 * compare_dents - compare directory entries by hash value 1178 * @a: first dent 1179 * @b: second dent 1180 * 1181 * When comparing the hash entries of @a to @b: 1182 * gt: returns 1 1183 * lt: returns -1 1184 * eq: returns 0 1185 */ 1186 1187 static int compare_dents(const void *a, const void *b) 1188 { 1189 const struct gfs2_dirent *dent_a, *dent_b; 1190 u32 hash_a, hash_b; 1191 int ret = 0; 1192 1193 dent_a = *(const struct gfs2_dirent **)a; 1194 hash_a = be32_to_cpu(dent_a->de_hash); 1195 1196 dent_b = *(const struct gfs2_dirent **)b; 1197 hash_b = be32_to_cpu(dent_b->de_hash); 1198 1199 if (hash_a > hash_b) 1200 ret = 1; 1201 else if (hash_a < hash_b) 1202 ret = -1; 1203 else { 1204 unsigned int len_a = be16_to_cpu(dent_a->de_name_len); 1205 unsigned int len_b = be16_to_cpu(dent_b->de_name_len); 1206 1207 if (len_a > len_b) 1208 ret = 1; 1209 else if (len_a < len_b) 1210 ret = -1; 1211 else 1212 ret = memcmp(dent_a + 1, dent_b + 1, len_a); 1213 } 1214 1215 return ret; 1216 } 1217 1218 /** 1219 * do_filldir_main - read out directory entries 1220 * @dip: The GFS2 inode 1221 * @ctx: what to feed the entries to 1222 * @darr: an array of struct gfs2_dirent pointers to read 1223 * @entries: the number of entries in darr 1224 * @copied: pointer to int that's non-zero if a entry has been copied out 1225 * 1226 * Jump through some hoops to make sure that if there are hash collsions, 1227 * they are read out at the beginning of a buffer. We want to minimize 1228 * the possibility that they will fall into different readdir buffers or 1229 * that someone will want to seek to that location. 1230 * 1231 * Returns: errno, >0 if the actor tells you to stop 1232 */ 1233 1234 static int do_filldir_main(struct gfs2_inode *dip, struct dir_context *ctx, 1235 const struct gfs2_dirent **darr, u32 entries, 1236 int *copied) 1237 { 1238 const struct gfs2_dirent *dent, *dent_next; 1239 u64 off, off_next; 1240 unsigned int x, y; 1241 int run = 0; 1242 1243 sort(darr, entries, sizeof(struct gfs2_dirent *), compare_dents, NULL); 1244 1245 dent_next = darr[0]; 1246 off_next = be32_to_cpu(dent_next->de_hash); 1247 off_next = gfs2_disk_hash2offset(off_next); 1248 1249 for (x = 0, y = 1; x < entries; x++, y++) { 1250 dent = dent_next; 1251 off = off_next; 1252 1253 if (y < entries) { 1254 dent_next = darr[y]; 1255 off_next = be32_to_cpu(dent_next->de_hash); 1256 off_next = gfs2_disk_hash2offset(off_next); 1257 1258 if (off < ctx->pos) 1259 continue; 1260 ctx->pos = off; 1261 1262 if (off_next == off) { 1263 if (*copied && !run) 1264 return 1; 1265 run = 1; 1266 } else 1267 run = 0; 1268 } else { 1269 if (off < ctx->pos) 1270 continue; 1271 ctx->pos = off; 1272 } 1273 1274 if (!dir_emit(ctx, (const char *)(dent + 1), 1275 be16_to_cpu(dent->de_name_len), 1276 be64_to_cpu(dent->de_inum.no_addr), 1277 be16_to_cpu(dent->de_type))) 1278 return 1; 1279 1280 *copied = 1; 1281 } 1282 1283 /* Increment the ctx->pos by one, so the next time we come into the 1284 do_filldir fxn, we get the next entry instead of the last one in the 1285 current leaf */ 1286 1287 ctx->pos++; 1288 1289 return 0; 1290 } 1291 1292 static void *gfs2_alloc_sort_buffer(unsigned size) 1293 { 1294 void *ptr = NULL; 1295 1296 if (size < KMALLOC_MAX_SIZE) 1297 ptr = kmalloc(size, GFP_NOFS | __GFP_NOWARN); 1298 if (!ptr) 1299 ptr = __vmalloc(size, GFP_NOFS, PAGE_KERNEL); 1300 return ptr; 1301 } 1302 1303 static void gfs2_free_sort_buffer(void *ptr) 1304 { 1305 if (is_vmalloc_addr(ptr)) 1306 vfree(ptr); 1307 else 1308 kfree(ptr); 1309 } 1310 1311 static int gfs2_dir_read_leaf(struct inode *inode, struct dir_context *ctx, 1312 int *copied, unsigned *depth, 1313 u64 leaf_no) 1314 { 1315 struct gfs2_inode *ip = GFS2_I(inode); 1316 struct gfs2_sbd *sdp = GFS2_SB(inode); 1317 struct buffer_head *bh; 1318 struct gfs2_leaf *lf; 1319 unsigned entries = 0, entries2 = 0; 1320 unsigned leaves = 0; 1321 const struct gfs2_dirent **darr, *dent; 1322 struct dirent_gather g; 1323 struct buffer_head **larr; 1324 int leaf = 0; 1325 int error, i; 1326 u64 lfn = leaf_no; 1327 1328 do { 1329 error = get_leaf(ip, lfn, &bh); 1330 if (error) 1331 goto out; 1332 lf = (struct gfs2_leaf *)bh->b_data; 1333 if (leaves == 0) 1334 *depth = be16_to_cpu(lf->lf_depth); 1335 entries += be16_to_cpu(lf->lf_entries); 1336 leaves++; 1337 lfn = be64_to_cpu(lf->lf_next); 1338 brelse(bh); 1339 } while(lfn); 1340 1341 if (!entries) 1342 return 0; 1343 1344 error = -ENOMEM; 1345 /* 1346 * The extra 99 entries are not normally used, but are a buffer 1347 * zone in case the number of entries in the leaf is corrupt. 1348 * 99 is the maximum number of entries that can fit in a single 1349 * leaf block. 1350 */ 1351 larr = gfs2_alloc_sort_buffer((leaves + entries + 99) * sizeof(void *)); 1352 if (!larr) 1353 goto out; 1354 darr = (const struct gfs2_dirent **)(larr + leaves); 1355 g.pdent = darr; 1356 g.offset = 0; 1357 lfn = leaf_no; 1358 1359 do { 1360 error = get_leaf(ip, lfn, &bh); 1361 if (error) 1362 goto out_free; 1363 lf = (struct gfs2_leaf *)bh->b_data; 1364 lfn = be64_to_cpu(lf->lf_next); 1365 if (lf->lf_entries) { 1366 entries2 += be16_to_cpu(lf->lf_entries); 1367 dent = gfs2_dirent_scan(inode, bh->b_data, bh->b_size, 1368 gfs2_dirent_gather, NULL, &g); 1369 error = PTR_ERR(dent); 1370 if (IS_ERR(dent)) 1371 goto out_free; 1372 if (entries2 != g.offset) { 1373 fs_warn(sdp, "Number of entries corrupt in dir " 1374 "leaf %llu, entries2 (%u) != " 1375 "g.offset (%u)\n", 1376 (unsigned long long)bh->b_blocknr, 1377 entries2, g.offset); 1378 1379 error = -EIO; 1380 goto out_free; 1381 } 1382 error = 0; 1383 larr[leaf++] = bh; 1384 } else { 1385 brelse(bh); 1386 } 1387 } while(lfn); 1388 1389 BUG_ON(entries2 != entries); 1390 error = do_filldir_main(ip, ctx, darr, entries, copied); 1391 out_free: 1392 for(i = 0; i < leaf; i++) 1393 brelse(larr[i]); 1394 gfs2_free_sort_buffer(larr); 1395 out: 1396 return error; 1397 } 1398 1399 /** 1400 * gfs2_dir_readahead - Issue read-ahead requests for leaf blocks. 1401 * 1402 * Note: we can't calculate each index like dir_e_read can because we don't 1403 * have the leaf, and therefore we don't have the depth, and therefore we 1404 * don't have the length. So we have to just read enough ahead to make up 1405 * for the loss of information. 1406 */ 1407 static void gfs2_dir_readahead(struct inode *inode, unsigned hsize, u32 index, 1408 struct file_ra_state *f_ra) 1409 { 1410 struct gfs2_inode *ip = GFS2_I(inode); 1411 struct gfs2_glock *gl = ip->i_gl; 1412 struct buffer_head *bh; 1413 u64 blocknr = 0, last; 1414 unsigned count; 1415 1416 /* First check if we've already read-ahead for the whole range. */ 1417 if (index + MAX_RA_BLOCKS < f_ra->start) 1418 return; 1419 1420 f_ra->start = max((pgoff_t)index, f_ra->start); 1421 for (count = 0; count < MAX_RA_BLOCKS; count++) { 1422 if (f_ra->start >= hsize) /* if exceeded the hash table */ 1423 break; 1424 1425 last = blocknr; 1426 blocknr = be64_to_cpu(ip->i_hash_cache[f_ra->start]); 1427 f_ra->start++; 1428 if (blocknr == last) 1429 continue; 1430 1431 bh = gfs2_getbuf(gl, blocknr, 1); 1432 if (trylock_buffer(bh)) { 1433 if (buffer_uptodate(bh)) { 1434 unlock_buffer(bh); 1435 brelse(bh); 1436 continue; 1437 } 1438 bh->b_end_io = end_buffer_read_sync; 1439 submit_bh(READA | REQ_META, bh); 1440 continue; 1441 } 1442 brelse(bh); 1443 } 1444 } 1445 1446 /** 1447 * dir_e_read - Reads the entries from a directory into a filldir buffer 1448 * @dip: dinode pointer 1449 * @ctx: actor to feed the entries to 1450 * 1451 * Returns: errno 1452 */ 1453 1454 static int dir_e_read(struct inode *inode, struct dir_context *ctx, 1455 struct file_ra_state *f_ra) 1456 { 1457 struct gfs2_inode *dip = GFS2_I(inode); 1458 u32 hsize, len = 0; 1459 u32 hash, index; 1460 __be64 *lp; 1461 int copied = 0; 1462 int error = 0; 1463 unsigned depth = 0; 1464 1465 hsize = 1 << dip->i_depth; 1466 hash = gfs2_dir_offset2hash(ctx->pos); 1467 index = hash >> (32 - dip->i_depth); 1468 1469 if (dip->i_hash_cache == NULL) 1470 f_ra->start = 0; 1471 lp = gfs2_dir_get_hash_table(dip); 1472 if (IS_ERR(lp)) 1473 return PTR_ERR(lp); 1474 1475 gfs2_dir_readahead(inode, hsize, index, f_ra); 1476 1477 while (index < hsize) { 1478 error = gfs2_dir_read_leaf(inode, ctx, 1479 &copied, &depth, 1480 be64_to_cpu(lp[index])); 1481 if (error) 1482 break; 1483 1484 len = 1 << (dip->i_depth - depth); 1485 index = (index & ~(len - 1)) + len; 1486 } 1487 1488 if (error > 0) 1489 error = 0; 1490 return error; 1491 } 1492 1493 int gfs2_dir_read(struct inode *inode, struct dir_context *ctx, 1494 struct file_ra_state *f_ra) 1495 { 1496 struct gfs2_inode *dip = GFS2_I(inode); 1497 struct gfs2_sbd *sdp = GFS2_SB(inode); 1498 struct dirent_gather g; 1499 const struct gfs2_dirent **darr, *dent; 1500 struct buffer_head *dibh; 1501 int copied = 0; 1502 int error; 1503 1504 if (!dip->i_entries) 1505 return 0; 1506 1507 if (dip->i_diskflags & GFS2_DIF_EXHASH) 1508 return dir_e_read(inode, ctx, f_ra); 1509 1510 if (!gfs2_is_stuffed(dip)) { 1511 gfs2_consist_inode(dip); 1512 return -EIO; 1513 } 1514 1515 error = gfs2_meta_inode_buffer(dip, &dibh); 1516 if (error) 1517 return error; 1518 1519 error = -ENOMEM; 1520 /* 96 is max number of dirents which can be stuffed into an inode */ 1521 darr = kmalloc(96 * sizeof(struct gfs2_dirent *), GFP_NOFS); 1522 if (darr) { 1523 g.pdent = darr; 1524 g.offset = 0; 1525 dent = gfs2_dirent_scan(inode, dibh->b_data, dibh->b_size, 1526 gfs2_dirent_gather, NULL, &g); 1527 if (IS_ERR(dent)) { 1528 error = PTR_ERR(dent); 1529 goto out; 1530 } 1531 if (dip->i_entries != g.offset) { 1532 fs_warn(sdp, "Number of entries corrupt in dir %llu, " 1533 "ip->i_entries (%u) != g.offset (%u)\n", 1534 (unsigned long long)dip->i_no_addr, 1535 dip->i_entries, 1536 g.offset); 1537 error = -EIO; 1538 goto out; 1539 } 1540 error = do_filldir_main(dip, ctx, darr, 1541 dip->i_entries, &copied); 1542 out: 1543 kfree(darr); 1544 } 1545 1546 if (error > 0) 1547 error = 0; 1548 1549 brelse(dibh); 1550 1551 return error; 1552 } 1553 1554 /** 1555 * gfs2_dir_search - Search a directory 1556 * @dip: The GFS2 dir inode 1557 * @name: The name we are looking up 1558 * @fail_on_exist: Fail if the name exists rather than looking it up 1559 * 1560 * This routine searches a directory for a file or another directory. 1561 * Assumes a glock is held on dip. 1562 * 1563 * Returns: errno 1564 */ 1565 1566 struct inode *gfs2_dir_search(struct inode *dir, const struct qstr *name, 1567 bool fail_on_exist) 1568 { 1569 struct buffer_head *bh; 1570 struct gfs2_dirent *dent; 1571 u64 addr, formal_ino; 1572 u16 dtype; 1573 1574 dent = gfs2_dirent_search(dir, name, gfs2_dirent_find, &bh); 1575 if (dent) { 1576 if (IS_ERR(dent)) 1577 return ERR_CAST(dent); 1578 dtype = be16_to_cpu(dent->de_type); 1579 addr = be64_to_cpu(dent->de_inum.no_addr); 1580 formal_ino = be64_to_cpu(dent->de_inum.no_formal_ino); 1581 brelse(bh); 1582 if (fail_on_exist) 1583 return ERR_PTR(-EEXIST); 1584 return gfs2_inode_lookup(dir->i_sb, dtype, addr, formal_ino, 0); 1585 } 1586 return ERR_PTR(-ENOENT); 1587 } 1588 1589 int gfs2_dir_check(struct inode *dir, const struct qstr *name, 1590 const struct gfs2_inode *ip) 1591 { 1592 struct buffer_head *bh; 1593 struct gfs2_dirent *dent; 1594 int ret = -ENOENT; 1595 1596 dent = gfs2_dirent_search(dir, name, gfs2_dirent_find, &bh); 1597 if (dent) { 1598 if (IS_ERR(dent)) 1599 return PTR_ERR(dent); 1600 if (ip) { 1601 if (be64_to_cpu(dent->de_inum.no_addr) != ip->i_no_addr) 1602 goto out; 1603 if (be64_to_cpu(dent->de_inum.no_formal_ino) != 1604 ip->i_no_formal_ino) 1605 goto out; 1606 if (unlikely(IF2DT(ip->i_inode.i_mode) != 1607 be16_to_cpu(dent->de_type))) { 1608 gfs2_consist_inode(GFS2_I(dir)); 1609 ret = -EIO; 1610 goto out; 1611 } 1612 } 1613 ret = 0; 1614 out: 1615 brelse(bh); 1616 } 1617 return ret; 1618 } 1619 1620 /** 1621 * dir_new_leaf - Add a new leaf onto hash chain 1622 * @inode: The directory 1623 * @name: The name we are adding 1624 * 1625 * This adds a new dir leaf onto an existing leaf when there is not 1626 * enough space to add a new dir entry. This is a last resort after 1627 * we've expanded the hash table to max size and also split existing 1628 * leaf blocks, so it will only occur for very large directories. 1629 * 1630 * The dist parameter is set to 1 for leaf blocks directly attached 1631 * to the hash table, 2 for one layer of indirection, 3 for two layers 1632 * etc. We are thus able to tell the difference between an old leaf 1633 * with dist set to zero (i.e. "don't know") and a new one where we 1634 * set this information for debug/fsck purposes. 1635 * 1636 * Returns: 0 on success, or -ve on error 1637 */ 1638 1639 static int dir_new_leaf(struct inode *inode, const struct qstr *name) 1640 { 1641 struct buffer_head *bh, *obh; 1642 struct gfs2_inode *ip = GFS2_I(inode); 1643 struct gfs2_leaf *leaf, *oleaf; 1644 u32 dist = 1; 1645 int error; 1646 u32 index; 1647 u64 bn; 1648 1649 index = name->hash >> (32 - ip->i_depth); 1650 error = get_first_leaf(ip, index, &obh); 1651 if (error) 1652 return error; 1653 do { 1654 dist++; 1655 oleaf = (struct gfs2_leaf *)obh->b_data; 1656 bn = be64_to_cpu(oleaf->lf_next); 1657 if (!bn) 1658 break; 1659 brelse(obh); 1660 error = get_leaf(ip, bn, &obh); 1661 if (error) 1662 return error; 1663 } while(1); 1664 1665 gfs2_trans_add_meta(ip->i_gl, obh); 1666 1667 leaf = new_leaf(inode, &bh, be16_to_cpu(oleaf->lf_depth)); 1668 if (!leaf) { 1669 brelse(obh); 1670 return -ENOSPC; 1671 } 1672 leaf->lf_dist = cpu_to_be32(dist); 1673 oleaf->lf_next = cpu_to_be64(bh->b_blocknr); 1674 brelse(bh); 1675 brelse(obh); 1676 1677 error = gfs2_meta_inode_buffer(ip, &bh); 1678 if (error) 1679 return error; 1680 gfs2_trans_add_meta(ip->i_gl, bh); 1681 gfs2_add_inode_blocks(&ip->i_inode, 1); 1682 gfs2_dinode_out(ip, bh->b_data); 1683 brelse(bh); 1684 return 0; 1685 } 1686 1687 /** 1688 * gfs2_dir_add - Add new filename into directory 1689 * @inode: The directory inode 1690 * @name: The new name 1691 * @nip: The GFS2 inode to be linked in to the directory 1692 * @da: The directory addition info 1693 * 1694 * If the call to gfs2_diradd_alloc_required resulted in there being 1695 * no need to allocate any new directory blocks, then it will contain 1696 * a pointer to the directory entry and the bh in which it resides. We 1697 * can use that without having to repeat the search. If there was no 1698 * free space, then we must now create more space. 1699 * 1700 * Returns: 0 on success, error code on failure 1701 */ 1702 1703 int gfs2_dir_add(struct inode *inode, const struct qstr *name, 1704 const struct gfs2_inode *nip, struct gfs2_diradd *da) 1705 { 1706 struct gfs2_inode *ip = GFS2_I(inode); 1707 struct buffer_head *bh = da->bh; 1708 struct gfs2_dirent *dent = da->dent; 1709 struct timespec tv; 1710 struct gfs2_leaf *leaf; 1711 int error; 1712 1713 while(1) { 1714 if (da->bh == NULL) { 1715 dent = gfs2_dirent_search(inode, name, 1716 gfs2_dirent_find_space, &bh); 1717 } 1718 if (dent) { 1719 if (IS_ERR(dent)) 1720 return PTR_ERR(dent); 1721 dent = gfs2_init_dirent(inode, dent, name, bh); 1722 gfs2_inum_out(nip, dent); 1723 dent->de_type = cpu_to_be16(IF2DT(nip->i_inode.i_mode)); 1724 tv = CURRENT_TIME; 1725 if (ip->i_diskflags & GFS2_DIF_EXHASH) { 1726 leaf = (struct gfs2_leaf *)bh->b_data; 1727 be16_add_cpu(&leaf->lf_entries, 1); 1728 leaf->lf_nsec = cpu_to_be32(tv.tv_nsec); 1729 leaf->lf_sec = cpu_to_be64(tv.tv_sec); 1730 } 1731 da->dent = NULL; 1732 da->bh = NULL; 1733 brelse(bh); 1734 ip->i_entries++; 1735 ip->i_inode.i_mtime = ip->i_inode.i_ctime = tv; 1736 if (S_ISDIR(nip->i_inode.i_mode)) 1737 inc_nlink(&ip->i_inode); 1738 mark_inode_dirty(inode); 1739 error = 0; 1740 break; 1741 } 1742 if (!(ip->i_diskflags & GFS2_DIF_EXHASH)) { 1743 error = dir_make_exhash(inode); 1744 if (error) 1745 break; 1746 continue; 1747 } 1748 error = dir_split_leaf(inode, name); 1749 if (error == 0) 1750 continue; 1751 if (error < 0) 1752 break; 1753 if (ip->i_depth < GFS2_DIR_MAX_DEPTH) { 1754 error = dir_double_exhash(ip); 1755 if (error) 1756 break; 1757 error = dir_split_leaf(inode, name); 1758 if (error < 0) 1759 break; 1760 if (error == 0) 1761 continue; 1762 } 1763 error = dir_new_leaf(inode, name); 1764 if (!error) 1765 continue; 1766 error = -ENOSPC; 1767 break; 1768 } 1769 return error; 1770 } 1771 1772 1773 /** 1774 * gfs2_dir_del - Delete a directory entry 1775 * @dip: The GFS2 inode 1776 * @filename: The filename 1777 * 1778 * Returns: 0 on success, error code on failure 1779 */ 1780 1781 int gfs2_dir_del(struct gfs2_inode *dip, const struct dentry *dentry) 1782 { 1783 const struct qstr *name = &dentry->d_name; 1784 struct gfs2_dirent *dent, *prev = NULL; 1785 struct buffer_head *bh; 1786 struct timespec tv = CURRENT_TIME; 1787 1788 /* Returns _either_ the entry (if its first in block) or the 1789 previous entry otherwise */ 1790 dent = gfs2_dirent_search(&dip->i_inode, name, gfs2_dirent_prev, &bh); 1791 if (!dent) { 1792 gfs2_consist_inode(dip); 1793 return -EIO; 1794 } 1795 if (IS_ERR(dent)) { 1796 gfs2_consist_inode(dip); 1797 return PTR_ERR(dent); 1798 } 1799 /* If not first in block, adjust pointers accordingly */ 1800 if (gfs2_dirent_find(dent, name, NULL) == 0) { 1801 prev = dent; 1802 dent = (struct gfs2_dirent *)((char *)dent + be16_to_cpu(prev->de_rec_len)); 1803 } 1804 1805 dirent_del(dip, bh, prev, dent); 1806 if (dip->i_diskflags & GFS2_DIF_EXHASH) { 1807 struct gfs2_leaf *leaf = (struct gfs2_leaf *)bh->b_data; 1808 u16 entries = be16_to_cpu(leaf->lf_entries); 1809 if (!entries) 1810 gfs2_consist_inode(dip); 1811 leaf->lf_entries = cpu_to_be16(--entries); 1812 leaf->lf_nsec = cpu_to_be32(tv.tv_nsec); 1813 leaf->lf_sec = cpu_to_be64(tv.tv_sec); 1814 } 1815 brelse(bh); 1816 1817 if (!dip->i_entries) 1818 gfs2_consist_inode(dip); 1819 dip->i_entries--; 1820 dip->i_inode.i_mtime = dip->i_inode.i_ctime = tv; 1821 if (S_ISDIR(dentry->d_inode->i_mode)) 1822 drop_nlink(&dip->i_inode); 1823 mark_inode_dirty(&dip->i_inode); 1824 1825 return 0; 1826 } 1827 1828 /** 1829 * gfs2_dir_mvino - Change inode number of directory entry 1830 * @dip: The GFS2 inode 1831 * @filename: 1832 * @new_inode: 1833 * 1834 * This routine changes the inode number of a directory entry. It's used 1835 * by rename to change ".." when a directory is moved. 1836 * Assumes a glock is held on dvp. 1837 * 1838 * Returns: errno 1839 */ 1840 1841 int gfs2_dir_mvino(struct gfs2_inode *dip, const struct qstr *filename, 1842 const struct gfs2_inode *nip, unsigned int new_type) 1843 { 1844 struct buffer_head *bh; 1845 struct gfs2_dirent *dent; 1846 int error; 1847 1848 dent = gfs2_dirent_search(&dip->i_inode, filename, gfs2_dirent_find, &bh); 1849 if (!dent) { 1850 gfs2_consist_inode(dip); 1851 return -EIO; 1852 } 1853 if (IS_ERR(dent)) 1854 return PTR_ERR(dent); 1855 1856 gfs2_trans_add_meta(dip->i_gl, bh); 1857 gfs2_inum_out(nip, dent); 1858 dent->de_type = cpu_to_be16(new_type); 1859 1860 if (dip->i_diskflags & GFS2_DIF_EXHASH) { 1861 brelse(bh); 1862 error = gfs2_meta_inode_buffer(dip, &bh); 1863 if (error) 1864 return error; 1865 gfs2_trans_add_meta(dip->i_gl, bh); 1866 } 1867 1868 dip->i_inode.i_mtime = dip->i_inode.i_ctime = CURRENT_TIME; 1869 gfs2_dinode_out(dip, bh->b_data); 1870 brelse(bh); 1871 return 0; 1872 } 1873 1874 /** 1875 * leaf_dealloc - Deallocate a directory leaf 1876 * @dip: the directory 1877 * @index: the hash table offset in the directory 1878 * @len: the number of pointers to this leaf 1879 * @leaf_no: the leaf number 1880 * @leaf_bh: buffer_head for the starting leaf 1881 * last_dealloc: 1 if this is the final dealloc for the leaf, else 0 1882 * 1883 * Returns: errno 1884 */ 1885 1886 static int leaf_dealloc(struct gfs2_inode *dip, u32 index, u32 len, 1887 u64 leaf_no, struct buffer_head *leaf_bh, 1888 int last_dealloc) 1889 { 1890 struct gfs2_sbd *sdp = GFS2_SB(&dip->i_inode); 1891 struct gfs2_leaf *tmp_leaf; 1892 struct gfs2_rgrp_list rlist; 1893 struct buffer_head *bh, *dibh; 1894 u64 blk, nblk; 1895 unsigned int rg_blocks = 0, l_blocks = 0; 1896 char *ht; 1897 unsigned int x, size = len * sizeof(u64); 1898 int error; 1899 1900 error = gfs2_rindex_update(sdp); 1901 if (error) 1902 return error; 1903 1904 memset(&rlist, 0, sizeof(struct gfs2_rgrp_list)); 1905 1906 ht = kzalloc(size, GFP_NOFS | __GFP_NOWARN); 1907 if (ht == NULL) 1908 ht = vzalloc(size); 1909 if (!ht) 1910 return -ENOMEM; 1911 1912 error = gfs2_quota_hold(dip, NO_UID_QUOTA_CHANGE, NO_GID_QUOTA_CHANGE); 1913 if (error) 1914 goto out; 1915 1916 /* Count the number of leaves */ 1917 bh = leaf_bh; 1918 1919 for (blk = leaf_no; blk; blk = nblk) { 1920 if (blk != leaf_no) { 1921 error = get_leaf(dip, blk, &bh); 1922 if (error) 1923 goto out_rlist; 1924 } 1925 tmp_leaf = (struct gfs2_leaf *)bh->b_data; 1926 nblk = be64_to_cpu(tmp_leaf->lf_next); 1927 if (blk != leaf_no) 1928 brelse(bh); 1929 1930 gfs2_rlist_add(dip, &rlist, blk); 1931 l_blocks++; 1932 } 1933 1934 gfs2_rlist_alloc(&rlist, LM_ST_EXCLUSIVE); 1935 1936 for (x = 0; x < rlist.rl_rgrps; x++) { 1937 struct gfs2_rgrpd *rgd; 1938 rgd = rlist.rl_ghs[x].gh_gl->gl_object; 1939 rg_blocks += rgd->rd_length; 1940 } 1941 1942 error = gfs2_glock_nq_m(rlist.rl_rgrps, rlist.rl_ghs); 1943 if (error) 1944 goto out_rlist; 1945 1946 error = gfs2_trans_begin(sdp, 1947 rg_blocks + (DIV_ROUND_UP(size, sdp->sd_jbsize) + 1) + 1948 RES_DINODE + RES_STATFS + RES_QUOTA, l_blocks); 1949 if (error) 1950 goto out_rg_gunlock; 1951 1952 bh = leaf_bh; 1953 1954 for (blk = leaf_no; blk; blk = nblk) { 1955 if (blk != leaf_no) { 1956 error = get_leaf(dip, blk, &bh); 1957 if (error) 1958 goto out_end_trans; 1959 } 1960 tmp_leaf = (struct gfs2_leaf *)bh->b_data; 1961 nblk = be64_to_cpu(tmp_leaf->lf_next); 1962 if (blk != leaf_no) 1963 brelse(bh); 1964 1965 gfs2_free_meta(dip, blk, 1); 1966 gfs2_add_inode_blocks(&dip->i_inode, -1); 1967 } 1968 1969 error = gfs2_dir_write_data(dip, ht, index * sizeof(u64), size); 1970 if (error != size) { 1971 if (error >= 0) 1972 error = -EIO; 1973 goto out_end_trans; 1974 } 1975 1976 error = gfs2_meta_inode_buffer(dip, &dibh); 1977 if (error) 1978 goto out_end_trans; 1979 1980 gfs2_trans_add_meta(dip->i_gl, dibh); 1981 /* On the last dealloc, make this a regular file in case we crash. 1982 (We don't want to free these blocks a second time.) */ 1983 if (last_dealloc) 1984 dip->i_inode.i_mode = S_IFREG; 1985 gfs2_dinode_out(dip, dibh->b_data); 1986 brelse(dibh); 1987 1988 out_end_trans: 1989 gfs2_trans_end(sdp); 1990 out_rg_gunlock: 1991 gfs2_glock_dq_m(rlist.rl_rgrps, rlist.rl_ghs); 1992 out_rlist: 1993 gfs2_rlist_free(&rlist); 1994 gfs2_quota_unhold(dip); 1995 out: 1996 if (is_vmalloc_addr(ht)) 1997 vfree(ht); 1998 else 1999 kfree(ht); 2000 return error; 2001 } 2002 2003 /** 2004 * gfs2_dir_exhash_dealloc - free all the leaf blocks in a directory 2005 * @dip: the directory 2006 * 2007 * Dealloc all on-disk directory leaves to FREEMETA state 2008 * Change on-disk inode type to "regular file" 2009 * 2010 * Returns: errno 2011 */ 2012 2013 int gfs2_dir_exhash_dealloc(struct gfs2_inode *dip) 2014 { 2015 struct buffer_head *bh; 2016 struct gfs2_leaf *leaf; 2017 u32 hsize, len; 2018 u32 index = 0, next_index; 2019 __be64 *lp; 2020 u64 leaf_no; 2021 int error = 0, last; 2022 2023 hsize = 1 << dip->i_depth; 2024 2025 lp = gfs2_dir_get_hash_table(dip); 2026 if (IS_ERR(lp)) 2027 return PTR_ERR(lp); 2028 2029 while (index < hsize) { 2030 leaf_no = be64_to_cpu(lp[index]); 2031 if (leaf_no) { 2032 error = get_leaf(dip, leaf_no, &bh); 2033 if (error) 2034 goto out; 2035 leaf = (struct gfs2_leaf *)bh->b_data; 2036 len = 1 << (dip->i_depth - be16_to_cpu(leaf->lf_depth)); 2037 2038 next_index = (index & ~(len - 1)) + len; 2039 last = ((next_index >= hsize) ? 1 : 0); 2040 error = leaf_dealloc(dip, index, len, leaf_no, bh, 2041 last); 2042 brelse(bh); 2043 if (error) 2044 goto out; 2045 index = next_index; 2046 } else 2047 index++; 2048 } 2049 2050 if (index != hsize) { 2051 gfs2_consist_inode(dip); 2052 error = -EIO; 2053 } 2054 2055 out: 2056 2057 return error; 2058 } 2059 2060 /** 2061 * gfs2_diradd_alloc_required - find if adding entry will require an allocation 2062 * @ip: the file being written to 2063 * @filname: the filename that's going to be added 2064 * @da: The structure to return dir alloc info 2065 * 2066 * Returns: 0 if ok, -ve on error 2067 */ 2068 2069 int gfs2_diradd_alloc_required(struct inode *inode, const struct qstr *name, 2070 struct gfs2_diradd *da) 2071 { 2072 struct gfs2_inode *ip = GFS2_I(inode); 2073 struct gfs2_sbd *sdp = GFS2_SB(inode); 2074 const unsigned int extra = sizeof(struct gfs2_dinode) - sizeof(struct gfs2_leaf); 2075 struct gfs2_dirent *dent; 2076 struct buffer_head *bh; 2077 2078 da->nr_blocks = 0; 2079 da->bh = NULL; 2080 da->dent = NULL; 2081 2082 dent = gfs2_dirent_search(inode, name, gfs2_dirent_find_space, &bh); 2083 if (!dent) { 2084 da->nr_blocks = sdp->sd_max_dirres; 2085 if (!(ip->i_diskflags & GFS2_DIF_EXHASH) && 2086 (GFS2_DIRENT_SIZE(name->len) < extra)) 2087 da->nr_blocks = 1; 2088 return 0; 2089 } 2090 if (IS_ERR(dent)) 2091 return PTR_ERR(dent); 2092 da->bh = bh; 2093 da->dent = dent; 2094 return 0; 2095 } 2096 2097