xref: /openbmc/linux/fs/gfs2/dir.c (revision 8730046c)
1 /*
2  * Copyright (C) Sistina Software, Inc.  1997-2003 All rights reserved.
3  * Copyright (C) 2004-2006 Red Hat, Inc.  All rights reserved.
4  *
5  * This copyrighted material is made available to anyone wishing to use,
6  * modify, copy, or redistribute it subject to the terms and conditions
7  * of the GNU General Public License version 2.
8  */
9 
10 /*
11  * Implements Extendible Hashing as described in:
12  *   "Extendible Hashing" by Fagin, et al in
13  *     __ACM Trans. on Database Systems__, Sept 1979.
14  *
15  *
16  * Here's the layout of dirents which is essentially the same as that of ext2
17  * within a single block. The field de_name_len is the number of bytes
18  * actually required for the name (no null terminator). The field de_rec_len
19  * is the number of bytes allocated to the dirent. The offset of the next
20  * dirent in the block is (dirent + dirent->de_rec_len). When a dirent is
21  * deleted, the preceding dirent inherits its allocated space, ie
22  * prev->de_rec_len += deleted->de_rec_len. Since the next dirent is obtained
23  * by adding de_rec_len to the current dirent, this essentially causes the
24  * deleted dirent to get jumped over when iterating through all the dirents.
25  *
26  * When deleting the first dirent in a block, there is no previous dirent so
27  * the field de_ino is set to zero to designate it as deleted. When allocating
28  * a dirent, gfs2_dirent_alloc iterates through the dirents in a block. If the
29  * first dirent has (de_ino == 0) and de_rec_len is large enough, this first
30  * dirent is allocated. Otherwise it must go through all the 'used' dirents
31  * searching for one in which the amount of total space minus the amount of
32  * used space will provide enough space for the new dirent.
33  *
34  * There are two types of blocks in which dirents reside. In a stuffed dinode,
35  * the dirents begin at offset sizeof(struct gfs2_dinode) from the beginning of
36  * the block.  In leaves, they begin at offset sizeof(struct gfs2_leaf) from the
37  * beginning of the leaf block. The dirents reside in leaves when
38  *
39  * dip->i_diskflags & GFS2_DIF_EXHASH is true
40  *
41  * Otherwise, the dirents are "linear", within a single stuffed dinode block.
42  *
43  * When the dirents are in leaves, the actual contents of the directory file are
44  * used as an array of 64-bit block pointers pointing to the leaf blocks. The
45  * dirents are NOT in the directory file itself. There can be more than one
46  * block pointer in the array that points to the same leaf. In fact, when a
47  * directory is first converted from linear to exhash, all of the pointers
48  * point to the same leaf.
49  *
50  * When a leaf is completely full, the size of the hash table can be
51  * doubled unless it is already at the maximum size which is hard coded into
52  * GFS2_DIR_MAX_DEPTH. After that, leaves are chained together in a linked list,
53  * but never before the maximum hash table size has been reached.
54  */
55 
56 #define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
57 
58 #include <linux/slab.h>
59 #include <linux/spinlock.h>
60 #include <linux/buffer_head.h>
61 #include <linux/sort.h>
62 #include <linux/gfs2_ondisk.h>
63 #include <linux/crc32.h>
64 #include <linux/vmalloc.h>
65 #include <linux/bio.h>
66 
67 #include "gfs2.h"
68 #include "incore.h"
69 #include "dir.h"
70 #include "glock.h"
71 #include "inode.h"
72 #include "meta_io.h"
73 #include "quota.h"
74 #include "rgrp.h"
75 #include "trans.h"
76 #include "bmap.h"
77 #include "util.h"
78 
79 #define IS_LEAF     1 /* Hashed (leaf) directory */
80 #define IS_DINODE   2 /* Linear (stuffed dinode block) directory */
81 
82 #define MAX_RA_BLOCKS 32 /* max read-ahead blocks */
83 
84 #define gfs2_disk_hash2offset(h) (((u64)(h)) >> 1)
85 #define gfs2_dir_offset2hash(p) ((u32)(((u64)(p)) << 1))
86 #define GFS2_HASH_INDEX_MASK 0xffffc000
87 #define GFS2_USE_HASH_FLAG 0x2000
88 
89 struct qstr gfs2_qdot __read_mostly;
90 struct qstr gfs2_qdotdot __read_mostly;
91 
92 typedef int (*gfs2_dscan_t)(const struct gfs2_dirent *dent,
93 			    const struct qstr *name, void *opaque);
94 
95 int gfs2_dir_get_new_buffer(struct gfs2_inode *ip, u64 block,
96 			    struct buffer_head **bhp)
97 {
98 	struct buffer_head *bh;
99 
100 	bh = gfs2_meta_new(ip->i_gl, block);
101 	gfs2_trans_add_meta(ip->i_gl, bh);
102 	gfs2_metatype_set(bh, GFS2_METATYPE_JD, GFS2_FORMAT_JD);
103 	gfs2_buffer_clear_tail(bh, sizeof(struct gfs2_meta_header));
104 	*bhp = bh;
105 	return 0;
106 }
107 
108 static int gfs2_dir_get_existing_buffer(struct gfs2_inode *ip, u64 block,
109 					struct buffer_head **bhp)
110 {
111 	struct buffer_head *bh;
112 	int error;
113 
114 	error = gfs2_meta_read(ip->i_gl, block, DIO_WAIT, 0, &bh);
115 	if (error)
116 		return error;
117 	if (gfs2_metatype_check(GFS2_SB(&ip->i_inode), bh, GFS2_METATYPE_JD)) {
118 		brelse(bh);
119 		return -EIO;
120 	}
121 	*bhp = bh;
122 	return 0;
123 }
124 
125 static int gfs2_dir_write_stuffed(struct gfs2_inode *ip, const char *buf,
126 				  unsigned int offset, unsigned int size)
127 {
128 	struct buffer_head *dibh;
129 	int error;
130 
131 	error = gfs2_meta_inode_buffer(ip, &dibh);
132 	if (error)
133 		return error;
134 
135 	gfs2_trans_add_meta(ip->i_gl, dibh);
136 	memcpy(dibh->b_data + offset + sizeof(struct gfs2_dinode), buf, size);
137 	if (ip->i_inode.i_size < offset + size)
138 		i_size_write(&ip->i_inode, offset + size);
139 	ip->i_inode.i_mtime = ip->i_inode.i_ctime = current_time(&ip->i_inode);
140 	gfs2_dinode_out(ip, dibh->b_data);
141 
142 	brelse(dibh);
143 
144 	return size;
145 }
146 
147 
148 
149 /**
150  * gfs2_dir_write_data - Write directory information to the inode
151  * @ip: The GFS2 inode
152  * @buf: The buffer containing information to be written
153  * @offset: The file offset to start writing at
154  * @size: The amount of data to write
155  *
156  * Returns: The number of bytes correctly written or error code
157  */
158 static int gfs2_dir_write_data(struct gfs2_inode *ip, const char *buf,
159 			       u64 offset, unsigned int size)
160 {
161 	struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode);
162 	struct buffer_head *dibh;
163 	u64 lblock, dblock;
164 	u32 extlen = 0;
165 	unsigned int o;
166 	int copied = 0;
167 	int error = 0;
168 	int new = 0;
169 
170 	if (!size)
171 		return 0;
172 
173 	if (gfs2_is_stuffed(ip) &&
174 	    offset + size <= sdp->sd_sb.sb_bsize - sizeof(struct gfs2_dinode))
175 		return gfs2_dir_write_stuffed(ip, buf, (unsigned int)offset,
176 					      size);
177 
178 	if (gfs2_assert_warn(sdp, gfs2_is_jdata(ip)))
179 		return -EINVAL;
180 
181 	if (gfs2_is_stuffed(ip)) {
182 		error = gfs2_unstuff_dinode(ip, NULL);
183 		if (error)
184 			return error;
185 	}
186 
187 	lblock = offset;
188 	o = do_div(lblock, sdp->sd_jbsize) + sizeof(struct gfs2_meta_header);
189 
190 	while (copied < size) {
191 		unsigned int amount;
192 		struct buffer_head *bh;
193 
194 		amount = size - copied;
195 		if (amount > sdp->sd_sb.sb_bsize - o)
196 			amount = sdp->sd_sb.sb_bsize - o;
197 
198 		if (!extlen) {
199 			new = 1;
200 			error = gfs2_extent_map(&ip->i_inode, lblock, &new,
201 						&dblock, &extlen);
202 			if (error)
203 				goto fail;
204 			error = -EIO;
205 			if (gfs2_assert_withdraw(sdp, dblock))
206 				goto fail;
207 		}
208 
209 		if (amount == sdp->sd_jbsize || new)
210 			error = gfs2_dir_get_new_buffer(ip, dblock, &bh);
211 		else
212 			error = gfs2_dir_get_existing_buffer(ip, dblock, &bh);
213 
214 		if (error)
215 			goto fail;
216 
217 		gfs2_trans_add_meta(ip->i_gl, bh);
218 		memcpy(bh->b_data + o, buf, amount);
219 		brelse(bh);
220 
221 		buf += amount;
222 		copied += amount;
223 		lblock++;
224 		dblock++;
225 		extlen--;
226 
227 		o = sizeof(struct gfs2_meta_header);
228 	}
229 
230 out:
231 	error = gfs2_meta_inode_buffer(ip, &dibh);
232 	if (error)
233 		return error;
234 
235 	if (ip->i_inode.i_size < offset + copied)
236 		i_size_write(&ip->i_inode, offset + copied);
237 	ip->i_inode.i_mtime = ip->i_inode.i_ctime = current_time(&ip->i_inode);
238 
239 	gfs2_trans_add_meta(ip->i_gl, dibh);
240 	gfs2_dinode_out(ip, dibh->b_data);
241 	brelse(dibh);
242 
243 	return copied;
244 fail:
245 	if (copied)
246 		goto out;
247 	return error;
248 }
249 
250 static int gfs2_dir_read_stuffed(struct gfs2_inode *ip, __be64 *buf,
251 				 unsigned int size)
252 {
253 	struct buffer_head *dibh;
254 	int error;
255 
256 	error = gfs2_meta_inode_buffer(ip, &dibh);
257 	if (!error) {
258 		memcpy(buf, dibh->b_data + sizeof(struct gfs2_dinode), size);
259 		brelse(dibh);
260 	}
261 
262 	return (error) ? error : size;
263 }
264 
265 
266 /**
267  * gfs2_dir_read_data - Read a data from a directory inode
268  * @ip: The GFS2 Inode
269  * @buf: The buffer to place result into
270  * @size: Amount of data to transfer
271  *
272  * Returns: The amount of data actually copied or the error
273  */
274 static int gfs2_dir_read_data(struct gfs2_inode *ip, __be64 *buf,
275 			      unsigned int size)
276 {
277 	struct gfs2_sbd *sdp = GFS2_SB(&ip->i_inode);
278 	u64 lblock, dblock;
279 	u32 extlen = 0;
280 	unsigned int o;
281 	int copied = 0;
282 	int error = 0;
283 
284 	if (gfs2_is_stuffed(ip))
285 		return gfs2_dir_read_stuffed(ip, buf, size);
286 
287 	if (gfs2_assert_warn(sdp, gfs2_is_jdata(ip)))
288 		return -EINVAL;
289 
290 	lblock = 0;
291 	o = do_div(lblock, sdp->sd_jbsize) + sizeof(struct gfs2_meta_header);
292 
293 	while (copied < size) {
294 		unsigned int amount;
295 		struct buffer_head *bh;
296 		int new;
297 
298 		amount = size - copied;
299 		if (amount > sdp->sd_sb.sb_bsize - o)
300 			amount = sdp->sd_sb.sb_bsize - o;
301 
302 		if (!extlen) {
303 			new = 0;
304 			error = gfs2_extent_map(&ip->i_inode, lblock, &new,
305 						&dblock, &extlen);
306 			if (error || !dblock)
307 				goto fail;
308 			BUG_ON(extlen < 1);
309 			bh = gfs2_meta_ra(ip->i_gl, dblock, extlen);
310 		} else {
311 			error = gfs2_meta_read(ip->i_gl, dblock, DIO_WAIT, 0, &bh);
312 			if (error)
313 				goto fail;
314 		}
315 		error = gfs2_metatype_check(sdp, bh, GFS2_METATYPE_JD);
316 		if (error) {
317 			brelse(bh);
318 			goto fail;
319 		}
320 		dblock++;
321 		extlen--;
322 		memcpy(buf, bh->b_data + o, amount);
323 		brelse(bh);
324 		buf += (amount/sizeof(__be64));
325 		copied += amount;
326 		lblock++;
327 		o = sizeof(struct gfs2_meta_header);
328 	}
329 
330 	return copied;
331 fail:
332 	return (copied) ? copied : error;
333 }
334 
335 /**
336  * gfs2_dir_get_hash_table - Get pointer to the dir hash table
337  * @ip: The inode in question
338  *
339  * Returns: The hash table or an error
340  */
341 
342 static __be64 *gfs2_dir_get_hash_table(struct gfs2_inode *ip)
343 {
344 	struct inode *inode = &ip->i_inode;
345 	int ret;
346 	u32 hsize;
347 	__be64 *hc;
348 
349 	BUG_ON(!(ip->i_diskflags & GFS2_DIF_EXHASH));
350 
351 	hc = ip->i_hash_cache;
352 	if (hc)
353 		return hc;
354 
355 	hsize = BIT(ip->i_depth);
356 	hsize *= sizeof(__be64);
357 	if (hsize != i_size_read(&ip->i_inode)) {
358 		gfs2_consist_inode(ip);
359 		return ERR_PTR(-EIO);
360 	}
361 
362 	hc = kmalloc(hsize, GFP_NOFS | __GFP_NOWARN);
363 	if (hc == NULL)
364 		hc = __vmalloc(hsize, GFP_NOFS, PAGE_KERNEL);
365 
366 	if (hc == NULL)
367 		return ERR_PTR(-ENOMEM);
368 
369 	ret = gfs2_dir_read_data(ip, hc, hsize);
370 	if (ret < 0) {
371 		kvfree(hc);
372 		return ERR_PTR(ret);
373 	}
374 
375 	spin_lock(&inode->i_lock);
376 	if (likely(!ip->i_hash_cache)) {
377 		ip->i_hash_cache = hc;
378 		hc = NULL;
379 	}
380 	spin_unlock(&inode->i_lock);
381 	kvfree(hc);
382 
383 	return ip->i_hash_cache;
384 }
385 
386 /**
387  * gfs2_dir_hash_inval - Invalidate dir hash
388  * @ip: The directory inode
389  *
390  * Must be called with an exclusive glock, or during glock invalidation.
391  */
392 void gfs2_dir_hash_inval(struct gfs2_inode *ip)
393 {
394 	__be64 *hc;
395 
396 	spin_lock(&ip->i_inode.i_lock);
397 	hc = ip->i_hash_cache;
398 	ip->i_hash_cache = NULL;
399 	spin_unlock(&ip->i_inode.i_lock);
400 
401 	kvfree(hc);
402 }
403 
404 static inline int gfs2_dirent_sentinel(const struct gfs2_dirent *dent)
405 {
406 	return dent->de_inum.no_addr == 0 || dent->de_inum.no_formal_ino == 0;
407 }
408 
409 static inline int __gfs2_dirent_find(const struct gfs2_dirent *dent,
410 				     const struct qstr *name, int ret)
411 {
412 	if (!gfs2_dirent_sentinel(dent) &&
413 	    be32_to_cpu(dent->de_hash) == name->hash &&
414 	    be16_to_cpu(dent->de_name_len) == name->len &&
415 	    memcmp(dent+1, name->name, name->len) == 0)
416 		return ret;
417 	return 0;
418 }
419 
420 static int gfs2_dirent_find(const struct gfs2_dirent *dent,
421 			    const struct qstr *name,
422 			    void *opaque)
423 {
424 	return __gfs2_dirent_find(dent, name, 1);
425 }
426 
427 static int gfs2_dirent_prev(const struct gfs2_dirent *dent,
428 			    const struct qstr *name,
429 			    void *opaque)
430 {
431 	return __gfs2_dirent_find(dent, name, 2);
432 }
433 
434 /*
435  * name->name holds ptr to start of block.
436  * name->len holds size of block.
437  */
438 static int gfs2_dirent_last(const struct gfs2_dirent *dent,
439 			    const struct qstr *name,
440 			    void *opaque)
441 {
442 	const char *start = name->name;
443 	const char *end = (const char *)dent + be16_to_cpu(dent->de_rec_len);
444 	if (name->len == (end - start))
445 		return 1;
446 	return 0;
447 }
448 
449 /* Look for the dirent that contains the offset specified in data. Once we
450  * find that dirent, there must be space available there for the new dirent */
451 static int gfs2_dirent_find_offset(const struct gfs2_dirent *dent,
452 				  const struct qstr *name,
453 				  void *ptr)
454 {
455 	unsigned required = GFS2_DIRENT_SIZE(name->len);
456 	unsigned actual = GFS2_DIRENT_SIZE(be16_to_cpu(dent->de_name_len));
457 	unsigned totlen = be16_to_cpu(dent->de_rec_len);
458 
459 	if (ptr < (void *)dent || ptr >= (void *)dent + totlen)
460 		return 0;
461 	if (gfs2_dirent_sentinel(dent))
462 		actual = 0;
463 	if (ptr < (void *)dent + actual)
464 		return -1;
465 	if ((void *)dent + totlen >= ptr + required)
466 		return 1;
467 	return -1;
468 }
469 
470 static int gfs2_dirent_find_space(const struct gfs2_dirent *dent,
471 				  const struct qstr *name,
472 				  void *opaque)
473 {
474 	unsigned required = GFS2_DIRENT_SIZE(name->len);
475 	unsigned actual = GFS2_DIRENT_SIZE(be16_to_cpu(dent->de_name_len));
476 	unsigned totlen = be16_to_cpu(dent->de_rec_len);
477 
478 	if (gfs2_dirent_sentinel(dent))
479 		actual = 0;
480 	if (totlen - actual >= required)
481 		return 1;
482 	return 0;
483 }
484 
485 struct dirent_gather {
486 	const struct gfs2_dirent **pdent;
487 	unsigned offset;
488 };
489 
490 static int gfs2_dirent_gather(const struct gfs2_dirent *dent,
491 			      const struct qstr *name,
492 			      void *opaque)
493 {
494 	struct dirent_gather *g = opaque;
495 	if (!gfs2_dirent_sentinel(dent)) {
496 		g->pdent[g->offset++] = dent;
497 	}
498 	return 0;
499 }
500 
501 /*
502  * Other possible things to check:
503  * - Inode located within filesystem size (and on valid block)
504  * - Valid directory entry type
505  * Not sure how heavy-weight we want to make this... could also check
506  * hash is correct for example, but that would take a lot of extra time.
507  * For now the most important thing is to check that the various sizes
508  * are correct.
509  */
510 static int gfs2_check_dirent(struct gfs2_dirent *dent, unsigned int offset,
511 			     unsigned int size, unsigned int len, int first)
512 {
513 	const char *msg = "gfs2_dirent too small";
514 	if (unlikely(size < sizeof(struct gfs2_dirent)))
515 		goto error;
516 	msg = "gfs2_dirent misaligned";
517 	if (unlikely(offset & 0x7))
518 		goto error;
519 	msg = "gfs2_dirent points beyond end of block";
520 	if (unlikely(offset + size > len))
521 		goto error;
522 	msg = "zero inode number";
523 	if (unlikely(!first && gfs2_dirent_sentinel(dent)))
524 		goto error;
525 	msg = "name length is greater than space in dirent";
526 	if (!gfs2_dirent_sentinel(dent) &&
527 	    unlikely(sizeof(struct gfs2_dirent)+be16_to_cpu(dent->de_name_len) >
528 		     size))
529 		goto error;
530 	return 0;
531 error:
532 	pr_warn("%s: %s (%s)\n",
533 		__func__, msg, first ? "first in block" : "not first in block");
534 	return -EIO;
535 }
536 
537 static int gfs2_dirent_offset(const void *buf)
538 {
539 	const struct gfs2_meta_header *h = buf;
540 	int offset;
541 
542 	BUG_ON(buf == NULL);
543 
544 	switch(be32_to_cpu(h->mh_type)) {
545 	case GFS2_METATYPE_LF:
546 		offset = sizeof(struct gfs2_leaf);
547 		break;
548 	case GFS2_METATYPE_DI:
549 		offset = sizeof(struct gfs2_dinode);
550 		break;
551 	default:
552 		goto wrong_type;
553 	}
554 	return offset;
555 wrong_type:
556 	pr_warn("%s: wrong block type %u\n", __func__, be32_to_cpu(h->mh_type));
557 	return -1;
558 }
559 
560 static struct gfs2_dirent *gfs2_dirent_scan(struct inode *inode, void *buf,
561 					    unsigned int len, gfs2_dscan_t scan,
562 					    const struct qstr *name,
563 					    void *opaque)
564 {
565 	struct gfs2_dirent *dent, *prev;
566 	unsigned offset;
567 	unsigned size;
568 	int ret = 0;
569 
570 	ret = gfs2_dirent_offset(buf);
571 	if (ret < 0)
572 		goto consist_inode;
573 
574 	offset = ret;
575 	prev = NULL;
576 	dent = buf + offset;
577 	size = be16_to_cpu(dent->de_rec_len);
578 	if (gfs2_check_dirent(dent, offset, size, len, 1))
579 		goto consist_inode;
580 	do {
581 		ret = scan(dent, name, opaque);
582 		if (ret)
583 			break;
584 		offset += size;
585 		if (offset == len)
586 			break;
587 		prev = dent;
588 		dent = buf + offset;
589 		size = be16_to_cpu(dent->de_rec_len);
590 		if (gfs2_check_dirent(dent, offset, size, len, 0))
591 			goto consist_inode;
592 	} while(1);
593 
594 	switch(ret) {
595 	case 0:
596 		return NULL;
597 	case 1:
598 		return dent;
599 	case 2:
600 		return prev ? prev : dent;
601 	default:
602 		BUG_ON(ret > 0);
603 		return ERR_PTR(ret);
604 	}
605 
606 consist_inode:
607 	gfs2_consist_inode(GFS2_I(inode));
608 	return ERR_PTR(-EIO);
609 }
610 
611 static int dirent_check_reclen(struct gfs2_inode *dip,
612 			       const struct gfs2_dirent *d, const void *end_p)
613 {
614 	const void *ptr = d;
615 	u16 rec_len = be16_to_cpu(d->de_rec_len);
616 
617 	if (unlikely(rec_len < sizeof(struct gfs2_dirent)))
618 		goto broken;
619 	ptr += rec_len;
620 	if (ptr < end_p)
621 		return rec_len;
622 	if (ptr == end_p)
623 		return -ENOENT;
624 broken:
625 	gfs2_consist_inode(dip);
626 	return -EIO;
627 }
628 
629 /**
630  * dirent_next - Next dirent
631  * @dip: the directory
632  * @bh: The buffer
633  * @dent: Pointer to list of dirents
634  *
635  * Returns: 0 on success, error code otherwise
636  */
637 
638 static int dirent_next(struct gfs2_inode *dip, struct buffer_head *bh,
639 		       struct gfs2_dirent **dent)
640 {
641 	struct gfs2_dirent *cur = *dent, *tmp;
642 	char *bh_end = bh->b_data + bh->b_size;
643 	int ret;
644 
645 	ret = dirent_check_reclen(dip, cur, bh_end);
646 	if (ret < 0)
647 		return ret;
648 
649 	tmp = (void *)cur + ret;
650 	ret = dirent_check_reclen(dip, tmp, bh_end);
651 	if (ret == -EIO)
652 		return ret;
653 
654         /* Only the first dent could ever have de_inum.no_addr == 0 */
655 	if (gfs2_dirent_sentinel(tmp)) {
656 		gfs2_consist_inode(dip);
657 		return -EIO;
658 	}
659 
660 	*dent = tmp;
661 	return 0;
662 }
663 
664 /**
665  * dirent_del - Delete a dirent
666  * @dip: The GFS2 inode
667  * @bh: The buffer
668  * @prev: The previous dirent
669  * @cur: The current dirent
670  *
671  */
672 
673 static void dirent_del(struct gfs2_inode *dip, struct buffer_head *bh,
674 		       struct gfs2_dirent *prev, struct gfs2_dirent *cur)
675 {
676 	u16 cur_rec_len, prev_rec_len;
677 
678 	if (gfs2_dirent_sentinel(cur)) {
679 		gfs2_consist_inode(dip);
680 		return;
681 	}
682 
683 	gfs2_trans_add_meta(dip->i_gl, bh);
684 
685 	/* If there is no prev entry, this is the first entry in the block.
686 	   The de_rec_len is already as big as it needs to be.  Just zero
687 	   out the inode number and return.  */
688 
689 	if (!prev) {
690 		cur->de_inum.no_addr = 0;
691 		cur->de_inum.no_formal_ino = 0;
692 		return;
693 	}
694 
695 	/*  Combine this dentry with the previous one.  */
696 
697 	prev_rec_len = be16_to_cpu(prev->de_rec_len);
698 	cur_rec_len = be16_to_cpu(cur->de_rec_len);
699 
700 	if ((char *)prev + prev_rec_len != (char *)cur)
701 		gfs2_consist_inode(dip);
702 	if ((char *)cur + cur_rec_len > bh->b_data + bh->b_size)
703 		gfs2_consist_inode(dip);
704 
705 	prev_rec_len += cur_rec_len;
706 	prev->de_rec_len = cpu_to_be16(prev_rec_len);
707 }
708 
709 
710 static struct gfs2_dirent *do_init_dirent(struct inode *inode,
711 					  struct gfs2_dirent *dent,
712 					  const struct qstr *name,
713 					  struct buffer_head *bh,
714 					  unsigned offset)
715 {
716 	struct gfs2_inode *ip = GFS2_I(inode);
717 	struct gfs2_dirent *ndent;
718 	unsigned totlen;
719 
720 	totlen = be16_to_cpu(dent->de_rec_len);
721 	BUG_ON(offset + name->len > totlen);
722 	gfs2_trans_add_meta(ip->i_gl, bh);
723 	ndent = (struct gfs2_dirent *)((char *)dent + offset);
724 	dent->de_rec_len = cpu_to_be16(offset);
725 	gfs2_qstr2dirent(name, totlen - offset, ndent);
726 	return ndent;
727 }
728 
729 
730 /*
731  * Takes a dent from which to grab space as an argument. Returns the
732  * newly created dent.
733  */
734 static struct gfs2_dirent *gfs2_init_dirent(struct inode *inode,
735 					    struct gfs2_dirent *dent,
736 					    const struct qstr *name,
737 					    struct buffer_head *bh)
738 {
739 	unsigned offset = 0;
740 
741 	if (!gfs2_dirent_sentinel(dent))
742 		offset = GFS2_DIRENT_SIZE(be16_to_cpu(dent->de_name_len));
743 	return do_init_dirent(inode, dent, name, bh, offset);
744 }
745 
746 static struct gfs2_dirent *gfs2_dirent_split_alloc(struct inode *inode,
747 						   struct buffer_head *bh,
748 						   const struct qstr *name,
749 						   void *ptr)
750 {
751 	struct gfs2_dirent *dent;
752 	dent = gfs2_dirent_scan(inode, bh->b_data, bh->b_size,
753 				gfs2_dirent_find_offset, name, ptr);
754 	if (!dent || IS_ERR(dent))
755 		return dent;
756 	return do_init_dirent(inode, dent, name, bh,
757 			      (unsigned)(ptr - (void *)dent));
758 }
759 
760 static int get_leaf(struct gfs2_inode *dip, u64 leaf_no,
761 		    struct buffer_head **bhp)
762 {
763 	int error;
764 
765 	error = gfs2_meta_read(dip->i_gl, leaf_no, DIO_WAIT, 0, bhp);
766 	if (!error && gfs2_metatype_check(GFS2_SB(&dip->i_inode), *bhp, GFS2_METATYPE_LF)) {
767 		/* pr_info("block num=%llu\n", leaf_no); */
768 		error = -EIO;
769 	}
770 
771 	return error;
772 }
773 
774 /**
775  * get_leaf_nr - Get a leaf number associated with the index
776  * @dip: The GFS2 inode
777  * @index:
778  * @leaf_out:
779  *
780  * Returns: 0 on success, error code otherwise
781  */
782 
783 static int get_leaf_nr(struct gfs2_inode *dip, u32 index,
784 		       u64 *leaf_out)
785 {
786 	__be64 *hash;
787 	int error;
788 
789 	hash = gfs2_dir_get_hash_table(dip);
790 	error = PTR_ERR_OR_ZERO(hash);
791 
792 	if (!error)
793 		*leaf_out = be64_to_cpu(*(hash + index));
794 
795 	return error;
796 }
797 
798 static int get_first_leaf(struct gfs2_inode *dip, u32 index,
799 			  struct buffer_head **bh_out)
800 {
801 	u64 leaf_no;
802 	int error;
803 
804 	error = get_leaf_nr(dip, index, &leaf_no);
805 	if (!error)
806 		error = get_leaf(dip, leaf_no, bh_out);
807 
808 	return error;
809 }
810 
811 static struct gfs2_dirent *gfs2_dirent_search(struct inode *inode,
812 					      const struct qstr *name,
813 					      gfs2_dscan_t scan,
814 					      struct buffer_head **pbh)
815 {
816 	struct buffer_head *bh;
817 	struct gfs2_dirent *dent;
818 	struct gfs2_inode *ip = GFS2_I(inode);
819 	int error;
820 
821 	if (ip->i_diskflags & GFS2_DIF_EXHASH) {
822 		struct gfs2_leaf *leaf;
823 		unsigned int hsize = BIT(ip->i_depth);
824 		unsigned int index;
825 		u64 ln;
826 		if (hsize * sizeof(u64) != i_size_read(inode)) {
827 			gfs2_consist_inode(ip);
828 			return ERR_PTR(-EIO);
829 		}
830 
831 		index = name->hash >> (32 - ip->i_depth);
832 		error = get_first_leaf(ip, index, &bh);
833 		if (error)
834 			return ERR_PTR(error);
835 		do {
836 			dent = gfs2_dirent_scan(inode, bh->b_data, bh->b_size,
837 						scan, name, NULL);
838 			if (dent)
839 				goto got_dent;
840 			leaf = (struct gfs2_leaf *)bh->b_data;
841 			ln = be64_to_cpu(leaf->lf_next);
842 			brelse(bh);
843 			if (!ln)
844 				break;
845 
846 			error = get_leaf(ip, ln, &bh);
847 		} while(!error);
848 
849 		return error ? ERR_PTR(error) : NULL;
850 	}
851 
852 
853 	error = gfs2_meta_inode_buffer(ip, &bh);
854 	if (error)
855 		return ERR_PTR(error);
856 	dent = gfs2_dirent_scan(inode, bh->b_data, bh->b_size, scan, name, NULL);
857 got_dent:
858 	if (unlikely(dent == NULL || IS_ERR(dent))) {
859 		brelse(bh);
860 		bh = NULL;
861 	}
862 	*pbh = bh;
863 	return dent;
864 }
865 
866 static struct gfs2_leaf *new_leaf(struct inode *inode, struct buffer_head **pbh, u16 depth)
867 {
868 	struct gfs2_inode *ip = GFS2_I(inode);
869 	unsigned int n = 1;
870 	u64 bn;
871 	int error;
872 	struct buffer_head *bh;
873 	struct gfs2_leaf *leaf;
874 	struct gfs2_dirent *dent;
875 	struct qstr name = { .name = "" };
876 	struct timespec tv = current_time(inode);
877 
878 	error = gfs2_alloc_blocks(ip, &bn, &n, 0, NULL);
879 	if (error)
880 		return NULL;
881 	bh = gfs2_meta_new(ip->i_gl, bn);
882 	if (!bh)
883 		return NULL;
884 
885 	gfs2_trans_add_unrevoke(GFS2_SB(inode), bn, 1);
886 	gfs2_trans_add_meta(ip->i_gl, bh);
887 	gfs2_metatype_set(bh, GFS2_METATYPE_LF, GFS2_FORMAT_LF);
888 	leaf = (struct gfs2_leaf *)bh->b_data;
889 	leaf->lf_depth = cpu_to_be16(depth);
890 	leaf->lf_entries = 0;
891 	leaf->lf_dirent_format = cpu_to_be32(GFS2_FORMAT_DE);
892 	leaf->lf_next = 0;
893 	leaf->lf_inode = cpu_to_be64(ip->i_no_addr);
894 	leaf->lf_dist = cpu_to_be32(1);
895 	leaf->lf_nsec = cpu_to_be32(tv.tv_nsec);
896 	leaf->lf_sec = cpu_to_be64(tv.tv_sec);
897 	memset(leaf->lf_reserved2, 0, sizeof(leaf->lf_reserved2));
898 	dent = (struct gfs2_dirent *)(leaf+1);
899 	gfs2_qstr2dirent(&name, bh->b_size - sizeof(struct gfs2_leaf), dent);
900 	*pbh = bh;
901 	return leaf;
902 }
903 
904 /**
905  * dir_make_exhash - Convert a stuffed directory into an ExHash directory
906  * @dip: The GFS2 inode
907  *
908  * Returns: 0 on success, error code otherwise
909  */
910 
911 static int dir_make_exhash(struct inode *inode)
912 {
913 	struct gfs2_inode *dip = GFS2_I(inode);
914 	struct gfs2_sbd *sdp = GFS2_SB(inode);
915 	struct gfs2_dirent *dent;
916 	struct qstr args;
917 	struct buffer_head *bh, *dibh;
918 	struct gfs2_leaf *leaf;
919 	int y;
920 	u32 x;
921 	__be64 *lp;
922 	u64 bn;
923 	int error;
924 
925 	error = gfs2_meta_inode_buffer(dip, &dibh);
926 	if (error)
927 		return error;
928 
929 	/*  Turn over a new leaf  */
930 
931 	leaf = new_leaf(inode, &bh, 0);
932 	if (!leaf)
933 		return -ENOSPC;
934 	bn = bh->b_blocknr;
935 
936 	gfs2_assert(sdp, dip->i_entries < BIT(16));
937 	leaf->lf_entries = cpu_to_be16(dip->i_entries);
938 
939 	/*  Copy dirents  */
940 
941 	gfs2_buffer_copy_tail(bh, sizeof(struct gfs2_leaf), dibh,
942 			     sizeof(struct gfs2_dinode));
943 
944 	/*  Find last entry  */
945 
946 	x = 0;
947 	args.len = bh->b_size - sizeof(struct gfs2_dinode) +
948 		   sizeof(struct gfs2_leaf);
949 	args.name = bh->b_data;
950 	dent = gfs2_dirent_scan(&dip->i_inode, bh->b_data, bh->b_size,
951 				gfs2_dirent_last, &args, NULL);
952 	if (!dent) {
953 		brelse(bh);
954 		brelse(dibh);
955 		return -EIO;
956 	}
957 	if (IS_ERR(dent)) {
958 		brelse(bh);
959 		brelse(dibh);
960 		return PTR_ERR(dent);
961 	}
962 
963 	/*  Adjust the last dirent's record length
964 	   (Remember that dent still points to the last entry.)  */
965 
966 	dent->de_rec_len = cpu_to_be16(be16_to_cpu(dent->de_rec_len) +
967 		sizeof(struct gfs2_dinode) -
968 		sizeof(struct gfs2_leaf));
969 
970 	brelse(bh);
971 
972 	/*  We're done with the new leaf block, now setup the new
973 	    hash table.  */
974 
975 	gfs2_trans_add_meta(dip->i_gl, dibh);
976 	gfs2_buffer_clear_tail(dibh, sizeof(struct gfs2_dinode));
977 
978 	lp = (__be64 *)(dibh->b_data + sizeof(struct gfs2_dinode));
979 
980 	for (x = sdp->sd_hash_ptrs; x--; lp++)
981 		*lp = cpu_to_be64(bn);
982 
983 	i_size_write(inode, sdp->sd_sb.sb_bsize / 2);
984 	gfs2_add_inode_blocks(&dip->i_inode, 1);
985 	dip->i_diskflags |= GFS2_DIF_EXHASH;
986 
987 	for (x = sdp->sd_hash_ptrs, y = -1; x; x >>= 1, y++) ;
988 	dip->i_depth = y;
989 
990 	gfs2_dinode_out(dip, dibh->b_data);
991 
992 	brelse(dibh);
993 
994 	return 0;
995 }
996 
997 /**
998  * dir_split_leaf - Split a leaf block into two
999  * @dip: The GFS2 inode
1000  * @index:
1001  * @leaf_no:
1002  *
1003  * Returns: 0 on success, error code on failure
1004  */
1005 
1006 static int dir_split_leaf(struct inode *inode, const struct qstr *name)
1007 {
1008 	struct gfs2_inode *dip = GFS2_I(inode);
1009 	struct buffer_head *nbh, *obh, *dibh;
1010 	struct gfs2_leaf *nleaf, *oleaf;
1011 	struct gfs2_dirent *dent = NULL, *prev = NULL, *next = NULL, *new;
1012 	u32 start, len, half_len, divider;
1013 	u64 bn, leaf_no;
1014 	__be64 *lp;
1015 	u32 index;
1016 	int x, moved = 0;
1017 	int error;
1018 
1019 	index = name->hash >> (32 - dip->i_depth);
1020 	error = get_leaf_nr(dip, index, &leaf_no);
1021 	if (error)
1022 		return error;
1023 
1024 	/*  Get the old leaf block  */
1025 	error = get_leaf(dip, leaf_no, &obh);
1026 	if (error)
1027 		return error;
1028 
1029 	oleaf = (struct gfs2_leaf *)obh->b_data;
1030 	if (dip->i_depth == be16_to_cpu(oleaf->lf_depth)) {
1031 		brelse(obh);
1032 		return 1; /* can't split */
1033 	}
1034 
1035 	gfs2_trans_add_meta(dip->i_gl, obh);
1036 
1037 	nleaf = new_leaf(inode, &nbh, be16_to_cpu(oleaf->lf_depth) + 1);
1038 	if (!nleaf) {
1039 		brelse(obh);
1040 		return -ENOSPC;
1041 	}
1042 	bn = nbh->b_blocknr;
1043 
1044 	/*  Compute the start and len of leaf pointers in the hash table.  */
1045 	len = BIT(dip->i_depth - be16_to_cpu(oleaf->lf_depth));
1046 	half_len = len >> 1;
1047 	if (!half_len) {
1048 		pr_warn("i_depth %u lf_depth %u index %u\n",
1049 			dip->i_depth, be16_to_cpu(oleaf->lf_depth), index);
1050 		gfs2_consist_inode(dip);
1051 		error = -EIO;
1052 		goto fail_brelse;
1053 	}
1054 
1055 	start = (index & ~(len - 1));
1056 
1057 	/* Change the pointers.
1058 	   Don't bother distinguishing stuffed from non-stuffed.
1059 	   This code is complicated enough already. */
1060 	lp = kmalloc(half_len * sizeof(__be64), GFP_NOFS);
1061 	if (!lp) {
1062 		error = -ENOMEM;
1063 		goto fail_brelse;
1064 	}
1065 
1066 	/*  Change the pointers  */
1067 	for (x = 0; x < half_len; x++)
1068 		lp[x] = cpu_to_be64(bn);
1069 
1070 	gfs2_dir_hash_inval(dip);
1071 
1072 	error = gfs2_dir_write_data(dip, (char *)lp, start * sizeof(u64),
1073 				    half_len * sizeof(u64));
1074 	if (error != half_len * sizeof(u64)) {
1075 		if (error >= 0)
1076 			error = -EIO;
1077 		goto fail_lpfree;
1078 	}
1079 
1080 	kfree(lp);
1081 
1082 	/*  Compute the divider  */
1083 	divider = (start + half_len) << (32 - dip->i_depth);
1084 
1085 	/*  Copy the entries  */
1086 	dent = (struct gfs2_dirent *)(obh->b_data + sizeof(struct gfs2_leaf));
1087 
1088 	do {
1089 		next = dent;
1090 		if (dirent_next(dip, obh, &next))
1091 			next = NULL;
1092 
1093 		if (!gfs2_dirent_sentinel(dent) &&
1094 		    be32_to_cpu(dent->de_hash) < divider) {
1095 			struct qstr str;
1096 			void *ptr = ((char *)dent - obh->b_data) + nbh->b_data;
1097 			str.name = (char*)(dent+1);
1098 			str.len = be16_to_cpu(dent->de_name_len);
1099 			str.hash = be32_to_cpu(dent->de_hash);
1100 			new = gfs2_dirent_split_alloc(inode, nbh, &str, ptr);
1101 			if (IS_ERR(new)) {
1102 				error = PTR_ERR(new);
1103 				break;
1104 			}
1105 
1106 			new->de_inum = dent->de_inum; /* No endian worries */
1107 			new->de_type = dent->de_type; /* No endian worries */
1108 			be16_add_cpu(&nleaf->lf_entries, 1);
1109 
1110 			dirent_del(dip, obh, prev, dent);
1111 
1112 			if (!oleaf->lf_entries)
1113 				gfs2_consist_inode(dip);
1114 			be16_add_cpu(&oleaf->lf_entries, -1);
1115 
1116 			if (!prev)
1117 				prev = dent;
1118 
1119 			moved = 1;
1120 		} else {
1121 			prev = dent;
1122 		}
1123 		dent = next;
1124 	} while (dent);
1125 
1126 	oleaf->lf_depth = nleaf->lf_depth;
1127 
1128 	error = gfs2_meta_inode_buffer(dip, &dibh);
1129 	if (!gfs2_assert_withdraw(GFS2_SB(&dip->i_inode), !error)) {
1130 		gfs2_trans_add_meta(dip->i_gl, dibh);
1131 		gfs2_add_inode_blocks(&dip->i_inode, 1);
1132 		gfs2_dinode_out(dip, dibh->b_data);
1133 		brelse(dibh);
1134 	}
1135 
1136 	brelse(obh);
1137 	brelse(nbh);
1138 
1139 	return error;
1140 
1141 fail_lpfree:
1142 	kfree(lp);
1143 
1144 fail_brelse:
1145 	brelse(obh);
1146 	brelse(nbh);
1147 	return error;
1148 }
1149 
1150 /**
1151  * dir_double_exhash - Double size of ExHash table
1152  * @dip: The GFS2 dinode
1153  *
1154  * Returns: 0 on success, error code on failure
1155  */
1156 
1157 static int dir_double_exhash(struct gfs2_inode *dip)
1158 {
1159 	struct buffer_head *dibh;
1160 	u32 hsize;
1161 	u32 hsize_bytes;
1162 	__be64 *hc;
1163 	__be64 *hc2, *h;
1164 	int x;
1165 	int error = 0;
1166 
1167 	hsize = BIT(dip->i_depth);
1168 	hsize_bytes = hsize * sizeof(__be64);
1169 
1170 	hc = gfs2_dir_get_hash_table(dip);
1171 	if (IS_ERR(hc))
1172 		return PTR_ERR(hc);
1173 
1174 	hc2 = kmalloc(hsize_bytes * 2, GFP_NOFS | __GFP_NOWARN);
1175 	if (hc2 == NULL)
1176 		hc2 = __vmalloc(hsize_bytes * 2, GFP_NOFS, PAGE_KERNEL);
1177 
1178 	if (!hc2)
1179 		return -ENOMEM;
1180 
1181 	h = hc2;
1182 	error = gfs2_meta_inode_buffer(dip, &dibh);
1183 	if (error)
1184 		goto out_kfree;
1185 
1186 	for (x = 0; x < hsize; x++) {
1187 		*h++ = *hc;
1188 		*h++ = *hc;
1189 		hc++;
1190 	}
1191 
1192 	error = gfs2_dir_write_data(dip, (char *)hc2, 0, hsize_bytes * 2);
1193 	if (error != (hsize_bytes * 2))
1194 		goto fail;
1195 
1196 	gfs2_dir_hash_inval(dip);
1197 	dip->i_hash_cache = hc2;
1198 	dip->i_depth++;
1199 	gfs2_dinode_out(dip, dibh->b_data);
1200 	brelse(dibh);
1201 	return 0;
1202 
1203 fail:
1204 	/* Replace original hash table & size */
1205 	gfs2_dir_write_data(dip, (char *)hc, 0, hsize_bytes);
1206 	i_size_write(&dip->i_inode, hsize_bytes);
1207 	gfs2_dinode_out(dip, dibh->b_data);
1208 	brelse(dibh);
1209 out_kfree:
1210 	kvfree(hc2);
1211 	return error;
1212 }
1213 
1214 /**
1215  * compare_dents - compare directory entries by hash value
1216  * @a: first dent
1217  * @b: second dent
1218  *
1219  * When comparing the hash entries of @a to @b:
1220  *   gt: returns 1
1221  *   lt: returns -1
1222  *   eq: returns 0
1223  */
1224 
1225 static int compare_dents(const void *a, const void *b)
1226 {
1227 	const struct gfs2_dirent *dent_a, *dent_b;
1228 	u32 hash_a, hash_b;
1229 	int ret = 0;
1230 
1231 	dent_a = *(const struct gfs2_dirent **)a;
1232 	hash_a = dent_a->de_cookie;
1233 
1234 	dent_b = *(const struct gfs2_dirent **)b;
1235 	hash_b = dent_b->de_cookie;
1236 
1237 	if (hash_a > hash_b)
1238 		ret = 1;
1239 	else if (hash_a < hash_b)
1240 		ret = -1;
1241 	else {
1242 		unsigned int len_a = be16_to_cpu(dent_a->de_name_len);
1243 		unsigned int len_b = be16_to_cpu(dent_b->de_name_len);
1244 
1245 		if (len_a > len_b)
1246 			ret = 1;
1247 		else if (len_a < len_b)
1248 			ret = -1;
1249 		else
1250 			ret = memcmp(dent_a + 1, dent_b + 1, len_a);
1251 	}
1252 
1253 	return ret;
1254 }
1255 
1256 /**
1257  * do_filldir_main - read out directory entries
1258  * @dip: The GFS2 inode
1259  * @ctx: what to feed the entries to
1260  * @darr: an array of struct gfs2_dirent pointers to read
1261  * @entries: the number of entries in darr
1262  * @copied: pointer to int that's non-zero if a entry has been copied out
1263  *
1264  * Jump through some hoops to make sure that if there are hash collsions,
1265  * they are read out at the beginning of a buffer.  We want to minimize
1266  * the possibility that they will fall into different readdir buffers or
1267  * that someone will want to seek to that location.
1268  *
1269  * Returns: errno, >0 if the actor tells you to stop
1270  */
1271 
1272 static int do_filldir_main(struct gfs2_inode *dip, struct dir_context *ctx,
1273 			   struct gfs2_dirent **darr, u32 entries,
1274 			   u32 sort_start, int *copied)
1275 {
1276 	const struct gfs2_dirent *dent, *dent_next;
1277 	u64 off, off_next;
1278 	unsigned int x, y;
1279 	int run = 0;
1280 
1281 	if (sort_start < entries)
1282 		sort(&darr[sort_start], entries - sort_start,
1283 		     sizeof(struct gfs2_dirent *), compare_dents, NULL);
1284 
1285 	dent_next = darr[0];
1286 	off_next = dent_next->de_cookie;
1287 
1288 	for (x = 0, y = 1; x < entries; x++, y++) {
1289 		dent = dent_next;
1290 		off = off_next;
1291 
1292 		if (y < entries) {
1293 			dent_next = darr[y];
1294 			off_next = dent_next->de_cookie;
1295 
1296 			if (off < ctx->pos)
1297 				continue;
1298 			ctx->pos = off;
1299 
1300 			if (off_next == off) {
1301 				if (*copied && !run)
1302 					return 1;
1303 				run = 1;
1304 			} else
1305 				run = 0;
1306 		} else {
1307 			if (off < ctx->pos)
1308 				continue;
1309 			ctx->pos = off;
1310 		}
1311 
1312 		if (!dir_emit(ctx, (const char *)(dent + 1),
1313 				be16_to_cpu(dent->de_name_len),
1314 				be64_to_cpu(dent->de_inum.no_addr),
1315 				be16_to_cpu(dent->de_type)))
1316 			return 1;
1317 
1318 		*copied = 1;
1319 	}
1320 
1321 	/* Increment the ctx->pos by one, so the next time we come into the
1322 	   do_filldir fxn, we get the next entry instead of the last one in the
1323 	   current leaf */
1324 
1325 	ctx->pos++;
1326 
1327 	return 0;
1328 }
1329 
1330 static void *gfs2_alloc_sort_buffer(unsigned size)
1331 {
1332 	void *ptr = NULL;
1333 
1334 	if (size < KMALLOC_MAX_SIZE)
1335 		ptr = kmalloc(size, GFP_NOFS | __GFP_NOWARN);
1336 	if (!ptr)
1337 		ptr = __vmalloc(size, GFP_NOFS, PAGE_KERNEL);
1338 	return ptr;
1339 }
1340 
1341 
1342 static int gfs2_set_cookies(struct gfs2_sbd *sdp, struct buffer_head *bh,
1343 			    unsigned leaf_nr, struct gfs2_dirent **darr,
1344 			    unsigned entries)
1345 {
1346 	int sort_id = -1;
1347 	int i;
1348 
1349 	for (i = 0; i < entries; i++) {
1350 		unsigned offset;
1351 
1352 		darr[i]->de_cookie = be32_to_cpu(darr[i]->de_hash);
1353 		darr[i]->de_cookie = gfs2_disk_hash2offset(darr[i]->de_cookie);
1354 
1355 		if (!sdp->sd_args.ar_loccookie)
1356 			continue;
1357 		offset = (char *)(darr[i]) -
1358 			 (bh->b_data + gfs2_dirent_offset(bh->b_data));
1359 		offset /= GFS2_MIN_DIRENT_SIZE;
1360 		offset += leaf_nr * sdp->sd_max_dents_per_leaf;
1361 		if (offset >= GFS2_USE_HASH_FLAG ||
1362 		    leaf_nr >= GFS2_USE_HASH_FLAG) {
1363 			darr[i]->de_cookie |= GFS2_USE_HASH_FLAG;
1364 			if (sort_id < 0)
1365 				sort_id = i;
1366 			continue;
1367 		}
1368 		darr[i]->de_cookie &= GFS2_HASH_INDEX_MASK;
1369 		darr[i]->de_cookie |= offset;
1370 	}
1371 	return sort_id;
1372 }
1373 
1374 
1375 static int gfs2_dir_read_leaf(struct inode *inode, struct dir_context *ctx,
1376 			      int *copied, unsigned *depth,
1377 			      u64 leaf_no)
1378 {
1379 	struct gfs2_inode *ip = GFS2_I(inode);
1380 	struct gfs2_sbd *sdp = GFS2_SB(inode);
1381 	struct buffer_head *bh;
1382 	struct gfs2_leaf *lf;
1383 	unsigned entries = 0, entries2 = 0;
1384 	unsigned leaves = 0, leaf = 0, offset, sort_offset;
1385 	struct gfs2_dirent **darr, *dent;
1386 	struct dirent_gather g;
1387 	struct buffer_head **larr;
1388 	int error, i, need_sort = 0, sort_id;
1389 	u64 lfn = leaf_no;
1390 
1391 	do {
1392 		error = get_leaf(ip, lfn, &bh);
1393 		if (error)
1394 			goto out;
1395 		lf = (struct gfs2_leaf *)bh->b_data;
1396 		if (leaves == 0)
1397 			*depth = be16_to_cpu(lf->lf_depth);
1398 		entries += be16_to_cpu(lf->lf_entries);
1399 		leaves++;
1400 		lfn = be64_to_cpu(lf->lf_next);
1401 		brelse(bh);
1402 	} while(lfn);
1403 
1404 	if (*depth < GFS2_DIR_MAX_DEPTH || !sdp->sd_args.ar_loccookie) {
1405 		need_sort = 1;
1406 		sort_offset = 0;
1407 	}
1408 
1409 	if (!entries)
1410 		return 0;
1411 
1412 	error = -ENOMEM;
1413 	/*
1414 	 * The extra 99 entries are not normally used, but are a buffer
1415 	 * zone in case the number of entries in the leaf is corrupt.
1416 	 * 99 is the maximum number of entries that can fit in a single
1417 	 * leaf block.
1418 	 */
1419 	larr = gfs2_alloc_sort_buffer((leaves + entries + 99) * sizeof(void *));
1420 	if (!larr)
1421 		goto out;
1422 	darr = (struct gfs2_dirent **)(larr + leaves);
1423 	g.pdent = (const struct gfs2_dirent **)darr;
1424 	g.offset = 0;
1425 	lfn = leaf_no;
1426 
1427 	do {
1428 		error = get_leaf(ip, lfn, &bh);
1429 		if (error)
1430 			goto out_free;
1431 		lf = (struct gfs2_leaf *)bh->b_data;
1432 		lfn = be64_to_cpu(lf->lf_next);
1433 		if (lf->lf_entries) {
1434 			offset = g.offset;
1435 			entries2 += be16_to_cpu(lf->lf_entries);
1436 			dent = gfs2_dirent_scan(inode, bh->b_data, bh->b_size,
1437 						gfs2_dirent_gather, NULL, &g);
1438 			error = PTR_ERR(dent);
1439 			if (IS_ERR(dent))
1440 				goto out_free;
1441 			if (entries2 != g.offset) {
1442 				fs_warn(sdp, "Number of entries corrupt in dir "
1443 						"leaf %llu, entries2 (%u) != "
1444 						"g.offset (%u)\n",
1445 					(unsigned long long)bh->b_blocknr,
1446 					entries2, g.offset);
1447 
1448 				error = -EIO;
1449 				goto out_free;
1450 			}
1451 			error = 0;
1452 			sort_id = gfs2_set_cookies(sdp, bh, leaf, &darr[offset],
1453 						   be16_to_cpu(lf->lf_entries));
1454 			if (!need_sort && sort_id >= 0) {
1455 				need_sort = 1;
1456 				sort_offset = offset + sort_id;
1457 			}
1458 			larr[leaf++] = bh;
1459 		} else {
1460 			larr[leaf++] = NULL;
1461 			brelse(bh);
1462 		}
1463 	} while(lfn);
1464 
1465 	BUG_ON(entries2 != entries);
1466 	error = do_filldir_main(ip, ctx, darr, entries, need_sort ?
1467 				sort_offset : entries, copied);
1468 out_free:
1469 	for(i = 0; i < leaf; i++)
1470 		if (larr[i])
1471 			brelse(larr[i]);
1472 	kvfree(larr);
1473 out:
1474 	return error;
1475 }
1476 
1477 /**
1478  * gfs2_dir_readahead - Issue read-ahead requests for leaf blocks.
1479  *
1480  * Note: we can't calculate each index like dir_e_read can because we don't
1481  * have the leaf, and therefore we don't have the depth, and therefore we
1482  * don't have the length. So we have to just read enough ahead to make up
1483  * for the loss of information.
1484  */
1485 static void gfs2_dir_readahead(struct inode *inode, unsigned hsize, u32 index,
1486 			       struct file_ra_state *f_ra)
1487 {
1488 	struct gfs2_inode *ip = GFS2_I(inode);
1489 	struct gfs2_glock *gl = ip->i_gl;
1490 	struct buffer_head *bh;
1491 	u64 blocknr = 0, last;
1492 	unsigned count;
1493 
1494 	/* First check if we've already read-ahead for the whole range. */
1495 	if (index + MAX_RA_BLOCKS < f_ra->start)
1496 		return;
1497 
1498 	f_ra->start = max((pgoff_t)index, f_ra->start);
1499 	for (count = 0; count < MAX_RA_BLOCKS; count++) {
1500 		if (f_ra->start >= hsize) /* if exceeded the hash table */
1501 			break;
1502 
1503 		last = blocknr;
1504 		blocknr = be64_to_cpu(ip->i_hash_cache[f_ra->start]);
1505 		f_ra->start++;
1506 		if (blocknr == last)
1507 			continue;
1508 
1509 		bh = gfs2_getbuf(gl, blocknr, 1);
1510 		if (trylock_buffer(bh)) {
1511 			if (buffer_uptodate(bh)) {
1512 				unlock_buffer(bh);
1513 				brelse(bh);
1514 				continue;
1515 			}
1516 			bh->b_end_io = end_buffer_read_sync;
1517 			submit_bh(REQ_OP_READ, REQ_RAHEAD | REQ_META, bh);
1518 			continue;
1519 		}
1520 		brelse(bh);
1521 	}
1522 }
1523 
1524 /**
1525  * dir_e_read - Reads the entries from a directory into a filldir buffer
1526  * @dip: dinode pointer
1527  * @ctx: actor to feed the entries to
1528  *
1529  * Returns: errno
1530  */
1531 
1532 static int dir_e_read(struct inode *inode, struct dir_context *ctx,
1533 		      struct file_ra_state *f_ra)
1534 {
1535 	struct gfs2_inode *dip = GFS2_I(inode);
1536 	u32 hsize, len = 0;
1537 	u32 hash, index;
1538 	__be64 *lp;
1539 	int copied = 0;
1540 	int error = 0;
1541 	unsigned depth = 0;
1542 
1543 	hsize = BIT(dip->i_depth);
1544 	hash = gfs2_dir_offset2hash(ctx->pos);
1545 	index = hash >> (32 - dip->i_depth);
1546 
1547 	if (dip->i_hash_cache == NULL)
1548 		f_ra->start = 0;
1549 	lp = gfs2_dir_get_hash_table(dip);
1550 	if (IS_ERR(lp))
1551 		return PTR_ERR(lp);
1552 
1553 	gfs2_dir_readahead(inode, hsize, index, f_ra);
1554 
1555 	while (index < hsize) {
1556 		error = gfs2_dir_read_leaf(inode, ctx,
1557 					   &copied, &depth,
1558 					   be64_to_cpu(lp[index]));
1559 		if (error)
1560 			break;
1561 
1562 		len = BIT(dip->i_depth - depth);
1563 		index = (index & ~(len - 1)) + len;
1564 	}
1565 
1566 	if (error > 0)
1567 		error = 0;
1568 	return error;
1569 }
1570 
1571 int gfs2_dir_read(struct inode *inode, struct dir_context *ctx,
1572 		  struct file_ra_state *f_ra)
1573 {
1574 	struct gfs2_inode *dip = GFS2_I(inode);
1575 	struct gfs2_sbd *sdp = GFS2_SB(inode);
1576 	struct dirent_gather g;
1577 	struct gfs2_dirent **darr, *dent;
1578 	struct buffer_head *dibh;
1579 	int copied = 0;
1580 	int error;
1581 
1582 	if (!dip->i_entries)
1583 		return 0;
1584 
1585 	if (dip->i_diskflags & GFS2_DIF_EXHASH)
1586 		return dir_e_read(inode, ctx, f_ra);
1587 
1588 	if (!gfs2_is_stuffed(dip)) {
1589 		gfs2_consist_inode(dip);
1590 		return -EIO;
1591 	}
1592 
1593 	error = gfs2_meta_inode_buffer(dip, &dibh);
1594 	if (error)
1595 		return error;
1596 
1597 	error = -ENOMEM;
1598 	/* 96 is max number of dirents which can be stuffed into an inode */
1599 	darr = kmalloc(96 * sizeof(struct gfs2_dirent *), GFP_NOFS);
1600 	if (darr) {
1601 		g.pdent = (const struct gfs2_dirent **)darr;
1602 		g.offset = 0;
1603 		dent = gfs2_dirent_scan(inode, dibh->b_data, dibh->b_size,
1604 					gfs2_dirent_gather, NULL, &g);
1605 		if (IS_ERR(dent)) {
1606 			error = PTR_ERR(dent);
1607 			goto out;
1608 		}
1609 		if (dip->i_entries != g.offset) {
1610 			fs_warn(sdp, "Number of entries corrupt in dir %llu, "
1611 				"ip->i_entries (%u) != g.offset (%u)\n",
1612 				(unsigned long long)dip->i_no_addr,
1613 				dip->i_entries,
1614 				g.offset);
1615 			error = -EIO;
1616 			goto out;
1617 		}
1618 		gfs2_set_cookies(sdp, dibh, 0, darr, dip->i_entries);
1619 		error = do_filldir_main(dip, ctx, darr,
1620 					dip->i_entries, 0, &copied);
1621 out:
1622 		kfree(darr);
1623 	}
1624 
1625 	if (error > 0)
1626 		error = 0;
1627 
1628 	brelse(dibh);
1629 
1630 	return error;
1631 }
1632 
1633 /**
1634  * gfs2_dir_search - Search a directory
1635  * @dip: The GFS2 dir inode
1636  * @name: The name we are looking up
1637  * @fail_on_exist: Fail if the name exists rather than looking it up
1638  *
1639  * This routine searches a directory for a file or another directory.
1640  * Assumes a glock is held on dip.
1641  *
1642  * Returns: errno
1643  */
1644 
1645 struct inode *gfs2_dir_search(struct inode *dir, const struct qstr *name,
1646 			      bool fail_on_exist)
1647 {
1648 	struct buffer_head *bh;
1649 	struct gfs2_dirent *dent;
1650 	u64 addr, formal_ino;
1651 	u16 dtype;
1652 
1653 	dent = gfs2_dirent_search(dir, name, gfs2_dirent_find, &bh);
1654 	if (dent) {
1655 		struct inode *inode;
1656 		u16 rahead;
1657 
1658 		if (IS_ERR(dent))
1659 			return ERR_CAST(dent);
1660 		dtype = be16_to_cpu(dent->de_type);
1661 		rahead = be16_to_cpu(dent->de_rahead);
1662 		addr = be64_to_cpu(dent->de_inum.no_addr);
1663 		formal_ino = be64_to_cpu(dent->de_inum.no_formal_ino);
1664 		brelse(bh);
1665 		if (fail_on_exist)
1666 			return ERR_PTR(-EEXIST);
1667 		inode = gfs2_inode_lookup(dir->i_sb, dtype, addr, formal_ino,
1668 					  GFS2_BLKST_FREE /* ignore */);
1669 		if (!IS_ERR(inode))
1670 			GFS2_I(inode)->i_rahead = rahead;
1671 		return inode;
1672 	}
1673 	return ERR_PTR(-ENOENT);
1674 }
1675 
1676 int gfs2_dir_check(struct inode *dir, const struct qstr *name,
1677 		   const struct gfs2_inode *ip)
1678 {
1679 	struct buffer_head *bh;
1680 	struct gfs2_dirent *dent;
1681 	int ret = -ENOENT;
1682 
1683 	dent = gfs2_dirent_search(dir, name, gfs2_dirent_find, &bh);
1684 	if (dent) {
1685 		if (IS_ERR(dent))
1686 			return PTR_ERR(dent);
1687 		if (ip) {
1688 			if (be64_to_cpu(dent->de_inum.no_addr) != ip->i_no_addr)
1689 				goto out;
1690 			if (be64_to_cpu(dent->de_inum.no_formal_ino) !=
1691 			    ip->i_no_formal_ino)
1692 				goto out;
1693 			if (unlikely(IF2DT(ip->i_inode.i_mode) !=
1694 			    be16_to_cpu(dent->de_type))) {
1695 				gfs2_consist_inode(GFS2_I(dir));
1696 				ret = -EIO;
1697 				goto out;
1698 			}
1699 		}
1700 		ret = 0;
1701 out:
1702 		brelse(bh);
1703 	}
1704 	return ret;
1705 }
1706 
1707 /**
1708  * dir_new_leaf - Add a new leaf onto hash chain
1709  * @inode: The directory
1710  * @name: The name we are adding
1711  *
1712  * This adds a new dir leaf onto an existing leaf when there is not
1713  * enough space to add a new dir entry. This is a last resort after
1714  * we've expanded the hash table to max size and also split existing
1715  * leaf blocks, so it will only occur for very large directories.
1716  *
1717  * The dist parameter is set to 1 for leaf blocks directly attached
1718  * to the hash table, 2 for one layer of indirection, 3 for two layers
1719  * etc. We are thus able to tell the difference between an old leaf
1720  * with dist set to zero (i.e. "don't know") and a new one where we
1721  * set this information for debug/fsck purposes.
1722  *
1723  * Returns: 0 on success, or -ve on error
1724  */
1725 
1726 static int dir_new_leaf(struct inode *inode, const struct qstr *name)
1727 {
1728 	struct buffer_head *bh, *obh;
1729 	struct gfs2_inode *ip = GFS2_I(inode);
1730 	struct gfs2_leaf *leaf, *oleaf;
1731 	u32 dist = 1;
1732 	int error;
1733 	u32 index;
1734 	u64 bn;
1735 
1736 	index = name->hash >> (32 - ip->i_depth);
1737 	error = get_first_leaf(ip, index, &obh);
1738 	if (error)
1739 		return error;
1740 	do {
1741 		dist++;
1742 		oleaf = (struct gfs2_leaf *)obh->b_data;
1743 		bn = be64_to_cpu(oleaf->lf_next);
1744 		if (!bn)
1745 			break;
1746 		brelse(obh);
1747 		error = get_leaf(ip, bn, &obh);
1748 		if (error)
1749 			return error;
1750 	} while(1);
1751 
1752 	gfs2_trans_add_meta(ip->i_gl, obh);
1753 
1754 	leaf = new_leaf(inode, &bh, be16_to_cpu(oleaf->lf_depth));
1755 	if (!leaf) {
1756 		brelse(obh);
1757 		return -ENOSPC;
1758 	}
1759 	leaf->lf_dist = cpu_to_be32(dist);
1760 	oleaf->lf_next = cpu_to_be64(bh->b_blocknr);
1761 	brelse(bh);
1762 	brelse(obh);
1763 
1764 	error = gfs2_meta_inode_buffer(ip, &bh);
1765 	if (error)
1766 		return error;
1767 	gfs2_trans_add_meta(ip->i_gl, bh);
1768 	gfs2_add_inode_blocks(&ip->i_inode, 1);
1769 	gfs2_dinode_out(ip, bh->b_data);
1770 	brelse(bh);
1771 	return 0;
1772 }
1773 
1774 static u16 gfs2_inode_ra_len(const struct gfs2_inode *ip)
1775 {
1776 	u64 where = ip->i_no_addr + 1;
1777 	if (ip->i_eattr == where)
1778 		return 1;
1779 	return 0;
1780 }
1781 
1782 /**
1783  * gfs2_dir_add - Add new filename into directory
1784  * @inode: The directory inode
1785  * @name: The new name
1786  * @nip: The GFS2 inode to be linked in to the directory
1787  * @da: The directory addition info
1788  *
1789  * If the call to gfs2_diradd_alloc_required resulted in there being
1790  * no need to allocate any new directory blocks, then it will contain
1791  * a pointer to the directory entry and the bh in which it resides. We
1792  * can use that without having to repeat the search. If there was no
1793  * free space, then we must now create more space.
1794  *
1795  * Returns: 0 on success, error code on failure
1796  */
1797 
1798 int gfs2_dir_add(struct inode *inode, const struct qstr *name,
1799 		 const struct gfs2_inode *nip, struct gfs2_diradd *da)
1800 {
1801 	struct gfs2_inode *ip = GFS2_I(inode);
1802 	struct buffer_head *bh = da->bh;
1803 	struct gfs2_dirent *dent = da->dent;
1804 	struct timespec tv;
1805 	struct gfs2_leaf *leaf;
1806 	int error;
1807 
1808 	while(1) {
1809 		if (da->bh == NULL) {
1810 			dent = gfs2_dirent_search(inode, name,
1811 						  gfs2_dirent_find_space, &bh);
1812 		}
1813 		if (dent) {
1814 			if (IS_ERR(dent))
1815 				return PTR_ERR(dent);
1816 			dent = gfs2_init_dirent(inode, dent, name, bh);
1817 			gfs2_inum_out(nip, dent);
1818 			dent->de_type = cpu_to_be16(IF2DT(nip->i_inode.i_mode));
1819 			dent->de_rahead = cpu_to_be16(gfs2_inode_ra_len(nip));
1820 			tv = current_time(&ip->i_inode);
1821 			if (ip->i_diskflags & GFS2_DIF_EXHASH) {
1822 				leaf = (struct gfs2_leaf *)bh->b_data;
1823 				be16_add_cpu(&leaf->lf_entries, 1);
1824 				leaf->lf_nsec = cpu_to_be32(tv.tv_nsec);
1825 				leaf->lf_sec = cpu_to_be64(tv.tv_sec);
1826 			}
1827 			da->dent = NULL;
1828 			da->bh = NULL;
1829 			brelse(bh);
1830 			ip->i_entries++;
1831 			ip->i_inode.i_mtime = ip->i_inode.i_ctime = tv;
1832 			if (S_ISDIR(nip->i_inode.i_mode))
1833 				inc_nlink(&ip->i_inode);
1834 			mark_inode_dirty(inode);
1835 			error = 0;
1836 			break;
1837 		}
1838 		if (!(ip->i_diskflags & GFS2_DIF_EXHASH)) {
1839 			error = dir_make_exhash(inode);
1840 			if (error)
1841 				break;
1842 			continue;
1843 		}
1844 		error = dir_split_leaf(inode, name);
1845 		if (error == 0)
1846 			continue;
1847 		if (error < 0)
1848 			break;
1849 		if (ip->i_depth < GFS2_DIR_MAX_DEPTH) {
1850 			error = dir_double_exhash(ip);
1851 			if (error)
1852 				break;
1853 			error = dir_split_leaf(inode, name);
1854 			if (error < 0)
1855 				break;
1856 			if (error == 0)
1857 				continue;
1858 		}
1859 		error = dir_new_leaf(inode, name);
1860 		if (!error)
1861 			continue;
1862 		error = -ENOSPC;
1863 		break;
1864 	}
1865 	return error;
1866 }
1867 
1868 
1869 /**
1870  * gfs2_dir_del - Delete a directory entry
1871  * @dip: The GFS2 inode
1872  * @filename: The filename
1873  *
1874  * Returns: 0 on success, error code on failure
1875  */
1876 
1877 int gfs2_dir_del(struct gfs2_inode *dip, const struct dentry *dentry)
1878 {
1879 	const struct qstr *name = &dentry->d_name;
1880 	struct gfs2_dirent *dent, *prev = NULL;
1881 	struct buffer_head *bh;
1882 	struct timespec tv = current_time(&dip->i_inode);
1883 
1884 	/* Returns _either_ the entry (if its first in block) or the
1885 	   previous entry otherwise */
1886 	dent = gfs2_dirent_search(&dip->i_inode, name, gfs2_dirent_prev, &bh);
1887 	if (!dent) {
1888 		gfs2_consist_inode(dip);
1889 		return -EIO;
1890 	}
1891 	if (IS_ERR(dent)) {
1892 		gfs2_consist_inode(dip);
1893 		return PTR_ERR(dent);
1894 	}
1895 	/* If not first in block, adjust pointers accordingly */
1896 	if (gfs2_dirent_find(dent, name, NULL) == 0) {
1897 		prev = dent;
1898 		dent = (struct gfs2_dirent *)((char *)dent + be16_to_cpu(prev->de_rec_len));
1899 	}
1900 
1901 	dirent_del(dip, bh, prev, dent);
1902 	if (dip->i_diskflags & GFS2_DIF_EXHASH) {
1903 		struct gfs2_leaf *leaf = (struct gfs2_leaf *)bh->b_data;
1904 		u16 entries = be16_to_cpu(leaf->lf_entries);
1905 		if (!entries)
1906 			gfs2_consist_inode(dip);
1907 		leaf->lf_entries = cpu_to_be16(--entries);
1908 		leaf->lf_nsec = cpu_to_be32(tv.tv_nsec);
1909 		leaf->lf_sec = cpu_to_be64(tv.tv_sec);
1910 	}
1911 	brelse(bh);
1912 
1913 	if (!dip->i_entries)
1914 		gfs2_consist_inode(dip);
1915 	dip->i_entries--;
1916 	dip->i_inode.i_mtime = dip->i_inode.i_ctime = tv;
1917 	if (d_is_dir(dentry))
1918 		drop_nlink(&dip->i_inode);
1919 	mark_inode_dirty(&dip->i_inode);
1920 
1921 	return 0;
1922 }
1923 
1924 /**
1925  * gfs2_dir_mvino - Change inode number of directory entry
1926  * @dip: The GFS2 inode
1927  * @filename:
1928  * @new_inode:
1929  *
1930  * This routine changes the inode number of a directory entry.  It's used
1931  * by rename to change ".." when a directory is moved.
1932  * Assumes a glock is held on dvp.
1933  *
1934  * Returns: errno
1935  */
1936 
1937 int gfs2_dir_mvino(struct gfs2_inode *dip, const struct qstr *filename,
1938 		   const struct gfs2_inode *nip, unsigned int new_type)
1939 {
1940 	struct buffer_head *bh;
1941 	struct gfs2_dirent *dent;
1942 	int error;
1943 
1944 	dent = gfs2_dirent_search(&dip->i_inode, filename, gfs2_dirent_find, &bh);
1945 	if (!dent) {
1946 		gfs2_consist_inode(dip);
1947 		return -EIO;
1948 	}
1949 	if (IS_ERR(dent))
1950 		return PTR_ERR(dent);
1951 
1952 	gfs2_trans_add_meta(dip->i_gl, bh);
1953 	gfs2_inum_out(nip, dent);
1954 	dent->de_type = cpu_to_be16(new_type);
1955 
1956 	if (dip->i_diskflags & GFS2_DIF_EXHASH) {
1957 		brelse(bh);
1958 		error = gfs2_meta_inode_buffer(dip, &bh);
1959 		if (error)
1960 			return error;
1961 		gfs2_trans_add_meta(dip->i_gl, bh);
1962 	}
1963 
1964 	dip->i_inode.i_mtime = dip->i_inode.i_ctime = current_time(&dip->i_inode);
1965 	gfs2_dinode_out(dip, bh->b_data);
1966 	brelse(bh);
1967 	return 0;
1968 }
1969 
1970 /**
1971  * leaf_dealloc - Deallocate a directory leaf
1972  * @dip: the directory
1973  * @index: the hash table offset in the directory
1974  * @len: the number of pointers to this leaf
1975  * @leaf_no: the leaf number
1976  * @leaf_bh: buffer_head for the starting leaf
1977  * last_dealloc: 1 if this is the final dealloc for the leaf, else 0
1978  *
1979  * Returns: errno
1980  */
1981 
1982 static int leaf_dealloc(struct gfs2_inode *dip, u32 index, u32 len,
1983 			u64 leaf_no, struct buffer_head *leaf_bh,
1984 			int last_dealloc)
1985 {
1986 	struct gfs2_sbd *sdp = GFS2_SB(&dip->i_inode);
1987 	struct gfs2_leaf *tmp_leaf;
1988 	struct gfs2_rgrp_list rlist;
1989 	struct buffer_head *bh, *dibh;
1990 	u64 blk, nblk;
1991 	unsigned int rg_blocks = 0, l_blocks = 0;
1992 	char *ht;
1993 	unsigned int x, size = len * sizeof(u64);
1994 	int error;
1995 
1996 	error = gfs2_rindex_update(sdp);
1997 	if (error)
1998 		return error;
1999 
2000 	memset(&rlist, 0, sizeof(struct gfs2_rgrp_list));
2001 
2002 	ht = kzalloc(size, GFP_NOFS | __GFP_NOWARN);
2003 	if (ht == NULL)
2004 		ht = __vmalloc(size, GFP_NOFS | __GFP_NOWARN | __GFP_ZERO,
2005 			       PAGE_KERNEL);
2006 	if (!ht)
2007 		return -ENOMEM;
2008 
2009 	error = gfs2_quota_hold(dip, NO_UID_QUOTA_CHANGE, NO_GID_QUOTA_CHANGE);
2010 	if (error)
2011 		goto out;
2012 
2013 	/*  Count the number of leaves  */
2014 	bh = leaf_bh;
2015 
2016 	for (blk = leaf_no; blk; blk = nblk) {
2017 		if (blk != leaf_no) {
2018 			error = get_leaf(dip, blk, &bh);
2019 			if (error)
2020 				goto out_rlist;
2021 		}
2022 		tmp_leaf = (struct gfs2_leaf *)bh->b_data;
2023 		nblk = be64_to_cpu(tmp_leaf->lf_next);
2024 		if (blk != leaf_no)
2025 			brelse(bh);
2026 
2027 		gfs2_rlist_add(dip, &rlist, blk);
2028 		l_blocks++;
2029 	}
2030 
2031 	gfs2_rlist_alloc(&rlist, LM_ST_EXCLUSIVE);
2032 
2033 	for (x = 0; x < rlist.rl_rgrps; x++) {
2034 		struct gfs2_rgrpd *rgd;
2035 		rgd = rlist.rl_ghs[x].gh_gl->gl_object;
2036 		rg_blocks += rgd->rd_length;
2037 	}
2038 
2039 	error = gfs2_glock_nq_m(rlist.rl_rgrps, rlist.rl_ghs);
2040 	if (error)
2041 		goto out_rlist;
2042 
2043 	error = gfs2_trans_begin(sdp,
2044 			rg_blocks + (DIV_ROUND_UP(size, sdp->sd_jbsize) + 1) +
2045 			RES_DINODE + RES_STATFS + RES_QUOTA, l_blocks);
2046 	if (error)
2047 		goto out_rg_gunlock;
2048 
2049 	bh = leaf_bh;
2050 
2051 	for (blk = leaf_no; blk; blk = nblk) {
2052 		if (blk != leaf_no) {
2053 			error = get_leaf(dip, blk, &bh);
2054 			if (error)
2055 				goto out_end_trans;
2056 		}
2057 		tmp_leaf = (struct gfs2_leaf *)bh->b_data;
2058 		nblk = be64_to_cpu(tmp_leaf->lf_next);
2059 		if (blk != leaf_no)
2060 			brelse(bh);
2061 
2062 		gfs2_free_meta(dip, blk, 1);
2063 		gfs2_add_inode_blocks(&dip->i_inode, -1);
2064 	}
2065 
2066 	error = gfs2_dir_write_data(dip, ht, index * sizeof(u64), size);
2067 	if (error != size) {
2068 		if (error >= 0)
2069 			error = -EIO;
2070 		goto out_end_trans;
2071 	}
2072 
2073 	error = gfs2_meta_inode_buffer(dip, &dibh);
2074 	if (error)
2075 		goto out_end_trans;
2076 
2077 	gfs2_trans_add_meta(dip->i_gl, dibh);
2078 	/* On the last dealloc, make this a regular file in case we crash.
2079 	   (We don't want to free these blocks a second time.)  */
2080 	if (last_dealloc)
2081 		dip->i_inode.i_mode = S_IFREG;
2082 	gfs2_dinode_out(dip, dibh->b_data);
2083 	brelse(dibh);
2084 
2085 out_end_trans:
2086 	gfs2_trans_end(sdp);
2087 out_rg_gunlock:
2088 	gfs2_glock_dq_m(rlist.rl_rgrps, rlist.rl_ghs);
2089 out_rlist:
2090 	gfs2_rlist_free(&rlist);
2091 	gfs2_quota_unhold(dip);
2092 out:
2093 	kvfree(ht);
2094 	return error;
2095 }
2096 
2097 /**
2098  * gfs2_dir_exhash_dealloc - free all the leaf blocks in a directory
2099  * @dip: the directory
2100  *
2101  * Dealloc all on-disk directory leaves to FREEMETA state
2102  * Change on-disk inode type to "regular file"
2103  *
2104  * Returns: errno
2105  */
2106 
2107 int gfs2_dir_exhash_dealloc(struct gfs2_inode *dip)
2108 {
2109 	struct buffer_head *bh;
2110 	struct gfs2_leaf *leaf;
2111 	u32 hsize, len;
2112 	u32 index = 0, next_index;
2113 	__be64 *lp;
2114 	u64 leaf_no;
2115 	int error = 0, last;
2116 
2117 	hsize = BIT(dip->i_depth);
2118 
2119 	lp = gfs2_dir_get_hash_table(dip);
2120 	if (IS_ERR(lp))
2121 		return PTR_ERR(lp);
2122 
2123 	while (index < hsize) {
2124 		leaf_no = be64_to_cpu(lp[index]);
2125 		if (leaf_no) {
2126 			error = get_leaf(dip, leaf_no, &bh);
2127 			if (error)
2128 				goto out;
2129 			leaf = (struct gfs2_leaf *)bh->b_data;
2130 			len = BIT(dip->i_depth - be16_to_cpu(leaf->lf_depth));
2131 
2132 			next_index = (index & ~(len - 1)) + len;
2133 			last = ((next_index >= hsize) ? 1 : 0);
2134 			error = leaf_dealloc(dip, index, len, leaf_no, bh,
2135 					     last);
2136 			brelse(bh);
2137 			if (error)
2138 				goto out;
2139 			index = next_index;
2140 		} else
2141 			index++;
2142 	}
2143 
2144 	if (index != hsize) {
2145 		gfs2_consist_inode(dip);
2146 		error = -EIO;
2147 	}
2148 
2149 out:
2150 
2151 	return error;
2152 }
2153 
2154 /**
2155  * gfs2_diradd_alloc_required - find if adding entry will require an allocation
2156  * @ip: the file being written to
2157  * @filname: the filename that's going to be added
2158  * @da: The structure to return dir alloc info
2159  *
2160  * Returns: 0 if ok, -ve on error
2161  */
2162 
2163 int gfs2_diradd_alloc_required(struct inode *inode, const struct qstr *name,
2164 			       struct gfs2_diradd *da)
2165 {
2166 	struct gfs2_inode *ip = GFS2_I(inode);
2167 	struct gfs2_sbd *sdp = GFS2_SB(inode);
2168 	const unsigned int extra = sizeof(struct gfs2_dinode) - sizeof(struct gfs2_leaf);
2169 	struct gfs2_dirent *dent;
2170 	struct buffer_head *bh;
2171 
2172 	da->nr_blocks = 0;
2173 	da->bh = NULL;
2174 	da->dent = NULL;
2175 
2176 	dent = gfs2_dirent_search(inode, name, gfs2_dirent_find_space, &bh);
2177 	if (!dent) {
2178 		da->nr_blocks = sdp->sd_max_dirres;
2179 		if (!(ip->i_diskflags & GFS2_DIF_EXHASH) &&
2180 		    (GFS2_DIRENT_SIZE(name->len) < extra))
2181 			da->nr_blocks = 1;
2182 		return 0;
2183 	}
2184 	if (IS_ERR(dent))
2185 		return PTR_ERR(dent);
2186 
2187 	if (da->save_loc) {
2188 		da->bh = bh;
2189 		da->dent = dent;
2190 	} else {
2191 		brelse(bh);
2192 	}
2193 	return 0;
2194 }
2195 
2196