xref: /openbmc/linux/fs/btrfs/extent_map.c (revision c8b978188c9a0fd3d535c13debd19d522b726f1f)
1d1310b2eSChris Mason #include <linux/err.h>
2a52d9a80SChris Mason #include <linux/gfp.h>
3d1310b2eSChris Mason #include <linux/slab.h>
4a52d9a80SChris Mason #include <linux/module.h>
5a52d9a80SChris Mason #include <linux/spinlock.h>
60a2118dfSJens Axboe #include <linux/version.h>
7d1310b2eSChris Mason #include <linux/hardirq.h>
8a52d9a80SChris Mason #include "extent_map.h"
9a52d9a80SChris Mason 
1086479a04SChris Mason /* temporary define until extent_map moves out of btrfs */
1186479a04SChris Mason struct kmem_cache *btrfs_cache_create(const char *name, size_t size,
1286479a04SChris Mason 				       unsigned long extra_flags,
1386479a04SChris Mason 				       void (*ctor)(void *, struct kmem_cache *,
1486479a04SChris Mason 						    unsigned long));
1586479a04SChris Mason 
16a52d9a80SChris Mason static struct kmem_cache *extent_map_cache;
17ca664626SChris Mason 
182f4cbe64SWyatt Banks int __init extent_map_init(void)
19a52d9a80SChris Mason {
2086479a04SChris Mason 	extent_map_cache = btrfs_cache_create("extent_map",
216d36dcd4SChris Mason 					    sizeof(struct extent_map), 0,
22a52d9a80SChris Mason 					    NULL);
232f4cbe64SWyatt Banks 	if (!extent_map_cache)
242f4cbe64SWyatt Banks 		return -ENOMEM;
252f4cbe64SWyatt Banks 	return 0;
26a52d9a80SChris Mason }
27a52d9a80SChris Mason 
2817636e03SChristian Hesse void extent_map_exit(void)
29a52d9a80SChris Mason {
30a52d9a80SChris Mason 	if (extent_map_cache)
31a52d9a80SChris Mason 		kmem_cache_destroy(extent_map_cache);
32a52d9a80SChris Mason }
33a52d9a80SChris Mason 
349d2423c5SChristoph Hellwig /**
359d2423c5SChristoph Hellwig  * extent_map_tree_init - initialize extent map tree
369d2423c5SChristoph Hellwig  * @tree:		tree to initialize
379d2423c5SChristoph Hellwig  * @mask:		flags for memory allocations during tree operations
389d2423c5SChristoph Hellwig  *
399d2423c5SChristoph Hellwig  * Initialize the extent tree @tree.  Should be called for each new inode
409d2423c5SChristoph Hellwig  * or other user of the extent_map interface.
419d2423c5SChristoph Hellwig  */
42d1310b2eSChris Mason void extent_map_tree_init(struct extent_map_tree *tree, gfp_t mask)
43a52d9a80SChris Mason {
44a52d9a80SChris Mason 	tree->map.rb_node = NULL;
45d1310b2eSChris Mason 	spin_lock_init(&tree->lock);
46a52d9a80SChris Mason }
47a52d9a80SChris Mason EXPORT_SYMBOL(extent_map_tree_init);
48a52d9a80SChris Mason 
499d2423c5SChristoph Hellwig /**
509d2423c5SChristoph Hellwig  * alloc_extent_map - allocate new extent map structure
519d2423c5SChristoph Hellwig  * @mask:	memory allocation flags
529d2423c5SChristoph Hellwig  *
539d2423c5SChristoph Hellwig  * Allocate a new extent_map structure.  The new structure is
549d2423c5SChristoph Hellwig  * returned with a reference count of one and needs to be
559d2423c5SChristoph Hellwig  * freed using free_extent_map()
569d2423c5SChristoph Hellwig  */
57a52d9a80SChris Mason struct extent_map *alloc_extent_map(gfp_t mask)
58a52d9a80SChris Mason {
59a52d9a80SChris Mason 	struct extent_map *em;
60a52d9a80SChris Mason 	em = kmem_cache_alloc(extent_map_cache, mask);
61a52d9a80SChris Mason 	if (!em || IS_ERR(em))
62a52d9a80SChris Mason 		return em;
63a52d9a80SChris Mason 	em->in_tree = 0;
64d1310b2eSChris Mason 	em->flags = 0;
65a52d9a80SChris Mason 	atomic_set(&em->refs, 1);
66a52d9a80SChris Mason 	return em;
67a52d9a80SChris Mason }
68a52d9a80SChris Mason EXPORT_SYMBOL(alloc_extent_map);
69a52d9a80SChris Mason 
709d2423c5SChristoph Hellwig /**
719d2423c5SChristoph Hellwig  * free_extent_map - drop reference count of an extent_map
729d2423c5SChristoph Hellwig  * @em:		extent map beeing releasead
739d2423c5SChristoph Hellwig  *
749d2423c5SChristoph Hellwig  * Drops the reference out on @em by one and free the structure
759d2423c5SChristoph Hellwig  * if the reference count hits zero.
769d2423c5SChristoph Hellwig  */
77a52d9a80SChris Mason void free_extent_map(struct extent_map *em)
78a52d9a80SChris Mason {
792bf5a725SChris Mason 	if (!em)
802bf5a725SChris Mason 		return;
81d1310b2eSChris Mason 	WARN_ON(atomic_read(&em->refs) == 0);
82a52d9a80SChris Mason 	if (atomic_dec_and_test(&em->refs)) {
83a52d9a80SChris Mason 		WARN_ON(em->in_tree);
84a52d9a80SChris Mason 		kmem_cache_free(extent_map_cache, em);
85a52d9a80SChris Mason 	}
86a52d9a80SChris Mason }
87a52d9a80SChris Mason EXPORT_SYMBOL(free_extent_map);
88a52d9a80SChris Mason 
89a52d9a80SChris Mason static struct rb_node *tree_insert(struct rb_root *root, u64 offset,
90a52d9a80SChris Mason 				   struct rb_node *node)
91a52d9a80SChris Mason {
92a52d9a80SChris Mason 	struct rb_node ** p = &root->rb_node;
93a52d9a80SChris Mason 	struct rb_node * parent = NULL;
94d1310b2eSChris Mason 	struct extent_map *entry;
95a52d9a80SChris Mason 
96a52d9a80SChris Mason 	while(*p) {
97a52d9a80SChris Mason 		parent = *p;
98d1310b2eSChris Mason 		entry = rb_entry(parent, struct extent_map, rb_node);
99d1310b2eSChris Mason 
100d1310b2eSChris Mason 		WARN_ON(!entry->in_tree);
101a52d9a80SChris Mason 
102a52d9a80SChris Mason 		if (offset < entry->start)
103a52d9a80SChris Mason 			p = &(*p)->rb_left;
104d1310b2eSChris Mason 		else if (offset >= extent_map_end(entry))
105a52d9a80SChris Mason 			p = &(*p)->rb_right;
106a52d9a80SChris Mason 		else
107a52d9a80SChris Mason 			return parent;
108a52d9a80SChris Mason 	}
109a52d9a80SChris Mason 
110d1310b2eSChris Mason 	entry = rb_entry(node, struct extent_map, rb_node);
111a52d9a80SChris Mason 	entry->in_tree = 1;
112a52d9a80SChris Mason 	rb_link_node(node, parent, p);
113a52d9a80SChris Mason 	rb_insert_color(node, root);
114a52d9a80SChris Mason 	return NULL;
115a52d9a80SChris Mason }
116a52d9a80SChris Mason 
117d352ac68SChris Mason /*
118d352ac68SChris Mason  * search through the tree for an extent_map with a given offset.  If
119d352ac68SChris Mason  * it can't be found, try to find some neighboring extents
120d352ac68SChris Mason  */
121a52d9a80SChris Mason static struct rb_node *__tree_search(struct rb_root *root, u64 offset,
1225f56406aSChris Mason 				     struct rb_node **prev_ret,
1235f56406aSChris Mason 				     struct rb_node **next_ret)
124a52d9a80SChris Mason {
125a52d9a80SChris Mason 	struct rb_node * n = root->rb_node;
126a52d9a80SChris Mason 	struct rb_node *prev = NULL;
1275f56406aSChris Mason 	struct rb_node *orig_prev = NULL;
128d1310b2eSChris Mason 	struct extent_map *entry;
129d1310b2eSChris Mason 	struct extent_map *prev_entry = NULL;
130a52d9a80SChris Mason 
131a52d9a80SChris Mason 	while(n) {
132d1310b2eSChris Mason 		entry = rb_entry(n, struct extent_map, rb_node);
133a52d9a80SChris Mason 		prev = n;
134a52d9a80SChris Mason 		prev_entry = entry;
135a52d9a80SChris Mason 
136d1310b2eSChris Mason 		WARN_ON(!entry->in_tree);
137d1310b2eSChris Mason 
138a52d9a80SChris Mason 		if (offset < entry->start)
139a52d9a80SChris Mason 			n = n->rb_left;
140d1310b2eSChris Mason 		else if (offset >= extent_map_end(entry))
141a52d9a80SChris Mason 			n = n->rb_right;
142a52d9a80SChris Mason 		else
143a52d9a80SChris Mason 			return n;
144a52d9a80SChris Mason 	}
1455f56406aSChris Mason 
1465f56406aSChris Mason 	if (prev_ret) {
1475f56406aSChris Mason 		orig_prev = prev;
148d1310b2eSChris Mason 		while(prev && offset >= extent_map_end(prev_entry)) {
149a52d9a80SChris Mason 			prev = rb_next(prev);
150d1310b2eSChris Mason 			prev_entry = rb_entry(prev, struct extent_map, rb_node);
151a52d9a80SChris Mason 		}
152a52d9a80SChris Mason 		*prev_ret = prev;
1535f56406aSChris Mason 		prev = orig_prev;
1545f56406aSChris Mason 	}
1555f56406aSChris Mason 
1565f56406aSChris Mason 	if (next_ret) {
157d1310b2eSChris Mason 		prev_entry = rb_entry(prev, struct extent_map, rb_node);
1585f56406aSChris Mason 		while(prev && offset < prev_entry->start) {
1595f56406aSChris Mason 			prev = rb_prev(prev);
160d1310b2eSChris Mason 			prev_entry = rb_entry(prev, struct extent_map, rb_node);
1615f56406aSChris Mason 		}
1625f56406aSChris Mason 		*next_ret = prev;
1635f56406aSChris Mason 	}
164a52d9a80SChris Mason 	return NULL;
165a52d9a80SChris Mason }
166a52d9a80SChris Mason 
167d352ac68SChris Mason /*
168d352ac68SChris Mason  * look for an offset in the tree, and if it can't be found, return
169d352ac68SChris Mason  * the first offset we can find smaller than 'offset'.
170d352ac68SChris Mason  */
171a52d9a80SChris Mason static inline struct rb_node *tree_search(struct rb_root *root, u64 offset)
172a52d9a80SChris Mason {
173a52d9a80SChris Mason 	struct rb_node *prev;
174a52d9a80SChris Mason 	struct rb_node *ret;
1755f56406aSChris Mason 	ret = __tree_search(root, offset, &prev, NULL);
176a52d9a80SChris Mason 	if (!ret)
177a52d9a80SChris Mason 		return prev;
178a52d9a80SChris Mason 	return ret;
179a52d9a80SChris Mason }
180a52d9a80SChris Mason 
181d352ac68SChris Mason /* check to see if two extent_map structs are adjacent and safe to merge */
182d1310b2eSChris Mason static int mergable_maps(struct extent_map *prev, struct extent_map *next)
183a52d9a80SChris Mason {
1847f3c74fbSChris Mason 	if (test_bit(EXTENT_FLAG_PINNED, &prev->flags))
1857f3c74fbSChris Mason 		return 0;
1867f3c74fbSChris Mason 
187*c8b97818SChris Mason 	/*
188*c8b97818SChris Mason 	 * don't merge compressed extents, we need to know their
189*c8b97818SChris Mason 	 * actual size
190*c8b97818SChris Mason 	 */
191*c8b97818SChris Mason 	if (test_bit(EXTENT_FLAG_COMPRESSED, &prev->flags))
192*c8b97818SChris Mason 		return 0;
193*c8b97818SChris Mason 
194d1310b2eSChris Mason 	if (extent_map_end(prev) == next->start &&
195d1310b2eSChris Mason 	    prev->flags == next->flags &&
196d1310b2eSChris Mason 	    prev->bdev == next->bdev &&
197d1310b2eSChris Mason 	    ((next->block_start == EXTENT_MAP_HOLE &&
198d1310b2eSChris Mason 	      prev->block_start == EXTENT_MAP_HOLE) ||
199d1310b2eSChris Mason 	     (next->block_start == EXTENT_MAP_INLINE &&
200d1310b2eSChris Mason 	      prev->block_start == EXTENT_MAP_INLINE) ||
201d1310b2eSChris Mason 	     (next->block_start == EXTENT_MAP_DELALLOC &&
202d1310b2eSChris Mason 	      prev->block_start == EXTENT_MAP_DELALLOC) ||
203d1310b2eSChris Mason 	     (next->block_start < EXTENT_MAP_LAST_BYTE - 1 &&
204d1310b2eSChris Mason 	      next->block_start == extent_map_block_end(prev)))) {
205d1310b2eSChris Mason 		return 1;
206d1310b2eSChris Mason 	}
207a52d9a80SChris Mason 	return 0;
208a52d9a80SChris Mason }
209a52d9a80SChris Mason 
2109d2423c5SChristoph Hellwig /**
2119d2423c5SChristoph Hellwig  * add_extent_mapping - add new extent map to the extent tree
2129d2423c5SChristoph Hellwig  * @tree:	tree to insert new map in
2139d2423c5SChristoph Hellwig  * @em:		map to insert
2149d2423c5SChristoph Hellwig  *
2159d2423c5SChristoph Hellwig  * Insert @em into @tree or perform a simple forward/backward merge with
2169d2423c5SChristoph Hellwig  * existing mappings.  The extent_map struct passed in will be inserted
2179d2423c5SChristoph Hellwig  * into the tree directly, with an additional reference taken, or a
2189d2423c5SChristoph Hellwig  * reference dropped if the merge attempt was sucessfull.
219a52d9a80SChris Mason  */
220a52d9a80SChris Mason int add_extent_mapping(struct extent_map_tree *tree,
221a52d9a80SChris Mason 		       struct extent_map *em)
222a52d9a80SChris Mason {
223a52d9a80SChris Mason 	int ret = 0;
224d1310b2eSChris Mason 	struct extent_map *merge = NULL;
225a52d9a80SChris Mason 	struct rb_node *rb;
2267c2fe32aSChris Mason 	struct extent_map *exist;
227a52d9a80SChris Mason 
2287c2fe32aSChris Mason 	exist = lookup_extent_mapping(tree, em->start, em->len);
2297c2fe32aSChris Mason 	if (exist) {
2307c2fe32aSChris Mason 		free_extent_map(exist);
2317c2fe32aSChris Mason 		ret = -EEXIST;
2327c2fe32aSChris Mason 		goto out;
2337c2fe32aSChris Mason 	}
23464f26f74SDavid Woodhouse 	assert_spin_locked(&tree->lock);
235d1310b2eSChris Mason 	rb = tree_insert(&tree->map, em->start, &em->rb_node);
236a52d9a80SChris Mason 	if (rb) {
237a52d9a80SChris Mason 		ret = -EEXIST;
238e6dcd2dcSChris Mason 		free_extent_map(merge);
239a52d9a80SChris Mason 		goto out;
240a52d9a80SChris Mason 	}
241a52d9a80SChris Mason 	atomic_inc(&em->refs);
242a52d9a80SChris Mason 	if (em->start != 0) {
243a52d9a80SChris Mason 		rb = rb_prev(&em->rb_node);
244a52d9a80SChris Mason 		if (rb)
245d1310b2eSChris Mason 			merge = rb_entry(rb, struct extent_map, rb_node);
246d1310b2eSChris Mason 		if (rb && mergable_maps(merge, em)) {
247d1310b2eSChris Mason 			em->start = merge->start;
248d1310b2eSChris Mason 			em->len += merge->len;
249*c8b97818SChris Mason 			em->block_len += merge->block_len;
250d1310b2eSChris Mason 			em->block_start = merge->block_start;
251d1310b2eSChris Mason 			merge->in_tree = 0;
252d1310b2eSChris Mason 			rb_erase(&merge->rb_node, &tree->map);
253d1310b2eSChris Mason 			free_extent_map(merge);
254a52d9a80SChris Mason 		}
255a52d9a80SChris Mason 	 }
256d1310b2eSChris Mason 	rb = rb_next(&em->rb_node);
257d1310b2eSChris Mason 	if (rb)
258d1310b2eSChris Mason 		merge = rb_entry(rb, struct extent_map, rb_node);
259d1310b2eSChris Mason 	if (rb && mergable_maps(em, merge)) {
260d1310b2eSChris Mason 		em->len += merge->len;
261*c8b97818SChris Mason 		em->block_len += merge->len;
262d1310b2eSChris Mason 		rb_erase(&merge->rb_node, &tree->map);
263d1310b2eSChris Mason 		merge->in_tree = 0;
264d1310b2eSChris Mason 		free_extent_map(merge);
265d1310b2eSChris Mason 	}
266a52d9a80SChris Mason out:
267a52d9a80SChris Mason 	return ret;
268a52d9a80SChris Mason }
269a52d9a80SChris Mason EXPORT_SYMBOL(add_extent_mapping);
270a52d9a80SChris Mason 
271d352ac68SChris Mason /* simple helper to do math around the end of an extent, handling wrap */
272d1310b2eSChris Mason static u64 range_end(u64 start, u64 len)
273d1310b2eSChris Mason {
274d1310b2eSChris Mason 	if (start + len < start)
275d1310b2eSChris Mason 		return (u64)-1;
276d1310b2eSChris Mason 	return start + len;
277d1310b2eSChris Mason }
278d1310b2eSChris Mason 
2799d2423c5SChristoph Hellwig /**
2809d2423c5SChristoph Hellwig  * lookup_extent_mapping - lookup extent_map
2819d2423c5SChristoph Hellwig  * @tree:	tree to lookup in
2829d2423c5SChristoph Hellwig  * @start:	byte offset to start the search
2839d2423c5SChristoph Hellwig  * @len:	length of the lookup range
2849d2423c5SChristoph Hellwig  *
2859d2423c5SChristoph Hellwig  * Find and return the first extent_map struct in @tree that intersects the
2869d2423c5SChristoph Hellwig  * [start, len] range.  There may be additional objects in the tree that
2879d2423c5SChristoph Hellwig  * intersect, so check the object returned carefully to make sure that no
2889d2423c5SChristoph Hellwig  * additional lookups are needed.
289a52d9a80SChris Mason  */
290a52d9a80SChris Mason struct extent_map *lookup_extent_mapping(struct extent_map_tree *tree,
291d1310b2eSChris Mason 					 u64 start, u64 len)
292a52d9a80SChris Mason {
293a52d9a80SChris Mason 	struct extent_map *em;
294a52d9a80SChris Mason 	struct rb_node *rb_node;
295306929f3SChristoph Hellwig 	struct rb_node *prev = NULL;
296306929f3SChristoph Hellwig 	struct rb_node *next = NULL;
297306929f3SChristoph Hellwig 	u64 end = range_end(start, len);
298306929f3SChristoph Hellwig 
29964f26f74SDavid Woodhouse 	assert_spin_locked(&tree->lock);
3005f56406aSChris Mason 	rb_node = __tree_search(&tree->map, start, &prev, &next);
3015f56406aSChris Mason 	if (!rb_node && prev) {
3025f56406aSChris Mason 		em = rb_entry(prev, struct extent_map, rb_node);
303d1310b2eSChris Mason 		if (end > em->start && start < extent_map_end(em))
3045f56406aSChris Mason 			goto found;
3055f56406aSChris Mason 	}
3065f56406aSChris Mason 	if (!rb_node && next) {
3075f56406aSChris Mason 		em = rb_entry(next, struct extent_map, rb_node);
308d1310b2eSChris Mason 		if (end > em->start && start < extent_map_end(em))
3095f56406aSChris Mason 			goto found;
3105f56406aSChris Mason 	}
311a52d9a80SChris Mason 	if (!rb_node) {
312a52d9a80SChris Mason 		em = NULL;
313a52d9a80SChris Mason 		goto out;
314a52d9a80SChris Mason 	}
315a52d9a80SChris Mason 	if (IS_ERR(rb_node)) {
316a52d9a80SChris Mason 		em = ERR_PTR(PTR_ERR(rb_node));
317a52d9a80SChris Mason 		goto out;
318a52d9a80SChris Mason 	}
319a52d9a80SChris Mason 	em = rb_entry(rb_node, struct extent_map, rb_node);
320d1310b2eSChris Mason 	if (end > em->start && start < extent_map_end(em))
321d1310b2eSChris Mason 		goto found;
322d1310b2eSChris Mason 
323a52d9a80SChris Mason 	em = NULL;
324a52d9a80SChris Mason 	goto out;
325d1310b2eSChris Mason 
3265f56406aSChris Mason found:
327a52d9a80SChris Mason 	atomic_inc(&em->refs);
328a52d9a80SChris Mason out:
329a52d9a80SChris Mason 	return em;
330a52d9a80SChris Mason }
331a52d9a80SChris Mason EXPORT_SYMBOL(lookup_extent_mapping);
332a52d9a80SChris Mason 
3339d2423c5SChristoph Hellwig /**
3349d2423c5SChristoph Hellwig  * remove_extent_mapping - removes an extent_map from the extent tree
3359d2423c5SChristoph Hellwig  * @tree:	extent tree to remove from
3369d2423c5SChristoph Hellwig  * @em:		extent map beeing removed
3379d2423c5SChristoph Hellwig  *
3389d2423c5SChristoph Hellwig  * Removes @em from @tree.  No reference counts are dropped, and no checks
3399d2423c5SChristoph Hellwig  * are done to see if the range is in use
340a52d9a80SChris Mason  */
341a52d9a80SChris Mason int remove_extent_mapping(struct extent_map_tree *tree, struct extent_map *em)
342a52d9a80SChris Mason {
343d1310b2eSChris Mason 	int ret = 0;
344a52d9a80SChris Mason 
3457f3c74fbSChris Mason 	WARN_ON(test_bit(EXTENT_FLAG_PINNED, &em->flags));
34664f26f74SDavid Woodhouse 	assert_spin_locked(&tree->lock);
347d1310b2eSChris Mason 	rb_erase(&em->rb_node, &tree->map);
348d1310b2eSChris Mason 	em->in_tree = 0;
349a52d9a80SChris Mason 	return ret;
350a52d9a80SChris Mason }
351a52d9a80SChris Mason EXPORT_SYMBOL(remove_extent_mapping);
352