xref: /openbmc/linux/fs/btrfs/extent_map.c (revision 64f26f745084872b916cd1bef6054e21b15c5784)
1d1310b2eSChris Mason #include <linux/err.h>
2a52d9a80SChris Mason #include <linux/gfp.h>
3d1310b2eSChris Mason #include <linux/slab.h>
4a52d9a80SChris Mason #include <linux/module.h>
5a52d9a80SChris Mason #include <linux/spinlock.h>
60a2118dfSJens Axboe #include <linux/version.h>
7d1310b2eSChris Mason #include <linux/hardirq.h>
8a52d9a80SChris Mason #include "extent_map.h"
9a52d9a80SChris Mason 
1086479a04SChris Mason /* temporary define until extent_map moves out of btrfs */
1186479a04SChris Mason struct kmem_cache *btrfs_cache_create(const char *name, size_t size,
1286479a04SChris Mason 				       unsigned long extra_flags,
1386479a04SChris Mason 				       void (*ctor)(void *, struct kmem_cache *,
1486479a04SChris Mason 						    unsigned long));
1586479a04SChris Mason 
16a52d9a80SChris Mason static struct kmem_cache *extent_map_cache;
17ca664626SChris Mason 
182f4cbe64SWyatt Banks int __init extent_map_init(void)
19a52d9a80SChris Mason {
2086479a04SChris Mason 	extent_map_cache = btrfs_cache_create("extent_map",
216d36dcd4SChris Mason 					    sizeof(struct extent_map), 0,
22a52d9a80SChris Mason 					    NULL);
232f4cbe64SWyatt Banks 	if (!extent_map_cache)
242f4cbe64SWyatt Banks 		return -ENOMEM;
252f4cbe64SWyatt Banks 	return 0;
26a52d9a80SChris Mason }
27a52d9a80SChris Mason 
2817636e03SChristian Hesse void extent_map_exit(void)
29a52d9a80SChris Mason {
30a52d9a80SChris Mason 	if (extent_map_cache)
31a52d9a80SChris Mason 		kmem_cache_destroy(extent_map_cache);
32a52d9a80SChris Mason }
33a52d9a80SChris Mason 
349d2423c5SChristoph Hellwig /**
359d2423c5SChristoph Hellwig  * extent_map_tree_init - initialize extent map tree
369d2423c5SChristoph Hellwig  * @tree:		tree to initialize
379d2423c5SChristoph Hellwig  * @mask:		flags for memory allocations during tree operations
389d2423c5SChristoph Hellwig  *
399d2423c5SChristoph Hellwig  * Initialize the extent tree @tree.  Should be called for each new inode
409d2423c5SChristoph Hellwig  * or other user of the extent_map interface.
419d2423c5SChristoph Hellwig  */
42d1310b2eSChris Mason void extent_map_tree_init(struct extent_map_tree *tree, gfp_t mask)
43a52d9a80SChris Mason {
44a52d9a80SChris Mason 	tree->map.rb_node = NULL;
45d1310b2eSChris Mason 	spin_lock_init(&tree->lock);
46a52d9a80SChris Mason }
47a52d9a80SChris Mason EXPORT_SYMBOL(extent_map_tree_init);
48a52d9a80SChris Mason 
499d2423c5SChristoph Hellwig /**
509d2423c5SChristoph Hellwig  * alloc_extent_map - allocate new extent map structure
519d2423c5SChristoph Hellwig  * @mask:	memory allocation flags
529d2423c5SChristoph Hellwig  *
539d2423c5SChristoph Hellwig  * Allocate a new extent_map structure.  The new structure is
549d2423c5SChristoph Hellwig  * returned with a reference count of one and needs to be
559d2423c5SChristoph Hellwig  * freed using free_extent_map()
569d2423c5SChristoph Hellwig  */
57a52d9a80SChris Mason struct extent_map *alloc_extent_map(gfp_t mask)
58a52d9a80SChris Mason {
59a52d9a80SChris Mason 	struct extent_map *em;
60a52d9a80SChris Mason 	em = kmem_cache_alloc(extent_map_cache, mask);
61a52d9a80SChris Mason 	if (!em || IS_ERR(em))
62a52d9a80SChris Mason 		return em;
63a52d9a80SChris Mason 	em->in_tree = 0;
64d1310b2eSChris Mason 	em->flags = 0;
65a52d9a80SChris Mason 	atomic_set(&em->refs, 1);
66a52d9a80SChris Mason 	return em;
67a52d9a80SChris Mason }
68a52d9a80SChris Mason EXPORT_SYMBOL(alloc_extent_map);
69a52d9a80SChris Mason 
709d2423c5SChristoph Hellwig /**
719d2423c5SChristoph Hellwig  * free_extent_map - drop reference count of an extent_map
729d2423c5SChristoph Hellwig  * @em:		extent map beeing releasead
739d2423c5SChristoph Hellwig  *
749d2423c5SChristoph Hellwig  * Drops the reference out on @em by one and free the structure
759d2423c5SChristoph Hellwig  * if the reference count hits zero.
769d2423c5SChristoph Hellwig  */
77a52d9a80SChris Mason void free_extent_map(struct extent_map *em)
78a52d9a80SChris Mason {
792bf5a725SChris Mason 	if (!em)
802bf5a725SChris Mason 		return;
81d1310b2eSChris Mason 	WARN_ON(atomic_read(&em->refs) == 0);
82a52d9a80SChris Mason 	if (atomic_dec_and_test(&em->refs)) {
83a52d9a80SChris Mason 		WARN_ON(em->in_tree);
84a52d9a80SChris Mason 		kmem_cache_free(extent_map_cache, em);
85a52d9a80SChris Mason 	}
86a52d9a80SChris Mason }
87a52d9a80SChris Mason EXPORT_SYMBOL(free_extent_map);
88a52d9a80SChris Mason 
89a52d9a80SChris Mason static struct rb_node *tree_insert(struct rb_root *root, u64 offset,
90a52d9a80SChris Mason 				   struct rb_node *node)
91a52d9a80SChris Mason {
92a52d9a80SChris Mason 	struct rb_node ** p = &root->rb_node;
93a52d9a80SChris Mason 	struct rb_node * parent = NULL;
94d1310b2eSChris Mason 	struct extent_map *entry;
95a52d9a80SChris Mason 
96a52d9a80SChris Mason 	while(*p) {
97a52d9a80SChris Mason 		parent = *p;
98d1310b2eSChris Mason 		entry = rb_entry(parent, struct extent_map, rb_node);
99d1310b2eSChris Mason 
100d1310b2eSChris Mason 		WARN_ON(!entry->in_tree);
101a52d9a80SChris Mason 
102a52d9a80SChris Mason 		if (offset < entry->start)
103a52d9a80SChris Mason 			p = &(*p)->rb_left;
104d1310b2eSChris Mason 		else if (offset >= extent_map_end(entry))
105a52d9a80SChris Mason 			p = &(*p)->rb_right;
106a52d9a80SChris Mason 		else
107a52d9a80SChris Mason 			return parent;
108a52d9a80SChris Mason 	}
109a52d9a80SChris Mason 
110d1310b2eSChris Mason 	entry = rb_entry(node, struct extent_map, rb_node);
111a52d9a80SChris Mason 	entry->in_tree = 1;
112a52d9a80SChris Mason 	rb_link_node(node, parent, p);
113a52d9a80SChris Mason 	rb_insert_color(node, root);
114a52d9a80SChris Mason 	return NULL;
115a52d9a80SChris Mason }
116a52d9a80SChris Mason 
117a52d9a80SChris Mason static struct rb_node *__tree_search(struct rb_root *root, u64 offset,
1185f56406aSChris Mason 				     struct rb_node **prev_ret,
1195f56406aSChris Mason 				     struct rb_node **next_ret)
120a52d9a80SChris Mason {
121a52d9a80SChris Mason 	struct rb_node * n = root->rb_node;
122a52d9a80SChris Mason 	struct rb_node *prev = NULL;
1235f56406aSChris Mason 	struct rb_node *orig_prev = NULL;
124d1310b2eSChris Mason 	struct extent_map *entry;
125d1310b2eSChris Mason 	struct extent_map *prev_entry = NULL;
126a52d9a80SChris Mason 
127a52d9a80SChris Mason 	while(n) {
128d1310b2eSChris Mason 		entry = rb_entry(n, struct extent_map, rb_node);
129a52d9a80SChris Mason 		prev = n;
130a52d9a80SChris Mason 		prev_entry = entry;
131a52d9a80SChris Mason 
132d1310b2eSChris Mason 		WARN_ON(!entry->in_tree);
133d1310b2eSChris Mason 
134a52d9a80SChris Mason 		if (offset < entry->start)
135a52d9a80SChris Mason 			n = n->rb_left;
136d1310b2eSChris Mason 		else if (offset >= extent_map_end(entry))
137a52d9a80SChris Mason 			n = n->rb_right;
138a52d9a80SChris Mason 		else
139a52d9a80SChris Mason 			return n;
140a52d9a80SChris Mason 	}
1415f56406aSChris Mason 
1425f56406aSChris Mason 	if (prev_ret) {
1435f56406aSChris Mason 		orig_prev = prev;
144d1310b2eSChris Mason 		while(prev && offset >= extent_map_end(prev_entry)) {
145a52d9a80SChris Mason 			prev = rb_next(prev);
146d1310b2eSChris Mason 			prev_entry = rb_entry(prev, struct extent_map, rb_node);
147a52d9a80SChris Mason 		}
148a52d9a80SChris Mason 		*prev_ret = prev;
1495f56406aSChris Mason 		prev = orig_prev;
1505f56406aSChris Mason 	}
1515f56406aSChris Mason 
1525f56406aSChris Mason 	if (next_ret) {
153d1310b2eSChris Mason 		prev_entry = rb_entry(prev, struct extent_map, rb_node);
1545f56406aSChris Mason 		while(prev && offset < prev_entry->start) {
1555f56406aSChris Mason 			prev = rb_prev(prev);
156d1310b2eSChris Mason 			prev_entry = rb_entry(prev, struct extent_map, rb_node);
1575f56406aSChris Mason 		}
1585f56406aSChris Mason 		*next_ret = prev;
1595f56406aSChris Mason 	}
160a52d9a80SChris Mason 	return NULL;
161a52d9a80SChris Mason }
162a52d9a80SChris Mason 
163a52d9a80SChris Mason static inline struct rb_node *tree_search(struct rb_root *root, u64 offset)
164a52d9a80SChris Mason {
165a52d9a80SChris Mason 	struct rb_node *prev;
166a52d9a80SChris Mason 	struct rb_node *ret;
1675f56406aSChris Mason 	ret = __tree_search(root, offset, &prev, NULL);
168a52d9a80SChris Mason 	if (!ret)
169a52d9a80SChris Mason 		return prev;
170a52d9a80SChris Mason 	return ret;
171a52d9a80SChris Mason }
172a52d9a80SChris Mason 
173d1310b2eSChris Mason static int mergable_maps(struct extent_map *prev, struct extent_map *next)
174a52d9a80SChris Mason {
1757f3c74fbSChris Mason 	if (test_bit(EXTENT_FLAG_PINNED, &prev->flags))
1767f3c74fbSChris Mason 		return 0;
1777f3c74fbSChris Mason 
178d1310b2eSChris Mason 	if (extent_map_end(prev) == next->start &&
179d1310b2eSChris Mason 	    prev->flags == next->flags &&
180d1310b2eSChris Mason 	    prev->bdev == next->bdev &&
181d1310b2eSChris Mason 	    ((next->block_start == EXTENT_MAP_HOLE &&
182d1310b2eSChris Mason 	      prev->block_start == EXTENT_MAP_HOLE) ||
183d1310b2eSChris Mason 	     (next->block_start == EXTENT_MAP_INLINE &&
184d1310b2eSChris Mason 	      prev->block_start == EXTENT_MAP_INLINE) ||
185d1310b2eSChris Mason 	     (next->block_start == EXTENT_MAP_DELALLOC &&
186d1310b2eSChris Mason 	      prev->block_start == EXTENT_MAP_DELALLOC) ||
187d1310b2eSChris Mason 	     (next->block_start < EXTENT_MAP_LAST_BYTE - 1 &&
188d1310b2eSChris Mason 	      next->block_start == extent_map_block_end(prev)))) {
189d1310b2eSChris Mason 		return 1;
190d1310b2eSChris Mason 	}
191a52d9a80SChris Mason 	return 0;
192a52d9a80SChris Mason }
193a52d9a80SChris Mason 
1949d2423c5SChristoph Hellwig /**
1959d2423c5SChristoph Hellwig  * add_extent_mapping - add new extent map to the extent tree
1969d2423c5SChristoph Hellwig  * @tree:	tree to insert new map in
1979d2423c5SChristoph Hellwig  * @em:		map to insert
1989d2423c5SChristoph Hellwig  *
1999d2423c5SChristoph Hellwig  * Insert @em into @tree or perform a simple forward/backward merge with
2009d2423c5SChristoph Hellwig  * existing mappings.  The extent_map struct passed in will be inserted
2019d2423c5SChristoph Hellwig  * into the tree directly, with an additional reference taken, or a
2029d2423c5SChristoph Hellwig  * reference dropped if the merge attempt was sucessfull.
203a52d9a80SChris Mason  */
204a52d9a80SChris Mason int add_extent_mapping(struct extent_map_tree *tree,
205a52d9a80SChris Mason 		       struct extent_map *em)
206a52d9a80SChris Mason {
207a52d9a80SChris Mason 	int ret = 0;
208d1310b2eSChris Mason 	struct extent_map *merge = NULL;
209a52d9a80SChris Mason 	struct rb_node *rb;
210a52d9a80SChris Mason 
211*64f26f74SDavid Woodhouse 	assert_spin_locked(&tree->lock);
212d1310b2eSChris Mason 	rb = tree_insert(&tree->map, em->start, &em->rb_node);
213a52d9a80SChris Mason 	if (rb) {
214a52d9a80SChris Mason 		ret = -EEXIST;
215e6dcd2dcSChris Mason 		free_extent_map(merge);
216a52d9a80SChris Mason 		goto out;
217a52d9a80SChris Mason 	}
218a52d9a80SChris Mason 	atomic_inc(&em->refs);
219a52d9a80SChris Mason 	if (em->start != 0) {
220a52d9a80SChris Mason 		rb = rb_prev(&em->rb_node);
221a52d9a80SChris Mason 		if (rb)
222d1310b2eSChris Mason 			merge = rb_entry(rb, struct extent_map, rb_node);
223d1310b2eSChris Mason 		if (rb && mergable_maps(merge, em)) {
224d1310b2eSChris Mason 			em->start = merge->start;
225d1310b2eSChris Mason 			em->len += merge->len;
226d1310b2eSChris Mason 			em->block_start = merge->block_start;
227d1310b2eSChris Mason 			merge->in_tree = 0;
228d1310b2eSChris Mason 			rb_erase(&merge->rb_node, &tree->map);
229d1310b2eSChris Mason 			free_extent_map(merge);
230a52d9a80SChris Mason 		}
231a52d9a80SChris Mason 	 }
232d1310b2eSChris Mason 	rb = rb_next(&em->rb_node);
233d1310b2eSChris Mason 	if (rb)
234d1310b2eSChris Mason 		merge = rb_entry(rb, struct extent_map, rb_node);
235d1310b2eSChris Mason 	if (rb && mergable_maps(em, merge)) {
236d1310b2eSChris Mason 		em->len += merge->len;
237d1310b2eSChris Mason 		rb_erase(&merge->rb_node, &tree->map);
238d1310b2eSChris Mason 		merge->in_tree = 0;
239d1310b2eSChris Mason 		free_extent_map(merge);
240d1310b2eSChris Mason 	}
241a52d9a80SChris Mason out:
242a52d9a80SChris Mason 	return ret;
243a52d9a80SChris Mason }
244a52d9a80SChris Mason EXPORT_SYMBOL(add_extent_mapping);
245a52d9a80SChris Mason 
246d1310b2eSChris Mason static u64 range_end(u64 start, u64 len)
247d1310b2eSChris Mason {
248d1310b2eSChris Mason 	if (start + len < start)
249d1310b2eSChris Mason 		return (u64)-1;
250d1310b2eSChris Mason 	return start + len;
251d1310b2eSChris Mason }
252d1310b2eSChris Mason 
2539d2423c5SChristoph Hellwig /**
2549d2423c5SChristoph Hellwig  * lookup_extent_mapping - lookup extent_map
2559d2423c5SChristoph Hellwig  * @tree:	tree to lookup in
2569d2423c5SChristoph Hellwig  * @start:	byte offset to start the search
2579d2423c5SChristoph Hellwig  * @len:	length of the lookup range
2589d2423c5SChristoph Hellwig  *
2599d2423c5SChristoph Hellwig  * Find and return the first extent_map struct in @tree that intersects the
2609d2423c5SChristoph Hellwig  * [start, len] range.  There may be additional objects in the tree that
2619d2423c5SChristoph Hellwig  * intersect, so check the object returned carefully to make sure that no
2629d2423c5SChristoph Hellwig  * additional lookups are needed.
263a52d9a80SChris Mason  */
264a52d9a80SChris Mason struct extent_map *lookup_extent_mapping(struct extent_map_tree *tree,
265d1310b2eSChris Mason 					 u64 start, u64 len)
266a52d9a80SChris Mason {
267a52d9a80SChris Mason 	struct extent_map *em;
268a52d9a80SChris Mason 	struct rb_node *rb_node;
269306929f3SChristoph Hellwig 	struct rb_node *prev = NULL;
270306929f3SChristoph Hellwig 	struct rb_node *next = NULL;
271306929f3SChristoph Hellwig 	u64 end = range_end(start, len);
272306929f3SChristoph Hellwig 
273*64f26f74SDavid Woodhouse 	assert_spin_locked(&tree->lock);
2745f56406aSChris Mason 	rb_node = __tree_search(&tree->map, start, &prev, &next);
2755f56406aSChris Mason 	if (!rb_node && prev) {
2765f56406aSChris Mason 		em = rb_entry(prev, struct extent_map, rb_node);
277d1310b2eSChris Mason 		if (end > em->start && start < extent_map_end(em))
2785f56406aSChris Mason 			goto found;
2795f56406aSChris Mason 	}
2805f56406aSChris Mason 	if (!rb_node && next) {
2815f56406aSChris Mason 		em = rb_entry(next, struct extent_map, rb_node);
282d1310b2eSChris Mason 		if (end > em->start && start < extent_map_end(em))
2835f56406aSChris Mason 			goto found;
2845f56406aSChris Mason 	}
285a52d9a80SChris Mason 	if (!rb_node) {
286a52d9a80SChris Mason 		em = NULL;
287a52d9a80SChris Mason 		goto out;
288a52d9a80SChris Mason 	}
289a52d9a80SChris Mason 	if (IS_ERR(rb_node)) {
290a52d9a80SChris Mason 		em = ERR_PTR(PTR_ERR(rb_node));
291a52d9a80SChris Mason 		goto out;
292a52d9a80SChris Mason 	}
293a52d9a80SChris Mason 	em = rb_entry(rb_node, struct extent_map, rb_node);
294d1310b2eSChris Mason 	if (end > em->start && start < extent_map_end(em))
295d1310b2eSChris Mason 		goto found;
296d1310b2eSChris Mason 
297a52d9a80SChris Mason 	em = NULL;
298a52d9a80SChris Mason 	goto out;
299d1310b2eSChris Mason 
3005f56406aSChris Mason found:
301a52d9a80SChris Mason 	atomic_inc(&em->refs);
302a52d9a80SChris Mason out:
303a52d9a80SChris Mason 	return em;
304a52d9a80SChris Mason }
305a52d9a80SChris Mason EXPORT_SYMBOL(lookup_extent_mapping);
306a52d9a80SChris Mason 
3079d2423c5SChristoph Hellwig /**
3089d2423c5SChristoph Hellwig  * remove_extent_mapping - removes an extent_map from the extent tree
3099d2423c5SChristoph Hellwig  * @tree:	extent tree to remove from
3109d2423c5SChristoph Hellwig  * @em:		extent map beeing removed
3119d2423c5SChristoph Hellwig  *
3129d2423c5SChristoph Hellwig  * Removes @em from @tree.  No reference counts are dropped, and no checks
3139d2423c5SChristoph Hellwig  * are done to see if the range is in use
314a52d9a80SChris Mason  */
315a52d9a80SChris Mason int remove_extent_mapping(struct extent_map_tree *tree, struct extent_map *em)
316a52d9a80SChris Mason {
317d1310b2eSChris Mason 	int ret = 0;
318a52d9a80SChris Mason 
3197f3c74fbSChris Mason 	WARN_ON(test_bit(EXTENT_FLAG_PINNED, &em->flags));
320*64f26f74SDavid Woodhouse 	assert_spin_locked(&tree->lock);
321d1310b2eSChris Mason 	rb_erase(&em->rb_node, &tree->map);
322d1310b2eSChris Mason 	em->in_tree = 0;
323a52d9a80SChris Mason 	return ret;
324a52d9a80SChris Mason }
325a52d9a80SChris Mason EXPORT_SYMBOL(remove_extent_mapping);
326