xref: /openbmc/linux/fs/ntfs3/bitmap.c (revision ae47ee5f)
1 // SPDX-License-Identifier: GPL-2.0
2 /*
3  *
4  * Copyright (C) 2019-2021 Paragon Software GmbH, All rights reserved.
5  *
6  * This code builds two trees of free clusters extents.
7  * Trees are sorted by start of extent and by length of extent.
8  * NTFS_MAX_WND_EXTENTS defines the maximum number of elements in trees.
9  * In extreme case code reads on-disk bitmap to find free clusters.
10  *
11  */
12 
13 #include <linux/blkdev.h>
14 #include <linux/buffer_head.h>
15 #include <linux/fs.h>
16 #include <linux/nls.h>
17 
18 #include "debug.h"
19 #include "ntfs.h"
20 #include "ntfs_fs.h"
21 
22 /*
23  * Maximum number of extents in tree.
24  */
25 #define NTFS_MAX_WND_EXTENTS (32u * 1024u)
26 
27 struct rb_node_key {
28 	struct rb_node node;
29 	size_t key;
30 };
31 
32 struct e_node {
33 	struct rb_node_key start; /* Tree sorted by start. */
34 	struct rb_node_key count; /* Tree sorted by len. */
35 };
36 
37 static int wnd_rescan(struct wnd_bitmap *wnd);
38 static struct buffer_head *wnd_map(struct wnd_bitmap *wnd, size_t iw);
39 static bool wnd_is_free_hlp(struct wnd_bitmap *wnd, size_t bit, size_t bits);
40 
41 static struct kmem_cache *ntfs_enode_cachep;
42 
43 int __init ntfs3_init_bitmap(void)
44 {
45 	ntfs_enode_cachep =
46 		kmem_cache_create("ntfs3_enode_cache", sizeof(struct e_node), 0,
47 				  SLAB_RECLAIM_ACCOUNT, NULL);
48 	return ntfs_enode_cachep ? 0 : -ENOMEM;
49 }
50 
51 void ntfs3_exit_bitmap(void)
52 {
53 	kmem_cache_destroy(ntfs_enode_cachep);
54 }
55 
56 static inline u32 wnd_bits(const struct wnd_bitmap *wnd, size_t i)
57 {
58 	return i + 1 == wnd->nwnd ? wnd->bits_last : wnd->sb->s_blocksize * 8;
59 }
60 
61 /*
62  * wnd_scan
63  *
64  * b_pos + b_len - biggest fragment.
65  * Scan range [wpos wbits) window @buf.
66  *
67  * Return: -1 if not found.
68  */
69 static size_t wnd_scan(const ulong *buf, size_t wbit, u32 wpos, u32 wend,
70 		       size_t to_alloc, size_t *prev_tail, size_t *b_pos,
71 		       size_t *b_len)
72 {
73 	while (wpos < wend) {
74 		size_t free_len;
75 		u32 free_bits, end;
76 		u32 used = find_next_zero_bit(buf, wend, wpos);
77 
78 		if (used >= wend) {
79 			if (*b_len < *prev_tail) {
80 				*b_pos = wbit - *prev_tail;
81 				*b_len = *prev_tail;
82 			}
83 
84 			*prev_tail = 0;
85 			return -1;
86 		}
87 
88 		if (used > wpos) {
89 			wpos = used;
90 			if (*b_len < *prev_tail) {
91 				*b_pos = wbit - *prev_tail;
92 				*b_len = *prev_tail;
93 			}
94 
95 			*prev_tail = 0;
96 		}
97 
98 		/*
99 		 * Now we have a fragment [wpos, wend) staring with 0.
100 		 */
101 		end = wpos + to_alloc - *prev_tail;
102 		free_bits = find_next_bit(buf, min(end, wend), wpos);
103 
104 		free_len = *prev_tail + free_bits - wpos;
105 
106 		if (*b_len < free_len) {
107 			*b_pos = wbit + wpos - *prev_tail;
108 			*b_len = free_len;
109 		}
110 
111 		if (free_len >= to_alloc)
112 			return wbit + wpos - *prev_tail;
113 
114 		if (free_bits >= wend) {
115 			*prev_tail += free_bits - wpos;
116 			return -1;
117 		}
118 
119 		wpos = free_bits + 1;
120 
121 		*prev_tail = 0;
122 	}
123 
124 	return -1;
125 }
126 
127 /*
128  * wnd_close - Frees all resources.
129  */
130 void wnd_close(struct wnd_bitmap *wnd)
131 {
132 	struct rb_node *node, *next;
133 
134 	kfree(wnd->free_bits);
135 	run_close(&wnd->run);
136 
137 	node = rb_first(&wnd->start_tree);
138 
139 	while (node) {
140 		next = rb_next(node);
141 		rb_erase(node, &wnd->start_tree);
142 		kmem_cache_free(ntfs_enode_cachep,
143 				rb_entry(node, struct e_node, start.node));
144 		node = next;
145 	}
146 }
147 
148 static struct rb_node *rb_lookup(struct rb_root *root, size_t v)
149 {
150 	struct rb_node **p = &root->rb_node;
151 	struct rb_node *r = NULL;
152 
153 	while (*p) {
154 		struct rb_node_key *k;
155 
156 		k = rb_entry(*p, struct rb_node_key, node);
157 		if (v < k->key) {
158 			p = &(*p)->rb_left;
159 		} else if (v > k->key) {
160 			r = &k->node;
161 			p = &(*p)->rb_right;
162 		} else {
163 			return &k->node;
164 		}
165 	}
166 
167 	return r;
168 }
169 
170 /*
171  * rb_insert_count - Helper function to insert special kind of 'count' tree.
172  */
173 static inline bool rb_insert_count(struct rb_root *root, struct e_node *e)
174 {
175 	struct rb_node **p = &root->rb_node;
176 	struct rb_node *parent = NULL;
177 	size_t e_ckey = e->count.key;
178 	size_t e_skey = e->start.key;
179 
180 	while (*p) {
181 		struct e_node *k =
182 			rb_entry(parent = *p, struct e_node, count.node);
183 
184 		if (e_ckey > k->count.key) {
185 			p = &(*p)->rb_left;
186 		} else if (e_ckey < k->count.key) {
187 			p = &(*p)->rb_right;
188 		} else if (e_skey < k->start.key) {
189 			p = &(*p)->rb_left;
190 		} else if (e_skey > k->start.key) {
191 			p = &(*p)->rb_right;
192 		} else {
193 			WARN_ON(1);
194 			return false;
195 		}
196 	}
197 
198 	rb_link_node(&e->count.node, parent, p);
199 	rb_insert_color(&e->count.node, root);
200 	return true;
201 }
202 
203 /*
204  * rb_insert_start - Helper function to insert special kind of 'count' tree.
205  */
206 static inline bool rb_insert_start(struct rb_root *root, struct e_node *e)
207 {
208 	struct rb_node **p = &root->rb_node;
209 	struct rb_node *parent = NULL;
210 	size_t e_skey = e->start.key;
211 
212 	while (*p) {
213 		struct e_node *k;
214 
215 		parent = *p;
216 
217 		k = rb_entry(parent, struct e_node, start.node);
218 		if (e_skey < k->start.key) {
219 			p = &(*p)->rb_left;
220 		} else if (e_skey > k->start.key) {
221 			p = &(*p)->rb_right;
222 		} else {
223 			WARN_ON(1);
224 			return false;
225 		}
226 	}
227 
228 	rb_link_node(&e->start.node, parent, p);
229 	rb_insert_color(&e->start.node, root);
230 	return true;
231 }
232 
233 /*
234  * wnd_add_free_ext - Adds a new extent of free space.
235  * @build:	1 when building tree.
236  */
237 static void wnd_add_free_ext(struct wnd_bitmap *wnd, size_t bit, size_t len,
238 			     bool build)
239 {
240 	struct e_node *e, *e0 = NULL;
241 	size_t ib, end_in = bit + len;
242 	struct rb_node *n;
243 
244 	if (build) {
245 		/* Use extent_min to filter too short extents. */
246 		if (wnd->count >= NTFS_MAX_WND_EXTENTS &&
247 		    len <= wnd->extent_min) {
248 			wnd->uptodated = -1;
249 			return;
250 		}
251 	} else {
252 		/* Try to find extent before 'bit'. */
253 		n = rb_lookup(&wnd->start_tree, bit);
254 
255 		if (!n) {
256 			n = rb_first(&wnd->start_tree);
257 		} else {
258 			e = rb_entry(n, struct e_node, start.node);
259 			n = rb_next(n);
260 			if (e->start.key + e->count.key == bit) {
261 				/* Remove left. */
262 				bit = e->start.key;
263 				len += e->count.key;
264 				rb_erase(&e->start.node, &wnd->start_tree);
265 				rb_erase(&e->count.node, &wnd->count_tree);
266 				wnd->count -= 1;
267 				e0 = e;
268 			}
269 		}
270 
271 		while (n) {
272 			size_t next_end;
273 
274 			e = rb_entry(n, struct e_node, start.node);
275 			next_end = e->start.key + e->count.key;
276 			if (e->start.key > end_in)
277 				break;
278 
279 			/* Remove right. */
280 			n = rb_next(n);
281 			len += next_end - end_in;
282 			end_in = next_end;
283 			rb_erase(&e->start.node, &wnd->start_tree);
284 			rb_erase(&e->count.node, &wnd->count_tree);
285 			wnd->count -= 1;
286 
287 			if (!e0)
288 				e0 = e;
289 			else
290 				kmem_cache_free(ntfs_enode_cachep, e);
291 		}
292 
293 		if (wnd->uptodated != 1) {
294 			/* Check bits before 'bit'. */
295 			ib = wnd->zone_bit == wnd->zone_end ||
296 					     bit < wnd->zone_end
297 				     ? 0
298 				     : wnd->zone_end;
299 
300 			while (bit > ib && wnd_is_free_hlp(wnd, bit - 1, 1)) {
301 				bit -= 1;
302 				len += 1;
303 			}
304 
305 			/* Check bits after 'end_in'. */
306 			ib = wnd->zone_bit == wnd->zone_end ||
307 					     end_in > wnd->zone_bit
308 				     ? wnd->nbits
309 				     : wnd->zone_bit;
310 
311 			while (end_in < ib && wnd_is_free_hlp(wnd, end_in, 1)) {
312 				end_in += 1;
313 				len += 1;
314 			}
315 		}
316 	}
317 	/* Insert new fragment. */
318 	if (wnd->count >= NTFS_MAX_WND_EXTENTS) {
319 		if (e0)
320 			kmem_cache_free(ntfs_enode_cachep, e0);
321 
322 		wnd->uptodated = -1;
323 
324 		/* Compare with smallest fragment. */
325 		n = rb_last(&wnd->count_tree);
326 		e = rb_entry(n, struct e_node, count.node);
327 		if (len <= e->count.key)
328 			goto out; /* Do not insert small fragments. */
329 
330 		if (build) {
331 			struct e_node *e2;
332 
333 			n = rb_prev(n);
334 			e2 = rb_entry(n, struct e_node, count.node);
335 			/* Smallest fragment will be 'e2->count.key'. */
336 			wnd->extent_min = e2->count.key;
337 		}
338 
339 		/* Replace smallest fragment by new one. */
340 		rb_erase(&e->start.node, &wnd->start_tree);
341 		rb_erase(&e->count.node, &wnd->count_tree);
342 		wnd->count -= 1;
343 	} else {
344 		e = e0 ? e0 : kmem_cache_alloc(ntfs_enode_cachep, GFP_ATOMIC);
345 		if (!e) {
346 			wnd->uptodated = -1;
347 			goto out;
348 		}
349 
350 		if (build && len <= wnd->extent_min)
351 			wnd->extent_min = len;
352 	}
353 	e->start.key = bit;
354 	e->count.key = len;
355 	if (len > wnd->extent_max)
356 		wnd->extent_max = len;
357 
358 	rb_insert_start(&wnd->start_tree, e);
359 	rb_insert_count(&wnd->count_tree, e);
360 	wnd->count += 1;
361 
362 out:;
363 }
364 
365 /*
366  * wnd_remove_free_ext - Remove a run from the cached free space.
367  */
368 static void wnd_remove_free_ext(struct wnd_bitmap *wnd, size_t bit, size_t len)
369 {
370 	struct rb_node *n, *n3;
371 	struct e_node *e, *e3;
372 	size_t end_in = bit + len;
373 	size_t end3, end, new_key, new_len, max_new_len;
374 
375 	/* Try to find extent before 'bit'. */
376 	n = rb_lookup(&wnd->start_tree, bit);
377 
378 	if (!n)
379 		return;
380 
381 	e = rb_entry(n, struct e_node, start.node);
382 	end = e->start.key + e->count.key;
383 
384 	new_key = new_len = 0;
385 	len = e->count.key;
386 
387 	/* Range [bit,end_in) must be inside 'e' or outside 'e' and 'n'. */
388 	if (e->start.key > bit)
389 		;
390 	else if (end_in <= end) {
391 		/* Range [bit,end_in) inside 'e'. */
392 		new_key = end_in;
393 		new_len = end - end_in;
394 		len = bit - e->start.key;
395 	} else if (bit > end) {
396 		bool bmax = false;
397 
398 		n3 = rb_next(n);
399 
400 		while (n3) {
401 			e3 = rb_entry(n3, struct e_node, start.node);
402 			if (e3->start.key >= end_in)
403 				break;
404 
405 			if (e3->count.key == wnd->extent_max)
406 				bmax = true;
407 
408 			end3 = e3->start.key + e3->count.key;
409 			if (end3 > end_in) {
410 				e3->start.key = end_in;
411 				rb_erase(&e3->count.node, &wnd->count_tree);
412 				e3->count.key = end3 - end_in;
413 				rb_insert_count(&wnd->count_tree, e3);
414 				break;
415 			}
416 
417 			n3 = rb_next(n3);
418 			rb_erase(&e3->start.node, &wnd->start_tree);
419 			rb_erase(&e3->count.node, &wnd->count_tree);
420 			wnd->count -= 1;
421 			kmem_cache_free(ntfs_enode_cachep, e3);
422 		}
423 		if (!bmax)
424 			return;
425 		n3 = rb_first(&wnd->count_tree);
426 		wnd->extent_max =
427 			n3 ? rb_entry(n3, struct e_node, count.node)->count.key
428 			   : 0;
429 		return;
430 	}
431 
432 	if (e->count.key != wnd->extent_max) {
433 		;
434 	} else if (rb_prev(&e->count.node)) {
435 		;
436 	} else {
437 		n3 = rb_next(&e->count.node);
438 		max_new_len = len > new_len ? len : new_len;
439 		if (!n3) {
440 			wnd->extent_max = max_new_len;
441 		} else {
442 			e3 = rb_entry(n3, struct e_node, count.node);
443 			wnd->extent_max = max(e3->count.key, max_new_len);
444 		}
445 	}
446 
447 	if (!len) {
448 		if (new_len) {
449 			e->start.key = new_key;
450 			rb_erase(&e->count.node, &wnd->count_tree);
451 			e->count.key = new_len;
452 			rb_insert_count(&wnd->count_tree, e);
453 		} else {
454 			rb_erase(&e->start.node, &wnd->start_tree);
455 			rb_erase(&e->count.node, &wnd->count_tree);
456 			wnd->count -= 1;
457 			kmem_cache_free(ntfs_enode_cachep, e);
458 		}
459 		goto out;
460 	}
461 	rb_erase(&e->count.node, &wnd->count_tree);
462 	e->count.key = len;
463 	rb_insert_count(&wnd->count_tree, e);
464 
465 	if (!new_len)
466 		goto out;
467 
468 	if (wnd->count >= NTFS_MAX_WND_EXTENTS) {
469 		wnd->uptodated = -1;
470 
471 		/* Get minimal extent. */
472 		e = rb_entry(rb_last(&wnd->count_tree), struct e_node,
473 			     count.node);
474 		if (e->count.key > new_len)
475 			goto out;
476 
477 		/* Replace minimum. */
478 		rb_erase(&e->start.node, &wnd->start_tree);
479 		rb_erase(&e->count.node, &wnd->count_tree);
480 		wnd->count -= 1;
481 	} else {
482 		e = kmem_cache_alloc(ntfs_enode_cachep, GFP_ATOMIC);
483 		if (!e)
484 			wnd->uptodated = -1;
485 	}
486 
487 	if (e) {
488 		e->start.key = new_key;
489 		e->count.key = new_len;
490 		rb_insert_start(&wnd->start_tree, e);
491 		rb_insert_count(&wnd->count_tree, e);
492 		wnd->count += 1;
493 	}
494 
495 out:
496 	if (!wnd->count && 1 != wnd->uptodated)
497 		wnd_rescan(wnd);
498 }
499 
500 /*
501  * wnd_rescan - Scan all bitmap. Used while initialization.
502  */
503 static int wnd_rescan(struct wnd_bitmap *wnd)
504 {
505 	int err = 0;
506 	size_t prev_tail = 0;
507 	struct super_block *sb = wnd->sb;
508 	struct ntfs_sb_info *sbi = sb->s_fs_info;
509 	u64 lbo, len = 0;
510 	u32 blocksize = sb->s_blocksize;
511 	u8 cluster_bits = sbi->cluster_bits;
512 	u32 wbits = 8 * sb->s_blocksize;
513 	u32 used, frb;
514 	const ulong *buf;
515 	size_t wpos, wbit, iw, vbo;
516 	struct buffer_head *bh = NULL;
517 	CLST lcn, clen;
518 
519 	wnd->uptodated = 0;
520 	wnd->extent_max = 0;
521 	wnd->extent_min = MINUS_ONE_T;
522 	wnd->total_zeroes = 0;
523 
524 	vbo = 0;
525 
526 	for (iw = 0; iw < wnd->nwnd; iw++) {
527 		if (iw + 1 == wnd->nwnd)
528 			wbits = wnd->bits_last;
529 
530 		if (wnd->inited) {
531 			if (!wnd->free_bits[iw]) {
532 				/* All ones. */
533 				if (prev_tail) {
534 					wnd_add_free_ext(wnd,
535 							 vbo * 8 - prev_tail,
536 							 prev_tail, true);
537 					prev_tail = 0;
538 				}
539 				goto next_wnd;
540 			}
541 			if (wbits == wnd->free_bits[iw]) {
542 				/* All zeroes. */
543 				prev_tail += wbits;
544 				wnd->total_zeroes += wbits;
545 				goto next_wnd;
546 			}
547 		}
548 
549 		if (!len) {
550 			u32 off = vbo & sbi->cluster_mask;
551 
552 			if (!run_lookup_entry(&wnd->run, vbo >> cluster_bits,
553 					      &lcn, &clen, NULL)) {
554 				err = -ENOENT;
555 				goto out;
556 			}
557 
558 			lbo = ((u64)lcn << cluster_bits) + off;
559 			len = ((u64)clen << cluster_bits) - off;
560 		}
561 
562 		bh = ntfs_bread(sb, lbo >> sb->s_blocksize_bits);
563 		if (!bh) {
564 			err = -EIO;
565 			goto out;
566 		}
567 
568 		buf = (ulong *)bh->b_data;
569 
570 		used = __bitmap_weight(buf, wbits);
571 		if (used < wbits) {
572 			frb = wbits - used;
573 			wnd->free_bits[iw] = frb;
574 			wnd->total_zeroes += frb;
575 		}
576 
577 		wpos = 0;
578 		wbit = vbo * 8;
579 
580 		if (wbit + wbits > wnd->nbits)
581 			wbits = wnd->nbits - wbit;
582 
583 		do {
584 			used = find_next_zero_bit(buf, wbits, wpos);
585 
586 			if (used > wpos && prev_tail) {
587 				wnd_add_free_ext(wnd, wbit + wpos - prev_tail,
588 						 prev_tail, true);
589 				prev_tail = 0;
590 			}
591 
592 			wpos = used;
593 
594 			if (wpos >= wbits) {
595 				/* No free blocks. */
596 				prev_tail = 0;
597 				break;
598 			}
599 
600 			frb = find_next_bit(buf, wbits, wpos);
601 			if (frb >= wbits) {
602 				/* Keep last free block. */
603 				prev_tail += frb - wpos;
604 				break;
605 			}
606 
607 			wnd_add_free_ext(wnd, wbit + wpos - prev_tail,
608 					 frb + prev_tail - wpos, true);
609 
610 			/* Skip free block and first '1'. */
611 			wpos = frb + 1;
612 			/* Reset previous tail. */
613 			prev_tail = 0;
614 		} while (wpos < wbits);
615 
616 next_wnd:
617 
618 		if (bh)
619 			put_bh(bh);
620 		bh = NULL;
621 
622 		vbo += blocksize;
623 		if (len) {
624 			len -= blocksize;
625 			lbo += blocksize;
626 		}
627 	}
628 
629 	/* Add last block. */
630 	if (prev_tail)
631 		wnd_add_free_ext(wnd, wnd->nbits - prev_tail, prev_tail, true);
632 
633 	/*
634 	 * Before init cycle wnd->uptodated was 0.
635 	 * If any errors or limits occurs while initialization then
636 	 * wnd->uptodated will be -1.
637 	 * If 'uptodated' is still 0 then Tree is really updated.
638 	 */
639 	if (!wnd->uptodated)
640 		wnd->uptodated = 1;
641 
642 	if (wnd->zone_bit != wnd->zone_end) {
643 		size_t zlen = wnd->zone_end - wnd->zone_bit;
644 
645 		wnd->zone_end = wnd->zone_bit;
646 		wnd_zone_set(wnd, wnd->zone_bit, zlen);
647 	}
648 
649 out:
650 	return err;
651 }
652 
653 int wnd_init(struct wnd_bitmap *wnd, struct super_block *sb, size_t nbits)
654 {
655 	int err;
656 	u32 blocksize = sb->s_blocksize;
657 	u32 wbits = blocksize * 8;
658 
659 	init_rwsem(&wnd->rw_lock);
660 
661 	wnd->sb = sb;
662 	wnd->nbits = nbits;
663 	wnd->total_zeroes = nbits;
664 	wnd->extent_max = MINUS_ONE_T;
665 	wnd->zone_bit = wnd->zone_end = 0;
666 	wnd->nwnd = bytes_to_block(sb, bitmap_size(nbits));
667 	wnd->bits_last = nbits & (wbits - 1);
668 	if (!wnd->bits_last)
669 		wnd->bits_last = wbits;
670 
671 	wnd->free_bits = kcalloc(wnd->nwnd, sizeof(u16), GFP_NOFS);
672 	if (!wnd->free_bits)
673 		return -ENOMEM;
674 
675 	err = wnd_rescan(wnd);
676 	if (err)
677 		return err;
678 
679 	wnd->inited = true;
680 
681 	return 0;
682 }
683 
684 /*
685  * wnd_map - Call sb_bread for requested window.
686  */
687 static struct buffer_head *wnd_map(struct wnd_bitmap *wnd, size_t iw)
688 {
689 	size_t vbo;
690 	CLST lcn, clen;
691 	struct super_block *sb = wnd->sb;
692 	struct ntfs_sb_info *sbi;
693 	struct buffer_head *bh;
694 	u64 lbo;
695 
696 	sbi = sb->s_fs_info;
697 	vbo = (u64)iw << sb->s_blocksize_bits;
698 
699 	if (!run_lookup_entry(&wnd->run, vbo >> sbi->cluster_bits, &lcn, &clen,
700 			      NULL)) {
701 		return ERR_PTR(-ENOENT);
702 	}
703 
704 	lbo = ((u64)lcn << sbi->cluster_bits) + (vbo & sbi->cluster_mask);
705 
706 	bh = ntfs_bread(wnd->sb, lbo >> sb->s_blocksize_bits);
707 	if (!bh)
708 		return ERR_PTR(-EIO);
709 
710 	return bh;
711 }
712 
713 /*
714  * wnd_set_free - Mark the bits range from bit to bit + bits as free.
715  */
716 int wnd_set_free(struct wnd_bitmap *wnd, size_t bit, size_t bits)
717 {
718 	int err = 0;
719 	struct super_block *sb = wnd->sb;
720 	size_t bits0 = bits;
721 	u32 wbits = 8 * sb->s_blocksize;
722 	size_t iw = bit >> (sb->s_blocksize_bits + 3);
723 	u32 wbit = bit & (wbits - 1);
724 	struct buffer_head *bh;
725 
726 	while (iw < wnd->nwnd && bits) {
727 		u32 tail, op;
728 		ulong *buf;
729 
730 		if (iw + 1 == wnd->nwnd)
731 			wbits = wnd->bits_last;
732 
733 		tail = wbits - wbit;
734 		op = tail < bits ? tail : bits;
735 
736 		bh = wnd_map(wnd, iw);
737 		if (IS_ERR(bh)) {
738 			err = PTR_ERR(bh);
739 			break;
740 		}
741 
742 		buf = (ulong *)bh->b_data;
743 
744 		lock_buffer(bh);
745 
746 		__bitmap_clear(buf, wbit, op);
747 
748 		wnd->free_bits[iw] += op;
749 
750 		set_buffer_uptodate(bh);
751 		mark_buffer_dirty(bh);
752 		unlock_buffer(bh);
753 		put_bh(bh);
754 
755 		wnd->total_zeroes += op;
756 		bits -= op;
757 		wbit = 0;
758 		iw += 1;
759 	}
760 
761 	wnd_add_free_ext(wnd, bit, bits0, false);
762 
763 	return err;
764 }
765 
766 /*
767  * wnd_set_used - Mark the bits range from bit to bit + bits as used.
768  */
769 int wnd_set_used(struct wnd_bitmap *wnd, size_t bit, size_t bits)
770 {
771 	int err = 0;
772 	struct super_block *sb = wnd->sb;
773 	size_t bits0 = bits;
774 	size_t iw = bit >> (sb->s_blocksize_bits + 3);
775 	u32 wbits = 8 * sb->s_blocksize;
776 	u32 wbit = bit & (wbits - 1);
777 	struct buffer_head *bh;
778 
779 	while (iw < wnd->nwnd && bits) {
780 		u32 tail, op;
781 		ulong *buf;
782 
783 		if (unlikely(iw + 1 == wnd->nwnd))
784 			wbits = wnd->bits_last;
785 
786 		tail = wbits - wbit;
787 		op = tail < bits ? tail : bits;
788 
789 		bh = wnd_map(wnd, iw);
790 		if (IS_ERR(bh)) {
791 			err = PTR_ERR(bh);
792 			break;
793 		}
794 		buf = (ulong *)bh->b_data;
795 
796 		lock_buffer(bh);
797 
798 		__bitmap_set(buf, wbit, op);
799 		wnd->free_bits[iw] -= op;
800 
801 		set_buffer_uptodate(bh);
802 		mark_buffer_dirty(bh);
803 		unlock_buffer(bh);
804 		put_bh(bh);
805 
806 		wnd->total_zeroes -= op;
807 		bits -= op;
808 		wbit = 0;
809 		iw += 1;
810 	}
811 
812 	if (!RB_EMPTY_ROOT(&wnd->start_tree))
813 		wnd_remove_free_ext(wnd, bit, bits0);
814 
815 	return err;
816 }
817 
818 /*
819  * wnd_is_free_hlp
820  *
821  * Return: True if all clusters [bit, bit+bits) are free (bitmap only).
822  */
823 static bool wnd_is_free_hlp(struct wnd_bitmap *wnd, size_t bit, size_t bits)
824 {
825 	struct super_block *sb = wnd->sb;
826 	size_t iw = bit >> (sb->s_blocksize_bits + 3);
827 	u32 wbits = 8 * sb->s_blocksize;
828 	u32 wbit = bit & (wbits - 1);
829 
830 	while (iw < wnd->nwnd && bits) {
831 		u32 tail, op;
832 
833 		if (unlikely(iw + 1 == wnd->nwnd))
834 			wbits = wnd->bits_last;
835 
836 		tail = wbits - wbit;
837 		op = tail < bits ? tail : bits;
838 
839 		if (wbits != wnd->free_bits[iw]) {
840 			bool ret;
841 			struct buffer_head *bh = wnd_map(wnd, iw);
842 
843 			if (IS_ERR(bh))
844 				return false;
845 
846 			ret = are_bits_clear((ulong *)bh->b_data, wbit, op);
847 
848 			put_bh(bh);
849 			if (!ret)
850 				return false;
851 		}
852 
853 		bits -= op;
854 		wbit = 0;
855 		iw += 1;
856 	}
857 
858 	return true;
859 }
860 
861 /*
862  * wnd_is_free
863  *
864  * Return: True if all clusters [bit, bit+bits) are free.
865  */
866 bool wnd_is_free(struct wnd_bitmap *wnd, size_t bit, size_t bits)
867 {
868 	bool ret;
869 	struct rb_node *n;
870 	size_t end;
871 	struct e_node *e;
872 
873 	if (RB_EMPTY_ROOT(&wnd->start_tree))
874 		goto use_wnd;
875 
876 	n = rb_lookup(&wnd->start_tree, bit);
877 	if (!n)
878 		goto use_wnd;
879 
880 	e = rb_entry(n, struct e_node, start.node);
881 
882 	end = e->start.key + e->count.key;
883 
884 	if (bit < end && bit + bits <= end)
885 		return true;
886 
887 use_wnd:
888 	ret = wnd_is_free_hlp(wnd, bit, bits);
889 
890 	return ret;
891 }
892 
893 /*
894  * wnd_is_used
895  *
896  * Return: True if all clusters [bit, bit+bits) are used.
897  */
898 bool wnd_is_used(struct wnd_bitmap *wnd, size_t bit, size_t bits)
899 {
900 	bool ret = false;
901 	struct super_block *sb = wnd->sb;
902 	size_t iw = bit >> (sb->s_blocksize_bits + 3);
903 	u32 wbits = 8 * sb->s_blocksize;
904 	u32 wbit = bit & (wbits - 1);
905 	size_t end;
906 	struct rb_node *n;
907 	struct e_node *e;
908 
909 	if (RB_EMPTY_ROOT(&wnd->start_tree))
910 		goto use_wnd;
911 
912 	end = bit + bits;
913 	n = rb_lookup(&wnd->start_tree, end - 1);
914 	if (!n)
915 		goto use_wnd;
916 
917 	e = rb_entry(n, struct e_node, start.node);
918 	if (e->start.key + e->count.key > bit)
919 		return false;
920 
921 use_wnd:
922 	while (iw < wnd->nwnd && bits) {
923 		u32 tail, op;
924 
925 		if (unlikely(iw + 1 == wnd->nwnd))
926 			wbits = wnd->bits_last;
927 
928 		tail = wbits - wbit;
929 		op = tail < bits ? tail : bits;
930 
931 		if (wnd->free_bits[iw]) {
932 			bool ret;
933 			struct buffer_head *bh = wnd_map(wnd, iw);
934 
935 			if (IS_ERR(bh))
936 				goto out;
937 
938 			ret = are_bits_set((ulong *)bh->b_data, wbit, op);
939 			put_bh(bh);
940 			if (!ret)
941 				goto out;
942 		}
943 
944 		bits -= op;
945 		wbit = 0;
946 		iw += 1;
947 	}
948 	ret = true;
949 
950 out:
951 	return ret;
952 }
953 
954 /*
955  * wnd_find - Look for free space.
956  *
957  * - flags - BITMAP_FIND_XXX flags
958  *
959  * Return: 0 if not found.
960  */
961 size_t wnd_find(struct wnd_bitmap *wnd, size_t to_alloc, size_t hint,
962 		size_t flags, size_t *allocated)
963 {
964 	struct super_block *sb;
965 	u32 wbits, wpos, wzbit, wzend;
966 	size_t fnd, max_alloc, b_len, b_pos;
967 	size_t iw, prev_tail, nwnd, wbit, ebit, zbit, zend;
968 	size_t to_alloc0 = to_alloc;
969 	const ulong *buf;
970 	const struct e_node *e;
971 	const struct rb_node *pr, *cr;
972 	u8 log2_bits;
973 	bool fbits_valid;
974 	struct buffer_head *bh;
975 
976 	/* Fast checking for available free space. */
977 	if (flags & BITMAP_FIND_FULL) {
978 		size_t zeroes = wnd_zeroes(wnd);
979 
980 		zeroes -= wnd->zone_end - wnd->zone_bit;
981 		if (zeroes < to_alloc0)
982 			goto no_space;
983 
984 		if (to_alloc0 > wnd->extent_max)
985 			goto no_space;
986 	} else {
987 		if (to_alloc > wnd->extent_max)
988 			to_alloc = wnd->extent_max;
989 	}
990 
991 	if (wnd->zone_bit <= hint && hint < wnd->zone_end)
992 		hint = wnd->zone_end;
993 
994 	max_alloc = wnd->nbits;
995 	b_len = b_pos = 0;
996 
997 	if (hint >= max_alloc)
998 		hint = 0;
999 
1000 	if (RB_EMPTY_ROOT(&wnd->start_tree)) {
1001 		if (wnd->uptodated == 1) {
1002 			/* Extents tree is updated -> No free space. */
1003 			goto no_space;
1004 		}
1005 		goto scan_bitmap;
1006 	}
1007 
1008 	e = NULL;
1009 	if (!hint)
1010 		goto allocate_biggest;
1011 
1012 	/* Use hint: Enumerate extents by start >= hint. */
1013 	pr = NULL;
1014 	cr = wnd->start_tree.rb_node;
1015 
1016 	for (;;) {
1017 		e = rb_entry(cr, struct e_node, start.node);
1018 
1019 		if (e->start.key == hint)
1020 			break;
1021 
1022 		if (e->start.key < hint) {
1023 			pr = cr;
1024 			cr = cr->rb_right;
1025 			if (!cr)
1026 				break;
1027 			continue;
1028 		}
1029 
1030 		cr = cr->rb_left;
1031 		if (!cr) {
1032 			e = pr ? rb_entry(pr, struct e_node, start.node) : NULL;
1033 			break;
1034 		}
1035 	}
1036 
1037 	if (!e)
1038 		goto allocate_biggest;
1039 
1040 	if (e->start.key + e->count.key > hint) {
1041 		/* We have found extension with 'hint' inside. */
1042 		size_t len = e->start.key + e->count.key - hint;
1043 
1044 		if (len >= to_alloc && hint + to_alloc <= max_alloc) {
1045 			fnd = hint;
1046 			goto found;
1047 		}
1048 
1049 		if (!(flags & BITMAP_FIND_FULL)) {
1050 			if (len > to_alloc)
1051 				len = to_alloc;
1052 
1053 			if (hint + len <= max_alloc) {
1054 				fnd = hint;
1055 				to_alloc = len;
1056 				goto found;
1057 			}
1058 		}
1059 	}
1060 
1061 allocate_biggest:
1062 	/* Allocate from biggest free extent. */
1063 	e = rb_entry(rb_first(&wnd->count_tree), struct e_node, count.node);
1064 	if (e->count.key != wnd->extent_max)
1065 		wnd->extent_max = e->count.key;
1066 
1067 	if (e->count.key < max_alloc) {
1068 		if (e->count.key >= to_alloc) {
1069 			;
1070 		} else if (flags & BITMAP_FIND_FULL) {
1071 			if (e->count.key < to_alloc0) {
1072 				/* Biggest free block is less then requested. */
1073 				goto no_space;
1074 			}
1075 			to_alloc = e->count.key;
1076 		} else if (-1 != wnd->uptodated) {
1077 			to_alloc = e->count.key;
1078 		} else {
1079 			/* Check if we can use more bits. */
1080 			size_t op, max_check;
1081 			struct rb_root start_tree;
1082 
1083 			memcpy(&start_tree, &wnd->start_tree,
1084 			       sizeof(struct rb_root));
1085 			memset(&wnd->start_tree, 0, sizeof(struct rb_root));
1086 
1087 			max_check = e->start.key + to_alloc;
1088 			if (max_check > max_alloc)
1089 				max_check = max_alloc;
1090 			for (op = e->start.key + e->count.key; op < max_check;
1091 			     op++) {
1092 				if (!wnd_is_free(wnd, op, 1))
1093 					break;
1094 			}
1095 			memcpy(&wnd->start_tree, &start_tree,
1096 			       sizeof(struct rb_root));
1097 			to_alloc = op - e->start.key;
1098 		}
1099 
1100 		/* Prepare to return. */
1101 		fnd = e->start.key;
1102 		if (e->start.key + to_alloc > max_alloc)
1103 			to_alloc = max_alloc - e->start.key;
1104 		goto found;
1105 	}
1106 
1107 	if (wnd->uptodated == 1) {
1108 		/* Extents tree is updated -> no free space. */
1109 		goto no_space;
1110 	}
1111 
1112 	b_len = e->count.key;
1113 	b_pos = e->start.key;
1114 
1115 scan_bitmap:
1116 	sb = wnd->sb;
1117 	log2_bits = sb->s_blocksize_bits + 3;
1118 
1119 	/* At most two ranges [hint, max_alloc) + [0, hint). */
1120 Again:
1121 
1122 	/* TODO: Optimize request for case nbits > wbits. */
1123 	iw = hint >> log2_bits;
1124 	wbits = sb->s_blocksize * 8;
1125 	wpos = hint & (wbits - 1);
1126 	prev_tail = 0;
1127 	fbits_valid = true;
1128 
1129 	if (max_alloc == wnd->nbits) {
1130 		nwnd = wnd->nwnd;
1131 	} else {
1132 		size_t t = max_alloc + wbits - 1;
1133 
1134 		nwnd = likely(t > max_alloc) ? (t >> log2_bits) : wnd->nwnd;
1135 	}
1136 
1137 	/* Enumerate all windows. */
1138 	for (; iw < nwnd; iw++) {
1139 		wbit = iw << log2_bits;
1140 
1141 		if (!wnd->free_bits[iw]) {
1142 			if (prev_tail > b_len) {
1143 				b_pos = wbit - prev_tail;
1144 				b_len = prev_tail;
1145 			}
1146 
1147 			/* Skip full used window. */
1148 			prev_tail = 0;
1149 			wpos = 0;
1150 			continue;
1151 		}
1152 
1153 		if (unlikely(iw + 1 == nwnd)) {
1154 			if (max_alloc == wnd->nbits) {
1155 				wbits = wnd->bits_last;
1156 			} else {
1157 				size_t t = max_alloc & (wbits - 1);
1158 
1159 				if (t) {
1160 					wbits = t;
1161 					fbits_valid = false;
1162 				}
1163 			}
1164 		}
1165 
1166 		if (wnd->zone_end > wnd->zone_bit) {
1167 			ebit = wbit + wbits;
1168 			zbit = max(wnd->zone_bit, wbit);
1169 			zend = min(wnd->zone_end, ebit);
1170 
1171 			/* Here we have a window [wbit, ebit) and zone [zbit, zend). */
1172 			if (zend <= zbit) {
1173 				/* Zone does not overlap window. */
1174 			} else {
1175 				wzbit = zbit - wbit;
1176 				wzend = zend - wbit;
1177 
1178 				/* Zone overlaps window. */
1179 				if (wnd->free_bits[iw] == wzend - wzbit) {
1180 					prev_tail = 0;
1181 					wpos = 0;
1182 					continue;
1183 				}
1184 
1185 				/* Scan two ranges window: [wbit, zbit) and [zend, ebit). */
1186 				bh = wnd_map(wnd, iw);
1187 
1188 				if (IS_ERR(bh)) {
1189 					/* TODO: Error */
1190 					prev_tail = 0;
1191 					wpos = 0;
1192 					continue;
1193 				}
1194 
1195 				buf = (ulong *)bh->b_data;
1196 
1197 				/* Scan range [wbit, zbit). */
1198 				if (wpos < wzbit) {
1199 					/* Scan range [wpos, zbit). */
1200 					fnd = wnd_scan(buf, wbit, wpos, wzbit,
1201 						       to_alloc, &prev_tail,
1202 						       &b_pos, &b_len);
1203 					if (fnd != MINUS_ONE_T) {
1204 						put_bh(bh);
1205 						goto found;
1206 					}
1207 				}
1208 
1209 				prev_tail = 0;
1210 
1211 				/* Scan range [zend, ebit). */
1212 				if (wzend < wbits) {
1213 					fnd = wnd_scan(buf, wbit,
1214 						       max(wzend, wpos), wbits,
1215 						       to_alloc, &prev_tail,
1216 						       &b_pos, &b_len);
1217 					if (fnd != MINUS_ONE_T) {
1218 						put_bh(bh);
1219 						goto found;
1220 					}
1221 				}
1222 
1223 				wpos = 0;
1224 				put_bh(bh);
1225 				continue;
1226 			}
1227 		}
1228 
1229 		/* Current window does not overlap zone. */
1230 		if (!wpos && fbits_valid && wnd->free_bits[iw] == wbits) {
1231 			/* Window is empty. */
1232 			if (prev_tail + wbits >= to_alloc) {
1233 				fnd = wbit + wpos - prev_tail;
1234 				goto found;
1235 			}
1236 
1237 			/* Increase 'prev_tail' and process next window. */
1238 			prev_tail += wbits;
1239 			wpos = 0;
1240 			continue;
1241 		}
1242 
1243 		/* Read window. */
1244 		bh = wnd_map(wnd, iw);
1245 		if (IS_ERR(bh)) {
1246 			// TODO: Error.
1247 			prev_tail = 0;
1248 			wpos = 0;
1249 			continue;
1250 		}
1251 
1252 		buf = (ulong *)bh->b_data;
1253 
1254 		/* Scan range [wpos, eBits). */
1255 		fnd = wnd_scan(buf, wbit, wpos, wbits, to_alloc, &prev_tail,
1256 			       &b_pos, &b_len);
1257 		put_bh(bh);
1258 		if (fnd != MINUS_ONE_T)
1259 			goto found;
1260 	}
1261 
1262 	if (b_len < prev_tail) {
1263 		/* The last fragment. */
1264 		b_len = prev_tail;
1265 		b_pos = max_alloc - prev_tail;
1266 	}
1267 
1268 	if (hint) {
1269 		/*
1270 		 * We have scanned range [hint max_alloc).
1271 		 * Prepare to scan range [0 hint + to_alloc).
1272 		 */
1273 		size_t nextmax = hint + to_alloc;
1274 
1275 		if (likely(nextmax >= hint) && nextmax < max_alloc)
1276 			max_alloc = nextmax;
1277 		hint = 0;
1278 		goto Again;
1279 	}
1280 
1281 	if (!b_len)
1282 		goto no_space;
1283 
1284 	wnd->extent_max = b_len;
1285 
1286 	if (flags & BITMAP_FIND_FULL)
1287 		goto no_space;
1288 
1289 	fnd = b_pos;
1290 	to_alloc = b_len;
1291 
1292 found:
1293 	if (flags & BITMAP_FIND_MARK_AS_USED) {
1294 		/* TODO: Optimize remove extent (pass 'e'?). */
1295 		if (wnd_set_used(wnd, fnd, to_alloc))
1296 			goto no_space;
1297 	} else if (wnd->extent_max != MINUS_ONE_T &&
1298 		   to_alloc > wnd->extent_max) {
1299 		wnd->extent_max = to_alloc;
1300 	}
1301 
1302 	*allocated = fnd;
1303 	return to_alloc;
1304 
1305 no_space:
1306 	return 0;
1307 }
1308 
1309 /*
1310  * wnd_extend - Extend bitmap ($MFT bitmap).
1311  */
1312 int wnd_extend(struct wnd_bitmap *wnd, size_t new_bits)
1313 {
1314 	int err;
1315 	struct super_block *sb = wnd->sb;
1316 	struct ntfs_sb_info *sbi = sb->s_fs_info;
1317 	u32 blocksize = sb->s_blocksize;
1318 	u32 wbits = blocksize * 8;
1319 	u32 b0, new_last;
1320 	size_t bits, iw, new_wnd;
1321 	size_t old_bits = wnd->nbits;
1322 	u16 *new_free;
1323 
1324 	if (new_bits <= old_bits)
1325 		return -EINVAL;
1326 
1327 	/* Align to 8 byte boundary. */
1328 	new_wnd = bytes_to_block(sb, bitmap_size(new_bits));
1329 	new_last = new_bits & (wbits - 1);
1330 	if (!new_last)
1331 		new_last = wbits;
1332 
1333 	if (new_wnd != wnd->nwnd) {
1334 		new_free = kmalloc(new_wnd * sizeof(u16), GFP_NOFS);
1335 		if (!new_free)
1336 			return -ENOMEM;
1337 
1338 		if (new_free != wnd->free_bits)
1339 			memcpy(new_free, wnd->free_bits,
1340 			       wnd->nwnd * sizeof(short));
1341 		memset(new_free + wnd->nwnd, 0,
1342 		       (new_wnd - wnd->nwnd) * sizeof(short));
1343 		kfree(wnd->free_bits);
1344 		wnd->free_bits = new_free;
1345 	}
1346 
1347 	/* Zero bits [old_bits,new_bits). */
1348 	bits = new_bits - old_bits;
1349 	b0 = old_bits & (wbits - 1);
1350 
1351 	for (iw = old_bits >> (sb->s_blocksize_bits + 3); bits; iw += 1) {
1352 		u32 op;
1353 		size_t frb;
1354 		u64 vbo, lbo, bytes;
1355 		struct buffer_head *bh;
1356 		ulong *buf;
1357 
1358 		if (iw + 1 == new_wnd)
1359 			wbits = new_last;
1360 
1361 		op = b0 + bits > wbits ? wbits - b0 : bits;
1362 		vbo = (u64)iw * blocksize;
1363 
1364 		err = ntfs_vbo_to_lbo(sbi, &wnd->run, vbo, &lbo, &bytes);
1365 		if (err)
1366 			break;
1367 
1368 		bh = ntfs_bread(sb, lbo >> sb->s_blocksize_bits);
1369 		if (!bh)
1370 			return -EIO;
1371 
1372 		lock_buffer(bh);
1373 		buf = (ulong *)bh->b_data;
1374 
1375 		__bitmap_clear(buf, b0, blocksize * 8 - b0);
1376 		frb = wbits - __bitmap_weight(buf, wbits);
1377 		wnd->total_zeroes += frb - wnd->free_bits[iw];
1378 		wnd->free_bits[iw] = frb;
1379 
1380 		set_buffer_uptodate(bh);
1381 		mark_buffer_dirty(bh);
1382 		unlock_buffer(bh);
1383 		/* err = sync_dirty_buffer(bh); */
1384 
1385 		b0 = 0;
1386 		bits -= op;
1387 	}
1388 
1389 	wnd->nbits = new_bits;
1390 	wnd->nwnd = new_wnd;
1391 	wnd->bits_last = new_last;
1392 
1393 	wnd_add_free_ext(wnd, old_bits, new_bits - old_bits, false);
1394 
1395 	return 0;
1396 }
1397 
1398 void wnd_zone_set(struct wnd_bitmap *wnd, size_t lcn, size_t len)
1399 {
1400 	size_t zlen;
1401 
1402 	zlen = wnd->zone_end - wnd->zone_bit;
1403 	if (zlen)
1404 		wnd_add_free_ext(wnd, wnd->zone_bit, zlen, false);
1405 
1406 	if (!RB_EMPTY_ROOT(&wnd->start_tree) && len)
1407 		wnd_remove_free_ext(wnd, lcn, len);
1408 
1409 	wnd->zone_bit = lcn;
1410 	wnd->zone_end = lcn + len;
1411 }
1412 
1413 int ntfs_trim_fs(struct ntfs_sb_info *sbi, struct fstrim_range *range)
1414 {
1415 	int err = 0;
1416 	struct super_block *sb = sbi->sb;
1417 	struct wnd_bitmap *wnd = &sbi->used.bitmap;
1418 	u32 wbits = 8 * sb->s_blocksize;
1419 	CLST len = 0, lcn = 0, done = 0;
1420 	CLST minlen = bytes_to_cluster(sbi, range->minlen);
1421 	CLST lcn_from = bytes_to_cluster(sbi, range->start);
1422 	size_t iw = lcn_from >> (sb->s_blocksize_bits + 3);
1423 	u32 wbit = lcn_from & (wbits - 1);
1424 	const ulong *buf;
1425 	CLST lcn_to;
1426 
1427 	if (!minlen)
1428 		minlen = 1;
1429 
1430 	if (range->len == (u64)-1)
1431 		lcn_to = wnd->nbits;
1432 	else
1433 		lcn_to = bytes_to_cluster(sbi, range->start + range->len);
1434 
1435 	down_read_nested(&wnd->rw_lock, BITMAP_MUTEX_CLUSTERS);
1436 
1437 	for (; iw < wnd->nbits; iw++, wbit = 0) {
1438 		CLST lcn_wnd = iw * wbits;
1439 		struct buffer_head *bh;
1440 
1441 		if (lcn_wnd > lcn_to)
1442 			break;
1443 
1444 		if (!wnd->free_bits[iw])
1445 			continue;
1446 
1447 		if (iw + 1 == wnd->nwnd)
1448 			wbits = wnd->bits_last;
1449 
1450 		if (lcn_wnd + wbits > lcn_to)
1451 			wbits = lcn_to - lcn_wnd;
1452 
1453 		bh = wnd_map(wnd, iw);
1454 		if (IS_ERR(bh)) {
1455 			err = PTR_ERR(bh);
1456 			break;
1457 		}
1458 
1459 		buf = (ulong *)bh->b_data;
1460 
1461 		for (; wbit < wbits; wbit++) {
1462 			if (!test_bit(wbit, buf)) {
1463 				if (!len)
1464 					lcn = lcn_wnd + wbit;
1465 				len += 1;
1466 				continue;
1467 			}
1468 			if (len >= minlen) {
1469 				err = ntfs_discard(sbi, lcn, len);
1470 				if (err)
1471 					goto out;
1472 				done += len;
1473 			}
1474 			len = 0;
1475 		}
1476 		put_bh(bh);
1477 	}
1478 
1479 	/* Process the last fragment. */
1480 	if (len >= minlen) {
1481 		err = ntfs_discard(sbi, lcn, len);
1482 		if (err)
1483 			goto out;
1484 		done += len;
1485 	}
1486 
1487 out:
1488 	range->len = (u64)done << sbi->cluster_bits;
1489 
1490 	up_read(&wnd->rw_lock);
1491 
1492 	return err;
1493 }
1494