Merge tag 'xfs-6.4-rc1-fixes' of git://git.kernel.org/pub/scm/fs/xfs/xfs-linux
[linux-block.git] / fs / ntfs3 / bitmap.c
CommitLineData
3f3b442b
KK
1// SPDX-License-Identifier: GPL-2.0
2/*
3 *
4 * Copyright (C) 2019-2021 Paragon Software GmbH, All rights reserved.
5 *
6 * This code builds two trees of free clusters extents.
7 * Trees are sorted by start of extent and by length of extent.
8 * NTFS_MAX_WND_EXTENTS defines the maximum number of elements in trees.
e8b8e97f 9 * In extreme case code reads on-disk bitmap to find free clusters.
3f3b442b
KK
10 *
11 */
12
3f3b442b
KK
13#include <linux/buffer_head.h>
14#include <linux/fs.h>
6e3331ee 15#include <linux/kernel.h>
3f3b442b 16
3f3b442b
KK
17#include "ntfs.h"
18#include "ntfs_fs.h"
19
20/*
21 * Maximum number of extents in tree.
22 */
23#define NTFS_MAX_WND_EXTENTS (32u * 1024u)
24
25struct rb_node_key {
26 struct rb_node node;
27 size_t key;
28};
29
3f3b442b 30struct e_node {
e8b8e97f
KA
31 struct rb_node_key start; /* Tree sorted by start. */
32 struct rb_node_key count; /* Tree sorted by len. */
3f3b442b
KK
33};
34
35static int wnd_rescan(struct wnd_bitmap *wnd);
36static struct buffer_head *wnd_map(struct wnd_bitmap *wnd, size_t iw);
37static bool wnd_is_free_hlp(struct wnd_bitmap *wnd, size_t bit, size_t bits);
38
39static struct kmem_cache *ntfs_enode_cachep;
40
41int __init ntfs3_init_bitmap(void)
42{
96de65a9
KK
43 ntfs_enode_cachep = kmem_cache_create("ntfs3_enode_cache",
44 sizeof(struct e_node), 0,
45 SLAB_RECLAIM_ACCOUNT, NULL);
3f3b442b
KK
46 return ntfs_enode_cachep ? 0 : -ENOMEM;
47}
48
49void ntfs3_exit_bitmap(void)
50{
51 kmem_cache_destroy(ntfs_enode_cachep);
52}
53
3f3b442b 54/*
e8b8e97f
KA
55 * wnd_scan
56 *
57 * b_pos + b_len - biggest fragment.
58 * Scan range [wpos wbits) window @buf.
59 *
60 * Return: -1 if not found.
3f3b442b 61 */
08811ba5 62static size_t wnd_scan(const void *buf, size_t wbit, u32 wpos, u32 wend,
3f3b442b
KK
63 size_t to_alloc, size_t *prev_tail, size_t *b_pos,
64 size_t *b_len)
65{
66 while (wpos < wend) {
67 size_t free_len;
68 u32 free_bits, end;
095d8ce6 69 u32 used = find_next_zero_bit_le(buf, wend, wpos);
3f3b442b
KK
70
71 if (used >= wend) {
72 if (*b_len < *prev_tail) {
73 *b_pos = wbit - *prev_tail;
74 *b_len = *prev_tail;
75 }
76
77 *prev_tail = 0;
78 return -1;
79 }
80
81 if (used > wpos) {
82 wpos = used;
83 if (*b_len < *prev_tail) {
84 *b_pos = wbit - *prev_tail;
85 *b_len = *prev_tail;
86 }
87
88 *prev_tail = 0;
89 }
90
91 /*
e8b8e97f 92 * Now we have a fragment [wpos, wend) staring with 0.
3f3b442b
KK
93 */
94 end = wpos + to_alloc - *prev_tail;
095d8ce6 95 free_bits = find_next_bit_le(buf, min(end, wend), wpos);
3f3b442b
KK
96
97 free_len = *prev_tail + free_bits - wpos;
98
99 if (*b_len < free_len) {
100 *b_pos = wbit + wpos - *prev_tail;
101 *b_len = free_len;
102 }
103
104 if (free_len >= to_alloc)
105 return wbit + wpos - *prev_tail;
106
107 if (free_bits >= wend) {
108 *prev_tail += free_bits - wpos;
109 return -1;
110 }
111
112 wpos = free_bits + 1;
113
114 *prev_tail = 0;
115 }
116
117 return -1;
118}
119
120/*
e8b8e97f 121 * wnd_close - Frees all resources.
3f3b442b
KK
122 */
123void wnd_close(struct wnd_bitmap *wnd)
124{
125 struct rb_node *node, *next;
126
195c52bd 127 kfree(wnd->free_bits);
3f3b442b
KK
128 run_close(&wnd->run);
129
130 node = rb_first(&wnd->start_tree);
131
132 while (node) {
133 next = rb_next(node);
134 rb_erase(node, &wnd->start_tree);
135 kmem_cache_free(ntfs_enode_cachep,
136 rb_entry(node, struct e_node, start.node));
137 node = next;
138 }
139}
140
141static struct rb_node *rb_lookup(struct rb_root *root, size_t v)
142{
143 struct rb_node **p = &root->rb_node;
144 struct rb_node *r = NULL;
145
146 while (*p) {
147 struct rb_node_key *k;
148
149 k = rb_entry(*p, struct rb_node_key, node);
150 if (v < k->key) {
151 p = &(*p)->rb_left;
152 } else if (v > k->key) {
153 r = &k->node;
154 p = &(*p)->rb_right;
155 } else {
156 return &k->node;
157 }
158 }
159
160 return r;
161}
162
163/*
e8b8e97f 164 * rb_insert_count - Helper function to insert special kind of 'count' tree.
3f3b442b
KK
165 */
166static inline bool rb_insert_count(struct rb_root *root, struct e_node *e)
167{
168 struct rb_node **p = &root->rb_node;
169 struct rb_node *parent = NULL;
170 size_t e_ckey = e->count.key;
171 size_t e_skey = e->start.key;
172
173 while (*p) {
174 struct e_node *k =
175 rb_entry(parent = *p, struct e_node, count.node);
176
177 if (e_ckey > k->count.key) {
178 p = &(*p)->rb_left;
179 } else if (e_ckey < k->count.key) {
180 p = &(*p)->rb_right;
181 } else if (e_skey < k->start.key) {
182 p = &(*p)->rb_left;
183 } else if (e_skey > k->start.key) {
184 p = &(*p)->rb_right;
185 } else {
186 WARN_ON(1);
187 return false;
188 }
189 }
190
191 rb_link_node(&e->count.node, parent, p);
192 rb_insert_color(&e->count.node, root);
193 return true;
194}
195
196/*
e8b8e97f 197 * rb_insert_start - Helper function to insert special kind of 'count' tree.
3f3b442b
KK
198 */
199static inline bool rb_insert_start(struct rb_root *root, struct e_node *e)
200{
201 struct rb_node **p = &root->rb_node;
202 struct rb_node *parent = NULL;
203 size_t e_skey = e->start.key;
204
205 while (*p) {
206 struct e_node *k;
207
208 parent = *p;
209
210 k = rb_entry(parent, struct e_node, start.node);
211 if (e_skey < k->start.key) {
212 p = &(*p)->rb_left;
213 } else if (e_skey > k->start.key) {
214 p = &(*p)->rb_right;
215 } else {
216 WARN_ON(1);
217 return false;
218 }
219 }
220
221 rb_link_node(&e->start.node, parent, p);
222 rb_insert_color(&e->start.node, root);
223 return true;
224}
225
226/*
e8b8e97f
KA
227 * wnd_add_free_ext - Adds a new extent of free space.
228 * @build: 1 when building tree.
3f3b442b
KK
229 */
230static void wnd_add_free_ext(struct wnd_bitmap *wnd, size_t bit, size_t len,
231 bool build)
232{
233 struct e_node *e, *e0 = NULL;
234 size_t ib, end_in = bit + len;
235 struct rb_node *n;
236
237 if (build) {
e8b8e97f 238 /* Use extent_min to filter too short extents. */
3f3b442b
KK
239 if (wnd->count >= NTFS_MAX_WND_EXTENTS &&
240 len <= wnd->extent_min) {
241 wnd->uptodated = -1;
242 return;
243 }
244 } else {
e8b8e97f 245 /* Try to find extent before 'bit'. */
3f3b442b
KK
246 n = rb_lookup(&wnd->start_tree, bit);
247
248 if (!n) {
249 n = rb_first(&wnd->start_tree);
250 } else {
251 e = rb_entry(n, struct e_node, start.node);
252 n = rb_next(n);
253 if (e->start.key + e->count.key == bit) {
e8b8e97f 254 /* Remove left. */
3f3b442b
KK
255 bit = e->start.key;
256 len += e->count.key;
257 rb_erase(&e->start.node, &wnd->start_tree);
258 rb_erase(&e->count.node, &wnd->count_tree);
259 wnd->count -= 1;
260 e0 = e;
261 }
262 }
263
264 while (n) {
265 size_t next_end;
266
267 e = rb_entry(n, struct e_node, start.node);
268 next_end = e->start.key + e->count.key;
269 if (e->start.key > end_in)
270 break;
271
e8b8e97f 272 /* Remove right. */
3f3b442b
KK
273 n = rb_next(n);
274 len += next_end - end_in;
275 end_in = next_end;
276 rb_erase(&e->start.node, &wnd->start_tree);
277 rb_erase(&e->count.node, &wnd->count_tree);
278 wnd->count -= 1;
279
280 if (!e0)
281 e0 = e;
282 else
283 kmem_cache_free(ntfs_enode_cachep, e);
284 }
285
286 if (wnd->uptodated != 1) {
e8b8e97f 287 /* Check bits before 'bit'. */
3f3b442b 288 ib = wnd->zone_bit == wnd->zone_end ||
96de65a9
KK
289 bit < wnd->zone_end ?
290 0 :
291 wnd->zone_end;
3f3b442b
KK
292
293 while (bit > ib && wnd_is_free_hlp(wnd, bit - 1, 1)) {
294 bit -= 1;
295 len += 1;
296 }
297
e8b8e97f 298 /* Check bits after 'end_in'. */
3f3b442b 299 ib = wnd->zone_bit == wnd->zone_end ||
96de65a9
KK
300 end_in > wnd->zone_bit ?
301 wnd->nbits :
302 wnd->zone_bit;
3f3b442b
KK
303
304 while (end_in < ib && wnd_is_free_hlp(wnd, end_in, 1)) {
305 end_in += 1;
306 len += 1;
307 }
308 }
309 }
e8b8e97f 310 /* Insert new fragment. */
3f3b442b
KK
311 if (wnd->count >= NTFS_MAX_WND_EXTENTS) {
312 if (e0)
313 kmem_cache_free(ntfs_enode_cachep, e0);
314
315 wnd->uptodated = -1;
316
e8b8e97f 317 /* Compare with smallest fragment. */
3f3b442b
KK
318 n = rb_last(&wnd->count_tree);
319 e = rb_entry(n, struct e_node, count.node);
320 if (len <= e->count.key)
e8b8e97f 321 goto out; /* Do not insert small fragments. */
3f3b442b
KK
322
323 if (build) {
324 struct e_node *e2;
325
326 n = rb_prev(n);
327 e2 = rb_entry(n, struct e_node, count.node);
e8b8e97f 328 /* Smallest fragment will be 'e2->count.key'. */
3f3b442b
KK
329 wnd->extent_min = e2->count.key;
330 }
331
e8b8e97f 332 /* Replace smallest fragment by new one. */
3f3b442b
KK
333 rb_erase(&e->start.node, &wnd->start_tree);
334 rb_erase(&e->count.node, &wnd->count_tree);
335 wnd->count -= 1;
336 } else {
337 e = e0 ? e0 : kmem_cache_alloc(ntfs_enode_cachep, GFP_ATOMIC);
338 if (!e) {
339 wnd->uptodated = -1;
340 goto out;
341 }
342
343 if (build && len <= wnd->extent_min)
344 wnd->extent_min = len;
345 }
346 e->start.key = bit;
347 e->count.key = len;
348 if (len > wnd->extent_max)
349 wnd->extent_max = len;
350
351 rb_insert_start(&wnd->start_tree, e);
352 rb_insert_count(&wnd->count_tree, e);
353 wnd->count += 1;
354
355out:;
356}
357
358/*
e8b8e97f 359 * wnd_remove_free_ext - Remove a run from the cached free space.
3f3b442b
KK
360 */
361static void wnd_remove_free_ext(struct wnd_bitmap *wnd, size_t bit, size_t len)
362{
363 struct rb_node *n, *n3;
364 struct e_node *e, *e3;
365 size_t end_in = bit + len;
366 size_t end3, end, new_key, new_len, max_new_len;
367
e8b8e97f 368 /* Try to find extent before 'bit'. */
3f3b442b
KK
369 n = rb_lookup(&wnd->start_tree, bit);
370
371 if (!n)
372 return;
373
374 e = rb_entry(n, struct e_node, start.node);
375 end = e->start.key + e->count.key;
376
377 new_key = new_len = 0;
378 len = e->count.key;
379
e8b8e97f 380 /* Range [bit,end_in) must be inside 'e' or outside 'e' and 'n'. */
3f3b442b
KK
381 if (e->start.key > bit)
382 ;
383 else if (end_in <= end) {
e8b8e97f 384 /* Range [bit,end_in) inside 'e'. */
3f3b442b
KK
385 new_key = end_in;
386 new_len = end - end_in;
387 len = bit - e->start.key;
388 } else if (bit > end) {
389 bool bmax = false;
390
391 n3 = rb_next(n);
392
393 while (n3) {
394 e3 = rb_entry(n3, struct e_node, start.node);
395 if (e3->start.key >= end_in)
396 break;
397
398 if (e3->count.key == wnd->extent_max)
399 bmax = true;
400
401 end3 = e3->start.key + e3->count.key;
402 if (end3 > end_in) {
403 e3->start.key = end_in;
404 rb_erase(&e3->count.node, &wnd->count_tree);
405 e3->count.key = end3 - end_in;
406 rb_insert_count(&wnd->count_tree, e3);
407 break;
408 }
409
410 n3 = rb_next(n3);
411 rb_erase(&e3->start.node, &wnd->start_tree);
412 rb_erase(&e3->count.node, &wnd->count_tree);
413 wnd->count -= 1;
414 kmem_cache_free(ntfs_enode_cachep, e3);
415 }
416 if (!bmax)
417 return;
418 n3 = rb_first(&wnd->count_tree);
419 wnd->extent_max =
96de65a9
KK
420 n3 ? rb_entry(n3, struct e_node, count.node)->count.key :
421 0;
3f3b442b
KK
422 return;
423 }
424
425 if (e->count.key != wnd->extent_max) {
426 ;
427 } else if (rb_prev(&e->count.node)) {
428 ;
429 } else {
430 n3 = rb_next(&e->count.node);
6e3331ee 431 max_new_len = max(len, new_len);
3f3b442b
KK
432 if (!n3) {
433 wnd->extent_max = max_new_len;
434 } else {
435 e3 = rb_entry(n3, struct e_node, count.node);
436 wnd->extent_max = max(e3->count.key, max_new_len);
437 }
438 }
439
440 if (!len) {
441 if (new_len) {
442 e->start.key = new_key;
443 rb_erase(&e->count.node, &wnd->count_tree);
444 e->count.key = new_len;
445 rb_insert_count(&wnd->count_tree, e);
446 } else {
447 rb_erase(&e->start.node, &wnd->start_tree);
448 rb_erase(&e->count.node, &wnd->count_tree);
449 wnd->count -= 1;
450 kmem_cache_free(ntfs_enode_cachep, e);
451 }
452 goto out;
453 }
454 rb_erase(&e->count.node, &wnd->count_tree);
455 e->count.key = len;
456 rb_insert_count(&wnd->count_tree, e);
457
458 if (!new_len)
459 goto out;
460
461 if (wnd->count >= NTFS_MAX_WND_EXTENTS) {
462 wnd->uptodated = -1;
463
e8b8e97f 464 /* Get minimal extent. */
3f3b442b
KK
465 e = rb_entry(rb_last(&wnd->count_tree), struct e_node,
466 count.node);
467 if (e->count.key > new_len)
468 goto out;
469
e8b8e97f 470 /* Replace minimum. */
3f3b442b
KK
471 rb_erase(&e->start.node, &wnd->start_tree);
472 rb_erase(&e->count.node, &wnd->count_tree);
473 wnd->count -= 1;
474 } else {
475 e = kmem_cache_alloc(ntfs_enode_cachep, GFP_ATOMIC);
476 if (!e)
477 wnd->uptodated = -1;
478 }
479
480 if (e) {
481 e->start.key = new_key;
482 e->count.key = new_len;
483 rb_insert_start(&wnd->start_tree, e);
484 rb_insert_count(&wnd->count_tree, e);
485 wnd->count += 1;
486 }
487
488out:
489 if (!wnd->count && 1 != wnd->uptodated)
490 wnd_rescan(wnd);
491}
492
493/*
e8b8e97f 494 * wnd_rescan - Scan all bitmap. Used while initialization.
3f3b442b
KK
495 */
496static int wnd_rescan(struct wnd_bitmap *wnd)
497{
498 int err = 0;
499 size_t prev_tail = 0;
500 struct super_block *sb = wnd->sb;
501 struct ntfs_sb_info *sbi = sb->s_fs_info;
502 u64 lbo, len = 0;
503 u32 blocksize = sb->s_blocksize;
504 u8 cluster_bits = sbi->cluster_bits;
505 u32 wbits = 8 * sb->s_blocksize;
506 u32 used, frb;
3f3b442b
KK
507 size_t wpos, wbit, iw, vbo;
508 struct buffer_head *bh = NULL;
509 CLST lcn, clen;
510
511 wnd->uptodated = 0;
512 wnd->extent_max = 0;
513 wnd->extent_min = MINUS_ONE_T;
514 wnd->total_zeroes = 0;
515
516 vbo = 0;
517
518 for (iw = 0; iw < wnd->nwnd; iw++) {
519 if (iw + 1 == wnd->nwnd)
520 wbits = wnd->bits_last;
521
522 if (wnd->inited) {
523 if (!wnd->free_bits[iw]) {
e8b8e97f 524 /* All ones. */
3f3b442b
KK
525 if (prev_tail) {
526 wnd_add_free_ext(wnd,
527 vbo * 8 - prev_tail,
528 prev_tail, true);
529 prev_tail = 0;
530 }
531 goto next_wnd;
532 }
533 if (wbits == wnd->free_bits[iw]) {
e8b8e97f 534 /* All zeroes. */
3f3b442b
KK
535 prev_tail += wbits;
536 wnd->total_zeroes += wbits;
537 goto next_wnd;
538 }
539 }
540
541 if (!len) {
542 u32 off = vbo & sbi->cluster_mask;
543
544 if (!run_lookup_entry(&wnd->run, vbo >> cluster_bits,
545 &lcn, &clen, NULL)) {
546 err = -ENOENT;
547 goto out;
548 }
549
550 lbo = ((u64)lcn << cluster_bits) + off;
551 len = ((u64)clen << cluster_bits) - off;
552 }
553
554 bh = ntfs_bread(sb, lbo >> sb->s_blocksize_bits);
555 if (!bh) {
556 err = -EIO;
557 goto out;
558 }
559
08811ba5 560 used = ntfs_bitmap_weight_le(bh->b_data, wbits);
3f3b442b
KK
561 if (used < wbits) {
562 frb = wbits - used;
563 wnd->free_bits[iw] = frb;
564 wnd->total_zeroes += frb;
565 }
566
567 wpos = 0;
568 wbit = vbo * 8;
569
570 if (wbit + wbits > wnd->nbits)
571 wbits = wnd->nbits - wbit;
572
573 do {
08811ba5 574 used = find_next_zero_bit_le(bh->b_data, wbits, wpos);
3f3b442b
KK
575
576 if (used > wpos && prev_tail) {
577 wnd_add_free_ext(wnd, wbit + wpos - prev_tail,
578 prev_tail, true);
579 prev_tail = 0;
580 }
581
582 wpos = used;
583
584 if (wpos >= wbits) {
e8b8e97f 585 /* No free blocks. */
3f3b442b
KK
586 prev_tail = 0;
587 break;
588 }
589
08811ba5 590 frb = find_next_bit_le(bh->b_data, wbits, wpos);
3f3b442b 591 if (frb >= wbits) {
e8b8e97f 592 /* Keep last free block. */
3f3b442b
KK
593 prev_tail += frb - wpos;
594 break;
595 }
596
597 wnd_add_free_ext(wnd, wbit + wpos - prev_tail,
598 frb + prev_tail - wpos, true);
599
e8b8e97f 600 /* Skip free block and first '1'. */
3f3b442b 601 wpos = frb + 1;
e8b8e97f 602 /* Reset previous tail. */
3f3b442b
KK
603 prev_tail = 0;
604 } while (wpos < wbits);
605
606next_wnd:
607
608 if (bh)
609 put_bh(bh);
610 bh = NULL;
611
612 vbo += blocksize;
613 if (len) {
614 len -= blocksize;
615 lbo += blocksize;
616 }
617 }
618
e8b8e97f 619 /* Add last block. */
3f3b442b
KK
620 if (prev_tail)
621 wnd_add_free_ext(wnd, wnd->nbits - prev_tail, prev_tail, true);
622
623 /*
e8b8e97f 624 * Before init cycle wnd->uptodated was 0.
3f3b442b 625 * If any errors or limits occurs while initialization then
e8b8e97f
KA
626 * wnd->uptodated will be -1.
627 * If 'uptodated' is still 0 then Tree is really updated.
3f3b442b
KK
628 */
629 if (!wnd->uptodated)
630 wnd->uptodated = 1;
631
632 if (wnd->zone_bit != wnd->zone_end) {
633 size_t zlen = wnd->zone_end - wnd->zone_bit;
634
635 wnd->zone_end = wnd->zone_bit;
636 wnd_zone_set(wnd, wnd->zone_bit, zlen);
637 }
638
639out:
640 return err;
641}
642
3f3b442b
KK
643int wnd_init(struct wnd_bitmap *wnd, struct super_block *sb, size_t nbits)
644{
645 int err;
646 u32 blocksize = sb->s_blocksize;
647 u32 wbits = blocksize * 8;
648
649 init_rwsem(&wnd->rw_lock);
650
651 wnd->sb = sb;
652 wnd->nbits = nbits;
653 wnd->total_zeroes = nbits;
654 wnd->extent_max = MINUS_ONE_T;
655 wnd->zone_bit = wnd->zone_end = 0;
656 wnd->nwnd = bytes_to_block(sb, bitmap_size(nbits));
657 wnd->bits_last = nbits & (wbits - 1);
658 if (!wnd->bits_last)
659 wnd->bits_last = wbits;
660
6827d50b
KK
661 wnd->free_bits =
662 kcalloc(wnd->nwnd, sizeof(u16), GFP_NOFS | __GFP_NOWARN);
3f3b442b
KK
663 if (!wnd->free_bits)
664 return -ENOMEM;
665
666 err = wnd_rescan(wnd);
667 if (err)
668 return err;
669
670 wnd->inited = true;
671
672 return 0;
673}
674
675/*
e8b8e97f 676 * wnd_map - Call sb_bread for requested window.
3f3b442b
KK
677 */
678static struct buffer_head *wnd_map(struct wnd_bitmap *wnd, size_t iw)
679{
680 size_t vbo;
681 CLST lcn, clen;
682 struct super_block *sb = wnd->sb;
683 struct ntfs_sb_info *sbi;
684 struct buffer_head *bh;
685 u64 lbo;
686
687 sbi = sb->s_fs_info;
688 vbo = (u64)iw << sb->s_blocksize_bits;
689
690 if (!run_lookup_entry(&wnd->run, vbo >> sbi->cluster_bits, &lcn, &clen,
691 NULL)) {
692 return ERR_PTR(-ENOENT);
693 }
694
695 lbo = ((u64)lcn << sbi->cluster_bits) + (vbo & sbi->cluster_mask);
696
697 bh = ntfs_bread(wnd->sb, lbo >> sb->s_blocksize_bits);
698 if (!bh)
699 return ERR_PTR(-EIO);
700
701 return bh;
702}
703
704/*
e8b8e97f 705 * wnd_set_free - Mark the bits range from bit to bit + bits as free.
3f3b442b
KK
706 */
707int wnd_set_free(struct wnd_bitmap *wnd, size_t bit, size_t bits)
708{
709 int err = 0;
710 struct super_block *sb = wnd->sb;
711 size_t bits0 = bits;
712 u32 wbits = 8 * sb->s_blocksize;
713 size_t iw = bit >> (sb->s_blocksize_bits + 3);
714 u32 wbit = bit & (wbits - 1);
715 struct buffer_head *bh;
716
717 while (iw < wnd->nwnd && bits) {
718 u32 tail, op;
3f3b442b
KK
719
720 if (iw + 1 == wnd->nwnd)
721 wbits = wnd->bits_last;
722
723 tail = wbits - wbit;
6e3331ee 724 op = min_t(u32, tail, bits);
3f3b442b
KK
725
726 bh = wnd_map(wnd, iw);
727 if (IS_ERR(bh)) {
728 err = PTR_ERR(bh);
729 break;
730 }
731
3f3b442b
KK
732 lock_buffer(bh);
733
08811ba5 734 ntfs_bitmap_clear_le(bh->b_data, wbit, op);
3f3b442b
KK
735
736 wnd->free_bits[iw] += op;
737
738 set_buffer_uptodate(bh);
739 mark_buffer_dirty(bh);
740 unlock_buffer(bh);
741 put_bh(bh);
742
743 wnd->total_zeroes += op;
744 bits -= op;
745 wbit = 0;
746 iw += 1;
747 }
748
749 wnd_add_free_ext(wnd, bit, bits0, false);
750
751 return err;
752}
753
754/*
e8b8e97f 755 * wnd_set_used - Mark the bits range from bit to bit + bits as used.
3f3b442b
KK
756 */
757int wnd_set_used(struct wnd_bitmap *wnd, size_t bit, size_t bits)
758{
759 int err = 0;
760 struct super_block *sb = wnd->sb;
761 size_t bits0 = bits;
762 size_t iw = bit >> (sb->s_blocksize_bits + 3);
763 u32 wbits = 8 * sb->s_blocksize;
764 u32 wbit = bit & (wbits - 1);
765 struct buffer_head *bh;
766
767 while (iw < wnd->nwnd && bits) {
768 u32 tail, op;
3f3b442b
KK
769
770 if (unlikely(iw + 1 == wnd->nwnd))
771 wbits = wnd->bits_last;
772
773 tail = wbits - wbit;
6e3331ee 774 op = min_t(u32, tail, bits);
3f3b442b
KK
775
776 bh = wnd_map(wnd, iw);
777 if (IS_ERR(bh)) {
778 err = PTR_ERR(bh);
779 break;
780 }
3f3b442b
KK
781
782 lock_buffer(bh);
783
08811ba5 784 ntfs_bitmap_set_le(bh->b_data, wbit, op);
3f3b442b
KK
785 wnd->free_bits[iw] -= op;
786
787 set_buffer_uptodate(bh);
788 mark_buffer_dirty(bh);
789 unlock_buffer(bh);
790 put_bh(bh);
791
792 wnd->total_zeroes -= op;
793 bits -= op;
794 wbit = 0;
795 iw += 1;
796 }
797
798 if (!RB_EMPTY_ROOT(&wnd->start_tree))
799 wnd_remove_free_ext(wnd, bit, bits0);
800
801 return err;
802}
803
ec5fc720
KK
804/*
805 * wnd_set_used_safe - Mark the bits range from bit to bit + bits as used.
806 *
807 * Unlikely wnd_set_used/wnd_set_free this function is not full trusted.
808 * It scans every bit in bitmap and marks free bit as used.
809 * @done - how many bits were marked as used.
810 *
811 * NOTE: normally *done should be 0.
812 */
813int wnd_set_used_safe(struct wnd_bitmap *wnd, size_t bit, size_t bits,
814 size_t *done)
815{
816 size_t i, from = 0, len = 0;
817 int err = 0;
818
819 *done = 0;
820 for (i = 0; i < bits; i++) {
821 if (wnd_is_free(wnd, bit + i, 1)) {
822 if (!len)
823 from = bit + i;
824 len += 1;
825 } else if (len) {
826 err = wnd_set_used(wnd, from, len);
827 *done += len;
828 len = 0;
829 if (err)
830 break;
831 }
832 }
833
834 if (len) {
835 /* last fragment. */
836 err = wnd_set_used(wnd, from, len);
837 *done += len;
838 }
839 return err;
840}
841
3f3b442b
KK
842/*
843 * wnd_is_free_hlp
844 *
e8b8e97f 845 * Return: True if all clusters [bit, bit+bits) are free (bitmap only).
3f3b442b
KK
846 */
847static bool wnd_is_free_hlp(struct wnd_bitmap *wnd, size_t bit, size_t bits)
848{
849 struct super_block *sb = wnd->sb;
850 size_t iw = bit >> (sb->s_blocksize_bits + 3);
851 u32 wbits = 8 * sb->s_blocksize;
852 u32 wbit = bit & (wbits - 1);
853
854 while (iw < wnd->nwnd && bits) {
855 u32 tail, op;
856
857 if (unlikely(iw + 1 == wnd->nwnd))
858 wbits = wnd->bits_last;
859
860 tail = wbits - wbit;
6e3331ee 861 op = min_t(u32, tail, bits);
3f3b442b
KK
862
863 if (wbits != wnd->free_bits[iw]) {
864 bool ret;
865 struct buffer_head *bh = wnd_map(wnd, iw);
866
867 if (IS_ERR(bh))
868 return false;
869
08811ba5 870 ret = are_bits_clear(bh->b_data, wbit, op);
3f3b442b
KK
871
872 put_bh(bh);
873 if (!ret)
874 return false;
875 }
876
877 bits -= op;
878 wbit = 0;
879 iw += 1;
880 }
881
882 return true;
883}
884
885/*
886 * wnd_is_free
887 *
e8b8e97f 888 * Return: True if all clusters [bit, bit+bits) are free.
3f3b442b
KK
889 */
890bool wnd_is_free(struct wnd_bitmap *wnd, size_t bit, size_t bits)
891{
892 bool ret;
893 struct rb_node *n;
894 size_t end;
895 struct e_node *e;
896
897 if (RB_EMPTY_ROOT(&wnd->start_tree))
898 goto use_wnd;
899
900 n = rb_lookup(&wnd->start_tree, bit);
901 if (!n)
902 goto use_wnd;
903
904 e = rb_entry(n, struct e_node, start.node);
905
906 end = e->start.key + e->count.key;
907
908 if (bit < end && bit + bits <= end)
909 return true;
910
911use_wnd:
912 ret = wnd_is_free_hlp(wnd, bit, bits);
913
914 return ret;
915}
916
917/*
918 * wnd_is_used
919 *
e8b8e97f 920 * Return: True if all clusters [bit, bit+bits) are used.
3f3b442b
KK
921 */
922bool wnd_is_used(struct wnd_bitmap *wnd, size_t bit, size_t bits)
923{
924 bool ret = false;
925 struct super_block *sb = wnd->sb;
926 size_t iw = bit >> (sb->s_blocksize_bits + 3);
927 u32 wbits = 8 * sb->s_blocksize;
928 u32 wbit = bit & (wbits - 1);
929 size_t end;
930 struct rb_node *n;
931 struct e_node *e;
932
933 if (RB_EMPTY_ROOT(&wnd->start_tree))
934 goto use_wnd;
935
936 end = bit + bits;
937 n = rb_lookup(&wnd->start_tree, end - 1);
938 if (!n)
939 goto use_wnd;
940
941 e = rb_entry(n, struct e_node, start.node);
942 if (e->start.key + e->count.key > bit)
943 return false;
944
945use_wnd:
946 while (iw < wnd->nwnd && bits) {
947 u32 tail, op;
948
949 if (unlikely(iw + 1 == wnd->nwnd))
950 wbits = wnd->bits_last;
951
952 tail = wbits - wbit;
6e3331ee 953 op = min_t(u32, tail, bits);
3f3b442b
KK
954
955 if (wnd->free_bits[iw]) {
956 bool ret;
957 struct buffer_head *bh = wnd_map(wnd, iw);
958
959 if (IS_ERR(bh))
960 goto out;
961
08811ba5 962 ret = are_bits_set(bh->b_data, wbit, op);
3f3b442b
KK
963 put_bh(bh);
964 if (!ret)
965 goto out;
966 }
967
968 bits -= op;
969 wbit = 0;
970 iw += 1;
971 }
972 ret = true;
973
974out:
975 return ret;
976}
977
978/*
e8b8e97f
KA
979 * wnd_find - Look for free space.
980 *
3f3b442b
KK
981 * - flags - BITMAP_FIND_XXX flags
982 *
e8b8e97f 983 * Return: 0 if not found.
3f3b442b
KK
984 */
985size_t wnd_find(struct wnd_bitmap *wnd, size_t to_alloc, size_t hint,
986 size_t flags, size_t *allocated)
987{
988 struct super_block *sb;
989 u32 wbits, wpos, wzbit, wzend;
990 size_t fnd, max_alloc, b_len, b_pos;
991 size_t iw, prev_tail, nwnd, wbit, ebit, zbit, zend;
992 size_t to_alloc0 = to_alloc;
3f3b442b
KK
993 const struct e_node *e;
994 const struct rb_node *pr, *cr;
995 u8 log2_bits;
996 bool fbits_valid;
997 struct buffer_head *bh;
998
e8b8e97f 999 /* Fast checking for available free space. */
3f3b442b
KK
1000 if (flags & BITMAP_FIND_FULL) {
1001 size_t zeroes = wnd_zeroes(wnd);
1002
1003 zeroes -= wnd->zone_end - wnd->zone_bit;
1004 if (zeroes < to_alloc0)
1005 goto no_space;
1006
1007 if (to_alloc0 > wnd->extent_max)
1008 goto no_space;
1009 } else {
1010 if (to_alloc > wnd->extent_max)
1011 to_alloc = wnd->extent_max;
1012 }
1013
1014 if (wnd->zone_bit <= hint && hint < wnd->zone_end)
1015 hint = wnd->zone_end;
1016
1017 max_alloc = wnd->nbits;
1018 b_len = b_pos = 0;
1019
1020 if (hint >= max_alloc)
1021 hint = 0;
1022
1023 if (RB_EMPTY_ROOT(&wnd->start_tree)) {
1024 if (wnd->uptodated == 1) {
e8b8e97f 1025 /* Extents tree is updated -> No free space. */
3f3b442b
KK
1026 goto no_space;
1027 }
1028 goto scan_bitmap;
1029 }
1030
1031 e = NULL;
1032 if (!hint)
1033 goto allocate_biggest;
1034
e8b8e97f 1035 /* Use hint: Enumerate extents by start >= hint. */
3f3b442b
KK
1036 pr = NULL;
1037 cr = wnd->start_tree.rb_node;
1038
1039 for (;;) {
1040 e = rb_entry(cr, struct e_node, start.node);
1041
1042 if (e->start.key == hint)
1043 break;
1044
1045 if (e->start.key < hint) {
1046 pr = cr;
1047 cr = cr->rb_right;
1048 if (!cr)
1049 break;
1050 continue;
1051 }
1052
1053 cr = cr->rb_left;
1054 if (!cr) {
1055 e = pr ? rb_entry(pr, struct e_node, start.node) : NULL;
1056 break;
1057 }
1058 }
1059
1060 if (!e)
1061 goto allocate_biggest;
1062
1063 if (e->start.key + e->count.key > hint) {
e8b8e97f 1064 /* We have found extension with 'hint' inside. */
3f3b442b
KK
1065 size_t len = e->start.key + e->count.key - hint;
1066
1067 if (len >= to_alloc && hint + to_alloc <= max_alloc) {
1068 fnd = hint;
1069 goto found;
1070 }
1071
1072 if (!(flags & BITMAP_FIND_FULL)) {
1073 if (len > to_alloc)
1074 len = to_alloc;
1075
1076 if (hint + len <= max_alloc) {
1077 fnd = hint;
1078 to_alloc = len;
1079 goto found;
1080 }
1081 }
1082 }
1083
1084allocate_biggest:
e8b8e97f 1085 /* Allocate from biggest free extent. */
3f3b442b
KK
1086 e = rb_entry(rb_first(&wnd->count_tree), struct e_node, count.node);
1087 if (e->count.key != wnd->extent_max)
1088 wnd->extent_max = e->count.key;
1089
1090 if (e->count.key < max_alloc) {
1091 if (e->count.key >= to_alloc) {
1092 ;
1093 } else if (flags & BITMAP_FIND_FULL) {
1094 if (e->count.key < to_alloc0) {
e8b8e97f 1095 /* Biggest free block is less then requested. */
3f3b442b
KK
1096 goto no_space;
1097 }
1098 to_alloc = e->count.key;
1099 } else if (-1 != wnd->uptodated) {
1100 to_alloc = e->count.key;
1101 } else {
e8b8e97f 1102 /* Check if we can use more bits. */
3f3b442b
KK
1103 size_t op, max_check;
1104 struct rb_root start_tree;
1105
1106 memcpy(&start_tree, &wnd->start_tree,
1107 sizeof(struct rb_root));
1108 memset(&wnd->start_tree, 0, sizeof(struct rb_root));
1109
1110 max_check = e->start.key + to_alloc;
1111 if (max_check > max_alloc)
1112 max_check = max_alloc;
1113 for (op = e->start.key + e->count.key; op < max_check;
1114 op++) {
1115 if (!wnd_is_free(wnd, op, 1))
1116 break;
1117 }
1118 memcpy(&wnd->start_tree, &start_tree,
1119 sizeof(struct rb_root));
1120 to_alloc = op - e->start.key;
1121 }
1122
e8b8e97f 1123 /* Prepare to return. */
3f3b442b
KK
1124 fnd = e->start.key;
1125 if (e->start.key + to_alloc > max_alloc)
1126 to_alloc = max_alloc - e->start.key;
1127 goto found;
1128 }
1129
1130 if (wnd->uptodated == 1) {
e8b8e97f 1131 /* Extents tree is updated -> no free space. */
3f3b442b
KK
1132 goto no_space;
1133 }
1134
1135 b_len = e->count.key;
1136 b_pos = e->start.key;
1137
1138scan_bitmap:
1139 sb = wnd->sb;
1140 log2_bits = sb->s_blocksize_bits + 3;
1141
d3624466 1142 /* At most two ranges [hint, max_alloc) + [0, hint). */
3f3b442b
KK
1143Again:
1144
e8b8e97f 1145 /* TODO: Optimize request for case nbits > wbits. */
3f3b442b
KK
1146 iw = hint >> log2_bits;
1147 wbits = sb->s_blocksize * 8;
1148 wpos = hint & (wbits - 1);
1149 prev_tail = 0;
1150 fbits_valid = true;
1151
1152 if (max_alloc == wnd->nbits) {
1153 nwnd = wnd->nwnd;
1154 } else {
1155 size_t t = max_alloc + wbits - 1;
1156
1157 nwnd = likely(t > max_alloc) ? (t >> log2_bits) : wnd->nwnd;
1158 }
1159
e8b8e97f 1160 /* Enumerate all windows. */
3f3b442b
KK
1161 for (; iw < nwnd; iw++) {
1162 wbit = iw << log2_bits;
1163
1164 if (!wnd->free_bits[iw]) {
1165 if (prev_tail > b_len) {
1166 b_pos = wbit - prev_tail;
1167 b_len = prev_tail;
1168 }
1169
e8b8e97f 1170 /* Skip full used window. */
3f3b442b
KK
1171 prev_tail = 0;
1172 wpos = 0;
1173 continue;
1174 }
1175
1176 if (unlikely(iw + 1 == nwnd)) {
1177 if (max_alloc == wnd->nbits) {
1178 wbits = wnd->bits_last;
1179 } else {
1180 size_t t = max_alloc & (wbits - 1);
1181
1182 if (t) {
1183 wbits = t;
1184 fbits_valid = false;
1185 }
1186 }
1187 }
1188
1189 if (wnd->zone_end > wnd->zone_bit) {
1190 ebit = wbit + wbits;
1191 zbit = max(wnd->zone_bit, wbit);
1192 zend = min(wnd->zone_end, ebit);
1193
e8b8e97f 1194 /* Here we have a window [wbit, ebit) and zone [zbit, zend). */
3f3b442b 1195 if (zend <= zbit) {
e8b8e97f 1196 /* Zone does not overlap window. */
3f3b442b
KK
1197 } else {
1198 wzbit = zbit - wbit;
1199 wzend = zend - wbit;
1200
e8b8e97f 1201 /* Zone overlaps window. */
3f3b442b
KK
1202 if (wnd->free_bits[iw] == wzend - wzbit) {
1203 prev_tail = 0;
1204 wpos = 0;
1205 continue;
1206 }
1207
e8b8e97f 1208 /* Scan two ranges window: [wbit, zbit) and [zend, ebit). */
3f3b442b
KK
1209 bh = wnd_map(wnd, iw);
1210
1211 if (IS_ERR(bh)) {
e8b8e97f 1212 /* TODO: Error */
3f3b442b
KK
1213 prev_tail = 0;
1214 wpos = 0;
1215 continue;
1216 }
1217
e8b8e97f 1218 /* Scan range [wbit, zbit). */
3f3b442b 1219 if (wpos < wzbit) {
e8b8e97f 1220 /* Scan range [wpos, zbit). */
08811ba5
KK
1221 fnd = wnd_scan(bh->b_data, wbit, wpos,
1222 wzbit, to_alloc,
1223 &prev_tail, &b_pos,
1224 &b_len);
3f3b442b
KK
1225 if (fnd != MINUS_ONE_T) {
1226 put_bh(bh);
1227 goto found;
1228 }
1229 }
1230
1231 prev_tail = 0;
1232
e8b8e97f 1233 /* Scan range [zend, ebit). */
3f3b442b 1234 if (wzend < wbits) {
08811ba5 1235 fnd = wnd_scan(bh->b_data, wbit,
3f3b442b
KK
1236 max(wzend, wpos), wbits,
1237 to_alloc, &prev_tail,
1238 &b_pos, &b_len);
1239 if (fnd != MINUS_ONE_T) {
1240 put_bh(bh);
1241 goto found;
1242 }
1243 }
1244
1245 wpos = 0;
1246 put_bh(bh);
1247 continue;
1248 }
1249 }
1250
e8b8e97f 1251 /* Current window does not overlap zone. */
3f3b442b 1252 if (!wpos && fbits_valid && wnd->free_bits[iw] == wbits) {
e8b8e97f 1253 /* Window is empty. */
3f3b442b
KK
1254 if (prev_tail + wbits >= to_alloc) {
1255 fnd = wbit + wpos - prev_tail;
1256 goto found;
1257 }
1258
e8b8e97f 1259 /* Increase 'prev_tail' and process next window. */
3f3b442b
KK
1260 prev_tail += wbits;
1261 wpos = 0;
1262 continue;
1263 }
1264
d3624466 1265 /* Read window. */
3f3b442b
KK
1266 bh = wnd_map(wnd, iw);
1267 if (IS_ERR(bh)) {
e8b8e97f 1268 // TODO: Error.
3f3b442b
KK
1269 prev_tail = 0;
1270 wpos = 0;
1271 continue;
1272 }
1273
e8b8e97f 1274 /* Scan range [wpos, eBits). */
08811ba5
KK
1275 fnd = wnd_scan(bh->b_data, wbit, wpos, wbits, to_alloc,
1276 &prev_tail, &b_pos, &b_len);
3f3b442b
KK
1277 put_bh(bh);
1278 if (fnd != MINUS_ONE_T)
1279 goto found;
1280 }
1281
1282 if (b_len < prev_tail) {
e8b8e97f 1283 /* The last fragment. */
3f3b442b
KK
1284 b_len = prev_tail;
1285 b_pos = max_alloc - prev_tail;
1286 }
1287
1288 if (hint) {
1289 /*
e8b8e97f
KA
1290 * We have scanned range [hint max_alloc).
1291 * Prepare to scan range [0 hint + to_alloc).
3f3b442b
KK
1292 */
1293 size_t nextmax = hint + to_alloc;
1294
1295 if (likely(nextmax >= hint) && nextmax < max_alloc)
1296 max_alloc = nextmax;
1297 hint = 0;
1298 goto Again;
1299 }
1300
1301 if (!b_len)
1302 goto no_space;
1303
1304 wnd->extent_max = b_len;
1305
1306 if (flags & BITMAP_FIND_FULL)
1307 goto no_space;
1308
1309 fnd = b_pos;
1310 to_alloc = b_len;
1311
1312found:
1313 if (flags & BITMAP_FIND_MARK_AS_USED) {
e8b8e97f 1314 /* TODO: Optimize remove extent (pass 'e'?). */
3f3b442b
KK
1315 if (wnd_set_used(wnd, fnd, to_alloc))
1316 goto no_space;
1317 } else if (wnd->extent_max != MINUS_ONE_T &&
1318 to_alloc > wnd->extent_max) {
1319 wnd->extent_max = to_alloc;
1320 }
1321
1322 *allocated = fnd;
1323 return to_alloc;
1324
1325no_space:
1326 return 0;
1327}
1328
1329/*
e8b8e97f 1330 * wnd_extend - Extend bitmap ($MFT bitmap).
3f3b442b
KK
1331 */
1332int wnd_extend(struct wnd_bitmap *wnd, size_t new_bits)
1333{
1334 int err;
1335 struct super_block *sb = wnd->sb;
1336 struct ntfs_sb_info *sbi = sb->s_fs_info;
1337 u32 blocksize = sb->s_blocksize;
1338 u32 wbits = blocksize * 8;
1339 u32 b0, new_last;
1340 size_t bits, iw, new_wnd;
1341 size_t old_bits = wnd->nbits;
1342 u16 *new_free;
1343
1344 if (new_bits <= old_bits)
1345 return -EINVAL;
1346
e8b8e97f 1347 /* Align to 8 byte boundary. */
3f3b442b
KK
1348 new_wnd = bytes_to_block(sb, bitmap_size(new_bits));
1349 new_last = new_bits & (wbits - 1);
1350 if (!new_last)
1351 new_last = wbits;
1352
1353 if (new_wnd != wnd->nwnd) {
92f017c4 1354 new_free = kmalloc_array(new_wnd, sizeof(u16), GFP_NOFS);
3f3b442b
KK
1355 if (!new_free)
1356 return -ENOMEM;
1357
548744f8 1358 memcpy(new_free, wnd->free_bits, wnd->nwnd * sizeof(short));
3f3b442b
KK
1359 memset(new_free + wnd->nwnd, 0,
1360 (new_wnd - wnd->nwnd) * sizeof(short));
195c52bd 1361 kfree(wnd->free_bits);
3f3b442b
KK
1362 wnd->free_bits = new_free;
1363 }
1364
e8b8e97f 1365 /* Zero bits [old_bits,new_bits). */
3f3b442b
KK
1366 bits = new_bits - old_bits;
1367 b0 = old_bits & (wbits - 1);
1368
1369 for (iw = old_bits >> (sb->s_blocksize_bits + 3); bits; iw += 1) {
1370 u32 op;
1371 size_t frb;
1372 u64 vbo, lbo, bytes;
1373 struct buffer_head *bh;
3f3b442b
KK
1374
1375 if (iw + 1 == new_wnd)
1376 wbits = new_last;
1377
1378 op = b0 + bits > wbits ? wbits - b0 : bits;
1379 vbo = (u64)iw * blocksize;
1380
1381 err = ntfs_vbo_to_lbo(sbi, &wnd->run, vbo, &lbo, &bytes);
1382 if (err)
1383 break;
1384
1385 bh = ntfs_bread(sb, lbo >> sb->s_blocksize_bits);
1386 if (!bh)
1387 return -EIO;
1388
1389 lock_buffer(bh);
3f3b442b 1390
08811ba5
KK
1391 ntfs_bitmap_clear_le(bh->b_data, b0, blocksize * 8 - b0);
1392 frb = wbits - ntfs_bitmap_weight_le(bh->b_data, wbits);
3f3b442b
KK
1393 wnd->total_zeroes += frb - wnd->free_bits[iw];
1394 wnd->free_bits[iw] = frb;
1395
1396 set_buffer_uptodate(bh);
1397 mark_buffer_dirty(bh);
1398 unlock_buffer(bh);
e8b8e97f 1399 /* err = sync_dirty_buffer(bh); */
3f3b442b
KK
1400
1401 b0 = 0;
1402 bits -= op;
1403 }
1404
1405 wnd->nbits = new_bits;
1406 wnd->nwnd = new_wnd;
1407 wnd->bits_last = new_last;
1408
1409 wnd_add_free_ext(wnd, old_bits, new_bits - old_bits, false);
1410
1411 return 0;
1412}
1413
3f3b442b
KK
1414void wnd_zone_set(struct wnd_bitmap *wnd, size_t lcn, size_t len)
1415{
54033c13 1416 size_t zlen = wnd->zone_end - wnd->zone_bit;
3f3b442b 1417
3f3b442b
KK
1418 if (zlen)
1419 wnd_add_free_ext(wnd, wnd->zone_bit, zlen, false);
1420
1421 if (!RB_EMPTY_ROOT(&wnd->start_tree) && len)
1422 wnd_remove_free_ext(wnd, lcn, len);
1423
1424 wnd->zone_bit = lcn;
1425 wnd->zone_end = lcn + len;
1426}
1427
1428int ntfs_trim_fs(struct ntfs_sb_info *sbi, struct fstrim_range *range)
1429{
1430 int err = 0;
1431 struct super_block *sb = sbi->sb;
1432 struct wnd_bitmap *wnd = &sbi->used.bitmap;
1433 u32 wbits = 8 * sb->s_blocksize;
1434 CLST len = 0, lcn = 0, done = 0;
1435 CLST minlen = bytes_to_cluster(sbi, range->minlen);
1436 CLST lcn_from = bytes_to_cluster(sbi, range->start);
1437 size_t iw = lcn_from >> (sb->s_blocksize_bits + 3);
1438 u32 wbit = lcn_from & (wbits - 1);
3f3b442b
KK
1439 CLST lcn_to;
1440
1441 if (!minlen)
1442 minlen = 1;
1443
1444 if (range->len == (u64)-1)
1445 lcn_to = wnd->nbits;
1446 else
1447 lcn_to = bytes_to_cluster(sbi, range->start + range->len);
1448
1449 down_read_nested(&wnd->rw_lock, BITMAP_MUTEX_CLUSTERS);
1450
557d1967 1451 for (; iw < wnd->nwnd; iw++, wbit = 0) {
3f3b442b
KK
1452 CLST lcn_wnd = iw * wbits;
1453 struct buffer_head *bh;
1454
1455 if (lcn_wnd > lcn_to)
1456 break;
1457
1458 if (!wnd->free_bits[iw])
1459 continue;
1460
1461 if (iw + 1 == wnd->nwnd)
1462 wbits = wnd->bits_last;
1463
1464 if (lcn_wnd + wbits > lcn_to)
1465 wbits = lcn_to - lcn_wnd;
1466
1467 bh = wnd_map(wnd, iw);
1468 if (IS_ERR(bh)) {
1469 err = PTR_ERR(bh);
1470 break;
1471 }
1472
3f3b442b 1473 for (; wbit < wbits; wbit++) {
08811ba5 1474 if (!test_bit_le(wbit, bh->b_data)) {
3f3b442b
KK
1475 if (!len)
1476 lcn = lcn_wnd + wbit;
1477 len += 1;
1478 continue;
1479 }
1480 if (len >= minlen) {
1481 err = ntfs_discard(sbi, lcn, len);
1482 if (err)
1483 goto out;
1484 done += len;
1485 }
1486 len = 0;
1487 }
1488 put_bh(bh);
1489 }
1490
e8b8e97f 1491 /* Process the last fragment. */
3f3b442b
KK
1492 if (len >= minlen) {
1493 err = ntfs_discard(sbi, lcn, len);
1494 if (err)
1495 goto out;
1496 done += len;
1497 }
1498
1499out:
1500 range->len = (u64)done << sbi->cluster_bits;
1501
1502 up_read(&wnd->rw_lock);
1503
1504 return err;
1505}
88a8d0d2 1506
08811ba5
KK
1507#if BITS_PER_LONG == 64
1508typedef __le64 bitmap_ulong;
1509#define cpu_to_ul(x) cpu_to_le64(x)
1510#define ul_to_cpu(x) le64_to_cpu(x)
1511#else
1512typedef __le32 bitmap_ulong;
1513#define cpu_to_ul(x) cpu_to_le32(x)
1514#define ul_to_cpu(x) le32_to_cpu(x)
1515#endif
1516
1517void ntfs_bitmap_set_le(void *map, unsigned int start, int len)
88a8d0d2 1518{
08811ba5 1519 bitmap_ulong *p = (bitmap_ulong *)map + BIT_WORD(start);
88a8d0d2
TK
1520 const unsigned int size = start + len;
1521 int bits_to_set = BITS_PER_LONG - (start % BITS_PER_LONG);
08811ba5 1522 bitmap_ulong mask_to_set = cpu_to_ul(BITMAP_FIRST_WORD_MASK(start));
88a8d0d2
TK
1523
1524 while (len - bits_to_set >= 0) {
1525 *p |= mask_to_set;
1526 len -= bits_to_set;
1527 bits_to_set = BITS_PER_LONG;
08811ba5 1528 mask_to_set = cpu_to_ul(~0UL);
88a8d0d2
TK
1529 p++;
1530 }
1531 if (len) {
08811ba5 1532 mask_to_set &= cpu_to_ul(BITMAP_LAST_WORD_MASK(size));
88a8d0d2
TK
1533 *p |= mask_to_set;
1534 }
1535}
1536
08811ba5 1537void ntfs_bitmap_clear_le(void *map, unsigned int start, int len)
88a8d0d2 1538{
08811ba5 1539 bitmap_ulong *p = (bitmap_ulong *)map + BIT_WORD(start);
88a8d0d2
TK
1540 const unsigned int size = start + len;
1541 int bits_to_clear = BITS_PER_LONG - (start % BITS_PER_LONG);
08811ba5 1542 bitmap_ulong mask_to_clear = cpu_to_ul(BITMAP_FIRST_WORD_MASK(start));
88a8d0d2
TK
1543
1544 while (len - bits_to_clear >= 0) {
1545 *p &= ~mask_to_clear;
1546 len -= bits_to_clear;
1547 bits_to_clear = BITS_PER_LONG;
08811ba5 1548 mask_to_clear = cpu_to_ul(~0UL);
88a8d0d2
TK
1549 p++;
1550 }
1551 if (len) {
08811ba5 1552 mask_to_clear &= cpu_to_ul(BITMAP_LAST_WORD_MASK(size));
88a8d0d2
TK
1553 *p &= ~mask_to_clear;
1554 }
1555}
08811ba5
KK
1556
1557unsigned int ntfs_bitmap_weight_le(const void *bitmap, int bits)
1558{
1559 const ulong *bmp = bitmap;
1560 unsigned int k, lim = bits / BITS_PER_LONG;
1561 unsigned int w = 0;
1562
1563 for (k = 0; k < lim; k++)
1564 w += hweight_long(bmp[k]);
1565
1566 if (bits % BITS_PER_LONG) {
1567 w += hweight_long(ul_to_cpu(((bitmap_ulong *)bitmap)[k]) &
1568 BITMAP_LAST_WORD_MASK(bits));
1569 }
1570
1571 return w;
1572}