Commit | Line | Data |
---|---|---|
7c1a000d | 1 | // SPDX-License-Identifier: GPL-2.0 |
a28ef1f5 CY |
2 | /* |
3 | * f2fs extent cache support | |
4 | * | |
5 | * Copyright (c) 2015 Motorola Mobility | |
6 | * Copyright (c) 2015 Samsung Electronics | |
7 | * Authors: Jaegeuk Kim <jaegeuk@kernel.org> | |
8 | * Chao Yu <chao2.yu@samsung.com> | |
a28ef1f5 CY |
9 | */ |
10 | ||
11 | #include <linux/fs.h> | |
12 | #include <linux/f2fs_fs.h> | |
13 | ||
14 | #include "f2fs.h" | |
15 | #include "node.h" | |
16 | #include <trace/events/f2fs.h> | |
17 | ||
54c2258c CY |
18 | static struct rb_entry *__lookup_rb_tree_fast(struct rb_entry *cached_re, |
19 | unsigned int ofs) | |
20 | { | |
21 | if (cached_re) { | |
22 | if (cached_re->ofs <= ofs && | |
23 | cached_re->ofs + cached_re->len > ofs) { | |
24 | return cached_re; | |
25 | } | |
26 | } | |
27 | return NULL; | |
28 | } | |
29 | ||
30 | static struct rb_entry *__lookup_rb_tree_slow(struct rb_root *root, | |
31 | unsigned int ofs) | |
32 | { | |
33 | struct rb_node *node = root->rb_node; | |
34 | struct rb_entry *re; | |
35 | ||
36 | while (node) { | |
37 | re = rb_entry(node, struct rb_entry, rb_node); | |
38 | ||
39 | if (ofs < re->ofs) | |
40 | node = node->rb_left; | |
41 | else if (ofs >= re->ofs + re->len) | |
42 | node = node->rb_right; | |
43 | else | |
44 | return re; | |
45 | } | |
46 | return NULL; | |
47 | } | |
48 | ||
4d57b86d | 49 | struct rb_entry *f2fs_lookup_rb_tree(struct rb_root *root, |
54c2258c CY |
50 | struct rb_entry *cached_re, unsigned int ofs) |
51 | { | |
52 | struct rb_entry *re; | |
53 | ||
54 | re = __lookup_rb_tree_fast(cached_re, ofs); | |
55 | if (!re) | |
56 | return __lookup_rb_tree_slow(root, ofs); | |
57 | ||
58 | return re; | |
59 | } | |
60 | ||
4d57b86d | 61 | struct rb_node **f2fs_lookup_rb_tree_for_insert(struct f2fs_sb_info *sbi, |
54c2258c CY |
62 | struct rb_root *root, struct rb_node **parent, |
63 | unsigned int ofs) | |
64 | { | |
65 | struct rb_node **p = &root->rb_node; | |
66 | struct rb_entry *re; | |
67 | ||
68 | while (*p) { | |
69 | *parent = *p; | |
70 | re = rb_entry(*parent, struct rb_entry, rb_node); | |
71 | ||
72 | if (ofs < re->ofs) | |
73 | p = &(*p)->rb_left; | |
74 | else if (ofs >= re->ofs + re->len) | |
75 | p = &(*p)->rb_right; | |
76 | else | |
77 | f2fs_bug_on(sbi, 1); | |
78 | } | |
79 | ||
80 | return p; | |
81 | } | |
82 | ||
83 | /* | |
84 | * lookup rb entry in position of @ofs in rb-tree, | |
85 | * if hit, return the entry, otherwise, return NULL | |
86 | * @prev_ex: extent before ofs | |
87 | * @next_ex: extent after ofs | |
88 | * @insert_p: insert point for new extent at ofs | |
89 | * in order to simpfy the insertion after. | |
90 | * tree must stay unchanged between lookup and insertion. | |
91 | */ | |
4d57b86d | 92 | struct rb_entry *f2fs_lookup_rb_tree_ret(struct rb_root *root, |
54c2258c CY |
93 | struct rb_entry *cached_re, |
94 | unsigned int ofs, | |
95 | struct rb_entry **prev_entry, | |
96 | struct rb_entry **next_entry, | |
97 | struct rb_node ***insert_p, | |
004b6862 CY |
98 | struct rb_node **insert_parent, |
99 | bool force) | |
54c2258c CY |
100 | { |
101 | struct rb_node **pnode = &root->rb_node; | |
102 | struct rb_node *parent = NULL, *tmp_node; | |
103 | struct rb_entry *re = cached_re; | |
104 | ||
105 | *insert_p = NULL; | |
106 | *insert_parent = NULL; | |
107 | *prev_entry = NULL; | |
108 | *next_entry = NULL; | |
109 | ||
110 | if (RB_EMPTY_ROOT(root)) | |
111 | return NULL; | |
112 | ||
113 | if (re) { | |
114 | if (re->ofs <= ofs && re->ofs + re->len > ofs) | |
115 | goto lookup_neighbors; | |
116 | } | |
117 | ||
118 | while (*pnode) { | |
119 | parent = *pnode; | |
120 | re = rb_entry(*pnode, struct rb_entry, rb_node); | |
121 | ||
122 | if (ofs < re->ofs) | |
123 | pnode = &(*pnode)->rb_left; | |
124 | else if (ofs >= re->ofs + re->len) | |
125 | pnode = &(*pnode)->rb_right; | |
126 | else | |
127 | goto lookup_neighbors; | |
128 | } | |
129 | ||
130 | *insert_p = pnode; | |
131 | *insert_parent = parent; | |
132 | ||
133 | re = rb_entry(parent, struct rb_entry, rb_node); | |
134 | tmp_node = parent; | |
135 | if (parent && ofs > re->ofs) | |
136 | tmp_node = rb_next(parent); | |
137 | *next_entry = rb_entry_safe(tmp_node, struct rb_entry, rb_node); | |
138 | ||
139 | tmp_node = parent; | |
140 | if (parent && ofs < re->ofs) | |
141 | tmp_node = rb_prev(parent); | |
142 | *prev_entry = rb_entry_safe(tmp_node, struct rb_entry, rb_node); | |
143 | return NULL; | |
144 | ||
145 | lookup_neighbors: | |
004b6862 | 146 | if (ofs == re->ofs || force) { |
54c2258c CY |
147 | /* lookup prev node for merging backward later */ |
148 | tmp_node = rb_prev(&re->rb_node); | |
149 | *prev_entry = rb_entry_safe(tmp_node, struct rb_entry, rb_node); | |
150 | } | |
004b6862 | 151 | if (ofs == re->ofs + re->len - 1 || force) { |
54c2258c CY |
152 | /* lookup next node for merging frontward later */ |
153 | tmp_node = rb_next(&re->rb_node); | |
154 | *next_entry = rb_entry_safe(tmp_node, struct rb_entry, rb_node); | |
155 | } | |
156 | return re; | |
157 | } | |
158 | ||
4d57b86d | 159 | bool f2fs_check_rb_tree_consistence(struct f2fs_sb_info *sbi, |
df0f6b44 CY |
160 | struct rb_root *root) |
161 | { | |
162 | #ifdef CONFIG_F2FS_CHECK_FS | |
163 | struct rb_node *cur = rb_first(root), *next; | |
164 | struct rb_entry *cur_re, *next_re; | |
165 | ||
166 | if (!cur) | |
167 | return true; | |
168 | ||
169 | while (cur) { | |
170 | next = rb_next(cur); | |
171 | if (!next) | |
172 | return true; | |
173 | ||
174 | cur_re = rb_entry(cur, struct rb_entry, rb_node); | |
175 | next_re = rb_entry(next, struct rb_entry, rb_node); | |
176 | ||
177 | if (cur_re->ofs + cur_re->len > next_re->ofs) { | |
178 | f2fs_msg(sbi->sb, KERN_INFO, "inconsistent rbtree, " | |
179 | "cur(%u, %u) next(%u, %u)", | |
180 | cur_re->ofs, cur_re->len, | |
181 | next_re->ofs, next_re->len); | |
182 | return false; | |
183 | } | |
184 | ||
185 | cur = next; | |
186 | } | |
187 | #endif | |
188 | return true; | |
189 | } | |
190 | ||
a28ef1f5 CY |
191 | static struct kmem_cache *extent_tree_slab; |
192 | static struct kmem_cache *extent_node_slab; | |
193 | ||
194 | static struct extent_node *__attach_extent_node(struct f2fs_sb_info *sbi, | |
195 | struct extent_tree *et, struct extent_info *ei, | |
196 | struct rb_node *parent, struct rb_node **p) | |
197 | { | |
198 | struct extent_node *en; | |
199 | ||
200 | en = kmem_cache_alloc(extent_node_slab, GFP_ATOMIC); | |
201 | if (!en) | |
202 | return NULL; | |
203 | ||
204 | en->ei = *ei; | |
205 | INIT_LIST_HEAD(&en->list); | |
201ef5e0 | 206 | en->et = et; |
a28ef1f5 CY |
207 | |
208 | rb_link_node(&en->rb_node, parent, p); | |
209 | rb_insert_color(&en->rb_node, &et->root); | |
68e35385 | 210 | atomic_inc(&et->node_cnt); |
a28ef1f5 CY |
211 | atomic_inc(&sbi->total_ext_node); |
212 | return en; | |
213 | } | |
214 | ||
215 | static void __detach_extent_node(struct f2fs_sb_info *sbi, | |
216 | struct extent_tree *et, struct extent_node *en) | |
217 | { | |
218 | rb_erase(&en->rb_node, &et->root); | |
68e35385 | 219 | atomic_dec(&et->node_cnt); |
a28ef1f5 CY |
220 | atomic_dec(&sbi->total_ext_node); |
221 | ||
222 | if (et->cached_en == en) | |
223 | et->cached_en = NULL; | |
a03f01f2 HP |
224 | kmem_cache_free(extent_node_slab, en); |
225 | } | |
226 | ||
227 | /* | |
228 | * Flow to release an extent_node: | |
229 | * 1. list_del_init | |
230 | * 2. __detach_extent_node | |
231 | * 3. kmem_cache_free. | |
232 | */ | |
233 | static void __release_extent_node(struct f2fs_sb_info *sbi, | |
234 | struct extent_tree *et, struct extent_node *en) | |
235 | { | |
236 | spin_lock(&sbi->extent_lock); | |
201ef5e0 HP |
237 | f2fs_bug_on(sbi, list_empty(&en->list)); |
238 | list_del_init(&en->list); | |
a03f01f2 HP |
239 | spin_unlock(&sbi->extent_lock); |
240 | ||
241 | __detach_extent_node(sbi, et, en); | |
a28ef1f5 CY |
242 | } |
243 | ||
244 | static struct extent_tree *__grab_extent_tree(struct inode *inode) | |
245 | { | |
246 | struct f2fs_sb_info *sbi = F2FS_I_SB(inode); | |
247 | struct extent_tree *et; | |
248 | nid_t ino = inode->i_ino; | |
249 | ||
5e8256ac | 250 | mutex_lock(&sbi->extent_tree_lock); |
a28ef1f5 CY |
251 | et = radix_tree_lookup(&sbi->extent_tree_root, ino); |
252 | if (!et) { | |
253 | et = f2fs_kmem_cache_alloc(extent_tree_slab, GFP_NOFS); | |
254 | f2fs_radix_tree_insert(&sbi->extent_tree_root, ino, et); | |
255 | memset(et, 0, sizeof(struct extent_tree)); | |
256 | et->ino = ino; | |
257 | et->root = RB_ROOT; | |
258 | et->cached_en = NULL; | |
259 | rwlock_init(&et->lock); | |
137d09f0 | 260 | INIT_LIST_HEAD(&et->list); |
68e35385 | 261 | atomic_set(&et->node_cnt, 0); |
7441ccef | 262 | atomic_inc(&sbi->total_ext_tree); |
74fd8d99 JK |
263 | } else { |
264 | atomic_dec(&sbi->total_zombie_tree); | |
137d09f0 | 265 | list_del_init(&et->list); |
a28ef1f5 | 266 | } |
5e8256ac | 267 | mutex_unlock(&sbi->extent_tree_lock); |
a28ef1f5 CY |
268 | |
269 | /* never died until evict_inode */ | |
270 | F2FS_I(inode)->extent_tree = et; | |
271 | ||
272 | return et; | |
273 | } | |
274 | ||
a6f78345 CY |
275 | static struct extent_node *__init_extent_tree(struct f2fs_sb_info *sbi, |
276 | struct extent_tree *et, struct extent_info *ei) | |
a28ef1f5 CY |
277 | { |
278 | struct rb_node **p = &et->root.rb_node; | |
a28ef1f5 CY |
279 | struct extent_node *en; |
280 | ||
a6f78345 | 281 | en = __attach_extent_node(sbi, et, ei, NULL, p); |
a28ef1f5 CY |
282 | if (!en) |
283 | return NULL; | |
a6f78345 CY |
284 | |
285 | et->largest = en->ei; | |
a28ef1f5 CY |
286 | et->cached_en = en; |
287 | return en; | |
288 | } | |
289 | ||
290 | static unsigned int __free_extent_tree(struct f2fs_sb_info *sbi, | |
201ef5e0 | 291 | struct extent_tree *et) |
a28ef1f5 CY |
292 | { |
293 | struct rb_node *node, *next; | |
294 | struct extent_node *en; | |
68e35385 | 295 | unsigned int count = atomic_read(&et->node_cnt); |
a28ef1f5 CY |
296 | |
297 | node = rb_first(&et->root); | |
298 | while (node) { | |
299 | next = rb_next(node); | |
300 | en = rb_entry(node, struct extent_node, rb_node); | |
201ef5e0 | 301 | __release_extent_node(sbi, et, en); |
a28ef1f5 CY |
302 | node = next; |
303 | } | |
304 | ||
68e35385 | 305 | return count - atomic_read(&et->node_cnt); |
a28ef1f5 CY |
306 | } |
307 | ||
b430f726 | 308 | static void __drop_largest_extent(struct extent_tree *et, |
41a099de | 309 | pgoff_t fofs, unsigned int len) |
a28ef1f5 | 310 | { |
b430f726 ZZ |
311 | if (fofs < et->largest.fofs + et->largest.len && |
312 | fofs + len > et->largest.fofs) { | |
313 | et->largest.len = 0; | |
314 | et->largest_updated = true; | |
205b9822 | 315 | } |
a28ef1f5 CY |
316 | } |
317 | ||
ed3d1256 | 318 | /* return true, if inode page is changed */ |
dad48e73 | 319 | static bool __f2fs_init_extent_tree(struct inode *inode, struct f2fs_extent *i_ext) |
a28ef1f5 CY |
320 | { |
321 | struct f2fs_sb_info *sbi = F2FS_I_SB(inode); | |
322 | struct extent_tree *et; | |
323 | struct extent_node *en; | |
324 | struct extent_info ei; | |
325 | ||
ed3d1256 JK |
326 | if (!f2fs_may_extent_tree(inode)) { |
327 | /* drop largest extent */ | |
328 | if (i_ext && i_ext->len) { | |
329 | i_ext->len = 0; | |
330 | return true; | |
331 | } | |
332 | return false; | |
333 | } | |
a28ef1f5 CY |
334 | |
335 | et = __grab_extent_tree(inode); | |
336 | ||
ed3d1256 JK |
337 | if (!i_ext || !i_ext->len) |
338 | return false; | |
a28ef1f5 | 339 | |
bd933d4f | 340 | get_extent_info(&ei, i_ext); |
a28ef1f5 CY |
341 | |
342 | write_lock(&et->lock); | |
68e35385 | 343 | if (atomic_read(&et->node_cnt)) |
a28ef1f5 CY |
344 | goto out; |
345 | ||
a6f78345 | 346 | en = __init_extent_tree(sbi, et, &ei); |
a28ef1f5 CY |
347 | if (en) { |
348 | spin_lock(&sbi->extent_lock); | |
349 | list_add_tail(&en->list, &sbi->extent_list); | |
350 | spin_unlock(&sbi->extent_lock); | |
351 | } | |
352 | out: | |
353 | write_unlock(&et->lock); | |
ed3d1256 | 354 | return false; |
a28ef1f5 CY |
355 | } |
356 | ||
dad48e73 YH |
357 | bool f2fs_init_extent_tree(struct inode *inode, struct f2fs_extent *i_ext) |
358 | { | |
359 | bool ret = __f2fs_init_extent_tree(inode, i_ext); | |
360 | ||
361 | if (!F2FS_I(inode)->extent_tree) | |
362 | set_inode_flag(inode, FI_NO_EXTENT); | |
363 | ||
364 | return ret; | |
365 | } | |
366 | ||
a28ef1f5 CY |
367 | static bool f2fs_lookup_extent_tree(struct inode *inode, pgoff_t pgofs, |
368 | struct extent_info *ei) | |
369 | { | |
370 | struct f2fs_sb_info *sbi = F2FS_I_SB(inode); | |
371 | struct extent_tree *et = F2FS_I(inode)->extent_tree; | |
372 | struct extent_node *en; | |
373 | bool ret = false; | |
374 | ||
375 | f2fs_bug_on(sbi, !et); | |
376 | ||
377 | trace_f2fs_lookup_extent_tree_start(inode, pgofs); | |
378 | ||
379 | read_lock(&et->lock); | |
380 | ||
381 | if (et->largest.fofs <= pgofs && | |
382 | et->largest.fofs + et->largest.len > pgofs) { | |
383 | *ei = et->largest; | |
384 | ret = true; | |
91c481ff | 385 | stat_inc_largest_node_hit(sbi); |
a28ef1f5 CY |
386 | goto out; |
387 | } | |
388 | ||
4d57b86d | 389 | en = (struct extent_node *)f2fs_lookup_rb_tree(&et->root, |
54c2258c CY |
390 | (struct rb_entry *)et->cached_en, pgofs); |
391 | if (!en) | |
392 | goto out; | |
393 | ||
394 | if (en == et->cached_en) | |
395 | stat_inc_cached_node_hit(sbi); | |
396 | else | |
397 | stat_inc_rbtree_node_hit(sbi); | |
398 | ||
399 | *ei = en->ei; | |
400 | spin_lock(&sbi->extent_lock); | |
401 | if (!list_empty(&en->list)) { | |
402 | list_move_tail(&en->list, &sbi->extent_list); | |
403 | et->cached_en = en; | |
a28ef1f5 | 404 | } |
54c2258c CY |
405 | spin_unlock(&sbi->extent_lock); |
406 | ret = true; | |
a28ef1f5 | 407 | out: |
727edac5 | 408 | stat_inc_total_hit(sbi); |
a28ef1f5 CY |
409 | read_unlock(&et->lock); |
410 | ||
411 | trace_f2fs_lookup_extent_tree_end(inode, pgofs, ei); | |
412 | return ret; | |
413 | } | |
414 | ||
b430f726 | 415 | static struct extent_node *__try_merge_extent_node(struct f2fs_sb_info *sbi, |
0f825ee6 | 416 | struct extent_tree *et, struct extent_info *ei, |
0f825ee6 | 417 | struct extent_node *prev_ex, |
ef05e221 | 418 | struct extent_node *next_ex) |
0f825ee6 | 419 | { |
0f825ee6 | 420 | struct extent_node *en = NULL; |
0f825ee6 FL |
421 | |
422 | if (prev_ex && __is_back_mergeable(ei, &prev_ex->ei)) { | |
0f825ee6 FL |
423 | prev_ex->ei.len += ei->len; |
424 | ei = &prev_ex->ei; | |
425 | en = prev_ex; | |
426 | } | |
ef05e221 | 427 | |
0f825ee6 | 428 | if (next_ex && __is_front_mergeable(ei, &next_ex->ei)) { |
0f825ee6 FL |
429 | next_ex->ei.fofs = ei->fofs; |
430 | next_ex->ei.blk = ei->blk; | |
431 | next_ex->ei.len += ei->len; | |
7855eba4 YH |
432 | if (en) |
433 | __release_extent_node(sbi, et, prev_ex); | |
434 | ||
0f825ee6 FL |
435 | en = next_ex; |
436 | } | |
ef05e221 | 437 | |
43a2fa18 JK |
438 | if (!en) |
439 | return NULL; | |
440 | ||
b430f726 | 441 | __try_update_largest_extent(et, en); |
43a2fa18 JK |
442 | |
443 | spin_lock(&sbi->extent_lock); | |
42926744 | 444 | if (!list_empty(&en->list)) { |
43a2fa18 | 445 | list_move_tail(&en->list, &sbi->extent_list); |
42926744 JK |
446 | et->cached_en = en; |
447 | } | |
43a2fa18 | 448 | spin_unlock(&sbi->extent_lock); |
ef05e221 CY |
449 | return en; |
450 | } | |
451 | ||
b430f726 | 452 | static struct extent_node *__insert_extent_tree(struct f2fs_sb_info *sbi, |
ef05e221 CY |
453 | struct extent_tree *et, struct extent_info *ei, |
454 | struct rb_node **insert_p, | |
455 | struct rb_node *insert_parent) | |
456 | { | |
8fe326cb | 457 | struct rb_node **p; |
ef05e221 CY |
458 | struct rb_node *parent = NULL; |
459 | struct extent_node *en = NULL; | |
0f825ee6 FL |
460 | |
461 | if (insert_p && insert_parent) { | |
462 | parent = insert_parent; | |
463 | p = insert_p; | |
464 | goto do_insert; | |
465 | } | |
466 | ||
4d57b86d | 467 | p = f2fs_lookup_rb_tree_for_insert(sbi, &et->root, &parent, ei->fofs); |
0f825ee6 FL |
468 | do_insert: |
469 | en = __attach_extent_node(sbi, et, ei, parent, p); | |
470 | if (!en) | |
471 | return NULL; | |
ef05e221 | 472 | |
b430f726 | 473 | __try_update_largest_extent(et, en); |
43a2fa18 JK |
474 | |
475 | /* update in global extent list */ | |
476 | spin_lock(&sbi->extent_lock); | |
477 | list_add_tail(&en->list, &sbi->extent_list); | |
42926744 | 478 | et->cached_en = en; |
43a2fa18 | 479 | spin_unlock(&sbi->extent_lock); |
0f825ee6 FL |
480 | return en; |
481 | } | |
482 | ||
317e1300 | 483 | static void f2fs_update_extent_tree_range(struct inode *inode, |
19b2c30d | 484 | pgoff_t fofs, block_t blkaddr, unsigned int len) |
a28ef1f5 CY |
485 | { |
486 | struct f2fs_sb_info *sbi = F2FS_I_SB(inode); | |
487 | struct extent_tree *et = F2FS_I(inode)->extent_tree; | |
4d1fa815 | 488 | struct extent_node *en = NULL, *en1 = NULL; |
19b2c30d | 489 | struct extent_node *prev_en = NULL, *next_en = NULL; |
a28ef1f5 | 490 | struct extent_info ei, dei, prev; |
0f825ee6 | 491 | struct rb_node **insert_p = NULL, *insert_parent = NULL; |
19b2c30d CY |
492 | unsigned int end = fofs + len; |
493 | unsigned int pos = (unsigned int)fofs; | |
b430f726 | 494 | bool updated = false; |
a28ef1f5 CY |
495 | |
496 | if (!et) | |
317e1300 | 497 | return; |
a28ef1f5 | 498 | |
744288c7 CY |
499 | trace_f2fs_update_extent_tree_range(inode, fofs, blkaddr, len); |
500 | ||
a28ef1f5 CY |
501 | write_lock(&et->lock); |
502 | ||
91942321 | 503 | if (is_inode_flag_set(inode, FI_NO_EXTENT)) { |
a28ef1f5 | 504 | write_unlock(&et->lock); |
317e1300 | 505 | return; |
a28ef1f5 CY |
506 | } |
507 | ||
508 | prev = et->largest; | |
509 | dei.len = 0; | |
510 | ||
4d1fa815 FL |
511 | /* |
512 | * drop largest extent before lookup, in case it's already | |
513 | * been shrunk from extent tree | |
514 | */ | |
b430f726 | 515 | __drop_largest_extent(et, fofs, len); |
a28ef1f5 | 516 | |
19b2c30d | 517 | /* 1. lookup first extent node in range [fofs, fofs + len - 1] */ |
4d57b86d | 518 | en = (struct extent_node *)f2fs_lookup_rb_tree_ret(&et->root, |
54c2258c CY |
519 | (struct rb_entry *)et->cached_en, fofs, |
520 | (struct rb_entry **)&prev_en, | |
521 | (struct rb_entry **)&next_en, | |
004b6862 | 522 | &insert_p, &insert_parent, false); |
4d1fa815 FL |
523 | if (!en) |
524 | en = next_en; | |
19b2c30d CY |
525 | |
526 | /* 2. invlidate all extent nodes in range [fofs, fofs + len - 1] */ | |
4d1fa815 FL |
527 | while (en && en->ei.fofs < end) { |
528 | unsigned int org_end; | |
529 | int parts = 0; /* # of parts current extent split into */ | |
19b2c30d | 530 | |
4d1fa815 | 531 | next_en = en1 = NULL; |
19b2c30d CY |
532 | |
533 | dei = en->ei; | |
4d1fa815 FL |
534 | org_end = dei.fofs + dei.len; |
535 | f2fs_bug_on(sbi, pos >= org_end); | |
19b2c30d | 536 | |
4d1fa815 FL |
537 | if (pos > dei.fofs && pos - dei.fofs >= F2FS_MIN_EXTENT_LEN) { |
538 | en->ei.len = pos - en->ei.fofs; | |
539 | prev_en = en; | |
540 | parts = 1; | |
541 | } | |
19b2c30d | 542 | |
4d1fa815 FL |
543 | if (end < org_end && org_end - end >= F2FS_MIN_EXTENT_LEN) { |
544 | if (parts) { | |
545 | set_extent_info(&ei, end, | |
546 | end - dei.fofs + dei.blk, | |
547 | org_end - end); | |
b430f726 | 548 | en1 = __insert_extent_tree(sbi, et, &ei, |
4d1fa815 FL |
549 | NULL, NULL); |
550 | next_en = en1; | |
551 | } else { | |
552 | en->ei.fofs = end; | |
553 | en->ei.blk += end - dei.fofs; | |
554 | en->ei.len -= end - dei.fofs; | |
555 | next_en = en; | |
19b2c30d | 556 | } |
4d1fa815 | 557 | parts++; |
19b2c30d CY |
558 | } |
559 | ||
4d1fa815 FL |
560 | if (!next_en) { |
561 | struct rb_node *node = rb_next(&en->rb_node); | |
19b2c30d | 562 | |
ed0b5620 GT |
563 | next_en = rb_entry_safe(node, struct extent_node, |
564 | rb_node); | |
a28ef1f5 CY |
565 | } |
566 | ||
4abd3f5a | 567 | if (parts) |
b430f726 | 568 | __try_update_largest_extent(et, en); |
4abd3f5a | 569 | else |
a03f01f2 | 570 | __release_extent_node(sbi, et, en); |
19b2c30d CY |
571 | |
572 | /* | |
4d1fa815 FL |
573 | * if original extent is split into zero or two parts, extent |
574 | * tree has been altered by deletion or insertion, therefore | |
575 | * invalidate pointers regard to tree. | |
19b2c30d | 576 | */ |
4d1fa815 FL |
577 | if (parts != 1) { |
578 | insert_p = NULL; | |
579 | insert_parent = NULL; | |
a28ef1f5 | 580 | } |
4d1fa815 | 581 | en = next_en; |
a28ef1f5 CY |
582 | } |
583 | ||
a28ef1f5 CY |
584 | /* 3. update extent in extent cache */ |
585 | if (blkaddr) { | |
19b2c30d CY |
586 | |
587 | set_extent_info(&ei, fofs, blkaddr, len); | |
b430f726 ZZ |
588 | if (!__try_merge_extent_node(sbi, et, &ei, prev_en, next_en)) |
589 | __insert_extent_tree(sbi, et, &ei, | |
ef05e221 | 590 | insert_p, insert_parent); |
a28ef1f5 CY |
591 | |
592 | /* give up extent_cache, if split and small updates happen */ | |
593 | if (dei.len >= 1 && | |
594 | prev.len < F2FS_MIN_EXTENT_LEN && | |
595 | et->largest.len < F2FS_MIN_EXTENT_LEN) { | |
b430f726 ZZ |
596 | et->largest.len = 0; |
597 | et->largest_updated = true; | |
91942321 | 598 | set_inode_flag(inode, FI_NO_EXTENT); |
a28ef1f5 | 599 | } |
19b2c30d | 600 | } |
a28ef1f5 | 601 | |
91942321 | 602 | if (is_inode_flag_set(inode, FI_NO_EXTENT)) |
201ef5e0 | 603 | __free_extent_tree(sbi, et); |
a28ef1f5 | 604 | |
b430f726 ZZ |
605 | if (et->largest_updated) { |
606 | et->largest_updated = false; | |
607 | updated = true; | |
608 | } | |
609 | ||
a28ef1f5 | 610 | write_unlock(&et->lock); |
b430f726 ZZ |
611 | |
612 | if (updated) | |
613 | f2fs_mark_inode_dirty_sync(inode, true); | |
a28ef1f5 CY |
614 | } |
615 | ||
616 | unsigned int f2fs_shrink_extent_tree(struct f2fs_sb_info *sbi, int nr_shrink) | |
617 | { | |
137d09f0 | 618 | struct extent_tree *et, *next; |
201ef5e0 | 619 | struct extent_node *en; |
a28ef1f5 CY |
620 | unsigned int node_cnt = 0, tree_cnt = 0; |
621 | int remained; | |
622 | ||
623 | if (!test_opt(sbi, EXTENT_CACHE)) | |
624 | return 0; | |
625 | ||
74fd8d99 JK |
626 | if (!atomic_read(&sbi->total_zombie_tree)) |
627 | goto free_node; | |
628 | ||
5e8256ac | 629 | if (!mutex_trylock(&sbi->extent_tree_lock)) |
a28ef1f5 CY |
630 | goto out; |
631 | ||
632 | /* 1. remove unreferenced extent tree */ | |
137d09f0 | 633 | list_for_each_entry_safe(et, next, &sbi->zombie_list, list) { |
9b72a388 CY |
634 | if (atomic_read(&et->node_cnt)) { |
635 | write_lock(&et->lock); | |
201ef5e0 | 636 | node_cnt += __free_extent_tree(sbi, et); |
9b72a388 CY |
637 | write_unlock(&et->lock); |
638 | } | |
201ef5e0 | 639 | f2fs_bug_on(sbi, atomic_read(&et->node_cnt)); |
137d09f0 JK |
640 | list_del_init(&et->list); |
641 | radix_tree_delete(&sbi->extent_tree_root, et->ino); | |
642 | kmem_cache_free(extent_tree_slab, et); | |
643 | atomic_dec(&sbi->total_ext_tree); | |
644 | atomic_dec(&sbi->total_zombie_tree); | |
645 | tree_cnt++; | |
a28ef1f5 | 646 | |
137d09f0 JK |
647 | if (node_cnt + tree_cnt >= nr_shrink) |
648 | goto unlock_out; | |
6fe2bc95 | 649 | cond_resched(); |
a28ef1f5 | 650 | } |
5e8256ac | 651 | mutex_unlock(&sbi->extent_tree_lock); |
a28ef1f5 | 652 | |
74fd8d99 | 653 | free_node: |
a28ef1f5 | 654 | /* 2. remove LRU extent entries */ |
5e8256ac | 655 | if (!mutex_trylock(&sbi->extent_tree_lock)) |
a28ef1f5 CY |
656 | goto out; |
657 | ||
658 | remained = nr_shrink - (node_cnt + tree_cnt); | |
659 | ||
660 | spin_lock(&sbi->extent_lock); | |
201ef5e0 HP |
661 | for (; remained > 0; remained--) { |
662 | if (list_empty(&sbi->extent_list)) | |
a28ef1f5 | 663 | break; |
201ef5e0 HP |
664 | en = list_first_entry(&sbi->extent_list, |
665 | struct extent_node, list); | |
666 | et = en->et; | |
667 | if (!write_trylock(&et->lock)) { | |
668 | /* refresh this extent node's position in extent list */ | |
669 | list_move_tail(&en->list, &sbi->extent_list); | |
670 | continue; | |
671 | } | |
a28ef1f5 | 672 | |
201ef5e0 HP |
673 | list_del_init(&en->list); |
674 | spin_unlock(&sbi->extent_lock); | |
9b72a388 | 675 | |
201ef5e0 | 676 | __detach_extent_node(sbi, et, en); |
a28ef1f5 | 677 | |
201ef5e0 HP |
678 | write_unlock(&et->lock); |
679 | node_cnt++; | |
680 | spin_lock(&sbi->extent_lock); | |
a28ef1f5 | 681 | } |
201ef5e0 HP |
682 | spin_unlock(&sbi->extent_lock); |
683 | ||
a28ef1f5 | 684 | unlock_out: |
5e8256ac | 685 | mutex_unlock(&sbi->extent_tree_lock); |
a28ef1f5 CY |
686 | out: |
687 | trace_f2fs_shrink_extent_tree(sbi, node_cnt, tree_cnt); | |
688 | ||
689 | return node_cnt + tree_cnt; | |
690 | } | |
691 | ||
692 | unsigned int f2fs_destroy_extent_node(struct inode *inode) | |
693 | { | |
694 | struct f2fs_sb_info *sbi = F2FS_I_SB(inode); | |
695 | struct extent_tree *et = F2FS_I(inode)->extent_tree; | |
696 | unsigned int node_cnt = 0; | |
697 | ||
9b72a388 | 698 | if (!et || !atomic_read(&et->node_cnt)) |
a28ef1f5 CY |
699 | return 0; |
700 | ||
701 | write_lock(&et->lock); | |
201ef5e0 | 702 | node_cnt = __free_extent_tree(sbi, et); |
a28ef1f5 CY |
703 | write_unlock(&et->lock); |
704 | ||
705 | return node_cnt; | |
706 | } | |
707 | ||
5f281fab JK |
708 | void f2fs_drop_extent_tree(struct inode *inode) |
709 | { | |
710 | struct f2fs_sb_info *sbi = F2FS_I_SB(inode); | |
711 | struct extent_tree *et = F2FS_I(inode)->extent_tree; | |
b430f726 | 712 | bool updated = false; |
5f281fab | 713 | |
bf617f7a CY |
714 | if (!f2fs_may_extent_tree(inode)) |
715 | return; | |
716 | ||
5f281fab JK |
717 | set_inode_flag(inode, FI_NO_EXTENT); |
718 | ||
719 | write_lock(&et->lock); | |
720 | __free_extent_tree(sbi, et); | |
b430f726 ZZ |
721 | if (et->largest.len) { |
722 | et->largest.len = 0; | |
723 | updated = true; | |
724 | } | |
5f281fab | 725 | write_unlock(&et->lock); |
b430f726 ZZ |
726 | if (updated) |
727 | f2fs_mark_inode_dirty_sync(inode, true); | |
5f281fab JK |
728 | } |
729 | ||
a28ef1f5 CY |
730 | void f2fs_destroy_extent_tree(struct inode *inode) |
731 | { | |
732 | struct f2fs_sb_info *sbi = F2FS_I_SB(inode); | |
733 | struct extent_tree *et = F2FS_I(inode)->extent_tree; | |
734 | unsigned int node_cnt = 0; | |
735 | ||
736 | if (!et) | |
737 | return; | |
738 | ||
68e35385 CY |
739 | if (inode->i_nlink && !is_bad_inode(inode) && |
740 | atomic_read(&et->node_cnt)) { | |
5e8256ac | 741 | mutex_lock(&sbi->extent_tree_lock); |
137d09f0 | 742 | list_add_tail(&et->list, &sbi->zombie_list); |
74fd8d99 | 743 | atomic_inc(&sbi->total_zombie_tree); |
5e8256ac | 744 | mutex_unlock(&sbi->extent_tree_lock); |
a28ef1f5 CY |
745 | return; |
746 | } | |
747 | ||
748 | /* free all extent info belong to this extent tree */ | |
749 | node_cnt = f2fs_destroy_extent_node(inode); | |
750 | ||
751 | /* delete extent tree entry in radix tree */ | |
5e8256ac | 752 | mutex_lock(&sbi->extent_tree_lock); |
68e35385 | 753 | f2fs_bug_on(sbi, atomic_read(&et->node_cnt)); |
a28ef1f5 CY |
754 | radix_tree_delete(&sbi->extent_tree_root, inode->i_ino); |
755 | kmem_cache_free(extent_tree_slab, et); | |
7441ccef | 756 | atomic_dec(&sbi->total_ext_tree); |
5e8256ac | 757 | mutex_unlock(&sbi->extent_tree_lock); |
a28ef1f5 CY |
758 | |
759 | F2FS_I(inode)->extent_tree = NULL; | |
760 | ||
761 | trace_f2fs_destroy_extent_tree(inode, node_cnt); | |
762 | } | |
763 | ||
764 | bool f2fs_lookup_extent_cache(struct inode *inode, pgoff_t pgofs, | |
765 | struct extent_info *ei) | |
766 | { | |
767 | if (!f2fs_may_extent_tree(inode)) | |
768 | return false; | |
769 | ||
770 | return f2fs_lookup_extent_tree(inode, pgofs, ei); | |
771 | } | |
772 | ||
773 | void f2fs_update_extent_cache(struct dnode_of_data *dn) | |
774 | { | |
a28ef1f5 | 775 | pgoff_t fofs; |
f28b3434 | 776 | block_t blkaddr; |
a28ef1f5 CY |
777 | |
778 | if (!f2fs_may_extent_tree(dn->inode)) | |
779 | return; | |
780 | ||
f28b3434 CY |
781 | if (dn->data_blkaddr == NEW_ADDR) |
782 | blkaddr = NULL_ADDR; | |
783 | else | |
784 | blkaddr = dn->data_blkaddr; | |
19b2c30d | 785 | |
4d57b86d | 786 | fofs = f2fs_start_bidx_of_node(ofs_of_node(dn->node_page), dn->inode) + |
81ca7350 | 787 | dn->ofs_in_node; |
ee6d182f | 788 | f2fs_update_extent_tree_range(dn->inode, fofs, blkaddr, 1); |
19b2c30d CY |
789 | } |
790 | ||
791 | void f2fs_update_extent_cache_range(struct dnode_of_data *dn, | |
792 | pgoff_t fofs, block_t blkaddr, unsigned int len) | |
793 | ||
794 | { | |
795 | if (!f2fs_may_extent_tree(dn->inode)) | |
796 | return; | |
797 | ||
ee6d182f | 798 | f2fs_update_extent_tree_range(dn->inode, fofs, blkaddr, len); |
a28ef1f5 CY |
799 | } |
800 | ||
4d57b86d | 801 | void f2fs_init_extent_cache_info(struct f2fs_sb_info *sbi) |
a28ef1f5 CY |
802 | { |
803 | INIT_RADIX_TREE(&sbi->extent_tree_root, GFP_NOIO); | |
5e8256ac | 804 | mutex_init(&sbi->extent_tree_lock); |
a28ef1f5 CY |
805 | INIT_LIST_HEAD(&sbi->extent_list); |
806 | spin_lock_init(&sbi->extent_lock); | |
7441ccef | 807 | atomic_set(&sbi->total_ext_tree, 0); |
137d09f0 | 808 | INIT_LIST_HEAD(&sbi->zombie_list); |
74fd8d99 | 809 | atomic_set(&sbi->total_zombie_tree, 0); |
a28ef1f5 CY |
810 | atomic_set(&sbi->total_ext_node, 0); |
811 | } | |
812 | ||
4d57b86d | 813 | int __init f2fs_create_extent_cache(void) |
a28ef1f5 CY |
814 | { |
815 | extent_tree_slab = f2fs_kmem_cache_create("f2fs_extent_tree", | |
816 | sizeof(struct extent_tree)); | |
817 | if (!extent_tree_slab) | |
818 | return -ENOMEM; | |
819 | extent_node_slab = f2fs_kmem_cache_create("f2fs_extent_node", | |
820 | sizeof(struct extent_node)); | |
821 | if (!extent_node_slab) { | |
822 | kmem_cache_destroy(extent_tree_slab); | |
823 | return -ENOMEM; | |
824 | } | |
825 | return 0; | |
826 | } | |
827 | ||
4d57b86d | 828 | void f2fs_destroy_extent_cache(void) |
a28ef1f5 CY |
829 | { |
830 | kmem_cache_destroy(extent_node_slab); | |
831 | kmem_cache_destroy(extent_tree_slab); | |
832 | } |