Merge branch 'next' into for-linus
[linux-block.git] / fs / xfs / xfs_refcount_item.c
CommitLineData
0b61f8a4 1// SPDX-License-Identifier: GPL-2.0+
baf4bcac
DW
2/*
3 * Copyright (C) 2016 Oracle. All Rights Reserved.
baf4bcac 4 * Author: Darrick J. Wong <darrick.wong@oracle.com>
baf4bcac
DW
5 */
6#include "xfs.h"
7#include "xfs_fs.h"
8#include "xfs_format.h"
9#include "xfs_log_format.h"
10#include "xfs_trans_resv.h"
f997ee21 11#include "xfs_bit.h"
b31c2bdc 12#include "xfs_shared.h"
baf4bcac 13#include "xfs_mount.h"
f997ee21 14#include "xfs_defer.h"
baf4bcac
DW
15#include "xfs_trans.h"
16#include "xfs_trans_priv.h"
baf4bcac
DW
17#include "xfs_refcount_item.h"
18#include "xfs_log.h"
f997ee21 19#include "xfs_refcount.h"
a5155b87 20#include "xfs_error.h"
9b4467e9 21#include "xfs_log_priv.h"
86ffa471 22#include "xfs_log_recover.h"
baf4bcac 23
182696fb
DW
24struct kmem_cache *xfs_cui_cache;
25struct kmem_cache *xfs_cud_cache;
baf4bcac 26
c57ed2f5
DW
27static const struct xfs_item_ops xfs_cui_item_ops;
28
baf4bcac
DW
29static inline struct xfs_cui_log_item *CUI_ITEM(struct xfs_log_item *lip)
30{
31 return container_of(lip, struct xfs_cui_log_item, cui_item);
32}
33
9b4467e9 34STATIC void
baf4bcac
DW
35xfs_cui_item_free(
36 struct xfs_cui_log_item *cuip)
37{
c230a4a8 38 kmem_free(cuip->cui_item.li_lv_shadow);
baf4bcac
DW
39 if (cuip->cui_format.cui_nextents > XFS_CUI_MAX_FAST_EXTENTS)
40 kmem_free(cuip);
41 else
182696fb 42 kmem_cache_free(xfs_cui_cache, cuip);
baf4bcac
DW
43}
44
0612d116
DC
45/*
46 * Freeing the CUI requires that we remove it from the AIL if it has already
47 * been placed there. However, the CUI may not yet have been placed in the AIL
48 * when called by xfs_cui_release() from CUD processing due to the ordering of
49 * committed vs unpin operations in bulk insert operations. Hence the reference
50 * count to ensure only the last caller frees the CUI.
51 */
c57ed2f5 52STATIC void
0612d116
DC
53xfs_cui_release(
54 struct xfs_cui_log_item *cuip)
55{
56 ASSERT(atomic_read(&cuip->cui_refcount) > 0);
3512fc1e
DC
57 if (!atomic_dec_and_test(&cuip->cui_refcount))
58 return;
59
60 xfs_trans_ail_delete(&cuip->cui_item, 0);
61 xfs_cui_item_free(cuip);
0612d116
DC
62}
63
64
baf4bcac
DW
65STATIC void
66xfs_cui_item_size(
67 struct xfs_log_item *lip,
68 int *nvecs,
69 int *nbytes)
70{
71 struct xfs_cui_log_item *cuip = CUI_ITEM(lip);
72
73 *nvecs += 1;
74 *nbytes += xfs_cui_log_format_sizeof(cuip->cui_format.cui_nextents);
75}
76
77/*
78 * This is called to fill in the vector of log iovecs for the
79 * given cui log item. We use only 1 iovec, and we point that
80 * at the cui_log_format structure embedded in the cui item.
81 * It is at this point that we assert that all of the extent
82 * slots in the cui item have been filled.
83 */
84STATIC void
85xfs_cui_item_format(
86 struct xfs_log_item *lip,
87 struct xfs_log_vec *lv)
88{
89 struct xfs_cui_log_item *cuip = CUI_ITEM(lip);
90 struct xfs_log_iovec *vecp = NULL;
91
92 ASSERT(atomic_read(&cuip->cui_next_extent) ==
93 cuip->cui_format.cui_nextents);
94
95 cuip->cui_format.cui_type = XFS_LI_CUI;
96 cuip->cui_format.cui_size = 1;
97
98 xlog_copy_iovec(lv, &vecp, XLOG_REG_TYPE_CUI_FORMAT, &cuip->cui_format,
99 xfs_cui_log_format_sizeof(cuip->cui_format.cui_nextents));
100}
101
baf4bcac
DW
102/*
103 * The unpin operation is the last place an CUI is manipulated in the log. It is
104 * either inserted in the AIL or aborted in the event of a log I/O error. In
105 * either case, the CUI transaction has been successfully committed to make it
106 * this far. Therefore, we expect whoever committed the CUI to either construct
107 * and commit the CUD or drop the CUD's reference in the event of error. Simply
108 * drop the log's CUI reference now that the log is done with it.
109 */
110STATIC void
111xfs_cui_item_unpin(
112 struct xfs_log_item *lip,
113 int remove)
114{
115 struct xfs_cui_log_item *cuip = CUI_ITEM(lip);
116
117 xfs_cui_release(cuip);
118}
119
baf4bcac
DW
120/*
121 * The CUI has been either committed or aborted if the transaction has been
122 * cancelled. If the transaction was cancelled, an CUD isn't going to be
123 * constructed and thus we free the CUI here directly.
124 */
125STATIC void
ddf92053 126xfs_cui_item_release(
baf4bcac
DW
127 struct xfs_log_item *lip)
128{
ddf92053 129 xfs_cui_release(CUI_ITEM(lip));
baf4bcac
DW
130}
131
baf4bcac
DW
132/*
133 * Allocate and initialize an cui item with the given number of extents.
134 */
9b4467e9 135STATIC struct xfs_cui_log_item *
baf4bcac
DW
136xfs_cui_init(
137 struct xfs_mount *mp,
138 uint nextents)
139
140{
141 struct xfs_cui_log_item *cuip;
142
143 ASSERT(nextents > 0);
144 if (nextents > XFS_CUI_MAX_FAST_EXTENTS)
145 cuip = kmem_zalloc(xfs_cui_log_item_sizeof(nextents),
707e0dda 146 0);
baf4bcac 147 else
182696fb 148 cuip = kmem_cache_zalloc(xfs_cui_cache,
32a2b11f 149 GFP_KERNEL | __GFP_NOFAIL);
baf4bcac
DW
150
151 xfs_log_item_init(mp, &cuip->cui_item, XFS_LI_CUI, &xfs_cui_item_ops);
152 cuip->cui_format.cui_nextents = nextents;
153 cuip->cui_format.cui_id = (uintptr_t)(void *)cuip;
154 atomic_set(&cuip->cui_next_extent, 0);
155 atomic_set(&cuip->cui_refcount, 2);
156
157 return cuip;
158}
159
baf4bcac
DW
160static inline struct xfs_cud_log_item *CUD_ITEM(struct xfs_log_item *lip)
161{
162 return container_of(lip, struct xfs_cud_log_item, cud_item);
163}
164
165STATIC void
166xfs_cud_item_size(
167 struct xfs_log_item *lip,
168 int *nvecs,
169 int *nbytes)
170{
171 *nvecs += 1;
172 *nbytes += sizeof(struct xfs_cud_log_format);
173}
174
175/*
176 * This is called to fill in the vector of log iovecs for the
177 * given cud log item. We use only 1 iovec, and we point that
178 * at the cud_log_format structure embedded in the cud item.
179 * It is at this point that we assert that all of the extent
180 * slots in the cud item have been filled.
181 */
182STATIC void
183xfs_cud_item_format(
184 struct xfs_log_item *lip,
185 struct xfs_log_vec *lv)
186{
187 struct xfs_cud_log_item *cudp = CUD_ITEM(lip);
188 struct xfs_log_iovec *vecp = NULL;
189
190 cudp->cud_format.cud_type = XFS_LI_CUD;
191 cudp->cud_format.cud_size = 1;
192
193 xlog_copy_iovec(lv, &vecp, XLOG_REG_TYPE_CUD_FORMAT, &cudp->cud_format,
194 sizeof(struct xfs_cud_log_format));
195}
196
baf4bcac
DW
197/*
198 * The CUD is either committed or aborted if the transaction is cancelled. If
199 * the transaction is cancelled, drop our reference to the CUI and free the
200 * CUD.
201 */
202STATIC void
ddf92053 203xfs_cud_item_release(
baf4bcac
DW
204 struct xfs_log_item *lip)
205{
206 struct xfs_cud_log_item *cudp = CUD_ITEM(lip);
207
ddf92053 208 xfs_cui_release(cudp->cud_cuip);
c230a4a8 209 kmem_free(cudp->cud_item.li_lv_shadow);
182696fb 210 kmem_cache_free(xfs_cud_cache, cudp);
baf4bcac
DW
211}
212
c23ab603
DC
213static struct xfs_log_item *
214xfs_cud_item_intent(
215 struct xfs_log_item *lip)
216{
217 return &CUD_ITEM(lip)->cud_cuip->cui_item;
218}
219
baf4bcac 220static const struct xfs_item_ops xfs_cud_item_ops = {
f5b81200
DC
221 .flags = XFS_ITEM_RELEASE_WHEN_COMMITTED |
222 XFS_ITEM_INTENT_DONE,
baf4bcac
DW
223 .iop_size = xfs_cud_item_size,
224 .iop_format = xfs_cud_item_format,
ddf92053 225 .iop_release = xfs_cud_item_release,
c23ab603 226 .iop_intent = xfs_cud_item_intent,
baf4bcac
DW
227};
228
effd5e96 229static struct xfs_cud_log_item *
ebeb8e06
CH
230xfs_trans_get_cud(
231 struct xfs_trans *tp,
baf4bcac 232 struct xfs_cui_log_item *cuip)
baf4bcac 233{
ebeb8e06 234 struct xfs_cud_log_item *cudp;
baf4bcac 235
182696fb 236 cudp = kmem_cache_zalloc(xfs_cud_cache, GFP_KERNEL | __GFP_NOFAIL);
ebeb8e06
CH
237 xfs_log_item_init(tp->t_mountp, &cudp->cud_item, XFS_LI_CUD,
238 &xfs_cud_item_ops);
baf4bcac
DW
239 cudp->cud_cuip = cuip;
240 cudp->cud_format.cud_cui_id = cuip->cui_format.cui_id;
241
ebeb8e06 242 xfs_trans_add_item(tp, &cudp->cud_item);
baf4bcac
DW
243 return cudp;
244}
f997ee21 245
effd5e96
CH
246/*
247 * Finish an refcount update and log it to the CUD. Note that the
248 * transaction is marked dirty regardless of whether the refcount
249 * update succeeds or fails to support the CUI/CUD lifecycle rules.
250 */
251static int
252xfs_trans_log_finish_refcount_update(
253 struct xfs_trans *tp,
254 struct xfs_cud_log_item *cudp,
0b11553e 255 struct xfs_refcount_intent *ri,
effd5e96
CH
256 struct xfs_btree_cur **pcur)
257{
258 int error;
259
0b11553e 260 error = xfs_refcount_finish_one(tp, ri, pcur);
effd5e96
CH
261
262 /*
263 * Mark the transaction dirty, even on error. This ensures the
264 * transaction is aborted, which:
265 *
266 * 1.) releases the CUI and frees the CUD
267 * 2.) shuts down the filesystem
268 */
bb7b1c9c 269 tp->t_flags |= XFS_TRANS_DIRTY | XFS_TRANS_HAS_INTENT_DONE;
effd5e96
CH
270 set_bit(XFS_LI_DIRTY, &cudp->cud_item.li_flags);
271
272 return error;
273}
274
275/* Sort refcount intents by AG. */
276static int
277xfs_refcount_update_diff_items(
278 void *priv,
4f0f586b
ST
279 const struct list_head *a,
280 const struct list_head *b)
effd5e96
CH
281{
282 struct xfs_mount *mp = priv;
283 struct xfs_refcount_intent *ra;
284 struct xfs_refcount_intent *rb;
285
286 ra = container_of(a, struct xfs_refcount_intent, ri_list);
287 rb = container_of(b, struct xfs_refcount_intent, ri_list);
288 return XFS_FSB_TO_AGNO(mp, ra->ri_startblock) -
289 XFS_FSB_TO_AGNO(mp, rb->ri_startblock);
290}
291
effd5e96
CH
292/* Set the phys extent flags for this reverse mapping. */
293static void
294xfs_trans_set_refcount_flags(
01a3af22 295 struct xfs_phys_extent *pmap,
effd5e96
CH
296 enum xfs_refcount_intent_type type)
297{
01a3af22 298 pmap->pe_flags = 0;
effd5e96
CH
299 switch (type) {
300 case XFS_REFCOUNT_INCREASE:
301 case XFS_REFCOUNT_DECREASE:
302 case XFS_REFCOUNT_ALLOC_COW:
303 case XFS_REFCOUNT_FREE_COW:
01a3af22 304 pmap->pe_flags |= type;
effd5e96
CH
305 break;
306 default:
307 ASSERT(0);
308 }
309}
310
311/* Log refcount updates in the intent item. */
312STATIC void
313xfs_refcount_update_log_item(
314 struct xfs_trans *tp,
c1f09188 315 struct xfs_cui_log_item *cuip,
01a3af22 316 struct xfs_refcount_intent *ri)
effd5e96 317{
effd5e96 318 uint next_extent;
01a3af22 319 struct xfs_phys_extent *pmap;
effd5e96 320
effd5e96
CH
321 tp->t_flags |= XFS_TRANS_DIRTY;
322 set_bit(XFS_LI_DIRTY, &cuip->cui_item.li_flags);
323
324 /*
325 * atomic_inc_return gives us the value after the increment;
326 * we want to use it as an array index so we need to subtract 1 from
327 * it.
328 */
329 next_extent = atomic_inc_return(&cuip->cui_next_extent) - 1;
330 ASSERT(next_extent < cuip->cui_format.cui_nextents);
01a3af22
DW
331 pmap = &cuip->cui_format.cui_extents[next_extent];
332 pmap->pe_startblock = ri->ri_startblock;
333 pmap->pe_len = ri->ri_blockcount;
334 xfs_trans_set_refcount_flags(pmap, ri->ri_type);
effd5e96
CH
335}
336
13a83333 337static struct xfs_log_item *
c1f09188
CH
338xfs_refcount_update_create_intent(
339 struct xfs_trans *tp,
340 struct list_head *items,
d367a868
CH
341 unsigned int count,
342 bool sort)
c1f09188
CH
343{
344 struct xfs_mount *mp = tp->t_mountp;
345 struct xfs_cui_log_item *cuip = xfs_cui_init(mp, count);
01a3af22 346 struct xfs_refcount_intent *ri;
c1f09188
CH
347
348 ASSERT(count > 0);
349
350 xfs_trans_add_item(tp, &cuip->cui_item);
d367a868
CH
351 if (sort)
352 list_sort(mp, items, xfs_refcount_update_diff_items);
01a3af22
DW
353 list_for_each_entry(ri, items, ri_list)
354 xfs_refcount_update_log_item(tp, cuip, ri);
13a83333 355 return &cuip->cui_item;
c1f09188
CH
356}
357
effd5e96 358/* Get an CUD so we can process all the deferred refcount updates. */
f09d167c 359static struct xfs_log_item *
effd5e96
CH
360xfs_refcount_update_create_done(
361 struct xfs_trans *tp,
13a83333 362 struct xfs_log_item *intent,
effd5e96
CH
363 unsigned int count)
364{
f09d167c 365 return &xfs_trans_get_cud(tp, CUI_ITEM(intent))->cud_item;
effd5e96
CH
366}
367
368/* Process a deferred refcount update. */
369STATIC int
370xfs_refcount_update_finish_item(
371 struct xfs_trans *tp,
f09d167c 372 struct xfs_log_item *done,
effd5e96 373 struct list_head *item,
3ec1b26c 374 struct xfs_btree_cur **state)
effd5e96 375{
0b11553e 376 struct xfs_refcount_intent *ri;
effd5e96
CH
377 int error;
378
0b11553e
DW
379 ri = container_of(item, struct xfs_refcount_intent, ri_list);
380 error = xfs_trans_log_finish_refcount_update(tp, CUD_ITEM(done), ri,
381 state);
3ec1b26c 382
effd5e96 383 /* Did we run out of reservation? Requeue what we didn't finish. */
0b11553e
DW
384 if (!error && ri->ri_blockcount > 0) {
385 ASSERT(ri->ri_type == XFS_REFCOUNT_INCREASE ||
386 ri->ri_type == XFS_REFCOUNT_DECREASE);
effd5e96
CH
387 return -EAGAIN;
388 }
0b11553e 389 kmem_cache_free(xfs_refcount_intent_cache, ri);
effd5e96
CH
390 return error;
391}
392
effd5e96
CH
393/* Abort all pending CUIs. */
394STATIC void
395xfs_refcount_update_abort_intent(
13a83333 396 struct xfs_log_item *intent)
effd5e96 397{
13a83333 398 xfs_cui_release(CUI_ITEM(intent));
effd5e96
CH
399}
400
401/* Cancel a deferred refcount update. */
402STATIC void
403xfs_refcount_update_cancel_item(
404 struct list_head *item)
405{
01a3af22 406 struct xfs_refcount_intent *ri;
effd5e96 407
01a3af22
DW
408 ri = container_of(item, struct xfs_refcount_intent, ri_list);
409 kmem_cache_free(xfs_refcount_intent_cache, ri);
effd5e96
CH
410}
411
412const struct xfs_defer_op_type xfs_refcount_update_defer_type = {
413 .max_items = XFS_CUI_MAX_FAST_EXTENTS,
effd5e96
CH
414 .create_intent = xfs_refcount_update_create_intent,
415 .abort_intent = xfs_refcount_update_abort_intent,
effd5e96
CH
416 .create_done = xfs_refcount_update_create_done,
417 .finish_item = xfs_refcount_update_finish_item,
3ec1b26c 418 .finish_cleanup = xfs_refcount_finish_one_cleanup,
effd5e96
CH
419 .cancel_item = xfs_refcount_update_cancel_item,
420};
421
ed64f834
DW
422/* Is this recovered CUI ok? */
423static inline bool
424xfs_cui_validate_phys(
425 struct xfs_mount *mp,
01a3af22 426 struct xfs_phys_extent *pmap)
ed64f834 427{
38c26bfd 428 if (!xfs_has_reflink(mp))
da5de110
DW
429 return false;
430
01a3af22 431 if (pmap->pe_flags & ~XFS_REFCOUNT_EXTENT_FLAGS)
0d79781a 432 return false;
ed64f834 433
01a3af22 434 switch (pmap->pe_flags & XFS_REFCOUNT_EXTENT_TYPE_MASK) {
ed64f834
DW
435 case XFS_REFCOUNT_INCREASE:
436 case XFS_REFCOUNT_DECREASE:
437 case XFS_REFCOUNT_ALLOC_COW:
438 case XFS_REFCOUNT_FREE_COW:
ed64f834
DW
439 break;
440 default:
0d79781a 441 return false;
ed64f834 442 }
0d79781a 443
01a3af22 444 return xfs_verify_fsbext(mp, pmap->pe_startblock, pmap->pe_len);
ed64f834
DW
445}
446
f997ee21
DW
447/*
448 * Process a refcount update intent item that was recovered from the log.
449 * We need to update the refcountbt.
450 */
c57ed2f5 451STATIC int
96b60f82
DW
452xfs_cui_item_recover(
453 struct xfs_log_item *lip,
e6fff81e 454 struct list_head *capture_list)
f997ee21 455{
96b60f82 456 struct xfs_cui_log_item *cuip = CUI_ITEM(lip);
33ba6129
DW
457 struct xfs_cud_log_item *cudp;
458 struct xfs_trans *tp;
459 struct xfs_btree_cur *rcur = NULL;
d86142dd 460 struct xfs_mount *mp = lip->li_log->l_mp;
96b60f82 461 unsigned int refc_type;
33ba6129 462 bool requeue_only = false;
96b60f82
DW
463 int i;
464 int error = 0;
f997ee21 465
f997ee21
DW
466 /*
467 * First check the validity of the extents described by the
468 * CUI. If any are bad, then assume that all are bad and
469 * just toss the CUI.
470 */
471 for (i = 0; i < cuip->cui_format.cui_nextents; i++) {
ed64f834
DW
472 if (!xfs_cui_validate_phys(mp,
473 &cuip->cui_format.cui_extents[i])) {
474 XFS_CORRUPTION_ERROR(__func__, XFS_ERRLEVEL_LOW, mp,
475 &cuip->cui_format,
476 sizeof(cuip->cui_format));
895e196f 477 return -EFSCORRUPTED;
ed64f834 478 }
f997ee21
DW
479 }
480
33ba6129
DW
481 /*
482 * Under normal operation, refcount updates are deferred, so we
483 * wouldn't be adding them directly to a transaction. All
484 * refcount updates manage reservation usage internally and
485 * dynamically by deferring work that won't fit in the
486 * transaction. Normally, any work that needs to be deferred
487 * gets attached to the same defer_ops that scheduled the
488 * refcount update. However, we're in log recovery here, so we
b63da6c8 489 * use the passed in defer_ops and to finish up any work that
b31c2bdc
DW
490 * doesn't fit. We need to reserve enough blocks to handle a
491 * full btree split on either end of the refcount range.
33ba6129 492 */
b31c2bdc
DW
493 error = xfs_trans_alloc(mp, &M_RES(mp)->tr_itruncate,
494 mp->m_refc_maxlevels * 2, 0, XFS_TRANS_RESERVE, &tp);
33ba6129
DW
495 if (error)
496 return error;
e6fff81e 497
33ba6129
DW
498 cudp = xfs_trans_get_cud(tp, cuip);
499
33ba6129 500 for (i = 0; i < cuip->cui_format.cui_nextents; i++) {
0b11553e 501 struct xfs_refcount_intent fake = { };
01a3af22 502 struct xfs_phys_extent *pmap;
0b11553e 503
01a3af22
DW
504 pmap = &cuip->cui_format.cui_extents[i];
505 refc_type = pmap->pe_flags & XFS_REFCOUNT_EXTENT_TYPE_MASK;
33ba6129
DW
506 switch (refc_type) {
507 case XFS_REFCOUNT_INCREASE:
508 case XFS_REFCOUNT_DECREASE:
509 case XFS_REFCOUNT_ALLOC_COW:
510 case XFS_REFCOUNT_FREE_COW:
0b11553e 511 fake.ri_type = refc_type;
33ba6129
DW
512 break;
513 default:
950f0d50
DW
514 XFS_CORRUPTION_ERROR(__func__, XFS_ERRLEVEL_LOW, mp,
515 &cuip->cui_format,
516 sizeof(cuip->cui_format));
33ba6129
DW
517 error = -EFSCORRUPTED;
518 goto abort_error;
519 }
0b11553e 520
01a3af22
DW
521 fake.ri_startblock = pmap->pe_startblock;
522 fake.ri_blockcount = pmap->pe_len;
0b11553e 523 if (!requeue_only)
33ba6129 524 error = xfs_trans_log_finish_refcount_update(tp, cudp,
0b11553e 525 &fake, &rcur);
43059d54
DW
526 if (error == -EFSCORRUPTED)
527 XFS_CORRUPTION_ERROR(__func__, XFS_ERRLEVEL_LOW, mp,
950f0d50
DW
528 &cuip->cui_format,
529 sizeof(cuip->cui_format));
33ba6129
DW
530 if (error)
531 goto abort_error;
532
533 /* Requeue what we didn't finish. */
0b11553e
DW
534 if (fake.ri_blockcount > 0) {
535 struct xfs_bmbt_irec irec = {
536 .br_startblock = fake.ri_startblock,
537 .br_blockcount = fake.ri_blockcount,
538 };
539
540 switch (fake.ri_type) {
33ba6129 541 case XFS_REFCOUNT_INCREASE:
74b4c5d4 542 xfs_refcount_increase_extent(tp, &irec);
33ba6129
DW
543 break;
544 case XFS_REFCOUNT_DECREASE:
74b4c5d4 545 xfs_refcount_decrease_extent(tp, &irec);
33ba6129 546 break;
174edb0e 547 case XFS_REFCOUNT_ALLOC_COW:
74b4c5d4 548 xfs_refcount_alloc_cow_extent(tp,
174edb0e
DW
549 irec.br_startblock,
550 irec.br_blockcount);
551 break;
552 case XFS_REFCOUNT_FREE_COW:
74b4c5d4 553 xfs_refcount_free_cow_extent(tp,
174edb0e
DW
554 irec.br_startblock,
555 irec.br_blockcount);
556 break;
33ba6129
DW
557 default:
558 ASSERT(0);
559 }
33ba6129
DW
560 requeue_only = true;
561 }
562 }
563
564 xfs_refcount_finish_one_cleanup(tp, rcur, error);
512edfac 565 return xfs_defer_ops_capture_and_commit(tp, capture_list);
33ba6129
DW
566
567abort_error:
568 xfs_refcount_finish_one_cleanup(tp, rcur, error);
33ba6129 569 xfs_trans_cancel(tp);
f997ee21
DW
570 return error;
571}
86ffa471 572
154c733a
DW
573STATIC bool
574xfs_cui_item_match(
575 struct xfs_log_item *lip,
576 uint64_t intent_id)
577{
578 return CUI_ITEM(lip)->cui_format.cui_id == intent_id;
579}
580
4e919af7
DW
581/* Relog an intent item to push the log tail forward. */
582static struct xfs_log_item *
583xfs_cui_item_relog(
584 struct xfs_log_item *intent,
585 struct xfs_trans *tp)
586{
587 struct xfs_cud_log_item *cudp;
588 struct xfs_cui_log_item *cuip;
01a3af22 589 struct xfs_phys_extent *pmap;
4e919af7
DW
590 unsigned int count;
591
592 count = CUI_ITEM(intent)->cui_format.cui_nextents;
01a3af22 593 pmap = CUI_ITEM(intent)->cui_format.cui_extents;
4e919af7
DW
594
595 tp->t_flags |= XFS_TRANS_DIRTY;
596 cudp = xfs_trans_get_cud(tp, CUI_ITEM(intent));
597 set_bit(XFS_LI_DIRTY, &cudp->cud_item.li_flags);
598
599 cuip = xfs_cui_init(tp->t_mountp, count);
01a3af22 600 memcpy(cuip->cui_format.cui_extents, pmap, count * sizeof(*pmap));
4e919af7
DW
601 atomic_set(&cuip->cui_next_extent, count);
602 xfs_trans_add_item(tp, &cuip->cui_item);
603 set_bit(XFS_LI_DIRTY, &cuip->cui_item.li_flags);
604 return &cuip->cui_item;
605}
606
c57ed2f5 607static const struct xfs_item_ops xfs_cui_item_ops = {
f5b81200 608 .flags = XFS_ITEM_INTENT,
c57ed2f5
DW
609 .iop_size = xfs_cui_item_size,
610 .iop_format = xfs_cui_item_format,
611 .iop_unpin = xfs_cui_item_unpin,
612 .iop_release = xfs_cui_item_release,
613 .iop_recover = xfs_cui_item_recover,
154c733a 614 .iop_match = xfs_cui_item_match,
4e919af7 615 .iop_relog = xfs_cui_item_relog,
c57ed2f5
DW
616};
617
a38935c0 618static inline void
9b4467e9 619xfs_cui_copy_format(
a38935c0
DW
620 struct xfs_cui_log_format *dst,
621 const struct xfs_cui_log_format *src)
9b4467e9 622{
a38935c0 623 unsigned int i;
9b4467e9 624
a38935c0 625 memcpy(dst, src, offsetof(struct xfs_cui_log_format, cui_extents));
9b4467e9 626
a38935c0
DW
627 for (i = 0; i < src->cui_nextents; i++)
628 memcpy(&dst->cui_extents[i], &src->cui_extents[i],
629 sizeof(struct xfs_phys_extent));
9b4467e9
DW
630}
631
632/*
633 * This routine is called to create an in-core extent refcount update
634 * item from the cui format structure which was logged on disk.
635 * It allocates an in-core cui, copies the extents from the format
636 * structure into it, and adds the cui to the AIL with the given
637 * LSN.
638 */
639STATIC int
640xlog_recover_cui_commit_pass2(
641 struct xlog *log,
642 struct list_head *buffer_list,
643 struct xlog_recover_item *item,
644 xfs_lsn_t lsn)
645{
9b4467e9
DW
646 struct xfs_mount *mp = log->l_mp;
647 struct xfs_cui_log_item *cuip;
648 struct xfs_cui_log_format *cui_formatp;
a38935c0 649 size_t len;
9b4467e9
DW
650
651 cui_formatp = item->ri_buf[0].i_addr;
652
a38935c0 653 if (item->ri_buf[0].i_len < xfs_cui_log_format_sizeof(0)) {
950f0d50
DW
654 XFS_CORRUPTION_ERROR(__func__, XFS_ERRLEVEL_LOW, mp,
655 item->ri_buf[0].i_addr, item->ri_buf[0].i_len);
a38935c0
DW
656 return -EFSCORRUPTED;
657 }
658
659 len = xfs_cui_log_format_sizeof(cui_formatp->cui_nextents);
660 if (item->ri_buf[0].i_len != len) {
950f0d50
DW
661 XFS_CORRUPTION_ERROR(__func__, XFS_ERRLEVEL_LOW, mp,
662 item->ri_buf[0].i_addr, item->ri_buf[0].i_len);
a38935c0 663 return -EFSCORRUPTED;
9b4467e9 664 }
a38935c0
DW
665
666 cuip = xfs_cui_init(mp, cui_formatp->cui_nextents);
667 xfs_cui_copy_format(&cuip->cui_format, cui_formatp);
9b4467e9 668 atomic_set(&cuip->cui_next_extent, cui_formatp->cui_nextents);
9b4467e9 669 /*
86a37174
DW
670 * Insert the intent into the AIL directly and drop one reference so
671 * that finishing or canceling the work will drop the other.
9b4467e9 672 */
86a37174 673 xfs_trans_ail_insert(log->l_ailp, &cuip->cui_item, lsn);
9b4467e9
DW
674 xfs_cui_release(cuip);
675 return 0;
676}
677
86ffa471
DW
678const struct xlog_recover_item_ops xlog_cui_item_ops = {
679 .item_type = XFS_LI_CUI,
9b4467e9 680 .commit_pass2 = xlog_recover_cui_commit_pass2,
86ffa471
DW
681};
682
9b4467e9
DW
683/*
684 * This routine is called when an CUD format structure is found in a committed
685 * transaction in the log. Its purpose is to cancel the corresponding CUI if it
686 * was still in the log. To do this it searches the AIL for the CUI with an id
687 * equal to that in the CUD format structure. If we find it we drop the CUD
688 * reference, which removes the CUI from the AIL and frees it.
689 */
690STATIC int
691xlog_recover_cud_commit_pass2(
692 struct xlog *log,
693 struct list_head *buffer_list,
694 struct xlog_recover_item *item,
695 xfs_lsn_t lsn)
696{
697 struct xfs_cud_log_format *cud_formatp;
9b4467e9
DW
698
699 cud_formatp = item->ri_buf[0].i_addr;
700 if (item->ri_buf[0].i_len != sizeof(struct xfs_cud_log_format)) {
950f0d50
DW
701 XFS_CORRUPTION_ERROR(__func__, XFS_ERRLEVEL_LOW, log->l_mp,
702 item->ri_buf[0].i_addr, item->ri_buf[0].i_len);
9b4467e9
DW
703 return -EFSCORRUPTED;
704 }
9b4467e9 705
154c733a 706 xlog_recover_release_intent(log, XFS_LI_CUI, cud_formatp->cud_cui_id);
9b4467e9
DW
707 return 0;
708}
709
86ffa471
DW
710const struct xlog_recover_item_ops xlog_cud_item_ops = {
711 .item_type = XFS_LI_CUD,
9b4467e9 712 .commit_pass2 = xlog_recover_cud_commit_pass2,
86ffa471 713};