bcachefs: Deduplicate keys in the journal before replay
authorKent Overstreet <kent.overstreet@gmail.com>
Fri, 12 Apr 2019 02:39:39 +0000 (22:39 -0400)
committerKent Overstreet <kent.overstreet@linux.dev>
Sun, 22 Oct 2023 21:08:20 +0000 (17:08 -0400)
Signed-off-by: Kent Overstreet <kent.overstreet@linux.dev>
fs/bcachefs/alloc_background.c
fs/bcachefs/alloc_background.h
fs/bcachefs/btree_gc.c
fs/bcachefs/btree_gc.h
fs/bcachefs/ec.c
fs/bcachefs/ec.h
fs/bcachefs/recovery.c
fs/bcachefs/recovery.h

index b3a8ff0b1daa2ab7b3f785dac32d5163d68a7c1a..5c8cebc443d1b7ddf56e081f9ab50396d15fc622 100644 (file)
@@ -12,7 +12,7 @@
 #include "debug.h"
 #include "ec.h"
 #include "error.h"
-#include "journal_io.h"
+#include "recovery.h"
 #include "trace.h"
 
 #include <linux/kthread.h>
@@ -261,13 +261,13 @@ static void bch2_alloc_read_key(struct bch_fs *c, struct bkey_s_c k)
        percpu_up_read(&c->mark_lock);
 }
 
-int bch2_alloc_read(struct bch_fs *c, struct list_head *journal_replay_list)
+int bch2_alloc_read(struct bch_fs *c, struct journal_keys *journal_keys)
 {
-       struct journal_replay *r;
        struct btree_trans trans;
        struct btree_iter *iter;
        struct bkey_s_c k;
        struct bch_dev *ca;
+       struct journal_key *j;
        unsigned i;
        int ret;
 
@@ -282,14 +282,9 @@ int bch2_alloc_read(struct bch_fs *c, struct list_head *journal_replay_list)
        if (ret)
                return ret;
 
-       list_for_each_entry(r, journal_replay_list, list) {
-               struct bkey_i *k, *n;
-               struct jset_entry *entry;
-
-               for_each_jset_key(k, n, entry, &r->j)
-                       if (entry->btree_id == BTREE_ID_ALLOC)
-                               bch2_alloc_read_key(c, bkey_i_to_s_c(k));
-       }
+       for_each_journal_key(*journal_keys, j)
+               if (j->btree_id == BTREE_ID_ALLOC)
+                       bch2_alloc_read_key(c, bkey_i_to_s_c(j->k));
 
        percpu_down_write(&c->mark_lock);
        bch2_dev_usage_from_buckets(c);
index 25d7426613daff2542af9e6b002205b6f3a633de..b75c56a5dae0107074de04b67f3f865c66b154f5 100644 (file)
@@ -25,7 +25,8 @@ void bch2_alloc_to_text(struct printbuf *, struct bch_fs *, struct bkey_s_c);
        .val_to_text    = bch2_alloc_to_text,           \
 }
 
-int bch2_alloc_read(struct bch_fs *, struct list_head *);
+struct journal_keys;
+int bch2_alloc_read(struct bch_fs *, struct journal_keys *);
 int bch2_alloc_replay_key(struct bch_fs *, struct bkey_i *);
 
 static inline void bch2_wake_allocator(struct bch_dev *ca)
index 84a0bb9202c46173ed1f229bcc3af16c76e5d1e7..cf0a2f4b22af5b1fb39c148d07f2f59d42b240f4 100644 (file)
@@ -19,9 +19,9 @@
 #include "error.h"
 #include "extents.h"
 #include "journal.h"
-#include "journal_io.h"
 #include "keylist.h"
 #include "move.h"
+#include "recovery.h"
 #include "replicas.h"
 #include "super-io.h"
 #include "trace.h"
@@ -273,7 +273,7 @@ static inline int btree_id_gc_phase_cmp(enum btree_id l, enum btree_id r)
                (int) btree_id_to_gc_phase(r);
 }
 
-static int bch2_gc_btrees(struct bch_fs *c, struct list_head *journal,
+static int bch2_gc_btrees(struct bch_fs *c, struct journal_keys *journal_keys,
                          bool initial, bool metadata_only)
 {
        enum btree_id ids[BTREE_ID_NR];
@@ -292,22 +292,18 @@ static int bch2_gc_btrees(struct bch_fs *c, struct list_head *journal,
                if (ret)
                        return ret;
 
-               if (journal && !metadata_only &&
+               if (journal_keys && !metadata_only &&
                    btree_node_type_needs_gc(type)) {
-                       struct bkey_i *k, *n;
-                       struct jset_entry *j;
-                       struct journal_replay *r;
+                       struct journal_key *j;
                        int ret;
 
-                       list_for_each_entry(r, journal, list)
-                               for_each_jset_key(k, n, j, &r->j) {
-                                       if (type == __btree_node_type(j->level, j->btree_id)) {
-                                               ret = bch2_gc_mark_key(c,
-                                                       bkey_i_to_s_c(k),
-                                                       &max_stale, initial);
-                                               if (ret)
-                                                       return ret;
-                                       }
+                       for_each_journal_key(*journal_keys, j)
+                               if (j->btree_id == id) {
+                                       ret = bch2_gc_mark_key(c,
+                                               bkey_i_to_s_c(j->k),
+                                               &max_stale, initial);
+                                       if (ret)
+                                               return ret;
                                }
                }
        }
@@ -695,7 +691,7 @@ static int bch2_gc_start(struct bch_fs *c,
  *    move around - if references move backwards in the ordering GC
  *    uses, GC could skip past them
  */
-int bch2_gc(struct bch_fs *c, struct list_head *journal,
+int bch2_gc(struct bch_fs *c, struct journal_keys *journal_keys,
            bool initial, bool metadata_only)
 {
        struct bch_dev *ca;
@@ -716,7 +712,7 @@ again:
 
        bch2_mark_superblocks(c);
 
-       ret = bch2_gc_btrees(c, journal, initial, metadata_only);
+       ret = bch2_gc_btrees(c, journal_keys, initial, metadata_only);
        if (ret)
                goto out;
 
index b7982e64b2352dc3072e678d0f272cd715174d7e..bd5f2752954fede861e8cb5170d9909cc8e41da0 100644 (file)
@@ -5,7 +5,9 @@
 #include "btree_types.h"
 
 void bch2_coalesce(struct bch_fs *);
-int bch2_gc(struct bch_fs *, struct list_head *, bool, bool);
+
+struct journal_keys;
+int bch2_gc(struct bch_fs *, struct journal_keys *, bool, bool);
 void bch2_gc_thread_stop(struct bch_fs *);
 int bch2_gc_thread_start(struct bch_fs *);
 void bch2_mark_dev_superblock(struct bch_fs *, struct bch_dev *, unsigned);
index 6a357e5b652e70f8f439e457f1c31b76bc1ebb50..47d197ed5c99b050907554fe8e55a782dd3dbccb 100644 (file)
@@ -12,8 +12,8 @@
 #include "ec.h"
 #include "error.h"
 #include "io.h"
-#include "journal_io.h"
 #include "keylist.h"
+#include "recovery.h"
 #include "super-io.h"
 #include "util.h"
 
@@ -1235,9 +1235,9 @@ static void bch2_stripe_read_key(struct bch_fs *c, struct bkey_s_c k)
        bch2_mark_key(c, k, true, 0, NULL, 0, 0);
 }
 
-int bch2_stripes_read(struct bch_fs *c, struct list_head *journal_replay_list)
+int bch2_stripes_read(struct bch_fs *c, struct journal_keys *journal_keys)
 {
-       struct journal_replay *r;
+       struct journal_key *i;
        struct btree_trans trans;
        struct btree_iter *iter;
        struct bkey_s_c k;
@@ -1258,14 +1258,9 @@ int bch2_stripes_read(struct bch_fs *c, struct list_head *journal_replay_list)
        if (ret)
                return ret;
 
-       list_for_each_entry(r, journal_replay_list, list) {
-               struct bkey_i *k, *n;
-               struct jset_entry *entry;
-
-               for_each_jset_key(k, n, entry, &r->j)
-                       if (entry->btree_id == BTREE_ID_EC)
-                               bch2_stripe_read_key(c, bkey_i_to_s_c(k));
-       }
+       for_each_journal_key(*journal_keys, i)
+               if (i->btree_id == BTREE_ID_EC)
+                       bch2_stripe_read_key(c, bkey_i_to_s_c(i->k));
 
        return 0;
 }
index b048244a4a4593cd067423f62aa12fafb28d18bc..8d9fbfd19f66f3ca5069e9ef58a48f82fb227b7a 100644 (file)
@@ -150,7 +150,8 @@ void bch2_ec_stop_dev(struct bch_fs *, struct bch_dev *);
 
 void bch2_ec_flush_new_stripes(struct bch_fs *);
 
-int bch2_stripes_read(struct bch_fs *, struct list_head *);
+struct journal_keys;
+int bch2_stripes_read(struct bch_fs *, struct journal_keys *);
 int bch2_stripes_write(struct bch_fs *, unsigned, bool *);
 
 int bch2_ec_mem_alloc(struct bch_fs *, bool);
index 2e849135195d2abac3d0fe776f8c0d3b359a7927..5bfb38c4290f0bf5558eacbb4f6d142f119550ad 100644 (file)
@@ -24,9 +24,9 @@
 
 #define QSTR(n) { { { .len = strlen(n) } }, .name = n }
 
-/* journal replay: */
+/* sort and dedup all keys in the journal: */
 
-static void bch2_journal_entries_free(struct list_head *list)
+static void journal_entries_free(struct list_head *list)
 {
 
        while (!list_empty(list)) {
@@ -38,6 +38,168 @@ static void bch2_journal_entries_free(struct list_head *list)
        }
 }
 
+static int journal_sort_key_cmp(const void *_l, const void *_r)
+{
+       const struct journal_key *l = _l;
+       const struct journal_key *r = _r;
+
+       return cmp_int(l->btree_id, r->btree_id) ?:
+               bkey_cmp(l->pos, r->pos) ?:
+               cmp_int(l->journal_seq, r->journal_seq) ?:
+               cmp_int(l->journal_offset, r->journal_offset);
+}
+
+static int journal_sort_seq_cmp(const void *_l, const void *_r)
+{
+       const struct journal_key *l = _l;
+       const struct journal_key *r = _r;
+
+       return cmp_int(l->journal_seq, r->journal_seq) ?:
+               cmp_int(l->btree_id, r->btree_id) ?:
+               bkey_cmp(l->pos, r->pos);
+}
+
+static void journal_keys_sift(struct journal_keys *keys, struct journal_key *i)
+{
+       while (i + 1 < keys->d + keys->nr &&
+              journal_sort_key_cmp(i, i + 1) > 0) {
+               swap(i[0], i[1]);
+               i++;
+       }
+}
+
+static void journal_keys_free(struct journal_keys *keys)
+{
+       struct journal_key *i;
+
+       for_each_journal_key(*keys, i)
+               if (i->allocated)
+                       kfree(i->k);
+       kvfree(keys->d);
+       keys->d = NULL;
+       keys->nr = 0;
+}
+
+static struct journal_keys journal_keys_sort(struct list_head *journal_entries)
+{
+       struct journal_replay *p;
+       struct jset_entry *entry;
+       struct bkey_i *k, *_n;
+       struct journal_keys keys = { NULL }, keys_deduped = { NULL };
+       struct journal_key *i;
+       size_t nr_keys = 0;
+
+       list_for_each_entry(p, journal_entries, list)
+               for_each_jset_key(k, _n, entry, &p->j)
+                       nr_keys++;
+
+       keys.journal_seq_base = keys_deduped.journal_seq_base =
+               le64_to_cpu(list_first_entry(journal_entries,
+                                            struct journal_replay,
+                                            list)->j.seq);
+
+       keys.d = kvmalloc(sizeof(keys.d[0]) * nr_keys, GFP_KERNEL);
+       if (!keys.d)
+               goto err;
+
+       keys_deduped.d = kvmalloc(sizeof(keys.d[0]) * nr_keys * 2, GFP_KERNEL);
+       if (!keys_deduped.d)
+               goto err;
+
+       list_for_each_entry(p, journal_entries, list)
+               for_each_jset_key(k, _n, entry, &p->j)
+                       keys.d[keys.nr++] = (struct journal_key) {
+                               .btree_id       = entry->btree_id,
+                               .pos            = bkey_start_pos(&k->k),
+                               .k              = k,
+                               .journal_seq    = le64_to_cpu(p->j.seq) -
+                                       keys.journal_seq_base,
+                               .journal_offset = k->_data - p->j._data,
+                       };
+
+       sort(keys.d, nr_keys, sizeof(keys.d[0]), journal_sort_key_cmp, NULL);
+
+       i = keys.d;
+       while (i < keys.d + keys.nr) {
+               if (i + 1 < keys.d + keys.nr &&
+                   i[0].btree_id == i[1].btree_id &&
+                   !bkey_cmp(i[0].pos, i[1].pos)) {
+                       if (bkey_cmp(i[0].k->k.p, i[1].k->k.p) <= 0) {
+                               i++;
+                       } else {
+                               bch2_cut_front(i[1].k->k.p, i[0].k);
+                               i[0].pos = i[1].k->k.p;
+                               journal_keys_sift(&keys, i);
+                       }
+                       continue;
+               }
+
+               if (i + 1 < keys.d + keys.nr &&
+                   i[0].btree_id == i[1].btree_id &&
+                   bkey_cmp(i[0].k->k.p, bkey_start_pos(&i[1].k->k)) > 0) {
+                       if ((cmp_int(i[0].journal_seq, i[1].journal_seq) ?:
+                            cmp_int(i[0].journal_offset, i[1].journal_offset)) < 0) {
+                               if (bkey_cmp(i[0].k->k.p, i[1].k->k.p) <= 0) {
+                                       bch2_cut_back(bkey_start_pos(&i[1].k->k), &i[0].k->k);
+                               } else {
+                                       struct bkey_i *split =
+                                               kmalloc(bkey_bytes(i[0].k), GFP_KERNEL);
+
+                                       if (!split)
+                                               goto err;
+
+                                       bkey_copy(split, i[0].k);
+                                       bch2_cut_back(bkey_start_pos(&i[1].k->k), &split->k);
+                                       keys_deduped.d[keys_deduped.nr++] = (struct journal_key) {
+                                               .btree_id       = i[0].btree_id,
+                                               .allocated      = true,
+                                               .pos            = bkey_start_pos(&split->k),
+                                               .k              = split,
+                                               .journal_seq    = i[0].journal_seq,
+                                               .journal_offset = i[0].journal_offset,
+                                       };
+
+                                       bch2_cut_front(i[1].k->k.p, i[0].k);
+                                       i[0].pos = i[1].k->k.p;
+                                       journal_keys_sift(&keys, i);
+                                       continue;
+                               }
+                       } else {
+                               if (bkey_cmp(i[0].k->k.p, i[1].k->k.p) >= 0) {
+                                       i[1] = i[0];
+                                       i++;
+                                       continue;
+                               } else {
+                                       bch2_cut_front(i[0].k->k.p, i[1].k);
+                                       i[1].pos = i[0].k->k.p;
+                                       journal_keys_sift(&keys, i + 1);
+                                       continue;
+                               }
+                       }
+               }
+
+               keys_deduped.d[keys_deduped.nr++] = *i++;
+       }
+
+       kvfree(keys.d);
+       return keys_deduped;
+err:
+       journal_keys_free(&keys_deduped);
+       kvfree(keys.d);
+       return (struct journal_keys) { NULL };
+}
+
+/* journal replay: */
+
+static void replay_now_at(struct journal *j, u64 seq)
+{
+       BUG_ON(seq < j->replay_journal_seq);
+       BUG_ON(seq > j->replay_journal_seq_end);
+
+       while (j->replay_journal_seq < seq)
+               bch2_journal_pin_put(j, j->replay_journal_seq++);
+}
+
 static int bch2_extent_replay_key(struct bch_fs *c, struct bkey_i *k)
 {
        struct btree_trans trans;
@@ -100,54 +262,42 @@ static int bch2_extent_replay_key(struct bch_fs *c, struct bkey_i *k)
        return ret;
 }
 
-static int bch2_journal_replay_key(struct bch_fs *c, enum btree_id btree_id,
-                                  struct bkey_i *k)
-{
-       switch (btree_id) {
-       case BTREE_ID_ALLOC:
-               return bch2_alloc_replay_key(c, k);
-       case BTREE_ID_EXTENTS:
-               return bch2_extent_replay_key(c, k);
-       default:
-               return bch2_btree_insert(c, btree_id, k,
-                                        NULL, NULL,
-                                        BTREE_INSERT_NOFAIL|
-                                        BTREE_INSERT_LAZY_RW|
-                                        BTREE_INSERT_JOURNAL_REPLAY|
-                                        BTREE_INSERT_NOMARK);
-       }
-}
-
-static void replay_now_at(struct journal *j, u64 seq)
-{
-       BUG_ON(seq < j->replay_journal_seq);
-       BUG_ON(seq > j->replay_journal_seq_end);
-
-       while (j->replay_journal_seq < seq)
-               bch2_journal_pin_put(j, j->replay_journal_seq++);
-}
-
-static int bch2_journal_replay(struct bch_fs *c, struct list_head *list)
+static int bch2_journal_replay(struct bch_fs *c,
+                              struct journal_keys keys)
 {
        struct journal *j = &c->journal;
-       struct bkey_i *k, *_n;
-       struct jset_entry *entry;
-       struct journal_replay *i, *n;
-       int ret = 0;
+       struct journal_key *i;
+       int ret;
 
-       list_for_each_entry_safe(i, n, list, list) {
-               replay_now_at(j, le64_to_cpu(i->j.seq));
+       sort(keys.d, keys.nr, sizeof(keys.d[0]), journal_sort_seq_cmp, NULL);
 
-               for_each_jset_key(k, _n, entry, &i->j) {
-                       ret = bch2_journal_replay_key(c, entry->btree_id, k);
-                       if (ret) {
-                               bch_err(c, "journal replay: error %d while replaying key",
-                                       ret);
-                               goto err;
-                       }
+       for_each_journal_key(keys, i) {
+               replay_now_at(j, keys.journal_seq_base + i->journal_seq);
+
+               switch (i->btree_id) {
+               case BTREE_ID_ALLOC:
+                       ret = bch2_alloc_replay_key(c, i->k);
+                       break;
+               case BTREE_ID_EXTENTS:
+                       ret = bch2_extent_replay_key(c, i->k);
+                       break;
+               default:
+                       ret = bch2_btree_insert(c, i->btree_id, i->k,
+                                               NULL, NULL,
+                                               BTREE_INSERT_NOFAIL|
+                                               BTREE_INSERT_LAZY_RW|
+                                               BTREE_INSERT_JOURNAL_REPLAY|
+                                               BTREE_INSERT_NOMARK);
+                       break;
+               }
 
-                       cond_resched();
+               if (ret) {
+                       bch_err(c, "journal replay: error %d while replaying key",
+                               ret);
+                       return ret;
                }
+
+               cond_resched();
        }
 
        replay_now_at(j, j->replay_journal_seq_end);
@@ -155,10 +305,7 @@ static int bch2_journal_replay(struct bch_fs *c, struct list_head *list)
 
        bch2_journal_set_replay_done(j);
        bch2_journal_flush_all_pins(j);
-       ret = bch2_journal_error(j);
-err:
-       bch2_journal_entries_free(list);
-       return ret;
+       return bch2_journal_error(j);
 }
 
 static bool journal_empty(struct list_head *journal)
@@ -475,7 +622,8 @@ int bch2_fs_recovery(struct bch_fs *c)
        const char *err = "cannot allocate memory";
        struct bch_sb_field_clean *clean = NULL;
        u64 journal_seq;
-       LIST_HEAD(journal);
+       LIST_HEAD(journal_entries);
+       struct journal_keys journal_keys = { NULL };
        int ret;
 
        if (c->sb.clean)
@@ -496,20 +644,27 @@ int bch2_fs_recovery(struct bch_fs *c)
        if (!c->sb.clean || c->opts.fsck) {
                struct jset *j;
 
-               ret = bch2_journal_read(c, &journal);
+               ret = bch2_journal_read(c, &journal_entries);
                if (ret)
                        goto err;
 
-               fsck_err_on(c->sb.clean && !journal_empty(&journal), c,
+               fsck_err_on(c->sb.clean && !journal_empty(&journal_entries), c,
                            "filesystem marked clean but journal not empty");
 
-               if (!c->sb.clean && list_empty(&journal)){
+               if (!c->sb.clean && list_empty(&journal_entries)) {
                        bch_err(c, "no journal entries found");
                        ret = BCH_FSCK_REPAIR_IMPOSSIBLE;
                        goto err;
                }
 
-               j = &list_last_entry(&journal, struct journal_replay, list)->j;
+               journal_keys = journal_keys_sort(&journal_entries);
+               if (!journal_keys.d) {
+                       ret = -ENOMEM;
+                       goto err;
+               }
+
+               j = &list_last_entry(&journal_entries,
+                                    struct journal_replay, list)->j;
 
                ret = verify_superblock_clean(c, &clean, j);
                if (ret)
@@ -520,7 +675,7 @@ int bch2_fs_recovery(struct bch_fs *c)
                journal_seq = le64_to_cpu(clean->journal_seq) + 1;
        }
 
-       ret = journal_replay_early(c, clean, &journal);
+       ret = journal_replay_early(c, clean, &journal_entries);
        if (ret)
                goto err;
 
@@ -538,11 +693,13 @@ int bch2_fs_recovery(struct bch_fs *c)
 
        ret = bch2_blacklist_table_initialize(c);
 
-       ret = verify_journal_entries_not_blacklisted_or_missing(c, &journal);
+       ret = verify_journal_entries_not_blacklisted_or_missing(c,
+                                               &journal_entries);
        if (ret)
                goto err;
 
-       ret = bch2_fs_journal_start(&c->journal, journal_seq, &journal);
+       ret = bch2_fs_journal_start(&c->journal, journal_seq,
+                                   &journal_entries);
        if (ret)
                goto err;
 
@@ -551,12 +708,12 @@ int bch2_fs_recovery(struct bch_fs *c)
                goto err;
 
        err = "error reading allocation information";
-       ret = bch2_alloc_read(c, &journal);
+       ret = bch2_alloc_read(c, &journal_keys);
        if (ret)
                goto err;
 
        bch_verbose(c, "starting stripes_read");
-       ret = bch2_stripes_read(c, &journal);
+       ret = bch2_stripes_read(c, &journal_keys);
        if (ret)
                goto err;
        bch_verbose(c, "stripes_read done");
@@ -568,7 +725,7 @@ int bch2_fs_recovery(struct bch_fs *c)
            test_bit(BCH_FS_REBUILD_REPLICAS, &c->flags)) {
                bch_verbose(c, "starting mark and sweep:");
                err = "error in recovery";
-               ret = bch2_gc(c, &journal, true, false);
+               ret = bch2_gc(c, &journal_keys, true, false);
                if (ret)
                        goto err;
                bch_verbose(c, "mark and sweep done");
@@ -589,7 +746,7 @@ int bch2_fs_recovery(struct bch_fs *c)
 
        bch_verbose(c, "starting journal replay:");
        err = "journal replay failed";
-       ret = bch2_journal_replay(c, &journal);
+       ret = bch2_journal_replay(c, journal_keys);
        if (ret)
                goto err;
        bch_verbose(c, "journal replay done");
@@ -629,7 +786,8 @@ int bch2_fs_recovery(struct bch_fs *c)
            c->journal_seq_blacklist_table->nr > 128)
                queue_work(system_long_wq, &c->journal_seq_blacklist_gc_work);
 out:
-       bch2_journal_entries_free(&journal);
+       journal_keys_free(&journal_keys);
+       journal_entries_free(&journal_entries);
        kfree(clean);
        return ret;
 err:
index 912929117c3733ff79a5d15882a2a5b345ed632e..a69260d6165a148b917a4ef2bae4abc78814a104 100644 (file)
@@ -2,6 +2,22 @@
 #ifndef _BCACHEFS_RECOVERY_H
 #define _BCACHEFS_RECOVERY_H
 
+struct journal_keys {
+       struct journal_key {
+               enum btree_id   btree_id:8;
+               unsigned        allocated:1;
+               struct bpos     pos;
+               struct bkey_i   *k;
+               u32             journal_seq;
+               u32             journal_offset;
+       }                       *d;
+       size_t                  nr;
+       u64                     journal_seq_base;
+};
+
+#define for_each_journal_key(keys, i)                          \
+       for (i = (keys).d; i < (keys).d + (keys).nr; (i)++)
+
 int bch2_fs_recovery(struct bch_fs *);
 int bch2_fs_initialize(struct bch_fs *);