Merge tag 'bcachefs-2024-05-19' of https://evilpiepirate.org/git/bcachefs
[linux-2.6-block.git] / fs / bcachefs / sb-members.h
index 5bf27d30ca2969017c6d33edde8c99ce1a36cb7b..dd93192ec065e8ecfb5d2ffb32e87f5d24599308 100644 (file)
@@ -29,19 +29,6 @@ static inline bool bch2_dev_is_readable(struct bch_dev *ca)
                ca->mi.state != BCH_MEMBER_STATE_failed;
 }
 
-static inline bool bch2_dev_get_ioref(struct bch_dev *ca, int rw)
-{
-       if (!percpu_ref_tryget(&ca->io_ref))
-               return false;
-
-       if (ca->mi.state == BCH_MEMBER_STATE_rw ||
-           (ca->mi.state == BCH_MEMBER_STATE_ro && rw == READ))
-               return true;
-
-       percpu_ref_put(&ca->io_ref);
-       return false;
-}
-
 static inline unsigned dev_mask_nr(const struct bch_devs_mask *devs)
 {
        return bitmap_weight(devs->d, BCH_SB_MEMBERS_MAX);
@@ -105,14 +92,41 @@ static inline struct bch_dev *__bch2_next_dev(struct bch_fs *c, struct bch_dev *
        for (struct bch_dev *_ca = NULL;                                \
             (_ca = __bch2_next_dev((_c), _ca, (_mask)));)
 
-static inline struct bch_dev *bch2_get_next_dev(struct bch_fs *c, struct bch_dev *ca)
+static inline void bch2_dev_get(struct bch_dev *ca)
+{
+#ifdef CONFIG_BCACHEFS_DEBUG
+       BUG_ON(atomic_long_inc_return(&ca->ref) <= 1L);
+#else
+       percpu_ref_get(&ca->ref);
+#endif
+}
+
+static inline void __bch2_dev_put(struct bch_dev *ca)
+{
+#ifdef CONFIG_BCACHEFS_DEBUG
+       long r = atomic_long_dec_return(&ca->ref);
+       if (r < (long) !ca->dying)
+               panic("bch_dev->ref underflow, last put: %pS\n", (void *) ca->last_put);
+       ca->last_put = _THIS_IP_;
+       if (!r)
+               complete(&ca->ref_completion);
+#else
+       percpu_ref_put(&ca->ref);
+#endif
+}
+
+static inline void bch2_dev_put(struct bch_dev *ca)
 {
-       rcu_read_lock();
        if (ca)
-               percpu_ref_put(&ca->ref);
+               __bch2_dev_put(ca);
+}
 
+static inline struct bch_dev *bch2_get_next_dev(struct bch_fs *c, struct bch_dev *ca)
+{
+       rcu_read_lock();
+       bch2_dev_put(ca);
        if ((ca = __bch2_next_dev(c, ca, NULL)))
-               percpu_ref_get(&ca->ref);
+               bch2_dev_get(ca);
        rcu_read_unlock();
 
        return ca;
@@ -158,26 +172,113 @@ static inline struct bch_dev *bch2_get_next_online_dev(struct bch_fs *c,
 #define for_each_readable_member(c, ca)                                \
        __for_each_online_member(c, ca, BIT( BCH_MEMBER_STATE_rw)|BIT(BCH_MEMBER_STATE_ro))
 
-/*
- * If a key exists that references a device, the device won't be going away and
- * we can omit rcu_read_lock():
- */
-static inline struct bch_dev *bch_dev_bkey_exists(const struct bch_fs *c, unsigned idx)
+static inline bool bch2_dev_exists(const struct bch_fs *c, unsigned dev)
 {
-       EBUG_ON(idx >= c->sb.nr_devices || !c->devs[idx]);
+       return dev < c->sb.nr_devices && c->devs[dev];
+}
 
-       return rcu_dereference_check(c->devs[idx], 1);
+static inline bool bucket_valid(const struct bch_dev *ca, u64 b)
+{
+       return b - ca->mi.first_bucket < ca->mi.nbuckets_minus_first;
 }
 
-static inline struct bch_dev *bch_dev_locked(struct bch_fs *c, unsigned idx)
+static inline struct bch_dev *bch2_dev_have_ref(const struct bch_fs *c, unsigned dev)
 {
-       EBUG_ON(idx >= c->sb.nr_devices || !c->devs[idx]);
+       EBUG_ON(!bch2_dev_exists(c, dev));
+
+       return rcu_dereference_check(c->devs[dev], 1);
+}
 
-       return rcu_dereference_protected(c->devs[idx],
+static inline struct bch_dev *bch2_dev_locked(struct bch_fs *c, unsigned dev)
+{
+       EBUG_ON(!bch2_dev_exists(c, dev));
+
+       return rcu_dereference_protected(c->devs[dev],
                                         lockdep_is_held(&c->sb_lock) ||
                                         lockdep_is_held(&c->state_lock));
 }
 
+static inline struct bch_dev *bch2_dev_rcu(struct bch_fs *c, unsigned dev)
+{
+       return c && dev < c->sb.nr_devices
+               ? rcu_dereference(c->devs[dev])
+               : NULL;
+}
+
+static inline struct bch_dev *bch2_dev_tryget_noerror(struct bch_fs *c, unsigned dev)
+{
+       rcu_read_lock();
+       struct bch_dev *ca = bch2_dev_rcu(c, dev);
+       if (ca)
+               bch2_dev_get(ca);
+       rcu_read_unlock();
+       return ca;
+}
+
+void bch2_dev_missing(struct bch_fs *, unsigned);
+
+static inline struct bch_dev *bch2_dev_tryget(struct bch_fs *c, unsigned dev)
+{
+       struct bch_dev *ca = bch2_dev_tryget_noerror(c, dev);
+       if (!ca)
+               bch2_dev_missing(c, dev);
+       return ca;
+}
+
+static inline struct bch_dev *bch2_dev_bucket_tryget_noerror(struct bch_fs *c, struct bpos bucket)
+{
+       struct bch_dev *ca = bch2_dev_tryget_noerror(c, bucket.inode);
+       if (ca && !bucket_valid(ca, bucket.offset)) {
+               bch2_dev_put(ca);
+               ca = NULL;
+       }
+       return ca;
+}
+
+void bch2_dev_bucket_missing(struct bch_fs *, struct bpos);
+
+static inline struct bch_dev *bch2_dev_bucket_tryget(struct bch_fs *c, struct bpos bucket)
+{
+       struct bch_dev *ca = bch2_dev_bucket_tryget_noerror(c, bucket);
+       if (!ca)
+               bch2_dev_bucket_missing(c, bucket);
+       return ca;
+}
+
+static inline struct bch_dev *bch2_dev_iterate_noerror(struct bch_fs *c, struct bch_dev *ca, unsigned dev_idx)
+{
+       if (ca && ca->dev_idx == dev_idx)
+               return ca;
+       bch2_dev_put(ca);
+       return bch2_dev_tryget_noerror(c, dev_idx);
+}
+
+static inline struct bch_dev *bch2_dev_iterate(struct bch_fs *c, struct bch_dev *ca, unsigned dev_idx)
+{
+       if (ca && ca->dev_idx == dev_idx)
+               return ca;
+       bch2_dev_put(ca);
+       return bch2_dev_tryget(c, dev_idx);
+}
+
+static inline struct bch_dev *bch2_dev_get_ioref(struct bch_fs *c, unsigned dev, int rw)
+{
+       rcu_read_lock();
+       struct bch_dev *ca = bch2_dev_rcu(c, dev);
+       if (ca && !percpu_ref_tryget(&ca->io_ref))
+               ca = NULL;
+       rcu_read_unlock();
+
+       if (ca &&
+           (ca->mi.state == BCH_MEMBER_STATE_rw ||
+           (ca->mi.state == BCH_MEMBER_STATE_ro && rw == READ)))
+               return ca;
+
+       if (ca)
+               percpu_ref_put(&ca->io_ref);
+       return NULL;
+}
+
 /* XXX kill, move to struct bch_fs */
 static inline struct bch_devs_mask bch2_online_devs(struct bch_fs *c)
 {
@@ -192,16 +293,16 @@ static inline struct bch_devs_mask bch2_online_devs(struct bch_fs *c)
 extern const struct bch_sb_field_ops bch_sb_field_ops_members_v1;
 extern const struct bch_sb_field_ops bch_sb_field_ops_members_v2;
 
-static inline bool bch2_member_exists(struct bch_member *m)
+static inline bool bch2_member_alive(struct bch_member *m)
 {
        return !bch2_is_zero(&m->uuid, sizeof(m->uuid));
 }
 
-static inline bool bch2_dev_exists(struct bch_sb *sb, unsigned dev)
+static inline bool bch2_member_exists(struct bch_sb *sb, unsigned dev)
 {
        if (dev < sb->nr_devices) {
                struct bch_member m = bch2_sb_member_get(sb, dev);
-               return bch2_member_exists(&m);
+               return bch2_member_alive(&m);
        }
        return false;
 }
@@ -210,6 +311,8 @@ static inline struct bch_member_cpu bch2_mi_to_cpu(struct bch_member *mi)
 {
        return (struct bch_member_cpu) {
                .nbuckets       = le64_to_cpu(mi->nbuckets),
+               .nbuckets_minus_first = le64_to_cpu(mi->nbuckets) -
+                       le16_to_cpu(mi->first_bucket),
                .first_bucket   = le16_to_cpu(mi->first_bucket),
                .bucket_size    = le16_to_cpu(mi->bucket_size),
                .group          = BCH_MEMBER_GROUP(mi),
@@ -220,7 +323,7 @@ static inline struct bch_member_cpu bch2_mi_to_cpu(struct bch_member *mi)
                        ? BCH_MEMBER_DURABILITY(mi) - 1
                        : 1,
                .freespace_initialized = BCH_MEMBER_FREESPACE_INITIALIZED(mi),
-               .valid          = bch2_member_exists(mi),
+               .valid          = bch2_member_alive(mi),
                .btree_bitmap_shift     = mi->btree_bitmap_shift,
                .btree_allocated_bitmap = le64_to_cpu(mi->btree_allocated_bitmap),
        };