bcachefs: Add more assertions for locking btree iterators out of order
authorKent Overstreet <kent.overstreet@gmail.com>
Fri, 3 Sep 2021 21:18:57 +0000 (17:18 -0400)
committerKent Overstreet <kent.overstreet@linux.dev>
Sun, 22 Oct 2023 21:09:11 +0000 (17:09 -0400)
btree_path_traverse_all() traverses btree iterators in sorted order, and
thus shouldn't see transaction restarts due to potential deadlocks - but
sometimes we do. This patch adds some more assertions and tracks some
more state to help track this down.

Signed-off-by: Kent Overstreet <kent.overstreet@gmail.com>
fs/bcachefs/btree_iter.c
fs/bcachefs/btree_iter.h
fs/bcachefs/btree_key_cache.c
fs/bcachefs/btree_locking.h
fs/bcachefs/btree_types.h

index a856f5e90727360d1f756fb3a54c32a43c1ae19d..edb4084f7b9092768a79a2ecc9e0cc3f92a164aa 100644 (file)
@@ -17,8 +17,6 @@
 
 #include <linux/prefetch.h>
 
-static inline void btree_trans_sort_paths(struct btree_trans *);
-
 static inline void btree_path_list_remove(struct btree_trans *, struct btree_path *);
 static inline void btree_path_list_add(struct btree_trans *, struct btree_path *,
                                       struct btree_path *);
@@ -159,7 +157,7 @@ bool __bch2_btree_node_relock(struct btree_trans *trans,
        if (six_relock_type(&b->c.lock, want, path->l[level].lock_seq) ||
            (btree_node_lock_seq_matches(path, b, level) &&
             btree_node_lock_increment(trans, b, level, want))) {
-               mark_btree_node_locked(path, level, want);
+               mark_btree_node_locked(trans, path, level, want);
                return true;
        } else {
                return false;
@@ -195,7 +193,7 @@ static bool bch2_btree_node_upgrade(struct btree_trans *trans,
 
        return false;
 success:
-       mark_btree_node_intent_locked(path, level);
+       mark_btree_node_intent_locked(trans, path, level);
        return true;
 }
 
@@ -1078,7 +1076,7 @@ void bch2_trans_node_add(struct btree_trans *trans, struct btree *b)
                        t = btree_lock_want(path, b->c.level);
                        if (t != BTREE_NODE_UNLOCKED) {
                                six_lock_increment(&b->c.lock, (enum six_lock_type) t);
-                               mark_btree_node_locked(path, b->c.level, (enum six_lock_type) t);
+                               mark_btree_node_locked(trans, path, b->c.level, (enum six_lock_type) t);
                        }
 
                        btree_path_level_init(trans, path, b);
@@ -1167,7 +1165,7 @@ static inline int btree_path_lock_root(struct btree_trans *trans,
                        for (i = path->level + 1; i < BTREE_MAX_DEPTH; i++)
                                path->l[i].b = NULL;
 
-                       mark_btree_node_locked(path, path->level, lock_type);
+                       mark_btree_node_locked(trans, path, path->level, lock_type);
                        btree_path_level_init(trans, path, b);
                        return 0;
                }
@@ -1259,7 +1257,7 @@ static __always_inline int btree_path_down(struct btree_trans *trans,
        if (unlikely(ret))
                goto err;
 
-       mark_btree_node_locked(path, level, lock_type);
+       mark_btree_node_locked(trans, path, level, lock_type);
        btree_path_level_init(trans, path, b);
 
        if (tmp.k->k.type == KEY_TYPE_btree_ptr_v2 &&
@@ -1340,6 +1338,9 @@ retry_all:
                path = trans->paths + trans->sorted[i];
 
                EBUG_ON(!(trans->paths_allocated & (1ULL << path->idx)));
+#ifdef CONFIG_BCACHEFS_DEBUG
+               trans->traverse_all_idx = path->idx;
+#endif
 
                ret = btree_path_traverse_one(trans, path, 0, _THIS_IP_);
                if (ret)
@@ -1361,6 +1362,9 @@ retry_all:
 out:
        bch2_btree_cache_cannibalize_unlock(c);
 
+#ifdef CONFIG_BCACHEFS_DEBUG
+       trans->traverse_all_idx = U8_MAX;
+#endif
        trans->in_traverse_all = false;
 
        trace_trans_traverse_all(trans->ip, trace_ip);
@@ -2319,13 +2323,9 @@ static void btree_trans_verify_sorted_refs(struct btree_trans *trans)
                BUG_ON(trans->paths[idx].sorted_idx != i);
        }
 }
-#else
-static inline void btree_trans_verify_sorted_refs(struct btree_trans *trans) {}
-#endif
 
 static void btree_trans_verify_sorted(struct btree_trans *trans)
 {
-#ifdef CONFIG_BCACHEFS_DEBUG
        struct btree_path *path, *prev = NULL;
        unsigned i;
 
@@ -2333,14 +2333,22 @@ static void btree_trans_verify_sorted(struct btree_trans *trans)
                BUG_ON(prev && btree_path_cmp(prev, path) > 0);
                prev = path;
        }
-#endif
 }
+#else
+static inline void btree_trans_verify_sorted_refs(struct btree_trans *trans) {}
+static inline void btree_trans_verify_sorted(struct btree_trans *trans) {}
+#endif
 
-static noinline void __btree_trans_sort_paths(struct btree_trans *trans)
+void __bch2_btree_trans_sort_paths(struct btree_trans *trans)
 {
        int i, l = 0, r = trans->nr_sorted, inc = 1;
        bool swapped;
 
+       btree_trans_verify_sorted_refs(trans);
+
+       if (trans->paths_sorted)
+               goto out;
+
        /*
         * Cocktail shaker sort: this is efficient because iterators will be
         * mostly sorteda.
@@ -2368,22 +2376,10 @@ static noinline void __btree_trans_sort_paths(struct btree_trans *trans)
        } while (swapped);
 
        trans->paths_sorted = true;
-
+out:
        btree_trans_verify_sorted(trans);
 }
 
-static inline void btree_trans_sort_paths(struct btree_trans *trans)
-{
-       btree_trans_verify_sorted_refs(trans);
-
-       if (trans->paths_sorted) {
-               if (IS_ENABLED(CONFIG_BCACHEFS_DEBUG))
-                       btree_trans_verify_sorted(trans);
-               return;
-       }
-       __btree_trans_sort_paths(trans);
-}
-
 static inline void btree_path_list_remove(struct btree_trans *trans,
                                          struct btree_path *path)
 {
@@ -2410,7 +2406,7 @@ static inline void btree_path_list_add(struct btree_trans *trans,
 {
        unsigned i;
 
-       path->sorted_idx = pos ? pos->sorted_idx + 1 : 0;
+       path->sorted_idx = pos ? pos->sorted_idx + 1 : trans->nr_sorted;
 
 #ifdef CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS
        memmove_u64s_up_small(trans->sorted + path->sorted_idx + 1,
index 983d611224580fbac654d749a07decc15df1a3e3..19ce6a6ece7d59093f6c3ca5f465058653152900 100644 (file)
@@ -57,6 +57,16 @@ static inline int btree_iter_err(const struct btree_iter *iter)
 
 /* Iterate over paths within a transaction: */
 
+void __bch2_btree_trans_sort_paths(struct btree_trans *);
+
+static inline void btree_trans_sort_paths(struct btree_trans *trans)
+{
+       if (!IS_ENABLED(CONFIG_BCACHEFS_DEBUG) &&
+           trans->paths_sorted)
+               return;
+       __bch2_btree_trans_sort_paths(trans);
+}
+
 static inline struct btree_path *
 __trans_next_path(struct btree_trans *trans, unsigned idx)
 {
index 9bdc2c3f21bf4869bf9af0ca45c8080edca6cef2..9d3a64f37f09af27fb13290b3daa8273fca72ce8 100644 (file)
@@ -297,7 +297,7 @@ retry:
                if (!ck)
                        goto retry;
 
-               mark_btree_node_locked(path, 0, SIX_LOCK_intent);
+               mark_btree_node_locked(trans, path, 0, SIX_LOCK_intent);
                path->locks_want = 1;
        } else {
                enum six_lock_type lock_want = __btree_lock_want(path, 0);
@@ -319,7 +319,7 @@ retry:
                        goto retry;
                }
 
-               mark_btree_node_locked(path, 0, lock_want);
+               mark_btree_node_locked(trans, path, 0, lock_want);
        }
 
        path->l[0].lock_seq     = ck->c.lock.state.seq;
index d05689180c63e855cbd4d34ae303206cb2194e09..ff58870311f3580d7f01ba566a70455d0fe7e50b 100644 (file)
@@ -57,7 +57,8 @@ static inline void mark_btree_node_unlocked(struct btree_path *path,
        path->nodes_intent_locked &= ~(1 << level);
 }
 
-static inline void mark_btree_node_locked(struct btree_path *path,
+static inline void mark_btree_node_locked(struct btree_trans *trans,
+                                         struct btree_path *path,
                                          unsigned level,
                                          enum six_lock_type type)
 {
@@ -67,12 +68,20 @@ static inline void mark_btree_node_locked(struct btree_path *path,
 
        path->nodes_locked |= 1 << level;
        path->nodes_intent_locked |= type << level;
+#ifdef CONFIG_BCACHEFS_DEBUG
+       path->ip_locked = _RET_IP_;
+       btree_trans_sort_paths(trans);
+       BUG_ON(trans->in_traverse_all &&
+              trans->traverse_all_idx != U8_MAX &&
+              path->sorted_idx > trans->paths[trans->traverse_all_idx].sorted_idx);
+#endif
 }
 
-static inline void mark_btree_node_intent_locked(struct btree_path *path,
+static inline void mark_btree_node_intent_locked(struct btree_trans *trans,
+                                                struct btree_path *path,
                                                 unsigned level)
 {
-       mark_btree_node_locked(path, level, SIX_LOCK_intent);
+       mark_btree_node_locked(trans, path, level, SIX_LOCK_intent);
 }
 
 static inline enum six_lock_type __btree_lock_want(struct btree_path *path, int level)
@@ -111,6 +120,9 @@ static inline void __bch2_btree_path_unlock(struct btree_path *path)
 
        while (path->nodes_locked)
                btree_node_unlock(path, __ffs(path->nodes_locked));
+#ifdef CONFIG_BCACHEFS_DEBUG
+       path->ip_locked = 0;
+#endif
 }
 
 static inline enum bch_time_stats lock_to_time_stat(enum six_lock_type type)
index b7cded2095fffa4069694ee0a0ded9608e4bda48..ce64d3ad768bb59f28f2d2584a48e4f03e6921e5 100644 (file)
@@ -255,6 +255,7 @@ struct btree_path {
        }                       l[BTREE_MAX_DEPTH];
 #ifdef CONFIG_BCACHEFS_DEBUG
        unsigned long           ip_allocated;
+       unsigned long           ip_locked;
 #endif
 };
 
@@ -368,6 +369,7 @@ struct btree_trans {
        struct bpos             locking_pos;
        u8                      locking_btree_id;
        u8                      locking_level;
+       u8                      traverse_all_idx;
        pid_t                   pid;
 #endif
        unsigned long           ip;