]> www.infradead.org Git - users/hch/misc.git/commitdiff
btrfs: implement ref_tracker for delayed_nodes
authorLeo Martins <loemra.dev@gmail.com>
Tue, 12 Aug 2025 23:04:39 +0000 (16:04 -0700)
committerDavid Sterba <dsterba@suse.com>
Mon, 22 Sep 2025 08:54:32 +0000 (10:54 +0200)
Add ref_tracker infrastructure for struct btrfs_delayed_node.

It is a response to the largest btrfs related crash in our fleet.  We're
seeing soft lockups in btrfs_kill_all_delayed_nodes() that seem to be a
result of delayed_nodes not being released properly.

A ref_tracker object is allocated on reference count increases and freed
on reference count decreases. The ref_tracker object stores a stack
trace of where it is allocated. The ref_tracker_dir object is embedded
in btrfs_delayed_node and keeps track of all current and some old/freed
ref_tracker objects. When a leak is detected we can print the stack
traces for all ref_trackers that have not yet been freed.

Here is a common example of taking a reference to a delayed_node and
freeing it with ref_tracker.

    struct btrfs_ref_tracker tracker;
    struct btrfs_delayed_node *node;

    node = btrfs_get_delayed_node(inode, &tracker);
    // use delayed_node...
    btrfs_release_delayed_node(node, &tracker);

There are two special cases where the delayed_node reference is "long
lived", meaning that the thread that takes the reference and the thread
that releases the reference are different. The 'inode_cache_tracker'
tracks the delayed_node stored in btrfs_inode. The 'node_list_tracker'
tracks the delayed_node stored in the btrfs_delayed_root
node_list/prepare_list. These trackers are embedded in the
btrfs_delayed_node.

btrfs_ref_tracker and btrfs_ref_tracker_dir are wrappers that either
compile to the corresponding ref_tracker structs or empty structs
depending on CONFIG_BTRFS_DEBUG. There are also btrfs wrappers for
the ref_tracker API.

Signed-off-by: Leo Martins <loemra.dev@gmail.com>
Reviewed-by: David Sterba <dsterba@suse.com>
Signed-off-by: David Sterba <dsterba@suse.com>
fs/btrfs/Kconfig
fs/btrfs/delayed-inode.c
fs/btrfs/delayed-inode.h

index 86d791210c393fe604d96faff9d1b0e383aec78a..4438637c8900cdfe64d6d00404d547785930639e 100644 (file)
@@ -62,6 +62,7 @@ config BTRFS_FS_RUN_SANITY_TESTS
 config BTRFS_DEBUG
        bool "Btrfs debugging support"
        depends on BTRFS_FS
+       select REF_TRACKER if STACKTRACE_SUPPORT
        help
          Enable run-time debugging support for the btrfs filesystem.
 
index c0c1ddd46b67a23a50607f6be14eb2fa20c359eb..364814642a91d6b0f4efc68ccfbe5a84c3cb830b 100644 (file)
@@ -57,6 +57,7 @@ static inline void btrfs_init_delayed_node(
        delayed_node->root = root;
        delayed_node->inode_id = inode_id;
        refcount_set(&delayed_node->refs, 0);
+       btrfs_delayed_node_ref_tracker_dir_init(delayed_node);
        delayed_node->ins_root = RB_ROOT_CACHED;
        delayed_node->del_root = RB_ROOT_CACHED;
        mutex_init(&delayed_node->mutex);
@@ -65,7 +66,8 @@ static inline void btrfs_init_delayed_node(
 }
 
 static struct btrfs_delayed_node *btrfs_get_delayed_node(
-               struct btrfs_inode *btrfs_inode)
+               struct btrfs_inode *btrfs_inode,
+               struct btrfs_ref_tracker *tracker)
 {
        struct btrfs_root *root = btrfs_inode->root;
        u64 ino = btrfs_ino(btrfs_inode);
@@ -74,6 +76,7 @@ static struct btrfs_delayed_node *btrfs_get_delayed_node(
        node = READ_ONCE(btrfs_inode->delayed_node);
        if (node) {
                refcount_inc(&node->refs);
+               btrfs_delayed_node_ref_tracker_alloc(node, tracker, GFP_NOFS);
                return node;
        }
 
@@ -83,6 +86,7 @@ static struct btrfs_delayed_node *btrfs_get_delayed_node(
        if (node) {
                if (btrfs_inode->delayed_node) {
                        refcount_inc(&node->refs);      /* can be accessed */
+                       btrfs_delayed_node_ref_tracker_alloc(node, tracker, GFP_ATOMIC);
                        BUG_ON(btrfs_inode->delayed_node != node);
                        xa_unlock(&root->delayed_nodes);
                        return node;
@@ -106,6 +110,9 @@ static struct btrfs_delayed_node *btrfs_get_delayed_node(
                 */
                if (refcount_inc_not_zero(&node->refs)) {
                        refcount_inc(&node->refs);
+                       btrfs_delayed_node_ref_tracker_alloc(node, tracker, GFP_ATOMIC);
+                       btrfs_delayed_node_ref_tracker_alloc(node, &node->inode_cache_tracker,
+                                                            GFP_ATOMIC);
                        btrfs_inode->delayed_node = node;
                } else {
                        node = NULL;
@@ -126,7 +133,8 @@ static struct btrfs_delayed_node *btrfs_get_delayed_node(
  * Return the delayed node, or error pointer on failure.
  */
 static struct btrfs_delayed_node *btrfs_get_or_create_delayed_node(
-               struct btrfs_inode *btrfs_inode)
+               struct btrfs_inode *btrfs_inode,
+               struct btrfs_ref_tracker *tracker)
 {
        struct btrfs_delayed_node *node;
        struct btrfs_root *root = btrfs_inode->root;
@@ -135,7 +143,7 @@ static struct btrfs_delayed_node *btrfs_get_or_create_delayed_node(
        void *ptr;
 
 again:
-       node = btrfs_get_delayed_node(btrfs_inode);
+       node = btrfs_get_delayed_node(btrfs_inode, tracker);
        if (node)
                return node;
 
@@ -144,12 +152,10 @@ again:
                return ERR_PTR(-ENOMEM);
        btrfs_init_delayed_node(node, root, ino);
 
-       /* Cached in the inode and can be accessed. */
-       refcount_set(&node->refs, 2);
-
        /* Allocate and reserve the slot, from now it can return a NULL from xa_load(). */
        ret = xa_reserve(&root->delayed_nodes, ino, GFP_NOFS);
        if (ret == -ENOMEM) {
+               btrfs_delayed_node_ref_tracker_dir_exit(node);
                kmem_cache_free(delayed_node_cache, node);
                return ERR_PTR(-ENOMEM);
        }
@@ -158,6 +164,7 @@ again:
        if (ptr) {
                /* Somebody inserted it, go back and read it. */
                xa_unlock(&root->delayed_nodes);
+               btrfs_delayed_node_ref_tracker_dir_exit(node);
                kmem_cache_free(delayed_node_cache, node);
                node = NULL;
                goto again;
@@ -166,6 +173,12 @@ again:
        ASSERT(xa_err(ptr) != -EINVAL);
        ASSERT(xa_err(ptr) != -ENOMEM);
        ASSERT(ptr == NULL);
+
+       /* Cached in the inode and can be accessed. */
+       refcount_set(&node->refs, 2);
+       btrfs_delayed_node_ref_tracker_alloc(node, tracker, GFP_ATOMIC);
+       btrfs_delayed_node_ref_tracker_alloc(node, &node->inode_cache_tracker, GFP_ATOMIC);
+
        btrfs_inode->delayed_node = node;
        xa_unlock(&root->delayed_nodes);
 
@@ -191,6 +204,8 @@ static void btrfs_queue_delayed_node(struct btrfs_delayed_root *root,
                list_add_tail(&node->n_list, &root->node_list);
                list_add_tail(&node->p_list, &root->prepare_list);
                refcount_inc(&node->refs);      /* inserted into list */
+               btrfs_delayed_node_ref_tracker_alloc(node, &node->node_list_tracker,
+                                                    GFP_ATOMIC);
                root->nodes++;
                set_bit(BTRFS_DELAYED_NODE_IN_LIST, &node->flags);
        }
@@ -204,6 +219,7 @@ static void btrfs_dequeue_delayed_node(struct btrfs_delayed_root *root,
        spin_lock(&root->lock);
        if (test_bit(BTRFS_DELAYED_NODE_IN_LIST, &node->flags)) {
                root->nodes--;
+               btrfs_delayed_node_ref_tracker_free(node, &node->node_list_tracker);
                refcount_dec(&node->refs);      /* not in the list */
                list_del_init(&node->n_list);
                if (!list_empty(&node->p_list))
@@ -214,22 +230,26 @@ static void btrfs_dequeue_delayed_node(struct btrfs_delayed_root *root,
 }
 
 static struct btrfs_delayed_node *btrfs_first_delayed_node(
-                       struct btrfs_delayed_root *delayed_root)
+                       struct btrfs_delayed_root *delayed_root,
+                       struct btrfs_ref_tracker *tracker)
 {
        struct btrfs_delayed_node *node;
 
        spin_lock(&delayed_root->lock);
        node = list_first_entry_or_null(&delayed_root->node_list,
                                        struct btrfs_delayed_node, n_list);
-       if (node)
+       if (node) {
                refcount_inc(&node->refs);
+               btrfs_delayed_node_ref_tracker_alloc(node, tracker, GFP_ATOMIC);
+       }
        spin_unlock(&delayed_root->lock);
 
        return node;
 }
 
 static struct btrfs_delayed_node *btrfs_next_delayed_node(
-                                               struct btrfs_delayed_node *node)
+                                               struct btrfs_delayed_node *node,
+                                               struct btrfs_ref_tracker *tracker)
 {
        struct btrfs_delayed_root *delayed_root;
        struct list_head *p;
@@ -249,6 +269,7 @@ static struct btrfs_delayed_node *btrfs_next_delayed_node(
 
        next = list_entry(p, struct btrfs_delayed_node, n_list);
        refcount_inc(&next->refs);
+       btrfs_delayed_node_ref_tracker_alloc(next, tracker, GFP_ATOMIC);
 out:
        spin_unlock(&delayed_root->lock);
 
@@ -257,7 +278,7 @@ out:
 
 static void __btrfs_release_delayed_node(
                                struct btrfs_delayed_node *delayed_node,
-                               int mod)
+                               int mod, struct btrfs_ref_tracker *tracker)
 {
        struct btrfs_delayed_root *delayed_root;
 
@@ -273,6 +294,7 @@ static void __btrfs_release_delayed_node(
                btrfs_dequeue_delayed_node(delayed_root, delayed_node);
        mutex_unlock(&delayed_node->mutex);
 
+       btrfs_delayed_node_ref_tracker_free(delayed_node, tracker);
        if (refcount_dec_and_test(&delayed_node->refs)) {
                struct btrfs_root *root = delayed_node->root;
 
@@ -282,17 +304,20 @@ static void __btrfs_release_delayed_node(
                 * back up.  We can delete it now.
                 */
                ASSERT(refcount_read(&delayed_node->refs) == 0);
+               btrfs_delayed_node_ref_tracker_dir_exit(delayed_node);
                kmem_cache_free(delayed_node_cache, delayed_node);
        }
 }
 
-static inline void btrfs_release_delayed_node(struct btrfs_delayed_node *node)
+static inline void btrfs_release_delayed_node(struct btrfs_delayed_node *node,
+                                             struct btrfs_ref_tracker *tracker)
 {
-       __btrfs_release_delayed_node(node, 0);
+       __btrfs_release_delayed_node(node, 0, tracker);
 }
 
 static struct btrfs_delayed_node *btrfs_first_prepared_delayed_node(
-                                       struct btrfs_delayed_root *delayed_root)
+                                       struct btrfs_delayed_root *delayed_root,
+                                       struct btrfs_ref_tracker *tracker)
 {
        struct btrfs_delayed_node *node;
 
@@ -302,6 +327,7 @@ static struct btrfs_delayed_node *btrfs_first_prepared_delayed_node(
        if (node) {
                list_del_init(&node->p_list);
                refcount_inc(&node->refs);
+               btrfs_delayed_node_ref_tracker_alloc(node, tracker, GFP_ATOMIC);
        }
        spin_unlock(&delayed_root->lock);
 
@@ -309,9 +335,10 @@ static struct btrfs_delayed_node *btrfs_first_prepared_delayed_node(
 }
 
 static inline void btrfs_release_prepared_delayed_node(
-                                       struct btrfs_delayed_node *node)
+                                       struct btrfs_delayed_node *node,
+                                       struct btrfs_ref_tracker *tracker)
 {
-       __btrfs_release_delayed_node(node, 1);
+       __btrfs_release_delayed_node(node, 1, tracker);
 }
 
 static struct btrfs_delayed_item *btrfs_alloc_delayed_item(u16 data_len,
@@ -1126,6 +1153,7 @@ static int __btrfs_run_delayed_items(struct btrfs_trans_handle *trans, int nr)
        struct btrfs_fs_info *fs_info = trans->fs_info;
        struct btrfs_delayed_root *delayed_root;
        struct btrfs_delayed_node *curr_node, *prev_node;
+       struct btrfs_ref_tracker curr_delayed_node_tracker, prev_delayed_node_tracker;
        struct btrfs_path *path;
        struct btrfs_block_rsv *block_rsv;
        int ret = 0;
@@ -1143,7 +1171,7 @@ static int __btrfs_run_delayed_items(struct btrfs_trans_handle *trans, int nr)
 
        delayed_root = fs_info->delayed_root;
 
-       curr_node = btrfs_first_delayed_node(delayed_root);
+       curr_node = btrfs_first_delayed_node(delayed_root, &curr_delayed_node_tracker);
        while (curr_node && (!count || nr--)) {
                ret = __btrfs_commit_inode_delayed_items(trans, path,
                                                         curr_node);
@@ -1153,7 +1181,8 @@ static int __btrfs_run_delayed_items(struct btrfs_trans_handle *trans, int nr)
                }
 
                prev_node = curr_node;
-               curr_node = btrfs_next_delayed_node(curr_node);
+               prev_delayed_node_tracker = curr_delayed_node_tracker;
+               curr_node = btrfs_next_delayed_node(curr_node, &curr_delayed_node_tracker);
                /*
                 * See the comment below about releasing path before releasing
                 * node. If the commit of delayed items was successful the path
@@ -1161,7 +1190,7 @@ static int __btrfs_run_delayed_items(struct btrfs_trans_handle *trans, int nr)
                 * point to locked extent buffers (a leaf at the very least).
                 */
                ASSERT(path->nodes[0] == NULL);
-               btrfs_release_delayed_node(prev_node);
+               btrfs_release_delayed_node(prev_node, &prev_delayed_node_tracker);
        }
 
        /*
@@ -1174,7 +1203,7 @@ static int __btrfs_run_delayed_items(struct btrfs_trans_handle *trans, int nr)
        btrfs_free_path(path);
 
        if (curr_node)
-               btrfs_release_delayed_node(curr_node);
+               btrfs_release_delayed_node(curr_node, &curr_delayed_node_tracker);
        trans->block_rsv = block_rsv;
 
        return ret;
@@ -1193,7 +1222,9 @@ int btrfs_run_delayed_items_nr(struct btrfs_trans_handle *trans, int nr)
 int btrfs_commit_inode_delayed_items(struct btrfs_trans_handle *trans,
                                     struct btrfs_inode *inode)
 {
-       struct btrfs_delayed_node *delayed_node = btrfs_get_delayed_node(inode);
+       struct btrfs_ref_tracker delayed_node_tracker;
+       struct btrfs_delayed_node *delayed_node =
+               btrfs_get_delayed_node(inode, &delayed_node_tracker);
        BTRFS_PATH_AUTO_FREE(path);
        struct btrfs_block_rsv *block_rsv;
        int ret;
@@ -1204,14 +1235,14 @@ int btrfs_commit_inode_delayed_items(struct btrfs_trans_handle *trans,
        mutex_lock(&delayed_node->mutex);
        if (!delayed_node->count) {
                mutex_unlock(&delayed_node->mutex);
-               btrfs_release_delayed_node(delayed_node);
+               btrfs_release_delayed_node(delayed_node, &delayed_node_tracker);
                return 0;
        }
        mutex_unlock(&delayed_node->mutex);
 
        path = btrfs_alloc_path();
        if (!path) {
-               btrfs_release_delayed_node(delayed_node);
+               btrfs_release_delayed_node(delayed_node, &delayed_node_tracker);
                return -ENOMEM;
        }
 
@@ -1220,7 +1251,7 @@ int btrfs_commit_inode_delayed_items(struct btrfs_trans_handle *trans,
 
        ret = __btrfs_commit_inode_delayed_items(trans, path, delayed_node);
 
-       btrfs_release_delayed_node(delayed_node);
+       btrfs_release_delayed_node(delayed_node, &delayed_node_tracker);
        trans->block_rsv = block_rsv;
 
        return ret;
@@ -1230,18 +1261,20 @@ int btrfs_commit_inode_delayed_inode(struct btrfs_inode *inode)
 {
        struct btrfs_fs_info *fs_info = inode->root->fs_info;
        struct btrfs_trans_handle *trans;
-       struct btrfs_delayed_node *delayed_node = btrfs_get_delayed_node(inode);
+       struct btrfs_ref_tracker delayed_node_tracker;
+       struct btrfs_delayed_node *delayed_node;
        struct btrfs_path *path;
        struct btrfs_block_rsv *block_rsv;
        int ret;
 
+       delayed_node = btrfs_get_delayed_node(inode, &delayed_node_tracker);
        if (!delayed_node)
                return 0;
 
        mutex_lock(&delayed_node->mutex);
        if (!test_bit(BTRFS_DELAYED_NODE_INODE_DIRTY, &delayed_node->flags)) {
                mutex_unlock(&delayed_node->mutex);
-               btrfs_release_delayed_node(delayed_node);
+               btrfs_release_delayed_node(delayed_node, &delayed_node_tracker);
                return 0;
        }
        mutex_unlock(&delayed_node->mutex);
@@ -1275,7 +1308,7 @@ trans_out:
        btrfs_end_transaction(trans);
        btrfs_btree_balance_dirty(fs_info);
 out:
-       btrfs_release_delayed_node(delayed_node);
+       btrfs_release_delayed_node(delayed_node, &delayed_node_tracker);
 
        return ret;
 }
@@ -1289,7 +1322,8 @@ void btrfs_remove_delayed_node(struct btrfs_inode *inode)
                return;
 
        inode->delayed_node = NULL;
-       btrfs_release_delayed_node(delayed_node);
+
+       btrfs_release_delayed_node(delayed_node, &delayed_node->inode_cache_tracker);
 }
 
 struct btrfs_async_delayed_work {
@@ -1305,6 +1339,7 @@ static void btrfs_async_run_delayed_root(struct btrfs_work *work)
        struct btrfs_trans_handle *trans;
        struct btrfs_path *path;
        struct btrfs_delayed_node *delayed_node = NULL;
+       struct btrfs_ref_tracker delayed_node_tracker;
        struct btrfs_root *root;
        struct btrfs_block_rsv *block_rsv;
        int total_done = 0;
@@ -1321,7 +1356,8 @@ static void btrfs_async_run_delayed_root(struct btrfs_work *work)
                    BTRFS_DELAYED_BACKGROUND / 2)
                        break;
 
-               delayed_node = btrfs_first_prepared_delayed_node(delayed_root);
+               delayed_node = btrfs_first_prepared_delayed_node(delayed_root,
+                                                                &delayed_node_tracker);
                if (!delayed_node)
                        break;
 
@@ -1330,7 +1366,8 @@ static void btrfs_async_run_delayed_root(struct btrfs_work *work)
                trans = btrfs_join_transaction(root);
                if (IS_ERR(trans)) {
                        btrfs_release_path(path);
-                       btrfs_release_prepared_delayed_node(delayed_node);
+                       btrfs_release_prepared_delayed_node(delayed_node,
+                                                           &delayed_node_tracker);
                        total_done++;
                        continue;
                }
@@ -1345,7 +1382,8 @@ static void btrfs_async_run_delayed_root(struct btrfs_work *work)
                btrfs_btree_balance_dirty_nodelay(root->fs_info);
 
                btrfs_release_path(path);
-               btrfs_release_prepared_delayed_node(delayed_node);
+               btrfs_release_prepared_delayed_node(delayed_node,
+                                                   &delayed_node_tracker);
                total_done++;
 
        } while ((async_work->nr == 0 && total_done < BTRFS_DELAYED_WRITEBACK)
@@ -1377,10 +1415,15 @@ static int btrfs_wq_run_delayed_node(struct btrfs_delayed_root *delayed_root,
 
 void btrfs_assert_delayed_root_empty(struct btrfs_fs_info *fs_info)
 {
-       struct btrfs_delayed_node *node = btrfs_first_delayed_node(fs_info->delayed_root);
+       struct btrfs_ref_tracker delayed_node_tracker;
+       struct btrfs_delayed_node *node;
 
-       if (WARN_ON(node))
+       node = btrfs_first_delayed_node( fs_info->delayed_root, &delayed_node_tracker);
+       if (WARN_ON(node)) {
+               btrfs_delayed_node_ref_tracker_free(node,
+                                                   &delayed_node_tracker);
                refcount_dec(&node->refs);
+       }
 }
 
 static bool could_end_wait(struct btrfs_delayed_root *delayed_root, int seq)
@@ -1454,13 +1497,14 @@ int btrfs_insert_delayed_dir_index(struct btrfs_trans_handle *trans,
        struct btrfs_fs_info *fs_info = trans->fs_info;
        const unsigned int leaf_data_size = BTRFS_LEAF_DATA_SIZE(fs_info);
        struct btrfs_delayed_node *delayed_node;
+       struct btrfs_ref_tracker delayed_node_tracker;
        struct btrfs_delayed_item *delayed_item;
        struct btrfs_dir_item *dir_item;
        bool reserve_leaf_space;
        u32 data_len;
        int ret;
 
-       delayed_node = btrfs_get_or_create_delayed_node(dir);
+       delayed_node = btrfs_get_or_create_delayed_node(dir, &delayed_node_tracker);
        if (IS_ERR(delayed_node))
                return PTR_ERR(delayed_node);
 
@@ -1536,7 +1580,7 @@ int btrfs_insert_delayed_dir_index(struct btrfs_trans_handle *trans,
        mutex_unlock(&delayed_node->mutex);
 
 release_node:
-       btrfs_release_delayed_node(delayed_node);
+       btrfs_release_delayed_node(delayed_node, &delayed_node_tracker);
        return ret;
 }
 
@@ -1591,10 +1635,11 @@ int btrfs_delete_delayed_dir_index(struct btrfs_trans_handle *trans,
                                   struct btrfs_inode *dir, u64 index)
 {
        struct btrfs_delayed_node *node;
+       struct btrfs_ref_tracker delayed_node_tracker;
        struct btrfs_delayed_item *item;
        int ret;
 
-       node = btrfs_get_or_create_delayed_node(dir);
+       node = btrfs_get_or_create_delayed_node(dir, &delayed_node_tracker);
        if (IS_ERR(node))
                return PTR_ERR(node);
 
@@ -1635,14 +1680,16 @@ int btrfs_delete_delayed_dir_index(struct btrfs_trans_handle *trans,
        }
        mutex_unlock(&node->mutex);
 end:
-       btrfs_release_delayed_node(node);
+       btrfs_release_delayed_node(node, &delayed_node_tracker);
        return ret;
 }
 
 int btrfs_inode_delayed_dir_index_count(struct btrfs_inode *inode)
 {
-       struct btrfs_delayed_node *delayed_node = btrfs_get_delayed_node(inode);
+       struct btrfs_ref_tracker delayed_node_tracker;
+       struct btrfs_delayed_node *delayed_node;
 
+       delayed_node = btrfs_get_delayed_node(inode, &delayed_node_tracker);
        if (!delayed_node)
                return -ENOENT;
 
@@ -1652,12 +1699,12 @@ int btrfs_inode_delayed_dir_index_count(struct btrfs_inode *inode)
         * is updated now. So we needn't lock the delayed node.
         */
        if (!delayed_node->index_cnt) {
-               btrfs_release_delayed_node(delayed_node);
+               btrfs_release_delayed_node(delayed_node, &delayed_node_tracker);
                return -EINVAL;
        }
 
        inode->index_cnt = delayed_node->index_cnt;
-       btrfs_release_delayed_node(delayed_node);
+       btrfs_release_delayed_node(delayed_node, &delayed_node_tracker);
        return 0;
 }
 
@@ -1668,8 +1715,9 @@ bool btrfs_readdir_get_delayed_items(struct btrfs_inode *inode,
 {
        struct btrfs_delayed_node *delayed_node;
        struct btrfs_delayed_item *item;
+       struct btrfs_ref_tracker delayed_node_tracker;
 
-       delayed_node = btrfs_get_delayed_node(inode);
+       delayed_node = btrfs_get_delayed_node(inode, &delayed_node_tracker);
        if (!delayed_node)
                return false;
 
@@ -1704,6 +1752,7 @@ bool btrfs_readdir_get_delayed_items(struct btrfs_inode *inode,
         * insert/delete delayed items in this period. So we also needn't
         * requeue or dequeue this delayed node.
         */
+       btrfs_delayed_node_ref_tracker_free(delayed_node, &delayed_node_tracker);
        refcount_dec(&delayed_node->refs);
 
        return true;
@@ -1844,17 +1893,18 @@ static void fill_stack_inode_item(struct btrfs_trans_handle *trans,
 int btrfs_fill_inode(struct btrfs_inode *inode, u32 *rdev)
 {
        struct btrfs_delayed_node *delayed_node;
+       struct btrfs_ref_tracker delayed_node_tracker;
        struct btrfs_inode_item *inode_item;
        struct inode *vfs_inode = &inode->vfs_inode;
 
-       delayed_node = btrfs_get_delayed_node(inode);
+       delayed_node = btrfs_get_delayed_node(inode, &delayed_node_tracker);
        if (!delayed_node)
                return -ENOENT;
 
        mutex_lock(&delayed_node->mutex);
        if (!test_bit(BTRFS_DELAYED_NODE_INODE_DIRTY, &delayed_node->flags)) {
                mutex_unlock(&delayed_node->mutex);
-               btrfs_release_delayed_node(delayed_node);
+               btrfs_release_delayed_node(delayed_node, &delayed_node_tracker);
                return -ENOENT;
        }
 
@@ -1892,7 +1942,7 @@ int btrfs_fill_inode(struct btrfs_inode *inode, u32 *rdev)
                inode->index_cnt = (u64)-1;
 
        mutex_unlock(&delayed_node->mutex);
-       btrfs_release_delayed_node(delayed_node);
+       btrfs_release_delayed_node(delayed_node, &delayed_node_tracker);
        return 0;
 }
 
@@ -1901,9 +1951,10 @@ int btrfs_delayed_update_inode(struct btrfs_trans_handle *trans,
 {
        struct btrfs_root *root = inode->root;
        struct btrfs_delayed_node *delayed_node;
+       struct btrfs_ref_tracker delayed_node_tracker;
        int ret = 0;
 
-       delayed_node = btrfs_get_or_create_delayed_node(inode);
+       delayed_node = btrfs_get_or_create_delayed_node(inode, &delayed_node_tracker);
        if (IS_ERR(delayed_node))
                return PTR_ERR(delayed_node);
 
@@ -1923,7 +1974,7 @@ int btrfs_delayed_update_inode(struct btrfs_trans_handle *trans,
        atomic_inc(&root->fs_info->delayed_root->items);
 release_node:
        mutex_unlock(&delayed_node->mutex);
-       btrfs_release_delayed_node(delayed_node);
+       btrfs_release_delayed_node(delayed_node, &delayed_node_tracker);
        return ret;
 }
 
@@ -1931,6 +1982,7 @@ int btrfs_delayed_delete_inode_ref(struct btrfs_inode *inode)
 {
        struct btrfs_fs_info *fs_info = inode->root->fs_info;
        struct btrfs_delayed_node *delayed_node;
+       struct btrfs_ref_tracker delayed_node_tracker;
 
        /*
         * we don't do delayed inode updates during log recovery because it
@@ -1940,7 +1992,7 @@ int btrfs_delayed_delete_inode_ref(struct btrfs_inode *inode)
        if (test_bit(BTRFS_FS_LOG_RECOVERING, &fs_info->flags))
                return -EAGAIN;
 
-       delayed_node = btrfs_get_or_create_delayed_node(inode);
+       delayed_node = btrfs_get_or_create_delayed_node(inode, &delayed_node_tracker);
        if (IS_ERR(delayed_node))
                return PTR_ERR(delayed_node);
 
@@ -1967,7 +2019,7 @@ int btrfs_delayed_delete_inode_ref(struct btrfs_inode *inode)
        atomic_inc(&fs_info->delayed_root->items);
 release_node:
        mutex_unlock(&delayed_node->mutex);
-       btrfs_release_delayed_node(delayed_node);
+       btrfs_release_delayed_node(delayed_node, &delayed_node_tracker);
        return 0;
 }
 
@@ -2011,19 +2063,21 @@ static void __btrfs_kill_delayed_node(struct btrfs_delayed_node *delayed_node)
 void btrfs_kill_delayed_inode_items(struct btrfs_inode *inode)
 {
        struct btrfs_delayed_node *delayed_node;
+       struct btrfs_ref_tracker delayed_node_tracker;
 
-       delayed_node = btrfs_get_delayed_node(inode);
+       delayed_node = btrfs_get_delayed_node(inode, &delayed_node_tracker);
        if (!delayed_node)
                return;
 
        __btrfs_kill_delayed_node(delayed_node);
-       btrfs_release_delayed_node(delayed_node);
+       btrfs_release_delayed_node(delayed_node, &delayed_node_tracker);
 }
 
 void btrfs_kill_all_delayed_nodes(struct btrfs_root *root)
 {
        unsigned long index = 0;
        struct btrfs_delayed_node *delayed_nodes[8];
+       struct btrfs_ref_tracker delayed_node_trackers[8];
 
        while (1) {
                struct btrfs_delayed_node *node;
@@ -2042,6 +2096,9 @@ void btrfs_kill_all_delayed_nodes(struct btrfs_root *root)
                         * about to be removed from the tree in the loop below
                         */
                        if (refcount_inc_not_zero(&node->refs)) {
+                               btrfs_delayed_node_ref_tracker_alloc(node,
+                                                    &delayed_node_trackers[count],
+                                                    GFP_ATOMIC);
                                delayed_nodes[count] = node;
                                count++;
                        }
@@ -2053,7 +2110,8 @@ void btrfs_kill_all_delayed_nodes(struct btrfs_root *root)
 
                for (int i = 0; i < count; i++) {
                        __btrfs_kill_delayed_node(delayed_nodes[i]);
-                       btrfs_release_delayed_node(delayed_nodes[i]);
+                       btrfs_release_delayed_node(delayed_nodes[i],
+                                                  &delayed_node_trackers[i]);
                }
        }
 }
@@ -2061,14 +2119,17 @@ void btrfs_kill_all_delayed_nodes(struct btrfs_root *root)
 void btrfs_destroy_delayed_inodes(struct btrfs_fs_info *fs_info)
 {
        struct btrfs_delayed_node *curr_node, *prev_node;
+       struct btrfs_ref_tracker curr_delayed_node_tracker, prev_delayed_node_tracker;
 
-       curr_node = btrfs_first_delayed_node(fs_info->delayed_root);
+       curr_node = btrfs_first_delayed_node(fs_info->delayed_root,
+                                            &curr_delayed_node_tracker);
        while (curr_node) {
                __btrfs_kill_delayed_node(curr_node);
 
                prev_node = curr_node;
-               curr_node = btrfs_next_delayed_node(curr_node);
-               btrfs_release_delayed_node(prev_node);
+               prev_delayed_node_tracker = curr_delayed_node_tracker;
+               curr_node = btrfs_next_delayed_node(curr_node, &curr_delayed_node_tracker);
+               btrfs_release_delayed_node(prev_node, &prev_delayed_node_tracker);
        }
 }
 
@@ -2078,8 +2139,9 @@ void btrfs_log_get_delayed_items(struct btrfs_inode *inode,
 {
        struct btrfs_delayed_node *node;
        struct btrfs_delayed_item *item;
+       struct btrfs_ref_tracker delayed_node_tracker;
 
-       node = btrfs_get_delayed_node(inode);
+       node = btrfs_get_delayed_node(inode, &delayed_node_tracker);
        if (!node)
                return;
 
@@ -2137,6 +2199,7 @@ void btrfs_log_get_delayed_items(struct btrfs_inode *inode,
         * delete delayed items.
         */
        ASSERT(refcount_read(&node->refs) > 1);
+       btrfs_delayed_node_ref_tracker_free(node, &delayed_node_tracker);
        refcount_dec(&node->refs);
 }
 
@@ -2147,8 +2210,9 @@ void btrfs_log_put_delayed_items(struct btrfs_inode *inode,
        struct btrfs_delayed_node *node;
        struct btrfs_delayed_item *item;
        struct btrfs_delayed_item *next;
+       struct btrfs_ref_tracker delayed_node_tracker;
 
-       node = btrfs_get_delayed_node(inode);
+       node = btrfs_get_delayed_node(inode, &delayed_node_tracker);
        if (!node)
                return;
 
@@ -2180,5 +2244,6 @@ void btrfs_log_put_delayed_items(struct btrfs_inode *inode,
         * delete delayed items.
         */
        ASSERT(refcount_read(&node->refs) > 1);
+       btrfs_delayed_node_ref_tracker_free(node, &delayed_node_tracker);
        refcount_dec(&node->refs);
 }
index e6e763ad2d421ffff8abdf01074deeefcf4bfc9b..93149c32d3fa41875d84bddba03d5db4038260b2 100644 (file)
@@ -16,6 +16,7 @@
 #include <linux/fs.h>
 #include <linux/atomic.h>
 #include <linux/refcount.h>
+#include <linux/ref_tracker.h>
 #include "ctree.h"
 
 struct btrfs_disk_key;
@@ -44,6 +45,22 @@ struct btrfs_delayed_root {
        wait_queue_head_t wait;
 };
 
+struct btrfs_ref_tracker_dir {
+#ifdef CONFIG_BTRFS_DEBUG
+       struct ref_tracker_dir dir;
+#else
+       struct {} tracker;
+#endif
+};
+
+struct btrfs_ref_tracker {
+#ifdef CONFIG_BTRFS_DEBUG
+       struct ref_tracker *tracker;
+#else
+       struct {} tracker;
+#endif
+};
+
 #define BTRFS_DELAYED_NODE_IN_LIST     0
 #define BTRFS_DELAYED_NODE_INODE_DIRTY 1
 #define BTRFS_DELAYED_NODE_DEL_IREF    2
@@ -78,6 +95,12 @@ struct btrfs_delayed_node {
         * actual number of leaves we end up using. Protected by @mutex.
         */
        u32 index_item_leaves;
+       /* Track all references to this delayed node. */
+       struct btrfs_ref_tracker_dir ref_dir;
+       /* Track delayed node reference stored in node list. */
+       struct btrfs_ref_tracker node_list_tracker;
+       /* Track delayed node reference stored in inode cache. */
+       struct btrfs_ref_tracker inode_cache_tracker;
 };
 
 struct btrfs_delayed_item {
@@ -169,4 +192,51 @@ void __cold btrfs_delayed_inode_exit(void);
 /* for debugging */
 void btrfs_assert_delayed_root_empty(struct btrfs_fs_info *fs_info);
 
+#define BTRFS_DELAYED_NODE_REF_TRACKER_QUARANTINE_COUNT                16
+#define BTRFS_DELAYED_NODE_REF_TRACKER_DISPLAY_LIMIT           16
+
+#ifdef CONFIG_BTRFS_DEBUG
+static inline void btrfs_delayed_node_ref_tracker_dir_init(struct btrfs_delayed_node *node)
+{
+       ref_tracker_dir_init(&node->ref_dir.dir,
+                            BTRFS_DELAYED_NODE_REF_TRACKER_QUARANTINE_COUNT,
+                            "delayed_node");
+}
+
+static inline void btrfs_delayed_node_ref_tracker_dir_exit(struct btrfs_delayed_node *node)
+{
+       ref_tracker_dir_exit(&node->ref_dir.dir);
+}
+
+static inline int btrfs_delayed_node_ref_tracker_alloc(struct btrfs_delayed_node *node,
+                                                      struct btrfs_ref_tracker *tracker,
+                                                      gfp_t gfp)
+{
+       return ref_tracker_alloc(&node->ref_dir.dir, &tracker->tracker, gfp);
+}
+
+static inline int btrfs_delayed_node_ref_tracker_free(struct btrfs_delayed_node *node,
+                                                     struct btrfs_ref_tracker *tracker)
+{
+       return ref_tracker_free(&node->ref_dir.dir, &tracker->tracker);
+}
+#else
+static inline void btrfs_delayed_node_ref_tracker_dir_init(struct btrfs_delayed_node *node) { }
+
+static inline void btrfs_delayed_node_ref_tracker_dir_exit(struct btrfs_delayed_node *node) { }
+
+static inline int btrfs_delayed_node_ref_tracker_alloc(struct btrfs_delayed_node *node,
+                                                      struct btrfs_ref_tracker *tracker,
+                                                      gfp_t gfp)
+{
+       return 0;
+}
+
+static inline int btrfs_delayed_node_ref_tracker_free(struct btrfs_delayed_node *node,
+                                                     struct btrfs_ref_tracker *tracker)
+{
+       return 0;
+}
+#endif
+
 #endif