goto out;
                }
 
-               __btrfs_tree_lock(left, BTRFS_NESTING_LEFT);
+               btrfs_tree_lock_nested(left, BTRFS_NESTING_LEFT);
                wret = btrfs_cow_block(trans, root, left,
                                       parent, pslot - 1, &left,
                                       BTRFS_NESTING_LEFT_COW);
                        goto out;
                }
 
-               __btrfs_tree_lock(right, BTRFS_NESTING_RIGHT);
+               btrfs_tree_lock_nested(right, BTRFS_NESTING_RIGHT);
                wret = btrfs_cow_block(trans, root, right,
                                       parent, pslot + 1, &right,
                                       BTRFS_NESTING_RIGHT_COW);
                if (IS_ERR(left))
                        return PTR_ERR(left);
 
-               __btrfs_tree_lock(left, BTRFS_NESTING_LEFT);
+               btrfs_tree_lock_nested(left, BTRFS_NESTING_LEFT);
 
                left_nr = btrfs_header_nritems(left);
                if (left_nr >= BTRFS_NODEPTRS_PER_BLOCK(fs_info) - 1) {
                if (IS_ERR(right))
                        return PTR_ERR(right);
 
-               __btrfs_tree_lock(right, BTRFS_NESTING_RIGHT);
+               btrfs_tree_lock_nested(right, BTRFS_NESTING_RIGHT);
 
                right_nr = btrfs_header_nritems(right);
                if (right_nr >= BTRFS_NODEPTRS_PER_BLOCK(fs_info) - 1) {
        if (IS_ERR(right))
                return PTR_ERR(right);
 
-       __btrfs_tree_lock(right, BTRFS_NESTING_RIGHT);
+       btrfs_tree_lock_nested(right, BTRFS_NESTING_RIGHT);
 
        free_space = btrfs_leaf_free_space(right);
        if (free_space < data_size)
        if (IS_ERR(left))
                return PTR_ERR(left);
 
-       __btrfs_tree_lock(left, BTRFS_NESTING_LEFT);
+       btrfs_tree_lock_nested(left, BTRFS_NESTING_LEFT);
 
        free_space = btrfs_leaf_free_space(left);
        if (free_space < data_size) {
 
         */
        btrfs_set_buffer_lockdep_class(lockdep_owner, buf, level);
 
-       __btrfs_tree_lock(buf, nest);
+       btrfs_tree_lock_nested(buf, nest);
        btrfs_clear_buffer_dirty(trans, buf);
        clear_bit(EXTENT_BUFFER_STALE, &buf->bflags);
        clear_bit(EXTENT_BUFFER_ZONED_ZEROOUT, &buf->bflags);
 
  */
 
 /*
- * __btrfs_tree_read_lock - lock extent buffer for read
+ * btrfs_tree_read_lock_nested - lock extent buffer for read
  * @eb:                the eb to be locked
  * @nest:      the nesting level to be used for lockdep
  *
  * This takes the read lock on the extent buffer, using the specified nesting
  * level for lockdep purposes.
  */
-void __btrfs_tree_read_lock(struct extent_buffer *eb, enum btrfs_lock_nesting nest)
+void btrfs_tree_read_lock_nested(struct extent_buffer *eb, enum btrfs_lock_nesting nest)
 {
        u64 start_ns = 0;
 
  *
  * Returns with the eb->lock write locked.
  */
-void __btrfs_tree_lock(struct extent_buffer *eb, enum btrfs_lock_nesting nest)
+void btrfs_tree_lock_nested(struct extent_buffer *eb, enum btrfs_lock_nesting nest)
        __acquires(&eb->lock)
 {
        u64 start_ns = 0;
 
 static_assert(BTRFS_NESTING_MAX <= MAX_LOCKDEP_SUBCLASSES,
              "too many lock subclasses defined");
 
-void __btrfs_tree_lock(struct extent_buffer *eb, enum btrfs_lock_nesting nest);
+void btrfs_tree_lock_nested(struct extent_buffer *eb, enum btrfs_lock_nesting nest);
 
 static inline void btrfs_tree_lock(struct extent_buffer *eb)
 {
-       __btrfs_tree_lock(eb, BTRFS_NESTING_NORMAL);
+       btrfs_tree_lock_nested(eb, BTRFS_NESTING_NORMAL);
 }
 
 void btrfs_tree_unlock(struct extent_buffer *eb);
 
-void __btrfs_tree_read_lock(struct extent_buffer *eb, enum btrfs_lock_nesting nest);
+void btrfs_tree_read_lock_nested(struct extent_buffer *eb, enum btrfs_lock_nesting nest);
 
 static inline void btrfs_tree_read_lock(struct extent_buffer *eb)
 {
-       __btrfs_tree_read_lock(eb, BTRFS_NESTING_NORMAL);
+       btrfs_tree_read_lock_nested(eb, BTRFS_NESTING_NORMAL);
 }
 
 void btrfs_tree_read_unlock(struct extent_buffer *eb);