bch2_write_ref_put(c, BCH_WRITE_REF_gc_gens);
 }
 
-void bch2_fs_btree_gc_exit(struct bch_fs *c)
-{
-}
-
-int bch2_fs_btree_gc_init(struct bch_fs *c)
+void bch2_fs_btree_gc_init_early(struct bch_fs *c)
 {
        seqcount_init(&c->gc_pos_lock);
        INIT_WORK(&c->gc_gens_work, bch2_gc_gens_work);
 
        init_rwsem(&c->gc_lock);
        mutex_init(&c->gc_gens_lock);
-       return 0;
 }
 
 int bch2_gc_gens(struct bch_fs *);
 void bch2_gc_gens_async(struct bch_fs *);
 
-void bch2_fs_btree_gc_exit(struct bch_fs *);
-int bch2_fs_btree_gc_init(struct bch_fs *);
+void bch2_fs_btree_gc_init_early(struct bch_fs *);
 
 #endif /* _BCACHEFS_BTREE_GC_H */
 
        darray_exit(&wb->inc.keys);
 }
 
-int bch2_fs_btree_write_buffer_init(struct bch_fs *c)
+void bch2_fs_btree_write_buffer_init_early(struct bch_fs *c)
 {
        struct btree_write_buffer *wb = &c->btree_write_buffer;
 
        mutex_init(&wb->inc.lock);
        mutex_init(&wb->flushing.lock);
        INIT_WORK(&wb->flush_work, bch2_btree_write_buffer_flush_work);
+}
+
+int bch2_fs_btree_write_buffer_init(struct bch_fs *c)
+{
+       struct btree_write_buffer *wb = &c->btree_write_buffer;
 
        /* Will be resized by journal as needed: */
        unsigned initial_size = 1 << 16;
 
 
 int bch2_btree_write_buffer_resize(struct bch_fs *, size_t);
 void bch2_fs_btree_write_buffer_exit(struct bch_fs *);
+void bch2_fs_btree_write_buffer_init_early(struct bch_fs *);
 int bch2_fs_btree_write_buffer_init(struct bch_fs *);
 
 #endif /* _BCACHEFS_BTREE_WRITE_BUFFER_H */
 
        free_fifo(&j->pin);
 }
 
-int bch2_fs_journal_init(struct journal *j)
+void bch2_fs_journal_init_early(struct journal *j)
 {
        static struct lock_class_key res_key;
 
        atomic64_set(&j->reservations.counter,
                ((union journal_res_state)
                 { .cur_entry_offset = JOURNAL_ENTRY_CLOSED_VAL }).v);
+}
 
+int bch2_fs_journal_init(struct journal *j)
+{
        j->free_buf_size = j->buf_size_want = JOURNAL_ENTRY_SIZE_MIN;
        j->free_buf = kvmalloc(j->free_buf_size, GFP_KERNEL);
        if (!j->free_buf)
 
 void bch2_dev_journal_exit(struct bch_dev *);
 int bch2_dev_journal_init(struct bch_dev *, struct bch_sb *);
 void bch2_fs_journal_exit(struct journal *);
+void bch2_fs_journal_init_early(struct journal *);
 int bch2_fs_journal_init(struct journal *);
 
 #endif /* _BCACHEFS_JOURNAL_H */
 
                        BUG_ON(atomic_read(&l->l[j]));
 }
 
-int bch2_fs_nocow_locking_init(struct bch_fs *c)
+void bch2_fs_nocow_locking_init_early(struct bch_fs *c)
 {
        struct bucket_nocow_lock_table *t = &c->nocow_locks;
 
        for (struct nocow_lock_bucket *l = t->l; l < t->l + ARRAY_SIZE(t->l); l++)
                spin_lock_init(&l->lock);
-
-       return 0;
 }
 
 void bch2_nocow_locks_to_text(struct printbuf *, struct bucket_nocow_lock_table *);
 
 void bch2_fs_nocow_locking_exit(struct bch_fs *);
-int bch2_fs_nocow_locking_init(struct bch_fs *);
+void bch2_fs_nocow_locking_init_early(struct bch_fs *);
 
 #endif /* _BCACHEFS_NOCOW_LOCKING_H */
 
        return ret;
 }
 
-int bch2_fs_subvolumes_init(struct bch_fs *c)
+void bch2_fs_subvolumes_init_early(struct bch_fs *c)
 {
        INIT_WORK(&c->snapshot_delete_work, bch2_delete_dead_snapshots_work);
        INIT_WORK(&c->snapshot_wait_for_pagecache_and_delete_work,
                  bch2_subvolume_wait_for_pagecache_and_delete);
        mutex_init(&c->snapshots_unlinked_lock);
-       return 0;
 }
 
 int bch2_initialize_subvolumes(struct bch_fs *);
 int bch2_fs_upgrade_for_subvolumes(struct bch_fs *);
 
-int bch2_fs_subvolumes_init(struct bch_fs *);
+void bch2_fs_subvolumes_init_early(struct bch_fs *);
 
 #endif /* _BCACHEFS_SUBVOLUME_H */
 
        bch2_io_clock_exit(&c->io_clock[WRITE]);
        bch2_io_clock_exit(&c->io_clock[READ]);
        bch2_fs_compress_exit(c);
-       bch2_fs_btree_gc_exit(c);
        bch2_journal_keys_put_initial(c);
        bch2_find_btree_nodes_exit(&c->found_btree_nodes);
        BUG_ON(atomic_read(&c->journal_keys.ref));
        bch2_fs_allocator_background_init(c);
        bch2_fs_allocator_foreground_init(c);
        bch2_fs_btree_cache_init_early(&c->btree_cache);
+       bch2_fs_btree_gc_init_early(c);
        bch2_fs_btree_interior_update_init_early(c);
        bch2_fs_btree_iter_init_early(c);
        bch2_fs_btree_key_cache_init_early(&c->btree_key_cache);
+       bch2_fs_btree_write_buffer_init_early(c);
        bch2_fs_copygc_init(c);
        bch2_fs_ec_init_early(c);
+       bch2_fs_journal_init_early(&c->journal);
        bch2_fs_journal_keys_init(c);
        bch2_fs_move_init(c);
+       bch2_fs_nocow_locking_init_early(c);
        bch2_fs_quota_init(c);
        bch2_fs_rebalance_init(c);
        bch2_fs_sb_errors_init_early(c);
+       bch2_fs_subvolumes_init_early(c);
 
        INIT_LIST_HEAD(&c->list);
 
 
        ret =
            bch2_fs_btree_cache_init(c) ?:
-           bch2_fs_btree_gc_init(c) ?:
            bch2_fs_btree_iter_init(c) ?:
            bch2_fs_btree_interior_update_init(c) ?:
            bch2_fs_btree_key_cache_init(&c->btree_key_cache) ?:
            bch2_fs_io_read_init(c) ?:
            bch2_fs_io_write_init(c) ?:
            bch2_fs_journal_init(&c->journal) ?:
-           bch2_fs_nocow_locking_init(c) ?:
            bch2_fs_sb_errors_init(c) ?:
-           bch2_fs_subvolumes_init(c) ?:
            bch2_fs_vfs_init(c);
        if (ret)
                goto err;