queue_for_each_hw_ctx(q, hctx, i) {
                if (hctx->flags & BLK_MQ_F_BLOCKING)
-                       synchronize_srcu(hctx->queue_rq_srcu);
+                       synchronize_srcu(hctx->srcu);
                else
                        rcu = true;
        }
        if (!(hctx->flags & BLK_MQ_F_BLOCKING))
                rcu_read_unlock();
        else
-               srcu_read_unlock(hctx->queue_rq_srcu, srcu_idx);
+               srcu_read_unlock(hctx->srcu, srcu_idx);
 }
 
 static void hctx_lock(struct blk_mq_hw_ctx *hctx, int *srcu_idx)
        if (!(hctx->flags & BLK_MQ_F_BLOCKING))
                rcu_read_lock();
        else
-               *srcu_idx = srcu_read_lock(hctx->queue_rq_srcu);
+               *srcu_idx = srcu_read_lock(hctx->srcu);
 }
 
 static void blk_mq_rq_update_aborted_gstate(struct request *rq, u64 gstate)
                        if (!(hctx->flags & BLK_MQ_F_BLOCKING))
                                has_rcu = true;
                        else
-                               synchronize_srcu(hctx->queue_rq_srcu);
+                               synchronize_srcu(hctx->srcu);
 
                        hctx->nr_expired = 0;
                }
                set->ops->exit_hctx(hctx, hctx_idx);
 
        if (hctx->flags & BLK_MQ_F_BLOCKING)
-               cleanup_srcu_struct(hctx->queue_rq_srcu);
+               cleanup_srcu_struct(hctx->srcu);
 
        blk_mq_remove_cpuhp(hctx);
        blk_free_flush_queue(hctx->fq);
                goto free_fq;
 
        if (hctx->flags & BLK_MQ_F_BLOCKING)
-               init_srcu_struct(hctx->queue_rq_srcu);
+               init_srcu_struct(hctx->srcu);
 
        blk_mq_debugfs_register_hctx(q, hctx);
 
 {
        int hw_ctx_size = sizeof(struct blk_mq_hw_ctx);
 
-       BUILD_BUG_ON(ALIGN(offsetof(struct blk_mq_hw_ctx, queue_rq_srcu),
+       BUILD_BUG_ON(ALIGN(offsetof(struct blk_mq_hw_ctx, srcu),
                           __alignof__(struct blk_mq_hw_ctx)) !=
                     sizeof(struct blk_mq_hw_ctx));