struct blk_stat_callback *cb)
 {
        unsigned int bucket;
+       unsigned long flags;
        int cpu;
 
        for_each_possible_cpu(cpu) {
                        blk_rq_stat_init(&cpu_stat[bucket]);
        }
 
-       spin_lock(&q->stats->lock);
+       spin_lock_irqsave(&q->stats->lock, flags);
        list_add_tail_rcu(&cb->list, &q->stats->callbacks);
        blk_queue_flag_set(QUEUE_FLAG_STATS, q);
-       spin_unlock(&q->stats->lock);
+       spin_unlock_irqrestore(&q->stats->lock, flags);
 }
 
 void blk_stat_remove_callback(struct request_queue *q,
                              struct blk_stat_callback *cb)
 {
-       spin_lock(&q->stats->lock);
+       unsigned long flags;
+
+       spin_lock_irqsave(&q->stats->lock, flags);
        list_del_rcu(&cb->list);
        if (list_empty(&q->stats->callbacks) && !q->stats->enable_accounting)
                blk_queue_flag_clear(QUEUE_FLAG_STATS, q);
-       spin_unlock(&q->stats->lock);
+       spin_unlock_irqrestore(&q->stats->lock, flags);
 
        del_timer_sync(&cb->timer);
 }
 
 void blk_stat_enable_accounting(struct request_queue *q)
 {
-       spin_lock(&q->stats->lock);
+       unsigned long flags;
+
+       spin_lock_irqsave(&q->stats->lock, flags);
        q->stats->enable_accounting = true;
        blk_queue_flag_set(QUEUE_FLAG_STATS, q);
-       spin_unlock(&q->stats->lock);
+       spin_unlock_irqrestore(&q->stats->lock, flags);
 }
 EXPORT_SYMBOL_GPL(blk_stat_enable_accounting);