* registered by blk_register_queue().
         */
        q->bypass_depth = 1;
-       __set_bit(QUEUE_FLAG_BYPASS, &q->queue_flags);
+       queue_flag_set_unlocked(QUEUE_FLAG_BYPASS, q);
 
        init_waitqueue_head(&q->mq_freeze_wq);
 
 
        q->queue_flags |= QUEUE_FLAG_MQ_DEFAULT;
 
        if (!(set->flags & BLK_MQ_F_SG_MERGE))
-               q->queue_flags |= 1 << QUEUE_FLAG_NO_SG_MERGE;
+               queue_flag_set_unlocked(QUEUE_FLAG_NO_SG_MERGE, q);
 
        q->sg_reserved_size = INT_MAX;
 
 
 {
        spin_lock_irq(q->queue_lock);
        if (queueable)
-               clear_bit(QUEUE_FLAG_FLUSH_NQ, &q->queue_flags);
+               queue_flag_clear(QUEUE_FLAG_FLUSH_NQ, q);
        else
-               set_bit(QUEUE_FLAG_FLUSH_NQ, &q->queue_flags);
+               queue_flag_set(QUEUE_FLAG_FLUSH_NQ, q);
        spin_unlock_irq(q->queue_lock);
 }
 EXPORT_SYMBOL_GPL(blk_queue_flush_queueable);
 
 
        spin_lock(&q->stats->lock);
        list_add_tail_rcu(&cb->list, &q->stats->callbacks);
-       set_bit(QUEUE_FLAG_STATS, &q->queue_flags);
+       queue_flag_set(QUEUE_FLAG_STATS, q);
        spin_unlock(&q->stats->lock);
 }
 EXPORT_SYMBOL_GPL(blk_stat_add_callback);
        spin_lock(&q->stats->lock);
        list_del_rcu(&cb->list);
        if (list_empty(&q->stats->callbacks) && !q->stats->enable_accounting)
-               clear_bit(QUEUE_FLAG_STATS, &q->queue_flags);
+               queue_flag_clear(QUEUE_FLAG_STATS, q);
        spin_unlock(&q->stats->lock);
 
        del_timer_sync(&cb->timer);
 {
        spin_lock(&q->stats->lock);
        q->stats->enable_accounting = true;
-       set_bit(QUEUE_FLAG_STATS, &q->queue_flags);
+       queue_flag_set(QUEUE_FLAG_STATS, q);
        spin_unlock(&q->stats->lock);
 }