QUEUE_FLAG_NO_SG_MERGE has been killed, so kill BLK_MQ_F_SG_MERGE too.
Reviewed-by: Christoph Hellwig <hch@lst.de>
Reviewed-by: Omar Sandoval <osandov@fb.com>
Signed-off-by: Ming Lei <ming.lei@redhat.com>
Signed-off-by: Jens Axboe <axboe@kernel.dk>
 static const char *const hctx_flag_name[] = {
        HCTX_FLAG_NAME(SHOULD_MERGE),
        HCTX_FLAG_NAME(TAG_SHARED),
-       HCTX_FLAG_NAME(SG_MERGE),
        HCTX_FLAG_NAME(BLOCKING),
        HCTX_FLAG_NAME(NO_SCHED),
 };
 
        lo->tag_set.queue_depth = 128;
        lo->tag_set.numa_node = NUMA_NO_NODE;
        lo->tag_set.cmd_size = sizeof(struct loop_cmd);
-       lo->tag_set.flags = BLK_MQ_F_SHOULD_MERGE | BLK_MQ_F_SG_MERGE;
+       lo->tag_set.flags = BLK_MQ_F_SHOULD_MERGE;
        lo->tag_set.driver_data = lo;
 
        err = blk_mq_alloc_tag_set(&lo->tag_set);
 
        nbd->tag_set.numa_node = NUMA_NO_NODE;
        nbd->tag_set.cmd_size = sizeof(struct nbd_cmd);
        nbd->tag_set.flags = BLK_MQ_F_SHOULD_MERGE |
-               BLK_MQ_F_SG_MERGE | BLK_MQ_F_BLOCKING;
+               BLK_MQ_F_BLOCKING;
        nbd->tag_set.driver_data = nbd;
 
        err = blk_mq_alloc_tag_set(&nbd->tag_set);
 
        rbd_dev->tag_set.ops = &rbd_mq_ops;
        rbd_dev->tag_set.queue_depth = rbd_dev->opts->queue_depth;
        rbd_dev->tag_set.numa_node = NUMA_NO_NODE;
-       rbd_dev->tag_set.flags = BLK_MQ_F_SHOULD_MERGE | BLK_MQ_F_SG_MERGE;
+       rbd_dev->tag_set.flags = BLK_MQ_F_SHOULD_MERGE;
        rbd_dev->tag_set.nr_hw_queues = 1;
        rbd_dev->tag_set.cmd_size = sizeof(struct work_struct);
 
 
                skdev->sgs_per_request * sizeof(struct scatterlist);
        skdev->tag_set.numa_node = NUMA_NO_NODE;
        skdev->tag_set.flags = BLK_MQ_F_SHOULD_MERGE |
-               BLK_MQ_F_SG_MERGE |
                BLK_ALLOC_POLICY_TO_MQ_FLAG(BLK_TAG_ALLOC_FIFO);
        skdev->tag_set.driver_data = skdev;
        rc = blk_mq_alloc_tag_set(&skdev->tag_set);
 
        } else
                info->tag_set.queue_depth = BLK_RING_SIZE(info);
        info->tag_set.numa_node = NUMA_NO_NODE;
-       info->tag_set.flags = BLK_MQ_F_SHOULD_MERGE | BLK_MQ_F_SG_MERGE;
+       info->tag_set.flags = BLK_MQ_F_SHOULD_MERGE;
        info->tag_set.cmd_size = sizeof(struct blkif_req);
        info->tag_set.driver_data = info;
 
 
        md->tag_set->ops = &dm_mq_ops;
        md->tag_set->queue_depth = dm_get_blk_mq_queue_depth();
        md->tag_set->numa_node = md->numa_node_id;
-       md->tag_set->flags = BLK_MQ_F_SHOULD_MERGE | BLK_MQ_F_SG_MERGE;
+       md->tag_set->flags = BLK_MQ_F_SHOULD_MERGE;
        md->tag_set->nr_hw_queues = dm_get_blk_mq_nr_hw_queues();
        md->tag_set->driver_data = md;
 
 
        else
                mq->tag_set.queue_depth = MMC_QUEUE_DEPTH;
        mq->tag_set.numa_node = NUMA_NO_NODE;
-       mq->tag_set.flags = BLK_MQ_F_SHOULD_MERGE | BLK_MQ_F_SG_MERGE |
-                           BLK_MQ_F_BLOCKING;
+       mq->tag_set.flags = BLK_MQ_F_SHOULD_MERGE | BLK_MQ_F_BLOCKING;
        mq->tag_set.nr_hw_queues = 1;
        mq->tag_set.cmd_size = sizeof(struct mmc_queue_req);
        mq->tag_set.driver_data = mq;
 
        shost->tag_set.queue_depth = shost->can_queue;
        shost->tag_set.cmd_size = cmd_size;
        shost->tag_set.numa_node = NUMA_NO_NODE;
-       shost->tag_set.flags = BLK_MQ_F_SHOULD_MERGE | BLK_MQ_F_SG_MERGE;
+       shost->tag_set.flags = BLK_MQ_F_SHOULD_MERGE;
        shost->tag_set.flags |=
                BLK_ALLOC_POLICY_TO_MQ_FLAG(shost->hostt->tag_alloc_policy);
        shost->tag_set.driver_data = shost;
 
 enum {
        BLK_MQ_F_SHOULD_MERGE   = 1 << 0,
        BLK_MQ_F_TAG_SHARED     = 1 << 1,
-       BLK_MQ_F_SG_MERGE       = 1 << 2,
        BLK_MQ_F_BLOCKING       = 1 << 5,
        BLK_MQ_F_NO_SCHED       = 1 << 6,
        BLK_MQ_F_ALLOC_POLICY_START_BIT = 8,