Some drivers use it now, others just set the limits field manually.
But in preparation for splitting this into a hard and soft limit,
ensure that they all call the proper function for setting the hw
limit for discards.
Reviewed-by: Jeff Moyer <jmoyer@redhat.com>
Signed-off-by: Jens Axboe <axboe@fb.com>
        blk_queue_physical_block_size(brd->brd_queue, PAGE_SIZE);
 
        brd->brd_queue->limits.discard_granularity = PAGE_SIZE;
-       brd->brd_queue->limits.max_discard_sectors = UINT_MAX;
+       blk_queue_max_discard_sectors(brd->brd_queue, UINT_MAX);
        brd->brd_queue->limits.discard_zeroes_data = 1;
        queue_flag_set_unlocked(QUEUE_FLAG_DISCARD, brd->brd_queue);
 
 
                        /* For now, don't allow more than one activity log extent worth of data
                         * to be discarded in one go. We may need to rework drbd_al_begin_io()
                         * to allow for even larger discard ranges */
-                       q->limits.max_discard_sectors = DRBD_MAX_DISCARD_SECTORS;
+                       blk_queue_max_discard_sectors(q, DRBD_MAX_DISCARD_SECTORS);
 
                        queue_flag_set_unlocked(QUEUE_FLAG_DISCARD, q);
                        /* REALLY? Is stacking secdiscard "legal"? */
                        if (blk_queue_secdiscard(b))
                                queue_flag_set_unlocked(QUEUE_FLAG_SECDISCARD, q);
                } else {
-                       q->limits.max_discard_sectors = 0;
+                       blk_queue_max_discard_sectors(q, 0);
                        queue_flag_clear_unlocked(QUEUE_FLAG_DISCARD, q);
                        queue_flag_clear_unlocked(QUEUE_FLAG_SECDISCARD, q);
                }
 
            lo->lo_encrypt_key_size) {
                q->limits.discard_granularity = 0;
                q->limits.discard_alignment = 0;
-               q->limits.max_discard_sectors = 0;
+               blk_queue_max_discard_sectors(q, 0);
                q->limits.discard_zeroes_data = 0;
                queue_flag_clear_unlocked(QUEUE_FLAG_DISCARD, q);
                return;
 
        q->limits.discard_granularity = inode->i_sb->s_blocksize;
        q->limits.discard_alignment = 0;
-       q->limits.max_discard_sectors = UINT_MAX >> 9;
+       blk_queue_max_discard_sectors(q, UINT_MAX >> 9);
        q->limits.discard_zeroes_data = 1;
        queue_flag_set_unlocked(QUEUE_FLAG_DISCARD, q);
 }
 
                queue_flag_set_unlocked(QUEUE_FLAG_NONROT, disk->queue);
                queue_flag_clear_unlocked(QUEUE_FLAG_ADD_RANDOM, disk->queue);
                disk->queue->limits.discard_granularity = 512;
-               disk->queue->limits.max_discard_sectors = UINT_MAX;
+               blk_queue_max_discard_sectors(disk->queue, UINT_MAX);
                disk->queue->limits.discard_zeroes_data = 0;
                blk_queue_max_hw_sectors(disk->queue, 65536);
                disk->queue->limits.max_sectors = 256;
 
        ns->queue->limits.discard_zeroes_data = 0;
        ns->queue->limits.discard_alignment = logical_block_size;
        ns->queue->limits.discard_granularity = logical_block_size;
-       ns->queue->limits.max_discard_sectors = 0xffffffff;
+       blk_queue_max_discard_sectors(ns->queue, 0xffffffff);
        queue_flag_set_unlocked(QUEUE_FLAG_DISCARD, ns->queue);
 }
 
 
        queue_flag_set_unlocked(QUEUE_FLAG_DISCARD, q);
        q->limits.discard_granularity = segment_size;
        q->limits.discard_alignment = segment_size;
-       q->limits.max_discard_sectors = segment_size / SECTOR_SIZE;
+       blk_queue_max_discard_sectors(q, segment_size / SECTOR_SIZE);
        q->limits.discard_zeroes_data = 1;
 
        blk_queue_merge_bvec(q, rbd_merge_bvec);
 
        /* DISCARD Flag initialization. */
        q->limits.discard_granularity = 8192;
        q->limits.discard_alignment = 0;
-       q->limits.max_discard_sectors = UINT_MAX >> 9;
+       blk_queue_max_discard_sectors(q, UINT_MAX >> 9);
        q->limits.discard_zeroes_data = 1;
        queue_flag_set_unlocked(QUEUE_FLAG_DISCARD, q);
        queue_flag_set_unlocked(QUEUE_FLAG_NONROT, q);
 
        blk_queue_io_min(zram->disk->queue, PAGE_SIZE);
        blk_queue_io_opt(zram->disk->queue, PAGE_SIZE);
        zram->disk->queue->limits.discard_granularity = PAGE_SIZE;
-       zram->disk->queue->limits.max_discard_sectors = UINT_MAX;
+       blk_queue_max_discard_sectors(zram->disk->queue, UINT_MAX);
        /*
         * zram_bio_discard() will clear all logical blocks if logical block
         * size is identical with physical block size(PAGE_SIZE). But if it is
 
        q->limits.max_sectors           = UINT_MAX;
        q->limits.max_segment_size      = UINT_MAX;
        q->limits.max_segments          = BIO_MAX_PAGES;
-       q->limits.max_discard_sectors   = UINT_MAX;
+       blk_queue_max_discard_sectors(q, UINT_MAX);
        q->limits.discard_granularity   = 512;
        q->limits.io_min                = block_size;
        q->limits.logical_block_size    = block_size;
 
                return;
 
        queue_flag_set_unlocked(QUEUE_FLAG_DISCARD, q);
-       q->limits.max_discard_sectors = max_discard;
+       blk_queue_max_discard_sectors(q, max_discard);
        if (card->erased_byte == 0 && !mmc_can_discard(card))
                q->limits.discard_zeroes_data = 1;
        q->limits.discard_granularity = card->pref_erase << 9;
 
 
        if (tr->discard) {
                queue_flag_set_unlocked(QUEUE_FLAG_DISCARD, new->rq);
-               new->rq->limits.max_discard_sectors = UINT_MAX;
+               blk_queue_max_discard_sectors(new->rq, UINT_MAX);
        }
 
        gd->queue = new->rq;
 
        switch (mode) {
 
        case SD_LBP_DISABLE:
-               q->limits.max_discard_sectors = 0;
+               blk_queue_max_discard_sectors(q, 0);
                queue_flag_clear_unlocked(QUEUE_FLAG_DISCARD, q);
                return;
 
                break;
        }
 
-       q->limits.max_discard_sectors = max_blocks * (logical_block_size >> 9);
+       blk_queue_max_discard_sectors(q, max_blocks * (logical_block_size >> 9));
        queue_flag_set_unlocked(QUEUE_FLAG_DISCARD, q);
 }