sector_t offset)
{
struct request_queue *q = rq->q;
+ unsigned int max_sectors;
if (blk_rq_is_passthrough(rq))
return q->limits.max_hw_sectors;
- if (!q->limits.chunk_sectors ||
- req_op(rq) == REQ_OP_DISCARD ||
- req_op(rq) == REQ_OP_SECURE_ERASE)
- return blk_queue_get_max_sectors(q, req_op(rq));
-
- return min(blk_max_size_offset(q, offset, 0),
- blk_queue_get_max_sectors(q, req_op(rq)));
+ max_sectors = blk_queue_get_max_sectors(q, req_op(rq));
+ if (q->limits.chunk_sectors &&
+ req_op(rq) != REQ_OP_DISCARD &&
+ req_op(rq) != REQ_OP_SECURE_ERASE)
+ max_sectors = min(max_sectors,
+ chunk_size_left(offset, q->limits.chunk_sectors));
+ return max_sectors;
}
static inline int ll_new_hw_segment(struct request *req, struct bio *bio,