return BLK_MQ_RQ_QUEUE_OK;
 }
 
+static inline void nvme_setup_write_zeroes(struct nvme_ns *ns,
+               struct request *req, struct nvme_command *cmnd)
+{
+       struct nvme_write_zeroes_cmd *write_zeroes = &cmnd->write_zeroes;
+
+       memset(cmnd, 0, sizeof(*cmnd));
+       write_zeroes->opcode = nvme_cmd_write_zeroes;
+       write_zeroes->nsid = cpu_to_le32(ns->ns_id);
+       write_zeroes->slba =
+               cpu_to_le64(nvme_block_nr(ns, blk_rq_pos(req)));
+       write_zeroes->length =
+               cpu_to_le16((blk_rq_bytes(req) >> ns->lba_shift) - 1);
+       write_zeroes->control = 0;
+}
+
 static inline void nvme_setup_rw(struct nvme_ns *ns, struct request *req,
                struct nvme_command *cmnd)
 {
                nvme_setup_flush(ns, cmd);
        else if (req_op(req) == REQ_OP_DISCARD)
                ret = nvme_setup_discard(ns, req, cmd);
+       else if (req_op(req) == REQ_OP_WRITE_ZEROES)
+               nvme_setup_write_zeroes(ns, req, cmd);
        else
                nvme_setup_rw(ns, req, cmd);
 
 
        if (ns->ctrl->oncs & NVME_CTRL_ONCS_DSM)
                nvme_config_discard(ns);
+       if (ns->ctrl->oncs & NVME_CTRL_ONCS_WRITE_ZEROES)
+               blk_queue_max_write_zeroes_sectors(ns->queue,
+                               ((u32)(USHRT_MAX + 1) * bs) >> 9);
+
        blk_mq_unfreeze_queue(disk->queue);
 }