static int nvmet_passthru_map_sg(struct nvmet_req *req, struct request *rq)
 {
        struct scatterlist *sg;
-       int op_flags = 0;
        struct bio *bio;
        int i, ret;
 
        if (req->sg_cnt > BIO_MAX_PAGES)
                return -EINVAL;
 
-       if (req->cmd->common.opcode == nvme_cmd_flush)
-               op_flags = REQ_FUA;
-       else if (nvme_is_write(req->cmd))
-               op_flags = REQ_SYNC | REQ_IDLE;
-
        bio = bio_alloc(GFP_KERNEL, req->sg_cnt);
        bio->bi_end_io = bio_put;
-       bio->bi_opf = req_op(rq) | op_flags;
+       bio->bi_opf = req_op(rq);
 
        for_each_sg(req->sg, sg, req->sg_cnt, i) {
                if (bio_add_pc_page(rq->q, bio, sg_page(sg), sg->length,