if (req->q->queuedata)
                req->timeout = NVME_IO_TIMEOUT;
        else /* no queuedata implies admin queue */
-               req->timeout = ADMIN_TIMEOUT;
+               req->timeout = NVME_ADMIN_TIMEOUT;
 
        req->cmd_flags |= REQ_FAILFAST_DRIVER;
        nvme_clear_nvme_request(req);
        cmd.common.cdw10 = cpu_to_le32(((u32)secp) << 24 | ((u32)spsp) << 8);
        cmd.common.cdw11 = cpu_to_le32(len);
 
-       return __nvme_submit_sync_cmd(ctrl->admin_q, &cmd, NULL, buffer, len,
-                                     ADMIN_TIMEOUT, NVME_QID_ANY, 1, 0, false);
+       return __nvme_submit_sync_cmd(ctrl->admin_q, &cmd, NULL, buffer, len, 0,
+                       NVME_QID_ANY, 1, 0, false);
 }
 EXPORT_SYMBOL_GPL(nvme_sec_submit);
 #endif /* CONFIG_BLK_SED_OPAL */
 
                            ctrl->lport->ops->fcprqst_priv_sz);
        ctrl->admin_tag_set.driver_data = ctrl;
        ctrl->admin_tag_set.nr_hw_queues = 1;
-       ctrl->admin_tag_set.timeout = ADMIN_TIMEOUT;
+       ctrl->admin_tag_set.timeout = NVME_ADMIN_TIMEOUT;
        ctrl->admin_tag_set.flags = BLK_MQ_F_NO_SCHED;
 
        ret = blk_mq_alloc_tag_set(&ctrl->admin_tag_set);
 
 #define NVME_IO_TIMEOUT        (nvme_io_timeout * HZ)
 
 extern unsigned int admin_timeout;
-#define ADMIN_TIMEOUT  (admin_timeout * HZ)
+#define NVME_ADMIN_TIMEOUT     (admin_timeout * HZ)
 
 #define NVME_DEFAULT_KATO      5
 #define NVME_KATO_GRACE                10
 
                dev->admin_tagset.nr_hw_queues = 1;
 
                dev->admin_tagset.queue_depth = NVME_AQ_MQ_TAG_DEPTH;
-               dev->admin_tagset.timeout = ADMIN_TIMEOUT;
+               dev->admin_tagset.timeout = NVME_ADMIN_TIMEOUT;
                dev->admin_tagset.numa_node = dev->ctrl.numa_node;
                dev->admin_tagset.cmd_size = sizeof(struct nvme_iod);
                dev->admin_tagset.flags = BLK_MQ_F_NO_SCHED;
        unsigned long timeout;
 
  retry:
-       timeout = ADMIN_TIMEOUT;
+       timeout = NVME_ADMIN_TIMEOUT;
        while (nr_queues > 0) {
                if (nvme_delete_queue(&dev->queues[nr_queues], opcode))
                        break;
 
                                NVME_RDMA_DATA_SGL_SIZE;
                set->driver_data = ctrl;
                set->nr_hw_queues = 1;
-               set->timeout = ADMIN_TIMEOUT;
+               set->timeout = NVME_ADMIN_TIMEOUT;
                set->flags = BLK_MQ_F_NO_SCHED;
        } else {
                set = &ctrl->tag_set;
 
                set->cmd_size = sizeof(struct nvme_tcp_request);
                set->driver_data = ctrl;
                set->nr_hw_queues = 1;
-               set->timeout = ADMIN_TIMEOUT;
+               set->timeout = NVME_ADMIN_TIMEOUT;
        } else {
                set = &ctrl->tag_set;
                memset(set, 0, sizeof(*set));
 
                NVME_INLINE_SG_CNT * sizeof(struct scatterlist);
        ctrl->admin_tag_set.driver_data = ctrl;
        ctrl->admin_tag_set.nr_hw_queues = 1;
-       ctrl->admin_tag_set.timeout = ADMIN_TIMEOUT;
+       ctrl->admin_tag_set.timeout = NVME_ADMIN_TIMEOUT;
        ctrl->admin_tag_set.flags = BLK_MQ_F_NO_SCHED;
 
        ctrl->queues[0].ctrl = ctrl;