+++ /dev/null
-// SPDX-License-Identifier: GPL-2.0
-/*
- * Copyright 2021 Marvell. All rights reserved.
- */
-#define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
-/* Kernel includes */
-#include <linux/kernel.h>
-#include <linux/module.h>
-
-/* Driver includes */
-#include "tcp-offload.h"
-
-static LIST_HEAD(nvme_tcp_ofld_devices);
-static DEFINE_MUTEX(nvme_tcp_ofld_devices_mutex);
-static LIST_HEAD(nvme_tcp_ofld_ctrl_list);
-static DEFINE_MUTEX(nvme_tcp_ofld_ctrl_mutex);
-static struct blk_mq_ops nvme_tcp_ofld_admin_mq_ops;
-static struct blk_mq_ops nvme_tcp_ofld_mq_ops;
-
-static inline struct nvme_tcp_ofld_ctrl *to_tcp_ofld_ctrl(struct nvme_ctrl *nctrl)
-{
-       return container_of(nctrl, struct nvme_tcp_ofld_ctrl, nctrl);
-}
-
-static inline int nvme_tcp_ofld_qid(struct nvme_tcp_ofld_queue *queue)
-{
-       return queue - queue->ctrl->queues;
-}
-
-/**
- * nvme_tcp_ofld_register_dev() - NVMeTCP Offload Library registration
- * function.
- * @dev:       NVMeTCP offload device instance to be registered to the
- *             common tcp offload instance.
- *
- * API function that registers the type of vendor specific driver
- * being implemented to the common NVMe over TCP offload library. Part of
- * the overall init sequence of starting up an offload driver.
- */
-int nvme_tcp_ofld_register_dev(struct nvme_tcp_ofld_dev *dev)
-{
-       struct nvme_tcp_ofld_ops *ops = dev->ops;
-
-       if (!ops->claim_dev ||
-           !ops->setup_ctrl ||
-           !ops->release_ctrl ||
-           !ops->create_queue ||
-           !ops->drain_queue ||
-           !ops->destroy_queue ||
-           !ops->poll_queue ||
-           !ops->send_req)
-               return -EINVAL;
-
-       mutex_lock(&nvme_tcp_ofld_devices_mutex);
-       list_add_tail(&dev->entry, &nvme_tcp_ofld_devices);
-       mutex_unlock(&nvme_tcp_ofld_devices_mutex);
-
-       return 0;
-}
-EXPORT_SYMBOL_GPL(nvme_tcp_ofld_register_dev);
-
-/**
- * nvme_tcp_ofld_unregister_dev() - NVMeTCP Offload Library unregistration
- * function.
- * @dev:       NVMeTCP offload device instance to be unregistered from the
- *             common tcp offload instance.
- *
- * API function that unregisters the type of vendor specific driver being
- * implemented from the common NVMe over TCP offload library.
- * Part of the overall exit sequence of unloading the implemented driver.
- */
-void nvme_tcp_ofld_unregister_dev(struct nvme_tcp_ofld_dev *dev)
-{
-       mutex_lock(&nvme_tcp_ofld_devices_mutex);
-       list_del(&dev->entry);
-       mutex_unlock(&nvme_tcp_ofld_devices_mutex);
-}
-EXPORT_SYMBOL_GPL(nvme_tcp_ofld_unregister_dev);
-
-/**
- * nvme_tcp_ofld_error_recovery() - NVMeTCP Offload library error recovery.
- * function.
- * @nctrl:     NVMe controller instance to change to resetting.
- *
- * API function that change the controller state to resseting.
- * Part of the overall controller reset sequence.
- */
-void nvme_tcp_ofld_error_recovery(struct nvme_ctrl *nctrl)
-{
-       if (!nvme_change_ctrl_state(nctrl, NVME_CTRL_RESETTING))
-               return;
-
-       queue_work(nvme_reset_wq, &to_tcp_ofld_ctrl(nctrl)->err_work);
-}
-EXPORT_SYMBOL_GPL(nvme_tcp_ofld_error_recovery);
-
-/**
- * nvme_tcp_ofld_report_queue_err() - NVMeTCP Offload report error event
- * callback function. Pointed to by nvme_tcp_ofld_queue->report_err.
- * @queue:     NVMeTCP offload queue instance on which the error has occurred.
- *
- * API function that allows the vendor specific offload driver to reports errors
- * to the common offload layer, to invoke error recovery.
- */
-int nvme_tcp_ofld_report_queue_err(struct nvme_tcp_ofld_queue *queue)
-{
-       pr_err("nvme-tcp-offload queue error\n");
-       nvme_tcp_ofld_error_recovery(&queue->ctrl->nctrl);
-
-       return 0;
-}
-
-/**
- * nvme_tcp_ofld_req_done() - NVMeTCP Offload request done callback
- * function. Pointed to by nvme_tcp_ofld_req->done.
- * Handles both NVME_TCP_F_DATA_SUCCESS flag and NVMe CQ.
- * @req:       NVMeTCP offload request to complete.
- * @result:     The nvme_result.
- * @status:     The completion status.
- *
- * API function that allows the vendor specific offload driver to report request
- * completions to the common offload layer.
- */
-void nvme_tcp_ofld_req_done(struct nvme_tcp_ofld_req *req,
-                           union nvme_result *result,
-                           __le16 status)
-{
-       struct request *rq = blk_mq_rq_from_pdu(req);
-
-       if (!nvme_try_complete_req(rq, cpu_to_le16(status << 1), *result))
-               nvme_complete_rq(rq);
-}
-
-/**
- * nvme_tcp_ofld_async_req_done() - NVMeTCP Offload request done callback
- * function for async request. Pointed to by nvme_tcp_ofld_req->done.
- * Handles both NVME_TCP_F_DATA_SUCCESS flag and NVMe CQ.
- * @req:       NVMeTCP offload request to complete.
- * @result:     The nvme_result.
- * @status:     The completion status.
- *
- * API function that allows the vendor specific offload driver to report request
- * completions to the common offload layer.
- */
-void nvme_tcp_ofld_async_req_done(struct nvme_tcp_ofld_req *req,
-                                 union nvme_result *result, __le16 status)
-{
-       struct nvme_tcp_ofld_queue *queue = req->queue;
-       struct nvme_tcp_ofld_ctrl *ctrl = queue->ctrl;
-
-       nvme_complete_async_event(&ctrl->nctrl, status, result);
-}
-
-static struct nvme_tcp_ofld_dev *
-nvme_tcp_ofld_lookup_dev(struct nvme_tcp_ofld_ctrl *ctrl)
-{
-       struct nvme_tcp_ofld_dev *dev;
-
-       mutex_lock(&nvme_tcp_ofld_devices_mutex);
-       list_for_each_entry(dev, &nvme_tcp_ofld_devices, entry) {
-               if (dev->ops->claim_dev(dev, ctrl))
-                       goto out;
-       }
-
-       dev = NULL;
-out:
-       mutex_unlock(&nvme_tcp_ofld_devices_mutex);
-
-       return dev;
-}
-
-static struct blk_mq_tag_set *
-nvme_tcp_ofld_alloc_tagset(struct nvme_ctrl *nctrl, bool admin)
-{
-       struct nvme_tcp_ofld_ctrl *ctrl = to_tcp_ofld_ctrl(nctrl);
-       struct blk_mq_tag_set *set;
-       int rc;
-
-       if (admin) {
-               set = &ctrl->admin_tag_set;
-               memset(set, 0, sizeof(*set));
-               set->ops = &nvme_tcp_ofld_admin_mq_ops;
-               set->queue_depth = NVME_AQ_MQ_TAG_DEPTH;
-               set->reserved_tags = NVMF_RESERVED_TAGS;
-               set->numa_node = nctrl->numa_node;
-               set->flags = BLK_MQ_F_BLOCKING;
-               set->cmd_size = sizeof(struct nvme_tcp_ofld_req);
-               set->driver_data = ctrl;
-               set->nr_hw_queues = 1;
-               set->timeout = NVME_ADMIN_TIMEOUT;
-       } else {
-               set = &ctrl->tag_set;
-               memset(set, 0, sizeof(*set));
-               set->ops = &nvme_tcp_ofld_mq_ops;
-               set->queue_depth = nctrl->sqsize + 1;
-               set->reserved_tags = NVMF_RESERVED_TAGS;
-               set->numa_node = nctrl->numa_node;
-               set->flags = BLK_MQ_F_SHOULD_MERGE;
-               set->cmd_size = sizeof(struct nvme_tcp_ofld_req);
-               set->driver_data = ctrl;
-               set->nr_hw_queues = nctrl->queue_count - 1;
-               set->timeout = NVME_IO_TIMEOUT;
-               set->nr_maps = nctrl->opts->nr_poll_queues ? HCTX_MAX_TYPES : 2;
-       }
-
-       rc = blk_mq_alloc_tag_set(set);
-       if (rc)
-               return ERR_PTR(rc);
-
-       return set;
-}
-
-static void __nvme_tcp_ofld_stop_queue(struct nvme_tcp_ofld_queue *queue)
-{
-       queue->dev->ops->drain_queue(queue);
-}
-
-static void nvme_tcp_ofld_stop_queue(struct nvme_ctrl *nctrl, int qid)
-{
-       struct nvme_tcp_ofld_ctrl *ctrl = to_tcp_ofld_ctrl(nctrl);
-       struct nvme_tcp_ofld_queue *queue = &ctrl->queues[qid];
-
-       mutex_lock(&queue->queue_lock);
-       if (test_and_clear_bit(NVME_TCP_OFLD_Q_LIVE, &queue->flags))
-               __nvme_tcp_ofld_stop_queue(queue);
-       mutex_unlock(&queue->queue_lock);
-}
-
-static void nvme_tcp_ofld_stop_io_queues(struct nvme_ctrl *ctrl)
-{
-       int i;
-
-       for (i = 1; i < ctrl->queue_count; i++)
-               nvme_tcp_ofld_stop_queue(ctrl, i);
-}
-
-static void __nvme_tcp_ofld_free_queue(struct nvme_tcp_ofld_queue *queue)
-{
-       queue->dev->ops->destroy_queue(queue);
-}
-
-static void nvme_tcp_ofld_free_queue(struct nvme_ctrl *nctrl, int qid)
-{
-       struct nvme_tcp_ofld_ctrl *ctrl = to_tcp_ofld_ctrl(nctrl);
-       struct nvme_tcp_ofld_queue *queue = &ctrl->queues[qid];
-
-       if (test_and_clear_bit(NVME_TCP_OFLD_Q_ALLOCATED, &queue->flags)) {
-               __nvme_tcp_ofld_free_queue(queue);
-               mutex_destroy(&queue->queue_lock);
-       }
-}
-
-static void
-nvme_tcp_ofld_free_io_queues(struct nvme_ctrl *nctrl)
-{
-       int i;
-
-       for (i = 1; i < nctrl->queue_count; i++)
-               nvme_tcp_ofld_free_queue(nctrl, i);
-}
-
-static void nvme_tcp_ofld_destroy_io_queues(struct nvme_ctrl *nctrl, bool remove)
-{
-       nvme_tcp_ofld_stop_io_queues(nctrl);
-       if (remove) {
-               blk_cleanup_queue(nctrl->connect_q);
-               blk_mq_free_tag_set(nctrl->tagset);
-       }
-       nvme_tcp_ofld_free_io_queues(nctrl);
-}
-
-static void nvme_tcp_ofld_destroy_admin_queue(struct nvme_ctrl *nctrl, bool remove)
-{
-       nvme_tcp_ofld_stop_queue(nctrl, 0);
-       if (remove) {
-               blk_cleanup_queue(nctrl->admin_q);
-               blk_cleanup_queue(nctrl->fabrics_q);
-               blk_mq_free_tag_set(nctrl->admin_tagset);
-       }
-       nvme_tcp_ofld_free_queue(nctrl, 0);
-}
-
-static int nvme_tcp_ofld_start_queue(struct nvme_ctrl *nctrl, int qid)
-{
-       struct nvme_tcp_ofld_ctrl *ctrl = to_tcp_ofld_ctrl(nctrl);
-       struct nvme_tcp_ofld_queue *queue = &ctrl->queues[qid];
-       int rc;
-
-       queue = &ctrl->queues[qid];
-       if (qid) {
-               queue->cmnd_capsule_len = nctrl->ioccsz * 16;
-               rc = nvmf_connect_io_queue(nctrl, qid, false);
-       } else {
-               queue->cmnd_capsule_len = sizeof(struct nvme_command) + NVME_TCP_ADMIN_CCSZ;
-               rc = nvmf_connect_admin_queue(nctrl);
-       }
-
-       if (!rc) {
-               set_bit(NVME_TCP_OFLD_Q_LIVE, &queue->flags);
-       } else {
-               if (test_bit(NVME_TCP_OFLD_Q_ALLOCATED, &queue->flags))
-                       __nvme_tcp_ofld_stop_queue(queue);
-               dev_err(nctrl->device,
-                       "failed to connect queue: %d ret=%d\n", qid, rc);
-       }
-
-       return rc;
-}
-
-static int nvme_tcp_ofld_configure_admin_queue(struct nvme_ctrl *nctrl,
-                                              bool new)
-{
-       struct nvme_tcp_ofld_ctrl *ctrl = to_tcp_ofld_ctrl(nctrl);
-       struct nvme_tcp_ofld_queue *queue = &ctrl->queues[0];
-       int rc;
-
-       mutex_init(&queue->queue_lock);
-
-       rc = ctrl->dev->ops->create_queue(queue, 0, NVME_AQ_DEPTH);
-       if (rc)
-               return rc;
-
-       set_bit(NVME_TCP_OFLD_Q_ALLOCATED, &queue->flags);
-       if (new) {
-               nctrl->admin_tagset =
-                               nvme_tcp_ofld_alloc_tagset(nctrl, true);
-               if (IS_ERR(nctrl->admin_tagset)) {
-                       rc = PTR_ERR(nctrl->admin_tagset);
-                       nctrl->admin_tagset = NULL;
-                       goto out_free_queue;
-               }
-
-               nctrl->fabrics_q = blk_mq_init_queue(nctrl->admin_tagset);
-               if (IS_ERR(nctrl->fabrics_q)) {
-                       rc = PTR_ERR(nctrl->fabrics_q);
-                       nctrl->fabrics_q = NULL;
-                       goto out_free_tagset;
-               }
-
-               nctrl->admin_q = blk_mq_init_queue(nctrl->admin_tagset);
-               if (IS_ERR(nctrl->admin_q)) {
-                       rc = PTR_ERR(nctrl->admin_q);
-                       nctrl->admin_q = NULL;
-                       goto out_cleanup_fabrics_q;
-               }
-       }
-
-       rc = nvme_tcp_ofld_start_queue(nctrl, 0);
-       if (rc)
-               goto out_cleanup_queue;
-
-       rc = nvme_enable_ctrl(nctrl);
-       if (rc)
-               goto out_stop_queue;
-
-       blk_mq_unquiesce_queue(nctrl->admin_q);
-
-       rc = nvme_init_ctrl_finish(nctrl);
-       if (rc)
-               goto out_quiesce_queue;
-
-       return 0;
-
-out_quiesce_queue:
-       blk_mq_quiesce_queue(nctrl->admin_q);
-       blk_sync_queue(nctrl->admin_q);
-out_stop_queue:
-       nvme_tcp_ofld_stop_queue(nctrl, 0);
-       nvme_cancel_admin_tagset(nctrl);
-out_cleanup_queue:
-       if (new)
-               blk_cleanup_queue(nctrl->admin_q);
-out_cleanup_fabrics_q:
-       if (new)
-               blk_cleanup_queue(nctrl->fabrics_q);
-out_free_tagset:
-       if (new)
-               blk_mq_free_tag_set(nctrl->admin_tagset);
-out_free_queue:
-       nvme_tcp_ofld_free_queue(nctrl, 0);
-
-       return rc;
-}
-
-static unsigned int nvme_tcp_ofld_nr_io_queues(struct nvme_ctrl *nctrl)
-{
-       struct nvme_tcp_ofld_ctrl *ctrl = to_tcp_ofld_ctrl(nctrl);
-       struct nvme_tcp_ofld_dev *dev = ctrl->dev;
-       u32 hw_vectors = dev->num_hw_vectors;
-       u32 nr_write_queues, nr_poll_queues;
-       u32 nr_io_queues, nr_total_queues;
-
-       nr_io_queues = min3(nctrl->opts->nr_io_queues, num_online_cpus(),
-                           hw_vectors);
-       nr_write_queues = min3(nctrl->opts->nr_write_queues, num_online_cpus(),
-                              hw_vectors);
-       nr_poll_queues = min3(nctrl->opts->nr_poll_queues, num_online_cpus(),
-                             hw_vectors);
-
-       nr_total_queues = nr_io_queues + nr_write_queues + nr_poll_queues;
-
-       return nr_total_queues;
-}
-
-static void
-nvme_tcp_ofld_set_io_queues(struct nvme_ctrl *nctrl, unsigned int nr_io_queues)
-{
-       struct nvme_tcp_ofld_ctrl *ctrl = to_tcp_ofld_ctrl(nctrl);
-       struct nvmf_ctrl_options *opts = nctrl->opts;
-
-       if (opts->nr_write_queues && opts->nr_io_queues < nr_io_queues) {
-               /*
-                * separate read/write queues
-                * hand out dedicated default queues only after we have
-                * sufficient read queues.
-                */
-               ctrl->io_queues[HCTX_TYPE_READ] = opts->nr_io_queues;
-               nr_io_queues -= ctrl->io_queues[HCTX_TYPE_READ];
-               ctrl->io_queues[HCTX_TYPE_DEFAULT] =
-                       min(opts->nr_write_queues, nr_io_queues);
-               nr_io_queues -= ctrl->io_queues[HCTX_TYPE_DEFAULT];
-       } else {
-               /*
-                * shared read/write queues
-                * either no write queues were requested, or we don't have
-                * sufficient queue count to have dedicated default queues.
-                */
-               ctrl->io_queues[HCTX_TYPE_DEFAULT] =
-                       min(opts->nr_io_queues, nr_io_queues);
-               nr_io_queues -= ctrl->io_queues[HCTX_TYPE_DEFAULT];
-       }
-
-       if (opts->nr_poll_queues && nr_io_queues) {
-               /* map dedicated poll queues only if we have queues left */
-               ctrl->io_queues[HCTX_TYPE_POLL] =
-                       min(opts->nr_poll_queues, nr_io_queues);
-       }
-}
-
-static int nvme_tcp_ofld_create_io_queues(struct nvme_ctrl *nctrl)
-{
-       struct nvme_tcp_ofld_ctrl *ctrl = to_tcp_ofld_ctrl(nctrl);
-       int i, rc;
-
-       for (i = 1; i < nctrl->queue_count; i++) {
-               mutex_init(&ctrl->queues[i].queue_lock);
-
-               rc = ctrl->dev->ops->create_queue(&ctrl->queues[i],
-                                                 i, nctrl->sqsize + 1);
-               if (rc)
-                       goto out_free_queues;
-
-               set_bit(NVME_TCP_OFLD_Q_ALLOCATED, &ctrl->queues[i].flags);
-       }
-
-       return 0;
-
-out_free_queues:
-       for (i--; i >= 1; i--)
-               nvme_tcp_ofld_free_queue(nctrl, i);
-
-       return rc;
-}
-
-static int nvme_tcp_ofld_alloc_io_queues(struct nvme_ctrl *nctrl)
-{
-       unsigned int nr_io_queues;
-       int rc;
-
-       nr_io_queues = nvme_tcp_ofld_nr_io_queues(nctrl);
-       rc = nvme_set_queue_count(nctrl, &nr_io_queues);
-       if (rc)
-               return rc;
-
-       nctrl->queue_count = nr_io_queues + 1;
-       if (nctrl->queue_count < 2) {
-               dev_err(nctrl->device,
-                       "unable to set any I/O queues\n");
-
-               return -ENOMEM;
-       }
-
-       dev_info(nctrl->device, "creating %d I/O queues.\n", nr_io_queues);
-       nvme_tcp_ofld_set_io_queues(nctrl, nr_io_queues);
-
-       return nvme_tcp_ofld_create_io_queues(nctrl);
-}
-
-static int nvme_tcp_ofld_start_io_queues(struct nvme_ctrl *nctrl)
-{
-       int i, rc = 0;
-
-       for (i = 1; i < nctrl->queue_count; i++) {
-               rc = nvme_tcp_ofld_start_queue(nctrl, i);
-               if (rc)
-                       goto out_stop_queues;
-       }
-
-       return 0;
-
-out_stop_queues:
-       for (i--; i >= 1; i--)
-               nvme_tcp_ofld_stop_queue(nctrl, i);
-
-       return rc;
-}
-
-static int
-nvme_tcp_ofld_configure_io_queues(struct nvme_ctrl *nctrl, bool new)
-{
-       int rc = nvme_tcp_ofld_alloc_io_queues(nctrl);
-
-       if (rc)
-               return rc;
-
-       if (new) {
-               nctrl->tagset = nvme_tcp_ofld_alloc_tagset(nctrl, false);
-               if (IS_ERR(nctrl->tagset)) {
-                       rc = PTR_ERR(nctrl->tagset);
-                       nctrl->tagset = NULL;
-                       goto out_free_io_queues;
-               }
-
-               nctrl->connect_q = blk_mq_init_queue(nctrl->tagset);
-               if (IS_ERR(nctrl->connect_q)) {
-                       rc = PTR_ERR(nctrl->connect_q);
-                       nctrl->connect_q = NULL;
-                       goto out_free_tag_set;
-               }
-       }
-
-       rc = nvme_tcp_ofld_start_io_queues(nctrl);
-       if (rc)
-               goto out_cleanup_connect_q;
-
-       if (!new) {
-               nvme_start_queues(nctrl);
-               if (!nvme_wait_freeze_timeout(nctrl, NVME_IO_TIMEOUT)) {
-                       /*
-                        * If we timed out waiting for freeze we are likely to
-                        * be stuck.  Fail the controller initialization just
-                        * to be safe.
-                        */
-                       rc = -ENODEV;
-                       goto out_wait_freeze_timed_out;
-               }
-               blk_mq_update_nr_hw_queues(nctrl->tagset, nctrl->queue_count - 1);
-               nvme_unfreeze(nctrl);
-       }
-
-       return 0;
-
-out_wait_freeze_timed_out:
-       nvme_stop_queues(nctrl);
-       nvme_sync_io_queues(nctrl);
-       nvme_tcp_ofld_stop_io_queues(nctrl);
-out_cleanup_connect_q:
-       nvme_cancel_tagset(nctrl);
-       if (new)
-               blk_cleanup_queue(nctrl->connect_q);
-out_free_tag_set:
-       if (new)
-               blk_mq_free_tag_set(nctrl->tagset);
-out_free_io_queues:
-       nvme_tcp_ofld_free_io_queues(nctrl);
-
-       return rc;
-}
-
-static void nvme_tcp_ofld_reconnect_or_remove(struct nvme_ctrl *nctrl)
-{
-       /* If we are resetting/deleting then do nothing */
-       if (nctrl->state != NVME_CTRL_CONNECTING) {
-               WARN_ON_ONCE(nctrl->state == NVME_CTRL_NEW ||
-                            nctrl->state == NVME_CTRL_LIVE);
-
-               return;
-       }
-
-       if (nvmf_should_reconnect(nctrl)) {
-               dev_info(nctrl->device, "Reconnecting in %d seconds...\n",
-                        nctrl->opts->reconnect_delay);
-               queue_delayed_work(nvme_wq,
-                                  &to_tcp_ofld_ctrl(nctrl)->connect_work,
-                                  nctrl->opts->reconnect_delay * HZ);
-       } else {
-               dev_info(nctrl->device, "Removing controller...\n");
-               nvme_delete_ctrl(nctrl);
-       }
-}
-
-static int
-nvme_tcp_ofld_init_admin_hctx(struct blk_mq_hw_ctx *hctx, void *data,
-                             unsigned int hctx_idx)
-{
-       struct nvme_tcp_ofld_ctrl *ctrl = data;
-
-       hctx->driver_data = &ctrl->queues[0];
-
-       return 0;
-}
-
-static int nvme_tcp_ofld_setup_ctrl(struct nvme_ctrl *nctrl, bool new)
-{
-       struct nvme_tcp_ofld_ctrl *ctrl = to_tcp_ofld_ctrl(nctrl);
-       struct nvmf_ctrl_options *opts = nctrl->opts;
-       int rc = 0;
-
-       rc = ctrl->dev->ops->setup_ctrl(ctrl);
-       if (rc)
-               return rc;
-
-       rc = nvme_tcp_ofld_configure_admin_queue(nctrl, new);
-       if (rc)
-               goto out_release_ctrl;
-
-       if (nctrl->icdoff) {
-               dev_err(nctrl->device, "icdoff is not supported!\n");
-               rc = -EINVAL;
-               goto destroy_admin;
-       }
-
-       if (!(nctrl->sgls & ((1 << 0) | (1 << 1)))) {
-               dev_err(nctrl->device, "Mandatory sgls are not supported!\n");
-               goto destroy_admin;
-       }
-
-       if (opts->queue_size > nctrl->sqsize + 1)
-               dev_warn(nctrl->device,
-                        "queue_size %zu > ctrl sqsize %u, clamping down\n",
-                        opts->queue_size, nctrl->sqsize + 1);
-
-       if (nctrl->sqsize + 1 > nctrl->maxcmd) {
-               dev_warn(nctrl->device,
-                        "sqsize %u > ctrl maxcmd %u, clamping down\n",
-                        nctrl->sqsize + 1, nctrl->maxcmd);
-               nctrl->sqsize = nctrl->maxcmd - 1;
-       }
-
-       if (nctrl->queue_count > 1) {
-               rc = nvme_tcp_ofld_configure_io_queues(nctrl, new);
-               if (rc)
-                       goto destroy_admin;
-       }
-
-       if (!nvme_change_ctrl_state(nctrl, NVME_CTRL_LIVE)) {
-               /*
-                * state change failure is ok if we started ctrl delete,
-                * unless we're during creation of a new controller to
-                * avoid races with teardown flow.
-                */
-               WARN_ON_ONCE(nctrl->state != NVME_CTRL_DELETING &&
-                            nctrl->state != NVME_CTRL_DELETING_NOIO);
-               WARN_ON_ONCE(new);
-               rc = -EINVAL;
-               goto destroy_io;
-       }
-
-       nvme_start_ctrl(nctrl);
-
-       return 0;
-
-destroy_io:
-       if (nctrl->queue_count > 1) {
-               nvme_stop_queues(nctrl);
-               nvme_sync_io_queues(nctrl);
-               nvme_tcp_ofld_stop_io_queues(nctrl);
-               nvme_cancel_tagset(nctrl);
-               nvme_tcp_ofld_destroy_io_queues(nctrl, new);
-       }
-destroy_admin:
-       blk_mq_quiesce_queue(nctrl->admin_q);
-       blk_sync_queue(nctrl->admin_q);
-       nvme_tcp_ofld_stop_queue(nctrl, 0);
-       nvme_cancel_admin_tagset(nctrl);
-       nvme_tcp_ofld_destroy_admin_queue(nctrl, new);
-out_release_ctrl:
-       ctrl->dev->ops->release_ctrl(ctrl);
-
-       return rc;
-}
-
-static int
-nvme_tcp_ofld_check_dev_opts(struct nvmf_ctrl_options *opts,
-                            struct nvme_tcp_ofld_ops *ofld_ops)
-{
-       unsigned int nvme_tcp_ofld_opt_mask = NVMF_ALLOWED_OPTS |
-                       ofld_ops->allowed_opts | ofld_ops->required_opts;
-       struct nvmf_ctrl_options dev_opts_mask;
-
-       if (opts->mask & ~nvme_tcp_ofld_opt_mask) {
-               pr_warn("One or more nvmf options missing from ofld drvr %s.\n",
-                       ofld_ops->name);
-
-               dev_opts_mask.mask = nvme_tcp_ofld_opt_mask;
-
-               return nvmf_check_required_opts(&dev_opts_mask, opts->mask);
-       }
-
-       return 0;
-}
-
-static void nvme_tcp_ofld_free_ctrl(struct nvme_ctrl *nctrl)
-{
-       struct nvme_tcp_ofld_ctrl *ctrl = to_tcp_ofld_ctrl(nctrl);
-       struct nvme_tcp_ofld_dev *dev = ctrl->dev;
-
-       if (list_empty(&ctrl->list))
-               goto free_ctrl;
-
-       ctrl->dev->ops->release_ctrl(ctrl);
-
-       mutex_lock(&nvme_tcp_ofld_ctrl_mutex);
-       list_del(&ctrl->list);
-       mutex_unlock(&nvme_tcp_ofld_ctrl_mutex);
-
-       nvmf_free_options(nctrl->opts);
-free_ctrl:
-       module_put(dev->ops->module);
-       kfree(ctrl->queues);
-       kfree(ctrl);
-}
-
-static void nvme_tcp_ofld_set_sg_null(struct nvme_command *c)
-{
-       struct nvme_sgl_desc *sg = &c->common.dptr.sgl;
-
-       sg->addr = 0;
-       sg->length = 0;
-       sg->type = (NVME_TRANSPORT_SGL_DATA_DESC << 4) | NVME_SGL_FMT_TRANSPORT_A;
-}
-
-inline void nvme_tcp_ofld_set_sg_inline(struct nvme_tcp_ofld_queue *queue,
-                                       struct nvme_command *c, u32 data_len)
-{
-       struct nvme_sgl_desc *sg = &c->common.dptr.sgl;
-
-       sg->addr = cpu_to_le64(queue->ctrl->nctrl.icdoff);
-       sg->length = cpu_to_le32(data_len);
-       sg->type = (NVME_SGL_FMT_DATA_DESC << 4) | NVME_SGL_FMT_OFFSET;
-}
-
-static void nvme_tcp_ofld_map_data(struct nvme_command *c, u32 data_len)
-{
-       struct nvme_sgl_desc *sg = &c->common.dptr.sgl;
-
-       sg->addr = 0;
-       sg->length = cpu_to_le32(data_len);
-       sg->type = (NVME_TRANSPORT_SGL_DATA_DESC << 4) | NVME_SGL_FMT_TRANSPORT_A;
-}
-
-static void nvme_tcp_ofld_submit_async_event(struct nvme_ctrl *arg)
-{
-       struct nvme_tcp_ofld_ctrl *ctrl = to_tcp_ofld_ctrl(arg);
-       struct nvme_tcp_ofld_queue *queue = &ctrl->queues[0];
-       struct nvme_tcp_ofld_dev *dev = queue->dev;
-       struct nvme_tcp_ofld_ops *ops = dev->ops;
-
-       ctrl->async_req.nvme_cmd.common.opcode = nvme_admin_async_event;
-       ctrl->async_req.nvme_cmd.common.command_id = NVME_AQ_BLK_MQ_DEPTH;
-       ctrl->async_req.nvme_cmd.common.flags |= NVME_CMD_SGL_METABUF;
-
-       nvme_tcp_ofld_set_sg_null(&ctrl->async_req.nvme_cmd);
-
-       ctrl->async_req.async = true;
-       ctrl->async_req.queue = queue;
-       ctrl->async_req.done = nvme_tcp_ofld_async_req_done;
-
-       ops->send_req(&ctrl->async_req);
-}
-
-static void
-nvme_tcp_ofld_teardown_admin_queue(struct nvme_ctrl *nctrl, bool remove)
-{
-       blk_mq_quiesce_queue(nctrl->admin_q);
-       blk_sync_queue(nctrl->admin_q);
-
-       nvme_tcp_ofld_stop_queue(nctrl, 0);
-       nvme_cancel_admin_tagset(nctrl);
-
-       if (remove)
-               blk_mq_unquiesce_queue(nctrl->admin_q);
-
-       nvme_tcp_ofld_destroy_admin_queue(nctrl, remove);
-}
-
-static void
-nvme_tcp_ofld_teardown_io_queues(struct nvme_ctrl *nctrl, bool remove)
-{
-       if (nctrl->queue_count <= 1)
-               return;
-
-       blk_mq_quiesce_queue(nctrl->admin_q);
-       nvme_start_freeze(nctrl);
-       nvme_stop_queues(nctrl);
-       nvme_sync_io_queues(nctrl);
-       nvme_tcp_ofld_stop_io_queues(nctrl);
-       nvme_cancel_tagset(nctrl);
-
-       if (remove)
-               nvme_start_queues(nctrl);
-
-       nvme_tcp_ofld_destroy_io_queues(nctrl, remove);
-}
-
-static void nvme_tcp_ofld_reconnect_ctrl_work(struct work_struct *work)
-{
-       struct nvme_tcp_ofld_ctrl *ctrl =
-                               container_of(to_delayed_work(work),
-                                            struct nvme_tcp_ofld_ctrl,
-                                            connect_work);
-       struct nvme_ctrl *nctrl = &ctrl->nctrl;
-
-       ++nctrl->nr_reconnects;
-
-       if (nvme_tcp_ofld_setup_ctrl(nctrl, false))
-               goto requeue;
-
-       dev_info(nctrl->device, "Successfully reconnected (%d attempt)\n",
-                nctrl->nr_reconnects);
-
-       nctrl->nr_reconnects = 0;
-
-       return;
-
-requeue:
-       dev_info(nctrl->device, "Failed reconnect attempt %d\n",
-                nctrl->nr_reconnects);
-       nvme_tcp_ofld_reconnect_or_remove(nctrl);
-}
-
-static void nvme_tcp_ofld_error_recovery_work(struct work_struct *work)
-{
-       struct nvme_tcp_ofld_ctrl *ctrl =
-               container_of(work, struct nvme_tcp_ofld_ctrl, err_work);
-       struct nvme_ctrl *nctrl = &ctrl->nctrl;
-
-       nvme_stop_keep_alive(nctrl);
-       nvme_tcp_ofld_teardown_io_queues(nctrl, false);
-       /* unquiesce to fail fast pending requests */
-       nvme_start_queues(nctrl);
-       nvme_tcp_ofld_teardown_admin_queue(nctrl, false);
-       blk_mq_unquiesce_queue(nctrl->admin_q);
-
-       if (!nvme_change_ctrl_state(nctrl, NVME_CTRL_CONNECTING)) {
-               /* state change failure is ok if we started nctrl delete */
-               WARN_ON_ONCE(nctrl->state != NVME_CTRL_DELETING &&
-                            nctrl->state != NVME_CTRL_DELETING_NOIO);
-
-               return;
-       }
-
-       nvme_tcp_ofld_reconnect_or_remove(nctrl);
-}
-
-static void
-nvme_tcp_ofld_teardown_ctrl(struct nvme_ctrl *nctrl, bool shutdown)
-{
-       struct nvme_tcp_ofld_ctrl *ctrl = to_tcp_ofld_ctrl(nctrl);
-
-       cancel_work_sync(&ctrl->err_work);
-       cancel_delayed_work_sync(&ctrl->connect_work);
-       nvme_tcp_ofld_teardown_io_queues(nctrl, shutdown);
-       blk_mq_quiesce_queue(nctrl->admin_q);
-       if (shutdown)
-               nvme_shutdown_ctrl(nctrl);
-       else
-               nvme_disable_ctrl(nctrl);
-       nvme_tcp_ofld_teardown_admin_queue(nctrl, shutdown);
-}
-
-static void nvme_tcp_ofld_delete_ctrl(struct nvme_ctrl *nctrl)
-{
-       nvme_tcp_ofld_teardown_ctrl(nctrl, true);
-}
-
-static void nvme_tcp_ofld_reset_ctrl_work(struct work_struct *work)
-{
-       struct nvme_ctrl *nctrl =
-               container_of(work, struct nvme_ctrl, reset_work);
-
-       nvme_stop_ctrl(nctrl);
-       nvme_tcp_ofld_teardown_ctrl(nctrl, false);
-
-       if (!nvme_change_ctrl_state(nctrl, NVME_CTRL_CONNECTING)) {
-               /* state change failure is ok if we started ctrl delete */
-               WARN_ON_ONCE(nctrl->state != NVME_CTRL_DELETING &&
-                            nctrl->state != NVME_CTRL_DELETING_NOIO);
-
-               return;
-       }
-
-       if (nvme_tcp_ofld_setup_ctrl(nctrl, false))
-               goto out_fail;
-
-       return;
-
-out_fail:
-       ++nctrl->nr_reconnects;
-       nvme_tcp_ofld_reconnect_or_remove(nctrl);
-}
-
-static int
-nvme_tcp_ofld_init_request(struct blk_mq_tag_set *set,
-                          struct request *rq,
-                          unsigned int hctx_idx,
-                          unsigned int numa_node)
-{
-       struct nvme_tcp_ofld_req *req = blk_mq_rq_to_pdu(rq);
-       struct nvme_tcp_ofld_ctrl *ctrl = set->driver_data;
-       int qid;
-
-       qid = (set == &ctrl->tag_set) ? hctx_idx + 1 : 0;
-       req->queue = &ctrl->queues[qid];
-       nvme_req(rq)->ctrl = &ctrl->nctrl;
-       nvme_req(rq)->cmd = &req->nvme_cmd;
-       req->done = nvme_tcp_ofld_req_done;
-
-       return 0;
-}
-
-inline size_t nvme_tcp_ofld_inline_data_size(struct nvme_tcp_ofld_queue *queue)
-{
-       return queue->cmnd_capsule_len - sizeof(struct nvme_command);
-}
-EXPORT_SYMBOL_GPL(nvme_tcp_ofld_inline_data_size);
-
-static blk_status_t
-nvme_tcp_ofld_queue_rq(struct blk_mq_hw_ctx *hctx,
-                      const struct blk_mq_queue_data *bd)
-{
-       struct nvme_tcp_ofld_req *req = blk_mq_rq_to_pdu(bd->rq);
-       struct nvme_tcp_ofld_queue *queue = hctx->driver_data;
-       struct nvme_tcp_ofld_ctrl *ctrl = queue->ctrl;
-       struct nvme_ns *ns = hctx->queue->queuedata;
-       struct nvme_tcp_ofld_dev *dev = queue->dev;
-       struct nvme_tcp_ofld_ops *ops = dev->ops;
-       struct nvme_command *nvme_cmd;
-       struct request *rq = bd->rq;
-       bool queue_ready;
-       u32 data_len;
-       int rc;
-
-       queue_ready = test_bit(NVME_TCP_OFLD_Q_LIVE, &queue->flags);
-
-       req->async = false;
-
-       if (!nvme_check_ready(&ctrl->nctrl, rq, queue_ready))
-               return nvme_fail_nonready_command(&ctrl->nctrl, rq);
-
-       rc = nvme_setup_cmd(ns, rq);
-       if (unlikely(rc))
-               return rc;
-
-       blk_mq_start_request(rq);
-
-       nvme_cmd = &req->nvme_cmd;
-       nvme_cmd->common.flags |= NVME_CMD_SGL_METABUF;
-
-       data_len = blk_rq_nr_phys_segments(rq) ? blk_rq_payload_bytes(rq) : 0;
-       if (!data_len)
-               nvme_tcp_ofld_set_sg_null(&req->nvme_cmd);
-       else if ((rq_data_dir(rq) == WRITE) &&
-                data_len <= nvme_tcp_ofld_inline_data_size(queue))
-               nvme_tcp_ofld_set_sg_inline(queue, nvme_cmd, data_len);
-       else
-               nvme_tcp_ofld_map_data(nvme_cmd, data_len);
-
-       rc = ops->send_req(req);
-       if (unlikely(rc))
-               return rc;
-
-       return BLK_STS_OK;
-}
-
-static void
-nvme_tcp_ofld_exit_request(struct blk_mq_tag_set *set,
-                          struct request *rq, unsigned int hctx_idx)
-{
-       /*
-        * Nothing is allocated in nvme_tcp_ofld_init_request,
-        * hence empty.
-        */
-}
-
-static int
-nvme_tcp_ofld_init_hctx(struct blk_mq_hw_ctx *hctx, void *data,
-                       unsigned int hctx_idx)
-{
-       struct nvme_tcp_ofld_ctrl *ctrl = data;
-
-       hctx->driver_data = &ctrl->queues[hctx_idx + 1];
-
-       return 0;
-}
-
-static int nvme_tcp_ofld_map_queues(struct blk_mq_tag_set *set)
-{
-       struct nvme_tcp_ofld_ctrl *ctrl = set->driver_data;
-       struct nvmf_ctrl_options *opts = ctrl->nctrl.opts;
-
-       if (opts->nr_write_queues && ctrl->io_queues[HCTX_TYPE_READ]) {
-               /* separate read/write queues */
-               set->map[HCTX_TYPE_DEFAULT].nr_queues =
-                       ctrl->io_queues[HCTX_TYPE_DEFAULT];
-               set->map[HCTX_TYPE_DEFAULT].queue_offset = 0;
-               set->map[HCTX_TYPE_READ].nr_queues =
-                       ctrl->io_queues[HCTX_TYPE_READ];
-               set->map[HCTX_TYPE_READ].queue_offset =
-                       ctrl->io_queues[HCTX_TYPE_DEFAULT];
-       } else {
-               /* shared read/write queues */
-               set->map[HCTX_TYPE_DEFAULT].nr_queues =
-                       ctrl->io_queues[HCTX_TYPE_DEFAULT];
-               set->map[HCTX_TYPE_DEFAULT].queue_offset = 0;
-               set->map[HCTX_TYPE_READ].nr_queues =
-                       ctrl->io_queues[HCTX_TYPE_DEFAULT];
-               set->map[HCTX_TYPE_READ].queue_offset = 0;
-       }
-       blk_mq_map_queues(&set->map[HCTX_TYPE_DEFAULT]);
-       blk_mq_map_queues(&set->map[HCTX_TYPE_READ]);
-
-       if (opts->nr_poll_queues && ctrl->io_queues[HCTX_TYPE_POLL]) {
-               /* map dedicated poll queues only if we have queues left */
-               set->map[HCTX_TYPE_POLL].nr_queues =
-                               ctrl->io_queues[HCTX_TYPE_POLL];
-               set->map[HCTX_TYPE_POLL].queue_offset =
-                       ctrl->io_queues[HCTX_TYPE_DEFAULT] +
-                       ctrl->io_queues[HCTX_TYPE_READ];
-               blk_mq_map_queues(&set->map[HCTX_TYPE_POLL]);
-       }
-
-       dev_info(ctrl->nctrl.device,
-                "mapped %d/%d/%d default/read/poll queues.\n",
-                ctrl->io_queues[HCTX_TYPE_DEFAULT],
-                ctrl->io_queues[HCTX_TYPE_READ],
-                ctrl->io_queues[HCTX_TYPE_POLL]);
-
-       return 0;
-}
-
-static int nvme_tcp_ofld_poll(struct blk_mq_hw_ctx *hctx)
-{
-       struct nvme_tcp_ofld_queue *queue = hctx->driver_data;
-       struct nvme_tcp_ofld_dev *dev = queue->dev;
-       struct nvme_tcp_ofld_ops *ops = dev->ops;
-
-       return ops->poll_queue(queue);
-}
-
-static void nvme_tcp_ofld_complete_timed_out(struct request *rq)
-{
-       struct nvme_tcp_ofld_req *req = blk_mq_rq_to_pdu(rq);
-       struct nvme_ctrl *nctrl = &req->queue->ctrl->nctrl;
-
-       nvme_tcp_ofld_stop_queue(nctrl, nvme_tcp_ofld_qid(req->queue));
-       if (blk_mq_request_started(rq) && !blk_mq_request_completed(rq)) {
-               nvme_req(rq)->status = NVME_SC_HOST_ABORTED_CMD;
-               blk_mq_complete_request(rq);
-       }
-}
-
-static enum blk_eh_timer_return nvme_tcp_ofld_timeout(struct request *rq, bool reserved)
-{
-       struct nvme_tcp_ofld_req *req = blk_mq_rq_to_pdu(rq);
-       struct nvme_tcp_ofld_ctrl *ctrl = req->queue->ctrl;
-
-       dev_warn(ctrl->nctrl.device,
-                "queue %d: timeout request %#x type %d\n",
-                nvme_tcp_ofld_qid(req->queue), rq->tag, req->nvme_cmd.common.opcode);
-
-       if (ctrl->nctrl.state != NVME_CTRL_LIVE) {
-               /*
-                * If we are resetting, connecting or deleting we should
-                * complete immediately because we may block controller
-                * teardown or setup sequence
-                * - ctrl disable/shutdown fabrics requests
-                * - connect requests
-                * - initialization admin requests
-                * - I/O requests that entered after unquiescing and
-                *   the controller stopped responding
-                *
-                * All other requests should be cancelled by the error
-                * recovery work, so it's fine that we fail it here.
-                */
-               nvme_tcp_ofld_complete_timed_out(rq);
-
-               return BLK_EH_DONE;
-       }
-
-       nvme_tcp_ofld_error_recovery(&ctrl->nctrl);
-
-       return BLK_EH_RESET_TIMER;
-}
-
-static struct blk_mq_ops nvme_tcp_ofld_mq_ops = {
-       .queue_rq       = nvme_tcp_ofld_queue_rq,
-       .complete       = nvme_complete_rq,
-       .init_request   = nvme_tcp_ofld_init_request,
-       .exit_request   = nvme_tcp_ofld_exit_request,
-       .init_hctx      = nvme_tcp_ofld_init_hctx,
-       .timeout        = nvme_tcp_ofld_timeout,
-       .map_queues     = nvme_tcp_ofld_map_queues,
-       .poll           = nvme_tcp_ofld_poll,
-};
-
-static struct blk_mq_ops nvme_tcp_ofld_admin_mq_ops = {
-       .queue_rq       = nvme_tcp_ofld_queue_rq,
-       .complete       = nvme_complete_rq,
-       .init_request   = nvme_tcp_ofld_init_request,
-       .exit_request   = nvme_tcp_ofld_exit_request,
-       .init_hctx      = nvme_tcp_ofld_init_admin_hctx,
-       .timeout        = nvme_tcp_ofld_timeout,
-};
-
-static const struct nvme_ctrl_ops nvme_tcp_ofld_ctrl_ops = {
-       .name                   = "tcp_offload",
-       .module                 = THIS_MODULE,
-       .flags                  = NVME_F_FABRICS,
-       .reg_read32             = nvmf_reg_read32,
-       .reg_read64             = nvmf_reg_read64,
-       .reg_write32            = nvmf_reg_write32,
-       .free_ctrl              = nvme_tcp_ofld_free_ctrl,
-       .submit_async_event     = nvme_tcp_ofld_submit_async_event,
-       .delete_ctrl            = nvme_tcp_ofld_delete_ctrl,
-       .get_address            = nvmf_get_address,
-};
-
-static bool
-nvme_tcp_ofld_existing_controller(struct nvmf_ctrl_options *opts)
-{
-       struct nvme_tcp_ofld_ctrl *ctrl;
-       bool found = false;
-
-       mutex_lock(&nvme_tcp_ofld_ctrl_mutex);
-       list_for_each_entry(ctrl, &nvme_tcp_ofld_ctrl_list, list) {
-               found = nvmf_ip_options_match(&ctrl->nctrl, opts);
-               if (found)
-                       break;
-       }
-       mutex_unlock(&nvme_tcp_ofld_ctrl_mutex);
-
-       return found;
-}
-
-static struct nvme_ctrl *
-nvme_tcp_ofld_create_ctrl(struct device *ndev, struct nvmf_ctrl_options *opts)
-{
-       struct nvme_tcp_ofld_queue *queue;
-       struct nvme_tcp_ofld_ctrl *ctrl;
-       struct nvme_tcp_ofld_dev *dev;
-       struct nvme_ctrl *nctrl;
-       int i, rc = 0;
-
-       ctrl = kzalloc(sizeof(*ctrl), GFP_KERNEL);
-       if (!ctrl)
-               return ERR_PTR(-ENOMEM);
-
-       INIT_LIST_HEAD(&ctrl->list);
-       nctrl = &ctrl->nctrl;
-       nctrl->opts = opts;
-       nctrl->queue_count = opts->nr_io_queues + opts->nr_write_queues +
-                            opts->nr_poll_queues + 1;
-       nctrl->sqsize = opts->queue_size - 1;
-       nctrl->kato = opts->kato;
-       INIT_DELAYED_WORK(&ctrl->connect_work,
-                         nvme_tcp_ofld_reconnect_ctrl_work);
-       INIT_WORK(&ctrl->err_work, nvme_tcp_ofld_error_recovery_work);
-       INIT_WORK(&nctrl->reset_work, nvme_tcp_ofld_reset_ctrl_work);
-       if (!(opts->mask & NVMF_OPT_TRSVCID)) {
-               opts->trsvcid =
-                       kstrdup(__stringify(NVME_TCP_DISC_PORT), GFP_KERNEL);
-               if (!opts->trsvcid) {
-                       rc = -ENOMEM;
-                       goto out_free_ctrl;
-               }
-               opts->mask |= NVMF_OPT_TRSVCID;
-       }
-
-       rc = inet_pton_with_scope(&init_net, AF_UNSPEC, opts->traddr,
-                                 opts->trsvcid,
-                                 &ctrl->conn_params.remote_ip_addr);
-       if (rc) {
-               pr_err("malformed address passed: %s:%s\n",
-                      opts->traddr, opts->trsvcid);
-               goto out_free_ctrl;
-       }
-
-       if (opts->mask & NVMF_OPT_HOST_TRADDR) {
-               rc = inet_pton_with_scope(&init_net, AF_UNSPEC,
-                                         opts->host_traddr, NULL,
-                                         &ctrl->conn_params.local_ip_addr);
-               if (rc) {
-                       pr_err("malformed src address passed: %s\n",
-                              opts->host_traddr);
-                       goto out_free_ctrl;
-               }
-       }
-
-       if (!opts->duplicate_connect &&
-           nvme_tcp_ofld_existing_controller(opts)) {
-               rc = -EALREADY;
-               goto out_free_ctrl;
-       }
-
-       /* Find device that can reach the dest addr */
-       dev = nvme_tcp_ofld_lookup_dev(ctrl);
-       if (!dev) {
-               pr_info("no device found for addr %s:%s.\n",
-                       opts->traddr, opts->trsvcid);
-               rc = -EINVAL;
-               goto out_free_ctrl;
-       }
-
-       /* Increase driver refcnt */
-       if (!try_module_get(dev->ops->module)) {
-               pr_err("try_module_get failed\n");
-               dev = NULL;
-               goto out_free_ctrl;
-       }
-
-       rc = nvme_tcp_ofld_check_dev_opts(opts, dev->ops);
-       if (rc)
-               goto out_module_put;
-
-       ctrl->dev = dev;
-
-       if (ctrl->dev->ops->max_hw_sectors)
-               nctrl->max_hw_sectors = ctrl->dev->ops->max_hw_sectors;
-       if (ctrl->dev->ops->max_segments)
-               nctrl->max_segments = ctrl->dev->ops->max_segments;
-
-       ctrl->queues = kcalloc(nctrl->queue_count,
-                              sizeof(struct nvme_tcp_ofld_queue),
-                              GFP_KERNEL);
-       if (!ctrl->queues) {
-               rc = -ENOMEM;
-               goto out_module_put;
-       }
-
-       for (i = 0; i < nctrl->queue_count; ++i) {
-               queue = &ctrl->queues[i];
-               queue->ctrl = ctrl;
-               queue->dev = dev;
-               queue->report_err = nvme_tcp_ofld_report_queue_err;
-       }
-
-       rc = nvme_init_ctrl(nctrl, ndev, &nvme_tcp_ofld_ctrl_ops, 0);
-       if (rc)
-               goto out_free_queues;
-
-       if (!nvme_change_ctrl_state(nctrl, NVME_CTRL_CONNECTING)) {
-               WARN_ON_ONCE(1);
-               rc = -EINTR;
-               goto out_uninit_ctrl;
-       }
-
-       rc = nvme_tcp_ofld_setup_ctrl(nctrl, true);
-       if (rc)
-               goto out_uninit_ctrl;
-
-       dev_info(nctrl->device, "new ctrl: NQN \"%s\", addr %pISp\n",
-                opts->subsysnqn, &ctrl->conn_params.remote_ip_addr);
-
-       mutex_lock(&nvme_tcp_ofld_ctrl_mutex);
-       list_add_tail(&ctrl->list, &nvme_tcp_ofld_ctrl_list);
-       mutex_unlock(&nvme_tcp_ofld_ctrl_mutex);
-
-       return nctrl;
-
-out_uninit_ctrl:
-       nvme_uninit_ctrl(nctrl);
-       nvme_put_ctrl(nctrl);
-out_free_queues:
-       kfree(ctrl->queues);
-out_module_put:
-       module_put(dev->ops->module);
-out_free_ctrl:
-       kfree(ctrl);
-
-       return ERR_PTR(rc);
-}
-
-static struct nvmf_transport_ops nvme_tcp_ofld_transport = {
-       .name           = "tcp_offload",
-       .module         = THIS_MODULE,
-       .required_opts  = NVMF_OPT_TRADDR,
-       .allowed_opts   = NVMF_OPT_TRSVCID | NVMF_OPT_NR_WRITE_QUEUES  |
-                         NVMF_OPT_HOST_TRADDR | NVMF_OPT_CTRL_LOSS_TMO |
-                         NVMF_OPT_RECONNECT_DELAY | NVMF_OPT_HDR_DIGEST |
-                         NVMF_OPT_DATA_DIGEST | NVMF_OPT_NR_POLL_QUEUES |
-                         NVMF_OPT_TOS,
-       .create_ctrl    = nvme_tcp_ofld_create_ctrl,
-};
-
-static int __init nvme_tcp_ofld_init_module(void)
-{
-       nvmf_register_transport(&nvme_tcp_ofld_transport);
-
-       return 0;
-}
-
-static void __exit nvme_tcp_ofld_cleanup_module(void)
-{
-       struct nvme_tcp_ofld_ctrl *ctrl;
-
-       nvmf_unregister_transport(&nvme_tcp_ofld_transport);
-
-       mutex_lock(&nvme_tcp_ofld_ctrl_mutex);
-       list_for_each_entry(ctrl, &nvme_tcp_ofld_ctrl_list, list)
-               nvme_delete_ctrl(&ctrl->nctrl);
-       mutex_unlock(&nvme_tcp_ofld_ctrl_mutex);
-       flush_workqueue(nvme_delete_wq);
-}
-
-module_init(nvme_tcp_ofld_init_module);
-module_exit(nvme_tcp_ofld_cleanup_module);
-MODULE_LICENSE("GPL v2");