]> www.infradead.org Git - nvme.git/commitdiff
nvmet: pci-epf: always fully initialize completion entries
authorDamien Le Moal <dlemoal@kernel.org>
Fri, 11 Apr 2025 01:42:09 +0000 (10:42 +0900)
committerChristoph Hellwig <hch@lst.de>
Wed, 16 Apr 2025 05:37:36 +0000 (07:37 +0200)
For a command that is normally processed through the command request
execute() function, the completion entry for the command is initialized
by __nvmet_req_complete() and nvmet_pci_epf_cq_work() only needs to set
the status field and the phase of the completion entry before posting
the entry to the completion queue.

However, for commands that are failed due to an internal error (e.g. the
command data buffer allocation fails), the command request execute()
function is not called and __nvmet_req_complete() is never executed for
the command, leaving the command completion entry uninitialized. For
such command failed before calling req->execute(), the host ends up
seeing completion entries with an invalid submission queue ID and
command ID.

Avoid such issue by always fully initilizing a command completion entry
in nvmet_pci_epf_cq_work(), setting the entry submission queue head, ID
and command ID.

Fixes: 0faa0fe6f90e ("nvmet: New NVMe PCI endpoint function target driver")
Signed-off-by: Damien Le Moal <dlemoal@kernel.org>
Reviewed-by: Niklas Cassel <cassel@kernel.org>
Signed-off-by: Christoph Hellwig <hch@lst.de>
drivers/nvme/target/pci-epf.c

index 51c27b32248d0137fd3a557728eaab72fc7fca0c..43296c05319c35c31d09878df09e92259ccc59af 100644 (file)
@@ -1648,16 +1648,17 @@ static int nvmet_pci_epf_process_sq(struct nvmet_pci_epf_ctrl *ctrl,
 {
        struct nvmet_pci_epf_iod *iod;
        int ret, n = 0;
+       u16 head = sq->head;
 
        sq->tail = nvmet_pci_epf_bar_read32(ctrl, sq->db);
-       while (sq->head != sq->tail && (!ctrl->sq_ab || n < ctrl->sq_ab)) {
+       while (head != sq->tail && (!ctrl->sq_ab || n < ctrl->sq_ab)) {
                iod = nvmet_pci_epf_alloc_iod(sq);
                if (!iod)
                        break;
 
                /* Get the NVMe command submitted by the host. */
                ret = nvmet_pci_epf_transfer(ctrl, &iod->cmd,
-                                            sq->pci_addr + sq->head * sq->qes,
+                                            sq->pci_addr + head * sq->qes,
                                             sq->qes, DMA_FROM_DEVICE);
                if (ret) {
                        /* Not much we can do... */
@@ -1666,12 +1667,13 @@ static int nvmet_pci_epf_process_sq(struct nvmet_pci_epf_ctrl *ctrl,
                }
 
                dev_dbg(ctrl->dev, "SQ[%u]: head %u, tail %u, command %s\n",
-                       sq->qid, sq->head, sq->tail,
+                       sq->qid, head, sq->tail,
                        nvmet_pci_epf_iod_name(iod));
 
-               sq->head++;
-               if (sq->head == sq->depth)
-                       sq->head = 0;
+               head++;
+               if (head == sq->depth)
+                       head = 0;
+               WRITE_ONCE(sq->head, head);
                n++;
 
                queue_work_on(WORK_CPU_UNBOUND, sq->iod_wq, &iod->work);
@@ -1761,8 +1763,17 @@ static void nvmet_pci_epf_cq_work(struct work_struct *work)
                if (!iod)
                        break;
 
-               /* Post the IOD completion entry. */
+               /*
+                * Post the IOD completion entry. If the IOD request was
+                * executed (req->execute() called), the CQE is already
+                * initialized. However, the IOD may have been failed before
+                * that, leaving the CQE not properly initialized. So always
+                * initialize it here.
+                */
                cqe = &iod->cqe;
+               cqe->sq_head = cpu_to_le16(READ_ONCE(iod->sq->head));
+               cqe->sq_id = cpu_to_le16(iod->sq->qid);
+               cqe->command_id = iod->cmd.common.command_id;
                cqe->status = cpu_to_le16((iod->status << 1) | cq->phase);
 
                dev_dbg(ctrl->dev,