While HBA reset is in progress, if IO/TM completion is received for the
same IO then IO/TM completion path releases the driver private resources
associated with IO. This fix prevents releasing the resources in
IO and TM completion path if HBA reset is in progress.
Signed-off-by: Narsimhulu Musini <nmusini@cisco.com>
Signed-off-by: Sesidhar Baddela <sebaddel@cisco.com>
Signed-off-by: Martin K. Petersen <martin.petersen@oracle.com>
 #define SNIC_DEV_RST_NOTSUP            BIT(25)
 #define SNIC_SCSI_CLEANUP              BIT(26)
 #define SNIC_HOST_RESET_ISSUED         BIT(27)
+#define SNIC_HOST_RESET_CMD_TERM       \
+       (SNIC_DEV_RST_NOTSUP | SNIC_SCSI_CLEANUP | SNIC_HOST_RESET_ISSUED)
 
 #define SNIC_ABTS_TIMEOUT              30000           /* msec */
 #define SNIC_LUN_RESET_TIMEOUT         30000           /* msec */
        SNIC_MSIX_INTR_MAX,
 };
 
+#define SNIC_INTRHDLR_NAMSZ    (2 * IFNAMSIZ)
 struct snic_msix_entry {
        int requested;
-       char devname[IFNAMSIZ];
+       char devname[SNIC_INTRHDLR_NAMSZ];
        irqreturn_t (*isr)(int, void *);
        void *devid;
 };
 
                      sc->device->lun, sc, sc->cmnd[0], snic_cmd_tag(sc),
                      CMD_FLAGS(sc), rqi);
 
+       if (CMD_FLAGS(sc) & SNIC_HOST_RESET_CMD_TERM) {
+               spin_unlock_irqrestore(io_lock, flags);
+
+               return;
+       }
+
        SNIC_BUG_ON(rqi != (struct snic_req_info *)ctx);
        WARN_ON_ONCE(req);
        if (!rqi) {
 
        io_lock = snic_io_lock_hash(snic, sc);
        spin_lock_irqsave(io_lock, flags);
+       if (CMD_FLAGS(sc) & SNIC_HOST_RESET_CMD_TERM) {
+               spin_unlock_irqrestore(io_lock, flags);
+
+               return ret;
+       }
        rqi = (struct snic_req_info *) CMD_SP(sc);
        WARN_ON_ONCE(!rqi);