struct list_head *done_q)
 {
        struct scsi_cmnd *scmd, *next;
+       struct scsi_device *sdev;
 
        list_for_each_entry_safe(scmd, next, work_q, eh_entry) {
                sdev_printk(KERN_INFO, scmd->device, "Device offlined - "
                            "not ready after error recovery\n");
-               scsi_device_set_state(scmd->device, SDEV_OFFLINE);
+               sdev = scmd->device;
+
+               mutex_lock(&sdev->state_mutex);
+               scsi_device_set_state(sdev, SDEV_OFFLINE);
+               mutex_unlock(&sdev->state_mutex);
+
                scsi_eh_finish_cmd(scmd, done_q);
        }
        return;
 
 int
 scsi_device_quiesce(struct scsi_device *sdev)
 {
-       int err = scsi_device_set_state(sdev, SDEV_QUIESCE);
+       int err;
+
+       mutex_lock(&sdev->state_mutex);
+       err = scsi_device_set_state(sdev, SDEV_QUIESCE);
+       mutex_unlock(&sdev->state_mutex);
+
        if (err)
                return err;
 
         * so assume the state is being managed elsewhere (for example
         * device deleted during suspend)
         */
-       if (sdev->sdev_state != SDEV_QUIESCE ||
-           scsi_device_set_state(sdev, SDEV_RUNNING))
-               return;
-       scsi_run_queue(sdev->request_queue);
+       mutex_lock(&sdev->state_mutex);
+       if (sdev->sdev_state == SDEV_QUIESCE &&
+           scsi_device_set_state(sdev, SDEV_RUNNING) == 0)
+               scsi_run_queue(sdev->request_queue);
+       mutex_unlock(&sdev->state_mutex);
 }
 EXPORT_SYMBOL(scsi_device_resume);
 
        struct request_queue *q = sdev->request_queue;
        int err;
 
+       mutex_lock(&sdev->state_mutex);
        err = scsi_internal_device_block_nowait(sdev);
        if (err == 0) {
                if (q->mq_ops)
                else
                        scsi_wait_for_queuecommand(sdev);
        }
+       mutex_unlock(&sdev->state_mutex);
+
        return err;
 }
  
 static int scsi_internal_device_unblock(struct scsi_device *sdev,
                                        enum scsi_device_state new_state)
 {
-       return scsi_internal_device_unblock_nowait(sdev, new_state);
+       int ret;
+
+       mutex_lock(&sdev->state_mutex);
+       ret = scsi_internal_device_unblock_nowait(sdev, new_state);
+       mutex_unlock(&sdev->state_mutex);
+
+       return ret;
 }
 
 static void
 
        sdev->id = starget->id;
        sdev->lun = lun;
        sdev->channel = starget->channel;
+       mutex_init(&sdev->state_mutex);
        sdev->sdev_state = SDEV_CREATED;
        INIT_LIST_HEAD(&sdev->siblings);
        INIT_LIST_HEAD(&sdev->same_target_siblings);
 
        /* set the device running here so that slave configure
         * may do I/O */
+       mutex_lock(&sdev->state_mutex);
        ret = scsi_device_set_state(sdev, SDEV_RUNNING);
-       if (ret) {
+       if (ret)
                ret = scsi_device_set_state(sdev, SDEV_BLOCK);
+       mutex_unlock(&sdev->state_mutex);
 
-               if (ret) {
-                       sdev_printk(KERN_ERR, sdev,
-                                   "in wrong state %s to complete scan\n",
-                                   scsi_device_state_name(sdev->sdev_state));
-                       return SCSI_SCAN_NO_RESPONSE;
-               }
+       if (ret) {
+               sdev_printk(KERN_ERR, sdev,
+                           "in wrong state %s to complete scan\n",
+                           scsi_device_state_name(sdev->sdev_state));
+               return SCSI_SCAN_NO_RESPONSE;
        }
 
        if (*bflags & BLIST_MS_192_BYTES_FOR_3F)
 
 store_state_field(struct device *dev, struct device_attribute *attr,
                  const char *buf, size_t count)
 {
-       int i;
+       int i, ret;
        struct scsi_device *sdev = to_scsi_device(dev);
        enum scsi_device_state state = 0;
 
        if (!state)
                return -EINVAL;
 
-       if (scsi_device_set_state(sdev, state))
-               return -EINVAL;
-       return count;
+       mutex_lock(&sdev->state_mutex);
+       ret = scsi_device_set_state(sdev, state);
+       mutex_unlock(&sdev->state_mutex);
+
+       return ret == 0 ? count : -EINVAL;
 }
 
 static ssize_t
 void __scsi_remove_device(struct scsi_device *sdev)
 {
        struct device *dev = &sdev->sdev_gendev;
+       int res;
 
        /*
         * This cleanup path is not reentrant and while it is impossible
                return;
 
        if (sdev->is_visible) {
-               if (scsi_device_set_state(sdev, SDEV_CANCEL) != 0)
+               /*
+                * If scsi_internal_target_block() is running concurrently,
+                * wait until it has finished before changing the device state.
+                */
+               mutex_lock(&sdev->state_mutex);
+               res = scsi_device_set_state(sdev, SDEV_CANCEL);
+               mutex_unlock(&sdev->state_mutex);
+
+               if (res != 0)
                        return;
 
                bsg_unregister_queue(sdev->request_queue);
         * scsi_run_queue() invocations have finished before tearing down the
         * device.
         */
+       mutex_lock(&sdev->state_mutex);
        scsi_device_set_state(sdev, SDEV_DEL);
+       mutex_unlock(&sdev->state_mutex);
+
        blk_cleanup_queue(sdev->request_queue);
        cancel_work_sync(&sdev->requeue_work);
 
 
                 * invoking scsi_target_unblock() won't change the state of
                 * these devices into running so do that explicitly.
                 */
-               spin_lock_irq(shost->host_lock);
-               __shost_for_each_device(sdev, shost)
+               shost_for_each_device(sdev, shost) {
+                       mutex_lock(&sdev->state_mutex);
                        if (sdev->sdev_state == SDEV_OFFLINE)
                                sdev->sdev_state = SDEV_RUNNING;
-               spin_unlock_irq(shost->host_lock);
+                       mutex_unlock(&sdev->state_mutex);
+               }
        } else if (rport->state == SRP_RPORT_RUNNING) {
                /*
                 * srp_reconnect_rport() has been invoked with fast_io_fail
 
 static int sd_eh_action(struct scsi_cmnd *scmd, int eh_disp)
 {
        struct scsi_disk *sdkp = scsi_disk(scmd->request->rq_disk);
+       struct scsi_device *sdev = scmd->device;
 
-       if (!scsi_device_online(scmd->device) ||
+       if (!scsi_device_online(sdev) ||
            !scsi_medium_access_command(scmd) ||
            host_byte(scmd->result) != DID_TIME_OUT ||
            eh_disp != SUCCESS)
        if (sdkp->medium_access_timed_out >= sdkp->max_medium_access_timeouts) {
                scmd_printk(KERN_ERR, scmd,
                            "Medium access timeout failure. Offlining disk!\n");
-               scsi_device_set_state(scmd->device, SDEV_OFFLINE);
+               mutex_lock(&sdev->state_mutex);
+               scsi_device_set_state(sdev, SDEV_OFFLINE);
+               mutex_unlock(&sdev->state_mutex);
 
                return SUCCESS;
        }
 
        void                    *handler_data;
 
        unsigned char           access_state;
+       struct mutex            state_mutex;
        enum scsi_device_state sdev_state;
        unsigned long           sdev_data[0];
 } __attribute__((aligned(sizeof(unsigned long))));