struct amdgpu_device *adev = ring->adev;
        struct amdgpu_kiq *kiq = &adev->gfx.kiq;
        struct amdgpu_ring *kiq_ring = &kiq->ring;
+       unsigned long flags;
 
        if (!kiq->pmf || !kiq->pmf->kiq_unmap_queues)
                return -EINVAL;
 
-       if (amdgpu_ring_alloc(kiq_ring, kiq->pmf->unmap_queues_size))
+       spin_lock_irqsave(&kiq->ring_lock, flags);
+
+       if (amdgpu_ring_alloc(kiq_ring, kiq->pmf->unmap_queues_size)) {
+               spin_unlock_irqrestore(&kiq->ring_lock, flags);
                return -ENOMEM;
+       }
 
        /* assert preemption condition */
        amdgpu_ring_set_preempt_cond_exec(ring, false);
                                   ++ring->trail_seq);
        amdgpu_ring_commit(kiq_ring);
 
+       spin_unlock_irqrestore(&kiq->ring_lock, flags);
+
        /* poll the trailing fence */
        for (i = 0; i < adev->usec_timeout; i++) {
                if (ring->trail_seq ==