]> www.infradead.org Git - users/jedix/linux-maple.git/commitdiff
drm/amdgpu/gfx10: remap queue after reset successfully
authorJiadong Zhu <Jiadong.Zhu@amd.com>
Fri, 14 Jun 2024 05:46:36 +0000 (13:46 +0800)
committerAlex Deucher <alexander.deucher@amd.com>
Mon, 2 Sep 2024 15:40:44 +0000 (11:40 -0400)
Kiq command unmap_queues only does the dequeueing action.
We have to map the queue back with clean mqd.

v2: fix up error handling (Alex)

Acked-by: Vitaly Prosyak <vitaly.prosyak@amd.com>
Signed-off-by: Jiadong Zhu <Jiadong.Zhu@amd.com>
Reviewed-by: Alex Deucher <alexander.deucher@amd.com>
Signed-off-by: Alex Deucher <alexander.deucher@amd.com>
drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c

index 7eff395627325a0da96762d10b63498c88a05f68..8038b62feb84288b7ab34a609f953658f3b6135e 100644 (file)
@@ -7030,13 +7030,13 @@ static int gfx_v10_0_kiq_init_queue(struct amdgpu_ring *ring)
        return 0;
 }
 
-static int gfx_v10_0_kcq_init_queue(struct amdgpu_ring *ring)
+static int gfx_v10_0_kcq_init_queue(struct amdgpu_ring *ring, bool restore)
 {
        struct amdgpu_device *adev = ring->adev;
        struct v10_compute_mqd *mqd = ring->mqd_ptr;
        int mqd_idx = ring - &adev->gfx.compute_ring[0];
 
-       if (!amdgpu_in_reset(adev) && !adev->in_suspend) {
+       if (!restore && !amdgpu_in_reset(adev) && !adev->in_suspend) {
                memset((void *)mqd, 0, sizeof(*mqd));
                mutex_lock(&adev->srbm_mutex);
                nv_grbm_select(adev, ring->me, ring->pipe, ring->queue, 0);
@@ -7098,7 +7098,7 @@ static int gfx_v10_0_kcq_resume(struct amdgpu_device *adev)
                        goto done;
                r = amdgpu_bo_kmap(ring->mqd_obj, (void **)&ring->mqd_ptr);
                if (!r) {
-                       r = gfx_v10_0_kcq_init_queue(ring);
+                       r = gfx_v10_0_kcq_init_queue(ring, false);
                        amdgpu_bo_kunmap(ring->mqd_obj);
                        ring->mqd_ptr = NULL;
                }
@@ -9483,25 +9483,49 @@ static int gfx_v10_0_reset_kcq(struct amdgpu_ring *ring,
        spin_lock_irqsave(&kiq->ring_lock, flags);
 
        if (amdgpu_ring_alloc(kiq_ring, kiq->pmf->unmap_queues_size)) {
-               spin_unlock_irqrestore(&kiq->ring_lock, flags);
-               return -ENOMEM;
+               r = -ENOMEM;
+               goto out;
        }
 
        kiq->pmf->kiq_unmap_queues(kiq_ring, ring, RESET_QUEUES,
                                   0, 0);
        amdgpu_ring_commit(kiq_ring);
 
-       spin_unlock_irqrestore(&kiq->ring_lock, flags);
+       r = amdgpu_ring_test_ring(kiq_ring);
+       if (r)
+               goto out;
+
+       r = amdgpu_bo_reserve(ring->mqd_obj, false);
+       if (unlikely(r != 0)) {
+               dev_err(adev->dev, "fail to resv mqd_obj\n");
+               goto out;
+       }
+       r = amdgpu_bo_kmap(ring->mqd_obj, (void **)&ring->mqd_ptr);
+       if (!r) {
+               r = gfx_v10_0_kcq_init_queue(ring, true);
+               amdgpu_bo_kunmap(ring->mqd_obj);
+               ring->mqd_ptr = NULL;
+       }
+       amdgpu_bo_unreserve(ring->mqd_obj);
+       if (r) {
+               dev_err(adev->dev, "fail to unresv mqd_obj\n");
+               goto out;
+       }
+
+       if (amdgpu_ring_alloc(kiq_ring, kiq->pmf->map_queues_size)) {
+               r = -ENOMEM;
+               goto out;
+       }
+       kiq->pmf->kiq_map_queues(kiq_ring, ring);
+       amdgpu_ring_commit(kiq_ring);
 
        r = amdgpu_ring_test_ring(kiq_ring);
+
+out:
+       spin_unlock_irqrestore(&kiq->ring_lock, flags);
        if (r)
                return r;
 
-       /* reset the ring */
-       ring->wptr = 0;
-       *ring->wptr_cpu_addr = 0;
-       amdgpu_ring_clear_ring(ring);
-
        return amdgpu_ring_test_ring(ring);
 }