]> www.infradead.org Git - users/hch/configfs.git/commitdiff
drm/amdgpu/mes: add multiple mes ring instances support
authorJack Xiao <Jack.Xiao@amd.com>
Wed, 7 Aug 2024 03:53:35 +0000 (11:53 +0800)
committerAlex Deucher <alexander.deucher@amd.com>
Tue, 13 Aug 2024 14:29:25 +0000 (10:29 -0400)
Add multiple mes ring instances in mes structure to support
multiple mes pipes.

Signed-off-by: Jack Xiao <Jack.Xiao@amd.com>
Acked-by: Alex Deucher <alexander.deucher@amd.com>
Reviewed-by: Hawking Zhang <Hawking.Zhang@amd.com>
Signed-off-by: Alex Deucher <alexander.deucher@amd.com>
drivers/gpu/drm/amd/amdgpu/amdgpu_gfx.c
drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.c
drivers/gpu/drm/amd/amdgpu/amdgpu_mes.c
drivers/gpu/drm/amd/amdgpu/amdgpu_mes.h
drivers/gpu/drm/amd/amdgpu/amdgpu_virt.c
drivers/gpu/drm/amd/amdgpu/gmc_v11_0.c
drivers/gpu/drm/amd/amdgpu/gmc_v12_0.c
drivers/gpu/drm/amd/amdgpu/mes_v11_0.c
drivers/gpu/drm/amd/amdgpu/mes_v12_0.c

index 5c9f36f01db0f4c135d1980ee2c1dffb681592bf..28bd2098a65e4e01c71a0305e937f934ae113879 100644 (file)
@@ -998,7 +998,7 @@ uint32_t amdgpu_kiq_rreg(struct amdgpu_device *adev, uint32_t reg, uint32_t xcc_
        if (amdgpu_device_skip_hw_access(adev))
                return 0;
 
-       if (adev->mes.ring.sched.ready)
+       if (adev->mes.ring[0].sched.ready)
                return amdgpu_mes_rreg(adev, reg);
 
        BUG_ON(!ring->funcs->emit_rreg);
@@ -1071,7 +1071,7 @@ void amdgpu_kiq_wreg(struct amdgpu_device *adev, uint32_t reg, uint32_t v, uint3
        if (amdgpu_device_skip_hw_access(adev))
                return;
 
-       if (adev->mes.ring.sched.ready) {
+       if (adev->mes.ring[0].sched.ready) {
                amdgpu_mes_wreg(adev, reg, v);
                return;
        }
index c0265902565670e98d28b8f015380840b06eb703..b49b3650fd62173b15c03d38889e63949b5b5bf5 100644 (file)
@@ -589,7 +589,8 @@ int amdgpu_gmc_allocate_vm_inv_eng(struct amdgpu_device *adev)
                ring = adev->rings[i];
                vmhub = ring->vm_hub;
 
-               if (ring == &adev->mes.ring ||
+               if (ring == &adev->mes.ring[0] ||
+                   ring == &adev->mes.ring[1] ||
                    ring == &adev->umsch_mm.ring)
                        continue;
 
@@ -761,7 +762,7 @@ void amdgpu_gmc_fw_reg_write_reg_wait(struct amdgpu_device *adev,
        unsigned long flags;
        uint32_t seq;
 
-       if (adev->mes.ring.sched.ready) {
+       if (adev->mes.ring[0].sched.ready) {
                amdgpu_mes_reg_write_reg_wait(adev, reg0, reg1,
                                              ref, mask);
                return;
index b2a9df20291347018775b4cc0cde27fd5f1ef878..be2156bf025226ef2483311c36166e8a2984f38d 100644 (file)
@@ -135,9 +135,11 @@ int amdgpu_mes_init(struct amdgpu_device *adev)
        idr_init(&adev->mes.queue_id_idr);
        ida_init(&adev->mes.doorbell_ida);
        spin_lock_init(&adev->mes.queue_id_lock);
-       spin_lock_init(&adev->mes.ring_lock);
        mutex_init(&adev->mes.mutex_hidden);
 
+       for (i = 0; i < AMDGPU_MAX_MES_PIPES; i++)
+               spin_lock_init(&adev->mes.ring_lock[i]);
+
        adev->mes.total_max_queue = AMDGPU_FENCE_MES_QUEUE_ID_MASK;
        adev->mes.vmid_mask_mmhub = 0xffffff00;
        adev->mes.vmid_mask_gfxhub = 0xffffff00;
index 174283a0fc07e9485733b27097ac831220430714..d87d068952e01275283e890d19ea2bc9aab5d7fd 100644 (file)
@@ -82,8 +82,8 @@ struct amdgpu_mes {
        uint64_t                        default_process_quantum;
        uint64_t                        default_gang_quantum;
 
-       struct amdgpu_ring              ring;
-       spinlock_t                      ring_lock;
+       struct amdgpu_ring              ring[AMDGPU_MAX_MES_PIPES];
+       spinlock_t                      ring_lock[AMDGPU_MAX_MES_PIPES];
 
        const struct firmware           *fw[AMDGPU_MAX_MES_PIPES];
 
index 111c380f929b597d65f32bc3f3a0cc62973a8dac..b287a82e6177e9b3614595ada9eba8f721fc1363 100644 (file)
@@ -858,7 +858,7 @@ void amdgpu_virt_post_reset(struct amdgpu_device *adev)
                adev->gfx.is_poweron = false;
        }
 
-       adev->mes.ring.sched.ready = false;
+       adev->mes.ring[0].sched.ready = false;
 }
 
 bool amdgpu_virt_fw_load_skip_check(struct amdgpu_device *adev, uint32_t ucode_id)
index b88a6fa173b3628904ea1f8778bb9b0b18589b76..2797fd84432b22d6ded588a047f40219187a282a 100644 (file)
@@ -231,7 +231,7 @@ static void gmc_v11_0_flush_gpu_tlb(struct amdgpu_device *adev, uint32_t vmid,
        /* This is necessary for SRIOV as well as for GFXOFF to function
         * properly under bare metal
         */
-       if ((adev->gfx.kiq[0].ring.sched.ready || adev->mes.ring.sched.ready) &&
+       if ((adev->gfx.kiq[0].ring.sched.ready || adev->mes.ring[0].sched.ready) &&
            (amdgpu_sriov_runtime(adev) || !amdgpu_sriov_vf(adev))) {
                amdgpu_gmc_fw_reg_write_reg_wait(adev, req, ack, inv_req,
                                                 1 << vmid, GET_INST(GC, 0));
index 26efce9aa4109f30b08798f55aece979114e9673..edcb5351f8cca7ce8f302cacd04189c12bf1cf1a 100644 (file)
@@ -299,7 +299,7 @@ static void gmc_v12_0_flush_gpu_tlb(struct amdgpu_device *adev, uint32_t vmid,
        /* This is necessary for SRIOV as well as for GFXOFF to function
         * properly under bare metal
         */
-       if ((adev->gfx.kiq[0].ring.sched.ready || adev->mes.ring.sched.ready) &&
+       if ((adev->gfx.kiq[0].ring.sched.ready || adev->mes.ring[0].sched.ready) &&
            (amdgpu_sriov_runtime(adev) || !amdgpu_sriov_vf(adev))) {
                struct amdgpu_vmhub *hub = &adev->vmhub[vmhub];
                const unsigned eng = 17;
index 61b8cb39826d04c524bba691461c265283464014..4c7899e527fed4b16c40247c12c9d3e3107c8668 100644 (file)
@@ -162,7 +162,7 @@ static int mes_v11_0_submit_pkt_and_poll_completion(struct amdgpu_mes *mes,
        union MESAPI__QUERY_MES_STATUS mes_status_pkt;
        signed long timeout = 3000000; /* 3000 ms */
        struct amdgpu_device *adev = mes->adev;
-       struct amdgpu_ring *ring = &mes->ring;
+       struct amdgpu_ring *ring = &mes->ring[0];
        struct MES_API_STATUS *api_status;
        union MESAPI__MISC *x_pkt = pkt;
        const char *op_str, *misc_op_str;
@@ -191,7 +191,7 @@ static int mes_v11_0_submit_pkt_and_poll_completion(struct amdgpu_mes *mes,
        status_ptr = (u64 *)&adev->wb.wb[status_offset];
        *status_ptr = 0;
 
-       spin_lock_irqsave(&mes->ring_lock, flags);
+       spin_lock_irqsave(&mes->ring_lock[0], flags);
        r = amdgpu_ring_alloc(ring, (size + sizeof(mes_status_pkt)) / 4);
        if (r)
                goto error_unlock_free;
@@ -221,7 +221,7 @@ static int mes_v11_0_submit_pkt_and_poll_completion(struct amdgpu_mes *mes,
                                   sizeof(mes_status_pkt) / 4);
 
        amdgpu_ring_commit(ring);
-       spin_unlock_irqrestore(&mes->ring_lock, flags);
+       spin_unlock_irqrestore(&mes->ring_lock[0], flags);
 
        op_str = mes_v11_0_get_op_string(x_pkt);
        misc_op_str = mes_v11_0_get_misc_op_string(x_pkt);
@@ -263,7 +263,7 @@ error_undo:
        amdgpu_ring_undo(ring);
 
 error_unlock_free:
-       spin_unlock_irqrestore(&mes->ring_lock, flags);
+       spin_unlock_irqrestore(&mes->ring_lock[0], flags);
 
 error_wb_free:
        amdgpu_device_wb_free(adev, status_offset);
@@ -1058,7 +1058,7 @@ static int mes_v11_0_kiq_enable_queue(struct amdgpu_device *adev)
                return r;
        }
 
-       kiq->pmf->kiq_map_queues(kiq_ring, &adev->mes.ring);
+       kiq->pmf->kiq_map_queues(kiq_ring, &adev->mes.ring[0]);
 
        return amdgpu_ring_test_helper(kiq_ring);
 }
@@ -1072,7 +1072,7 @@ static int mes_v11_0_queue_init(struct amdgpu_device *adev,
        if (pipe == AMDGPU_MES_KIQ_PIPE)
                ring = &adev->gfx.kiq[0].ring;
        else if (pipe == AMDGPU_MES_SCHED_PIPE)
-               ring = &adev->mes.ring;
+               ring = &adev->mes.ring[0];
        else
                BUG();
 
@@ -1114,7 +1114,7 @@ static int mes_v11_0_ring_init(struct amdgpu_device *adev)
 {
        struct amdgpu_ring *ring;
 
-       ring = &adev->mes.ring;
+       ring = &adev->mes.ring[0];
 
        ring->funcs = &mes_v11_0_ring_funcs;
 
@@ -1167,7 +1167,7 @@ static int mes_v11_0_mqd_sw_init(struct amdgpu_device *adev,
        if (pipe == AMDGPU_MES_KIQ_PIPE)
                ring = &adev->gfx.kiq[0].ring;
        else if (pipe == AMDGPU_MES_SCHED_PIPE)
-               ring = &adev->mes.ring;
+               ring = &adev->mes.ring[0];
        else
                BUG();
 
@@ -1259,12 +1259,12 @@ static int mes_v11_0_sw_fini(void *handle)
                              &adev->gfx.kiq[0].ring.mqd_gpu_addr,
                              &adev->gfx.kiq[0].ring.mqd_ptr);
 
-       amdgpu_bo_free_kernel(&adev->mes.ring.mqd_obj,
-                             &adev->mes.ring.mqd_gpu_addr,
-                             &adev->mes.ring.mqd_ptr);
+       amdgpu_bo_free_kernel(&adev->mes.ring[0].mqd_obj,
+                             &adev->mes.ring[0].mqd_gpu_addr,
+                             &adev->mes.ring[0].mqd_ptr);
 
        amdgpu_ring_fini(&adev->gfx.kiq[0].ring);
-       amdgpu_ring_fini(&adev->mes.ring);
+       amdgpu_ring_fini(&adev->mes.ring[0]);
 
        if (adev->firmware.load_type == AMDGPU_FW_LOAD_DIRECT) {
                mes_v11_0_free_ucode_buffers(adev, AMDGPU_MES_KIQ_PIPE);
@@ -1375,9 +1375,9 @@ failure:
 
 static int mes_v11_0_kiq_hw_fini(struct amdgpu_device *adev)
 {
-       if (adev->mes.ring.sched.ready) {
-               mes_v11_0_kiq_dequeue(&adev->mes.ring);
-               adev->mes.ring.sched.ready = false;
+       if (adev->mes.ring[0].sched.ready) {
+               mes_v11_0_kiq_dequeue(&adev->mes.ring[0]);
+               adev->mes.ring[0].sched.ready = false;
        }
 
        if (amdgpu_sriov_vf(adev)) {
@@ -1395,7 +1395,7 @@ static int mes_v11_0_hw_init(void *handle)
        int r;
        struct amdgpu_device *adev = (struct amdgpu_device *)handle;
 
-       if (adev->mes.ring.sched.ready)
+       if (adev->mes.ring[0].sched.ready)
                goto out;
 
        if (!adev->enable_mes_kiq) {
@@ -1440,7 +1440,7 @@ out:
         * with MES enabled.
         */
        adev->gfx.kiq[0].ring.sched.ready = false;
-       adev->mes.ring.sched.ready = true;
+       adev->mes.ring[0].sched.ready = true;
 
        return 0;
 
index 5e06a982eb543e6b864f04b9167063191f2f1ed5..ac6209a0029c64f9ca7cadc88c6c10ec3684053c 100644 (file)
@@ -148,7 +148,7 @@ static int mes_v12_0_submit_pkt_and_poll_completion(struct amdgpu_mes *mes,
        union MESAPI__QUERY_MES_STATUS mes_status_pkt;
        signed long timeout = 3000000; /* 3000 ms */
        struct amdgpu_device *adev = mes->adev;
-       struct amdgpu_ring *ring = &mes->ring;
+       struct amdgpu_ring *ring = &mes->ring[0];
        struct MES_API_STATUS *api_status;
        union MESAPI__MISC *x_pkt = pkt;
        const char *op_str, *misc_op_str;
@@ -177,7 +177,7 @@ static int mes_v12_0_submit_pkt_and_poll_completion(struct amdgpu_mes *mes,
        status_ptr = (u64 *)&adev->wb.wb[status_offset];
        *status_ptr = 0;
 
-       spin_lock_irqsave(&mes->ring_lock, flags);
+       spin_lock_irqsave(&mes->ring_lock[0], flags);
        r = amdgpu_ring_alloc(ring, (size + sizeof(mes_status_pkt)) / 4);
        if (r)
                goto error_unlock_free;
@@ -207,7 +207,7 @@ static int mes_v12_0_submit_pkt_and_poll_completion(struct amdgpu_mes *mes,
                                   sizeof(mes_status_pkt) / 4);
 
        amdgpu_ring_commit(ring);
-       spin_unlock_irqrestore(&mes->ring_lock, flags);
+       spin_unlock_irqrestore(&mes->ring_lock[0], flags);
 
        op_str = mes_v12_0_get_op_string(x_pkt);
        misc_op_str = mes_v12_0_get_misc_op_string(x_pkt);
@@ -249,7 +249,7 @@ error_undo:
        amdgpu_ring_undo(ring);
 
 error_unlock_free:
-       spin_unlock_irqrestore(&mes->ring_lock, flags);
+       spin_unlock_irqrestore(&mes->ring_lock[0], flags);
 
 error_wb_free:
        amdgpu_device_wb_free(adev, status_offset);
@@ -1128,7 +1128,7 @@ static int mes_v12_0_kiq_enable_queue(struct amdgpu_device *adev)
                return r;
        }
 
-       kiq->pmf->kiq_map_queues(kiq_ring, &adev->mes.ring);
+       kiq->pmf->kiq_map_queues(kiq_ring, &adev->mes.ring[0]);
 
        r = amdgpu_ring_test_ring(kiq_ring);
        if (r) {
@@ -1147,7 +1147,7 @@ static int mes_v12_0_queue_init(struct amdgpu_device *adev,
        if (pipe == AMDGPU_MES_KIQ_PIPE)
                ring = &adev->gfx.kiq[0].ring;
        else if (pipe == AMDGPU_MES_SCHED_PIPE)
-               ring = &adev->mes.ring;
+               ring = &adev->mes.ring[0];
        else
                BUG();
 
@@ -1193,7 +1193,7 @@ static int mes_v12_0_ring_init(struct amdgpu_device *adev)
 {
        struct amdgpu_ring *ring;
 
-       ring = &adev->mes.ring;
+       ring = &adev->mes.ring[0];
 
        ring->funcs = &mes_v12_0_ring_funcs;
 
@@ -1246,7 +1246,7 @@ static int mes_v12_0_mqd_sw_init(struct amdgpu_device *adev,
        if (pipe == AMDGPU_MES_KIQ_PIPE)
                ring = &adev->gfx.kiq[0].ring;
        else if (pipe == AMDGPU_MES_SCHED_PIPE)
-               ring = &adev->mes.ring;
+               ring = &adev->mes.ring[0];
        else
                BUG();
 
@@ -1335,12 +1335,12 @@ static int mes_v12_0_sw_fini(void *handle)
                              &adev->gfx.kiq[0].ring.mqd_gpu_addr,
                              &adev->gfx.kiq[0].ring.mqd_ptr);
 
-       amdgpu_bo_free_kernel(&adev->mes.ring.mqd_obj,
-                             &adev->mes.ring.mqd_gpu_addr,
-                             &adev->mes.ring.mqd_ptr);
+       amdgpu_bo_free_kernel(&adev->mes.ring[0].mqd_obj,
+                             &adev->mes.ring[0].mqd_gpu_addr,
+                             &adev->mes.ring[0].mqd_ptr);
 
        amdgpu_ring_fini(&adev->gfx.kiq[0].ring);
-       amdgpu_ring_fini(&adev->mes.ring);
+       amdgpu_ring_fini(&adev->mes.ring[0]);
 
        if (adev->firmware.load_type == AMDGPU_FW_LOAD_DIRECT) {
                mes_v12_0_free_ucode_buffers(adev, AMDGPU_MES_KIQ_PIPE);
@@ -1384,7 +1384,7 @@ static void mes_v12_0_kiq_dequeue_sched(struct amdgpu_device *adev)
        soc21_grbm_select(adev, 0, 0, 0, 0);
        mutex_unlock(&adev->srbm_mutex);
 
-       adev->mes.ring.sched.ready = false;
+       adev->mes.ring[0].sched.ready = false;
 }
 
 static void mes_v12_0_kiq_setting(struct amdgpu_ring *ring)
@@ -1448,9 +1448,9 @@ failure:
 
 static int mes_v12_0_kiq_hw_fini(struct amdgpu_device *adev)
 {
-       if (adev->mes.ring.sched.ready) {
+       if (adev->mes.ring[0].sched.ready) {
                mes_v12_0_kiq_dequeue_sched(adev);
-               adev->mes.ring.sched.ready = false;
+               adev->mes.ring[0].sched.ready = false;
        }
 
        mes_v12_0_enable(adev, false);
@@ -1463,7 +1463,7 @@ static int mes_v12_0_hw_init(void *handle)
        int r;
        struct amdgpu_device *adev = (struct amdgpu_device *)handle;
 
-       if (adev->mes.ring.sched.ready)
+       if (adev->mes.ring[0].sched.ready)
                goto out;
 
        if (!adev->enable_mes_kiq || adev->enable_uni_mes) {
@@ -1515,7 +1515,7 @@ out:
         * with MES enabled.
         */
        adev->gfx.kiq[0].ring.sched.ready = false;
-       adev->mes.ring.sched.ready = true;
+       adev->mes.ring[0].sched.ready = true;
 
        return 0;