]> www.infradead.org Git - users/jedix/linux-maple.git/commitdiff
drm/xe/vf: Rebase exec queue parallel commands during migration recovery
authorTomasz Lis <tomasz.lis@intel.com>
Sat, 2 Aug 2025 03:10:45 +0000 (05:10 +0200)
committerMichał Winiarski <michal.winiarski@intel.com>
Mon, 4 Aug 2025 14:47:12 +0000 (16:47 +0200)
Parallel exec queues have an additional command streamer buffer which holds
a GGTT reference to data within context status. The GGTT references have to
be fixed after VF migration.

v2: Properly handle nop entry, verify if parsing goes ok
v3: Improve error/warn logging, add propagation of errors,
 give names to magic offsets

Signed-off-by: Tomasz Lis <tomasz.lis@intel.com>
Cc: Michal Winiarski <michal.winiarski@intel.com>
Reviewed-by: Michał Winiarski <michal.winiarski@intel.com>
Link: https://lore.kernel.org/r/20250802031045.1127138-9-tomasz.lis@intel.com
Signed-off-by: Michał Winiarski <michal.winiarski@intel.com>
drivers/gpu/drm/xe/abi/guc_actions_abi.h
drivers/gpu/drm/xe/xe_guc_submit.c

index 81eb046aeebfef4fe684e2481bd857b5b23010a3..d8cf68a0516d5bf0c4a94fb1b4f06b89193fab2f 100644 (file)
@@ -193,6 +193,14 @@ enum xe_guc_register_context_multi_lrc_param_offsets {
        XE_GUC_REGISTER_CONTEXT_MULTI_LRC_MSG_MIN_LEN = 11,
 };
 
+enum xe_guc_context_wq_item_offsets {
+       XE_GUC_CONTEXT_WQ_HEADER_DATA_0_TYPE_LEN = 0,
+       XE_GUC_CONTEXT_WQ_EL_INFO_DATA_1_CTX_DESC_LOW,
+       XE_GUC_CONTEXT_WQ_EL_INFO_DATA_2_GUCCTX_RINGTAIL_FREEZEPOCS,
+       XE_GUC_CONTEXT_WQ_EL_INFO_DATA_3_WI_FENCE_ID,
+       XE_GUC_CONTEXT_WQ_EL_CHILD_LIST_DATA_4_RINGTAIL,
+};
+
 enum xe_guc_report_status {
        XE_GUC_REPORT_STATUS_UNKNOWN = 0x0,
        XE_GUC_REPORT_STATUS_ACKED = 0x1,
index 20f5b982026474c069a2b870fc147537f4c2bfd1..554a25ca77d34af7912055699b1a5d9edaee5c55 100644 (file)
@@ -671,12 +671,18 @@ static void wq_item_append(struct xe_exec_queue *q)
        if (wq_wait_for_space(q, wqi_size))
                return;
 
+       xe_gt_assert(guc_to_gt(guc), i == XE_GUC_CONTEXT_WQ_HEADER_DATA_0_TYPE_LEN);
        wqi[i++] = FIELD_PREP(WQ_TYPE_MASK, WQ_TYPE_MULTI_LRC) |
                FIELD_PREP(WQ_LEN_MASK, len_dw);
+       xe_gt_assert(guc_to_gt(guc), i == XE_GUC_CONTEXT_WQ_EL_INFO_DATA_1_CTX_DESC_LOW);
        wqi[i++] = xe_lrc_descriptor(q->lrc[0]);
+       xe_gt_assert(guc_to_gt(guc), i ==
+                    XE_GUC_CONTEXT_WQ_EL_INFO_DATA_2_GUCCTX_RINGTAIL_FREEZEPOCS);
        wqi[i++] = FIELD_PREP(WQ_GUC_ID_MASK, q->guc->id) |
                FIELD_PREP(WQ_RING_TAIL_MASK, q->lrc[0]->ring.tail / sizeof(u64));
+       xe_gt_assert(guc_to_gt(guc), i == XE_GUC_CONTEXT_WQ_EL_INFO_DATA_3_WI_FENCE_ID);
        wqi[i++] = 0;
+       xe_gt_assert(guc_to_gt(guc), i == XE_GUC_CONTEXT_WQ_EL_CHILD_LIST_DATA_4_RINGTAIL);
        for (j = 1; j < q->width; ++j) {
                struct xe_lrc *lrc = q->lrc[j];
 
@@ -697,6 +703,50 @@ static void wq_item_append(struct xe_exec_queue *q)
        parallel_write(xe, map, wq_desc.tail, q->guc->wqi_tail);
 }
 
+static int wq_items_rebase(struct xe_exec_queue *q)
+{
+       struct xe_guc *guc = exec_queue_to_guc(q);
+       struct xe_device *xe = guc_to_xe(guc);
+       struct iosys_map map = xe_lrc_parallel_map(q->lrc[0]);
+       int i = q->guc->wqi_head;
+
+       /* the ring starts after a header struct */
+       iosys_map_incr(&map, offsetof(struct guc_submit_parallel_scratch, wq[0]));
+
+       while ((i % WQ_SIZE) != (q->guc->wqi_tail % WQ_SIZE)) {
+               u32 len_dw, type, val;
+
+               if (drm_WARN_ON_ONCE(&xe->drm, i < 0 || i > 2 * WQ_SIZE))
+                       break;
+
+               val = xe_map_rd_ring_u32(xe, &map, i / sizeof(u32) +
+                                        XE_GUC_CONTEXT_WQ_HEADER_DATA_0_TYPE_LEN,
+                                        WQ_SIZE / sizeof(u32));
+               len_dw = FIELD_GET(WQ_LEN_MASK, val);
+               type = FIELD_GET(WQ_TYPE_MASK, val);
+
+               if (drm_WARN_ON_ONCE(&xe->drm, len_dw >= WQ_SIZE / sizeof(u32)))
+                       break;
+
+               if (type == WQ_TYPE_MULTI_LRC) {
+                       val = xe_lrc_descriptor(q->lrc[0]);
+                       xe_map_wr_ring_u32(xe, &map, i / sizeof(u32) +
+                                          XE_GUC_CONTEXT_WQ_EL_INFO_DATA_1_CTX_DESC_LOW,
+                                          WQ_SIZE / sizeof(u32), val);
+               } else if (drm_WARN_ON_ONCE(&xe->drm, type != WQ_TYPE_NOOP)) {
+                       break;
+               }
+
+               i += (len_dw + 1) * sizeof(u32);
+       }
+
+       if ((i % WQ_SIZE) != (q->guc->wqi_tail % WQ_SIZE)) {
+               xe_gt_err(q->gt, "Exec queue fixups incomplete - wqi parse failed\n");
+               return -EBADMSG;
+       }
+       return 0;
+}
+
 #define RESUME_PENDING ~0x0ull
 static void submit_exec_queue(struct xe_exec_queue *q)
 {
@@ -2549,6 +2599,10 @@ int xe_guc_contexts_hwsp_rebase(struct xe_guc *guc, void *scratch)
                err = xe_exec_queue_contexts_hwsp_rebase(q, scratch);
                if (err)
                        break;
+               if (xe_exec_queue_is_parallel(q))
+                       err = wq_items_rebase(q);
+               if (err)
+                       break;
        }
        mutex_unlock(&guc->submission_state.lock);