]> www.infradead.org Git - users/jedix/linux-maple.git/commitdiff
drm/xe: prevent UAF around preempt fence
authorMatthew Auld <matthew.auld@intel.com>
Wed, 14 Aug 2024 11:01:30 +0000 (12:01 +0100)
committerRodrigo Vivi <rodrigo.vivi@intel.com>
Mon, 19 Aug 2024 14:39:28 +0000 (10:39 -0400)
The fence lock is part of the queue, therefore in the current design
anything locking the fence should then also hold a ref to the queue to
prevent the queue from being freed.

However, currently it looks like we signal the fence and then drop the
queue ref, but if something is waiting on the fence, the waiter is
kicked to wake up at some later point, where upon waking up it first
grabs the lock before checking the fence state. But if we have already
dropped the queue ref, then the lock might already be freed as part of
the queue, leading to uaf.

To prevent this, move the fence lock into the fence itself so we don't
run into lifetime issues. Alternative might be to have device level
lock, or only release the queue in the fence release callback, however
that might require pushing to another worker to avoid locking issues.

Fixes: dd08ebf6c352 ("drm/xe: Introduce a new DRM driver for Intel GPUs")
References: https://gitlab.freedesktop.org/drm/xe/kernel/-/issues/2454
References: https://gitlab.freedesktop.org/drm/xe/kernel/-/issues/2342
References: https://gitlab.freedesktop.org/drm/xe/kernel/-/issues/2020
Signed-off-by: Matthew Auld <matthew.auld@intel.com>
Cc: Matthew Brost <matthew.brost@intel.com>
Cc: <stable@vger.kernel.org> # v6.8+
Reviewed-by: Matthew Brost <matthew.brost@intel.com>
Link: https://patchwork.freedesktop.org/patch/msgid/20240814110129.825847-2-matthew.auld@intel.com
(cherry picked from commit 7116c35aacedc38be6d15bd21b2fc936eed0008b)
Signed-off-by: Rodrigo Vivi <rodrigo.vivi@intel.com>
drivers/gpu/drm/xe/xe_exec_queue.c
drivers/gpu/drm/xe/xe_exec_queue_types.h
drivers/gpu/drm/xe/xe_preempt_fence.c
drivers/gpu/drm/xe/xe_preempt_fence_types.h

index 16f24f4a7062d8cc51fb1bc8c850e9d2f386b958..9731dcd0b1bde9cffdcfac70fe4578bd7e640364 100644 (file)
@@ -643,7 +643,6 @@ int xe_exec_queue_create_ioctl(struct drm_device *dev, void *data,
 
                if (xe_vm_in_preempt_fence_mode(vm)) {
                        q->lr.context = dma_fence_context_alloc(1);
-                       spin_lock_init(&q->lr.lock);
 
                        err = xe_vm_add_compute_exec_queue(vm, q);
                        if (XE_IOCTL_DBG(xe, err))
index a35ce24c97982c880aec69b6dffe3151b3e8ef93..f6ee0ae80fd63dbe63ba461ec60c029bc5534bb5 100644 (file)
@@ -126,8 +126,6 @@ struct xe_exec_queue {
                u32 seqno;
                /** @lr.link: link into VM's list of exec queues */
                struct list_head link;
-               /** @lr.lock: preemption fences lock */
-               spinlock_t lock;
        } lr;
 
        /** @ops: submission backend exec queue operations */
index e8b8ae5c6485e3798d93e3f8aa072e482e850cfb..c453f45328b1c57eac65f32341aee489bc071379 100644 (file)
@@ -128,8 +128,9 @@ xe_preempt_fence_arm(struct xe_preempt_fence *pfence, struct xe_exec_queue *q,
 {
        list_del_init(&pfence->link);
        pfence->q = xe_exec_queue_get(q);
+       spin_lock_init(&pfence->lock);
        dma_fence_init(&pfence->base, &preempt_fence_ops,
-                     &q->lr.lock, context, seqno);
+                     &pfence->lock, context, seqno);
 
        return &pfence->base;
 }
index b54b5c29b5331e7a0c445abbc10551afa76525fd..312c3372a49f902c3b6a38c1fa26aacb78c1b67b 100644 (file)
@@ -25,6 +25,8 @@ struct xe_preempt_fence {
        struct xe_exec_queue *q;
        /** @preempt_work: work struct which issues preemption */
        struct work_struct preempt_work;
+       /** @lock: dma-fence fence lock */
+       spinlock_t lock;
        /** @error: preempt fence is in error state */
        int error;
 };