Commit de8548813824 ("drm/panthor: Add the scheduler logical block")
handled destruction of a group's queues' drm scheduler entities early
into the group destruction procedure.

However, that races with the group submit ioctl, because by the time
entities are destroyed (through the group destroy ioctl), the submission
procedure might've already obtained a group handle, and therefore the
ability to push jobs into entities. This is met with a DRM error message
within the drm scheduler core as a situation that should never occur.

Fix by deferring drm scheduler entity destruction to queue release time.

Fixes: de8548813824 ("drm/panthor: Add the scheduler logical block")
Signed-off-by: Adrián Larumbe <[email protected]>
---
 drivers/gpu/drm/panthor/panthor_sched.c | 8 +-------
 1 file changed, 1 insertion(+), 7 deletions(-)

diff --git a/drivers/gpu/drm/panthor/panthor_sched.c 
b/drivers/gpu/drm/panthor/panthor_sched.c
index 0cc9055f4ee5..f5e01cb16cfc 100644
--- a/drivers/gpu/drm/panthor/panthor_sched.c
+++ b/drivers/gpu/drm/panthor/panthor_sched.c
@@ -898,8 +898,7 @@ static void group_free_queue(struct panthor_group *group, 
struct panthor_queue *
        if (IS_ERR_OR_NULL(queue))
                return;
 
-       if (queue->entity.fence_context)
-               drm_sched_entity_destroy(&queue->entity);
+       drm_sched_entity_destroy(&queue->entity);
 
        if (queue->scheduler.ops)
                drm_sched_fini(&queue->scheduler);
@@ -3609,11 +3608,6 @@ int panthor_group_destroy(struct panthor_file *pfile, 
u32 group_handle)
        if (!group)
                return -EINVAL;
 
-       for (u32 i = 0; i < group->queue_count; i++) {
-               if (group->queues[i])
-                       drm_sched_entity_destroy(&group->queues[i]->entity);
-       }
-
        mutex_lock(&sched->reset.lock);
        mutex_lock(&sched->lock);
        group->destroyed = true;
-- 
2.51.0

Reply via email to