Re: [PATCH 1/8] drm: execution context for GEM buffers v2
Am 2022-06-09 um 03:18 schrieb Christian König: Am 09.06.22 um 02:05 schrieb Felix Kuehling: [SNIP] + * + * ret = drm_exec_lock(&exec, boB, 1); Where is drm_exec_lock? It's not in this patch. I've renamed this function to drm_exec_prepare_obj() but forgot to update the documentation. Going to fix this, thanks. + * drm_exec_continue_on_contention(&exec); + * if (ret) + * goto error; + * } + * + * drm_exec_for_each_locked_object(&exec, index, obj) { + * dma_resv_add_fence(obj->resv, fence, DMA_RESV_USAGE_READ); + * ... + * } + * drm_exec_fini(&exec); + * + * See struct dma_exec for more details. + */ + +/* Dummy value used to initially enter the retry loop */ +#define DRM_EXEC_DUMMY (void*)~0 + +/* Initialize the drm_exec_objects container */ +static void drm_exec_objects_init(struct drm_exec_objects *container) +{ + container->objects = kmalloc(PAGE_SIZE, GFP_KERNEL); Should this be kvmalloc? You're using kvrealloc and kvfree elsewhere. The initial number of objects tracked should be small enough so that we can use kmalloc() directly. Ok. It just seems weird to allocate with kmalloc and then kvrealloc the same pointer. But if that's safe it doesn't matter. BTW, if the caller already knows the number of BOs it wants to add, would is make sense to add that as a parameter to drm_exec_objects_init to save you all the reallocs? [SNIP] +/** + * drm_exec_prepare_obj - prepare a GEM object for use + * @exec: the drm_exec object with the state + * @obj: the GEM object to prepare + * @num_fences: how many fences to reserve + * + * Prepare a GEM object for use by locking it and reserving fence slots. All + * succesfully locked objects are put into the locked container. Duplicates + * detected as well and automatically moved into the duplicates container. + * + * Returns: -EDEADLK if a contention is detected, -ENOMEM when memory + * allocation failed and zero for success. + */ +int drm_exec_prepare_obj(struct drm_exec *exec, struct drm_gem_object *obj, + unsigned int num_fences) +{ + int ret; + + ret = drm_exec_lock_contended(exec); If this succeeds, it won't reserve any fence slots for object. Is that a problem? No, the contended object should be re-presented later on and then we allocate the fence slots. Ok. Along with all other objects, because drm_exec_cleanup will have cleaned up the lists. + if (unlikely(ret == -EALREADY)) { + struct drm_exec_objects *container = &exec->duplicates; + + /* + * If this is the first locked GEM object it was most likely + * just contended. So don't add it to the duplicates, just + * reserve the fence slots. I don't understand this. Seems a bit arbitrary. Is it even legal to try to add the same object twice? I thought duplicates was for different objects that share the same reservation, not actually the same object on the same list twice. Yes, it's legal to try the same object twice. That can easily happen when userspace specifies the same handle twice for a submission. The other possibility is that we had a contention, backed off and then locked the contention first and then re-tried everything else once more. Maybe you meant to compare with exec->locked.objects[exec->locked.num_objects-1], assuming that drm_exec_lock_contended just succeeded locking a previously contended object, and the caller retried locking that same object again? Yes, that's pretty much the idea. But then exec->locked.num_objects is just 1 so that's equal to checking exec->locked.num_objects[0]. Ok. I missed that after drm_exec_continue_on_contention the drm_exec_cleanup will clean up the lists and we start over locking all the objects. Why can't drm_exec_cleanup just lock the contended object itself? I think that would make it more obvious that the contended object is always the first one that gets locked and added to the list. Thanks, Felix On the other hand we would also catch cases where we lock the same object multiple times directly behind each other. Need to think about that a bit. + */ + if (exec->locked.num_objects && exec->locked.objects[0] == obj) + container = NULL; + + ret = drm_exec_obj_locked(container, obj, num_fences); + if (ret) + return ret; + + } else if (unlikely(ret)) { + return ret; + + } else { + ret = drm_exec_obj_locked(&exec->locked, obj, num_fences); + if (ret) + goto error_unlock; + } + + drm_gem_object_get(obj); The container already gets a reference to obj. What is this extra reference for? And where does it get dropped? Need to double check this, might be that it's just a leftover. Thanks, Christian. Regards, Felix
Re: [PATCH 1/8] drm: execution context for GEM buffers v2
Am 09.06.22 um 02:05 schrieb Felix Kuehling: [SNIP] + * + * ret = drm_exec_lock(&exec, boB, 1); Where is drm_exec_lock? It's not in this patch. I've renamed this function to drm_exec_prepare_obj() but forgot to update the documentation. Going to fix this, thanks. + * drm_exec_continue_on_contention(&exec); + * if (ret) + * goto error; + * } + * + * drm_exec_for_each_locked_object(&exec, index, obj) { + * dma_resv_add_fence(obj->resv, fence, DMA_RESV_USAGE_READ); + * ... + * } + * drm_exec_fini(&exec); + * + * See struct dma_exec for more details. + */ + +/* Dummy value used to initially enter the retry loop */ +#define DRM_EXEC_DUMMY (void*)~0 + +/* Initialize the drm_exec_objects container */ +static void drm_exec_objects_init(struct drm_exec_objects *container) +{ + container->objects = kmalloc(PAGE_SIZE, GFP_KERNEL); Should this be kvmalloc? You're using kvrealloc and kvfree elsewhere. The initial number of objects tracked should be small enough so that we can use kmalloc() directly. [SNIP] +/** + * drm_exec_prepare_obj - prepare a GEM object for use + * @exec: the drm_exec object with the state + * @obj: the GEM object to prepare + * @num_fences: how many fences to reserve + * + * Prepare a GEM object for use by locking it and reserving fence slots. All + * succesfully locked objects are put into the locked container. Duplicates + * detected as well and automatically moved into the duplicates container. + * + * Returns: -EDEADLK if a contention is detected, -ENOMEM when memory + * allocation failed and zero for success. + */ +int drm_exec_prepare_obj(struct drm_exec *exec, struct drm_gem_object *obj, + unsigned int num_fences) +{ + int ret; + + ret = drm_exec_lock_contended(exec); If this succeeds, it won't reserve any fence slots for object. Is that a problem? No, the contended object should be re-presented later on and then we allocate the fence slots. + if (unlikely(ret == -EALREADY)) { + struct drm_exec_objects *container = &exec->duplicates; + + /* + * If this is the first locked GEM object it was most likely + * just contended. So don't add it to the duplicates, just + * reserve the fence slots. I don't understand this. Seems a bit arbitrary. Is it even legal to try to add the same object twice? I thought duplicates was for different objects that share the same reservation, not actually the same object on the same list twice. Yes, it's legal to try the same object twice. That can easily happen when userspace specifies the same handle twice for a submission. The other possibility is that we had a contention, backed off and then locked the contention first and then re-tried everything else once more. Maybe you meant to compare with exec->locked.objects[exec->locked.num_objects-1], assuming that drm_exec_lock_contended just succeeded locking a previously contended object, and the caller retried locking that same object again? Yes, that's pretty much the idea. But then exec->locked.num_objects is just 1 so that's equal to checking exec->locked.num_objects[0]. On the other hand we would also catch cases where we lock the same object multiple times directly behind each other. Need to think about that a bit. + */ + if (exec->locked.num_objects && exec->locked.objects[0] == obj) + container = NULL; + + ret = drm_exec_obj_locked(container, obj, num_fences); + if (ret) + return ret; + + } else if (unlikely(ret)) { + return ret; + + } else { + ret = drm_exec_obj_locked(&exec->locked, obj, num_fences); + if (ret) + goto error_unlock; + } + + drm_gem_object_get(obj); The container already gets a reference to obj. What is this extra reference for? And where does it get dropped? Need to double check this, might be that it's just a leftover. Thanks, Christian. Regards, Felix
Re: [PATCH 1/8] drm: execution context for GEM buffers v2
On 2022-05-04 03:47, Christian König wrote: This adds the infrastructure for an execution context for GEM buffers which is similar to the existinc TTMs execbuf util and intended to replace it in the long term. The basic functionality is that we abstracts the necessary loop to lock many different GEM buffers with automated deadlock and duplicate handling. v2: drop xarray and use dynamic resized array instead, the locking overhead is unecessary and measureable. Signed-off-by: Christian König I finally got around to catching up with this thread. See some questions and comments inline. --- Documentation/gpu/drm-mm.rst | 12 ++ drivers/gpu/drm/Kconfig | 7 + drivers/gpu/drm/Makefile | 2 + drivers/gpu/drm/drm_exec.c | 295 +++ include/drm/drm_exec.h | 144 + 5 files changed, 460 insertions(+) create mode 100644 drivers/gpu/drm/drm_exec.c create mode 100644 include/drm/drm_exec.h diff --git a/Documentation/gpu/drm-mm.rst b/Documentation/gpu/drm-mm.rst index f32ccce5722d..bf7dd2a78e9b 100644 --- a/Documentation/gpu/drm-mm.rst +++ b/Documentation/gpu/drm-mm.rst @@ -493,6 +493,18 @@ DRM Sync Objects .. kernel-doc:: drivers/gpu/drm/drm_syncobj.c :export: +DRM Execution context += + +.. kernel-doc:: drivers/gpu/drm/drm_exec.c + :doc: Overview + +.. kernel-doc:: include/drm/drm_exec.h + :internal: + +.. kernel-doc:: drivers/gpu/drm/drm_exec.c + :export: + GPU Scheduler = diff --git a/drivers/gpu/drm/Kconfig b/drivers/gpu/drm/Kconfig index e88c497fa010..1b35c10df263 100644 --- a/drivers/gpu/drm/Kconfig +++ b/drivers/gpu/drm/Kconfig @@ -179,6 +179,12 @@ config DRM_TTM GPU memory types. Will be enabled automatically if a device driver uses it. +config DRM_EXEC + tristate + depends on DRM + help + Execution context for command submissions + config DRM_BUDDY tristate depends on DRM @@ -252,6 +258,7 @@ config DRM_AMDGPU select DRM_SCHED select DRM_TTM select DRM_TTM_HELPER + select DRM_EXEC select POWER_SUPPLY select HWMON select BACKLIGHT_CLASS_DEVICE diff --git a/drivers/gpu/drm/Makefile b/drivers/gpu/drm/Makefile index 15fe3163f822..ee8573b683f3 100644 --- a/drivers/gpu/drm/Makefile +++ b/drivers/gpu/drm/Makefile @@ -37,6 +37,8 @@ obj-$(CONFIG_DRM_PANEL_ORIENTATION_QUIRKS) += drm_panel_orientation_quirks.o # # Memory-management helpers # +# +obj-$(CONFIG_DRM_EXEC) += drm_exec.o obj-$(CONFIG_DRM_BUDDY) += drm_buddy.o diff --git a/drivers/gpu/drm/drm_exec.c b/drivers/gpu/drm/drm_exec.c new file mode 100644 index ..ed2106c22786 --- /dev/null +++ b/drivers/gpu/drm/drm_exec.c @@ -0,0 +1,295 @@ +/* SPDX-License-Identifier: GPL-2.0 OR MIT */ + +#include +#include +#include + +/** + * DOC: Overview + * + * This component mainly abstracts the retry loop necessary for locking + * multiple GEM objects while preparing hardware operations (e.g. command + * submissions, page table updates etc..). + * + * If a contention is detected while locking a GEM object the cleanup procedure + * unlocks all previously locked GEM objects and locks the contended one first + * before locking any further objects. + * + * After an object is locked fences slots can optionally be reserved on the + * dma_resv object inside the GEM object. + * + * A typical usage pattern should look like this:: + * + * struct drm_gem_object *obj; + * struct drm_exec exec; + * unsigned long index; + * int ret; + * + * drm_exec_init(&exec, true); + * drm_exec_while_not_all_locked(&exec) { + * ret = drm_exec_prepare_obj(&exec, boA, 1); + * drm_exec_continue_on_contention(&exec); + * if (ret) + * goto error; + * + * ret = drm_exec_lock(&exec, boB, 1); Where is drm_exec_lock? It's not in this patch. + * drm_exec_continue_on_contention(&exec); + * if (ret) + * goto error; + * } + * + * drm_exec_for_each_locked_object(&exec, index, obj) { + * dma_resv_add_fence(obj->resv, fence, DMA_RESV_USAGE_READ); + * ... + * } + * drm_exec_fini(&exec); + * + * See struct dma_exec for more details. + */ + +/* Dummy value used to initially enter the retry loop */ +#define DRM_EXEC_DUMMY (void*)~0 + +/* Initialize the drm_exec_objects container */ +static void drm_exec_objects_init(struct drm_exec_objects *container) +{ + container->objects = kmalloc(PAGE_SIZE, GFP_KERNEL); Should this be kvmalloc? You're using kvrealloc and kvfree elsewhere. + + /* If allocation here fails, just delay that till the first use */ + container->max_objects = container->objects ? + PAGE_SIZE / sizeof(void *) : 0; + container->num_objects = 0; +} + +/* Cleanup the drm_exec_objects container
Re: [PATCH 1/8] drm: execution context for GEM buffers v2
On Mon, May 09, 2022 at 05:01:33PM +0200, Christian König wrote: > Am 09.05.22 um 16:31 schrieb Daniel Vetter: > > On Wed, May 04, 2022 at 09:47:32AM +0200, Christian König wrote: > > > [SNIP] > > > +/* Make sure we have enough room and add an object the container */ > > > +static int drm_exec_objects_add(struct drm_exec_objects *container, > > > + struct drm_gem_object *obj) > > > +{ > > > + if (unlikely(container->num_objects == container->max_objects)) { > > > + size_t size = container->max_objects * sizeof(void *); > > > + void *tmp; > > > + > > > + tmp = kvrealloc(container->objects, size, size + PAGE_SIZE, > > > + GFP_KERNEL); > > > + if (!tmp) > > > + return -ENOMEM; > > > + > > > + container->objects = tmp; > > > + container->max_objects += PAGE_SIZE / sizeof(void *); > > Might be worth it to inquire the actual allocation size here, since if > > it's kmalloc the generic buckets only cover doubling of sizes, so once > > it's big it goes up a lot quicker than PAGE_SIZE. > > > > But also krealloc checks this internally already so maybe better to not > > break the abstraction. > > How can I actually do this? ksize() only works with kmalloc(). > > Or do we had a function to figure out if vmalloc or kmalloc was used by > kvrealloc()? kvfree has a is_vmalloc_addr so it would boil down to open-code that a bit. Probably not worth the trouble really, otoh looking at kvrealloc it doesn't use krealloc underneath, so it's not doing that check. Maybe we should just push that check into kvrealloc for the !vmalloc_addr case. > > > [SNIP] > > > +/** > > > + * drm_exec_cleanup - cleanup when contention is detected > > > + * @exec: the drm_exec object to cleanup > > > + * > > > + * Cleanup the current state and return true if we should stay inside > > > the retry > > > + * loop, false if there wasn't any contention detected and we can keep > > > the > > > + * objects locked. > > > + */ > > > +bool drm_exec_cleanup(struct drm_exec *exec) > > > +{ > > > + if (likely(!exec->contended)) { > > > + ww_acquire_done(&exec->ticket); > > > + return false; > > > + } > > > + > > > + if (likely(exec->contended == DRM_EXEC_DUMMY)) { > > > + exec->contended = NULL; > > > + ww_acquire_init(&exec->ticket, &reservation_ww_class); > > Not sure why this is here instead of in _init()? I thought you're playing > > some really dangerous tricks with re-initting the acquire ctx, which would > > at least be questionable, but does not look like that. > > That was my initial design, but the problem with this approach is that all > locks taken between drm_exec_init() and the loop suddenly have a lockdep > dependency on reservation_ww_class. And that in turn goes boom immediately. > > Took me a moment to realize what's wrong with that as well. Uh crap, indeed. I think minimally this needs to be document, but personally I'm leaning towards drm_exec_prepare_init(), which does this explicitly. I do agree we need this split, especially so we can eventually add helpers for bo lookup, or maybe userptr/hmm prep and things like that, which all has to be outside of the acquire_ctx. > > [SNIP] > > +/** > > + * drm_exec_has_duplicates - check for duplicated GEM object > > + * @exec: drm_exec object > > + * > > + * Return true if the drm_exec object has encountered some already locked > > GEM > > + * objects while trying to lock them. This can happen if multiple GEM > > objects > > + * share the same underlying resv object. > > + */ > > +static inline bool drm_exec_has_duplicates(struct drm_exec *exec) > > +{ > > + return exec->duplicates.num_objects > 0; > > Definitely an aside, but in our i915 efforts to get rid of temporary pins > > we run into some fun where the eviction code couldn't differentiate from > > memory we need reserved for the CS and memory we just keep locked because > > we evicted it and fun stuff like that. So maybe we need a bit more > > tracking here eventually, but that's only when we have this somehow glued > > into ttm eviction code. > > Hehe, yeah that's what I was thinking about as well. But then I though one > step at a time. > > > Also the even more massive step would be to glue this into dma-buf so you > > can do dynamic dma-buf eviction and still keep track of all the buffers. I > > think with some clever pointer tagging and a bit more indirection we could > > nest drm_exec structures (so that a driver could insert it's entire > > drm_exec structure with a drm_exec-level callback for handling refcounting > > and stuff like that). > > I considered in which component to put this quite a bit as well, but then > intentionally decided against DMA-buf. > > One major reason was that not all buffers which needs to be locked this way > are actually exported as DMA-buf. > > Another reason is that DMA-buf doesn't necessary need a concept of an > execution context. As far as I
Re: [PATCH 1/8] drm: execution context for GEM buffers v2
Am 09.05.22 um 16:31 schrieb Daniel Vetter: On Wed, May 04, 2022 at 09:47:32AM +0200, Christian König wrote: [SNIP] +/* Make sure we have enough room and add an object the container */ +static int drm_exec_objects_add(struct drm_exec_objects *container, + struct drm_gem_object *obj) +{ + if (unlikely(container->num_objects == container->max_objects)) { + size_t size = container->max_objects * sizeof(void *); + void *tmp; + + tmp = kvrealloc(container->objects, size, size + PAGE_SIZE, + GFP_KERNEL); + if (!tmp) + return -ENOMEM; + + container->objects = tmp; + container->max_objects += PAGE_SIZE / sizeof(void *); Might be worth it to inquire the actual allocation size here, since if it's kmalloc the generic buckets only cover doubling of sizes, so once it's big it goes up a lot quicker than PAGE_SIZE. But also krealloc checks this internally already so maybe better to not break the abstraction. How can I actually do this? ksize() only works with kmalloc(). Or do we had a function to figure out if vmalloc or kmalloc was used by kvrealloc()? [SNIP] +/** + * drm_exec_cleanup - cleanup when contention is detected + * @exec: the drm_exec object to cleanup + * + * Cleanup the current state and return true if we should stay inside the retry + * loop, false if there wasn't any contention detected and we can keep the + * objects locked. + */ +bool drm_exec_cleanup(struct drm_exec *exec) +{ + if (likely(!exec->contended)) { + ww_acquire_done(&exec->ticket); + return false; + } + + if (likely(exec->contended == DRM_EXEC_DUMMY)) { + exec->contended = NULL; + ww_acquire_init(&exec->ticket, &reservation_ww_class); Not sure why this is here instead of in _init()? I thought you're playing some really dangerous tricks with re-initting the acquire ctx, which would at least be questionable, but does not look like that. That was my initial design, but the problem with this approach is that all locks taken between drm_exec_init() and the loop suddenly have a lockdep dependency on reservation_ww_class. And that in turn goes boom immediately. Took me a moment to realize what's wrong with that as well. [SNIP] +/** + * drm_exec_has_duplicates - check for duplicated GEM object + * @exec: drm_exec object + * + * Return true if the drm_exec object has encountered some already locked GEM + * objects while trying to lock them. This can happen if multiple GEM objects + * share the same underlying resv object. + */ +static inline bool drm_exec_has_duplicates(struct drm_exec *exec) +{ + return exec->duplicates.num_objects > 0; Definitely an aside, but in our i915 efforts to get rid of temporary pins we run into some fun where the eviction code couldn't differentiate from memory we need reserved for the CS and memory we just keep locked because we evicted it and fun stuff like that. So maybe we need a bit more tracking here eventually, but that's only when we have this somehow glued into ttm eviction code. Hehe, yeah that's what I was thinking about as well. But then I though one step at a time. Also the even more massive step would be to glue this into dma-buf so you can do dynamic dma-buf eviction and still keep track of all the buffers. I think with some clever pointer tagging and a bit more indirection we could nest drm_exec structures (so that a driver could insert it's entire drm_exec structure with a drm_exec-level callback for handling refcounting and stuff like that). I considered in which component to put this quite a bit as well, but then intentionally decided against DMA-buf. One major reason was that not all buffers which needs to be locked this way are actually exported as DMA-buf. Another reason is that DMA-buf doesn't necessary need a concept of an execution context. As far as I can see that's something GPU/DRM driver specific. So anyway I think this all looks good, just one more thing before I think we should land this: gem helpers in drm_gem_lock_reservations() has something which is practically compatible already, except that you bulk-add the entire set of objects. I think if you add a bulk-prepare function then we could also replace all those. Maybe even nicer if the bulk-prepare takes the array of handles and does the handle lookup too, but at least something which can subsititue drm_gem_lock_reservations with drm_exec would be nice to validate the helpers a bit more and really make sure we only have one of them left. I was considering that as well, but then also thought one step at a time. Not sure if it's possible to look up handles without running into some locking fun, thought. Thanks for the review, Christian. Thoughts? -Daniel +} + +void drm_exec_init(struct drm_exec *exec, bool interruptible); +void drm_exec_fini(
Re: [PATCH 1/8] drm: execution context for GEM buffers v2
On Wed, May 04, 2022 at 09:47:32AM +0200, Christian König wrote: > This adds the infrastructure for an execution context for GEM buffers > which is similar to the existinc TTMs execbuf util and intended to replace > it in the long term. > > The basic functionality is that we abstracts the necessary loop to lock > many different GEM buffers with automated deadlock and duplicate handling. > > v2: drop xarray and use dynamic resized array instead, the locking > overhead is unecessary and measureable. > > Signed-off-by: Christian König > --- > Documentation/gpu/drm-mm.rst | 12 ++ > drivers/gpu/drm/Kconfig | 7 + > drivers/gpu/drm/Makefile | 2 + > drivers/gpu/drm/drm_exec.c | 295 +++ > include/drm/drm_exec.h | 144 + > 5 files changed, 460 insertions(+) > create mode 100644 drivers/gpu/drm/drm_exec.c > create mode 100644 include/drm/drm_exec.h > > diff --git a/Documentation/gpu/drm-mm.rst b/Documentation/gpu/drm-mm.rst > index f32ccce5722d..bf7dd2a78e9b 100644 > --- a/Documentation/gpu/drm-mm.rst > +++ b/Documentation/gpu/drm-mm.rst > @@ -493,6 +493,18 @@ DRM Sync Objects > .. kernel-doc:: drivers/gpu/drm/drm_syncobj.c > :export: > > +DRM Execution context > += > + > +.. kernel-doc:: drivers/gpu/drm/drm_exec.c > + :doc: Overview > + > +.. kernel-doc:: include/drm/drm_exec.h > + :internal: > + > +.. kernel-doc:: drivers/gpu/drm/drm_exec.c > + :export: > + > GPU Scheduler > = > > diff --git a/drivers/gpu/drm/Kconfig b/drivers/gpu/drm/Kconfig > index e88c497fa010..1b35c10df263 100644 > --- a/drivers/gpu/drm/Kconfig > +++ b/drivers/gpu/drm/Kconfig > @@ -179,6 +179,12 @@ config DRM_TTM > GPU memory types. Will be enabled automatically if a device driver > uses it. > > +config DRM_EXEC > + tristate > + depends on DRM > + help > + Execution context for command submissions > + > config DRM_BUDDY > tristate > depends on DRM > @@ -252,6 +258,7 @@ config DRM_AMDGPU > select DRM_SCHED > select DRM_TTM > select DRM_TTM_HELPER > + select DRM_EXEC > select POWER_SUPPLY > select HWMON > select BACKLIGHT_CLASS_DEVICE > diff --git a/drivers/gpu/drm/Makefile b/drivers/gpu/drm/Makefile > index 15fe3163f822..ee8573b683f3 100644 > --- a/drivers/gpu/drm/Makefile > +++ b/drivers/gpu/drm/Makefile > @@ -37,6 +37,8 @@ obj-$(CONFIG_DRM_PANEL_ORIENTATION_QUIRKS) += > drm_panel_orientation_quirks.o > # > # Memory-management helpers > # > +# > +obj-$(CONFIG_DRM_EXEC) += drm_exec.o > > obj-$(CONFIG_DRM_BUDDY) += drm_buddy.o > > diff --git a/drivers/gpu/drm/drm_exec.c b/drivers/gpu/drm/drm_exec.c > new file mode 100644 > index ..ed2106c22786 > --- /dev/null > +++ b/drivers/gpu/drm/drm_exec.c > @@ -0,0 +1,295 @@ > +/* SPDX-License-Identifier: GPL-2.0 OR MIT */ > + > +#include > +#include > +#include > + > +/** > + * DOC: Overview > + * > + * This component mainly abstracts the retry loop necessary for locking > + * multiple GEM objects while preparing hardware operations (e.g. command > + * submissions, page table updates etc..). > + * > + * If a contention is detected while locking a GEM object the cleanup > procedure > + * unlocks all previously locked GEM objects and locks the contended one > first > + * before locking any further objects. > + * > + * After an object is locked fences slots can optionally be reserved on the > + * dma_resv object inside the GEM object. > + * > + * A typical usage pattern should look like this:: > + * > + * struct drm_gem_object *obj; > + * struct drm_exec exec; > + * unsigned long index; > + * int ret; > + * > + * drm_exec_init(&exec, true); > + * drm_exec_while_not_all_locked(&exec) { > + * ret = drm_exec_prepare_obj(&exec, boA, 1); > + * drm_exec_continue_on_contention(&exec); > + * if (ret) > + * goto error; > + * > + * ret = drm_exec_lock(&exec, boB, 1); > + * drm_exec_continue_on_contention(&exec); > + * if (ret) > + * goto error; > + * } > + * > + * drm_exec_for_each_locked_object(&exec, index, obj) { > + * dma_resv_add_fence(obj->resv, fence, DMA_RESV_USAGE_READ); > + * ... > + * } > + * drm_exec_fini(&exec); > + * > + * See struct dma_exec for more details. > + */ > + > +/* Dummy value used to initially enter the retry loop */ > +#define DRM_EXEC_DUMMY (void*)~0 > + > +/* Initialize the drm_exec_objects container */ > +static void drm_exec_objects_init(struct drm_exec_objects *container) > +{ > + container->objects = kmalloc(PAGE_SIZE, GFP_KERNEL); > + > + /* If allocation here fails, just delay that till the first use */ > + container->max_objects = container->objects ? > + PAGE_SIZE / sizeof(void *) : 0; > + container->num_objects = 0; > +} > + > +/* Cleanup the drm_exec_objects container */ > +
[PATCH 1/8] drm: execution context for GEM buffers v2
This adds the infrastructure for an execution context for GEM buffers which is similar to the existinc TTMs execbuf util and intended to replace it in the long term. The basic functionality is that we abstracts the necessary loop to lock many different GEM buffers with automated deadlock and duplicate handling. v2: drop xarray and use dynamic resized array instead, the locking overhead is unecessary and measureable. Signed-off-by: Christian König --- Documentation/gpu/drm-mm.rst | 12 ++ drivers/gpu/drm/Kconfig | 7 + drivers/gpu/drm/Makefile | 2 + drivers/gpu/drm/drm_exec.c | 295 +++ include/drm/drm_exec.h | 144 + 5 files changed, 460 insertions(+) create mode 100644 drivers/gpu/drm/drm_exec.c create mode 100644 include/drm/drm_exec.h diff --git a/Documentation/gpu/drm-mm.rst b/Documentation/gpu/drm-mm.rst index f32ccce5722d..bf7dd2a78e9b 100644 --- a/Documentation/gpu/drm-mm.rst +++ b/Documentation/gpu/drm-mm.rst @@ -493,6 +493,18 @@ DRM Sync Objects .. kernel-doc:: drivers/gpu/drm/drm_syncobj.c :export: +DRM Execution context += + +.. kernel-doc:: drivers/gpu/drm/drm_exec.c + :doc: Overview + +.. kernel-doc:: include/drm/drm_exec.h + :internal: + +.. kernel-doc:: drivers/gpu/drm/drm_exec.c + :export: + GPU Scheduler = diff --git a/drivers/gpu/drm/Kconfig b/drivers/gpu/drm/Kconfig index e88c497fa010..1b35c10df263 100644 --- a/drivers/gpu/drm/Kconfig +++ b/drivers/gpu/drm/Kconfig @@ -179,6 +179,12 @@ config DRM_TTM GPU memory types. Will be enabled automatically if a device driver uses it. +config DRM_EXEC + tristate + depends on DRM + help + Execution context for command submissions + config DRM_BUDDY tristate depends on DRM @@ -252,6 +258,7 @@ config DRM_AMDGPU select DRM_SCHED select DRM_TTM select DRM_TTM_HELPER + select DRM_EXEC select POWER_SUPPLY select HWMON select BACKLIGHT_CLASS_DEVICE diff --git a/drivers/gpu/drm/Makefile b/drivers/gpu/drm/Makefile index 15fe3163f822..ee8573b683f3 100644 --- a/drivers/gpu/drm/Makefile +++ b/drivers/gpu/drm/Makefile @@ -37,6 +37,8 @@ obj-$(CONFIG_DRM_PANEL_ORIENTATION_QUIRKS) += drm_panel_orientation_quirks.o # # Memory-management helpers # +# +obj-$(CONFIG_DRM_EXEC) += drm_exec.o obj-$(CONFIG_DRM_BUDDY) += drm_buddy.o diff --git a/drivers/gpu/drm/drm_exec.c b/drivers/gpu/drm/drm_exec.c new file mode 100644 index ..ed2106c22786 --- /dev/null +++ b/drivers/gpu/drm/drm_exec.c @@ -0,0 +1,295 @@ +/* SPDX-License-Identifier: GPL-2.0 OR MIT */ + +#include +#include +#include + +/** + * DOC: Overview + * + * This component mainly abstracts the retry loop necessary for locking + * multiple GEM objects while preparing hardware operations (e.g. command + * submissions, page table updates etc..). + * + * If a contention is detected while locking a GEM object the cleanup procedure + * unlocks all previously locked GEM objects and locks the contended one first + * before locking any further objects. + * + * After an object is locked fences slots can optionally be reserved on the + * dma_resv object inside the GEM object. + * + * A typical usage pattern should look like this:: + * + * struct drm_gem_object *obj; + * struct drm_exec exec; + * unsigned long index; + * int ret; + * + * drm_exec_init(&exec, true); + * drm_exec_while_not_all_locked(&exec) { + * ret = drm_exec_prepare_obj(&exec, boA, 1); + * drm_exec_continue_on_contention(&exec); + * if (ret) + * goto error; + * + * ret = drm_exec_lock(&exec, boB, 1); + * drm_exec_continue_on_contention(&exec); + * if (ret) + * goto error; + * } + * + * drm_exec_for_each_locked_object(&exec, index, obj) { + * dma_resv_add_fence(obj->resv, fence, DMA_RESV_USAGE_READ); + * ... + * } + * drm_exec_fini(&exec); + * + * See struct dma_exec for more details. + */ + +/* Dummy value used to initially enter the retry loop */ +#define DRM_EXEC_DUMMY (void*)~0 + +/* Initialize the drm_exec_objects container */ +static void drm_exec_objects_init(struct drm_exec_objects *container) +{ + container->objects = kmalloc(PAGE_SIZE, GFP_KERNEL); + + /* If allocation here fails, just delay that till the first use */ + container->max_objects = container->objects ? + PAGE_SIZE / sizeof(void *) : 0; + container->num_objects = 0; +} + +/* Cleanup the drm_exec_objects container */ +static void drm_exec_objects_fini(struct drm_exec_objects *container) +{ + kvfree(container->objects); +} + +/* Make sure we have enough room and add an object the container */ +static int drm_exec_objects_add(struct drm_exec_objects *container, + struct drm_gem_
[PATCH 1/8] drm: execution context for GEM buffers v2
This adds the infrastructure for an execution context for GEM buffers which is similar to the existinc TTMs execbuf util and intended to replace it in the long term. The basic functionality is that we abstracts the necessary loop to lock many different GEM buffers with automated deadlock and duplicate handling. v2: drop xarray and use dynamic resized array instead, the locking overhead is unecessary and measureable. Signed-off-by: Christian König --- Documentation/gpu/drm-mm.rst | 12 ++ drivers/gpu/drm/Kconfig | 7 + drivers/gpu/drm/Makefile | 2 + drivers/gpu/drm/drm_exec.c | 295 +++ include/drm/drm_exec.h | 144 + 5 files changed, 460 insertions(+) create mode 100644 drivers/gpu/drm/drm_exec.c create mode 100644 include/drm/drm_exec.h diff --git a/Documentation/gpu/drm-mm.rst b/Documentation/gpu/drm-mm.rst index f32ccce5722d..bf7dd2a78e9b 100644 --- a/Documentation/gpu/drm-mm.rst +++ b/Documentation/gpu/drm-mm.rst @@ -493,6 +493,18 @@ DRM Sync Objects .. kernel-doc:: drivers/gpu/drm/drm_syncobj.c :export: +DRM Execution context += + +.. kernel-doc:: drivers/gpu/drm/drm_exec.c + :doc: Overview + +.. kernel-doc:: include/drm/drm_exec.h + :internal: + +.. kernel-doc:: drivers/gpu/drm/drm_exec.c + :export: + GPU Scheduler = diff --git a/drivers/gpu/drm/Kconfig b/drivers/gpu/drm/Kconfig index e88c497fa010..1b35c10df263 100644 --- a/drivers/gpu/drm/Kconfig +++ b/drivers/gpu/drm/Kconfig @@ -179,6 +179,12 @@ config DRM_TTM GPU memory types. Will be enabled automatically if a device driver uses it. +config DRM_EXEC + tristate + depends on DRM + help + Execution context for command submissions + config DRM_BUDDY tristate depends on DRM @@ -252,6 +258,7 @@ config DRM_AMDGPU select DRM_SCHED select DRM_TTM select DRM_TTM_HELPER + select DRM_EXEC select POWER_SUPPLY select HWMON select BACKLIGHT_CLASS_DEVICE diff --git a/drivers/gpu/drm/Makefile b/drivers/gpu/drm/Makefile index 15fe3163f822..ee8573b683f3 100644 --- a/drivers/gpu/drm/Makefile +++ b/drivers/gpu/drm/Makefile @@ -37,6 +37,8 @@ obj-$(CONFIG_DRM_PANEL_ORIENTATION_QUIRKS) += drm_panel_orientation_quirks.o # # Memory-management helpers # +# +obj-$(CONFIG_DRM_EXEC) += drm_exec.o obj-$(CONFIG_DRM_BUDDY) += drm_buddy.o diff --git a/drivers/gpu/drm/drm_exec.c b/drivers/gpu/drm/drm_exec.c new file mode 100644 index ..ed2106c22786 --- /dev/null +++ b/drivers/gpu/drm/drm_exec.c @@ -0,0 +1,295 @@ +/* SPDX-License-Identifier: GPL-2.0 OR MIT */ + +#include +#include +#include + +/** + * DOC: Overview + * + * This component mainly abstracts the retry loop necessary for locking + * multiple GEM objects while preparing hardware operations (e.g. command + * submissions, page table updates etc..). + * + * If a contention is detected while locking a GEM object the cleanup procedure + * unlocks all previously locked GEM objects and locks the contended one first + * before locking any further objects. + * + * After an object is locked fences slots can optionally be reserved on the + * dma_resv object inside the GEM object. + * + * A typical usage pattern should look like this:: + * + * struct drm_gem_object *obj; + * struct drm_exec exec; + * unsigned long index; + * int ret; + * + * drm_exec_init(&exec, true); + * drm_exec_while_not_all_locked(&exec) { + * ret = drm_exec_prepare_obj(&exec, boA, 1); + * drm_exec_continue_on_contention(&exec); + * if (ret) + * goto error; + * + * ret = drm_exec_lock(&exec, boB, 1); + * drm_exec_continue_on_contention(&exec); + * if (ret) + * goto error; + * } + * + * drm_exec_for_each_locked_object(&exec, index, obj) { + * dma_resv_add_fence(obj->resv, fence, DMA_RESV_USAGE_READ); + * ... + * } + * drm_exec_fini(&exec); + * + * See struct dma_exec for more details. + */ + +/* Dummy value used to initially enter the retry loop */ +#define DRM_EXEC_DUMMY (void*)~0 + +/* Initialize the drm_exec_objects container */ +static void drm_exec_objects_init(struct drm_exec_objects *container) +{ + container->objects = kmalloc(PAGE_SIZE, GFP_KERNEL); + + /* If allocation here fails, just delay that till the first use */ + container->max_objects = container->objects ? + PAGE_SIZE / sizeof(void *) : 0; + container->num_objects = 0; +} + +/* Cleanup the drm_exec_objects container */ +static void drm_exec_objects_fini(struct drm_exec_objects *container) +{ + kvfree(container->objects); +} + +/* Make sure we have enough room and add an object the container */ +static int drm_exec_objects_add(struct drm_exec_objects *container, + struct drm_gem_