Message ID | 20210805104705.862416-14-daniel.vetter@ffwll.ch (mailing list archive) |
---|---|
State | New, archived |
Headers | show |
Series | [v5,01/20] drm/sched: Split drm_sched_job_init | expand |
On Thu, Aug 05, 2021 at 12:46:58PM +0200, Daniel Vetter wrote: > Integrated into the scheduler now and all users converted over. > > Signed-off-by: Daniel Vetter <daniel.vetter@intel.com> > Cc: Maarten Lankhorst <maarten.lankhorst@linux.intel.com> > Cc: Maxime Ripard <mripard@kernel.org> > Cc: Thomas Zimmermann <tzimmermann@suse.de> > Cc: David Airlie <airlied@linux.ie> > Cc: Daniel Vetter <daniel@ffwll.ch> > Cc: Sumit Semwal <sumit.semwal@linaro.org> > Cc: "Christian König" <christian.koenig@amd.com> > Cc: linux-media@vger.kernel.org > Cc: linaro-mm-sig@lists.linaro.org Some acks would be great here. -Daniel > --- > drivers/gpu/drm/drm_gem.c | 96 --------------------------------------- > include/drm/drm_gem.h | 5 -- > 2 files changed, 101 deletions(-) > > diff --git a/drivers/gpu/drm/drm_gem.c b/drivers/gpu/drm/drm_gem.c > index 09c820045859..37e2e2820f08 100644 > --- a/drivers/gpu/drm/drm_gem.c > +++ b/drivers/gpu/drm/drm_gem.c > @@ -1272,99 +1272,3 @@ drm_gem_unlock_reservations(struct drm_gem_object **objs, int count, > ww_acquire_fini(acquire_ctx); > } > EXPORT_SYMBOL(drm_gem_unlock_reservations); > - > -/** > - * drm_gem_fence_array_add - Adds the fence to an array of fences to be > - * waited on, deduplicating fences from the same context. > - * > - * @fence_array: array of dma_fence * for the job to block on. > - * @fence: the dma_fence to add to the list of dependencies. > - * > - * This functions consumes the reference for @fence both on success and error > - * cases. > - * > - * Returns: > - * 0 on success, or an error on failing to expand the array. > - */ > -int drm_gem_fence_array_add(struct xarray *fence_array, > - struct dma_fence *fence) > -{ > - struct dma_fence *entry; > - unsigned long index; > - u32 id = 0; > - int ret; > - > - if (!fence) > - return 0; > - > - /* Deduplicate if we already depend on a fence from the same context. > - * This lets the size of the array of deps scale with the number of > - * engines involved, rather than the number of BOs. > - */ > - xa_for_each(fence_array, index, entry) { > - if (entry->context != fence->context) > - continue; > - > - if (dma_fence_is_later(fence, entry)) { > - dma_fence_put(entry); > - xa_store(fence_array, index, fence, GFP_KERNEL); > - } else { > - dma_fence_put(fence); > - } > - return 0; > - } > - > - ret = xa_alloc(fence_array, &id, fence, xa_limit_32b, GFP_KERNEL); > - if (ret != 0) > - dma_fence_put(fence); > - > - return ret; > -} > -EXPORT_SYMBOL(drm_gem_fence_array_add); > - > -/** > - * drm_gem_fence_array_add_implicit - Adds the implicit dependencies tracked > - * in the GEM object's reservation object to an array of dma_fences for use in > - * scheduling a rendering job. > - * > - * This should be called after drm_gem_lock_reservations() on your array of > - * GEM objects used in the job but before updating the reservations with your > - * own fences. > - * > - * @fence_array: array of dma_fence * for the job to block on. > - * @obj: the gem object to add new dependencies from. > - * @write: whether the job might write the object (so we need to depend on > - * shared fences in the reservation object). > - */ > -int drm_gem_fence_array_add_implicit(struct xarray *fence_array, > - struct drm_gem_object *obj, > - bool write) > -{ > - int ret; > - struct dma_fence **fences; > - unsigned int i, fence_count; > - > - if (!write) { > - struct dma_fence *fence = > - dma_resv_get_excl_unlocked(obj->resv); > - > - return drm_gem_fence_array_add(fence_array, fence); > - } > - > - ret = dma_resv_get_fences(obj->resv, NULL, > - &fence_count, &fences); > - if (ret || !fence_count) > - return ret; > - > - for (i = 0; i < fence_count; i++) { > - ret = drm_gem_fence_array_add(fence_array, fences[i]); > - if (ret) > - break; > - } > - > - for (; i < fence_count; i++) > - dma_fence_put(fences[i]); > - kfree(fences); > - return ret; > -} > -EXPORT_SYMBOL(drm_gem_fence_array_add_implicit); > diff --git a/include/drm/drm_gem.h b/include/drm/drm_gem.h > index 35e7f44c2a75..e55a767188af 100644 > --- a/include/drm/drm_gem.h > +++ b/include/drm/drm_gem.h > @@ -407,11 +407,6 @@ int drm_gem_lock_reservations(struct drm_gem_object **objs, int count, > struct ww_acquire_ctx *acquire_ctx); > void drm_gem_unlock_reservations(struct drm_gem_object **objs, int count, > struct ww_acquire_ctx *acquire_ctx); > -int drm_gem_fence_array_add(struct xarray *fence_array, > - struct dma_fence *fence); > -int drm_gem_fence_array_add_implicit(struct xarray *fence_array, > - struct drm_gem_object *obj, > - bool write); > int drm_gem_dumb_map_offset(struct drm_file *file, struct drm_device *dev, > u32 handle, u64 *offset); > > -- > 2.32.0 >
diff --git a/drivers/gpu/drm/drm_gem.c b/drivers/gpu/drm/drm_gem.c index 09c820045859..37e2e2820f08 100644 --- a/drivers/gpu/drm/drm_gem.c +++ b/drivers/gpu/drm/drm_gem.c @@ -1272,99 +1272,3 @@ drm_gem_unlock_reservations(struct drm_gem_object **objs, int count, ww_acquire_fini(acquire_ctx); } EXPORT_SYMBOL(drm_gem_unlock_reservations); - -/** - * drm_gem_fence_array_add - Adds the fence to an array of fences to be - * waited on, deduplicating fences from the same context. - * - * @fence_array: array of dma_fence * for the job to block on. - * @fence: the dma_fence to add to the list of dependencies. - * - * This functions consumes the reference for @fence both on success and error - * cases. - * - * Returns: - * 0 on success, or an error on failing to expand the array. - */ -int drm_gem_fence_array_add(struct xarray *fence_array, - struct dma_fence *fence) -{ - struct dma_fence *entry; - unsigned long index; - u32 id = 0; - int ret; - - if (!fence) - return 0; - - /* Deduplicate if we already depend on a fence from the same context. - * This lets the size of the array of deps scale with the number of - * engines involved, rather than the number of BOs. - */ - xa_for_each(fence_array, index, entry) { - if (entry->context != fence->context) - continue; - - if (dma_fence_is_later(fence, entry)) { - dma_fence_put(entry); - xa_store(fence_array, index, fence, GFP_KERNEL); - } else { - dma_fence_put(fence); - } - return 0; - } - - ret = xa_alloc(fence_array, &id, fence, xa_limit_32b, GFP_KERNEL); - if (ret != 0) - dma_fence_put(fence); - - return ret; -} -EXPORT_SYMBOL(drm_gem_fence_array_add); - -/** - * drm_gem_fence_array_add_implicit - Adds the implicit dependencies tracked - * in the GEM object's reservation object to an array of dma_fences for use in - * scheduling a rendering job. - * - * This should be called after drm_gem_lock_reservations() on your array of - * GEM objects used in the job but before updating the reservations with your - * own fences. - * - * @fence_array: array of dma_fence * for the job to block on. - * @obj: the gem object to add new dependencies from. - * @write: whether the job might write the object (so we need to depend on - * shared fences in the reservation object). - */ -int drm_gem_fence_array_add_implicit(struct xarray *fence_array, - struct drm_gem_object *obj, - bool write) -{ - int ret; - struct dma_fence **fences; - unsigned int i, fence_count; - - if (!write) { - struct dma_fence *fence = - dma_resv_get_excl_unlocked(obj->resv); - - return drm_gem_fence_array_add(fence_array, fence); - } - - ret = dma_resv_get_fences(obj->resv, NULL, - &fence_count, &fences); - if (ret || !fence_count) - return ret; - - for (i = 0; i < fence_count; i++) { - ret = drm_gem_fence_array_add(fence_array, fences[i]); - if (ret) - break; - } - - for (; i < fence_count; i++) - dma_fence_put(fences[i]); - kfree(fences); - return ret; -} -EXPORT_SYMBOL(drm_gem_fence_array_add_implicit); diff --git a/include/drm/drm_gem.h b/include/drm/drm_gem.h index 35e7f44c2a75..e55a767188af 100644 --- a/include/drm/drm_gem.h +++ b/include/drm/drm_gem.h @@ -407,11 +407,6 @@ int drm_gem_lock_reservations(struct drm_gem_object **objs, int count, struct ww_acquire_ctx *acquire_ctx); void drm_gem_unlock_reservations(struct drm_gem_object **objs, int count, struct ww_acquire_ctx *acquire_ctx); -int drm_gem_fence_array_add(struct xarray *fence_array, - struct dma_fence *fence); -int drm_gem_fence_array_add_implicit(struct xarray *fence_array, - struct drm_gem_object *obj, - bool write); int drm_gem_dumb_map_offset(struct drm_file *file, struct drm_device *dev, u32 handle, u64 *offset);
Integrated into the scheduler now and all users converted over. Signed-off-by: Daniel Vetter <daniel.vetter@intel.com> Cc: Maarten Lankhorst <maarten.lankhorst@linux.intel.com> Cc: Maxime Ripard <mripard@kernel.org> Cc: Thomas Zimmermann <tzimmermann@suse.de> Cc: David Airlie <airlied@linux.ie> Cc: Daniel Vetter <daniel@ffwll.ch> Cc: Sumit Semwal <sumit.semwal@linaro.org> Cc: "Christian König" <christian.koenig@amd.com> Cc: linux-media@vger.kernel.org Cc: linaro-mm-sig@lists.linaro.org --- drivers/gpu/drm/drm_gem.c | 96 --------------------------------------- include/drm/drm_gem.h | 5 -- 2 files changed, 101 deletions(-)