diff mbox series

[36/59] drm/ttm: add wrapper to get manager from bdev.

Message ID 20200804025632.3868079-37-airlied@gmail.com (mailing list archive)
State New, archived
Headers show
Series ttm misc cleanups, mem refactoring, rename objects. (v2) | expand

Commit Message

Dave Airlie Aug. 4, 2020, 2:56 a.m. UTC
From: Dave Airlie <airlied@redhat.com>

This will allow different abstractions later.

Signed-off-by: Dave Airlie <airlied@redhat.com>
---
 drivers/gpu/drm/ttm/ttm_bo.c      | 34 +++++++++++++++----------------
 drivers/gpu/drm/ttm/ttm_bo_util.c | 20 +++++++++---------
 drivers/gpu/drm/ttm/ttm_bo_vm.c   |  2 +-
 include/drm/ttm/ttm_bo_driver.h   |  6 ++++++
 4 files changed, 34 insertions(+), 28 deletions(-)

Comments

Christian König Aug. 4, 2020, 11:25 a.m. UTC | #1
Am 04.08.20 um 04:56 schrieb Dave Airlie:
> From: Dave Airlie <airlied@redhat.com>
>
> This will allow different abstractions later.
>
> Signed-off-by: Dave Airlie <airlied@redhat.com>

Acked-by: Christian König <christian.koenig@amd.com>

> ---
>   drivers/gpu/drm/ttm/ttm_bo.c      | 34 +++++++++++++++----------------
>   drivers/gpu/drm/ttm/ttm_bo_util.c | 20 +++++++++---------
>   drivers/gpu/drm/ttm/ttm_bo_vm.c   |  2 +-
>   include/drm/ttm/ttm_bo_driver.h   |  6 ++++++
>   4 files changed, 34 insertions(+), 28 deletions(-)
>
> diff --git a/drivers/gpu/drm/ttm/ttm_bo.c b/drivers/gpu/drm/ttm/ttm_bo.c
> index ebecb796dd49..8777c323e7de 100644
> --- a/drivers/gpu/drm/ttm/ttm_bo.c
> +++ b/drivers/gpu/drm/ttm/ttm_bo.c
> @@ -108,7 +108,7 @@ static void ttm_bo_mem_space_debug(struct ttm_buffer_object *bo,
>   			return;
>   		drm_printf(&p, "  placement[%d]=0x%08X (%d)\n",
>   			   i, placement->placement[i].flags, mem_type);
> -		man = &bo->bdev->man[mem_type];
> +		man = ttm_manager_type(bo->bdev, mem_type);
>   		ttm_mem_type_manager_debug(man, &p);
>   	}
>   }
> @@ -157,7 +157,7 @@ static void ttm_bo_add_mem_to_lru(struct ttm_buffer_object *bo,
>   	if (mem->placement & TTM_PL_FLAG_NO_EVICT)
>   		return;
>   
> -	man = &bdev->man[mem->mem_type];
> +	man = ttm_manager_type(bdev, mem->mem_type);
>   	list_add_tail(&bo->lru, &man->lru[bo->priority]);
>   
>   	if (man->use_tt && bo->ttm &&
> @@ -232,7 +232,7 @@ void ttm_bo_bulk_move_lru_tail(struct ttm_lru_bulk_move *bulk)
>   		dma_resv_assert_held(pos->first->base.resv);
>   		dma_resv_assert_held(pos->last->base.resv);
>   
> -		man = &pos->first->bdev->man[TTM_PL_TT];
> +		man = ttm_manager_type(pos->first->bdev, TTM_PL_TT);
>   		list_bulk_move_tail(&man->lru[i], &pos->first->lru,
>   				    &pos->last->lru);
>   	}
> @@ -247,7 +247,7 @@ void ttm_bo_bulk_move_lru_tail(struct ttm_lru_bulk_move *bulk)
>   		dma_resv_assert_held(pos->first->base.resv);
>   		dma_resv_assert_held(pos->last->base.resv);
>   
> -		man = &pos->first->bdev->man[TTM_PL_VRAM];
> +		man = ttm_manager_type(pos->first->bdev, TTM_PL_VRAM);
>   		list_bulk_move_tail(&man->lru[i], &pos->first->lru,
>   				    &pos->last->lru);
>   	}
> @@ -273,8 +273,8 @@ static int ttm_bo_handle_move_mem(struct ttm_buffer_object *bo,
>   				  struct ttm_operation_ctx *ctx)
>   {
>   	struct ttm_bo_device *bdev = bo->bdev;
> -	struct ttm_mem_type_manager *old_man = &bdev->man[bo->mem.mem_type];
> -	struct ttm_mem_type_manager *new_man = &bdev->man[mem->mem_type];
> +	struct ttm_mem_type_manager *old_man = ttm_manager_type(bdev, bo->mem.mem_type);
> +	struct ttm_mem_type_manager *new_man = ttm_manager_type(bdev, mem->mem_type);
>   	int ret;
>   
>   	ret = ttm_mem_io_lock(old_man, true);
> @@ -340,7 +340,7 @@ static int ttm_bo_handle_move_mem(struct ttm_buffer_object *bo,
>   	return 0;
>   
>   out_err:
> -	new_man = &bdev->man[bo->mem.mem_type];
> +	new_man = ttm_manager_type(bdev, bo->mem.mem_type);
>   	if (!new_man->use_tt) {
>   		ttm_tt_destroy(bo->ttm);
>   		bo->ttm = NULL;
> @@ -552,7 +552,7 @@ static void ttm_bo_release(struct kref *kref)
>   	struct ttm_buffer_object *bo =
>   	    container_of(kref, struct ttm_buffer_object, kref);
>   	struct ttm_bo_device *bdev = bo->bdev;
> -	struct ttm_mem_type_manager *man = &bdev->man[bo->mem.mem_type];
> +	struct ttm_mem_type_manager *man = ttm_manager_type(bdev, bo->mem.mem_type);
>   	size_t acc_size = bo->acc_size;
>   	int ret;
>   
> @@ -844,7 +844,7 @@ static int ttm_bo_mem_get(struct ttm_buffer_object *bo,
>   			  const struct ttm_place *place,
>   			  struct ttm_mem_reg *mem)
>   {
> -	struct ttm_mem_type_manager *man = &bo->bdev->man[mem->mem_type];
> +	struct ttm_mem_type_manager *man = ttm_manager_type(bo->bdev, mem->mem_type);
>   
>   	mem->mm_node = NULL;
>   	if (!man->func || !man->func->get_node)
> @@ -855,7 +855,7 @@ static int ttm_bo_mem_get(struct ttm_buffer_object *bo,
>   
>   void ttm_bo_mem_put(struct ttm_buffer_object *bo, struct ttm_mem_reg *mem)
>   {
> -	struct ttm_mem_type_manager *man = &bo->bdev->man[mem->mem_type];
> +	struct ttm_mem_type_manager *man = ttm_manager_type(bo->bdev, mem->mem_type);
>   
>   	if (!man->func || !man->func->put_node)
>   		return;
> @@ -912,7 +912,7 @@ static int ttm_bo_mem_force_space(struct ttm_buffer_object *bo,
>   				  struct ttm_operation_ctx *ctx)
>   {
>   	struct ttm_bo_device *bdev = bo->bdev;
> -	struct ttm_mem_type_manager *man = &bdev->man[mem->mem_type];
> +	struct ttm_mem_type_manager *man = ttm_manager_type(bdev, mem->mem_type);
>   	struct ww_acquire_ctx *ticket;
>   	int ret;
>   
> @@ -1002,7 +1002,7 @@ static int ttm_bo_mem_placement(struct ttm_buffer_object *bo,
>   	if (ret)
>   		return ret;
>   
> -	man = &bdev->man[mem_type];
> +	man = ttm_manager_type(bdev, mem_type);
>   	if (!man->has_type || !man->use_type)
>   		return -EBUSY;
>   
> @@ -1065,7 +1065,7 @@ int ttm_bo_mem_space(struct ttm_buffer_object *bo,
>   		if (unlikely(ret))
>   			goto error;
>   
> -		man = &bdev->man[mem->mem_type];
> +		man = ttm_manager_type(bdev, mem->mem_type);
>   		ret = ttm_bo_add_move_fence(bo, man, mem, ctx->no_wait_gpu);
>   		if (unlikely(ret)) {
>   			ttm_bo_mem_put(bo, mem);
> @@ -1455,7 +1455,7 @@ EXPORT_SYMBOL(ttm_mem_type_manager_force_list_clean);
>   
>   int ttm_bo_evict_mm(struct ttm_bo_device *bdev, unsigned mem_type)
>   {
> -	struct ttm_mem_type_manager *man = &bdev->man[mem_type];
> +	struct ttm_mem_type_manager *man = ttm_manager_type(bdev, mem_type);
>   
>   	if (mem_type == 0 || mem_type >= TTM_NUM_MEM_TYPES) {
>   		pr_err("Illegal memory manager memory type %u\n", mem_type);
> @@ -1558,7 +1558,7 @@ int ttm_bo_device_release(struct ttm_bo_device *bdev)
>   	unsigned i;
>   	struct ttm_mem_type_manager *man;
>   
> -	man = &bdev->man[TTM_PL_SYSTEM];
> +	man = ttm_manager_type(bdev, TTM_PL_SYSTEM);
>   	ttm_mem_type_manager_disable(man);
>   
>   	mutex_lock(&ttm_global_mutex);
> @@ -1585,7 +1585,7 @@ EXPORT_SYMBOL(ttm_bo_device_release);
>   
>   static void ttm_bo_init_sysman(struct ttm_bo_device *bdev)
>   {
> -	struct ttm_mem_type_manager *man = &bdev->man[TTM_PL_SYSTEM];
> +	struct ttm_mem_type_manager *man = ttm_manager_type(bdev, TTM_PL_SYSTEM);
>   
>   	/*
>   	 * Initialize the system memory buffer type.
> @@ -1649,7 +1649,7 @@ void ttm_bo_unmap_virtual_locked(struct ttm_buffer_object *bo)
>   void ttm_bo_unmap_virtual(struct ttm_buffer_object *bo)
>   {
>   	struct ttm_bo_device *bdev = bo->bdev;
> -	struct ttm_mem_type_manager *man = &bdev->man[bo->mem.mem_type];
> +	struct ttm_mem_type_manager *man = ttm_manager_type(bdev, bo->mem.mem_type);
>   
>   	ttm_mem_io_lock(man, false);
>   	ttm_bo_unmap_virtual_locked(bo);
> diff --git a/drivers/gpu/drm/ttm/ttm_bo_util.c b/drivers/gpu/drm/ttm/ttm_bo_util.c
> index 1f502be0b646..879c8ded0cd8 100644
> --- a/drivers/gpu/drm/ttm/ttm_bo_util.c
> +++ b/drivers/gpu/drm/ttm/ttm_bo_util.c
> @@ -129,7 +129,7 @@ static int ttm_mem_io_evict(struct ttm_mem_type_manager *man)
>   int ttm_mem_io_reserve(struct ttm_bo_device *bdev,
>   		       struct ttm_mem_reg *mem)
>   {
> -	struct ttm_mem_type_manager *man = &bdev->man[mem->mem_type];
> +	struct ttm_mem_type_manager *man = ttm_manager_type(bdev, mem->mem_type);
>   	int ret;
>   
>   	if (mem->bus.io_reserved_count++)
> @@ -162,7 +162,7 @@ void ttm_mem_io_free(struct ttm_bo_device *bdev,
>   
>   int ttm_mem_io_reserve_vm(struct ttm_buffer_object *bo)
>   {
> -	struct ttm_mem_type_manager *man = &bo->bdev->man[bo->mem.mem_type];
> +	struct ttm_mem_type_manager *man = ttm_manager_type(bo->bdev, bo->mem.mem_type);
>   	struct ttm_mem_reg *mem = &bo->mem;
>   	int ret;
>   
> @@ -195,7 +195,7 @@ static int ttm_mem_reg_ioremap(struct ttm_bo_device *bdev,
>   			       struct ttm_mem_reg *mem,
>   			       void **virtual)
>   {
> -	struct ttm_mem_type_manager *man = &bdev->man[mem->mem_type];
> +	struct ttm_mem_type_manager *man = ttm_manager_type(bdev, mem->mem_type);
>   	int ret;
>   	void *addr;
>   
> @@ -232,7 +232,7 @@ static void ttm_mem_reg_iounmap(struct ttm_bo_device *bdev,
>   {
>   	struct ttm_mem_type_manager *man;
>   
> -	man = &bdev->man[mem->mem_type];
> +	man = ttm_manager_type(bdev, mem->mem_type);
>   
>   	if (virtual && mem->bus.addr == NULL)
>   		iounmap(virtual);
> @@ -303,7 +303,7 @@ int ttm_bo_move_memcpy(struct ttm_buffer_object *bo,
>   		       struct ttm_mem_reg *new_mem)
>   {
>   	struct ttm_bo_device *bdev = bo->bdev;
> -	struct ttm_mem_type_manager *man = &bdev->man[new_mem->mem_type];
> +	struct ttm_mem_type_manager *man = ttm_manager_type(bdev, new_mem->mem_type);
>   	struct ttm_tt *ttm = bo->ttm;
>   	struct ttm_mem_reg *old_mem = &bo->mem;
>   	struct ttm_mem_reg old_copy = *old_mem;
> @@ -571,7 +571,7 @@ int ttm_bo_kmap(struct ttm_buffer_object *bo,
>   		struct ttm_bo_kmap_obj *map)
>   {
>   	struct ttm_mem_type_manager *man =
> -		&bo->bdev->man[bo->mem.mem_type];
> +		ttm_manager_type(bo->bdev, bo->mem.mem_type);
>   	unsigned long offset, size;
>   	int ret;
>   
> @@ -601,7 +601,7 @@ void ttm_bo_kunmap(struct ttm_bo_kmap_obj *map)
>   {
>   	struct ttm_buffer_object *bo = map->bo;
>   	struct ttm_mem_type_manager *man =
> -		&bo->bdev->man[bo->mem.mem_type];
> +		ttm_manager_type(bo->bdev, bo->mem.mem_type);
>   
>   	if (!map->virtual)
>   		return;
> @@ -634,7 +634,7 @@ int ttm_bo_move_accel_cleanup(struct ttm_buffer_object *bo,
>   			      struct ttm_mem_reg *new_mem)
>   {
>   	struct ttm_bo_device *bdev = bo->bdev;
> -	struct ttm_mem_type_manager *man = &bdev->man[new_mem->mem_type];
> +	struct ttm_mem_type_manager *man = ttm_manager_type(bdev, new_mem->mem_type);
>   	struct ttm_mem_reg *old_mem = &bo->mem;
>   	int ret;
>   	struct ttm_buffer_object *ghost_obj;
> @@ -697,8 +697,8 @@ int ttm_bo_pipeline_move(struct ttm_buffer_object *bo,
>   	struct ttm_bo_device *bdev = bo->bdev;
>   	struct ttm_mem_reg *old_mem = &bo->mem;
>   
> -	struct ttm_mem_type_manager *from = &bdev->man[old_mem->mem_type];
> -	struct ttm_mem_type_manager *to = &bdev->man[new_mem->mem_type];
> +	struct ttm_mem_type_manager *from = ttm_manager_type(bdev, old_mem->mem_type);
> +	struct ttm_mem_type_manager *to = ttm_manager_type(bdev, new_mem->mem_type);
>   
>   	int ret;
>   
> diff --git a/drivers/gpu/drm/ttm/ttm_bo_vm.c b/drivers/gpu/drm/ttm/ttm_bo_vm.c
> index 468a0eb9e632..5ae679184eb5 100644
> --- a/drivers/gpu/drm/ttm/ttm_bo_vm.c
> +++ b/drivers/gpu/drm/ttm/ttm_bo_vm.c
> @@ -282,7 +282,7 @@ vm_fault_t ttm_bo_vm_fault_reserved(struct vm_fault *vmf,
>   	vm_fault_t ret = VM_FAULT_NOPAGE;
>   	unsigned long address = vmf->address;
>   	struct ttm_mem_type_manager *man =
> -		&bdev->man[bo->mem.mem_type];
> +		ttm_manager_type(bdev, bo->mem.mem_type);
>   
>   	/*
>   	 * Refuse to fault imported pages. This should be handled
> diff --git a/include/drm/ttm/ttm_bo_driver.h b/include/drm/ttm/ttm_bo_driver.h
> index 8cc39cd55a14..e80deee3ae99 100644
> --- a/include/drm/ttm/ttm_bo_driver.h
> +++ b/include/drm/ttm/ttm_bo_driver.h
> @@ -444,6 +444,12 @@ struct ttm_bo_device {
>   	bool no_retry;
>   };
>   
> +static inline struct ttm_mem_type_manager *ttm_manager_type(struct ttm_bo_device *bdev,
> +							    int mem_type)
> +{
> +	return &bdev->man[mem_type];
> +}
> +
>   /**
>    * struct ttm_lru_bulk_move_pos
>    *
Ben Skeggs Aug. 5, 2020, 5:47 a.m. UTC | #2
On Tue, 4 Aug 2020 at 21:25, Christian König <christian.koenig@amd.com> wrote:
>
> Am 04.08.20 um 04:56 schrieb Dave Airlie:
> > From: Dave Airlie <airlied@redhat.com>
> >
> > This will allow different abstractions later.
> >
> > Signed-off-by: Dave Airlie <airlied@redhat.com>
>
> Acked-by: Christian König <christian.koenig@amd.com>
Reviewed-by: Ben Skeggs <bskeggs@redhat.com>

>
> > ---
> >   drivers/gpu/drm/ttm/ttm_bo.c      | 34 +++++++++++++++----------------
> >   drivers/gpu/drm/ttm/ttm_bo_util.c | 20 +++++++++---------
> >   drivers/gpu/drm/ttm/ttm_bo_vm.c   |  2 +-
> >   include/drm/ttm/ttm_bo_driver.h   |  6 ++++++
> >   4 files changed, 34 insertions(+), 28 deletions(-)
> >
> > diff --git a/drivers/gpu/drm/ttm/ttm_bo.c b/drivers/gpu/drm/ttm/ttm_bo.c
> > index ebecb796dd49..8777c323e7de 100644
> > --- a/drivers/gpu/drm/ttm/ttm_bo.c
> > +++ b/drivers/gpu/drm/ttm/ttm_bo.c
> > @@ -108,7 +108,7 @@ static void ttm_bo_mem_space_debug(struct ttm_buffer_object *bo,
> >                       return;
> >               drm_printf(&p, "  placement[%d]=0x%08X (%d)\n",
> >                          i, placement->placement[i].flags, mem_type);
> > -             man = &bo->bdev->man[mem_type];
> > +             man = ttm_manager_type(bo->bdev, mem_type);
> >               ttm_mem_type_manager_debug(man, &p);
> >       }
> >   }
> > @@ -157,7 +157,7 @@ static void ttm_bo_add_mem_to_lru(struct ttm_buffer_object *bo,
> >       if (mem->placement & TTM_PL_FLAG_NO_EVICT)
> >               return;
> >
> > -     man = &bdev->man[mem->mem_type];
> > +     man = ttm_manager_type(bdev, mem->mem_type);
> >       list_add_tail(&bo->lru, &man->lru[bo->priority]);
> >
> >       if (man->use_tt && bo->ttm &&
> > @@ -232,7 +232,7 @@ void ttm_bo_bulk_move_lru_tail(struct ttm_lru_bulk_move *bulk)
> >               dma_resv_assert_held(pos->first->base.resv);
> >               dma_resv_assert_held(pos->last->base.resv);
> >
> > -             man = &pos->first->bdev->man[TTM_PL_TT];
> > +             man = ttm_manager_type(pos->first->bdev, TTM_PL_TT);
> >               list_bulk_move_tail(&man->lru[i], &pos->first->lru,
> >                                   &pos->last->lru);
> >       }
> > @@ -247,7 +247,7 @@ void ttm_bo_bulk_move_lru_tail(struct ttm_lru_bulk_move *bulk)
> >               dma_resv_assert_held(pos->first->base.resv);
> >               dma_resv_assert_held(pos->last->base.resv);
> >
> > -             man = &pos->first->bdev->man[TTM_PL_VRAM];
> > +             man = ttm_manager_type(pos->first->bdev, TTM_PL_VRAM);
> >               list_bulk_move_tail(&man->lru[i], &pos->first->lru,
> >                                   &pos->last->lru);
> >       }
> > @@ -273,8 +273,8 @@ static int ttm_bo_handle_move_mem(struct ttm_buffer_object *bo,
> >                                 struct ttm_operation_ctx *ctx)
> >   {
> >       struct ttm_bo_device *bdev = bo->bdev;
> > -     struct ttm_mem_type_manager *old_man = &bdev->man[bo->mem.mem_type];
> > -     struct ttm_mem_type_manager *new_man = &bdev->man[mem->mem_type];
> > +     struct ttm_mem_type_manager *old_man = ttm_manager_type(bdev, bo->mem.mem_type);
> > +     struct ttm_mem_type_manager *new_man = ttm_manager_type(bdev, mem->mem_type);
> >       int ret;
> >
> >       ret = ttm_mem_io_lock(old_man, true);
> > @@ -340,7 +340,7 @@ static int ttm_bo_handle_move_mem(struct ttm_buffer_object *bo,
> >       return 0;
> >
> >   out_err:
> > -     new_man = &bdev->man[bo->mem.mem_type];
> > +     new_man = ttm_manager_type(bdev, bo->mem.mem_type);
> >       if (!new_man->use_tt) {
> >               ttm_tt_destroy(bo->ttm);
> >               bo->ttm = NULL;
> > @@ -552,7 +552,7 @@ static void ttm_bo_release(struct kref *kref)
> >       struct ttm_buffer_object *bo =
> >           container_of(kref, struct ttm_buffer_object, kref);
> >       struct ttm_bo_device *bdev = bo->bdev;
> > -     struct ttm_mem_type_manager *man = &bdev->man[bo->mem.mem_type];
> > +     struct ttm_mem_type_manager *man = ttm_manager_type(bdev, bo->mem.mem_type);
> >       size_t acc_size = bo->acc_size;
> >       int ret;
> >
> > @@ -844,7 +844,7 @@ static int ttm_bo_mem_get(struct ttm_buffer_object *bo,
> >                         const struct ttm_place *place,
> >                         struct ttm_mem_reg *mem)
> >   {
> > -     struct ttm_mem_type_manager *man = &bo->bdev->man[mem->mem_type];
> > +     struct ttm_mem_type_manager *man = ttm_manager_type(bo->bdev, mem->mem_type);
> >
> >       mem->mm_node = NULL;
> >       if (!man->func || !man->func->get_node)
> > @@ -855,7 +855,7 @@ static int ttm_bo_mem_get(struct ttm_buffer_object *bo,
> >
> >   void ttm_bo_mem_put(struct ttm_buffer_object *bo, struct ttm_mem_reg *mem)
> >   {
> > -     struct ttm_mem_type_manager *man = &bo->bdev->man[mem->mem_type];
> > +     struct ttm_mem_type_manager *man = ttm_manager_type(bo->bdev, mem->mem_type);
> >
> >       if (!man->func || !man->func->put_node)
> >               return;
> > @@ -912,7 +912,7 @@ static int ttm_bo_mem_force_space(struct ttm_buffer_object *bo,
> >                                 struct ttm_operation_ctx *ctx)
> >   {
> >       struct ttm_bo_device *bdev = bo->bdev;
> > -     struct ttm_mem_type_manager *man = &bdev->man[mem->mem_type];
> > +     struct ttm_mem_type_manager *man = ttm_manager_type(bdev, mem->mem_type);
> >       struct ww_acquire_ctx *ticket;
> >       int ret;
> >
> > @@ -1002,7 +1002,7 @@ static int ttm_bo_mem_placement(struct ttm_buffer_object *bo,
> >       if (ret)
> >               return ret;
> >
> > -     man = &bdev->man[mem_type];
> > +     man = ttm_manager_type(bdev, mem_type);
> >       if (!man->has_type || !man->use_type)
> >               return -EBUSY;
> >
> > @@ -1065,7 +1065,7 @@ int ttm_bo_mem_space(struct ttm_buffer_object *bo,
> >               if (unlikely(ret))
> >                       goto error;
> >
> > -             man = &bdev->man[mem->mem_type];
> > +             man = ttm_manager_type(bdev, mem->mem_type);
> >               ret = ttm_bo_add_move_fence(bo, man, mem, ctx->no_wait_gpu);
> >               if (unlikely(ret)) {
> >                       ttm_bo_mem_put(bo, mem);
> > @@ -1455,7 +1455,7 @@ EXPORT_SYMBOL(ttm_mem_type_manager_force_list_clean);
> >
> >   int ttm_bo_evict_mm(struct ttm_bo_device *bdev, unsigned mem_type)
> >   {
> > -     struct ttm_mem_type_manager *man = &bdev->man[mem_type];
> > +     struct ttm_mem_type_manager *man = ttm_manager_type(bdev, mem_type);
> >
> >       if (mem_type == 0 || mem_type >= TTM_NUM_MEM_TYPES) {
> >               pr_err("Illegal memory manager memory type %u\n", mem_type);
> > @@ -1558,7 +1558,7 @@ int ttm_bo_device_release(struct ttm_bo_device *bdev)
> >       unsigned i;
> >       struct ttm_mem_type_manager *man;
> >
> > -     man = &bdev->man[TTM_PL_SYSTEM];
> > +     man = ttm_manager_type(bdev, TTM_PL_SYSTEM);
> >       ttm_mem_type_manager_disable(man);
> >
> >       mutex_lock(&ttm_global_mutex);
> > @@ -1585,7 +1585,7 @@ EXPORT_SYMBOL(ttm_bo_device_release);
> >
> >   static void ttm_bo_init_sysman(struct ttm_bo_device *bdev)
> >   {
> > -     struct ttm_mem_type_manager *man = &bdev->man[TTM_PL_SYSTEM];
> > +     struct ttm_mem_type_manager *man = ttm_manager_type(bdev, TTM_PL_SYSTEM);
> >
> >       /*
> >        * Initialize the system memory buffer type.
> > @@ -1649,7 +1649,7 @@ void ttm_bo_unmap_virtual_locked(struct ttm_buffer_object *bo)
> >   void ttm_bo_unmap_virtual(struct ttm_buffer_object *bo)
> >   {
> >       struct ttm_bo_device *bdev = bo->bdev;
> > -     struct ttm_mem_type_manager *man = &bdev->man[bo->mem.mem_type];
> > +     struct ttm_mem_type_manager *man = ttm_manager_type(bdev, bo->mem.mem_type);
> >
> >       ttm_mem_io_lock(man, false);
> >       ttm_bo_unmap_virtual_locked(bo);
> > diff --git a/drivers/gpu/drm/ttm/ttm_bo_util.c b/drivers/gpu/drm/ttm/ttm_bo_util.c
> > index 1f502be0b646..879c8ded0cd8 100644
> > --- a/drivers/gpu/drm/ttm/ttm_bo_util.c
> > +++ b/drivers/gpu/drm/ttm/ttm_bo_util.c
> > @@ -129,7 +129,7 @@ static int ttm_mem_io_evict(struct ttm_mem_type_manager *man)
> >   int ttm_mem_io_reserve(struct ttm_bo_device *bdev,
> >                      struct ttm_mem_reg *mem)
> >   {
> > -     struct ttm_mem_type_manager *man = &bdev->man[mem->mem_type];
> > +     struct ttm_mem_type_manager *man = ttm_manager_type(bdev, mem->mem_type);
> >       int ret;
> >
> >       if (mem->bus.io_reserved_count++)
> > @@ -162,7 +162,7 @@ void ttm_mem_io_free(struct ttm_bo_device *bdev,
> >
> >   int ttm_mem_io_reserve_vm(struct ttm_buffer_object *bo)
> >   {
> > -     struct ttm_mem_type_manager *man = &bo->bdev->man[bo->mem.mem_type];
> > +     struct ttm_mem_type_manager *man = ttm_manager_type(bo->bdev, bo->mem.mem_type);
> >       struct ttm_mem_reg *mem = &bo->mem;
> >       int ret;
> >
> > @@ -195,7 +195,7 @@ static int ttm_mem_reg_ioremap(struct ttm_bo_device *bdev,
> >                              struct ttm_mem_reg *mem,
> >                              void **virtual)
> >   {
> > -     struct ttm_mem_type_manager *man = &bdev->man[mem->mem_type];
> > +     struct ttm_mem_type_manager *man = ttm_manager_type(bdev, mem->mem_type);
> >       int ret;
> >       void *addr;
> >
> > @@ -232,7 +232,7 @@ static void ttm_mem_reg_iounmap(struct ttm_bo_device *bdev,
> >   {
> >       struct ttm_mem_type_manager *man;
> >
> > -     man = &bdev->man[mem->mem_type];
> > +     man = ttm_manager_type(bdev, mem->mem_type);
> >
> >       if (virtual && mem->bus.addr == NULL)
> >               iounmap(virtual);
> > @@ -303,7 +303,7 @@ int ttm_bo_move_memcpy(struct ttm_buffer_object *bo,
> >                      struct ttm_mem_reg *new_mem)
> >   {
> >       struct ttm_bo_device *bdev = bo->bdev;
> > -     struct ttm_mem_type_manager *man = &bdev->man[new_mem->mem_type];
> > +     struct ttm_mem_type_manager *man = ttm_manager_type(bdev, new_mem->mem_type);
> >       struct ttm_tt *ttm = bo->ttm;
> >       struct ttm_mem_reg *old_mem = &bo->mem;
> >       struct ttm_mem_reg old_copy = *old_mem;
> > @@ -571,7 +571,7 @@ int ttm_bo_kmap(struct ttm_buffer_object *bo,
> >               struct ttm_bo_kmap_obj *map)
> >   {
> >       struct ttm_mem_type_manager *man =
> > -             &bo->bdev->man[bo->mem.mem_type];
> > +             ttm_manager_type(bo->bdev, bo->mem.mem_type);
> >       unsigned long offset, size;
> >       int ret;
> >
> > @@ -601,7 +601,7 @@ void ttm_bo_kunmap(struct ttm_bo_kmap_obj *map)
> >   {
> >       struct ttm_buffer_object *bo = map->bo;
> >       struct ttm_mem_type_manager *man =
> > -             &bo->bdev->man[bo->mem.mem_type];
> > +             ttm_manager_type(bo->bdev, bo->mem.mem_type);
> >
> >       if (!map->virtual)
> >               return;
> > @@ -634,7 +634,7 @@ int ttm_bo_move_accel_cleanup(struct ttm_buffer_object *bo,
> >                             struct ttm_mem_reg *new_mem)
> >   {
> >       struct ttm_bo_device *bdev = bo->bdev;
> > -     struct ttm_mem_type_manager *man = &bdev->man[new_mem->mem_type];
> > +     struct ttm_mem_type_manager *man = ttm_manager_type(bdev, new_mem->mem_type);
> >       struct ttm_mem_reg *old_mem = &bo->mem;
> >       int ret;
> >       struct ttm_buffer_object *ghost_obj;
> > @@ -697,8 +697,8 @@ int ttm_bo_pipeline_move(struct ttm_buffer_object *bo,
> >       struct ttm_bo_device *bdev = bo->bdev;
> >       struct ttm_mem_reg *old_mem = &bo->mem;
> >
> > -     struct ttm_mem_type_manager *from = &bdev->man[old_mem->mem_type];
> > -     struct ttm_mem_type_manager *to = &bdev->man[new_mem->mem_type];
> > +     struct ttm_mem_type_manager *from = ttm_manager_type(bdev, old_mem->mem_type);
> > +     struct ttm_mem_type_manager *to = ttm_manager_type(bdev, new_mem->mem_type);
> >
> >       int ret;
> >
> > diff --git a/drivers/gpu/drm/ttm/ttm_bo_vm.c b/drivers/gpu/drm/ttm/ttm_bo_vm.c
> > index 468a0eb9e632..5ae679184eb5 100644
> > --- a/drivers/gpu/drm/ttm/ttm_bo_vm.c
> > +++ b/drivers/gpu/drm/ttm/ttm_bo_vm.c
> > @@ -282,7 +282,7 @@ vm_fault_t ttm_bo_vm_fault_reserved(struct vm_fault *vmf,
> >       vm_fault_t ret = VM_FAULT_NOPAGE;
> >       unsigned long address = vmf->address;
> >       struct ttm_mem_type_manager *man =
> > -             &bdev->man[bo->mem.mem_type];
> > +             ttm_manager_type(bdev, bo->mem.mem_type);
> >
> >       /*
> >        * Refuse to fault imported pages. This should be handled
> > diff --git a/include/drm/ttm/ttm_bo_driver.h b/include/drm/ttm/ttm_bo_driver.h
> > index 8cc39cd55a14..e80deee3ae99 100644
> > --- a/include/drm/ttm/ttm_bo_driver.h
> > +++ b/include/drm/ttm/ttm_bo_driver.h
> > @@ -444,6 +444,12 @@ struct ttm_bo_device {
> >       bool no_retry;
> >   };
> >
> > +static inline struct ttm_mem_type_manager *ttm_manager_type(struct ttm_bo_device *bdev,
> > +                                                         int mem_type)
> > +{
> > +     return &bdev->man[mem_type];
> > +}
> > +
> >   /**
> >    * struct ttm_lru_bulk_move_pos
> >    *
>
> _______________________________________________
> dri-devel mailing list
> dri-devel@lists.freedesktop.org
> https://lists.freedesktop.org/mailman/listinfo/dri-devel
diff mbox series

Patch

diff --git a/drivers/gpu/drm/ttm/ttm_bo.c b/drivers/gpu/drm/ttm/ttm_bo.c
index ebecb796dd49..8777c323e7de 100644
--- a/drivers/gpu/drm/ttm/ttm_bo.c
+++ b/drivers/gpu/drm/ttm/ttm_bo.c
@@ -108,7 +108,7 @@  static void ttm_bo_mem_space_debug(struct ttm_buffer_object *bo,
 			return;
 		drm_printf(&p, "  placement[%d]=0x%08X (%d)\n",
 			   i, placement->placement[i].flags, mem_type);
-		man = &bo->bdev->man[mem_type];
+		man = ttm_manager_type(bo->bdev, mem_type);
 		ttm_mem_type_manager_debug(man, &p);
 	}
 }
@@ -157,7 +157,7 @@  static void ttm_bo_add_mem_to_lru(struct ttm_buffer_object *bo,
 	if (mem->placement & TTM_PL_FLAG_NO_EVICT)
 		return;
 
-	man = &bdev->man[mem->mem_type];
+	man = ttm_manager_type(bdev, mem->mem_type);
 	list_add_tail(&bo->lru, &man->lru[bo->priority]);
 
 	if (man->use_tt && bo->ttm &&
@@ -232,7 +232,7 @@  void ttm_bo_bulk_move_lru_tail(struct ttm_lru_bulk_move *bulk)
 		dma_resv_assert_held(pos->first->base.resv);
 		dma_resv_assert_held(pos->last->base.resv);
 
-		man = &pos->first->bdev->man[TTM_PL_TT];
+		man = ttm_manager_type(pos->first->bdev, TTM_PL_TT);
 		list_bulk_move_tail(&man->lru[i], &pos->first->lru,
 				    &pos->last->lru);
 	}
@@ -247,7 +247,7 @@  void ttm_bo_bulk_move_lru_tail(struct ttm_lru_bulk_move *bulk)
 		dma_resv_assert_held(pos->first->base.resv);
 		dma_resv_assert_held(pos->last->base.resv);
 
-		man = &pos->first->bdev->man[TTM_PL_VRAM];
+		man = ttm_manager_type(pos->first->bdev, TTM_PL_VRAM);
 		list_bulk_move_tail(&man->lru[i], &pos->first->lru,
 				    &pos->last->lru);
 	}
@@ -273,8 +273,8 @@  static int ttm_bo_handle_move_mem(struct ttm_buffer_object *bo,
 				  struct ttm_operation_ctx *ctx)
 {
 	struct ttm_bo_device *bdev = bo->bdev;
-	struct ttm_mem_type_manager *old_man = &bdev->man[bo->mem.mem_type];
-	struct ttm_mem_type_manager *new_man = &bdev->man[mem->mem_type];
+	struct ttm_mem_type_manager *old_man = ttm_manager_type(bdev, bo->mem.mem_type);
+	struct ttm_mem_type_manager *new_man = ttm_manager_type(bdev, mem->mem_type);
 	int ret;
 
 	ret = ttm_mem_io_lock(old_man, true);
@@ -340,7 +340,7 @@  static int ttm_bo_handle_move_mem(struct ttm_buffer_object *bo,
 	return 0;
 
 out_err:
-	new_man = &bdev->man[bo->mem.mem_type];
+	new_man = ttm_manager_type(bdev, bo->mem.mem_type);
 	if (!new_man->use_tt) {
 		ttm_tt_destroy(bo->ttm);
 		bo->ttm = NULL;
@@ -552,7 +552,7 @@  static void ttm_bo_release(struct kref *kref)
 	struct ttm_buffer_object *bo =
 	    container_of(kref, struct ttm_buffer_object, kref);
 	struct ttm_bo_device *bdev = bo->bdev;
-	struct ttm_mem_type_manager *man = &bdev->man[bo->mem.mem_type];
+	struct ttm_mem_type_manager *man = ttm_manager_type(bdev, bo->mem.mem_type);
 	size_t acc_size = bo->acc_size;
 	int ret;
 
@@ -844,7 +844,7 @@  static int ttm_bo_mem_get(struct ttm_buffer_object *bo,
 			  const struct ttm_place *place,
 			  struct ttm_mem_reg *mem)
 {
-	struct ttm_mem_type_manager *man = &bo->bdev->man[mem->mem_type];
+	struct ttm_mem_type_manager *man = ttm_manager_type(bo->bdev, mem->mem_type);
 
 	mem->mm_node = NULL;
 	if (!man->func || !man->func->get_node)
@@ -855,7 +855,7 @@  static int ttm_bo_mem_get(struct ttm_buffer_object *bo,
 
 void ttm_bo_mem_put(struct ttm_buffer_object *bo, struct ttm_mem_reg *mem)
 {
-	struct ttm_mem_type_manager *man = &bo->bdev->man[mem->mem_type];
+	struct ttm_mem_type_manager *man = ttm_manager_type(bo->bdev, mem->mem_type);
 
 	if (!man->func || !man->func->put_node)
 		return;
@@ -912,7 +912,7 @@  static int ttm_bo_mem_force_space(struct ttm_buffer_object *bo,
 				  struct ttm_operation_ctx *ctx)
 {
 	struct ttm_bo_device *bdev = bo->bdev;
-	struct ttm_mem_type_manager *man = &bdev->man[mem->mem_type];
+	struct ttm_mem_type_manager *man = ttm_manager_type(bdev, mem->mem_type);
 	struct ww_acquire_ctx *ticket;
 	int ret;
 
@@ -1002,7 +1002,7 @@  static int ttm_bo_mem_placement(struct ttm_buffer_object *bo,
 	if (ret)
 		return ret;
 
-	man = &bdev->man[mem_type];
+	man = ttm_manager_type(bdev, mem_type);
 	if (!man->has_type || !man->use_type)
 		return -EBUSY;
 
@@ -1065,7 +1065,7 @@  int ttm_bo_mem_space(struct ttm_buffer_object *bo,
 		if (unlikely(ret))
 			goto error;
 
-		man = &bdev->man[mem->mem_type];
+		man = ttm_manager_type(bdev, mem->mem_type);
 		ret = ttm_bo_add_move_fence(bo, man, mem, ctx->no_wait_gpu);
 		if (unlikely(ret)) {
 			ttm_bo_mem_put(bo, mem);
@@ -1455,7 +1455,7 @@  EXPORT_SYMBOL(ttm_mem_type_manager_force_list_clean);
 
 int ttm_bo_evict_mm(struct ttm_bo_device *bdev, unsigned mem_type)
 {
-	struct ttm_mem_type_manager *man = &bdev->man[mem_type];
+	struct ttm_mem_type_manager *man = ttm_manager_type(bdev, mem_type);
 
 	if (mem_type == 0 || mem_type >= TTM_NUM_MEM_TYPES) {
 		pr_err("Illegal memory manager memory type %u\n", mem_type);
@@ -1558,7 +1558,7 @@  int ttm_bo_device_release(struct ttm_bo_device *bdev)
 	unsigned i;
 	struct ttm_mem_type_manager *man;
 
-	man = &bdev->man[TTM_PL_SYSTEM];
+	man = ttm_manager_type(bdev, TTM_PL_SYSTEM);
 	ttm_mem_type_manager_disable(man);
 
 	mutex_lock(&ttm_global_mutex);
@@ -1585,7 +1585,7 @@  EXPORT_SYMBOL(ttm_bo_device_release);
 
 static void ttm_bo_init_sysman(struct ttm_bo_device *bdev)
 {
-	struct ttm_mem_type_manager *man = &bdev->man[TTM_PL_SYSTEM];
+	struct ttm_mem_type_manager *man = ttm_manager_type(bdev, TTM_PL_SYSTEM);
 
 	/*
 	 * Initialize the system memory buffer type.
@@ -1649,7 +1649,7 @@  void ttm_bo_unmap_virtual_locked(struct ttm_buffer_object *bo)
 void ttm_bo_unmap_virtual(struct ttm_buffer_object *bo)
 {
 	struct ttm_bo_device *bdev = bo->bdev;
-	struct ttm_mem_type_manager *man = &bdev->man[bo->mem.mem_type];
+	struct ttm_mem_type_manager *man = ttm_manager_type(bdev, bo->mem.mem_type);
 
 	ttm_mem_io_lock(man, false);
 	ttm_bo_unmap_virtual_locked(bo);
diff --git a/drivers/gpu/drm/ttm/ttm_bo_util.c b/drivers/gpu/drm/ttm/ttm_bo_util.c
index 1f502be0b646..879c8ded0cd8 100644
--- a/drivers/gpu/drm/ttm/ttm_bo_util.c
+++ b/drivers/gpu/drm/ttm/ttm_bo_util.c
@@ -129,7 +129,7 @@  static int ttm_mem_io_evict(struct ttm_mem_type_manager *man)
 int ttm_mem_io_reserve(struct ttm_bo_device *bdev,
 		       struct ttm_mem_reg *mem)
 {
-	struct ttm_mem_type_manager *man = &bdev->man[mem->mem_type];
+	struct ttm_mem_type_manager *man = ttm_manager_type(bdev, mem->mem_type);
 	int ret;
 
 	if (mem->bus.io_reserved_count++)
@@ -162,7 +162,7 @@  void ttm_mem_io_free(struct ttm_bo_device *bdev,
 
 int ttm_mem_io_reserve_vm(struct ttm_buffer_object *bo)
 {
-	struct ttm_mem_type_manager *man = &bo->bdev->man[bo->mem.mem_type];
+	struct ttm_mem_type_manager *man = ttm_manager_type(bo->bdev, bo->mem.mem_type);
 	struct ttm_mem_reg *mem = &bo->mem;
 	int ret;
 
@@ -195,7 +195,7 @@  static int ttm_mem_reg_ioremap(struct ttm_bo_device *bdev,
 			       struct ttm_mem_reg *mem,
 			       void **virtual)
 {
-	struct ttm_mem_type_manager *man = &bdev->man[mem->mem_type];
+	struct ttm_mem_type_manager *man = ttm_manager_type(bdev, mem->mem_type);
 	int ret;
 	void *addr;
 
@@ -232,7 +232,7 @@  static void ttm_mem_reg_iounmap(struct ttm_bo_device *bdev,
 {
 	struct ttm_mem_type_manager *man;
 
-	man = &bdev->man[mem->mem_type];
+	man = ttm_manager_type(bdev, mem->mem_type);
 
 	if (virtual && mem->bus.addr == NULL)
 		iounmap(virtual);
@@ -303,7 +303,7 @@  int ttm_bo_move_memcpy(struct ttm_buffer_object *bo,
 		       struct ttm_mem_reg *new_mem)
 {
 	struct ttm_bo_device *bdev = bo->bdev;
-	struct ttm_mem_type_manager *man = &bdev->man[new_mem->mem_type];
+	struct ttm_mem_type_manager *man = ttm_manager_type(bdev, new_mem->mem_type);
 	struct ttm_tt *ttm = bo->ttm;
 	struct ttm_mem_reg *old_mem = &bo->mem;
 	struct ttm_mem_reg old_copy = *old_mem;
@@ -571,7 +571,7 @@  int ttm_bo_kmap(struct ttm_buffer_object *bo,
 		struct ttm_bo_kmap_obj *map)
 {
 	struct ttm_mem_type_manager *man =
-		&bo->bdev->man[bo->mem.mem_type];
+		ttm_manager_type(bo->bdev, bo->mem.mem_type);
 	unsigned long offset, size;
 	int ret;
 
@@ -601,7 +601,7 @@  void ttm_bo_kunmap(struct ttm_bo_kmap_obj *map)
 {
 	struct ttm_buffer_object *bo = map->bo;
 	struct ttm_mem_type_manager *man =
-		&bo->bdev->man[bo->mem.mem_type];
+		ttm_manager_type(bo->bdev, bo->mem.mem_type);
 
 	if (!map->virtual)
 		return;
@@ -634,7 +634,7 @@  int ttm_bo_move_accel_cleanup(struct ttm_buffer_object *bo,
 			      struct ttm_mem_reg *new_mem)
 {
 	struct ttm_bo_device *bdev = bo->bdev;
-	struct ttm_mem_type_manager *man = &bdev->man[new_mem->mem_type];
+	struct ttm_mem_type_manager *man = ttm_manager_type(bdev, new_mem->mem_type);
 	struct ttm_mem_reg *old_mem = &bo->mem;
 	int ret;
 	struct ttm_buffer_object *ghost_obj;
@@ -697,8 +697,8 @@  int ttm_bo_pipeline_move(struct ttm_buffer_object *bo,
 	struct ttm_bo_device *bdev = bo->bdev;
 	struct ttm_mem_reg *old_mem = &bo->mem;
 
-	struct ttm_mem_type_manager *from = &bdev->man[old_mem->mem_type];
-	struct ttm_mem_type_manager *to = &bdev->man[new_mem->mem_type];
+	struct ttm_mem_type_manager *from = ttm_manager_type(bdev, old_mem->mem_type);
+	struct ttm_mem_type_manager *to = ttm_manager_type(bdev, new_mem->mem_type);
 
 	int ret;
 
diff --git a/drivers/gpu/drm/ttm/ttm_bo_vm.c b/drivers/gpu/drm/ttm/ttm_bo_vm.c
index 468a0eb9e632..5ae679184eb5 100644
--- a/drivers/gpu/drm/ttm/ttm_bo_vm.c
+++ b/drivers/gpu/drm/ttm/ttm_bo_vm.c
@@ -282,7 +282,7 @@  vm_fault_t ttm_bo_vm_fault_reserved(struct vm_fault *vmf,
 	vm_fault_t ret = VM_FAULT_NOPAGE;
 	unsigned long address = vmf->address;
 	struct ttm_mem_type_manager *man =
-		&bdev->man[bo->mem.mem_type];
+		ttm_manager_type(bdev, bo->mem.mem_type);
 
 	/*
 	 * Refuse to fault imported pages. This should be handled
diff --git a/include/drm/ttm/ttm_bo_driver.h b/include/drm/ttm/ttm_bo_driver.h
index 8cc39cd55a14..e80deee3ae99 100644
--- a/include/drm/ttm/ttm_bo_driver.h
+++ b/include/drm/ttm/ttm_bo_driver.h
@@ -444,6 +444,12 @@  struct ttm_bo_device {
 	bool no_retry;
 };
 
+static inline struct ttm_mem_type_manager *ttm_manager_type(struct ttm_bo_device *bdev,
+							    int mem_type)
+{
+	return &bdev->man[mem_type];
+}
+
 /**
  * struct ttm_lru_bulk_move_pos
  *