Message ID | 20200731040520.3701599-27-airlied@gmail.com (mailing list archive) |
---|---|
State | New, archived |
Headers | show |
Series | ttm mem manager refactoring. | expand |
Hi Am 31.07.20 um 06:04 schrieb Dave Airlie: > From: Dave Airlie <airlied@redhat.com> > > This will allow different abstractions later. You should consider moving this patch to the beginning of the series, so that patches 1 to 25 can benefit from it. Best regards Thomas > > Signed-off-by: Dave Airlie <airlied@redhat.com> > --- > drivers/gpu/drm/ttm/ttm_bo.c | 34 +++++++++++++++---------------- > drivers/gpu/drm/ttm/ttm_bo_util.c | 20 +++++++++--------- > drivers/gpu/drm/ttm/ttm_bo_vm.c | 2 +- > include/drm/ttm/ttm_bo_driver.h | 6 ++++++ > 4 files changed, 34 insertions(+), 28 deletions(-) > > diff --git a/drivers/gpu/drm/ttm/ttm_bo.c b/drivers/gpu/drm/ttm/ttm_bo.c > index 07c653374f15..7c6389ea067f 100644 > --- a/drivers/gpu/drm/ttm/ttm_bo.c > +++ b/drivers/gpu/drm/ttm/ttm_bo.c > @@ -80,7 +80,7 @@ static inline int ttm_mem_type_from_place(const struct ttm_place *place, > static void ttm_mem_type_debug(struct ttm_bo_device *bdev, struct drm_printer *p, > int mem_type) > { > - struct ttm_mem_type_manager *man = &bdev->man[mem_type]; > + struct ttm_mem_type_manager *man = ttm_manager_type(bdev, mem_type); > > drm_printf(p, " has_type: %d\n", man->has_type); > drm_printf(p, " use_type: %d\n", man->use_type); > @@ -156,7 +156,7 @@ static void ttm_bo_add_mem_to_lru(struct ttm_buffer_object *bo, > if (mem->placement & TTM_PL_FLAG_NO_EVICT) > return; > > - man = &bdev->man[mem->mem_type]; > + man = ttm_manager_type(bdev, mem->mem_type); > list_add_tail(&bo->lru, &man->lru[bo->priority]); > > if (man->use_tt && bo->ttm && > @@ -231,7 +231,7 @@ void ttm_bo_bulk_move_lru_tail(struct ttm_lru_bulk_move *bulk) > dma_resv_assert_held(pos->first->base.resv); > dma_resv_assert_held(pos->last->base.resv); > > - man = &pos->first->bdev->man[TTM_PL_TT]; > + man = ttm_manager_type(pos->first->bdev, TTM_PL_TT); > list_bulk_move_tail(&man->lru[i], &pos->first->lru, > &pos->last->lru); > } > @@ -246,7 +246,7 @@ void ttm_bo_bulk_move_lru_tail(struct ttm_lru_bulk_move *bulk) > dma_resv_assert_held(pos->first->base.resv); > dma_resv_assert_held(pos->last->base.resv); > > - man = &pos->first->bdev->man[TTM_PL_VRAM]; > + man = ttm_manager_type(pos->first->bdev, TTM_PL_VRAM); > list_bulk_move_tail(&man->lru[i], &pos->first->lru, > &pos->last->lru); > } > @@ -272,8 +272,8 @@ static int ttm_bo_handle_move_mem(struct ttm_buffer_object *bo, > struct ttm_operation_ctx *ctx) > { > struct ttm_bo_device *bdev = bo->bdev; > - struct ttm_mem_type_manager *old_man = &bdev->man[bo->mem.mem_type]; > - struct ttm_mem_type_manager *new_man = &bdev->man[mem->mem_type]; > + struct ttm_mem_type_manager *old_man = ttm_manager_type(bdev, bo->mem.mem_type); > + struct ttm_mem_type_manager *new_man = ttm_manager_type(bdev, mem->mem_type); > int ret; > > ret = ttm_mem_io_lock(old_man, true); > @@ -338,7 +338,7 @@ static int ttm_bo_handle_move_mem(struct ttm_buffer_object *bo, > return 0; > > out_err: > - new_man = &bdev->man[bo->mem.mem_type]; > + new_man = ttm_manager_type(bdev, bo->mem.mem_type); > if (!new_man->use_tt) { > ttm_tt_destroy(bo->ttm); > bo->ttm = NULL; > @@ -550,7 +550,7 @@ static void ttm_bo_release(struct kref *kref) > struct ttm_buffer_object *bo = > container_of(kref, struct ttm_buffer_object, kref); > struct ttm_bo_device *bdev = bo->bdev; > - struct ttm_mem_type_manager *man = &bdev->man[bo->mem.mem_type]; > + struct ttm_mem_type_manager *man = ttm_manager_type(bdev, bo->mem.mem_type); > size_t acc_size = bo->acc_size; > int ret; > > @@ -838,7 +838,7 @@ static int ttm_bo_mem_get(struct ttm_buffer_object *bo, > const struct ttm_place *place, > struct ttm_mem_reg *mem) > { > - struct ttm_mem_type_manager *man = &bo->bdev->man[mem->mem_type]; > + struct ttm_mem_type_manager *man = ttm_manager_type(bo->bdev, mem->mem_type); > > mem->mm_node = NULL; > if (!man->func || !man->func->get_node) > @@ -849,7 +849,7 @@ static int ttm_bo_mem_get(struct ttm_buffer_object *bo, > > void ttm_bo_mem_put(struct ttm_buffer_object *bo, struct ttm_mem_reg *mem) > { > - struct ttm_mem_type_manager *man = &bo->bdev->man[mem->mem_type]; > + struct ttm_mem_type_manager *man = ttm_manager_type(bo->bdev, mem->mem_type); > > if (!man->func || !man->func->put_node) > return; > @@ -906,7 +906,7 @@ static int ttm_bo_mem_force_space(struct ttm_buffer_object *bo, > struct ttm_operation_ctx *ctx) > { > struct ttm_bo_device *bdev = bo->bdev; > - struct ttm_mem_type_manager *man = &bdev->man[mem->mem_type]; > + struct ttm_mem_type_manager *man = ttm_manager_type(bdev, mem->mem_type); > struct ww_acquire_ctx *ticket; > int ret; > > @@ -996,7 +996,7 @@ static int ttm_bo_mem_placement(struct ttm_buffer_object *bo, > if (ret) > return ret; > > - man = &bdev->man[mem_type]; > + man = ttm_manager_type(bdev, mem_type); > if (!man->has_type || !man->use_type) > return -EBUSY; > > @@ -1059,7 +1059,7 @@ int ttm_bo_mem_space(struct ttm_buffer_object *bo, > if (unlikely(ret)) > goto error; > > - man = &bdev->man[mem->mem_type]; > + man = ttm_manager_type(bdev, mem->mem_type); > ret = ttm_bo_add_move_fence(bo, man, mem, ctx->no_wait_gpu); > if (unlikely(ret)) { > ttm_bo_mem_put(bo, mem); > @@ -1448,7 +1448,7 @@ EXPORT_SYMBOL(ttm_bo_force_list_clean); > > int ttm_bo_evict_mm(struct ttm_bo_device *bdev, unsigned mem_type) > { > - struct ttm_mem_type_manager *man = &bdev->man[mem_type]; > + struct ttm_mem_type_manager *man = ttm_manager_type(bdev, mem_type); > > if (mem_type == 0 || mem_type >= TTM_NUM_MEM_TYPES) { > pr_err("Illegal memory manager memory type %u\n", mem_type); > @@ -1551,7 +1551,7 @@ int ttm_bo_device_release(struct ttm_bo_device *bdev) > unsigned i; > struct ttm_mem_type_manager *man; > > - man = &bdev->man[TTM_PL_SYSTEM]; > + man = ttm_manager_type(bdev, TTM_PL_SYSTEM); > ttm_bo_disable_mm(man); > > mutex_lock(&ttm_global_mutex); > @@ -1578,7 +1578,7 @@ EXPORT_SYMBOL(ttm_bo_device_release); > > static void ttm_bo_init_sysman(struct ttm_bo_device *bdev) > { > - struct ttm_mem_type_manager *man = &bdev->man[TTM_PL_SYSTEM]; > + struct ttm_mem_type_manager *man = ttm_manager_type(bdev, TTM_PL_SYSTEM); > > /* > * Initialize the system memory buffer type. > @@ -1642,7 +1642,7 @@ void ttm_bo_unmap_virtual_locked(struct ttm_buffer_object *bo) > void ttm_bo_unmap_virtual(struct ttm_buffer_object *bo) > { > struct ttm_bo_device *bdev = bo->bdev; > - struct ttm_mem_type_manager *man = &bdev->man[bo->mem.mem_type]; > + struct ttm_mem_type_manager *man = ttm_manager_type(bdev, bo->mem.mem_type); > > ttm_mem_io_lock(man, false); > ttm_bo_unmap_virtual_locked(bo); > diff --git a/drivers/gpu/drm/ttm/ttm_bo_util.c b/drivers/gpu/drm/ttm/ttm_bo_util.c > index 1f502be0b646..879c8ded0cd8 100644 > --- a/drivers/gpu/drm/ttm/ttm_bo_util.c > +++ b/drivers/gpu/drm/ttm/ttm_bo_util.c > @@ -129,7 +129,7 @@ static int ttm_mem_io_evict(struct ttm_mem_type_manager *man) > int ttm_mem_io_reserve(struct ttm_bo_device *bdev, > struct ttm_mem_reg *mem) > { > - struct ttm_mem_type_manager *man = &bdev->man[mem->mem_type]; > + struct ttm_mem_type_manager *man = ttm_manager_type(bdev, mem->mem_type); > int ret; > > if (mem->bus.io_reserved_count++) > @@ -162,7 +162,7 @@ void ttm_mem_io_free(struct ttm_bo_device *bdev, > > int ttm_mem_io_reserve_vm(struct ttm_buffer_object *bo) > { > - struct ttm_mem_type_manager *man = &bo->bdev->man[bo->mem.mem_type]; > + struct ttm_mem_type_manager *man = ttm_manager_type(bo->bdev, bo->mem.mem_type); > struct ttm_mem_reg *mem = &bo->mem; > int ret; > > @@ -195,7 +195,7 @@ static int ttm_mem_reg_ioremap(struct ttm_bo_device *bdev, > struct ttm_mem_reg *mem, > void **virtual) > { > - struct ttm_mem_type_manager *man = &bdev->man[mem->mem_type]; > + struct ttm_mem_type_manager *man = ttm_manager_type(bdev, mem->mem_type); > int ret; > void *addr; > > @@ -232,7 +232,7 @@ static void ttm_mem_reg_iounmap(struct ttm_bo_device *bdev, > { > struct ttm_mem_type_manager *man; > > - man = &bdev->man[mem->mem_type]; > + man = ttm_manager_type(bdev, mem->mem_type); > > if (virtual && mem->bus.addr == NULL) > iounmap(virtual); > @@ -303,7 +303,7 @@ int ttm_bo_move_memcpy(struct ttm_buffer_object *bo, > struct ttm_mem_reg *new_mem) > { > struct ttm_bo_device *bdev = bo->bdev; > - struct ttm_mem_type_manager *man = &bdev->man[new_mem->mem_type]; > + struct ttm_mem_type_manager *man = ttm_manager_type(bdev, new_mem->mem_type); > struct ttm_tt *ttm = bo->ttm; > struct ttm_mem_reg *old_mem = &bo->mem; > struct ttm_mem_reg old_copy = *old_mem; > @@ -571,7 +571,7 @@ int ttm_bo_kmap(struct ttm_buffer_object *bo, > struct ttm_bo_kmap_obj *map) > { > struct ttm_mem_type_manager *man = > - &bo->bdev->man[bo->mem.mem_type]; > + ttm_manager_type(bo->bdev, bo->mem.mem_type); > unsigned long offset, size; > int ret; > > @@ -601,7 +601,7 @@ void ttm_bo_kunmap(struct ttm_bo_kmap_obj *map) > { > struct ttm_buffer_object *bo = map->bo; > struct ttm_mem_type_manager *man = > - &bo->bdev->man[bo->mem.mem_type]; > + ttm_manager_type(bo->bdev, bo->mem.mem_type); > > if (!map->virtual) > return; > @@ -634,7 +634,7 @@ int ttm_bo_move_accel_cleanup(struct ttm_buffer_object *bo, > struct ttm_mem_reg *new_mem) > { > struct ttm_bo_device *bdev = bo->bdev; > - struct ttm_mem_type_manager *man = &bdev->man[new_mem->mem_type]; > + struct ttm_mem_type_manager *man = ttm_manager_type(bdev, new_mem->mem_type); > struct ttm_mem_reg *old_mem = &bo->mem; > int ret; > struct ttm_buffer_object *ghost_obj; > @@ -697,8 +697,8 @@ int ttm_bo_pipeline_move(struct ttm_buffer_object *bo, > struct ttm_bo_device *bdev = bo->bdev; > struct ttm_mem_reg *old_mem = &bo->mem; > > - struct ttm_mem_type_manager *from = &bdev->man[old_mem->mem_type]; > - struct ttm_mem_type_manager *to = &bdev->man[new_mem->mem_type]; > + struct ttm_mem_type_manager *from = ttm_manager_type(bdev, old_mem->mem_type); > + struct ttm_mem_type_manager *to = ttm_manager_type(bdev, new_mem->mem_type); > > int ret; > > diff --git a/drivers/gpu/drm/ttm/ttm_bo_vm.c b/drivers/gpu/drm/ttm/ttm_bo_vm.c > index 468a0eb9e632..5ae679184eb5 100644 > --- a/drivers/gpu/drm/ttm/ttm_bo_vm.c > +++ b/drivers/gpu/drm/ttm/ttm_bo_vm.c > @@ -282,7 +282,7 @@ vm_fault_t ttm_bo_vm_fault_reserved(struct vm_fault *vmf, > vm_fault_t ret = VM_FAULT_NOPAGE; > unsigned long address = vmf->address; > struct ttm_mem_type_manager *man = > - &bdev->man[bo->mem.mem_type]; > + ttm_manager_type(bdev, bo->mem.mem_type); > > /* > * Refuse to fault imported pages. This should be handled > diff --git a/include/drm/ttm/ttm_bo_driver.h b/include/drm/ttm/ttm_bo_driver.h > index 9d066529ca61..ec25451b503f 100644 > --- a/include/drm/ttm/ttm_bo_driver.h > +++ b/include/drm/ttm/ttm_bo_driver.h > @@ -444,6 +444,12 @@ struct ttm_bo_device { > bool no_retry; > }; > > +static inline struct ttm_mem_type_manager *ttm_manager_type(struct ttm_bo_device *bdev, > + int mem_type) > +{ > + return &bdev->man[mem_type]; > +} > + > /** > * struct ttm_lru_bulk_move_pos > * >
On Fri, 31 Jul 2020 at 17:14, Thomas Zimmermann <tzimmermann@suse.de> wrote: > > Hi > > Am 31.07.20 um 06:04 schrieb Dave Airlie: > > From: Dave Airlie <airlied@redhat.com> > > > > This will allow different abstractions later. > > You should consider moving this patch to the beginning of the series, so > that patches 1 to 25 can benefit from it. I did consider it, but I'd have to move all the follow on patches as well, and it got messy in rebase land, and I started introducing errors, so I left it alone, and it's not necessary until the patch that changes it's definition anyways. Dave.
On Fri, Jul 31, 2020 at 05:21:38PM +1000, Dave Airlie wrote: > On Fri, 31 Jul 2020 at 17:14, Thomas Zimmermann <tzimmermann@suse.de> wrote: > > > > Hi > > > > Am 31.07.20 um 06:04 schrieb Dave Airlie: > > > From: Dave Airlie <airlied@redhat.com> > > > > > > This will allow different abstractions later. > > > > You should consider moving this patch to the beginning of the series, so > > that patches 1 to 25 can benefit from it. > > I did consider it, but I'd have to move all the follow on patches as > well, and it got messy in rebase land, and I started introducing > errors, so I left it alone, and it's not necessary until the patch > that changes it's definition anyways. I like this ordering a lot better, it gives a clear separation between the different steps. Some code gets touched multiple times, but interleaving the demidlayering would only make this worse I think. -Daniel > > Dave. > _______________________________________________ > dri-devel mailing list > dri-devel@lists.freedesktop.org > https://lists.freedesktop.org/mailman/listinfo/dri-devel
Am 31.07.20 um 06:04 schrieb Dave Airlie: > From: Dave Airlie <airlied@redhat.com> > > This will allow different abstractions later. Mhm, I'm questioning if this is really worth it and not just overkill. Running "sed -i 's/&bdev->man/bdev->man/'" on all drivers when we make the switch to a pointer should do it as well. Christian. > > Signed-off-by: Dave Airlie <airlied@redhat.com> > --- > drivers/gpu/drm/ttm/ttm_bo.c | 34 +++++++++++++++---------------- > drivers/gpu/drm/ttm/ttm_bo_util.c | 20 +++++++++--------- > drivers/gpu/drm/ttm/ttm_bo_vm.c | 2 +- > include/drm/ttm/ttm_bo_driver.h | 6 ++++++ > 4 files changed, 34 insertions(+), 28 deletions(-) > > diff --git a/drivers/gpu/drm/ttm/ttm_bo.c b/drivers/gpu/drm/ttm/ttm_bo.c > index 07c653374f15..7c6389ea067f 100644 > --- a/drivers/gpu/drm/ttm/ttm_bo.c > +++ b/drivers/gpu/drm/ttm/ttm_bo.c > @@ -80,7 +80,7 @@ static inline int ttm_mem_type_from_place(const struct ttm_place *place, > static void ttm_mem_type_debug(struct ttm_bo_device *bdev, struct drm_printer *p, > int mem_type) > { > - struct ttm_mem_type_manager *man = &bdev->man[mem_type]; > + struct ttm_mem_type_manager *man = ttm_manager_type(bdev, mem_type); > > drm_printf(p, " has_type: %d\n", man->has_type); > drm_printf(p, " use_type: %d\n", man->use_type); > @@ -156,7 +156,7 @@ static void ttm_bo_add_mem_to_lru(struct ttm_buffer_object *bo, > if (mem->placement & TTM_PL_FLAG_NO_EVICT) > return; > > - man = &bdev->man[mem->mem_type]; > + man = ttm_manager_type(bdev, mem->mem_type); > list_add_tail(&bo->lru, &man->lru[bo->priority]); > > if (man->use_tt && bo->ttm && > @@ -231,7 +231,7 @@ void ttm_bo_bulk_move_lru_tail(struct ttm_lru_bulk_move *bulk) > dma_resv_assert_held(pos->first->base.resv); > dma_resv_assert_held(pos->last->base.resv); > > - man = &pos->first->bdev->man[TTM_PL_TT]; > + man = ttm_manager_type(pos->first->bdev, TTM_PL_TT); > list_bulk_move_tail(&man->lru[i], &pos->first->lru, > &pos->last->lru); > } > @@ -246,7 +246,7 @@ void ttm_bo_bulk_move_lru_tail(struct ttm_lru_bulk_move *bulk) > dma_resv_assert_held(pos->first->base.resv); > dma_resv_assert_held(pos->last->base.resv); > > - man = &pos->first->bdev->man[TTM_PL_VRAM]; > + man = ttm_manager_type(pos->first->bdev, TTM_PL_VRAM); > list_bulk_move_tail(&man->lru[i], &pos->first->lru, > &pos->last->lru); > } > @@ -272,8 +272,8 @@ static int ttm_bo_handle_move_mem(struct ttm_buffer_object *bo, > struct ttm_operation_ctx *ctx) > { > struct ttm_bo_device *bdev = bo->bdev; > - struct ttm_mem_type_manager *old_man = &bdev->man[bo->mem.mem_type]; > - struct ttm_mem_type_manager *new_man = &bdev->man[mem->mem_type]; > + struct ttm_mem_type_manager *old_man = ttm_manager_type(bdev, bo->mem.mem_type); > + struct ttm_mem_type_manager *new_man = ttm_manager_type(bdev, mem->mem_type); > int ret; > > ret = ttm_mem_io_lock(old_man, true); > @@ -338,7 +338,7 @@ static int ttm_bo_handle_move_mem(struct ttm_buffer_object *bo, > return 0; > > out_err: > - new_man = &bdev->man[bo->mem.mem_type]; > + new_man = ttm_manager_type(bdev, bo->mem.mem_type); > if (!new_man->use_tt) { > ttm_tt_destroy(bo->ttm); > bo->ttm = NULL; > @@ -550,7 +550,7 @@ static void ttm_bo_release(struct kref *kref) > struct ttm_buffer_object *bo = > container_of(kref, struct ttm_buffer_object, kref); > struct ttm_bo_device *bdev = bo->bdev; > - struct ttm_mem_type_manager *man = &bdev->man[bo->mem.mem_type]; > + struct ttm_mem_type_manager *man = ttm_manager_type(bdev, bo->mem.mem_type); > size_t acc_size = bo->acc_size; > int ret; > > @@ -838,7 +838,7 @@ static int ttm_bo_mem_get(struct ttm_buffer_object *bo, > const struct ttm_place *place, > struct ttm_mem_reg *mem) > { > - struct ttm_mem_type_manager *man = &bo->bdev->man[mem->mem_type]; > + struct ttm_mem_type_manager *man = ttm_manager_type(bo->bdev, mem->mem_type); > > mem->mm_node = NULL; > if (!man->func || !man->func->get_node) > @@ -849,7 +849,7 @@ static int ttm_bo_mem_get(struct ttm_buffer_object *bo, > > void ttm_bo_mem_put(struct ttm_buffer_object *bo, struct ttm_mem_reg *mem) > { > - struct ttm_mem_type_manager *man = &bo->bdev->man[mem->mem_type]; > + struct ttm_mem_type_manager *man = ttm_manager_type(bo->bdev, mem->mem_type); > > if (!man->func || !man->func->put_node) > return; > @@ -906,7 +906,7 @@ static int ttm_bo_mem_force_space(struct ttm_buffer_object *bo, > struct ttm_operation_ctx *ctx) > { > struct ttm_bo_device *bdev = bo->bdev; > - struct ttm_mem_type_manager *man = &bdev->man[mem->mem_type]; > + struct ttm_mem_type_manager *man = ttm_manager_type(bdev, mem->mem_type); > struct ww_acquire_ctx *ticket; > int ret; > > @@ -996,7 +996,7 @@ static int ttm_bo_mem_placement(struct ttm_buffer_object *bo, > if (ret) > return ret; > > - man = &bdev->man[mem_type]; > + man = ttm_manager_type(bdev, mem_type); > if (!man->has_type || !man->use_type) > return -EBUSY; > > @@ -1059,7 +1059,7 @@ int ttm_bo_mem_space(struct ttm_buffer_object *bo, > if (unlikely(ret)) > goto error; > > - man = &bdev->man[mem->mem_type]; > + man = ttm_manager_type(bdev, mem->mem_type); > ret = ttm_bo_add_move_fence(bo, man, mem, ctx->no_wait_gpu); > if (unlikely(ret)) { > ttm_bo_mem_put(bo, mem); > @@ -1448,7 +1448,7 @@ EXPORT_SYMBOL(ttm_bo_force_list_clean); > > int ttm_bo_evict_mm(struct ttm_bo_device *bdev, unsigned mem_type) > { > - struct ttm_mem_type_manager *man = &bdev->man[mem_type]; > + struct ttm_mem_type_manager *man = ttm_manager_type(bdev, mem_type); > > if (mem_type == 0 || mem_type >= TTM_NUM_MEM_TYPES) { > pr_err("Illegal memory manager memory type %u\n", mem_type); > @@ -1551,7 +1551,7 @@ int ttm_bo_device_release(struct ttm_bo_device *bdev) > unsigned i; > struct ttm_mem_type_manager *man; > > - man = &bdev->man[TTM_PL_SYSTEM]; > + man = ttm_manager_type(bdev, TTM_PL_SYSTEM); > ttm_bo_disable_mm(man); > > mutex_lock(&ttm_global_mutex); > @@ -1578,7 +1578,7 @@ EXPORT_SYMBOL(ttm_bo_device_release); > > static void ttm_bo_init_sysman(struct ttm_bo_device *bdev) > { > - struct ttm_mem_type_manager *man = &bdev->man[TTM_PL_SYSTEM]; > + struct ttm_mem_type_manager *man = ttm_manager_type(bdev, TTM_PL_SYSTEM); > > /* > * Initialize the system memory buffer type. > @@ -1642,7 +1642,7 @@ void ttm_bo_unmap_virtual_locked(struct ttm_buffer_object *bo) > void ttm_bo_unmap_virtual(struct ttm_buffer_object *bo) > { > struct ttm_bo_device *bdev = bo->bdev; > - struct ttm_mem_type_manager *man = &bdev->man[bo->mem.mem_type]; > + struct ttm_mem_type_manager *man = ttm_manager_type(bdev, bo->mem.mem_type); > > ttm_mem_io_lock(man, false); > ttm_bo_unmap_virtual_locked(bo); > diff --git a/drivers/gpu/drm/ttm/ttm_bo_util.c b/drivers/gpu/drm/ttm/ttm_bo_util.c > index 1f502be0b646..879c8ded0cd8 100644 > --- a/drivers/gpu/drm/ttm/ttm_bo_util.c > +++ b/drivers/gpu/drm/ttm/ttm_bo_util.c > @@ -129,7 +129,7 @@ static int ttm_mem_io_evict(struct ttm_mem_type_manager *man) > int ttm_mem_io_reserve(struct ttm_bo_device *bdev, > struct ttm_mem_reg *mem) > { > - struct ttm_mem_type_manager *man = &bdev->man[mem->mem_type]; > + struct ttm_mem_type_manager *man = ttm_manager_type(bdev, mem->mem_type); > int ret; > > if (mem->bus.io_reserved_count++) > @@ -162,7 +162,7 @@ void ttm_mem_io_free(struct ttm_bo_device *bdev, > > int ttm_mem_io_reserve_vm(struct ttm_buffer_object *bo) > { > - struct ttm_mem_type_manager *man = &bo->bdev->man[bo->mem.mem_type]; > + struct ttm_mem_type_manager *man = ttm_manager_type(bo->bdev, bo->mem.mem_type); > struct ttm_mem_reg *mem = &bo->mem; > int ret; > > @@ -195,7 +195,7 @@ static int ttm_mem_reg_ioremap(struct ttm_bo_device *bdev, > struct ttm_mem_reg *mem, > void **virtual) > { > - struct ttm_mem_type_manager *man = &bdev->man[mem->mem_type]; > + struct ttm_mem_type_manager *man = ttm_manager_type(bdev, mem->mem_type); > int ret; > void *addr; > > @@ -232,7 +232,7 @@ static void ttm_mem_reg_iounmap(struct ttm_bo_device *bdev, > { > struct ttm_mem_type_manager *man; > > - man = &bdev->man[mem->mem_type]; > + man = ttm_manager_type(bdev, mem->mem_type); > > if (virtual && mem->bus.addr == NULL) > iounmap(virtual); > @@ -303,7 +303,7 @@ int ttm_bo_move_memcpy(struct ttm_buffer_object *bo, > struct ttm_mem_reg *new_mem) > { > struct ttm_bo_device *bdev = bo->bdev; > - struct ttm_mem_type_manager *man = &bdev->man[new_mem->mem_type]; > + struct ttm_mem_type_manager *man = ttm_manager_type(bdev, new_mem->mem_type); > struct ttm_tt *ttm = bo->ttm; > struct ttm_mem_reg *old_mem = &bo->mem; > struct ttm_mem_reg old_copy = *old_mem; > @@ -571,7 +571,7 @@ int ttm_bo_kmap(struct ttm_buffer_object *bo, > struct ttm_bo_kmap_obj *map) > { > struct ttm_mem_type_manager *man = > - &bo->bdev->man[bo->mem.mem_type]; > + ttm_manager_type(bo->bdev, bo->mem.mem_type); > unsigned long offset, size; > int ret; > > @@ -601,7 +601,7 @@ void ttm_bo_kunmap(struct ttm_bo_kmap_obj *map) > { > struct ttm_buffer_object *bo = map->bo; > struct ttm_mem_type_manager *man = > - &bo->bdev->man[bo->mem.mem_type]; > + ttm_manager_type(bo->bdev, bo->mem.mem_type); > > if (!map->virtual) > return; > @@ -634,7 +634,7 @@ int ttm_bo_move_accel_cleanup(struct ttm_buffer_object *bo, > struct ttm_mem_reg *new_mem) > { > struct ttm_bo_device *bdev = bo->bdev; > - struct ttm_mem_type_manager *man = &bdev->man[new_mem->mem_type]; > + struct ttm_mem_type_manager *man = ttm_manager_type(bdev, new_mem->mem_type); > struct ttm_mem_reg *old_mem = &bo->mem; > int ret; > struct ttm_buffer_object *ghost_obj; > @@ -697,8 +697,8 @@ int ttm_bo_pipeline_move(struct ttm_buffer_object *bo, > struct ttm_bo_device *bdev = bo->bdev; > struct ttm_mem_reg *old_mem = &bo->mem; > > - struct ttm_mem_type_manager *from = &bdev->man[old_mem->mem_type]; > - struct ttm_mem_type_manager *to = &bdev->man[new_mem->mem_type]; > + struct ttm_mem_type_manager *from = ttm_manager_type(bdev, old_mem->mem_type); > + struct ttm_mem_type_manager *to = ttm_manager_type(bdev, new_mem->mem_type); > > int ret; > > diff --git a/drivers/gpu/drm/ttm/ttm_bo_vm.c b/drivers/gpu/drm/ttm/ttm_bo_vm.c > index 468a0eb9e632..5ae679184eb5 100644 > --- a/drivers/gpu/drm/ttm/ttm_bo_vm.c > +++ b/drivers/gpu/drm/ttm/ttm_bo_vm.c > @@ -282,7 +282,7 @@ vm_fault_t ttm_bo_vm_fault_reserved(struct vm_fault *vmf, > vm_fault_t ret = VM_FAULT_NOPAGE; > unsigned long address = vmf->address; > struct ttm_mem_type_manager *man = > - &bdev->man[bo->mem.mem_type]; > + ttm_manager_type(bdev, bo->mem.mem_type); > > /* > * Refuse to fault imported pages. This should be handled > diff --git a/include/drm/ttm/ttm_bo_driver.h b/include/drm/ttm/ttm_bo_driver.h > index 9d066529ca61..ec25451b503f 100644 > --- a/include/drm/ttm/ttm_bo_driver.h > +++ b/include/drm/ttm/ttm_bo_driver.h > @@ -444,6 +444,12 @@ struct ttm_bo_device { > bool no_retry; > }; > > +static inline struct ttm_mem_type_manager *ttm_manager_type(struct ttm_bo_device *bdev, > + int mem_type) > +{ > + return &bdev->man[mem_type]; > +} > + > /** > * struct ttm_lru_bulk_move_pos > *
On Fri, 31 Jul 2020 at 23:23, Christian König <christian.koenig@amd.com> wrote: > > Am 31.07.20 um 06:04 schrieb Dave Airlie: > > From: Dave Airlie <airlied@redhat.com> > > > > This will allow different abstractions later. > > Mhm, I'm questioning if this is really worth it and not just overkill. > > Running "sed -i 's/&bdev->man/bdev->man/'" on all drivers when we make > the switch to a pointer should do it as well. > It's a good place to hook debugging in if you get something rogue accessing unregistered bdev->man ptrs as well. I'd rather not sed the interface later, it's nice to transition things cleanly between states. Dave.
diff --git a/drivers/gpu/drm/ttm/ttm_bo.c b/drivers/gpu/drm/ttm/ttm_bo.c index 07c653374f15..7c6389ea067f 100644 --- a/drivers/gpu/drm/ttm/ttm_bo.c +++ b/drivers/gpu/drm/ttm/ttm_bo.c @@ -80,7 +80,7 @@ static inline int ttm_mem_type_from_place(const struct ttm_place *place, static void ttm_mem_type_debug(struct ttm_bo_device *bdev, struct drm_printer *p, int mem_type) { - struct ttm_mem_type_manager *man = &bdev->man[mem_type]; + struct ttm_mem_type_manager *man = ttm_manager_type(bdev, mem_type); drm_printf(p, " has_type: %d\n", man->has_type); drm_printf(p, " use_type: %d\n", man->use_type); @@ -156,7 +156,7 @@ static void ttm_bo_add_mem_to_lru(struct ttm_buffer_object *bo, if (mem->placement & TTM_PL_FLAG_NO_EVICT) return; - man = &bdev->man[mem->mem_type]; + man = ttm_manager_type(bdev, mem->mem_type); list_add_tail(&bo->lru, &man->lru[bo->priority]); if (man->use_tt && bo->ttm && @@ -231,7 +231,7 @@ void ttm_bo_bulk_move_lru_tail(struct ttm_lru_bulk_move *bulk) dma_resv_assert_held(pos->first->base.resv); dma_resv_assert_held(pos->last->base.resv); - man = &pos->first->bdev->man[TTM_PL_TT]; + man = ttm_manager_type(pos->first->bdev, TTM_PL_TT); list_bulk_move_tail(&man->lru[i], &pos->first->lru, &pos->last->lru); } @@ -246,7 +246,7 @@ void ttm_bo_bulk_move_lru_tail(struct ttm_lru_bulk_move *bulk) dma_resv_assert_held(pos->first->base.resv); dma_resv_assert_held(pos->last->base.resv); - man = &pos->first->bdev->man[TTM_PL_VRAM]; + man = ttm_manager_type(pos->first->bdev, TTM_PL_VRAM); list_bulk_move_tail(&man->lru[i], &pos->first->lru, &pos->last->lru); } @@ -272,8 +272,8 @@ static int ttm_bo_handle_move_mem(struct ttm_buffer_object *bo, struct ttm_operation_ctx *ctx) { struct ttm_bo_device *bdev = bo->bdev; - struct ttm_mem_type_manager *old_man = &bdev->man[bo->mem.mem_type]; - struct ttm_mem_type_manager *new_man = &bdev->man[mem->mem_type]; + struct ttm_mem_type_manager *old_man = ttm_manager_type(bdev, bo->mem.mem_type); + struct ttm_mem_type_manager *new_man = ttm_manager_type(bdev, mem->mem_type); int ret; ret = ttm_mem_io_lock(old_man, true); @@ -338,7 +338,7 @@ static int ttm_bo_handle_move_mem(struct ttm_buffer_object *bo, return 0; out_err: - new_man = &bdev->man[bo->mem.mem_type]; + new_man = ttm_manager_type(bdev, bo->mem.mem_type); if (!new_man->use_tt) { ttm_tt_destroy(bo->ttm); bo->ttm = NULL; @@ -550,7 +550,7 @@ static void ttm_bo_release(struct kref *kref) struct ttm_buffer_object *bo = container_of(kref, struct ttm_buffer_object, kref); struct ttm_bo_device *bdev = bo->bdev; - struct ttm_mem_type_manager *man = &bdev->man[bo->mem.mem_type]; + struct ttm_mem_type_manager *man = ttm_manager_type(bdev, bo->mem.mem_type); size_t acc_size = bo->acc_size; int ret; @@ -838,7 +838,7 @@ static int ttm_bo_mem_get(struct ttm_buffer_object *bo, const struct ttm_place *place, struct ttm_mem_reg *mem) { - struct ttm_mem_type_manager *man = &bo->bdev->man[mem->mem_type]; + struct ttm_mem_type_manager *man = ttm_manager_type(bo->bdev, mem->mem_type); mem->mm_node = NULL; if (!man->func || !man->func->get_node) @@ -849,7 +849,7 @@ static int ttm_bo_mem_get(struct ttm_buffer_object *bo, void ttm_bo_mem_put(struct ttm_buffer_object *bo, struct ttm_mem_reg *mem) { - struct ttm_mem_type_manager *man = &bo->bdev->man[mem->mem_type]; + struct ttm_mem_type_manager *man = ttm_manager_type(bo->bdev, mem->mem_type); if (!man->func || !man->func->put_node) return; @@ -906,7 +906,7 @@ static int ttm_bo_mem_force_space(struct ttm_buffer_object *bo, struct ttm_operation_ctx *ctx) { struct ttm_bo_device *bdev = bo->bdev; - struct ttm_mem_type_manager *man = &bdev->man[mem->mem_type]; + struct ttm_mem_type_manager *man = ttm_manager_type(bdev, mem->mem_type); struct ww_acquire_ctx *ticket; int ret; @@ -996,7 +996,7 @@ static int ttm_bo_mem_placement(struct ttm_buffer_object *bo, if (ret) return ret; - man = &bdev->man[mem_type]; + man = ttm_manager_type(bdev, mem_type); if (!man->has_type || !man->use_type) return -EBUSY; @@ -1059,7 +1059,7 @@ int ttm_bo_mem_space(struct ttm_buffer_object *bo, if (unlikely(ret)) goto error; - man = &bdev->man[mem->mem_type]; + man = ttm_manager_type(bdev, mem->mem_type); ret = ttm_bo_add_move_fence(bo, man, mem, ctx->no_wait_gpu); if (unlikely(ret)) { ttm_bo_mem_put(bo, mem); @@ -1448,7 +1448,7 @@ EXPORT_SYMBOL(ttm_bo_force_list_clean); int ttm_bo_evict_mm(struct ttm_bo_device *bdev, unsigned mem_type) { - struct ttm_mem_type_manager *man = &bdev->man[mem_type]; + struct ttm_mem_type_manager *man = ttm_manager_type(bdev, mem_type); if (mem_type == 0 || mem_type >= TTM_NUM_MEM_TYPES) { pr_err("Illegal memory manager memory type %u\n", mem_type); @@ -1551,7 +1551,7 @@ int ttm_bo_device_release(struct ttm_bo_device *bdev) unsigned i; struct ttm_mem_type_manager *man; - man = &bdev->man[TTM_PL_SYSTEM]; + man = ttm_manager_type(bdev, TTM_PL_SYSTEM); ttm_bo_disable_mm(man); mutex_lock(&ttm_global_mutex); @@ -1578,7 +1578,7 @@ EXPORT_SYMBOL(ttm_bo_device_release); static void ttm_bo_init_sysman(struct ttm_bo_device *bdev) { - struct ttm_mem_type_manager *man = &bdev->man[TTM_PL_SYSTEM]; + struct ttm_mem_type_manager *man = ttm_manager_type(bdev, TTM_PL_SYSTEM); /* * Initialize the system memory buffer type. @@ -1642,7 +1642,7 @@ void ttm_bo_unmap_virtual_locked(struct ttm_buffer_object *bo) void ttm_bo_unmap_virtual(struct ttm_buffer_object *bo) { struct ttm_bo_device *bdev = bo->bdev; - struct ttm_mem_type_manager *man = &bdev->man[bo->mem.mem_type]; + struct ttm_mem_type_manager *man = ttm_manager_type(bdev, bo->mem.mem_type); ttm_mem_io_lock(man, false); ttm_bo_unmap_virtual_locked(bo); diff --git a/drivers/gpu/drm/ttm/ttm_bo_util.c b/drivers/gpu/drm/ttm/ttm_bo_util.c index 1f502be0b646..879c8ded0cd8 100644 --- a/drivers/gpu/drm/ttm/ttm_bo_util.c +++ b/drivers/gpu/drm/ttm/ttm_bo_util.c @@ -129,7 +129,7 @@ static int ttm_mem_io_evict(struct ttm_mem_type_manager *man) int ttm_mem_io_reserve(struct ttm_bo_device *bdev, struct ttm_mem_reg *mem) { - struct ttm_mem_type_manager *man = &bdev->man[mem->mem_type]; + struct ttm_mem_type_manager *man = ttm_manager_type(bdev, mem->mem_type); int ret; if (mem->bus.io_reserved_count++) @@ -162,7 +162,7 @@ void ttm_mem_io_free(struct ttm_bo_device *bdev, int ttm_mem_io_reserve_vm(struct ttm_buffer_object *bo) { - struct ttm_mem_type_manager *man = &bo->bdev->man[bo->mem.mem_type]; + struct ttm_mem_type_manager *man = ttm_manager_type(bo->bdev, bo->mem.mem_type); struct ttm_mem_reg *mem = &bo->mem; int ret; @@ -195,7 +195,7 @@ static int ttm_mem_reg_ioremap(struct ttm_bo_device *bdev, struct ttm_mem_reg *mem, void **virtual) { - struct ttm_mem_type_manager *man = &bdev->man[mem->mem_type]; + struct ttm_mem_type_manager *man = ttm_manager_type(bdev, mem->mem_type); int ret; void *addr; @@ -232,7 +232,7 @@ static void ttm_mem_reg_iounmap(struct ttm_bo_device *bdev, { struct ttm_mem_type_manager *man; - man = &bdev->man[mem->mem_type]; + man = ttm_manager_type(bdev, mem->mem_type); if (virtual && mem->bus.addr == NULL) iounmap(virtual); @@ -303,7 +303,7 @@ int ttm_bo_move_memcpy(struct ttm_buffer_object *bo, struct ttm_mem_reg *new_mem) { struct ttm_bo_device *bdev = bo->bdev; - struct ttm_mem_type_manager *man = &bdev->man[new_mem->mem_type]; + struct ttm_mem_type_manager *man = ttm_manager_type(bdev, new_mem->mem_type); struct ttm_tt *ttm = bo->ttm; struct ttm_mem_reg *old_mem = &bo->mem; struct ttm_mem_reg old_copy = *old_mem; @@ -571,7 +571,7 @@ int ttm_bo_kmap(struct ttm_buffer_object *bo, struct ttm_bo_kmap_obj *map) { struct ttm_mem_type_manager *man = - &bo->bdev->man[bo->mem.mem_type]; + ttm_manager_type(bo->bdev, bo->mem.mem_type); unsigned long offset, size; int ret; @@ -601,7 +601,7 @@ void ttm_bo_kunmap(struct ttm_bo_kmap_obj *map) { struct ttm_buffer_object *bo = map->bo; struct ttm_mem_type_manager *man = - &bo->bdev->man[bo->mem.mem_type]; + ttm_manager_type(bo->bdev, bo->mem.mem_type); if (!map->virtual) return; @@ -634,7 +634,7 @@ int ttm_bo_move_accel_cleanup(struct ttm_buffer_object *bo, struct ttm_mem_reg *new_mem) { struct ttm_bo_device *bdev = bo->bdev; - struct ttm_mem_type_manager *man = &bdev->man[new_mem->mem_type]; + struct ttm_mem_type_manager *man = ttm_manager_type(bdev, new_mem->mem_type); struct ttm_mem_reg *old_mem = &bo->mem; int ret; struct ttm_buffer_object *ghost_obj; @@ -697,8 +697,8 @@ int ttm_bo_pipeline_move(struct ttm_buffer_object *bo, struct ttm_bo_device *bdev = bo->bdev; struct ttm_mem_reg *old_mem = &bo->mem; - struct ttm_mem_type_manager *from = &bdev->man[old_mem->mem_type]; - struct ttm_mem_type_manager *to = &bdev->man[new_mem->mem_type]; + struct ttm_mem_type_manager *from = ttm_manager_type(bdev, old_mem->mem_type); + struct ttm_mem_type_manager *to = ttm_manager_type(bdev, new_mem->mem_type); int ret; diff --git a/drivers/gpu/drm/ttm/ttm_bo_vm.c b/drivers/gpu/drm/ttm/ttm_bo_vm.c index 468a0eb9e632..5ae679184eb5 100644 --- a/drivers/gpu/drm/ttm/ttm_bo_vm.c +++ b/drivers/gpu/drm/ttm/ttm_bo_vm.c @@ -282,7 +282,7 @@ vm_fault_t ttm_bo_vm_fault_reserved(struct vm_fault *vmf, vm_fault_t ret = VM_FAULT_NOPAGE; unsigned long address = vmf->address; struct ttm_mem_type_manager *man = - &bdev->man[bo->mem.mem_type]; + ttm_manager_type(bdev, bo->mem.mem_type); /* * Refuse to fault imported pages. This should be handled diff --git a/include/drm/ttm/ttm_bo_driver.h b/include/drm/ttm/ttm_bo_driver.h index 9d066529ca61..ec25451b503f 100644 --- a/include/drm/ttm/ttm_bo_driver.h +++ b/include/drm/ttm/ttm_bo_driver.h @@ -444,6 +444,12 @@ struct ttm_bo_device { bool no_retry; }; +static inline struct ttm_mem_type_manager *ttm_manager_type(struct ttm_bo_device *bdev, + int mem_type) +{ + return &bdev->man[mem_type]; +} + /** * struct ttm_lru_bulk_move_pos *