diff mbox

[5/6] drm/msm: support for an arbitrary number of address spaces

Message ID 20170613184948.8278-6-robdclark@gmail.com (mailing list archive)
State New, archived
Headers show

Commit Message

Rob Clark June 13, 2017, 6:49 p.m. UTC
It means we have to do a list traversal where we once had an index into
a table.  But the list will normally have one or two entries.

Signed-off-by: Rob Clark <robdclark@gmail.com>
---
 drivers/gpu/drm/msm/msm_gem.c | 138 +++++++++++++++++++++++++++++-------------
 drivers/gpu/drm/msm/msm_gem.h |   4 +-
 2 files changed, 99 insertions(+), 43 deletions(-)

Comments

Jordan Crouse June 13, 2017, 7:40 p.m. UTC | #1
On Tue, Jun 13, 2017 at 02:49:47PM -0400, Rob Clark wrote:
> It means we have to do a list traversal where we once had an index into
> a table.  But the list will normally have one or two entries.
> 
> Signed-off-by: Rob Clark <robdclark@gmail.com>

I dig the rename - makes more sense.

Acked-by: Jordan Crouse <jcrouse@codeaurora.org>

> ---
>  drivers/gpu/drm/msm/msm_gem.c | 138 +++++++++++++++++++++++++++++-------------
>  drivers/gpu/drm/msm/msm_gem.h |   4 +-
>  2 files changed, 99 insertions(+), 43 deletions(-)
> 
> diff --git a/drivers/gpu/drm/msm/msm_gem.c b/drivers/gpu/drm/msm/msm_gem.c
> index 754432c..410368f 100644
> --- a/drivers/gpu/drm/msm/msm_gem.c
> +++ b/drivers/gpu/drm/msm/msm_gem.c
> @@ -283,21 +283,59 @@ uint64_t msm_gem_mmap_offset(struct drm_gem_object *obj)
>  	return offset;
>  }
>  
> +static struct msm_gem_vma *add_vma(struct drm_gem_object *obj,
> +		struct msm_gem_address_space *aspace)
> +{
> +	struct msm_gem_object *msm_obj = to_msm_bo(obj);
> +	struct msm_gem_vma *vma;
> +
> +	vma = kzalloc(sizeof(*vma), GFP_KERNEL);
> +	if (!vma)
> +		return ERR_PTR(-ENOMEM);
> +
> +	vma->aspace = aspace;
> +
> +	list_add_tail(&vma->list, &msm_obj->vmas);
> +
> +	return vma;
> +}
> +
> +static struct msm_gem_vma *lookup_vma(struct drm_gem_object *obj,
> +		struct msm_gem_address_space *aspace)
> +{
> +	struct msm_gem_object *msm_obj = to_msm_bo(obj);
> +	struct msm_gem_vma *vma;
> +
> +	WARN_ON(!mutex_is_locked(&obj->dev->struct_mutex));
> +
> +	list_for_each_entry(vma, &msm_obj->vmas, list) {
> +		if (vma->aspace == aspace)
> +			return vma;
> +	}
> +
> +	return NULL;
> +}
> +
> +static void del_vma(struct msm_gem_vma *vma)
> +{
> +	if (!vma)
> +		return;
> +
> +	list_del(&vma->list);
> +	kfree(vma);
> +}
> +
>  static void
>  put_iova(struct drm_gem_object *obj)
>  {
> -	struct drm_device *dev = obj->dev;
> -	struct msm_drm_private *priv = obj->dev->dev_private;
>  	struct msm_gem_object *msm_obj = to_msm_bo(obj);
> -	int id;
> +	struct msm_gem_vma *vma, *tmp;
>  
> -	WARN_ON(!mutex_is_locked(&dev->struct_mutex));
> +	WARN_ON(!mutex_is_locked(&obj->dev->struct_mutex));
>  
> -	for (id = 0; id < ARRAY_SIZE(msm_obj->domain); id++) {
> -		if (!priv->aspace[id])
> -			continue;
> -		msm_gem_unmap_vma(priv->aspace[id],
> -				&msm_obj->domain[id], msm_obj->sgt);
> +	list_for_each_entry_safe(vma, tmp, &msm_obj->vmas, list) {
> +		msm_gem_unmap_vma(vma->aspace, vma, msm_obj->sgt);
> +		del_vma(vma);
>  	}
>  }
>  
> @@ -312,24 +350,37 @@ int msm_gem_get_iova_locked(struct drm_gem_object *obj,
>  		struct msm_gem_address_space *aspace, uint64_t *iova)
>  {
>  	struct msm_gem_object *msm_obj = to_msm_bo(obj);
> -	int id = aspace ? aspace->id : 0;
> +	struct msm_gem_vma *vma;
>  	int ret = 0;
>  
>  	WARN_ON(!mutex_is_locked(&obj->dev->struct_mutex));
>  
> -	if (!msm_obj->domain[id].iova) {
> -		struct msm_drm_private *priv = obj->dev->dev_private;
> -		struct page **pages = get_pages(obj);
> +	vma = lookup_vma(obj, aspace);
>  
> -		if (IS_ERR(pages))
> -			return PTR_ERR(pages);
> +	if (!vma) {
> +		struct page **pages;
> +
> +		vma = add_vma(obj, aspace);
> +		if (IS_ERR(vma))
> +			return PTR_ERR(vma);
> +
> +		pages = get_pages(obj);
> +		if (IS_ERR(pages)) {
> +			ret = PTR_ERR(pages);
> +			goto fail;
> +		}
>  
> -		ret = msm_gem_map_vma(priv->aspace[id], &msm_obj->domain[id],
> -				msm_obj->sgt, obj->size >> PAGE_SHIFT);
> +		ret = msm_gem_map_vma(aspace, vma, msm_obj->sgt,
> +				obj->size >> PAGE_SHIFT);
> +		if (ret)
> +			goto fail;
>  	}
>  
> -	if (!ret)
> -		*iova = msm_obj->domain[id].iova;
> +	*iova = vma->iova;
> +	return 0;
> +
> +fail:
> +	del_vma(vma);
>  
>  	return ret;
>  }
> @@ -338,22 +389,12 @@ int msm_gem_get_iova_locked(struct drm_gem_object *obj,
>  int msm_gem_get_iova(struct drm_gem_object *obj,
>  		struct msm_gem_address_space *aspace, uint64_t *iova)
>  {
> -	struct msm_gem_object *msm_obj = to_msm_bo(obj);
> -	int id = aspace ? aspace->id : 0;
>  	int ret;
>  
> -	/* this is safe right now because we don't unmap until the
> -	 * bo is deleted:
> -	 */
> -	if (msm_obj->domain[id].iova) {
> -		might_lock(&obj->dev->struct_mutex);
> -		*iova = msm_obj->domain[id].iova;
> -		return 0;
> -	}
> -
>  	mutex_lock(&obj->dev->struct_mutex);
>  	ret = msm_gem_get_iova_locked(obj, aspace, iova);
>  	mutex_unlock(&obj->dev->struct_mutex);
> +
>  	return ret;
>  }
>  
> @@ -363,10 +404,14 @@ int msm_gem_get_iova(struct drm_gem_object *obj,
>  uint64_t msm_gem_iova(struct drm_gem_object *obj,
>  		struct msm_gem_address_space *aspace)
>  {
> -	struct msm_gem_object *msm_obj = to_msm_bo(obj);
> -	int id = aspace ? aspace->id : 0;
> -	WARN_ON(!msm_obj->domain[id].iova);
> -	return msm_obj->domain[id].iova;
> +	struct msm_gem_vma *vma;
> +
> +	mutex_lock(&obj->dev->struct_mutex);
> +	vma = lookup_vma(obj, aspace);
> +	mutex_unlock(&obj->dev->struct_mutex);
> +	WARN_ON(!vma);
> +
> +	return vma ? vma->iova : 0;
>  }
>  
>  void msm_gem_put_iova(struct drm_gem_object *obj,
> @@ -624,11 +669,10 @@ void msm_gem_describe(struct drm_gem_object *obj, struct seq_file *m)
>  	struct msm_gem_object *msm_obj = to_msm_bo(obj);
>  	struct reservation_object *robj = msm_obj->resv;
>  	struct reservation_object_list *fobj;
> -	struct msm_drm_private *priv = obj->dev->dev_private;
>  	struct dma_fence *fence;
> +	struct msm_gem_vma *vma;
>  	uint64_t off = drm_vma_node_start(&obj->vma_node);
>  	const char *madv;
> -	unsigned id;
>  
>  	WARN_ON(!mutex_is_locked(&obj->dev->struct_mutex));
>  
> @@ -650,8 +694,9 @@ void msm_gem_describe(struct drm_gem_object *obj, struct seq_file *m)
>  			obj->name, kref_read(&obj->refcount),
>  			off, msm_obj->vaddr);
>  
> -	for (id = 0; id < priv->num_aspaces; id++)
> -		seq_printf(m, " %08llx", msm_obj->domain[id].iova);
> +	/* FIXME: we need to print the address space here too */
> +	list_for_each_entry(vma, &msm_obj->vmas, list)
> +		seq_printf(m, " %08llx", vma->iova);
>  
>  	seq_printf(m, " %zu%s\n", obj->size, madv);
>  
> @@ -788,6 +833,8 @@ static int msm_gem_new_impl(struct drm_device *dev,
>  	}
>  
>  	INIT_LIST_HEAD(&msm_obj->submit_entry);
> +	INIT_LIST_HEAD(&msm_obj->vmas);
> +
>  	list_add_tail(&msm_obj->mm_list, &priv->inactive_list);
>  
>  	*obj = &msm_obj->base;
> @@ -826,19 +873,26 @@ struct drm_gem_object *msm_gem_new(struct drm_device *dev,
>  		goto fail;
>  
>  	if (use_vram) {
> -		struct msm_gem_object *msm_obj = to_msm_bo(obj);
> +		struct msm_gem_vma *vma;
>  		struct page **pages;
>  
> -		msm_obj->vram_node = &msm_obj->domain[0].node;
> +		vma = add_vma(obj, NULL);
> +		if (IS_ERR(vma)) {
> +			ret = PTR_ERR(vma);
> +			goto fail;
> +		}
> +
> +		to_msm_bo(obj)->vram_node = &vma->node;
> +
>  		drm_gem_private_object_init(dev, obj, size);
>  
> -		msm_obj->pages = get_pages(obj);
>  		pages = get_pages(obj);
>  		if (IS_ERR(pages)) {
>  			ret = PTR_ERR(pages);
>  			goto fail;
>  		}
> -		msm_obj->domain[0].iova = physaddr(obj);
> +
> +		vma->iova = physaddr(obj);
>  	} else {
>  		ret = drm_gem_object_init(dev, obj, size);
>  		if (ret)
> diff --git a/drivers/gpu/drm/msm/msm_gem.h b/drivers/gpu/drm/msm/msm_gem.h
> index 4b4b352..ff468da 100644
> --- a/drivers/gpu/drm/msm/msm_gem.h
> +++ b/drivers/gpu/drm/msm/msm_gem.h
> @@ -39,6 +39,8 @@ struct msm_gem_address_space {
>  struct msm_gem_vma {
>  	struct drm_mm_node node;
>  	uint64_t iova;
> +	struct msm_gem_address_space *aspace;
> +	struct list_head list;    /* node in msm_gem_object::vmas */
>  };
>  
>  struct msm_gem_object {
> @@ -78,7 +80,7 @@ struct msm_gem_object {
>  	struct sg_table *sgt;
>  	void *vaddr;
>  
> -	struct msm_gem_vma domain[NUM_DOMAINS];
> +	struct list_head vmas;    /* list of msm_gem_vma */
>  
>  	/* normally (resv == &_resv) except for imported bo's */
>  	struct reservation_object *resv;
> -- 
> 2.9.4
>
diff mbox

Patch

diff --git a/drivers/gpu/drm/msm/msm_gem.c b/drivers/gpu/drm/msm/msm_gem.c
index 754432c..410368f 100644
--- a/drivers/gpu/drm/msm/msm_gem.c
+++ b/drivers/gpu/drm/msm/msm_gem.c
@@ -283,21 +283,59 @@  uint64_t msm_gem_mmap_offset(struct drm_gem_object *obj)
 	return offset;
 }
 
+static struct msm_gem_vma *add_vma(struct drm_gem_object *obj,
+		struct msm_gem_address_space *aspace)
+{
+	struct msm_gem_object *msm_obj = to_msm_bo(obj);
+	struct msm_gem_vma *vma;
+
+	vma = kzalloc(sizeof(*vma), GFP_KERNEL);
+	if (!vma)
+		return ERR_PTR(-ENOMEM);
+
+	vma->aspace = aspace;
+
+	list_add_tail(&vma->list, &msm_obj->vmas);
+
+	return vma;
+}
+
+static struct msm_gem_vma *lookup_vma(struct drm_gem_object *obj,
+		struct msm_gem_address_space *aspace)
+{
+	struct msm_gem_object *msm_obj = to_msm_bo(obj);
+	struct msm_gem_vma *vma;
+
+	WARN_ON(!mutex_is_locked(&obj->dev->struct_mutex));
+
+	list_for_each_entry(vma, &msm_obj->vmas, list) {
+		if (vma->aspace == aspace)
+			return vma;
+	}
+
+	return NULL;
+}
+
+static void del_vma(struct msm_gem_vma *vma)
+{
+	if (!vma)
+		return;
+
+	list_del(&vma->list);
+	kfree(vma);
+}
+
 static void
 put_iova(struct drm_gem_object *obj)
 {
-	struct drm_device *dev = obj->dev;
-	struct msm_drm_private *priv = obj->dev->dev_private;
 	struct msm_gem_object *msm_obj = to_msm_bo(obj);
-	int id;
+	struct msm_gem_vma *vma, *tmp;
 
-	WARN_ON(!mutex_is_locked(&dev->struct_mutex));
+	WARN_ON(!mutex_is_locked(&obj->dev->struct_mutex));
 
-	for (id = 0; id < ARRAY_SIZE(msm_obj->domain); id++) {
-		if (!priv->aspace[id])
-			continue;
-		msm_gem_unmap_vma(priv->aspace[id],
-				&msm_obj->domain[id], msm_obj->sgt);
+	list_for_each_entry_safe(vma, tmp, &msm_obj->vmas, list) {
+		msm_gem_unmap_vma(vma->aspace, vma, msm_obj->sgt);
+		del_vma(vma);
 	}
 }
 
@@ -312,24 +350,37 @@  int msm_gem_get_iova_locked(struct drm_gem_object *obj,
 		struct msm_gem_address_space *aspace, uint64_t *iova)
 {
 	struct msm_gem_object *msm_obj = to_msm_bo(obj);
-	int id = aspace ? aspace->id : 0;
+	struct msm_gem_vma *vma;
 	int ret = 0;
 
 	WARN_ON(!mutex_is_locked(&obj->dev->struct_mutex));
 
-	if (!msm_obj->domain[id].iova) {
-		struct msm_drm_private *priv = obj->dev->dev_private;
-		struct page **pages = get_pages(obj);
+	vma = lookup_vma(obj, aspace);
 
-		if (IS_ERR(pages))
-			return PTR_ERR(pages);
+	if (!vma) {
+		struct page **pages;
+
+		vma = add_vma(obj, aspace);
+		if (IS_ERR(vma))
+			return PTR_ERR(vma);
+
+		pages = get_pages(obj);
+		if (IS_ERR(pages)) {
+			ret = PTR_ERR(pages);
+			goto fail;
+		}
 
-		ret = msm_gem_map_vma(priv->aspace[id], &msm_obj->domain[id],
-				msm_obj->sgt, obj->size >> PAGE_SHIFT);
+		ret = msm_gem_map_vma(aspace, vma, msm_obj->sgt,
+				obj->size >> PAGE_SHIFT);
+		if (ret)
+			goto fail;
 	}
 
-	if (!ret)
-		*iova = msm_obj->domain[id].iova;
+	*iova = vma->iova;
+	return 0;
+
+fail:
+	del_vma(vma);
 
 	return ret;
 }
@@ -338,22 +389,12 @@  int msm_gem_get_iova_locked(struct drm_gem_object *obj,
 int msm_gem_get_iova(struct drm_gem_object *obj,
 		struct msm_gem_address_space *aspace, uint64_t *iova)
 {
-	struct msm_gem_object *msm_obj = to_msm_bo(obj);
-	int id = aspace ? aspace->id : 0;
 	int ret;
 
-	/* this is safe right now because we don't unmap until the
-	 * bo is deleted:
-	 */
-	if (msm_obj->domain[id].iova) {
-		might_lock(&obj->dev->struct_mutex);
-		*iova = msm_obj->domain[id].iova;
-		return 0;
-	}
-
 	mutex_lock(&obj->dev->struct_mutex);
 	ret = msm_gem_get_iova_locked(obj, aspace, iova);
 	mutex_unlock(&obj->dev->struct_mutex);
+
 	return ret;
 }
 
@@ -363,10 +404,14 @@  int msm_gem_get_iova(struct drm_gem_object *obj,
 uint64_t msm_gem_iova(struct drm_gem_object *obj,
 		struct msm_gem_address_space *aspace)
 {
-	struct msm_gem_object *msm_obj = to_msm_bo(obj);
-	int id = aspace ? aspace->id : 0;
-	WARN_ON(!msm_obj->domain[id].iova);
-	return msm_obj->domain[id].iova;
+	struct msm_gem_vma *vma;
+
+	mutex_lock(&obj->dev->struct_mutex);
+	vma = lookup_vma(obj, aspace);
+	mutex_unlock(&obj->dev->struct_mutex);
+	WARN_ON(!vma);
+
+	return vma ? vma->iova : 0;
 }
 
 void msm_gem_put_iova(struct drm_gem_object *obj,
@@ -624,11 +669,10 @@  void msm_gem_describe(struct drm_gem_object *obj, struct seq_file *m)
 	struct msm_gem_object *msm_obj = to_msm_bo(obj);
 	struct reservation_object *robj = msm_obj->resv;
 	struct reservation_object_list *fobj;
-	struct msm_drm_private *priv = obj->dev->dev_private;
 	struct dma_fence *fence;
+	struct msm_gem_vma *vma;
 	uint64_t off = drm_vma_node_start(&obj->vma_node);
 	const char *madv;
-	unsigned id;
 
 	WARN_ON(!mutex_is_locked(&obj->dev->struct_mutex));
 
@@ -650,8 +694,9 @@  void msm_gem_describe(struct drm_gem_object *obj, struct seq_file *m)
 			obj->name, kref_read(&obj->refcount),
 			off, msm_obj->vaddr);
 
-	for (id = 0; id < priv->num_aspaces; id++)
-		seq_printf(m, " %08llx", msm_obj->domain[id].iova);
+	/* FIXME: we need to print the address space here too */
+	list_for_each_entry(vma, &msm_obj->vmas, list)
+		seq_printf(m, " %08llx", vma->iova);
 
 	seq_printf(m, " %zu%s\n", obj->size, madv);
 
@@ -788,6 +833,8 @@  static int msm_gem_new_impl(struct drm_device *dev,
 	}
 
 	INIT_LIST_HEAD(&msm_obj->submit_entry);
+	INIT_LIST_HEAD(&msm_obj->vmas);
+
 	list_add_tail(&msm_obj->mm_list, &priv->inactive_list);
 
 	*obj = &msm_obj->base;
@@ -826,19 +873,26 @@  struct drm_gem_object *msm_gem_new(struct drm_device *dev,
 		goto fail;
 
 	if (use_vram) {
-		struct msm_gem_object *msm_obj = to_msm_bo(obj);
+		struct msm_gem_vma *vma;
 		struct page **pages;
 
-		msm_obj->vram_node = &msm_obj->domain[0].node;
+		vma = add_vma(obj, NULL);
+		if (IS_ERR(vma)) {
+			ret = PTR_ERR(vma);
+			goto fail;
+		}
+
+		to_msm_bo(obj)->vram_node = &vma->node;
+
 		drm_gem_private_object_init(dev, obj, size);
 
-		msm_obj->pages = get_pages(obj);
 		pages = get_pages(obj);
 		if (IS_ERR(pages)) {
 			ret = PTR_ERR(pages);
 			goto fail;
 		}
-		msm_obj->domain[0].iova = physaddr(obj);
+
+		vma->iova = physaddr(obj);
 	} else {
 		ret = drm_gem_object_init(dev, obj, size);
 		if (ret)
diff --git a/drivers/gpu/drm/msm/msm_gem.h b/drivers/gpu/drm/msm/msm_gem.h
index 4b4b352..ff468da 100644
--- a/drivers/gpu/drm/msm/msm_gem.h
+++ b/drivers/gpu/drm/msm/msm_gem.h
@@ -39,6 +39,8 @@  struct msm_gem_address_space {
 struct msm_gem_vma {
 	struct drm_mm_node node;
 	uint64_t iova;
+	struct msm_gem_address_space *aspace;
+	struct list_head list;    /* node in msm_gem_object::vmas */
 };
 
 struct msm_gem_object {
@@ -78,7 +80,7 @@  struct msm_gem_object {
 	struct sg_table *sgt;
 	void *vaddr;
 
-	struct msm_gem_vma domain[NUM_DOMAINS];
+	struct list_head vmas;    /* list of msm_gem_vma */
 
 	/* normally (resv == &_resv) except for imported bo's */
 	struct reservation_object *resv;