[PATCH 5/6] drm/msm: support for an arbitrary number of address spaces

Jordan Crouse jcrouse at codeaurora.org
Tue Jun 13 19:40:06 UTC 2017


On Tue, Jun 13, 2017 at 02:49:47PM -0400, Rob Clark wrote:
> It means we have to do a list traversal where we once had an index into
> a table.  But the list will normally have one or two entries.
> 
> Signed-off-by: Rob Clark <robdclark at gmail.com>

I dig the rename - makes more sense.

Acked-by: Jordan Crouse <jcrouse at codeaurora.org>

> ---
>  drivers/gpu/drm/msm/msm_gem.c | 138 +++++++++++++++++++++++++++++-------------
>  drivers/gpu/drm/msm/msm_gem.h |   4 +-
>  2 files changed, 99 insertions(+), 43 deletions(-)
> 
> diff --git a/drivers/gpu/drm/msm/msm_gem.c b/drivers/gpu/drm/msm/msm_gem.c
> index 754432c..410368f 100644
> --- a/drivers/gpu/drm/msm/msm_gem.c
> +++ b/drivers/gpu/drm/msm/msm_gem.c
> @@ -283,21 +283,59 @@ uint64_t msm_gem_mmap_offset(struct drm_gem_object *obj)
>  	return offset;
>  }
>  
> +static struct msm_gem_vma *add_vma(struct drm_gem_object *obj,
> +		struct msm_gem_address_space *aspace)
> +{
> +	struct msm_gem_object *msm_obj = to_msm_bo(obj);
> +	struct msm_gem_vma *vma;
> +
> +	vma = kzalloc(sizeof(*vma), GFP_KERNEL);
> +	if (!vma)
> +		return ERR_PTR(-ENOMEM);
> +
> +	vma->aspace = aspace;
> +
> +	list_add_tail(&vma->list, &msm_obj->vmas);
> +
> +	return vma;
> +}
> +
> +static struct msm_gem_vma *lookup_vma(struct drm_gem_object *obj,
> +		struct msm_gem_address_space *aspace)
> +{
> +	struct msm_gem_object *msm_obj = to_msm_bo(obj);
> +	struct msm_gem_vma *vma;
> +
> +	WARN_ON(!mutex_is_locked(&obj->dev->struct_mutex));
> +
> +	list_for_each_entry(vma, &msm_obj->vmas, list) {
> +		if (vma->aspace == aspace)
> +			return vma;
> +	}
> +
> +	return NULL;
> +}
> +
> +static void del_vma(struct msm_gem_vma *vma)
> +{
> +	if (!vma)
> +		return;
> +
> +	list_del(&vma->list);
> +	kfree(vma);
> +}
> +
>  static void
>  put_iova(struct drm_gem_object *obj)
>  {
> -	struct drm_device *dev = obj->dev;
> -	struct msm_drm_private *priv = obj->dev->dev_private;
>  	struct msm_gem_object *msm_obj = to_msm_bo(obj);
> -	int id;
> +	struct msm_gem_vma *vma, *tmp;
>  
> -	WARN_ON(!mutex_is_locked(&dev->struct_mutex));
> +	WARN_ON(!mutex_is_locked(&obj->dev->struct_mutex));
>  
> -	for (id = 0; id < ARRAY_SIZE(msm_obj->domain); id++) {
> -		if (!priv->aspace[id])
> -			continue;
> -		msm_gem_unmap_vma(priv->aspace[id],
> -				&msm_obj->domain[id], msm_obj->sgt);
> +	list_for_each_entry_safe(vma, tmp, &msm_obj->vmas, list) {
> +		msm_gem_unmap_vma(vma->aspace, vma, msm_obj->sgt);
> +		del_vma(vma);
>  	}
>  }
>  
> @@ -312,24 +350,37 @@ int msm_gem_get_iova_locked(struct drm_gem_object *obj,
>  		struct msm_gem_address_space *aspace, uint64_t *iova)
>  {
>  	struct msm_gem_object *msm_obj = to_msm_bo(obj);
> -	int id = aspace ? aspace->id : 0;
> +	struct msm_gem_vma *vma;
>  	int ret = 0;
>  
>  	WARN_ON(!mutex_is_locked(&obj->dev->struct_mutex));
>  
> -	if (!msm_obj->domain[id].iova) {
> -		struct msm_drm_private *priv = obj->dev->dev_private;
> -		struct page **pages = get_pages(obj);
> +	vma = lookup_vma(obj, aspace);
>  
> -		if (IS_ERR(pages))
> -			return PTR_ERR(pages);
> +	if (!vma) {
> +		struct page **pages;
> +
> +		vma = add_vma(obj, aspace);
> +		if (IS_ERR(vma))
> +			return PTR_ERR(vma);
> +
> +		pages = get_pages(obj);
> +		if (IS_ERR(pages)) {
> +			ret = PTR_ERR(pages);
> +			goto fail;
> +		}
>  
> -		ret = msm_gem_map_vma(priv->aspace[id], &msm_obj->domain[id],
> -				msm_obj->sgt, obj->size >> PAGE_SHIFT);
> +		ret = msm_gem_map_vma(aspace, vma, msm_obj->sgt,
> +				obj->size >> PAGE_SHIFT);
> +		if (ret)
> +			goto fail;
>  	}
>  
> -	if (!ret)
> -		*iova = msm_obj->domain[id].iova;
> +	*iova = vma->iova;
> +	return 0;
> +
> +fail:
> +	del_vma(vma);
>  
>  	return ret;
>  }
> @@ -338,22 +389,12 @@ int msm_gem_get_iova_locked(struct drm_gem_object *obj,
>  int msm_gem_get_iova(struct drm_gem_object *obj,
>  		struct msm_gem_address_space *aspace, uint64_t *iova)
>  {
> -	struct msm_gem_object *msm_obj = to_msm_bo(obj);
> -	int id = aspace ? aspace->id : 0;
>  	int ret;
>  
> -	/* this is safe right now because we don't unmap until the
> -	 * bo is deleted:
> -	 */
> -	if (msm_obj->domain[id].iova) {
> -		might_lock(&obj->dev->struct_mutex);
> -		*iova = msm_obj->domain[id].iova;
> -		return 0;
> -	}
> -
>  	mutex_lock(&obj->dev->struct_mutex);
>  	ret = msm_gem_get_iova_locked(obj, aspace, iova);
>  	mutex_unlock(&obj->dev->struct_mutex);
> +
>  	return ret;
>  }
>  
> @@ -363,10 +404,14 @@ int msm_gem_get_iova(struct drm_gem_object *obj,
>  uint64_t msm_gem_iova(struct drm_gem_object *obj,
>  		struct msm_gem_address_space *aspace)
>  {
> -	struct msm_gem_object *msm_obj = to_msm_bo(obj);
> -	int id = aspace ? aspace->id : 0;
> -	WARN_ON(!msm_obj->domain[id].iova);
> -	return msm_obj->domain[id].iova;
> +	struct msm_gem_vma *vma;
> +
> +	mutex_lock(&obj->dev->struct_mutex);
> +	vma = lookup_vma(obj, aspace);
> +	mutex_unlock(&obj->dev->struct_mutex);
> +	WARN_ON(!vma);
> +
> +	return vma ? vma->iova : 0;
>  }
>  
>  void msm_gem_put_iova(struct drm_gem_object *obj,
> @@ -624,11 +669,10 @@ void msm_gem_describe(struct drm_gem_object *obj, struct seq_file *m)
>  	struct msm_gem_object *msm_obj = to_msm_bo(obj);
>  	struct reservation_object *robj = msm_obj->resv;
>  	struct reservation_object_list *fobj;
> -	struct msm_drm_private *priv = obj->dev->dev_private;
>  	struct dma_fence *fence;
> +	struct msm_gem_vma *vma;
>  	uint64_t off = drm_vma_node_start(&obj->vma_node);
>  	const char *madv;
> -	unsigned id;
>  
>  	WARN_ON(!mutex_is_locked(&obj->dev->struct_mutex));
>  
> @@ -650,8 +694,9 @@ void msm_gem_describe(struct drm_gem_object *obj, struct seq_file *m)
>  			obj->name, kref_read(&obj->refcount),
>  			off, msm_obj->vaddr);
>  
> -	for (id = 0; id < priv->num_aspaces; id++)
> -		seq_printf(m, " %08llx", msm_obj->domain[id].iova);
> +	/* FIXME: we need to print the address space here too */
> +	list_for_each_entry(vma, &msm_obj->vmas, list)
> +		seq_printf(m, " %08llx", vma->iova);
>  
>  	seq_printf(m, " %zu%s\n", obj->size, madv);
>  
> @@ -788,6 +833,8 @@ static int msm_gem_new_impl(struct drm_device *dev,
>  	}
>  
>  	INIT_LIST_HEAD(&msm_obj->submit_entry);
> +	INIT_LIST_HEAD(&msm_obj->vmas);
> +
>  	list_add_tail(&msm_obj->mm_list, &priv->inactive_list);
>  
>  	*obj = &msm_obj->base;
> @@ -826,19 +873,26 @@ struct drm_gem_object *msm_gem_new(struct drm_device *dev,
>  		goto fail;
>  
>  	if (use_vram) {
> -		struct msm_gem_object *msm_obj = to_msm_bo(obj);
> +		struct msm_gem_vma *vma;
>  		struct page **pages;
>  
> -		msm_obj->vram_node = &msm_obj->domain[0].node;
> +		vma = add_vma(obj, NULL);
> +		if (IS_ERR(vma)) {
> +			ret = PTR_ERR(vma);
> +			goto fail;
> +		}
> +
> +		to_msm_bo(obj)->vram_node = &vma->node;
> +
>  		drm_gem_private_object_init(dev, obj, size);
>  
> -		msm_obj->pages = get_pages(obj);
>  		pages = get_pages(obj);
>  		if (IS_ERR(pages)) {
>  			ret = PTR_ERR(pages);
>  			goto fail;
>  		}
> -		msm_obj->domain[0].iova = physaddr(obj);
> +
> +		vma->iova = physaddr(obj);
>  	} else {
>  		ret = drm_gem_object_init(dev, obj, size);
>  		if (ret)
> diff --git a/drivers/gpu/drm/msm/msm_gem.h b/drivers/gpu/drm/msm/msm_gem.h
> index 4b4b352..ff468da 100644
> --- a/drivers/gpu/drm/msm/msm_gem.h
> +++ b/drivers/gpu/drm/msm/msm_gem.h
> @@ -39,6 +39,8 @@ struct msm_gem_address_space {
>  struct msm_gem_vma {
>  	struct drm_mm_node node;
>  	uint64_t iova;
> +	struct msm_gem_address_space *aspace;
> +	struct list_head list;    /* node in msm_gem_object::vmas */
>  };
>  
>  struct msm_gem_object {
> @@ -78,7 +80,7 @@ struct msm_gem_object {
>  	struct sg_table *sgt;
>  	void *vaddr;
>  
> -	struct msm_gem_vma domain[NUM_DOMAINS];
> +	struct list_head vmas;    /* list of msm_gem_vma */
>  
>  	/* normally (resv == &_resv) except for imported bo's */
>  	struct reservation_object *resv;
> -- 
> 2.9.4
> 

-- 
The Qualcomm Innovation Center, Inc. is a member of Code Aurora Forum,
a Linux Foundation Collaborative Project


More information about the dri-devel mailing list