[PATCH 4/9] drm/msm/gem: Drop PAGE_SHIFT for address space mm

Dmitry Baryshkov dmitry.baryshkov at linaro.org
Tue Mar 29 23:12:05 UTC 2022


On Wed, 30 Mar 2022 at 02:00, Rob Clark <robdclark at gmail.com> wrote:
>
> From: Rob Clark <robdclark at chromium.org>
>
> Get rid of all the unnecessary conversion between address/size and page
> offsets.  It just confuses things.

Reviewed-by: Dmitry Baryshkov <dmitry.baryshkov at linaro.org>



>
> Signed-off-by: Rob Clark <robdclark at chromium.org>
> ---
>  drivers/gpu/drm/msm/adreno/a6xx_gmu.c |  2 +-
>  drivers/gpu/drm/msm/msm_gem.c         |  5 ++---
>  drivers/gpu/drm/msm/msm_gem.h         |  4 ++--
>  drivers/gpu/drm/msm/msm_gem_vma.c     | 16 ++++++++--------
>  4 files changed, 13 insertions(+), 14 deletions(-)
>
> diff --git a/drivers/gpu/drm/msm/adreno/a6xx_gmu.c b/drivers/gpu/drm/msm/adreno/a6xx_gmu.c
> index 3e325e2a2b1b..9f76f5b15759 100644
> --- a/drivers/gpu/drm/msm/adreno/a6xx_gmu.c
> +++ b/drivers/gpu/drm/msm/adreno/a6xx_gmu.c
> @@ -1172,7 +1172,7 @@ static int a6xx_gmu_memory_alloc(struct a6xx_gmu *gmu, struct a6xx_gmu_bo *bo,
>                 return PTR_ERR(bo->obj);
>
>         ret = msm_gem_get_and_pin_iova_range(bo->obj, gmu->aspace, &bo->iova,
> -               range_start >> PAGE_SHIFT, range_end >> PAGE_SHIFT);
> +                                            range_start, range_end);
>         if (ret) {
>                 drm_gem_object_put(bo->obj);
>                 return ret;
> diff --git a/drivers/gpu/drm/msm/msm_gem.c b/drivers/gpu/drm/msm/msm_gem.c
> index f96d1dc72021..f4b68bb28a4d 100644
> --- a/drivers/gpu/drm/msm/msm_gem.c
> +++ b/drivers/gpu/drm/msm/msm_gem.c
> @@ -392,7 +392,7 @@ static int get_iova_locked(struct drm_gem_object *obj,
>                 if (IS_ERR(vma))
>                         return PTR_ERR(vma);
>
> -               ret = msm_gem_init_vma(aspace, vma, obj->size >> PAGE_SHIFT,
> +               ret = msm_gem_init_vma(aspace, vma, obj->size,
>                         range_start, range_end);
>                 if (ret) {
>                         del_vma(vma);
> @@ -434,8 +434,7 @@ static int msm_gem_pin_iova(struct drm_gem_object *obj,
>         if (IS_ERR(pages))
>                 return PTR_ERR(pages);
>
> -       ret = msm_gem_map_vma(aspace, vma, prot,
> -                       msm_obj->sgt, obj->size >> PAGE_SHIFT);
> +       ret = msm_gem_map_vma(aspace, vma, prot, msm_obj->sgt, obj->size);
>
>         if (!ret)
>                 msm_obj->pin_count++;
> diff --git a/drivers/gpu/drm/msm/msm_gem.h b/drivers/gpu/drm/msm/msm_gem.h
> index 1b7f0f0b88bf..090c3b1a6d9a 100644
> --- a/drivers/gpu/drm/msm/msm_gem.h
> +++ b/drivers/gpu/drm/msm/msm_gem.h
> @@ -59,7 +59,7 @@ struct msm_gem_vma {
>  };
>
>  int msm_gem_init_vma(struct msm_gem_address_space *aspace,
> -               struct msm_gem_vma *vma, int npages,
> +               struct msm_gem_vma *vma, int size,
>                 u64 range_start, u64 range_end);
>  bool msm_gem_vma_inuse(struct msm_gem_vma *vma);
>  void msm_gem_purge_vma(struct msm_gem_address_space *aspace,
> @@ -68,7 +68,7 @@ void msm_gem_unmap_vma(struct msm_gem_address_space *aspace,
>                 struct msm_gem_vma *vma);
>  int msm_gem_map_vma(struct msm_gem_address_space *aspace,
>                 struct msm_gem_vma *vma, int prot,
> -               struct sg_table *sgt, int npages);
> +               struct sg_table *sgt, int size);
>  void msm_gem_close_vma(struct msm_gem_address_space *aspace,
>                 struct msm_gem_vma *vma);
>
> diff --git a/drivers/gpu/drm/msm/msm_gem_vma.c b/drivers/gpu/drm/msm/msm_gem_vma.c
> index dc2ae097805e..4949899f1fc7 100644
> --- a/drivers/gpu/drm/msm/msm_gem_vma.c
> +++ b/drivers/gpu/drm/msm/msm_gem_vma.c
> @@ -46,7 +46,7 @@ bool msm_gem_vma_inuse(struct msm_gem_vma *vma)
>  void msm_gem_purge_vma(struct msm_gem_address_space *aspace,
>                 struct msm_gem_vma *vma)
>  {
> -       unsigned size = vma->node.size << PAGE_SHIFT;
> +       unsigned size = vma->node.size;
>
>         /* Print a message if we try to purge a vma in use */
>         if (GEM_WARN_ON(msm_gem_vma_inuse(vma)))
> @@ -73,9 +73,8 @@ void msm_gem_unmap_vma(struct msm_gem_address_space *aspace,
>  int
>  msm_gem_map_vma(struct msm_gem_address_space *aspace,
>                 struct msm_gem_vma *vma, int prot,
> -               struct sg_table *sgt, int npages)
> +               struct sg_table *sgt, int size)
>  {
> -       unsigned size = npages << PAGE_SHIFT;
>         int ret = 0;
>
>         if (GEM_WARN_ON(!vma->iova))
> @@ -120,7 +119,7 @@ void msm_gem_close_vma(struct msm_gem_address_space *aspace,
>
>  /* Initialize a new vma and allocate an iova for it */
>  int msm_gem_init_vma(struct msm_gem_address_space *aspace,
> -               struct msm_gem_vma *vma, int npages,
> +               struct msm_gem_vma *vma, int size,
>                 u64 range_start, u64 range_end)
>  {
>         int ret;
> @@ -129,14 +128,15 @@ int msm_gem_init_vma(struct msm_gem_address_space *aspace,
>                 return -EBUSY;
>
>         spin_lock(&aspace->lock);
> -       ret = drm_mm_insert_node_in_range(&aspace->mm, &vma->node, npages, 0,
> -               0, range_start, range_end, 0);
> +       ret = drm_mm_insert_node_in_range(&aspace->mm, &vma->node,
> +                                         size, PAGE_SIZE, 0,
> +                                         range_start, range_end, 0);
>         spin_unlock(&aspace->lock);
>
>         if (ret)
>                 return ret;
>
> -       vma->iova = vma->node.start << PAGE_SHIFT;
> +       vma->iova = vma->node.start;
>         vma->mapped = false;
>
>         kref_get(&aspace->kref);
> @@ -161,7 +161,7 @@ msm_gem_address_space_create(struct msm_mmu *mmu, const char *name,
>         aspace->name = name;
>         aspace->mmu = mmu;
>
> -       drm_mm_init(&aspace->mm, va_start >> PAGE_SHIFT, size >> PAGE_SHIFT);
> +       drm_mm_init(&aspace->mm, va_start, size);
>
>         kref_init(&aspace->kref);
>
> --
> 2.35.1
>


-- 
With best wishes
Dmitry


More information about the dri-devel mailing list