[PATCH 4/7] drm/amdgpu: use the AGP aperture for system memory access v2

Zhang, Jerry (Junwei) Jerry.Zhang at amd.com
Thu Aug 30 03:20:29 UTC 2018


On 08/29/2018 10:08 PM, Christian König wrote:
> Start to use the old AGP aperture for system memory access.
>
> v2: Move that to amdgpu_ttm_alloc_gart
>
> Signed-off-by: Christian König <christian.koenig at amd.com>
> ---
>   drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.c | 23 ++++++++++
>   drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.h |  1 +
>   drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c | 58 ++++++++++++++-----------
>   3 files changed, 57 insertions(+), 25 deletions(-)
>
> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.c
> index 1d201fd3f4af..65aee57b35fe 100644
> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.c
> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.c
> @@ -79,6 +79,29 @@ uint64_t amdgpu_gmc_pd_addr(struct amdgpu_bo *bo)
>   	return pd_addr;
>   }
>
> +/**
> + * amdgpu_gmc_agp_addr - return the address in the AGP address space
> + *
> + * @tbo: TTM BO which needs the address, must be in GTT domain
> + *
> + * Tries to figure out how to access the BO through the AGP aperture. Returns
> + * AMDGPU_BO_INVALID_OFFSET if that is not possible.
> + */
> +uint64_t amdgpu_gmc_agp_addr(struct ttm_buffer_object *bo)
> +{
> +	struct amdgpu_device *adev = amdgpu_ttm_adev(bo->bdev);
> +	struct ttm_dma_tt *ttm;
> +
> +	if (bo->num_pages != 1 || bo->ttm->caching_state == tt_cached)
> +		return AMDGPU_BO_INVALID_OFFSET;

If GTT bo size is 1 page, it will also access in AGP address space?

Jerry
> +
> +	ttm = container_of(bo->ttm, struct ttm_dma_tt, ttm);
> +	if (ttm->dma_address[0] + PAGE_SIZE >= adev->gmc.agp_size)
> +		return AMDGPU_BO_INVALID_OFFSET;
> +
> +	return adev->gmc.agp_start + ttm->dma_address[0];
> +}
> +
>   /**
>    * amdgpu_gmc_vram_location - try to find VRAM location
>    *
> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.h b/drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.h
> index c9985e7dc9e5..265ca415c64c 100644
> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.h
> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_gmc.h
> @@ -163,6 +163,7 @@ static inline uint64_t amdgpu_gmc_sign_extend(uint64_t addr)
>   void amdgpu_gmc_get_pde_for_bo(struct amdgpu_bo *bo, int level,
>   			       uint64_t *addr, uint64_t *flags);
>   uint64_t amdgpu_gmc_pd_addr(struct amdgpu_bo *bo);
> +uint64_t amdgpu_gmc_agp_addr(struct ttm_buffer_object *bo);
>   void amdgpu_gmc_vram_location(struct amdgpu_device *adev, struct amdgpu_gmc *mc,
>   			      u64 base);
>   void amdgpu_gmc_gart_location(struct amdgpu_device *adev,
> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
> index d9f3201c9e5c..8a158ee922f7 100644
> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_ttm.c
> @@ -1081,41 +1081,49 @@ int amdgpu_ttm_alloc_gart(struct ttm_buffer_object *bo)
>   	struct ttm_mem_reg tmp;
>   	struct ttm_placement placement;
>   	struct ttm_place placements;
> -	uint64_t flags;
> +	uint64_t addr, flags;
>   	int r;
>
>   	if (bo->mem.start != AMDGPU_BO_INVALID_OFFSET)
>   		return 0;
>
> -	/* allocate GART space */
> -	tmp = bo->mem;
> -	tmp.mm_node = NULL;
> -	placement.num_placement = 1;
> -	placement.placement = &placements;
> -	placement.num_busy_placement = 1;
> -	placement.busy_placement = &placements;
> -	placements.fpfn = 0;
> -	placements.lpfn = adev->gmc.gart_size >> PAGE_SHIFT;
> -	placements.flags = (bo->mem.placement & ~TTM_PL_MASK_MEM) |
> -		TTM_PL_FLAG_TT;
> +	addr = amdgpu_gmc_agp_addr(bo);
> +	if (addr != AMDGPU_BO_INVALID_OFFSET) {
> +		bo->mem.start = addr >> PAGE_SHIFT;
> +	} else {
>
> -	r = ttm_bo_mem_space(bo, &placement, &tmp, &ctx);
> -	if (unlikely(r))
> -		return r;
> +		/* allocate GART space */
> +		tmp = bo->mem;
> +		tmp.mm_node = NULL;
> +		placement.num_placement = 1;
> +		placement.placement = &placements;
> +		placement.num_busy_placement = 1;
> +		placement.busy_placement = &placements;
> +		placements.fpfn = 0;
> +		placements.lpfn = adev->gmc.gart_size >> PAGE_SHIFT;
> +		placements.flags = (bo->mem.placement & ~TTM_PL_MASK_MEM) |
> +			TTM_PL_FLAG_TT;
> +
> +		r = ttm_bo_mem_space(bo, &placement, &tmp, &ctx);
> +		if (unlikely(r))
> +			return r;
>
> -	/* compute PTE flags for this buffer object */
> -	flags = amdgpu_ttm_tt_pte_flags(adev, bo->ttm, &tmp);
> +		/* compute PTE flags for this buffer object */
> +		flags = amdgpu_ttm_tt_pte_flags(adev, bo->ttm, &tmp);
>
> -	/* Bind pages */
> -	gtt->offset = ((u64)tmp.start << PAGE_SHIFT) - adev->gmc.gart_start;
> -	r = amdgpu_ttm_gart_bind(adev, bo, flags);
> -	if (unlikely(r)) {
> -		ttm_bo_mem_put(bo, &tmp);
> -		return r;
> +		/* Bind pages */
> +		gtt->offset = ((u64)tmp.start << PAGE_SHIFT) -
> +			adev->gmc.gart_start;
> +		r = amdgpu_ttm_gart_bind(adev, bo, flags);
> +		if (unlikely(r)) {
> +			ttm_bo_mem_put(bo, &tmp);
> +			return r;
> +		}
> +
> +		ttm_bo_mem_put(bo, &bo->mem);
> +		bo->mem = tmp;
>   	}
>
> -	ttm_bo_mem_put(bo, &bo->mem);
> -	bo->mem = tmp;
>   	bo->offset = (bo->mem.start << PAGE_SHIFT) +
>   		bo->bdev->man[bo->mem.mem_type].gpu_offset;
>
>


More information about the amd-gfx mailing list