[PATCH 21/59] drm/vmwgfx/ttm: switch gmrid allocator to new init paths.

daniel at ffwll.ch daniel at ffwll.ch
Wed Aug 5 09:00:15 UTC 2020


On Tue, Aug 04, 2020 at 12:55:54PM +1000, Dave Airlie wrote:
> From: Dave Airlie <airlied at redhat.com>
> 
> Signed-off-by: Dave Airlie <airlied at redhat.com>
> ---
>  drivers/gpu/drm/vmwgfx/vmwgfx_drv.c           | 17 ++++-----------
>  drivers/gpu/drm/vmwgfx/vmwgfx_drv.h           |  2 +-
>  drivers/gpu/drm/vmwgfx/vmwgfx_gmrid_manager.c | 21 ++++++++++++-------
>  3 files changed, 19 insertions(+), 21 deletions(-)
> 
> diff --git a/drivers/gpu/drm/vmwgfx/vmwgfx_drv.c b/drivers/gpu/drm/vmwgfx/vmwgfx_drv.c
> index 5a889df2de03..364d5f3ff9a8 100644
> --- a/drivers/gpu/drm/vmwgfx/vmwgfx_drv.c
> +++ b/drivers/gpu/drm/vmwgfx/vmwgfx_drv.c
> @@ -896,14 +896,10 @@ static int vmw_driver_load(struct drm_device *dev, unsigned long chipset)
>  	 *  slots as well as the bo size.
>  	 */
>  	dev_priv->has_gmr = true;
> -	dev_priv->bdev.man[VMW_PL_GMR].func = &vmw_gmrid_manager_func;
> -	dev_priv->bdev.man[VMW_PL_GMR].available_caching = TTM_PL_FLAG_CACHED;
> -	dev_priv->bdev.man[VMW_PL_GMR].default_caching = TTM_PL_FLAG_CACHED;
>  	/* TODO: This is most likely not correct */
> -	dev_priv->bdev.man[VMW_PL_GMR].use_tt = true;
>  	if (((dev_priv->capabilities & (SVGA_CAP_GMR | SVGA_CAP_GMR2)) == 0) ||
> -	    refuse_dma || ttm_bo_init_mm(&dev_priv->bdev, VMW_PL_GMR,
> -					 VMW_PL_GMR) != 0) {
> +	    refuse_dma ||
> +	    vmw_gmrid_man_init(dev_priv, VMW_PL_GMR) != 0) {
>  		DRM_INFO("No GMR memory available. "
>  			 "Graphics memory resources are very limited.\n");
>  		dev_priv->has_gmr = false;
> @@ -911,13 +907,8 @@ static int vmw_driver_load(struct drm_device *dev, unsigned long chipset)
>  
>  	if (dev_priv->capabilities & SVGA_CAP_GBOBJECTS && !refuse_dma) {
>  		dev_priv->has_mob = true;
> -		dev_priv->bdev.man[VMW_PL_MOB].func = &vmw_gmrid_manager_func;
> -		dev_priv->bdev.man[VMW_PL_MOB].available_caching = TTM_PL_FLAG_CACHED;
> -		dev_priv->bdev.man[VMW_PL_MOB].default_caching = TTM_PL_FLAG_CACHED;
> -		/* TODO: This is most likely not correct */
> -		dev_priv->bdev.man[VMW_PL_MOB].use_tt = true;
> -		if (ttm_bo_init_mm(&dev_priv->bdev, VMW_PL_MOB,
> -				   VMW_PL_MOB) != 0) {
> +
> +		if (vmw_gmrid_man_init(dev_priv, VMW_PL_MOB) != 0) {
>  			DRM_INFO("No MOB memory available. "
>  				 "3D will be disabled.\n");
>  			dev_priv->has_mob = false;
> diff --git a/drivers/gpu/drm/vmwgfx/vmwgfx_drv.h b/drivers/gpu/drm/vmwgfx/vmwgfx_drv.h
> index 10b681725a53..8f319dd6cdb4 100644
> --- a/drivers/gpu/drm/vmwgfx/vmwgfx_drv.h
> +++ b/drivers/gpu/drm/vmwgfx/vmwgfx_drv.h
> @@ -1221,7 +1221,7 @@ int vmw_overlay_num_free_overlays(struct vmw_private *dev_priv);
>   * GMR Id manager
>   */
>  
> -extern const struct ttm_mem_type_manager_func vmw_gmrid_manager_func;
> +int vmw_gmrid_man_init(struct vmw_private *dev_priv, int type);
>  
>  /**
>   * Prime - vmwgfx_prime.c
> diff --git a/drivers/gpu/drm/vmwgfx/vmwgfx_gmrid_manager.c b/drivers/gpu/drm/vmwgfx/vmwgfx_gmrid_manager.c
> index fb1bf4dd91d1..141fb14e3583 100644
> --- a/drivers/gpu/drm/vmwgfx/vmwgfx_gmrid_manager.c
> +++ b/drivers/gpu/drm/vmwgfx/vmwgfx_gmrid_manager.c
> @@ -94,22 +94,28 @@ static void vmw_gmrid_man_put_node(struct ttm_mem_type_manager *man,
>  	}
>  }
>  
> -static int vmw_gmrid_man_init(struct ttm_mem_type_manager *man,
> -			      unsigned long p_size)
> +static const struct ttm_mem_type_manager_func vmw_gmrid_manager_func;
> +
> +int vmw_gmrid_man_init(struct vmw_private *dev_priv, int type)
>  {
> -	struct vmw_private *dev_priv =
> -		container_of(man->bdev, struct vmw_private, bdev);
> +	struct ttm_mem_type_manager *man = &dev_priv->bdev.man[type];
>  	struct vmwgfx_gmrid_man *gman =
>  		kzalloc(sizeof(*gman), GFP_KERNEL);
>  
>  	if (unlikely(!gman))
>  		return -ENOMEM;
>  
> +	man->func = &vmw_gmrid_manager_func;
> +	man->available_caching = TTM_PL_FLAG_CACHED;
> +	man->default_caching = TTM_PL_FLAG_CACHED;
> +	/* TODO: This is most likely not correct */
> +	man->use_tt = true;
> +	ttm_mem_type_manager_init(&dev_priv->bdev, man, 0);
>  	spin_lock_init(&gman->lock);
>  	gman->used_gmr_pages = 0;
>  	ida_init(&gman->gmr_ida);
>  
> -	switch (p_size) {
> +	switch (type) {
>  	case VMW_PL_GMR:
>  		gman->max_gmr_ids = dev_priv->max_gmr_ids;
>  		gman->max_gmr_pages = dev_priv->max_gmr_pages;
> @@ -122,6 +128,8 @@ static int vmw_gmrid_man_init(struct ttm_mem_type_manager *man,
>  		BUG();
>  	}
>  	man->priv = (void *) gman;
> +
> +	ttm_mem_type_manager_set_used(man, true);
>  	return 0;
>  }
>  
> @@ -137,8 +145,7 @@ static int vmw_gmrid_man_takedown(struct ttm_mem_type_manager *man)
>  	return 0;
>  }
>  
> -const struct ttm_mem_type_manager_func vmw_gmrid_manager_func = {
> -	.init = vmw_gmrid_man_init,
> +static const struct ttm_mem_type_manager_func vmw_gmrid_manager_func = {
>  	.takedown = vmw_gmrid_man_takedown,
>  	.get_node = vmw_gmrid_man_get_node,
>  	.put_node = vmw_gmrid_man_put_node,


Reviewed-by: Daniel Vetter <daniel.vetter at ffwll.ch>

> -- 
> 2.26.2
> 
> _______________________________________________
> dri-devel mailing list
> dri-devel at lists.freedesktop.org
> https://lists.freedesktop.org/mailman/listinfo/dri-devel

-- 
Daniel Vetter
Software Engineer, Intel Corporation
http://blog.ffwll.ch


More information about the dri-devel mailing list