[PATCH 5/5] nouveau/ttm/PCIe: Use dma_addr if TTM has set it.

Konrad Rzeszutek Wilk konrad.wilk at oracle.com
Thu Jan 27 13:22:12 PST 2011


On Fri, Jan 07, 2011 at 12:11:44PM -0500, Konrad Rzeszutek Wilk wrote:
> If the TTM layer has used the DMA API to setup pages that are
> TTM_PAGE_FLAG_DMA32 (look at patch titled: "ttm: Utilize the dma_addr_t
> array for pages that are to in DMA32 pool."), lets use it
> when programming the GART in the PCIe type cards.
> 
> This patch skips doing the pci_map_page (and pci_unmap_page) if
> there is a DMA addresses passed in for that page. If the dma_address
> is zero (or DMA_ERROR_CODE), then we continue on with our old
> behaviour.

Hey Ben and Jerome,

I should have CC-ed you guys earlier but missed that and instead just
CC-ed the mailing list. I was wondering what your thoughts are
about this patchset? Thomas took a look at the patchset and he is OK
but more eyes never hurt.


> 
> Signed-off-by: Konrad Rzeszutek Wilk <konrad.wilk at oracle.com>
> ---
>  drivers/gpu/drm/nouveau/nouveau_sgdma.c |   28 +++++++++++++++++++++-------
>  1 files changed, 21 insertions(+), 7 deletions(-)
> 
> diff --git a/drivers/gpu/drm/nouveau/nouveau_sgdma.c b/drivers/gpu/drm/nouveau/nouveau_sgdma.c
> index edc140a..bbdd982 100644
> --- a/drivers/gpu/drm/nouveau/nouveau_sgdma.c
> +++ b/drivers/gpu/drm/nouveau/nouveau_sgdma.c
> @@ -12,6 +12,7 @@ struct nouveau_sgdma_be {
>  	struct drm_device *dev;
>  
>  	dma_addr_t *pages;
> +	bool *ttm_alloced;
>  	unsigned nr_pages;
>  
>  	unsigned pte_start;
> @@ -35,15 +36,25 @@ nouveau_sgdma_populate(struct ttm_backend *be, unsigned long num_pages,
>  	if (!nvbe->pages)
>  		return -ENOMEM;
>  
> +	nvbe->ttm_alloced = kmalloc(sizeof(bool) * num_pages, GFP_KERNEL);
> +	if (!nvbe->ttm_alloced)
> +		return -ENOMEM;
> +
>  	nvbe->nr_pages = 0;
>  	while (num_pages--) {
> -		nvbe->pages[nvbe->nr_pages] =
> -			pci_map_page(dev->pdev, pages[nvbe->nr_pages], 0,
> +		if (dma_addrs[nvbe->nr_pages] != DMA_ERROR_CODE) {
> +			nvbe->pages[nvbe->nr_pages] =
> +					dma_addrs[nvbe->nr_pages];
> +		 	nvbe->ttm_alloced[nvbe->nr_pages] = true;
> +		} else {
> +			nvbe->pages[nvbe->nr_pages] =
> +				pci_map_page(dev->pdev, pages[nvbe->nr_pages], 0,
>  				     PAGE_SIZE, PCI_DMA_BIDIRECTIONAL);
> -		if (pci_dma_mapping_error(dev->pdev,
> -					  nvbe->pages[nvbe->nr_pages])) {
> -			be->func->clear(be);
> -			return -EFAULT;
> +			if (pci_dma_mapping_error(dev->pdev,
> +						  nvbe->pages[nvbe->nr_pages])) {
> +				be->func->clear(be);
> +				return -EFAULT;
> +			}
>  		}
>  
>  		nvbe->nr_pages++;
> @@ -66,11 +77,14 @@ nouveau_sgdma_clear(struct ttm_backend *be)
>  			be->func->unbind(be);
>  
>  		while (nvbe->nr_pages--) {
> -			pci_unmap_page(dev->pdev, nvbe->pages[nvbe->nr_pages],
> +			if (!nvbe->ttm_alloced[nvbe->nr_pages])
> +				pci_unmap_page(dev->pdev, nvbe->pages[nvbe->nr_pages],
>  				       PAGE_SIZE, PCI_DMA_BIDIRECTIONAL);
>  		}
>  		kfree(nvbe->pages);
> +		kfree(nvbe->ttm_alloced);
>  		nvbe->pages = NULL;
> +		nvbe->ttm_alloced = NULL;
>  		nvbe->nr_pages = 0;
>  	}
>  }
> -- 
> 1.7.1


More information about the dri-devel mailing list