[v4,1/3] drm/prime: use dma length macro when mapping sg
Marek Szyprowski
m.szyprowski at samsung.com
Fri Mar 27 07:54:53 UTC 2020
Hi All,
On 2020-03-25 10:07, Shane Francis wrote:
> As dma_map_sg can reorganize scatter-gather lists in a
> way that can cause some later segments to be empty we should
> always use the sg_dma_len macro to fetch the actual length.
>
> This could now be 0 and not need to be mapped to a page or
> address array
>
> Signed-off-by: Shane Francis <bigbeeshane at gmail.com>
> Reviewed-by: Michael J. Ruhl <michael.j.ruhl at intel.com>
This patch landed in linux-next 20200326 and it causes a kernel panic on
various Exynos SoC based boards.
> ---
> drivers/gpu/drm/drm_prime.c | 2 +-
> 1 file changed, 1 insertion(+), 1 deletion(-)
>
> diff --git a/drivers/gpu/drm/drm_prime.c b/drivers/gpu/drm/drm_prime.c
> index 86d9b0e45c8c..1de2cde2277c 100644
> --- a/drivers/gpu/drm/drm_prime.c
> +++ b/drivers/gpu/drm/drm_prime.c
> @@ -967,7 +967,7 @@ int drm_prime_sg_to_page_addr_arrays(struct sg_table *sgt, struct page **pages,
>
> index = 0;
> for_each_sg(sgt->sgl, sg, sgt->nents, count) {
> - len = sg->length;
> + len = sg_dma_len(sg);
> page = sg_page(sg);
> addr = sg_dma_address(sg);
>
Sorry, but this code is wrong :(
The scatterlist elements (sg) describes memory chunks in physical memory
and in the DMA (IO virtual) space. However in general, you cannot assume
1:1 mapping between them. If you access sg_page(sg) (basically
sg->page), you must match it with sg->length. When you access
sg_dma_address(sg) (again, in most cases it is sg->dma_address), then
you must match it with sg_dma_len(sg). The sg->dma_address might not be
the dma address of the sg->page.
In some cases (when IOMMU is available, it performs aggregation of the
scatterlist chunks and a few other, minor requirements), the whole
scatterlist might be mapped into contiguous DMA address space and filled
only to the first sg element.
The proper way to iterate over a scatterlists to get both the pages and
the DMA addresses assigned to them is:
int drm_prime_sg_to_page_addr_arrays(struct sg_table *sgt, struct page
**pages,
dma_addr_t *addrs, int max_entries)
{
unsigned count;
struct scatterlist *sg;
struct page *page;
u32 page_len, page_index;
dma_addr_t addr;
u32 dma_len, dma_index;
page_index = 0;
dma_index = 0;
for_each_sg(sgt->sgl, sg, sgt->nents, count) {
page_len = sg->length;
page = sg_page(sg);
dma_len = sg_dma_len(sg);
addr = sg_dma_address(sg);
while (pages && page_len > 0) {
if (WARN_ON(page_index >= max_entries))
return -1;
pages[page_index] = page;
page++;
page_len -= PAGE_SIZE;
page_index++;
}
while (addrs && dma_len > 0) {
if (WARN_ON(dma_index >= max_entries))
return -1;
addrs[dma_index] = addr;
addr += PAGE_SIZE;
dma_len -= PAGE_SIZE;
dma_index++;
}
}
return 0;
}
I will send a patch in a few minutes with the above fixed code.
Best regards
--
Marek Szyprowski, PhD
Samsung R&D Institute Poland
More information about the dri-devel
mailing list