[igt-dev] [PATCH i-g-t v5 1/3] lib/intel_bufops: Introduce buffer operations

Katarzyna Dec katarzyna.dec at intel.com
Wed Jan 8 11:58:28 UTC 2020


On Wed, Jan 08, 2020 at 08:54:09AM +0100, Zbigniew Kempczyński wrote:
> Different GENs supports different tile surfaces. Older GENs
> have HW fences to allow X / Y surface tiling / detiling.
> Newer GENs have to tile / detile such surface in software.
> 
> To make test developer life easier this code adds buffer
> operations (short buf_ops) to use appropriate functions allowing
> copying linear buffer to BO and from BO to linear buffer
> regardless GPU generation and tiling/swizzling within BO. For GENs
> having fences support preference is to use them if they are available
> (X / Y tiling is probed on buf_ops initialization).
> 
> Note: on Gen2 code supports only HW tiling at the moment.
> 
> Signed-off-by: Zbigniew Kempczyński <zbigniew.kempczynski at intel.com>
> Cc: Chris Wilson <chris at chris-wilson.co.uk>
> Cc: Katarzyna Dec <katarzyna.dec at intel.com>
> ---
>  .../igt-gpu-tools/igt-gpu-tools-docs.xml      |    1 +
>  lib/Makefile.sources                          |    2 +
>  lib/intel_bufops.c                            | 1120 +++++++++++++++++
>  lib/intel_bufops.h                            |   84 ++
>  lib/meson.build                               |    1 +
>  5 files changed, 1208 insertions(+)
>  create mode 100644 lib/intel_bufops.c
>  create mode 100644 lib/intel_bufops.h
> 
> diff --git a/docs/reference/igt-gpu-tools/igt-gpu-tools-docs.xml b/docs/reference/igt-gpu-tools/igt-gpu-tools-docs.xml
> index 454d64e5..aa9fef20 100644
> --- a/docs/reference/igt-gpu-tools/igt-gpu-tools-docs.xml
> +++ b/docs/reference/igt-gpu-tools/igt-gpu-tools-docs.xml
> @@ -42,6 +42,7 @@
>      <xi:include href="xml/igt_vgem.xml"/>
>      <xi:include href="xml/igt_x86.xml"/>
>      <xi:include href="xml/intel_batchbuffer.xml"/>
> +    <xi:include href="xml/intel_bufops.xml"/>
>      <xi:include href="xml/intel_chipset.xml"/>
>      <xi:include href="xml/intel_io.xml"/>
>      <xi:include href="xml/ioctl_wrappers.xml"/>
> diff --git a/lib/Makefile.sources b/lib/Makefile.sources
> index 5dd3962e..da34eae7 100644
> --- a/lib/Makefile.sources
> +++ b/lib/Makefile.sources
> @@ -67,6 +67,8 @@ lib_source_list =	 	\
>  	intel_aub.h		\
>  	intel_batchbuffer.c	\
>  	intel_batchbuffer.h	\
> +	intel_bufops.c		\
> +	intel_bufops.h		\
>  	intel_chipset.c		\
>  	intel_chipset.h		\
>  	intel_device_info.c	\
> diff --git a/lib/intel_bufops.c b/lib/intel_bufops.c
> new file mode 100644
> index 00000000..ffd3494a
> --- /dev/null
> +++ b/lib/intel_bufops.c
> @@ -0,0 +1,1120 @@
> +/*
> + * Copyright © 2020 Intel Corporation
> + *
> + * Permission is hereby granted, free of charge, to any person obtaining a
> + * copy of this software and associated documentation files (the "Software"),
> + * to deal in the Software without restriction, including without limitation
> + * the rights to use, copy, modify, merge, publish, distribute, sublicense,
> + * and/or sell copies of the Software, and to permit persons to whom the
> + * Software is furnished to do so, subject to the following conditions:
> + *
> + * The above copyright notice and this permission notice (including the next
> + * paragraph) shall be included in all copies or substantial portions of the
> + * Software.
> + *
> + * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
> + * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
> + * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.  IN NO EVENT SHALL
> + * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
> + * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
> + * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS
> + * IN THE SOFTWARE.
> + *
> + */
> +
> +#include <sys/ioctl.h>
> +#include "igt.h"
> +#include "igt_x86.h"
> +#include "intel_bufops.h"
> +
> +/**
> + * SECTION:intel_bufops
> + * @short_description: Buffer operation on tiled surfaces
> + * @title: Buffer operations
> + * @include: igt.h
> + *
> + * # Buffer operations
> + *
> + * Intel GPU devices supports different set of tiled surfaces.
> + * Checking each time what tile formats are supports is cumbersome and
> + * error prone.
> + *
> + * Buffer operation (buf_ops) provide a wrapper to conditional code which
> + * can be used without worrying of implementation details giving:
> + * - copy linear to tiled buffer
> + * - copy tiled buffer to linear
> + *
> + * Following code order should be used (linear is plain memory with some
> + * image data):
> + *
> + * |[<!-- language="c" -->
> + * struct buf_ops *bops;
> + * struct intel_buf ibuf;
> + * ...
> + * bops = buf_ops_create(fd);
> + * intel_buf_init(bops, &ibuf, 512, 512, 32, I915_TILING_X, false);
> + * ...
> + * linear_to_intel_buf(bops, &ibuf, linear);
> + * ...
> + * intel_buf_to_linear(bops, &ibuf, linear);
> + * ...
> + * intel_buf_close(bops, &ibuf);
> + * ...
> + * buf_ops_destroy(bops);
> + * ]|
> + *
> + * Calling buf_ops_create(fd) probes hardware capabilities (supported fences,
> + * swizzling) and returns opaque pointer to buf_ops. From now on
> + * intel_buf_to_linear() and linear_to_intel_buf() will choose appropriate
> + * function to do the job.
> + *
> + * Note: bufops doesn't support SW tiling code yet.
> + */
> +
> +//#define BUFOPS_DEBUGGING
> +#ifdef BUFOPS_DEBUGGING
> +#define DEBUG(...) printf(__VA_ARGS__)
> +#define DEBUGFN() DEBUG("\t -> %s()\n", __FUNCTION__)
> +#else
> +#define DEBUG(...)
> +#define DEBUGFN()
> +#endif
> +
> +#define TILE_DEF(x) (1 << (x))
> +#define TILE_NONE   TILE_DEF(I915_TILING_NONE)
> +#define TILE_X      TILE_DEF(I915_TILING_X)
> +#define TILE_Y      TILE_DEF(I915_TILING_Y)
> +#define TILE_Yf     TILE_DEF(I915_TILING_Yf)
> +#define TILE_Ys     TILE_DEF(I915_TILING_Ys)
> +
> +typedef void (*bo_copy)(struct buf_ops *, struct intel_buf *, uint32_t *);
> +
> +struct buf_ops {
> +	int fd;
> +	int gen_start;
> +	int gen_end;
> +	int intel_gen;
> +	uint32_t supported_tiles;
> +	uint32_t supported_hw_tiles;
> +	uint32_t swizzle_x;
> +	uint32_t swizzle_y;
> +	bo_copy linear_to;
> +	bo_copy linear_to_x;
> +	bo_copy linear_to_y;
> +	bo_copy linear_to_yf;
> +	bo_copy linear_to_ys;
> +	bo_copy to_linear;
> +	bo_copy x_to_linear;
> +	bo_copy y_to_linear;
> +	bo_copy yf_to_linear;
> +	bo_copy ys_to_linear;
> +};
> +
> +static const char *tiling_str(uint32_t tiling)
> +{
> +	switch (tiling) {
> +	case I915_TILING_NONE: return "NONE";
> +	case I915_TILING_X:    return "X";
> +	case I915_TILING_Y:    return "Y";
> +	case I915_TILING_Yf:   return "Yf";
> +	case I915_TILING_Ys:   return "Ys";
> +	default:               return "UNKNOWN";
> +	}
> +}
> +
> +static const char *bool_str(bool v)
> +{
> +	return v ? "yes" : "no";
> +}
> +
> +static inline bool is_hw_tiling_supported(struct buf_ops *bops, uint32_t tiling)
> +{
> +	return bops->supported_hw_tiles & TILE_DEF(tiling);
> +}
> +
> +static inline bool is_tiling_supported(struct buf_ops *bops, uint32_t tiling)
> +{
> +	return bops->supported_tiles & TILE_DEF(tiling);
> +}
> +
> +static int __gem_get_tiling(int fd, struct drm_i915_gem_get_tiling *arg)
> +{
> +	int err;
> +
> +	err = 0;
> +	if (igt_ioctl(fd, DRM_IOCTL_I915_GEM_GET_TILING, arg))
> +		err = -errno;
> +	errno = 0;
> +
> +	return err;
> +}
> +
> +static bool __get_tiling(int fd, uint32_t handle, uint32_t *tiling,
> +			 uint32_t *swizzle)
> +{
> +	struct drm_i915_gem_get_tiling get_tiling;
> +
> +	memset(&get_tiling, 0, sizeof(get_tiling));
> +	get_tiling.handle = handle;
> +
> +	if (__gem_get_tiling(fd, &get_tiling) != 0)
> +		return false;
> +
> +	*tiling = get_tiling.tiling_mode;
> +	*swizzle = get_tiling.swizzle_mode;
> +	igt_debug("buf tiling: %s, swizzle: %x, phys_swizzle: %x\n",
> +		  tiling_str(get_tiling.tiling_mode),
> +		  get_tiling.swizzle_mode,
> +		  get_tiling.phys_swizzle_mode);
> +
> +	return get_tiling.phys_swizzle_mode == get_tiling.swizzle_mode;
> +}
> +
> +static int __set_tiling(int fd, uint32_t handle, uint32_t tiling,
> +			uint32_t stride)
> +{
> +	struct drm_i915_gem_set_tiling st;
> +
> +	st.handle = handle;
> +	st.tiling_mode = tiling;
> +	st.stride = tiling ? stride : 0;
> +
> +	return ioctl(fd, DRM_IOCTL_I915_GEM_SET_TILING, &st);
> +}
> +
> +static void set_hw_tiled(struct buf_ops *bops, struct intel_buf *buf)
> +{
> +	int ret;
> +
> +	if (buf->tiling != I915_TILING_X && buf->tiling != I915_TILING_Y)
> +		return;
> +
> +	if (!buf_ops_has_hw_fence(bops, buf->tiling))
> +		return;
> +
> +	ret = __set_tiling(bops->fd, buf->handle, buf->tiling, buf->stride);
> +	igt_assert(ret == 0);
> +}
> +
> +static unsigned long swizzle_bit(unsigned int bit, unsigned long offset)
> +{
> +	return (offset & (1ul << bit)) >> (bit - 6);
> +}
> +
> +static unsigned long swizzle_addr(void *ptr, uint32_t swizzle)
> +{
> +	unsigned long addr = to_user_pointer(ptr);
> +
> +	switch (swizzle) {
> +	case I915_BIT_6_SWIZZLE_NONE:
> +		return addr;
> +	case I915_BIT_6_SWIZZLE_9:
> +		return addr ^ swizzle_bit(9, addr);
> +	case I915_BIT_6_SWIZZLE_9_10:
> +		return addr ^ swizzle_bit(9, addr) ^ swizzle_bit(10, addr);
> +	case I915_BIT_6_SWIZZLE_9_11:
> +		return addr ^ swizzle_bit(9, addr) ^ swizzle_bit(11, addr);
> +	case I915_BIT_6_SWIZZLE_9_10_11:
> +		return (addr ^
> +			swizzle_bit(9, addr) ^
> +			swizzle_bit(10, addr) ^
> +			swizzle_bit(11, addr));
> +
> +	case I915_BIT_6_SWIZZLE_UNKNOWN:
> +	case I915_BIT_6_SWIZZLE_9_17:
> +	case I915_BIT_6_SWIZZLE_9_10_17:
> +	default:
> +		/* If we hit this case, we need to implement support for the
> +		 * appropriate swizzling method. */
> +		igt_require(false);
> +		return addr;
> +	}
> +}
> +
> +static void *x_ptr(void *ptr,
> +		   unsigned int x, unsigned int y,
> +		   unsigned int stride, unsigned int cpp)
> +{
> +	const int tile_width = 512;
> +	const int tile_height = 8;
> +	const int tile_size = tile_width * tile_height;
> +	int tile_x, tile_y;
> +	int offset_x, offset_y, pos;
> +
> +	x *= cpp;
> +	tile_x = x / tile_width;
> +	tile_y = y / tile_height;
> +	offset_x = (tile_x * tile_size);
> +	offset_y = (tile_y * stride * tile_height);
> +
> +	pos = offset_y + offset_x +
> +			(y % tile_height * tile_width) + (x % tile_width);
> +
> +	return ptr + pos;
> +}
> +
> +static void *y_ptr(void *ptr,
> +		   unsigned int x, unsigned int y,
> +		   unsigned int stride, unsigned int cpp)
> +{
> +	const int tile_width = 128;
> +	const int tile_height = 32;
> +	const int owords = 16;
> +	const int tile_size = tile_width * tile_height;
> +	int tile_x, tile_y;
> +	int offset_x, offset_y, pos;
> +	int shift_x, shift_y;
> +
> +	x *= cpp;
> +	tile_x = x / tile_width;
> +	tile_y = y / tile_height;
> +	offset_x = tile_x * tile_size;
> +	offset_y = tile_y * stride * tile_height;
> +	shift_x = x % owords + (x % tile_width) / owords * tile_width * cpp;
> +	shift_y = y % tile_height * owords;
> +
> +	pos = offset_y + offset_x + shift_x + shift_y;
> +
> +	return ptr + pos;
> +}
> +
> +static void *yf_ptr(void *ptr,
> +		    unsigned int x, unsigned int y,
> +		    unsigned int stride, unsigned int cpp)
> +{
> +	const int tile_size = 4 * 1024;
> +	const int tile_width = 128;
> +	int row_size = (stride / tile_width) * tile_size;
> +
> +	x *= cpp; /* convert to Byte offset */
> +
> +
> +	/*
> +	 * Within a 4k Yf tile, the byte swizzling pattern is
> +	 * msb......lsb
> +	 * xyxyxyyyxxxx
> +	 * The tiles themselves are laid out in row major order.
> +	 */
> +	return ptr +
> +			((x & 0xf) * 1) + /* 4x1 pixels(32bpp) = 16B */
> +			((y & 0x3) * 16) + /* 4x4 pixels = 64B */
> +			(((y & 0x4) >> 2) * 64) + /* 1x2 64B blocks */
> +			(((x & 0x10) >> 4) * 128) + /* 2x2 64B blocks = 256B block */
> +			(((y & 0x8) >> 3) * 256) + /* 2x1 256B blocks */
> +			(((x & 0x20) >> 5) * 512) + /* 2x2 256B blocks */
> +			(((y & 0x10) >> 4) * 1024) + /* 4x2 256 blocks */
> +			(((x & 0x40) >> 6) * 2048) + /* 4x4 256B blocks = 4k tile */
> +			(((x & ~0x7f) >> 7) * tile_size) + /* row of tiles */
> +			(((y & ~0x1f) >> 5) * row_size);
> +}
> +
> +typedef void *(*fn_ptr)(void *, unsigned int, unsigned int,
> +			unsigned int, unsigned int);
> +static fn_ptr __get_tile_fn_ptr(int tiling)
> +{
> +	fn_ptr fn = NULL;
> +
> +	switch (tiling) {
> +	case I915_TILING_X:
> +		fn = x_ptr;
> +		break;
> +	case I915_TILING_Y:
> +		fn = y_ptr;
> +		break;
> +	case I915_TILING_Yf:
> +		fn = yf_ptr;
> +		break;
> +	case I915_TILING_Ys:
> +		/* To be implemented */
> +		break;
> +	}
> +
> +	igt_assert_f(fn, "Can't find tile function for tiling: %d\n", tiling);
> +
> +	return fn;
> +}
> +
> +static void __copy_linear_to(int fd, struct intel_buf *buf,
> +			     const uint32_t *linear,
> +			     int tiling, uint32_t swizzle)
> +{
> +	int height = intel_buf_height(buf);
> +	int width = intel_buf_width(buf);
> +	fn_ptr fn = __get_tile_fn_ptr(tiling);
> +	void *map;
> +
> +	gem_set_domain(fd, buf->handle,
> +		       I915_GEM_DOMAIN_CPU, I915_GEM_DOMAIN_CPU);
> +
> +	map = __gem_mmap_offset__cpu(fd, buf->handle, 0,
> +				     buf->size, PROT_READ | PROT_WRITE);
> +	if (!map)
> +		map = gem_mmap__cpu(fd, buf->handle, 0,
> +				    buf->size, PROT_READ | PROT_WRITE);
> +
> +	for (int y = 0; y < height; y++) {
> +		for (int x = 0; x < width; x++) {
> +			uint32_t *ptr = fn(map, x, y, buf->stride, buf->bpp/8);
> +
> +			if (swizzle)
> +				ptr = from_user_pointer(swizzle_addr(ptr,
> +								     swizzle));
> +			*ptr = linear[y * width + x];
> +		}
> +	}
> +
> +	munmap(map, buf->size);
> +}
> +
> +static void copy_linear_to_x(struct buf_ops *bops, struct intel_buf *buf,
> +			     uint32_t *linear)
> +{
> +	DEBUGFN();
> +	__copy_linear_to(bops->fd, buf, linear, I915_TILING_X, bops->swizzle_x);
> +}
> +
> +static void copy_linear_to_y(struct buf_ops *bops, struct intel_buf *buf,
> +			     uint32_t *linear)
> +{
> +	DEBUGFN();
> +	__copy_linear_to(bops->fd, buf, linear, I915_TILING_Y, bops->swizzle_y);
> +}
> +
> +static void copy_linear_to_yf(struct buf_ops *bops, struct intel_buf *buf,
> +			      uint32_t *linear)
> +{
> +	DEBUGFN();
> +	__copy_linear_to(bops->fd, buf, linear, I915_TILING_Yf, 0);
> +}
> +
> +static void copy_linear_to_ys(struct buf_ops *bops, struct intel_buf *buf,
> +			      uint32_t *linear)
> +{
> +	DEBUGFN();
> +	__copy_linear_to(bops->fd, buf, linear, I915_TILING_Ys, 0);
> +}
> +
> +static void __copy_to_linear(int fd, struct intel_buf *buf,
> +			     uint32_t *linear, int tiling, uint32_t swizzle)
> +{
> +	int height = intel_buf_height(buf);
> +	int width = intel_buf_width(buf);
> +	fn_ptr fn = __get_tile_fn_ptr(tiling);
> +	void *map;
> +
> +	gem_set_domain(fd, buf->handle, I915_GEM_DOMAIN_CPU, 0);
> +
> +	map = __gem_mmap_offset__cpu(fd, buf->handle, 0, buf->size, PROT_READ);
> +	if (!map)
> +		map = gem_mmap__cpu(fd, buf->handle, 0, buf->size, PROT_READ);
> +
> +	for (int y = 0; y < height; y++) {
> +		for (int x = 0; x < width; x++) {
> +			uint32_t *ptr = fn(map, x, y, buf->stride, buf->bpp/8);
> +
> +			if (swizzle)
> +				ptr = from_user_pointer(swizzle_addr(ptr,
> +								     swizzle));
> +			linear[y * width + x] = *ptr;
> +		}
> +	}
> +
> +	munmap(map, buf->size);
> +}
> +
> +static void copy_x_to_linear(struct buf_ops *bops, struct intel_buf *buf,
> +			     uint32_t *linear)
> +{
> +	DEBUGFN();
> +	__copy_to_linear(bops->fd, buf, linear, I915_TILING_X, bops->swizzle_x);
> +}
> +
> +static void copy_y_to_linear(struct buf_ops *bops, struct intel_buf *buf,
> +			     uint32_t *linear)
> +{
> +	DEBUGFN();
> +	__copy_to_linear(bops->fd, buf, linear, I915_TILING_Y, bops->swizzle_y);
> +}
> +
> +static void copy_yf_to_linear(struct buf_ops *bops, struct intel_buf *buf,
> +			      uint32_t *linear)
> +{
> +	DEBUGFN();
> +	__copy_to_linear(bops->fd, buf, linear, I915_TILING_Yf, 0);
> +}
> +
> +static void copy_ys_to_linear(struct buf_ops *bops, struct intel_buf *buf,
> +			      uint32_t *linear)
> +{
> +	DEBUGFN();
> +	__copy_to_linear(bops->fd, buf, linear, I915_TILING_Ys, 0);
> +}
> +
> +static void copy_linear_to_gtt(struct buf_ops *bops, struct intel_buf *buf,
> +			       uint32_t *linear)
> +{
> +	void *map;
> +
> +	DEBUGFN();
> +
> +	gem_set_domain(bops->fd, buf->handle,
> +		       I915_GEM_DOMAIN_GTT, I915_GEM_DOMAIN_GTT);
> +
> +	map = gem_mmap__gtt(bops->fd, buf->handle,
> +			    buf->size, PROT_READ | PROT_WRITE);
> +
> +	memcpy(map, linear, buf->size);
> +
> +	munmap(map, buf->size);
> +}
> +
> +static void copy_gtt_to_linear(struct buf_ops *bops, struct intel_buf *buf,
> +			       uint32_t *linear)
> +{
> +	void *map;
> +
> +	DEBUGFN();
> +
> +	gem_set_domain(bops->fd, buf->handle,
> +		       I915_GEM_DOMAIN_GTT, 0);
> +
> +	map = gem_mmap__gtt(bops->fd, buf->handle,
> +			    buf->size, PROT_READ);
> +
> +	igt_memcpy_from_wc(linear, map, buf->size);
> +
> +	munmap(map, buf->size);
> +}
> +
> +static void copy_linear_to_wc(struct buf_ops *bops, struct intel_buf *buf,
> +			      uint32_t *linear)
> +{
> +	void *map;
> +
> +	DEBUGFN();
> +
> +	gem_set_domain(bops->fd, buf->handle,
> +		       I915_GEM_DOMAIN_GTT, I915_GEM_DOMAIN_GTT);
> +
> +	map = __gem_mmap_offset__wc(bops->fd, buf->handle, 0,
> +				    buf->size, PROT_READ | PROT_WRITE);
> +	if (!map)
> +		map = gem_mmap__wc(bops->fd, buf->handle, 0,
> +				   buf->size, PROT_READ | PROT_WRITE);
> +
> +	memcpy(map, linear, buf->size);
> +
> +	munmap(map, buf->size);
> +}
> +
> +static void copy_wc_to_linear(struct buf_ops *bops, struct intel_buf *buf,
> +			      uint32_t *linear)
> +{
> +	void *map;
> +
> +	DEBUGFN();
> +
> +	gem_set_domain(bops->fd, buf->handle,
> +		       I915_GEM_DOMAIN_GTT, 0);
> +
> +	map = __gem_mmap_offset__wc(bops->fd, buf->handle, 0,
> +				    buf->size, PROT_READ);
> +	if (!map)
> +		map = gem_mmap__wc(bops->fd, buf->handle, 0,
> +				   buf->size, PROT_READ);
> +
> +	igt_memcpy_from_wc(linear, map, buf->size);
> +
> +	munmap(map, buf->size);
> +}
> +
> +static void copy_linear_to_cpu(struct buf_ops *bops, struct intel_buf *buf,
> +			       uint32_t *linear)
> +{
> +	void *map;
> +
> +	DEBUGFN();
> +
> +	gem_set_domain(bops->fd, buf->handle,
> +		       I915_GEM_DOMAIN_CPU, I915_GEM_DOMAIN_CPU);
> +
> +	map = __gem_mmap_offset__cpu(bops->fd, buf->handle, 0,
> +				     buf->size, PROT_READ | PROT_WRITE);
> +	if (!map)
> +		map = gem_mmap__cpu(bops->fd, buf->handle, 0,
> +				    buf->size, PROT_READ | PROT_WRITE);
> +
> +	memcpy(map, linear, buf->size);
> +
> +	munmap(map, buf->size);
> +}
> +
> +static void copy_cpu_to_linear(struct buf_ops *bops, struct intel_buf *buf,
> +			       uint32_t *linear)
> +{
> +	void *map;
> +
> +	DEBUGFN();
> +
> +	gem_set_domain(bops->fd, buf->handle,
> +		       I915_GEM_DOMAIN_CPU, 0);
> +
> +	map = __gem_mmap_offset__cpu(bops->fd, buf->handle, 0,
> +				     buf->size, PROT_READ);
> +	if (!map)
> +		map = gem_mmap__cpu(bops->fd, buf->handle, 0,
> +				    buf->size, PROT_READ);
> +
> +	memcpy(linear, map, buf->size);
> +
> +	munmap(map, buf->size);
> +}
> +
> +void intel_buf_to_linear(struct buf_ops *bops, struct intel_buf *buf,
> +			 uint32_t *linear)
> +{
> +	igt_assert(bops);
> +
> +	switch (buf->tiling) {
> +	case I915_TILING_NONE:
> +		igt_assert(bops->to_linear);
> +		bops->to_linear(bops, buf, linear);
> +		break;
> +	case I915_TILING_X:
> +		igt_assert(bops->x_to_linear);
> +		bops->x_to_linear(bops, buf, linear);
> +		break;
> +	case I915_TILING_Y:
> +		igt_assert(bops->y_to_linear);
> +		bops->y_to_linear(bops, buf, linear);
> +		break;
> +	case I915_TILING_Yf:
> +		igt_assert(bops->yf_to_linear);
> +		bops->yf_to_linear(bops, buf, linear);
> +		break;
> +	case I915_TILING_Ys:
> +		igt_assert(bops->ys_to_linear);
> +		bops->ys_to_linear(bops, buf, linear);
> +		break;
> +	}
> +}
> +
> +void linear_to_intel_buf(struct buf_ops *bops, struct intel_buf *buf,
> +			 uint32_t *linear)
> +{
> +	igt_assert(bops);
> +
> +	switch (buf->tiling) {
> +	case I915_TILING_NONE:
> +		igt_assert(bops->linear_to);
> +		bops->linear_to(bops, buf, linear);
> +		break;
> +	case I915_TILING_X:
> +		igt_assert(bops->linear_to_x);
> +		bops->linear_to_x(bops, buf, linear);
> +		break;
> +	case I915_TILING_Y:
> +		igt_assert(bops->linear_to_y);
> +		bops->linear_to_y(bops, buf, linear);
> +		break;
> +	case I915_TILING_Yf:
> +		igt_assert(bops->linear_to_yf);
> +		bops->linear_to_yf(bops, buf, linear);
> +		break;
> +	case I915_TILING_Ys:
> +		igt_assert(bops->linear_to_ys);
> +		bops->linear_to_ys(bops, buf, linear);
> +		break;
> +	}
> +}
> +
> +static void __intel_buf_init(struct buf_ops *bops,
> +			     uint32_t handle,
> +			     struct intel_buf *buf,
> +			     int width, int height, int bpp,
> +			     uint32_t req_tiling, bool compression)
> +{
> +	uint32_t tiling = req_tiling;
> +	uint32_t size;
> +
> +	igt_assert(bops);
> +	igt_assert(buf);
> +	igt_assert(width > 0 && height > 0);
> +	igt_assert(bpp == 8 || bpp == 16 || bpp == 32);
> +
> +	memset(buf, 0, sizeof(*buf));
> +
> +	if (compression) {
> +		int aux_width, aux_height;
> +
> +		igt_require(bops->intel_gen >= 9);
> +		igt_assert(req_tiling == I915_TILING_Y ||
> +			   req_tiling == I915_TILING_Yf);
> +
> +		/*
> +		 * On GEN12+ we align the main surface to 4 * 4 main surface
> +		 * tiles, which is 64kB. These 16 tiles are mapped by 4 AUX
> +		 * CCS units, that is 4 * 64 bytes. These 4 CCS units are in
> +		 * turn mapped by one L1 AUX page table entry.
> +		 */
> +		if (bops->intel_gen >= 12)
> +			buf->stride = ALIGN(width * (bpp / 8), 128 * 4);
> +		else
> +			buf->stride = ALIGN(width * (bpp / 8), 128);
> +
> +		if (bops->intel_gen >= 12)
> +			height = ALIGN(height, 4 * 32);
> +
> +		buf->size = buf->stride * height;
> +		buf->tiling = tiling;
> +		buf->bpp = bpp;
> +		buf->compression = compression;
> +
> +		aux_width = intel_buf_aux_width(bops->intel_gen, buf);
> +		aux_height = intel_buf_aux_height(bops->intel_gen, buf);
> +
> +		buf->aux.offset = buf->stride * ALIGN(height, 32);
> +		buf->aux.stride = aux_width;
> +
> +		size = buf->aux.offset + aux_width * aux_height;
> +
> +	} else {
> +		buf->stride = ALIGN(width * (bpp / 8), 128);
> +		buf->size = buf->stride * height;
> +		buf->tiling = tiling;
> +		buf->bpp = bpp;
> +
> +		size = buf->stride * ALIGN(height, 32);
> +	}
> +
> +	if (handle)
> +		buf->handle = handle;
> +	else
> +		buf->handle = gem_create(bops->fd, size);
> +
> +	set_hw_tiled(bops, buf);
> +}
> +
> +/**
> + * intel_buf_init
> + * @bops: pointer to buf_ops
> + * @buf: pointer to intel_buf structure to be filled
> + * @width: surface width
> + * @height: surface height
> + * @bpp: bits-per-pixel (8 / 16 / 32)
> + * @tiling: surface tiling
> + * @compression: surface compression flag on / off
> + *
> + * Function creates new BO within intel_buf structure and fills all
> + * structure fields.
> + *
> + * Note. For X / Y if GPU supports fences HW tiling is configured.
> + */
> +void intel_buf_init(struct buf_ops *bops,
> +		    struct intel_buf *buf,
> +		    int width, int height, int bpp,
> +		    uint32_t tiling, bool compression)
> +{
> +	__intel_buf_init(bops, 0, buf, width, height, bpp, tiling,
> +			 compression);
> +}
> +
> +/**
> + * intel_buf_close
> + * @bops: pointer to buf_ops
> + * @buf: pointer to intel_buf structure
> + *
> + * Function closes gem BO inside intel_buf.
> + */
> +void intel_buf_close(struct buf_ops *bops, struct intel_buf *buf)
> +{
> +	igt_assert(bops);
> +	igt_assert(buf);
> +
> +	gem_close(bops->fd, buf->handle);
> +}
> +
> +/**
> + * intel_buf_init_using_handle
> + * @bops: pointer to buf_ops
> + * @handle: BO handle created by the caller
> + * @buf: pointer to intel_buf structure to be filled
> + * @width: surface width
> + * @height: surface height
> + * @bpp: bits-per-pixel (8 / 16 / 32)
> + * @tiling: surface tiling
> + * @compression: surface compression flag on / off
> + *
> + * Function configures BO handle within intel_buf structure passed by the caller
> + * (with all its metadata - width, height, ...). Useful if BO was created
> + * outside.
> + *
> + * Note: intel_buf_close() can be used to close the BO handle, but caller
> + * must be aware to not close the BO twice.
> + */
> +void intel_buf_init_using_handle(struct buf_ops *bops,
> +				 uint32_t handle,
> +				 struct intel_buf *buf,
> +				 int width, int height, int bpp,
> +				 uint32_t req_tiling, bool compression)
> +{
> +	__intel_buf_init(bops, handle, buf, width, height, bpp, req_tiling,
> +			 compression);
> +}
> +
> +#define DEFAULT_BUFOPS(__gen_start, __gen_end) \
> +	.gen_start          = __gen_start, \
> +	.gen_end            = __gen_end, \
> +	.supported_hw_tiles = TILE_X | TILE_Y, \
> +	.linear_to_x        = copy_linear_to_gtt, \
> +	.linear_to_y        = copy_linear_to_gtt, \
> +	.linear_to_yf       = copy_linear_to_yf, \
> +	.linear_to_ys       = copy_linear_to_ys, \
> +	.x_to_linear        = copy_gtt_to_linear, \
> +	.y_to_linear        = copy_gtt_to_linear, \
> +	.yf_to_linear       = copy_yf_to_linear, \
> +	.ys_to_linear       = copy_ys_to_linear
> +
> +struct buf_ops buf_ops_arr[] = {
> +	/* Generations 2 - 8 */
> +{
> +	DEFAULT_BUFOPS(2, 8),
> +	.supported_tiles    = TILE_NONE | TILE_X | TILE_Y,
> +	.linear_to          = copy_linear_to_cpu,
> +	.to_linear          = copy_cpu_to_linear,
> +},
> +/* Generations 9 - 11 */
> +{
> +	DEFAULT_BUFOPS(9, 11),
> +	.supported_tiles    = TILE_NONE | TILE_X | TILE_Y | TILE_Yf,
> +	.linear_to          = copy_linear_to_cpu,
> +	.to_linear          = copy_cpu_to_linear,
> +},
> +/* Generation 12 */
> +{
> +	DEFAULT_BUFOPS(12, 12),
> +	.supported_tiles   = TILE_NONE | TILE_X | TILE_Y | TILE_Yf | TILE_Ys,
> +	.linear_to         = copy_linear_to_wc,
> +	.to_linear         = copy_wc_to_linear,
> +},
> +};
> +
> +static bool probe_hw_tiling(struct buf_ops *bops, uint32_t tiling)
> +{
> +	uint64_t size = 256 * 256;
> +	uint32_t handle, buf_tiling, buf_swizzle;
> +	uint32_t stride;
> +	int ret;
> +	bool is_set = false;
> +
> +	if (tiling == I915_TILING_X)
> +		stride = 512;
> +	else if (tiling == I915_TILING_Y)
> +		stride = 128;
> +	else
> +		return false;
> +
> +	handle = gem_create(bops->fd, size);
> +
> +	/* Single shot, if no fences are available we fail immediately */
> +	ret = __set_tiling(bops->fd, handle, tiling, stride);
> +	if (ret)
> +		goto end;
> +
> +	is_set = __get_tiling(bops->fd, handle, &buf_tiling, &buf_swizzle);
> +	if (is_set) {
> +		if (tiling == I915_TILING_X)
> +			bops->swizzle_x = buf_swizzle;
> +		else if (tiling == I915_TILING_Y)
> +			bops->swizzle_y = buf_swizzle;
> +	}
> +end:
> +	gem_close(bops->fd, handle);
> +
> +	return is_set;
> +}
> +
> +static void *alloc_aligned(uint64_t size)
> +{
> +	void *p;
> +
> +	igt_assert_eq(posix_memalign(&p, 16, size), 0);
> +
> +	return p;
> +}
> +
> +/*
> + * Simple idempotency test between HW -> SW and SW -> HW BO.
> + */
> +static void idempotency_selftest(struct buf_ops *bops, uint32_t tiling)
> +{
> +	struct intel_buf buf;
> +	uint8_t *linear_in, *linear_out, *map;
> +	int i;
> +	const int width = 512;
> +	const int height = 512;
> +	const int bpp = 32;
> +	const int size = width * height * bpp / 8;
> +	bool software_tiling = false;
> +
> +	if (!is_hw_tiling_supported(bops, tiling))
> +		return;
> +
> +	linear_in = alloc_aligned(size);
> +	linear_out = alloc_aligned(size);
> +
> +	for (i = 0; i < size; i++)
> +		linear_in[i] = i % 253; /* prime chosen intentionally */
> +
> +	do {
> +		igt_debug("Checking idempotency, SW: %s, HW: %s, tiling: %s\n",
> +			  bool_str(software_tiling),
> +			  bool_str(!software_tiling),
> +			  tiling_str(tiling));
> +		intel_buf_init(bops, &buf, width, height, bpp, tiling, false);
> +		buf_ops_set_software_tiling(bops, tiling, software_tiling);
> +
> +		linear_to_intel_buf(bops, &buf, (uint32_t *) linear_in);
> +		map = __gem_mmap_offset__cpu(bops->fd, buf.handle, 0,
> +					   buf.size, PROT_READ);
> +		if (!map)
> +			map = gem_mmap__cpu(bops->fd, buf.handle, 0,
> +					    buf.size, PROT_READ);
> +		igt_assert(map);
> +		igt_assert(memcmp(linear_in, map, size));
> +
> +		buf_ops_set_software_tiling(bops, tiling, !software_tiling);
> +		intel_buf_to_linear(bops, &buf, (uint32_t *) linear_out);
> +		igt_assert(!memcmp(linear_in, linear_out, size));
> +
> +		munmap(map, size);
> +		intel_buf_close(bops, &buf);
> +
> +		software_tiling = !software_tiling;
> +	} while (software_tiling);
> +
> +	igt_debug("Idempotency for %s tiling OK\n", tiling_str(tiling));
> +	buf_ops_set_software_tiling(bops, tiling, false);
> +}
> +
> +/**
> + * buf_ops_create
> + * @fd: device filedescriptor
> + *
> + * Create buf_ops structure depending on fd-device capabilities.
> + *
> + * Returns: opaque pointer to buf_ops.
> + *
> + */
> +struct buf_ops *buf_ops_create(int fd)
> +{
> +	struct buf_ops *bops = calloc(1, sizeof(*bops));
> +	uint32_t devid;
> +	int generation;
> +
> +	igt_assert(bops);
> +
> +	devid = intel_get_drm_devid(fd);
> +	generation = intel_gen(devid);
> +
> +	/* Predefined settings */
> +	for (int i = 0; i < ARRAY_SIZE(buf_ops_arr); i++) {
> +		if (generation >= buf_ops_arr[i].gen_start &&
> +				generation <= buf_ops_arr[i].gen_end) {
> +			memcpy(bops, &buf_ops_arr[i], sizeof(*bops));
> +			bops->fd = fd;
> +			bops->intel_gen = generation;
> +			igt_debug("generation: %d, supported tiles: 0x%02x\n",
> +				  generation, bops->supported_tiles);
> +			break;
> +		}
> +	}
> +
> +	igt_assert(bops->intel_gen);
> +
> +	/*
> +	 * Warning!
> +	 *
> +	 * Gen2 software tiling/detiling is not supported! (yet).
> +	 *
> +	 * If you are brave hero with an access to Gen2 you can save the world.
> +	 * Until then we're doomed to use only hardware (de)tiling.
> +	 *
> +	 * Ok, you have been warned.
> +	 */
> +	if (bops->intel_gen == 2) {
> +		igt_warn("Gen2 detected. HW (de)tiling support only.");
> +		return bops;
> +	}
> +
> +	/* Let's probe X and Y hw tiling support */
> +	if (is_hw_tiling_supported(bops, I915_TILING_X)) {
> +		bool supported = probe_hw_tiling(bops, I915_TILING_X);
> +
> +		igt_debug("X fence support: %s\n", bool_str(supported));
> +		if (!supported) {
> +			bops->supported_hw_tiles &= ~TILE_X;
> +			bops->linear_to_x = copy_linear_to_x;
> +			bops->x_to_linear = copy_x_to_linear;
> +		}
> +	}
> +
> +	if (is_hw_tiling_supported(bops, I915_TILING_Y)) {
> +		bool supported = probe_hw_tiling(bops, I915_TILING_Y);
> +
> +		igt_debug("Y fence support: %s\n", bool_str(supported));
> +		if (!supported) {
> +			bops->supported_hw_tiles &= ~TILE_Y;
> +			bops->linear_to_y = copy_linear_to_y;
> +			bops->y_to_linear = copy_y_to_linear;
> +		}
> +	}
> +
> +	/* Disable other tiling format functions if not supported */
> +	if (!is_tiling_supported(bops, I915_TILING_Yf)) {
> +		igt_debug("Yf format not supported\n");
> +		bops->linear_to_yf = NULL;
> +		bops->yf_to_linear = NULL;
> +	}
> +
> +	if (!is_tiling_supported(bops, I915_TILING_Ys)) {
> +		igt_debug("Ys format not supported\n");
> +		bops->linear_to_ys = NULL;
> +		bops->ys_to_linear = NULL;
> +	}
> +
> +	idempotency_selftest(bops, I915_TILING_X);
> +	idempotency_selftest(bops, I915_TILING_Y);
> +
> +	return bops;
> +}
> +
> +/**
> + * buf_ops_destroy
> + * @bops: pointer to buf_ops
> + *
> + * Function frees buf_ops structure.
> + */
> +void buf_ops_destroy(struct buf_ops *bops)
> +{
> +	igt_assert(bops);
> +	free(bops);
> +}
> +
> +/**
> + * buf_ops_set_software_tiling
> + * @bops: pointer to buf_ops
> + * @tiling: surface tiling
> + * @use_software_tiling: if true use software copying methods, otherwise
> + * use hardware (via gtt)
> + *
> + * Function allows switch X / Y surfaces to software / hardware copying methods
> + * which honors tiling and swizzling.
> + *
> + * Returns:
> + * false - switch wasn't possible.
> + * true - switch to software / hardware method succeed.
> + */
> +bool buf_ops_set_software_tiling(struct buf_ops *bops,
> +				 uint32_t tiling,
> +				 bool use_software_tiling)
> +{
> +	bool was_changed = true;
> +
> +	igt_assert(bops);
> +
> +	/* Until appropriate code is added we don't support SW tiling on Gen2 */
> +	if (bops->intel_gen == 2) {
> +		igt_warn("Change to software tiling on Gen2 is not supported!");
> +		return false;
> +	}
> +
> +	switch (tiling) {
> +	case I915_TILING_X:
> +		if (use_software_tiling) {
> +			igt_debug("-> change X to SW\n");
> +			bops->linear_to_x = copy_linear_to_x;
> +			bops->x_to_linear = copy_x_to_linear;
> +		} else {
> +			if (is_hw_tiling_supported(bops, I915_TILING_X)) {
> +				igt_debug("-> change X to HW\n");
> +				bops->linear_to_x = copy_linear_to_gtt;
> +				bops->x_to_linear = copy_gtt_to_linear;
> +			} else {
> +				igt_debug("-> X cannot be changed to HW\n");
> +				was_changed = false;
> +			}
> +		}
> +		break;
> +
> +	case I915_TILING_Y:
> +		if (use_software_tiling) {
> +			igt_debug("-> change Y to SW\n");
> +			bops->linear_to_y = copy_linear_to_y;
> +			bops->y_to_linear = copy_y_to_linear;
> +		} else {
> +			if (is_hw_tiling_supported(bops, I915_TILING_Y)) {
> +				igt_debug("-> change Y to HW\n");
> +				bops->linear_to_y = copy_linear_to_gtt;
> +				bops->y_to_linear = copy_gtt_to_linear;
> +			} else {
> +				igt_debug("-> Y cannot be changed to HW\n");
> +				was_changed = false;
> +			}
> +		}
> +		break;
> +
> +	default:
> +		igt_warn("Invalid tiling: %d\n", tiling);
> +		was_changed = false;
> +	}
> +
> +	return was_changed;
> +}
> +
> +/**
> + * buf_ops_has_hw_fence
> + * @bops: pointer to buf_ops
> + * @tiling: surface tiling
> + *
> + * Function checks if surface with tiling has HW fences which can be used
> + * to copy it via gtt.
> + *
> + * Returns:
> + * false - fence for tiling is not supported.
> + * true - fence for tiling is supported.
> + */
> +bool buf_ops_has_hw_fence(struct buf_ops *bops, uint32_t tiling)
> +{
> +	uint32_t tile_mask = TILE_DEF(tiling);
> +
> +	igt_assert(bops);
> +
> +	if (tile_mask & bops->supported_hw_tiles)
> +		return true;
> +
> +	return false;
> +}
> +
> +/**
> + * buf_ops_has_tiling_support
> + * @bops: pointer to buf_ops
> + * @tiling: surface tiling
> + *
> + * Function checks capabilities to handle surfaces with tiling in GPU.
> + *
> + * Returns:
> + * false - GPU does not support tiling.
> + * true - GPU supports tiling.
> + */
> +bool buf_ops_has_tiling_support(struct buf_ops *bops, uint32_t tiling)
> +{
> +	uint32_t tile_mask = TILE_DEF(tiling);
> +
> +	igt_assert(bops);
> +
> +	if (tile_mask & bops->supported_tiles)
> +		return true;
> +
> +	return false;
> +}
> diff --git a/lib/intel_bufops.h b/lib/intel_bufops.h
> new file mode 100644
> index 00000000..c1d1e92d
> --- /dev/null
> +++ b/lib/intel_bufops.h
> @@ -0,0 +1,84 @@
> +#ifndef __INTEL_BUFOPS_H__
> +#define __INTEL_BUFOPS_H__
> +
> +#include <stdint.h>
> +
> +struct buf_ops;
> +
> +struct intel_buf {
> +	uint32_t handle;
> +	uint32_t stride;
> +	uint32_t tiling;
> +	uint32_t bpp;
> +	uint32_t size;
> +	uint32_t compression;
> +	struct {
> +		uint32_t offset;
> +		uint32_t stride;
> +	} aux;
> +};
> +
> +inline unsigned int intel_buf_width(const struct intel_buf *buf)
> +{
> +	return buf->stride/(buf->bpp / 8);
> +}
> +
> +inline unsigned int intel_buf_height(const struct intel_buf *buf)
> +{
> +	return buf->size/buf->stride;
> +}
> +
> +inline unsigned int intel_buf_aux_width(int gen, const struct intel_buf *buf)
> +{
> +	/*
> +	 * GEN12+: The AUX CCS unit size is 64 bytes mapping 4 main surface
> +	 * tiles. Thus the width of the CCS unit is 4*32=128 pixels on the
> +	 * main surface.
> +	 */
> +	if (gen >= 12)
> +		return DIV_ROUND_UP(intel_buf_width(buf), 128) * 64;
> +
> +	return DIV_ROUND_UP(intel_buf_width(buf), 1024) * 128;
> +}
> +
> +inline unsigned int intel_buf_aux_height(int gen, const struct intel_buf *buf)
> +{
> +	/*
> +	 * GEN12+: The AUX CCS unit size is 64 bytes mapping 4 main surface
> +	 * tiles. Thus the height of the CCS unit is 32 pixel rows on the main
> +	 * surface.
> +	 */
> +	if (gen >= 12)
> +		return DIV_ROUND_UP(intel_buf_height(buf), 32);
> +
> +	return DIV_ROUND_UP(intel_buf_height(buf), 512) * 32;
> +}
> +
> +struct buf_ops *buf_ops_create(int fd);
> +void buf_ops_destroy(struct buf_ops *bops);
> +
> +bool buf_ops_set_software_tiling(struct buf_ops *bops,
> +				 uint32_t tiling,
> +				 bool use_software_tiling);
> +
> +void intel_buf_to_linear(struct buf_ops *bops, struct intel_buf *buf,
> +			 uint32_t *linear);
> +
> +void linear_to_intel_buf(struct buf_ops *bops, struct intel_buf *buf,
> +			 uint32_t *linear);
> +
> +bool buf_ops_has_hw_fence(struct buf_ops *bops, uint32_t tiling);
> +bool buf_ops_has_tiling_support(struct buf_ops *bops, uint32_t tiling);
> +
> +void intel_buf_init(struct buf_ops *bops, struct intel_buf *buf,
> +		    int width, int height, int bpp,
> +		    uint32_t tiling, bool compression);
> +void intel_buf_close(struct buf_ops *bops, struct intel_buf *buf);
> +
> +void intel_buf_init_using_handle(struct buf_ops *bops,
> +				 uint32_t handle,
> +				 struct intel_buf *buf,
> +				 int width, int height, int bpp,
> +				 uint32_t req_tiling, bool compression);
Quite a code! I didn't noticed anything to change.
LGTM: Ack-by: Katarzyna Dec <katarzyna.dec at intel.com>
Kasia :)

> +
> +#endif
> diff --git a/lib/meson.build b/lib/meson.build
> index 57eb7d93..cf3e35da 100644
> --- a/lib/meson.build
> +++ b/lib/meson.build
> @@ -27,6 +27,7 @@ lib_sources = [
>  	'igt_x86.c',
>  	'instdone.c',
>  	'intel_batchbuffer.c',
> +	'intel_bufops.c',
>  	'intel_chipset.c',
>  	'intel_device_info.c',
>  	'intel_os.c',
> -- 
> 2.23.0
> 


More information about the igt-dev mailing list