[Intel-gfx] [PATCH i-g-t 1/2] lib/intel_batchbuffer: Add blitter copy using XY_SRC_COPY_BLT
Chris Wilson
chris at chris-wilson.co.uk
Thu Jan 9 14:43:55 UTC 2020
Quoting Janusz Krzysztofik (2020-01-09 14:01:24)
> From: Vanshidhar Konda <vanshidhar.r.konda at intel.com>
>
> Add a method that uses the XY_SRC_COPY_BLT instruction for copying
> buffers using the blitter engine.
>
> v2: Use uint32_t for parameters; fix stride for Gen2/3
> v3: Taken over from Vanshi by Janusz,
> - skip upper bits of src and dst addresses on Gen < 8
>
> Signed-off-by: Vanshidhar Konda <vanshidhar.r.konda at intel.com>
> Signed-off-by: Janusz Krzysztofik <janusz.krzysztofik at linux.intel.com>
> ---
> lib/intel_batchbuffer.c | 185 ++++++++++++++++++++++++++++++++++++++++
> lib/intel_batchbuffer.h | 21 +++++
> 2 files changed, 206 insertions(+)
>
> diff --git a/lib/intel_batchbuffer.c b/lib/intel_batchbuffer.c
> index 07de5cbb..a56cd5a7 100644
> --- a/lib/intel_batchbuffer.c
> +++ b/lib/intel_batchbuffer.c
> @@ -45,6 +45,12 @@
>
> #include <i915_drm.h>
>
> +#define MI_FLUSH_DW (0x26 << 23)
> +
> +#define BCS_SWCTRL 0x22200
> +#define BCS_SRC_Y (1 << 0)
> +#define BCS_DST_Y (1 << 1)
> +
> /**
> * SECTION:intel_batchbuffer
> * @short_description: Batchbuffer and blitter support
> @@ -660,6 +666,185 @@ static void exec_blit(int fd,
> gem_execbuf(fd, &exec);
> }
>
> +static uint32_t src_copy_dword0(uint32_t src_tiling, uint32_t dst_tiling,
> + uint32_t bpp, uint32_t device_gen)
> +{
> + uint32_t dword0 = 0;
> +
> + dword0 |= XY_SRC_COPY_BLT_CMD;
> + if (bpp == 32)
> + dword0 |= XY_SRC_COPY_BLT_WRITE_RGB |
> + XY_SRC_COPY_BLT_WRITE_ALPHA;
> +
> + if (device_gen >= 4 && src_tiling)
> + dword0 |= XY_SRC_COPY_BLT_SRC_TILED;
> + if (device_gen >= 4 && dst_tiling)
> + dword0 |= XY_SRC_COPY_BLT_DST_TILED;
> +
> + return dword0;
> +}
> +
> +static uint32_t src_copy_dword1(uint32_t dst_pitch, uint32_t bpp)
> +{
> + uint32_t dword1 = 0;
> +
> + switch (bpp) {
> + case 8:
> + break;
> + case 16:
> + dword1 |= (1 << 24); /* Only support 565 color */
> + break;
> + case 32:
> + dword1 |= (3 << 24);
> + break;
> + default:
> + igt_assert(0);
> + }
> +
> + dword1 |= 0xcc << 16;
> + dword1 |= dst_pitch;
> +
> + return dword1;
> +}
> +/**
> + * igt_blitter_src_copy__raw:
> + * @fd: file descriptor of the i915 driver
> + * @src_handle: GEM handle of the source buffer
> + * @src_delta: offset into the source GEM bo, in bytes
> + * @src_stride: Stride (in bytes) of the source buffer
> + * @src_tiling: Tiling mode of the source buffer
> + * @src_x: X coordinate of the source region to copy
> + * @src_y: Y coordinate of the source region to copy
> + * @width: Width of the region to copy
> + * @height: Height of the region to copy
> + * @bpp: source and destination bits per pixel
> + * @dst_handle: GEM handle of the destination buffer
> + * @dst_delta: offset into the destination GEM bo, in bytes
> + * @dst_stride: Stride (in bytes) of the destination buffer
> + * @dst_tiling: Tiling mode of the destination buffer
> + * @dst_x: X coordinate of destination
> + * @dst_y: Y coordinate of destination
> + *
> + */
> +void igt_blitter_src_copy__raw(int fd,
Just call it igt_blitter_copy(). It should be the shorter, more often
useful variant. _fast is the special case version.
These functions should not be intel_batchbuffer.c. Either you do intend
them to be vendor neutral, and they could indeed become so, or they
should not be called igt_.
> + /* src */
> + uint32_t src_handle,
> + uint32_t src_delta,
> + uint32_t src_stride,
> + uint32_t src_tiling,
> + uint32_t src_x, uint32_t src_y,
> +
> + /* size */
> + uint32_t width, uint32_t height,
> +
> + /* bpp */
> + uint32_t bpp,
> +
> + /* dst */
> + uint32_t dst_handle,
> + uint32_t dst_delta,
> + uint32_t dst_stride,
> + uint32_t dst_tiling,
> + uint32_t dst_x, uint32_t dst_y)
> +{
> + uint32_t batch[32];
> + struct drm_i915_gem_exec_object2 objs[3];
> + struct drm_i915_gem_relocation_entry relocs[2];
> + uint32_t batch_handle;
> + uint32_t src_pitch, dst_pitch;
> + uint32_t dst_reloc_offset, src_reloc_offset;
> + int i = 0;
> + uint32_t gen = intel_gen(intel_get_drm_devid(fd));
> + const bool has_64b_reloc = gen >= 8;
> +
> + memset(batch, 0, sizeof(batch));
> +
> + igt_assert((src_tiling == I915_TILING_NONE) ||
> + (src_tiling == I915_TILING_X) ||
> + (src_tiling == I915_TILING_Y));
> + igt_assert((dst_tiling == I915_TILING_NONE) ||
> + (dst_tiling == I915_TILING_X) ||
> + (dst_tiling == I915_TILING_Y));
> +
> + src_pitch = (gen >= 4 && src_tiling) ? src_stride / 4 : src_stride;
> + dst_pitch = (gen >= 4 && dst_tiling) ? dst_stride / 4 : dst_stride;
> +
> + CHECK_RANGE(src_x); CHECK_RANGE(src_y);
> + CHECK_RANGE(dst_x); CHECK_RANGE(dst_y);
> + CHECK_RANGE(width); CHECK_RANGE(height);
> + CHECK_RANGE(src_x + width); CHECK_RANGE(src_y + height);
> + CHECK_RANGE(dst_x + width); CHECK_RANGE(dst_y + height);
> + CHECK_RANGE(src_pitch); CHECK_RANGE(dst_pitch);
> +
> + if ((src_tiling | dst_tiling) >= I915_TILING_Y) {
> + unsigned int mask;
> +
> + batch[i++] = MI_LOAD_REGISTER_IMM;
> + batch[i++] = BCS_SWCTRL;
> +
> + mask = (BCS_SRC_Y | BCS_DST_Y) << 16;
> + if (src_tiling == I915_TILING_Y)
> + mask |= BCS_SRC_Y;
> + if (dst_tiling == I915_TILING_Y)
> + mask |= BCS_DST_Y;
> + batch[i++] = mask;
> + }
> +
> + batch[i] = src_copy_dword0(src_tiling, dst_tiling, bpp, gen);
> + batch[i++] |= 6 + 2 * has_64b_reloc;
> + batch[i++] = src_copy_dword1(dst_pitch, bpp);
> + batch[i++] = (dst_y << 16) | dst_x; /* dst x1,y1 */
> + batch[i++] = ((dst_y + height) << 16) | (dst_x + width); /* dst x2,y2 */
> + dst_reloc_offset = i;
> + batch[i++] = dst_delta; /* dst address lower bits */
> + if (has_64b_reloc)
> + batch[i++] = 0; /* dst address upper bits */
It is better do the relocations inline and use the computed address from
the relocation routine so that the value in the batch is known to be
equal to value passed to the kernel.
> + batch[i++] = (src_y << 16) | src_x; /* src x1,y1 */
> + batch[i++] = src_pitch;
> + src_reloc_offset = i;
> + batch[i++] = src_delta; /* src address lower bits */
> + if (has_64b_reloc)
> + batch[i++] = 0; /* src address upper bits */
> +
> + if ((src_tiling | dst_tiling) >= I915_TILING_Y) {
> + igt_assert(gen >= 6);
> + batch[i++] = MI_FLUSH_DW | 2;
> + batch[i++] = 0;
> + batch[i++] = 0;
> + batch[i++] = 0;
> +
> + batch[i++] = MI_LOAD_REGISTER_IMM;
> + batch[i++] = BCS_SWCTRL;
> + batch[i++] = (BCS_SRC_Y | BCS_DST_Y) << 16;
> + }
> +
> + batch[i++] = MI_BATCH_BUFFER_END;
> + batch[i++] = MI_NOOP;
> +
> + igt_assert(i <= ARRAY_SIZE(batch));
> +
> + batch_handle = gem_create(fd, 4096);
> + gem_write(fd, batch_handle, 0, batch, sizeof(batch));
> +
> + fill_relocation(&relocs[0], dst_handle, dst_delta, dst_reloc_offset,
> + I915_GEM_DOMAIN_RENDER, I915_GEM_DOMAIN_RENDER);
> + fill_relocation(&relocs[1], src_handle, src_delta, src_reloc_offset,
> + I915_GEM_DOMAIN_RENDER, 0);
> +
> + fill_object(&objs[0], dst_handle, NULL, 0);
> + fill_object(&objs[1], src_handle, NULL, 0);
> + fill_object(&objs[2], batch_handle, relocs, 2);
> +
> + if (dst_tiling)
> + objs[0].flags |= EXEC_OBJECT_NEEDS_FENCE;
obj[0].flags |= EXEC_OBJECT_WRITE;
and then you can would be eligible to use execbuf.flags |= I915_EXEC_NORELOC;
But given you have no history, for it to work you could preseed the
obj.offset to sensible values rather than have all 3 three in the same
location. And you could fix the code to I915_EXEC_LUT_HANDLE.
> + if (src_tiling)
> + objs[1].flags |= EXEC_OBJECT_NEEDS_FENCE;
> +
> + exec_blit(fd, objs, 3, ARRAY_SIZE(batch));
exec_blit() is broken as it does not apply to <gen6.
> +
> + gem_close(fd, batch_handle);
> +}
More information about the Intel-gfx
mailing list