[igt-dev] [PATCH i-g-t 1/3] lib: Move common gpgpu/media functions to gpu_fill library
Daniele Ceraolo Spurio
daniele.ceraolospurio at intel.com
Wed Mar 28 18:50:57 UTC 2018
On 28/03/18 06:32, Szwichtenberg, Radoslaw wrote:
> On Tue, 2018-03-27 at 16:52 -0700, Daniele Ceraolo Spurio wrote:
>>
>> On 27/03/18 06:36, Katarzyna Dec wrote:
>>> Gpgpu_fill and media_fill libraries are very similar and many
>>> functions can be shared. I have created library gpu_fill with
>>> all functions needed for implementing gpgpu_fill and media_fill
>>> tests for all Gens. Duplicates, e.g. where only name was changed,
>>> were removed. The earliest common function remained.
>>>
>>> Signed-off-by: Katarzyna Dec <katarzyna.dec at intel.com>
>>> Cc: Lukasz Kalamarz <lukasz.kalamarz at intel.com>
>>> Cc: Daniele Ceraolo Spurio <daniele.ceraolospurio at intel.com>
>>> ---
>>> lib/Makefile.sources | 2 +
>>> lib/gpgpu_fill.c | 567 +------------------------------------
>>> lib/gpu_fill.c | 729
>>> ++++++++++++++++++++++++++++++++++++++++++++++++
>>> lib/gpu_fill.h | 141 ++++++++++
>>> lib/media_fill_gen7.c | 271 +-----------------
>>> lib/media_fill_gen8.c | 297 +-------------------
>>> lib/media_fill_gen8lp.c | 289 +------------------
>>> lib/media_fill_gen9.c | 304 +-------------------
>>> lib/meson.build | 1 +
>>> 9 files changed, 890 insertions(+), 1711 deletions(-)
>>> create mode 100644 lib/gpu_fill.c
>>> create mode 100644 lib/gpu_fill.h
>>>
>>
>> This patch has a very big diff and it isn't a 1:1 move of code so I
>> suggest to break it up a little to make it easier to review.
> Looks like almost 1:1 move to me. Differences in this code is one function
> unified cause its body is exactly the same (gen7_fill_kernel) and small
> modification in gen8_emit_state_base_address.
>
> An approach taken here looks like phased one: moving everything to one place
> (immediately removing some duplication for exactly same functions) and as second
> phase reworking similar functions that have very little differences (as you
> suggest below). Indeed this patch is quite big though it doesn't have too many
> functional changes so should be quite easy to review. For reworking functions
> that are very similar and are spread across many files (like fill_surface_state)
My worry here was with small differences, like the one you've pointed
out in gen8_emit_state_base_address, getting lost in the noise. Also,
why move (and then even update in patch 2) some functions we already
know we can get rid of, like gen8lp_emit_media_objects? makes more sense
getting rid of it first IMO. Anyway if the maintainers are happy with 1
big patch I won't oppose it, but at least the batch_* functions should
be moved with (and in some cases merged with) all the other
intel_batchbuffer related functions and not go in the new gpu_fill file.
Daniele
> I think this way will allow easier review for remaining changes. I think that
> those changes should be part of this series though.
>
> Radek
>> You could
>> start with moving the batch_* functions, which btw probably should be
>> moved to intel_batchbuffer.<h/c>. There are copies of those batch_*
>> functions in other files as well (e.g rendercopy), so once you move them
>> it'd be nice to move them from everywhere. Note that some of these
>> functions seem to be similar to existing ones (e.g. batch_copy and
>> intel_batchbuffer_data), so we can probably remove extra duplication by
>> merging some of these.
>>
>> Also, I think we could remove some duplication before doing the move,
>> which could make things more modular and reduce the diff of the move
>> itself.
>> For example, the only difference between media_fill_gen8.c and
>> media_fill_gen8lp.c is that in the !lp case we emit an extra
>> GEN8_MEDIA_STATE_FLUSH. Instead of duplicating the whole file, we can
>> just add a !IS_CHERRYVIEW(batch->devid) check before emitting that
>> instruction and eliminate gen8lp_media_fillfunc and relative file entirely.
>> Similarly, the diff between the gen8 and gen9 file is only in the
>> gen9_emit_state_base_address and gen9_media_fillfunc functions. The
>> first has a very small difference compared to the gen8 function so we
>> can probably merge the 2 using a IS_GEN() check to apply the diff, while
>> the latter we can just copy in the gen8 file and thus get rid of the
>> gen9 file as well.
>>
>> Daniele
>>
>>> diff --git a/lib/Makefile.sources b/lib/Makefile.sources
>>> index 3d37ef1d..690a1d35 100644
>>> --- a/lib/Makefile.sources
>>> +++ b/lib/Makefile.sources
>>> @@ -55,6 +55,8 @@ lib_source_list = \
>>> intel_reg.h \
>>> ioctl_wrappers.c \
>>> ioctl_wrappers.h \
>>> + gpu_fill.h \
>>> + gpu_fill.c \
>>> media_fill.h \
>>> media_fill_gen7.c \
>>> media_fill_gen8.c \
>>> diff --git a/lib/gpgpu_fill.c b/lib/gpgpu_fill.c
>>> index 4d98643d..68cbac5e 100644
>>> --- a/lib/gpgpu_fill.c
>>> +++ b/lib/gpgpu_fill.c
>>> @@ -34,6 +34,7 @@
>>> #include "gen7_media.h"
>>> #include "gen8_media.h"
>>> #include "gpgpu_fill.h"
>>> +#include "gpu_fill.h"
>>>
>>> /* shaders/gpgpu/gpgpu_fill.gxa */
>>> static const uint32_t gen7_gpgpu_kernel[][4] = {
>>> @@ -75,572 +76,6 @@ static const uint32_t gen9_gpgpu_kernel[][4] = {
>>> { 0x07800031, 0x20000a40, 0x06000e00, 0x82000010 },
>>> };
>>>
>>> -static uint32_t
>>> -batch_used(struct intel_batchbuffer *batch)
>>> -{
>>> - return batch->ptr - batch->buffer;
>>> -}
>>> -
>>> -static uint32_t
>>> -batch_align(struct intel_batchbuffer *batch, uint32_t align)
>>> -{
>>> - uint32_t offset = batch_used(batch);
>>> - offset = ALIGN(offset, align);
>>> - batch->ptr = batch->buffer + offset;
>>> - return offset;
>>> -}
>>> -
>>> -static void *
>>> -batch_alloc(struct intel_batchbuffer *batch, uint32_t size, uint32_t align)
>>> -{
>>> - uint32_t offset = batch_align(batch, align);
>>> - batch->ptr += size;
>>> - return memset(batch->buffer + offset, 0, size);
>>> -}
>>> -
>>> -static uint32_t
>>> -batch_offset(struct intel_batchbuffer *batch, void *ptr)
>>> -{
>>> - return (uint8_t *)ptr - batch->buffer;
>>> -}
>>> -
>>> -static uint32_t
>>> -batch_copy(struct intel_batchbuffer *batch, const void *ptr, uint32_t size,
>>> - uint32_t align)
>>> -{
>>> - return batch_offset(batch, memcpy(batch_alloc(batch, size, align),
>>> ptr, size));
>>> -}
>>> -
>>> -static void
>>> -gen7_render_flush(struct intel_batchbuffer *batch, uint32_t batch_end)
>>> -{
>>> - int ret;
>>> -
>>> - ret = drm_intel_bo_subdata(batch->bo, 0, 4096, batch->buffer);
>>> - if (ret == 0)
>>> - ret = drm_intel_bo_mrb_exec(batch->bo, batch_end,
>>> - NULL, 0, 0, 0);
>>> - igt_assert(ret == 0);
>>> -}
>>> -
>>> -static uint32_t
>>> -gen7_fill_curbe_buffer_data(struct intel_batchbuffer *batch, uint8_t color)
>>> -{
>>> - uint8_t *curbe_buffer;
>>> - uint32_t offset;
>>> -
>>> - curbe_buffer = batch_alloc(batch, sizeof(uint32_t) * 8, 64);
>>> - offset = batch_offset(batch, curbe_buffer);
>>> - *curbe_buffer = color;
>>> -
>>> - return offset;
>>> -}
>>> -
>>> -static uint32_t
>>> -gen7_fill_surface_state(struct intel_batchbuffer *batch,
>>> - struct igt_buf *buf,
>>> - uint32_t format,
>>> - int is_dst)
>>> -{
>>> - struct gen7_surface_state *ss;
>>> - uint32_t write_domain, read_domain, offset;
>>> - int ret;
>>> -
>>> - if (is_dst) {
>>> - write_domain = read_domain = I915_GEM_DOMAIN_RENDER;
>>> - } else {
>>> - write_domain = 0;
>>> - read_domain = I915_GEM_DOMAIN_SAMPLER;
>>> - }
>>> -
>>> - ss = batch_alloc(batch, sizeof(*ss), 64);
>>> - offset = batch_offset(batch, ss);
>>> -
>>> - ss->ss0.surface_type = GEN7_SURFACE_2D;
>>> - ss->ss0.surface_format = format;
>>> - ss->ss0.render_cache_read_write = 1;
>>> -
>>> - if (buf->tiling == I915_TILING_X)
>>> - ss->ss0.tiled_mode = 2;
>>> - else if (buf->tiling == I915_TILING_Y)
>>> - ss->ss0.tiled_mode = 3;
>>> -
>>> - ss->ss1.base_addr = buf->bo->offset;
>>> - ret = drm_intel_bo_emit_reloc(batch->bo,
>>> - batch_offset(batch, ss) + 4,
>>> - buf->bo, 0,
>>> - read_domain, write_domain);
>>> - igt_assert(ret == 0);
>>> -
>>> - ss->ss2.height = igt_buf_height(buf) - 1;
>>> - ss->ss2.width = igt_buf_width(buf) - 1;
>>> -
>>> - ss->ss3.pitch = buf->stride - 1;
>>> -
>>> - ss->ss7.shader_chanel_select_r = 4;
>>> - ss->ss7.shader_chanel_select_g = 5;
>>> - ss->ss7.shader_chanel_select_b = 6;
>>> - ss->ss7.shader_chanel_select_a = 7;
>>> -
>>> - return offset;
>>> -}
>>> -
>>> -static uint32_t
>>> -gen8_fill_surface_state(struct intel_batchbuffer *batch,
>>> - struct igt_buf *buf,
>>> - uint32_t format,
>>> - int is_dst)
>>> -{
>>> - struct gen8_surface_state *ss;
>>> - uint32_t write_domain, read_domain, offset;
>>> - int ret;
>>> -
>>> - if (is_dst) {
>>> - write_domain = read_domain = I915_GEM_DOMAIN_RENDER;
>>> - } else {
>>> - write_domain = 0;
>>> - read_domain = I915_GEM_DOMAIN_SAMPLER;
>>> - }
>>> -
>>> - ss = batch_alloc(batch, sizeof(*ss), 64);
>>> - offset = batch_offset(batch, ss);
>>> -
>>> - ss->ss0.surface_type = GEN8_SURFACE_2D;
>>> - ss->ss0.surface_format = format;
>>> - ss->ss0.render_cache_read_write = 1;
>>> - ss->ss0.vertical_alignment = 1; /* align 4 */
>>> - ss->ss0.horizontal_alignment = 1; /* align 4 */
>>> -
>>> - if (buf->tiling == I915_TILING_X)
>>> - ss->ss0.tiled_mode = 2;
>>> - else if (buf->tiling == I915_TILING_Y)
>>> - ss->ss0.tiled_mode = 3;
>>> -
>>> - ss->ss8.base_addr = buf->bo->offset;
>>> -
>>> - ret = drm_intel_bo_emit_reloc(batch->bo,
>>> - batch_offset(batch, ss) + 8 * 4,
>>> - buf->bo, 0,
>>> - read_domain, write_domain);
>>> - igt_assert_eq(ret, 0);
>>> -
>>> - ss->ss2.height = igt_buf_height(buf) - 1;
>>> - ss->ss2.width = igt_buf_width(buf) - 1;
>>> - ss->ss3.pitch = buf->stride - 1;
>>> -
>>> - ss->ss7.shader_chanel_select_r = 4;
>>> - ss->ss7.shader_chanel_select_g = 5;
>>> - ss->ss7.shader_chanel_select_b = 6;
>>> - ss->ss7.shader_chanel_select_a = 7;
>>> -
>>> - return offset;
>>> -
>>> -}
>>> -
>>> -static uint32_t
>>> -gen7_fill_binding_table(struct intel_batchbuffer *batch,
>>> - struct igt_buf *dst)
>>> -{
>>> - uint32_t *binding_table, offset;
>>> -
>>> - binding_table = batch_alloc(batch, 32, 64);
>>> - offset = batch_offset(batch, binding_table);
>>> -
>>> - binding_table[0] = gen7_fill_surface_state(batch, dst,
>>> GEN7_SURFACEFORMAT_R8_UNORM, 1);
>>> -
>>> - return offset;
>>> -}
>>> -
>>> -static uint32_t
>>> -gen8_fill_binding_table(struct intel_batchbuffer *batch,
>>> - struct igt_buf *dst)
>>> -{
>>> - uint32_t *binding_table, offset;
>>> -
>>> - binding_table = batch_alloc(batch, 32, 64);
>>> - offset = batch_offset(batch, binding_table);
>>> -
>>> - binding_table[0] = gen8_fill_surface_state(batch, dst,
>>> GEN8_SURFACEFORMAT_R8_UNORM, 1);
>>> -
>>> - return offset;
>>> -}
>>> -
>>> -static uint32_t
>>> -gen7_fill_gpgpu_kernel(struct intel_batchbuffer *batch,
>>> - const uint32_t kernel[][4],
>>> - size_t size)
>>> -{
>>> - uint32_t offset;
>>> -
>>> - offset = batch_copy(batch, kernel, size, 64);
>>> -
>>> - return offset;
>>> -}
>>> -
>>> -static uint32_t
>>> -gen7_fill_interface_descriptor(struct intel_batchbuffer *batch, struct
>>> igt_buf *dst,
>>> - const uint32_t kernel[][4], size_t size)
>>> -{
>>> - struct gen7_interface_descriptor_data *idd;
>>> - uint32_t offset;
>>> - uint32_t binding_table_offset, kernel_offset;
>>> -
>>> - binding_table_offset = gen7_fill_binding_table(batch, dst);
>>> - kernel_offset = gen7_fill_gpgpu_kernel(batch, kernel, size);
>>> -
>>> - idd = batch_alloc(batch, sizeof(*idd), 64);
>>> - offset = batch_offset(batch, idd);
>>> -
>>> - idd->desc0.kernel_start_pointer = (kernel_offset >> 6);
>>> -
>>> - idd->desc1.single_program_flow = 1;
>>> - idd->desc1.floating_point_mode = GEN7_FLOATING_POINT_IEEE_754;
>>> -
>>> - idd->desc2.sampler_count = 0; /* 0 samplers used */
>>> - idd->desc2.sampler_state_pointer = 0;
>>> -
>>> - idd->desc3.binding_table_entry_count = 0;
>>> - idd->desc3.binding_table_pointer = (binding_table_offset >> 5);
>>> -
>>> - idd->desc4.constant_urb_entry_read_offset = 0;
>>> - idd->desc4.constant_urb_entry_read_length = 1; /* grf 1 */
>>> -
>>> - return offset;
>>> -}
>>> -
>>> -static uint32_t
>>> -gen8_fill_interface_descriptor(struct intel_batchbuffer *batch, struct
>>> igt_buf *dst,
>>> - const uint32_t kernel[][4], size_t size)
>>> -{
>>> - struct gen8_interface_descriptor_data *idd;
>>> - uint32_t offset;
>>> - uint32_t binding_table_offset, kernel_offset;
>>> -
>>> - binding_table_offset = gen8_fill_binding_table(batch, dst);
>>> - kernel_offset = gen7_fill_gpgpu_kernel(batch, kernel, size);
>>> -
>>> - idd = batch_alloc(batch, sizeof(*idd), 64);
>>> - offset = batch_offset(batch, idd);
>>> -
>>> - idd->desc0.kernel_start_pointer = (kernel_offset >> 6);
>>> -
>>> - idd->desc2.single_program_flow = 1;
>>> - idd->desc2.floating_point_mode = GEN8_FLOATING_POINT_IEEE_754;
>>> -
>>> - idd->desc3.sampler_count = 0; /* 0 samplers used */
>>> - idd->desc3.sampler_state_pointer = 0;
>>> -
>>> - idd->desc4.binding_table_entry_count = 0;
>>> - idd->desc4.binding_table_pointer = (binding_table_offset >> 5);
>>> -
>>> - idd->desc5.constant_urb_entry_read_offset = 0;
>>> - idd->desc5.constant_urb_entry_read_length = 1; /* grf 1 */
>>> -
>>> - return offset;
>>> -}
>>> -
>>> -static void
>>> -gen7_emit_state_base_address(struct intel_batchbuffer *batch)
>>> -{
>>> - OUT_BATCH(GEN7_STATE_BASE_ADDRESS | (10 - 2));
>>> -
>>> - /* general */
>>> - OUT_BATCH(0);
>>> -
>>> - /* surface */
>>> - OUT_RELOC(batch->bo, I915_GEM_DOMAIN_INSTRUCTION, 0,
>>> BASE_ADDRESS_MODIFY);
>>> -
>>> - /* dynamic */
>>> - OUT_RELOC(batch->bo, I915_GEM_DOMAIN_INSTRUCTION, 0,
>>> BASE_ADDRESS_MODIFY);
>>> -
>>> - /* indirect */
>>> - OUT_BATCH(0);
>>> -
>>> - /* instruction */
>>> - OUT_RELOC(batch->bo, I915_GEM_DOMAIN_INSTRUCTION, 0,
>>> BASE_ADDRESS_MODIFY);
>>> -
>>> - /* general/dynamic/indirect/instruction access Bound */
>>> - OUT_BATCH(0);
>>> - OUT_BATCH(0 | BASE_ADDRESS_MODIFY);
>>> - OUT_BATCH(0);
>>> - OUT_BATCH(0 | BASE_ADDRESS_MODIFY);
>>> -}
>>> -
>>> -static void
>>> -gen8_emit_state_base_address(struct intel_batchbuffer *batch)
>>> -{
>>> - OUT_BATCH(GEN8_STATE_BASE_ADDRESS | (16 - 2));
>>> -
>>> - /* general */
>>> - OUT_BATCH(0 | (0x78 << 4) | (0 << 1) | BASE_ADDRESS_MODIFY);
>>> - OUT_BATCH(0);
>>> -
>>> - /* stateless data port */
>>> - OUT_BATCH(0 | BASE_ADDRESS_MODIFY);
>>> -
>>> - /* surface */
>>> - OUT_RELOC(batch->bo, I915_GEM_DOMAIN_SAMPLER, 0,
>>> BASE_ADDRESS_MODIFY);
>>> -
>>> - /* dynamic */
>>> - OUT_RELOC(batch->bo, I915_GEM_DOMAIN_RENDER |
>>> I915_GEM_DOMAIN_INSTRUCTION,
>>> - 0, BASE_ADDRESS_MODIFY);
>>> -
>>> - /* indirect */
>>> - OUT_BATCH(0);
>>> - OUT_BATCH(0 );
>>> -
>>> - /* instruction */
>>> - OUT_RELOC(batch->bo, I915_GEM_DOMAIN_INSTRUCTION, 0,
>>> BASE_ADDRESS_MODIFY);
>>> -
>>> - /* general state buffer size */
>>> - OUT_BATCH(0xfffff000 | 1);
>>> - /* dynamic state buffer size */
>>> - OUT_BATCH(1 << 12 | 1);
>>> - /* indirect object buffer size */
>>> - OUT_BATCH(0xfffff000 | 1);
>>> - /* intruction buffer size, must set modify enable bit, otherwise it
>>> may result in GPU hang */
>>> - OUT_BATCH(1 << 12 | 1);
>>> -}
>>> -
>>> -static void
>>> -gen9_emit_state_base_address(struct intel_batchbuffer *batch)
>>> -{
>>> - OUT_BATCH(GEN8_STATE_BASE_ADDRESS | (19 - 2));
>>> -
>>> - /* general */
>>> - OUT_BATCH(0 | BASE_ADDRESS_MODIFY);
>>> - OUT_BATCH(0);
>>> -
>>> - /* stateless data port */
>>> - OUT_BATCH(0 | BASE_ADDRESS_MODIFY);
>>> -
>>> - /* surface */
>>> - OUT_RELOC(batch->bo, I915_GEM_DOMAIN_SAMPLER, 0,
>>> BASE_ADDRESS_MODIFY);
>>> -
>>> - /* dynamic */
>>> - OUT_RELOC(batch->bo, I915_GEM_DOMAIN_RENDER |
>>> I915_GEM_DOMAIN_INSTRUCTION,
>>> - 0, BASE_ADDRESS_MODIFY);
>>> -
>>> - /* indirect */
>>> - OUT_BATCH(0);
>>> - OUT_BATCH(0);
>>> -
>>> - /* instruction */
>>> - OUT_RELOC(batch->bo, I915_GEM_DOMAIN_INSTRUCTION, 0,
>>> BASE_ADDRESS_MODIFY);
>>> -
>>> - /* general state buffer size */
>>> - OUT_BATCH(0xfffff000 | 1);
>>> - /* dynamic state buffer size */
>>> - OUT_BATCH(1 << 12 | 1);
>>> - /* indirect object buffer size */
>>> - OUT_BATCH(0xfffff000 | 1);
>>> - /* intruction buffer size, must set modify enable bit, otherwise it
>>> may result in GPU hang */
>>> - OUT_BATCH(1 << 12 | 1);
>>> -
>>> - /* Bindless surface state base address */
>>> - OUT_BATCH(0 | BASE_ADDRESS_MODIFY);
>>> - OUT_BATCH(0);
>>> - OUT_BATCH(0xfffff000);
>>> -}
>>> -
>>> -static void
>>> -gen7_emit_vfe_state_gpgpu(struct intel_batchbuffer *batch)
>>> -{
>>> - OUT_BATCH(GEN7_MEDIA_VFE_STATE | (8 - 2));
>>> -
>>> - /* scratch buffer */
>>> - OUT_BATCH(0);
>>> -
>>> - /* number of threads & urb entries */
>>> - OUT_BATCH(1 << 16 | /* max num of threads */
>>> - 0 << 8 | /* num of URB entry */
>>> - 1 << 2); /* GPGPU mode */
>>> -
>>> - OUT_BATCH(0);
>>> -
>>> - /* urb entry size & curbe size */
>>> - OUT_BATCH(0 << 16 | /* URB entry size in 256 bits unit */
>>> - 1); /* CURBE entry size in 256 bits unit
>>> */
>>> -
>>> - /* scoreboard */
>>> - OUT_BATCH(0);
>>> - OUT_BATCH(0);
>>> - OUT_BATCH(0);
>>> -}
>>> -
>>> -static void
>>> -gen8_emit_vfe_state_gpgpu(struct intel_batchbuffer *batch)
>>> -{
>>> - OUT_BATCH(GEN8_MEDIA_VFE_STATE | (9 - 2));
>>> -
>>> - /* scratch buffer */
>>> - OUT_BATCH(0);
>>> - OUT_BATCH(0);
>>> -
>>> - /* number of threads & urb entries */
>>> - OUT_BATCH(1 << 16 | 1 << 8);
>>> -
>>> - OUT_BATCH(0);
>>> -
>>> - /* urb entry size & curbe size */
>>> - OUT_BATCH(0 << 16 | 1);
>>> -
>>> - /* scoreboard */
>>> - OUT_BATCH(0);
>>> - OUT_BATCH(0);
>>> - OUT_BATCH(0);
>>> -}
>>> -
>>> -static void
>>> -gen7_emit_curbe_load(struct intel_batchbuffer *batch, uint32_t
>>> curbe_buffer)
>>> -{
>>> - OUT_BATCH(GEN7_MEDIA_CURBE_LOAD | (4 - 2));
>>> - OUT_BATCH(0);
>>> - /* curbe total data length */
>>> - OUT_BATCH(64);
>>> - /* curbe data start address, is relative to the dynamics base
>>> address */
>>> - OUT_BATCH(curbe_buffer);
>>> -}
>>> -
>>> -static void
>>> -gen7_emit_interface_descriptor_load(struct intel_batchbuffer *batch,
>>> uint32_t interface_descriptor)
>>> -{
>>> - OUT_BATCH(GEN7_MEDIA_INTERFACE_DESCRIPTOR_LOAD | (4 - 2));
>>> - OUT_BATCH(0);
>>> - /* interface descriptor data length */
>>> - OUT_BATCH(sizeof(struct gen7_interface_descriptor_data));
>>> - /* interface descriptor address, is relative to the dynamics base
>>> address */
>>> - OUT_BATCH(interface_descriptor);
>>> -}
>>> -
>>> -static void
>>> -gen8_emit_interface_descriptor_load(struct intel_batchbuffer *batch,
>>> uint32_t interface_descriptor)
>>> -{
>>> - OUT_BATCH(GEN8_MEDIA_INTERFACE_DESCRIPTOR_LOAD | (4 - 2));
>>> - OUT_BATCH(0);
>>> - /* interface descriptor data length */
>>> - OUT_BATCH(sizeof(struct gen8_interface_descriptor_data));
>>> - /* interface descriptor address, is relative to the dynamics base
>>> address */
>>> - OUT_BATCH(interface_descriptor);
>>> -}
>>> -
>>> -static void
>>> -gen7_emit_gpgpu_walk(struct intel_batchbuffer *batch,
>>> - unsigned x, unsigned y,
>>> - unsigned width, unsigned height)
>>> -{
>>> - uint32_t x_dim, y_dim, tmp, right_mask;
>>> -
>>> - /*
>>> - * Simply do SIMD16 based dispatch, so every thread uses
>>> - * SIMD16 channels.
>>> - *
>>> - * Define our own thread group size, e.g 16x1 for every group, then
>>> - * will have 1 thread each group in SIMD16 dispatch. So thread
>>> - * width/height/depth are all 1.
>>> - *
>>> - * Then thread group X = width / 16 (aligned to 16)
>>> - * thread group Y = height;
>>> - */
>>> - x_dim = (width + 15) / 16;
>>> - y_dim = height;
>>> -
>>> - tmp = width & 15;
>>> - if (tmp == 0)
>>> - right_mask = (1 << 16) - 1;
>>> - else
>>> - right_mask = (1 << tmp) - 1;
>>> -
>>> - OUT_BATCH(GEN7_GPGPU_WALKER | 9);
>>> -
>>> - /* interface descriptor offset */
>>> - OUT_BATCH(0);
>>> -
>>> - /* SIMD size, thread w/h/d */
>>> - OUT_BATCH(1 << 30 | /* SIMD16 */
>>> - 0 << 16 | /* depth:1 */
>>> - 0 << 8 | /* height:1 */
>>> - 0); /* width:1 */
>>> -
>>> - /* thread group X */
>>> - OUT_BATCH(0);
>>> - OUT_BATCH(x_dim);
>>> -
>>> - /* thread group Y */
>>> - OUT_BATCH(0);
>>> - OUT_BATCH(y_dim);
>>> -
>>> - /* thread group Z */
>>> - OUT_BATCH(0);
>>> - OUT_BATCH(1);
>>> -
>>> - /* right mask */
>>> - OUT_BATCH(right_mask);
>>> -
>>> - /* bottom mask, height 1, always 0xffffffff */
>>> - OUT_BATCH(0xffffffff);
>>> -}
>>> -
>>> -static void
>>> -gen8_emit_gpgpu_walk(struct intel_batchbuffer *batch,
>>> - unsigned x, unsigned y,
>>> - unsigned width, unsigned height)
>>> -{
>>> - uint32_t x_dim, y_dim, tmp, right_mask;
>>> -
>>> - /*
>>> - * Simply do SIMD16 based dispatch, so every thread uses
>>> - * SIMD16 channels.
>>> - *
>>> - * Define our own thread group size, e.g 16x1 for every group, then
>>> - * will have 1 thread each group in SIMD16 dispatch. So thread
>>> - * width/height/depth are all 1.
>>> - *
>>> - * Then thread group X = width / 16 (aligned to 16)
>>> - * thread group Y = height;
>>> - */
>>> - x_dim = (width + 15) / 16;
>>> - y_dim = height;
>>> -
>>> - tmp = width & 15;
>>> - if (tmp == 0)
>>> - right_mask = (1 << 16) - 1;
>>> - else
>>> - right_mask = (1 << tmp) - 1;
>>> -
>>> - OUT_BATCH(GEN7_GPGPU_WALKER | 13);
>>> -
>>> - OUT_BATCH(0); /* kernel offset */
>>> - OUT_BATCH(0); /* indirect data length */
>>> - OUT_BATCH(0); /* indirect data offset */
>>> -
>>> - /* SIMD size, thread w/h/d */
>>> - OUT_BATCH(1 << 30 | /* SIMD16 */
>>> - 0 << 16 | /* depth:1 */
>>> - 0 << 8 | /* height:1 */
>>> - 0); /* width:1 */
>>> -
>>> - /* thread group X */
>>> - OUT_BATCH(0);
>>> - OUT_BATCH(0);
>>> - OUT_BATCH(x_dim);
>>> -
>>> - /* thread group Y */
>>> - OUT_BATCH(0);
>>> - OUT_BATCH(0);
>>> - OUT_BATCH(y_dim);
>>> -
>>> - /* thread group Z */
>>> - OUT_BATCH(0);
>>> - OUT_BATCH(1);
>>> -
>>> - /* right mask */
>>> - OUT_BATCH(right_mask);
>>> -
>>> - /* bottom mask, height 1, always 0xffffffff */
>>> - OUT_BATCH(0xffffffff);
>>> -}
>>> -
>>> /*
>>> * This sets up the gpgpu pipeline,
>>> *
>>> diff --git a/lib/gpu_fill.c b/lib/gpu_fill.c
>>> new file mode 100644
>>> index 00000000..f4563112
>>> --- /dev/null
>>> +++ b/lib/gpu_fill.c
>>> @@ -0,0 +1,729 @@
>>> +#include <intel_bufmgr.h>
>>> +#include <i915_drm.h>
>>> +
>>> +#include "intel_reg.h"
>>> +#include "drmtest.h"
>>> +#include "gpu_fill.h"
>>> +#include <assert.h>
>>> +#include "gen7_media.h"
>>> +#include "gen8_media.h"
>>> +
>>> +
>>> +uint32_t
>>> +batch_used(struct intel_batchbuffer *batch)
>>> +{
>>> + return batch->ptr - batch->buffer;
>>> +}
>>> +
>>> +uint32_t
>>> +batch_align(struct intel_batchbuffer *batch, uint32_t align)
>>> +{
>>> + uint32_t offset = batch_used(batch);
>>> + offset = ALIGN(offset, align);
>>> + batch->ptr = batch->buffer + offset;
>>> + return offset;
>>> +}
>>> +
>>> +void *
>>> +batch_alloc(struct intel_batchbuffer *batch, uint32_t size, uint32_t align)
>>> +{
>>> + uint32_t offset = batch_align(batch, align);
>>> + batch->ptr += size;
>>> + return memset(batch->buffer + offset, 0, size);
>>> +}
>>> +
>>> +uint32_t
>>> +batch_offset(struct intel_batchbuffer *batch, void *ptr)
>>> +{
>>> + return (uint8_t *)ptr - batch->buffer;
>>> +}
>>> +
>>> +uint32_t
>>> +batch_copy(struct intel_batchbuffer *batch, const void *ptr, uint32_t size,
>>> uint32_t align)
>>> +{
>>> + return batch_offset(batch, memcpy(batch_alloc(batch, size, align),
>>> ptr, size));
>>> +}
>>> +
>>> +void
>>> +gen7_render_flush(struct intel_batchbuffer *batch, uint32_t batch_end)
>>> +{
>>> + int ret;
>>> +
>>> + ret = drm_intel_bo_subdata(batch->bo, 0, 4096, batch->buffer);
>>> + if (ret == 0)
>>> + ret = drm_intel_bo_mrb_exec(batch->bo, batch_end,
>>> + NULL, 0, 0, 0);
>>> + igt_assert(ret == 0);
>>> +}
>>> +
>>> +uint32_t
>>> +gen7_fill_curbe_buffer_data(struct intel_batchbuffer *batch,
>>> + uint8_t color)
>>> +{
>>> + uint8_t *curbe_buffer;
>>> + uint32_t offset;
>>> +
>>> + curbe_buffer = batch_alloc(batch, sizeof(uint32_t) * 8, 64);
>>> + offset = batch_offset(batch, curbe_buffer);
>>> + *curbe_buffer = color;
>>> +
>>> + return offset;
>>> +}
>>> +
>>> +uint32_t
>>> +gen7_fill_surface_state(struct intel_batchbuffer *batch,
>>> + struct igt_buf *buf,
>>> + uint32_t format,
>>> + int is_dst)
>>> +{
>>> + struct gen7_surface_state *ss;
>>> + uint32_t write_domain, read_domain, offset;
>>> + int ret;
>>> +
>>> + if (is_dst) {
>>> + write_domain = read_domain = I915_GEM_DOMAIN_RENDER;
>>> + } else {
>>> + write_domain = 0;
>>> + read_domain = I915_GEM_DOMAIN_SAMPLER;
>>> + }
>>> +
>>> + ss = batch_alloc(batch, sizeof(*ss), 64);
>>> + offset = batch_offset(batch, ss);
>>> +
>>> + ss->ss0.surface_type = GEN7_SURFACE_2D;
>>> + ss->ss0.surface_format = format;
>>> + ss->ss0.render_cache_read_write = 1;
>>> +
>>> + if (buf->tiling == I915_TILING_X)
>>> + ss->ss0.tiled_mode = 2;
>>> + else if (buf->tiling == I915_TILING_Y)
>>> + ss->ss0.tiled_mode = 3;
>>> +
>>> + ss->ss1.base_addr = buf->bo->offset;
>>> + ret = drm_intel_bo_emit_reloc(batch->bo,
>>> + batch_offset(batch, ss) + 4,
>>> + buf->bo, 0,
>>> + read_domain, write_domain);
>>> + igt_assert(ret == 0);
>>> +
>>> + ss->ss2.height = igt_buf_height(buf) - 1;
>>> + ss->ss2.width = igt_buf_width(buf) - 1;
>>> +
>>> + ss->ss3.pitch = buf->stride - 1;
>>> +
>>> + ss->ss7.shader_chanel_select_r = 4;
>>> + ss->ss7.shader_chanel_select_g = 5;
>>> + ss->ss7.shader_chanel_select_b = 6;
>>> + ss->ss7.shader_chanel_select_a = 7;
>>> +
>>> + return offset;
>>> +}
>>> +
>>> +uint32_t
>>> +gen7_fill_binding_table(struct intel_batchbuffer *batch,
>>> + struct igt_buf *dst)
>>> +{
>>> + uint32_t *binding_table, offset;
>>> +
>>> + binding_table = batch_alloc(batch, 32, 64);
>>> + offset = batch_offset(batch, binding_table);
>>> +
>>> + binding_table[0] = gen7_fill_surface_state(batch, dst,
>>> GEN7_SURFACEFORMAT_R8_UNORM, 1);
>>> +
>>> + return offset;
>>> +}
>>> +
>>> +uint32_t
>>> +gen7_fill_kernel(struct intel_batchbuffer *batch,
>>> + const uint32_t kernel[][4],
>>> + size_t size)
>>> +{
>>> + uint32_t offset;
>>> +
>>> + offset = batch_copy(batch, kernel, size, 64);
>>> +
>>> + return offset;
>>> +}
>>> +
>>> +uint32_t
>>> +gen7_fill_interface_descriptor(struct intel_batchbuffer *batch, struct
>>> igt_buf *dst,
>>> + const uint32_t kernel[][4], size_t size)
>>> +{
>>> + struct gen7_interface_descriptor_data *idd;
>>> + uint32_t offset;
>>> + uint32_t binding_table_offset, kernel_offset;
>>> +
>>> + binding_table_offset = gen7_fill_binding_table(batch, dst);
>>> + kernel_offset = gen7_fill_kernel(batch, kernel, size);
>>> +
>>> + idd = batch_alloc(batch, sizeof(*idd), 64);
>>> + offset = batch_offset(batch, idd);
>>> +
>>> + idd->desc0.kernel_start_pointer = (kernel_offset >> 6);
>>> +
>>> + idd->desc1.single_program_flow = 1;
>>> + idd->desc1.floating_point_mode = GEN7_FLOATING_POINT_IEEE_754;
>>> +
>>> + idd->desc2.sampler_count = 0; /* 0 samplers used */
>>> + idd->desc2.sampler_state_pointer = 0;
>>> +
>>> + idd->desc3.binding_table_entry_count = 0;
>>> + idd->desc3.binding_table_pointer = (binding_table_offset >> 5);
>>> +
>>> + idd->desc4.constant_urb_entry_read_offset = 0;
>>> + idd->desc4.constant_urb_entry_read_length = 1; /* grf 1 */
>>> +
>>> + return offset;
>>> +}
>>> +
>>> +void
>>> +gen7_emit_state_base_address(struct intel_batchbuffer *batch)
>>> +{
>>> + OUT_BATCH(GEN7_STATE_BASE_ADDRESS | (10 - 2));
>>> +
>>> + /* general */
>>> + OUT_BATCH(0);
>>> +
>>> + /* surface */
>>> + OUT_RELOC(batch->bo, I915_GEM_DOMAIN_INSTRUCTION, 0,
>>> BASE_ADDRESS_MODIFY);
>>> +
>>> + /* dynamic */
>>> + OUT_RELOC(batch->bo, I915_GEM_DOMAIN_INSTRUCTION, 0,
>>> BASE_ADDRESS_MODIFY);
>>> +
>>> + /* indirect */
>>> + OUT_BATCH(0);
>>> +
>>> + /* instruction */
>>> + OUT_RELOC(batch->bo, I915_GEM_DOMAIN_INSTRUCTION, 0,
>>> BASE_ADDRESS_MODIFY);
>>> +
>>> + /* general/dynamic/indirect/instruction access Bound */
>>> + OUT_BATCH(0);
>>> + OUT_BATCH(0 | BASE_ADDRESS_MODIFY);
>>> + OUT_BATCH(0);
>>> + OUT_BATCH(0 | BASE_ADDRESS_MODIFY);
>>> +}
>>> +
>>> +void
>>> +gen7_emit_vfe_state(struct intel_batchbuffer *batch)
>>> +{
>>> + OUT_BATCH(GEN7_MEDIA_VFE_STATE | (8 - 2));
>>> +
>>> + /* scratch buffer */
>>> + OUT_BATCH(0);
>>> +
>>> + /* number of threads & urb entries */
>>> + OUT_BATCH(1 << 16 |
>>> + 2 << 8);
>>> +
>>> + OUT_BATCH(0);
>>> +
>>> + /* urb entry size & curbe size */
>>> + OUT_BATCH(2 << 16 | /* in 256 bits unit */
>>> + 2); /* in 256 bits unit */
>>> +
>>> + /* scoreboard */
>>> + OUT_BATCH(0);
>>> + OUT_BATCH(0);
>>> + OUT_BATCH(0);
>>> +}
>>> +
>>> +void
>>> +gen7_emit_vfe_state_gpgpu(struct intel_batchbuffer *batch)
>>> +{
>>> + OUT_BATCH(GEN7_MEDIA_VFE_STATE | (8 - 2));
>>> +
>>> + /* scratch buffer */
>>> + OUT_BATCH(0);
>>> +
>>> + /* number of threads & urb entries */
>>> + OUT_BATCH(1 << 16 | /* max num of threads */
>>> + 0 << 8 | /* num of URB entry */
>>> + 1 << 2); /* GPGPU mode */
>>> +
>>> + OUT_BATCH(0);
>>> +
>>> + /* urb entry size & curbe size */
>>> + OUT_BATCH(0 << 16 | /* URB entry size in 256 bits unit */
>>> + 1); /* CURBE entry size in 256 bits unit
>>> */
>>> +
>>> + /* scoreboard */
>>> + OUT_BATCH(0);
>>> + OUT_BATCH(0);
>>> + OUT_BATCH(0);
>>> +}
>>> +
>>> +void
>>> +gen7_emit_curbe_load(struct intel_batchbuffer *batch, uint32_t
>>> curbe_buffer)
>>> +{
>>> + OUT_BATCH(GEN7_MEDIA_CURBE_LOAD | (4 - 2));
>>> + OUT_BATCH(0);
>>> + /* curbe total data length */
>>> + OUT_BATCH(64);
>>> + /* curbe data start address, is relative to the dynamics base
>>> address */
>>> + OUT_BATCH(curbe_buffer);
>>> +}
>>> +
>>> +void
>>> +gen7_emit_interface_descriptor_load(struct intel_batchbuffer *batch,
>>> uint32_t interface_descriptor)
>>> +{
>>> + OUT_BATCH(GEN7_MEDIA_INTERFACE_DESCRIPTOR_LOAD | (4 - 2));
>>> + OUT_BATCH(0);
>>> + /* interface descriptor data length */
>>> + OUT_BATCH(sizeof(struct gen7_interface_descriptor_data));
>>> + /* interface descriptor address, is relative to the dynamics base
>>> address */
>>> + OUT_BATCH(interface_descriptor);
>>> +}
>>> +
>>> +void
>>> +gen7_emit_media_objects(struct intel_batchbuffer *batch,
>>> + unsigned x, unsigned y,
>>> + unsigned width, unsigned height)
>>> +{
>>> + int i, j;
>>> +
>>> + for (i = 0; i < width / 16; i++) {
>>> + for (j = 0; j < height / 16; j++) {
>>> + OUT_BATCH(GEN7_MEDIA_OBJECT | (8 - 2));
>>> +
>>> + /* interface descriptor offset */
>>> + OUT_BATCH(0);
>>> +
>>> + /* without indirect data */
>>> + OUT_BATCH(0);
>>> + OUT_BATCH(0);
>>> +
>>> + /* scoreboard */
>>> + OUT_BATCH(0);
>>> + OUT_BATCH(0);
>>> +
>>> + /* inline data (xoffset, yoffset) */
>>> + OUT_BATCH(x + i * 16);
>>> + OUT_BATCH(y + j * 16);
>>> + }
>>> + }
>>> +}
>>> +
>>> +uint32_t
>>> +gen8_fill_surface_state(struct intel_batchbuffer *batch,
>>> + struct igt_buf *buf,
>>> + uint32_t format,
>>> + int is_dst)
>>> +{
>>> + struct gen8_surface_state *ss;
>>> + uint32_t write_domain, read_domain, offset;
>>> + int ret;
>>> +
>>> + if (is_dst) {
>>> + write_domain = read_domain = I915_GEM_DOMAIN_RENDER;
>>> + } else {
>>> + write_domain = 0;
>>> + read_domain = I915_GEM_DOMAIN_SAMPLER;
>>> + }
>>> +
>>> + ss = batch_alloc(batch, sizeof(*ss), 64);
>>> + offset = batch_offset(batch, ss);
>>> +
>>> + ss->ss0.surface_type = GEN8_SURFACE_2D;
>>> + ss->ss0.surface_format = format;
>>> + ss->ss0.render_cache_read_write = 1;
>>> + ss->ss0.vertical_alignment = 1; /* align 4 */
>>> + ss->ss0.horizontal_alignment = 1; /* align 4 */
>>> +
>>> + if (buf->tiling == I915_TILING_X)
>>> + ss->ss0.tiled_mode = 2;
>>> + else if (buf->tiling == I915_TILING_Y)
>>> + ss->ss0.tiled_mode = 3;
>>> +
>>> + ss->ss8.base_addr = buf->bo->offset;
>>> +
>>> + ret = drm_intel_bo_emit_reloc(batch->bo,
>>> + batch_offset(batch, ss) + 8 * 4,
>>> + buf->bo, 0,
>>> + read_domain, write_domain);
>>> + igt_assert(ret == 0);
>>> +
>>> + ss->ss2.height = igt_buf_height(buf) - 1;
>>> + ss->ss2.width = igt_buf_width(buf) - 1;
>>> + ss->ss3.pitch = buf->stride - 1;
>>> +
>>> + ss->ss7.shader_chanel_select_r = 4;
>>> + ss->ss7.shader_chanel_select_g = 5;
>>> + ss->ss7.shader_chanel_select_b = 6;
>>> + ss->ss7.shader_chanel_select_a = 7;
>>> +
>>> + return offset;
>>> +}
>>> +
>>> +uint32_t
>>> +gen8_fill_binding_table(struct intel_batchbuffer *batch,
>>> + struct igt_buf *dst)
>>> +{
>>> + uint32_t *binding_table, offset;
>>> +
>>> + binding_table = batch_alloc(batch, 32, 64);
>>> + offset = batch_offset(batch, binding_table);
>>> +
>>> + binding_table[0] = gen8_fill_surface_state(batch, dst,
>>> GEN8_SURFACEFORMAT_R8_UNORM, 1);
>>> +
>>> + return offset;
>>> +}
>>> +
>>> +uint32_t
>>> +gen8_fill_interface_descriptor(struct intel_batchbuffer *batch, struct
>>> igt_buf *dst, const uint32_t kernel[][4], size_t size)
>>> +{
>>> + struct gen8_interface_descriptor_data *idd;
>>> + uint32_t offset;
>>> + uint32_t binding_table_offset, kernel_offset;
>>> +
>>> + binding_table_offset = gen8_fill_binding_table(batch, dst);
>>> + kernel_offset = gen7_fill_kernel(batch, kernel, sizeof(kernel));
>>> +
>>> + idd = batch_alloc(batch, sizeof(*idd), 64);
>>> + offset = batch_offset(batch, idd);
>>> +
>>> + idd->desc0.kernel_start_pointer = (kernel_offset >> 6);
>>> +
>>> + idd->desc2.single_program_flow = 1;
>>> + idd->desc2.floating_point_mode = GEN8_FLOATING_POINT_IEEE_754;
>>> +
>>> + idd->desc3.sampler_count = 0; /* 0 samplers used */
>>> + idd->desc3.sampler_state_pointer = 0;
>>> +
>>> + idd->desc4.binding_table_entry_count = 0;
>>> + idd->desc4.binding_table_pointer = (binding_table_offset >> 5);
>>> +
>>> + idd->desc5.constant_urb_entry_read_offset = 0;
>>> + idd->desc5.constant_urb_entry_read_length = 1; /* grf 1 */
>>> +
>>> + return offset;
>>> +}
>>> +
>>> +void
>>> +gen8_emit_state_base_address(struct intel_batchbuffer *batch)
>>> +{
>>> + OUT_BATCH(GEN8_STATE_BASE_ADDRESS | (16 - 2));
>>> +
>>> + /* general */
>>> + OUT_BATCH(0 | BASE_ADDRESS_MODIFY);
>>> + OUT_BATCH(0);
>>> +
>>> + /* stateless data port */
>>> + OUT_BATCH(0 | BASE_ADDRESS_MODIFY);
>>> +
>>> + /* surface */
>>> + OUT_RELOC(batch->bo, I915_GEM_DOMAIN_SAMPLER, 0,
>>> BASE_ADDRESS_MODIFY);
>>> +
>>> + /* dynamic */
>>> + OUT_RELOC(batch->bo, I915_GEM_DOMAIN_RENDER |
>>> I915_GEM_DOMAIN_INSTRUCTION,
>>> + 0, BASE_ADDRESS_MODIFY);
>>> +
>>> + /* indirect */
>>> + OUT_BATCH(0);
>>> + OUT_BATCH(0);
>>> +
>>> + /* instruction */
>>> + OUT_RELOC(batch->bo, I915_GEM_DOMAIN_INSTRUCTION, 0,
>>> BASE_ADDRESS_MODIFY);
>>> +
>>> + /* general state buffer size */
>>> + OUT_BATCH(0xfffff000 | 1);
>>> + /* dynamic state buffer size */
>>> + OUT_BATCH(1 << 12 | 1);
>>> + /* indirect object buffer size */
>>> + OUT_BATCH(0xfffff000 | 1);
>>> + /* intruction buffer size, must set modify enable bit, otherwise it
>>> may result in GPU hang */
>>> + OUT_BATCH(1 << 12 | 1);
>>> +}
>>> +
>>> +void
>>> +gen8_emit_vfe_state(struct intel_batchbuffer *batch)
>>> +{
>>> + OUT_BATCH(GEN8_MEDIA_VFE_STATE | (9 - 2));
>>> +
>>> + /* scratch buffer */
>>> + OUT_BATCH(0);
>>> + OUT_BATCH(0);
>>> +
>>> + /* number of threads & urb entries */
>>> + OUT_BATCH(1 << 16 |
>>> + 2 << 8);
>>> +
>>> + OUT_BATCH(0);
>>> +
>>> + /* urb entry size & curbe size */
>>> + OUT_BATCH(2 << 16 |
>>> + 2);
>>> +
>>> + /* scoreboard */
>>> + OUT_BATCH(0);
>>> + OUT_BATCH(0);
>>> + OUT_BATCH(0);
>>> +}
>>> +
>>> +void
>>> +gen8_emit_vfe_state_gpgpu(struct intel_batchbuffer *batch)
>>> +{
>>> + OUT_BATCH(GEN8_MEDIA_VFE_STATE | (9 - 2));
>>> +
>>> + /* scratch buffer */
>>> + OUT_BATCH(0);
>>> + OUT_BATCH(0);
>>> +
>>> + /* number of threads & urb entries */
>>> + OUT_BATCH(1 << 16 | 1 << 8);
>>> +
>>> + OUT_BATCH(0);
>>> +
>>> + /* urb entry size & curbe size */
>>> + OUT_BATCH(0 << 16 | 1);
>>> +
>>> + /* scoreboard */
>>> + OUT_BATCH(0);
>>> + OUT_BATCH(0);
>>> + OUT_BATCH(0);
>>> +}
>>> +
>>> +void
>>> +gen8_emit_curbe_load(struct intel_batchbuffer *batch, uint32_t
>>> curbe_buffer)
>>> +{
>>> + OUT_BATCH(GEN8_MEDIA_CURBE_LOAD | (4 - 2));
>>> + OUT_BATCH(0);
>>> + /* curbe total data length */
>>> + OUT_BATCH(64);
>>> + /* curbe data start address, is relative to the dynamics base
>>> address */
>>> + OUT_BATCH(curbe_buffer);
>>> +}
>>> +
>>> +void
>>> +gen8_emit_interface_descriptor_load(struct intel_batchbuffer *batch,
>>> uint32_t interface_descriptor)
>>> +{
>>> + OUT_BATCH(GEN8_MEDIA_INTERFACE_DESCRIPTOR_LOAD | (4 - 2));
>>> + OUT_BATCH(0);
>>> + /* interface descriptor data length */
>>> + OUT_BATCH(sizeof(struct gen8_interface_descriptor_data));
>>> + /* interface descriptor address, is relative to the dynamics base
>>> address */
>>> + OUT_BATCH(interface_descriptor);
>>> +}
>>> +
>>> +void
>>> +gen8_emit_media_state_flush(struct intel_batchbuffer *batch)
>>> +{
>>> + OUT_BATCH(GEN8_MEDIA_STATE_FLUSH | (2 - 2));
>>> + OUT_BATCH(0);
>>> +}
>>> +
>>> +void
>>> +gen8_emit_media_objects(struct intel_batchbuffer *batch,
>>> + unsigned x, unsigned y,
>>> + unsigned width, unsigned height)
>>> +{
>>> + int i, j;
>>> +
>>> + for (i = 0; i < width / 16; i++) {
>>> + for (j = 0; j < height / 16; j++) {
>>> + OUT_BATCH(GEN8_MEDIA_OBJECT | (8 - 2));
>>> +
>>> + /* interface descriptor offset */
>>> + OUT_BATCH(0);
>>> +
>>> + /* without indirect data */
>>> + OUT_BATCH(0);
>>> + OUT_BATCH(0);
>>> +
>>> + /* scoreboard */
>>> + OUT_BATCH(0);
>>> + OUT_BATCH(0);
>>> +
>>> + /* inline data (xoffset, yoffset) */
>>> + OUT_BATCH(x + i * 16);
>>> + OUT_BATCH(y + j * 16);
>>> + gen8_emit_media_state_flush(batch);
>>> + }
>>> + }
>>> +}
>>> +
>>> +void
>>> +gen8lp_emit_media_objects(struct intel_batchbuffer *batch,
>>> + unsigned x, unsigned y,
>>> + unsigned width, unsigned height)
>>> +{
>>> + int i, j;
>>> +
>>> + for (i = 0; i < width / 16; i++) {
>>> + for (j = 0; j < height / 16; j++) {
>>> + OUT_BATCH(GEN8_MEDIA_OBJECT | (8 - 2));
>>> +
>>> + /* interface descriptor offset */
>>> + OUT_BATCH(0);
>>> +
>>> + /* without indirect data */
>>> + OUT_BATCH(0);
>>> + OUT_BATCH(0);
>>> +
>>> + /* scoreboard */
>>> + OUT_BATCH(0);
>>> + OUT_BATCH(0);
>>> +
>>> + /* inline data (xoffset, yoffset) */
>>> + OUT_BATCH(x + i * 16);
>>> + OUT_BATCH(y + j * 16);
>>> + }
>>> + }
>>> +}
>>> +
>>> +void
>>> +gen9_emit_state_base_address(struct intel_batchbuffer *batch)
>>> +{
>>> + OUT_BATCH(GEN8_STATE_BASE_ADDRESS | (19 - 2));
>>> +
>>> + /* general */
>>> + OUT_BATCH(0 | BASE_ADDRESS_MODIFY);
>>> + OUT_BATCH(0);
>>> +
>>> + /* stateless data port */
>>> + OUT_BATCH(0 | BASE_ADDRESS_MODIFY);
>>> +
>>> + /* surface */
>>> + OUT_RELOC(batch->bo, I915_GEM_DOMAIN_SAMPLER, 0,
>>> BASE_ADDRESS_MODIFY);
>>> +
>>> + /* dynamic */
>>> + OUT_RELOC(batch->bo, I915_GEM_DOMAIN_RENDER |
>>> I915_GEM_DOMAIN_INSTRUCTION,
>>> + 0, BASE_ADDRESS_MODIFY);
>>> +
>>> + /* indirect */
>>> + OUT_BATCH(0);
>>> + OUT_BATCH(0);
>>> +
>>> + /* instruction */
>>> + OUT_RELOC(batch->bo, I915_GEM_DOMAIN_INSTRUCTION, 0,
>>> BASE_ADDRESS_MODIFY);
>>> +
>>> + /* general state buffer size */
>>> + OUT_BATCH(0xfffff000 | 1);
>>> + /* dynamic state buffer size */
>>> + OUT_BATCH(1 << 12 | 1);
>>> + /* indirect object buffer size */
>>> + OUT_BATCH(0xfffff000 | 1);
>>> + /* intruction buffer size, must set modify enable bit, otherwise it
>>> may result in GPU hang */
>>> + OUT_BATCH(1 << 12 | 1);
>>> +
>>> + /* Bindless surface state base address */
>>> + OUT_BATCH(0 | BASE_ADDRESS_MODIFY);
>>> + OUT_BATCH(0);
>>> + OUT_BATCH(0xfffff000);
>>> +}
>>> +
>>> +void
>>> +gen7_emit_gpgpu_walk(struct intel_batchbuffer *batch,
>>> + unsigned x, unsigned y,
>>> + unsigned width, unsigned height)
>>> +{
>>> + uint32_t x_dim, y_dim, tmp, right_mask;
>>> +
>>> + /*
>>> + * Simply do SIMD16 based dispatch, so every thread uses
>>> + * SIMD16 channels.
>>> + *
>>> + * Define our own thread group size, e.g 16x1 for every group, then
>>> + * will have 1 thread each group in SIMD16 dispatch. So thread
>>> + * width/height/depth are all 1.
>>> + *
>>> + * Then thread group X = width / 16 (aligned to 16)
>>> + * thread group Y = height;
>>> + */
>>> + x_dim = (width + 15) / 16;
>>> + y_dim = height;
>>> +
>>> + tmp = width & 15;
>>> + if (tmp == 0)
>>> + right_mask = (1 << 16) - 1;
>>> + else
>>> + right_mask = (1 << tmp) - 1;
>>> +
>>> + OUT_BATCH(GEN7_GPGPU_WALKER | 9);
>>> +
>>> + /* interface descriptor offset */
>>> + OUT_BATCH(0);
>>> +
>>> + /* SIMD size, thread w/h/d */
>>> + OUT_BATCH(1 << 30 | /* SIMD16 */
>>> + 0 << 16 | /* depth:1 */
>>> + 0 << 8 | /* height:1 */
>>> + 0); /* width:1 */
>>> +
>>> + /* thread group X */
>>> + OUT_BATCH(0);
>>> + OUT_BATCH(x_dim);
>>> +
>>> + /* thread group Y */
>>> + OUT_BATCH(0);
>>> + OUT_BATCH(y_dim);
>>> +
>>> + /* thread group Z */
>>> + OUT_BATCH(0);
>>> + OUT_BATCH(1);
>>> +
>>> + /* right mask */
>>> + OUT_BATCH(right_mask);
>>> +
>>> + /* bottom mask, height 1, always 0xffffffff */
>>> + OUT_BATCH(0xffffffff);
>>> +}
>>> +
>>> +void
>>> +gen8_emit_gpgpu_walk(struct intel_batchbuffer *batch,
>>> + unsigned x, unsigned y,
>>> + unsigned width, unsigned height)
>>> +{
>>> + uint32_t x_dim, y_dim, tmp, right_mask;
>>> +
>>> + /*
>>> + * Simply do SIMD16 based dispatch, so every thread uses
>>> + * SIMD16 channels.
>>> + *
>>> + * Define our own thread group size, e.g 16x1 for every group, then
>>> + * will have 1 thread each group in SIMD16 dispatch. So thread
>>> + * width/height/depth are all 1.
>>> + *
>>> + * Then thread group X = width / 16 (aligned to 16)
>>> + * thread group Y = height;
>>> + */
>>> + x_dim = (width + 15) / 16;
>>> + y_dim = height;
>>> +
>>> + tmp = width & 15;
>>> + if (tmp == 0)
>>> + right_mask = (1 << 16) - 1;
>>> + else
>>> + right_mask = (1 << tmp) - 1;
>>> +
>>> + OUT_BATCH(GEN7_GPGPU_WALKER | 13);
>>> +
>>> + OUT_BATCH(0); /* kernel offset */
>>> + OUT_BATCH(0); /* indirect data length */
>>> + OUT_BATCH(0); /* indirect data offset */
>>> +
>>> + /* SIMD size, thread w/h/d */
>>> + OUT_BATCH(1 << 30 | /* SIMD16 */
>>> + 0 << 16 | /* depth:1 */
>>> + 0 << 8 | /* height:1 */
>>> + 0); /* width:1 */
>>> +
>>> + /* thread group X */
>>> + OUT_BATCH(0);
>>> + OUT_BATCH(0);
>>> + OUT_BATCH(x_dim);
>>> +
>>> + /* thread group Y */
>>> + OUT_BATCH(0);
>>> + OUT_BATCH(0);
>>> + OUT_BATCH(y_dim);
>>> +
>>> + /* thread group Z */
>>> + OUT_BATCH(0);
>>> + OUT_BATCH(1);
>>> +
>>> + /* right mask */
>>> + OUT_BATCH(right_mask);
>>> +
>>> + /* bottom mask, height 1, always 0xffffffff */
>>> + OUT_BATCH(0xffffffff);
>>> +}
>>> diff --git a/lib/gpu_fill.h b/lib/gpu_fill.h
>>> new file mode 100644
>>> index 00000000..a271ce6e
>>> --- /dev/null
>>> +++ b/lib/gpu_fill.h
>>> @@ -0,0 +1,141 @@
>>> +/*
>>> + * Copyright © 2018 Intel Corporation
>>> + *
>>> + * Permission is hereby granted, free of charge, to any person obtaining a
>>> + * copy of this software and associated documentation files (the
>>> "Software"),
>>> + * to deal in the Software without restriction, including without
>>> limitation
>>> + * the rights to use, copy, modify, merge, publish, distribute, sublicense,
>>> + * and/or sell copies of the Software, and to permit persons to whom the
>>> + * Software is furnished to do so, subject to the following conditions:
>>> + *
>>> + * The above copyright notice and this permission notice (including the
>>> next
>>> + * paragraph) shall be included in all copies or substantial portions of
>>> the
>>> + * Software.
>>> + *
>>> + * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS
>>> OR
>>> + * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
>>> + * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
>>> + * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR
>>> OTHER
>>> + * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
>>> + * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
>>> DEALINGS
>>> + * IN THE SOFTWARE.
>>> + *
>>> + */
>>> +
>>> +#ifndef GPU_FILL_H
>>> +#define GPU_FILL_H
>>> +
>>> +uint32_t
>>> +batch_used(struct intel_batchbuffer *batch);
>>> +
>>> +uint32_t
>>> +batch_align(struct intel_batchbuffer *batch, uint32_t align);
>>> +
>>> +void *
>>> +batch_alloc(struct intel_batchbuffer *batch, uint32_t size, uint32_t
>>> align);
>>> +
>>> +uint32_t
>>> +batch_offset(struct intel_batchbuffer *batch, void *ptr);
>>> +
>>> +uint32_t
>>> +batch_copy(struct intel_batchbuffer *batch, const void *ptr, uint32_t size,
>>> uint32_t align);
>>> +
>>> +void
>>> +gen7_render_flush(struct intel_batchbuffer *batch, uint32_t batch_end);
>>> +
>>> +uint32_t
>>> +gen7_fill_curbe_buffer_data(struct intel_batchbuffer *batch,
>>> + uint8_t color);
>>> +
>>> +uint32_t
>>> +gen7_fill_surface_state(struct intel_batchbuffer *batch,
>>> + struct igt_buf *buf,
>>> + uint32_t format,
>>> + int is_dst);
>>> +
>>> +uint32_t
>>> +gen7_fill_binding_table(struct intel_batchbuffer *batch,
>>> + struct igt_buf *dst);
>>> +uint32_t
>>> +gen7_fill_kernel(struct intel_batchbuffer *batch,
>>> + const uint32_t kernel[][4],
>>> + size_t size);
>>> +
>>> +uint32_t
>>> +gen7_fill_interface_descriptor(struct intel_batchbuffer *batch, struct
>>> igt_buf *dst,
>>> + const uint32_t kernel[][4], size_t size);
>>> +
>>> +void
>>> +gen7_emit_state_base_address(struct intel_batchbuffer *batch);
>>> +
>>> +void
>>> +gen7_emit_vfe_state(struct intel_batchbuffer *batch);
>>> +
>>> +void
>>> +gen7_emit_vfe_state_gpgpu(struct intel_batchbuffer *batch);
>>> +
>>> +void
>>> +gen7_emit_curbe_load(struct intel_batchbuffer *batch, uint32_t
>>> curbe_buffer);
>>> +
>>> +void
>>> +gen7_emit_interface_descriptor_load(struct intel_batchbuffer *batch,
>>> uint32_t interface_descriptor);
>>> +
>>> +void
>>> +gen7_emit_media_objects(struct intel_batchbuffer *batch,
>>> + unsigned x, unsigned y,
>>> + unsigned width, unsigned height);
>>> +
>>> +uint32_t
>>> +gen8_fill_surface_state(struct intel_batchbuffer *batch,
>>> + struct igt_buf *buf,
>>> + uint32_t format,
>>> + int is_dst);
>>> +
>>> +uint32_t
>>> +gen8_fill_binding_table(struct intel_batchbuffer *batch,
>>> + struct igt_buf *dst);
>>> +
>>> +uint32_t
>>> +gen8_fill_interface_descriptor(struct intel_batchbuffer *batch, struct
>>> igt_buf *dst, const uint32_t kernel[][4], size_t size);
>>> +
>>> +void
>>> +gen8_emit_state_base_address(struct intel_batchbuffer *batch);
>>> +
>>> +void
>>> +gen8_emit_vfe_state(struct intel_batchbuffer *batch);
>>> +
>>> +void
>>> +gen8_emit_vfe_state_gpgpu(struct intel_batchbuffer *batch);
>>> +
>>> +void
>>> +gen8_emit_curbe_load(struct intel_batchbuffer *batch, uint32_t
>>> curbe_buffer);
>>> +
>>> +void
>>> +gen8_emit_interface_descriptor_load(struct intel_batchbuffer *batch,
>>> uint32_t interface_descriptor);
>>> +
>>> +void
>>> +gen8_emit_media_state_flush(struct intel_batchbuffer *batch);
>>> +
>>> +void
>>> +gen8_emit_media_objects(struct intel_batchbuffer *batch,
>>> + unsigned x, unsigned y,
>>> + unsigned width, unsigned height);
>>> +
>>> +void
>>> +gen8lp_emit_media_objects(struct intel_batchbuffer *batch,
>>> + unsigned x, unsigned y,
>>> + unsigned width, unsigned height);
>>> +
>>> +void
>>> +gen9_emit_state_base_address(struct intel_batchbuffer *batch);
>>> +
>>> +void
>>> +gen7_emit_gpgpu_walk(struct intel_batchbuffer *batch,
>>> + unsigned x, unsigned y,
>>> + unsigned width, unsigned height);
>>> +void
>>> +gen8_emit_gpgpu_walk(struct intel_batchbuffer *batch,
>>> + unsigned x, unsigned y,
>>> + unsigned width, unsigned height);
>>> +
>>> +#endif /* GPU_FILL_H */
>>> diff --git a/lib/media_fill_gen7.c b/lib/media_fill_gen7.c
>>> index 6fb44798..c97555a6 100644
>>> --- a/lib/media_fill_gen7.c
>>> +++ b/lib/media_fill_gen7.c
>>> @@ -5,7 +5,7 @@
>>> #include "gen7_media.h"
>>> #include "intel_reg.h"
>>> #include "drmtest.h"
>>> -
>>> +#include "gpu_fill.h"
>>> #include <assert.h>
>>>
>>> static const uint32_t media_kernel[][4] = {
>>> @@ -22,275 +22,6 @@ static const uint32_t media_kernel[][4] = {
>>> { 0x07800031, 0x20001ca8, 0x00000e00, 0x82000010 },
>>> };
>>>
>>> -static uint32_t
>>> -batch_used(struct intel_batchbuffer *batch)
>>> -{
>>> - return batch->ptr - batch->buffer;
>>> -}
>>> -
>>> -static uint32_t
>>> -batch_align(struct intel_batchbuffer *batch, uint32_t align)
>>> -{
>>> - uint32_t offset = batch_used(batch);
>>> - offset = ALIGN(offset, align);
>>> - batch->ptr = batch->buffer + offset;
>>> - return offset;
>>> -}
>>> -
>>> -static void *
>>> -batch_alloc(struct intel_batchbuffer *batch, uint32_t size, uint32_t align)
>>> -{
>>> - uint32_t offset = batch_align(batch, align);
>>> - batch->ptr += size;
>>> - return memset(batch->buffer + offset, 0, size);
>>> -}
>>> -
>>> -static uint32_t
>>> -batch_offset(struct intel_batchbuffer *batch, void *ptr)
>>> -{
>>> - return (uint8_t *)ptr - batch->buffer;
>>> -}
>>> -
>>> -static uint32_t
>>> -batch_copy(struct intel_batchbuffer *batch, const void *ptr, uint32_t size,
>>> uint32_t align)
>>> -{
>>> - return batch_offset(batch, memcpy(batch_alloc(batch, size, align),
>>> ptr, size));
>>> -}
>>> -
>>> -static void
>>> -gen7_render_flush(struct intel_batchbuffer *batch, uint32_t batch_end)
>>> -{
>>> - int ret;
>>> -
>>> - ret = drm_intel_bo_subdata(batch->bo, 0, 4096, batch->buffer);
>>> - if (ret == 0)
>>> - ret = drm_intel_bo_mrb_exec(batch->bo, batch_end,
>>> - NULL, 0, 0, 0);
>>> - igt_assert(ret == 0);
>>> -}
>>> -
>>> -static uint32_t
>>> -gen7_fill_curbe_buffer_data(struct intel_batchbuffer *batch,
>>> - uint8_t color)
>>> -{
>>> - uint8_t *curbe_buffer;
>>> - uint32_t offset;
>>> -
>>> - curbe_buffer = batch_alloc(batch, sizeof(uint32_t) * 8, 64);
>>> - offset = batch_offset(batch, curbe_buffer);
>>> - *curbe_buffer = color;
>>> -
>>> - return offset;
>>> -}
>>> -
>>> -static uint32_t
>>> -gen7_fill_surface_state(struct intel_batchbuffer *batch,
>>> - struct igt_buf *buf,
>>> - uint32_t format,
>>> - int is_dst)
>>> -{
>>> - struct gen7_surface_state *ss;
>>> - uint32_t write_domain, read_domain, offset;
>>> - int ret;
>>> -
>>> - if (is_dst) {
>>> - write_domain = read_domain = I915_GEM_DOMAIN_RENDER;
>>> - } else {
>>> - write_domain = 0;
>>> - read_domain = I915_GEM_DOMAIN_SAMPLER;
>>> - }
>>> -
>>> - ss = batch_alloc(batch, sizeof(*ss), 64);
>>> - offset = batch_offset(batch, ss);
>>> -
>>> - ss->ss0.surface_type = GEN7_SURFACE_2D;
>>> - ss->ss0.surface_format = format;
>>> - ss->ss0.render_cache_read_write = 1;
>>> -
>>> - if (buf->tiling == I915_TILING_X)
>>> - ss->ss0.tiled_mode = 2;
>>> - else if (buf->tiling == I915_TILING_Y)
>>> - ss->ss0.tiled_mode = 3;
>>> -
>>> - ss->ss1.base_addr = buf->bo->offset;
>>> - ret = drm_intel_bo_emit_reloc(batch->bo,
>>> - batch_offset(batch, ss) + 4,
>>> - buf->bo, 0,
>>> - read_domain, write_domain);
>>> - igt_assert(ret == 0);
>>> -
>>> - ss->ss2.height = igt_buf_height(buf) - 1;
>>> - ss->ss2.width = igt_buf_width(buf) - 1;
>>> -
>>> - ss->ss3.pitch = buf->stride - 1;
>>> -
>>> - ss->ss7.shader_chanel_select_r = 4;
>>> - ss->ss7.shader_chanel_select_g = 5;
>>> - ss->ss7.shader_chanel_select_b = 6;
>>> - ss->ss7.shader_chanel_select_a = 7;
>>> -
>>> - return offset;
>>> -}
>>> -
>>> -static uint32_t
>>> -gen7_fill_binding_table(struct intel_batchbuffer *batch,
>>> - struct igt_buf *dst)
>>> -{
>>> - uint32_t *binding_table, offset;
>>> -
>>> - binding_table = batch_alloc(batch, 32, 64);
>>> - offset = batch_offset(batch, binding_table);
>>> -
>>> - binding_table[0] = gen7_fill_surface_state(batch, dst,
>>> GEN7_SURFACEFORMAT_R8_UNORM, 1);
>>> -
>>> - return offset;
>>> -}
>>> -
>>> -static uint32_t
>>> -gen7_fill_media_kernel(struct intel_batchbuffer *batch,
>>> - const uint32_t kernel[][4],
>>> - size_t size)
>>> -{
>>> - uint32_t offset;
>>> -
>>> - offset = batch_copy(batch, kernel, size, 64);
>>> -
>>> - return offset;
>>> -}
>>> -
>>> -static uint32_t
>>> -gen7_fill_interface_descriptor(struct intel_batchbuffer *batch, struct
>>> igt_buf *dst,
>>> - const uint32_t kernel[][4], size_t size)
>>> -{
>>> - struct gen7_interface_descriptor_data *idd;
>>> - uint32_t offset;
>>> - uint32_t binding_table_offset, kernel_offset;
>>> -
>>> - binding_table_offset = gen7_fill_binding_table(batch, dst);
>>> - kernel_offset = gen7_fill_media_kernel(batch, kernel, size);
>>> -
>>> - idd = batch_alloc(batch, sizeof(*idd), 64);
>>> - offset = batch_offset(batch, idd);
>>> -
>>> - idd->desc0.kernel_start_pointer = (kernel_offset >> 6);
>>> -
>>> - idd->desc1.single_program_flow = 1;
>>> - idd->desc1.floating_point_mode = GEN7_FLOATING_POINT_IEEE_754;
>>> -
>>> - idd->desc2.sampler_count = 0; /* 0 samplers used */
>>> - idd->desc2.sampler_state_pointer = 0;
>>> -
>>> - idd->desc3.binding_table_entry_count = 0;
>>> - idd->desc3.binding_table_pointer = (binding_table_offset >> 5);
>>> -
>>> - idd->desc4.constant_urb_entry_read_offset = 0;
>>> - idd->desc4.constant_urb_entry_read_length = 1; /* grf 1 */
>>> -
>>> - return offset;
>>> -}
>>> -
>>> -static void
>>> -gen7_emit_state_base_address(struct intel_batchbuffer *batch)
>>> -{
>>> - OUT_BATCH(GEN7_STATE_BASE_ADDRESS | (10 - 2));
>>> -
>>> - /* general */
>>> - OUT_BATCH(0);
>>> -
>>> - /* surface */
>>> - OUT_RELOC(batch->bo, I915_GEM_DOMAIN_INSTRUCTION, 0,
>>> BASE_ADDRESS_MODIFY);
>>> -
>>> - /* dynamic */
>>> - OUT_RELOC(batch->bo, I915_GEM_DOMAIN_INSTRUCTION, 0,
>>> BASE_ADDRESS_MODIFY);
>>> -
>>> - /* indirect */
>>> - OUT_BATCH(0);
>>> -
>>> - /* instruction */
>>> - OUT_RELOC(batch->bo, I915_GEM_DOMAIN_INSTRUCTION, 0,
>>> BASE_ADDRESS_MODIFY);
>>> -
>>> - /* general/dynamic/indirect/instruction access Bound */
>>> - OUT_BATCH(0);
>>> - OUT_BATCH(0 | BASE_ADDRESS_MODIFY);
>>> - OUT_BATCH(0);
>>> - OUT_BATCH(0 | BASE_ADDRESS_MODIFY);
>>> -}
>>> -
>>> -static void
>>> -gen7_emit_vfe_state(struct intel_batchbuffer *batch)
>>> -{
>>> - OUT_BATCH(GEN7_MEDIA_VFE_STATE | (8 - 2));
>>> -
>>> - /* scratch buffer */
>>> - OUT_BATCH(0);
>>> -
>>> - /* number of threads & urb entries */
>>> - OUT_BATCH(1 << 16 |
>>> - 2 << 8);
>>> -
>>> - OUT_BATCH(0);
>>> -
>>> - /* urb entry size & curbe size */
>>> - OUT_BATCH(2 << 16 | /* in 256 bits unit */
>>> - 2); /* in 256 bits unit */
>>> -
>>> - /* scoreboard */
>>> - OUT_BATCH(0);
>>> - OUT_BATCH(0);
>>> - OUT_BATCH(0);
>>> -}
>>> -
>>> -static void
>>> -gen7_emit_curbe_load(struct intel_batchbuffer *batch, uint32_t
>>> curbe_buffer)
>>> -{
>>> - OUT_BATCH(GEN7_MEDIA_CURBE_LOAD | (4 - 2));
>>> - OUT_BATCH(0);
>>> - /* curbe total data length */
>>> - OUT_BATCH(64);
>>> - /* curbe data start address, is relative to the dynamics base
>>> address */
>>> - OUT_BATCH(curbe_buffer);
>>> -}
>>> -
>>> -static void
>>> -gen7_emit_interface_descriptor_load(struct intel_batchbuffer *batch,
>>> uint32_t interface_descriptor)
>>> -{
>>> - OUT_BATCH(GEN7_MEDIA_INTERFACE_DESCRIPTOR_LOAD | (4 - 2));
>>> - OUT_BATCH(0);
>>> - /* interface descriptor data length */
>>> - OUT_BATCH(sizeof(struct gen7_interface_descriptor_data));
>>> - /* interface descriptor address, is relative to the dynamics base
>>> address */
>>> - OUT_BATCH(interface_descriptor);
>>> -}
>>> -
>>> -static void
>>> -gen7_emit_media_objects(struct intel_batchbuffer *batch,
>>> - unsigned x, unsigned y,
>>> - unsigned width, unsigned height)
>>> -{
>>> - int i, j;
>>> -
>>> - for (i = 0; i < width / 16; i++) {
>>> - for (j = 0; j < height / 16; j++) {
>>> - OUT_BATCH(GEN7_MEDIA_OBJECT | (8 - 2));
>>> -
>>> - /* interface descriptor offset */
>>> - OUT_BATCH(0);
>>> -
>>> - /* without indirect data */
>>> - OUT_BATCH(0);
>>> - OUT_BATCH(0);
>>> -
>>> - /* scoreboard */
>>> - OUT_BATCH(0);
>>> - OUT_BATCH(0);
>>> -
>>> - /* inline data (xoffset, yoffset) */
>>> - OUT_BATCH(x + i * 16);
>>> - OUT_BATCH(y + j * 16);
>>> - }
>>> - }
>>> -}
>>> -
>>> /*
>>> * This sets up the media pipeline,
>>> *
>>> diff --git a/lib/media_fill_gen8.c b/lib/media_fill_gen8.c
>>> index 4a8fe5a2..88c7dbdd 100644
>>> --- a/lib/media_fill_gen8.c
>>> +++ b/lib/media_fill_gen8.c
>>> @@ -5,7 +5,7 @@
>>> #include "gen8_media.h"
>>> #include "intel_reg.h"
>>> #include "drmtest.h"
>>> -
>>> +#include "gpu_fill.h"
>>> #include <assert.h>
>>>
>>>
>>> @@ -23,294 +23,6 @@ static const uint32_t media_kernel[][4] = {
>>> { 0x07800031, 0x20000a40, 0x0e000e00, 0x82000010 },
>>> };
>>>
>>> -static uint32_t
>>> -batch_used(struct intel_batchbuffer *batch)
>>> -{
>>> - return batch->ptr - batch->buffer;
>>> -}
>>> -
>>> -static uint32_t
>>> -batch_align(struct intel_batchbuffer *batch, uint32_t align)
>>> -{
>>> - uint32_t offset = batch_used(batch);
>>> - offset = ALIGN(offset, align);
>>> - batch->ptr = batch->buffer + offset;
>>> - return offset;
>>> -}
>>> -
>>> -static void *
>>> -batch_alloc(struct intel_batchbuffer *batch, uint32_t size, uint32_t align)
>>> -{
>>> - uint32_t offset = batch_align(batch, align);
>>> - batch->ptr += size;
>>> - return memset(batch->buffer + offset, 0, size);
>>> -}
>>> -
>>> -static uint32_t
>>> -batch_offset(struct intel_batchbuffer *batch, void *ptr)
>>> -{
>>> - return (uint8_t *)ptr - batch->buffer;
>>> -}
>>> -
>>> -static uint32_t
>>> -batch_copy(struct intel_batchbuffer *batch, const void *ptr, uint32_t size,
>>> uint32_t align)
>>> -{
>>> - return batch_offset(batch, memcpy(batch_alloc(batch, size, align),
>>> ptr, size));
>>> -}
>>> -
>>> -static void
>>> -gen8_render_flush(struct intel_batchbuffer *batch, uint32_t batch_end)
>>> -{
>>> - int ret;
>>> -
>>> - ret = drm_intel_bo_subdata(batch->bo, 0, 4096, batch->buffer);
>>> - if (ret == 0)
>>> - ret = drm_intel_bo_mrb_exec(batch->bo, batch_end,
>>> - NULL, 0, 0, 0);
>>> - igt_assert(ret == 0);
>>> -}
>>> -
>>> -static uint32_t
>>> -gen8_fill_curbe_buffer_data(struct intel_batchbuffer *batch,
>>> - uint8_t color)
>>> -{
>>> - uint8_t *curbe_buffer;
>>> - uint32_t offset;
>>> -
>>> - curbe_buffer = batch_alloc(batch, sizeof(uint32_t) * 8, 64);
>>> - offset = batch_offset(batch, curbe_buffer);
>>> - *curbe_buffer = color;
>>> -
>>> - return offset;
>>> -}
>>> -
>>> -static uint32_t
>>> -gen8_fill_surface_state(struct intel_batchbuffer *batch,
>>> - struct igt_buf *buf,
>>> - uint32_t format,
>>> - int is_dst)
>>> -{
>>> - struct gen8_surface_state *ss;
>>> - uint32_t write_domain, read_domain, offset;
>>> - int ret;
>>> -
>>> - if (is_dst) {
>>> - write_domain = read_domain = I915_GEM_DOMAIN_RENDER;
>>> - } else {
>>> - write_domain = 0;
>>> - read_domain = I915_GEM_DOMAIN_SAMPLER;
>>> - }
>>> -
>>> - ss = batch_alloc(batch, sizeof(*ss), 64);
>>> - offset = batch_offset(batch, ss);
>>> -
>>> - ss->ss0.surface_type = GEN8_SURFACE_2D;
>>> - ss->ss0.surface_format = format;
>>> - ss->ss0.render_cache_read_write = 1;
>>> - ss->ss0.vertical_alignment = 1; /* align 4 */
>>> - ss->ss0.horizontal_alignment = 1; /* align 4 */
>>> -
>>> - if (buf->tiling == I915_TILING_X)
>>> - ss->ss0.tiled_mode = 2;
>>> - else if (buf->tiling == I915_TILING_Y)
>>> - ss->ss0.tiled_mode = 3;
>>> -
>>> - ss->ss8.base_addr = buf->bo->offset;
>>> -
>>> - ret = drm_intel_bo_emit_reloc(batch->bo,
>>> - batch_offset(batch, ss) + 8 * 4,
>>> - buf->bo, 0,
>>> - read_domain, write_domain);
>>> - igt_assert(ret == 0);
>>> -
>>> - ss->ss2.height = igt_buf_height(buf) - 1;
>>> - ss->ss2.width = igt_buf_width(buf) - 1;
>>> - ss->ss3.pitch = buf->stride - 1;
>>> -
>>> - ss->ss7.shader_chanel_select_r = 4;
>>> - ss->ss7.shader_chanel_select_g = 5;
>>> - ss->ss7.shader_chanel_select_b = 6;
>>> - ss->ss7.shader_chanel_select_a = 7;
>>> -
>>> - return offset;
>>> -}
>>> -
>>> -static uint32_t
>>> -gen8_fill_binding_table(struct intel_batchbuffer *batch,
>>> - struct igt_buf *dst)
>>> -{
>>> - uint32_t *binding_table, offset;
>>> -
>>> - binding_table = batch_alloc(batch, 32, 64);
>>> - offset = batch_offset(batch, binding_table);
>>> -
>>> - binding_table[0] = gen8_fill_surface_state(batch, dst,
>>> GEN8_SURFACEFORMAT_R8_UNORM, 1);
>>> -
>>> - return offset;
>>> -}
>>> -
>>> -static uint32_t
>>> -gen8_fill_media_kernel(struct intel_batchbuffer *batch,
>>> - const uint32_t kernel[][4],
>>> - size_t size)
>>> -{
>>> - uint32_t offset;
>>> -
>>> - offset = batch_copy(batch, kernel, size, 64);
>>> -
>>> - return offset;
>>> -}
>>> -
>>> -static uint32_t
>>> -gen8_fill_interface_descriptor(struct intel_batchbuffer *batch, struct
>>> igt_buf *dst)
>>> -{
>>> - struct gen8_interface_descriptor_data *idd;
>>> - uint32_t offset;
>>> - uint32_t binding_table_offset, kernel_offset;
>>> -
>>> - binding_table_offset = gen8_fill_binding_table(batch, dst);
>>> - kernel_offset = gen8_fill_media_kernel(batch, media_kernel,
>>> sizeof(media_kernel));
>>> -
>>> - idd = batch_alloc(batch, sizeof(*idd), 64);
>>> - offset = batch_offset(batch, idd);
>>> -
>>> - idd->desc0.kernel_start_pointer = (kernel_offset >> 6);
>>> -
>>> - idd->desc2.single_program_flow = 1;
>>> - idd->desc2.floating_point_mode = GEN8_FLOATING_POINT_IEEE_754;
>>> -
>>> - idd->desc3.sampler_count = 0; /* 0 samplers used */
>>> - idd->desc3.sampler_state_pointer = 0;
>>> -
>>> - idd->desc4.binding_table_entry_count = 0;
>>> - idd->desc4.binding_table_pointer = (binding_table_offset >> 5);
>>> -
>>> - idd->desc5.constant_urb_entry_read_offset = 0;
>>> - idd->desc5.constant_urb_entry_read_length = 1; /* grf 1 */
>>> -
>>> - return offset;
>>> -}
>>> -
>>> -static void
>>> -gen8_emit_state_base_address(struct intel_batchbuffer *batch)
>>> -{
>>> - OUT_BATCH(GEN8_STATE_BASE_ADDRESS | (16 - 2));
>>> -
>>> - /* general */
>>> - OUT_BATCH(0 | BASE_ADDRESS_MODIFY);
>>> - OUT_BATCH(0);
>>> -
>>> - /* stateless data port */
>>> - OUT_BATCH(0 | BASE_ADDRESS_MODIFY);
>>> -
>>> - /* surface */
>>> - OUT_RELOC(batch->bo, I915_GEM_DOMAIN_SAMPLER, 0,
>>> BASE_ADDRESS_MODIFY);
>>> -
>>> - /* dynamic */
>>> - OUT_RELOC(batch->bo, I915_GEM_DOMAIN_RENDER |
>>> I915_GEM_DOMAIN_INSTRUCTION,
>>> - 0, BASE_ADDRESS_MODIFY);
>>> -
>>> - /* indirect */
>>> - OUT_BATCH(0);
>>> - OUT_BATCH(0);
>>> -
>>> - /* instruction */
>>> - OUT_RELOC(batch->bo, I915_GEM_DOMAIN_INSTRUCTION, 0,
>>> BASE_ADDRESS_MODIFY);
>>> -
>>> - /* general state buffer size */
>>> - OUT_BATCH(0xfffff000 | 1);
>>> - /* dynamic state buffer size */
>>> - OUT_BATCH(1 << 12 | 1);
>>> - /* indirect object buffer size */
>>> - OUT_BATCH(0xfffff000 | 1);
>>> - /* intruction buffer size, must set modify enable bit, otherwise it
>>> may result in GPU hang */
>>> - OUT_BATCH(1 << 12 | 1);
>>> -}
>>> -
>>> -static void
>>> -gen8_emit_vfe_state(struct intel_batchbuffer *batch)
>>> -{
>>> - OUT_BATCH(GEN8_MEDIA_VFE_STATE | (9 - 2));
>>> -
>>> - /* scratch buffer */
>>> - OUT_BATCH(0);
>>> - OUT_BATCH(0);
>>> -
>>> - /* number of threads & urb entries */
>>> - OUT_BATCH(1 << 16 |
>>> - 2 << 8);
>>> -
>>> - OUT_BATCH(0);
>>> -
>>> - /* urb entry size & curbe size */
>>> - OUT_BATCH(2 << 16 |
>>> - 2);
>>> -
>>> - /* scoreboard */
>>> - OUT_BATCH(0);
>>> - OUT_BATCH(0);
>>> - OUT_BATCH(0);
>>> -}
>>> -
>>> -static void
>>> -gen8_emit_curbe_load(struct intel_batchbuffer *batch, uint32_t
>>> curbe_buffer)
>>> -{
>>> - OUT_BATCH(GEN8_MEDIA_CURBE_LOAD | (4 - 2));
>>> - OUT_BATCH(0);
>>> - /* curbe total data length */
>>> - OUT_BATCH(64);
>>> - /* curbe data start address, is relative to the dynamics base
>>> address */
>>> - OUT_BATCH(curbe_buffer);
>>> -}
>>> -
>>> -static void
>>> -gen8_emit_interface_descriptor_load(struct intel_batchbuffer *batch,
>>> uint32_t interface_descriptor)
>>> -{
>>> - OUT_BATCH(GEN8_MEDIA_INTERFACE_DESCRIPTOR_LOAD | (4 - 2));
>>> - OUT_BATCH(0);
>>> - /* interface descriptor data length */
>>> - OUT_BATCH(sizeof(struct gen8_interface_descriptor_data));
>>> - /* interface descriptor address, is relative to the dynamics base
>>> address */
>>> - OUT_BATCH(interface_descriptor);
>>> -}
>>> -
>>> -static void
>>> -gen8_emit_media_state_flush(struct intel_batchbuffer *batch)
>>> -{
>>> - OUT_BATCH(GEN8_MEDIA_STATE_FLUSH | (2 - 2));
>>> - OUT_BATCH(0);
>>> -}
>>> -
>>> -static void
>>> -gen8_emit_media_objects(struct intel_batchbuffer *batch,
>>> - unsigned x, unsigned y,
>>> - unsigned width, unsigned height)
>>> -{
>>> - int i, j;
>>> -
>>> - for (i = 0; i < width / 16; i++) {
>>> - for (j = 0; j < height / 16; j++) {
>>> - OUT_BATCH(GEN8_MEDIA_OBJECT | (8 - 2));
>>> -
>>> - /* interface descriptor offset */
>>> - OUT_BATCH(0);
>>> -
>>> - /* without indirect data */
>>> - OUT_BATCH(0);
>>> - OUT_BATCH(0);
>>> -
>>> - /* scoreboard */
>>> - OUT_BATCH(0);
>>> - OUT_BATCH(0);
>>> -
>>> - /* inline data (xoffset, yoffset) */
>>> - OUT_BATCH(x + i * 16);
>>> - OUT_BATCH(y + j * 16);
>>> - gen8_emit_media_state_flush(batch);
>>> - }
>>> - }
>>> -}
>>> -
>>> /*
>>> * This sets up the media pipeline,
>>> *
>>> @@ -348,8 +60,9 @@ gen8_media_fillfunc(struct intel_batchbuffer *batch,
>>> /* setup states */
>>> batch->ptr = &batch->buffer[BATCH_STATE_SPLIT];
>>>
>>> - curbe_buffer = gen8_fill_curbe_buffer_data(batch, color);
>>> - interface_descriptor = gen8_fill_interface_descriptor(batch, dst);
>>> + curbe_buffer = gen7_fill_curbe_buffer_data(batch, color);
>>> + interface_descriptor = gen8_fill_interface_descriptor(batch, dst,
>>> + media_kernel,
>>> sizeof(media_kernel));
>>> igt_assert(batch->ptr < &batch->buffer[4095]);
>>>
>>> /* media pipeline */
>>> @@ -370,6 +83,6 @@ gen8_media_fillfunc(struct intel_batchbuffer *batch,
>>> batch_end = batch_align(batch, 8);
>>> igt_assert(batch_end < BATCH_STATE_SPLIT);
>>>
>>> - gen8_render_flush(batch, batch_end);
>>> + gen7_render_flush(batch, batch_end);
>>> intel_batchbuffer_reset(batch);
>>> }
>>> diff --git a/lib/media_fill_gen8lp.c b/lib/media_fill_gen8lp.c
>>> index 1f8a4adc..b30d96a3 100644
>>> --- a/lib/media_fill_gen8lp.c
>>> +++ b/lib/media_fill_gen8lp.c
>>> @@ -5,7 +5,7 @@
>>> #include "gen8_media.h"
>>> #include "intel_reg.h"
>>> #include "drmtest.h"
>>> -
>>> +#include "gpu_fill.h"
>>> #include <assert.h>
>>>
>>>
>>> @@ -23,286 +23,6 @@ static const uint32_t media_kernel[][4] = {
>>> { 0x07800031, 0x20000a40, 0x0e000e00, 0x82000010 },
>>> };
>>>
>>> -static uint32_t
>>> -batch_used(struct intel_batchbuffer *batch)
>>> -{
>>> - return batch->ptr - batch->buffer;
>>> -}
>>> -
>>> -static uint32_t
>>> -batch_align(struct intel_batchbuffer *batch, uint32_t align)
>>> -{
>>> - uint32_t offset = batch_used(batch);
>>> - offset = ALIGN(offset, align);
>>> - batch->ptr = batch->buffer + offset;
>>> - return offset;
>>> -}
>>> -
>>> -static void *
>>> -batch_alloc(struct intel_batchbuffer *batch, uint32_t size, uint32_t align)
>>> -{
>>> - uint32_t offset = batch_align(batch, align);
>>> - batch->ptr += size;
>>> - return memset(batch->buffer + offset, 0, size);
>>> -}
>>> -
>>> -static uint32_t
>>> -batch_offset(struct intel_batchbuffer *batch, void *ptr)
>>> -{
>>> - return (uint8_t *)ptr - batch->buffer;
>>> -}
>>> -
>>> -static uint32_t
>>> -batch_copy(struct intel_batchbuffer *batch, const void *ptr, uint32_t size,
>>> uint32_t align)
>>> -{
>>> - return batch_offset(batch, memcpy(batch_alloc(batch, size, align),
>>> ptr, size));
>>> -}
>>> -
>>> -static void
>>> -gen8_render_flush(struct intel_batchbuffer *batch, uint32_t batch_end)
>>> -{
>>> - int ret;
>>> -
>>> - ret = drm_intel_bo_subdata(batch->bo, 0, 4096, batch->buffer);
>>> - if (ret == 0)
>>> - ret = drm_intel_bo_mrb_exec(batch->bo, batch_end,
>>> - NULL, 0, 0, 0);
>>> - igt_assert(ret == 0);
>>> -}
>>> -
>>> -static uint32_t
>>> -gen8_fill_curbe_buffer_data(struct intel_batchbuffer *batch,
>>> - uint8_t color)
>>> -{
>>> - uint8_t *curbe_buffer;
>>> - uint32_t offset;
>>> -
>>> - curbe_buffer = batch_alloc(batch, sizeof(uint32_t) * 8, 64);
>>> - offset = batch_offset(batch, curbe_buffer);
>>> - *curbe_buffer = color;
>>> -
>>> - return offset;
>>> -}
>>> -
>>> -static uint32_t
>>> -gen8_fill_surface_state(struct intel_batchbuffer *batch,
>>> - struct igt_buf *buf,
>>> - uint32_t format,
>>> - int is_dst)
>>> -{
>>> - struct gen8_surface_state *ss;
>>> - uint32_t write_domain, read_domain, offset;
>>> - int ret;
>>> -
>>> - if (is_dst) {
>>> - write_domain = read_domain = I915_GEM_DOMAIN_RENDER;
>>> - } else {
>>> - write_domain = 0;
>>> - read_domain = I915_GEM_DOMAIN_SAMPLER;
>>> - }
>>> -
>>> - ss = batch_alloc(batch, sizeof(*ss), 64);
>>> - offset = batch_offset(batch, ss);
>>> -
>>> - ss->ss0.surface_type = GEN8_SURFACE_2D;
>>> - ss->ss0.surface_format = format;
>>> - ss->ss0.render_cache_read_write = 1;
>>> - ss->ss0.vertical_alignment = 1; /* align 4 */
>>> - ss->ss0.horizontal_alignment = 1; /* align 4 */
>>> -
>>> - if (buf->tiling == I915_TILING_X)
>>> - ss->ss0.tiled_mode = 2;
>>> - else if (buf->tiling == I915_TILING_Y)
>>> - ss->ss0.tiled_mode = 3;
>>> -
>>> - ss->ss8.base_addr = buf->bo->offset;
>>> -
>>> - ret = drm_intel_bo_emit_reloc(batch->bo,
>>> - batch_offset(batch, ss) + 8 * 4,
>>> - buf->bo, 0,
>>> - read_domain, write_domain);
>>> - igt_assert(ret == 0);
>>> -
>>> - ss->ss2.height = igt_buf_height(buf) - 1;
>>> - ss->ss2.width = igt_buf_width(buf) - 1;
>>> - ss->ss3.pitch = buf->stride - 1;
>>> -
>>> - ss->ss7.shader_chanel_select_r = 4;
>>> - ss->ss7.shader_chanel_select_g = 5;
>>> - ss->ss7.shader_chanel_select_b = 6;
>>> - ss->ss7.shader_chanel_select_a = 7;
>>> -
>>> - return offset;
>>> -}
>>> -
>>> -static uint32_t
>>> -gen8_fill_binding_table(struct intel_batchbuffer *batch,
>>> - struct igt_buf *dst)
>>> -{
>>> - uint32_t *binding_table, offset;
>>> -
>>> - binding_table = batch_alloc(batch, 32, 64);
>>> - offset = batch_offset(batch, binding_table);
>>> -
>>> - binding_table[0] = gen8_fill_surface_state(batch, dst,
>>> GEN8_SURFACEFORMAT_R8_UNORM, 1);
>>> -
>>> - return offset;
>>> -}
>>> -
>>> -static uint32_t
>>> -gen8_fill_media_kernel(struct intel_batchbuffer *batch,
>>> - const uint32_t kernel[][4],
>>> - size_t size)
>>> -{
>>> - uint32_t offset;
>>> -
>>> - offset = batch_copy(batch, kernel, size, 64);
>>> -
>>> - return offset;
>>> -}
>>> -
>>> -static uint32_t
>>> -gen8_fill_interface_descriptor(struct intel_batchbuffer *batch, struct
>>> igt_buf *dst)
>>> -{
>>> - struct gen8_interface_descriptor_data *idd;
>>> - uint32_t offset;
>>> - uint32_t binding_table_offset, kernel_offset;
>>> -
>>> - binding_table_offset = gen8_fill_binding_table(batch, dst);
>>> - kernel_offset = gen8_fill_media_kernel(batch, media_kernel,
>>> sizeof(media_kernel));
>>> -
>>> - idd = batch_alloc(batch, sizeof(*idd), 64);
>>> - offset = batch_offset(batch, idd);
>>> -
>>> - idd->desc0.kernel_start_pointer = (kernel_offset >> 6);
>>> -
>>> - idd->desc2.single_program_flow = 1;
>>> - idd->desc2.floating_point_mode = GEN8_FLOATING_POINT_IEEE_754;
>>> -
>>> - idd->desc3.sampler_count = 0; /* 0 samplers used */
>>> - idd->desc3.sampler_state_pointer = 0;
>>> -
>>> - idd->desc4.binding_table_entry_count = 0;
>>> - idd->desc4.binding_table_pointer = (binding_table_offset >> 5);
>>> -
>>> - idd->desc5.constant_urb_entry_read_offset = 0;
>>> - idd->desc5.constant_urb_entry_read_length = 1; /* grf 1 */
>>> -
>>> - return offset;
>>> -}
>>> -
>>> -static void
>>> -gen8_emit_state_base_address(struct intel_batchbuffer *batch)
>>> -{
>>> - OUT_BATCH(GEN8_STATE_BASE_ADDRESS | (16 - 2));
>>> -
>>> - /* general */
>>> - OUT_BATCH(0 | BASE_ADDRESS_MODIFY);
>>> - OUT_BATCH(0);
>>> -
>>> - /* stateless data port */
>>> - OUT_BATCH(0 | BASE_ADDRESS_MODIFY);
>>> -
>>> - /* surface */
>>> - OUT_RELOC(batch->bo, I915_GEM_DOMAIN_SAMPLER, 0,
>>> BASE_ADDRESS_MODIFY);
>>> -
>>> - /* dynamic */
>>> - OUT_RELOC(batch->bo, I915_GEM_DOMAIN_RENDER |
>>> I915_GEM_DOMAIN_INSTRUCTION,
>>> - 0, BASE_ADDRESS_MODIFY);
>>> -
>>> - /* indirect */
>>> - OUT_BATCH(0);
>>> - OUT_BATCH(0);
>>> -
>>> - /* instruction */
>>> - OUT_RELOC(batch->bo, I915_GEM_DOMAIN_INSTRUCTION, 0,
>>> BASE_ADDRESS_MODIFY);
>>> -
>>> - /* general state buffer size */
>>> - OUT_BATCH(0xfffff000 | 1);
>>> - /* dynamic state buffer size */
>>> - OUT_BATCH(1 << 12 | 1);
>>> - /* indirect object buffer size */
>>> - OUT_BATCH(0xfffff000 | 1);
>>> - /* intruction buffer size, must set modify enable bit, otherwise it
>>> may result in GPU hang */
>>> - OUT_BATCH(1 << 12 | 1);
>>> -}
>>> -
>>> -static void
>>> -gen8_emit_vfe_state(struct intel_batchbuffer *batch)
>>> -{
>>> - OUT_BATCH(GEN8_MEDIA_VFE_STATE | (9 - 2));
>>> -
>>> - /* scratch buffer */
>>> - OUT_BATCH(0);
>>> - OUT_BATCH(0);
>>> -
>>> - /* number of threads & urb entries */
>>> - OUT_BATCH(1 << 16 |
>>> - 2 << 8);
>>> -
>>> - OUT_BATCH(0);
>>> -
>>> - /* urb entry size & curbe size */
>>> - OUT_BATCH(2 << 16 |
>>> - 2);
>>> -
>>> - /* scoreboard */
>>> - OUT_BATCH(0);
>>> - OUT_BATCH(0);
>>> - OUT_BATCH(0);
>>> -}
>>> -
>>> -static void
>>> -gen8_emit_curbe_load(struct intel_batchbuffer *batch, uint32_t
>>> curbe_buffer)
>>> -{
>>> - OUT_BATCH(GEN8_MEDIA_CURBE_LOAD | (4 - 2));
>>> - OUT_BATCH(0);
>>> - /* curbe total data length */
>>> - OUT_BATCH(64);
>>> - /* curbe data start address, is relative to the dynamics base
>>> address */
>>> - OUT_BATCH(curbe_buffer);
>>> -}
>>> -
>>> -static void
>>> -gen8_emit_interface_descriptor_load(struct intel_batchbuffer *batch,
>>> uint32_t interface_descriptor)
>>> -{
>>> - OUT_BATCH(GEN8_MEDIA_INTERFACE_DESCRIPTOR_LOAD | (4 - 2));
>>> - OUT_BATCH(0);
>>> - /* interface descriptor data length */
>>> - OUT_BATCH(sizeof(struct gen8_interface_descriptor_data));
>>> - /* interface descriptor address, is relative to the dynamics base
>>> address */
>>> - OUT_BATCH(interface_descriptor);
>>> -}
>>> -
>>> -static void
>>> -gen8lp_emit_media_objects(struct intel_batchbuffer *batch,
>>> - unsigned x, unsigned y,
>>> - unsigned width, unsigned height)
>>> -{
>>> - int i, j;
>>> -
>>> - for (i = 0; i < width / 16; i++) {
>>> - for (j = 0; j < height / 16; j++) {
>>> - OUT_BATCH(GEN8_MEDIA_OBJECT | (8 - 2));
>>> -
>>> - /* interface descriptor offset */
>>> - OUT_BATCH(0);
>>> -
>>> - /* without indirect data */
>>> - OUT_BATCH(0);
>>> - OUT_BATCH(0);
>>> -
>>> - /* scoreboard */
>>> - OUT_BATCH(0);
>>> - OUT_BATCH(0);
>>> -
>>> - /* inline data (xoffset, yoffset) */
>>> - OUT_BATCH(x + i * 16);
>>> - OUT_BATCH(y + j * 16);
>>> - }
>>> - }
>>> -}
>>> -
>>> /*
>>> * This sets up the media pipeline,
>>> *
>>> @@ -340,8 +60,9 @@ gen8lp_media_fillfunc(struct intel_batchbuffer *batch,
>>> /* setup states */
>>> batch->ptr = &batch->buffer[BATCH_STATE_SPLIT];
>>>
>>> - curbe_buffer = gen8_fill_curbe_buffer_data(batch, color);
>>> - interface_descriptor = gen8_fill_interface_descriptor(batch, dst);
>>> + curbe_buffer = gen7_fill_curbe_buffer_data(batch, color);
>>> + interface_descriptor = gen8_fill_interface_descriptor(batch, dst,
>>> + media_kernel,
>>> sizeof(media_kernel));
>>> igt_assert(batch->ptr < &batch->buffer[4095]);
>>>
>>> /* media pipeline */
>>> @@ -362,6 +83,6 @@ gen8lp_media_fillfunc(struct intel_batchbuffer *batch,
>>> batch_end = batch_align(batch, 8);
>>> igt_assert(batch_end < BATCH_STATE_SPLIT);
>>>
>>> - gen8_render_flush(batch, batch_end);
>>> + gen7_render_flush(batch, batch_end);
>>> intel_batchbuffer_reset(batch);
>>> }
>>> diff --git a/lib/media_fill_gen9.c b/lib/media_fill_gen9.c
>>> index 3fd21819..e5d94487 100644
>>> --- a/lib/media_fill_gen9.c
>>> +++ b/lib/media_fill_gen9.c
>>> @@ -4,11 +4,9 @@
>>> #include "media_fill.h"
>>> #include "gen8_media.h"
>>> #include "intel_reg.h"
>>> -
>>> +#include "gpu_fill.h"
>>> #include <assert.h>
>>>
>>> -#define ALIGN(x, y) (((x) + (y)-1) & ~((y)-1))
>>> -
>>> static const uint32_t media_kernel[][4] = {
>>> { 0x00400001, 0x20202288, 0x00000020, 0x00000000 },
>>> { 0x00600001, 0x20800208, 0x008d0000, 0x00000000 },
>>> @@ -23,299 +21,6 @@ static const uint32_t media_kernel[][4] = {
>>> { 0x07800031, 0x20000a40, 0x0e000e00, 0x82000010 },
>>> };
>>>
>>> -static uint32_t
>>> -batch_used(struct intel_batchbuffer *batch)
>>> -{
>>> - return batch->ptr - batch->buffer;
>>> -}
>>> -
>>> -static uint32_t
>>> -batch_align(struct intel_batchbuffer *batch, uint32_t align)
>>> -{
>>> - uint32_t offset = batch_used(batch);
>>> - offset = ALIGN(offset, align);
>>> - batch->ptr = batch->buffer + offset;
>>> - return offset;
>>> -}
>>> -
>>> -static void *
>>> -batch_alloc(struct intel_batchbuffer *batch, uint32_t size, uint32_t align)
>>> -{
>>> - uint32_t offset = batch_align(batch, align);
>>> - batch->ptr += size;
>>> - return memset(batch->buffer + offset, 0, size);
>>> -}
>>> -
>>> -static uint32_t
>>> -batch_offset(struct intel_batchbuffer *batch, void *ptr)
>>> -{
>>> - return (uint8_t *)ptr - batch->buffer;
>>> -}
>>> -
>>> -static uint32_t
>>> -batch_copy(struct intel_batchbuffer *batch, const void *ptr, uint32_t size,
>>> uint32_t align)
>>> -{
>>> - return batch_offset(batch, memcpy(batch_alloc(batch, size, align),
>>> ptr, size));
>>> -}
>>> -
>>> -static void
>>> -gen8_render_flush(struct intel_batchbuffer *batch, uint32_t batch_end)
>>> -{
>>> - int ret;
>>> -
>>> - ret = drm_intel_bo_subdata(batch->bo, 0, 4096, batch->buffer);
>>> - if (ret == 0)
>>> - ret = drm_intel_bo_mrb_exec(batch->bo, batch_end,
>>> - NULL, 0, 0, 0);
>>> - assert(ret == 0);
>>> -}
>>> -
>>> -static uint32_t
>>> -gen8_fill_curbe_buffer_data(struct intel_batchbuffer *batch,
>>> - uint8_t color)
>>> -{
>>> - uint8_t *curbe_buffer;
>>> - uint32_t offset;
>>> -
>>> - curbe_buffer = batch_alloc(batch, sizeof(uint32_t) * 8, 64);
>>> - offset = batch_offset(batch, curbe_buffer);
>>> - *curbe_buffer = color;
>>> -
>>> - return offset;
>>> -}
>>> -
>>> -static uint32_t
>>> -gen8_fill_surface_state(struct intel_batchbuffer *batch,
>>> - struct igt_buf *buf,
>>> - uint32_t format,
>>> - int is_dst)
>>> -{
>>> - struct gen8_surface_state *ss;
>>> - uint32_t write_domain, read_domain, offset;
>>> - int ret;
>>> -
>>> - if (is_dst) {
>>> - write_domain = read_domain = I915_GEM_DOMAIN_RENDER;
>>> - } else {
>>> - write_domain = 0;
>>> - read_domain = I915_GEM_DOMAIN_SAMPLER;
>>> - }
>>> -
>>> - ss = batch_alloc(batch, sizeof(*ss), 64);
>>> - offset = batch_offset(batch, ss);
>>> -
>>> - ss->ss0.surface_type = GEN8_SURFACE_2D;
>>> - ss->ss0.surface_format = format;
>>> - ss->ss0.render_cache_read_write = 1;
>>> - ss->ss0.vertical_alignment = 1; /* align 4 */
>>> - ss->ss0.horizontal_alignment = 1; /* align 4 */
>>> -
>>> - if (buf->tiling == I915_TILING_X)
>>> - ss->ss0.tiled_mode = 2;
>>> - else if (buf->tiling == I915_TILING_Y)
>>> - ss->ss0.tiled_mode = 3;
>>> -
>>> - ss->ss8.base_addr = buf->bo->offset;
>>> -
>>> - ret = drm_intel_bo_emit_reloc(batch->bo,
>>> - batch_offset(batch, ss) + 8 * 4,
>>> - buf->bo, 0,
>>> - read_domain, write_domain);
>>> - assert(ret == 0);
>>> -
>>> - ss->ss2.height = igt_buf_height(buf) - 1;
>>> - ss->ss2.width = igt_buf_width(buf) - 1;
>>> - ss->ss3.pitch = buf->stride - 1;
>>> -
>>> - ss->ss7.shader_chanel_select_r = 4;
>>> - ss->ss7.shader_chanel_select_g = 5;
>>> - ss->ss7.shader_chanel_select_b = 6;
>>> - ss->ss7.shader_chanel_select_a = 7;
>>> -
>>> - return offset;
>>> -}
>>> -
>>> -static uint32_t
>>> -gen8_fill_binding_table(struct intel_batchbuffer *batch,
>>> - struct igt_buf *dst)
>>> -{
>>> - uint32_t *binding_table, offset;
>>> -
>>> - binding_table = batch_alloc(batch, 32, 64);
>>> - offset = batch_offset(batch, binding_table);
>>> -
>>> - binding_table[0] = gen8_fill_surface_state(batch, dst,
>>> GEN8_SURFACEFORMAT_R8_UNORM, 1);
>>> -
>>> - return offset;
>>> -}
>>> -
>>> -static uint32_t
>>> -gen8_fill_media_kernel(struct intel_batchbuffer *batch,
>>> - const uint32_t kernel[][4],
>>> - size_t size)
>>> -{
>>> - uint32_t offset;
>>> -
>>> - offset = batch_copy(batch, kernel, size, 64);
>>> -
>>> - return offset;
>>> -}
>>> -
>>> -static uint32_t
>>> -gen8_fill_interface_descriptor(struct intel_batchbuffer *batch, struct
>>> igt_buf *dst)
>>> -{
>>> - struct gen8_interface_descriptor_data *idd;
>>> - uint32_t offset;
>>> - uint32_t binding_table_offset, kernel_offset;
>>> -
>>> - binding_table_offset = gen8_fill_binding_table(batch, dst);
>>> - kernel_offset = gen8_fill_media_kernel(batch, media_kernel,
>>> sizeof(media_kernel));
>>> -
>>> - idd = batch_alloc(batch, sizeof(*idd), 64);
>>> - offset = batch_offset(batch, idd);
>>> -
>>> - idd->desc0.kernel_start_pointer = (kernel_offset >> 6);
>>> -
>>> - idd->desc2.single_program_flow = 1;
>>> - idd->desc2.floating_point_mode = GEN8_FLOATING_POINT_IEEE_754;
>>> -
>>> - idd->desc3.sampler_count = 0; /* 0 samplers used */
>>> - idd->desc3.sampler_state_pointer = 0;
>>> -
>>> - idd->desc4.binding_table_entry_count = 0;
>>> - idd->desc4.binding_table_pointer = (binding_table_offset >> 5);
>>> -
>>> - idd->desc5.constant_urb_entry_read_offset = 0;
>>> - idd->desc5.constant_urb_entry_read_length = 1; /* grf 1 */
>>> -
>>> - return offset;
>>> -}
>>> -
>>> -static void
>>> -gen9_emit_state_base_address(struct intel_batchbuffer *batch)
>>> -{
>>> - OUT_BATCH(GEN8_STATE_BASE_ADDRESS | (19 - 2));
>>> -
>>> - /* general */
>>> - OUT_BATCH(0 | BASE_ADDRESS_MODIFY);
>>> - OUT_BATCH(0);
>>> -
>>> - /* stateless data port */
>>> - OUT_BATCH(0 | BASE_ADDRESS_MODIFY);
>>> -
>>> - /* surface */
>>> - OUT_RELOC(batch->bo, I915_GEM_DOMAIN_SAMPLER, 0,
>>> BASE_ADDRESS_MODIFY);
>>> -
>>> - /* dynamic */
>>> - OUT_RELOC(batch->bo, I915_GEM_DOMAIN_RENDER |
>>> I915_GEM_DOMAIN_INSTRUCTION,
>>> - 0, BASE_ADDRESS_MODIFY);
>>> -
>>> - /* indirect */
>>> - OUT_BATCH(0);
>>> - OUT_BATCH(0);
>>> -
>>> - /* instruction */
>>> - OUT_RELOC(batch->bo, I915_GEM_DOMAIN_INSTRUCTION, 0,
>>> BASE_ADDRESS_MODIFY);
>>> -
>>> - /* general state buffer size */
>>> - OUT_BATCH(0xfffff000 | 1);
>>> - /* dynamic state buffer size */
>>> - OUT_BATCH(1 << 12 | 1);
>>> - /* indirect object buffer size */
>>> - OUT_BATCH(0xfffff000 | 1);
>>> - /* intruction buffer size, must set modify enable bit, otherwise it
>>> may result in GPU hang */
>>> - OUT_BATCH(1 << 12 | 1);
>>> -
>>> - /* Bindless surface state base address */
>>> - OUT_BATCH(0 | BASE_ADDRESS_MODIFY);
>>> - OUT_BATCH(0);
>>> - OUT_BATCH(0xfffff000);
>>> -}
>>> -
>>> -static void
>>> -gen8_emit_vfe_state(struct intel_batchbuffer *batch)
>>> -{
>>> - OUT_BATCH(GEN8_MEDIA_VFE_STATE | (9 - 2));
>>> -
>>> - /* scratch buffer */
>>> - OUT_BATCH(0);
>>> - OUT_BATCH(0);
>>> -
>>> - /* number of threads & urb entries */
>>> - OUT_BATCH(1 << 16 |
>>> - 2 << 8);
>>> -
>>> - OUT_BATCH(0);
>>> -
>>> - /* urb entry size & curbe size */
>>> - OUT_BATCH(2 << 16 |
>>> - 2);
>>> -
>>> - /* scoreboard */
>>> - OUT_BATCH(0);
>>> - OUT_BATCH(0);
>>> - OUT_BATCH(0);
>>> -}
>>> -
>>> -static void
>>> -gen8_emit_curbe_load(struct intel_batchbuffer *batch, uint32_t
>>> curbe_buffer)
>>> -{
>>> - OUT_BATCH(GEN8_MEDIA_CURBE_LOAD | (4 - 2));
>>> - OUT_BATCH(0);
>>> - /* curbe total data length */
>>> - OUT_BATCH(64);
>>> - /* curbe data start address, is relative to the dynamics base
>>> address */
>>> - OUT_BATCH(curbe_buffer);
>>> -}
>>> -
>>> -static void
>>> -gen8_emit_interface_descriptor_load(struct intel_batchbuffer *batch,
>>> uint32_t interface_descriptor)
>>> -{
>>> - OUT_BATCH(GEN8_MEDIA_INTERFACE_DESCRIPTOR_LOAD | (4 - 2));
>>> - OUT_BATCH(0);
>>> - /* interface descriptor data length */
>>> - OUT_BATCH(sizeof(struct gen8_interface_descriptor_data));
>>> - /* interface descriptor address, is relative to the dynamics base
>>> address */
>>> - OUT_BATCH(interface_descriptor);
>>> -}
>>> -
>>> -static void
>>> -gen8_emit_media_state_flush(struct intel_batchbuffer *batch)
>>> -{
>>> - OUT_BATCH(GEN8_MEDIA_STATE_FLUSH | (2 - 2));
>>> - OUT_BATCH(0);
>>> -}
>>> -
>>> -static void
>>> -gen8_emit_media_objects(struct intel_batchbuffer *batch,
>>> - unsigned x, unsigned y,
>>> - unsigned width, unsigned height)
>>> -{
>>> - int i, j;
>>> -
>>> - for (i = 0; i < width / 16; i++) {
>>> - for (j = 0; j < height / 16; j++) {
>>> - OUT_BATCH(GEN8_MEDIA_OBJECT | (8 - 2));
>>> -
>>> - /* interface descriptor offset */
>>> - OUT_BATCH(0);
>>> -
>>> - /* without indirect data */
>>> - OUT_BATCH(0);
>>> - OUT_BATCH(0);
>>> -
>>> - /* scoreboard */
>>> - OUT_BATCH(0);
>>> - OUT_BATCH(0);
>>> -
>>> - /* inline data (xoffset, yoffset) */
>>> - OUT_BATCH(x + i * 16);
>>> - OUT_BATCH(y + j * 16);
>>> - gen8_emit_media_state_flush(batch);
>>> - }
>>> - }
>>> -}
>>> -
>>> /*
>>> * This sets up the media pipeline,
>>> *
>>> @@ -353,8 +58,9 @@ gen9_media_fillfunc(struct intel_batchbuffer *batch,
>>> /* setup states */
>>> batch->ptr = &batch->buffer[BATCH_STATE_SPLIT];
>>>
>>> - curbe_buffer = gen8_fill_curbe_buffer_data(batch, color);
>>> - interface_descriptor = gen8_fill_interface_descriptor(batch, dst);
>>> + curbe_buffer = gen7_fill_curbe_buffer_data(batch, color);
>>> + interface_descriptor = gen8_fill_interface_descriptor(batch, dst,
>>> + media_kernel,
>>> sizeof(media_kernel));
>>> assert(batch->ptr < &batch->buffer[4095]);
>>>
>>> /* media pipeline */
>>> @@ -387,6 +93,6 @@ gen9_media_fillfunc(struct intel_batchbuffer *batch,
>>> batch_end = batch_align(batch, 8);
>>> assert(batch_end < BATCH_STATE_SPLIT);
>>>
>>> - gen8_render_flush(batch, batch_end);
>>> + gen7_render_flush(batch, batch_end);
>>> intel_batchbuffer_reset(batch);
>>> }
>>> diff --git a/lib/meson.build b/lib/meson.build
>>> index b3b8b14a..38c3f107 100644
>>> --- a/lib/meson.build
>>> +++ b/lib/meson.build
>>> @@ -24,6 +24,7 @@ lib_sources = [
>>> 'intel_os.c',
>>> 'intel_mmio.c',
>>> 'ioctl_wrappers.c',
>>> + 'gpu_fill.c',
>>> 'media_fill_gen7.c',
>>> 'media_fill_gen8.c',
>>> 'media_fill_gen8lp.c',
>>>
>>
>> _______________________________________________
>> igt-dev mailing list
>> igt-dev at lists.freedesktop.org
>> https://lists.freedesktop.org/mailman/listinfo/igt-dev
More information about the igt-dev
mailing list