[igt-dev] [PATCH i-g-t 1/3] lib: Move common gpgpu/media functions to gpu_fill library

Katarzyna Dec katarzyna.dec at intel.com
Wed Apr 4 13:25:27 UTC 2018


On Tue, Mar 27, 2018 at 04:52:34PM -0700, Daniele Ceraolo Spurio wrote:
> 
> 
> On 27/03/18 06:36, Katarzyna Dec wrote:
> > Gpgpu_fill and media_fill libraries are very similar and many
> > functions can be shared. I have created library gpu_fill with
> > all functions needed for implementing gpgpu_fill and media_fill
> > tests for all Gens. Duplicates, e.g. where only name was changed,
> > were removed. The earliest common function remained.
> > 
> > Signed-off-by: Katarzyna Dec <katarzyna.dec at intel.com>
> > Cc: Lukasz Kalamarz <lukasz.kalamarz at intel.com>
> > Cc: Daniele Ceraolo Spurio <daniele.ceraolospurio at intel.com>
> > ---
> >   lib/Makefile.sources    |   2 +
> >   lib/gpgpu_fill.c        | 567 +------------------------------------
> >   lib/gpu_fill.c          | 729 ++++++++++++++++++++++++++++++++++++++++++++++++
> >   lib/gpu_fill.h          | 141 ++++++++++
> >   lib/media_fill_gen7.c   | 271 +-----------------
> >   lib/media_fill_gen8.c   | 297 +-------------------
> >   lib/media_fill_gen8lp.c | 289 +------------------
> >   lib/media_fill_gen9.c   | 304 +-------------------
> >   lib/meson.build         |   1 +
> >   9 files changed, 890 insertions(+), 1711 deletions(-)
> >   create mode 100644 lib/gpu_fill.c
> >   create mode 100644 lib/gpu_fill.h
> > 
> 
> This patch has a very big diff and it isn't a 1:1 move of code so I suggest
> to break it up a little to make it easier to review. You could start with
> moving the batch_* functions, which btw probably should be moved to
> intel_batchbuffer.<h/c>. There are copies of those batch_* functions in
> other files as well (e.g rendercopy), so once you move them it'd be nice to
> move them from everywhere. Note that some of these functions seem to be
> similar to existing ones (e.g. batch_copy and intel_batchbuffer_data), so we
> can probably remove extra duplication by merging some of these.
That is a good point. Lukasz Kalamarz started to make changes like that and we
will include them into this series.
> 
> Also, I think we could remove some duplication before doing the move, which
> could make things more modular and reduce the diff of the move itself. For
> example, the only difference between media_fill_gen8.c and
> media_fill_gen8lp.c is that in the !lp case we emit an extra
> GEN8_MEDIA_STATE_FLUSH. Instead of duplicating the whole file, we can just
> add a !IS_CHERRYVIEW(batch->devid) check before emitting that instruction
> and eliminate gen8lp_media_fillfunc and relative file entirely.
I will try to do that.
> Similarly, the diff between the gen8 and gen9 file is only in the
> gen9_emit_state_base_address and gen9_media_fillfunc functions. The first
> has a very small difference compared to the gen8 function so we can probably
> merge the 2 using a IS_GEN() check to apply the diff, while the latter we
> can just copy in the gen8 file and thus get rid of the gen9 file as well.
> 
I will compare funtions for gen8 and gen9 and will see what is worth unifying.

Thanks for the feedback!
Kasia
> Daniele
> 
> > diff --git a/lib/Makefile.sources b/lib/Makefile.sources
> > index 3d37ef1d..690a1d35 100644
> > --- a/lib/Makefile.sources
> > +++ b/lib/Makefile.sources
> > @@ -55,6 +55,8 @@ lib_source_list =	 	\
> >   	intel_reg.h		\
> >   	ioctl_wrappers.c	\
> >   	ioctl_wrappers.h	\
> > +	gpu_fill.h		\
> > +	gpu_fill.c		\
> >   	media_fill.h            \
> >   	media_fill_gen7.c       \
> >   	media_fill_gen8.c       \
> > diff --git a/lib/gpgpu_fill.c b/lib/gpgpu_fill.c
> > index 4d98643d..68cbac5e 100644
> > --- a/lib/gpgpu_fill.c
> > +++ b/lib/gpgpu_fill.c
> > @@ -34,6 +34,7 @@
> >   #include "gen7_media.h"
> >   #include "gen8_media.h"
> >   #include "gpgpu_fill.h"
> > +#include "gpu_fill.h"
> >   /* shaders/gpgpu/gpgpu_fill.gxa */
> >   static const uint32_t gen7_gpgpu_kernel[][4] = {
> > @@ -75,572 +76,6 @@ static const uint32_t gen9_gpgpu_kernel[][4] = {
> >   	{ 0x07800031, 0x20000a40, 0x06000e00, 0x82000010 },
> >   };
> > -static uint32_t
> > -batch_used(struct intel_batchbuffer *batch)
> > -{
> > -	return batch->ptr - batch->buffer;
> > -}
> > -
> > -static uint32_t
> > -batch_align(struct intel_batchbuffer *batch, uint32_t align)
> > -{
> > -	uint32_t offset = batch_used(batch);
> > -	offset = ALIGN(offset, align);
> > -	batch->ptr = batch->buffer + offset;
> > -	return offset;
> > -}
> > -
> > -static void *
> > -batch_alloc(struct intel_batchbuffer *batch, uint32_t size, uint32_t align)
> > -{
> > -	uint32_t offset = batch_align(batch, align);
> > -	batch->ptr += size;
> > -	return memset(batch->buffer + offset, 0, size);
> > -}
> > -
> > -static uint32_t
> > -batch_offset(struct intel_batchbuffer *batch, void *ptr)
> > -{
> > -	return (uint8_t *)ptr - batch->buffer;
> > -}
> > -
> > -static uint32_t
> > -batch_copy(struct intel_batchbuffer *batch, const void *ptr, uint32_t size,
> > -	   uint32_t align)
> > -{
> > -	return batch_offset(batch, memcpy(batch_alloc(batch, size, align), ptr, size));
> > -}
> > -
> > -static void
> > -gen7_render_flush(struct intel_batchbuffer *batch, uint32_t batch_end)
> > -{
> > -	int ret;
> > -
> > -	ret = drm_intel_bo_subdata(batch->bo, 0, 4096, batch->buffer);
> > -	if (ret == 0)
> > -		ret = drm_intel_bo_mrb_exec(batch->bo, batch_end,
> > -					NULL, 0, 0, 0);
> > -	igt_assert(ret == 0);
> > -}
> > -
> > -static uint32_t
> > -gen7_fill_curbe_buffer_data(struct intel_batchbuffer *batch, uint8_t color)
> > -{
> > -	uint8_t *curbe_buffer;
> > -	uint32_t offset;
> > -
> > -	curbe_buffer = batch_alloc(batch, sizeof(uint32_t) * 8, 64);
> > -	offset = batch_offset(batch, curbe_buffer);
> > -	*curbe_buffer = color;
> > -
> > -	return offset;
> > -}
> > -
> > -static uint32_t
> > -gen7_fill_surface_state(struct intel_batchbuffer *batch,
> > -			struct igt_buf *buf,
> > -			uint32_t format,
> > -			int is_dst)
> > -{
> > -	struct gen7_surface_state *ss;
> > -	uint32_t write_domain, read_domain, offset;
> > -	int ret;
> > -
> > -	if (is_dst) {
> > -		write_domain = read_domain = I915_GEM_DOMAIN_RENDER;
> > -	} else {
> > -		write_domain = 0;
> > -		read_domain = I915_GEM_DOMAIN_SAMPLER;
> > -	}
> > -
> > -	ss = batch_alloc(batch, sizeof(*ss), 64);
> > -	offset = batch_offset(batch, ss);
> > -
> > -	ss->ss0.surface_type = GEN7_SURFACE_2D;
> > -	ss->ss0.surface_format = format;
> > -	ss->ss0.render_cache_read_write = 1;
> > -
> > -	if (buf->tiling == I915_TILING_X)
> > -		ss->ss0.tiled_mode = 2;
> > -	else if (buf->tiling == I915_TILING_Y)
> > -		ss->ss0.tiled_mode = 3;
> > -
> > -	ss->ss1.base_addr = buf->bo->offset;
> > -	ret = drm_intel_bo_emit_reloc(batch->bo,
> > -				batch_offset(batch, ss) + 4,
> > -				buf->bo, 0,
> > -				read_domain, write_domain);
> > -	igt_assert(ret == 0);
> > -
> > -	ss->ss2.height = igt_buf_height(buf) - 1;
> > -	ss->ss2.width  = igt_buf_width(buf) - 1;
> > -
> > -	ss->ss3.pitch  = buf->stride - 1;
> > -
> > -	ss->ss7.shader_chanel_select_r = 4;
> > -	ss->ss7.shader_chanel_select_g = 5;
> > -	ss->ss7.shader_chanel_select_b = 6;
> > -	ss->ss7.shader_chanel_select_a = 7;
> > -
> > -	return offset;
> > -}
> > -
> > -static uint32_t
> > -gen8_fill_surface_state(struct intel_batchbuffer *batch,
> > -			struct igt_buf *buf,
> > -			uint32_t format,
> > -			int is_dst)
> > -{
> > -	struct gen8_surface_state *ss;
> > -	uint32_t write_domain, read_domain, offset;
> > -	int ret;
> > -
> > -	if (is_dst) {
> > -		write_domain = read_domain = I915_GEM_DOMAIN_RENDER;
> > -	} else {
> > -		write_domain = 0;
> > -		read_domain = I915_GEM_DOMAIN_SAMPLER;
> > -	}
> > -
> > -	ss = batch_alloc(batch, sizeof(*ss), 64);
> > -	offset = batch_offset(batch, ss);
> > -
> > -	ss->ss0.surface_type = GEN8_SURFACE_2D;
> > -	ss->ss0.surface_format = format;
> > -	ss->ss0.render_cache_read_write = 1;
> > -	ss->ss0.vertical_alignment = 1; /* align 4 */
> > -	ss->ss0.horizontal_alignment = 1; /* align 4 */
> > -
> > -	if (buf->tiling == I915_TILING_X)
> > -		ss->ss0.tiled_mode = 2;
> > -	else if (buf->tiling == I915_TILING_Y)
> > -		ss->ss0.tiled_mode = 3;
> > -
> > -	ss->ss8.base_addr = buf->bo->offset;
> > -
> > -	ret = drm_intel_bo_emit_reloc(batch->bo,
> > -				batch_offset(batch, ss) + 8 * 4,
> > -				buf->bo, 0,
> > -				read_domain, write_domain);
> > -	igt_assert_eq(ret, 0);
> > -
> > -	ss->ss2.height = igt_buf_height(buf) - 1;
> > -	ss->ss2.width  = igt_buf_width(buf) - 1;
> > -	ss->ss3.pitch  = buf->stride - 1;
> > -
> > -	ss->ss7.shader_chanel_select_r = 4;
> > -	ss->ss7.shader_chanel_select_g = 5;
> > -	ss->ss7.shader_chanel_select_b = 6;
> > -	ss->ss7.shader_chanel_select_a = 7;
> > -
> > -	return offset;
> > -
> > -}
> > -
> > -static uint32_t
> > -gen7_fill_binding_table(struct intel_batchbuffer *batch,
> > -			struct igt_buf *dst)
> > -{
> > -	uint32_t *binding_table, offset;
> > -
> > -	binding_table = batch_alloc(batch, 32, 64);
> > -	offset = batch_offset(batch, binding_table);
> > -
> > -	binding_table[0] = gen7_fill_surface_state(batch, dst, GEN7_SURFACEFORMAT_R8_UNORM, 1);
> > -
> > -	return offset;
> > -}
> > -
> > -static uint32_t
> > -gen8_fill_binding_table(struct intel_batchbuffer *batch,
> > -			struct igt_buf *dst)
> > -{
> > -	uint32_t *binding_table, offset;
> > -
> > -	binding_table = batch_alloc(batch, 32, 64);
> > -	offset = batch_offset(batch, binding_table);
> > -
> > -	binding_table[0] = gen8_fill_surface_state(batch, dst, GEN8_SURFACEFORMAT_R8_UNORM, 1);
> > -
> > -	return offset;
> > -}
> > -
> > -static uint32_t
> > -gen7_fill_gpgpu_kernel(struct intel_batchbuffer *batch,
> > -		const uint32_t kernel[][4],
> > -		size_t size)
> > -{
> > -	uint32_t offset;
> > -
> > -	offset = batch_copy(batch, kernel, size, 64);
> > -
> > -	return offset;
> > -}
> > -
> > -static uint32_t
> > -gen7_fill_interface_descriptor(struct intel_batchbuffer *batch, struct igt_buf *dst,
> > -			       const uint32_t kernel[][4], size_t size)
> > -{
> > -	struct gen7_interface_descriptor_data *idd;
> > -	uint32_t offset;
> > -	uint32_t binding_table_offset, kernel_offset;
> > -
> > -	binding_table_offset = gen7_fill_binding_table(batch, dst);
> > -	kernel_offset = gen7_fill_gpgpu_kernel(batch, kernel, size);
> > -
> > -	idd = batch_alloc(batch, sizeof(*idd), 64);
> > -	offset = batch_offset(batch, idd);
> > -
> > -	idd->desc0.kernel_start_pointer = (kernel_offset >> 6);
> > -
> > -	idd->desc1.single_program_flow = 1;
> > -	idd->desc1.floating_point_mode = GEN7_FLOATING_POINT_IEEE_754;
> > -
> > -	idd->desc2.sampler_count = 0;      /* 0 samplers used */
> > -	idd->desc2.sampler_state_pointer = 0;
> > -
> > -	idd->desc3.binding_table_entry_count = 0;
> > -	idd->desc3.binding_table_pointer = (binding_table_offset >> 5);
> > -
> > -	idd->desc4.constant_urb_entry_read_offset = 0;
> > -	idd->desc4.constant_urb_entry_read_length = 1; /* grf 1 */
> > -
> > -	return offset;
> > -}
> > -
> > -static uint32_t
> > -gen8_fill_interface_descriptor(struct intel_batchbuffer *batch, struct igt_buf *dst,
> > -			       const uint32_t kernel[][4], size_t size)
> > -{
> > -	struct gen8_interface_descriptor_data *idd;
> > -	uint32_t offset;
> > -	uint32_t binding_table_offset, kernel_offset;
> > -
> > -	binding_table_offset = gen8_fill_binding_table(batch, dst);
> > -	kernel_offset = gen7_fill_gpgpu_kernel(batch, kernel, size);
> > -
> > -	idd = batch_alloc(batch, sizeof(*idd), 64);
> > -	offset = batch_offset(batch, idd);
> > -
> > -	idd->desc0.kernel_start_pointer = (kernel_offset >> 6);
> > -
> > -	idd->desc2.single_program_flow = 1;
> > -	idd->desc2.floating_point_mode = GEN8_FLOATING_POINT_IEEE_754;
> > -
> > -	idd->desc3.sampler_count = 0;      /* 0 samplers used */
> > -	idd->desc3.sampler_state_pointer = 0;
> > -
> > -	idd->desc4.binding_table_entry_count = 0;
> > -	idd->desc4.binding_table_pointer = (binding_table_offset >> 5);
> > -
> > -	idd->desc5.constant_urb_entry_read_offset = 0;
> > -	idd->desc5.constant_urb_entry_read_length = 1; /* grf 1 */
> > -
> > -	return offset;
> > -}
> > -
> > -static void
> > -gen7_emit_state_base_address(struct intel_batchbuffer *batch)
> > -{
> > -	OUT_BATCH(GEN7_STATE_BASE_ADDRESS | (10 - 2));
> > -
> > -	/* general */
> > -	OUT_BATCH(0);
> > -
> > -	/* surface */
> > -	OUT_RELOC(batch->bo, I915_GEM_DOMAIN_INSTRUCTION, 0, BASE_ADDRESS_MODIFY);
> > -
> > -	/* dynamic */
> > -	OUT_RELOC(batch->bo, I915_GEM_DOMAIN_INSTRUCTION, 0, BASE_ADDRESS_MODIFY);
> > -
> > -	/* indirect */
> > -	OUT_BATCH(0);
> > -
> > -	/* instruction */
> > -	OUT_RELOC(batch->bo, I915_GEM_DOMAIN_INSTRUCTION, 0, BASE_ADDRESS_MODIFY);
> > -
> > -	/* general/dynamic/indirect/instruction access Bound */
> > -	OUT_BATCH(0);
> > -	OUT_BATCH(0 | BASE_ADDRESS_MODIFY);
> > -	OUT_BATCH(0);
> > -	OUT_BATCH(0 | BASE_ADDRESS_MODIFY);
> > -}
> > -
> > -static void
> > -gen8_emit_state_base_address(struct intel_batchbuffer *batch)
> > -{
> > -	OUT_BATCH(GEN8_STATE_BASE_ADDRESS | (16 - 2));
> > -
> > -	/* general */
> > -	OUT_BATCH(0 | (0x78 << 4) | (0 << 1) |  BASE_ADDRESS_MODIFY);
> > -	OUT_BATCH(0);
> > -
> > -	/* stateless data port */
> > -	OUT_BATCH(0 | BASE_ADDRESS_MODIFY);
> > -
> > -	/* surface */
> > -	OUT_RELOC(batch->bo, I915_GEM_DOMAIN_SAMPLER, 0, BASE_ADDRESS_MODIFY);
> > -
> > -	/* dynamic */
> > -	OUT_RELOC(batch->bo, I915_GEM_DOMAIN_RENDER | I915_GEM_DOMAIN_INSTRUCTION,
> > -		  0, BASE_ADDRESS_MODIFY);
> > -
> > -	/* indirect */
> > -	OUT_BATCH(0);
> > -	OUT_BATCH(0 );
> > -
> > -	/* instruction */
> > -	OUT_RELOC(batch->bo, I915_GEM_DOMAIN_INSTRUCTION, 0, BASE_ADDRESS_MODIFY);
> > -
> > -	/* general state buffer size */
> > -	OUT_BATCH(0xfffff000 | 1);
> > -	/* dynamic state buffer size */
> > -	OUT_BATCH(1 << 12 | 1);
> > -	/* indirect object buffer size */
> > -	OUT_BATCH(0xfffff000 | 1);
> > -	/* intruction buffer size, must set modify enable bit, otherwise it may result in GPU hang */
> > -	OUT_BATCH(1 << 12 | 1);
> > -}
> > -
> > -static void
> > -gen9_emit_state_base_address(struct intel_batchbuffer *batch)
> > -{
> > -	OUT_BATCH(GEN8_STATE_BASE_ADDRESS | (19 - 2));
> > -
> > -	/* general */
> > -	OUT_BATCH(0 | BASE_ADDRESS_MODIFY);
> > -	OUT_BATCH(0);
> > -
> > -	/* stateless data port */
> > -	OUT_BATCH(0 | BASE_ADDRESS_MODIFY);
> > -
> > -	/* surface */
> > -	OUT_RELOC(batch->bo, I915_GEM_DOMAIN_SAMPLER, 0, BASE_ADDRESS_MODIFY);
> > -
> > -	/* dynamic */
> > -	OUT_RELOC(batch->bo, I915_GEM_DOMAIN_RENDER | I915_GEM_DOMAIN_INSTRUCTION,
> > -		0, BASE_ADDRESS_MODIFY);
> > -
> > -	/* indirect */
> > -	OUT_BATCH(0);
> > -	OUT_BATCH(0);
> > -
> > -	/* instruction */
> > -	OUT_RELOC(batch->bo, I915_GEM_DOMAIN_INSTRUCTION, 0, BASE_ADDRESS_MODIFY);
> > -
> > -	/* general state buffer size */
> > -	OUT_BATCH(0xfffff000 | 1);
> > -	/* dynamic state buffer size */
> > -	OUT_BATCH(1 << 12 | 1);
> > -	/* indirect object buffer size */
> > -	OUT_BATCH(0xfffff000 | 1);
> > -	/* intruction buffer size, must set modify enable bit, otherwise it may result in GPU hang */
> > -	OUT_BATCH(1 << 12 | 1);
> > -
> > -	/* Bindless surface state base address */
> > -	OUT_BATCH(0 | BASE_ADDRESS_MODIFY);
> > -	OUT_BATCH(0);
> > -	OUT_BATCH(0xfffff000);
> > -}
> > -
> > -static void
> > -gen7_emit_vfe_state_gpgpu(struct intel_batchbuffer *batch)
> > -{
> > -	OUT_BATCH(GEN7_MEDIA_VFE_STATE | (8 - 2));
> > -
> > -	/* scratch buffer */
> > -	OUT_BATCH(0);
> > -
> > -	/* number of threads & urb entries */
> > -	OUT_BATCH(1 << 16 | /* max num of threads */
> > -		  0 << 8 | /* num of URB entry */
> > -		  1 << 2); /* GPGPU mode */
> > -
> > -	OUT_BATCH(0);
> > -
> > -	/* urb entry size & curbe size */
> > -	OUT_BATCH(0 << 16 | 	/* URB entry size in 256 bits unit */
> > -		  1);		/* CURBE entry size in 256 bits unit */
> > -
> > -	/* scoreboard */
> > -	OUT_BATCH(0);
> > -	OUT_BATCH(0);
> > -	OUT_BATCH(0);
> > -}
> > -
> > -static void
> > -gen8_emit_vfe_state_gpgpu(struct intel_batchbuffer *batch)
> > -{
> > -	OUT_BATCH(GEN8_MEDIA_VFE_STATE | (9 - 2));
> > -
> > -	/* scratch buffer */
> > -	OUT_BATCH(0);
> > -	OUT_BATCH(0);
> > -
> > -	/* number of threads & urb entries */
> > -	OUT_BATCH(1 << 16 | 1 << 8);
> > -
> > -	OUT_BATCH(0);
> > -
> > -	/* urb entry size & curbe size */
> > -	OUT_BATCH(0 << 16 | 1);
> > -
> > -	/* scoreboard */
> > -	OUT_BATCH(0);
> > -	OUT_BATCH(0);
> > -	OUT_BATCH(0);
> > -}
> > -
> > -static void
> > -gen7_emit_curbe_load(struct intel_batchbuffer *batch, uint32_t curbe_buffer)
> > -{
> > -	OUT_BATCH(GEN7_MEDIA_CURBE_LOAD | (4 - 2));
> > -	OUT_BATCH(0);
> > -	/* curbe total data length */
> > -	OUT_BATCH(64);
> > -	/* curbe data start address, is relative to the dynamics base address */
> > -	OUT_BATCH(curbe_buffer);
> > -}
> > -
> > -static void
> > -gen7_emit_interface_descriptor_load(struct intel_batchbuffer *batch, uint32_t interface_descriptor)
> > -{
> > -	OUT_BATCH(GEN7_MEDIA_INTERFACE_DESCRIPTOR_LOAD | (4 - 2));
> > -	OUT_BATCH(0);
> > -	/* interface descriptor data length */
> > -	OUT_BATCH(sizeof(struct gen7_interface_descriptor_data));
> > -	/* interface descriptor address, is relative to the dynamics base address */
> > -	OUT_BATCH(interface_descriptor);
> > -}
> > -
> > -static void
> > -gen8_emit_interface_descriptor_load(struct intel_batchbuffer *batch, uint32_t interface_descriptor)
> > -{
> > -	OUT_BATCH(GEN8_MEDIA_INTERFACE_DESCRIPTOR_LOAD | (4 - 2));
> > -	OUT_BATCH(0);
> > -	/* interface descriptor data length */
> > -	OUT_BATCH(sizeof(struct gen8_interface_descriptor_data));
> > -	/* interface descriptor address, is relative to the dynamics base address */
> > -	OUT_BATCH(interface_descriptor);
> > -}
> > -
> > -static void
> > -gen7_emit_gpgpu_walk(struct intel_batchbuffer *batch,
> > -		     unsigned x, unsigned y,
> > -		     unsigned width, unsigned height)
> > -{
> > -	uint32_t x_dim, y_dim, tmp, right_mask;
> > -
> > -	/*
> > -	 * Simply do SIMD16 based dispatch, so every thread uses
> > -	 * SIMD16 channels.
> > -	 *
> > -	 * Define our own thread group size, e.g 16x1 for every group, then
> > -	 * will have 1 thread each group in SIMD16 dispatch. So thread
> > -	 * width/height/depth are all 1.
> > -	 *
> > -	 * Then thread group X = width / 16 (aligned to 16)
> > -	 * thread group Y = height;
> > -	 */
> > -	x_dim = (width + 15) / 16;
> > -	y_dim = height;
> > -
> > -	tmp = width & 15;
> > -	if (tmp == 0)
> > -		right_mask = (1 << 16) - 1;
> > -	else
> > -		right_mask = (1 << tmp) - 1;
> > -
> > -	OUT_BATCH(GEN7_GPGPU_WALKER | 9);
> > -
> > -	/* interface descriptor offset */
> > -	OUT_BATCH(0);
> > -
> > -	/* SIMD size, thread w/h/d */
> > -	OUT_BATCH(1 << 30 | /* SIMD16 */
> > -		  0 << 16 | /* depth:1 */
> > -		  0 << 8 | /* height:1 */
> > -		  0); /* width:1 */
> > -
> > -	/* thread group X */
> > -	OUT_BATCH(0);
> > -	OUT_BATCH(x_dim);
> > -
> > -	/* thread group Y */
> > -	OUT_BATCH(0);
> > -	OUT_BATCH(y_dim);
> > -
> > -	/* thread group Z */
> > -	OUT_BATCH(0);
> > -	OUT_BATCH(1);
> > -
> > -	/* right mask */
> > -	OUT_BATCH(right_mask);
> > -
> > -	/* bottom mask, height 1, always 0xffffffff */
> > -	OUT_BATCH(0xffffffff);
> > -}
> > -
> > -static void
> > -gen8_emit_gpgpu_walk(struct intel_batchbuffer *batch,
> > -		     unsigned x, unsigned y,
> > -		     unsigned width, unsigned height)
> > -{
> > -	uint32_t x_dim, y_dim, tmp, right_mask;
> > -
> > -	/*
> > -	 * Simply do SIMD16 based dispatch, so every thread uses
> > -	 * SIMD16 channels.
> > -	 *
> > -	 * Define our own thread group size, e.g 16x1 for every group, then
> > -	 * will have 1 thread each group in SIMD16 dispatch. So thread
> > -	 * width/height/depth are all 1.
> > -	 *
> > -	 * Then thread group X = width / 16 (aligned to 16)
> > -	 * thread group Y = height;
> > -	 */
> > -	x_dim = (width + 15) / 16;
> > -	y_dim = height;
> > -
> > -	tmp = width & 15;
> > -	if (tmp == 0)
> > -		right_mask = (1 << 16) - 1;
> > -	else
> > -		right_mask = (1 << tmp) - 1;
> > -
> > -	OUT_BATCH(GEN7_GPGPU_WALKER | 13);
> > -
> > -	OUT_BATCH(0); /* kernel offset */
> > -	OUT_BATCH(0); /* indirect data length */
> > -	OUT_BATCH(0); /* indirect data offset */
> > -
> > -	/* SIMD size, thread w/h/d */
> > -	OUT_BATCH(1 << 30 | /* SIMD16 */
> > -		  0 << 16 | /* depth:1 */
> > -		  0 << 8 | /* height:1 */
> > -		  0); /* width:1 */
> > -
> > -	/* thread group X */
> > -	OUT_BATCH(0);
> > -	OUT_BATCH(0);
> > -	OUT_BATCH(x_dim);
> > -
> > -	/* thread group Y */
> > -	OUT_BATCH(0);
> > -	OUT_BATCH(0);
> > -	OUT_BATCH(y_dim);
> > -
> > -	/* thread group Z */
> > -	OUT_BATCH(0);
> > -	OUT_BATCH(1);
> > -
> > -	/* right mask */
> > -	OUT_BATCH(right_mask);
> > -
> > -	/* bottom mask, height 1, always 0xffffffff */
> > -	OUT_BATCH(0xffffffff);
> > -}
> > -
> >   /*
> >    * This sets up the gpgpu pipeline,
> >    *
> > diff --git a/lib/gpu_fill.c b/lib/gpu_fill.c
> > new file mode 100644
> > index 00000000..f4563112
> > --- /dev/null
> > +++ b/lib/gpu_fill.c
> > @@ -0,0 +1,729 @@
> > +#include <intel_bufmgr.h>
> > +#include <i915_drm.h>
> > +
> > +#include "intel_reg.h"
> > +#include "drmtest.h"
> > +#include "gpu_fill.h"
> > +#include <assert.h>
> > +#include "gen7_media.h"
> > +#include "gen8_media.h"
> > +
> > +
> > +uint32_t
> > +batch_used(struct intel_batchbuffer *batch)
> > +{
> > +	return batch->ptr - batch->buffer;
> > +}
> > +
> > +uint32_t
> > +batch_align(struct intel_batchbuffer *batch, uint32_t align)
> > +{
> > +	uint32_t offset = batch_used(batch);
> > +	offset = ALIGN(offset, align);
> > +	batch->ptr = batch->buffer + offset;
> > +	return offset;
> > +}
> > +
> > +void *
> > +batch_alloc(struct intel_batchbuffer *batch, uint32_t size, uint32_t align)
> > +{
> > +	uint32_t offset = batch_align(batch, align);
> > +	batch->ptr += size;
> > +	return memset(batch->buffer + offset, 0, size);
> > +}
> > +
> > +uint32_t
> > +batch_offset(struct intel_batchbuffer *batch, void *ptr)
> > +{
> > +	return (uint8_t *)ptr - batch->buffer;
> > +}
> > +
> > +uint32_t
> > +batch_copy(struct intel_batchbuffer *batch, const void *ptr, uint32_t size, uint32_t align)
> > +{
> > +	return batch_offset(batch, memcpy(batch_alloc(batch, size, align), ptr, size));
> > +}
> > +
> > +void
> > +gen7_render_flush(struct intel_batchbuffer *batch, uint32_t batch_end)
> > +{
> > +	int ret;
> > +
> > +	ret = drm_intel_bo_subdata(batch->bo, 0, 4096, batch->buffer);
> > +	if (ret == 0)
> > +		ret = drm_intel_bo_mrb_exec(batch->bo, batch_end,
> > +					NULL, 0, 0, 0);
> > +	igt_assert(ret == 0);
> > +}
> > +
> > +uint32_t
> > +gen7_fill_curbe_buffer_data(struct intel_batchbuffer *batch,
> > +			uint8_t color)
> > +{
> > +	uint8_t *curbe_buffer;
> > +	uint32_t offset;
> > +
> > +	curbe_buffer = batch_alloc(batch, sizeof(uint32_t) * 8, 64);
> > +	offset = batch_offset(batch, curbe_buffer);
> > +	*curbe_buffer = color;
> > +
> > +	return offset;
> > +}
> > +
> > +uint32_t
> > +gen7_fill_surface_state(struct intel_batchbuffer *batch,
> > +			struct igt_buf *buf,
> > +			uint32_t format,
> > +			int is_dst)
> > +{
> > +	struct gen7_surface_state *ss;
> > +	uint32_t write_domain, read_domain, offset;
> > +	int ret;
> > +
> > +	if (is_dst) {
> > +		write_domain = read_domain = I915_GEM_DOMAIN_RENDER;
> > +	} else {
> > +		write_domain = 0;
> > +		read_domain = I915_GEM_DOMAIN_SAMPLER;
> > +	}
> > +
> > +	ss = batch_alloc(batch, sizeof(*ss), 64);
> > +	offset = batch_offset(batch, ss);
> > +
> > +	ss->ss0.surface_type = GEN7_SURFACE_2D;
> > +	ss->ss0.surface_format = format;
> > +	ss->ss0.render_cache_read_write = 1;
> > +
> > +	if (buf->tiling == I915_TILING_X)
> > +		ss->ss0.tiled_mode = 2;
> > +	else if (buf->tiling == I915_TILING_Y)
> > +		ss->ss0.tiled_mode = 3;
> > +
> > +	ss->ss1.base_addr = buf->bo->offset;
> > +	ret = drm_intel_bo_emit_reloc(batch->bo,
> > +				batch_offset(batch, ss) + 4,
> > +				buf->bo, 0,
> > +				read_domain, write_domain);
> > +	igt_assert(ret == 0);
> > +
> > +	ss->ss2.height = igt_buf_height(buf) - 1;
> > +	ss->ss2.width  = igt_buf_width(buf) - 1;
> > +
> > +	ss->ss3.pitch  = buf->stride - 1;
> > +
> > +	ss->ss7.shader_chanel_select_r = 4;
> > +	ss->ss7.shader_chanel_select_g = 5;
> > +	ss->ss7.shader_chanel_select_b = 6;
> > +	ss->ss7.shader_chanel_select_a = 7;
> > +
> > +	return offset;
> > +}
> > +
> > +uint32_t
> > +gen7_fill_binding_table(struct intel_batchbuffer *batch,
> > +			struct igt_buf *dst)
> > +{
> > +	uint32_t *binding_table, offset;
> > +
> > +	binding_table = batch_alloc(batch, 32, 64);
> > +	offset = batch_offset(batch, binding_table);
> > +
> > +	binding_table[0] = gen7_fill_surface_state(batch, dst, GEN7_SURFACEFORMAT_R8_UNORM, 1);
> > +
> > +	return offset;
> > +}
> > +
> > +uint32_t
> > +gen7_fill_kernel(struct intel_batchbuffer *batch,
> > +		const uint32_t kernel[][4],
> > +		size_t size)
> > +{
> > +	uint32_t offset;
> > +
> > +	offset = batch_copy(batch, kernel, size, 64);
> > +
> > +	return offset;
> > +}
> > +
> > +uint32_t
> > +gen7_fill_interface_descriptor(struct intel_batchbuffer *batch, struct igt_buf *dst,
> > +			       const uint32_t kernel[][4], size_t size)
> > +{
> > +	struct gen7_interface_descriptor_data *idd;
> > +	uint32_t offset;
> > +	uint32_t binding_table_offset, kernel_offset;
> > +
> > +	binding_table_offset = gen7_fill_binding_table(batch, dst);
> > +	kernel_offset = gen7_fill_kernel(batch, kernel, size);
> > +
> > +	idd = batch_alloc(batch, sizeof(*idd), 64);
> > +	offset = batch_offset(batch, idd);
> > +
> > +	idd->desc0.kernel_start_pointer = (kernel_offset >> 6);
> > +
> > +	idd->desc1.single_program_flow = 1;
> > +	idd->desc1.floating_point_mode = GEN7_FLOATING_POINT_IEEE_754;
> > +
> > +	idd->desc2.sampler_count = 0;      /* 0 samplers used */
> > +	idd->desc2.sampler_state_pointer = 0;
> > +
> > +	idd->desc3.binding_table_entry_count = 0;
> > +	idd->desc3.binding_table_pointer = (binding_table_offset >> 5);
> > +
> > +	idd->desc4.constant_urb_entry_read_offset = 0;
> > +	idd->desc4.constant_urb_entry_read_length = 1; /* grf 1 */
> > +
> > +	return offset;
> > +}
> > +
> > +void
> > +gen7_emit_state_base_address(struct intel_batchbuffer *batch)
> > +{
> > +	OUT_BATCH(GEN7_STATE_BASE_ADDRESS | (10 - 2));
> > +
> > +	/* general */
> > +	OUT_BATCH(0);
> > +
> > +	/* surface */
> > +	OUT_RELOC(batch->bo, I915_GEM_DOMAIN_INSTRUCTION, 0, BASE_ADDRESS_MODIFY);
> > +
> > +	/* dynamic */
> > +	OUT_RELOC(batch->bo, I915_GEM_DOMAIN_INSTRUCTION, 0, BASE_ADDRESS_MODIFY);
> > +
> > +	/* indirect */
> > +	OUT_BATCH(0);
> > +
> > +	/* instruction */
> > +	OUT_RELOC(batch->bo, I915_GEM_DOMAIN_INSTRUCTION, 0, BASE_ADDRESS_MODIFY);
> > +
> > +	/* general/dynamic/indirect/instruction access Bound */
> > +	OUT_BATCH(0);
> > +	OUT_BATCH(0 | BASE_ADDRESS_MODIFY);
> > +	OUT_BATCH(0);
> > +	OUT_BATCH(0 | BASE_ADDRESS_MODIFY);
> > +}
> > +
> > +void
> > +gen7_emit_vfe_state(struct intel_batchbuffer *batch)
> > +{
> > +	OUT_BATCH(GEN7_MEDIA_VFE_STATE | (8 - 2));
> > +
> > +	/* scratch buffer */
> > +	OUT_BATCH(0);
> > +
> > +	/* number of threads & urb entries */
> > +	OUT_BATCH(1 << 16 |
> > +		2 << 8);
> > +
> > +	OUT_BATCH(0);
> > +
> > +	/* urb entry size & curbe size */
> > +	OUT_BATCH(2 << 16 | 	/* in 256 bits unit */
> > +		2);		/* in 256 bits unit */
> > +
> > +	/* scoreboard */
> > +	OUT_BATCH(0);
> > +	OUT_BATCH(0);
> > +	OUT_BATCH(0);
> > +}
> > +
> > +void
> > +gen7_emit_vfe_state_gpgpu(struct intel_batchbuffer *batch)
> > +{
> > +	OUT_BATCH(GEN7_MEDIA_VFE_STATE | (8 - 2));
> > +
> > +	/* scratch buffer */
> > +	OUT_BATCH(0);
> > +
> > +	/* number of threads & urb entries */
> > +	OUT_BATCH(1 << 16 | /* max num of threads */
> > +		  0 << 8 | /* num of URB entry */
> > +		  1 << 2); /* GPGPU mode */
> > +
> > +	OUT_BATCH(0);
> > +
> > +	/* urb entry size & curbe size */
> > +	OUT_BATCH(0 << 16 | 	/* URB entry size in 256 bits unit */
> > +		  1);		/* CURBE entry size in 256 bits unit */
> > +
> > +	/* scoreboard */
> > +	OUT_BATCH(0);
> > +	OUT_BATCH(0);
> > +	OUT_BATCH(0);
> > +}
> > +
> > +void
> > +gen7_emit_curbe_load(struct intel_batchbuffer *batch, uint32_t curbe_buffer)
> > +{
> > +	OUT_BATCH(GEN7_MEDIA_CURBE_LOAD | (4 - 2));
> > +	OUT_BATCH(0);
> > +	/* curbe total data length */
> > +	OUT_BATCH(64);
> > +	/* curbe data start address, is relative to the dynamics base address */
> > +	OUT_BATCH(curbe_buffer);
> > +}
> > +
> > +void
> > +gen7_emit_interface_descriptor_load(struct intel_batchbuffer *batch, uint32_t interface_descriptor)
> > +{
> > +	OUT_BATCH(GEN7_MEDIA_INTERFACE_DESCRIPTOR_LOAD | (4 - 2));
> > +	OUT_BATCH(0);
> > +	/* interface descriptor data length */
> > +	OUT_BATCH(sizeof(struct gen7_interface_descriptor_data));
> > +	/* interface descriptor address, is relative to the dynamics base address */
> > +	OUT_BATCH(interface_descriptor);
> > +}
> > +
> > +void
> > +gen7_emit_media_objects(struct intel_batchbuffer *batch,
> > +			unsigned x, unsigned y,
> > +			unsigned width, unsigned height)
> > +{
> > +	int i, j;
> > +
> > +	for (i = 0; i < width / 16; i++) {
> > +		for (j = 0; j < height / 16; j++) {
> > +			OUT_BATCH(GEN7_MEDIA_OBJECT | (8 - 2));
> > +
> > +			/* interface descriptor offset */
> > +			OUT_BATCH(0);
> > +
> > +			/* without indirect data */
> > +			OUT_BATCH(0);
> > +			OUT_BATCH(0);
> > +
> > +			/* scoreboard */
> > +			OUT_BATCH(0);
> > +			OUT_BATCH(0);
> > +
> > +			/* inline data (xoffset, yoffset) */
> > +			OUT_BATCH(x + i * 16);
> > +			OUT_BATCH(y + j * 16);
> > +		}
> > +	}
> > +}
> > +
> > +uint32_t
> > +gen8_fill_surface_state(struct intel_batchbuffer *batch,
> > +			struct igt_buf *buf,
> > +			uint32_t format,
> > +			int is_dst)
> > +{
> > +	struct gen8_surface_state *ss;
> > +	uint32_t write_domain, read_domain, offset;
> > +	int ret;
> > +
> > +	if (is_dst) {
> > +		write_domain = read_domain = I915_GEM_DOMAIN_RENDER;
> > +	} else {
> > +		write_domain = 0;
> > +		read_domain = I915_GEM_DOMAIN_SAMPLER;
> > +	}
> > +
> > +	ss = batch_alloc(batch, sizeof(*ss), 64);
> > +	offset = batch_offset(batch, ss);
> > +
> > +	ss->ss0.surface_type = GEN8_SURFACE_2D;
> > +	ss->ss0.surface_format = format;
> > +	ss->ss0.render_cache_read_write = 1;
> > +	ss->ss0.vertical_alignment = 1; /* align 4 */
> > +	ss->ss0.horizontal_alignment = 1; /* align 4 */
> > +
> > +	if (buf->tiling == I915_TILING_X)
> > +		ss->ss0.tiled_mode = 2;
> > +	else if (buf->tiling == I915_TILING_Y)
> > +		ss->ss0.tiled_mode = 3;
> > +
> > +	ss->ss8.base_addr = buf->bo->offset;
> > +
> > +	ret = drm_intel_bo_emit_reloc(batch->bo,
> > +				batch_offset(batch, ss) + 8 * 4,
> > +				buf->bo, 0,
> > +				read_domain, write_domain);
> > +	igt_assert(ret == 0);
> > +
> > +	ss->ss2.height = igt_buf_height(buf) - 1;
> > +	ss->ss2.width  = igt_buf_width(buf) - 1;
> > +	ss->ss3.pitch  = buf->stride - 1;
> > +
> > +	ss->ss7.shader_chanel_select_r = 4;
> > +	ss->ss7.shader_chanel_select_g = 5;
> > +	ss->ss7.shader_chanel_select_b = 6;
> > +	ss->ss7.shader_chanel_select_a = 7;
> > +
> > +	return offset;
> > +}
> > +
> > +uint32_t
> > +gen8_fill_binding_table(struct intel_batchbuffer *batch,
> > +			struct igt_buf *dst)
> > +{
> > +	uint32_t *binding_table, offset;
> > +
> > +	binding_table = batch_alloc(batch, 32, 64);
> > +	offset = batch_offset(batch, binding_table);
> > +
> > +	binding_table[0] = gen8_fill_surface_state(batch, dst, GEN8_SURFACEFORMAT_R8_UNORM, 1);
> > +
> > +	return offset;
> > +}
> > +
> > +uint32_t
> > +gen8_fill_interface_descriptor(struct intel_batchbuffer *batch, struct igt_buf *dst, const uint32_t kernel[][4], size_t size)
> > +{
> > +	struct gen8_interface_descriptor_data *idd;
> > +	uint32_t offset;
> > +	uint32_t binding_table_offset, kernel_offset;
> > +
> > +	binding_table_offset = gen8_fill_binding_table(batch, dst);
> > +	kernel_offset = gen7_fill_kernel(batch, kernel, sizeof(kernel));
> > +
> > +	idd = batch_alloc(batch, sizeof(*idd), 64);
> > +	offset = batch_offset(batch, idd);
> > +
> > +	idd->desc0.kernel_start_pointer = (kernel_offset >> 6);
> > +
> > +	idd->desc2.single_program_flow = 1;
> > +	idd->desc2.floating_point_mode = GEN8_FLOATING_POINT_IEEE_754;
> > +
> > +	idd->desc3.sampler_count = 0;      /* 0 samplers used */
> > +	idd->desc3.sampler_state_pointer = 0;
> > +
> > +	idd->desc4.binding_table_entry_count = 0;
> > +	idd->desc4.binding_table_pointer = (binding_table_offset >> 5);
> > +
> > +	idd->desc5.constant_urb_entry_read_offset = 0;
> > +	idd->desc5.constant_urb_entry_read_length = 1; /* grf 1 */
> > +
> > +	return offset;
> > +}
> > +
> > +void
> > +gen8_emit_state_base_address(struct intel_batchbuffer *batch)
> > +{
> > +	OUT_BATCH(GEN8_STATE_BASE_ADDRESS | (16 - 2));
> > +
> > +	/* general */
> > +	OUT_BATCH(0 | BASE_ADDRESS_MODIFY);
> > +	OUT_BATCH(0);
> > +
> > +	/* stateless data port */
> > +	OUT_BATCH(0 | BASE_ADDRESS_MODIFY);
> > +
> > +	/* surface */
> > +	OUT_RELOC(batch->bo, I915_GEM_DOMAIN_SAMPLER, 0, BASE_ADDRESS_MODIFY);
> > +
> > +	/* dynamic */
> > +	OUT_RELOC(batch->bo, I915_GEM_DOMAIN_RENDER | I915_GEM_DOMAIN_INSTRUCTION,
> > +		0, BASE_ADDRESS_MODIFY);
> > +
> > +	/* indirect */
> > +	OUT_BATCH(0);
> > +	OUT_BATCH(0);
> > +
> > +	/* instruction */
> > +	OUT_RELOC(batch->bo, I915_GEM_DOMAIN_INSTRUCTION, 0, BASE_ADDRESS_MODIFY);
> > +
> > +	/* general state buffer size */
> > +	OUT_BATCH(0xfffff000 | 1);
> > +	/* dynamic state buffer size */
> > +	OUT_BATCH(1 << 12 | 1);
> > +	/* indirect object buffer size */
> > +	OUT_BATCH(0xfffff000 | 1);
> > +	/* intruction buffer size, must set modify enable bit, otherwise it may result in GPU hang */
> > +	OUT_BATCH(1 << 12 | 1);
> > +}
> > +
> > +void
> > +gen8_emit_vfe_state(struct intel_batchbuffer *batch)
> > +{
> > +	OUT_BATCH(GEN8_MEDIA_VFE_STATE | (9 - 2));
> > +
> > +	/* scratch buffer */
> > +	OUT_BATCH(0);
> > +	OUT_BATCH(0);
> > +
> > +	/* number of threads & urb entries */
> > +	OUT_BATCH(1 << 16 |
> > +		2 << 8);
> > +
> > +	OUT_BATCH(0);
> > +
> > +	/* urb entry size & curbe size */
> > +	OUT_BATCH(2 << 16 |
> > +		2);
> > +
> > +	/* scoreboard */
> > +	OUT_BATCH(0);
> > +	OUT_BATCH(0);
> > +	OUT_BATCH(0);
> > +}
> > +
> > +void
> > +gen8_emit_vfe_state_gpgpu(struct intel_batchbuffer *batch)
> > +{
> > +	OUT_BATCH(GEN8_MEDIA_VFE_STATE | (9 - 2));
> > +
> > +	/* scratch buffer */
> > +	OUT_BATCH(0);
> > +	OUT_BATCH(0);
> > +
> > +	/* number of threads & urb entries */
> > +	OUT_BATCH(1 << 16 | 1 << 8);
> > +
> > +	OUT_BATCH(0);
> > +
> > +	/* urb entry size & curbe size */
> > +	OUT_BATCH(0 << 16 | 1);
> > +
> > +	/* scoreboard */
> > +	OUT_BATCH(0);
> > +	OUT_BATCH(0);
> > +	OUT_BATCH(0);
> > +}
> > +
> > +void
> > +gen8_emit_curbe_load(struct intel_batchbuffer *batch, uint32_t curbe_buffer)
> > +{
> > +	OUT_BATCH(GEN8_MEDIA_CURBE_LOAD | (4 - 2));
> > +	OUT_BATCH(0);
> > +	/* curbe total data length */
> > +	OUT_BATCH(64);
> > +	/* curbe data start address, is relative to the dynamics base address */
> > +	OUT_BATCH(curbe_buffer);
> > +}
> > +
> > +void
> > +gen8_emit_interface_descriptor_load(struct intel_batchbuffer *batch, uint32_t interface_descriptor)
> > +{
> > +	OUT_BATCH(GEN8_MEDIA_INTERFACE_DESCRIPTOR_LOAD | (4 - 2));
> > +	OUT_BATCH(0);
> > +	/* interface descriptor data length */
> > +	OUT_BATCH(sizeof(struct gen8_interface_descriptor_data));
> > +	/* interface descriptor address, is relative to the dynamics base address */
> > +	OUT_BATCH(interface_descriptor);
> > +}
> > +
> > +void
> > +gen8_emit_media_state_flush(struct intel_batchbuffer *batch)
> > +{
> > +	OUT_BATCH(GEN8_MEDIA_STATE_FLUSH | (2 - 2));
> > +	OUT_BATCH(0);
> > +}
> > +
> > +void
> > +gen8_emit_media_objects(struct intel_batchbuffer *batch,
> > +			unsigned x, unsigned y,
> > +			unsigned width, unsigned height)
> > +{
> > +	int i, j;
> > +
> > +	for (i = 0; i < width / 16; i++) {
> > +		for (j = 0; j < height / 16; j++) {
> > +			OUT_BATCH(GEN8_MEDIA_OBJECT | (8 - 2));
> > +
> > +			/* interface descriptor offset */
> > +			OUT_BATCH(0);
> > +
> > +			/* without indirect data */
> > +			OUT_BATCH(0);
> > +			OUT_BATCH(0);
> > +
> > +			/* scoreboard */
> > +			OUT_BATCH(0);
> > +			OUT_BATCH(0);
> > +
> > +			/* inline data (xoffset, yoffset) */
> > +			OUT_BATCH(x + i * 16);
> > +			OUT_BATCH(y + j * 16);
> > +			gen8_emit_media_state_flush(batch);
> > +		}
> > +	}
> > +}
> > +
> > +void
> > +gen8lp_emit_media_objects(struct intel_batchbuffer *batch,
> > +			unsigned x, unsigned y,
> > +			unsigned width, unsigned height)
> > +{
> > +	int i, j;
> > +
> > +	for (i = 0; i < width / 16; i++) {
> > +		for (j = 0; j < height / 16; j++) {
> > +			OUT_BATCH(GEN8_MEDIA_OBJECT | (8 - 2));
> > +
> > +			/* interface descriptor offset */
> > +			OUT_BATCH(0);
> > +
> > +			/* without indirect data */
> > +			OUT_BATCH(0);
> > +			OUT_BATCH(0);
> > +
> > +			/* scoreboard */
> > +			OUT_BATCH(0);
> > +			OUT_BATCH(0);
> > +
> > +			/* inline data (xoffset, yoffset) */
> > +			OUT_BATCH(x + i * 16);
> > +			OUT_BATCH(y + j * 16);
> > +		}
> > +	}
> > +}
> > +
> > +void
> > +gen9_emit_state_base_address(struct intel_batchbuffer *batch)
> > +{
> > +	OUT_BATCH(GEN8_STATE_BASE_ADDRESS | (19 - 2));
> > +
> > +	/* general */
> > +	OUT_BATCH(0 | BASE_ADDRESS_MODIFY);
> > +	OUT_BATCH(0);
> > +
> > +	/* stateless data port */
> > +	OUT_BATCH(0 | BASE_ADDRESS_MODIFY);
> > +
> > +	/* surface */
> > +	OUT_RELOC(batch->bo, I915_GEM_DOMAIN_SAMPLER, 0, BASE_ADDRESS_MODIFY);
> > +
> > +	/* dynamic */
> > +	OUT_RELOC(batch->bo, I915_GEM_DOMAIN_RENDER | I915_GEM_DOMAIN_INSTRUCTION,
> > +		0, BASE_ADDRESS_MODIFY);
> > +
> > +	/* indirect */
> > +	OUT_BATCH(0);
> > +	OUT_BATCH(0);
> > +
> > +	/* instruction */
> > +	OUT_RELOC(batch->bo, I915_GEM_DOMAIN_INSTRUCTION, 0, BASE_ADDRESS_MODIFY);
> > +
> > +	/* general state buffer size */
> > +	OUT_BATCH(0xfffff000 | 1);
> > +	/* dynamic state buffer size */
> > +	OUT_BATCH(1 << 12 | 1);
> > +	/* indirect object buffer size */
> > +	OUT_BATCH(0xfffff000 | 1);
> > +	/* intruction buffer size, must set modify enable bit, otherwise it may result in GPU hang */
> > +	OUT_BATCH(1 << 12 | 1);
> > +
> > +	/* Bindless surface state base address */
> > +	OUT_BATCH(0 | BASE_ADDRESS_MODIFY);
> > +	OUT_BATCH(0);
> > +	OUT_BATCH(0xfffff000);
> > +}
> > +
> > +void
> > +gen7_emit_gpgpu_walk(struct intel_batchbuffer *batch,
> > +		     unsigned x, unsigned y,
> > +		     unsigned width, unsigned height)
> > +{
> > +	uint32_t x_dim, y_dim, tmp, right_mask;
> > +
> > +	/*
> > +	 * Simply do SIMD16 based dispatch, so every thread uses
> > +	 * SIMD16 channels.
> > +	 *
> > +	 * Define our own thread group size, e.g 16x1 for every group, then
> > +	 * will have 1 thread each group in SIMD16 dispatch. So thread
> > +	 * width/height/depth are all 1.
> > +	 *
> > +	 * Then thread group X = width / 16 (aligned to 16)
> > +	 * thread group Y = height;
> > +	 */
> > +	x_dim = (width + 15) / 16;
> > +	y_dim = height;
> > +
> > +	tmp = width & 15;
> > +	if (tmp == 0)
> > +		right_mask = (1 << 16) - 1;
> > +	else
> > +		right_mask = (1 << tmp) - 1;
> > +
> > +	OUT_BATCH(GEN7_GPGPU_WALKER | 9);
> > +
> > +	/* interface descriptor offset */
> > +	OUT_BATCH(0);
> > +
> > +	/* SIMD size, thread w/h/d */
> > +	OUT_BATCH(1 << 30 | /* SIMD16 */
> > +		  0 << 16 | /* depth:1 */
> > +		  0 << 8 | /* height:1 */
> > +		  0); /* width:1 */
> > +
> > +	/* thread group X */
> > +	OUT_BATCH(0);
> > +	OUT_BATCH(x_dim);
> > +
> > +	/* thread group Y */
> > +	OUT_BATCH(0);
> > +	OUT_BATCH(y_dim);
> > +
> > +	/* thread group Z */
> > +	OUT_BATCH(0);
> > +	OUT_BATCH(1);
> > +
> > +	/* right mask */
> > +	OUT_BATCH(right_mask);
> > +
> > +	/* bottom mask, height 1, always 0xffffffff */
> > +	OUT_BATCH(0xffffffff);
> > +}
> > +
> > +void
> > +gen8_emit_gpgpu_walk(struct intel_batchbuffer *batch,
> > +		     unsigned x, unsigned y,
> > +		     unsigned width, unsigned height)
> > +{
> > +	uint32_t x_dim, y_dim, tmp, right_mask;
> > +
> > +	/*
> > +	 * Simply do SIMD16 based dispatch, so every thread uses
> > +	 * SIMD16 channels.
> > +	 *
> > +	 * Define our own thread group size, e.g 16x1 for every group, then
> > +	 * will have 1 thread each group in SIMD16 dispatch. So thread
> > +	 * width/height/depth are all 1.
> > +	 *
> > +	 * Then thread group X = width / 16 (aligned to 16)
> > +	 * thread group Y = height;
> > +	 */
> > +	x_dim = (width + 15) / 16;
> > +	y_dim = height;
> > +
> > +	tmp = width & 15;
> > +	if (tmp == 0)
> > +		right_mask = (1 << 16) - 1;
> > +	else
> > +		right_mask = (1 << tmp) - 1;
> > +
> > +	OUT_BATCH(GEN7_GPGPU_WALKER | 13);
> > +
> > +	OUT_BATCH(0); /* kernel offset */
> > +	OUT_BATCH(0); /* indirect data length */
> > +	OUT_BATCH(0); /* indirect data offset */
> > +
> > +	/* SIMD size, thread w/h/d */
> > +	OUT_BATCH(1 << 30 | /* SIMD16 */
> > +		  0 << 16 | /* depth:1 */
> > +		  0 << 8 | /* height:1 */
> > +		  0); /* width:1 */
> > +
> > +	/* thread group X */
> > +	OUT_BATCH(0);
> > +	OUT_BATCH(0);
> > +	OUT_BATCH(x_dim);
> > +
> > +	/* thread group Y */
> > +	OUT_BATCH(0);
> > +	OUT_BATCH(0);
> > +	OUT_BATCH(y_dim);
> > +
> > +	/* thread group Z */
> > +	OUT_BATCH(0);
> > +	OUT_BATCH(1);
> > +
> > +	/* right mask */
> > +	OUT_BATCH(right_mask);
> > +
> > +	/* bottom mask, height 1, always 0xffffffff */
> > +	OUT_BATCH(0xffffffff);
> > +}
> > diff --git a/lib/gpu_fill.h b/lib/gpu_fill.h
> > new file mode 100644
> > index 00000000..a271ce6e
> > --- /dev/null
> > +++ b/lib/gpu_fill.h
> > @@ -0,0 +1,141 @@
> > +/*
> > + * Copyright © 2018 Intel Corporation
> > + *
> > + * Permission is hereby granted, free of charge, to any person obtaining a
> > + * copy of this software and associated documentation files (the "Software"),
> > + * to deal in the Software without restriction, including without limitation
> > + * the rights to use, copy, modify, merge, publish, distribute, sublicense,
> > + * and/or sell copies of the Software, and to permit persons to whom the
> > + * Software is furnished to do so, subject to the following conditions:
> > + *
> > + * The above copyright notice and this permission notice (including the next
> > + * paragraph) shall be included in all copies or substantial portions of the
> > + * Software.
> > + *
> > + * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
> > + * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
> > + * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.  IN NO EVENT SHALL
> > + * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
> > + * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
> > + * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS
> > + * IN THE SOFTWARE.
> > + *
> > + */
> > +
> > +#ifndef GPU_FILL_H
> > +#define GPU_FILL_H
> > +
> > +uint32_t
> > +batch_used(struct intel_batchbuffer *batch);
> > +
> > +uint32_t
> > +batch_align(struct intel_batchbuffer *batch, uint32_t align);
> > +
> > +void *
> > +batch_alloc(struct intel_batchbuffer *batch, uint32_t size, uint32_t align);
> > +
> > +uint32_t
> > +batch_offset(struct intel_batchbuffer *batch, void *ptr);
> > +
> > +uint32_t
> > +batch_copy(struct intel_batchbuffer *batch, const void *ptr, uint32_t size, uint32_t align);
> > +
> > +void
> > +gen7_render_flush(struct intel_batchbuffer *batch, uint32_t batch_end);
> > +
> > +uint32_t
> > +gen7_fill_curbe_buffer_data(struct intel_batchbuffer *batch,
> > +			uint8_t color);
> > +
> > +uint32_t
> > +gen7_fill_surface_state(struct intel_batchbuffer *batch,
> > +			struct igt_buf *buf,
> > +			uint32_t format,
> > +			int is_dst);
> > +
> > +uint32_t
> > +gen7_fill_binding_table(struct intel_batchbuffer *batch,
> > +			struct igt_buf *dst);
> > +uint32_t
> > +gen7_fill_kernel(struct intel_batchbuffer *batch,
> > +		const uint32_t kernel[][4],
> > +		size_t size);
> > +
> > +uint32_t
> > +gen7_fill_interface_descriptor(struct intel_batchbuffer *batch, struct igt_buf *dst,
> > +			       const uint32_t kernel[][4], size_t size);
> > +
> > +void
> > +gen7_emit_state_base_address(struct intel_batchbuffer *batch);
> > +
> > +void
> > +gen7_emit_vfe_state(struct intel_batchbuffer *batch);
> > +
> > +void
> > +gen7_emit_vfe_state_gpgpu(struct intel_batchbuffer *batch);
> > +
> > +void
> > +gen7_emit_curbe_load(struct intel_batchbuffer *batch, uint32_t curbe_buffer);
> > +
> > +void
> > +gen7_emit_interface_descriptor_load(struct intel_batchbuffer *batch, uint32_t interface_descriptor);
> > +
> > +void
> > +gen7_emit_media_objects(struct intel_batchbuffer *batch,
> > +			unsigned x, unsigned y,
> > +			unsigned width, unsigned height);
> > +
> > +uint32_t
> > +gen8_fill_surface_state(struct intel_batchbuffer *batch,
> > +			struct igt_buf *buf,
> > +			uint32_t format,
> > +			int is_dst);
> > +
> > +uint32_t
> > +gen8_fill_binding_table(struct intel_batchbuffer *batch,
> > +			struct igt_buf *dst);
> > +
> > +uint32_t
> > +gen8_fill_interface_descriptor(struct intel_batchbuffer *batch, struct igt_buf *dst, const uint32_t kernel[][4], size_t size);
> > +
> > +void
> > +gen8_emit_state_base_address(struct intel_batchbuffer *batch);
> > +
> > +void
> > +gen8_emit_vfe_state(struct intel_batchbuffer *batch);
> > +
> > +void
> > +gen8_emit_vfe_state_gpgpu(struct intel_batchbuffer *batch);
> > +
> > +void
> > +gen8_emit_curbe_load(struct intel_batchbuffer *batch, uint32_t curbe_buffer);
> > +
> > +void
> > +gen8_emit_interface_descriptor_load(struct intel_batchbuffer *batch, uint32_t interface_descriptor);
> > +
> > +void
> > +gen8_emit_media_state_flush(struct intel_batchbuffer *batch);
> > +
> > +void
> > +gen8_emit_media_objects(struct intel_batchbuffer *batch,
> > +			unsigned x, unsigned y,
> > +			unsigned width, unsigned height);
> > +
> > +void
> > +gen8lp_emit_media_objects(struct intel_batchbuffer *batch,
> > +			unsigned x, unsigned y,
> > +			unsigned width, unsigned height);
> > +
> > +void
> > +gen9_emit_state_base_address(struct intel_batchbuffer *batch);
> > +
> > +void
> > +gen7_emit_gpgpu_walk(struct intel_batchbuffer *batch,
> > +		     unsigned x, unsigned y,
> > +		     unsigned width, unsigned height);
> > +void
> > +gen8_emit_gpgpu_walk(struct intel_batchbuffer *batch,
> > +		     unsigned x, unsigned y,
> > +		     unsigned width, unsigned height);
> > +
> > +#endif /* GPU_FILL_H */
> > diff --git a/lib/media_fill_gen7.c b/lib/media_fill_gen7.c
> > index 6fb44798..c97555a6 100644
> > --- a/lib/media_fill_gen7.c
> > +++ b/lib/media_fill_gen7.c
> > @@ -5,7 +5,7 @@
> >   #include "gen7_media.h"
> >   #include "intel_reg.h"
> >   #include "drmtest.h"
> > -
> > +#include "gpu_fill.h"
> >   #include <assert.h>
> >   static const uint32_t media_kernel[][4] = {
> > @@ -22,275 +22,6 @@ static const uint32_t media_kernel[][4] = {
> >   	{ 0x07800031, 0x20001ca8, 0x00000e00, 0x82000010 },
> >   };
> > -static uint32_t
> > -batch_used(struct intel_batchbuffer *batch)
> > -{
> > -	return batch->ptr - batch->buffer;
> > -}
> > -
> > -static uint32_t
> > -batch_align(struct intel_batchbuffer *batch, uint32_t align)
> > -{
> > -	uint32_t offset = batch_used(batch);
> > -	offset = ALIGN(offset, align);
> > -	batch->ptr = batch->buffer + offset;
> > -	return offset;
> > -}
> > -
> > -static void *
> > -batch_alloc(struct intel_batchbuffer *batch, uint32_t size, uint32_t align)
> > -{
> > -	uint32_t offset = batch_align(batch, align);
> > -	batch->ptr += size;
> > -	return memset(batch->buffer + offset, 0, size);
> > -}
> > -
> > -static uint32_t
> > -batch_offset(struct intel_batchbuffer *batch, void *ptr)
> > -{
> > -	return (uint8_t *)ptr - batch->buffer;
> > -}
> > -
> > -static uint32_t
> > -batch_copy(struct intel_batchbuffer *batch, const void *ptr, uint32_t size, uint32_t align)
> > -{
> > -	return batch_offset(batch, memcpy(batch_alloc(batch, size, align), ptr, size));
> > -}
> > -
> > -static void
> > -gen7_render_flush(struct intel_batchbuffer *batch, uint32_t batch_end)
> > -{
> > -	int ret;
> > -
> > -	ret = drm_intel_bo_subdata(batch->bo, 0, 4096, batch->buffer);
> > -	if (ret == 0)
> > -		ret = drm_intel_bo_mrb_exec(batch->bo, batch_end,
> > -					NULL, 0, 0, 0);
> > -	igt_assert(ret == 0);
> > -}
> > -
> > -static uint32_t
> > -gen7_fill_curbe_buffer_data(struct intel_batchbuffer *batch,
> > -			uint8_t color)
> > -{
> > -	uint8_t *curbe_buffer;
> > -	uint32_t offset;
> > -
> > -	curbe_buffer = batch_alloc(batch, sizeof(uint32_t) * 8, 64);
> > -	offset = batch_offset(batch, curbe_buffer);
> > -	*curbe_buffer = color;
> > -
> > -	return offset;
> > -}
> > -
> > -static uint32_t
> > -gen7_fill_surface_state(struct intel_batchbuffer *batch,
> > -			struct igt_buf *buf,
> > -			uint32_t format,
> > -			int is_dst)
> > -{
> > -	struct gen7_surface_state *ss;
> > -	uint32_t write_domain, read_domain, offset;
> > -	int ret;
> > -
> > -	if (is_dst) {
> > -		write_domain = read_domain = I915_GEM_DOMAIN_RENDER;
> > -	} else {
> > -		write_domain = 0;
> > -		read_domain = I915_GEM_DOMAIN_SAMPLER;
> > -	}
> > -
> > -	ss = batch_alloc(batch, sizeof(*ss), 64);
> > -	offset = batch_offset(batch, ss);
> > -
> > -	ss->ss0.surface_type = GEN7_SURFACE_2D;
> > -	ss->ss0.surface_format = format;
> > -	ss->ss0.render_cache_read_write = 1;
> > -
> > -	if (buf->tiling == I915_TILING_X)
> > -		ss->ss0.tiled_mode = 2;
> > -	else if (buf->tiling == I915_TILING_Y)
> > -		ss->ss0.tiled_mode = 3;
> > -
> > -	ss->ss1.base_addr = buf->bo->offset;
> > -	ret = drm_intel_bo_emit_reloc(batch->bo,
> > -				batch_offset(batch, ss) + 4,
> > -				buf->bo, 0,
> > -				read_domain, write_domain);
> > -	igt_assert(ret == 0);
> > -
> > -	ss->ss2.height = igt_buf_height(buf) - 1;
> > -	ss->ss2.width  = igt_buf_width(buf) - 1;
> > -
> > -	ss->ss3.pitch  = buf->stride - 1;
> > -
> > -	ss->ss7.shader_chanel_select_r = 4;
> > -	ss->ss7.shader_chanel_select_g = 5;
> > -	ss->ss7.shader_chanel_select_b = 6;
> > -	ss->ss7.shader_chanel_select_a = 7;
> > -
> > -	return offset;
> > -}
> > -
> > -static uint32_t
> > -gen7_fill_binding_table(struct intel_batchbuffer *batch,
> > -			struct igt_buf *dst)
> > -{
> > -	uint32_t *binding_table, offset;
> > -
> > -	binding_table = batch_alloc(batch, 32, 64);
> > -	offset = batch_offset(batch, binding_table);
> > -
> > -	binding_table[0] = gen7_fill_surface_state(batch, dst, GEN7_SURFACEFORMAT_R8_UNORM, 1);
> > -
> > -	return offset;
> > -}
> > -
> > -static uint32_t
> > -gen7_fill_media_kernel(struct intel_batchbuffer *batch,
> > -		const uint32_t kernel[][4],
> > -		size_t size)
> > -{
> > -	uint32_t offset;
> > -
> > -	offset = batch_copy(batch, kernel, size, 64);
> > -
> > -	return offset;
> > -}
> > -
> > -static uint32_t
> > -gen7_fill_interface_descriptor(struct intel_batchbuffer *batch, struct igt_buf *dst,
> > -			       const uint32_t kernel[][4], size_t size)
> > -{
> > -	struct gen7_interface_descriptor_data *idd;
> > -	uint32_t offset;
> > -	uint32_t binding_table_offset, kernel_offset;
> > -
> > -	binding_table_offset = gen7_fill_binding_table(batch, dst);
> > -	kernel_offset = gen7_fill_media_kernel(batch, kernel, size);
> > -
> > -	idd = batch_alloc(batch, sizeof(*idd), 64);
> > -	offset = batch_offset(batch, idd);
> > -
> > -	idd->desc0.kernel_start_pointer = (kernel_offset >> 6);
> > -
> > -	idd->desc1.single_program_flow = 1;
> > -	idd->desc1.floating_point_mode = GEN7_FLOATING_POINT_IEEE_754;
> > -
> > -	idd->desc2.sampler_count = 0;      /* 0 samplers used */
> > -	idd->desc2.sampler_state_pointer = 0;
> > -
> > -	idd->desc3.binding_table_entry_count = 0;
> > -	idd->desc3.binding_table_pointer = (binding_table_offset >> 5);
> > -
> > -	idd->desc4.constant_urb_entry_read_offset = 0;
> > -	idd->desc4.constant_urb_entry_read_length = 1; /* grf 1 */
> > -
> > -	return offset;
> > -}
> > -
> > -static void
> > -gen7_emit_state_base_address(struct intel_batchbuffer *batch)
> > -{
> > -	OUT_BATCH(GEN7_STATE_BASE_ADDRESS | (10 - 2));
> > -
> > -	/* general */
> > -	OUT_BATCH(0);
> > -
> > -	/* surface */
> > -	OUT_RELOC(batch->bo, I915_GEM_DOMAIN_INSTRUCTION, 0, BASE_ADDRESS_MODIFY);
> > -
> > -	/* dynamic */
> > -	OUT_RELOC(batch->bo, I915_GEM_DOMAIN_INSTRUCTION, 0, BASE_ADDRESS_MODIFY);
> > -
> > -	/* indirect */
> > -	OUT_BATCH(0);
> > -
> > -	/* instruction */
> > -	OUT_RELOC(batch->bo, I915_GEM_DOMAIN_INSTRUCTION, 0, BASE_ADDRESS_MODIFY);
> > -
> > -	/* general/dynamic/indirect/instruction access Bound */
> > -	OUT_BATCH(0);
> > -	OUT_BATCH(0 | BASE_ADDRESS_MODIFY);
> > -	OUT_BATCH(0);
> > -	OUT_BATCH(0 | BASE_ADDRESS_MODIFY);
> > -}
> > -
> > -static void
> > -gen7_emit_vfe_state(struct intel_batchbuffer *batch)
> > -{
> > -	OUT_BATCH(GEN7_MEDIA_VFE_STATE | (8 - 2));
> > -
> > -	/* scratch buffer */
> > -	OUT_BATCH(0);
> > -
> > -	/* number of threads & urb entries */
> > -	OUT_BATCH(1 << 16 |
> > -		2 << 8);
> > -
> > -	OUT_BATCH(0);
> > -
> > -	/* urb entry size & curbe size */
> > -	OUT_BATCH(2 << 16 | 	/* in 256 bits unit */
> > -		2);		/* in 256 bits unit */
> > -
> > -	/* scoreboard */
> > -	OUT_BATCH(0);
> > -	OUT_BATCH(0);
> > -	OUT_BATCH(0);
> > -}
> > -
> > -static void
> > -gen7_emit_curbe_load(struct intel_batchbuffer *batch, uint32_t curbe_buffer)
> > -{
> > -	OUT_BATCH(GEN7_MEDIA_CURBE_LOAD | (4 - 2));
> > -	OUT_BATCH(0);
> > -	/* curbe total data length */
> > -	OUT_BATCH(64);
> > -	/* curbe data start address, is relative to the dynamics base address */
> > -	OUT_BATCH(curbe_buffer);
> > -}
> > -
> > -static void
> > -gen7_emit_interface_descriptor_load(struct intel_batchbuffer *batch, uint32_t interface_descriptor)
> > -{
> > -	OUT_BATCH(GEN7_MEDIA_INTERFACE_DESCRIPTOR_LOAD | (4 - 2));
> > -	OUT_BATCH(0);
> > -	/* interface descriptor data length */
> > -	OUT_BATCH(sizeof(struct gen7_interface_descriptor_data));
> > -	/* interface descriptor address, is relative to the dynamics base address */
> > -	OUT_BATCH(interface_descriptor);
> > -}
> > -
> > -static void
> > -gen7_emit_media_objects(struct intel_batchbuffer *batch,
> > -			unsigned x, unsigned y,
> > -			unsigned width, unsigned height)
> > -{
> > -	int i, j;
> > -
> > -	for (i = 0; i < width / 16; i++) {
> > -		for (j = 0; j < height / 16; j++) {
> > -			OUT_BATCH(GEN7_MEDIA_OBJECT | (8 - 2));
> > -
> > -			/* interface descriptor offset */
> > -			OUT_BATCH(0);
> > -
> > -			/* without indirect data */
> > -			OUT_BATCH(0);
> > -			OUT_BATCH(0);
> > -
> > -			/* scoreboard */
> > -			OUT_BATCH(0);
> > -			OUT_BATCH(0);
> > -
> > -			/* inline data (xoffset, yoffset) */
> > -			OUT_BATCH(x + i * 16);
> > -			OUT_BATCH(y + j * 16);
> > -		}
> > -	}
> > -}
> > -
> >   /*
> >    * This sets up the media pipeline,
> >    *
> > diff --git a/lib/media_fill_gen8.c b/lib/media_fill_gen8.c
> > index 4a8fe5a2..88c7dbdd 100644
> > --- a/lib/media_fill_gen8.c
> > +++ b/lib/media_fill_gen8.c
> > @@ -5,7 +5,7 @@
> >   #include "gen8_media.h"
> >   #include "intel_reg.h"
> >   #include "drmtest.h"
> > -
> > +#include "gpu_fill.h"
> >   #include <assert.h>
> > @@ -23,294 +23,6 @@ static const uint32_t media_kernel[][4] = {
> >   	{ 0x07800031, 0x20000a40, 0x0e000e00, 0x82000010 },
> >   };
> > -static uint32_t
> > -batch_used(struct intel_batchbuffer *batch)
> > -{
> > -	return batch->ptr - batch->buffer;
> > -}
> > -
> > -static uint32_t
> > -batch_align(struct intel_batchbuffer *batch, uint32_t align)
> > -{
> > -	uint32_t offset = batch_used(batch);
> > -	offset = ALIGN(offset, align);
> > -	batch->ptr = batch->buffer + offset;
> > -	return offset;
> > -}
> > -
> > -static void *
> > -batch_alloc(struct intel_batchbuffer *batch, uint32_t size, uint32_t align)
> > -{
> > -	uint32_t offset = batch_align(batch, align);
> > -	batch->ptr += size;
> > -	return memset(batch->buffer + offset, 0, size);
> > -}
> > -
> > -static uint32_t
> > -batch_offset(struct intel_batchbuffer *batch, void *ptr)
> > -{
> > -	return (uint8_t *)ptr - batch->buffer;
> > -}
> > -
> > -static uint32_t
> > -batch_copy(struct intel_batchbuffer *batch, const void *ptr, uint32_t size, uint32_t align)
> > -{
> > -	return batch_offset(batch, memcpy(batch_alloc(batch, size, align), ptr, size));
> > -}
> > -
> > -static void
> > -gen8_render_flush(struct intel_batchbuffer *batch, uint32_t batch_end)
> > -{
> > -	int ret;
> > -
> > -	ret = drm_intel_bo_subdata(batch->bo, 0, 4096, batch->buffer);
> > -	if (ret == 0)
> > -		ret = drm_intel_bo_mrb_exec(batch->bo, batch_end,
> > -					NULL, 0, 0, 0);
> > -	igt_assert(ret == 0);
> > -}
> > -
> > -static uint32_t
> > -gen8_fill_curbe_buffer_data(struct intel_batchbuffer *batch,
> > -			uint8_t color)
> > -{
> > -	uint8_t *curbe_buffer;
> > -	uint32_t offset;
> > -
> > -	curbe_buffer = batch_alloc(batch, sizeof(uint32_t) * 8, 64);
> > -	offset = batch_offset(batch, curbe_buffer);
> > -	*curbe_buffer = color;
> > -
> > -	return offset;
> > -}
> > -
> > -static uint32_t
> > -gen8_fill_surface_state(struct intel_batchbuffer *batch,
> > -			struct igt_buf *buf,
> > -			uint32_t format,
> > -			int is_dst)
> > -{
> > -	struct gen8_surface_state *ss;
> > -	uint32_t write_domain, read_domain, offset;
> > -	int ret;
> > -
> > -	if (is_dst) {
> > -		write_domain = read_domain = I915_GEM_DOMAIN_RENDER;
> > -	} else {
> > -		write_domain = 0;
> > -		read_domain = I915_GEM_DOMAIN_SAMPLER;
> > -	}
> > -
> > -	ss = batch_alloc(batch, sizeof(*ss), 64);
> > -	offset = batch_offset(batch, ss);
> > -
> > -	ss->ss0.surface_type = GEN8_SURFACE_2D;
> > -	ss->ss0.surface_format = format;
> > -	ss->ss0.render_cache_read_write = 1;
> > -	ss->ss0.vertical_alignment = 1; /* align 4 */
> > -	ss->ss0.horizontal_alignment = 1; /* align 4 */
> > -
> > -	if (buf->tiling == I915_TILING_X)
> > -		ss->ss0.tiled_mode = 2;
> > -	else if (buf->tiling == I915_TILING_Y)
> > -		ss->ss0.tiled_mode = 3;
> > -
> > -	ss->ss8.base_addr = buf->bo->offset;
> > -
> > -	ret = drm_intel_bo_emit_reloc(batch->bo,
> > -				batch_offset(batch, ss) + 8 * 4,
> > -				buf->bo, 0,
> > -				read_domain, write_domain);
> > -	igt_assert(ret == 0);
> > -
> > -	ss->ss2.height = igt_buf_height(buf) - 1;
> > -	ss->ss2.width  = igt_buf_width(buf) - 1;
> > -	ss->ss3.pitch  = buf->stride - 1;
> > -
> > -	ss->ss7.shader_chanel_select_r = 4;
> > -	ss->ss7.shader_chanel_select_g = 5;
> > -	ss->ss7.shader_chanel_select_b = 6;
> > -	ss->ss7.shader_chanel_select_a = 7;
> > -
> > -	return offset;
> > -}
> > -
> > -static uint32_t
> > -gen8_fill_binding_table(struct intel_batchbuffer *batch,
> > -			struct igt_buf *dst)
> > -{
> > -	uint32_t *binding_table, offset;
> > -
> > -	binding_table = batch_alloc(batch, 32, 64);
> > -	offset = batch_offset(batch, binding_table);
> > -
> > -	binding_table[0] = gen8_fill_surface_state(batch, dst, GEN8_SURFACEFORMAT_R8_UNORM, 1);
> > -
> > -	return offset;
> > -}
> > -
> > -static uint32_t
> > -gen8_fill_media_kernel(struct intel_batchbuffer *batch,
> > -		const uint32_t kernel[][4],
> > -		size_t size)
> > -{
> > -	uint32_t offset;
> > -
> > -	offset = batch_copy(batch, kernel, size, 64);
> > -
> > -	return offset;
> > -}
> > -
> > -static uint32_t
> > -gen8_fill_interface_descriptor(struct intel_batchbuffer *batch, struct igt_buf *dst)
> > -{
> > -	struct gen8_interface_descriptor_data *idd;
> > -	uint32_t offset;
> > -	uint32_t binding_table_offset, kernel_offset;
> > -
> > -	binding_table_offset = gen8_fill_binding_table(batch, dst);
> > -	kernel_offset = gen8_fill_media_kernel(batch, media_kernel, sizeof(media_kernel));
> > -
> > -	idd = batch_alloc(batch, sizeof(*idd), 64);
> > -	offset = batch_offset(batch, idd);
> > -
> > -	idd->desc0.kernel_start_pointer = (kernel_offset >> 6);
> > -
> > -	idd->desc2.single_program_flow = 1;
> > -	idd->desc2.floating_point_mode = GEN8_FLOATING_POINT_IEEE_754;
> > -
> > -	idd->desc3.sampler_count = 0;      /* 0 samplers used */
> > -	idd->desc3.sampler_state_pointer = 0;
> > -
> > -	idd->desc4.binding_table_entry_count = 0;
> > -	idd->desc4.binding_table_pointer = (binding_table_offset >> 5);
> > -
> > -	idd->desc5.constant_urb_entry_read_offset = 0;
> > -	idd->desc5.constant_urb_entry_read_length = 1; /* grf 1 */
> > -
> > -	return offset;
> > -}
> > -
> > -static void
> > -gen8_emit_state_base_address(struct intel_batchbuffer *batch)
> > -{
> > -	OUT_BATCH(GEN8_STATE_BASE_ADDRESS | (16 - 2));
> > -
> > -	/* general */
> > -	OUT_BATCH(0 | BASE_ADDRESS_MODIFY);
> > -	OUT_BATCH(0);
> > -
> > -	/* stateless data port */
> > -	OUT_BATCH(0 | BASE_ADDRESS_MODIFY);
> > -
> > -	/* surface */
> > -	OUT_RELOC(batch->bo, I915_GEM_DOMAIN_SAMPLER, 0, BASE_ADDRESS_MODIFY);
> > -
> > -	/* dynamic */
> > -	OUT_RELOC(batch->bo, I915_GEM_DOMAIN_RENDER | I915_GEM_DOMAIN_INSTRUCTION,
> > -		0, BASE_ADDRESS_MODIFY);
> > -
> > -	/* indirect */
> > -	OUT_BATCH(0);
> > -	OUT_BATCH(0);
> > -
> > -	/* instruction */
> > -	OUT_RELOC(batch->bo, I915_GEM_DOMAIN_INSTRUCTION, 0, BASE_ADDRESS_MODIFY);
> > -
> > -	/* general state buffer size */
> > -	OUT_BATCH(0xfffff000 | 1);
> > -	/* dynamic state buffer size */
> > -	OUT_BATCH(1 << 12 | 1);
> > -	/* indirect object buffer size */
> > -	OUT_BATCH(0xfffff000 | 1);
> > -	/* intruction buffer size, must set modify enable bit, otherwise it may result in GPU hang */
> > -	OUT_BATCH(1 << 12 | 1);
> > -}
> > -
> > -static void
> > -gen8_emit_vfe_state(struct intel_batchbuffer *batch)
> > -{
> > -	OUT_BATCH(GEN8_MEDIA_VFE_STATE | (9 - 2));
> > -
> > -	/* scratch buffer */
> > -	OUT_BATCH(0);
> > -	OUT_BATCH(0);
> > -
> > -	/* number of threads & urb entries */
> > -	OUT_BATCH(1 << 16 |
> > -		2 << 8);
> > -
> > -	OUT_BATCH(0);
> > -
> > -	/* urb entry size & curbe size */
> > -	OUT_BATCH(2 << 16 |
> > -		2);
> > -
> > -	/* scoreboard */
> > -	OUT_BATCH(0);
> > -	OUT_BATCH(0);
> > -	OUT_BATCH(0);
> > -}
> > -
> > -static void
> > -gen8_emit_curbe_load(struct intel_batchbuffer *batch, uint32_t curbe_buffer)
> > -{
> > -	OUT_BATCH(GEN8_MEDIA_CURBE_LOAD | (4 - 2));
> > -	OUT_BATCH(0);
> > -	/* curbe total data length */
> > -	OUT_BATCH(64);
> > -	/* curbe data start address, is relative to the dynamics base address */
> > -	OUT_BATCH(curbe_buffer);
> > -}
> > -
> > -static void
> > -gen8_emit_interface_descriptor_load(struct intel_batchbuffer *batch, uint32_t interface_descriptor)
> > -{
> > -	OUT_BATCH(GEN8_MEDIA_INTERFACE_DESCRIPTOR_LOAD | (4 - 2));
> > -	OUT_BATCH(0);
> > -	/* interface descriptor data length */
> > -	OUT_BATCH(sizeof(struct gen8_interface_descriptor_data));
> > -	/* interface descriptor address, is relative to the dynamics base address */
> > -	OUT_BATCH(interface_descriptor);
> > -}
> > -
> > -static void
> > -gen8_emit_media_state_flush(struct intel_batchbuffer *batch)
> > -{
> > -	OUT_BATCH(GEN8_MEDIA_STATE_FLUSH | (2 - 2));
> > -	OUT_BATCH(0);
> > -}
> > -
> > -static void
> > -gen8_emit_media_objects(struct intel_batchbuffer *batch,
> > -			unsigned x, unsigned y,
> > -			unsigned width, unsigned height)
> > -{
> > -	int i, j;
> > -
> > -	for (i = 0; i < width / 16; i++) {
> > -		for (j = 0; j < height / 16; j++) {
> > -			OUT_BATCH(GEN8_MEDIA_OBJECT | (8 - 2));
> > -
> > -			/* interface descriptor offset */
> > -			OUT_BATCH(0);
> > -
> > -			/* without indirect data */
> > -			OUT_BATCH(0);
> > -			OUT_BATCH(0);
> > -
> > -			/* scoreboard */
> > -			OUT_BATCH(0);
> > -			OUT_BATCH(0);
> > -
> > -			/* inline data (xoffset, yoffset) */
> > -			OUT_BATCH(x + i * 16);
> > -			OUT_BATCH(y + j * 16);
> > -			gen8_emit_media_state_flush(batch);
> > -		}
> > -	}
> > -}
> > -
> >   /*
> >    * This sets up the media pipeline,
> >    *
> > @@ -348,8 +60,9 @@ gen8_media_fillfunc(struct intel_batchbuffer *batch,
> >   	/* setup states */
> >   	batch->ptr = &batch->buffer[BATCH_STATE_SPLIT];
> > -	curbe_buffer = gen8_fill_curbe_buffer_data(batch, color);
> > -	interface_descriptor = gen8_fill_interface_descriptor(batch, dst);
> > +	curbe_buffer = gen7_fill_curbe_buffer_data(batch, color);
> > +	interface_descriptor = gen8_fill_interface_descriptor(batch, dst,
> > +					media_kernel, sizeof(media_kernel));
> >   	igt_assert(batch->ptr < &batch->buffer[4095]);
> >   	/* media pipeline */
> > @@ -370,6 +83,6 @@ gen8_media_fillfunc(struct intel_batchbuffer *batch,
> >   	batch_end = batch_align(batch, 8);
> >   	igt_assert(batch_end < BATCH_STATE_SPLIT);
> > -	gen8_render_flush(batch, batch_end);
> > +	gen7_render_flush(batch, batch_end);
> >   	intel_batchbuffer_reset(batch);
> >   }
> > diff --git a/lib/media_fill_gen8lp.c b/lib/media_fill_gen8lp.c
> > index 1f8a4adc..b30d96a3 100644
> > --- a/lib/media_fill_gen8lp.c
> > +++ b/lib/media_fill_gen8lp.c
> > @@ -5,7 +5,7 @@
> >   #include "gen8_media.h"
> >   #include "intel_reg.h"
> >   #include "drmtest.h"
> > -
> > +#include "gpu_fill.h"
> >   #include <assert.h>
> > @@ -23,286 +23,6 @@ static const uint32_t media_kernel[][4] = {
> >   	{ 0x07800031, 0x20000a40, 0x0e000e00, 0x82000010 },
> >   };
> > -static uint32_t
> > -batch_used(struct intel_batchbuffer *batch)
> > -{
> > -	return batch->ptr - batch->buffer;
> > -}
> > -
> > -static uint32_t
> > -batch_align(struct intel_batchbuffer *batch, uint32_t align)
> > -{
> > -	uint32_t offset = batch_used(batch);
> > -	offset = ALIGN(offset, align);
> > -	batch->ptr = batch->buffer + offset;
> > -	return offset;
> > -}
> > -
> > -static void *
> > -batch_alloc(struct intel_batchbuffer *batch, uint32_t size, uint32_t align)
> > -{
> > -	uint32_t offset = batch_align(batch, align);
> > -	batch->ptr += size;
> > -	return memset(batch->buffer + offset, 0, size);
> > -}
> > -
> > -static uint32_t
> > -batch_offset(struct intel_batchbuffer *batch, void *ptr)
> > -{
> > -	return (uint8_t *)ptr - batch->buffer;
> > -}
> > -
> > -static uint32_t
> > -batch_copy(struct intel_batchbuffer *batch, const void *ptr, uint32_t size, uint32_t align)
> > -{
> > -	return batch_offset(batch, memcpy(batch_alloc(batch, size, align), ptr, size));
> > -}
> > -
> > -static void
> > -gen8_render_flush(struct intel_batchbuffer *batch, uint32_t batch_end)
> > -{
> > -	int ret;
> > -
> > -	ret = drm_intel_bo_subdata(batch->bo, 0, 4096, batch->buffer);
> > -	if (ret == 0)
> > -		ret = drm_intel_bo_mrb_exec(batch->bo, batch_end,
> > -					NULL, 0, 0, 0);
> > -	igt_assert(ret == 0);
> > -}
> > -
> > -static uint32_t
> > -gen8_fill_curbe_buffer_data(struct intel_batchbuffer *batch,
> > -			uint8_t color)
> > -{
> > -	uint8_t *curbe_buffer;
> > -	uint32_t offset;
> > -
> > -	curbe_buffer = batch_alloc(batch, sizeof(uint32_t) * 8, 64);
> > -	offset = batch_offset(batch, curbe_buffer);
> > -	*curbe_buffer = color;
> > -
> > -	return offset;
> > -}
> > -
> > -static uint32_t
> > -gen8_fill_surface_state(struct intel_batchbuffer *batch,
> > -			struct igt_buf *buf,
> > -			uint32_t format,
> > -			int is_dst)
> > -{
> > -	struct gen8_surface_state *ss;
> > -	uint32_t write_domain, read_domain, offset;
> > -	int ret;
> > -
> > -	if (is_dst) {
> > -		write_domain = read_domain = I915_GEM_DOMAIN_RENDER;
> > -	} else {
> > -		write_domain = 0;
> > -		read_domain = I915_GEM_DOMAIN_SAMPLER;
> > -	}
> > -
> > -	ss = batch_alloc(batch, sizeof(*ss), 64);
> > -	offset = batch_offset(batch, ss);
> > -
> > -	ss->ss0.surface_type = GEN8_SURFACE_2D;
> > -	ss->ss0.surface_format = format;
> > -	ss->ss0.render_cache_read_write = 1;
> > -	ss->ss0.vertical_alignment = 1; /* align 4 */
> > -	ss->ss0.horizontal_alignment = 1; /* align 4 */
> > -
> > -	if (buf->tiling == I915_TILING_X)
> > -		ss->ss0.tiled_mode = 2;
> > -	else if (buf->tiling == I915_TILING_Y)
> > -		ss->ss0.tiled_mode = 3;
> > -
> > -	ss->ss8.base_addr = buf->bo->offset;
> > -
> > -	ret = drm_intel_bo_emit_reloc(batch->bo,
> > -				batch_offset(batch, ss) + 8 * 4,
> > -				buf->bo, 0,
> > -				read_domain, write_domain);
> > -	igt_assert(ret == 0);
> > -
> > -	ss->ss2.height = igt_buf_height(buf) - 1;
> > -	ss->ss2.width  = igt_buf_width(buf) - 1;
> > -	ss->ss3.pitch  = buf->stride - 1;
> > -
> > -	ss->ss7.shader_chanel_select_r = 4;
> > -	ss->ss7.shader_chanel_select_g = 5;
> > -	ss->ss7.shader_chanel_select_b = 6;
> > -	ss->ss7.shader_chanel_select_a = 7;
> > -
> > -	return offset;
> > -}
> > -
> > -static uint32_t
> > -gen8_fill_binding_table(struct intel_batchbuffer *batch,
> > -			struct igt_buf *dst)
> > -{
> > -	uint32_t *binding_table, offset;
> > -
> > -	binding_table = batch_alloc(batch, 32, 64);
> > -	offset = batch_offset(batch, binding_table);
> > -
> > -	binding_table[0] = gen8_fill_surface_state(batch, dst, GEN8_SURFACEFORMAT_R8_UNORM, 1);
> > -
> > -	return offset;
> > -}
> > -
> > -static uint32_t
> > -gen8_fill_media_kernel(struct intel_batchbuffer *batch,
> > -		const uint32_t kernel[][4],
> > -		size_t size)
> > -{
> > -	uint32_t offset;
> > -
> > -	offset = batch_copy(batch, kernel, size, 64);
> > -
> > -	return offset;
> > -}
> > -
> > -static uint32_t
> > -gen8_fill_interface_descriptor(struct intel_batchbuffer *batch, struct igt_buf *dst)
> > -{
> > -	struct gen8_interface_descriptor_data *idd;
> > -	uint32_t offset;
> > -	uint32_t binding_table_offset, kernel_offset;
> > -
> > -	binding_table_offset = gen8_fill_binding_table(batch, dst);
> > -	kernel_offset = gen8_fill_media_kernel(batch, media_kernel, sizeof(media_kernel));
> > -
> > -	idd = batch_alloc(batch, sizeof(*idd), 64);
> > -	offset = batch_offset(batch, idd);
> > -
> > -	idd->desc0.kernel_start_pointer = (kernel_offset >> 6);
> > -
> > -	idd->desc2.single_program_flow = 1;
> > -	idd->desc2.floating_point_mode = GEN8_FLOATING_POINT_IEEE_754;
> > -
> > -	idd->desc3.sampler_count = 0;      /* 0 samplers used */
> > -	idd->desc3.sampler_state_pointer = 0;
> > -
> > -	idd->desc4.binding_table_entry_count = 0;
> > -	idd->desc4.binding_table_pointer = (binding_table_offset >> 5);
> > -
> > -	idd->desc5.constant_urb_entry_read_offset = 0;
> > -	idd->desc5.constant_urb_entry_read_length = 1; /* grf 1 */
> > -
> > -	return offset;
> > -}
> > -
> > -static void
> > -gen8_emit_state_base_address(struct intel_batchbuffer *batch)
> > -{
> > -	OUT_BATCH(GEN8_STATE_BASE_ADDRESS | (16 - 2));
> > -
> > -	/* general */
> > -	OUT_BATCH(0 | BASE_ADDRESS_MODIFY);
> > -	OUT_BATCH(0);
> > -
> > -	/* stateless data port */
> > -	OUT_BATCH(0 | BASE_ADDRESS_MODIFY);
> > -
> > -	/* surface */
> > -	OUT_RELOC(batch->bo, I915_GEM_DOMAIN_SAMPLER, 0, BASE_ADDRESS_MODIFY);
> > -
> > -	/* dynamic */
> > -	OUT_RELOC(batch->bo, I915_GEM_DOMAIN_RENDER | I915_GEM_DOMAIN_INSTRUCTION,
> > -		0, BASE_ADDRESS_MODIFY);
> > -
> > -	/* indirect */
> > -	OUT_BATCH(0);
> > -	OUT_BATCH(0);
> > -
> > -	/* instruction */
> > -	OUT_RELOC(batch->bo, I915_GEM_DOMAIN_INSTRUCTION, 0, BASE_ADDRESS_MODIFY);
> > -
> > -	/* general state buffer size */
> > -	OUT_BATCH(0xfffff000 | 1);
> > -	/* dynamic state buffer size */
> > -	OUT_BATCH(1 << 12 | 1);
> > -	/* indirect object buffer size */
> > -	OUT_BATCH(0xfffff000 | 1);
> > -	/* intruction buffer size, must set modify enable bit, otherwise it may result in GPU hang */
> > -	OUT_BATCH(1 << 12 | 1);
> > -}
> > -
> > -static void
> > -gen8_emit_vfe_state(struct intel_batchbuffer *batch)
> > -{
> > -	OUT_BATCH(GEN8_MEDIA_VFE_STATE | (9 - 2));
> > -
> > -	/* scratch buffer */
> > -	OUT_BATCH(0);
> > -	OUT_BATCH(0);
> > -
> > -	/* number of threads & urb entries */
> > -	OUT_BATCH(1 << 16 |
> > -		2 << 8);
> > -
> > -	OUT_BATCH(0);
> > -
> > -	/* urb entry size & curbe size */
> > -	OUT_BATCH(2 << 16 |
> > -		2);
> > -
> > -	/* scoreboard */
> > -	OUT_BATCH(0);
> > -	OUT_BATCH(0);
> > -	OUT_BATCH(0);
> > -}
> > -
> > -static void
> > -gen8_emit_curbe_load(struct intel_batchbuffer *batch, uint32_t curbe_buffer)
> > -{
> > -	OUT_BATCH(GEN8_MEDIA_CURBE_LOAD | (4 - 2));
> > -	OUT_BATCH(0);
> > -	/* curbe total data length */
> > -	OUT_BATCH(64);
> > -	/* curbe data start address, is relative to the dynamics base address */
> > -	OUT_BATCH(curbe_buffer);
> > -}
> > -
> > -static void
> > -gen8_emit_interface_descriptor_load(struct intel_batchbuffer *batch, uint32_t interface_descriptor)
> > -{
> > -	OUT_BATCH(GEN8_MEDIA_INTERFACE_DESCRIPTOR_LOAD | (4 - 2));
> > -	OUT_BATCH(0);
> > -	/* interface descriptor data length */
> > -	OUT_BATCH(sizeof(struct gen8_interface_descriptor_data));
> > -	/* interface descriptor address, is relative to the dynamics base address */
> > -	OUT_BATCH(interface_descriptor);
> > -}
> > -
> > -static void
> > -gen8lp_emit_media_objects(struct intel_batchbuffer *batch,
> > -			unsigned x, unsigned y,
> > -			unsigned width, unsigned height)
> > -{
> > -	int i, j;
> > -
> > -	for (i = 0; i < width / 16; i++) {
> > -		for (j = 0; j < height / 16; j++) {
> > -			OUT_BATCH(GEN8_MEDIA_OBJECT | (8 - 2));
> > -
> > -			/* interface descriptor offset */
> > -			OUT_BATCH(0);
> > -
> > -			/* without indirect data */
> > -			OUT_BATCH(0);
> > -			OUT_BATCH(0);
> > -
> > -			/* scoreboard */
> > -			OUT_BATCH(0);
> > -			OUT_BATCH(0);
> > -
> > -			/* inline data (xoffset, yoffset) */
> > -			OUT_BATCH(x + i * 16);
> > -			OUT_BATCH(y + j * 16);
> > -		}
> > -	}
> > -}
> > -
> >   /*
> >    * This sets up the media pipeline,
> >    *
> > @@ -340,8 +60,9 @@ gen8lp_media_fillfunc(struct intel_batchbuffer *batch,
> >   	/* setup states */
> >   	batch->ptr = &batch->buffer[BATCH_STATE_SPLIT];
> > -	curbe_buffer = gen8_fill_curbe_buffer_data(batch, color);
> > -	interface_descriptor = gen8_fill_interface_descriptor(batch, dst);
> > +	curbe_buffer = gen7_fill_curbe_buffer_data(batch, color);
> > +	interface_descriptor = gen8_fill_interface_descriptor(batch, dst,
> > +					media_kernel, sizeof(media_kernel));
> >   	igt_assert(batch->ptr < &batch->buffer[4095]);
> >   	/* media pipeline */
> > @@ -362,6 +83,6 @@ gen8lp_media_fillfunc(struct intel_batchbuffer *batch,
> >   	batch_end = batch_align(batch, 8);
> >   	igt_assert(batch_end < BATCH_STATE_SPLIT);
> > -	gen8_render_flush(batch, batch_end);
> > +	gen7_render_flush(batch, batch_end);
> >   	intel_batchbuffer_reset(batch);
> >   }
> > diff --git a/lib/media_fill_gen9.c b/lib/media_fill_gen9.c
> > index 3fd21819..e5d94487 100644
> > --- a/lib/media_fill_gen9.c
> > +++ b/lib/media_fill_gen9.c
> > @@ -4,11 +4,9 @@
> >   #include "media_fill.h"
> >   #include "gen8_media.h"
> >   #include "intel_reg.h"
> > -
> > +#include "gpu_fill.h"
> >   #include <assert.h>
> > -#define ALIGN(x, y) (((x) + (y)-1) & ~((y)-1))
> > -
> >   static const uint32_t media_kernel[][4] = {
> >   	{ 0x00400001, 0x20202288, 0x00000020, 0x00000000 },
> >   	{ 0x00600001, 0x20800208, 0x008d0000, 0x00000000 },
> > @@ -23,299 +21,6 @@ static const uint32_t media_kernel[][4] = {
> >   	{ 0x07800031, 0x20000a40, 0x0e000e00, 0x82000010 },
> >   };
> > -static uint32_t
> > -batch_used(struct intel_batchbuffer *batch)
> > -{
> > -	return batch->ptr - batch->buffer;
> > -}
> > -
> > -static uint32_t
> > -batch_align(struct intel_batchbuffer *batch, uint32_t align)
> > -{
> > -	uint32_t offset = batch_used(batch);
> > -	offset = ALIGN(offset, align);
> > -	batch->ptr = batch->buffer + offset;
> > -	return offset;
> > -}
> > -
> > -static void *
> > -batch_alloc(struct intel_batchbuffer *batch, uint32_t size, uint32_t align)
> > -{
> > -	uint32_t offset = batch_align(batch, align);
> > -	batch->ptr += size;
> > -	return memset(batch->buffer + offset, 0, size);
> > -}
> > -
> > -static uint32_t
> > -batch_offset(struct intel_batchbuffer *batch, void *ptr)
> > -{
> > -	return (uint8_t *)ptr - batch->buffer;
> > -}
> > -
> > -static uint32_t
> > -batch_copy(struct intel_batchbuffer *batch, const void *ptr, uint32_t size, uint32_t align)
> > -{
> > -	return batch_offset(batch, memcpy(batch_alloc(batch, size, align), ptr, size));
> > -}
> > -
> > -static void
> > -gen8_render_flush(struct intel_batchbuffer *batch, uint32_t batch_end)
> > -{
> > -	int ret;
> > -
> > -	ret = drm_intel_bo_subdata(batch->bo, 0, 4096, batch->buffer);
> > -	if (ret == 0)
> > -		ret = drm_intel_bo_mrb_exec(batch->bo, batch_end,
> > -					NULL, 0, 0, 0);
> > -	assert(ret == 0);
> > -}
> > -
> > -static uint32_t
> > -gen8_fill_curbe_buffer_data(struct intel_batchbuffer *batch,
> > -			uint8_t color)
> > -{
> > -	uint8_t *curbe_buffer;
> > -	uint32_t offset;
> > -
> > -	curbe_buffer = batch_alloc(batch, sizeof(uint32_t) * 8, 64);
> > -	offset = batch_offset(batch, curbe_buffer);
> > -	*curbe_buffer = color;
> > -
> > -	return offset;
> > -}
> > -
> > -static uint32_t
> > -gen8_fill_surface_state(struct intel_batchbuffer *batch,
> > -			struct igt_buf *buf,
> > -			uint32_t format,
> > -			int is_dst)
> > -{
> > -	struct gen8_surface_state *ss;
> > -	uint32_t write_domain, read_domain, offset;
> > -	int ret;
> > -
> > -	if (is_dst) {
> > -		write_domain = read_domain = I915_GEM_DOMAIN_RENDER;
> > -	} else {
> > -		write_domain = 0;
> > -		read_domain = I915_GEM_DOMAIN_SAMPLER;
> > -	}
> > -
> > -	ss = batch_alloc(batch, sizeof(*ss), 64);
> > -	offset = batch_offset(batch, ss);
> > -
> > -	ss->ss0.surface_type = GEN8_SURFACE_2D;
> > -	ss->ss0.surface_format = format;
> > -	ss->ss0.render_cache_read_write = 1;
> > -	ss->ss0.vertical_alignment = 1; /* align 4 */
> > -	ss->ss0.horizontal_alignment = 1; /* align 4 */
> > -
> > -	if (buf->tiling == I915_TILING_X)
> > -		ss->ss0.tiled_mode = 2;
> > -	else if (buf->tiling == I915_TILING_Y)
> > -		ss->ss0.tiled_mode = 3;
> > -
> > -	ss->ss8.base_addr = buf->bo->offset;
> > -
> > -	ret = drm_intel_bo_emit_reloc(batch->bo,
> > -				batch_offset(batch, ss) + 8 * 4,
> > -				buf->bo, 0,
> > -				read_domain, write_domain);
> > -	assert(ret == 0);
> > -
> > -	ss->ss2.height = igt_buf_height(buf) - 1;
> > -	ss->ss2.width  = igt_buf_width(buf) - 1;
> > -	ss->ss3.pitch  = buf->stride - 1;
> > -
> > -	ss->ss7.shader_chanel_select_r = 4;
> > -	ss->ss7.shader_chanel_select_g = 5;
> > -	ss->ss7.shader_chanel_select_b = 6;
> > -	ss->ss7.shader_chanel_select_a = 7;
> > -
> > -	return offset;
> > -}
> > -
> > -static uint32_t
> > -gen8_fill_binding_table(struct intel_batchbuffer *batch,
> > -			struct igt_buf *dst)
> > -{
> > -	uint32_t *binding_table, offset;
> > -
> > -	binding_table = batch_alloc(batch, 32, 64);
> > -	offset = batch_offset(batch, binding_table);
> > -
> > -	binding_table[0] = gen8_fill_surface_state(batch, dst, GEN8_SURFACEFORMAT_R8_UNORM, 1);
> > -
> > -	return offset;
> > -}
> > -
> > -static uint32_t
> > -gen8_fill_media_kernel(struct intel_batchbuffer *batch,
> > -		const uint32_t kernel[][4],
> > -		size_t size)
> > -{
> > -	uint32_t offset;
> > -
> > -	offset = batch_copy(batch, kernel, size, 64);
> > -
> > -	return offset;
> > -}
> > -
> > -static uint32_t
> > -gen8_fill_interface_descriptor(struct intel_batchbuffer *batch, struct igt_buf *dst)
> > -{
> > -	struct gen8_interface_descriptor_data *idd;
> > -	uint32_t offset;
> > -	uint32_t binding_table_offset, kernel_offset;
> > -
> > -	binding_table_offset = gen8_fill_binding_table(batch, dst);
> > -	kernel_offset = gen8_fill_media_kernel(batch, media_kernel, sizeof(media_kernel));
> > -
> > -	idd = batch_alloc(batch, sizeof(*idd), 64);
> > -	offset = batch_offset(batch, idd);
> > -
> > -	idd->desc0.kernel_start_pointer = (kernel_offset >> 6);
> > -
> > -	idd->desc2.single_program_flow = 1;
> > -	idd->desc2.floating_point_mode = GEN8_FLOATING_POINT_IEEE_754;
> > -
> > -	idd->desc3.sampler_count = 0;      /* 0 samplers used */
> > -	idd->desc3.sampler_state_pointer = 0;
> > -
> > -	idd->desc4.binding_table_entry_count = 0;
> > -	idd->desc4.binding_table_pointer = (binding_table_offset >> 5);
> > -
> > -	idd->desc5.constant_urb_entry_read_offset = 0;
> > -	idd->desc5.constant_urb_entry_read_length = 1; /* grf 1 */
> > -
> > -	return offset;
> > -}
> > -
> > -static void
> > -gen9_emit_state_base_address(struct intel_batchbuffer *batch)
> > -{
> > -	OUT_BATCH(GEN8_STATE_BASE_ADDRESS | (19 - 2));
> > -
> > -	/* general */
> > -	OUT_BATCH(0 | BASE_ADDRESS_MODIFY);
> > -	OUT_BATCH(0);
> > -
> > -	/* stateless data port */
> > -	OUT_BATCH(0 | BASE_ADDRESS_MODIFY);
> > -
> > -	/* surface */
> > -	OUT_RELOC(batch->bo, I915_GEM_DOMAIN_SAMPLER, 0, BASE_ADDRESS_MODIFY);
> > -
> > -	/* dynamic */
> > -	OUT_RELOC(batch->bo, I915_GEM_DOMAIN_RENDER | I915_GEM_DOMAIN_INSTRUCTION,
> > -		0, BASE_ADDRESS_MODIFY);
> > -
> > -	/* indirect */
> > -	OUT_BATCH(0);
> > -	OUT_BATCH(0);
> > -
> > -	/* instruction */
> > -	OUT_RELOC(batch->bo, I915_GEM_DOMAIN_INSTRUCTION, 0, BASE_ADDRESS_MODIFY);
> > -
> > -	/* general state buffer size */
> > -	OUT_BATCH(0xfffff000 | 1);
> > -	/* dynamic state buffer size */
> > -	OUT_BATCH(1 << 12 | 1);
> > -	/* indirect object buffer size */
> > -	OUT_BATCH(0xfffff000 | 1);
> > -	/* intruction buffer size, must set modify enable bit, otherwise it may result in GPU hang */
> > -	OUT_BATCH(1 << 12 | 1);
> > -
> > -	/* Bindless surface state base address */
> > -	OUT_BATCH(0 | BASE_ADDRESS_MODIFY);
> > -	OUT_BATCH(0);
> > -	OUT_BATCH(0xfffff000);
> > -}
> > -
> > -static void
> > -gen8_emit_vfe_state(struct intel_batchbuffer *batch)
> > -{
> > -	OUT_BATCH(GEN8_MEDIA_VFE_STATE | (9 - 2));
> > -
> > -	/* scratch buffer */
> > -	OUT_BATCH(0);
> > -	OUT_BATCH(0);
> > -
> > -	/* number of threads & urb entries */
> > -	OUT_BATCH(1 << 16 |
> > -		2 << 8);
> > -
> > -	OUT_BATCH(0);
> > -
> > -	/* urb entry size & curbe size */
> > -	OUT_BATCH(2 << 16 |
> > -		2);
> > -
> > -	/* scoreboard */
> > -	OUT_BATCH(0);
> > -	OUT_BATCH(0);
> > -	OUT_BATCH(0);
> > -}
> > -
> > -static void
> > -gen8_emit_curbe_load(struct intel_batchbuffer *batch, uint32_t curbe_buffer)
> > -{
> > -	OUT_BATCH(GEN8_MEDIA_CURBE_LOAD | (4 - 2));
> > -	OUT_BATCH(0);
> > -	/* curbe total data length */
> > -	OUT_BATCH(64);
> > -	/* curbe data start address, is relative to the dynamics base address */
> > -	OUT_BATCH(curbe_buffer);
> > -}
> > -
> > -static void
> > -gen8_emit_interface_descriptor_load(struct intel_batchbuffer *batch, uint32_t interface_descriptor)
> > -{
> > -	OUT_BATCH(GEN8_MEDIA_INTERFACE_DESCRIPTOR_LOAD | (4 - 2));
> > -	OUT_BATCH(0);
> > -	/* interface descriptor data length */
> > -	OUT_BATCH(sizeof(struct gen8_interface_descriptor_data));
> > -	/* interface descriptor address, is relative to the dynamics base address */
> > -	OUT_BATCH(interface_descriptor);
> > -}
> > -
> > -static void
> > -gen8_emit_media_state_flush(struct intel_batchbuffer *batch)
> > -{
> > -	OUT_BATCH(GEN8_MEDIA_STATE_FLUSH | (2 - 2));
> > -	OUT_BATCH(0);
> > -}
> > -
> > -static void
> > -gen8_emit_media_objects(struct intel_batchbuffer *batch,
> > -			unsigned x, unsigned y,
> > -			unsigned width, unsigned height)
> > -{
> > -	int i, j;
> > -
> > -	for (i = 0; i < width / 16; i++) {
> > -		for (j = 0; j < height / 16; j++) {
> > -			OUT_BATCH(GEN8_MEDIA_OBJECT | (8 - 2));
> > -
> > -			/* interface descriptor offset */
> > -			OUT_BATCH(0);
> > -
> > -			/* without indirect data */
> > -			OUT_BATCH(0);
> > -			OUT_BATCH(0);
> > -
> > -			/* scoreboard */
> > -			OUT_BATCH(0);
> > -			OUT_BATCH(0);
> > -
> > -			/* inline data (xoffset, yoffset) */
> > -			OUT_BATCH(x + i * 16);
> > -			OUT_BATCH(y + j * 16);
> > -			gen8_emit_media_state_flush(batch);
> > -		}
> > -	}
> > -}
> > -
> >   /*
> >    * This sets up the media pipeline,
> >    *
> > @@ -353,8 +58,9 @@ gen9_media_fillfunc(struct intel_batchbuffer *batch,
> >   	/* setup states */
> >   	batch->ptr = &batch->buffer[BATCH_STATE_SPLIT];
> > -	curbe_buffer = gen8_fill_curbe_buffer_data(batch, color);
> > -	interface_descriptor = gen8_fill_interface_descriptor(batch, dst);
> > +	curbe_buffer = gen7_fill_curbe_buffer_data(batch, color);
> > +	interface_descriptor = gen8_fill_interface_descriptor(batch, dst,
> > +					media_kernel, sizeof(media_kernel));
> >   	assert(batch->ptr < &batch->buffer[4095]);
> >   	/* media pipeline */
> > @@ -387,6 +93,6 @@ gen9_media_fillfunc(struct intel_batchbuffer *batch,
> >   	batch_end = batch_align(batch, 8);
> >   	assert(batch_end < BATCH_STATE_SPLIT);
> > -	gen8_render_flush(batch, batch_end);
> > +	gen7_render_flush(batch, batch_end);
> >   	intel_batchbuffer_reset(batch);
> >   }
> > diff --git a/lib/meson.build b/lib/meson.build
> > index b3b8b14a..38c3f107 100644
> > --- a/lib/meson.build
> > +++ b/lib/meson.build
> > @@ -24,6 +24,7 @@ lib_sources = [
> >   	'intel_os.c',
> >   	'intel_mmio.c',
> >   	'ioctl_wrappers.c',
> > +	'gpu_fill.c',
> >   	'media_fill_gen7.c',
> >   	'media_fill_gen8.c',
> >   	'media_fill_gen8lp.c',
> > 


More information about the igt-dev mailing list