[Intel-gfx] [RFC PATCH 2/2] drm/i915/gen7: Clear all EU/L3 residual contexts
Chris Wilson
chris at chris-wilson.co.uk
Tue Jan 14 22:35:56 UTC 2020
Quoting Akeem G Abodunrin (2020-01-14 14:51:36)
> From: Prathap Kumar Valsan <prathap.kumar.valsan at intel.com>
>
> On gen7 and gen7.5 devices, there could be leftover data residuals in
> EU/L3 from the retiring context. This patch introduces workaround to clear
> that residual contexts, by submitting a batch buffer with dedicated HW
> context to the GPU with ring allocation for each context switching.
>
> Signed-off-by: Mika Kuoppala <mika.kuoppala at linux.intel.com>
> Signed-off-by: Prathap Kumar Valsan <prathap.kumar.valsan at intel.com>
> Signed-off-by: Akeem G Abodunrin <akeem.g.abodunrin at intel.com>
> Cc: Chris Wilson <chris.p.wilson at intel.com>
> Cc: Balestrieri Francesco <francesco.balestrieri at intel.com>
> Cc: Bloomfield Jon <jon.bloomfield at intel.com>
> Cc: Dutt Sudeep <sudeep.dutt at intel.com>
> ---
> drivers/gpu/drm/i915/Makefile | 1 +
> drivers/gpu/drm/i915/gt/gen7_renderclear.c | 514 ++++++++++++++++++
> drivers/gpu/drm/i915/gt/gen7_renderclear.h | 15 +
> drivers/gpu/drm/i915/gt/intel_gpu_commands.h | 17 +-
> .../gpu/drm/i915/gt/intel_ring_submission.c | 3 +-
> drivers/gpu/drm/i915/i915_utils.h | 5 +
> 6 files changed, 551 insertions(+), 4 deletions(-)
> create mode 100644 drivers/gpu/drm/i915/gt/gen7_renderclear.c
> create mode 100644 drivers/gpu/drm/i915/gt/gen7_renderclear.h
>
> diff --git a/drivers/gpu/drm/i915/Makefile b/drivers/gpu/drm/i915/Makefile
> index b8c5f8934dbd..e5386871f015 100644
> --- a/drivers/gpu/drm/i915/Makefile
> +++ b/drivers/gpu/drm/i915/Makefile
> @@ -75,6 +75,7 @@ gt-y += \
> gt/debugfs_gt.o \
> gt/debugfs_gt_pm.o \
> gt/gen6_ppgtt.o \
> + gt/gen7_renderclear.o \
> gt/gen8_ppgtt.o \
> gt/intel_breadcrumbs.o \
> gt/intel_context.o \
> diff --git a/drivers/gpu/drm/i915/gt/gen7_renderclear.c b/drivers/gpu/drm/i915/gt/gen7_renderclear.c
> new file mode 100644
> index 000000000000..3e9fc2c05fbb
> --- /dev/null
> +++ b/drivers/gpu/drm/i915/gt/gen7_renderclear.c
> @@ -0,0 +1,514 @@
> +// SPDX-License-Identifier: MIT
> +/*
> + * Copyright © 2019 Intel Corporation
> + */
> +
> +#include "gen7_renderclear.h"
> +#include "i915_drv.h"
> +#include "i915_utils.h"
> +#include "intel_gpu_commands.h"
> +
> +#define MAX_URB_ENTRIES 64
> +#define STATE_SIZE (4 * 1024)
> +
> +/* Media CB Kernel for gen7 devices */
> +static const u32 cb7_kernel[][4] = {
> + { 0x00000001, 0x26020128, 0x00000024, 0x00000000 },
> + { 0x00000040, 0x20280c21, 0x00000028, 0x00000001 },
> + { 0x01000010, 0x20000c20, 0x0000002c, 0x00000000 },
> + { 0x00010220, 0x34001c00, 0x00001400, 0x0000002c },
> + { 0x00600001, 0x20600061, 0x00000000, 0x00000000 },
> + { 0x00000008, 0x20601c85, 0x00000e00, 0x0000000c },
> + { 0x00000005, 0x20601ca5, 0x00000060, 0x00000001 },
> + { 0x00000008, 0x20641c85, 0x00000e00, 0x0000000d },
> + { 0x00000005, 0x20641ca5, 0x00000064, 0x00000003 },
> + { 0x00000041, 0x207424a5, 0x00000064, 0x00000034 },
> + { 0x00000040, 0x206014a5, 0x00000060, 0x00000074 },
> + { 0x00000008, 0x20681c85, 0x00000e00, 0x00000008 },
> + { 0x00000005, 0x20681ca5, 0x00000068, 0x0000000f },
> + { 0x00000041, 0x20701ca5, 0x00000060, 0x00000010 },
> + { 0x00000040, 0x206814a5, 0x00000068, 0x00000070 },
> + { 0x00600001, 0x20a00061, 0x00000000, 0x00000000 },
> + { 0x00000005, 0x206c1c85, 0x00000e00, 0x00000007 },
> + { 0x00000041, 0x206c1ca5, 0x0000006c, 0x00000004 },
> + { 0x00600001, 0x20800021, 0x008d0000, 0x00000000 },
> + { 0x00000001, 0x20800021, 0x0000006c, 0x00000000 },
> + { 0x00000001, 0x20840021, 0x00000068, 0x00000000 },
> + { 0x00000001, 0x20880061, 0x00000000, 0x00000003 },
> + { 0x00000005, 0x208c0d21, 0x00000086, 0xffffffff },
> + { 0x05600032, 0x20a01fa1, 0x008d0080, 0x02190001 },
> + { 0x00000040, 0x20a01ca5, 0x000000a0, 0x00000001 },
> + { 0x05600032, 0x20a01fa1, 0x008d0080, 0x040a8001 },
> + { 0x02000040, 0x20281c21, 0x00000028, 0xffffffff },
> + { 0x00010220, 0x34001c00, 0x00001400, 0xfffffffc },
> + { 0x00000001, 0x26020128, 0x00000024, 0x00000000 },
> + { 0x00000001, 0x220000e4, 0x00000000, 0x00000000 },
> + { 0x00000001, 0x220801ec, 0x00000000, 0x007f007f },
> + { 0x00600001, 0x20400021, 0x008d0000, 0x00000000 },
> + { 0x00600001, 0x2fe00021, 0x008d0000, 0x00000000 },
> + { 0x00200001, 0x20400121, 0x00450020, 0x00000000 },
> + { 0x00000001, 0x20480061, 0x00000000, 0x000f000f },
> + { 0x00000005, 0x204c0d21, 0x00000046, 0xffffffef },
> + { 0x00800001, 0x20600061, 0x00000000, 0x00000000 },
> + { 0x00800001, 0x20800061, 0x00000000, 0x00000000 },
> + { 0x00800001, 0x20a00061, 0x00000000, 0x00000000 },
> + { 0x00800001, 0x20c00061, 0x00000000, 0x00000000 },
> + { 0x00800001, 0x20e00061, 0x00000000, 0x00000000 },
> + { 0x00800001, 0x21000061, 0x00000000, 0x00000000 },
> + { 0x00800001, 0x21200061, 0x00000000, 0x00000000 },
> + { 0x00800001, 0x21400061, 0x00000000, 0x00000000 },
> + { 0x05600032, 0x20001fa0, 0x008d0040, 0x120a8000 },
> + { 0x00000040, 0x20402d21, 0x00000020, 0x00100010 },
> + { 0x05600032, 0x20001fa0, 0x008d0040, 0x120a8000 },
> + { 0x02000040, 0x22083d8c, 0x00000208, 0xffffffff },
> + { 0x00800001, 0xa0000109, 0x00000602, 0x00000000 },
> + { 0x00000040, 0x22001c84, 0x00000200, 0x00000020 },
> + { 0x00010220, 0x34001c00, 0x00001400, 0xfffffff8 },
> + { 0x07600032, 0x20001fa0, 0x008d0fe0, 0x82000010 },
> + { 0x00000000, 0x00000000, 0x00000000, 0x00000000 },
> + { 0x00000000, 0x00000000, 0x00000000, 0x00000000 },
> + { 0x00000000, 0x00000000, 0x00000000, 0x00000000 },
> + { 0x00000000, 0x00000000, 0x00000000, 0x00000000 },
> + { 0x00000000, 0x00000000, 0x00000000, 0x00000000 },
> + { 0x00000000, 0x00000000, 0x00000000, 0x00000000 },
> + { 0x00000000, 0x00000000, 0x00000000, 0x00000000 },
> + { 0x00000000, 0x00000000, 0x00000000, 0x00000000 },
> +};
> +
> +/* Media CB Kernel for gen7.5 devices */
> +static const u32 cb75_kernel[][4] = {
> + { 0x00000001, 0x26020128, 0x00000024, 0x00000000 },
> + { 0x00000040, 0x20280c21, 0x00000028, 0x00000001 },
> + { 0x01000010, 0x20000c20, 0x0000002c, 0x00000000 },
> + { 0x00010220, 0x34001c00, 0x00001400, 0x00000160 },
> + { 0x00600001, 0x20600061, 0x00000000, 0x00000000 },
> + { 0x00000008, 0x20601c85, 0x00000e00, 0x0000000c },
> + { 0x00000005, 0x20601ca5, 0x00000060, 0x00000001 },
> + { 0x00000008, 0x20641c85, 0x00000e00, 0x0000000d },
> + { 0x00000005, 0x20641ca5, 0x00000064, 0x00000003 },
> + { 0x00000041, 0x207424a5, 0x00000064, 0x00000034 },
> + { 0x00000040, 0x206014a5, 0x00000060, 0x00000074 },
> + { 0x00000008, 0x20681c85, 0x00000e00, 0x00000008 },
> + { 0x00000005, 0x20681ca5, 0x00000068, 0x0000000f },
> + { 0x00000041, 0x20701ca5, 0x00000060, 0x00000010 },
> + { 0x00000040, 0x206814a5, 0x00000068, 0x00000070 },
> + { 0x00600001, 0x20a00061, 0x00000000, 0x00000000 },
> + { 0x00000005, 0x206c1c85, 0x00000e00, 0x00000007 },
> + { 0x00000041, 0x206c1ca5, 0x0000006c, 0x00000004 },
> + { 0x00600001, 0x20800021, 0x008d0000, 0x00000000 },
> + { 0x00000001, 0x20800021, 0x0000006c, 0x00000000 },
> + { 0x00000001, 0x20840021, 0x00000068, 0x00000000 },
> + { 0x00000001, 0x20880061, 0x00000000, 0x00000003 },
> + { 0x00000005, 0x208c0d21, 0x00000086, 0xffffffff },
> + { 0x05600032, 0x20a01fa1, 0x008d0080, 0x02190001 },
> + { 0x00000040, 0x20a01ca5, 0x000000a0, 0x00000001 },
> + { 0x05600032, 0x20a01fa1, 0x008d0080, 0x040a8001 },
> + { 0x02000040, 0x20281c21, 0x00000028, 0xffffffff },
> + { 0x00010220, 0x34001c00, 0x00001400, 0xffffffe0 },
> + { 0x00000001, 0x26020128, 0x00000024, 0x00000000 },
> + { 0x00000001, 0x220000e4, 0x00000000, 0x00000000 },
> + { 0x00000001, 0x220801ec, 0x00000000, 0x007f007f },
> + { 0x00600001, 0x20400021, 0x008d0000, 0x00000000 },
> + { 0x00600001, 0x2fe00021, 0x008d0000, 0x00000000 },
> + { 0x00200001, 0x20400121, 0x00450020, 0x00000000 },
> + { 0x00000001, 0x20480061, 0x00000000, 0x000f000f },
> + { 0x00000005, 0x204c0d21, 0x00000046, 0xffffffef },
> + { 0x00800001, 0x20600061, 0x00000000, 0x00000000 },
> + { 0x00800001, 0x20800061, 0x00000000, 0x00000000 },
> + { 0x00800001, 0x20a00061, 0x00000000, 0x00000000 },
> + { 0x00800001, 0x20c00061, 0x00000000, 0x00000000 },
> + { 0x00800001, 0x20e00061, 0x00000000, 0x00000000 },
> + { 0x00800001, 0x21000061, 0x00000000, 0x00000000 },
> + { 0x00800001, 0x21200061, 0x00000000, 0x00000000 },
> + { 0x00800001, 0x21400061, 0x00000000, 0x00000000 },
> + { 0x05600032, 0x20001fa0, 0x008d0040, 0x120a8000 },
> + { 0x00000040, 0x20402d21, 0x00000020, 0x00100010 },
> + { 0x05600032, 0x20001fa0, 0x008d0040, 0x120a8000 },
> + { 0x02000040, 0x22083d8c, 0x00000208, 0xffffffff },
> + { 0x00800001, 0xa0000109, 0x00000602, 0x00000000 },
> + { 0x00000040, 0x22001c84, 0x00000200, 0x00000020 },
> + { 0x00010220, 0x34001c00, 0x00001400, 0xffffffc0 },
> + { 0x07600032, 0x20001fa0, 0x008d0fe0, 0x82000010 },
> + { 0x00000000, 0x00000000, 0x00000000, 0x00000000 },
> + { 0x00000000, 0x00000000, 0x00000000, 0x00000000 },
> + { 0x00000000, 0x00000000, 0x00000000, 0x00000000 },
> + { 0x00000000, 0x00000000, 0x00000000, 0x00000000 },
> + { 0x00000000, 0x00000000, 0x00000000, 0x00000000 },
> + { 0x00000000, 0x00000000, 0x00000000, 0x00000000 },
> + { 0x00000000, 0x00000000, 0x00000000, 0x00000000 },
> + { 0x00000000, 0x00000000, 0x00000000, 0x00000000 },
> +};
> +
> +struct cb_kernel {
> + const void *data;
> + u32 size;
> +};
> +
> +#define CB_KERNEL(name) { .data = (name), .size = sizeof(name) }
> +
> +static const struct cb_kernel cb_kernel_gen7 = CB_KERNEL(cb7_kernel);
> +static const struct cb_kernel cb_kernel_hsw = CB_KERNEL(cb75_kernel);
> +
> +struct batch_chunk {
> + struct i915_vma *vma;
> + u32 offset;
> + u32 *start;
> + u32 *end;
> + u32 max_items;
> +};
> +
> +struct batch_vals {
> + struct drm_i915_private *i915;
Never set or used.
> + u32 max_primitives;
> + u32 max_urb_entries;
> + u32 cmd_size;
> + u32 state_size;
> + u32 state_start;
> + u32 batch_size;
> + u32 surface_height;
> + u32 surface_width;
> + u32 scratch_size;
> + u32 max_size;
> +};
> +
> +static void
> +batch_get_defaults(struct drm_i915_private *i915, struct batch_vals *bv)
> +{
> + if (IS_HASWELL(i915)) {
> + bv->max_primitives = 280;
> + bv->max_urb_entries = MAX_URB_ENTRIES;
> + bv->surface_height = 16 * 16;
> + bv->surface_width = 32 * 2 * 16;
> + } else {
> + bv->max_primitives = 128;
> + bv->max_urb_entries = MAX_URB_ENTRIES / 2;
> + bv->surface_height = 16 * 8;
> + bv->surface_width = 32 * 16;
> + }
> + bv->cmd_size = bv->max_primitives * 4096;
> + bv->state_size = STATE_SIZE;
> + bv->state_start = bv->cmd_size;
> + bv->batch_size = bv->cmd_size + bv->state_size;
> + bv->scratch_size = bv->surface_height * bv->surface_width;
> + bv->max_size = bv->batch_size + bv->scratch_size;
> +}
> +
> +static void batch_init(struct batch_chunk *bc,
> + struct i915_vma *vma,
> + u32 *start, u32 offset, u32 max_bytes)
> +{
> + bc->vma = vma;
> + bc->offset = offset;
> + bc->start = start + bc->offset / sizeof(*bc->start);
> + bc->end = bc->start;
> + bc->max_items = max_bytes / sizeof(*bc->start);
> +}
> +
> +static u32 batch_offset(const struct batch_chunk *bc, u32 *cs)
> +{
> + return (cs - bc->start) * sizeof(*bc->start) + bc->offset;
> +}
> +
> +static u32 batch_addr(const struct batch_chunk *bc)
> +{
> + return bc->vma->node.start;
> +}
> +
> +static void batch_add(struct batch_chunk *bc, const u32 d)
> +{
> + GEM_DEBUG_WARN_ON((bc->end - bc->start) >= bc->max_items);
> + *bc->end++ = d;
> +}
> +
> +static u32 *batch_alloc_items(struct batch_chunk *bc, u32 align, u32 items)
> +{
> + u32 *map;
> +
> + if (align) {
> + u32 *end = ptr_align(bc->end, align);
> +
> + memset32(bc->end, 0, (end - bc->end) / sizeof(u32));
end and bc->end are both u32, so we are already taking sizeof(u32) into
account. Just
memset32(bc->end, 0, end - bc->end);
(Good job we cleared the whole buffer just in case.)
> + bc->end = end;
> + }
> +
> + map = bc->end;
> + bc->end += items;
> +
> + return map;
> +}
> +
> +static u32 *batch_alloc_bytes(struct batch_chunk *bc, u32 align, u32 bytes)
> +{
> + GEM_BUG_ON(!IS_ALIGNED(bytes, sizeof(*bc->start)));
> + return batch_alloc_items(bc, align, bytes / sizeof(*bc->start));
> +}
> +
> +static u32
> +gen7_fill_surface_state(struct batch_chunk *state,
> + const u32 dst_offset,
> + const struct batch_vals *bv)
> +{
> + u32 surface_h = bv->surface_height;
> + u32 surface_w = bv->surface_width;
> + u32 *cs = batch_alloc_items(state, 32, 8);
> + u32 offset = batch_offset(state, cs);
> +
> +#define SURFACE_2D 1
> +#define SURFACEFORMAT_B8G8R8A8_UNORM 0x0C0
> +#define RENDER_CACHE_READ_WRITE 1
> +
> + *cs++ = SURFACE_2D << 29 |
> + (SURFACEFORMAT_B8G8R8A8_UNORM << 18) |
> + (RENDER_CACHE_READ_WRITE << 8);
> +
> + *cs++ = batch_addr(state) + dst_offset;
> +
> + *cs++ = ((surface_h / 4 - 1) << 16) | (surface_w / 4 - 1);
> + *cs++ = surface_w;
> + *cs++ = 0;
> + *cs++ = 0;
> + *cs++ = 0;
> +#define SHADER_CHANNELS(r, g, b, a) \
> + (((r) << 25) | ((g) << 22) | ((b) << 19) | ((a) << 16))
> + *cs++ = SHADER_CHANNELS(4, 5, 6, 7);
A useful debug trick would be to finish each packet with
batch_advance(state, cs);
#define batch_advance(X, CS) GEM_BUG_ON((X)->end != (CS))
> +
> + return offset;
> +}
> +
> +static u32
> +gen7_fill_binding_table(struct batch_chunk *state,
> + const struct batch_vals *bv)
> +{
> + u32 *cs = batch_alloc_items(state, 32, 8);
> + u32 offset = batch_offset(state, cs);
> + u32 surface_start;
> +
> + surface_start = gen7_fill_surface_state(state, bv->batch_size, bv);
> + *cs++ = surface_start - state->offset;
> + *cs++ = 0;
> + *cs++ = 0;
> + *cs++ = 0;
> + *cs++ = 0;
> + *cs++ = 0;
> + *cs++ = 0;
> + *cs++ = 0;
> +
> + return offset;
> +}
> +
> +static u32
> +gen7_fill_kernel_data(struct batch_chunk *state,
> + const u32 *data,
> + const u32 size)
> +{
> + return batch_offset(state,
> + memcpy(batch_alloc_bytes(state, 64, size),
> + data, size));
> +}
> +
> +static u32
> +gen7_fill_interface_descriptor(struct batch_chunk *state,
> + const struct batch_vals *bv,
> + const struct cb_kernel *kernel,
> + unsigned int count)
> +{
> + u32 *cs = batch_alloc_items(state, 32, 8 * count);
> + u32 offset = batch_offset(state, cs);
> +
> + *cs++ = gen7_fill_kernel_data(state, kernel->data, kernel->size);
> + *cs++ = (1 << 7) | (1 << 13);
> + *cs++ = 0;
> + *cs++ = (gen7_fill_binding_table(state, bv) - state->offset) | 1;
> + *cs++ = 0;
> + *cs++ = 0;
> + *cs++ = 0;
> + *cs++ = 0;
> + /* 1 - 63dummy idds */
> + memset32(cs, 0x00, (count - 1) * 8);
> +
> + return offset;
> +}
> +
> +static void
> +gen7_emit_state_base_address(struct batch_chunk *batch,
> + u32 surface_state_base)
> +{
> + u32 *cs = batch_alloc_items(batch, 0, 12);
> +
> + *cs++ = STATE_BASE_ADDRESS | (12 - 2);
> + /* general */
> + *cs++ = batch_addr(batch) | BASE_ADDRESS_MODIFY;
> + /* surface */
> + *cs++ = batch_addr(batch) | surface_state_base | BASE_ADDRESS_MODIFY;
> + /* dynamic */
> + *cs++ = batch_addr(batch) | BASE_ADDRESS_MODIFY;
> + /* indirect */
> + *cs++ = batch_addr(batch) | BASE_ADDRESS_MODIFY;
> + /* instruction */
> + *cs++ = batch_addr(batch) | BASE_ADDRESS_MODIFY;
> +
> + /* general/dynamic/indirect/instruction access Bound */
> + *cs++ = 0;
> + *cs++ = BASE_ADDRESS_MODIFY;
> + *cs++ = 0;
> + *cs++ = BASE_ADDRESS_MODIFY;
> + *cs++ = 0;
> + *cs++ = 0;
> +}
> +
> +static void
> +gen7_emit_vfe_state(struct batch_chunk *batch,
> + const struct batch_vals *bv,
> + u32 urb_size, u32 curbe_size,
> + u32 mode)
> +{
> + u32 urb_entries = bv->max_urb_entries;
> + u32 threads = bv->max_primitives - 1;
> + u32 *cs = batch_alloc_items(batch, 32, 8);
> +
> + *cs++ = MEDIA_VFE_STATE | (8 - 2);
> +
> + /* scratch buffer */
> + *cs++ = 0;
> +
> + /* number of threads & urb entries */
> + *cs++ = threads << 16 |
> + urb_entries << 8 |
> + mode << 2; /* GPGPU vs media mode */
*cs++ = threads << 16 | urb_entries << 8 | mode << 2;
Only the comment overflows, rewrite the comment.
> +
> + *cs++ = 0;
> +
> + /* urb entry size & curbe size */
> + *cs++ = urb_size << 16 | /* in 256 bits unit */
> + curbe_size; /* in 256 bits unit */
You could just say in 256b units once, and pull this onto one line.
> +
> + /* scoreboard */
> + *cs++ = 0;
> + *cs++ = 0;
> + *cs++ = 0;
> +}
> +
> +static void
> +gen7_emit_interface_descriptor_load(struct batch_chunk *batch,
> + const u32 interface_descriptor,
> + unsigned int count)
> +{
> + u32 *cs = batch_alloc_items(batch, 8, 4);
> +
> + *cs++ = MEDIA_INTERFACE_DESCRIPTOR_LOAD | (4 - 2);
> + *cs++ = 0;
> + *cs++ = count * 8 * sizeof(*cs);
> +
> + /* interface descriptor address, is relative to the dynamics base
> + * address
> + */
/*
* This is the style we use for block
* comments.
*/
> + *cs++ = interface_descriptor;
> +}
> +
> +static void
> +gen7_emit_media_object(struct batch_chunk *batch,
> + unsigned int media_object_index)
> +{
> + unsigned int x_offset = (media_object_index % 16) * 64;
> + unsigned int y_offset = (media_object_index / 16) * 16;
> + unsigned int inline_data_size;
> + unsigned int media_batch_size;
> + unsigned int i;
> + u32 *cs;
> +
> + inline_data_size = 112 * 8;
> + media_batch_size = inline_data_size + 6;
> +
> + cs = batch_alloc_items(batch, 8, media_batch_size);
> +
> + *cs++ = MEDIA_OBJECT | (media_batch_size - 2);
> +
> + /* interface descriptor offset */
> + *cs++ = 0;
> +
> + /* without indirect data */
> + *cs++ = 0;
> + *cs++ = 0;
> +
> + /* scoreboard */
> + *cs++ = 0;
> + *cs++ = 0;
> +
> + /* inline */
> + *cs++ = (y_offset << 16) | (x_offset);
> + *cs++ = 0;
> + *cs++ = 0x1E00;
> + for (i = 3; i < inline_data_size; i++)
> + *cs++ = 0;
> +}
> +
> +static void gen7_emit_pipeline_flush(struct batch_chunk *batch)
> +{
> + u32 *cs = batch_alloc_items(batch, 0, 5);
> +
> + *cs++ = GFX_OP_PIPE_CONTROL(5);
> + *cs++ = PIPE_CONTROL_STATE_CACHE_INVALIDATE |
> + PIPE_CONTROL_GLOBAL_GTT_IVB;
> + *cs++ = 0;
> + *cs++ = 0;
> + *cs++ = 0;
> +}
> +
> +static void emit_batch(struct i915_vma * const vma,
> + u32 *start,
> + const struct batch_vals *bv)
> +{
> + struct drm_i915_private *i915 = vma->vm->i915;
> + unsigned int desc_count = 64;
> + const u32 urb_size = 112;
> + struct batch_chunk cmds, state;
> + u32 interface_descriptor;
> + unsigned int i;
> +
> + batch_init(&cmds, vma, start, 0, bv->cmd_size);
> + batch_init(&state, vma, start, bv->state_start, bv->state_size);
> +
> + interface_descriptor =
> + gen7_fill_interface_descriptor(&state, bv,
> + IS_HASWELL(i915) ?
> + &cb_kernel_hsw : &cb_kernel_gen7,
> + desc_count);
> + gen7_emit_pipeline_flush(&cmds);
> + batch_add(&cmds, PIPELINE_SELECT | PIPELINE_SELECT_MEDIA);
> + batch_add(&cmds, MI_NOOP);
> + gen7_emit_state_base_address(&cmds, interface_descriptor);
> + gen7_emit_pipeline_flush(&cmds);
> +
> + gen7_emit_vfe_state(&cmds, bv, urb_size - 1, 0, 0);
> +
> + gen7_emit_interface_descriptor_load(&cmds,
> + interface_descriptor,
> + desc_count);
> +
> + for (i = 0; i < bv->max_primitives; i++)
> + gen7_emit_media_object(&cmds, i);
> +
> + batch_add(&cmds, MI_BATCH_BUFFER_END);
> +}
> +
> +int gen7_setup_clear_gpr_bb(struct intel_engine_cs * const engine,
> + struct i915_vma * const vma)
> +{
> + struct batch_vals bv;
> + u32 *batch;
> +
> + batch_get_defaults(engine->i915, &bv);
> + if (!vma)
> + return bv.max_size;
GEM_BUG_ON(vma->obj->base.size < bv.max_size);
Yeah, we might revisit why this doesn't return the populated vma directly,
since we have the vm available in engine->kernel_context->vm
> +
> + batch = i915_gem_object_pin_map(vma->obj, I915_MAP_WC);
> + if (IS_ERR(batch))
> + return PTR_ERR(batch);
> +
> + emit_batch(vma, memset(batch, 0, bv.max_size), &bv);
> +
> + i915_gem_object_flush_map(vma->obj);
> + i915_gem_object_unpin_map(vma->obj);
> +
> + return 0;
> +}
More information about the Intel-gfx
mailing list