[igt-dev] [PATCH 4/6] tests/amdgpu: add vce decoder test

vitaly.prosyak at amd.com vitaly.prosyak at amd.com
Tue Jul 4 17:45:00 UTC 2023


From: Vitaly Prosyak <vitaly.prosyak at amd.com>

Ported and refactored drmlib vce_tests.c and  the following is done:

1. Remove everywhere global variables.
2. Reuse common functions from amd_mmd_shared.h
3. Properly formatted code to meet igt guidelines.

Signed-off-by: Vitaly Prosyak <vitaly.prosyak at amd.com>
---
 tests/amdgpu/amd_vce_dec.c | 602 +++++++++++++++++++++++++++++++++++++
 tests/amdgpu/meson.build   |   1 +
 2 files changed, 603 insertions(+)
 create mode 100644 tests/amdgpu/amd_vce_dec.c

diff --git a/tests/amdgpu/amd_vce_dec.c b/tests/amdgpu/amd_vce_dec.c
new file mode 100644
index 000000000..364534955
--- /dev/null
+++ b/tests/amdgpu/amd_vce_dec.c
@@ -0,0 +1,602 @@
+// SPDX-License-Identifier: MIT
+// Copyright 2023 Advanced Micro Devices, Inc.
+// Copyright 2014 Advanced Micro Devices, Inc.
+
+#include "lib/amdgpu/amd_mmd_shared.h"
+#include "lib/amdgpu/amd_mmd_vce_ib.h"
+
+#define FW_53_0_03 ((53 << 24) | (0 << 16) | (03 << 8))
+
+struct amdgpu_vce_encode {
+	unsigned int width;
+	unsigned int height;
+	struct amdgpu_mmd_bo vbuf;
+	struct amdgpu_mmd_bo bs[2];
+	struct amdgpu_mmd_bo fb[2];
+	struct amdgpu_mmd_bo cpb;
+	unsigned int ib_len;
+	bool two_instance;
+	struct amdgpu_mmd_bo mvrefbuf;
+	struct amdgpu_mmd_bo mvb;
+	unsigned int mvbuf_size;
+};
+
+
+static bool
+is_vce_tests_enable(amdgpu_device_handle device_handle, uint32_t family_id,
+		uint32_t chip_id, uint32_t chip_rev, bool *is_mv_supported)
+{
+	uint64_t ids_flags;
+	uint32_t version, feature;
+	int r;
+	struct amdgpu_gpu_info gpu_info = {0};
+
+	r = amdgpu_query_gpu_info(device_handle, &gpu_info);
+	igt_assert_eq(r, 0);
+	ids_flags = gpu_info.ids_flags;
+
+	amdgpu_query_firmware_version(device_handle, AMDGPU_INFO_FW_VCE, 0,
+					  0, &version, &feature);
+
+	if (family_id >= AMDGPU_FAMILY_RV || family_id == AMDGPU_FAMILY_SI ||
+			is_gfx_pipe_removed(family_id, chip_id, chip_rev)) {
+		igt_info("\n\nThe ASIC NOT support VCE, tests are disabled\n");
+		return false;
+	}
+
+	if (!(chip_id == (chip_rev + 0x3C) || /* FIJI */
+			chip_id == (chip_rev + 0x50) || /* Polaris 10*/
+			chip_id == (chip_rev + 0x5A) || /* Polaris 11*/
+			chip_id == (chip_rev + 0x64) || /* Polaris 12*/
+			(family_id >= AMDGPU_FAMILY_AI && !ids_flags))) /* dGPU > Polaris */
+		igt_info("\n\nThe ASIC NOT support VCE MV, tests are disabled\n");
+	else if (version < FW_53_0_03)
+		igt_info("\n\nThe ASIC FW version NOT support VCE MV, tests are disabled\n");
+	else
+		*is_mv_supported = true;
+
+	return true;
+}
+
+static void
+amdgpu_cs_vce_create(amdgpu_device_handle device_handle,
+		struct amdgpu_vce_encode *enc, struct mmd_context *context, bool is_mv_supported)
+{
+	unsigned int align = (context->family_id >= AMDGPU_FAMILY_AI) ? 256 : 16;
+	int len, r;
+
+	enc->width = vce_create[6];
+	enc->height = vce_create[7];
+
+	context->num_resources  = 0;
+	alloc_resource(device_handle, &enc->fb[0], 4096, AMDGPU_GEM_DOMAIN_GTT);
+	context->resources[context->num_resources++] = enc->fb[0].handle;
+	context->resources[context->num_resources++] = context->ib_handle;
+
+	len = 0;
+	memcpy(context->ib_cpu, vce_session, sizeof(vce_session));
+	len += sizeof(vce_session) / 4;
+	memcpy((context->ib_cpu + len), vce_taskinfo, sizeof(vce_taskinfo));
+	len += sizeof(vce_taskinfo) / 4;
+	memcpy((context->ib_cpu + len), vce_create, sizeof(vce_create));
+	context->ib_cpu[len + 8] = ALIGN(enc->width, align);
+	context->ib_cpu[len + 9] = ALIGN(enc->width, align);
+	if (is_mv_supported == true) {/* disableTwoInstance */
+		if (context->family_id >= AMDGPU_FAMILY_AI)
+			context->ib_cpu[len + 11] = 0x01000001;
+		else
+			context->ib_cpu[len + 11] = 0x01000201;
+	}
+	len += sizeof(vce_create) / 4;
+	memcpy((context->ib_cpu + len), vce_feedback, sizeof(vce_feedback));
+	context->ib_cpu[len + 2] = enc->fb[0].addr >> 32;
+	context->ib_cpu[len + 3] = enc->fb[0].addr;
+	len += sizeof(vce_feedback) / 4;
+
+	r = submit(device_handle, context, len, AMDGPU_HW_IP_VCE);
+	igt_assert_eq(r, 0);
+
+	free_resource(&enc->fb[0]);
+}
+
+static void
+amdgpu_cs_vce_config(amdgpu_device_handle device_handle,
+		struct mmd_context *context, bool is_mv_supported)
+{
+	int len = 0, r;
+
+	memcpy((context->ib_cpu + len), vce_session, sizeof(vce_session));
+	len += sizeof(vce_session) / 4;
+	memcpy((context->ib_cpu + len), vce_taskinfo, sizeof(vce_taskinfo));
+	context->ib_cpu[len + 3] = 2;
+	context->ib_cpu[len + 6] = 0xffffffff;
+	len += sizeof(vce_taskinfo) / 4;
+	memcpy((context->ib_cpu + len), vce_rate_ctrl, sizeof(vce_rate_ctrl));
+	len += sizeof(vce_rate_ctrl) / 4;
+	memcpy((context->ib_cpu + len), vce_config_ext, sizeof(vce_config_ext));
+	len += sizeof(vce_config_ext) / 4;
+	memcpy((context->ib_cpu + len), vce_motion_est, sizeof(vce_motion_est));
+	len += sizeof(vce_motion_est) / 4;
+	memcpy((context->ib_cpu + len), vce_rdo, sizeof(vce_rdo));
+	len += sizeof(vce_rdo) / 4;
+	memcpy((context->ib_cpu + len), vce_pic_ctrl, sizeof(vce_pic_ctrl));
+	if (is_mv_supported == true)
+		context->ib_cpu[len + 27] = 0x00000001; /* encSliceMode */
+	len += sizeof(vce_pic_ctrl) / 4;
+
+	r = submit(device_handle, context, len, AMDGPU_HW_IP_VCE);
+	igt_assert_eq(r, 0);
+}
+
+static void amdgpu_cs_vce_encode_idr(amdgpu_device_handle device_handle,
+		struct mmd_context *context, struct amdgpu_vce_encode *enc)
+{
+
+	uint64_t luma_offset, chroma_offset;
+	unsigned int align = (context->family_id >= AMDGPU_FAMILY_AI) ? 256 : 16;
+	unsigned int luma_size = ALIGN(enc->width, align) * ALIGN(enc->height, 16);
+	int len = 0, i, r;
+
+	luma_offset = enc->vbuf.addr;
+	chroma_offset = luma_offset + luma_size;
+
+	memcpy((context->ib_cpu + len), vce_session, sizeof(vce_session));
+	len += sizeof(vce_session) / 4;
+	memcpy((context->ib_cpu + len), vce_taskinfo, sizeof(vce_taskinfo));
+	len += sizeof(vce_taskinfo) / 4;
+	memcpy((context->ib_cpu + len), vce_bs_buffer, sizeof(vce_bs_buffer));
+	context->ib_cpu[len + 2] = enc->bs[0].addr >> 32;
+	context->ib_cpu[len + 3] = enc->bs[0].addr;
+	len += sizeof(vce_bs_buffer) / 4;
+	memcpy((context->ib_cpu + len), vce_context_buffer, sizeof(vce_context_buffer));
+	context->ib_cpu[len + 2] = enc->cpb.addr >> 32;
+	context->ib_cpu[len + 3] = enc->cpb.addr;
+	len += sizeof(vce_context_buffer) / 4;
+	memcpy((context->ib_cpu + len), vce_aux_buffer, sizeof(vce_aux_buffer));
+	for (i = 0; i <  8; ++i)
+		context->ib_cpu[len + 2 + i] = luma_size * 1.5 * (i + 2);
+	for (i = 0; i <  8; ++i)
+		context->ib_cpu[len + 10 + i] = luma_size * 1.5;
+	len += sizeof(vce_aux_buffer) / 4;
+	memcpy((context->ib_cpu + len), vce_feedback, sizeof(vce_feedback));
+	context->ib_cpu[len + 2] = enc->fb[0].addr >> 32;
+	context->ib_cpu[len + 3] = enc->fb[0].addr;
+	len += sizeof(vce_feedback) / 4;
+	memcpy((context->ib_cpu + len), vce_encode, sizeof(vce_encode));
+	context->ib_cpu[len + 9] = luma_offset >> 32;
+	context->ib_cpu[len + 10] = luma_offset;
+	context->ib_cpu[len + 11] = chroma_offset >> 32;
+	context->ib_cpu[len + 12] = chroma_offset;
+	context->ib_cpu[len + 14] = ALIGN(enc->width, align);
+	context->ib_cpu[len + 15] = ALIGN(enc->width, align);
+	context->ib_cpu[len + 73] = luma_size * 1.5;
+	context->ib_cpu[len + 74] = luma_size * 2.5;
+	len += sizeof(vce_encode) / 4;
+	enc->ib_len = len;
+	if (!enc->two_instance) {
+		r = submit(device_handle, context, len, AMDGPU_HW_IP_VCE);
+		igt_assert_eq(r, 0);
+	}
+}
+
+static void amdgpu_cs_vce_encode_p(amdgpu_device_handle device_handle,
+		struct mmd_context *context, struct amdgpu_vce_encode *enc)
+{
+	uint64_t luma_offset, chroma_offset;
+	int len, i, r;
+	unsigned int align = (context->family_id >= AMDGPU_FAMILY_AI) ? 256 : 16;
+	unsigned int luma_size = ALIGN(enc->width, align) * ALIGN(enc->height, 16);
+
+	len = (enc->two_instance) ? enc->ib_len : 0;
+	luma_offset = enc->vbuf.addr;
+	chroma_offset = luma_offset + luma_size;
+
+	if (!enc->two_instance) {
+		memcpy((context->ib_cpu + len), vce_session, sizeof(vce_session));
+		len += sizeof(vce_session) / 4;
+	}
+	memcpy((context->ib_cpu + len), vce_taskinfo, sizeof(vce_taskinfo));
+	len += sizeof(vce_taskinfo) / 4;
+	memcpy((context->ib_cpu + len), vce_bs_buffer, sizeof(vce_bs_buffer));
+	context->ib_cpu[len + 2] = enc->bs[1].addr >> 32;
+	context->ib_cpu[len + 3] = enc->bs[1].addr;
+	len += sizeof(vce_bs_buffer) / 4;
+	memcpy((context->ib_cpu + len), vce_context_buffer, sizeof(vce_context_buffer));
+	context->ib_cpu[len + 2] = enc->cpb.addr >> 32;
+	context->ib_cpu[len + 3] = enc->cpb.addr;
+	len += sizeof(vce_context_buffer) / 4;
+	memcpy((context->ib_cpu + len), vce_aux_buffer, sizeof(vce_aux_buffer));
+	for (i = 0; i <  8; ++i)
+		context->ib_cpu[len + 2 + i] = luma_size * 1.5 * (i + 2);
+	for (i = 0; i <  8; ++i)
+		context->ib_cpu[len + 10 + i] = luma_size * 1.5;
+	len += sizeof(vce_aux_buffer) / 4;
+	memcpy((context->ib_cpu + len), vce_feedback, sizeof(vce_feedback));
+	context->ib_cpu[len + 2] = enc->fb[1].addr >> 32;
+	context->ib_cpu[len + 3] = enc->fb[1].addr;
+	len += sizeof(vce_feedback) / 4;
+	memcpy((context->ib_cpu + len), vce_encode, sizeof(vce_encode));
+	context->ib_cpu[len + 2] = 0;
+	context->ib_cpu[len + 9] = luma_offset >> 32;
+	context->ib_cpu[len + 10] = luma_offset;
+	context->ib_cpu[len + 11] = chroma_offset >> 32;
+	context->ib_cpu[len + 12] = chroma_offset;
+	context->ib_cpu[len + 14] = ALIGN(enc->width, align);
+	context->ib_cpu[len + 15] = ALIGN(enc->width, align);
+	context->ib_cpu[len + 18] = 0;
+	context->ib_cpu[len + 19] = 0;
+	context->ib_cpu[len + 56] = 3;
+	context->ib_cpu[len + 57] = 0;
+	context->ib_cpu[len + 58] = 0;
+	context->ib_cpu[len + 59] = luma_size * 1.5;
+	context->ib_cpu[len + 60] = luma_size * 2.5;
+	context->ib_cpu[len + 73] = 0;
+	context->ib_cpu[len + 74] = luma_size;
+	context->ib_cpu[len + 81] = 1;
+	context->ib_cpu[len + 82] = 1;
+	len += sizeof(vce_encode) / 4;
+
+	r = submit(device_handle, context, len, AMDGPU_HW_IP_VCE);
+	igt_assert_eq(r, 0);
+}
+
+static void check_result(struct amdgpu_vce_encode *enc)
+{
+	uint64_t sum;
+	uint32_t s[2] = {180325, 15946};
+	uint32_t *ptr, size;
+	int i, j, r;
+
+	for (i = 0; i < 2; ++i) {
+		r = amdgpu_bo_cpu_map(enc->fb[i].handle, (void **)&enc->fb[i].ptr);
+		igt_assert_eq(r, 0);
+		ptr = (uint32_t *)enc->fb[i].ptr;
+		size = ptr[4] - ptr[9];
+		r = amdgpu_bo_cpu_unmap(enc->fb[i].handle);
+		igt_assert_eq(r, 0);
+		r = amdgpu_bo_cpu_map(enc->bs[i].handle, (void **)&enc->bs[i].ptr);
+		igt_assert_eq(r, 0);
+		for (j = 0, sum = 0; j < size; ++j)
+			sum += enc->bs[i].ptr[j];
+		igt_assert_eq(sum, s[i]);
+		r = amdgpu_bo_cpu_unmap(enc->bs[i].handle);
+		igt_assert_eq(r, 0);
+	}
+}
+
+static void
+amdgpu_cs_vce_encode(amdgpu_device_handle device_handle, struct mmd_context *context,
+		struct amdgpu_vce_encode *enc, bool is_mv_supported)
+{
+	uint32_t vbuf_size, bs_size = 0x154000, cpb_size;
+	unsigned int align = (context->family_id >= AMDGPU_FAMILY_AI) ? 256 : 16;
+	int i, r;
+
+	vbuf_size = ALIGN(enc->width, align) * ALIGN(enc->height, 16) * 1.5;
+	cpb_size = vbuf_size * 10;
+	context->num_resources = 0;
+	alloc_resource(device_handle, &enc->fb[0], 4096, AMDGPU_GEM_DOMAIN_GTT);
+	context->resources[context->num_resources++] = enc->fb[0].handle;
+	alloc_resource(device_handle, &enc->fb[1], 4096, AMDGPU_GEM_DOMAIN_GTT);
+	context->resources[context->num_resources++] = enc->fb[1].handle;
+	alloc_resource(device_handle, &enc->bs[0], bs_size, AMDGPU_GEM_DOMAIN_GTT);
+	context->resources[context->num_resources++] = enc->bs[0].handle;
+	alloc_resource(device_handle, &enc->bs[1], bs_size, AMDGPU_GEM_DOMAIN_GTT);
+	context->resources[context->num_resources++] = enc->bs[1].handle;
+	alloc_resource(device_handle, &enc->vbuf, vbuf_size, AMDGPU_GEM_DOMAIN_VRAM);
+	context->resources[context->num_resources++] = enc->vbuf.handle;
+	alloc_resource(device_handle, &enc->cpb, cpb_size, AMDGPU_GEM_DOMAIN_VRAM);
+	context->resources[context->num_resources++] = enc->cpb.handle;
+	context->resources[context->num_resources++] = context->ib_handle;
+
+	r = amdgpu_bo_cpu_map(enc->vbuf.handle, (void **)&enc->vbuf.ptr);
+	igt_assert_eq(r, 0);
+
+	memset(enc->vbuf.ptr, 0, vbuf_size);
+	for (i = 0; i < enc->height; ++i) {
+		memcpy(enc->vbuf.ptr, (frame + i * enc->width), enc->width);
+		enc->vbuf.ptr += ALIGN(enc->width, align);
+	}
+	for (i = 0; i < enc->height / 2; ++i) {
+		memcpy(enc->vbuf.ptr, ((frame + enc->height * enc->width) +
+				i * enc->width), enc->width);
+		enc->vbuf.ptr += ALIGN(enc->width, align);
+	}
+
+	r = amdgpu_bo_cpu_unmap(enc->vbuf.handle);
+	igt_assert_eq(r, 0);
+
+	amdgpu_cs_vce_config(device_handle, context, is_mv_supported);
+
+	if (context->family_id >= AMDGPU_FAMILY_VI) {
+		vce_taskinfo[3] = 3;
+		amdgpu_cs_vce_encode_idr(device_handle, context, enc);
+		amdgpu_cs_vce_encode_p(device_handle, context, enc);
+		check_result(enc);
+
+		/* two pipes */
+		vce_encode[16] = 0;
+		amdgpu_cs_vce_encode_idr(device_handle, context, enc);
+		amdgpu_cs_vce_encode_p(device_handle, context, enc);
+		check_result(enc);
+
+		/* two instances */
+		if (context->vce_harvest_config == 0) {
+			enc->two_instance = true;
+			vce_taskinfo[2] = 0x83;
+			vce_taskinfo[4] = 1;
+			amdgpu_cs_vce_encode_idr(device_handle, context, enc);
+			vce_taskinfo[2] = 0xffffffff;
+			vce_taskinfo[4] = 2;
+			amdgpu_cs_vce_encode_p(device_handle, context, enc);
+			check_result(enc);
+		}
+	} else {
+		vce_taskinfo[3] = 3;
+		vce_encode[16] = 0;
+		amdgpu_cs_vce_encode_idr(device_handle, context, enc);
+		amdgpu_cs_vce_encode_p(device_handle, context, enc);
+		check_result(enc);
+	}
+
+	free_resource(&enc->fb[0]);
+	free_resource(&enc->fb[1]);
+	free_resource(&enc->bs[0]);
+	free_resource(&enc->bs[1]);
+	free_resource(&enc->vbuf);
+	free_resource(&enc->cpb);
+}
+
+static void amdgpu_cs_vce_mv(amdgpu_device_handle device_handle,
+		struct mmd_context *context, struct amdgpu_vce_encode *enc)
+{
+	uint64_t luma_offset, chroma_offset;
+	uint64_t mv_ref_luma_offset;
+	unsigned int align = (context->family_id >= AMDGPU_FAMILY_AI) ? 256 : 16;
+	unsigned int luma_size = ALIGN(enc->width, align) * ALIGN(enc->height, 16);
+	int len = 0, i, r;
+
+	luma_offset = enc->vbuf.addr;
+	chroma_offset = luma_offset + luma_size;
+	mv_ref_luma_offset = enc->mvrefbuf.addr;
+
+	memcpy((context->ib_cpu + len), vce_session, sizeof(vce_session));
+	len += sizeof(vce_session) / 4;
+	memcpy((context->ib_cpu + len), vce_taskinfo, sizeof(vce_taskinfo));
+	len += sizeof(vce_taskinfo) / 4;
+	memcpy((context->ib_cpu + len), vce_bs_buffer, sizeof(vce_bs_buffer));
+	context->ib_cpu[len + 2] = enc->bs[0].addr >> 32;
+	context->ib_cpu[len + 3] = enc->bs[0].addr;
+	len += sizeof(vce_bs_buffer) / 4;
+	memcpy((context->ib_cpu + len), vce_context_buffer, sizeof(vce_context_buffer));
+	context->ib_cpu[len + 2] = enc->cpb.addr >> 32;
+	context->ib_cpu[len + 3] = enc->cpb.addr;
+	len += sizeof(vce_context_buffer) / 4;
+	memcpy((context->ib_cpu + len), vce_aux_buffer, sizeof(vce_aux_buffer));
+	for (i = 0; i <  8; ++i)
+		context->ib_cpu[len + 2 + i] = luma_size * 1.5 * (i + 2);
+	for (i = 0; i <  8; ++i)
+		context->ib_cpu[len + 10 + i] = luma_size * 1.5;
+	len += sizeof(vce_aux_buffer) / 4;
+	memcpy((context->ib_cpu + len), vce_feedback, sizeof(vce_feedback));
+	context->ib_cpu[len + 2] = enc->fb[0].addr >> 32;
+	context->ib_cpu[len + 3] = enc->fb[0].addr;
+	len += sizeof(vce_feedback) / 4;
+	memcpy((context->ib_cpu + len), vce_mv_buffer, sizeof(vce_mv_buffer));
+	context->ib_cpu[len + 2] = mv_ref_luma_offset >> 32;
+	context->ib_cpu[len + 3] = mv_ref_luma_offset;
+	context->ib_cpu[len + 4] = ALIGN(enc->width, align);
+	context->ib_cpu[len + 5] = ALIGN(enc->width, align);
+	context->ib_cpu[len + 6] = luma_size;
+	context->ib_cpu[len + 7] = enc->mvb.addr >> 32;
+	context->ib_cpu[len + 8] = enc->mvb.addr;
+	len += sizeof(vce_mv_buffer) / 4;
+	memcpy((context->ib_cpu + len), vce_encode, sizeof(vce_encode));
+	context->ib_cpu[len + 2] = 0;
+	context->ib_cpu[len + 3] = 0;
+	context->ib_cpu[len + 4] = 0x154000;
+	context->ib_cpu[len + 9] = luma_offset >> 32;
+	context->ib_cpu[len + 10] = luma_offset;
+	context->ib_cpu[len + 11] = chroma_offset >> 32;
+	context->ib_cpu[len + 12] = chroma_offset;
+	context->ib_cpu[len + 13] = ALIGN(enc->height, 16);
+	context->ib_cpu[len + 14] = ALIGN(enc->width, align);
+	context->ib_cpu[len + 15] = ALIGN(enc->width, align);
+	/* encDisableMBOffloading-encDisableTwoPipeMode-encInputPicArrayMode-encInputPicAddrMode */
+	context->ib_cpu[len + 16] = 0x01010000;
+	context->ib_cpu[len + 18] = 0; /* encPicType */
+	context->ib_cpu[len + 19] = 0; /* encIdrFlag */
+	context->ib_cpu[len + 20] = 0; /* encIdrPicId */
+	context->ib_cpu[len + 21] = 0; /* encMGSKeyPic */
+	context->ib_cpu[len + 22] = 0; /* encReferenceFlag */
+	context->ib_cpu[len + 23] = 0; /* encTemporalLayerIndex */
+	context->ib_cpu[len + 55] = 0; /* pictureStructure */
+	context->ib_cpu[len + 56] = 0; /* encPicType -ref[0] */
+	context->ib_cpu[len + 61] = 0; /* pictureStructure */
+	context->ib_cpu[len + 62] = 0; /* encPicType -ref[1] */
+	context->ib_cpu[len + 67] = 0; /* pictureStructure */
+	context->ib_cpu[len + 68] = 0; /* encPicType -ref1 */
+	context->ib_cpu[len + 81] = 1; /* frameNumber */
+	context->ib_cpu[len + 82] = 2; /* pictureOrderCount */
+	context->ib_cpu[len + 83] = 0xffffffff; /* numIPicRemainInRCGOP */
+	context->ib_cpu[len + 84] = 0xffffffff; /* numPPicRemainInRCGOP */
+	context->ib_cpu[len + 85] = 0xffffffff; /* numBPicRemainInRCGOP */
+	context->ib_cpu[len + 86] = 0xffffffff; /* numIRPicRemainInRCGOP */
+	context->ib_cpu[len + 87] = 0; /* remainedIntraRefreshPictures */
+	len += sizeof(vce_encode) / 4;
+
+	enc->ib_len = len;
+	r = submit(device_handle, context, len, AMDGPU_HW_IP_VCE);
+	igt_assert_eq(r, 0);
+}
+
+static void check_mv_result(struct amdgpu_vce_encode *enc)
+{
+	uint64_t sum;
+	uint32_t s = 140790;
+	int j, r;
+
+	r = amdgpu_bo_cpu_map(enc->fb[0].handle, (void **)&enc->fb[0].ptr);
+	igt_assert_eq(r, 0);
+	r = amdgpu_bo_cpu_unmap(enc->fb[0].handle);
+	igt_assert_eq(r, 0);
+	r = amdgpu_bo_cpu_map(enc->mvb.handle, (void **)&enc->mvb.ptr);
+	igt_assert_eq(r, 0);
+	for (j = 0, sum = 0; j < enc->mvbuf_size; ++j)
+		sum += enc->mvb.ptr[j];
+	igt_assert_eq(sum, s);
+	r = amdgpu_bo_cpu_unmap(enc->mvb.handle);
+	igt_assert_eq(r, 0);
+}
+
+static void
+amdgpu_cs_vce_encode_mv(amdgpu_device_handle device_handle, struct mmd_context *context,
+		struct amdgpu_vce_encode *enc, bool is_mv_supported)
+{
+	uint32_t vbuf_size, bs_size = 0x154000, cpb_size;
+	unsigned int align = (context->family_id >= AMDGPU_FAMILY_AI) ? 256 : 16;
+	int i, r;
+
+	vbuf_size = ALIGN(enc->width, align) * ALIGN(enc->height, 16) * 1.5;
+	enc->mvbuf_size = ALIGN(enc->width, 16) * ALIGN(enc->height, 16) / 8;
+	cpb_size = vbuf_size * 10;
+	context->num_resources = 0;
+	alloc_resource(device_handle, &enc->fb[0], 4096, AMDGPU_GEM_DOMAIN_GTT);
+	context->resources[context->num_resources++] = enc->fb[0].handle;
+	alloc_resource(device_handle, &enc->bs[0], bs_size, AMDGPU_GEM_DOMAIN_GTT);
+	context->resources[context->num_resources++] = enc->bs[0].handle;
+	alloc_resource(device_handle, &enc->mvb, enc->mvbuf_size, AMDGPU_GEM_DOMAIN_GTT);
+	context->resources[context->num_resources++] = enc->mvb.handle;
+	alloc_resource(device_handle, &enc->vbuf, vbuf_size, AMDGPU_GEM_DOMAIN_VRAM);
+	context->resources[context->num_resources++] = enc->vbuf.handle;
+	alloc_resource(device_handle, &enc->mvrefbuf, vbuf_size, AMDGPU_GEM_DOMAIN_VRAM);
+	context->resources[context->num_resources++] = enc->mvrefbuf.handle;
+	alloc_resource(device_handle, &enc->cpb, cpb_size, AMDGPU_GEM_DOMAIN_VRAM);
+	context->resources[context->num_resources++] = enc->cpb.handle;
+	context->resources[context->num_resources++] = context->ib_handle;
+
+	r = amdgpu_bo_cpu_map(enc->vbuf.handle, (void **)&enc->vbuf.ptr);
+	igt_assert_eq(r, 0);
+
+	memset(enc->vbuf.ptr, 0, vbuf_size);
+	for (i = 0; i < enc->height; ++i) {
+		memcpy(enc->vbuf.ptr, (frame + i * enc->width), enc->width);
+		enc->vbuf.ptr += ALIGN(enc->width, align);
+	}
+	for (i = 0; i < enc->height / 2; ++i) {
+		memcpy(enc->vbuf.ptr, ((frame + enc->height * enc->width) + i * enc->width), enc->width);
+		enc->vbuf.ptr += ALIGN(enc->width, align);
+	}
+
+	r = amdgpu_bo_cpu_unmap(enc->vbuf.handle);
+	igt_assert_eq(r, 0);
+
+	r = amdgpu_bo_cpu_map(enc->mvrefbuf.handle, (void **)&enc->mvrefbuf.ptr);
+	igt_assert_eq(r, 0);
+
+	memset(enc->mvrefbuf.ptr, 0, vbuf_size);
+	for (i = 0; i < enc->height; ++i) {
+		memcpy(enc->mvrefbuf.ptr, (frame + (enc->height - i - 1) * enc->width), enc->width);
+		enc->mvrefbuf.ptr += ALIGN(enc->width, align);
+	}
+	for (i = 0; i < enc->height / 2; ++i) {
+		memcpy(enc->mvrefbuf.ptr,
+		((frame + enc->height * enc->width) + (enc->height / 2 - i - 1) * enc->width), enc->width);
+		enc->mvrefbuf.ptr += ALIGN(enc->width, align);
+	}
+
+	r = amdgpu_bo_cpu_unmap(enc->mvrefbuf.handle);
+	igt_assert_eq(r, 0);
+
+	amdgpu_cs_vce_config(device_handle, context, is_mv_supported);
+
+	vce_taskinfo[3] = 3;
+	amdgpu_cs_vce_mv(device_handle, context, enc);
+	check_mv_result(enc);
+
+	free_resource(&enc->fb[0]);
+	free_resource(&enc->bs[0]);
+	free_resource(&enc->vbuf);
+	free_resource(&enc->cpb);
+	free_resource(&enc->mvrefbuf);
+	free_resource(&enc->mvb);
+}
+
+static void
+amdgpu_cs_vce_destroy(amdgpu_device_handle device_handle, struct mmd_context *context,
+		struct amdgpu_vce_encode *enc)
+{
+	int len, r;
+
+	context->num_resources  = 0;
+	alloc_resource(device_handle, &enc->fb[0], 4096, AMDGPU_GEM_DOMAIN_GTT);
+	context->resources[context->num_resources++] = enc->fb[0].handle;
+	context->resources[context->num_resources++] = context->ib_handle;
+
+	len = 0;
+	memcpy(context->ib_cpu, vce_session, sizeof(vce_session));
+	len += sizeof(vce_session) / 4;
+	memcpy((context->ib_cpu + len), vce_taskinfo, sizeof(vce_taskinfo));
+	context->ib_cpu[len + 3] = 1;
+	len += sizeof(vce_taskinfo) / 4;
+	memcpy((context->ib_cpu + len), vce_feedback, sizeof(vce_feedback));
+	context->ib_cpu[len + 2] = enc->fb[0].addr >> 32;
+	context->ib_cpu[len + 3] = enc->fb[0].addr;
+	len += sizeof(vce_feedback) / 4;
+	memcpy((context->ib_cpu + len), vce_destroy, sizeof(vce_destroy));
+	len += sizeof(vce_destroy) / 4;
+
+	r = submit(device_handle, context, len, AMDGPU_HW_IP_VCE);
+	igt_assert_eq(r, 0);
+
+	free_resource(&enc->fb[0]);
+}
+
+igt_main
+{
+	amdgpu_device_handle device;
+	struct mmd_context context = {};
+	struct amdgpu_vce_encode enc = {};
+	int fd = -1;
+	bool is_mv_supported = false;
+
+	igt_fixture {
+		uint32_t major, minor;
+		int err;
+
+		fd = drm_open_driver(DRIVER_AMDGPU);
+		err = amdgpu_device_initialize(fd, &major, &minor, &device);
+		igt_require(err == 0);
+		igt_info("Initialized amdgpu, driver version %d.%d\n",
+			 major, minor);
+		err = mmd_context_init(device, &context);
+		igt_require(err == 0);
+		igt_skip_on(!is_vce_tests_enable(device, context.family_id, context.chip_id,
+				context.chip_rev, &is_mv_supported));
+	}
+	igt_describe("Test whether vce enc is created");
+	igt_subtest("amdgpu_cs_vce_create")
+	amdgpu_cs_vce_create(device, &enc, &context, is_mv_supported);
+
+	igt_describe("Test whether vce enc encodes");
+	igt_subtest("amdgpu_cs_vce_encode")
+	amdgpu_cs_vce_encode(device, &context, &enc,  is_mv_supported);
+
+	if (is_mv_supported) {
+		igt_describe("Test whether vce enc encodes mv");
+		igt_subtest("amdgpu_cs_vce_encode_mv")
+		amdgpu_cs_vce_encode_mv(device, &context, &enc, is_mv_supported);
+	}
+
+	igt_describe("Test whether vce enc is destroyed");
+	igt_subtest("amdgpu_cs_vce_destroy")
+	amdgpu_cs_vce_destroy(device, &context, &enc);
+
+
+	igt_fixture {
+		mmd_context_clean(device, &context);
+		amdgpu_device_deinitialize(device);
+		drm_close_driver(fd);
+	}
+
+}
diff --git a/tests/amdgpu/meson.build b/tests/amdgpu/meson.build
index 576333aab..e6e6eaabf 100644
--- a/tests/amdgpu/meson.build
+++ b/tests/amdgpu/meson.build
@@ -26,6 +26,7 @@ if libdrm_amdgpu.found()
 			  'amd_psr',
 			  'amd_uvd_dec',
 			  'amd_uvd_enc',
+			  'amd_vce_dec',
 			  'amd_vrr_range',
 			]
 	amdgpu_deps += libdrm_amdgpu
-- 
2.25.1



More information about the igt-dev mailing list