[igt-dev] [PATCH] tests/amdgpu: add vcn decoder and encoder tests
vitaly.prosyak at amd.com
vitaly.prosyak at amd.com
Fri Jul 7 13:09:00 UTC 2023
From: Vitaly Prosyak <vitaly.prosyak at amd.com>
Ported and refactored drmlib vcn_tests.c and the following is done:
1. Remove everywhere global variables.
2. Reuse common functions from amd_mmd_shared.h
3. Properly formatted code to meet igt guidelines.
Signed-off-by: Vitaly Prosyak <vitaly.prosyak at amd.com>
Acked-by: Leo Liu <leo.liu at amd.com>
---
tests/amdgpu/amd_vcn.c | 1465 ++++++++++++++++++++++++++++++++++++++
tests/amdgpu/meson.build | 1 +
2 files changed, 1466 insertions(+)
create mode 100644 tests/amdgpu/amd_vcn.c
diff --git a/tests/amdgpu/amd_vcn.c b/tests/amdgpu/amd_vcn.c
new file mode 100644
index 000000000..b0601702b
--- /dev/null
+++ b/tests/amdgpu/amd_vcn.c
@@ -0,0 +1,1465 @@
+// SPDX-License-Identifier: MIT
+// Copyright 2023 Advanced Micro Devices, Inc.
+// Copyright 2014 Advanced Micro Devices, Inc.
+
+#include "lib/amdgpu/amd_mmd_shared.h"
+
+#define DECODE_CMD_MSG_BUFFER 0x00000000
+#define DECODE_CMD_DPB_BUFFER 0x00000001
+#define DECODE_CMD_DECODING_TARGET_BUFFER 0x00000002
+#define DECODE_CMD_FEEDBACK_BUFFER 0x00000003
+#define DECODE_CMD_PROB_TBL_BUFFER 0x00000004
+#define DECODE_CMD_SESSION_CONTEXT_BUFFER 0x00000005
+#define DECODE_CMD_BITSTREAM_BUFFER 0x00000100
+#define DECODE_CMD_IT_SCALING_TABLE_BUFFER 0x00000204
+#define DECODE_CMD_CONTEXT_BUFFER 0x00000206
+
+#define DECODE_IB_PARAM_DECODE_BUFFER (0x00000001)
+
+#define DECODE_CMDBUF_FLAGS_MSG_BUFFER (0x00000001)
+#define DECODE_CMDBUF_FLAGS_DPB_BUFFER (0x00000002)
+#define DECODE_CMDBUF_FLAGS_BITSTREAM_BUFFER (0x00000004)
+#define DECODE_CMDBUF_FLAGS_DECODING_TARGET_BUFFER (0x00000008)
+#define DECODE_CMDBUF_FLAGS_FEEDBACK_BUFFER (0x00000010)
+#define DECODE_CMDBUF_FLAGS_IT_SCALING_BUFFER (0x00000200)
+#define DECODE_CMDBUF_FLAGS_CONTEXT_BUFFER (0x00000800)
+#define DECODE_CMDBUF_FLAGS_PROB_TBL_BUFFER (0x00001000)
+#define DECODE_CMDBUF_FLAGS_SESSION_CONTEXT_BUFFER (0x00100000)
+
+#define H264_NAL_TYPE_NON_IDR_SLICE 1
+#define H264_NAL_TYPE_DP_A_SLICE 2
+#define H264_NAL_TYPE_DP_B_SLICE 3
+#define H264_NAL_TYPE_DP_C_SLICE 0x4
+#define H264_NAL_TYPE_IDR_SLICE 0x5
+#define H264_NAL_TYPE_SEI 0x6
+#define H264_NAL_TYPE_SEQ_PARAM 0x7
+#define H264_NAL_TYPE_PIC_PARAM 0x8
+#define H264_NAL_TYPE_ACCESS_UNIT 0x9
+#define H264_NAL_TYPE_END_OF_SEQ 0xa
+#define H264_NAL_TYPE_END_OF_STREAM 0xb
+#define H264_NAL_TYPE_FILLER_DATA 0xc
+#define H264_NAL_TYPE_SEQ_EXTENSION 0xd
+
+#define H264_START_CODE 0x000001
+
+
+struct rvcn_decode_buffer {
+ unsigned int valid_buf_flag;
+ unsigned int msg_buffer_address_hi;
+ unsigned int msg_buffer_address_lo;
+ unsigned int dpb_buffer_address_hi;
+ unsigned int dpb_buffer_address_lo;
+ unsigned int target_buffer_address_hi;
+ unsigned int target_buffer_address_lo;
+ unsigned int session_contex_buffer_address_hi;
+ unsigned int session_contex_buffer_address_lo;
+ unsigned int bitstream_buffer_address_hi;
+ unsigned int bitstream_buffer_address_lo;
+ unsigned int context_buffer_address_hi;
+ unsigned int context_buffer_address_lo;
+ unsigned int feedback_buffer_address_hi;
+ unsigned int feedback_buffer_address_lo;
+ unsigned int luma_hist_buffer_address_hi;
+ unsigned int luma_hist_buffer_address_lo;
+ unsigned int prob_tbl_buffer_address_hi;
+ unsigned int prob_tbl_buffer_address_lo;
+ unsigned int sclr_coeff_buffer_address_hi;
+ unsigned int sclr_coeff_buffer_address_lo;
+ unsigned int it_sclr_table_buffer_address_hi;
+ unsigned int it_sclr_table_buffer_address_lo;
+ unsigned int sclr_target_buffer_address_hi;
+ unsigned int sclr_target_buffer_address_lo;
+ unsigned int cenc_size_info_buffer_address_hi;
+ unsigned int cenc_size_info_buffer_address_lo;
+ unsigned int mpeg2_pic_param_buffer_address_hi;
+ unsigned int mpeg2_pic_param_buffer_address_lo;
+ unsigned int mpeg2_mb_control_buffer_address_hi;
+ unsigned int mpeg2_mb_control_buffer_address_lo;
+ unsigned int mpeg2_idct_coeff_buffer_address_hi;
+ unsigned int mpeg2_idct_coeff_buffer_address_lo;
+};
+
+struct rvcn_decode_ib_package {
+ unsigned int package_size;
+ unsigned int package_type;
+};
+
+
+struct amdgpu_vcn_reg {
+ uint32_t data0;
+ uint32_t data1;
+ uint32_t cmd;
+ uint32_t nop;
+ uint32_t cntl;
+};
+
+struct bufferInfo {
+ uint32_t numOfBitsInBuffer;
+ const uint8_t *decBuffer;
+ uint8_t decData;
+ uint32_t decBufferSize;
+ const uint8_t *end;
+};
+
+struct h264_decode {
+ uint8_t profile;
+ uint8_t level_idc;
+ uint8_t nal_ref_idc;
+ uint8_t nal_unit_type;
+ uint32_t pic_width, pic_height;
+ uint32_t slice_type;
+};
+
+struct vcn_context {
+ struct amdgpu_mmd_bo enc_buf;
+ struct amdgpu_mmd_bo cpb_buf;
+ uint32_t enc_task_id;
+ uint32_t *ib_checksum;
+ uint32_t *ib_size_in_dw;
+ uint32_t gWidth, gHeight, gSliceType;
+ struct rvcn_decode_buffer *decode_buffer;
+};
+
+static struct amdgpu_vcn_reg reg[] = {
+ {0x81c4, 0x81c5, 0x81c3, 0x81ff, 0x81c6},
+ {0x504, 0x505, 0x503, 0x53f, 0x506},
+ {0x10, 0x11, 0xf, 0x29, 0x26d},
+};
+
+static uint32_t
+bs_read_u(struct bufferInfo *bufinfo, int n);
+
+static uint32_t
+bs_read_ue(struct bufferInfo *bufinfo);
+
+static bool
+is_vcn_tests_enable(amdgpu_device_handle device_handle, struct mmd_context *context)
+{
+ struct drm_amdgpu_info_hw_ip info;
+ int r;
+
+ r = amdgpu_query_hw_ip_info(device_handle, AMDGPU_HW_IP_VCN_ENC, 0, &info);
+
+ if (r)
+ return false;
+
+ context->vcn_ip_version_major = info.hw_ip_version_major;
+ context->vcn_ip_version_minor = info.hw_ip_version_minor;
+ context->enc_ring = !!info.available_rings;
+ /* in vcn 4.0 it re-uses encoding queue as unified queue */
+ if (context->vcn_ip_version_major >= 4) {
+ context->vcn_unified_ring = true;
+ context->vcn_dec_sw_ring = true;
+ context->dec_ring = context->enc_ring;
+ } else {
+ r = amdgpu_query_hw_ip_info(device_handle, AMDGPU_HW_IP_VCN_DEC, 0, &info);
+ context->dec_ring = !!info.available_rings;
+ }
+
+ if (!(context->dec_ring || context->enc_ring) ||
+ (context->family_id < AMDGPU_FAMILY_RV &&
+ (context->family_id == AMDGPU_FAMILY_AI &&
+ (context->chip_id - context->chip_rev) < 0x32))) { /* Arcturus */
+ igt_info("The ASIC does NOT support VCN, vcn test is disabled\n");
+ return false;
+ }
+
+ if (context->family_id == AMDGPU_FAMILY_AI)
+ context->enc_ring = false;
+
+ if (!context->dec_ring) {
+ igt_info("VCN Tests DEC create disable");
+ igt_info("VCN Tests DEC decode disable");
+ igt_info("VCN Tests DEC destroy disable");
+ }
+
+ if (!context->enc_ring) {
+ igt_info("VCN Tests ENC create disable");
+ igt_info("VCN Tests ENC encode disable");
+ igt_info("VCN Tests ENC destroy disable");
+ }
+
+ if (context->vcn_ip_version_major == 1) {
+ context->vcn_reg_index = 0;
+ } else if (context->vcn_ip_version_major == 2 && context->vcn_ip_version_minor == 0) {
+ context->vcn_reg_index = 1;
+ } else if ((context->vcn_ip_version_major == 2 && context->vcn_ip_version_minor >= 5) ||
+ context->vcn_ip_version_major == 3) {
+ context->vcn_reg_index = 2;
+ }
+
+ return true;
+}
+
+static void
+amdgpu_cs_sq_head(struct vcn_context *v_context, uint32_t *base, int *offset, bool enc)
+{
+ /* signature */
+ *(base + (*offset)++) = 0x00000010;
+ *(base + (*offset)++) = 0x30000002;
+ v_context->ib_checksum = base + (*offset)++;
+ v_context->ib_size_in_dw = base + (*offset)++;
+
+ /* engine info */
+ *(base + (*offset)++) = 0x00000010;
+ *(base + (*offset)++) = 0x30000001;
+ *(base + (*offset)++) = enc ? 2 : 3;
+ *(base + (*offset)++) = 0x00000000;
+}
+
+static void
+amdgpu_cs_sq_ib_tail(struct vcn_context *v_context, uint32_t *end)
+{
+ uint32_t size_in_dw;
+ uint32_t checksum = 0;
+
+ /* if the pointers are invalid, no need to process */
+ if (v_context->ib_checksum == NULL || v_context->ib_size_in_dw == NULL)
+ return;
+
+ size_in_dw = end - v_context->ib_size_in_dw - 1;
+ *v_context->ib_size_in_dw = size_in_dw;
+ *(v_context->ib_size_in_dw + 4) = size_in_dw * sizeof(uint32_t);
+
+ for (int i = 0; i < size_in_dw; i++)
+ checksum += *(v_context->ib_checksum + 2 + i);
+
+ *v_context->ib_checksum = checksum;
+
+ v_context->ib_checksum = NULL;
+ v_context->ib_size_in_dw = NULL;
+}
+
+static void
+vcn_dec_cmd(struct mmd_context *context, struct vcn_context *v_context,
+ uint64_t addr, unsigned int cmd, int *idx)
+{
+ if (context->vcn_dec_sw_ring == false) {
+ context->ib_cpu[(*idx)++] = reg[context->vcn_reg_index].data0;
+ context->ib_cpu[(*idx)++] = addr;
+ context->ib_cpu[(*idx)++] = reg[context->vcn_reg_index].data1;
+ context->ib_cpu[(*idx)++] = addr >> 32;
+ context->ib_cpu[(*idx)++] = reg[context->vcn_reg_index].cmd;
+ context->ib_cpu[(*idx)++] = cmd << 1;
+ return;
+ }
+
+ /* Support decode software ring message */
+ if (!(*idx)) {
+ struct rvcn_decode_ib_package *ib_header;
+
+ if (context->vcn_unified_ring)
+ amdgpu_cs_sq_head(v_context, context->ib_cpu, idx, false);
+
+ ib_header = (struct rvcn_decode_ib_package *)&context->ib_cpu[*idx];
+ ib_header->package_size = sizeof(struct rvcn_decode_buffer) +
+ sizeof(struct rvcn_decode_ib_package);
+
+ (*idx)++;
+ ib_header->package_type = (DECODE_IB_PARAM_DECODE_BUFFER);
+ (*idx)++;
+
+ v_context->decode_buffer = (struct rvcn_decode_buffer *)&(context->ib_cpu[*idx]);
+ *idx += sizeof(struct rvcn_decode_buffer) / 4;
+ memset(v_context->decode_buffer, 0, sizeof(struct rvcn_decode_buffer));
+ }
+
+ switch (cmd) {
+ case DECODE_CMD_MSG_BUFFER:
+ v_context->decode_buffer->valid_buf_flag |= DECODE_CMDBUF_FLAGS_MSG_BUFFER;
+ v_context->decode_buffer->msg_buffer_address_hi = (addr >> 32);
+ v_context->decode_buffer->msg_buffer_address_lo = (addr);
+ break;
+ case DECODE_CMD_DPB_BUFFER:
+ v_context->decode_buffer->valid_buf_flag |= (DECODE_CMDBUF_FLAGS_DPB_BUFFER);
+ v_context->decode_buffer->dpb_buffer_address_hi = (addr >> 32);
+ v_context->decode_buffer->dpb_buffer_address_lo = (addr);
+ break;
+ case DECODE_CMD_DECODING_TARGET_BUFFER:
+ v_context->decode_buffer->valid_buf_flag |= (DECODE_CMDBUF_FLAGS_DECODING_TARGET_BUFFER);
+ v_context->decode_buffer->target_buffer_address_hi = (addr >> 32);
+ v_context->decode_buffer->target_buffer_address_lo = (addr);
+ break;
+ case DECODE_CMD_FEEDBACK_BUFFER:
+ v_context->decode_buffer->valid_buf_flag |= (DECODE_CMDBUF_FLAGS_FEEDBACK_BUFFER);
+ v_context->decode_buffer->feedback_buffer_address_hi = (addr >> 32);
+ v_context->decode_buffer->feedback_buffer_address_lo = (addr);
+ break;
+ case DECODE_CMD_PROB_TBL_BUFFER:
+ v_context->decode_buffer->valid_buf_flag |= (DECODE_CMDBUF_FLAGS_PROB_TBL_BUFFER);
+ v_context->decode_buffer->prob_tbl_buffer_address_hi = (addr >> 32);
+ v_context->decode_buffer->prob_tbl_buffer_address_lo = (addr);
+ break;
+ case DECODE_CMD_SESSION_CONTEXT_BUFFER:
+ v_context->decode_buffer->valid_buf_flag |= (DECODE_CMDBUF_FLAGS_SESSION_CONTEXT_BUFFER);
+ v_context->decode_buffer->session_contex_buffer_address_hi = (addr >> 32);
+ v_context->decode_buffer->session_contex_buffer_address_lo = (addr);
+ break;
+ case DECODE_CMD_BITSTREAM_BUFFER:
+ v_context->decode_buffer->valid_buf_flag |= (DECODE_CMDBUF_FLAGS_BITSTREAM_BUFFER);
+ v_context->decode_buffer->bitstream_buffer_address_hi = (addr >> 32);
+ v_context->decode_buffer->bitstream_buffer_address_lo = (addr);
+ break;
+ case DECODE_CMD_IT_SCALING_TABLE_BUFFER:
+ v_context->decode_buffer->valid_buf_flag |= (DECODE_CMDBUF_FLAGS_IT_SCALING_BUFFER);
+ v_context->decode_buffer->it_sclr_table_buffer_address_hi = (addr >> 32);
+ v_context->decode_buffer->it_sclr_table_buffer_address_lo = (addr);
+ break;
+ case DECODE_CMD_CONTEXT_BUFFER:
+ v_context->decode_buffer->valid_buf_flag |= (DECODE_CMDBUF_FLAGS_CONTEXT_BUFFER);
+ v_context->decode_buffer->context_buffer_address_hi = (addr >> 32);
+ v_context->decode_buffer->context_buffer_address_lo = (addr);
+ break;
+ default:
+ igt_info("Not Supported!\n");
+ }
+}
+
+static void
+amdgpu_cs_vcn_dec_create(amdgpu_device_handle device_handle, struct mmd_context *context,
+ struct vcn_context *v_context)
+{
+ struct amdgpu_mmd_bo msg_buf;
+ unsigned int ip;
+ int len, r;
+
+ context->num_resources = 0;
+ alloc_resource(device_handle, &msg_buf, 4096, AMDGPU_GEM_DOMAIN_GTT);
+ context->resources[context->num_resources++] = msg_buf.handle;
+ context->resources[context->num_resources++] = context->ib_handle;
+ r = amdgpu_bo_cpu_map(msg_buf.handle, (void **)&msg_buf.ptr);
+ igt_assert_eq(r, 0);
+
+ memset(msg_buf.ptr, 0, 4096);
+ memcpy(msg_buf.ptr, vcn_dec_create_msg, sizeof(vcn_dec_create_msg));
+
+ len = 0;
+ if (context->vcn_dec_sw_ring == true)
+ vcn_dec_cmd(context, v_context, msg_buf.addr, 0, &len);
+ else {
+ context->ib_cpu[len++] = reg[context->vcn_reg_index].data0;
+ context->ib_cpu[len++] = msg_buf.addr;
+ context->ib_cpu[len++] = reg[context->vcn_reg_index].data1;
+ context->ib_cpu[len++] = msg_buf.addr >> 32;
+ context->ib_cpu[len++] = reg[context->vcn_reg_index].cmd;
+ context->ib_cpu[len++] = 0;
+ for (; len % 16; ) {
+ context->ib_cpu[len++] = reg[context->vcn_reg_index].nop;
+ context->ib_cpu[len++] = 0;
+ }
+ }
+
+ if (context->vcn_unified_ring) {
+ amdgpu_cs_sq_ib_tail(v_context, context->ib_cpu + len);
+ ip = AMDGPU_HW_IP_VCN_ENC;
+ } else
+ ip = AMDGPU_HW_IP_VCN_DEC;
+
+ r = submit(device_handle, context, len, ip);
+
+ igt_assert_eq(r, 0);
+
+ free_resource(&msg_buf);
+}
+
+static void
+amdgpu_cs_vcn_dec_decode(amdgpu_device_handle device_handle, struct mmd_context *context,
+ struct vcn_context *v_context)
+{
+ const unsigned int dpb_size = 15923584, dt_size = 737280;
+ uint64_t msg_addr, fb_addr, bs_addr, dpb_addr, ctx_addr, dt_addr, it_addr, sum;
+ struct amdgpu_mmd_bo dec_buf;
+ int size, len, i, r;
+ unsigned int ip;
+ uint8_t *dec;
+
+ size = 4*1024; /* msg */
+ size += 4*1024; /* fb */
+ size += 4096; /*it_scaling_table*/
+ size += ALIGN(sizeof(uvd_bitstream), 4*1024);
+ size += ALIGN(dpb_size, 4*1024);
+ size += ALIGN(dt_size, 4*1024);
+
+ context->num_resources = 0;
+ alloc_resource(device_handle, &dec_buf, size, AMDGPU_GEM_DOMAIN_GTT);
+ context->resources[context->num_resources++] = dec_buf.handle;
+ context->resources[context->num_resources++] = context->ib_handle;
+
+ r = amdgpu_bo_cpu_map(dec_buf.handle, (void **)&dec_buf.ptr);
+ dec = dec_buf.ptr;
+
+ igt_assert_eq(r, 0);
+ memset(dec_buf.ptr, 0, size);
+ memcpy(dec_buf.ptr, vcn_dec_decode_msg, sizeof(vcn_dec_decode_msg));
+ memcpy(dec_buf.ptr + sizeof(vcn_dec_decode_msg),
+ avc_decode_msg, sizeof(avc_decode_msg));
+
+ dec += 4*1024;
+ memcpy(dec, feedback_msg, sizeof(feedback_msg));
+ dec += 4*1024;
+ memcpy(dec, uvd_it_scaling_table, sizeof(uvd_it_scaling_table));
+
+ dec += 4*1024;
+ memcpy(dec, uvd_bitstream, sizeof(uvd_bitstream));
+
+ dec += ALIGN(sizeof(uvd_bitstream), 4*1024);
+
+ dec += ALIGN(dpb_size, 4*1024);
+
+ msg_addr = dec_buf.addr;
+ fb_addr = msg_addr + 4*1024;
+ it_addr = fb_addr + 4*1024;
+ bs_addr = it_addr + 4*1024;
+ dpb_addr = ALIGN(bs_addr + sizeof(uvd_bitstream), 4*1024);
+ ctx_addr = ALIGN(dpb_addr + 0x006B9400, 4*1024);
+ dt_addr = ALIGN(dpb_addr + dpb_size, 4*1024);
+
+ len = 0;
+ vcn_dec_cmd(context, v_context, msg_addr, 0x0, &len);
+ vcn_dec_cmd(context, v_context, dpb_addr, 0x1, &len);
+ vcn_dec_cmd(context, v_context, dt_addr, 0x2, &len);
+ vcn_dec_cmd(context, v_context, fb_addr, 0x3, &len);
+ vcn_dec_cmd(context, v_context, bs_addr, 0x100, &len);
+ vcn_dec_cmd(context, v_context, it_addr, 0x204, &len);
+ vcn_dec_cmd(context, v_context, ctx_addr, 0x206, &len);
+
+ if (context->vcn_dec_sw_ring == false) {
+ context->ib_cpu[len++] = reg[context->vcn_reg_index].cntl;
+ context->ib_cpu[len++] = 0x1;
+ for (; len % 16; ) {
+ context->ib_cpu[len++] = reg[context->vcn_reg_index].nop;
+ context->ib_cpu[len++] = 0;
+ }
+ }
+
+ if (context->vcn_unified_ring) {
+ amdgpu_cs_sq_ib_tail(v_context, context->ib_cpu + len);
+ ip = AMDGPU_HW_IP_VCN_ENC;
+ } else
+ ip = AMDGPU_HW_IP_VCN_DEC;
+
+ r = submit(device_handle, context, len, ip);
+ igt_assert_eq(r, 0);
+
+ for (i = 0, sum = 0; i < dt_size; ++i)
+ sum += dec[i];
+
+ igt_assert_eq(sum, SUM_DECODE);
+
+ free_resource(&dec_buf);
+}
+
+static void
+amdgpu_cs_vcn_dec_destroy(amdgpu_device_handle device_handle,
+ struct mmd_context *context, struct vcn_context *v_context)
+{
+ struct amdgpu_mmd_bo msg_buf;
+ unsigned int ip;
+ int len, r;
+
+ context->num_resources = 0;
+ alloc_resource(device_handle, &msg_buf, 1024, AMDGPU_GEM_DOMAIN_GTT);
+ context->resources[context->num_resources++] = msg_buf.handle;
+ context->resources[context->num_resources++] = context->ib_handle;
+
+ r = amdgpu_bo_cpu_map(msg_buf.handle, (void **)&msg_buf.ptr);
+ igt_assert_eq(r, 0);
+
+ memset(msg_buf.ptr, 0, 1024);
+ memcpy(msg_buf.ptr, vcn_dec_destroy_msg, sizeof(vcn_dec_destroy_msg));
+
+ len = 0;
+ if (context->vcn_dec_sw_ring == true)
+ vcn_dec_cmd(context, v_context, msg_buf.addr, 0, &len);
+ else {
+ context->ib_cpu[len++] = reg[context->vcn_reg_index].data0;
+ context->ib_cpu[len++] = msg_buf.addr;
+ context->ib_cpu[len++] = reg[context->vcn_reg_index].data1;
+ context->ib_cpu[len++] = msg_buf.addr >> 32;
+ context->ib_cpu[len++] = reg[context->vcn_reg_index].cmd;
+ context->ib_cpu[len++] = 0;
+ for (; len % 16; ) {
+ context->ib_cpu[len++] = reg[context->vcn_reg_index].nop;
+ context->ib_cpu[len++] = 0;
+ }
+ }
+
+ if (context->vcn_unified_ring) {
+ amdgpu_cs_sq_ib_tail(v_context, context->ib_cpu + len);
+ ip = AMDGPU_HW_IP_VCN_ENC;
+ } else
+ ip = AMDGPU_HW_IP_VCN_DEC;
+
+ r = submit(device_handle, context, len, ip);
+ igt_assert_eq(r, 0);
+
+ free_resource(&msg_buf);
+}
+
+static void
+amdgpu_cs_vcn_enc_create(amdgpu_device_handle device_handle,
+ struct mmd_context *context, struct vcn_context *v_context)
+{
+ int len, r;
+ uint32_t *p_task_size = NULL;
+ uint32_t task_offset = 0, st_offset;
+ uint32_t *st_size = NULL;
+ unsigned int width = 160, height = 128, buf_size;
+ uint32_t fw_maj = 1, fw_min = 9;
+
+ if (context->vcn_ip_version_major == 2) {
+ fw_maj = 1;
+ fw_min = 1;
+ } else if (context->vcn_ip_version_major == 3) {
+ fw_maj = 1;
+ fw_min = 0;
+ }
+
+ v_context->gWidth = width;
+ v_context->gHeight = height;
+ buf_size = ALIGN(width, 256) * ALIGN(height, 32) * 3 / 2;
+ v_context->enc_task_id = 1;
+
+ context->num_resources = 0;
+ alloc_resource(device_handle, &v_context->enc_buf, 128 * 1024, AMDGPU_GEM_DOMAIN_GTT);
+ alloc_resource(device_handle, &v_context->cpb_buf, buf_size * 2, AMDGPU_GEM_DOMAIN_GTT);
+ context->resources[context->num_resources++] = v_context->enc_buf.handle;
+ context->resources[context->num_resources++] = v_context->cpb_buf.handle;
+ context->resources[context->num_resources++] = context->ib_handle;
+
+ r = amdgpu_bo_cpu_map(v_context->enc_buf.handle, (void **)&v_context->enc_buf.ptr);
+ memset(v_context->enc_buf.ptr, 0, 128 * 1024);
+ r = amdgpu_bo_cpu_unmap(v_context->enc_buf.handle);
+
+ r = amdgpu_bo_cpu_map(v_context->cpb_buf.handle, (void **)&v_context->enc_buf.ptr);
+ memset(v_context->enc_buf.ptr, 0, buf_size * 2);
+ r = amdgpu_bo_cpu_unmap(v_context->cpb_buf.handle);
+
+ len = 0;
+
+ if (context->vcn_unified_ring)
+ amdgpu_cs_sq_head(v_context, context->ib_cpu, &len, true);
+
+ /* session info */
+ st_offset = len;
+ st_size = &context->ib_cpu[len++]; /* size */
+ context->ib_cpu[len++] = 0x00000001; /* RENCODE_IB_PARAM_SESSION_INFO */
+ context->ib_cpu[len++] = ((fw_maj << 16) | (fw_min << 0));
+ context->ib_cpu[len++] = v_context->enc_buf.addr >> 32;
+ context->ib_cpu[len++] = v_context->enc_buf.addr;
+ context->ib_cpu[len++] = 1; /* RENCODE_ENGINE_TYPE_ENCODE; */
+ *st_size = (len - st_offset) * 4;
+
+ /* task info */
+ task_offset = len;
+ st_offset = len;
+ st_size = &context->ib_cpu[len++]; /* size */
+ context->ib_cpu[len++] = 0x00000002; /* RENCODE_IB_PARAM_TASK_INFO */
+ p_task_size = &context->ib_cpu[len++];
+ context->ib_cpu[len++] = v_context->enc_task_id++; /* task_id */
+ context->ib_cpu[len++] = 0; /* feedback */
+ *st_size = (len - st_offset) * 4;
+
+ /* op init */
+ st_offset = len;
+ st_size = &context->ib_cpu[len++]; /* size */
+ context->ib_cpu[len++] = 0x01000001; /* RENCODE_IB_OP_INITIALIZE */
+ *st_size = (len - st_offset) * 4;
+
+ /* session_init */
+ st_offset = len;
+ st_size = &context->ib_cpu[len++]; /* size */
+ context->ib_cpu[len++] = 0x00000003; /* RENCODE_IB_PARAM_SESSION_INIT */
+ context->ib_cpu[len++] = 1; /* RENCODE_ENCODE_STANDARD_H264 */
+ context->ib_cpu[len++] = width;
+ context->ib_cpu[len++] = height;
+ context->ib_cpu[len++] = 0;
+ context->ib_cpu[len++] = 0;
+ context->ib_cpu[len++] = 0; /* pre encode mode */
+ context->ib_cpu[len++] = 0; /* chroma enabled : false */
+ *st_size = (len - st_offset) * 4;
+
+ /* slice control */
+ st_offset = len;
+ st_size = &context->ib_cpu[len++]; /* size */
+ context->ib_cpu[len++] = 0x00200001; /* RENCODE_H264_IB_PARAM_SLICE_CONTROL */
+ context->ib_cpu[len++] = 0; /* RENCODE_H264_SLICE_CONTROL_MODE_FIXED_MBS */
+ context->ib_cpu[len++] = ALIGN(width, 16) / 16 * ALIGN(height, 16) / 16;
+ *st_size = (len - st_offset) * 4;
+
+ /* enc spec misc */
+ st_offset = len;
+ st_size = &context->ib_cpu[len++]; /* size */
+ context->ib_cpu[len++] = 0x00200002; /* RENCODE_H264_IB_PARAM_SPEC_MISC */
+ context->ib_cpu[len++] = 0; /* constrained intra pred flag */
+ context->ib_cpu[len++] = 0; /* cabac enable */
+ context->ib_cpu[len++] = 0; /* cabac init idc */
+ context->ib_cpu[len++] = 1; /* half pel enabled */
+ context->ib_cpu[len++] = 1; /* quarter pel enabled */
+ context->ib_cpu[len++] = 100; /* BASELINE profile */
+ context->ib_cpu[len++] = 11; /* level */
+ if (context->vcn_ip_version_major == 3) {
+ context->ib_cpu[len++] = 0; /* b_picture_enabled */
+ context->ib_cpu[len++] = 0; /* weighted_bipred_idc */
+ }
+ *st_size = (len - st_offset) * 4;
+
+ /* deblocking filter */
+ st_offset = len;
+ st_size = &context->ib_cpu[len++]; /* size */
+ context->ib_cpu[len++] = 0x00200004; /* RENCODE_H264_IB_PARAM_DEBLOCKING_FILTER */
+ context->ib_cpu[len++] = 0; /* disable deblocking filter idc */
+ context->ib_cpu[len++] = 0; /* alpha c0 offset */
+ context->ib_cpu[len++] = 0; /* tc offset */
+ context->ib_cpu[len++] = 0; /* cb offset */
+ context->ib_cpu[len++] = 0; /* cr offset */
+ *st_size = (len - st_offset) * 4;
+
+ /* layer control */
+ st_offset = len;
+ st_size = &context->ib_cpu[len++]; /* size */
+ context->ib_cpu[len++] = 0x00000004; /* RENCODE_IB_PARAM_LAYER_CONTROL */
+ context->ib_cpu[len++] = 1; /* max temporal layer */
+ context->ib_cpu[len++] = 1; /* no of temporal layer */
+ *st_size = (len - st_offset) * 4;
+
+ /* rc_session init */
+ st_offset = len;
+ st_size = &context->ib_cpu[len++]; /* size */
+ context->ib_cpu[len++] = 0x00000006; /* RENCODE_IB_PARAM_RATE_CONTROL_SESSION_INIT */
+ context->ib_cpu[len++] = 0; /* rate control */
+ context->ib_cpu[len++] = 48; /* vbv buffer level */
+ *st_size = (len - st_offset) * 4;
+
+ /* quality params */
+ st_offset = len;
+ st_size = &context->ib_cpu[len++]; /* size */
+ context->ib_cpu[len++] = 0x00000009; /* RENCODE_IB_PARAM_QUALITY_PARAMS */
+ context->ib_cpu[len++] = 0; /* vbaq mode */
+ context->ib_cpu[len++] = 0; /* scene change sensitivity */
+ context->ib_cpu[len++] = 0; /* scene change min idr interval */
+ context->ib_cpu[len++] = 0;
+ if (context->vcn_ip_version_major == 3)
+ context->ib_cpu[len++] = 0;
+ *st_size = (len - st_offset) * 4;
+
+ /* layer select */
+ st_offset = len;
+ st_size = &context->ib_cpu[len++]; /* size */
+ context->ib_cpu[len++] = 0x00000005; /* RENCODE_IB_PARAM_LAYER_SELECT */
+ context->ib_cpu[len++] = 0; /* temporal layer */
+ *st_size = (len - st_offset) * 4;
+
+ /* rc layer init */
+ st_offset = len;
+ st_size = &context->ib_cpu[len++]; /* size */
+ context->ib_cpu[len++] = 0x00000007; /* RENCODE_IB_PARAM_RATE_CONTROL_LAYER_INIT */
+ context->ib_cpu[len++] = 0;
+ context->ib_cpu[len++] = 0;
+ context->ib_cpu[len++] = 25;
+ context->ib_cpu[len++] = 1;
+ context->ib_cpu[len++] = 0x01312d00;
+ context->ib_cpu[len++] = 0;
+ context->ib_cpu[len++] = 0;
+ context->ib_cpu[len++] = 0;
+ *st_size = (len - st_offset) * 4;
+
+ /* layer select */
+ st_offset = len;
+ st_size = &context->ib_cpu[len++]; /* size */
+ context->ib_cpu[len++] = 0x00000005; /* RENCODE_IB_PARAM_LAYER_SELECT */
+ context->ib_cpu[len++] = 0; /* temporal layer */
+ *st_size = (len - st_offset) * 4;
+
+ /* rc per pic */
+ st_offset = len;
+ st_size = &context->ib_cpu[len++]; /* size */
+ context->ib_cpu[len++] = 0x00000008; /* RENCODE_IB_PARAM_RATE_CONTROL_PER_PICTURE */
+ context->ib_cpu[len++] = 20;
+ context->ib_cpu[len++] = 0;
+ context->ib_cpu[len++] = 51;
+ context->ib_cpu[len++] = 0;
+ context->ib_cpu[len++] = 1;
+ context->ib_cpu[len++] = 0;
+ context->ib_cpu[len++] = 1;
+ *st_size = (len - st_offset) * 4;
+
+ /* op init rc */
+ st_offset = len;
+ st_size = &context->ib_cpu[len++]; /* size */
+ context->ib_cpu[len++] = 0x01000004; /* RENCODE_IB_OP_INIT_RC */
+ *st_size = (len - st_offset) * 4;
+
+ /* op init rc vbv */
+ st_offset = len;
+ st_size = &context->ib_cpu[len++]; /* size */
+ context->ib_cpu[len++] = 0x01000005; /* RENCODE_IB_OP_INIT_RC_VBV_BUFFER_LEVEL */
+ *st_size = (len - st_offset) * 4;
+
+ *p_task_size = (len - task_offset) * 4;
+
+ if (context->vcn_unified_ring)
+ amdgpu_cs_sq_ib_tail(v_context, context->ib_cpu + len);
+
+ r = submit(device_handle, context, len, AMDGPU_HW_IP_VCN_ENC);
+ igt_assert_eq(r, 0);
+}
+
+static int32_t
+h264_se(struct bufferInfo *bufInfo)
+{
+ uint32_t ret, temp;
+
+ ret = bs_read_ue(bufInfo);
+ if ((ret & 0x1) == 0) {
+ ret >>= 1;
+ temp = 0 - ret;
+ return temp;
+ }
+
+ return (ret + 1) >> 1;
+}
+
+static void
+h264_check_0s(struct bufferInfo *bufInfo, int count)
+{
+ uint32_t val;
+
+ val = bs_read_u(bufInfo, count);
+ if (val != 0)
+ igt_info("field error - %d bits should be 0 is %x\n", count, val);
+}
+
+static int
+bs_eof(struct bufferInfo *bufinfo)
+{
+ if (bufinfo->decBuffer >= bufinfo->end)
+ return 1;
+ else
+ return 0;
+}
+
+static uint32_t
+bs_read_u1(struct bufferInfo *bufinfo)
+{
+ uint32_t r = 0;
+ uint32_t temp = 0;
+
+ bufinfo->numOfBitsInBuffer--;
+ if (!bs_eof(bufinfo)) {
+ temp = (((bufinfo->decData)) >> bufinfo->numOfBitsInBuffer);
+ r = temp & 0x01;
+ }
+
+ if (bufinfo->numOfBitsInBuffer == 0) {
+ bufinfo->decBuffer++;
+ bufinfo->decData = *bufinfo->decBuffer;
+ bufinfo->numOfBitsInBuffer = 8;
+ }
+
+ return r;
+}
+
+static uint32_t
+bs_read_u(struct bufferInfo *bufinfo, int n)
+{
+ uint32_t r = 0;
+ int i;
+
+ for (i = 0; i < n; i++)
+ r |= (bs_read_u1(bufinfo) << (n - i - 1));
+
+ return r;
+}
+
+static uint32_t
+bs_read_ue(struct bufferInfo *bufinfo)
+{
+ int32_t r = 0;
+ int i = 0;
+
+ while ((bs_read_u1(bufinfo) == 0) && (i < 32) && (!bs_eof(bufinfo)))
+ i++;
+ r = bs_read_u(bufinfo, i);
+ r += (1 << i) - 1;
+ return r;
+}
+
+static uint32_t
+remove_03(uint8_t *bptr, uint32_t len)
+{
+ uint32_t nal_len = 0;
+
+ while (nal_len + 2 < len) {
+ if (bptr[0] == 0 && bptr[1] == 0 && bptr[2] == 3) {
+ bptr += 2;
+ nal_len += 2;
+ len--;
+ memmove(bptr, bptr + 1, len - nal_len);
+ } else {
+ bptr++;
+ nal_len++;
+ }
+ }
+ return len;
+}
+
+static void
+scaling_list(uint32_t ix, uint32_t sizeOfScalingList, struct bufferInfo *bufInfo)
+{
+ uint32_t lastScale = 8, nextScale = 8;
+ uint32_t jx;
+ int deltaScale;
+
+ for (jx = 0; jx < sizeOfScalingList; jx++) {
+ if (nextScale != 0) {
+ deltaScale = h264_se(bufInfo);
+ nextScale = (lastScale + deltaScale + 256) % 256;
+ }
+ if (nextScale == 0)
+ ;
+ else
+ lastScale = nextScale;
+ }
+}
+
+static void
+h264_parse_sequence_parameter_set(struct h264_decode *dec, struct bufferInfo *bufInfo)
+{
+ uint32_t temp, seq_scaling_matrix_present_flag, pic_order_cnt_type,
+ PicWidthInMbs, PicHeightInMapUnits, frame_mbs_only_flag;
+
+ dec->profile = bs_read_u(bufInfo, 8);
+ bs_read_u(bufInfo, 1); /* constaint_set0_flag */
+ bs_read_u(bufInfo, 1); /* constaint_set1_flag */
+ bs_read_u(bufInfo, 1); /* constaint_set2_flag */
+ bs_read_u(bufInfo, 1); /* constaint_set3_flag */
+ bs_read_u(bufInfo, 1); /* constaint_set4_flag */
+ bs_read_u(bufInfo, 1); /* constaint_set5_flag */
+
+
+ h264_check_0s(bufInfo, 2);
+ dec->level_idc = bs_read_u(bufInfo, 8);
+ bs_read_ue(bufInfo); /* SPS id*/
+ if (dec->profile == 100 || dec->profile == 110 ||
+ dec->profile == 122 || dec->profile == 144) {
+ uint32_t chroma_format_idc = bs_read_ue(bufInfo);
+
+ if (chroma_format_idc == 3)
+ bs_read_u(bufInfo, 1); /* residual_colour_transform_flag */
+
+ bs_read_ue(bufInfo); /* bit_depth_luma_minus8 */
+ bs_read_ue(bufInfo); /* bit_depth_chroma_minus8 */
+ bs_read_u(bufInfo, 1); /* qpprime_y_zero_transform_bypass_flag */
+ seq_scaling_matrix_present_flag = bs_read_u(bufInfo, 1);
+
+ if (seq_scaling_matrix_present_flag) {
+ for (uint32_t ix = 0; ix < 8; ix++) {
+ temp = bs_read_u(bufInfo, 1);
+ if (temp)
+ scaling_list(ix, ix < 6 ? 16 : 64, bufInfo);
+ }
+ }
+ }
+
+ bs_read_ue(bufInfo); /* log2_max_frame_num_minus4 */
+ pic_order_cnt_type = bs_read_ue(bufInfo);
+
+ if (pic_order_cnt_type == 0) {
+ bs_read_ue(bufInfo); /* log2_max_pic_order_cnt_lsb_minus4 */
+ } else if (pic_order_cnt_type == 1) {
+ bs_read_u(bufInfo, 1); /* delta_pic_order_always_zero_flag */
+ h264_se(bufInfo); /* offset_for_non_ref_pic */
+ h264_se(bufInfo); /* offset_for_top_to_bottom_field */
+ temp = bs_read_ue(bufInfo);
+ for (uint32_t ix = 0; ix < temp; ix++)
+ h264_se(bufInfo); /* offset_for_ref_frame[index] */
+ }
+ bs_read_ue(bufInfo); /* num_ref_frames */
+ bs_read_u(bufInfo, 1); /* gaps_in_frame_num_flag */
+ PicWidthInMbs = bs_read_ue(bufInfo) + 1;
+
+ dec->pic_width = PicWidthInMbs * 16;
+ PicHeightInMapUnits = bs_read_ue(bufInfo) + 1;
+
+ dec->pic_height = PicHeightInMapUnits * 16;
+ frame_mbs_only_flag = bs_read_u(bufInfo, 1);
+ if (!frame_mbs_only_flag)
+ bs_read_u(bufInfo, 1); /* mb_adaptive_frame_field_flag */
+ bs_read_u(bufInfo, 1); /* direct_8x8_inference_flag */
+ temp = bs_read_u(bufInfo, 1);
+ if (temp) {
+ bs_read_ue(bufInfo); /* frame_crop_left_offset */
+ bs_read_ue(bufInfo); /* frame_crop_right_offset */
+ bs_read_ue(bufInfo); /* frame_crop_top_offset */
+ bs_read_ue(bufInfo); /* frame_crop_bottom_offset */
+ }
+ temp = bs_read_u(bufInfo, 1); /* VUI Parameters */
+}
+
+static void
+h264_slice_header(struct h264_decode *dec, struct bufferInfo *bufInfo)
+{
+ uint32_t temp;
+
+ bs_read_ue(bufInfo); /* first_mb_in_slice */
+ temp = bs_read_ue(bufInfo);
+ dec->slice_type = ((temp > 5) ? (temp - 5) : temp);
+}
+
+static uint8_t
+h264_parse_nal(struct h264_decode *dec, struct bufferInfo *bufInfo)
+{
+ uint8_t type = 0;
+
+ h264_check_0s(bufInfo, 1);
+ dec->nal_ref_idc = bs_read_u(bufInfo, 2);
+ dec->nal_unit_type = type = bs_read_u(bufInfo, 5);
+ switch (type) {
+ case H264_NAL_TYPE_NON_IDR_SLICE:
+ case H264_NAL_TYPE_IDR_SLICE:
+ h264_slice_header(dec, bufInfo);
+ break;
+ case H264_NAL_TYPE_SEQ_PARAM:
+ h264_parse_sequence_parameter_set(dec, bufInfo);
+ break;
+ case H264_NAL_TYPE_PIC_PARAM:
+ case H264_NAL_TYPE_SEI:
+ case H264_NAL_TYPE_ACCESS_UNIT:
+ case H264_NAL_TYPE_SEQ_EXTENSION:
+ /* NOP */
+ break;
+ default:
+ igt_info("Nal type unknown %d\n", type);
+ break;
+ }
+ return type;
+}
+
+static uint32_t
+h264_find_next_start_code(uint8_t *pBuf, uint32_t bufLen)
+{
+ uint32_t val;
+ uint32_t offset, startBytes;
+
+ offset = startBytes = 0;
+ if (pBuf[0] == 0 && pBuf[1] == 0 && pBuf[2] == 0 && pBuf[3] == 1) {
+ pBuf += 4;
+ offset = 4;
+ startBytes = 1;
+ } else if (pBuf[0] == 0 && pBuf[1] == 0 && pBuf[2] == 1) {
+ pBuf += 3;
+ offset = 3;
+ startBytes = 1;
+ }
+ val = 0xffffffff;
+ while (offset < bufLen - 3) {
+ val <<= 8;
+ val |= *pBuf++;
+ offset++;
+ if (val == H264_START_CODE)
+ return offset - 4;
+
+ if ((val & 0x00ffffff) == H264_START_CODE)
+ return offset - 3;
+ }
+ if (bufLen - offset <= 3 && startBytes == 0) {
+ startBytes = 0;
+ return 0;
+ }
+
+ return offset;
+}
+
+static int
+verify_checksum(struct vcn_context *v_context, uint8_t *buffer, uint32_t buffer_size)
+{
+ uint32_t buffer_pos = 0;
+ struct h264_decode dec;
+ int done = 0;
+
+ memset(&dec, 0, sizeof(struct h264_decode));
+ do {
+ uint32_t ret;
+
+ ret = h264_find_next_start_code(buffer + buffer_pos,
+ buffer_size - buffer_pos);
+ if (ret == 0) {
+ done = 1;
+ if (buffer_pos == 0)
+ igt_info("couldn't find start code in buffer from 0\n");
+ } else {
+ /* have a complete NAL from buffer_pos to end */
+ if (ret > 3) {
+ uint32_t nal_len;
+ struct bufferInfo bufinfo;
+
+ nal_len = remove_03(buffer + buffer_pos, ret);
+ bufinfo.decBuffer = buffer + buffer_pos + (buffer[buffer_pos + 2] == 1 ? 3 : 4);
+ bufinfo.decBufferSize = (nal_len - (buffer[buffer_pos + 2] == 1 ? 3 : 4)) * 8;
+ bufinfo.end = buffer + buffer_pos + nal_len;
+ bufinfo.numOfBitsInBuffer = 8;
+ bufinfo.decData = *bufinfo.decBuffer;
+ h264_parse_nal(&dec, &bufinfo);
+ }
+ buffer_pos += ret; /* buffer_pos points to next code */
+ }
+ } while (done == 0);
+
+ if ((dec.pic_width == v_context->gWidth) &&
+ (dec.pic_height == v_context->gHeight) &&
+ (dec.slice_type == v_context->gSliceType))
+ return 0;
+ else
+ return -1;
+}
+
+static void
+check_result(struct vcn_context *v_context, struct amdgpu_mmd_bo fb_buf,
+ struct amdgpu_mmd_bo bs_buf, int frame_type)
+{
+ uint32_t *fb_ptr;
+ uint8_t *bs_ptr;
+ uint32_t size;
+ int r;
+/* uint64_t s[3] = {0, 1121279001727, 1059312481445}; */
+
+ r = amdgpu_bo_cpu_map(fb_buf.handle, (void **)&fb_buf.ptr);
+ igt_assert_eq(r, 0);
+ fb_ptr = (uint32_t *)fb_buf.ptr;
+ size = fb_ptr[6];
+ r = amdgpu_bo_cpu_unmap(fb_buf.handle);
+ igt_assert_eq(r, 0);
+ r = amdgpu_bo_cpu_map(bs_buf.handle, (void **)&bs_buf.ptr);
+ igt_assert_eq(r, 0);
+
+ bs_ptr = (uint8_t *)bs_buf.ptr;
+ r = verify_checksum(v_context, bs_ptr, size);
+ igt_assert_eq(r, 0);
+ r = amdgpu_bo_cpu_unmap(bs_buf.handle);
+
+ igt_assert_eq(r, 0);
+}
+
+static void
+amdgpu_cs_vcn_enc_encode_frame(amdgpu_device_handle device_handle,
+ struct mmd_context *context, struct vcn_context *v_context, int frame_type)
+{
+ struct amdgpu_mmd_bo bs_buf, fb_buf, vbv_buf;
+ int len, r, i;
+ unsigned int width = 160, height = 128, buf_size;
+ uint32_t *p_task_size = NULL;
+ uint32_t task_offset = 0, st_offset;
+ uint32_t *st_size = NULL;
+ uint32_t fw_maj = 1, fw_min = 9;
+
+ if (context->vcn_ip_version_major == 2) {
+ fw_maj = 1;
+ fw_min = 1;
+ } else if (context->vcn_ip_version_major == 3) {
+ fw_maj = 1;
+ fw_min = 0;
+ }
+ v_context->gSliceType = frame_type;
+ buf_size = ALIGN(width, 256) * ALIGN(height, 32) * 3 / 2;
+
+ context->num_resources = 0;
+ alloc_resource(device_handle, &bs_buf, 4096, AMDGPU_GEM_DOMAIN_GTT);
+ alloc_resource(device_handle, &fb_buf, 4096, AMDGPU_GEM_DOMAIN_GTT);
+ alloc_resource(device_handle, &vbv_buf, buf_size, AMDGPU_GEM_DOMAIN_GTT);
+ context->resources[context->num_resources++] = v_context->enc_buf.handle;
+ context->resources[context->num_resources++] = v_context->cpb_buf.handle;
+ context->resources[context->num_resources++] = bs_buf.handle;
+ context->resources[context->num_resources++] = fb_buf.handle;
+ context->resources[context->num_resources++] = vbv_buf.handle;
+ context->resources[context->num_resources++] = context->ib_handle;
+
+
+ r = amdgpu_bo_cpu_map(bs_buf.handle, (void **)&bs_buf.ptr);
+ memset(bs_buf.ptr, 0, 4096);
+ r = amdgpu_bo_cpu_unmap(bs_buf.handle);
+
+ r = amdgpu_bo_cpu_map(fb_buf.handle, (void **)&fb_buf.ptr);
+ memset(fb_buf.ptr, 0, 4096);
+ r = amdgpu_bo_cpu_unmap(fb_buf.handle);
+
+ r = amdgpu_bo_cpu_map(vbv_buf.handle, (void **)&vbv_buf.ptr);
+ igt_assert_eq(r, 0);
+
+ for (i = 0; i < ALIGN(height, 32) * 3 / 2; i++)
+ memcpy(vbv_buf.ptr + i * ALIGN(width, 256), frame + i * width, width);
+
+ r = amdgpu_bo_cpu_unmap(vbv_buf.handle);
+ igt_assert_eq(r, 0);
+
+ len = 0;
+
+ if (context->vcn_unified_ring)
+ amdgpu_cs_sq_head(v_context, context->ib_cpu, &len, true);
+
+ /* session info */
+ st_offset = len;
+ st_size = &context->ib_cpu[len++]; /* size */
+ context->ib_cpu[len++] = 0x00000001; /* RENCODE_IB_PARAM_SESSION_INFO */
+ context->ib_cpu[len++] = ((fw_maj << 16) | (fw_min << 0));
+ context->ib_cpu[len++] = v_context->enc_buf.addr >> 32;
+ context->ib_cpu[len++] = v_context->enc_buf.addr;
+ context->ib_cpu[len++] = 1; /* RENCODE_ENGINE_TYPE_ENCODE */;
+ *st_size = (len - st_offset) * 4;
+
+ /* task info */
+ task_offset = len;
+ st_offset = len;
+ st_size = &context->ib_cpu[len++]; /* size */
+ context->ib_cpu[len++] = 0x00000002; /* RENCODE_IB_PARAM_TASK_INFO */
+ p_task_size = &context->ib_cpu[len++];
+ context->ib_cpu[len++] = v_context->enc_task_id++; /* task_id */
+ context->ib_cpu[len++] = 1; /* feedback */
+ *st_size = (len - st_offset) * 4;
+
+ if (frame_type == 2) {
+ /* sps */
+ st_offset = len;
+ st_size = &context->ib_cpu[len++]; /* size */
+ if (context->vcn_ip_version_major == 1)
+ context->ib_cpu[len++] = 0x00000020; /* RENCODE_IB_PARAM_DIRECT_OUTPUT_NALU vcn 1 */
+ else
+ context->ib_cpu[len++] = 0x0000000a; /* RENCODE_IB_PARAM_DIRECT_OUTPUT_NALU vcn 2,3 */
+ context->ib_cpu[len++] = 0x00000002; /* RENCODE_DIRECT_OUTPUT_NALU_TYPE_SPS */
+ context->ib_cpu[len++] = 0x00000011; /* sps len */
+ context->ib_cpu[len++] = 0x00000001; /* start code */
+ context->ib_cpu[len++] = 0x6764440b;
+ context->ib_cpu[len++] = 0xac54c284;
+ context->ib_cpu[len++] = 0x68078442;
+ context->ib_cpu[len++] = 0x37000000;
+ *st_size = (len - st_offset) * 4;
+
+ /* pps */
+ st_offset = len;
+ st_size = &context->ib_cpu[len++]; /* size */
+ if (context->vcn_ip_version_major == 1)
+ context->ib_cpu[len++] = 0x00000020; /* RENCODE_IB_PARAM_DIRECT_OUTPUT_NALU vcn 1*/
+ else
+ context->ib_cpu[len++] = 0x0000000a; /* RENCODE_IB_PARAM_DIRECT_OUTPUT_NALU vcn 2,3*/
+ context->ib_cpu[len++] = 0x00000003; /* RENCODE_DIRECT_OUTPUT_NALU_TYPE_PPS */
+ context->ib_cpu[len++] = 0x00000008; /* pps len */
+ context->ib_cpu[len++] = 0x00000001; /* start code */
+ context->ib_cpu[len++] = 0x68ce3c80;
+ *st_size = (len - st_offset) * 4;
+ }
+
+ /* slice header */
+ st_offset = len;
+ st_size = &context->ib_cpu[len++]; /* size */
+ if (context->vcn_ip_version_major == 1)
+ context->ib_cpu[len++] = 0x0000000a; /* RENCODE_IB_PARAM_SLICE_HEADER vcn 1 */
+ else
+ context->ib_cpu[len++] = 0x0000000b; /* RENCODE_IB_PARAM_SLICE_HEADER vcn 2,3 */
+ if (frame_type == 2) {
+ context->ib_cpu[len++] = 0x65000000;
+ context->ib_cpu[len++] = 0x11040000;
+ } else {
+ context->ib_cpu[len++] = 0x41000000;
+ context->ib_cpu[len++] = 0x34210000;
+ }
+ context->ib_cpu[len++] = 0xe0000000;
+ for (i = 0; i < 13; i++)
+ context->ib_cpu[len++] = 0x00000000;
+
+ context->ib_cpu[len++] = 0x00000001;
+ context->ib_cpu[len++] = 0x00000008;
+ context->ib_cpu[len++] = 0x00020000;
+ context->ib_cpu[len++] = 0x00000000;
+ context->ib_cpu[len++] = 0x00000001;
+ context->ib_cpu[len++] = 0x00000015;
+ context->ib_cpu[len++] = 0x00020001;
+ context->ib_cpu[len++] = 0x00000000;
+ context->ib_cpu[len++] = 0x00000001;
+ context->ib_cpu[len++] = 0x00000003;
+ for (i = 0; i < 22; i++)
+ context->ib_cpu[len++] = 0x00000000;
+
+ *st_size = (len - st_offset) * 4;
+
+ /* encode params */
+ st_offset = len;
+ st_size = &context->ib_cpu[len++]; /* size */
+ if (context->vcn_ip_version_major == 1)
+ context->ib_cpu[len++] = 0x0000000b; /* RENCODE_IB_PARAM_ENCODE_PARAMS vcn 1*/
+ else
+ context->ib_cpu[len++] = 0x0000000f; /* RENCODE_IB_PARAM_ENCODE_PARAMS vcn 2,3*/
+ context->ib_cpu[len++] = frame_type;
+ context->ib_cpu[len++] = 0x0001f000;
+ context->ib_cpu[len++] = vbv_buf.addr >> 32;
+ context->ib_cpu[len++] = vbv_buf.addr;
+ context->ib_cpu[len++] = (vbv_buf.addr + ALIGN(width, 256) * ALIGN(height, 32)) >> 32;
+ context->ib_cpu[len++] = vbv_buf.addr + ALIGN(width, 256) * ALIGN(height, 32);
+ context->ib_cpu[len++] = 0x00000100;
+ context->ib_cpu[len++] = 0x00000080;
+ context->ib_cpu[len++] = 0x00000000;
+ context->ib_cpu[len++] = 0xffffffff;
+ context->ib_cpu[len++] = 0x00000000;
+ *st_size = (len - st_offset) * 4;
+
+ /* encode params h264 */
+ st_offset = len;
+ st_size = &context->ib_cpu[len++]; /* size */
+ context->ib_cpu[len++] = 0x00200003; /* RENCODE_H264_IB_PARAM_ENCODE_PARAMS */
+ if (context->vcn_ip_version_major != 3) {
+ context->ib_cpu[len++] = 0x00000000;
+ context->ib_cpu[len++] = 0x00000000;
+ context->ib_cpu[len++] = 0x00000000;
+ context->ib_cpu[len++] = 0xffffffff;
+ } else {
+ context->ib_cpu[len++] = 0x00000000;
+ context->ib_cpu[len++] = 0x00000000;
+ context->ib_cpu[len++] = 0x00000000;
+ context->ib_cpu[len++] = 0x00000000;
+ context->ib_cpu[len++] = 0x00000000;
+ context->ib_cpu[len++] = 0x00000000;
+ context->ib_cpu[len++] = 0x00000000;
+ context->ib_cpu[len++] = 0xffffffff;
+ context->ib_cpu[len++] = 0x00000000;
+ context->ib_cpu[len++] = 0x00000000;
+ context->ib_cpu[len++] = 0x00000000;
+ context->ib_cpu[len++] = 0x00000000;
+ context->ib_cpu[len++] = 0xffffffff;
+ context->ib_cpu[len++] = 0x00000000;
+ context->ib_cpu[len++] = 0x00000000;
+ context->ib_cpu[len++] = 0x00000000;
+ context->ib_cpu[len++] = 0x00000000;
+ }
+ *st_size = (len - st_offset) * 4;
+
+ /* encode context */
+ st_offset = len;
+ st_size = &context->ib_cpu[len++]; /* size */
+ if (context->vcn_ip_version_major == 1)
+ context->ib_cpu[len++] = 0x0000000d; /* ENCODE_CONTEXT_BUFFER vcn 1 */
+ else
+ context->ib_cpu[len++] = 0x00000011; /* ENCODE_CONTEXT_BUFFER vcn 2,3 */
+ context->ib_cpu[len++] = v_context->cpb_buf.addr >> 32;
+ context->ib_cpu[len++] = v_context->cpb_buf.addr;
+ context->ib_cpu[len++] = 0x00000000; /* swizzle mode */
+ context->ib_cpu[len++] = 0x00000100; /* luma pitch */
+ context->ib_cpu[len++] = 0x00000100; /* chroma pitch */
+ context->ib_cpu[len++] = 0x00000003; /* no reconstructed picture */
+ context->ib_cpu[len++] = 0x00000000; /* reconstructed pic 1 luma offset */
+ context->ib_cpu[len++] = ALIGN(width, 256) * ALIGN(height, 32); /* pic1 chroma offset */
+ context->ib_cpu[len++] = ALIGN(width, 256) * ALIGN(height, 32) * 3 / 2; /* pic2 luma offset */
+ context->ib_cpu[len++] = ALIGN(width, 256) * ALIGN(height, 32) * 5 / 2; /* pic2 chroma offset */
+
+ for (i = 0; i < 136; i++)
+ context->ib_cpu[len++] = 0x00000000;
+ *st_size = (len - st_offset) * 4;
+
+ /* bitstream buffer */
+ st_offset = len;
+ st_size = &context->ib_cpu[len++]; /* size */
+ if (context->vcn_ip_version_major == 1)
+ context->ib_cpu[len++] = 0x0000000e; /* VIDEO_BITSTREAM_BUFFER vcn 1 */
+ else
+ context->ib_cpu[len++] = 0x00000012; /* VIDEO_BITSTREAM_BUFFER vcn 2,3 */
+ context->ib_cpu[len++] = 0x00000000; /* mode */
+ context->ib_cpu[len++] = bs_buf.addr >> 32;
+ context->ib_cpu[len++] = bs_buf.addr;
+ context->ib_cpu[len++] = 0x0001f000;
+ context->ib_cpu[len++] = 0x00000000;
+ *st_size = (len - st_offset) * 4;
+
+ /* feedback */
+ st_offset = len;
+ st_size = &context->ib_cpu[len++]; /* size */
+ if (context->vcn_ip_version_major == 1)
+ context->ib_cpu[len++] = 0x00000010; /* FEEDBACK_BUFFER vcn 1 */
+ else
+ context->ib_cpu[len++] = 0x00000015; /* FEEDBACK_BUFFER vcn 2,3 */
+ context->ib_cpu[len++] = 0x00000000;
+ context->ib_cpu[len++] = fb_buf.addr >> 32;
+ context->ib_cpu[len++] = fb_buf.addr;
+ context->ib_cpu[len++] = 0x00000010;
+ context->ib_cpu[len++] = 0x00000028;
+ *st_size = (len - st_offset) * 4;
+
+ /* intra refresh */
+ st_offset = len;
+ st_size = &context->ib_cpu[len++];
+ if (context->vcn_ip_version_major == 1)
+ context->ib_cpu[len++] = 0x0000000c; /* INTRA_REFRESH vcn 1 */
+ else
+ context->ib_cpu[len++] = 0x00000010; /* INTRA_REFRESH vcn 2,3 */
+ context->ib_cpu[len++] = 0x00000000;
+ context->ib_cpu[len++] = 0x00000000;
+ context->ib_cpu[len++] = 0x00000000;
+ *st_size = (len - st_offset) * 4;
+
+ if (context->vcn_ip_version_major != 1) {
+ /* Input Format */
+ st_offset = len;
+ st_size = &context->ib_cpu[len++];
+ context->ib_cpu[len++] = 0x0000000c;
+ context->ib_cpu[len++] = 0x00000000; /* RENCODE_COLOR_VOLUME_G22_BT709 */
+ context->ib_cpu[len++] = 0x00000000;
+ context->ib_cpu[len++] = 0x00000000;
+ context->ib_cpu[len++] = 0x00000000;
+ context->ib_cpu[len++] = 0x00000000;
+ context->ib_cpu[len++] = 0x00000000; /* RENCODE_COLOR_BIT_DEPTH_8_BIT */
+ context->ib_cpu[len++] = 0x00000000; /* RENCODE_COLOR_PACKING_FORMAT_NV12 */
+ *st_size = (len - st_offset) * 4;
+
+ /* Output Format */
+ st_offset = len;
+ st_size = &context->ib_cpu[len++];
+ context->ib_cpu[len++] = 0x0000000d;
+ context->ib_cpu[len++] = 0x00000000; /* RENCODE_COLOR_VOLUME_G22_BT709 */
+ context->ib_cpu[len++] = 0x00000000;
+ context->ib_cpu[len++] = 0x00000000;
+ context->ib_cpu[len++] = 0x00000000; /* RENCODE_COLOR_BIT_DEPTH_8_BIT */
+ *st_size = (len - st_offset) * 4;
+ }
+ /* op_speed */
+ st_offset = len;
+ st_size = &context->ib_cpu[len++];
+ context->ib_cpu[len++] = 0x01000006; /* SPEED_ENCODING_MODE */
+ *st_size = (len - st_offset) * 4;
+
+ /* op_enc */
+ st_offset = len;
+ st_size = &context->ib_cpu[len++];
+ context->ib_cpu[len++] = 0x01000003;
+ *st_size = (len - st_offset) * 4;
+
+ *p_task_size = (len - task_offset) * 4;
+
+ if (context->vcn_unified_ring)
+ amdgpu_cs_sq_ib_tail(v_context, context->ib_cpu + len);
+
+ r = submit(device_handle, context, len, AMDGPU_HW_IP_VCN_ENC);
+ igt_assert_eq(r, 0);
+
+ /* check result */
+ check_result(v_context, fb_buf, bs_buf, frame_type);
+
+ free_resource(&fb_buf);
+ free_resource(&bs_buf);
+ free_resource(&vbv_buf);
+}
+
+static void
+amdgpu_cs_vcn_enc_encode(amdgpu_device_handle device_handle,
+ struct mmd_context *context, struct vcn_context *v_context)
+{
+ amdgpu_cs_vcn_enc_encode_frame(device_handle, context, v_context, 2); /* IDR frame */
+}
+
+static void
+amdgpu_cs_vcn_enc_destroy(amdgpu_device_handle device_handle, struct mmd_context *context,
+ struct vcn_context *v_context)
+{
+ int len = 0, r;
+ uint32_t *p_task_size = NULL;
+ uint32_t task_offset = 0, st_offset;
+ uint32_t *st_size = NULL;
+ uint32_t fw_maj = 1, fw_min = 9;
+
+ if (context->vcn_ip_version_major == 2) {
+ fw_maj = 1;
+ fw_min = 1;
+ } else if (context->vcn_ip_version_major == 3) {
+ fw_maj = 1;
+ fw_min = 0;
+ }
+
+ context->num_resources = 0;
+ //alloc_resource(&v_context->enc_buf, 128 * 1024, AMDGPU_GEM_DOMAIN_GTT);
+ context->resources[context->num_resources++] = v_context->enc_buf.handle;
+ context->resources[context->num_resources++] = context->ib_handle;
+
+ if (context->vcn_unified_ring)
+ amdgpu_cs_sq_head(v_context, context->ib_cpu, &len, true);
+
+ /* session info */
+ st_offset = len;
+ st_size = &context->ib_cpu[len++]; /* size */
+ context->ib_cpu[len++] = 0x00000001; /* RENCODE_IB_PARAM_SESSION_INFO */
+ context->ib_cpu[len++] = ((fw_maj << 16) | (fw_min << 0));
+ context->ib_cpu[len++] = v_context->enc_buf.addr >> 32;
+ context->ib_cpu[len++] = v_context->enc_buf.addr;
+ context->ib_cpu[len++] = 1; /* RENCODE_ENGINE_TYPE_ENCODE; */
+ *st_size = (len - st_offset) * 4;
+
+ /* task info */
+ task_offset = len;
+ st_offset = len;
+ st_size = &context->ib_cpu[len++]; /* size */
+ context->ib_cpu[len++] = 0x00000002; /* RENCODE_IB_PARAM_TASK_INFO */
+ p_task_size = &context->ib_cpu[len++];
+ context->ib_cpu[len++] = v_context->enc_task_id++; /* task_id */
+ context->ib_cpu[len++] = 0; /* feedback */
+ *st_size = (len - st_offset) * 4;
+
+ /* op close */
+ st_offset = len;
+ st_size = &context->ib_cpu[len++];
+ context->ib_cpu[len++] = 0x01000002; /* RENCODE_IB_OP_CLOSE_SESSION */
+ *st_size = (len - st_offset) * 4;
+
+ *p_task_size = (len - task_offset) * 4;
+
+ if (context->vcn_unified_ring)
+ amdgpu_cs_sq_ib_tail(v_context, context->ib_cpu + len);
+
+ r = submit(device_handle, context, len, AMDGPU_HW_IP_VCN_ENC);
+ igt_assert_eq(r, 0);
+
+ free_resource(&v_context->cpb_buf);
+ free_resource(&v_context->enc_buf);
+}
+
+igt_main
+{
+ amdgpu_device_handle device;
+ struct mmd_context context = {};
+ struct vcn_context v_context = {};
+ int fd = -1;
+
+ igt_fixture {
+ uint32_t major, minor;
+ int err;
+
+ fd = drm_open_driver(DRIVER_AMDGPU);
+ err = amdgpu_device_initialize(fd, &major, &minor, &device);
+ igt_require(err == 0);
+ igt_info("Initialized amdgpu, driver version %d.%d\n",
+ major, minor);
+ err = mmd_context_init(device, &context);
+ igt_require(err == 0);
+ igt_skip_on(!is_vcn_tests_enable(device, &context));
+ igt_skip_on_f(!context.dec_ring && !context.enc_ring, "vcn no dec and enc rings\n");
+ }
+
+ igt_describe("Test whether vcn dec is created, decodes, destroyed");
+ igt_subtest_with_dynamic("amdgpu_cs_vcn_dec_create, decode, destroy") {
+ if (context.dec_ring) {
+ amdgpu_cs_vcn_dec_create(device, &context, &v_context);
+ amdgpu_cs_vcn_dec_decode(device, &context, &v_context);
+ amdgpu_cs_vcn_dec_destroy(device, &context, &v_context);
+ }
+ }
+
+ igt_describe("Test whether vcn enc is created, decodes, destroyed");
+ igt_subtest_with_dynamic("amdgpu_cs_vcn_enc_create, encode, destroy") {
+ if (context.enc_ring) {
+ amdgpu_cs_vcn_enc_create(device, &context, &v_context);
+ amdgpu_cs_vcn_enc_encode(device, &context, &v_context);
+ amdgpu_cs_vcn_enc_destroy(device, &context, &v_context);
+ }
+ }
+
+ igt_fixture {
+ mmd_context_clean(device, &context);
+ amdgpu_device_deinitialize(device);
+ drm_close_driver(fd);
+ }
+
+}
diff --git a/tests/amdgpu/meson.build b/tests/amdgpu/meson.build
index e6e6eaabf..d09a4b6e7 100644
--- a/tests/amdgpu/meson.build
+++ b/tests/amdgpu/meson.build
@@ -27,6 +27,7 @@ if libdrm_amdgpu.found()
'amd_uvd_dec',
'amd_uvd_enc',
'amd_vce_dec',
+ 'amd_vcn',
'amd_vrr_range',
]
amdgpu_deps += libdrm_amdgpu
--
2.25.1
More information about the igt-dev
mailing list