[PATCH libdrm] drm/amdgpu: add new low overhead command submission API. (v2)

Christian König deathsimple at vodafone.de
Tue Jul 18 13:59:41 UTC 2017


Am 18.07.2017 um 05:52 schrieb Dave Airlie:
> From: Dave Airlie <airlied at redhat.com>
>
> This just sends chunks to the kernel API for a single command
> stream.
>
> This should provide a more future proof and extensible API
> for command submission.
>
> v2: use amdgpu_bo_list_handle, add two helper functions to
> access bo and context internals.
>
> Signed-off-by: Dave Airlie <airlied at redhat.com>

Reviewed-by: Christian König <christian.koenig at amd.com>

> ---
>   amdgpu/amdgpu.h    | 30 ++++++++++++++++++++++++++++++
>   amdgpu/amdgpu_cs.c | 47 +++++++++++++++++++++++++++++++++++++++++++++++
>   2 files changed, 77 insertions(+)
>
> diff --git a/amdgpu/amdgpu.h b/amdgpu/amdgpu.h
> index 183f974..238b1aa 100644
> --- a/amdgpu/amdgpu.h
> +++ b/amdgpu/amdgpu.h
> @@ -1382,6 +1382,36 @@ int amdgpu_cs_import_syncobj(amdgpu_device_handle dev,
>   			     int shared_fd,
>   			     uint32_t *syncobj);
>   
> +/**
> + *  Submit raw command submission to kernel
> + *
> + * \param   dev	       - \c [in] device handle
> + * \param   context    - \c [in] context handle for context id
> + * \param   bo_list_handle - \c [in] request bo list handle (0 for none)
> + * \param   num_chunks - \c [in] number of CS chunks to submit
> + * \param   chunks     - \c [in] array of CS chunks
> + * \param   seq_no     - \c [out] output sequence number for submission.
> + *
> + * \return   0 on success\n
> + *          <0 - Negative POSIX Error code
> + *
> + */
> +struct drm_amdgpu_cs_chunk;
> +struct drm_amdgpu_cs_chunk_dep;
> +struct drm_amdgpu_cs_chunk_data;
> +
> +int amdgpu_cs_submit_raw(amdgpu_device_handle dev,
> +			 amdgpu_context_handle context,
> +			 amdgpu_bo_list_handle bo_list_handle,
> +			 int num_chunks,
> +			 struct drm_amdgpu_cs_chunk *chunks,
> +			 uint64_t *seq_no);
> +
> +void amdgpu_cs_chunk_fence_to_dep(struct amdgpu_cs_fence *fence,
> +				  struct drm_amdgpu_cs_chunk_dep *dep);
> +void amdgpu_cs_chunk_fence_info_to_data(struct amdgpu_cs_fence_info *fence_info,
> +					struct drm_amdgpu_cs_chunk_data *data);
> +
>   #ifdef __cplusplus
>   }
>   #endif
> diff --git a/amdgpu/amdgpu_cs.c b/amdgpu/amdgpu_cs.c
> index 722fd75..dfba875 100644
> --- a/amdgpu/amdgpu_cs.c
> +++ b/amdgpu/amdgpu_cs.c
> @@ -634,3 +634,50 @@ int amdgpu_cs_import_syncobj(amdgpu_device_handle dev,
>   
>   	return drmSyncobjFDToHandle(dev->fd, shared_fd, handle);
>   }
> +
> +int amdgpu_cs_submit_raw(amdgpu_device_handle dev,
> +			 amdgpu_context_handle context,
> +			 amdgpu_bo_list_handle bo_list_handle,
> +			 int num_chunks,
> +			 struct drm_amdgpu_cs_chunk *chunks,
> +			 uint64_t *seq_no)
> +{
> +	union drm_amdgpu_cs cs = {0};
> +	uint64_t *chunk_array;
> +	int i, r;
> +	if (num_chunks == 0)
> +		return -EINVAL;
> +
> +	chunk_array = alloca(sizeof(uint64_t) * num_chunks);
> +	for (i = 0; i < num_chunks; i++)
> +		chunk_array[i] = (uint64_t)(uintptr_t)&chunks[i];
> +	cs.in.chunks = (uint64_t)(uintptr_t)chunk_array;
> +	cs.in.ctx_id = context->id;
> +	cs.in.bo_list_handle = bo_list_handle ? bo_list_handle->handle : 0;
> +	cs.in.num_chunks = num_chunks;
> +	r = drmCommandWriteRead(dev->fd, DRM_AMDGPU_CS,
> +				&cs, sizeof(cs));
> +	if (r)
> +		return r;
> +
> +	if (seq_no)
> +		*seq_no = cs.out.handle;
> +	return 0;
> +}
> +
> +void amdgpu_cs_chunk_fence_info_to_data(struct amdgpu_cs_fence_info *fence_info,
> +					struct drm_amdgpu_cs_chunk_data *data)
> +{
> +	data->fence_data.handle = fence_info->handle->handle;
> +	data->fence_data.offset = fence_info->offset * sizeof(uint64_t);
> +}
> +
> +void amdgpu_cs_chunk_fence_to_dep(struct amdgpu_cs_fence *fence,
> +				  struct drm_amdgpu_cs_chunk_dep *dep)
> +{
> +	dep->ip_type = fence->ip_type;
> +	dep->ip_instance = fence->ip_instance;
> +	dep->ring = fence->ring;
> +	dep->ctx_id = fence->context->id;
> +	dep->handle = fence->fence;
> +}




More information about the dri-devel mailing list