[virglrenderer-devel] [PATCH 6/9] renderer: expose compute shaders to the host. (v2)
Dave Airlie
airlied at gmail.com
Tue Jul 31 22:28:18 UTC 2018
From: Dave Airlie <airlied at redhat.com>
Add a capability and some queries to expose compute shaders
to the host.
v2: expose more limits, check feature on launch grid
---
src/virgl_hw.h | 5 +++++
src/vrend_renderer.c | 23 +++++++++++++++++++++++
2 files changed, 28 insertions(+)
diff --git a/src/virgl_hw.h b/src/virgl_hw.h
index 69b9394..661be1e 100644
--- a/src/virgl_hw.h
+++ b/src/virgl_hw.h
@@ -224,6 +224,7 @@ enum virgl_formats {
#define VIRGL_CAP_TGSI_PRECISE (1 << 4)
#define VIRGL_CAP_TXQS (1 << 5)
#define VIRGL_CAP_MEMORY_BARRIER (1 << 6)
+#define VIRGL_CAP_COMPUTE_SHADER (1 << 7)
/* virgl bind flags - these are compatible with mesa 10.5 gallium.
* but are fixed, no other should be passed to virgl either.
@@ -337,6 +338,10 @@ struct virgl_caps_v2 {
uint32_t max_shader_image_frag_compute;
uint32_t max_shader_image_other_stages;
uint32_t max_image_samples;
+ uint32_t max_compute_work_group_invocations;
+ uint32_t max_compute_shared_memory_size;
+ uint32_t max_compute_grid_size[3];
+ uint32_t max_compute_block_size[3];
};
union virgl_caps {
diff --git a/src/vrend_renderer.c b/src/vrend_renderer.c
index f1571e0..3241ac3 100644
--- a/src/vrend_renderer.c
+++ b/src/vrend_renderer.c
@@ -94,6 +94,7 @@ enum features_id
feat_base_instance,
feat_barrier,
feat_bit_encoding,
+ feat_compute_shader,
feat_copy_image,
feat_conditional_render_inverted,
feat_cube_map_array,
@@ -150,6 +151,7 @@ static const struct {
[feat_base_instance] = { 42, UNAVAIL, { "GL_ARB_base_instance", "GL_EXT_base_instance" } },
[feat_barrier] = { 42, 31, {} },
[feat_bit_encoding] = { 33, UNAVAIL, { "GL_ARB_shader_bit_encoding" } },
+ [feat_compute_shader] = { 43, 31, { "GL_ARB_compute_shader" } },
[feat_copy_image] = { 43, 32, { "GL_ARB_copy_image", "GL_EXT_copy_image", "GL_OES_copy_image" } },
[feat_conditional_render_inverted] = { 45, UNAVAIL, { "GL_ARB_conditional_render_inverted" } },
[feat_cube_map_array] = { 40, UNAVAIL, { "GL_ARB_texture_cube_map_array", "GL_EXT_texture_cube_map_array", "GL_OES_texture_cube_map_array" } },
@@ -2898,6 +2900,10 @@ int vrend_create_shader(struct vrend_context *ctx,
type == PIPE_SHADER_TESS_EVAL))
return EINVAL;
+ if (!has_feature(feat_compute_shader) &&
+ type == PIPE_SHADER_COMPUTE)
+ return EINVAL;
+
if (offlen & VIRGL_OBJ_SHADER_OFFSET_CONT)
new_shader = false;
else if (((offlen + 3) / 4) > pkt_length)
@@ -3936,6 +3942,10 @@ void vrend_launch_grid(struct vrend_context *ctx,
{
bool new_program = false;
struct vrend_resource *indirect_res = NULL;
+
+ if (!has_feature(feat_compute_shader))
+ return;
+
if (ctx->sub->cs_shader_dirty) {
struct vrend_linked_shader_program *prog;
bool same_prog, cs_dirty;
@@ -8164,6 +8174,19 @@ void vrend_renderer_fill_caps(uint32_t set, UNUSED uint32_t version,
if (gl_ver >= 44)
glGetIntegerv(GL_MAX_VERTEX_ATTRIB_STRIDE, (GLint*)&caps->v2.max_vertex_attrib_stride);
+ if (has_feature(feat_compute_shader)) {
+ glGetIntegerv(GL_MAX_COMPUTE_WORK_GROUP_INVOCATIONS, (GLint*)&caps->v2.max_compute_work_group_invocations);
+ glGetIntegerv(GL_MAX_COMPUTE_SHARED_MEMORY_SIZE, (GLint*)&caps->v2.max_compute_shared_memory_size);
+ glGetIntegeri_v(GL_MAX_COMPUTE_WORK_GROUP_COUNT, 0, (GLint*)&caps->v2.max_compute_grid_size[0]);
+ glGetIntegeri_v(GL_MAX_COMPUTE_WORK_GROUP_COUNT, 1, (GLint*)&caps->v2.max_compute_grid_size[1]);
+ glGetIntegeri_v(GL_MAX_COMPUTE_WORK_GROUP_COUNT, 2, (GLint*)&caps->v2.max_compute_grid_size[2]);
+ glGetIntegeri_v(GL_MAX_COMPUTE_WORK_GROUP_SIZE, 0, (GLint*)&caps->v2.max_compute_block_size[0]);
+ glGetIntegeri_v(GL_MAX_COMPUTE_WORK_GROUP_SIZE, 1, (GLint*)&caps->v2.max_compute_block_size[1]);
+ glGetIntegeri_v(GL_MAX_COMPUTE_WORK_GROUP_SIZE, 2, (GLint*)&caps->v2.max_compute_block_size[2]);
+
+ caps->v2.capability_bits |= VIRGL_CAP_COMPUTE_SHADER;
+ }
+
if (has_feature(feat_texture_view))
caps->v2.capability_bits |= VIRGL_CAP_TEXTURE_VIEW;
--
2.14.3
More information about the virglrenderer-devel
mailing list