Mesa (master): nir: Add a lowering pass for non-uniform resource access

GitLab Mirror gitlab-mirror at kemper.freedesktop.org
Mon Mar 25 20:00:47 UTC 2019


Module: Mesa
Branch: master
Commit: 3bd54576415130465f096d73b7940dfbe02bb71b
URL:    http://cgit.freedesktop.org/mesa/mesa/commit/?id=3bd54576415130465f096d73b7940dfbe02bb71b

Author: Jason Ekstrand <jason.ekstrand at intel.com>
Date:   Wed Feb 27 14:36:44 2019 -0600

nir: Add a lowering pass for non-uniform resource access

Reviewed-by: Bas Nieuwenhuizen <bas at basnieuwenhuizen.nl>

---

 src/compiler/Makefile.sources                   |   1 +
 src/compiler/nir/meson.build                    |   1 +
 src/compiler/nir/nir.h                          |  16 ++
 src/compiler/nir/nir_lower_non_uniform_access.c | 265 ++++++++++++++++++++++++
 src/compiler/shader_enums.h                     |   3 +
 5 files changed, 286 insertions(+)

diff --git a/src/compiler/Makefile.sources b/src/compiler/Makefile.sources
index 5ac4d0dc297..e542d86a37a 100644
--- a/src/compiler/Makefile.sources
+++ b/src/compiler/Makefile.sources
@@ -255,6 +255,7 @@ NIR_FILES = \
 	nir/nir_lower_io_to_temporaries.c \
 	nir/nir_lower_io_to_scalar.c \
 	nir/nir_lower_io_to_vector.c \
+	nir/nir_lower_non_uniform_access.c \
 	nir/nir_lower_packing.c \
 	nir/nir_lower_passthrough_edgeflags.c \
 	nir/nir_lower_patch_vertices.c \
diff --git a/src/compiler/nir/meson.build b/src/compiler/nir/meson.build
index 510e99c779b..eecc19e7a6e 100644
--- a/src/compiler/nir/meson.build
+++ b/src/compiler/nir/meson.build
@@ -136,6 +136,7 @@ files_libnir = files(
   'nir_lower_io_to_temporaries.c',
   'nir_lower_io_to_scalar.c',
   'nir_lower_io_to_vector.c',
+  'nir_lower_non_uniform_access.c',
   'nir_lower_packing.c',
   'nir_lower_passthrough_edgeflags.c',
   'nir_lower_patch_vertices.c',
diff --git a/src/compiler/nir/nir.h b/src/compiler/nir/nir.h
index 1033b545d6a..40c5dcfe4ac 100644
--- a/src/compiler/nir/nir.h
+++ b/src/compiler/nir/nir.h
@@ -1485,6 +1485,12 @@ typedef struct {
    /* gather offsets */
    int8_t tg4_offsets[4][2];
 
+   /* True if the texture index or handle is not dynamically uniform */
+   bool texture_non_uniform;
+
+   /* True if the sampler index or handle is not dynamically uniform */
+   bool sampler_non_uniform;
+
    /** The texture index
     *
     * If this texture instruction has a nir_tex_src_texture_offset source,
@@ -3259,6 +3265,16 @@ typedef struct nir_lower_tex_options {
 bool nir_lower_tex(nir_shader *shader,
                    const nir_lower_tex_options *options);
 
+enum nir_lower_non_uniform_access_type {
+   nir_lower_non_uniform_ubo_access     = (1 << 0),
+   nir_lower_non_uniform_ssbo_access    = (1 << 1),
+   nir_lower_non_uniform_texture_access = (1 << 2),
+   nir_lower_non_uniform_image_access   = (1 << 3),
+};
+
+bool nir_lower_non_uniform_access(nir_shader *shader,
+                                  enum nir_lower_non_uniform_access_type);
+
 bool nir_lower_idiv(nir_shader *shader);
 
 bool nir_lower_clip_vs(nir_shader *shader, unsigned ucp_enables, bool use_vars);
diff --git a/src/compiler/nir/nir_lower_non_uniform_access.c b/src/compiler/nir/nir_lower_non_uniform_access.c
new file mode 100644
index 00000000000..6aa11f9bebe
--- /dev/null
+++ b/src/compiler/nir/nir_lower_non_uniform_access.c
@@ -0,0 +1,265 @@
+/*
+ * Copyright © 2019 Intel Corporation
+ *
+ * Permission is hereby granted, free of charge, to any person obtaining a
+ * copy of this software and associated documentation files (the "Software"),
+ * to deal in the Software without restriction, including without limitation
+ * the rights to use, copy, modify, merge, publish, distribute, sublicense,
+ * and/or sell copies of the Software, and to permit persons to whom the
+ * Software is furnished to do so, subject to the following conditions:
+ *
+ * The above copyright notice and this permission notice (including the next
+ * paragraph) shall be included in all copies or substantial portions of the
+ * Software.
+ *
+ * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
+ * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
+ * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.  IN NO EVENT SHALL
+ * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
+ * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
+ * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS
+ * IN THE SOFTWARE.
+ */
+
+#include "nir.h"
+#include "nir_builder.h"
+
+static nir_ssa_def *
+read_first_invocation(nir_builder *b, nir_ssa_def *x)
+{
+   nir_intrinsic_instr *first =
+      nir_intrinsic_instr_create(b->shader,
+                                 nir_intrinsic_read_first_invocation);
+   first->num_components = x->num_components;
+   first->src[0] = nir_src_for_ssa(x);
+   nir_ssa_dest_init(&first->instr, &first->dest,
+                     x->num_components, x->bit_size, NULL);
+   return &first->dest.ssa;
+}
+
+static bool
+lower_non_uniform_tex_access(nir_builder *b, nir_tex_instr *tex)
+{
+   if (!tex->texture_non_uniform && !tex->sampler_non_uniform)
+      return false;
+
+   /* We can have at most one texture and one sampler handle */
+   nir_ssa_def *handles[2];
+   unsigned handle_count = 0;
+   for (unsigned i = 0; i < tex->num_srcs; i++) {
+      switch (tex->src[i].src_type) {
+      case nir_tex_src_texture_offset:
+      case nir_tex_src_texture_handle:
+         if (!tex->texture_non_uniform)
+            continue;
+         break;
+
+      case nir_tex_src_sampler_offset:
+      case nir_tex_src_sampler_handle:
+         if (!tex->sampler_non_uniform)
+            continue;
+         break;
+
+      default:
+         continue;
+      }
+
+      assert(tex->src[i].src.is_ssa);
+      assert(tex->src[i].src.ssa->num_components == 1);
+      assert(handle_count < 2);
+      handles[handle_count++] = tex->src[i].src.ssa;
+   }
+
+   if (handle_count == 0)
+      return false;
+
+   b->cursor = nir_instr_remove(&tex->instr);
+
+   nir_push_loop(b);
+
+   nir_ssa_def *all_equal_first = nir_imm_true(b);
+   for (unsigned i = 0; i < handle_count; i++) {
+      nir_ssa_def *equal_first =
+         nir_ieq(b, read_first_invocation(b, handles[i]), handles[i]);
+      all_equal_first = nir_iand(b, all_equal_first, equal_first);
+   }
+
+   nir_push_if(b, all_equal_first);
+
+   nir_builder_instr_insert(b, &tex->instr);
+   nir_jump(b, nir_jump_break);
+
+   return true;
+}
+
+static bool
+lower_non_uniform_access_intrin(nir_builder *b, nir_intrinsic_instr *intrin,
+                                unsigned handle_src)
+{
+   if (!(nir_intrinsic_access(intrin) & ACCESS_NON_UNIFORM))
+      return false;
+
+   /* If it's constant, it's automatically uniform; don't bother. */
+   if (nir_src_is_const(intrin->src[handle_src]))
+      return false;
+
+   b->cursor = nir_instr_remove(&intrin->instr);
+
+   nir_push_loop(b);
+
+   assert(intrin->src[handle_src].is_ssa);
+   assert(intrin->src[handle_src].ssa->num_components == 1);
+   nir_ssa_def *handle = intrin->src[handle_src].ssa;
+
+   nir_push_if(b, nir_ieq(b, read_first_invocation(b, handle), handle));
+
+   nir_builder_instr_insert(b, &intrin->instr);
+   nir_jump(b, nir_jump_break);
+
+   return true;
+}
+
+static bool
+nir_lower_non_uniform_access_impl(nir_function_impl *impl,
+                                  enum nir_lower_non_uniform_access_type types)
+{
+   bool progress = false;
+
+   nir_builder b;
+   nir_builder_init(&b, impl);
+
+   nir_foreach_block(block, impl) {
+      nir_foreach_instr(instr, block) {
+         switch (instr->type) {
+         case nir_instr_type_tex: {
+            nir_tex_instr *tex = nir_instr_as_tex(instr);
+            if ((types & nir_lower_non_uniform_texture_access) &&
+                lower_non_uniform_tex_access(&b, tex))
+               progress = true;
+            break;
+         }
+
+         case nir_instr_type_intrinsic: {
+            nir_intrinsic_instr *intrin = nir_instr_as_intrinsic(instr);
+            switch (intrin->intrinsic) {
+            case nir_intrinsic_load_ubo:
+               if ((types & nir_lower_non_uniform_ubo_access) &&
+                   lower_non_uniform_access_intrin(&b, intrin, 0))
+                  progress = true;
+               break;
+
+            case nir_intrinsic_load_ssbo:
+            case nir_intrinsic_ssbo_atomic_add:
+            case nir_intrinsic_ssbo_atomic_imin:
+            case nir_intrinsic_ssbo_atomic_umin:
+            case nir_intrinsic_ssbo_atomic_imax:
+            case nir_intrinsic_ssbo_atomic_umax:
+            case nir_intrinsic_ssbo_atomic_and:
+            case nir_intrinsic_ssbo_atomic_or:
+            case nir_intrinsic_ssbo_atomic_xor:
+            case nir_intrinsic_ssbo_atomic_exchange:
+            case nir_intrinsic_ssbo_atomic_comp_swap:
+            case nir_intrinsic_ssbo_atomic_fadd:
+            case nir_intrinsic_ssbo_atomic_fmin:
+            case nir_intrinsic_ssbo_atomic_fmax:
+            case nir_intrinsic_ssbo_atomic_fcomp_swap:
+               if ((types & nir_lower_non_uniform_ssbo_access) &&
+                   lower_non_uniform_access_intrin(&b, intrin, 0))
+                  progress = true;
+               break;
+
+            case nir_intrinsic_store_ssbo:
+               /* SSBO Stores put the index in the second source */
+               if ((types & nir_lower_non_uniform_ssbo_access) &&
+                   lower_non_uniform_access_intrin(&b, intrin, 1))
+                  progress = true;
+               break;
+
+            case nir_intrinsic_image_load:
+            case nir_intrinsic_image_store:
+            case nir_intrinsic_image_atomic_add:
+            case nir_intrinsic_image_atomic_min:
+            case nir_intrinsic_image_atomic_max:
+            case nir_intrinsic_image_atomic_and:
+            case nir_intrinsic_image_atomic_or:
+            case nir_intrinsic_image_atomic_xor:
+            case nir_intrinsic_image_atomic_exchange:
+            case nir_intrinsic_image_atomic_comp_swap:
+            case nir_intrinsic_image_atomic_fadd:
+            case nir_intrinsic_image_size:
+            case nir_intrinsic_image_samples:
+            case nir_intrinsic_bindless_image_load:
+            case nir_intrinsic_bindless_image_store:
+            case nir_intrinsic_bindless_image_atomic_add:
+            case nir_intrinsic_bindless_image_atomic_min:
+            case nir_intrinsic_bindless_image_atomic_max:
+            case nir_intrinsic_bindless_image_atomic_and:
+            case nir_intrinsic_bindless_image_atomic_or:
+            case nir_intrinsic_bindless_image_atomic_xor:
+            case nir_intrinsic_bindless_image_atomic_exchange:
+            case nir_intrinsic_bindless_image_atomic_comp_swap:
+            case nir_intrinsic_bindless_image_atomic_fadd:
+            case nir_intrinsic_bindless_image_size:
+            case nir_intrinsic_bindless_image_samples:
+               if ((types & nir_lower_non_uniform_image_access) &&
+                   lower_non_uniform_access_intrin(&b, intrin, 0))
+                  progress = true;
+               break;
+
+            default:
+               /* Nothing to do */
+               break;
+            }
+            break;
+         }
+
+         default:
+            /* Nothing to do */
+            break;
+         }
+      }
+   }
+
+   if (progress)
+      nir_metadata_preserve(impl, nir_metadata_none);
+
+   return progress;
+}
+
+/**
+ * Lowers non-uniform resource access by using a loop
+ *
+ * This pass lowers non-uniform resource access by using subgroup operations
+ * and a loop.  Most hardware requires things like textures and UBO access
+ * operations to happen on a dynamically uniform (or at least subgroup
+ * uniform) resource.  This pass allows for non-uniform access by placing the
+ * texture instruction in a loop that looks something like this:
+ *
+ * loop {
+ *    bool tex_eq_first = readFirstInvocationARB(texture) == texture;
+ *    bool smp_eq_first = readFirstInvocationARB(sampler) == sampler;
+ *    if (tex_eq_first && smp_eq_first) {
+ *       res = texture(texture, sampler, ...);
+ *       break;
+ *    }
+ * }
+ *
+ * Fortunately, because the instruction is immediately followed by the only
+ * break in the loop, the block containing the instruction dominates the end
+ * of the loop.  Therefore, it's safe to move the instruction into the loop
+ * without fixing up SSA in any way.
+ */
+bool
+nir_lower_non_uniform_access(nir_shader *shader,
+                             enum nir_lower_non_uniform_access_type types)
+{
+   bool progress = false;
+
+   nir_foreach_function(function, shader) {
+      if (function->impl &&
+          nir_lower_non_uniform_access_impl(function->impl, types))
+         progress = true;
+   }
+
+   return progress;
+}
diff --git a/src/compiler/shader_enums.h b/src/compiler/shader_enums.h
index fe26d03a920..883b2825146 100644
--- a/src/compiler/shader_enums.h
+++ b/src/compiler/shader_enums.h
@@ -716,6 +716,9 @@ enum gl_access_qualifier
    ACCESS_VOLATILE      = (1 << 2),
    ACCESS_NON_READABLE  = (1 << 3),
    ACCESS_NON_WRITEABLE = (1 << 4),
+
+   /** The access may use a non-uniform buffer or image index */
+   ACCESS_NON_UNIFORM   = (1 << 5),
 };
 
 /**




More information about the mesa-commit mailing list