[committed] Support lastprivate(conditional:) on combined for simd
Jakub Jelinek
jakub@redhat.com
Tue Jun 4 12:50:00 GMT 2019
Hi!
The following patch adds support for lastprivate(conditional:) on for simd.
Bootstrapped/regtested on x86_64-linux and i686-linux, committed to trunk.
2019-06-04 Jakub Jelinek <jakub@redhat.com>
* gimplify.c (gimplify_scan_omp_clauses): Don't sorry_at on lastprivate
conditional on combined for simd.
* omp-low.c (struct omp_context): Add combined_into_simd_safelen0
member.
(lower_rec_input_clauses): For gimple_omp_for_combined_into_p max_vf 1
constructs, don't remove lastprivate_conditional_map, but instead set
ctx->combined_into_simd_safelen0 and adjust hash_map, so that it points
to parent construct temporaries.
(lower_lastprivate_clauses): Handle ctx->combined_into_simd_safelen0
like !ctx->lastprivate_conditional_map.
(lower_omp_1) <case GIMPLE_ASSIGN>: If up->combined_into_simd_safelen0,
use up->outer context instead of up.
* omp-expand.c (expand_omp_for_generic): Perform cond_var bump even if
gimple_omp_for_combined_p.
(expand_omp_for_static_nochunk): Likewise.
(expand_omp_for_static_chunk): Add forgotten cond_var bump that was
probably moved over into expand_omp_for_generic rather than being copied
there.
gcc/cp/
* cp-tree.h (CP_OMP_CLAUSE_INFO): Allow for any clauses up to _condvar_
instead of only up to linear.
gcc/testsuite/
* c-c++-common/gomp/lastprivate-conditional-2.c (foo): Don't expect
a sorry_at on any of the clauses.
libgomp/
* testsuite/libgomp.c-c++-common/lastprivate-conditional-7.c: New test.
* testsuite/libgomp.c-c++-common/lastprivate-conditional-8.c: New test.
* testsuite/libgomp.c-c++-common/lastprivate-conditional-9.c: New test.
* testsuite/libgomp.c-c++-common/lastprivate-conditional-10.c: New test.
--- gcc/gimplify.c.jj 2019-05-31 23:37:51.066938833 +0200
+++ gcc/gimplify.c 2019-06-03 10:22:30.272829191 +0200
@@ -8145,31 +8145,7 @@ gimplify_scan_omp_clauses (tree *list_p,
OMP_CLAUSE_LASTPRIVATE_CONDITIONAL (c) = 0;
}
if (OMP_CLAUSE_LASTPRIVATE_CONDITIONAL (c))
- {
- splay_tree_node n = NULL;
- if (code == OMP_SIMD
- && outer_ctx
- && outer_ctx->region_type == ORT_WORKSHARE)
- {
- n = splay_tree_lookup (outer_ctx->variables,
- (splay_tree_key) decl);
- if (n == NULL
- && outer_ctx->outer_context
- && (outer_ctx->outer_context->region_type
- == ORT_COMBINED_PARALLEL))
- n = splay_tree_lookup (outer_ctx->outer_context->variables,
- (splay_tree_key) decl);
- }
- if (n && (n->value & GOVD_LASTPRIVATE_CONDITIONAL) != 0)
- {
- sorry_at (OMP_CLAUSE_LOCATION (c),
- "%<conditional%> modifier on %<lastprivate%> "
- "clause not supported yet");
- OMP_CLAUSE_LASTPRIVATE_CONDITIONAL (c) = 0;
- }
- else
- flags |= GOVD_LASTPRIVATE_CONDITIONAL;
- }
+ flags |= GOVD_LASTPRIVATE_CONDITIONAL;
if (outer_ctx
&& (outer_ctx->region_type == ORT_COMBINED_PARALLEL
|| ((outer_ctx->region_type & ORT_COMBINED_TEAMS)
--- gcc/omp-low.c.jj 2019-05-31 23:37:51.069938786 +0200
+++ gcc/omp-low.c 2019-06-03 15:12:45.535613313 +0200
@@ -137,6 +137,10 @@ struct omp_context
/* True if this construct can be cancelled. */
bool cancellable;
+
+ /* True if lower_omp_1 should look up lastprivate conditional in parent
+ context. */
+ bool combined_into_simd_safelen0;
};
static splay_tree all_contexts;
@@ -4816,6 +4820,8 @@ lower_rec_input_clauses (tree clauses, g
void_node);
gimple_seq tseq = NULL;
gimplify_and_add (x, &tseq);
+ if (ctx->outer)
+ lower_omp (&tseq, ctx->outer);
gimple_seq_add_seq (&llist[1], tseq);
}
if (y)
@@ -5278,11 +5284,31 @@ lower_rec_input_clauses (tree clauses, g
sctx.is_simt = false;
if (ctx->lastprivate_conditional_map)
{
- /* When not vectorized, treat lastprivate(conditional:) like
- normal lastprivate, as there will be just one simd lane
- writing the privatized variable. */
- delete ctx->lastprivate_conditional_map;
- ctx->lastprivate_conditional_map = NULL;
+ if (gimple_omp_for_combined_into_p (ctx->stmt))
+ {
+ /* Signal to lower_omp_1 that it should use parent context. */
+ ctx->combined_into_simd_safelen0 = true;
+ for (c = clauses; c ; c = OMP_CLAUSE_CHAIN (c))
+ if (OMP_CLAUSE_CODE (c) == OMP_CLAUSE_LASTPRIVATE
+ && OMP_CLAUSE_LASTPRIVATE_CONDITIONAL (c))
+ {
+ tree o = lookup_decl (OMP_CLAUSE_DECL (c), ctx);
+ tree *v
+ = ctx->lastprivate_conditional_map->get (o);
+ tree po = lookup_decl (OMP_CLAUSE_DECL (c), ctx->outer);
+ tree *pv
+ = ctx->outer->lastprivate_conditional_map->get (po);
+ *v = *pv;
+ }
+ }
+ else
+ {
+ /* When not vectorized, treat lastprivate(conditional:) like
+ normal lastprivate, as there will be just one simd lane
+ writing the privatized variable. */
+ delete ctx->lastprivate_conditional_map;
+ ctx->lastprivate_conditional_map = NULL;
+ }
}
}
@@ -5652,7 +5678,8 @@ lower_lastprivate_clauses (tree clauses,
if (OMP_CLAUSE_CODE (c) == OMP_CLAUSE_LASTPRIVATE
&& OMP_CLAUSE_LASTPRIVATE_CONDITIONAL (c)
- && ctx->lastprivate_conditional_map)
+ && ctx->lastprivate_conditional_map
+ && !ctx->combined_into_simd_safelen0)
{
gcc_assert (body_p);
if (simduid)
@@ -10812,6 +10839,8 @@ lower_omp_1 (gimple_stmt_iterator *gsi_p
if (tree *v = up->lastprivate_conditional_map->get (lhs))
{
tree clauses;
+ if (up->combined_into_simd_safelen0)
+ up = up->outer;
if (gimple_code (up->stmt) == GIMPLE_OMP_FOR)
clauses = gimple_omp_for_clauses (up->stmt);
else
--- gcc/omp-expand.c.jj 2019-05-31 23:37:51.071938754 +0200
+++ gcc/omp-expand.c 2019-06-03 12:04:20.342277395 +0200
@@ -3257,6 +3257,25 @@ expand_omp_for_generic (struct omp_regio
vmain = gimple_omp_continue_control_use (cont_stmt);
vback = gimple_omp_continue_control_def (cont_stmt);
+ if (cond_var)
+ {
+ tree itype = TREE_TYPE (cond_var);
+ tree t2;
+ if ((fd->ordered && fd->collapse == 1)
+ || bias
+ || POINTER_TYPE_P (type)
+ || TREE_CODE (fd->loop.n1) != INTEGER_CST
+ || fd->loop.cond_code != LT_EXPR)
+ t2 = build_int_cst (itype, 1);
+ else
+ t2 = fold_convert (itype, fd->loop.step);
+ t2 = fold_build2 (PLUS_EXPR, itype, cond_var, t2);
+ t2 = force_gimple_operand_gsi (&gsi, t2, false,
+ NULL_TREE, true, GSI_SAME_STMT);
+ assign_stmt = gimple_build_assign (cond_var, t2);
+ gsi_insert_before (&gsi, assign_stmt, GSI_SAME_STMT);
+ }
+
if (!gimple_omp_for_combined_p (fd->for_stmt))
{
if (POINTER_TYPE_P (type))
@@ -3270,25 +3289,6 @@ expand_omp_for_generic (struct omp_regio
assign_stmt = gimple_build_assign (vback, t);
gsi_insert_before (&gsi, assign_stmt, GSI_SAME_STMT);
- if (cond_var)
- {
- tree itype = TREE_TYPE (cond_var);
- tree t2;
- if ((fd->ordered && fd->collapse == 1)
- || bias
- || POINTER_TYPE_P (type)
- || TREE_CODE (fd->loop.n1) != INTEGER_CST
- || fd->loop.cond_code != LT_EXPR)
- t2 = build_int_cst (itype, 1);
- else
- t2 = fold_convert (itype, fd->loop.step);
- t2 = fold_build2 (PLUS_EXPR, itype, cond_var, t2);
- t2 = force_gimple_operand_gsi (&gsi, t2, false,
- NULL_TREE, true, GSI_SAME_STMT);
- assign_stmt = gimple_build_assign (cond_var, t2);
- gsi_insert_before (&gsi, assign_stmt, GSI_SAME_STMT);
- }
-
if (fd->ordered && counts[fd->collapse - 1] == NULL_TREE)
{
tree tem;
@@ -3962,6 +3962,23 @@ expand_omp_for_static_nochunk (struct om
vmain = gimple_omp_continue_control_use (cont_stmt);
vback = gimple_omp_continue_control_def (cont_stmt);
+ if (cond_var)
+ {
+ tree itype = TREE_TYPE (cond_var);
+ tree t2;
+ if (POINTER_TYPE_P (type)
+ || TREE_CODE (n1) != INTEGER_CST
+ || fd->loop.cond_code != LT_EXPR)
+ t2 = build_int_cst (itype, 1);
+ else
+ t2 = fold_convert (itype, step);
+ t2 = fold_build2 (PLUS_EXPR, itype, cond_var, t2);
+ t2 = force_gimple_operand_gsi (&gsi, t2, false,
+ NULL_TREE, true, GSI_SAME_STMT);
+ assign_stmt = gimple_build_assign (cond_var, t2);
+ gsi_insert_before (&gsi, assign_stmt, GSI_SAME_STMT);
+ }
+
if (!gimple_omp_for_combined_p (fd->for_stmt))
{
if (POINTER_TYPE_P (type))
@@ -3975,23 +3992,6 @@ expand_omp_for_static_nochunk (struct om
assign_stmt = gimple_build_assign (vback, t);
gsi_insert_before (&gsi, assign_stmt, GSI_SAME_STMT);
- if (cond_var)
- {
- tree itype = TREE_TYPE (cond_var);
- tree t2;
- if (POINTER_TYPE_P (type)
- || TREE_CODE (n1) != INTEGER_CST
- || fd->loop.cond_code != LT_EXPR)
- t2 = build_int_cst (itype, 1);
- else
- t2 = fold_convert (itype, step);
- t2 = fold_build2 (PLUS_EXPR, itype, cond_var, t2);
- t2 = force_gimple_operand_gsi (&gsi, t2, false,
- NULL_TREE, true, GSI_SAME_STMT);
- assign_stmt = gimple_build_assign (cond_var, t2);
- gsi_insert_before (&gsi, assign_stmt, GSI_SAME_STMT);
- }
-
t = build2 (fd->loop.cond_code, boolean_type_node,
DECL_P (vback) && TREE_ADDRESSABLE (vback)
? t : vback, e);
@@ -4607,6 +4607,23 @@ expand_omp_for_static_chunk (struct omp_
vmain = gimple_omp_continue_control_use (cont_stmt);
vback = gimple_omp_continue_control_def (cont_stmt);
+ if (cond_var)
+ {
+ tree itype = TREE_TYPE (cond_var);
+ tree t2;
+ if (POINTER_TYPE_P (type)
+ || TREE_CODE (n1) != INTEGER_CST
+ || fd->loop.cond_code != LT_EXPR)
+ t2 = build_int_cst (itype, 1);
+ else
+ t2 = fold_convert (itype, step);
+ t2 = fold_build2 (PLUS_EXPR, itype, cond_var, t2);
+ t2 = force_gimple_operand_gsi (&gsi, t2, false,
+ NULL_TREE, true, GSI_SAME_STMT);
+ assign_stmt = gimple_build_assign (cond_var, t2);
+ gsi_insert_before (&gsi, assign_stmt, GSI_SAME_STMT);
+ }
+
if (!gimple_omp_for_combined_p (fd->for_stmt))
{
if (POINTER_TYPE_P (type))
--- gcc/cp/cp-tree.h.jj 2019-05-31 23:35:38.461038702 +0200
+++ gcc/cp/cp-tree.h 2019-06-03 12:39:21.453395398 +0200
@@ -4924,7 +4924,7 @@ more_aggr_init_expr_args_p (const aggr_i
See semantics.c for details. */
#define CP_OMP_CLAUSE_INFO(NODE) \
TREE_TYPE (OMP_CLAUSE_RANGE_CHECK (NODE, OMP_CLAUSE_PRIVATE, \
- OMP_CLAUSE_LINEAR))
+ OMP_CLAUSE__CONDTEMP_))
/* Nonzero if this transaction expression's body contains statements. */
#define TRANSACTION_EXPR_IS_STMT(NODE) \
--- gcc/testsuite/c-c++-common/gomp/lastprivate-conditional-2.c.jj 2019-05-31 23:37:51.071938754 +0200
+++ gcc/testsuite/c-c++-common/gomp/lastprivate-conditional-2.c 2019-06-03 12:25:31.233599215 +0200
@@ -13,7 +13,7 @@ foo (int *p)
if (p[i])
b = i;
#pragma omp parallel
- #pragma omp for simd lastprivate (conditional: c) /* { dg-message "not supported yet" } */
+ #pragma omp for simd lastprivate (conditional: c)
for (i = 0; i < 32; i++)
if (p[i])
c = i;
@@ -21,7 +21,7 @@ foo (int *p)
for (i = 0; i < 32; i++)
if (p[i])
d = i;
- #pragma omp parallel for simd lastprivate (conditional: e) /* { dg-message "not supported yet" } */
+ #pragma omp parallel for simd lastprivate (conditional: e)
for (i = 0; i < 32; i++)
if (p[i])
e = i;
--- libgomp/testsuite/libgomp.c-c++-common/lastprivate-conditional-7.c.jj 2019-06-03 12:30:01.087307199 +0200
+++ libgomp/testsuite/libgomp.c-c++-common/lastprivate-conditional-7.c 2019-06-03 12:57:18.597254143 +0200
@@ -0,0 +1,63 @@
+/* { dg-do run } */
+/* { dg-additional-options "-O2 -fdump-tree-vect-details" } */
+/* { dg-additional-options "-mavx" { target avx_runtime } } */
+/* { dg-final { scan-tree-dump-times "vectorized 1 loops" 3 "vect" { target avx_runtime } } } */
+
+int v, x;
+
+__attribute__((noipa)) void
+foo (int *a)
+{
+ #pragma omp for simd lastprivate (conditional: x) schedule(simd: static)
+ for (int i = 0; i < 128; i++)
+ if (a[i])
+ x = a[i];
+}
+
+__attribute__((noipa)) void
+bar (int *a, int *b)
+{
+ #pragma omp for simd lastprivate (conditional: x, v) schedule(static, 16)
+ for (int i = 16; i < 128; ++i)
+ {
+ if (a[i])
+ x = a[i];
+ if (b[i])
+ v = b[i] + 10;
+ }
+}
+
+__attribute__((noipa)) void
+baz (int *a)
+{
+ #pragma omp for simd lastprivate (conditional: x) schedule(simd: dynamic, 16)
+ for (int i = 0; i < 128; i++)
+ if (a[i])
+ x = a[i] + 5;
+}
+
+int
+main ()
+{
+ int a[128], b[128], i;
+ for (i = 0; i < 128; i++)
+ {
+ a[i] = ((i % 11) == 2) ? i + 10 : 0;
+ asm volatile ("" : "+g" (i));
+ b[i] = ((i % 13) == 5) ? i * 2 : 0;
+ }
+ #pragma omp parallel
+ foo (a);
+ if (x != 133)
+ __builtin_abort ();
+ x = -3;
+ #pragma omp parallel
+ bar (b, a);
+ if (x != 244 || v != 143)
+ __builtin_abort ();
+ #pragma omp parallel
+ baz (b);
+ if (x != 249)
+ __builtin_abort ();
+ return 0;
+}
--- libgomp/testsuite/libgomp.c-c++-common/lastprivate-conditional-8.c.jj 2019-06-03 13:01:22.102373429 +0200
+++ libgomp/testsuite/libgomp.c-c++-common/lastprivate-conditional-8.c 2019-06-03 13:02:38.295159149 +0200
@@ -0,0 +1,60 @@
+/* { dg-do run } */
+/* { dg-additional-options "-O2 -fdump-tree-vect-details" } */
+/* { dg-additional-options "-mavx" { target avx_runtime } } */
+/* { dg-final { scan-tree-dump-times "vectorized 1 loops" 3 "vect" { target avx_runtime } } } */
+
+int v, x;
+
+__attribute__((noipa)) int
+foo (int *a)
+{
+ #pragma omp parallel for simd lastprivate (conditional: x) schedule(simd: static)
+ for (int i = 0; i < 128; i++)
+ if (a[i])
+ x = a[i];
+ return x;
+}
+
+__attribute__((noipa)) int
+bar (int *a, int *b)
+{
+ #pragma omp parallel
+ #pragma omp for simd lastprivate (conditional: x, v) schedule(static, 16)
+ for (int i = 16; i < 128; ++i)
+ {
+ if (a[i])
+ x = a[i];
+ if (b[i])
+ v = b[i] + 10;
+ }
+ return x;
+}
+
+__attribute__((noipa)) int
+baz (int *a)
+{
+ #pragma omp parallel for simd lastprivate (conditional: x) schedule(simd: dynamic, 16)
+ for (int i = 0; i < 128; i++)
+ if (a[i])
+ x = a[i] + 5;
+ return x;
+}
+
+int
+main ()
+{
+ int a[128], b[128], i;
+ for (i = 0; i < 128; i++)
+ {
+ a[i] = ((i % 11) == 2) ? i + 10 : 0;
+ asm volatile ("" : "+g" (i));
+ b[i] = ((i % 13) == 5) ? i * 2 : 0;
+ }
+ if (foo (a) != 133)
+ __builtin_abort ();
+ if (bar (b, a) != 244 || v != 143)
+ __builtin_abort ();
+ if (baz (b) != 249)
+ __builtin_abort ();
+ return 0;
+}
--- libgomp/testsuite/libgomp.c-c++-common/lastprivate-conditional-9.c.jj 2019-06-03 13:10:44.860421306 +0200
+++ libgomp/testsuite/libgomp.c-c++-common/lastprivate-conditional-9.c 2019-06-03 13:11:32.053675094 +0200
@@ -0,0 +1,60 @@
+/* { dg-do run } */
+
+int v, x;
+
+__attribute__((noipa)) void
+foo (int *a)
+{
+ #pragma omp for simd lastprivate (conditional: x) schedule(simd: static) if (simd: 0)
+ for (int i = 0; i < 128; i++)
+ if (a[i])
+ x = a[i];
+}
+
+__attribute__((noipa)) void
+bar (int *a, int *b)
+{
+ #pragma omp for simd lastprivate (conditional: x, v) schedule(static, 16) simdlen(1)
+ for (int i = 16; i < 128; ++i)
+ {
+ if (a[i])
+ x = a[i];
+ if (b[i])
+ v = b[i] + 10;
+ }
+}
+
+__attribute__((noipa)) void
+baz (int *a)
+{
+ #pragma omp for simd lastprivate (conditional: x) schedule(simd: dynamic, 16) if (0)
+ for (int i = 0; i < 128; i++)
+ if (a[i])
+ x = a[i] + 5;
+}
+
+int
+main ()
+{
+ int a[128], b[128], i;
+ for (i = 0; i < 128; i++)
+ {
+ a[i] = ((i % 11) == 2) ? i + 10 : 0;
+ asm volatile ("" : "+g" (i));
+ b[i] = ((i % 13) == 5) ? i * 2 : 0;
+ }
+ #pragma omp parallel
+ foo (a);
+ if (x != 133)
+ __builtin_abort ();
+ x = -3;
+ #pragma omp parallel
+ bar (b, a);
+ if (x != 244 || v != 143)
+ __builtin_abort ();
+ #pragma omp parallel
+ baz (b);
+ if (x != 249)
+ __builtin_abort ();
+ return 0;
+}
--- libgomp/testsuite/libgomp.c-c++-common/lastprivate-conditional-10.c.jj 2019-06-03 15:17:33.368060408 +0200
+++ libgomp/testsuite/libgomp.c-c++-common/lastprivate-conditional-10.c 2019-06-03 15:19:11.717504728 +0200
@@ -0,0 +1,57 @@
+/* { dg-do run } */
+
+int v, x;
+
+__attribute__((noipa)) int
+foo (int *a)
+{
+ #pragma omp parallel for simd lastprivate (conditional: x) schedule(simd: static) if (simd: 0)
+ for (int i = 0; i < 128; i++)
+ if (a[i])
+ x = a[i];
+ return x;
+}
+
+__attribute__((noipa)) int
+bar (int *a, int *b)
+{
+ #pragma omp parallel
+ #pragma omp for simd lastprivate (conditional: x, v) schedule(static, 16) simdlen (1)
+ for (int i = 16; i < 128; ++i)
+ {
+ if (a[i])
+ x = a[i];
+ if (b[i])
+ v = b[i] + 10;
+ }
+ return x;
+}
+
+__attribute__((noipa)) int
+baz (int *a)
+{
+ #pragma omp parallel for simd if (simd: 0) lastprivate (conditional: x) schedule(simd: dynamic, 16)
+ for (int i = 0; i < 128; i++)
+ if (a[i])
+ x = a[i] + 5;
+ return x;
+}
+
+int
+main ()
+{
+ int a[128], b[128], i;
+ for (i = 0; i < 128; i++)
+ {
+ a[i] = ((i % 11) == 2) ? i + 10 : 0;
+ asm volatile ("" : "+g" (i));
+ b[i] = ((i % 13) == 5) ? i * 2 : 0;
+ }
+ if (foo (a) != 133)
+ __builtin_abort ();
+ if (bar (b, a) != 244 || v != 143)
+ __builtin_abort ();
+ if (baz (b) != 249)
+ __builtin_abort ();
+ return 0;
+}
Jakub
More information about the Gcc-patches
mailing list