Message ID | 7c4bb857-a078-0c43-d853-6729cfa07221@codesourcery.com |
---|---|
State | New |
Headers | show |
Series | OpenMP: Disable GPU threads when only teams are used | expand |
On Thu, Jul 02, 2020 at 05:15:20PM +0100, Andrew Stubbs wrote: > This patch, originally by Kwok, auto-adjusts the default OpenMP target > arguments to set num_threads(1) when there are no parallel regions. There > may still be multiple teams in this case. > > The result is that libgomp will not attempt to launch GPU threads that will > never get used. > > OK to commit? That doesn't look safe to me. My understanding of the patch is that it looks for parallel construct lexically in the target region, but that isn't sufficient, one can do that only if the target region can't encounter a parallel construct in the target region (i.e. the body and all functions that are called from it at runtime). void foo () { #pragma omp distribute parallel for simd for (int i = 0; i < 10000000; i++) do_something; } extern void baz (); // function that calls foo, unconditionally or conditionally #pragma omp declare target to (foo, baz) void bar () { #pragma omp target teams baz (); } Perhaps one could ignore some builtin calls but it would need to be ones where one can assume there will be no OpenMP code in them. Also, it needs to avoid doing the optimization if there is or might indirectly be called omp_get_thread_limit (), because if the optimization forces thread_limit (1), that means that omp_get_thread_limit () in the region will also return 1 rather than the expected value. Jakub
On 02/07/2020 18:00, Jakub Jelinek wrote: > On Thu, Jul 02, 2020 at 05:15:20PM +0100, Andrew Stubbs wrote: >> This patch, originally by Kwok, auto-adjusts the default OpenMP target >> arguments to set num_threads(1) when there are no parallel regions. There >> may still be multiple teams in this case. >> >> The result is that libgomp will not attempt to launch GPU threads that will >> never get used. >> >> OK to commit? > > That doesn't look safe to me. > My understanding of the patch is that it looks for parallel construct > lexically in the target region, but that isn't sufficient, one can do that > only if the target region can't encounter a parallel construct in the target > region (i.e. the body and all functions that are called from it at runtime). OpenMP is complicated. :-( Is it normally expected that the runtime will always launch the maximum number of threads, just in case? There's a cost to both launching and running excess threads that it would be nice to avoid, but the real point of the optimization is that launching fewer threads allows us to launch more teams. AMD GPUs usually allow us to run 2040 or 2400 wavefronts simultaneously, so if we're running 15 unused threads for each team then we're limiting ourselves to 60 or 64 teams. If we limit each team to 1 thread then we can run the full 2040 or 2400 teams. Potentially, that's a 16x speed improvement on kernels that happen to not use parallel regions. I would like to be able to do this, but it appears that the region data is insufficient for complex cases. Can you suggest a good way to solve this? > Perhaps one could ignore some builtin calls but it would need to be ones > where one can assume there will be no OpenMP code in them. > > Also, it needs to avoid doing the optimization if there is or might > indirectly be called omp_get_thread_limit (), because if the optimization > forces thread_limit (1), that means that omp_get_thread_limit () in the > region will also return 1 rather than the expected value. Would that not be the correct answer, if the number of threads actually has been limited to 1? Thanks for the prompt review. Andrew
On Thu, Jul 02, 2020 at 10:16:25PM +0100, Andrew Stubbs wrote: > On 02/07/2020 18:00, Jakub Jelinek wrote: > > On Thu, Jul 02, 2020 at 05:15:20PM +0100, Andrew Stubbs wrote: > > > This patch, originally by Kwok, auto-adjusts the default OpenMP target > > > arguments to set num_threads(1) when there are no parallel regions. There > > > may still be multiple teams in this case. > > > > > > The result is that libgomp will not attempt to launch GPU threads that will > > > never get used. > > > > > > OK to commit? > > > > That doesn't look safe to me. > > My understanding of the patch is that it looks for parallel construct > > lexically in the target region, but that isn't sufficient, one can do that > > only if the target region can't encounter a parallel construct in the target > > region (i.e. the body and all functions that are called from it at runtime). > > OpenMP is complicated. :-( And it is and getting worse. > Is it normally expected that the runtime will always launch the maximum > number of threads, just in case? That is an implementation detail, the OpenMP model doesn't require that. The question is whether when encountering the parallel you can ask for more threads or not. E.g. on the host or in the host fallback, that is the case, we can just pthread_create as many threads as needed, for PTX there is the theoretical possibility to use dynamic parallelism, but I think it doesn't really work well and there were major problems with that. Anyway, I'd think OpenMP code that will only do teams and not parallel paralelism will be very rare in practice, it is true that in our testsuite we have probably a lot of tests for that but those are artificial tests. If somebody wants to get as much as possible from the hw, one should use all of teams, parallel and simd parallelism. If the user put an explicit thread_limit clause, I'd just trust the user what he is doing. If not, it is implementation defined what the maximum will be, but I'd say using a maximum of 1 if we don't find a parallel construct lexically nested is not a good default, even when it can be conforming. Because a reasonable application will have the parallel parallelism burried in one or more of the functions it calls, or if not, will use explicit thread_limit(1). If you want to perform some IPA analysis for this and tweak the default thread_limit based on what it (conservatively) finds out, I have nothing against that. Jakub
OpenMP: Disable GPU threads when only teams are used gcc/ * omp-expand.c (contains_threads): New. (get_target_arguments): Add region argument. Set number of threads to one if region does not contain threads. (expand_omp_target): Add extra argument in call to get_target_arguments. Co-Authored-By: Andrew Stubbs <ams@codesourcery.com> diff --git a/gcc/omp-expand.c b/gcc/omp-expand.c index 0f07e51f7e8..6afe18d5ee0 100644 --- a/gcc/omp-expand.c +++ b/gcc/omp-expand.c @@ -8461,10 +8461,22 @@ push_target_argument_according_to_value (gimple_stmt_iterator *gsi, int device, } } +static bool +contains_threads (struct omp_region *region) +{ + if (!region) + return false; + + return region->type == GIMPLE_OMP_PARALLEL + || contains_threads (region->inner) + || contains_threads (region->next); +} + /* Create an array of arguments that is then passed to GOMP_target. */ static tree -get_target_arguments (gimple_stmt_iterator *gsi, gomp_target *tgt_stmt) +get_target_arguments (gimple_stmt_iterator *gsi, gomp_target *tgt_stmt, + struct omp_region *region) { auto_vec <tree, 6> args; tree clauses = gimple_omp_target_clauses (tgt_stmt); @@ -8481,6 +8493,11 @@ get_target_arguments (gimple_stmt_iterator *gsi, gomp_target *tgt_stmt) t = OMP_CLAUSE_THREAD_LIMIT_EXPR (c); else t = integer_minus_one_node; + + if (tree_int_cst_equal (t, integer_zero_node) + && !contains_threads (region->inner)) + t = integer_one_node; + push_target_argument_according_to_value (gsi, GOMP_TARGET_ARG_DEVICE_ALL, GOMP_TARGET_ARG_THREAD_LIMIT, t, &args); @@ -8994,7 +9011,7 @@ expand_omp_target (struct omp_region *region) depend = build_int_cst (ptr_type_node, 0); args.quick_push (depend); if (start_ix == BUILT_IN_GOMP_TARGET) - args.quick_push (get_target_arguments (&gsi, entry_stmt)); + args.quick_push (get_target_arguments (&gsi, entry_stmt, region)); break; case BUILT_IN_GOACC_PARALLEL: if (lookup_attribute ("oacc serial", DECL_ATTRIBUTES (child_fn)) != NULL)