]> git.ipfire.org Git - thirdparty/gcc.git/blob - libgomp/parallel.c
Add prange entries in gimple-range-op.cc.
[thirdparty/gcc.git] / libgomp / parallel.c
1 /* Copyright (C) 2005-2020 Free Software Foundation, Inc.
2 Contributed by Richard Henderson <rth@redhat.com>.
3
4 This file is part of the GNU Offloading and Multi Processing Library
5 (libgomp).
6
7 Libgomp is free software; you can redistribute it and/or modify it
8 under the terms of the GNU General Public License as published by
9 the Free Software Foundation; either version 3, or (at your option)
10 any later version.
11
12 Libgomp is distributed in the hope that it will be useful, but WITHOUT ANY
13 WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS
14 FOR A PARTICULAR PURPOSE. See the GNU General Public License for
15 more details.
16
17 Under Section 7 of GPL version 3, you are granted additional
18 permissions described in the GCC Runtime Library Exception, version
19 3.1, as published by the Free Software Foundation.
20
21 You should have received a copy of the GNU General Public License and
22 a copy of the GCC Runtime Library Exception along with this program;
23 see the files COPYING3 and COPYING.RUNTIME respectively. If not, see
24 <http://www.gnu.org/licenses/>. */
25
26 /* This file handles the (bare) PARALLEL construct. */
27
28 #include "libgomp.h"
29 #include <limits.h>
30
31
32 /* Determine the number of threads to be launched for a PARALLEL construct.
33 This algorithm is explicitly described in OpenMP 3.0 section 2.4.1.
34 SPECIFIED is a combination of the NUM_THREADS clause and the IF clause.
35 If the IF clause is false, SPECIFIED is forced to 1. When NUM_THREADS
36 is not present, SPECIFIED is 0. */
37
38 unsigned
39 gomp_resolve_num_threads (unsigned specified, unsigned count)
40 {
41 struct gomp_thread *thr = gomp_thread ();
42 struct gomp_task_icv *icv;
43 unsigned threads_requested, max_num_threads, num_threads;
44 unsigned long busy;
45 struct gomp_thread_pool *pool;
46
47 icv = gomp_icv (false);
48
49 if (specified == 1)
50 return 1;
51 else if (thr->ts.active_level >= 1 && !icv->nest_var)
52 return 1;
53 else if (thr->ts.active_level >= gomp_max_active_levels_var)
54 return 1;
55
56 /* If NUM_THREADS not specified, use nthreads_var. */
57 if (specified == 0)
58 threads_requested = icv->nthreads_var;
59 else
60 threads_requested = specified;
61
62 max_num_threads = threads_requested;
63
64 /* If dynamic threads are enabled, bound the number of threads
65 that we launch. */
66 if (icv->dyn_var)
67 {
68 unsigned dyn = gomp_dynamic_max_threads ();
69 if (dyn < max_num_threads)
70 max_num_threads = dyn;
71
72 /* Optimization for parallel sections. */
73 if (count && count < max_num_threads)
74 max_num_threads = count;
75 }
76
77 /* UINT_MAX stands for infinity. */
78 if (__builtin_expect (icv->thread_limit_var == UINT_MAX, 1)
79 || max_num_threads == 1)
80 return max_num_threads;
81
82 /* The threads_busy counter lives in thread_pool, if there
83 isn't a thread_pool yet, there must be just one thread
84 in the contention group. If thr->team is NULL, this isn't
85 nested parallel, so there is just one thread in the
86 contention group as well, no need to handle it atomically. */
87 pool = thr->thread_pool;
88 if (thr->ts.team == NULL || pool == NULL)
89 {
90 num_threads = max_num_threads;
91 if (num_threads > icv->thread_limit_var)
92 num_threads = icv->thread_limit_var;
93 if (pool)
94 pool->threads_busy = num_threads;
95 return num_threads;
96 }
97
98 #ifdef HAVE_SYNC_BUILTINS
99 do
100 {
101 busy = pool->threads_busy;
102 num_threads = max_num_threads;
103 if (icv->thread_limit_var - busy + 1 < num_threads)
104 num_threads = icv->thread_limit_var - busy + 1;
105 }
106 while (__sync_val_compare_and_swap (&pool->threads_busy,
107 busy, busy + num_threads - 1)
108 != busy);
109 #else
110 gomp_mutex_lock (&gomp_managed_threads_lock);
111 num_threads = max_num_threads;
112 busy = pool->threads_busy;
113 if (icv->thread_limit_var - busy + 1 < num_threads)
114 num_threads = icv->thread_limit_var - busy + 1;
115 pool->threads_busy += num_threads - 1;
116 gomp_mutex_unlock (&gomp_managed_threads_lock);
117 #endif
118
119 return num_threads;
120 }
121
122 void
123 GOMP_parallel_start (void (*fn) (void *), void *data, unsigned num_threads)
124 {
125 num_threads = gomp_resolve_num_threads (num_threads, 0);
126 gomp_team_start (fn, data, num_threads, 0, gomp_new_team (num_threads),
127 NULL);
128 }
129
130 void
131 GOMP_parallel_end (void)
132 {
133 struct gomp_task_icv *icv = gomp_icv (false);
134 if (__builtin_expect (icv->thread_limit_var != UINT_MAX, 0))
135 {
136 struct gomp_thread *thr = gomp_thread ();
137 struct gomp_team *team = thr->ts.team;
138 unsigned int nthreads = team ? team->nthreads : 1;
139 gomp_team_end ();
140 if (nthreads > 1)
141 {
142 /* If not nested, there is just one thread in the
143 contention group left, no need for atomicity. */
144 if (thr->ts.team == NULL)
145 thr->thread_pool->threads_busy = 1;
146 else
147 {
148 #ifdef HAVE_SYNC_BUILTINS
149 __sync_fetch_and_add (&thr->thread_pool->threads_busy,
150 1UL - nthreads);
151 #else
152 gomp_mutex_lock (&gomp_managed_threads_lock);
153 thr->thread_pool->threads_busy -= nthreads - 1;
154 gomp_mutex_unlock (&gomp_managed_threads_lock);
155 #endif
156 }
157 }
158 }
159 else
160 gomp_team_end ();
161 }
162 ialias (GOMP_parallel_end)
163
164 void
165 GOMP_parallel (void (*fn) (void *), void *data, unsigned num_threads,
166 unsigned int flags)
167 {
168 num_threads = gomp_resolve_num_threads (num_threads, 0);
169 gomp_team_start (fn, data, num_threads, flags, gomp_new_team (num_threads),
170 NULL);
171 fn (data);
172 ialias_call (GOMP_parallel_end) ();
173 }
174
175 unsigned
176 GOMP_parallel_reductions (void (*fn) (void *), void *data,
177 unsigned num_threads, unsigned int flags)
178 {
179 struct gomp_taskgroup *taskgroup;
180 num_threads = gomp_resolve_num_threads (num_threads, 0);
181 uintptr_t *rdata = *(uintptr_t **)data;
182 taskgroup = gomp_parallel_reduction_register (rdata, num_threads);
183 gomp_team_start (fn, data, num_threads, flags, gomp_new_team (num_threads),
184 taskgroup);
185 fn (data);
186 ialias_call (GOMP_parallel_end) ();
187 gomp_sem_destroy (&taskgroup->taskgroup_sem);
188 free (taskgroup);
189 return num_threads;
190 }
191
192 bool
193 GOMP_cancellation_point (int which)
194 {
195 if (!gomp_cancel_var)
196 return false;
197
198 struct gomp_thread *thr = gomp_thread ();
199 struct gomp_team *team = thr->ts.team;
200 if (which & (GOMP_CANCEL_LOOP | GOMP_CANCEL_SECTIONS))
201 {
202 if (team == NULL)
203 return false;
204 return team->work_share_cancelled != 0;
205 }
206 else if (which & GOMP_CANCEL_TASKGROUP)
207 {
208 if (thr->task->taskgroup)
209 {
210 if (thr->task->taskgroup->cancelled)
211 return true;
212 if (thr->task->taskgroup->workshare
213 && thr->task->taskgroup->prev
214 && thr->task->taskgroup->prev->cancelled)
215 return true;
216 }
217 /* FALLTHRU into the GOMP_CANCEL_PARALLEL case,
218 as #pragma omp cancel parallel also cancels all explicit
219 tasks. */
220 }
221 if (team)
222 return gomp_team_barrier_cancelled (&team->barrier);
223 return false;
224 }
225 ialias (GOMP_cancellation_point)
226
227 bool
228 GOMP_cancel (int which, bool do_cancel)
229 {
230 if (!gomp_cancel_var)
231 return false;
232
233 if (!do_cancel)
234 return ialias_call (GOMP_cancellation_point) (which);
235
236 struct gomp_thread *thr = gomp_thread ();
237 struct gomp_team *team = thr->ts.team;
238 if (which & (GOMP_CANCEL_LOOP | GOMP_CANCEL_SECTIONS))
239 {
240 /* In orphaned worksharing region, all we want to cancel
241 is current thread. */
242 if (team != NULL)
243 team->work_share_cancelled = 1;
244 return true;
245 }
246 else if (which & GOMP_CANCEL_TASKGROUP)
247 {
248 if (thr->task->taskgroup)
249 {
250 struct gomp_taskgroup *taskgroup = thr->task->taskgroup;
251 if (taskgroup->workshare && taskgroup->prev)
252 taskgroup = taskgroup->prev;
253 if (!taskgroup->cancelled)
254 {
255 gomp_mutex_lock (&team->task_lock);
256 taskgroup->cancelled = true;
257 gomp_mutex_unlock (&team->task_lock);
258 }
259 }
260 return true;
261 }
262 team->team_cancelled = 1;
263 gomp_team_barrier_cancel (team);
264 return true;
265 }
266 \f
267 /* The public OpenMP API for thread and team related inquiries. */
268
269 int
270 omp_get_num_threads (void)
271 {
272 struct gomp_team *team = gomp_thread ()->ts.team;
273 return team ? team->nthreads : 1;
274 }
275
276 int
277 omp_get_thread_num (void)
278 {
279 return gomp_thread ()->ts.team_id;
280 }
281
282 /* This wasn't right for OpenMP 2.5. Active region used to be non-zero
283 when the IF clause doesn't evaluate to false, starting with OpenMP 3.0
284 it is non-zero with more than one thread in the team. */
285
286 int
287 omp_in_parallel (void)
288 {
289 return gomp_thread ()->ts.active_level > 0;
290 }
291
292 int
293 omp_get_level (void)
294 {
295 return gomp_thread ()->ts.level;
296 }
297
298 int
299 omp_get_ancestor_thread_num (int level)
300 {
301 struct gomp_team_state *ts = &gomp_thread ()->ts;
302 if (level < 0 || level > ts->level)
303 return -1;
304 for (level = ts->level - level; level > 0; --level)
305 ts = &ts->team->prev_ts;
306 return ts->team_id;
307 }
308
309 int
310 omp_get_team_size (int level)
311 {
312 struct gomp_team_state *ts = &gomp_thread ()->ts;
313 if (level < 0 || level > ts->level)
314 return -1;
315 for (level = ts->level - level; level > 0; --level)
316 ts = &ts->team->prev_ts;
317 if (ts->team == NULL)
318 return 1;
319 else
320 return ts->team->nthreads;
321 }
322
323 int
324 omp_get_active_level (void)
325 {
326 return gomp_thread ()->ts.active_level;
327 }
328
329 ialias (omp_get_num_threads)
330 ialias (omp_get_thread_num)
331 ialias (omp_in_parallel)
332 ialias (omp_get_level)
333 ialias (omp_get_ancestor_thread_num)
334 ialias (omp_get_team_size)
335 ialias (omp_get_active_level)