]> git.ipfire.org Git - thirdparty/gcc.git/blob - libcilkrts/runtime/scheduler.h
[Patch AArch64] Fixup floating point division with -march=armv8-a+nosimd
[thirdparty/gcc.git] / libcilkrts / runtime / scheduler.h
1 /* scheduler.h -*-C++-*-
2 *
3 *************************************************************************
4 *
5 * Copyright (C) 2009-2016, Intel Corporation
6 * All rights reserved.
7 *
8 * Redistribution and use in source and binary forms, with or without
9 * modification, are permitted provided that the following conditions
10 * are met:
11 *
12 * * Redistributions of source code must retain the above copyright
13 * notice, this list of conditions and the following disclaimer.
14 * * Redistributions in binary form must reproduce the above copyright
15 * notice, this list of conditions and the following disclaimer in
16 * the documentation and/or other materials provided with the
17 * distribution.
18 * * Neither the name of Intel Corporation nor the names of its
19 * contributors may be used to endorse or promote products derived
20 * from this software without specific prior written permission.
21 *
22 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
23 * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
24 * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
25 * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
26 * HOLDER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT,
27 * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING,
28 * BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS
29 * OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED
30 * AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
31 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY
32 * WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
33 * POSSIBILITY OF SUCH DAMAGE.
34 *
35 * *********************************************************************
36 *
37 * PLEASE NOTE: This file is a downstream copy of a file mainitained in
38 * a repository at cilkplus.org. Changes made to this file that are not
39 * submitted through the contribution process detailed at
40 * http://www.cilkplus.org/submit-cilk-contribution will be lost the next
41 * time that a new version is released. Changes only submitted to the
42 * GNU compiler collection or posted to the git repository at
43 * https://bitbucket.org/intelcilkruntime/intel-cilk-runtime.git are
44 * not tracked.
45 *
46 * We welcome your contributions to this open source project. Thank you
47 * for your assistance in helping us improve Cilk Plus.
48 **************************************************************************/
49
50 /**
51 * @file scheduler.h
52 *
53 * @brief scheduler.h declares routines for the Intel Cilk Plus scheduler,
54 * making it the heart of the Intel Cilk Plus implementation.
55 */
56
57 #ifndef INCLUDED_SCHEDULER_DOT_H
58 #define INCLUDED_SCHEDULER_DOT_H
59
60 #include <cilk/common.h>
61 #include <internal/abi.h>
62
63 #include "rts-common.h"
64 #include "full_frame.h"
65 #include "reducer_impl.h"
66 #include "global_state.h"
67
68 #ifdef CILK_RECORD_REPLAY
69 #include "record-replay.h"
70 #endif
71
72 __CILKRTS_BEGIN_EXTERN_C
73
74
75 /**
76 * @brief Flag to disable parallel reductions.
77 *
78 * Set to 0 to allow parallel reductions.
79 */
80 #define DISABLE_PARALLEL_REDUCERS 0
81
82 /**
83 * @brief Debugging level for parallel reductions.
84 *
85 * Print debugging messages and assertions for parallel reducers. 0 is
86 * no debugging. A higher value generates more output.
87 */
88
89 #ifndef REDPAR_DEBUG
90 #define REDPAR_DEBUG 0
91 #endif
92
93 /**
94 * @brief Lock the worker mutex to allow exclusive access to the
95 * values in the @c __cilkrts_worker and local_state structures.
96 *
97 * @pre @c w->l->do_not_steal must not be set. Essentially this
98 * condition asserts that the worker is not locked recursively.
99 *
100 * @param w The worker to lock.
101 */
102 COMMON_PORTABLE
103 void __cilkrts_worker_lock(__cilkrts_worker *w);
104
105 /**
106 * @brief Unlock the worker mutex.
107 *
108 * @pre @c w->l->do_not_steal must be set. Essentially this condition
109 * asserts that the worker has been previously locked.
110 *
111 * @param w The worker to unlock.
112 */
113 COMMON_PORTABLE
114 void __cilkrts_worker_unlock(__cilkrts_worker *w);
115
116 /**
117 * @brief Push the next full frame to be made active in this worker
118 * and increment its join counter.
119 *
120 * __cilkrts_push_next_frame and pop_next_frame work on a one-element queue.
121 * This queue is used to communicate across the runtime from the code that
122 * wants to activate a frame to the code that can actually begin execution
123 * on that frame. They are asymetrical in that push increments the join
124 * counter but pop does not decrement it. Rather, a single push/pop
125 * combination makes a frame active and increments its join counter once.
126 *
127 * @note A system worker may chose to push work onto a user worker if
128 * the work is the continuation from a sync which only the user worker
129 * may complete.
130 *
131 * @param w The worker which the frame is to be pushed onto.
132 * @param ff The full_frame which is to be continued by the worker.
133 */
134 COMMON_PORTABLE
135 void __cilkrts_push_next_frame(__cilkrts_worker *w,
136 full_frame *ff);
137
138 /**
139 * @brief Sync on this worker.
140 *
141 * If this worker is the last to reach the sync, execution may resume
142 * on this worker after the sync.
143 *
144 * If this worker is not the last spawned child to reach the sync,
145 * then execution is suspended and the worker will re-enter the
146 * scheduling loop, looking for work it can steal.
147 *
148 * This function will jump into the runtime to switch to the scheduling
149 * stack to implement most of its logic.
150 *
151 * @param w The worker which is executing the sync.
152 * @param sf The __cilkrts_stack_frame containing the sync.
153 */
154 COMMON_PORTABLE
155 NORETURN __cilkrts_c_sync(__cilkrts_worker *w,
156 __cilkrts_stack_frame *sf);
157
158 /**
159 * @brief Worker @c w completely promotes its own deque, simulating the case
160 * where the whole deque is stolen.
161 *
162 * We use this mechanism to force the allocation of new storage for
163 * reducers for race-detection purposes.
164 *
165 * This method is called from the reducer lookup logic when
166 * @c g->force_reduce is set.
167 *
168 * @warning Use of "force_reduce" is known to have bugs when run with
169 * more than 1 worker.
170 *
171 * @param w The worker which is to have all entries in its deque
172 * promoted to full frames.
173 */
174 COMMON_PORTABLE
175 void __cilkrts_promote_own_deque(__cilkrts_worker *w);
176
177 /**
178 * Called when a spawned function attempts to return and
179 * __cilkrts_undo_detach() fails. This can happen for two reasons:
180 *
181 * @li If another worker is considering stealing our parent, it bumps the
182 * exception pointer while it did so, which will cause __cilkrts_undo_detach()
183 * to fail. If the other worker didn't complete the steal of our parent, we
184 * still may be able to return to it, either because the steal attempt failed,
185 * or we won the race for the tail pointer.
186 *
187 * @li If the function's parent has been stolen then we cannot return. Instead
188 * we'll longjmp into the runtime to switch onto the scheduling stack to
189 * execute do_return_from_spawn() and determine what to do. Either this
190 * worker is the last one to the sync, in which case we need to jump to the
191 * sync, or this worker is not the last one to the sync, in which case we'll
192 * abandon this work and jump to the scheduling loop to search for more work
193 * we can steal.
194 *
195 * @param w The worker which attempting to return from a spawn to
196 * a stolen parent.
197 * @param returning_sf The stack frame which is returning.
198 */
199 COMMON_PORTABLE
200 void __cilkrts_c_THE_exception_check(__cilkrts_worker *w,
201 __cilkrts_stack_frame *returning_sf);
202
203 /**
204 * @brief Return an exception to an stolen parent.
205 *
206 * Used by the gcc implementation of exceptions to return an exception
207 * to a stolen parent
208 *
209 * @param w The worker which attempting to return from a spawn with an
210 * exception to a stolen parent.
211 * @param returning_sf The stack frame which is returning.
212 */
213 COMMON_PORTABLE
214 NORETURN __cilkrts_exception_from_spawn(__cilkrts_worker *w,
215 __cilkrts_stack_frame *returning_sf);
216
217 /**
218 * @brief Used by the Windows implementations of exceptions to migrate an exception
219 * across fibers.
220 *
221 * Call this function when an exception has been thrown and has to
222 * traverse across a steal. The exception has already been wrapped
223 * up, so all that remains is to longjmp() into the continuation,
224 * sync, and re-raise it.
225 *
226 * @param sf The __cilkrts_stack_frame for the frame that is attempting to
227 * return an exception to a stolen parent.
228 */
229 void __cilkrts_migrate_exception (__cilkrts_stack_frame *sf);
230
231 /**
232 * @brief Return from a call, not a spawn, where this frame has ever
233 * been stolen.
234 *
235 * @param w The worker that is returning from a frame which was ever stolen.
236 */
237 COMMON_PORTABLE
238 void __cilkrts_return(__cilkrts_worker *w);
239
240 /**
241 * @brief Special return from the initial frame.
242 *
243 * This method will be called from @c __cilkrts_leave_frame if
244 * @c CILK_FRAME_LAST is set.
245 *
246 * This function will do the things necessary to cleanup, and unbind the
247 * thread from the Intel Cilk Plus runtime. If this is the last user
248 * worker unbinding from the runtime, all system worker threads will be
249 * suspended.
250 *
251 * @pre @c w must be the currently executing worker, and must be a user
252 * worker.
253 *
254 * @param w The worker that's returning from the initial frame.
255 */
256 COMMON_PORTABLE
257 void __cilkrts_c_return_from_initial(__cilkrts_worker *w);
258
259 /**
260 * @brief Used by exception handling code to pop an entry from the
261 * worker's deque.
262 *
263 * @param w Worker to pop the entry from
264 *
265 * @return __cilkrts_stack_frame of parent call
266 * @return NULL if the deque is empty
267 */
268 COMMON_PORTABLE
269 __cilkrts_stack_frame *__cilkrts_pop_tail(__cilkrts_worker *w);
270
271 /**
272 * @brief Modifies the worker's protected_tail to prevent frames from
273 * being stolen.
274 *
275 * The Dekker protocol has been extended to only steal if head+1 is also
276 * less than protected_tail.
277 *
278 * @param w The worker to be modified.
279 * @param new_protected_tail The new setting for protected_tail, or NULL if the
280 * entire deque is to be protected
281 *
282 * @return Previous value of protected tail.
283 */
284 COMMON_PORTABLE
285 __cilkrts_stack_frame *volatile *__cilkrts_disallow_stealing(
286 __cilkrts_worker *w,
287 __cilkrts_stack_frame *volatile *new_protected_tail);
288
289 /**
290 * @brief Restores the protected tail to a previous state, possibly
291 * allowing frames to be stolen.
292 *
293 * @param w The worker to be modified.
294 * @param saved_protected_tail A previous setting for protected_tail that is
295 * to be restored
296 */
297 COMMON_PORTABLE
298 void __cilkrts_restore_stealing(
299 __cilkrts_worker *w,
300 __cilkrts_stack_frame *volatile *saved_protected_tail);
301
302 /**
303 * @brief Initialize a @c __cilkrts_worker.
304 *
305 * @note The memory for the worker must have been allocated outside
306 * this call.
307 *
308 * @param g The global_state_t.
309 * @param self The index into the global_state's array of workers for this
310 * worker, or -1 if this worker was allocated from the heap and cannot be
311 * stolen from.
312 * @param w The worker to be initialized.
313 *
314 * @return The initialized __cilkrts_worker.
315 */
316 COMMON_PORTABLE
317 __cilkrts_worker *make_worker(global_state_t *g,
318 int self,
319 __cilkrts_worker *w);
320
321 /**
322 * @brief Free up any resources allocated for a worker.
323 *
324 * @note The memory for the @c __cilkrts_worker itself must be
325 * deallocated outside this call.
326 *
327 * @param w The worker to be destroyed.
328 */
329 COMMON_PORTABLE
330 void destroy_worker (__cilkrts_worker *w);
331
332 /**
333 * @brief Initialize the runtime.
334 *
335 * If necessary, allocates and initializes the global state. If
336 * necessary, unsuspends the system workers.
337 *
338 * @param start Specifies whether the workers are to be unsuspended if
339 * they are suspended. Allows __cilkrts_init() to start up the runtime without
340 * releasing the system threads.
341 */
342 COMMON_PORTABLE
343 void __cilkrts_init_internal(int start);
344
345 /**
346 * @brief Part of the sequence to shutdown the runtime.
347 *
348 * Specifically, this call frees the @c global_state_t for the runtime.
349 *
350 * @param g The global_state_t.
351 */
352 COMMON_PORTABLE
353 void __cilkrts_deinit_internal(global_state_t *g);
354
355 /**
356 * Obsolete. We no longer need to import or export reducer maps.
357 */
358 COMMON_PORTABLE
359 cilkred_map *__cilkrts_xchg_reducer(
360 __cilkrts_worker *w, cilkred_map *newmap) cilk_nothrow;
361
362 /**
363 * @brief Called when a user thread is bound to the runtime.
364 *
365 * If this action increments the count of bound user threads from 0 to
366 * 1, the system worker threads are unsuspended.
367 *
368 * If this action increments the count of bound user threads from 0 to
369 * 1, the system worker threads are unsuspended.
370 *
371 * @pre Global lock must be held.
372 * @param g The runtime global state.
373 */
374 COMMON_PORTABLE
375 void __cilkrts_enter_cilk(global_state_t *g);
376
377 /**
378 * @brief Called when a user thread is unbound from the runtime.
379 *
380 * If this action decrements the count of bound user threads to 0, the
381 * system worker threads are suspended.
382 *
383 *
384 * @pre Global lock must be held.
385 *
386 * @param g The runtime global state.
387 */
388 COMMON_PORTABLE
389 void __cilkrts_leave_cilk(global_state_t *g);
390
391
392 /**
393 * @brief cilk_fiber_proc that runs the main scheduler loop on a
394 * user worker.
395 *
396 * @pre fiber's owner field should be set to the correct __cilkrts_worker
397 * @pre fiber must be a user worker.
398 *
399 * @param fiber The scheduling fiber object.
400 */
401 void scheduler_fiber_proc_for_user_worker(cilk_fiber *fiber);
402
403
404 /**
405 * @brief Prints out Cilk runtime statistics.
406 *
407 * @param g The runtime global state.
408 *
409 * This method is useful only for debugging purposes. No guarantees
410 * are made as to the validity of this data. :)
411 */
412 COMMON_PORTABLE
413 void __cilkrts_dump_stats_to_stderr(global_state_t *g);
414
415 #ifdef CILK_RECORD_REPLAY
416 COMMON_PORTABLE
417 char * walk_pedigree_nodes(char *p, const __cilkrts_pedigree *pnode);
418
419 /**
420 * @brief Used by exception handling code to simulate the popping of
421 * an entry from the worker's deque.
422 *
423 * @param w Worker whose deque we want to check
424 *
425 * @return @c __cilkrts_stack_frame of parent call
426 * @return NULL if the deque is empty
427 */
428 COMMON_PORTABLE
429 __cilkrts_stack_frame *simulate_pop_tail(__cilkrts_worker *w);
430
431 #endif
432
433 __CILKRTS_END_EXTERN_C
434
435 #endif // ! defined(INCLUDED_SCHEDULER_DOT_H)