]> git.ipfire.org Git - thirdparty/gcc.git/blob - libstdc++-v3/include/bits/atomic_wait.h
libstdc++: Minor codegen improvement for atomic wait spinloop
[thirdparty/gcc.git] / libstdc++-v3 / include / bits / atomic_wait.h
1 // -*- C++ -*- header.
2
3 // Copyright (C) 2020-2022 Free Software Foundation, Inc.
4 //
5 // This file is part of the GNU ISO C++ Library. This library is free
6 // software; you can redistribute it and/or modify it under the
7 // terms of the GNU General Public License as published by the
8 // Free Software Foundation; either version 3, or (at your option)
9 // any later version.
10
11 // This library is distributed in the hope that it will be useful,
12 // but WITHOUT ANY WARRANTY; without even the implied warranty of
13 // MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
14 // GNU General Public License for more details.
15
16 // Under Section 7 of GPL version 3, you are granted additional
17 // permissions described in the GCC Runtime Library Exception, version
18 // 3.1, as published by the Free Software Foundation.
19
20 // You should have received a copy of the GNU General Public License and
21 // a copy of the GCC Runtime Library Exception along with this program;
22 // see the files COPYING3 and COPYING.RUNTIME respectively. If not, see
23 // <http://www.gnu.org/licenses/>.
24
25 /** @file bits/atomic_wait.h
26 * This is an internal header file, included by other library headers.
27 * Do not attempt to use it directly. @headername{atomic}
28 */
29
30 #ifndef _GLIBCXX_ATOMIC_WAIT_H
31 #define _GLIBCXX_ATOMIC_WAIT_H 1
32
33 #pragma GCC system_header
34
35 #include <bits/c++config.h>
36 #if defined _GLIBCXX_HAS_GTHREADS || defined _GLIBCXX_HAVE_LINUX_FUTEX
37 #include <bits/functional_hash.h>
38 #include <bits/gthr.h>
39 #include <ext/numeric_traits.h>
40
41 #ifdef _GLIBCXX_HAVE_LINUX_FUTEX
42 # include <cerrno>
43 # include <climits>
44 # include <unistd.h>
45 # include <syscall.h>
46 # include <bits/functexcept.h>
47 #endif
48
49 # include <bits/std_mutex.h> // std::mutex, std::__condvar
50
51 #define __cpp_lib_atomic_wait 201907L
52
53 namespace std _GLIBCXX_VISIBILITY(default)
54 {
55 _GLIBCXX_BEGIN_NAMESPACE_VERSION
56 namespace __detail
57 {
58 #ifdef _GLIBCXX_HAVE_LINUX_FUTEX
59 #define _GLIBCXX_HAVE_PLATFORM_WAIT 1
60 using __platform_wait_t = int;
61 static constexpr size_t __platform_wait_alignment = 4;
62 #else
63 // define _GLIBCX_HAVE_PLATFORM_WAIT and implement __platform_wait()
64 // and __platform_notify() if there is a more efficient primitive supported
65 // by the platform (e.g. __ulock_wait()/__ulock_wake()) which is better than
66 // a mutex/condvar based wait.
67 using __platform_wait_t = uint64_t;
68 static constexpr size_t __platform_wait_alignment
69 = __alignof__(__platform_wait_t);
70 #endif
71 } // namespace __detail
72
73 template<typename _Tp>
74 inline constexpr bool __platform_wait_uses_type
75 #ifdef _GLIBCXX_HAVE_PLATFORM_WAIT
76 = is_scalar_v<_Tp>
77 && ((sizeof(_Tp) == sizeof(__detail::__platform_wait_t))
78 && (alignof(_Tp*) >= __detail::__platform_wait_alignment));
79 #else
80 = false;
81 #endif
82
83 namespace __detail
84 {
85 #ifdef _GLIBCXX_HAVE_LINUX_FUTEX
86 enum class __futex_wait_flags : int
87 {
88 #ifdef _GLIBCXX_HAVE_LINUX_FUTEX_PRIVATE
89 __private_flag = 128,
90 #else
91 __private_flag = 0,
92 #endif
93 __wait = 0,
94 __wake = 1,
95 __wait_bitset = 9,
96 __wake_bitset = 10,
97 __wait_private = __wait | __private_flag,
98 __wake_private = __wake | __private_flag,
99 __wait_bitset_private = __wait_bitset | __private_flag,
100 __wake_bitset_private = __wake_bitset | __private_flag,
101 __bitset_match_any = -1
102 };
103
104 template<typename _Tp>
105 void
106 __platform_wait(const _Tp* __addr, __platform_wait_t __val) noexcept
107 {
108 auto __e = syscall (SYS_futex, static_cast<const void*>(__addr),
109 static_cast<int>(__futex_wait_flags::__wait_private),
110 __val, nullptr);
111 if (!__e || errno == EAGAIN)
112 return;
113 if (errno != EINTR)
114 __throw_system_error(errno);
115 }
116
117 template<typename _Tp>
118 void
119 __platform_notify(const _Tp* __addr, bool __all) noexcept
120 {
121 syscall (SYS_futex, static_cast<const void*>(__addr),
122 static_cast<int>(__futex_wait_flags::__wake_private),
123 __all ? INT_MAX : 1);
124 }
125 #endif
126
127 inline void
128 __thread_yield() noexcept
129 {
130 #if defined _GLIBCXX_HAS_GTHREADS && defined _GLIBCXX_USE_SCHED_YIELD
131 __gthread_yield();
132 #endif
133 }
134
135 inline void
136 __thread_relax() noexcept
137 {
138 #if defined __i386__ || defined __x86_64__
139 __builtin_ia32_pause();
140 #else
141 __thread_yield();
142 #endif
143 }
144
145 constexpr auto __atomic_spin_count_relax = 12;
146 constexpr auto __atomic_spin_count = 16;
147
148 struct __default_spin_policy
149 {
150 bool
151 operator()() const noexcept
152 { return false; }
153 };
154
155 template<typename _Pred,
156 typename _Spin = __default_spin_policy>
157 bool
158 __atomic_spin(_Pred& __pred, _Spin __spin = _Spin{ }) noexcept
159 {
160 for (auto __i = 0; __i < __atomic_spin_count; ++__i)
161 {
162 if (__pred())
163 return true;
164
165 if (__i < __atomic_spin_count_relax)
166 __detail::__thread_relax();
167 else
168 __detail::__thread_yield();
169 }
170
171 while (__spin())
172 {
173 if (__pred())
174 return true;
175 }
176
177 return false;
178 }
179
180 // return true if equal
181 template<typename _Tp>
182 bool __atomic_compare(const _Tp& __a, const _Tp& __b)
183 {
184 // TODO make this do the correct padding bit ignoring comparison
185 return __builtin_memcmp(&__a, &__b, sizeof(_Tp)) == 0;
186 }
187
188 struct __waiter_pool_base
189 {
190 // Don't use std::hardware_destructive_interference_size here because we
191 // don't want the layout of library types to depend on compiler options.
192 static constexpr auto _S_align = 64;
193
194 alignas(_S_align) __platform_wait_t _M_wait = 0;
195
196 #ifndef _GLIBCXX_HAVE_PLATFORM_WAIT
197 mutex _M_mtx;
198 #endif
199
200 alignas(_S_align) __platform_wait_t _M_ver = 0;
201
202 #ifndef _GLIBCXX_HAVE_PLATFORM_WAIT
203 __condvar _M_cv;
204 #endif
205 __waiter_pool_base() = default;
206
207 void
208 _M_enter_wait() noexcept
209 { __atomic_fetch_add(&_M_wait, 1, __ATOMIC_SEQ_CST); }
210
211 void
212 _M_leave_wait() noexcept
213 { __atomic_fetch_sub(&_M_wait, 1, __ATOMIC_RELEASE); }
214
215 bool
216 _M_waiting() const noexcept
217 {
218 __platform_wait_t __res;
219 __atomic_load(&_M_wait, &__res, __ATOMIC_SEQ_CST);
220 return __res != 0;
221 }
222
223 void
224 _M_notify(const __platform_wait_t* __addr, bool __all, bool __bare) noexcept
225 {
226 if (!(__bare || _M_waiting()))
227 return;
228
229 #ifdef _GLIBCXX_HAVE_PLATFORM_WAIT
230 __platform_notify(__addr, __all);
231 #else
232 if (__all)
233 _M_cv.notify_all();
234 else
235 _M_cv.notify_one();
236 #endif
237 }
238
239 static __waiter_pool_base&
240 _S_for(const void* __addr) noexcept
241 {
242 constexpr uintptr_t __ct = 16;
243 static __waiter_pool_base __w[__ct];
244 auto __key = (uintptr_t(__addr) >> 2) % __ct;
245 return __w[__key];
246 }
247 };
248
249 struct __waiter_pool : __waiter_pool_base
250 {
251 void
252 _M_do_wait(const __platform_wait_t* __addr, __platform_wait_t __old) noexcept
253 {
254 #ifdef _GLIBCXX_HAVE_PLATFORM_WAIT
255 __platform_wait(__addr, __old);
256 #else
257 __platform_wait_t __val;
258 __atomic_load(__addr, &__val, __ATOMIC_SEQ_CST);
259 if (__val == __old)
260 {
261 lock_guard<mutex> __l(_M_mtx);
262 _M_cv.wait(_M_mtx);
263 }
264 #endif // __GLIBCXX_HAVE_PLATFORM_WAIT
265 }
266 };
267
268 template<typename _Tp>
269 struct __waiter_base
270 {
271 using __waiter_type = _Tp;
272
273 __waiter_type& _M_w;
274 __platform_wait_t* _M_addr;
275
276 template<typename _Up>
277 static __platform_wait_t*
278 _S_wait_addr(const _Up* __a, __platform_wait_t* __b)
279 {
280 if constexpr (__platform_wait_uses_type<_Up>)
281 return reinterpret_cast<__platform_wait_t*>(const_cast<_Up*>(__a));
282 else
283 return __b;
284 }
285
286 static __waiter_type&
287 _S_for(const void* __addr) noexcept
288 {
289 static_assert(sizeof(__waiter_type) == sizeof(__waiter_pool_base));
290 auto& res = __waiter_pool_base::_S_for(__addr);
291 return reinterpret_cast<__waiter_type&>(res);
292 }
293
294 template<typename _Up>
295 explicit __waiter_base(const _Up* __addr) noexcept
296 : _M_w(_S_for(__addr))
297 , _M_addr(_S_wait_addr(__addr, &_M_w._M_ver))
298 { }
299
300 bool
301 _M_laundered() const
302 { return _M_addr == &_M_w._M_ver; }
303
304 void
305 _M_notify(bool __all, bool __bare = false)
306 {
307 if (_M_laundered())
308 {
309 __atomic_fetch_add(_M_addr, 1, __ATOMIC_SEQ_CST);
310 __all = true;
311 }
312 _M_w._M_notify(_M_addr, __all, __bare);
313 }
314
315 template<typename _Up, typename _ValFn,
316 typename _Spin = __default_spin_policy>
317 static bool
318 _S_do_spin_v(__platform_wait_t* __addr,
319 const _Up& __old, _ValFn __vfn,
320 __platform_wait_t& __val,
321 _Spin __spin = _Spin{ })
322 {
323 auto const __pred = [=]
324 { return !__detail::__atomic_compare(__old, __vfn()); };
325
326 if constexpr (__platform_wait_uses_type<_Up>)
327 {
328 __builtin_memcpy(&__val, &__old, sizeof(__val));
329 }
330 else
331 {
332 __atomic_load(__addr, &__val, __ATOMIC_ACQUIRE);
333 }
334 return __atomic_spin(__pred, __spin);
335 }
336
337 template<typename _Up, typename _ValFn,
338 typename _Spin = __default_spin_policy>
339 bool
340 _M_do_spin_v(const _Up& __old, _ValFn __vfn,
341 __platform_wait_t& __val,
342 _Spin __spin = _Spin{ })
343 { return _S_do_spin_v(_M_addr, __old, __vfn, __val, __spin); }
344
345 template<typename _Pred,
346 typename _Spin = __default_spin_policy>
347 static bool
348 _S_do_spin(const __platform_wait_t* __addr,
349 _Pred __pred,
350 __platform_wait_t& __val,
351 _Spin __spin = _Spin{ })
352 {
353 __atomic_load(__addr, &__val, __ATOMIC_ACQUIRE);
354 return __atomic_spin(__pred, __spin);
355 }
356
357 template<typename _Pred,
358 typename _Spin = __default_spin_policy>
359 bool
360 _M_do_spin(_Pred __pred, __platform_wait_t& __val,
361 _Spin __spin = _Spin{ })
362 { return _S_do_spin(_M_addr, __pred, __val, __spin); }
363 };
364
365 template<typename _EntersWait>
366 struct __waiter : __waiter_base<__waiter_pool>
367 {
368 using __base_type = __waiter_base<__waiter_pool>;
369
370 template<typename _Tp>
371 explicit __waiter(const _Tp* __addr) noexcept
372 : __base_type(__addr)
373 {
374 if constexpr (_EntersWait::value)
375 _M_w._M_enter_wait();
376 }
377
378 ~__waiter()
379 {
380 if constexpr (_EntersWait::value)
381 _M_w._M_leave_wait();
382 }
383
384 template<typename _Tp, typename _ValFn>
385 void
386 _M_do_wait_v(_Tp __old, _ValFn __vfn)
387 {
388 do
389 {
390 __platform_wait_t __val;
391 if (__base_type::_M_do_spin_v(__old, __vfn, __val))
392 return;
393 __base_type::_M_w._M_do_wait(__base_type::_M_addr, __val);
394 }
395 while (__detail::__atomic_compare(__old, __vfn()));
396 }
397
398 template<typename _Pred>
399 void
400 _M_do_wait(_Pred __pred) noexcept
401 {
402 do
403 {
404 __platform_wait_t __val;
405 if (__base_type::_M_do_spin(__pred, __val))
406 return;
407 __base_type::_M_w._M_do_wait(__base_type::_M_addr, __val);
408 }
409 while (!__pred());
410 }
411 };
412
413 using __enters_wait = __waiter<std::true_type>;
414 using __bare_wait = __waiter<std::false_type>;
415 } // namespace __detail
416
417 template<typename _Tp, typename _ValFn>
418 void
419 __atomic_wait_address_v(const _Tp* __addr, _Tp __old,
420 _ValFn __vfn) noexcept
421 {
422 __detail::__enters_wait __w(__addr);
423 __w._M_do_wait_v(__old, __vfn);
424 }
425
426 template<typename _Tp, typename _Pred>
427 void
428 __atomic_wait_address(const _Tp* __addr, _Pred __pred) noexcept
429 {
430 __detail::__enters_wait __w(__addr);
431 __w._M_do_wait(__pred);
432 }
433
434 // This call is to be used by atomic types which track contention externally
435 template<typename _Pred>
436 void
437 __atomic_wait_address_bare(const __detail::__platform_wait_t* __addr,
438 _Pred __pred) noexcept
439 {
440 #ifdef _GLIBCXX_HAVE_PLATFORM_WAIT
441 do
442 {
443 __detail::__platform_wait_t __val;
444 if (__detail::__bare_wait::_S_do_spin(__addr, __pred, __val))
445 return;
446 __detail::__platform_wait(__addr, __val);
447 }
448 while (!__pred());
449 #else // !_GLIBCXX_HAVE_PLATFORM_WAIT
450 __detail::__bare_wait __w(__addr);
451 __w._M_do_wait(__pred);
452 #endif
453 }
454
455 template<typename _Tp>
456 void
457 __atomic_notify_address(const _Tp* __addr, bool __all) noexcept
458 {
459 __detail::__bare_wait __w(__addr);
460 __w._M_notify(__all);
461 }
462
463 // This call is to be used by atomic types which track contention externally
464 inline void
465 __atomic_notify_address_bare(const __detail::__platform_wait_t* __addr,
466 bool __all) noexcept
467 {
468 #ifdef _GLIBCXX_HAVE_PLATFORM_WAIT
469 __detail::__platform_notify(__addr, __all);
470 #else
471 __detail::__bare_wait __w(__addr);
472 __w._M_notify(__all, true);
473 #endif
474 }
475 _GLIBCXX_END_NAMESPACE_VERSION
476 } // namespace std
477 #endif // GTHREADS || LINUX_FUTEX
478 #endif // _GLIBCXX_ATOMIC_WAIT_H