source: Daodan/MSYS2/mingw32/include/c++/11.2.0/bits/atomic_wait.h@ 1181

Last change on this file since 1181 was 1166, checked in by rossy, 3 years ago

Daodan: Replace MinGW build env with an up-to-date MSYS2 env

File size: 12.2 KB
RevLine 
[1166]1// -*- C++ -*- header.
2
3// Copyright (C) 2020-2021 Free Software Foundation, Inc.
4//
5// This file is part of the GNU ISO C++ Library. This library is free
6// software; you can redistribute it and/or modify it under the
7// terms of the GNU General Public License as published by the
8// Free Software Foundation; either version 3, or (at your option)
9// any later version.
10
11// This library is distributed in the hope that it will be useful,
12// but WITHOUT ANY WARRANTY; without even the implied warranty of
13// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
14// GNU General Public License for more details.
15
16// Under Section 7 of GPL version 3, you are granted additional
17// permissions described in the GCC Runtime Library Exception, version
18// 3.1, as published by the Free Software Foundation.
19
20// You should have received a copy of the GNU General Public License and
21// a copy of the GCC Runtime Library Exception along with this program;
22// see the files COPYING3 and COPYING.RUNTIME respectively. If not, see
23// <http://www.gnu.org/licenses/>.
24
25/** @file bits/atomic_wait.h
26 * This is an internal header file, included by other library headers.
27 * Do not attempt to use it directly. @headername{atomic}
28 */
29
30#ifndef _GLIBCXX_ATOMIC_WAIT_H
31#define _GLIBCXX_ATOMIC_WAIT_H 1
32
33#pragma GCC system_header
34
35#include <bits/c++config.h>
36#if defined _GLIBCXX_HAS_GTHREADS || defined _GLIBCXX_HAVE_LINUX_FUTEX
37#include <bits/functional_hash.h>
38#include <bits/gthr.h>
39#include <ext/numeric_traits.h>
40
41#ifdef _GLIBCXX_HAVE_LINUX_FUTEX
42# include <cerrno>
43# include <climits>
44# include <unistd.h>
45# include <syscall.h>
46# include <bits/functexcept.h>
47#endif
48
49# include <bits/std_mutex.h> // std::mutex, std::__condvar
50
51#define __cpp_lib_atomic_wait 201907L
52
53namespace std _GLIBCXX_VISIBILITY(default)
54{
55_GLIBCXX_BEGIN_NAMESPACE_VERSION
56 namespace __detail
57 {
58#ifdef _GLIBCXX_HAVE_LINUX_FUTEX
59 using __platform_wait_t = int;
60 static constexpr size_t __platform_wait_alignment = 4;
61#else
62 using __platform_wait_t = uint64_t;
63 static constexpr size_t __platform_wait_alignment
64 = __alignof__(__platform_wait_t);
65#endif
66 } // namespace __detail
67
68 template<typename _Tp>
69 inline constexpr bool __platform_wait_uses_type
70#ifdef _GLIBCXX_HAVE_PLATFORM_WAIT
71 = is_scalar_v<_Tp>
72 && ((sizeof(_Tp) == sizeof(__detail::__platform_wait_t))
73 && (alignof(_Tp*) >= __platform_wait_alignment));
74#else
75 = false;
76#endif
77
78 namespace __detail
79 {
80#ifdef _GLIBCXX_HAVE_LINUX_FUTEX
81#define _GLIBCXX_HAVE_PLATFORM_WAIT 1
82 enum class __futex_wait_flags : int
83 {
84#ifdef _GLIBCXX_HAVE_LINUX_FUTEX_PRIVATE
85 __private_flag = 128,
86#else
87 __private_flag = 0,
88#endif
89 __wait = 0,
90 __wake = 1,
91 __wait_bitset = 9,
92 __wake_bitset = 10,
93 __wait_private = __wait | __private_flag,
94 __wake_private = __wake | __private_flag,
95 __wait_bitset_private = __wait_bitset | __private_flag,
96 __wake_bitset_private = __wake_bitset | __private_flag,
97 __bitset_match_any = -1
98 };
99
100 template<typename _Tp>
101 void
102 __platform_wait(const _Tp* __addr, __platform_wait_t __val) noexcept
103 {
104 auto __e = syscall (SYS_futex, static_cast<const void*>(__addr),
105 static_cast<int>(__futex_wait_flags::__wait_private),
106 __val, nullptr);
107 if (!__e || errno == EAGAIN)
108 return;
109 if (errno != EINTR)
110 __throw_system_error(errno);
111 }
112
113 template<typename _Tp>
114 void
115 __platform_notify(const _Tp* __addr, bool __all) noexcept
116 {
117 syscall (SYS_futex, static_cast<const void*>(__addr),
118 static_cast<int>(__futex_wait_flags::__wake_private),
119 __all ? INT_MAX : 1);
120 }
121#else
122// define _GLIBCX_HAVE_PLATFORM_WAIT and implement __platform_wait()
123// and __platform_notify() if there is a more efficient primitive supported
124// by the platform (e.g. __ulock_wait()/__ulock_wake()) which is better than
125// a mutex/condvar based wait
126#endif
127
128 inline void
129 __thread_yield() noexcept
130 {
131#if defined _GLIBCXX_HAS_GTHREADS && defined _GLIBCXX_USE_SCHED_YIELD
132 __gthread_yield();
133#endif
134 }
135
136 inline void
137 __thread_relax() noexcept
138 {
139#if defined __i386__ || defined __x86_64__
140 __builtin_ia32_pause();
141#else
142 __thread_yield();
143#endif
144 }
145
146 constexpr auto __atomic_spin_count_1 = 12;
147 constexpr auto __atomic_spin_count_2 = 4;
148
149 struct __default_spin_policy
150 {
151 bool
152 operator()() const noexcept
153 { return false; }
154 };
155
156 template<typename _Pred,
157 typename _Spin = __default_spin_policy>
158 bool
159 __atomic_spin(_Pred& __pred, _Spin __spin = _Spin{ }) noexcept
160 {
161 for (auto __i = 0; __i < __atomic_spin_count_1; ++__i)
162 {
163 if (__pred())
164 return true;
165 __detail::__thread_relax();
166 }
167
168 for (auto __i = 0; __i < __atomic_spin_count_2; ++__i)
169 {
170 if (__pred())
171 return true;
172 __detail::__thread_yield();
173 }
174
175 while (__spin())
176 {
177 if (__pred())
178 return true;
179 }
180
181 return false;
182 }
183
184 // return true if equal
185 template<typename _Tp>
186 bool __atomic_compare(const _Tp& __a, const _Tp& __b)
187 {
188 // TODO make this do the correct padding bit ignoring comparison
189 return __builtin_memcmp(&__a, &__b, sizeof(_Tp)) == 0;
190 }
191
192 struct __waiter_pool_base
193 {
194#ifdef __cpp_lib_hardware_interference_size
195 static constexpr auto _S_align = hardware_destructive_interference_size;
196#else
197 static constexpr auto _S_align = 64;
198#endif
199
200 alignas(_S_align) __platform_wait_t _M_wait = 0;
201
202#ifndef _GLIBCXX_HAVE_PLATFORM_WAIT
203 mutex _M_mtx;
204#endif
205
206 alignas(_S_align) __platform_wait_t _M_ver = 0;
207
208#ifndef _GLIBCXX_HAVE_PLATFORM_WAIT
209 __condvar _M_cv;
210#endif
211 __waiter_pool_base() = default;
212
213 void
214 _M_enter_wait() noexcept
215 { __atomic_fetch_add(&_M_wait, 1, __ATOMIC_ACQ_REL); }
216
217 void
218 _M_leave_wait() noexcept
219 { __atomic_fetch_sub(&_M_wait, 1, __ATOMIC_ACQ_REL); }
220
221 bool
222 _M_waiting() const noexcept
223 {
224 __platform_wait_t __res;
225 __atomic_load(&_M_wait, &__res, __ATOMIC_ACQUIRE);
226 return __res > 0;
227 }
228
229 void
230 _M_notify(const __platform_wait_t* __addr, bool __all, bool __bare) noexcept
231 {
232 if (!(__bare || _M_waiting()))
233 return;
234
235#ifdef _GLIBCXX_HAVE_PLATFORM_WAIT
236 __platform_notify(__addr, __all);
237#else
238 if (__all)
239 _M_cv.notify_all();
240 else
241 _M_cv.notify_one();
242#endif
243 }
244
245 static __waiter_pool_base&
246 _S_for(const void* __addr) noexcept
247 {
248 constexpr uintptr_t __ct = 16;
249 static __waiter_pool_base __w[__ct];
250 auto __key = (uintptr_t(__addr) >> 2) % __ct;
251 return __w[__key];
252 }
253 };
254
255 struct __waiter_pool : __waiter_pool_base
256 {
257 void
258 _M_do_wait(const __platform_wait_t* __addr, __platform_wait_t __old) noexcept
259 {
260#ifdef _GLIBCXX_HAVE_PLATFORM_WAIT
261 __platform_wait(__addr, __old);
262#else
263 __platform_wait_t __val;
264 __atomic_load(__addr, &__val, __ATOMIC_RELAXED);
265 if (__val == __old)
266 {
267 lock_guard<mutex> __l(_M_mtx);
268 _M_cv.wait(_M_mtx);
269 }
270#endif // __GLIBCXX_HAVE_PLATFORM_WAIT
271 }
272 };
273
274 template<typename _Tp>
275 struct __waiter_base
276 {
277 using __waiter_type = _Tp;
278
279 __waiter_type& _M_w;
280 __platform_wait_t* _M_addr;
281
282 template<typename _Up>
283 static __platform_wait_t*
284 _S_wait_addr(const _Up* __a, __platform_wait_t* __b)
285 {
286 if constexpr (__platform_wait_uses_type<_Up>)
287 return reinterpret_cast<__platform_wait_t*>(const_cast<_Up*>(__a));
288 else
289 return __b;
290 }
291
292 static __waiter_type&
293 _S_for(const void* __addr) noexcept
294 {
295 static_assert(sizeof(__waiter_type) == sizeof(__waiter_pool_base));
296 auto& res = __waiter_pool_base::_S_for(__addr);
297 return reinterpret_cast<__waiter_type&>(res);
298 }
299
300 template<typename _Up>
301 explicit __waiter_base(const _Up* __addr) noexcept
302 : _M_w(_S_for(__addr))
303 , _M_addr(_S_wait_addr(__addr, &_M_w._M_ver))
304 { }
305
306 bool
307 _M_laundered() const
308 { return _M_addr == &_M_w._M_ver; }
309
310 void
311 _M_notify(bool __all, bool __bare = false)
312 {
313 if (_M_laundered())
314 {
315 __atomic_fetch_add(_M_addr, 1, __ATOMIC_ACQ_REL);
316 __all = true;
317 }
318 _M_w._M_notify(_M_addr, __all, __bare);
319 }
320
321 template<typename _Up, typename _ValFn,
322 typename _Spin = __default_spin_policy>
323 static bool
324 _S_do_spin_v(__platform_wait_t* __addr,
325 const _Up& __old, _ValFn __vfn,
326 __platform_wait_t& __val,
327 _Spin __spin = _Spin{ })
328 {
329 auto const __pred = [=]
330 { return !__detail::__atomic_compare(__old, __vfn()); };
331
332 if constexpr (__platform_wait_uses_type<_Up>)
333 {
334 __val == __old;
335 }
336 else
337 {
338 __atomic_load(__addr, &__val, __ATOMIC_RELAXED);
339 }
340 return __atomic_spin(__pred, __spin);
341 }
342
343 template<typename _Up, typename _ValFn,
344 typename _Spin = __default_spin_policy>
345 bool
346 _M_do_spin_v(const _Up& __old, _ValFn __vfn,
347 __platform_wait_t& __val,
348 _Spin __spin = _Spin{ })
349 { return _S_do_spin_v(_M_addr, __old, __vfn, __val, __spin); }
350
351 template<typename _Pred,
352 typename _Spin = __default_spin_policy>
353 static bool
354 _S_do_spin(const __platform_wait_t* __addr,
355 _Pred __pred,
356 __platform_wait_t& __val,
357 _Spin __spin = _Spin{ })
358 {
359 __atomic_load(__addr, &__val, __ATOMIC_RELAXED);
360 return __atomic_spin(__pred, __spin);
361 }
362
363 template<typename _Pred,
364 typename _Spin = __default_spin_policy>
365 bool
366 _M_do_spin(_Pred __pred, __platform_wait_t& __val,
367 _Spin __spin = _Spin{ })
368 { return _S_do_spin(_M_addr, __pred, __val, __spin); }
369 };
370
371 template<typename _EntersWait>
372 struct __waiter : __waiter_base<__waiter_pool>
373 {
374 using __base_type = __waiter_base<__waiter_pool>;
375
376 template<typename _Tp>
377 explicit __waiter(const _Tp* __addr) noexcept
378 : __base_type(__addr)
379 {
380 if constexpr (_EntersWait::value)
381 _M_w._M_enter_wait();
382 }
383
384 ~__waiter()
385 {
386 if constexpr (_EntersWait::value)
387 _M_w._M_leave_wait();
388 }
389
390 template<typename _Tp, typename _ValFn>
391 void
392 _M_do_wait_v(_Tp __old, _ValFn __vfn)
393 {
394 __platform_wait_t __val;
395 if (__base_type::_M_do_spin_v(__old, __vfn, __val))
396 return;
397
398 do
399 {
400 __base_type::_M_w._M_do_wait(__base_type::_M_addr, __val);
401 }
402 while (__detail::__atomic_compare(__old, __vfn()));
403 }
404
405 template<typename _Pred>
406 void
407 _M_do_wait(_Pred __pred) noexcept
408 {
409 do
410 {
411 __platform_wait_t __val;
412 if (__base_type::_M_do_spin(__pred, __val))
413 return;
414 __base_type::_M_w._M_do_wait(__base_type::_M_addr, __val);
415 }
416 while (!__pred());
417 }
418 };
419
420 using __enters_wait = __waiter<std::true_type>;
421 using __bare_wait = __waiter<std::false_type>;
422 } // namespace __detail
423
424 template<typename _Tp, typename _ValFn>
425 void
426 __atomic_wait_address_v(const _Tp* __addr, _Tp __old,
427 _ValFn __vfn) noexcept
428 {
429 __detail::__enters_wait __w(__addr);
430 __w._M_do_wait_v(__old, __vfn);
431 }
432
433 template<typename _Tp, typename _Pred>
434 void
435 __atomic_wait_address(const _Tp* __addr, _Pred __pred) noexcept
436 {
437 __detail::__enters_wait __w(__addr);
438 __w._M_do_wait(__pred);
439 }
440
441 // This call is to be used by atomic types which track contention externally
442 template<typename _Pred>
443 void
444 __atomic_wait_address_bare(const __detail::__platform_wait_t* __addr,
445 _Pred __pred) noexcept
446 {
447#ifdef _GLIBCXX_HAVE_PLATFORM_WAIT
448 do
449 {
450 __detail::__platform_wait_t __val;
451 if (__detail::__bare_wait::_S_do_spin(__addr, __pred, __val))
452 return;
453 __detail::__platform_wait(__addr, __val);
454 }
455 while (!__pred());
456#else // !_GLIBCXX_HAVE_PLATFORM_WAIT
457 __detail::__bare_wait __w(__addr);
458 __w._M_do_wait(__pred);
459#endif
460 }
461
462 template<typename _Tp>
463 void
464 __atomic_notify_address(const _Tp* __addr, bool __all) noexcept
465 {
466 __detail::__bare_wait __w(__addr);
467 __w._M_notify(__all);
468 }
469
470 // This call is to be used by atomic types which track contention externally
471 inline void
472 __atomic_notify_address_bare(const __detail::__platform_wait_t* __addr,
473 bool __all) noexcept
474 {
475#ifdef _GLIBCXX_HAVE_PLATFORM_WAIT
476 __detail::__platform_notify(__addr, __all);
477#else
478 __detail::__bare_wait __w(__addr);
479 __w._M_notify(__all, true);
480#endif
481 }
482_GLIBCXX_END_NAMESPACE_VERSION
483} // namespace std
484#endif // GTHREADS || LINUX_FUTEX
485#endif // _GLIBCXX_ATOMIC_WAIT_H
Note: See TracBrowser for help on using the repository browser.