xref: /netbsd-src/external/gpl3/gcc/dist/libstdc++-v3/include/std/mutex (revision c42dbd0ed2e61fe6eda8590caa852ccf34719964)
1// <mutex> -*- C++ -*-
2
3// Copyright (C) 2003-2022 Free Software Foundation, Inc.
4//
5// This file is part of the GNU ISO C++ Library.  This library is free
6// software; you can redistribute it and/or modify it under the
7// terms of the GNU General Public License as published by the
8// Free Software Foundation; either version 3, or (at your option)
9// any later version.
10
11// This library is distributed in the hope that it will be useful,
12// but WITHOUT ANY WARRANTY; without even the implied warranty of
13// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
14// GNU General Public License for more details.
15
16// Under Section 7 of GPL version 3, you are granted additional
17// permissions described in the GCC Runtime Library Exception, version
18// 3.1, as published by the Free Software Foundation.
19
20// You should have received a copy of the GNU General Public License and
21// a copy of the GCC Runtime Library Exception along with this program;
22// see the files COPYING3 and COPYING.RUNTIME respectively.  If not, see
23// <http://www.gnu.org/licenses/>.
24
25/** @file include/mutex
26 *  This is a Standard C++ Library header.
27 */
28
29#ifndef _GLIBCXX_MUTEX
30#define _GLIBCXX_MUTEX 1
31
32#pragma GCC system_header
33
34#if __cplusplus < 201103L
35# include <bits/c++0x_warning.h>
36#else
37
38#include <tuple>
39#include <exception>
40#include <type_traits>
41#include <system_error>
42#include <bits/chrono.h>
43#include <bits/std_mutex.h>
44#include <bits/unique_lock.h>
45#if ! _GTHREAD_USE_MUTEX_TIMEDLOCK
46# include <condition_variable>
47# include <thread>
48#endif
49#include <ext/atomicity.h>     // __gnu_cxx::__is_single_threaded
50
51#if defined _GLIBCXX_HAS_GTHREADS && ! defined _GLIBCXX_HAVE_TLS
52# include <bits/std_function.h>  // std::function
53#endif
54
55namespace std _GLIBCXX_VISIBILITY(default)
56{
57_GLIBCXX_BEGIN_NAMESPACE_VERSION
58
59  /**
60   * @addtogroup mutexes
61   * @{
62   */
63
64#ifdef _GLIBCXX_HAS_GTHREADS
65
66  // Common base class for std::recursive_mutex and std::recursive_timed_mutex
67  class __recursive_mutex_base
68  {
69  protected:
70    typedef __gthread_recursive_mutex_t		__native_type;
71
72    __recursive_mutex_base(const __recursive_mutex_base&) = delete;
73    __recursive_mutex_base& operator=(const __recursive_mutex_base&) = delete;
74
75#ifdef __GTHREAD_RECURSIVE_MUTEX_INIT
76    __native_type  _M_mutex = __GTHREAD_RECURSIVE_MUTEX_INIT;
77
78    __recursive_mutex_base() = default;
79#else
80    __native_type  _M_mutex;
81
82    __recursive_mutex_base()
83    {
84      // XXX EAGAIN, ENOMEM, EPERM, EBUSY(may), EINVAL(may)
85      __GTHREAD_RECURSIVE_MUTEX_INIT_FUNCTION(&_M_mutex);
86    }
87
88    ~__recursive_mutex_base()
89    { __gthread_recursive_mutex_destroy(&_M_mutex); }
90#endif
91  };
92
93  /// The standard recursive mutex type.
94  class recursive_mutex : private __recursive_mutex_base
95  {
96  public:
97    typedef __native_type* 			native_handle_type;
98
99    recursive_mutex() = default;
100    ~recursive_mutex() = default;
101
102    recursive_mutex(const recursive_mutex&) = delete;
103    recursive_mutex& operator=(const recursive_mutex&) = delete;
104
105    void
106    lock()
107    {
108      int __e = __gthread_recursive_mutex_lock(&_M_mutex);
109
110      // EINVAL, EAGAIN, EBUSY, EINVAL, EDEADLK(may)
111      if (__e)
112	__throw_system_error(__e);
113    }
114
115    bool
116    try_lock() noexcept
117    {
118      // XXX EINVAL, EAGAIN, EBUSY
119      return !__gthread_recursive_mutex_trylock(&_M_mutex);
120    }
121
122    void
123    unlock()
124    {
125      // XXX EINVAL, EAGAIN, EBUSY
126      __gthread_recursive_mutex_unlock(&_M_mutex);
127    }
128
129    native_handle_type
130    native_handle() noexcept
131    { return &_M_mutex; }
132  };
133
134#if _GTHREAD_USE_MUTEX_TIMEDLOCK
135  template<typename _Derived>
136    class __timed_mutex_impl
137    {
138    protected:
139      template<typename _Rep, typename _Period>
140	bool
141	_M_try_lock_for(const chrono::duration<_Rep, _Period>& __rtime)
142	{
143#if _GLIBCXX_USE_PTHREAD_MUTEX_CLOCKLOCK
144	  using __clock = chrono::steady_clock;
145#else
146	  using __clock = chrono::system_clock;
147#endif
148
149	  auto __rt = chrono::duration_cast<__clock::duration>(__rtime);
150	  if (ratio_greater<__clock::period, _Period>())
151	    ++__rt;
152	  return _M_try_lock_until(__clock::now() + __rt);
153	}
154
155      template<typename _Duration>
156	bool
157	_M_try_lock_until(const chrono::time_point<chrono::system_clock,
158						   _Duration>& __atime)
159	{
160	  auto __s = chrono::time_point_cast<chrono::seconds>(__atime);
161	  auto __ns = chrono::duration_cast<chrono::nanoseconds>(__atime - __s);
162
163	  __gthread_time_t __ts = {
164	    static_cast<std::time_t>(__s.time_since_epoch().count()),
165	    static_cast<long>(__ns.count())
166	  };
167
168	  return static_cast<_Derived*>(this)->_M_timedlock(__ts);
169	}
170
171#ifdef _GLIBCXX_USE_PTHREAD_MUTEX_CLOCKLOCK
172      template<typename _Duration>
173	bool
174	_M_try_lock_until(const chrono::time_point<chrono::steady_clock,
175						   _Duration>& __atime)
176	{
177	  auto __s = chrono::time_point_cast<chrono::seconds>(__atime);
178	  auto __ns = chrono::duration_cast<chrono::nanoseconds>(__atime - __s);
179
180	  __gthread_time_t __ts = {
181	    static_cast<std::time_t>(__s.time_since_epoch().count()),
182	    static_cast<long>(__ns.count())
183	  };
184
185	  return static_cast<_Derived*>(this)->_M_clocklock(CLOCK_MONOTONIC,
186							    __ts);
187	}
188#endif
189
190      template<typename _Clock, typename _Duration>
191	bool
192	_M_try_lock_until(const chrono::time_point<_Clock, _Duration>& __atime)
193	{
194#if __cplusplus > 201703L
195	  static_assert(chrono::is_clock_v<_Clock>);
196#endif
197	  // The user-supplied clock may not tick at the same rate as
198	  // steady_clock, so we must loop in order to guarantee that
199	  // the timeout has expired before returning false.
200	  auto __now = _Clock::now();
201	  do {
202	    auto __rtime = __atime - __now;
203	    if (_M_try_lock_for(__rtime))
204	      return true;
205	    __now = _Clock::now();
206	  } while (__atime > __now);
207	  return false;
208	}
209    };
210
211  /// The standard timed mutex type.
212  class timed_mutex
213  : private __mutex_base, public __timed_mutex_impl<timed_mutex>
214  {
215  public:
216    typedef __native_type* 		  	native_handle_type;
217
218    timed_mutex() = default;
219    ~timed_mutex() = default;
220
221    timed_mutex(const timed_mutex&) = delete;
222    timed_mutex& operator=(const timed_mutex&) = delete;
223
224    void
225    lock()
226    {
227      int __e = __gthread_mutex_lock(&_M_mutex);
228
229      // EINVAL, EAGAIN, EBUSY, EINVAL, EDEADLK(may)
230      if (__e)
231	__throw_system_error(__e);
232    }
233
234    bool
235    try_lock() noexcept
236    {
237      // XXX EINVAL, EAGAIN, EBUSY
238      return !__gthread_mutex_trylock(&_M_mutex);
239    }
240
241    template <class _Rep, class _Period>
242      bool
243      try_lock_for(const chrono::duration<_Rep, _Period>& __rtime)
244      { return _M_try_lock_for(__rtime); }
245
246    template <class _Clock, class _Duration>
247      bool
248      try_lock_until(const chrono::time_point<_Clock, _Duration>& __atime)
249      { return _M_try_lock_until(__atime); }
250
251    void
252    unlock()
253    {
254      // XXX EINVAL, EAGAIN, EBUSY
255      __gthread_mutex_unlock(&_M_mutex);
256    }
257
258    native_handle_type
259    native_handle() noexcept
260    { return &_M_mutex; }
261
262    private:
263      friend class __timed_mutex_impl<timed_mutex>;
264
265      bool
266      _M_timedlock(const __gthread_time_t& __ts)
267      { return !__gthread_mutex_timedlock(&_M_mutex, &__ts); }
268
269#if _GLIBCXX_USE_PTHREAD_MUTEX_CLOCKLOCK
270      bool
271      _M_clocklock(clockid_t clockid, const __gthread_time_t& __ts)
272      { return !pthread_mutex_clocklock(&_M_mutex, clockid, &__ts); }
273#endif
274  };
275
276  /// recursive_timed_mutex
277  class recursive_timed_mutex
278  : private __recursive_mutex_base,
279    public __timed_mutex_impl<recursive_timed_mutex>
280  {
281  public:
282    typedef __native_type* 			native_handle_type;
283
284    recursive_timed_mutex() = default;
285    ~recursive_timed_mutex() = default;
286
287    recursive_timed_mutex(const recursive_timed_mutex&) = delete;
288    recursive_timed_mutex& operator=(const recursive_timed_mutex&) = delete;
289
290    void
291    lock()
292    {
293      int __e = __gthread_recursive_mutex_lock(&_M_mutex);
294
295      // EINVAL, EAGAIN, EBUSY, EINVAL, EDEADLK(may)
296      if (__e)
297	__throw_system_error(__e);
298    }
299
300    bool
301    try_lock() noexcept
302    {
303      // XXX EINVAL, EAGAIN, EBUSY
304      return !__gthread_recursive_mutex_trylock(&_M_mutex);
305    }
306
307    template <class _Rep, class _Period>
308      bool
309      try_lock_for(const chrono::duration<_Rep, _Period>& __rtime)
310      { return _M_try_lock_for(__rtime); }
311
312    template <class _Clock, class _Duration>
313      bool
314      try_lock_until(const chrono::time_point<_Clock, _Duration>& __atime)
315      { return _M_try_lock_until(__atime); }
316
317    void
318    unlock()
319    {
320      // XXX EINVAL, EAGAIN, EBUSY
321      __gthread_recursive_mutex_unlock(&_M_mutex);
322    }
323
324    native_handle_type
325    native_handle() noexcept
326    { return &_M_mutex; }
327
328    private:
329      friend class __timed_mutex_impl<recursive_timed_mutex>;
330
331      bool
332      _M_timedlock(const __gthread_time_t& __ts)
333      { return !__gthread_recursive_mutex_timedlock(&_M_mutex, &__ts); }
334
335#ifdef _GLIBCXX_USE_PTHREAD_MUTEX_CLOCKLOCK
336      bool
337      _M_clocklock(clockid_t clockid, const __gthread_time_t& __ts)
338      { return !pthread_mutex_clocklock(&_M_mutex, clockid, &__ts); }
339#endif
340  };
341
342#else // !_GTHREAD_USE_MUTEX_TIMEDLOCK
343
344  /// timed_mutex
345  class timed_mutex
346  {
347    mutex		_M_mut;
348    condition_variable	_M_cv;
349    bool		_M_locked = false;
350
351  public:
352
353    timed_mutex() = default;
354    ~timed_mutex() { __glibcxx_assert( !_M_locked ); }
355
356    timed_mutex(const timed_mutex&) = delete;
357    timed_mutex& operator=(const timed_mutex&) = delete;
358
359    void
360    lock()
361    {
362      unique_lock<mutex> __lk(_M_mut);
363      _M_cv.wait(__lk, [&]{ return !_M_locked; });
364      _M_locked = true;
365    }
366
367    bool
368    try_lock()
369    {
370      lock_guard<mutex> __lk(_M_mut);
371      if (_M_locked)
372	return false;
373      _M_locked = true;
374      return true;
375    }
376
377    template<typename _Rep, typename _Period>
378      bool
379      try_lock_for(const chrono::duration<_Rep, _Period>& __rtime)
380      {
381	unique_lock<mutex> __lk(_M_mut);
382	if (!_M_cv.wait_for(__lk, __rtime, [&]{ return !_M_locked; }))
383	  return false;
384	_M_locked = true;
385	return true;
386      }
387
388    template<typename _Clock, typename _Duration>
389      bool
390      try_lock_until(const chrono::time_point<_Clock, _Duration>& __atime)
391      {
392	unique_lock<mutex> __lk(_M_mut);
393	if (!_M_cv.wait_until(__lk, __atime, [&]{ return !_M_locked; }))
394	  return false;
395	_M_locked = true;
396	return true;
397      }
398
399    void
400    unlock()
401    {
402      lock_guard<mutex> __lk(_M_mut);
403      __glibcxx_assert( _M_locked );
404      _M_locked = false;
405      _M_cv.notify_one();
406    }
407  };
408
409  /// recursive_timed_mutex
410  class recursive_timed_mutex
411  {
412    mutex		_M_mut;
413    condition_variable	_M_cv;
414    thread::id		_M_owner;
415    unsigned		_M_count = 0;
416
417    // Predicate type that tests whether the current thread can lock a mutex.
418    struct _Can_lock
419    {
420      // Returns true if the mutex is unlocked or is locked by _M_caller.
421      bool
422      operator()() const noexcept
423      { return _M_mx->_M_count == 0 || _M_mx->_M_owner == _M_caller; }
424
425      const recursive_timed_mutex* _M_mx;
426      thread::id _M_caller;
427    };
428
429  public:
430
431    recursive_timed_mutex() = default;
432    ~recursive_timed_mutex() { __glibcxx_assert( _M_count == 0 ); }
433
434    recursive_timed_mutex(const recursive_timed_mutex&) = delete;
435    recursive_timed_mutex& operator=(const recursive_timed_mutex&) = delete;
436
437    void
438    lock()
439    {
440      auto __id = this_thread::get_id();
441      _Can_lock __can_lock{this, __id};
442      unique_lock<mutex> __lk(_M_mut);
443      _M_cv.wait(__lk, __can_lock);
444      if (_M_count == -1u)
445	__throw_system_error(EAGAIN); // [thread.timedmutex.recursive]/3
446      _M_owner = __id;
447      ++_M_count;
448    }
449
450    bool
451    try_lock()
452    {
453      auto __id = this_thread::get_id();
454      _Can_lock __can_lock{this, __id};
455      lock_guard<mutex> __lk(_M_mut);
456      if (!__can_lock())
457	return false;
458      if (_M_count == -1u)
459	return false;
460      _M_owner = __id;
461      ++_M_count;
462      return true;
463    }
464
465    template<typename _Rep, typename _Period>
466      bool
467      try_lock_for(const chrono::duration<_Rep, _Period>& __rtime)
468      {
469	auto __id = this_thread::get_id();
470	_Can_lock __can_lock{this, __id};
471	unique_lock<mutex> __lk(_M_mut);
472	if (!_M_cv.wait_for(__lk, __rtime, __can_lock))
473	  return false;
474	if (_M_count == -1u)
475	  return false;
476	_M_owner = __id;
477	++_M_count;
478	return true;
479      }
480
481    template<typename _Clock, typename _Duration>
482      bool
483      try_lock_until(const chrono::time_point<_Clock, _Duration>& __atime)
484      {
485	auto __id = this_thread::get_id();
486	_Can_lock __can_lock{this, __id};
487	unique_lock<mutex> __lk(_M_mut);
488	if (!_M_cv.wait_until(__lk, __atime, __can_lock))
489	  return false;
490	if (_M_count == -1u)
491	  return false;
492	_M_owner = __id;
493	++_M_count;
494	return true;
495      }
496
497    void
498    unlock()
499    {
500      lock_guard<mutex> __lk(_M_mut);
501      __glibcxx_assert( _M_owner == this_thread::get_id() );
502      __glibcxx_assert( _M_count > 0 );
503      if (--_M_count == 0)
504	{
505	  _M_owner = {};
506	  _M_cv.notify_one();
507	}
508    }
509  };
510
511#endif
512#endif // _GLIBCXX_HAS_GTHREADS
513
514  /// @cond undocumented
515  namespace __detail
516  {
517    // Lock the last lockable, after all previous ones are locked.
518    template<typename _Lockable>
519      inline int
520      __try_lock_impl(_Lockable& __l)
521      {
522	if (unique_lock<_Lockable> __lock{__l, try_to_lock})
523	  {
524	    __lock.release();
525	    return -1;
526	  }
527	else
528	  return 0;
529      }
530
531    // Lock each lockable in turn.
532    // Use iteration if all lockables are the same type, recursion otherwise.
533    template<typename _L0, typename... _Lockables>
534      inline int
535      __try_lock_impl(_L0& __l0, _Lockables&... __lockables)
536      {
537#if __cplusplus >= 201703L
538	if constexpr ((is_same_v<_L0, _Lockables> && ...))
539	  {
540	    constexpr int _Np = 1 + sizeof...(_Lockables);
541	    unique_lock<_L0> __locks[_Np] = {
542		{__l0, defer_lock}, {__lockables, defer_lock}...
543	    };
544	    for (int __i = 0; __i < _Np; ++__i)
545	      {
546		if (!__locks[__i].try_lock())
547		  {
548		    const int __failed = __i;
549		    while (__i--)
550		      __locks[__i].unlock();
551		    return __failed;
552		  }
553	      }
554	    for (auto& __l : __locks)
555	      __l.release();
556	    return -1;
557	  }
558	else
559#endif
560	if (unique_lock<_L0> __lock{__l0, try_to_lock})
561	  {
562	    int __idx = __detail::__try_lock_impl(__lockables...);
563	    if (__idx == -1)
564	      {
565		__lock.release();
566		return -1;
567	      }
568	    return __idx + 1;
569	  }
570	else
571	  return 0;
572      }
573
574  } // namespace __detail
575  /// @endcond
576
577  /** @brief Generic try_lock.
578   *  @param __l1 Meets Lockable requirements (try_lock() may throw).
579   *  @param __l2 Meets Lockable requirements (try_lock() may throw).
580   *  @param __l3 Meets Lockable requirements (try_lock() may throw).
581   *  @return Returns -1 if all try_lock() calls return true. Otherwise returns
582   *          a 0-based index corresponding to the argument that returned false.
583   *  @post Either all arguments are locked, or none will be.
584   *
585   *  Sequentially calls try_lock() on each argument.
586   */
587  template<typename _L1, typename _L2, typename... _L3>
588    inline int
589    try_lock(_L1& __l1, _L2& __l2, _L3&... __l3)
590    {
591      return __detail::__try_lock_impl(__l1, __l2, __l3...);
592    }
593
594  /// @cond undocumented
595  namespace __detail
596  {
597    // This function can recurse up to N levels deep, for N = 1+sizeof...(L1).
598    // On each recursion the lockables are rotated left one position,
599    // e.g. depth 0: l0, l1, l2; depth 1: l1, l2, l0; depth 2: l2, l0, l1.
600    // When a call to l_i.try_lock() fails it recurses/returns to depth=i
601    // so that l_i is the first argument, and then blocks until l_i is locked.
602    template<typename _L0, typename... _L1>
603      void
604      __lock_impl(int& __i, int __depth, _L0& __l0, _L1&... __l1)
605      {
606	while (__i >= __depth)
607	  {
608	    if (__i == __depth)
609	      {
610		int __failed = 1; // index that couldn't be locked
611		{
612		  unique_lock<_L0> __first(__l0);
613		  __failed += __detail::__try_lock_impl(__l1...);
614		  if (!__failed)
615		    {
616		      __i = -1; // finished
617		      __first.release();
618		      return;
619		    }
620		}
621#if defined _GLIBCXX_HAS_GTHREADS && defined _GLIBCXX_USE_SCHED_YIELD
622		__gthread_yield();
623#endif
624		constexpr auto __n = 1 + sizeof...(_L1);
625		__i = (__depth + __failed) % __n;
626	      }
627	    else // rotate left until l_i is first.
628	      __detail::__lock_impl(__i, __depth + 1, __l1..., __l0);
629	  }
630      }
631
632  } // namespace __detail
633  /// @endcond
634
635  /** @brief Generic lock.
636   *  @param __l1 Meets Lockable requirements (try_lock() may throw).
637   *  @param __l2 Meets Lockable requirements (try_lock() may throw).
638   *  @param __l3 Meets Lockable requirements (try_lock() may throw).
639   *  @throw An exception thrown by an argument's lock() or try_lock() member.
640   *  @post All arguments are locked.
641   *
642   *  All arguments are locked via a sequence of calls to lock(), try_lock()
643   *  and unlock().  If this function exits via an exception any locks that
644   *  were obtained will be released.
645   */
646  template<typename _L1, typename _L2, typename... _L3>
647    void
648    lock(_L1& __l1, _L2& __l2, _L3&... __l3)
649    {
650#if __cplusplus >= 201703L
651      if constexpr (is_same_v<_L1, _L2> && (is_same_v<_L1, _L3> && ...))
652	{
653	  constexpr int _Np = 2 + sizeof...(_L3);
654	  unique_lock<_L1> __locks[] = {
655	      {__l1, defer_lock}, {__l2, defer_lock}, {__l3, defer_lock}...
656	  };
657	  int __first = 0;
658	  do {
659	    __locks[__first].lock();
660	    for (int __j = 1; __j < _Np; ++__j)
661	      {
662		const int __idx = (__first + __j) % _Np;
663		if (!__locks[__idx].try_lock())
664		  {
665		    for (int __k = __j; __k != 0; --__k)
666		      __locks[(__first + __k - 1) % _Np].unlock();
667		    __first = __idx;
668		    break;
669		  }
670	      }
671	  } while (!__locks[__first].owns_lock());
672
673	  for (auto& __l : __locks)
674	    __l.release();
675	}
676      else
677#endif
678	{
679	  int __i = 0;
680	  __detail::__lock_impl(__i, 0, __l1, __l2, __l3...);
681	}
682    }
683
684#if __cplusplus >= 201703L
685#define __cpp_lib_scoped_lock 201703L
686  /** @brief A scoped lock type for multiple lockable objects.
687   *
688   * A scoped_lock controls mutex ownership within a scope, releasing
689   * ownership in the destructor.
690   */
691  template<typename... _MutexTypes>
692    class scoped_lock
693    {
694    public:
695      explicit scoped_lock(_MutexTypes&... __m) : _M_devices(std::tie(__m...))
696      { std::lock(__m...); }
697
698      explicit scoped_lock(adopt_lock_t, _MutexTypes&... __m) noexcept
699      : _M_devices(std::tie(__m...))
700      { } // calling thread owns mutex
701
702      ~scoped_lock()
703      { std::apply([](auto&... __m) { (__m.unlock(), ...); }, _M_devices); }
704
705      scoped_lock(const scoped_lock&) = delete;
706      scoped_lock& operator=(const scoped_lock&) = delete;
707
708    private:
709      tuple<_MutexTypes&...> _M_devices;
710    };
711
712  template<>
713    class scoped_lock<>
714    {
715    public:
716      explicit scoped_lock() = default;
717      explicit scoped_lock(adopt_lock_t) noexcept { }
718      ~scoped_lock() = default;
719
720      scoped_lock(const scoped_lock&) = delete;
721      scoped_lock& operator=(const scoped_lock&) = delete;
722    };
723
724  template<typename _Mutex>
725    class scoped_lock<_Mutex>
726    {
727    public:
728      using mutex_type = _Mutex;
729
730      explicit scoped_lock(mutex_type& __m) : _M_device(__m)
731      { _M_device.lock(); }
732
733      explicit scoped_lock(adopt_lock_t, mutex_type& __m) noexcept
734      : _M_device(__m)
735      { } // calling thread owns mutex
736
737      ~scoped_lock()
738      { _M_device.unlock(); }
739
740      scoped_lock(const scoped_lock&) = delete;
741      scoped_lock& operator=(const scoped_lock&) = delete;
742
743    private:
744      mutex_type&  _M_device;
745    };
746#endif // C++17
747
748#ifdef _GLIBCXX_HAS_GTHREADS
749  /// Flag type used by std::call_once
750  struct once_flag
751  {
752    constexpr once_flag() noexcept = default;
753
754    /// Deleted copy constructor
755    once_flag(const once_flag&) = delete;
756    /// Deleted assignment operator
757    once_flag& operator=(const once_flag&) = delete;
758
759  private:
760    // For gthreads targets a pthread_once_t is used with pthread_once, but
761    // for most targets this doesn't work correctly for exceptional executions.
762    __gthread_once_t _M_once = __GTHREAD_ONCE_INIT;
763
764    struct _Prepare_execution;
765
766    template<typename _Callable, typename... _Args>
767      friend void
768      call_once(once_flag& __once, _Callable&& __f, _Args&&... __args);
769  };
770
771  /// @cond undocumented
772# ifdef _GLIBCXX_HAVE_TLS
773  // If TLS is available use thread-local state for the type-erased callable
774  // that is being run by std::call_once in the current thread.
775  extern __thread void* __once_callable;
776  extern __thread void (*__once_call)();
777
778  // RAII type to set up state for pthread_once call.
779  struct once_flag::_Prepare_execution
780  {
781    template<typename _Callable>
782      explicit
783      _Prepare_execution(_Callable& __c)
784      {
785	// Store address in thread-local pointer:
786	__once_callable = std::__addressof(__c);
787	// Trampoline function to invoke the closure via thread-local pointer:
788	__once_call = [] { (*static_cast<_Callable*>(__once_callable))(); };
789      }
790
791    ~_Prepare_execution()
792    {
793      // PR libstdc++/82481
794      __once_callable = nullptr;
795      __once_call = nullptr;
796    }
797
798    _Prepare_execution(const _Prepare_execution&) = delete;
799    _Prepare_execution& operator=(const _Prepare_execution&) = delete;
800  };
801
802# else
803  // Without TLS use a global std::mutex and store the callable in a
804  // global std::function.
805  extern function<void()> __once_functor;
806
807  extern void
808  __set_once_functor_lock_ptr(unique_lock<mutex>*);
809
810  extern mutex&
811  __get_once_mutex();
812
813  // RAII type to set up state for pthread_once call.
814  struct once_flag::_Prepare_execution
815  {
816    template<typename _Callable>
817      explicit
818      _Prepare_execution(_Callable& __c)
819      {
820	// Store the callable in the global std::function
821	__once_functor = __c;
822	__set_once_functor_lock_ptr(&_M_functor_lock);
823      }
824
825    ~_Prepare_execution()
826    {
827      if (_M_functor_lock)
828	__set_once_functor_lock_ptr(nullptr);
829    }
830
831  private:
832    // XXX This deadlocks if used recursively (PR 97949)
833    unique_lock<mutex> _M_functor_lock{__get_once_mutex()};
834
835    _Prepare_execution(const _Prepare_execution&) = delete;
836    _Prepare_execution& operator=(const _Prepare_execution&) = delete;
837  };
838# endif
839  /// @endcond
840
841  // This function is passed to pthread_once by std::call_once.
842  // It runs __once_call() or __once_functor().
843  extern "C" void __once_proxy(void);
844
845  /// Invoke a callable and synchronize with other calls using the same flag
846  template<typename _Callable, typename... _Args>
847    void
848    call_once(once_flag& __once, _Callable&& __f, _Args&&... __args)
849    {
850      // Closure type that runs the function
851      auto __callable = [&] {
852	  std::__invoke(std::forward<_Callable>(__f),
853			std::forward<_Args>(__args)...);
854      };
855
856      once_flag::_Prepare_execution __exec(__callable);
857
858      // XXX pthread_once does not reset the flag if an exception is thrown.
859      if (int __e = __gthread_once(&__once._M_once, &__once_proxy))
860	__throw_system_error(__e);
861    }
862
863#else // _GLIBCXX_HAS_GTHREADS
864
865  /// Flag type used by std::call_once
866  struct once_flag
867  {
868    constexpr once_flag() noexcept = default;
869
870    /// Deleted copy constructor
871    once_flag(const once_flag&) = delete;
872    /// Deleted assignment operator
873    once_flag& operator=(const once_flag&) = delete;
874
875  private:
876    // There are two different std::once_flag interfaces, abstracting four
877    // different implementations.
878    // The single-threaded interface uses the _M_activate() and _M_finish(bool)
879    // functions, which start and finish an active execution respectively.
880    // See [thread.once.callonce] in C++11 for the definition of
881    // active/passive/returning/exceptional executions.
882    enum _Bits : int { _Init = 0, _Active = 1, _Done = 2 };
883
884    int _M_once = _Bits::_Init;
885
886    // Check to see if all executions will be passive now.
887    bool
888    _M_passive() const noexcept;
889
890    // Attempts to begin an active execution.
891    bool _M_activate();
892
893    // Must be called to complete an active execution.
894    // The argument is true if the active execution was a returning execution,
895    // false if it was an exceptional execution.
896    void _M_finish(bool __returning) noexcept;
897
898    // RAII helper to call _M_finish.
899    struct _Active_execution
900    {
901      explicit _Active_execution(once_flag& __flag) : _M_flag(__flag) { }
902
903      ~_Active_execution() { _M_flag._M_finish(_M_returning); }
904
905      _Active_execution(const _Active_execution&) = delete;
906      _Active_execution& operator=(const _Active_execution&) = delete;
907
908      once_flag& _M_flag;
909      bool _M_returning = false;
910    };
911
912    template<typename _Callable, typename... _Args>
913      friend void
914      call_once(once_flag& __once, _Callable&& __f, _Args&&... __args);
915  };
916
917  // Inline definitions of std::once_flag members for single-threaded targets.
918
919  inline bool
920  once_flag::_M_passive() const noexcept
921  { return _M_once == _Bits::_Done; }
922
923  inline bool
924  once_flag::_M_activate()
925  {
926    if (_M_once == _Bits::_Init) [[__likely__]]
927      {
928	_M_once = _Bits::_Active;
929	return true;
930      }
931    else if (_M_passive()) // Caller should have checked this already.
932      return false;
933    else
934      __throw_system_error(EDEADLK);
935  }
936
937  inline void
938  once_flag::_M_finish(bool __returning) noexcept
939  { _M_once = __returning ? _Bits::_Done : _Bits::_Init; }
940
941  /// Invoke a callable and synchronize with other calls using the same flag
942  template<typename _Callable, typename... _Args>
943    inline void
944    call_once(once_flag& __once, _Callable&& __f, _Args&&... __args)
945    {
946      if (__once._M_passive())
947	return;
948      else if (__once._M_activate())
949	{
950	  once_flag::_Active_execution __exec(__once);
951
952	  // _GLIBCXX_RESOLVE_LIB_DEFECTS
953	  // 2442. call_once() shouldn't DECAY_COPY()
954	  std::__invoke(std::forward<_Callable>(__f),
955			std::forward<_Args>(__args)...);
956
957	  // __f(__args...) did not throw
958	  __exec._M_returning = true;
959	}
960    }
961#endif // _GLIBCXX_HAS_GTHREADS
962
963  /// @} group mutexes
964_GLIBCXX_END_NAMESPACE_VERSION
965} // namespace
966
967#endif // C++11
968
969#endif // _GLIBCXX_MUTEX
970