xref: /netbsd-src/external/gpl3/gcc/dist/libstdc++-v3/include/std/stop_token (revision b1e838363e3c6fc78a55519254d99869742dd33c)
1// <stop_token> -*- C++ -*-
2
3// Copyright (C) 2019-2022 Free Software Foundation, Inc.
4//
5// This file is part of the GNU ISO C++ Library.  This library is free
6// software; you can redistribute it and/or modify it under the
7// terms of the GNU General Public License as published by the
8// Free Software Foundation; either version 3, or (at your option)
9// any later version.
10
11// This library is distributed in the hope that it will be useful,
12// but WITHOUT ANY WARRANTY; without even the implied warranty of
13// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
14// GNU General Public License for more details.
15
16// Under Section 7 of GPL version 3, you are granted additional
17// permissions described in the GCC Runtime Library Exception, version
18// 3.1, as published by the Free Software Foundation.
19
20// You should have received a copy of the GNU General Public License and
21// a copy of the GCC Runtime Library Exception along with this program;
22// see the files COPYING3 and COPYING.RUNTIME respectively.  If not, see
23// <http://www.gnu.org/licenses/>.
24
25/** @file include/stop_token
26 *  This is a Standard C++ Library header.
27 */
28
29#ifndef _GLIBCXX_STOP_TOKEN
30#define _GLIBCXX_STOP_TOKEN
31
32#if __cplusplus > 201703L
33
34#include <atomic>
35#include <bits/std_thread.h>
36
37#include <semaphore>
38
39#define __cpp_lib_jthread 201911L
40
41namespace std _GLIBCXX_VISIBILITY(default)
42{
43_GLIBCXX_BEGIN_NAMESPACE_VERSION
44
45  /// Tag type indicating a stop_source should have no shared-stop-state.
46  struct nostopstate_t { explicit nostopstate_t() = default; };
47  inline constexpr nostopstate_t nostopstate{};
48
49  class stop_source;
50
51  /// Allow testing whether a stop request has been made on a `stop_source`.
52  class stop_token
53  {
54  public:
55    stop_token() noexcept = default;
56
57    stop_token(const stop_token&) noexcept = default;
58    stop_token(stop_token&&) noexcept = default;
59
60    ~stop_token() = default;
61
62    stop_token&
63    operator=(const stop_token&) noexcept = default;
64
65    stop_token&
66    operator=(stop_token&&) noexcept = default;
67
68    [[nodiscard]]
69    bool
70    stop_possible() const noexcept
71    {
72      return static_cast<bool>(_M_state) && _M_state->_M_stop_possible();
73    }
74
75    [[nodiscard]]
76    bool
77    stop_requested() const noexcept
78    {
79      return static_cast<bool>(_M_state) && _M_state->_M_stop_requested();
80    }
81
82    void
83    swap(stop_token& __rhs) noexcept
84    { _M_state.swap(__rhs._M_state); }
85
86    [[nodiscard]]
87    friend bool
88    operator==(const stop_token& __a, const stop_token& __b)
89    { return __a._M_state == __b._M_state; }
90
91    friend void
92    swap(stop_token& __lhs, stop_token& __rhs) noexcept
93    { __lhs.swap(__rhs); }
94
95  private:
96    friend class stop_source;
97    template<typename _Callback>
98      friend class stop_callback;
99
100    static void
101    _S_yield() noexcept
102    {
103#if defined __i386__ || defined __x86_64__
104      __builtin_ia32_pause();
105#endif
106      this_thread::yield();
107    }
108
109#ifndef __cpp_lib_semaphore
110    struct binary_semaphore
111    {
112      explicit binary_semaphore(int __d) : _M_counter(__d > 0) { }
113
114      void release() { _M_counter.fetch_add(1, memory_order::release); }
115
116      void acquire()
117      {
118	int __old = 1;
119	while (!_M_counter.compare_exchange_weak(__old, 0,
120						 memory_order::acquire,
121						 memory_order::relaxed))
122	  {
123	    __old = 1;
124	    _S_yield();
125	  }
126      }
127
128      atomic<int> _M_counter;
129    };
130#endif
131
132    struct _Stop_cb
133    {
134      using __cb_type = void(_Stop_cb*) noexcept;
135      __cb_type* _M_callback;
136      _Stop_cb* _M_prev = nullptr;
137      _Stop_cb* _M_next = nullptr;
138      bool* _M_destroyed = nullptr;
139      binary_semaphore _M_done{0};
140
141      [[__gnu__::__nonnull__]]
142      explicit
143      _Stop_cb(__cb_type* __cb)
144      : _M_callback(__cb)
145      { }
146
147      void _M_run() noexcept { _M_callback(this); }
148    };
149
150    struct _Stop_state_t
151    {
152      using value_type = uint32_t;
153      static constexpr value_type _S_stop_requested_bit = 1;
154      static constexpr value_type _S_locked_bit = 2;
155      static constexpr value_type _S_ssrc_counter_inc = 4;
156
157      std::atomic<value_type> _M_owners{1};
158      std::atomic<value_type> _M_value{_S_ssrc_counter_inc};
159      _Stop_cb* _M_head = nullptr;
160      std::thread::id _M_requester;
161
162      _Stop_state_t() = default;
163
164      bool
165      _M_stop_possible() noexcept
166      {
167	// true if a stop request has already been made or there are still
168	// stop_source objects that would allow one to be made.
169	return _M_value.load(memory_order::acquire) & ~_S_locked_bit;
170      }
171
172      bool
173      _M_stop_requested() noexcept
174      {
175	return _M_value.load(memory_order::acquire) & _S_stop_requested_bit;
176      }
177
178      void
179      _M_add_owner() noexcept
180      {
181	_M_owners.fetch_add(1, memory_order::relaxed);
182      }
183
184      void
185      _M_release_ownership() noexcept
186      {
187	if (_M_owners.fetch_sub(1, memory_order::acq_rel) == 1)
188	  delete this;
189      }
190
191      void
192      _M_add_ssrc() noexcept
193      {
194	_M_value.fetch_add(_S_ssrc_counter_inc, memory_order::relaxed);
195      }
196
197      void
198      _M_sub_ssrc() noexcept
199      {
200	_M_value.fetch_sub(_S_ssrc_counter_inc, memory_order::release);
201      }
202
203      // Obtain lock.
204      void
205      _M_lock() noexcept
206      {
207	// Can use relaxed loads to get the current value.
208	// The successful call to _M_try_lock is an acquire operation.
209	auto __old = _M_value.load(memory_order::relaxed);
210	while (!_M_try_lock(__old, memory_order::relaxed))
211	  { }
212      }
213
214      // Precondition: calling thread holds the lock.
215      void
216      _M_unlock() noexcept
217      {
218	_M_value.fetch_sub(_S_locked_bit, memory_order::release);
219      }
220
221      bool
222      _M_request_stop() noexcept
223      {
224	// obtain lock and set stop_requested bit
225	auto __old = _M_value.load(memory_order::acquire);
226	do
227	  {
228	    if (__old & _S_stop_requested_bit) // stop request already made
229	      return false;
230	  }
231	while (!_M_try_lock_and_stop(__old));
232
233	_M_requester = this_thread::get_id();
234
235	while (_M_head)
236	  {
237	    bool __last_cb;
238	    _Stop_cb* __cb = _M_head;
239	    _M_head = _M_head->_M_next;
240	    if (_M_head)
241	      {
242		_M_head->_M_prev = nullptr;
243		__last_cb = false;
244	      }
245	    else
246	      __last_cb = true;
247
248	    // Allow other callbacks to be unregistered while __cb runs.
249	    _M_unlock();
250
251	    bool __destroyed = false;
252	    __cb->_M_destroyed = &__destroyed;
253
254	    // run callback
255	    __cb->_M_run();
256
257	    if (!__destroyed)
258	      {
259		__cb->_M_destroyed = nullptr;
260
261		// synchronize with destructor of stop_callback that owns *__cb
262		if (!__gnu_cxx::__is_single_threaded())
263		  __cb->_M_done.release();
264	      }
265
266	    // Avoid relocking if we already know there are no more callbacks.
267	    if (__last_cb)
268	      return true;
269
270	    _M_lock();
271	  }
272
273	_M_unlock();
274	return true;
275      }
276
277      [[__gnu__::__nonnull__]]
278      bool
279      _M_register_callback(_Stop_cb* __cb) noexcept
280      {
281	auto __old = _M_value.load(memory_order::acquire);
282	do
283	  {
284	    if (__old & _S_stop_requested_bit) // stop request already made
285	      {
286		__cb->_M_run(); // run synchronously
287		return false;
288	      }
289
290	    if (__old < _S_ssrc_counter_inc) // no stop_source owns *this
291	      // No need to register callback if no stop request can be made.
292	      // Returning false also means the stop_callback does not share
293	      // ownership of this state, but that's not observable.
294	      return false;
295	  }
296	while (!_M_try_lock(__old));
297
298        __cb->_M_next = _M_head;
299        if (_M_head)
300          {
301            _M_head->_M_prev = __cb;
302          }
303        _M_head = __cb;
304	_M_unlock();
305        return true;
306      }
307
308      // Called by ~stop_callback just before destroying *__cb.
309      [[__gnu__::__nonnull__]]
310      void
311      _M_remove_callback(_Stop_cb* __cb)
312      {
313	_M_lock();
314
315        if (__cb == _M_head)
316          {
317            _M_head = _M_head->_M_next;
318            if (_M_head)
319	      _M_head->_M_prev = nullptr;
320	    _M_unlock();
321	    return;
322          }
323	else if (__cb->_M_prev)
324          {
325            __cb->_M_prev->_M_next = __cb->_M_next;
326            if (__cb->_M_next)
327	      __cb->_M_next->_M_prev = __cb->_M_prev;
328	    _M_unlock();
329	    return;
330          }
331
332	_M_unlock();
333
334	// Callback is not in the list, so must have been removed by a call to
335	// _M_request_stop.
336
337	// Despite appearances there is no data race on _M_requester. The only
338	// write to it happens before the callback is removed from the list,
339	// and removing it from the list happens before this read.
340	if (!(_M_requester == this_thread::get_id()))
341	  {
342	    // Synchronize with completion of callback.
343	    __cb->_M_done.acquire();
344	    // Safe for ~stop_callback to destroy *__cb now.
345	    return;
346	  }
347
348	if (__cb->_M_destroyed)
349	  *__cb->_M_destroyed = true;
350      }
351
352      // Try to obtain the lock.
353      // Returns true if the lock is acquired (with memory order acquire).
354      // Otherwise, sets __curval = _M_value.load(__failure) and returns false.
355      // Might fail spuriously, so must be called in a loop.
356      bool
357      _M_try_lock(value_type& __curval,
358		  memory_order __failure = memory_order::acquire) noexcept
359      {
360	return _M_do_try_lock(__curval, 0, memory_order::acquire, __failure);
361      }
362
363      // Try to obtain the lock to make a stop request.
364      // Returns true if the lock is acquired and the _S_stop_requested_bit is
365      // set (with memory order acq_rel so that other threads see the request).
366      // Otherwise, sets __curval = _M_value.load(memory_order::acquire) and
367      // returns false.
368      // Might fail spuriously, so must be called in a loop.
369      bool
370      _M_try_lock_and_stop(value_type& __curval) noexcept
371      {
372	return _M_do_try_lock(__curval, _S_stop_requested_bit,
373			      memory_order::acq_rel, memory_order::acquire);
374      }
375
376      bool
377      _M_do_try_lock(value_type& __curval, value_type __newbits,
378		     memory_order __success, memory_order __failure) noexcept
379      {
380	if (__curval & _S_locked_bit)
381	  {
382	    _S_yield();
383	    __curval = _M_value.load(__failure);
384	    return false;
385	  }
386	__newbits |= _S_locked_bit;
387	return _M_value.compare_exchange_weak(__curval, __curval | __newbits,
388					      __success, __failure);
389      }
390    };
391
392    struct _Stop_state_ref
393    {
394      _Stop_state_ref() = default;
395
396      [[__gnu__::__access__(__none__, 2)]]
397      explicit
398      _Stop_state_ref(const stop_source&)
399      : _M_ptr(new _Stop_state_t())
400      { }
401
402      _Stop_state_ref(const _Stop_state_ref& __other) noexcept
403      : _M_ptr(__other._M_ptr)
404      {
405	if (_M_ptr)
406	  _M_ptr->_M_add_owner();
407      }
408
409      _Stop_state_ref(_Stop_state_ref&& __other) noexcept
410      : _M_ptr(__other._M_ptr)
411      {
412	__other._M_ptr = nullptr;
413      }
414
415      _Stop_state_ref&
416      operator=(const _Stop_state_ref& __other) noexcept
417      {
418	if (auto __ptr = __other._M_ptr; __ptr != _M_ptr)
419	  {
420	    if (__ptr)
421	      __ptr->_M_add_owner();
422	    if (_M_ptr)
423	      _M_ptr->_M_release_ownership();
424	    _M_ptr = __ptr;
425	  }
426	return *this;
427      }
428
429      _Stop_state_ref&
430      operator=(_Stop_state_ref&& __other) noexcept
431      {
432	_Stop_state_ref(std::move(__other)).swap(*this);
433	return *this;
434      }
435
436      ~_Stop_state_ref()
437      {
438	if (_M_ptr)
439	  _M_ptr->_M_release_ownership();
440      }
441
442      void
443      swap(_Stop_state_ref& __other) noexcept
444      { std::swap(_M_ptr, __other._M_ptr); }
445
446      explicit operator bool() const noexcept { return _M_ptr != nullptr; }
447
448      _Stop_state_t* operator->() const noexcept { return _M_ptr; }
449
450#if __cpp_impl_three_way_comparison >= 201907L
451      friend bool
452      operator==(const _Stop_state_ref&, const _Stop_state_ref&) = default;
453#else
454      friend bool
455      operator==(const _Stop_state_ref& __lhs, const _Stop_state_ref& __rhs)
456      noexcept
457      { return __lhs._M_ptr == __rhs._M_ptr; }
458
459      friend bool
460      operator!=(const _Stop_state_ref& __lhs, const _Stop_state_ref& __rhs)
461      noexcept
462      { return __lhs._M_ptr != __rhs._M_ptr; }
463#endif
464
465    private:
466      _Stop_state_t* _M_ptr = nullptr;
467    };
468
469    _Stop_state_ref _M_state;
470
471    explicit
472    stop_token(const _Stop_state_ref& __state) noexcept
473    : _M_state{__state}
474    { }
475  };
476
477  /// A type that allows a stop request to be made.
478  class stop_source
479  {
480  public:
481    stop_source() : _M_state(*this)
482    { }
483
484    explicit stop_source(std::nostopstate_t) noexcept
485    { }
486
487    stop_source(const stop_source& __other) noexcept
488    : _M_state(__other._M_state)
489    {
490      if (_M_state)
491	_M_state->_M_add_ssrc();
492    }
493
494    stop_source(stop_source&&) noexcept = default;
495
496    stop_source&
497    operator=(const stop_source& __other) noexcept
498    {
499      if (_M_state != __other._M_state)
500	{
501	  stop_source __sink(std::move(*this));
502	  _M_state = __other._M_state;
503	  if (_M_state)
504	    _M_state->_M_add_ssrc();
505	}
506      return *this;
507    }
508
509    stop_source&
510    operator=(stop_source&&) noexcept = default;
511
512    ~stop_source()
513    {
514      if (_M_state)
515	_M_state->_M_sub_ssrc();
516    }
517
518    [[nodiscard]]
519    bool
520    stop_possible() const noexcept
521    {
522      return static_cast<bool>(_M_state);
523    }
524
525    [[nodiscard]]
526    bool
527    stop_requested() const noexcept
528    {
529      return static_cast<bool>(_M_state) && _M_state->_M_stop_requested();
530    }
531
532    bool
533    request_stop() const noexcept
534    {
535      if (stop_possible())
536        return _M_state->_M_request_stop();
537      return false;
538    }
539
540    [[nodiscard]]
541    stop_token
542    get_token() const noexcept
543    {
544      return stop_token{_M_state};
545    }
546
547    void
548    swap(stop_source& __other) noexcept
549    {
550      _M_state.swap(__other._M_state);
551    }
552
553    [[nodiscard]]
554    friend bool
555    operator==(const stop_source& __a, const stop_source& __b) noexcept
556    {
557      return __a._M_state == __b._M_state;
558    }
559
560    friend void
561    swap(stop_source& __lhs, stop_source& __rhs) noexcept
562    {
563      __lhs.swap(__rhs);
564    }
565
566  private:
567    stop_token::_Stop_state_ref _M_state;
568  };
569
570  /// A wrapper for callbacks to be run when a stop request is made.
571  template<typename _Callback>
572    class [[nodiscard]] stop_callback
573    {
574      static_assert(is_nothrow_destructible_v<_Callback>);
575      static_assert(is_invocable_v<_Callback>);
576
577    public:
578      using callback_type = _Callback;
579
580      template<typename _Cb,
581               enable_if_t<is_constructible_v<_Callback, _Cb>, int> = 0>
582        explicit
583	stop_callback(const stop_token& __token, _Cb&& __cb)
584        noexcept(is_nothrow_constructible_v<_Callback, _Cb>)
585	: _M_cb(std::forward<_Cb>(__cb))
586        {
587	  if (auto __state = __token._M_state)
588	    {
589	      if (__state->_M_register_callback(&_M_cb))
590		_M_state.swap(__state);
591	    }
592        }
593
594      template<typename _Cb,
595               enable_if_t<is_constructible_v<_Callback, _Cb>, int> = 0>
596        explicit
597	stop_callback(stop_token&& __token, _Cb&& __cb)
598        noexcept(is_nothrow_constructible_v<_Callback, _Cb>)
599	: _M_cb(std::forward<_Cb>(__cb))
600	{
601	  if (auto& __state = __token._M_state)
602	    {
603	      if (__state->_M_register_callback(&_M_cb))
604		_M_state.swap(__state);
605	    }
606	}
607
608      ~stop_callback()
609      {
610	if (_M_state)
611	  {
612	    _M_state->_M_remove_callback(&_M_cb);
613	  }
614      }
615
616      stop_callback(const stop_callback&) = delete;
617      stop_callback& operator=(const stop_callback&) = delete;
618      stop_callback(stop_callback&&) = delete;
619      stop_callback& operator=(stop_callback&&) = delete;
620
621    private:
622      struct _Cb_impl : stop_token::_Stop_cb
623      {
624	template<typename _Cb>
625	  explicit
626	  _Cb_impl(_Cb&& __cb)
627	  : _Stop_cb(&_S_execute),
628	    _M_cb(std::forward<_Cb>(__cb))
629	  { }
630
631	_Callback _M_cb;
632
633	[[__gnu__::__nonnull__]]
634	static void
635	_S_execute(_Stop_cb* __that) noexcept
636	{
637	  _Callback& __cb = static_cast<_Cb_impl*>(__that)->_M_cb;
638	  std::forward<_Callback>(__cb)();
639	}
640      };
641
642      _Cb_impl _M_cb;
643      stop_token::_Stop_state_ref _M_state;
644    };
645
646  template<typename _Callback>
647    stop_callback(stop_token, _Callback) -> stop_callback<_Callback>;
648
649_GLIBCXX_END_NAMESPACE_VERSION
650} // namespace
651#endif // __cplusplus > 201703L
652#endif // _GLIBCXX_STOP_TOKEN
653