xref: /dflybsd-src/contrib/gcc-8.0/libstdc++-v3/include/ext/mt_allocator.h (revision 38fd149817dfbff97799f62fcb70be98c4e32523)
1*38fd1498Szrj // MT-optimized allocator -*- C++ -*-
2*38fd1498Szrj 
3*38fd1498Szrj // Copyright (C) 2003-2018 Free Software Foundation, Inc.
4*38fd1498Szrj //
5*38fd1498Szrj // This file is part of the GNU ISO C++ Library.  This library is free
6*38fd1498Szrj // software; you can redistribute it and/or modify it under the
7*38fd1498Szrj // terms of the GNU General Public License as published by the
8*38fd1498Szrj // Free Software Foundation; either version 3, or (at your option)
9*38fd1498Szrj // any later version.
10*38fd1498Szrj 
11*38fd1498Szrj // This library is distributed in the hope that it will be useful,
12*38fd1498Szrj // but WITHOUT ANY WARRANTY; without even the implied warranty of
13*38fd1498Szrj // MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
14*38fd1498Szrj // GNU General Public License for more details.
15*38fd1498Szrj 
16*38fd1498Szrj // Under Section 7 of GPL version 3, you are granted additional
17*38fd1498Szrj // permissions described in the GCC Runtime Library Exception, version
18*38fd1498Szrj // 3.1, as published by the Free Software Foundation.
19*38fd1498Szrj 
20*38fd1498Szrj // You should have received a copy of the GNU General Public License and
21*38fd1498Szrj // a copy of the GCC Runtime Library Exception along with this program;
22*38fd1498Szrj // see the files COPYING3 and COPYING.RUNTIME respectively.  If not, see
23*38fd1498Szrj // <http://www.gnu.org/licenses/>.
24*38fd1498Szrj 
25*38fd1498Szrj /** @file ext/mt_allocator.h
26*38fd1498Szrj  *  This file is a GNU extension to the Standard C++ Library.
27*38fd1498Szrj  */
28*38fd1498Szrj 
29*38fd1498Szrj #ifndef _MT_ALLOCATOR_H
30*38fd1498Szrj #define _MT_ALLOCATOR_H 1
31*38fd1498Szrj 
32*38fd1498Szrj #include <new>
33*38fd1498Szrj #include <cstdlib>
34*38fd1498Szrj #include <bits/functexcept.h>
35*38fd1498Szrj #include <ext/atomicity.h>
36*38fd1498Szrj #include <bits/move.h>
37*38fd1498Szrj #if __cplusplus >= 201103L
38*38fd1498Szrj #include <type_traits>
39*38fd1498Szrj #endif
40*38fd1498Szrj 
_GLIBCXX_VISIBILITY(default)41*38fd1498Szrj namespace __gnu_cxx _GLIBCXX_VISIBILITY(default)
42*38fd1498Szrj {
43*38fd1498Szrj _GLIBCXX_BEGIN_NAMESPACE_VERSION
44*38fd1498Szrj 
45*38fd1498Szrj   using std::size_t;
46*38fd1498Szrj   using std::ptrdiff_t;
47*38fd1498Szrj 
48*38fd1498Szrj   typedef void (*__destroy_handler)(void*);
49*38fd1498Szrj 
50*38fd1498Szrj   /// Base class for pool object.
51*38fd1498Szrj   struct __pool_base
52*38fd1498Szrj   {
53*38fd1498Szrj     // Using short int as type for the binmap implies we are never
54*38fd1498Szrj     // caching blocks larger than 32768 with this allocator.
55*38fd1498Szrj     typedef unsigned short int _Binmap_type;
56*38fd1498Szrj 
57*38fd1498Szrj     // Variables used to configure the behavior of the allocator,
58*38fd1498Szrj     // assigned and explained in detail below.
59*38fd1498Szrj     struct _Tune
60*38fd1498Szrj      {
61*38fd1498Szrj       // Compile time constants for the default _Tune values.
62*38fd1498Szrj       enum { _S_align = 8 };
63*38fd1498Szrj       enum { _S_max_bytes = 128 };
64*38fd1498Szrj       enum { _S_min_bin = 8 };
65*38fd1498Szrj       enum { _S_chunk_size = 4096 - 4 * sizeof(void*) };
66*38fd1498Szrj       enum { _S_max_threads = 4096 };
67*38fd1498Szrj       enum { _S_freelist_headroom = 10 };
68*38fd1498Szrj 
69*38fd1498Szrj       // Alignment needed.
70*38fd1498Szrj       // NB: In any case must be >= sizeof(_Block_record), that
71*38fd1498Szrj       // is 4 on 32 bit machines and 8 on 64 bit machines.
72*38fd1498Szrj       size_t	_M_align;
73*38fd1498Szrj 
74*38fd1498Szrj       // Allocation requests (after round-up to power of 2) below
75*38fd1498Szrj       // this value will be handled by the allocator. A raw new/
76*38fd1498Szrj       // call will be used for requests larger than this value.
77*38fd1498Szrj       // NB: Must be much smaller than _M_chunk_size and in any
78*38fd1498Szrj       // case <= 32768.
79*38fd1498Szrj       size_t	_M_max_bytes;
80*38fd1498Szrj 
81*38fd1498Szrj       // Size in bytes of the smallest bin.
82*38fd1498Szrj       // NB: Must be a power of 2 and >= _M_align (and of course
83*38fd1498Szrj       // much smaller than _M_max_bytes).
84*38fd1498Szrj       size_t	_M_min_bin;
85*38fd1498Szrj 
86*38fd1498Szrj       // In order to avoid fragmenting and minimize the number of
87*38fd1498Szrj       // new() calls we always request new memory using this
88*38fd1498Szrj       // value. Based on previous discussions on the libstdc++
89*38fd1498Szrj       // mailing list we have chosen the value below.
90*38fd1498Szrj       // See http://gcc.gnu.org/ml/libstdc++/2001-07/msg00077.html
91*38fd1498Szrj       // NB: At least one order of magnitude > _M_max_bytes.
92*38fd1498Szrj       size_t	_M_chunk_size;
93*38fd1498Szrj 
94*38fd1498Szrj       // The maximum number of supported threads. For
95*38fd1498Szrj       // single-threaded operation, use one. Maximum values will
96*38fd1498Szrj       // vary depending on details of the underlying system. (For
97*38fd1498Szrj       // instance, Linux 2.4.18 reports 4070 in
98*38fd1498Szrj       // /proc/sys/kernel/threads-max, while Linux 2.6.6 reports
99*38fd1498Szrj       // 65534)
100*38fd1498Szrj       size_t 	_M_max_threads;
101*38fd1498Szrj 
102*38fd1498Szrj       // Each time a deallocation occurs in a threaded application
103*38fd1498Szrj       // we make sure that there are no more than
104*38fd1498Szrj       // _M_freelist_headroom % of used memory on the freelist. If
105*38fd1498Szrj       // the number of additional records is more than
106*38fd1498Szrj       // _M_freelist_headroom % of the freelist, we move these
107*38fd1498Szrj       // records back to the global pool.
108*38fd1498Szrj       size_t 	_M_freelist_headroom;
109*38fd1498Szrj 
110*38fd1498Szrj       // Set to true forces all allocations to use new().
111*38fd1498Szrj       bool 	_M_force_new;
112*38fd1498Szrj 
113*38fd1498Szrj       explicit
114*38fd1498Szrj       _Tune()
115*38fd1498Szrj       : _M_align(_S_align), _M_max_bytes(_S_max_bytes), _M_min_bin(_S_min_bin),
116*38fd1498Szrj       _M_chunk_size(_S_chunk_size), _M_max_threads(_S_max_threads),
117*38fd1498Szrj       _M_freelist_headroom(_S_freelist_headroom),
118*38fd1498Szrj       _M_force_new(std::getenv("GLIBCXX_FORCE_NEW") ? true : false)
119*38fd1498Szrj       { }
120*38fd1498Szrj 
121*38fd1498Szrj       explicit
122*38fd1498Szrj       _Tune(size_t __align, size_t __maxb, size_t __minbin, size_t __chunk,
123*38fd1498Szrj 	    size_t __maxthreads, size_t __headroom, bool __force)
124*38fd1498Szrj       : _M_align(__align), _M_max_bytes(__maxb), _M_min_bin(__minbin),
125*38fd1498Szrj       _M_chunk_size(__chunk), _M_max_threads(__maxthreads),
126*38fd1498Szrj       _M_freelist_headroom(__headroom), _M_force_new(__force)
127*38fd1498Szrj       { }
128*38fd1498Szrj     };
129*38fd1498Szrj 
130*38fd1498Szrj     struct _Block_address
131*38fd1498Szrj     {
132*38fd1498Szrj       void* 			_M_initial;
133*38fd1498Szrj       _Block_address* 		_M_next;
134*38fd1498Szrj     };
135*38fd1498Szrj 
136*38fd1498Szrj     const _Tune&
137*38fd1498Szrj     _M_get_options() const
138*38fd1498Szrj     { return _M_options; }
139*38fd1498Szrj 
140*38fd1498Szrj     void
141*38fd1498Szrj     _M_set_options(_Tune __t)
142*38fd1498Szrj     {
143*38fd1498Szrj       if (!_M_init)
144*38fd1498Szrj 	_M_options = __t;
145*38fd1498Szrj     }
146*38fd1498Szrj 
147*38fd1498Szrj     bool
148*38fd1498Szrj     _M_check_threshold(size_t __bytes)
149*38fd1498Szrj     { return __bytes > _M_options._M_max_bytes || _M_options._M_force_new; }
150*38fd1498Szrj 
151*38fd1498Szrj     size_t
152*38fd1498Szrj     _M_get_binmap(size_t __bytes)
153*38fd1498Szrj     { return _M_binmap[__bytes]; }
154*38fd1498Szrj 
155*38fd1498Szrj     size_t
156*38fd1498Szrj     _M_get_align()
157*38fd1498Szrj     { return _M_options._M_align; }
158*38fd1498Szrj 
159*38fd1498Szrj     explicit
160*38fd1498Szrj     __pool_base()
161*38fd1498Szrj     : _M_options(_Tune()), _M_binmap(0), _M_init(false) { }
162*38fd1498Szrj 
163*38fd1498Szrj     explicit
164*38fd1498Szrj     __pool_base(const _Tune& __options)
165*38fd1498Szrj     : _M_options(__options), _M_binmap(0), _M_init(false) { }
166*38fd1498Szrj 
167*38fd1498Szrj   private:
168*38fd1498Szrj     explicit
169*38fd1498Szrj     __pool_base(const __pool_base&);
170*38fd1498Szrj 
171*38fd1498Szrj     __pool_base&
172*38fd1498Szrj     operator=(const __pool_base&);
173*38fd1498Szrj 
174*38fd1498Szrj   protected:
175*38fd1498Szrj     // Configuration options.
176*38fd1498Szrj     _Tune 	       		_M_options;
177*38fd1498Szrj 
178*38fd1498Szrj     _Binmap_type* 		_M_binmap;
179*38fd1498Szrj 
180*38fd1498Szrj     // Configuration of the pool object via _M_options can happen
181*38fd1498Szrj     // after construction but before initialization. After
182*38fd1498Szrj     // initialization is complete, this variable is set to true.
183*38fd1498Szrj     bool 			_M_init;
184*38fd1498Szrj   };
185*38fd1498Szrj 
186*38fd1498Szrj 
187*38fd1498Szrj   /**
188*38fd1498Szrj    *  @brief  Data describing the underlying memory pool, parameterized on
189*38fd1498Szrj    *  threading support.
190*38fd1498Szrj    */
191*38fd1498Szrj   template<bool _Thread>
192*38fd1498Szrj     class __pool;
193*38fd1498Szrj 
194*38fd1498Szrj   /// Specialization for single thread.
195*38fd1498Szrj   template<>
196*38fd1498Szrj     class __pool<false> : public __pool_base
197*38fd1498Szrj     {
198*38fd1498Szrj     public:
199*38fd1498Szrj       union _Block_record
200*38fd1498Szrj       {
201*38fd1498Szrj 	// Points to the block_record of the next free block.
202*38fd1498Szrj 	_Block_record* 			_M_next;
203*38fd1498Szrj       };
204*38fd1498Szrj 
205*38fd1498Szrj       struct _Bin_record
206*38fd1498Szrj       {
207*38fd1498Szrj 	// An "array" of pointers to the first free block.
208*38fd1498Szrj 	_Block_record**			_M_first;
209*38fd1498Szrj 
210*38fd1498Szrj 	// A list of the initial addresses of all allocated blocks.
211*38fd1498Szrj 	_Block_address*		     	_M_address;
212*38fd1498Szrj       };
213*38fd1498Szrj 
214*38fd1498Szrj       void
215*38fd1498Szrj       _M_initialize_once()
216*38fd1498Szrj       {
217*38fd1498Szrj 	if (__builtin_expect(_M_init == false, false))
218*38fd1498Szrj 	  _M_initialize();
219*38fd1498Szrj       }
220*38fd1498Szrj 
221*38fd1498Szrj       void
222*38fd1498Szrj       _M_destroy() throw();
223*38fd1498Szrj 
224*38fd1498Szrj       char*
225*38fd1498Szrj       _M_reserve_block(size_t __bytes, const size_t __thread_id);
226*38fd1498Szrj 
227*38fd1498Szrj       void
228*38fd1498Szrj       _M_reclaim_block(char* __p, size_t __bytes) throw ();
229*38fd1498Szrj 
230*38fd1498Szrj       size_t
231*38fd1498Szrj       _M_get_thread_id() { return 0; }
232*38fd1498Szrj 
233*38fd1498Szrj       const _Bin_record&
234*38fd1498Szrj       _M_get_bin(size_t __which)
235*38fd1498Szrj       { return _M_bin[__which]; }
236*38fd1498Szrj 
237*38fd1498Szrj       void
238*38fd1498Szrj       _M_adjust_freelist(const _Bin_record&, _Block_record*, size_t)
239*38fd1498Szrj       { }
240*38fd1498Szrj 
241*38fd1498Szrj       explicit __pool()
242*38fd1498Szrj       : _M_bin(0), _M_bin_size(1) { }
243*38fd1498Szrj 
244*38fd1498Szrj       explicit __pool(const __pool_base::_Tune& __tune)
245*38fd1498Szrj       : __pool_base(__tune), _M_bin(0), _M_bin_size(1) { }
246*38fd1498Szrj 
247*38fd1498Szrj     private:
248*38fd1498Szrj       // An "array" of bin_records each of which represents a specific
249*38fd1498Szrj       // power of 2 size. Memory to this "array" is allocated in
250*38fd1498Szrj       // _M_initialize().
251*38fd1498Szrj       _Bin_record*		 _M_bin;
252*38fd1498Szrj 
253*38fd1498Szrj       // Actual value calculated in _M_initialize().
254*38fd1498Szrj       size_t 	       	     	_M_bin_size;
255*38fd1498Szrj 
256*38fd1498Szrj       void
257*38fd1498Szrj       _M_initialize();
258*38fd1498Szrj   };
259*38fd1498Szrj 
260*38fd1498Szrj #ifdef __GTHREADS
261*38fd1498Szrj   /// Specialization for thread enabled, via gthreads.h.
262*38fd1498Szrj   template<>
263*38fd1498Szrj     class __pool<true> : public __pool_base
264*38fd1498Szrj     {
265*38fd1498Szrj     public:
266*38fd1498Szrj       // Each requesting thread is assigned an id ranging from 1 to
267*38fd1498Szrj       // _S_max_threads. Thread id 0 is used as a global memory pool.
268*38fd1498Szrj       // In order to get constant performance on the thread assignment
269*38fd1498Szrj       // routine, we keep a list of free ids. When a thread first
270*38fd1498Szrj       // requests memory we remove the first record in this list and
271*38fd1498Szrj       // stores the address in a __gthread_key. When initializing the
272*38fd1498Szrj       // __gthread_key we specify a destructor. When this destructor
273*38fd1498Szrj       // (i.e. the thread dies) is called, we return the thread id to
274*38fd1498Szrj       // the front of this list.
275*38fd1498Szrj       struct _Thread_record
276*38fd1498Szrj       {
277*38fd1498Szrj 	// Points to next free thread id record. NULL if last record in list.
278*38fd1498Szrj 	_Thread_record*			_M_next;
279*38fd1498Szrj 
280*38fd1498Szrj 	// Thread id ranging from 1 to _S_max_threads.
281*38fd1498Szrj 	size_t                          _M_id;
282*38fd1498Szrj       };
283*38fd1498Szrj 
284*38fd1498Szrj       union _Block_record
285*38fd1498Szrj       {
286*38fd1498Szrj 	// Points to the block_record of the next free block.
287*38fd1498Szrj 	_Block_record*			_M_next;
288*38fd1498Szrj 
289*38fd1498Szrj 	// The thread id of the thread which has requested this block.
290*38fd1498Szrj 	size_t                          _M_thread_id;
291*38fd1498Szrj       };
292*38fd1498Szrj 
293*38fd1498Szrj       struct _Bin_record
294*38fd1498Szrj       {
295*38fd1498Szrj 	// An "array" of pointers to the first free block for each
296*38fd1498Szrj 	// thread id. Memory to this "array" is allocated in
297*38fd1498Szrj 	// _S_initialize() for _S_max_threads + global pool 0.
298*38fd1498Szrj 	_Block_record**			_M_first;
299*38fd1498Szrj 
300*38fd1498Szrj 	// A list of the initial addresses of all allocated blocks.
301*38fd1498Szrj 	_Block_address*		     	_M_address;
302*38fd1498Szrj 
303*38fd1498Szrj 	// An "array" of counters used to keep track of the amount of
304*38fd1498Szrj 	// blocks that are on the freelist/used for each thread id.
305*38fd1498Szrj 	// - Note that the second part of the allocated _M_used "array"
306*38fd1498Szrj 	//   actually hosts (atomic) counters of reclaimed blocks:  in
307*38fd1498Szrj 	//   _M_reserve_block and in _M_reclaim_block those numbers are
308*38fd1498Szrj 	//   subtracted from the first ones to obtain the actual size
309*38fd1498Szrj 	//   of the "working set" of the given thread.
310*38fd1498Szrj 	// - Memory to these "arrays" is allocated in _S_initialize()
311*38fd1498Szrj 	//   for _S_max_threads + global pool 0.
312*38fd1498Szrj 	size_t*				_M_free;
313*38fd1498Szrj 	size_t*			        _M_used;
314*38fd1498Szrj 
315*38fd1498Szrj 	// Each bin has its own mutex which is used to ensure data
316*38fd1498Szrj 	// integrity while changing "ownership" on a block.  The mutex
317*38fd1498Szrj 	// is initialized in _S_initialize().
318*38fd1498Szrj 	__gthread_mutex_t*              _M_mutex;
319*38fd1498Szrj       };
320*38fd1498Szrj 
321*38fd1498Szrj       // XXX GLIBCXX_ABI Deprecated
322*38fd1498Szrj       void
323*38fd1498Szrj       _M_initialize(__destroy_handler);
324*38fd1498Szrj 
325*38fd1498Szrj       void
326*38fd1498Szrj       _M_initialize_once()
327*38fd1498Szrj       {
328*38fd1498Szrj 	if (__builtin_expect(_M_init == false, false))
329*38fd1498Szrj 	  _M_initialize();
330*38fd1498Szrj       }
331*38fd1498Szrj 
332*38fd1498Szrj       void
333*38fd1498Szrj       _M_destroy() throw();
334*38fd1498Szrj 
335*38fd1498Szrj       char*
336*38fd1498Szrj       _M_reserve_block(size_t __bytes, const size_t __thread_id);
337*38fd1498Szrj 
338*38fd1498Szrj       void
339*38fd1498Szrj       _M_reclaim_block(char* __p, size_t __bytes) throw ();
340*38fd1498Szrj 
341*38fd1498Szrj       const _Bin_record&
342*38fd1498Szrj       _M_get_bin(size_t __which)
343*38fd1498Szrj       { return _M_bin[__which]; }
344*38fd1498Szrj 
345*38fd1498Szrj       void
346*38fd1498Szrj       _M_adjust_freelist(const _Bin_record& __bin, _Block_record* __block,
347*38fd1498Szrj 			 size_t __thread_id)
348*38fd1498Szrj       {
349*38fd1498Szrj 	if (__gthread_active_p())
350*38fd1498Szrj 	  {
351*38fd1498Szrj 	    __block->_M_thread_id = __thread_id;
352*38fd1498Szrj 	    --__bin._M_free[__thread_id];
353*38fd1498Szrj 	    ++__bin._M_used[__thread_id];
354*38fd1498Szrj 	  }
355*38fd1498Szrj       }
356*38fd1498Szrj 
357*38fd1498Szrj       // XXX GLIBCXX_ABI Deprecated
358*38fd1498Szrj       void
359*38fd1498Szrj       _M_destroy_thread_key(void*) throw ();
360*38fd1498Szrj 
361*38fd1498Szrj       size_t
362*38fd1498Szrj       _M_get_thread_id();
363*38fd1498Szrj 
364*38fd1498Szrj       explicit __pool()
365*38fd1498Szrj       : _M_bin(0), _M_bin_size(1), _M_thread_freelist(0)
366*38fd1498Szrj       { }
367*38fd1498Szrj 
368*38fd1498Szrj       explicit __pool(const __pool_base::_Tune& __tune)
369*38fd1498Szrj       : __pool_base(__tune), _M_bin(0), _M_bin_size(1),
370*38fd1498Szrj 	_M_thread_freelist(0)
371*38fd1498Szrj       { }
372*38fd1498Szrj 
373*38fd1498Szrj     private:
374*38fd1498Szrj       // An "array" of bin_records each of which represents a specific
375*38fd1498Szrj       // power of 2 size. Memory to this "array" is allocated in
376*38fd1498Szrj       // _M_initialize().
377*38fd1498Szrj       _Bin_record*		_M_bin;
378*38fd1498Szrj 
379*38fd1498Szrj       // Actual value calculated in _M_initialize().
380*38fd1498Szrj       size_t 	       	     	_M_bin_size;
381*38fd1498Szrj 
382*38fd1498Szrj       _Thread_record* 		_M_thread_freelist;
383*38fd1498Szrj       void*			_M_thread_freelist_initial;
384*38fd1498Szrj 
385*38fd1498Szrj       void
386*38fd1498Szrj       _M_initialize();
387*38fd1498Szrj     };
388*38fd1498Szrj #endif
389*38fd1498Szrj 
390*38fd1498Szrj   template<template <bool> class _PoolTp, bool _Thread>
391*38fd1498Szrj     struct __common_pool
392*38fd1498Szrj     {
393*38fd1498Szrj       typedef _PoolTp<_Thread> 		pool_type;
394*38fd1498Szrj 
395*38fd1498Szrj       static pool_type&
396*38fd1498Szrj       _S_get_pool()
397*38fd1498Szrj       {
398*38fd1498Szrj 	static pool_type _S_pool;
399*38fd1498Szrj 	return _S_pool;
400*38fd1498Szrj       }
401*38fd1498Szrj     };
402*38fd1498Szrj 
403*38fd1498Szrj   template<template <bool> class _PoolTp, bool _Thread>
404*38fd1498Szrj     struct __common_pool_base;
405*38fd1498Szrj 
406*38fd1498Szrj   template<template <bool> class _PoolTp>
407*38fd1498Szrj     struct __common_pool_base<_PoolTp, false>
408*38fd1498Szrj     : public __common_pool<_PoolTp, false>
409*38fd1498Szrj     {
410*38fd1498Szrj       using  __common_pool<_PoolTp, false>::_S_get_pool;
411*38fd1498Szrj 
412*38fd1498Szrj       static void
413*38fd1498Szrj       _S_initialize_once()
414*38fd1498Szrj       {
415*38fd1498Szrj 	static bool __init;
416*38fd1498Szrj 	if (__builtin_expect(__init == false, false))
417*38fd1498Szrj 	  {
418*38fd1498Szrj 	    _S_get_pool()._M_initialize_once();
419*38fd1498Szrj 	    __init = true;
420*38fd1498Szrj 	  }
421*38fd1498Szrj       }
422*38fd1498Szrj     };
423*38fd1498Szrj 
424*38fd1498Szrj #ifdef __GTHREADS
425*38fd1498Szrj   template<template <bool> class _PoolTp>
426*38fd1498Szrj     struct __common_pool_base<_PoolTp, true>
427*38fd1498Szrj     : public __common_pool<_PoolTp, true>
428*38fd1498Szrj     {
429*38fd1498Szrj       using  __common_pool<_PoolTp, true>::_S_get_pool;
430*38fd1498Szrj 
431*38fd1498Szrj       static void
432*38fd1498Szrj       _S_initialize()
433*38fd1498Szrj       { _S_get_pool()._M_initialize_once(); }
434*38fd1498Szrj 
435*38fd1498Szrj       static void
436*38fd1498Szrj       _S_initialize_once()
437*38fd1498Szrj       {
438*38fd1498Szrj 	static bool __init;
439*38fd1498Szrj 	if (__builtin_expect(__init == false, false))
440*38fd1498Szrj 	  {
441*38fd1498Szrj 	    if (__gthread_active_p())
442*38fd1498Szrj 	      {
443*38fd1498Szrj 		// On some platforms, __gthread_once_t is an aggregate.
444*38fd1498Szrj 		static __gthread_once_t __once = __GTHREAD_ONCE_INIT;
445*38fd1498Szrj 		__gthread_once(&__once, _S_initialize);
446*38fd1498Szrj 	      }
447*38fd1498Szrj 
448*38fd1498Szrj 	    // Double check initialization. May be necessary on some
449*38fd1498Szrj 	    // systems for proper construction when not compiling with
450*38fd1498Szrj 	    // thread flags.
451*38fd1498Szrj 	    _S_get_pool()._M_initialize_once();
452*38fd1498Szrj 	    __init = true;
453*38fd1498Szrj 	  }
454*38fd1498Szrj       }
455*38fd1498Szrj     };
456*38fd1498Szrj #endif
457*38fd1498Szrj 
458*38fd1498Szrj   /// Policy for shared __pool objects.
459*38fd1498Szrj   template<template <bool> class _PoolTp, bool _Thread>
460*38fd1498Szrj     struct __common_pool_policy : public __common_pool_base<_PoolTp, _Thread>
461*38fd1498Szrj     {
462*38fd1498Szrj       template<typename _Tp1, template <bool> class _PoolTp1 = _PoolTp,
463*38fd1498Szrj 	       bool _Thread1 = _Thread>
464*38fd1498Szrj         struct _M_rebind
465*38fd1498Szrj         { typedef __common_pool_policy<_PoolTp1, _Thread1> other; };
466*38fd1498Szrj 
467*38fd1498Szrj       using  __common_pool_base<_PoolTp, _Thread>::_S_get_pool;
468*38fd1498Szrj       using  __common_pool_base<_PoolTp, _Thread>::_S_initialize_once;
469*38fd1498Szrj   };
470*38fd1498Szrj 
471*38fd1498Szrj 
472*38fd1498Szrj   template<typename _Tp, template <bool> class _PoolTp, bool _Thread>
473*38fd1498Szrj     struct __per_type_pool
474*38fd1498Szrj     {
475*38fd1498Szrj       typedef _Tp 			value_type;
476*38fd1498Szrj       typedef _PoolTp<_Thread> 		pool_type;
477*38fd1498Szrj 
478*38fd1498Szrj       static pool_type&
479*38fd1498Szrj       _S_get_pool()
480*38fd1498Szrj       {
481*38fd1498Szrj 	// Sane defaults for the _PoolTp.
482*38fd1498Szrj 	typedef typename pool_type::_Block_record _Block_record;
483*38fd1498Szrj 	const static size_t __a = (__alignof__(_Tp) >= sizeof(_Block_record)
484*38fd1498Szrj 				   ? __alignof__(_Tp) : sizeof(_Block_record));
485*38fd1498Szrj 
486*38fd1498Szrj 	typedef typename __pool_base::_Tune _Tune;
487*38fd1498Szrj 	static _Tune _S_tune(__a, sizeof(_Tp) * 64,
488*38fd1498Szrj 			     sizeof(_Tp) * 2 >= __a ? sizeof(_Tp) * 2 : __a,
489*38fd1498Szrj 			     sizeof(_Tp) * size_t(_Tune::_S_chunk_size),
490*38fd1498Szrj 			     _Tune::_S_max_threads,
491*38fd1498Szrj 			     _Tune::_S_freelist_headroom,
492*38fd1498Szrj 			     std::getenv("GLIBCXX_FORCE_NEW") ? true : false);
493*38fd1498Szrj 	static pool_type _S_pool(_S_tune);
494*38fd1498Szrj 	return _S_pool;
495*38fd1498Szrj       }
496*38fd1498Szrj     };
497*38fd1498Szrj 
498*38fd1498Szrj   template<typename _Tp, template <bool> class _PoolTp, bool _Thread>
499*38fd1498Szrj     struct __per_type_pool_base;
500*38fd1498Szrj 
501*38fd1498Szrj   template<typename _Tp, template <bool> class _PoolTp>
502*38fd1498Szrj     struct __per_type_pool_base<_Tp, _PoolTp, false>
503*38fd1498Szrj     : public __per_type_pool<_Tp, _PoolTp, false>
504*38fd1498Szrj     {
505*38fd1498Szrj       using  __per_type_pool<_Tp, _PoolTp, false>::_S_get_pool;
506*38fd1498Szrj 
507*38fd1498Szrj       static void
508*38fd1498Szrj       _S_initialize_once()
509*38fd1498Szrj       {
510*38fd1498Szrj 	static bool __init;
511*38fd1498Szrj 	if (__builtin_expect(__init == false, false))
512*38fd1498Szrj 	  {
513*38fd1498Szrj 	    _S_get_pool()._M_initialize_once();
514*38fd1498Szrj 	    __init = true;
515*38fd1498Szrj 	  }
516*38fd1498Szrj       }
517*38fd1498Szrj     };
518*38fd1498Szrj 
519*38fd1498Szrj  #ifdef __GTHREADS
520*38fd1498Szrj  template<typename _Tp, template <bool> class _PoolTp>
521*38fd1498Szrj     struct __per_type_pool_base<_Tp, _PoolTp, true>
522*38fd1498Szrj     : public __per_type_pool<_Tp, _PoolTp, true>
523*38fd1498Szrj     {
524*38fd1498Szrj       using  __per_type_pool<_Tp, _PoolTp, true>::_S_get_pool;
525*38fd1498Szrj 
526*38fd1498Szrj       static void
527*38fd1498Szrj       _S_initialize()
528*38fd1498Szrj       { _S_get_pool()._M_initialize_once(); }
529*38fd1498Szrj 
530*38fd1498Szrj       static void
531*38fd1498Szrj       _S_initialize_once()
532*38fd1498Szrj       {
533*38fd1498Szrj 	static bool __init;
534*38fd1498Szrj 	if (__builtin_expect(__init == false, false))
535*38fd1498Szrj 	  {
536*38fd1498Szrj 	    if (__gthread_active_p())
537*38fd1498Szrj 	      {
538*38fd1498Szrj 		// On some platforms, __gthread_once_t is an aggregate.
539*38fd1498Szrj 		static __gthread_once_t __once = __GTHREAD_ONCE_INIT;
540*38fd1498Szrj 		__gthread_once(&__once, _S_initialize);
541*38fd1498Szrj 	      }
542*38fd1498Szrj 
543*38fd1498Szrj 	    // Double check initialization. May be necessary on some
544*38fd1498Szrj 	    // systems for proper construction when not compiling with
545*38fd1498Szrj 	    // thread flags.
546*38fd1498Szrj 	    _S_get_pool()._M_initialize_once();
547*38fd1498Szrj 	    __init = true;
548*38fd1498Szrj 	  }
549*38fd1498Szrj       }
550*38fd1498Szrj     };
551*38fd1498Szrj #endif
552*38fd1498Szrj 
553*38fd1498Szrj   /// Policy for individual __pool objects.
554*38fd1498Szrj   template<typename _Tp, template <bool> class _PoolTp, bool _Thread>
555*38fd1498Szrj     struct __per_type_pool_policy
556*38fd1498Szrj     : public __per_type_pool_base<_Tp, _PoolTp, _Thread>
557*38fd1498Szrj     {
558*38fd1498Szrj       template<typename _Tp1, template <bool> class _PoolTp1 = _PoolTp,
559*38fd1498Szrj 	       bool _Thread1 = _Thread>
560*38fd1498Szrj         struct _M_rebind
561*38fd1498Szrj         { typedef __per_type_pool_policy<_Tp1, _PoolTp1, _Thread1> other; };
562*38fd1498Szrj 
563*38fd1498Szrj       using  __per_type_pool_base<_Tp, _PoolTp, _Thread>::_S_get_pool;
564*38fd1498Szrj       using  __per_type_pool_base<_Tp, _PoolTp, _Thread>::_S_initialize_once;
565*38fd1498Szrj   };
566*38fd1498Szrj 
567*38fd1498Szrj 
568*38fd1498Szrj   /// Base class for _Tp dependent member functions.
569*38fd1498Szrj   template<typename _Tp>
570*38fd1498Szrj     class __mt_alloc_base
571*38fd1498Szrj     {
572*38fd1498Szrj     public:
573*38fd1498Szrj       typedef size_t                    size_type;
574*38fd1498Szrj       typedef ptrdiff_t                 difference_type;
575*38fd1498Szrj       typedef _Tp*                      pointer;
576*38fd1498Szrj       typedef const _Tp*                const_pointer;
577*38fd1498Szrj       typedef _Tp&                      reference;
578*38fd1498Szrj       typedef const _Tp&                const_reference;
579*38fd1498Szrj       typedef _Tp                       value_type;
580*38fd1498Szrj 
581*38fd1498Szrj #if __cplusplus >= 201103L
582*38fd1498Szrj       // _GLIBCXX_RESOLVE_LIB_DEFECTS
583*38fd1498Szrj       // 2103. propagate_on_container_move_assignment
584*38fd1498Szrj       typedef std::true_type propagate_on_container_move_assignment;
585*38fd1498Szrj #endif
586*38fd1498Szrj 
587*38fd1498Szrj       pointer
588*38fd1498Szrj       address(reference __x) const _GLIBCXX_NOEXCEPT
589*38fd1498Szrj       { return std::__addressof(__x); }
590*38fd1498Szrj 
591*38fd1498Szrj       const_pointer
592*38fd1498Szrj       address(const_reference __x) const _GLIBCXX_NOEXCEPT
593*38fd1498Szrj       { return std::__addressof(__x); }
594*38fd1498Szrj 
595*38fd1498Szrj       size_type
596*38fd1498Szrj       max_size() const _GLIBCXX_USE_NOEXCEPT
597*38fd1498Szrj       { return size_t(-1) / sizeof(_Tp); }
598*38fd1498Szrj 
599*38fd1498Szrj #if __cplusplus >= 201103L
600*38fd1498Szrj       template<typename _Up, typename... _Args>
601*38fd1498Szrj         void
602*38fd1498Szrj         construct(_Up* __p, _Args&&... __args)
603*38fd1498Szrj 	{ ::new((void *)__p) _Up(std::forward<_Args>(__args)...); }
604*38fd1498Szrj 
605*38fd1498Szrj       template<typename _Up>
606*38fd1498Szrj         void
607*38fd1498Szrj         destroy(_Up* __p) { __p->~_Up(); }
608*38fd1498Szrj #else
609*38fd1498Szrj       // _GLIBCXX_RESOLVE_LIB_DEFECTS
610*38fd1498Szrj       // 402. wrong new expression in [some_] allocator::construct
611*38fd1498Szrj       void
612*38fd1498Szrj       construct(pointer __p, const _Tp& __val)
613*38fd1498Szrj       { ::new((void *)__p) _Tp(__val); }
614*38fd1498Szrj 
615*38fd1498Szrj       void
616*38fd1498Szrj       destroy(pointer __p) { __p->~_Tp(); }
617*38fd1498Szrj #endif
618*38fd1498Szrj     };
619*38fd1498Szrj 
620*38fd1498Szrj #ifdef __GTHREADS
621*38fd1498Szrj #define __thread_default true
622*38fd1498Szrj #else
623*38fd1498Szrj #define __thread_default false
624*38fd1498Szrj #endif
625*38fd1498Szrj 
626*38fd1498Szrj   /**
627*38fd1498Szrj    *  @brief  This is a fixed size (power of 2) allocator which - when
628*38fd1498Szrj    *  compiled with thread support - will maintain one freelist per
629*38fd1498Szrj    *  size per thread plus a @a global one. Steps are taken to limit
630*38fd1498Szrj    *  the per thread freelist sizes (by returning excess back to
631*38fd1498Szrj    *  the @a global list).
632*38fd1498Szrj    *  @ingroup allocators
633*38fd1498Szrj    *
634*38fd1498Szrj    *  Further details:
635*38fd1498Szrj    *  https://gcc.gnu.org/onlinedocs/libstdc++/manual/mt_allocator.html
636*38fd1498Szrj    */
637*38fd1498Szrj   template<typename _Tp,
638*38fd1498Szrj 	   typename _Poolp = __common_pool_policy<__pool, __thread_default> >
639*38fd1498Szrj     class __mt_alloc : public __mt_alloc_base<_Tp>
640*38fd1498Szrj     {
641*38fd1498Szrj     public:
642*38fd1498Szrj       typedef size_t                    	size_type;
643*38fd1498Szrj       typedef ptrdiff_t                 	difference_type;
644*38fd1498Szrj       typedef _Tp*                      	pointer;
645*38fd1498Szrj       typedef const _Tp*                	const_pointer;
646*38fd1498Szrj       typedef _Tp&                      	reference;
647*38fd1498Szrj       typedef const _Tp&                	const_reference;
648*38fd1498Szrj       typedef _Tp                       	value_type;
649*38fd1498Szrj       typedef _Poolp      			__policy_type;
650*38fd1498Szrj       typedef typename _Poolp::pool_type	__pool_type;
651*38fd1498Szrj 
652*38fd1498Szrj       template<typename _Tp1, typename _Poolp1 = _Poolp>
653*38fd1498Szrj         struct rebind
654*38fd1498Szrj         {
655*38fd1498Szrj 	  typedef typename _Poolp1::template _M_rebind<_Tp1>::other pol_type;
656*38fd1498Szrj 	  typedef __mt_alloc<_Tp1, pol_type> other;
657*38fd1498Szrj 	};
658*38fd1498Szrj 
659*38fd1498Szrj       __mt_alloc() _GLIBCXX_USE_NOEXCEPT { }
660*38fd1498Szrj 
661*38fd1498Szrj       __mt_alloc(const __mt_alloc&) _GLIBCXX_USE_NOEXCEPT { }
662*38fd1498Szrj 
663*38fd1498Szrj       template<typename _Tp1, typename _Poolp1>
664*38fd1498Szrj         __mt_alloc(const __mt_alloc<_Tp1, _Poolp1>&) _GLIBCXX_USE_NOEXCEPT { }
665*38fd1498Szrj 
666*38fd1498Szrj       ~__mt_alloc() _GLIBCXX_USE_NOEXCEPT { }
667*38fd1498Szrj 
668*38fd1498Szrj       pointer
669*38fd1498Szrj       allocate(size_type __n, const void* = 0);
670*38fd1498Szrj 
671*38fd1498Szrj       void
672*38fd1498Szrj       deallocate(pointer __p, size_type __n);
673*38fd1498Szrj 
674*38fd1498Szrj       const __pool_base::_Tune
675*38fd1498Szrj       _M_get_options()
676*38fd1498Szrj       {
677*38fd1498Szrj 	// Return a copy, not a reference, for external consumption.
678*38fd1498Szrj 	return __policy_type::_S_get_pool()._M_get_options();
679*38fd1498Szrj       }
680*38fd1498Szrj 
681*38fd1498Szrj       void
682*38fd1498Szrj       _M_set_options(__pool_base::_Tune __t)
683*38fd1498Szrj       { __policy_type::_S_get_pool()._M_set_options(__t); }
684*38fd1498Szrj     };
685*38fd1498Szrj 
686*38fd1498Szrj   template<typename _Tp, typename _Poolp>
687*38fd1498Szrj     typename __mt_alloc<_Tp, _Poolp>::pointer
688*38fd1498Szrj     __mt_alloc<_Tp, _Poolp>::
689*38fd1498Szrj     allocate(size_type __n, const void*)
690*38fd1498Szrj     {
691*38fd1498Szrj       if (__n > this->max_size())
692*38fd1498Szrj 	std::__throw_bad_alloc();
693*38fd1498Szrj 
694*38fd1498Szrj #if __cpp_aligned_new
695*38fd1498Szrj       // Types with extended alignment are handled by operator new/delete.
696*38fd1498Szrj       if (alignof(_Tp) > __STDCPP_DEFAULT_NEW_ALIGNMENT__)
697*38fd1498Szrj 	{
698*38fd1498Szrj 	  std::align_val_t __al = std::align_val_t(alignof(_Tp));
699*38fd1498Szrj 	  return static_cast<_Tp*>(::operator new(__n * sizeof(_Tp), __al));
700*38fd1498Szrj 	}
701*38fd1498Szrj #endif
702*38fd1498Szrj 
703*38fd1498Szrj       __policy_type::_S_initialize_once();
704*38fd1498Szrj 
705*38fd1498Szrj       // Requests larger than _M_max_bytes are handled by operator
706*38fd1498Szrj       // new/delete directly.
707*38fd1498Szrj       __pool_type& __pool = __policy_type::_S_get_pool();
708*38fd1498Szrj       const size_t __bytes = __n * sizeof(_Tp);
709*38fd1498Szrj       if (__pool._M_check_threshold(__bytes))
710*38fd1498Szrj 	{
711*38fd1498Szrj 	  void* __ret = ::operator new(__bytes);
712*38fd1498Szrj 	  return static_cast<_Tp*>(__ret);
713*38fd1498Szrj 	}
714*38fd1498Szrj 
715*38fd1498Szrj       // Round up to power of 2 and figure out which bin to use.
716*38fd1498Szrj       const size_t __which = __pool._M_get_binmap(__bytes);
717*38fd1498Szrj       const size_t __thread_id = __pool._M_get_thread_id();
718*38fd1498Szrj 
719*38fd1498Szrj       // Find out if we have blocks on our freelist.  If so, go ahead
720*38fd1498Szrj       // and use them directly without having to lock anything.
721*38fd1498Szrj       char* __c;
722*38fd1498Szrj       typedef typename __pool_type::_Bin_record _Bin_record;
723*38fd1498Szrj       const _Bin_record& __bin = __pool._M_get_bin(__which);
724*38fd1498Szrj       if (__bin._M_first[__thread_id])
725*38fd1498Szrj 	{
726*38fd1498Szrj 	  // Already reserved.
727*38fd1498Szrj 	  typedef typename __pool_type::_Block_record _Block_record;
728*38fd1498Szrj 	  _Block_record* __block = __bin._M_first[__thread_id];
729*38fd1498Szrj 	  __bin._M_first[__thread_id] = __block->_M_next;
730*38fd1498Szrj 
731*38fd1498Szrj 	  __pool._M_adjust_freelist(__bin, __block, __thread_id);
732*38fd1498Szrj 	  __c = reinterpret_cast<char*>(__block) + __pool._M_get_align();
733*38fd1498Szrj 	}
734*38fd1498Szrj       else
735*38fd1498Szrj 	{
736*38fd1498Szrj 	  // Null, reserve.
737*38fd1498Szrj 	  __c = __pool._M_reserve_block(__bytes, __thread_id);
738*38fd1498Szrj 	}
739*38fd1498Szrj       return static_cast<_Tp*>(static_cast<void*>(__c));
740*38fd1498Szrj     }
741*38fd1498Szrj 
742*38fd1498Szrj   template<typename _Tp, typename _Poolp>
743*38fd1498Szrj     void
744*38fd1498Szrj     __mt_alloc<_Tp, _Poolp>::
745*38fd1498Szrj     deallocate(pointer __p, size_type __n)
746*38fd1498Szrj     {
747*38fd1498Szrj       if (__builtin_expect(__p != 0, true))
748*38fd1498Szrj 	{
749*38fd1498Szrj #if __cpp_aligned_new
750*38fd1498Szrj 	  // Types with extended alignment are handled by operator new/delete.
751*38fd1498Szrj 	  if (alignof(_Tp) > __STDCPP_DEFAULT_NEW_ALIGNMENT__)
752*38fd1498Szrj 	    {
753*38fd1498Szrj 	      ::operator delete(__p, std::align_val_t(alignof(_Tp)));
754*38fd1498Szrj 	      return;
755*38fd1498Szrj 	    }
756*38fd1498Szrj #endif
757*38fd1498Szrj 
758*38fd1498Szrj 	  // Requests larger than _M_max_bytes are handled by
759*38fd1498Szrj 	  // operators new/delete directly.
760*38fd1498Szrj 	  __pool_type& __pool = __policy_type::_S_get_pool();
761*38fd1498Szrj 	  const size_t __bytes = __n * sizeof(_Tp);
762*38fd1498Szrj 	  if (__pool._M_check_threshold(__bytes))
763*38fd1498Szrj 	    ::operator delete(__p);
764*38fd1498Szrj 	  else
765*38fd1498Szrj 	    __pool._M_reclaim_block(reinterpret_cast<char*>(__p), __bytes);
766*38fd1498Szrj 	}
767*38fd1498Szrj     }
768*38fd1498Szrj 
769*38fd1498Szrj   template<typename _Tp, typename _Poolp>
770*38fd1498Szrj     inline bool
771*38fd1498Szrj     operator==(const __mt_alloc<_Tp, _Poolp>&, const __mt_alloc<_Tp, _Poolp>&)
772*38fd1498Szrj     { return true; }
773*38fd1498Szrj 
774*38fd1498Szrj   template<typename _Tp, typename _Poolp>
775*38fd1498Szrj     inline bool
776*38fd1498Szrj     operator!=(const __mt_alloc<_Tp, _Poolp>&, const __mt_alloc<_Tp, _Poolp>&)
777*38fd1498Szrj     { return false; }
778*38fd1498Szrj 
779*38fd1498Szrj #undef __thread_default
780*38fd1498Szrj 
781*38fd1498Szrj _GLIBCXX_END_NAMESPACE_VERSION
782*38fd1498Szrj } // namespace
783*38fd1498Szrj 
784*38fd1498Szrj #endif
785