xref: /dflybsd-src/contrib/gcc-8.0/libgomp/loop_ull.c (revision 38fd149817dfbff97799f62fcb70be98c4e32523)
1*38fd1498Szrj /* Copyright (C) 2005-2018 Free Software Foundation, Inc.
2*38fd1498Szrj    Contributed by Richard Henderson <rth@redhat.com>.
3*38fd1498Szrj 
4*38fd1498Szrj    This file is part of the GNU Offloading and Multi Processing Library
5*38fd1498Szrj    (libgomp).
6*38fd1498Szrj 
7*38fd1498Szrj    Libgomp is free software; you can redistribute it and/or modify it
8*38fd1498Szrj    under the terms of the GNU General Public License as published by
9*38fd1498Szrj    the Free Software Foundation; either version 3, or (at your option)
10*38fd1498Szrj    any later version.
11*38fd1498Szrj 
12*38fd1498Szrj    Libgomp is distributed in the hope that it will be useful, but WITHOUT ANY
13*38fd1498Szrj    WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS
14*38fd1498Szrj    FOR A PARTICULAR PURPOSE.  See the GNU General Public License for
15*38fd1498Szrj    more details.
16*38fd1498Szrj 
17*38fd1498Szrj    Under Section 7 of GPL version 3, you are granted additional
18*38fd1498Szrj    permissions described in the GCC Runtime Library Exception, version
19*38fd1498Szrj    3.1, as published by the Free Software Foundation.
20*38fd1498Szrj 
21*38fd1498Szrj    You should have received a copy of the GNU General Public License and
22*38fd1498Szrj    a copy of the GCC Runtime Library Exception along with this program;
23*38fd1498Szrj    see the files COPYING3 and COPYING.RUNTIME respectively.  If not, see
24*38fd1498Szrj    <http://www.gnu.org/licenses/>.  */
25*38fd1498Szrj 
26*38fd1498Szrj /* This file handles the LOOP (FOR/DO) construct.  */
27*38fd1498Szrj 
28*38fd1498Szrj #include <limits.h>
29*38fd1498Szrj #include <stdlib.h>
30*38fd1498Szrj #include "libgomp.h"
31*38fd1498Szrj 
32*38fd1498Szrj typedef unsigned long long gomp_ull;
33*38fd1498Szrj 
34*38fd1498Szrj /* Initialize the given work share construct from the given arguments.  */
35*38fd1498Szrj 
36*38fd1498Szrj static inline void
gomp_loop_ull_init(struct gomp_work_share * ws,bool up,gomp_ull start,gomp_ull end,gomp_ull incr,enum gomp_schedule_type sched,gomp_ull chunk_size)37*38fd1498Szrj gomp_loop_ull_init (struct gomp_work_share *ws, bool up, gomp_ull start,
38*38fd1498Szrj 		    gomp_ull end, gomp_ull incr, enum gomp_schedule_type sched,
39*38fd1498Szrj 		    gomp_ull chunk_size)
40*38fd1498Szrj {
41*38fd1498Szrj   ws->sched = sched;
42*38fd1498Szrj   ws->chunk_size_ull = chunk_size;
43*38fd1498Szrj   /* Canonicalize loops that have zero iterations to ->next == ->end.  */
44*38fd1498Szrj   ws->end_ull = ((up && start > end) || (!up && start < end))
45*38fd1498Szrj 		? start : end;
46*38fd1498Szrj   ws->incr_ull = incr;
47*38fd1498Szrj   ws->next_ull = start;
48*38fd1498Szrj   ws->mode = 0;
49*38fd1498Szrj   if (sched == GFS_DYNAMIC)
50*38fd1498Szrj     {
51*38fd1498Szrj       ws->chunk_size_ull *= incr;
52*38fd1498Szrj 
53*38fd1498Szrj #if defined HAVE_SYNC_BUILTINS && defined __LP64__
54*38fd1498Szrj       {
55*38fd1498Szrj 	/* For dynamic scheduling prepare things to make each iteration
56*38fd1498Szrj 	   faster.  */
57*38fd1498Szrj 	struct gomp_thread *thr = gomp_thread ();
58*38fd1498Szrj 	struct gomp_team *team = thr->ts.team;
59*38fd1498Szrj 	long nthreads = team ? team->nthreads : 1;
60*38fd1498Szrj 
61*38fd1498Szrj 	if (__builtin_expect (up, 1))
62*38fd1498Szrj 	  {
63*38fd1498Szrj 	    /* Cheap overflow protection.  */
64*38fd1498Szrj 	    if (__builtin_expect ((nthreads | ws->chunk_size_ull)
65*38fd1498Szrj 				  < 1ULL << (sizeof (gomp_ull)
66*38fd1498Szrj 					     * __CHAR_BIT__ / 2 - 1), 1))
67*38fd1498Szrj 	      ws->mode = ws->end_ull < (__LONG_LONG_MAX__ * 2ULL + 1
68*38fd1498Szrj 					- (nthreads + 1) * ws->chunk_size_ull);
69*38fd1498Szrj 	  }
70*38fd1498Szrj 	/* Cheap overflow protection.  */
71*38fd1498Szrj 	else if (__builtin_expect ((nthreads | -ws->chunk_size_ull)
72*38fd1498Szrj 				   < 1ULL << (sizeof (gomp_ull)
73*38fd1498Szrj 					      * __CHAR_BIT__ / 2 - 1), 1))
74*38fd1498Szrj 	  ws->mode = ws->end_ull > ((nthreads + 1) * -ws->chunk_size_ull
75*38fd1498Szrj 				    - (__LONG_LONG_MAX__ * 2ULL + 1));
76*38fd1498Szrj       }
77*38fd1498Szrj #endif
78*38fd1498Szrj     }
79*38fd1498Szrj   if (!up)
80*38fd1498Szrj     ws->mode |= 2;
81*38fd1498Szrj }
82*38fd1498Szrj 
83*38fd1498Szrj /* The *_start routines are called when first encountering a loop construct
84*38fd1498Szrj    that is not bound directly to a parallel construct.  The first thread
85*38fd1498Szrj    that arrives will create the work-share construct; subsequent threads
86*38fd1498Szrj    will see the construct exists and allocate work from it.
87*38fd1498Szrj 
88*38fd1498Szrj    START, END, INCR are the bounds of the loop; due to the restrictions of
89*38fd1498Szrj    OpenMP, these values must be the same in every thread.  This is not
90*38fd1498Szrj    verified (nor is it entirely verifiable, since START is not necessarily
91*38fd1498Szrj    retained intact in the work-share data structure).  CHUNK_SIZE is the
92*38fd1498Szrj    scheduling parameter; again this must be identical in all threads.
93*38fd1498Szrj 
94*38fd1498Szrj    Returns true if there's any work for this thread to perform.  If so,
95*38fd1498Szrj    *ISTART and *IEND are filled with the bounds of the iteration block
96*38fd1498Szrj    allocated to this thread.  Returns false if all work was assigned to
97*38fd1498Szrj    other threads prior to this thread's arrival.  */
98*38fd1498Szrj 
99*38fd1498Szrj static bool
gomp_loop_ull_static_start(bool up,gomp_ull start,gomp_ull end,gomp_ull incr,gomp_ull chunk_size,gomp_ull * istart,gomp_ull * iend)100*38fd1498Szrj gomp_loop_ull_static_start (bool up, gomp_ull start, gomp_ull end,
101*38fd1498Szrj 			    gomp_ull incr, gomp_ull chunk_size,
102*38fd1498Szrj 			    gomp_ull *istart, gomp_ull *iend)
103*38fd1498Szrj {
104*38fd1498Szrj   struct gomp_thread *thr = gomp_thread ();
105*38fd1498Szrj 
106*38fd1498Szrj   thr->ts.static_trip = 0;
107*38fd1498Szrj   if (gomp_work_share_start (false))
108*38fd1498Szrj     {
109*38fd1498Szrj       gomp_loop_ull_init (thr->ts.work_share, up, start, end, incr,
110*38fd1498Szrj 			  GFS_STATIC, chunk_size);
111*38fd1498Szrj       gomp_work_share_init_done ();
112*38fd1498Szrj     }
113*38fd1498Szrj 
114*38fd1498Szrj   return !gomp_iter_ull_static_next (istart, iend);
115*38fd1498Szrj }
116*38fd1498Szrj 
117*38fd1498Szrj static bool
gomp_loop_ull_dynamic_start(bool up,gomp_ull start,gomp_ull end,gomp_ull incr,gomp_ull chunk_size,gomp_ull * istart,gomp_ull * iend)118*38fd1498Szrj gomp_loop_ull_dynamic_start (bool up, gomp_ull start, gomp_ull end,
119*38fd1498Szrj 			     gomp_ull incr, gomp_ull chunk_size,
120*38fd1498Szrj 			     gomp_ull *istart, gomp_ull *iend)
121*38fd1498Szrj {
122*38fd1498Szrj   struct gomp_thread *thr = gomp_thread ();
123*38fd1498Szrj   bool ret;
124*38fd1498Szrj 
125*38fd1498Szrj   if (gomp_work_share_start (false))
126*38fd1498Szrj     {
127*38fd1498Szrj       gomp_loop_ull_init (thr->ts.work_share, up, start, end, incr,
128*38fd1498Szrj 			  GFS_DYNAMIC, chunk_size);
129*38fd1498Szrj       gomp_work_share_init_done ();
130*38fd1498Szrj     }
131*38fd1498Szrj 
132*38fd1498Szrj #if defined HAVE_SYNC_BUILTINS && defined __LP64__
133*38fd1498Szrj   ret = gomp_iter_ull_dynamic_next (istart, iend);
134*38fd1498Szrj #else
135*38fd1498Szrj   gomp_mutex_lock (&thr->ts.work_share->lock);
136*38fd1498Szrj   ret = gomp_iter_ull_dynamic_next_locked (istart, iend);
137*38fd1498Szrj   gomp_mutex_unlock (&thr->ts.work_share->lock);
138*38fd1498Szrj #endif
139*38fd1498Szrj 
140*38fd1498Szrj   return ret;
141*38fd1498Szrj }
142*38fd1498Szrj 
143*38fd1498Szrj static bool
gomp_loop_ull_guided_start(bool up,gomp_ull start,gomp_ull end,gomp_ull incr,gomp_ull chunk_size,gomp_ull * istart,gomp_ull * iend)144*38fd1498Szrj gomp_loop_ull_guided_start (bool up, gomp_ull start, gomp_ull end,
145*38fd1498Szrj 			    gomp_ull incr, gomp_ull chunk_size,
146*38fd1498Szrj 			    gomp_ull *istart, gomp_ull *iend)
147*38fd1498Szrj {
148*38fd1498Szrj   struct gomp_thread *thr = gomp_thread ();
149*38fd1498Szrj   bool ret;
150*38fd1498Szrj 
151*38fd1498Szrj   if (gomp_work_share_start (false))
152*38fd1498Szrj     {
153*38fd1498Szrj       gomp_loop_ull_init (thr->ts.work_share, up, start, end, incr,
154*38fd1498Szrj 			  GFS_GUIDED, chunk_size);
155*38fd1498Szrj       gomp_work_share_init_done ();
156*38fd1498Szrj     }
157*38fd1498Szrj 
158*38fd1498Szrj #if defined HAVE_SYNC_BUILTINS && defined __LP64__
159*38fd1498Szrj   ret = gomp_iter_ull_guided_next (istart, iend);
160*38fd1498Szrj #else
161*38fd1498Szrj   gomp_mutex_lock (&thr->ts.work_share->lock);
162*38fd1498Szrj   ret = gomp_iter_ull_guided_next_locked (istart, iend);
163*38fd1498Szrj   gomp_mutex_unlock (&thr->ts.work_share->lock);
164*38fd1498Szrj #endif
165*38fd1498Szrj 
166*38fd1498Szrj   return ret;
167*38fd1498Szrj }
168*38fd1498Szrj 
169*38fd1498Szrj bool
GOMP_loop_ull_runtime_start(bool up,gomp_ull start,gomp_ull end,gomp_ull incr,gomp_ull * istart,gomp_ull * iend)170*38fd1498Szrj GOMP_loop_ull_runtime_start (bool up, gomp_ull start, gomp_ull end,
171*38fd1498Szrj 			     gomp_ull incr, gomp_ull *istart, gomp_ull *iend)
172*38fd1498Szrj {
173*38fd1498Szrj   struct gomp_task_icv *icv = gomp_icv (false);
174*38fd1498Szrj   switch (icv->run_sched_var)
175*38fd1498Szrj     {
176*38fd1498Szrj     case GFS_STATIC:
177*38fd1498Szrj       return gomp_loop_ull_static_start (up, start, end, incr,
178*38fd1498Szrj 					 icv->run_sched_chunk_size,
179*38fd1498Szrj 					 istart, iend);
180*38fd1498Szrj     case GFS_DYNAMIC:
181*38fd1498Szrj       return gomp_loop_ull_dynamic_start (up, start, end, incr,
182*38fd1498Szrj 					  icv->run_sched_chunk_size,
183*38fd1498Szrj 					  istart, iend);
184*38fd1498Szrj     case GFS_GUIDED:
185*38fd1498Szrj       return gomp_loop_ull_guided_start (up, start, end, incr,
186*38fd1498Szrj 					 icv->run_sched_chunk_size,
187*38fd1498Szrj 					 istart, iend);
188*38fd1498Szrj     case GFS_AUTO:
189*38fd1498Szrj       /* For now map to schedule(static), later on we could play with feedback
190*38fd1498Szrj 	 driven choice.  */
191*38fd1498Szrj       return gomp_loop_ull_static_start (up, start, end, incr,
192*38fd1498Szrj 					 0, istart, iend);
193*38fd1498Szrj     default:
194*38fd1498Szrj       abort ();
195*38fd1498Szrj     }
196*38fd1498Szrj }
197*38fd1498Szrj 
198*38fd1498Szrj /* The *_ordered_*_start routines are similar.  The only difference is that
199*38fd1498Szrj    this work-share construct is initialized to expect an ORDERED section.  */
200*38fd1498Szrj 
201*38fd1498Szrj static bool
gomp_loop_ull_ordered_static_start(bool up,gomp_ull start,gomp_ull end,gomp_ull incr,gomp_ull chunk_size,gomp_ull * istart,gomp_ull * iend)202*38fd1498Szrj gomp_loop_ull_ordered_static_start (bool up, gomp_ull start, gomp_ull end,
203*38fd1498Szrj 				    gomp_ull incr, gomp_ull chunk_size,
204*38fd1498Szrj 				    gomp_ull *istart, gomp_ull *iend)
205*38fd1498Szrj {
206*38fd1498Szrj   struct gomp_thread *thr = gomp_thread ();
207*38fd1498Szrj 
208*38fd1498Szrj   thr->ts.static_trip = 0;
209*38fd1498Szrj   if (gomp_work_share_start (true))
210*38fd1498Szrj     {
211*38fd1498Szrj       gomp_loop_ull_init (thr->ts.work_share, up, start, end, incr,
212*38fd1498Szrj 			  GFS_STATIC, chunk_size);
213*38fd1498Szrj       gomp_ordered_static_init ();
214*38fd1498Szrj       gomp_work_share_init_done ();
215*38fd1498Szrj     }
216*38fd1498Szrj 
217*38fd1498Szrj   return !gomp_iter_ull_static_next (istart, iend);
218*38fd1498Szrj }
219*38fd1498Szrj 
220*38fd1498Szrj static bool
gomp_loop_ull_ordered_dynamic_start(bool up,gomp_ull start,gomp_ull end,gomp_ull incr,gomp_ull chunk_size,gomp_ull * istart,gomp_ull * iend)221*38fd1498Szrj gomp_loop_ull_ordered_dynamic_start (bool up, gomp_ull start, gomp_ull end,
222*38fd1498Szrj 				     gomp_ull incr, gomp_ull chunk_size,
223*38fd1498Szrj 				     gomp_ull *istart, gomp_ull *iend)
224*38fd1498Szrj {
225*38fd1498Szrj   struct gomp_thread *thr = gomp_thread ();
226*38fd1498Szrj   bool ret;
227*38fd1498Szrj 
228*38fd1498Szrj   if (gomp_work_share_start (true))
229*38fd1498Szrj     {
230*38fd1498Szrj       gomp_loop_ull_init (thr->ts.work_share, up, start, end, incr,
231*38fd1498Szrj 			  GFS_DYNAMIC, chunk_size);
232*38fd1498Szrj       gomp_mutex_lock (&thr->ts.work_share->lock);
233*38fd1498Szrj       gomp_work_share_init_done ();
234*38fd1498Szrj     }
235*38fd1498Szrj   else
236*38fd1498Szrj     gomp_mutex_lock (&thr->ts.work_share->lock);
237*38fd1498Szrj 
238*38fd1498Szrj   ret = gomp_iter_ull_dynamic_next_locked (istart, iend);
239*38fd1498Szrj   if (ret)
240*38fd1498Szrj     gomp_ordered_first ();
241*38fd1498Szrj   gomp_mutex_unlock (&thr->ts.work_share->lock);
242*38fd1498Szrj 
243*38fd1498Szrj   return ret;
244*38fd1498Szrj }
245*38fd1498Szrj 
246*38fd1498Szrj static bool
gomp_loop_ull_ordered_guided_start(bool up,gomp_ull start,gomp_ull end,gomp_ull incr,gomp_ull chunk_size,gomp_ull * istart,gomp_ull * iend)247*38fd1498Szrj gomp_loop_ull_ordered_guided_start (bool up, gomp_ull start, gomp_ull end,
248*38fd1498Szrj 				    gomp_ull incr, gomp_ull chunk_size,
249*38fd1498Szrj 				    gomp_ull *istart, gomp_ull *iend)
250*38fd1498Szrj {
251*38fd1498Szrj   struct gomp_thread *thr = gomp_thread ();
252*38fd1498Szrj   bool ret;
253*38fd1498Szrj 
254*38fd1498Szrj   if (gomp_work_share_start (true))
255*38fd1498Szrj     {
256*38fd1498Szrj       gomp_loop_ull_init (thr->ts.work_share, up, start, end, incr,
257*38fd1498Szrj 			  GFS_GUIDED, chunk_size);
258*38fd1498Szrj       gomp_mutex_lock (&thr->ts.work_share->lock);
259*38fd1498Szrj       gomp_work_share_init_done ();
260*38fd1498Szrj     }
261*38fd1498Szrj   else
262*38fd1498Szrj     gomp_mutex_lock (&thr->ts.work_share->lock);
263*38fd1498Szrj 
264*38fd1498Szrj   ret = gomp_iter_ull_guided_next_locked (istart, iend);
265*38fd1498Szrj   if (ret)
266*38fd1498Szrj     gomp_ordered_first ();
267*38fd1498Szrj   gomp_mutex_unlock (&thr->ts.work_share->lock);
268*38fd1498Szrj 
269*38fd1498Szrj   return ret;
270*38fd1498Szrj }
271*38fd1498Szrj 
272*38fd1498Szrj bool
GOMP_loop_ull_ordered_runtime_start(bool up,gomp_ull start,gomp_ull end,gomp_ull incr,gomp_ull * istart,gomp_ull * iend)273*38fd1498Szrj GOMP_loop_ull_ordered_runtime_start (bool up, gomp_ull start, gomp_ull end,
274*38fd1498Szrj 				     gomp_ull incr, gomp_ull *istart,
275*38fd1498Szrj 				     gomp_ull *iend)
276*38fd1498Szrj {
277*38fd1498Szrj   struct gomp_task_icv *icv = gomp_icv (false);
278*38fd1498Szrj   switch (icv->run_sched_var)
279*38fd1498Szrj     {
280*38fd1498Szrj     case GFS_STATIC:
281*38fd1498Szrj       return gomp_loop_ull_ordered_static_start (up, start, end, incr,
282*38fd1498Szrj 						 icv->run_sched_chunk_size,
283*38fd1498Szrj 						 istart, iend);
284*38fd1498Szrj     case GFS_DYNAMIC:
285*38fd1498Szrj       return gomp_loop_ull_ordered_dynamic_start (up, start, end, incr,
286*38fd1498Szrj 						  icv->run_sched_chunk_size,
287*38fd1498Szrj 						  istart, iend);
288*38fd1498Szrj     case GFS_GUIDED:
289*38fd1498Szrj       return gomp_loop_ull_ordered_guided_start (up, start, end, incr,
290*38fd1498Szrj 						 icv->run_sched_chunk_size,
291*38fd1498Szrj 						 istart, iend);
292*38fd1498Szrj     case GFS_AUTO:
293*38fd1498Szrj       /* For now map to schedule(static), later on we could play with feedback
294*38fd1498Szrj 	 driven choice.  */
295*38fd1498Szrj       return gomp_loop_ull_ordered_static_start (up, start, end, incr,
296*38fd1498Szrj 						 0, istart, iend);
297*38fd1498Szrj     default:
298*38fd1498Szrj       abort ();
299*38fd1498Szrj     }
300*38fd1498Szrj }
301*38fd1498Szrj 
302*38fd1498Szrj /* The *_doacross_*_start routines are similar.  The only difference is that
303*38fd1498Szrj    this work-share construct is initialized to expect an ORDERED(N) - DOACROSS
304*38fd1498Szrj    section, and the worksharing loop iterates always from 0 to COUNTS[0] - 1
305*38fd1498Szrj    and other COUNTS array elements tell the library number of iterations
306*38fd1498Szrj    in the ordered inner loops.  */
307*38fd1498Szrj 
308*38fd1498Szrj static bool
gomp_loop_ull_doacross_static_start(unsigned ncounts,gomp_ull * counts,gomp_ull chunk_size,gomp_ull * istart,gomp_ull * iend)309*38fd1498Szrj gomp_loop_ull_doacross_static_start (unsigned ncounts, gomp_ull *counts,
310*38fd1498Szrj 				     gomp_ull chunk_size, gomp_ull *istart,
311*38fd1498Szrj 				     gomp_ull *iend)
312*38fd1498Szrj {
313*38fd1498Szrj   struct gomp_thread *thr = gomp_thread ();
314*38fd1498Szrj 
315*38fd1498Szrj   thr->ts.static_trip = 0;
316*38fd1498Szrj   if (gomp_work_share_start (false))
317*38fd1498Szrj     {
318*38fd1498Szrj       gomp_loop_ull_init (thr->ts.work_share, true, 0, counts[0], 1,
319*38fd1498Szrj 			  GFS_STATIC, chunk_size);
320*38fd1498Szrj       gomp_doacross_ull_init (ncounts, counts, chunk_size);
321*38fd1498Szrj       gomp_work_share_init_done ();
322*38fd1498Szrj     }
323*38fd1498Szrj 
324*38fd1498Szrj   return !gomp_iter_ull_static_next (istart, iend);
325*38fd1498Szrj }
326*38fd1498Szrj 
327*38fd1498Szrj static bool
gomp_loop_ull_doacross_dynamic_start(unsigned ncounts,gomp_ull * counts,gomp_ull chunk_size,gomp_ull * istart,gomp_ull * iend)328*38fd1498Szrj gomp_loop_ull_doacross_dynamic_start (unsigned ncounts, gomp_ull *counts,
329*38fd1498Szrj 				      gomp_ull chunk_size, gomp_ull *istart,
330*38fd1498Szrj 				      gomp_ull *iend)
331*38fd1498Szrj {
332*38fd1498Szrj   struct gomp_thread *thr = gomp_thread ();
333*38fd1498Szrj   bool ret;
334*38fd1498Szrj 
335*38fd1498Szrj   if (gomp_work_share_start (false))
336*38fd1498Szrj     {
337*38fd1498Szrj       gomp_loop_ull_init (thr->ts.work_share, true, 0, counts[0], 1,
338*38fd1498Szrj 			  GFS_DYNAMIC, chunk_size);
339*38fd1498Szrj       gomp_doacross_ull_init (ncounts, counts, chunk_size);
340*38fd1498Szrj       gomp_work_share_init_done ();
341*38fd1498Szrj     }
342*38fd1498Szrj 
343*38fd1498Szrj #if defined HAVE_SYNC_BUILTINS && defined __LP64__
344*38fd1498Szrj   ret = gomp_iter_ull_dynamic_next (istart, iend);
345*38fd1498Szrj #else
346*38fd1498Szrj   gomp_mutex_lock (&thr->ts.work_share->lock);
347*38fd1498Szrj   ret = gomp_iter_ull_dynamic_next_locked (istart, iend);
348*38fd1498Szrj   gomp_mutex_unlock (&thr->ts.work_share->lock);
349*38fd1498Szrj #endif
350*38fd1498Szrj 
351*38fd1498Szrj   return ret;
352*38fd1498Szrj }
353*38fd1498Szrj 
354*38fd1498Szrj static bool
gomp_loop_ull_doacross_guided_start(unsigned ncounts,gomp_ull * counts,gomp_ull chunk_size,gomp_ull * istart,gomp_ull * iend)355*38fd1498Szrj gomp_loop_ull_doacross_guided_start (unsigned ncounts, gomp_ull *counts,
356*38fd1498Szrj 				     gomp_ull chunk_size, gomp_ull *istart,
357*38fd1498Szrj 				     gomp_ull *iend)
358*38fd1498Szrj {
359*38fd1498Szrj   struct gomp_thread *thr = gomp_thread ();
360*38fd1498Szrj   bool ret;
361*38fd1498Szrj 
362*38fd1498Szrj   if (gomp_work_share_start (false))
363*38fd1498Szrj     {
364*38fd1498Szrj       gomp_loop_ull_init (thr->ts.work_share, true, 0, counts[0], 1,
365*38fd1498Szrj 			  GFS_GUIDED, chunk_size);
366*38fd1498Szrj       gomp_doacross_ull_init (ncounts, counts, chunk_size);
367*38fd1498Szrj       gomp_work_share_init_done ();
368*38fd1498Szrj     }
369*38fd1498Szrj 
370*38fd1498Szrj #if defined HAVE_SYNC_BUILTINS && defined __LP64__
371*38fd1498Szrj   ret = gomp_iter_ull_guided_next (istart, iend);
372*38fd1498Szrj #else
373*38fd1498Szrj   gomp_mutex_lock (&thr->ts.work_share->lock);
374*38fd1498Szrj   ret = gomp_iter_ull_guided_next_locked (istart, iend);
375*38fd1498Szrj   gomp_mutex_unlock (&thr->ts.work_share->lock);
376*38fd1498Szrj #endif
377*38fd1498Szrj 
378*38fd1498Szrj   return ret;
379*38fd1498Szrj }
380*38fd1498Szrj 
381*38fd1498Szrj bool
GOMP_loop_ull_doacross_runtime_start(unsigned ncounts,gomp_ull * counts,gomp_ull * istart,gomp_ull * iend)382*38fd1498Szrj GOMP_loop_ull_doacross_runtime_start (unsigned ncounts, gomp_ull *counts,
383*38fd1498Szrj 				      gomp_ull *istart, gomp_ull *iend)
384*38fd1498Szrj {
385*38fd1498Szrj   struct gomp_task_icv *icv = gomp_icv (false);
386*38fd1498Szrj   switch (icv->run_sched_var)
387*38fd1498Szrj     {
388*38fd1498Szrj     case GFS_STATIC:
389*38fd1498Szrj       return gomp_loop_ull_doacross_static_start (ncounts, counts,
390*38fd1498Szrj 						  icv->run_sched_chunk_size,
391*38fd1498Szrj 						  istart, iend);
392*38fd1498Szrj     case GFS_DYNAMIC:
393*38fd1498Szrj       return gomp_loop_ull_doacross_dynamic_start (ncounts, counts,
394*38fd1498Szrj 						   icv->run_sched_chunk_size,
395*38fd1498Szrj 						   istart, iend);
396*38fd1498Szrj     case GFS_GUIDED:
397*38fd1498Szrj       return gomp_loop_ull_doacross_guided_start (ncounts, counts,
398*38fd1498Szrj 						  icv->run_sched_chunk_size,
399*38fd1498Szrj 						  istart, iend);
400*38fd1498Szrj     case GFS_AUTO:
401*38fd1498Szrj       /* For now map to schedule(static), later on we could play with feedback
402*38fd1498Szrj 	 driven choice.  */
403*38fd1498Szrj       return gomp_loop_ull_doacross_static_start (ncounts, counts,
404*38fd1498Szrj 						  0, istart, iend);
405*38fd1498Szrj     default:
406*38fd1498Szrj       abort ();
407*38fd1498Szrj     }
408*38fd1498Szrj }
409*38fd1498Szrj 
410*38fd1498Szrj /* The *_next routines are called when the thread completes processing of
411*38fd1498Szrj    the iteration block currently assigned to it.  If the work-share
412*38fd1498Szrj    construct is bound directly to a parallel construct, then the iteration
413*38fd1498Szrj    bounds may have been set up before the parallel.  In which case, this
414*38fd1498Szrj    may be the first iteration for the thread.
415*38fd1498Szrj 
416*38fd1498Szrj    Returns true if there is work remaining to be performed; *ISTART and
417*38fd1498Szrj    *IEND are filled with a new iteration block.  Returns false if all work
418*38fd1498Szrj    has been assigned.  */
419*38fd1498Szrj 
420*38fd1498Szrj static bool
gomp_loop_ull_static_next(gomp_ull * istart,gomp_ull * iend)421*38fd1498Szrj gomp_loop_ull_static_next (gomp_ull *istart, gomp_ull *iend)
422*38fd1498Szrj {
423*38fd1498Szrj   return !gomp_iter_ull_static_next (istart, iend);
424*38fd1498Szrj }
425*38fd1498Szrj 
426*38fd1498Szrj static bool
gomp_loop_ull_dynamic_next(gomp_ull * istart,gomp_ull * iend)427*38fd1498Szrj gomp_loop_ull_dynamic_next (gomp_ull *istart, gomp_ull *iend)
428*38fd1498Szrj {
429*38fd1498Szrj   bool ret;
430*38fd1498Szrj 
431*38fd1498Szrj #if defined HAVE_SYNC_BUILTINS && defined __LP64__
432*38fd1498Szrj   ret = gomp_iter_ull_dynamic_next (istart, iend);
433*38fd1498Szrj #else
434*38fd1498Szrj   struct gomp_thread *thr = gomp_thread ();
435*38fd1498Szrj   gomp_mutex_lock (&thr->ts.work_share->lock);
436*38fd1498Szrj   ret = gomp_iter_ull_dynamic_next_locked (istart, iend);
437*38fd1498Szrj   gomp_mutex_unlock (&thr->ts.work_share->lock);
438*38fd1498Szrj #endif
439*38fd1498Szrj 
440*38fd1498Szrj   return ret;
441*38fd1498Szrj }
442*38fd1498Szrj 
443*38fd1498Szrj static bool
gomp_loop_ull_guided_next(gomp_ull * istart,gomp_ull * iend)444*38fd1498Szrj gomp_loop_ull_guided_next (gomp_ull *istart, gomp_ull *iend)
445*38fd1498Szrj {
446*38fd1498Szrj   bool ret;
447*38fd1498Szrj 
448*38fd1498Szrj #if defined HAVE_SYNC_BUILTINS && defined __LP64__
449*38fd1498Szrj   ret = gomp_iter_ull_guided_next (istart, iend);
450*38fd1498Szrj #else
451*38fd1498Szrj   struct gomp_thread *thr = gomp_thread ();
452*38fd1498Szrj   gomp_mutex_lock (&thr->ts.work_share->lock);
453*38fd1498Szrj   ret = gomp_iter_ull_guided_next_locked (istart, iend);
454*38fd1498Szrj   gomp_mutex_unlock (&thr->ts.work_share->lock);
455*38fd1498Szrj #endif
456*38fd1498Szrj 
457*38fd1498Szrj   return ret;
458*38fd1498Szrj }
459*38fd1498Szrj 
460*38fd1498Szrj bool
GOMP_loop_ull_runtime_next(gomp_ull * istart,gomp_ull * iend)461*38fd1498Szrj GOMP_loop_ull_runtime_next (gomp_ull *istart, gomp_ull *iend)
462*38fd1498Szrj {
463*38fd1498Szrj   struct gomp_thread *thr = gomp_thread ();
464*38fd1498Szrj 
465*38fd1498Szrj   switch (thr->ts.work_share->sched)
466*38fd1498Szrj     {
467*38fd1498Szrj     case GFS_STATIC:
468*38fd1498Szrj     case GFS_AUTO:
469*38fd1498Szrj       return gomp_loop_ull_static_next (istart, iend);
470*38fd1498Szrj     case GFS_DYNAMIC:
471*38fd1498Szrj       return gomp_loop_ull_dynamic_next (istart, iend);
472*38fd1498Szrj     case GFS_GUIDED:
473*38fd1498Szrj       return gomp_loop_ull_guided_next (istart, iend);
474*38fd1498Szrj     default:
475*38fd1498Szrj       abort ();
476*38fd1498Szrj     }
477*38fd1498Szrj }
478*38fd1498Szrj 
479*38fd1498Szrj /* The *_ordered_*_next routines are called when the thread completes
480*38fd1498Szrj    processing of the iteration block currently assigned to it.
481*38fd1498Szrj 
482*38fd1498Szrj    Returns true if there is work remaining to be performed; *ISTART and
483*38fd1498Szrj    *IEND are filled with a new iteration block.  Returns false if all work
484*38fd1498Szrj    has been assigned.  */
485*38fd1498Szrj 
486*38fd1498Szrj static bool
gomp_loop_ull_ordered_static_next(gomp_ull * istart,gomp_ull * iend)487*38fd1498Szrj gomp_loop_ull_ordered_static_next (gomp_ull *istart, gomp_ull *iend)
488*38fd1498Szrj {
489*38fd1498Szrj   struct gomp_thread *thr = gomp_thread ();
490*38fd1498Szrj   int test;
491*38fd1498Szrj 
492*38fd1498Szrj   gomp_ordered_sync ();
493*38fd1498Szrj   gomp_mutex_lock (&thr->ts.work_share->lock);
494*38fd1498Szrj   test = gomp_iter_ull_static_next (istart, iend);
495*38fd1498Szrj   if (test >= 0)
496*38fd1498Szrj     gomp_ordered_static_next ();
497*38fd1498Szrj   gomp_mutex_unlock (&thr->ts.work_share->lock);
498*38fd1498Szrj 
499*38fd1498Szrj   return test == 0;
500*38fd1498Szrj }
501*38fd1498Szrj 
502*38fd1498Szrj static bool
gomp_loop_ull_ordered_dynamic_next(gomp_ull * istart,gomp_ull * iend)503*38fd1498Szrj gomp_loop_ull_ordered_dynamic_next (gomp_ull *istart, gomp_ull *iend)
504*38fd1498Szrj {
505*38fd1498Szrj   struct gomp_thread *thr = gomp_thread ();
506*38fd1498Szrj   bool ret;
507*38fd1498Szrj 
508*38fd1498Szrj   gomp_ordered_sync ();
509*38fd1498Szrj   gomp_mutex_lock (&thr->ts.work_share->lock);
510*38fd1498Szrj   ret = gomp_iter_ull_dynamic_next_locked (istart, iend);
511*38fd1498Szrj   if (ret)
512*38fd1498Szrj     gomp_ordered_next ();
513*38fd1498Szrj   else
514*38fd1498Szrj     gomp_ordered_last ();
515*38fd1498Szrj   gomp_mutex_unlock (&thr->ts.work_share->lock);
516*38fd1498Szrj 
517*38fd1498Szrj   return ret;
518*38fd1498Szrj }
519*38fd1498Szrj 
520*38fd1498Szrj static bool
gomp_loop_ull_ordered_guided_next(gomp_ull * istart,gomp_ull * iend)521*38fd1498Szrj gomp_loop_ull_ordered_guided_next (gomp_ull *istart, gomp_ull *iend)
522*38fd1498Szrj {
523*38fd1498Szrj   struct gomp_thread *thr = gomp_thread ();
524*38fd1498Szrj   bool ret;
525*38fd1498Szrj 
526*38fd1498Szrj   gomp_ordered_sync ();
527*38fd1498Szrj   gomp_mutex_lock (&thr->ts.work_share->lock);
528*38fd1498Szrj   ret = gomp_iter_ull_guided_next_locked (istart, iend);
529*38fd1498Szrj   if (ret)
530*38fd1498Szrj     gomp_ordered_next ();
531*38fd1498Szrj   else
532*38fd1498Szrj     gomp_ordered_last ();
533*38fd1498Szrj   gomp_mutex_unlock (&thr->ts.work_share->lock);
534*38fd1498Szrj 
535*38fd1498Szrj   return ret;
536*38fd1498Szrj }
537*38fd1498Szrj 
538*38fd1498Szrj bool
GOMP_loop_ull_ordered_runtime_next(gomp_ull * istart,gomp_ull * iend)539*38fd1498Szrj GOMP_loop_ull_ordered_runtime_next (gomp_ull *istart, gomp_ull *iend)
540*38fd1498Szrj {
541*38fd1498Szrj   struct gomp_thread *thr = gomp_thread ();
542*38fd1498Szrj 
543*38fd1498Szrj   switch (thr->ts.work_share->sched)
544*38fd1498Szrj     {
545*38fd1498Szrj     case GFS_STATIC:
546*38fd1498Szrj     case GFS_AUTO:
547*38fd1498Szrj       return gomp_loop_ull_ordered_static_next (istart, iend);
548*38fd1498Szrj     case GFS_DYNAMIC:
549*38fd1498Szrj       return gomp_loop_ull_ordered_dynamic_next (istart, iend);
550*38fd1498Szrj     case GFS_GUIDED:
551*38fd1498Szrj       return gomp_loop_ull_ordered_guided_next (istart, iend);
552*38fd1498Szrj     default:
553*38fd1498Szrj       abort ();
554*38fd1498Szrj     }
555*38fd1498Szrj }
556*38fd1498Szrj 
557*38fd1498Szrj /* We use static functions above so that we're sure that the "runtime"
558*38fd1498Szrj    function can defer to the proper routine without interposition.  We
559*38fd1498Szrj    export the static function with a strong alias when possible, or with
560*38fd1498Szrj    a wrapper function otherwise.  */
561*38fd1498Szrj 
562*38fd1498Szrj #ifdef HAVE_ATTRIBUTE_ALIAS
563*38fd1498Szrj extern __typeof(gomp_loop_ull_static_start) GOMP_loop_ull_static_start
564*38fd1498Szrj 	__attribute__((alias ("gomp_loop_ull_static_start")));
565*38fd1498Szrj extern __typeof(gomp_loop_ull_dynamic_start) GOMP_loop_ull_dynamic_start
566*38fd1498Szrj 	__attribute__((alias ("gomp_loop_ull_dynamic_start")));
567*38fd1498Szrj extern __typeof(gomp_loop_ull_guided_start) GOMP_loop_ull_guided_start
568*38fd1498Szrj 	__attribute__((alias ("gomp_loop_ull_guided_start")));
569*38fd1498Szrj extern __typeof(gomp_loop_ull_dynamic_start) GOMP_loop_ull_nonmonotonic_dynamic_start
570*38fd1498Szrj 	__attribute__((alias ("gomp_loop_ull_dynamic_start")));
571*38fd1498Szrj extern __typeof(gomp_loop_ull_guided_start) GOMP_loop_ull_nonmonotonic_guided_start
572*38fd1498Szrj 	__attribute__((alias ("gomp_loop_ull_guided_start")));
573*38fd1498Szrj 
574*38fd1498Szrj extern __typeof(gomp_loop_ull_ordered_static_start) GOMP_loop_ull_ordered_static_start
575*38fd1498Szrj 	__attribute__((alias ("gomp_loop_ull_ordered_static_start")));
576*38fd1498Szrj extern __typeof(gomp_loop_ull_ordered_dynamic_start) GOMP_loop_ull_ordered_dynamic_start
577*38fd1498Szrj 	__attribute__((alias ("gomp_loop_ull_ordered_dynamic_start")));
578*38fd1498Szrj extern __typeof(gomp_loop_ull_ordered_guided_start) GOMP_loop_ull_ordered_guided_start
579*38fd1498Szrj 	__attribute__((alias ("gomp_loop_ull_ordered_guided_start")));
580*38fd1498Szrj 
581*38fd1498Szrj extern __typeof(gomp_loop_ull_doacross_static_start) GOMP_loop_ull_doacross_static_start
582*38fd1498Szrj 	__attribute__((alias ("gomp_loop_ull_doacross_static_start")));
583*38fd1498Szrj extern __typeof(gomp_loop_ull_doacross_dynamic_start) GOMP_loop_ull_doacross_dynamic_start
584*38fd1498Szrj 	__attribute__((alias ("gomp_loop_ull_doacross_dynamic_start")));
585*38fd1498Szrj extern __typeof(gomp_loop_ull_doacross_guided_start) GOMP_loop_ull_doacross_guided_start
586*38fd1498Szrj 	__attribute__((alias ("gomp_loop_ull_doacross_guided_start")));
587*38fd1498Szrj 
588*38fd1498Szrj extern __typeof(gomp_loop_ull_static_next) GOMP_loop_ull_static_next
589*38fd1498Szrj 	__attribute__((alias ("gomp_loop_ull_static_next")));
590*38fd1498Szrj extern __typeof(gomp_loop_ull_dynamic_next) GOMP_loop_ull_dynamic_next
591*38fd1498Szrj 	__attribute__((alias ("gomp_loop_ull_dynamic_next")));
592*38fd1498Szrj extern __typeof(gomp_loop_ull_guided_next) GOMP_loop_ull_guided_next
593*38fd1498Szrj 	__attribute__((alias ("gomp_loop_ull_guided_next")));
594*38fd1498Szrj extern __typeof(gomp_loop_ull_dynamic_next) GOMP_loop_ull_nonmonotonic_dynamic_next
595*38fd1498Szrj 	__attribute__((alias ("gomp_loop_ull_dynamic_next")));
596*38fd1498Szrj extern __typeof(gomp_loop_ull_guided_next) GOMP_loop_ull_nonmonotonic_guided_next
597*38fd1498Szrj 	__attribute__((alias ("gomp_loop_ull_guided_next")));
598*38fd1498Szrj 
599*38fd1498Szrj extern __typeof(gomp_loop_ull_ordered_static_next) GOMP_loop_ull_ordered_static_next
600*38fd1498Szrj 	__attribute__((alias ("gomp_loop_ull_ordered_static_next")));
601*38fd1498Szrj extern __typeof(gomp_loop_ull_ordered_dynamic_next) GOMP_loop_ull_ordered_dynamic_next
602*38fd1498Szrj 	__attribute__((alias ("gomp_loop_ull_ordered_dynamic_next")));
603*38fd1498Szrj extern __typeof(gomp_loop_ull_ordered_guided_next) GOMP_loop_ull_ordered_guided_next
604*38fd1498Szrj 	__attribute__((alias ("gomp_loop_ull_ordered_guided_next")));
605*38fd1498Szrj #else
606*38fd1498Szrj bool
GOMP_loop_ull_static_start(bool up,gomp_ull start,gomp_ull end,gomp_ull incr,gomp_ull chunk_size,gomp_ull * istart,gomp_ull * iend)607*38fd1498Szrj GOMP_loop_ull_static_start (bool up, gomp_ull start, gomp_ull end,
608*38fd1498Szrj 			    gomp_ull incr, gomp_ull chunk_size,
609*38fd1498Szrj 			    gomp_ull *istart, gomp_ull *iend)
610*38fd1498Szrj {
611*38fd1498Szrj   return gomp_loop_ull_static_start (up, start, end, incr, chunk_size, istart,
612*38fd1498Szrj 				     iend);
613*38fd1498Szrj }
614*38fd1498Szrj 
615*38fd1498Szrj bool
GOMP_loop_ull_dynamic_start(bool up,gomp_ull start,gomp_ull end,gomp_ull incr,gomp_ull chunk_size,gomp_ull * istart,gomp_ull * iend)616*38fd1498Szrj GOMP_loop_ull_dynamic_start (bool up, gomp_ull start, gomp_ull end,
617*38fd1498Szrj 			     gomp_ull incr, gomp_ull chunk_size,
618*38fd1498Szrj 			     gomp_ull *istart, gomp_ull *iend)
619*38fd1498Szrj {
620*38fd1498Szrj   return gomp_loop_ull_dynamic_start (up, start, end, incr, chunk_size, istart,
621*38fd1498Szrj 				      iend);
622*38fd1498Szrj }
623*38fd1498Szrj 
624*38fd1498Szrj bool
GOMP_loop_ull_guided_start(bool up,gomp_ull start,gomp_ull end,gomp_ull incr,gomp_ull chunk_size,gomp_ull * istart,gomp_ull * iend)625*38fd1498Szrj GOMP_loop_ull_guided_start (bool up, gomp_ull start, gomp_ull end,
626*38fd1498Szrj 			    gomp_ull incr, gomp_ull chunk_size,
627*38fd1498Szrj 			    gomp_ull *istart, gomp_ull *iend)
628*38fd1498Szrj {
629*38fd1498Szrj   return gomp_loop_ull_guided_start (up, start, end, incr, chunk_size, istart,
630*38fd1498Szrj 				     iend);
631*38fd1498Szrj }
632*38fd1498Szrj 
633*38fd1498Szrj bool
GOMP_loop_ull_nonmonotonic_dynamic_start(bool up,gomp_ull start,gomp_ull end,gomp_ull incr,gomp_ull chunk_size,gomp_ull * istart,gomp_ull * iend)634*38fd1498Szrj GOMP_loop_ull_nonmonotonic_dynamic_start (bool up, gomp_ull start,
635*38fd1498Szrj 					  gomp_ull end, gomp_ull incr,
636*38fd1498Szrj 					  gomp_ull chunk_size,
637*38fd1498Szrj 					  gomp_ull *istart, gomp_ull *iend)
638*38fd1498Szrj {
639*38fd1498Szrj   return gomp_loop_ull_dynamic_start (up, start, end, incr, chunk_size, istart,
640*38fd1498Szrj 				      iend);
641*38fd1498Szrj }
642*38fd1498Szrj 
643*38fd1498Szrj bool
GOMP_loop_ull_nonmonotonic_guided_start(bool up,gomp_ull start,gomp_ull end,gomp_ull incr,gomp_ull chunk_size,gomp_ull * istart,gomp_ull * iend)644*38fd1498Szrj GOMP_loop_ull_nonmonotonic_guided_start (bool up, gomp_ull start, gomp_ull end,
645*38fd1498Szrj 					 gomp_ull incr, gomp_ull chunk_size,
646*38fd1498Szrj 					 gomp_ull *istart, gomp_ull *iend)
647*38fd1498Szrj {
648*38fd1498Szrj   return gomp_loop_ull_guided_start (up, start, end, incr, chunk_size, istart,
649*38fd1498Szrj 				     iend);
650*38fd1498Szrj }
651*38fd1498Szrj 
652*38fd1498Szrj bool
GOMP_loop_ull_ordered_static_start(bool up,gomp_ull start,gomp_ull end,gomp_ull incr,gomp_ull chunk_size,gomp_ull * istart,gomp_ull * iend)653*38fd1498Szrj GOMP_loop_ull_ordered_static_start (bool up, gomp_ull start, gomp_ull end,
654*38fd1498Szrj 				    gomp_ull incr, gomp_ull chunk_size,
655*38fd1498Szrj 				    gomp_ull *istart, gomp_ull *iend)
656*38fd1498Szrj {
657*38fd1498Szrj   return gomp_loop_ull_ordered_static_start (up, start, end, incr, chunk_size,
658*38fd1498Szrj 					     istart, iend);
659*38fd1498Szrj }
660*38fd1498Szrj 
661*38fd1498Szrj bool
GOMP_loop_ull_ordered_dynamic_start(bool up,gomp_ull start,gomp_ull end,gomp_ull incr,gomp_ull chunk_size,gomp_ull * istart,gomp_ull * iend)662*38fd1498Szrj GOMP_loop_ull_ordered_dynamic_start (bool up, gomp_ull start, gomp_ull end,
663*38fd1498Szrj 				     gomp_ull incr, gomp_ull chunk_size,
664*38fd1498Szrj 				     gomp_ull *istart, gomp_ull *iend)
665*38fd1498Szrj {
666*38fd1498Szrj   return gomp_loop_ull_ordered_dynamic_start (up, start, end, incr, chunk_size,
667*38fd1498Szrj 					      istart, iend);
668*38fd1498Szrj }
669*38fd1498Szrj 
670*38fd1498Szrj bool
GOMP_loop_ull_ordered_guided_start(bool up,gomp_ull start,gomp_ull end,gomp_ull incr,gomp_ull chunk_size,gomp_ull * istart,gomp_ull * iend)671*38fd1498Szrj GOMP_loop_ull_ordered_guided_start (bool up, gomp_ull start, gomp_ull end,
672*38fd1498Szrj 				    gomp_ull incr, gomp_ull chunk_size,
673*38fd1498Szrj 				    gomp_ull *istart, gomp_ull *iend)
674*38fd1498Szrj {
675*38fd1498Szrj   return gomp_loop_ull_ordered_guided_start (up, start, end, incr, chunk_size,
676*38fd1498Szrj 					     istart, iend);
677*38fd1498Szrj }
678*38fd1498Szrj 
679*38fd1498Szrj bool
GOMP_loop_ull_doacross_static_start(unsigned ncounts,gomp_ull * counts,gomp_ull chunk_size,gomp_ull * istart,gomp_ull * iend)680*38fd1498Szrj GOMP_loop_ull_doacross_static_start (unsigned ncounts, gomp_ull *counts,
681*38fd1498Szrj 				     gomp_ull chunk_size, gomp_ull *istart,
682*38fd1498Szrj 				     gomp_ull *iend)
683*38fd1498Szrj {
684*38fd1498Szrj   return gomp_loop_ull_doacross_static_start (ncounts, counts, chunk_size,
685*38fd1498Szrj 					      istart, iend);
686*38fd1498Szrj }
687*38fd1498Szrj 
688*38fd1498Szrj bool
GOMP_loop_ull_doacross_dynamic_start(unsigned ncounts,gomp_ull * counts,gomp_ull chunk_size,gomp_ull * istart,gomp_ull * iend)689*38fd1498Szrj GOMP_loop_ull_doacross_dynamic_start (unsigned ncounts, gomp_ull *counts,
690*38fd1498Szrj 				      gomp_ull chunk_size, gomp_ull *istart,
691*38fd1498Szrj 				      gomp_ull *iend)
692*38fd1498Szrj {
693*38fd1498Szrj   return gomp_loop_ull_doacross_dynamic_start (ncounts, counts, chunk_size,
694*38fd1498Szrj 					       istart, iend);
695*38fd1498Szrj }
696*38fd1498Szrj 
697*38fd1498Szrj bool
GOMP_loop_ull_doacross_guided_start(unsigned ncounts,gomp_ull * counts,gomp_ull chunk_size,gomp_ull * istart,gomp_ull * iend)698*38fd1498Szrj GOMP_loop_ull_doacross_guided_start (unsigned ncounts, gomp_ull *counts,
699*38fd1498Szrj 				     gomp_ull chunk_size, gomp_ull *istart,
700*38fd1498Szrj 				     gomp_ull *iend)
701*38fd1498Szrj {
702*38fd1498Szrj   return gomp_loop_ull_doacross_guided_start (ncounts, counts, chunk_size,
703*38fd1498Szrj 					      istart, iend);
704*38fd1498Szrj }
705*38fd1498Szrj 
706*38fd1498Szrj bool
GOMP_loop_ull_static_next(gomp_ull * istart,gomp_ull * iend)707*38fd1498Szrj GOMP_loop_ull_static_next (gomp_ull *istart, gomp_ull *iend)
708*38fd1498Szrj {
709*38fd1498Szrj   return gomp_loop_ull_static_next (istart, iend);
710*38fd1498Szrj }
711*38fd1498Szrj 
712*38fd1498Szrj bool
GOMP_loop_ull_dynamic_next(gomp_ull * istart,gomp_ull * iend)713*38fd1498Szrj GOMP_loop_ull_dynamic_next (gomp_ull *istart, gomp_ull *iend)
714*38fd1498Szrj {
715*38fd1498Szrj   return gomp_loop_ull_dynamic_next (istart, iend);
716*38fd1498Szrj }
717*38fd1498Szrj 
718*38fd1498Szrj bool
GOMP_loop_ull_guided_next(gomp_ull * istart,gomp_ull * iend)719*38fd1498Szrj GOMP_loop_ull_guided_next (gomp_ull *istart, gomp_ull *iend)
720*38fd1498Szrj {
721*38fd1498Szrj   return gomp_loop_ull_guided_next (istart, iend);
722*38fd1498Szrj }
723*38fd1498Szrj 
724*38fd1498Szrj bool
GOMP_loop_ull_nonmonotonic_dynamic_next(gomp_ull * istart,gomp_ull * iend)725*38fd1498Szrj GOMP_loop_ull_nonmonotonic_dynamic_next (gomp_ull *istart, gomp_ull *iend)
726*38fd1498Szrj {
727*38fd1498Szrj   return gomp_loop_ull_dynamic_next (istart, iend);
728*38fd1498Szrj }
729*38fd1498Szrj 
730*38fd1498Szrj bool
GOMP_loop_ull_nonmonotonic_guided_next(gomp_ull * istart,gomp_ull * iend)731*38fd1498Szrj GOMP_loop_ull_nonmonotonic_guided_next (gomp_ull *istart, gomp_ull *iend)
732*38fd1498Szrj {
733*38fd1498Szrj   return gomp_loop_ull_guided_next (istart, iend);
734*38fd1498Szrj }
735*38fd1498Szrj 
736*38fd1498Szrj bool
GOMP_loop_ull_ordered_static_next(gomp_ull * istart,gomp_ull * iend)737*38fd1498Szrj GOMP_loop_ull_ordered_static_next (gomp_ull *istart, gomp_ull *iend)
738*38fd1498Szrj {
739*38fd1498Szrj   return gomp_loop_ull_ordered_static_next (istart, iend);
740*38fd1498Szrj }
741*38fd1498Szrj 
742*38fd1498Szrj bool
GOMP_loop_ull_ordered_dynamic_next(gomp_ull * istart,gomp_ull * iend)743*38fd1498Szrj GOMP_loop_ull_ordered_dynamic_next (gomp_ull *istart, gomp_ull *iend)
744*38fd1498Szrj {
745*38fd1498Szrj   return gomp_loop_ull_ordered_dynamic_next (istart, iend);
746*38fd1498Szrj }
747*38fd1498Szrj 
748*38fd1498Szrj bool
GOMP_loop_ull_ordered_guided_next(gomp_ull * istart,gomp_ull * iend)749*38fd1498Szrj GOMP_loop_ull_ordered_guided_next (gomp_ull *istart, gomp_ull *iend)
750*38fd1498Szrj {
751*38fd1498Szrj   return gomp_loop_ull_ordered_guided_next (istart, iend);
752*38fd1498Szrj }
753*38fd1498Szrj #endif
754