blob: 491af74da40f849a86e46f2dc42c6b8ddae1caf3 [file] [log] [blame]
/* Copyright (C) 2005-2022 Free Software Foundation, Inc.
Contributed by Richard Henderson <rth@redhat.com>.
This file is part of the GNU Offloading and Multi Processing Library
(libgomp).
Libgomp is free software; you can redistribute it and/or modify it
under the terms of the GNU General Public License as published by
the Free Software Foundation; either version 3, or (at your option)
any later version.
Libgomp is distributed in the hope that it will be useful, but WITHOUT ANY
WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS
FOR A PARTICULAR PURPOSE. See the GNU General Public License for
more details.
Under Section 7 of GPL version 3, you are granted additional
permissions described in the GCC Runtime Library Exception, version
3.1, as published by the Free Software Foundation.
You should have received a copy of the GNU General Public License and
a copy of the GCC Runtime Library Exception along with this program;
see the files COPYING3 and COPYING.RUNTIME respectively. If not, see
<http://www.gnu.org/licenses/>. */
/* This file contains routines for managing work-share iteration, both
for loops and sections. */
#include "libgomp.h"
#include <stdlib.h>
typedef unsigned long long gomp_ull;
/* This function implements the STATIC scheduling method. The caller should
iterate *pstart <= x < *pend. Return zero if there are more iterations
to perform; nonzero if not. Return less than 0 if this thread had
received the absolutely last iteration. */
int
gomp_iter_ull_static_next (gomp_ull *pstart, gomp_ull *pend)
{
struct gomp_thread *thr = gomp_thread ();
struct gomp_team *team = thr->ts.team;
struct gomp_work_share *ws = thr->ts.work_share;
unsigned long nthreads = team ? team->nthreads : 1;
if (thr->ts.static_trip == -1)
return -1;
/* Quick test for degenerate teams and orphaned constructs. */
if (nthreads == 1)
{
*pstart = ws->next_ull;
*pend = ws->end_ull;
thr->ts.static_trip = -1;
return ws->next_ull == ws->end_ull;
}
/* We interpret chunk_size zero as "unspecified", which means that we
should break up the iterations such that each thread makes only one
trip through the outer loop. */
if (ws->chunk_size_ull == 0)
{
gomp_ull n, q, i, t, s0, e0, s, e;
if (thr->ts.static_trip > 0)
return 1;
/* Compute the total number of iterations. */
if (__builtin_expect (ws->mode, 0) == 0)
n = (ws->end_ull - ws->next_ull + ws->incr_ull - 1) / ws->incr_ull;
else
n = (ws->next_ull - ws->end_ull - ws->incr_ull - 1) / -ws->incr_ull;
i = thr->ts.team_id;
/* Compute the "zero-based" start and end points. That is, as
if the loop began at zero and incremented by one. */
q = n / nthreads;
t = n % nthreads;
if (i < t)
{
t = 0;
q++;
}
s0 = q * i + t;
e0 = s0 + q;
/* Notice when no iterations allocated for this thread. */
if (s0 >= e0)
{
thr->ts.static_trip = 1;
return 1;
}
/* Transform these to the actual start and end numbers. */
s = s0 * ws->incr_ull + ws->next_ull;
e = e0 * ws->incr_ull + ws->next_ull;
*pstart = s;
*pend = e;
thr->ts.static_trip = (e0 == n ? -1 : 1);
return 0;
}
else
{
gomp_ull n, s0, e0, i, c, s, e;
/* Otherwise, each thread gets exactly chunk_size iterations
(if available) each time through the loop. */
if (__builtin_expect (ws->mode, 0) == 0)
n = (ws->end_ull - ws->next_ull + ws->incr_ull - 1) / ws->incr_ull;
else
n = (ws->next_ull - ws->end_ull - ws->incr_ull - 1) / -ws->incr_ull;
i = thr->ts.team_id;
c = ws->chunk_size_ull;
/* Initial guess is a C sized chunk positioned nthreads iterations
in, offset by our thread number. */
s0 = (thr->ts.static_trip * (gomp_ull) nthreads + i) * c;
e0 = s0 + c;
/* Detect overflow. */
if (s0 >= n)
return 1;
if (e0 > n)
e0 = n;
/* Transform these to the actual start and end numbers. */
s = s0 * ws->incr_ull + ws->next_ull;
e = e0 * ws->incr_ull + ws->next_ull;
*pstart = s;
*pend = e;
if (e0 == n)
thr->ts.static_trip = -1;
else
thr->ts.static_trip++;
return 0;
}
}
/* This function implements the DYNAMIC scheduling method. Arguments are
as for gomp_iter_ull_static_next. This function must be called with
ws->lock held. */
bool
gomp_iter_ull_dynamic_next_locked (gomp_ull *pstart, gomp_ull *pend)
{
struct gomp_thread *thr = gomp_thread ();
struct gomp_work_share *ws = thr->ts.work_share;
gomp_ull start, end, chunk, left;
start = ws->next_ull;
if (start == ws->end_ull)
return false;
chunk = ws->chunk_size_ull;
left = ws->end_ull - start;
if (__builtin_expect (ws->mode & 2, 0))
{
if (chunk < left)
chunk = left;
}
else
{
if (chunk > left)
chunk = left;
}
end = start + chunk;
ws->next_ull = end;
*pstart = start;
*pend = end;
return true;
}
#if defined HAVE_SYNC_BUILTINS && defined __LP64__
/* Similar, but doesn't require the lock held, and uses compare-and-swap
instead. Note that the only memory value that changes is ws->next_ull. */
bool
gomp_iter_ull_dynamic_next (gomp_ull *pstart, gomp_ull *pend)
{
struct gomp_thread *thr = gomp_thread ();
struct gomp_work_share *ws = thr->ts.work_share;
gomp_ull start, end, nend, chunk;
end = ws->end_ull;
chunk = ws->chunk_size_ull;
if (__builtin_expect (ws->mode & 1, 1))
{
gomp_ull tmp = __sync_fetch_and_add (&ws->next_ull, chunk);
if (__builtin_expect (ws->mode & 2, 0) == 0)
{
if (tmp >= end)
return false;
nend = tmp + chunk;
if (nend > end)
nend = end;
*pstart = tmp;
*pend = nend;
return true;
}
else
{
if (tmp <= end)
return false;
nend = tmp + chunk;
if (nend < end)
nend = end;
*pstart = tmp;
*pend = nend;
return true;
}
}
start = __atomic_load_n (&ws->next_ull, MEMMODEL_RELAXED);
while (1)
{
gomp_ull left = end - start;
gomp_ull tmp;
if (start == end)
return false;
if (__builtin_expect (ws->mode & 2, 0))
{
if (chunk < left)
chunk = left;
}
else
{
if (chunk > left)
chunk = left;
}
nend = start + chunk;
tmp = __sync_val_compare_and_swap (&ws->next_ull, start, nend);
if (__builtin_expect (tmp == start, 1))
break;
start = tmp;
}
*pstart = start;
*pend = nend;
return true;
}
#endif /* HAVE_SYNC_BUILTINS */
/* This function implements the GUIDED scheduling method. Arguments are
as for gomp_iter_ull_static_next. This function must be called with the
work share lock held. */
bool
gomp_iter_ull_guided_next_locked (gomp_ull *pstart, gomp_ull *pend)
{
struct gomp_thread *thr = gomp_thread ();
struct gomp_work_share *ws = thr->ts.work_share;
struct gomp_team *team = thr->ts.team;
gomp_ull nthreads = team ? team->nthreads : 1;
gomp_ull n, q;
gomp_ull start, end;
if (ws->next_ull == ws->end_ull)
return false;
start = ws->next_ull;
if (__builtin_expect (ws->mode, 0) == 0)
n = (ws->end_ull - start) / ws->incr_ull;
else
n = (start - ws->end_ull) / -ws->incr_ull;
q = (n + nthreads - 1) / nthreads;
if (q < ws->chunk_size_ull)
q = ws->chunk_size_ull;
if (q <= n)
end = start + q * ws->incr_ull;
else
end = ws->end_ull;
ws->next_ull = end;
*pstart = start;
*pend = end;
return true;
}
#if defined HAVE_SYNC_BUILTINS && defined __LP64__
/* Similar, but doesn't require the lock held, and uses compare-and-swap
instead. Note that the only memory value that changes is ws->next_ull. */
bool
gomp_iter_ull_guided_next (gomp_ull *pstart, gomp_ull *pend)
{
struct gomp_thread *thr = gomp_thread ();
struct gomp_work_share *ws = thr->ts.work_share;
struct gomp_team *team = thr->ts.team;
gomp_ull nthreads = team ? team->nthreads : 1;
gomp_ull start, end, nend, incr;
gomp_ull chunk_size;
start = __atomic_load_n (&ws->next_ull, MEMMODEL_RELAXED);
end = ws->end_ull;
incr = ws->incr_ull;
chunk_size = ws->chunk_size_ull;
while (1)
{
gomp_ull n, q;
gomp_ull tmp;
if (start == end)
return false;
if (__builtin_expect (ws->mode, 0) == 0)
n = (end - start) / incr;
else
n = (start - end) / -incr;
q = (n + nthreads - 1) / nthreads;
if (q < chunk_size)
q = chunk_size;
if (__builtin_expect (q <= n, 1))
nend = start + q * incr;
else
nend = end;
tmp = __sync_val_compare_and_swap (&ws->next_ull, start, nend);
if (__builtin_expect (tmp == start, 1))
break;
start = tmp;
}
*pstart = start;
*pend = nend;
return true;
}
#endif /* HAVE_SYNC_BUILTINS */