123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259 |
- /* Copyright (C) 2005-2022 Free Software Foundation, Inc.
- Contributed by Richard Henderson <rth@redhat.com>.
- This file is part of the GNU Offloading and Multi Processing Library
- (libgomp).
- Libgomp is free software; you can redistribute it and/or modify it
- under the terms of the GNU General Public License as published by
- the Free Software Foundation; either version 3, or (at your option)
- any later version.
- Libgomp is distributed in the hope that it will be useful, but WITHOUT ANY
- WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS
- FOR A PARTICULAR PURPOSE. See the GNU General Public License for
- more details.
- Under Section 7 of GPL version 3, you are granted additional
- permissions described in the GCC Runtime Library Exception, version
- 3.1, as published by the Free Software Foundation.
- You should have received a copy of the GNU General Public License and
- a copy of the GCC Runtime Library Exception along with this program;
- see the files COPYING3 and COPYING.RUNTIME respectively. If not, see
- <http://www.gnu.org/licenses/>. */
- /* This file handles the SECTIONS construct. */
- #include "libgomp.h"
- #include <string.h>
- ialias_redirect (GOMP_taskgroup_reduction_register)
- /* Initialize the given work share construct from the given arguments. */
- static inline void
- gomp_sections_init (struct gomp_work_share *ws, unsigned count)
- {
- ws->sched = GFS_DYNAMIC;
- ws->chunk_size = 1;
- ws->end = count + 1L;
- ws->incr = 1;
- ws->next = 1;
- #ifdef HAVE_SYNC_BUILTINS
- /* Prepare things to make each iteration faster. */
- if (sizeof (long) > sizeof (unsigned))
- ws->mode = 1;
- else
- {
- struct gomp_thread *thr = gomp_thread ();
- struct gomp_team *team = thr->ts.team;
- long nthreads = team ? team->nthreads : 1;
- ws->mode = ((nthreads | ws->end)
- < 1UL << (sizeof (long) * __CHAR_BIT__ / 2 - 1));
- }
- #else
- ws->mode = 0;
- #endif
- }
- /* This routine is called when first encountering a sections construct
- that is not bound directly to a parallel construct. The first thread
- that arrives will create the work-share construct; subsequent threads
- will see the construct exists and allocate work from it.
- COUNT is the number of sections in this construct.
- Returns the 1-based section number for this thread to perform, or 0 if
- all work was assigned to other threads prior to this thread's arrival. */
- unsigned
- GOMP_sections_start (unsigned count)
- {
- struct gomp_thread *thr = gomp_thread ();
- long s, e, ret;
- if (gomp_work_share_start (0))
- {
- gomp_sections_init (thr->ts.work_share, count);
- gomp_work_share_init_done ();
- }
- #ifdef HAVE_SYNC_BUILTINS
- if (gomp_iter_dynamic_next (&s, &e))
- ret = s;
- else
- ret = 0;
- #else
- gomp_mutex_lock (&thr->ts.work_share->lock);
- if (gomp_iter_dynamic_next_locked (&s, &e))
- ret = s;
- else
- ret = 0;
- gomp_mutex_unlock (&thr->ts.work_share->lock);
- #endif
- return ret;
- }
- unsigned
- GOMP_sections2_start (unsigned count, uintptr_t *reductions, void **mem)
- {
- struct gomp_thread *thr = gomp_thread ();
- long s, e, ret;
- if (reductions)
- gomp_workshare_taskgroup_start ();
- if (gomp_work_share_start (0))
- {
- gomp_sections_init (thr->ts.work_share, count);
- if (reductions)
- {
- GOMP_taskgroup_reduction_register (reductions);
- thr->task->taskgroup->workshare = true;
- thr->ts.work_share->task_reductions = reductions;
- }
- if (mem)
- {
- uintptr_t size = (uintptr_t) *mem;
- #define INLINE_ORDERED_TEAM_IDS_OFF \
- ((offsetof (struct gomp_work_share, inline_ordered_team_ids) \
- + __alignof__ (long long) - 1) & ~(__alignof__ (long long) - 1))
- if (size > (sizeof (struct gomp_work_share)
- - INLINE_ORDERED_TEAM_IDS_OFF))
- *mem
- = (void *) (thr->ts.work_share->ordered_team_ids
- = gomp_malloc_cleared (size));
- else
- *mem = memset (((char *) thr->ts.work_share)
- + INLINE_ORDERED_TEAM_IDS_OFF, '\0', size);
- }
- gomp_work_share_init_done ();
- }
- else
- {
- if (reductions)
- {
- uintptr_t *first_reductions = thr->ts.work_share->task_reductions;
- gomp_workshare_task_reduction_register (reductions,
- first_reductions);
- }
- if (mem)
- {
- if ((offsetof (struct gomp_work_share, inline_ordered_team_ids)
- & (__alignof__ (long long) - 1)) == 0)
- *mem = (void *) thr->ts.work_share->ordered_team_ids;
- else
- {
- uintptr_t p = (uintptr_t) thr->ts.work_share->ordered_team_ids;
- p += __alignof__ (long long) - 1;
- p &= ~(__alignof__ (long long) - 1);
- *mem = (void *) p;
- }
- }
- }
- #ifdef HAVE_SYNC_BUILTINS
- if (gomp_iter_dynamic_next (&s, &e))
- ret = s;
- else
- ret = 0;
- #else
- gomp_mutex_lock (&thr->ts.work_share->lock);
- if (gomp_iter_dynamic_next_locked (&s, &e))
- ret = s;
- else
- ret = 0;
- gomp_mutex_unlock (&thr->ts.work_share->lock);
- #endif
- return ret;
- }
- /* This routine is called when the thread completes processing of the
- section currently assigned to it. If the work-share construct is
- bound directly to a parallel construct, then the construct may have
- been set up before the parallel. In which case, this may be the
- first iteration for the thread.
- Returns the 1-based section number for this thread to perform, or 0 if
- all work was assigned to other threads prior to this thread's arrival. */
- unsigned
- GOMP_sections_next (void)
- {
- long s, e, ret;
- #ifdef HAVE_SYNC_BUILTINS
- if (gomp_iter_dynamic_next (&s, &e))
- ret = s;
- else
- ret = 0;
- #else
- struct gomp_thread *thr = gomp_thread ();
- gomp_mutex_lock (&thr->ts.work_share->lock);
- if (gomp_iter_dynamic_next_locked (&s, &e))
- ret = s;
- else
- ret = 0;
- gomp_mutex_unlock (&thr->ts.work_share->lock);
- #endif
- return ret;
- }
- /* This routine pre-initializes a work-share construct to avoid one
- synchronization once we get into the loop. */
- void
- GOMP_parallel_sections_start (void (*fn) (void *), void *data,
- unsigned num_threads, unsigned count)
- {
- struct gomp_team *team;
- num_threads = gomp_resolve_num_threads (num_threads, count);
- team = gomp_new_team (num_threads);
- gomp_sections_init (&team->work_shares[0], count);
- gomp_team_start (fn, data, num_threads, 0, team, NULL);
- }
- ialias_redirect (GOMP_parallel_end)
- void
- GOMP_parallel_sections (void (*fn) (void *), void *data,
- unsigned num_threads, unsigned count, unsigned flags)
- {
- struct gomp_team *team;
- num_threads = gomp_resolve_num_threads (num_threads, count);
- team = gomp_new_team (num_threads);
- gomp_sections_init (&team->work_shares[0], count);
- gomp_team_start (fn, data, num_threads, flags, team, NULL);
- fn (data);
- GOMP_parallel_end ();
- }
- /* The GOMP_section_end* routines are called after the thread is told
- that all sections are complete. The first two versions synchronize
- all threads; the nowait version does not. */
- void
- GOMP_sections_end (void)
- {
- gomp_work_share_end ();
- }
- bool
- GOMP_sections_end_cancel (void)
- {
- return gomp_work_share_end_cancel ();
- }
- void
- GOMP_sections_end_nowait (void)
- {
- gomp_work_share_end_nowait ();
- }
|