Home | History | Annotate | Line # | Download | only in linux
bar.h revision 1.12
      1  1.12  mrg /* Copyright (C) 2005-2022 Free Software Foundation, Inc.
      2   1.1  mrg    Contributed by Richard Henderson <rth (at) redhat.com>.
      3   1.1  mrg 
      4   1.5  mrg    This file is part of the GNU Offloading and Multi Processing Library
      5   1.5  mrg    (libgomp).
      6   1.1  mrg 
      7   1.1  mrg    Libgomp is free software; you can redistribute it and/or modify it
      8   1.1  mrg    under the terms of the GNU General Public License as published by
      9   1.1  mrg    the Free Software Foundation; either version 3, or (at your option)
     10   1.1  mrg    any later version.
     11   1.1  mrg 
     12   1.1  mrg    Libgomp is distributed in the hope that it will be useful, but WITHOUT ANY
     13   1.1  mrg    WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS
     14   1.1  mrg    FOR A PARTICULAR PURPOSE.  See the GNU General Public License for
     15   1.1  mrg    more details.
     16   1.1  mrg 
     17   1.1  mrg    Under Section 7 of GPL version 3, you are granted additional
     18   1.1  mrg    permissions described in the GCC Runtime Library Exception, version
     19   1.1  mrg    3.1, as published by the Free Software Foundation.
     20   1.1  mrg 
     21   1.1  mrg    You should have received a copy of the GNU General Public License and
     22   1.1  mrg    a copy of the GCC Runtime Library Exception along with this program;
     23   1.1  mrg    see the files COPYING3 and COPYING.RUNTIME respectively.  If not, see
     24   1.1  mrg    <http://www.gnu.org/licenses/>.  */
     25   1.1  mrg 
     26   1.1  mrg /* This is a Linux specific implementation of a barrier synchronization
     27   1.1  mrg    mechanism for libgomp.  This type is private to the library.  This
     28   1.1  mrg    implementation uses atomic instructions and the futex syscall.  */
     29   1.1  mrg 
     30   1.1  mrg #ifndef GOMP_BARRIER_H
     31   1.1  mrg #define GOMP_BARRIER_H 1
     32   1.1  mrg 
     33   1.1  mrg #include "mutex.h"
     34   1.1  mrg 
     35   1.1  mrg typedef struct
     36   1.1  mrg {
     37   1.1  mrg   /* Make sure total/generation is in a mostly read cacheline, while
     38   1.1  mrg      awaited in a separate cacheline.  */
     39   1.1  mrg   unsigned total __attribute__((aligned (64)));
     40   1.1  mrg   unsigned generation;
     41   1.1  mrg   unsigned awaited __attribute__((aligned (64)));
     42   1.5  mrg   unsigned awaited_final;
     43   1.1  mrg } gomp_barrier_t;
     44   1.5  mrg 
     45   1.1  mrg typedef unsigned int gomp_barrier_state_t;
     46   1.1  mrg 
     47   1.5  mrg /* The generation field contains a counter in the high bits, with a few
     48   1.5  mrg    low bits dedicated to flags.  Note that TASK_PENDING and WAS_LAST can
     49   1.5  mrg    share space because WAS_LAST is never stored back to generation.  */
     50   1.5  mrg #define BAR_TASK_PENDING	1
     51   1.5  mrg #define BAR_WAS_LAST		1
     52   1.5  mrg #define BAR_WAITING_FOR_TASK	2
     53   1.5  mrg #define BAR_CANCELLED		4
     54   1.5  mrg #define BAR_INCR		8
     55   1.5  mrg 
     56   1.1  mrg static inline void gomp_barrier_init (gomp_barrier_t *bar, unsigned count)
     57   1.1  mrg {
     58   1.1  mrg   bar->total = count;
     59   1.1  mrg   bar->awaited = count;
     60   1.5  mrg   bar->awaited_final = count;
     61   1.1  mrg   bar->generation = 0;
     62   1.1  mrg }
     63   1.1  mrg 
     64   1.1  mrg static inline void gomp_barrier_reinit (gomp_barrier_t *bar, unsigned count)
     65   1.1  mrg {
     66   1.3  mrg   __atomic_add_fetch (&bar->awaited, count - bar->total, MEMMODEL_ACQ_REL);
     67   1.1  mrg   bar->total = count;
     68   1.1  mrg }
     69   1.1  mrg 
     70   1.1  mrg static inline void gomp_barrier_destroy (gomp_barrier_t *bar)
     71   1.1  mrg {
     72   1.1  mrg }
     73   1.1  mrg 
     74   1.1  mrg extern void gomp_barrier_wait (gomp_barrier_t *);
     75   1.1  mrg extern void gomp_barrier_wait_last (gomp_barrier_t *);
     76   1.1  mrg extern void gomp_barrier_wait_end (gomp_barrier_t *, gomp_barrier_state_t);
     77   1.1  mrg extern void gomp_team_barrier_wait (gomp_barrier_t *);
     78   1.5  mrg extern void gomp_team_barrier_wait_final (gomp_barrier_t *);
     79   1.1  mrg extern void gomp_team_barrier_wait_end (gomp_barrier_t *,
     80   1.1  mrg 					gomp_barrier_state_t);
     81   1.5  mrg extern bool gomp_team_barrier_wait_cancel (gomp_barrier_t *);
     82   1.5  mrg extern bool gomp_team_barrier_wait_cancel_end (gomp_barrier_t *,
     83   1.5  mrg 					       gomp_barrier_state_t);
     84   1.1  mrg extern void gomp_team_barrier_wake (gomp_barrier_t *, int);
     85   1.5  mrg struct gomp_team;
     86   1.5  mrg extern void gomp_team_barrier_cancel (struct gomp_team *);
     87   1.1  mrg 
     88   1.1  mrg static inline gomp_barrier_state_t
     89   1.1  mrg gomp_barrier_wait_start (gomp_barrier_t *bar)
     90   1.1  mrg {
     91   1.5  mrg   unsigned int ret = __atomic_load_n (&bar->generation, MEMMODEL_ACQUIRE);
     92   1.5  mrg   ret &= -BAR_INCR | BAR_CANCELLED;
     93   1.3  mrg   /* A memory barrier is needed before exiting from the various forms
     94   1.3  mrg      of gomp_barrier_wait, to satisfy OpenMP API version 3.1 section
     95   1.3  mrg      2.8.6 flush Construct, which says there is an implicit flush during
     96   1.3  mrg      a barrier region.  This is a convenient place to add the barrier,
     97   1.3  mrg      so we use MEMMODEL_ACQ_REL here rather than MEMMODEL_ACQUIRE.  */
     98   1.5  mrg   if (__atomic_add_fetch (&bar->awaited, -1, MEMMODEL_ACQ_REL) == 0)
     99   1.5  mrg     ret |= BAR_WAS_LAST;
    100   1.5  mrg   return ret;
    101   1.5  mrg }
    102   1.5  mrg 
    103   1.5  mrg static inline gomp_barrier_state_t
    104   1.5  mrg gomp_barrier_wait_cancel_start (gomp_barrier_t *bar)
    105   1.5  mrg {
    106   1.5  mrg   return gomp_barrier_wait_start (bar);
    107   1.5  mrg }
    108   1.5  mrg 
    109   1.5  mrg /* This is like gomp_barrier_wait_start, except it decrements
    110   1.5  mrg    bar->awaited_final rather than bar->awaited and should be used
    111   1.5  mrg    for the gomp_team_end barrier only.  */
    112   1.5  mrg static inline gomp_barrier_state_t
    113   1.5  mrg gomp_barrier_wait_final_start (gomp_barrier_t *bar)
    114   1.5  mrg {
    115   1.5  mrg   unsigned int ret = __atomic_load_n (&bar->generation, MEMMODEL_ACQUIRE);
    116   1.5  mrg   ret &= -BAR_INCR | BAR_CANCELLED;
    117   1.5  mrg   /* See above gomp_barrier_wait_start comment.  */
    118   1.5  mrg   if (__atomic_add_fetch (&bar->awaited_final, -1, MEMMODEL_ACQ_REL) == 0)
    119   1.5  mrg     ret |= BAR_WAS_LAST;
    120   1.1  mrg   return ret;
    121   1.1  mrg }
    122   1.1  mrg 
    123   1.1  mrg static inline bool
    124   1.1  mrg gomp_barrier_last_thread (gomp_barrier_state_t state)
    125   1.1  mrg {
    126   1.5  mrg   return state & BAR_WAS_LAST;
    127   1.1  mrg }
    128   1.1  mrg 
    129   1.1  mrg /* All the inlines below must be called with team->task_lock
    130   1.1  mrg    held.  */
    131   1.1  mrg 
    132   1.1  mrg static inline void
    133   1.1  mrg gomp_team_barrier_set_task_pending (gomp_barrier_t *bar)
    134   1.1  mrg {
    135   1.5  mrg   bar->generation |= BAR_TASK_PENDING;
    136   1.1  mrg }
    137   1.1  mrg 
    138   1.1  mrg static inline void
    139   1.1  mrg gomp_team_barrier_clear_task_pending (gomp_barrier_t *bar)
    140   1.1  mrg {
    141   1.5  mrg   bar->generation &= ~BAR_TASK_PENDING;
    142   1.1  mrg }
    143   1.1  mrg 
    144   1.1  mrg static inline void
    145   1.1  mrg gomp_team_barrier_set_waiting_for_tasks (gomp_barrier_t *bar)
    146   1.1  mrg {
    147   1.5  mrg   bar->generation |= BAR_WAITING_FOR_TASK;
    148   1.1  mrg }
    149   1.1  mrg 
    150   1.1  mrg static inline bool
    151   1.1  mrg gomp_team_barrier_waiting_for_tasks (gomp_barrier_t *bar)
    152   1.1  mrg {
    153   1.5  mrg   return (bar->generation & BAR_WAITING_FOR_TASK) != 0;
    154   1.5  mrg }
    155   1.5  mrg 
    156   1.5  mrg static inline bool
    157   1.5  mrg gomp_team_barrier_cancelled (gomp_barrier_t *bar)
    158   1.5  mrg {
    159   1.5  mrg   return __builtin_expect ((bar->generation & BAR_CANCELLED) != 0, 0);
    160   1.1  mrg }
    161   1.1  mrg 
    162   1.1  mrg static inline void
    163   1.1  mrg gomp_team_barrier_done (gomp_barrier_t *bar, gomp_barrier_state_t state)
    164   1.1  mrg {
    165   1.5  mrg   bar->generation = (state & -BAR_INCR) + BAR_INCR;
    166   1.1  mrg }
    167   1.1  mrg 
    168   1.1  mrg #endif /* GOMP_BARRIER_H */
    169