| 1 | // <barrier> -*- C++ -*- | 
 
 
 
 
 | 2 |  | 
 
 
 
 
 | 3 | // Copyright (C) 2020-2021 Free Software Foundation, Inc. | 
 
 
 
 
 | 4 | // | 
 
 
 
 
 | 5 | // This file is part of the GNU ISO C++ Library.  This library is free | 
 
 
 
 
 | 6 | // software; you can redistribute it and/or modify it under the | 
 
 
 
 
 | 7 | // terms of the GNU General Public License as published by the | 
 
 
 
 
 | 8 | // Free Software Foundation; either version 3, or (at your option) | 
 
 
 
 
 | 9 | // any later version. | 
 
 
 
 
 | 10 |  | 
 
 
 
 
 | 11 | // This library is distributed in the hope that it will be useful, | 
 
 
 
 
 | 12 | // but WITHOUT ANY WARRANTY; without even the implied warranty of | 
 
 
 
 
 | 13 | // MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the | 
 
 
 
 
 | 14 | // GNU General Public License for more details. | 
 
 
 
 
 | 15 |  | 
 
 
 
 
 | 16 | // You should have received a copy of the GNU General Public License along | 
 
 
 
 
 | 17 | // with this library; see the file COPYING3.  If not see | 
 
 
 
 
 | 18 | // <http://www.gnu.org/licenses/>. | 
 
 
 
 
 | 19 |  | 
 
 
 
 
 | 20 | // This implementation is based on libcxx/include/barrier | 
 
 
 
 
 | 21 | //===-- barrier.h --------------------------------------------------===// | 
 
 
 
 
 | 22 | // | 
 
 
 
 
 | 23 | // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. | 
 
 
 
 
 | 24 | // See https://llvm.org/LICENSE.txt for license information. | 
 
 
 
 
 | 25 | // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception | 
 
 
 
 
 | 26 | // | 
 
 
 
 
 | 27 | //===---------------------------------------------------------------===// | 
 
 
 
 
 | 28 |  | 
 
 
 
 
 | 29 | /** @file include/barrier | 
 
 
 
 
 | 30 | *  This is a Standard C++ Library header. | 
 
 
 
 
 | 31 | */ | 
 
 
 
 
 | 32 |  | 
 
 
 
 
 | 33 | #ifndef _GLIBCXX_BARRIER | 
 
 
 
 
 | 34 | #define _GLIBCXX_BARRIER 1 | 
 
 
 
 
 | 35 |  | 
 
 
 
 
 | 36 | #pragma GCC system_header | 
 
 
 
 
 | 37 |  | 
 
 
 
 
 | 38 | #if __cplusplus > 201703L | 
 
 
 
 
 | 39 | #include <bits/atomic_base.h> | 
 
 
 
 
 | 40 | #if __cpp_lib_atomic_wait && __cpp_aligned_new | 
 
 
 
 
 | 41 | #include <bits/std_thread.h> | 
 
 
 
 
 | 42 | #include <bits/unique_ptr.h> | 
 
 
 
 
 | 43 |  | 
 
 
 
 
 | 44 | #include <array> | 
 
 
 
 
 | 45 |  | 
 
 
 
 
 | 46 | #define __cpp_lib_barrier 201907L | 
 
 
 
 
 | 47 |  | 
 
 
 
 
 | 48 | namespace std _GLIBCXX_VISIBILITY(default) | 
 
 
 
 
 | 49 | { | 
 
 
 
 
 | 50 | _GLIBCXX_BEGIN_NAMESPACE_VERSION | 
 
 
 
 
 | 51 |  | 
 
 
 
 
 | 52 | struct __empty_completion | 
 
 
 
 
 | 53 | { | 
 
 
 
 
 | 54 | _GLIBCXX_ALWAYS_INLINE void | 
 
 
 
 
 | 55 | operator()() noexcept | 
 
 
 
 
 | 56 | { } | 
 
 
 
 
 | 57 | }; | 
 
 
 
 
 | 58 |  | 
 
 
 
 
 | 59 | /* | 
 
 
 
 
 | 60 |  | 
 
 
 
 
 | 61 | The default implementation of __tree_barrier is a classic tree barrier. | 
 
 
 
 
 | 62 |  | 
 
 
 
 
 | 63 | It looks different from literature pseudocode for two main reasons: | 
 
 
 
 
 | 64 | 1. Threads that call into std::barrier functions do not provide indices, | 
 
 
 
 
 | 65 | so a numbering step is added before the actual barrier algorithm, | 
 
 
 
 
 | 66 | appearing as an N+1 round to the N rounds of the tree barrier. | 
 
 
 
 
 | 67 | 2. A great deal of attention has been paid to avoid cache line thrashing | 
 
 
 
 
 | 68 | by flattening the tree structure into cache-line sized arrays, that | 
 
 
 
 
 | 69 | are indexed in an efficient way. | 
 
 
 
 
 | 70 |  | 
 
 
 
 
 | 71 | */ | 
 
 
 
 
 | 72 |  | 
 
 
 
 
 | 73 | enum class __barrier_phase_t : unsigned char { }; | 
 
 
 
 
 | 74 |  | 
 
 
 
 
 | 75 | template<typename _CompletionF> | 
 
 
 
 
 | 76 | class __tree_barrier | 
 
 
 
 
 | 77 | { | 
 
 
 
 
 | 78 | using __atomic_phase_ref_t = std::__atomic_ref<__barrier_phase_t>; | 
 
 
 
 
 | 79 | using __atomic_phase_const_ref_t = std::__atomic_ref<const __barrier_phase_t>; | 
 
 
 
 
 | 80 | static constexpr auto __phase_alignment = | 
 
 
 
 
 | 81 | __atomic_phase_ref_t::required_alignment; | 
 
 
 
 
 | 82 |  | 
 
 
 
 
 | 83 | using __tickets_t = std::array<__barrier_phase_t, 64>; | 
 
 
 
 
 | 84 | struct alignas(64) /* naturally-align the heap state */ __state_t | 
 
 
 
 
 | 85 | { | 
 
 
 
 
 | 86 | alignas(__phase_alignment) __tickets_t __tickets; | 
 
 
 
 
 | 87 | }; | 
 
 
 
 
 | 88 |  | 
 
 
 
 
 | 89 | ptrdiff_t _M_expected; | 
 
 
 
 
 | 90 | unique_ptr<__state_t[]> _M_state; | 
 
 
 
 
 | 91 | __atomic_base<ptrdiff_t> _M_expected_adjustment; | 
 
 
 
 
 | 92 | _CompletionF _M_completion; | 
 
 
 
 
 | 93 |  | 
 
 
 
 
 | 94 | alignas(__phase_alignment) __barrier_phase_t  _M_phase; | 
 
 
 
 
 | 95 |  | 
 
 
 
 
 | 96 | bool | 
 
 
 
 
 | 97 | _M_arrive(__barrier_phase_t __old_phase, size_t __current) | 
 
 
 
 
 | 98 | { | 
 
 
 
 
 | 99 | const auto __old_phase_val = static_cast<unsigned char>(__old_phase); | 
 
 
 
 
 | 100 | const auto __half_step = | 
 
 
 
 
 | 101 | static_cast<__barrier_phase_t>(__old_phase_val + 1); | 
 
 
 
 
 | 102 | const auto __full_step = | 
 
 
 
 
 | 103 | static_cast<__barrier_phase_t>(__old_phase_val + 2); | 
 
 
 
 
 | 104 |  | 
 
 
 
 
 | 105 | size_t __current_expected = _M_expected; | 
 
 
 
 
 | 106 | __current %= ((_M_expected + 1) >> 1); | 
 
 
 
 
 | 107 |  | 
 
 
 
 
 | 108 | for (int __round = 0; ; ++__round) | 
 
 
 
 
 | 109 | { | 
 
 
 
 
 | 110 | if (__current_expected <= 1) | 
 
 
 
 
 | 111 | return true; | 
 
 
 
 
 | 112 | size_t const __end_node = ((__current_expected + 1) >> 1), | 
 
 
 
 
 | 113 | __last_node = __end_node - 1; | 
 
 
 
 
 | 114 | for ( ; ; ++__current) | 
 
 
 
 
 | 115 | { | 
 
 
 
 
 | 116 | if (__current == __end_node) | 
 
 
 
 
 | 117 | __current = 0; | 
 
 
 
 
 | 118 | auto __expect = __old_phase; | 
 
 
 
 
 | 119 | __atomic_phase_ref_t __phase(_M_state[__current] | 
 
 
 
 
 | 120 | .__tickets[__round]); | 
 
 
 
 
 | 121 | if (__current == __last_node && (__current_expected & 1)) | 
 
 
 
 
 | 122 | { | 
 
 
 
 
 | 123 | if (__phase.compare_exchange_strong(__expect, __full_step, | 
 
 
 
 
 | 124 | memory_order_acq_rel)) | 
 
 
 
 
 | 125 | break;     // I'm 1 in 1, go to next __round | 
 
 
 
 
 | 126 | } | 
 
 
 
 
 | 127 | else if (__phase.compare_exchange_strong(__expect, __half_step, | 
 
 
 
 
 | 128 | memory_order_acq_rel)) | 
 
 
 
 
 | 129 | { | 
 
 
 
 
 | 130 | return false; // I'm 1 in 2, done with arrival | 
 
 
 
 
 | 131 | } | 
 
 
 
 
 | 132 | else if (__expect == __half_step) | 
 
 
 
 
 | 133 | { | 
 
 
 
 
 | 134 | if (__phase.compare_exchange_strong(__expect, __full_step, | 
 
 
 
 
 | 135 | memory_order_acq_rel)) | 
 
 
 
 
 | 136 | break;    // I'm 2 in 2, go to next __round | 
 
 
 
 
 | 137 | } | 
 
 
 
 
 | 138 | } | 
 
 
 
 
 | 139 | __current_expected = __last_node + 1; | 
 
 
 
 
 | 140 | __current >>= 1; | 
 
 
 
 
 | 141 | } | 
 
 
 
 
 | 142 | } | 
 
 
 
 
 | 143 |  | 
 
 
 
 
 | 144 | public: | 
 
 
 
 
 | 145 | using arrival_token = __barrier_phase_t; | 
 
 
 
 
 | 146 |  | 
 
 
 
 
 | 147 | static constexpr ptrdiff_t | 
 
 
 
 
 | 148 | max() noexcept | 
 
 
 
 
 | 149 | { return __PTRDIFF_MAX__; } | 
 
 
 
 
 | 150 |  | 
 
 
 
 
 | 151 | __tree_barrier(ptrdiff_t __expected, _CompletionF __completion) | 
 
 
 
 
 | 152 | : _M_expected(__expected), _M_expected_adjustment(0), | 
 
 
 
 
 | 153 | _M_completion(move(__completion)), | 
 
 
 
 
 | 154 | _M_phase(static_cast<__barrier_phase_t>(0)) | 
 
 
 
 
 | 155 | { | 
 
 
 
 
 | 156 | size_t const __count = (_M_expected + 1) >> 1; | 
 
 
 
 
 | 157 |  | 
 
 
 
 
 | 158 | _M_state = std::make_unique<__state_t[]>(__count); | 
 
 
 
 
 | 159 | } | 
 
 
 
 
 | 160 |  | 
 
 
 
 
 | 161 | [[nodiscard]] arrival_token | 
 
 
 
 
 | 162 | arrive(ptrdiff_t __update) | 
 
 
 
 
 | 163 | { | 
 
 
 
 
 | 164 | std::hash<std::thread::id> __hasher; | 
 
 
 
 
 | 165 | size_t __current = __hasher(std::this_thread::get_id()); | 
 
 
 
 
 | 166 | __atomic_phase_ref_t __phase(_M_phase); | 
 
 
 
 
 | 167 | const auto __old_phase = __phase.load(memory_order_relaxed); | 
 
 
 
 
 | 168 | const auto __cur = static_cast<unsigned char>(__old_phase); | 
 
 
 
 
 | 169 | for(; __update; --__update) | 
 
 
 
 
 | 170 | { | 
 
 
 
 
 | 171 | if(_M_arrive(__old_phase, __current)) | 
 
 
 
 
 | 172 | { | 
 
 
 
 
 | 173 | _M_completion(); | 
 
 
 
 
 | 174 | _M_expected += _M_expected_adjustment.load(memory_order_relaxed); | 
 
 
 
 
 | 175 | _M_expected_adjustment.store(0, memory_order_relaxed); | 
 
 
 
 
 | 176 | auto __new_phase = static_cast<__barrier_phase_t>(__cur + 2); | 
 
 
 
 
 | 177 | __phase.store(__new_phase, memory_order_release); | 
 
 
 
 
 | 178 | __phase.notify_all(); | 
 
 
 
 
 | 179 | } | 
 
 
 
 
 | 180 | } | 
 
 
 
 
 | 181 | return __old_phase; | 
 
 
 
 
 | 182 | } | 
 
 
 
 
 | 183 |  | 
 
 
 
 
 | 184 | void | 
 
 
 
 
 | 185 | wait(arrival_token&& __old_phase) const | 
 
 
 
 
 | 186 | { | 
 
 
 
 
 | 187 | __atomic_phase_const_ref_t __phase(_M_phase); | 
 
 
 
 
 | 188 | auto const __test_fn = [=] | 
 
 
 
 
 | 189 | { | 
 
 
 
 
 | 190 | return __phase.load(memory_order_acquire) != __old_phase; | 
 
 
 
 
 | 191 | }; | 
 
 
 
 
 | 192 | std::__atomic_wait_address(&_M_phase, __test_fn); | 
 
 
 
 
 | 193 | } | 
 
 
 
 
 | 194 |  | 
 
 
 
 
 | 195 | void | 
 
 
 
 
 | 196 | arrive_and_drop() | 
 
 
 
 
 | 197 | { | 
 
 
 
 
 | 198 | _M_expected_adjustment.fetch_sub(1, memory_order_relaxed); | 
 
 
 
 
 | 199 | (void)arrive(1); | 
 
 
 
 
 | 200 | } | 
 
 
 
 
 | 201 | }; | 
 
 
 
 
 | 202 |  | 
 
 
 
 
 | 203 | template<typename _CompletionF = __empty_completion> | 
 
 
 
 
 | 204 | class barrier | 
 
 
 
 
 | 205 | { | 
 
 
 
 
 | 206 | // Note, we may introduce a "central" barrier algorithm at some point | 
 
 
 
 
 | 207 | // for more space constrained targets | 
 
 
 
 
 | 208 | using __algorithm_t = __tree_barrier<_CompletionF>; | 
 
 
 
 
 | 209 | __algorithm_t _M_b; | 
 
 
 
 
 | 210 |  | 
 
 
 
 
 | 211 | public: | 
 
 
 
 
 | 212 | class arrival_token final | 
 
 
 
 
 | 213 | { | 
 
 
 
 
 | 214 | public: | 
 
 
 
 
 | 215 | arrival_token(arrival_token&&) = default; | 
 
 
 
 
 | 216 | arrival_token& operator=(arrival_token&&) = default; | 
 
 
 
 
 | 217 | ~arrival_token() = default; | 
 
 
 
 
 | 218 |  | 
 
 
 
 
 | 219 | private: | 
 
 
 
 
 | 220 | friend class barrier; | 
 
 
 
 
 | 221 | using __token = typename __algorithm_t::arrival_token; | 
 
 
 
 
 | 222 | explicit arrival_token(__token __tok) noexcept : _M_tok(__tok) { } | 
 
 
 
 
 | 223 | __token _M_tok; | 
 
 
 
 
 | 224 | }; | 
 
 
 
 
 | 225 |  | 
 
 
 
 
 | 226 | static constexpr ptrdiff_t | 
 
 
 
 
 | 227 | max() noexcept | 
 
 
 
 
 | 228 | { return __algorithm_t::max(); } | 
 
 
 
 
 | 229 |  | 
 
 
 
 
 | 230 | explicit | 
 
 
 
 
 | 231 | barrier(ptrdiff_t __count, _CompletionF __completion = _CompletionF()) | 
 
 
 
 
 | 232 | : _M_b(__count, std::move(__completion)) | 
 
 
 
 
 | 233 | { } | 
 
 
 
 
 | 234 |  | 
 
 
 
 
 | 235 | barrier(barrier const&) = delete; | 
 
 
 
 
 | 236 | barrier& operator=(barrier const&) = delete; | 
 
 
 
 
 | 237 |  | 
 
 
 
 
 | 238 | [[nodiscard]] arrival_token | 
 
 
 
 
 | 239 | arrive(ptrdiff_t __update = 1) | 
 
 
 
 
 | 240 | { return arrival_token{_M_b.arrive(__update)}; } | 
 
 
 
 
 | 241 |  | 
 
 
 
 
 | 242 | void | 
 
 
 
 
 | 243 | wait(arrival_token&& __phase) const | 
 
 
 
 
 | 244 | { _M_b.wait(std::move(__phase._M_tok)); } | 
 
 
 
 
 | 245 |  | 
 
 
 
 
 | 246 | void | 
 
 
 
 
 | 247 | arrive_and_wait() | 
 
 
 
 
 | 248 | { wait(arrive()); } | 
 
 
 
 
 | 249 |  | 
 
 
 
 
 | 250 | void | 
 
 
 
 
 | 251 | arrive_and_drop() | 
 
 
 
 
 | 252 | { _M_b.arrive_and_drop(); } | 
 
 
 
 
 | 253 | }; | 
 
 
 
 
 | 254 |  | 
 
 
 
 
 | 255 | _GLIBCXX_END_NAMESPACE_VERSION | 
 
 
 
 
 | 256 | } // namespace | 
 
 
 
 
 | 257 | #endif // __cpp_lib_atomic_wait && __cpp_aligned_new | 
 
 
 
 
 | 258 | #endif // __cplusplus > 201703L | 
 
 
 
 
 | 259 | #endif // _GLIBCXX_BARRIER |