cuda::barrier

Defined in header <cuda/barrier>:

template <cuda::thread_scope Scope,
          typename CompletionFunction = /* unspecified */>
class cuda::barrier;

The class template cuda::barrier is an extended form of cuda::std::barrier that takes an additional cuda::thread_scope argument.

If !(scope == thread_block_scope && __isShared(this)), then the semantics are the same as cuda::std::barrier, otherwise, see below.

The cuda::barrier class templates extends cuda::std::barrier with the following additional operations:

cuda::barrier::init

Initialize a cuda::barrier.

cuda::device::barrier_native_handle

Get the native handle to a cuda::barrier.

cuda::device::barrier_arrive_tx

Arrive on a cuda::barrier<cuda::thread_scope_block> with transaction count update.

cuda::device::barrier_expect_tx

Update transaction count of cuda::barrier<cuda::thread_scope_block>.

If scope == thread_scope_block && __isShared(this), then the semantics of [thread.barrier.class] of ISO/IEC IS 14882 (the C++ Standard) are modified as follows:

A barrier is a thread coordination mechanism whose lifetime consists of a sequence of barrier phases, where each phase allows at most an expected number of threads to block until the expected number of threads and the expected number of transaction-based asynchronous operations arrive at the barrier. Each barrier phase consists of the following steps:

  1. The expected count is decremented by each call to arrive,arrive_and_drop, or cuda::device::barrier_arrive_tx.

  2. The transaction count is incremented by each call to cuda::device::barrier_arrive_tx and decremented by the completion of transaction-based asynchronous operations such as cuda::memcpy_async_tx.

  3. Exactly once after both the expected count and the transaction count reach zero, a thread executes the completion step during its call to arrive, arrive_and_drop, cuda::device::barrier_arrive_tx, or wait, except that it is implementation-defined whether the step executes if no thread calls wait.

  4. When the completion step finishes, the expected count is reset to what was specified by the expected argument to the constructor, possibly adjusted by calls to arrive_and_drop, and the next phase starts.

Concurrent invocations of the member functions of barrier and the non-member barrier APIs in cuda::device, other than its destructor, do not introduce data races. The member functions arrive and arrive_and_drop, and the non-member function cuda::device::barrier_arrive_tx, execute atomically.

NVCC __shared__ Initialization Warnings

When using libcu++ with NVCC, a __shared__ cuda::barrier will lead to the following warning because __shared__ variables are not initialized:

warning: dynamic initialization is not supported for a function-scope static
__shared__ variable within a __device__/__global__ function

It can be silenced using #pragma nv_diag_suppress static_var_with_dynamic_init.

To properly initialize a __shared__ cuda::barrier, use the cuda::barrier::init friend function.

Concurrency Restrictions

An object of type cuda::barrier or cuda::std::barrier shall not be accessed concurrently by CPU and GPU threads unless:

Note, for objects of scopes other than cuda::thread_scope_system this is a data-race, and therefore also prohibited regardless of memory characteristics.

Under CUDA Compute Capability 8 (Ampere) or above, when an object of type cuda::barrier<thread_scope_block> is placed in __shared__ memory, the member function arrive performs a reduction of the arrival count among coalesced threads followed by the arrival operation in one thread. Programs shall ensure that this transformation would not introduce errors, for example relative to the requirements of thread.barrier.class paragraph 12 of ISO/IEC IS 14882 (the C++ Standard).

Under CUDA Compute Capability 6 (Pascal) or prior, an object of type cuda::barrier or cuda::std::barrier may not be used.

Shared memory barriers with transaction count

In addition to the arrival count, a cuda::barrier<thread_scope_block> object located in shared memory supports a tx-count, which is used for tracking the completion of some asynchronous memory operations or transactions. The tx-count tracks the number of asynchronous transactions, in units specified by the asynchronous memory operation (typically bytes), that are outstanding and yet to be complete. This capability is exposed, starting with the Hopper architecture (CUDA Compute Capability 9).

The tx-count of cuda::barrier must be set to the total amount of asynchronous memory operations, in units as specified by the asynchronous operations, to be tracked by the current phase. This can be achieved with the cuda::device::barrier_arrive_tx function call.

Upon completion of each of the asynchronous operations, the tx-count of the cuda::barrier will be updated and thus progress the cuda::barrier towards the completion of the current phase. This may complete the current phase.

Implementation-Defined Behavior

For each cuda::thread_scope S and CompletionFunction F, the value of cuda::barrier<S, F>::max() is as follows:

cuda::thread_scope S

CompletionFunction F

barrier<S, F>::max()

cuda::thread_scope_block

Default or user-provided

(1 << 20) - 1

Not cuda::thread_scope_block

Default

cuda::std::numeric_limits<cuda::std::int32_t>::max()

Not cuda::thread_scope_block

User-provided

cuda::std::numeric_limits<cuda::std::ptrdiff_t>::max()

Example

#include <cuda/barrier>

__global__ void example_kernel() {
  // This barrier is suitable for all threads in the system.
  cuda::barrier<cuda::thread_scope_system> a(10);

  // This barrier has the same type as the previous one (`a`).
  cuda::std::barrier<> b(10);

  // This barrier is suitable for all threads on the current processor (e.g. GPU).
  cuda::barrier<cuda::thread_scope_device> c(10);

  // This barrier is suitable for all threads in the same thread block.
  cuda::barrier<cuda::thread_scope_block> d(10);
}

See it on Godbolt