st#
PTX ISA: st
st.global.b8#
// st.space.b8 [addr], src; // PTX ISA 10, SM_50
// .space = { .global }
template <typename B8, enable_if_t<sizeof(B8) == 1, bool> = true>
__device__ static inline void st(
cuda::ptx::space_global_t,
B8* addr,
B8 src);
st.global.b16#
// st.space.b16 [addr], src; // PTX ISA 10, SM_50
// .space = { .global }
template <typename B16, enable_if_t<sizeof(B16) == 2, bool> = true>
__device__ static inline void st(
cuda::ptx::space_global_t,
B16* addr,
B16 src);
st.global.b32#
// st.space.b32 [addr], src; // PTX ISA 10, SM_50
// .space = { .global }
template <typename B32, enable_if_t<sizeof(B32) == 4, bool> = true>
__device__ static inline void st(
cuda::ptx::space_global_t,
B32* addr,
B32 src);
st.global.b64#
// st.space.b64 [addr], src; // PTX ISA 10, SM_50
// .space = { .global }
template <typename B64, enable_if_t<sizeof(B64) == 8, bool> = true>
__device__ static inline void st(
cuda::ptx::space_global_t,
B64* addr,
B64 src);
st.global.b128#
// st.space.b128 [addr], src; // PTX ISA 83, SM_70
// .space = { .global }
template <typename B128, enable_if_t<sizeof(B128) == 16, bool> = true>
__device__ static inline void st(
cuda::ptx::space_global_t,
B128* addr,
B128 src);
st.global.v4.b64#
// st.space.v4.b64 [addr], src; // PTX ISA 88, SM_100
// .space = { .global }
template <typename B256, enable_if_t<sizeof(B256) == 32, bool> = true>
__device__ static inline void st(
cuda::ptx::space_global_t,
B256* addr,
B256 src);
st.global.L2::cache_hint.b8#
// st.space.L2::cache_hint.b8 [addr], src, cache_policy; // PTX ISA 74, SM_80
// .space = { .global }
template <typename B8, enable_if_t<sizeof(B8) == 1, bool> = true>
__device__ static inline void st_L2_cache_hint(
cuda::ptx::space_global_t,
B8* addr,
B8 src,
uint64_t cache_policy);
st.global.L2::cache_hint.b16#
// st.space.L2::cache_hint.b16 [addr], src, cache_policy; // PTX ISA 74, SM_80
// .space = { .global }
template <typename B16, enable_if_t<sizeof(B16) == 2, bool> = true>
__device__ static inline void st_L2_cache_hint(
cuda::ptx::space_global_t,
B16* addr,
B16 src,
uint64_t cache_policy);
st.global.L2::cache_hint.b32#
// st.space.L2::cache_hint.b32 [addr], src, cache_policy; // PTX ISA 74, SM_80
// .space = { .global }
template <typename B32, enable_if_t<sizeof(B32) == 4, bool> = true>
__device__ static inline void st_L2_cache_hint(
cuda::ptx::space_global_t,
B32* addr,
B32 src,
uint64_t cache_policy);
st.global.L2::cache_hint.b64#
// st.space.L2::cache_hint.b64 [addr], src, cache_policy; // PTX ISA 74, SM_80
// .space = { .global }
template <typename B64, enable_if_t<sizeof(B64) == 8, bool> = true>
__device__ static inline void st_L2_cache_hint(
cuda::ptx::space_global_t,
B64* addr,
B64 src,
uint64_t cache_policy);
st.global.L2::cache_hint.b128#
// st.space.L2::cache_hint.b128 [addr], src, cache_policy; // PTX ISA 83, SM_80
// .space = { .global }
template <typename B128, enable_if_t<sizeof(B128) == 16, bool> = true>
__device__ static inline void st_L2_cache_hint(
cuda::ptx::space_global_t,
B128* addr,
B128 src,
uint64_t cache_policy);
st.global.L2::cache_hint.v4.b64#
// st.space.L2::cache_hint.v4.b64 [addr], src, cache_policy; // PTX ISA 88, SM_100
// .space = { .global }
template <typename B256, enable_if_t<sizeof(B256) == 32, bool> = true>
__device__ static inline void st_L2_cache_hint(
cuda::ptx::space_global_t,
B256* addr,
B256 src,
uint64_t cache_policy);
st.global.L1::evict_first.b8#
// st.space.L1::evict_first.b8 [addr], src; // PTX ISA 74, SM_70
// .space = { .global }
template <typename B8, enable_if_t<sizeof(B8) == 1, bool> = true>
__device__ static inline void st_L1_evict_first(
cuda::ptx::space_global_t,
B8* addr,
B8 src);
st.global.L1::evict_first.b16#
// st.space.L1::evict_first.b16 [addr], src; // PTX ISA 74, SM_70
// .space = { .global }
template <typename B16, enable_if_t<sizeof(B16) == 2, bool> = true>
__device__ static inline void st_L1_evict_first(
cuda::ptx::space_global_t,
B16* addr,
B16 src);
st.global.L1::evict_first.b32#
// st.space.L1::evict_first.b32 [addr], src; // PTX ISA 74, SM_70
// .space = { .global }
template <typename B32, enable_if_t<sizeof(B32) == 4, bool> = true>
__device__ static inline void st_L1_evict_first(
cuda::ptx::space_global_t,
B32* addr,
B32 src);
st.global.L1::evict_first.b64#
// st.space.L1::evict_first.b64 [addr], src; // PTX ISA 74, SM_70
// .space = { .global }
template <typename B64, enable_if_t<sizeof(B64) == 8, bool> = true>
__device__ static inline void st_L1_evict_first(
cuda::ptx::space_global_t,
B64* addr,
B64 src);
st.global.L1::evict_first.b128#
// st.space.L1::evict_first.b128 [addr], src; // PTX ISA 83, SM_70
// .space = { .global }
template <typename B128, enable_if_t<sizeof(B128) == 16, bool> = true>
__device__ static inline void st_L1_evict_first(
cuda::ptx::space_global_t,
B128* addr,
B128 src);
st.global.L1::evict_first.v4.b64#
// st.space.L1::evict_first.v4.b64 [addr], src; // PTX ISA 88, SM_100
// .space = { .global }
template <typename B256, enable_if_t<sizeof(B256) == 32, bool> = true>
__device__ static inline void st_L1_evict_first(
cuda::ptx::space_global_t,
B256* addr,
B256 src);
st.global.L1::evict_first.L2::cache_hint.b8#
// st.space.L1::evict_first.L2::cache_hint.b8 [addr], src, cache_policy; // PTX ISA 74, SM_80
// .space = { .global }
template <typename B8, enable_if_t<sizeof(B8) == 1, bool> = true>
__device__ static inline void st_L1_evict_first_L2_cache_hint(
cuda::ptx::space_global_t,
B8* addr,
B8 src,
uint64_t cache_policy);
st.global.L1::evict_first.L2::cache_hint.b16#
// st.space.L1::evict_first.L2::cache_hint.b16 [addr], src, cache_policy; // PTX ISA 74, SM_80
// .space = { .global }
template <typename B16, enable_if_t<sizeof(B16) == 2, bool> = true>
__device__ static inline void st_L1_evict_first_L2_cache_hint(
cuda::ptx::space_global_t,
B16* addr,
B16 src,
uint64_t cache_policy);
st.global.L1::evict_first.L2::cache_hint.b32#
// st.space.L1::evict_first.L2::cache_hint.b32 [addr], src, cache_policy; // PTX ISA 74, SM_80
// .space = { .global }
template <typename B32, enable_if_t<sizeof(B32) == 4, bool> = true>
__device__ static inline void st_L1_evict_first_L2_cache_hint(
cuda::ptx::space_global_t,
B32* addr,
B32 src,
uint64_t cache_policy);
st.global.L1::evict_first.L2::cache_hint.b64#
// st.space.L1::evict_first.L2::cache_hint.b64 [addr], src, cache_policy; // PTX ISA 74, SM_80
// .space = { .global }
template <typename B64, enable_if_t<sizeof(B64) == 8, bool> = true>
__device__ static inline void st_L1_evict_first_L2_cache_hint(
cuda::ptx::space_global_t,
B64* addr,
B64 src,
uint64_t cache_policy);
st.global.L1::evict_first.L2::cache_hint.b128#
// st.space.L1::evict_first.L2::cache_hint.b128 [addr], src, cache_policy; // PTX ISA 83, SM_80
// .space = { .global }
template <typename B128, enable_if_t<sizeof(B128) == 16, bool> = true>
__device__ static inline void st_L1_evict_first_L2_cache_hint(
cuda::ptx::space_global_t,
B128* addr,
B128 src,
uint64_t cache_policy);
st.global.L1::evict_first.L2::cache_hint.v4.b64#
// st.space.L1::evict_first.L2::cache_hint.v4.b64 [addr], src, cache_policy; // PTX ISA 88, SM_100
// .space = { .global }
template <typename B256, enable_if_t<sizeof(B256) == 32, bool> = true>
__device__ static inline void st_L1_evict_first_L2_cache_hint(
cuda::ptx::space_global_t,
B256* addr,
B256 src,
uint64_t cache_policy);
st.global.L1::evict_last.b8#
// st.space.L1::evict_last.b8 [addr], src; // PTX ISA 74, SM_70
// .space = { .global }
template <typename B8, enable_if_t<sizeof(B8) == 1, bool> = true>
__device__ static inline void st_L1_evict_last(
cuda::ptx::space_global_t,
B8* addr,
B8 src);
st.global.L1::evict_last.b16#
// st.space.L1::evict_last.b16 [addr], src; // PTX ISA 74, SM_70
// .space = { .global }
template <typename B16, enable_if_t<sizeof(B16) == 2, bool> = true>
__device__ static inline void st_L1_evict_last(
cuda::ptx::space_global_t,
B16* addr,
B16 src);
st.global.L1::evict_last.b32#
// st.space.L1::evict_last.b32 [addr], src; // PTX ISA 74, SM_70
// .space = { .global }
template <typename B32, enable_if_t<sizeof(B32) == 4, bool> = true>
__device__ static inline void st_L1_evict_last(
cuda::ptx::space_global_t,
B32* addr,
B32 src);
st.global.L1::evict_last.b64#
// st.space.L1::evict_last.b64 [addr], src; // PTX ISA 74, SM_70
// .space = { .global }
template <typename B64, enable_if_t<sizeof(B64) == 8, bool> = true>
__device__ static inline void st_L1_evict_last(
cuda::ptx::space_global_t,
B64* addr,
B64 src);
st.global.L1::evict_last.b128#
// st.space.L1::evict_last.b128 [addr], src; // PTX ISA 83, SM_70
// .space = { .global }
template <typename B128, enable_if_t<sizeof(B128) == 16, bool> = true>
__device__ static inline void st_L1_evict_last(
cuda::ptx::space_global_t,
B128* addr,
B128 src);
st.global.L1::evict_last.v4.b64#
// st.space.L1::evict_last.v4.b64 [addr], src; // PTX ISA 88, SM_100
// .space = { .global }
template <typename B256, enable_if_t<sizeof(B256) == 32, bool> = true>
__device__ static inline void st_L1_evict_last(
cuda::ptx::space_global_t,
B256* addr,
B256 src);
st.global.L1::evict_last.L2::cache_hint.b8#
// st.space.L1::evict_last.L2::cache_hint.b8 [addr], src, cache_policy; // PTX ISA 74, SM_80
// .space = { .global }
template <typename B8, enable_if_t<sizeof(B8) == 1, bool> = true>
__device__ static inline void st_L1_evict_last_L2_cache_hint(
cuda::ptx::space_global_t,
B8* addr,
B8 src,
uint64_t cache_policy);
st.global.L1::evict_last.L2::cache_hint.b16#
// st.space.L1::evict_last.L2::cache_hint.b16 [addr], src, cache_policy; // PTX ISA 74, SM_80
// .space = { .global }
template <typename B16, enable_if_t<sizeof(B16) == 2, bool> = true>
__device__ static inline void st_L1_evict_last_L2_cache_hint(
cuda::ptx::space_global_t,
B16* addr,
B16 src,
uint64_t cache_policy);
st.global.L1::evict_last.L2::cache_hint.b32#
// st.space.L1::evict_last.L2::cache_hint.b32 [addr], src, cache_policy; // PTX ISA 74, SM_80
// .space = { .global }
template <typename B32, enable_if_t<sizeof(B32) == 4, bool> = true>
__device__ static inline void st_L1_evict_last_L2_cache_hint(
cuda::ptx::space_global_t,
B32* addr,
B32 src,
uint64_t cache_policy);
st.global.L1::evict_last.L2::cache_hint.b64#
// st.space.L1::evict_last.L2::cache_hint.b64 [addr], src, cache_policy; // PTX ISA 74, SM_80
// .space = { .global }
template <typename B64, enable_if_t<sizeof(B64) == 8, bool> = true>
__device__ static inline void st_L1_evict_last_L2_cache_hint(
cuda::ptx::space_global_t,
B64* addr,
B64 src,
uint64_t cache_policy);
st.global.L1::evict_last.L2::cache_hint.b128#
// st.space.L1::evict_last.L2::cache_hint.b128 [addr], src, cache_policy; // PTX ISA 83, SM_80
// .space = { .global }
template <typename B128, enable_if_t<sizeof(B128) == 16, bool> = true>
__device__ static inline void st_L1_evict_last_L2_cache_hint(
cuda::ptx::space_global_t,
B128* addr,
B128 src,
uint64_t cache_policy);
st.global.L1::evict_last.L2::cache_hint.v4.b64#
// st.space.L1::evict_last.L2::cache_hint.v4.b64 [addr], src, cache_policy; // PTX ISA 88, SM_100
// .space = { .global }
template <typename B256, enable_if_t<sizeof(B256) == 32, bool> = true>
__device__ static inline void st_L1_evict_last_L2_cache_hint(
cuda::ptx::space_global_t,
B256* addr,
B256 src,
uint64_t cache_policy);
st.global.L1::no_allocate.b8#
// st.space.L1::no_allocate.b8 [addr], src; // PTX ISA 74, SM_70
// .space = { .global }
template <typename B8, enable_if_t<sizeof(B8) == 1, bool> = true>
__device__ static inline void st_L1_no_allocate(
cuda::ptx::space_global_t,
B8* addr,
B8 src);
st.global.L1::no_allocate.b16#
// st.space.L1::no_allocate.b16 [addr], src; // PTX ISA 74, SM_70
// .space = { .global }
template <typename B16, enable_if_t<sizeof(B16) == 2, bool> = true>
__device__ static inline void st_L1_no_allocate(
cuda::ptx::space_global_t,
B16* addr,
B16 src);
st.global.L1::no_allocate.b32#
// st.space.L1::no_allocate.b32 [addr], src; // PTX ISA 74, SM_70
// .space = { .global }
template <typename B32, enable_if_t<sizeof(B32) == 4, bool> = true>
__device__ static inline void st_L1_no_allocate(
cuda::ptx::space_global_t,
B32* addr,
B32 src);
st.global.L1::no_allocate.b64#
// st.space.L1::no_allocate.b64 [addr], src; // PTX ISA 74, SM_70
// .space = { .global }
template <typename B64, enable_if_t<sizeof(B64) == 8, bool> = true>
__device__ static inline void st_L1_no_allocate(
cuda::ptx::space_global_t,
B64* addr,
B64 src);
st.global.L1::no_allocate.b128#
// st.space.L1::no_allocate.b128 [addr], src; // PTX ISA 83, SM_70
// .space = { .global }
template <typename B128, enable_if_t<sizeof(B128) == 16, bool> = true>
__device__ static inline void st_L1_no_allocate(
cuda::ptx::space_global_t,
B128* addr,
B128 src);
st.global.L1::no_allocate.v4.b64#
// st.space.L1::no_allocate.v4.b64 [addr], src; // PTX ISA 88, SM_100
// .space = { .global }
template <typename B256, enable_if_t<sizeof(B256) == 32, bool> = true>
__device__ static inline void st_L1_no_allocate(
cuda::ptx::space_global_t,
B256* addr,
B256 src);
st.global.L1::no_allocate.L2::cache_hint.b8#
// st.space.L1::no_allocate.L2::cache_hint.b8 [addr], src, cache_policy; // PTX ISA 74, SM_80
// .space = { .global }
template <typename B8, enable_if_t<sizeof(B8) == 1, bool> = true>
__device__ static inline void st_L1_no_allocate_L2_cache_hint(
cuda::ptx::space_global_t,
B8* addr,
B8 src,
uint64_t cache_policy);
st.global.L1::no_allocate.L2::cache_hint.b16#
// st.space.L1::no_allocate.L2::cache_hint.b16 [addr], src, cache_policy; // PTX ISA 74, SM_80
// .space = { .global }
template <typename B16, enable_if_t<sizeof(B16) == 2, bool> = true>
__device__ static inline void st_L1_no_allocate_L2_cache_hint(
cuda::ptx::space_global_t,
B16* addr,
B16 src,
uint64_t cache_policy);
st.global.L1::no_allocate.L2::cache_hint.b32#
// st.space.L1::no_allocate.L2::cache_hint.b32 [addr], src, cache_policy; // PTX ISA 74, SM_80
// .space = { .global }
template <typename B32, enable_if_t<sizeof(B32) == 4, bool> = true>
__device__ static inline void st_L1_no_allocate_L2_cache_hint(
cuda::ptx::space_global_t,
B32* addr,
B32 src,
uint64_t cache_policy);
st.global.L1::no_allocate.L2::cache_hint.b64#
// st.space.L1::no_allocate.L2::cache_hint.b64 [addr], src, cache_policy; // PTX ISA 74, SM_80
// .space = { .global }
template <typename B64, enable_if_t<sizeof(B64) == 8, bool> = true>
__device__ static inline void st_L1_no_allocate_L2_cache_hint(
cuda::ptx::space_global_t,
B64* addr,
B64 src,
uint64_t cache_policy);
st.global.L1::no_allocate.L2::cache_hint.b128#
// st.space.L1::no_allocate.L2::cache_hint.b128 [addr], src, cache_policy; // PTX ISA 83, SM_80
// .space = { .global }
template <typename B128, enable_if_t<sizeof(B128) == 16, bool> = true>
__device__ static inline void st_L1_no_allocate_L2_cache_hint(
cuda::ptx::space_global_t,
B128* addr,
B128 src,
uint64_t cache_policy);
st.global.L1::no_allocate.L2::cache_hint.v4.b64#
// st.space.L1::no_allocate.L2::cache_hint.v4.b64 [addr], src, cache_policy; // PTX ISA 88, SM_100
// .space = { .global }
template <typename B256, enable_if_t<sizeof(B256) == 32, bool> = true>
__device__ static inline void st_L1_no_allocate_L2_cache_hint(
cuda::ptx::space_global_t,
B256* addr,
B256 src,
uint64_t cache_policy);