blob: 4af3fdf2bd294cd441e0068d0911336c3cfdd0c6 [file] [log] [blame]
/**
* The atomic module provides basic support for lock-free
* concurrent programming.
*
* Copyright: Copyright Sean Kelly 2005 - 2016.
* License: $(LINK2 http://www.boost.org/LICENSE_1_0.txt, Boost License 1.0)
* Authors: Sean Kelly, Alex Rønne Petersen, Manu Evans
* Source: $(DRUNTIMESRC core/_atomic.d)
*/
module core.atomic;
import core.internal.atomic;
import core.internal.attributes : betterC;
import core.internal.traits : hasUnsharedIndirections;
/**
* Specifies the memory ordering semantics of an atomic operation.
*
* See_Also:
* $(HTTP en.cppreference.com/w/cpp/atomic/memory_order)
*/
enum MemoryOrder
{
/**
* Not sequenced.
* Corresponds to $(LINK2 https://llvm.org/docs/Atomics.html#monotonic, LLVM AtomicOrdering.Monotonic)
* and C++11/C11 `memory_order_relaxed`.
*/
raw = 0,
/**
* Hoist-load + hoist-store barrier.
* Corresponds to $(LINK2 https://llvm.org/docs/Atomics.html#acquire, LLVM AtomicOrdering.Acquire)
* and C++11/C11 `memory_order_acquire`.
*/
acq = 2,
/**
* Sink-load + sink-store barrier.
* Corresponds to $(LINK2 https://llvm.org/docs/Atomics.html#release, LLVM AtomicOrdering.Release)
* and C++11/C11 `memory_order_release`.
*/
rel = 3,
/**
* Acquire + release barrier.
* Corresponds to $(LINK2 https://llvm.org/docs/Atomics.html#acquirerelease, LLVM AtomicOrdering.AcquireRelease)
* and C++11/C11 `memory_order_acq_rel`.
*/
acq_rel = 4,
/**
* Fully sequenced (acquire + release). Corresponds to
* $(LINK2 https://llvm.org/docs/Atomics.html#sequentiallyconsistent, LLVM AtomicOrdering.SequentiallyConsistent)
* and C++11/C11 `memory_order_seq_cst`.
*/
seq = 5,
}
/**
* Loads 'val' from memory and returns it. The memory barrier specified
* by 'ms' is applied to the operation, which is fully sequenced by
* default. Valid memory orders are MemoryOrder.raw, MemoryOrder.acq,
* and MemoryOrder.seq.
*
* Params:
* val = The target variable.
*
* Returns:
* The value of 'val'.
*/
T atomicLoad(MemoryOrder ms = MemoryOrder.seq, T)(ref return scope const T val) pure nothrow @nogc @trusted
if (!is(T == shared U, U) && !is(T == shared inout U, U) && !is(T == shared const U, U))
{
static if (__traits(isFloating, T))
{
alias IntTy = IntForFloat!T;
IntTy r = core.internal.atomic.atomicLoad!ms(cast(IntTy*)&val);
return *cast(T*)&r;
}
else
return core.internal.atomic.atomicLoad!ms(cast(T*)&val);
}
/// Ditto
T atomicLoad(MemoryOrder ms = MemoryOrder.seq, T)(ref return scope shared const T val) pure nothrow @nogc @trusted
if (!hasUnsharedIndirections!T)
{
import core.internal.traits : hasUnsharedIndirections;
static assert(!hasUnsharedIndirections!T, "Copying `" ~ shared(const(T)).stringof ~ "` would violate shared.");
return atomicLoad!ms(*cast(T*)&val);
}
/// Ditto
TailShared!T atomicLoad(MemoryOrder ms = MemoryOrder.seq, T)(ref shared const T val) pure nothrow @nogc @trusted
if (hasUnsharedIndirections!T)
{
// HACK: DEPRECATE THIS FUNCTION, IT IS INVALID TO DO ATOMIC LOAD OF SHARED CLASS
// this is here because code exists in the wild that does this...
return core.internal.atomic.atomicLoad!ms(cast(TailShared!T*)&val);
}
/**
* Writes 'newval' into 'val'. The memory barrier specified by 'ms' is
* applied to the operation, which is fully sequenced by default.
* Valid memory orders are MemoryOrder.raw, MemoryOrder.rel, and
* MemoryOrder.seq.
*
* Params:
* val = The target variable.
* newval = The value to store.
*/
void atomicStore(MemoryOrder ms = MemoryOrder.seq, T, V)(ref T val, V newval) pure nothrow @nogc @trusted
if (!is(T == shared) && !is(V == shared))
{
import core.internal.traits : hasElaborateCopyConstructor;
static assert (!hasElaborateCopyConstructor!T, "`T` may not have an elaborate copy: atomic operations override regular copying semantics.");
// resolve implicit conversions
T arg = newval;
static if (__traits(isFloating, T))
{
alias IntTy = IntForFloat!T;
core.internal.atomic.atomicStore!ms(cast(IntTy*)&val, *cast(IntTy*)&arg);
}
else
core.internal.atomic.atomicStore!ms(&val, arg);
}
/// Ditto
void atomicStore(MemoryOrder ms = MemoryOrder.seq, T, V)(ref shared T val, V newval) pure nothrow @nogc @trusted
if (!is(T == class))
{
static if (is (V == shared U, U))
alias Thunk = U;
else
{
import core.internal.traits : hasUnsharedIndirections;
static assert(!hasUnsharedIndirections!V, "Copying argument `" ~ V.stringof ~ " newval` to `" ~ shared(T).stringof ~ " here` would violate shared.");
alias Thunk = V;
}
atomicStore!ms(*cast(T*)&val, *cast(Thunk*)&newval);
}
/// Ditto
void atomicStore(MemoryOrder ms = MemoryOrder.seq, T, V)(ref shared T val, shared V newval) pure nothrow @nogc @trusted
if (is(T == class))
{
static assert (is (V : T), "Can't assign `newval` of type `shared " ~ V.stringof ~ "` to `shared " ~ T.stringof ~ "`.");
core.internal.atomic.atomicStore!ms(cast(T*)&val, cast(V)newval);
}
/**
* Atomically adds `mod` to the value referenced by `val` and returns the value `val` held previously.
* This operation is both lock-free and atomic.
*
* Params:
* val = Reference to the value to modify.
* mod = The value to add.
*
* Returns:
* The value held previously by `val`.
*/
T atomicFetchAdd(MemoryOrder ms = MemoryOrder.seq, T)(ref return scope T val, size_t mod) pure nothrow @nogc @trusted
if ((__traits(isIntegral, T) || is(T == U*, U)) && !is(T == shared))
in (atomicValueIsProperlyAligned(val))
{
static if (is(T == U*, U))
return cast(T)core.internal.atomic.atomicFetchAdd!ms(cast(size_t*)&val, mod * U.sizeof);
else
return core.internal.atomic.atomicFetchAdd!ms(&val, cast(T)mod);
}
/// Ditto
T atomicFetchAdd(MemoryOrder ms = MemoryOrder.seq, T)(ref return scope shared T val, size_t mod) pure nothrow @nogc @trusted
if (__traits(isIntegral, T) || is(T == U*, U))
in (atomicValueIsProperlyAligned(val))
{
return atomicFetchAdd!ms(*cast(T*)&val, mod);
}
/**
* Atomically subtracts `mod` from the value referenced by `val` and returns the value `val` held previously.
* This operation is both lock-free and atomic.
*
* Params:
* val = Reference to the value to modify.
* mod = The value to subtract.
*
* Returns:
* The value held previously by `val`.
*/
T atomicFetchSub(MemoryOrder ms = MemoryOrder.seq, T)(ref return scope T val, size_t mod) pure nothrow @nogc @trusted
if ((__traits(isIntegral, T) || is(T == U*, U)) && !is(T == shared))
in (atomicValueIsProperlyAligned(val))
{
static if (is(T == U*, U))
return cast(T)core.internal.atomic.atomicFetchSub!ms(cast(size_t*)&val, mod * U.sizeof);
else
return core.internal.atomic.atomicFetchSub!ms(&val, cast(T)mod);
}
/// Ditto
T atomicFetchSub(MemoryOrder ms = MemoryOrder.seq, T)(ref return scope shared T val, size_t mod) pure nothrow @nogc @trusted
if (__traits(isIntegral, T) || is(T == U*, U))
in (atomicValueIsProperlyAligned(val))
{
return atomicFetchSub!ms(*cast(T*)&val, mod);
}
/**
* Exchange `exchangeWith` with the memory referenced by `here`.
* This operation is both lock-free and atomic.
*
* Params:
* here = The address of the destination variable.
* exchangeWith = The value to exchange.
*
* Returns:
* The value held previously by `here`.
*/
T atomicExchange(MemoryOrder ms = MemoryOrder.seq,T,V)(T* here, V exchangeWith) pure nothrow @nogc @trusted
if (!is(T == shared) && !is(V == shared))
in (atomicPtrIsProperlyAligned(here), "Argument `here` is not properly aligned")
{
// resolve implicit conversions
T arg = exchangeWith;
static if (__traits(isFloating, T))
{
alias IntTy = IntForFloat!T;
IntTy r = core.internal.atomic.atomicExchange!ms(cast(IntTy*)here, *cast(IntTy*)&arg);
return *cast(shared(T)*)&r;
}
else
return core.internal.atomic.atomicExchange!ms(here, arg);
}
/// Ditto
TailShared!T atomicExchange(MemoryOrder ms = MemoryOrder.seq,T,V)(shared(T)* here, V exchangeWith) pure nothrow @nogc @trusted
if (!is(T == class) && !is(T == interface))
in (atomicPtrIsProperlyAligned(here), "Argument `here` is not properly aligned")
{
static if (is (V == shared U, U))
alias Thunk = U;
else
{
import core.internal.traits : hasUnsharedIndirections;
static assert(!hasUnsharedIndirections!V, "Copying `exchangeWith` of type `" ~ V.stringof ~ "` to `" ~ shared(T).stringof ~ "` would violate shared.");
alias Thunk = V;
}
return atomicExchange!ms(cast(T*)here, *cast(Thunk*)&exchangeWith);
}
/// Ditto
shared(T) atomicExchange(MemoryOrder ms = MemoryOrder.seq,T,V)(shared(T)* here, shared(V) exchangeWith) pure nothrow @nogc @trusted
if (is(T == class) || is(T == interface))
in (atomicPtrIsProperlyAligned(here), "Argument `here` is not properly aligned")
{
static assert (is (V : T), "Can't assign `exchangeWith` of type `" ~ shared(V).stringof ~ "` to `" ~ shared(T).stringof ~ "`.");
return cast(shared)core.internal.atomic.atomicExchange!ms(cast(T*)here, cast(V)exchangeWith);
}
/**
* Performs either compare-and-set or compare-and-swap (or exchange).
*
* There are two categories of overloads in this template:
* The first category does a simple compare-and-set.
* The comparison value (`ifThis`) is treated as an rvalue.
*
* The second category does a compare-and-swap (a.k.a. compare-and-exchange),
* and expects `ifThis` to be a pointer type, where the previous value
* of `here` will be written.
*
* This operation is both lock-free and atomic.
*
* Params:
* here = The address of the destination variable.
* writeThis = The value to store.
* ifThis = The comparison value.
*
* Returns:
* true if the store occurred, false if not.
*/
template cas(MemoryOrder succ = MemoryOrder.seq, MemoryOrder fail = MemoryOrder.seq)
{
/// Compare-and-set for non-shared values
bool cas(T, V1, V2)(T* here, V1 ifThis, V2 writeThis) pure nothrow @nogc @trusted
if (!is(T == shared) && is(T : V1))
in (atomicPtrIsProperlyAligned(here), "Argument `here` is not properly aligned")
{
// resolve implicit conversions
const T arg1 = ifThis;
T arg2 = writeThis;
static if (__traits(isFloating, T))
{
alias IntTy = IntForFloat!T;
return atomicCompareExchangeStrongNoResult!(succ, fail)(
cast(IntTy*)here, *cast(IntTy*)&arg1, *cast(IntTy*)&arg2);
}
else
return atomicCompareExchangeStrongNoResult!(succ, fail)(here, arg1, arg2);
}
/// Compare-and-set for shared value type
bool cas(T, V1, V2)(shared(T)* here, V1 ifThis, V2 writeThis) pure nothrow @nogc @trusted
if (!is(T == class) && (is(T : V1) || is(shared T : V1)))
in (atomicPtrIsProperlyAligned(here), "Argument `here` is not properly aligned")
{
static if (is (V1 == shared U1, U1))
alias Thunk1 = U1;
else
alias Thunk1 = V1;
static if (is (V2 == shared U2, U2))
alias Thunk2 = U2;
else
{
import core.internal.traits : hasUnsharedIndirections;
static assert(!hasUnsharedIndirections!V2,
"Copying `" ~ V2.stringof ~ "* writeThis` to `" ~
shared(T).stringof ~ "* here` would violate shared.");
alias Thunk2 = V2;
}
return cas(cast(T*)here, *cast(Thunk1*)&ifThis, *cast(Thunk2*)&writeThis);
}
/// Compare-and-set for `shared` reference type (`class`)
bool cas(T, V1, V2)(shared(T)* here, shared(V1) ifThis, shared(V2) writeThis)
pure nothrow @nogc @trusted
if (is(T == class))
in (atomicPtrIsProperlyAligned(here), "Argument `here` is not properly aligned")
{
return atomicCompareExchangeStrongNoResult!(succ, fail)(
cast(T*)here, cast(V1)ifThis, cast(V2)writeThis);
}
/// Compare-and-exchange for non-`shared` types
bool cas(T, V)(T* here, T* ifThis, V writeThis) pure nothrow @nogc @trusted
if (!is(T == shared) && !is(V == shared))
in (atomicPtrIsProperlyAligned(here), "Argument `here` is not properly aligned")
{
// resolve implicit conversions
T arg1 = writeThis;
static if (__traits(isFloating, T))
{
alias IntTy = IntForFloat!T;
return atomicCompareExchangeStrong!(succ, fail)(
cast(IntTy*)here, cast(IntTy*)ifThis, *cast(IntTy*)&writeThis);
}
else
return atomicCompareExchangeStrong!(succ, fail)(here, ifThis, writeThis);
}
/// Compare and exchange for mixed-`shared`ness types
bool cas(T, V1, V2)(shared(T)* here, V1* ifThis, V2 writeThis) pure nothrow @nogc @trusted
if (!is(T == class) && (is(T : V1) || is(shared T : V1)))
in (atomicPtrIsProperlyAligned(here), "Argument `here` is not properly aligned")
{
static if (is (V1 == shared U1, U1))
alias Thunk1 = U1;
else
{
import core.internal.traits : hasUnsharedIndirections;
static assert(!hasUnsharedIndirections!V1,
"Copying `" ~ shared(T).stringof ~ "* here` to `" ~
V1.stringof ~ "* ifThis` would violate shared.");
alias Thunk1 = V1;
}
static if (is (V2 == shared U2, U2))
alias Thunk2 = U2;
else
{
import core.internal.traits : hasUnsharedIndirections;
static assert(!hasUnsharedIndirections!V2,
"Copying `" ~ V2.stringof ~ "* writeThis` to `" ~
shared(T).stringof ~ "* here` would violate shared.");
alias Thunk2 = V2;
}
static assert (is(T : Thunk1),
"Mismatching types for `here` and `ifThis`: `" ~
shared(T).stringof ~ "` and `" ~ V1.stringof ~ "`.");
return cas(cast(T*)here, cast(Thunk1*)ifThis, *cast(Thunk2*)&writeThis);
}
/// Compare-and-exchange for `class`
bool cas(T, V)(shared(T)* here, shared(T)* ifThis, shared(V) writeThis)
pure nothrow @nogc @trusted
if (is(T == class))
in (atomicPtrIsProperlyAligned(here), "Argument `here` is not properly aligned")
{
return atomicCompareExchangeStrong!(succ, fail)(
cast(T*)here, cast(T*)ifThis, cast(V)writeThis);
}
}
/**
* Stores 'writeThis' to the memory referenced by 'here' if the value
* referenced by 'here' is equal to 'ifThis'.
* The 'weak' version of cas may spuriously fail. It is recommended to
* use `casWeak` only when `cas` would be used in a loop.
* This operation is both
* lock-free and atomic.
*
* Params:
* here = The address of the destination variable.
* writeThis = The value to store.
* ifThis = The comparison value.
*
* Returns:
* true if the store occurred, false if not.
*/
bool casWeak(MemoryOrder succ = MemoryOrder.seq,MemoryOrder fail = MemoryOrder.seq,T,V1,V2)(T* here, V1 ifThis, V2 writeThis) pure nothrow @nogc @trusted
if (!is(T == shared) && is(T : V1))
in (atomicPtrIsProperlyAligned(here), "Argument `here` is not properly aligned")
{
// resolve implicit conversions
T arg1 = ifThis;
T arg2 = writeThis;
static if (__traits(isFloating, T))
{
alias IntTy = IntForFloat!T;
return atomicCompareExchangeWeakNoResult!(succ, fail)(cast(IntTy*)here, *cast(IntTy*)&arg1, *cast(IntTy*)&arg2);
}
else
return atomicCompareExchangeWeakNoResult!(succ, fail)(here, arg1, arg2);
}
/// Ditto
bool casWeak(MemoryOrder succ = MemoryOrder.seq,MemoryOrder fail = MemoryOrder.seq,T,V1,V2)(shared(T)* here, V1 ifThis, V2 writeThis) pure nothrow @nogc @trusted
if (!is(T == class) && (is(T : V1) || is(shared T : V1)))
in (atomicPtrIsProperlyAligned(here), "Argument `here` is not properly aligned")
{
static if (is (V1 == shared U1, U1))
alias Thunk1 = U1;
else
alias Thunk1 = V1;
static if (is (V2 == shared U2, U2))
alias Thunk2 = U2;
else
{
import core.internal.traits : hasUnsharedIndirections;
static assert(!hasUnsharedIndirections!V2, "Copying `" ~ V2.stringof ~ "* writeThis` to `" ~ shared(T).stringof ~ "* here` would violate shared.");
alias Thunk2 = V2;
}
return casWeak!(succ, fail)(cast(T*)here, *cast(Thunk1*)&ifThis, *cast(Thunk2*)&writeThis);
}
/// Ditto
bool casWeak(MemoryOrder succ = MemoryOrder.seq,MemoryOrder fail = MemoryOrder.seq,T,V1,V2)(shared(T)* here, shared(V1) ifThis, shared(V2) writeThis) pure nothrow @nogc @trusted
if (is(T == class))
in (atomicPtrIsProperlyAligned(here), "Argument `here` is not properly aligned")
{
return atomicCompareExchangeWeakNoResult!(succ, fail)(cast(T*)here, cast(V1)ifThis, cast(V2)writeThis);
}
/**
* Stores 'writeThis' to the memory referenced by 'here' if the value
* referenced by 'here' is equal to the value referenced by 'ifThis'.
* The prior value referenced by 'here' is written to `ifThis` and
* returned to the user.
* The 'weak' version of cas may spuriously fail. It is recommended to
* use `casWeak` only when `cas` would be used in a loop.
* This operation is both lock-free and atomic.
*
* Params:
* here = The address of the destination variable.
* writeThis = The value to store.
* ifThis = The address of the value to compare, and receives the prior value of `here` as output.
*
* Returns:
* true if the store occurred, false if not.
*/
bool casWeak(MemoryOrder succ = MemoryOrder.seq,MemoryOrder fail = MemoryOrder.seq,T,V)(T* here, T* ifThis, V writeThis) pure nothrow @nogc @trusted
if (!is(T == shared S, S) && !is(V == shared U, U))
in (atomicPtrIsProperlyAligned(here), "Argument `here` is not properly aligned")
{
// resolve implicit conversions
T arg1 = writeThis;
static if (__traits(isFloating, T))
{
alias IntTy = IntForFloat!T;
return atomicCompareExchangeWeak!(succ, fail)(cast(IntTy*)here, cast(IntTy*)ifThis, *cast(IntTy*)&writeThis);
}
else
return atomicCompareExchangeWeak!(succ, fail)(here, ifThis, writeThis);
}
/// Ditto
bool casWeak(MemoryOrder succ = MemoryOrder.seq,MemoryOrder fail = MemoryOrder.seq,T,V1,V2)(shared(T)* here, V1* ifThis, V2 writeThis) pure nothrow @nogc @trusted
if (!is(T == class) && (is(T : V1) || is(shared T : V1)))
in (atomicPtrIsProperlyAligned(here), "Argument `here` is not properly aligned")
{
static if (is (V1 == shared U1, U1))
alias Thunk1 = U1;
else
{
import core.internal.traits : hasUnsharedIndirections;
static assert(!hasUnsharedIndirections!V1, "Copying `" ~ shared(T).stringof ~ "* here` to `" ~ V1.stringof ~ "* ifThis` would violate shared.");
alias Thunk1 = V1;
}
static if (is (V2 == shared U2, U2))
alias Thunk2 = U2;
else
{
import core.internal.traits : hasUnsharedIndirections;
static assert(!hasUnsharedIndirections!V2, "Copying `" ~ V2.stringof ~ "* writeThis` to `" ~ shared(T).stringof ~ "* here` would violate shared.");
alias Thunk2 = V2;
}
static assert (is(T : Thunk1), "Mismatching types for `here` and `ifThis`: `" ~ shared(T).stringof ~ "` and `" ~ V1.stringof ~ "`.");
return casWeak!(succ, fail)(cast(T*)here, cast(Thunk1*)ifThis, *cast(Thunk2*)&writeThis);
}
/// Ditto
bool casWeak(MemoryOrder succ = MemoryOrder.seq,MemoryOrder fail = MemoryOrder.seq,T,V)(shared(T)* here, shared(T)* ifThis, shared(V) writeThis) pure nothrow @nogc @trusted
if (is(T == class))
in (atomicPtrIsProperlyAligned(here), "Argument `here` is not properly aligned")
{
return atomicCompareExchangeWeak!(succ, fail)(cast(T*)here, cast(T*)ifThis, cast(V)writeThis);
}
/**
* Inserts a full load/store memory fence (on platforms that need it). This ensures
* that all loads and stores before a call to this function are executed before any
* loads and stores after the call.
*/
void atomicFence(MemoryOrder order = MemoryOrder.seq)() pure nothrow @nogc @safe
{
core.internal.atomic.atomicFence!order();
}
/**
* Gives a hint to the processor that the calling thread is in a 'spin-wait' loop,
* allowing to more efficiently allocate resources.
*/
void pause() pure nothrow @nogc @safe
{
core.internal.atomic.pause();
}
/**
* Performs the binary operation 'op' on val using 'mod' as the modifier.
*
* Params:
* val = The target variable.
* mod = The modifier to apply.
*
* Returns:
* The result of the operation.
*/
TailShared!T atomicOp(string op, T, V1)(ref shared T val, V1 mod) pure nothrow @nogc @safe
if (__traits(compiles, mixin("*cast(T*)&val" ~ op ~ "mod")))
in (atomicValueIsProperlyAligned(val))
{
// binary operators
//
// + - * / % ^^ &
// | ^ << >> >>> ~ in
// == != < <= > >=
static if (op == "+" || op == "-" || op == "*" || op == "/" ||
op == "%" || op == "^^" || op == "&" || op == "|" ||
op == "^" || op == "<<" || op == ">>" || op == ">>>" ||
op == "~" || // skip "in"
op == "==" || op == "!=" || op == "<" || op == "<=" ||
op == ">" || op == ">=")
{
T get = atomicLoad!(MemoryOrder.raw, T)(val);
mixin("return get " ~ op ~ " mod;");
}
else
// assignment operators
//
// += -= *= /= %= ^^= &=
// |= ^= <<= >>= >>>= ~=
static if (op == "+=" && __traits(isIntegral, T) && __traits(isIntegral, V1) && T.sizeof <= size_t.sizeof && V1.sizeof <= size_t.sizeof)
{
return cast(T)(atomicFetchAdd(val, mod) + mod);
}
else static if (op == "-=" && __traits(isIntegral, T) && __traits(isIntegral, V1) && T.sizeof <= size_t.sizeof && V1.sizeof <= size_t.sizeof)
{
return cast(T)(atomicFetchSub(val, mod) - mod);
}
else static if (op == "+=" || op == "-=" || op == "*=" || op == "/=" ||
op == "%=" || op == "^^=" || op == "&=" || op == "|=" ||
op == "^=" || op == "<<=" || op == ">>=" || op == ">>>=") // skip "~="
{
T set, get = atomicLoad!(MemoryOrder.raw, T)(val);
do
{
set = get;
mixin("set " ~ op ~ " mod;");
} while (!casWeakByRef(val, get, set));
return set;
}
else
{
static assert(false, "Operation not supported.");
}
}
version (D_InlineAsm_X86)
{
enum has64BitXCHG = false;
enum has64BitCAS = true;
enum has128BitCAS = false;
}
else version (D_InlineAsm_X86_64)
{
enum has64BitXCHG = true;
enum has64BitCAS = true;
enum has128BitCAS = true;
}
else version (GNU)
{
import gcc.config;
enum has64BitCAS = GNU_Have_64Bit_Atomics;
enum has64BitXCHG = GNU_Have_64Bit_Atomics;
enum has128BitCAS = GNU_Have_LibAtomic;
}
else
{
enum has64BitXCHG = false;
enum has64BitCAS = false;
enum has128BitCAS = false;
}
private
{
bool atomicValueIsProperlyAligned(T)(ref T val) pure nothrow @nogc @trusted
{
return atomicPtrIsProperlyAligned(&val);
}
bool atomicPtrIsProperlyAligned(T)(T* ptr) pure nothrow @nogc @safe
{
// NOTE: Strictly speaking, the x86 supports atomic operations on
// unaligned values. However, this is far slower than the
// common case, so such behavior should be prohibited.
static if (T.sizeof > size_t.sizeof)
{
version (X86)
{
// cmpxchg8b only requires 4-bytes alignment
return cast(size_t)ptr % size_t.sizeof == 0;
}
else
{
// e.g., x86_64 cmpxchg16b requires 16-bytes alignment
return cast(size_t)ptr % T.sizeof == 0;
}
}
else
{
return cast(size_t)ptr % T.sizeof == 0;
}
}
template IntForFloat(F)
if (__traits(isFloating, F))
{
static if (F.sizeof == 4)
alias IntForFloat = uint;
else static if (F.sizeof == 8)
alias IntForFloat = ulong;
else
static assert (false, "Invalid floating point type: " ~ F.stringof ~ ", only support `float` and `double`.");
}
template IntForStruct(S)
if (is(S == struct))
{
static if (S.sizeof == 1)
alias IntForFloat = ubyte;
else static if (F.sizeof == 2)
alias IntForFloat = ushort;
else static if (F.sizeof == 4)
alias IntForFloat = uint;
else static if (F.sizeof == 8)
alias IntForFloat = ulong;
else static if (F.sizeof == 16)
alias IntForFloat = ulong[2]; // TODO: what's the best type here? slice/delegates pass in registers...
else
static assert (ValidateStruct!S);
}
template ValidateStruct(S)
if (is(S == struct))
{
import core.internal.traits : hasElaborateAssign;
// `(x & (x-1)) == 0` checks that x is a power of 2.
static assert (S.sizeof <= size_t.sizeof * 2
&& (S.sizeof & (S.sizeof - 1)) == 0,
S.stringof ~ " has invalid size for atomic operations.");
static assert (!hasElaborateAssign!S, S.stringof ~ " may not have an elaborate assignment when used with atomic operations.");
enum ValidateStruct = true;
}
// TODO: it'd be nice if we had @trusted scopes; we could remove this...
bool casWeakByRef(T,V1,V2)(ref T value, ref V1 ifThis, V2 writeThis) pure nothrow @nogc @trusted
{
return casWeak(&value, &ifThis, writeThis);
}
/* Construct a type with a shared tail, and if possible with an unshared
head. */
template TailShared(U) if (!is(U == shared))
{
alias TailShared = .TailShared!(shared U);
}
template TailShared(S) if (is(S == shared))
{
// Get the unshared variant of S.
static if (is(S U == shared U)) {}
else static assert(false, "Should never be triggered. The `static " ~
"if` declares `U` as the unshared version of the shared type " ~
"`S`. `S` is explicitly declared as shared, so getting `U` " ~
"should always work.");
static if (is(S : U))
alias TailShared = U;
else static if (is(S == struct))
{
enum implName = () {
/* Start with "_impl". If S has a field with that name, append
underscores until the clash is resolved. */
string name = "_impl";
string[] fieldNames;
static foreach (alias field; S.tupleof)
{
fieldNames ~= __traits(identifier, field);
}
static bool canFind(string[] haystack, string needle)
{
foreach (candidate; haystack)
{
if (candidate == needle) return true;
}
return false;
}
while (canFind(fieldNames, name)) name ~= "_";
return name;
} ();
struct TailShared
{
static foreach (i, alias field; S.tupleof)
{
/* On @trusted: This is casting the field from shared(Foo)
to TailShared!Foo. The cast is safe because the field has
been loaded and is not shared anymore. */
mixin("
@trusted @property
ref " ~ __traits(identifier, field) ~ "()
{
alias R = TailShared!(typeof(field));
return * cast(R*) &" ~ implName ~ ".tupleof[i];
}
");
}
mixin("
S " ~ implName ~ ";
alias " ~ implName ~ " this;
");
}
}
else
alias TailShared = S;
}
@safe unittest
{
// No tail (no indirections) -> fully unshared.
static assert(is(TailShared!int == int));
static assert(is(TailShared!(shared int) == int));
static struct NoIndir { int i; }
static assert(is(TailShared!NoIndir == NoIndir));
static assert(is(TailShared!(shared NoIndir) == NoIndir));
// Tail can be independently shared or is already -> tail-shared.
static assert(is(TailShared!(int*) == shared(int)*));
static assert(is(TailShared!(shared int*) == shared(int)*));
static assert(is(TailShared!(shared(int)*) == shared(int)*));
static assert(is(TailShared!(int[]) == shared(int)[]));
static assert(is(TailShared!(shared int[]) == shared(int)[]));
static assert(is(TailShared!(shared(int)[]) == shared(int)[]));
static struct S1 { shared int* p; }
static assert(is(TailShared!S1 == S1));
static assert(is(TailShared!(shared S1) == S1));
static struct S2 { shared(int)* p; }
static assert(is(TailShared!S2 == S2));
static assert(is(TailShared!(shared S2) == S2));
// Tail follows shared-ness of head -> fully shared.
static class C { int i; }
static assert(is(TailShared!C == shared C));
static assert(is(TailShared!(shared C) == shared C));
/* However, structs get a wrapper that has getters which cast to
TailShared. */
static struct S3 { int* p; int _impl; int _impl_; int _impl__; }
static assert(!is(TailShared!S3 : S3));
static assert(is(TailShared!S3 : shared S3));
static assert(is(TailShared!(shared S3) == TailShared!S3));
static struct S4 { shared(int)** p; }
static assert(!is(TailShared!S4 : S4));
static assert(is(TailShared!S4 : shared S4));
static assert(is(TailShared!(shared S4) == TailShared!S4));
}
}
////////////////////////////////////////////////////////////////////////////////
// Unit Tests
////////////////////////////////////////////////////////////////////////////////
version (CoreUnittest)
{
version (D_LP64)
{
enum hasDWCAS = has128BitCAS;
}
else
{
enum hasDWCAS = has64BitCAS;
}
void testXCHG(T)(T val) pure nothrow @nogc @trusted
in
{
assert(val !is T.init);
}
do
{
T base = cast(T)null;
shared(T) atom = cast(shared(T))null;
assert(base !is val, T.stringof);
assert(atom is base, T.stringof);
assert(atomicExchange(&atom, val) is base, T.stringof);
assert(atom is val, T.stringof);
}
void testCAS(T)(T val) pure nothrow @nogc @trusted
in
{
assert(val !is T.init);
}
do
{
T base = cast(T)null;
shared(T) atom = cast(shared(T))null;
assert(base !is val, T.stringof);
assert(atom is base, T.stringof);
assert(cas(&atom, base, val), T.stringof);
assert(atom is val, T.stringof);
assert(!cas(&atom, base, base), T.stringof);
assert(atom is val, T.stringof);
atom = cast(shared(T))null;
shared(T) arg = base;
assert(cas(&atom, &arg, val), T.stringof);
assert(arg is base, T.stringof);
assert(atom is val, T.stringof);
arg = base;
assert(!cas(&atom, &arg, base), T.stringof);
assert(arg is val, T.stringof);
assert(atom is val, T.stringof);
}
void testLoadStore(MemoryOrder ms = MemoryOrder.seq, T)(T val = T.init + 1) pure nothrow @nogc @trusted
{
T base = cast(T) 0;
shared(T) atom = cast(T) 0;
assert(base !is val);
assert(atom is base);
atomicStore!(ms)(atom, val);
base = atomicLoad!(ms)(atom);
assert(base is val, T.stringof);
assert(atom is val);
}
void testType(T)(T val = T.init + 1) pure nothrow @nogc @safe
{
static if (T.sizeof < 8 || has64BitXCHG)
testXCHG!(T)(val);
testCAS!(T)(val);
testLoadStore!(MemoryOrder.seq, T)(val);
testLoadStore!(MemoryOrder.raw, T)(val);
}
@betterC @safe pure nothrow unittest
{
testType!(bool)();
testType!(byte)();
testType!(ubyte)();
testType!(short)();
testType!(ushort)();
testType!(int)();
testType!(uint)();
}
@safe pure nothrow unittest
{
testType!(shared int*)();
static interface Inter {}
static class KlassImpl : Inter {}
testXCHG!(shared Inter)(new shared(KlassImpl));
testCAS!(shared Inter)(new shared(KlassImpl));
static class Klass {}
testXCHG!(shared Klass)(new shared(Klass));
testCAS!(shared Klass)(new shared(Klass));
testXCHG!(shared int)(42);
testType!(float)(0.1f);
static if (has64BitCAS)
{
testType!(double)(0.1);
testType!(long)();
testType!(ulong)();
}
static if (has128BitCAS)
{
() @trusted
{
align(16) struct Big { long a, b; }
shared(Big) atom;
shared(Big) base;
shared(Big) arg;
shared(Big) val = Big(1, 2);
assert(cas(&atom, arg, val), Big.stringof);
assert(atom is val, Big.stringof);
assert(!cas(&atom, arg, val), Big.stringof);
assert(atom is val, Big.stringof);
atom = Big();
assert(cas(&atom, &arg, val), Big.stringof);
assert(arg is base, Big.stringof);
assert(atom is val, Big.stringof);
arg = Big();
assert(!cas(&atom, &arg, base), Big.stringof);
assert(arg is val, Big.stringof);
assert(atom is val, Big.stringof);
}();
}
shared(size_t) i;
atomicOp!"+="(i, cast(size_t) 1);
assert(i == 1);
atomicOp!"-="(i, cast(size_t) 1);
assert(i == 0);
shared float f = 0.1f;
atomicOp!"+="(f, 0.1f);
assert(f > 0.1999f && f < 0.2001f);
static if (has64BitCAS)
{
shared double d = 0.1;
atomicOp!"+="(d, 0.1);
assert(d > 0.1999 && d < 0.2001);
}
}
@betterC pure nothrow unittest
{
static if (has128BitCAS)
{
struct DoubleValue
{
long value1;
long value2;
}
align(16) shared DoubleValue a;
atomicStore(a, DoubleValue(1,2));
assert(a.value1 == 1 && a.value2 ==2);
while (!cas(&a, DoubleValue(1,2), DoubleValue(3,4))){}
assert(a.value1 == 3 && a.value2 ==4);
align(16) DoubleValue b = atomicLoad(a);
assert(b.value1 == 3 && b.value2 ==4);
}
static if (hasDWCAS)
{
static struct List { size_t gen; List* next; }
shared(List) head;
assert(cas(&head, shared(List)(0, null), shared(List)(1, cast(List*)1)));
assert(head.gen == 1);
assert(cast(size_t)head.next == 1);
}
// https://issues.dlang.org/show_bug.cgi?id=20629
static struct Struct
{
uint a, b;
}
shared Struct s1 = Struct(1, 2);
atomicStore(s1, Struct(3, 4));
assert(cast(uint) s1.a == 3);
assert(cast(uint) s1.b == 4);
}
// https://issues.dlang.org/show_bug.cgi?id=20844
static if (hasDWCAS)
{
debug: // tests CAS in-contract
pure nothrow unittest
{
import core.exception : AssertError;
align(16) shared ubyte[2 * size_t.sizeof + 1] data;
auto misalignedPointer = cast(size_t[2]*) &data[1];
size_t[2] x;
try
cas(misalignedPointer, x, x);
catch (AssertError)
return;
assert(0, "should have failed");
}
}
@betterC pure nothrow @nogc @safe unittest
{
int a;
if (casWeak!(MemoryOrder.acq_rel, MemoryOrder.raw)(&a, 0, 4))
assert(a == 4);
}
@betterC pure nothrow unittest
{
static struct S { int val; }
auto s = shared(S)(1);
shared(S*) ptr;
// head unshared
shared(S)* ifThis = null;
shared(S)* writeThis = &s;
assert(ptr is null);
assert(cas(&ptr, ifThis, writeThis));
assert(ptr is writeThis);
// head shared
shared(S*) ifThis2 = writeThis;
shared(S*) writeThis2 = null;
assert(cas(&ptr, ifThis2, writeThis2));
assert(ptr is null);
}
unittest
{
import core.thread;
// Use heap memory to ensure an optimizing
// compiler doesn't put things in registers.
uint* x = new uint();
bool* f = new bool();
uint* r = new uint();
auto thr = new Thread(()
{
while (!*f)
{
}
atomicFence();
*r = *x;
});
thr.start();
*x = 42;
atomicFence();
*f = true;
atomicFence();
thr.join();
assert(*r == 42);
}
// === atomicFetchAdd and atomicFetchSub operations ====
@betterC pure nothrow @nogc @safe unittest
{
shared ubyte u8 = 1;
shared ushort u16 = 2;
shared uint u32 = 3;
shared byte i8 = 5;
shared short i16 = 6;
shared int i32 = 7;
assert(atomicOp!"+="(u8, 8) == 9);
assert(atomicOp!"+="(u16, 8) == 10);
assert(atomicOp!"+="(u32, 8) == 11);
assert(atomicOp!"+="(i8, 8) == 13);
assert(atomicOp!"+="(i16, 8) == 14);
assert(atomicOp!"+="(i32, 8) == 15);
version (D_LP64)
{
shared ulong u64 = 4;
shared long i64 = 8;
assert(atomicOp!"+="(u64, 8) == 12);
assert(atomicOp!"+="(i64, 8) == 16);
}
}
@betterC pure nothrow @nogc unittest
{
byte[10] byteArray = [1, 3, 5, 7, 9, 11, 13, 15, 17, 19];
ulong[10] ulongArray = [2, 4, 6, 8, 10, 12, 14, 16, 19, 20];
{
auto array = byteArray;
byte* ptr = &array[0];
byte* prevPtr = atomicFetchAdd(ptr, 3);
assert(prevPtr == &array[0]);
assert(*prevPtr == 1);
assert(*ptr == 7);
}
{
auto array = ulongArray;
ulong* ptr = &array[0];
ulong* prevPtr = atomicFetchAdd(ptr, 3);
assert(prevPtr == &array[0]);
assert(*prevPtr == 2);
assert(*ptr == 8);
}
}
@betterC pure nothrow @nogc @safe unittest
{
shared ubyte u8 = 1;
shared ushort u16 = 2;
shared uint u32 = 3;
shared byte i8 = 5;
shared short i16 = 6;
shared int i32 = 7;
assert(atomicOp!"-="(u8, 1) == 0);
assert(atomicOp!"-="(u16, 1) == 1);
assert(atomicOp!"-="(u32, 1) == 2);
assert(atomicOp!"-="(i8, 1) == 4);
assert(atomicOp!"-="(i16, 1) == 5);
assert(atomicOp!"-="(i32, 1) == 6);
version (D_LP64)
{
shared ulong u64 = 4;
shared long i64 = 8;
assert(atomicOp!"-="(u64, 1) == 3);
assert(atomicOp!"-="(i64, 1) == 7);
}
}
@betterC pure nothrow @nogc unittest
{
byte[10] byteArray = [1, 3, 5, 7, 9, 11, 13, 15, 17, 19];
ulong[10] ulongArray = [2, 4, 6, 8, 10, 12, 14, 16, 19, 20];
{
auto array = byteArray;
byte* ptr = &array[5];
byte* prevPtr = atomicFetchSub(ptr, 4);
assert(prevPtr == &array[5]);
assert(*prevPtr == 11);
assert(*ptr == 3); // https://issues.dlang.org/show_bug.cgi?id=21578
}
{
auto array = ulongArray;
ulong* ptr = &array[5];
ulong* prevPtr = atomicFetchSub(ptr, 4);
assert(prevPtr == &array[5]);
assert(*prevPtr == 12);
assert(*ptr == 4); // https://issues.dlang.org/show_bug.cgi?id=21578
}
}
@betterC pure nothrow @nogc @safe unittest // issue 16651
{
shared ulong a = 2;
uint b = 1;
atomicOp!"-="(a, b);
assert(a == 1);
shared uint c = 2;
ubyte d = 1;
atomicOp!"-="(c, d);
assert(c == 1);
}
pure nothrow @safe unittest // issue 16230
{
shared int i;
static assert(is(typeof(atomicLoad(i)) == int));
shared int* p;
static assert(is(typeof(atomicLoad(p)) == shared(int)*));
shared int[] a;
static if (__traits(compiles, atomicLoad(a)))
{
static assert(is(typeof(atomicLoad(a)) == shared(int)[]));
}
static struct S { int* _impl; }
shared S s;
static assert(is(typeof(atomicLoad(s)) : shared S));
static assert(is(typeof(atomicLoad(s)._impl) == shared(int)*));
auto u = atomicLoad(s);
assert(u._impl is null);
u._impl = new shared int(42);
assert(atomicLoad(*u._impl) == 42);
static struct S2 { S s; }
shared S2 s2;
static assert(is(typeof(atomicLoad(s2).s) == TailShared!S));
static struct S3 { size_t head; int* tail; }
shared S3 s3;
static if (__traits(compiles, atomicLoad(s3)))
{
static assert(is(typeof(atomicLoad(s3).head) == size_t));
static assert(is(typeof(atomicLoad(s3).tail) == shared(int)*));
}
static class C { int i; }
shared C c;
static assert(is(typeof(atomicLoad(c)) == shared C));
static struct NoIndirections { int i; }
shared NoIndirections n;
static assert(is(typeof(atomicLoad(n)) == NoIndirections));
}
unittest // Issue 21631
{
shared uint si1 = 45;
shared uint si2 = 38;
shared uint* psi = &si1;
assert((&psi).cas(cast(const) psi, &si2));
}
}