1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
|
#pragma once
namespace Balau {
namespace Atomic {
#if (__GNUC__ >= 5) || ((__GNUC__ == 4) && ((__GNUC_MINOR__ >= 1)))
// gcc version of the atomic operations
template <class T> T Or(volatile T * ptr, T mask) { __sync_or_and_fetch(ptr, mask); }
template <class T> T And(volatile T * ptr, T mask) { __sync_and_and_fetch(ptr, mask); }
template <class T> T Xor(volatile T * ptr, T mask) { __sync_xor_and_fetch(ptr, mask); }
template <class T> T Nand(volatile T * ptr, T mask) { __sync_nand_and_fetch(ptr, mask); }
template <class T> T Increment(volatile T * ptr, T delta = 1) { __sync_add_and_fetch(ptr, delta); }
template <class T> T Decrement(volatile T * ptr, T delta = 1) { __sync_sub_and_fetch(ptr, delta); }
namespace Prefetch {
template <class T> T Or(volatile T * ptr, T mask) { __sync_fetch_and_or(ptr, mask); }
template <class T> T And(volatile T * ptr, T mask) { __sync_fetch_and_and(ptr, mask); }
template <class T> T Xor(volatile T * ptr, T mask) { __sync_fetch_and_xor(ptr, mask); }
template <class T> T Nand(volatile T * ptr, T mask) { __sync_fetch_and_nand(ptr, mask); }
template <class T> T Increment(volatile T * ptr, T delta = 1) { __sync_fetch_and_add(ptr, delta); }
template <class T> T Decrement(volatile T * ptr, T delta = 1) { __sync_fetch_and_sub(ptr, delta); }
};
template <class T> T CmpXChgVal(volatile T * ptr, const T xch, const T cmp) { return __sync_val_compare_and_swap(ptr, cmp, xch); }
template <class T> bool CmpXChgBool(volatile T * ptr, const T xch, const T cmp) { return __sync_bool_compare_and_swap(ptr, cmp, xch); }
template <class T> T Exchange32(volatile T * ptr, const T exchange) {
#if defined(i386) || defined (__x86_64)
__asm__ __volatile__("lock xchgl %0, (%1)" : "+r"(exchange) : "r"(ptr));
return exchange;
#else
T p;
do { p = *ptr; } while (!__sync_bool_compare_and_swap(ptr, p, exchange));
return p;
#endif
}
template <class T> T Exchange64(volatile T * ptr, const T exchange) {
#if defined(i386) || defined (__x86_64)
__asm__ __volatile__("lock xchgq %0, (%1)" : "+r"(exchange) : "r"(ptr));
return exchange;
#else
T p;
do { p = *ptr; } while (!__sync_bool_compare_and_swap(ptr, p, exchange));
return p;
#endif
}
#else
#ifdef _MSVC
// Visual Studio version of the atomic operations
#error MSVC not yet implemented.
#else
#error No known platform for atomic operations.
#endif
#endif
template <class T> T * ExchangePtr(T * volatile * ptr, const T * exchange) {
#if defined (__x86_64)
return Exchange64(ptr, exchange);
#else
return Exchange32(ptr, exchange);
#endif
}
};
};
|