1
0
mirror of https://github.com/RPCS3/rpcs3.git synced 2024-11-25 12:12:50 +01:00

Implement atomic_t<>::release

More relaxed store with release memory order
This commit is contained in:
Nekotekina 2019-01-28 18:39:39 +03:00
parent 50922faac9
commit 2b66abaf10
4 changed files with 60 additions and 26 deletions

View File

@ -4,7 +4,7 @@
#include <functional>
// Helper class, provides access to compiler-specific atomic intrinsics
template<typename T, std::size_t Size = sizeof(T)>
template <typename T, std::size_t Size = sizeof(T)>
struct atomic_storage
{
static_assert(sizeof(T) <= 16 && sizeof(T) == alignof(T), "atomic_storage<> error: invalid type");
@ -29,6 +29,11 @@ struct atomic_storage
__atomic_store(&dest, &value, __ATOMIC_SEQ_CST);
}
static inline void release(T& dest, T value)
{
__atomic_store(&dest, &value, __ATOMIC_RELEASE);
}
static inline T exchange(T& dest, T value)
{
T result;
@ -176,7 +181,7 @@ struct atomic_storage
/* The rest: ugly MSVC intrinsics + inline asm implementations */
template<typename T>
template <typename T>
struct atomic_storage<T, 1> : atomic_storage<T, 0>
{
#ifdef _MSC_VER
@ -200,6 +205,12 @@ struct atomic_storage<T, 1> : atomic_storage<T, 0>
_InterlockedExchange8((volatile char*)&dest, (char&)value);
}
static inline void release(T& dest, T value)
{
_ReadWriteBarrier();
*(volatile char*)&dest = (char&)value;
}
static inline T exchange(T& dest, T value)
{
char r = _InterlockedExchange8((volatile char*)&dest, (char&)value);
@ -232,7 +243,7 @@ struct atomic_storage<T, 1> : atomic_storage<T, 0>
#endif
};
template<typename T>
template <typename T>
struct atomic_storage<T, 2> : atomic_storage<T, 0>
{
#ifdef _MSC_VER
@ -256,6 +267,12 @@ struct atomic_storage<T, 2> : atomic_storage<T, 0>
_InterlockedExchange16((volatile short*)&dest, (short&)value);
}
static inline void release(T& dest, T value)
{
_ReadWriteBarrier();
*(volatile short*)&dest = (short&)value;
}
static inline T exchange(T& dest, T value)
{
short r = _InterlockedExchange16((volatile short*)&dest, (short&)value);
@ -324,7 +341,7 @@ struct atomic_storage<T, 2> : atomic_storage<T, 0>
#endif
};
template<typename T>
template <typename T>
struct atomic_storage<T, 4> : atomic_storage<T, 0>
{
#ifdef _MSC_VER
@ -348,6 +365,12 @@ struct atomic_storage<T, 4> : atomic_storage<T, 0>
_InterlockedExchange((volatile long*)&dest, (long&)value);
}
static inline void release(T& dest, T value)
{
_ReadWriteBarrier();
*(volatile long*)&dest = (long&)value;
}
static inline T exchange(T& dest, T value)
{
long r = _InterlockedExchange((volatile long*)&dest, (long&)value);
@ -423,7 +446,7 @@ struct atomic_storage<T, 4> : atomic_storage<T, 0>
#endif
};
template<typename T>
template <typename T>
struct atomic_storage<T, 8> : atomic_storage<T, 0>
{
#ifdef _MSC_VER
@ -447,6 +470,12 @@ struct atomic_storage<T, 8> : atomic_storage<T, 0>
_InterlockedExchange64((volatile llong*)&dest, (llong&)value);
}
static inline void release(T& dest, T value)
{
_ReadWriteBarrier();
*(volatile llong*)&dest = (llong&)value;
}
static inline T exchange(T& dest, T value)
{
llong r = _InterlockedExchange64((volatile llong*)&dest, (llong&)value);
@ -525,7 +554,7 @@ struct atomic_storage<T, 8> : atomic_storage<T, 0>
#endif
};
template<typename T>
template <typename T>
struct atomic_storage<T, 16> : atomic_storage<T, 0>
{
#ifdef _MSC_VER
@ -550,6 +579,14 @@ struct atomic_storage<T, 16> : atomic_storage<T, 0>
while (!_InterlockedCompareExchange128((volatile llong*)&dest, hi, lo, cmp));
}
static inline void release(T& dest, T value)
{
llong lo = *(llong*)&value;
llong hi = *((llong*)&value + 1);
llong cmp[2]{ *(volatile llong*)&dest, *((volatile llong*)&dest + 1) };
while (!_InterlockedCompareExchange128((volatile llong*)&dest, hi, lo, cmp));
}
static inline T exchange(T& dest, T value)
{
llong lo = *(llong*)&value;
@ -764,6 +801,12 @@ public:
return rhs;
}
// Atomically write data with release memory order (faster on x86)
void release(const type& rhs)
{
atomic_storage<type>::release(m_data, rhs);
}
// Atomically replace data with value, return previous data value
type exchange(const type& rhs)
{

View File

@ -543,11 +543,7 @@ void spu_recompiler_base::dispatch(spu_thread& spu, void*, u8* rip)
// If code verification failed from a patched patchpoint, clear it with a single NOP
if (rip)
{
#ifdef _MSC_VER
*(volatile u64*)(rip) = 0x841f0f;
#else
__atomic_store_n(reinterpret_cast<u64*>(rip), 0x841f0f, __ATOMIC_RELAXED);
#endif
atomic_storage<u64>::release(*reinterpret_cast<u64*>(rip), 0x841f0f);
}
// Second attempt (recover from the recursion after repeated unsuccessful trampoline call)
@ -580,7 +576,11 @@ void spu_recompiler_base::branch(spu_thread& spu, void*, u8* rip)
// Overwrite jump to this function with jump to the compiled function
const s64 rel = reinterpret_cast<u64>(func) - reinterpret_cast<u64>(rip) - 5;
alignas(8) u8 bytes[8];
union
{
u8 bytes[8];
u64 result;
};
if (rel >= INT32_MIN && rel <= INT32_MAX)
{
@ -609,11 +609,7 @@ void spu_recompiler_base::branch(spu_thread& spu, void*, u8* rip)
std::memset(bytes + 3, 0x00, 5);
}
#ifdef _MSC_VER
*(volatile u64*)(rip) = *reinterpret_cast<u64*>(+bytes);
#else
__atomic_store_n(reinterpret_cast<u64*>(rip), *reinterpret_cast<u64*>(+bytes), __ATOMIC_RELAXED);
#endif
atomic_storage<u64>::release(*reinterpret_cast<u64*>(rip), result);
}
std::vector<u32> spu_recompiler_base::block(const be_t<u32>* ls, u32 entry_point)

View File

@ -245,12 +245,7 @@ public:
void set_value(u32 value, bool count = true)
{
const u64 new_data = u64{count} << off_count | value;
#ifdef _MSC_VER
const_cast<volatile u64&>(data.raw()) = new_data;
#else
__atomic_store_n(&data.raw(), new_data, __ATOMIC_RELAXED);
#endif
data.release(u64{count} << off_count | value);
}
u32 get_value()

View File

@ -85,7 +85,7 @@ namespace vm
if (!lock && lock.compare_and_swap_test(0, lock_info))
{
return &lock;
}
}
}
}
}
@ -256,7 +256,7 @@ namespace vm
{
const u64 value = lock;
// Test beginning address
// Test beginning address
if (static_cast<u32>(value) > addr)
{
break;
@ -295,7 +295,7 @@ namespace vm
writer_lock::~writer_lock()
{
g_addr_lock.raw() = 0;
g_addr_lock.release(0);
g_mutex.unlock();
}