mirror of
https://github.com/RPCS3/rpcs3.git
synced 2024-11-22 10:42:36 +01:00
cf13dee1f3
Should fix some deadlock issues.
330 lines
6.2 KiB
C++
330 lines
6.2 KiB
C++
#pragma once
|
|
|
|
#include <mutex>
|
|
#include "types.h"
|
|
#include "util/atomic.hpp"
|
|
|
|
// Shared mutex with small size (u32).
|
|
class shared_mutex final
|
|
{
|
|
enum : u32
|
|
{
|
|
c_one = 1u << 14, // Fixed-point 1.0 value (one writer, max_readers = c_one - 1)
|
|
c_sig = 1u << 30,
|
|
c_err = 1u << 31,
|
|
c_vip = 1u << 7,
|
|
};
|
|
|
|
atomic_t<u32> m_value{};
|
|
|
|
void imp_lock_shared(u32 val);
|
|
void imp_unlock_shared(u32 old);
|
|
void imp_lock_low(u32 val);
|
|
void imp_unlock_low(u32 old);
|
|
void imp_lock_vip(u32 val);
|
|
void imp_unlock_vip(u32 old);
|
|
void imp_wait();
|
|
void imp_signal();
|
|
void imp_lock(u32 val);
|
|
void imp_unlock(u32 old);
|
|
void imp_lock_upgrade();
|
|
void imp_lock_unlock();
|
|
|
|
public:
|
|
constexpr shared_mutex() = default;
|
|
|
|
bool try_lock_shared()
|
|
{
|
|
const u32 value = m_value.load();
|
|
|
|
// Conditional increment
|
|
return value < c_one - 1 && m_value.compare_and_swap_test(value, value + 1);
|
|
}
|
|
|
|
void lock_shared()
|
|
{
|
|
const u32 value = m_value.load();
|
|
|
|
if (UNLIKELY(value >= c_one - 1 || !m_value.compare_and_swap_test(value, value + 1)))
|
|
{
|
|
imp_lock_shared(value);
|
|
}
|
|
}
|
|
|
|
void lock_shared_hle()
|
|
{
|
|
const u32 value = m_value.load();
|
|
|
|
if (LIKELY(value < c_one - 1))
|
|
{
|
|
u32 old = value;
|
|
if (LIKELY(atomic_storage<u32>::compare_exchange_hle_acq(m_value.raw(), old, value + 1)))
|
|
{
|
|
return;
|
|
}
|
|
}
|
|
|
|
imp_lock_shared(value);
|
|
}
|
|
|
|
void unlock_shared()
|
|
{
|
|
// Unconditional decrement (can result in broken state)
|
|
const u32 value = m_value.fetch_sub(1);
|
|
|
|
if (UNLIKELY(value >= c_one))
|
|
{
|
|
imp_unlock_shared(value);
|
|
}
|
|
}
|
|
|
|
void unlock_shared_hle()
|
|
{
|
|
const u32 value = atomic_storage<u32>::fetch_add_hle_rel(m_value.raw(), -1);
|
|
|
|
if (UNLIKELY(value >= c_one))
|
|
{
|
|
imp_unlock_shared(value);
|
|
}
|
|
}
|
|
|
|
bool try_lock_low()
|
|
{
|
|
const u32 value = m_value.load();
|
|
|
|
// Conditional increment
|
|
return value < c_vip - 1 && m_value.compare_and_swap_test(value, value + 1);
|
|
}
|
|
|
|
void lock_low()
|
|
{
|
|
const u32 value = m_value.load();
|
|
|
|
if (UNLIKELY(value >= c_vip - 1 || !m_value.compare_and_swap_test(value, value + 1)))
|
|
{
|
|
imp_lock_low(value);
|
|
}
|
|
}
|
|
|
|
void unlock_low()
|
|
{
|
|
// Unconditional decrement (can result in broken state)
|
|
const u32 value = m_value.fetch_sub(1);
|
|
|
|
if (UNLIKELY(value >= c_one))
|
|
{
|
|
imp_unlock_low(value);
|
|
}
|
|
}
|
|
|
|
bool try_lock_vip()
|
|
{
|
|
const u32 value = m_value.load();
|
|
|
|
// Conditional increment
|
|
return (value < c_one - 1 || value & (c_one - c_vip)) && (value % c_vip) == 0 && m_value.compare_and_swap_test(value, value + c_vip);
|
|
}
|
|
|
|
void lock_vip()
|
|
{
|
|
const u32 value = m_value.load();
|
|
|
|
if (UNLIKELY((value >= c_one - 1 && !(value & (c_one - c_vip))) || (value % c_vip) || !m_value.compare_and_swap_test(value, value + c_vip)))
|
|
{
|
|
imp_lock_vip(value);
|
|
}
|
|
}
|
|
|
|
void unlock_vip()
|
|
{
|
|
// Unconditional decrement (can result in broken state)
|
|
const u32 value = m_value.fetch_sub(c_vip);
|
|
|
|
if (UNLIKELY(value >= c_one))
|
|
{
|
|
imp_unlock_vip(value);
|
|
}
|
|
}
|
|
|
|
bool try_lock()
|
|
{
|
|
return m_value.compare_and_swap_test(0, c_one);
|
|
}
|
|
|
|
void lock()
|
|
{
|
|
const u32 value = m_value.compare_and_swap(0, c_one);
|
|
|
|
if (UNLIKELY(value))
|
|
{
|
|
imp_lock(value);
|
|
}
|
|
}
|
|
|
|
void lock_hle()
|
|
{
|
|
u32 value = 0;
|
|
|
|
if (UNLIKELY(!atomic_storage<u32>::compare_exchange_hle_acq(m_value.raw(), value, c_one)))
|
|
{
|
|
imp_lock(value);
|
|
}
|
|
}
|
|
|
|
void unlock()
|
|
{
|
|
// Unconditional decrement (can result in broken state)
|
|
const u32 value = m_value.fetch_sub(c_one);
|
|
|
|
if (UNLIKELY(value != c_one))
|
|
{
|
|
imp_unlock(value);
|
|
}
|
|
}
|
|
|
|
void unlock_hle()
|
|
{
|
|
const u32 value = atomic_storage<u32>::fetch_add_hle_rel(m_value.raw(), 0u - c_one);
|
|
|
|
if (UNLIKELY(value != c_one))
|
|
{
|
|
imp_unlock(value);
|
|
}
|
|
}
|
|
|
|
bool try_lock_upgrade()
|
|
{
|
|
const u32 value = m_value.load();
|
|
|
|
// Conditional increment, try to convert a single reader into a writer, ignoring other writers
|
|
return (value + c_one - 1) % c_one == 0 && m_value.compare_and_swap_test(value, value + c_one - 1);
|
|
}
|
|
|
|
void lock_upgrade()
|
|
{
|
|
if (UNLIKELY(!try_lock_upgrade()))
|
|
{
|
|
imp_lock_upgrade();
|
|
}
|
|
}
|
|
|
|
void lock_downgrade()
|
|
{
|
|
// Convert to reader lock (can result in broken state)
|
|
m_value -= c_one - 1;
|
|
}
|
|
|
|
void lock_downgrade_to_vip()
|
|
{
|
|
// Convert to vip lock (can result in broken state)
|
|
m_value -= c_one - c_vip;
|
|
}
|
|
|
|
// Optimized wait for lockability without locking, relaxed
|
|
void lock_unlock()
|
|
{
|
|
if (UNLIKELY(m_value != 0))
|
|
{
|
|
imp_lock_unlock();
|
|
}
|
|
}
|
|
|
|
// Check whether can immediately obtain an exclusive (writer) lock
|
|
bool is_free() const
|
|
{
|
|
return m_value.load() == 0;
|
|
}
|
|
|
|
// Check whether can immediately obtain a shared (reader) lock
|
|
bool is_lockable() const
|
|
{
|
|
return m_value.load() < c_one - 1;
|
|
}
|
|
|
|
// Special purpose logic
|
|
bool downgrade_unique_vip_lock_to_low_or_unlock();
|
|
};
|
|
|
|
// Simplified shared (reader) lock implementation. Mutually incompatible with low_lock and vip_lock.
|
|
class reader_lock final
|
|
{
|
|
shared_mutex& m_mutex;
|
|
bool m_upgraded = false;
|
|
|
|
public:
|
|
reader_lock(const reader_lock&) = delete;
|
|
|
|
reader_lock& operator=(const reader_lock&) = delete;
|
|
|
|
explicit reader_lock(shared_mutex& mutex)
|
|
: m_mutex(mutex)
|
|
{
|
|
m_mutex.lock_shared();
|
|
}
|
|
|
|
// One-way lock upgrade; note that the observed state could have been changed
|
|
void upgrade()
|
|
{
|
|
if (!m_upgraded)
|
|
{
|
|
m_mutex.lock_upgrade();
|
|
m_upgraded = true;
|
|
}
|
|
}
|
|
|
|
// Try to upgrade; if it succeeds, the observed state has NOT been changed
|
|
bool try_upgrade()
|
|
{
|
|
return m_upgraded || (m_upgraded = m_mutex.try_lock_upgrade());
|
|
}
|
|
|
|
~reader_lock()
|
|
{
|
|
m_upgraded ? m_mutex.unlock() : m_mutex.unlock_shared();
|
|
}
|
|
};
|
|
|
|
// Special shared (reader) lock, mutually exclusive with vip locks. Mutually incompatible with normal shared (reader) lock.
|
|
class low_lock final
|
|
{
|
|
shared_mutex& m_mutex;
|
|
|
|
public:
|
|
low_lock(const low_lock&) = delete;
|
|
|
|
low_lock& operator=(const low_lock&) = delete;
|
|
|
|
explicit low_lock(shared_mutex& mutex)
|
|
: m_mutex(mutex)
|
|
{
|
|
m_mutex.lock_low();
|
|
}
|
|
|
|
~low_lock()
|
|
{
|
|
m_mutex.unlock_low();
|
|
}
|
|
};
|
|
|
|
// Special shared (reader) lock, mutually exclusive with low locks. Mutually incompatible with normal shared (reader) lock.
|
|
class vip_lock final
|
|
{
|
|
shared_mutex& m_mutex;
|
|
|
|
public:
|
|
vip_lock(const vip_lock&) = delete;
|
|
|
|
vip_lock& operator=(const vip_lock&) = delete;
|
|
|
|
explicit vip_lock(shared_mutex& mutex)
|
|
: m_mutex(mutex)
|
|
{
|
|
m_mutex.lock_vip();
|
|
}
|
|
|
|
~vip_lock()
|
|
{
|
|
m_mutex.unlock_vip();
|
|
}
|
|
};
|