1
0
mirror of https://github.com/RPCS3/rpcs3.git synced 2024-11-26 04:32:35 +01:00
rpcs3/Utilities/GNU.h

391 lines
11 KiB
C
Raw Normal View History

2013-11-19 11:30:58 +01:00
#pragma once
2014-10-07 15:35:44 +02:00
#include <emmintrin.h>
#ifdef _WIN32
#define thread_local __declspec(thread)
#elif __APPLE__
#define thread_local __thread
#endif
2014-07-14 21:15:30 +02:00
#ifdef _WIN32
2014-08-22 16:21:55 +02:00
#define __noinline __declspec(noinline)
2014-07-14 21:15:30 +02:00
#else
2014-08-22 16:21:55 +02:00
#define __noinline __attribute__((noinline))
2014-07-14 21:15:30 +02:00
#endif
#ifdef _WIN32
#define __safebuffers __declspec(safebuffers)
#else
#define __safebuffers
#endif
2014-07-01 17:34:25 +02:00
template<size_t size>
2014-09-19 02:19:22 +02:00
void strcpy_trunc(char(&dst)[size], const std::string& src)
2014-07-01 17:34:25 +02:00
{
const size_t count = (src.size() >= size) ? size - 1 /* truncation */ : src.size();
memcpy(dst, src.c_str(), count);
dst[count] = 0;
}
2014-09-19 02:19:22 +02:00
template<size_t size, size_t rsize>
void strcpy_trunc(char(&dst)[size], const char(&src)[rsize])
{
const size_t count = (rsize >= size) ? size - 1 /* truncation */ : rsize;
memcpy(dst, src, count);
dst[count] = 0;
}
2013-11-19 11:30:58 +01:00
#if defined(__GNUG__)
#include <cmath>
#include <stdlib.h>
#include <cstdint>
#ifndef __APPLE__
#include <malloc.h>
#endif
#define _fpclass(x) std::fpclassify(x)
2013-11-19 11:30:58 +01:00
#define __forceinline __attribute__((always_inline))
#define _byteswap_ushort(x) __builtin_bswap16(x)
#define _byteswap_ulong(x) __builtin_bswap32(x)
#define _byteswap_uint64(x) __builtin_bswap64(x)
#define INFINITE 0xFFFFFFFF
2014-02-21 17:13:57 +01:00
#define _CRT_ALIGN(x) __attribute__((aligned(x)))
inline uint64_t __umulh(uint64_t a, uint64_t b)
{
uint64_t result;
__asm__("mulq %[b]" : "=d" (result) : [a] "a" (a), [b] "rm" (b));
return result;
}
inline int64_t __mulh(int64_t a, int64_t b)
{
int64_t result;
__asm__("imulq %[b]" : "=d" (result) : [a] "a" (a), [b] "rm" (b));
return result;
}
2014-07-12 09:02:39 +02:00
void * _aligned_malloc(size_t size, size_t alignment);
2014-07-12 09:02:39 +02:00
#ifdef __APPLE__
2014-04-29 22:10:42 +02:00
int clock_gettime(int foo, struct timespec *ts);
#define wxIsNaN(x) ((x) != (x))
#ifndef CLOCK_MONOTONIC
#define CLOCK_MONOTONIC 0
2014-04-29 22:10:42 +02:00
#endif /* !CLOCK_MONOTONIC */
2014-07-12 09:02:39 +02:00
#endif /* __APPLE__ */
2014-04-29 22:10:42 +02:00
#define _aligned_free free
#define DWORD int32_t
2013-11-19 11:30:58 +01:00
#endif
#ifndef InterlockedCompareExchange
static __forceinline uint8_t InterlockedCompareExchange(volatile uint8_t* dest, uint8_t exch, uint8_t comp)
{
2014-10-10 20:19:14 +02:00
#if defined(__GNUG__)
return __sync_val_compare_and_swap(dest, comp, exch);
#else
return _InterlockedCompareExchange8((volatile char*)dest, exch, comp);
2014-10-10 20:19:14 +02:00
#endif
}
2014-09-23 16:27:18 +02:00
static __forceinline uint16_t InterlockedCompareExchange(volatile uint16_t* dest, uint16_t exch, uint16_t comp)
{
2014-10-10 20:19:14 +02:00
#if defined(__GNUG__)
return __sync_val_compare_and_swap(dest, comp, exch);
#else
2014-09-23 16:27:18 +02:00
return _InterlockedCompareExchange16((volatile short*)dest, exch, comp);
2014-10-10 20:19:14 +02:00
#endif
2014-09-23 16:27:18 +02:00
}
static __forceinline uint32_t InterlockedCompareExchange(volatile uint32_t* dest, uint32_t exch, uint32_t comp)
{
2014-10-10 20:19:14 +02:00
#if defined(__GNUG__)
return __sync_val_compare_and_swap(dest, comp, exch);
#else
return _InterlockedCompareExchange((volatile long*)dest, exch, comp);
2014-10-10 20:19:14 +02:00
#endif
}
static __forceinline uint64_t InterlockedCompareExchange(volatile uint64_t* dest, uint64_t exch, uint64_t comp)
{
2014-10-10 20:19:14 +02:00
#if defined(__GNUG__)
return __sync_val_compare_and_swap(dest, comp, exch);
#else
return _InterlockedCompareExchange64((volatile long long*)dest, exch, comp);
2014-10-10 20:19:14 +02:00
#endif
}
#endif
2014-10-10 20:19:14 +02:00
static __forceinline bool InterlockedCompareExchangeTest(volatile uint8_t* dest, uint8_t exch, uint8_t comp)
{
#if defined(__GNUG__)
return __sync_bool_compare_and_swap(dest, comp, exch);
#else
2014-10-16 19:07:41 +02:00
return (uint8_t)_InterlockedCompareExchange8((volatile char*)dest, exch, comp) == comp;
2014-10-10 20:19:14 +02:00
#endif
}
static __forceinline bool InterlockedCompareExchangeTest(volatile uint16_t* dest, uint16_t exch, uint16_t comp)
{
#if defined(__GNUG__)
return __sync_bool_compare_and_swap(dest, comp, exch);
#else
2014-10-16 19:07:41 +02:00
return (uint16_t)_InterlockedCompareExchange16((volatile short*)dest, exch, comp) == comp;
2014-10-10 20:19:14 +02:00
#endif
}
static __forceinline bool InterlockedCompareExchangeTest(volatile uint32_t* dest, uint32_t exch, uint32_t comp)
{
#if defined(__GNUG__)
return __sync_bool_compare_and_swap(dest, comp, exch);
#else
2014-10-16 19:07:41 +02:00
return (uint32_t)_InterlockedCompareExchange((volatile long*)dest, exch, comp) == comp;
2014-10-10 20:19:14 +02:00
#endif
}
static __forceinline bool InterlockedCompareExchangeTest(volatile uint64_t* dest, uint64_t exch, uint64_t comp)
{
#if defined(__GNUG__)
return __sync_bool_compare_and_swap(dest, comp, exch);
#else
2014-10-16 19:07:41 +02:00
return (uint64_t)_InterlockedCompareExchange64((volatile long long*)dest, exch, comp) == comp;
2014-10-10 20:19:14 +02:00
#endif
}
#ifndef InterlockedExchange
static __forceinline uint8_t InterlockedExchange(volatile uint8_t* dest, uint8_t value)
{
2014-10-10 20:19:14 +02:00
#if defined(__GNUG__)
return __sync_lock_test_and_set(dest, value);
#else
return _InterlockedExchange8((volatile char*)dest, value);
2014-10-10 20:19:14 +02:00
#endif
}
2014-09-23 16:27:18 +02:00
static __forceinline uint16_t InterlockedExchange(volatile uint16_t* dest, uint16_t value)
{
2014-10-10 20:19:14 +02:00
#if defined(__GNUG__)
return __sync_lock_test_and_set(dest, value);
#else
2014-09-23 16:27:18 +02:00
return _InterlockedExchange16((volatile short*)dest, value);
2014-10-10 20:19:14 +02:00
#endif
2014-09-23 16:27:18 +02:00
}
static __forceinline uint32_t InterlockedExchange(volatile uint32_t* dest, uint32_t value)
{
2014-10-10 20:19:14 +02:00
#if defined(__GNUG__)
return __sync_lock_test_and_set(dest, value);
#else
return _InterlockedExchange((volatile long*)dest, value);
2014-10-10 20:19:14 +02:00
#endif
}
static __forceinline uint64_t InterlockedExchange(volatile uint64_t* dest, uint64_t value)
{
2014-10-10 20:19:14 +02:00
#if defined(__GNUG__)
return __sync_lock_test_and_set(dest, value);
#else
return _InterlockedExchange64((volatile long long*)dest, value);
2014-10-10 20:19:14 +02:00
#endif
}
#endif
#ifndef InterlockedOr
static __forceinline uint8_t InterlockedOr(volatile uint8_t* dest, uint8_t value)
{
2014-10-10 20:19:14 +02:00
#if defined(__GNUG__)
return __sync_fetch_and_or(dest, value);
#else
return _InterlockedOr8((volatile char*)dest, value);
2014-10-10 20:19:14 +02:00
#endif
}
2014-09-23 16:27:18 +02:00
static __forceinline uint16_t InterlockedOr(volatile uint16_t* dest, uint16_t value)
{
2014-10-10 20:19:14 +02:00
#if defined(__GNUG__)
return __sync_fetch_and_or(dest, value);
#else
2014-09-23 16:27:18 +02:00
return _InterlockedOr16((volatile short*)dest, value);
2014-10-10 20:19:14 +02:00
#endif
2014-09-23 16:27:18 +02:00
}
static __forceinline uint32_t InterlockedOr(volatile uint32_t* dest, uint32_t value)
{
2014-10-10 20:19:14 +02:00
#if defined(__GNUG__)
return __sync_fetch_and_or(dest, value);
#else
return _InterlockedOr((volatile long*)dest, value);
2014-10-10 20:19:14 +02:00
#endif
}
static __forceinline uint64_t InterlockedOr(volatile uint64_t* dest, uint64_t value)
{
2014-10-10 20:19:14 +02:00
#if defined(__GNUG__)
return __sync_fetch_and_or(dest, value);
#else
return _InterlockedOr64((volatile long long*)dest, value);
2014-10-10 20:19:14 +02:00
#endif
}
#endif
#ifndef InterlockedAnd
static __forceinline uint8_t InterlockedAnd(volatile uint8_t* dest, uint8_t value)
{
2014-10-10 20:19:14 +02:00
#if defined(__GNUG__)
return __sync_fetch_and_and(dest, value);
#else
return _InterlockedAnd8((volatile char*)dest, value);
2014-10-10 20:19:14 +02:00
#endif
}
2014-09-23 16:27:18 +02:00
static __forceinline uint16_t InterlockedAnd(volatile uint16_t* dest, uint16_t value)
{
2014-10-10 20:19:14 +02:00
#if defined(__GNUG__)
return __sync_fetch_and_and(dest, value);
#else
2014-09-23 16:27:18 +02:00
return _InterlockedAnd16((volatile short*)dest, value);
2014-10-10 20:19:14 +02:00
#endif
2014-09-23 16:27:18 +02:00
}
static __forceinline uint32_t InterlockedAnd(volatile uint32_t* dest, uint32_t value)
{
2014-10-10 20:19:14 +02:00
#if defined(__GNUG__)
return __sync_fetch_and_and(dest, value);
#else
return _InterlockedAnd((volatile long*)dest, value);
2014-10-10 20:19:14 +02:00
#endif
}
static __forceinline uint64_t InterlockedAnd(volatile uint64_t* dest, uint64_t value)
{
2014-10-10 20:19:14 +02:00
#if defined(__GNUG__)
return __sync_fetch_and_and(dest, value);
#else
return _InterlockedAnd64((volatile long long*)dest, value);
2014-10-10 20:19:14 +02:00
#endif
}
#endif
#ifndef InterlockedXor
static __forceinline uint8_t InterlockedXor(volatile uint8_t* dest, uint8_t value)
{
2014-10-10 20:19:14 +02:00
#if defined(__GNUG__)
return __sync_fetch_and_xor(dest, value);
#else
return _InterlockedXor8((volatile char*)dest, value);
2014-10-10 20:19:14 +02:00
#endif
}
2014-09-23 16:27:18 +02:00
static __forceinline uint16_t InterlockedXor(volatile uint16_t* dest, uint16_t value)
{
2014-10-10 20:19:14 +02:00
#if defined(__GNUG__)
return __sync_fetch_and_xor(dest, value);
#else
2014-09-23 16:27:18 +02:00
return _InterlockedXor16((volatile short*)dest, value);
2014-10-10 20:19:14 +02:00
#endif
2014-09-23 16:27:18 +02:00
}
static __forceinline uint32_t InterlockedXor(volatile uint32_t* dest, uint32_t value)
{
2014-10-10 20:19:14 +02:00
#if defined(__GNUG__)
return __sync_fetch_and_xor(dest, value);
#else
return _InterlockedXor((volatile long*)dest, value);
2014-10-10 20:19:14 +02:00
#endif
}
static __forceinline uint64_t InterlockedXor(volatile uint64_t* dest, uint64_t value)
{
2014-10-10 20:19:14 +02:00
#if defined(__GNUG__)
return __sync_fetch_and_xor(dest, value);
#else
return _InterlockedXor64((volatile long long*)dest, value);
2014-10-10 20:19:14 +02:00
#endif
}
#endif
2014-09-27 20:49:33 +02:00
static __forceinline uint32_t cntlz32(uint32_t arg)
{
2015-01-12 22:32:53 +01:00
#if defined(_MSC_VER)
2014-09-27 20:49:33 +02:00
unsigned long res;
if (!_BitScanReverse(&res, arg))
{
return 32;
}
else
{
return res ^ 31;
}
2015-01-12 22:32:53 +01:00
#else
if (arg)
{
return __builtin_clzll((uint64_t)arg) - 32;
}
else
{
return 32;
}
2014-09-27 20:49:33 +02:00
#endif
}
static __forceinline uint64_t cntlz64(uint64_t arg)
{
2015-01-12 22:32:53 +01:00
#if defined(_MSC_VER)
2014-09-27 20:49:33 +02:00
unsigned long res;
if (!_BitScanReverse64(&res, arg))
{
return 64;
}
else
{
return res ^ 63;
}
2015-01-12 22:32:53 +01:00
#else
if (arg)
{
return __builtin_clzll(arg);
}
else
{
return 64;
}
2014-09-27 20:49:33 +02:00
#endif
}
2014-10-07 15:35:44 +02:00
2014-10-08 16:26:08 +02:00
// compare 16 packed unsigned bytes (greater than)
2015-03-26 19:42:12 +01:00
inline __m128i sse_cmpgt_epu8(__m128i A, __m128i B)
2014-10-07 15:35:44 +02:00
{
// (A xor 0x80) > (B xor 0x80)
2015-03-26 19:42:12 +01:00
const auto sign = _mm_set1_epi32(0x80808080);
return _mm_cmpgt_epi8(_mm_xor_si128(A, sign), _mm_xor_si128(B, sign));
2014-10-07 15:35:44 +02:00
}
2015-03-26 19:42:12 +01:00
inline __m128i sse_cmpgt_epu16(__m128i A, __m128i B)
2014-10-07 15:35:44 +02:00
{
2015-03-26 19:42:12 +01:00
const auto sign = _mm_set1_epi32(0x80008000);
return _mm_cmpgt_epi16(_mm_xor_si128(A, sign), _mm_xor_si128(B, sign));
}
inline __m128i sse_cmpgt_epu32(__m128i A, __m128i B)
{
const auto sign = _mm_set1_epi32(0x80000000);
return _mm_cmpgt_epi32(_mm_xor_si128(A, sign), _mm_xor_si128(B, sign));
2014-10-07 15:35:44 +02:00
}
2015-03-29 13:00:10 +02:00
inline __m128 sse_exp2_ps(__m128 A)
{
const auto x0 = _mm_max_ps(_mm_min_ps(A, _mm_set1_ps(127.4999961f)), _mm_set1_ps(-127.4999961f));
const auto x1 = _mm_add_ps(x0, _mm_set1_ps(0.5f));
const auto x2 = _mm_sub_epi32(_mm_cvtps_epi32(x1), _mm_and_si128(_mm_castps_si128(_mm_cmpnlt_ps(_mm_setzero_ps(), x1)), _mm_set1_epi32(1)));
const auto x3 = _mm_sub_ps(x0, _mm_cvtepi32_ps(x2));
const auto x4 = _mm_mul_ps(x3, x3);
const auto x5 = _mm_mul_ps(x3, _mm_add_ps(_mm_mul_ps(_mm_add_ps(_mm_mul_ps(x4, _mm_set1_ps(0.023093347705f)), _mm_set1_ps(20.20206567f)), x4), _mm_set1_ps(1513.906801f)));
const auto x6 = _mm_mul_ps(x5, _mm_rcp_ps(_mm_sub_ps(_mm_add_ps(_mm_mul_ps(_mm_set1_ps(233.1842117f), x4), _mm_set1_ps(4368.211667f)), x5)));
return _mm_mul_ps(_mm_add_ps(_mm_add_ps(x6, x6), _mm_set1_ps(1.0f)), _mm_castsi128_ps(_mm_slli_epi32(_mm_add_epi32(x2, _mm_set1_epi32(127)), 23)));
}
inline __m128 sse_log2_ps(__m128 A)
{
const auto _1 = _mm_set1_ps(1.0f);
const auto _c = _mm_set1_ps(1.442695040f);
const auto x0 = _mm_max_ps(A, _mm_castsi128_ps(_mm_set1_epi32(0x00800000)));
const auto x1 = _mm_or_ps(_mm_and_ps(x0, _mm_castsi128_ps(_mm_set1_epi32(0x807fffff))), _1);
const auto x2 = _mm_rcp_ps(_mm_add_ps(x1, _1));
const auto x3 = _mm_mul_ps(_mm_sub_ps(x1, _1), x2);
const auto x4 = _mm_add_ps(x3, x3);
const auto x5 = _mm_mul_ps(x4, x4);
const auto x6 = _mm_add_ps(_mm_mul_ps(_mm_add_ps(_mm_mul_ps(_mm_set1_ps(-0.7895802789f), x5), _mm_set1_ps(16.38666457f)), x5), _mm_set1_ps(-64.1409953f));
const auto x7 = _mm_rcp_ps(_mm_add_ps(_mm_mul_ps(_mm_add_ps(_mm_mul_ps(_mm_set1_ps(-35.67227983f), x5), _mm_set1_ps(312.0937664f)), x5), _mm_set1_ps(-769.6919436f)));
const auto x8 = _mm_cvtepi32_ps(_mm_sub_epi32(_mm_srli_epi32(_mm_castps_si128(x0), 23), _mm_set1_epi32(127)));
return _mm_add_ps(_mm_mul_ps(_mm_mul_ps(_mm_mul_ps(_mm_mul_ps(x5, x6), x7), x4), _c), _mm_add_ps(_mm_mul_ps(x4, _c), x8));
}