// Copyright 2009-2021 Intel Corporation // SPDX-License-Identifier: Apache-2.0 #pragma once #include "platform.h" #if defined(__WIN32__) #include #endif #if defined(__ARM_NEON) #include "../simd/arm/emulation.h" #else #include #endif #if defined(__BMI__) && defined(__GNUC__) && !defined(__INTEL_COMPILER) #if !defined(_tzcnt_u32) #define _tzcnt_u32 __tzcnt_u32 #endif #if !defined(_tzcnt_u64) #define _tzcnt_u64 __tzcnt_u64 #endif #endif #if defined(__LZCNT__) #if !defined(_lzcnt_u32) #define _lzcnt_u32 __lzcnt32 #endif #if !defined(_lzcnt_u64) #define _lzcnt_u64 __lzcnt64 #endif #endif #if defined(__WIN32__) // -- GODOT start -- #if !defined(NOMINMAX) // -- GODOT end -- #define NOMINMAX // -- GODOT start -- #endif #include "windows.h" // -- GODOT end -- #endif /* normally defined in pmmintrin.h, but we always need this */ #if !defined(_MM_SET_DENORMALS_ZERO_MODE) #define _MM_DENORMALS_ZERO_ON (0x0040) #define _MM_DENORMALS_ZERO_OFF (0x0000) #define _MM_DENORMALS_ZERO_MASK (0x0040) #define _MM_SET_DENORMALS_ZERO_MODE(x) (_mm_setcsr((_mm_getcsr() & ~_MM_DENORMALS_ZERO_MASK) | (x))) #endif namespace embree { //////////////////////////////////////////////////////////////////////////////// /// Windows Platform //////////////////////////////////////////////////////////////////////////////// #if defined(__WIN32__) __forceinline size_t read_tsc() { LARGE_INTEGER li; QueryPerformanceCounter(&li); return (size_t)li.QuadPart; } __forceinline int bsf(int v) { #if defined(__AVX2__) return _tzcnt_u32(v); #else unsigned long r = 0; _BitScanForward(&r,v); return r; #endif } __forceinline unsigned bsf(unsigned v) { #if defined(__AVX2__) return _tzcnt_u32(v); #else unsigned long r = 0; _BitScanForward(&r,v); return r; #endif } #if defined(__X86_64__) __forceinline size_t bsf(size_t v) { #if defined(__AVX2__) return _tzcnt_u64(v); #else unsigned long r = 0; _BitScanForward64(&r,v); return r; #endif } #endif __forceinline int bscf(int& v) { int i = bsf(v); v &= v-1; return i; } __forceinline unsigned bscf(unsigned& v) { unsigned i = bsf(v); v &= v-1; return i; } #if defined(__X86_64__) __forceinline size_t bscf(size_t& v) { size_t i = bsf(v); v &= v-1; return i; } #endif __forceinline int bsr(int v) { #if defined(__AVX2__) return 31 - _lzcnt_u32(v); #else unsigned long r = 0; _BitScanReverse(&r,v); return r; #endif } __forceinline unsigned bsr(unsigned v) { #if defined(__AVX2__) return 31 - _lzcnt_u32(v); #else unsigned long r = 0; _BitScanReverse(&r,v); return r; #endif } #if defined(__X86_64__) __forceinline size_t bsr(size_t v) { #if defined(__AVX2__) return 63 -_lzcnt_u64(v); #else unsigned long r = 0; _BitScanReverse64(&r, v); return r; #endif } #endif __forceinline int lzcnt(const int x) { #if defined(__AVX2__) return _lzcnt_u32(x); #else if (unlikely(x == 0)) return 32; return 31 - bsr(x); #endif } __forceinline int btc(int v, int i) { long r = v; _bittestandcomplement(&r,i); return r; } __forceinline int bts(int v, int i) { long r = v; _bittestandset(&r,i); return r; } __forceinline int btr(int v, int i) { long r = v; _bittestandreset(&r,i); return r; } #if defined(__X86_64__) __forceinline size_t btc(size_t v, size_t i) { size_t r = v; _bittestandcomplement64((__int64*)&r,i); return r; } __forceinline size_t bts(size_t v, size_t i) { __int64 r = v; _bittestandset64(&r,i); return r; } __forceinline size_t btr(size_t v, size_t i) { __int64 r = v; _bittestandreset64(&r,i); return r; } #endif __forceinline int32_t atomic_cmpxchg(volatile int32_t* p, const int32_t c, const int32_t v) { return _InterlockedCompareExchange((volatile long*)p,v,c); } //////////////////////////////////////////////////////////////////////////////// /// Unix Platform //////////////////////////////////////////////////////////////////////////////// #else #if defined(__i386__) && defined(__PIC__) __forceinline void __cpuid(int out[4], int op) { asm volatile ("xchg{l}\t{%%}ebx, %1\n\t" "cpuid\n\t" "xchg{l}\t{%%}ebx, %1\n\t" : "=a"(out[0]), "=r"(out[1]), "=c"(out[2]), "=d"(out[3]) : "0"(op)); } __forceinline void __cpuid_count(int out[4], int op1, int op2) { asm volatile ("xchg{l}\t{%%}ebx, %1\n\t" "cpuid\n\t" "xchg{l}\t{%%}ebx, %1\n\t" : "=a" (out[0]), "=r" (out[1]), "=c" (out[2]), "=d" (out[3]) : "0" (op1), "2" (op2)); } #elif defined(__X86_ASM__) __forceinline void __cpuid(int out[4], int op) { asm volatile ("cpuid" : "=a"(out[0]), "=b"(out[1]), "=c"(out[2]), "=d"(out[3]) : "a"(op)); } __forceinline void __cpuid_count(int out[4], int op1, int op2) { asm volatile ("cpuid" : "=a"(out[0]), "=b"(out[1]), "=c"(out[2]), "=d"(out[3]) : "a"(op1), "c"(op2)); } #endif __forceinline uint64_t read_tsc() { #if defined(__X86_ASM__) uint32_t high,low; asm volatile ("rdtsc" : "=d"(high), "=a"(low)); return (((uint64_t)high) << 32) + (uint64_t)low; #else /* Not supported yet, meaning measuring traversal cost per pixel does not work. */ return 0; #endif } __forceinline int bsf(int v) { #if defined(__AVX2__) return _tzcnt_u32(v); #elif defined(__X86_ASM__) int r = 0; asm ("bsf %1,%0" : "=r"(r) : "r"(v)); return r; #else return __builtin_ctz(v); #endif } #if defined(__64BIT__) __forceinline unsigned bsf(unsigned v) { #if defined(__AVX2__) return _tzcnt_u32(v); #elif defined(__X86_ASM__) unsigned r = 0; asm ("bsf %1,%0" : "=r"(r) : "r"(v)); return r; #else return __builtin_ctz(v); #endif } #endif __forceinline size_t bsf(size_t v) { #if defined(__AVX2__) #if defined(__X86_64__) return _tzcnt_u64(v); #else return _tzcnt_u32(v); #endif #elif defined(__X86_ASM__) size_t r = 0; asm ("bsf %1,%0" : "=r"(r) : "r"(v)); return r; #else return __builtin_ctzl(v); #endif } __forceinline int bscf(int& v) { int i = bsf(v); v &= v-1; return i; } #if defined(__64BIT__) __forceinline unsigned int bscf(unsigned int& v) { unsigned int i = bsf(v); v &= v-1; return i; } #endif __forceinline size_t bscf(size_t& v) { size_t i = bsf(v); v &= v-1; return i; } __forceinline int bsr(int v) { #if defined(__AVX2__) return 31 - _lzcnt_u32(v); #elif defined(__X86_ASM__) int r = 0; asm ("bsr %1,%0" : "=r"(r) : "r"(v)); return r; #else return __builtin_clz(v) ^ 31; #endif } #if defined(__64BIT__) __forceinline unsigned bsr(unsigned v) { #if defined(__AVX2__) return 31 - _lzcnt_u32(v); #elif defined(__X86_ASM__) unsigned r = 0; asm ("bsr %1,%0" : "=r"(r) : "r"(v)); return r; #else return __builtin_clz(v) ^ 31; #endif } #endif __forceinline size_t bsr(size_t v) { #if defined(__AVX2__) #if defined(__X86_64__) return 63 - _lzcnt_u64(v); #else return 31 - _lzcnt_u32(v); #endif #elif defined(__X86_ASM__) size_t r = 0; asm ("bsr %1,%0" : "=r"(r) : "r"(v)); return r; #else return (sizeof(v) * 8 - 1) - __builtin_clzl(v); #endif } __forceinline int lzcnt(const int x) { #if defined(__AVX2__) return _lzcnt_u32(x); #else if (unlikely(x == 0)) return 32; return 31 - bsr(x); #endif } __forceinline size_t blsr(size_t v) { #if defined(__AVX2__) #if defined(__INTEL_COMPILER) return _blsr_u64(v); #else #if defined(__X86_64__) return __blsr_u64(v); #else return __blsr_u32(v); #endif #endif #else return v & (v-1); #endif } __forceinline int btc(int v, int i) { #if defined(__X86_ASM__) int r = 0; asm ("btc %1,%0" : "=r"(r) : "r"(i), "0"(v) : "flags" ); return r; #else return (v ^ (1 << i)); #endif } __forceinline int bts(int v, int i) { #if defined(__X86_ASM__) int r = 0; asm ("bts %1,%0" : "=r"(r) : "r"(i), "0"(v) : "flags"); return r; #else return (v | (v << i)); #endif } __forceinline int btr(int v, int i) { #if defined(__X86_ASM__) int r = 0; asm ("btr %1,%0" : "=r"(r) : "r"(i), "0"(v) : "flags"); return r; #else return (v & ~(v << i)); #endif } __forceinline size_t btc(size_t v, size_t i) { #if defined(__X86_ASM__) size_t r = 0; asm ("btc %1,%0" : "=r"(r) : "r"(i), "0"(v) : "flags" ); return r; #else return (v ^ (1 << i)); #endif } __forceinline size_t bts(size_t v, size_t i) { #if defined(__X86_ASM__) size_t r = 0; asm ("bts %1,%0" : "=r"(r) : "r"(i), "0"(v) : "flags"); return r; #else return (v | (v << i)); #endif } __forceinline size_t btr(size_t v, size_t i) { #if defined(__X86_ASM__) size_t r = 0; asm ("btr %1,%0" : "=r"(r) : "r"(i), "0"(v) : "flags"); return r; #else return (v & ~(v << i)); #endif } __forceinline int32_t atomic_cmpxchg(int32_t volatile* value, int32_t comparand, const int32_t input) { return __sync_val_compare_and_swap(value, comparand, input); } #endif //////////////////////////////////////////////////////////////////////////////// /// All Platforms //////////////////////////////////////////////////////////////////////////////// #if defined(__clang__) || defined(__GNUC__) #if !defined(_mm_undefined_ps) __forceinline __m128 _mm_undefined_ps() { return _mm_setzero_ps(); } #endif #if !defined(_mm_undefined_si128) __forceinline __m128i _mm_undefined_si128() { return _mm_setzero_si128(); } #endif #if !defined(_mm256_undefined_ps) && defined(__AVX__) __forceinline __m256 _mm256_undefined_ps() { return _mm256_setzero_ps(); } #endif #if !defined(_mm256_undefined_si256) && defined(__AVX__) __forceinline __m256i _mm256_undefined_si256() { return _mm256_setzero_si256(); } #endif #if !defined(_mm512_undefined_ps) && defined(__AVX512F__) __forceinline __m512 _mm512_undefined_ps() { return _mm512_setzero_ps(); } #endif #if !defined(_mm512_undefined_epi32) && defined(__AVX512F__) __forceinline __m512i _mm512_undefined_epi32() { return _mm512_setzero_si512(); } #endif #endif #if defined(__SSE4_2__) __forceinline int popcnt(int in) { return _mm_popcnt_u32(in); } __forceinline unsigned popcnt(unsigned in) { return _mm_popcnt_u32(in); } #if defined(__64BIT__) __forceinline size_t popcnt(size_t in) { return _mm_popcnt_u64(in); } #endif #endif #if defined(__X86_ASM__) __forceinline uint64_t rdtsc() { int dummy[4]; __cpuid(dummy,0); uint64_t clock = read_tsc(); __cpuid(dummy,0); return clock; } #endif __forceinline void pause_cpu(const size_t N = 8) { for (size_t i=0; i