| 123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659 |
- // ======================================================================== //
- // Copyright 2009-2017 Intel Corporation //
- // //
- // Licensed under the Apache License, Version 2.0 (the "License"); //
- // you may not use this file except in compliance with the License. //
- // You may obtain a copy of the License at //
- // //
- // http://www.apache.org/licenses/LICENSE-2.0 //
- // //
- // Unless required by applicable law or agreed to in writing, software //
- // distributed under the License is distributed on an "AS IS" BASIS, //
- // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. //
- // See the License for the specific language governing permissions and //
- // limitations under the License. //
- // ======================================================================== //
- #pragma once
- namespace embree
- {
- /* 8-wide AVX float type */
- template<>
- struct vfloat<8>
- {
- typedef vboolf8 Bool;
- typedef vint8 Int;
- typedef vfloat8 Float;
- enum { size = 8 }; // number of SIMD elements
- union { __m256 v; float f[8]; int i[8]; }; // data
- ////////////////////////////////////////////////////////////////////////////////
- /// Constructors, Assignment & Cast Operators
- ////////////////////////////////////////////////////////////////////////////////
- __forceinline vfloat ( ) {}
- __forceinline vfloat ( const vfloat8& other ) { v = other.v; }
- __forceinline vfloat8& operator=( const vfloat8& other ) { v = other.v; return *this; }
- __forceinline vfloat( const __m256 a ) : v(a) {}
- __forceinline operator const __m256&( void ) const { return v; }
- __forceinline operator __m256&( void ) { return v; }
- __forceinline explicit vfloat( const vfloat4& a ) : v(_mm256_insertf128_ps(_mm256_castps128_ps256(a),a,1)) {}
- __forceinline vfloat( const vfloat4& a, const vfloat4& b ) : v(_mm256_insertf128_ps(_mm256_castps128_ps256(a),b,1)) {}
- __forceinline explicit vfloat( const char* const a ) : v(_mm256_loadu_ps((const float*)a)) {}
- __forceinline vfloat( const float& a ) : v(_mm256_broadcast_ss(&a)) {}
- __forceinline vfloat( float a, float b) : v(_mm256_set_ps(b, a, b, a, b, a, b, a)) {}
- __forceinline vfloat( float a, float b, float c, float d ) : v(_mm256_set_ps(d, c, b, a, d, c, b, a)) {}
- __forceinline vfloat( float a, float b, float c, float d, float e, float f, float g, float h ) : v(_mm256_set_ps(h, g, f, e, d, c, b, a)) {}
- __forceinline explicit vfloat( const __m256i a ) : v(_mm256_cvtepi32_ps(a)) {}
- ////////////////////////////////////////////////////////////////////////////////
- /// Constants
- ////////////////////////////////////////////////////////////////////////////////
- __forceinline vfloat( ZeroTy ) : v(_mm256_setzero_ps()) {}
- __forceinline vfloat( OneTy ) : v(_mm256_set1_ps(1.0f)) {}
- __forceinline vfloat( PosInfTy ) : v(_mm256_set1_ps(pos_inf)) {}
- __forceinline vfloat( NegInfTy ) : v(_mm256_set1_ps(neg_inf)) {}
- __forceinline vfloat( StepTy ) : v(_mm256_set_ps(7.0f, 6.0f, 5.0f, 4.0f, 3.0f, 2.0f, 1.0f, 0.0f)) {}
- __forceinline vfloat( NaNTy ) : v(_mm256_set1_ps(nan)) {}
- ////////////////////////////////////////////////////////////////////////////////
- /// Loads and Stores
- ////////////////////////////////////////////////////////////////////////////////
- static __forceinline vfloat8 broadcast( const void* const a ) {
- return _mm256_broadcast_ss((float*)a);
- }
- static __forceinline vfloat8 broadcast2( const float* const a, const float* const b ) {
- #if defined(__INTEL_COMPILER)
- const vfloat8 v0 = _mm256_broadcast_ss(a);
- const vfloat8 v1 = _mm256_broadcast_ss(b);
- return _mm256_blend_ps(v1, v0, 0xf);
- #else
- return _mm256_set_ps(*b,*b,*b,*b,*a,*a,*a,*a);
- #endif
- }
- static __forceinline const vfloat8 broadcast4f(const vfloat4* ptr) {
- return _mm256_broadcast_ps((__m128*)ptr);
- }
- static __forceinline vfloat8 load( const unsigned char* const ptr ) {
- #if defined(__AVX2__)
- return _mm256_cvtepi32_ps(_mm256_cvtepu8_epi32(_mm_loadu_si128((__m128i*)ptr)));
- #else
- return vfloat8(vfloat4::load(ptr),vfloat4::load(ptr+4));
- #endif
- }
-
- static __forceinline vfloat8 load ( const void* const ptr) { return _mm256_load_ps((float*)ptr); }
- static __forceinline vfloat8 loadu( const void* const ptr) { return _mm256_loadu_ps((float*)ptr); }
- static __forceinline void store ( void* ptr, const vfloat8& v) { return _mm256_store_ps((float*)ptr,v); }
- static __forceinline void storeu( void* ptr, const vfloat8& v) { return _mm256_storeu_ps((float*)ptr,v); }
- #if defined(__AVX512VL__)
- static __forceinline vfloat8 load ( const vboolf8& mask, const void* const ptr) { return _mm256_mask_load_ps (_mm256_setzero_ps(),mask,(float*)ptr); }
- static __forceinline vfloat8 loadu( const vboolf8& mask, const void* const ptr) { return _mm256_mask_loadu_ps(_mm256_setzero_ps(),mask,(float*)ptr); }
- static __forceinline void store ( const vboolf8& mask, void* ptr, const vfloat8& v) { _mm256_mask_store_ps ((float*)ptr,mask,v); }
- static __forceinline void storeu( const vboolf8& mask, void* ptr, const vfloat8& v) { _mm256_mask_storeu_ps((float*)ptr,mask,v); }
- #else
- static __forceinline vfloat8 load ( const vboolf8& mask, const void* const ptr) { return _mm256_maskload_ps((float*)ptr,(__m256i)mask); }
- static __forceinline vfloat8 loadu( const vboolf8& mask, const void* const ptr) { return _mm256_maskload_ps((float*)ptr,(__m256i)mask); }
- static __forceinline void store ( const vboolf8& mask, void* ptr, const vfloat8& v) { _mm256_maskstore_ps((float*)ptr,(__m256i)mask,v); }
- static __forceinline void storeu( const vboolf8& mask, void* ptr, const vfloat8& v) { _mm256_maskstore_ps((float*)ptr,(__m256i)mask,v); }
- #endif
-
- #if defined(__AVX2__)
- static __forceinline vfloat8 load_nt(void* ptr) {
- return _mm256_castsi256_ps(_mm256_stream_load_si256((__m256i*)ptr));
- }
- #endif
-
- static __forceinline void store_nt(void* ptr, const vfloat8& v) {
- _mm256_stream_ps((float*)ptr,v);
- }
- template<int scale>
- static __forceinline void scatter(const vboolf8& mask, void* ptr, const vint8& ofs, const vfloat8& v)
- {
- #if defined(__AVX512VL__)
- _mm256_mask_i32scatter_ps(ptr,mask,ofs,v,scale);
- #else
- if (likely(mask[0])) *(float*)(((char*)ptr)+scale*ofs[0]) = v[0];
- if (likely(mask[1])) *(float*)(((char*)ptr)+scale*ofs[1]) = v[1];
- if (likely(mask[2])) *(float*)(((char*)ptr)+scale*ofs[2]) = v[2];
- if (likely(mask[3])) *(float*)(((char*)ptr)+scale*ofs[3]) = v[3];
- if (likely(mask[4])) *(float*)(((char*)ptr)+scale*ofs[4]) = v[4];
- if (likely(mask[5])) *(float*)(((char*)ptr)+scale*ofs[5]) = v[5];
- if (likely(mask[6])) *(float*)(((char*)ptr)+scale*ofs[6]) = v[6];
- if (likely(mask[7])) *(float*)(((char*)ptr)+scale*ofs[7]) = v[7];
- #endif
- }
- static __forceinline void store(const vboolf8& mask, char* ptr, const vint8& ofs, const vfloat8& v) {
- scatter<1>(mask,ptr,ofs,v);
- }
- static __forceinline void store(const vboolf8& mask, float* ptr, const vint8& ofs, const vfloat8& v) {
- scatter<4>(mask,ptr,ofs,v);
- }
- ////////////////////////////////////////////////////////////////////////////////
- /// Array Access
- ////////////////////////////////////////////////////////////////////////////////
- __forceinline const float& operator []( const size_t index ) const { assert(index < 8); return f[index]; }
- __forceinline float& operator []( const size_t index ) { assert(index < 8); return f[index]; }
- };
- ////////////////////////////////////////////////////////////////////////////////
- /// Unary Operators
- ////////////////////////////////////////////////////////////////////////////////
- __forceinline const vfloat8 asFloat ( const vint8& a ) { return _mm256_castsi256_ps(a); }
- __forceinline const vint8 asInt ( const vfloat8& a ) { return _mm256_castps_si256(a); }
- __forceinline const vfloat8 operator +( const vfloat8& a ) { return a; }
- __forceinline const vfloat8 operator -( const vfloat8& a ) {
- const __m256 mask = _mm256_castsi256_ps(_mm256_set1_epi32(0x80000000));
- return _mm256_xor_ps(a.v, mask);
- }
- __forceinline const vfloat8 abs ( const vfloat8& a ) {
- const __m256 mask = _mm256_castsi256_ps(_mm256_set1_epi32(0x7fffffff));
- return _mm256_and_ps(a.v, mask);
- }
- __forceinline const vfloat8 sign ( const vfloat8& a ) { return _mm256_blendv_ps(vfloat8(one), -vfloat8(one), _mm256_cmp_ps(a, vfloat8(zero), _CMP_NGE_UQ )); }
- __forceinline const vfloat8 signmsk ( const vfloat8& a ) { return _mm256_and_ps(a.v,_mm256_castsi256_ps(_mm256_set1_epi32(0x80000000))); }
- __forceinline const vfloat8 rcp(const vfloat8& a)
- {
- #if defined(__AVX512VL__)
- const vfloat8 r = _mm256_rcp14_ps(a.v);
- #else
- const vfloat8 r = _mm256_rcp_ps(a.v);
- #endif
- #if defined(__AVX2__)
- return _mm256_mul_ps(r,_mm256_fnmadd_ps(r, a, vfloat8(2.0f)));
- #else
- return _mm256_mul_ps(r,_mm256_sub_ps(vfloat8(2.0f), _mm256_mul_ps(r, a)));
- #endif
- }
- __forceinline const vfloat8 sqr ( const vfloat8& a ) { return _mm256_mul_ps(a,a); }
- __forceinline const vfloat8 sqrt ( const vfloat8& a ) { return _mm256_sqrt_ps(a.v); }
- __forceinline const vfloat8 rsqrt( const vfloat8& a )
- {
- #if defined(__AVX512VL__)
- const vfloat8 r = _mm256_rsqrt14_ps(a.v);
- #else
- const vfloat8 r = _mm256_rsqrt_ps(a.v);
- #endif
- #if defined(__AVX2__)
- return _mm256_fmadd_ps(_mm256_set1_ps(1.5f), r,
- _mm256_mul_ps(_mm256_mul_ps(_mm256_mul_ps(a, _mm256_set1_ps(-0.5f)), r), _mm256_mul_ps(r, r)));
- #else
- return _mm256_add_ps(_mm256_mul_ps(_mm256_set1_ps(1.5f), r),
- _mm256_mul_ps(_mm256_mul_ps(_mm256_mul_ps(a, _mm256_set1_ps(-0.5f)), r), _mm256_mul_ps(r, r)));
- #endif
- }
- ////////////////////////////////////////////////////////////////////////////////
- /// Binary Operators
- ////////////////////////////////////////////////////////////////////////////////
- __forceinline const vfloat8 operator +( const vfloat8& a, const vfloat8& b ) { return _mm256_add_ps(a.v, b.v); }
- __forceinline const vfloat8 operator +( const vfloat8& a, const float b ) { return a + vfloat8(b); }
- __forceinline const vfloat8 operator +( const float a, const vfloat8& b ) { return vfloat8(a) + b; }
- __forceinline const vfloat8 operator -( const vfloat8& a, const vfloat8& b ) { return _mm256_sub_ps(a.v, b.v); }
- __forceinline const vfloat8 operator -( const vfloat8& a, const float b ) { return a - vfloat8(b); }
- __forceinline const vfloat8 operator -( const float a, const vfloat8& b ) { return vfloat8(a) - b; }
- __forceinline const vfloat8 operator *( const vfloat8& a, const vfloat8& b ) { return _mm256_mul_ps(a.v, b.v); }
- __forceinline const vfloat8 operator *( const vfloat8& a, const float b ) { return a * vfloat8(b); }
- __forceinline const vfloat8 operator *( const float a, const vfloat8& b ) { return vfloat8(a) * b; }
- __forceinline const vfloat8 operator /( const vfloat8& a, const vfloat8& b ) { return _mm256_div_ps(a.v, b.v); }
- __forceinline const vfloat8 operator /( const vfloat8& a, const float b ) { return a / vfloat8(b); }
- __forceinline const vfloat8 operator /( const float a, const vfloat8& b ) { return vfloat8(a) / b; }
- __forceinline const vfloat8 operator^( const vfloat8& a, const vfloat8& b ) { return _mm256_xor_ps(a.v,b.v); }
- __forceinline const vfloat8 operator^( const vfloat8& a, const vint8& b ) { return _mm256_xor_ps(a.v,_mm256_castsi256_ps(b.v)); }
- __forceinline const vfloat8 operator&( const vfloat8& a, const vfloat8& b ) { return _mm256_and_ps(a.v,b.v); }
- __forceinline const vfloat8 min( const vfloat8& a, const vfloat8& b ) { return _mm256_min_ps(a.v, b.v); }
- __forceinline const vfloat8 min( const vfloat8& a, const float b ) { return _mm256_min_ps(a.v, vfloat8(b)); }
- __forceinline const vfloat8 min( const float a, const vfloat8& b ) { return _mm256_min_ps(vfloat8(a), b.v); }
- __forceinline const vfloat8 max( const vfloat8& a, const vfloat8& b ) { return _mm256_max_ps(a.v, b.v); }
- __forceinline const vfloat8 max( const vfloat8& a, const float b ) { return _mm256_max_ps(a.v, vfloat8(b)); }
- __forceinline const vfloat8 max( const float a, const vfloat8& b ) { return _mm256_max_ps(vfloat8(a), b.v); }
- #if defined (__AVX2__)
- __forceinline vfloat8 mini(const vfloat8& a, const vfloat8& b) {
- const vint8 ai = _mm256_castps_si256(a);
- const vint8 bi = _mm256_castps_si256(b);
- const vint8 ci = _mm256_min_epi32(ai,bi);
- return _mm256_castsi256_ps(ci);
- }
- __forceinline vfloat8 maxi(const vfloat8& a, const vfloat8& b) {
- const vint8 ai = _mm256_castps_si256(a);
- const vint8 bi = _mm256_castps_si256(b);
- const vint8 ci = _mm256_max_epi32(ai,bi);
- return _mm256_castsi256_ps(ci);
- }
- __forceinline vfloat8 minui(const vfloat8& a, const vfloat8& b) {
- const vint8 ai = _mm256_castps_si256(a);
- const vint8 bi = _mm256_castps_si256(b);
- const vint8 ci = _mm256_min_epu32(ai,bi);
- return _mm256_castsi256_ps(ci);
- }
- __forceinline vfloat8 maxui(const vfloat8& a, const vfloat8& b) {
- const vint8 ai = _mm256_castps_si256(a);
- const vint8 bi = _mm256_castps_si256(b);
- const vint8 ci = _mm256_max_epu32(ai,bi);
- return _mm256_castsi256_ps(ci);
- }
- #else
- __forceinline vfloat8 mini(const vfloat8& a, const vfloat8& b) {
- return min(a,b);
- }
- __forceinline vfloat8 maxi(const vfloat8& a, const vfloat8& b) {
- return max(a,b);
- }
- #endif
- ////////////////////////////////////////////////////////////////////////////////
- /// Ternary Operators
- ////////////////////////////////////////////////////////////////////////////////
- #if defined(__AVX2__)
- __forceinline const vfloat8 madd ( const vfloat8& a, const vfloat8& b, const vfloat8& c) { return _mm256_fmadd_ps(a,b,c); }
- __forceinline const vfloat8 msub ( const vfloat8& a, const vfloat8& b, const vfloat8& c) { return _mm256_fmsub_ps(a,b,c); }
- __forceinline const vfloat8 nmadd ( const vfloat8& a, const vfloat8& b, const vfloat8& c) { return _mm256_fnmadd_ps(a,b,c); }
- __forceinline const vfloat8 nmsub ( const vfloat8& a, const vfloat8& b, const vfloat8& c) { return _mm256_fnmsub_ps(a,b,c); }
- #else
- __forceinline const vfloat8 madd ( const vfloat8& a, const vfloat8& b, const vfloat8& c) { return a*b+c; }
- __forceinline const vfloat8 msub ( const vfloat8& a, const vfloat8& b, const vfloat8& c) { return a*b-c; }
- __forceinline const vfloat8 nmadd ( const vfloat8& a, const vfloat8& b, const vfloat8& c) { return -a*b+c;}
- __forceinline const vfloat8 nmsub ( const vfloat8& a, const vfloat8& b, const vfloat8& c) { return -a*b-c; }
- #endif
- ////////////////////////////////////////////////////////////////////////////////
- /// Assignment Operators
- ////////////////////////////////////////////////////////////////////////////////
- __forceinline vfloat8& operator +=( vfloat8& a, const vfloat8& b ) { return a = a + b; }
- __forceinline vfloat8& operator +=( vfloat8& a, const float b ) { return a = a + b; }
- __forceinline vfloat8& operator -=( vfloat8& a, const vfloat8& b ) { return a = a - b; }
- __forceinline vfloat8& operator -=( vfloat8& a, const float b ) { return a = a - b; }
- __forceinline vfloat8& operator *=( vfloat8& a, const vfloat8& b ) { return a = a * b; }
- __forceinline vfloat8& operator *=( vfloat8& a, const float b ) { return a = a * b; }
- __forceinline vfloat8& operator /=( vfloat8& a, const vfloat8& b ) { return a = a / b; }
- __forceinline vfloat8& operator /=( vfloat8& a, const float b ) { return a = a / b; }
- ////////////////////////////////////////////////////////////////////////////////
- /// Comparison Operators + Select
- ////////////////////////////////////////////////////////////////////////////////
- #if defined(__AVX512VL__)
- __forceinline const vboolf8 operator ==( const vfloat8& a, const vfloat8& b ) { return _mm256_cmp_ps_mask(a, b, _MM_CMPINT_EQ); }
- __forceinline const vboolf8 operator !=( const vfloat8& a, const vfloat8& b ) { return _mm256_cmp_ps_mask(a, b, _MM_CMPINT_NE); }
- __forceinline const vboolf8 operator < ( const vfloat8& a, const vfloat8& b ) { return _mm256_cmp_ps_mask(a, b, _MM_CMPINT_LT); }
- __forceinline const vboolf8 operator >=( const vfloat8& a, const vfloat8& b ) { return _mm256_cmp_ps_mask(a, b, _MM_CMPINT_GE); }
- __forceinline const vboolf8 operator > ( const vfloat8& a, const vfloat8& b ) { return _mm256_cmp_ps_mask(a, b, _MM_CMPINT_GT); }
- __forceinline const vboolf8 operator <=( const vfloat8& a, const vfloat8& b ) { return _mm256_cmp_ps_mask(a, b, _MM_CMPINT_LE); }
- __forceinline const vfloat8 select( const vboolf8& m, const vfloat8& t, const vfloat8& f ) {
- return _mm256_mask_blend_ps(m, f, t);
- }
- #else
- __forceinline const vboolf8 operator ==( const vfloat8& a, const vfloat8& b ) { return _mm256_cmp_ps(a, b, _CMP_EQ_OQ ); }
- __forceinline const vboolf8 operator !=( const vfloat8& a, const vfloat8& b ) { return _mm256_cmp_ps(a, b, _CMP_NEQ_OQ); }
- __forceinline const vboolf8 operator < ( const vfloat8& a, const vfloat8& b ) { return _mm256_cmp_ps(a, b, _CMP_LT_OQ ); }
- __forceinline const vboolf8 operator >=( const vfloat8& a, const vfloat8& b ) { return _mm256_cmp_ps(a, b, _CMP_GE_OQ ); }
- __forceinline const vboolf8 operator > ( const vfloat8& a, const vfloat8& b ) { return _mm256_cmp_ps(a, b, _CMP_GT_OQ ); }
- __forceinline const vboolf8 operator <=( const vfloat8& a, const vfloat8& b ) { return _mm256_cmp_ps(a, b, _CMP_LE_OQ ); }
- __forceinline const vfloat8 select( const vboolf8& m, const vfloat8& t, const vfloat8& f ) {
- return _mm256_blendv_ps(f, t, m);
- }
- #endif
- template<int mask>
- __forceinline const vfloat8 select( const vfloat8& t, const vfloat8& f ) {
- return _mm256_blend_ps(f, t, mask);
- }
- __forceinline const vboolf8 operator ==( const vfloat8& a, const float& b ) { return a == vfloat8(b); }
- __forceinline const vboolf8 operator ==( const float& a, const vfloat8& b ) { return vfloat8(a) == b; }
- __forceinline const vboolf8 operator !=( const vfloat8& a, const float& b ) { return a != vfloat8(b); }
- __forceinline const vboolf8 operator !=( const float& a, const vfloat8& b ) { return vfloat8(a) != b; }
- __forceinline const vboolf8 operator < ( const vfloat8& a, const float& b ) { return a < vfloat8(b); }
- __forceinline const vboolf8 operator < ( const float& a, const vfloat8& b ) { return vfloat8(a) < b; }
- __forceinline const vboolf8 operator >=( const vfloat8& a, const float& b ) { return a >= vfloat8(b); }
- __forceinline const vboolf8 operator >=( const float& a, const vfloat8& b ) { return vfloat8(a) >= b; }
- __forceinline const vboolf8 operator > ( const vfloat8& a, const float& b ) { return a > vfloat8(b); }
- __forceinline const vboolf8 operator > ( const float& a, const vfloat8& b ) { return vfloat8(a) > b; }
- __forceinline const vboolf8 operator <=( const vfloat8& a, const float& b ) { return a <= vfloat8(b); }
- __forceinline const vboolf8 operator <=( const float& a, const vfloat8& b ) { return vfloat8(a) <= b; }
- __forceinline vboolf8 eq(const vfloat8& a, const vfloat8& b) { return a == b; }
- __forceinline vboolf8 ne(const vfloat8& a, const vfloat8& b) { return a != b; }
- __forceinline vboolf8 lt(const vfloat8& a, const vfloat8& b) { return a < b; }
- __forceinline vboolf8 ge(const vfloat8& a, const vfloat8& b) { return a >= b; }
- __forceinline vboolf8 gt(const vfloat8& a, const vfloat8& b) { return a > b; }
- __forceinline vboolf8 le(const vfloat8& a, const vfloat8& b) { return a <= b; }
- #if defined(__AVX512VL__)
- __forceinline vboolf8 eq(const vboolf8& mask, const vfloat8& a, const vfloat8& b) { return _mm256_mask_cmp_ps_mask(mask, a, b, _MM_CMPINT_EQ); }
- __forceinline vboolf8 ne(const vboolf8& mask, const vfloat8& a, const vfloat8& b) { return _mm256_mask_cmp_ps_mask(mask, a, b, _MM_CMPINT_NE); }
- __forceinline vboolf8 lt(const vboolf8& mask, const vfloat8& a, const vfloat8& b) { return _mm256_mask_cmp_ps_mask(mask, a, b, _MM_CMPINT_LT); }
- __forceinline vboolf8 ge(const vboolf8& mask, const vfloat8& a, const vfloat8& b) { return _mm256_mask_cmp_ps_mask(mask, a, b, _MM_CMPINT_GE); }
- __forceinline vboolf8 gt(const vboolf8& mask, const vfloat8& a, const vfloat8& b) { return _mm256_mask_cmp_ps_mask(mask, a, b, _MM_CMPINT_GT); }
- __forceinline vboolf8 le(const vboolf8& mask, const vfloat8& a, const vfloat8& b) { return _mm256_mask_cmp_ps_mask(mask, a, b, _MM_CMPINT_LE); }
- #else
- __forceinline vboolf8 eq(const vboolf8& mask, const vfloat8& a, const vfloat8& b) { return mask & (a == b); }
- __forceinline vboolf8 ne(const vboolf8& mask, const vfloat8& a, const vfloat8& b) { return mask & (a != b); }
- __forceinline vboolf8 lt(const vboolf8& mask, const vfloat8& a, const vfloat8& b) { return mask & (a < b); }
- __forceinline vboolf8 ge(const vboolf8& mask, const vfloat8& a, const vfloat8& b) { return mask & (a >= b); }
- __forceinline vboolf8 gt(const vboolf8& mask, const vfloat8& a, const vfloat8& b) { return mask & (a > b); }
- __forceinline vboolf8 le(const vboolf8& mask, const vfloat8& a, const vfloat8& b) { return mask & (a <= b); }
- #endif
- __forceinline vfloat8 lerp(const vfloat8& a, const vfloat8& b, const vfloat8& t) {
- return madd(t,b-a,a);
- }
- __forceinline bool isvalid ( const vfloat8& v ) {
- return all((v > vfloat8(-FLT_LARGE)) & (v < vfloat8(+FLT_LARGE)));
- }
- __forceinline bool is_finite ( const vfloat8& a ) {
- return all((a >= vfloat8(-FLT_MAX)) & (a <= vfloat8(+FLT_MAX)));
- }
- __forceinline bool is_finite ( const vboolf8& valid, const vfloat8& a ) {
- return all(valid, (a >= vfloat8(-FLT_MAX)) & (a <= vfloat8(+FLT_MAX)));
- }
-
- ////////////////////////////////////////////////////////////////////////////////
- /// Rounding Functions
- ////////////////////////////////////////////////////////////////////////////////
- __forceinline const vfloat8 floor( const vfloat8& a ) { return _mm256_round_ps(a, _MM_FROUND_TO_NEG_INF ); }
- __forceinline const vfloat8 ceil ( const vfloat8& a ) { return _mm256_round_ps(a, _MM_FROUND_TO_POS_INF ); }
- __forceinline const vfloat8 trunc( const vfloat8& a ) { return _mm256_round_ps(a, _MM_FROUND_TO_ZERO ); }
- __forceinline const vfloat8 frac ( const vfloat8& a ) { return a-floor(a); }
- ////////////////////////////////////////////////////////////////////////////////
- /// Movement/Shifting/Shuffling Functions
- ////////////////////////////////////////////////////////////////////////////////
- __forceinline vfloat8 unpacklo( const vfloat8& a, const vfloat8& b ) { return _mm256_unpacklo_ps(a.v, b.v); }
- __forceinline vfloat8 unpackhi( const vfloat8& a, const vfloat8& b ) { return _mm256_unpackhi_ps(a.v, b.v); }
- template<size_t i> __forceinline const vfloat8 shuffle( const vfloat8& a ) {
- return _mm256_permute_ps(a, _MM_SHUFFLE(i, i, i, i));
- }
- template<size_t i0, size_t i1> __forceinline const vfloat8 shuffle4( const vfloat8& a ) {
- return _mm256_permute2f128_ps(a, a, (i1 << 4) | (i0 << 0));
- }
- template<size_t i0, size_t i1> __forceinline const vfloat8 shuffle4( const vfloat8& a, const vfloat8& b) {
- return _mm256_permute2f128_ps(a, b, (i1 << 4) | (i0 << 0));
- }
- template<size_t i0, size_t i1, size_t i2, size_t i3> __forceinline const vfloat8 shuffle( const vfloat8& a ) {
- return _mm256_permute_ps(a, _MM_SHUFFLE(i3, i2, i1, i0));
- }
- template<size_t i0, size_t i1, size_t i2, size_t i3> __forceinline const vfloat8 shuffle( const vfloat8& a, const vfloat8& b ) {
- return _mm256_shuffle_ps(a, b, _MM_SHUFFLE(i3, i2, i1, i0));
- }
- template<> __forceinline const vfloat8 shuffle<0, 0, 2, 2>( const vfloat8& b ) { return _mm256_moveldup_ps(b); }
- template<> __forceinline const vfloat8 shuffle<1, 1, 3, 3>( const vfloat8& b ) { return _mm256_movehdup_ps(b); }
- template<> __forceinline const vfloat8 shuffle<0, 1, 0, 1>( const vfloat8& b ) { return _mm256_castpd_ps(_mm256_movedup_pd(_mm256_castps_pd(b))); }
- __forceinline const vfloat8 broadcast(const float* ptr) { return _mm256_broadcast_ss(ptr); }
- template<size_t i> __forceinline const vfloat8 insert4(const vfloat8& a, const vfloat4& b) { return _mm256_insertf128_ps(a, b, i); }
- template<size_t i> __forceinline const vfloat4 extract4 (const vfloat8& a) { return _mm256_extractf128_ps(a, i); }
- template<> __forceinline const vfloat4 extract4<0>(const vfloat8& a) { return _mm256_castps256_ps128(a); }
- __forceinline float toScalar(const vfloat8& a) { return _mm_cvtss_f32(_mm256_castps256_ps128(a)); }
- __forceinline vfloat8 assign( const vfloat4& a ) { return _mm256_castps128_ps256(a); }
- #if defined (__AVX2__)
- __forceinline vfloat8 permute(const vfloat8 &a, const __m256i &index) {
- return _mm256_permutevar8x32_ps(a,index);
- }
- #endif
- #if defined(__AVX512VL__)
- template<int i>
- __forceinline vfloat8 align_shift_right(const vfloat8 &a, const vfloat8 &b) {
- return _mm256_castsi256_ps(_mm256_alignr_epi32(_mm256_castps_si256(a), _mm256_castps_si256(b), i));
- }
- #endif
- #if defined (__AVX_I__)
- template<const int mode>
- __forceinline vint4 convert_to_hf16(const vfloat8 &a) {
- return _mm256_cvtps_ph(a,mode);
- }
- __forceinline vfloat8 convert_from_hf16(const vint4 &a) {
- return _mm256_cvtph_ps(a);
- }
- #endif
- __forceinline vfloat4 broadcast4f( const vfloat8& a, const size_t k ) {
- return vfloat4::broadcast(&a[k]);
- }
- __forceinline vfloat8 broadcast8f( const vfloat8& a, const size_t k ) {
- return vfloat8::broadcast(&a[k]);
- }
- #if defined(__AVX512VL__)
- __forceinline vfloat8 shift_right_1( const vfloat8& x) {
- return align_shift_right<1>(zero,x);
- }
- #else
- __forceinline vfloat8 shift_right_1( const vfloat8& x) {
- const vfloat8 t0 = shuffle<1,2,3,0>(x);
- const vfloat8 t1 = shuffle4<1,0>(t0);
- return _mm256_blend_ps(t0,t1,0x88);
- }
- #endif
- __forceinline vint8 floori (const vfloat8& a) {
- return vint8(floor(a));
- }
- ////////////////////////////////////////////////////////////////////////////////
- /// Transpose
- ////////////////////////////////////////////////////////////////////////////////
- __forceinline void transpose(const vfloat8& r0, const vfloat8& r1, const vfloat8& r2, const vfloat8& r3, vfloat8& c0, vfloat8& c1, vfloat8& c2, vfloat8& c3)
- {
- vfloat8 l02 = unpacklo(r0,r2);
- vfloat8 h02 = unpackhi(r0,r2);
- vfloat8 l13 = unpacklo(r1,r3);
- vfloat8 h13 = unpackhi(r1,r3);
- c0 = unpacklo(l02,l13);
- c1 = unpackhi(l02,l13);
- c2 = unpacklo(h02,h13);
- c3 = unpackhi(h02,h13);
- }
- __forceinline void transpose(const vfloat8& r0, const vfloat8& r1, const vfloat8& r2, const vfloat8& r3, vfloat8& c0, vfloat8& c1, vfloat8& c2)
- {
- vfloat8 l02 = unpacklo(r0,r2);
- vfloat8 h02 = unpackhi(r0,r2);
- vfloat8 l13 = unpacklo(r1,r3);
- vfloat8 h13 = unpackhi(r1,r3);
- c0 = unpacklo(l02,l13);
- c1 = unpackhi(l02,l13);
- c2 = unpacklo(h02,h13);
- }
- __forceinline void transpose(const vfloat8& r0, const vfloat8& r1, const vfloat8& r2, const vfloat8& r3, const vfloat8& r4, const vfloat8& r5, const vfloat8& r6, const vfloat8& r7,
- vfloat8& c0, vfloat8& c1, vfloat8& c2, vfloat8& c3, vfloat8& c4, vfloat8& c5, vfloat8& c6, vfloat8& c7)
- {
- vfloat8 h0,h1,h2,h3; transpose(r0,r1,r2,r3,h0,h1,h2,h3);
- vfloat8 h4,h5,h6,h7; transpose(r4,r5,r6,r7,h4,h5,h6,h7);
- c0 = shuffle4<0,2>(h0,h4);
- c1 = shuffle4<0,2>(h1,h5);
- c2 = shuffle4<0,2>(h2,h6);
- c3 = shuffle4<0,2>(h3,h7);
- c4 = shuffle4<1,3>(h0,h4);
- c5 = shuffle4<1,3>(h1,h5);
- c6 = shuffle4<1,3>(h2,h6);
- c7 = shuffle4<1,3>(h3,h7);
- }
- __forceinline void transpose(const vfloat4& r0, const vfloat4& r1, const vfloat4& r2, const vfloat4& r3, const vfloat4& r4, const vfloat4& r5, const vfloat4& r6, const vfloat4& r7,
- vfloat8& c0, vfloat8& c1, vfloat8& c2)
- {
- transpose(vfloat8(r0,r4), vfloat8(r1,r5), vfloat8(r2,r6), vfloat8(r3,r7), c0, c1, c2);
- }
- ////////////////////////////////////////////////////////////////////////////////
- /// Reductions
- ////////////////////////////////////////////////////////////////////////////////
- __forceinline const vfloat8 vreduce_min2(const vfloat8& v) { return min(v,shuffle<1,0,3,2>(v)); }
- __forceinline const vfloat8 vreduce_min4(const vfloat8& v) { vfloat8 v1 = vreduce_min2(v); return min(v1,shuffle<2,3,0,1>(v1)); }
- __forceinline const vfloat8 vreduce_min (const vfloat8& v) { vfloat8 v1 = vreduce_min4(v); return min(v1,shuffle4<1,0>(v1)); }
- __forceinline const vfloat8 vreduce_max2(const vfloat8& v) { return max(v,shuffle<1,0,3,2>(v)); }
- __forceinline const vfloat8 vreduce_max4(const vfloat8& v) { vfloat8 v1 = vreduce_max2(v); return max(v1,shuffle<2,3,0,1>(v1)); }
- __forceinline const vfloat8 vreduce_max (const vfloat8& v) { vfloat8 v1 = vreduce_max4(v); return max(v1,shuffle4<1,0>(v1)); }
- __forceinline const vfloat8 vreduce_add2(const vfloat8& v) { return v + shuffle<1,0,3,2>(v); }
- __forceinline const vfloat8 vreduce_add4(const vfloat8& v) { vfloat8 v1 = vreduce_add2(v); return v1 + shuffle<2,3,0,1>(v1); }
- __forceinline const vfloat8 vreduce_add (const vfloat8& v) { vfloat8 v1 = vreduce_add4(v); return v1 + shuffle4<1,0>(v1); }
- __forceinline float reduce_min(const vfloat8& v) { return toScalar(vreduce_min(v)); }
- __forceinline float reduce_max(const vfloat8& v) { return toScalar(vreduce_max(v)); }
- __forceinline float reduce_add(const vfloat8& v) { return toScalar(vreduce_add(v)); }
- __forceinline size_t select_min(const vboolf8& valid, const vfloat8& v)
- {
- const vfloat8 a = select(valid,v,vfloat8(pos_inf));
- const vbool8 valid_min = valid & (a == vreduce_min(a));
- return __bsf(movemask(any(valid_min) ? valid_min : valid));
- }
- __forceinline size_t select_max(const vboolf8& valid, const vfloat8& v)
- {
- const vfloat8 a = select(valid,v,vfloat8(neg_inf));
- const vbool8 valid_max = valid & (a == vreduce_max(a));
- return __bsf(movemask(any(valid_max) ? valid_max : valid));
- }
- ////////////////////////////////////////////////////////////////////////////////
- /// Euclidian Space Operators (pairs of Vec3fa's)
- ////////////////////////////////////////////////////////////////////////////////
- //__forceinline vfloat8 dot ( const vfloat8& a, const vfloat8& b ) {
- // return vreduce_add4(a*b);
- //}
- __forceinline vfloat8 dot ( const vfloat8& a, const vfloat8& b ) {
- return _mm256_dp_ps(a,b,0x7F);
- }
- __forceinline vfloat8 cross ( const vfloat8& a, const vfloat8& b )
- {
- const vfloat8 a0 = a;
- const vfloat8 b0 = shuffle<1,2,0,3>(b);
- const vfloat8 a1 = shuffle<1,2,0,3>(a);
- const vfloat8 b1 = b;
- return shuffle<1,2,0,3>(msub(a0,b0,a1*b1));
- }
- //__forceinline float sqr_length ( const vfloat<8>& a ) { return dot(a,a); }
- //__forceinline float rcp_length ( const vfloat<8>& a ) { return rsqrt(dot(a,a)); }
- //__forceinline float rcp_length2( const vfloat<8>& a ) { return rcp(dot(a,a)); }
- //__forceinline float length ( const vfloat<8>& a ) { return sqrt(dot(a,a)); }
- __forceinline vfloat<8> normalize( const vfloat<8>& a ) { return a*rsqrt(dot(a,a)); }
- //__forceinline float distance ( const vfloat<8>& a, const vfloat<8>& b ) { return length(a-b); }
- //__forceinline float halfArea ( const vfloat<8>& d ) { return madd(d.x,(d.y+d.z),d.y*d.z); }
- //__forceinline float area ( const vfloat<8>& d ) { return 2.0f*halfArea(d); }
- //__forceinline vfloat<8> reflect ( const vfloat<8>& V, const vfloat<8>& N ) { return 2.0f*dot(V,N)*N-V; }
- //__forceinline vfloat<8> normalize_safe( const vfloat<8>& a ) {
- // const float d = dot(a,a); if (unlikely(d == 0.0f)) return a; else return a*rsqrt(d);
- //}
- ////////////////////////////////////////////////////////////////////////////////
- /// In Register Sorting
- ////////////////////////////////////////////////////////////////////////////////
- __forceinline vfloat8 sortNetwork(const vfloat8& v)
- {
- const vfloat8 a0 = v;
- const vfloat8 b0 = shuffle<1,0,3,2>(a0);
- const vfloat8 c0 = min(a0,b0);
- const vfloat8 d0 = max(a0,b0);
- const vfloat8 a1 = select<0x99 /* 0b10011001 */>(c0,d0);
- const vfloat8 b1 = shuffle<2,3,0,1>(a1);
- const vfloat8 c1 = min(a1,b1);
- const vfloat8 d1 = max(a1,b1);
- const vfloat8 a2 = select<0xc3 /* 0b11000011 */>(c1,d1);
- const vfloat8 b2 = shuffle<1,0,3,2>(a2);
- const vfloat8 c2 = min(a2,b2);
- const vfloat8 d2 = max(a2,b2);
- const vfloat8 a3 = select<0xa5 /* 0b10100101 */>(c2,d2);
- const vfloat8 b3 = shuffle4<1,0>(a3);
- const vfloat8 c3 = min(a3,b3);
- const vfloat8 d3 = max(a3,b3);
- const vfloat8 a4 = select<0xf /* 0b00001111 */>(c3,d3);
- const vfloat8 b4 = shuffle<2,3,0,1>(a4);
- const vfloat8 c4 = min(a4,b4);
- const vfloat8 d4 = max(a4,b4);
- const vfloat8 a5 = select<0x33 /* 0b00110011 */>(c4,d4);
- const vfloat8 b5 = shuffle<1,0,3,2>(a5);
- const vfloat8 c5 = min(a5,b5);
- const vfloat8 d5 = max(a5,b5);
- const vfloat8 a6 = select<0x55 /* 0b01010101 */>(c5,d5);
- return a6;
- }
- ////////////////////////////////////////////////////////////////////////////////
- /// Output Operators
- ////////////////////////////////////////////////////////////////////////////////
- inline std::ostream& operator<<(std::ostream& cout, const vfloat8& a) {
- return cout << "<" << a[0] << ", " << a[1] << ", " << a[2] << ", " << a[3] << ", " << a[4] << ", " << a[5] << ", " << a[6] << ", " << a[7] << ">";
- }
- }
|