|
@@ -4,103 +4,135 @@ package simd_x86
|
|
|
import "core:intrinsics"
|
|
|
import "core:simd"
|
|
|
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_pause :: #force_inline proc "c" () {
|
|
|
pause()
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_clflush :: #force_inline proc "c" (p: rawptr) {
|
|
|
clflush(p)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_lfence :: #force_inline proc "c" () {
|
|
|
lfence()
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_mfence :: #force_inline proc "c" () {
|
|
|
mfence()
|
|
|
}
|
|
|
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_add_epi8 :: #force_inline proc "c" (a, b: __m128i) -> __m128i {
|
|
|
return transmute(__m128i)simd.add(transmute(i8x16)a, transmute(i8x16)b)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_add_epi16 :: #force_inline proc "c" (a, b: __m128i) -> __m128i {
|
|
|
return transmute(__m128i)simd.add(transmute(i16x8)a, transmute(i16x8)b)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_add_epi32 :: #force_inline proc "c" (a, b: __m128i) -> __m128i {
|
|
|
return transmute(__m128i)simd.add(transmute(i32x4)a, transmute(i32x4)b)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_add_epi64 :: #force_inline proc "c" (a, b: __m128i) -> __m128i {
|
|
|
return transmute(__m128i)simd.add(transmute(i64x2)a, transmute(i64x2)b)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_adds_epi8 :: #force_inline proc "c" (a, b: __m128i) -> __m128i {
|
|
|
return transmute(__m128i)simd.add_sat(transmute(i8x16)a, transmute(i8x16)b)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_adds_epi16 :: #force_inline proc "c" (a, b: __m128i) -> __m128i {
|
|
|
return transmute(__m128i)simd.add_sat(transmute(i16x8)a, transmute(i16x8)b)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_adds_epu8 :: #force_inline proc "c" (a, b: __m128i) -> __m128i {
|
|
|
return transmute(__m128i)simd.add_sat(transmute(u8x16)a, transmute(u8x16)b)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_adds_epu16 :: #force_inline proc "c" (a, b: __m128i) -> __m128i {
|
|
|
return transmute(__m128i)simd.add_sat(transmute(u16x8)a, transmute(u16x8)b)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_avg_epu8 :: #force_inline proc "c" (a, b: __m128i) -> __m128i {
|
|
|
return transmute(__m128i)pavgb(transmute(u8x16)a, transmute(u8x16)b)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_avg_epu16 :: #force_inline proc "c" (a, b: __m128i) -> __m128i {
|
|
|
return transmute(__m128i)pavgw(transmute(u16x8)a, transmute(u16x8)b)
|
|
|
}
|
|
|
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_madd_epi16 :: #force_inline proc "c" (a, b: __m128i) -> __m128i {
|
|
|
return transmute(__m128i)pmaddwd(transmute(i16x8)a, transmute(i16x8)b)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_max_epi16 :: #force_inline proc "c" (a, b: __m128i) -> __m128i {
|
|
|
return transmute(__m128i)pmaxsw(transmute(i16x8)a, transmute(i16x8)b)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_max_epu8 :: #force_inline proc "c" (a, b: __m128i) -> __m128i {
|
|
|
return transmute(__m128i)pmaxub(transmute(u8x16)a, transmute(u8x16)b)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_min_epi16 :: #force_inline proc "c" (a, b: __m128i) -> __m128i {
|
|
|
return transmute(__m128i)pminsw(transmute(i16x8)a, transmute(i16x8)b)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_min_epu8 :: #force_inline proc "c" (a, b: __m128i) -> __m128i {
|
|
|
return transmute(__m128i)pminub(transmute(u8x16)a, transmute(u8x16)b)
|
|
|
}
|
|
|
|
|
|
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_mulhi_epi16 :: #force_inline proc "c" (a, b: __m128i) -> __m128i {
|
|
|
return transmute(__m128i)pmulhw(transmute(i16x8)a, transmute(i16x8)b)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_mulhi_epu16 :: #force_inline proc "c" (a, b: __m128i) -> __m128i {
|
|
|
return transmute(__m128i)pmulhuw(transmute(u16x8)a, transmute(u16x8)b)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_mullo_epi16 :: #force_inline proc "c" (a, b: __m128i) -> __m128i {
|
|
|
return transmute(__m128i)simd.mul(transmute(i16x8)a, transmute(i16x8)b)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_mul_epu32 :: #force_inline proc "c" (a, b: __m128i) -> __m128i {
|
|
|
return transmute(__m128i)pmuludq(transmute(u32x4)a, transmute(u32x4)b)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_sad_epu8 :: #force_inline proc "c" (a, b: __m128i) -> __m128i {
|
|
|
return transmute(__m128i)psadbw(transmute(u8x16)a, transmute(u8x16)b)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_sub_epi8 :: #force_inline proc "c" (a, b: __m128i) -> __m128i {
|
|
|
return transmute(__m128i)simd.sub(transmute(i8x16)a, transmute(i8x16)b)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_sub_epi16 :: #force_inline proc "c" (a, b: __m128i) -> __m128i {
|
|
|
return transmute(__m128i)simd.sub(transmute(i16x8)a, transmute(i16x8)b)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_sub_epi32 :: #force_inline proc "c" (a, b: __m128i) -> __m128i {
|
|
|
return transmute(__m128i)simd.sub(transmute(i32x4)a, transmute(i32x4)b)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_sub_epi64 :: #force_inline proc "c" (a, b: __m128i) -> __m128i {
|
|
|
return transmute(__m128i)simd.sub(transmute(i64x2)a, transmute(i64x2)b)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_subs_epi8 :: #force_inline proc "c" (a, b: __m128i) -> __m128i {
|
|
|
return transmute(__m128i)simd.sub_sat(transmute(i8x16)a, transmute(i8x16)b)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_subs_epi16 :: #force_inline proc "c" (a, b: __m128i) -> __m128i {
|
|
|
return transmute(__m128i)simd.sub_sat(transmute(i16x8)a, transmute(i16x8)b)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_subs_epu8 :: #force_inline proc "c" (a, b: __m128i) -> __m128i {
|
|
|
return transmute(__m128i)simd.sub_sat(transmute(u8x16)a, transmute(u8x16)b)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_subs_epu16 :: #force_inline proc "c" (a, b: __m128i) -> __m128i {
|
|
|
return transmute(__m128i)simd.sub_sat(transmute(u16x8)a, transmute(u16x8)b)
|
|
|
}
|
|
@@ -108,6 +140,7 @@ _mm_subs_epu16 :: #force_inline proc "c" (a, b: __m128i) -> __m128i {
|
|
|
|
|
|
|
|
|
@(private)
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_slli_si128_impl :: #force_inline proc "c" (a: __m128i, $IMM8: u32) -> __m128i {
|
|
|
shift :: IMM8 & 0xff
|
|
|
|
|
@@ -134,6 +167,7 @@ _mm_slli_si128_impl :: #force_inline proc "c" (a: __m128i, $IMM8: u32) -> __m128
|
|
|
}
|
|
|
|
|
|
@(private)
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_srli_si128_impl :: #force_inline proc "c" (a: __m128i, $IMM8: u32) -> __m128i {
|
|
|
shift :: IMM8
|
|
|
return transmute(__m128i)simd.shuffle(
|
|
@@ -159,203 +193,264 @@ _mm_srli_si128_impl :: #force_inline proc "c" (a: __m128i, $IMM8: u32) -> __m128
|
|
|
}
|
|
|
|
|
|
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_slli_si128 :: #force_inline proc "c" (a: __m128i, $IMM8: u32) -> __m128i {
|
|
|
return _mm_slli_si128_impl(a, IMM8)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_bslli_si128 :: #force_inline proc "c" (a: __m128i, $IMM8: u32) -> __m128i {
|
|
|
return _mm_slli_si128_impl(a, IMM8)
|
|
|
}
|
|
|
|
|
|
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_bsrli_si128 :: #force_inline proc "c" (a: __m128i, $IMM8: u32) -> __m128i {
|
|
|
return _mm_srli_si128_impl(a, IMM8)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_slli_epi16 :: #force_inline proc "c" (a: __m128i, $IMM8: u32) -> __m128i {
|
|
|
return transmute(__m128i)pslliw(transmute(i16x8)a, IMM8)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_sll_epi16 :: #force_inline proc "c" (a, count: __m128i) -> __m128i {
|
|
|
return transmute(__m128i)psllw(transmute(i16x8)a, transmute(i16x8)count)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_slli_epi32 :: #force_inline proc "c" (a: __m128i, $IMM8: u32) -> __m128i {
|
|
|
return transmute(__m128i)psllid(transmute(i32x4)a, IMM8)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_sll_epi32 :: #force_inline proc "c" (a, count: __m128i) -> __m128i {
|
|
|
return transmute(__m128i)pslld(transmute(i32x4)a, transmute(i32x4)count)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_slli_epi64 :: #force_inline proc "c" (a: __m128i, $IMM8: u32) -> __m128i {
|
|
|
return transmute(__m128i)pslliq(transmute(i64x2)a, IMM8)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_sll_epi64 :: #force_inline proc "c" (a, count: __m128i) -> __m128i {
|
|
|
return transmute(__m128i)psllq(transmute(i64x2)a, transmute(i64x2)count)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_srai_epi16 :: #force_inline proc "c" (a: __m128i, $IMM8: u32) -> __m128i {
|
|
|
return transmute(__m128i)psraiw(transmute(i16x8)a. IMM8)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_sra_epi16 :: #force_inline proc "c" (a, count: __m128i) -> __m128i {
|
|
|
return transmute(__m128i)psraw(transmute(i16x8)a, transmute(i16x8)count)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_srai_epi32 :: #force_inline proc "c" (a: __m128i, $IMM8: u32) -> __m128i {
|
|
|
return transmute(__m128i)psraid(transmute(i32x4)a, IMM8)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_sra_epi32 :: #force_inline proc "c" (a, count: __m128i) -> __m128i {
|
|
|
return transmute(__m128i)psrad(transmute(i32x4)a, transmute(i32x4)count)
|
|
|
}
|
|
|
|
|
|
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_srli_si128 :: #force_inline proc "c" (a: __m128i, $IMM8: u32) -> __m128i {
|
|
|
return _mm_srli_si128_impl(a, IMM8)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_srli_epi16 :: #force_inline proc "c" (a: __m128i, $IMM8: u32) -> __m128i {
|
|
|
return transmute(__m128i)psrliw(transmute(i16x8)a. IMM8)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_srl_epi16 :: #force_inline proc "c" (a, count: __m128i) -> __m128i {
|
|
|
return transmute(__m128i)psrlw(transmute(i16x8)a, transmute(i16x8)count)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_srli_epi32 :: #force_inline proc "c" (a: __m128i, $IMM8: u32) -> __m128i {
|
|
|
return transmute(__m128i)psrlid(transmute(i32x4)a, IMM8)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_srl_epi32 :: #force_inline proc "c" (a, count: __m128i) -> __m128i {
|
|
|
return transmute(__m128i)psrld(transmute(i32x4)a, transmute(i32x4)count)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_srli_epi64 :: #force_inline proc "c" (a: __m128i, $IMM8: u32) -> __m128i {
|
|
|
return transmute(__m128i)psrliq(transmute(i64x2)a, IMM8)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_srl_epi64 :: #force_inline proc "c" (a, count: __m128i) -> __m128i {
|
|
|
return transmute(__m128i)psrlq(transmute(i64x2)a, transmute(i64x2)count)
|
|
|
}
|
|
|
|
|
|
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_and_si128 :: #force_inline proc "c" (a, b: __m128i) -> __m128i {
|
|
|
return simd.and(a, b)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_andnot_si128 :: #force_inline proc "c" (a, b: __m128i) -> __m128i {
|
|
|
return simd.and_not(b, a)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_or_si128 :: #force_inline proc "c" (a, b: __m128i) -> __m128i {
|
|
|
return simd.or(a, b)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_xor_si128 :: #force_inline proc "c" (a, b: __m128i) -> __m128i {
|
|
|
return simd.xor(a, b)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_cmpeq_epi8 :: #force_inline proc "c" (a, b: __m128i) -> __m128i {
|
|
|
return transmute(__m128i)simd.lanes_eq(transmute(i8x16)a, transmute(i8x16)b)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_cmpeq_epi16 :: #force_inline proc "c" (a, b: __m128i) -> __m128i {
|
|
|
return transmute(__m128i)simd.lanes_eq(transmute(i16x8)a, transmute(i16x8)b)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_cmpeq_epi32 :: #force_inline proc "c" (a, b: __m128i) -> __m128i {
|
|
|
return transmute(__m128i)simd.lanes_eq(transmute(i32x4)a, transmute(i32x4)b)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_cmpgt_epi8 :: #force_inline proc "c" (a, b: __m128i) -> __m128i {
|
|
|
return transmute(__m128i)simd.lanes_gt(transmute(i8x16)a, transmute(i8x16)b)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_cmpgt_epi16 :: #force_inline proc "c" (a, b: __m128i) -> __m128i {
|
|
|
return transmute(__m128i)simd.lanes_gt(transmute(i16x8)a, transmute(i16x8)b)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_cmpgt_epi32 :: #force_inline proc "c" (a, b: __m128i) -> __m128i {
|
|
|
return transmute(__m128i)simd.lanes_gt(transmute(i32x4)a, transmute(i32x4)b)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_cmplt_epi8 :: #force_inline proc "c" (a, b: __m128i) -> __m128i {
|
|
|
return transmute(__m128i)simd.lanes_lt(transmute(i8x16)a, transmute(i8x16)b)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_cmplt_epi16 :: #force_inline proc "c" (a, b: __m128i) -> __m128i {
|
|
|
return transmute(__m128i)simd.lanes_lt(transmute(i16x8)a, transmute(i16x8)b)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_cmplt_epi32 :: #force_inline proc "c" (a, b: __m128i) -> __m128i {
|
|
|
return transmute(__m128i)simd.lanes_lt(transmute(i32x4)a, transmute(i32x4)b)
|
|
|
}
|
|
|
|
|
|
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_cvtepi32_pd :: #force_inline proc "c" (a: __m128i) -> __m128d {
|
|
|
v := transmute(i32x4)a
|
|
|
return cast(__m128d)simd.shuffle(v, v, 0, 1)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_cvtsi32_sd :: #force_inline proc "c" (a: __m128d, b: i32) -> __m128d {
|
|
|
return simd.replace(a, 0, f64(b))
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_cvtepi32_ps :: #force_inline proc "c" (a: __m128i) -> __m128 {
|
|
|
return cvtdq2ps(transmute(i32x4)a)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_cvtps_epi32 :: #force_inline proc "c" (a: __m128) -> __m128i {
|
|
|
return transmute(__m128i)cvtps2dq(a)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_cvtsi32_si128 :: #force_inline proc "c" (a: i32) -> __m128i {
|
|
|
return transmute(__m128i)i32x4{a, 0, 0, 0}
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_cvtsi128_si32 :: #force_inline proc "c" (a: __m128i) -> i32 {
|
|
|
return simd.extract(transmute(i32x4)a, 0)
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_set_epi64x :: #force_inline proc "c" (e1, e0: i64) -> __m128i {
|
|
|
return transmute(__m128i)i64x2{e0, e1}
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_set_epi32 :: #force_inline proc "c" (e3, e2, e1, e0: i32) -> __m128i {
|
|
|
return transmute(__m128i)i32x4{e0, e1, e2, e3}
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_set_epi16 :: #force_inline proc "c" (e7, e6, e5, e4, e3, e2, e1, e0: i16) -> __m128i {
|
|
|
return transmute(__m128i)i16x8{e0, e1, e2, e3, e4, e5, e6, e7}
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_set_epi8 :: #force_inline proc "c" (e15, e14, e13, e12, e11, e10, e9, e8, e7, e6, e5, e4, e3, e2, e1, e0: i8) -> __m128i {
|
|
|
return transmute(__m128i)i8x16{e0, e1, e2, e3, e4, e5, e6, e7, e8, e9, e10, e11, e12, e13, e14, e15}
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_set1_epi64x :: #force_inline proc "c" (a: i64) -> __m128i {
|
|
|
return _mm_set_epi64x(a, a)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_set1_epi32 :: #force_inline proc "c" (a: i32) -> __m128i {
|
|
|
return _mm_set_epi32(a, a, a, a)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_set1_epi16 :: #force_inline proc "c" (a: i16) -> __m128i {
|
|
|
return _mm_set_epi16(a, a, a, a, a, a, a, a)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_set1_epi8 :: #force_inline proc "c" (a: i8) -> __m128i {
|
|
|
return _mm_set_epi8(a, a, a, a, a, a, a, a, a, a, a, a, a, a, a, a)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_setr_epi32 :: #force_inline proc "c" (e3, e2, e1, e0: i32) -> __m128i {
|
|
|
return _mm_set_epi32(e0, e1, e2, e3)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_setr_epi16 :: #force_inline proc "c" (e7, e6, e5, e4, e3, e2, e1, e0: i16) -> __m128i {
|
|
|
return _mm_set_epi16(e0, e1, e2, e3, e4, e5, e6, e7)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_setr_epi8 :: #force_inline proc "c" (e15, e14, e13, e12, e11, e10, e9, e8, e7, e6, e5, e4, e3, e2, e1, e0: i8) -> __m128i {
|
|
|
return _mm_set_epi8(e0, e1, e2, e3, e4, e5, e6, e7, e8, e9, e10, e11, e12, e13, e14, e15)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_setzero_si128 :: #force_inline proc "c" () -> __m128i {
|
|
|
return _mm_set1_epi64x(0)
|
|
|
}
|
|
|
|
|
|
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_loadl_epi64 :: #force_inline proc "c" (mem_addr: ^__m128i) -> __m128i {
|
|
|
return _mm_set_epi64x(0, intrinsics.unaligned_load((^i64)(mem_addr)))
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_load_si128 :: #force_inline proc "c" (mem_addr: ^__m128i) -> __m128i {
|
|
|
return mem_addr^
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_loadu_si128 :: #force_inline proc "c" (mem_addr: ^__m128i) -> __m128i {
|
|
|
dst := _mm_undefined_si128()
|
|
|
intrinsics.mem_copy_non_overlapping(&dst, mem_addr, size_of(__m128i))
|
|
|
return dst
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_maskmoveu_si128 :: #force_inline proc "c" (a, mask: __m128i, mem_addr: rawptr) {
|
|
|
maskmovdqu(transmute(i8x16)a, transmute(i8x16)mask, mem_addr)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_store_si128 :: #force_inline proc "c" (mem_addr: ^__m128i, a: __m128i) {
|
|
|
mem_addr^ = a
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_storeu_si128 :: #force_inline proc "c" (mem_addr: ^__m128i, a: __m128i) {
|
|
|
storeudq(mem_addr, a)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_storel_epi64 :: #force_inline proc "c" (mem_addr: ^__m128i, a: __m128i) {
|
|
|
a := a
|
|
|
intrinsics.mem_copy_non_overlapping(mem_addr, &a, 8)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_stream_si128 :: #force_inline proc "c" (mem_addr: ^__m128i, a: __m128i) {
|
|
|
intrinsics.non_temporal_store(mem_addr, a)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_stream_si32 :: #force_inline proc "c" (mem_addr: ^i32, a: i32) {
|
|
|
intrinsics.non_temporal_store(mem_addr, a)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_move_epi64 :: #force_inline proc "c" (a: __m128i) -> __m128i {
|
|
|
zero := _mm_setzero_si128()
|
|
|
return transmute(__m128i)simd.shuffle(transmute(i64x2)a, transmute(i64x2)zero, 0, 2)
|
|
@@ -364,24 +459,31 @@ _mm_move_epi64 :: #force_inline proc "c" (a: __m128i) -> __m128i {
|
|
|
|
|
|
|
|
|
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_packs_epi16 :: #force_inline proc "c" (a, b: __m128i) -> __m128i {
|
|
|
return transmute(__m128i)packsswb(transmute(i16x8)a, transmute(i16x8)b)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_packs_epi32 :: #force_inline proc "c" (a, b: __m128i) -> __m128i {
|
|
|
return transmute(__m128i)packssdw(transmute(i32x4)a, transmute(i32x4)b)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_packus_epi16 :: #force_inline proc "c" (a, b: __m128i) -> __m128i {
|
|
|
return transmute(__m128i)packuswb(transmute(i16x8)a, transmute(i16x8)b)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_extract_epi16 :: #force_inline proc "c" (a: __m128i, $IMM8: u32) -> i32 {
|
|
|
return i32(simd.extract(transmute(u16x8)a, IMM8))
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_insert_epi16 :: #force_inline proc "c" (a: __m128i, i: i32, $IMM8: u32) -> __m128i {
|
|
|
return i32(simd.replace(transmute(u16x8)a, IMM8, i16(i)))
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_movemask_epi8 :: #force_inline proc "c" (a: __m128i) -> i32 {
|
|
|
return pmovmskb(transmute(i8x16)a)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_shuffle_epi32 :: #force_inline proc "c" (a: __m128i, $IMM8: u32) -> __m128i {
|
|
|
v := transmute(i32x4)a
|
|
|
return transmute(__m128i)simd.shuffle(
|
|
@@ -393,6 +495,7 @@ _mm_shuffle_epi32 :: #force_inline proc "c" (a: __m128i, $IMM8: u32) -> __m128i
|
|
|
(IMM8 >> 6) & 0b11,
|
|
|
)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_shufflehi_epi16 :: #force_inline proc "c" (a: __m128i, $IMM8: u32) -> __m128i {
|
|
|
v := transmute(i16x8)a
|
|
|
return transmute(__m128i)simd.shuffle(
|
|
@@ -408,6 +511,7 @@ _mm_shufflehi_epi16 :: #force_inline proc "c" (a: __m128i, $IMM8: u32) -> __m128
|
|
|
((IMM8 >> 6) & 0b11) + 4,
|
|
|
)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_shufflelo_epi16 :: #force_inline proc "c" (a: __m128i, $IMM8: u32) -> __m128i {
|
|
|
v := transmute(i16x8)a
|
|
|
return transmute(__m128i)simd.shuffle(
|
|
@@ -423,6 +527,7 @@ _mm_shufflelo_epi16 :: #force_inline proc "c" (a: __m128i, $IMM8: u32) -> __m128
|
|
|
7,
|
|
|
)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_unpackhi_epi8 :: #force_inline proc "c" (a, b: __m128i) -> __m128i {
|
|
|
return transmute(__m128i)simd.shuffle(
|
|
|
transmute(i8x16)a,
|
|
@@ -430,15 +535,19 @@ _mm_unpackhi_epi8 :: #force_inline proc "c" (a, b: __m128i) -> __m128i {
|
|
|
8, 24, 9, 25, 10, 26, 11, 27, 12, 28, 13, 29, 14, 30, 15, 31,
|
|
|
)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_unpackhi_epi16 :: #force_inline proc "c" (a, b: __m128i) -> __m128i {
|
|
|
return transmute(__m128i)simd.shuffle(transmute(i16x8)a, transmute(i16x8)b, 4, 12, 5, 13, 6, 14, 7, 15)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_unpackhi_epi32 :: #force_inline proc "c" (a, b: __m128i) -> __m128i {
|
|
|
return transmute(__m128i)simd.shuffle(transmute(i32x4)a, transmute(i32x4)b, 2, 6, 3, 7)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_unpackhi_epi64 :: #force_inline proc "c" (a, b: __m128i) -> __m128i {
|
|
|
return transmute(__m128i)simd.shuffle(transmute(i64x2)a, transmute(i64x2)b, 1, 3)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_unpacklo_epi8 :: #force_inline proc "c" (a, b: __m128i) -> __m128i {
|
|
|
return transmute(__m128i)simd.shuffle(
|
|
|
transmute(i8x16)a,
|
|
@@ -446,12 +555,15 @@ _mm_unpacklo_epi8 :: #force_inline proc "c" (a, b: __m128i) -> __m128i {
|
|
|
0, 16, 1, 17, 2, 18, 3, 19, 4, 20, 5, 21, 6, 22, 7, 23,
|
|
|
)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_unpacklo_epi16 :: #force_inline proc "c" (a, b: __m128i) -> __m128i {
|
|
|
return transmute(__m128i)simd.shuffle(transmute(i16x8)a, transmute(i16x8)b, 0, 8, 1, 9, 2, 10, 3, 11)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_unpacklo_epi32 :: #force_inline proc "c" (a, b: __m128i) -> __m128i {
|
|
|
return transmute(__m128i)simd.shuffle(transmute(i32x4)a, transmute(i32x4)b, 0, 4, 1, 5)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_unpacklo_epi64 :: #force_inline proc "c" (a, b: __m128i) -> __m128i {
|
|
|
return transmute(__m128i)simd.shuffle(transmute(i64x2)a, transmute(i64x2)b, 0, 2)
|
|
|
}
|
|
@@ -459,57 +571,75 @@ _mm_unpacklo_epi64 :: #force_inline proc "c" (a, b: __m128i) -> __m128i {
|
|
|
|
|
|
|
|
|
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_add_sd :: #force_inline proc "c" (a, b: __m128d) -> __m128d {
|
|
|
return simd.replace(a, 0, _mm_cvtsd_f64(a) + _mm_cvtsd_f64(b))
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_add_pd :: #force_inline proc "c" (a, b: __m128d) -> __m128d {
|
|
|
return simd.add(a, b)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_div_sd :: #force_inline proc "c" (a, b: __m128d) -> __m128d {
|
|
|
return simd.replace(a, 0, _mm_cvtsd_f64(a) / _mm_cvtsd_f64(b))
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_div_pd :: #force_inline proc "c" (a, b: __m128d) -> __m128d {
|
|
|
return simd.div(a, b)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_max_sd :: #force_inline proc "c" (a, b: __m128d) -> __m128d {
|
|
|
return maxsd(a, b)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_max_pd :: #force_inline proc "c" (a, b: __m128d) -> __m128d {
|
|
|
return maxpd(a, b)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_min_sd :: #force_inline proc "c" (a, b: __m128d) -> __m128d {
|
|
|
return minsd(a, b)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_min_pd :: #force_inline proc "c" (a, b: __m128d) -> __m128d {
|
|
|
return minpd(a, b)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_mul_sd :: #force_inline proc "c" (a, b: __m128d) -> __m128d {
|
|
|
return simd.replace(a, 0, _mm_cvtsd_f64(a) * _mm_cvtsd_f64(b))
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_mul_pd :: #force_inline proc "c" (a, b: __m128d) -> __m128d {
|
|
|
return simd.mul(a, b)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_sqrt_sd :: #force_inline proc "c" (a, b: __m128d) -> __m128d {
|
|
|
return simd.replace(a, 0, _mm_cvtsd_f64(sqrtsd(b)))
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_sqrt_pd :: #force_inline proc "c" (a: __m128d) -> __m128d {
|
|
|
return simd.sqrt(a)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_sub_sd :: #force_inline proc "c" (a, b: __m128d) -> __m128d {
|
|
|
return simd.replace(a, 0, _mm_cvtsd_f64(a) - _mm_cvtsd_f64(b))
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_sub_pd :: #force_inline proc "c" (a, b: __m128d) -> __m128d {
|
|
|
return simd.sub(a, b)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_and_pd :: #force_inline proc "c" (a, b: __m128d) -> __m128d {
|
|
|
return transmute(__m128d)_mm_and_si128(transmute(__m128i)a, transmute(__m128i)b)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_andnot_pd :: #force_inline proc "c" (a, b: __m128d) -> __m128d {
|
|
|
return transmute(__m128d)_mm_andnot_si128(transmute(__m128i)a, transmute(__m128i)b)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_or_pd :: #force_inline proc "c" (a, b: __m128d) -> __m128d {
|
|
|
return transmute(__m128d)_mm_or_si128(transmute(__m128i)a, transmute(__m128i)b)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_xor_pd :: #force_inline proc "c" (a, b: __m128d) -> __m128d {
|
|
|
return transmute(__m128d)_mm_xor_si128(transmute(__m128i)a, transmute(__m128i)b)
|
|
|
}
|
|
@@ -517,111 +647,147 @@ _mm_xor_pd :: #force_inline proc "c" (a, b: __m128d) -> __m128d {
|
|
|
|
|
|
|
|
|
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_cmpeq_sd :: #force_inline proc "c" (a, b: __m128d) -> __m128d {
|
|
|
return cmpsd(a, b, 0)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_cmplt_sd :: #force_inline proc "c" (a, b: __m128d) -> __m128d {
|
|
|
return cmpsd(a, b, 1)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_cmple_sd :: #force_inline proc "c" (a, b: __m128d) -> __m128d {
|
|
|
return cmpsd(a, b, 2)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_cmpgt_sd :: #force_inline proc "c" (a, b: __m128d) -> __m128d {
|
|
|
return simd.replace(_mm_cmplt_sd(b, a), 1, simd.extract(a, 1))
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_cmpge_sd :: #force_inline proc "c" (a, b: __m128d) -> __m128d {
|
|
|
return simd.replace(_mm_cmple_sd(b, a), 1, simd.extract(a, 1))
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_cmpord_sd :: #force_inline proc "c" (a, b: __m128d) -> __m128d {
|
|
|
return cmpsd(a, b, 7)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_cmpunord_sd :: #force_inline proc "c" (a, b: __m128d) -> __m128d {
|
|
|
return cmpsd(a, b, 3)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_cmpneq_sd :: #force_inline proc "c" (a, b: __m128d) -> __m128d {
|
|
|
return cmpsd(a, b, 4)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_cmpnlt_sd :: #force_inline proc "c" (a, b: __m128d) -> __m128d {
|
|
|
return cmpsd(a, b, 5)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_cmpnle_sd :: #force_inline proc "c" (a, b: __m128d) -> __m128d {
|
|
|
return cmpsd(a, b, 6)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_cmpngt_sd :: #force_inline proc "c" (a, b: __m128d) -> __m128d {
|
|
|
return simd.replace(_mm_cmpnlt_sd(b, a), 1, simd.extract(a, 1))
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_cmpnge_sd :: #force_inline proc "c" (a, b: __m128d) -> __m128d {
|
|
|
return simd.replace(_mm_cmpnle_sd(b, a), 1, simd.extract(a, 1))
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_cmpeq_pd :: #force_inline proc "c" (a, b: __m128d) -> __m128d {
|
|
|
return cmppd(a, b, 0)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_cmplt_pd :: #force_inline proc "c" (a, b: __m128d) -> __m128d {
|
|
|
return cmppd(a, b, 1)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_cmple_pd :: #force_inline proc "c" (a, b: __m128d) -> __m128d {
|
|
|
return cmppd(a, b, 2)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_cmpgt_pd :: #force_inline proc "c" (a, b: __m128d) -> __m128d {
|
|
|
return _mm_cmplt_pd(b, a)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_cmpge_pd :: #force_inline proc "c" (a, b: __m128d) -> __m128d {
|
|
|
return _mm_cmple_pd(b, a)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_cmpord_pd :: #force_inline proc "c" (a, b: __m128d) -> __m128d {
|
|
|
return cmppd(a, b, 7)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_cmpunord_pd :: #force_inline proc "c" (a, b: __m128d) -> __m128d {
|
|
|
return cmppd(a, b, 3)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_cmpneq_pd :: #force_inline proc "c" (a, b: __m128d) -> __m128d {
|
|
|
return cmppd(a, b, 4)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_cmpnlt_pd :: #force_inline proc "c" (a, b: __m128d) -> __m128d {
|
|
|
return cmppd(a, b, 5)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_cmpnle_pd :: #force_inline proc "c" (a, b: __m128d) -> __m128d {
|
|
|
return cmppd(a, b, 6)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_cmpngt_pd :: #force_inline proc "c" (a, b: __m128d) -> __m128d {
|
|
|
return _mm_cmpnlt_pd(b, a)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_cmpnge_pd :: #force_inline proc "c" (a, b: __m128d) -> __m128d {
|
|
|
return _mm_cmpnle_pd(b, a)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_comieq_sd :: #force_inline proc "c" (a, b: __m128d) -> i32 {
|
|
|
return comieqsd(a, b)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_comilt_sd :: #force_inline proc "c" (a, b: __m128d) -> i32 {
|
|
|
return comiltsd(a, b)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_comile_sd :: #force_inline proc "c" (a, b: __m128d) -> i32 {
|
|
|
return comilesd(a, b)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_comigt_sd :: #force_inline proc "c" (a, b: __m128d) -> i32 {
|
|
|
return comigtsd(a, b)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_comige_sd :: #force_inline proc "c" (a, b: __m128d) -> i32 {
|
|
|
return comigesd(a, b)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_comineq_sd :: #force_inline proc "c" (a, b: __m128d) -> i32 {
|
|
|
return comineqsd(a, b)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_ucomieq_sd :: #force_inline proc "c" (a, b: __m128d) -> i32 {
|
|
|
return ucomieqsd(a, b)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_ucomilt_sd :: #force_inline proc "c" (a, b: __m128d) -> i32 {
|
|
|
return ucomiltsd(a, b)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_ucomile_sd :: #force_inline proc "c" (a, b: __m128d) -> i32 {
|
|
|
return ucomilesd(a, b)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_ucomigt_sd :: #force_inline proc "c" (a, b: __m128d) -> i32 {
|
|
|
return ucomigtsd(a, b)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_ucomige_sd :: #force_inline proc "c" (a, b: __m128d) -> i32 {
|
|
|
return ucomigesd(a, b)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_ucomineq_sd :: #force_inline proc "c" (a, b: __m128d) -> i32 {
|
|
|
return ucomineqsd(a, b)
|
|
|
}
|
|
@@ -630,115 +796,151 @@ _mm_ucomineq_sd :: #force_inline proc "c" (a, b: __m128d) -> i32 {
|
|
|
|
|
|
|
|
|
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_cvtpd_ps :: #force_inline proc "c" (a: __m128d) -> __m128 {
|
|
|
return cvtpd2ps(a)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_cvtps_pd :: #force_inline proc "c" (a: __m128) -> __m128d {
|
|
|
return cvtps2pd(a)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_cvtpd_epi32 :: #force_inline proc "c" (a: __m128d) -> __m128i {
|
|
|
return transmute(__m128i)cvtpd2dq(a)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_cvtsd_si32 :: #force_inline proc "c" (a: __m128d) -> i32 {
|
|
|
return cvtsd2si(a)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_cvtsd_ss :: #force_inline proc "c" (a, b: __m128d) -> __m128 {
|
|
|
return cvtsd2ss(a, b)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_cvtsd_f64 :: #force_inline proc "c" (a: __m128d) -> f64 {
|
|
|
return simd.extract(a, 0)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_cvtss_sd :: #force_inline proc "c" (a, b: __m128) -> __m128d {
|
|
|
return cvtss2sd(a, b)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_cvttpd_epi32 :: #force_inline proc "c" (a: __m128d) -> __m128i {
|
|
|
return transmute(__m128i)cvttpd2dq(a)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_cvttsd_si32 :: #force_inline proc "c" (a: __m128d) -> i32 {
|
|
|
return cvttsd2si(a)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_cvttps_epi32 :: #force_inline proc "c" (a: __m128) -> __m128i {
|
|
|
return transmute(__m128i)cvttps2dq(a)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_set_sd :: #force_inline proc "c" (a: f64) -> __m128d {
|
|
|
return _mm_set_pd(0.0, a)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_set1_pd :: #force_inline proc "c" (a: f64) -> __m128d {
|
|
|
return _mm_set_pd(a, a)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_set_pd1 :: #force_inline proc "c" (a: f64) -> __m128d {
|
|
|
return _mm_set_pd(a, a)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_set_pd :: #force_inline proc "c" (a: f64, b: f64) -> __m128d {
|
|
|
return __m128d{b, a}
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_setr_pd :: #force_inline proc "c" (a: f64, b: f64) -> __m128d {
|
|
|
return _mm_set_pd(b, a)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_setzero_pd :: #force_inline proc "c" () -> __m128d {
|
|
|
return _mm_set_pd(0.0, 0.0)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_movemask_pd :: #force_inline proc "c" (a: __m128d) -> i32 {
|
|
|
return movmskpd(a)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_load_pd :: #force_inline proc "c" (mem_addr: ^f64) -> __m128d {
|
|
|
return (^__m128d)(mem_addr)^
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_load_sd :: #force_inline proc "c" (mem_addr: ^f64) -> __m128d {
|
|
|
return _mm_setr_pd(mem_addr^, 0.)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_loadh_pd :: #force_inline proc "c" (a: __m128d, mem_addr: ^f64) -> __m128d {
|
|
|
return _mm_setr_pd(simd.extract(a, 0), mem_addr^)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_loadl_pd :: #force_inline proc "c" (a: __m128d, mem_addr: ^f64) -> __m128d {
|
|
|
return _mm_setr_pd(mem_addr^, simd.extract(a, 1))
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_stream_pd :: #force_inline proc "c" (mem_addr: ^f64, a: __m128d) {
|
|
|
intrinsics.non_temporal_store((^__m128d)(mem_addr), a)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_store_sd :: #force_inline proc "c" (mem_addr: ^f64, a: __m128d) {
|
|
|
mem_addr^ = simd.extract(a, 0)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_store_pd :: #force_inline proc "c" (mem_addr: ^f64, a: __m128d) {
|
|
|
(^__m128d)(mem_addr)^ = a
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_storeu_pd :: #force_inline proc "c" (mem_addr: ^f64, a: __m128d) {
|
|
|
storeupd(mem_addr, a)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_store1_pd :: #force_inline proc "c" (mem_addr: ^f64, a: __m128d) {
|
|
|
(^__m128d)(mem_addr)^ = simd.shuffle(a, a, 0, 0)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_store_pd1 :: #force_inline proc "c" (mem_addr: ^f64, a: __m128d) {
|
|
|
(^__m128d)(mem_addr)^ = simd.shuffle(a, a, 0, 0)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_storer_pd :: #force_inline proc "c" (mem_addr: ^f64, a: __m128d) {
|
|
|
(^__m128d)(mem_addr)^ = simd.shuffle(a, a, 1, 0)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_storeh_pd :: #force_inline proc "c" (mem_addr: ^f64, a: __m128d) {
|
|
|
mem_addr^ = simd.extract(a, 1)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_storel_pd :: #force_inline proc "c" (mem_addr: ^f64, a: __m128d) {
|
|
|
mem_addr^ = simd.extract(a, 0)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_load1_pd :: #force_inline proc "c" (mem_addr: ^f64) -> __m128d {
|
|
|
d := mem_addr^
|
|
|
return _mm_setr_pd(d, d)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_load_pd1 :: #force_inline proc "c" (mem_addr: ^f64) -> __m128d {
|
|
|
return _mm_load1_pd(mem_addr)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_loadr_pd :: #force_inline proc "c" (mem_addr: ^f64) -> __m128d {
|
|
|
a := _mm_load_pd(mem_addr)
|
|
|
return simd.shuffle(a, a, 1, 0)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_loadu_pd :: #force_inline proc "c" (mem_addr: ^f64) -> __m128d {
|
|
|
dst := _mm_undefined_pd()
|
|
|
intrinsics.mem_copy_non_overlapping(&dst, mem_addr, size_of(__m128d))
|
|
|
return dst
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_shuffle_pd :: #force_inline proc "c" (a, b: __m128d, $MASK: u32) -> __m128d {
|
|
|
return simd.shuffle(a, b, MASK&0b1, ((MASK>>1)&0b1) + 2)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_move_sd :: #force_inline proc "c" (a, b: __m128d) -> __m128d {
|
|
|
return _mm_setr_pd(simd.extract(b, 0), simd.extract(a, 1))
|
|
|
}
|
|
@@ -746,71 +948,92 @@ _mm_move_sd :: #force_inline proc "c" (a, b: __m128d) -> __m128d {
|
|
|
|
|
|
|
|
|
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_castpd_ps :: #force_inline proc "c" (a: __m128d) -> __m128 {
|
|
|
return transmute(__m128)a
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_castpd_si128 :: #force_inline proc "c" (a: __m128d) -> __m128i {
|
|
|
return transmute(__m128i)a
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_castps_pd :: #force_inline proc "c" (a: __m128) -> __m128d {
|
|
|
return transmute(__m128d)a
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_castps_si128 :: #force_inline proc "c" (a: __m128) -> __m128i {
|
|
|
return transmute(__m128i)a
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_castsi128_pd :: #force_inline proc "c" (a: __m128i) -> __m128d {
|
|
|
return transmute(__m128d)a
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_castsi128_ps :: #force_inline proc "c" (a: __m128i) -> __m128 {
|
|
|
return transmute(__m128)a
|
|
|
}
|
|
|
|
|
|
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_undefined_pd :: #force_inline proc "c" () -> __m128d {
|
|
|
return __m128d{0, 0}
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_undefined_si128 :: #force_inline proc "c" () -> __m128i {
|
|
|
return __m128i{0, 0}
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_unpackhi_pd :: #force_inline proc "c" (a, b: __m128d) -> __m128d {
|
|
|
return simd.shuffle(a, b, 1, 3)
|
|
|
}
|
|
|
+@(enable_target_feature="sse2")
|
|
|
_mm_unpacklo_pd :: #force_inline proc "c" (a, b: __m128d) -> __m128d {
|
|
|
return simd.shuffle(a, b, 0, 2)
|
|
|
}
|
|
|
|
|
|
|
|
|
when ODIN_ARCH == .amd64 {
|
|
|
+ @(enable_target_feature="sse2")
|
|
|
_mm_cvtsd_si64 :: #force_inline proc "c" (a: __m128d) -> i64 {
|
|
|
return cvtsd2si64(a)
|
|
|
}
|
|
|
+ @(enable_target_feature="sse2")
|
|
|
_mm_cvtsd_si64x :: #force_inline proc "c" (a: __m128d) -> i64 {
|
|
|
return _mm_cvtsd_si64(a)
|
|
|
}
|
|
|
+ @(enable_target_feature="sse2")
|
|
|
_mm_cvttsd_si64 :: #force_inline proc "c" (a: __m128d) -> i64 {
|
|
|
return cvttsd2si64(a)
|
|
|
}
|
|
|
+ @(enable_target_feature="sse2")
|
|
|
_mm_cvttsd_si64x :: #force_inline proc "c" (a: __m128d) -> i64 {
|
|
|
return _mm_cvttsd_si64(a)
|
|
|
}
|
|
|
+ @(enable_target_feature="sse2")
|
|
|
_mm_stream_si64 :: #force_inline proc "c" (mem_addr: ^i64, a: i64) {
|
|
|
intrinsics.non_temporal_store(mem_addr, a)
|
|
|
}
|
|
|
+ @(enable_target_feature="sse2")
|
|
|
_mm_cvtsi64_si128 :: #force_inline proc "c" (a: i64) -> __m128i {
|
|
|
return _mm_set_epi64x(0, a)
|
|
|
}
|
|
|
+ @(enable_target_feature="sse2")
|
|
|
_mm_cvtsi64x_si128 :: #force_inline proc "c" (a: i64) -> __m128i {
|
|
|
return _mm_cvtsi64_si128(a)
|
|
|
}
|
|
|
+ @(enable_target_feature="sse2")
|
|
|
_mm_cvtsi128_si64 :: #force_inline proc "c" (a: __m128i) -> i64 {
|
|
|
return simd.extract(transmute(i64x2)a, 0)
|
|
|
}
|
|
|
+ @(enable_target_feature="sse2")
|
|
|
_mm_cvtsi128_si64x :: #force_inline proc "c" (a: __m128i) -> i64 {
|
|
|
return _mm_cvtsi128_si64(a)
|
|
|
}
|
|
|
+ @(enable_target_feature="sse2")
|
|
|
_mm_cvtsi64_sd :: #force_inline proc "c" (a: __m128d, b: i64) -> __m128d {
|
|
|
return simd.replace(a, 0, f64(b))
|
|
|
}
|
|
|
+ @(enable_target_feature="sse2")
|
|
|
_mm_cvtsi64x_sd :: #force_inline proc "c" (a: __m128d, b: i64) -> __m128d {
|
|
|
return _mm_cvtsi64_sd(a, b)
|
|
|
}
|