Просмотр исходного кода

Uniformalize low level SIMD API

Christophe Riccio 9 лет назад
Родитель
Сommit
0ee3a79bfd
5 измененных файлов с 88 добавлено и 71 удалено
  1. 1 1
      glm/detail/func_geometric_simd.inl
  2. 1 1
      glm/detail/func_matrix_simd.inl
  3. 49 32
      glm/simd/common.h
  4. 18 18
      glm/simd/geometric.h
  5. 19 19
      glm/simd/matrix.h

+ 1 - 1
glm/detail/func_geometric_simd.inl

@@ -8,7 +8,7 @@ namespace detail
 	{
 		GLM_FUNC_QUALIFIER static float call(tvec4<float, P> const& x, tvec4<float, P> const& y)
 		{
-			__m128 const dot0 = glm_dot_ss(x.data, y.data);
+			__m128 const dot0 = glm_f32v1_dot(x.data, y.data);
 			return _mm_cvtss_f32(dot0);
 		}
 	};

+ 1 - 1
glm/detail/func_matrix_simd.inl

@@ -15,7 +15,7 @@ namespace detail
 			GLM_FUNC_QUALIFIER static tmat4x4<float, P> call(tmat4x4<float, P> const& m)
 			{
 				tmat4x4<float, P> Result(uninitialize);
-				glm_inverse_ps(*reinterpret_cast<__m128 const(*)[4]>(&m[0].data), *reinterpret_cast<__m128(*)[4]>(&Result[0].data));
+				glm_f32m4_inverse(*reinterpret_cast<__m128 const(*)[4]>(&m[0].data), *reinterpret_cast<__m128(*)[4]>(&Result[0].data));
 				return Result;
 			}
 		};

+ 49 - 32
glm/simd/common.h

@@ -10,17 +10,15 @@ static const __m128 GLM_VAR_USED glm_minus_one = _mm_set_ps1(-1.0f);
 static const __m128 GLM_VAR_USED glm_two = _mm_set_ps1(2.0f);
 static const __m128 GLM_VAR_USED glm_three = _mm_set_ps1(3.0f);
 
-static const __m128 GLM_VAR_USED glm_epi32_sign_mask = _mm_castsi128_ps(_mm_set1_epi32(static_cast<int>(0x80000000)));
 static const __m128 GLM_VAR_USED glm_ps_2pow23 = _mm_set_ps1(8388608.0f);
-static const __m128 GLM_VAR_USED glm_ps_1 = _mm_set_ps1(1.0f);
 
-GLM_FUNC_QUALIFIER __m128 glm_abs_ps(__m128 x)
+GLM_FUNC_QUALIFIER __m128 glm_f32v4_abs(__m128 x)
 {
 	return _mm_and_ps(x, _mm_castsi128_ps(_mm_set1_epi32(0x7FFFFFFF)));
 }
 
 //sign
-GLM_FUNC_QUALIFIER __m128 glm_sgn_ps(__m128 x)
+GLM_FUNC_QUALIFIER __m128 glm_f32v4_sgn(__m128 x)
 {
 	__m128 const Cmp0 = _mm_cmplt_ps(x, glm_zero);
 	__m128 const Cmp1 = _mm_cmpgt_ps(x, glm_zero);
@@ -30,9 +28,10 @@ GLM_FUNC_QUALIFIER __m128 glm_sgn_ps(__m128 x)
 }
 
 //round
-GLM_FUNC_QUALIFIER __m128 glm_rnd_ps(__m128 x)
+GLM_FUNC_QUALIFIER __m128 glm_f32v4_rnd(__m128 x)
 {
-	__m128 const and0 = _mm_and_ps(glm_epi32_sign_mask, x);
+	__m128 const sgn0 = _mm_castsi128_ps(_mm_set1_epi32(static_cast<int>(0x80000000)));
+	__m128 const and0 = _mm_and_ps(sgn0, x);
 	__m128 const or0 = _mm_or_ps(and0, glm_ps_2pow23);
 	__m128 const add0 = _mm_add_ps(x, or0);
 	__m128 const sub0 = _mm_sub_ps(add0, or0);
@@ -40,11 +39,11 @@ GLM_FUNC_QUALIFIER __m128 glm_rnd_ps(__m128 x)
 }
 
 //floor
-GLM_FUNC_QUALIFIER __m128 glm_flr_ps(__m128 x)
+GLM_FUNC_QUALIFIER __m128 glm_f32v4_flr(__m128 x)
 {
-	__m128 const rnd0 = glm_rnd_ps(x);
+	__m128 const rnd0 = glm_f32v4_rnd(x);
 	__m128 const cmp0 = _mm_cmplt_ps(x, rnd0);
-	__m128 const and0 = _mm_and_ps(cmp0, glm_ps_1);
+	__m128 const and0 = _mm_and_ps(cmp0, _mm_set1_ps(1.0f));
 	__m128 const sub0 = _mm_sub_ps(rnd0, and0);
 	return sub0;
 }
@@ -56,50 +55,51 @@ GLM_FUNC_QUALIFIER __m128 glm_flr_ps(__m128 x)
 //}
 
 //roundEven
-GLM_FUNC_QUALIFIER __m128 glm_rde_ps(__m128 x)
+GLM_FUNC_QUALIFIER __m128 glm_f32v4_rde(__m128 x)
 {
-	__m128 const and0 = _mm_and_ps(glm_epi32_sign_mask, x);
+	__m128 const sgn0 = _mm_castsi128_ps(_mm_set1_epi32(static_cast<int>(0x80000000)));
+	__m128 const and0 = _mm_and_ps(sgn0, x);
 	__m128 const or0 = _mm_or_ps(and0, glm_ps_2pow23);
 	__m128 const add0 = _mm_add_ps(x, or0);
 	__m128 const sub0 = _mm_sub_ps(add0, or0);
 	return sub0;
 }
 
-GLM_FUNC_QUALIFIER __m128 glm_ceil_ps(__m128 x)
+GLM_FUNC_QUALIFIER __m128 glm_f32v4_ceil(__m128 x)
 {
-	__m128 const rnd0 = glm_rnd_ps(x);
+	__m128 const rnd0 = glm_f32v4_rnd(x);
 	__m128 const cmp0 = _mm_cmpgt_ps(x, rnd0);
-	__m128 const and0 = _mm_and_ps(cmp0, glm_ps_1);
+	__m128 const and0 = _mm_and_ps(cmp0, _mm_set1_ps(1.0f));
 	__m128 const add0 = _mm_add_ps(rnd0, and0);
 	return add0;
 }
 
-GLM_FUNC_QUALIFIER __m128 glm_frc_ps(__m128 x)
+GLM_FUNC_QUALIFIER __m128 glm_f32v4_frc(__m128 x)
 {
-	__m128 const flr0 = glm_flr_ps(x);
+	__m128 const flr0 = glm_f32v4_flr(x);
 	__m128 const sub0 = _mm_sub_ps(x, flr0);
 	return sub0;
 }
 
-GLM_FUNC_QUALIFIER __m128 glm_mod_ps(__m128 x, __m128 y)
+GLM_FUNC_QUALIFIER __m128 glm_f32v4_mod(__m128 x, __m128 y)
 {
 	__m128 const div0 = _mm_div_ps(x, y);
-	__m128 const flr0 = glm_flr_ps(div0);
+	__m128 const flr0 = glm_f32v4_flr(div0);
 	__m128 const mul0 = _mm_mul_ps(y, flr0);
 	__m128 const sub0 = _mm_sub_ps(x, mul0);
 	return sub0;
 }
 
-GLM_FUNC_QUALIFIER __m128 glm_clp_ps(__m128 v, __m128 minVal, __m128 maxVal)
+GLM_FUNC_QUALIFIER __m128 glm_f32v4_clp(__m128 v, __m128 minVal, __m128 maxVal)
 {
 	__m128 const min0 = _mm_min_ps(v, maxVal);
 	__m128 const max0 = _mm_max_ps(min0, minVal);
 	return max0;
 }
 
-GLM_FUNC_QUALIFIER __m128 glm_mix_ps(__m128 v1, __m128 v2, __m128 a)
+GLM_FUNC_QUALIFIER __m128 glm_f32v4_mix(__m128 v1, __m128 v2, __m128 a)
 {
-	__m128 const sub0 = _mm_sub_ps(glm_one, a);
+	__m128 const sub0 = _mm_sub_ps(_mm_set1_ps(1.0f), a);
 	__m128 const mul0 = _mm_mul_ps(v1, sub0);
 	__m128 const mul1 = _mm_mul_ps(v2, a);
 	__m128 const add0 = _mm_add_ps(mul0, mul1);
@@ -107,19 +107,19 @@ GLM_FUNC_QUALIFIER __m128 glm_mix_ps(__m128 v1, __m128 v2, __m128 a)
 }
 
 //step
-GLM_FUNC_QUALIFIER __m128 glm_stp_ps(__m128 edge, __m128 x)
+GLM_FUNC_QUALIFIER __m128 glm_f32v4_stp(__m128 edge, __m128 x)
 {
 	__m128 const cmp = _mm_cmple_ps(x, edge);
-	return _mm_movemask_ps(cmp) == 0 ? glm_one : glm_zero;
+	return _mm_movemask_ps(cmp) == 0 ? _mm_set1_ps(1.0f) : _mm_set1_ps(0.0f);
 }
 
 // smoothstep
-GLM_FUNC_QUALIFIER __m128 glm_ssp_ps(__m128 edge0, __m128 edge1, __m128 x)
+GLM_FUNC_QUALIFIER __m128 glm_f32v4_ssp(__m128 edge0, __m128 edge1, __m128 x)
 {
 	__m128 const sub0 = _mm_sub_ps(x, edge0);
 	__m128 const sub1 = _mm_sub_ps(edge1, edge0);
 	__m128 const div0 = _mm_sub_ps(sub0, sub1);
-	__m128 const clp0 = glm_clp_ps(div0, glm_zero, glm_one);
+	__m128 const clp0 = glm_f32v4_clp(div0, _mm_set1_ps(0.0f), _mm_set1_ps(1.0f));
 	__m128 const mul0 = _mm_mul_ps(glm_two, clp0);
 	__m128 const sub2 = _mm_sub_ps(glm_three, mul0);
 	__m128 const mul1 = _mm_mul_ps(clp0, clp0);
@@ -128,7 +128,7 @@ GLM_FUNC_QUALIFIER __m128 glm_ssp_ps(__m128 edge0, __m128 edge1, __m128 x)
 }
 
 // Agner Fog method
-GLM_FUNC_QUALIFIER __m128 glm_nan_ps(__m128 x)
+GLM_FUNC_QUALIFIER __m128 glm_f32v4_nan(__m128 x)
 {
 	__m128i const t1 = _mm_castps_si128(x);						// reinterpret as 32-bit integer
 	__m128i const t2 = _mm_sll_epi32(t1, _mm_cvtsi32_si128(1));	// shift out sign bit
@@ -142,7 +142,7 @@ GLM_FUNC_QUALIFIER __m128 glm_nan_ps(__m128 x)
 }
 
 // Agner Fog method
-GLM_FUNC_QUALIFIER __m128 glm_inf_ps(__m128 x)
+GLM_FUNC_QUALIFIER __m128 glm_f32v4_inf(__m128 x)
 {
 	__m128i const t1 = _mm_castps_si128(x);											// reinterpret as 32-bit integer
 	__m128i const t2 = _mm_sll_epi32(t1, _mm_cvtsi32_si128(1));						// shift out sign bit
@@ -151,10 +151,27 @@ GLM_FUNC_QUALIFIER __m128 glm_inf_ps(__m128 x)
 
 // SSE scalar reciprocal sqrt using rsqrt op, plus one Newton-Rhaphson iteration
 // By Elan Ruskin, http://assemblyrequired.crashworks.org/
-GLM_FUNC_QUALIFIER __m128 glm_sqrt_wip_ss(__m128 x)
+GLM_FUNC_QUALIFIER __m128 glm_f32v1_sqrt_wip(__m128 x)
 {
-	__m128 const recip = _mm_rsqrt_ss(x);  // "estimate" opcode
-	__m128 const halfrecip = _mm_mul_ss(glm_half, recip);
-	__m128 const threeminus_xrr = _mm_sub_ss(glm_three, _mm_mul_ss(x, _mm_mul_ss(recip, recip)));
-	return _mm_mul_ss(halfrecip, threeminus_xrr);
+	__m128 const Rcp0 = _mm_rsqrt_ss(x);  // "estimate" opcode
+	__m128 const Mul0 = _mm_mul_ss(_mm_set1_ps(0.5f), Rcp0);
+	__m128 const Mul1 = _mm_mul_ss(Rcp0, Rcp0);
+	__m128 const Mul2 = _mm_mul_ss(x, Mul1);
+	__m128 const Sub0 = _mm_sub_ss(_mm_set1_ps(3.0f), Mul2);
+	__m128 const Mul3 = _mm_mul_ss(Mul0, Sub0);
+	return Mul3;
 }
+
+// SSE scalar reciprocal sqrt using rsqrt op, plus one Newton-Rhaphson iteration
+// By Elan Ruskin, http://assemblyrequired.crashworks.org/
+GLM_FUNC_QUALIFIER __m128 glm_f32v4_sqrt_wip(__m128 x)
+{
+	__m128 const Rcp0 = _mm_rsqrt_ps(x);  // "estimate" opcode
+	__m128 const Mul0 = _mm_mul_ps(_mm_set1_ps(0.5f), Rcp0);
+	__m128 const Mul1 = _mm_mul_ps(Mul0, Mul0);
+	__m128 const Mul2 = _mm_mul_ps(x, Mul1);
+	__m128 const Sub0 = _mm_sub_ps(_mm_set1_ps(3.0f), Mul2);
+	__m128 const Mul3 = _mm_mul_ps(Mul0, Sub0);
+	return Mul3;
+}
+

+ 18 - 18
glm/simd/geometric.h

@@ -5,7 +5,7 @@
 
 #include "common.h"
 
-GLM_FUNC_QUALIFIER __m128 glm_dot_ps(__m128 v1, __m128 v2)
+GLM_FUNC_QUALIFIER __m128 glm_f32v4_dot(__m128 v1, __m128 v2)
 {
 #	if GLM_ARCH & GLM_ARCH_AVX
 		return _mm_dp_ps(v1, v2, 0xff);
@@ -24,7 +24,7 @@ GLM_FUNC_QUALIFIER __m128 glm_dot_ps(__m128 v1, __m128 v2)
 #	endif
 }
 
-GLM_FUNC_QUALIFIER __m128 glm_dot_ss(__m128 v1, __m128 v2)
+GLM_FUNC_QUALIFIER __m128 glm_f32v1_dot(__m128 v1, __m128 v2)
 {
 #	if GLM_ARCH & GLM_ARCH_AVX
 		return _mm_dp_ps(v1, v2, 0xff);
@@ -43,21 +43,21 @@ GLM_FUNC_QUALIFIER __m128 glm_dot_ss(__m128 v1, __m128 v2)
 #	endif
 }
 
-GLM_FUNC_QUALIFIER __m128 glm_len_ps(__m128 x)
+GLM_FUNC_QUALIFIER __m128 glm_f32v4_len(__m128 x)
 {
-	__m128 const dot0 = glm_dot_ps(x, x);
+	__m128 const dot0 = glm_f32v4_dot(x, x);
 	__m128 const sqt0 = _mm_sqrt_ps(dot0);
 	return sqt0;
 }
 
-GLM_FUNC_QUALIFIER __m128 glm_dst_ps(__m128 p0, __m128 p1)
+GLM_FUNC_QUALIFIER __m128 glm_f32v4_dst(__m128 p0, __m128 p1)
 {
 	__m128 sub0 = _mm_sub_ps(p0, p1);
-	__m128 len0 = glm_len_ps(sub0);
+	__m128 len0 = glm_f32v4_len(sub0);
 	return len0;
 }
 
-GLM_FUNC_QUALIFIER __m128 glm_xpd_ps(__m128 v1, __m128 v2)
+GLM_FUNC_QUALIFIER __m128 glm_f32v4_xpd(__m128 v1, __m128 v2)
 {
 	__m128 swp0 = _mm_shuffle_ps(v1, v1, _MM_SHUFFLE(3, 0, 2, 1));
 	__m128 swp1 = _mm_shuffle_ps(v1, v1, _MM_SHUFFLE(3, 1, 0, 2));
@@ -69,43 +69,43 @@ GLM_FUNC_QUALIFIER __m128 glm_xpd_ps(__m128 v1, __m128 v2)
 	return sub0;
 }
 
-GLM_FUNC_QUALIFIER __m128 glm_nrm_ps(__m128 v)
+GLM_FUNC_QUALIFIER __m128 glm_f32v4_nrm(__m128 v)
 {
-	__m128 dot0 = glm_dot_ps(v, v);
+	__m128 dot0 = glm_f32v4_dot(v, v);
 	__m128 isr0 = _mm_rsqrt_ps(dot0);
 	__m128 mul0 = _mm_mul_ps(v, isr0);
 	return mul0;
 }
 
-GLM_FUNC_QUALIFIER __m128 glm_ffd_ps(__m128 N, __m128 I, __m128 Nref)
+GLM_FUNC_QUALIFIER __m128 glm_f32v4_ffd(__m128 N, __m128 I, __m128 Nref)
 {
-	__m128 dot0 = glm_dot_ps(Nref, I);
-	__m128 sgn0 = glm_sgn_ps(dot0);
+	__m128 dot0 = glm_f32v4_dot(Nref, I);
+	__m128 sgn0 = glm_f32v4_sgn(dot0);
 	__m128 mul0 = _mm_mul_ps(sgn0, glm_minus_one);
 	__m128 mul1 = _mm_mul_ps(N, mul0);
 	return mul1;
 }
 
-GLM_FUNC_QUALIFIER __m128 glm_rfe_ps(__m128 I, __m128 N)
+GLM_FUNC_QUALIFIER __m128 glm_f32v4_rfe(__m128 I, __m128 N)
 {
-	__m128 dot0 = glm_dot_ps(N, I);
+	__m128 dot0 = glm_f32v4_dot(N, I);
 	__m128 mul0 = _mm_mul_ps(N, dot0);
 	__m128 mul1 = _mm_mul_ps(mul0, glm_two);
 	__m128 sub0 = _mm_sub_ps(I, mul1);
 	return sub0;
 }
 
-GLM_FUNC_QUALIFIER __m128 glm_rfa_ps(__m128 I, __m128 N, __m128 eta)
+GLM_FUNC_QUALIFIER __m128 glm_f32v4_rfa(__m128 I, __m128 N, __m128 eta)
 {
-	__m128 dot0 = glm_dot_ps(N, I);
+	__m128 dot0 = glm_f32v4_dot(N, I);
 	__m128 mul0 = _mm_mul_ps(eta, eta);
 	__m128 mul1 = _mm_mul_ps(dot0, dot0);
 	__m128 sub0 = _mm_sub_ps(glm_one, mul0);
 	__m128 sub1 = _mm_sub_ps(glm_one, mul1);
 	__m128 mul2 = _mm_mul_ps(sub0, sub1);
 	
-	if(_mm_movemask_ps(_mm_cmplt_ss(mul2, glm_zero)) == 0)
-		return glm_zero;
+	if(_mm_movemask_ps(_mm_cmplt_ss(mul2, _mm_set1_ps(0.0f))) == 0)
+		return _mm_set1_ps(0.0f);
 
 	__m128 sqt0 = _mm_sqrt_ps(mul2);
 	__m128 mul3 = _mm_mul_ps(eta, dot0);

+ 19 - 19
glm/simd/matrix.h

@@ -9,7 +9,7 @@ static const __m128 GLM_VAR_USED _m128_rad_ps = _mm_set_ps1(3.141592653589793238
 static const __m128 GLM_VAR_USED _m128_deg_ps = _mm_set_ps1(180.f / 3.141592653589793238462643383279f);
 
 template <typename matType>
-GLM_FUNC_QUALIFIER matType glm_comp_mul_ps
+GLM_FUNC_QUALIFIER matType glm_comp_mul_f32m4
 (
 	__m128 const in1[4],
 	__m128 const in2[4],
@@ -22,7 +22,7 @@ GLM_FUNC_QUALIFIER matType glm_comp_mul_ps
 	out[3] = _mm_mul_ps(in1[3], in2[3]);
 }
 
-GLM_FUNC_QUALIFIER void glm_add_ps(__m128 const in1[4], __m128 const in2[4], __m128 out[4])
+GLM_FUNC_QUALIFIER void glm_add_f32m4(__m128 const in1[4], __m128 const in2[4], __m128 out[4])
 {
 	{
 		out[0] = _mm_add_ps(in1[0], in2[0]);
@@ -32,7 +32,7 @@ GLM_FUNC_QUALIFIER void glm_add_ps(__m128 const in1[4], __m128 const in2[4], __m
 	}
 }
 
-GLM_FUNC_QUALIFIER void glm_sub_ps(__m128 const in1[4], __m128 const in2[4], __m128 out[4])
+GLM_FUNC_QUALIFIER void glm_sub_f32v4(__m128 const in1[4], __m128 const in2[4], __m128 out[4])
 {
 	{
 		out[0] = _mm_sub_ps(in1[0], in2[0]);
@@ -42,7 +42,7 @@ GLM_FUNC_QUALIFIER void glm_sub_ps(__m128 const in1[4], __m128 const in2[4], __m
 	}
 }
 
-GLM_FUNC_QUALIFIER __m128 glm_mul_ps(__m128 const m[4], __m128 v)
+GLM_FUNC_QUALIFIER __m128 glm_mul_f32v4(__m128 const m[4], __m128 v)
 {
 	__m128 v0 = _mm_shuffle_ps(v, v, _MM_SHUFFLE(0, 0, 0, 0));
 	__m128 v1 = _mm_shuffle_ps(v, v, _MM_SHUFFLE(1, 1, 1, 1));
@@ -61,7 +61,7 @@ GLM_FUNC_QUALIFIER __m128 glm_mul_ps(__m128 const m[4], __m128 v)
 	return a2;
 }
 
-GLM_FUNC_QUALIFIER __m128 glm_mul_ps(__m128 v, __m128 const m[4])
+GLM_FUNC_QUALIFIER __m128 glm_mul_f32v4(__m128 v, __m128 const m[4])
 {
 	__m128 i0 = m[0];
 	__m128 i1 = m[1];
@@ -88,7 +88,7 @@ GLM_FUNC_QUALIFIER __m128 glm_mul_ps(__m128 v, __m128 const m[4])
 	return f2;
 }
 
-GLM_FUNC_QUALIFIER void glm_mul_ps(__m128 const in1[4], __m128 const in2[4], __m128 out[4])
+GLM_FUNC_QUALIFIER void glm_mul_f32v4(__m128 const in1[4], __m128 const in2[4], __m128 out[4])
 {
 	{
 		__m128 e0 = _mm_shuffle_ps(in2[0], in2[0], _MM_SHUFFLE(0, 0, 0, 0));
@@ -164,7 +164,7 @@ GLM_FUNC_QUALIFIER void glm_mul_ps(__m128 const in1[4], __m128 const in2[4], __m
 	}
 }
 
-GLM_FUNC_QUALIFIER void glm_transpose_ps(__m128 const in[4], __m128 out[4])
+GLM_FUNC_QUALIFIER void glm_transpose_f32m4(__m128 const in[4], __m128 out[4])
 {
 	__m128 tmp0 = _mm_shuffle_ps(in[0], in[1], 0x44);
 	__m128 tmp2 = _mm_shuffle_ps(in[0], in[1], 0xEE);
@@ -177,7 +177,7 @@ GLM_FUNC_QUALIFIER void glm_transpose_ps(__m128 const in[4], __m128 out[4])
 	out[3] = _mm_shuffle_ps(tmp2, tmp3, 0xDD);
 }
 
-GLM_FUNC_QUALIFIER __m128 glm_slow_det_ps(__m128 const in[4])
+GLM_FUNC_QUALIFIER __m128 glm_det_highp_f32m4(__m128 const in[4])
 {
 	__m128 Fac0;
 	{
@@ -387,11 +387,11 @@ GLM_FUNC_QUALIFIER __m128 glm_slow_det_ps(__m128 const in[4])
 	//						+ m[0][1] * Inverse[1][0]
 	//						+ m[0][2] * Inverse[2][0]
 	//						+ m[0][3] * Inverse[3][0];
-	__m128 Det0 = glm_dot_ps(in[0], Row2);
+	__m128 Det0 = glm_f32v4_dot(in[0], Row2);
 	return Det0;
 }
 
-GLM_FUNC_QUALIFIER __m128 glm_detd_ps(__m128 const m[4])
+GLM_FUNC_QUALIFIER __m128 glm_detd_f32m4(__m128 const m[4])
 {
 	// _mm_castsi128_ps(_mm_shuffle_epi32(_mm_castps_si128(
 
@@ -451,10 +451,10 @@ GLM_FUNC_QUALIFIER __m128 glm_detd_ps(__m128 const m[4])
 	//	 + m[0][2] * DetCof[2]
 	//	 + m[0][3] * DetCof[3];
 
-	return glm_dot_ps(m[0], DetCof);
+	return glm_f32v4_dot(m[0], DetCof);
 }
 
-GLM_FUNC_QUALIFIER __m128 glm_det_ps(__m128 const m[4])
+GLM_FUNC_QUALIFIER __m128 glm_det_f32m4(__m128 const m[4])
 {
 	// _mm_castsi128_ps(_mm_shuffle_epi32(_mm_castps_si128(add)
 
@@ -514,10 +514,10 @@ GLM_FUNC_QUALIFIER __m128 glm_det_ps(__m128 const m[4])
 	//	 + m[0][2] * DetCof[2]
 	//	 + m[0][3] * DetCof[3];
 
-	return glm_dot_ps(m[0], DetCof);
+	return glm_f32v4_dot(m[0], DetCof);
 }
 
-GLM_FUNC_QUALIFIER void glm_inverse_ps(__m128 const in[4], __m128 out[4])
+GLM_FUNC_QUALIFIER void glm_f32m4_inverse(__m128 const in[4], __m128 out[4])
 {
 	__m128 Fac0;
 	{
@@ -727,7 +727,7 @@ GLM_FUNC_QUALIFIER void glm_inverse_ps(__m128 const in[4], __m128 out[4])
 	//						+ m[0][1] * Inverse[1][0] 
 	//						+ m[0][2] * Inverse[2][0] 
 	//						+ m[0][3] * Inverse[3][0];
-	__m128 Det0 = glm_dot_ps(in[0], Row2);
+	__m128 Det0 = glm_f32v4_dot(in[0], Row2);
 	__m128 Rcp0 = _mm_div_ps(glm_one, Det0);
 	//__m128 Rcp0 = _mm_rcp_ps(Det0);
 
@@ -738,7 +738,7 @@ GLM_FUNC_QUALIFIER void glm_inverse_ps(__m128 const in[4], __m128 out[4])
 	out[3] = _mm_mul_ps(Inv3, Rcp0);
 }
 
-GLM_FUNC_QUALIFIER void inverse_fast_ps(__m128 const in[4], __m128 out[4])
+GLM_FUNC_QUALIFIER void glm_lowp_f32v4_inverse(__m128 const in[4], __m128 out[4])
 {
 	__m128 Fac0;
 	{
@@ -948,7 +948,7 @@ GLM_FUNC_QUALIFIER void inverse_fast_ps(__m128 const in[4], __m128 out[4])
 	//						+ m[0][1] * Inverse[1][0] 
 	//						+ m[0][2] * Inverse[2][0] 
 	//						+ m[0][3] * Inverse[3][0];
-	__m128 Det0 = glm_dot_ps(in[0], Row2);
+	__m128 Det0 = glm_f32v4_dot(in[0], Row2);
 	__m128 Rcp0 = _mm_rcp_ps(Det0);
 	//__m128 Rcp0 = _mm_div_ps(one, Det0);
 	//	Inverse /= Determinant;
@@ -958,7 +958,7 @@ GLM_FUNC_QUALIFIER void inverse_fast_ps(__m128 const in[4], __m128 out[4])
 	out[3] = _mm_mul_ps(Inv3, Rcp0);
 }
 /*
-GLM_FUNC_QUALIFIER void glm_rotate_ps(__m128 const in[4], float Angle, float const v[3], __m128 out[4])
+GLM_FUNC_QUALIFIER void glm_f32m4_rotate(__m128 const in[4], float Angle, float const v[3], __m128 out[4])
 {
 	float a = glm::radians(Angle);
 	float c = cos(a);
@@ -1028,7 +1028,7 @@ GLM_FUNC_QUALIFIER void glm_rotate_ps(__m128 const in[4], float Angle, float con
 	sse_mul_ps(in, Result, out);
 }
 */
-GLM_FUNC_QUALIFIER void glm_outer_ps(__m128 const & c, __m128 const & r, __m128 out[4])
+GLM_FUNC_QUALIFIER void glm_f32m4_outer(__m128 const & c, __m128 const & r, __m128 out[4])
 {
 	out[0] = _mm_mul_ps(c, _mm_shuffle_ps(r, r, _MM_SHUFFLE(0, 0, 0, 0)));
 	out[1] = _mm_mul_ps(c, _mm_shuffle_ps(r, r, _MM_SHUFFLE(1, 1, 1, 1)));