// // Copyright (c) 2008-2017 the Urho3D project. // // Permission is hereby granted, free of charge, to any person obtaining a copy // of this software and associated documentation files (the "Software"), to deal // in the Software without restriction, including without limitation the rights // to use, copy, modify, merge, publish, distribute, sublicense, and/or sell // copies of the Software, and to permit persons to whom the Software is // furnished to do so, subject to the following conditions: // // The above copyright notice and this permission notice shall be included in // all copies or substantial portions of the Software. // // THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR // IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, // FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE // AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER // LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, // OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN // THE SOFTWARE. // #pragma once #include "../Math/Quaternion.h" #include "../Math/Vector4.h" #ifdef ATOMIC_SSE #include #endif namespace Atomic { class Matrix3x4; /// 4x4 matrix for arbitrary linear transforms including projection. class ATOMIC_API Matrix4 { public: /// Construct an identity matrix. Matrix4() #ifndef ATOMIC_SSE :m00_(1.0f), m01_(0.0f), m02_(0.0f), m03_(0.0f), m10_(0.0f), m11_(1.0f), m12_(0.0f), m13_(0.0f), m20_(0.0f), m21_(0.0f), m22_(1.0f), m23_(0.0f), m30_(0.0f), m31_(0.0f), m32_(0.0f), m33_(1.0f) #endif { #ifdef ATOMIC_SSE _mm_storeu_ps(&m00_, _mm_set_ps(0.f, 0.f, 0.f, 1.f)); _mm_storeu_ps(&m10_, _mm_set_ps(0.f, 0.f, 1.f, 0.f)); _mm_storeu_ps(&m20_, _mm_set_ps(0.f, 1.f, 0.f, 0.f)); _mm_storeu_ps(&m30_, _mm_set_ps(1.f, 0.f, 0.f, 0.f)); #endif } /// Copy-construct from another matrix. Matrix4(const Matrix4& matrix) #ifndef ATOMIC_SSE :m00_(matrix.m00_), m01_(matrix.m01_), m02_(matrix.m02_), m03_(matrix.m03_), m10_(matrix.m10_), m11_(matrix.m11_), m12_(matrix.m12_), m13_(matrix.m13_), m20_(matrix.m20_), m21_(matrix.m21_), m22_(matrix.m22_), m23_(matrix.m23_), m30_(matrix.m30_), m31_(matrix.m31_), m32_(matrix.m32_), m33_(matrix.m33_) #endif { #ifdef ATOMIC_SSE _mm_storeu_ps(&m00_, _mm_loadu_ps(&matrix.m00_)); _mm_storeu_ps(&m10_, _mm_loadu_ps(&matrix.m10_)); _mm_storeu_ps(&m20_, _mm_loadu_ps(&matrix.m20_)); _mm_storeu_ps(&m30_, _mm_loadu_ps(&matrix.m30_)); #endif } /// Copy-construct from a 3x3 matrix and set the extra elements to identity. Matrix4(const Matrix3& matrix) : m00_(matrix.m00_), m01_(matrix.m01_), m02_(matrix.m02_), m03_(0.0f), m10_(matrix.m10_), m11_(matrix.m11_), m12_(matrix.m12_), m13_(0.0f), m20_(matrix.m20_), m21_(matrix.m21_), m22_(matrix.m22_), m23_(0.0f), m30_(0.0f), m31_(0.0f), m32_(0.0f), m33_(1.0f) { } /// Construct from values. Matrix4(float v00, float v01, float v02, float v03, float v10, float v11, float v12, float v13, float v20, float v21, float v22, float v23, float v30, float v31, float v32, float v33) : m00_(v00), m01_(v01), m02_(v02), m03_(v03), m10_(v10), m11_(v11), m12_(v12), m13_(v13), m20_(v20), m21_(v21), m22_(v22), m23_(v23), m30_(v30), m31_(v31), m32_(v32), m33_(v33) { } /// Construct from a float array. explicit Matrix4(const float* data) #ifndef ATOMIC_SSE :m00_(data[0]), m01_(data[1]), m02_(data[2]), m03_(data[3]), m10_(data[4]), m11_(data[5]), m12_(data[6]), m13_(data[7]), m20_(data[8]), m21_(data[9]), m22_(data[10]), m23_(data[11]), m30_(data[12]), m31_(data[13]), m32_(data[14]), m33_(data[15]) #endif { #ifdef ATOMIC_SSE _mm_storeu_ps(&m00_, _mm_loadu_ps(data)); _mm_storeu_ps(&m10_, _mm_loadu_ps(data + 4)); _mm_storeu_ps(&m20_, _mm_loadu_ps(data + 8)); _mm_storeu_ps(&m30_, _mm_loadu_ps(data + 12)); #endif } /// Assign from another matrix. Matrix4& operator =(const Matrix4& rhs) { #ifdef ATOMIC_SSE _mm_storeu_ps(&m00_, _mm_loadu_ps(&rhs.m00_)); _mm_storeu_ps(&m10_, _mm_loadu_ps(&rhs.m10_)); _mm_storeu_ps(&m20_, _mm_loadu_ps(&rhs.m20_)); _mm_storeu_ps(&m30_, _mm_loadu_ps(&rhs.m30_)); #else m00_ = rhs.m00_; m01_ = rhs.m01_; m02_ = rhs.m02_; m03_ = rhs.m03_; m10_ = rhs.m10_; m11_ = rhs.m11_; m12_ = rhs.m12_; m13_ = rhs.m13_; m20_ = rhs.m20_; m21_ = rhs.m21_; m22_ = rhs.m22_; m23_ = rhs.m23_; m30_ = rhs.m30_; m31_ = rhs.m31_; m32_ = rhs.m32_; m33_ = rhs.m33_; #endif return *this; } /// Assign from a 3x3 matrix. Set the extra elements to identity. Matrix4& operator =(const Matrix3& rhs) { m00_ = rhs.m00_; m01_ = rhs.m01_; m02_ = rhs.m02_; m03_ = 0.0f; m10_ = rhs.m10_; m11_ = rhs.m11_; m12_ = rhs.m12_; m13_ = 0.0f; m20_ = rhs.m20_; m21_ = rhs.m21_; m22_ = rhs.m22_; m23_ = 0.0f; m30_ = 0.0f; m31_ = 0.0f; m32_ = 0.0f; m33_ = 1.0f; return *this; } /// Test for equality with another matrix without epsilon. bool operator ==(const Matrix4& rhs) const { #ifdef ATOMIC_SSE __m128 c0 = _mm_cmpeq_ps(_mm_loadu_ps(&m00_), _mm_loadu_ps(&rhs.m00_)); __m128 c1 = _mm_cmpeq_ps(_mm_loadu_ps(&m10_), _mm_loadu_ps(&rhs.m10_)); c0 = _mm_and_ps(c0, c1); __m128 c2 = _mm_cmpeq_ps(_mm_loadu_ps(&m20_), _mm_loadu_ps(&rhs.m20_)); __m128 c3 = _mm_cmpeq_ps(_mm_loadu_ps(&m30_), _mm_loadu_ps(&rhs.m30_)); c2 = _mm_and_ps(c2, c3); c0 = _mm_and_ps(c0, c2); __m128 hi = _mm_movehl_ps(c0, c0); c0 = _mm_and_ps(c0, hi); hi = _mm_shuffle_ps(c0, c0, _MM_SHUFFLE(1, 1, 1, 1)); c0 = _mm_and_ps(c0, hi); return _mm_cvtsi128_si32(_mm_castps_si128(c0)) == -1; #else const float* leftData = Data(); const float* rightData = rhs.Data(); for (unsigned i = 0; i < 16; ++i) { if (leftData[i] != rightData[i]) return false; } return true; #endif } /// Test for inequality with another matrix without epsilon. bool operator !=(const Matrix4& rhs) const { return !(*this == rhs); } /// Multiply a Vector3 which is assumed to represent position. Vector3 operator *(const Vector3& rhs) const { #ifdef ATOMIC_SSE __m128 vec = _mm_set_ps(1.f, rhs.z_, rhs.y_, rhs.x_); __m128 r0 = _mm_mul_ps(_mm_loadu_ps(&m00_), vec); __m128 r1 = _mm_mul_ps(_mm_loadu_ps(&m10_), vec); __m128 t0 = _mm_unpacklo_ps(r0, r1); __m128 t1 = _mm_unpackhi_ps(r0, r1); t0 = _mm_add_ps(t0, t1); __m128 r2 = _mm_mul_ps(_mm_loadu_ps(&m20_), vec); __m128 r3 = _mm_mul_ps(_mm_loadu_ps(&m30_), vec); __m128 t2 = _mm_unpacklo_ps(r2, r3); __m128 t3 = _mm_unpackhi_ps(r2, r3); t2 = _mm_add_ps(t2, t3); vec = _mm_add_ps(_mm_movelh_ps(t0, t2), _mm_movehl_ps(t2, t0)); vec = _mm_div_ps(vec, _mm_shuffle_ps(vec, vec, _MM_SHUFFLE(3, 3, 3, 3))); return Vector3( _mm_cvtss_f32(vec), _mm_cvtss_f32(_mm_shuffle_ps(vec, vec, _MM_SHUFFLE(1, 1, 1, 1))), _mm_cvtss_f32(_mm_movehl_ps(vec, vec))); #else float invW = 1.0f / (m30_ * rhs.x_ + m31_ * rhs.y_ + m32_ * rhs.z_ + m33_); return Vector3( (m00_ * rhs.x_ + m01_ * rhs.y_ + m02_ * rhs.z_ + m03_) * invW, (m10_ * rhs.x_ + m11_ * rhs.y_ + m12_ * rhs.z_ + m13_) * invW, (m20_ * rhs.x_ + m21_ * rhs.y_ + m22_ * rhs.z_ + m23_) * invW ); #endif } /// Multiply a Vector4. Vector4 operator *(const Vector4& rhs) const { #ifdef ATOMIC_SSE __m128 vec = _mm_loadu_ps(&rhs.x_); __m128 r0 = _mm_mul_ps(_mm_loadu_ps(&m00_), vec); __m128 r1 = _mm_mul_ps(_mm_loadu_ps(&m10_), vec); __m128 t0 = _mm_unpacklo_ps(r0, r1); __m128 t1 = _mm_unpackhi_ps(r0, r1); t0 = _mm_add_ps(t0, t1); __m128 r2 = _mm_mul_ps(_mm_loadu_ps(&m20_), vec); __m128 r3 = _mm_mul_ps(_mm_loadu_ps(&m30_), vec); __m128 t2 = _mm_unpacklo_ps(r2, r3); __m128 t3 = _mm_unpackhi_ps(r2, r3); t2 = _mm_add_ps(t2, t3); vec = _mm_add_ps(_mm_movelh_ps(t0, t2), _mm_movehl_ps(t2, t0)); Vector4 ret; _mm_storeu_ps(&ret.x_, vec); return ret; #else return Vector4( m00_ * rhs.x_ + m01_ * rhs.y_ + m02_ * rhs.z_ + m03_ * rhs.w_, m10_ * rhs.x_ + m11_ * rhs.y_ + m12_ * rhs.z_ + m13_ * rhs.w_, m20_ * rhs.x_ + m21_ * rhs.y_ + m22_ * rhs.z_ + m23_ * rhs.w_, m30_ * rhs.x_ + m31_ * rhs.y_ + m32_ * rhs.z_ + m33_ * rhs.w_ ); #endif } /// Add a matrix. Matrix4 operator +(const Matrix4& rhs) const { #ifdef ATOMIC_SSE Matrix4 ret; _mm_storeu_ps(&ret.m00_, _mm_add_ps(_mm_loadu_ps(&m00_), _mm_loadu_ps(&rhs.m00_))); _mm_storeu_ps(&ret.m10_, _mm_add_ps(_mm_loadu_ps(&m10_), _mm_loadu_ps(&rhs.m10_))); _mm_storeu_ps(&ret.m20_, _mm_add_ps(_mm_loadu_ps(&m20_), _mm_loadu_ps(&rhs.m20_))); _mm_storeu_ps(&ret.m30_, _mm_add_ps(_mm_loadu_ps(&m30_), _mm_loadu_ps(&rhs.m30_))); return ret; #else return Matrix4( m00_ + rhs.m00_, m01_ + rhs.m01_, m02_ + rhs.m02_, m03_ + rhs.m03_, m10_ + rhs.m10_, m11_ + rhs.m11_, m12_ + rhs.m12_, m13_ + rhs.m13_, m20_ + rhs.m20_, m21_ + rhs.m21_, m22_ + rhs.m22_, m23_ + rhs.m23_, m30_ + rhs.m30_, m31_ + rhs.m31_, m32_ + rhs.m32_, m33_ + rhs.m33_ ); #endif } /// Subtract a matrix. Matrix4 operator -(const Matrix4& rhs) const { #ifdef ATOMIC_SSE Matrix4 ret; _mm_storeu_ps(&ret.m00_, _mm_sub_ps(_mm_loadu_ps(&m00_), _mm_loadu_ps(&rhs.m00_))); _mm_storeu_ps(&ret.m10_, _mm_sub_ps(_mm_loadu_ps(&m10_), _mm_loadu_ps(&rhs.m10_))); _mm_storeu_ps(&ret.m20_, _mm_sub_ps(_mm_loadu_ps(&m20_), _mm_loadu_ps(&rhs.m20_))); _mm_storeu_ps(&ret.m30_, _mm_sub_ps(_mm_loadu_ps(&m30_), _mm_loadu_ps(&rhs.m30_))); return ret; #else return Matrix4( m00_ - rhs.m00_, m01_ - rhs.m01_, m02_ - rhs.m02_, m03_ - rhs.m03_, m10_ - rhs.m10_, m11_ - rhs.m11_, m12_ - rhs.m12_, m13_ - rhs.m13_, m20_ - rhs.m20_, m21_ - rhs.m21_, m22_ - rhs.m22_, m23_ - rhs.m23_, m30_ - rhs.m30_, m31_ - rhs.m31_, m32_ - rhs.m32_, m33_ - rhs.m33_ ); #endif } /// Multiply with a scalar. Matrix4 operator *(float rhs) const { #ifdef ATOMIC_SSE Matrix4 ret; const __m128 mul = _mm_set1_ps(rhs); _mm_storeu_ps(&ret.m00_, _mm_mul_ps(_mm_loadu_ps(&m00_), mul)); _mm_storeu_ps(&ret.m10_, _mm_mul_ps(_mm_loadu_ps(&m10_), mul)); _mm_storeu_ps(&ret.m20_, _mm_mul_ps(_mm_loadu_ps(&m20_), mul)); _mm_storeu_ps(&ret.m30_, _mm_mul_ps(_mm_loadu_ps(&m30_), mul)); return ret; #else return Matrix4( m00_ * rhs, m01_ * rhs, m02_ * rhs, m03_ * rhs, m10_ * rhs, m11_ * rhs, m12_ * rhs, m13_ * rhs, m20_ * rhs, m21_ * rhs, m22_ * rhs, m23_ * rhs, m30_ * rhs, m31_ * rhs, m32_ * rhs, m33_ * rhs ); #endif } /// Multiply a matrix. Matrix4 operator *(const Matrix4& rhs) const { #ifdef ATOMIC_SSE Matrix4 out; __m128 r0 = _mm_loadu_ps(&rhs.m00_); __m128 r1 = _mm_loadu_ps(&rhs.m10_); __m128 r2 = _mm_loadu_ps(&rhs.m20_); __m128 r3 = _mm_loadu_ps(&rhs.m30_); __m128 l = _mm_loadu_ps(&m00_); __m128 t0 = _mm_mul_ps(_mm_shuffle_ps(l, l, _MM_SHUFFLE(0, 0, 0, 0)), r0); __m128 t1 = _mm_mul_ps(_mm_shuffle_ps(l, l, _MM_SHUFFLE(1, 1, 1, 1)), r1); __m128 t2 = _mm_mul_ps(_mm_shuffle_ps(l, l, _MM_SHUFFLE(2, 2, 2, 2)), r2); __m128 t3 = _mm_mul_ps(_mm_shuffle_ps(l, l, _MM_SHUFFLE(3, 3, 3, 3)), r3); _mm_storeu_ps(&out.m00_, _mm_add_ps(_mm_add_ps(t0, t1), _mm_add_ps(t2, t3))); l = _mm_loadu_ps(&m10_); t0 = _mm_mul_ps(_mm_shuffle_ps(l, l, _MM_SHUFFLE(0, 0, 0, 0)), r0); t1 = _mm_mul_ps(_mm_shuffle_ps(l, l, _MM_SHUFFLE(1, 1, 1, 1)), r1); t2 = _mm_mul_ps(_mm_shuffle_ps(l, l, _MM_SHUFFLE(2, 2, 2, 2)), r2); t3 = _mm_mul_ps(_mm_shuffle_ps(l, l, _MM_SHUFFLE(3, 3, 3, 3)), r3); _mm_storeu_ps(&out.m10_, _mm_add_ps(_mm_add_ps(t0, t1), _mm_add_ps(t2, t3))); l = _mm_loadu_ps(&m20_); t0 = _mm_mul_ps(_mm_shuffle_ps(l, l, _MM_SHUFFLE(0, 0, 0, 0)), r0); t1 = _mm_mul_ps(_mm_shuffle_ps(l, l, _MM_SHUFFLE(1, 1, 1, 1)), r1); t2 = _mm_mul_ps(_mm_shuffle_ps(l, l, _MM_SHUFFLE(2, 2, 2, 2)), r2); t3 = _mm_mul_ps(_mm_shuffle_ps(l, l, _MM_SHUFFLE(3, 3, 3, 3)), r3); _mm_storeu_ps(&out.m20_, _mm_add_ps(_mm_add_ps(t0, t1), _mm_add_ps(t2, t3))); l = _mm_loadu_ps(&m30_); t0 = _mm_mul_ps(_mm_shuffle_ps(l, l, _MM_SHUFFLE(0, 0, 0, 0)), r0); t1 = _mm_mul_ps(_mm_shuffle_ps(l, l, _MM_SHUFFLE(1, 1, 1, 1)), r1); t2 = _mm_mul_ps(_mm_shuffle_ps(l, l, _MM_SHUFFLE(2, 2, 2, 2)), r2); t3 = _mm_mul_ps(_mm_shuffle_ps(l, l, _MM_SHUFFLE(3, 3, 3, 3)), r3); _mm_storeu_ps(&out.m30_, _mm_add_ps(_mm_add_ps(t0, t1), _mm_add_ps(t2, t3))); return out; #else return Matrix4( m00_ * rhs.m00_ + m01_ * rhs.m10_ + m02_ * rhs.m20_ + m03_ * rhs.m30_, m00_ * rhs.m01_ + m01_ * rhs.m11_ + m02_ * rhs.m21_ + m03_ * rhs.m31_, m00_ * rhs.m02_ + m01_ * rhs.m12_ + m02_ * rhs.m22_ + m03_ * rhs.m32_, m00_ * rhs.m03_ + m01_ * rhs.m13_ + m02_ * rhs.m23_ + m03_ * rhs.m33_, m10_ * rhs.m00_ + m11_ * rhs.m10_ + m12_ * rhs.m20_ + m13_ * rhs.m30_, m10_ * rhs.m01_ + m11_ * rhs.m11_ + m12_ * rhs.m21_ + m13_ * rhs.m31_, m10_ * rhs.m02_ + m11_ * rhs.m12_ + m12_ * rhs.m22_ + m13_ * rhs.m32_, m10_ * rhs.m03_ + m11_ * rhs.m13_ + m12_ * rhs.m23_ + m13_ * rhs.m33_, m20_ * rhs.m00_ + m21_ * rhs.m10_ + m22_ * rhs.m20_ + m23_ * rhs.m30_, m20_ * rhs.m01_ + m21_ * rhs.m11_ + m22_ * rhs.m21_ + m23_ * rhs.m31_, m20_ * rhs.m02_ + m21_ * rhs.m12_ + m22_ * rhs.m22_ + m23_ * rhs.m32_, m20_ * rhs.m03_ + m21_ * rhs.m13_ + m22_ * rhs.m23_ + m23_ * rhs.m33_, m30_ * rhs.m00_ + m31_ * rhs.m10_ + m32_ * rhs.m20_ + m33_ * rhs.m30_, m30_ * rhs.m01_ + m31_ * rhs.m11_ + m32_ * rhs.m21_ + m33_ * rhs.m31_, m30_ * rhs.m02_ + m31_ * rhs.m12_ + m32_ * rhs.m22_ + m33_ * rhs.m32_, m30_ * rhs.m03_ + m31_ * rhs.m13_ + m32_ * rhs.m23_ + m33_ * rhs.m33_ ); #endif } /// Multiply with a 3x4 matrix. Matrix4 operator *(const Matrix3x4& rhs) const; /// Set translation elements. void SetTranslation(const Vector3& translation) { m03_ = translation.x_; m13_ = translation.y_; m23_ = translation.z_; } /// Set rotation elements from a 3x3 matrix. void SetRotation(const Matrix3& rotation) { m00_ = rotation.m00_; m01_ = rotation.m01_; m02_ = rotation.m02_; m10_ = rotation.m10_; m11_ = rotation.m11_; m12_ = rotation.m12_; m20_ = rotation.m20_; m21_ = rotation.m21_; m22_ = rotation.m22_; } /// Set scaling elements. void SetScale(const Vector3& scale) { m00_ = scale.x_; m11_ = scale.y_; m22_ = scale.z_; } /// Set uniform scaling elements. void SetScale(float scale) { m00_ = scale; m11_ = scale; m22_ = scale; } /// Return the combined rotation and scaling matrix. Matrix3 ToMatrix3() const { return Matrix3( m00_, m01_, m02_, m10_, m11_, m12_, m20_, m21_, m22_ ); } /// Return the rotation matrix with scaling removed. Matrix3 RotationMatrix() const { Vector3 invScale( 1.0f / sqrtf(m00_ * m00_ + m10_ * m10_ + m20_ * m20_), 1.0f / sqrtf(m01_ * m01_ + m11_ * m11_ + m21_ * m21_), 1.0f / sqrtf(m02_ * m02_ + m12_ * m12_ + m22_ * m22_) ); return ToMatrix3().Scaled(invScale); } /// Return the translation part. Vector3 Translation() const { return Vector3( m03_, m13_, m23_ ); } /// Return the rotation part. Quaternion Rotation() const { return Quaternion(RotationMatrix()); } /// Return the scaling part. Vector3 Scale() const { return Vector3( sqrtf(m00_ * m00_ + m10_ * m10_ + m20_ * m20_), sqrtf(m01_ * m01_ + m11_ * m11_ + m21_ * m21_), sqrtf(m02_ * m02_ + m12_ * m12_ + m22_ * m22_) ); } /// Return the scaling part with the sign. Reference rotation matrix is required to avoid ambiguity. Vector3 SignedScale(const Matrix3& rotation) const { return Vector3( rotation.m00_ * m00_ + rotation.m10_ * m10_ + rotation.m20_ * m20_, rotation.m01_ * m01_ + rotation.m11_ * m11_ + rotation.m21_ * m21_, rotation.m02_ * m02_ + rotation.m12_ * m12_ + rotation.m22_ * m22_ ); } /// Return transposed. Matrix4 Transpose() const { #ifdef ATOMIC_SSE __m128 m0 = _mm_loadu_ps(&m00_); __m128 m1 = _mm_loadu_ps(&m10_); __m128 m2 = _mm_loadu_ps(&m20_); __m128 m3 = _mm_loadu_ps(&m30_); _MM_TRANSPOSE4_PS(m0, m1, m2, m3); Matrix4 out; _mm_storeu_ps(&out.m00_, m0); _mm_storeu_ps(&out.m10_, m1); _mm_storeu_ps(&out.m20_, m2); _mm_storeu_ps(&out.m30_, m3); return out; #else return Matrix4( m00_, m10_, m20_, m30_, m01_, m11_, m21_, m31_, m02_, m12_, m22_, m32_, m03_, m13_, m23_, m33_ ); #endif } /// Test for equality with another matrix with epsilon. bool Equals(const Matrix4& rhs) const { const float* leftData = Data(); const float* rightData = rhs.Data(); for (unsigned i = 0; i < 16; ++i) { if (!Atomic::Equals(leftData[i], rightData[i])) return false; } return true; } /// Return decomposition to translation, rotation and scale. void Decompose(Vector3& translation, Quaternion& rotation, Vector3& scale) const; /// Return inverse. Matrix4 Inverse() const; /// Return float data. const float* Data() const { return &m00_; } /// Return matrix element. float Element(unsigned i, unsigned j) const { return Data()[i * 4 + j]; } /// Return matrix row. Vector4 Row(unsigned i) const { return Vector4(Element(i, 0), Element(i, 1), Element(i, 2), Element(i, 3)); } /// Return matrix column. Vector4 Column(unsigned j) const { return Vector4(Element(0, j), Element(1, j), Element(2, j), Element(3, j)); } /// Return as string. String ToString() const; float m00_; float m01_; float m02_; float m03_; float m10_; float m11_; float m12_; float m13_; float m20_; float m21_; float m22_; float m23_; float m30_; float m31_; float m32_; float m33_; /// Bulk transpose matrices. static void BulkTranspose(float* dest, const float* src, unsigned count) { for (unsigned i = 0; i < count; ++i) { #ifdef ATOMIC_SSE __m128 m0 = _mm_loadu_ps(src); __m128 m1 = _mm_loadu_ps(src + 4); __m128 m2 = _mm_loadu_ps(src + 8); __m128 m3 = _mm_loadu_ps(src + 12); _MM_TRANSPOSE4_PS(m0, m1, m2, m3); _mm_storeu_ps(dest, m0); _mm_storeu_ps(dest + 4, m1); _mm_storeu_ps(dest + 8, m2); _mm_storeu_ps(dest + 12, m3); #else dest[0] = src[0]; dest[1] = src[4]; dest[2] = src[8]; dest[3] = src[12]; dest[4] = src[1]; dest[5] = src[5]; dest[6] = src[9]; dest[7] = src[13]; dest[8] = src[2]; dest[9] = src[6]; dest[10] = src[10]; dest[11] = src[14]; dest[12] = src[3]; dest[13] = src[7]; dest[14] = src[11]; dest[15] = src[15]; #endif dest += 16; src += 16; } } /// Zero matrix. static const Matrix4 ZERO; /// Identity matrix. static const Matrix4 IDENTITY; }; /// Multiply a 4x4 matrix with a scalar. inline Matrix4 operator *(float lhs, const Matrix4& rhs) { return rhs * lhs; } }