| 123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717 |
- //
- // Copyright (c) 2008-2017 the Urho3D project.
- //
- // Permission is hereby granted, free of charge, to any person obtaining a copy
- // of this software and associated documentation files (the "Software"), to deal
- // in the Software without restriction, including without limitation the rights
- // to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
- // copies of the Software, and to permit persons to whom the Software is
- // furnished to do so, subject to the following conditions:
- //
- // The above copyright notice and this permission notice shall be included in
- // all copies or substantial portions of the Software.
- //
- // THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
- // IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
- // FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
- // AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
- // LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
- // OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
- // THE SOFTWARE.
- //
- #pragma once
- #include "../Math/Quaternion.h"
- #include "../Math/Vector4.h"
- #ifdef ATOMIC_SSE
- #include <emmintrin.h>
- #endif
- namespace Atomic
- {
- class Matrix3x4;
- /// 4x4 matrix for arbitrary linear transforms including projection.
- class ATOMIC_API Matrix4
- {
- public:
- /// Construct an identity matrix.
- Matrix4()
- #ifndef ATOMIC_SSE
- :m00_(1.0f),
- m01_(0.0f),
- m02_(0.0f),
- m03_(0.0f),
- m10_(0.0f),
- m11_(1.0f),
- m12_(0.0f),
- m13_(0.0f),
- m20_(0.0f),
- m21_(0.0f),
- m22_(1.0f),
- m23_(0.0f),
- m30_(0.0f),
- m31_(0.0f),
- m32_(0.0f),
- m33_(1.0f)
- #endif
- {
- #ifdef ATOMIC_SSE
- _mm_storeu_ps(&m00_, _mm_set_ps(0.f, 0.f, 0.f, 1.f));
- _mm_storeu_ps(&m10_, _mm_set_ps(0.f, 0.f, 1.f, 0.f));
- _mm_storeu_ps(&m20_, _mm_set_ps(0.f, 1.f, 0.f, 0.f));
- _mm_storeu_ps(&m30_, _mm_set_ps(1.f, 0.f, 0.f, 0.f));
- #endif
- }
- /// Copy-construct from another matrix.
- Matrix4(const Matrix4& matrix)
- #ifndef ATOMIC_SSE
- :m00_(matrix.m00_),
- m01_(matrix.m01_),
- m02_(matrix.m02_),
- m03_(matrix.m03_),
- m10_(matrix.m10_),
- m11_(matrix.m11_),
- m12_(matrix.m12_),
- m13_(matrix.m13_),
- m20_(matrix.m20_),
- m21_(matrix.m21_),
- m22_(matrix.m22_),
- m23_(matrix.m23_),
- m30_(matrix.m30_),
- m31_(matrix.m31_),
- m32_(matrix.m32_),
- m33_(matrix.m33_)
- #endif
- {
- #ifdef ATOMIC_SSE
- _mm_storeu_ps(&m00_, _mm_loadu_ps(&matrix.m00_));
- _mm_storeu_ps(&m10_, _mm_loadu_ps(&matrix.m10_));
- _mm_storeu_ps(&m20_, _mm_loadu_ps(&matrix.m20_));
- _mm_storeu_ps(&m30_, _mm_loadu_ps(&matrix.m30_));
- #endif
- }
- /// Copy-construct from a 3x3 matrix and set the extra elements to identity.
- Matrix4(const Matrix3& matrix) :
- m00_(matrix.m00_),
- m01_(matrix.m01_),
- m02_(matrix.m02_),
- m03_(0.0f),
- m10_(matrix.m10_),
- m11_(matrix.m11_),
- m12_(matrix.m12_),
- m13_(0.0f),
- m20_(matrix.m20_),
- m21_(matrix.m21_),
- m22_(matrix.m22_),
- m23_(0.0f),
- m30_(0.0f),
- m31_(0.0f),
- m32_(0.0f),
- m33_(1.0f)
- {
- }
- /// Construct from values.
- Matrix4(float v00, float v01, float v02, float v03,
- float v10, float v11, float v12, float v13,
- float v20, float v21, float v22, float v23,
- float v30, float v31, float v32, float v33) :
- m00_(v00),
- m01_(v01),
- m02_(v02),
- m03_(v03),
- m10_(v10),
- m11_(v11),
- m12_(v12),
- m13_(v13),
- m20_(v20),
- m21_(v21),
- m22_(v22),
- m23_(v23),
- m30_(v30),
- m31_(v31),
- m32_(v32),
- m33_(v33)
- {
- }
- /// Construct from a float array.
- explicit Matrix4(const float* data)
- #ifndef ATOMIC_SSE
- :m00_(data[0]),
- m01_(data[1]),
- m02_(data[2]),
- m03_(data[3]),
- m10_(data[4]),
- m11_(data[5]),
- m12_(data[6]),
- m13_(data[7]),
- m20_(data[8]),
- m21_(data[9]),
- m22_(data[10]),
- m23_(data[11]),
- m30_(data[12]),
- m31_(data[13]),
- m32_(data[14]),
- m33_(data[15])
- #endif
- {
- #ifdef ATOMIC_SSE
- _mm_storeu_ps(&m00_, _mm_loadu_ps(data));
- _mm_storeu_ps(&m10_, _mm_loadu_ps(data + 4));
- _mm_storeu_ps(&m20_, _mm_loadu_ps(data + 8));
- _mm_storeu_ps(&m30_, _mm_loadu_ps(data + 12));
- #endif
- }
- /// Assign from another matrix.
- Matrix4& operator =(const Matrix4& rhs)
- {
- #ifdef ATOMIC_SSE
- _mm_storeu_ps(&m00_, _mm_loadu_ps(&rhs.m00_));
- _mm_storeu_ps(&m10_, _mm_loadu_ps(&rhs.m10_));
- _mm_storeu_ps(&m20_, _mm_loadu_ps(&rhs.m20_));
- _mm_storeu_ps(&m30_, _mm_loadu_ps(&rhs.m30_));
- #else
- m00_ = rhs.m00_;
- m01_ = rhs.m01_;
- m02_ = rhs.m02_;
- m03_ = rhs.m03_;
- m10_ = rhs.m10_;
- m11_ = rhs.m11_;
- m12_ = rhs.m12_;
- m13_ = rhs.m13_;
- m20_ = rhs.m20_;
- m21_ = rhs.m21_;
- m22_ = rhs.m22_;
- m23_ = rhs.m23_;
- m30_ = rhs.m30_;
- m31_ = rhs.m31_;
- m32_ = rhs.m32_;
- m33_ = rhs.m33_;
- #endif
- return *this;
- }
- /// Assign from a 3x3 matrix. Set the extra elements to identity.
- Matrix4& operator =(const Matrix3& rhs)
- {
- m00_ = rhs.m00_;
- m01_ = rhs.m01_;
- m02_ = rhs.m02_;
- m03_ = 0.0f;
- m10_ = rhs.m10_;
- m11_ = rhs.m11_;
- m12_ = rhs.m12_;
- m13_ = 0.0f;
- m20_ = rhs.m20_;
- m21_ = rhs.m21_;
- m22_ = rhs.m22_;
- m23_ = 0.0f;
- m30_ = 0.0f;
- m31_ = 0.0f;
- m32_ = 0.0f;
- m33_ = 1.0f;
- return *this;
- }
- /// Test for equality with another matrix without epsilon.
- bool operator ==(const Matrix4& rhs) const
- {
- #ifdef ATOMIC_SSE
- __m128 c0 = _mm_cmpeq_ps(_mm_loadu_ps(&m00_), _mm_loadu_ps(&rhs.m00_));
- __m128 c1 = _mm_cmpeq_ps(_mm_loadu_ps(&m10_), _mm_loadu_ps(&rhs.m10_));
- c0 = _mm_and_ps(c0, c1);
- __m128 c2 = _mm_cmpeq_ps(_mm_loadu_ps(&m20_), _mm_loadu_ps(&rhs.m20_));
- __m128 c3 = _mm_cmpeq_ps(_mm_loadu_ps(&m30_), _mm_loadu_ps(&rhs.m30_));
- c2 = _mm_and_ps(c2, c3);
- c0 = _mm_and_ps(c0, c2);
- __m128 hi = _mm_movehl_ps(c0, c0);
- c0 = _mm_and_ps(c0, hi);
- hi = _mm_shuffle_ps(c0, c0, _MM_SHUFFLE(1, 1, 1, 1));
- c0 = _mm_and_ps(c0, hi);
- return _mm_cvtsi128_si32(_mm_castps_si128(c0)) == -1;
- #else
- const float* leftData = Data();
- const float* rightData = rhs.Data();
- for (unsigned i = 0; i < 16; ++i)
- {
- if (leftData[i] != rightData[i])
- return false;
- }
- return true;
- #endif
- }
- /// Test for inequality with another matrix without epsilon.
- bool operator !=(const Matrix4& rhs) const { return !(*this == rhs); }
- /// Multiply a Vector3 which is assumed to represent position.
- Vector3 operator *(const Vector3& rhs) const
- {
- #ifdef ATOMIC_SSE
- __m128 vec = _mm_set_ps(1.f, rhs.z_, rhs.y_, rhs.x_);
- __m128 r0 = _mm_mul_ps(_mm_loadu_ps(&m00_), vec);
- __m128 r1 = _mm_mul_ps(_mm_loadu_ps(&m10_), vec);
- __m128 t0 = _mm_unpacklo_ps(r0, r1);
- __m128 t1 = _mm_unpackhi_ps(r0, r1);
- t0 = _mm_add_ps(t0, t1);
- __m128 r2 = _mm_mul_ps(_mm_loadu_ps(&m20_), vec);
- __m128 r3 = _mm_mul_ps(_mm_loadu_ps(&m30_), vec);
- __m128 t2 = _mm_unpacklo_ps(r2, r3);
- __m128 t3 = _mm_unpackhi_ps(r2, r3);
- t2 = _mm_add_ps(t2, t3);
- vec = _mm_add_ps(_mm_movelh_ps(t0, t2), _mm_movehl_ps(t2, t0));
- vec = _mm_div_ps(vec, _mm_shuffle_ps(vec, vec, _MM_SHUFFLE(3, 3, 3, 3)));
- return Vector3(
- _mm_cvtss_f32(vec),
- _mm_cvtss_f32(_mm_shuffle_ps(vec, vec, _MM_SHUFFLE(1, 1, 1, 1))),
- _mm_cvtss_f32(_mm_movehl_ps(vec, vec)));
- #else
- float invW = 1.0f / (m30_ * rhs.x_ + m31_ * rhs.y_ + m32_ * rhs.z_ + m33_);
- return Vector3(
- (m00_ * rhs.x_ + m01_ * rhs.y_ + m02_ * rhs.z_ + m03_) * invW,
- (m10_ * rhs.x_ + m11_ * rhs.y_ + m12_ * rhs.z_ + m13_) * invW,
- (m20_ * rhs.x_ + m21_ * rhs.y_ + m22_ * rhs.z_ + m23_) * invW
- );
- #endif
- }
- /// Multiply a Vector4.
- Vector4 operator *(const Vector4& rhs) const
- {
- #ifdef ATOMIC_SSE
- __m128 vec = _mm_loadu_ps(&rhs.x_);
- __m128 r0 = _mm_mul_ps(_mm_loadu_ps(&m00_), vec);
- __m128 r1 = _mm_mul_ps(_mm_loadu_ps(&m10_), vec);
- __m128 t0 = _mm_unpacklo_ps(r0, r1);
- __m128 t1 = _mm_unpackhi_ps(r0, r1);
- t0 = _mm_add_ps(t0, t1);
- __m128 r2 = _mm_mul_ps(_mm_loadu_ps(&m20_), vec);
- __m128 r3 = _mm_mul_ps(_mm_loadu_ps(&m30_), vec);
- __m128 t2 = _mm_unpacklo_ps(r2, r3);
- __m128 t3 = _mm_unpackhi_ps(r2, r3);
- t2 = _mm_add_ps(t2, t3);
- vec = _mm_add_ps(_mm_movelh_ps(t0, t2), _mm_movehl_ps(t2, t0));
- Vector4 ret;
- _mm_storeu_ps(&ret.x_, vec);
- return ret;
- #else
- return Vector4(
- m00_ * rhs.x_ + m01_ * rhs.y_ + m02_ * rhs.z_ + m03_ * rhs.w_,
- m10_ * rhs.x_ + m11_ * rhs.y_ + m12_ * rhs.z_ + m13_ * rhs.w_,
- m20_ * rhs.x_ + m21_ * rhs.y_ + m22_ * rhs.z_ + m23_ * rhs.w_,
- m30_ * rhs.x_ + m31_ * rhs.y_ + m32_ * rhs.z_ + m33_ * rhs.w_
- );
- #endif
- }
- /// Add a matrix.
- Matrix4 operator +(const Matrix4& rhs) const
- {
- #ifdef ATOMIC_SSE
- Matrix4 ret;
- _mm_storeu_ps(&ret.m00_, _mm_add_ps(_mm_loadu_ps(&m00_), _mm_loadu_ps(&rhs.m00_)));
- _mm_storeu_ps(&ret.m10_, _mm_add_ps(_mm_loadu_ps(&m10_), _mm_loadu_ps(&rhs.m10_)));
- _mm_storeu_ps(&ret.m20_, _mm_add_ps(_mm_loadu_ps(&m20_), _mm_loadu_ps(&rhs.m20_)));
- _mm_storeu_ps(&ret.m30_, _mm_add_ps(_mm_loadu_ps(&m30_), _mm_loadu_ps(&rhs.m30_)));
- return ret;
- #else
- return Matrix4(
- m00_ + rhs.m00_,
- m01_ + rhs.m01_,
- m02_ + rhs.m02_,
- m03_ + rhs.m03_,
- m10_ + rhs.m10_,
- m11_ + rhs.m11_,
- m12_ + rhs.m12_,
- m13_ + rhs.m13_,
- m20_ + rhs.m20_,
- m21_ + rhs.m21_,
- m22_ + rhs.m22_,
- m23_ + rhs.m23_,
- m30_ + rhs.m30_,
- m31_ + rhs.m31_,
- m32_ + rhs.m32_,
- m33_ + rhs.m33_
- );
- #endif
- }
- /// Subtract a matrix.
- Matrix4 operator -(const Matrix4& rhs) const
- {
- #ifdef ATOMIC_SSE
- Matrix4 ret;
- _mm_storeu_ps(&ret.m00_, _mm_sub_ps(_mm_loadu_ps(&m00_), _mm_loadu_ps(&rhs.m00_)));
- _mm_storeu_ps(&ret.m10_, _mm_sub_ps(_mm_loadu_ps(&m10_), _mm_loadu_ps(&rhs.m10_)));
- _mm_storeu_ps(&ret.m20_, _mm_sub_ps(_mm_loadu_ps(&m20_), _mm_loadu_ps(&rhs.m20_)));
- _mm_storeu_ps(&ret.m30_, _mm_sub_ps(_mm_loadu_ps(&m30_), _mm_loadu_ps(&rhs.m30_)));
- return ret;
- #else
- return Matrix4(
- m00_ - rhs.m00_,
- m01_ - rhs.m01_,
- m02_ - rhs.m02_,
- m03_ - rhs.m03_,
- m10_ - rhs.m10_,
- m11_ - rhs.m11_,
- m12_ - rhs.m12_,
- m13_ - rhs.m13_,
- m20_ - rhs.m20_,
- m21_ - rhs.m21_,
- m22_ - rhs.m22_,
- m23_ - rhs.m23_,
- m30_ - rhs.m30_,
- m31_ - rhs.m31_,
- m32_ - rhs.m32_,
- m33_ - rhs.m33_
- );
- #endif
- }
- /// Multiply with a scalar.
- Matrix4 operator *(float rhs) const
- {
- #ifdef ATOMIC_SSE
- Matrix4 ret;
- const __m128 mul = _mm_set1_ps(rhs);
- _mm_storeu_ps(&ret.m00_, _mm_mul_ps(_mm_loadu_ps(&m00_), mul));
- _mm_storeu_ps(&ret.m10_, _mm_mul_ps(_mm_loadu_ps(&m10_), mul));
- _mm_storeu_ps(&ret.m20_, _mm_mul_ps(_mm_loadu_ps(&m20_), mul));
- _mm_storeu_ps(&ret.m30_, _mm_mul_ps(_mm_loadu_ps(&m30_), mul));
- return ret;
- #else
- return Matrix4(
- m00_ * rhs,
- m01_ * rhs,
- m02_ * rhs,
- m03_ * rhs,
- m10_ * rhs,
- m11_ * rhs,
- m12_ * rhs,
- m13_ * rhs,
- m20_ * rhs,
- m21_ * rhs,
- m22_ * rhs,
- m23_ * rhs,
- m30_ * rhs,
- m31_ * rhs,
- m32_ * rhs,
- m33_ * rhs
- );
- #endif
- }
- /// Multiply a matrix.
- Matrix4 operator *(const Matrix4& rhs) const
- {
- #ifdef ATOMIC_SSE
- Matrix4 out;
- __m128 r0 = _mm_loadu_ps(&rhs.m00_);
- __m128 r1 = _mm_loadu_ps(&rhs.m10_);
- __m128 r2 = _mm_loadu_ps(&rhs.m20_);
- __m128 r3 = _mm_loadu_ps(&rhs.m30_);
- __m128 l = _mm_loadu_ps(&m00_);
- __m128 t0 = _mm_mul_ps(_mm_shuffle_ps(l, l, _MM_SHUFFLE(0, 0, 0, 0)), r0);
- __m128 t1 = _mm_mul_ps(_mm_shuffle_ps(l, l, _MM_SHUFFLE(1, 1, 1, 1)), r1);
- __m128 t2 = _mm_mul_ps(_mm_shuffle_ps(l, l, _MM_SHUFFLE(2, 2, 2, 2)), r2);
- __m128 t3 = _mm_mul_ps(_mm_shuffle_ps(l, l, _MM_SHUFFLE(3, 3, 3, 3)), r3);
- _mm_storeu_ps(&out.m00_, _mm_add_ps(_mm_add_ps(t0, t1), _mm_add_ps(t2, t3)));
- l = _mm_loadu_ps(&m10_);
- t0 = _mm_mul_ps(_mm_shuffle_ps(l, l, _MM_SHUFFLE(0, 0, 0, 0)), r0);
- t1 = _mm_mul_ps(_mm_shuffle_ps(l, l, _MM_SHUFFLE(1, 1, 1, 1)), r1);
- t2 = _mm_mul_ps(_mm_shuffle_ps(l, l, _MM_SHUFFLE(2, 2, 2, 2)), r2);
- t3 = _mm_mul_ps(_mm_shuffle_ps(l, l, _MM_SHUFFLE(3, 3, 3, 3)), r3);
- _mm_storeu_ps(&out.m10_, _mm_add_ps(_mm_add_ps(t0, t1), _mm_add_ps(t2, t3)));
- l = _mm_loadu_ps(&m20_);
- t0 = _mm_mul_ps(_mm_shuffle_ps(l, l, _MM_SHUFFLE(0, 0, 0, 0)), r0);
- t1 = _mm_mul_ps(_mm_shuffle_ps(l, l, _MM_SHUFFLE(1, 1, 1, 1)), r1);
- t2 = _mm_mul_ps(_mm_shuffle_ps(l, l, _MM_SHUFFLE(2, 2, 2, 2)), r2);
- t3 = _mm_mul_ps(_mm_shuffle_ps(l, l, _MM_SHUFFLE(3, 3, 3, 3)), r3);
- _mm_storeu_ps(&out.m20_, _mm_add_ps(_mm_add_ps(t0, t1), _mm_add_ps(t2, t3)));
- l = _mm_loadu_ps(&m30_);
- t0 = _mm_mul_ps(_mm_shuffle_ps(l, l, _MM_SHUFFLE(0, 0, 0, 0)), r0);
- t1 = _mm_mul_ps(_mm_shuffle_ps(l, l, _MM_SHUFFLE(1, 1, 1, 1)), r1);
- t2 = _mm_mul_ps(_mm_shuffle_ps(l, l, _MM_SHUFFLE(2, 2, 2, 2)), r2);
- t3 = _mm_mul_ps(_mm_shuffle_ps(l, l, _MM_SHUFFLE(3, 3, 3, 3)), r3);
- _mm_storeu_ps(&out.m30_, _mm_add_ps(_mm_add_ps(t0, t1), _mm_add_ps(t2, t3)));
- return out;
- #else
- return Matrix4(
- m00_ * rhs.m00_ + m01_ * rhs.m10_ + m02_ * rhs.m20_ + m03_ * rhs.m30_,
- m00_ * rhs.m01_ + m01_ * rhs.m11_ + m02_ * rhs.m21_ + m03_ * rhs.m31_,
- m00_ * rhs.m02_ + m01_ * rhs.m12_ + m02_ * rhs.m22_ + m03_ * rhs.m32_,
- m00_ * rhs.m03_ + m01_ * rhs.m13_ + m02_ * rhs.m23_ + m03_ * rhs.m33_,
- m10_ * rhs.m00_ + m11_ * rhs.m10_ + m12_ * rhs.m20_ + m13_ * rhs.m30_,
- m10_ * rhs.m01_ + m11_ * rhs.m11_ + m12_ * rhs.m21_ + m13_ * rhs.m31_,
- m10_ * rhs.m02_ + m11_ * rhs.m12_ + m12_ * rhs.m22_ + m13_ * rhs.m32_,
- m10_ * rhs.m03_ + m11_ * rhs.m13_ + m12_ * rhs.m23_ + m13_ * rhs.m33_,
- m20_ * rhs.m00_ + m21_ * rhs.m10_ + m22_ * rhs.m20_ + m23_ * rhs.m30_,
- m20_ * rhs.m01_ + m21_ * rhs.m11_ + m22_ * rhs.m21_ + m23_ * rhs.m31_,
- m20_ * rhs.m02_ + m21_ * rhs.m12_ + m22_ * rhs.m22_ + m23_ * rhs.m32_,
- m20_ * rhs.m03_ + m21_ * rhs.m13_ + m22_ * rhs.m23_ + m23_ * rhs.m33_,
- m30_ * rhs.m00_ + m31_ * rhs.m10_ + m32_ * rhs.m20_ + m33_ * rhs.m30_,
- m30_ * rhs.m01_ + m31_ * rhs.m11_ + m32_ * rhs.m21_ + m33_ * rhs.m31_,
- m30_ * rhs.m02_ + m31_ * rhs.m12_ + m32_ * rhs.m22_ + m33_ * rhs.m32_,
- m30_ * rhs.m03_ + m31_ * rhs.m13_ + m32_ * rhs.m23_ + m33_ * rhs.m33_
- );
- #endif
- }
- /// Multiply with a 3x4 matrix.
- Matrix4 operator *(const Matrix3x4& rhs) const;
- /// Set translation elements.
- void SetTranslation(const Vector3& translation)
- {
- m03_ = translation.x_;
- m13_ = translation.y_;
- m23_ = translation.z_;
- }
- /// Set rotation elements from a 3x3 matrix.
- void SetRotation(const Matrix3& rotation)
- {
- m00_ = rotation.m00_;
- m01_ = rotation.m01_;
- m02_ = rotation.m02_;
- m10_ = rotation.m10_;
- m11_ = rotation.m11_;
- m12_ = rotation.m12_;
- m20_ = rotation.m20_;
- m21_ = rotation.m21_;
- m22_ = rotation.m22_;
- }
- /// Set scaling elements.
- void SetScale(const Vector3& scale)
- {
- m00_ = scale.x_;
- m11_ = scale.y_;
- m22_ = scale.z_;
- }
- /// Set uniform scaling elements.
- void SetScale(float scale)
- {
- m00_ = scale;
- m11_ = scale;
- m22_ = scale;
- }
- /// Return the combined rotation and scaling matrix.
- Matrix3 ToMatrix3() const
- {
- return Matrix3(
- m00_,
- m01_,
- m02_,
- m10_,
- m11_,
- m12_,
- m20_,
- m21_,
- m22_
- );
- }
- /// Return the rotation matrix with scaling removed.
- Matrix3 RotationMatrix() const
- {
- Vector3 invScale(
- 1.0f / sqrtf(m00_ * m00_ + m10_ * m10_ + m20_ * m20_),
- 1.0f / sqrtf(m01_ * m01_ + m11_ * m11_ + m21_ * m21_),
- 1.0f / sqrtf(m02_ * m02_ + m12_ * m12_ + m22_ * m22_)
- );
- return ToMatrix3().Scaled(invScale);
- }
- /// Return the translation part.
- Vector3 Translation() const
- {
- return Vector3(
- m03_,
- m13_,
- m23_
- );
- }
- /// Return the rotation part.
- Quaternion Rotation() const { return Quaternion(RotationMatrix()); }
- /// Return the scaling part.
- Vector3 Scale() const
- {
- return Vector3(
- sqrtf(m00_ * m00_ + m10_ * m10_ + m20_ * m20_),
- sqrtf(m01_ * m01_ + m11_ * m11_ + m21_ * m21_),
- sqrtf(m02_ * m02_ + m12_ * m12_ + m22_ * m22_)
- );
- }
- /// Return the scaling part with the sign. Reference rotation matrix is required to avoid ambiguity.
- Vector3 SignedScale(const Matrix3& rotation) const
- {
- return Vector3(
- rotation.m00_ * m00_ + rotation.m10_ * m10_ + rotation.m20_ * m20_,
- rotation.m01_ * m01_ + rotation.m11_ * m11_ + rotation.m21_ * m21_,
- rotation.m02_ * m02_ + rotation.m12_ * m12_ + rotation.m22_ * m22_
- );
- }
- /// Return transposed.
- Matrix4 Transpose() const
- {
- #ifdef ATOMIC_SSE
- __m128 m0 = _mm_loadu_ps(&m00_);
- __m128 m1 = _mm_loadu_ps(&m10_);
- __m128 m2 = _mm_loadu_ps(&m20_);
- __m128 m3 = _mm_loadu_ps(&m30_);
- _MM_TRANSPOSE4_PS(m0, m1, m2, m3);
- Matrix4 out;
- _mm_storeu_ps(&out.m00_, m0);
- _mm_storeu_ps(&out.m10_, m1);
- _mm_storeu_ps(&out.m20_, m2);
- _mm_storeu_ps(&out.m30_, m3);
- return out;
- #else
- return Matrix4(
- m00_,
- m10_,
- m20_,
- m30_,
- m01_,
- m11_,
- m21_,
- m31_,
- m02_,
- m12_,
- m22_,
- m32_,
- m03_,
- m13_,
- m23_,
- m33_
- );
- #endif
- }
- /// Test for equality with another matrix with epsilon.
- bool Equals(const Matrix4& rhs) const
- {
- const float* leftData = Data();
- const float* rightData = rhs.Data();
- for (unsigned i = 0; i < 16; ++i)
- {
- if (!Atomic::Equals(leftData[i], rightData[i]))
- return false;
- }
- return true;
- }
- /// Return decomposition to translation, rotation and scale.
- void Decompose(Vector3& translation, Quaternion& rotation, Vector3& scale) const;
- /// Return inverse.
- Matrix4 Inverse() const;
- /// Return float data.
- const float* Data() const { return &m00_; }
- /// Return matrix element.
- float Element(unsigned i, unsigned j) const { return Data()[i * 4 + j]; }
- /// Return matrix row.
- Vector4 Row(unsigned i) const { return Vector4(Element(i, 0), Element(i, 1), Element(i, 2), Element(i, 3)); }
- /// Return matrix column.
- Vector4 Column(unsigned j) const { return Vector4(Element(0, j), Element(1, j), Element(2, j), Element(3, j)); }
- /// Return as string.
- String ToString() const;
- float m00_;
- float m01_;
- float m02_;
- float m03_;
- float m10_;
- float m11_;
- float m12_;
- float m13_;
- float m20_;
- float m21_;
- float m22_;
- float m23_;
- float m30_;
- float m31_;
- float m32_;
- float m33_;
- /// Bulk transpose matrices.
- static void BulkTranspose(float* dest, const float* src, unsigned count)
- {
- for (unsigned i = 0; i < count; ++i)
- {
- #ifdef ATOMIC_SSE
- __m128 m0 = _mm_loadu_ps(src);
- __m128 m1 = _mm_loadu_ps(src + 4);
- __m128 m2 = _mm_loadu_ps(src + 8);
- __m128 m3 = _mm_loadu_ps(src + 12);
- _MM_TRANSPOSE4_PS(m0, m1, m2, m3);
- _mm_storeu_ps(dest, m0);
- _mm_storeu_ps(dest + 4, m1);
- _mm_storeu_ps(dest + 8, m2);
- _mm_storeu_ps(dest + 12, m3);
- #else
- dest[0] = src[0];
- dest[1] = src[4];
- dest[2] = src[8];
- dest[3] = src[12];
- dest[4] = src[1];
- dest[5] = src[5];
- dest[6] = src[9];
- dest[7] = src[13];
- dest[8] = src[2];
- dest[9] = src[6];
- dest[10] = src[10];
- dest[11] = src[14];
- dest[12] = src[3];
- dest[13] = src[7];
- dest[14] = src[11];
- dest[15] = src[15];
- #endif
- dest += 16;
- src += 16;
- }
- }
- /// Zero matrix.
- static const Matrix4 ZERO;
- /// Identity matrix.
- static const Matrix4 IDENTITY;
- };
- /// Multiply a 4x4 matrix with a scalar.
- inline Matrix4 operator *(float lhs, const Matrix4& rhs) { return rhs * lhs; }
- }
|