123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729730731732733734735736737738739740741742743744745746747748749750751752753754755756757758759760761762763764765766767768769770771772773774775776777778779780781782783784785786787788789790791792793794795796797798799800801802803804805806807808809810811812813814815816817818819820821822823824825826827828829830831832833834835836837838839840841842843844845846847848849850851852853854855856857858859860861862863864865866867868869870871872873874875876877878879880881882883884885886887888889890891892893894895896897898899900901902903904905906907908909910911912913914915916917918919920921922923924925926927928929930931932933934935936937938939940941942943944945946947948949950951952953954955956957958959960961962963964965966967968969970971972973974975976977978979980981982983984985986987988989990991992993994995996997998999100010011002100310041005100610071008100910101011101210131014101510161017101810191020102110221023102410251026102710281029103010311032103310341035103610371038103910401041104210431044104510461047104810491050105110521053105410551056105710581059106010611062106310641065106610671068106910701071107210731074107510761077107810791080108110821083108410851086108710881089109010911092109310941095109610971098109911001101110211031104110511061107110811091110111111121113111411151116111711181119112011211122112311241125112611271128112911301131113211331134113511361137113811391140114111421143114411451146114711481149115011511152115311541155115611571158115911601161116211631164116511661167116811691170117111721173117411751176117711781179118011811182118311841185118611871188118911901191119211931194119511961197119811991200120112021203120412051206120712081209121012111212121312141215121612171218121912201221122212231224122512261227122812291230123112321233123412351236123712381239124012411242124312441245124612471248124912501251125212531254125512561257125812591260126112621263126412651266126712681269127012711272127312741275127612771278127912801281128212831284128512861287128812891290129112921293129412951296129712981299130013011302130313041305130613071308130913101311131213131314131513161317131813191320132113221323132413251326132713281329133013311332133313341335133613371338133913401341134213431344134513461347134813491350135113521353135413551356135713581359136013611362136313641365136613671368136913701371137213731374137513761377137813791380138113821383138413851386138713881389139013911392139313941395139613971398139914001401140214031404140514061407140814091410141114121413141414151416141714181419142014211422142314241425142614271428142914301431143214331434143514361437143814391440144114421443144414451446144714481449145014511452145314541455145614571458145914601461146214631464146514661467146814691470147114721473147414751476147714781479148014811482148314841485148614871488148914901491149214931494149514961497149814991500150115021503150415051506150715081509151015111512151315141515151615171518151915201521152215231524152515261527152815291530153115321533153415351536153715381539154015411542154315441545154615471548154915501551155215531554155515561557155815591560156115621563156415651566156715681569157015711572157315741575157615771578157915801581158215831584158515861587158815891590159115921593159415951596159715981599160016011602160316041605160616071608160916101611161216131614161516161617161816191620162116221623162416251626162716281629163016311632163316341635163616371638163916401641164216431644164516461647164816491650165116521653165416551656165716581659166016611662166316641665166616671668166916701671167216731674167516761677167816791680168116821683168416851686168716881689169016911692169316941695169616971698169917001701170217031704170517061707170817091710171117121713171417151716171717181719172017211722172317241725172617271728172917301731173217331734173517361737173817391740174117421743174417451746174717481749175017511752175317541755175617571758175917601761176217631764176517661767176817691770177117721773177417751776177717781779178017811782178317841785178617871788178917901791179217931794179517961797179817991800180118021803180418051806180718081809181018111812181318141815181618171818181918201821182218231824182518261827182818291830183118321833183418351836183718381839184018411842184318441845184618471848184918501851185218531854185518561857185818591860186118621863186418651866186718681869187018711872187318741875187618771878187918801881188218831884188518861887188818891890189118921893189418951896189718981899190019011902190319041905190619071908190919101911191219131914191519161917191819191920192119221923192419251926192719281929193019311932193319341935193619371938193919401941194219431944194519461947194819491950195119521953195419551956195719581959196019611962196319641965196619671968196919701971197219731974197519761977197819791980198119821983 |
- // basisu_containers.h
- #pragma once
- #include <stdlib.h>
- #include <stdio.h>
- #include <stdint.h>
- #include <assert.h>
- #include <algorithm>
- #if defined(__linux__) && !defined(ANDROID)
- // Only for malloc_usable_size() in basisu_containers_impl.h
- #include <malloc.h>
- #define HAS_MALLOC_USABLE_SIZE 1
- #endif
- // Set to 1 to always check vector operator[], front(), and back() even in release.
- #define BASISU_VECTOR_FORCE_CHECKING 0
- // If 1, the vector container will not query the CRT to get the size of resized memory blocks.
- #define BASISU_VECTOR_DETERMINISTIC 1
- #ifdef _MSC_VER
- #define BASISU_FORCE_INLINE __forceinline
- #else
- #define BASISU_FORCE_INLINE inline
- #endif
- namespace basisu
- {
- enum { cInvalidIndex = -1 };
- namespace helpers
- {
- inline bool is_power_of_2(uint32_t x) { return x && ((x & (x - 1U)) == 0U); }
- inline bool is_power_of_2(uint64_t x) { return x && ((x & (x - 1U)) == 0U); }
- template<class T> const T& minimum(const T& a, const T& b) { return (b < a) ? b : a; }
- template<class T> const T& maximum(const T& a, const T& b) { return (a < b) ? b : a; }
- inline uint32_t floor_log2i(uint32_t v)
- {
- uint32_t l = 0;
- while (v > 1U)
- {
- v >>= 1;
- l++;
- }
- return l;
- }
- inline uint32_t next_pow2(uint32_t val)
- {
- val--;
- val |= val >> 16;
- val |= val >> 8;
- val |= val >> 4;
- val |= val >> 2;
- val |= val >> 1;
- return val + 1;
- }
- inline uint64_t next_pow2(uint64_t val)
- {
- val--;
- val |= val >> 32;
- val |= val >> 16;
- val |= val >> 8;
- val |= val >> 4;
- val |= val >> 2;
- val |= val >> 1;
- return val + 1;
- }
- } // namespace helpers
- template <typename T>
- inline T* construct(T* p)
- {
- return new (static_cast<void*>(p)) T;
- }
- template <typename T, typename U>
- inline T* construct(T* p, const U& init)
- {
- return new (static_cast<void*>(p)) T(init);
- }
- template <typename T>
- inline void construct_array(T* p, size_t n)
- {
- T* q = p + n;
- for (; p != q; ++p)
- new (static_cast<void*>(p)) T;
- }
- template <typename T, typename U>
- inline void construct_array(T* p, size_t n, const U& init)
- {
- T* q = p + n;
- for (; p != q; ++p)
- new (static_cast<void*>(p)) T(init);
- }
- template <typename T>
- inline void destruct(T* p)
- {
- (void)p;
- p->~T();
- }
- template <typename T> inline void destruct_array(T* p, size_t n)
- {
- T* q = p + n;
- for (; p != q; ++p)
- p->~T();
- }
- template<typename T> struct int_traits { enum { cMin = INT32_MIN, cMax = INT32_MAX, cSigned = true }; };
- template<> struct int_traits<int8_t> { enum { cMin = INT8_MIN, cMax = INT8_MAX, cSigned = true }; };
- template<> struct int_traits<int16_t> { enum { cMin = INT16_MIN, cMax = INT16_MAX, cSigned = true }; };
- template<> struct int_traits<int32_t> { enum { cMin = INT32_MIN, cMax = INT32_MAX, cSigned = true }; };
- template<> struct int_traits<uint8_t> { enum { cMin = 0, cMax = UINT8_MAX, cSigned = false }; };
- template<> struct int_traits<uint16_t> { enum { cMin = 0, cMax = UINT16_MAX, cSigned = false }; };
- template<> struct int_traits<uint32_t> { enum { cMin = 0, cMax = UINT32_MAX, cSigned = false }; };
- template<typename T>
- struct scalar_type
- {
- enum { cFlag = false };
- static inline void construct(T* p) { basisu::construct(p); }
- static inline void construct(T* p, const T& init) { basisu::construct(p, init); }
- static inline void construct_array(T* p, size_t n) { basisu::construct_array(p, n); }
- static inline void destruct(T* p) { basisu::destruct(p); }
- static inline void destruct_array(T* p, size_t n) { basisu::destruct_array(p, n); }
- };
- template<typename T> struct scalar_type<T*>
- {
- enum { cFlag = true };
- static inline void construct(T** p) { memset(p, 0, sizeof(T*)); }
- static inline void construct(T** p, T* init) { *p = init; }
- static inline void construct_array(T** p, size_t n) { memset(p, 0, sizeof(T*) * n); }
- static inline void destruct(T** p) { p; }
- static inline void destruct_array(T** p, size_t n) { p, n; }
- };
- #define BASISU_DEFINE_BUILT_IN_TYPE(X) \
- template<> struct scalar_type<X> { \
- enum { cFlag = true }; \
- static inline void construct(X* p) { memset(p, 0, sizeof(X)); } \
- static inline void construct(X* p, const X& init) { memcpy(p, &init, sizeof(X)); } \
- static inline void construct_array(X* p, size_t n) { memset(p, 0, sizeof(X) * n); } \
- static inline void destruct(X* p) { p; } \
- static inline void destruct_array(X* p, size_t n) { p, n; } };
- BASISU_DEFINE_BUILT_IN_TYPE(bool)
- BASISU_DEFINE_BUILT_IN_TYPE(char)
- BASISU_DEFINE_BUILT_IN_TYPE(unsigned char)
- BASISU_DEFINE_BUILT_IN_TYPE(short)
- BASISU_DEFINE_BUILT_IN_TYPE(unsigned short)
- BASISU_DEFINE_BUILT_IN_TYPE(int)
- BASISU_DEFINE_BUILT_IN_TYPE(unsigned int)
- BASISU_DEFINE_BUILT_IN_TYPE(long)
- BASISU_DEFINE_BUILT_IN_TYPE(unsigned long)
- #ifdef __GNUC__
- BASISU_DEFINE_BUILT_IN_TYPE(long long)
- BASISU_DEFINE_BUILT_IN_TYPE(unsigned long long)
- #else
- BASISU_DEFINE_BUILT_IN_TYPE(__int64)
- BASISU_DEFINE_BUILT_IN_TYPE(unsigned __int64)
- #endif
- BASISU_DEFINE_BUILT_IN_TYPE(float)
- BASISU_DEFINE_BUILT_IN_TYPE(double)
- BASISU_DEFINE_BUILT_IN_TYPE(long double)
- #undef BASISU_DEFINE_BUILT_IN_TYPE
- template<typename T>
- struct bitwise_movable { enum { cFlag = false }; };
- #define BASISU_DEFINE_BITWISE_MOVABLE(Q) template<> struct bitwise_movable<Q> { enum { cFlag = true }; };
- template<typename T>
- struct bitwise_copyable { enum { cFlag = false }; };
- #define BASISU_DEFINE_BITWISE_COPYABLE(Q) template<> struct bitwise_copyable<Q> { enum { cFlag = true }; };
- #define BASISU_IS_POD(T) __is_pod(T)
- #define BASISU_IS_SCALAR_TYPE(T) (scalar_type<T>::cFlag)
- #if defined(__GNUC__) && __GNUC__<5
- #define BASISU_IS_TRIVIALLY_COPYABLE(...) __has_trivial_copy(__VA_ARGS__)
- #else
- #define BASISU_IS_TRIVIALLY_COPYABLE(...) std::is_trivially_copyable<__VA_ARGS__>::value
- #endif
- // TODO: clean this up
- #define BASISU_IS_BITWISE_COPYABLE(T) (BASISU_IS_SCALAR_TYPE(T) || BASISU_IS_POD(T) || BASISU_IS_TRIVIALLY_COPYABLE(T) || (bitwise_copyable<T>::cFlag))
- #define BASISU_IS_BITWISE_COPYABLE_OR_MOVABLE(T) (BASISU_IS_BITWISE_COPYABLE(T) || (bitwise_movable<T>::cFlag))
- #define BASISU_HAS_DESTRUCTOR(T) ((!scalar_type<T>::cFlag) && (!__is_pod(T)))
- typedef char(&yes_t)[1];
- typedef char(&no_t)[2];
- template <class U> yes_t class_test(int U::*);
- template <class U> no_t class_test(...);
- template <class T> struct is_class
- {
- enum { value = (sizeof(class_test<T>(0)) == sizeof(yes_t)) };
- };
- template <typename T> struct is_pointer
- {
- enum { value = false };
- };
- template <typename T> struct is_pointer<T*>
- {
- enum { value = true };
- };
- struct empty_type { };
- BASISU_DEFINE_BITWISE_COPYABLE(empty_type);
- BASISU_DEFINE_BITWISE_MOVABLE(empty_type);
- template<typename T> struct rel_ops
- {
- friend bool operator!=(const T& x, const T& y) { return (!(x == y)); }
- friend bool operator> (const T& x, const T& y) { return (y < x); }
- friend bool operator<=(const T& x, const T& y) { return (!(y < x)); }
- friend bool operator>=(const T& x, const T& y) { return (!(x < y)); }
- };
- struct elemental_vector
- {
- void* m_p;
- uint32_t m_size;
- uint32_t m_capacity;
- typedef void (*object_mover)(void* pDst, void* pSrc, uint32_t num);
- bool increase_capacity(uint32_t min_new_capacity, bool grow_hint, uint32_t element_size, object_mover pRelocate, bool nofail);
- };
- template<typename T>
- class vector : public rel_ops< vector<T> >
- {
- public:
- typedef T* iterator;
- typedef const T* const_iterator;
- typedef T value_type;
- typedef T& reference;
- typedef const T& const_reference;
- typedef T* pointer;
- typedef const T* const_pointer;
- inline vector() :
- m_p(NULL),
- m_size(0),
- m_capacity(0)
- {
- }
- inline vector(uint32_t n, const T& init) :
- m_p(NULL),
- m_size(0),
- m_capacity(0)
- {
- increase_capacity(n, false);
- construct_array(m_p, n, init);
- m_size = n;
- }
- inline vector(const vector& other) :
- m_p(NULL),
- m_size(0),
- m_capacity(0)
- {
- increase_capacity(other.m_size, false);
- m_size = other.m_size;
- if (BASISU_IS_BITWISE_COPYABLE(T))
- {
- if ((m_p) && (other.m_p))
- memcpy(m_p, other.m_p, m_size * sizeof(T));
- }
- else
- {
- T* pDst = m_p;
- const T* pSrc = other.m_p;
- for (uint32_t i = m_size; i > 0; i--)
- construct(pDst++, *pSrc++);
- }
- }
- inline explicit vector(size_t size) :
- m_p(NULL),
- m_size(0),
- m_capacity(0)
- {
- resize(size);
- }
- inline ~vector()
- {
- if (m_p)
- {
- scalar_type<T>::destruct_array(m_p, m_size);
- free(m_p);
- }
- }
- inline vector& operator= (const vector& other)
- {
- if (this == &other)
- return *this;
- if (m_capacity >= other.m_size)
- resize(0);
- else
- {
- clear();
- increase_capacity(other.m_size, false);
- }
- if (BASISU_IS_BITWISE_COPYABLE(T))
- {
- if ((m_p) && (other.m_p))
- memcpy(m_p, other.m_p, other.m_size * sizeof(T));
- }
- else
- {
- T* pDst = m_p;
- const T* pSrc = other.m_p;
- for (uint32_t i = other.m_size; i > 0; i--)
- construct(pDst++, *pSrc++);
- }
- m_size = other.m_size;
- return *this;
- }
- BASISU_FORCE_INLINE const T* begin() const { return m_p; }
- BASISU_FORCE_INLINE T* begin() { return m_p; }
- BASISU_FORCE_INLINE const T* end() const { return m_p + m_size; }
- BASISU_FORCE_INLINE T* end() { return m_p + m_size; }
- BASISU_FORCE_INLINE bool empty() const { return !m_size; }
- BASISU_FORCE_INLINE uint32_t size() const { return m_size; }
- BASISU_FORCE_INLINE uint32_t size_in_bytes() const { return m_size * sizeof(T); }
- BASISU_FORCE_INLINE uint32_t capacity() const { return m_capacity; }
- // operator[] will assert on out of range indices, but in final builds there is (and will never be) any range checking on this method.
- //BASISU_FORCE_INLINE const T& operator[] (uint32_t i) const { assert(i < m_size); return m_p[i]; }
- //BASISU_FORCE_INLINE T& operator[] (uint32_t i) { assert(i < m_size); return m_p[i]; }
-
- #if !BASISU_VECTOR_FORCE_CHECKING
- BASISU_FORCE_INLINE const T& operator[] (size_t i) const { assert(i < m_size); return m_p[i]; }
- BASISU_FORCE_INLINE T& operator[] (size_t i) { assert(i < m_size); return m_p[i]; }
- #else
- BASISU_FORCE_INLINE const T& operator[] (size_t i) const
- {
- if (i >= m_size)
- {
- fprintf(stderr, "operator[] invalid index: %u, max entries %u, type size %u\n", (uint32_t)i, m_size, (uint32_t)sizeof(T));
- abort();
- }
- return m_p[i];
- }
- BASISU_FORCE_INLINE T& operator[] (size_t i)
- {
- if (i >= m_size)
- {
- fprintf(stderr, "operator[] invalid index: %u, max entries %u, type size %u\n", (uint32_t)i, m_size, (uint32_t)sizeof(T));
- abort();
- }
- return m_p[i];
- }
- #endif
- // at() always includes range checking, even in final builds, unlike operator [].
- // The first element is returned if the index is out of range.
- BASISU_FORCE_INLINE const T& at(size_t i) const { assert(i < m_size); return (i >= m_size) ? m_p[0] : m_p[i]; }
- BASISU_FORCE_INLINE T& at(size_t i) { assert(i < m_size); return (i >= m_size) ? m_p[0] : m_p[i]; }
-
- #if !BASISU_VECTOR_FORCE_CHECKING
- BASISU_FORCE_INLINE const T& front() const { assert(m_size); return m_p[0]; }
- BASISU_FORCE_INLINE T& front() { assert(m_size); return m_p[0]; }
- BASISU_FORCE_INLINE const T& back() const { assert(m_size); return m_p[m_size - 1]; }
- BASISU_FORCE_INLINE T& back() { assert(m_size); return m_p[m_size - 1]; }
- #else
- BASISU_FORCE_INLINE const T& front() const
- {
- if (!m_size)
- {
- fprintf(stderr, "front: vector is empty, type size %u\n", (uint32_t)sizeof(T));
- abort();
- }
- return m_p[0];
- }
- BASISU_FORCE_INLINE T& front()
- {
- if (!m_size)
- {
- fprintf(stderr, "front: vector is empty, type size %u\n", (uint32_t)sizeof(T));
- abort();
- }
- return m_p[0];
- }
- BASISU_FORCE_INLINE const T& back() const
- {
- if(!m_size)
- {
- fprintf(stderr, "back: vector is empty, type size %u\n", (uint32_t)sizeof(T));
- abort();
- }
- return m_p[m_size - 1];
- }
- BASISU_FORCE_INLINE T& back()
- {
- if (!m_size)
- {
- fprintf(stderr, "back: vector is empty, type size %u\n", (uint32_t)sizeof(T));
- abort();
- }
- return m_p[m_size - 1];
- }
- #endif
- BASISU_FORCE_INLINE const T* get_ptr() const { return m_p; }
- BASISU_FORCE_INLINE T* get_ptr() { return m_p; }
- BASISU_FORCE_INLINE const T* data() const { return m_p; }
- BASISU_FORCE_INLINE T* data() { return m_p; }
- // clear() sets the container to empty, then frees the allocated block.
- inline void clear()
- {
- if (m_p)
- {
- scalar_type<T>::destruct_array(m_p, m_size);
- free(m_p);
- m_p = NULL;
- m_size = 0;
- m_capacity = 0;
- }
- }
- inline void clear_no_destruction()
- {
- if (m_p)
- {
- free(m_p);
- m_p = NULL;
- m_size = 0;
- m_capacity = 0;
- }
- }
- inline void reserve(size_t new_capacity_size_t)
- {
- if (new_capacity_size_t > UINT32_MAX)
- {
- assert(0);
- return;
- }
- uint32_t new_capacity = (uint32_t)new_capacity_size_t;
- if (new_capacity > m_capacity)
- increase_capacity(new_capacity, false);
- else if (new_capacity < m_capacity)
- {
- // Must work around the lack of a "decrease_capacity()" method.
- // This case is rare enough in practice that it's probably not worth implementing an optimized in-place resize.
- vector tmp;
- tmp.increase_capacity(helpers::maximum(m_size, new_capacity), false);
- tmp = *this;
- swap(tmp);
- }
- }
- inline bool try_reserve(size_t new_capacity_size_t)
- {
- if (new_capacity_size_t > UINT32_MAX)
- {
- assert(0);
- return false;
- }
- uint32_t new_capacity = (uint32_t)new_capacity_size_t;
- if (new_capacity > m_capacity)
- {
- if (!increase_capacity(new_capacity, false))
- return false;
- }
- else if (new_capacity < m_capacity)
- {
- // Must work around the lack of a "decrease_capacity()" method.
- // This case is rare enough in practice that it's probably not worth implementing an optimized in-place resize.
- vector tmp;
- tmp.increase_capacity(helpers::maximum(m_size, new_capacity), false);
- tmp = *this;
- swap(tmp);
- }
- return true;
- }
- // resize(0) sets the container to empty, but does not free the allocated block.
- inline void resize(size_t new_size_size_t, bool grow_hint = false)
- {
- if (new_size_size_t > UINT32_MAX)
- {
- assert(0);
- return;
- }
- uint32_t new_size = (uint32_t)new_size_size_t;
- if (m_size != new_size)
- {
- if (new_size < m_size)
- scalar_type<T>::destruct_array(m_p + new_size, m_size - new_size);
- else
- {
- if (new_size > m_capacity)
- increase_capacity(new_size, (new_size == (m_size + 1)) || grow_hint);
- scalar_type<T>::construct_array(m_p + m_size, new_size - m_size);
- }
- m_size = new_size;
- }
- }
- inline bool try_resize(size_t new_size_size_t, bool grow_hint = false)
- {
- if (new_size_size_t > UINT32_MAX)
- {
- assert(0);
- return false;
- }
- uint32_t new_size = (uint32_t)new_size_size_t;
- if (m_size != new_size)
- {
- if (new_size < m_size)
- scalar_type<T>::destruct_array(m_p + new_size, m_size - new_size);
- else
- {
- if (new_size > m_capacity)
- {
- if (!increase_capacity(new_size, (new_size == (m_size + 1)) || grow_hint, true))
- return false;
- }
- scalar_type<T>::construct_array(m_p + m_size, new_size - m_size);
- }
- m_size = new_size;
- }
- return true;
- }
- // If size >= capacity/2, reset() sets the container's size to 0 but doesn't free the allocated block (because the container may be similarly loaded in the future).
- // Otherwise it blows away the allocated block. See http://www.codercorner.com/blog/?p=494
- inline void reset()
- {
- if (m_size >= (m_capacity >> 1))
- resize(0);
- else
- clear();
- }
- inline T* enlarge(uint32_t i)
- {
- uint32_t cur_size = m_size;
- resize(cur_size + i, true);
- return get_ptr() + cur_size;
- }
- inline T* try_enlarge(uint32_t i)
- {
- uint32_t cur_size = m_size;
- if (!try_resize(cur_size + i, true))
- return NULL;
- return get_ptr() + cur_size;
- }
- BASISU_FORCE_INLINE void push_back(const T& obj)
- {
- assert(!m_p || (&obj < m_p) || (&obj >= (m_p + m_size)));
- if (m_size >= m_capacity)
- increase_capacity(m_size + 1, true);
- scalar_type<T>::construct(m_p + m_size, obj);
- m_size++;
- }
- inline bool try_push_back(const T& obj)
- {
- assert(!m_p || (&obj < m_p) || (&obj >= (m_p + m_size)));
- if (m_size >= m_capacity)
- {
- if (!increase_capacity(m_size + 1, true, true))
- return false;
- }
- scalar_type<T>::construct(m_p + m_size, obj);
- m_size++;
- return true;
- }
- inline void push_back_value(T obj)
- {
- if (m_size >= m_capacity)
- increase_capacity(m_size + 1, true);
- scalar_type<T>::construct(m_p + m_size, obj);
- m_size++;
- }
- inline void pop_back()
- {
- assert(m_size);
- if (m_size)
- {
- m_size--;
- scalar_type<T>::destruct(&m_p[m_size]);
- }
- }
- inline void insert(uint32_t index, const T* p, uint32_t n)
- {
- assert(index <= m_size);
- if (!n)
- return;
- const uint32_t orig_size = m_size;
- resize(m_size + n, true);
- const uint32_t num_to_move = orig_size - index;
- if (BASISU_IS_BITWISE_COPYABLE(T))
- {
- // This overwrites the destination object bits, but bitwise copyable means we don't need to worry about destruction.
- memmove(m_p + index + n, m_p + index, sizeof(T) * num_to_move);
- }
- else
- {
- const T* pSrc = m_p + orig_size - 1;
- T* pDst = const_cast<T*>(pSrc) + n;
- for (uint32_t i = 0; i < num_to_move; i++)
- {
- assert((pDst - m_p) < (int)m_size);
- *pDst-- = *pSrc--;
- }
- }
- T* pDst = m_p + index;
- if (BASISU_IS_BITWISE_COPYABLE(T))
- {
- // This copies in the new bits, overwriting the existing objects, which is OK for copyable types that don't need destruction.
- memcpy(pDst, p, sizeof(T) * n);
- }
- else
- {
- for (uint32_t i = 0; i < n; i++)
- {
- assert((pDst - m_p) < (int)m_size);
- *pDst++ = *p++;
- }
- }
- }
- inline void insert(T* p, const T& obj)
- {
- int64_t ofs = p - begin();
- if ((ofs < 0) || (ofs > UINT32_MAX))
- {
- assert(0);
- return;
- }
- insert((uint32_t)ofs, &obj, 1);
- }
- // push_front() isn't going to be very fast - it's only here for usability.
- inline void push_front(const T& obj)
- {
- insert(0, &obj, 1);
- }
- vector& append(const vector& other)
- {
- if (other.m_size)
- insert(m_size, &other[0], other.m_size);
- return *this;
- }
- vector& append(const T* p, uint32_t n)
- {
- if (n)
- insert(m_size, p, n);
- return *this;
- }
-
- inline void erase(uint32_t start, uint32_t n)
- {
- assert((start + n) <= m_size);
- if ((start + n) > m_size)
- return;
- if (!n)
- return;
- const uint32_t num_to_move = m_size - (start + n);
- T* pDst = m_p + start;
- const T* pSrc = m_p + start + n;
- if (BASISU_IS_BITWISE_COPYABLE_OR_MOVABLE(T))
- {
- // This test is overly cautious.
- if ((!BASISU_IS_BITWISE_COPYABLE(T)) || (BASISU_HAS_DESTRUCTOR(T)))
- {
- // Type has been marked explictly as bitwise movable, which means we can move them around but they may need to be destructed.
- // First destroy the erased objects.
- scalar_type<T>::destruct_array(pDst, n);
- }
- // Copy "down" the objects to preserve, filling in the empty slots.
- memmove(pDst, pSrc, num_to_move * sizeof(T));
- }
- else
- {
- // Type is not bitwise copyable or movable.
- // Move them down one at a time by using the equals operator, and destroying anything that's left over at the end.
- T* pDst_end = pDst + num_to_move;
- while (pDst != pDst_end)
- *pDst++ = *pSrc++;
- scalar_type<T>::destruct_array(pDst_end, n);
- }
- m_size -= n;
- }
- inline void erase(uint32_t index)
- {
- erase(index, 1);
- }
- inline void erase(T* p)
- {
- assert((p >= m_p) && (p < (m_p + m_size)));
- erase(static_cast<uint32_t>(p - m_p));
- }
- inline void erase(T *pFirst, T *pEnd)
- {
- assert(pFirst <= pEnd);
- assert(pFirst >= begin() && pFirst <= end());
- assert(pEnd >= begin() && pEnd <= end());
- int64_t ofs = pFirst - begin();
- if ((ofs < 0) || (ofs > UINT32_MAX))
- {
- assert(0);
- return;
- }
- int64_t n = pEnd - pFirst;
- if ((n < 0) || (n > UINT32_MAX))
- {
- assert(0);
- return;
- }
- erase((uint32_t)ofs, (uint32_t)n);
- }
- void erase_unordered(uint32_t index)
- {
- assert(index < m_size);
- if ((index + 1) < m_size)
- (*this)[index] = back();
- pop_back();
- }
- inline bool operator== (const vector& rhs) const
- {
- if (m_size != rhs.m_size)
- return false;
- else if (m_size)
- {
- if (scalar_type<T>::cFlag)
- return memcmp(m_p, rhs.m_p, sizeof(T) * m_size) == 0;
- else
- {
- const T* pSrc = m_p;
- const T* pDst = rhs.m_p;
- for (uint32_t i = m_size; i; i--)
- if (!(*pSrc++ == *pDst++))
- return false;
- }
- }
- return true;
- }
- inline bool operator< (const vector& rhs) const
- {
- const uint32_t min_size = helpers::minimum(m_size, rhs.m_size);
- const T* pSrc = m_p;
- const T* pSrc_end = m_p + min_size;
- const T* pDst = rhs.m_p;
- while ((pSrc < pSrc_end) && (*pSrc == *pDst))
- {
- pSrc++;
- pDst++;
- }
- if (pSrc < pSrc_end)
- return *pSrc < *pDst;
- return m_size < rhs.m_size;
- }
- inline void swap(vector& other)
- {
- std::swap(m_p, other.m_p);
- std::swap(m_size, other.m_size);
- std::swap(m_capacity, other.m_capacity);
- }
- inline void sort()
- {
- std::sort(begin(), end());
- }
- inline void unique()
- {
- if (!empty())
- {
- sort();
- resize(std::unique(begin(), end()) - begin());
- }
- }
- inline void reverse()
- {
- uint32_t j = m_size >> 1;
- for (uint32_t i = 0; i < j; i++)
- std::swap(m_p[i], m_p[m_size - 1 - i]);
- }
- inline int find(const T& key) const
- {
- const T* p = m_p;
- const T* p_end = m_p + m_size;
- uint32_t index = 0;
- while (p != p_end)
- {
- if (key == *p)
- return index;
- p++;
- index++;
- }
- return cInvalidIndex;
- }
- inline int find_sorted(const T& key) const
- {
- if (m_size)
- {
- // Uniform binary search - Knuth Algorithm 6.2.1 U, unrolled twice.
- int i = ((m_size + 1) >> 1) - 1;
- int m = m_size;
- for (; ; )
- {
- assert(i >= 0 && i < (int)m_size);
- const T* pKey_i = m_p + i;
- int cmp = key < *pKey_i;
- #if defined(_DEBUG) || defined(DEBUG)
- int cmp2 = *pKey_i < key;
- assert((cmp != cmp2) || (key == *pKey_i));
- #endif
- if ((!cmp) && (key == *pKey_i)) return i;
- m >>= 1;
- if (!m) break;
- cmp = -cmp;
- i += (((m + 1) >> 1) ^ cmp) - cmp;
- if (i < 0)
- break;
- assert(i >= 0 && i < (int)m_size);
- pKey_i = m_p + i;
- cmp = key < *pKey_i;
- #if defined(_DEBUG) || defined(DEBUG)
- cmp2 = *pKey_i < key;
- assert((cmp != cmp2) || (key == *pKey_i));
- #endif
- if ((!cmp) && (key == *pKey_i)) return i;
- m >>= 1;
- if (!m) break;
- cmp = -cmp;
- i += (((m + 1) >> 1) ^ cmp) - cmp;
- if (i < 0)
- break;
- }
- }
- return cInvalidIndex;
- }
- template<typename Q>
- inline int find_sorted(const T& key, Q less_than) const
- {
- if (m_size)
- {
- // Uniform binary search - Knuth Algorithm 6.2.1 U, unrolled twice.
- int i = ((m_size + 1) >> 1) - 1;
- int m = m_size;
- for (; ; )
- {
- assert(i >= 0 && i < (int)m_size);
- const T* pKey_i = m_p + i;
- int cmp = less_than(key, *pKey_i);
- if ((!cmp) && (!less_than(*pKey_i, key))) return i;
- m >>= 1;
- if (!m) break;
- cmp = -cmp;
- i += (((m + 1) >> 1) ^ cmp) - cmp;
- if (i < 0)
- break;
- assert(i >= 0 && i < (int)m_size);
- pKey_i = m_p + i;
- cmp = less_than(key, *pKey_i);
- if ((!cmp) && (!less_than(*pKey_i, key))) return i;
- m >>= 1;
- if (!m) break;
- cmp = -cmp;
- i += (((m + 1) >> 1) ^ cmp) - cmp;
- if (i < 0)
- break;
- }
- }
- return cInvalidIndex;
- }
- inline uint32_t count_occurences(const T& key) const
- {
- uint32_t c = 0;
- const T* p = m_p;
- const T* p_end = m_p + m_size;
- while (p != p_end)
- {
- if (key == *p)
- c++;
- p++;
- }
- return c;
- }
- inline void set_all(const T& o)
- {
- if ((sizeof(T) == 1) && (scalar_type<T>::cFlag))
- memset(m_p, *reinterpret_cast<const uint8_t*>(&o), m_size);
- else
- {
- T* pDst = m_p;
- T* pDst_end = pDst + m_size;
- while (pDst != pDst_end)
- *pDst++ = o;
- }
- }
- // Caller assumes ownership of the heap block associated with the container. Container is cleared.
- inline void* assume_ownership()
- {
- T* p = m_p;
- m_p = NULL;
- m_size = 0;
- m_capacity = 0;
- return p;
- }
- // Caller is granting ownership of the indicated heap block.
- // Block must have size constructed elements, and have enough room for capacity elements.
- // The block must have been allocated using malloc().
- // Important: This method is used in Basis Universal. If you change how this container allocates memory, you'll need to change any users of this method.
- inline bool grant_ownership(T* p, uint32_t size, uint32_t capacity)
- {
- // To to prevent the caller from obviously shooting themselves in the foot.
- if (((p + capacity) > m_p) && (p < (m_p + m_capacity)))
- {
- // Can grant ownership of a block inside the container itself!
- assert(0);
- return false;
- }
- if (size > capacity)
- {
- assert(0);
- return false;
- }
- if (!p)
- {
- if (capacity)
- {
- assert(0);
- return false;
- }
- }
- else if (!capacity)
- {
- assert(0);
- return false;
- }
- clear();
- m_p = p;
- m_size = size;
- m_capacity = capacity;
- return true;
- }
- private:
- T* m_p;
- uint32_t m_size;
- uint32_t m_capacity;
- template<typename Q> struct is_vector { enum { cFlag = false }; };
- template<typename Q> struct is_vector< vector<Q> > { enum { cFlag = true }; };
- static void object_mover(void* pDst_void, void* pSrc_void, uint32_t num)
- {
- T* pSrc = static_cast<T*>(pSrc_void);
- T* const pSrc_end = pSrc + num;
- T* pDst = static_cast<T*>(pDst_void);
- while (pSrc != pSrc_end)
- {
- // placement new
- new (static_cast<void*>(pDst)) T(*pSrc);
- pSrc->~T();
- ++pSrc;
- ++pDst;
- }
- }
- inline bool increase_capacity(uint32_t min_new_capacity, bool grow_hint, bool nofail = false)
- {
- return reinterpret_cast<elemental_vector*>(this)->increase_capacity(
- min_new_capacity, grow_hint, sizeof(T),
- (BASISU_IS_BITWISE_COPYABLE_OR_MOVABLE(T) || (is_vector<T>::cFlag)) ? NULL : object_mover, nofail);
- }
- };
- template<typename T> struct bitwise_movable< vector<T> > { enum { cFlag = true }; };
- // Hash map
- template <typename T>
- struct hasher
- {
- inline size_t operator() (const T& key) const { return static_cast<size_t>(key); }
- };
- template <typename T>
- struct equal_to
- {
- inline bool operator()(const T& a, const T& b) const { return a == b; }
- };
- // Important: The Hasher and Equals objects must be bitwise movable!
- template<typename Key, typename Value = empty_type, typename Hasher = hasher<Key>, typename Equals = equal_to<Key> >
- class hash_map
- {
- public:
- class iterator;
- class const_iterator;
-
- private:
- friend class iterator;
- friend class const_iterator;
- enum state
- {
- cStateInvalid = 0,
- cStateValid = 1
- };
- enum
- {
- cMinHashSize = 4U
- };
- public:
- typedef hash_map<Key, Value, Hasher, Equals> hash_map_type;
- typedef std::pair<Key, Value> value_type;
- typedef Key key_type;
- typedef Value referent_type;
- typedef Hasher hasher_type;
- typedef Equals equals_type;
- hash_map() :
- m_hash_shift(32), m_num_valid(0), m_grow_threshold(0)
- {
- }
- hash_map(const hash_map& other) :
- m_values(other.m_values),
- m_hash_shift(other.m_hash_shift),
- m_hasher(other.m_hasher),
- m_equals(other.m_equals),
- m_num_valid(other.m_num_valid),
- m_grow_threshold(other.m_grow_threshold)
- {
- }
- hash_map& operator= (const hash_map& other)
- {
- if (this == &other)
- return *this;
- clear();
- m_values = other.m_values;
- m_hash_shift = other.m_hash_shift;
- m_num_valid = other.m_num_valid;
- m_grow_threshold = other.m_grow_threshold;
- m_hasher = other.m_hasher;
- m_equals = other.m_equals;
- return *this;
- }
- inline ~hash_map()
- {
- clear();
- }
- const Equals& get_equals() const { return m_equals; }
- Equals& get_equals() { return m_equals; }
- void set_equals(const Equals& equals) { m_equals = equals; }
- const Hasher& get_hasher() const { return m_hasher; }
- Hasher& get_hasher() { return m_hasher; }
- void set_hasher(const Hasher& hasher) { m_hasher = hasher; }
- inline void clear()
- {
- if (!m_values.empty())
- {
- if (BASISU_HAS_DESTRUCTOR(Key) || BASISU_HAS_DESTRUCTOR(Value))
- {
- node* p = &get_node(0);
- node* p_end = p + m_values.size();
- uint32_t num_remaining = m_num_valid;
- while (p != p_end)
- {
- if (p->state)
- {
- destruct_value_type(p);
- num_remaining--;
- if (!num_remaining)
- break;
- }
- p++;
- }
- }
- m_values.clear_no_destruction();
- m_hash_shift = 32;
- m_num_valid = 0;
- m_grow_threshold = 0;
- }
- }
- inline void reset()
- {
- if (!m_num_valid)
- return;
- if (BASISU_HAS_DESTRUCTOR(Key) || BASISU_HAS_DESTRUCTOR(Value))
- {
- node* p = &get_node(0);
- node* p_end = p + m_values.size();
- uint32_t num_remaining = m_num_valid;
- while (p != p_end)
- {
- if (p->state)
- {
- destruct_value_type(p);
- p->state = cStateInvalid;
- num_remaining--;
- if (!num_remaining)
- break;
- }
- p++;
- }
- }
- else if (sizeof(node) <= 32)
- {
- memset(&m_values[0], 0, m_values.size_in_bytes());
- }
- else
- {
- node* p = &get_node(0);
- node* p_end = p + m_values.size();
- uint32_t num_remaining = m_num_valid;
- while (p != p_end)
- {
- if (p->state)
- {
- p->state = cStateInvalid;
- num_remaining--;
- if (!num_remaining)
- break;
- }
- p++;
- }
- }
- m_num_valid = 0;
- }
- inline uint32_t size()
- {
- return m_num_valid;
- }
- inline uint32_t get_table_size()
- {
- return m_values.size();
- }
- inline bool empty()
- {
- return !m_num_valid;
- }
- inline void reserve(uint32_t new_capacity)
- {
- uint64_t new_hash_size = helpers::maximum(1U, new_capacity);
- new_hash_size = new_hash_size * 2ULL;
- if (!helpers::is_power_of_2(new_hash_size))
- new_hash_size = helpers::next_pow2(new_hash_size);
- new_hash_size = helpers::maximum<uint64_t>(cMinHashSize, new_hash_size);
- new_hash_size = helpers::minimum<uint64_t>(0x80000000UL, new_hash_size);
- if (new_hash_size > m_values.size())
- rehash((uint32_t)new_hash_size);
- }
-
- class iterator
- {
- friend class hash_map<Key, Value, Hasher, Equals>;
- friend class hash_map<Key, Value, Hasher, Equals>::const_iterator;
- public:
- inline iterator() : m_pTable(NULL), m_index(0) { }
- inline iterator(hash_map_type& table, uint32_t index) : m_pTable(&table), m_index(index) { }
- inline iterator(const iterator& other) : m_pTable(other.m_pTable), m_index(other.m_index) { }
- inline iterator& operator= (const iterator& other)
- {
- m_pTable = other.m_pTable;
- m_index = other.m_index;
- return *this;
- }
- // post-increment
- inline iterator operator++(int)
- {
- iterator result(*this);
- ++*this;
- return result;
- }
- // pre-increment
- inline iterator& operator++()
- {
- probe();
- return *this;
- }
- inline value_type& operator*() const { return *get_cur(); }
- inline value_type* operator->() const { return get_cur(); }
- inline bool operator == (const iterator& b) const { return (m_pTable == b.m_pTable) && (m_index == b.m_index); }
- inline bool operator != (const iterator& b) const { return !(*this == b); }
- inline bool operator == (const const_iterator& b) const { return (m_pTable == b.m_pTable) && (m_index == b.m_index); }
- inline bool operator != (const const_iterator& b) const { return !(*this == b); }
- private:
- hash_map_type* m_pTable;
- uint32_t m_index;
- inline value_type* get_cur() const
- {
- assert(m_pTable && (m_index < m_pTable->m_values.size()));
- assert(m_pTable->get_node_state(m_index) == cStateValid);
- return &m_pTable->get_node(m_index);
- }
- inline void probe()
- {
- assert(m_pTable);
- m_index = m_pTable->find_next(m_index);
- }
- };
- class const_iterator
- {
- friend class hash_map<Key, Value, Hasher, Equals>;
- friend class hash_map<Key, Value, Hasher, Equals>::iterator;
- public:
- inline const_iterator() : m_pTable(NULL), m_index(0) { }
- inline const_iterator(const hash_map_type& table, uint32_t index) : m_pTable(&table), m_index(index) { }
- inline const_iterator(const iterator& other) : m_pTable(other.m_pTable), m_index(other.m_index) { }
- inline const_iterator(const const_iterator& other) : m_pTable(other.m_pTable), m_index(other.m_index) { }
- inline const_iterator& operator= (const const_iterator& other)
- {
- m_pTable = other.m_pTable;
- m_index = other.m_index;
- return *this;
- }
- inline const_iterator& operator= (const iterator& other)
- {
- m_pTable = other.m_pTable;
- m_index = other.m_index;
- return *this;
- }
- // post-increment
- inline const_iterator operator++(int)
- {
- const_iterator result(*this);
- ++*this;
- return result;
- }
- // pre-increment
- inline const_iterator& operator++()
- {
- probe();
- return *this;
- }
- inline const value_type& operator*() const { return *get_cur(); }
- inline const value_type* operator->() const { return get_cur(); }
- inline bool operator == (const const_iterator& b) const { return (m_pTable == b.m_pTable) && (m_index == b.m_index); }
- inline bool operator != (const const_iterator& b) const { return !(*this == b); }
- inline bool operator == (const iterator& b) const { return (m_pTable == b.m_pTable) && (m_index == b.m_index); }
- inline bool operator != (const iterator& b) const { return !(*this == b); }
- private:
- const hash_map_type* m_pTable;
- uint32_t m_index;
- inline const value_type* get_cur() const
- {
- assert(m_pTable && (m_index < m_pTable->m_values.size()));
- assert(m_pTable->get_node_state(m_index) == cStateValid);
- return &m_pTable->get_node(m_index);
- }
- inline void probe()
- {
- assert(m_pTable);
- m_index = m_pTable->find_next(m_index);
- }
- };
- inline const_iterator begin() const
- {
- if (!m_num_valid)
- return end();
- return const_iterator(*this, find_next(UINT32_MAX));
- }
- inline const_iterator end() const
- {
- return const_iterator(*this, m_values.size());
- }
- inline iterator begin()
- {
- if (!m_num_valid)
- return end();
- return iterator(*this, find_next(UINT32_MAX));
- }
- inline iterator end()
- {
- return iterator(*this, m_values.size());
- }
- // insert_result.first will always point to inserted key/value (or the already existing key/value).
- // insert_resutt.second will be true if a new key/value was inserted, or false if the key already existed (in which case first will point to the already existing value).
- typedef std::pair<iterator, bool> insert_result;
- inline insert_result insert(const Key& k, const Value& v = Value())
- {
- insert_result result;
- if (!insert_no_grow(result, k, v))
- {
- grow();
- // This must succeed.
- if (!insert_no_grow(result, k, v))
- {
- fprintf(stderr, "insert() failed");
- abort();
- }
- }
- return result;
- }
- inline insert_result insert(const value_type& v)
- {
- return insert(v.first, v.second);
- }
- inline const_iterator find(const Key& k) const
- {
- return const_iterator(*this, find_index(k));
- }
- inline iterator find(const Key& k)
- {
- return iterator(*this, find_index(k));
- }
- inline bool erase(const Key& k)
- {
- uint32_t i = find_index(k);
- if (i >= m_values.size())
- return false;
- node* pDst = &get_node(i);
- destruct_value_type(pDst);
- pDst->state = cStateInvalid;
- m_num_valid--;
- for (; ; )
- {
- uint32_t r, j = i;
- node* pSrc = pDst;
- do
- {
- if (!i)
- {
- i = m_values.size() - 1;
- pSrc = &get_node(i);
- }
- else
- {
- i--;
- pSrc--;
- }
- if (!pSrc->state)
- return true;
- r = hash_key(pSrc->first);
- } while ((i <= r && r < j) || (r < j && j < i) || (j < i && i <= r));
- move_node(pDst, pSrc);
- pDst = pSrc;
- }
- }
- inline void swap(hash_map_type& other)
- {
- m_values.swap(other.m_values);
- std::swap(m_hash_shift, other.m_hash_shift);
- std::swap(m_num_valid, other.m_num_valid);
- std::swap(m_grow_threshold, other.m_grow_threshold);
- std::swap(m_hasher, other.m_hasher);
- std::swap(m_equals, other.m_equals);
- }
- private:
- struct node : public value_type
- {
- uint8_t state;
- };
- static inline void construct_value_type(value_type* pDst, const Key& k, const Value& v)
- {
- if (BASISU_IS_BITWISE_COPYABLE(Key))
- memcpy(&pDst->first, &k, sizeof(Key));
- else
- scalar_type<Key>::construct(&pDst->first, k);
- if (BASISU_IS_BITWISE_COPYABLE(Value))
- memcpy(&pDst->second, &v, sizeof(Value));
- else
- scalar_type<Value>::construct(&pDst->second, v);
- }
- static inline void construct_value_type(value_type* pDst, const value_type* pSrc)
- {
- if ((BASISU_IS_BITWISE_COPYABLE(Key)) && (BASISU_IS_BITWISE_COPYABLE(Value)))
- {
- memcpy(pDst, pSrc, sizeof(value_type));
- }
- else
- {
- if (BASISU_IS_BITWISE_COPYABLE(Key))
- memcpy(&pDst->first, &pSrc->first, sizeof(Key));
- else
- scalar_type<Key>::construct(&pDst->first, pSrc->first);
- if (BASISU_IS_BITWISE_COPYABLE(Value))
- memcpy(&pDst->second, &pSrc->second, sizeof(Value));
- else
- scalar_type<Value>::construct(&pDst->second, pSrc->second);
- }
- }
- static inline void destruct_value_type(value_type* p)
- {
- scalar_type<Key>::destruct(&p->first);
- scalar_type<Value>::destruct(&p->second);
- }
- // Moves *pSrc to *pDst efficiently.
- // pDst should NOT be constructed on entry.
- static inline void move_node(node* pDst, node* pSrc, bool update_src_state = true)
- {
- assert(!pDst->state);
- if (BASISU_IS_BITWISE_COPYABLE_OR_MOVABLE(Key) && BASISU_IS_BITWISE_COPYABLE_OR_MOVABLE(Value))
- {
- memcpy(pDst, pSrc, sizeof(node));
- }
- else
- {
- if (BASISU_IS_BITWISE_COPYABLE_OR_MOVABLE(Key))
- memcpy(&pDst->first, &pSrc->first, sizeof(Key));
- else
- {
- scalar_type<Key>::construct(&pDst->first, pSrc->first);
- scalar_type<Key>::destruct(&pSrc->first);
- }
- if (BASISU_IS_BITWISE_COPYABLE_OR_MOVABLE(Value))
- memcpy(&pDst->second, &pSrc->second, sizeof(Value));
- else
- {
- scalar_type<Value>::construct(&pDst->second, pSrc->second);
- scalar_type<Value>::destruct(&pSrc->second);
- }
- pDst->state = cStateValid;
- }
- if (update_src_state)
- pSrc->state = cStateInvalid;
- }
- struct raw_node
- {
- inline raw_node()
- {
- node* p = reinterpret_cast<node*>(this);
- p->state = cStateInvalid;
- }
- inline ~raw_node()
- {
- node* p = reinterpret_cast<node*>(this);
- if (p->state)
- hash_map_type::destruct_value_type(p);
- }
- inline raw_node(const raw_node& other)
- {
- node* pDst = reinterpret_cast<node*>(this);
- const node* pSrc = reinterpret_cast<const node*>(&other);
- if (pSrc->state)
- {
- hash_map_type::construct_value_type(pDst, pSrc);
- pDst->state = cStateValid;
- }
- else
- pDst->state = cStateInvalid;
- }
- inline raw_node& operator= (const raw_node& rhs)
- {
- if (this == &rhs)
- return *this;
- node* pDst = reinterpret_cast<node*>(this);
- const node* pSrc = reinterpret_cast<const node*>(&rhs);
- if (pSrc->state)
- {
- if (pDst->state)
- {
- pDst->first = pSrc->first;
- pDst->second = pSrc->second;
- }
- else
- {
- hash_map_type::construct_value_type(pDst, pSrc);
- pDst->state = cStateValid;
- }
- }
- else if (pDst->state)
- {
- hash_map_type::destruct_value_type(pDst);
- pDst->state = cStateInvalid;
- }
- return *this;
- }
- uint8_t m_bits[sizeof(node)];
- };
- typedef basisu::vector<raw_node> node_vector;
- node_vector m_values;
- uint32_t m_hash_shift;
- Hasher m_hasher;
- Equals m_equals;
- uint32_t m_num_valid;
- uint32_t m_grow_threshold;
- inline uint32_t hash_key(const Key& k) const
- {
- assert((1U << (32U - m_hash_shift)) == m_values.size());
- uint32_t hash = static_cast<uint32_t>(m_hasher(k));
- // Fibonacci hashing
- hash = (2654435769U * hash) >> m_hash_shift;
- assert(hash < m_values.size());
- return hash;
- }
- inline const node& get_node(uint32_t index) const
- {
- return *reinterpret_cast<const node*>(&m_values[index]);
- }
- inline node& get_node(uint32_t index)
- {
- return *reinterpret_cast<node*>(&m_values[index]);
- }
- inline state get_node_state(uint32_t index) const
- {
- return static_cast<state>(get_node(index).state);
- }
- inline void set_node_state(uint32_t index, bool valid)
- {
- get_node(index).state = valid;
- }
- inline void grow()
- {
- uint64_t n = m_values.size() * 3ULL; // was * 2
-
- if (!helpers::is_power_of_2(n))
- n = helpers::next_pow2(n);
- if (n > 0x80000000UL)
- n = 0x80000000UL;
- rehash(helpers::maximum<uint32_t>(cMinHashSize, (uint32_t)n));
- }
- inline void rehash(uint32_t new_hash_size)
- {
- assert(new_hash_size >= m_num_valid);
- assert(helpers::is_power_of_2(new_hash_size));
- if ((new_hash_size < m_num_valid) || (new_hash_size == m_values.size()))
- return;
- hash_map new_map;
- new_map.m_values.resize(new_hash_size);
- new_map.m_hash_shift = 32U - helpers::floor_log2i(new_hash_size);
- assert(new_hash_size == (1U << (32U - new_map.m_hash_shift)));
- new_map.m_grow_threshold = UINT_MAX;
- node* pNode = reinterpret_cast<node*>(m_values.begin());
- node* pNode_end = pNode + m_values.size();
- while (pNode != pNode_end)
- {
- if (pNode->state)
- {
- new_map.move_into(pNode);
- if (new_map.m_num_valid == m_num_valid)
- break;
- }
- pNode++;
- }
- new_map.m_grow_threshold = (new_hash_size + 1U) >> 1U;
- m_values.clear_no_destruction();
- m_hash_shift = 32;
- swap(new_map);
- }
- inline uint32_t find_next(uint32_t index) const
- {
- index++;
- if (index >= m_values.size())
- return index;
- const node* pNode = &get_node(index);
- for (; ; )
- {
- if (pNode->state)
- break;
- if (++index >= m_values.size())
- break;
- pNode++;
- }
- return index;
- }
- inline uint32_t find_index(const Key& k) const
- {
- if (m_num_valid)
- {
- uint32_t index = hash_key(k);
- const node* pNode = &get_node(index);
- if (pNode->state)
- {
- if (m_equals(pNode->first, k))
- return index;
- const uint32_t orig_index = index;
- for (; ; )
- {
- if (!index)
- {
- index = m_values.size() - 1;
- pNode = &get_node(index);
- }
- else
- {
- index--;
- pNode--;
- }
- if (index == orig_index)
- break;
- if (!pNode->state)
- break;
- if (m_equals(pNode->first, k))
- return index;
- }
- }
- }
- return m_values.size();
- }
- inline bool insert_no_grow(insert_result& result, const Key& k, const Value& v = Value())
- {
- if (!m_values.size())
- return false;
- uint32_t index = hash_key(k);
- node* pNode = &get_node(index);
- if (pNode->state)
- {
- if (m_equals(pNode->first, k))
- {
- result.first = iterator(*this, index);
- result.second = false;
- return true;
- }
- const uint32_t orig_index = index;
- for (; ; )
- {
- if (!index)
- {
- index = m_values.size() - 1;
- pNode = &get_node(index);
- }
- else
- {
- index--;
- pNode--;
- }
- if (orig_index == index)
- return false;
- if (!pNode->state)
- break;
- if (m_equals(pNode->first, k))
- {
- result.first = iterator(*this, index);
- result.second = false;
- return true;
- }
- }
- }
- if (m_num_valid >= m_grow_threshold)
- return false;
- construct_value_type(pNode, k, v);
- pNode->state = cStateValid;
- m_num_valid++;
- assert(m_num_valid <= m_values.size());
- result.first = iterator(*this, index);
- result.second = true;
- return true;
- }
- inline void move_into(node* pNode)
- {
- uint32_t index = hash_key(pNode->first);
- node* pDst_node = &get_node(index);
- if (pDst_node->state)
- {
- const uint32_t orig_index = index;
- for (; ; )
- {
- if (!index)
- {
- index = m_values.size() - 1;
- pDst_node = &get_node(index);
- }
- else
- {
- index--;
- pDst_node--;
- }
- if (index == orig_index)
- {
- assert(false);
- return;
- }
- if (!pDst_node->state)
- break;
- }
- }
- move_node(pDst_node, pNode, false);
- m_num_valid++;
- }
- };
- template<typename Key, typename Value, typename Hasher, typename Equals>
- struct bitwise_movable< hash_map<Key, Value, Hasher, Equals> > { enum { cFlag = true }; };
-
- #if BASISU_HASHMAP_TEST
- extern void hash_map_test();
- #endif
-
- } // namespace basisu
- namespace std
- {
- template<typename T>
- inline void swap(basisu::vector<T>& a, basisu::vector<T>& b)
- {
- a.swap(b);
- }
- template<typename Key, typename Value, typename Hasher, typename Equals>
- inline void swap(basisu::hash_map<Key, Value, Hasher, Equals>& a, basisu::hash_map<Key, Value, Hasher, Equals>& b)
- {
- a.swap(b);
- }
- } // namespace std
|