SDL_atomic.h 18 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495
  1. /*
  2. Simple DirectMedia Layer
  3. Copyright (C) 1997-2024 Sam Lantinga <[email protected]>
  4. This software is provided 'as-is', without any express or implied
  5. warranty. In no event will the authors be held liable for any damages
  6. arising from the use of this software.
  7. Permission is granted to anyone to use this software for any purpose,
  8. including commercial applications, and to alter it and redistribute it
  9. freely, subject to the following restrictions:
  10. 1. The origin of this software must not be misrepresented; you must not
  11. claim that you wrote the original software. If you use this software
  12. in a product, an acknowledgment in the product documentation would be
  13. appreciated but is not required.
  14. 2. Altered source versions must be plainly marked as such, and must not be
  15. misrepresented as being the original software.
  16. 3. This notice may not be removed or altered from any source distribution.
  17. */
  18. /**
  19. * \file SDL_atomic.h
  20. *
  21. * Atomic operations.
  22. *
  23. * IMPORTANT:
  24. * If you are not an expert in concurrent lockless programming, you should
  25. * not be using any functions in this file. You should be protecting your
  26. * data structures with full mutexes instead.
  27. *
  28. * Seriously, here be dragons!
  29. * ^^^^^^^^^^^^^^^^^^^^^^^^^^^
  30. *
  31. * You can find out a little more about lockless programming and the
  32. * subtle issues that can arise here:
  33. * https://learn.microsoft.com/en-us/windows/win32/dxtecharts/lockless-programming
  34. *
  35. * There's also lots of good information here:
  36. * http://www.1024cores.net/home/lock-free-algorithms
  37. * http://preshing.com/
  38. *
  39. * These operations may or may not actually be implemented using
  40. * processor specific atomic operations. When possible they are
  41. * implemented as true processor specific atomic operations. When that
  42. * is not possible the are implemented using locks that *do* use the
  43. * available atomic operations.
  44. *
  45. * All of the atomic operations that modify memory are full memory barriers.
  46. */
  47. #ifndef SDL_atomic_h_
  48. #define SDL_atomic_h_
  49. #include <SDL3/SDL_stdinc.h>
  50. #include <SDL3/SDL_platform_defines.h>
  51. #include <SDL3/SDL_begin_code.h>
  52. /* Set up for C function definitions, even when using C++ */
  53. #ifdef __cplusplus
  54. extern "C" {
  55. #endif
  56. /**
  57. * An atomic spinlock.
  58. *
  59. * The atomic locks are efficient spinlocks using CPU instructions,
  60. * but are vulnerable to starvation and can spin forever if a thread
  61. * holding a lock has been terminated. For this reason you should
  62. * minimize the code executed inside an atomic lock and never do
  63. * expensive things like API or system calls while holding them.
  64. *
  65. * They are also vulnerable to starvation if the thread holding
  66. * the lock is lower priority than other threads and doesn't get
  67. * scheduled. In general you should use mutexes instead, since
  68. * they have better performance and contention behavior.
  69. *
  70. * The atomic locks are not safe to lock recursively.
  71. *
  72. * Porting Note:
  73. * The spin lock functions and type are required and can not be
  74. * emulated because they are used in the atomic emulation code.
  75. */
  76. typedef int SDL_SpinLock;
  77. /**
  78. * Try to lock a spin lock by setting it to a non-zero value.
  79. *
  80. * ***Please note that spinlocks are dangerous if you don't know what you're
  81. * doing. Please be careful using any sort of spinlock!***
  82. *
  83. * \param lock a pointer to a lock variable
  84. * \returns SDL_TRUE if the lock succeeded, SDL_FALSE if the lock is already
  85. * held.
  86. *
  87. * \since This function is available since SDL 3.0.0.
  88. *
  89. * \sa SDL_LockSpinlock
  90. * \sa SDL_UnlockSpinlock
  91. */
  92. extern DECLSPEC SDL_bool SDLCALL SDL_TryLockSpinlock(SDL_SpinLock *lock);
  93. /**
  94. * Lock a spin lock by setting it to a non-zero value.
  95. *
  96. * ***Please note that spinlocks are dangerous if you don't know what you're
  97. * doing. Please be careful using any sort of spinlock!***
  98. *
  99. * \param lock a pointer to a lock variable
  100. *
  101. * \since This function is available since SDL 3.0.0.
  102. *
  103. * \sa SDL_TryLockSpinlock
  104. * \sa SDL_UnlockSpinlock
  105. */
  106. extern DECLSPEC void SDLCALL SDL_LockSpinlock(SDL_SpinLock *lock);
  107. /**
  108. * Unlock a spin lock by setting it to 0.
  109. *
  110. * Always returns immediately.
  111. *
  112. * ***Please note that spinlocks are dangerous if you don't know what you're
  113. * doing. Please be careful using any sort of spinlock!***
  114. *
  115. * \param lock a pointer to a lock variable
  116. *
  117. * \since This function is available since SDL 3.0.0.
  118. *
  119. * \sa SDL_LockSpinlock
  120. * \sa SDL_TryLockSpinlock
  121. */
  122. extern DECLSPEC void SDLCALL SDL_UnlockSpinlock(SDL_SpinLock *lock);
  123. #ifdef SDL_WIKI_DOCUMENTATION_SECTION
  124. /**
  125. * Mark a compiler barrier.
  126. *
  127. * A compiler barrier prevents the compiler from reordering reads and writes
  128. * to globally visible variables across the call.
  129. *
  130. * This macro only prevents the compiler from reordering reads and writes, it
  131. * does not prevent the CPU from reordering reads and writes. However, all of
  132. * the atomic operations that modify memory are full memory barriers.
  133. *
  134. * \threadsafety Obviously this macro is safe to use from any thread at any
  135. * time, but if you find yourself needing this, you are probably
  136. * dealing with some very sensitive code; be careful!
  137. *
  138. * \since This macro is available since SDL 3.0.0.
  139. */
  140. #define SDL_CompilerBarrier() DoCompilerSpecificReadWriteBarrier()
  141. #elif defined(_MSC_VER) && (_MSC_VER > 1200) && !defined(__clang__)
  142. void _ReadWriteBarrier(void);
  143. #pragma intrinsic(_ReadWriteBarrier)
  144. #define SDL_CompilerBarrier() _ReadWriteBarrier()
  145. #elif (defined(__GNUC__) && !defined(SDL_PLATFORM_EMSCRIPTEN)) || (defined(__SUNPRO_C) && (__SUNPRO_C >= 0x5120))
  146. /* This is correct for all CPUs when using GCC or Solaris Studio 12.1+. */
  147. #define SDL_CompilerBarrier() __asm__ __volatile__ ("" : : : "memory")
  148. #elif defined(__WATCOMC__)
  149. extern __inline void SDL_CompilerBarrier(void);
  150. #pragma aux SDL_CompilerBarrier = "" parm [] modify exact [];
  151. #else
  152. #define SDL_CompilerBarrier() \
  153. { SDL_SpinLock _tmp = 0; SDL_LockSpinlock(&_tmp); SDL_UnlockSpinlock(&_tmp); }
  154. #endif
  155. /**
  156. * Insert a memory release barrier.
  157. *
  158. * Memory barriers are designed to prevent reads and writes from being
  159. * reordered by the compiler and being seen out of order on multi-core CPUs.
  160. *
  161. * A typical pattern would be for thread A to write some data and a flag, and
  162. * for thread B to read the flag and get the data. In this case you would
  163. * insert a release barrier between writing the data and the flag,
  164. * guaranteeing that the data write completes no later than the flag is
  165. * written, and you would insert an acquire barrier between reading the flag
  166. * and reading the data, to ensure that all the reads associated with the flag
  167. * have completed.
  168. *
  169. * In this pattern you should always see a release barrier paired with an
  170. * acquire barrier and you should gate the data reads/writes with a single
  171. * flag variable.
  172. *
  173. * For more information on these semantics, take a look at the blog post:
  174. * http://preshing.com/20120913/acquire-and-release-semantics
  175. *
  176. * \threadsafety Obviously this macro is safe to use from any thread at any
  177. * time, but if you find yourself needing this, you are probably
  178. * dealing with some very sensitive code; be careful!
  179. *
  180. * \since This function is available since SDL 3.0.0.
  181. */
  182. extern DECLSPEC void SDLCALL SDL_MemoryBarrierReleaseFunction(void);
  183. /**
  184. * Insert a memory acquire barrier.
  185. *
  186. * Please refer to SDL_MemoryBarrierReleaseFunction for the details!
  187. *
  188. * \threadsafety Obviously this function is safe to use from any thread at any
  189. * time, but if you find yourself needing this, you are probably
  190. * dealing with some very sensitive code; be careful!
  191. *
  192. * \since This function is available since SDL 3.0.0.
  193. *
  194. * \sa SDL_MemoryBarrierReleaseFunction
  195. */
  196. extern DECLSPEC void SDLCALL SDL_MemoryBarrierAcquireFunction(void);
  197. /* !!! FIXME: this should have documentation! */
  198. #if defined(__GNUC__) && (defined(__powerpc__) || defined(__ppc__))
  199. #define SDL_MemoryBarrierRelease() __asm__ __volatile__ ("lwsync" : : : "memory")
  200. #define SDL_MemoryBarrierAcquire() __asm__ __volatile__ ("lwsync" : : : "memory")
  201. #elif defined(__GNUC__) && defined(__aarch64__)
  202. #define SDL_MemoryBarrierRelease() __asm__ __volatile__ ("dmb ish" : : : "memory")
  203. #define SDL_MemoryBarrierAcquire() __asm__ __volatile__ ("dmb ish" : : : "memory")
  204. #elif defined(__GNUC__) && defined(__arm__)
  205. #if 0 /* defined(SDL_PLATFORM_LINUX) || defined(SDL_PLATFORM_ANDROID) */
  206. /* Information from:
  207. https://chromium.googlesource.com/chromium/chromium/+/trunk/base/atomicops_internals_arm_gcc.h#19
  208. The Linux kernel provides a helper function which provides the right code for a memory barrier,
  209. hard-coded at address 0xffff0fa0
  210. */
  211. typedef void (*SDL_KernelMemoryBarrierFunc)();
  212. #define SDL_MemoryBarrierRelease() ((SDL_KernelMemoryBarrierFunc)0xffff0fa0)()
  213. #define SDL_MemoryBarrierAcquire() ((SDL_KernelMemoryBarrierFunc)0xffff0fa0)()
  214. #else
  215. #if defined(__ARM_ARCH_7__) || defined(__ARM_ARCH_7A__) || defined(__ARM_ARCH_7EM__) || defined(__ARM_ARCH_7R__) || defined(__ARM_ARCH_7M__) || defined(__ARM_ARCH_7S__) || defined(__ARM_ARCH_8A__)
  216. #define SDL_MemoryBarrierRelease() __asm__ __volatile__ ("dmb ish" : : : "memory")
  217. #define SDL_MemoryBarrierAcquire() __asm__ __volatile__ ("dmb ish" : : : "memory")
  218. #elif defined(__ARM_ARCH_6__) || defined(__ARM_ARCH_6J__) || defined(__ARM_ARCH_6K__) || defined(__ARM_ARCH_6T2__) || defined(__ARM_ARCH_6Z__) || defined(__ARM_ARCH_6ZK__)
  219. #ifdef __thumb__
  220. /* The mcr instruction isn't available in thumb mode, use real functions */
  221. #define SDL_MEMORY_BARRIER_USES_FUNCTION
  222. #define SDL_MemoryBarrierRelease() SDL_MemoryBarrierReleaseFunction()
  223. #define SDL_MemoryBarrierAcquire() SDL_MemoryBarrierAcquireFunction()
  224. #else
  225. #define SDL_MemoryBarrierRelease() __asm__ __volatile__ ("mcr p15, 0, %0, c7, c10, 5" : : "r"(0) : "memory")
  226. #define SDL_MemoryBarrierAcquire() __asm__ __volatile__ ("mcr p15, 0, %0, c7, c10, 5" : : "r"(0) : "memory")
  227. #endif /* __thumb__ */
  228. #else
  229. #define SDL_MemoryBarrierRelease() __asm__ __volatile__ ("" : : : "memory")
  230. #define SDL_MemoryBarrierAcquire() __asm__ __volatile__ ("" : : : "memory")
  231. #endif /* SDL_PLATFORM_LINUX || SDL_PLATFORM_ANDROID */
  232. #endif /* __GNUC__ && __arm__ */
  233. #else
  234. #if (defined(__SUNPRO_C) && (__SUNPRO_C >= 0x5120))
  235. /* This is correct for all CPUs on Solaris when using Solaris Studio 12.1+. */
  236. #include <mbarrier.h>
  237. #define SDL_MemoryBarrierRelease() __machine_rel_barrier()
  238. #define SDL_MemoryBarrierAcquire() __machine_acq_barrier()
  239. #else
  240. /* This is correct for the x86 and x64 CPUs, and we'll expand this over time. */
  241. #define SDL_MemoryBarrierRelease() SDL_CompilerBarrier()
  242. #define SDL_MemoryBarrierAcquire() SDL_CompilerBarrier()
  243. #endif
  244. #endif
  245. /* "REP NOP" is PAUSE, coded for tools that don't know it by that name. */
  246. #ifdef SDL_WIKI_DOCUMENTATION_SECTION
  247. /**
  248. * A macro to insert a CPU-specific "pause" instruction into the program.
  249. *
  250. * This can be useful in busy-wait loops, as it serves as a hint to the CPU
  251. * as to the program's intent; some CPUs can use this to do more efficient
  252. * processing. On some platforms, this doesn't do anything, so using this
  253. * macro might just be a harmless no-op.
  254. *
  255. * Note that if you are busy-waiting, there are often more-efficient
  256. * approaches with other synchronization primitives: mutexes, semaphores,
  257. * condition variables, etc.
  258. *
  259. * \threadsafety This macro is safe to use from any thread.
  260. *
  261. * \since This macro is available since SDL 3.0.0.
  262. */
  263. #define SDL_CPUPauseInstruction() DoACPUPauseInACompilerAndArchitectureSpecificWay
  264. #elif (defined(__GNUC__) || defined(__clang__)) && (defined(__i386__) || defined(__x86_64__))
  265. #define SDL_CPUPauseInstruction() __asm__ __volatile__("pause\n") /* Some assemblers can't do REP NOP, so go with PAUSE. */
  266. #elif (defined(__arm__) && defined(__ARM_ARCH) && __ARM_ARCH >= 7) || defined(__aarch64__)
  267. #define SDL_CPUPauseInstruction() __asm__ __volatile__("yield" ::: "memory")
  268. #elif (defined(__powerpc__) || defined(__powerpc64__))
  269. #define SDL_CPUPauseInstruction() __asm__ __volatile__("or 27,27,27");
  270. #elif (defined(__riscv) && __riscv_xlen == 64)
  271. #define SDL_CPUPauseInstruction() __asm__ __volatile__(".insn i 0x0F, 0, x0, x0, 0x010");
  272. #elif defined(_MSC_VER) && (defined(_M_IX86) || defined(_M_X64))
  273. #define SDL_CPUPauseInstruction() _mm_pause() /* this is actually "rep nop" and not a SIMD instruction. No inline asm in MSVC x86-64! */
  274. #elif defined(_MSC_VER) && (defined(_M_ARM) || defined(_M_ARM64))
  275. #define SDL_CPUPauseInstruction() __yield()
  276. #elif defined(__WATCOMC__) && defined(__386__)
  277. extern __inline void SDL_CPUPauseInstruction(void);
  278. #pragma aux SDL_CPUPauseInstruction = ".686p" ".xmm2" "pause"
  279. #else
  280. #define SDL_CPUPauseInstruction()
  281. #endif
  282. /**
  283. * A type representing an atomic integer value.
  284. *
  285. * This can be used to manage a value that is synchronized across multiple
  286. * CPUs without a race condition; when an app sets a value with SDL_AtomicSet
  287. * all other threads, regardless of the CPU it is running on, will see that
  288. * value when retrieved with SDL_AtomicGet, regardless of CPU caches, etc.
  289. *
  290. * This is also useful for atomic compare-and-swap operations: a thread
  291. * can change the value as long as its current value matches expectations.
  292. * When done in a loop, one can guarantee data consistency across threads
  293. * without a lock (but the usual warnings apply: if you don't know what
  294. * you're doing, or you don't do it carefully, you can confidently cause
  295. * any number of disasters with this, so in most cases, you _should_
  296. * use a mutex instead of this!).
  297. *
  298. * This is a struct so people don't accidentally use numeric operations on
  299. * it directly. You have to use SDL_Atomic* functions.
  300. *
  301. * \since This struct is available since SDL 3.0.0.
  302. *
  303. * \sa SDL_AtomicCompareAndSwap
  304. * \sa SDL_AtomicGet
  305. * \sa SDL_AtomicSet
  306. * \sa SDL_AtomicAdd
  307. */
  308. typedef struct SDL_AtomicInt { int value; } SDL_AtomicInt;
  309. /**
  310. * Set an atomic variable to a new value if it is currently an old value.
  311. *
  312. * ***Note: If you don't know what this function is for, you shouldn't use
  313. * it!***
  314. *
  315. * \param a a pointer to an SDL_AtomicInt variable to be modified
  316. * \param oldval the old value
  317. * \param newval the new value
  318. * \returns SDL_TRUE if the atomic variable was set, SDL_FALSE otherwise.
  319. *
  320. * \since This function is available since SDL 3.0.0.
  321. *
  322. * \sa SDL_AtomicCompareAndSwapPointer
  323. */
  324. extern DECLSPEC SDL_bool SDLCALL SDL_AtomicCompareAndSwap(SDL_AtomicInt *a, int oldval, int newval);
  325. /**
  326. * Set an atomic variable to a value.
  327. *
  328. * This function also acts as a full memory barrier.
  329. *
  330. * ***Note: If you don't know what this function is for, you shouldn't use
  331. * it!***
  332. *
  333. * \param a a pointer to an SDL_AtomicInt variable to be modified
  334. * \param v the desired value
  335. * \returns the previous value of the atomic variable.
  336. *
  337. * \since This function is available since SDL 3.0.0.
  338. *
  339. * \sa SDL_AtomicGet
  340. */
  341. extern DECLSPEC int SDLCALL SDL_AtomicSet(SDL_AtomicInt *a, int v);
  342. /**
  343. * Get the value of an atomic variable.
  344. *
  345. * ***Note: If you don't know what this function is for, you shouldn't use
  346. * it!***
  347. *
  348. * \param a a pointer to an SDL_AtomicInt variable
  349. * \returns the current value of an atomic variable.
  350. *
  351. * \since This function is available since SDL 3.0.0.
  352. *
  353. * \sa SDL_AtomicSet
  354. */
  355. extern DECLSPEC int SDLCALL SDL_AtomicGet(SDL_AtomicInt *a);
  356. /**
  357. * Add to an atomic variable.
  358. *
  359. * This function also acts as a full memory barrier.
  360. *
  361. * ***Note: If you don't know what this function is for, you shouldn't use
  362. * it!***
  363. *
  364. * \param a a pointer to an SDL_AtomicInt variable to be modified
  365. * \param v the desired value to add
  366. * \returns the previous value of the atomic variable.
  367. *
  368. * \since This function is available since SDL 3.0.0.
  369. *
  370. * \sa SDL_AtomicDecRef
  371. * \sa SDL_AtomicIncRef
  372. */
  373. extern DECLSPEC int SDLCALL SDL_AtomicAdd(SDL_AtomicInt *a, int v);
  374. #ifndef SDL_AtomicIncRef
  375. /**
  376. * Increment an atomic variable used as a reference count.
  377. *
  378. * ***Note: If you don't know what this macro is for, you shouldn't use it!***
  379. *
  380. * \param a a pointer to an SDL_AtomicInt to increment.
  381. * \returns the previous value of the atomic variable.
  382. *
  383. * \since This macro is available since SDL 3.0.0.
  384. *
  385. * \sa SDL_AtomicDecRef
  386. */
  387. #define SDL_AtomicIncRef(a) SDL_AtomicAdd(a, 1)
  388. #endif
  389. #ifndef SDL_AtomicDecRef
  390. /**
  391. * Decrement an atomic variable used as a reference count.
  392. *
  393. * ***Note: If you don't know what this macro is for, you shouldn't use it!***
  394. *
  395. * \param a a pointer to an SDL_AtomicInt to increment.
  396. * \returns SDL_TRUE if the variable reached zero after decrementing,
  397. * SDL_FALSE otherwise
  398. *
  399. * \since This macro is available since SDL 3.0.0.
  400. *
  401. * \sa SDL_AtomicIncRef
  402. */
  403. #define SDL_AtomicDecRef(a) (SDL_AtomicAdd(a, -1) == 1)
  404. #endif
  405. /**
  406. * Set a pointer to a new value if it is currently an old value.
  407. *
  408. * ***Note: If you don't know what this function is for, you shouldn't use
  409. * it!***
  410. *
  411. * \param a a pointer to a pointer
  412. * \param oldval the old pointer value
  413. * \param newval the new pointer value
  414. * \returns SDL_TRUE if the pointer was set, SDL_FALSE otherwise.
  415. *
  416. * \since This function is available since SDL 3.0.0.
  417. *
  418. * \sa SDL_AtomicCompareAndSwap
  419. * \sa SDL_AtomicGetPtr
  420. * \sa SDL_AtomicSetPtr
  421. */
  422. extern DECLSPEC SDL_bool SDLCALL SDL_AtomicCompareAndSwapPointer(void **a, void *oldval, void *newval);
  423. /**
  424. * Set a pointer to a value atomically.
  425. *
  426. * ***Note: If you don't know what this function is for, you shouldn't use
  427. * it!***
  428. *
  429. * \param a a pointer to a pointer
  430. * \param v the desired pointer value
  431. * \returns the previous value of the pointer.
  432. *
  433. * \since This function is available since SDL 3.0.0.
  434. *
  435. * \sa SDL_AtomicCompareAndSwapPointer
  436. * \sa SDL_AtomicGetPtr
  437. */
  438. extern DECLSPEC void* SDLCALL SDL_AtomicSetPtr(void **a, void* v);
  439. /**
  440. * Get the value of a pointer atomically.
  441. *
  442. * ***Note: If you don't know what this function is for, you shouldn't use
  443. * it!***
  444. *
  445. * \param a a pointer to a pointer
  446. * \returns the current value of a pointer.
  447. *
  448. * \since This function is available since SDL 3.0.0.
  449. *
  450. * \sa SDL_AtomicCompareAndSwapPointer
  451. * \sa SDL_AtomicSetPtr
  452. */
  453. extern DECLSPEC void* SDLCALL SDL_AtomicGetPtr(void **a);
  454. /* Ends C function definitions when using C++ */
  455. #ifdef __cplusplus
  456. }
  457. #endif
  458. #include <SDL3/SDL_close_code.h>
  459. #endif /* SDL_atomic_h_ */