|
@@ -0,0 +1,412 @@
|
|
|
+#load "win32.odin"
|
|
|
+
|
|
|
+assume :: proc(cond: bool) #foreign "llvm.assume"
|
|
|
+
|
|
|
+__debug_trap :: proc() #foreign "llvm.debugtrap"
|
|
|
+__trap :: proc() #foreign "llvm.trap"
|
|
|
+read_cycle_counter :: proc() -> u64 #foreign "llvm.readcyclecounter"
|
|
|
+
|
|
|
+bit_reverse16 :: proc(b: u16) -> u16 #foreign "llvm.bitreverse.i16"
|
|
|
+bit_reverse32 :: proc(b: u32) -> u32 #foreign "llvm.bitreverse.i32"
|
|
|
+bit_reverse64 :: proc(b: u64) -> u64 #foreign "llvm.bitreverse.i64"
|
|
|
+
|
|
|
+byte_swap16 :: proc(b: u16) -> u16 #foreign "llvm.bswap.i16"
|
|
|
+byte_swap32 :: proc(b: u32) -> u32 #foreign "llvm.bswap.i32"
|
|
|
+byte_swap64 :: proc(b: u64) -> u64 #foreign "llvm.bswap.i64"
|
|
|
+
|
|
|
+fmuladd_f32 :: proc(a, b, c: f32) -> f32 #foreign "llvm.fmuladd.f32"
|
|
|
+fmuladd_f64 :: proc(a, b, c: f64) -> f64 #foreign "llvm.fmuladd.f64"
|
|
|
+
|
|
|
+// TODO(bill): make custom heap procedures
|
|
|
+heap_alloc :: proc(len: int) -> rawptr #foreign "malloc"
|
|
|
+heap_dealloc :: proc(ptr: rawptr) #foreign "free"
|
|
|
+
|
|
|
+memory_zero :: proc(data: rawptr, len: int) {
|
|
|
+ d := slice_ptr(data as ^byte, len)
|
|
|
+ for i := 0; i < len; i++ {
|
|
|
+ d[i] = 0
|
|
|
+ }
|
|
|
+}
|
|
|
+
|
|
|
+memory_compare :: proc(dst, src: rawptr, len: int) -> int {
|
|
|
+ s1, s2: ^byte = dst, src
|
|
|
+ for i := 0; i < len; i++ {
|
|
|
+ a := ptr_offset(s1, i)^
|
|
|
+ b := ptr_offset(s2, i)^
|
|
|
+ if a != b {
|
|
|
+ return (a - b) as int
|
|
|
+ }
|
|
|
+ }
|
|
|
+ return 0
|
|
|
+}
|
|
|
+
|
|
|
+memory_copy :: proc(dst, src: rawptr, n: int) #inline {
|
|
|
+ if dst == src {
|
|
|
+ return
|
|
|
+ }
|
|
|
+
|
|
|
+ v128b :: type {4}u32
|
|
|
+ compile_assert(align_of(v128b) == 16)
|
|
|
+
|
|
|
+ d, s: ^byte = dst, src
|
|
|
+
|
|
|
+ for ; s as uint % 16 != 0 && n != 0; n-- {
|
|
|
+ d^ = s^
|
|
|
+ d, s = ptr_offset(d, 1), ptr_offset(s, 1)
|
|
|
+ }
|
|
|
+
|
|
|
+ if d as uint % 16 == 0 {
|
|
|
+ for ; n >= 16; d, s, n = ptr_offset(d, 16), ptr_offset(s, 16), n-16 {
|
|
|
+ (d as ^v128b)^ = (s as ^v128b)^
|
|
|
+ }
|
|
|
+
|
|
|
+ if n&8 != 0 {
|
|
|
+ (d as ^u64)^ = (s as ^u64)^
|
|
|
+ d, s = ptr_offset(d, 8), ptr_offset(s, 8)
|
|
|
+ }
|
|
|
+ if n&4 != 0 {
|
|
|
+ (d as ^u32)^ = (s as ^u32)^;
|
|
|
+ d, s = ptr_offset(d, 4), ptr_offset(s, 4)
|
|
|
+ }
|
|
|
+ if n&2 != 0 {
|
|
|
+ (d as ^u16)^ = (s as ^u16)^
|
|
|
+ d, s = ptr_offset(d, 2), ptr_offset(s, 2)
|
|
|
+ }
|
|
|
+ if n&1 != 0 {
|
|
|
+ d^ = s^
|
|
|
+ d, s = ptr_offset(d, 1), ptr_offset(s, 1)
|
|
|
+ }
|
|
|
+ return;
|
|
|
+ }
|
|
|
+
|
|
|
+ // IMPORTANT NOTE(bill): Little endian only
|
|
|
+ LS :: proc(a, b: u32) -> u32 #inline { return a << b }
|
|
|
+ RS :: proc(a, b: u32) -> u32 #inline { return a >> b }
|
|
|
+ /* NOTE(bill): Big endian version
|
|
|
+ LS :: proc(a, b: u32) -> u32 #inline { return a >> b; }
|
|
|
+ RS :: proc(a, b: u32) -> u32 #inline { return a << b; }
|
|
|
+ */
|
|
|
+
|
|
|
+ w, x: u32
|
|
|
+
|
|
|
+ if d as uint % 4 == 1 {
|
|
|
+ w = (s as ^u32)^
|
|
|
+ d^ = s^; d = ptr_offset(d, 1); s = ptr_offset(s, 1)
|
|
|
+ d^ = s^; d = ptr_offset(d, 1); s = ptr_offset(s, 1)
|
|
|
+ d^ = s^; d = ptr_offset(d, 1); s = ptr_offset(s, 1)
|
|
|
+ n -= 3
|
|
|
+
|
|
|
+ for n > 16 {
|
|
|
+ d32 := d as ^u32
|
|
|
+ s32 := ptr_offset(s, 1) as ^u32
|
|
|
+ x = s32^; d32^ = LS(w, 24) | RS(x, 8)
|
|
|
+ d32, s32 = ptr_offset(d32, 1), ptr_offset(s32, 1)
|
|
|
+ w = s32^; d32^ = LS(x, 24) | RS(w, 8)
|
|
|
+ d32, s32 = ptr_offset(d32, 1), ptr_offset(s32, 1)
|
|
|
+ x = s32^; d32^ = LS(w, 24) | RS(x, 8)
|
|
|
+ d32, s32 = ptr_offset(d32, 1), ptr_offset(s32, 1)
|
|
|
+ w = s32^; d32^ = LS(x, 24) | RS(w, 8)
|
|
|
+ d32, s32 = ptr_offset(d32, 1), ptr_offset(s32, 1)
|
|
|
+
|
|
|
+ d, s, n = ptr_offset(d, 16), ptr_offset(s, 16), n-16
|
|
|
+ }
|
|
|
+
|
|
|
+ } else if d as uint % 4 == 2 {
|
|
|
+ w = (s as ^u32)^
|
|
|
+ d^ = s^; d = ptr_offset(d, 1); s = ptr_offset(s, 1)
|
|
|
+ d^ = s^; d = ptr_offset(d, 1); s = ptr_offset(s, 1)
|
|
|
+ n -= 2
|
|
|
+
|
|
|
+ for n > 17 {
|
|
|
+ d32 := d as ^u32
|
|
|
+ s32 := ptr_offset(s, 2) as ^u32
|
|
|
+ x = s32^; d32^ = LS(w, 16) | RS(x, 16)
|
|
|
+ d32, s32 = ptr_offset(d32, 1), ptr_offset(s32, 1)
|
|
|
+ w = s32^; d32^ = LS(x, 16) | RS(w, 16)
|
|
|
+ d32, s32 = ptr_offset(d32, 1), ptr_offset(s32, 1)
|
|
|
+ x = s32^; d32^ = LS(w, 16) | RS(x, 16)
|
|
|
+ d32, s32 = ptr_offset(d32, 1), ptr_offset(s32, 1)
|
|
|
+ w = s32^; d32^ = LS(x, 16) | RS(w, 16)
|
|
|
+ d32, s32 = ptr_offset(d32, 1), ptr_offset(s32, 1)
|
|
|
+
|
|
|
+ d, s, n = ptr_offset(d, 16), ptr_offset(s, 16), n-16
|
|
|
+ }
|
|
|
+
|
|
|
+ } else if d as uint % 4 == 3 {
|
|
|
+ w = (s as ^u32)^
|
|
|
+ d^ = s^
|
|
|
+ n -= 1
|
|
|
+
|
|
|
+ for n > 18 {
|
|
|
+ d32 := d as ^u32
|
|
|
+ s32 := ptr_offset(s, 3) as ^u32
|
|
|
+ x = s32^; d32^ = LS(w, 8) | RS(x, 24)
|
|
|
+ d32, s32 = ptr_offset(d32, 1), ptr_offset(s32, 1)
|
|
|
+ w = s32^; d32^ = LS(x, 8) | RS(w, 24)
|
|
|
+ d32, s32 = ptr_offset(d32, 1), ptr_offset(s32, 1)
|
|
|
+ x = s32^; d32^ = LS(w, 8) | RS(x, 24)
|
|
|
+ d32, s32 = ptr_offset(d32, 1), ptr_offset(s32, 1)
|
|
|
+ w = s32^; d32^ = LS(x, 8) | RS(w, 24)
|
|
|
+ d32, s32 = ptr_offset(d32, 1), ptr_offset(s32, 1)
|
|
|
+
|
|
|
+ d, s, n = ptr_offset(d, 16), ptr_offset(s, 16), n-16
|
|
|
+ }
|
|
|
+ }
|
|
|
+
|
|
|
+ if n&16 != 0 {
|
|
|
+ (d as ^v128b)^ = (s as ^v128b)^
|
|
|
+ d, s = ptr_offset(d, 16), ptr_offset(s, 16)
|
|
|
+ }
|
|
|
+ if n&8 != 0 {
|
|
|
+ (d as ^u64)^ = (s as ^u64)^
|
|
|
+ d, s = ptr_offset(d, 8), ptr_offset(s, 8)
|
|
|
+ }
|
|
|
+ if n&4 != 0 {
|
|
|
+ (d as ^u32)^ = (s as ^u32)^;
|
|
|
+ d, s = ptr_offset(d, 4), ptr_offset(s, 4)
|
|
|
+ }
|
|
|
+ if n&2 != 0 {
|
|
|
+ (d as ^u16)^ = (s as ^u16)^
|
|
|
+ d, s = ptr_offset(d, 2), ptr_offset(s, 2)
|
|
|
+ }
|
|
|
+ if n&1 != 0 {
|
|
|
+ d^ = s^
|
|
|
+ }
|
|
|
+}
|
|
|
+
|
|
|
+memory_move :: proc(dst, src: rawptr, n: int) #inline {
|
|
|
+ d, s: ^byte = dst, src
|
|
|
+ if d == s {
|
|
|
+ return
|
|
|
+ }
|
|
|
+ if d >= ptr_offset(s, n) || ptr_offset(d, n) <= s {
|
|
|
+ memory_copy(d, s, n)
|
|
|
+ return
|
|
|
+ }
|
|
|
+
|
|
|
+ // TODO(bill): Vectorize the shit out of this
|
|
|
+ if d < s {
|
|
|
+ if s as int % size_of(int) == d as int % size_of(int) {
|
|
|
+ for d as int % size_of(int) != 0 {
|
|
|
+ if n == 0 {
|
|
|
+ return
|
|
|
+ }
|
|
|
+ n--
|
|
|
+ d^ = s^
|
|
|
+ d, s = ptr_offset(d, 1), ptr_offset(s, 1)
|
|
|
+ }
|
|
|
+ di, si := d as ^int, s as ^int
|
|
|
+ for n >= size_of(int) {
|
|
|
+ di^ = si^
|
|
|
+ di, si = ptr_offset(di, 1), ptr_offset(si, 1)
|
|
|
+ n -= size_of(int)
|
|
|
+ }
|
|
|
+ }
|
|
|
+ for ; n > 0; n-- {
|
|
|
+ d^ = s^
|
|
|
+ d, s = ptr_offset(d, 1), ptr_offset(s, 1)
|
|
|
+ }
|
|
|
+ } else {
|
|
|
+ if s as int % size_of(int) == d as int % size_of(int) {
|
|
|
+ for ptr_offset(d, n) as int % size_of(int) != 0 {
|
|
|
+ if n == 0 {
|
|
|
+ return
|
|
|
+ }
|
|
|
+ n--
|
|
|
+ d^ = s^
|
|
|
+ d, s = ptr_offset(d, 1), ptr_offset(s, 1)
|
|
|
+ }
|
|
|
+ for n >= size_of(int) {
|
|
|
+ n -= size_of(int)
|
|
|
+ di := ptr_offset(d, n) as ^int
|
|
|
+ si := ptr_offset(s, n) as ^int
|
|
|
+ di^ = si^
|
|
|
+ }
|
|
|
+ for ; n > 0; n-- {
|
|
|
+ d^ = s^
|
|
|
+ d, s = ptr_offset(d, 1), ptr_offset(s, 1)
|
|
|
+ }
|
|
|
+ }
|
|
|
+ for n > 0 {
|
|
|
+ n--
|
|
|
+ dn := ptr_offset(d, n)
|
|
|
+ sn := ptr_offset(s, n)
|
|
|
+ dn^ = sn^
|
|
|
+ }
|
|
|
+ }
|
|
|
+}
|
|
|
+
|
|
|
+__string_eq :: proc(a, b: string) -> bool {
|
|
|
+ if len(a) != len(b) {
|
|
|
+ return false
|
|
|
+ }
|
|
|
+ if ^a[0] == ^b[0] {
|
|
|
+ return true
|
|
|
+ }
|
|
|
+ return memory_compare(^a[0], ^b[0], len(a)) == 0
|
|
|
+}
|
|
|
+
|
|
|
+__string_cmp :: proc(a, b : string) -> int {
|
|
|
+ min_len := len(a)
|
|
|
+ if len(b) < min_len {
|
|
|
+ min_len = len(b)
|
|
|
+ }
|
|
|
+ for i := 0; i < min_len; i++ {
|
|
|
+ x := a[i]
|
|
|
+ y := b[i]
|
|
|
+ if x < y {
|
|
|
+ return -1
|
|
|
+ } else if x > y {
|
|
|
+ return +1
|
|
|
+ }
|
|
|
+ }
|
|
|
+
|
|
|
+ if len(a) < len(b) {
|
|
|
+ return -1
|
|
|
+ } else if len(a) > len(b) {
|
|
|
+ return +1
|
|
|
+ }
|
|
|
+ return 0
|
|
|
+}
|
|
|
+
|
|
|
+__string_ne :: proc(a, b : string) -> bool #inline { return !__string_eq(a, b) }
|
|
|
+__string_lt :: proc(a, b : string) -> bool #inline { return __string_cmp(a, b) < 0 }
|
|
|
+__string_gt :: proc(a, b : string) -> bool #inline { return __string_cmp(a, b) > 0 }
|
|
|
+__string_le :: proc(a, b : string) -> bool #inline { return __string_cmp(a, b) <= 0 }
|
|
|
+__string_ge :: proc(a, b : string) -> bool #inline { return __string_cmp(a, b) >= 0 }
|
|
|
+
|
|
|
+
|
|
|
+
|
|
|
+
|
|
|
+Allocation_Mode :: type enum {
|
|
|
+ ALLOC,
|
|
|
+ DEALLOC,
|
|
|
+ DEALLOC_ALL,
|
|
|
+ RESIZE,
|
|
|
+}
|
|
|
+
|
|
|
+
|
|
|
+
|
|
|
+Allocator_Proc :: type proc(allocator_data: rawptr, mode: Allocation_Mode,
|
|
|
+ size, alignment: int,
|
|
|
+ old_memory: rawptr, old_size: int, flags: u64) -> rawptr
|
|
|
+
|
|
|
+Allocator :: type struct {
|
|
|
+ procedure: Allocator_Proc;
|
|
|
+ data: rawptr
|
|
|
+}
|
|
|
+
|
|
|
+
|
|
|
+Context :: type struct {
|
|
|
+ thread_ptr: rawptr
|
|
|
+
|
|
|
+ user_data: rawptr
|
|
|
+ user_index: int
|
|
|
+
|
|
|
+ allocator: Allocator
|
|
|
+}
|
|
|
+
|
|
|
+#thread_local context: Context
|
|
|
+
|
|
|
+DEFAULT_ALIGNMENT :: 2*size_of(int)
|
|
|
+
|
|
|
+
|
|
|
+__check_context :: proc() {
|
|
|
+ if context.allocator.procedure == null {
|
|
|
+ context.allocator = __default_allocator()
|
|
|
+ }
|
|
|
+ if context.thread_ptr == null {
|
|
|
+ // TODO(bill):
|
|
|
+ // context.thread_ptr = current_thread_pointer()
|
|
|
+ }
|
|
|
+}
|
|
|
+
|
|
|
+
|
|
|
+alloc :: proc(size: int) -> rawptr #inline { return alloc_align(size, DEFAULT_ALIGNMENT) }
|
|
|
+
|
|
|
+alloc_align :: proc(size, alignment: int) -> rawptr #inline {
|
|
|
+ __check_context()
|
|
|
+ a := context.allocator
|
|
|
+ return a.procedure(a.data, Allocation_Mode.ALLOC, size, alignment, null, 0, 0)
|
|
|
+}
|
|
|
+
|
|
|
+dealloc :: proc(ptr: rawptr) #inline {
|
|
|
+ __check_context()
|
|
|
+ a := context.allocator
|
|
|
+ _ = a.procedure(a.data, Allocation_Mode.DEALLOC, 0, 0, ptr, 0, 0)
|
|
|
+}
|
|
|
+dealloc_all :: proc(ptr: rawptr) #inline {
|
|
|
+ __check_context()
|
|
|
+ a := context.allocator
|
|
|
+ _ = a.procedure(a.data, Allocation_Mode.DEALLOC_ALL, 0, 0, ptr, 0, 0)
|
|
|
+}
|
|
|
+
|
|
|
+
|
|
|
+resize :: proc(ptr: rawptr, old_size, new_size: int) -> rawptr #inline { return resize_align(ptr, old_size, new_size, DEFAULT_ALIGNMENT) }
|
|
|
+resize_align :: proc(ptr: rawptr, old_size, new_size, alignment: int) -> rawptr #inline {
|
|
|
+ __check_context()
|
|
|
+ a := context.allocator
|
|
|
+ return a.procedure(a.data, Allocation_Mode.RESIZE, new_size, alignment, ptr, old_size, 0)
|
|
|
+}
|
|
|
+
|
|
|
+
|
|
|
+
|
|
|
+default_resize_align :: proc(old_memory: rawptr, old_size, new_size, alignment: int) -> rawptr {
|
|
|
+ if old_memory == null {
|
|
|
+ return alloc_align(new_size, alignment)
|
|
|
+ }
|
|
|
+
|
|
|
+ if new_size == 0 {
|
|
|
+ dealloc(old_memory)
|
|
|
+ return null
|
|
|
+ }
|
|
|
+
|
|
|
+ if new_size == old_size {
|
|
|
+ return old_memory
|
|
|
+ }
|
|
|
+
|
|
|
+ new_memory := alloc_align(new_size, alignment)
|
|
|
+ if new_memory == null {
|
|
|
+ return null
|
|
|
+ }
|
|
|
+
|
|
|
+ memory_copy(new_memory, old_memory, min(old_size, new_size));
|
|
|
+ dealloc(old_memory)
|
|
|
+ return new_memory
|
|
|
+}
|
|
|
+
|
|
|
+
|
|
|
+__default_allocator_proc :: proc(allocator_data: rawptr, mode: Allocation_Mode,
|
|
|
+ size, alignment: int,
|
|
|
+ old_memory: rawptr, old_size: int, flags: u64) -> rawptr {
|
|
|
+ using Allocation_Mode
|
|
|
+ match mode {
|
|
|
+ case ALLOC:
|
|
|
+ return heap_alloc(size)
|
|
|
+ case RESIZE:
|
|
|
+ return default_resize_align(old_memory, old_size, size, alignment)
|
|
|
+ case DEALLOC:
|
|
|
+ heap_dealloc(old_memory)
|
|
|
+ case DEALLOC_ALL:
|
|
|
+ // NOTE(bill): Does nothing
|
|
|
+ }
|
|
|
+
|
|
|
+ return null
|
|
|
+}
|
|
|
+
|
|
|
+__default_allocator :: proc() -> Allocator {
|
|
|
+ return Allocator{
|
|
|
+ __default_allocator_proc,
|
|
|
+ null,
|
|
|
+ }
|
|
|
+}
|
|
|
+
|
|
|
+
|
|
|
+
|
|
|
+
|
|
|
+__assert :: proc(msg: string) {
|
|
|
+ file_write(file_get_standard(File_Standard.ERROR), msg as []byte)
|
|
|
+ // TODO(bill): Which is better?
|
|
|
+ // __trap()
|
|
|
+ __debug_trap()
|
|
|
+}
|