Browse Source

Comment out many mutex guards in `type_(size|align)_of_internal`

gingerBill 2 years ago
parent
commit
7ffffeeccc
2 changed files with 21 additions and 22 deletions
  1. 10 11
      src/thread_pool.cpp
  2. 11 11
      src/types.cpp

+ 10 - 11
src/thread_pool.cpp

@@ -30,7 +30,7 @@ gb_internal void thread_pool_init(ThreadPool *pool, gbAllocator const &a, isize
 	slice_init(&pool->threads, a, thread_count + 1);
 
 	// NOTE: this needs to be initialized before any thread starts
-	pool->running.store(true);
+	pool->running.store(true, std::memory_order_seq_cst);
 
 	// setup the main thread
 	thread_init(pool, &pool->threads[0], 0);
@@ -43,7 +43,7 @@ gb_internal void thread_pool_init(ThreadPool *pool, gbAllocator const &a, isize
 }
 
 gb_internal void thread_pool_destroy(ThreadPool *pool) {
-	pool->running.store(false);
+	pool->running.store(false, std::memory_order_seq_cst);
 
 	for_array_off(i, 1, pool->threads) {
 		Thread *t = &pool->threads[i];
@@ -114,7 +114,7 @@ gb_internal bool thread_pool_add_task(ThreadPool *pool, WorkerTaskProc *proc, vo
 gb_internal void thread_pool_wait(ThreadPool *pool) {
 	WorkerTask task;
 
-	while (pool->tasks_left.load()) {
+	while (pool->tasks_left.load(std::memory_order_acquire)) {
 		// if we've got tasks on our queue, run them
 		while (thread_pool_queue_pop(current_thread, &task)) {
 			task.do_work(task.data);
@@ -126,7 +126,7 @@ gb_internal void thread_pool_wait(ThreadPool *pool) {
 		// This *must* be executed in this order, so the futex wakes immediately
 		// if rem_tasks has changed since we checked last, otherwise the program
 		// will permanently sleep
-		Footex rem_tasks = pool->tasks_left.load();
+		Footex rem_tasks = pool->tasks_left.load(std::memory_order_acquire);
 		if (rem_tasks == 0) {
 			return;
 		}
@@ -141,7 +141,7 @@ gb_internal THREAD_PROC(thread_pool_thread_proc) {
 	ThreadPool *pool = current_thread->pool;
 	// debugf("worker id: %td\n", current_thread->idx);
 
-	while (pool->running.load()) {
+	while (pool->running.load(std::memory_order_seq_cst)) {
 		// If we've got tasks to process, work through them
 		usize finished_tasks = 0;
 		i32 state;
@@ -152,30 +152,29 @@ gb_internal THREAD_PROC(thread_pool_thread_proc) {
 
 			finished_tasks += 1;
 		}
-		if (finished_tasks > 0 && pool->tasks_left.load() == 0) {
+		if (finished_tasks > 0 && pool->tasks_left.load(std::memory_order_acquire) == 0) {
 			futex_signal(&pool->tasks_left);
 		}
 
 		// If there's still work somewhere and we don't have it, steal it
-		if (pool->tasks_left.load()) {
+		if (pool->tasks_left.load(std::memory_order_acquire)) {
 			usize idx = cast(usize)current_thread->idx;
 			for_array(i, pool->threads) {
-				if (pool->tasks_left.load() == 0) {
+				if (pool->tasks_left.load(std::memory_order_acquire) == 0) {
 					break;
 				}
 
 				idx = (idx + 1) % cast(usize)pool->threads.count;
 
 				Thread *thread = &pool->threads.data[idx];
-				WorkerTask task;
+				WorkerTask task, another_task;
 				if (!thread_pool_queue_pop(thread, &task)) {
 					continue;
 				}
-
 				task.do_work(task.data);
 				pool->tasks_left.fetch_sub(1, std::memory_order_release);
 
-				if (pool->tasks_left.load() == 0) {
+				if (pool->tasks_left.load(std::memory_order_acquire) == 0) {
 					futex_signal(&pool->tasks_left);
 				}
 

+ 11 - 11
src/types.cpp

@@ -3406,7 +3406,7 @@ gb_internal i64 type_align_of_internal(Type *t, TypePath *path) {
 	} break;
 
 	case Type_Array: {
-		MUTEX_GUARD(&g_type_mutex);
+		// MUTEX_GUARD(&g_type_mutex);
 
 		Type *elem = t->Array.elem;
 		bool pop = type_path_push(path, elem);
@@ -3419,7 +3419,7 @@ gb_internal i64 type_align_of_internal(Type *t, TypePath *path) {
 	}
 
 	case Type_EnumeratedArray: {
-		MUTEX_GUARD(&g_type_mutex);
+		// MUTEX_GUARD(&g_type_mutex);
 
 		Type *elem = t->EnumeratedArray.elem;
 		bool pop = type_path_push(path, elem);
@@ -3440,7 +3440,7 @@ gb_internal i64 type_align_of_internal(Type *t, TypePath *path) {
 
 
 	case Type_Tuple: {
-		MUTEX_GUARD(&g_type_mutex);
+		// MUTEX_GUARD(&g_type_mutex);
 
 		i64 max = 1;
 		for_array(i, t->Tuple.variables) {
@@ -3465,7 +3465,7 @@ gb_internal i64 type_align_of_internal(Type *t, TypePath *path) {
 			return gb_max(t->Union.custom_align, 1);
 		}
 
-		MUTEX_GUARD(&g_type_mutex);
+		// MUTEX_GUARD(&g_type_mutex);
 
 		i64 max = 1;
 		for_array(i, t->Union.variants) {
@@ -3492,7 +3492,7 @@ gb_internal i64 type_align_of_internal(Type *t, TypePath *path) {
 			return 1;
 		}
 
-		MUTEX_GUARD(&g_type_mutex);
+		// MUTEX_GUARD(&g_type_mutex);
 
 		i64 max = 1;
 		for_array(i, t->Struct.fields) {
@@ -3605,7 +3605,7 @@ gb_internal i64 type_size_of_internal(Type *t, TypePath *path) {
 
 	switch (t->kind) {
 	case Type_Named: {
-		MUTEX_GUARD(&g_type_mutex);
+		// MUTEX_GUARD(&g_type_mutex);
 
 		bool pop = type_path_push(path, t);
 		if (path->failure) {
@@ -3644,7 +3644,7 @@ gb_internal i64 type_size_of_internal(Type *t, TypePath *path) {
 		return build_context.word_size*2;
 
 	case Type_Array: {
-		MUTEX_GUARD(&g_type_mutex);
+		// MUTEX_GUARD(&g_type_mutex);
 
 		i64 count, align, size, alignment;
 		count = t->Array.count;
@@ -3661,7 +3661,7 @@ gb_internal i64 type_size_of_internal(Type *t, TypePath *path) {
 	} break;
 
 	case Type_EnumeratedArray: {
-		MUTEX_GUARD(&g_type_mutex);
+		// MUTEX_GUARD(&g_type_mutex);
 
 		i64 count, align, size, alignment;
 		count = t->EnumeratedArray.count;
@@ -3695,7 +3695,7 @@ gb_internal i64 type_size_of_internal(Type *t, TypePath *path) {
 		return (1 + 1 + 2)*build_context.word_size;
 
 	case Type_Tuple: {
-		MUTEX_GUARD(&g_type_mutex);
+		// MUTEX_GUARD(&g_type_mutex);
 
 		i64 count, align, size;
 		count = t->Tuple.variables.count;
@@ -3715,7 +3715,7 @@ gb_internal i64 type_size_of_internal(Type *t, TypePath *path) {
 		if (t->Union.variants.count == 0) {
 			return 0;
 		}
-		MUTEX_GUARD(&g_type_mutex);
+		// MUTEX_GUARD(&g_type_mutex);
 
 		i64 align = type_align_of_internal(t, path);
 		if (path->failure) {
@@ -3754,7 +3754,7 @@ gb_internal i64 type_size_of_internal(Type *t, TypePath *path) {
 
 
 	case Type_Struct: {
-		MUTEX_GUARD(&g_type_mutex);
+		// MUTEX_GUARD(&g_type_mutex);
 
 		if (t->Struct.is_raw_union) {
 			i64 count = t->Struct.fields.count;