Browse Source

Improve thread pool (volatile hints, etc)

gingerBill 6 years ago
parent
commit
1348d8a8cd
3 changed files with 45 additions and 22 deletions
  1. 4 4
      src/gb/gb.h
  2. 1 1
      src/parser.cpp
  3. 40 17
      src/thread_pool.cpp

+ 4 - 4
src/gb/gb.h

@@ -978,10 +978,10 @@ typedef struct gbThread {
 	pthread_t     posix_handle;
 #endif
 
-	gbThreadProc *proc;
-	void *        user_data;
-	isize         user_index;
-	isize         return_value;
+	gbThreadProc * proc;
+	void *         user_data;
+	isize          user_index;
+	isize volatile return_value;
 
 	gbSemaphore   semaphore;
 	isize         stack_size;

+ 1 - 1
src/parser.cpp

@@ -4822,7 +4822,7 @@ ParseFileError parse_packages(Parser *p, String init_filename) {
 	thread_pool_kick_and_wait(&parser_thread_pool);
 
 	// NOTE(bill): Get the last error and use that
-	for (isize i = parser_thread_pool.threads.count-1; i >= 0; i--) {
+	for (isize i = parser_thread_pool.thread_count-1; i >= 0; i--) {
 		gbThread *t = &parser_thread_pool.threads[i];
 		ParseFileError err = cast(ParseFileError)t->return_value;
 		if (err != ParseFile_None) {

+ 40 - 17
src/thread_pool.cpp

@@ -16,8 +16,14 @@ struct ThreadPool {
 	gbAtomic32  processing_work_count;
 	bool        is_running;
 
-	Array<WorkerTask> tasks;
-	Array<gbThread> threads;
+	gbAllocator allocator;
+
+	WorkerTask *tasks;
+	isize volatile task_count;
+	isize volatile task_capacity;
+
+	gbThread *threads;
+	isize thread_count;
 
 	char worker_prefix[10];
 	i32 worker_prefix_len;
@@ -33,8 +39,12 @@ void thread_pool_kick_and_wait(ThreadPool *pool);
 GB_THREAD_PROC(worker_thread_internal);
 
 void thread_pool_init(ThreadPool *pool, gbAllocator const &a, isize thread_count, char const *worker_prefix) {
-	pool->tasks = array_make<WorkerTask>(a, 0, 1024);
-	pool->threads = array_make<gbThread>(a, thread_count);
+	pool->allocator = a;
+	pool->task_count = 0;
+	pool->task_capacity = 1024;
+	pool->tasks = gb_alloc_array(a, WorkerTask, pool->task_capacity);
+	pool->threads = gb_alloc_array(a, gbThread, thread_count);
+	pool->thread_count = thread_count;
 	gb_mutex_init(&pool->task_mutex);
 	gb_mutex_init(&pool->mutex);
 	gb_semaphore_init(&pool->semaphore);
@@ -48,7 +58,7 @@ void thread_pool_init(ThreadPool *pool, gbAllocator const &a, isize thread_count
 		pool->worker_prefix_len = worker_prefix_len;
 	}
 
-	for_array(i, pool->threads) {
+	for (isize i = 0; i < pool->thread_count; i++) {
 		gbThread *t = &pool->threads[i];
 		gb_thread_init(t);
 		t->user_index = i;
@@ -63,7 +73,7 @@ void thread_pool_init(ThreadPool *pool, gbAllocator const &a, isize thread_count
 }
 
 void thread_pool_start(ThreadPool *pool) {
-	for_array(i, pool->threads) {
+	for (isize i = 0; i < pool->thread_count; i++) {
 		gbThread *t = &pool->threads[i];
 		gb_thread_start(t, worker_thread_internal, pool);
 	}
@@ -72,11 +82,11 @@ void thread_pool_start(ThreadPool *pool) {
 void thread_pool_join(ThreadPool *pool) {
 	pool->is_running = false;
 
-	for_array(i, pool->threads) {
+	for (isize i = 0; i < pool->thread_count; i++) {
 		gb_semaphore_release(&pool->semaphore);
 	}
 
-	for_array(i, pool->threads) {
+	for (isize i = 0; i < pool->thread_count; i++) {
 		gbThread *t = &pool->threads[i];
 		gb_thread_join(t);
 	}
@@ -89,18 +99,30 @@ void thread_pool_destroy(ThreadPool *pool) {
 	gb_semaphore_destroy(&pool->semaphore);
 	gb_mutex_destroy(&pool->mutex);
 	gb_mutex_destroy(&pool->task_mutex);
-	array_free(&pool->threads);
-	array_free(&pool->tasks);
+	gb_free(pool->allocator, pool->threads);
+	pool->thread_count = 0;
+	gb_free(pool->allocator, pool->tasks);
+	pool->task_count = 0;
+	pool->task_capacity = 0;
+
 }
 
 
 void thread_pool_add_task(ThreadPool *pool, WorkerTaskProc *proc, void *data) {
 	gb_mutex_lock(&pool->task_mutex);
 
+	if (pool->task_count == pool->task_capacity) {
+		isize new_cap = 2*pool->task_capacity + 8;
+		WorkerTask *new_tasks = gb_alloc_array(pool->allocator, WorkerTask, new_cap);
+		gb_memmove(new_tasks, pool->tasks, pool->task_count*gb_size_of(WorkerTask));
+		pool->tasks = new_tasks;
+		pool->task_capacity = new_cap;
+	}
 	WorkerTask task = {};
 	task.do_work = proc;
 	task.data = data;
-	array_add(&pool->tasks, task);
+
+	pool->tasks[pool->task_count++] = task;
 
 	gb_mutex_unlock(&pool->task_mutex);
 
@@ -108,19 +130,20 @@ void thread_pool_add_task(ThreadPool *pool, WorkerTaskProc *proc, void *data) {
 }
 
 void thread_pool_kick(ThreadPool *pool) {
-	if (pool->tasks.count > 0) {
-		isize count = gb_min(pool->tasks.count, pool->threads.count);
+	gb_mutex_lock(&pool->task_mutex);
+	if (pool->task_count > 0) {
+		isize count = gb_max(pool->task_count, pool->thread_count);
 		for (isize i = 0; i < count; i++) {
 			gb_semaphore_post(&pool->semaphore, 1);
 		}
 	}
-
+	gb_mutex_unlock(&pool->task_mutex);
 }
 void thread_pool_kick_and_wait(ThreadPool *pool) {
 	thread_pool_kick(pool);
 
 	isize return_value = 0;
-	while (pool->tasks.count > 0 || gb_atomic32_load(&pool->processing_work_count) != 0) {
+	while (pool->task_count > 0 || gb_atomic32_load(&pool->processing_work_count) != 0) {
 		gb_yield();
 	}
 
@@ -138,9 +161,9 @@ GB_THREAD_PROC(worker_thread_internal) {
 		bool got_task = false;
 
 		if (gb_mutex_try_lock(&pool->task_mutex)) {
-			if (pool->tasks.count > 0) {
+			if (pool->task_count > 0) {
 				gb_atomic32_fetch_add(&pool->processing_work_count, +1);
-				task = array_pop(&pool->tasks);
+				task = pool->tasks[--pool->task_count];
 				got_task = true;
 			}
 			gb_mutex_unlock(&pool->task_mutex);