From 7ffffeecccc6a1fa1b26238f8ed4608d93ec9bb0 Mon Sep 17 00:00:00 2001 From: gingerBill Date: Mon, 2 Jan 2023 21:35:40 +0000 Subject: [PATCH] Comment out many mutex guards in `type_(size|align)_of_internal` --- src/thread_pool.cpp | 21 ++++++++++----------- src/types.cpp | 22 +++++++++++----------- 2 files changed, 21 insertions(+), 22 deletions(-) diff --git a/src/thread_pool.cpp b/src/thread_pool.cpp index a429e47ff..b89e00454 100644 --- a/src/thread_pool.cpp +++ b/src/thread_pool.cpp @@ -30,7 +30,7 @@ gb_internal void thread_pool_init(ThreadPool *pool, gbAllocator const &a, isize slice_init(&pool->threads, a, thread_count + 1); // NOTE: this needs to be initialized before any thread starts - pool->running.store(true); + pool->running.store(true, std::memory_order_seq_cst); // setup the main thread thread_init(pool, &pool->threads[0], 0); @@ -43,7 +43,7 @@ gb_internal void thread_pool_init(ThreadPool *pool, gbAllocator const &a, isize } gb_internal void thread_pool_destroy(ThreadPool *pool) { - pool->running.store(false); + pool->running.store(false, std::memory_order_seq_cst); for_array_off(i, 1, pool->threads) { Thread *t = &pool->threads[i]; @@ -114,7 +114,7 @@ gb_internal bool thread_pool_add_task(ThreadPool *pool, WorkerTaskProc *proc, vo gb_internal void thread_pool_wait(ThreadPool *pool) { WorkerTask task; - while (pool->tasks_left.load()) { + while (pool->tasks_left.load(std::memory_order_acquire)) { // if we've got tasks on our queue, run them while (thread_pool_queue_pop(current_thread, &task)) { task.do_work(task.data); @@ -126,7 +126,7 @@ gb_internal void thread_pool_wait(ThreadPool *pool) { // This *must* be executed in this order, so the futex wakes immediately // if rem_tasks has changed since we checked last, otherwise the program // will permanently sleep - Footex rem_tasks = pool->tasks_left.load(); + Footex rem_tasks = pool->tasks_left.load(std::memory_order_acquire); if (rem_tasks == 0) { return; } @@ -141,7 +141,7 @@ gb_internal THREAD_PROC(thread_pool_thread_proc) { ThreadPool *pool = current_thread->pool; // debugf("worker id: %td\n", current_thread->idx); - while (pool->running.load()) { + while (pool->running.load(std::memory_order_seq_cst)) { // If we've got tasks to process, work through them usize finished_tasks = 0; i32 state; @@ -152,30 +152,29 @@ gb_internal THREAD_PROC(thread_pool_thread_proc) { finished_tasks += 1; } - if (finished_tasks > 0 && pool->tasks_left.load() == 0) { + if (finished_tasks > 0 && pool->tasks_left.load(std::memory_order_acquire) == 0) { futex_signal(&pool->tasks_left); } // If there's still work somewhere and we don't have it, steal it - if (pool->tasks_left.load()) { + if (pool->tasks_left.load(std::memory_order_acquire)) { usize idx = cast(usize)current_thread->idx; for_array(i, pool->threads) { - if (pool->tasks_left.load() == 0) { + if (pool->tasks_left.load(std::memory_order_acquire) == 0) { break; } idx = (idx + 1) % cast(usize)pool->threads.count; Thread *thread = &pool->threads.data[idx]; - WorkerTask task; + WorkerTask task, another_task; if (!thread_pool_queue_pop(thread, &task)) { continue; } - task.do_work(task.data); pool->tasks_left.fetch_sub(1, std::memory_order_release); - if (pool->tasks_left.load() == 0) { + if (pool->tasks_left.load(std::memory_order_acquire) == 0) { futex_signal(&pool->tasks_left); } diff --git a/src/types.cpp b/src/types.cpp index c49f43f7c..1e2d85ac6 100644 --- a/src/types.cpp +++ b/src/types.cpp @@ -3406,7 +3406,7 @@ gb_internal i64 type_align_of_internal(Type *t, TypePath *path) { } break; case Type_Array: { - MUTEX_GUARD(&g_type_mutex); + // MUTEX_GUARD(&g_type_mutex); Type *elem = t->Array.elem; bool pop = type_path_push(path, elem); @@ -3419,7 +3419,7 @@ gb_internal i64 type_align_of_internal(Type *t, TypePath *path) { } case Type_EnumeratedArray: { - MUTEX_GUARD(&g_type_mutex); + // MUTEX_GUARD(&g_type_mutex); Type *elem = t->EnumeratedArray.elem; bool pop = type_path_push(path, elem); @@ -3440,7 +3440,7 @@ gb_internal i64 type_align_of_internal(Type *t, TypePath *path) { case Type_Tuple: { - MUTEX_GUARD(&g_type_mutex); + // MUTEX_GUARD(&g_type_mutex); i64 max = 1; for_array(i, t->Tuple.variables) { @@ -3465,7 +3465,7 @@ gb_internal i64 type_align_of_internal(Type *t, TypePath *path) { return gb_max(t->Union.custom_align, 1); } - MUTEX_GUARD(&g_type_mutex); + // MUTEX_GUARD(&g_type_mutex); i64 max = 1; for_array(i, t->Union.variants) { @@ -3492,7 +3492,7 @@ gb_internal i64 type_align_of_internal(Type *t, TypePath *path) { return 1; } - MUTEX_GUARD(&g_type_mutex); + // MUTEX_GUARD(&g_type_mutex); i64 max = 1; for_array(i, t->Struct.fields) { @@ -3605,7 +3605,7 @@ gb_internal i64 type_size_of_internal(Type *t, TypePath *path) { switch (t->kind) { case Type_Named: { - MUTEX_GUARD(&g_type_mutex); + // MUTEX_GUARD(&g_type_mutex); bool pop = type_path_push(path, t); if (path->failure) { @@ -3644,7 +3644,7 @@ gb_internal i64 type_size_of_internal(Type *t, TypePath *path) { return build_context.word_size*2; case Type_Array: { - MUTEX_GUARD(&g_type_mutex); + // MUTEX_GUARD(&g_type_mutex); i64 count, align, size, alignment; count = t->Array.count; @@ -3661,7 +3661,7 @@ gb_internal i64 type_size_of_internal(Type *t, TypePath *path) { } break; case Type_EnumeratedArray: { - MUTEX_GUARD(&g_type_mutex); + // MUTEX_GUARD(&g_type_mutex); i64 count, align, size, alignment; count = t->EnumeratedArray.count; @@ -3695,7 +3695,7 @@ gb_internal i64 type_size_of_internal(Type *t, TypePath *path) { return (1 + 1 + 2)*build_context.word_size; case Type_Tuple: { - MUTEX_GUARD(&g_type_mutex); + // MUTEX_GUARD(&g_type_mutex); i64 count, align, size; count = t->Tuple.variables.count; @@ -3715,7 +3715,7 @@ gb_internal i64 type_size_of_internal(Type *t, TypePath *path) { if (t->Union.variants.count == 0) { return 0; } - MUTEX_GUARD(&g_type_mutex); + // MUTEX_GUARD(&g_type_mutex); i64 align = type_align_of_internal(t, path); if (path->failure) { @@ -3754,7 +3754,7 @@ gb_internal i64 type_size_of_internal(Type *t, TypePath *path) { case Type_Struct: { - MUTEX_GUARD(&g_type_mutex); + // MUTEX_GUARD(&g_type_mutex); if (t->Struct.is_raw_union) { i64 count = t->Struct.fields.count;