Comment out many mutex guards in type_(size|align)_of_internal

This commit is contained in:
gingerBill
2023-01-02 21:35:40 +00:00
parent f16d8e77b3
commit 7ffffeeccc
2 changed files with 21 additions and 22 deletions

View File

@@ -30,7 +30,7 @@ gb_internal void thread_pool_init(ThreadPool *pool, gbAllocator const &a, isize
slice_init(&pool->threads, a, thread_count + 1);
// NOTE: this needs to be initialized before any thread starts
pool->running.store(true);
pool->running.store(true, std::memory_order_seq_cst);
// setup the main thread
thread_init(pool, &pool->threads[0], 0);
@@ -43,7 +43,7 @@ gb_internal void thread_pool_init(ThreadPool *pool, gbAllocator const &a, isize
}
gb_internal void thread_pool_destroy(ThreadPool *pool) {
pool->running.store(false);
pool->running.store(false, std::memory_order_seq_cst);
for_array_off(i, 1, pool->threads) {
Thread *t = &pool->threads[i];
@@ -114,7 +114,7 @@ gb_internal bool thread_pool_add_task(ThreadPool *pool, WorkerTaskProc *proc, vo
gb_internal void thread_pool_wait(ThreadPool *pool) {
WorkerTask task;
while (pool->tasks_left.load()) {
while (pool->tasks_left.load(std::memory_order_acquire)) {
// if we've got tasks on our queue, run them
while (thread_pool_queue_pop(current_thread, &task)) {
task.do_work(task.data);
@@ -126,7 +126,7 @@ gb_internal void thread_pool_wait(ThreadPool *pool) {
// This *must* be executed in this order, so the futex wakes immediately
// if rem_tasks has changed since we checked last, otherwise the program
// will permanently sleep
Footex rem_tasks = pool->tasks_left.load();
Footex rem_tasks = pool->tasks_left.load(std::memory_order_acquire);
if (rem_tasks == 0) {
return;
}
@@ -141,7 +141,7 @@ gb_internal THREAD_PROC(thread_pool_thread_proc) {
ThreadPool *pool = current_thread->pool;
// debugf("worker id: %td\n", current_thread->idx);
while (pool->running.load()) {
while (pool->running.load(std::memory_order_seq_cst)) {
// If we've got tasks to process, work through them
usize finished_tasks = 0;
i32 state;
@@ -152,30 +152,29 @@ gb_internal THREAD_PROC(thread_pool_thread_proc) {
finished_tasks += 1;
}
if (finished_tasks > 0 && pool->tasks_left.load() == 0) {
if (finished_tasks > 0 && pool->tasks_left.load(std::memory_order_acquire) == 0) {
futex_signal(&pool->tasks_left);
}
// If there's still work somewhere and we don't have it, steal it
if (pool->tasks_left.load()) {
if (pool->tasks_left.load(std::memory_order_acquire)) {
usize idx = cast(usize)current_thread->idx;
for_array(i, pool->threads) {
if (pool->tasks_left.load() == 0) {
if (pool->tasks_left.load(std::memory_order_acquire) == 0) {
break;
}
idx = (idx + 1) % cast(usize)pool->threads.count;
Thread *thread = &pool->threads.data[idx];
WorkerTask task;
WorkerTask task, another_task;
if (!thread_pool_queue_pop(thread, &task)) {
continue;
}
task.do_work(task.data);
pool->tasks_left.fetch_sub(1, std::memory_order_release);
if (pool->tasks_left.load() == 0) {
if (pool->tasks_left.load(std::memory_order_acquire) == 0) {
futex_signal(&pool->tasks_left);
}

View File

@@ -3406,7 +3406,7 @@ gb_internal i64 type_align_of_internal(Type *t, TypePath *path) {
} break;
case Type_Array: {
MUTEX_GUARD(&g_type_mutex);
// MUTEX_GUARD(&g_type_mutex);
Type *elem = t->Array.elem;
bool pop = type_path_push(path, elem);
@@ -3419,7 +3419,7 @@ gb_internal i64 type_align_of_internal(Type *t, TypePath *path) {
}
case Type_EnumeratedArray: {
MUTEX_GUARD(&g_type_mutex);
// MUTEX_GUARD(&g_type_mutex);
Type *elem = t->EnumeratedArray.elem;
bool pop = type_path_push(path, elem);
@@ -3440,7 +3440,7 @@ gb_internal i64 type_align_of_internal(Type *t, TypePath *path) {
case Type_Tuple: {
MUTEX_GUARD(&g_type_mutex);
// MUTEX_GUARD(&g_type_mutex);
i64 max = 1;
for_array(i, t->Tuple.variables) {
@@ -3465,7 +3465,7 @@ gb_internal i64 type_align_of_internal(Type *t, TypePath *path) {
return gb_max(t->Union.custom_align, 1);
}
MUTEX_GUARD(&g_type_mutex);
// MUTEX_GUARD(&g_type_mutex);
i64 max = 1;
for_array(i, t->Union.variants) {
@@ -3492,7 +3492,7 @@ gb_internal i64 type_align_of_internal(Type *t, TypePath *path) {
return 1;
}
MUTEX_GUARD(&g_type_mutex);
// MUTEX_GUARD(&g_type_mutex);
i64 max = 1;
for_array(i, t->Struct.fields) {
@@ -3605,7 +3605,7 @@ gb_internal i64 type_size_of_internal(Type *t, TypePath *path) {
switch (t->kind) {
case Type_Named: {
MUTEX_GUARD(&g_type_mutex);
// MUTEX_GUARD(&g_type_mutex);
bool pop = type_path_push(path, t);
if (path->failure) {
@@ -3644,7 +3644,7 @@ gb_internal i64 type_size_of_internal(Type *t, TypePath *path) {
return build_context.word_size*2;
case Type_Array: {
MUTEX_GUARD(&g_type_mutex);
// MUTEX_GUARD(&g_type_mutex);
i64 count, align, size, alignment;
count = t->Array.count;
@@ -3661,7 +3661,7 @@ gb_internal i64 type_size_of_internal(Type *t, TypePath *path) {
} break;
case Type_EnumeratedArray: {
MUTEX_GUARD(&g_type_mutex);
// MUTEX_GUARD(&g_type_mutex);
i64 count, align, size, alignment;
count = t->EnumeratedArray.count;
@@ -3695,7 +3695,7 @@ gb_internal i64 type_size_of_internal(Type *t, TypePath *path) {
return (1 + 1 + 2)*build_context.word_size;
case Type_Tuple: {
MUTEX_GUARD(&g_type_mutex);
// MUTEX_GUARD(&g_type_mutex);
i64 count, align, size;
count = t->Tuple.variables.count;
@@ -3715,7 +3715,7 @@ gb_internal i64 type_size_of_internal(Type *t, TypePath *path) {
if (t->Union.variants.count == 0) {
return 0;
}
MUTEX_GUARD(&g_type_mutex);
// MUTEX_GUARD(&g_type_mutex);
i64 align = type_align_of_internal(t, path);
if (path->failure) {
@@ -3754,7 +3754,7 @@ gb_internal i64 type_size_of_internal(Type *t, TypePath *path) {
case Type_Struct: {
MUTEX_GUARD(&g_type_mutex);
// MUTEX_GUARD(&g_type_mutex);
if (t->Struct.is_raw_union) {
i64 count = t->Struct.fields.count;