mirror of
https://github.com/odin-lang/Odin.git
synced 2026-04-27 00:33:56 +00:00
Merge pull request #5707 from janga-perlind/lucas/timing-stuff
Improve OSX threading performance
This commit is contained in:
@@ -19,6 +19,11 @@ enum GrabState {
|
||||
Grab_Failed = 2,
|
||||
};
|
||||
|
||||
enum BroadcastWaitState {
|
||||
Nobody_Waiting = 0,
|
||||
Someone_Waiting = 1,
|
||||
};
|
||||
|
||||
struct ThreadPool {
|
||||
gbAllocator threads_allocator;
|
||||
Slice<Thread> threads;
|
||||
@@ -54,8 +59,8 @@ gb_internal void thread_pool_destroy(ThreadPool *pool) {
|
||||
|
||||
for_array_off(i, 1, pool->threads) {
|
||||
Thread *t = &pool->threads[i];
|
||||
pool->tasks_available.fetch_add(1, std::memory_order_acquire);
|
||||
futex_broadcast(&pool->tasks_available);
|
||||
pool->tasks_available.store(Nobody_Waiting);
|
||||
futex_broadcast(&t->pool->tasks_available);
|
||||
thread_join_and_destroy(t);
|
||||
}
|
||||
|
||||
@@ -87,8 +92,10 @@ void thread_pool_queue_push(Thread *thread, WorkerTask task) {
|
||||
thread->queue.bottom.store(bot + 1, std::memory_order_relaxed);
|
||||
|
||||
thread->pool->tasks_left.fetch_add(1, std::memory_order_release);
|
||||
thread->pool->tasks_available.fetch_add(1, std::memory_order_relaxed);
|
||||
futex_broadcast(&thread->pool->tasks_available);
|
||||
i32 state = Someone_Waiting;
|
||||
if (thread->pool->tasks_available.compare_exchange_strong(state, Nobody_Waiting)) {
|
||||
futex_broadcast(&thread->pool->tasks_available);
|
||||
}
|
||||
}
|
||||
|
||||
GrabState thread_pool_queue_take(Thread *thread, WorkerTask *task) {
|
||||
@@ -230,12 +237,13 @@ gb_internal THREAD_PROC(thread_pool_thread_proc) {
|
||||
}
|
||||
|
||||
// if we've done all our work, and there's nothing to steal, go to sleep
|
||||
state = pool->tasks_available.load(std::memory_order_acquire);
|
||||
pool->tasks_available.store(Someone_Waiting);
|
||||
if (!pool->running) { break; }
|
||||
futex_wait(&pool->tasks_available, state);
|
||||
futex_wait(&pool->tasks_available, Someone_Waiting);
|
||||
|
||||
main_loop_continue:;
|
||||
}
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
|
||||
@@ -195,7 +195,13 @@ gb_internal void mutex_lock(RecursiveMutex *m) {
|
||||
// inside the lock
|
||||
return;
|
||||
}
|
||||
futex_wait(&m->owner, prev_owner);
|
||||
|
||||
// NOTE(lucas): we are doing spin lock since futex signal is expensive on OSX. The recursive locks are
|
||||
// very short lived so we don't hit this mega often and I see no perform regression on windows (with
|
||||
// a performance uplift on OSX).
|
||||
|
||||
//futex_wait(&m->owner, prev_owner);
|
||||
yield_thread();
|
||||
}
|
||||
}
|
||||
gb_internal bool mutex_try_lock(RecursiveMutex *m) {
|
||||
@@ -216,7 +222,9 @@ gb_internal void mutex_unlock(RecursiveMutex *m) {
|
||||
return;
|
||||
}
|
||||
m->owner.exchange(0, std::memory_order_release);
|
||||
futex_signal(&m->owner);
|
||||
// NOTE(lucas): see comment about spin lock in mutex_lock above
|
||||
|
||||
// futex_signal(&m->owner);
|
||||
// outside the lock
|
||||
}
|
||||
|
||||
@@ -423,44 +431,28 @@ gb_internal void semaphore_wait(Semaphore *s) {
|
||||
}
|
||||
|
||||
struct RwMutex {
|
||||
BlockingMutex lock;
|
||||
Condition cond;
|
||||
int32_t readers;
|
||||
// TODO(bill): make this a proper RW mutex
|
||||
BlockingMutex mutex;
|
||||
};
|
||||
|
||||
gb_internal void rw_mutex_lock(RwMutex *m) {
|
||||
mutex_lock(&m->lock);
|
||||
while (m->readers != 0) {
|
||||
condition_wait(&m->cond, &m->lock);
|
||||
}
|
||||
mutex_lock(&m->mutex);
|
||||
}
|
||||
gb_internal bool rw_mutex_try_lock(RwMutex *m) {
|
||||
// TODO(bill): rw_mutex_try_lock
|
||||
rw_mutex_lock(m);
|
||||
return true;
|
||||
return mutex_try_lock(&m->mutex);
|
||||
}
|
||||
gb_internal void rw_mutex_unlock(RwMutex *m) {
|
||||
condition_signal(&m->cond);
|
||||
mutex_unlock(&m->lock);
|
||||
mutex_unlock(&m->mutex);
|
||||
}
|
||||
|
||||
gb_internal void rw_mutex_shared_lock(RwMutex *m) {
|
||||
mutex_lock(&m->lock);
|
||||
m->readers += 1;
|
||||
mutex_unlock(&m->lock);
|
||||
mutex_lock(&m->mutex);
|
||||
}
|
||||
gb_internal bool rw_mutex_try_shared_lock(RwMutex *m) {
|
||||
// TODO(bill): rw_mutex_try_shared_lock
|
||||
rw_mutex_shared_lock(m);
|
||||
return true;
|
||||
return mutex_try_lock(&m->mutex);
|
||||
}
|
||||
gb_internal void rw_mutex_shared_unlock(RwMutex *m) {
|
||||
mutex_lock(&m->lock);
|
||||
m->readers -= 1;
|
||||
if (m->readers == 0) {
|
||||
condition_signal(&m->cond);
|
||||
}
|
||||
mutex_unlock(&m->lock);
|
||||
mutex_unlock(&m->mutex);
|
||||
}
|
||||
#endif
|
||||
|
||||
|
||||
Reference in New Issue
Block a user