diff --git a/src/asset_cache.c b/src/asset_cache.c index 93d8403b..940359ab 100644 --- a/src/asset_cache.c +++ b/src/asset_cache.c @@ -77,6 +77,7 @@ INTERNAL void refresh_dbg_table(void) INTERNAL struct asset *asset_cache_get_slot_locked(struct sys_lock *lock, struct string key, u64 hash) { sys_assert_locked_s(lock, &G.lookup_mutex); + (UNUSED)lock; u64 index = hash % ARRAY_COUNT(G.lookup); while (true) { diff --git a/src/mixer.c b/src/mixer.c index 16448561..75a65082 100644 --- a/src/mixer.c +++ b/src/mixer.c @@ -105,6 +105,7 @@ INTERNAL struct track *track_from_handle(struct mixer_track_handle handle) INTERNAL struct track *track_alloc_locked(struct sys_lock *lock, struct sound *sound) { sys_assert_locked_e(lock, &G.mutex); + (UNUSED)lock; struct track *track = NULL; if (G.track_first_free) { @@ -143,6 +144,7 @@ INTERNAL struct track *track_alloc_locked(struct sys_lock *lock, struct sound *s INTERNAL void track_release_locked(struct sys_lock *lock, struct track *track) { sys_assert_locked_e(lock, &G.mutex); + (UNUSED)lock; /* Remove from playing list */ struct track *prev = track->prev; diff --git a/src/sys_win32.c b/src/sys_win32.c index 0002b72b..da3e64e7 100644 --- a/src/sys_win32.c +++ b/src/sys_win32.c @@ -1580,6 +1580,7 @@ struct thread_local_store *sys_thread_get_thread_local_store(void) INTERNAL struct win32_thread *win32_thread_alloc_locked(struct sys_lock *lock) { sys_assert_locked_e(lock, &G.threads_mutex); + (UNUSED)lock; struct win32_thread *t = NULL; if (G.threads_first_free) { @@ -1605,6 +1606,7 @@ INTERNAL struct win32_thread *win32_thread_alloc_locked(struct sys_lock *lock) INTERNAL void win32_thread_release_locked(struct sys_lock *lock, struct win32_thread *t) { sys_assert_locked_e(lock, &G.threads_mutex); + (UNUSED)lock; if (t->prev) { t->prev->next = t->next; @@ -1628,6 +1630,7 @@ INTERNAL void win32_thread_release_locked(struct sys_lock *lock, struct win32_th INTERNAL struct win32_thread *win32_thread_from_sys_thread_locked(struct sys_lock *lock, struct sys_thread st) { sys_assert_locked_s(lock, &G.threads_mutex); + (UNUSED)lock; u64 gen = st.handle[0]; struct win32_thread *t = (struct win32_thread *)st.handle[1]; @@ -1641,6 +1644,7 @@ INTERNAL struct win32_thread *win32_thread_from_sys_thread_locked(struct sys_loc INTERNAL struct sys_thread sys_thread_from_win32_thread_locked(struct sys_lock *lock, struct win32_thread *t) { sys_assert_locked_s(lock, &G.threads_mutex); + (UNUSED)lock; return (struct sys_thread) { .handle[0] = t->gen, diff --git a/src/work.c b/src/work.c index a57caa39..cd405465 100644 --- a/src/work.c +++ b/src/work.c @@ -166,6 +166,7 @@ INTERNAL struct work *work_alloc_locked(struct sys_lock *lock) { __prof; sys_assert_locked_e(lock, &G.mutex); + (UNUSED)lock; struct work *work = NULL; @@ -192,6 +193,7 @@ INTERNAL struct work *work_alloc_locked(struct sys_lock *lock) INTERNAL void work_release_locked(struct sys_lock *lock, struct work *work) { sys_assert_locked_e(lock, &G.mutex); + (UNUSED)lock; work->next_free = G.free_work_head; G.free_work_head = work; @@ -201,6 +203,7 @@ INTERNAL void work_release_locked(struct sys_lock *lock, struct work *work) INTERNAL struct work_handle work_to_handle_locked(struct sys_lock *lock, struct work *work) { sys_assert_locked_e(lock, &G.mutex); + (UNUSED)lock; return (struct work_handle) { .work = work, @@ -211,6 +214,7 @@ INTERNAL struct work_handle work_to_handle_locked(struct sys_lock *lock, struct INTERNAL struct work_task *task_alloc_locked(struct sys_lock *lock) { sys_assert_locked_e(lock, &G.mutex); + (UNUSED)lock; struct work_task *task = NULL; @@ -231,6 +235,7 @@ INTERNAL struct work_task *task_alloc_locked(struct sys_lock *lock) INTERNAL void task_release_locked(struct sys_lock *lock, struct work_task *task) { sys_assert_locked_e(lock, &G.mutex); + (UNUSED)lock; task->next_free = G.free_task_head; G.free_task_head = task; @@ -244,6 +249,7 @@ INTERNAL void work_schedule_locked(struct sys_lock *lock, struct work *work) { __prof; sys_assert_locked_e(lock, &G.mutex); + (UNUSED)lock; enum work_priority priority = work->priority; @@ -282,6 +288,7 @@ INTERNAL void work_unschedule_locked(struct sys_lock *lock, struct work *work) { __prof; sys_assert_locked_e(lock, &G.mutex); + (UNUSED)lock; struct work *prev = (struct work *)work->prev_scheduled; struct work *next = (struct work *)work->next_scheduled; @@ -556,6 +563,7 @@ struct work_handle work_slate_end_and_help(struct work_slate *ws, enum work_prio INTERNAL struct work *work_from_handle_locked(struct sys_lock *lock, struct work_handle handle) { sys_assert_locked_e(lock, &G.mutex); + (UNUSED)lock; struct work *work = handle.work; if (work->gen != handle.gen) {