add (UNUSED) after lock assertions

This commit is contained in:
jacob 2024-08-09 15:02:13 -05:00
parent 8c26381533
commit 094b64078b
4 changed files with 15 additions and 0 deletions

View File

@ -77,6 +77,7 @@ INTERNAL void refresh_dbg_table(void)
INTERNAL struct asset *asset_cache_get_slot_locked(struct sys_lock *lock, struct string key, u64 hash) INTERNAL struct asset *asset_cache_get_slot_locked(struct sys_lock *lock, struct string key, u64 hash)
{ {
sys_assert_locked_s(lock, &G.lookup_mutex); sys_assert_locked_s(lock, &G.lookup_mutex);
(UNUSED)lock;
u64 index = hash % ARRAY_COUNT(G.lookup); u64 index = hash % ARRAY_COUNT(G.lookup);
while (true) { while (true) {

View File

@ -105,6 +105,7 @@ INTERNAL struct track *track_from_handle(struct mixer_track_handle handle)
INTERNAL struct track *track_alloc_locked(struct sys_lock *lock, struct sound *sound) INTERNAL struct track *track_alloc_locked(struct sys_lock *lock, struct sound *sound)
{ {
sys_assert_locked_e(lock, &G.mutex); sys_assert_locked_e(lock, &G.mutex);
(UNUSED)lock;
struct track *track = NULL; struct track *track = NULL;
if (G.track_first_free) { if (G.track_first_free) {
@ -143,6 +144,7 @@ INTERNAL struct track *track_alloc_locked(struct sys_lock *lock, struct sound *s
INTERNAL void track_release_locked(struct sys_lock *lock, struct track *track) INTERNAL void track_release_locked(struct sys_lock *lock, struct track *track)
{ {
sys_assert_locked_e(lock, &G.mutex); sys_assert_locked_e(lock, &G.mutex);
(UNUSED)lock;
/* Remove from playing list */ /* Remove from playing list */
struct track *prev = track->prev; struct track *prev = track->prev;

View File

@ -1580,6 +1580,7 @@ struct thread_local_store *sys_thread_get_thread_local_store(void)
INTERNAL struct win32_thread *win32_thread_alloc_locked(struct sys_lock *lock) INTERNAL struct win32_thread *win32_thread_alloc_locked(struct sys_lock *lock)
{ {
sys_assert_locked_e(lock, &G.threads_mutex); sys_assert_locked_e(lock, &G.threads_mutex);
(UNUSED)lock;
struct win32_thread *t = NULL; struct win32_thread *t = NULL;
if (G.threads_first_free) { if (G.threads_first_free) {
@ -1605,6 +1606,7 @@ INTERNAL struct win32_thread *win32_thread_alloc_locked(struct sys_lock *lock)
INTERNAL void win32_thread_release_locked(struct sys_lock *lock, struct win32_thread *t) INTERNAL void win32_thread_release_locked(struct sys_lock *lock, struct win32_thread *t)
{ {
sys_assert_locked_e(lock, &G.threads_mutex); sys_assert_locked_e(lock, &G.threads_mutex);
(UNUSED)lock;
if (t->prev) { if (t->prev) {
t->prev->next = t->next; t->prev->next = t->next;
@ -1628,6 +1630,7 @@ INTERNAL void win32_thread_release_locked(struct sys_lock *lock, struct win32_th
INTERNAL struct win32_thread *win32_thread_from_sys_thread_locked(struct sys_lock *lock, struct sys_thread st) INTERNAL struct win32_thread *win32_thread_from_sys_thread_locked(struct sys_lock *lock, struct sys_thread st)
{ {
sys_assert_locked_s(lock, &G.threads_mutex); sys_assert_locked_s(lock, &G.threads_mutex);
(UNUSED)lock;
u64 gen = st.handle[0]; u64 gen = st.handle[0];
struct win32_thread *t = (struct win32_thread *)st.handle[1]; struct win32_thread *t = (struct win32_thread *)st.handle[1];
@ -1641,6 +1644,7 @@ INTERNAL struct win32_thread *win32_thread_from_sys_thread_locked(struct sys_loc
INTERNAL struct sys_thread sys_thread_from_win32_thread_locked(struct sys_lock *lock, struct win32_thread *t) INTERNAL struct sys_thread sys_thread_from_win32_thread_locked(struct sys_lock *lock, struct win32_thread *t)
{ {
sys_assert_locked_s(lock, &G.threads_mutex); sys_assert_locked_s(lock, &G.threads_mutex);
(UNUSED)lock;
return (struct sys_thread) { return (struct sys_thread) {
.handle[0] = t->gen, .handle[0] = t->gen,

View File

@ -166,6 +166,7 @@ INTERNAL struct work *work_alloc_locked(struct sys_lock *lock)
{ {
__prof; __prof;
sys_assert_locked_e(lock, &G.mutex); sys_assert_locked_e(lock, &G.mutex);
(UNUSED)lock;
struct work *work = NULL; struct work *work = NULL;
@ -192,6 +193,7 @@ INTERNAL struct work *work_alloc_locked(struct sys_lock *lock)
INTERNAL void work_release_locked(struct sys_lock *lock, struct work *work) INTERNAL void work_release_locked(struct sys_lock *lock, struct work *work)
{ {
sys_assert_locked_e(lock, &G.mutex); sys_assert_locked_e(lock, &G.mutex);
(UNUSED)lock;
work->next_free = G.free_work_head; work->next_free = G.free_work_head;
G.free_work_head = work; G.free_work_head = work;
@ -201,6 +203,7 @@ INTERNAL void work_release_locked(struct sys_lock *lock, struct work *work)
INTERNAL struct work_handle work_to_handle_locked(struct sys_lock *lock, struct work *work) INTERNAL struct work_handle work_to_handle_locked(struct sys_lock *lock, struct work *work)
{ {
sys_assert_locked_e(lock, &G.mutex); sys_assert_locked_e(lock, &G.mutex);
(UNUSED)lock;
return (struct work_handle) { return (struct work_handle) {
.work = work, .work = work,
@ -211,6 +214,7 @@ INTERNAL struct work_handle work_to_handle_locked(struct sys_lock *lock, struct
INTERNAL struct work_task *task_alloc_locked(struct sys_lock *lock) INTERNAL struct work_task *task_alloc_locked(struct sys_lock *lock)
{ {
sys_assert_locked_e(lock, &G.mutex); sys_assert_locked_e(lock, &G.mutex);
(UNUSED)lock;
struct work_task *task = NULL; struct work_task *task = NULL;
@ -231,6 +235,7 @@ INTERNAL struct work_task *task_alloc_locked(struct sys_lock *lock)
INTERNAL void task_release_locked(struct sys_lock *lock, struct work_task *task) INTERNAL void task_release_locked(struct sys_lock *lock, struct work_task *task)
{ {
sys_assert_locked_e(lock, &G.mutex); sys_assert_locked_e(lock, &G.mutex);
(UNUSED)lock;
task->next_free = G.free_task_head; task->next_free = G.free_task_head;
G.free_task_head = task; G.free_task_head = task;
@ -244,6 +249,7 @@ INTERNAL void work_schedule_locked(struct sys_lock *lock, struct work *work)
{ {
__prof; __prof;
sys_assert_locked_e(lock, &G.mutex); sys_assert_locked_e(lock, &G.mutex);
(UNUSED)lock;
enum work_priority priority = work->priority; enum work_priority priority = work->priority;
@ -282,6 +288,7 @@ INTERNAL void work_unschedule_locked(struct sys_lock *lock, struct work *work)
{ {
__prof; __prof;
sys_assert_locked_e(lock, &G.mutex); sys_assert_locked_e(lock, &G.mutex);
(UNUSED)lock;
struct work *prev = (struct work *)work->prev_scheduled; struct work *prev = (struct work *)work->prev_scheduled;
struct work *next = (struct work *)work->next_scheduled; struct work *next = (struct work *)work->next_scheduled;
@ -556,6 +563,7 @@ struct work_handle work_slate_end_and_help(struct work_slate *ws, enum work_prio
INTERNAL struct work *work_from_handle_locked(struct sys_lock *lock, struct work_handle handle) INTERNAL struct work *work_from_handle_locked(struct sys_lock *lock, struct work_handle handle)
{ {
sys_assert_locked_e(lock, &G.mutex); sys_assert_locked_e(lock, &G.mutex);
(UNUSED)lock;
struct work *work = handle.work; struct work *work = handle.work;
if (work->gen != handle.gen) { if (work->gen != handle.gen) {