Lines Matching refs:l_mutex
400 l->l_mutex = l->l_cpu->ci_schedstate.spc_lwplock; in lwp_ctor()
837 KASSERT(l2->l_mutex == l2->l_cpu->ci_schedstate.spc_lwplock); in lwp_create()
1045 lock = prev->l_mutex; in lwp_startup()
1545 kmutex_t *cur = l->l_mutex; in lwp_locked()
1556 kmutex_t *oldmtx = l->l_mutex; in lwp_setlock()
1560 atomic_store_release(&l->l_mutex, mtx); in lwp_setlock()
1575 old = l->l_mutex; in lwp_unlock_to()
1576 atomic_store_release(&l->l_mutex, mtx); in lwp_unlock_to()
1586 if (!mutex_tryenter(old = atomic_load_consume(&l->l_mutex))) in lwp_trylock()
1588 if (__predict_true(atomic_load_relaxed(&l->l_mutex) == old)) in lwp_trylock()
1598 KASSERT(mutex_owned(l->l_mutex)); in lwp_unsleep()
1608 kmutex_t *old = atomic_load_consume(&l->l_mutex); in lwp_lock()
1615 while (__predict_false(atomic_load_relaxed(&l->l_mutex) != old)) { in lwp_lock()
1617 old = atomic_load_consume(&l->l_mutex); in lwp_lock()
1629 mutex_spin_exit(l->l_mutex); in lwp_unlock()
1636 KASSERT(mutex_owned(l->l_mutex)); in lwp_changepri()
1648 KASSERT(mutex_owned(l->l_mutex)); in lwp_lendpri()
1659 KASSERT(mutex_owned(l->l_mutex)); in lwp_eprio()