Searched refs:spc_mutex (Results 1 – 8 of 8) sorted by relevance
/netbsd-src/sys/sys/ |
H A D | lwp.h | 411 mutex_spin_enter(ci->ci_schedstate.spc_mutex); in spc_lock() 417 mutex_spin_exit(ci->ci_schedstate.spc_mutex); in spc_unlock() 428 mutex_spin_enter(spc1->spc_mutex); in spc_dlock() 429 mutex_spin_enter(spc2->spc_mutex); in spc_dlock() 431 mutex_spin_enter(spc2->spc_mutex); in spc_dlock() 432 mutex_spin_enter(spc1->spc_mutex); in spc_dlock()
|
H A D | sched.h | 159 struct kmutex *spc_mutex; /* (: lock on below, runnable LWPs */ member
|
/netbsd-src/sys/kern/ |
H A D | kern_synch.c | 565 KASSERT(lwp_locked(newl, spc->spc_mutex)); in nextlwp() 620 KASSERT(mutex_owned(curcpu()->ci_schedstate.spc_mutex)); in mi_switch() 668 lwp_setlock(l, spc->spc_mutex); in mi_switch() 731 if (l->l_mutex == spc->spc_mutex) { in mi_switch() 742 mutex_spin_exit(spc->spc_mutex); in mi_switch() 867 mutex_spin_exit(spc->spc_mutex); in mi_switch() 895 KASSERT(l->l_mutex != l->l_cpu->ci_schedstate.spc_mutex); in setrunnable() 951 oldlock = lwp_setlock(l, l->l_cpu->ci_schedstate.spc_mutex); in setrunnable() 1066 KASSERT(lwp_locked(l, spc->spc_mutex)); in sched_changepri() 1095 KASSERT(lwp_locked(l, spc->spc_mutex)); in sched_lendpri()
|
H A D | kern_runq.c | 153 if (spc->spc_mutex != NULL) { in sched_cpuattach() 165 spc->spc_mutex = mutex_obj_alloc(MUTEX_DEFAULT, IPL_SCHED); in sched_cpuattach() 182 * Put an LWP onto a run queue. The LWP must be locked by spc_mutex for 195 KASSERT(lwp_locked(l, l->l_cpu->ci_schedstate.spc_mutex)); in sched_enqueue() 267 KASSERT(lwp_locked(l, spc->spc_mutex)); in sched_dequeue() 317 * will be for concurrency reasons, spc_mutex will be dropped before return. 328 KASSERT(mutex_owned(spc->spc_mutex)); in sched_resched_cpu() 412 * specified, and ideally it will be for concurrency reasons, spc_mutex will 420 KASSERT(lwp_locked(l, ci->ci_schedstate.spc_mutex)); in sched_resched_lwp() 700 lwp_unlock_to(l, curspc->spc_mutex); in sched_catchlwp() [all...] |
H A D | sched_4bsd.c | 124 KASSERT(l->l_mutex != spc->spc_mutex); in sched_tick()
|
H A D | sched_m2.c | 306 KASSERT(l->l_mutex != spc->spc_mutex); in sched_tick()
|
H A D | kern_sleepq.c | 176 lwp_setlock(l, spc->spc_mutex); in sleepq_remove()
|
/netbsd-src/sys/rump/librump/rumpkern/ |
H A D | scheduler.c | 174 ci->ci_schedstate.spc_mutex = in rump_scheduler_init() 381 l->l_mutex = rcpu->rcpu_ci->ci_schedstate.spc_mutex; in rump_schedule_cpu_interlock() 405 KASSERT(l->l_mutex == l->l_cpu->ci_schedstate.spc_mutex); in rump_unschedule()
|