Home
last modified time | relevance | path

Searched refs:spc_mutex (Results 1 – 8 of 8) sorted by relevance

/netbsd-src/sys/sys/
H A Dlwp.h411 mutex_spin_enter(ci->ci_schedstate.spc_mutex); in spc_lock()
417 mutex_spin_exit(ci->ci_schedstate.spc_mutex); in spc_unlock()
428 mutex_spin_enter(spc1->spc_mutex); in spc_dlock()
429 mutex_spin_enter(spc2->spc_mutex); in spc_dlock()
431 mutex_spin_enter(spc2->spc_mutex); in spc_dlock()
432 mutex_spin_enter(spc1->spc_mutex); in spc_dlock()
H A Dsched.h159 struct kmutex *spc_mutex; /* (: lock on below, runnable LWPs */ member
/netbsd-src/sys/kern/
H A Dkern_synch.c565 KASSERT(lwp_locked(newl, spc->spc_mutex)); in nextlwp()
620 KASSERT(mutex_owned(curcpu()->ci_schedstate.spc_mutex)); in mi_switch()
668 lwp_setlock(l, spc->spc_mutex); in mi_switch()
731 if (l->l_mutex == spc->spc_mutex) { in mi_switch()
742 mutex_spin_exit(spc->spc_mutex); in mi_switch()
867 mutex_spin_exit(spc->spc_mutex); in mi_switch()
895 KASSERT(l->l_mutex != l->l_cpu->ci_schedstate.spc_mutex); in setrunnable()
951 oldlock = lwp_setlock(l, l->l_cpu->ci_schedstate.spc_mutex); in setrunnable()
1066 KASSERT(lwp_locked(l, spc->spc_mutex)); in sched_changepri()
1095 KASSERT(lwp_locked(l, spc->spc_mutex)); in sched_lendpri()
H A Dkern_runq.c153 if (spc->spc_mutex != NULL) { in sched_cpuattach()
165 spc->spc_mutex = mutex_obj_alloc(MUTEX_DEFAULT, IPL_SCHED); in sched_cpuattach()
182 * Put an LWP onto a run queue. The LWP must be locked by spc_mutex for
195 KASSERT(lwp_locked(l, l->l_cpu->ci_schedstate.spc_mutex)); in sched_enqueue()
267 KASSERT(lwp_locked(l, spc->spc_mutex)); in sched_dequeue()
317 * will be for concurrency reasons, spc_mutex will be dropped before return.
328 KASSERT(mutex_owned(spc->spc_mutex)); in sched_resched_cpu()
412 * specified, and ideally it will be for concurrency reasons, spc_mutex will
420 KASSERT(lwp_locked(l, ci->ci_schedstate.spc_mutex)); in sched_resched_lwp()
700 lwp_unlock_to(l, curspc->spc_mutex); in sched_catchlwp()
[all...]
H A Dsched_4bsd.c124 KASSERT(l->l_mutex != spc->spc_mutex); in sched_tick()
H A Dsched_m2.c306 KASSERT(l->l_mutex != spc->spc_mutex); in sched_tick()
H A Dkern_sleepq.c176 lwp_setlock(l, spc->spc_mutex); in sleepq_remove()
/netbsd-src/sys/rump/librump/rumpkern/
H A Dscheduler.c174 ci->ci_schedstate.spc_mutex = in rump_scheduler_init()
381 l->l_mutex = rcpu->rcpu_ci->ci_schedstate.spc_mutex; in rump_schedule_cpu_interlock()
405 KASSERT(l->l_mutex == l->l_cpu->ci_schedstate.spc_mutex); in rump_unschedule()