Lines Matching defs:spc
88 struct schedstate_percpu *spc = &ci->ci_schedstate;
93 if (spc->spc_schedflags & SPCF_SEENRR || count >= 2) {
99 atomic_setbits_int(&spc->spc_schedflags,
102 atomic_setbits_int(&spc->spc_schedflags,
107 if (spc->spc_nrun || spc->spc_schedflags & SPCF_SHOULDYIELD)
344 struct schedstate_percpu *spc = &curcpu()->ci_schedstate;
370 if (ISSET(spc->spc_schedflags, SPCF_ITIMER)) {
371 atomic_clearbits_int(&spc->spc_schedflags, SPCF_ITIMER);
372 clockintr_cancel(&spc->spc_itimer);
374 if (ISSET(spc->spc_schedflags, SPCF_PROFCLOCK)) {
375 atomic_clearbits_int(&spc->spc_schedflags, SPCF_PROFCLOCK);
376 clockintr_cancel(&spc->spc_profclock);
383 atomic_clearbits_int(&spc->spc_schedflags, SPCF_SWITCHCLEAR);
420 spc = &p->p_cpu->ci_schedstate;
424 atomic_setbits_int(&spc->spc_schedflags, SPCF_ITIMER);
425 clockintr_advance(&spc->spc_itimer, hardclock_period);
428 atomic_setbits_int(&spc->spc_schedflags, SPCF_PROFCLOCK);
429 clockintr_advance(&spc->spc_profclock, profclock_period);
432 nanouptime(&spc->spc_runtime);
528 struct schedstate_percpu *spc = &ci->ci_schedstate;
531 if (p == spc->spc_idleproc || spc->spc_spinning)