Home | Sort by: relevance | last modified time | path |
/src/sys/kern/ | |
kern_runq.c | 717 struct schedstate_percpu *spc, *tspc; local in function:sched_idle_migrate 749 tspc = &tci->ci_schedstate; 759 } else if (!mutex_tryenter(tspc->spc_mutex)) { 774 lwp_setlock(l, tspc->spc_mutex); 795 struct schedstate_percpu *spc, *tspc; local in function:sched_steal 799 tspc = &tci->ci_schedstate; 800 if (atomic_load_relaxed(&tspc->spc_mcount) != 0 && 801 spc->spc_psid == tspc->spc_psid) { 819 struct schedstate_percpu *spc, *tspc; local in function:sched_idle 882 tspc = &inner->ci_schedstate 909 struct schedstate_percpu *tspc; local in function:sched_preempted [all...] |
kern_lwp.c | 1375 struct schedstate_percpu *tspc; local in function:lwp_migrate 1403 tspc = &tci->ci_schedstate; 1416 lwp_unlock_to(l, tspc->spc_lwplock); |