/src/lib/librumpuser/ |
rumpuser_sp.c | 138 LIST_ENTRY(prefork) pf_spcentries; /* linked from forking spc */ 148 waitresp(struct spclient *spc, struct respwait *rw) 153 pthread_mutex_lock(&spc->spc_mtx); 154 sendunlockl(spc); 155 while (!rw->rw_done && spc->spc_state != SPCSTATE_DYING) { 156 pthread_cond_wait(&rw->rw_cv, &spc->spc_mtx); 158 TAILQ_REMOVE(&spc->spc_respwait, rw, rw_entries); 159 spcstate = spc->spc_state; 160 pthread_mutex_unlock(&spc->spc_mtx); 195 lwproc_rfork(struct spclient *spc, int flags, const char *comm 560 struct spclient *spc = &spclist[idx]; local in function:serv_handledisco 594 struct spclient *spc; local in function:serv_shutdown 785 struct spclient *spc = arg; local in function:sp_copyin 826 struct spclient *spc = arg; local in function:sp_copyout 859 struct spclient *spc = arg; local in function:rumpuser_sp_anonmmap 888 struct spclient *spc = arg; local in function:rumpuser_sp_raise 1182 struct spclient *spc; local in function:spserver 1375 struct spclient *spc = arg; local in function:rumpuser_sp_fini [all...] |
sp_common.c | 256 spcresetbuf(struct spclient *spc) 259 spc->spc_buf = NULL; 260 spc->spc_off = 0; 264 spcfreebuf(struct spclient *spc) 267 free(spc->spc_buf); 268 spcresetbuf(spc); 272 sendlockl(struct spclient *spc) 275 while (spc->spc_ostatus != SPCSTATUS_FREE) { 276 spc->spc_ostatus = SPCSTATUS_WANTED; 277 pthread_cond_wait(&spc->spc_cv, &spc->spc_mtx) [all...] |
/src/sys/dev/pci/ |
sti_pci.c | 98 struct sti_pci_softc *spc = device_private(self); local in function:sti_pci_attach 102 spc->sc_dev = self; 104 spc->sc_pc = paa->pa_pc; 105 spc->sc_tag = paa->pa_tag; 106 spc->sc_base.sc_dev = self; 107 spc->sc_base.sc_enable_rom = sti_pci_enable_rom; 108 spc->sc_base.sc_disable_rom = sti_pci_disable_rom; 112 if (sti_check_rom(spc, paa) != 0) 116 ret = sti_pci_is_console(paa, spc->sc_base. bases); 118 spc->sc_base.sc_flags |= STI_CONSOLE 129 struct sti_pci_softc *spc = device_private(dev); local in function:sti_pci_end_attach 396 struct sti_pci_softc *spc = device_private(sc->sc_dev); local in function:sti_pci_enable_rom 422 struct sti_pci_softc *spc = device_private(sc->sc_dev); local in function:sti_pci_disable_rom [all...] |
/src/sys/kern/ |
kern_runq.c | 138 struct schedstate_percpu *spc; local in function:sched_cpuattach 143 spc = &ci->ci_schedstate; 144 spc->spc_nextpkg = ci; 146 if (spc->spc_lwplock == NULL) { 147 spc->spc_lwplock = mutex_obj_alloc(MUTEX_DEFAULT, IPL_SCHED); 151 lwp0.l_mutex = spc->spc_lwplock; 153 if (spc->spc_mutex != NULL) { 159 size = roundup2(sizeof(spc->spc_queue[0]) * PRI_COUNT, coherency_unit) + 162 spc->spc_queue = (void *)roundup2((uintptr_t)p, coherency_unit); 165 spc->spc_mutex = mutex_obj_alloc(MUTEX_DEFAULT, IPL_SCHED) 188 struct schedstate_percpu *spc; local in function:sched_enqueue 262 struct schedstate_percpu *spc; local in function:sched_dequeue 322 struct schedstate_percpu *spc; local in function:sched_resched_cpu 455 const struct schedstate_percpu *spc = &ci->ci_schedstate; local in function:sched_migratable 476 struct schedstate_percpu *spc = &curcpu()->ci_schedstate; local in function:sched_nextpkg 569 struct schedstate_percpu *spc; local in function:sched_takecpu 655 struct schedstate_percpu *spc, *curspc; local in function:sched_catchlwp 717 struct schedstate_percpu *spc, *tspc; local in function:sched_idle_migrate 795 struct schedstate_percpu *spc, *tspc; local in function:sched_steal 819 struct schedstate_percpu *spc, *tspc; local in function:sched_idle 1076 struct schedstate_percpu *spc; local in function:sched_nextlwp 1113 const struct schedstate_percpu *spc; local in function:sched_curcpu_runnable_p 1188 struct schedstate_percpu *spc; local in function:sched_print_runqueue [all...] |
kern_cpu.c | 312 struct schedstate_percpu *spc, *mspc = NULL; local in function:cpu_xc_offline 322 spc = &ci->ci_schedstate; 324 spc->spc_flags |= SPCF_OFFLINE; 382 spc->spc_flags &= ~SPCF_OFFLINE; 389 struct schedstate_percpu *spc; local in function:cpu_xc_online 394 spc = &ci->ci_schedstate; 396 spc->spc_flags &= ~SPCF_OFFLINE; 403 struct schedstate_percpu *spc; local in function:cpu_setstate 410 spc = &ci->ci_schedstate; 415 if ((spc->spc_flags & SPCF_OFFLINE) == 0 495 struct schedstate_percpu *spc; local in function:cpu_xc_intr 507 struct schedstate_percpu *spc; local in function:cpu_xc_nointr 519 struct schedstate_percpu *spc; local in function:cpu_setintr [all...] |
kern_idle.c | 48 struct schedstate_percpu *spc; local in function:idle_loop 54 spc = &ci->ci_schedstate; 55 KASSERT(lwp_locked(l, spc->spc_lwplock)); 59 spc->spc_flags |= SPCF_RUNNING; 83 if ((spc->spc_flags & SPCF_OFFLINE) == 0) {
|
kern_clock.c | 414 struct schedstate_percpu *spc = &ci->ci_schedstate; local in function:statclock 425 if (spc->spc_psdiv != psdiv) { 426 spc->spc_psdiv = psdiv; 427 spc->spc_pscnt = psdiv; 449 if (--spc->spc_pscnt > 0) { 460 spc->spc_cp_time[CP_NICE]++; 462 spc->spc_cp_time[CP_USER]++; 489 if (--spc->spc_pscnt > 0) { 510 spc->spc_cp_time[CP_INTR]++; 513 spc->spc_cp_time[CP_SYS]++ [all...] |
sched_4bsd.c | 104 struct schedstate_percpu *spc = &ci->ci_schedstate; local in function:sched_tick 108 spc->spc_ticks = sched_rrticks; 113 /* spc now unlocked */ 124 KASSERT(l->l_mutex != spc->spc_mutex); 134 if (spc->spc_flags & SPCF_SHOULDYIELD) { 141 } else if (spc->spc_flags & SPCF_SEENRR) { 148 spc->spc_flags |= SPCF_SHOULDYIELD; 156 spc->spc_flags |= SPCF_SHOULDYIELD; 158 spc->spc_flags |= SPCF_SEENRR; 166 /* spc now unlocked * [all...] |
kern_synch.c | 549 nextlwp(struct cpu_info *ci, struct schedstate_percpu *spc) 565 KASSERT(lwp_locked(newl, spc->spc_mutex)); 570 spc->spc_curpriority = lwp_eprio(newl); 571 spc->spc_flags &= ~(SPCF_SWITCHCLEAR | SPCF_IDLE); 572 lwp_setlock(newl, spc->spc_lwplock); 580 spc->spc_curpriority = PRI_IDLE; 581 spc->spc_flags = (spc->spc_flags & ~SPCF_SWITCHCLEAR) | 611 struct schedstate_percpu *spc; local in function:mi_switch 631 spc = &ci->ci_schedstate 1057 struct schedstate_percpu *spc; local in function:sched_changepri 1086 struct schedstate_percpu *spc; local in function:sched_lendpri [all...] |
sys_pset.c | 222 struct schedstate_percpu *spc; local in function:kern_pset_destroy 224 spc = &ci->ci_schedstate; 225 if (spc->spc_psid != psid) 227 spc->spc_psid = PS_NONE; 306 struct schedstate_percpu *spc = NULL; local in function:sys_pset_assign 325 spc = ispc; 338 opsid = spc->spc_psid; 350 if (spc->spc_psid == psid) 356 if (psid != PS_NONE && ((spc->spc_flags & SPCF_OFFLINE) || 357 (nnone == 1 && spc->spc_psid == PS_NONE))) [all...] |
sched_m2.c | 280 struct schedstate_percpu *spc = &l->l_cpu->ci_schedstate; local in function:sched_oncpu 285 spc->spc_ticks = l->l_sched.timeslice; 298 struct schedstate_percpu *spc = &ci->ci_schedstate; local in function:sched_tick 306 KASSERT(l->l_mutex != spc->spc_mutex); 314 spc->spc_ticks = l->l_sched.timeslice; 339 if (lwp_eprio(l) <= spc->spc_maxpriority || l->l_target_cpu) { 340 spc->spc_flags |= SPCF_SHOULDYIELD; 343 /* spc now unlocked */ 345 spc->spc_ticks = l->l_sched.timeslice;
|
subr_interrupt.c | 66 struct schedstate_percpu *spc; local in function:interrupt_shield_xcall 71 spc = &ci->ci_schedstate; 75 spc->spc_flags &= ~SPCF_NOINTR; 77 spc->spc_flags |= SPCF_NOINTR; 88 struct schedstate_percpu *spc; local in function:interrupt_shield 96 spc = &ci->ci_schedstate; 98 if ((spc->spc_flags & SPCF_NOINTR) == 0) 101 if ((spc->spc_flags & SPCF_NOINTR) != 0) 114 spc->spc_lastmod = time_second;
|
/src/lib/librumpclient/ |
rumpclient.c | 137 send_with_recon(struct spclient *spc, struct iovec *iov, size_t iovlen) 145 rv = dosend(spc, iov, iovlen); 220 cliwaitresp(struct spclient *spc, struct respwait *rw, sigset_t *mask, 226 pthread_mutex_lock(&spc->spc_mtx); 228 sendunlockl(spc); 229 mygen = spc->spc_generation; 233 if (__predict_false(spc->spc_generation != mygen || !imalive)) 237 if (spc->spc_istatus == SPCSTATUS_FREE) { 240 spc->spc_istatus = SPCSTATUS_BUSY; 241 pthread_mutex_unlock(&spc->spc_mtx) [all...] |
/src/etc/rc.d/ |
bootconf.sh | 35 spc="" 44 echo -n "${spc}[${name}]" 46 echo -n "${spc}${name}" 48 spc=" "
|
/src/sys/dev/pcmcia/ |
spc_pcmcia.c | 106 struct spc_softc *spc = &sc->sc_spc; local in function:spc_pcmcia_attach 112 spc->sc_dev = self; 122 spc->sc_iot = cfe->iospace[0].handle.iot; 123 spc->sc_ioh = cfe->iospace[0].handle.ioh; 129 spc->sc_initiator = 7; /* XXX */ 130 spc->sc_adapter.adapt_enable = spc_pcmcia_enable; 131 spc->sc_adapter.adapt_refcnt = 1; 136 spc_attach(spc); 137 scsipi_adapter_delref(&spc->sc_adapter);
|
/src/sys/external/bsd/gnu-efi/dist/lib/ |
print.c | 323 POOL_PRINT *spc; local in function:_SPrint 325 spc = Context; 332 if (spc->len + len > spc->maxlen) { 333 len = spc->maxlen - spc->len; 340 CopyMem (spc->str + spc->len, Buffer, len * sizeof(CHAR16)); 341 spc->len += len; 347 if (spc->len < spc->maxlen) 365 POOL_PRINT *spc; local in function:_PoolPrint 456 POOL_PRINT spc; local in function:UnicodeVSPrint 530 POOL_PRINT spc; local in function:VPoolPrint [all...] |
/src/usr.sbin/diskpart/ |
diskpart.c | 124 int spc, def, part, layout, j, ch; local in function:main 176 spc = dp->d_secpercyl; 188 threshold = howmany(spc, badsecttable); 199 dp->d_ncylinders = howmany(totsize, spc); 200 badsecttable = spc * dp->d_ncylinders - totsize; 212 curcyl += howmany(defpart[def][part], spc); 231 numcyls[part] = howmany(defpart[def][part], spc); 239 defpart[def][PART('f')] = numcyls[PART('f')] * spc - badsecttable; 240 defpart[def][PART('g')] = numcyls[PART('g')] * spc - badsecttable; 241 defpart[def][PART('c')] = numcyls[PART('c')] * spc; [all...] |
/src/sbin/newfs_msdos/ |
mkfs_msdos.c | 120 u_int8_t spc; /* sectors per cluster */ member in struct:bsbpb 164 u_int spc; /* sectors per cluster */ member in struct:bpb 182 { .bps = a, .spc = b, .res = c, .nft = d, .rde = e, \ 340 if (bpb.spc == 0) { /* set defaults */ 344 bpb.spc = 1; /* use 512 bytes */ 348 bpb.spc = 1; /* use 512 bytes */ 352 bpb.spc = 8; /* use 4k */ 357 while (bpb.spc < 128 && x < x1) { 359 bpb.spc *= 2; 424 bpb.spc = o.block_size / bpb.bps [all...] |
/src/distrib/utils/embedded/conf/ |
riscv.conf | 36 local spc=2048 37 local cylinders=$(( ${totalsize} / ${spc} )) 47 sectors/cylinder: ${spc}
|
evbmips.conf | 38 local spc=2048 39 local cylinders=$(( ${totalsize} / ${spc} )) 49 sectors/cylinder: ${spc}
|
x86.conf | 26 local spc=2048 27 local cylinders=$(( ${totalsize} / ${spc} )) 37 sectors/cylinder: ${spc}
|
evbarm.conf | 38 local spc=2048 39 local cylinders=$(( ${totalsize} / ${spc} )) 49 sectors/cylinder: ${spc}
|
evbppc.conf | 38 local spc=2048 39 local cylinders=$(( ${totalsize} / ${spc} )) 49 sectors/cylinder: ${spc}
|
/src/lib/libcurses/ |
slk.c | 733 size_t spc, len, width, x; local in function:__slk_set_finalise 737 spc = screen->slk_label_len; 756 if (width + w > spc) 766 if (len > spc) 767 len = spc; 776 x = (spc - width) / 2; 777 if (x + width > spc) 781 x = spc - width; 791 spc -= x; 796 spc -= width [all...] |
/src/sys/arch/dreamcast/dreamcast/ |
machdep.c | 264 intc_intr(int ssr, int spc, int ssp) 283 cf.spc = spc;
|