/src/sys/arch/sparc/sparc/ |
cache_print.h | 37 static void cache_printf_backend(struct cacheinfo *ci, const char *cpuname); 40 cache_printf_backend(struct cacheinfo *ci, const char *cpuname) 43 if (ci->c_flags & CACHE_TRAPPAGEBUG) 48 if (ci->c_totalsize == 0) { 53 if (ci->c_split) { 56 printf("%s", (ci->c_physical ? "physical " : "")); 57 if (ci->ic_totalsize > 0) { 59 ci->ic_totalsize/1024, ci->ic_linesize); 62 if (ci->dc_totalsize > 0) [all...] |
/src/tests/usr.bin/xlint/lint1/ |
d_gcc_compound_statements2.c | 20 struct cpu_info *ci; 21 __asm__ volatile("movl %%fs:4,%0":"=r" (ci)); 22 ci; 29 struct cpu_info ci = { 0 }; local in function:statement_expr_with_only_stmt 31 if (ci.bar > 0) 32 ci.bar++; 33 ci;
|
/src/sys/rump/librump/rumpkern/arch/generic/ |
rump_generic_cpu.c | 41 rump_cpu_attach(struct cpu_info *ci) 46 ci->ci_next = rumpcpu_info_list; 47 rumpcpu_info_list = ci; 49 ci->ci_index = nattached++; 51 kcpuset_set(kcpuset_attached, cpu_index(ci)); 52 kcpuset_set(kcpuset_running, cpu_index(ci));
|
/src/sys/arch/arm/pic/ |
pic_splfuncs.c | 55 struct cpu_info * const ci = curcpu(); local in function:_splraise 56 const int oldipl = ci->ci_cpl; 58 if (newipl > ci->ci_cpl) { 59 pic_set_priority(ci, newipl); 67 struct cpu_info * const ci = curcpu(); local in function:_spllower 68 const int oldipl = ci->ci_cpl; 69 KASSERT(panicstr || newipl <= ci->ci_cpl); 70 if (newipl < ci->ci_cpl) { 72 ci->ci_intr_depth++; 74 ci->ci_intr_depth-- 85 struct cpu_info * const ci = curcpu(); local in function:splx [all...] |
/src/sys/arch/mips/mips/ |
ipifuncs.c | 68 ipi_nop(struct cpu_info *ci) 77 ipi_ast(struct cpu_info *ci) 79 ci->ci_onproc->l_md.md_astpending = 1; 83 ipi_shootdown(struct cpu_info *ci) 89 ipi_syncicache(struct cpu_info *ci) 91 pmap_tlb_syncicache_wanted(ci); 96 ipi_kpreempt(struct cpu_info *ci) 119 ipi_process(struct cpu_info *ci, uint64_t ipi_mask) 124 ci->ci_evcnt_per_ipi[IPI_NOP].ev_count++; 125 ipi_nop(ci); [all...] |
mips3_clockintr.c | 68 struct cpu_info * const ci = curcpu(); local in function:mips3_clockintr 71 ci->ci_ev_count_compare.ev_count++; 73 KASSERT((ci->ci_cycles_per_hz & ~(0xffffffff)) == 0); 74 ci->ci_next_cp0_clk_intr += (uint32_t)(ci->ci_cycles_per_hz & 0xffffffff); 75 mips3_cp0_compare_write(ci->ci_next_cp0_clk_intr); 84 if ((ci->ci_next_cp0_clk_intr - new_cnt) & 0x80000000) { 86 ci->ci_next_cp0_clk_intr = new_cnt + curcpu()->ci_cycles_per_hz; 87 mips3_cp0_compare_write(ci->ci_next_cp0_clk_intr); 106 struct cpu_info * const ci = curcpu() local in function:mips3_init_cp0_clocks [all...] |
mips_softint.c | 62 struct cpu_info * const ci = l->l_cpu; local in function:softint_init_md 65 ci->ci_softlwps[si_level] = l; 100 ci->ci_softints ^= SOFTINT_##level##_MASK; \ 101 softint_fast_dispatch(ci->ci_softlwps[SOFTINT_##level], \ 103 KASSERTMSG(ci->ci_cpl == IPL_HIGH, "cpl (%d) != HIGH", ci->ci_cpl); \ 110 struct cpu_info * const ci = curcpu(); local in function:softint_process 115 KASSERT(ci->ci_cpl == IPL_HIGH); 117 KASSERTMSG(ci->ci_mtx_count == 0, 119 __func__, cpu_index(ci), ci, ci->ci_mtx_count) [all...] |
/src/sys/arch/xen/x86/ |
xen_ipi.c | 74 static void xen_ipi_kpreempt(struct cpu_info *ci, struct intrframe *); 97 struct cpu_info *ci; local in function:xen_ipi_handler 99 ci = curcpu(); 101 KASSERT(ci == arg); 102 pending = atomic_swap_32(&ci->ci_ipis, 0); 108 ci->ci_ipi_events[bit].ev_count++; 110 (*xen_ipifunc[bit])(ci, regs); 126 struct cpu_info *ci; local in function:xen_ipi_init 129 ci = curcpu(); 131 vcpu = ci->ci_vcpuid 192 struct cpu_info *ci, *self = curcpu(); local in function:xen_broadcast_ipi [all...] |
cpu.c | 183 static void cpu_boot_secondary(struct cpu_info *ci); 184 static void cpu_start_secondary(struct cpu_info *ci); 199 struct cpu_info *ci; local in function:cpu_attach 214 ptr = (uintptr_t)kmem_zalloc(sizeof(*ci) + CACHE_LINE_SIZE - 1, 216 ci = (struct cpu_info *)roundup2(ptr, CACHE_LINE_SIZE); 217 ci->ci_curldt = -1; 223 tmp->ci_next = ci; 225 ci = &phycpu_info_primary; 228 ci->ci_self = ci; 255 struct cpu_info *ci = sc->sc_info; local in function:cpu_rescan 277 struct cpu_info *ci = sc->sc_info; local in function:cpu_childdetached 368 struct cpu_info *ci; local in function:cpu_attach_common 591 struct cpu_info *ci; local in function:cpu_boot_secondary_processors 629 struct cpu_info *ci; local in function:cpu_init_idle_lwps 702 struct cpu_info *ci = (struct cpu_info *)v; local in function:cpu_hatch 767 struct cpu_info *ci; local in function:cpu_debug_dump 1108 struct cpu_info *ci = curcpu(); local in function:cpu_load_pmap [all...] |
/src/sys/arch/vax/vax/ |
multicpu.c | 82 struct cpu_info *ci; local in function:cpu_slavesetup 89 ci = kmem_zalloc(sizeof(*ci), KM_SLEEP); 90 device_set_private(self, ci); 91 ci->ci_dev = self; 92 ci->ci_slotid = slotid; 93 ci->ci_cpuid = device_unit(self); 104 ci->ci_istack = istackbase + PAGE_SIZE; 105 SIMPLEQ_INSERT_TAIL(&cpus, ci, ci_next); 108 cq->cq_ci = ci; 121 struct cpu_info *ci; local in function:cpu_send_ipi 155 struct cpu_info * const ci = curcpu(); local in function:cpu_handle_ipi [all...] |
/src/sys/arch/hppa/hppa/ |
ipifuncs.c | 61 hppa_ipi_init(struct cpu_info *ci) 63 struct cpu_softc *sc = ci->ci_softc; 79 struct cpu_info *ci = curcpu(); local in function:hppa_ipi_intr 80 struct cpu_softc *sc = ci->ci_softc; 85 ipi_pending = atomic_swap_ulong(&ci->ci_ipi, 0); 105 hppa_ipi_send(struct cpu_info *ci, u_long ipi) 108 KASSERT(ci->ci_flags & CPUF_RUNNING); 110 atomic_or_ulong(&ci->ci_ipi, (1L << ipi)); 117 * that the store to ci->ci_ipi above must happen before 126 cpu = (struct iomod *)(ci->ci_hpa) 136 struct cpu_info *ci; local in function:hppa_ipi_broadcast 156 struct cpu_info *ci = curcpu(); local in function:hppa_ipi_halt [all...] |
/src/sys/rump/librump/rumpkern/arch/x86/ |
rump_x86_cpu.c | 48 rump_cpu_attach(struct cpu_info *ci) 52 ci->ci_flags |= CPUF_PRIMARY; 55 ci->ci_next = cpu_info_list; 56 cpu_info_list = ci; 58 kcpuset_set(kcpuset_attached, cpu_index(ci)); 59 kcpuset_set(kcpuset_running, cpu_index(ci));
|
/src/sys/arch/sbmips/sbmips/ |
cpu.c | 87 struct cpu_info *ci; local in function:cpu_attach 115 ci = curcpu(); 116 ci->ci_cpu_freq = 50000000 * plldiv; 118 ci->ci_divisor_delay = (ci->ci_cpu_freq + 500000) / 1000000; 120 ci->ci_cycles_per_hz = (ci->ci_cpu_freq + hz / 2 ) / hz; 123 ci->ci_cpu_freq / 1000000, 124 (ci->ci_cpu_freq % 1000000) / 10000, 125 ci->ci_cycles_per_hz, ci->ci_divisor_delay) [all...] |
/src/sys/sys/ |
cctr.h | 52 #define cc_hardclock(ci) \ 54 if ((ci)->ci_cc.cc_cal_ticks && \ 55 ++(ci)->ci_cc.cc_ticks == (ci)->ci_cc.cc_cal_ticks) { \ 56 (ci)->ci_cc.cc_ticks = 0; \ 57 cc_calibrate_cpu((ci)); \
|
/src/sys/arch/x86/x86/ |
cpu.c | 211 static void cpu_boot_secondary(struct cpu_info *ci); 212 static void cpu_start_secondary(struct cpu_info *ci); 248 cpu_pcpuarea_init(struct cpu_info *ci) 258 base = (vaddr_t)&pcpuarea->ent[cpu_index(ci)]; 277 cpu_vm_init(struct cpu_info *ci) 291 cai = &ci->ci_cinfo[i]; 331 aprint_debug_dev(ci->ci_dev, "%d page colors\n", ncolors); 334 pmap_tlb_cpu_init(ci); 336 pmap_vpage_cpu_init(ci); 345 struct cpu_info *ci; local in function:cpu_attach 572 struct cpu_info *ci = sc->sc_info; local in function:cpu_rescan 624 struct cpu_info *ci = sc->sc_info; local in function:cpu_childdetached 754 struct cpu_info *ci; local in function:cpu_boot_secondary_processors 804 struct cpu_info *ci; local in function:cpu_init_idle_lwps 911 struct cpu_info *ci = (struct cpu_info *)v; local in function:cpu_hatch 1071 struct cpu_info *ci; local in function:cpu_debug_dump 1329 struct cpu_info *ci = sc->sc_info; local in function:cpu_stop 1358 struct cpu_info *ci = sc->sc_info; local in function:cpu_suspend 1373 struct cpu_info *ci = sc->sc_info; local in function:cpu_resume 1404 struct cpu_info *ci = sc->sc_info; local in function:cpu_shutdown 1513 struct cpu_info *ci = curcpu(); local in function:x86_cpu_idle_mwait 1527 struct cpu_info *ci = curcpu(); local in function:x86_cpu_idle_halt 1555 struct cpu_info *ci = curcpu(); local in function:cpu_load_pmap [all...] |
ipi.c | 100 x86_send_ipi(struct cpu_info *ci, int ipimask) 106 if (__predict_false((ci->ci_flags & CPUF_RUNNING) == 0)) 111 n = atomic_cas_32(&ci->ci_ipis, o, o | ipimask); 119 ret = x86_ipi(LAPIC_IPI_VECTOR, ci->ci_cpuid, LAPIC_DLMODE_FIXED); 124 device_xname(ci->ci_dev)); 134 struct cpu_info *ci, *self = curcpu(); local in function:x86_broadcast_ipi 138 for (CPU_INFO_FOREACH(cii, ci)) { 139 if (ci == self) 141 if ((ci->ci_flags & CPUF_RUNNING) == 0) 143 atomic_or_32(&ci->ci_ipis, ipimask) 155 struct cpu_info *ci = curcpu(); local in function:x86_ipi_handler [all...] |
/src/sys/kern/ |
subr_cpu.c | 103 struct cpu_info *ci; local in function:mi_cpu_init 111 ci = curcpu(); 112 cpu_topology_fake1(ci); 173 cpu_topology_set(struct cpu_info *ci, u_int package_id, u_int core_id, 179 ci->ci_package_id = package_id; 180 ci->ci_core_id = core_id; 181 ci->ci_smt_id = smt_id; 182 ci->ci_numa_id = numa_id; 183 for (rel = 0; rel < __arraycount(ci->ci_sibling); rel++) { 184 ci->ci_sibling[rel] = ci 227 struct cpu_info *ci, *ci2; local in function:cpu_topology_dump 292 struct cpu_info *ci; local in function:cpu_topology_fake 310 struct cpu_info *ci, *ci2, *ci3; local in function:cpu_topology_init 492 struct cpu_info *ci; local in function:cpu_count_sync [all...] |
kern_cpu.c | 129 mi_cpu_attach(struct cpu_info *ci) 135 if ((ci->ci_index = ncpu) >= maxcpus) 137 kcpuset_set(kcpuset_attached, cpu_index(ci)); 142 kcpuset_create(&ci->ci_kcpuset, true); 143 kcpuset_set(ci->ci_kcpuset, cpu_index(ci)); 145 TAILQ_INIT(&ci->ci_data.cpu_ld_locks); 146 __cpu_simple_lock_init(&ci->ci_data.cpu_ld_lock); 149 snprintf(ci->ci_data.cpu_name, sizeof(ci->ci_data.cpu_name), "cpu%d" 197 struct cpu_info *ci; local in function:cpuctl_ioctl 289 struct cpu_info *ci; local in function:cpu_lookup [all...] |
/src/sys/arch/evbmips/evbmips/ |
cpu.c | 64 struct cpu_info * const ci = curcpu(); local in function:cpu_attach 66 ci->ci_dev = self; 67 device_set_private(self, ci); 70 ci->ci_cpu_freq / 1000000, 71 (ci->ci_cpu_freq % 1000000) / 10000, 72 ci->ci_cycles_per_hz, ci->ci_divisor_delay); 76 cpu_attach_common(self, ci);
|
interrupt.c | 53 struct cpu_info * const ci = curcpu(); local in function:cpu_intr 56 const int mtx_count = ci->ci_mtx_count; 57 const u_int biglock_count = ci->ci_biglock_count; 60 KASSERT(ci->ci_cpl == IPL_HIGH); 63 ci->ci_data.cpu_nintr++; 70 KASSERTMSG(ci->ci_cpl == ipl, 71 "%s: cpl (%d) != ipl (%d)", __func__, ci->ci_cpl, ipl); 77 .intr = (ci->ci_idepth > 1) 96 KASSERT(biglock_count == ci->ci_biglock_count); 98 KASSERT(mtx_count == ci->ci_mtx_count) [all...] |
/src/lib/libc/citrus/ |
citrus_iconv.c | 139 close_shared(struct _citrus_iconv_shared *ci) 141 if (ci) { 142 if (ci->ci_module) { 143 if (ci->ci_ops) { 144 if (ci->ci_closure) 145 (*ci->ci_ops->io_uninit_shared)(ci); 146 free(ci->ci_ops); 148 _citrus_unload_module(ci->ci_module); 150 free(ci); 160 struct _citrus_iconv_shared *ci; local in function:open_shared 264 struct _citrus_iconv_shared * ci; local in function:get_shared 335 struct _citrus_iconv_shared *ci = NULL; local in function:_citrus_iconv_open [all...] |
/src/sys/arch/powerpc/powerpc/ |
softint_machdep.c | 55 softint_deliver(struct cpu_info *ci, int ipl) 58 KASSERT(ci->ci_data.cpu_softints & (1 << ipl)); 59 ci->ci_data.cpu_softints ^= 1 << ipl; 60 softint_fast_dispatch(ci->ci_softlwps[si_level], ipl); 61 KASSERTMSG(ci->ci_cpl == IPL_HIGH, 62 "%s: cpl (%d) != HIGH", __func__, ci->ci_cpl); 66 powerpc_softint(struct cpu_info *ci, int old_ipl, vaddr_t pc) 71 KASSERTMSG(ci->ci_idepth == -1, 73 cpu_index(ci), ci->ci_idepth) 109 struct cpu_info * const ci = l->l_cpu; local in function:powerpc_softint_init_md 119 struct cpu_info * const ci = curcpu(); local in function:powerpc_softint_trigger [all...] |
/src/usr.sbin/cpuctl/arch/ |
i386.c | 673 disable_tsc(struct cpu_info *ci) 675 if (ci->ci_feat_val[0] & CPUID_TSC) { 676 ci->ci_feat_val[0] &= ~CPUID_TSC; 682 amd_family5_setup(struct cpu_info *ci) 685 switch (ci->ci_model) { 693 if (ci->ci_feat_val[0] & CPUID_APIC) 694 ci->ci_feat_val[0] = 695 (ci->ci_feat_val[0] & ~CPUID_APIC) | CPUID_PGE; 705 cyrix6x86_cpu_setup(struct cpu_info *ci) 712 if (ci->ci_signature != 0x552 1975 struct cpu_info *ci, cistore; local in function:identifycpu 2593 struct cpu_info ci; local in function:ucodeupdate_check [all...] |
/src/sys/arch/riscv/riscv/ |
ipifuncs.c | 60 ipi_nop(struct cpu_info *ci) 69 ipi_ast(struct cpu_info *ci) 71 ci->ci_onproc->l_md.md_astpending = 1; 76 ipi_kpreempt(struct cpu_info *ci) 99 ipi_process(struct cpu_info *ci, unsigned long ipi_mask) 104 ci->ci_evcnt_per_ipi[IPI_NOP].ev_count++; 105 ipi_nop(ci); 108 ci->ci_evcnt_per_ipi[IPI_AST].ev_count++; 109 ipi_ast(ci); 112 ci->ci_evcnt_per_ipi[IPI_SUSPEND].ev_count++ [all...] |
clock_machdep.c | 94 struct cpu_info * const ci = curcpu(); local in function:riscv_timer_init 96 evcnt_attach_dynamic(&ci->ci_ev_timer, EVCNT_TYPE_INTR, 97 NULL, device_xname(ci->ci_dev), "timer"); 99 ci->ci_lastintr = csr_time_read(); 100 uint64_t next = ci->ci_lastintr + timer_ticks_per_hz; 101 ci->ci_lastintr_scheduled = next; 106 if (cpu_index(ci) == 0) { 116 struct cpu_info * const ci = curcpu(); local in function:riscv_timer_intr 123 ci->ci_lastintr = now; 124 ci->ci_ev_timer.ev_count++ [all...] |