| /src/sys/arch/sparc/sparc/ | 
| cache_print.h | 37 static void cache_printf_backend(struct cacheinfo *ci, const char *cpuname); 40 cache_printf_backend(struct cacheinfo *ci, const char *cpuname)
 43 	if (ci->c_flags & CACHE_TRAPPAGEBUG)
 48 	if (ci->c_totalsize == 0) {
 53 	if (ci->c_split) {
 56 		printf("%s", (ci->c_physical ? "physical " : ""));
 57 		if (ci->ic_totalsize > 0) {
 59 			    ci->ic_totalsize/1024, ci->ic_linesize);
 62 		if (ci->dc_totalsize > 0)
 [all...]
 | 
| /src/tests/usr.bin/xlint/lint1/ | 
| d_gcc_compound_statements2.c | 20 	    struct cpu_info *ci; 21 	    __asm__ volatile("movl %%fs:4,%0":"=r" (ci));
 22 	    ci;
 29 	struct cpu_info ci = { 0 };  local in function:statement_expr_with_only_stmt
 31 		if (ci.bar > 0)
 32 			ci.bar++;
 33 		ci;
 
 | 
| /src/sys/rump/librump/rumpkern/arch/generic/ | 
| rump_generic_cpu.c | 41 rump_cpu_attach(struct cpu_info *ci) 46 	ci->ci_next = rumpcpu_info_list;
 47 	rumpcpu_info_list = ci;
 49 	ci->ci_index = nattached++;
 51 	kcpuset_set(kcpuset_attached, cpu_index(ci));
 52 	kcpuset_set(kcpuset_running, cpu_index(ci));
 
 | 
| /src/sys/arch/arm/pic/ | 
| pic_splfuncs.c | 55 	struct cpu_info * const ci = curcpu();  local in function:_splraise 56 	const int oldipl = ci->ci_cpl;
 58 	if (newipl > ci->ci_cpl) {
 59 		pic_set_priority(ci, newipl);
 67 	struct cpu_info * const ci = curcpu();  local in function:_spllower
 68 	const int oldipl = ci->ci_cpl;
 69 	KASSERT(panicstr || newipl <= ci->ci_cpl);
 70 	if (newipl < ci->ci_cpl) {
 72 		ci->ci_intr_depth++;
 74 		ci->ci_intr_depth--
 85  struct cpu_info * const ci = curcpu();  local in function:splx
 [all...]
 | 
| /src/sys/arch/mips/mips/ | 
| ipifuncs.c | 68 ipi_nop(struct cpu_info *ci) 77 ipi_ast(struct cpu_info *ci)
 79 	ci->ci_onproc->l_md.md_astpending = 1;
 83 ipi_shootdown(struct cpu_info *ci)
 89 ipi_syncicache(struct cpu_info *ci)
 91 	pmap_tlb_syncicache_wanted(ci);
 96 ipi_kpreempt(struct cpu_info *ci)
 119 ipi_process(struct cpu_info *ci, uint64_t ipi_mask)
 124 		ci->ci_evcnt_per_ipi[IPI_NOP].ev_count++;
 125 		ipi_nop(ci);
 [all...]
 | 
| mips3_clockintr.c | 68 	struct cpu_info * const ci = curcpu();  local in function:mips3_clockintr 71 	ci->ci_ev_count_compare.ev_count++;
 73 	KASSERT((ci->ci_cycles_per_hz & ~(0xffffffff)) == 0);
 74 	ci->ci_next_cp0_clk_intr += (uint32_t)(ci->ci_cycles_per_hz & 0xffffffff);
 75 	mips3_cp0_compare_write(ci->ci_next_cp0_clk_intr);
 84 	if ((ci->ci_next_cp0_clk_intr - new_cnt) & 0x80000000) {
 86 		ci->ci_next_cp0_clk_intr = new_cnt + curcpu()->ci_cycles_per_hz;
 87 		mips3_cp0_compare_write(ci->ci_next_cp0_clk_intr);
 106 	struct cpu_info * const ci = curcpu()  local in function:mips3_init_cp0_clocks
 [all...]
 | 
| mips_softint.c | 62 	struct cpu_info * const ci = l->l_cpu;  local in function:softint_init_md 65 	ci->ci_softlwps[si_level] = l;
 100 		ci->ci_softints ^= SOFTINT_##level##_MASK; \
 101 		softint_fast_dispatch(ci->ci_softlwps[SOFTINT_##level], \
 103 		KASSERTMSG(ci->ci_cpl == IPL_HIGH, "cpl (%d) != HIGH", ci->ci_cpl); \
 110 	struct cpu_info * const ci = curcpu();  local in function:softint_process
 115 	KASSERT(ci->ci_cpl == IPL_HIGH);
 117 	KASSERTMSG(ci->ci_mtx_count == 0,
 119 	     __func__, cpu_index(ci), ci, ci->ci_mtx_count)
 [all...]
 | 
| /src/sys/arch/xen/x86/ | 
| xen_ipi.c | 74 static void xen_ipi_kpreempt(struct cpu_info *ci, struct intrframe *); 97 	struct cpu_info *ci;  local in function:xen_ipi_handler
 99 	ci = curcpu();
 101 	KASSERT(ci == arg);
 102 	pending = atomic_swap_32(&ci->ci_ipis, 0);
 108 		ci->ci_ipi_events[bit].ev_count++;
 110 			(*xen_ipifunc[bit])(ci, regs);
 126 	struct cpu_info *ci;  local in function:xen_ipi_init
 129 	ci = curcpu();
 131 	vcpu = ci->ci_vcpuid
 192  struct cpu_info *ci, *self = curcpu();  local in function:xen_broadcast_ipi
 [all...]
 | 
| cpu.c | 183 static void    	cpu_boot_secondary(struct cpu_info *ci); 184 static void    	cpu_start_secondary(struct cpu_info *ci);
 199 	struct cpu_info *ci;  local in function:cpu_attach
 214 		ptr = (uintptr_t)kmem_zalloc(sizeof(*ci) + CACHE_LINE_SIZE - 1,
 216 		ci = (struct cpu_info *)roundup2(ptr, CACHE_LINE_SIZE);
 217 		ci->ci_curldt = -1;
 223 		tmp->ci_next = ci;
 225 		ci = &phycpu_info_primary;
 228 	ci->ci_self = ci;
 255  struct cpu_info *ci = sc->sc_info;  local in function:cpu_rescan
 277  struct cpu_info *ci = sc->sc_info;  local in function:cpu_childdetached
 368  struct cpu_info *ci;  local in function:cpu_attach_common
 591  struct cpu_info *ci;  local in function:cpu_boot_secondary_processors
 629  struct cpu_info *ci;  local in function:cpu_init_idle_lwps
 702  struct cpu_info *ci = (struct cpu_info *)v;  local in function:cpu_hatch
 767  struct cpu_info *ci;  local in function:cpu_debug_dump
 1108  struct cpu_info *ci = curcpu();  local in function:cpu_load_pmap
 [all...]
 | 
| /src/sys/arch/vax/vax/ | 
| multicpu.c | 82 	struct cpu_info *ci;  local in function:cpu_slavesetup 89 	ci = kmem_zalloc(sizeof(*ci), KM_SLEEP);
 90 	device_set_private(self, ci);
 91 	ci->ci_dev = self;
 92 	ci->ci_slotid = slotid;
 93 	ci->ci_cpuid = device_unit(self);
 104 	ci->ci_istack = istackbase + PAGE_SIZE;
 105 	SIMPLEQ_INSERT_TAIL(&cpus, ci, ci_next);
 108 	cq->cq_ci = ci;
 121  struct cpu_info *ci;  local in function:cpu_send_ipi
 155  struct cpu_info * const ci = curcpu();  local in function:cpu_handle_ipi
 [all...]
 | 
| /src/sys/arch/hppa/hppa/ | 
| ipifuncs.c | 61 hppa_ipi_init(struct cpu_info *ci) 63 	struct cpu_softc *sc = ci->ci_softc;
 79 	struct cpu_info *ci = curcpu();  local in function:hppa_ipi_intr
 80 	struct cpu_softc *sc = ci->ci_softc;
 85 	ipi_pending = atomic_swap_ulong(&ci->ci_ipi, 0);
 105 hppa_ipi_send(struct cpu_info *ci, u_long ipi)
 108 	KASSERT(ci->ci_flags & CPUF_RUNNING);
 110 	atomic_or_ulong(&ci->ci_ipi, (1L << ipi));
 117 	 * that the store to ci->ci_ipi above must happen before
 126 	cpu = (struct iomod *)(ci->ci_hpa)
 136  struct cpu_info *ci;  local in function:hppa_ipi_broadcast
 156  struct cpu_info *ci = curcpu();  local in function:hppa_ipi_halt
 [all...]
 | 
| /src/sys/rump/librump/rumpkern/arch/x86/ | 
| rump_x86_cpu.c | 48 rump_cpu_attach(struct cpu_info *ci) 52 		ci->ci_flags |= CPUF_PRIMARY;
 55 	ci->ci_next = cpu_info_list;
 56 	cpu_info_list = ci;
 58 	kcpuset_set(kcpuset_attached, cpu_index(ci));
 59 	kcpuset_set(kcpuset_running, cpu_index(ci));
 
 | 
| /src/sys/arch/sbmips/sbmips/ | 
| cpu.c | 87 	struct cpu_info *ci;  local in function:cpu_attach 115 		ci = curcpu();
 116 		ci->ci_cpu_freq = 50000000 * plldiv;
 118 		ci->ci_divisor_delay = (ci->ci_cpu_freq + 500000) / 1000000;
 120 		ci->ci_cycles_per_hz = (ci->ci_cpu_freq + hz / 2 ) / hz;
 123 		    ci->ci_cpu_freq / 1000000,
 124 		    (ci->ci_cpu_freq % 1000000) / 10000,
 125 		    ci->ci_cycles_per_hz, ci->ci_divisor_delay)
 [all...]
 | 
| /src/sys/sys/ | 
| cctr.h | 52 #define	cc_hardclock(ci)						\ 54 	if ((ci)->ci_cc.cc_cal_ticks &&					\
 55 	    ++(ci)->ci_cc.cc_ticks == (ci)->ci_cc.cc_cal_ticks) {	\
 56 		(ci)->ci_cc.cc_ticks = 0;				\
 57 		cc_calibrate_cpu((ci));					\
 
 | 
| /src/sys/arch/x86/x86/ | 
| cpu.c | 211 static void	cpu_boot_secondary(struct cpu_info *ci); 212 static void	cpu_start_secondary(struct cpu_info *ci);
 248 cpu_pcpuarea_init(struct cpu_info *ci)
 258 	base = (vaddr_t)&pcpuarea->ent[cpu_index(ci)];
 277 cpu_vm_init(struct cpu_info *ci)
 291 		cai = &ci->ci_cinfo[i];
 331 	aprint_debug_dev(ci->ci_dev, "%d page colors\n", ncolors);
 334 	pmap_tlb_cpu_init(ci);
 336 	pmap_vpage_cpu_init(ci);
 345 	struct cpu_info *ci;  local in function:cpu_attach
 572  struct cpu_info *ci = sc->sc_info;  local in function:cpu_rescan
 624  struct cpu_info *ci = sc->sc_info;  local in function:cpu_childdetached
 754  struct cpu_info *ci;  local in function:cpu_boot_secondary_processors
 804  struct cpu_info *ci;  local in function:cpu_init_idle_lwps
 911  struct cpu_info *ci = (struct cpu_info *)v;  local in function:cpu_hatch
 1071  struct cpu_info *ci;  local in function:cpu_debug_dump
 1329  struct cpu_info *ci = sc->sc_info;  local in function:cpu_stop
 1358  struct cpu_info *ci = sc->sc_info;  local in function:cpu_suspend
 1373  struct cpu_info *ci = sc->sc_info;  local in function:cpu_resume
 1404  struct cpu_info *ci = sc->sc_info;  local in function:cpu_shutdown
 1513  struct cpu_info *ci = curcpu();  local in function:x86_cpu_idle_mwait
 1527  struct cpu_info *ci = curcpu();  local in function:x86_cpu_idle_halt
 1555  struct cpu_info *ci = curcpu();  local in function:cpu_load_pmap
 [all...]
 | 
| ipi.c | 100 x86_send_ipi(struct cpu_info *ci, int ipimask) 106 	if (__predict_false((ci->ci_flags & CPUF_RUNNING) == 0))
 111 		n = atomic_cas_32(&ci->ci_ipis, o, o | ipimask);
 119 		ret = x86_ipi(LAPIC_IPI_VECTOR, ci->ci_cpuid, LAPIC_DLMODE_FIXED);
 124 			    device_xname(ci->ci_dev));
 134 	struct cpu_info *ci, *self = curcpu();  local in function:x86_broadcast_ipi
 138 	for (CPU_INFO_FOREACH(cii, ci)) {
 139 		if (ci == self)
 141 		if ((ci->ci_flags & CPUF_RUNNING) == 0)
 143 		atomic_or_32(&ci->ci_ipis, ipimask)
 155  struct cpu_info *ci = curcpu();  local in function:x86_ipi_handler
 [all...]
 | 
| /src/sys/kern/ | 
| subr_cpu.c | 103 	struct cpu_info *ci;  local in function:mi_cpu_init 111 	ci = curcpu();
 112 	cpu_topology_fake1(ci);
 173 cpu_topology_set(struct cpu_info *ci, u_int package_id, u_int core_id,
 179 	ci->ci_package_id = package_id;
 180 	ci->ci_core_id = core_id;
 181 	ci->ci_smt_id = smt_id;
 182 	ci->ci_numa_id = numa_id;
 183 	for (rel = 0; rel < __arraycount(ci->ci_sibling); rel++) {
 184 		ci->ci_sibling[rel] = ci
 227  struct cpu_info *ci, *ci2;  local in function:cpu_topology_dump
 292  struct cpu_info *ci;  local in function:cpu_topology_fake
 310  struct cpu_info *ci, *ci2, *ci3;  local in function:cpu_topology_init
 492  struct cpu_info *ci;  local in function:cpu_count_sync
 [all...]
 | 
| kern_cpu.c | 129 mi_cpu_attach(struct cpu_info *ci) 135 	if ((ci->ci_index = ncpu) >= maxcpus)
 137 	kcpuset_set(kcpuset_attached, cpu_index(ci));
 142 	kcpuset_create(&ci->ci_kcpuset, true);
 143 	kcpuset_set(ci->ci_kcpuset, cpu_index(ci));
 145 	TAILQ_INIT(&ci->ci_data.cpu_ld_locks);
 146 	__cpu_simple_lock_init(&ci->ci_data.cpu_ld_lock);
 149 	snprintf(ci->ci_data.cpu_name, sizeof(ci->ci_data.cpu_name), "cpu%d"
 197  struct cpu_info *ci;  local in function:cpuctl_ioctl
 289  struct cpu_info *ci;  local in function:cpu_lookup
 [all...]
 | 
| /src/sys/arch/evbmips/evbmips/ | 
| cpu.c | 64 	struct cpu_info * const ci = curcpu();  local in function:cpu_attach 66 	ci->ci_dev = self;
 67 	device_set_private(self, ci);
 70 	    ci->ci_cpu_freq / 1000000,
 71 	    (ci->ci_cpu_freq % 1000000) / 10000,
 72 	    ci->ci_cycles_per_hz, ci->ci_divisor_delay);
 76 	cpu_attach_common(self, ci);
 
 | 
| interrupt.c | 53 	struct cpu_info * const ci = curcpu();  local in function:cpu_intr 56 	const int mtx_count = ci->ci_mtx_count;
 57 	const u_int biglock_count = ci->ci_biglock_count;
 60 	KASSERT(ci->ci_cpl == IPL_HIGH);
 63 	ci->ci_data.cpu_nintr++;
 70 		KASSERTMSG(ci->ci_cpl == ipl,
 71 		    "%s: cpl (%d) != ipl (%d)", __func__, ci->ci_cpl, ipl);
 77 			.intr = (ci->ci_idepth > 1)
 96 		KASSERT(biglock_count == ci->ci_biglock_count);
 98 		KASSERT(mtx_count == ci->ci_mtx_count)
 [all...]
 | 
| /src/lib/libc/citrus/ | 
| citrus_iconv.c | 139 close_shared(struct _citrus_iconv_shared *ci) 141 	if (ci) {
 142 		if (ci->ci_module) {
 143 			if (ci->ci_ops) {
 144 				if (ci->ci_closure)
 145 					(*ci->ci_ops->io_uninit_shared)(ci);
 146 				free(ci->ci_ops);
 148 			_citrus_unload_module(ci->ci_module);
 150 		free(ci);
 160  struct _citrus_iconv_shared *ci;  local in function:open_shared
 264  struct _citrus_iconv_shared * ci;  local in function:get_shared
 335  struct _citrus_iconv_shared *ci = NULL;  local in function:_citrus_iconv_open
 [all...]
 | 
| /src/sys/arch/powerpc/powerpc/ | 
| softint_machdep.c | 55 softint_deliver(struct cpu_info *ci, int ipl) 58 	KASSERT(ci->ci_data.cpu_softints & (1 << ipl));
 59 	ci->ci_data.cpu_softints ^= 1 << ipl;
 60 	softint_fast_dispatch(ci->ci_softlwps[si_level], ipl);
 61 	KASSERTMSG(ci->ci_cpl == IPL_HIGH,
 62 	    "%s: cpl (%d) != HIGH", __func__, ci->ci_cpl);
 66 powerpc_softint(struct cpu_info *ci, int old_ipl, vaddr_t pc)
 71 	KASSERTMSG(ci->ci_idepth == -1,
 73 	     cpu_index(ci), ci->ci_idepth)
 109  struct cpu_info * const ci = l->l_cpu;  local in function:powerpc_softint_init_md
 119  struct cpu_info * const ci = curcpu();  local in function:powerpc_softint_trigger
 [all...]
 | 
| /src/usr.sbin/cpuctl/arch/ | 
| i386.c | 673 disable_tsc(struct cpu_info *ci) 675 	if (ci->ci_feat_val[0] & CPUID_TSC) {
 676 		ci->ci_feat_val[0] &= ~CPUID_TSC;
 682 amd_family5_setup(struct cpu_info *ci)
 685 	switch (ci->ci_model) {
 693 		if (ci->ci_feat_val[0] & CPUID_APIC)
 694 			ci->ci_feat_val[0] =
 695 			    (ci->ci_feat_val[0] & ~CPUID_APIC) | CPUID_PGE;
 705 cyrix6x86_cpu_setup(struct cpu_info *ci)
 712 	if (ci->ci_signature != 0x552
 1975  struct cpu_info *ci, cistore;  local in function:identifycpu
 2593  struct cpu_info ci;  local in function:ucodeupdate_check
 [all...]
 | 
| /src/sys/arch/riscv/riscv/ | 
| ipifuncs.c | 60 ipi_nop(struct cpu_info *ci) 69 ipi_ast(struct cpu_info *ci)
 71 	ci->ci_onproc->l_md.md_astpending = 1;
 76 ipi_kpreempt(struct cpu_info *ci)
 99 ipi_process(struct cpu_info *ci, unsigned long ipi_mask)
 104 		ci->ci_evcnt_per_ipi[IPI_NOP].ev_count++;
 105 		ipi_nop(ci);
 108 		ci->ci_evcnt_per_ipi[IPI_AST].ev_count++;
 109 		ipi_ast(ci);
 112 		ci->ci_evcnt_per_ipi[IPI_SUSPEND].ev_count++
 [all...]
 | 
| clock_machdep.c | 94 	struct cpu_info * const ci = curcpu();  local in function:riscv_timer_init 96 	evcnt_attach_dynamic(&ci->ci_ev_timer, EVCNT_TYPE_INTR,
 97 	    NULL, device_xname(ci->ci_dev), "timer");
 99 	ci->ci_lastintr = csr_time_read();
 100 	uint64_t next = ci->ci_lastintr + timer_ticks_per_hz;
 101 	ci->ci_lastintr_scheduled = next;
 106 	if (cpu_index(ci) == 0) {
 116 	struct cpu_info * const ci = curcpu();  local in function:riscv_timer_intr
 123 	ci->ci_lastintr = now;
 124 	ci->ci_ev_timer.ev_count++
 [all...]
 |