HomeSort by: relevance | last modified time | path
    Searched refs:ci (Results 1 - 25 of 448) sorted by relevancy

1 2 3 4 5 6 7 8 91011>>

  /src/sys/arch/sparc/sparc/
cache_print.h 37 static void cache_printf_backend(struct cacheinfo *ci, const char *cpuname);
40 cache_printf_backend(struct cacheinfo *ci, const char *cpuname)
43 if (ci->c_flags & CACHE_TRAPPAGEBUG)
48 if (ci->c_totalsize == 0) {
53 if (ci->c_split) {
56 printf("%s", (ci->c_physical ? "physical " : ""));
57 if (ci->ic_totalsize > 0) {
59 ci->ic_totalsize/1024, ci->ic_linesize);
62 if (ci->dc_totalsize > 0)
    [all...]
  /src/tests/usr.bin/xlint/lint1/
d_gcc_compound_statements2.c 20 struct cpu_info *ci;
21 __asm__ volatile("movl %%fs:4,%0":"=r" (ci));
22 ci;
29 struct cpu_info ci = { 0 }; local in function:statement_expr_with_only_stmt
31 if (ci.bar > 0)
32 ci.bar++;
33 ci;
  /src/sys/rump/librump/rumpkern/arch/generic/
rump_generic_cpu.c 41 rump_cpu_attach(struct cpu_info *ci)
46 ci->ci_next = rumpcpu_info_list;
47 rumpcpu_info_list = ci;
49 ci->ci_index = nattached++;
51 kcpuset_set(kcpuset_attached, cpu_index(ci));
52 kcpuset_set(kcpuset_running, cpu_index(ci));
  /src/sys/arch/arm/pic/
pic_splfuncs.c 55 struct cpu_info * const ci = curcpu(); local in function:_splraise
56 const int oldipl = ci->ci_cpl;
58 if (newipl > ci->ci_cpl) {
59 pic_set_priority(ci, newipl);
67 struct cpu_info * const ci = curcpu(); local in function:_spllower
68 const int oldipl = ci->ci_cpl;
69 KASSERT(panicstr || newipl <= ci->ci_cpl);
70 if (newipl < ci->ci_cpl) {
72 ci->ci_intr_depth++;
74 ci->ci_intr_depth--
85 struct cpu_info * const ci = curcpu(); local in function:splx
    [all...]
  /src/sys/arch/mips/mips/
ipifuncs.c 68 ipi_nop(struct cpu_info *ci)
77 ipi_ast(struct cpu_info *ci)
79 ci->ci_onproc->l_md.md_astpending = 1;
83 ipi_shootdown(struct cpu_info *ci)
89 ipi_syncicache(struct cpu_info *ci)
91 pmap_tlb_syncicache_wanted(ci);
96 ipi_kpreempt(struct cpu_info *ci)
119 ipi_process(struct cpu_info *ci, uint64_t ipi_mask)
124 ci->ci_evcnt_per_ipi[IPI_NOP].ev_count++;
125 ipi_nop(ci);
    [all...]
mips3_clockintr.c 68 struct cpu_info * const ci = curcpu(); local in function:mips3_clockintr
71 ci->ci_ev_count_compare.ev_count++;
73 KASSERT((ci->ci_cycles_per_hz & ~(0xffffffff)) == 0);
74 ci->ci_next_cp0_clk_intr += (uint32_t)(ci->ci_cycles_per_hz & 0xffffffff);
75 mips3_cp0_compare_write(ci->ci_next_cp0_clk_intr);
84 if ((ci->ci_next_cp0_clk_intr - new_cnt) & 0x80000000) {
86 ci->ci_next_cp0_clk_intr = new_cnt + curcpu()->ci_cycles_per_hz;
87 mips3_cp0_compare_write(ci->ci_next_cp0_clk_intr);
106 struct cpu_info * const ci = curcpu() local in function:mips3_init_cp0_clocks
    [all...]
mips_softint.c 62 struct cpu_info * const ci = l->l_cpu; local in function:softint_init_md
65 ci->ci_softlwps[si_level] = l;
100 ci->ci_softints ^= SOFTINT_##level##_MASK; \
101 softint_fast_dispatch(ci->ci_softlwps[SOFTINT_##level], \
103 KASSERTMSG(ci->ci_cpl == IPL_HIGH, "cpl (%d) != HIGH", ci->ci_cpl); \
110 struct cpu_info * const ci = curcpu(); local in function:softint_process
115 KASSERT(ci->ci_cpl == IPL_HIGH);
117 KASSERTMSG(ci->ci_mtx_count == 0,
119 __func__, cpu_index(ci), ci, ci->ci_mtx_count)
    [all...]
  /src/sys/arch/xen/x86/
xen_ipi.c 74 static void xen_ipi_kpreempt(struct cpu_info *ci, struct intrframe *);
97 struct cpu_info *ci; local in function:xen_ipi_handler
99 ci = curcpu();
101 KASSERT(ci == arg);
102 pending = atomic_swap_32(&ci->ci_ipis, 0);
108 ci->ci_ipi_events[bit].ev_count++;
110 (*xen_ipifunc[bit])(ci, regs);
126 struct cpu_info *ci; local in function:xen_ipi_init
129 ci = curcpu();
131 vcpu = ci->ci_vcpuid
192 struct cpu_info *ci, *self = curcpu(); local in function:xen_broadcast_ipi
    [all...]
cpu.c 183 static void cpu_boot_secondary(struct cpu_info *ci);
184 static void cpu_start_secondary(struct cpu_info *ci);
199 struct cpu_info *ci; local in function:cpu_attach
214 ptr = (uintptr_t)kmem_zalloc(sizeof(*ci) + CACHE_LINE_SIZE - 1,
216 ci = (struct cpu_info *)roundup2(ptr, CACHE_LINE_SIZE);
217 ci->ci_curldt = -1;
223 tmp->ci_next = ci;
225 ci = &phycpu_info_primary;
228 ci->ci_self = ci;
255 struct cpu_info *ci = sc->sc_info; local in function:cpu_rescan
277 struct cpu_info *ci = sc->sc_info; local in function:cpu_childdetached
368 struct cpu_info *ci; local in function:cpu_attach_common
591 struct cpu_info *ci; local in function:cpu_boot_secondary_processors
629 struct cpu_info *ci; local in function:cpu_init_idle_lwps
702 struct cpu_info *ci = (struct cpu_info *)v; local in function:cpu_hatch
767 struct cpu_info *ci; local in function:cpu_debug_dump
1108 struct cpu_info *ci = curcpu(); local in function:cpu_load_pmap
    [all...]
  /src/sys/arch/vax/vax/
multicpu.c 82 struct cpu_info *ci; local in function:cpu_slavesetup
89 ci = kmem_zalloc(sizeof(*ci), KM_SLEEP);
90 device_set_private(self, ci);
91 ci->ci_dev = self;
92 ci->ci_slotid = slotid;
93 ci->ci_cpuid = device_unit(self);
104 ci->ci_istack = istackbase + PAGE_SIZE;
105 SIMPLEQ_INSERT_TAIL(&cpus, ci, ci_next);
108 cq->cq_ci = ci;
121 struct cpu_info *ci; local in function:cpu_send_ipi
155 struct cpu_info * const ci = curcpu(); local in function:cpu_handle_ipi
    [all...]
  /src/sys/arch/hppa/hppa/
ipifuncs.c 61 hppa_ipi_init(struct cpu_info *ci)
63 struct cpu_softc *sc = ci->ci_softc;
79 struct cpu_info *ci = curcpu(); local in function:hppa_ipi_intr
80 struct cpu_softc *sc = ci->ci_softc;
85 ipi_pending = atomic_swap_ulong(&ci->ci_ipi, 0);
105 hppa_ipi_send(struct cpu_info *ci, u_long ipi)
108 KASSERT(ci->ci_flags & CPUF_RUNNING);
110 atomic_or_ulong(&ci->ci_ipi, (1L << ipi));
117 * that the store to ci->ci_ipi above must happen before
126 cpu = (struct iomod *)(ci->ci_hpa)
136 struct cpu_info *ci; local in function:hppa_ipi_broadcast
156 struct cpu_info *ci = curcpu(); local in function:hppa_ipi_halt
    [all...]
  /src/sys/rump/librump/rumpkern/arch/x86/
rump_x86_cpu.c 48 rump_cpu_attach(struct cpu_info *ci)
52 ci->ci_flags |= CPUF_PRIMARY;
55 ci->ci_next = cpu_info_list;
56 cpu_info_list = ci;
58 kcpuset_set(kcpuset_attached, cpu_index(ci));
59 kcpuset_set(kcpuset_running, cpu_index(ci));
  /src/sys/arch/sbmips/sbmips/
cpu.c 87 struct cpu_info *ci; local in function:cpu_attach
115 ci = curcpu();
116 ci->ci_cpu_freq = 50000000 * plldiv;
118 ci->ci_divisor_delay = (ci->ci_cpu_freq + 500000) / 1000000;
120 ci->ci_cycles_per_hz = (ci->ci_cpu_freq + hz / 2 ) / hz;
123 ci->ci_cpu_freq / 1000000,
124 (ci->ci_cpu_freq % 1000000) / 10000,
125 ci->ci_cycles_per_hz, ci->ci_divisor_delay)
    [all...]
  /src/sys/sys/
cctr.h 52 #define cc_hardclock(ci) \
54 if ((ci)->ci_cc.cc_cal_ticks && \
55 ++(ci)->ci_cc.cc_ticks == (ci)->ci_cc.cc_cal_ticks) { \
56 (ci)->ci_cc.cc_ticks = 0; \
57 cc_calibrate_cpu((ci)); \
  /src/sys/arch/x86/x86/
cpu.c 211 static void cpu_boot_secondary(struct cpu_info *ci);
212 static void cpu_start_secondary(struct cpu_info *ci);
248 cpu_pcpuarea_init(struct cpu_info *ci)
258 base = (vaddr_t)&pcpuarea->ent[cpu_index(ci)];
277 cpu_vm_init(struct cpu_info *ci)
291 cai = &ci->ci_cinfo[i];
331 aprint_debug_dev(ci->ci_dev, "%d page colors\n", ncolors);
334 pmap_tlb_cpu_init(ci);
336 pmap_vpage_cpu_init(ci);
345 struct cpu_info *ci; local in function:cpu_attach
572 struct cpu_info *ci = sc->sc_info; local in function:cpu_rescan
624 struct cpu_info *ci = sc->sc_info; local in function:cpu_childdetached
754 struct cpu_info *ci; local in function:cpu_boot_secondary_processors
804 struct cpu_info *ci; local in function:cpu_init_idle_lwps
911 struct cpu_info *ci = (struct cpu_info *)v; local in function:cpu_hatch
1071 struct cpu_info *ci; local in function:cpu_debug_dump
1329 struct cpu_info *ci = sc->sc_info; local in function:cpu_stop
1358 struct cpu_info *ci = sc->sc_info; local in function:cpu_suspend
1373 struct cpu_info *ci = sc->sc_info; local in function:cpu_resume
1404 struct cpu_info *ci = sc->sc_info; local in function:cpu_shutdown
1513 struct cpu_info *ci = curcpu(); local in function:x86_cpu_idle_mwait
1527 struct cpu_info *ci = curcpu(); local in function:x86_cpu_idle_halt
1555 struct cpu_info *ci = curcpu(); local in function:cpu_load_pmap
    [all...]
ipi.c 100 x86_send_ipi(struct cpu_info *ci, int ipimask)
106 if (__predict_false((ci->ci_flags & CPUF_RUNNING) == 0))
111 n = atomic_cas_32(&ci->ci_ipis, o, o | ipimask);
119 ret = x86_ipi(LAPIC_IPI_VECTOR, ci->ci_cpuid, LAPIC_DLMODE_FIXED);
124 device_xname(ci->ci_dev));
134 struct cpu_info *ci, *self = curcpu(); local in function:x86_broadcast_ipi
138 for (CPU_INFO_FOREACH(cii, ci)) {
139 if (ci == self)
141 if ((ci->ci_flags & CPUF_RUNNING) == 0)
143 atomic_or_32(&ci->ci_ipis, ipimask)
155 struct cpu_info *ci = curcpu(); local in function:x86_ipi_handler
    [all...]
  /src/sys/kern/
subr_cpu.c 103 struct cpu_info *ci; local in function:mi_cpu_init
111 ci = curcpu();
112 cpu_topology_fake1(ci);
173 cpu_topology_set(struct cpu_info *ci, u_int package_id, u_int core_id,
179 ci->ci_package_id = package_id;
180 ci->ci_core_id = core_id;
181 ci->ci_smt_id = smt_id;
182 ci->ci_numa_id = numa_id;
183 for (rel = 0; rel < __arraycount(ci->ci_sibling); rel++) {
184 ci->ci_sibling[rel] = ci
227 struct cpu_info *ci, *ci2; local in function:cpu_topology_dump
292 struct cpu_info *ci; local in function:cpu_topology_fake
310 struct cpu_info *ci, *ci2, *ci3; local in function:cpu_topology_init
492 struct cpu_info *ci; local in function:cpu_count_sync
    [all...]
kern_cpu.c 129 mi_cpu_attach(struct cpu_info *ci)
135 if ((ci->ci_index = ncpu) >= maxcpus)
137 kcpuset_set(kcpuset_attached, cpu_index(ci));
142 kcpuset_create(&ci->ci_kcpuset, true);
143 kcpuset_set(ci->ci_kcpuset, cpu_index(ci));
145 TAILQ_INIT(&ci->ci_data.cpu_ld_locks);
146 __cpu_simple_lock_init(&ci->ci_data.cpu_ld_lock);
149 snprintf(ci->ci_data.cpu_name, sizeof(ci->ci_data.cpu_name), "cpu%d"
197 struct cpu_info *ci; local in function:cpuctl_ioctl
289 struct cpu_info *ci; local in function:cpu_lookup
    [all...]
  /src/sys/arch/evbmips/evbmips/
cpu.c 64 struct cpu_info * const ci = curcpu(); local in function:cpu_attach
66 ci->ci_dev = self;
67 device_set_private(self, ci);
70 ci->ci_cpu_freq / 1000000,
71 (ci->ci_cpu_freq % 1000000) / 10000,
72 ci->ci_cycles_per_hz, ci->ci_divisor_delay);
76 cpu_attach_common(self, ci);
interrupt.c 53 struct cpu_info * const ci = curcpu(); local in function:cpu_intr
56 const int mtx_count = ci->ci_mtx_count;
57 const u_int biglock_count = ci->ci_biglock_count;
60 KASSERT(ci->ci_cpl == IPL_HIGH);
63 ci->ci_data.cpu_nintr++;
70 KASSERTMSG(ci->ci_cpl == ipl,
71 "%s: cpl (%d) != ipl (%d)", __func__, ci->ci_cpl, ipl);
77 .intr = (ci->ci_idepth > 1)
96 KASSERT(biglock_count == ci->ci_biglock_count);
98 KASSERT(mtx_count == ci->ci_mtx_count)
    [all...]
  /src/lib/libc/citrus/
citrus_iconv.c 139 close_shared(struct _citrus_iconv_shared *ci)
141 if (ci) {
142 if (ci->ci_module) {
143 if (ci->ci_ops) {
144 if (ci->ci_closure)
145 (*ci->ci_ops->io_uninit_shared)(ci);
146 free(ci->ci_ops);
148 _citrus_unload_module(ci->ci_module);
150 free(ci);
160 struct _citrus_iconv_shared *ci; local in function:open_shared
264 struct _citrus_iconv_shared * ci; local in function:get_shared
335 struct _citrus_iconv_shared *ci = NULL; local in function:_citrus_iconv_open
    [all...]
  /src/sys/arch/powerpc/powerpc/
softint_machdep.c 55 softint_deliver(struct cpu_info *ci, int ipl)
58 KASSERT(ci->ci_data.cpu_softints & (1 << ipl));
59 ci->ci_data.cpu_softints ^= 1 << ipl;
60 softint_fast_dispatch(ci->ci_softlwps[si_level], ipl);
61 KASSERTMSG(ci->ci_cpl == IPL_HIGH,
62 "%s: cpl (%d) != HIGH", __func__, ci->ci_cpl);
66 powerpc_softint(struct cpu_info *ci, int old_ipl, vaddr_t pc)
71 KASSERTMSG(ci->ci_idepth == -1,
73 cpu_index(ci), ci->ci_idepth)
109 struct cpu_info * const ci = l->l_cpu; local in function:powerpc_softint_init_md
119 struct cpu_info * const ci = curcpu(); local in function:powerpc_softint_trigger
    [all...]
  /src/usr.sbin/cpuctl/arch/
i386.c 673 disable_tsc(struct cpu_info *ci)
675 if (ci->ci_feat_val[0] & CPUID_TSC) {
676 ci->ci_feat_val[0] &= ~CPUID_TSC;
682 amd_family5_setup(struct cpu_info *ci)
685 switch (ci->ci_model) {
693 if (ci->ci_feat_val[0] & CPUID_APIC)
694 ci->ci_feat_val[0] =
695 (ci->ci_feat_val[0] & ~CPUID_APIC) | CPUID_PGE;
705 cyrix6x86_cpu_setup(struct cpu_info *ci)
712 if (ci->ci_signature != 0x552
1975 struct cpu_info *ci, cistore; local in function:identifycpu
2593 struct cpu_info ci; local in function:ucodeupdate_check
    [all...]
  /src/sys/arch/riscv/riscv/
ipifuncs.c 60 ipi_nop(struct cpu_info *ci)
69 ipi_ast(struct cpu_info *ci)
71 ci->ci_onproc->l_md.md_astpending = 1;
76 ipi_kpreempt(struct cpu_info *ci)
99 ipi_process(struct cpu_info *ci, unsigned long ipi_mask)
104 ci->ci_evcnt_per_ipi[IPI_NOP].ev_count++;
105 ipi_nop(ci);
108 ci->ci_evcnt_per_ipi[IPI_AST].ev_count++;
109 ipi_ast(ci);
112 ci->ci_evcnt_per_ipi[IPI_SUSPEND].ev_count++
    [all...]
clock_machdep.c 94 struct cpu_info * const ci = curcpu(); local in function:riscv_timer_init
96 evcnt_attach_dynamic(&ci->ci_ev_timer, EVCNT_TYPE_INTR,
97 NULL, device_xname(ci->ci_dev), "timer");
99 ci->ci_lastintr = csr_time_read();
100 uint64_t next = ci->ci_lastintr + timer_ticks_per_hz;
101 ci->ci_lastintr_scheduled = next;
106 if (cpu_index(ci) == 0) {
116 struct cpu_info * const ci = curcpu(); local in function:riscv_timer_intr
123 ci->ci_lastintr = now;
124 ci->ci_ev_timer.ev_count++
    [all...]

Completed in 35 milliseconds

1 2 3 4 5 6 7 8 91011>>