Home | History | Annotate | Line # | Download | only in tprof
      1 /* $NetBSD: tprof_armv8.c,v 1.20 2023/04/11 10:07:12 msaitoh Exp $ */
      2 
      3 /*-
      4  * Copyright (c) 2018 Jared McNeill <jmcneill (at) invisible.ca>
      5  * All rights reserved.
      6  *
      7  * Redistribution and use in source and binary forms, with or without
      8  * modification, are permitted provided that the following conditions
      9  * are met:
     10  * 1. Redistributions of source code must retain the above copyright
     11  *    notice, this list of conditions and the following disclaimer.
     12  * 2. Redistributions in binary form must reproduce the above copyright
     13  *    notice, this list of conditions and the following disclaimer in the
     14  *    documentation and/or other materials provided with the distribution.
     15  *
     16  * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR
     17  * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
     18  * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
     19  * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT,
     20  * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING,
     21  * BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
     22  * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED
     23  * AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY,
     24  * OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
     25  * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
     26  * SUCH DAMAGE.
     27  */
     28 
     29 #include <sys/cdefs.h>
     30 __KERNEL_RCSID(0, "$NetBSD: tprof_armv8.c,v 1.20 2023/04/11 10:07:12 msaitoh Exp $");
     31 
     32 #include <sys/param.h>
     33 #include <sys/bus.h>
     34 #include <sys/cpu.h>
     35 #include <sys/percpu.h>
     36 #include <sys/xcall.h>
     37 
     38 #include <dev/tprof/tprof.h>
     39 
     40 #include <arm/armreg.h>
     41 #include <arm/cpufunc.h>
     42 
     43 #include <dev/tprof/tprof_armv8.h>
     44 
     45 static u_int counter_bitwidth;
     46 
     47 /*
     48  * armv8 can handle up to 31 event counters,
     49  * PMCR_EL0.N counters are actually available.
     50  */
     51 
     52 static bool
     53 armv8_pmu_event_implemented(uint16_t event)
     54 {
     55 	uint64_t eid[2];
     56 
     57 	if (event >= 64)
     58 		return false;
     59 
     60 	eid[0] = reg_pmceid0_el0_read();
     61 	eid[1] = reg_pmceid1_el0_read();
     62 
     63 	/* The low 32bits of PMCEID[01]_EL0 contain the common events 0 to n */
     64 	const u_int idx = event / 32;
     65 	const u_int bit = event % 32;
     66 
     67 	if (eid[idx] & __BIT(bit))
     68 		return true;
     69 
     70 	return false;
     71 }
     72 
     73 static void
     74 armv8_pmu_set_pmevtyper(u_int counter, uint64_t val)
     75 {
     76 	reg_pmselr_el0_write(counter);
     77 	isb();
     78 	reg_pmxevtyper_el0_write(val);
     79 }
     80 
     81 static inline void
     82 armv8_pmu_set_pmevcntr(u_int counter, uint64_t val)
     83 {
     84 	reg_pmselr_el0_write(counter);
     85 	isb();
     86 	reg_pmxevcntr_el0_write(val);
     87 }
     88 
     89 static inline uint64_t
     90 armv8_pmu_get_pmevcntr(u_int counter)
     91 {
     92 	reg_pmselr_el0_write(counter);
     93 	isb();
     94 	return reg_pmxevcntr_el0_read();
     95 }
     96 
     97 /* Read and write at once */
     98 static inline uint64_t
     99 armv8_pmu_getset_pmevcntr(u_int counter, uint64_t val)
    100 {
    101 	uint64_t c;
    102 
    103 	reg_pmselr_el0_write(counter);
    104 	isb();
    105 	c = reg_pmxevcntr_el0_read();
    106 	reg_pmxevcntr_el0_write(val);
    107 	return c;
    108 }
    109 
    110 static uint32_t
    111 armv8_pmu_ncounters(void)
    112 {
    113 	return __SHIFTOUT(reg_pmcr_el0_read(), PMCR_N);
    114 }
    115 
    116 static u_int
    117 armv8_pmu_counter_bitwidth(u_int counter)
    118 {
    119 	return counter_bitwidth;
    120 }
    121 
    122 static uint64_t
    123 armv8_pmu_counter_estimate_freq(u_int counter)
    124 {
    125 	return curcpu()->ci_data.cpu_cc_freq;
    126 }
    127 
    128 static int
    129 armv8_pmu_valid_event(u_int counter, const tprof_param_t *param)
    130 {
    131 	if (!armv8_pmu_event_implemented(param->p_event)) {
    132 		printf("%s: event %#" PRIx64 " not implemented on this CPU\n",
    133 		    __func__, param->p_event);
    134 		return EINVAL;
    135 	}
    136 	return 0;
    137 }
    138 
    139 static void
    140 armv8_pmu_configure_event(u_int counter, const tprof_param_t *param)
    141 {
    142 	/* Disable event counter */
    143 	reg_pmcntenclr_el0_write(__BIT(counter) & PMCNTEN_P);
    144 
    145 	/* Disable overflow interrupts */
    146 	reg_pmintenclr_el1_write(__BIT(counter) & PMINTEN_P);
    147 
    148 	/* Configure event counter */
    149 	uint64_t pmevtyper = __SHIFTIN(param->p_event, PMEVTYPER_EVTCOUNT);
    150 	if (!ISSET(param->p_flags, TPROF_PARAM_USER))
    151 		pmevtyper |= PMEVTYPER_U;
    152 	if (!ISSET(param->p_flags, TPROF_PARAM_KERN))
    153 		pmevtyper |= PMEVTYPER_P;
    154 	armv8_pmu_set_pmevtyper(counter, pmevtyper);
    155 
    156 	if (ISSET(param->p_flags, TPROF_PARAM_PROFILE) ||
    157 	    counter_bitwidth != 64) {
    158 		/* Enable overflow interrupts */
    159 		reg_pmintenset_el1_write(__BIT(counter) & PMINTEN_P);
    160 	}
    161 
    162 	/* Clear overflow flag */
    163 	reg_pmovsclr_el0_write(__BIT(counter) & PMOVS_P);
    164 
    165 	/* Reset the counter */
    166 	armv8_pmu_set_pmevcntr(counter, param->p_value);
    167 }
    168 
    169 static void
    170 armv8_pmu_start(tprof_countermask_t runmask)
    171 {
    172 	/* Enable event counters */
    173 	reg_pmcntenset_el0_write(runmask & PMCNTEN_P);
    174 
    175 	/*
    176 	 * PMCR.E is shared with PMCCNTR_EL0 and event counters.
    177 	 * It is set here in case PMCCNTR_EL0 is not used in the system.
    178 	 */
    179 	reg_pmcr_el0_write(reg_pmcr_el0_read() | PMCR_E);
    180 }
    181 
    182 static void
    183 armv8_pmu_stop(tprof_countermask_t stopmask)
    184 {
    185 	/* Disable event counter */
    186 	reg_pmcntenclr_el0_write(stopmask & PMCNTEN_P);
    187 }
    188 
    189 /* XXX: argument of armv8_pmu_intr() */
    190 extern struct tprof_backend *tprof_backend;
    191 static void *pmu_intr_arg;
    192 
    193 int
    194 armv8_pmu_intr(void *priv)
    195 {
    196 	const struct trapframe * const tf = priv;
    197 	tprof_backend_softc_t *sc = pmu_intr_arg;
    198 	tprof_frame_info_t tfi;
    199 	int bit;
    200 	const uint32_t pmovs = reg_pmovsset_el0_read();
    201 
    202 	uint64_t *counters_offset =
    203 	    percpu_getptr_remote(sc->sc_ctr_offset_percpu, curcpu());
    204 	uint32_t mask = pmovs;
    205 	while ((bit = ffs(mask)) != 0) {
    206 		bit--;
    207 		CLR(mask, __BIT(bit));
    208 
    209 		if (ISSET(sc->sc_ctr_prof_mask, __BIT(bit))) {
    210 			/* Account for the counter, and reset */
    211 			uint64_t ctr = armv8_pmu_getset_pmevcntr(bit,
    212 			    sc->sc_count[bit].ctr_counter_reset_val);
    213 			counters_offset[bit] +=
    214 			    sc->sc_count[bit].ctr_counter_val + ctr;
    215 
    216 			/* Record a sample */
    217 			tfi.tfi_pc = tf->tf_pc;
    218 			tfi.tfi_counter = bit;
    219 			tfi.tfi_inkernel =
    220 			    tfi.tfi_pc >= VM_MIN_KERNEL_ADDRESS &&
    221 			    tfi.tfi_pc < VM_MAX_KERNEL_ADDRESS;
    222 			tprof_sample(NULL, &tfi);
    223 		} else if (ISSET(sc->sc_ctr_ovf_mask, __BIT(bit))) {
    224 			/* Counter has overflowed */
    225 			counters_offset[bit] += __BIT(32);
    226 		}
    227 	}
    228 	reg_pmovsclr_el0_write(pmovs);
    229 
    230 	return 1;
    231 }
    232 
    233 static uint32_t
    234 armv8_pmu_ident(void)
    235 {
    236 	return TPROF_IDENT_ARMV8_GENERIC;
    237 }
    238 
    239 static const tprof_backend_ops_t tprof_armv8_pmu_ops = {
    240 	.tbo_ident = armv8_pmu_ident,
    241 	.tbo_ncounters = armv8_pmu_ncounters,
    242 	.tbo_counter_bitwidth = armv8_pmu_counter_bitwidth,
    243 	.tbo_counter_read = armv8_pmu_get_pmevcntr,
    244 	.tbo_counter_estimate_freq = armv8_pmu_counter_estimate_freq,
    245 	.tbo_valid_event = armv8_pmu_valid_event,
    246 	.tbo_configure_event = armv8_pmu_configure_event,
    247 	.tbo_start = armv8_pmu_start,
    248 	.tbo_stop = armv8_pmu_stop,
    249 	.tbo_establish = NULL,
    250 	.tbo_disestablish = NULL,
    251 };
    252 
    253 static void
    254 armv8_pmu_init_cpu(void *arg1, void *arg2)
    255 {
    256 	/* Disable EL0 access to performance monitors */
    257 	reg_pmuserenr_el0_write(0);
    258 
    259 	/* Disable interrupts */
    260 	reg_pmintenclr_el1_write(PMINTEN_P);
    261 
    262 	/* Disable event counters */
    263 	reg_pmcntenclr_el0_write(PMCNTEN_P);
    264 }
    265 
    266 bool
    267 armv8_pmu_detect(void)
    268 {
    269 	const uint64_t dfr0 = reg_id_aa64dfr0_el1_read();
    270 	const u_int pmuver = __SHIFTOUT(dfr0, ID_AA64DFR0_EL1_PMUVER);
    271 
    272 	return pmuver != ID_AA64DFR0_EL1_PMUVER_NONE &&
    273 	       pmuver != ID_AA64DFR0_EL1_PMUVER_IMPL;
    274 }
    275 
    276 int
    277 armv8_pmu_init(void)
    278 {
    279 	int error, ncounters;
    280 
    281 	KASSERT(armv8_pmu_detect());
    282 
    283 	ncounters = armv8_pmu_ncounters();
    284 	if (ncounters == 0)
    285 		return ENOTSUP;
    286 
    287 	/* Is 64bit event counter available? */
    288 	const uint64_t dfr0 = reg_id_aa64dfr0_el1_read();
    289 	const u_int pmuver = __SHIFTOUT(dfr0, ID_AA64DFR0_EL1_PMUVER);
    290 	if (pmuver >= ID_AA64DFR0_EL1_PMUVER_V3P5 &&
    291 	    ISSET(reg_pmcr_el0_read(), PMCR_LP))
    292 		counter_bitwidth = 64;
    293 	else
    294 		counter_bitwidth = 32;
    295 
    296 	uint64_t xc = xc_broadcast(0, armv8_pmu_init_cpu, NULL, NULL);
    297 	xc_wait(xc);
    298 
    299 	error = tprof_backend_register("tprof_armv8", &tprof_armv8_pmu_ops,
    300 	    TPROF_BACKEND_VERSION);
    301 	if (error == 0) {
    302 		/* XXX: for argument of armv8_pmu_intr() */
    303 		pmu_intr_arg = tprof_backend;
    304 	}
    305 
    306 	return error;
    307 }
    308