xref: /netbsd-src/sys/dev/tprof/tprof_armv8.c (revision 5f2f42719cd62ff11fd913b40b7ce19f07c4fd25)
1 /* $NetBSD: tprof_armv8.c,v 1.14 2022/05/16 09:42:32 jmcneill Exp $ */
2 
3 /*-
4  * Copyright (c) 2018 Jared McNeill <jmcneill@invisible.ca>
5  * All rights reserved.
6  *
7  * Redistribution and use in source and binary forms, with or without
8  * modification, are permitted provided that the following conditions
9  * are met:
10  * 1. Redistributions of source code must retain the above copyright
11  *    notice, this list of conditions and the following disclaimer.
12  * 2. Redistributions in binary form must reproduce the above copyright
13  *    notice, this list of conditions and the following disclaimer in the
14  *    documentation and/or other materials provided with the distribution.
15  *
16  * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR
17  * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
18  * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
19  * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT,
20  * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING,
21  * BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
22  * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED
23  * AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY,
24  * OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
25  * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
26  * SUCH DAMAGE.
27  */
28 
29 #include <sys/cdefs.h>
30 __KERNEL_RCSID(0, "$NetBSD: tprof_armv8.c,v 1.14 2022/05/16 09:42:32 jmcneill Exp $");
31 
32 #include <sys/param.h>
33 #include <sys/bus.h>
34 #include <sys/cpu.h>
35 #include <sys/xcall.h>
36 
37 #include <dev/tprof/tprof.h>
38 
39 #include <arm/armreg.h>
40 #include <arm/cpufunc.h>
41 
42 #include <dev/tprof/tprof_armv8.h>
43 
44 static tprof_param_t armv8_pmu_param;
45 static const u_int armv8_pmu_counter = 0;
46 static uint32_t counter_val;
47 static uint32_t counter_reset_val;
48 
49 static bool
50 armv8_pmu_event_implemented(uint16_t event)
51 {
52 	uint64_t eid[2];
53 
54 	if (event >= 64)
55 		return false;
56 
57 	eid[0] = reg_pmceid0_el0_read();
58 	eid[1] = reg_pmceid1_el0_read();
59 
60 	/* The low 32bits of PMCEID[01]_EL0 contain the common events 0 to n */
61 	const u_int idx = event / 32;
62 	const u_int bit = event % 32;
63 
64 	if (eid[idx] & __BIT(bit))
65 		return true;
66 
67 	return false;
68 }
69 
70 static void
71 armv8_pmu_set_pmevtyper(u_int counter, uint64_t val)
72 {
73 	reg_pmselr_el0_write(counter);
74 	isb();
75 	reg_pmxevtyper_el0_write(val);
76 }
77 
78 static void
79 armv8_pmu_set_pmevcntr(u_int counter, uint32_t val)
80 {
81 	reg_pmselr_el0_write(counter);
82 	isb();
83 	reg_pmxevcntr_el0_write(val);
84 }
85 
86 static void
87 armv8_pmu_start_cpu(void *arg1, void *arg2)
88 {
89 	const uint32_t counter_mask = __BIT(armv8_pmu_counter);
90 	uint64_t pmevtyper;
91 
92 	/* Disable event counter */
93 	reg_pmcntenclr_el0_write(counter_mask);
94 
95 	/* Configure event counter */
96 	pmevtyper = __SHIFTIN(armv8_pmu_param.p_event, PMEVTYPER_EVTCOUNT);
97 	if (!ISSET(armv8_pmu_param.p_flags, TPROF_PARAM_USER))
98 		pmevtyper |= PMEVTYPER_U;
99 	if (!ISSET(armv8_pmu_param.p_flags, TPROF_PARAM_KERN))
100 		pmevtyper |= PMEVTYPER_P;
101 
102 	armv8_pmu_set_pmevtyper(armv8_pmu_counter, pmevtyper);
103 
104 	/* Enable overflow interrupts */
105 	reg_pmintenset_el1_write(counter_mask);
106 
107 	/* Clear overflow flag */
108 	reg_pmovsclr_el0_write(counter_mask);
109 
110 	/* Initialize event counter value */
111 	armv8_pmu_set_pmevcntr(armv8_pmu_counter, counter_reset_val);
112 
113 	/* Enable event counter */
114 	reg_pmcntenset_el0_write(counter_mask);
115 	reg_pmcr_el0_write(PMCR_E);
116 }
117 
118 static void
119 armv8_pmu_stop_cpu(void *arg1, void *arg2)
120 {
121 	const uint32_t counter_mask = __BIT(armv8_pmu_counter);
122 
123 	/* Disable overflow interrupts */
124 	reg_pmintenclr_el1_write(counter_mask);
125 
126 	/* Disable event counter */
127 	reg_pmcntenclr_el0_write(counter_mask);
128 	reg_pmcr_el0_write(0);
129 }
130 
131 static uint64_t
132 armv8_pmu_estimate_freq(void)
133 {
134 	uint64_t cpufreq = curcpu()->ci_data.cpu_cc_freq;
135 	uint64_t freq = 10000;
136 
137 	counter_val = cpufreq / freq;
138 	if (counter_val == 0)
139 		counter_val = 4000000000ULL / freq;
140 
141 	return freq;
142 }
143 
144 static uint32_t
145 armv8_pmu_ident(void)
146 {
147 	return TPROF_IDENT_ARMV8_GENERIC;
148 }
149 
150 static int
151 armv8_pmu_start(const tprof_param_t *param)
152 {
153 	/* PMCR.N of 0 means that no event counters are available */
154 	if (__SHIFTOUT(reg_pmcr_el0_read(), PMCR_N) == 0) {
155 		return EINVAL;
156 	}
157 
158 	if (!armv8_pmu_event_implemented(param->p_event)) {
159 		printf("%s: event %#" PRIx64 " not implemented on this CPU\n",
160 		    __func__, param->p_event);
161 		return EINVAL;
162 	}
163 
164 	counter_reset_val = -counter_val + 1;
165 
166 	armv8_pmu_param = *param;
167 	uint64_t xc = xc_broadcast(0, armv8_pmu_start_cpu, NULL, NULL);
168 	xc_wait(xc);
169 
170 	return 0;
171 }
172 
173 static void
174 armv8_pmu_stop(const tprof_param_t *param)
175 {
176 	uint64_t xc;
177 
178 	xc = xc_broadcast(0, armv8_pmu_stop_cpu, NULL, NULL);
179 	xc_wait(xc);
180 }
181 
182 static const tprof_backend_ops_t tprof_armv8_pmu_ops = {
183 	.tbo_estimate_freq = armv8_pmu_estimate_freq,
184 	.tbo_ident = armv8_pmu_ident,
185 	.tbo_start = armv8_pmu_start,
186 	.tbo_stop = armv8_pmu_stop,
187 };
188 
189 int
190 armv8_pmu_intr(void *priv)
191 {
192 	const struct trapframe * const tf = priv;
193 	const uint32_t counter_mask = __BIT(armv8_pmu_counter);
194 	tprof_frame_info_t tfi;
195 
196 	const uint32_t pmovs = reg_pmovsset_el0_read();
197 	if ((pmovs & counter_mask) != 0) {
198 		tfi.tfi_pc = tf->tf_pc;
199 		tfi.tfi_inkernel = tfi.tfi_pc >= VM_MIN_KERNEL_ADDRESS &&
200 		    tfi.tfi_pc < VM_MAX_KERNEL_ADDRESS;
201 		tprof_sample(NULL, &tfi);
202 
203 		armv8_pmu_set_pmevcntr(armv8_pmu_counter, counter_reset_val);
204 	}
205 	reg_pmovsclr_el0_write(pmovs);
206 
207 	return 1;
208 }
209 
210 static void
211 armv8_pmu_init_cpu(void *arg1, void *arg2)
212 {
213 	/* Disable EL0 access to performance monitors */
214 	reg_pmuserenr_el0_write(0);
215 
216 	/* Disable interrupts */
217 	reg_pmintenclr_el1_write(~0U);
218 
219 	/* Disable event counters */
220 	reg_pmcntenclr_el0_write(PMCNTEN_P);
221 }
222 
223 bool
224 armv8_pmu_detect(void)
225 {
226 	const uint64_t dfr0 = reg_id_aa64dfr0_el1_read();
227 	const u_int pmuver = __SHIFTOUT(dfr0, ID_AA64DFR0_EL1_PMUVER);
228 
229 	return pmuver != ID_AA64DFR0_EL1_PMUVER_NONE &&
230 	       pmuver != ID_AA64DFR0_EL1_PMUVER_IMPL;
231 }
232 
233 int
234 armv8_pmu_init(void)
235 {
236 	KASSERT(armv8_pmu_detect());
237 
238 	uint64_t xc = xc_broadcast(0, armv8_pmu_init_cpu, NULL, NULL);
239 	xc_wait(xc);
240 
241 	return tprof_backend_register("tprof_armv8", &tprof_armv8_pmu_ops,
242 	    TPROF_BACKEND_VERSION);
243 }
244