xref: /netbsd-src/sys/kern/subr_prof.c (revision d20841bb642898112fe68f0ad3f7b26dddf56f07)
1 /*	$NetBSD: subr_prof.c,v 1.29 2003/12/04 19:38:23 atatat Exp $	*/
2 
3 /*-
4  * Copyright (c) 1982, 1986, 1993
5  *	The Regents of the University of California.  All rights reserved.
6  *
7  * Redistribution and use in source and binary forms, with or without
8  * modification, are permitted provided that the following conditions
9  * are met:
10  * 1. Redistributions of source code must retain the above copyright
11  *    notice, this list of conditions and the following disclaimer.
12  * 2. Redistributions in binary form must reproduce the above copyright
13  *    notice, this list of conditions and the following disclaimer in the
14  *    documentation and/or other materials provided with the distribution.
15  * 3. Neither the name of the University nor the names of its contributors
16  *    may be used to endorse or promote products derived from this software
17  *    without specific prior written permission.
18  *
19  * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
20  * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
21  * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
22  * ARE DISCLAIMED.  IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
23  * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
24  * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
25  * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
26  * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
27  * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
28  * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
29  * SUCH DAMAGE.
30  *
31  *	@(#)subr_prof.c	8.4 (Berkeley) 2/14/95
32  */
33 
34 #include <sys/cdefs.h>
35 __KERNEL_RCSID(0, "$NetBSD: subr_prof.c,v 1.29 2003/12/04 19:38:23 atatat Exp $");
36 
37 #include <sys/param.h>
38 #include <sys/systm.h>
39 #include <sys/kernel.h>
40 #include <sys/proc.h>
41 #include <sys/user.h>
42 #include <sys/mount.h>
43 #include <sys/sa.h>
44 #include <sys/syscallargs.h>
45 #include <sys/sysctl.h>
46 
47 #include <machine/cpu.h>
48 
49 #ifdef GPROF
50 #include <sys/malloc.h>
51 #include <sys/gmon.h>
52 
53 MALLOC_DEFINE(M_GPROF, "gprof", "kernel profiling buffer");
54 
55 /*
56  * Froms is actually a bunch of unsigned shorts indexing tos
57  */
58 struct gmonparam _gmonparam = { GMON_PROF_OFF };
59 
60 /* Actual start of the kernel text segment. */
61 extern char kernel_text[];
62 
63 extern char etext[];
64 
65 
66 void
67 kmstartup()
68 {
69 	char *cp;
70 	struct gmonparam *p = &_gmonparam;
71 	/*
72 	 * Round lowpc and highpc to multiples of the density we're using
73 	 * so the rest of the scaling (here and in gprof) stays in ints.
74 	 */
75 	p->lowpc = ROUNDDOWN(((u_long)kernel_text),
76 		HISTFRACTION * sizeof(HISTCOUNTER));
77 	p->highpc = ROUNDUP((u_long)etext,
78 		HISTFRACTION * sizeof(HISTCOUNTER));
79 	p->textsize = p->highpc - p->lowpc;
80 	printf("Profiling kernel, textsize=%ld [%lx..%lx]\n",
81 	       p->textsize, p->lowpc, p->highpc);
82 	p->kcountsize = p->textsize / HISTFRACTION;
83 	p->hashfraction = HASHFRACTION;
84 	p->fromssize = p->textsize / HASHFRACTION;
85 	p->tolimit = p->textsize * ARCDENSITY / 100;
86 	if (p->tolimit < MINARCS)
87 		p->tolimit = MINARCS;
88 	else if (p->tolimit > MAXARCS)
89 		p->tolimit = MAXARCS;
90 	p->tossize = p->tolimit * sizeof(struct tostruct);
91 	cp = (char *)malloc(p->kcountsize + p->fromssize + p->tossize,
92 	    M_GPROF, M_NOWAIT);
93 	if (cp == 0) {
94 		printf("No memory for profiling.\n");
95 		return;
96 	}
97 	memset(cp, 0, p->kcountsize + p->tossize + p->fromssize);
98 	p->tos = (struct tostruct *)cp;
99 	cp += p->tossize;
100 	p->kcount = (u_short *)cp;
101 	cp += p->kcountsize;
102 	p->froms = (u_short *)cp;
103 }
104 
105 /*
106  * Return kernel profiling information.
107  */
108 /*
109  * sysctl helper routine for kern.profiling subtree.  enables/disables
110  * kernel profiling and gives out copies of the profiling data.
111  */
112 static int
113 sysctl_kern_profiling(SYSCTLFN_ARGS)
114 {
115 	struct gmonparam *gp = &_gmonparam;
116 	int error;
117 	struct sysctlnode node;
118 
119 	node = *rnode;
120 
121 	switch (node.sysctl_num) {
122 	case GPROF_STATE:
123 		node.sysctl_data = &gp->state;
124 		break;
125 	case GPROF_COUNT:
126 		node.sysctl_data = gp->kcount;
127 		node.sysctl_size = gp->kcountsize;
128 		break;
129 	case GPROF_FROMS:
130 		node.sysctl_data = gp->froms;
131 		node.sysctl_size = gp->fromssize;
132 		break;
133 	case GPROF_TOS:
134 		node.sysctl_data = gp->tos;
135 		node.sysctl_size = gp->tossize;
136 		break;
137 	case GPROF_GMONPARAM:
138 		node.sysctl_data = gp;
139 		node.sysctl_size = sizeof(*gp);
140 		break;
141 	default:
142 		return (EOPNOTSUPP);
143 	}
144 
145 	error = sysctl_lookup(SYSCTLFN_CALL(&node));
146 	if (error || newp == NULL)
147 		return (error);
148 
149 	if (node.sysctl_num == GPROF_STATE) {
150 		if (gp->state == GMON_PROF_OFF)
151 			stopprofclock(&proc0);
152 		else
153 			startprofclock(&proc0);
154 	}
155 
156 	return (0);
157 }
158 
159 SYSCTL_SETUP(sysctl_kern_gprof_setup, "sysctl kern.profiling subtree setup")
160 {
161 
162 	sysctl_createv(SYSCTL_PERMANENT,
163 		       CTLTYPE_NODE, "kern", NULL,
164 		       NULL, 0, NULL, 0,
165 		       CTL_KERN, CTL_EOL);
166 	sysctl_createv(SYSCTL_PERMANENT,
167 		       CTLTYPE_NODE, "profiling", NULL,
168 		       NULL, 0, NULL, 0,
169 		       CTL_KERN, KERN_PROF, CTL_EOL);
170 
171 	sysctl_createv(SYSCTL_PERMANENT|SYSCTL_READWRITE,
172 		       CTLTYPE_INT, "state", NULL,
173 		       sysctl_kern_profiling, 0, NULL, 0,
174 		       CTL_KERN, KERN_PROF, GPROF_STATE, CTL_EOL);
175 	sysctl_createv(SYSCTL_PERMANENT|SYSCTL_READWRITE,
176 		       CTLTYPE_STRUCT, "count", NULL,
177 		       sysctl_kern_profiling, 0, NULL, 0,
178 		       CTL_KERN, KERN_PROF, GPROF_COUNT, CTL_EOL);
179 	sysctl_createv(SYSCTL_PERMANENT|SYSCTL_READWRITE,
180 		       CTLTYPE_STRUCT, "froms", NULL,
181 		       sysctl_kern_profiling, 0, NULL, 0,
182 		       CTL_KERN, KERN_PROF, GPROF_FROMS, CTL_EOL);
183 	sysctl_createv(SYSCTL_PERMANENT|SYSCTL_READWRITE,
184 		       CTLTYPE_STRUCT, "tos", NULL,
185 		       sysctl_kern_profiling, 0, NULL, 0,
186 		       CTL_KERN, KERN_PROF, GPROF_TOS, CTL_EOL);
187 	sysctl_createv(SYSCTL_PERMANENT,
188 		       CTLTYPE_STRUCT, "gmonparam", NULL,
189 		       sysctl_kern_profiling, 0, NULL, 0,
190 		       CTL_KERN, KERN_PROF, GPROF_GMONPARAM, CTL_EOL);
191 }
192 #endif /* GPROF */
193 
194 /*
195  * Profiling system call.
196  *
197  * The scale factor is a fixed point number with 16 bits of fraction, so that
198  * 1.0 is represented as 0x10000.  A scale factor of 0 turns off profiling.
199  */
200 /* ARGSUSED */
201 int
202 sys_profil(l, v, retval)
203 	struct lwp *l;
204 	void *v;
205 	register_t *retval;
206 {
207 	struct sys_profil_args /* {
208 		syscallarg(caddr_t) samples;
209 		syscallarg(u_int) size;
210 		syscallarg(u_int) offset;
211 		syscallarg(u_int) scale;
212 	} */ *uap = v;
213 	struct proc *p = l->l_proc;
214 	struct uprof *upp;
215 	int s;
216 
217 	if (SCARG(uap, scale) > (1 << 16))
218 		return (EINVAL);
219 	if (SCARG(uap, scale) == 0) {
220 		stopprofclock(p);
221 		return (0);
222 	}
223 	upp = &p->p_stats->p_prof;
224 
225 	/* Block profile interrupts while changing state. */
226 	s = splstatclock();
227 	upp->pr_off = SCARG(uap, offset);
228 	upp->pr_scale = SCARG(uap, scale);
229 	upp->pr_base = SCARG(uap, samples);
230 	upp->pr_size = SCARG(uap, size);
231 	startprofclock(p);
232 	splx(s);
233 
234 	return (0);
235 }
236 
237 /*
238  * Scale is a fixed-point number with the binary point 16 bits
239  * into the value, and is <= 1.0.  pc is at most 32 bits, so the
240  * intermediate result is at most 48 bits.
241  */
242 #define	PC_TO_INDEX(pc, prof) \
243 	((int)(((u_quad_t)((pc) - (prof)->pr_off) * \
244 	    (u_quad_t)((prof)->pr_scale)) >> 16) & ~1)
245 
246 /*
247  * Collect user-level profiling statistics; called on a profiling tick,
248  * when a process is running in user-mode.  This routine may be called
249  * from an interrupt context.  We try to update the user profiling buffers
250  * cheaply with fuswintr() and suswintr().  If that fails, we revert to
251  * an AST that will vector us to trap() with a context in which copyin
252  * and copyout will work.  Trap will then call addupc_task().
253  *
254  * Note that we may (rarely) not get around to the AST soon enough, and
255  * lose profile ticks when the next tick overwrites this one, but in this
256  * case the system is overloaded and the profile is probably already
257  * inaccurate.
258  */
259 void
260 addupc_intr(p, pc)
261 	struct proc *p;
262 	u_long pc;
263 {
264 	struct uprof *prof;
265 	caddr_t addr;
266 	u_int i;
267 	int v;
268 
269 	prof = &p->p_stats->p_prof;
270 	if (pc < prof->pr_off ||
271 	    (i = PC_TO_INDEX(pc, prof)) >= prof->pr_size)
272 		return;			/* out of range; ignore */
273 
274 	addr = prof->pr_base + i;
275 	if ((v = fuswintr(addr)) == -1 || suswintr(addr, v + 1) == -1) {
276 		prof->pr_addr = pc;
277 		prof->pr_ticks++;
278 		need_proftick(p);
279 	}
280 }
281 
282 /*
283  * Much like before, but we can afford to take faults here.  If the
284  * update fails, we simply turn off profiling.
285  */
286 void
287 addupc_task(p, pc, ticks)
288 	struct proc *p;
289 	u_long pc;
290 	u_int ticks;
291 {
292 	struct uprof *prof;
293 	caddr_t addr;
294 	u_int i;
295 	u_short v;
296 
297 	/* Testing P_PROFIL may be unnecessary, but is certainly safe. */
298 	if ((p->p_flag & P_PROFIL) == 0 || ticks == 0)
299 		return;
300 
301 	prof = &p->p_stats->p_prof;
302 	if (pc < prof->pr_off ||
303 	    (i = PC_TO_INDEX(pc, prof)) >= prof->pr_size)
304 		return;
305 
306 	addr = prof->pr_base + i;
307 	if (copyin(addr, (caddr_t)&v, sizeof(v)) == 0) {
308 		v += ticks;
309 		if (copyout((caddr_t)&v, addr, sizeof(v)) == 0)
310 			return;
311 	}
312 	stopprofclock(p);
313 }
314