xref: /dflybsd-src/sys/vm/vm_meter.c (revision 0d5acd7467c4e95f792ef49fceb3ab8e917ce86b)
1 /*
2  * (MPSAFE)
3  *
4  * Copyright (c) 1982, 1986, 1989, 1993
5  *	The Regents of the University of California.  All rights reserved.
6  *
7  * Redistribution and use in source and binary forms, with or without
8  * modification, are permitted provided that the following conditions
9  * are met:
10  * 1. Redistributions of source code must retain the above copyright
11  *    notice, this list of conditions and the following disclaimer.
12  * 2. Redistributions in binary form must reproduce the above copyright
13  *    notice, this list of conditions and the following disclaimer in the
14  *    documentation and/or other materials provided with the distribution.
15  * 3. Neither the name of the University nor the names of its contributors
16  *    may be used to endorse or promote products derived from this software
17  *    without specific prior written permission.
18  *
19  * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
20  * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
21  * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
22  * ARE DISCLAIMED.  IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
23  * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
24  * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
25  * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
26  * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
27  * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
28  * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
29  * SUCH DAMAGE.
30  *
31  *	@(#)vm_meter.c	8.4 (Berkeley) 1/4/94
32  * $FreeBSD: src/sys/vm/vm_meter.c,v 1.34.2.7 2002/10/10 19:28:22 dillon Exp $
33  * $DragonFly: src/sys/vm/vm_meter.c,v 1.15 2008/04/28 18:04:08 dillon Exp $
34  */
35 
36 #include <sys/param.h>
37 #include <sys/proc.h>
38 #include <sys/systm.h>
39 #include <sys/kernel.h>
40 #include <sys/resource.h>
41 #include <sys/vmmeter.h>
42 
43 #include <vm/vm.h>
44 #include <vm/vm_page.h>
45 #include <vm/vm_extern.h>
46 #include <vm/vm_param.h>
47 #include <sys/lock.h>
48 #include <vm/pmap.h>
49 #include <vm/vm_map.h>
50 #include <vm/vm_object.h>
51 #include <sys/sysctl.h>
52 
53 struct vmstats vmstats;
54 
55 static int maxslp = MAXSLP;
56 
57 SYSCTL_UINT(_vm, VM_V_FREE_MIN, v_free_min,
58 	CTLFLAG_RW, &vmstats.v_free_min, 0,
59 	"Minimum number of pages desired free");
60 SYSCTL_UINT(_vm, VM_V_FREE_TARGET, v_free_target,
61 	CTLFLAG_RW, &vmstats.v_free_target, 0,
62 	"Number of pages desired free");
63 SYSCTL_UINT(_vm, VM_V_FREE_RESERVED, v_free_reserved,
64 	CTLFLAG_RW, &vmstats.v_free_reserved, 0,
65 	"Number of pages reserved for deadlock");
66 SYSCTL_UINT(_vm, VM_V_INACTIVE_TARGET, v_inactive_target,
67 	CTLFLAG_RW, &vmstats.v_inactive_target, 0,
68 	"Number of pages desired inactive");
69 SYSCTL_UINT(_vm, VM_V_CACHE_MIN, v_cache_min,
70 	CTLFLAG_RW, &vmstats.v_cache_min, 0,
71 	"Min number of pages desired on cache queue");
72 SYSCTL_UINT(_vm, VM_V_CACHE_MAX, v_cache_max,
73 	CTLFLAG_RW, &vmstats.v_cache_max, 0,
74 	"Max number of pages in cached obj");
75 SYSCTL_UINT(_vm, VM_V_PAGEOUT_FREE_MIN, v_pageout_free_min,
76 	CTLFLAG_RW, &vmstats.v_pageout_free_min, 0,
77 	"Min number pages reserved for kernel");
78 SYSCTL_UINT(_vm, OID_AUTO, v_free_severe,
79 	CTLFLAG_RW, &vmstats.v_free_severe, 0, "");
80 
81 SYSCTL_STRUCT(_vm, VM_LOADAVG, loadavg, CTLFLAG_RD,
82     &averunnable, loadavg, "Machine loadaverage history");
83 
84 static int do_vmtotal_callback(struct proc *p, void *data);
85 
86 /*
87  * No requirements.
88  */
89 static int
90 do_vmtotal(SYSCTL_HANDLER_ARGS)
91 {
92 	struct vmtotal total;
93 	struct vmtotal *totalp;
94 	struct vm_object marker;
95 	vm_object_t object;
96 	long collisions;
97 	int burst;
98 
99 	bzero(&total, sizeof(total));
100 	totalp = &total;
101 	bzero(&marker, sizeof(marker));
102 	marker.type = OBJT_MARKER;
103 	collisions = vmobj_token.t_collisions;
104 
105 #if 0
106 	/*
107 	 * Mark all objects as inactive.
108 	 */
109 	lwkt_gettoken(&vmobj_token);
110 	for (object = TAILQ_FIRST(&vm_object_list);
111 	    object != NULL;
112 	    object = TAILQ_NEXT(object,object_list)) {
113 		if (object->type == OBJT_MARKER)
114 			continue;
115 		vm_object_clear_flag(object, OBJ_ACTIVE);
116 	}
117 	lwkt_reltoken(&vmobj_token);
118 #endif
119 
120 	/*
121 	 * Calculate process statistics.
122 	 */
123 	allproc_scan(do_vmtotal_callback, totalp);
124 
125 	/*
126 	 * Calculate object memory usage statistics.
127 	 */
128 	lwkt_gettoken(&vmobj_token);
129 	TAILQ_INSERT_HEAD(&vm_object_list, &marker, object_list);
130 	burst = 0;
131 
132 	for (object = TAILQ_FIRST(&vm_object_list);
133 	    object != NULL;
134 	    object = TAILQ_NEXT(object, object_list)) {
135 		/*
136 		 * devices, like /dev/mem, will badly skew our totals.
137 		 * markers aren't real objects.
138 		 */
139 		if (object->type == OBJT_MARKER)
140 			continue;
141 		if (object->type == OBJT_DEVICE)
142 			continue;
143 		if (object->type == OBJT_MGTDEVICE)
144 			continue;
145 		if (object->size >= 0x7FFFFFFF) {
146 			/*
147 			 * Probably unbounded anonymous memory (really
148 			 * bounded by related vm_map_entry structures which
149 			 * we do not have access to in this loop).
150 			 */
151 			totalp->t_vm += object->resident_page_count;
152 		} else {
153 			/*
154 			 * It's questionable how useful this is but...
155 			 */
156 			totalp->t_vm += object->size;
157 		}
158 		totalp->t_rm += object->resident_page_count;
159 		if (object->flags & OBJ_ACTIVE) {
160 			totalp->t_avm += object->size;
161 			totalp->t_arm += object->resident_page_count;
162 		}
163 		if (object->shadow_count > 1) {
164 			/* shared object */
165 			totalp->t_vmshr += object->size;
166 			totalp->t_rmshr += object->resident_page_count;
167 			if (object->flags & OBJ_ACTIVE) {
168 				totalp->t_avmshr += object->size;
169 				totalp->t_armshr += object->resident_page_count;
170 			}
171 		}
172 
173 		/*
174 		 * Don't waste time unnecessarily
175 		 */
176 		if (++burst < 25)
177 			continue;
178 		burst = 0;
179 
180 		/*
181 		 * Don't hog the vmobj_token if someone else wants it.
182 		 */
183 		TAILQ_REMOVE(&vm_object_list, &marker, object_list);
184 		TAILQ_INSERT_AFTER(&vm_object_list, object,
185 				   &marker, object_list);
186 		object = &marker;
187 		if (collisions != vmobj_token.t_collisions) {
188 			tsleep(&vm_object_list, 0, "breath", 1);
189 			collisions = vmobj_token.t_collisions;
190 		} else {
191 			lwkt_yield();
192 		}
193 	}
194 
195 	TAILQ_REMOVE(&vm_object_list, &marker, object_list);
196 	lwkt_reltoken(&vmobj_token);
197 
198 	totalp->t_free = vmstats.v_free_count + vmstats.v_cache_count;
199 
200 	return (sysctl_handle_opaque(oidp, totalp, sizeof total, req));
201 }
202 
203 /*
204  * The caller must hold proc_token.
205  */
206 static int
207 do_vmtotal_callback(struct proc *p, void *data)
208 {
209 	struct vmtotal *totalp = data;
210 	struct lwp *lp;
211 
212 	if (p->p_flags & P_SYSTEM)
213 		return(0);
214 
215 	FOREACH_LWP_IN_PROC(lp, p) {
216 		switch (lp->lwp_stat) {
217 		case LSSTOP:
218 		case LSSLEEP:
219 			if ((p->p_flags & P_SWAPPEDOUT) == 0) {
220 				if ((lp->lwp_flags & LWP_SINTR) == 0)
221 					totalp->t_dw++;
222 				else if (lp->lwp_slptime < maxslp)
223 					totalp->t_sl++;
224 			} else if (lp->lwp_slptime < maxslp) {
225 				totalp->t_sw++;
226 			}
227 			if (lp->lwp_slptime >= maxslp)
228 				return(0);
229 			break;
230 
231 		case LSRUN:
232 			if (p->p_flags & P_SWAPPEDOUT)
233 				totalp->t_sw++;
234 			else
235 				totalp->t_rq++;
236 			if (p->p_stat == SIDL)
237 				return(0);
238 			break;
239 
240 		default:
241 			return (0);
242 		}
243 
244 		/*
245 		 * Set while in vm_fault()
246 		 */
247 		if (lp->lwp_flags & LWP_PAGING)
248 			totalp->t_pw++;
249 	}
250 	return(0);
251 }
252 
253 /*
254  * No requirements.
255  */
256 static int
257 do_vmstats(SYSCTL_HANDLER_ARGS)
258 {
259 	struct vmstats vms = vmstats;
260 	return (sysctl_handle_opaque(oidp, &vms, sizeof(vms), req));
261 }
262 
263 /*
264  * No requirements.
265  */
266 static int
267 do_vmmeter(SYSCTL_HANDLER_ARGS)
268 {
269 	int boffset = offsetof(struct vmmeter, vmmeter_uint_begin);
270 	int eoffset = offsetof(struct vmmeter, vmmeter_uint_end);
271 	struct vmmeter vmm;
272 	int i;
273 
274 	bzero(&vmm, sizeof(vmm));
275 	for (i = 0; i < ncpus; ++i) {
276 		int off;
277 		struct globaldata *gd = globaldata_find(i);
278 
279 		for (off = boffset; off <= eoffset; off += sizeof(u_int)) {
280 			*(u_int *)((char *)&vmm + off) +=
281 				*(u_int *)((char *)&gd->gd_cnt + off);
282 		}
283 
284 	}
285 	vmm.v_intr += vmm.v_ipi + vmm.v_timer;
286 	return (sysctl_handle_opaque(oidp, &vmm, sizeof(vmm), req));
287 }
288 
289 /*
290  * vcnt() -	accumulate statistics from the cnt structure for each cpu
291  *
292  *	The vmmeter structure is now per-cpu as well as global.  Those
293  *	statistics which can be kept on a per-cpu basis (to avoid cache
294  *	stalls between cpus) can be moved to the per-cpu vmmeter.  Remaining
295  *	statistics, such as v_free_reserved, are left in the global
296  *	structure.
297  *
298  * (sysctl_oid *oidp, void *arg1, int arg2, struct sysctl_req *req)
299  *
300  * No requirements.
301  */
302 static int
303 vcnt(SYSCTL_HANDLER_ARGS)
304 {
305 	int i;
306 	int count = 0;
307 	int offset = arg2;
308 
309 	for (i = 0; i < ncpus; ++i) {
310 		struct globaldata *gd = globaldata_find(i);
311 		count += *(int *)((char *)&gd->gd_cnt + offset);
312 	}
313 	return(SYSCTL_OUT(req, &count, sizeof(int)));
314 }
315 
316 /*
317  * No requirements.
318  */
319 static int
320 vcnt_intr(SYSCTL_HANDLER_ARGS)
321 {
322 	int i;
323 	int count = 0;
324 
325 	for (i = 0; i < ncpus; ++i) {
326 		struct globaldata *gd = globaldata_find(i);
327 
328 		count += gd->gd_cnt.v_intr + gd->gd_cnt.v_ipi +
329 			 gd->gd_cnt.v_timer;
330 	}
331 	return(SYSCTL_OUT(req, &count, sizeof(int)));
332 }
333 
334 #define VMMETEROFF(var)	offsetof(struct vmmeter, var)
335 
336 SYSCTL_PROC(_vm, OID_AUTO, vmtotal, CTLTYPE_OPAQUE|CTLFLAG_RD,
337     0, sizeof(struct vmtotal), do_vmtotal, "S,vmtotal",
338     "System virtual memory aggregate");
339 SYSCTL_PROC(_vm, OID_AUTO, vmstats, CTLTYPE_OPAQUE|CTLFLAG_RD,
340     0, sizeof(struct vmstats), do_vmstats, "S,vmstats",
341     "System virtual memory statistics");
342 SYSCTL_PROC(_vm, OID_AUTO, vmmeter, CTLTYPE_OPAQUE|CTLFLAG_RD,
343     0, sizeof(struct vmmeter), do_vmmeter, "S,vmmeter",
344     "System statistics");
345 SYSCTL_NODE(_vm, OID_AUTO, stats, CTLFLAG_RW, 0, "VM meter stats");
346 SYSCTL_NODE(_vm_stats, OID_AUTO, sys, CTLFLAG_RW, 0, "VM meter sys stats");
347 SYSCTL_NODE(_vm_stats, OID_AUTO, vm, CTLFLAG_RW, 0, "VM meter vm stats");
348 SYSCTL_NODE(_vm_stats, OID_AUTO, misc, CTLFLAG_RW, 0, "VM meter misc stats");
349 
350 SYSCTL_PROC(_vm_stats_sys, OID_AUTO, v_swtch, CTLTYPE_UINT|CTLFLAG_RD,
351 	0, VMMETEROFF(v_swtch), vcnt, "IU", "Context switches");
352 SYSCTL_PROC(_vm_stats_sys, OID_AUTO, v_intrans_coll, CTLTYPE_UINT|CTLFLAG_RD,
353 	0, VMMETEROFF(v_intrans_coll), vcnt, "IU", "Intransit map collisions (total)");
354 SYSCTL_PROC(_vm_stats_sys, OID_AUTO, v_intrans_wait, CTLTYPE_UINT|CTLFLAG_RD,
355 	0, VMMETEROFF(v_intrans_wait), vcnt, "IU", "Intransit map collisions which blocked");
356 SYSCTL_PROC(_vm_stats_sys, OID_AUTO, v_forwarded_ints, CTLTYPE_UINT|CTLFLAG_RD,
357 	0, VMMETEROFF(v_forwarded_ints), vcnt, "IU", "Forwarded interrupts due to MP lock");
358 SYSCTL_PROC(_vm_stats_sys, OID_AUTO, v_forwarded_hits, CTLTYPE_UINT|CTLFLAG_RD,
359 	0, VMMETEROFF(v_forwarded_hits), vcnt, "IU", "Forwarded hits due to MP lock");
360 SYSCTL_PROC(_vm_stats_sys, OID_AUTO, v_forwarded_misses, CTLTYPE_UINT|CTLFLAG_RD,
361 	0, VMMETEROFF(v_forwarded_misses), vcnt, "IU", "Forwarded misses due to MP lock");
362 SYSCTL_PROC(_vm_stats_sys, OID_AUTO, v_trap, CTLTYPE_UINT|CTLFLAG_RD,
363 	0, VMMETEROFF(v_trap), vcnt, "IU", "Traps");
364 SYSCTL_PROC(_vm_stats_sys, OID_AUTO, v_syscall, CTLTYPE_UINT|CTLFLAG_RD,
365 	0, VMMETEROFF(v_syscall), vcnt, "IU", "Syscalls");
366 SYSCTL_PROC(_vm_stats_sys, OID_AUTO, v_intr, CTLTYPE_UINT|CTLFLAG_RD,
367 	0, VMMETEROFF(v_intr), vcnt_intr, "IU", "Hardware interrupts");
368 SYSCTL_PROC(_vm_stats_sys, OID_AUTO, v_ipi, CTLTYPE_UINT|CTLFLAG_RD,
369 	0, VMMETEROFF(v_ipi), vcnt, "IU", "Inter-processor interrupts");
370 SYSCTL_PROC(_vm_stats_sys, OID_AUTO, v_timer, CTLTYPE_UINT|CTLFLAG_RD,
371 	0, VMMETEROFF(v_timer), vcnt, "IU", "LAPIC timer interrupts");
372 SYSCTL_PROC(_vm_stats_sys, OID_AUTO, v_soft, CTLTYPE_UINT|CTLFLAG_RD,
373 	0, VMMETEROFF(v_soft), vcnt, "IU", "Software interrupts");
374 SYSCTL_PROC(_vm_stats_vm, OID_AUTO, v_vm_faults, CTLTYPE_UINT|CTLFLAG_RD,
375 	0, VMMETEROFF(v_vm_faults), vcnt, "IU", "VM faults");
376 SYSCTL_PROC(_vm_stats_vm, OID_AUTO, v_cow_faults, CTLTYPE_UINT|CTLFLAG_RD,
377 	0, VMMETEROFF(v_cow_faults), vcnt, "IU", "COW faults");
378 SYSCTL_PROC(_vm_stats_vm, OID_AUTO, v_cow_optim, CTLTYPE_UINT|CTLFLAG_RD,
379 	0, VMMETEROFF(v_cow_optim), vcnt, "IU", "Optimized COW faults");
380 SYSCTL_PROC(_vm_stats_vm, OID_AUTO, v_zfod, CTLTYPE_UINT|CTLFLAG_RD,
381 	0, VMMETEROFF(v_zfod), vcnt, "IU", "Zero fill");
382 SYSCTL_PROC(_vm_stats_vm, OID_AUTO, v_ozfod, CTLTYPE_UINT|CTLFLAG_RD,
383 	0, VMMETEROFF(v_ozfod), vcnt, "IU", "Optimized zero fill");
384 SYSCTL_PROC(_vm_stats_vm, OID_AUTO, v_swapin, CTLTYPE_UINT|CTLFLAG_RD,
385 	0, VMMETEROFF(v_swapin), vcnt, "IU", "Swapin operations");
386 SYSCTL_PROC(_vm_stats_vm, OID_AUTO, v_swapout, CTLTYPE_UINT|CTLFLAG_RD,
387 	0, VMMETEROFF(v_swapout), vcnt, "IU", "Swapout operations");
388 SYSCTL_PROC(_vm_stats_vm, OID_AUTO, v_swappgsin, CTLTYPE_UINT|CTLFLAG_RD,
389 	0, VMMETEROFF(v_swappgsin), vcnt, "IU", "Swapin pages");
390 SYSCTL_PROC(_vm_stats_vm, OID_AUTO, v_swappgsout, CTLTYPE_UINT|CTLFLAG_RD,
391 	0, VMMETEROFF(v_swappgsout), vcnt, "IU", "Swapout pages");
392 SYSCTL_PROC(_vm_stats_vm, OID_AUTO, v_vnodein, CTLTYPE_UINT|CTLFLAG_RD,
393 	0, VMMETEROFF(v_vnodein), vcnt, "IU", "Vnodein operations");
394 SYSCTL_PROC(_vm_stats_vm, OID_AUTO, v_vnodeout, CTLTYPE_UINT|CTLFLAG_RD,
395 	0, VMMETEROFF(v_vnodeout), vcnt, "IU", "Vnodeout operations");
396 SYSCTL_PROC(_vm_stats_vm, OID_AUTO, v_vnodepgsin, CTLTYPE_UINT|CTLFLAG_RD,
397 	0, VMMETEROFF(v_vnodepgsin), vcnt, "IU", "Vnodein pages");
398 SYSCTL_PROC(_vm_stats_vm, OID_AUTO, v_vnodepgsout, CTLTYPE_UINT|CTLFLAG_RD,
399 	0, VMMETEROFF(v_vnodepgsout), vcnt, "IU", "Vnodeout pages");
400 SYSCTL_PROC(_vm_stats_vm, OID_AUTO, v_intrans, CTLTYPE_UINT|CTLFLAG_RD,
401 	0, VMMETEROFF(v_intrans), vcnt, "IU", "In transit page blocking");
402 SYSCTL_PROC(_vm_stats_vm, OID_AUTO, v_reactivated, CTLTYPE_UINT|CTLFLAG_RD,
403 	0, VMMETEROFF(v_reactivated), vcnt, "IU", "Reactivated pages");
404 SYSCTL_PROC(_vm_stats_vm, OID_AUTO, v_pdwakeups, CTLTYPE_UINT|CTLFLAG_RD,
405 	0, VMMETEROFF(v_pdwakeups), vcnt, "IU", "Pagedaemon wakeups");
406 SYSCTL_PROC(_vm_stats_vm, OID_AUTO, v_ppwakeups, CTLTYPE_UINT|CTLFLAG_RD,
407 	0, VMMETEROFF(v_ppwakeups), vcnt, "IU", "vm_wait wakeups");
408 SYSCTL_PROC(_vm_stats_vm, OID_AUTO, v_pdpages, CTLTYPE_UINT|CTLFLAG_RD,
409 	0, VMMETEROFF(v_pdpages), vcnt, "IU", "Pagedaemon page scans");
410 SYSCTL_PROC(_vm_stats_vm, OID_AUTO, v_dfree, CTLTYPE_UINT|CTLFLAG_RD,
411 	0, VMMETEROFF(v_dfree), vcnt, "IU", "Pages freed by daemon");
412 SYSCTL_PROC(_vm_stats_vm, OID_AUTO, v_pfree, CTLTYPE_UINT|CTLFLAG_RD,
413 	0, VMMETEROFF(v_pfree), vcnt, "IU", "Pages freed by exiting processes");
414 SYSCTL_PROC(_vm_stats_vm, OID_AUTO, v_tfree, CTLTYPE_UINT|CTLFLAG_RD,
415 	0, VMMETEROFF(v_tfree), vcnt, "IU", "Total pages freed");
416 SYSCTL_PROC(_vm_stats_vm, OID_AUTO, v_forks, CTLTYPE_UINT|CTLFLAG_RD,
417 	0, VMMETEROFF(v_forks), vcnt, "IU", "Number of fork() calls");
418 SYSCTL_PROC(_vm_stats_vm, OID_AUTO, v_vforks, CTLTYPE_UINT|CTLFLAG_RD,
419 	0, VMMETEROFF(v_vforks), vcnt, "IU", "Number of vfork() calls");
420 SYSCTL_PROC(_vm_stats_vm, OID_AUTO, v_rforks, CTLTYPE_UINT|CTLFLAG_RD,
421 	0, VMMETEROFF(v_rforks), vcnt, "IU", "Number of rfork() calls");
422 SYSCTL_PROC(_vm_stats_vm, OID_AUTO, v_kthreads, CTLTYPE_UINT|CTLFLAG_RD,
423 	0, VMMETEROFF(v_kthreads), vcnt, "IU", "Number of fork() calls by kernel");
424 SYSCTL_PROC(_vm_stats_vm, OID_AUTO, v_forkpages, CTLTYPE_UINT|CTLFLAG_RD,
425 	0, VMMETEROFF(v_forkpages), vcnt, "IU", "VM pages affected by fork()");
426 SYSCTL_PROC(_vm_stats_vm, OID_AUTO, v_vforkpages, CTLTYPE_UINT|CTLFLAG_RD,
427 	0, VMMETEROFF(v_vforkpages), vcnt, "IU", "VM pages affected by vfork()");
428 SYSCTL_PROC(_vm_stats_vm, OID_AUTO, v_rforkpages, CTLTYPE_UINT|CTLFLAG_RD,
429 	0, VMMETEROFF(v_rforkpages), vcnt, "IU", "VM pages affected by rfork()");
430 SYSCTL_PROC(_vm_stats_vm, OID_AUTO, v_kthreadpages, CTLTYPE_UINT|CTLFLAG_RD,
431 	0, VMMETEROFF(v_kthreadpages), vcnt, "IU", "VM pages affected by fork() by kernel");
432 
433 SYSCTL_UINT(_vm_stats_vm, OID_AUTO,
434 	v_page_size, CTLFLAG_RD, &vmstats.v_page_size, 0,
435 	"Page size in bytes");
436 SYSCTL_UINT(_vm_stats_vm, OID_AUTO,
437 	v_page_count, CTLFLAG_RD, &vmstats.v_page_count, 0,
438 	"Total number of pages in system");
439 SYSCTL_UINT(_vm_stats_vm, OID_AUTO,
440 	v_free_reserved, CTLFLAG_RD, &vmstats.v_free_reserved, 0,
441 	"Number of pages reserved for deadlock");
442 SYSCTL_UINT(_vm_stats_vm, OID_AUTO,
443 	v_free_target, CTLFLAG_RD, &vmstats.v_free_target, 0,
444 	"Number of pages desired free");
445 SYSCTL_UINT(_vm_stats_vm, OID_AUTO,
446 	v_free_min, CTLFLAG_RD, &vmstats.v_free_min, 0,
447 	"Minimum number of pages desired free");
448 SYSCTL_UINT(_vm_stats_vm, OID_AUTO,
449 	v_free_count, CTLFLAG_RD, &vmstats.v_free_count, 0,
450 	"Number of pages free");
451 SYSCTL_UINT(_vm_stats_vm, OID_AUTO,
452 	v_wire_count, CTLFLAG_RD, &vmstats.v_wire_count, 0,
453 	"Number of pages wired down");
454 SYSCTL_UINT(_vm_stats_vm, OID_AUTO,
455 	v_active_count, CTLFLAG_RD, &vmstats.v_active_count, 0,
456 	"Number of pages active");
457 SYSCTL_UINT(_vm_stats_vm, OID_AUTO,
458 	v_inactive_target, CTLFLAG_RD, &vmstats.v_inactive_target, 0,
459 	"Number of pages desired inactive");
460 SYSCTL_UINT(_vm_stats_vm, OID_AUTO,
461 	v_inactive_count, CTLFLAG_RD, &vmstats.v_inactive_count, 0,
462 	"Number of pages inactive");
463 SYSCTL_UINT(_vm_stats_vm, OID_AUTO,
464 	v_cache_count, CTLFLAG_RD, &vmstats.v_cache_count, 0,
465 	"Number of pages on buffer cache queue");
466 SYSCTL_UINT(_vm_stats_vm, OID_AUTO,
467 	v_cache_min, CTLFLAG_RD, &vmstats.v_cache_min, 0,
468 	"Min number of pages desired on cache queue");
469 SYSCTL_UINT(_vm_stats_vm, OID_AUTO,
470 	v_cache_max, CTLFLAG_RD, &vmstats.v_cache_max, 0,
471 	"Max number of pages in cached obj");
472 SYSCTL_UINT(_vm_stats_vm, OID_AUTO,
473 	v_pageout_free_min, CTLFLAG_RD, &vmstats.v_pageout_free_min, 0,
474 	"Min number pages reserved for kernel");
475 SYSCTL_UINT(_vm_stats_vm, OID_AUTO,
476 	v_interrupt_free_min, CTLFLAG_RD, &vmstats.v_interrupt_free_min, 0,
477 	"Reserved number of pages for int code");
478 SYSCTL_INT(_vm_stats_misc, OID_AUTO,
479 	zero_page_count, CTLFLAG_RD, &vm_page_zero_count, 0,
480 	"Number of zeroing pages");
481 
482 /*
483  * No requirements.
484  */
485 static int
486 do_vmmeter_pcpu(SYSCTL_HANDLER_ARGS)
487 {
488 	int boffset = offsetof(struct vmmeter, vmmeter_uint_begin);
489 	int eoffset = offsetof(struct vmmeter, vmmeter_uint_end);
490 	struct globaldata *gd = arg1;
491 	struct vmmeter vmm;
492 	int off;
493 
494 	bzero(&vmm, sizeof(vmm));
495 	for (off = boffset; off <= eoffset; off += sizeof(u_int)) {
496 		*(u_int *)((char *)&vmm + off) +=
497 			*(u_int *)((char *)&gd->gd_cnt + off);
498 	}
499 	vmm.v_intr += vmm.v_ipi + vmm.v_timer;
500 	return (sysctl_handle_opaque(oidp, &vmm, sizeof(vmm), req));
501 }
502 
503 /*
504  * Called from the low level boot code only.
505  */
506 static void
507 vmmeter_init(void *dummy __unused)
508 {
509 	int i;
510 
511 	for (i = 0; i < ncpus; ++i) {
512 		struct sysctl_ctx_list *ctx;
513 		struct sysctl_oid *oid;
514 		struct globaldata *gd;
515 		char name[32];
516 
517 		ksnprintf(name, sizeof(name), "cpu%d", i);
518 
519 		ctx = kmalloc(sizeof(*ctx), M_TEMP, M_WAITOK);
520 		sysctl_ctx_init(ctx);
521 		oid = SYSCTL_ADD_NODE(ctx, SYSCTL_STATIC_CHILDREN(_vm),
522 				      OID_AUTO, name, CTLFLAG_RD, 0, "");
523 
524 		gd = globaldata_find(i);
525 		SYSCTL_ADD_PROC(ctx, SYSCTL_CHILDREN(oid), OID_AUTO,
526 				"vmmeter", CTLTYPE_OPAQUE|CTLFLAG_RD,
527 				gd, sizeof(struct vmmeter), do_vmmeter_pcpu,
528 				"S,vmmeter", "System per-cpu statistics");
529 	}
530 }
531 SYSINIT(vmmeter, SI_SUB_PSEUDO, SI_ORDER_ANY, vmmeter_init, 0);
532