xref: /dragonfly/sys/vm/vm_meter.c (revision fb151170)
1 /*
2  * (MPSAFE)
3  *
4  * Copyright (c) 1982, 1986, 1989, 1993
5  *	The Regents of the University of California.  All rights reserved.
6  *
7  * Redistribution and use in source and binary forms, with or without
8  * modification, are permitted provided that the following conditions
9  * are met:
10  * 1. Redistributions of source code must retain the above copyright
11  *    notice, this list of conditions and the following disclaimer.
12  * 2. Redistributions in binary form must reproduce the above copyright
13  *    notice, this list of conditions and the following disclaimer in the
14  *    documentation and/or other materials provided with the distribution.
15  * 3. All advertising materials mentioning features or use of this software
16  *    must display the following acknowledgement:
17  *	This product includes software developed by the University of
18  *	California, Berkeley and its contributors.
19  * 4. Neither the name of the University nor the names of its contributors
20  *    may be used to endorse or promote products derived from this software
21  *    without specific prior written permission.
22  *
23  * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
24  * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
25  * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
26  * ARE DISCLAIMED.  IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
27  * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
28  * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
29  * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
30  * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
31  * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
32  * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
33  * SUCH DAMAGE.
34  *
35  *	@(#)vm_meter.c	8.4 (Berkeley) 1/4/94
36  * $FreeBSD: src/sys/vm/vm_meter.c,v 1.34.2.7 2002/10/10 19:28:22 dillon Exp $
37  * $DragonFly: src/sys/vm/vm_meter.c,v 1.15 2008/04/28 18:04:08 dillon Exp $
38  */
39 
40 #include <sys/param.h>
41 #include <sys/proc.h>
42 #include <sys/systm.h>
43 #include <sys/kernel.h>
44 #include <sys/resource.h>
45 #include <sys/vmmeter.h>
46 
47 #include <vm/vm.h>
48 #include <vm/vm_page.h>
49 #include <vm/vm_extern.h>
50 #include <vm/vm_param.h>
51 #include <sys/lock.h>
52 #include <vm/pmap.h>
53 #include <vm/vm_map.h>
54 #include <vm/vm_object.h>
55 #include <sys/sysctl.h>
56 
57 struct vmstats vmstats;
58 
59 static int maxslp = MAXSLP;
60 
61 SYSCTL_UINT(_vm, VM_V_FREE_MIN, v_free_min,
62 	CTLFLAG_RW, &vmstats.v_free_min, 0,
63 	"Minimum number of pages desired free");
64 SYSCTL_UINT(_vm, VM_V_FREE_TARGET, v_free_target,
65 	CTLFLAG_RW, &vmstats.v_free_target, 0,
66 	"Number of pages desired free");
67 SYSCTL_UINT(_vm, VM_V_FREE_RESERVED, v_free_reserved,
68 	CTLFLAG_RW, &vmstats.v_free_reserved, 0,
69 	"Number of pages reserved for deadlock");
70 SYSCTL_UINT(_vm, VM_V_INACTIVE_TARGET, v_inactive_target,
71 	CTLFLAG_RW, &vmstats.v_inactive_target, 0,
72 	"Number of pages desired inactive");
73 SYSCTL_UINT(_vm, VM_V_CACHE_MIN, v_cache_min,
74 	CTLFLAG_RW, &vmstats.v_cache_min, 0,
75 	"Min number of pages desired on cache queue");
76 SYSCTL_UINT(_vm, VM_V_CACHE_MAX, v_cache_max,
77 	CTLFLAG_RW, &vmstats.v_cache_max, 0,
78 	"Max number of pages in cached obj");
79 SYSCTL_UINT(_vm, VM_V_PAGEOUT_FREE_MIN, v_pageout_free_min,
80 	CTLFLAG_RW, &vmstats.v_pageout_free_min, 0,
81 	"Min number pages reserved for kernel");
82 SYSCTL_UINT(_vm, OID_AUTO, v_free_severe,
83 	CTLFLAG_RW, &vmstats.v_free_severe, 0, "");
84 
85 SYSCTL_STRUCT(_vm, VM_LOADAVG, loadavg, CTLFLAG_RD,
86     &averunnable, loadavg, "Machine loadaverage history");
87 
88 static int do_vmtotal_callback(struct proc *p, void *data);
89 
90 /*
91  * No requirements.
92  */
93 static int
94 do_vmtotal(SYSCTL_HANDLER_ARGS)
95 {
96 	struct vmtotal total;
97 	struct vmtotal *totalp;
98 	vm_object_t object;
99 
100 	bzero(&total, sizeof(total));
101 	totalp = &total;
102 
103 	/*
104 	 * Mark all objects as inactive.
105 	 */
106 	lwkt_gettoken(&vmobj_token);
107 	for (object = TAILQ_FIRST(&vm_object_list);
108 	    object != NULL;
109 	    object = TAILQ_NEXT(object,object_list)) {
110 		if (object->type == OBJT_MARKER)
111 			continue;
112 		vm_object_clear_flag(object, OBJ_ACTIVE);
113 	}
114 	lwkt_reltoken(&vmobj_token);
115 
116 	/*
117 	 * Calculate process statistics.
118 	 */
119 	allproc_scan(do_vmtotal_callback, totalp);
120 
121 	/*
122 	 * Calculate object memory usage statistics.
123 	 */
124 	lwkt_gettoken(&vmobj_token);
125 	for (object = TAILQ_FIRST(&vm_object_list);
126 	    object != NULL;
127 	    object = TAILQ_NEXT(object, object_list)) {
128 		/*
129 		 * devices, like /dev/mem, will badly skew our totals.
130 		 * markers aren't real objects.
131 		 */
132 		if (object->type == OBJT_MARKER)
133 			continue;
134 		if (object->type == OBJT_DEVICE)
135 			continue;
136 		if (object->size >= 0x7FFFFFFF) {
137 			/*
138 			 * Probably unbounded anonymous memory (really
139 			 * bounded by related vm_map_entry structures which
140 			 * we do not have access to in this loop).
141 			 */
142 			totalp->t_vm += object->resident_page_count;
143 		} else {
144 			/*
145 			 * It's questionable how useful this is but...
146 			 */
147 			totalp->t_vm += object->size;
148 		}
149 		totalp->t_rm += object->resident_page_count;
150 		if (object->flags & OBJ_ACTIVE) {
151 			totalp->t_avm += object->size;
152 			totalp->t_arm += object->resident_page_count;
153 		}
154 		if (object->shadow_count > 1) {
155 			/* shared object */
156 			totalp->t_vmshr += object->size;
157 			totalp->t_rmshr += object->resident_page_count;
158 			if (object->flags & OBJ_ACTIVE) {
159 				totalp->t_avmshr += object->size;
160 				totalp->t_armshr += object->resident_page_count;
161 			}
162 		}
163 	}
164 	lwkt_reltoken(&vmobj_token);
165 	totalp->t_free = vmstats.v_free_count + vmstats.v_cache_count;
166 
167 	return (sysctl_handle_opaque(oidp, totalp, sizeof total, req));
168 }
169 
170 /*
171  * The caller must hold proc_token.
172  */
173 static int
174 do_vmtotal_callback(struct proc *p, void *data)
175 {
176 	struct vmtotal *totalp = data;
177 	struct lwp *lp;
178 
179 	if (p->p_flags & P_SYSTEM)
180 		return(0);
181 
182 	FOREACH_LWP_IN_PROC(lp, p) {
183 		switch (lp->lwp_stat) {
184 		case LSSTOP:
185 		case LSSLEEP:
186 			if ((p->p_flags & P_SWAPPEDOUT) == 0) {
187 				if ((lp->lwp_flags & LWP_SINTR) == 0)
188 					totalp->t_dw++;
189 				else if (lp->lwp_slptime < maxslp)
190 					totalp->t_sl++;
191 			} else if (lp->lwp_slptime < maxslp) {
192 				totalp->t_sw++;
193 			}
194 			if (lp->lwp_slptime >= maxslp)
195 				return(0);
196 			break;
197 
198 		case LSRUN:
199 			if (p->p_flags & P_SWAPPEDOUT)
200 				totalp->t_sw++;
201 			else
202 				totalp->t_rq++;
203 			if (p->p_stat == SIDL)
204 				return(0);
205 			break;
206 
207 		default:
208 			return (0);
209 		}
210 
211 		/*
212 		 * Set while in vm_fault()
213 		 */
214 		if (lp->lwp_flags & LWP_PAGING)
215 			totalp->t_pw++;
216 	}
217 	return(0);
218 }
219 
220 /*
221  * No requirements.
222  */
223 static int
224 do_vmstats(SYSCTL_HANDLER_ARGS)
225 {
226 	struct vmstats vms = vmstats;
227 	return (sysctl_handle_opaque(oidp, &vms, sizeof(vms), req));
228 }
229 
230 /*
231  * No requirements.
232  */
233 static int
234 do_vmmeter(SYSCTL_HANDLER_ARGS)
235 {
236 	int boffset = offsetof(struct vmmeter, vmmeter_uint_begin);
237 	int eoffset = offsetof(struct vmmeter, vmmeter_uint_end);
238 	struct vmmeter vmm;
239 	int i;
240 
241 	bzero(&vmm, sizeof(vmm));
242 	for (i = 0; i < ncpus; ++i) {
243 		int off;
244 		struct globaldata *gd = globaldata_find(i);
245 
246 		for (off = boffset; off <= eoffset; off += sizeof(u_int)) {
247 			*(u_int *)((char *)&vmm + off) +=
248 				*(u_int *)((char *)&gd->gd_cnt + off);
249 		}
250 
251 	}
252 	vmm.v_intr += vmm.v_ipi + vmm.v_timer;
253 	return (sysctl_handle_opaque(oidp, &vmm, sizeof(vmm), req));
254 }
255 
256 /*
257  * vcnt() -	accumulate statistics from the cnt structure for each cpu
258  *
259  *	The vmmeter structure is now per-cpu as well as global.  Those
260  *	statistics which can be kept on a per-cpu basis (to avoid cache
261  *	stalls between cpus) can be moved to the per-cpu vmmeter.  Remaining
262  *	statistics, such as v_free_reserved, are left in the global
263  *	structure.
264  *
265  * (sysctl_oid *oidp, void *arg1, int arg2, struct sysctl_req *req)
266  *
267  * No requirements.
268  */
269 static int
270 vcnt(SYSCTL_HANDLER_ARGS)
271 {
272 	int i;
273 	int count = 0;
274 	int offset = arg2;
275 
276 	for (i = 0; i < ncpus; ++i) {
277 		struct globaldata *gd = globaldata_find(i);
278 		count += *(int *)((char *)&gd->gd_cnt + offset);
279 	}
280 	return(SYSCTL_OUT(req, &count, sizeof(int)));
281 }
282 
283 /*
284  * No requirements.
285  */
286 static int
287 vcnt_intr(SYSCTL_HANDLER_ARGS)
288 {
289 	int i;
290 	int count = 0;
291 
292 	for (i = 0; i < ncpus; ++i) {
293 		struct globaldata *gd = globaldata_find(i);
294 
295 		count += gd->gd_cnt.v_intr + gd->gd_cnt.v_ipi +
296 			 gd->gd_cnt.v_timer;
297 	}
298 	return(SYSCTL_OUT(req, &count, sizeof(int)));
299 }
300 
301 #define VMMETEROFF(var)	offsetof(struct vmmeter, var)
302 
303 SYSCTL_PROC(_vm, OID_AUTO, vmtotal, CTLTYPE_OPAQUE|CTLFLAG_RD,
304     0, sizeof(struct vmtotal), do_vmtotal, "S,vmtotal",
305     "System virtual memory aggregate");
306 SYSCTL_PROC(_vm, OID_AUTO, vmstats, CTLTYPE_OPAQUE|CTLFLAG_RD,
307     0, sizeof(struct vmstats), do_vmstats, "S,vmstats",
308     "System virtual memory statistics");
309 SYSCTL_PROC(_vm, OID_AUTO, vmmeter, CTLTYPE_OPAQUE|CTLFLAG_RD,
310     0, sizeof(struct vmmeter), do_vmmeter, "S,vmmeter",
311     "System statistics");
312 SYSCTL_NODE(_vm, OID_AUTO, stats, CTLFLAG_RW, 0, "VM meter stats");
313 SYSCTL_NODE(_vm_stats, OID_AUTO, sys, CTLFLAG_RW, 0, "VM meter sys stats");
314 SYSCTL_NODE(_vm_stats, OID_AUTO, vm, CTLFLAG_RW, 0, "VM meter vm stats");
315 SYSCTL_NODE(_vm_stats, OID_AUTO, misc, CTLFLAG_RW, 0, "VM meter misc stats");
316 
317 SYSCTL_PROC(_vm_stats_sys, OID_AUTO, v_swtch, CTLTYPE_UINT|CTLFLAG_RD,
318 	0, VMMETEROFF(v_swtch), vcnt, "IU", "Context switches");
319 SYSCTL_PROC(_vm_stats_sys, OID_AUTO, v_intrans_coll, CTLTYPE_UINT|CTLFLAG_RD,
320 	0, VMMETEROFF(v_intrans_coll), vcnt, "IU", "Intransit map collisions (total)");
321 SYSCTL_PROC(_vm_stats_sys, OID_AUTO, v_intrans_wait, CTLTYPE_UINT|CTLFLAG_RD,
322 	0, VMMETEROFF(v_intrans_wait), vcnt, "IU", "Intransit map collisions which blocked");
323 SYSCTL_PROC(_vm_stats_sys, OID_AUTO, v_forwarded_ints, CTLTYPE_UINT|CTLFLAG_RD,
324 	0, VMMETEROFF(v_forwarded_ints), vcnt, "IU", "Forwarded interrupts due to MP lock");
325 SYSCTL_PROC(_vm_stats_sys, OID_AUTO, v_forwarded_hits, CTLTYPE_UINT|CTLFLAG_RD,
326 	0, VMMETEROFF(v_forwarded_hits), vcnt, "IU", "Forwarded hits due to MP lock");
327 SYSCTL_PROC(_vm_stats_sys, OID_AUTO, v_forwarded_misses, CTLTYPE_UINT|CTLFLAG_RD,
328 	0, VMMETEROFF(v_forwarded_misses), vcnt, "IU", "Forwarded misses due to MP lock");
329 SYSCTL_PROC(_vm_stats_sys, OID_AUTO, v_trap, CTLTYPE_UINT|CTLFLAG_RD,
330 	0, VMMETEROFF(v_trap), vcnt, "IU", "Traps");
331 SYSCTL_PROC(_vm_stats_sys, OID_AUTO, v_syscall, CTLTYPE_UINT|CTLFLAG_RD,
332 	0, VMMETEROFF(v_syscall), vcnt, "IU", "Syscalls");
333 SYSCTL_PROC(_vm_stats_sys, OID_AUTO, v_intr, CTLTYPE_UINT|CTLFLAG_RD,
334 	0, VMMETEROFF(v_intr), vcnt_intr, "IU", "Hardware interrupts");
335 SYSCTL_PROC(_vm_stats_sys, OID_AUTO, v_ipi, CTLTYPE_UINT|CTLFLAG_RD,
336 	0, VMMETEROFF(v_ipi), vcnt, "IU", "Inter-processor interrupts");
337 SYSCTL_PROC(_vm_stats_sys, OID_AUTO, v_timer, CTLTYPE_UINT|CTLFLAG_RD,
338 	0, VMMETEROFF(v_timer), vcnt, "IU", "LAPIC timer interrupts");
339 SYSCTL_PROC(_vm_stats_sys, OID_AUTO, v_soft, CTLTYPE_UINT|CTLFLAG_RD,
340 	0, VMMETEROFF(v_soft), vcnt, "IU", "Software interrupts");
341 SYSCTL_PROC(_vm_stats_vm, OID_AUTO, v_vm_faults, CTLTYPE_UINT|CTLFLAG_RD,
342 	0, VMMETEROFF(v_vm_faults), vcnt, "IU", "VM faults");
343 SYSCTL_PROC(_vm_stats_vm, OID_AUTO, v_cow_faults, CTLTYPE_UINT|CTLFLAG_RD,
344 	0, VMMETEROFF(v_cow_faults), vcnt, "IU", "COW faults");
345 SYSCTL_PROC(_vm_stats_vm, OID_AUTO, v_cow_optim, CTLTYPE_UINT|CTLFLAG_RD,
346 	0, VMMETEROFF(v_cow_optim), vcnt, "IU", "Optimized COW faults");
347 SYSCTL_PROC(_vm_stats_vm, OID_AUTO, v_zfod, CTLTYPE_UINT|CTLFLAG_RD,
348 	0, VMMETEROFF(v_zfod), vcnt, "IU", "Zero fill");
349 SYSCTL_PROC(_vm_stats_vm, OID_AUTO, v_ozfod, CTLTYPE_UINT|CTLFLAG_RD,
350 	0, VMMETEROFF(v_ozfod), vcnt, "IU", "Optimized zero fill");
351 SYSCTL_PROC(_vm_stats_vm, OID_AUTO, v_swapin, CTLTYPE_UINT|CTLFLAG_RD,
352 	0, VMMETEROFF(v_swapin), vcnt, "IU", "Swapin operations");
353 SYSCTL_PROC(_vm_stats_vm, OID_AUTO, v_swapout, CTLTYPE_UINT|CTLFLAG_RD,
354 	0, VMMETEROFF(v_swapout), vcnt, "IU", "Swapout operations");
355 SYSCTL_PROC(_vm_stats_vm, OID_AUTO, v_swappgsin, CTLTYPE_UINT|CTLFLAG_RD,
356 	0, VMMETEROFF(v_swappgsin), vcnt, "IU", "Swapin pages");
357 SYSCTL_PROC(_vm_stats_vm, OID_AUTO, v_swappgsout, CTLTYPE_UINT|CTLFLAG_RD,
358 	0, VMMETEROFF(v_swappgsout), vcnt, "IU", "Swapout pages");
359 SYSCTL_PROC(_vm_stats_vm, OID_AUTO, v_vnodein, CTLTYPE_UINT|CTLFLAG_RD,
360 	0, VMMETEROFF(v_vnodein), vcnt, "IU", "Vnodein operations");
361 SYSCTL_PROC(_vm_stats_vm, OID_AUTO, v_vnodeout, CTLTYPE_UINT|CTLFLAG_RD,
362 	0, VMMETEROFF(v_vnodeout), vcnt, "IU", "Vnodeout operations");
363 SYSCTL_PROC(_vm_stats_vm, OID_AUTO, v_vnodepgsin, CTLTYPE_UINT|CTLFLAG_RD,
364 	0, VMMETEROFF(v_vnodepgsin), vcnt, "IU", "Vnodein pages");
365 SYSCTL_PROC(_vm_stats_vm, OID_AUTO, v_vnodepgsout, CTLTYPE_UINT|CTLFLAG_RD,
366 	0, VMMETEROFF(v_vnodepgsout), vcnt, "IU", "Vnodeout pages");
367 SYSCTL_PROC(_vm_stats_vm, OID_AUTO, v_intrans, CTLTYPE_UINT|CTLFLAG_RD,
368 	0, VMMETEROFF(v_intrans), vcnt, "IU", "In transit page blocking");
369 SYSCTL_PROC(_vm_stats_vm, OID_AUTO, v_reactivated, CTLTYPE_UINT|CTLFLAG_RD,
370 	0, VMMETEROFF(v_reactivated), vcnt, "IU", "Reactivated pages");
371 SYSCTL_PROC(_vm_stats_vm, OID_AUTO, v_pdwakeups, CTLTYPE_UINT|CTLFLAG_RD,
372 	0, VMMETEROFF(v_pdwakeups), vcnt, "IU", "Pagedaemon wakeups");
373 SYSCTL_PROC(_vm_stats_vm, OID_AUTO, v_ppwakeups, CTLTYPE_UINT|CTLFLAG_RD,
374 	0, VMMETEROFF(v_ppwakeups), vcnt, "IU", "vm_wait wakeups");
375 SYSCTL_PROC(_vm_stats_vm, OID_AUTO, v_pdpages, CTLTYPE_UINT|CTLFLAG_RD,
376 	0, VMMETEROFF(v_pdpages), vcnt, "IU", "Pagedaemon page scans");
377 SYSCTL_PROC(_vm_stats_vm, OID_AUTO, v_dfree, CTLTYPE_UINT|CTLFLAG_RD,
378 	0, VMMETEROFF(v_dfree), vcnt, "IU", "Pages freed by daemon");
379 SYSCTL_PROC(_vm_stats_vm, OID_AUTO, v_pfree, CTLTYPE_UINT|CTLFLAG_RD,
380 	0, VMMETEROFF(v_pfree), vcnt, "IU", "Pages freed by exiting processes");
381 SYSCTL_PROC(_vm_stats_vm, OID_AUTO, v_tfree, CTLTYPE_UINT|CTLFLAG_RD,
382 	0, VMMETEROFF(v_tfree), vcnt, "IU", "Total pages freed");
383 SYSCTL_PROC(_vm_stats_vm, OID_AUTO, v_forks, CTLTYPE_UINT|CTLFLAG_RD,
384 	0, VMMETEROFF(v_forks), vcnt, "IU", "Number of fork() calls");
385 SYSCTL_PROC(_vm_stats_vm, OID_AUTO, v_vforks, CTLTYPE_UINT|CTLFLAG_RD,
386 	0, VMMETEROFF(v_vforks), vcnt, "IU", "Number of vfork() calls");
387 SYSCTL_PROC(_vm_stats_vm, OID_AUTO, v_rforks, CTLTYPE_UINT|CTLFLAG_RD,
388 	0, VMMETEROFF(v_rforks), vcnt, "IU", "Number of rfork() calls");
389 SYSCTL_PROC(_vm_stats_vm, OID_AUTO, v_kthreads, CTLTYPE_UINT|CTLFLAG_RD,
390 	0, VMMETEROFF(v_kthreads), vcnt, "IU", "Number of fork() calls by kernel");
391 SYSCTL_PROC(_vm_stats_vm, OID_AUTO, v_forkpages, CTLTYPE_UINT|CTLFLAG_RD,
392 	0, VMMETEROFF(v_forkpages), vcnt, "IU", "VM pages affected by fork()");
393 SYSCTL_PROC(_vm_stats_vm, OID_AUTO, v_vforkpages, CTLTYPE_UINT|CTLFLAG_RD,
394 	0, VMMETEROFF(v_vforkpages), vcnt, "IU", "VM pages affected by vfork()");
395 SYSCTL_PROC(_vm_stats_vm, OID_AUTO, v_rforkpages, CTLTYPE_UINT|CTLFLAG_RD,
396 	0, VMMETEROFF(v_rforkpages), vcnt, "IU", "VM pages affected by rfork()");
397 SYSCTL_PROC(_vm_stats_vm, OID_AUTO, v_kthreadpages, CTLTYPE_UINT|CTLFLAG_RD,
398 	0, VMMETEROFF(v_kthreadpages), vcnt, "IU", "VM pages affected by fork() by kernel");
399 
400 SYSCTL_UINT(_vm_stats_vm, OID_AUTO,
401 	v_page_size, CTLFLAG_RD, &vmstats.v_page_size, 0,
402 	"Page size in bytes");
403 SYSCTL_UINT(_vm_stats_vm, OID_AUTO,
404 	v_page_count, CTLFLAG_RD, &vmstats.v_page_count, 0,
405 	"Total number of pages in system");
406 SYSCTL_UINT(_vm_stats_vm, OID_AUTO,
407 	v_free_reserved, CTLFLAG_RD, &vmstats.v_free_reserved, 0,
408 	"Number of pages reserved for deadlock");
409 SYSCTL_UINT(_vm_stats_vm, OID_AUTO,
410 	v_free_target, CTLFLAG_RD, &vmstats.v_free_target, 0,
411 	"Number of pages desired free");
412 SYSCTL_UINT(_vm_stats_vm, OID_AUTO,
413 	v_free_min, CTLFLAG_RD, &vmstats.v_free_min, 0,
414 	"Minimum number of pages desired free");
415 SYSCTL_UINT(_vm_stats_vm, OID_AUTO,
416 	v_free_count, CTLFLAG_RD, &vmstats.v_free_count, 0,
417 	"Number of pages free");
418 SYSCTL_UINT(_vm_stats_vm, OID_AUTO,
419 	v_wire_count, CTLFLAG_RD, &vmstats.v_wire_count, 0,
420 	"Number of pages wired down");
421 SYSCTL_UINT(_vm_stats_vm, OID_AUTO,
422 	v_active_count, CTLFLAG_RD, &vmstats.v_active_count, 0,
423 	"Number of pages active");
424 SYSCTL_UINT(_vm_stats_vm, OID_AUTO,
425 	v_inactive_target, CTLFLAG_RD, &vmstats.v_inactive_target, 0,
426 	"Number of pages desired inactive");
427 SYSCTL_UINT(_vm_stats_vm, OID_AUTO,
428 	v_inactive_count, CTLFLAG_RD, &vmstats.v_inactive_count, 0,
429 	"Number of pages inactive");
430 SYSCTL_UINT(_vm_stats_vm, OID_AUTO,
431 	v_cache_count, CTLFLAG_RD, &vmstats.v_cache_count, 0,
432 	"Number of pages on buffer cache queue");
433 SYSCTL_UINT(_vm_stats_vm, OID_AUTO,
434 	v_cache_min, CTLFLAG_RD, &vmstats.v_cache_min, 0,
435 	"Min number of pages desired on cache queue");
436 SYSCTL_UINT(_vm_stats_vm, OID_AUTO,
437 	v_cache_max, CTLFLAG_RD, &vmstats.v_cache_max, 0,
438 	"Max number of pages in cached obj");
439 SYSCTL_UINT(_vm_stats_vm, OID_AUTO,
440 	v_pageout_free_min, CTLFLAG_RD, &vmstats.v_pageout_free_min, 0,
441 	"Min number pages reserved for kernel");
442 SYSCTL_UINT(_vm_stats_vm, OID_AUTO,
443 	v_interrupt_free_min, CTLFLAG_RD, &vmstats.v_interrupt_free_min, 0,
444 	"Reserved number of pages for int code");
445 SYSCTL_INT(_vm_stats_misc, OID_AUTO,
446 	zero_page_count, CTLFLAG_RD, &vm_page_zero_count, 0,
447 	"Number of zeroing pages");
448 
449 /*
450  * No requirements.
451  */
452 static int
453 do_vmmeter_pcpu(SYSCTL_HANDLER_ARGS)
454 {
455 	int boffset = offsetof(struct vmmeter, vmmeter_uint_begin);
456 	int eoffset = offsetof(struct vmmeter, vmmeter_uint_end);
457 	struct globaldata *gd = arg1;
458 	struct vmmeter vmm;
459 	int off;
460 
461 	bzero(&vmm, sizeof(vmm));
462 	for (off = boffset; off <= eoffset; off += sizeof(u_int)) {
463 		*(u_int *)((char *)&vmm + off) +=
464 			*(u_int *)((char *)&gd->gd_cnt + off);
465 	}
466 	vmm.v_intr += vmm.v_ipi + vmm.v_timer;
467 	return (sysctl_handle_opaque(oidp, &vmm, sizeof(vmm), req));
468 }
469 
470 /*
471  * Called from the low level boot code only.
472  */
473 static void
474 vmmeter_init(void *dummy __unused)
475 {
476 	int i;
477 
478 	for (i = 0; i < ncpus; ++i) {
479 		struct sysctl_ctx_list *ctx;
480 		struct sysctl_oid *oid;
481 		struct globaldata *gd;
482 		char name[32];
483 
484 		ksnprintf(name, sizeof(name), "cpu%d", i);
485 
486 		ctx = kmalloc(sizeof(*ctx), M_TEMP, M_WAITOK);
487 		sysctl_ctx_init(ctx);
488 		oid = SYSCTL_ADD_NODE(ctx, SYSCTL_STATIC_CHILDREN(_vm),
489 				      OID_AUTO, name, CTLFLAG_RD, 0, "");
490 
491 		gd = globaldata_find(i);
492 		SYSCTL_ADD_PROC(ctx, SYSCTL_CHILDREN(oid), OID_AUTO,
493 				"vmmeter", CTLTYPE_OPAQUE|CTLFLAG_RD,
494 				gd, sizeof(struct vmmeter), do_vmmeter_pcpu,
495 				"S,vmmeter", "System per-cpu statistics");
496 	}
497 }
498 SYSINIT(vmmeter, SI_SUB_PSEUDO, SI_ORDER_ANY, vmmeter_init, 0);
499