1/*-
2 * SPDX-License-Identifier: BSD-3-Clause
3 *
4 * Copyright (c) 1982, 1986, 1989, 1993
5 *	The Regents of the University of California.  All rights reserved.
6 *
7 * Redistribution and use in source and binary forms, with or without
8 * modification, are permitted provided that the following conditions
9 * are met:
10 * 1. Redistributions of source code must retain the above copyright
11 *    notice, this list of conditions and the following disclaimer.
12 * 2. Redistributions in binary form must reproduce the above copyright
13 *    notice, this list of conditions and the following disclaimer in the
14 *    documentation and/or other materials provided with the distribution.
15 * 3. Neither the name of the University nor the names of its contributors
16 *    may be used to endorse or promote products derived from this software
17 *    without specific prior written permission.
18 *
19 * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
20 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
21 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
22 * ARE DISCLAIMED.  IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
23 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
24 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
25 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
26 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
27 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
28 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
29 * SUCH DAMAGE.
30 */
31
32#include <sys/param.h>
33#include <sys/systm.h>
34#include <sys/kernel.h>
35#include <sys/lock.h>
36#include <sys/malloc.h>
37#include <sys/mutex.h>
38#include <sys/proc.h>
39#include <sys/resource.h>
40#include <sys/rwlock.h>
41#include <sys/sx.h>
42#include <sys/vmmeter.h>
43#include <sys/smp.h>
44
45#include <vm/vm.h>
46#include <vm/vm_page.h>
47#include <vm/vm_extern.h>
48#include <vm/vm_param.h>
49#include <vm/vm_phys.h>
50#include <vm/vm_pagequeue.h>
51#include <vm/pmap.h>
52#include <vm/vm_map.h>
53#include <vm/vm_object.h>
54#include <sys/sysctl.h>
55
56struct vmmeter __read_mostly vm_cnt = {
57	.v_swtch = EARLY_COUNTER,
58	.v_trap = EARLY_COUNTER,
59	.v_syscall = EARLY_COUNTER,
60	.v_intr = EARLY_COUNTER,
61	.v_soft = EARLY_COUNTER,
62	.v_vm_faults = EARLY_COUNTER,
63	.v_io_faults = EARLY_COUNTER,
64	.v_cow_faults = EARLY_COUNTER,
65	.v_cow_optim = EARLY_COUNTER,
66	.v_zfod = EARLY_COUNTER,
67	.v_ozfod = EARLY_COUNTER,
68	.v_swapin = EARLY_COUNTER,
69	.v_swapout = EARLY_COUNTER,
70	.v_swappgsin = EARLY_COUNTER,
71	.v_swappgsout = EARLY_COUNTER,
72	.v_vnodein = EARLY_COUNTER,
73	.v_vnodeout = EARLY_COUNTER,
74	.v_vnodepgsin = EARLY_COUNTER,
75	.v_vnodepgsout = EARLY_COUNTER,
76	.v_intrans = EARLY_COUNTER,
77	.v_reactivated = EARLY_COUNTER,
78	.v_pdwakeups = EARLY_COUNTER,
79	.v_pdpages = EARLY_COUNTER,
80	.v_pdshortfalls = EARLY_COUNTER,
81	.v_dfree = EARLY_COUNTER,
82	.v_pfree = EARLY_COUNTER,
83	.v_tfree = EARLY_COUNTER,
84	.v_forks = EARLY_COUNTER,
85	.v_vforks = EARLY_COUNTER,
86	.v_rforks = EARLY_COUNTER,
87	.v_kthreads = EARLY_COUNTER,
88	.v_forkpages = EARLY_COUNTER,
89	.v_vforkpages = EARLY_COUNTER,
90	.v_rforkpages = EARLY_COUNTER,
91	.v_kthreadpages = EARLY_COUNTER,
92	.v_wire_count = EARLY_COUNTER,
93};
94
95u_long __exclusive_cache_line vm_user_wire_count;
96
97static void
98vmcounter_startup(void)
99{
100	counter_u64_t *cnt = (counter_u64_t *)&vm_cnt;
101
102	COUNTER_ARRAY_ALLOC(cnt, VM_METER_NCOUNTERS, M_WAITOK);
103}
104SYSINIT(counter, SI_SUB_KMEM, SI_ORDER_FIRST, vmcounter_startup, NULL);
105
106SYSCTL_UINT(_vm, VM_V_FREE_MIN, v_free_min,
107	CTLFLAG_RW, &vm_cnt.v_free_min, 0, "Minimum low-free-pages threshold");
108SYSCTL_UINT(_vm, VM_V_FREE_TARGET, v_free_target,
109	CTLFLAG_RW, &vm_cnt.v_free_target, 0, "Desired free pages");
110SYSCTL_UINT(_vm, VM_V_FREE_RESERVED, v_free_reserved,
111	CTLFLAG_RW, &vm_cnt.v_free_reserved, 0, "Pages reserved for deadlock");
112SYSCTL_UINT(_vm, VM_V_INACTIVE_TARGET, v_inactive_target,
113	CTLFLAG_RW, &vm_cnt.v_inactive_target, 0, "Pages desired inactive");
114SYSCTL_UINT(_vm, VM_V_PAGEOUT_FREE_MIN, v_pageout_free_min,
115	CTLFLAG_RW, &vm_cnt.v_pageout_free_min, 0, "Min pages reserved for kernel");
116SYSCTL_UINT(_vm, OID_AUTO, v_free_severe,
117	CTLFLAG_RW, &vm_cnt.v_free_severe, 0, "Severe page depletion point");
118
119static int
120sysctl_vm_loadavg(SYSCTL_HANDLER_ARGS)
121{
122
123#ifdef SCTL_MASK32
124	uint32_t la[4];
125
126	if (req->flags & SCTL_MASK32) {
127		la[0] = averunnable.ldavg[0];
128		la[1] = averunnable.ldavg[1];
129		la[2] = averunnable.ldavg[2];
130		la[3] = averunnable.fscale;
131		return SYSCTL_OUT(req, la, sizeof(la));
132	} else
133#endif
134		return SYSCTL_OUT(req, &averunnable, sizeof(averunnable));
135}
136SYSCTL_PROC(_vm, VM_LOADAVG, loadavg, CTLTYPE_STRUCT | CTLFLAG_RD |
137    CTLFLAG_MPSAFE, NULL, 0, sysctl_vm_loadavg, "S,loadavg",
138    "Machine loadaverage history");
139
140#if defined(COMPAT_FREEBSD11)
141struct vmtotal11 {
142	int16_t	t_rq;
143	int16_t	t_dw;
144	int16_t	t_pw;
145	int16_t	t_sl;
146	int16_t	t_sw;
147	int32_t	t_vm;
148	int32_t	t_avm;
149	int32_t	t_rm;
150	int32_t	t_arm;
151	int32_t	t_vmshr;
152	int32_t	t_avmshr;
153	int32_t	t_rmshr;
154	int32_t	t_armshr;
155	int32_t	t_free;
156};
157#endif
158
159static int
160vmtotal(SYSCTL_HANDLER_ARGS)
161{
162	struct vmtotal total;
163#if defined(COMPAT_FREEBSD11)
164	struct vmtotal11 total11;
165#endif
166	vm_object_t object;
167	struct proc *p;
168	struct thread *td;
169
170	if (req->oldptr == NULL) {
171#if defined(COMPAT_FREEBSD11)
172		if (curproc->p_osrel < P_OSREL_VMTOTAL64)
173			return (SYSCTL_OUT(req, NULL, sizeof(total11)));
174#endif
175		return (SYSCTL_OUT(req, NULL, sizeof(total)));
176	}
177	bzero(&total, sizeof(total));
178
179	/*
180	 * Calculate process statistics.
181	 */
182	sx_slock(&allproc_lock);
183	FOREACH_PROC_IN_SYSTEM(p) {
184		if ((p->p_flag & P_SYSTEM) != 0)
185			continue;
186		PROC_LOCK(p);
187		if (p->p_state != PRS_NEW) {
188			FOREACH_THREAD_IN_PROC(p, td) {
189				thread_lock(td);
190				switch (TD_GET_STATE(td)) {
191				case TDS_INHIBITED:
192					if (TD_IS_SWAPPED(td))
193						total.t_sw++;
194					else if (TD_IS_SLEEPING(td)) {
195						if (td->td_priority <= PZERO)
196							total.t_dw++;
197						else
198							total.t_sl++;
199					}
200					break;
201				case TDS_CAN_RUN:
202					total.t_sw++;
203					break;
204				case TDS_RUNQ:
205				case TDS_RUNNING:
206					total.t_rq++;
207					break;
208				default:
209					break;
210				}
211				thread_unlock(td);
212			}
213		}
214		PROC_UNLOCK(p);
215	}
216	sx_sunlock(&allproc_lock);
217	/*
218	 * Calculate object memory usage statistics.
219	 */
220	mtx_lock(&vm_object_list_mtx);
221	TAILQ_FOREACH(object, &vm_object_list, object_list) {
222		/*
223		 * Perform unsynchronized reads on the object.  In
224		 * this case, the lack of synchronization should not
225		 * impair the accuracy of the reported statistics.
226		 */
227		if ((object->flags & OBJ_FICTITIOUS) != 0) {
228			/*
229			 * Devices, like /dev/mem, will badly skew our totals.
230			 */
231			continue;
232		}
233		if (object->ref_count == 0) {
234			/*
235			 * Also skip unreferenced objects, including
236			 * vnodes representing mounted file systems.
237			 */
238			continue;
239		}
240		if (object->ref_count == 1 &&
241		    (object->flags & (OBJ_ANON | OBJ_SWAP)) == OBJ_SWAP) {
242			/*
243			 * Also skip otherwise unreferenced swap
244			 * objects backing tmpfs vnodes, and POSIX or
245			 * SysV shared memory.
246			 */
247			continue;
248		}
249		total.t_vm += object->size;
250		total.t_rm += object->resident_page_count;
251		if (vm_object_is_active(object)) {
252			total.t_avm += object->size;
253			total.t_arm += object->resident_page_count;
254		}
255		if (object->shadow_count > 1) {
256			/* shared object */
257			total.t_vmshr += object->size;
258			total.t_rmshr += object->resident_page_count;
259			if (vm_object_is_active(object)) {
260				total.t_avmshr += object->size;
261				total.t_armshr += object->resident_page_count;
262			}
263		}
264	}
265	mtx_unlock(&vm_object_list_mtx);
266	total.t_pw = vm_wait_count();
267	total.t_free = vm_free_count();
268#if defined(COMPAT_FREEBSD11)
269	/* sysctl(8) allocates twice as much memory as reported by sysctl(3) */
270	if (curproc->p_osrel < P_OSREL_VMTOTAL64 && (req->oldlen ==
271	    sizeof(total11) || req->oldlen == 2 * sizeof(total11))) {
272		bzero(&total11, sizeof(total11));
273		total11.t_rq = total.t_rq;
274		total11.t_dw = total.t_dw;
275		total11.t_pw = total.t_pw;
276		total11.t_sl = total.t_sl;
277		total11.t_sw = total.t_sw;
278		total11.t_vm = total.t_vm;	/* truncate */
279		total11.t_avm = total.t_avm;	/* truncate */
280		total11.t_rm = total.t_rm;	/* truncate */
281		total11.t_arm = total.t_arm;	/* truncate */
282		total11.t_vmshr = total.t_vmshr;	/* truncate */
283		total11.t_avmshr = total.t_avmshr;	/* truncate */
284		total11.t_rmshr = total.t_rmshr;	/* truncate */
285		total11.t_armshr = total.t_armshr;	/* truncate */
286		total11.t_free = total.t_free;		/* truncate */
287		return (SYSCTL_OUT(req, &total11, sizeof(total11)));
288	}
289#endif
290	return (SYSCTL_OUT(req, &total, sizeof(total)));
291}
292
293SYSCTL_PROC(_vm, VM_TOTAL, vmtotal, CTLTYPE_OPAQUE | CTLFLAG_RD |
294    CTLFLAG_MPSAFE, NULL, 0, vmtotal, "S,vmtotal",
295    "System virtual memory statistics");
296SYSCTL_NODE(_vm, OID_AUTO, stats, CTLFLAG_RW | CTLFLAG_MPSAFE, 0,
297    "VM meter stats");
298static SYSCTL_NODE(_vm_stats, OID_AUTO, sys, CTLFLAG_RW | CTLFLAG_MPSAFE, 0,
299    "VM meter sys stats");
300static SYSCTL_NODE(_vm_stats, OID_AUTO, vm, CTLFLAG_RW | CTLFLAG_MPSAFE, 0,
301    "VM meter vm stats");
302SYSCTL_NODE(_vm_stats, OID_AUTO, misc, CTLFLAG_RW | CTLFLAG_MPSAFE, 0,
303    "VM meter misc stats");
304
305static int
306sysctl_handle_vmstat(SYSCTL_HANDLER_ARGS)
307{
308	uint64_t val;
309#ifdef COMPAT_FREEBSD11
310	uint32_t val32;
311#endif
312
313	val = counter_u64_fetch(*(counter_u64_t *)arg1);
314#ifdef COMPAT_FREEBSD11
315	if (req->oldlen == sizeof(val32)) {
316		val32 = val;		/* truncate */
317		return (SYSCTL_OUT(req, &val32, sizeof(val32)));
318	}
319#endif
320	return (SYSCTL_OUT(req, &val, sizeof(val)));
321}
322
323#define	VM_STATS(parent, var, descr) \
324    SYSCTL_OID(parent, OID_AUTO, var, CTLTYPE_U64 | CTLFLAG_MPSAFE | \
325    CTLFLAG_RD, &vm_cnt.var, 0, sysctl_handle_vmstat, "QU", descr)
326#define	VM_STATS_VM(var, descr)		VM_STATS(_vm_stats_vm, var, descr)
327#define	VM_STATS_SYS(var, descr)	VM_STATS(_vm_stats_sys, var, descr)
328
329VM_STATS_SYS(v_swtch, "Context switches");
330VM_STATS_SYS(v_trap, "Traps");
331VM_STATS_SYS(v_syscall, "System calls");
332VM_STATS_SYS(v_intr, "Device interrupts");
333VM_STATS_SYS(v_soft, "Software interrupts");
334VM_STATS_VM(v_vm_faults, "Address memory faults");
335VM_STATS_VM(v_io_faults, "Page faults requiring I/O");
336VM_STATS_VM(v_cow_faults, "Copy-on-write faults");
337VM_STATS_VM(v_cow_optim, "Optimized COW faults");
338VM_STATS_VM(v_zfod, "Pages zero-filled on demand");
339VM_STATS_VM(v_ozfod, "Optimized zero fill pages");
340VM_STATS_VM(v_swapin, "Swap pager pageins");
341VM_STATS_VM(v_swapout, "Swap pager pageouts");
342VM_STATS_VM(v_swappgsin, "Swap pages swapped in");
343VM_STATS_VM(v_swappgsout, "Swap pages swapped out");
344VM_STATS_VM(v_vnodein, "Vnode pager pageins");
345VM_STATS_VM(v_vnodeout, "Vnode pager pageouts");
346VM_STATS_VM(v_vnodepgsin, "Vnode pages paged in");
347VM_STATS_VM(v_vnodepgsout, "Vnode pages paged out");
348VM_STATS_VM(v_intrans, "In transit page faults");
349VM_STATS_VM(v_reactivated, "Pages reactivated by pagedaemon");
350VM_STATS_VM(v_pdwakeups, "Pagedaemon wakeups");
351VM_STATS_VM(v_pdshortfalls, "Page reclamation shortfalls");
352VM_STATS_VM(v_dfree, "Pages freed by pagedaemon");
353VM_STATS_VM(v_pfree, "Pages freed by exiting processes");
354VM_STATS_VM(v_tfree, "Total pages freed");
355VM_STATS_VM(v_forks, "Number of fork() calls");
356VM_STATS_VM(v_vforks, "Number of vfork() calls");
357VM_STATS_VM(v_rforks, "Number of rfork() calls");
358VM_STATS_VM(v_kthreads, "Number of fork() calls by kernel");
359VM_STATS_VM(v_forkpages, "VM pages affected by fork()");
360VM_STATS_VM(v_vforkpages, "VM pages affected by vfork()");
361VM_STATS_VM(v_rforkpages, "VM pages affected by rfork()");
362VM_STATS_VM(v_kthreadpages, "VM pages affected by fork() by kernel");
363
364static int
365sysctl_handle_vmstat_proc(SYSCTL_HANDLER_ARGS)
366{
367	u_int (*fn)(void);
368	uint32_t val;
369
370	fn = arg1;
371	val = fn();
372	return (SYSCTL_OUT(req, &val, sizeof(val)));
373}
374
375#define	VM_STATS_PROC(var, descr, fn) \
376    SYSCTL_OID(_vm_stats_vm, OID_AUTO, var, CTLTYPE_U32 | CTLFLAG_MPSAFE | \
377    CTLFLAG_RD, fn, 0, sysctl_handle_vmstat_proc, "IU", descr)
378
379#define	VM_STATS_UINT(var, descr)	\
380    SYSCTL_UINT(_vm_stats_vm, OID_AUTO, var, CTLFLAG_RD, &vm_cnt.var, 0, descr)
381#define	VM_STATS_ULONG(var, descr)	\
382    SYSCTL_ULONG(_vm_stats_vm, OID_AUTO, var, CTLFLAG_RD, &vm_cnt.var, 0, descr)
383
384VM_STATS_UINT(v_page_size, "Page size in bytes");
385VM_STATS_UINT(v_page_count, "Total number of pages in system");
386VM_STATS_UINT(v_free_reserved, "Pages reserved for deadlock");
387VM_STATS_UINT(v_free_target, "Pages desired free");
388VM_STATS_UINT(v_free_min, "Minimum low-free-pages threshold");
389VM_STATS_PROC(v_free_count, "Free pages", vm_free_count);
390VM_STATS_PROC(v_wire_count, "Wired pages", vm_wire_count);
391VM_STATS_PROC(v_active_count, "Active pages", vm_active_count);
392VM_STATS_UINT(v_inactive_target, "Desired inactive pages");
393VM_STATS_PROC(v_inactive_count, "Inactive pages", vm_inactive_count);
394VM_STATS_PROC(v_laundry_count, "Pages eligible for laundering",
395    vm_laundry_count);
396VM_STATS_UINT(v_pageout_free_min, "Min pages reserved for kernel");
397VM_STATS_UINT(v_interrupt_free_min, "Reserved pages for interrupt code");
398VM_STATS_UINT(v_free_severe, "Severe page depletion point");
399
400SYSCTL_ULONG(_vm_stats_vm, OID_AUTO, v_user_wire_count, CTLFLAG_RD,
401    &vm_user_wire_count, 0, "User-wired virtual memory");
402
403#ifdef COMPAT_FREEBSD11
404/*
405 * Provide compatibility sysctls for the benefit of old utilities which exit
406 * with an error if they cannot be found.
407 */
408SYSCTL_UINT(_vm_stats_vm, OID_AUTO, v_cache_count, CTLFLAG_RD,
409    SYSCTL_NULL_UINT_PTR, 0, "Dummy for compatibility");
410SYSCTL_UINT(_vm_stats_vm, OID_AUTO, v_tcached, CTLFLAG_RD,
411    SYSCTL_NULL_UINT_PTR, 0, "Dummy for compatibility");
412#endif
413
414u_int
415vm_free_count(void)
416{
417	u_int v;
418	int i;
419
420	v = 0;
421	for (i = 0; i < vm_ndomains; i++)
422		v += vm_dom[i].vmd_free_count;
423
424	return (v);
425}
426
427static u_int
428vm_pagequeue_count(int pq)
429{
430	u_int v;
431	int i;
432
433	v = 0;
434	for (i = 0; i < vm_ndomains; i++)
435		v += vm_dom[i].vmd_pagequeues[pq].pq_cnt;
436
437	return (v);
438}
439
440u_int
441vm_active_count(void)
442{
443
444	return (vm_pagequeue_count(PQ_ACTIVE));
445}
446
447u_int
448vm_inactive_count(void)
449{
450
451	return (vm_pagequeue_count(PQ_INACTIVE));
452}
453
454u_int
455vm_laundry_count(void)
456{
457
458	return (vm_pagequeue_count(PQ_LAUNDRY));
459}
460
461static int
462sysctl_vm_pdpages(SYSCTL_HANDLER_ARGS)
463{
464	struct vm_pagequeue *pq;
465	uint64_t ret;
466	int dom, i;
467
468	ret = counter_u64_fetch(vm_cnt.v_pdpages);
469	for (dom = 0; dom < vm_ndomains; dom++)
470		for (i = 0; i < PQ_COUNT; i++) {
471			pq = &VM_DOMAIN(dom)->vmd_pagequeues[i];
472			ret += pq->pq_pdpages;
473		}
474	return (SYSCTL_OUT(req, &ret, sizeof(ret)));
475}
476SYSCTL_PROC(_vm_stats_vm, OID_AUTO, v_pdpages,
477    CTLTYPE_U64 | CTLFLAG_MPSAFE | CTLFLAG_RD, NULL, 0, sysctl_vm_pdpages, "QU",
478    "Pages analyzed by pagedaemon");
479
480static void
481vm_domain_stats_init(struct vm_domain *vmd, struct sysctl_oid *parent)
482{
483	struct sysctl_oid *oid;
484
485	vmd->vmd_oid = SYSCTL_ADD_NODE(NULL, SYSCTL_CHILDREN(parent), OID_AUTO,
486	    vmd->vmd_name, CTLFLAG_RD | CTLFLAG_MPSAFE, NULL, "");
487	oid = SYSCTL_ADD_NODE(NULL, SYSCTL_CHILDREN(vmd->vmd_oid), OID_AUTO,
488	    "stats", CTLFLAG_RD | CTLFLAG_MPSAFE, NULL, "");
489	SYSCTL_ADD_UINT(NULL, SYSCTL_CHILDREN(oid), OID_AUTO,
490	    "free_count", CTLFLAG_RD, &vmd->vmd_free_count, 0,
491	    "Free pages");
492	SYSCTL_ADD_UINT(NULL, SYSCTL_CHILDREN(oid), OID_AUTO,
493	    "active", CTLFLAG_RD, &vmd->vmd_pagequeues[PQ_ACTIVE].pq_cnt, 0,
494	    "Active pages");
495	SYSCTL_ADD_U64(NULL, SYSCTL_CHILDREN(oid), OID_AUTO,
496	    "actpdpgs", CTLFLAG_RD,
497	    &vmd->vmd_pagequeues[PQ_ACTIVE].pq_pdpages, 0,
498	    "Active pages scanned by the page daemon");
499	SYSCTL_ADD_UINT(NULL, SYSCTL_CHILDREN(oid), OID_AUTO,
500	    "inactive", CTLFLAG_RD, &vmd->vmd_pagequeues[PQ_INACTIVE].pq_cnt, 0,
501	    "Inactive pages");
502	SYSCTL_ADD_U64(NULL, SYSCTL_CHILDREN(oid), OID_AUTO,
503	    "inactpdpgs", CTLFLAG_RD,
504	    &vmd->vmd_pagequeues[PQ_INACTIVE].pq_pdpages, 0,
505	    "Inactive pages scanned by the page daemon");
506	SYSCTL_ADD_UINT(NULL, SYSCTL_CHILDREN(oid), OID_AUTO,
507	    "laundry", CTLFLAG_RD, &vmd->vmd_pagequeues[PQ_LAUNDRY].pq_cnt, 0,
508	    "laundry pages");
509	SYSCTL_ADD_U64(NULL, SYSCTL_CHILDREN(oid), OID_AUTO,
510	    "laundpdpgs", CTLFLAG_RD,
511	    &vmd->vmd_pagequeues[PQ_LAUNDRY].pq_pdpages, 0,
512	    "Laundry pages scanned by the page daemon");
513	SYSCTL_ADD_UINT(NULL, SYSCTL_CHILDREN(oid), OID_AUTO, "unswappable",
514	    CTLFLAG_RD, &vmd->vmd_pagequeues[PQ_UNSWAPPABLE].pq_cnt, 0,
515	    "Unswappable pages");
516	SYSCTL_ADD_U64(NULL, SYSCTL_CHILDREN(oid), OID_AUTO,
517	    "unswppdpgs", CTLFLAG_RD,
518	    &vmd->vmd_pagequeues[PQ_UNSWAPPABLE].pq_pdpages, 0,
519	    "Unswappable pages scanned by the page daemon");
520	SYSCTL_ADD_UINT(NULL, SYSCTL_CHILDREN(oid), OID_AUTO,
521	    "inactive_target", CTLFLAG_RD, &vmd->vmd_inactive_target, 0,
522	    "Target inactive pages");
523	SYSCTL_ADD_UINT(NULL, SYSCTL_CHILDREN(oid), OID_AUTO,
524	    "free_target", CTLFLAG_RD, &vmd->vmd_free_target, 0,
525	    "Target free pages");
526	SYSCTL_ADD_UINT(NULL, SYSCTL_CHILDREN(oid), OID_AUTO,
527	    "free_reserved", CTLFLAG_RD, &vmd->vmd_free_reserved, 0,
528	    "Reserved free pages");
529	SYSCTL_ADD_UINT(NULL, SYSCTL_CHILDREN(oid), OID_AUTO,
530	    "free_min", CTLFLAG_RD, &vmd->vmd_free_min, 0,
531	    "Minimum free pages");
532	SYSCTL_ADD_UINT(NULL, SYSCTL_CHILDREN(oid), OID_AUTO,
533	    "free_severe", CTLFLAG_RD, &vmd->vmd_free_severe, 0,
534	    "Severe free pages");
535	SYSCTL_ADD_UINT(NULL, SYSCTL_CHILDREN(oid), OID_AUTO,
536	    "inactive_pps", CTLFLAG_RD, &vmd->vmd_inactive_pps, 0,
537	    "inactive pages freed/second");
538
539}
540
541static void
542vm_stats_init(void *arg __unused)
543{
544	struct sysctl_oid *oid;
545	int i;
546
547	oid = SYSCTL_ADD_NODE(NULL, SYSCTL_STATIC_CHILDREN(_vm), OID_AUTO,
548	    "domain", CTLFLAG_RD | CTLFLAG_MPSAFE, NULL, "");
549	for (i = 0; i < vm_ndomains; i++)
550		vm_domain_stats_init(VM_DOMAIN(i), oid);
551}
552
553SYSINIT(vmstats_init, SI_SUB_VM_CONF, SI_ORDER_FIRST, vm_stats_init, NULL);
554