1 /* $NetBSD: bufcache.c,v 1.21 2008/01/24 17:32:58 ad Exp $ */ 2 3 /*- 4 * Copyright (c) 1999 The NetBSD Foundation, Inc. 5 * All rights reserved. 6 * 7 * This code is derived from software contributed to The NetBSD Foundation 8 * by Simon Burge. 9 * 10 * Redistribution and use in source and binary forms, with or without 11 * modification, are permitted provided that the following conditions 12 * are met: 13 * 1. Redistributions of source code must retain the above copyright 14 * notice, this list of conditions and the following disclaimer. 15 * 2. Redistributions in binary form must reproduce the above copyright 16 * notice, this list of conditions and the following disclaimer in the 17 * documentation and/or other materials provided with the distribution. 18 * 3. All advertising materials mentioning features or use of this software 19 * must display the following acknowledgement: 20 * This product includes software developed by the NetBSD 21 * Foundation, Inc. and its contributors. 22 * 4. Neither the name of The NetBSD Foundation nor the names of its 23 * contributors may be used to endorse or promote products derived 24 * from this software without specific prior written permission. 25 * 26 * THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS 27 * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED 28 * TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR 29 * PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS 30 * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR 31 * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF 32 * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS 33 * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN 34 * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) 35 * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE 36 * POSSIBILITY OF SUCH DAMAGE. 37 */ 38 39 #include <sys/cdefs.h> 40 #ifndef lint 41 __RCSID("$NetBSD: bufcache.c,v 1.21 2008/01/24 17:32:58 ad Exp $"); 42 #endif /* not lint */ 43 44 #include <sys/param.h> 45 #include <sys/buf.h> 46 #include <sys/mount.h> 47 #include <sys/sysctl.h> 48 #include <sys/vnode.h> 49 50 #include <uvm/uvm_extern.h> 51 52 #include <err.h> 53 #include <errno.h> 54 #include <inttypes.h> 55 #include <math.h> 56 #include <stdlib.h> 57 #include <string.h> 58 #include <unistd.h> 59 #include <stdbool.h> 60 61 #include <miscfs/specfs/specdev.h> 62 63 #include "systat.h" 64 #include "extern.h" 65 66 #define VCACHE_SIZE 50 67 #define PAGEINFO_ROWS 5 68 69 struct vcache { 70 int vc_age; 71 struct vnode *vc_addr; 72 struct vnode vc_node; 73 }; 74 75 struct ml_entry { 76 u_int ml_count; 77 u_long ml_size; 78 u_long ml_valid; 79 struct mount *ml_addr; 80 LIST_ENTRY(ml_entry) ml_entries; 81 struct mount ml_mount; 82 }; 83 84 static struct nlist namelist[] = { 85 #define X_BUFMEM 0 86 { .n_name = "_bufmem" }, 87 { .n_name = NULL }, 88 }; 89 90 static struct vcache vcache[VCACHE_SIZE]; 91 static LIST_HEAD(mount_list, ml_entry) mount_list; 92 93 static u_long bufmem; 94 static u_int nbuf, pgwidth, kbwidth; 95 static struct uvmexp_sysctl uvmexp; 96 97 static void vc_init(void); 98 static void ml_init(void); 99 static struct vnode *vc_lookup(struct vnode *); 100 static struct mount *ml_lookup(struct mount *, int, int); 101 static void fetchuvmexp(void); 102 103 104 WINDOW * 105 openbufcache(void) 106 { 107 108 return (subwin(stdscr, -1, 0, 5, 0)); 109 } 110 111 void 112 closebufcache(WINDOW *w) 113 { 114 115 if (w == NULL) 116 return; 117 wclear(w); 118 wrefresh(w); 119 delwin(w); 120 ml_init(); /* Clear out mount list */ 121 } 122 123 void 124 labelbufcache(void) 125 { 126 int i; 127 128 for (i = 0; i <= PAGEINFO_ROWS; i++) { 129 wmove(wnd, i, 0); 130 wclrtoeol(wnd); 131 } 132 mvwaddstr(wnd, PAGEINFO_ROWS + 1, 0, "File System Bufs used" 133 " % kB in use % Bufsize kB % Util %"); 134 wclrtoeol(wnd); 135 } 136 137 void 138 showbufcache(void) 139 { 140 int tbuf, i, lastrow; 141 double tvalid, tsize; 142 struct ml_entry *ml; 143 144 NREAD(X_BUFMEM, &bufmem, sizeof(bufmem)); 145 146 mvwprintw(wnd, 0, 0, 147 " %*d metadata buffers using %*ld kBytes of " 148 "memory (%2.0f%%).", 149 pgwidth, nbuf, kbwidth, bufmem / 1024, 150 ((bufmem * 100.0) + 0.5) / getpagesize() / uvmexp.npages); 151 wclrtoeol(wnd); 152 mvwprintw(wnd, 1, 0, 153 " %*" PRIu64 " pages for cached file data using %*" 154 PRIu64 " kBytes of memory (%2.0f%%).", 155 pgwidth, uvmexp.filepages, 156 kbwidth, uvmexp.filepages * getpagesize() / 1024, 157 (uvmexp.filepages * 100 + 0.5) / uvmexp.npages); 158 wclrtoeol(wnd); 159 mvwprintw(wnd, 2, 0, 160 " %*" PRIu64 " pages for executables using %*" 161 PRIu64 " kBytes of memory (%2.0f%%).", 162 pgwidth, uvmexp.execpages, 163 kbwidth, uvmexp.execpages * getpagesize() / 1024, 164 (uvmexp.execpages * 100 + 0.5) / uvmexp.npages); 165 wclrtoeol(wnd); 166 mvwprintw(wnd, 3, 0, 167 " %*" PRIu64 " pages for anon (non-file) data %*" 168 PRIu64 " kBytes of memory (%2.0f%%).", 169 pgwidth, uvmexp.anonpages, 170 kbwidth, uvmexp.anonpages * getpagesize() / 1024, 171 (uvmexp.anonpages * 100 + 0.5) / uvmexp.npages); 172 wclrtoeol(wnd); 173 mvwprintw(wnd, 4, 0, 174 " %*" PRIu64 " free pages %*" 175 PRIu64 " kBytes of memory (%2.0f%%).", 176 pgwidth, uvmexp.free, 177 kbwidth, uvmexp.free * getpagesize() / 1024, 178 (uvmexp.free * 100 + 0.5) / uvmexp.npages); 179 wclrtoeol(wnd); 180 181 if (nbuf == 0 || bufmem == 0) { 182 wclrtobot(wnd); 183 return; 184 } 185 186 tbuf = 0; 187 tvalid = tsize = 0; 188 lastrow = PAGEINFO_ROWS + 2; /* Leave room for header. */ 189 for (i = lastrow, ml = LIST_FIRST(&mount_list); ml != NULL; 190 i++, ml = LIST_NEXT(ml, ml_entries)) { 191 192 int cnt = ml->ml_count; 193 double v = ml->ml_valid; 194 double s = ml->ml_size; 195 196 /* Display in window if enough room. */ 197 if (i < getmaxy(wnd) - 2) { 198 mvwprintw(wnd, i, 0, "%-20.20s", ml->ml_addr == NULL ? 199 "NULL" : ml->ml_mount.mnt_stat.f_mntonname); 200 wprintw(wnd, 201 " %6d %3d %8ld %3.0f %8ld %3.0f %3.0f", 202 cnt, (100 * cnt) / nbuf, 203 (long)(v/1024), 100 * v / bufmem, 204 (long)(s/1024), 100 * s / bufmem, 205 100 * v / s); 206 wclrtoeol(wnd); 207 lastrow = i; 208 } 209 210 /* Update statistics. */ 211 tbuf += cnt; 212 tvalid += v; 213 tsize += s; 214 } 215 216 wclrtobot(wnd); 217 mvwprintw(wnd, lastrow + 2, 0, 218 "%-20s %6d %3d %8ld %3.0f %8ld %3.0f %3.0f", 219 "Total:", tbuf, (100 * tbuf) / nbuf, 220 (long)(tvalid/1024), 100 * tvalid / bufmem, 221 (long)(tsize/1024), 100 * tsize / bufmem, 222 tsize != 0 ? ((100 * tvalid) / tsize) : 0); 223 } 224 225 int 226 initbufcache(void) 227 { 228 if (namelist[0].n_type == 0) { 229 if (kvm_nlist(kd, namelist)) { 230 nlisterr(namelist); 231 return(0); 232 } 233 } 234 235 fetchuvmexp(); 236 pgwidth = (int)(floor(log10((double)uvmexp.npages)) + 1); 237 kbwidth = (int)(floor(log10(uvmexp.npages * getpagesize() / 1024.0)) + 238 1); 239 240 return(1); 241 } 242 243 static void 244 fetchuvmexp(void) 245 { 246 int mib[2]; 247 size_t size; 248 249 /* Re-read pages used for vnodes & executables */ 250 size = sizeof(uvmexp); 251 mib[0] = CTL_VM; 252 mib[1] = VM_UVMEXP2; 253 if (sysctl(mib, 2, &uvmexp, &size, NULL, 0) < 0) { 254 error("can't get uvmexp: %s\n", strerror(errno)); 255 memset(&uvmexp, 0, sizeof(uvmexp)); 256 } 257 } 258 259 void 260 fetchbufcache(void) 261 { 262 int count; 263 struct buf_sysctl *bp, *buffers; 264 struct vnode *vn; 265 struct mount *mt; 266 struct ml_entry *ml; 267 int mib[6]; 268 size_t size; 269 int extraslop = 0; 270 271 /* Re-read pages used for vnodes & executables */ 272 fetchuvmexp(); 273 274 /* Initialise vnode cache and mount list. */ 275 vc_init(); 276 ml_init(); 277 278 /* Get metadata buffers */ 279 size = 0; 280 buffers = NULL; 281 mib[0] = CTL_KERN; 282 mib[1] = KERN_BUF; 283 mib[2] = KERN_BUF_ALL; 284 mib[3] = KERN_BUF_ALL; 285 mib[4] = (int)sizeof(struct buf_sysctl); 286 mib[5] = INT_MAX; /* we want them all */ 287 again: 288 if (sysctl(mib, 6, NULL, &size, NULL, 0) < 0) { 289 error("can't get buffers size: %s\n", strerror(errno)); 290 return; 291 } 292 if (size == 0) 293 return; 294 295 size += extraslop * sizeof(struct buf_sysctl); 296 buffers = malloc(size); 297 if (buffers == NULL) { 298 error("can't allocate buffers: %s\n", strerror(errno)); 299 return; 300 } 301 if (sysctl(mib, 6, buffers, &size, NULL, 0) < 0) { 302 free(buffers); 303 if (extraslop == 0) { 304 extraslop = 100; 305 goto again; 306 } 307 error("can't get buffers: %s\n", strerror(errno)); 308 return; 309 } 310 311 nbuf = size / sizeof(struct buf_sysctl); 312 for (bp = buffers; bp < buffers + nbuf; bp++) { 313 if (UINT64TOPTR(bp->b_vp) != NULL) { 314 struct mount *mp; 315 vn = vc_lookup(UINT64TOPTR(bp->b_vp)); 316 if (vn == NULL) 317 break; 318 319 mp = vn->v_mount; 320 /* 321 * References to mounted-on vnodes should be 322 * counted towards the mounted filesystem. 323 */ 324 if (vn->v_type == VBLK && vn->v_specnode != NULL) { 325 specnode_t sn; 326 specdev_t sd; 327 if (!KREAD(vn->v_specnode, &sn, sizeof(sn))) 328 continue; 329 if (!KREAD(sn.sn_dev, &sd, sizeof(sd))) 330 continue; 331 if (sd.sd_mountpoint) 332 mp = sd.sd_mountpoint; 333 } 334 if (mp != NULL) 335 mt = ml_lookup(mp, 336 bp->b_bufsize, 337 bp->b_bcount); 338 } 339 } 340 341 /* simple sort - there's not that many entries */ 342 do { 343 if ((ml = LIST_FIRST(&mount_list)) == NULL || 344 LIST_NEXT(ml, ml_entries) == NULL) 345 break; 346 347 count = 0; 348 for (ml = LIST_FIRST(&mount_list); ml != NULL; 349 ml = LIST_NEXT(ml, ml_entries)) { 350 if (LIST_NEXT(ml, ml_entries) == NULL) 351 break; 352 if (ml->ml_count < LIST_NEXT(ml, ml_entries)->ml_count) { 353 ml = LIST_NEXT(ml, ml_entries); 354 LIST_REMOVE(ml, ml_entries); 355 LIST_INSERT_HEAD(&mount_list, ml, ml_entries); 356 count++; 357 } 358 } 359 } while (count != 0); 360 361 free(buffers); 362 } 363 364 static void 365 vc_init(void) 366 { 367 int i; 368 369 /* vc_addr == NULL for unused cache entry. */ 370 for (i = 0; i < VCACHE_SIZE; i++) 371 vcache[i].vc_addr = NULL; 372 } 373 374 static void 375 ml_init(void) 376 { 377 struct ml_entry *ml; 378 379 /* Throw out the current mount list and start again. */ 380 while ((ml = LIST_FIRST(&mount_list)) != NULL) { 381 LIST_REMOVE(ml, ml_entries); 382 free(ml); 383 } 384 } 385 386 387 static struct vnode * 388 vc_lookup(struct vnode *vaddr) 389 { 390 struct vnode *ret; 391 size_t i, oldest; 392 393 ret = NULL; 394 oldest = 0; 395 for (i = 0; i < VCACHE_SIZE; i++) { 396 if (vcache[i].vc_addr == NULL) 397 break; 398 vcache[i].vc_age++; 399 if (vcache[i].vc_age < vcache[oldest].vc_age) 400 oldest = i; 401 if (vcache[i].vc_addr == vaddr) { 402 vcache[i].vc_age = 0; 403 ret = &vcache[i].vc_node; 404 } 405 } 406 407 /* Find an entry in the cache? */ 408 if (ret != NULL) 409 return(ret); 410 411 /* Go past the end of the cache? */ 412 if (i >= VCACHE_SIZE) 413 i = oldest; 414 415 /* Read in new vnode and reset age counter. */ 416 if (KREAD(vaddr, &vcache[i].vc_node, sizeof(struct vnode)) == 0) 417 return NULL; 418 vcache[i].vc_addr = vaddr; 419 vcache[i].vc_age = 0; 420 421 return(&vcache[i].vc_node); 422 } 423 424 static struct mount * 425 ml_lookup(struct mount *maddr, int size, int valid) 426 { 427 struct ml_entry *ml; 428 429 for (ml = LIST_FIRST(&mount_list); ml != NULL; 430 ml = LIST_NEXT(ml, ml_entries)) 431 if (ml->ml_addr == maddr) { 432 ml->ml_count++; 433 ml->ml_size += size; 434 ml->ml_valid += valid; 435 if (ml->ml_addr == NULL) 436 return(NULL); 437 else 438 return(&ml->ml_mount); 439 } 440 441 if ((ml = malloc(sizeof(struct ml_entry))) == NULL) { 442 error("out of memory"); 443 die(0); 444 } 445 LIST_INSERT_HEAD(&mount_list, ml, ml_entries); 446 ml->ml_count = 1; 447 ml->ml_size = size; 448 ml->ml_valid = valid; 449 ml->ml_addr = maddr; 450 if (maddr == NULL) 451 return(NULL); 452 453 KREAD(maddr, &ml->ml_mount, sizeof(struct mount)); 454 return(&ml->ml_mount); 455 } 456