1 /* $NetBSD: umap_vnops.c,v 1.60 2017/06/04 08:02:26 hannken Exp $ */ 2 3 /* 4 * Copyright (c) 1992, 1993 5 * The Regents of the University of California. All rights reserved. 6 * 7 * This code is derived from software donated to Berkeley by 8 * the UCLA Ficus project. 9 * 10 * Redistribution and use in source and binary forms, with or without 11 * modification, are permitted provided that the following conditions 12 * are met: 13 * 1. Redistributions of source code must retain the above copyright 14 * notice, this list of conditions and the following disclaimer. 15 * 2. Redistributions in binary form must reproduce the above copyright 16 * notice, this list of conditions and the following disclaimer in the 17 * documentation and/or other materials provided with the distribution. 18 * 3. Neither the name of the University nor the names of its contributors 19 * may be used to endorse or promote products derived from this software 20 * without specific prior written permission. 21 * 22 * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND 23 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE 24 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE 25 * ARE DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE 26 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL 27 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS 28 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) 29 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT 30 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY 31 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF 32 * SUCH DAMAGE. 33 * 34 * @(#)umap_vnops.c 8.6 (Berkeley) 5/22/95 35 */ 36 37 /* 38 * Umap Layer 39 */ 40 41 #include <sys/cdefs.h> 42 __KERNEL_RCSID(0, "$NetBSD: umap_vnops.c,v 1.60 2017/06/04 08:02:26 hannken Exp $"); 43 44 #include <sys/param.h> 45 #include <sys/systm.h> 46 #include <sys/time.h> 47 #include <sys/vnode.h> 48 #include <sys/mount.h> 49 #include <sys/namei.h> 50 #include <sys/buf.h> 51 #include <sys/kauth.h> 52 53 #include <miscfs/umapfs/umap.h> 54 #include <miscfs/genfs/genfs.h> 55 #include <miscfs/genfs/layer_extern.h> 56 57 /* 58 * Note: If the LAYERFS_MBYPASSDEBUG flag is set, it is possible 59 * that the debug printing will bomb out, because kauth routines 60 * do not handle NOCRED or FSCRED like other credentials and end 61 * up dereferencing an inappropriate pointer. 62 * 63 * That should be fixed in kauth rather than here. 64 */ 65 66 int umap_lookup(void *); 67 int umap_getattr(void *); 68 int umap_print(void *); 69 int umap_rename(void *); 70 71 /* 72 * Global vfs data structures 73 */ 74 /* 75 * XXX - strategy, bwrite are hand coded currently. They should 76 * go away with a merged buffer/block cache. 77 * 78 */ 79 int (**umap_vnodeop_p)(void *); 80 const struct vnodeopv_entry_desc umap_vnodeop_entries[] = { 81 { &vop_default_desc, umap_bypass }, 82 83 { &vop_lookup_desc, umap_lookup }, 84 { &vop_getattr_desc, umap_getattr }, 85 { &vop_print_desc, umap_print }, 86 { &vop_rename_desc, umap_rename }, 87 88 { &vop_fsync_desc, layer_fsync }, 89 { &vop_inactive_desc, layer_inactive }, 90 { &vop_reclaim_desc, layer_reclaim }, 91 { &vop_open_desc, layer_open }, 92 { &vop_close_desc, layer_close }, 93 { &vop_setattr_desc, layer_setattr }, 94 { &vop_access_desc, layer_access }, 95 { &vop_remove_desc, layer_remove }, 96 { &vop_revoke_desc, layer_revoke }, 97 { &vop_rmdir_desc, layer_rmdir }, 98 99 { &vop_bmap_desc, layer_bmap }, 100 { &vop_getpages_desc, layer_getpages }, 101 { &vop_putpages_desc, layer_putpages }, 102 103 { NULL, NULL } 104 }; 105 const struct vnodeopv_desc umapfs_vnodeop_opv_desc = 106 { &umap_vnodeop_p, umap_vnodeop_entries }; 107 108 /* 109 * This is the 08-June-1999 bypass routine. 110 * See layer_vnops.c:layer_bypass for more details. 111 */ 112 int 113 umap_bypass(void *v) 114 { 115 struct vop_generic_args /* { 116 struct vnodeop_desc *a_desc; 117 <other random data follows, presumably> 118 } */ *ap = v; 119 int (**our_vnodeop_p)(void *); 120 kauth_cred_t *credpp = NULL, credp = 0; 121 kauth_cred_t savecredp = 0, savecompcredp = 0; 122 kauth_cred_t compcredp = 0; 123 struct vnode **this_vp_p; 124 int error; 125 struct vnode *old_vps[VDESC_MAX_VPS], *vp0; 126 struct vnode **vps_p[VDESC_MAX_VPS]; 127 struct vnode ***vppp; 128 struct vnodeop_desc *descp = ap->a_desc; 129 int reles, i, flags; 130 struct componentname **compnamepp = 0; 131 132 #ifdef DIAGNOSTIC 133 /* 134 * We require at least one vp. 135 */ 136 if (descp->vdesc_vp_offsets == NULL || 137 descp->vdesc_vp_offsets[0] == VDESC_NO_OFFSET) 138 panic("%s: no vp's in map.\n", __func__); 139 #endif 140 141 vps_p[0] = 142 VOPARG_OFFSETTO(struct vnode**, descp->vdesc_vp_offsets[0], ap); 143 vp0 = *vps_p[0]; 144 flags = MOUNTTOUMAPMOUNT(vp0->v_mount)->umapm_flags; 145 our_vnodeop_p = vp0->v_op; 146 147 if (flags & LAYERFS_MBYPASSDEBUG) 148 printf("%s: %s\n", __func__, descp->vdesc_name); 149 150 /* 151 * Map the vnodes going in. 152 * Later, we'll invoke the operation based on 153 * the first mapped vnode's operation vector. 154 */ 155 reles = descp->vdesc_flags; 156 for (i = 0; i < VDESC_MAX_VPS; reles >>= 1, i++) { 157 if (descp->vdesc_vp_offsets[i] == VDESC_NO_OFFSET) 158 break; /* bail out at end of list */ 159 vps_p[i] = this_vp_p = 160 VOPARG_OFFSETTO(struct vnode**, descp->vdesc_vp_offsets[i], 161 ap); 162 /* 163 * We're not guaranteed that any but the first vnode 164 * are of our type. Check for and don't map any 165 * that aren't. (We must always map first vp or vclean fails.) 166 */ 167 if (i && (*this_vp_p == NULL || 168 (*this_vp_p)->v_op != our_vnodeop_p)) { 169 old_vps[i] = NULL; 170 } else { 171 old_vps[i] = *this_vp_p; 172 *(vps_p[i]) = UMAPVPTOLOWERVP(*this_vp_p); 173 /* 174 * XXX - Several operations have the side effect 175 * of vrele'ing their vp's. We must account for 176 * that. (This should go away in the future.) 177 */ 178 if (reles & VDESC_VP0_WILLRELE) 179 vref(*this_vp_p); 180 } 181 182 } 183 184 /* 185 * Fix the credentials. (That's the purpose of this layer.) 186 */ 187 188 if (descp->vdesc_cred_offset != VDESC_NO_OFFSET) { 189 190 credpp = VOPARG_OFFSETTO(kauth_cred_t*, 191 descp->vdesc_cred_offset, ap); 192 193 /* Save old values */ 194 195 savecredp = *credpp; 196 if (savecredp != NOCRED && savecredp != FSCRED) 197 *credpp = kauth_cred_dup(savecredp); 198 credp = *credpp; 199 200 if ((flags & LAYERFS_MBYPASSDEBUG) && 201 kauth_cred_geteuid(credp) != 0) 202 printf("umap_bypass: user was %d, group %d\n", 203 kauth_cred_geteuid(credp), kauth_cred_getegid(credp)); 204 205 /* Map all ids in the credential structure. */ 206 207 umap_mapids(vp0->v_mount, credp); 208 209 if ((flags & LAYERFS_MBYPASSDEBUG) && 210 kauth_cred_geteuid(credp) != 0) 211 printf("umap_bypass: user now %d, group %d\n", 212 kauth_cred_geteuid(credp), kauth_cred_getegid(credp)); 213 } 214 215 /* BSD often keeps a credential in the componentname structure 216 * for speed. If there is one, it better get mapped, too. 217 */ 218 219 if (descp->vdesc_componentname_offset != VDESC_NO_OFFSET) { 220 221 compnamepp = VOPARG_OFFSETTO(struct componentname**, 222 descp->vdesc_componentname_offset, ap); 223 224 savecompcredp = (*compnamepp)->cn_cred; 225 if (savecompcredp != NOCRED && savecompcredp != FSCRED) 226 (*compnamepp)->cn_cred = kauth_cred_dup(savecompcredp); 227 compcredp = (*compnamepp)->cn_cred; 228 229 if ((flags & LAYERFS_MBYPASSDEBUG) && 230 kauth_cred_geteuid(compcredp) != 0) 231 printf("umap_bypass: component credit user was %d, group %d\n", 232 kauth_cred_geteuid(compcredp), kauth_cred_getegid(compcredp)); 233 234 /* Map all ids in the credential structure. */ 235 236 umap_mapids(vp0->v_mount, compcredp); 237 238 if ((flags & LAYERFS_MBYPASSDEBUG) && 239 kauth_cred_geteuid(compcredp) != 0) 240 printf("umap_bypass: component credit user now %d, group %d\n", 241 kauth_cred_geteuid(compcredp), kauth_cred_getegid(compcredp)); 242 } 243 244 /* 245 * Call the operation on the lower layer 246 * with the modified argument structure. 247 */ 248 error = VCALL(*vps_p[0], descp->vdesc_offset, ap); 249 250 /* 251 * Maintain the illusion of call-by-value 252 * by restoring vnodes in the argument structure 253 * to their original value. 254 */ 255 reles = descp->vdesc_flags; 256 for (i = 0; i < VDESC_MAX_VPS; reles >>= 1, i++) { 257 if (descp->vdesc_vp_offsets[i] == VDESC_NO_OFFSET) 258 break; /* bail out at end of list */ 259 if (old_vps[i]) { 260 *(vps_p[i]) = old_vps[i]; 261 if (reles & VDESC_VP0_WILLRELE) 262 vrele(*(vps_p[i])); 263 } 264 } 265 266 /* 267 * Map the possible out-going vpp 268 * (Assumes that the lower layer always returns 269 * a VREF'ed vpp unless it gets an error.) 270 */ 271 if (descp->vdesc_vpp_offset != VDESC_NO_OFFSET && !error) { 272 vppp = VOPARG_OFFSETTO(struct vnode***, 273 descp->vdesc_vpp_offset, ap); 274 /* 275 * Only vop_lookup, vop_create, vop_makedir, vop_mknod 276 * and vop_symlink return vpp's. vop_lookup doesn't call bypass 277 * as a lookup on "." would generate a locking error. 278 * So all the calls which get us here have a unlocked vpp. :-) 279 */ 280 error = layer_node_create(old_vps[0]->v_mount, **vppp, *vppp); 281 if (error) { 282 vrele(**vppp); 283 **vppp = NULL; 284 } 285 } 286 287 /* 288 * Free duplicate cred structure and restore old one. 289 */ 290 if (descp->vdesc_cred_offset != VDESC_NO_OFFSET) { 291 if ((flags & LAYERFS_MBYPASSDEBUG) && credp && 292 kauth_cred_geteuid(credp) != 0) 293 printf("umap_bypass: returning-user was %d\n", 294 kauth_cred_geteuid(credp)); 295 296 if (savecredp != NOCRED && savecredp != FSCRED && credpp) { 297 kauth_cred_free(credp); 298 *credpp = savecredp; 299 if ((flags & LAYERFS_MBYPASSDEBUG) && credpp && 300 kauth_cred_geteuid(*credpp) != 0) 301 printf("umap_bypass: returning-user now %d\n\n", 302 kauth_cred_geteuid(savecredp)); 303 } 304 } 305 306 if (descp->vdesc_componentname_offset != VDESC_NO_OFFSET) { 307 if ((flags & LAYERFS_MBYPASSDEBUG) && compcredp && 308 kauth_cred_geteuid(compcredp) != 0) 309 printf("umap_bypass: returning-component-user was %d\n", 310 kauth_cred_geteuid(compcredp)); 311 312 if (savecompcredp != NOCRED && savecompcredp != FSCRED) { 313 kauth_cred_free(compcredp); 314 (*compnamepp)->cn_cred = savecompcredp; 315 if ((flags & LAYERFS_MBYPASSDEBUG) && savecompcredp && 316 kauth_cred_geteuid(savecompcredp) != 0) 317 printf("umap_bypass: returning-component-user now %d\n", 318 kauth_cred_geteuid(savecompcredp)); 319 } 320 } 321 322 return (error); 323 } 324 325 /* 326 * This is based on the 08-June-1999 bypass routine. 327 * See layer_vnops.c:layer_bypass for more details. 328 */ 329 int 330 umap_lookup(void *v) 331 { 332 struct vop_lookup_v2_args /* { 333 struct vnodeop_desc *a_desc; 334 struct vnode * a_dvp; 335 struct vnode ** a_vpp; 336 struct componentname * a_cnp; 337 } */ *ap = v; 338 struct componentname *cnp = ap->a_cnp; 339 kauth_cred_t savecompcredp = NULL; 340 kauth_cred_t compcredp = NULL; 341 struct vnode *dvp, *vp, *ldvp; 342 struct mount *mp; 343 int error; 344 int flags, cnf = cnp->cn_flags; 345 346 dvp = ap->a_dvp; 347 mp = dvp->v_mount; 348 349 if ((cnf & ISLASTCN) && (dvp->v_mount->mnt_flag & MNT_RDONLY) && 350 (cnp->cn_nameiop == DELETE || cnp->cn_nameiop == RENAME)) 351 return (EROFS); 352 353 flags = MOUNTTOUMAPMOUNT(mp)->umapm_flags; 354 ldvp = UMAPVPTOLOWERVP(dvp); 355 356 if (flags & LAYERFS_MBYPASSDEBUG) 357 printf("umap_lookup\n"); 358 359 /* 360 * Fix the credentials. (That's the purpose of this layer.) 361 * 362 * BSD often keeps a credential in the componentname structure 363 * for speed. If there is one, it better get mapped, too. 364 */ 365 366 if ((savecompcredp = cnp->cn_cred)) { 367 compcredp = kauth_cred_dup(savecompcredp); 368 cnp->cn_cred = compcredp; 369 370 if ((flags & LAYERFS_MBYPASSDEBUG) && 371 kauth_cred_geteuid(compcredp) != 0) 372 printf("umap_lookup: component credit user was %d, group %d\n", 373 kauth_cred_geteuid(compcredp), kauth_cred_getegid(compcredp)); 374 375 /* Map all ids in the credential structure. */ 376 umap_mapids(mp, compcredp); 377 } 378 379 if ((flags & LAYERFS_MBYPASSDEBUG) && compcredp && 380 kauth_cred_geteuid(compcredp) != 0) 381 printf("umap_lookup: component credit user now %d, group %d\n", 382 kauth_cred_geteuid(compcredp), kauth_cred_getegid(compcredp)); 383 384 ap->a_dvp = ldvp; 385 error = VCALL(ldvp, ap->a_desc->vdesc_offset, ap); 386 vp = *ap->a_vpp; 387 *ap->a_vpp = NULL; 388 389 if (error == EJUSTRETURN && (cnf & ISLASTCN) && 390 (dvp->v_mount->mnt_flag & MNT_RDONLY) && 391 (cnp->cn_nameiop == CREATE || cnp->cn_nameiop == RENAME)) 392 error = EROFS; 393 394 /* Do locking fixup as appropriate. See layer_lookup() for info */ 395 if (ldvp == vp) { 396 *ap->a_vpp = dvp; 397 vref(dvp); 398 vrele(vp); 399 } else if (vp != NULL) { 400 error = layer_node_create(mp, vp, ap->a_vpp); 401 if (error) { 402 vrele(vp); 403 } 404 } 405 406 /* 407 * Free duplicate cred structure and restore old one. 408 */ 409 if ((flags & LAYERFS_MBYPASSDEBUG) && compcredp && 410 kauth_cred_geteuid(compcredp) != 0) 411 printf("umap_lookup: returning-component-user was %d\n", 412 kauth_cred_geteuid(compcredp)); 413 414 if (savecompcredp != NOCRED && savecompcredp != FSCRED) { 415 if (compcredp) 416 kauth_cred_free(compcredp); 417 cnp->cn_cred = savecompcredp; 418 if ((flags & LAYERFS_MBYPASSDEBUG) && savecompcredp && 419 kauth_cred_geteuid(savecompcredp) != 0) 420 printf("umap_lookup: returning-component-user now %d\n", 421 kauth_cred_geteuid(savecompcredp)); 422 } 423 424 return (error); 425 } 426 427 /* 428 * We handle getattr to change the fsid. 429 */ 430 int 431 umap_getattr(void *v) 432 { 433 struct vop_getattr_args /* { 434 struct vnode *a_vp; 435 struct vattr *a_vap; 436 kauth_cred_t a_cred; 437 struct lwp *a_l; 438 } */ *ap = v; 439 uid_t uid; 440 gid_t gid; 441 int error, tmpid, nentries, gnentries, flags; 442 u_long (*mapdata)[2]; 443 u_long (*gmapdata)[2]; 444 struct vnode **vp1p; 445 const struct vnodeop_desc *descp = ap->a_desc; 446 447 if ((error = umap_bypass(ap)) != 0) 448 return (error); 449 /* Requires that arguments be restored. */ 450 ap->a_vap->va_fsid = ap->a_vp->v_mount->mnt_stat.f_fsidx.__fsid_val[0]; 451 452 flags = MOUNTTOUMAPMOUNT(ap->a_vp->v_mount)->umapm_flags; 453 /* 454 * Umap needs to map the uid and gid returned by a stat 455 * into the proper values for this site. This involves 456 * finding the returned uid in the mapping information, 457 * translating it into the uid on the other end, 458 * and filling in the proper field in the vattr 459 * structure pointed to by ap->a_vap. The group 460 * is easier, since currently all groups will be 461 * translate to the NULLGROUP. 462 */ 463 464 /* Find entry in map */ 465 466 uid = ap->a_vap->va_uid; 467 gid = ap->a_vap->va_gid; 468 if ((flags & LAYERFS_MBYPASSDEBUG)) 469 printf("umap_getattr: mapped uid = %d, mapped gid = %d\n", uid, 470 gid); 471 472 vp1p = VOPARG_OFFSETTO(struct vnode**, descp->vdesc_vp_offsets[0], ap); 473 nentries = MOUNTTOUMAPMOUNT((*vp1p)->v_mount)->info_nentries; 474 mapdata = (MOUNTTOUMAPMOUNT((*vp1p)->v_mount)->info_mapdata); 475 gnentries = MOUNTTOUMAPMOUNT((*vp1p)->v_mount)->info_gnentries; 476 gmapdata = (MOUNTTOUMAPMOUNT((*vp1p)->v_mount)->info_gmapdata); 477 478 /* Reverse map the uid for the vnode. Since it's a reverse 479 map, we can't use umap_mapids() to do it. */ 480 481 tmpid = umap_reverse_findid(uid, mapdata, nentries); 482 483 if (tmpid != -1) { 484 ap->a_vap->va_uid = (uid_t) tmpid; 485 if ((flags & LAYERFS_MBYPASSDEBUG)) 486 printf("umap_getattr: original uid = %d\n", uid); 487 } else 488 ap->a_vap->va_uid = (uid_t) NOBODY; 489 490 /* Reverse map the gid for the vnode. */ 491 492 tmpid = umap_reverse_findid(gid, gmapdata, gnentries); 493 494 if (tmpid != -1) { 495 ap->a_vap->va_gid = (gid_t) tmpid; 496 if ((flags & LAYERFS_MBYPASSDEBUG)) 497 printf("umap_getattr: original gid = %d\n", gid); 498 } else 499 ap->a_vap->va_gid = (gid_t) NULLGROUP; 500 501 return (0); 502 } 503 504 int 505 umap_print(void *v) 506 { 507 struct vop_print_args /* { 508 struct vnode *a_vp; 509 } */ *ap = v; 510 struct vnode *vp = ap->a_vp; 511 printf("\ttag VT_UMAPFS, vp=%p, lowervp=%p\n", vp, 512 UMAPVPTOLOWERVP(vp)); 513 return (0); 514 } 515 516 int 517 umap_rename(void *v) 518 { 519 struct vop_rename_args /* { 520 struct vnode *a_fdvp; 521 struct vnode *a_fvp; 522 struct componentname *a_fcnp; 523 struct vnode *a_tdvp; 524 struct vnode *a_tvp; 525 struct componentname *a_tcnp; 526 } */ *ap = v; 527 int error, flags; 528 struct componentname *compnamep; 529 kauth_cred_t compcredp, savecompcredp; 530 struct vnode *vp; 531 struct vnode *tvp; 532 533 /* 534 * Rename is irregular, having two componentname structures. 535 * We need to map the cre in the second structure, 536 * and then bypass takes care of the rest. 537 */ 538 539 vp = ap->a_fdvp; 540 flags = MOUNTTOUMAPMOUNT(vp->v_mount)->umapm_flags; 541 compnamep = ap->a_tcnp; 542 compcredp = compnamep->cn_cred; 543 544 savecompcredp = compcredp; 545 compcredp = compnamep->cn_cred = kauth_cred_dup(savecompcredp); 546 547 if ((flags & LAYERFS_MBYPASSDEBUG) && 548 kauth_cred_geteuid(compcredp) != 0) 549 printf("umap_rename: rename component credit user was %d, group %d\n", 550 kauth_cred_geteuid(compcredp), kauth_cred_getegid(compcredp)); 551 552 /* Map all ids in the credential structure. */ 553 554 umap_mapids(vp->v_mount, compcredp); 555 556 if ((flags & LAYERFS_MBYPASSDEBUG) && 557 kauth_cred_geteuid(compcredp) != 0) 558 printf("umap_rename: rename component credit user now %d, group %d\n", 559 kauth_cred_geteuid(compcredp), kauth_cred_getegid(compcredp)); 560 561 tvp = ap->a_tvp; 562 if (tvp) { 563 if (tvp->v_mount != vp->v_mount) 564 tvp = NULL; 565 else 566 vref(tvp); 567 } 568 error = umap_bypass(ap); 569 if (tvp) { 570 if (error == 0) 571 VTOLAYER(tvp)->layer_flags |= LAYERFS_REMOVED; 572 vrele(tvp); 573 } 574 575 /* Restore the additional mapped componentname cred structure. */ 576 577 kauth_cred_free(compcredp); 578 compnamep->cn_cred = savecompcredp; 579 580 return error; 581 } 582