Lines Matching defs:vp

295 	struct vnode *vp;
300 vp = v->a_vp;
302 node = VP_TO_TMPFS_NODE(vp);
318 KASSERT(vp->v_type != VREG || (node->tn_reg.tn_aobj->flags &
320 vnode_create_vobject(vp, node->tn_size, v->a_td);
325 if (error == 0 && fp != NULL && vp->v_type == VREG) {
336 struct vnode *vp = v->a_vp;
339 tmpfs_update(vp);
364 struct vnode *vp;
369 vp = v->a_vp;
370 node = VP_TO_TMPFS_NODE_SMR(vp);
384 tmpfs_access_locked(struct vnode *vp, struct tmpfs_node *node,
389 ASSERT_VOP_LOCKED(vp,
396 return (vaccess(vp->v_type, node->tn_mode, node->tn_uid, node->tn_gid,
403 struct vnode *vp = v->a_vp;
405 struct tmpfs_node *node = VP_TO_TMPFS_NODE(vp);
416 switch (vp->v_type) {
423 (vp->v_mount->mnt_flag & MNT_RDONLY) != 0)
440 return (tmpfs_access_locked(vp, node, accmode, cred));
446 struct vnode *vp = v->a_vp;
451 node = VP_TO_TMPFS_NODE(vp);
453 tmpfs_update_getattr(vp);
459 sb->st_dev = vp->v_mount->mnt_stat.f_fsid.val[0];
461 sb->st_mode = node->tn_mode | VTTOIF(vp->v_type);
465 sb->st_rdev = (vp->v_type == VBLK || vp->v_type == VCHR) ?
480 if (vp->v_type == VREG) {
501 struct vnode *vp = v->a_vp;
505 node = VP_TO_TMPFS_NODE(vp);
507 tmpfs_update_getattr(vp);
509 vap->va_type = vp->v_type;
514 vap->va_fsid = vp->v_mount->mnt_stat.f_fsid.val[0];
524 vap->va_rdev = (vp->v_type == VBLK || vp->v_type == VCHR) ?
526 if (vp->v_type == VREG) {
547 struct vnode *vp = v->a_vp;
554 ASSERT_VOP_IN_SEQC(vp);
570 error = tmpfs_chflags(vp, vap->va_flags, cred, td);
573 error = tmpfs_chsize(vp, vap->va_size, cred, td);
576 error = tmpfs_chown(vp, vap->va_uid, vap->va_gid, cred, td);
579 error = tmpfs_chmod(vp, vap->va_mode, cred, td);
587 error = tmpfs_chtimes(vp, vap, cred, td);
594 tmpfs_update(vp);
602 struct vnode *vp;
606 vp = v->a_vp;
607 if (vp->v_type != VREG)
612 node = VP_TO_TMPFS_NODE(vp);
613 tmpfs_set_accessed(VFS_TO_TMPFS(vp->v_mount), node);
620 struct vnode *vp;
626 vp = v->a_vp;
627 VNPASS((vn_irflag_read(vp) & VIRF_PGREAD) != 0, vp);
635 node = VP_TO_TMPFS_NODE_SMR(vp);
647 if (!VN_IS_DOOMED(vp)) {
663 struct vnode *vp;
671 vp = v->a_vp;
675 node = VP_TO_TMPFS_NODE(vp);
678 if (uio->uio_offset < 0 || vp->v_type != VREG)
684 error = vn_rlimit_fsizex(vp, uio, VFS_TO_TMPFS(vp->v_mount)->
692 error = tmpfs_reg_resize(vp, uio->uio_offset + uio->uio_resid,
704 vn_seqc_write_begin(vp);
706 vn_seqc_write_end(vp);
710 (void)tmpfs_reg_resize(vp, oldsize, TRUE);
729 struct vnode *vp = v->a_vp;
731 tmpfs_check_mtime(vp);
732 tmpfs_update(vp);
741 struct vnode *vp = v->a_vp;
749 if (vp->v_type == VDIR) {
755 node = VP_TO_TMPFS_NODE(vp);
756 tmp = VFS_TO_TMPFS(vp->v_mount);
790 struct vnode *vp = v->a_vp;
797 MPASS(dvp != vp); /* XXX When can this be false? */
798 node = VP_TO_TMPFS_NODE(vp);
815 error = tmpfs_alloc_dirent(VFS_TO_TMPFS(vp->v_mount), node,
825 /* vp link count has changed, so update node times. */
827 tmpfs_update(vp);
1324 struct vnode *vp = v->a_vp;
1335 node = VP_TO_TMPFS_DIR(vp);
1359 /* Get the directory entry associated with node (vp). This was
1377 tmpfs_dir_clear_whiteouts(vp);
1398 cache_vop_rmdir(dvp, vp);
1439 struct vnode *vp;
1448 vp = va->a_vp;
1455 if (vp->v_type != VDIR)
1459 node = VP_TO_TMPFS_DIR(vp);
1460 tm = VFS_TO_TMPFS(vp->v_mount);
1499 struct vnode *vp = v->a_vp;
1506 MPASS(vp->v_type == VLNK);
1508 node = VP_TO_TMPFS_NODE(vp);
1512 tmpfs_set_accessed(VFS_TO_TMPFS(vp->v_mount), node);
1526 struct vnode *vp;
1530 vp = v->a_vp;
1531 node = VP_TO_TMPFS_NODE_SMR(vp);
1546 struct vnode *vp;
1549 vp = v->a_vp;
1550 node = VP_TO_TMPFS_NODE(vp);
1552 vrecycle(vp);
1554 tmpfs_check_mtime(vp);
1561 struct vnode *vp;
1565 vp = ap->a_vp;
1566 node = VP_TO_TMPFS_NODE(vp);
1569 if (vp->v_type == VREG) {
1570 obj = vp->v_object;
1582 struct vnode *vp;
1587 vp = v->a_vp;
1588 node = VP_TO_TMPFS_NODE(vp);
1589 tmp = VFS_TO_TMPFS(vp->v_mount);
1591 if (vp->v_type == VREG)
1592 tmpfs_destroy_vobject(vp, node->tn_reg.tn_aobj);
1593 vp->v_object = NULL;
1597 tmpfs_free_vp(vp);
1616 MPASS(vp->v_data == NULL);
1623 struct vnode *vp = v->a_vp;
1627 node = VP_TO_TMPFS_NODE(vp);
1635 if (vp->v_type == VFIFO)
1636 fifo_printinfo(vp);
1646 struct vnode *vp = v->a_vp;
1668 if (vp->v_type == VDIR || vp->v_type == VFIFO)
1764 tmpfs_vptocnp_fill(struct vnode *vp, struct tmpfs_node *tn,
1770 error = vn_vget_ino_gen(vp, tmpfs_vn_get_ino_alloc, tnp, LK_SHARED,
1786 if (vp != *dvp)
1789 if (vp != *dvp)
1792 vrele(vp);
1800 struct vnode *vp, **dvp;
1808 vp = ap->a_vp;
1813 tm = VFS_TO_TMPFS(vp->v_mount);
1814 tn = VP_TO_TMPFS_NODE(vp);
1820 error = tmpfs_vptocnp_fill(vp, tn, tn->tn_dir.tn_parent, buf,
1847 error = tmpfs_vptocnp_fill(vp, tn, tnp, buf, buflen,
1853 if (VN_IS_DOOMED(vp)) {
1907 struct vnode *vp = ap->a_vp;
1915 node = VP_TO_TMPFS_NODE(vp);
1916 tmp = VFS_TO_TMPFS(vp->v_mount);
1948 struct vnode *vp = ap->a_vp;
1954 node = VP_TO_TMPFS_NODE(vp);
1986 struct vnode *vp = ap->a_vp;
1991 node = VP_TO_TMPFS_NODE(vp);
2022 struct vnode *vp = ap->a_vp;
2032 node = VP_TO_TMPFS_NODE(vp);
2033 tmp = VFS_TO_TMPFS(vp->v_mount);
2115 tmpfs_seek_datahole(struct vnode *vp, off_t *off, bool seekdata)
2122 if (vp->v_type != VREG)
2124 tn = VP_TO_TMPFS_NODE(vp);
2147 struct vnode *vp = ap->a_vp;
2153 error = vn_lock(vp, LK_SHARED);
2158 error = tmpfs_seek_datahole(vp, (off_t *)ap->a_data,
2160 VOP_UNLOCK(vp);