/* $NetBSD: chfs_vnops.c,v 1.33 2017/05/26 14:21:02 riastradh Exp $ */ /*- * Copyright (c) 2010 Department of Software Engineering, * University of Szeged, Hungary * Copyright (C) 2010 Tamas Toth * Copyright (C) 2010 Adam Hoka * All rights reserved. * * This code is derived from software contributed to The NetBSD Foundation * by the Department of Software Engineering, University of Szeged, Hungary * * Redistribution and use in source and binary forms, with or without * modification, are permitted provided that the following conditions * are met: * 1. Redistributions of source code must retain the above copyright * notice, this list of conditions and the following disclaimer. * 2. Redistributions in binary form must reproduce the above copyright * notice, this list of conditions and the following disclaimer in the * documentation and/or other materials provided with the distribution. * * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT, * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, * BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED * AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, * OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF * SUCH DAMAGE. */ #include #include #include #include #include #include #include #include #include #include #include #include #include "chfs.h" #define READ_S "chfs_read" int chfs_lookup(void *v) { struct vnode *dvp = ((struct vop_lookup_v2_args *) v)->a_dvp; struct vnode **vpp = ((struct vop_lookup_v2_args *) v)->a_vpp; struct componentname *cnp = ((struct vop_lookup_v2_args *) v)->a_cnp; int error; struct chfs_inode* ip; struct ufsmount* ump; struct chfs_mount* chmp; struct chfs_vnode_cache* chvc; struct chfs_dirent* fd; dbg("lookup(): %s\n", cnp->cn_nameptr); KASSERT(VOP_ISLOCKED(dvp)); *vpp = NULL; /* Check accessibility of requested node as a first step. */ error = VOP_ACCESS(dvp, VEXEC, cnp->cn_cred); if (error != 0) { goto out; } /* If requesting the last path component on a read-only file system * with a write operation, deny it. */ if ((cnp->cn_flags & ISLASTCN) && (dvp->v_mount->mnt_flag & MNT_RDONLY) && (cnp->cn_nameiop == DELETE || cnp->cn_nameiop == RENAME)) { error = EROFS; goto out; } /* Avoid doing a linear scan of the directory if the requested * directory/name couple is already in the cache. */ if (cache_lookup(dvp, cnp->cn_nameptr, cnp->cn_namelen, cnp->cn_nameiop, cnp->cn_flags, NULL, vpp)) { return (*vpp == NULLVP ? ENOENT : 0); } ip = VTOI(dvp); ump = VFSTOUFS(dvp->v_mount); chmp = ump->um_chfs; if (ip->ino == 0) { ip->ino = ++chmp->chm_max_vno; } mutex_enter(&chmp->chm_lock_vnocache); chvc = chfs_vnode_cache_get(chmp, ip->ino); mutex_exit(&chmp->chm_lock_vnocache); /* We cannot be requesting the parent directory of the root node. */ KASSERT(IMPLIES(ip->ch_type == CHT_DIR && chvc->pvno == chvc->vno, !(cnp->cn_flags & ISDOTDOT))); if (cnp->cn_flags & ISDOTDOT) { VOP_UNLOCK(dvp); error = VFS_VGET(dvp->v_mount, ip->chvc->pvno, vpp); vn_lock(dvp, LK_EXCLUSIVE | LK_RETRY); } else if (cnp->cn_namelen == 1 && cnp->cn_nameptr[0] == '.') { vref(dvp); *vpp = dvp; error = 0; } else { fd = chfs_dir_lookup(ip, cnp); if (fd == NULL) { dbg("fd null\n"); /* The entry was not found in the directory. * This is OK if we are creating or renaming an * entry and are working on the last component of * the path name. */ if ((cnp->cn_flags & ISLASTCN) && (cnp->cn_nameiop == CREATE || cnp->cn_nameiop == RENAME)) { error = VOP_ACCESS(dvp, VWRITE, cnp->cn_cred); if (error) { dbg("after the entry was not found in dir\n"); goto out; } dbg("return EJUSTRETURN\n"); error = EJUSTRETURN; } else { error = ENOENT; } } else { /* If we are not at the last path component and * found a non-directory or non-link entry (which * may itself be pointing to a directory), raise * an error. */ if ((fd->type != CHT_DIR && fd->type != CHT_LNK) && !(cnp->cn_flags & ISLASTCN)) { error = ENOTDIR; goto out; } dbg("vno@allocating new vnode: %llu\n", (unsigned long long)fd->vno); error = VFS_VGET(dvp->v_mount, fd->vno, vpp); } } /* Store the result of this lookup in the cache. Avoid this if the * request was for creation, as it does not improve timings on * emprical tests. */ if (cnp->cn_nameiop != CREATE && (cnp->cn_flags & ISDOTDOT) == 0) { cache_enter(dvp, *vpp, cnp->cn_nameptr, cnp->cn_namelen, cnp->cn_flags); } out: /* If there were no errors, *vpp cannot be NULL. */ KASSERT(IFF(error == 0, *vpp != NULL)); KASSERT(VOP_ISLOCKED(dvp)); if (error) return error; if (*vpp != dvp) VOP_UNLOCK(*vpp); return 0; } /* --------------------------------------------------------------------- */ int chfs_create(void *v) { struct vop_create_v3_args /* { struct vnode *a_dvp; struct vnode **a_vpp; struct componentname *a_cnp; struct vattr *a_vap; } */*ap = v; int error, mode; dbg("create()\n"); mode = MAKEIMODE(ap->a_vap->va_type, ap->a_vap->va_mode); if ((mode & IFMT) == 0) { if (ap->a_vap->va_type == VREG) mode |= IFREG; if (ap->a_vap->va_type == VSOCK) mode |= IFSOCK; } error = chfs_makeinode(mode, ap->a_dvp, ap->a_vpp, ap->a_cnp, ap->a_vap->va_type); if (error) { dbg("error: %d\n", error); return error; } VN_KNOTE(ap->a_dvp, NOTE_WRITE); return 0; } /* --------------------------------------------------------------------- */ int chfs_mknod(void *v) { struct vnode *dvp = ((struct vop_mknod_v3_args *) v)->a_dvp; struct vnode **vpp = ((struct vop_mknod_v3_args *) v)->a_vpp; struct componentname *cnp = ((struct vop_mknod_v3_args *) v)->a_cnp; struct vattr *vap = ((struct vop_mknod_v3_args *) v)->a_vap; int mode, err = 0; struct chfs_inode *ip; struct vnode *vp; struct ufsmount *ump; struct chfs_mount *chmp; struct chfs_full_dnode *fd; struct buf *bp; int len; dbg("mknod()\n"); ump = VFSTOUFS(dvp->v_mount); chmp = ump->um_chfs; /* Check type of node. */ if (vap->va_type != VBLK && vap->va_type != VCHR && vap->va_type != VFIFO) return EINVAL; vp = *vpp; mode = MAKEIMODE(vap->va_type, vap->va_mode); if ((mode & IFMT) == 0) { switch (vap->va_type) { case VBLK: mode |= IFBLK; break; case VCHR: mode |= IFCHR; break; case VFIFO: mode |= IFIFO; break; default: break; } } /* Create a new node. */ err = chfs_makeinode(mode, dvp, &vp, cnp, vap->va_type); ip = VTOI(vp); if (vap->va_rdev != VNOVAL) ip->rdev = vap->va_rdev; if (vap->va_type == VFIFO) vp->v_op = chfs_fifoop_p; else { vp->v_op = chfs_specop_p; spec_node_init(vp, ip->rdev); } if (err) return err; /* Device is written out as a data node. */ len = sizeof(dev_t); chfs_set_vnode_size(vp, len); bp = getiobuf(vp, true); bp->b_bufsize = bp->b_resid = len; bp->b_data = kmem_alloc(len, KM_SLEEP); memcpy(bp->b_data, &ip->rdev, len); bp->b_blkno = 0; fd = chfs_alloc_full_dnode(); mutex_enter(&chmp->chm_lock_mountfields); err = chfs_write_flash_dnode(chmp, vp, bp, fd); if (err) { mutex_exit(&chmp->chm_lock_mountfields); kmem_free(bp->b_data, len); return err; } /* Add data node to the inode. */ err = chfs_add_full_dnode_to_inode(chmp, ip, fd); if (err) { mutex_exit(&chmp->chm_lock_mountfields); kmem_free(bp->b_data, len); return err; } mutex_exit(&chmp->chm_lock_mountfields); *vpp = vp; kmem_free(bp->b_data, len); putiobuf(bp); return 0; } /* --------------------------------------------------------------------- */ int chfs_open(void *v) { struct vnode *vp = ((struct vop_open_args *) v)->a_vp; int mode = ((struct vop_open_args *) v)->a_mode; dbg("open()\n"); int error; struct chfs_inode *ip; KASSERT(VOP_ISLOCKED(vp)); ip = VTOI(vp); KASSERT(vp->v_size == ip->size); if (ip->chvc->nlink < 1) { error = ENOENT; goto out; } /* If the file is marked append-only, deny write requests. */ if (ip->flags & APPEND && (mode & (FWRITE | O_APPEND)) == FWRITE) error = EPERM; else error = 0; out: KASSERT(VOP_ISLOCKED(vp)); return error; } /* --------------------------------------------------------------------- */ int chfs_close(void *v) { struct vnode *vp = ((struct vop_close_args *) v)->a_vp; dbg("close()\n"); struct chfs_inode *ip; KASSERT(VOP_ISLOCKED(vp)); ip = VTOI(vp); if (ip->chvc->nlink > 0) { chfs_update(vp, NULL, NULL, UPDATE_CLOSE); } return 0; } /* --------------------------------------------------------------------- */ int chfs_access(void *v) { struct vnode *vp = ((struct vop_access_args *) v)->a_vp; int mode = ((struct vop_access_args *) v)->a_mode; kauth_cred_t cred = ((struct vop_access_args *) v)->a_cred; dbg("access()\n"); struct chfs_inode *ip = VTOI(vp); if (mode & VWRITE) { switch (vp->v_type) { case VLNK: case VDIR: case VREG: if (vp->v_mount->mnt_flag & MNT_RDONLY) return (EROFS); break; case VBLK: case VCHR: case VSOCK: case VFIFO: break; default: break; } } if (mode & VWRITE && ip->flags & IMMUTABLE) return (EPERM); return kauth_authorize_vnode(cred, KAUTH_ACCESS_ACTION(mode, vp->v_type, ip->mode & ALLPERMS), vp, NULL, genfs_can_access(vp->v_type, ip->mode & ALLPERMS, ip->uid, ip->gid, mode, cred)); } /* --------------------------------------------------------------------- */ int chfs_getattr(void *v) { struct vnode *vp = ((struct vop_getattr_args *) v)->a_vp; struct vattr *vap = ((struct vop_getattr_args *) v)->a_vap; struct chfs_inode *ip = VTOI(vp); dbg("getattr()\n"); KASSERT(vp->v_size == ip->size); vattr_null(vap); CHFS_ITIMES(ip, NULL, NULL, NULL); vap->va_type = CHTTOVT(ip->ch_type); vap->va_mode = ip->mode & ALLPERMS; vap->va_nlink = ip->chvc->nlink; vap->va_uid = ip->uid; vap->va_gid = ip->gid; vap->va_fsid = ip->dev; vap->va_fileid = ip->ino; vap->va_size = ip->size; vap->va_blocksize = PAGE_SIZE; vap->va_atime.tv_sec = ip->atime; vap->va_atime.tv_nsec = 0; vap->va_mtime.tv_sec = ip->mtime; vap->va_mtime.tv_nsec = 0; vap->va_ctime.tv_sec = ip->ctime; vap->va_ctime.tv_nsec = 0; vap->va_gen = ip->version; vap->va_flags = ip->flags; vap->va_rdev = ip->rdev; vap->va_bytes = round_page(ip->size); vap->va_filerev = VNOVAL; vap->va_vaflags = 0; vap->va_spare = VNOVAL; return 0; } /* --------------------------------------------------------------------- */ /* Note: modelled after tmpfs's same function */ int chfs_setattr(void *v) { struct vnode *vp = ((struct vop_setattr_args *) v)->a_vp; struct vattr *vap = ((struct vop_setattr_args *) v)->a_vap; kauth_cred_t cred = ((struct vop_setattr_args *) v)->a_cred; struct chfs_inode *ip; struct ufsmount *ump = VFSTOUFS(vp->v_mount); struct chfs_mount *chmp = ump->um_chfs; int error = 0; dbg("setattr()\n"); KASSERT(VOP_ISLOCKED(vp)); ip = VTOI(vp); /* Abort if any unsettable attribute is given. */ if (vap->va_type != VNON || vap->va_nlink != VNOVAL || vap->va_fsid != VNOVAL || vap->va_fileid != VNOVAL || vap->va_blocksize != VNOVAL /*|| GOODTIME(&vap->va_ctime)*/ || vap->va_gen != VNOVAL || vap->va_rdev != VNOVAL || vap->va_bytes != VNOVAL) { return EINVAL; } /* set flags */ if (error == 0 && (vap->va_flags != VNOVAL)) { error = chfs_chflags(vp, vap->va_flags, cred); return error; } if (ip->flags & (IMMUTABLE | APPEND)) { error = EPERM; return error; } /* set size */ if (error == 0 && (vap->va_size != VNOVAL)) { error = chfs_chsize(vp, vap->va_size, cred); if (error) return error; } /* set owner */ if (error == 0 && (vap->va_uid != VNOVAL || vap->va_gid != VNOVAL)) { error = chfs_chown(vp, vap->va_uid, vap->va_gid, cred); if (error) return error; } /* set mode */ if (error == 0 && (vap->va_mode != VNOVAL)) { error = chfs_chmod(vp, vap->va_mode, cred); if (error) return error; } /* set time */ if (vap->va_atime.tv_sec != VNOVAL || vap->va_mtime.tv_sec != VNOVAL) { error = kauth_authorize_vnode(cred, KAUTH_VNODE_WRITE_TIMES, vp, NULL, genfs_can_chtimes(vp, vap->va_vaflags, ip->uid, cred)); if (error) return error; if (vap->va_atime.tv_sec != VNOVAL) ip->iflag |= IN_ACCESS; if (vap->va_mtime.tv_sec != VNOVAL) ip->iflag |= IN_CHANGE | IN_UPDATE; error = chfs_update(vp, &vap->va_atime, &vap->va_mtime, UPDATE_WAIT); if (error) return error; } /* Write it out. */ mutex_enter(&chmp->chm_lock_mountfields); error = chfs_write_flash_vnode(chmp, ip, ALLOC_NORMAL); mutex_exit(&chmp->chm_lock_mountfields); return error; } int chfs_chmod(struct vnode *vp, int mode, kauth_cred_t cred) { struct chfs_inode *ip = VTOI(vp); int error; dbg("chmod\n"); error = kauth_authorize_vnode(cred, KAUTH_VNODE_WRITE_SECURITY, vp, NULL, genfs_can_chmod(vp->v_type, cred, ip->uid, ip->gid, mode)); if (error) return error; ip->mode &= ~ALLPERMS; ip->mode |= (mode & ALLPERMS); ip->iflag |= IN_CHANGE; error = chfs_update(vp, NULL, NULL, UPDATE_WAIT); if (error) return error; return 0; } int chfs_chown(struct vnode *vp, uid_t uid, gid_t gid, kauth_cred_t cred) { struct chfs_inode *ip = VTOI(vp); int error; dbg("chown\n"); if (uid == (uid_t)VNOVAL) uid = ip->uid; if (gid == (gid_t)VNOVAL) gid = ip->gid; error = kauth_authorize_vnode(cred, KAUTH_VNODE_CHANGE_OWNERSHIP, vp, NULL, genfs_can_chown(cred, ip->uid, ip->gid, uid, gid)); if (error) return error; ip->gid = gid; ip->uid = uid; ip->iflag |= IN_CHANGE; error = chfs_update(vp, NULL, NULL, UPDATE_WAIT); if (error) return error; return 0; } /* --------------------------------------------------------------------- */ /* calculates ((off_t)blk * chmp->chm_chm_fs_bsize) */ #define chfs_lblktosize(chmp, blk) \ (((off_t)(blk)) << (chmp)->chm_fs_bshift) /* calculates (loc % chmp->chm_chm_fs_bsize) */ #define chfs_blkoff(chmp, loc) \ ((loc) & (chmp)->chm_fs_qbmask) /* calculates (loc / chmp->chm_chm_fs_bsize) */ #define chfs_lblkno(chmp, loc) \ ((loc) >> (chmp)->chm_fs_bshift) /* calculates roundup(size, chmp->chm_chm_fs_fsize) */ #define chfs_fragroundup(chmp, size) \ (((size) + (chmp)->chm_fs_qfmask) & (chmp)->chm_fs_fmask) #define chfs_blksize(chmp, ip, lbn) \ (((lbn) >= UFS_NDADDR || (ip)->size >= chfs_lblktosize(chmp, (lbn) + 1)) \ ? (chmp)->chm_fs_bsize \ : (chfs_fragroundup(chmp, chfs_blkoff(chmp, (ip)->size)))) /* calculates roundup(size, chmp->chm_chm_fs_bsize) */ #define chfs_blkroundup(chmp, size) \ (((size) + (chmp)->chm_fs_qbmask) & (chmp)->chm_fs_bmask) /* from ffs read */ int chfs_read(void *v) { struct vop_read_args /* { struct vnode *a_vp; struct uio *a_uio; int a_ioflag; kauth_cred_t a_cred; } */ *ap = v; struct vnode *vp; struct chfs_inode *ip; struct uio *uio; struct ufsmount *ump; struct buf *bp; struct chfs_mount *chmp; daddr_t lbn, nextlbn; off_t bytesinfile; long size, xfersize, blkoffset; int error, ioflag; vsize_t bytelen; bool usepc = false; dbg("chfs_read\n"); vp = ap->a_vp; ip = VTOI(vp); ump = ip->ump; uio = ap->a_uio; ioflag = ap->a_ioflag; error = 0; dbg("ip->size:%llu\n", (unsigned long long)ip->size); #ifdef DIAGNOSTIC if (uio->uio_rw != UIO_READ) panic("%s: mode", READ_S); if (vp->v_type == VLNK) { if (ip->size < ump->um_maxsymlinklen) panic("%s: short symlink", READ_S); } else if (vp->v_type != VREG && vp->v_type != VDIR) panic("%s: type %d", READ_S, vp->v_type); #endif chmp = ip->chmp; if ((u_int64_t)uio->uio_offset > ump->um_maxfilesize) return (EFBIG); if (uio->uio_resid == 0) return (0); if (uio->uio_offset >= ip->size) goto out; usepc = vp->v_type == VREG; bytelen = 0; if (usepc) { const int advice = IO_ADV_DECODE(ap->a_ioflag); while (uio->uio_resid > 0) { if (ioflag & IO_DIRECT) { genfs_directio(vp, uio, ioflag); } bytelen = MIN(ip->size - uio->uio_offset, uio->uio_resid); if (bytelen == 0) break; error = ubc_uiomove(&vp->v_uobj, uio, bytelen, advice, UBC_READ | UBC_PARTIALOK | (UBC_WANT_UNMAP(vp) ? UBC_UNMAP : 0)); if (error) break; } goto out; } dbg("start reading\n"); for (error = 0, bp = NULL; uio->uio_resid > 0; bp = NULL) { bytesinfile = ip->size - uio->uio_offset; if (bytesinfile <= 0) break; lbn = chfs_lblkno(chmp, uio->uio_offset); nextlbn = lbn + 1; size = chfs_blksize(chmp, ip, lbn); blkoffset = chfs_blkoff(chmp, uio->uio_offset); xfersize = MIN(MIN(chmp->chm_fs_bsize - blkoffset, uio->uio_resid), bytesinfile); if (chfs_lblktosize(chmp, nextlbn) >= ip->size) { error = bread(vp, lbn, size, 0, &bp); dbg("after bread\n"); } else { int nextsize = chfs_blksize(chmp, ip, nextlbn); dbg("size: %ld\n", size); error = breadn(vp, lbn, size, &nextlbn, &nextsize, 1, 0, &bp); dbg("after breadN\n"); } if (error) break; /* * We should only get non-zero b_resid when an I/O error * has occurred, which should cause us to break above. * However, if the short read did not cause an error, * then we want to ensure that we do not uiomove bad * or uninitialized data. */ size -= bp->b_resid; if (size < xfersize) { if (size == 0) break; xfersize = size; } dbg("uiomove\n"); error = uiomove((char *)bp->b_data + blkoffset, xfersize, uio); if (error) break; brelse(bp, 0); } if (bp != NULL) brelse(bp, 0); out: // FIXME HACK ip->ino = ip->chvc->vno; if (!(vp->v_mount->mnt_flag & MNT_NOATIME)) { ip->iflag |= IN_ACCESS; if ((ap->a_ioflag & IO_SYNC) == IO_SYNC) { if (error) { return error; } error = chfs_update(vp, NULL, NULL, UPDATE_WAIT); } } dbg("[END]\n"); return (error); } /* --------------------------------------------------------------------- */ /* from ffs write */ int chfs_write(void *v) { struct vop_write_args /* { struct vnode *a_vp; struct uio *a_uio; int a_ioflag; kauth_cred_t a_cred; } */ *ap = v; struct vnode *vp ; struct uio *uio; struct chfs_inode *ip; struct chfs_mount *chmp; struct lwp *l; kauth_cred_t cred; off_t osize, origoff, oldoff, preallocoff, endallocoff, nsize; int blkoffset, error, flags, ioflag, resid; int aflag; int extended=0; vsize_t bytelen; bool async; struct ufsmount *ump; cred = ap->a_cred; ioflag = ap->a_ioflag; uio = ap->a_uio; vp = ap->a_vp; ip = VTOI(vp); ump = ip->ump; dbg("write\n"); KASSERT(vp->v_size == ip->size); switch (vp->v_type) { case VREG: if (ioflag & IO_APPEND) uio->uio_offset = ip->size; if ((ip->flags & APPEND) && uio->uio_offset != ip->size) return (EPERM); /* FALLTHROUGH */ case VLNK: break; case VDIR: if ((ioflag & IO_SYNC) == 0) panic("chfs_write: nonsync dir write"); break; default: panic("chfs_write: type"); } chmp = ip->chmp; if (uio->uio_offset < 0 || (u_int64_t)uio->uio_offset + uio->uio_resid > ump->um_maxfilesize) { dbg("uio->uio_offset = %lld | uio->uio_offset + " "uio->uio_resid (%llu) > ump->um_maxfilesize (%lld)\n", (long long)uio->uio_offset, (uint64_t)uio->uio_offset + uio->uio_resid, (long long)ump->um_maxfilesize); return (EFBIG); } /* * Maybe this should be above the vnode op call, but so long as * file servers have no limits, I don't think it matters. */ l = curlwp; if (vp->v_type == VREG && l && uio->uio_offset + uio->uio_resid > l->l_proc->p_rlimit[RLIMIT_FSIZE].rlim_cur) { mutex_enter(proc_lock); psignal(l->l_proc, SIGXFSZ); mutex_exit(proc_lock); return (EFBIG); } if (uio->uio_resid == 0) return (0); flags = ioflag & IO_SYNC ? B_SYNC : 0; async = vp->v_mount->mnt_flag & MNT_ASYNC; origoff = uio->uio_offset; resid = uio->uio_resid; osize = ip->size; error = 0; preallocoff = round_page(chfs_blkroundup(chmp, MAX(osize, uio->uio_offset))); aflag = ioflag & IO_SYNC ? B_SYNC : 0; nsize = MAX(osize, uio->uio_offset + uio->uio_resid); endallocoff = nsize - chfs_blkoff(chmp, nsize); /* * if we're increasing the file size, deal with expanding * the fragment if there is one. */ if (nsize > osize && chfs_lblkno(chmp, osize) < UFS_NDADDR && chfs_lblkno(chmp, osize) != chfs_lblkno(chmp, nsize) && chfs_blkroundup(chmp, osize) != osize) { off_t eob; eob = chfs_blkroundup(chmp, osize); uvm_vnp_setwritesize(vp, eob); error = ufs_balloc_range(vp, osize, eob - osize, cred, aflag); if (error) goto out; if (flags & B_SYNC) { mutex_enter(vp->v_interlock); VOP_PUTPAGES(vp, trunc_page(osize & chmp->chm_fs_bmask), round_page(eob), PGO_CLEANIT | PGO_SYNCIO | PGO_JOURNALLOCKED); } } while (uio->uio_resid > 0) { int ubc_flags = UBC_WRITE; bool overwrite; /* if we're overwrite a whole block */ off_t newoff; if (ioflag & IO_DIRECT) { genfs_directio(vp, uio, ioflag | IO_JOURNALLOCKED); } oldoff = uio->uio_offset; blkoffset = chfs_blkoff(chmp, uio->uio_offset); bytelen = MIN(chmp->chm_fs_bsize - blkoffset, uio->uio_resid); if (bytelen == 0) { break; } /* * if we're filling in a hole, allocate the blocks now and * initialize the pages first. if we're extending the file, * we can safely allocate blocks without initializing pages * since the new blocks will be inaccessible until the write * is complete. */ overwrite = uio->uio_offset >= preallocoff && uio->uio_offset < endallocoff; if (!overwrite && (vp->v_vflag & VV_MAPPED) == 0 && chfs_blkoff(chmp, uio->uio_offset) == 0 && (uio->uio_offset & PAGE_MASK) == 0) { vsize_t len; len = trunc_page(bytelen); len -= chfs_blkoff(chmp, len); if (len > 0) { overwrite = true; bytelen = len; } } newoff = oldoff + bytelen; if (vp->v_size < newoff) { uvm_vnp_setwritesize(vp, newoff); } if (!overwrite) { error = ufs_balloc_range(vp, uio->uio_offset, bytelen, cred, aflag); if (error) break; } else { genfs_node_wrlock(vp); error = GOP_ALLOC(vp, uio->uio_offset, bytelen, aflag, cred); genfs_node_unlock(vp); if (error) break; ubc_flags |= UBC_FAULTBUSY; } /* * copy the data. */ ubc_flags |= UBC_WANT_UNMAP(vp) ? UBC_UNMAP : 0; error = ubc_uiomove(&vp->v_uobj, uio, bytelen, IO_ADV_DECODE(ioflag), ubc_flags); /* * update UVM's notion of the size now that we've * copied the data into the vnode's pages. * * we should update the size even when uiomove failed. */ if (vp->v_size < newoff) { uvm_vnp_setsize(vp, newoff); extended = 1; } if (error) break; /* * flush what we just wrote if necessary. * XXXUBC simplistic async flushing. */ if (!async && oldoff >> 16 != uio->uio_offset >> 16) { mutex_enter(vp->v_interlock); error = VOP_PUTPAGES(vp, (oldoff >> 16) << 16, (uio->uio_offset >> 16) << 16, PGO_CLEANIT | PGO_JOURNALLOCKED); if (error) break; } } out: if (error == 0 && ioflag & IO_SYNC) { mutex_enter(vp->v_interlock); error = VOP_PUTPAGES(vp, trunc_page(origoff & chmp->chm_fs_bmask), round_page(chfs_blkroundup(chmp, uio->uio_offset)), PGO_CLEANIT | PGO_SYNCIO | PGO_JOURNALLOCKED); } ip->iflag |= IN_CHANGE | IN_UPDATE; if (resid > uio->uio_resid && ap->a_cred) { if (ip->mode & ISUID) { if (kauth_authorize_vnode(ap->a_cred, KAUTH_VNODE_RETAIN_SUID, vp, NULL, EPERM) != 0) ip->mode &= ~ISUID; } if (ip->mode & ISGID) { if (kauth_authorize_vnode(ap->a_cred, KAUTH_VNODE_RETAIN_SGID, vp, NULL, EPERM) != 0) ip->mode &= ~ISGID; } } if (resid > uio->uio_resid) VN_KNOTE(vp, NOTE_WRITE | (extended ? NOTE_EXTEND : 0)); if (error) { (void) UFS_TRUNCATE(vp, osize, ioflag & IO_SYNC, ap->a_cred); uio->uio_offset -= resid - uio->uio_resid; uio->uio_resid = resid; } else if (resid > uio->uio_resid && (ioflag & IO_SYNC) == IO_SYNC) error = UFS_UPDATE(vp, NULL, NULL, UPDATE_WAIT); //FIXME HACK chfs_set_vnode_size(vp, vp->v_size); KASSERT(vp->v_size == ip->size); mutex_enter(&chmp->chm_lock_mountfields); error = chfs_write_flash_vnode(chmp, ip, ALLOC_NORMAL); mutex_exit(&chmp->chm_lock_mountfields); return (error); } /* --------------------------------------------------------------------- */ int chfs_fsync(void *v) { struct vop_fsync_args /* { struct vnode *a_vp; kauth_cred_t a_cred; int a_flags; off_t offlo; off_t offhi; } */ *ap = v; struct vnode *vp = ap->a_vp; if (ap->a_flags & FSYNC_CACHE) { return ENODEV; } vflushbuf(vp, ap->a_flags); return 0; } /* --------------------------------------------------------------------- */ int chfs_remove(void *v) { struct vnode *dvp = ((struct vop_remove_v2_args *) v)->a_dvp; struct vnode *vp = ((struct vop_remove_v2_args *) v)->a_vp; struct componentname *cnp = (((struct vop_remove_v2_args *) v)->a_cnp); dbg("remove\n"); KASSERT(VOP_ISLOCKED(dvp)); KASSERT(VOP_ISLOCKED(vp)); struct chfs_inode *ip = VTOI(vp); struct chfs_inode *parent = VTOI(dvp); int error = 0; if (vp->v_type == VDIR || (ip->flags & (IMMUTABLE | APPEND)) || (parent->flags & APPEND)) { error = EPERM; goto out; } KASSERT(ip->chvc->vno != ip->chvc->pvno); error = chfs_do_unlink(ip, parent, cnp->cn_nameptr, cnp->cn_namelen); out: vput(vp); return error; } /* --------------------------------------------------------------------- */ int chfs_link(void *v) { struct vnode *dvp = ((struct vop_link_v2_args *) v)->a_dvp; struct vnode *vp = ((struct vop_link_v2_args *) v)->a_vp; struct componentname *cnp = ((struct vop_link_v2_args *) v)->a_cnp; struct chfs_inode *ip, *parent; int error = 0; if (vp->v_type == VDIR) { VOP_ABORTOP(dvp, cnp); error = EISDIR; goto out; } if (dvp->v_mount != vp->v_mount) { VOP_ABORTOP(dvp, cnp); error = EXDEV; goto out; } if (dvp != vp && (error = vn_lock(vp, LK_EXCLUSIVE))) { VOP_ABORTOP(dvp, cnp); goto out; } parent = VTOI(dvp); ip = VTOI(vp); error = chfs_do_link(ip, parent, cnp->cn_nameptr, cnp->cn_namelen, ip->ch_type); if (dvp != vp) VOP_UNLOCK(vp); out: return error; } /* --------------------------------------------------------------------- */ int chfs_rename(void *v) { struct vnode *fdvp = ((struct vop_rename_args *) v)->a_fdvp; struct vnode *fvp = ((struct vop_rename_args *) v)->a_fvp; struct componentname *fcnp = ((struct vop_rename_args *) v)->a_fcnp; struct vnode *tdvp = ((struct vop_rename_args *) v)->a_tdvp; struct vnode *tvp = ((struct vop_rename_args *) v)->a_tvp; struct componentname *tcnp = ((struct vop_rename_args *) v)->a_tcnp; struct chfs_inode *oldparent, *old; struct chfs_inode *newparent; struct chfs_dirent *fd; struct chfs_inode *ip; int error = 0; dbg("rename\n"); KASSERT(VOP_ISLOCKED(tdvp)); KASSERT(IMPLIES(tvp != NULL, VOP_ISLOCKED(tvp) == LK_EXCLUSIVE)); oldparent = VTOI(fdvp); old = VTOI(fvp); newparent = VTOI(tdvp); if (tvp) { dbg("tvp not null\n"); ip = VTOI(tvp); if (tvp->v_type == VDIR) { TAILQ_FOREACH(fd, &ip->dents, fds) { if (fd->vno) { error = ENOTEMPTY; goto out_unlocked; } } } error = chfs_do_unlink(ip, newparent, tcnp->cn_nameptr, tcnp->cn_namelen); vput(tvp); } VFS_VGET(tdvp->v_mount, old->ino, &tvp); ip = VTOI(tvp); /* link new */ error = chfs_do_link(ip, newparent, tcnp->cn_nameptr, tcnp->cn_namelen, ip->ch_type); /* remove old */ error = chfs_do_unlink(old, oldparent, fcnp->cn_nameptr, fcnp->cn_namelen); out_unlocked: /* Release target nodes. */ if (tdvp == tvp) vrele(tdvp); else vput(tdvp); if (tvp != NULL) vput(tvp); /* Release source nodes. */ vrele(fdvp); vrele(fvp); return error; } /* --------------------------------------------------------------------- */ int chfs_mkdir(void *v) { struct vnode *dvp = ((struct vop_mkdir_v3_args *) v)->a_dvp; struct vnode **vpp = ((struct vop_mkdir_v3_args *)v)->a_vpp; struct componentname *cnp = ((struct vop_mkdir_v3_args *) v)->a_cnp; struct vattr *vap = ((struct vop_mkdir_v3_args *) v)->a_vap; dbg("mkdir()\n"); int mode; mode = vap->va_mode & ACCESSPERMS; if ((mode & IFMT) == 0) { mode |= IFDIR; } KASSERT(vap->va_type == VDIR); return chfs_makeinode(mode, dvp, vpp, cnp, VDIR); } /* --------------------------------------------------------------------- */ int chfs_rmdir(void *v) { struct vnode *dvp = ((struct vop_rmdir_v2_args *) v)->a_dvp; struct vnode *vp = ((struct vop_rmdir_v2_args *) v)->a_vp; struct componentname *cnp = ((struct vop_rmdir_v2_args *) v)->a_cnp; dbg("rmdir()\n"); KASSERT(VOP_ISLOCKED(dvp)); KASSERT(VOP_ISLOCKED(vp)); struct chfs_inode *ip = VTOI(vp); struct chfs_inode *parent = VTOI(dvp); struct chfs_dirent *fd; int error = 0; if (vp->v_type != VDIR) { error = ENOTDIR; goto out; } KASSERT(ip->chvc->vno != ip->chvc->pvno); TAILQ_FOREACH(fd, &ip->dents, fds) { if (fd->vno) { error = ENOTEMPTY; goto out; } } error = chfs_do_unlink(ip, parent, cnp->cn_nameptr, cnp->cn_namelen); out: vput(vp); return error; } /* --------------------------------------------------------------------- */ int chfs_symlink(void *v) { struct vnode *dvp = ((struct vop_symlink_v3_args *) v)->a_dvp; struct vnode **vpp = ((struct vop_symlink_v3_args *) v)->a_vpp; struct componentname *cnp = ((struct vop_symlink_v3_args *) v)->a_cnp; struct vattr *vap = ((struct vop_symlink_v3_args *) v)->a_vap; char *target = ((struct vop_symlink_v3_args *) v)->a_target; struct ufsmount *ump; struct chfs_mount *chmp; struct vnode *vp; struct chfs_inode *ip; int len, err; struct chfs_full_dnode *fd; struct buf *bp; dbg("symlink()\n"); ump = VFSTOUFS(dvp->v_mount); chmp = ump->um_chfs; err = chfs_makeinode(IFLNK | vap->va_mode, dvp, vpp, cnp, VLNK); if (err) return (err); VN_KNOTE(dvp, NOTE_WRITE); vp = *vpp; len = strlen(target); ip = VTOI(vp); /* TODO max symlink len instead of "100" */ if (len < 100) { /* symlink path stored as a data node */ ip->target = kmem_alloc(len, KM_SLEEP); memcpy(ip->target, target, len); chfs_set_vnode_size(vp, len); ip->iflag |= IN_CHANGE | IN_UPDATE; bp = getiobuf(vp, true); bp->b_bufsize = bp->b_resid = len; bp->b_data = kmem_alloc(len, KM_SLEEP); memcpy(bp->b_data, target, len); bp->b_blkno = 0; fd = chfs_alloc_full_dnode(); mutex_enter(&chmp->chm_lock_mountfields); /* write out the data node */ err = chfs_write_flash_dnode(chmp, vp, bp, fd); if (err) { mutex_exit(&chmp->chm_lock_mountfields); goto out; } /* add it to the inode */ err = chfs_add_full_dnode_to_inode(chmp, ip, fd); if (err) { mutex_exit(&chmp->chm_lock_mountfields); goto out; } mutex_exit(&chmp->chm_lock_mountfields); kmem_free(bp->b_data, len); putiobuf(bp); uvm_vnp_setsize(vp, len); } else { err = ufs_bufio(UIO_WRITE, vp, target, len, (off_t)0, IO_NODELOCKED, cnp->cn_cred, (size_t *)0, NULL); } out: if (err) vrele(vp); return (err); } /* --------------------------------------------------------------------- */ int chfs_readdir(void *v) { struct vnode *vp = ((struct vop_readdir_args *) v)->a_vp; struct uio *uio = ((struct vop_readdir_args *) v)->a_uio; int *eofflag = ((struct vop_readdir_args *) v)->a_eofflag; int error = 0; off_t skip, offset; struct chfs_inode *ip; struct chfs_dirent *fd; struct ufsmount *ump; struct chfs_mount *chmp; struct chfs_vnode_cache *chvc; KASSERT(VOP_ISLOCKED(vp)); /* This operation only makes sense on directory nodes. */ if (vp->v_type != VDIR) { error = ENOTDIR; goto out; } ip = VTOI(vp); /* uiomove in chfs_filldir automatically increments the * uio_offset by an arbitrary size, so we discard any change * to uio_offset and set it to our own value on return */ offset = uio->uio_offset; /* Add this entry. */ if (offset == CHFS_OFFSET_DOT) { error = chfs_filldir(uio, ip->ino, ".", 1, CHT_DIR); if (error == -1) { error = 0; goto outok; } else if (error != 0) goto outok; offset = CHFS_OFFSET_DOTDOT; } /* Add parent entry. */ if (offset == CHFS_OFFSET_DOTDOT) { ump = VFSTOUFS(vp->v_mount); chmp = ump->um_chfs; mutex_enter(&chmp->chm_lock_vnocache); chvc = chfs_vnode_cache_get(chmp, ip->ino); mutex_exit(&chmp->chm_lock_vnocache); error = chfs_filldir(uio, chvc->pvno, "..", 2, CHT_DIR); if (error == -1) { error = 0; goto outok; } else if (error != 0) { goto outok; } /* Has child or not? */ if (TAILQ_EMPTY(&ip->dents)) { offset = CHFS_OFFSET_EOF; } else { offset = CHFS_OFFSET_FIRST; } } if (offset != CHFS_OFFSET_EOF) { /* Child entries. */ skip = offset - CHFS_OFFSET_FIRST; TAILQ_FOREACH(fd, &ip->dents, fds) { /* seek to offset by skipping items */ /* XXX race conditions by changed dirent? */ if (skip > 0) { skip--; continue; } if (fd->vno != 0) { error = chfs_filldir(uio, fd->vno, fd->name, fd->nsize, fd->type); if (error == -1) { error = 0; goto outok; } else if (error != 0) { dbg("err %d\n", error); goto outok; } } offset++; } } offset = CHFS_OFFSET_EOF; outok: uio->uio_offset = offset; if (eofflag != NULL) { *eofflag = (error == 0 && uio->uio_offset == CHFS_OFFSET_EOF); } out: KASSERT(VOP_ISLOCKED(vp)); return error; } /* --------------------------------------------------------------------- */ int chfs_readlink(void *v) { struct vnode *vp = ((struct vop_readlink_args *) v)->a_vp; struct uio *uio = ((struct vop_readlink_args *) v)->a_uio; kauth_cred_t cred = ((struct vop_readlink_args *) v)->a_cred; struct chfs_inode *ip = VTOI(vp); dbg("readlink()\n"); /* TODO max symlink len instead of "100" */ if (ip->size < 100) { uiomove(ip->target, ip->size, uio); return (0); } return (UFS_BUFRD(vp, uio, 0, cred)); } /* --------------------------------------------------------------------- */ int chfs_inactive(void *v) { struct vnode *vp = ((struct vop_inactive_v2_args *) v)->a_vp; struct chfs_inode *ip = VTOI(vp); struct chfs_vnode_cache *chvc; dbg("inactive | vno: %llu\n", (unsigned long long)ip->ino); KASSERT(VOP_ISLOCKED(vp)); /* Reclaim only if there is no link to the node. */ if (ip->ino) { chvc = ip->chvc; if (chvc->nlink) *((struct vop_inactive_v2_args *) v)->a_recycle = 0; } else { *((struct vop_inactive_v2_args *) v)->a_recycle = 1; } return 0; } /* --------------------------------------------------------------------- */ int chfs_reclaim(void *v) { struct vop_reclaim_v2_args *ap = v; struct vnode *vp = ap->a_vp; struct chfs_inode *ip = VTOI(vp); struct chfs_mount *chmp = ip->chmp; struct chfs_dirent *fd; VOP_UNLOCK(vp); mutex_enter(&chmp->chm_lock_mountfields); mutex_enter(&chmp->chm_lock_vnocache); ip->chvc->state = VNO_STATE_CHECKEDABSENT; mutex_exit(&chmp->chm_lock_vnocache); chfs_update(vp, NULL, NULL, UPDATE_CLOSE); /* Clean fragments. */ chfs_kill_fragtree(chmp, &ip->fragtree); /* Clean dirents. */ fd = TAILQ_FIRST(&ip->dents); while (fd) { TAILQ_REMOVE(&ip->dents, fd, fds); chfs_free_dirent(fd); fd = TAILQ_FIRST(&ip->dents); } cache_purge(vp); if (ip->devvp) { vrele(ip->devvp); ip->devvp = 0; } genfs_node_destroy(vp); pool_put(&chfs_inode_pool, vp->v_data); vp->v_data = NULL; mutex_exit(&chmp->chm_lock_mountfields); return (0); } /* --------------------------------------------------------------------- */ int chfs_advlock(void *v) { return 0; } /* --------------------------------------------------------------------- */ int chfs_strategy(void *v) { struct vop_strategy_args /* { const struct vnodeop_desc *a_desc; struct vnode *a_vp; struct buf *a_bp; } */ *ap = v; struct chfs_full_dnode *fd; struct buf *bp = ap->a_bp; struct vnode *vp = ap->a_vp; struct chfs_inode *ip = VTOI(vp); struct chfs_mount *chmp = ip->chmp; int read = (bp->b_flags & B_READ) ? 1 : 0; int err = 0; if (read) { err = chfs_read_data(chmp, vp, bp); } else { mutex_enter(&chmp->chm_lock_mountfields); fd = chfs_alloc_full_dnode(); err = chfs_write_flash_dnode(chmp, vp, bp, fd); if (err) { mutex_exit(&chmp->chm_lock_mountfields); goto out; } ip = VTOI(vp); err = chfs_add_full_dnode_to_inode(chmp, ip, fd); mutex_exit(&chmp->chm_lock_mountfields); } out: biodone(bp); return err; } int chfs_bmap(void *v) { struct vop_bmap_args /* { struct vnode *a_vp; daddr_t a_bn; struct vnode **a_vpp; daddr_t *a_bnp; int *a_runp; int *a_runb; } */ *ap = v; if (ap->a_vpp != NULL) *ap->a_vpp = ap->a_vp; if (ap->a_bnp != NULL) *ap->a_bnp = ap->a_bn; if (ap->a_runp != NULL) *ap->a_runp = 0; return (0); } /* * vnode operations vector used for files stored in a chfs file system. */ int (**chfs_vnodeop_p)(void *); const struct vnodeopv_entry_desc chfs_vnodeop_entries[] = { { &vop_default_desc, vn_default_error }, { &vop_lookup_desc, chfs_lookup }, { &vop_create_desc, chfs_create }, { &vop_mknod_desc, chfs_mknod }, { &vop_open_desc, chfs_open }, { &vop_close_desc, chfs_close }, { &vop_access_desc, chfs_access }, { &vop_getattr_desc, chfs_getattr }, { &vop_setattr_desc, chfs_setattr }, { &vop_read_desc, chfs_read }, { &vop_write_desc, chfs_write }, { &vop_fallocate_desc, genfs_eopnotsupp }, { &vop_fdiscard_desc, genfs_eopnotsupp }, { &vop_ioctl_desc, genfs_enoioctl }, { &vop_fcntl_desc, genfs_fcntl }, { &vop_poll_desc, genfs_poll }, { &vop_kqfilter_desc, genfs_kqfilter }, { &vop_revoke_desc, genfs_revoke }, { &vop_mmap_desc, genfs_mmap }, { &vop_fsync_desc, chfs_fsync }, { &vop_seek_desc, genfs_seek }, { &vop_remove_desc, chfs_remove }, { &vop_link_desc, chfs_link }, { &vop_rename_desc, chfs_rename }, { &vop_mkdir_desc, chfs_mkdir }, { &vop_rmdir_desc, chfs_rmdir }, { &vop_symlink_desc, chfs_symlink }, { &vop_readdir_desc, chfs_readdir }, { &vop_readlink_desc, chfs_readlink }, { &vop_abortop_desc, genfs_abortop }, { &vop_inactive_desc, chfs_inactive }, { &vop_reclaim_desc, chfs_reclaim }, { &vop_lock_desc, genfs_lock }, { &vop_unlock_desc, genfs_unlock }, { &vop_bmap_desc, chfs_bmap }, { &vop_strategy_desc, chfs_strategy }, { &vop_print_desc, ufs_print }, { &vop_pathconf_desc, ufs_pathconf }, { &vop_islocked_desc, genfs_islocked }, { &vop_advlock_desc, chfs_advlock }, { &vop_bwrite_desc, vn_bwrite }, { &vop_getpages_desc, genfs_getpages }, { &vop_putpages_desc, genfs_putpages }, { NULL, NULL } }; const struct vnodeopv_desc chfs_vnodeop_opv_desc = { &chfs_vnodeop_p, chfs_vnodeop_entries }; /* --------------------------------------------------------------------- */ /* * vnode operations vector used for special devices stored in a chfs * file system. */ int (**chfs_specop_p)(void *); const struct vnodeopv_entry_desc chfs_specop_entries[] = { { &vop_default_desc, vn_default_error }, { &vop_lookup_desc, spec_lookup }, { &vop_create_desc, spec_create }, { &vop_mknod_desc, spec_mknod }, { &vop_open_desc, spec_open }, { &vop_close_desc, ufsspec_close }, { &vop_access_desc, chfs_access }, { &vop_getattr_desc, chfs_getattr }, { &vop_setattr_desc, chfs_setattr }, { &vop_read_desc, chfs_read }, { &vop_write_desc, chfs_write }, { &vop_fallocate_desc, spec_fallocate }, { &vop_fdiscard_desc, spec_fdiscard }, { &vop_ioctl_desc, spec_ioctl }, { &vop_fcntl_desc, genfs_fcntl }, { &vop_poll_desc, spec_poll }, { &vop_kqfilter_desc, spec_kqfilter }, { &vop_revoke_desc, spec_revoke }, { &vop_mmap_desc, spec_mmap }, { &vop_fsync_desc, spec_fsync }, { &vop_seek_desc, spec_seek }, { &vop_remove_desc, spec_remove }, { &vop_link_desc, spec_link }, { &vop_rename_desc, spec_rename }, { &vop_mkdir_desc, spec_mkdir }, { &vop_rmdir_desc, spec_rmdir }, { &vop_symlink_desc, spec_symlink }, { &vop_readdir_desc, spec_readdir }, { &vop_readlink_desc, spec_readlink }, { &vop_abortop_desc, spec_abortop }, { &vop_inactive_desc, chfs_inactive }, { &vop_reclaim_desc, chfs_reclaim }, { &vop_lock_desc, genfs_lock }, { &vop_unlock_desc, genfs_unlock }, { &vop_bmap_desc, spec_bmap }, { &vop_strategy_desc, spec_strategy }, { &vop_print_desc, ufs_print }, { &vop_pathconf_desc, spec_pathconf }, { &vop_islocked_desc, genfs_islocked }, { &vop_advlock_desc, spec_advlock }, { &vop_bwrite_desc, vn_bwrite }, { &vop_getpages_desc, spec_getpages }, { &vop_putpages_desc, spec_putpages }, { NULL, NULL } }; const struct vnodeopv_desc chfs_specop_opv_desc = { &chfs_specop_p, chfs_specop_entries }; /* --------------------------------------------------------------------- */ /* * vnode operations vector used for fifos stored in a chfs file system. */ int (**chfs_fifoop_p)(void *); const struct vnodeopv_entry_desc chfs_fifoop_entries[] = { { &vop_default_desc, vn_default_error }, { &vop_lookup_desc, vn_fifo_bypass }, { &vop_create_desc, vn_fifo_bypass }, { &vop_mknod_desc, vn_fifo_bypass }, { &vop_open_desc, vn_fifo_bypass }, { &vop_close_desc, ufsfifo_close }, { &vop_access_desc, chfs_access }, { &vop_getattr_desc, chfs_getattr }, { &vop_setattr_desc, chfs_setattr }, { &vop_read_desc, ufsfifo_read }, { &vop_write_desc, ufsfifo_write }, { &vop_fallocate_desc, vn_fifo_bypass }, { &vop_fdiscard_desc, vn_fifo_bypass }, { &vop_ioctl_desc, vn_fifo_bypass }, { &vop_fcntl_desc, genfs_fcntl }, { &vop_poll_desc, vn_fifo_bypass }, { &vop_kqfilter_desc, vn_fifo_bypass }, { &vop_revoke_desc, vn_fifo_bypass }, { &vop_mmap_desc, vn_fifo_bypass }, { &vop_fsync_desc, vn_fifo_bypass }, { &vop_seek_desc, vn_fifo_bypass }, { &vop_remove_desc, vn_fifo_bypass }, { &vop_link_desc, vn_fifo_bypass }, { &vop_rename_desc, vn_fifo_bypass }, { &vop_mkdir_desc, vn_fifo_bypass }, { &vop_rmdir_desc, vn_fifo_bypass }, { &vop_symlink_desc, vn_fifo_bypass }, { &vop_readdir_desc, vn_fifo_bypass }, { &vop_readlink_desc, vn_fifo_bypass }, { &vop_abortop_desc, vn_fifo_bypass }, { &vop_inactive_desc, chfs_inactive }, { &vop_reclaim_desc, chfs_reclaim }, { &vop_lock_desc, genfs_lock }, { &vop_unlock_desc, genfs_unlock }, { &vop_bmap_desc, vn_fifo_bypass }, { &vop_strategy_desc, vn_fifo_bypass }, { &vop_print_desc, ufs_print }, { &vop_pathconf_desc, vn_fifo_bypass }, { &vop_islocked_desc, genfs_islocked }, { &vop_advlock_desc, vn_fifo_bypass }, { &vop_bwrite_desc, genfs_nullop }, { &vop_getpages_desc, genfs_badop }, { &vop_putpages_desc, vn_fifo_bypass }, { NULL, NULL } }; const struct vnodeopv_desc chfs_fifoop_opv_desc = { &chfs_fifoop_p, chfs_fifoop_entries };