+ switch (FILEGLOB_DTYPE(fp->fp_glob)) {
+ case DTYPE_VNODE:
+ case DTYPE_PIPE:
+ case DTYPE_SOCKET:
+ break;
+ default:
+ error = EINVAL;
+ goto out;
+ }
+ if (rw == UIO_WRITE && !(fp->f_flag & FWRITE)) {
+ error = EBADF;
+ goto out;
+ }
+
+ if (rw == UIO_READ && !(fp->f_flag & FREAD)) {
+ error = EBADF;
+ goto out;
+ }
+
+ context.vc_ucred = fp->fp_glob->fg_cred;
+
+ if (UIO_SEG_IS_USER_SPACE(segflg)) {
+ spacetype = proc_is64bit(p) ? UIO_USERSPACE64 : UIO_USERSPACE32;
+ } else {
+ spacetype = UIO_SYSSPACE;
+ }
+
+ auio = uio_createwithbuffer(1, offset, spacetype, rw, &uio_buf[0], sizeof(uio_buf));
+
+ uio_addiov(auio, (user_addr_t)base, (user_size_t)len);
+
+ if (!(io_flg & IO_APPEND)) {
+ flags = FOF_OFFSET;
+ }
+
+ if (rw == UIO_WRITE) {
+ user_ssize_t orig_resid = uio_resid(auio);
+ error = fo_write(fp, auio, flags, &context);
+ if (uio_resid(auio) < orig_resid) {
+ os_atomic_or(&fp->fp_glob->fg_flag, FWASWRITTEN, relaxed);
+ }
+ } else {
+ error = fo_read(fp, auio, flags, &context);
+ }
+
+ if (aresid) {
+ *aresid = uio_resid(auio);
+ } else if (uio_resid(auio) && error == 0) {
+ error = EIO;
+ }
+out:
+ fp_drop(p, fd, fp, 0);
+ return error;
+}
+
+
+
+/*
+ * sys_dup
+ *
+ * Description: Duplicate a file descriptor.
+ *
+ * Parameters: p Process performing the dup
+ * uap->fd The fd to dup
+ * retval Pointer to the call return area
+ *
+ * Returns: 0 Success
+ * !0 Errno
+ *
+ * Implicit returns:
+ * *retval (modified) The new descriptor
+ */
+int
+sys_dup(proc_t p, struct dup_args *uap, int32_t *retval)
+{
+ struct filedesc *fdp = p->p_fd;
+ int old = uap->fd;
+ int new, error;
+ struct fileproc *fp;
+
+ proc_fdlock(p);
+ if ((error = fp_lookup(p, old, &fp, 1))) {
+ proc_fdunlock(p);
+ return error;
+ }
+ if (fp_isguarded(fp, GUARD_DUP)) {
+ error = fp_guard_exception(p, old, fp, kGUARD_EXC_DUP);
+ (void) fp_drop(p, old, fp, 1);
+ proc_fdunlock(p);
+ return error;
+ }
+ if ((error = fdalloc(p, 0, &new))) {
+ fp_drop(p, old, fp, 1);
+ proc_fdunlock(p);
+ return error;
+ }
+ error = finishdup(p, fdp, old, new, 0, retval);
+ fp_drop(p, old, fp, 1);
+ proc_fdunlock(p);
+
+ if (ENTR_SHOULDTRACE && FILEGLOB_DTYPE(fp->fp_glob) == DTYPE_SOCKET) {
+ KERNEL_ENERGYTRACE(kEnTrActKernSocket, DBG_FUNC_START,
+ new, 0, (int64_t)VM_KERNEL_ADDRPERM(fp->f_data));
+ }
+
+ return error;
+}
+
+/*
+ * sys_dup2
+ *
+ * Description: Duplicate a file descriptor to a particular value.
+ *
+ * Parameters: p Process performing the dup
+ * uap->from The fd to dup
+ * uap->to The fd to dup it to
+ * retval Pointer to the call return area
+ *
+ * Returns: 0 Success
+ * !0 Errno
+ *
+ * Implicit returns:
+ * *retval (modified) The new descriptor
+ */
+int
+sys_dup2(proc_t p, struct dup2_args *uap, int32_t *retval)
+{
+ return dup2(p, uap->from, uap->to, retval);
+}
+
+int
+dup2(proc_t p, int old, int new, int *retval)
+{
+ struct filedesc *fdp = p->p_fd;
+ struct fileproc *fp, *nfp;
+ int i, error;
+ rlim_t nofile = proc_limitgetcur(p, RLIMIT_NOFILE, TRUE);
+
+ proc_fdlock(p);
+
+startover:
+ if ((error = fp_lookup(p, old, &fp, 1))) {
+ proc_fdunlock(p);
+ return error;
+ }
+ if (fp_isguarded(fp, GUARD_DUP)) {
+ error = fp_guard_exception(p, old, fp, kGUARD_EXC_DUP);
+ (void) fp_drop(p, old, fp, 1);
+ proc_fdunlock(p);
+ return error;
+ }
+ if (new < 0 ||
+ (rlim_t)new >= nofile ||
+ new >= maxfilesperproc) {
+ fp_drop(p, old, fp, 1);
+ proc_fdunlock(p);
+ return EBADF;
+ }
+ if (old == new) {
+ fp_drop(p, old, fp, 1);
+ *retval = new;
+ proc_fdunlock(p);
+ return 0;
+ }
+ if (new < 0 || new >= fdp->fd_nfiles) {
+ if ((error = fdalloc(p, new, &i))) {
+ fp_drop(p, old, fp, 1);
+ proc_fdunlock(p);
+ return error;
+ }
+ if (new != i) {
+ fdrelse(p, i);
+ goto closeit;
+ }
+ } else {
+closeit:
+ if ((fdp->fd_ofileflags[new] & UF_RESERVED) == UF_RESERVED) {
+ fp_drop(p, old, fp, 1);
+ procfdtbl_waitfd(p, new);
+#if DIAGNOSTIC
+ proc_fdlock_assert(p, LCK_MTX_ASSERT_OWNED);
+#endif
+ goto startover;
+ }
+
+ if ((nfp = fdp->fd_ofiles[new]) != NULL) {
+ if (fp_isguarded(nfp, GUARD_CLOSE)) {
+ fp_drop(p, old, fp, 1);
+ error = fp_guard_exception(p,
+ new, nfp, kGUARD_EXC_CLOSE);
+ proc_fdunlock(p);
+ return error;
+ }
+ (void)fp_close_and_unlock(p, new, nfp, FD_DUP2RESV);
+ proc_fdlock(p);
+ assert(fdp->fd_ofileflags[new] & UF_RESERVED);
+ } else {
+#if DIAGNOSTIC
+ if (fdp->fd_ofiles[new] != NULL) {
+ panic("dup2: no ref on fileproc %d", new);
+ }
+#endif
+ procfdtbl_reservefd(p, new);
+ }
+ }
+#if DIAGNOSTIC
+ if (fdp->fd_ofiles[new] != 0) {
+ panic("dup2: overwriting fd_ofiles with new %d", new);
+ }
+ if ((fdp->fd_ofileflags[new] & UF_RESERVED) == 0) {
+ panic("dup2: unreserved fileflags with new %d", new);
+ }
+#endif
+ error = finishdup(p, fdp, old, new, 0, retval);
+ fp_drop(p, old, fp, 1);
+ proc_fdunlock(p);
+
+ return error;
+}
+
+
+/*
+ * fcntl
+ *
+ * Description: The file control system call.
+ *
+ * Parameters: p Process performing the fcntl
+ * uap->fd The fd to operate against
+ * uap->cmd The command to perform
+ * uap->arg Pointer to the command argument
+ * retval Pointer to the call return area
+ *
+ * Returns: 0 Success
+ * !0 Errno (see fcntl_nocancel)
+ *
+ * Implicit returns:
+ * *retval (modified) fcntl return value (if any)
+ *
+ * Notes: This system call differs from fcntl_nocancel() in that it
+ * tests for cancellation prior to performing a potentially
+ * blocking operation.
+ */
+int
+sys_fcntl(proc_t p, struct fcntl_args *uap, int32_t *retval)
+{
+ __pthread_testcancel(1);
+ return sys_fcntl_nocancel(p, (struct fcntl_nocancel_args *)uap, retval);
+}
+
+#define ACCOUNT_OPENFROM_ENTITLEMENT \
+ "com.apple.private.vfs.role-account-openfrom"
+
+/*
+ * sys_fcntl_nocancel
+ *
+ * Description: A non-cancel-testing file control system call.
+ *
+ * Parameters: p Process performing the fcntl
+ * uap->fd The fd to operate against
+ * uap->cmd The command to perform
+ * uap->arg Pointer to the command argument
+ * retval Pointer to the call return area
+ *
+ * Returns: 0 Success
+ * EINVAL
+ * fp_lookup:EBADF Bad file descriptor
+ * [F_DUPFD]
+ * fdalloc:EMFILE
+ * fdalloc:ENOMEM
+ * finishdup:EBADF
+ * finishdup:ENOMEM
+ * [F_SETOWN]
+ * ESRCH
+ * [F_SETLK]
+ * EBADF
+ * EOVERFLOW
+ * copyin:EFAULT
+ * vnode_getwithref:???
+ * VNOP_ADVLOCK:???
+ * msleep:ETIMEDOUT
+ * [F_GETLK]
+ * EBADF
+ * EOVERFLOW
+ * copyin:EFAULT
+ * copyout:EFAULT
+ * vnode_getwithref:???
+ * VNOP_ADVLOCK:???
+ * [F_PREALLOCATE]
+ * EBADF
+ * EINVAL
+ * copyin:EFAULT
+ * copyout:EFAULT
+ * vnode_getwithref:???
+ * VNOP_ALLOCATE:???
+ * [F_SETSIZE,F_RDADVISE]
+ * EBADF
+ * EINVAL
+ * copyin:EFAULT
+ * vnode_getwithref:???
+ * [F_RDAHEAD,F_NOCACHE]
+ * EBADF
+ * vnode_getwithref:???
+ * [???]
+ *
+ * Implicit returns:
+ * *retval (modified) fcntl return value (if any)
+ */
+int
+sys_fcntl_nocancel(proc_t p, struct fcntl_nocancel_args *uap, int32_t *retval)
+{
+ int fd = uap->fd;
+ struct filedesc *fdp = p->p_fd;
+ struct fileproc *fp;
+ char *pop;
+ struct vnode *vp = NULLVP; /* for AUDIT_ARG() at end */
+ unsigned int oflags, nflags;
+ int i, tmp, error, error2, flg = 0;
+ struct flock fl = {};
+ struct flocktimeout fltimeout;
+ struct timespec *timeout = NULL;
+ struct vfs_context context;
+ off_t offset;
+ int newmin;
+ daddr64_t lbn, bn;
+ unsigned int fflag;
+ user_addr_t argp;
+ boolean_t is64bit;
+ rlim_t nofile;
+ int has_entitlement = 0;
+
+ AUDIT_ARG(fd, uap->fd);
+ AUDIT_ARG(cmd, uap->cmd);
+
+ nofile = proc_limitgetcur(p, RLIMIT_NOFILE, TRUE);
+
+ proc_fdlock(p);
+ if ((error = fp_lookup(p, fd, &fp, 1))) {
+ proc_fdunlock(p);
+ return error;
+ }
+ context.vc_thread = current_thread();
+ context.vc_ucred = fp->f_cred;
+
+ is64bit = proc_is64bit(p);
+ if (is64bit) {
+ argp = uap->arg;
+ } else {
+ /*
+ * Since the arg parameter is defined as a long but may be
+ * either a long or a pointer we must take care to handle
+ * sign extension issues. Our sys call munger will sign
+ * extend a long when we are called from a 32-bit process.
+ * Since we can never have an address greater than 32-bits
+ * from a 32-bit process we lop off the top 32-bits to avoid
+ * getting the wrong address
+ */
+ argp = CAST_USER_ADDR_T((uint32_t)uap->arg);
+ }
+
+#if CONFIG_MACF
+ error = mac_file_check_fcntl(proc_ucred(p), fp->fp_glob, uap->cmd,
+ uap->arg);
+ if (error) {
+ goto out;
+ }
+#endif
+
+ pop = &fdp->fd_ofileflags[fd];
+
+ switch (uap->cmd) {
+ case F_DUPFD:
+ case F_DUPFD_CLOEXEC:
+ if (fp_isguarded(fp, GUARD_DUP)) {
+ error = fp_guard_exception(p, fd, fp, kGUARD_EXC_DUP);
+ goto out;
+ }
+ newmin = CAST_DOWN_EXPLICIT(int, uap->arg); /* arg is an int, so we won't lose bits */
+ AUDIT_ARG(value32, newmin);
+ if ((rlim_t)newmin >= nofile ||
+ newmin >= maxfilesperproc) {
+ error = EINVAL;
+ goto out;
+ }
+ if ((error = fdalloc(p, newmin, &i))) {
+ goto out;
+ }
+ error = finishdup(p, fdp, fd, i,
+ uap->cmd == F_DUPFD_CLOEXEC ? UF_EXCLOSE : 0, retval);
+ goto out;
+
+ case F_GETFD:
+ *retval = (*pop & UF_EXCLOSE)? FD_CLOEXEC : 0;
+ error = 0;
+ goto out;
+
+ case F_SETFD:
+ AUDIT_ARG(value32, (uint32_t)uap->arg);
+ if (uap->arg & FD_CLOEXEC) {
+ *pop |= UF_EXCLOSE;
+ } else {
+ if (fp_isguarded(fp, 0)) {
+ error = fp_guard_exception(p,
+ fd, fp, kGUARD_EXC_NOCLOEXEC);
+ goto out;
+ }
+ *pop &= ~UF_EXCLOSE;
+ }
+ error = 0;
+ goto out;
+
+ case F_GETFL:
+ *retval = OFLAGS(fp->f_flag);
+ error = 0;
+ goto out;
+
+ case F_SETFL:
+ // FIXME (rdar://54898652)
+ //
+ // this code is broken if fnctl(F_SETFL), ioctl() are
+ // called concurrently for the same fileglob.
+
+ tmp = CAST_DOWN_EXPLICIT(int, uap->arg); /* arg is an int, so we won't lose bits */
+ AUDIT_ARG(value32, tmp);
+
+ os_atomic_rmw_loop(&fp->f_flag, oflags, nflags, relaxed, {
+ nflags = oflags & ~FCNTLFLAGS;
+ nflags |= FFLAGS(tmp) & FCNTLFLAGS;
+ });
+ tmp = nflags & FNONBLOCK;
+ error = fo_ioctl(fp, FIONBIO, (caddr_t)&tmp, &context);
+ if (error) {
+ goto out;
+ }
+ tmp = nflags & FASYNC;
+ error = fo_ioctl(fp, FIOASYNC, (caddr_t)&tmp, &context);
+ if (!error) {
+ goto out;
+ }
+ os_atomic_andnot(&fp->f_flag, FNONBLOCK, relaxed);
+ tmp = 0;
+ (void)fo_ioctl(fp, FIONBIO, (caddr_t)&tmp, &context);
+ goto out;
+
+ case F_GETOWN:
+ if (fp->f_type == DTYPE_SOCKET) {
+ *retval = ((struct socket *)fp->f_data)->so_pgid;
+ error = 0;
+ goto out;
+ }
+ error = fo_ioctl(fp, TIOCGPGRP, (caddr_t)retval, &context);
+ *retval = -*retval;
+ goto out;
+
+ case F_SETOWN:
+ tmp = CAST_DOWN_EXPLICIT(pid_t, uap->arg); /* arg is an int, so we won't lose bits */
+ AUDIT_ARG(value32, tmp);
+ if (fp->f_type == DTYPE_SOCKET) {
+ ((struct socket *)fp->f_data)->so_pgid = tmp;
+ error = 0;
+ goto out;
+ }
+ if (fp->f_type == DTYPE_PIPE) {
+ error = fo_ioctl(fp, TIOCSPGRP, (caddr_t)&tmp, &context);
+ goto out;
+ }
+
+ if (tmp <= 0) {
+ tmp = -tmp;
+ } else {
+ proc_t p1 = proc_find(tmp);
+ if (p1 == 0) {
+ error = ESRCH;
+ goto out;
+ }
+ tmp = (int)p1->p_pgrpid;
+ proc_rele(p1);
+ }
+ error = fo_ioctl(fp, TIOCSPGRP, (caddr_t)&tmp, &context);
+ goto out;
+
+ case F_SETNOSIGPIPE:
+ tmp = CAST_DOWN_EXPLICIT(int, uap->arg);
+ if (fp->f_type == DTYPE_SOCKET) {
+#if SOCKETS
+ error = sock_setsockopt((struct socket *)fp->f_data,
+ SOL_SOCKET, SO_NOSIGPIPE, &tmp, sizeof(tmp));
+#else
+ error = EINVAL;
+#endif
+ } else {
+ struct fileglob *fg = fp->fp_glob;
+
+ lck_mtx_lock_spin(&fg->fg_lock);
+ if (tmp) {
+ fg->fg_lflags |= FG_NOSIGPIPE;
+ } else {
+ fg->fg_lflags &= ~FG_NOSIGPIPE;
+ }
+ lck_mtx_unlock(&fg->fg_lock);
+ error = 0;
+ }
+ goto out;
+
+ case F_GETNOSIGPIPE:
+ if (fp->f_type == DTYPE_SOCKET) {
+#if SOCKETS
+ int retsize = sizeof(*retval);
+ error = sock_getsockopt((struct socket *)fp->f_data,
+ SOL_SOCKET, SO_NOSIGPIPE, retval, &retsize);
+#else
+ error = EINVAL;
+#endif
+ } else {
+ *retval = (fp->fp_glob->fg_lflags & FG_NOSIGPIPE) ?
+ 1 : 0;
+ error = 0;
+ }
+ goto out;
+
+ case F_SETCONFINED:
+ /*
+ * If this is the only reference to this fglob in the process
+ * and it's already marked as close-on-fork then mark it as
+ * (immutably) "confined" i.e. any fd that points to it will
+ * forever be close-on-fork, and attempts to use an IPC
+ * mechanism to move the descriptor elsewhere will fail.
+ */
+ if (CAST_DOWN_EXPLICIT(int, uap->arg)) {
+ struct fileglob *fg = fp->fp_glob;
+
+ lck_mtx_lock_spin(&fg->fg_lock);
+ if (fg->fg_lflags & FG_CONFINED) {
+ error = 0;
+ } else if (1 != os_ref_get_count_raw(&fg->fg_count)) {
+ error = EAGAIN; /* go close the dup .. */
+ } else if (UF_FORKCLOSE == (*pop & UF_FORKCLOSE)) {
+ fg->fg_lflags |= FG_CONFINED;
+ error = 0;
+ } else {
+ error = EBADF; /* open without O_CLOFORK? */
+ }
+ lck_mtx_unlock(&fg->fg_lock);
+ } else {
+ /*
+ * Other subsystems may have built on the immutability
+ * of FG_CONFINED; clearing it may be tricky.
+ */
+ error = EPERM; /* immutable */
+ }
+ goto out;
+
+ case F_GETCONFINED:
+ *retval = (fp->fp_glob->fg_lflags & FG_CONFINED) ? 1 : 0;
+ error = 0;
+ goto out;
+
+ case F_SETLKWTIMEOUT:
+ case F_SETLKW:
+ case F_OFD_SETLKWTIMEOUT:
+ case F_OFD_SETLKW:
+ flg |= F_WAIT;
+ OS_FALLTHROUGH;
+
+ case F_SETLK:
+ case F_OFD_SETLK:
+ if (fp->f_type != DTYPE_VNODE) {
+ error = EBADF;
+ goto out;
+ }
+ vp = (struct vnode *)fp->f_data;
+
+ fflag = fp->f_flag;
+ offset = fp->f_offset;
+ proc_fdunlock(p);
+
+ /* Copy in the lock structure */
+ if (F_SETLKWTIMEOUT == uap->cmd ||
+ F_OFD_SETLKWTIMEOUT == uap->cmd) {
+ error = copyin(argp, (caddr_t) &fltimeout, sizeof(fltimeout));
+ if (error) {
+ goto outdrop;
+ }
+ fl = fltimeout.fl;
+ timeout = &fltimeout.timeout;
+ } else {
+ error = copyin(argp, (caddr_t)&fl, sizeof(fl));
+ if (error) {
+ goto outdrop;
+ }
+ }
+
+ /* Check starting byte and ending byte for EOVERFLOW in SEEK_CUR */
+ /* and ending byte for EOVERFLOW in SEEK_SET */
+ error = check_file_seek_range(&fl, offset);
+ if (error) {
+ goto outdrop;
+ }
+
+ if ((error = vnode_getwithref(vp))) {
+ goto outdrop;
+ }
+ if (fl.l_whence == SEEK_CUR) {
+ fl.l_start += offset;
+ }
+
+#if CONFIG_MACF
+ error = mac_file_check_lock(proc_ucred(p), fp->fp_glob,
+ F_SETLK, &fl);
+ if (error) {
+ (void)vnode_put(vp);
+ goto outdrop;
+ }
+#endif
+ switch (uap->cmd) {
+ case F_OFD_SETLK:
+ case F_OFD_SETLKW:
+ case F_OFD_SETLKWTIMEOUT:
+ flg |= F_OFD_LOCK;
+ switch (fl.l_type) {
+ case F_RDLCK:
+ if ((fflag & FREAD) == 0) {
+ error = EBADF;
+ break;
+ }
+ error = VNOP_ADVLOCK(vp, (caddr_t)fp->fp_glob,
+ F_SETLK, &fl, flg, &context, timeout);
+ break;
+ case F_WRLCK:
+ if ((fflag & FWRITE) == 0) {
+ error = EBADF;
+ break;
+ }
+ error = VNOP_ADVLOCK(vp, (caddr_t)fp->fp_glob,
+ F_SETLK, &fl, flg, &context, timeout);
+ break;
+ case F_UNLCK:
+ error = VNOP_ADVLOCK(vp, (caddr_t)fp->fp_glob,
+ F_UNLCK, &fl, F_OFD_LOCK, &context,
+ timeout);
+ break;
+ default:
+ error = EINVAL;
+ break;
+ }
+ if (0 == error &&
+ (F_RDLCK == fl.l_type || F_WRLCK == fl.l_type)) {
+ struct fileglob *fg = fp->fp_glob;
+
+ /*
+ * arrange F_UNLCK on last close (once
+ * set, FG_HAS_OFDLOCK is immutable)
+ */
+ if ((fg->fg_lflags & FG_HAS_OFDLOCK) == 0) {
+ lck_mtx_lock_spin(&fg->fg_lock);
+ fg->fg_lflags |= FG_HAS_OFDLOCK;
+ lck_mtx_unlock(&fg->fg_lock);
+ }
+ }
+ break;
+ default:
+ flg |= F_POSIX;
+ switch (fl.l_type) {
+ case F_RDLCK:
+ if ((fflag & FREAD) == 0) {
+ error = EBADF;
+ break;
+ }
+ // XXX UInt32 unsafe for LP64 kernel
+ os_atomic_or(&p->p_ladvflag, P_LADVLOCK, relaxed);
+ error = VNOP_ADVLOCK(vp, (caddr_t)p,
+ F_SETLK, &fl, flg, &context, timeout);
+ break;
+ case F_WRLCK:
+ if ((fflag & FWRITE) == 0) {
+ error = EBADF;
+ break;
+ }
+ // XXX UInt32 unsafe for LP64 kernel
+ os_atomic_or(&p->p_ladvflag, P_LADVLOCK, relaxed);
+ error = VNOP_ADVLOCK(vp, (caddr_t)p,
+ F_SETLK, &fl, flg, &context, timeout);
+ break;
+ case F_UNLCK:
+ error = VNOP_ADVLOCK(vp, (caddr_t)p,
+ F_UNLCK, &fl, F_POSIX, &context, timeout);
+ break;
+ default:
+ error = EINVAL;
+ break;
+ }
+ break;
+ }
+ (void) vnode_put(vp);
+ goto outdrop;
+
+ case F_GETLK:
+ case F_OFD_GETLK:
+ case F_GETLKPID:
+ case F_OFD_GETLKPID:
+ if (fp->f_type != DTYPE_VNODE) {
+ error = EBADF;
+ goto out;
+ }
+ vp = (struct vnode *)fp->f_data;
+
+ offset = fp->f_offset;
+ proc_fdunlock(p);
+
+ /* Copy in the lock structure */
+ error = copyin(argp, (caddr_t)&fl, sizeof(fl));
+ if (error) {
+ goto outdrop;
+ }
+
+ /* Check starting byte and ending byte for EOVERFLOW in SEEK_CUR */
+ /* and ending byte for EOVERFLOW in SEEK_SET */
+ error = check_file_seek_range(&fl, offset);
+ if (error) {
+ goto outdrop;
+ }
+
+ if ((fl.l_whence == SEEK_SET) && (fl.l_start < 0)) {
+ error = EINVAL;
+ goto outdrop;
+ }
+
+ switch (fl.l_type) {
+ case F_RDLCK:
+ case F_UNLCK:
+ case F_WRLCK:
+ break;
+ default:
+ error = EINVAL;
+ goto outdrop;
+ }
+
+ switch (fl.l_whence) {
+ case SEEK_CUR:
+ case SEEK_SET:
+ case SEEK_END:
+ break;
+ default:
+ error = EINVAL;
+ goto outdrop;
+ }
+
+ if ((error = vnode_getwithref(vp)) == 0) {
+ if (fl.l_whence == SEEK_CUR) {
+ fl.l_start += offset;
+ }
+
+#if CONFIG_MACF
+ error = mac_file_check_lock(proc_ucred(p), fp->fp_glob,
+ uap->cmd, &fl);
+ if (error == 0)
+#endif
+ switch (uap->cmd) {
+ case F_OFD_GETLK:
+ error = VNOP_ADVLOCK(vp, (caddr_t)fp->fp_glob,
+ F_GETLK, &fl, F_OFD_LOCK, &context, NULL);
+ break;
+ case F_OFD_GETLKPID:
+ error = VNOP_ADVLOCK(vp, (caddr_t)fp->fp_glob,
+ F_GETLKPID, &fl, F_OFD_LOCK, &context, NULL);
+ break;
+ default:
+ error = VNOP_ADVLOCK(vp, (caddr_t)p,
+ uap->cmd, &fl, F_POSIX, &context, NULL);
+ break;
+ }
+
+ (void)vnode_put(vp);
+
+ if (error == 0) {
+ error = copyout((caddr_t)&fl, argp, sizeof(fl));
+ }
+ }
+ goto outdrop;
+
+ case F_PREALLOCATE: {
+ fstore_t alloc_struct; /* structure for allocate command */
+ u_int32_t alloc_flags = 0;
+
+ if (fp->f_type != DTYPE_VNODE) {
+ error = EBADF;
+ goto out;
+ }
+
+ vp = (struct vnode *)fp->f_data;
+ proc_fdunlock(p);
+
+ /* make sure that we have write permission */
+ if ((fp->f_flag & FWRITE) == 0) {
+ error = EBADF;
+ goto outdrop;
+ }
+
+ error = copyin(argp, (caddr_t)&alloc_struct, sizeof(alloc_struct));
+ if (error) {
+ goto outdrop;
+ }
+
+ /* now set the space allocated to 0 */
+ alloc_struct.fst_bytesalloc = 0;
+
+ /*
+ * Do some simple parameter checking
+ */
+
+ /* set up the flags */
+
+ alloc_flags |= PREALLOCATE;
+
+ if (alloc_struct.fst_flags & F_ALLOCATECONTIG) {
+ alloc_flags |= ALLOCATECONTIG;
+ }
+
+ if (alloc_struct.fst_flags & F_ALLOCATEALL) {
+ alloc_flags |= ALLOCATEALL;
+ }
+
+ /*
+ * Do any position mode specific stuff. The only
+ * position mode supported now is PEOFPOSMODE
+ */
+
+ switch (alloc_struct.fst_posmode) {
+ case F_PEOFPOSMODE:
+ if (alloc_struct.fst_offset != 0) {
+ error = EINVAL;
+ goto outdrop;
+ }
+
+ alloc_flags |= ALLOCATEFROMPEOF;
+ break;
+
+ case F_VOLPOSMODE:
+ if (alloc_struct.fst_offset <= 0) {
+ error = EINVAL;
+ goto outdrop;
+ }
+
+ alloc_flags |= ALLOCATEFROMVOL;
+ break;
+
+ default: {
+ error = EINVAL;
+ goto outdrop;
+ }
+ }
+ if ((error = vnode_getwithref(vp)) == 0) {
+ /*
+ * call allocate to get the space
+ */
+ error = VNOP_ALLOCATE(vp, alloc_struct.fst_length, alloc_flags,
+ &alloc_struct.fst_bytesalloc, alloc_struct.fst_offset,
+ &context);
+ (void)vnode_put(vp);
+
+ error2 = copyout((caddr_t)&alloc_struct, argp, sizeof(alloc_struct));
+
+ if (error == 0) {
+ error = error2;
+ }
+ }
+ goto outdrop;
+ }
+ case F_PUNCHHOLE: {
+ fpunchhole_t args;
+
+ if (fp->f_type != DTYPE_VNODE) {
+ error = EBADF;
+ goto out;
+ }
+
+ vp = (struct vnode *)fp->f_data;
+ proc_fdunlock(p);
+
+ /* need write permissions */
+ if ((fp->f_flag & FWRITE) == 0) {
+ error = EPERM;
+ goto outdrop;
+ }
+
+ if ((error = copyin(argp, (caddr_t)&args, sizeof(args)))) {
+ goto outdrop;
+ }
+
+ if ((error = vnode_getwithref(vp))) {
+ goto outdrop;
+ }
+
+#if CONFIG_MACF
+ if ((error = mac_vnode_check_write(&context, fp->fp_glob->fg_cred, vp))) {
+ (void)vnode_put(vp);
+ goto outdrop;
+ }
+#endif
+
+ error = VNOP_IOCTL(vp, F_PUNCHHOLE, (caddr_t)&args, 0, &context);
+ (void)vnode_put(vp);
+
+ goto outdrop;
+ }
+ case F_TRIM_ACTIVE_FILE: {
+ ftrimactivefile_t args;
+
+ if (priv_check_cred(kauth_cred_get(), PRIV_TRIM_ACTIVE_FILE, 0)) {
+ error = EACCES;
+ goto out;
+ }
+
+ if (fp->f_type != DTYPE_VNODE) {
+ error = EBADF;
+ goto out;
+ }
+
+ vp = (struct vnode *)fp->f_data;
+ proc_fdunlock(p);
+
+ /* need write permissions */
+ if ((fp->f_flag & FWRITE) == 0) {
+ error = EPERM;
+ goto outdrop;
+ }
+
+ if ((error = copyin(argp, (caddr_t)&args, sizeof(args)))) {
+ goto outdrop;
+ }
+
+ if ((error = vnode_getwithref(vp))) {
+ goto outdrop;
+ }
+
+ error = VNOP_IOCTL(vp, F_TRIM_ACTIVE_FILE, (caddr_t)&args, 0, &context);
+ (void)vnode_put(vp);
+
+ goto outdrop;
+ }
+ case F_SPECULATIVE_READ: {
+ fspecread_t args;
+
+ if (fp->f_type != DTYPE_VNODE) {
+ error = EBADF;
+ goto out;
+ }
+
+ vp = (struct vnode *)fp->f_data;
+ proc_fdunlock(p);
+
+ if ((error = copyin(argp, (caddr_t)&args, sizeof(args)))) {
+ goto outdrop;
+ }
+
+ /* Discard invalid offsets or lengths */
+ if ((args.fsr_offset < 0) || (args.fsr_length < 0)) {
+ error = EINVAL;
+ goto outdrop;
+ }
+
+ /*
+ * Round the file offset down to a page-size boundary (or to 0).
+ * The filesystem will need to round the length up to the end of the page boundary
+ * or to the EOF of the file.
+ */
+ uint64_t foff = (((uint64_t)args.fsr_offset) & ~((uint64_t)PAGE_MASK));
+ uint64_t foff_delta = args.fsr_offset - foff;
+ args.fsr_offset = (off_t) foff;
+
+ /*
+ * Now add in the delta to the supplied length. Since we may have adjusted the
+ * offset, increase it by the amount that we adjusted.
+ */
+ args.fsr_length += foff_delta;
+
+ if ((error = vnode_getwithref(vp))) {
+ goto outdrop;
+ }
+ error = VNOP_IOCTL(vp, F_SPECULATIVE_READ, (caddr_t)&args, 0, &context);
+ (void)vnode_put(vp);
+
+ goto outdrop;
+ }
+ case F_SETSIZE:
+ if (fp->f_type != DTYPE_VNODE) {
+ error = EBADF;
+ goto out;
+ }
+ vp = (struct vnode *)fp->f_data;
+ proc_fdunlock(p);
+
+ error = copyin(argp, (caddr_t)&offset, sizeof(off_t));
+ if (error) {
+ goto outdrop;
+ }
+ AUDIT_ARG(value64, offset);
+
+ error = vnode_getwithref(vp);
+ if (error) {
+ goto outdrop;
+ }
+
+#if CONFIG_MACF
+ error = mac_vnode_check_truncate(&context,
+ fp->fp_glob->fg_cred, vp);
+ if (error) {
+ (void)vnode_put(vp);
+ goto outdrop;
+ }
+#endif
+ /*
+ * Make sure that we are root. Growing a file
+ * without zero filling the data is a security hole.
+ */
+ if (!kauth_cred_issuser(kauth_cred_get())) {
+ error = EACCES;
+ } else {
+ /*
+ * Require privilege to change file size without zerofill,
+ * else will change the file size and zerofill it.
+ */
+ error = priv_check_cred(kauth_cred_get(), PRIV_VFS_SETSIZE, 0);
+ if (error == 0) {
+ error = vnode_setsize(vp, offset, IO_NOZEROFILL, &context);
+ } else {
+ error = vnode_setsize(vp, offset, 0, &context);
+ }
+
+#if CONFIG_MACF
+ if (error == 0) {
+ mac_vnode_notify_truncate(&context, fp->fp_glob->fg_cred, vp);
+ }
+#endif
+ }
+
+ (void)vnode_put(vp);
+ goto outdrop;
+
+ case F_RDAHEAD:
+ if (fp->f_type != DTYPE_VNODE) {
+ error = EBADF;
+ goto out;
+ }
+ if (uap->arg) {
+ os_atomic_andnot(&fp->fp_glob->fg_flag, FNORDAHEAD, relaxed);
+ } else {
+ os_atomic_or(&fp->fp_glob->fg_flag, FNORDAHEAD, relaxed);
+ }
+ goto out;
+
+ case F_NOCACHE:
+ if (fp->f_type != DTYPE_VNODE) {
+ error = EBADF;
+ goto out;
+ }
+ if (uap->arg) {
+ os_atomic_or(&fp->fp_glob->fg_flag, FNOCACHE, relaxed);
+ } else {
+ os_atomic_andnot(&fp->fp_glob->fg_flag, FNOCACHE, relaxed);
+ }
+ goto out;
+
+ case F_NODIRECT:
+ if (fp->f_type != DTYPE_VNODE) {
+ error = EBADF;
+ goto out;
+ }
+ if (uap->arg) {
+ os_atomic_or(&fp->fp_glob->fg_flag, FNODIRECT, relaxed);
+ } else {
+ os_atomic_andnot(&fp->fp_glob->fg_flag, FNODIRECT, relaxed);
+ }
+ goto out;
+
+ case F_SINGLE_WRITER:
+ if (fp->f_type != DTYPE_VNODE) {
+ error = EBADF;
+ goto out;
+ }
+ if (uap->arg) {
+ os_atomic_or(&fp->fp_glob->fg_flag, FSINGLE_WRITER, relaxed);
+ } else {
+ os_atomic_andnot(&fp->fp_glob->fg_flag, FSINGLE_WRITER, relaxed);
+ }
+ goto out;
+
+ case F_GLOBAL_NOCACHE:
+ if (fp->f_type != DTYPE_VNODE) {
+ error = EBADF;
+ goto out;
+ }
+ vp = (struct vnode *)fp->f_data;
+ proc_fdunlock(p);
+
+ if ((error = vnode_getwithref(vp)) == 0) {
+ *retval = vnode_isnocache(vp);
+
+ if (uap->arg) {
+ vnode_setnocache(vp);
+ } else {
+ vnode_clearnocache(vp);
+ }
+
+ (void)vnode_put(vp);
+ }
+ goto outdrop;
+
+ case F_CHECK_OPENEVT:
+ if (fp->f_type != DTYPE_VNODE) {
+ error = EBADF;
+ goto out;
+ }
+ vp = (struct vnode *)fp->f_data;
+ proc_fdunlock(p);
+
+ if ((error = vnode_getwithref(vp)) == 0) {
+ *retval = vnode_is_openevt(vp);
+
+ if (uap->arg) {
+ vnode_set_openevt(vp);
+ } else {
+ vnode_clear_openevt(vp);
+ }
+
+ (void)vnode_put(vp);
+ }
+ goto outdrop;
+
+ case F_RDADVISE: {
+ struct radvisory ra_struct;
+
+ if (fp->f_type != DTYPE_VNODE) {
+ error = EBADF;
+ goto out;
+ }
+ vp = (struct vnode *)fp->f_data;
+ proc_fdunlock(p);
+
+ if ((error = copyin(argp, (caddr_t)&ra_struct, sizeof(ra_struct)))) {
+ goto outdrop;
+ }
+ if (ra_struct.ra_offset < 0 || ra_struct.ra_count < 0) {
+ error = EINVAL;
+ goto outdrop;
+ }
+ if ((error = vnode_getwithref(vp)) == 0) {
+ error = VNOP_IOCTL(vp, F_RDADVISE, (caddr_t)&ra_struct, 0, &context);
+
+ (void)vnode_put(vp);
+ }
+ goto outdrop;
+ }
+
+ case F_FLUSH_DATA:
+
+ if (fp->f_type != DTYPE_VNODE) {
+ error = EBADF;
+ goto out;
+ }
+ vp = (struct vnode *)fp->f_data;
+ proc_fdunlock(p);
+
+ if ((error = vnode_getwithref(vp)) == 0) {
+ error = VNOP_FSYNC(vp, MNT_NOWAIT, &context);
+
+ (void)vnode_put(vp);
+ }
+ goto outdrop;
+
+ case F_LOG2PHYS:
+ case F_LOG2PHYS_EXT: {
+ struct log2phys l2p_struct = {}; /* structure for allocate command */
+ int devBlockSize;
+
+ off_t file_offset = 0;
+ size_t a_size = 0;
+ size_t run = 0;
+
+ if (uap->cmd == F_LOG2PHYS_EXT) {
+ error = copyin(argp, (caddr_t)&l2p_struct, sizeof(l2p_struct));
+ if (error) {
+ goto out;
+ }
+ file_offset = l2p_struct.l2p_devoffset;
+ } else {
+ file_offset = fp->f_offset;
+ }
+ if (fp->f_type != DTYPE_VNODE) {
+ error = EBADF;
+ goto out;
+ }
+ vp = (struct vnode *)fp->f_data;
+ proc_fdunlock(p);
+ if ((error = vnode_getwithref(vp))) {
+ goto outdrop;
+ }
+ error = VNOP_OFFTOBLK(vp, file_offset, &lbn);
+ if (error) {
+ (void)vnode_put(vp);
+ goto outdrop;
+ }
+ error = VNOP_BLKTOOFF(vp, lbn, &offset);
+ if (error) {
+ (void)vnode_put(vp);
+ goto outdrop;
+ }
+ devBlockSize = vfs_devblocksize(vnode_mount(vp));
+ if (uap->cmd == F_LOG2PHYS_EXT) {
+ if (l2p_struct.l2p_contigbytes < 0) {
+ vnode_put(vp);
+ error = EINVAL;
+ goto outdrop;
+ }
+
+ a_size = (size_t)MIN((uint64_t)l2p_struct.l2p_contigbytes, SIZE_MAX);
+ } else {
+ a_size = devBlockSize;
+ }
+
+ error = VNOP_BLOCKMAP(vp, offset, a_size, &bn, &run, NULL, 0, &context);
+
+ (void)vnode_put(vp);
+
+ if (!error) {
+ l2p_struct.l2p_flags = 0; /* for now */
+ if (uap->cmd == F_LOG2PHYS_EXT) {
+ l2p_struct.l2p_contigbytes = run - (file_offset - offset);
+ } else {
+ l2p_struct.l2p_contigbytes = 0; /* for now */
+ }
+
+ /*
+ * The block number being -1 suggests that the file offset is not backed
+ * by any real blocks on-disk. As a result, just let it be passed back up wholesale.
+ */
+ if (bn == -1) {
+ /* Don't multiply it by the block size */
+ l2p_struct.l2p_devoffset = bn;
+ } else {
+ l2p_struct.l2p_devoffset = bn * devBlockSize;
+ l2p_struct.l2p_devoffset += file_offset - offset;
+ }
+ error = copyout((caddr_t)&l2p_struct, argp, sizeof(l2p_struct));
+ }
+ goto outdrop;
+ }
+ case F_GETPATH:
+ case F_GETPATH_NOFIRMLINK: {
+ char *pathbufp;
+ int pathlen;
+
+ if (fp->f_type != DTYPE_VNODE) {
+ error = EBADF;
+ goto out;
+ }
+ vp = (struct vnode *)fp->f_data;
+ proc_fdunlock(p);
+
+ pathlen = MAXPATHLEN;
+ pathbufp = zalloc(ZV_NAMEI);
+
+ if ((error = vnode_getwithref(vp)) == 0) {
+ if (uap->cmd == F_GETPATH_NOFIRMLINK) {
+ error = vn_getpath_ext(vp, NULL, pathbufp, &pathlen, VN_GETPATH_NO_FIRMLINK);
+ } else {
+ error = vn_getpath(vp, pathbufp, &pathlen);
+ }
+ (void)vnode_put(vp);
+
+ if (error == 0) {
+ error = copyout((caddr_t)pathbufp, argp, pathlen);
+ }
+ }
+ zfree(ZV_NAMEI, pathbufp);
+ goto outdrop;
+ }
+
+ case F_PATHPKG_CHECK: {
+ char *pathbufp;
+ size_t pathlen;
+
+ if (fp->f_type != DTYPE_VNODE) {
+ error = EBADF;
+ goto out;
+ }
+ vp = (struct vnode *)fp->f_data;
+ proc_fdunlock(p);
+
+ pathlen = MAXPATHLEN;
+ pathbufp = zalloc(ZV_NAMEI);
+
+ if ((error = copyinstr(argp, pathbufp, MAXPATHLEN, &pathlen)) == 0) {
+ if ((error = vnode_getwithref(vp)) == 0) {
+ AUDIT_ARG(text, pathbufp);
+ error = vn_path_package_check(vp, pathbufp, (int)pathlen, retval);
+
+ (void)vnode_put(vp);
+ }
+ }
+ zfree(ZV_NAMEI, pathbufp);
+ goto outdrop;
+ }
+
+ case F_CHKCLEAN: // used by regression tests to see if all dirty pages got cleaned by fsync()
+ case F_FULLFSYNC: // fsync + flush the journal + DKIOCSYNCHRONIZE
+ case F_BARRIERFSYNC: // fsync + barrier
+ case F_FREEZE_FS: // freeze all other fs operations for the fs of this fd
+ case F_THAW_FS: { // thaw all frozen fs operations for the fs of this fd
+ if (fp->f_type != DTYPE_VNODE) {
+ error = EBADF;
+ goto out;
+ }
+ vp = (struct vnode *)fp->f_data;
+ proc_fdunlock(p);
+
+ if ((error = vnode_getwithref(vp)) == 0) {
+ error = VNOP_IOCTL(vp, uap->cmd, (caddr_t)NULL, 0, &context);
+
+ (void)vnode_put(vp);
+ }
+ break;
+ }
+
+ /*
+ * SPI (private) for opening a file starting from a dir fd
+ */
+ case F_OPENFROM: {
+ struct user_fopenfrom fopen;
+ struct vnode_attr va;
+ struct nameidata nd;
+ int cmode;
+
+ /* Check if this isn't a valid file descriptor */
+ if ((fp->f_type != DTYPE_VNODE) ||
+ (fp->f_flag & FREAD) == 0) {
+ error = EBADF;
+ goto out;
+ }
+ vp = (struct vnode *)fp->f_data;
+ proc_fdunlock(p);
+
+ if (vnode_getwithref(vp)) {
+ error = ENOENT;
+ goto outdrop;
+ }
+
+ /* Only valid for directories */
+ if (vp->v_type != VDIR) {
+ vnode_put(vp);
+ error = ENOTDIR;
+ goto outdrop;
+ }
+
+ /*
+ * Only entitled apps may use the credentials of the thread
+ * that opened the file descriptor.
+ * Non-entitled threads will use their own context.
+ */
+ if (IOTaskHasEntitlement(current_task(), ACCOUNT_OPENFROM_ENTITLEMENT)) {
+ has_entitlement = 1;
+ }
+
+ /* Get flags, mode and pathname arguments. */
+ if (IS_64BIT_PROCESS(p)) {
+ error = copyin(argp, &fopen, sizeof(fopen));
+ } else {
+ struct user32_fopenfrom fopen32;
+
+ error = copyin(argp, &fopen32, sizeof(fopen32));
+ fopen.o_flags = fopen32.o_flags;
+ fopen.o_mode = fopen32.o_mode;
+ fopen.o_pathname = CAST_USER_ADDR_T(fopen32.o_pathname);
+ }
+ if (error) {
+ vnode_put(vp);
+ goto outdrop;
+ }
+ AUDIT_ARG(fflags, fopen.o_flags);
+ AUDIT_ARG(mode, fopen.o_mode);
+ VATTR_INIT(&va);
+ /* Mask off all but regular access permissions */
+ cmode = ((fopen.o_mode & ~fdp->fd_cmask) & ALLPERMS) & ~S_ISTXT;
+ VATTR_SET(&va, va_mode, cmode & ACCESSPERMS);
+
+ /* Start the lookup relative to the file descriptor's vnode. */
+ NDINIT(&nd, LOOKUP, OP_OPEN, USEDVP | FOLLOW | AUDITVNPATH1, UIO_USERSPACE,
+ fopen.o_pathname, has_entitlement ? &context : vfs_context_current());
+ nd.ni_dvp = vp;
+
+ error = open1(has_entitlement ? &context : vfs_context_current(),
+ &nd, fopen.o_flags, &va, fileproc_alloc_init, NULL, retval);
+
+ vnode_put(vp);
+ break;
+ }
+ /*
+ * SPI (private) for unlinking a file starting from a dir fd
+ */
+ case F_UNLINKFROM: {
+ user_addr_t pathname;
+
+ /* Check if this isn't a valid file descriptor */
+ if ((fp->f_type != DTYPE_VNODE) ||
+ (fp->f_flag & FREAD) == 0) {
+ error = EBADF;
+ goto out;
+ }
+ vp = (struct vnode *)fp->f_data;
+ proc_fdunlock(p);
+
+ if (vnode_getwithref(vp)) {
+ error = ENOENT;
+ goto outdrop;
+ }
+
+ /* Only valid for directories */
+ if (vp->v_type != VDIR) {
+ vnode_put(vp);
+ error = ENOTDIR;
+ goto outdrop;
+ }
+
+ /*
+ * Only entitled apps may use the credentials of the thread
+ * that opened the file descriptor.
+ * Non-entitled threads will use their own context.
+ */
+ if (IOTaskHasEntitlement(current_task(), ACCOUNT_OPENFROM_ENTITLEMENT)) {
+ has_entitlement = 1;
+ }
+
+ /* Get flags, mode and pathname arguments. */
+ if (IS_64BIT_PROCESS(p)) {
+ pathname = (user_addr_t)argp;
+ } else {
+ pathname = CAST_USER_ADDR_T(argp);
+ }
+
+ /* Start the lookup relative to the file descriptor's vnode. */
+ error = unlink1(has_entitlement ? &context : vfs_context_current(),
+ vp, pathname, UIO_USERSPACE, 0);
+
+ vnode_put(vp);
+ break;
+ }
+
+ case F_ADDSIGS:
+ case F_ADDFILESIGS:
+ case F_ADDFILESIGS_FOR_DYLD_SIM:
+ case F_ADDFILESIGS_RETURN:
+ case F_ADDFILESIGS_INFO:
+ {
+ struct cs_blob *blob = NULL;
+ struct user_fsignatures fs;
+ kern_return_t kr;
+ vm_offset_t kernel_blob_addr;
+ vm_size_t kernel_blob_size;
+ int blob_add_flags = 0;
+ const size_t sizeof_fs = (uap->cmd == F_ADDFILESIGS_INFO ?
+ offsetof(struct user_fsignatures, fs_cdhash /* first output element */) :
+ offsetof(struct user_fsignatures, fs_fsignatures_size /* compat */));
+
+ if (fp->f_type != DTYPE_VNODE) {
+ error = EBADF;
+ goto out;
+ }
+ vp = (struct vnode *)fp->f_data;
+ proc_fdunlock(p);
+
+ if (uap->cmd == F_ADDFILESIGS_FOR_DYLD_SIM) {
+ blob_add_flags |= MAC_VNODE_CHECK_DYLD_SIM;
+ if ((p->p_csflags & CS_KILL) == 0) {
+ proc_lock(p);
+ p->p_csflags |= CS_KILL;
+ proc_unlock(p);
+ }
+ }
+
+ error = vnode_getwithref(vp);
+ if (error) {
+ goto outdrop;
+ }
+
+ if (IS_64BIT_PROCESS(p)) {
+ error = copyin(argp, &fs, sizeof_fs);
+ } else {
+ if (uap->cmd == F_ADDFILESIGS_INFO) {
+ error = EINVAL;
+ vnode_put(vp);
+ goto outdrop;
+ }
+
+ struct user32_fsignatures fs32;
+
+ error = copyin(argp, &fs32, sizeof(fs32));
+ fs.fs_file_start = fs32.fs_file_start;
+ fs.fs_blob_start = CAST_USER_ADDR_T(fs32.fs_blob_start);
+ fs.fs_blob_size = fs32.fs_blob_size;
+ }
+
+ if (error) {
+ vnode_put(vp);
+ goto outdrop;
+ }
+
+ /*
+ * First check if we have something loaded a this offset
+ */
+ blob = ubc_cs_blob_get(vp, CPU_TYPE_ANY, CPU_SUBTYPE_ANY, fs.fs_file_start);
+ if (blob != NULL) {
+ /* If this is for dyld_sim revalidate the blob */
+ if (uap->cmd == F_ADDFILESIGS_FOR_DYLD_SIM) {
+ error = ubc_cs_blob_revalidate(vp, blob, NULL, blob_add_flags, proc_platform(p));
+ if (error) {
+ blob = NULL;
+ if (error != EAGAIN) {
+ vnode_put(vp);
+ goto outdrop;
+ }
+ }
+ }
+ }
+
+ if (blob == NULL) {
+ /*
+ * An arbitrary limit, to prevent someone from mapping in a 20GB blob. This should cover
+ * our use cases for the immediate future, but note that at the time of this commit, some
+ * platforms are nearing 2MB blob sizes (with a prior soft limit of 2.5MB).
+ *
+ * We should consider how we can manage this more effectively; the above means that some
+ * platforms are using megabytes of memory for signing data; it merely hasn't crossed the
+ * threshold considered ridiculous at the time of this change.
+ */
+#define CS_MAX_BLOB_SIZE (40ULL * 1024ULL * 1024ULL)
+ if (fs.fs_blob_size > CS_MAX_BLOB_SIZE) {
+ error = E2BIG;
+ vnode_put(vp);
+ goto outdrop;
+ }
+
+ kernel_blob_size = CAST_DOWN(vm_size_t, fs.fs_blob_size);
+ kr = ubc_cs_blob_allocate(&kernel_blob_addr, &kernel_blob_size);
+ if (kr != KERN_SUCCESS || kernel_blob_size < fs.fs_blob_size) {
+ error = ENOMEM;
+ vnode_put(vp);
+ goto outdrop;
+ }
+
+ if (uap->cmd == F_ADDSIGS) {
+ error = copyin(fs.fs_blob_start,
+ (void *) kernel_blob_addr,
+ fs.fs_blob_size);
+ } else { /* F_ADDFILESIGS || F_ADDFILESIGS_RETURN || F_ADDFILESIGS_FOR_DYLD_SIM || F_ADDFILESIGS_INFO */
+ int resid;
+
+ error = vn_rdwr(UIO_READ,
+ vp,
+ (caddr_t) kernel_blob_addr,
+ (int)kernel_blob_size,
+ fs.fs_file_start + fs.fs_blob_start,
+ UIO_SYSSPACE,
+ 0,
+ kauth_cred_get(),
+ &resid,
+ p);
+ if ((error == 0) && resid) {
+ /* kernel_blob_size rounded to a page size, but signature may be at end of file */
+ memset((void *)(kernel_blob_addr + (kernel_blob_size - resid)), 0x0, resid);
+ }
+ }
+
+ if (error) {
+ ubc_cs_blob_deallocate(kernel_blob_addr,
+ kernel_blob_size);
+ vnode_put(vp);
+ goto outdrop;
+ }
+
+ blob = NULL;
+ error = ubc_cs_blob_add(vp,
+ proc_platform(p),
+ CPU_TYPE_ANY, /* not for a specific architecture */
+ CPU_SUBTYPE_ANY,
+ fs.fs_file_start,
+ &kernel_blob_addr,
+ kernel_blob_size,
+ NULL,
+ blob_add_flags,
+ &blob);
+
+ /* ubc_blob_add() has consumed "kernel_blob_addr" if it is zeroed */
+ if (error) {
+ if (kernel_blob_addr) {
+ ubc_cs_blob_deallocate(kernel_blob_addr,
+ kernel_blob_size);
+ }
+ vnode_put(vp);
+ goto outdrop;
+ } else {
+#if CHECK_CS_VALIDATION_BITMAP
+ ubc_cs_validation_bitmap_allocate( vp );
+#endif
+ }
+ }
+
+ if (uap->cmd == F_ADDFILESIGS_RETURN || uap->cmd == F_ADDFILESIGS_FOR_DYLD_SIM ||
+ uap->cmd == F_ADDFILESIGS_INFO) {
+ /*
+ * The first element of the structure is a
+ * off_t that happen to have the same size for
+ * all archs. Lets overwrite that.
+ */
+ off_t end_offset = 0;
+ if (blob) {
+ end_offset = blob->csb_end_offset;
+ }
+ error = copyout(&end_offset, argp, sizeof(end_offset));
+
+ if (error) {
+ vnode_put(vp);
+ goto outdrop;
+ }
+ }
+
+ if (uap->cmd == F_ADDFILESIGS_INFO) {
+ /* Return information. What we copy out depends on the size of the
+ * passed in structure, to keep binary compatibility. */
+
+ if (fs.fs_fsignatures_size >= sizeof(struct user_fsignatures)) {
+ // enough room for fs_cdhash[20]+fs_hash_type
+
+ if (blob != NULL) {
+ error = copyout(blob->csb_cdhash,
+ (vm_address_t)argp + offsetof(struct user_fsignatures, fs_cdhash),
+ USER_FSIGNATURES_CDHASH_LEN);
+ if (error) {
+ vnode_put(vp);
+ goto outdrop;
+ }
+ int hashtype = cs_hash_type(blob->csb_hashtype);
+ error = copyout(&hashtype,
+ (vm_address_t)argp + offsetof(struct user_fsignatures, fs_hash_type),
+ sizeof(int));
+ if (error) {
+ vnode_put(vp);
+ goto outdrop;
+ }
+ }
+ }
+ }
+
+ (void) vnode_put(vp);
+ break;
+ }
+#if CONFIG_SUPPLEMENTAL_SIGNATURES
+ case F_ADDFILESUPPL:
+ {
+ struct vnode *ivp;
+ struct cs_blob *blob = NULL;
+ struct user_fsupplement fs;
+ int orig_fd;
+ struct fileproc* orig_fp = NULL;
+ kern_return_t kr;
+ vm_offset_t kernel_blob_addr;
+ vm_size_t kernel_blob_size;
+
+ if (!IS_64BIT_PROCESS(p)) {
+ error = EINVAL;
+ goto out; // drop fp and unlock fds
+ }
+
+ if (fp->f_type != DTYPE_VNODE) {
+ error = EBADF;
+ goto out;
+ }
+
+ error = copyin(argp, &fs, sizeof(fs));
+ if (error) {
+ goto out;
+ }
+
+ orig_fd = fs.fs_orig_fd;
+ if ((error = fp_lookup(p, orig_fd, &orig_fp, 1))) {
+ printf("CODE SIGNING: Failed to find original file for supplemental signature attachment\n");
+ goto out;
+ }
+
+ if (orig_fp->f_type != DTYPE_VNODE) {
+ error = EBADF;
+ fp_drop(p, orig_fd, orig_fp, 1);
+ goto out;
+ }
+
+ ivp = (struct vnode *)orig_fp->f_data;
+
+ vp = (struct vnode *)fp->f_data;
+
+ proc_fdunlock(p);
+
+ error = vnode_getwithref(ivp);
+ if (error) {
+ fp_drop(p, orig_fd, orig_fp, 0);
+ goto outdrop; //drop fp
+ }
+
+ error = vnode_getwithref(vp);
+ if (error) {
+ vnode_put(ivp);
+ fp_drop(p, orig_fd, orig_fp, 0);
+ goto outdrop;
+ }
+
+ if (fs.fs_blob_size > CS_MAX_BLOB_SIZE) {
+ error = E2BIG;
+ goto dropboth; // drop iocounts on vp and ivp, drop orig_fp then drop fp via outdrop
+ }
+
+ kernel_blob_size = CAST_DOWN(vm_size_t, fs.fs_blob_size);
+ kr = ubc_cs_blob_allocate(&kernel_blob_addr, &kernel_blob_size);
+ if (kr != KERN_SUCCESS) {
+ error = ENOMEM;
+ goto dropboth;
+ }
+
+ int resid;
+ error = vn_rdwr(UIO_READ, vp,
+ (caddr_t)kernel_blob_addr, (int)kernel_blob_size,
+ fs.fs_file_start + fs.fs_blob_start,
+ UIO_SYSSPACE, 0,
+ kauth_cred_get(), &resid, p);
+ if ((error == 0) && resid) {
+ /* kernel_blob_size rounded to a page size, but signature may be at end of file */
+ memset((void *)(kernel_blob_addr + (kernel_blob_size - resid)), 0x0, resid);
+ }
+
+ if (error) {
+ ubc_cs_blob_deallocate(kernel_blob_addr,
+ kernel_blob_size);
+ goto dropboth;
+ }
+
+ error = ubc_cs_blob_add_supplement(vp, ivp, fs.fs_file_start,
+ &kernel_blob_addr, kernel_blob_size, &blob);
+
+ /* ubc_blob_add_supplement() has consumed kernel_blob_addr if it is zeroed */
+ if (error) {
+ if (kernel_blob_addr) {
+ ubc_cs_blob_deallocate(kernel_blob_addr,
+ kernel_blob_size);
+ }
+ goto dropboth;
+ }
+ vnode_put(ivp);
+ vnode_put(vp);
+ fp_drop(p, orig_fd, orig_fp, 0);
+ break;
+
+dropboth:
+ vnode_put(ivp);
+ vnode_put(vp);
+ fp_drop(p, orig_fd, orig_fp, 0);
+ goto outdrop;
+ }
+#endif
+ case F_GETCODEDIR:
+ case F_FINDSIGS: {
+ error = ENOTSUP;
+ goto out;
+ }
+ case F_CHECK_LV: {
+ struct fileglob *fg;
+ fchecklv_t lv = {};
+
+ if (fp->f_type != DTYPE_VNODE) {
+ error = EBADF;
+ goto out;
+ }
+ fg = fp->fp_glob;
+ proc_fdunlock(p);
+
+ if (IS_64BIT_PROCESS(p)) {
+ error = copyin(argp, &lv, sizeof(lv));
+ } else {
+ struct user32_fchecklv lv32 = {};
+
+ error = copyin(argp, &lv32, sizeof(lv32));
+ lv.lv_file_start = lv32.lv_file_start;
+ lv.lv_error_message = (void *)(uintptr_t)lv32.lv_error_message;
+ lv.lv_error_message_size = lv32.lv_error_message_size;
+ }
+ if (error) {
+ goto outdrop;
+ }
+
+#if CONFIG_MACF
+ error = mac_file_check_library_validation(p, fg, lv.lv_file_start,
+ (user_long_t)lv.lv_error_message, lv.lv_error_message_size);
+#endif
+
+ break;
+ }
+ case F_GETSIGSINFO: {
+ struct cs_blob *blob = NULL;
+ fgetsigsinfo_t sigsinfo = {};
+
+ if (fp->f_type != DTYPE_VNODE) {
+ error = EBADF;
+ goto out;
+ }
+ vp = (struct vnode *)fp->f_data;
+ proc_fdunlock(p);
+
+ error = vnode_getwithref(vp);
+ if (error) {
+ goto outdrop;
+ }
+
+ error = copyin(argp, &sigsinfo, sizeof(sigsinfo));
+ if (error) {
+ vnode_put(vp);
+ goto outdrop;
+ }
+
+ blob = ubc_cs_blob_get(vp, CPU_TYPE_ANY, CPU_SUBTYPE_ANY, sigsinfo.fg_file_start);
+ if (blob == NULL) {
+ error = ENOENT;
+ vnode_put(vp);
+ goto outdrop;
+ }
+ switch (sigsinfo.fg_info_request) {
+ case GETSIGSINFO_PLATFORM_BINARY:
+ sigsinfo.fg_sig_is_platform = blob->csb_platform_binary;
+ error = copyout(&sigsinfo.fg_sig_is_platform,
+ (vm_address_t)argp + offsetof(struct fgetsigsinfo, fg_sig_is_platform),
+ sizeof(sigsinfo.fg_sig_is_platform));
+ if (error) {
+ vnode_put(vp);
+ goto outdrop;
+ }
+ break;
+ default:
+ error = EINVAL;
+ vnode_put(vp);
+ goto outdrop;
+ }
+ vnode_put(vp);
+ break;
+ }
+#if CONFIG_PROTECT
+ case F_GETPROTECTIONCLASS: {
+ if (fp->f_type != DTYPE_VNODE) {
+ error = EBADF;
+ goto out;
+ }
+ vp = (struct vnode *)fp->f_data;
+
+ proc_fdunlock(p);
+
+ if (vnode_getwithref(vp)) {
+ error = ENOENT;
+ goto outdrop;
+ }
+
+ struct vnode_attr va;
+
+ VATTR_INIT(&va);
+ VATTR_WANTED(&va, va_dataprotect_class);
+ error = VNOP_GETATTR(vp, &va, &context);
+ if (!error) {
+ if (VATTR_IS_SUPPORTED(&va, va_dataprotect_class)) {
+ *retval = va.va_dataprotect_class;
+ } else {
+ error = ENOTSUP;
+ }
+ }
+
+ vnode_put(vp);
+ break;
+ }
+
+ case F_SETPROTECTIONCLASS: {
+ /* tmp must be a valid PROTECTION_CLASS_* */
+ tmp = CAST_DOWN_EXPLICIT(uint32_t, uap->arg);
+
+ if (fp->f_type != DTYPE_VNODE) {
+ error = EBADF;
+ goto out;
+ }
+ vp = (struct vnode *)fp->f_data;
+
+ proc_fdunlock(p);
+
+ if (vnode_getwithref(vp)) {
+ error = ENOENT;
+ goto outdrop;
+ }
+
+ /* Only go forward if you have write access */
+ vfs_context_t ctx = vfs_context_current();
+ if (vnode_authorize(vp, NULLVP, (KAUTH_VNODE_ACCESS | KAUTH_VNODE_WRITE_DATA), ctx) != 0) {
+ vnode_put(vp);
+ error = EBADF;
+ goto outdrop;
+ }
+
+ struct vnode_attr va;
+
+ VATTR_INIT(&va);
+ VATTR_SET(&va, va_dataprotect_class, tmp);
+
+ error = VNOP_SETATTR(vp, &va, ctx);
+
+ vnode_put(vp);
+ break;
+ }
+
+ case F_TRANSCODEKEY: {
+ if (fp->f_type != DTYPE_VNODE) {
+ error = EBADF;
+ goto out;
+ }
+
+ vp = (struct vnode *)fp->f_data;
+ proc_fdunlock(p);
+
+ if (vnode_getwithref(vp)) {
+ error = ENOENT;
+ goto outdrop;
+ }
+
+ cp_key_t k = {
+ .len = CP_MAX_WRAPPEDKEYSIZE,
+ };
+
+ k.key = kheap_alloc(KHEAP_TEMP, CP_MAX_WRAPPEDKEYSIZE, Z_WAITOK | Z_ZERO);
+ if (k.key == NULL) {
+ error = ENOMEM;
+ } else {
+ error = VNOP_IOCTL(vp, F_TRANSCODEKEY, (caddr_t)&k, 1, &context);
+ }
+
+ vnode_put(vp);
+
+ if (error == 0) {
+ error = copyout(k.key, argp, k.len);
+ *retval = k.len;
+ }
+
+ kheap_free(KHEAP_TEMP, k.key, CP_MAX_WRAPPEDKEYSIZE);
+
+ break;
+ }
+
+ case F_GETPROTECTIONLEVEL: {
+ if (fp->f_type != DTYPE_VNODE) {
+ error = EBADF;
+ goto out;
+ }
+
+ vp = (struct vnode*) fp->f_data;
+ proc_fdunlock(p);
+
+ if (vnode_getwithref(vp)) {
+ error = ENOENT;
+ goto outdrop;
+ }
+
+ error = VNOP_IOCTL(vp, F_GETPROTECTIONLEVEL, (caddr_t)retval, 0, &context);
+
+ vnode_put(vp);
+ break;
+ }
+
+ case F_GETDEFAULTPROTLEVEL: {
+ if (fp->f_type != DTYPE_VNODE) {
+ error = EBADF;
+ goto out;
+ }
+
+ vp = (struct vnode*) fp->f_data;
+ proc_fdunlock(p);
+
+ if (vnode_getwithref(vp)) {
+ error = ENOENT;
+ goto outdrop;
+ }
+
+ /*
+ * if cp_get_major_vers fails, error will be set to proper errno
+ * and cp_version will still be 0.
+ */
+
+ error = VNOP_IOCTL(vp, F_GETDEFAULTPROTLEVEL, (caddr_t)retval, 0, &context);
+
+ vnode_put(vp);
+ break;
+ }
+
+#endif /* CONFIG_PROTECT */
+
+ case F_MOVEDATAEXTENTS: {
+ struct fileproc *fp2 = NULL;
+ struct vnode *src_vp = NULLVP;
+ struct vnode *dst_vp = NULLVP;
+ /* We need to grab the 2nd FD out of the argments before moving on. */
+ int fd2 = CAST_DOWN_EXPLICIT(int32_t, uap->arg);
+
+ error = priv_check_cred(kauth_cred_get(), PRIV_VFS_MOVE_DATA_EXTENTS, 0);
+ if (error) {
+ goto out;
+ }
+
+ if (fp->f_type != DTYPE_VNODE) {
+ error = EBADF;
+ goto out;
+ }
+
+ /*
+ * For now, special case HFS+ and APFS only, since this
+ * is SPI.
+ */
+ src_vp = (struct vnode *)fp->f_data;
+ if (src_vp->v_tag != VT_HFS && src_vp->v_tag != VT_APFS) {
+ error = ENOTSUP;
+ goto out;
+ }
+
+ /*
+ * Get the references before we start acquiring iocounts on the vnodes,
+ * while we still hold the proc fd lock
+ */
+ if ((error = fp_lookup(p, fd2, &fp2, 1))) {
+ error = EBADF;
+ goto out;
+ }
+ if (fp2->f_type != DTYPE_VNODE) {
+ fp_drop(p, fd2, fp2, 1);
+ error = EBADF;
+ goto out;
+ }
+ dst_vp = (struct vnode *)fp2->f_data;
+ if (dst_vp->v_tag != VT_HFS && dst_vp->v_tag != VT_APFS) {
+ fp_drop(p, fd2, fp2, 1);
+ error = ENOTSUP;
+ goto out;
+ }
+
+#if CONFIG_MACF
+ /* Re-do MAC checks against the new FD, pass in a fake argument */
+ error = mac_file_check_fcntl(proc_ucred(p), fp2->fp_glob, uap->cmd, 0);
+ if (error) {
+ fp_drop(p, fd2, fp2, 1);
+ goto out;
+ }
+#endif
+ /* Audit the 2nd FD */
+ AUDIT_ARG(fd, fd2);
+
+ proc_fdunlock(p);
+
+ if (vnode_getwithref(src_vp)) {
+ fp_drop(p, fd2, fp2, 0);
+ error = ENOENT;
+ goto outdrop;
+ }
+ if (vnode_getwithref(dst_vp)) {
+ vnode_put(src_vp);
+ fp_drop(p, fd2, fp2, 0);
+ error = ENOENT;
+ goto outdrop;
+ }
+
+ /*
+ * Basic asserts; validate they are not the same and that
+ * both live on the same filesystem.
+ */
+ if (dst_vp == src_vp) {
+ vnode_put(src_vp);
+ vnode_put(dst_vp);
+ fp_drop(p, fd2, fp2, 0);
+ error = EINVAL;
+ goto outdrop;
+ }
+
+ if (dst_vp->v_mount != src_vp->v_mount) {
+ vnode_put(src_vp);
+ vnode_put(dst_vp);
+ fp_drop(p, fd2, fp2, 0);
+ error = EXDEV;
+ goto outdrop;
+ }
+
+ /* Now we have a legit pair of FDs. Go to work */
+
+ /* Now check for write access to the target files */
+ if (vnode_authorize(src_vp, NULLVP,
+ (KAUTH_VNODE_ACCESS | KAUTH_VNODE_WRITE_DATA), &context) != 0) {
+ vnode_put(src_vp);
+ vnode_put(dst_vp);
+ fp_drop(p, fd2, fp2, 0);
+ error = EBADF;
+ goto outdrop;
+ }
+
+ if (vnode_authorize(dst_vp, NULLVP,
+ (KAUTH_VNODE_ACCESS | KAUTH_VNODE_WRITE_DATA), &context) != 0) {
+ vnode_put(src_vp);
+ vnode_put(dst_vp);
+ fp_drop(p, fd2, fp2, 0);
+ error = EBADF;
+ goto outdrop;
+ }
+
+ /* Verify that both vps point to files and not directories */
+ if (!vnode_isreg(src_vp) || !vnode_isreg(dst_vp)) {
+ error = EINVAL;
+ vnode_put(src_vp);
+ vnode_put(dst_vp);
+ fp_drop(p, fd2, fp2, 0);
+ goto outdrop;
+ }
+
+ /*
+ * The exchangedata syscall handler passes in 0 for the flags to VNOP_EXCHANGE.
+ * We'll pass in our special bit indicating that the new behavior is expected
+ */
+
+ error = VNOP_EXCHANGE(src_vp, dst_vp, FSOPT_EXCHANGE_DATA_ONLY, &context);
+
+ vnode_put(src_vp);
+ vnode_put(dst_vp);
+ fp_drop(p, fd2, fp2, 0);
+ break;
+ }
+
+ /*
+ * SPI for making a file compressed.
+ */
+ case F_MAKECOMPRESSED: {
+ uint32_t gcounter = CAST_DOWN_EXPLICIT(uint32_t, uap->arg);
+
+ if (fp->f_type != DTYPE_VNODE) {
+ error = EBADF;
+ goto out;
+ }
+
+ vp = (struct vnode*) fp->f_data;
+ proc_fdunlock(p);
+
+ /* get the vnode */
+ if (vnode_getwithref(vp)) {
+ error = ENOENT;
+ goto outdrop;
+ }
+
+ /* Is it a file? */
+ if ((vnode_isreg(vp) == 0) && (vnode_islnk(vp) == 0)) {
+ vnode_put(vp);
+ error = EBADF;
+ goto outdrop;
+ }
+
+ /* invoke ioctl to pass off to FS */
+ /* Only go forward if you have write access */
+ vfs_context_t ctx = vfs_context_current();
+ if (vnode_authorize(vp, NULLVP, (KAUTH_VNODE_ACCESS | KAUTH_VNODE_WRITE_DATA), ctx) != 0) {
+ vnode_put(vp);
+ error = EBADF;
+ goto outdrop;
+ }
+
+ error = VNOP_IOCTL(vp, uap->cmd, (caddr_t)&gcounter, 0, &context);
+
+ vnode_put(vp);
+ break;
+ }
+
+ /*
+ * SPI (private) for indicating to a filesystem that subsequent writes to
+ * the open FD will written to the Fastflow.
+ */
+ case F_SET_GREEDY_MODE:
+ /* intentionally drop through to the same handler as F_SETSTATIC.
+ * both fcntls should pass the argument and their selector into VNOP_IOCTL.
+ */
+
+ /*
+ * SPI (private) for indicating to a filesystem that subsequent writes to
+ * the open FD will represent static content.
+ */
+ case F_SETSTATICCONTENT: {
+ caddr_t ioctl_arg = NULL;
+
+ if (uap->arg) {
+ ioctl_arg = (caddr_t) 1;
+ }
+
+ if (fp->f_type != DTYPE_VNODE) {
+ error = EBADF;
+ goto out;
+ }
+ vp = (struct vnode *)fp->f_data;
+ proc_fdunlock(p);
+
+ error = vnode_getwithref(vp);
+ if (error) {
+ error = ENOENT;
+ goto outdrop;
+ }
+
+ /* Only go forward if you have write access */
+ vfs_context_t ctx = vfs_context_current();
+ if (vnode_authorize(vp, NULLVP, (KAUTH_VNODE_ACCESS | KAUTH_VNODE_WRITE_DATA), ctx) != 0) {
+ vnode_put(vp);
+ error = EBADF;
+ goto outdrop;
+ }
+
+ error = VNOP_IOCTL(vp, uap->cmd, ioctl_arg, 0, &context);
+ (void)vnode_put(vp);
+
+ break;
+ }
+
+ /*
+ * SPI (private) for indicating to the lower level storage driver that the
+ * subsequent writes should be of a particular IO type (burst, greedy, static),
+ * or other flavors that may be necessary.
+ */
+ case F_SETIOTYPE: {
+ caddr_t param_ptr;
+ uint32_t param;
+
+ if (uap->arg) {
+ /* extract 32 bits of flags from userland */
+ param_ptr = (caddr_t) uap->arg;
+ param = (uint32_t) param_ptr;
+ } else {
+ /* If no argument is specified, error out */
+ error = EINVAL;
+ goto out;
+ }
+
+ /*
+ * Validate the different types of flags that can be specified:
+ * all of them are mutually exclusive for now.
+ */
+ switch (param) {
+ case F_IOTYPE_ISOCHRONOUS:
+ break;
+
+ default:
+ error = EINVAL;
+ goto out;
+ }
+
+
+ if (fp->f_type != DTYPE_VNODE) {
+ error = EBADF;
+ goto out;
+ }
+ vp = (struct vnode *)fp->f_data;
+ proc_fdunlock(p);
+
+ error = vnode_getwithref(vp);
+ if (error) {
+ error = ENOENT;
+ goto outdrop;
+ }
+
+ /* Only go forward if you have write access */
+ vfs_context_t ctx = vfs_context_current();
+ if (vnode_authorize(vp, NULLVP, (KAUTH_VNODE_ACCESS | KAUTH_VNODE_WRITE_DATA), ctx) != 0) {
+ vnode_put(vp);
+ error = EBADF;
+ goto outdrop;
+ }
+
+ error = VNOP_IOCTL(vp, uap->cmd, param_ptr, 0, &context);
+ (void)vnode_put(vp);
+
+ break;
+ }
+
+ /*
+ * Set the vnode pointed to by 'fd'
+ * and tag it as the (potentially future) backing store
+ * for another filesystem
+ */
+ case F_SETBACKINGSTORE: {
+ if (fp->f_type != DTYPE_VNODE) {
+ error = EBADF;
+ goto out;
+ }
+
+ vp = (struct vnode *)fp->f_data;
+
+ if (vp->v_tag != VT_HFS) {
+ error = EINVAL;
+ goto out;
+ }
+ proc_fdunlock(p);
+
+ if (vnode_getwithref(vp)) {
+ error = ENOENT;
+ goto outdrop;
+ }
+
+ /* only proceed if you have write access */
+ vfs_context_t ctx = vfs_context_current();
+ if (vnode_authorize(vp, NULLVP, (KAUTH_VNODE_ACCESS | KAUTH_VNODE_WRITE_DATA), ctx) != 0) {
+ vnode_put(vp);
+ error = EBADF;
+ goto outdrop;
+ }
+
+
+ /* If arg != 0, set, otherwise unset */
+ if (uap->arg) {
+ error = VNOP_IOCTL(vp, uap->cmd, (caddr_t)1, 0, &context);
+ } else {
+ error = VNOP_IOCTL(vp, uap->cmd, (caddr_t)NULL, 0, &context);
+ }
+
+ vnode_put(vp);
+ break;
+ }
+
+ /*
+ * like F_GETPATH, but special semantics for
+ * the mobile time machine handler.
+ */
+ case F_GETPATH_MTMINFO: {
+ char *pathbufp;
+ int pathlen;
+
+ if (fp->f_type != DTYPE_VNODE) {
+ error = EBADF;
+ goto out;
+ }
+ vp = (struct vnode *)fp->f_data;
+ proc_fdunlock(p);
+
+ pathlen = MAXPATHLEN;
+ pathbufp = zalloc(ZV_NAMEI);
+
+ if ((error = vnode_getwithref(vp)) == 0) {
+ int backingstore = 0;
+
+ /* Check for error from vn_getpath before moving on */
+ if ((error = vn_getpath(vp, pathbufp, &pathlen)) == 0) {
+ if (vp->v_tag == VT_HFS) {
+ error = VNOP_IOCTL(vp, uap->cmd, (caddr_t) &backingstore, 0, &context);
+ }
+ (void)vnode_put(vp);
+
+ if (error == 0) {
+ error = copyout((caddr_t)pathbufp, argp, pathlen);
+ }
+ if (error == 0) {
+ /*
+ * If the copyout was successful, now check to ensure
+ * that this vnode is not a BACKINGSTORE vnode. mtmd
+ * wants the path regardless.
+ */
+ if (backingstore) {
+ error = EBUSY;
+ }
+ }
+ } else {
+ (void)vnode_put(vp);
+ }
+ }
+
+ zfree(ZV_NAMEI, pathbufp);
+ goto outdrop;
+ }
+
+#if DEBUG || DEVELOPMENT
+ case F_RECYCLE:
+ if (fp->f_type != DTYPE_VNODE) {
+ error = EBADF;
+ goto out;
+ }
+ vp = (struct vnode *)fp->f_data;
+ proc_fdunlock(p);
+
+ vnode_recycle(vp);
+ break;
+#endif
+
+ default:
+ /*
+ * This is an fcntl() that we d not recognize at this level;
+ * if this is a vnode, we send it down into the VNOP_IOCTL
+ * for this vnode; this can include special devices, and will
+ * effectively overload fcntl() to send ioctl()'s.
+ */
+ if ((uap->cmd & IOC_VOID) && (uap->cmd & IOC_INOUT)) {
+ error = EINVAL;
+ goto out;
+ }
+
+ /* Catch any now-invalid fcntl() selectors */
+ switch (uap->cmd) {
+ case (int)APFSIOC_REVERT_TO_SNAPSHOT:
+ case (int)FSIOC_FIOSEEKHOLE:
+ case (int)FSIOC_FIOSEEKDATA:
+ case (int)FSIOC_CAS_BSDFLAGS:
+ case HFS_GET_BOOT_INFO:
+ case HFS_SET_BOOT_INFO:
+ case FIOPINSWAP:
+ case F_MARKDEPENDENCY:
+ case TIOCREVOKE:
+ case TIOCREVOKECLEAR:
+ error = EINVAL;
+ goto out;
+ default:
+ break;
+ }
+
+ if (fp->f_type != DTYPE_VNODE) {
+ error = EBADF;
+ goto out;
+ }
+ vp = (struct vnode *)fp->f_data;
+ proc_fdunlock(p);
+
+ if ((error = vnode_getwithref(vp)) == 0) {
+#define STK_PARAMS 128
+ char stkbuf[STK_PARAMS] = {0};
+ unsigned int size;
+ caddr_t data, memp;
+ /*
+ * For this to work properly, we have to copy in the
+ * ioctl() cmd argument if there is one; we must also
+ * check that a command parameter, if present, does
+ * not exceed the maximum command length dictated by
+ * the number of bits we have available in the command
+ * to represent a structure length. Finally, we have
+ * to copy the results back out, if it is that type of
+ * ioctl().
+ */
+ size = IOCPARM_LEN(uap->cmd);
+ if (size > IOCPARM_MAX) {
+ (void)vnode_put(vp);
+ error = EINVAL;
+ break;
+ }
+
+ memp = NULL;
+ if (size > sizeof(stkbuf)) {
+ memp = (caddr_t)kheap_alloc(KHEAP_TEMP, size, Z_WAITOK);
+ if (memp == 0) {
+ (void)vnode_put(vp);
+ error = ENOMEM;
+ goto outdrop;
+ }
+ data = memp;
+ } else {
+ data = &stkbuf[0];
+ }
+
+ if (uap->cmd & IOC_IN) {
+ if (size) {
+ /* structure */
+ error = copyin(argp, data, size);
+ if (error) {
+ (void)vnode_put(vp);
+ if (memp) {
+ kheap_free(KHEAP_TEMP, memp, size);
+ }
+ goto outdrop;
+ }
+
+ /* Bzero the section beyond that which was needed */
+ if (size <= sizeof(stkbuf)) {
+ bzero((((uint8_t*)data) + size), (sizeof(stkbuf) - size));
+ }
+ } else {
+ /* int */
+ if (is64bit) {
+ *(user_addr_t *)data = argp;
+ } else {
+ *(uint32_t *)data = (uint32_t)argp;
+ }
+ };
+ } else if ((uap->cmd & IOC_OUT) && size) {
+ /*
+ * Zero the buffer so the user always
+ * gets back something deterministic.
+ */
+ bzero(data, size);
+ } else if (uap->cmd & IOC_VOID) {
+ if (is64bit) {
+ *(user_addr_t *)data = argp;
+ } else {
+ *(uint32_t *)data = (uint32_t)argp;
+ }
+ }
+
+ error = VNOP_IOCTL(vp, uap->cmd, CAST_DOWN(caddr_t, data), 0, &context);
+
+ (void)vnode_put(vp);
+
+ /* Copy any output data to user */
+ if (error == 0 && (uap->cmd & IOC_OUT) && size) {
+ error = copyout(data, argp, size);
+ }
+ if (memp) {
+ kheap_free(KHEAP_TEMP, memp, size);
+ }
+ }
+ break;
+ }
+
+outdrop:
+ AUDIT_ARG(vnpath_withref, vp, ARG_VNODE1);
+ fp_drop(p, fd, fp, 0);
+ return error;
+out:
+ fp_drop(p, fd, fp, 1);
+ proc_fdunlock(p);
+ return error;
+}
+
+
+/*
+ * finishdup
+ *
+ * Description: Common code for dup, dup2, and fcntl(F_DUPFD).
+ *
+ * Parameters: p Process performing the dup
+ * old The fd to dup
+ * new The fd to dup it to
+ * fd_flags Flags to augment the new fd
+ * retval Pointer to the call return area
+ *
+ * Returns: 0 Success
+ * EBADF
+ * ENOMEM
+ *
+ * Implicit returns:
+ * *retval (modified) The new descriptor
+ *
+ * Locks: Assumes proc_fdlock for process pointing to fdp is held by
+ * the caller
+ *
+ * Notes: This function may drop and reacquire this lock; it is unsafe
+ * for a caller to assume that other state protected by the lock
+ * has not been subsequently changed out from under it.
+ */
+int
+finishdup(proc_t p,
+ struct filedesc *fdp, int old, int new, int fd_flags, int32_t *retval)
+{
+ struct fileproc *nfp;
+ struct fileproc *ofp;
+#if CONFIG_MACF
+ int error;
+#endif
+
+#if DIAGNOSTIC
+ proc_fdlock_assert(p, LCK_MTX_ASSERT_OWNED);
+#endif
+ if ((ofp = fdp->fd_ofiles[old]) == NULL ||
+ (fdp->fd_ofileflags[old] & UF_RESERVED)) {
+ fdrelse(p, new);
+ return EBADF;
+ }
+
+#if CONFIG_MACF
+ error = mac_file_check_dup(proc_ucred(p), ofp->fp_glob, new);
+ if (error) {
+ fdrelse(p, new);
+ return error;
+ }
+#endif
+
+ proc_fdunlock(p);
+
+ nfp = fileproc_alloc_init(NULL);
+
+ proc_fdlock(p);
+
+ if (nfp == NULL) {
+ fdrelse(p, new);
+ return ENOMEM;
+ }
+
+ fg_ref(p, ofp->fp_glob);
+ nfp->fp_glob = ofp->fp_glob;
+
+#if DIAGNOSTIC
+ if (fdp->fd_ofiles[new] != 0) {
+ panic("finishdup: overwriting fd_ofiles with new %d", new);
+ }
+ if ((fdp->fd_ofileflags[new] & UF_RESERVED) == 0) {
+ panic("finishdup: unreserved fileflags with new %d", new);
+ }
+#endif
+
+ if (new > fdp->fd_lastfile) {
+ fdp->fd_lastfile = new;
+ }
+ *fdflags(p, new) |= fd_flags;
+ procfdtbl_releasefd(p, new, nfp);
+ *retval = new;
+ return 0;
+}
+
+
+/*
+ * sys_close
+ *
+ * Description: The implementation of the close(2) system call
+ *
+ * Parameters: p Process in whose per process file table
+ * the close is to occur
+ * uap->fd fd to be closed
+ * retval <unused>
+ *
+ * Returns: 0 Success
+ * fp_lookup:EBADF Bad file descriptor
+ * fp_guard_exception:??? Guarded file descriptor
+ * close_internal:EBADF
+ * close_internal:??? Anything returnable by a per-fileops
+ * close function
+ */
+int
+sys_close(proc_t p, struct close_args *uap, __unused int32_t *retval)
+{
+ __pthread_testcancel(1);
+ return close_nocancel(p, uap->fd);
+}
+
+int
+sys_close_nocancel(proc_t p, struct close_nocancel_args *uap, __unused int32_t *retval)
+{
+ return close_nocancel(p, uap->fd);
+}
+
+int
+close_nocancel(proc_t p, int fd)
+{
+ struct fileproc *fp;
+
+ AUDIT_SYSCLOSE(p, fd);
+
+ proc_fdlock(p);
+ if ((fp = fp_get_noref_locked(p, fd)) == NULL) {
+ proc_fdunlock(p);
+ return EBADF;
+ }
+
+ if (fp_isguarded(fp, GUARD_CLOSE)) {
+ int error = fp_guard_exception(p, fd, fp, kGUARD_EXC_CLOSE);
+ proc_fdunlock(p);
+ return error;
+ }
+
+ return fp_close_and_unlock(p, fd, fp, 0);
+}
+
+
+int
+fp_close_and_unlock(proc_t p, int fd, struct fileproc *fp, int flags)
+{
+ struct filedesc *fdp = p->p_fd;
+ struct fileglob *fg = fp->fp_glob;
+
+#if DIAGNOSTIC
+ proc_fdlock_assert(p, LCK_MTX_ASSERT_OWNED);
+#endif
+
+ /*
+ * Keep most people from finding the filedesc while we are closing it.
+ *
+ * Callers are:
+ *
+ * - dup2() which always waits for UF_RESERVED to clear
+ *
+ * - close/guarded_close/... who will fail the fileproc lookup if
+ * UF_RESERVED is set,
+ *
+ * - fdexec()/fdfree() who only run once all threads in the proc
+ * are properly canceled, hence no fileproc in this proc should
+ * be in flux.
+ *
+ * Which means that neither UF_RESERVED nor UF_CLOSING should be set.
+ *
+ * Callers of fp_get_noref_locked_with_iocount() can still find
+ * this entry so that they can drop their I/O reference despite
+ * not having remembered the fileproc pointer (namely select() and
+ * file_drop()).
+ */
+ if (p->p_fd->fd_ofileflags[fd] & (UF_RESERVED | UF_CLOSING)) {
+ panic("%s: called with fileproc in flux (%d/:%p)",
+ __func__, fd, fp);
+ }
+ p->p_fd->fd_ofileflags[fd] |= (UF_RESERVED | UF_CLOSING);
+
+ if ((fp->fp_flags & FP_AIOISSUED) || kauth_authorize_fileop_has_listeners()) {
+ proc_fdunlock(p);
+
+ if ((FILEGLOB_DTYPE(fg) == DTYPE_VNODE) && kauth_authorize_fileop_has_listeners()) {
+ /*
+ * call out to allow 3rd party notification of close.
+ * Ignore result of kauth_authorize_fileop call.
+ */
+ if (vnode_getwithref((vnode_t)fg->fg_data) == 0) {
+ u_int fileop_flags = 0;
+ if (fg->fg_flag & FWASWRITTEN) {
+ fileop_flags |= KAUTH_FILEOP_CLOSE_MODIFIED;
+ }
+ kauth_authorize_fileop(fg->fg_cred, KAUTH_FILEOP_CLOSE,
+ (uintptr_t)fg->fg_data, (uintptr_t)fileop_flags);
+#if CONFIG_MACF
+ mac_file_notify_close(proc_ucred(p), fp->fp_glob);
+#endif
+ vnode_put((vnode_t)fg->fg_data);
+ }
+ }
+ if (fp->fp_flags & FP_AIOISSUED) {
+ /*
+ * cancel all async IO requests that can be cancelled.
+ */
+ _aio_close( p, fd );
+ }
+
+ proc_fdlock(p);
+ }
+
+ if (fd < fdp->fd_knlistsize) {
+ knote_fdclose(p, fd);
+ }
+
+ fileproc_drain(p, fp);
+
+ if (flags & FD_DUP2RESV) {
+ fdp->fd_ofiles[fd] = NULL;
+ fdp->fd_ofileflags[fd] &= ~(UF_CLOSING | UF_EXCLOSE | UF_FORKCLOSE);
+ } else {
+ fdrelse(p, fd);
+ }
+
+ proc_fdunlock(p);
+
+ if (ENTR_SHOULDTRACE && FILEGLOB_DTYPE(fg) == DTYPE_SOCKET) {
+ KERNEL_ENERGYTRACE(kEnTrActKernSocket, DBG_FUNC_END,
+ fd, 0, (int64_t)VM_KERNEL_ADDRPERM(fg->fg_data));
+ }
+
+ fileproc_free(fp);
+
+ return fg_drop(p, fg);
+}
+
+
+/*
+ * fstat
+ *
+ * Description: Return status information about a file descriptor.
+ *
+ * Parameters: p The process doing the fstat
+ * fd The fd to stat
+ * ub The user stat buffer
+ * xsecurity The user extended security
+ * buffer, or 0 if none
+ * xsecurity_size The size of xsecurity, or 0
+ * if no xsecurity
+ * isstat64 Flag to indicate 64 bit version
+ * for inode size, etc.
+ *
+ * Returns: 0 Success
+ * EBADF
+ * EFAULT
+ * fp_lookup:EBADF Bad file descriptor
+ * vnode_getwithref:???
+ * copyout:EFAULT
+ * vnode_getwithref:???
+ * vn_stat:???
+ * soo_stat:???
+ * pipe_stat:???
+ * pshm_stat:???
+ * kqueue_stat:???
+ *
+ * Notes: Internal implementation for all other fstat() related
+ * functions
+ *
+ * XXX switch on node type is bogus; need a stat in struct
+ * XXX fileops instead.
+ */
+static int
+fstat(proc_t p, int fd, user_addr_t ub, user_addr_t xsecurity, user_addr_t xsecurity_size, int isstat64)
+{
+ struct fileproc *fp;
+ union {
+ struct stat sb;
+ struct stat64 sb64;
+ } source;
+ union {
+ struct user64_stat user64_sb;
+ struct user32_stat user32_sb;
+ struct user64_stat64 user64_sb64;
+ struct user32_stat64 user32_sb64;
+ } dest;
+ int error, my_size;
+ file_type_t type;
+ caddr_t data;
+ kauth_filesec_t fsec;
+ user_size_t xsecurity_bufsize;
+ vfs_context_t ctx = vfs_context_current();
+ void * sbptr;
+
+
+ AUDIT_ARG(fd, fd);
+
+ if ((error = fp_lookup(p, fd, &fp, 0)) != 0) {
+ return error;
+ }
+ type = fp->f_type;
+ data = fp->f_data;
+ fsec = KAUTH_FILESEC_NONE;
+
+ sbptr = (void *)&source;
+
+ switch (type) {
+ case DTYPE_VNODE:
+ if ((error = vnode_getwithref((vnode_t)data)) == 0) {
+ /*
+ * If the caller has the file open, and is not
+ * requesting extended security information, we are
+ * going to let them get the basic stat information.
+ */
+ if (xsecurity == USER_ADDR_NULL) {
+ error = vn_stat_noauth((vnode_t)data, sbptr, NULL, isstat64, 0, ctx,
+ fp->fp_glob->fg_cred);
+ } else {
+ error = vn_stat((vnode_t)data, sbptr, &fsec, isstat64, 0, ctx);
+ }
+
+ AUDIT_ARG(vnpath, (struct vnode *)data, ARG_VNODE1);
+ (void)vnode_put((vnode_t)data);
+ }
+ break;
+
+#if SOCKETS
+ case DTYPE_SOCKET:
+ error = soo_stat((struct socket *)data, sbptr, isstat64);
+ break;
+#endif /* SOCKETS */
+
+ case DTYPE_PIPE:
+ error = pipe_stat((void *)data, sbptr, isstat64);
+ break;
+
+ case DTYPE_PSXSHM:
+ error = pshm_stat((void *)data, sbptr, isstat64);
+ break;
+
+ case DTYPE_KQUEUE:
+ error = kqueue_stat((void *)data, sbptr, isstat64, p);
+ break;
+
+ default:
+ error = EBADF;
+ goto out;
+ }
+ if (error == 0) {
+ caddr_t sbp;
+
+ if (isstat64 != 0) {
+ source.sb64.st_lspare = 0;
+ source.sb64.st_qspare[0] = 0LL;
+ source.sb64.st_qspare[1] = 0LL;
+
+ if (IS_64BIT_PROCESS(current_proc())) {
+ munge_user64_stat64(&source.sb64, &dest.user64_sb64);
+ my_size = sizeof(dest.user64_sb64);
+ sbp = (caddr_t)&dest.user64_sb64;
+ } else {
+ munge_user32_stat64(&source.sb64, &dest.user32_sb64);
+ my_size = sizeof(dest.user32_sb64);
+ sbp = (caddr_t)&dest.user32_sb64;
+ }
+ } else {
+ source.sb.st_lspare = 0;
+ source.sb.st_qspare[0] = 0LL;
+ source.sb.st_qspare[1] = 0LL;
+ if (IS_64BIT_PROCESS(current_proc())) {
+ munge_user64_stat(&source.sb, &dest.user64_sb);
+ my_size = sizeof(dest.user64_sb);
+ sbp = (caddr_t)&dest.user64_sb;
+ } else {
+ munge_user32_stat(&source.sb, &dest.user32_sb);
+ my_size = sizeof(dest.user32_sb);
+ sbp = (caddr_t)&dest.user32_sb;
+ }
+ }
+
+ error = copyout(sbp, ub, my_size);
+ }
+
+ /* caller wants extended security information? */
+ if (xsecurity != USER_ADDR_NULL) {
+ /* did we get any? */
+ if (fsec == KAUTH_FILESEC_NONE) {
+ if (susize(xsecurity_size, 0) != 0) {
+ error = EFAULT;
+ goto out;
+ }
+ } else {
+ /* find the user buffer size */
+ xsecurity_bufsize = fusize(xsecurity_size);
+
+ /* copy out the actual data size */
+ if (susize(xsecurity_size, KAUTH_FILESEC_COPYSIZE(fsec)) != 0) {
+ error = EFAULT;
+ goto out;
+ }
+
+ /* if the caller supplied enough room, copy out to it */
+ if (xsecurity_bufsize >= KAUTH_FILESEC_COPYSIZE(fsec)) {
+ error = copyout(fsec, xsecurity, KAUTH_FILESEC_COPYSIZE(fsec));
+ }
+ }
+ }
+out:
+ fp_drop(p, fd, fp, 0);
+ if (fsec != NULL) {
+ kauth_filesec_free(fsec);
+ }
+ return error;
+}
+
+
+/*
+ * sys_fstat_extended
+ *
+ * Description: Extended version of fstat supporting returning extended
+ * security information
+ *
+ * Parameters: p The process doing the fstat
+ * uap->fd The fd to stat
+ * uap->ub The user stat buffer
+ * uap->xsecurity The user extended security
+ * buffer, or 0 if none
+ * uap->xsecurity_size The size of xsecurity, or 0
+ *
+ * Returns: 0 Success
+ * !0 Errno (see fstat)
+ */
+int
+sys_fstat_extended(proc_t p, struct fstat_extended_args *uap, __unused int32_t *retval)
+{
+ return fstat(p, uap->fd, uap->ub, uap->xsecurity, uap->xsecurity_size, 0);
+}
+
+
+/*
+ * sys_fstat
+ *
+ * Description: Get file status for the file associated with fd
+ *
+ * Parameters: p The process doing the fstat
+ * uap->fd The fd to stat
+ * uap->ub The user stat buffer
+ *
+ * Returns: 0 Success
+ * !0 Errno (see fstat)
+ */
+int
+sys_fstat(proc_t p, struct fstat_args *uap, __unused int32_t *retval)
+{
+ return fstat(p, uap->fd, uap->ub, 0, 0, 0);
+}
+
+
+/*
+ * sys_fstat64_extended
+ *
+ * Description: Extended version of fstat64 supporting returning extended
+ * security information
+ *
+ * Parameters: p The process doing the fstat
+ * uap->fd The fd to stat
+ * uap->ub The user stat buffer
+ * uap->xsecurity The user extended security
+ * buffer, or 0 if none
+ * uap->xsecurity_size The size of xsecurity, or 0
+ *
+ * Returns: 0 Success
+ * !0 Errno (see fstat)
+ */
+int
+sys_fstat64_extended(proc_t p, struct fstat64_extended_args *uap, __unused int32_t *retval)
+{
+ return fstat(p, uap->fd, uap->ub, uap->xsecurity, uap->xsecurity_size, 1);
+}
+
+
+/*
+ * sys_fstat64
+ *
+ * Description: Get 64 bit version of the file status for the file associated
+ * with fd
+ *
+ * Parameters: p The process doing the fstat
+ * uap->fd The fd to stat
+ * uap->ub The user stat buffer
+ *
+ * Returns: 0 Success
+ * !0 Errno (see fstat)
+ */
+int
+sys_fstat64(proc_t p, struct fstat64_args *uap, __unused int32_t *retval)
+{
+ return fstat(p, uap->fd, uap->ub, 0, 0, 1);
+}
+
+
+/*
+ * sys_fpathconf
+ *
+ * Description: Return pathconf information about a file descriptor.
+ *
+ * Parameters: p Process making the request
+ * uap->fd fd to get information about
+ * uap->name Name of information desired
+ * retval Pointer to the call return area
+ *
+ * Returns: 0 Success
+ * EINVAL
+ * fp_lookup:EBADF Bad file descriptor
+ * vnode_getwithref:???
+ * vn_pathconf:???
+ *
+ * Implicit returns:
+ * *retval (modified) Returned information (numeric)
+ */
+int
+sys_fpathconf(proc_t p, struct fpathconf_args *uap, int32_t *retval)
+{
+ int fd = uap->fd;
+ struct fileproc *fp;
+ struct vnode *vp;
+ int error = 0;
+ file_type_t type;
+ caddr_t data;
+
+
+ AUDIT_ARG(fd, uap->fd);
+ if ((error = fp_lookup(p, fd, &fp, 0))) {
+ return error;
+ }
+ type = fp->f_type;
+ data = fp->f_data;
+
+ switch (type) {
+ case DTYPE_SOCKET:
+ if (uap->name != _PC_PIPE_BUF) {
+ error = EINVAL;
+ goto out;
+ }
+ *retval = PIPE_BUF;
+ error = 0;
+ goto out;
+
+ case DTYPE_PIPE:
+ if (uap->name != _PC_PIPE_BUF) {
+ error = EINVAL;
+ goto out;
+ }
+ *retval = PIPE_BUF;
+ error = 0;
+ goto out;
+
+ case DTYPE_VNODE:
+ vp = (struct vnode *)data;
+
+ if ((error = vnode_getwithref(vp)) == 0) {
+ AUDIT_ARG(vnpath, vp, ARG_VNODE1);
+
+ error = vn_pathconf(vp, uap->name, retval, vfs_context_current());
+
+ (void)vnode_put(vp);
+ }
+ goto out;
+
+ default:
+ error = EINVAL;
+ goto out;
+ }
+ /*NOTREACHED*/
+out:
+ fp_drop(p, fd, fp, 0);
+ return error;
+}
+
+/*
+ * Statistics counter for the number of times a process calling fdalloc()
+ * has resulted in an expansion of the per process open file table.
+ *
+ * XXX This would likely be of more use if it were per process
+ */
+int fdexpand;
+
+
+/*
+ * fdalloc
+ *
+ * Description: Allocate a file descriptor for the process.
+ *
+ * Parameters: p Process to allocate the fd in
+ * want The fd we would prefer to get
+ * result Pointer to fd we got
+ *
+ * Returns: 0 Success
+ * EMFILE
+ * ENOMEM
+ *
+ * Implicit returns:
+ * *result (modified) The fd which was allocated
+ */
+int
+fdalloc(proc_t p, int want, int *result)
+{
+ struct filedesc *fdp = p->p_fd;
+ int i;
+ int last, numfiles, oldnfiles;
+ struct fileproc **newofiles, **ofiles;
+ char *newofileflags;
+ rlim_t lim;
+ rlim_t nofile = proc_limitgetcur(p, RLIMIT_NOFILE, TRUE);
+
+ nofile = MIN(nofile, INT_MAX);
+
+ /*
+ * Search for a free descriptor starting at the higher
+ * of want or fd_freefile. If that fails, consider
+ * expanding the ofile array.
+ */
+#if DIAGNOSTIC
+ proc_fdlock_assert(p, LCK_MTX_ASSERT_OWNED);
+#endif
+
+ lim = MIN(nofile, maxfilesperproc);
+ for (;;) {
+ last = (int)MIN((unsigned int)fdp->fd_nfiles, (unsigned int)lim);
+ if ((i = want) < fdp->fd_freefile) {
+ i = fdp->fd_freefile;
+ }
+ for (; i < last; i++) {
+ if (fdp->fd_ofiles[i] == NULL && !(fdp->fd_ofileflags[i] & UF_RESERVED)) {
+ procfdtbl_reservefd(p, i);
+ if (i > fdp->fd_lastfile) {
+ fdp->fd_lastfile = i;
+ }
+ if (want <= fdp->fd_freefile) {
+ fdp->fd_freefile = i;
+ }
+ *result = i;
+ return 0;
+ }
+ }
+
+ /*
+ * No space in current array. Expand?
+ */
+ if ((rlim_t)fdp->fd_nfiles >= lim) {
+ return EMFILE;
+ }
+ if (fdp->fd_nfiles < NDEXTENT) {
+ numfiles = NDEXTENT;
+ } else {
+ numfiles = 2 * fdp->fd_nfiles;
+ }
+ /* Enforce lim */
+ if ((rlim_t)numfiles > lim) {
+ numfiles = (int)lim;
+ }
+ proc_fdunlock(p);
+ newofiles = kheap_alloc(KM_OFILETABL, numfiles * OFILESIZE,
+ Z_WAITOK);
+ proc_fdlock(p);
+ if (newofiles == NULL) {
+ return ENOMEM;
+ }
+ if (fdp->fd_nfiles >= numfiles) {
+ kheap_free(KM_OFILETABL, newofiles, numfiles * OFILESIZE);
+ continue;
+ }
+ newofileflags = (char *) &newofiles[numfiles];
+ /*
+ * Copy the existing ofile and ofileflags arrays
+ * and zero the new portion of each array.
+ */
+ oldnfiles = fdp->fd_nfiles;
+ (void) memcpy(newofiles, fdp->fd_ofiles,
+ oldnfiles * sizeof(*fdp->fd_ofiles));
+ (void) memset(&newofiles[oldnfiles], 0,
+ (numfiles - oldnfiles) * sizeof(*fdp->fd_ofiles));
+
+ (void) memcpy(newofileflags, fdp->fd_ofileflags,
+ oldnfiles * sizeof(*fdp->fd_ofileflags));
+ (void) memset(&newofileflags[oldnfiles], 0,
+ (numfiles - oldnfiles) *
+ sizeof(*fdp->fd_ofileflags));
+ ofiles = fdp->fd_ofiles;
+ fdp->fd_ofiles = newofiles;
+ fdp->fd_ofileflags = newofileflags;
+ fdp->fd_nfiles = numfiles;
+ kheap_free(KM_OFILETABL, ofiles, oldnfiles * OFILESIZE);
+ fdexpand++;
+ }
+}
+
+
+/*
+ * fdavail
+ *
+ * Description: Check to see whether n user file descriptors are available
+ * to the process p.
+ *
+ * Parameters: p Process to check in
+ * n The number of fd's desired
+ *
+ * Returns: 0 No
+ * 1 Yes
+ *
+ * Locks: Assumes proc_fdlock for process is held by the caller
+ *
+ * Notes: The answer only remains valid so long as the proc_fdlock is
+ * held by the caller.
+ */
+int
+fdavail(proc_t p, int n)
+{
+ struct filedesc *fdp = p->p_fd;
+ struct fileproc **fpp;
+ char *flags;
+ int i;
+ int lim;
+ rlim_t nofile = proc_limitgetcur(p, RLIMIT_NOFILE, TRUE);
+
+ lim = (int)MIN(nofile, maxfilesperproc);
+ if ((i = lim - fdp->fd_nfiles) > 0 && (n -= i) <= 0) {
+ return 1;
+ }
+ fpp = &fdp->fd_ofiles[fdp->fd_freefile];
+ flags = &fdp->fd_ofileflags[fdp->fd_freefile];
+ for (i = fdp->fd_nfiles - fdp->fd_freefile; --i >= 0; fpp++, flags++) {
+ if (*fpp == NULL && !(*flags & UF_RESERVED) && --n <= 0) {
+ return 1;
+ }
+ }
+ return 0;
+}
+
+
+struct fileproc *
+fp_get_noref_locked(proc_t p, int fd)
+{
+ struct filedesc *fdp = p->p_fd;
+ struct fileproc *fp;
+
+ if (fd < 0 || fd >= fdp->fd_nfiles ||
+ (fp = fdp->fd_ofiles[fd]) == NULL ||
+ (fdp->fd_ofileflags[fd] & UF_RESERVED)) {
+ return NULL;
+ }
+ return fp;
+}
+
+struct fileproc *
+fp_get_noref_locked_with_iocount(proc_t p, int fd)
+{
+ struct filedesc *fdp = p->p_fd;
+ struct fileproc *fp = NULL;
+
+ if (fd < 0 || fd >= fdp->fd_nfiles ||
+ (fp = fdp->fd_ofiles[fd]) == NULL ||
+ os_ref_get_count(&fp->fp_iocount) <= 1 ||
+ ((fdp->fd_ofileflags[fd] & UF_RESERVED) &&
+ !(fdp->fd_ofileflags[fd] & UF_CLOSING))) {
+ panic("%s: caller without an ioccount on fileproc (%d/:%p)",
+ __func__, fd, fp);
+ }
+
+ return fp;
+}
+
+int
+fp_get_ftype(proc_t p, int fd, file_type_t ftype, int err, struct fileproc **fpp)
+{
+ struct filedesc *fdp = p->p_fd;
+ struct fileproc *fp;
+
+ proc_fdlock_spin(p);
+ if (fd < 0 || fd >= fdp->fd_nfiles ||
+ (fp = fdp->fd_ofiles[fd]) == NULL ||
+ (fdp->fd_ofileflags[fd] & UF_RESERVED)) {
+ proc_fdunlock(p);
+ return EBADF;
+ }
+
+ if (fp->f_type != ftype) {
+ proc_fdunlock(p);
+ return err;
+ }
+
+ os_ref_retain_locked(&fp->fp_iocount);
+ proc_fdunlock(p);
+
+ *fpp = fp;
+ return 0;
+}
+
+
+/*
+ * fp_getfvp
+ *
+ * Description: Get fileproc and vnode pointer for a given fd from the per
+ * process open file table of the specified process, and if
+ * successful, increment the fp_iocount
+ *
+ * Parameters: p Process in which fd lives
+ * fd fd to get information for
+ * resultfp Pointer to result fileproc
+ * pointer area, or 0 if none
+ * resultvp Pointer to result vnode pointer
+ * area, or 0 if none
+ *
+ * Returns: 0 Success
+ * EBADF Bad file descriptor
+ * ENOTSUP fd does not refer to a vnode
+ *
+ * Implicit returns:
+ * *resultfp (modified) Fileproc pointer
+ * *resultvp (modified) vnode pointer
+ *
+ * Notes: The resultfp and resultvp fields are optional, and may be
+ * independently specified as NULL to skip returning information
+ *
+ * Locks: Internally takes and releases proc_fdlock
+ */
+int
+fp_getfvp(proc_t p, int fd, struct fileproc **resultfp, struct vnode **resultvp)
+{
+ struct fileproc *fp;
+ int error;
+
+ error = fp_get_ftype(p, fd, DTYPE_VNODE, ENOTSUP, &fp);
+ if (error == 0) {
+ if (resultfp) {
+ *resultfp = fp;
+ }
+ if (resultvp) {
+ *resultvp = (struct vnode *)fp->f_data;
+ }
+ }
+
+ return error;
+}
+
+
+/*
+ * fp_get_pipe_id
+ *
+ * Description: Get pipe id for a given fd from the per process open file table
+ * of the specified process.
+ *
+ * Parameters: p Process in which fd lives
+ * fd fd to get information for
+ * result_pipe_id Pointer to result pipe id
+ *
+ * Returns: 0 Success
+ * EIVAL NULL pointer arguments passed
+ * fp_lookup:EBADF Bad file descriptor
+ * ENOTSUP fd does not refer to a pipe
+ *
+ * Implicit returns:
+ * *result_pipe_id (modified) pipe id
+ *
+ * Locks: Internally takes and releases proc_fdlock
+ */
+int
+fp_get_pipe_id(proc_t p, int fd, uint64_t *result_pipe_id)
+{
+ struct fileproc *fp = FILEPROC_NULL;
+ struct fileglob *fg = NULL;
+ int error = 0;
+
+ if (p == NULL || result_pipe_id == NULL) {
+ return EINVAL;
+ }
+
+ proc_fdlock(p);
+ if ((error = fp_lookup(p, fd, &fp, 1))) {
+ proc_fdunlock(p);
+ return error;
+ }
+ fg = fp->fp_glob;
+
+ if (FILEGLOB_DTYPE(fg) == DTYPE_PIPE) {
+ *result_pipe_id = pipe_id((struct pipe*)fg->fg_data);
+ } else {
+ error = ENOTSUP;
+ }
+
+ fp_drop(p, fd, fp, 1);
+ proc_fdunlock(p);
+ return error;
+}
+
+
+/*
+ * fp_lookup
+ *
+ * Description: Get fileproc pointer for a given fd from the per process
+ * open file table of the specified process and if successful,
+ * increment the fp_iocount
+ *
+ * Parameters: p Process in which fd lives
+ * fd fd to get information for
+ * resultfp Pointer to result fileproc
+ * pointer area, or 0 if none
+ * locked !0 if the caller holds the
+ * proc_fdlock, 0 otherwise
+ *
+ * Returns: 0 Success
+ * EBADF Bad file descriptor
+ *
+ * Implicit returns:
+ * *resultfp (modified) Fileproc pointer
+ *
+ * Locks: If the argument 'locked' is non-zero, then the caller is
+ * expected to have taken and held the proc_fdlock; if it is
+ * zero, than this routine internally takes and drops this lock.
+ */
+int
+fp_lookup(proc_t p, int fd, struct fileproc **resultfp, int locked)
+{
+ struct filedesc *fdp = p->p_fd;
+ struct fileproc *fp;
+
+ if (!locked) {
+ proc_fdlock_spin(p);
+ }
+ if (fd < 0 || fdp == NULL || fd >= fdp->fd_nfiles ||
+ (fp = fdp->fd_ofiles[fd]) == NULL ||
+ (fdp->fd_ofileflags[fd] & UF_RESERVED)) {
+ if (!locked) {
+ proc_fdunlock(p);
+ }
+ return EBADF;
+ }
+ os_ref_retain_locked(&fp->fp_iocount);
+
+ if (resultfp) {
+ *resultfp = fp;
+ }
+ if (!locked) {
+ proc_fdunlock(p);
+ }
+
+ return 0;
+}
+
+
+/*
+ * fp_tryswap
+ *
+ * Description: Swap the fileproc pointer for a given fd with a new
+ * fileproc pointer in the per-process open file table of
+ * the specified process. The fdlock must be held at entry.
+ * Iff the swap is successful, the old fileproc pointer is freed.
+ *
+ * Parameters: p Process containing the fd
+ * fd The fd of interest
+ * nfp Pointer to the newfp
+ *
+ * Returns: 0 Success
+ * EBADF Bad file descriptor
+ * EINTR Interrupted
+ * EKEEPLOOKING Other references were active, try again.
+ */
+int
+fp_tryswap(proc_t p, int fd, struct fileproc *nfp)
+{
+ struct fileproc *fp;
+ int error;
+
+ proc_fdlock_assert(p, LCK_MTX_ASSERT_OWNED);
+
+ if (0 != (error = fp_lookup(p, fd, &fp, 1))) {
+ return error;
+ }
+ /*
+ * At this point, our caller (change_guardedfd_np) has
+ * one fp_iocount reference, and we just took another
+ * one to begin the replacement.
+ * fp and nfp have a +1 reference from allocation.
+ * Thus if no-one else is looking, fp_iocount should be 3.
+ */
+ if (os_ref_get_count(&fp->fp_iocount) < 3 ||
+ 1 != os_ref_get_count(&nfp->fp_iocount)) {
+ panic("%s: fp_iocount", __func__);
+ } else if (3 == os_ref_get_count(&fp->fp_iocount)) {
+ /* Copy the contents of *fp, preserving the "type" of *nfp */
+
+ nfp->fp_flags = (nfp->fp_flags & FP_TYPEMASK) |
+ (fp->fp_flags & ~FP_TYPEMASK);
+ os_ref_retain_locked(&nfp->fp_iocount);
+ os_ref_retain_locked(&nfp->fp_iocount);
+ nfp->fp_glob = fp->fp_glob;
+ nfp->fp_wset = fp->fp_wset;
+
+ p->p_fd->fd_ofiles[fd] = nfp;
+ fp_drop(p, fd, nfp, 1);
+
+ os_ref_release_live(&fp->fp_iocount);
+ os_ref_release_live(&fp->fp_iocount);
+ fileproc_free(fp);
+ } else {
+ /*
+ * Wait for all other active references to evaporate.
+ */
+ p->p_fpdrainwait = 1;
+ error = msleep(&p->p_fpdrainwait, &p->p_fdmlock,
+ PRIBIO | PCATCH, "tryswap fpdrain", NULL);
+ if (0 == error) {
+ /*
+ * Return an "internal" errno to trigger a full
+ * reevaluation of the change-guard attempt.
+ */
+ error = EKEEPLOOKING;
+ }
+ (void) fp_drop(p, fd, fp, 1);
+ }
+ return error;
+}
+
+
+/*
+ * fp_drop
+ *
+ * Description: Drop the I/O reference previously taken by calling fp_lookup
+ * et. al.
+ *
+ * Parameters: p Process in which the fd lives
+ * fd fd associated with the fileproc
+ * fp fileproc on which to set the
+ * flag and drop the reference
+ * locked flag to internally take and
+ * drop proc_fdlock if it is not
+ * already held by the caller
+ *
+ * Returns: 0 Success
+ * EBADF Bad file descriptor
+ *
+ * Locks: This function internally takes and drops the proc_fdlock for
+ * the supplied process if 'locked' is non-zero, and assumes that
+ * the caller already holds this lock if 'locked' is non-zero.
+ *
+ * Notes: The fileproc must correspond to the fd in the supplied proc
+ */
+int
+fp_drop(proc_t p, int fd, struct fileproc *fp, int locked)
+{
+ struct filedesc *fdp = p->p_fd;
+ int needwakeup = 0;
+
+ if (!locked) {
+ proc_fdlock_spin(p);
+ }
+ if ((fp == FILEPROC_NULL) && (fd < 0 || fd >= fdp->fd_nfiles ||
+ (fp = fdp->fd_ofiles[fd]) == NULL ||
+ ((fdp->fd_ofileflags[fd] & UF_RESERVED) &&
+ !(fdp->fd_ofileflags[fd] & UF_CLOSING)))) {
+ if (!locked) {
+ proc_fdunlock(p);
+ }
+ return EBADF;
+ }
+
+ if (1 == os_ref_release_locked(&fp->fp_iocount)) {
+ if (fp->fp_flags & FP_SELCONFLICT) {
+ fp->fp_flags &= ~FP_SELCONFLICT;
+ }
+
+ if (p->p_fpdrainwait) {
+ p->p_fpdrainwait = 0;
+ needwakeup = 1;
+ }
+ }
+ if (!locked) {
+ proc_fdunlock(p);
+ }
+ if (needwakeup) {
+ wakeup(&p->p_fpdrainwait);
+ }
+
+ return 0;
+}
+
+
+/*
+ * file_vnode
+ *
+ * Description: Given an fd, look it up in the current process's per process
+ * open file table, and return its internal vnode pointer.
+ *
+ * Parameters: fd fd to obtain vnode from
+ * vpp pointer to vnode return area
+ *
+ * Returns: 0 Success
+ * EINVAL The fd does not refer to a
+ * vnode fileproc entry
+ * fp_lookup:EBADF Bad file descriptor
+ *
+ * Implicit returns:
+ * *vpp (modified) Returned vnode pointer
+ *
+ * Locks: This function internally takes and drops the proc_fdlock for
+ * the current process
+ *
+ * Notes: If successful, this function increments the fp_iocount on the
+ * fd's corresponding fileproc.
+ *
+ * The fileproc referenced is not returned; because of this, care
+ * must be taken to not drop the last reference (e.g. by closing
+ * the file). This is inherently unsafe, since the reference may
+ * not be recoverable from the vnode, if there is a subsequent
+ * close that destroys the associate fileproc. The caller should
+ * therefore retain their own reference on the fileproc so that
+ * the fp_iocount can be dropped subsequently. Failure to do this
+ * can result in the returned pointer immediately becoming invalid
+ * following the call.
+ *
+ * Use of this function is discouraged.
+ */
+int
+file_vnode(int fd, struct vnode **vpp)
+{
+ return file_vnode_withvid(fd, vpp, NULL);
+}
+
+/*
+ * file_vnode_withvid
+ *
+ * Description: Given an fd, look it up in the current process's per process
+ * open file table, and return its internal vnode pointer.
+ *
+ * Parameters: fd fd to obtain vnode from
+ * vpp pointer to vnode return area
+ * vidp pointer to vid of the returned vnode
+ *
+ * Returns: 0 Success
+ * EINVAL The fd does not refer to a
+ * vnode fileproc entry
+ * fp_lookup:EBADF Bad file descriptor
+ *
+ * Implicit returns:
+ * *vpp (modified) Returned vnode pointer
+ *
+ * Locks: This function internally takes and drops the proc_fdlock for
+ * the current process
+ *
+ * Notes: If successful, this function increments the fp_iocount on the
+ * fd's corresponding fileproc.
+ *
+ * The fileproc referenced is not returned; because of this, care
+ * must be taken to not drop the last reference (e.g. by closing
+ * the file). This is inherently unsafe, since the reference may
+ * not be recoverable from the vnode, if there is a subsequent
+ * close that destroys the associate fileproc. The caller should
+ * therefore retain their own reference on the fileproc so that
+ * the fp_iocount can be dropped subsequently. Failure to do this
+ * can result in the returned pointer immediately becoming invalid
+ * following the call.
+ *
+ * Use of this function is discouraged.
+ */
+int
+file_vnode_withvid(int fd, struct vnode **vpp, uint32_t *vidp)
+{
+ struct fileproc *fp;
+ int error;
+
+ error = fp_get_ftype(current_proc(), fd, DTYPE_VNODE, EINVAL, &fp);
+ if (error == 0) {
+ if (vpp) {
+ *vpp = fp->f_data;
+ }
+ if (vidp) {
+ *vidp = vnode_vid(fp->f_data);
+ }
+ }
+ return error;
+}
+
+/*
+ * file_socket
+ *
+ * Description: Given an fd, look it up in the current process's per process
+ * open file table, and return its internal socket pointer.
+ *
+ * Parameters: fd fd to obtain vnode from
+ * sp pointer to socket return area
+ *
+ * Returns: 0 Success
+ * ENOTSOCK Not a socket
+ * fp_lookup:EBADF Bad file descriptor
+ *
+ * Implicit returns:
+ * *sp (modified) Returned socket pointer
+ *
+ * Locks: This function internally takes and drops the proc_fdlock for
+ * the current process
+ *
+ * Notes: If successful, this function increments the fp_iocount on the
+ * fd's corresponding fileproc.
+ *
+ * The fileproc referenced is not returned; because of this, care
+ * must be taken to not drop the last reference (e.g. by closing
+ * the file). This is inherently unsafe, since the reference may
+ * not be recoverable from the socket, if there is a subsequent
+ * close that destroys the associate fileproc. The caller should
+ * therefore retain their own reference on the fileproc so that
+ * the fp_iocount can be dropped subsequently. Failure to do this
+ * can result in the returned pointer immediately becoming invalid
+ * following the call.
+ *
+ * Use of this function is discouraged.
+ */
+int
+file_socket(int fd, struct socket **sp)
+{
+ struct fileproc *fp;
+ int error;
+
+ error = fp_get_ftype(current_proc(), fd, DTYPE_SOCKET, ENOTSOCK, &fp);
+ if (error == 0) {
+ if (sp) {
+ *sp = (struct socket *)fp->f_data;
+ }
+ }
+ return error;
+}
+
+
+/*
+ * file_flags
+ *
+ * Description: Given an fd, look it up in the current process's per process
+ * open file table, and return its fileproc's flags field.
+ *
+ * Parameters: fd fd whose flags are to be
+ * retrieved
+ * flags pointer to flags data area
+ *
+ * Returns: 0 Success
+ * ENOTSOCK Not a socket
+ * fp_lookup:EBADF Bad file descriptor
+ *
+ * Implicit returns:
+ * *flags (modified) Returned flags field
+ *
+ * Locks: This function internally takes and drops the proc_fdlock for
+ * the current process
+ */
+int
+file_flags(int fd, int *flags)
+{
+ proc_t p = current_proc();
+ struct fileproc *fp;
+ int error = EBADF;
+
+ proc_fdlock_spin(p);
+ fp = fp_get_noref_locked(p, fd);
+ if (fp) {
+ *flags = (int)fp->f_flag;
+ error = 0;
+ }
+ proc_fdunlock(p);
+
+ return error;
+}
+
+
+/*
+ * file_drop
+ *
+ * Description: Drop an iocount reference on an fd, and wake up any waiters
+ * for draining (i.e. blocked in fileproc_drain() called during
+ * the last attempt to close a file).
+ *
+ * Parameters: fd fd on which an ioreference is
+ * to be dropped
+ *
+ * Returns: 0 Success
+ *
+ * Description: Given an fd, look it up in the current process's per process
+ * open file table, and drop it's fileproc's fp_iocount by one
+ *
+ * Notes: This is intended as a corresponding operation to the functions
+ * file_vnode() and file_socket() operations.
+ *
+ * If the caller can't possibly hold an I/O reference,
+ * this function will panic the kernel rather than allowing
+ * for memory corruption. Callers should always call this
+ * because they acquired an I/O reference on this file before.
+ *
+ * Use of this function is discouraged.
+ */
+int
+file_drop(int fd)
+{
+ struct fileproc *fp;
+ proc_t p = current_proc();
+ int needwakeup = 0;
+
+ proc_fdlock_spin(p);
+ fp = fp_get_noref_locked_with_iocount(p, fd);
+
+ if (1 == os_ref_release_locked(&fp->fp_iocount)) {
+ if (fp->fp_flags & FP_SELCONFLICT) {
+ fp->fp_flags &= ~FP_SELCONFLICT;
+ }
+
+ if (p->p_fpdrainwait) {
+ p->p_fpdrainwait = 0;
+ needwakeup = 1;
+ }
+ }
+ proc_fdunlock(p);
+
+ if (needwakeup) {
+ wakeup(&p->p_fpdrainwait);
+ }
+ return 0;
+}
+
+
+
+/*
+ * falloc_withalloc
+ *
+ * Create a new open file structure and allocate
+ * a file descriptor for the process that refers to it.
+ *
+ * Returns: 0 Success
+ *
+ * Description: Allocate an entry in the per process open file table and
+ * return the corresponding fileproc and fd.
+ *
+ * Parameters: p The process in whose open file
+ * table the fd is to be allocated
+ * resultfp Pointer to fileproc pointer
+ * return area
+ * resultfd Pointer to fd return area
+ * ctx VFS context
+ * fp_zalloc fileproc allocator to use
+ * crarg allocator args
+ *
+ * Returns: 0 Success
+ * ENFILE Too many open files in system
+ * fdalloc:EMFILE Too many open files in process
+ * fdalloc:ENOMEM M_OFILETABL zone exhausted
+ * ENOMEM fp_zone or fg_zone zone
+ * exhausted
+ *
+ * Implicit returns:
+ * *resultfd (modified) Returned fileproc pointer
+ * *resultfd (modified) Returned fd
+ *
+ * Notes: This function takes separate process and context arguments
+ * solely to support kern_exec.c; otherwise, it would take
+ * neither, and use the vfs_context_current() routine internally.
+ */
+int
+falloc_withalloc(proc_t p, struct fileproc **resultfp, int *resultfd,
+ vfs_context_t ctx, fp_allocfn_t fp_zalloc, void *crarg)
+{
+ struct fileproc *fp;
+ struct fileglob *fg;
+ int error, nfd;
+
+ /* Make sure we don't go beyond the system-wide limit */
+ if (nfiles >= maxfiles) {
+ tablefull("file");
+ return ENFILE;
+ }
+
+ proc_fdlock(p);
+
+ /* fdalloc will make sure the process stays below per-process limit */
+ if ((error = fdalloc(p, 0, &nfd))) {
+ proc_fdunlock(p);
+ return error;
+ }
+
+#if CONFIG_MACF
+ error = mac_file_check_create(proc_ucred(p));
+ if (error) {
+ proc_fdunlock(p);
+ return error;
+ }
+#endif
+
+ /*
+ * Allocate a new file descriptor.
+ * If the process has file descriptor zero open, add to the list
+ * of open files at that point, otherwise put it at the front of
+ * the list of open files.
+ */
+ proc_fdunlock(p);
+
+ fp = (*fp_zalloc)(crarg);
+ if (fp == NULL) {
+ return ENOMEM;
+ }
+ fg = zalloc_flags(fg_zone, Z_WAITOK | Z_ZERO);
+ lck_mtx_init(&fg->fg_lock, &file_lck_grp, LCK_ATTR_NULL);
+
+ os_ref_retain_locked(&fp->fp_iocount);
+ os_ref_init_raw(&fg->fg_count, &f_refgrp);
+ fg->fg_ops = &uninitops;
+ fp->fp_glob = fg;
+#if CONFIG_MACF
+ mac_file_label_init(fg);
+#endif
+
+ kauth_cred_ref(ctx->vc_ucred);
+
+ fp->f_cred = ctx->vc_ucred;
+
+#if CONFIG_MACF
+ mac_file_label_associate(fp->f_cred, fg);
+#endif
+
+ os_atomic_inc(&nfiles, relaxed);
+
+ proc_fdlock(p);
+
+ p->p_fd->fd_ofiles[nfd] = fp;
+
+ proc_fdunlock(p);
+
+ if (resultfp) {
+ *resultfp = fp;
+ }
+ if (resultfd) {
+ *resultfd = nfd;
+ }
+
+ return 0;
+}
+
+int
+falloc(proc_t p, struct fileproc **resultfp, int *resultfd, vfs_context_t ctx)
+{
+ return falloc_withalloc(p, resultfp, resultfd, ctx,
+ fileproc_alloc_init, NULL);
+}