]> git.saurik.com Git - apple/xnu.git/blobdiff - bsd/kern/sys_pipe.c
xnu-7195.81.3.tar.gz
[apple/xnu.git] / bsd / kern / sys_pipe.c
index c6725337b5b87fd024a1e95685df2d271af7baa3..508a92c87fa8d360722dc529b7475179b7d663c2 100644 (file)
  *    are met.
  */
 /*
- * Copyright (c) 2003-2014 Apple Inc. All rights reserved.
+ * Copyright (c) 2003-2020 Apple Inc. All rights reserved.
  *
  * @APPLE_OSREFERENCE_LICENSE_HEADER_START@
- * 
+ *
  * This file contains Original Code and/or Modifications of Original Code
  * as defined in and that are subject to the Apple Public Source License
  * Version 2.0 (the 'License'). You may not use this file except in
  * unlawful or unlicensed copies of an Apple operating system, or to
  * circumvent, violate, or enable the circumvention or violation of, any
  * terms of an Apple operating system software license agreement.
- * 
+ *
  * Please obtain a copy of the License at
  * http://www.opensource.apple.com/apsl/ and read it before using this file.
- * 
+ *
  * The Original Code and all software distributed under the License are
  * distributed on an 'AS IS' basis, WITHOUT WARRANTY OF ANY KIND, EITHER
  * EXPRESS OR IMPLIED, AND APPLE HEREBY DISCLAIMS ALL SUCH WARRANTIES,
@@ -40,7 +40,7 @@
  * FITNESS FOR A PARTICULAR PURPOSE, QUIET ENJOYMENT OR NON-INFRINGEMENT.
  * Please see the License for the specific language governing rights and
  * limitations under the License.
- * 
+ *
  * @APPLE_OSREFERENCE_LICENSE_HEADER_END@
  */
 /*
  * do.
  *
  * Pipes are implemented as circular buffers. Following are the valid states in pipes operations
- *  
+ *
  *      _________________________________
  * 1.  |_________________________________| r=w, c=0
- * 
+ *
  *      _________________________________
  * 2.  |__r:::::wc_______________________| r <= w , c > 0
  *
  *  a-z define the steps in a program flow
  *  1-4 are the states as defined aboe
  *  Action: is what file operation is done on the pipe
- *  
+ *
  *  Current:None  Action: initialize with size M=200
  *  a. State 1 ( r=0, w=0, c=0)
- *  
+ *
  *  Current: a    Action: write(100) (w < M)
  *  b. State 2 (r=0, w=100, c=100)
- *  
+ *
  *  Current: b    Action: write(100) (w = M-w)
  *  c. State 4 (r=0,w=0,c=200)
- *  
+ *
  *  Current: b    Action: read(70)  ( r < c )
  *  d. State 2(r=70,w=100,c=30)
- *  
+ *
  *  Current: d   Action: write(75) ( w < (m-w))
  *  e. State 2 (r=70,w=175,c=105)
- *  
+ *
  *  Current: d    Action: write(110) ( w > (m-w))
  *  f. State 3 (r=70,w=10,c=140)
- *  
+ *
  *  Current: d   Action: read(30) (r >= c )
  *  g. State 1 (r=100,w=100,c=0)
- *  
+ *
  */
 
 /*
  * dynamically change to larger sizes based on usage. The buffer size is never
  * reduced. The total amount of kernel memory used is governed by maxpipekva.
  * In case of dynamic expansion limit is reached, the output thread is blocked
- * until the pipe buffer empties enough to continue. 
+ * until the pipe buffer empties enough to continue.
  *
  * In order to limit the resource use of pipes, two sysctls exist:
  *
  * kern.ipc.maxpipekva - This is a hard limit on the amount of pageable
- * address space available to us in pipe_map. 
+ * address space available to us in pipe_map.
  *
  * Memory usage may be monitored through the sysctls
  * kern.ipc.pipes, kern.ipc.pipekva.
 #include <security/mac_framework.h>
 #endif
 
-#define f_flag f_fglob->fg_flag
-#define f_msgcount f_fglob->fg_msgcount
-#define f_cred f_fglob->fg_cred
-#define f_ops f_fglob->fg_ops
-#define f_offset f_fglob->fg_offset
-#define f_data f_fglob->fg_data
+#define f_flag fp_glob->fg_flag
+#define f_ops fp_glob->fg_ops
+#define f_data fp_glob->fg_data
+
+struct pipepair {
+       lck_mtx_t     pp_mtx;
+       struct pipe   pp_rpipe;
+       struct pipe   pp_wpipe;
+       uint64_t      pp_pipe_id;       /* unique ID shared by both pipe ends */
+};
+
+#define PIPE_PAIR(pipe) \
+               __container_of(PIPE_MTX(pipe), struct pipepair, pp_mtx)
 
 /*
- * interfaces to the outside world exported through file operations 
+ * interfaces to the outside world exported through file operations
  */
 static int pipe_read(struct fileproc *fp, struct uio *uio,
-               int flags, vfs_context_t ctx);
+    int flags, vfs_context_t ctx);
 static int pipe_write(struct fileproc *fp, struct uio *uio,
-               int flags, vfs_context_t ctx);
+    int flags, vfs_context_t ctx);
 static int pipe_close(struct fileglob *fg, vfs_context_t ctx);
 static int pipe_select(struct fileproc *fp, int which, void * wql,
-               vfs_context_t ctx);
+    vfs_context_t ctx);
 static int pipe_kqfilter(struct fileproc *fp, struct knote *kn,
-               struct kevent_internal_s *kev, vfs_context_t ctx);
+    struct kevent_qos_s *kev);
 static int pipe_ioctl(struct fileproc *fp, u_long cmd, caddr_t data,
-               vfs_context_t ctx);
-static int pipe_drain(struct fileproc *fp,vfs_context_t ctx);
+    vfs_context_t ctx);
+static int pipe_drain(struct fileproc *fp, vfs_context_t ctx);
 
 static const struct fileops pipeops = {
-       .fo_type = DTYPE_PIPE,
-       .fo_read = pipe_read,
-       .fo_write = pipe_write,
-       .fo_ioctl = pipe_ioctl,
-       .fo_select = pipe_select,
-       .fo_close = pipe_close,
+       .fo_type     = DTYPE_PIPE,
+       .fo_read     = pipe_read,
+       .fo_write    = pipe_write,
+       .fo_ioctl    = pipe_ioctl,
+       .fo_select   = pipe_select,
+       .fo_close    = pipe_close,
+       .fo_drain    = pipe_drain,
        .fo_kqfilter = pipe_kqfilter,
-       .fo_drain = pipe_drain,
 };
 
 static void filt_pipedetach(struct knote *kn);
 
+static int filt_pipenotsup(struct knote *kn, long hint);
+static int filt_pipenotsuptouch(struct knote *kn, struct kevent_qos_s *kev);
+static int filt_pipenotsupprocess(struct knote *kn, struct kevent_qos_s *kev);
+
 static int filt_piperead(struct knote *kn, long hint);
-static int filt_pipereadtouch(struct knote *kn, struct kevent_internal_s *kev);
-static int filt_pipereadprocess(struct knote *kn, struct filt_process_s *data, struct kevent_internal_s *kev);
+static int filt_pipereadtouch(struct knote *kn, struct kevent_qos_s *kev);
+static int filt_pipereadprocess(struct knote *kn, struct kevent_qos_s *kev);
 
 static int filt_pipewrite(struct knote *kn, long hint);
-static int filt_pipewritetouch(struct knote *kn, struct kevent_internal_s *kev);
-static int filt_pipewriteprocess(struct knote *kn, struct filt_process_s *data, struct kevent_internal_s *kev);
+static int filt_pipewritetouch(struct knote *kn, struct kevent_qos_s *kev);
+static int filt_pipewriteprocess(struct knote *kn, struct kevent_qos_s *kev);
+
+SECURITY_READ_ONLY_EARLY(struct filterops) pipe_nfiltops = {
+       .f_isfd    = 1,
+       .f_detach  = filt_pipedetach,
+       .f_event   = filt_pipenotsup,
+       .f_touch   = filt_pipenotsuptouch,
+       .f_process = filt_pipenotsupprocess,
+};
 
 SECURITY_READ_ONLY_EARLY(struct filterops) pipe_rfiltops = {
-        .f_isfd = 1,
-        .f_detach = filt_pipedetach,
-        .f_event = filt_piperead,
-       .f_touch = filt_pipereadtouch,
+       .f_isfd    = 1,
+       .f_detach  = filt_pipedetach,
+       .f_event   = filt_piperead,
+       .f_touch   = filt_pipereadtouch,
        .f_process = filt_pipereadprocess,
 };
 
 SECURITY_READ_ONLY_EARLY(struct filterops) pipe_wfiltops = {
-        .f_isfd = 1,
-        .f_detach = filt_pipedetach,
-        .f_event = filt_pipewrite,
-       .f_touch = filt_pipewritetouch,
+       .f_isfd    = 1,
+       .f_detach  = filt_pipedetach,
+       .f_event   = filt_pipewrite,
+       .f_touch   = filt_pipewritetouch,
        .f_process = filt_pipewriteprocess,
 };
 
+#if PIPE_SYSCTLS
 static int nbigpipe;      /* for compatibility sake. no longer used */
+#endif
 static int amountpipes;   /* total number of pipes in system */
 static int amountpipekva; /* total memory used by pipes */
 
+static _Atomic uint64_t pipe_unique_id = 1;
+
 int maxpipekva __attribute__((used)) = PIPE_KVAMAX;  /* allowing 16MB max. */
 
 #if PIPE_SYSCTLS
 SYSCTL_DECL(_kern_ipc);
 
-SYSCTL_INT(_kern_ipc, OID_AUTO, maxpipekva, CTLFLAG_RD|CTLFLAG_LOCKED,
-          &maxpipekva, 0, "Pipe KVA limit");
-SYSCTL_INT(_kern_ipc, OID_AUTO, maxpipekvawired, CTLFLAG_RW|CTLFLAG_LOCKED,
-          &maxpipekvawired, 0, "Pipe KVA wired limit");
-SYSCTL_INT(_kern_ipc, OID_AUTO, pipes, CTLFLAG_RD|CTLFLAG_LOCKED,
-          &amountpipes, 0, "Current # of pipes");
-SYSCTL_INT(_kern_ipc, OID_AUTO, bigpipes, CTLFLAG_RD|CTLFLAG_LOCKED,
-          &nbigpipe, 0, "Current # of big pipes");
-SYSCTL_INT(_kern_ipc, OID_AUTO, pipekva, CTLFLAG_RD|CTLFLAG_LOCKED,
-          &amountpipekva, 0, "Pipe KVA usage");
-SYSCTL_INT(_kern_ipc, OID_AUTO, pipekvawired, CTLFLAG_RD|CTLFLAG_LOCKED,
-          &amountpipekvawired, 0, "Pipe wired KVA usage");
+SYSCTL_INT(_kern_ipc, OID_AUTO, maxpipekva, CTLFLAG_RD | CTLFLAG_LOCKED,
+    &maxpipekva, 0, "Pipe KVA limit");
+SYSCTL_INT(_kern_ipc, OID_AUTO, maxpipekvawired, CTLFLAG_RW | CTLFLAG_LOCKED,
+    &maxpipekvawired, 0, "Pipe KVA wired limit");
+SYSCTL_INT(_kern_ipc, OID_AUTO, pipes, CTLFLAG_RD | CTLFLAG_LOCKED,
+    &amountpipes, 0, "Current # of pipes");
+SYSCTL_INT(_kern_ipc, OID_AUTO, bigpipes, CTLFLAG_RD | CTLFLAG_LOCKED,
+    &nbigpipe, 0, "Current # of big pipes");
+SYSCTL_INT(_kern_ipc, OID_AUTO, pipekva, CTLFLAG_RD | CTLFLAG_LOCKED,
+    &amountpipekva, 0, "Pipe KVA usage");
+SYSCTL_INT(_kern_ipc, OID_AUTO, pipekvawired, CTLFLAG_RD | CTLFLAG_LOCKED,
+    &amountpipekvawired, 0, "Pipe wired KVA usage");
 #endif
 
+static int pipepair_alloc(struct pipe **rpipe, struct pipe **wpipe);
 static void pipeclose(struct pipe *cpipe);
 static void pipe_free_kmem(struct pipe *cpipe);
-static int pipe_create(struct pipe **cpipep);
 static int pipespace(struct pipe *cpipe, int size);
 static int choose_pipespace(unsigned long current, unsigned long expected);
 static int expand_pipespace(struct pipe *p, int target_size);
@@ -245,70 +268,18 @@ static void pipeselwakeup(struct pipe *cpipe, struct pipe *spipe);
 static __inline int pipeio_lock(struct pipe *cpipe, int catch);
 static __inline void pipeio_unlock(struct pipe *cpipe);
 
-extern int postpipeevent(struct pipe *, int);
-extern void evpipefree(struct pipe *cpipe);
-
-static lck_grp_t       *pipe_mtx_grp;
-static lck_attr_t      *pipe_mtx_attr;
-static lck_grp_attr_t  *pipe_mtx_grp_attr;
-
-static zone_t pipe_zone;
+static LCK_GRP_DECLARE(pipe_mtx_grp, "pipe");
+static ZONE_DECLARE(pipe_zone, "pipe zone", sizeof(struct pipepair), ZC_NONE);
 
-#define MAX_PIPESIZE(pipe)             ( MAX(PIPE_SIZE, (pipe)->pipe_buffer.size) )
-
-#define        PIPE_GARBAGE_AGE_LIMIT          5000    /* In milliseconds */
-#define PIPE_GARBAGE_QUEUE_LIMIT       32000
-
-struct pipe_garbage {
-       struct pipe             *pg_pipe;
-       struct pipe_garbage     *pg_next;
-       uint64_t                pg_timestamp;
-};
-
-static zone_t pipe_garbage_zone;
-static struct pipe_garbage *pipe_garbage_head = NULL;
-static struct pipe_garbage *pipe_garbage_tail = NULL;
-static uint64_t pipe_garbage_age_limit = PIPE_GARBAGE_AGE_LIMIT;
-static int pipe_garbage_count = 0;
-static lck_mtx_t *pipe_garbage_lock;
-static void pipe_garbage_collect(struct pipe *cpipe);
+#define MAX_PIPESIZE(pipe)              ( MAX(PIPE_SIZE, (pipe)->pipe_buffer.size) )
 
 SYSINIT(vfs, SI_SUB_VFS, SI_ORDER_ANY, pipeinit, NULL);
 
-/* initial setup done at time of sysinit */
-void
-pipeinit(void)
-{
-       nbigpipe=0;
-       vm_size_t zone_size;
-       zone_size = 8192 * sizeof(struct pipe);
-        pipe_zone = zinit(sizeof(struct pipe), zone_size, 4096, "pipe zone");
-
-
-       /* allocate lock group attribute and group for pipe mutexes */
-       pipe_mtx_grp_attr = lck_grp_attr_alloc_init();
-       pipe_mtx_grp = lck_grp_alloc_init("pipe", pipe_mtx_grp_attr);
-
-       /* allocate the lock attribute for pipe mutexes */
-       pipe_mtx_attr = lck_attr_alloc_init();
-
-       /*
-        * Set up garbage collection for dead pipes
-        */
-       zone_size = (PIPE_GARBAGE_QUEUE_LIMIT + 20) *
-           sizeof(struct pipe_garbage);
-        pipe_garbage_zone = (zone_t)zinit(sizeof(struct pipe_garbage),
-           zone_size, 4096, "pipe garbage zone");
-       pipe_garbage_lock = lck_mtx_alloc_init(pipe_mtx_grp, pipe_mtx_attr);
-       
-}
-
-#ifndef        CONFIG_EMBEDDED
+#if defined(XNU_TARGET_OS_OSX)
 /* Bitmap for things to touch in pipe_touch() */
-#define        PIPE_ATIME      0x00000001      /* time of last access */
-#define        PIPE_MTIME      0x00000002      /* time of last modification */
-#define        PIPE_CTIME      0x00000004      /* time of last status change */
+#define PIPE_ATIME      0x00000001      /* time of last access */
+#define PIPE_MTIME      0x00000002      /* time of last modification */
+#define PIPE_CTIME      0x00000004      /* time of last status change */
 
 static void
 pipe_touch(struct pipe *tpipe, int touch)
@@ -334,16 +305,16 @@ pipe_touch(struct pipe *tpipe, int touch)
 }
 #endif
 
-static const unsigned int pipesize_blocks[] = {512,1024,2048,4096, 4096 * 2, PIPE_SIZE , PIPE_SIZE * 4 };
+static const unsigned int pipesize_blocks[] = {512, 1024, 2048, 4096, 4096 * 2, PIPE_SIZE, PIPE_SIZE * 4 };
 
-/* 
- * finds the right size from possible sizes in pipesize_blocks 
- * returns the size which matches max(current,expected) 
+/*
+ * finds the right size from possible sizes in pipesize_blocks
+ * returns the size which matches max(current,expected)
  */
-static int 
+static int
 choose_pipespace(unsigned long current, unsigned long expected)
 {
-       int i = sizeof(pipesize_blocks)/sizeof(unsigned int) -1;
+       int i = sizeof(pipesize_blocks) / sizeof(unsigned int) - 1;
        unsigned long target;
 
        /*
@@ -352,16 +323,16 @@ choose_pipespace(unsigned long current, unsigned long expected)
         */
        assert(PIPE_BUF == pipesize_blocks[0]);
 
-       if (expected > current) 
+       if (expected > current) {
                target = expected;
-       else
+       } else {
                target = current;
+       }
 
-       while ( i >0 && pipesize_blocks[i-1] > target) {
-               i=i-1;
-
+       while (i > 0 && pipesize_blocks[i - 1] > target) {
+               i = i - 1;
        }
-       
+
        return pipesize_blocks[i];
 }
 
@@ -373,27 +344,28 @@ choose_pipespace(unsigned long current, unsigned long expected)
  * Required: PIPE_LOCK and io lock to be held by caller.
  * returns 0 on success or no expansion possible
  */
-static int 
+static int
 expand_pipespace(struct pipe *p, int target_size)
 {
        struct pipe tmp, oldpipe;
        int error;
        tmp.pipe_buffer.buffer = 0;
-       
+
        if (p->pipe_buffer.size >= (unsigned) target_size) {
                return 0; /* the existing buffer is max size possible */
        }
-       
+
        /* create enough space in the target */
        error = pipespace(&tmp, target_size);
-       if (error != 0)
-               return (error);
+       if (error != 0) {
+               return error;
+       }
 
        oldpipe.pipe_buffer.buffer = p->pipe_buffer.buffer;
        oldpipe.pipe_buffer.size = p->pipe_buffer.size;
-       
+
        memcpy(tmp.pipe_buffer.buffer, p->pipe_buffer.buffer, p->pipe_buffer.size);
-       if (p->pipe_buffer.cnt > 0 && p->pipe_buffer.in <= p->pipe_buffer.out ){
+       if (p->pipe_buffer.cnt > 0 && p->pipe_buffer.in <= p->pipe_buffer.out{
                /* we are in State 3 and need extra copying for read to be consistent */
                memcpy(&tmp.pipe_buffer.buffer[p->pipe_buffer.size], p->pipe_buffer.buffer, p->pipe_buffer.size);
                p->pipe_buffer.in += p->pipe_buffer.size;
@@ -409,11 +381,11 @@ expand_pipespace(struct pipe *p, int target_size)
 
 /*
  * The pipe system call for the DTYPE_PIPE type of pipes
- * 
+ *
  * returns:
- *  FREAD  | fd0 | -->[struct rpipe] --> |~~buffer~~| \  
+ *  FREAD  | fd0 | -->[struct rpipe] --> |~~buffer~~| \
  *                                                    (pipe_mutex)
- *  FWRITE | fd1 | -->[struct wpipe] --X              / 
+ *  FWRITE | fd1 | -->[struct wpipe] --X              /
  */
 
 /* ARGSUSED */
@@ -422,47 +394,30 @@ pipe(proc_t p, __unused struct pipe_args *uap, int32_t *retval)
 {
        struct fileproc *rf, *wf;
        struct pipe *rpipe, *wpipe;
-       lck_mtx_t   *pmtx;
-       int fd, error;
-
-       if ((pmtx = lck_mtx_alloc_init(pipe_mtx_grp, pipe_mtx_attr)) == NULL)
-               return (ENOMEM);
-       
-       rpipe = wpipe = NULL;
-       if (pipe_create(&rpipe) || pipe_create(&wpipe)) {
-               error = ENFILE;
-               goto freepipes;
-       }
-        /*
-        * allocate the space for the normal I/O direction up
-        * front... we'll delay the allocation for the other
-        * direction until a write actually occurs (most likely it won't)...
-         */
-       error = pipespace(rpipe, choose_pipespace(rpipe->pipe_buffer.size, 0));
-        if (error)
-               goto freepipes;
-
-       TAILQ_INIT(&rpipe->pipe_evlist);
-       TAILQ_INIT(&wpipe->pipe_evlist);
+       int error;
 
-       error = falloc(p, &rf, &fd, vfs_context_current());
+       error = pipepair_alloc(&rpipe, &wpipe);
        if (error) {
-               goto freepipes;
+               return error;
        }
-       retval[0] = fd;
 
        /*
-        * for now we'll create half-duplex pipes(refer returns section above). 
+        * for now we'll create half-duplex pipes(refer returns section above).
         * this is what we've always supported..
         */
+
+       error = falloc(p, &rf, &retval[0], vfs_context_current());
+       if (error) {
+               goto freepipes;
+       }
        rf->f_flag = FREAD;
        rf->f_data = (caddr_t)rpipe;
        rf->f_ops = &pipeops;
 
-       error = falloc(p, &wf, &fd, vfs_context_current());
+       error = falloc(p, &wf, &retval[1], vfs_context_current());
        if (error) {
                fp_free(p, retval[0], rf);
-               goto freepipes;
+               goto freepipes;
        }
        wf->f_flag = FWRITE;
        wf->f_data = (caddr_t)wpipe;
@@ -470,10 +425,7 @@ pipe(proc_t p, __unused struct pipe_args *uap, int32_t *retval)
 
        rpipe->pipe_peer = wpipe;
        wpipe->pipe_peer = rpipe;
-       /* both structures share the same mutex */
-       rpipe->pipe_mtxp = wpipe->pipe_mtxp = pmtx; 
 
-       retval[1] = fd;
 #if CONFIG_MACF
        /*
         * XXXXXXXX SHOULD NOT HOLD FILE_LOCK() XXXXXXXXXXXX
@@ -493,75 +445,73 @@ pipe(proc_t p, __unused struct pipe_args *uap, int32_t *retval)
        fp_drop(p, retval[0], rf, 1);
        fp_drop(p, retval[1], wf, 1);
        proc_fdunlock(p);
-
-
-       return (0);
+       return 0;
 
 freepipes:
-       pipeclose(rpipe); 
-       pipeclose(wpipe); 
-       lck_mtx_free(pmtx, pipe_mtx_grp);
-
-       return (error);
+       pipeclose(rpipe);
+       pipeclose(wpipe);
+       return error;
 }
 
 int
 pipe_stat(struct pipe *cpipe, void *ub, int isstat64)
 {
 #if CONFIG_MACF
-        int error;
+       int error;
 #endif
-       int     pipe_size = 0;
-       int     pipe_count;
-       struct stat *sb = (struct stat *)0;     /* warning avoidance ; protected by isstat64 */
+       int     pipe_size = 0;
+       int     pipe_count;
+       struct stat *sb = (struct stat *)0;     /* warning avoidance ; protected by isstat64 */
        struct stat64 * sb64 = (struct stat64 *)0;  /* warning avoidance ; protected by isstat64 */
 
-       if (cpipe == NULL)
-               return (EBADF);
+       if (cpipe == NULL) {
+               return EBADF;
+       }
        PIPE_LOCK(cpipe);
 
 #if CONFIG_MACF
        error = mac_pipe_check_stat(kauth_cred_get(), cpipe);
        if (error) {
                PIPE_UNLOCK(cpipe);
-               return (error);
+               return error;
        }
 #endif
        if (cpipe->pipe_buffer.buffer == 0) {
-               /* must be stat'ing the write fd */
-               if (cpipe->pipe_peer) {
-                       /* the peer still exists, use it's info */
-                       pipe_size  = MAX_PIPESIZE(cpipe->pipe_peer);
+               /* must be stat'ing the write fd */
+               if (cpipe->pipe_peer) {
+                       /* the peer still exists, use it's info */
+                       pipe_size  = MAX_PIPESIZE(cpipe->pipe_peer);
                        pipe_count = cpipe->pipe_peer->pipe_buffer.cnt;
                } else {
                        pipe_count = 0;
                }
        } else {
-               pipe_size  = MAX_PIPESIZE(cpipe);
+               pipe_size  = MAX_PIPESIZE(cpipe);
                pipe_count = cpipe->pipe_buffer.cnt;
        }
        /*
         * since peer's buffer is setup ouside of lock
         * we might catch it in transient state
         */
-       if (pipe_size == 0)
+       if (pipe_size == 0) {
                pipe_size  = MAX(PIPE_SIZE, pipesize_blocks[0]);
+       }
 
        if (isstat64 != 0) {
-               sb64 = (struct stat64 *)ub;     
+               sb64 = (struct stat64 *)ub;
 
                bzero(sb64, sizeof(*sb64));
                sb64->st_mode = S_IFIFO | S_IRUSR | S_IWUSR | S_IRGRP | S_IWGRP;
                sb64->st_blksize = pipe_size;
                sb64->st_size = pipe_count;
                sb64->st_blocks = (sb64->st_size + sb64->st_blksize - 1) / sb64->st_blksize;
-       
+
                sb64->st_uid = kauth_getuid();
                sb64->st_gid = kauth_getgid();
-       
+
                sb64->st_atimespec.tv_sec  = cpipe->st_atimespec.tv_sec;
                sb64->st_atimespec.tv_nsec = cpipe->st_atimespec.tv_nsec;
-       
+
                sb64->st_mtimespec.tv_sec  = cpipe->st_mtimespec.tv_sec;
                sb64->st_mtimespec.tv_nsec = cpipe->st_mtimespec.tv_nsec;
 
@@ -569,26 +519,26 @@ pipe_stat(struct pipe *cpipe, void *ub, int isstat64)
                sb64->st_ctimespec.tv_nsec = cpipe->st_ctimespec.tv_nsec;
 
                /*
-               * Return a relatively unique inode number based on the current
-               * address of this pipe's struct pipe.  This number may be recycled
-               * relatively quickly.
-               */
-               sb64->st_ino = (ino64_t)VM_KERNEL_ADDRPERM((uintptr_t)cpipe);
+                * Return a relatively unique inode number based on the current
+                * address of this pipe's struct pipe.  This number may be recycled
+                * relatively quickly.
+                */
+               sb64->st_ino = (ino64_t)VM_KERNEL_ADDRHASH((uintptr_t)cpipe);
        } else {
-               sb = (struct stat *)ub; 
+               sb = (struct stat *)ub;
 
                bzero(sb, sizeof(*sb));
                sb->st_mode = S_IFIFO | S_IRUSR | S_IWUSR | S_IRGRP | S_IWGRP;
                sb->st_blksize = pipe_size;
                sb->st_size = pipe_count;
                sb->st_blocks = (sb->st_size + sb->st_blksize - 1) / sb->st_blksize;
-       
+
                sb->st_uid = kauth_getuid();
                sb->st_gid = kauth_getgid();
-       
+
                sb->st_atimespec.tv_sec  = cpipe->st_atimespec.tv_sec;
                sb->st_atimespec.tv_nsec = cpipe->st_atimespec.tv_nsec;
-       
+
                sb->st_mtimespec.tv_sec  = cpipe->st_mtimespec.tv_sec;
                sb->st_mtimespec.tv_nsec = cpipe->st_mtimespec.tv_nsec;
 
@@ -596,11 +546,11 @@ pipe_stat(struct pipe *cpipe, void *ub, int isstat64)
                sb->st_ctimespec.tv_nsec = cpipe->st_ctimespec.tv_nsec;
 
                /*
-               * Return a relatively unique inode number based on the current
-               * address of this pipe's struct pipe.  This number may be recycled
-               * relatively quickly.
-               */
-               sb->st_ino = (ino_t)VM_KERNEL_ADDRPERM((uintptr_t)cpipe);
+                * Return a relatively unique inode number based on the current
+                * address of this pipe's struct pipe.  This number may be recycled
+                * relatively quickly.
+                */
+               sb->st_ino = (ino_t)VM_KERNEL_ADDRHASH((uintptr_t)cpipe);
        }
        PIPE_UNLOCK(cpipe);
 
@@ -612,9 +562,14 @@ pipe_stat(struct pipe *cpipe, void *ub, int isstat64)
         * XXX is associated with pipes, since they are implemented via a
         * XXX struct fileops indirection rather than as FS objects.
         */
-       return (0);
+       return 0;
 }
 
+uint64_t
+pipe_id(struct pipe *p)
+{
+       return PIPE_PAIR(p)->pp_pipe_id;
+}
 
 /*
  * Allocate kva for pipe circular buffer, the space is pageable
@@ -627,11 +582,14 @@ pipespace(struct pipe *cpipe, int size)
 {
        vm_offset_t buffer;
 
-       if (size <= 0)
-               return(EINVAL);
+       if (size <= 0) {
+               return EINVAL;
+       }
 
-       if ((buffer = (vm_offset_t)kalloc(size)) == 0 )
-               return(ENOMEM);
+       buffer = (vm_offset_t)kheap_alloc(KHEAP_DATA_BUFFERS, size, Z_WAITOK);
+       if (!buffer) {
+               return ENOMEM;
+       }
 
        /* free old resources if we're resizing */
        pipe_free_kmem(cpipe);
@@ -644,34 +602,80 @@ pipespace(struct pipe *cpipe, int size)
        OSAddAtomic(1, &amountpipes);
        OSAddAtomic(cpipe->pipe_buffer.size, &amountpipekva);
 
-       return (0);
+       return 0;
 }
 
 /*
  * initialize and allocate VM and memory for pipe
  */
 static int
-pipe_create(struct pipe **cpipep)
+pipepair_alloc(struct pipe **rp_out, struct pipe **wp_out)
 {
-       struct pipe *cpipe;
-       cpipe = (struct pipe *)zalloc(pipe_zone);
+       struct pipepair *pp = zalloc(pipe_zone);
+       struct pipe *rpipe = &pp->pp_rpipe;
+       struct pipe *wpipe = &pp->pp_wpipe;
 
-       if ((*cpipep = cpipe) == NULL)
-               return (ENOMEM);
+       if (pp == NULL) {
+               return ENOMEM;
+       }
 
        /*
         * protect so pipespace or pipeclose don't follow a junk pointer
         * if pipespace() fails.
         */
-       bzero(cpipe, sizeof *cpipe);
+       bzero(pp, sizeof(struct pipepair));
+       pp->pp_pipe_id = os_atomic_inc_orig(&pipe_unique_id, relaxed);
+       lck_mtx_init(&pp->pp_mtx, &pipe_mtx_grp, LCK_ATTR_NULL);
 
-#ifndef        CONFIG_EMBEDDED
+       rpipe->pipe_mtxp = &pp->pp_mtx;
+       wpipe->pipe_mtxp = &pp->pp_mtx;
+
+#if defined(XNU_TARGET_OS_OSX)
        /* Initial times are all the time of creation of the pipe */
-       pipe_touch(cpipe, PIPE_ATIME | PIPE_MTIME | PIPE_CTIME);
+       pipe_touch(rpipe, PIPE_ATIME | PIPE_MTIME | PIPE_CTIME);
+       pipe_touch(wpipe, PIPE_ATIME | PIPE_MTIME | PIPE_CTIME);
 #endif
-       return (0);
+
+       /*
+        * allocate the space for the normal I/O direction up
+        * front... we'll delay the allocation for the other
+        * direction until a write actually occurs (most likely it won't)...
+        */
+       int error = pipespace(rpipe, choose_pipespace(rpipe->pipe_buffer.size, 0));
+       if (__improbable(error)) {
+               lck_mtx_destroy(&pp->pp_mtx, &pipe_mtx_grp);
+               zfree(pipe_zone, pp);
+               return error;
+       }
+
+       *rp_out = rpipe;
+       *wp_out = wpipe;
+       return 0;
 }
 
+static void
+pipepair_destroy_pipe(struct pipepair *pp, struct pipe *cpipe)
+{
+       bool can_free;
+
+       pipe_free_kmem(cpipe);
+
+       lck_mtx_lock(&pp->pp_mtx);
+       if (__improbable(cpipe->pipe_state & PIPE_DEAD)) {
+               panic("double free of pipe %p in pair %p", cpipe, pp);
+       }
+
+       cpipe->pipe_state |= PIPE_DEAD;
+
+       can_free = (pp->pp_rpipe.pipe_state & PIPE_DEAD) &&
+           (pp->pp_wpipe.pipe_state & PIPE_DEAD);
+       lck_mtx_unlock(&pp->pp_mtx);
+
+       if (can_free) {
+               lck_mtx_destroy(&pp->pp_mtx, &pipe_mtx_grp);
+               zfree(pipe_zone, pp);
+       }
+}
 
 /*
  * lock a pipe for I/O, blocking other access
@@ -683,12 +687,13 @@ pipeio_lock(struct pipe *cpipe, int catch)
        while (cpipe->pipe_state & PIPE_LOCKFL) {
                cpipe->pipe_state |= PIPE_LWANT;
                error = msleep(cpipe, PIPE_MTX(cpipe), catch ? (PRIBIO | PCATCH) : PRIBIO,
-                              "pipelk", 0);
-               if (error != 0) 
-                       return (error);
+                   "pipelk", 0);
+               if (error != 0) {
+                       return error;
+               }
        }
        cpipe->pipe_state |= PIPE_LOCKFL;
-       return (0);
+       return 0;
 }
 
 /*
@@ -714,17 +719,16 @@ pipeselwakeup(struct pipe *cpipe, struct pipe *spipe)
                cpipe->pipe_state &= ~PIPE_SEL;
                selwakeup(&cpipe->pipe_sel);
        }
-        if (cpipe->pipe_state & PIPE_KNOTE) 
-              KNOTE(&cpipe->pipe_sel.si_note, 1);
 
-       postpipeevent(cpipe, EV_RWBYTES);
+       KNOTE(&cpipe->pipe_sel.si_note, 1);
 
        if (spipe && (spipe->pipe_state & PIPE_ASYNC) && spipe->pipe_pgid) {
-               if (spipe->pipe_pgid < 0)
-                       gsignal(-spipe->pipe_pgid, SIGIO);
-               else 
-                       proc_signal(spipe->pipe_pgid, SIGIO);
-        }
+               if (spipe->pipe_pgid < 0) {
+                       gsignal(-spipe->pipe_pgid, SIGIO);
+               } else {
+                       proc_signal(spipe->pipe_pgid, SIGIO);
+               }
+       }
 }
 
 /*
@@ -734,7 +738,7 @@ pipeselwakeup(struct pipe *cpipe, struct pipe *spipe)
 /* ARGSUSED */
 static int
 pipe_read(struct fileproc *fp, struct uio *uio, __unused int flags,
-       __unused vfs_context_t ctx)
+    __unused vfs_context_t ctx)
 {
        struct pipe *rpipe = (struct pipe *)fp->f_data;
        int error;
@@ -745,13 +749,15 @@ pipe_read(struct fileproc *fp, struct uio *uio, __unused int flags,
        ++rpipe->pipe_busy;
 
        error = pipeio_lock(rpipe, 1);
-       if (error)
+       if (error) {
                goto unlocked_error;
+       }
 
 #if CONFIG_MACF
        error = mac_pipe_check_read(kauth_cred_get(), rpipe);
-       if (error)
+       if (error) {
                goto locked_error;
+       }
 #endif
 
 
@@ -762,30 +768,33 @@ pipe_read(struct fileproc *fp, struct uio *uio, __unused int flags,
                if (rpipe->pipe_buffer.cnt > 0) {
                        /*
                         * # bytes to read is min( bytes from read pointer until end of buffer,
-                        *                         total unread bytes, 
+                        *                         total unread bytes,
                         *                         user requested byte count)
                         */
                        size = rpipe->pipe_buffer.size - rpipe->pipe_buffer.out;
-                       if (size > rpipe->pipe_buffer.cnt)
+                       if (size > rpipe->pipe_buffer.cnt) {
                                size = rpipe->pipe_buffer.cnt;
-                       // LP64todo - fix this!
-                       if (size > (u_int) uio_resid(uio))
-                               size = (u_int) uio_resid(uio);
+                       }
+
+                       size = (u_int) MIN(INT_MAX, MIN((user_size_t)size,
+                           (user_size_t)uio_resid(uio)));
 
                        PIPE_UNLOCK(rpipe); /* we still hold io lock.*/
                        error = uiomove(
-                           &rpipe->pipe_buffer.buffer[rpipe->pipe_buffer.out],
-                           size, uio);
+                               &rpipe->pipe_buffer.buffer[rpipe->pipe_buffer.out],
+                               size, uio);
                        PIPE_LOCK(rpipe);
-                       if (error)
+                       if (error) {
                                break;
+                       }
 
                        rpipe->pipe_buffer.out += size;
-                       if (rpipe->pipe_buffer.out >= rpipe->pipe_buffer.size)
+                       if (rpipe->pipe_buffer.out >= rpipe->pipe_buffer.size) {
                                rpipe->pipe_buffer.out = 0;
+                       }
 
                        rpipe->pipe_buffer.cnt -= size;
-                       
+
                        /*
                         * If there is no more to read in the pipe, reset
                         * its pointers to the beginning.  This improves
@@ -801,7 +810,8 @@ pipe_read(struct fileproc *fp, struct uio *uio, __unused int flags,
                         * detect EOF condition
                         * read returns 0 on EOF, no need to set error
                         */
-                       if (rpipe->pipe_state & (PIPE_DRAIN | PIPE_EOF)) {
+                       if ((rpipe->pipe_state & (PIPE_DRAIN | PIPE_EOF)) ||
+                           (fileproc_get_vflags(fp) & FPV_DRAIN)) {
                                break;
                        }
 
@@ -816,11 +826,12 @@ pipe_read(struct fileproc *fp, struct uio *uio, __unused int flags,
                        /*
                         * Break if some data was read in previous iteration.
                         */
-                       if (nread > 0)
+                       if (nread > 0) {
                                break;
+                       }
 
                        /*
-                        * Unlock the pipe buffer for our remaining processing. 
+                        * Unlock the pipe buffer for our remaining processing.
                         * We will either break out with an error or we will
                         * sleep and relock to loop.
                         */
@@ -835,11 +846,13 @@ pipe_read(struct fileproc *fp, struct uio *uio, __unused int flags,
                        } else {
                                rpipe->pipe_state |= PIPE_WANTR;
                                error = msleep(rpipe, PIPE_MTX(rpipe), PRIBIO | PCATCH, "piperd", 0);
-                               if (error == 0)
-                                       error = pipeio_lock(rpipe, 1);
+                               if (error == 0) {
+                                       error = pipeio_lock(rpipe, 1);
+                               }
                        }
-                       if (error)
+                       if (error) {
                                goto unlocked_error;
+                       }
                }
        }
 #if CONFIG_MACF
@@ -854,7 +867,7 @@ unlocked_error:
         * PIPE_WANT processing only makes sense if pipe_busy is 0.
         */
        if ((rpipe->pipe_busy == 0) && (rpipe->pipe_state & PIPE_WANT)) {
-               rpipe->pipe_state &= ~(PIPE_WANT|PIPE_WANTW);
+               rpipe->pipe_state &= ~(PIPE_WANT | PIPE_WANTW);
                wakeup(rpipe);
        } else if (rpipe->pipe_buffer.cnt < rpipe->pipe_buffer.size) {
                /*
@@ -866,33 +879,37 @@ unlocked_error:
                }
        }
 
-       if ((rpipe->pipe_buffer.size - rpipe->pipe_buffer.cnt) > 0)
+       if ((rpipe->pipe_buffer.size - rpipe->pipe_buffer.cnt) > 0) {
                pipeselwakeup(rpipe, rpipe->pipe_peer);
+       }
 
-#ifndef        CONFIG_EMBEDDED
+#if defined(XNU_TARGET_OS_OSX)
        /* update last read time */
        pipe_touch(rpipe, PIPE_ATIME);
 #endif
 
        PIPE_UNLOCK(rpipe);
 
-       return (error);
+       return error;
 }
 
 /*
- * perform a write of n bytes into the read side of buffer. Since 
+ * perform a write of n bytes into the read side of buffer. Since
  * pipes are unidirectional a write is meant to be read by the otherside only.
  */
 static int
 pipe_write(struct fileproc *fp, struct uio *uio, __unused int flags,
-       __unused vfs_context_t ctx)
+    __unused vfs_context_t ctx)
 {
        int error = 0;
-       int orig_resid;
+       size_t orig_resid;
        int pipe_size;
        struct pipe *wpipe, *rpipe;
        // LP64todo - fix this!
-       orig_resid = uio_resid(uio);
+       orig_resid = (size_t)uio_resid(uio);
+       if (orig_resid > LONG_MAX) {
+               return EINVAL;
+       }
        int space;
 
        rpipe = (struct pipe *)fp->f_data;
@@ -903,15 +920,16 @@ pipe_write(struct fileproc *fp, struct uio *uio, __unused int flags,
        /*
         * detect loss of pipe read side, issue SIGPIPE if lost.
         */
-       if (wpipe == NULL || (wpipe->pipe_state & (PIPE_DRAIN | PIPE_EOF))) {
+       if (wpipe == NULL || (wpipe->pipe_state & (PIPE_DRAIN | PIPE_EOF)) ||
+           (fileproc_get_vflags(fp) & FPV_DRAIN)) {
                PIPE_UNLOCK(rpipe);
-               return (EPIPE);
+               return EPIPE;
        }
 #if CONFIG_MACF
        error = mac_pipe_check_write(kauth_cred_get(), wpipe);
        if (error) {
                PIPE_UNLOCK(rpipe);
-               return (error);
+               return error;
        }
 #endif
        ++wpipe->pipe_busy;
@@ -925,71 +943,72 @@ pipe_write(struct fileproc *fp, struct uio *uio, __unused int flags,
         * fd[1], so allocating space for both ends is a waste...
         */
 
-       if ( wpipe->pipe_buffer.buffer == 0 || ( 
-               (unsigned)orig_resid > wpipe->pipe_buffer.size - wpipe->pipe_buffer.cnt &&
-               amountpipekva < maxpipekva ) ) {
-
-               pipe_size = choose_pipespace(wpipe->pipe_buffer.size, wpipe->pipe_buffer.cnt + orig_resid);
+       if (wpipe->pipe_buffer.buffer == 0 || (
+                   (unsigned)orig_resid > wpipe->pipe_buffer.size - wpipe->pipe_buffer.cnt &&
+                   amountpipekva < maxpipekva)) {
+               pipe_size = choose_pipespace(wpipe->pipe_buffer.size, wpipe->pipe_buffer.cnt + orig_resid);
        }
        if (pipe_size) {
-               /*
+               /*
                 * need to do initial allocation or resizing of pipe
-                * holding both structure and io locks. 
+                * holding both structure and io locks.
                 */
                if ((error = pipeio_lock(wpipe, 1)) == 0) {
-                       if (wpipe->pipe_buffer.cnt == 0)                        
+                       if (wpipe->pipe_buffer.cnt == 0) {
                                error = pipespace(wpipe, pipe_size);
-                       else 
+                       } else {
                                error = expand_pipespace(wpipe, pipe_size);
-               
+                       }
+
                        pipeio_unlock(wpipe);
-                       
+
                        /* allocation failed */
-                       if (wpipe->pipe_buffer.buffer == 0)
-                               error = ENOMEM;
+                       if (wpipe->pipe_buffer.buffer == 0) {
+                               error = ENOMEM;
+                       }
                }
                if (error) {
-                       /*
+                       /*
                         * If an error occurred unbusy and return, waking up any pending
                         * readers.
                         */
-                       --wpipe->pipe_busy;
-                       if ((wpipe->pipe_busy == 0) && 
+                       --wpipe->pipe_busy;
+                       if ((wpipe->pipe_busy == 0) &&
                            (wpipe->pipe_state & PIPE_WANT)) {
-                               wpipe->pipe_state &= ~(PIPE_WANT | PIPE_WANTR);
+                               wpipe->pipe_state &= ~(PIPE_WANT | PIPE_WANTR);
                                wakeup(wpipe);
                        }
                        PIPE_UNLOCK(rpipe);
-                       return(error);
+                       return error;
                }
        }
 
        while (uio_resid(uio)) {
-
-       retrywrite:
+retrywrite:
                space = wpipe->pipe_buffer.size - wpipe->pipe_buffer.cnt;
 
                /* Writes of size <= PIPE_BUF must be atomic. */
-               if ((space < uio_resid(uio)) && (orig_resid <= PIPE_BUF))
+               if ((space < uio_resid(uio)) && (orig_resid <= PIPE_BUF)) {
                        space = 0;
+               }
 
                if (space > 0) {
+                       if ((error = pipeio_lock(wpipe, 1)) == 0) {
+                               size_t size;       /* Transfer size */
+                               size_t segsize;    /* first segment to transfer */
 
-                       if ((error = pipeio_lock(wpipe,1)) == 0) {
-                               int size;       /* Transfer size */
-                               int segsize;    /* first segment to transfer */
-
-                               if (wpipe->pipe_state & (PIPE_DRAIN | PIPE_EOF)) {
+                               if ((wpipe->pipe_state & (PIPE_DRAIN | PIPE_EOF)) ||
+                                   (fileproc_get_vflags(fp) & FPV_DRAIN)) {
                                        pipeio_unlock(wpipe);
-                                       error = EPIPE;
+                                       error = EPIPE;
                                        break;
                                }
-                               /* 
+                               /*
                                 * If a process blocked in pipeio_lock, our
                                 * value for space might be bad... the mutex
                                 * is dropped while we're blocked
                                 */
-                               if (space > (int)(wpipe->pipe_buffer.size - 
+                               if (space > (int)(wpipe->pipe_buffer.size -
                                    wpipe->pipe_buffer.cnt)) {
                                        pipeio_unlock(wpipe);
                                        goto retrywrite;
@@ -1000,47 +1019,53 @@ pipe_write(struct fileproc *fp, struct uio *uio, __unused int flags,
                                 * and free space in pipe buffer.
                                 */
                                // LP64todo - fix this!
-                               if (space > uio_resid(uio))
-                                       size = uio_resid(uio);
-                               else
+                               if (space > uio_resid(uio)) {
+                                       size = (size_t)uio_resid(uio);
+                                       if (size > LONG_MAX) {
+                                               panic("size greater than LONG_MAX");
+                                       }
+                               } else {
                                        size = space;
+                               }
                                /*
-                                * First segment to transfer is minimum of 
+                                * First segment to transfer is minimum of
                                 * transfer size and contiguous space in
                                 * pipe buffer.  If first segment to transfer
                                 * is less than the transfer size, we've got
                                 * a wraparound in the buffer.
                                 */
-                               segsize = wpipe->pipe_buffer.size - 
-                                       wpipe->pipe_buffer.in;
-                               if (segsize > size)
+                               segsize = wpipe->pipe_buffer.size -
+                                   wpipe->pipe_buffer.in;
+                               if (segsize > size) {
                                        segsize = size;
-                               
+                               }
+
                                /* Transfer first segment */
 
                                PIPE_UNLOCK(rpipe);
-                               error = uiomove(&wpipe->pipe_buffer.buffer[wpipe->pipe_buffer.in], 
-                                               segsize, uio);
+                               error = uiomove(&wpipe->pipe_buffer.buffer[wpipe->pipe_buffer.in],
+                                   (int)segsize, uio);
                                PIPE_LOCK(rpipe);
-                               
+
                                if (error == 0 && segsize < size) {
-                                       /* 
+                                       /*
                                         * Transfer remaining part now, to
                                         * support atomic writes.  Wraparound
                                         * happened. (State 3)
                                         */
-                                       if (wpipe->pipe_buffer.in + segsize != 
-                                           wpipe->pipe_buffer.size)
+                                       if (wpipe->pipe_buffer.in + segsize !=
+                                           wpipe->pipe_buffer.size) {
                                                panic("Expected pipe buffer "
                                                    "wraparound disappeared");
-                                               
+                                       }
+
                                        PIPE_UNLOCK(rpipe);
                                        error = uiomove(
-                                           &wpipe->pipe_buffer.buffer[0],
-                                           size - segsize, uio);
+                                               &wpipe->pipe_buffer.buffer[0],
+                                               (int)(size - segsize), uio);
                                        PIPE_LOCK(rpipe);
                                }
-                               /* 
+                               /*
                                 * readers never know to read until count is updated.
                                 */
                                if (error == 0) {
@@ -1049,24 +1074,25 @@ pipe_write(struct fileproc *fp, struct uio *uio, __unused int flags,
                                            wpipe->pipe_buffer.size) {
                                                if (wpipe->pipe_buffer.in !=
                                                    size - segsize +
-                                                   wpipe->pipe_buffer.size)
+                                                   wpipe->pipe_buffer.size) {
                                                        panic("Expected "
                                                            "wraparound bad");
-                                               wpipe->pipe_buffer.in = size -
-                                                   segsize;
+                                               }
+                                               wpipe->pipe_buffer.in = (unsigned int)(size -
+                                                   segsize);
                                        }
-                               
+
                                        wpipe->pipe_buffer.cnt += size;
                                        if (wpipe->pipe_buffer.cnt >
-                                           wpipe->pipe_buffer.size)
+                                           wpipe->pipe_buffer.size) {
                                                panic("Pipe buffer overflow");
-                               
+                                       }
                                }
                                pipeio_unlock(wpipe);
                        }
-                       if (error)
+                       if (error) {
                                break;
-
+                       }
                } else {
                        /*
                         * If the "read-side" has been blocked, wake it up now.
@@ -1075,23 +1101,25 @@ pipe_write(struct fileproc *fp, struct uio *uio, __unused int flags,
                                wpipe->pipe_state &= ~PIPE_WANTR;
                                wakeup(wpipe);
                        }
+
                        /*
-                        * don't block on non-blocking I/O
-                        * we'll do the pipeselwakeup on the way out
+                        * If read side wants to go away, we just issue a signal
+                        * to ourselves.
                         */
-                       if (fp->f_flag & FNONBLOCK) {
-                               error = EAGAIN;
+                       if ((wpipe->pipe_state & (PIPE_DRAIN | PIPE_EOF)) ||
+                           (fileproc_get_vflags(fp) & FPV_DRAIN)) {
+                               error = EPIPE;
                                break;
                        }
 
                        /*
-                        * If read side wants to go away, we just issue a signal
-                        * to ourselves.
+                        * don't block on non-blocking I/O
+                        * we'll do the pipeselwakeup on the way out
                         */
-                       if (wpipe->pipe_state & (PIPE_DRAIN | PIPE_EOF)) {
-                               error = EPIPE;
+                       if (fp->f_flag & FNONBLOCK) {
+                               error = EAGAIN;
                                break;
-                       }       
+                       }
 
                        /*
                         * We have no more space and have something to offer,
@@ -1103,8 +1131,9 @@ pipe_write(struct fileproc *fp, struct uio *uio, __unused int flags,
 
                        error = msleep(wpipe, PIPE_MTX(wpipe), PRIBIO | PCATCH, "pipewr", 0);
 
-                       if (error != 0)
+                       if (error != 0) {
                                break;
+                       }
                }
        }
        --wpipe->pipe_busy;
@@ -1128,14 +1157,14 @@ pipe_write(struct fileproc *fp, struct uio *uio, __unused int flags,
                pipeselwakeup(wpipe, wpipe);
        }
 
-#ifndef        CONFIG_EMBEDDED
+#if defined(XNU_TARGET_OS_OSX)
        /* Update modification, status change (# of bytes in pipe) times */
        pipe_touch(rpipe, PIPE_MTIME | PIPE_CTIME);
        pipe_touch(wpipe, PIPE_MTIME | PIPE_CTIME);
 #endif
        PIPE_UNLOCK(rpipe);
 
-       return (error);
+       return error;
 }
 
 /*
@@ -1144,7 +1173,7 @@ pipe_write(struct fileproc *fp, struct uio *uio, __unused int flags,
 /* ARGSUSED 3 */
 static int
 pipe_ioctl(struct fileproc *fp, u_long cmd, caddr_t data,
-       __unused vfs_context_t ctx)
+    __unused vfs_context_t ctx)
 {
        struct pipe *mpipe = (struct pipe *)fp->f_data;
 #if CONFIG_MACF
@@ -1158,15 +1187,14 @@ pipe_ioctl(struct fileproc *fp, u_long cmd, caddr_t data,
        if (error) {
                PIPE_UNLOCK(mpipe);
 
-               return (error);
+               return error;
        }
 #endif
 
        switch (cmd) {
-
        case FIONBIO:
                PIPE_UNLOCK(mpipe);
-               return (0);
+               return 0;
 
        case FIOASYNC:
                if (*(int *)data) {
@@ -1175,28 +1203,27 @@ pipe_ioctl(struct fileproc *fp, u_long cmd, caddr_t data,
                        mpipe->pipe_state &= ~PIPE_ASYNC;
                }
                PIPE_UNLOCK(mpipe);
-               return (0);
+               return 0;
 
        case FIONREAD:
                *(int *)data = mpipe->pipe_buffer.cnt;
                PIPE_UNLOCK(mpipe);
-               return (0);
+               return 0;
 
        case TIOCSPGRP:
                mpipe->pipe_pgid = *(int *)data;
 
                PIPE_UNLOCK(mpipe);
-               return (0);
+               return 0;
 
        case TIOCGPGRP:
                *(int *)data = mpipe->pipe_pgid;
 
                PIPE_UNLOCK(mpipe);
-               return (0);
-
+               return 0;
        }
        PIPE_UNLOCK(mpipe);
-       return (ENOTTY);
+       return ENOTTY;
 }
 
 
@@ -1207,13 +1234,14 @@ pipe_select(struct fileproc *fp, int which, void *wql, vfs_context_t ctx)
        struct pipe *wpipe;
        int    retnum = 0;
 
-       if (rpipe == NULL || rpipe == (struct pipe *)-1)
-               return (retnum);
+       if (rpipe == NULL || rpipe == (struct pipe *)-1) {
+               return retnum;
+       }
 
        PIPE_LOCK(rpipe);
 
        wpipe = rpipe->pipe_peer;
-       
+
 
 #if CONFIG_MACF
        /*
@@ -1223,44 +1251,44 @@ pipe_select(struct fileproc *fp, int which, void *wql, vfs_context_t ctx)
         */
        if (mac_pipe_check_select(vfs_context_ucred(ctx), rpipe, which)) {
                PIPE_UNLOCK(rpipe);
-               return (0);
+               return 0;
        }
 #endif
-        switch (which) {
-
-        case FREAD:
+       switch (which) {
+       case FREAD:
                if ((rpipe->pipe_state & PIPE_DIRECTW) ||
                    (rpipe->pipe_buffer.cnt > 0) ||
-                   (rpipe->pipe_state & (PIPE_DRAIN | PIPE_EOF))) {
-
-                       retnum = 1;
+                   (rpipe->pipe_state & (PIPE_DRAIN | PIPE_EOF)) ||
+                   (fileproc_get_vflags(fp) & FPV_DRAIN)) {
+                       retnum = 1;
                } else {
-                       rpipe->pipe_state |= PIPE_SEL;
-                       selrecord(vfs_context_proc(ctx), &rpipe->pipe_sel, wql);
+                       rpipe->pipe_state |= PIPE_SEL;
+                       selrecord(vfs_context_proc(ctx), &rpipe->pipe_sel, wql);
                }
                break;
 
-        case FWRITE:
-               if (wpipe)
+       case FWRITE:
+               if (wpipe) {
                        wpipe->pipe_state |= PIPE_WSELECT;
+               }
                if (wpipe == NULL || (wpipe->pipe_state & (PIPE_DRAIN | PIPE_EOF)) ||
+                   (fileproc_get_vflags(fp) & FPV_DRAIN) ||
                    (((wpipe->pipe_state & PIPE_DIRECTW) == 0) &&
-                    (MAX_PIPESIZE(wpipe) - wpipe->pipe_buffer.cnt) >= PIPE_BUF)) {
-
-                       retnum = 1;
+                   (MAX_PIPESIZE(wpipe) - wpipe->pipe_buffer.cnt) >= PIPE_BUF)) {
+                       retnum = 1;
                } else {
-                       wpipe->pipe_state |= PIPE_SEL;
+                       wpipe->pipe_state |= PIPE_SEL;
                        selrecord(vfs_context_proc(ctx), &wpipe->pipe_sel, wql);
                }
                break;
-        case 0:
-               rpipe->pipe_state |= PIPE_SEL;
+       case 0:
+               rpipe->pipe_state |= PIPE_SEL;
                selrecord(vfs_context_proc(ctx), &rpipe->pipe_sel, wql);
                break;
-        }
+       }
        PIPE_UNLOCK(rpipe);
 
-        return (retnum);
+       return retnum;
 }
 
 
@@ -1268,16 +1296,17 @@ pipe_select(struct fileproc *fp, int which, void *wql, vfs_context_t ctx)
 static int
 pipe_close(struct fileglob *fg, __unused vfs_context_t ctx)
 {
-        struct pipe *cpipe;
+       struct pipe *cpipe;
 
        proc_fdlock_spin(vfs_context_proc(ctx));
        cpipe = (struct pipe *)fg->fg_data;
        fg->fg_data = NULL;
        proc_fdunlock(vfs_context_proc(ctx));
-       if (cpipe)
-               pipeclose(cpipe);
+       if (cpipe) {
+               pipeclose(cpipe);
+       }
 
-       return (0);
+       return 0;
 }
 
 static void
@@ -1286,8 +1315,8 @@ pipe_free_kmem(struct pipe *cpipe)
        if (cpipe->pipe_buffer.buffer != NULL) {
                OSAddAtomic(-(cpipe->pipe_buffer.size), &amountpipekva);
                OSAddAtomic(-1, &amountpipes);
-               kfree((void *)cpipe->pipe_buffer.buffer,
-                         cpipe->pipe_buffer.size);
+               kheap_free(KHEAP_DATA_BUFFERS, cpipe->pipe_buffer.buffer,
+                   cpipe->pipe_buffer.size);
                cpipe->pipe_buffer.buffer = NULL;
                cpipe->pipe_buffer.size = 0;
        }
@@ -1301,12 +1330,7 @@ pipeclose(struct pipe *cpipe)
 {
        struct pipe *ppipe;
 
-       if (cpipe == NULL)
-               return;
-       /* partially created pipes won't have a valid mutex. */
-       if (PIPE_MTX(cpipe) != NULL)
-               PIPE_LOCK(cpipe);
-               
+       PIPE_LOCK(cpipe);
 
        /*
         * If the other side is blocked, wake it up saying that
@@ -1315,115 +1339,149 @@ pipeclose(struct pipe *cpipe)
        cpipe->pipe_state &= ~PIPE_DRAIN;
        cpipe->pipe_state |= PIPE_EOF;
        pipeselwakeup(cpipe, cpipe);
-       
+
        while (cpipe->pipe_busy) {
                cpipe->pipe_state |= PIPE_WANT;
 
                wakeup(cpipe);
-               msleep(cpipe, PIPE_MTX(cpipe), PRIBIO, "pipecl", 0);
+               msleep(cpipe, PIPE_MTX(cpipe), PRIBIO, "pipecl", 0);
        }
 
 #if CONFIG_MACF
        /*
         * Free the shared pipe label only after the two ends are disconnected.
         */
-       if (cpipe->pipe_label != NULL && cpipe->pipe_peer == NULL)
+       if (cpipe->pipe_label != NULL && cpipe->pipe_peer == NULL) {
                mac_pipe_label_destroy(cpipe);
+       }
 #endif
 
        /*
         * Disconnect from peer
         */
        if ((ppipe = cpipe->pipe_peer) != NULL) {
-
                ppipe->pipe_state &= ~(PIPE_DRAIN);
                ppipe->pipe_state |= PIPE_EOF;
 
                pipeselwakeup(ppipe, ppipe);
                wakeup(ppipe);
 
-               if (cpipe->pipe_state & PIPE_KNOTE)
-                       KNOTE(&ppipe->pipe_sel.si_note, 1);
-
-               postpipeevent(ppipe, EV_RCLOSED);
+               KNOTE(&ppipe->pipe_sel.si_note, 1);
 
                ppipe->pipe_peer = NULL;
        }
-       evpipefree(cpipe);
 
        /*
         * free resources
         */
-       if (PIPE_MTX(cpipe) != NULL) {
-               if (ppipe != NULL) {
-                       /*
-                        * since the mutex is shared and the peer is still
-                        * alive, we need to release the mutex, not free it
-                        */
-                       PIPE_UNLOCK(cpipe);
-               } else {
-                       /*
-                        * peer is gone, so we're the sole party left with
-                        * interest in this mutex... unlock and free it
-                        */
-                       PIPE_UNLOCK(cpipe);
-                       lck_mtx_free(PIPE_MTX(cpipe), pipe_mtx_grp);
-               }
+
+       PIPE_UNLOCK(cpipe);
+
+       pipepair_destroy_pipe(PIPE_PAIR(cpipe), cpipe);
+}
+
+static int64_t
+filt_pipelowwat(struct knote *kn, struct pipe *rpipe, int64_t def_lowwat)
+{
+       if ((kn->kn_sfflags & NOTE_LOWAT) == 0) {
+               return def_lowwat;
        }
-       pipe_free_kmem(cpipe);
-       if (cpipe->pipe_state & PIPE_WSELECT) {
-               pipe_garbage_collect(cpipe);
-       } else {
-               zfree(pipe_zone, cpipe);
-               pipe_garbage_collect(NULL);
+       if (rpipe->pipe_buffer.size && kn->kn_sdata > MAX_PIPESIZE(rpipe)) {
+               return MAX_PIPESIZE(rpipe);
        }
-
+       return MAX(kn->kn_sdata, def_lowwat);
 }
 
-/*ARGSUSED*/
 static int
-filt_piperead_common(struct knote *kn, struct pipe *rpipe)
+filt_pipe_draincommon(struct knote *kn, struct pipe *rpipe)
 {
-       struct pipe *wpipe;
-       int    retval;
-
-       /*
-        * we're being called back via the KNOTE post
-        * we made in pipeselwakeup, and we already hold the mutex...
-        */
+       struct pipe *wpipe = rpipe->pipe_peer;
 
-       wpipe = rpipe->pipe_peer;
-       kn->kn_data = rpipe->pipe_buffer.cnt;
        if ((rpipe->pipe_state & (PIPE_DRAIN | PIPE_EOF)) ||
            (wpipe == NULL) || (wpipe->pipe_state & (PIPE_DRAIN | PIPE_EOF))) {
                kn->kn_flags |= EV_EOF;
-               retval = 1;
+               return 1;
+       }
+
+       return 0;
+}
+
+static int
+filt_pipenotsup(struct knote *kn, long hint)
+{
+#pragma unused(hint)
+       struct pipe *rpipe = kn->kn_hook;
+
+       return filt_pipe_draincommon(kn, rpipe);
+}
+
+static int
+filt_pipenotsuptouch(struct knote *kn, struct kevent_qos_s *kev)
+{
+       struct pipe *rpipe = kn->kn_hook;
+       int res;
+
+       PIPE_LOCK(rpipe);
+
+       /* accept new kevent data (and save off lowat threshold and flag) */
+       kn->kn_sfflags = kev->fflags;
+       kn->kn_sdata = kev->data;
+
+       /* determine if any event is now deemed fired */
+       res = filt_pipe_draincommon(kn, rpipe);
+
+       PIPE_UNLOCK(rpipe);
+
+       return res;
+}
+
+static int
+filt_pipenotsupprocess(struct knote *kn, struct kevent_qos_s *kev)
+{
+       struct pipe *rpipe = kn->kn_hook;
+       int res;
+
+       PIPE_LOCK(rpipe);
+       res = filt_pipe_draincommon(kn, rpipe);
+       if (res) {
+               knote_fill_kevent(kn, kev, 0);
+       }
+       PIPE_UNLOCK(rpipe);
+
+       return res;
+}
+
+/*ARGSUSED*/
+static int
+filt_piperead_common(struct knote *kn, struct kevent_qos_s *kev, struct pipe *rpipe)
+{
+       int64_t data = rpipe->pipe_buffer.cnt;
+       int res = 0;
+
+       if (filt_pipe_draincommon(kn, rpipe)) {
+               res = 1;
        } else {
-               int64_t lowwat = 1;
-               if (kn->kn_sfflags & NOTE_LOWAT) {
-                       if (rpipe->pipe_buffer.size && kn->kn_sdata > MAX_PIPESIZE(rpipe))
-                               lowwat = MAX_PIPESIZE(rpipe);
-                       else if (kn->kn_sdata > lowwat)
-                               lowwat = kn->kn_sdata;
-               }
-               retval = kn->kn_data >= lowwat;
+               res = data >= filt_pipelowwat(kn, rpipe, 1);
+       }
+       if (res && kev) {
+               knote_fill_kevent(kn, kev, data);
        }
-       return (retval);
+       return res;
 }
 
 static int
 filt_piperead(struct knote *kn, long hint)
 {
 #pragma unused(hint)
-       struct pipe *rpipe = (struct pipe *)kn->kn_fp->f_data;
+       struct pipe *rpipe = kn->kn_hook;
 
-       return filt_piperead_common(kn, rpipe);
+       return filt_piperead_common(kn, NULL, rpipe);
 }
 
 static int
-filt_pipereadtouch(struct knote *kn, struct kevent_internal_s *kev)
+filt_pipereadtouch(struct knote *kn, struct kevent_qos_s *kev)
 {
-       struct pipe *rpipe = (struct pipe *)kn->kn_fp->f_data;
+       struct pipe *rpipe = kn->kn_hook;
        int retval;
 
        PIPE_LOCK(rpipe);
@@ -1433,7 +1491,7 @@ filt_pipereadtouch(struct knote *kn, struct kevent_internal_s *kev)
        kn->kn_sfflags = kev->fflags;
 
        /* identify if any events are now fired */
-       retval = filt_piperead_common(kn, rpipe);
+       retval = filt_piperead_common(kn, NULL, rpipe);
 
        PIPE_UNLOCK(rpipe);
 
@@ -1441,54 +1499,35 @@ filt_pipereadtouch(struct knote *kn, struct kevent_internal_s *kev)
 }
 
 static int
-filt_pipereadprocess(struct knote *kn, struct filt_process_s *data, struct kevent_internal_s *kev)
+filt_pipereadprocess(struct knote *kn, struct kevent_qos_s *kev)
 {
-#pragma unused(data)
-       struct pipe *rpipe = (struct pipe *)kn->kn_fp->f_data;
+       struct pipe *rpipe = kn->kn_hook;
        int    retval;
 
        PIPE_LOCK(rpipe);
-       retval = filt_piperead_common(kn, rpipe);
-       if (retval) {
-               *kev = kn->kn_kevent;
-               if (kn->kn_flags & EV_CLEAR) {
-                       kn->kn_fflags = 0;
-                       kn->kn_data = 0;
-               }
-       }
+       retval = filt_piperead_common(kn, kev, rpipe);
        PIPE_UNLOCK(rpipe);
 
-       return (retval);
+       return retval;
 }
 
 /*ARGSUSED*/
 static int
-filt_pipewrite_common(struct knote *kn, struct pipe *rpipe)
+filt_pipewrite_common(struct knote *kn, struct kevent_qos_s *kev, struct pipe *rpipe)
 {
-       struct pipe *wpipe;
+       int64_t data = 0;
+       int res = 0;
 
-       /*
-        * we're being called back via the KNOTE post
-        * we made in pipeselwakeup, and we already hold the mutex...
-        */
-       wpipe = rpipe->pipe_peer;
-
-       if ((wpipe == NULL) || (wpipe->pipe_state & (PIPE_DRAIN | PIPE_EOF))) {
-               kn->kn_data = 0;
-               kn->kn_flags |= EV_EOF; 
-               return (1);
+       if (filt_pipe_draincommon(kn, rpipe)) {
+               res = 1;
+       } else {
+               data = MAX_PIPESIZE(rpipe) - rpipe->pipe_buffer.cnt;
+               res = data >= filt_pipelowwat(kn, rpipe, PIPE_BUF);
        }
-       kn->kn_data = MAX_PIPESIZE(wpipe) - wpipe->pipe_buffer.cnt;
-
-       int64_t lowwat = PIPE_BUF;
-       if (kn->kn_sfflags & NOTE_LOWAT) {
-               if (wpipe->pipe_buffer.size && kn->kn_sdata > MAX_PIPESIZE(wpipe))
-                       lowwat = MAX_PIPESIZE(wpipe);
-               else if (kn->kn_sdata > lowwat)
-                       lowwat = kn->kn_sdata;
+       if (res && kev) {
+               knote_fill_kevent(kn, kev, data);
        }
-
-       return (kn->kn_data >= lowwat);
+       return res;
 }
 
 /*ARGSUSED*/
@@ -1496,16 +1535,16 @@ static int
 filt_pipewrite(struct knote *kn, long hint)
 {
 #pragma unused(hint)
-       struct pipe *rpipe = (struct pipe *)kn->kn_fp->f_data;
+       struct pipe *rpipe = kn->kn_hook;
 
-       return filt_pipewrite_common(kn, rpipe);
+       return filt_pipewrite_common(kn, NULL, rpipe);
 }
 
 
 static int
-filt_pipewritetouch(struct knote *kn, struct kevent_internal_s *kev)
+filt_pipewritetouch(struct knote *kn, struct kevent_qos_s *kev)
 {
-       struct pipe *rpipe = (struct pipe *)kn->kn_fp->f_data;
+       struct pipe *rpipe = kn->kn_hook;
        int res;
 
        PIPE_LOCK(rpipe);
@@ -1515,7 +1554,7 @@ filt_pipewritetouch(struct knote *kn, struct kevent_internal_s *kev)
        kn->kn_sdata = kev->data;
 
        /* determine if any event is now deemed fired */
-       res = filt_pipewrite_common(kn, rpipe);
+       res = filt_pipewrite_common(kn, NULL, rpipe);
 
        PIPE_UNLOCK(rpipe);
 
@@ -1523,21 +1562,13 @@ filt_pipewritetouch(struct knote *kn, struct kevent_internal_s *kev)
 }
 
 static int
-filt_pipewriteprocess(struct knote *kn, struct filt_process_s *data, struct kevent_internal_s *kev)
+filt_pipewriteprocess(struct knote *kn, struct kevent_qos_s *kev)
 {
-#pragma unused(data)
-       struct pipe *rpipe = (struct pipe *)kn->kn_fp->f_data;
+       struct pipe *rpipe = kn->kn_hook;
        int res;
 
        PIPE_LOCK(rpipe);
-       res = filt_pipewrite_common(kn, rpipe);
-       if (res) {
-               *kev = kn->kn_kevent;
-               if (kn->kn_flags & EV_CLEAR) {
-                       kn->kn_fflags = 0;
-                       kn->kn_data = 0;
-               }
-       }
+       res = filt_pipewrite_common(kn, kev, rpipe);
        PIPE_UNLOCK(rpipe);
 
        return res;
@@ -1545,10 +1576,11 @@ filt_pipewriteprocess(struct knote *kn, struct filt_process_s *data, struct keve
 
 /*ARGSUSED*/
 static int
-pipe_kqfilter(__unused struct fileproc *fp, struct knote *kn,
-               __unused struct kevent_internal_s *kev, __unused vfs_context_t ctx)
+pipe_kqfilter(struct fileproc *fp, struct knote *kn,
+    __unused struct kevent_qos_s *kev)
 {
-       struct pipe *cpipe = (struct pipe *)kn->kn_fp->f_data;
+       struct pipe *cpipe = (struct pipe *)fp->f_data;
+       struct pipe *rpipe = &PIPE_PAIR(cpipe)->pp_rpipe;
        int res;
 
        PIPE_LOCK(cpipe);
@@ -1558,49 +1590,56 @@ pipe_kqfilter(__unused struct fileproc *fp, struct knote *kn,
         * XXX process credential should have a persistent reference on it
         * XXX before being passed in here.
         */
-       if (mac_pipe_check_kqfilter(vfs_context_ucred(ctx), kn, cpipe) != 0) {
+       kauth_cred_t cred = vfs_context_ucred(vfs_context_current());
+       if (mac_pipe_check_kqfilter(cred, kn, cpipe) != 0) {
                PIPE_UNLOCK(cpipe);
-               kn->kn_flags = EV_ERROR;
-               kn->kn_data = EPERM;
+               knote_set_error(kn, EPERM);
                return 0;
        }
 #endif
 
+       /*
+        * FreeBSD will fail the attach with EPIPE if the peer pipe is detached,
+        * however, this isn't a programming error as the other side closing
+        * could race with the kevent registration.
+        *
+        * Attach should only fail for programming mistakes else it will break
+        * libdispatch.
+        *
+        * Like FreeBSD, have a "Neutered" filter that will not fire until
+        * the pipe dies if the wrong filter is attached to the wrong end.
+        *
+        * Knotes are always attached to the "rpipe".
+        */
        switch (kn->kn_filter) {
        case EVFILT_READ:
-               kn->kn_filtid = EVFILTID_PIPE_R;
-
-               /* determine initial state */
-               res = filt_piperead_common(kn, cpipe);
+               if (fp->f_flag & FREAD) {
+                       kn->kn_filtid = EVFILTID_PIPE_R;
+                       res = filt_piperead_common(kn, NULL, rpipe);
+               } else {
+                       kn->kn_filtid = EVFILTID_PIPE_N;
+                       res = filt_pipe_draincommon(kn, rpipe);
+               }
                break;
 
        case EVFILT_WRITE:
-               kn->kn_filtid = EVFILTID_PIPE_W;
-
-               if (cpipe->pipe_peer == NULL) {
-                       /*
-                        * other end of pipe has been closed
-                        */
-                       PIPE_UNLOCK(cpipe);
-                       kn->kn_flags = EV_ERROR;
-                       kn->kn_data = EPIPE;
-                       return 0;
+               if (fp->f_flag & FWRITE) {
+                       kn->kn_filtid = EVFILTID_PIPE_W;
+                       res = filt_pipewrite_common(kn, NULL, rpipe);
+               } else {
+                       kn->kn_filtid = EVFILTID_PIPE_N;
+                       res = filt_pipe_draincommon(kn, rpipe);
                }
-               if (cpipe->pipe_peer)
-               cpipe = cpipe->pipe_peer;
-
-               /* determine inital state */
-               res = filt_pipewrite_common(kn, cpipe);
                break;
+
        default:
-               PIPE_UNLOCK(cpipe);
-               kn->kn_flags = EV_ERROR;
-               kn->kn_data = EINVAL;
+               PIPE_UNLOCK(cpipe);
+               knote_set_error(kn, EINVAL);
                return 0;
        }
 
-       if (KNOTE_ATTACH(&cpipe->pipe_sel.si_note, kn))
-               cpipe->pipe_state |= PIPE_KNOTE;
+       kn->kn_hook = rpipe;
+       KNOTE_ATTACH(&rpipe->pipe_sel.si_note, kn);
 
        PIPE_UNLOCK(cpipe);
        return res;
@@ -1610,20 +1649,10 @@ static void
 filt_pipedetach(struct knote *kn)
 {
        struct pipe *cpipe = (struct pipe *)kn->kn_fp->f_data;
+       struct pipe *rpipe = &PIPE_PAIR(cpipe)->pp_rpipe;
 
        PIPE_LOCK(cpipe);
-
-       if (kn->kn_filter == EVFILT_WRITE) {
-               if (cpipe->pipe_peer == NULL) {
-                       PIPE_UNLOCK(cpipe);
-                       return;
-               }
-               cpipe = cpipe->pipe_peer;
-       }
-       if (cpipe->pipe_state & PIPE_KNOTE) {
-               if (KNOTE_DETACH(&cpipe->pipe_sel.si_note, kn))
-                       cpipe->pipe_state &= ~PIPE_KNOTE;
-       }
+       KNOTE_DETACH(&rpipe->pipe_sel.si_note, kn);
        PIPE_UNLOCK(cpipe);
 }
 
@@ -1631,47 +1660,49 @@ int
 fill_pipeinfo(struct pipe * cpipe, struct pipe_info * pinfo)
 {
 #if CONFIG_MACF
-        int error;
+       int error;
 #endif
        struct timespec now;
        struct vinfo_stat * ub;
        int pipe_size = 0;
        int pipe_count;
 
-       if (cpipe == NULL)
-               return (EBADF);
+       if (cpipe == NULL) {
+               return EBADF;
+       }
        PIPE_LOCK(cpipe);
 
 #if CONFIG_MACF
        error = mac_pipe_check_stat(kauth_cred_get(), cpipe);
        if (error) {
                PIPE_UNLOCK(cpipe);
-               return (error);
+               return error;
        }
 #endif
        if (cpipe->pipe_buffer.buffer == 0) {
-               /*
+               /*
                 * must be stat'ing the write fd
                 */
-               if (cpipe->pipe_peer) {
-                       /*
+               if (cpipe->pipe_peer) {
+                       /*
                         * the peer still exists, use it's info
                         */
-                       pipe_size  = MAX_PIPESIZE(cpipe->pipe_peer);
+                       pipe_size  = MAX_PIPESIZE(cpipe->pipe_peer);
                        pipe_count = cpipe->pipe_peer->pipe_buffer.cnt;
                } else {
                        pipe_count = 0;
                }
        } else {
-               pipe_size  = MAX_PIPESIZE(cpipe);
+               pipe_size  = MAX_PIPESIZE(cpipe);
                pipe_count = cpipe->pipe_buffer.cnt;
        }
        /*
         * since peer's buffer is setup ouside of lock
         * we might catch it in transient state
         */
-       if (pipe_size == 0)
+       if (pipe_size == 0) {
                pipe_size  = PIPE_SIZE;
+       }
 
        ub = &pinfo->pipe_stat;
 
@@ -1679,8 +1710,9 @@ fill_pipeinfo(struct pipe * cpipe, struct pipe_info * pinfo)
        ub->vst_mode = S_IFIFO | S_IRUSR | S_IWUSR | S_IRGRP | S_IWGRP;
        ub->vst_blksize = pipe_size;
        ub->vst_size = pipe_count;
-       if (ub->vst_blksize != 0)
+       if (ub->vst_blksize != 0) {
                ub->vst_blocks = (ub->vst_size + ub->vst_blksize - 1) / ub->vst_blksize;
+       }
        ub->vst_nlink = 1;
 
        ub->vst_uid = kauth_getuid();
@@ -1701,113 +1733,51 @@ fill_pipeinfo(struct pipe * cpipe, struct pipe_info * pinfo)
         * XXX (st_dev, st_ino) should be unique.
         */
 
-       pinfo->pipe_handle = (uint64_t)VM_KERNEL_ADDRPERM((uintptr_t)cpipe);
-       pinfo->pipe_peerhandle = (uint64_t)VM_KERNEL_ADDRPERM((uintptr_t)(cpipe->pipe_peer));
+       pinfo->pipe_handle = (uint64_t)VM_KERNEL_ADDRHASH((uintptr_t)cpipe);
+       pinfo->pipe_peerhandle = (uint64_t)VM_KERNEL_ADDRHASH((uintptr_t)(cpipe->pipe_peer));
        pinfo->pipe_status = cpipe->pipe_state;
 
        PIPE_UNLOCK(cpipe);
 
-       return (0);
+       return 0;
 }
 
 
-static int 
+static int
 pipe_drain(struct fileproc *fp, __unused vfs_context_t ctx)
 {
-
        /* Note: fdlock already held */
-       struct pipe *ppipe, *cpipe = (struct pipe *)(fp->f_fglob->fg_data);
+       struct pipe *ppipe, *cpipe = (struct pipe *)(fp->fp_glob->fg_data);
+       boolean_t drain_pipe = FALSE;
+
+       /* Check if the pipe is going away */
+       lck_mtx_lock_spin(&fp->fp_glob->fg_lock);
+       if (os_ref_get_count_raw(&fp->fp_glob->fg_count) == 1) {
+               drain_pipe = TRUE;
+       }
+       lck_mtx_unlock(&fp->fp_glob->fg_lock);
 
        if (cpipe) {
                PIPE_LOCK(cpipe);
-               cpipe->pipe_state |= PIPE_DRAIN; 
-               cpipe->pipe_state &= ~(PIPE_WANTR | PIPE_WANTW);
+
+               if (drain_pipe) {
+                       cpipe->pipe_state |= PIPE_DRAIN;
+                       cpipe->pipe_state &= ~(PIPE_WANTR | PIPE_WANTW);
+               }
                wakeup(cpipe);
-               
+
                /* Must wake up peer: a writer sleeps on the read side */
                if ((ppipe = cpipe->pipe_peer)) {
-                       ppipe->pipe_state |= PIPE_DRAIN;
-                       ppipe->pipe_state &= ~(PIPE_WANTR | PIPE_WANTW);
+                       if (drain_pipe) {
+                               ppipe->pipe_state |= PIPE_DRAIN;
+                               ppipe->pipe_state &= ~(PIPE_WANTR | PIPE_WANTW);
+                       }
                        wakeup(ppipe);
                }
-               
+
                PIPE_UNLOCK(cpipe);
                return 0;
        }
 
        return 1;
 }
-
-
- /*
- * When a thread sets a write-select on a pipe, it creates an implicit,
- * untracked dependency between that thread and the peer of the pipe
- * on which the select is set.  If the peer pipe is closed and freed
- * before the select()ing thread wakes up, the system will panic as
- * it attempts to unwind the dangling select().  To avoid that panic,
- * we notice whenever a dangerous select() is set on a pipe, and
- * defer the final deletion of the pipe until that select()s are all
- * resolved.  Since we can't currently detect exactly when that
- * resolution happens, we use a simple garbage collection queue to 
- * reap the at-risk pipes 'later'.
- */
-static void
-pipe_garbage_collect(struct pipe *cpipe)
-{
-       uint64_t old, now;
-       struct pipe_garbage *pgp;
-
-       /* Convert msecs to nsecs and then to abstime */
-       old = pipe_garbage_age_limit * 1000000;
-       nanoseconds_to_absolutetime(old, &old);
-
-       lck_mtx_lock(pipe_garbage_lock);
-
-       /* Free anything that's been on the queue for <mumble> seconds */
-       now = mach_absolute_time();
-       old = now - old;
-       while ((pgp = pipe_garbage_head) && pgp->pg_timestamp < old) {
-               pipe_garbage_head = pgp->pg_next;
-               if (pipe_garbage_head == NULL)
-                       pipe_garbage_tail = NULL;
-               pipe_garbage_count--;
-               zfree(pipe_zone, pgp->pg_pipe);
-               zfree(pipe_garbage_zone, pgp);
-       }
-
-       /* Add the new pipe (if any) to the tail of the garbage queue */
-       if (cpipe) {
-               cpipe->pipe_state = PIPE_DEAD;
-               pgp = (struct pipe_garbage *)zalloc(pipe_garbage_zone);
-               if (pgp == NULL) {
-                       /*
-                        * We're too low on memory to garbage collect the
-                        * pipe.  Freeing it runs the risk of panicing the
-                        * system.  All we can do is leak it and leave
-                        * a breadcrumb behind.  The good news, such as it
-                        * is, is that this will probably never happen.
-                        * We will probably hit the panic below first.
-                        */
-                       printf("Leaking pipe %p - no room left in the queue",
-                           cpipe);
-                       lck_mtx_unlock(pipe_garbage_lock);
-                       return;
-               }
-
-               pgp->pg_pipe = cpipe;
-               pgp->pg_timestamp = now;
-               pgp->pg_next = NULL;
-
-               if (pipe_garbage_tail)
-                       pipe_garbage_tail->pg_next = pgp;
-               pipe_garbage_tail = pgp;
-               if (pipe_garbage_head == NULL)
-                       pipe_garbage_head = pipe_garbage_tail;
-
-               if (pipe_garbage_count++ >= PIPE_GARBAGE_QUEUE_LIMIT)
-                       panic("Length of pipe garbage queue exceeded %d",
-                           PIPE_GARBAGE_QUEUE_LIMIT);
-       }
-       lck_mtx_unlock(pipe_garbage_lock);
-}
-