* are met.
*/
/*
- * Copyright (c) 2003-2014 Apple Inc. All rights reserved.
+ * Copyright (c) 2003-2020 Apple Inc. All rights reserved.
*
* @APPLE_OSREFERENCE_LICENSE_HEADER_START@
- *
+ *
* This file contains Original Code and/or Modifications of Original Code
* as defined in and that are subject to the Apple Public Source License
* Version 2.0 (the 'License'). You may not use this file except in
* unlawful or unlicensed copies of an Apple operating system, or to
* circumvent, violate, or enable the circumvention or violation of, any
* terms of an Apple operating system software license agreement.
- *
+ *
* Please obtain a copy of the License at
* http://www.opensource.apple.com/apsl/ and read it before using this file.
- *
+ *
* The Original Code and all software distributed under the License are
* distributed on an 'AS IS' basis, WITHOUT WARRANTY OF ANY KIND, EITHER
* EXPRESS OR IMPLIED, AND APPLE HEREBY DISCLAIMS ALL SUCH WARRANTIES,
* FITNESS FOR A PARTICULAR PURPOSE, QUIET ENJOYMENT OR NON-INFRINGEMENT.
* Please see the License for the specific language governing rights and
* limitations under the License.
- *
+ *
* @APPLE_OSREFERENCE_LICENSE_HEADER_END@
*/
/*
* do.
*
* Pipes are implemented as circular buffers. Following are the valid states in pipes operations
- *
+ *
* _________________________________
* 1. |_________________________________| r=w, c=0
- *
+ *
* _________________________________
* 2. |__r:::::wc_______________________| r <= w , c > 0
*
* a-z define the steps in a program flow
* 1-4 are the states as defined aboe
* Action: is what file operation is done on the pipe
- *
+ *
* Current:None Action: initialize with size M=200
* a. State 1 ( r=0, w=0, c=0)
- *
+ *
* Current: a Action: write(100) (w < M)
* b. State 2 (r=0, w=100, c=100)
- *
+ *
* Current: b Action: write(100) (w = M-w)
* c. State 4 (r=0,w=0,c=200)
- *
+ *
* Current: b Action: read(70) ( r < c )
* d. State 2(r=70,w=100,c=30)
- *
+ *
* Current: d Action: write(75) ( w < (m-w))
* e. State 2 (r=70,w=175,c=105)
- *
+ *
* Current: d Action: write(110) ( w > (m-w))
* f. State 3 (r=70,w=10,c=140)
- *
+ *
* Current: d Action: read(30) (r >= c )
* g. State 1 (r=100,w=100,c=0)
- *
+ *
*/
/*
* dynamically change to larger sizes based on usage. The buffer size is never
* reduced. The total amount of kernel memory used is governed by maxpipekva.
* In case of dynamic expansion limit is reached, the output thread is blocked
- * until the pipe buffer empties enough to continue.
+ * until the pipe buffer empties enough to continue.
*
* In order to limit the resource use of pipes, two sysctls exist:
*
* kern.ipc.maxpipekva - This is a hard limit on the amount of pageable
- * address space available to us in pipe_map.
+ * address space available to us in pipe_map.
*
* Memory usage may be monitored through the sysctls
* kern.ipc.pipes, kern.ipc.pipekva.
#include <security/mac_framework.h>
#endif
-#define f_flag f_fglob->fg_flag
-#define f_msgcount f_fglob->fg_msgcount
-#define f_cred f_fglob->fg_cred
-#define f_ops f_fglob->fg_ops
-#define f_offset f_fglob->fg_offset
-#define f_data f_fglob->fg_data
+#define f_flag fp_glob->fg_flag
+#define f_ops fp_glob->fg_ops
+#define f_data fp_glob->fg_data
+
+struct pipepair {
+ lck_mtx_t pp_mtx;
+ struct pipe pp_rpipe;
+ struct pipe pp_wpipe;
+ uint64_t pp_pipe_id; /* unique ID shared by both pipe ends */
+};
+
+#define PIPE_PAIR(pipe) \
+ __container_of(PIPE_MTX(pipe), struct pipepair, pp_mtx)
/*
- * interfaces to the outside world exported through file operations
+ * interfaces to the outside world exported through file operations
*/
static int pipe_read(struct fileproc *fp, struct uio *uio,
- int flags, vfs_context_t ctx);
+ int flags, vfs_context_t ctx);
static int pipe_write(struct fileproc *fp, struct uio *uio,
- int flags, vfs_context_t ctx);
+ int flags, vfs_context_t ctx);
static int pipe_close(struct fileglob *fg, vfs_context_t ctx);
static int pipe_select(struct fileproc *fp, int which, void * wql,
- vfs_context_t ctx);
+ vfs_context_t ctx);
static int pipe_kqfilter(struct fileproc *fp, struct knote *kn,
- struct kevent_internal_s *kev, vfs_context_t ctx);
+ struct kevent_qos_s *kev);
static int pipe_ioctl(struct fileproc *fp, u_long cmd, caddr_t data,
- vfs_context_t ctx);
-static int pipe_drain(struct fileproc *fp,vfs_context_t ctx);
+ vfs_context_t ctx);
+static int pipe_drain(struct fileproc *fp, vfs_context_t ctx);
static const struct fileops pipeops = {
- .fo_type = DTYPE_PIPE,
- .fo_read = pipe_read,
- .fo_write = pipe_write,
- .fo_ioctl = pipe_ioctl,
- .fo_select = pipe_select,
- .fo_close = pipe_close,
+ .fo_type = DTYPE_PIPE,
+ .fo_read = pipe_read,
+ .fo_write = pipe_write,
+ .fo_ioctl = pipe_ioctl,
+ .fo_select = pipe_select,
+ .fo_close = pipe_close,
+ .fo_drain = pipe_drain,
.fo_kqfilter = pipe_kqfilter,
- .fo_drain = pipe_drain,
};
static void filt_pipedetach(struct knote *kn);
+static int filt_pipenotsup(struct knote *kn, long hint);
+static int filt_pipenotsuptouch(struct knote *kn, struct kevent_qos_s *kev);
+static int filt_pipenotsupprocess(struct knote *kn, struct kevent_qos_s *kev);
+
static int filt_piperead(struct knote *kn, long hint);
-static int filt_pipereadtouch(struct knote *kn, struct kevent_internal_s *kev);
-static int filt_pipereadprocess(struct knote *kn, struct filt_process_s *data, struct kevent_internal_s *kev);
+static int filt_pipereadtouch(struct knote *kn, struct kevent_qos_s *kev);
+static int filt_pipereadprocess(struct knote *kn, struct kevent_qos_s *kev);
static int filt_pipewrite(struct knote *kn, long hint);
-static int filt_pipewritetouch(struct knote *kn, struct kevent_internal_s *kev);
-static int filt_pipewriteprocess(struct knote *kn, struct filt_process_s *data, struct kevent_internal_s *kev);
+static int filt_pipewritetouch(struct knote *kn, struct kevent_qos_s *kev);
+static int filt_pipewriteprocess(struct knote *kn, struct kevent_qos_s *kev);
+
+SECURITY_READ_ONLY_EARLY(struct filterops) pipe_nfiltops = {
+ .f_isfd = 1,
+ .f_detach = filt_pipedetach,
+ .f_event = filt_pipenotsup,
+ .f_touch = filt_pipenotsuptouch,
+ .f_process = filt_pipenotsupprocess,
+};
SECURITY_READ_ONLY_EARLY(struct filterops) pipe_rfiltops = {
- .f_isfd = 1,
- .f_detach = filt_pipedetach,
- .f_event = filt_piperead,
- .f_touch = filt_pipereadtouch,
+ .f_isfd = 1,
+ .f_detach = filt_pipedetach,
+ .f_event = filt_piperead,
+ .f_touch = filt_pipereadtouch,
.f_process = filt_pipereadprocess,
};
SECURITY_READ_ONLY_EARLY(struct filterops) pipe_wfiltops = {
- .f_isfd = 1,
- .f_detach = filt_pipedetach,
- .f_event = filt_pipewrite,
- .f_touch = filt_pipewritetouch,
+ .f_isfd = 1,
+ .f_detach = filt_pipedetach,
+ .f_event = filt_pipewrite,
+ .f_touch = filt_pipewritetouch,
.f_process = filt_pipewriteprocess,
};
+#if PIPE_SYSCTLS
static int nbigpipe; /* for compatibility sake. no longer used */
+#endif
static int amountpipes; /* total number of pipes in system */
static int amountpipekva; /* total memory used by pipes */
+static _Atomic uint64_t pipe_unique_id = 1;
+
int maxpipekva __attribute__((used)) = PIPE_KVAMAX; /* allowing 16MB max. */
#if PIPE_SYSCTLS
SYSCTL_DECL(_kern_ipc);
-SYSCTL_INT(_kern_ipc, OID_AUTO, maxpipekva, CTLFLAG_RD|CTLFLAG_LOCKED,
- &maxpipekva, 0, "Pipe KVA limit");
-SYSCTL_INT(_kern_ipc, OID_AUTO, maxpipekvawired, CTLFLAG_RW|CTLFLAG_LOCKED,
- &maxpipekvawired, 0, "Pipe KVA wired limit");
-SYSCTL_INT(_kern_ipc, OID_AUTO, pipes, CTLFLAG_RD|CTLFLAG_LOCKED,
- &amountpipes, 0, "Current # of pipes");
-SYSCTL_INT(_kern_ipc, OID_AUTO, bigpipes, CTLFLAG_RD|CTLFLAG_LOCKED,
- &nbigpipe, 0, "Current # of big pipes");
-SYSCTL_INT(_kern_ipc, OID_AUTO, pipekva, CTLFLAG_RD|CTLFLAG_LOCKED,
- &amountpipekva, 0, "Pipe KVA usage");
-SYSCTL_INT(_kern_ipc, OID_AUTO, pipekvawired, CTLFLAG_RD|CTLFLAG_LOCKED,
- &amountpipekvawired, 0, "Pipe wired KVA usage");
+SYSCTL_INT(_kern_ipc, OID_AUTO, maxpipekva, CTLFLAG_RD | CTLFLAG_LOCKED,
+ &maxpipekva, 0, "Pipe KVA limit");
+SYSCTL_INT(_kern_ipc, OID_AUTO, maxpipekvawired, CTLFLAG_RW | CTLFLAG_LOCKED,
+ &maxpipekvawired, 0, "Pipe KVA wired limit");
+SYSCTL_INT(_kern_ipc, OID_AUTO, pipes, CTLFLAG_RD | CTLFLAG_LOCKED,
+ &amountpipes, 0, "Current # of pipes");
+SYSCTL_INT(_kern_ipc, OID_AUTO, bigpipes, CTLFLAG_RD | CTLFLAG_LOCKED,
+ &nbigpipe, 0, "Current # of big pipes");
+SYSCTL_INT(_kern_ipc, OID_AUTO, pipekva, CTLFLAG_RD | CTLFLAG_LOCKED,
+ &amountpipekva, 0, "Pipe KVA usage");
+SYSCTL_INT(_kern_ipc, OID_AUTO, pipekvawired, CTLFLAG_RD | CTLFLAG_LOCKED,
+ &amountpipekvawired, 0, "Pipe wired KVA usage");
#endif
+static int pipepair_alloc(struct pipe **rpipe, struct pipe **wpipe);
static void pipeclose(struct pipe *cpipe);
static void pipe_free_kmem(struct pipe *cpipe);
-static int pipe_create(struct pipe **cpipep);
static int pipespace(struct pipe *cpipe, int size);
static int choose_pipespace(unsigned long current, unsigned long expected);
static int expand_pipespace(struct pipe *p, int target_size);
static __inline int pipeio_lock(struct pipe *cpipe, int catch);
static __inline void pipeio_unlock(struct pipe *cpipe);
-extern int postpipeevent(struct pipe *, int);
-extern void evpipefree(struct pipe *cpipe);
-
-static lck_grp_t *pipe_mtx_grp;
-static lck_attr_t *pipe_mtx_attr;
-static lck_grp_attr_t *pipe_mtx_grp_attr;
-
-static zone_t pipe_zone;
+static LCK_GRP_DECLARE(pipe_mtx_grp, "pipe");
+static ZONE_DECLARE(pipe_zone, "pipe zone", sizeof(struct pipepair), ZC_NONE);
-#define MAX_PIPESIZE(pipe) ( MAX(PIPE_SIZE, (pipe)->pipe_buffer.size) )
-
-#define PIPE_GARBAGE_AGE_LIMIT 5000 /* In milliseconds */
-#define PIPE_GARBAGE_QUEUE_LIMIT 32000
-
-struct pipe_garbage {
- struct pipe *pg_pipe;
- struct pipe_garbage *pg_next;
- uint64_t pg_timestamp;
-};
-
-static zone_t pipe_garbage_zone;
-static struct pipe_garbage *pipe_garbage_head = NULL;
-static struct pipe_garbage *pipe_garbage_tail = NULL;
-static uint64_t pipe_garbage_age_limit = PIPE_GARBAGE_AGE_LIMIT;
-static int pipe_garbage_count = 0;
-static lck_mtx_t *pipe_garbage_lock;
-static void pipe_garbage_collect(struct pipe *cpipe);
+#define MAX_PIPESIZE(pipe) ( MAX(PIPE_SIZE, (pipe)->pipe_buffer.size) )
SYSINIT(vfs, SI_SUB_VFS, SI_ORDER_ANY, pipeinit, NULL);
-/* initial setup done at time of sysinit */
-void
-pipeinit(void)
-{
- nbigpipe=0;
- vm_size_t zone_size;
-
- zone_size = 8192 * sizeof(struct pipe);
- pipe_zone = zinit(sizeof(struct pipe), zone_size, 4096, "pipe zone");
-
-
- /* allocate lock group attribute and group for pipe mutexes */
- pipe_mtx_grp_attr = lck_grp_attr_alloc_init();
- pipe_mtx_grp = lck_grp_alloc_init("pipe", pipe_mtx_grp_attr);
-
- /* allocate the lock attribute for pipe mutexes */
- pipe_mtx_attr = lck_attr_alloc_init();
-
- /*
- * Set up garbage collection for dead pipes
- */
- zone_size = (PIPE_GARBAGE_QUEUE_LIMIT + 20) *
- sizeof(struct pipe_garbage);
- pipe_garbage_zone = (zone_t)zinit(sizeof(struct pipe_garbage),
- zone_size, 4096, "pipe garbage zone");
- pipe_garbage_lock = lck_mtx_alloc_init(pipe_mtx_grp, pipe_mtx_attr);
-
-}
-
-#ifndef CONFIG_EMBEDDED
+#if defined(XNU_TARGET_OS_OSX)
/* Bitmap for things to touch in pipe_touch() */
-#define PIPE_ATIME 0x00000001 /* time of last access */
-#define PIPE_MTIME 0x00000002 /* time of last modification */
-#define PIPE_CTIME 0x00000004 /* time of last status change */
+#define PIPE_ATIME 0x00000001 /* time of last access */
+#define PIPE_MTIME 0x00000002 /* time of last modification */
+#define PIPE_CTIME 0x00000004 /* time of last status change */
static void
pipe_touch(struct pipe *tpipe, int touch)
}
#endif
-static const unsigned int pipesize_blocks[] = {512,1024,2048,4096, 4096 * 2, PIPE_SIZE , PIPE_SIZE * 4 };
+static const unsigned int pipesize_blocks[] = {512, 1024, 2048, 4096, 4096 * 2, PIPE_SIZE, PIPE_SIZE * 4 };
-/*
- * finds the right size from possible sizes in pipesize_blocks
- * returns the size which matches max(current,expected)
+/*
+ * finds the right size from possible sizes in pipesize_blocks
+ * returns the size which matches max(current,expected)
*/
-static int
+static int
choose_pipespace(unsigned long current, unsigned long expected)
{
- int i = sizeof(pipesize_blocks)/sizeof(unsigned int) -1;
+ int i = sizeof(pipesize_blocks) / sizeof(unsigned int) - 1;
unsigned long target;
/*
*/
assert(PIPE_BUF == pipesize_blocks[0]);
- if (expected > current)
+ if (expected > current) {
target = expected;
- else
+ } else {
target = current;
+ }
- while ( i >0 && pipesize_blocks[i-1] > target) {
- i=i-1;
-
+ while (i > 0 && pipesize_blocks[i - 1] > target) {
+ i = i - 1;
}
-
+
return pipesize_blocks[i];
}
* Required: PIPE_LOCK and io lock to be held by caller.
* returns 0 on success or no expansion possible
*/
-static int
+static int
expand_pipespace(struct pipe *p, int target_size)
{
struct pipe tmp, oldpipe;
int error;
tmp.pipe_buffer.buffer = 0;
-
+
if (p->pipe_buffer.size >= (unsigned) target_size) {
return 0; /* the existing buffer is max size possible */
}
-
+
/* create enough space in the target */
error = pipespace(&tmp, target_size);
- if (error != 0)
- return (error);
+ if (error != 0) {
+ return error;
+ }
oldpipe.pipe_buffer.buffer = p->pipe_buffer.buffer;
oldpipe.pipe_buffer.size = p->pipe_buffer.size;
-
+
memcpy(tmp.pipe_buffer.buffer, p->pipe_buffer.buffer, p->pipe_buffer.size);
- if (p->pipe_buffer.cnt > 0 && p->pipe_buffer.in <= p->pipe_buffer.out ){
+ if (p->pipe_buffer.cnt > 0 && p->pipe_buffer.in <= p->pipe_buffer.out) {
/* we are in State 3 and need extra copying for read to be consistent */
memcpy(&tmp.pipe_buffer.buffer[p->pipe_buffer.size], p->pipe_buffer.buffer, p->pipe_buffer.size);
p->pipe_buffer.in += p->pipe_buffer.size;
/*
* The pipe system call for the DTYPE_PIPE type of pipes
- *
+ *
* returns:
- * FREAD | fd0 | -->[struct rpipe] --> |~~buffer~~| \
+ * FREAD | fd0 | -->[struct rpipe] --> |~~buffer~~| \
* (pipe_mutex)
- * FWRITE | fd1 | -->[struct wpipe] --X /
+ * FWRITE | fd1 | -->[struct wpipe] --X /
*/
/* ARGSUSED */
{
struct fileproc *rf, *wf;
struct pipe *rpipe, *wpipe;
- lck_mtx_t *pmtx;
- int fd, error;
-
- if ((pmtx = lck_mtx_alloc_init(pipe_mtx_grp, pipe_mtx_attr)) == NULL)
- return (ENOMEM);
-
- rpipe = wpipe = NULL;
- if (pipe_create(&rpipe) || pipe_create(&wpipe)) {
- error = ENFILE;
- goto freepipes;
- }
- /*
- * allocate the space for the normal I/O direction up
- * front... we'll delay the allocation for the other
- * direction until a write actually occurs (most likely it won't)...
- */
- error = pipespace(rpipe, choose_pipespace(rpipe->pipe_buffer.size, 0));
- if (error)
- goto freepipes;
-
- TAILQ_INIT(&rpipe->pipe_evlist);
- TAILQ_INIT(&wpipe->pipe_evlist);
+ int error;
- error = falloc(p, &rf, &fd, vfs_context_current());
+ error = pipepair_alloc(&rpipe, &wpipe);
if (error) {
- goto freepipes;
+ return error;
}
- retval[0] = fd;
/*
- * for now we'll create half-duplex pipes(refer returns section above).
+ * for now we'll create half-duplex pipes(refer returns section above).
* this is what we've always supported..
*/
+
+ error = falloc(p, &rf, &retval[0], vfs_context_current());
+ if (error) {
+ goto freepipes;
+ }
rf->f_flag = FREAD;
rf->f_data = (caddr_t)rpipe;
rf->f_ops = &pipeops;
- error = falloc(p, &wf, &fd, vfs_context_current());
+ error = falloc(p, &wf, &retval[1], vfs_context_current());
if (error) {
fp_free(p, retval[0], rf);
- goto freepipes;
+ goto freepipes;
}
wf->f_flag = FWRITE;
wf->f_data = (caddr_t)wpipe;
rpipe->pipe_peer = wpipe;
wpipe->pipe_peer = rpipe;
- /* both structures share the same mutex */
- rpipe->pipe_mtxp = wpipe->pipe_mtxp = pmtx;
- retval[1] = fd;
#if CONFIG_MACF
/*
* XXXXXXXX SHOULD NOT HOLD FILE_LOCK() XXXXXXXXXXXX
fp_drop(p, retval[0], rf, 1);
fp_drop(p, retval[1], wf, 1);
proc_fdunlock(p);
-
-
- return (0);
+ return 0;
freepipes:
- pipeclose(rpipe);
- pipeclose(wpipe);
- lck_mtx_free(pmtx, pipe_mtx_grp);
-
- return (error);
+ pipeclose(rpipe);
+ pipeclose(wpipe);
+ return error;
}
int
pipe_stat(struct pipe *cpipe, void *ub, int isstat64)
{
#if CONFIG_MACF
- int error;
+ int error;
#endif
- int pipe_size = 0;
- int pipe_count;
- struct stat *sb = (struct stat *)0; /* warning avoidance ; protected by isstat64 */
+ int pipe_size = 0;
+ int pipe_count;
+ struct stat *sb = (struct stat *)0; /* warning avoidance ; protected by isstat64 */
struct stat64 * sb64 = (struct stat64 *)0; /* warning avoidance ; protected by isstat64 */
- if (cpipe == NULL)
- return (EBADF);
+ if (cpipe == NULL) {
+ return EBADF;
+ }
PIPE_LOCK(cpipe);
#if CONFIG_MACF
error = mac_pipe_check_stat(kauth_cred_get(), cpipe);
if (error) {
PIPE_UNLOCK(cpipe);
- return (error);
+ return error;
}
#endif
if (cpipe->pipe_buffer.buffer == 0) {
- /* must be stat'ing the write fd */
- if (cpipe->pipe_peer) {
- /* the peer still exists, use it's info */
- pipe_size = MAX_PIPESIZE(cpipe->pipe_peer);
+ /* must be stat'ing the write fd */
+ if (cpipe->pipe_peer) {
+ /* the peer still exists, use it's info */
+ pipe_size = MAX_PIPESIZE(cpipe->pipe_peer);
pipe_count = cpipe->pipe_peer->pipe_buffer.cnt;
} else {
pipe_count = 0;
}
} else {
- pipe_size = MAX_PIPESIZE(cpipe);
+ pipe_size = MAX_PIPESIZE(cpipe);
pipe_count = cpipe->pipe_buffer.cnt;
}
/*
* since peer's buffer is setup ouside of lock
* we might catch it in transient state
*/
- if (pipe_size == 0)
+ if (pipe_size == 0) {
pipe_size = MAX(PIPE_SIZE, pipesize_blocks[0]);
+ }
if (isstat64 != 0) {
- sb64 = (struct stat64 *)ub;
+ sb64 = (struct stat64 *)ub;
bzero(sb64, sizeof(*sb64));
sb64->st_mode = S_IFIFO | S_IRUSR | S_IWUSR | S_IRGRP | S_IWGRP;
sb64->st_blksize = pipe_size;
sb64->st_size = pipe_count;
sb64->st_blocks = (sb64->st_size + sb64->st_blksize - 1) / sb64->st_blksize;
-
+
sb64->st_uid = kauth_getuid();
sb64->st_gid = kauth_getgid();
-
+
sb64->st_atimespec.tv_sec = cpipe->st_atimespec.tv_sec;
sb64->st_atimespec.tv_nsec = cpipe->st_atimespec.tv_nsec;
-
+
sb64->st_mtimespec.tv_sec = cpipe->st_mtimespec.tv_sec;
sb64->st_mtimespec.tv_nsec = cpipe->st_mtimespec.tv_nsec;
sb64->st_ctimespec.tv_nsec = cpipe->st_ctimespec.tv_nsec;
/*
- * Return a relatively unique inode number based on the current
- * address of this pipe's struct pipe. This number may be recycled
- * relatively quickly.
- */
- sb64->st_ino = (ino64_t)VM_KERNEL_ADDRPERM((uintptr_t)cpipe);
+ * Return a relatively unique inode number based on the current
+ * address of this pipe's struct pipe. This number may be recycled
+ * relatively quickly.
+ */
+ sb64->st_ino = (ino64_t)VM_KERNEL_ADDRHASH((uintptr_t)cpipe);
} else {
- sb = (struct stat *)ub;
+ sb = (struct stat *)ub;
bzero(sb, sizeof(*sb));
sb->st_mode = S_IFIFO | S_IRUSR | S_IWUSR | S_IRGRP | S_IWGRP;
sb->st_blksize = pipe_size;
sb->st_size = pipe_count;
sb->st_blocks = (sb->st_size + sb->st_blksize - 1) / sb->st_blksize;
-
+
sb->st_uid = kauth_getuid();
sb->st_gid = kauth_getgid();
-
+
sb->st_atimespec.tv_sec = cpipe->st_atimespec.tv_sec;
sb->st_atimespec.tv_nsec = cpipe->st_atimespec.tv_nsec;
-
+
sb->st_mtimespec.tv_sec = cpipe->st_mtimespec.tv_sec;
sb->st_mtimespec.tv_nsec = cpipe->st_mtimespec.tv_nsec;
sb->st_ctimespec.tv_nsec = cpipe->st_ctimespec.tv_nsec;
/*
- * Return a relatively unique inode number based on the current
- * address of this pipe's struct pipe. This number may be recycled
- * relatively quickly.
- */
- sb->st_ino = (ino_t)VM_KERNEL_ADDRPERM((uintptr_t)cpipe);
+ * Return a relatively unique inode number based on the current
+ * address of this pipe's struct pipe. This number may be recycled
+ * relatively quickly.
+ */
+ sb->st_ino = (ino_t)VM_KERNEL_ADDRHASH((uintptr_t)cpipe);
}
PIPE_UNLOCK(cpipe);
* XXX is associated with pipes, since they are implemented via a
* XXX struct fileops indirection rather than as FS objects.
*/
- return (0);
+ return 0;
}
+uint64_t
+pipe_id(struct pipe *p)
+{
+ return PIPE_PAIR(p)->pp_pipe_id;
+}
/*
* Allocate kva for pipe circular buffer, the space is pageable
{
vm_offset_t buffer;
- if (size <= 0)
- return(EINVAL);
+ if (size <= 0) {
+ return EINVAL;
+ }
- if ((buffer = (vm_offset_t)kalloc(size)) == 0 )
- return(ENOMEM);
+ buffer = (vm_offset_t)kheap_alloc(KHEAP_DATA_BUFFERS, size, Z_WAITOK);
+ if (!buffer) {
+ return ENOMEM;
+ }
/* free old resources if we're resizing */
pipe_free_kmem(cpipe);
OSAddAtomic(1, &amountpipes);
OSAddAtomic(cpipe->pipe_buffer.size, &amountpipekva);
- return (0);
+ return 0;
}
/*
* initialize and allocate VM and memory for pipe
*/
static int
-pipe_create(struct pipe **cpipep)
+pipepair_alloc(struct pipe **rp_out, struct pipe **wp_out)
{
- struct pipe *cpipe;
- cpipe = (struct pipe *)zalloc(pipe_zone);
+ struct pipepair *pp = zalloc(pipe_zone);
+ struct pipe *rpipe = &pp->pp_rpipe;
+ struct pipe *wpipe = &pp->pp_wpipe;
- if ((*cpipep = cpipe) == NULL)
- return (ENOMEM);
+ if (pp == NULL) {
+ return ENOMEM;
+ }
/*
* protect so pipespace or pipeclose don't follow a junk pointer
* if pipespace() fails.
*/
- bzero(cpipe, sizeof *cpipe);
+ bzero(pp, sizeof(struct pipepair));
+ pp->pp_pipe_id = os_atomic_inc_orig(&pipe_unique_id, relaxed);
+ lck_mtx_init(&pp->pp_mtx, &pipe_mtx_grp, LCK_ATTR_NULL);
-#ifndef CONFIG_EMBEDDED
+ rpipe->pipe_mtxp = &pp->pp_mtx;
+ wpipe->pipe_mtxp = &pp->pp_mtx;
+
+#if defined(XNU_TARGET_OS_OSX)
/* Initial times are all the time of creation of the pipe */
- pipe_touch(cpipe, PIPE_ATIME | PIPE_MTIME | PIPE_CTIME);
+ pipe_touch(rpipe, PIPE_ATIME | PIPE_MTIME | PIPE_CTIME);
+ pipe_touch(wpipe, PIPE_ATIME | PIPE_MTIME | PIPE_CTIME);
#endif
- return (0);
+
+ /*
+ * allocate the space for the normal I/O direction up
+ * front... we'll delay the allocation for the other
+ * direction until a write actually occurs (most likely it won't)...
+ */
+ int error = pipespace(rpipe, choose_pipespace(rpipe->pipe_buffer.size, 0));
+ if (__improbable(error)) {
+ lck_mtx_destroy(&pp->pp_mtx, &pipe_mtx_grp);
+ zfree(pipe_zone, pp);
+ return error;
+ }
+
+ *rp_out = rpipe;
+ *wp_out = wpipe;
+ return 0;
}
+static void
+pipepair_destroy_pipe(struct pipepair *pp, struct pipe *cpipe)
+{
+ bool can_free;
+
+ pipe_free_kmem(cpipe);
+
+ lck_mtx_lock(&pp->pp_mtx);
+ if (__improbable(cpipe->pipe_state & PIPE_DEAD)) {
+ panic("double free of pipe %p in pair %p", cpipe, pp);
+ }
+
+ cpipe->pipe_state |= PIPE_DEAD;
+
+ can_free = (pp->pp_rpipe.pipe_state & PIPE_DEAD) &&
+ (pp->pp_wpipe.pipe_state & PIPE_DEAD);
+ lck_mtx_unlock(&pp->pp_mtx);
+
+ if (can_free) {
+ lck_mtx_destroy(&pp->pp_mtx, &pipe_mtx_grp);
+ zfree(pipe_zone, pp);
+ }
+}
/*
* lock a pipe for I/O, blocking other access
while (cpipe->pipe_state & PIPE_LOCKFL) {
cpipe->pipe_state |= PIPE_LWANT;
error = msleep(cpipe, PIPE_MTX(cpipe), catch ? (PRIBIO | PCATCH) : PRIBIO,
- "pipelk", 0);
- if (error != 0)
- return (error);
+ "pipelk", 0);
+ if (error != 0) {
+ return error;
+ }
}
cpipe->pipe_state |= PIPE_LOCKFL;
- return (0);
+ return 0;
}
/*
cpipe->pipe_state &= ~PIPE_SEL;
selwakeup(&cpipe->pipe_sel);
}
- if (cpipe->pipe_state & PIPE_KNOTE)
- KNOTE(&cpipe->pipe_sel.si_note, 1);
- postpipeevent(cpipe, EV_RWBYTES);
+ KNOTE(&cpipe->pipe_sel.si_note, 1);
if (spipe && (spipe->pipe_state & PIPE_ASYNC) && spipe->pipe_pgid) {
- if (spipe->pipe_pgid < 0)
- gsignal(-spipe->pipe_pgid, SIGIO);
- else
- proc_signal(spipe->pipe_pgid, SIGIO);
- }
+ if (spipe->pipe_pgid < 0) {
+ gsignal(-spipe->pipe_pgid, SIGIO);
+ } else {
+ proc_signal(spipe->pipe_pgid, SIGIO);
+ }
+ }
}
/*
/* ARGSUSED */
static int
pipe_read(struct fileproc *fp, struct uio *uio, __unused int flags,
- __unused vfs_context_t ctx)
+ __unused vfs_context_t ctx)
{
struct pipe *rpipe = (struct pipe *)fp->f_data;
int error;
++rpipe->pipe_busy;
error = pipeio_lock(rpipe, 1);
- if (error)
+ if (error) {
goto unlocked_error;
+ }
#if CONFIG_MACF
error = mac_pipe_check_read(kauth_cred_get(), rpipe);
- if (error)
+ if (error) {
goto locked_error;
+ }
#endif
if (rpipe->pipe_buffer.cnt > 0) {
/*
* # bytes to read is min( bytes from read pointer until end of buffer,
- * total unread bytes,
+ * total unread bytes,
* user requested byte count)
*/
size = rpipe->pipe_buffer.size - rpipe->pipe_buffer.out;
- if (size > rpipe->pipe_buffer.cnt)
+ if (size > rpipe->pipe_buffer.cnt) {
size = rpipe->pipe_buffer.cnt;
- // LP64todo - fix this!
- if (size > (u_int) uio_resid(uio))
- size = (u_int) uio_resid(uio);
+ }
+
+ size = (u_int) MIN(INT_MAX, MIN((user_size_t)size,
+ (user_size_t)uio_resid(uio)));
PIPE_UNLOCK(rpipe); /* we still hold io lock.*/
error = uiomove(
- &rpipe->pipe_buffer.buffer[rpipe->pipe_buffer.out],
- size, uio);
+ &rpipe->pipe_buffer.buffer[rpipe->pipe_buffer.out],
+ size, uio);
PIPE_LOCK(rpipe);
- if (error)
+ if (error) {
break;
+ }
rpipe->pipe_buffer.out += size;
- if (rpipe->pipe_buffer.out >= rpipe->pipe_buffer.size)
+ if (rpipe->pipe_buffer.out >= rpipe->pipe_buffer.size) {
rpipe->pipe_buffer.out = 0;
+ }
rpipe->pipe_buffer.cnt -= size;
-
+
/*
* If there is no more to read in the pipe, reset
* its pointers to the beginning. This improves
* detect EOF condition
* read returns 0 on EOF, no need to set error
*/
- if (rpipe->pipe_state & (PIPE_DRAIN | PIPE_EOF)) {
+ if ((rpipe->pipe_state & (PIPE_DRAIN | PIPE_EOF)) ||
+ (fileproc_get_vflags(fp) & FPV_DRAIN)) {
break;
}
/*
* Break if some data was read in previous iteration.
*/
- if (nread > 0)
+ if (nread > 0) {
break;
+ }
/*
- * Unlock the pipe buffer for our remaining processing.
+ * Unlock the pipe buffer for our remaining processing.
* We will either break out with an error or we will
* sleep and relock to loop.
*/
} else {
rpipe->pipe_state |= PIPE_WANTR;
error = msleep(rpipe, PIPE_MTX(rpipe), PRIBIO | PCATCH, "piperd", 0);
- if (error == 0)
- error = pipeio_lock(rpipe, 1);
+ if (error == 0) {
+ error = pipeio_lock(rpipe, 1);
+ }
}
- if (error)
+ if (error) {
goto unlocked_error;
+ }
}
}
#if CONFIG_MACF
* PIPE_WANT processing only makes sense if pipe_busy is 0.
*/
if ((rpipe->pipe_busy == 0) && (rpipe->pipe_state & PIPE_WANT)) {
- rpipe->pipe_state &= ~(PIPE_WANT|PIPE_WANTW);
+ rpipe->pipe_state &= ~(PIPE_WANT | PIPE_WANTW);
wakeup(rpipe);
} else if (rpipe->pipe_buffer.cnt < rpipe->pipe_buffer.size) {
/*
}
}
- if ((rpipe->pipe_buffer.size - rpipe->pipe_buffer.cnt) > 0)
+ if ((rpipe->pipe_buffer.size - rpipe->pipe_buffer.cnt) > 0) {
pipeselwakeup(rpipe, rpipe->pipe_peer);
+ }
-#ifndef CONFIG_EMBEDDED
+#if defined(XNU_TARGET_OS_OSX)
/* update last read time */
pipe_touch(rpipe, PIPE_ATIME);
#endif
PIPE_UNLOCK(rpipe);
- return (error);
+ return error;
}
/*
- * perform a write of n bytes into the read side of buffer. Since
+ * perform a write of n bytes into the read side of buffer. Since
* pipes are unidirectional a write is meant to be read by the otherside only.
*/
static int
pipe_write(struct fileproc *fp, struct uio *uio, __unused int flags,
- __unused vfs_context_t ctx)
+ __unused vfs_context_t ctx)
{
int error = 0;
- int orig_resid;
+ size_t orig_resid;
int pipe_size;
struct pipe *wpipe, *rpipe;
// LP64todo - fix this!
- orig_resid = uio_resid(uio);
+ orig_resid = (size_t)uio_resid(uio);
+ if (orig_resid > LONG_MAX) {
+ return EINVAL;
+ }
int space;
rpipe = (struct pipe *)fp->f_data;
/*
* detect loss of pipe read side, issue SIGPIPE if lost.
*/
- if (wpipe == NULL || (wpipe->pipe_state & (PIPE_DRAIN | PIPE_EOF))) {
+ if (wpipe == NULL || (wpipe->pipe_state & (PIPE_DRAIN | PIPE_EOF)) ||
+ (fileproc_get_vflags(fp) & FPV_DRAIN)) {
PIPE_UNLOCK(rpipe);
- return (EPIPE);
+ return EPIPE;
}
#if CONFIG_MACF
error = mac_pipe_check_write(kauth_cred_get(), wpipe);
if (error) {
PIPE_UNLOCK(rpipe);
- return (error);
+ return error;
}
#endif
++wpipe->pipe_busy;
* fd[1], so allocating space for both ends is a waste...
*/
- if ( wpipe->pipe_buffer.buffer == 0 || (
- (unsigned)orig_resid > wpipe->pipe_buffer.size - wpipe->pipe_buffer.cnt &&
- amountpipekva < maxpipekva ) ) {
-
- pipe_size = choose_pipespace(wpipe->pipe_buffer.size, wpipe->pipe_buffer.cnt + orig_resid);
+ if (wpipe->pipe_buffer.buffer == 0 || (
+ (unsigned)orig_resid > wpipe->pipe_buffer.size - wpipe->pipe_buffer.cnt &&
+ amountpipekva < maxpipekva)) {
+ pipe_size = choose_pipespace(wpipe->pipe_buffer.size, wpipe->pipe_buffer.cnt + orig_resid);
}
if (pipe_size) {
- /*
+ /*
* need to do initial allocation or resizing of pipe
- * holding both structure and io locks.
+ * holding both structure and io locks.
*/
if ((error = pipeio_lock(wpipe, 1)) == 0) {
- if (wpipe->pipe_buffer.cnt == 0)
+ if (wpipe->pipe_buffer.cnt == 0) {
error = pipespace(wpipe, pipe_size);
- else
+ } else {
error = expand_pipespace(wpipe, pipe_size);
-
+ }
+
pipeio_unlock(wpipe);
-
+
/* allocation failed */
- if (wpipe->pipe_buffer.buffer == 0)
- error = ENOMEM;
+ if (wpipe->pipe_buffer.buffer == 0) {
+ error = ENOMEM;
+ }
}
if (error) {
- /*
+ /*
* If an error occurred unbusy and return, waking up any pending
* readers.
*/
- --wpipe->pipe_busy;
- if ((wpipe->pipe_busy == 0) &&
+ --wpipe->pipe_busy;
+ if ((wpipe->pipe_busy == 0) &&
(wpipe->pipe_state & PIPE_WANT)) {
- wpipe->pipe_state &= ~(PIPE_WANT | PIPE_WANTR);
+ wpipe->pipe_state &= ~(PIPE_WANT | PIPE_WANTR);
wakeup(wpipe);
}
PIPE_UNLOCK(rpipe);
- return(error);
+ return error;
}
}
while (uio_resid(uio)) {
-
- retrywrite:
+retrywrite:
space = wpipe->pipe_buffer.size - wpipe->pipe_buffer.cnt;
/* Writes of size <= PIPE_BUF must be atomic. */
- if ((space < uio_resid(uio)) && (orig_resid <= PIPE_BUF))
+ if ((space < uio_resid(uio)) && (orig_resid <= PIPE_BUF)) {
space = 0;
+ }
if (space > 0) {
+ if ((error = pipeio_lock(wpipe, 1)) == 0) {
+ size_t size; /* Transfer size */
+ size_t segsize; /* first segment to transfer */
- if ((error = pipeio_lock(wpipe,1)) == 0) {
- int size; /* Transfer size */
- int segsize; /* first segment to transfer */
-
- if (wpipe->pipe_state & (PIPE_DRAIN | PIPE_EOF)) {
+ if ((wpipe->pipe_state & (PIPE_DRAIN | PIPE_EOF)) ||
+ (fileproc_get_vflags(fp) & FPV_DRAIN)) {
pipeio_unlock(wpipe);
- error = EPIPE;
+ error = EPIPE;
break;
}
- /*
+ /*
* If a process blocked in pipeio_lock, our
* value for space might be bad... the mutex
* is dropped while we're blocked
*/
- if (space > (int)(wpipe->pipe_buffer.size -
+ if (space > (int)(wpipe->pipe_buffer.size -
wpipe->pipe_buffer.cnt)) {
pipeio_unlock(wpipe);
goto retrywrite;
* and free space in pipe buffer.
*/
// LP64todo - fix this!
- if (space > uio_resid(uio))
- size = uio_resid(uio);
- else
+ if (space > uio_resid(uio)) {
+ size = (size_t)uio_resid(uio);
+ if (size > LONG_MAX) {
+ panic("size greater than LONG_MAX");
+ }
+ } else {
size = space;
+ }
/*
- * First segment to transfer is minimum of
+ * First segment to transfer is minimum of
* transfer size and contiguous space in
* pipe buffer. If first segment to transfer
* is less than the transfer size, we've got
* a wraparound in the buffer.
*/
- segsize = wpipe->pipe_buffer.size -
- wpipe->pipe_buffer.in;
- if (segsize > size)
+ segsize = wpipe->pipe_buffer.size -
+ wpipe->pipe_buffer.in;
+ if (segsize > size) {
segsize = size;
-
+ }
+
/* Transfer first segment */
PIPE_UNLOCK(rpipe);
- error = uiomove(&wpipe->pipe_buffer.buffer[wpipe->pipe_buffer.in],
- segsize, uio);
+ error = uiomove(&wpipe->pipe_buffer.buffer[wpipe->pipe_buffer.in],
+ (int)segsize, uio);
PIPE_LOCK(rpipe);
-
+
if (error == 0 && segsize < size) {
- /*
+ /*
* Transfer remaining part now, to
* support atomic writes. Wraparound
* happened. (State 3)
*/
- if (wpipe->pipe_buffer.in + segsize !=
- wpipe->pipe_buffer.size)
+ if (wpipe->pipe_buffer.in + segsize !=
+ wpipe->pipe_buffer.size) {
panic("Expected pipe buffer "
"wraparound disappeared");
-
+ }
+
PIPE_UNLOCK(rpipe);
error = uiomove(
- &wpipe->pipe_buffer.buffer[0],
- size - segsize, uio);
+ &wpipe->pipe_buffer.buffer[0],
+ (int)(size - segsize), uio);
PIPE_LOCK(rpipe);
}
- /*
+ /*
* readers never know to read until count is updated.
*/
if (error == 0) {
wpipe->pipe_buffer.size) {
if (wpipe->pipe_buffer.in !=
size - segsize +
- wpipe->pipe_buffer.size)
+ wpipe->pipe_buffer.size) {
panic("Expected "
"wraparound bad");
- wpipe->pipe_buffer.in = size -
- segsize;
+ }
+ wpipe->pipe_buffer.in = (unsigned int)(size -
+ segsize);
}
-
+
wpipe->pipe_buffer.cnt += size;
if (wpipe->pipe_buffer.cnt >
- wpipe->pipe_buffer.size)
+ wpipe->pipe_buffer.size) {
panic("Pipe buffer overflow");
-
+ }
}
pipeio_unlock(wpipe);
}
- if (error)
+ if (error) {
break;
-
+ }
} else {
/*
* If the "read-side" has been blocked, wake it up now.
wpipe->pipe_state &= ~PIPE_WANTR;
wakeup(wpipe);
}
+
/*
- * don't block on non-blocking I/O
- * we'll do the pipeselwakeup on the way out
+ * If read side wants to go away, we just issue a signal
+ * to ourselves.
*/
- if (fp->f_flag & FNONBLOCK) {
- error = EAGAIN;
+ if ((wpipe->pipe_state & (PIPE_DRAIN | PIPE_EOF)) ||
+ (fileproc_get_vflags(fp) & FPV_DRAIN)) {
+ error = EPIPE;
break;
}
/*
- * If read side wants to go away, we just issue a signal
- * to ourselves.
+ * don't block on non-blocking I/O
+ * we'll do the pipeselwakeup on the way out
*/
- if (wpipe->pipe_state & (PIPE_DRAIN | PIPE_EOF)) {
- error = EPIPE;
+ if (fp->f_flag & FNONBLOCK) {
+ error = EAGAIN;
break;
- }
+ }
/*
* We have no more space and have something to offer,
error = msleep(wpipe, PIPE_MTX(wpipe), PRIBIO | PCATCH, "pipewr", 0);
- if (error != 0)
+ if (error != 0) {
break;
+ }
}
}
--wpipe->pipe_busy;
pipeselwakeup(wpipe, wpipe);
}
-#ifndef CONFIG_EMBEDDED
+#if defined(XNU_TARGET_OS_OSX)
/* Update modification, status change (# of bytes in pipe) times */
pipe_touch(rpipe, PIPE_MTIME | PIPE_CTIME);
pipe_touch(wpipe, PIPE_MTIME | PIPE_CTIME);
#endif
PIPE_UNLOCK(rpipe);
- return (error);
+ return error;
}
/*
/* ARGSUSED 3 */
static int
pipe_ioctl(struct fileproc *fp, u_long cmd, caddr_t data,
- __unused vfs_context_t ctx)
+ __unused vfs_context_t ctx)
{
struct pipe *mpipe = (struct pipe *)fp->f_data;
#if CONFIG_MACF
if (error) {
PIPE_UNLOCK(mpipe);
- return (error);
+ return error;
}
#endif
switch (cmd) {
-
case FIONBIO:
PIPE_UNLOCK(mpipe);
- return (0);
+ return 0;
case FIOASYNC:
if (*(int *)data) {
mpipe->pipe_state &= ~PIPE_ASYNC;
}
PIPE_UNLOCK(mpipe);
- return (0);
+ return 0;
case FIONREAD:
*(int *)data = mpipe->pipe_buffer.cnt;
PIPE_UNLOCK(mpipe);
- return (0);
+ return 0;
case TIOCSPGRP:
mpipe->pipe_pgid = *(int *)data;
PIPE_UNLOCK(mpipe);
- return (0);
+ return 0;
case TIOCGPGRP:
*(int *)data = mpipe->pipe_pgid;
PIPE_UNLOCK(mpipe);
- return (0);
-
+ return 0;
}
PIPE_UNLOCK(mpipe);
- return (ENOTTY);
+ return ENOTTY;
}
struct pipe *wpipe;
int retnum = 0;
- if (rpipe == NULL || rpipe == (struct pipe *)-1)
- return (retnum);
+ if (rpipe == NULL || rpipe == (struct pipe *)-1) {
+ return retnum;
+ }
PIPE_LOCK(rpipe);
wpipe = rpipe->pipe_peer;
-
+
#if CONFIG_MACF
/*
*/
if (mac_pipe_check_select(vfs_context_ucred(ctx), rpipe, which)) {
PIPE_UNLOCK(rpipe);
- return (0);
+ return 0;
}
#endif
- switch (which) {
-
- case FREAD:
+ switch (which) {
+ case FREAD:
if ((rpipe->pipe_state & PIPE_DIRECTW) ||
(rpipe->pipe_buffer.cnt > 0) ||
- (rpipe->pipe_state & (PIPE_DRAIN | PIPE_EOF))) {
-
- retnum = 1;
+ (rpipe->pipe_state & (PIPE_DRAIN | PIPE_EOF)) ||
+ (fileproc_get_vflags(fp) & FPV_DRAIN)) {
+ retnum = 1;
} else {
- rpipe->pipe_state |= PIPE_SEL;
- selrecord(vfs_context_proc(ctx), &rpipe->pipe_sel, wql);
+ rpipe->pipe_state |= PIPE_SEL;
+ selrecord(vfs_context_proc(ctx), &rpipe->pipe_sel, wql);
}
break;
- case FWRITE:
- if (wpipe)
+ case FWRITE:
+ if (wpipe) {
wpipe->pipe_state |= PIPE_WSELECT;
+ }
if (wpipe == NULL || (wpipe->pipe_state & (PIPE_DRAIN | PIPE_EOF)) ||
+ (fileproc_get_vflags(fp) & FPV_DRAIN) ||
(((wpipe->pipe_state & PIPE_DIRECTW) == 0) &&
- (MAX_PIPESIZE(wpipe) - wpipe->pipe_buffer.cnt) >= PIPE_BUF)) {
-
- retnum = 1;
+ (MAX_PIPESIZE(wpipe) - wpipe->pipe_buffer.cnt) >= PIPE_BUF)) {
+ retnum = 1;
} else {
- wpipe->pipe_state |= PIPE_SEL;
+ wpipe->pipe_state |= PIPE_SEL;
selrecord(vfs_context_proc(ctx), &wpipe->pipe_sel, wql);
}
break;
- case 0:
- rpipe->pipe_state |= PIPE_SEL;
+ case 0:
+ rpipe->pipe_state |= PIPE_SEL;
selrecord(vfs_context_proc(ctx), &rpipe->pipe_sel, wql);
break;
- }
+ }
PIPE_UNLOCK(rpipe);
- return (retnum);
+ return retnum;
}
static int
pipe_close(struct fileglob *fg, __unused vfs_context_t ctx)
{
- struct pipe *cpipe;
+ struct pipe *cpipe;
proc_fdlock_spin(vfs_context_proc(ctx));
cpipe = (struct pipe *)fg->fg_data;
fg->fg_data = NULL;
proc_fdunlock(vfs_context_proc(ctx));
- if (cpipe)
- pipeclose(cpipe);
+ if (cpipe) {
+ pipeclose(cpipe);
+ }
- return (0);
+ return 0;
}
static void
if (cpipe->pipe_buffer.buffer != NULL) {
OSAddAtomic(-(cpipe->pipe_buffer.size), &amountpipekva);
OSAddAtomic(-1, &amountpipes);
- kfree((void *)cpipe->pipe_buffer.buffer,
- cpipe->pipe_buffer.size);
+ kheap_free(KHEAP_DATA_BUFFERS, cpipe->pipe_buffer.buffer,
+ cpipe->pipe_buffer.size);
cpipe->pipe_buffer.buffer = NULL;
cpipe->pipe_buffer.size = 0;
}
{
struct pipe *ppipe;
- if (cpipe == NULL)
- return;
- /* partially created pipes won't have a valid mutex. */
- if (PIPE_MTX(cpipe) != NULL)
- PIPE_LOCK(cpipe);
-
+ PIPE_LOCK(cpipe);
/*
* If the other side is blocked, wake it up saying that
cpipe->pipe_state &= ~PIPE_DRAIN;
cpipe->pipe_state |= PIPE_EOF;
pipeselwakeup(cpipe, cpipe);
-
+
while (cpipe->pipe_busy) {
cpipe->pipe_state |= PIPE_WANT;
wakeup(cpipe);
- msleep(cpipe, PIPE_MTX(cpipe), PRIBIO, "pipecl", 0);
+ msleep(cpipe, PIPE_MTX(cpipe), PRIBIO, "pipecl", 0);
}
#if CONFIG_MACF
/*
* Free the shared pipe label only after the two ends are disconnected.
*/
- if (cpipe->pipe_label != NULL && cpipe->pipe_peer == NULL)
+ if (cpipe->pipe_label != NULL && cpipe->pipe_peer == NULL) {
mac_pipe_label_destroy(cpipe);
+ }
#endif
/*
* Disconnect from peer
*/
if ((ppipe = cpipe->pipe_peer) != NULL) {
-
ppipe->pipe_state &= ~(PIPE_DRAIN);
ppipe->pipe_state |= PIPE_EOF;
pipeselwakeup(ppipe, ppipe);
wakeup(ppipe);
- if (cpipe->pipe_state & PIPE_KNOTE)
- KNOTE(&ppipe->pipe_sel.si_note, 1);
-
- postpipeevent(ppipe, EV_RCLOSED);
+ KNOTE(&ppipe->pipe_sel.si_note, 1);
ppipe->pipe_peer = NULL;
}
- evpipefree(cpipe);
/*
* free resources
*/
- if (PIPE_MTX(cpipe) != NULL) {
- if (ppipe != NULL) {
- /*
- * since the mutex is shared and the peer is still
- * alive, we need to release the mutex, not free it
- */
- PIPE_UNLOCK(cpipe);
- } else {
- /*
- * peer is gone, so we're the sole party left with
- * interest in this mutex... unlock and free it
- */
- PIPE_UNLOCK(cpipe);
- lck_mtx_free(PIPE_MTX(cpipe), pipe_mtx_grp);
- }
+
+ PIPE_UNLOCK(cpipe);
+
+ pipepair_destroy_pipe(PIPE_PAIR(cpipe), cpipe);
+}
+
+static int64_t
+filt_pipelowwat(struct knote *kn, struct pipe *rpipe, int64_t def_lowwat)
+{
+ if ((kn->kn_sfflags & NOTE_LOWAT) == 0) {
+ return def_lowwat;
}
- pipe_free_kmem(cpipe);
- if (cpipe->pipe_state & PIPE_WSELECT) {
- pipe_garbage_collect(cpipe);
- } else {
- zfree(pipe_zone, cpipe);
- pipe_garbage_collect(NULL);
+ if (rpipe->pipe_buffer.size && kn->kn_sdata > MAX_PIPESIZE(rpipe)) {
+ return MAX_PIPESIZE(rpipe);
}
-
+ return MAX(kn->kn_sdata, def_lowwat);
}
-/*ARGSUSED*/
static int
-filt_piperead_common(struct knote *kn, struct pipe *rpipe)
+filt_pipe_draincommon(struct knote *kn, struct pipe *rpipe)
{
- struct pipe *wpipe;
- int retval;
-
- /*
- * we're being called back via the KNOTE post
- * we made in pipeselwakeup, and we already hold the mutex...
- */
+ struct pipe *wpipe = rpipe->pipe_peer;
- wpipe = rpipe->pipe_peer;
- kn->kn_data = rpipe->pipe_buffer.cnt;
if ((rpipe->pipe_state & (PIPE_DRAIN | PIPE_EOF)) ||
(wpipe == NULL) || (wpipe->pipe_state & (PIPE_DRAIN | PIPE_EOF))) {
kn->kn_flags |= EV_EOF;
- retval = 1;
+ return 1;
+ }
+
+ return 0;
+}
+
+static int
+filt_pipenotsup(struct knote *kn, long hint)
+{
+#pragma unused(hint)
+ struct pipe *rpipe = kn->kn_hook;
+
+ return filt_pipe_draincommon(kn, rpipe);
+}
+
+static int
+filt_pipenotsuptouch(struct knote *kn, struct kevent_qos_s *kev)
+{
+ struct pipe *rpipe = kn->kn_hook;
+ int res;
+
+ PIPE_LOCK(rpipe);
+
+ /* accept new kevent data (and save off lowat threshold and flag) */
+ kn->kn_sfflags = kev->fflags;
+ kn->kn_sdata = kev->data;
+
+ /* determine if any event is now deemed fired */
+ res = filt_pipe_draincommon(kn, rpipe);
+
+ PIPE_UNLOCK(rpipe);
+
+ return res;
+}
+
+static int
+filt_pipenotsupprocess(struct knote *kn, struct kevent_qos_s *kev)
+{
+ struct pipe *rpipe = kn->kn_hook;
+ int res;
+
+ PIPE_LOCK(rpipe);
+ res = filt_pipe_draincommon(kn, rpipe);
+ if (res) {
+ knote_fill_kevent(kn, kev, 0);
+ }
+ PIPE_UNLOCK(rpipe);
+
+ return res;
+}
+
+/*ARGSUSED*/
+static int
+filt_piperead_common(struct knote *kn, struct kevent_qos_s *kev, struct pipe *rpipe)
+{
+ int64_t data = rpipe->pipe_buffer.cnt;
+ int res = 0;
+
+ if (filt_pipe_draincommon(kn, rpipe)) {
+ res = 1;
} else {
- int64_t lowwat = 1;
- if (kn->kn_sfflags & NOTE_LOWAT) {
- if (rpipe->pipe_buffer.size && kn->kn_sdata > MAX_PIPESIZE(rpipe))
- lowwat = MAX_PIPESIZE(rpipe);
- else if (kn->kn_sdata > lowwat)
- lowwat = kn->kn_sdata;
- }
- retval = kn->kn_data >= lowwat;
+ res = data >= filt_pipelowwat(kn, rpipe, 1);
+ }
+ if (res && kev) {
+ knote_fill_kevent(kn, kev, data);
}
- return (retval);
+ return res;
}
static int
filt_piperead(struct knote *kn, long hint)
{
#pragma unused(hint)
- struct pipe *rpipe = (struct pipe *)kn->kn_fp->f_data;
+ struct pipe *rpipe = kn->kn_hook;
- return filt_piperead_common(kn, rpipe);
+ return filt_piperead_common(kn, NULL, rpipe);
}
static int
-filt_pipereadtouch(struct knote *kn, struct kevent_internal_s *kev)
+filt_pipereadtouch(struct knote *kn, struct kevent_qos_s *kev)
{
- struct pipe *rpipe = (struct pipe *)kn->kn_fp->f_data;
+ struct pipe *rpipe = kn->kn_hook;
int retval;
PIPE_LOCK(rpipe);
kn->kn_sfflags = kev->fflags;
/* identify if any events are now fired */
- retval = filt_piperead_common(kn, rpipe);
+ retval = filt_piperead_common(kn, NULL, rpipe);
PIPE_UNLOCK(rpipe);
}
static int
-filt_pipereadprocess(struct knote *kn, struct filt_process_s *data, struct kevent_internal_s *kev)
+filt_pipereadprocess(struct knote *kn, struct kevent_qos_s *kev)
{
-#pragma unused(data)
- struct pipe *rpipe = (struct pipe *)kn->kn_fp->f_data;
+ struct pipe *rpipe = kn->kn_hook;
int retval;
PIPE_LOCK(rpipe);
- retval = filt_piperead_common(kn, rpipe);
- if (retval) {
- *kev = kn->kn_kevent;
- if (kn->kn_flags & EV_CLEAR) {
- kn->kn_fflags = 0;
- kn->kn_data = 0;
- }
- }
+ retval = filt_piperead_common(kn, kev, rpipe);
PIPE_UNLOCK(rpipe);
- return (retval);
+ return retval;
}
/*ARGSUSED*/
static int
-filt_pipewrite_common(struct knote *kn, struct pipe *rpipe)
+filt_pipewrite_common(struct knote *kn, struct kevent_qos_s *kev, struct pipe *rpipe)
{
- struct pipe *wpipe;
+ int64_t data = 0;
+ int res = 0;
- /*
- * we're being called back via the KNOTE post
- * we made in pipeselwakeup, and we already hold the mutex...
- */
- wpipe = rpipe->pipe_peer;
-
- if ((wpipe == NULL) || (wpipe->pipe_state & (PIPE_DRAIN | PIPE_EOF))) {
- kn->kn_data = 0;
- kn->kn_flags |= EV_EOF;
- return (1);
+ if (filt_pipe_draincommon(kn, rpipe)) {
+ res = 1;
+ } else {
+ data = MAX_PIPESIZE(rpipe) - rpipe->pipe_buffer.cnt;
+ res = data >= filt_pipelowwat(kn, rpipe, PIPE_BUF);
}
- kn->kn_data = MAX_PIPESIZE(wpipe) - wpipe->pipe_buffer.cnt;
-
- int64_t lowwat = PIPE_BUF;
- if (kn->kn_sfflags & NOTE_LOWAT) {
- if (wpipe->pipe_buffer.size && kn->kn_sdata > MAX_PIPESIZE(wpipe))
- lowwat = MAX_PIPESIZE(wpipe);
- else if (kn->kn_sdata > lowwat)
- lowwat = kn->kn_sdata;
+ if (res && kev) {
+ knote_fill_kevent(kn, kev, data);
}
-
- return (kn->kn_data >= lowwat);
+ return res;
}
/*ARGSUSED*/
filt_pipewrite(struct knote *kn, long hint)
{
#pragma unused(hint)
- struct pipe *rpipe = (struct pipe *)kn->kn_fp->f_data;
+ struct pipe *rpipe = kn->kn_hook;
- return filt_pipewrite_common(kn, rpipe);
+ return filt_pipewrite_common(kn, NULL, rpipe);
}
static int
-filt_pipewritetouch(struct knote *kn, struct kevent_internal_s *kev)
+filt_pipewritetouch(struct knote *kn, struct kevent_qos_s *kev)
{
- struct pipe *rpipe = (struct pipe *)kn->kn_fp->f_data;
+ struct pipe *rpipe = kn->kn_hook;
int res;
PIPE_LOCK(rpipe);
kn->kn_sdata = kev->data;
/* determine if any event is now deemed fired */
- res = filt_pipewrite_common(kn, rpipe);
+ res = filt_pipewrite_common(kn, NULL, rpipe);
PIPE_UNLOCK(rpipe);
}
static int
-filt_pipewriteprocess(struct knote *kn, struct filt_process_s *data, struct kevent_internal_s *kev)
+filt_pipewriteprocess(struct knote *kn, struct kevent_qos_s *kev)
{
-#pragma unused(data)
- struct pipe *rpipe = (struct pipe *)kn->kn_fp->f_data;
+ struct pipe *rpipe = kn->kn_hook;
int res;
PIPE_LOCK(rpipe);
- res = filt_pipewrite_common(kn, rpipe);
- if (res) {
- *kev = kn->kn_kevent;
- if (kn->kn_flags & EV_CLEAR) {
- kn->kn_fflags = 0;
- kn->kn_data = 0;
- }
- }
+ res = filt_pipewrite_common(kn, kev, rpipe);
PIPE_UNLOCK(rpipe);
return res;
/*ARGSUSED*/
static int
-pipe_kqfilter(__unused struct fileproc *fp, struct knote *kn,
- __unused struct kevent_internal_s *kev, __unused vfs_context_t ctx)
+pipe_kqfilter(struct fileproc *fp, struct knote *kn,
+ __unused struct kevent_qos_s *kev)
{
- struct pipe *cpipe = (struct pipe *)kn->kn_fp->f_data;
+ struct pipe *cpipe = (struct pipe *)fp->f_data;
+ struct pipe *rpipe = &PIPE_PAIR(cpipe)->pp_rpipe;
int res;
PIPE_LOCK(cpipe);
* XXX process credential should have a persistent reference on it
* XXX before being passed in here.
*/
- if (mac_pipe_check_kqfilter(vfs_context_ucred(ctx), kn, cpipe) != 0) {
+ kauth_cred_t cred = vfs_context_ucred(vfs_context_current());
+ if (mac_pipe_check_kqfilter(cred, kn, cpipe) != 0) {
PIPE_UNLOCK(cpipe);
- kn->kn_flags = EV_ERROR;
- kn->kn_data = EPERM;
+ knote_set_error(kn, EPERM);
return 0;
}
#endif
+ /*
+ * FreeBSD will fail the attach with EPIPE if the peer pipe is detached,
+ * however, this isn't a programming error as the other side closing
+ * could race with the kevent registration.
+ *
+ * Attach should only fail for programming mistakes else it will break
+ * libdispatch.
+ *
+ * Like FreeBSD, have a "Neutered" filter that will not fire until
+ * the pipe dies if the wrong filter is attached to the wrong end.
+ *
+ * Knotes are always attached to the "rpipe".
+ */
switch (kn->kn_filter) {
case EVFILT_READ:
- kn->kn_filtid = EVFILTID_PIPE_R;
-
- /* determine initial state */
- res = filt_piperead_common(kn, cpipe);
+ if (fp->f_flag & FREAD) {
+ kn->kn_filtid = EVFILTID_PIPE_R;
+ res = filt_piperead_common(kn, NULL, rpipe);
+ } else {
+ kn->kn_filtid = EVFILTID_PIPE_N;
+ res = filt_pipe_draincommon(kn, rpipe);
+ }
break;
case EVFILT_WRITE:
- kn->kn_filtid = EVFILTID_PIPE_W;
-
- if (cpipe->pipe_peer == NULL) {
- /*
- * other end of pipe has been closed
- */
- PIPE_UNLOCK(cpipe);
- kn->kn_flags = EV_ERROR;
- kn->kn_data = EPIPE;
- return 0;
+ if (fp->f_flag & FWRITE) {
+ kn->kn_filtid = EVFILTID_PIPE_W;
+ res = filt_pipewrite_common(kn, NULL, rpipe);
+ } else {
+ kn->kn_filtid = EVFILTID_PIPE_N;
+ res = filt_pipe_draincommon(kn, rpipe);
}
- if (cpipe->pipe_peer)
- cpipe = cpipe->pipe_peer;
-
- /* determine inital state */
- res = filt_pipewrite_common(kn, cpipe);
break;
+
default:
- PIPE_UNLOCK(cpipe);
- kn->kn_flags = EV_ERROR;
- kn->kn_data = EINVAL;
+ PIPE_UNLOCK(cpipe);
+ knote_set_error(kn, EINVAL);
return 0;
}
- if (KNOTE_ATTACH(&cpipe->pipe_sel.si_note, kn))
- cpipe->pipe_state |= PIPE_KNOTE;
+ kn->kn_hook = rpipe;
+ KNOTE_ATTACH(&rpipe->pipe_sel.si_note, kn);
PIPE_UNLOCK(cpipe);
return res;
filt_pipedetach(struct knote *kn)
{
struct pipe *cpipe = (struct pipe *)kn->kn_fp->f_data;
+ struct pipe *rpipe = &PIPE_PAIR(cpipe)->pp_rpipe;
PIPE_LOCK(cpipe);
-
- if (kn->kn_filter == EVFILT_WRITE) {
- if (cpipe->pipe_peer == NULL) {
- PIPE_UNLOCK(cpipe);
- return;
- }
- cpipe = cpipe->pipe_peer;
- }
- if (cpipe->pipe_state & PIPE_KNOTE) {
- if (KNOTE_DETACH(&cpipe->pipe_sel.si_note, kn))
- cpipe->pipe_state &= ~PIPE_KNOTE;
- }
+ KNOTE_DETACH(&rpipe->pipe_sel.si_note, kn);
PIPE_UNLOCK(cpipe);
}
fill_pipeinfo(struct pipe * cpipe, struct pipe_info * pinfo)
{
#if CONFIG_MACF
- int error;
+ int error;
#endif
struct timespec now;
struct vinfo_stat * ub;
int pipe_size = 0;
int pipe_count;
- if (cpipe == NULL)
- return (EBADF);
+ if (cpipe == NULL) {
+ return EBADF;
+ }
PIPE_LOCK(cpipe);
#if CONFIG_MACF
error = mac_pipe_check_stat(kauth_cred_get(), cpipe);
if (error) {
PIPE_UNLOCK(cpipe);
- return (error);
+ return error;
}
#endif
if (cpipe->pipe_buffer.buffer == 0) {
- /*
+ /*
* must be stat'ing the write fd
*/
- if (cpipe->pipe_peer) {
- /*
+ if (cpipe->pipe_peer) {
+ /*
* the peer still exists, use it's info
*/
- pipe_size = MAX_PIPESIZE(cpipe->pipe_peer);
+ pipe_size = MAX_PIPESIZE(cpipe->pipe_peer);
pipe_count = cpipe->pipe_peer->pipe_buffer.cnt;
} else {
pipe_count = 0;
}
} else {
- pipe_size = MAX_PIPESIZE(cpipe);
+ pipe_size = MAX_PIPESIZE(cpipe);
pipe_count = cpipe->pipe_buffer.cnt;
}
/*
* since peer's buffer is setup ouside of lock
* we might catch it in transient state
*/
- if (pipe_size == 0)
+ if (pipe_size == 0) {
pipe_size = PIPE_SIZE;
+ }
ub = &pinfo->pipe_stat;
ub->vst_mode = S_IFIFO | S_IRUSR | S_IWUSR | S_IRGRP | S_IWGRP;
ub->vst_blksize = pipe_size;
ub->vst_size = pipe_count;
- if (ub->vst_blksize != 0)
+ if (ub->vst_blksize != 0) {
ub->vst_blocks = (ub->vst_size + ub->vst_blksize - 1) / ub->vst_blksize;
+ }
ub->vst_nlink = 1;
ub->vst_uid = kauth_getuid();
* XXX (st_dev, st_ino) should be unique.
*/
- pinfo->pipe_handle = (uint64_t)VM_KERNEL_ADDRPERM((uintptr_t)cpipe);
- pinfo->pipe_peerhandle = (uint64_t)VM_KERNEL_ADDRPERM((uintptr_t)(cpipe->pipe_peer));
+ pinfo->pipe_handle = (uint64_t)VM_KERNEL_ADDRHASH((uintptr_t)cpipe);
+ pinfo->pipe_peerhandle = (uint64_t)VM_KERNEL_ADDRHASH((uintptr_t)(cpipe->pipe_peer));
pinfo->pipe_status = cpipe->pipe_state;
PIPE_UNLOCK(cpipe);
- return (0);
+ return 0;
}
-static int
+static int
pipe_drain(struct fileproc *fp, __unused vfs_context_t ctx)
{
-
/* Note: fdlock already held */
- struct pipe *ppipe, *cpipe = (struct pipe *)(fp->f_fglob->fg_data);
+ struct pipe *ppipe, *cpipe = (struct pipe *)(fp->fp_glob->fg_data);
+ boolean_t drain_pipe = FALSE;
+
+ /* Check if the pipe is going away */
+ lck_mtx_lock_spin(&fp->fp_glob->fg_lock);
+ if (os_ref_get_count_raw(&fp->fp_glob->fg_count) == 1) {
+ drain_pipe = TRUE;
+ }
+ lck_mtx_unlock(&fp->fp_glob->fg_lock);
if (cpipe) {
PIPE_LOCK(cpipe);
- cpipe->pipe_state |= PIPE_DRAIN;
- cpipe->pipe_state &= ~(PIPE_WANTR | PIPE_WANTW);
+
+ if (drain_pipe) {
+ cpipe->pipe_state |= PIPE_DRAIN;
+ cpipe->pipe_state &= ~(PIPE_WANTR | PIPE_WANTW);
+ }
wakeup(cpipe);
-
+
/* Must wake up peer: a writer sleeps on the read side */
if ((ppipe = cpipe->pipe_peer)) {
- ppipe->pipe_state |= PIPE_DRAIN;
- ppipe->pipe_state &= ~(PIPE_WANTR | PIPE_WANTW);
+ if (drain_pipe) {
+ ppipe->pipe_state |= PIPE_DRAIN;
+ ppipe->pipe_state &= ~(PIPE_WANTR | PIPE_WANTW);
+ }
wakeup(ppipe);
}
-
+
PIPE_UNLOCK(cpipe);
return 0;
}
return 1;
}
-
-
- /*
- * When a thread sets a write-select on a pipe, it creates an implicit,
- * untracked dependency between that thread and the peer of the pipe
- * on which the select is set. If the peer pipe is closed and freed
- * before the select()ing thread wakes up, the system will panic as
- * it attempts to unwind the dangling select(). To avoid that panic,
- * we notice whenever a dangerous select() is set on a pipe, and
- * defer the final deletion of the pipe until that select()s are all
- * resolved. Since we can't currently detect exactly when that
- * resolution happens, we use a simple garbage collection queue to
- * reap the at-risk pipes 'later'.
- */
-static void
-pipe_garbage_collect(struct pipe *cpipe)
-{
- uint64_t old, now;
- struct pipe_garbage *pgp;
-
- /* Convert msecs to nsecs and then to abstime */
- old = pipe_garbage_age_limit * 1000000;
- nanoseconds_to_absolutetime(old, &old);
-
- lck_mtx_lock(pipe_garbage_lock);
-
- /* Free anything that's been on the queue for <mumble> seconds */
- now = mach_absolute_time();
- old = now - old;
- while ((pgp = pipe_garbage_head) && pgp->pg_timestamp < old) {
- pipe_garbage_head = pgp->pg_next;
- if (pipe_garbage_head == NULL)
- pipe_garbage_tail = NULL;
- pipe_garbage_count--;
- zfree(pipe_zone, pgp->pg_pipe);
- zfree(pipe_garbage_zone, pgp);
- }
-
- /* Add the new pipe (if any) to the tail of the garbage queue */
- if (cpipe) {
- cpipe->pipe_state = PIPE_DEAD;
- pgp = (struct pipe_garbage *)zalloc(pipe_garbage_zone);
- if (pgp == NULL) {
- /*
- * We're too low on memory to garbage collect the
- * pipe. Freeing it runs the risk of panicing the
- * system. All we can do is leak it and leave
- * a breadcrumb behind. The good news, such as it
- * is, is that this will probably never happen.
- * We will probably hit the panic below first.
- */
- printf("Leaking pipe %p - no room left in the queue",
- cpipe);
- lck_mtx_unlock(pipe_garbage_lock);
- return;
- }
-
- pgp->pg_pipe = cpipe;
- pgp->pg_timestamp = now;
- pgp->pg_next = NULL;
-
- if (pipe_garbage_tail)
- pipe_garbage_tail->pg_next = pgp;
- pipe_garbage_tail = pgp;
- if (pipe_garbage_head == NULL)
- pipe_garbage_head = pipe_garbage_tail;
-
- if (pipe_garbage_count++ >= PIPE_GARBAGE_QUEUE_LIMIT)
- panic("Length of pipe garbage queue exceeded %d",
- PIPE_GARBAGE_QUEUE_LIMIT);
- }
- lck_mtx_unlock(pipe_garbage_lock);
-}
-