#include <netinet/mptcp_var.h>
#endif /* MULTIPATH */
+#define ROUNDUP(a, b) (((a) + ((b) - 1)) & (~((b) - 1)))
+
+#if DEBUG || DEVELOPMENT
+#define DEBUG_KERNEL_ADDRPERM(_v) (_v)
+#else
+#define DEBUG_KERNEL_ADDRPERM(_v) VM_KERNEL_ADDRPERM(_v)
+#endif
+
/* TODO: this should be in a header file somewhere */
extern char *proc_name_address(void *p);
static int filt_sowrite(struct knote *kn, long hint);
static void filt_sockdetach(struct knote *kn);
static int filt_sockev(struct knote *kn, long hint);
+static void filt_socktouch(struct knote *kn, struct kevent_internal_s *kev,
+ long type);
static int sooptcopyin_timeval(struct sockopt *, struct timeval *);
static int sooptcopyout_timeval(struct sockopt *, const struct timeval *);
.f_isfd = 1,
.f_detach = filt_sockdetach,
.f_event = filt_sockev,
+ .f_touch = filt_socktouch,
};
SYSCTL_DECL(_kern_ipc);
SYSCTL_INT(_kern_ipc, OID_AUTO, sosendjcl_ignore_capab,
CTLFLAG_RW | CTLFLAG_LOCKED, &sosendjcl_ignore_capab, 0, "");
+/*
+ * Set this to ignore SOF1_IF_2KCL and use big clusters for large
+ * writes on the socket for all protocols on any network interfaces.
+ * Be extra careful when setting this to 1, because sending down packets with
+ * clusters larger that 2 KB might lead to system panics or data corruption.
+ * When set to 0, the system will respect SOF1_IF_2KCL, which is set
+ * on the outgoing interface
+ * Set this to 1 for testing/debugging purposes only.
+ */
int sosendbigcl_ignore_capab = 0;
SYSCTL_INT(_kern_ipc, OID_AUTO, sosendbigcl_ignore_capab,
CTLFLAG_RW | CTLFLAG_LOCKED, &sosendbigcl_ignore_capab, 0, "");
SYSCTL_INT(_kern_ipc, OID_AUTO, sorestrictsend, CTLFLAG_RW | CTLFLAG_LOCKED,
&sorestrictsend, 0, "Enable outbound interface restrictions");
+int soreserveheadroom = 1;
+SYSCTL_INT(_kern_ipc, OID_AUTO, soreserveheadroom, CTLFLAG_RW | CTLFLAG_LOCKED,
+ &soreserveheadroom, 0, "To allocate contiguous datagram buffers");
+
extern struct inpcbinfo tcbinfo;
/* TODO: these should be in header file */
vm_size_t so_cache_zone_element_size;
-static int sodelayed_copy(struct socket *, struct uio *, struct mbuf **, user_ssize_t *);
+static int sodelayed_copy(struct socket *, struct uio *, struct mbuf **,
+ user_ssize_t *);
static void cached_sock_alloc(struct socket **, int);
static void cached_sock_free(struct socket *);
+/*
+ * Maximum of extended background idle sockets per process
+ * Set to zero to disable further setting of the option
+ */
+
+#define SO_IDLE_BK_IDLE_MAX_PER_PROC 1
+#define SO_IDLE_BK_IDLE_TIME 600
+#define SO_IDLE_BK_IDLE_RCV_HIWAT 131072
+
+struct soextbkidlestat soextbkidlestat;
+
+SYSCTL_UINT(_kern_ipc, OID_AUTO, maxextbkidleperproc,
+ CTLFLAG_RW | CTLFLAG_LOCKED, &soextbkidlestat.so_xbkidle_maxperproc, 0,
+ "Maximum of extended background idle sockets per process");
+
+SYSCTL_UINT(_kern_ipc, OID_AUTO, extbkidletime, CTLFLAG_RW | CTLFLAG_LOCKED,
+ &soextbkidlestat.so_xbkidle_time, 0,
+ "Time in seconds to keep extended background idle sockets");
+
+SYSCTL_UINT(_kern_ipc, OID_AUTO, extbkidlercvhiwat, CTLFLAG_RW | CTLFLAG_LOCKED,
+ &soextbkidlestat.so_xbkidle_rcvhiwat, 0,
+ "High water mark for extended background idle sockets");
+
+SYSCTL_STRUCT(_kern_ipc, OID_AUTO, extbkidlestat, CTLFLAG_RD | CTLFLAG_LOCKED,
+ &soextbkidlestat, soextbkidlestat, "");
+
+int so_set_extended_bk_idle(struct socket *, int);
+
/*
* SOTCDB_NO_DSCP is set by default, to prevent the networking stack from
* setting the DSCP code on the packet based on the service class; see
_CASSERT(sizeof(so_gencnt) == sizeof(uint64_t));
VERIFY(IS_P2ALIGNED(&so_gencnt, sizeof(uint32_t)));
+#ifdef __LP64__
+ _CASSERT(sizeof(struct sa_endpoints) == sizeof(struct user64_sa_endpoints));
+ _CASSERT(offsetof(struct sa_endpoints, sae_srcif) == offsetof(struct user64_sa_endpoints, sae_srcif));
+ _CASSERT(offsetof(struct sa_endpoints, sae_srcaddr) == offsetof(struct user64_sa_endpoints, sae_srcaddr));
+ _CASSERT(offsetof(struct sa_endpoints, sae_srcaddrlen) == offsetof(struct user64_sa_endpoints, sae_srcaddrlen));
+ _CASSERT(offsetof(struct sa_endpoints, sae_dstaddr) == offsetof(struct user64_sa_endpoints, sae_dstaddr));
+ _CASSERT(offsetof(struct sa_endpoints, sae_dstaddrlen) == offsetof(struct user64_sa_endpoints, sae_dstaddrlen));
+#else
+ _CASSERT(sizeof(struct sa_endpoints) == sizeof(struct user32_sa_endpoints));
+ _CASSERT(offsetof(struct sa_endpoints, sae_srcif) == offsetof(struct user32_sa_endpoints, sae_srcif));
+ _CASSERT(offsetof(struct sa_endpoints, sae_srcaddr) == offsetof(struct user32_sa_endpoints, sae_srcaddr));
+ _CASSERT(offsetof(struct sa_endpoints, sae_srcaddrlen) == offsetof(struct user32_sa_endpoints, sae_srcaddrlen));
+ _CASSERT(offsetof(struct sa_endpoints, sae_dstaddr) == offsetof(struct user32_sa_endpoints, sae_dstaddr));
+ _CASSERT(offsetof(struct sa_endpoints, sae_dstaddrlen) == offsetof(struct user32_sa_endpoints, sae_dstaddrlen));
+#endif
+
if (socketinit_done) {
printf("socketinit: already called...\n");
return;
so_cache_zone_element_size = (vm_size_t)(sizeof (struct socket) + 4
+ get_inpcb_str_size() + 4 + get_tcp_str_size());
- so_cache_zone = zinit(so_cache_zone_element_size,
+ so_cache_zone = zinit(so_cache_zone_element_size,
(120000 * so_cache_zone_element_size), 8192, "socache zone");
zone_change(so_cache_zone, Z_CALLERACCT, FALSE);
zone_change(so_cache_zone, Z_NOENCRYPT, TRUE);
zone_change(se_zone, Z_CALLERACCT, FALSE);
zone_change(se_zone, Z_EXPAND, TRUE);
+ bzero(&soextbkidlestat, sizeof(struct soextbkidlestat));
+ soextbkidlestat.so_xbkidle_maxperproc = SO_IDLE_BK_IDLE_MAX_PER_PROC;
+ soextbkidlestat.so_xbkidle_time = SO_IDLE_BK_IDLE_TIME;
+ soextbkidlestat.so_xbkidle_rcvhiwat = SO_IDLE_BK_IDLE_RCV_HIWAT;
in_pcbinit();
sflt_init();
bzero((caddr_t)*so, sizeof (struct socket));
/*
- * Define offsets for extra structures into our
- * single block of memory. Align extra structures
+ * Define offsets for extra structures into our
+ * single block of memory. Align extra structures
* on longword boundaries.
*/
(caddr_t)offset;
}
- (*so)->cached_in_sock_layer = true;
+ OSBitOrAtomic(SOF1_CACHED_IN_SOCK_LAYER, &(*so)->so_flags1);
}
static void
#if NECP
static void
-so_update_necp_policy(struct socket *so, struct sockaddr *override_local_addr, struct sockaddr *override_remote_addr)
+so_update_necp_policy(struct socket *so, struct sockaddr *override_local_addr,
+ struct sockaddr *override_remote_addr)
{
if (SOCK_DOM(so) == PF_INET || SOCK_DOM(so) == PF_INET6)
- inp_update_necp_policy(sotoinpcb(so), override_local_addr, override_remote_addr, 0);
+ inp_update_necp_policy(sotoinpcb(so), override_local_addr,
+ override_remote_addr, 0);
}
#endif /* NECP */
while (!STAILQ_EMPTY(&so_cache_head)) {
VERIFY(cached_sock_count > 0);
p = STAILQ_FIRST(&so_cache_head);
- if ((so_cache_time - p->cache_timestamp) <
+ if ((so_cache_time - p->cache_timestamp) <
SO_CACHE_TIME_LIMIT)
break;
* If this thread or task is marked to create backgrounded sockets,
* mark the socket as background.
*/
- if (proc_get_effective_thread_policy(current_thread(), TASK_POLICY_NEW_SOCKETS_BG)) {
+ if (proc_get_effective_thread_policy(current_thread(),
+ TASK_POLICY_NEW_SOCKETS_BG)) {
socket_set_traffic_mgt_flags(so, TRAFFIC_MGT_SO_BACKGROUND);
so->so_background_thread = current_thread();
}
so_update_last_owner_locked(so, p);
so_update_policy(so);
-
+
#if NECP
so_update_necp_policy(so, nam, NULL);
#endif /* NECP */
-
+
/*
* If this is a bind request on a socket that has been marked
* as inactive, reject it now before we go any further.
if (so->so_flags & SOF_DEFUNCT) {
error = EINVAL;
SODEFUNCTLOG(("%s[%d]: defunct so 0x%llx [%d,%d] (%d)\n",
- __func__, proc_pid(p), (uint64_t)VM_KERNEL_ADDRPERM(so),
+ __func__, proc_pid(p), (uint64_t)DEBUG_KERNEL_ADDRPERM(so),
SOCK_DOM(so), SOCK_TYPE(so), error));
goto out;
}
mac_socket_label_destroy(so);
#endif /* MAC_SOCKET */
- if (so->cached_in_sock_layer) {
+ if (so->so_flags1 & SOF1_CACHED_IN_SOCK_LAYER) {
cached_sock_free(so);
} else {
FREE_ZONE(so, sizeof (*so), so->so_zone);
so_update_last_owner_locked(so, p);
so_update_policy(so);
-
+
#if NECP
so_update_necp_policy(so, NULL, NULL);
#endif /* NECP */
-
+
if (so->so_proto == NULL) {
error = EINVAL;
goto out;
if (so->so_flags & SOF_DEFUNCT) {
SODEFUNCTLOG(("%s[%d]: defunct so 0x%llx [%d,%d] "
"(%d)\n", __func__, proc_pid(p),
- (uint64_t)VM_KERNEL_ADDRPERM(so),
+ (uint64_t)DEBUG_KERNEL_ADDRPERM(so),
SOCK_DOM(so), SOCK_TYPE(so), error));
}
goto out;
}
#endif /* CONTENT_FILTER */
+ if (so->so_flags1 & SOF1_EXTEND_BK_IDLE_INPROG) {
+ soresume(current_proc(), so, 1);
+ so->so_flags1 &= ~SOF1_EXTEND_BK_IDLE_WANTED;
+ }
+
if ((so->so_options & SO_ACCEPTCONN)) {
struct socket *sp, *sonext;
int socklock = 0;
if (so->so_flags & SOF_DEFUNCT) {
SODEFUNCTLOG(("%s[%d]: defunct so 0x%llx [%d,%d] "
"(%d)\n", __func__, proc_pid(p),
- (uint64_t)VM_KERNEL_ADDRPERM(so),
+ (uint64_t)DEBUG_KERNEL_ADDRPERM(so),
SOCK_DOM(so), SOCK_TYPE(so), error));
}
if (dolock)
int
soconnectxlocked(struct socket *so, struct sockaddr_list **src_sl,
struct sockaddr_list **dst_sl, struct proc *p, uint32_t ifscope,
- associd_t aid, connid_t *pcid, uint32_t flags, void *arg,
- uint32_t arglen)
+ sae_associd_t aid, sae_connid_t *pcid, uint32_t flags, void *arg,
+ uint32_t arglen, uio_t auio, user_ssize_t *bytes_written)
{
int error;
so_update_last_owner_locked(so, p);
so_update_policy(so);
-
+
/*
* If this is a listening socket or if this is a previously-accepted
* socket that has been marked as inactive, reject the connect request.
if (so->so_flags & SOF_DEFUNCT) {
SODEFUNCTLOG(("%s[%d]: defunct so 0x%llx [%d,%d] "
"(%d)\n", __func__, proc_pid(p),
- (uint64_t)VM_KERNEL_ADDRPERM(so),
+ (uint64_t)DEBUG_KERNEL_ADDRPERM(so),
SOCK_DOM(so), SOCK_TYPE(so), error));
}
return (error);
} else {
error = (*so->so_proto->pr_usrreqs->pru_connectx)
(so, src_sl, dst_sl, p, ifscope, aid, pcid,
- flags, arg, arglen);
+ flags, arg, arglen, auio, bytes_written);
}
}
}
int
-sodisconnectxlocked(struct socket *so, associd_t aid, connid_t cid)
+sodisconnectxlocked(struct socket *so, sae_associd_t aid, sae_connid_t cid)
{
int error;
}
int
-sodisconnectx(struct socket *so, associd_t aid, connid_t cid)
+sodisconnectx(struct socket *so, sae_associd_t aid, sae_connid_t cid)
{
int error;
}
int
-sopeelofflocked(struct socket *so, associd_t aid, struct socket **psop)
+sopeelofflocked(struct socket *so, sae_associd_t aid, struct socket **psop)
{
return ((*so->so_proto->pr_usrreqs->pru_peeloff)(so, aid, psop));
}
defunct:
error = EPIPE;
SODEFUNCTLOG(("%s[%d]: defunct so 0x%llx [%d,%d] (%d)\n",
- __func__, proc_selfpid(), (uint64_t)VM_KERNEL_ADDRPERM(so),
+ __func__, proc_selfpid(),
+ (uint64_t)DEBUG_KERNEL_ADDRPERM(so),
SOCK_DOM(so), SOCK_TYPE(so), error));
return (error);
}
cfil_sock_data_pending(&so->so_snd) != 0)
CFIL_LOG(LOG_INFO,
"so %llx ignore SS_CANTSENDMORE",
- (uint64_t)VM_KERNEL_ADDRPERM(so));
+ (uint64_t)DEBUG_KERNEL_ADDRPERM(so));
else
#endif /* CONTENT_FILTER */
return (EPIPE);
if ((so->so_state & SS_ISCONNECTED) == 0) {
if ((so->so_proto->pr_flags & PR_CONNREQUIRED) != 0) {
if (((so->so_state & SS_ISCONFIRMING) == 0) &&
- (resid != 0 || clen == 0)) {
+ (resid != 0 || clen == 0) &&
+ !(so->so_flags1 & SOF1_PRECONNECT_DATA)) {
#if MPTCP
- /*
- * MPTCP Fast Join sends data before the
+ /*
+ * MPTCP Fast Join sends data before the
* socket is truly connected.
*/
if ((so->so_flags & (SOF_MP_SUBFLOW |
SOF_MPTCP_FASTJOIN)) !=
(SOF_MP_SUBFLOW | SOF_MPTCP_FASTJOIN))
-#endif /* MPTCP */
+#endif /* MPTCP */
return (ENOTCONN);
}
} else if (addr == 0 && !(flags&MSG_HOLD)) {
ENOTCONN : EDESTADDRREQ);
}
}
+
if (so->so_flags & SOF_ENABLE_MSGS)
space = msgq_sbspace(so, control);
else
return (EMSGSIZE);
if ((space < resid + clen &&
- (atomic || space < (int32_t)so->so_snd.sb_lowat || space < clen)) ||
+ (atomic || (space < (int32_t)so->so_snd.sb_lowat) ||
+ space < clen)) ||
(so->so_type == SOCK_STREAM && so_wait_for_if_feedback(so))) {
+ /*
+ * don't block the connectx call when there's more data
+ * than can be copied.
+ */
+ if (so->so_flags1 & SOF1_PRECONNECT_DATA) {
+ if (space == 0) {
+ return (EWOULDBLOCK);
+ }
+ if (space < (int32_t)so->so_snd.sb_lowat) {
+ return (0);
+ }
+ }
if ((so->so_state & SS_NBIO) || (flags & MSG_NBIO) ||
assumelock) {
return (EWOULDBLOCK);
{
struct mbuf **mp;
struct mbuf *m, *freelist = NULL;
- user_ssize_t space, len, resid;
+ user_ssize_t space, len, resid, orig_resid;
int clen = 0, error, dontroute, mlen, sendflags;
int atomic = sosendallatonce(so) || top;
int sblocked = 0;
struct proc *p = current_proc();
struct mbuf *control_copy = NULL;
+ uint16_t headroom = 0;
+ boolean_t en_tracing = FALSE;
if (uio != NULL)
resid = uio_resid(uio);
socket_lock(so, 1);
+ /*
+ * trace if tracing & network (vs. unix) sockets & and
+ * non-loopback
+ */
+ if (ENTR_SHOULDTRACE &&
+ (SOCK_CHECK_DOM(so, AF_INET) || SOCK_CHECK_DOM(so, AF_INET6))) {
+ struct inpcb *inp = sotoinpcb(so);
+ if (inp->inp_last_outifp != NULL &&
+ !(inp->inp_last_outifp->if_flags & IFF_LOOPBACK)) {
+ en_tracing = TRUE;
+ KERNEL_ENERGYTRACE(kEnTrActKernSockWrite, DBG_FUNC_START,
+ VM_KERNEL_ADDRPERM(so),
+ ((so->so_state & SS_NBIO) ? kEnTrFlagNonBlocking : 0),
+ (int64_t)resid);
+ orig_resid = resid;
+ }
+ }
+
/*
* Re-injection should not affect process accounting
*/
if ((flags & MSG_SKIPCFIL) == 0) {
- so_update_last_owner_locked(so, p);
- so_update_policy(so);
-
+ so_update_last_owner_locked(so, p);
+ so_update_policy(so);
+
#if NECP
- so_update_necp_policy(so, NULL, addr);
+ so_update_necp_policy(so, NULL, addr);
#endif /* NECP */
}
-
+
if (so->so_type != SOCK_STREAM && (flags & MSG_OOB) != 0) {
error = EOPNOTSUPP;
socket_unlock(so, 1);
if (control != NULL)
clen = control->m_len;
+ if (soreserveheadroom != 0)
+ headroom = so->so_pktheadroom;
+
do {
error = sosendcheck(so, addr, resid, clen, atomic, flags,
&sblocked, control);
int bytes_to_copy;
boolean_t jumbocl;
boolean_t bigcl;
+ int bytes_to_alloc;
bytes_to_copy = imin(resid, space);
+ bytes_to_alloc = bytes_to_copy;
+ if (top == NULL)
+ bytes_to_alloc += headroom;
+
if (sosendminchain > 0)
chainlength = 0;
else
chainlength = sosendmaxchain;
/*
- * Use big 4 KB cluster only when outgoing
- * interface does not want 2 LB clusters
+ * Use big 4 KB cluster when the outgoing interface
+ * does not prefer 2 KB clusters
*/
- bigcl =
- !(so->so_flags1 & SOF1_IF_2KCL) ||
+ bigcl = !(so->so_flags1 & SOF1_IF_2KCL) ||
sosendbigcl_ignore_capab;
-
+
/*
* Attempt to use larger than system page-size
* clusters for large writes only if there is
* haven't yet consumed.
*/
if (freelist == NULL &&
- bytes_to_copy > MBIGCLBYTES &&
+ bytes_to_alloc > MBIGCLBYTES &&
jumbocl) {
num_needed =
- bytes_to_copy / M16KCLBYTES;
+ bytes_to_alloc / M16KCLBYTES;
- if ((bytes_to_copy -
+ if ((bytes_to_alloc -
(num_needed * M16KCLBYTES))
>= MINCLSIZE)
num_needed++;
}
if (freelist == NULL &&
- bytes_to_copy > MCLBYTES &&
+ bytes_to_alloc > MCLBYTES &&
bigcl) {
num_needed =
- bytes_to_copy / MBIGCLBYTES;
+ bytes_to_alloc / MBIGCLBYTES;
- if ((bytes_to_copy -
+ if ((bytes_to_alloc -
(num_needed * MBIGCLBYTES)) >=
MINCLSIZE)
num_needed++;
*/
}
- if (freelist == NULL &&
- bytes_to_copy > MINCLSIZE) {
+ /*
+ * Allocate a cluster as we want to
+ * avoid to split the data in more
+ * that one segment and using MINCLSIZE
+ * would lead us to allocate two mbufs
+ */
+ if (soreserveheadroom != 0 &&
+ freelist == NULL &&
+ ((top == NULL &&
+ bytes_to_alloc > _MHLEN) ||
+ bytes_to_alloc > _MLEN)) {
+ num_needed = ROUNDUP(bytes_to_alloc, MCLBYTES) /
+ MCLBYTES;
+ freelist =
+ m_getpackets_internal(
+ (unsigned int *)&num_needed,
+ hdrs_needed, M_WAIT, 0,
+ MCLBYTES);
+ /*
+ * Fall back to a single mbuf
+ * if allocation failed
+ */
+ } else if (freelist == NULL &&
+ bytes_to_alloc > MINCLSIZE) {
num_needed =
- bytes_to_copy / MCLBYTES;
+ bytes_to_alloc / MCLBYTES;
- if ((bytes_to_copy -
+ if ((bytes_to_alloc -
(num_needed * MCLBYTES)) >=
MINCLSIZE)
num_needed++;
* if allocation failed
*/
}
-
+ /*
+ * For datagram protocols, leave
+ * headroom for protocol headers
+ * in the first cluster of the chain
+ */
+ if (freelist != NULL && atomic &&
+ top == NULL && headroom > 0) {
+ freelist->m_data += headroom;
+ }
+
+ /*
+ * Fall back to regular mbufs without
+ * reserving the socket headroom
+ */
if (freelist == NULL) {
if (top == NULL)
MGETHDR(freelist,
m->m_next = NULL;
if ((m->m_flags & M_EXT))
- mlen = m->m_ext.ext_size;
+ mlen = m->m_ext.ext_size -
+ m_leadingspace(m);
else if ((m->m_flags & M_PKTHDR))
mlen =
MHLEN - m_leadingspace(m);
else
- mlen = MLEN;
+ mlen = MLEN - m_leadingspace(m);
len = imin(mlen, bytes_to_copy);
chainlength += len;
if (dontroute)
so->so_options |= SO_DONTROUTE;
- /* Compute flags here, for pru_send and NKEs */
+ /*
+ * Compute flags here, for pru_send and NKEs
+ *
+ * If the user set MSG_EOF, the protocol
+ * understands this flag and nothing left to
+ * send then use PRU_SEND_EOF instead of PRU_SEND.
+ */
sendflags = (flags & MSG_OOB) ? PRUS_OOB :
- /*
- * If the user set MSG_EOF, the protocol
- * understands this flag and nothing left to
- * send then use PRU_SEND_EOF instead of PRU_SEND.
- */
((flags & MSG_EOF) &&
- (so->so_proto->pr_flags & PR_IMPLOPCL) &&
- (resid <= 0)) ? PRUS_EOF :
- /* If there is more to send set PRUS_MORETOCOME */
- (resid > 0 && space > 0) ? PRUS_MORETOCOME : 0;
-
+ (so->so_proto->pr_flags & PR_IMPLOPCL) &&
+ (resid <= 0)) ? PRUS_EOF :
+ /* If there is more to send set PRUS_MORETOCOME */
+ (resid > 0 && space > 0) ? PRUS_MORETOCOME : 0;
+
if ((flags & MSG_SKIPCFIL) == 0) {
/*
* Socket filter processing
* Content filter processing
*/
error = cfil_sock_data_out(so, addr, top,
- control, (sendflags & MSG_OOB) ?
+ control, (sendflags & MSG_OOB) ?
sock_data_filt_flag_oob : 0);
if (error) {
if (error == EJUSTRETURN) {
if (control_copy != NULL)
m_freem(control_copy);
- KERNEL_DEBUG(DBG_FNC_SOSEND | DBG_FUNC_END, so, resid, so->so_snd.sb_cc,
- space, error);
+ /*
+ * One write has been done. This was enough. Get back to "normal"
+ * behavior.
+ */
+ if (so->so_flags1 & SOF1_PRECONNECT_DATA)
+ so->so_flags1 &= ~SOF1_PRECONNECT_DATA;
+
+ if (en_tracing) {
+ /* resid passed here is the bytes left in uio */
+ KERNEL_ENERGYTRACE(kEnTrActKernSockWrite, DBG_FUNC_END,
+ VM_KERNEL_ADDRPERM(so),
+ ((error == EWOULDBLOCK) ? kEnTrFlagNoWork : 0),
+ (int64_t)(orig_resid - resid));
+ }
+ KERNEL_DEBUG(DBG_FNC_SOSEND | DBG_FUNC_END, so, resid,
+ so->so_snd.sb_cc, space, error);
return (error);
}
+/*
+ * Supported only connected sockets (no address) without ancillary data
+ * (control mbuf) for atomic protocols
+ */
int
-sosend_list(struct socket *so, struct sockaddr *addr, struct uio **uioarray,
- u_int uiocnt, struct mbuf *top, struct mbuf *control, int flags)
+sosend_list(struct socket *so, struct uio **uioarray, u_int uiocnt, int flags)
{
struct mbuf *m, *freelist = NULL;
user_ssize_t len, resid;
- int clen = 0, error, dontroute, mlen;
- int atomic = sosendallatonce(so) || top;
+ int error, dontroute, mlen;
+ int atomic = sosendallatonce(so);
int sblocked = 0;
struct proc *p = current_proc();
u_int uiofirst = 0;
u_int uiolast = 0;
+ struct mbuf *top = NULL;
+ uint16_t headroom = 0;
+ boolean_t bigcl;
KERNEL_DEBUG((DBG_FNC_SOSEND_LIST | DBG_FUNC_START), so, uiocnt,
so->so_snd.sb_cc, so->so_snd.sb_lowat, so->so_snd.sb_hiwat);
error = EINVAL;
goto out;
}
- if (uioarray != NULL)
- resid = uio_array_resid(uioarray, uiocnt);
- else
- resid = mbuf_pkt_list_len(top);
+ resid = uio_array_resid(uioarray, uiocnt);
/*
* In theory resid should be unsigned.
error = EINVAL;
goto out;
}
- /*
- * Disallow functionality not currently supported
- * Note: Will need to treat arrays of addresses and controls
- */
- if (addr != NULL) {
- printf("%s addr not supported\n", __func__);
- error = EOPNOTSUPP;
- goto out;
- }
- if (control != NULL) {
- printf("%s control not supported\n", __func__);
- error = EOPNOTSUPP;
- goto out;
- }
socket_lock(so, 1);
so_update_last_owner_locked(so, p);
so_update_policy(so);
-
+
#if NECP
- so_update_necp_policy(so, NULL, addr);
+ so_update_necp_policy(so, NULL, NULL);
#endif /* NECP */
-
+
dontroute = (flags & MSG_DONTROUTE) &&
(so->so_options & SO_DONTROUTE) == 0 &&
(so->so_proto->pr_flags & PR_ATOMIC);
OSIncrementAtomicLong(&p->p_stats->p_ru.ru_msgsnd);
- if (control != NULL)
- clen = control->m_len;
-
- error = sosendcheck(so, addr, resid, clen, atomic, flags,
- &sblocked, control);
+ error = sosendcheck(so, NULL, resid, 0, atomic, flags,
+ &sblocked, NULL);
if (error)
goto release;
+ /*
+ * Use big 4 KB clusters when the outgoing interface does not prefer
+ * 2 KB clusters
+ */
+ bigcl = !(so->so_flags1 & SOF1_IF_2KCL) || sosendbigcl_ignore_capab;
+
+ if (soreserveheadroom != 0)
+ headroom = so->so_pktheadroom;
+
do {
int i;
+ int num_needed = 0;
+ int chainlength;
+ size_t maxpktlen = 0;
+ int bytes_to_alloc;
- if (uioarray == NULL) {
- /*
- * Data is prepackaged in "top".
- */
- resid = 0;
- } else {
- int num_needed = 0;
- int chainlength;
- size_t maxpktlen = 0;
+ if (sosendminchain > 0)
+ chainlength = 0;
+ else
+ chainlength = sosendmaxchain;
- if (sosendminchain > 0)
- chainlength = 0;
- else
- chainlength = sosendmaxchain;
+ socket_unlock(so, 0);
- socket_unlock(so, 0);
+ /*
+ * Find a set of uio that fit in a reasonable number
+ * of mbuf packets
+ */
+ for (i = uiofirst; i < uiocnt; i++) {
+ struct uio *auio = uioarray[i];
- /*
- * Find a set of uio that fit in a reasonable number
- * of mbuf packets
- */
- for (i = uiofirst; i < uiocnt; i++) {
- struct uio *auio = uioarray[i];
+ len = uio_resid(auio);
- len = uio_resid(auio);
+ /* Do nothing for empty messages */
+ if (len == 0)
+ continue;
- /* Do nothing for empty messages */
- if (len == 0)
- continue;
+ num_needed += 1;
+ uiolast += 1;
- num_needed += 1;
- uiolast += 1;
-
- if (len > maxpktlen)
- maxpktlen = len;
+ if (len > maxpktlen)
+ maxpktlen = len;
- chainlength += len;
- if (chainlength > sosendmaxchain)
- break;
- }
- /*
- * Nothing left to send
- */
- if (num_needed == 0) {
- socket_lock(so, 0);
+ chainlength += len;
+ if (chainlength > sosendmaxchain)
break;
- }
- /*
- * Allocate the mbuf packets at once
- */
+ }
+ /*
+ * Nothing left to send
+ */
+ if (num_needed == 0) {
+ socket_lock(so, 0);
+ break;
+ }
+ /*
+ * Allocate buffer large enough to include headroom space for
+ * network and link header
+ *
+ */
+ bytes_to_alloc = maxpktlen + headroom;
+
+ /*
+ * Allocate a single contiguous buffer of the smallest available
+ * size when possible
+ */
+ if (bytes_to_alloc > MCLBYTES &&
+ bytes_to_alloc <= MBIGCLBYTES && bigcl) {
+ freelist = m_getpackets_internal(
+ (unsigned int *)&num_needed,
+ num_needed, M_WAIT, 1,
+ MBIGCLBYTES);
+ } else if (bytes_to_alloc > _MHLEN &&
+ bytes_to_alloc <= MCLBYTES) {
+ freelist = m_getpackets_internal(
+ (unsigned int *)&num_needed,
+ num_needed, M_WAIT, 1,
+ MCLBYTES);
+ } else {
freelist = m_allocpacket_internal(
(unsigned int *)&num_needed,
- maxpktlen, NULL, M_WAIT, 1, 0);
+ bytes_to_alloc, NULL, M_WAIT, 1, 0);
+ }
+
+ if (freelist == NULL) {
+ socket_lock(so, 0);
+ error = ENOMEM;
+ goto release;
+ }
+ /*
+ * Copy each uio of the set into its own mbuf packet
+ */
+ for (i = uiofirst, m = freelist;
+ i < uiolast && m != NULL;
+ i++) {
+ int bytes_to_copy;
+ struct mbuf *n;
+ struct uio *auio = uioarray[i];
- if (freelist == NULL) {
- socket_lock(so, 0);
- error = ENOMEM;
- goto release;
- }
+ bytes_to_copy = uio_resid(auio);
+
+ /* Do nothing for empty messages */
+ if (bytes_to_copy == 0)
+ continue;
/*
- * Copy each uio of the set into its own mbuf packet
+ * Leave headroom for protocol headers
+ * in the first mbuf of the chain
*/
- for (i = uiofirst, m = freelist;
- i < uiolast && m != NULL;
- i++) {
- int bytes_to_copy;
- struct mbuf *n;
- struct uio *auio = uioarray[i];
-
- bytes_to_copy = uio_resid(auio);
-
- /* Do nothing for empty messages */
- if (bytes_to_copy == 0)
- continue;
-
- for (n = m; n != NULL; n = n->m_next) {
- mlen = mbuf_maxlen(n);
-
- len = imin(mlen, bytes_to_copy);
-
- /*
- * Note: uiomove() decrements the iovec
- * length
- */
- error = uiomove(mtod(n, caddr_t),
- len, auio);
- if (error != 0)
- break;
- n->m_len = len;
- m->m_pkthdr.len += len;
+ m->m_data += headroom;
+
+ for (n = m; n != NULL; n = n->m_next) {
+ if ((m->m_flags & M_EXT))
+ mlen = m->m_ext.ext_size -
+ m_leadingspace(m);
+ else if ((m->m_flags & M_PKTHDR))
+ mlen =
+ MHLEN - m_leadingspace(m);
+ else
+ mlen = MLEN - m_leadingspace(m);
+ len = imin(mlen, bytes_to_copy);
- VERIFY(m->m_pkthdr.len <= maxpktlen);
-
- bytes_to_copy -= len;
- resid -= len;
- }
- if (m->m_pkthdr.len == 0) {
- printf("%s so %llx pkt %llx len null\n",
- __func__,
- (uint64_t)VM_KERNEL_ADDRPERM(so),
- (uint64_t)VM_KERNEL_ADDRPERM(m));
- }
+ /*
+ * Note: uiomove() decrements the iovec
+ * length
+ */
+ error = uiomove(mtod(n, caddr_t),
+ len, auio);
if (error != 0)
break;
- m = m->m_nextpkt;
- }
+ n->m_len = len;
+ m->m_pkthdr.len += len;
- socket_lock(so, 0);
+ VERIFY(m->m_pkthdr.len <= maxpktlen);
- if (error)
- goto release;
- top = freelist;
- freelist = NULL;
+ bytes_to_copy -= len;
+ resid -= len;
+ }
+ if (m->m_pkthdr.len == 0) {
+ printf(
+ "%s:%d so %llx pkt %llx type %u len null\n",
+ __func__, __LINE__,
+ (uint64_t)DEBUG_KERNEL_ADDRPERM(so),
+ (uint64_t)DEBUG_KERNEL_ADDRPERM(m),
+ m->m_type);
+ }
+ if (error != 0)
+ break;
+ m = m->m_nextpkt;
}
+ socket_lock(so, 0);
+
+ if (error)
+ goto release;
+ top = freelist;
+ freelist = NULL;
+
if (dontroute)
so->so_options |= SO_DONTROUTE;
if ((flags & MSG_SKIPCFIL) == 0) {
struct mbuf **prevnextp = NULL;
-
+
for (i = uiofirst, m = top;
i < uiolast && m != NULL;
i++) {
/*
* Socket filter processing
*/
- error = sflt_data_out(so, addr, &m,
- &control, 0);
+ error = sflt_data_out(so, NULL, &m,
+ NULL, 0);
if (error != 0 && error != EJUSTRETURN)
goto release;
-
+
#if CONTENT_FILTER
if (error == 0) {
/*
* Content filter processing
*/
- error = cfil_sock_data_out(so, addr, m,
- control, 0);
+ error = cfil_sock_data_out(so, NULL, m,
+ NULL, 0);
if (error != 0 && error != EJUSTRETURN)
goto release;
}
*prevnextp = nextpkt;
else
top = nextpkt;
- }
-
+ }
+
m = nextpkt;
if (m != NULL)
prevnextp = &m->m_nextpkt;
}
if (top != NULL)
error = (*so->so_proto->pr_usrreqs->pru_send_list)
- (so, 0, top, addr, control, p);
+ (so, 0, top, NULL, NULL, p);
if (dontroute)
so->so_options &= ~SO_DONTROUTE;
- clen = 0;
top = NULL;
uiofirst = uiolast;
} while (resid > 0 && error == 0);
out:
if (top != NULL)
m_freem(top);
- if (control != NULL)
- m_freem(control);
if (freelist != NULL)
m_freem_list(freelist);
return (error);
}
+/*
+ * May return ERESTART when packet is dropped by MAC policy check
+ */
+static int
+soreceive_addr(struct proc *p, struct socket *so, struct sockaddr **psa,
+ int flags, struct mbuf **mp, struct mbuf **nextrecordp, int canwait)
+{
+ int error = 0;
+ struct mbuf *m = *mp;
+ struct mbuf *nextrecord = *nextrecordp;
+
+ KASSERT(m->m_type == MT_SONAME, ("receive 1a"));
+#if CONFIG_MACF_SOCKET_SUBSET
+ /*
+ * Call the MAC framework for policy checking if we're in
+ * the user process context and the socket isn't connected.
+ */
+ if (p != kernproc && !(so->so_state & SS_ISCONNECTED)) {
+ struct mbuf *m0 = m;
+ /*
+ * Dequeue this record (temporarily) from the receive
+ * list since we're about to drop the socket's lock
+ * where a new record may arrive and be appended to
+ * the list. Upon MAC policy failure, the record
+ * will be freed. Otherwise, we'll add it back to
+ * the head of the list. We cannot rely on SB_LOCK
+ * because append operation uses the socket's lock.
+ */
+ do {
+ m->m_nextpkt = NULL;
+ sbfree(&so->so_rcv, m);
+ m = m->m_next;
+ } while (m != NULL);
+ m = m0;
+ so->so_rcv.sb_mb = nextrecord;
+ SB_EMPTY_FIXUP(&so->so_rcv);
+ SBLASTRECORDCHK(&so->so_rcv, "soreceive 1a");
+ SBLASTMBUFCHK(&so->so_rcv, "soreceive 1a");
+ socket_unlock(so, 0);
+
+ if (mac_socket_check_received(proc_ucred(p), so,
+ mtod(m, struct sockaddr *)) != 0) {
+ /*
+ * MAC policy failure; free this record and
+ * process the next record (or block until
+ * one is available). We have adjusted sb_cc
+ * and sb_mbcnt above so there is no need to
+ * call sbfree() again.
+ */
+ m_freem(m);
+ /*
+ * Clear SB_LOCK but don't unlock the socket.
+ * Process the next record or wait for one.
+ */
+ socket_lock(so, 0);
+ sbunlock(&so->so_rcv, TRUE); /* stay locked */
+ error = ERESTART;
+ goto done;
+ }
+ socket_lock(so, 0);
+ /*
+ * If the socket has been defunct'd, drop it.
+ */
+ if (so->so_flags & SOF_DEFUNCT) {
+ m_freem(m);
+ error = ENOTCONN;
+ goto done;
+ }
+ /*
+ * Re-adjust the socket receive list and re-enqueue
+ * the record in front of any packets which may have
+ * been appended while we dropped the lock.
+ */
+ for (m = m0; m->m_next != NULL; m = m->m_next)
+ sballoc(&so->so_rcv, m);
+ sballoc(&so->so_rcv, m);
+ if (so->so_rcv.sb_mb == NULL) {
+ so->so_rcv.sb_lastrecord = m0;
+ so->so_rcv.sb_mbtail = m;
+ }
+ m = m0;
+ nextrecord = m->m_nextpkt = so->so_rcv.sb_mb;
+ so->so_rcv.sb_mb = m;
+ SBLASTRECORDCHK(&so->so_rcv, "soreceive 1b");
+ SBLASTMBUFCHK(&so->so_rcv, "soreceive 1b");
+ }
+#endif /* CONFIG_MACF_SOCKET_SUBSET */
+ if (psa != NULL) {
+ *psa = dup_sockaddr(mtod(m, struct sockaddr *), canwait);
+ if ((*psa == NULL) && (flags & MSG_NEEDSA)) {
+ error = EWOULDBLOCK;
+ goto done;
+ }
+ }
+ if (flags & MSG_PEEK) {
+ m = m->m_next;
+ } else {
+ sbfree(&so->so_rcv, m);
+ if (m->m_next == NULL && so->so_rcv.sb_cc != 0) {
+ panic("%s: about to create invalid socketbuf",
+ __func__);
+ /* NOTREACHED */
+ }
+ MFREE(m, so->so_rcv.sb_mb);
+ m = so->so_rcv.sb_mb;
+ if (m != NULL) {
+ m->m_nextpkt = nextrecord;
+ } else {
+ so->so_rcv.sb_mb = nextrecord;
+ SB_EMPTY_FIXUP(&so->so_rcv);
+ }
+ }
+done:
+ *mp = m;
+ *nextrecordp = nextrecord;
+
+ return (error);
+}
+
+/*
+ * Process one or more MT_CONTROL mbufs present before any data mbufs
+ * in the first mbuf chain on the socket buffer. If MSG_PEEK, we
+ * just copy the data; if !MSG_PEEK, we call into the protocol to
+ * perform externalization.
+ */
+static int
+soreceive_ctl(struct socket *so, struct mbuf **controlp, int flags,
+ struct mbuf **mp, struct mbuf **nextrecordp)
+{
+ int error = 0;
+ struct mbuf *cm = NULL, *cmn;
+ struct mbuf **cme = &cm;
+ struct sockbuf *sb_rcv = &so->so_rcv;
+ struct mbuf **msgpcm = NULL;
+ struct mbuf *m = *mp;
+ struct mbuf *nextrecord = *nextrecordp;
+ struct protosw *pr = so->so_proto;
+
+ /*
+ * Externalizing the control messages would require us to
+ * drop the socket's lock below. Once we re-acquire the
+ * lock, the mbuf chain might change. In order to preserve
+ * consistency, we unlink all control messages from the
+ * first mbuf chain in one shot and link them separately
+ * onto a different chain.
+ */
+ do {
+ if (flags & MSG_PEEK) {
+ if (controlp != NULL) {
+ if (*controlp == NULL) {
+ msgpcm = controlp;
+ }
+ *controlp = m_copy(m, 0, m->m_len);
+
+ /*
+ * If we failed to allocate an mbuf,
+ * release any previously allocated
+ * mbufs for control data. Return
+ * an error. Keep the mbufs in the
+ * socket as this is using
+ * MSG_PEEK flag.
+ */
+ if (*controlp == NULL) {
+ m_freem(*msgpcm);
+ error = ENOBUFS;
+ goto done;
+ }
+ controlp = &(*controlp)->m_next;
+ }
+ m = m->m_next;
+ } else {
+ m->m_nextpkt = NULL;
+ sbfree(sb_rcv, m);
+ sb_rcv->sb_mb = m->m_next;
+ m->m_next = NULL;
+ *cme = m;
+ cme = &(*cme)->m_next;
+ m = sb_rcv->sb_mb;
+ }
+ } while (m != NULL && m->m_type == MT_CONTROL);
+
+ if (!(flags & MSG_PEEK)) {
+ if (sb_rcv->sb_mb != NULL) {
+ sb_rcv->sb_mb->m_nextpkt = nextrecord;
+ } else {
+ sb_rcv->sb_mb = nextrecord;
+ SB_EMPTY_FIXUP(sb_rcv);
+ }
+ if (nextrecord == NULL)
+ sb_rcv->sb_lastrecord = m;
+ }
+
+ SBLASTRECORDCHK(&so->so_rcv, "soreceive ctl");
+ SBLASTMBUFCHK(&so->so_rcv, "soreceive ctl");
+
+ while (cm != NULL) {
+ int cmsg_type;
+
+ cmn = cm->m_next;
+ cm->m_next = NULL;
+ cmsg_type = mtod(cm, struct cmsghdr *)->cmsg_type;
+
+ /*
+ * Call the protocol to externalize SCM_RIGHTS message
+ * and return the modified message to the caller upon
+ * success. Otherwise, all other control messages are
+ * returned unmodified to the caller. Note that we
+ * only get into this loop if MSG_PEEK is not set.
+ */
+ if (pr->pr_domain->dom_externalize != NULL &&
+ cmsg_type == SCM_RIGHTS) {
+ /*
+ * Release socket lock: see 3903171. This
+ * would also allow more records to be appended
+ * to the socket buffer. We still have SB_LOCK
+ * set on it, so we can be sure that the head
+ * of the mbuf chain won't change.
+ */
+ socket_unlock(so, 0);
+ error = (*pr->pr_domain->dom_externalize)(cm);
+ socket_lock(so, 0);
+ } else {
+ error = 0;
+ }
+
+ if (controlp != NULL && error == 0) {
+ *controlp = cm;
+ controlp = &(*controlp)->m_next;
+ } else {
+ (void) m_free(cm);
+ }
+ cm = cmn;
+ }
+ /*
+ * Update the value of nextrecord in case we received new
+ * records when the socket was unlocked above for
+ * externalizing SCM_RIGHTS.
+ */
+ if (m != NULL)
+ nextrecord = sb_rcv->sb_mb->m_nextpkt;
+ else
+ nextrecord = sb_rcv->sb_mb;
+
+done:
+ *mp = m;
+ *nextrecordp = nextrecord;
+
+ return (error);
+}
+
/*
* Implement receive operations on a socket.
* We depend on the way that records are added to the sockbuf
int flags, error, offset;
user_ssize_t len;
struct protosw *pr = so->so_proto;
- int moff, type =0;
+ int moff, type = 0;
user_ssize_t orig_resid = uio_resid(uio);
user_ssize_t delayed_copy_len;
int can_delay;
int need_event;
struct proc *p = current_proc();
-
- KERNEL_DEBUG(DBG_FNC_SORECEIVE | DBG_FUNC_START, so, uio_resid(uio),
- so->so_rcv.sb_cc, so->so_rcv.sb_lowat, so->so_rcv.sb_hiwat);
+ boolean_t en_tracing = FALSE;
/*
* Sanity check on the length passed by caller as we are making 'int'
if (orig_resid < 0 || orig_resid > INT_MAX)
return (EINVAL);
+ KERNEL_DEBUG(DBG_FNC_SORECEIVE | DBG_FUNC_START, so,
+ uio_resid(uio), so->so_rcv.sb_cc, so->so_rcv.sb_lowat,
+ so->so_rcv.sb_hiwat);
+
socket_lock(so, 1);
so_update_last_owner_locked(so, p);
so_update_policy(so);
error = ENOTCONN;
SODEFUNCTLOG(("%s[%d]: defunct so 0x%llx [%d,%d] (%d)\n",
- __func__, proc_pid(p), (uint64_t)VM_KERNEL_ADDRPERM(so),
+ __func__, proc_pid(p), (uint64_t)DEBUG_KERNEL_ADDRPERM(so),
SOCK_DOM(so), SOCK_TYPE(so), error));
/*
* This socket should have been disconnected and flushed
return (error);
}
+ if ((so->so_flags1 & SOF1_PRECONNECT_DATA) &&
+ pr->pr_usrreqs->pru_preconnect) {
+ /*
+ * A user may set the CONNECT_RESUME_ON_READ_WRITE-flag but not
+ * calling write() right after this. *If* the app calls a read
+ * we do not want to block this read indefinetely. Thus,
+ * we trigger a connect so that the session gets initiated.
+ */
+ error = (*pr->pr_usrreqs->pru_preconnect)(so);
+
+ if (error) {
+ socket_unlock(so, 1);
+ return (error);
+ }
+ }
+
+ if (ENTR_SHOULDTRACE &&
+ (SOCK_CHECK_DOM(so, AF_INET) || SOCK_CHECK_DOM(so, AF_INET6))) {
+ /*
+ * enable energy tracing for inet sockets that go over
+ * non-loopback interfaces only.
+ */
+ struct inpcb *inp = sotoinpcb(so);
+ if (inp->inp_last_outifp != NULL &&
+ !(inp->inp_last_outifp->if_flags & IFF_LOOPBACK)) {
+ en_tracing = TRUE;
+ KERNEL_ENERGYTRACE(kEnTrActKernSockRead, DBG_FUNC_START,
+ VM_KERNEL_ADDRPERM(so),
+ ((so->so_state & SS_NBIO) ?
+ kEnTrFlagNonBlocking : 0),
+ (int64_t)orig_resid);
+ }
+ }
+
/*
* When SO_WANTOOBFLAG is set we try to get out-of-band data
* regardless of the flags argument. Here is the case were
}
}
socket_unlock(so, 1);
+ if (en_tracing) {
+ KERNEL_ENERGYTRACE(kEnTrActKernSockRead, DBG_FUNC_END,
+ VM_KERNEL_ADDRPERM(so), 0,
+ (int64_t)(orig_resid - uio_resid(uio)));
+ }
KERNEL_DEBUG(DBG_FNC_SORECEIVE | DBG_FUNC_END, error,
0, 0, 0, 0);
#ifdef MORE_LOCKING_DEBUG
if (so->so_usecount <= 1)
printf("soreceive: sblock so=0x%llx ref=%d on socket\n",
- (uint64_t)VM_KERNEL_ADDRPERM(so), so->so_usecount);
+ (uint64_t)DEBUG_KERNEL_ADDRPERM(so), so->so_usecount);
#endif
/*
* See if the socket has been closed (SS_NOFDREF|SS_CANTRCVMORE)
socket_unlock(so, 1);
KERNEL_DEBUG(DBG_FNC_SORECEIVE | DBG_FUNC_END, error,
0, 0, 0, 0);
+ if (en_tracing) {
+ KERNEL_ENERGYTRACE(kEnTrActKernSockRead, DBG_FUNC_END,
+ VM_KERNEL_ADDRPERM(so), 0,
+ (int64_t)(orig_resid - uio_resid(uio)));
+ }
return (error);
}
cfil_sock_data_pending(&so->so_rcv) != 0)
CFIL_LOG(LOG_INFO,
"so %llx ignore SS_CANTRCVMORE",
- (uint64_t)VM_KERNEL_ADDRPERM(so));
- else
+ (uint64_t)DEBUG_KERNEL_ADDRPERM(so));
+ else
#endif /* CONTENT_FILTER */
if (m != NULL)
goto dontblock;
}
if (uio_resid(uio) == 0)
goto release;
+
if ((so->so_state & SS_NBIO) ||
(flags & (MSG_DONTWAIT|MSG_NBIO))) {
error = EWOULDBLOCK;
socket_unlock(so, 1);
KERNEL_DEBUG(DBG_FNC_SORECEIVE | DBG_FUNC_END, error,
0, 0, 0, 0);
+ if (en_tracing) {
+ KERNEL_ENERGYTRACE(kEnTrActKernSockRead, DBG_FUNC_END,
+ VM_KERNEL_ADDRPERM(so), 0,
+ (int64_t)(orig_resid - uio_resid(uio)));
+ }
return (error);
}
goto restart;
SBLASTRECORDCHK(&so->so_rcv, "soreceive 1");
SBLASTMBUFCHK(&so->so_rcv, "soreceive 1");
nextrecord = m->m_nextpkt;
- if ((pr->pr_flags & PR_ADDR) && m->m_type == MT_SONAME) {
- KASSERT(m->m_type == MT_SONAME, ("receive 1a"));
-#if CONFIG_MACF_SOCKET_SUBSET
- /*
- * Call the MAC framework for policy checking if we're in
- * the user process context and the socket isn't connected.
- */
- if (p != kernproc && !(so->so_state & SS_ISCONNECTED)) {
- struct mbuf *m0 = m;
- /*
- * Dequeue this record (temporarily) from the receive
- * list since we're about to drop the socket's lock
- * where a new record may arrive and be appended to
- * the list. Upon MAC policy failure, the record
- * will be freed. Otherwise, we'll add it back to
- * the head of the list. We cannot rely on SB_LOCK
- * because append operation uses the socket's lock.
- */
- do {
- m->m_nextpkt = NULL;
- sbfree(&so->so_rcv, m);
- m = m->m_next;
- } while (m != NULL);
- m = m0;
- so->so_rcv.sb_mb = nextrecord;
- SB_EMPTY_FIXUP(&so->so_rcv);
- SBLASTRECORDCHK(&so->so_rcv, "soreceive 1a");
- SBLASTMBUFCHK(&so->so_rcv, "soreceive 1a");
- socket_unlock(so, 0);
- if (mac_socket_check_received(proc_ucred(p), so,
- mtod(m, struct sockaddr *)) != 0) {
- /*
- * MAC policy failure; free this record and
- * process the next record (or block until
- * one is available). We have adjusted sb_cc
- * and sb_mbcnt above so there is no need to
- * call sbfree() again.
- */
- do {
- m = m_free(m);
- } while (m != NULL);
- /*
- * Clear SB_LOCK but don't unlock the socket.
- * Process the next record or wait for one.
- */
- socket_lock(so, 0);
- sbunlock(&so->so_rcv, TRUE); /* stay locked */
- goto restart;
- }
- socket_lock(so, 0);
- /*
- * If the socket has been defunct'd, drop it.
- */
- if (so->so_flags & SOF_DEFUNCT) {
- m_freem(m);
- error = ENOTCONN;
- goto release;
- }
- /*
- * Re-adjust the socket receive list and re-enqueue
- * the record in front of any packets which may have
- * been appended while we dropped the lock.
- */
- for (m = m0; m->m_next != NULL; m = m->m_next)
- sballoc(&so->so_rcv, m);
- sballoc(&so->so_rcv, m);
- if (so->so_rcv.sb_mb == NULL) {
- so->so_rcv.sb_lastrecord = m0;
- so->so_rcv.sb_mbtail = m;
- }
- m = m0;
- nextrecord = m->m_nextpkt = so->so_rcv.sb_mb;
- so->so_rcv.sb_mb = m;
- SBLASTRECORDCHK(&so->so_rcv, "soreceive 1b");
- SBLASTMBUFCHK(&so->so_rcv, "soreceive 1b");
- }
-#endif /* CONFIG_MACF_SOCKET_SUBSET */
+ if ((pr->pr_flags & PR_ADDR) && m->m_type == MT_SONAME) {
+ error = soreceive_addr(p, so, psa, flags, &m, &nextrecord,
+ mp0 == NULL);
+ if (error == ERESTART)
+ goto restart;
+ else if (error != 0)
+ goto release;
orig_resid = 0;
- if (psa != NULL) {
- *psa = dup_sockaddr(mtod(m, struct sockaddr *),
- mp0 == NULL);
- if ((*psa == NULL) && (flags & MSG_NEEDSA)) {
- error = EWOULDBLOCK;
- goto release;
- }
- }
- if (flags & MSG_PEEK) {
- m = m->m_next;
- } else {
- sbfree(&so->so_rcv, m);
- if (m->m_next == NULL && so->so_rcv.sb_cc != 0) {
- panic("%s: about to create invalid socketbuf",
- __func__);
- /* NOTREACHED */
- }
- MFREE(m, so->so_rcv.sb_mb);
- m = so->so_rcv.sb_mb;
- if (m != NULL) {
- m->m_nextpkt = nextrecord;
- } else {
- so->so_rcv.sb_mb = nextrecord;
- SB_EMPTY_FIXUP(&so->so_rcv);
- }
- }
}
/*
* perform externalization.
*/
if (m != NULL && m->m_type == MT_CONTROL) {
- struct mbuf *cm = NULL, *cmn;
- struct mbuf **cme = &cm;
- struct sockbuf *sb_rcv = &so->so_rcv;
- struct mbuf **msgpcm = NULL;
-
- /*
- * Externalizing the control messages would require us to
- * drop the socket's lock below. Once we re-acquire the
- * lock, the mbuf chain might change. In order to preserve
- * consistency, we unlink all control messages from the
- * first mbuf chain in one shot and link them separately
- * onto a different chain.
- */
- do {
- if (flags & MSG_PEEK) {
- if (controlp != NULL) {
- if (*controlp == NULL) {
- msgpcm = controlp;
- }
- *controlp = m_copy(m, 0, m->m_len);
-
- /*
- * If we failed to allocate an mbuf,
- * release any previously allocated
- * mbufs for control data. Return
- * an error. Keep the mbufs in the
- * socket as this is using
- * MSG_PEEK flag.
- */
- if (*controlp == NULL) {
- m_freem(*msgpcm);
- error = ENOBUFS;
- goto release;
- }
- controlp = &(*controlp)->m_next;
- }
- m = m->m_next;
- } else {
- m->m_nextpkt = NULL;
- sbfree(sb_rcv, m);
- sb_rcv->sb_mb = m->m_next;
- m->m_next = NULL;
- *cme = m;
- cme = &(*cme)->m_next;
- m = sb_rcv->sb_mb;
- }
- } while (m != NULL && m->m_type == MT_CONTROL);
-
- if (!(flags & MSG_PEEK)) {
- if (sb_rcv->sb_mb != NULL) {
- sb_rcv->sb_mb->m_nextpkt = nextrecord;
- } else {
- sb_rcv->sb_mb = nextrecord;
- SB_EMPTY_FIXUP(sb_rcv);
- }
- if (nextrecord == NULL)
- sb_rcv->sb_lastrecord = m;
- }
-
- SBLASTRECORDCHK(&so->so_rcv, "soreceive ctl");
- SBLASTMBUFCHK(&so->so_rcv, "soreceive ctl");
-
- while (cm != NULL) {
- int cmsg_type;
-
- cmn = cm->m_next;
- cm->m_next = NULL;
- cmsg_type = mtod(cm, struct cmsghdr *)->cmsg_type;
-
- /*
- * Call the protocol to externalize SCM_RIGHTS message
- * and return the modified message to the caller upon
- * success. Otherwise, all other control messages are
- * returned unmodified to the caller. Note that we
- * only get into this loop if MSG_PEEK is not set.
- */
- if (pr->pr_domain->dom_externalize != NULL &&
- cmsg_type == SCM_RIGHTS) {
- /*
- * Release socket lock: see 3903171. This
- * would also allow more records to be appended
- * to the socket buffer. We still have SB_LOCK
- * set on it, so we can be sure that the head
- * of the mbuf chain won't change.
- */
- socket_unlock(so, 0);
- error = (*pr->pr_domain->dom_externalize)(cm);
- socket_lock(so, 0);
- } else {
- error = 0;
- }
-
- if (controlp != NULL && error == 0) {
- *controlp = cm;
- controlp = &(*controlp)->m_next;
- orig_resid = 0;
- } else {
- (void) m_free(cm);
- }
- cm = cmn;
- }
- /*
- * Update the value of nextrecord in case we received new
- * records when the socket was unlocked above for
- * externalizing SCM_RIGHTS.
- */
- if (m != NULL)
- nextrecord = sb_rcv->sb_mb->m_nextpkt;
- else
- nextrecord = sb_rcv->sb_mb;
+ error = soreceive_ctl(so, controlp, flags, &m, &nextrecord);
+ if (error != 0)
+ goto release;
orig_resid = 0;
}
sbunlock(&so->so_rcv, FALSE); /* will unlock socket */
+ if (en_tracing) {
+ KERNEL_ENERGYTRACE(kEnTrActKernSockRead, DBG_FUNC_END,
+ VM_KERNEL_ADDRPERM(so),
+ ((error == EWOULDBLOCK) ? kEnTrFlagNoWork : 0),
+ (int64_t)(orig_resid - uio_resid(uio)));
+ }
KERNEL_DEBUG(DBG_FNC_SORECEIVE | DBG_FUNC_END, so, uio_resid(uio),
so->so_rcv.sb_cc, 0, error);
return (error);
}
+static int
+sodelayed_copy_list(struct socket *so, struct recv_msg_elem *msgarray,
+ u_int uiocnt, struct mbuf **free_list, user_ssize_t *resid)
+{
+#pragma unused(so)
+ int error = 0;
+ struct mbuf *ml, *m;
+ int i = 0;
+ struct uio *auio;
+
+ for (ml = *free_list, i = 0; ml != NULL && i < uiocnt;
+ ml = ml->m_nextpkt, i++) {
+ auio = msgarray[i].uio;
+ for (m = ml; m != NULL; m = m->m_next) {
+ error = uiomove(mtod(m, caddr_t), m->m_len, auio);
+ if (error != 0)
+ goto out;
+ }
+ }
+out:
+ m_freem_list(*free_list);
+
+ *free_list = NULL;
+ *resid = 0;
+
+ return (error);
+}
+
int
-soreceive_list(struct socket *so, struct sockaddr **psa, struct uio **uioarray,
- u_int uiocnt, struct mbuf **mp0, struct mbuf **controlp, int *flagsp)
+soreceive_list(struct socket *so, struct recv_msg_elem *msgarray, u_int uiocnt,
+ int *flagsp)
{
- struct mbuf *m, **mp;
+ struct mbuf *m;
struct mbuf *nextrecord;
- struct mbuf *ml = NULL, *free_list = NULL;
- int flags, error, offset;
- user_ssize_t len;
+ struct mbuf *ml = NULL, *free_list = NULL, *free_tail = NULL;
+ int error;
+ user_ssize_t len, pktlen, delayed_copy_len = 0;
struct protosw *pr = so->so_proto;
- user_ssize_t orig_resid, resid;
+ user_ssize_t resid;
struct proc *p = current_proc();
struct uio *auio = NULL;
- int i = 0;
+ int npkts = 0;
int sblocked = 0;
+ struct sockaddr **psa = NULL;
+ struct mbuf **controlp = NULL;
+ int can_delay;
+ int flags;
+ struct mbuf *free_others = NULL;
KERNEL_DEBUG(DBG_FNC_SORECEIVE_LIST | DBG_FUNC_START,
so, uiocnt,
so->so_rcv.sb_cc, so->so_rcv.sb_lowat, so->so_rcv.sb_hiwat);
- mp = mp0;
- if (psa != NULL)
- *psa = NULL;
- if (controlp != NULL)
- *controlp = NULL;
- if (flagsp != NULL)
- flags = *flagsp &~ MSG_EOR;
- else
- flags = 0;
- /*
- * Disallow functionality not currently supported
- */
- if (mp0 != NULL) {
- printf("%s mp0 not supported\n", __func__);
- error = EOPNOTSUPP;
- goto out;
- }
- if (psa != NULL) {
- printf("%s sockaddr not supported\n", __func__);
- error = EOPNOTSUPP;
- goto out;
- }
- if (controlp != NULL) {
- printf("%s control not supported\n", __func__);
- error = EOPNOTSUPP;
- goto out;
- }
-
/*
* Sanity checks:
* - Only supports don't wait flags
* - Protocol must support packet chains
* - The uio array is NULL (should we panic?)
*/
- if (flags & ~(MSG_DONTWAIT | MSG_NBIO)) {
- printf("%s flags not supported\n", __func__);
- error = EOPNOTSUPP;
+ if (flagsp != NULL)
+ flags = *flagsp;
+ else
+ flags = 0;
+ if (flags & ~(MSG_PEEK | MSG_WAITALL | MSG_DONTWAIT | MSG_NEEDSA |
+ MSG_NBIO)) {
+ printf("%s invalid flags 0x%x\n", __func__, flags);
+ error = EINVAL;
goto out;
}
if (so->so_type != SOCK_DGRAM) {
error = EPROTONOSUPPORT;
goto out;
}
- if (uioarray == NULL) {
+ if (msgarray == NULL) {
printf("%s uioarray is NULL\n", __func__);
error = EINVAL;
goto out;
* Sanity check on the length passed by caller as we are making 'int'
* comparisons
*/
- resid = orig_resid = uio_array_resid(uioarray, uiocnt);
- if (orig_resid < 0 || orig_resid > INT_MAX) {
+ resid = recv_msg_array_resid(msgarray, uiocnt);
+ if (resid < 0 || resid > INT_MAX) {
error = EINVAL;
goto out;
}
+ if (!(flags & MSG_PEEK) && sorecvmincopy > 0)
+ can_delay = 1;
+ else
+ can_delay = 0;
+
socket_lock(so, 1);
so_update_last_owner_locked(so, p);
so_update_policy(so);
#if NECP
so_update_necp_policy(so, NULL, NULL);
#endif /* NECP */
-
+
/*
* If a recv attempt is made on a previously-accepted socket
* that has been marked as inactive (disconnected), reject
error = ENOTCONN;
SODEFUNCTLOG(("%s[%d]: defunct so 0x%llx [%d,%d] (%d)\n",
- __func__, proc_pid(p), (uint64_t)VM_KERNEL_ADDRPERM(so),
+ __func__, proc_pid(p), (uint64_t)DEBUG_KERNEL_ADDRPERM(so),
SOCK_DOM(so), SOCK_TYPE(so), error));
/*
* This socket should have been disconnected and flushed
sb_empty_assert(sb, __func__);
goto release;
}
- if (mp != NULL)
- *mp = NULL;
+
+next:
+ /*
+ * The uio may be empty
+ */
+ if (npkts >= uiocnt) {
+ error = 0;
+ goto release;
+ }
restart:
/*
* See if the socket has been closed (SS_NOFDREF|SS_CANTRCVMORE)
}
sblocked = 1;
- /*
- * Skip empty uio
- */
- auio = uioarray[i];
- while (uio_resid(auio) == 0) {
- i++;
- if (i >= uiocnt) {
- error = 0;
- goto release;
- }
- }
-
m = so->so_rcv.sb_mb;
/*
* Block awaiting more datagram if needed
*/
- if (m == NULL) {
+ if (m == NULL || (((flags & MSG_DONTWAIT) == 0 &&
+ (so->so_rcv.sb_cc < so->so_rcv.sb_lowat ||
+ ((flags & MSG_WAITALL) && npkts < uiocnt))))) {
/*
* Panic if we notice inconsistencies in the socket's
* receive list; both sb_mb and sb_cc should correctly
if (so->so_error) {
error = so->so_error;
+ if ((flags & MSG_PEEK) == 0)
+ so->so_error = 0;
goto release;
}
if (so->so_state & SS_CANTRCVMORE) {
}
/*
* Do not block if we got some data
- * Note: We could use MSG_WAITALL to wait
*/
- resid = uio_array_resid(uioarray, uiocnt);
- if (resid != orig_resid) {
+ if (free_list != NULL) {
error = 0;
goto release;
}
-
+
SBLASTRECORDCHK(&so->so_rcv, "soreceive sbwait 1");
SBLASTMBUFCHK(&so->so_rcv, "soreceive sbwait 1");
goto restart;
}
- if (m->m_pkthdr.len == 0) {
- printf("%s so %llx pkt %llx len is null\n",
- __func__,
- (uint64_t)VM_KERNEL_ADDRPERM(so),
- (uint64_t)VM_KERNEL_ADDRPERM(m));
- goto restart;
- }
OSIncrementAtomicLong(&p->p_stats->p_ru.ru_msgrcv);
SBLASTRECORDCHK(&so->so_rcv, "soreceive 1");
SBLASTMBUFCHK(&so->so_rcv, "soreceive 1");
/*
* Consume the current uio index as we have a datagram
*/
- i += 1;
+ auio = msgarray[npkts].uio;
+ resid = uio_resid(auio);
+ msgarray[npkts].which |= SOCK_MSG_DATA;
+ psa = (msgarray[npkts].which & SOCK_MSG_SA) ?
+ &msgarray[npkts].psa : NULL;
+ controlp = (msgarray[npkts].which & SOCK_MSG_CONTROL) ?
+ &msgarray[npkts].controlp : NULL;
+ npkts += 1;
nextrecord = m->m_nextpkt;
-#if SO_RECEIVE_LIST_SOCKADDR_NOT_YET
if ((pr->pr_flags & PR_ADDR) && m->m_type == MT_SONAME) {
- /*
- * to be adapted from soreceive()
- */
+ error = soreceive_addr(p, so, psa, flags, &m, &nextrecord, 1);
+ if (error == ERESTART)
+ goto restart;
+ else if (error != 0)
+ goto release;
}
-#endif /* SO_RECEIVE_LIST_SOCKADDR_NOT_YET */
-#if SO_RECEIVE_LIST_CONTROL_NOT_YET
- /*
- * Process one or more MT_CONTROL mbufs present before any data mbufs
- * in the first mbuf chain on the socket buffer. If MSG_PEEK, we
- * just copy the data; if !MSG_PEEK, we call into the protocol to
- * perform externalization.
- */
if (m != NULL && m->m_type == MT_CONTROL) {
- /*
- * to be adapted from soreceive()
- */
+ error = soreceive_ctl(so, controlp, flags, &m, &nextrecord);
+ if (error != 0)
+ goto release;
}
-#endif /* SO_RECEIVE_LIST_CONTROL_NOT_YET */
- offset = 0;
+ if (m->m_pkthdr.len == 0) {
+ printf("%s:%d so %llx pkt %llx type %u pktlen null\n",
+ __func__, __LINE__,
+ (uint64_t)DEBUG_KERNEL_ADDRPERM(so),
+ (uint64_t)DEBUG_KERNEL_ADDRPERM(m),
+ m->m_type);
+ }
/*
- * Loop to copy out the mbufs of the current record
+ * Loop to copy the mbufs of the current record
+ * Support zero length packets
*/
- while (m != NULL && uio_resid(auio) > 0 && error == 0) {
- len = uio_resid(auio);
-
+ ml = NULL;
+ pktlen = 0;
+ while (m != NULL && (len = resid - pktlen) >= 0 && error == 0) {
if (m->m_len == 0)
- printf("%s: so %llx m %llx m_len is 0\n",
- __func__,
- (uint64_t)VM_KERNEL_ADDRPERM(so),
- (uint64_t)VM_KERNEL_ADDRPERM(m));
-
+ panic("%p m_len zero", m);
+ if (m->m_type == 0)
+ panic("%p m_type zero", m);
/*
* Clip to the residual length
*/
if (len > m->m_len)
len = m->m_len;
+ pktlen += len;
/*
- * If mp is set, just pass back the mbufs.
- * Otherwise copy them out via the uio, then free.
+ * Copy the mbufs via the uio or delay the copy
* Sockbuf must be consistent here (points to current mbuf,
* it points to next record) when we drop priority;
* we must note any additions to the sockbuf when we
* block interrupts again.
*/
- if (mp != NULL) {
- uio_setresid(auio, (uio_resid(auio) - len));
- } else {
- SBLASTRECORDCHK(&so->so_rcv, "soreceive uiomove");
- SBLASTMBUFCHK(&so->so_rcv, "soreceive uiomove");
-
+ if (len > 0 && can_delay == 0) {
socket_unlock(so, 0);
error = uiomove(mtod(m, caddr_t), (int)len, auio);
socket_lock(so, 0);
-
if (error)
goto release;
+ } else {
+ delayed_copy_len += len;
}
+
if (len == m->m_len) {
/*
- * m was entirely copied
+ * m was entirely copied
*/
- nextrecord = m->m_nextpkt;
sbfree(&so->so_rcv, m);
+ nextrecord = m->m_nextpkt;
m->m_nextpkt = NULL;
/*
- * Move to m_next
+ * Set the first packet to the head of the free list
*/
- if (mp != NULL) {
- *mp = m;
- mp = &m->m_next;
- so->so_rcv.sb_mb = m = m->m_next;
- *mp = NULL;
- } else {
- if (free_list == NULL)
- free_list = m;
- else
- ml->m_next = m;
- ml = m;
- so->so_rcv.sb_mb = m = m->m_next;
- ml->m_next = NULL;
- ml->m_nextpkt = NULL;
+ if (free_list == NULL)
+ free_list = m;
+ /*
+ * Link current packet to tail of free list
+ */
+ if (ml == NULL) {
+ if (free_tail != NULL)
+ free_tail->m_nextpkt = m;
+ free_tail = m;
}
+ /*
+ * Link current mbuf to last mbuf of current packet
+ */
+ if (ml != NULL)
+ ml->m_next = m;
+ ml = m;
+
+ /*
+ * Move next buf to head of socket buffer
+ */
+ so->so_rcv.sb_mb = m = ml->m_next;
+ ml->m_next = NULL;
+
if (m != NULL) {
m->m_nextpkt = nextrecord;
if (nextrecord == NULL)
/*
* Stop the loop on partial copy
*/
- if (mp != NULL) {
- int copy_flag;
-
- if (flags & MSG_DONTWAIT)
- copy_flag = M_DONTWAIT;
- else
- copy_flag = M_WAIT;
- *mp = m_copym(m, 0, len, copy_flag);
- /*
- * Failed to allocate an mbuf?
- * Adjust uio_resid back, it was
- * adjusted down by len bytes which
- * we didn't copy over.
- */
- if (*mp == NULL) {
- uio_setresid(auio,
- (uio_resid(auio) + len));
- error = ENOMEM;
- break;
- }
- }
break;
}
}
if (so->so_usecount <= 1) {
panic("%s: after big while so=%llx ref=%d on socket\n",
__func__,
- (uint64_t)VM_KERNEL_ADDRPERM(so), so->so_usecount);
+ (uint64_t)DEBUG_KERNEL_ADDRPERM(so), so->so_usecount);
/* NOTREACHED */
}
#endif
*/
if (m != NULL) {
if (so->so_options & SO_DONTTRUNC) {
+ /*
+ * Copyout first the freelist then the partial mbuf
+ */
+ socket_unlock(so, 0);
+ if (delayed_copy_len)
+ error = sodelayed_copy_list(so, msgarray,
+ uiocnt, &free_list, &delayed_copy_len);
+
+ if (error == 0) {
+ error = uiomove(mtod(m, caddr_t), (int)len,
+ auio);
+ }
+ socket_lock(so, 0);
+ if (error)
+ goto release;
+
m->m_data += len;
m->m_len -= len;
so->so_rcv.sb_cc -= len;
* - There was no error
* - A packet was not truncated
* - We can still receive more data
- */
- if (i < uiocnt && error == 0 &&
- (flags & (MSG_RCVMORE | MSG_TRUNC)) == 0
- && (so->so_state & SS_CANTRCVMORE) == 0) {
+ */
+ if (npkts < uiocnt && error == 0 &&
+ (flags & (MSG_RCVMORE | MSG_TRUNC)) == 0 &&
+ (so->so_state & SS_CANTRCVMORE) == 0) {
sbunlock(&so->so_rcv, TRUE); /* keep socket locked */
sblocked = 0;
- goto restart;
+ goto next;
}
+ if (flagsp != NULL)
+ *flagsp |= flags;
release:
/*
* pru_rcvd may cause more data to be received if the socket lock
* is dropped so we set MSG_HAVEMORE now based on what we know.
- * That way the caller won't be surprised if it receives less data than requested.
+ * That way the caller won't be surprised if it receives less data
+ * than requested.
*/
if ((so->so_options & SO_WANTMORE) && so->so_rcv.sb_cc > 0)
flags |= MSG_HAVEMORE;
if (pr->pr_flags & PR_WANTRCVD && so->so_pcb)
(*pr->pr_usrreqs->pru_rcvd)(so, flags);
- if (flagsp != NULL)
- *flagsp |= flags;
if (sblocked)
sbunlock(&so->so_rcv, FALSE); /* will unlock socket */
else
socket_unlock(so, 1);
+
+ if (delayed_copy_len)
+ error = sodelayed_copy_list(so, msgarray, uiocnt,
+ &free_list, &delayed_copy_len);
out:
/*
- * Amortize the cost
+ * Amortize the cost of freeing the mbufs
*/
if (free_list != NULL)
m_freem_list(free_list);
+ if (free_others != NULL)
+ m_freem_list(free_others);
KERNEL_DEBUG(DBG_FNC_SORECEIVE_LIST | DBG_FUNC_END, error,
0, 0, 0, 0);
}
}
#endif /* CONTENT_FILTER */
-
+
error = soshutdownlock_final(so, how);
done:
if (so->so_snd.sb_flags & SB_UNIX) {
struct unpcb *unp =
(struct unpcb *)(so->so_pcb);
- if (unp != NULL && unp->unp_conn != NULL) {
+ if (unp != NULL &&
+ unp->unp_conn != NULL) {
hiwat += unp->unp_conn->unp_cc;
}
}
sowwakeup(so);
}
break;
- }
+ }
case SO_RCVLOWAT: {
int64_t data_len;
so->so_rcv.sb_lowat =
(optval > so->so_rcv.sb_hiwat) ?
so->so_rcv.sb_hiwat : optval;
- data_len = so->so_rcv.sb_cc
+ data_len = so->so_rcv.sb_cc
- so->so_rcv.sb_ctl;
if (data_len >= so->so_rcv.sb_lowat)
sorwakeup(so);
SODEFUNCTLOG(("%s[%d]: so 0x%llx [%s %s:%d -> "
"%s:%d] is now marked as %seligible for "
"defunct\n", __func__, proc_selfpid(),
- (uint64_t)VM_KERNEL_ADDRPERM(so),
+ (uint64_t)DEBUG_KERNEL_ADDRPERM(so),
(SOCK_TYPE(so) == SOCK_STREAM) ?
"TCP" : "UDP", inet_ntop(SOCK_DOM(so),
((SOCK_DOM(so) == PF_INET) ?
SODEFUNCTLOG(("%s[%d]: so 0x%llx [%d,%d] is "
"now marked as %seligible for defunct\n",
__func__, proc_selfpid(),
- (uint64_t)VM_KERNEL_ADDRPERM(so),
+ (uint64_t)DEBUG_KERNEL_ADDRPERM(so),
SOCK_DOM(so), SOCK_TYPE(so),
(so->so_flags & SOF_NODEFUNCT) ?
"not " : ""));
error = so_set_effective_uuid(so, euuid, sopt->sopt_p);
break;
}
-
+
#if NECP
case SO_NECP_ATTRIBUTES:
error = necp_set_socket_attributes(so, sopt);
break;
#endif /* MPTCP */
+ case SO_EXTENDED_BK_IDLE:
+ error = sooptcopyin(sopt, &optval, sizeof (optval),
+ sizeof (optval));
+ if (error == 0)
+ error = so_set_extended_bk_idle(so, optval);
+ break;
+
default:
error = ENOPROTOOPT;
break;
goto integer;
case SO_AWDL_UNRESTRICTED:
- if (SOCK_DOM(so) == PF_INET ||
+ if (SOCK_DOM(so) == PF_INET ||
SOCK_DOM(so) == PF_INET6) {
optval = inp_get_awdl_unrestricted(
sotoinpcb(so));
error = flow_divert_token_get(so, sopt);
break;
#endif /* FLOW_DIVERT */
-
+
#if NECP
case SO_NECP_ATTRIBUTES:
error = necp_get_socket_attributes(so, sopt);
sock_id = cfil_sock_id_from_socket(so);
- error = sooptcopyout(sopt, &sock_id,
+ error = sooptcopyout(sopt, &sock_id,
sizeof(cfil_sock_id_t));
break;
}
break;
}
optval = (so->so_flags & SOF_MPTCP_FASTJOIN);
- break;
+ /* Fixed along with rdar://19391339 */
+ goto integer;
#endif /* MPTCP */
+ case SO_EXTENDED_BK_IDLE:
+ optval = (so->so_flags1 & SOF1_EXTEND_BK_IDLE_WANTED);
+ goto integer;
+
default:
error = ENOPROTOOPT;
break;
/*
* If the caller explicitly asked for OOB results (e.g. poll()),
* save that off in the hookid field and reserve the kn_flags
- * EV_OOBAND bit for output only).
+ * EV_OOBAND bit for output only.
*/
if (kn->kn_flags & EV_OOBAND) {
kn->kn_flags &= ~EV_OOBAND;
case EVFILT_SOCK:
kn->kn_fop = &sock_filtops;
skl = &so->so_klist;
+ kn->kn_hookid = 0;
+ kn->kn_status |= KN_TOUCH;
break;
default:
socket_unlock(so, 1);
}
/* socket isn't a listener */
+ /*
+ * NOTE_LOWAT specifies new low water mark in data, i.e.
+ * the bytes of protocol data. We therefore exclude any
+ * control bytes.
+ */
kn->kn_data = so->so_rcv.sb_cc - so->so_rcv.sb_ctl;
+
/*
* Clear out EV_OOBAND that filt_soread may have set in the
* past.
*/
kn->kn_flags &= ~EV_OOBAND;
-
- if ((so->so_oobmark) || (so->so_state & SS_RCVATMARK)){
+ if ((so->so_oobmark) || (so->so_state & SS_RCVATMARK)) {
kn->kn_flags |= EV_OOBAND;
/*
* If caller registered explicit interest in OOB data,
return (1);
}
}
-
+
if ((so->so_state & SS_CANTRCVMORE)
#if CONTENT_FILTER
&& cfil_sock_data_pending(&so->so_rcv) == 0
}
int64_t lowwat = so->so_rcv.sb_lowat;
+ /*
+ * Ensure that when NOTE_LOWAT is used, the derived
+ * low water mark is bounded by socket's rcv buf's
+ * high and low water mark values.
+ */
if (kn->kn_sfflags & NOTE_LOWAT) {
if (kn->kn_sdata > so->so_rcv.sb_hiwat)
lowwat = so->so_rcv.sb_hiwat;
if ((hint & SO_FILT_HINT_LOCKED) == 0)
socket_unlock(so, 1);
- return (kn->kn_data >= lowwat);
+ /*
+ * The order below is important. Since NOTE_LOWAT
+ * overrides sb_lowat, check for NOTE_LOWAT case
+ * first.
+ */
+ if (kn->kn_sfflags & NOTE_LOWAT)
+ return (kn->kn_data >= lowwat);
+
+ return (so->so_rcv.sb_cc >= lowwat);
}
static void
ret = 1;
goto out;
}
- if (((so->so_state & SS_ISCONNECTED) == 0) &&
- (so->so_proto->pr_flags & PR_CONNREQUIRED)) {
+ if (!socanwrite(so)) {
ret = 0;
goto out;
}
+ if (so->so_flags1 & SOF1_PRECONNECT_DATA) {
+ ret = 1;
+ goto out;
+ }
int64_t lowwat = so->so_snd.sb_lowat;
if (kn->kn_sfflags & NOTE_LOWAT) {
if (kn->kn_sdata > so->so_snd.sb_hiwat)
}
#endif
else {
- return (1);
+ ret = 1;
+ goto out;
}
} else {
ret = 1;
int ret = 0, locked = 0;
struct socket *so = (struct socket *)kn->kn_fp->f_fglob->fg_data;
long ev_hint = (hint & SO_FILT_HINT_EV);
+ uint32_t level_trigger = 0;
if ((hint & SO_FILT_HINT_LOCKED) == 0) {
socket_lock(so, 1);
}
if (ev_hint & SO_FILT_HINT_CONNRESET) {
- if (kn->kn_sfflags & NOTE_CONNRESET)
- kn->kn_fflags |= NOTE_CONNRESET;
+ kn->kn_fflags |= NOTE_CONNRESET;
}
if (ev_hint & SO_FILT_HINT_TIMEOUT) {
- if (kn->kn_sfflags & NOTE_TIMEOUT)
- kn->kn_fflags |= NOTE_TIMEOUT;
+ kn->kn_fflags |= NOTE_TIMEOUT;
}
if (ev_hint & SO_FILT_HINT_NOSRCADDR) {
- if (kn->kn_sfflags & NOTE_NOSRCADDR)
- kn->kn_fflags |= NOTE_NOSRCADDR;
+ kn->kn_fflags |= NOTE_NOSRCADDR;
}
if (ev_hint & SO_FILT_HINT_IFDENIED) {
- if ((kn->kn_sfflags & NOTE_IFDENIED))
- kn->kn_fflags |= NOTE_IFDENIED;
+ kn->kn_fflags |= NOTE_IFDENIED;
}
if (ev_hint & SO_FILT_HINT_KEEPALIVE) {
- if (kn->kn_sfflags & NOTE_KEEPALIVE)
- kn->kn_fflags |= NOTE_KEEPALIVE;
+ kn->kn_fflags |= NOTE_KEEPALIVE;
}
if (ev_hint & SO_FILT_HINT_ADAPTIVE_WTIMO) {
- if (kn->kn_sfflags & NOTE_ADAPTIVE_WTIMO)
- kn->kn_fflags |= NOTE_ADAPTIVE_WTIMO;
+ kn->kn_fflags |= NOTE_ADAPTIVE_WTIMO;
}
if (ev_hint & SO_FILT_HINT_ADAPTIVE_RTIMO) {
- if (kn->kn_sfflags & NOTE_ADAPTIVE_RTIMO)
- kn->kn_fflags |= NOTE_ADAPTIVE_RTIMO;
+ kn->kn_fflags |= NOTE_ADAPTIVE_RTIMO;
}
- if (ev_hint & SO_FILT_HINT_CONNECTED) {
- if (kn->kn_sfflags & NOTE_CONNECTED)
- kn->kn_fflags |= NOTE_CONNECTED;
+ if ((ev_hint & SO_FILT_HINT_CONNECTED) ||
+ (so->so_state & SS_ISCONNECTED)) {
+ kn->kn_fflags |= NOTE_CONNECTED;
+ level_trigger |= NOTE_CONNECTED;
}
- if (ev_hint & SO_FILT_HINT_DISCONNECTED) {
- if (kn->kn_sfflags & NOTE_DISCONNECTED)
- kn->kn_fflags |= NOTE_DISCONNECTED;
+ if ((ev_hint & SO_FILT_HINT_DISCONNECTED) ||
+ (so->so_state & SS_ISDISCONNECTED)) {
+ kn->kn_fflags |= NOTE_DISCONNECTED;
+ level_trigger |= NOTE_DISCONNECTED;
}
if (ev_hint & SO_FILT_HINT_CONNINFO_UPDATED) {
if (so->so_proto != NULL &&
- (so->so_proto->pr_flags & PR_EVCONNINFO) &&
- (kn->kn_sfflags & NOTE_CONNINFO_UPDATED))
+ (so->so_proto->pr_flags & PR_EVCONNINFO))
kn->kn_fflags |= NOTE_CONNINFO_UPDATED;
}
- if ((kn->kn_sfflags & NOTE_READCLOSED) &&
- (so->so_state & SS_CANTRCVMORE)
+ if ((so->so_state & SS_CANTRCVMORE)
#if CONTENT_FILTER
- && cfil_sock_data_pending(&so->so_rcv) == 0
+ && cfil_sock_data_pending(&so->so_rcv) == 0
#endif /* CONTENT_FILTER */
- )
+ ) {
kn->kn_fflags |= NOTE_READCLOSED;
+ level_trigger |= NOTE_READCLOSED;
+ }
- if ((kn->kn_sfflags & NOTE_WRITECLOSED) &&
- (so->so_state & SS_CANTSENDMORE))
+ if (so->so_state & SS_CANTSENDMORE) {
kn->kn_fflags |= NOTE_WRITECLOSED;
+ level_trigger |= NOTE_WRITECLOSED;
+ }
- if ((kn->kn_sfflags & NOTE_SUSPEND) &&
- ((ev_hint & SO_FILT_HINT_SUSPEND) ||
- (so->so_flags & SOF_SUSPENDED))) {
+ if ((ev_hint & SO_FILT_HINT_SUSPEND) ||
+ (so->so_flags & SOF_SUSPENDED)) {
kn->kn_fflags &= ~(NOTE_SUSPEND | NOTE_RESUME);
+
+ /* If resume event was delivered before, reset it */
+ kn->kn_hookid &= ~NOTE_RESUME;
+
kn->kn_fflags |= NOTE_SUSPEND;
+ level_trigger |= NOTE_SUSPEND;
}
- if ((kn->kn_sfflags & NOTE_RESUME) &&
- ((ev_hint & SO_FILT_HINT_RESUME) ||
- (so->so_flags & SOF_SUSPENDED) == 0)) {
+ if ((ev_hint & SO_FILT_HINT_RESUME) ||
+ (so->so_flags & SOF_SUSPENDED) == 0) {
kn->kn_fflags &= ~(NOTE_SUSPEND | NOTE_RESUME);
+
+ /* If suspend event was delivered before, reset it */
+ kn->kn_hookid &= ~NOTE_SUSPEND;
+
kn->kn_fflags |= NOTE_RESUME;
+ level_trigger |= NOTE_RESUME;
}
if (so->so_error != 0) {
get_sockev_state(so, (u_int32_t *)&(kn->kn_data));
}
- if (kn->kn_fflags != 0)
+ /* Reset any events that are not requested on this knote */
+ kn->kn_fflags &= (kn->kn_sfflags & EVFILT_SOCK_ALL_MASK);
+ level_trigger &= (kn->kn_sfflags & EVFILT_SOCK_ALL_MASK);
+
+ /* Find the level triggerred events that are already delivered */
+ level_trigger &= kn->kn_hookid;
+ level_trigger &= EVFILT_SOCK_LEVEL_TRIGGER_MASK;
+
+ /* Do not deliver level triggerred events more than once */
+ if ((kn->kn_fflags & ~level_trigger) != 0)
ret = 1;
if (locked)
return (ret);
}
+static void
+filt_socktouch(struct knote *kn, struct kevent_internal_s *kev, long type)
+{
+#pragma unused(kev)
+ switch (type) {
+ case EVENT_REGISTER:
+ {
+ uint32_t changed_flags;
+ changed_flags = (kn->kn_sfflags ^ kn->kn_hookid);
+
+ /*
+ * Since we keep track of events that are already
+ * delivered, if any of those events are not requested
+ * anymore the state related to them can be reset
+ */
+ kn->kn_hookid &=
+ ~(changed_flags & EVFILT_SOCK_LEVEL_TRIGGER_MASK);
+ break;
+ }
+ case EVENT_PROCESS:
+ /*
+ * Store the state of the events being delivered. This
+ * state can be used to deliver level triggered events
+ * ateast once and still avoid waking up the application
+ * multiple times as long as the event is active.
+ */
+ if (kn->kn_fflags != 0)
+ kn->kn_hookid |= (kn->kn_fflags &
+ EVFILT_SOCK_LEVEL_TRIGGER_MASK);
+
+ /*
+ * NOTE_RESUME and NOTE_SUSPEND are an exception, deliver
+ * only one of them and remember the last one that was
+ * delivered last
+ */
+ if (kn->kn_fflags & NOTE_SUSPEND)
+ kn->kn_hookid &= ~NOTE_RESUME;
+ if (kn->kn_fflags & NOTE_RESUME)
+ kn->kn_hookid &= ~NOTE_SUSPEND;
+ break;
+ default:
+ break;
+ }
+}
+
void
get_sockev_state(struct socket *so, u_int32_t *statep)
{
SODEFUNCTLOG(("%s[%d]: (target pid %d level %d) "
"so 0x%llx [%d,%d] is not eligible for defunct "
"(%d)\n", __func__, proc_selfpid(), proc_pid(p),
- level, (uint64_t)VM_KERNEL_ADDRPERM(so),
+ level, (uint64_t)DEBUG_KERNEL_ADDRPERM(so),
SOCK_DOM(so), SOCK_TYPE(so), err));
return (err);
}
so->so_flags &= ~SOF_NODEFUNCT;
SODEFUNCTLOG(("%s[%d]: (target pid %d level %d) so 0x%llx "
"[%d,%d] defunct by force\n", __func__, proc_selfpid(),
- proc_pid(p), level, (uint64_t)VM_KERNEL_ADDRPERM(so),
+ proc_pid(p), level, (uint64_t)DEBUG_KERNEL_ADDRPERM(so),
SOCK_DOM(so), SOCK_TYPE(so)));
+ } else if (so->so_flags1 & SOF1_EXTEND_BK_IDLE_WANTED) {
+ struct inpcb *inp = (struct inpcb *)so->so_pcb;
+ struct ifnet *ifp = inp->inp_last_outifp;
+
+ if (ifp && IFNET_IS_CELLULAR(ifp)) {
+ OSIncrementAtomic(&soextbkidlestat.so_xbkidle_nocell);
+ } else if (so->so_flags & SOF_DELEGATED) {
+ OSIncrementAtomic(&soextbkidlestat.so_xbkidle_nodlgtd);
+ } else if (soextbkidlestat.so_xbkidle_time == 0) {
+ OSIncrementAtomic(&soextbkidlestat.so_xbkidle_notime);
+ } else if (noforce) {
+ OSIncrementAtomic(&soextbkidlestat.so_xbkidle_active);
+
+ so->so_flags1 |= SOF1_EXTEND_BK_IDLE_INPROG;
+ so->so_extended_bk_start = net_uptime();
+ OSBitOrAtomic(P_LXBKIDLEINPROG, &p->p_ladvflag);
+
+ inpcb_timer_sched(inp->inp_pcbinfo, INPCB_TIMER_LAZY);
+
+ err = EOPNOTSUPP;
+ SODEFUNCTLOG(("%s[%d]: (target pid %d level %d) "
+ "extend bk idle "
+ "so 0x%llx rcv hw %d cc %d\n",
+ __func__, proc_selfpid(), proc_pid(p),
+ level, (uint64_t)DEBUG_KERNEL_ADDRPERM(so),
+ so->so_rcv.sb_hiwat, so->so_rcv.sb_cc));
+ return (err);
+ } else {
+ OSIncrementAtomic(&soextbkidlestat.so_xbkidle_forced);
+ }
}
so->so_flags |= SOF_DEFUNCT;
done:
SODEFUNCTLOG(("%s[%d]: (target pid %d level %d) so 0x%llx [%d,%d] %s "
- "defunct\n", __func__, proc_selfpid(), proc_pid(p), level,
- (uint64_t)VM_KERNEL_ADDRPERM(so), SOCK_DOM(so), SOCK_TYPE(so),
- defunct ? "is already" : "marked as"));
+ "defunct%s\n", __func__, proc_selfpid(), proc_pid(p), level,
+ (uint64_t)DEBUG_KERNEL_ADDRPERM(so), SOCK_DOM(so), SOCK_TYPE(so),
+ defunct ? "is already" : "marked as",
+ (so->so_flags1 & SOF1_EXTEND_BK_IDLE_WANTED) ? " extbkidle" : ""));
return (err);
}
SODEFUNCTLOG(("%s[%d]: (target pid %d level %d) so 0x%llx [%s "
"%s:%d -> %s:%d] is now defunct [rcv_si 0x%x, snd_si 0x%x, "
"rcv_fl 0x%x, snd_fl 0x%x]\n", __func__, proc_selfpid(),
- proc_pid(p), level, (uint64_t)VM_KERNEL_ADDRPERM(so),
+ proc_pid(p), level, (uint64_t)DEBUG_KERNEL_ADDRPERM(so),
(SOCK_TYPE(so) == SOCK_STREAM) ? "TCP" : "UDP",
inet_ntop(SOCK_DOM(so), ((SOCK_DOM(so) == PF_INET) ?
(void *)&inp->inp_laddr.s_addr : (void *)&inp->in6p_laddr),
SODEFUNCTLOG(("%s[%d]: (target pid %d level %d) so 0x%llx "
"[%d,%d] is now defunct [rcv_si 0x%x, snd_si 0x%x, "
"rcv_fl 0x%x, snd_fl 0x%x]\n", __func__, proc_selfpid(),
- proc_pid(p), level, (uint64_t)VM_KERNEL_ADDRPERM(so),
+ proc_pid(p), level, (uint64_t)DEBUG_KERNEL_ADDRPERM(so),
SOCK_DOM(so), SOCK_TYPE(so), (uint32_t)rcv->sb_sel.si_flags,
(uint32_t)snd->sb_sel.si_flags, rcv->sb_flags,
snd->sb_flags));
return (0);
}
+int
+soresume(struct proc *p, struct socket *so, int locked)
+{
+ if (locked == 0)
+ socket_lock(so, 1);
+
+ if (so->so_flags1 & SOF1_EXTEND_BK_IDLE_INPROG) {
+ SODEFUNCTLOG(("%s[%d]: )target pid %d) so 0x%llx [%d,%d] "
+ "resumed from bk idle\n",
+ __func__, proc_selfpid(), proc_pid(p),
+ (uint64_t)DEBUG_KERNEL_ADDRPERM(so),
+ SOCK_DOM(so), SOCK_TYPE(so)));
+
+ so->so_flags1 &= ~SOF1_EXTEND_BK_IDLE_INPROG;
+ so->so_extended_bk_start = 0;
+ OSBitAndAtomic(~P_LXBKIDLEINPROG, &p->p_ladvflag);
+
+ OSIncrementAtomic(&soextbkidlestat.so_xbkidle_resumed);
+ OSDecrementAtomic(&soextbkidlestat.so_xbkidle_active);
+ VERIFY(soextbkidlestat.so_xbkidle_active >= 0);
+ }
+ if (locked == 0)
+ socket_unlock(so, 1);
+
+ return (0);
+}
+
+/*
+ * Does not attempt to account for sockets that are delegated from
+ * the current process
+ */
+int
+so_set_extended_bk_idle(struct socket *so, int optval)
+{
+ int error = 0;
+
+ if ((SOCK_DOM(so) != PF_INET && SOCK_DOM(so) != PF_INET6) ||
+ SOCK_PROTO(so) != IPPROTO_TCP) {
+ OSDecrementAtomic(&soextbkidlestat.so_xbkidle_notsupp);
+ error = EOPNOTSUPP;
+ } else if (optval == 0) {
+ so->so_flags1 &= ~SOF1_EXTEND_BK_IDLE_WANTED;
+
+ soresume(current_proc(), so, 1);
+ } else {
+ struct proc *p = current_proc();
+ int i;
+ struct filedesc *fdp;
+ int count = 0;
+
+ proc_fdlock(p);
+
+ fdp = p->p_fd;
+ for (i = 0; i < fdp->fd_nfiles; i++) {
+ struct fileproc *fp = fdp->fd_ofiles[i];
+ struct socket *so2;
+
+ if (fp == NULL ||
+ (fdp->fd_ofileflags[i] & UF_RESERVED) != 0 ||
+ FILEGLOB_DTYPE(fp->f_fglob) != DTYPE_SOCKET)
+ continue;
+
+ so2 = (struct socket *)fp->f_fglob->fg_data;
+ if (so != so2 &&
+ so2->so_flags1 & SOF1_EXTEND_BK_IDLE_WANTED)
+ count++;
+ if (count >= soextbkidlestat.so_xbkidle_maxperproc)
+ break;
+ }
+ if (count >= soextbkidlestat.so_xbkidle_maxperproc) {
+ OSIncrementAtomic(&soextbkidlestat.so_xbkidle_toomany);
+ error = EBUSY;
+ } else if (so->so_flags & SOF_DELEGATED) {
+ OSIncrementAtomic(&soextbkidlestat.so_xbkidle_nodlgtd);
+ error = EBUSY;
+ } else {
+ so->so_flags1 |= SOF1_EXTEND_BK_IDLE_WANTED;
+ OSIncrementAtomic(&soextbkidlestat.so_xbkidle_wantok);
+ }
+ SODEFUNCTLOG(("%s[%d]: so 0x%llx [%d,%d] "
+ "%s marked for extended bk idle\n",
+ __func__, proc_selfpid(),
+ (uint64_t)DEBUG_KERNEL_ADDRPERM(so),
+ SOCK_DOM(so), SOCK_TYPE(so),
+ (so->so_flags1 & SOF1_EXTEND_BK_IDLE_WANTED) ?
+ "is" : "not"));
+
+ proc_fdunlock(p);
+ }
+
+ return (error);
+}
+
+static void
+so_stop_extended_bk_idle(struct socket *so)
+{
+ so->so_flags1 &= ~SOF1_EXTEND_BK_IDLE_INPROG;
+ so->so_extended_bk_start = 0;
+
+ OSDecrementAtomic(&soextbkidlestat.so_xbkidle_active);
+ VERIFY(soextbkidlestat.so_xbkidle_active >= 0);
+ /*
+ * Force defunct
+ */
+ sosetdefunct(current_proc(), so,
+ SHUTDOWN_SOCKET_LEVEL_DISCONNECT_INTERNAL, FALSE);
+ if (so->so_flags & SOF_DEFUNCT) {
+ sodefunct(current_proc(), so,
+ SHUTDOWN_SOCKET_LEVEL_DISCONNECT_INTERNAL);
+ }
+}
+
+void
+so_drain_extended_bk_idle(struct socket *so)
+{
+ if (so && (so->so_flags1 & SOF1_EXTEND_BK_IDLE_INPROG)) {
+ /*
+ * Only penalize sockets that have outstanding data
+ */
+ if (so->so_rcv.sb_cc || so->so_snd.sb_cc) {
+ so_stop_extended_bk_idle(so);
+
+ OSIncrementAtomic(&soextbkidlestat.so_xbkidle_drained);
+ }
+ }
+}
+
+/*
+ * Return values tells if socket is still in extended background idle
+ */
+int
+so_check_extended_bk_idle_time(struct socket *so)
+{
+ int ret = 1;
+
+ if ((so->so_flags1 & SOF1_EXTEND_BK_IDLE_INPROG)) {
+ SODEFUNCTLOG(("%s[%d]: so 0x%llx [%d,%d]\n",
+ __func__, proc_selfpid(),
+ (uint64_t)DEBUG_KERNEL_ADDRPERM(so),
+ SOCK_DOM(so), SOCK_TYPE(so)));
+ if (net_uptime() - so->so_extended_bk_start >
+ soextbkidlestat.so_xbkidle_time) {
+ so_stop_extended_bk_idle(so);
+
+ OSIncrementAtomic(&soextbkidlestat.so_xbkidle_expired);
+
+ ret = 0;
+ } else {
+ struct inpcb *inp = (struct inpcb *)so->so_pcb;
+
+ inpcb_timer_sched(inp->inp_pcbinfo, INPCB_TIMER_LAZY);
+ OSIncrementAtomic(&soextbkidlestat.so_xbkidle_resched);
+ }
+ }
+
+ return (ret);
+}
+
+void
+resume_proc_sockets(proc_t p)
+{
+ if (p->p_ladvflag & P_LXBKIDLEINPROG) {
+ struct filedesc *fdp;
+ int i;
+
+ proc_fdlock(p);
+ fdp = p->p_fd;
+ for (i = 0; i < fdp->fd_nfiles; i++) {
+ struct fileproc *fp;
+ struct socket *so;
+
+ fp = fdp->fd_ofiles[i];
+ if (fp == NULL ||
+ (fdp->fd_ofileflags[i] & UF_RESERVED) != 0 ||
+ FILEGLOB_DTYPE(fp->f_fglob) != DTYPE_SOCKET)
+ continue;
+
+ so = (struct socket *)fp->f_fglob->fg_data;
+ (void) soresume(p, so, 0);
+ }
+ proc_fdunlock(p);
+
+ OSBitAndAtomic(~P_LXBKIDLEINPROG, &p->p_ladvflag);
+ }
+}
+
__private_extern__ int
so_set_recv_anyif(struct socket *so, int optval)
{
nocell_old = (so->so_restrictions & SO_RESTRICT_DENY_CELLULAR);
noexpensive_old = (so->so_restrictions & SO_RESTRICT_DENY_EXPENSIVE);
so->so_restrictions |= (vals & (SO_RESTRICT_DENY_IN |
- SO_RESTRICT_DENY_OUT | SO_RESTRICT_DENY_CELLULAR |
+ SO_RESTRICT_DENY_OUT | SO_RESTRICT_DENY_CELLULAR |
SO_RESTRICT_DENY_EXPENSIVE));
nocell_new = (so->so_restrictions & SO_RESTRICT_DENY_CELLULAR);
noexpensive_new = (so->so_restrictions & SO_RESTRICT_DENY_EXPENSIVE);
if (SOCK_DOM(so) == PF_INET) {
#endif /* !INET6 */
if (nocell_new - nocell_old != 0) {
- /* if deny cellular is now set, do what's needed for INPCB */
+ /*
+ * if deny cellular is now set, do what's needed
+ * for INPCB
+ */
inp_set_nocellular(sotoinpcb(so));
}
if (noexpensive_new - noexpensive_old != 0) {
so_get_restrictions(struct socket *so)
{
return (so->so_restrictions & (SO_RESTRICT_DENY_IN |
- SO_RESTRICT_DENY_OUT |
+ SO_RESTRICT_DENY_OUT |
SO_RESTRICT_DENY_CELLULAR | SO_RESTRICT_DENY_EXPENSIVE));
}
uuid_unparse(so->e_uuid, buf);
log(LOG_DEBUG, "%s[%s,%d]: so 0x%llx [%d,%d] epid %d (%s) "
"euuid %s%s\n", __func__, proc_name_address(p),
- proc_pid(p), (uint64_t)VM_KERNEL_ADDRPERM(so), SOCK_DOM(so),
- SOCK_TYPE(so), so->e_pid, proc_name_address(ep), buf,
+ proc_pid(p), (uint64_t)DEBUG_KERNEL_ADDRPERM(so),
+ SOCK_DOM(so), SOCK_TYPE(so),
+ so->e_pid, proc_name_address(ep), buf,
((so->so_flags & SOF_DELEGATED) ? " [delegated]" : ""));
} else if (error != 0 && net_io_policy_log) {
log(LOG_ERR, "%s[%s,%d]: so 0x%llx [%d,%d] epid %d (%s) "
"ERROR (%d)\n", __func__, proc_name_address(p),
- proc_pid(p), (uint64_t)VM_KERNEL_ADDRPERM(so), SOCK_DOM(so),
- SOCK_TYPE(so), epid, (ep == PROC_NULL) ? "PROC_NULL" :
+ proc_pid(p), (uint64_t)DEBUG_KERNEL_ADDRPERM(so),
+ SOCK_DOM(so), SOCK_TYPE(so),
+ epid, (ep == PROC_NULL) ? "PROC_NULL" :
proc_name_address(ep), error);
}
/* UUID must not be all-zeroes (reserved for kernel) */
if (uuid_is_null(euuid)) {
error = EINVAL;
- goto done;;
+ goto done;
}
/*
uuid_unparse(so->e_uuid, buf);
log(LOG_DEBUG, "%s[%s,%d]: so 0x%llx [%d,%d] epid %d "
"euuid %s%s\n", __func__, proc_name_address(p), proc_pid(p),
- (uint64_t)VM_KERNEL_ADDRPERM(so), SOCK_DOM(so),
+ (uint64_t)DEBUG_KERNEL_ADDRPERM(so), SOCK_DOM(so),
SOCK_TYPE(so), so->e_pid, buf,
((so->so_flags & SOF_DELEGATED) ? " [delegated]" : ""));
} else if (error != 0 && net_io_policy_log) {
uuid_unparse(euuid, buf);
log(LOG_DEBUG, "%s[%s,%d]: so 0x%llx [%d,%d] euuid %s "
"ERROR (%d)\n", __func__, proc_name_address(p), proc_pid(p),
- (uint64_t)VM_KERNEL_ADDRPERM(so), SOCK_DOM(so),
+ (uint64_t)DEBUG_KERNEL_ADDRPERM(so), SOCK_DOM(so),
SOCK_TYPE(so), buf, error);
}
}
void
-socket_post_kev_msg(uint32_t ev_code,
+socket_post_kev_msg(uint32_t ev_code,
struct kev_socket_event_data *ev_data,
uint32_t ev_datalen)
{
min(peersa->sa_len,
sizeof (ev.ev_data.kev_peername)));
socket_post_kev_msg(KEV_SOCKET_CLOSED,
- &ev.ev_data, sizeof (ev));
+ &ev.ev_data, sizeof (ev));
}
}
if (socksa != NULL)