X-Git-Url: https://git.saurik.com/apple/xnu.git/blobdiff_plain/fe8ab488e9161c46dd9885d58fc52996dc0249ff..ea3f04195ba4a5034c9c8e9b726d4f7ce96f1832:/bsd/netinet/in_pcb.c diff --git a/bsd/netinet/in_pcb.c b/bsd/netinet/in_pcb.c index e74dccbc3..d097b293f 100644 --- a/bsd/netinet/in_pcb.c +++ b/bsd/netinet/in_pcb.c @@ -1,5 +1,5 @@ /* - * Copyright (c) 2000-2014 Apple Inc. All rights reserved. + * Copyright (c) 2000-2019 Apple Inc. All rights reserved. * * @APPLE_OSREFERENCE_LICENSE_HEADER_START@ * @@ -78,6 +78,7 @@ #include #include #include +#include #include #include @@ -91,12 +92,15 @@ #include #include #include +#include #include +#include #include #include #include #include + #if INET6 #include #include @@ -108,35 +112,41 @@ #include #include +#include + #if NECP #include #endif -static lck_grp_t *inpcb_lock_grp; -static lck_attr_t *inpcb_lock_attr; -static lck_grp_attr_t *inpcb_lock_grp_attr; -decl_lck_mtx_data(static, inpcb_lock); /* global INPCB lock */ +#include +#include +#include + +#include + +extern const char *proc_name_address(struct proc *); + +static lck_grp_t *inpcb_lock_grp; +static lck_attr_t *inpcb_lock_attr; +static lck_grp_attr_t *inpcb_lock_grp_attr; +decl_lck_mtx_data(static, inpcb_lock); /* global INPCB lock */ decl_lck_mtx_data(static, inpcb_timeout_lock); static TAILQ_HEAD(, inpcbinfo) inpcb_head = TAILQ_HEAD_INITIALIZER(inpcb_head); -static u_int16_t inpcb_timeout_run = 0; /* INPCB timer is scheduled to run */ +static u_int16_t inpcb_timeout_run = 0; /* INPCB timer is scheduled to run */ static boolean_t inpcb_garbage_collecting = FALSE; /* gc timer is scheduled */ -static boolean_t inpcb_ticking = FALSE; /* "slow" timer is scheduled */ +static boolean_t inpcb_ticking = FALSE; /* "slow" timer is scheduled */ static boolean_t inpcb_fast_timer_on = FALSE; -/* - * If the total number of gc reqs is above a threshold, schedule - * garbage collect timer sooner - */ -static boolean_t inpcb_toomany_gcreq = FALSE; - -#define INPCB_GCREQ_THRESHOLD 50000 -#define INPCB_TOOMANY_GCREQ_TIMER (hz/10) /* 10 times a second */ +#define INPCB_GCREQ_THRESHOLD 50000 -static void inpcb_sched_timeout(struct timeval *); -static void inpcb_timeout(void *); -int inpcb_timeout_lazy = 10; /* 10 seconds leeway for lazy timers */ +static thread_call_t inpcb_thread_call, inpcb_fast_thread_call; +static void inpcb_sched_timeout(void); +static void inpcb_sched_lazy_timeout(void); +static void _inpcb_sched_timeout(unsigned int); +static void inpcb_timeout(void *, void *); +const int inpcb_timeout_lazy = 10; /* 10 seconds leeway for lazy timers */ extern int tvtohz(struct timeval *); #if CONFIG_PROC_UUID_POLICY @@ -146,21 +156,21 @@ static void inp_update_necp_want_app_policy(struct inpcb *, boolean_t); #endif /* NECP */ #endif /* !CONFIG_PROC_UUID_POLICY */ -#define DBG_FNC_PCB_LOOKUP NETDBG_CODE(DBG_NETTCP, (6 << 8)) -#define DBG_FNC_PCB_HLOOKUP NETDBG_CODE(DBG_NETTCP, ((6 << 8) | 1)) +#define DBG_FNC_PCB_LOOKUP NETDBG_CODE(DBG_NETTCP, (6 << 8)) +#define DBG_FNC_PCB_HLOOKUP NETDBG_CODE(DBG_NETTCP, ((6 << 8) | 1)) /* * These configure the range of local port addresses assigned to * "unspecified" outgoing connections/packets/whatever. */ -int ipport_lowfirstauto = IPPORT_RESERVED - 1; /* 1023 */ -int ipport_lowlastauto = IPPORT_RESERVEDSTART; /* 600 */ -int ipport_firstauto = IPPORT_HIFIRSTAUTO; /* 49152 */ -int ipport_lastauto = IPPORT_HILASTAUTO; /* 65535 */ -int ipport_hifirstauto = IPPORT_HIFIRSTAUTO; /* 49152 */ -int ipport_hilastauto = IPPORT_HILASTAUTO; /* 65535 */ - -#define RANGECHK(var, min, max) \ +int ipport_lowfirstauto = IPPORT_RESERVED - 1; /* 1023 */ +int ipport_lowlastauto = IPPORT_RESERVEDSTART; /* 600 */ +int ipport_firstauto = IPPORT_HIFIRSTAUTO; /* 49152 */ +int ipport_lastauto = IPPORT_HILASTAUTO; /* 65535 */ +int ipport_hifirstauto = IPPORT_HIFIRSTAUTO; /* 49152 */ +int ipport_hilastauto = IPPORT_HILASTAUTO; /* 65535 */ + +#define RANGECHK(var, min, max) \ if ((var) < (min)) { (var) = (min); } \ else if ((var) > (max)) { (var) = (max); } @@ -169,6 +179,20 @@ sysctl_net_ipport_check SYSCTL_HANDLER_ARGS { #pragma unused(arg1, arg2) int error; +#if (DEBUG | DEVELOPMENT) + int old_value = *(int *)oidp->oid_arg1; + /* + * For unit testing allow a non-superuser process with the + * proper entitlement to modify the variables + */ + if (req->newptr) { + if (proc_suser(current_proc()) != 0 && + (error = priv_check_cred(kauth_cred_get(), + PRIV_NETINET_RESERVEDPORT, 0))) { + return EPERM; + } + } +#endif /* (DEBUG | DEVELOPMENT) */ error = sysctl_handle_int(oidp, oidp->oid_arg1, oidp->oid_arg2, req); if (!error) { @@ -179,56 +203,86 @@ sysctl_net_ipport_check SYSCTL_HANDLER_ARGS RANGECHK(ipport_hifirstauto, IPPORT_RESERVED, USHRT_MAX); RANGECHK(ipport_hilastauto, IPPORT_RESERVED, USHRT_MAX); } - return (error); + +#if (DEBUG | DEVELOPMENT) + os_log(OS_LOG_DEFAULT, + "%s:%u sysctl net.restricted_port.verbose: %d -> %d)", + proc_best_name(current_proc()), proc_selfpid(), + old_value, *(int *)oidp->oid_arg1); +#endif /* (DEBUG | DEVELOPMENT) */ + + return error; } #undef RANGECHK SYSCTL_NODE(_net_inet_ip, IPPROTO_IP, portrange, - CTLFLAG_RW|CTLFLAG_LOCKED, 0, "IP Ports"); + CTLFLAG_RW | CTLFLAG_LOCKED, 0, "IP Ports"); + +#if (DEBUG | DEVELOPMENT) +#define CTLFAGS_IP_PORTRANGE (CTLTYPE_INT | CTLFLAG_RW | CTLFLAG_LOCKED | CTLFLAG_ANYBODY) +#else +#define CTLFAGS_IP_PORTRANGE (CTLTYPE_INT | CTLFLAG_RW | CTLFLAG_LOCKED) +#endif /* (DEBUG | DEVELOPMENT) */ SYSCTL_PROC(_net_inet_ip_portrange, OID_AUTO, lowfirst, - CTLTYPE_INT|CTLFLAG_RW | CTLFLAG_LOCKED, - &ipport_lowfirstauto, 0, &sysctl_net_ipport_check, "I", ""); + CTLFAGS_IP_PORTRANGE, + &ipport_lowfirstauto, 0, &sysctl_net_ipport_check, "I", ""); SYSCTL_PROC(_net_inet_ip_portrange, OID_AUTO, lowlast, - CTLTYPE_INT|CTLFLAG_RW | CTLFLAG_LOCKED, - &ipport_lowlastauto, 0, &sysctl_net_ipport_check, "I", ""); + CTLFAGS_IP_PORTRANGE, + &ipport_lowlastauto, 0, &sysctl_net_ipport_check, "I", ""); SYSCTL_PROC(_net_inet_ip_portrange, OID_AUTO, first, - CTLTYPE_INT|CTLFLAG_RW | CTLFLAG_LOCKED, - &ipport_firstauto, 0, &sysctl_net_ipport_check, "I", ""); + CTLFAGS_IP_PORTRANGE, + &ipport_firstauto, 0, &sysctl_net_ipport_check, "I", ""); SYSCTL_PROC(_net_inet_ip_portrange, OID_AUTO, last, - CTLTYPE_INT|CTLFLAG_RW | CTLFLAG_LOCKED, - &ipport_lastauto, 0, &sysctl_net_ipport_check, "I", ""); + CTLFAGS_IP_PORTRANGE, + &ipport_lastauto, 0, &sysctl_net_ipport_check, "I", ""); SYSCTL_PROC(_net_inet_ip_portrange, OID_AUTO, hifirst, - CTLTYPE_INT|CTLFLAG_RW | CTLFLAG_LOCKED, - &ipport_hifirstauto, 0, &sysctl_net_ipport_check, "I", ""); + CTLFAGS_IP_PORTRANGE, + &ipport_hifirstauto, 0, &sysctl_net_ipport_check, "I", ""); SYSCTL_PROC(_net_inet_ip_portrange, OID_AUTO, hilast, - CTLTYPE_INT|CTLFLAG_RW | CTLFLAG_LOCKED, - &ipport_hilastauto, 0, &sysctl_net_ipport_check, "I", ""); + CTLFAGS_IP_PORTRANGE, + &ipport_hilastauto, 0, &sysctl_net_ipport_check, "I", ""); + +static uint32_t apn_fallbk_debug = 0; +#define apn_fallbk_log(x) do { if (apn_fallbk_debug >= 1) log x; } while (0) + +#if CONFIG_EMBEDDED +static boolean_t apn_fallbk_enabled = TRUE; + +SYSCTL_DECL(_net_inet); +SYSCTL_NODE(_net_inet, OID_AUTO, apn_fallback, CTLFLAG_RW | CTLFLAG_LOCKED, 0, "APN Fallback"); +SYSCTL_UINT(_net_inet_apn_fallback, OID_AUTO, enable, CTLFLAG_RW | CTLFLAG_LOCKED, + &apn_fallbk_enabled, 0, "APN fallback enable"); +SYSCTL_UINT(_net_inet_apn_fallback, OID_AUTO, debug, CTLFLAG_RW | CTLFLAG_LOCKED, + &apn_fallbk_debug, 0, "APN fallback debug enable"); +#else +static boolean_t apn_fallbk_enabled = FALSE; +#endif -extern int udp_use_randomport; -extern int tcp_use_randomport; +extern int udp_use_randomport; +extern int tcp_use_randomport; /* Structs used for flowhash computation */ struct inp_flowhash_key_addr { union { - struct in_addr v4; + struct in_addr v4; struct in6_addr v6; - u_int8_t addr8[16]; - u_int16_t addr16[8]; - u_int32_t addr32[4]; + u_int8_t addr8[16]; + u_int16_t addr16[8]; + u_int32_t addr32[4]; } infha; }; struct inp_flowhash_key { - struct inp_flowhash_key_addr infh_laddr; - struct inp_flowhash_key_addr infh_faddr; - u_int32_t infh_lport; - u_int32_t infh_fport; - u_int32_t infh_af; - u_int32_t infh_proto; - u_int32_t infh_rand1; - u_int32_t infh_rand2; + struct inp_flowhash_key_addr infh_laddr; + struct inp_flowhash_key_addr infh_faddr; + u_int32_t infh_lport; + u_int32_t infh_fport; + u_int32_t infh_af; + u_int32_t infh_proto; + u_int32_t infh_rand1; + u_int32_t infh_rand2; }; static u_int32_t inp_hash_seed = 0; @@ -236,8 +290,8 @@ static u_int32_t inp_hash_seed = 0; static int infc_cmp(const struct inpcb *, const struct inpcb *); /* Flags used by inp_fc_getinp */ -#define INPFC_SOLOCKED 0x1 -#define INPFC_REMOVE 0x2 +#define INPFC_SOLOCKED 0x1 +#define INPFC_REMOVE 0x2 static struct inpcb *inp_fc_getinp(u_int32_t, u_int32_t); static void inp_fc_feedback(struct inpcb *); @@ -272,6 +326,13 @@ in_pcbinit(void) inpcb_lock_attr = lck_attr_alloc_init(); lck_mtx_init(&inpcb_lock, inpcb_lock_grp, inpcb_lock_attr); lck_mtx_init(&inpcb_timeout_lock, inpcb_lock_grp, inpcb_lock_attr); + inpcb_thread_call = thread_call_allocate_with_priority(inpcb_timeout, + NULL, THREAD_CALL_PRIORITY_KERNEL); + inpcb_fast_thread_call = thread_call_allocate_with_priority( + inpcb_timeout, NULL, THREAD_CALL_PRIORITY_KERNEL); + if (inpcb_thread_call == NULL || inpcb_fast_thread_call == NULL) { + panic("unable to alloc the inpcb thread call"); + } /* * Initialize data structures required to deliver @@ -284,22 +345,15 @@ in_pcbinit(void) lck_mtx_unlock(&inp_fc_lck); } -#define INPCB_HAVE_TIMER_REQ(req) (((req).intimer_lazy > 0) || \ +#define INPCB_HAVE_TIMER_REQ(req) (((req).intimer_lazy > 0) || \ ((req).intimer_fast > 0) || ((req).intimer_nodelay > 0)) static void -inpcb_timeout(void *arg) +inpcb_timeout(void *arg0, void *arg1) { -#pragma unused(arg) +#pragma unused(arg0, arg1) struct inpcbinfo *ipi; boolean_t t, gc; struct intimercount gccnt, tmcnt; - struct timeval leeway; - boolean_t toomany_gc = FALSE; - - if (arg != NULL) { - VERIFY(arg == &inpcb_toomany_gcreq); - toomany_gc = *(boolean_t *)arg; - } /* * Update coarse-grained networking timestamp (in sec.); the idea @@ -325,7 +379,7 @@ inpcb_timeout(void *arg) TAILQ_FOREACH(ipi, &inpcb_head, ipi_entry) { if (INPCB_HAVE_TIMER_REQ(ipi->ipi_gc_req)) { bzero(&ipi->ipi_gc_req, - sizeof(ipi->ipi_gc_req)); + sizeof(ipi->ipi_gc_req)); if (gc && ipi->ipi_gc != NULL) { ipi->ipi_gc(ipi); gccnt.intimer_lazy += @@ -338,12 +392,12 @@ inpcb_timeout(void *arg) } if (INPCB_HAVE_TIMER_REQ(ipi->ipi_timer_req)) { bzero(&ipi->ipi_timer_req, - sizeof(ipi->ipi_timer_req)); + sizeof(ipi->ipi_timer_req)); if (t && ipi->ipi_timer != NULL) { ipi->ipi_timer(ipi); tmcnt.intimer_lazy += ipi->ipi_timer_req.intimer_lazy; - tmcnt.intimer_lazy += + tmcnt.intimer_fast += ipi->ipi_timer_req.intimer_fast; tmcnt.intimer_nodelay += ipi->ipi_timer_req.intimer_nodelay; @@ -355,101 +409,104 @@ inpcb_timeout(void *arg) } /* lock was dropped above, so check first before overriding */ - if (!inpcb_garbage_collecting) + if (!inpcb_garbage_collecting) { inpcb_garbage_collecting = INPCB_HAVE_TIMER_REQ(gccnt); - if (!inpcb_ticking) + } + if (!inpcb_ticking) { inpcb_ticking = INPCB_HAVE_TIMER_REQ(tmcnt); + } /* re-arm the timer if there's work to do */ - if (toomany_gc) { - inpcb_toomany_gcreq = FALSE; - } else { - inpcb_timeout_run--; - VERIFY(inpcb_timeout_run >= 0 && inpcb_timeout_run < 2); - } + inpcb_timeout_run--; + VERIFY(inpcb_timeout_run >= 0 && inpcb_timeout_run < 2); - bzero(&leeway, sizeof(leeway)); - leeway.tv_sec = inpcb_timeout_lazy; - if (gccnt.intimer_nodelay > 0 || tmcnt.intimer_nodelay > 0) - inpcb_sched_timeout(NULL); - else if ((gccnt.intimer_fast + tmcnt.intimer_fast) <= 5) + if (gccnt.intimer_nodelay > 0 || tmcnt.intimer_nodelay > 0) { + inpcb_sched_timeout(); + } else if ((gccnt.intimer_fast + tmcnt.intimer_fast) <= 5) { /* be lazy when idle with little activity */ - inpcb_sched_timeout(&leeway); - else - inpcb_sched_timeout(NULL); + inpcb_sched_lazy_timeout(); + } else { + inpcb_sched_timeout(); + } lck_mtx_unlock(&inpcb_timeout_lock); } static void -inpcb_sched_timeout(struct timeval *leeway) +inpcb_sched_timeout(void) +{ + _inpcb_sched_timeout(0); +} + +static void +inpcb_sched_lazy_timeout(void) +{ + _inpcb_sched_timeout(inpcb_timeout_lazy); +} + +static void +_inpcb_sched_timeout(unsigned int offset) { - lck_mtx_assert(&inpcb_timeout_lock, LCK_MTX_ASSERT_OWNED); + uint64_t deadline, leeway; + clock_interval_to_deadline(1, NSEC_PER_SEC, &deadline); + LCK_MTX_ASSERT(&inpcb_timeout_lock, LCK_MTX_ASSERT_OWNED); if (inpcb_timeout_run == 0 && - (inpcb_garbage_collecting || inpcb_ticking)) { + (inpcb_garbage_collecting || inpcb_ticking)) { lck_mtx_convert_spin(&inpcb_timeout_lock); inpcb_timeout_run++; - if (leeway == NULL) { + if (offset == 0) { inpcb_fast_timer_on = TRUE; - timeout(inpcb_timeout, NULL, hz); + thread_call_enter_delayed(inpcb_thread_call, + deadline); } else { inpcb_fast_timer_on = FALSE; - timeout_with_leeway(inpcb_timeout, NULL, hz, - tvtohz(leeway)); + clock_interval_to_absolutetime_interval(offset, + NSEC_PER_SEC, &leeway); + thread_call_enter_delayed_with_leeway( + inpcb_thread_call, NULL, deadline, leeway, + THREAD_CALL_DELAY_LEEWAY); } } else if (inpcb_timeout_run == 1 && - leeway == NULL && !inpcb_fast_timer_on) { + offset == 0 && !inpcb_fast_timer_on) { /* * Since the request was for a fast timer but the * scheduled timer is a lazy timer, try to schedule - * another instance of fast timer also + * another instance of fast timer also. */ lck_mtx_convert_spin(&inpcb_timeout_lock); inpcb_timeout_run++; inpcb_fast_timer_on = TRUE; - timeout(inpcb_timeout, NULL, hz); + thread_call_enter_delayed(inpcb_fast_thread_call, deadline); } } void inpcb_gc_sched(struct inpcbinfo *ipi, u_int32_t type) { - struct timeval leeway; u_int32_t gccnt; + lck_mtx_lock_spin(&inpcb_timeout_lock); inpcb_garbage_collecting = TRUE; - gccnt = ipi->ipi_gc_req.intimer_nodelay + - ipi->ipi_gc_req.intimer_fast; + ipi->ipi_gc_req.intimer_fast; - if (gccnt > INPCB_GCREQ_THRESHOLD && !inpcb_toomany_gcreq) { - inpcb_toomany_gcreq = TRUE; - - /* - * There are toomany pcbs waiting to be garbage collected, - * schedule a much faster timeout in addition to - * the caller's request - */ - lck_mtx_convert_spin(&inpcb_timeout_lock); - timeout(inpcb_timeout, (void *)&inpcb_toomany_gcreq, - INPCB_TOOMANY_GCREQ_TIMER); + if (gccnt > INPCB_GCREQ_THRESHOLD) { + type = INPCB_TIMER_FAST; } switch (type) { case INPCB_TIMER_NODELAY: atomic_add_32(&ipi->ipi_gc_req.intimer_nodelay, 1); - inpcb_sched_timeout(NULL); + inpcb_sched_timeout(); break; case INPCB_TIMER_FAST: atomic_add_32(&ipi->ipi_gc_req.intimer_fast, 1); - inpcb_sched_timeout(NULL); + inpcb_sched_timeout(); break; default: atomic_add_32(&ipi->ipi_gc_req.intimer_lazy, 1); - leeway.tv_sec = inpcb_timeout_lazy; - leeway.tv_usec = 0; - inpcb_sched_timeout(&leeway); + inpcb_sched_lazy_timeout(); break; } lck_mtx_unlock(&inpcb_timeout_lock); @@ -458,23 +515,20 @@ inpcb_gc_sched(struct inpcbinfo *ipi, u_int32_t type) void inpcb_timer_sched(struct inpcbinfo *ipi, u_int32_t type) { - struct timeval leeway; lck_mtx_lock_spin(&inpcb_timeout_lock); inpcb_ticking = TRUE; switch (type) { case INPCB_TIMER_NODELAY: atomic_add_32(&ipi->ipi_timer_req.intimer_nodelay, 1); - inpcb_sched_timeout(NULL); + inpcb_sched_timeout(); break; case INPCB_TIMER_FAST: atomic_add_32(&ipi->ipi_timer_req.intimer_fast, 1); - inpcb_sched_timeout(NULL); + inpcb_sched_timeout(); break; default: atomic_add_32(&ipi->ipi_timer_req.intimer_lazy, 1); - leeway.tv_sec = inpcb_timeout_lazy; - leeway.tv_usec = 0; - inpcb_sched_timeout(&leeway); + inpcb_sched_lazy_timeout(); break; } lck_mtx_unlock(&inpcb_timeout_lock); @@ -505,16 +559,18 @@ in_pcbinfo_detach(struct inpcbinfo *ipi) lck_mtx_lock(&inpcb_lock); TAILQ_FOREACH(ipi0, &inpcb_head, ipi_entry) { - if (ipi0 == ipi) + if (ipi0 == ipi) { break; + } } - if (ipi0 != NULL) + if (ipi0 != NULL) { TAILQ_REMOVE(&inpcb_head, ipi0, ipi_entry); - else + } else { error = ENXIO; + } lck_mtx_unlock(&inpcb_lock); - return (error); + return error; } /* @@ -529,20 +585,21 @@ in_pcballoc(struct socket *so, struct inpcbinfo *pcbinfo, struct proc *p) { #pragma unused(p) struct inpcb *inp; - caddr_t temp; + caddr_t temp; #if CONFIG_MACF_NET int mac_error; #endif /* CONFIG_MACF_NET */ - if (!so->cached_in_sock_layer) { + if ((so->so_flags1 & SOF1_CACHED_IN_SOCK_LAYER) == 0) { inp = (struct inpcb *)zalloc(pcbinfo->ipi_zone); - if (inp == NULL) - return (ENOBUFS); - bzero((caddr_t)inp, sizeof (*inp)); + if (inp == NULL) { + return ENOBUFS; + } + bzero((caddr_t)inp, sizeof(*inp)); } else { inp = (struct inpcb *)(void *)so->so_saved_pcb; temp = inp->inp_saved_ppcb; - bzero((caddr_t)inp, sizeof (*inp)); + bzero((caddr_t)inp, sizeof(*inp)); inp->inp_saved_ppcb = temp; } @@ -552,48 +609,49 @@ in_pcballoc(struct socket *so, struct inpcbinfo *pcbinfo, struct proc *p) #if CONFIG_MACF_NET mac_error = mac_inpcb_label_init(inp, M_WAITOK); if (mac_error != 0) { - if (!so->cached_in_sock_layer) + if ((so->so_flags1 & SOF1_CACHED_IN_SOCK_LAYER) == 0) { zfree(pcbinfo->ipi_zone, inp); - return (mac_error); + } + return mac_error; } mac_inpcb_label_associate(so, inp); #endif /* CONFIG_MACF_NET */ /* make sure inp_stat is always 64-bit aligned */ inp->inp_stat = (struct inp_stat *)P2ROUNDUP(inp->inp_stat_store, - sizeof (u_int64_t)); + sizeof(u_int64_t)); if (((uintptr_t)inp->inp_stat - (uintptr_t)inp->inp_stat_store) + - sizeof (*inp->inp_stat) > sizeof (inp->inp_stat_store)) { + sizeof(*inp->inp_stat) > sizeof(inp->inp_stat_store)) { panic("%s: insufficient space to align inp_stat", __func__); /* NOTREACHED */ } /* make sure inp_cstat is always 64-bit aligned */ inp->inp_cstat = (struct inp_stat *)P2ROUNDUP(inp->inp_cstat_store, - sizeof (u_int64_t)); + sizeof(u_int64_t)); if (((uintptr_t)inp->inp_cstat - (uintptr_t)inp->inp_cstat_store) + - sizeof (*inp->inp_cstat) > sizeof (inp->inp_cstat_store)) { + sizeof(*inp->inp_cstat) > sizeof(inp->inp_cstat_store)) { panic("%s: insufficient space to align inp_cstat", __func__); /* NOTREACHED */ } /* make sure inp_wstat is always 64-bit aligned */ inp->inp_wstat = (struct inp_stat *)P2ROUNDUP(inp->inp_wstat_store, - sizeof (u_int64_t)); + sizeof(u_int64_t)); if (((uintptr_t)inp->inp_wstat - (uintptr_t)inp->inp_wstat_store) + - sizeof (*inp->inp_wstat) > sizeof (inp->inp_wstat_store)) { + sizeof(*inp->inp_wstat) > sizeof(inp->inp_wstat_store)) { panic("%s: insufficient space to align inp_wstat", __func__); /* NOTREACHED */ } /* make sure inp_Wstat is always 64-bit aligned */ inp->inp_Wstat = (struct inp_stat *)P2ROUNDUP(inp->inp_Wstat_store, - sizeof (u_int64_t)); + sizeof(u_int64_t)); if (((uintptr_t)inp->inp_Wstat - (uintptr_t)inp->inp_Wstat_store) + - sizeof (*inp->inp_Wstat) > sizeof (inp->inp_Wstat_store)) { + sizeof(*inp->inp_Wstat) > sizeof(inp->inp_Wstat_store)) { panic("%s: insufficient space to align inp_Wstat", __func__); /* NOTREACHED */ } - + so->so_pcb = (caddr_t)inp; if (so->so_proto->pr_flags & PR_PCBLOCK) { @@ -602,12 +660,17 @@ in_pcballoc(struct socket *so, struct inpcbinfo *pcbinfo, struct proc *p) } #if INET6 - if (SOCK_DOM(so) == PF_INET6 && !ip6_mapped_addr_on) + if (SOCK_DOM(so) == PF_INET6 && !ip6_mapped_addr_on) { inp->inp_flags |= IN6P_IPV6_V6ONLY; + } - if (ip6_auto_flowlabel) + if (ip6_auto_flowlabel) { inp->inp_flags |= IN6P_AUTOFLOWLABEL; + } #endif /* INET6 */ + if (intcoproc_unrestricted) { + inp->inp_flags2 |= INP2_INTCOPROC_ALLOWED; + } (void) inp_update_policy(inp); @@ -616,14 +679,14 @@ in_pcballoc(struct socket *so, struct inpcbinfo *pcbinfo, struct proc *p) LIST_INSERT_HEAD(pcbinfo->ipi_listhead, inp, inp_list); pcbinfo->ipi_count++; lck_rw_done(pcbinfo->ipi_lock); - return (0); + return 0; } /* * in_pcblookup_local_and_cleanup does everything * in_pcblookup_local does but it checks for a socket * that's going away. Since we know that the lock is - * held read+write when this funciton is called, we + * held read+write when this function is called, we * can safely dispose of this socket like the slow * timer would usually do and return NULL. This is * great for bind. @@ -641,19 +704,20 @@ in_pcblookup_local_and_cleanup(struct inpcbinfo *pcbinfo, struct in_addr laddr, if (inp != NULL && inp->inp_wantcnt == WNT_STOPUSING) { struct socket *so = inp->inp_socket; - lck_mtx_lock(&inp->inpcb_mtx); + socket_lock(so, 0); if (so->so_usecount == 0) { - if (inp->inp_state != INPCB_STATE_DEAD) + if (inp->inp_state != INPCB_STATE_DEAD) { in_pcbdetach(inp); - in_pcbdispose(inp); /* will unlock & destroy */ + } + in_pcbdispose(inp); /* will unlock & destroy */ inp = NULL; } else { - lck_mtx_unlock(&inp->inpcb_mtx); + socket_unlock(so, 0); } } - return (inp); + return inp; } static void @@ -665,20 +729,20 @@ in_pcb_conflict_post_msg(u_int16_t port) * who has set SOF_NOTIFYCONFLICT owns. */ struct kev_msg ev_msg; - struct kev_in_portinuse in_portinuse; + struct kev_in_portinuse in_portinuse; - bzero(&in_portinuse, sizeof (struct kev_in_portinuse)); - bzero(&ev_msg, sizeof (struct kev_msg)); - in_portinuse.port = ntohs(port); /* port in host order */ + bzero(&in_portinuse, sizeof(struct kev_in_portinuse)); + bzero(&ev_msg, sizeof(struct kev_msg)); + in_portinuse.port = ntohs(port); /* port in host order */ in_portinuse.req_pid = proc_selfpid(); ev_msg.vendor_code = KEV_VENDOR_APPLE; ev_msg.kev_class = KEV_NETWORK_CLASS; ev_msg.kev_subclass = KEV_INET_SUBCLASS; ev_msg.event_code = KEV_INET_PORTINUSE; ev_msg.dv[0].data_ptr = &in_portinuse; - ev_msg.dv[0].data_length = sizeof (struct kev_in_portinuse); + ev_msg.dv[0].data_length = sizeof(struct kev_in_portinuse); ev_msg.dv[1].data_length = 0; - kev_post_msg(&ev_msg); + dlil_post_complete_msg(NULL, &ev_msg); } /* @@ -708,23 +772,29 @@ in_pcbbind(struct inpcb *inp, struct sockaddr *nam, struct proc *p) struct in_addr laddr; struct ifnet *outif = NULL; - if (TAILQ_EMPTY(&in_ifaddrhead)) /* XXX broken! */ - return (EADDRNOTAVAIL); - if (inp->inp_lport != 0 || inp->inp_laddr.s_addr != INADDR_ANY) - return (EINVAL); - if (!(so->so_options & (SO_REUSEADDR|SO_REUSEPORT))) + if (TAILQ_EMPTY(&in_ifaddrhead)) { /* XXX broken! */ + return EADDRNOTAVAIL; + } + if (!(so->so_options & (SO_REUSEADDR | SO_REUSEPORT))) { wild = 1; - socket_unlock(so, 0); /* keep reference on socket */ - lck_rw_lock_exclusive(pcbinfo->ipi_lock); + } bzero(&laddr, sizeof(laddr)); - if (nam != NULL) { + socket_unlock(so, 0); /* keep reference on socket */ + lck_rw_lock_exclusive(pcbinfo->ipi_lock); + if (inp->inp_lport != 0 || inp->inp_laddr.s_addr != INADDR_ANY) { + /* another thread completed the bind */ + lck_rw_done(pcbinfo->ipi_lock); + socket_lock(so, 0); + return EINVAL; + } - if (nam->sa_len != sizeof (struct sockaddr_in)) { + if (nam != NULL) { + if (nam->sa_len != sizeof(struct sockaddr_in)) { lck_rw_done(pcbinfo->ipi_lock); socket_lock(so, 0); - return (EINVAL); + return EINVAL; } #if 0 /* @@ -734,7 +804,7 @@ in_pcbbind(struct inpcb *inp, struct sockaddr *nam, struct proc *p) if (nam->sa_family != AF_INET) { lck_rw_done(pcbinfo->ipi_lock); socket_lock(so, 0); - return (EAFNOSUPPORT); + return EAFNOSUPPORT; } #endif /* 0 */ lport = SIN(nam)->sin_port; @@ -747,23 +817,24 @@ in_pcbbind(struct inpcb *inp, struct sockaddr *nam, struct proc *p) * and a multicast address is bound on both * new and duplicated sockets. */ - if (so->so_options & SO_REUSEADDR) - reuseport = SO_REUSEADDR|SO_REUSEPORT; + if (so->so_options & SO_REUSEADDR) { + reuseport = SO_REUSEADDR | SO_REUSEPORT; + } } else if (SIN(nam)->sin_addr.s_addr != INADDR_ANY) { struct sockaddr_in sin; struct ifaddr *ifa; /* Sanitized for interface address searches */ - bzero(&sin, sizeof (sin)); + bzero(&sin, sizeof(sin)); sin.sin_family = AF_INET; - sin.sin_len = sizeof (struct sockaddr_in); + sin.sin_len = sizeof(struct sockaddr_in); sin.sin_addr.s_addr = SIN(nam)->sin_addr.s_addr; ifa = ifa_ifwithaddr(SA(&sin)); if (ifa == NULL) { lck_rw_done(pcbinfo->ipi_lock); socket_lock(so, 0); - return (EADDRNOTAVAIL); + return EADDRNOTAVAIL; } else { /* * Opportunistically determine the outbound @@ -779,11 +850,16 @@ in_pcbbind(struct inpcb *inp, struct sockaddr *nam, struct proc *p) IFA_REMREF(ifa); } } + + if (lport != 0) { struct inpcb *t; uid_t u; - if (ntohs(lport) < IPPORT_RESERVED) { +#if !CONFIG_EMBEDDED + if (ntohs(lport) < IPPORT_RESERVED && + SIN(nam)->sin_addr.s_addr != 0 && + !(inp->inp_flags2 & INP2_EXTERNAL_PORT)) { cred = kauth_cred_proc_ref(p); error = priv_check_cred(cred, PRIV_NETINET_RESERVEDPORT, 0); @@ -791,38 +867,57 @@ in_pcbbind(struct inpcb *inp, struct sockaddr *nam, struct proc *p) if (error != 0) { lck_rw_done(pcbinfo->ipi_lock); socket_lock(so, 0); - return (EACCES); + return EACCES; } } +#endif /* !CONFIG_EMBEDDED */ + /* + * Check wether the process is allowed to bind to a restricted port + */ + if (!current_task_can_use_restricted_in_port(lport, + so->so_proto->pr_protocol, PORT_FLAGS_BSD)) { + lck_rw_done(pcbinfo->ipi_lock); + socket_lock(so, 0); + return EADDRINUSE; + } + if (!IN_MULTICAST(ntohl(SIN(nam)->sin_addr.s_addr)) && (u = kauth_cred_getuid(so->so_cred)) != 0 && (t = in_pcblookup_local_and_cleanup( - inp->inp_pcbinfo, SIN(nam)->sin_addr, lport, - INPLOOKUP_WILDCARD)) != NULL && + inp->inp_pcbinfo, SIN(nam)->sin_addr, lport, + INPLOOKUP_WILDCARD)) != NULL && (SIN(nam)->sin_addr.s_addr != INADDR_ANY || t->inp_laddr.s_addr != INADDR_ANY || !(t->inp_socket->so_options & SO_REUSEPORT)) && (u != kauth_cred_getuid(t->inp_socket->so_cred)) && !(t->inp_socket->so_flags & SOF_REUSESHAREUID) && (SIN(nam)->sin_addr.s_addr != INADDR_ANY || - t->inp_laddr.s_addr != INADDR_ANY)) { + t->inp_laddr.s_addr != INADDR_ANY) && + (!(t->inp_flags2 & INP2_EXTERNAL_PORT) || + !(inp->inp_flags2 & INP2_EXTERNAL_PORT) || + uuid_compare(t->necp_client_uuid, inp->necp_client_uuid) != 0)) { if ((t->inp_socket->so_flags & SOF_NOTIFYCONFLICT) && - !(so->so_flags & SOF_NOTIFYCONFLICT)) + !(so->so_flags & SOF_NOTIFYCONFLICT)) { conflict = 1; + } lck_rw_done(pcbinfo->ipi_lock); - if (conflict) + if (conflict) { in_pcb_conflict_post_msg(lport); + } socket_lock(so, 0); - return (EADDRINUSE); + return EADDRINUSE; } t = in_pcblookup_local_and_cleanup(pcbinfo, SIN(nam)->sin_addr, lport, wild); if (t != NULL && - (reuseport & t->inp_socket->so_options) == 0) { + (reuseport & t->inp_socket->so_options) == 0 && + (!(t->inp_flags2 & INP2_EXTERNAL_PORT) || + !(inp->inp_flags2 & INP2_EXTERNAL_PORT) || + uuid_compare(t->necp_client_uuid, inp->necp_client_uuid) != 0)) { #if INET6 if (SIN(nam)->sin_addr.s_addr != INADDR_ANY || t->inp_laddr.s_addr != INADDR_ANY || @@ -830,18 +925,19 @@ in_pcbbind(struct inpcb *inp, struct sockaddr *nam, struct proc *p) SOCK_DOM(t->inp_socket) != PF_INET6) #endif /* INET6 */ { - if ((t->inp_socket->so_flags & SOF_NOTIFYCONFLICT) && - !(so->so_flags & SOF_NOTIFYCONFLICT)) + !(so->so_flags & SOF_NOTIFYCONFLICT)) { conflict = 1; + } lck_rw_done(pcbinfo->ipi_lock); - if (conflict) + if (conflict) { in_pcb_conflict_post_msg(lport); + } socket_lock(so, 0); - return (EADDRINUSE); + return EADDRINUSE; } } } @@ -850,6 +946,14 @@ in_pcbbind(struct inpcb *inp, struct sockaddr *nam, struct proc *p) if (lport == 0) { u_short first, last; int count; + bool found; + + /* + * Override wild = 1 for implicit bind (mainly used by connect) + * For implicit bind (lport == 0), we always use an unused port, + * so REUSEADDR|REUSEPORT don't apply + */ + wild = 1; randomport = (so->so_flags & SOF_BINDRANDOMPORT) || (so->so_type == SOCK_STREAM ? tcp_use_randomport : @@ -861,7 +965,7 @@ in_pcbbind(struct inpcb *inp, struct sockaddr *nam, struct proc *p) */ anonport = TRUE; if (inp->inp_flags & INP_HIGHPORT) { - first = ipport_hifirstauto; /* sysctl */ + first = ipport_hifirstauto; /* sysctl */ last = ipport_hilastauto; lastport = &pcbinfo->ipi_lasthi; } else if (inp->inp_flags & INP_LOWPORT) { @@ -872,20 +976,21 @@ in_pcbbind(struct inpcb *inp, struct sockaddr *nam, struct proc *p) if (error != 0) { lck_rw_done(pcbinfo->ipi_lock); socket_lock(so, 0); - return (error); + return error; } - first = ipport_lowfirstauto; /* 1023 */ - last = ipport_lowlastauto; /* 600 */ + first = ipport_lowfirstauto; /* 1023 */ + last = ipport_lowlastauto; /* 600 */ lastport = &pcbinfo->ipi_lastlow; } else { - first = ipport_firstauto; /* sysctl */ + first = ipport_firstauto; /* sysctl */ last = ipport_lastauto; lastport = &pcbinfo->ipi_lastport; } /* No point in randomizing if only one port is available */ - if (first == last) + if (first == last) { randomport = 0; + } /* * Simple check to ensure all ports are not used up causing * a deadlock here. @@ -894,59 +999,102 @@ in_pcbbind(struct inpcb *inp, struct sockaddr *nam, struct proc *p) * is not being tested on each round of the loop. */ if (first > last) { + struct in_addr lookup_addr; + /* * counting down */ if (randomport) { - read_random(&rand_port, sizeof (rand_port)); + read_frandom(&rand_port, sizeof(rand_port)); *lastport = first - (rand_port % (first - last)); } count = first - last; + lookup_addr = (laddr.s_addr != INADDR_ANY) ? laddr : + inp->inp_laddr; + + found = false; do { - if (count-- < 0) { /* completely used? */ + if (count-- < 0) { /* completely used? */ lck_rw_done(pcbinfo->ipi_lock); socket_lock(so, 0); - return (EADDRNOTAVAIL); + return EADDRNOTAVAIL; } --*lastport; - if (*lastport > first || *lastport < last) + if (*lastport > first || *lastport < last) { *lastport = first; + } lport = htons(*lastport); - } while (in_pcblookup_local_and_cleanup(pcbinfo, - ((laddr.s_addr != INADDR_ANY) ? laddr : - inp->inp_laddr), lport, wild)); + + /* + * Skip if this is a restricted port as we do not want to + * restricted ports as ephemeral + */ + if (IS_RESTRICTED_IN_PORT(lport)) { + continue; + } + + found = in_pcblookup_local_and_cleanup(pcbinfo, + lookup_addr, lport, wild) == NULL; + } while (!found); } else { + struct in_addr lookup_addr; + /* * counting up */ if (randomport) { - read_random(&rand_port, sizeof (rand_port)); + read_frandom(&rand_port, sizeof(rand_port)); *lastport = first + (rand_port % (first - last)); } count = last - first; + lookup_addr = (laddr.s_addr != INADDR_ANY) ? laddr : + inp->inp_laddr; + + found = false; do { - if (count-- < 0) { /* completely used? */ + if (count-- < 0) { /* completely used? */ lck_rw_done(pcbinfo->ipi_lock); socket_lock(so, 0); - return (EADDRNOTAVAIL); + return EADDRNOTAVAIL; } ++*lastport; - if (*lastport < first || *lastport > last) + if (*lastport < first || *lastport > last) { *lastport = first; + } lport = htons(*lastport); - } while (in_pcblookup_local_and_cleanup(pcbinfo, - ((laddr.s_addr != INADDR_ANY) ? laddr : - inp->inp_laddr), lport, wild)); + + /* + * Skip if this is a restricted port as we do not want to + * restricted ports as ephemeral + */ + if (IS_RESTRICTED_IN_PORT(lport)) { + continue; + } + + found = in_pcblookup_local_and_cleanup(pcbinfo, + lookup_addr, lport, wild) == NULL; + } while (!found); } } socket_lock(so, 0); + + /* + * We unlocked socket's protocol lock for a long time. + * The socket might have been dropped/defuncted. + * Checking if world has changed since. + */ + if (inp->inp_state == INPCB_STATE_DEAD) { + lck_rw_done(pcbinfo->ipi_lock); + return ECONNABORTED; + } + if (inp->inp_lport != 0 || inp->inp_laddr.s_addr != INADDR_ANY) { lck_rw_done(pcbinfo->ipi_lock); - return (EINVAL); + return EINVAL; } if (laddr.s_addr != INADDR_ANY) { @@ -954,22 +1102,189 @@ in_pcbbind(struct inpcb *inp, struct sockaddr *nam, struct proc *p) inp->inp_last_outifp = outif; } inp->inp_lport = lport; - if (anonport) + if (anonport) { inp->inp_flags |= INP_ANONPORT; + } if (in_pcbinshash(inp, 1) != 0) { inp->inp_laddr.s_addr = INADDR_ANY; inp->inp_last_outifp = NULL; inp->inp_lport = 0; - if (anonport) + if (anonport) { inp->inp_flags &= ~INP_ANONPORT; + } lck_rw_done(pcbinfo->ipi_lock); - return (EAGAIN); + return EAGAIN; } lck_rw_done(pcbinfo->ipi_lock); sflt_notify(so, sock_evt_bound, NULL); - return (0); + return 0; +} + +#define APN_FALLBACK_IP_FILTER(a) \ + (IN_LINKLOCAL(ntohl((a)->sin_addr.s_addr)) || \ + IN_LOOPBACK(ntohl((a)->sin_addr.s_addr)) || \ + IN_ZERONET(ntohl((a)->sin_addr.s_addr)) || \ + IN_MULTICAST(ntohl((a)->sin_addr.s_addr)) || \ + IN_PRIVATE(ntohl((a)->sin_addr.s_addr))) + +#define APN_FALLBACK_NOTIF_INTERVAL 2 /* Magic Number */ +static uint64_t last_apn_fallback = 0; + +static boolean_t +apn_fallback_required(proc_t proc, struct socket *so, struct sockaddr_in *p_dstv4) +{ + uint64_t timenow; + struct sockaddr_storage lookup_default_addr; + struct rtentry *rt = NULL; + + VERIFY(proc != NULL); + + if (apn_fallbk_enabled == FALSE) { + return FALSE; + } + + if (proc == kernproc) { + return FALSE; + } + + if (so && (so->so_options & SO_NOAPNFALLBK)) { + return FALSE; + } + + timenow = net_uptime(); + if ((timenow - last_apn_fallback) < APN_FALLBACK_NOTIF_INTERVAL) { + apn_fallbk_log((LOG_INFO, "APN fallback notification throttled.\n")); + return FALSE; + } + + if (p_dstv4 && APN_FALLBACK_IP_FILTER(p_dstv4)) { + return FALSE; + } + + /* Check if we have unscoped IPv6 default route through cellular */ + bzero(&lookup_default_addr, sizeof(lookup_default_addr)); + lookup_default_addr.ss_family = AF_INET6; + lookup_default_addr.ss_len = sizeof(struct sockaddr_in6); + + rt = rtalloc1((struct sockaddr *)&lookup_default_addr, 0, 0); + if (NULL == rt) { + apn_fallbk_log((LOG_INFO, "APN fallback notification could not find " + "unscoped default IPv6 route.\n")); + return FALSE; + } + + if (!IFNET_IS_CELLULAR(rt->rt_ifp)) { + rtfree(rt); + apn_fallbk_log((LOG_INFO, "APN fallback notification could not find " + "unscoped default IPv6 route through cellular interface.\n")); + return FALSE; + } + + /* + * We have a default IPv6 route, ensure that + * we do not have IPv4 default route before triggering + * the event + */ + rtfree(rt); + rt = NULL; + + bzero(&lookup_default_addr, sizeof(lookup_default_addr)); + lookup_default_addr.ss_family = AF_INET; + lookup_default_addr.ss_len = sizeof(struct sockaddr_in); + + rt = rtalloc1((struct sockaddr *)&lookup_default_addr, 0, 0); + + if (rt) { + rtfree(rt); + rt = NULL; + apn_fallbk_log((LOG_INFO, "APN fallback notification found unscoped " + "IPv4 default route!\n")); + return FALSE; + } + + { + /* + * We disable APN fallback if the binary is not a third-party app. + * Note that platform daemons use their process name as a + * bundle ID so we filter out bundle IDs without dots. + */ + const char *bundle_id = cs_identity_get(proc); + if (bundle_id == NULL || + bundle_id[0] == '\0' || + strchr(bundle_id, '.') == NULL || + strncmp(bundle_id, "com.apple.", sizeof("com.apple.") - 1) == 0) { + apn_fallbk_log((LOG_INFO, "Abort: APN fallback notification found first-" + "party bundle ID \"%s\"!\n", (bundle_id ? bundle_id : "NULL"))); + return FALSE; + } + } + + { + /* + * The Apple App Store IPv6 requirement started on + * June 1st, 2016 at 12:00:00 AM PDT. + * We disable APN fallback if the binary is more recent than that. + * We check both atime and birthtime since birthtime is not always supported. + */ + static const long ipv6_start_date = 1464764400L; + vfs_context_t context; + struct stat64 sb; + int vn_stat_error; + + bzero(&sb, sizeof(struct stat64)); + context = vfs_context_create(NULL); + vn_stat_error = vn_stat(proc->p_textvp, &sb, NULL, 1, 0, context); + (void)vfs_context_rele(context); + + if (vn_stat_error != 0 || + sb.st_atimespec.tv_sec >= ipv6_start_date || + sb.st_birthtimespec.tv_sec >= ipv6_start_date) { + apn_fallbk_log((LOG_INFO, "Abort: APN fallback notification found binary " + "too recent! (err %d atime %ld mtime %ld ctime %ld birthtime %ld)\n", + vn_stat_error, sb.st_atimespec.tv_sec, sb.st_mtimespec.tv_sec, + sb.st_ctimespec.tv_sec, sb.st_birthtimespec.tv_sec)); + return FALSE; + } + } + return TRUE; +} + +static void +apn_fallback_trigger(proc_t proc, struct socket *so) +{ + pid_t pid = 0; + struct kev_msg ev_msg; + struct kev_netevent_apnfallbk_data apnfallbk_data; + + last_apn_fallback = net_uptime(); + pid = proc_pid(proc); + uuid_t application_uuid; + uuid_clear(application_uuid); + proc_getexecutableuuid(proc, application_uuid, + sizeof(application_uuid)); + + bzero(&ev_msg, sizeof(struct kev_msg)); + ev_msg.vendor_code = KEV_VENDOR_APPLE; + ev_msg.kev_class = KEV_NETWORK_CLASS; + ev_msg.kev_subclass = KEV_NETEVENT_SUBCLASS; + ev_msg.event_code = KEV_NETEVENT_APNFALLBACK; + + bzero(&apnfallbk_data, sizeof(apnfallbk_data)); + + if (so->so_flags & SOF_DELEGATED) { + apnfallbk_data.epid = so->e_pid; + uuid_copy(apnfallbk_data.euuid, so->e_uuid); + } else { + apnfallbk_data.epid = so->last_pid; + uuid_copy(apnfallbk_data.euuid, so->last_uuid); + } + + ev_msg.dv[0].data_ptr = &apnfallbk_data; + ev_msg.dv[0].data_length = sizeof(apnfallbk_data); + kev_post_msg(&ev_msg); + apn_fallbk_log((LOG_INFO, "APN fallback notification issued.\n")); } /* @@ -994,7 +1309,7 @@ in_pcbbind(struct inpcb *inp, struct sockaddr *nam, struct proc *p) */ int in_pcbladdr(struct inpcb *inp, struct sockaddr *nam, struct in_addr *laddr, - unsigned int ifscope, struct ifnet **outif) + unsigned int ifscope, struct ifnet **outif, int raw) { struct route *ro = &inp->inp_route; struct in_ifaddr *ia = NULL; @@ -1002,14 +1317,18 @@ in_pcbladdr(struct inpcb *inp, struct sockaddr *nam, struct in_addr *laddr, int error = 0; boolean_t restricted = FALSE; - if (outif != NULL) + if (outif != NULL) { *outif = NULL; - if (nam->sa_len != sizeof (struct sockaddr_in)) - return (EINVAL); - if (SIN(nam)->sin_family != AF_INET) - return (EAFNOSUPPORT); - if (SIN(nam)->sin_port == 0) - return (EADDRNOTAVAIL); + } + if (nam->sa_len != sizeof(struct sockaddr_in)) { + return EINVAL; + } + if (SIN(nam)->sin_family != AF_INET) { + return EAFNOSUPPORT; + } + if (raw == 0 && SIN(nam)->sin_port == 0) { + return EADDRNOTAVAIL; + } /* * If the destination address is INADDR_ANY, @@ -1018,8 +1337,8 @@ in_pcbladdr(struct inpcb *inp, struct sockaddr *nam, struct in_addr *laddr, * and the primary interface supports broadcast, * choose the broadcast address for that interface. */ - if (SIN(nam)->sin_addr.s_addr == INADDR_ANY || - SIN(nam)->sin_addr.s_addr == (u_int32_t)INADDR_BROADCAST) { + if (raw == 0 && (SIN(nam)->sin_addr.s_addr == INADDR_ANY || + SIN(nam)->sin_addr.s_addr == (u_int32_t)INADDR_BROADCAST)) { lck_rw_lock_shared(in_ifaddr_rwlock); if (!TAILQ_EMPTY(&in_ifaddrhead)) { ia = TAILQ_FIRST(&in_ifaddrhead); @@ -1041,15 +1360,16 @@ in_pcbladdr(struct inpcb *inp, struct sockaddr *nam, struct in_addr *laddr, if (inp->inp_laddr.s_addr != INADDR_ANY) { VERIFY(ia == NULL); *laddr = inp->inp_laddr; - return (0); + return 0; } /* * If the ifscope is specified by the caller (e.g. IP_PKTINFO) * then it overrides the sticky ifscope set for the socket. */ - if (ifscope == IFSCOPE_NONE && (inp->inp_flags & INP_BOUND_IF)) + if (ifscope == IFSCOPE_NONE && (inp->inp_flags & INP_BOUND_IF)) { ifscope = inp->inp_boundifp->if_index; + } /* * If route is known or can be allocated now, @@ -1057,33 +1377,37 @@ in_pcbladdr(struct inpcb *inp, struct sockaddr *nam, struct in_addr *laddr, * Note that we should check the address family of the cached * destination, in case of sharing the cache with IPv6. */ - if (ro->ro_rt != NULL) + if (ro->ro_rt != NULL) { RT_LOCK_SPIN(ro->ro_rt); + } if (ROUTE_UNUSABLE(ro) || ro->ro_dst.sa_family != AF_INET || SIN(&ro->ro_dst)->sin_addr.s_addr != SIN(nam)->sin_addr.s_addr || (inp->inp_socket->so_options & SO_DONTROUTE)) { - if (ro->ro_rt != NULL) + if (ro->ro_rt != NULL) { RT_UNLOCK(ro->ro_rt); + } ROUTE_RELEASE(ro); } if (!(inp->inp_socket->so_options & SO_DONTROUTE) && (ro->ro_rt == NULL || ro->ro_rt->rt_ifp == NULL)) { - if (ro->ro_rt != NULL) + if (ro->ro_rt != NULL) { RT_UNLOCK(ro->ro_rt); + } ROUTE_RELEASE(ro); /* No route yet, so try to acquire one */ - bzero(&ro->ro_dst, sizeof (struct sockaddr_in)); + bzero(&ro->ro_dst, sizeof(struct sockaddr_in)); ro->ro_dst.sa_family = AF_INET; - ro->ro_dst.sa_len = sizeof (struct sockaddr_in); + ro->ro_dst.sa_len = sizeof(struct sockaddr_in); SIN(&ro->ro_dst)->sin_addr = SIN(nam)->sin_addr; rtalloc_scoped(ro, ifscope); - if (ro->ro_rt != NULL) + if (ro->ro_rt != NULL) { RT_LOCK_SPIN(ro->ro_rt); + } } /* Sanitized local copy for interface address searches */ - bzero(&sin, sizeof (sin)); + bzero(&sin, sizeof(sin)); sin.sin_family = AF_INET; - sin.sin_len = sizeof (struct sockaddr_in); + sin.sin_len = sizeof(struct sockaddr_in); sin.sin_addr.s_addr = SIN(nam)->sin_addr.s_addr; /* * If we did not find (or use) a route, assume dest is reachable @@ -1091,11 +1415,20 @@ in_pcbladdr(struct inpcb *inp, struct sockaddr *nam, struct in_addr *laddr, * interface to take the source address from. */ if (ro->ro_rt == NULL) { + proc_t proc = current_proc(); + VERIFY(ia == NULL); ia = ifatoia(ifa_ifwithdstaddr(SA(&sin))); - if (ia == NULL) + if (ia == NULL) { ia = ifatoia(ifa_ifwithnet_scoped(SA(&sin), ifscope)); + } error = ((ia == NULL) ? ENETUNREACH : 0); + + if (apn_fallback_required(proc, inp->inp_socket, + (void *)nam)) { + apn_fallback_trigger(proc, inp->inp_socket); + } + goto done; } RT_LOCK_ASSERT_HELD(ro->ro_rt); @@ -1121,6 +1454,21 @@ in_pcbladdr(struct inpcb *inp, struct sockaddr *nam, struct in_addr *laddr, RT_CONVERT_LOCK(ro->ro_rt); ia = ifatoia(ro->ro_rt->rt_ifa); IFA_ADDREF(&ia->ia_ifa); + + /* + * Mark the control block for notification of + * a possible flow that might undergo clat46 + * translation. + * + * We defer the decision to a later point when + * inpcb is being disposed off. + * The reason is that we only want to send notification + * if the flow was ever used to send data. + */ + if (IS_INTF_CLAT46(ro->ro_rt->rt_ifp)) { + inp->inp_flags2 |= INP2_CLAT46_FLOW; + } + RT_UNLOCK(ro->ro_rt); error = 0; } @@ -1137,15 +1485,18 @@ in_pcbladdr(struct inpcb *inp, struct sockaddr *nam, struct in_addr *laddr, */ VERIFY(ia == NULL); ia = ifatoia(ifa_ifwithdstaddr(SA(&sin))); - if (ia == NULL) + if (ia == NULL) { ia = ifatoia(ifa_ifwithaddr_scoped(SA(&sin), ifscope)); - if (ia == NULL) + } + if (ia == NULL) { ia = ifatoia(ifa_ifwithnet_scoped(SA(&sin), ifscope)); + } if (ia == NULL) { RT_LOCK(ro->ro_rt); ia = ifatoia(ro->ro_rt->rt_ifa); - if (ia != NULL) + if (ia != NULL) { IFA_ADDREF(&ia->ia_ifa); + } RT_UNLOCK(ro->ro_rt); } error = ((ia == NULL) ? ENETUNREACH : 0); @@ -1166,20 +1517,24 @@ done: if (imo->imo_multicast_ifp != NULL && (ia == NULL || ia->ia_ifp != imo->imo_multicast_ifp)) { ifp = imo->imo_multicast_ifp; - if (ia != NULL) + if (ia != NULL) { IFA_REMREF(&ia->ia_ifa); + } lck_rw_lock_shared(in_ifaddr_rwlock); TAILQ_FOREACH(ia, &in_ifaddrhead, ia_link) { - if (ia->ia_ifp == ifp) + if (ia->ia_ifp == ifp) { break; + } } - if (ia != NULL) + if (ia != NULL) { IFA_ADDREF(&ia->ia_ifa); + } lck_rw_done(in_ifaddr_rwlock); - if (ia == NULL) + if (ia == NULL) { error = EADDRNOTAVAIL; - else + } else { error = 0; + } } IMO_UNLOCK(imo); } @@ -1204,16 +1559,18 @@ done: if (outif != NULL) { struct ifnet *ifp; - if (ro->ro_rt != NULL) + if (ro->ro_rt != NULL) { ifp = ro->ro_rt->rt_ifp; - else + } else { ifp = ia->ia_ifp; + } VERIFY(ifp != NULL); IFA_CONVERT_LOCK(&ia->ia_ifa); - ifnet_reference(ifp); /* for caller */ - if (*outif != NULL) + ifnet_reference(ifp); /* for caller */ + if (*outif != NULL) { ifnet_release(*outif); + } *outif = ifp; } IFA_UNLOCK(&ia->ia_ifa); @@ -1229,7 +1586,7 @@ done: SO_FILT_HINT_IFDENIED)); } - return (error); + return error; } /* @@ -1252,11 +1609,18 @@ in_pcbconnect(struct inpcb *inp, struct sockaddr *nam, struct proc *p, int error; struct socket *so = inp->inp_socket; +#if CONTENT_FILTER + if (so) { + so->so_state_change_cnt++; + } +#endif + /* * Call inner routine, to assign local interface address. */ - if ((error = in_pcbladdr(inp, nam, &laddr, ifscope, outif)) != 0) - return (error); + if ((error = in_pcbladdr(inp, nam, &laddr, ifscope, outif, 0)) != 0) { + return error; + } socket_unlock(so, 0); pcb = in_pcblookup_hash(inp->inp_pcbinfo, sin->sin_addr, sin->sin_port, @@ -1269,18 +1633,20 @@ in_pcbconnect(struct inpcb *inp, struct sockaddr *nam, struct proc *p, * embryonic socket, it can get aborted if another thread is closing * the listener (radar 7947600). */ - if ((so->so_flags & SOF_ABORTED) != 0) - return (ECONNREFUSED); + if ((so->so_flags & SOF_ABORTED) != 0) { + return ECONNREFUSED; + } if (pcb != NULL) { in_pcb_checkstate(pcb, WNT_RELEASE, pcb == inp ? 1 : 0); - return (EADDRINUSE); + return EADDRINUSE; } if (inp->inp_laddr.s_addr == INADDR_ANY) { if (inp->inp_lport == 0) { error = in_pcbbind(inp, NULL, p); - if (error) - return (error); + if (error) { + return error; + } } if (!lck_rw_try_lock_exclusive(inp->inp_pcbinfo->ipi_lock)) { /* @@ -1296,6 +1662,17 @@ in_pcbconnect(struct inpcb *inp, struct sockaddr *nam, struct proc *p, inp->inp_last_outifp = (outif != NULL) ? *outif : NULL; inp->inp_flags |= INP_INADDR_ANY; } else { + /* + * Usage of IP_PKTINFO, without local port already + * speficified will cause kernel to panic, + * see rdar://problem/18508185. + * For now returning error to avoid a kernel panic + * This routines can be refactored and handle this better + * in future. + */ + if (inp->inp_lport == 0) { + return EINVAL; + } if (!lck_rw_try_lock_exclusive(inp->inp_pcbinfo->ipi_lock)) { /* * Lock inversion issue, mostly with udp @@ -1308,11 +1685,12 @@ in_pcbconnect(struct inpcb *inp, struct sockaddr *nam, struct proc *p, } inp->inp_faddr = sin->sin_addr; inp->inp_fport = sin->sin_port; - if (nstat_collect && SOCK_PROTO(so) == IPPROTO_UDP) + if (nstat_collect && SOCK_PROTO(so) == IPPROTO_UDP) { nstat_pcb_invalidate_cache(inp); + } in_pcbrehash(inp); lck_rw_done(inp->inp_pcbinfo->ipi_lock); - return (0); + return 0; } void @@ -1320,12 +1698,19 @@ in_pcbdisconnect(struct inpcb *inp) { struct socket *so = inp->inp_socket; - if (nstat_collect && SOCK_PROTO(so) == IPPROTO_UDP) + if (nstat_collect && SOCK_PROTO(so) == IPPROTO_UDP) { nstat_pcb_cache(inp); + } inp->inp_faddr.s_addr = INADDR_ANY; inp->inp_fport = 0; +#if CONTENT_FILTER + if (so) { + so->so_state_change_cnt++; + } +#endif + if (!lck_rw_try_lock_exclusive(inp->inp_pcbinfo->ipi_lock)) { /* lock inversion issue, mostly with udp multicast packets */ socket_unlock(so, 0); @@ -1340,8 +1725,9 @@ in_pcbdisconnect(struct inpcb *inp) * so check for SOF_MP_SUBFLOW socket flag before detaching the PCB; * when the socket is closed for real, SOF_MP_SUBFLOW would be cleared. */ - if (!(so->so_flags & SOF_MP_SUBFLOW) && (so->so_state & SS_NOFDREF)) + if (!(so->so_flags & SOF_MP_SUBFLOW) && (so->so_state & SS_NOFDREF)) { in_pcbdetach(inp); + } } void @@ -1355,20 +1741,34 @@ in_pcbdetach(struct inpcb *inp) inp, so, SOCK_PROTO(so)); /* NOTREACHED */ } - + #if IPSEC if (inp->inp_sp != NULL) { (void) ipsec4_delete_pcbpolicy(inp); } #endif /* IPSEC */ - + + if (inp->inp_stat != NULL && SOCK_PROTO(so) == IPPROTO_UDP) { + if (inp->inp_stat->rxpackets == 0 && inp->inp_stat->txpackets == 0) { + INC_ATOMIC_INT64_LIM(net_api_stats.nas_socket_inet_dgram_no_data); + } + } + /* * Let NetworkStatistics know this PCB is going away * before we detach it. */ - if (nstat_collect && - (SOCK_PROTO(so) == IPPROTO_TCP || SOCK_PROTO(so) == IPPROTO_UDP)) + if (nstat_collect && + (SOCK_PROTO(so) == IPPROTO_TCP || SOCK_PROTO(so) == IPPROTO_UDP)) { nstat_pcb_detach(inp); + } + + /* Free memory buffer held for generating keep alives */ + if (inp->inp_keepalive_data != NULL) { + FREE(inp->inp_keepalive_data, M_TEMP); + inp->inp_keepalive_data = NULL; + } + /* mark socket state as dead */ if (in_pcb_checkstate(inp, WNT_STOPUSING, 1) != WNT_STOPUSING) { panic("%s: so=%p proto=%d couldn't set to STOPUSING\n", @@ -1387,14 +1787,50 @@ in_pcbdetach(struct inpcb *inp) ROUTE_RELEASE(&inp->inp_route); imo = inp->inp_moptions; inp->inp_moptions = NULL; - if (imo != NULL) - IMO_REMREF(imo); sofreelastref(so, 0); inp->inp_state = INPCB_STATE_DEAD; + + /* + * Enqueue an event to send kernel event notification + * if the flow has to CLAT46 for data packets + */ + if (inp->inp_flags2 & INP2_CLAT46_FLOW) { + /* + * If there has been any exchange of data bytes + * over this flow. + * Schedule a notification to report that flow is + * using client side translation. + */ + if (inp->inp_stat != NULL && + (inp->inp_stat->txbytes != 0 || + inp->inp_stat->rxbytes != 0)) { + if (so->so_flags & SOF_DELEGATED) { + in6_clat46_event_enqueue_nwk_wq_entry( + IN6_CLAT46_EVENT_V4_FLOW, + so->e_pid, + so->e_uuid); + } else { + in6_clat46_event_enqueue_nwk_wq_entry( + IN6_CLAT46_EVENT_V4_FLOW, + so->last_pid, + so->last_uuid); + } + } + } + /* makes sure we're not called twice from so_close */ so->so_flags |= SOF_PCBCLEARING; inpcb_gc_sched(inp->inp_pcbinfo, INPCB_TIMER_FAST); + + /* + * See inp_join_group() for why we need to unlock + */ + if (imo != NULL) { + socket_unlock(so, 0); + IMO_REMREF(imo); + socket_lock(so, 0); + } } } @@ -1426,7 +1862,7 @@ in_pcbdispose(struct inpcb *inp) } } - lck_rw_assert(ipi->ipi_lock, LCK_RW_ASSERT_EXCLUSIVE); + LCK_RW_ASSERT(ipi->ipi_lock, LCK_RW_ASSERT_EXCLUSIVE); inp->inp_gencnt = ++ipi->ipi_gencnt; /* access ipi in in_pcbremlists */ @@ -1449,6 +1885,11 @@ in_pcbdispose(struct inpcb *inp) /* NOTREACHED */ } lck_mtx_unlock(&inp->inpcb_mtx); + +#if NECP + necp_inpcb_remove_cb(inp); +#endif /* NECP */ + lck_mtx_destroy(&inp->inpcb_mtx, ipi->ipi_lock_grp); } /* makes sure we're not called twice from so_close */ @@ -1459,13 +1900,16 @@ in_pcbdispose(struct inpcb *inp) #if CONFIG_MACF_NET mac_inpcb_label_destroy(inp); #endif /* CONFIG_MACF_NET */ +#if NECP + necp_inpcb_dispose(inp); +#endif /* NECP */ /* * In case there a route cached after a detach (possible * in the tcp case), make sure that it is freed before * we deallocate the structure. */ ROUTE_RELEASE(&inp->inp_route); - if (!so->cached_in_sock_layer) { + if ((so->so_flags1 & SOF1_CACHED_IN_SOCK_LAYER) == 0) { zfree(ipi->ipi_zone, inp); } sodealloc(so); @@ -1487,46 +1931,44 @@ in_getsockaddr(struct socket *so, struct sockaddr **nam) /* * Do the malloc first in case it blocks. */ - MALLOC(sin, struct sockaddr_in *, sizeof (*sin), M_SONAME, M_WAITOK); - if (sin == NULL) - return (ENOBUFS); - bzero(sin, sizeof (*sin)); + MALLOC(sin, struct sockaddr_in *, sizeof(*sin), M_SONAME, M_WAITOK); + if (sin == NULL) { + return ENOBUFS; + } + bzero(sin, sizeof(*sin)); sin->sin_family = AF_INET; - sin->sin_len = sizeof (*sin); + sin->sin_len = sizeof(*sin); if ((inp = sotoinpcb(so)) == NULL) { FREE(sin, M_SONAME); - return (EINVAL); + return EINVAL; } sin->sin_port = inp->inp_lport; sin->sin_addr = inp->inp_laddr; *nam = (struct sockaddr *)sin; - return (0); + return 0; } int -in_getsockaddr_s(struct socket *so, struct sockaddr_storage *ss) +in_getsockaddr_s(struct socket *so, struct sockaddr_in *ss) { - struct sockaddr_in *sin = SIN(ss); + struct sockaddr_in *sin = ss; struct inpcb *inp; VERIFY(ss != NULL); - bzero(ss, sizeof (*ss)); + bzero(ss, sizeof(*ss)); sin->sin_family = AF_INET; - sin->sin_len = sizeof (*sin); + sin->sin_len = sizeof(*sin); - if ((inp = sotoinpcb(so)) == NULL -#if NECP - || (necp_socket_should_use_flow_divert(inp)) -#endif /* NECP */ - ) - return (inp == NULL ? EINVAL : EPROTOTYPE); + if ((inp = sotoinpcb(so)) == NULL) { + return EINVAL; + } sin->sin_port = inp->inp_lport; sin->sin_addr = inp->inp_laddr; - return (0); + return 0; } int @@ -1538,47 +1980,23 @@ in_getpeeraddr(struct socket *so, struct sockaddr **nam) /* * Do the malloc first in case it blocks. */ - MALLOC(sin, struct sockaddr_in *, sizeof (*sin), M_SONAME, M_WAITOK); - if (sin == NULL) - return (ENOBUFS); - bzero((caddr_t)sin, sizeof (*sin)); + MALLOC(sin, struct sockaddr_in *, sizeof(*sin), M_SONAME, M_WAITOK); + if (sin == NULL) { + return ENOBUFS; + } + bzero((caddr_t)sin, sizeof(*sin)); sin->sin_family = AF_INET; - sin->sin_len = sizeof (*sin); + sin->sin_len = sizeof(*sin); if ((inp = sotoinpcb(so)) == NULL) { FREE(sin, M_SONAME); - return (EINVAL); + return EINVAL; } sin->sin_port = inp->inp_fport; sin->sin_addr = inp->inp_faddr; *nam = (struct sockaddr *)sin; - return (0); -} - -int -in_getpeeraddr_s(struct socket *so, struct sockaddr_storage *ss) -{ - struct sockaddr_in *sin = SIN(ss); - struct inpcb *inp; - - VERIFY(ss != NULL); - bzero(ss, sizeof (*ss)); - - sin->sin_family = AF_INET; - sin->sin_len = sizeof (*sin); - - if ((inp = sotoinpcb(so)) == NULL -#if NECP - || (necp_socket_should_use_flow_divert(inp)) -#endif /* NECP */ - ) { - return (inp == NULL ? EINVAL : EPROTOTYPE); - } - - sin->sin_port = inp->inp_fport; - sin->sin_addr = inp->inp_faddr; - return (0); + return 0; } void @@ -1591,14 +2009,17 @@ in_pcbnotifyall(struct inpcbinfo *pcbinfo, struct in_addr faddr, LIST_FOREACH(inp, pcbinfo->ipi_listhead, inp_list) { #if INET6 - if (!(inp->inp_vflag & INP_IPV4)) + if (!(inp->inp_vflag & INP_IPV4)) { continue; + } #endif /* INET6 */ if (inp->inp_faddr.s_addr != faddr.s_addr || - inp->inp_socket == NULL) + inp->inp_socket == NULL) { continue; - if (in_pcb_checkstate(inp, WNT_ACQUIRE, 0) == WNT_STOPUSING) + } + if (in_pcb_checkstate(inp, WNT_ACQUIRE, 0) == WNT_STOPUSING) { continue; + } socket_lock(inp->inp_socket, 1); (*notify)(inp, errno); (void) in_pcb_checkstate(inp, WNT_RELEASE, 1); @@ -1618,18 +2039,11 @@ in_losing(struct inpcb *inp) { boolean_t release = FALSE; struct rtentry *rt; - struct rt_addrinfo info; if ((rt = inp->inp_route.ro_rt) != NULL) { struct in_ifaddr *ia = NULL; - bzero((caddr_t)&info, sizeof (info)); RT_LOCK(rt); - info.rti_info[RTAX_DST] = - (struct sockaddr *)&inp->inp_route.ro_dst; - info.rti_info[RTAX_GATEWAY] = rt->rt_gateway; - info.rti_info[RTAX_NETMASK] = rt_mask(rt); - rt_missmsg(RTM_LOSING, &info, rt->rt_flags, 0); if (rt->rt_flags & RTF_DYNAMIC) { /* * Prevent another thread from modifying rt_key, @@ -1652,11 +2066,13 @@ in_losing(struct inpcb *inp) */ release = TRUE; } - if (ia != NULL) + if (ia != NULL) { IFA_REMREF(&ia->ia_ifa); + } } - if (rt == NULL || release) + if (rt == NULL || release) { ROUTE_RELEASE(&inp->inp_route); + } } /* @@ -1682,11 +2098,13 @@ in_rtchange(struct inpcb *inp, int errno) */ release = TRUE; } - if (ia != NULL) + if (ia != NULL) { IFA_REMREF(&ia->ia_ifa); + } } - if (rt == NULL || release) + if (rt == NULL || release) { ROUTE_RELEASE(&inp->inp_route); + } } /* @@ -1712,8 +2130,9 @@ in_pcblookup_local(struct inpcbinfo *pcbinfo, struct in_addr laddr, pcbinfo->ipi_hashmask)]; LIST_FOREACH(inp, head, inp_hash) { #if INET6 - if (!(inp->inp_vflag & INP_IPV4)) + if (!(inp->inp_vflag & INP_IPV4)) { continue; + } #endif /* INET6 */ if (inp->inp_faddr.s_addr == INADDR_ANY && inp->inp_laddr.s_addr == laddr.s_addr && @@ -1721,14 +2140,14 @@ in_pcblookup_local(struct inpcbinfo *pcbinfo, struct in_addr laddr, /* * Found. */ - return (inp); + return inp; } } /* * Not found. */ KERNEL_DEBUG(DBG_FNC_PCB_LOOKUP | DBG_FUNC_END, 0, 0, 0, 0, 0); - return (NULL); + return NULL; } else { struct inpcbporthead *porthash; struct inpcbport *phd; @@ -1742,8 +2161,9 @@ in_pcblookup_local(struct inpcbinfo *pcbinfo, struct in_addr laddr, porthash = &pcbinfo->ipi_porthashbase[INP_PCBPORTHASH(lport, pcbinfo->ipi_porthashmask)]; LIST_FOREACH(phd, porthash, phd_hash) { - if (phd->phd_port == lport) + if (phd->phd_port == lport) { break; + } } if (phd != NULL) { /* @@ -1753,20 +2173,24 @@ in_pcblookup_local(struct inpcbinfo *pcbinfo, struct in_addr laddr, LIST_FOREACH(inp, &phd->phd_pcblist, inp_portlist) { wildcard = 0; #if INET6 - if (!(inp->inp_vflag & INP_IPV4)) + if (!(inp->inp_vflag & INP_IPV4)) { continue; + } #endif /* INET6 */ - if (inp->inp_faddr.s_addr != INADDR_ANY) + if (inp->inp_faddr.s_addr != INADDR_ANY) { wildcard++; + } if (inp->inp_laddr.s_addr != INADDR_ANY) { - if (laddr.s_addr == INADDR_ANY) + if (laddr.s_addr == INADDR_ANY) { wildcard++; - else if (inp->inp_laddr.s_addr != - laddr.s_addr) + } else if (inp->inp_laddr.s_addr != + laddr.s_addr) { continue; + } } else { - if (laddr.s_addr != INADDR_ANY) + if (laddr.s_addr != INADDR_ANY) { wildcard++; + } } if (wildcard < matchwild) { match = inp; @@ -1779,7 +2203,7 @@ in_pcblookup_local(struct inpcbinfo *pcbinfo, struct in_addr laddr, } KERNEL_DEBUG(DBG_FNC_PCB_LOOKUP | DBG_FUNC_END, match, 0, 0, 0, 0); - return (match); + return match; } } @@ -1816,11 +2240,19 @@ in_pcblookup_hash_exists(struct inpcbinfo *pcbinfo, struct in_addr faddr, pcbinfo->ipi_hashmask)]; LIST_FOREACH(inp, head, inp_hash) { #if INET6 - if (!(inp->inp_vflag & INP_IPV4)) + if (!(inp->inp_vflag & INP_IPV4)) { continue; + } #endif /* INET6 */ - if (inp_restricted_recv(inp, ifp)) + if (inp_restricted_recv(inp, ifp)) { + continue; + } + +#if NECP + if (!necp_socket_is_allowed_to_recv_on_interface(inp, ifp)) { continue; + } +#endif /* NECP */ if (inp->inp_faddr.s_addr == faddr.s_addr && inp->inp_laddr.s_addr == laddr.s_addr && @@ -1831,12 +2263,12 @@ in_pcblookup_hash_exists(struct inpcbinfo *pcbinfo, struct in_addr faddr, * Found. */ *uid = kauth_cred_getuid( - inp->inp_socket->so_cred); + inp->inp_socket->so_cred); *gid = kauth_cred_getgid( - inp->inp_socket->so_cred); + inp->inp_socket->so_cred); } lck_rw_done(pcbinfo->ipi_lock); - return (found); + return found; } } @@ -1845,38 +2277,46 @@ in_pcblookup_hash_exists(struct inpcbinfo *pcbinfo, struct in_addr faddr, * Not found. */ lck_rw_done(pcbinfo->ipi_lock); - return (0); + return 0; } head = &pcbinfo->ipi_hashbase[INP_PCBHASH(INADDR_ANY, lport, 0, pcbinfo->ipi_hashmask)]; LIST_FOREACH(inp, head, inp_hash) { #if INET6 - if (!(inp->inp_vflag & INP_IPV4)) + if (!(inp->inp_vflag & INP_IPV4)) { continue; + } #endif /* INET6 */ - if (inp_restricted_recv(inp, ifp)) + if (inp_restricted_recv(inp, ifp)) { + continue; + } + +#if NECP + if (!necp_socket_is_allowed_to_recv_on_interface(inp, ifp)) { continue; + } +#endif /* NECP */ if (inp->inp_faddr.s_addr == INADDR_ANY && inp->inp_lport == lport) { if (inp->inp_laddr.s_addr == laddr.s_addr) { if ((found = (inp->inp_socket != NULL))) { *uid = kauth_cred_getuid( - inp->inp_socket->so_cred); + inp->inp_socket->so_cred); *gid = kauth_cred_getgid( - inp->inp_socket->so_cred); + inp->inp_socket->so_cred); } lck_rw_done(pcbinfo->ipi_lock); - return (found); + return found; } else if (inp->inp_laddr.s_addr == INADDR_ANY) { #if INET6 if (inp->inp_socket && - SOCK_CHECK_DOM(inp->inp_socket, PF_INET6)) + SOCK_CHECK_DOM(inp->inp_socket, PF_INET6)) { local_wild_mapped = inp; - else + } else #endif /* INET6 */ - local_wild = inp; + local_wild = inp; } } } @@ -1885,25 +2325,25 @@ in_pcblookup_hash_exists(struct inpcbinfo *pcbinfo, struct in_addr faddr, if (local_wild_mapped != NULL) { if ((found = (local_wild_mapped->inp_socket != NULL))) { *uid = kauth_cred_getuid( - local_wild_mapped->inp_socket->so_cred); + local_wild_mapped->inp_socket->so_cred); *gid = kauth_cred_getgid( - local_wild_mapped->inp_socket->so_cred); + local_wild_mapped->inp_socket->so_cred); } lck_rw_done(pcbinfo->ipi_lock); - return (found); + return found; } #endif /* INET6 */ lck_rw_done(pcbinfo->ipi_lock); - return (0); + return 0; } if ((found = (local_wild->inp_socket != NULL))) { *uid = kauth_cred_getuid( - local_wild->inp_socket->so_cred); + local_wild->inp_socket->so_cred); *gid = kauth_cred_getgid( - local_wild->inp_socket->so_cred); + local_wild->inp_socket->so_cred); } lck_rw_done(pcbinfo->ipi_lock); - return (found); + return found; } /* @@ -1935,11 +2375,19 @@ in_pcblookup_hash(struct inpcbinfo *pcbinfo, struct in_addr faddr, pcbinfo->ipi_hashmask)]; LIST_FOREACH(inp, head, inp_hash) { #if INET6 - if (!(inp->inp_vflag & INP_IPV4)) + if (!(inp->inp_vflag & INP_IPV4)) { continue; + } #endif /* INET6 */ - if (inp_restricted_recv(inp, ifp)) + if (inp_restricted_recv(inp, ifp)) { continue; + } + +#if NECP + if (!necp_socket_is_allowed_to_recv_on_interface(inp, ifp)) { + continue; + } +#endif /* NECP */ if (inp->inp_faddr.s_addr == faddr.s_addr && inp->inp_laddr.s_addr == laddr.s_addr && @@ -1951,11 +2399,11 @@ in_pcblookup_hash(struct inpcbinfo *pcbinfo, struct in_addr faddr, if (in_pcb_checkstate(inp, WNT_ACQUIRE, 0) != WNT_STOPUSING) { lck_rw_done(pcbinfo->ipi_lock); - return (inp); + return inp; } else { /* it's there but dead, say it isn't found */ lck_rw_done(pcbinfo->ipi_lock); - return (NULL); + return NULL; } } } @@ -1965,18 +2413,26 @@ in_pcblookup_hash(struct inpcbinfo *pcbinfo, struct in_addr faddr, * Not found. */ lck_rw_done(pcbinfo->ipi_lock); - return (NULL); + return NULL; } head = &pcbinfo->ipi_hashbase[INP_PCBHASH(INADDR_ANY, lport, 0, pcbinfo->ipi_hashmask)]; LIST_FOREACH(inp, head, inp_hash) { #if INET6 - if (!(inp->inp_vflag & INP_IPV4)) + if (!(inp->inp_vflag & INP_IPV4)) { continue; + } #endif /* INET6 */ - if (inp_restricted_recv(inp, ifp)) + if (inp_restricted_recv(inp, ifp)) { + continue; + } + +#if NECP + if (!necp_socket_is_allowed_to_recv_on_interface(inp, ifp)) { continue; + } +#endif /* NECP */ if (inp->inp_faddr.s_addr == INADDR_ANY && inp->inp_lport == lport) { @@ -1984,19 +2440,19 @@ in_pcblookup_hash(struct inpcbinfo *pcbinfo, struct in_addr faddr, if (in_pcb_checkstate(inp, WNT_ACQUIRE, 0) != WNT_STOPUSING) { lck_rw_done(pcbinfo->ipi_lock); - return (inp); + return inp; } else { /* it's dead; say it isn't found */ lck_rw_done(pcbinfo->ipi_lock); - return (NULL); + return NULL; } } else if (inp->inp_laddr.s_addr == INADDR_ANY) { #if INET6 - if (SOCK_CHECK_DOM(inp->inp_socket, PF_INET6)) + if (SOCK_CHECK_DOM(inp->inp_socket, PF_INET6)) { local_wild_mapped = inp; - else + } else #endif /* INET6 */ - local_wild = inp; + local_wild = inp; } } } @@ -2006,30 +2462,36 @@ in_pcblookup_hash(struct inpcbinfo *pcbinfo, struct in_addr faddr, if (in_pcb_checkstate(local_wild_mapped, WNT_ACQUIRE, 0) != WNT_STOPUSING) { lck_rw_done(pcbinfo->ipi_lock); - return (local_wild_mapped); + return local_wild_mapped; } else { /* it's dead; say it isn't found */ lck_rw_done(pcbinfo->ipi_lock); - return (NULL); + return NULL; } } #endif /* INET6 */ lck_rw_done(pcbinfo->ipi_lock); - return (NULL); + return NULL; } if (in_pcb_checkstate(local_wild, WNT_ACQUIRE, 0) != WNT_STOPUSING) { lck_rw_done(pcbinfo->ipi_lock); - return (local_wild); + return local_wild; } /* * It's either not found or is already dead. */ lck_rw_done(pcbinfo->ipi_lock); - return (NULL); + return NULL; } /* - * Insert PCB onto various hash lists. + * @brief Insert PCB onto various hash lists. + * + * @param inp Pointer to internet protocol control block + * @param locked Implies if ipi_lock (protecting pcb list) + * is already locked or not. + * + * @return int error on failure and 0 on success */ int in_pcbinshash(struct inpcb *inp, int locked) @@ -2049,23 +2511,29 @@ in_pcbinshash(struct inpcb *inp, int locked) socket_unlock(inp->inp_socket, 0); lck_rw_lock_exclusive(pcbinfo->ipi_lock); socket_lock(inp->inp_socket, 0); - if (inp->inp_state == INPCB_STATE_DEAD) { - /* - * The socket got dropped when - * it was unlocked - */ - lck_rw_done(pcbinfo->ipi_lock); - return (ECONNABORTED); - } } } + /* + * This routine or its caller may have given up + * socket's protocol lock briefly. + * During that time the socket may have been dropped. + * Safe-guarding against that. + */ + if (inp->inp_state == INPCB_STATE_DEAD) { + if (!locked) { + lck_rw_done(pcbinfo->ipi_lock); + } + return ECONNABORTED; + } + + #if INET6 - if (inp->inp_vflag & INP_IPV6) + if (inp->inp_vflag & INP_IPV6) { hashkey_faddr = inp->in6p_faddr.s6_addr32[3] /* XXX */; - else + } else #endif /* INET6 */ - hashkey_faddr = inp->inp_faddr.s_addr; + hashkey_faddr = inp->inp_faddr.s_addr; inp->inp_hash_element = INP_PCBHASH(hashkey_faddr, inp->inp_lport, inp->inp_fport, pcbinfo->ipi_hashmask); @@ -2079,22 +2547,22 @@ in_pcbinshash(struct inpcb *inp, int locked) * Go through port list and look for a head for this lport. */ LIST_FOREACH(phd, pcbporthash, phd_hash) { - if (phd->phd_port == inp->inp_lport) + if (phd->phd_port == inp->inp_lport) { break; + } } - VERIFY(inp->inp_state != INPCB_STATE_DEAD); - /* * If none exists, malloc one and tack it on. */ if (phd == NULL) { - MALLOC(phd, struct inpcbport *, sizeof (struct inpcbport), + MALLOC(phd, struct inpcbport *, sizeof(struct inpcbport), M_PCB, M_WAITOK); if (phd == NULL) { - if (!locked) + if (!locked) { lck_rw_done(pcbinfo->ipi_lock); - return (ENOBUFS); /* XXX */ + } + return ENOBUFS; /* XXX */ } phd->phd_port = inp->inp_lport; LIST_INIT(&phd->phd_pcblist); @@ -2102,20 +2570,23 @@ in_pcbinshash(struct inpcb *inp, int locked) } VERIFY(!(inp->inp_flags2 & INP2_INHASHLIST)); + + inp->inp_phd = phd; LIST_INSERT_HEAD(&phd->phd_pcblist, inp, inp_portlist); LIST_INSERT_HEAD(pcbhash, inp, inp_hash); inp->inp_flags2 |= INP2_INHASHLIST; - if (!locked) + if (!locked) { lck_rw_done(pcbinfo->ipi_lock); - + } + #if NECP // This call catches the original setting of the local address inp_update_necp_policy(inp, NULL, NULL, 0); #endif /* NECP */ - - return (0); + + return 0; } /* @@ -2131,11 +2602,11 @@ in_pcbrehash(struct inpcb *inp) u_int32_t hashkey_faddr; #if INET6 - if (inp->inp_vflag & INP_IPV6) + if (inp->inp_vflag & INP_IPV6) { hashkey_faddr = inp->in6p_faddr.s6_addr32[3] /* XXX */; - else + } else #endif /* INET6 */ - hashkey_faddr = inp->inp_faddr.s_addr; + hashkey_faddr = inp->inp_faddr.s_addr; inp->inp_hash_element = INP_PCBHASH(hashkey_faddr, inp->inp_lport, inp->inp_fport, inp->inp_pcbinfo->ipi_hashmask); @@ -2149,7 +2620,7 @@ in_pcbrehash(struct inpcb *inp) VERIFY(!(inp->inp_flags2 & INP2_INHASHLIST)); LIST_INSERT_HEAD(head, inp, inp_hash); inp->inp_flags2 |= INP2_INHASHLIST; - + #if NECP // This call catches updates to the remote addresses inp_update_necp_policy(inp, NULL, NULL, 0); @@ -2167,7 +2638,7 @@ in_pcbremlists(struct inpcb *inp) /* * Check if it's in hashlist -- an inp is placed in hashlist when - * it's local port gets assigned. So it should also be present + * it's local port gets assigned. So it should also be present * in the port list. */ if (inp->inp_flags2 & INP2_INHASHLIST) { @@ -2203,7 +2674,7 @@ in_pcbremlists(struct inpcb *inp) } if (inp->inp_flags2 & INP2_IN_FCTREE) { - inp_fc_getinp(inp->inp_flowhash, (INPFC_SOLOCKED|INPFC_REMOVE)); + inp_fc_getinp(inp->inp_flowhash, (INPFC_SOLOCKED | INPFC_REMOVE)); VERIFY(!(inp->inp_flags2 & INP2_IN_FCTREE)); } @@ -2233,8 +2704,9 @@ in_pcb_checkstate(struct inpcb *pcb, int mode, int locked) * STOPUSING, if success we're good, if it's in use, will * be marked later */ - if (locked == 0) + if (locked == 0) { socket_lock(pcb->inp_socket, 1); + } pcb->inp_state = INPCB_STATE_DEAD; stopusing: @@ -2243,21 +2715,22 @@ stopusing: __func__, pcb, pcb->inp_socket); /* NOTREACHED */ } - if (locked == 0) + if (locked == 0) { socket_unlock(pcb->inp_socket, 1); + } inpcb_gc_sched(pcb->inp_pcbinfo, INPCB_TIMER_FAST); origwant = *wantcnt; - if ((UInt16) origwant == 0xffff) /* should stop using */ - return (WNT_STOPUSING); + if ((UInt16) origwant == 0xffff) { /* should stop using */ + return WNT_STOPUSING; + } newwant = 0xffff; if ((UInt16) origwant == 0) { /* try to mark it as unsuable now */ OSCompareAndSwap(origwant, newwant, wantcnt); } - return (WNT_STOPUSING); - break; + return WNT_STOPUSING; case WNT_ACQUIRE: /* @@ -2269,20 +2742,20 @@ stopusing: origwant = *wantcnt; if ((UInt16) origwant == 0xffff) { /* should stop using */ - return (WNT_STOPUSING); + return WNT_STOPUSING; } newwant = origwant + 1; } while (!OSCompareAndSwap(origwant, newwant, wantcnt)); - return (WNT_ACQUIRE); - break; + return WNT_ACQUIRE; case WNT_RELEASE: /* * Release reference. If result is null and pcb state * is DEAD, set wanted bit to STOPUSING */ - if (locked == 0) + if (locked == 0) { socket_lock(pcb->inp_socket, 1); + } do { origwant = *wantcnt; @@ -2293,25 +2766,27 @@ stopusing: } if ((UInt16) origwant == 0xffff) { /* should stop using */ - if (locked == 0) + if (locked == 0) { socket_unlock(pcb->inp_socket, 1); - return (WNT_STOPUSING); + } + return WNT_STOPUSING; } newwant = origwant - 1; } while (!OSCompareAndSwap(origwant, newwant, wantcnt)); - if (pcb->inp_state == INPCB_STATE_DEAD) + if (pcb->inp_state == INPCB_STATE_DEAD) { goto stopusing; + } if (pcb->inp_socket->so_usecount < 0) { panic("%s: RELEASE pcb=%p so=%p usecount is negative\n", __func__, pcb, pcb->inp_socket); /* NOTREACHED */ } - if (locked == 0) + if (locked == 0) { socket_unlock(pcb->inp_socket, 1); - return (WNT_RELEASE); - break; + } + return WNT_RELEASE; default: panic("%s: so=%p not a valid state =%x\n", __func__, @@ -2320,7 +2795,7 @@ stopusing: } /* NOTREACHED */ - return (mode); + return mode; } /* @@ -2331,7 +2806,7 @@ stopusing: void inpcb_to_compat(struct inpcb *inp, struct inpcb_compat *inp_compat) { - bzero(inp_compat, sizeof (*inp_compat)); + bzero(inp_compat, sizeof(*inp_compat)); inp_compat->inp_fport = inp->inp_fport; inp_compat->inp_lport = inp->inp_lport; inp_compat->nat_owner = 0; @@ -2353,6 +2828,7 @@ inpcb_to_compat(struct inpcb *inp, struct inpcb_compat *inp_compat) inp_compat->inp_depend6.inp6_hops = inp->inp_depend6.inp6_hops; } +#if !CONFIG_EMBEDDED void inpcb_to_xinpcb64(struct inpcb *inp, struct xinpcb64 *xinp) { @@ -2372,6 +2848,7 @@ inpcb_to_xinpcb64(struct inpcb *inp, struct xinpcb64 *xinp) xinp->inp_depend6.inp6_ifindex = 0; xinp->inp_depend6.inp6_hops = inp->inp_depend6.inp6_hops; } +#endif /* !CONFIG_EMBEDDED */ /* * The following routines implement this scheme: @@ -2404,16 +2881,17 @@ inp_route_copyout(struct inpcb *inp, struct route *dst) { struct route *src = &inp->inp_route; - lck_mtx_assert(&inp->inpcb_mtx, LCK_MTX_ASSERT_OWNED); + socket_lock_assert_owned(inp->inp_socket); /* * If the route in the PCB is stale or not for IPv4, blow it away; * this is possible in the case of IPv4-mapped address case. */ - if (ROUTE_UNUSABLE(src) || rt_key(src->ro_rt)->sa_family != AF_INET) + if (ROUTE_UNUSABLE(src) || rt_key(src->ro_rt)->sa_family != AF_INET) { ROUTE_RELEASE(src); + } - route_copyout(dst, src, sizeof (*dst)); + route_copyout(dst, src, sizeof(*dst)); } void @@ -2421,17 +2899,18 @@ inp_route_copyin(struct inpcb *inp, struct route *src) { struct route *dst = &inp->inp_route; - lck_mtx_assert(&inp->inpcb_mtx, LCK_MTX_ASSERT_OWNED); + socket_lock_assert_owned(inp->inp_socket); /* Minor sanity check */ - if (src->ro_rt != NULL && rt_key(src->ro_rt)->sa_family != AF_INET) + if (src->ro_rt != NULL && rt_key(src->ro_rt)->sa_family != AF_INET) { panic("%s: wrong or corrupted route: %p", __func__, src); + } - route_copyin(src, dst, sizeof (*src)); + route_copyin(src, dst, sizeof(*src)); } /* - * Handler for setting IP_FORCE_OUT_IFP/IP_BOUND_IF/IPV6_BOUND_IF socket option. + * Handler for setting IP_BOUND_IF/IPV6_BOUND_IF socket option. */ int inp_bindif(struct inpcb *inp, unsigned int ifscope, struct ifnet **pifp) @@ -2442,7 +2921,7 @@ inp_bindif(struct inpcb *inp, unsigned int ifscope, struct ifnet **pifp) if ((ifscope > (unsigned)if_index) || (ifscope != IFSCOPE_NONE && (ifp = ifindex2ifnet[ifscope]) == NULL)) { ifnet_head_done(); - return (ENXIO); + return ENXIO; } ifnet_head_done(); @@ -2458,18 +2937,20 @@ inp_bindif(struct inpcb *inp, unsigned int ifscope, struct ifnet **pifp) * exact match for the embedded interface scope. */ inp->inp_boundifp = ifp; - if (inp->inp_boundifp == NULL) + if (inp->inp_boundifp == NULL) { inp->inp_flags &= ~INP_BOUND_IF; - else + } else { inp->inp_flags |= INP_BOUND_IF; + } /* Blow away any cached route in the PCB */ ROUTE_RELEASE(&inp->inp_route); - if (pifp != NULL) + if (pifp != NULL) { *pifp = ifp; + } - return (0); + return 0; } /* @@ -2516,6 +2997,15 @@ inp_set_noexpensive(struct inpcb *inp) ROUTE_RELEASE(&inp->inp_route); } +void +inp_set_noconstrained(struct inpcb *inp) +{ + inp->inp_flags2 |= INP2_NO_IFF_CONSTRAINED; + + /* Blow away any cached route in the PCB */ + ROUTE_RELEASE(&inp->inp_route); +} + void inp_set_awdl_unrestricted(struct inpcb *inp) { @@ -2540,6 +3030,30 @@ inp_clear_awdl_unrestricted(struct inpcb *inp) ROUTE_RELEASE(&inp->inp_route); } +void +inp_set_intcoproc_allowed(struct inpcb *inp) +{ + inp->inp_flags2 |= INP2_INTCOPROC_ALLOWED; + + /* Blow away any cached route in the PCB */ + ROUTE_RELEASE(&inp->inp_route); +} + +boolean_t +inp_get_intcoproc_allowed(struct inpcb *inp) +{ + return (inp->inp_flags2 & INP2_INTCOPROC_ALLOWED) ? TRUE : FALSE; +} + +void +inp_clear_intcoproc_allowed(struct inpcb *inp) +{ + inp->inp_flags2 &= ~INP2_INTCOPROC_ALLOWED; + + /* Blow away any cached route in the PCB */ + ROUTE_RELEASE(&inp->inp_route); +} + #if NECP /* * Called when PROC_UUID_NECP_APP_POLICY is set. @@ -2572,13 +3086,14 @@ inp_calc_flowhash(struct inpcb *inp) u_int32_t flowhash = 0; struct inpcb *tmp_inp = NULL; - if (inp_hash_seed == 0) + if (inp_hash_seed == 0) { inp_hash_seed = RandomULong(); + } - bzero(&fh, sizeof (fh)); + bzero(&fh, sizeof(fh)); - bcopy(&inp->inp_dependladdr, &fh.infh_laddr, sizeof (fh.infh_laddr)); - bcopy(&inp->inp_dependfaddr, &fh.infh_faddr, sizeof (fh.infh_faddr)); + bcopy(&inp->inp_dependladdr, &fh.infh_laddr, sizeof(fh.infh_laddr)); + bcopy(&inp->inp_dependfaddr, &fh.infh_faddr, sizeof(fh.infh_faddr)); fh.infh_lport = inp->inp_lport; fh.infh_fport = inp->inp_fport; @@ -2588,7 +3103,7 @@ inp_calc_flowhash(struct inpcb *inp) fh.infh_rand2 = RandomULong(); try_again: - flowhash = net_flowhash(&fh, sizeof (fh), inp_hash_seed); + flowhash = net_flowhash(&fh, sizeof(fh), inp_hash_seed); if (flowhash == 0) { /* try to get a non-zero flowhash */ inp_hash_seed = RandomULong(); @@ -2617,7 +3132,7 @@ try_again: inp->inp_flags2 |= INP2_IN_FCTREE; lck_mtx_unlock(&inp_fc_lck); - return (flowhash); + return flowhash; } void @@ -2627,8 +3142,9 @@ inp_flowadv(uint32_t flowhash) inp = inp_fc_getinp(flowhash, 0); - if (inp == NULL) + if (inp == NULL) { return; + } inp_fc_feedback(inp); } @@ -2638,8 +3154,8 @@ inp_flowadv(uint32_t flowhash) static inline int infc_cmp(const struct inpcb *inp1, const struct inpcb *inp2) { - return (memcmp(&(inp1->inp_flowhash), &(inp2->inp_flowhash), - sizeof(inp1->inp_flowhash))); + return memcmp(&(inp1->inp_flowhash), &(inp2->inp_flowhash), + sizeof(inp1->inp_flowhash)); } static struct inpcb * @@ -2654,23 +3170,24 @@ inp_fc_getinp(u_int32_t flowhash, u_int32_t flags) if (inp == NULL) { /* inp is not present, return */ lck_mtx_unlock(&inp_fc_lck); - return (NULL); + return NULL; } if (flags & INPFC_REMOVE) { RB_REMOVE(inp_fc_tree, &inp_fc_tree, inp); lck_mtx_unlock(&inp_fc_lck); - bzero(&(inp->infc_link), sizeof (inp->infc_link)); + bzero(&(inp->infc_link), sizeof(inp->infc_link)); inp->inp_flags2 &= ~INP2_IN_FCTREE; - return (NULL); + return NULL; } - if (in_pcb_checkstate(inp, WNT_ACQUIRE, locked) == WNT_STOPUSING) + if (in_pcb_checkstate(inp, WNT_ACQUIRE, locked) == WNT_STOPUSING) { inp = NULL; + } lck_mtx_unlock(&inp_fc_lck); - return (inp); + return inp; } static void @@ -2687,8 +3204,9 @@ inp_fc_feedback(struct inpcb *inp) return; } - if (inp->inp_sndinprog_cnt > 0) + if (inp->inp_sndinprog_cnt > 0) { inp->inp_flags |= INP_FC_FEEDBACK; + } /* * Return if the connection is not in flow-controlled state. @@ -2701,8 +3219,9 @@ inp_fc_feedback(struct inpcb *inp) } inp_reset_fc_state(inp); - if (SOCK_TYPE(so) == SOCK_STREAM) + if (SOCK_TYPE(so) == SOCK_STREAM) { inp_fc_unthrottle_tcp(inp); + } socket_unlock(so, 1); } @@ -2722,8 +3241,9 @@ inp_reset_fc_state(struct inpcb *inp) } /* Give a write wakeup to unblock the socket */ - if (needwakeup) + if (needwakeup) { sowwakeup(so); + } } int @@ -2737,14 +3257,16 @@ inp_set_fc_state(struct inpcb *inp, int advcode) * flow controlled state and receiving feedback from * the interface */ - if (inp->inp_flags & INP_FC_FEEDBACK) - return (0); + if (inp->inp_flags & INP_FC_FEEDBACK) { + return 0; + } inp->inp_flags &= ~(INP_FLOW_CONTROLLED | INP_FLOW_SUSPENDED); if ((tmp_inp = inp_fc_getinp(inp->inp_flowhash, INPFC_SOLOCKED)) != NULL) { - if (in_pcb_checkstate(tmp_inp, WNT_RELEASE, 1) == WNT_STOPUSING) - return (0); + if (in_pcb_checkstate(tmp_inp, WNT_RELEASE, 1) == WNT_STOPUSING) { + return 0; + } VERIFY(tmp_inp == inp); switch (advcode) { case FADV_FLOW_CONTROLLED: @@ -2759,9 +3281,9 @@ inp_set_fc_state(struct inpcb *inp, int advcode) inp->inp_socket->so_flags |= SOF_SUSPENDED; break; } - return (1); + return 1; } - return (0); + return 0; } /* @@ -2774,24 +3296,28 @@ inp_flush(struct inpcb *inp, int optval) struct ifnet *rtifp, *oifp; /* Either all classes or one of the valid ones */ - if (optval != SO_TC_ALL && !SO_VALID_TC(optval)) - return (EINVAL); + if (optval != SO_TC_ALL && !SO_VALID_TC(optval)) { + return EINVAL; + } /* We need a flow hash for identification */ - if (flowhash == 0) - return (0); + if (flowhash == 0) { + return 0; + } /* Grab the interfaces from the route and pcb */ rtifp = ((inp->inp_route.ro_rt != NULL) ? inp->inp_route.ro_rt->rt_ifp : NULL); oifp = inp->inp_last_outifp; - if (rtifp != NULL) + if (rtifp != NULL) { if_qflush_sc(rtifp, so_tc2msc(optval), flowhash, NULL, NULL, 0); - if (oifp != NULL && oifp != rtifp) + } + if (oifp != NULL && oifp != rtifp) { if_qflush_sc(oifp, so_tc2msc(optval), flowhash, NULL, NULL, 0); + } - return (0); + return 0; } /* @@ -2816,18 +3342,24 @@ inp_get_soprocinfo(struct inpcb *inp, struct so_procinfo *soprocinfo) struct socket *so = inp->inp_socket; soprocinfo->spi_pid = so->last_pid; - if (so->last_pid != 0) + strlcpy(&soprocinfo->spi_proc_name[0], &inp->inp_last_proc_name[0], + sizeof(soprocinfo->spi_proc_name)); + if (so->last_pid != 0) { uuid_copy(soprocinfo->spi_uuid, so->last_uuid); + } /* * When not delegated, the effective pid is the same as the real pid */ if (so->so_flags & SOF_DELEGATED) { + soprocinfo->spi_delegated = 1; soprocinfo->spi_epid = so->e_pid; - if (so->e_pid != 0) - uuid_copy(soprocinfo->spi_euuid, so->e_uuid); + uuid_copy(soprocinfo->spi_euuid, so->e_uuid); } else { + soprocinfo->spi_delegated = 0; soprocinfo->spi_epid = so->last_pid; } + strlcpy(&soprocinfo->spi_e_proc_name[0], &inp->inp_e_proc_name[0], + sizeof(soprocinfo->spi_e_proc_name)); } int @@ -2837,10 +3369,11 @@ inp_findinpcb_procinfo(struct inpcbinfo *pcbinfo, uint32_t flowhash, struct inpcb *inp = NULL; int found = 0; - bzero(soprocinfo, sizeof (struct so_procinfo)); + bzero(soprocinfo, sizeof(struct so_procinfo)); - if (!flowhash) - return (-1); + if (!flowhash) { + return -1; + } lck_rw_lock_shared(pcbinfo->ipi_lock); LIST_FOREACH(inp, pcbinfo->ipi_listhead, inp_list) { @@ -2854,7 +3387,7 @@ inp_findinpcb_procinfo(struct inpcbinfo *pcbinfo, uint32_t flowhash, } lck_rw_done(pcbinfo->ipi_lock); - return (found); + return found; } #if CONFIG_PROC_UUID_POLICY @@ -2952,9 +3485,9 @@ inp_update_necp_policy(struct inpcb *inp, struct sockaddr *override_local_addr, { necp_socket_find_policy_match(inp, override_local_addr, override_remote_addr, override_bound_interface); if (necp_socket_should_rescope(inp) && - inp->inp_lport == 0 && - inp->inp_laddr.s_addr == INADDR_ANY && - IN6_IS_ADDR_UNSPECIFIED(&inp->in6p_laddr)) { + inp->inp_lport == 0 && + inp->inp_laddr.s_addr == INADDR_ANY && + IN6_IS_ADDR_UNSPECIFIED(&inp->in6p_laddr)) { // If we should rescope, and the socket is not yet bound inp_bindif(inp, necp_socket_get_rescope_if_index(inp), NULL); } @@ -2969,28 +3502,40 @@ inp_update_policy(struct inpcb *inp) uint32_t pflags = 0; int32_t ogencnt; int err = 0; + uint8_t *lookup_uuid = NULL; if (!net_io_policy_uuid || - so == NULL || inp->inp_state == INPCB_STATE_DEAD) - return (0); + so == NULL || inp->inp_state == INPCB_STATE_DEAD) { + return 0; + } /* * Kernel-created sockets that aren't delegating other sockets * are currently exempted from UUID policy checks. */ - if (so->last_pid == 0 && !(so->so_flags & SOF_DELEGATED)) - return (0); + if (so->last_pid == 0 && !(so->so_flags & SOF_DELEGATED)) { + return 0; + } + +#if defined(XNU_TARGET_OS_OSX) + if (so->so_rpid > 0) { + lookup_uuid = so->so_ruuid; + } +#endif + if (lookup_uuid == NULL) { + lookup_uuid = ((so->so_flags & SOF_DELEGATED) ? so->e_uuid : so->last_uuid); + } ogencnt = so->so_policy_gencnt; - err = proc_uuid_policy_lookup(((so->so_flags & SOF_DELEGATED) ? - so->e_uuid : so->last_uuid), &pflags, &so->so_policy_gencnt); + err = proc_uuid_policy_lookup(lookup_uuid, &pflags, &so->so_policy_gencnt); /* * Discard cached generation count if the entry is gone (ENOENT), * so that we go thru the checks below. */ - if (err == ENOENT && ogencnt != 0) + if (err == ENOENT && ogencnt != 0) { so->so_policy_gencnt = 0; + } /* * If the generation count has changed, inspect the policy flags @@ -3015,50 +3560,86 @@ inp_update_policy(struct inpcb *inp) #endif /* NECP */ } - return ((err == ENOENT) ? 0 : err); + return (err == ENOENT) ? 0 : err; #else /* !CONFIG_PROC_UUID_POLICY */ #pragma unused(inp) - return (0); + return 0; #endif /* !CONFIG_PROC_UUID_POLICY */ } + +static unsigned int log_restricted; +SYSCTL_DECL(_net_inet); +SYSCTL_INT(_net_inet, OID_AUTO, log_restricted, + CTLFLAG_RW | CTLFLAG_LOCKED, &log_restricted, 0, + "Log network restrictions"); /* * Called when we need to enforce policy restrictions in the input path. * * Returns TRUE if we're not allowed to receive data, otherwise FALSE. */ -boolean_t -inp_restricted_recv(struct inpcb *inp, struct ifnet *ifp) +static boolean_t +_inp_restricted_recv(struct inpcb *inp, struct ifnet *ifp) { VERIFY(inp != NULL); /* * Inbound restrictions. */ - if (!sorestrictrecv) - return (FALSE); + if (!sorestrictrecv) { + return FALSE; + } - if (ifp == NULL) - return (FALSE); + if (ifp == NULL) { + return FALSE; + } + + if (IFNET_IS_CELLULAR(ifp) && INP_NO_CELLULAR(inp)) { + return TRUE; + } + + if (IFNET_IS_EXPENSIVE(ifp) && INP_NO_EXPENSIVE(inp)) { + return TRUE; + } + + if (IFNET_IS_CONSTRAINED(ifp) && INP_NO_CONSTRAINED(inp)) { + return TRUE; + } + + if (IFNET_IS_AWDL_RESTRICTED(ifp) && !INP_AWDL_UNRESTRICTED(inp)) { + return TRUE; + } + + if (!(ifp->if_eflags & IFEF_RESTRICTED_RECV)) { + return FALSE; + } - if (IFNET_IS_CELLULAR(ifp) && INP_NO_CELLULAR(inp)) - return (TRUE); + if (inp->inp_flags & INP_RECV_ANYIF) { + return FALSE; + } - if (IFNET_IS_EXPENSIVE(ifp) && INP_NO_EXPENSIVE(inp)) - return (TRUE); + if ((inp->inp_flags & INP_BOUND_IF) && inp->inp_boundifp == ifp) { + return FALSE; + } - if (IFNET_IS_AWDL_RESTRICTED(ifp) && !INP_AWDL_UNRESTRICTED(inp)) - return (TRUE); - - if (!(ifp->if_eflags & IFEF_RESTRICTED_RECV)) - return (FALSE); + if (IFNET_IS_INTCOPROC(ifp) && !INP_INTCOPROC_ALLOWED(inp)) { + return TRUE; + } - if (inp->inp_flags & INP_RECV_ANYIF) - return (FALSE); + return TRUE; +} - if ((inp->inp_flags & INP_BOUND_IF) && inp->inp_boundifp == ifp) - return (FALSE); +boolean_t +inp_restricted_recv(struct inpcb *inp, struct ifnet *ifp) +{ + boolean_t ret; - return (TRUE); + ret = _inp_restricted_recv(inp, ifp); + if (ret == TRUE && log_restricted) { + printf("pid %d (%s) is unable to receive packets on %s\n", + current_proc()->p_pid, proc_best_name(current_proc()), + ifp->if_xname); + } + return ret; } /* @@ -3066,28 +3647,199 @@ inp_restricted_recv(struct inpcb *inp, struct ifnet *ifp) * * Returns TRUE if we're not allowed to send data out, otherwise FALSE. */ -boolean_t -inp_restricted_send(struct inpcb *inp, struct ifnet *ifp) +static boolean_t +_inp_restricted_send(struct inpcb *inp, struct ifnet *ifp) { VERIFY(inp != NULL); /* * Outbound restrictions. */ - if (!sorestrictsend) - return (FALSE); + if (!sorestrictsend) { + return FALSE; + } + + if (ifp == NULL) { + return FALSE; + } + + if (IFNET_IS_CELLULAR(ifp) && INP_NO_CELLULAR(inp)) { + return TRUE; + } + + if (IFNET_IS_EXPENSIVE(ifp) && INP_NO_EXPENSIVE(inp)) { + return TRUE; + } + + if (IFNET_IS_CONSTRAINED(ifp) && INP_NO_CONSTRAINED(inp)) { + return TRUE; + } + + if (IFNET_IS_AWDL_RESTRICTED(ifp) && !INP_AWDL_UNRESTRICTED(inp)) { + return TRUE; + } - if (ifp == NULL) - return (FALSE); + if (IFNET_IS_INTCOPROC(ifp) && !INP_INTCOPROC_ALLOWED(inp)) { + return TRUE; + } + + return FALSE; +} + +boolean_t +inp_restricted_send(struct inpcb *inp, struct ifnet *ifp) +{ + boolean_t ret; + + ret = _inp_restricted_send(inp, ifp); + if (ret == TRUE && log_restricted) { + printf("pid %d (%s) is unable to transmit packets on %s\n", + current_proc()->p_pid, proc_best_name(current_proc()), + ifp->if_xname); + } + return ret; +} + +inline void +inp_count_sndbytes(struct inpcb *inp, u_int32_t th_ack) +{ + struct ifnet *ifp = inp->inp_last_outifp; + struct socket *so = inp->inp_socket; + if (ifp != NULL && !(so->so_flags & SOF_MP_SUBFLOW) && + (ifp->if_type == IFT_CELLULAR || IFNET_IS_WIFI(ifp))) { + int32_t unsent; - if (IFNET_IS_CELLULAR(ifp) && INP_NO_CELLULAR(inp)) - return (TRUE); + so->so_snd.sb_flags |= SB_SNDBYTE_CNT; - if (IFNET_IS_EXPENSIVE(ifp) && INP_NO_EXPENSIVE(inp)) - return (TRUE); + /* + * There can be data outstanding before the connection + * becomes established -- TFO case + */ + if (so->so_snd.sb_cc > 0) { + inp_incr_sndbytes_total(so, so->so_snd.sb_cc); + } - if (IFNET_IS_AWDL_RESTRICTED(ifp) && !INP_AWDL_UNRESTRICTED(inp)) - return (TRUE); + unsent = inp_get_sndbytes_allunsent(so, th_ack); + if (unsent > 0) { + inp_incr_sndbytes_unsent(so, unsent); + } + } +} + +inline void +inp_incr_sndbytes_total(struct socket *so, int32_t len) +{ + struct inpcb *inp = (struct inpcb *)so->so_pcb; + struct ifnet *ifp = inp->inp_last_outifp; + + if (ifp != NULL) { + VERIFY(ifp->if_sndbyte_total >= 0); + OSAddAtomic64(len, &ifp->if_sndbyte_total); + } +} + +inline void +inp_decr_sndbytes_total(struct socket *so, int32_t len) +{ + struct inpcb *inp = (struct inpcb *)so->so_pcb; + struct ifnet *ifp = inp->inp_last_outifp; + + if (ifp != NULL) { + VERIFY(ifp->if_sndbyte_total >= len); + OSAddAtomic64(-len, &ifp->if_sndbyte_total); + } +} + +inline void +inp_incr_sndbytes_unsent(struct socket *so, int32_t len) +{ + struct inpcb *inp = (struct inpcb *)so->so_pcb; + struct ifnet *ifp = inp->inp_last_outifp; + + if (ifp != NULL) { + VERIFY(ifp->if_sndbyte_unsent >= 0); + OSAddAtomic64(len, &ifp->if_sndbyte_unsent); + } +} + +inline void +inp_decr_sndbytes_unsent(struct socket *so, int32_t len) +{ + if (so == NULL || !(so->so_snd.sb_flags & SB_SNDBYTE_CNT)) { + return; + } + + struct inpcb *inp = (struct inpcb *)so->so_pcb; + struct ifnet *ifp = inp->inp_last_outifp; + + if (ifp != NULL) { + if (ifp->if_sndbyte_unsent >= len) { + OSAddAtomic64(-len, &ifp->if_sndbyte_unsent); + } else { + ifp->if_sndbyte_unsent = 0; + } + } +} + +inline void +inp_decr_sndbytes_allunsent(struct socket *so, u_int32_t th_ack) +{ + int32_t len; + + if (so == NULL || !(so->so_snd.sb_flags & SB_SNDBYTE_CNT)) { + return; + } + + len = inp_get_sndbytes_allunsent(so, th_ack); + inp_decr_sndbytes_unsent(so, len); +} + + +inline void +inp_set_activity_bitmap(struct inpcb *inp) +{ + in_stat_set_activity_bitmap(&inp->inp_nw_activity, net_uptime()); +} + +inline void +inp_get_activity_bitmap(struct inpcb *inp, activity_bitmap_t *ab) +{ + bcopy(&inp->inp_nw_activity, ab, sizeof(*ab)); +} + +void +inp_update_last_owner(struct socket *so, struct proc *p, struct proc *ep) +{ + struct inpcb *inp = (struct inpcb *)so->so_pcb; + + if (inp == NULL) { + return; + } + + if (p != NULL) { + strlcpy(&inp->inp_last_proc_name[0], proc_name_address(p), sizeof(inp->inp_last_proc_name)); + } + if (so->so_flags & SOF_DELEGATED) { + if (ep != NULL) { + strlcpy(&inp->inp_e_proc_name[0], proc_name_address(ep), sizeof(inp->inp_e_proc_name)); + } else { + inp->inp_e_proc_name[0] = 0; + } + } else { + inp->inp_e_proc_name[0] = 0; + } +} + +void +inp_copy_last_owner(struct socket *so, struct socket *head) +{ + struct inpcb *inp = (struct inpcb *)so->so_pcb; + struct inpcb *head_inp = (struct inpcb *)head->so_pcb; + + if (inp == NULL || head_inp == NULL) { + return; + } - return (FALSE); + strlcpy(&inp->inp_last_proc_name[0], &head_inp->inp_last_proc_name[0], sizeof(inp->inp_last_proc_name)); + strlcpy(&inp->inp_e_proc_name[0], &head_inp->inp_e_proc_name[0], sizeof(inp->inp_e_proc_name)); }