2 * Copyright (c) 2000-2016 Apple Inc. All rights reserved.
4 * @APPLE_OSREFERENCE_LICENSE_HEADER_START@
6 * This file contains Original Code and/or Modifications of Original Code
7 * as defined in and that are subject to the Apple Public Source License
8 * Version 2.0 (the 'License'). You may not use this file except in
9 * compliance with the License. The rights granted to you under the License
10 * may not be used to create, or enable the creation or redistribution of,
11 * unlawful or unlicensed copies of an Apple operating system, or to
12 * circumvent, violate, or enable the circumvention or violation of, any
13 * terms of an Apple operating system software license agreement.
15 * Please obtain a copy of the License at
16 * http://www.opensource.apple.com/apsl/ and read it before using this file.
18 * The Original Code and all software distributed under the License are
19 * distributed on an 'AS IS' basis, WITHOUT WARRANTY OF ANY KIND, EITHER
20 * EXPRESS OR IMPLIED, AND APPLE HEREBY DISCLAIMS ALL SUCH WARRANTIES,
21 * INCLUDING WITHOUT LIMITATION, ANY WARRANTIES OF MERCHANTABILITY,
22 * FITNESS FOR A PARTICULAR PURPOSE, QUIET ENJOYMENT OR NON-INFRINGEMENT.
23 * Please see the License for the specific language governing rights and
24 * limitations under the License.
26 * @APPLE_OSREFERENCE_LICENSE_HEADER_END@
29 * Copyright (c) 1982, 1986, 1991, 1993, 1995
30 * The Regents of the University of California. All rights reserved.
32 * Redistribution and use in source and binary forms, with or without
33 * modification, are permitted provided that the following conditions
35 * 1. Redistributions of source code must retain the above copyright
36 * notice, this list of conditions and the following disclaimer.
37 * 2. Redistributions in binary form must reproduce the above copyright
38 * notice, this list of conditions and the following disclaimer in the
39 * documentation and/or other materials provided with the distribution.
40 * 3. All advertising materials mentioning features or use of this software
41 * must display the following acknowledgement:
42 * This product includes software developed by the University of
43 * California, Berkeley and its contributors.
44 * 4. Neither the name of the University nor the names of its contributors
45 * may be used to endorse or promote products derived from this software
46 * without specific prior written permission.
48 * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
49 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
50 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
51 * ARE DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
52 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
53 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
54 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
55 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
56 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
57 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
60 * @(#)in_pcb.c 8.4 (Berkeley) 5/24/95
61 * $FreeBSD: src/sys/netinet/in_pcb.c,v 1.59.2.17 2001/08/13 16:26:17 ume Exp $
64 #include <sys/param.h>
65 #include <sys/systm.h>
66 #include <sys/malloc.h>
68 #include <sys/domain.h>
69 #include <sys/protosw.h>
70 #include <sys/socket.h>
71 #include <sys/socketvar.h>
73 #include <sys/kernel.h>
74 #include <sys/sysctl.h>
75 #include <sys/mcache.h>
76 #include <sys/kauth.h>
78 #include <sys/proc_uuid_policy.h>
79 #include <sys/syslog.h>
83 #include <libkern/OSAtomic.h>
84 #include <kern/locks.h>
86 #include <machine/limits.h>
88 #include <kern/zalloc.h>
91 #include <net/if_types.h>
92 #include <net/route.h>
93 #include <net/flowhash.h>
94 #include <net/flowadv.h>
95 #include <net/ntstat.h>
97 #include <netinet/in.h>
98 #include <netinet/in_pcb.h>
99 #include <netinet/in_var.h>
100 #include <netinet/ip_var.h>
102 #include <netinet/ip6.h>
103 #include <netinet6/ip6_var.h>
106 #include <sys/kdebug.h>
107 #include <sys/random.h>
109 #include <dev/random/randomdev.h>
110 #include <mach/boolean.h>
112 #include <pexpert/pexpert.h>
115 #include <net/necp.h>
118 #include <sys/stat.h>
120 #include <sys/vnode.h>
122 static lck_grp_t
*inpcb_lock_grp
;
123 static lck_attr_t
*inpcb_lock_attr
;
124 static lck_grp_attr_t
*inpcb_lock_grp_attr
;
125 decl_lck_mtx_data(static, inpcb_lock
); /* global INPCB lock */
126 decl_lck_mtx_data(static, inpcb_timeout_lock
);
128 static TAILQ_HEAD(, inpcbinfo
) inpcb_head
= TAILQ_HEAD_INITIALIZER(inpcb_head
);
130 static u_int16_t inpcb_timeout_run
= 0; /* INPCB timer is scheduled to run */
131 static boolean_t inpcb_garbage_collecting
= FALSE
; /* gc timer is scheduled */
132 static boolean_t inpcb_ticking
= FALSE
; /* "slow" timer is scheduled */
133 static boolean_t inpcb_fast_timer_on
= FALSE
;
134 static boolean_t intcoproc_unrestricted
= FALSE
;
136 extern char *proc_best_name(proc_t
);
139 * If the total number of gc reqs is above a threshold, schedule
140 * garbage collect timer sooner
142 static boolean_t inpcb_toomany_gcreq
= FALSE
;
144 #define INPCB_GCREQ_THRESHOLD 50000
146 static thread_call_t inpcb_thread_call
, inpcb_fast_thread_call
;
147 static void inpcb_sched_timeout(void);
148 static void inpcb_sched_lazy_timeout(void);
149 static void _inpcb_sched_timeout(unsigned int);
150 static void inpcb_timeout(void *, void *);
151 const int inpcb_timeout_lazy
= 10; /* 10 seconds leeway for lazy timers */
152 extern int tvtohz(struct timeval
*);
154 #if CONFIG_PROC_UUID_POLICY
155 static void inp_update_cellular_policy(struct inpcb
*, boolean_t
);
157 static void inp_update_necp_want_app_policy(struct inpcb
*, boolean_t
);
159 #endif /* !CONFIG_PROC_UUID_POLICY */
161 #define DBG_FNC_PCB_LOOKUP NETDBG_CODE(DBG_NETTCP, (6 << 8))
162 #define DBG_FNC_PCB_HLOOKUP NETDBG_CODE(DBG_NETTCP, ((6 << 8) | 1))
165 * These configure the range of local port addresses assigned to
166 * "unspecified" outgoing connections/packets/whatever.
168 int ipport_lowfirstauto
= IPPORT_RESERVED
- 1; /* 1023 */
169 int ipport_lowlastauto
= IPPORT_RESERVEDSTART
; /* 600 */
170 int ipport_firstauto
= IPPORT_HIFIRSTAUTO
; /* 49152 */
171 int ipport_lastauto
= IPPORT_HILASTAUTO
; /* 65535 */
172 int ipport_hifirstauto
= IPPORT_HIFIRSTAUTO
; /* 49152 */
173 int ipport_hilastauto
= IPPORT_HILASTAUTO
; /* 65535 */
175 #define RANGECHK(var, min, max) \
176 if ((var) < (min)) { (var) = (min); } \
177 else if ((var) > (max)) { (var) = (max); }
180 sysctl_net_ipport_check SYSCTL_HANDLER_ARGS
182 #pragma unused(arg1, arg2)
185 error
= sysctl_handle_int(oidp
, oidp
->oid_arg1
, oidp
->oid_arg2
, req
);
187 RANGECHK(ipport_lowfirstauto
, 1, IPPORT_RESERVED
- 1);
188 RANGECHK(ipport_lowlastauto
, 1, IPPORT_RESERVED
- 1);
189 RANGECHK(ipport_firstauto
, IPPORT_RESERVED
, USHRT_MAX
);
190 RANGECHK(ipport_lastauto
, IPPORT_RESERVED
, USHRT_MAX
);
191 RANGECHK(ipport_hifirstauto
, IPPORT_RESERVED
, USHRT_MAX
);
192 RANGECHK(ipport_hilastauto
, IPPORT_RESERVED
, USHRT_MAX
);
199 SYSCTL_NODE(_net_inet_ip
, IPPROTO_IP
, portrange
,
200 CTLFLAG_RW
|CTLFLAG_LOCKED
, 0, "IP Ports");
202 SYSCTL_PROC(_net_inet_ip_portrange
, OID_AUTO
, lowfirst
,
203 CTLTYPE_INT
|CTLFLAG_RW
| CTLFLAG_LOCKED
,
204 &ipport_lowfirstauto
, 0, &sysctl_net_ipport_check
, "I", "");
205 SYSCTL_PROC(_net_inet_ip_portrange
, OID_AUTO
, lowlast
,
206 CTLTYPE_INT
|CTLFLAG_RW
| CTLFLAG_LOCKED
,
207 &ipport_lowlastauto
, 0, &sysctl_net_ipport_check
, "I", "");
208 SYSCTL_PROC(_net_inet_ip_portrange
, OID_AUTO
, first
,
209 CTLTYPE_INT
|CTLFLAG_RW
| CTLFLAG_LOCKED
,
210 &ipport_firstauto
, 0, &sysctl_net_ipport_check
, "I", "");
211 SYSCTL_PROC(_net_inet_ip_portrange
, OID_AUTO
, last
,
212 CTLTYPE_INT
|CTLFLAG_RW
| CTLFLAG_LOCKED
,
213 &ipport_lastauto
, 0, &sysctl_net_ipport_check
, "I", "");
214 SYSCTL_PROC(_net_inet_ip_portrange
, OID_AUTO
, hifirst
,
215 CTLTYPE_INT
|CTLFLAG_RW
| CTLFLAG_LOCKED
,
216 &ipport_hifirstauto
, 0, &sysctl_net_ipport_check
, "I", "");
217 SYSCTL_PROC(_net_inet_ip_portrange
, OID_AUTO
, hilast
,
218 CTLTYPE_INT
|CTLFLAG_RW
| CTLFLAG_LOCKED
,
219 &ipport_hilastauto
, 0, &sysctl_net_ipport_check
, "I", "");
221 static uint32_t apn_fallbk_debug
= 0;
222 #define apn_fallbk_log(x) do { if (apn_fallbk_debug >= 1) log x; } while (0)
224 static boolean_t apn_fallbk_enabled
= FALSE
;
226 extern int udp_use_randomport
;
227 extern int tcp_use_randomport
;
229 /* Structs used for flowhash computation */
230 struct inp_flowhash_key_addr
{
240 struct inp_flowhash_key
{
241 struct inp_flowhash_key_addr infh_laddr
;
242 struct inp_flowhash_key_addr infh_faddr
;
243 u_int32_t infh_lport
;
244 u_int32_t infh_fport
;
246 u_int32_t infh_proto
;
247 u_int32_t infh_rand1
;
248 u_int32_t infh_rand2
;
251 static u_int32_t inp_hash_seed
= 0;
253 static int infc_cmp(const struct inpcb
*, const struct inpcb
*);
255 /* Flags used by inp_fc_getinp */
256 #define INPFC_SOLOCKED 0x1
257 #define INPFC_REMOVE 0x2
258 static struct inpcb
*inp_fc_getinp(u_int32_t
, u_int32_t
);
260 static void inp_fc_feedback(struct inpcb
*);
261 extern void tcp_remove_from_time_wait(struct inpcb
*inp
);
263 decl_lck_mtx_data(static, inp_fc_lck
);
265 RB_HEAD(inp_fc_tree
, inpcb
) inp_fc_tree
;
266 RB_PROTOTYPE(inp_fc_tree
, inpcb
, infc_link
, infc_cmp
);
267 RB_GENERATE(inp_fc_tree
, inpcb
, infc_link
, infc_cmp
);
270 * Use this inp as a key to find an inp in the flowhash tree.
271 * Accesses to it are protected by inp_fc_lck.
273 struct inpcb key_inp
;
276 * in_pcb.c: manage the Protocol Control Blocks.
282 static int inpcb_initialized
= 0;
284 VERIFY(!inpcb_initialized
);
285 inpcb_initialized
= 1;
287 inpcb_lock_grp_attr
= lck_grp_attr_alloc_init();
288 inpcb_lock_grp
= lck_grp_alloc_init("inpcb", inpcb_lock_grp_attr
);
289 inpcb_lock_attr
= lck_attr_alloc_init();
290 lck_mtx_init(&inpcb_lock
, inpcb_lock_grp
, inpcb_lock_attr
);
291 lck_mtx_init(&inpcb_timeout_lock
, inpcb_lock_grp
, inpcb_lock_attr
);
292 inpcb_thread_call
= thread_call_allocate_with_priority(inpcb_timeout
,
293 NULL
, THREAD_CALL_PRIORITY_KERNEL
);
294 inpcb_fast_thread_call
= thread_call_allocate_with_priority(
295 inpcb_timeout
, NULL
, THREAD_CALL_PRIORITY_KERNEL
);
296 if (inpcb_thread_call
== NULL
|| inpcb_fast_thread_call
== NULL
)
297 panic("unable to alloc the inpcb thread call");
300 * Initialize data structures required to deliver
303 lck_mtx_init(&inp_fc_lck
, inpcb_lock_grp
, inpcb_lock_attr
);
304 lck_mtx_lock(&inp_fc_lck
);
305 RB_INIT(&inp_fc_tree
);
306 bzero(&key_inp
, sizeof(key_inp
));
307 lck_mtx_unlock(&inp_fc_lck
);
309 PE_parse_boot_argn("intcoproc_unrestricted", &intcoproc_unrestricted
,
310 sizeof (intcoproc_unrestricted
));
313 #define INPCB_HAVE_TIMER_REQ(req) (((req).intimer_lazy > 0) || \
314 ((req).intimer_fast > 0) || ((req).intimer_nodelay > 0))
316 inpcb_timeout(void *arg0
, void *arg1
)
319 struct inpcbinfo
*ipi
;
321 struct intimercount gccnt
, tmcnt
;
322 boolean_t toomany_gc
= FALSE
;
325 VERIFY(arg1
== &inpcb_toomany_gcreq
);
326 toomany_gc
= *(boolean_t
*)arg1
;
330 * Update coarse-grained networking timestamp (in sec.); the idea
331 * is to piggy-back on the timeout callout to update the counter
332 * returnable via net_uptime().
336 bzero(&gccnt
, sizeof(gccnt
));
337 bzero(&tmcnt
, sizeof(tmcnt
));
339 lck_mtx_lock_spin(&inpcb_timeout_lock
);
340 gc
= inpcb_garbage_collecting
;
341 inpcb_garbage_collecting
= FALSE
;
344 inpcb_ticking
= FALSE
;
347 lck_mtx_unlock(&inpcb_timeout_lock
);
349 lck_mtx_lock(&inpcb_lock
);
350 TAILQ_FOREACH(ipi
, &inpcb_head
, ipi_entry
) {
351 if (INPCB_HAVE_TIMER_REQ(ipi
->ipi_gc_req
)) {
352 bzero(&ipi
->ipi_gc_req
,
353 sizeof(ipi
->ipi_gc_req
));
354 if (gc
&& ipi
->ipi_gc
!= NULL
) {
356 gccnt
.intimer_lazy
+=
357 ipi
->ipi_gc_req
.intimer_lazy
;
358 gccnt
.intimer_fast
+=
359 ipi
->ipi_gc_req
.intimer_fast
;
360 gccnt
.intimer_nodelay
+=
361 ipi
->ipi_gc_req
.intimer_nodelay
;
364 if (INPCB_HAVE_TIMER_REQ(ipi
->ipi_timer_req
)) {
365 bzero(&ipi
->ipi_timer_req
,
366 sizeof(ipi
->ipi_timer_req
));
367 if (t
&& ipi
->ipi_timer
!= NULL
) {
369 tmcnt
.intimer_lazy
+=
370 ipi
->ipi_timer_req
.intimer_lazy
;
371 tmcnt
.intimer_lazy
+=
372 ipi
->ipi_timer_req
.intimer_fast
;
373 tmcnt
.intimer_nodelay
+=
374 ipi
->ipi_timer_req
.intimer_nodelay
;
378 lck_mtx_unlock(&inpcb_lock
);
379 lck_mtx_lock_spin(&inpcb_timeout_lock
);
382 /* lock was dropped above, so check first before overriding */
383 if (!inpcb_garbage_collecting
)
384 inpcb_garbage_collecting
= INPCB_HAVE_TIMER_REQ(gccnt
);
386 inpcb_ticking
= INPCB_HAVE_TIMER_REQ(tmcnt
);
388 /* re-arm the timer if there's work to do */
390 inpcb_toomany_gcreq
= FALSE
;
393 VERIFY(inpcb_timeout_run
>= 0 && inpcb_timeout_run
< 2);
396 if (gccnt
.intimer_nodelay
> 0 || tmcnt
.intimer_nodelay
> 0)
397 inpcb_sched_timeout();
398 else if ((gccnt
.intimer_fast
+ tmcnt
.intimer_fast
) <= 5)
399 /* be lazy when idle with little activity */
400 inpcb_sched_lazy_timeout();
402 inpcb_sched_timeout();
404 lck_mtx_unlock(&inpcb_timeout_lock
);
408 inpcb_sched_timeout(void)
410 _inpcb_sched_timeout(0);
414 inpcb_sched_lazy_timeout(void)
416 _inpcb_sched_timeout(inpcb_timeout_lazy
);
420 _inpcb_sched_timeout(unsigned int offset
)
422 uint64_t deadline
, leeway
;
424 clock_interval_to_deadline(1, NSEC_PER_SEC
, &deadline
);
425 lck_mtx_assert(&inpcb_timeout_lock
, LCK_MTX_ASSERT_OWNED
);
426 if (inpcb_timeout_run
== 0 &&
427 (inpcb_garbage_collecting
|| inpcb_ticking
)) {
428 lck_mtx_convert_spin(&inpcb_timeout_lock
);
431 inpcb_fast_timer_on
= TRUE
;
432 thread_call_enter_delayed(inpcb_thread_call
,
435 inpcb_fast_timer_on
= FALSE
;
436 clock_interval_to_absolutetime_interval(offset
,
437 NSEC_PER_SEC
, &leeway
);
438 thread_call_enter_delayed_with_leeway(
439 inpcb_thread_call
, NULL
, deadline
, leeway
,
440 THREAD_CALL_DELAY_LEEWAY
);
442 } else if (inpcb_timeout_run
== 1 &&
443 offset
== 0 && !inpcb_fast_timer_on
) {
445 * Since the request was for a fast timer but the
446 * scheduled timer is a lazy timer, try to schedule
447 * another instance of fast timer also.
449 lck_mtx_convert_spin(&inpcb_timeout_lock
);
451 inpcb_fast_timer_on
= TRUE
;
452 thread_call_enter_delayed(inpcb_fast_thread_call
, deadline
);
457 inpcb_gc_sched(struct inpcbinfo
*ipi
, u_int32_t type
)
462 lck_mtx_lock_spin(&inpcb_timeout_lock
);
463 inpcb_garbage_collecting
= TRUE
;
464 gccnt
= ipi
->ipi_gc_req
.intimer_nodelay
+
465 ipi
->ipi_gc_req
.intimer_fast
;
467 if (gccnt
> INPCB_GCREQ_THRESHOLD
&& !inpcb_toomany_gcreq
) {
468 inpcb_toomany_gcreq
= TRUE
;
471 * There are toomany pcbs waiting to be garbage collected,
472 * schedule a much faster timeout in addition to
473 * the caller's request
475 lck_mtx_convert_spin(&inpcb_timeout_lock
);
476 clock_interval_to_deadline(100, NSEC_PER_MSEC
, &deadline
);
477 thread_call_enter1_delayed(inpcb_thread_call
,
478 &inpcb_toomany_gcreq
, deadline
);
482 case INPCB_TIMER_NODELAY
:
483 atomic_add_32(&ipi
->ipi_gc_req
.intimer_nodelay
, 1);
484 inpcb_sched_timeout();
486 case INPCB_TIMER_FAST
:
487 atomic_add_32(&ipi
->ipi_gc_req
.intimer_fast
, 1);
488 inpcb_sched_timeout();
491 atomic_add_32(&ipi
->ipi_gc_req
.intimer_lazy
, 1);
492 inpcb_sched_lazy_timeout();
495 lck_mtx_unlock(&inpcb_timeout_lock
);
499 inpcb_timer_sched(struct inpcbinfo
*ipi
, u_int32_t type
)
502 lck_mtx_lock_spin(&inpcb_timeout_lock
);
503 inpcb_ticking
= TRUE
;
505 case INPCB_TIMER_NODELAY
:
506 atomic_add_32(&ipi
->ipi_timer_req
.intimer_nodelay
, 1);
507 inpcb_sched_timeout();
509 case INPCB_TIMER_FAST
:
510 atomic_add_32(&ipi
->ipi_timer_req
.intimer_fast
, 1);
511 inpcb_sched_timeout();
514 atomic_add_32(&ipi
->ipi_timer_req
.intimer_lazy
, 1);
515 inpcb_sched_lazy_timeout();
518 lck_mtx_unlock(&inpcb_timeout_lock
);
522 in_pcbinfo_attach(struct inpcbinfo
*ipi
)
524 struct inpcbinfo
*ipi0
;
526 lck_mtx_lock(&inpcb_lock
);
527 TAILQ_FOREACH(ipi0
, &inpcb_head
, ipi_entry
) {
529 panic("%s: ipi %p already in the list\n",
534 TAILQ_INSERT_TAIL(&inpcb_head
, ipi
, ipi_entry
);
535 lck_mtx_unlock(&inpcb_lock
);
539 in_pcbinfo_detach(struct inpcbinfo
*ipi
)
541 struct inpcbinfo
*ipi0
;
544 lck_mtx_lock(&inpcb_lock
);
545 TAILQ_FOREACH(ipi0
, &inpcb_head
, ipi_entry
) {
550 TAILQ_REMOVE(&inpcb_head
, ipi0
, ipi_entry
);
553 lck_mtx_unlock(&inpcb_lock
);
559 * Allocate a PCB and associate it with the socket.
566 in_pcballoc(struct socket
*so
, struct inpcbinfo
*pcbinfo
, struct proc
*p
)
573 #endif /* CONFIG_MACF_NET */
575 if ((so
->so_flags1
& SOF1_CACHED_IN_SOCK_LAYER
) == 0) {
576 inp
= (struct inpcb
*)zalloc(pcbinfo
->ipi_zone
);
579 bzero((caddr_t
)inp
, sizeof (*inp
));
581 inp
= (struct inpcb
*)(void *)so
->so_saved_pcb
;
582 temp
= inp
->inp_saved_ppcb
;
583 bzero((caddr_t
)inp
, sizeof (*inp
));
584 inp
->inp_saved_ppcb
= temp
;
587 inp
->inp_gencnt
= ++pcbinfo
->ipi_gencnt
;
588 inp
->inp_pcbinfo
= pcbinfo
;
589 inp
->inp_socket
= so
;
591 mac_error
= mac_inpcb_label_init(inp
, M_WAITOK
);
592 if (mac_error
!= 0) {
593 if ((so
->so_flags1
& SOF1_CACHED_IN_SOCK_LAYER
) == 0)
594 zfree(pcbinfo
->ipi_zone
, inp
);
597 mac_inpcb_label_associate(so
, inp
);
598 #endif /* CONFIG_MACF_NET */
599 /* make sure inp_stat is always 64-bit aligned */
600 inp
->inp_stat
= (struct inp_stat
*)P2ROUNDUP(inp
->inp_stat_store
,
602 if (((uintptr_t)inp
->inp_stat
- (uintptr_t)inp
->inp_stat_store
) +
603 sizeof (*inp
->inp_stat
) > sizeof (inp
->inp_stat_store
)) {
604 panic("%s: insufficient space to align inp_stat", __func__
);
608 /* make sure inp_cstat is always 64-bit aligned */
609 inp
->inp_cstat
= (struct inp_stat
*)P2ROUNDUP(inp
->inp_cstat_store
,
611 if (((uintptr_t)inp
->inp_cstat
- (uintptr_t)inp
->inp_cstat_store
) +
612 sizeof (*inp
->inp_cstat
) > sizeof (inp
->inp_cstat_store
)) {
613 panic("%s: insufficient space to align inp_cstat", __func__
);
617 /* make sure inp_wstat is always 64-bit aligned */
618 inp
->inp_wstat
= (struct inp_stat
*)P2ROUNDUP(inp
->inp_wstat_store
,
620 if (((uintptr_t)inp
->inp_wstat
- (uintptr_t)inp
->inp_wstat_store
) +
621 sizeof (*inp
->inp_wstat
) > sizeof (inp
->inp_wstat_store
)) {
622 panic("%s: insufficient space to align inp_wstat", __func__
);
626 /* make sure inp_Wstat is always 64-bit aligned */
627 inp
->inp_Wstat
= (struct inp_stat
*)P2ROUNDUP(inp
->inp_Wstat_store
,
629 if (((uintptr_t)inp
->inp_Wstat
- (uintptr_t)inp
->inp_Wstat_store
) +
630 sizeof (*inp
->inp_Wstat
) > sizeof (inp
->inp_Wstat_store
)) {
631 panic("%s: insufficient space to align inp_Wstat", __func__
);
635 so
->so_pcb
= (caddr_t
)inp
;
637 if (so
->so_proto
->pr_flags
& PR_PCBLOCK
) {
638 lck_mtx_init(&inp
->inpcb_mtx
, pcbinfo
->ipi_lock_grp
,
639 pcbinfo
->ipi_lock_attr
);
643 if (SOCK_DOM(so
) == PF_INET6
&& !ip6_mapped_addr_on
)
644 inp
->inp_flags
|= IN6P_IPV6_V6ONLY
;
646 if (ip6_auto_flowlabel
)
647 inp
->inp_flags
|= IN6P_AUTOFLOWLABEL
;
649 if (intcoproc_unrestricted
)
650 inp
->inp_flags2
|= INP2_INTCOPROC_ALLOWED
;
652 (void) inp_update_policy(inp
);
654 lck_rw_lock_exclusive(pcbinfo
->ipi_lock
);
655 inp
->inp_gencnt
= ++pcbinfo
->ipi_gencnt
;
656 LIST_INSERT_HEAD(pcbinfo
->ipi_listhead
, inp
, inp_list
);
657 pcbinfo
->ipi_count
++;
658 lck_rw_done(pcbinfo
->ipi_lock
);
663 * in_pcblookup_local_and_cleanup does everything
664 * in_pcblookup_local does but it checks for a socket
665 * that's going away. Since we know that the lock is
666 * held read+write when this funciton is called, we
667 * can safely dispose of this socket like the slow
668 * timer would usually do and return NULL. This is
672 in_pcblookup_local_and_cleanup(struct inpcbinfo
*pcbinfo
, struct in_addr laddr
,
673 u_int lport_arg
, int wild_okay
)
677 /* Perform normal lookup */
678 inp
= in_pcblookup_local(pcbinfo
, laddr
, lport_arg
, wild_okay
);
680 /* Check if we found a match but it's waiting to be disposed */
681 if (inp
!= NULL
&& inp
->inp_wantcnt
== WNT_STOPUSING
) {
682 struct socket
*so
= inp
->inp_socket
;
684 lck_mtx_lock(&inp
->inpcb_mtx
);
686 if (so
->so_usecount
== 0) {
687 if (inp
->inp_state
!= INPCB_STATE_DEAD
)
689 in_pcbdispose(inp
); /* will unlock & destroy */
692 lck_mtx_unlock(&inp
->inpcb_mtx
);
700 in_pcb_conflict_post_msg(u_int16_t port
)
703 * Radar 5523020 send a kernel event notification if a
704 * non-participating socket tries to bind the port a socket
705 * who has set SOF_NOTIFYCONFLICT owns.
707 struct kev_msg ev_msg
;
708 struct kev_in_portinuse in_portinuse
;
710 bzero(&in_portinuse
, sizeof (struct kev_in_portinuse
));
711 bzero(&ev_msg
, sizeof (struct kev_msg
));
712 in_portinuse
.port
= ntohs(port
); /* port in host order */
713 in_portinuse
.req_pid
= proc_selfpid();
714 ev_msg
.vendor_code
= KEV_VENDOR_APPLE
;
715 ev_msg
.kev_class
= KEV_NETWORK_CLASS
;
716 ev_msg
.kev_subclass
= KEV_INET_SUBCLASS
;
717 ev_msg
.event_code
= KEV_INET_PORTINUSE
;
718 ev_msg
.dv
[0].data_ptr
= &in_portinuse
;
719 ev_msg
.dv
[0].data_length
= sizeof (struct kev_in_portinuse
);
720 ev_msg
.dv
[1].data_length
= 0;
721 dlil_post_complete_msg(NULL
, &ev_msg
);
725 * Bind an INPCB to an address and/or port. This routine should not alter
726 * the caller-supplied local address "nam".
729 * EADDRNOTAVAIL Address not available.
730 * EINVAL Invalid argument
731 * EAFNOSUPPORT Address family not supported [notdef]
732 * EACCES Permission denied
733 * EADDRINUSE Address in use
734 * EAGAIN Resource unavailable, try again
735 * priv_check_cred:EPERM Operation not permitted
738 in_pcbbind(struct inpcb
*inp
, struct sockaddr
*nam
, struct proc
*p
)
740 struct socket
*so
= inp
->inp_socket
;
741 unsigned short *lastport
;
742 struct inpcbinfo
*pcbinfo
= inp
->inp_pcbinfo
;
743 u_short lport
= 0, rand_port
= 0;
744 int wild
= 0, reuseport
= (so
->so_options
& SO_REUSEPORT
);
745 int error
, randomport
, conflict
= 0;
746 boolean_t anonport
= FALSE
;
748 struct in_addr laddr
;
749 struct ifnet
*outif
= NULL
;
751 if (TAILQ_EMPTY(&in_ifaddrhead
)) /* XXX broken! */
752 return (EADDRNOTAVAIL
);
753 if (inp
->inp_lport
!= 0 || inp
->inp_laddr
.s_addr
!= INADDR_ANY
)
755 if (!(so
->so_options
& (SO_REUSEADDR
|SO_REUSEPORT
)))
758 bzero(&laddr
, sizeof(laddr
));
760 socket_unlock(so
, 0); /* keep reference on socket */
761 lck_rw_lock_exclusive(pcbinfo
->ipi_lock
);
765 if (nam
->sa_len
!= sizeof (struct sockaddr_in
)) {
766 lck_rw_done(pcbinfo
->ipi_lock
);
772 * We should check the family, but old programs
773 * incorrectly fail to initialize it.
775 if (nam
->sa_family
!= AF_INET
) {
776 lck_rw_done(pcbinfo
->ipi_lock
);
778 return (EAFNOSUPPORT
);
781 lport
= SIN(nam
)->sin_port
;
783 if (IN_MULTICAST(ntohl(SIN(nam
)->sin_addr
.s_addr
))) {
785 * Treat SO_REUSEADDR as SO_REUSEPORT for multicast;
786 * allow complete duplication of binding if
787 * SO_REUSEPORT is set, or if SO_REUSEADDR is set
788 * and a multicast address is bound on both
789 * new and duplicated sockets.
791 if (so
->so_options
& SO_REUSEADDR
)
792 reuseport
= SO_REUSEADDR
|SO_REUSEPORT
;
793 } else if (SIN(nam
)->sin_addr
.s_addr
!= INADDR_ANY
) {
794 struct sockaddr_in sin
;
797 /* Sanitized for interface address searches */
798 bzero(&sin
, sizeof (sin
));
799 sin
.sin_family
= AF_INET
;
800 sin
.sin_len
= sizeof (struct sockaddr_in
);
801 sin
.sin_addr
.s_addr
= SIN(nam
)->sin_addr
.s_addr
;
803 ifa
= ifa_ifwithaddr(SA(&sin
));
805 lck_rw_done(pcbinfo
->ipi_lock
);
807 return (EADDRNOTAVAIL
);
810 * Opportunistically determine the outbound
811 * interface that may be used; this may not
812 * hold true if we end up using a route
813 * going over a different interface, e.g.
814 * when sending to a local address. This
815 * will get updated again after sending.
818 outif
= ifa
->ifa_ifp
;
827 if (ntohs(lport
) < IPPORT_RESERVED
) {
828 cred
= kauth_cred_proc_ref(p
);
829 error
= priv_check_cred(cred
,
830 PRIV_NETINET_RESERVEDPORT
, 0);
831 kauth_cred_unref(&cred
);
833 lck_rw_done(pcbinfo
->ipi_lock
);
838 if (!IN_MULTICAST(ntohl(SIN(nam
)->sin_addr
.s_addr
)) &&
839 (u
= kauth_cred_getuid(so
->so_cred
)) != 0 &&
840 (t
= in_pcblookup_local_and_cleanup(
841 inp
->inp_pcbinfo
, SIN(nam
)->sin_addr
, lport
,
842 INPLOOKUP_WILDCARD
)) != NULL
&&
843 (SIN(nam
)->sin_addr
.s_addr
!= INADDR_ANY
||
844 t
->inp_laddr
.s_addr
!= INADDR_ANY
||
845 !(t
->inp_socket
->so_options
& SO_REUSEPORT
)) &&
846 (u
!= kauth_cred_getuid(t
->inp_socket
->so_cred
)) &&
847 !(t
->inp_socket
->so_flags
& SOF_REUSESHAREUID
) &&
848 (SIN(nam
)->sin_addr
.s_addr
!= INADDR_ANY
||
849 t
->inp_laddr
.s_addr
!= INADDR_ANY
)) {
850 if ((t
->inp_socket
->so_flags
&
851 SOF_NOTIFYCONFLICT
) &&
852 !(so
->so_flags
& SOF_NOTIFYCONFLICT
))
855 lck_rw_done(pcbinfo
->ipi_lock
);
858 in_pcb_conflict_post_msg(lport
);
863 t
= in_pcblookup_local_and_cleanup(pcbinfo
,
864 SIN(nam
)->sin_addr
, lport
, wild
);
866 (reuseport
& t
->inp_socket
->so_options
) == 0) {
868 if (SIN(nam
)->sin_addr
.s_addr
!= INADDR_ANY
||
869 t
->inp_laddr
.s_addr
!= INADDR_ANY
||
870 SOCK_DOM(so
) != PF_INET6
||
871 SOCK_DOM(t
->inp_socket
) != PF_INET6
)
875 if ((t
->inp_socket
->so_flags
&
876 SOF_NOTIFYCONFLICT
) &&
877 !(so
->so_flags
& SOF_NOTIFYCONFLICT
))
880 lck_rw_done(pcbinfo
->ipi_lock
);
883 in_pcb_conflict_post_msg(lport
);
889 laddr
= SIN(nam
)->sin_addr
;
895 randomport
= (so
->so_flags
& SOF_BINDRANDOMPORT
) ||
896 (so
->so_type
== SOCK_STREAM
? tcp_use_randomport
:
900 * Even though this looks similar to the code in
901 * in6_pcbsetport, the v6 vs v4 checks are different.
904 if (inp
->inp_flags
& INP_HIGHPORT
) {
905 first
= ipport_hifirstauto
; /* sysctl */
906 last
= ipport_hilastauto
;
907 lastport
= &pcbinfo
->ipi_lasthi
;
908 } else if (inp
->inp_flags
& INP_LOWPORT
) {
909 cred
= kauth_cred_proc_ref(p
);
910 error
= priv_check_cred(cred
,
911 PRIV_NETINET_RESERVEDPORT
, 0);
912 kauth_cred_unref(&cred
);
914 lck_rw_done(pcbinfo
->ipi_lock
);
918 first
= ipport_lowfirstauto
; /* 1023 */
919 last
= ipport_lowlastauto
; /* 600 */
920 lastport
= &pcbinfo
->ipi_lastlow
;
922 first
= ipport_firstauto
; /* sysctl */
923 last
= ipport_lastauto
;
924 lastport
= &pcbinfo
->ipi_lastport
;
926 /* No point in randomizing if only one port is available */
931 * Simple check to ensure all ports are not used up causing
934 * We split the two cases (up and down) so that the direction
935 * is not being tested on each round of the loop.
942 read_random(&rand_port
, sizeof (rand_port
));
944 first
- (rand_port
% (first
- last
));
946 count
= first
- last
;
949 if (count
-- < 0) { /* completely used? */
950 lck_rw_done(pcbinfo
->ipi_lock
);
952 return (EADDRNOTAVAIL
);
955 if (*lastport
> first
|| *lastport
< last
)
957 lport
= htons(*lastport
);
958 } while (in_pcblookup_local_and_cleanup(pcbinfo
,
959 ((laddr
.s_addr
!= INADDR_ANY
) ? laddr
:
960 inp
->inp_laddr
), lport
, wild
));
966 read_random(&rand_port
, sizeof (rand_port
));
968 first
+ (rand_port
% (first
- last
));
970 count
= last
- first
;
973 if (count
-- < 0) { /* completely used? */
974 lck_rw_done(pcbinfo
->ipi_lock
);
976 return (EADDRNOTAVAIL
);
979 if (*lastport
< first
|| *lastport
> last
)
981 lport
= htons(*lastport
);
982 } while (in_pcblookup_local_and_cleanup(pcbinfo
,
983 ((laddr
.s_addr
!= INADDR_ANY
) ? laddr
:
984 inp
->inp_laddr
), lport
, wild
));
990 * We unlocked socket's protocol lock for a long time.
991 * The socket might have been dropped/defuncted.
992 * Checking if world has changed since.
994 if (inp
->inp_state
== INPCB_STATE_DEAD
) {
995 lck_rw_done(pcbinfo
->ipi_lock
);
996 return (ECONNABORTED
);
999 if (inp
->inp_lport
!= 0 || inp
->inp_laddr
.s_addr
!= INADDR_ANY
) {
1000 lck_rw_done(pcbinfo
->ipi_lock
);
1004 if (laddr
.s_addr
!= INADDR_ANY
) {
1005 inp
->inp_laddr
= laddr
;
1006 inp
->inp_last_outifp
= outif
;
1008 inp
->inp_lport
= lport
;
1010 inp
->inp_flags
|= INP_ANONPORT
;
1012 if (in_pcbinshash(inp
, 1) != 0) {
1013 inp
->inp_laddr
.s_addr
= INADDR_ANY
;
1014 inp
->inp_last_outifp
= NULL
;
1018 inp
->inp_flags
&= ~INP_ANONPORT
;
1019 lck_rw_done(pcbinfo
->ipi_lock
);
1022 lck_rw_done(pcbinfo
->ipi_lock
);
1023 sflt_notify(so
, sock_evt_bound
, NULL
);
1027 #define APN_FALLBACK_IP_FILTER(a) \
1028 (IN_LINKLOCAL(ntohl((a)->sin_addr.s_addr)) || \
1029 IN_LOOPBACK(ntohl((a)->sin_addr.s_addr)) || \
1030 IN_ZERONET(ntohl((a)->sin_addr.s_addr)) || \
1031 IN_MULTICAST(ntohl((a)->sin_addr.s_addr)) || \
1032 IN_PRIVATE(ntohl((a)->sin_addr.s_addr)))
1034 #define APN_FALLBACK_NOTIF_INTERVAL 2 /* Magic Number */
1035 static uint64_t last_apn_fallback
= 0;
1038 apn_fallback_required (proc_t proc
, struct socket
*so
, struct sockaddr_in
*p_dstv4
)
1041 struct sockaddr_storage lookup_default_addr
;
1042 struct rtentry
*rt
= NULL
;
1044 VERIFY(proc
!= NULL
);
1046 if (apn_fallbk_enabled
== FALSE
)
1049 if (proc
== kernproc
)
1052 if (so
&& (so
->so_options
& SO_NOAPNFALLBK
))
1055 timenow
= net_uptime();
1056 if ((timenow
- last_apn_fallback
) < APN_FALLBACK_NOTIF_INTERVAL
) {
1057 apn_fallbk_log((LOG_INFO
, "APN fallback notification throttled.\n"));
1061 if (p_dstv4
&& APN_FALLBACK_IP_FILTER(p_dstv4
))
1064 /* Check if we have unscoped IPv6 default route through cellular */
1065 bzero(&lookup_default_addr
, sizeof(lookup_default_addr
));
1066 lookup_default_addr
.ss_family
= AF_INET6
;
1067 lookup_default_addr
.ss_len
= sizeof(struct sockaddr_in6
);
1069 rt
= rtalloc1((struct sockaddr
*)&lookup_default_addr
, 0, 0);
1071 apn_fallbk_log((LOG_INFO
, "APN fallback notification could not find "
1072 "unscoped default IPv6 route.\n"));
1076 if (!IFNET_IS_CELLULAR(rt
->rt_ifp
)) {
1078 apn_fallbk_log((LOG_INFO
, "APN fallback notification could not find "
1079 "unscoped default IPv6 route through cellular interface.\n"));
1084 * We have a default IPv6 route, ensure that
1085 * we do not have IPv4 default route before triggering
1091 bzero(&lookup_default_addr
, sizeof(lookup_default_addr
));
1092 lookup_default_addr
.ss_family
= AF_INET
;
1093 lookup_default_addr
.ss_len
= sizeof(struct sockaddr_in
);
1095 rt
= rtalloc1((struct sockaddr
*)&lookup_default_addr
, 0, 0);
1100 apn_fallbk_log((LOG_INFO
, "APN fallback notification found unscoped "
1101 "IPv4 default route!\n"));
1107 * We disable APN fallback if the binary is not a third-party app.
1108 * Note that platform daemons use their process name as a
1109 * bundle ID so we filter out bundle IDs without dots.
1111 const char *bundle_id
= cs_identity_get(proc
);
1112 if (bundle_id
== NULL
||
1113 bundle_id
[0] == '\0' ||
1114 strchr(bundle_id
, '.') == NULL
||
1115 strncmp(bundle_id
, "com.apple.", sizeof("com.apple.") - 1) == 0) {
1116 apn_fallbk_log((LOG_INFO
, "Abort: APN fallback notification found first-"
1117 "party bundle ID \"%s\"!\n", (bundle_id
? bundle_id
: "NULL")));
1124 * The Apple App Store IPv6 requirement started on
1125 * June 1st, 2016 at 12:00:00 AM PDT.
1126 * We disable APN fallback if the binary is more recent than that.
1127 * We check both atime and birthtime since birthtime is not always supported.
1129 static const long ipv6_start_date
= 1464764400L;
1130 vfs_context_t context
;
1134 bzero(&sb
, sizeof(struct stat64
));
1135 context
= vfs_context_create(NULL
);
1136 vn_stat_error
= vn_stat(proc
->p_textvp
, &sb
, NULL
, 1, context
);
1137 (void)vfs_context_rele(context
);
1139 if (vn_stat_error
!= 0 ||
1140 sb
.st_atimespec
.tv_sec
>= ipv6_start_date
||
1141 sb
.st_birthtimespec
.tv_sec
>= ipv6_start_date
) {
1142 apn_fallbk_log((LOG_INFO
, "Abort: APN fallback notification found binary "
1143 "too recent! (err %d atime %ld mtime %ld ctime %ld birthtime %ld)\n",
1144 vn_stat_error
, sb
.st_atimespec
.tv_sec
, sb
.st_mtimespec
.tv_sec
,
1145 sb
.st_ctimespec
.tv_sec
, sb
.st_birthtimespec
.tv_sec
));
1153 apn_fallback_trigger(proc_t proc
)
1156 struct kev_msg ev_msg
;
1157 struct kev_netevent_apnfallbk_data apnfallbk_data
;
1159 last_apn_fallback
= net_uptime();
1160 pid
= proc_pid(proc
);
1161 uuid_t application_uuid
;
1162 uuid_clear(application_uuid
);
1163 proc_getexecutableuuid(proc
, application_uuid
,
1164 sizeof(application_uuid
));
1166 bzero(&ev_msg
, sizeof (struct kev_msg
));
1167 ev_msg
.vendor_code
= KEV_VENDOR_APPLE
;
1168 ev_msg
.kev_class
= KEV_NETWORK_CLASS
;
1169 ev_msg
.kev_subclass
= KEV_NETEVENT_SUBCLASS
;
1170 ev_msg
.event_code
= KEV_NETEVENT_APNFALLBACK
;
1172 bzero(&apnfallbk_data
, sizeof(apnfallbk_data
));
1173 apnfallbk_data
.epid
= pid
;
1174 uuid_copy(apnfallbk_data
.euuid
, application_uuid
);
1176 ev_msg
.dv
[0].data_ptr
= &apnfallbk_data
;
1177 ev_msg
.dv
[0].data_length
= sizeof(apnfallbk_data
);
1178 kev_post_msg(&ev_msg
);
1179 apn_fallbk_log((LOG_INFO
, "APN fallback notification issued.\n"));
1183 * Transform old in_pcbconnect() into an inner subroutine for new
1184 * in_pcbconnect(); do some validity-checking on the remote address
1185 * (in "nam") and then determine local host address (i.e., which
1186 * interface) to use to access that remote host.
1188 * This routine may alter the caller-supplied remote address "nam".
1190 * The caller may override the bound-to-interface setting of the socket
1191 * by specifying the ifscope parameter (e.g. from IP_PKTINFO.)
1193 * This routine might return an ifp with a reference held if the caller
1194 * provides a non-NULL outif, even in the error case. The caller is
1195 * responsible for releasing its reference.
1197 * Returns: 0 Success
1198 * EINVAL Invalid argument
1199 * EAFNOSUPPORT Address family not supported
1200 * EADDRNOTAVAIL Address not available
1203 in_pcbladdr(struct inpcb
*inp
, struct sockaddr
*nam
, struct in_addr
*laddr
,
1204 unsigned int ifscope
, struct ifnet
**outif
, int raw
)
1206 struct route
*ro
= &inp
->inp_route
;
1207 struct in_ifaddr
*ia
= NULL
;
1208 struct sockaddr_in sin
;
1210 boolean_t restricted
= FALSE
;
1214 if (nam
->sa_len
!= sizeof (struct sockaddr_in
))
1216 if (SIN(nam
)->sin_family
!= AF_INET
)
1217 return (EAFNOSUPPORT
);
1218 if (raw
== 0 && SIN(nam
)->sin_port
== 0)
1219 return (EADDRNOTAVAIL
);
1222 * If the destination address is INADDR_ANY,
1223 * use the primary local address.
1224 * If the supplied address is INADDR_BROADCAST,
1225 * and the primary interface supports broadcast,
1226 * choose the broadcast address for that interface.
1228 if (raw
== 0 && (SIN(nam
)->sin_addr
.s_addr
== INADDR_ANY
||
1229 SIN(nam
)->sin_addr
.s_addr
== (u_int32_t
)INADDR_BROADCAST
)) {
1230 lck_rw_lock_shared(in_ifaddr_rwlock
);
1231 if (!TAILQ_EMPTY(&in_ifaddrhead
)) {
1232 ia
= TAILQ_FIRST(&in_ifaddrhead
);
1233 IFA_LOCK_SPIN(&ia
->ia_ifa
);
1234 if (SIN(nam
)->sin_addr
.s_addr
== INADDR_ANY
) {
1235 SIN(nam
)->sin_addr
= IA_SIN(ia
)->sin_addr
;
1236 } else if (ia
->ia_ifp
->if_flags
& IFF_BROADCAST
) {
1237 SIN(nam
)->sin_addr
=
1238 SIN(&ia
->ia_broadaddr
)->sin_addr
;
1240 IFA_UNLOCK(&ia
->ia_ifa
);
1243 lck_rw_done(in_ifaddr_rwlock
);
1246 * Otherwise, if the socket has already bound the source, just use it.
1248 if (inp
->inp_laddr
.s_addr
!= INADDR_ANY
) {
1250 *laddr
= inp
->inp_laddr
;
1255 * If the ifscope is specified by the caller (e.g. IP_PKTINFO)
1256 * then it overrides the sticky ifscope set for the socket.
1258 if (ifscope
== IFSCOPE_NONE
&& (inp
->inp_flags
& INP_BOUND_IF
))
1259 ifscope
= inp
->inp_boundifp
->if_index
;
1262 * If route is known or can be allocated now,
1263 * our src addr is taken from the i/f, else punt.
1264 * Note that we should check the address family of the cached
1265 * destination, in case of sharing the cache with IPv6.
1267 if (ro
->ro_rt
!= NULL
)
1268 RT_LOCK_SPIN(ro
->ro_rt
);
1269 if (ROUTE_UNUSABLE(ro
) || ro
->ro_dst
.sa_family
!= AF_INET
||
1270 SIN(&ro
->ro_dst
)->sin_addr
.s_addr
!= SIN(nam
)->sin_addr
.s_addr
||
1271 (inp
->inp_socket
->so_options
& SO_DONTROUTE
)) {
1272 if (ro
->ro_rt
!= NULL
)
1273 RT_UNLOCK(ro
->ro_rt
);
1276 if (!(inp
->inp_socket
->so_options
& SO_DONTROUTE
) &&
1277 (ro
->ro_rt
== NULL
|| ro
->ro_rt
->rt_ifp
== NULL
)) {
1278 if (ro
->ro_rt
!= NULL
)
1279 RT_UNLOCK(ro
->ro_rt
);
1281 /* No route yet, so try to acquire one */
1282 bzero(&ro
->ro_dst
, sizeof (struct sockaddr_in
));
1283 ro
->ro_dst
.sa_family
= AF_INET
;
1284 ro
->ro_dst
.sa_len
= sizeof (struct sockaddr_in
);
1285 SIN(&ro
->ro_dst
)->sin_addr
= SIN(nam
)->sin_addr
;
1286 rtalloc_scoped(ro
, ifscope
);
1287 if (ro
->ro_rt
!= NULL
)
1288 RT_LOCK_SPIN(ro
->ro_rt
);
1290 /* Sanitized local copy for interface address searches */
1291 bzero(&sin
, sizeof (sin
));
1292 sin
.sin_family
= AF_INET
;
1293 sin
.sin_len
= sizeof (struct sockaddr_in
);
1294 sin
.sin_addr
.s_addr
= SIN(nam
)->sin_addr
.s_addr
;
1296 * If we did not find (or use) a route, assume dest is reachable
1297 * on a directly connected network and try to find a corresponding
1298 * interface to take the source address from.
1300 if (ro
->ro_rt
== NULL
) {
1301 proc_t proc
= current_proc();
1304 ia
= ifatoia(ifa_ifwithdstaddr(SA(&sin
)));
1306 ia
= ifatoia(ifa_ifwithnet_scoped(SA(&sin
), ifscope
));
1307 error
= ((ia
== NULL
) ? ENETUNREACH
: 0);
1309 if (apn_fallback_required(proc
, inp
->inp_socket
,
1311 apn_fallback_trigger(proc
);
1315 RT_LOCK_ASSERT_HELD(ro
->ro_rt
);
1317 * If the outgoing interface on the route found is not
1318 * a loopback interface, use the address from that interface.
1320 if (!(ro
->ro_rt
->rt_ifp
->if_flags
& IFF_LOOPBACK
)) {
1323 * If the route points to a cellular interface and the
1324 * caller forbids our using interfaces of such type,
1325 * pretend that there is no route.
1326 * Apply the same logic for expensive interfaces.
1328 if (inp_restricted_send(inp
, ro
->ro_rt
->rt_ifp
)) {
1329 RT_UNLOCK(ro
->ro_rt
);
1331 error
= EHOSTUNREACH
;
1334 /* Become a regular mutex */
1335 RT_CONVERT_LOCK(ro
->ro_rt
);
1336 ia
= ifatoia(ro
->ro_rt
->rt_ifa
);
1337 IFA_ADDREF(&ia
->ia_ifa
);
1338 RT_UNLOCK(ro
->ro_rt
);
1343 VERIFY(ro
->ro_rt
->rt_ifp
->if_flags
& IFF_LOOPBACK
);
1344 RT_UNLOCK(ro
->ro_rt
);
1346 * The outgoing interface is marked with 'loopback net', so a route
1347 * to ourselves is here.
1348 * Try to find the interface of the destination address and then
1349 * take the address from there. That interface is not necessarily
1350 * a loopback interface.
1353 ia
= ifatoia(ifa_ifwithdstaddr(SA(&sin
)));
1355 ia
= ifatoia(ifa_ifwithaddr_scoped(SA(&sin
), ifscope
));
1357 ia
= ifatoia(ifa_ifwithnet_scoped(SA(&sin
), ifscope
));
1360 ia
= ifatoia(ro
->ro_rt
->rt_ifa
);
1362 IFA_ADDREF(&ia
->ia_ifa
);
1363 RT_UNLOCK(ro
->ro_rt
);
1365 error
= ((ia
== NULL
) ? ENETUNREACH
: 0);
1369 * If the destination address is multicast and an outgoing
1370 * interface has been set as a multicast option, use the
1371 * address of that interface as our source address.
1373 if (IN_MULTICAST(ntohl(SIN(nam
)->sin_addr
.s_addr
)) &&
1374 inp
->inp_moptions
!= NULL
) {
1375 struct ip_moptions
*imo
;
1378 imo
= inp
->inp_moptions
;
1380 if (imo
->imo_multicast_ifp
!= NULL
&& (ia
== NULL
||
1381 ia
->ia_ifp
!= imo
->imo_multicast_ifp
)) {
1382 ifp
= imo
->imo_multicast_ifp
;
1384 IFA_REMREF(&ia
->ia_ifa
);
1385 lck_rw_lock_shared(in_ifaddr_rwlock
);
1386 TAILQ_FOREACH(ia
, &in_ifaddrhead
, ia_link
) {
1387 if (ia
->ia_ifp
== ifp
)
1391 IFA_ADDREF(&ia
->ia_ifa
);
1392 lck_rw_done(in_ifaddr_rwlock
);
1394 error
= EADDRNOTAVAIL
;
1401 * Don't do pcblookup call here; return interface in laddr
1402 * and exit to caller, that will do the lookup.
1406 * If the source address belongs to a cellular interface
1407 * and the socket forbids our using interfaces of such
1408 * type, pretend that there is no source address.
1409 * Apply the same logic for expensive interfaces.
1411 IFA_LOCK_SPIN(&ia
->ia_ifa
);
1412 if (inp_restricted_send(inp
, ia
->ia_ifa
.ifa_ifp
)) {
1413 IFA_UNLOCK(&ia
->ia_ifa
);
1414 error
= EHOSTUNREACH
;
1416 } else if (error
== 0) {
1417 *laddr
= ia
->ia_addr
.sin_addr
;
1418 if (outif
!= NULL
) {
1421 if (ro
->ro_rt
!= NULL
)
1422 ifp
= ro
->ro_rt
->rt_ifp
;
1426 VERIFY(ifp
!= NULL
);
1427 IFA_CONVERT_LOCK(&ia
->ia_ifa
);
1428 ifnet_reference(ifp
); /* for caller */
1430 ifnet_release(*outif
);
1433 IFA_UNLOCK(&ia
->ia_ifa
);
1435 IFA_UNLOCK(&ia
->ia_ifa
);
1437 IFA_REMREF(&ia
->ia_ifa
);
1441 if (restricted
&& error
== EHOSTUNREACH
) {
1442 soevent(inp
->inp_socket
, (SO_FILT_HINT_LOCKED
|
1443 SO_FILT_HINT_IFDENIED
));
1451 * Connect from a socket to a specified address.
1452 * Both address and port must be specified in argument sin.
1453 * If don't have a local address for this socket yet,
1456 * The caller may override the bound-to-interface setting of the socket
1457 * by specifying the ifscope parameter (e.g. from IP_PKTINFO.)
1460 in_pcbconnect(struct inpcb
*inp
, struct sockaddr
*nam
, struct proc
*p
,
1461 unsigned int ifscope
, struct ifnet
**outif
)
1463 struct in_addr laddr
;
1464 struct sockaddr_in
*sin
= (struct sockaddr_in
*)(void *)nam
;
1467 struct socket
*so
= inp
->inp_socket
;
1470 * Call inner routine, to assign local interface address.
1472 if ((error
= in_pcbladdr(inp
, nam
, &laddr
, ifscope
, outif
, 0)) != 0)
1475 socket_unlock(so
, 0);
1476 pcb
= in_pcblookup_hash(inp
->inp_pcbinfo
, sin
->sin_addr
, sin
->sin_port
,
1477 inp
->inp_laddr
.s_addr
? inp
->inp_laddr
: laddr
,
1478 inp
->inp_lport
, 0, NULL
);
1482 * Check if the socket is still in a valid state. When we unlock this
1483 * embryonic socket, it can get aborted if another thread is closing
1484 * the listener (radar 7947600).
1486 if ((so
->so_flags
& SOF_ABORTED
) != 0)
1487 return (ECONNREFUSED
);
1490 in_pcb_checkstate(pcb
, WNT_RELEASE
, pcb
== inp
? 1 : 0);
1491 return (EADDRINUSE
);
1493 if (inp
->inp_laddr
.s_addr
== INADDR_ANY
) {
1494 if (inp
->inp_lport
== 0) {
1495 error
= in_pcbbind(inp
, NULL
, p
);
1499 if (!lck_rw_try_lock_exclusive(inp
->inp_pcbinfo
->ipi_lock
)) {
1501 * Lock inversion issue, mostly with udp
1502 * multicast packets.
1504 socket_unlock(so
, 0);
1505 lck_rw_lock_exclusive(inp
->inp_pcbinfo
->ipi_lock
);
1508 inp
->inp_laddr
= laddr
;
1509 /* no reference needed */
1510 inp
->inp_last_outifp
= (outif
!= NULL
) ? *outif
: NULL
;
1511 inp
->inp_flags
|= INP_INADDR_ANY
;
1514 * Usage of IP_PKTINFO, without local port already
1515 * speficified will cause kernel to panic,
1516 * see rdar://problem/18508185.
1517 * For now returning error to avoid a kernel panic
1518 * This routines can be refactored and handle this better
1521 if (inp
->inp_lport
== 0)
1523 if (!lck_rw_try_lock_exclusive(inp
->inp_pcbinfo
->ipi_lock
)) {
1525 * Lock inversion issue, mostly with udp
1526 * multicast packets.
1528 socket_unlock(so
, 0);
1529 lck_rw_lock_exclusive(inp
->inp_pcbinfo
->ipi_lock
);
1533 inp
->inp_faddr
= sin
->sin_addr
;
1534 inp
->inp_fport
= sin
->sin_port
;
1535 if (nstat_collect
&& SOCK_PROTO(so
) == IPPROTO_UDP
)
1536 nstat_pcb_invalidate_cache(inp
);
1538 lck_rw_done(inp
->inp_pcbinfo
->ipi_lock
);
1543 in_pcbdisconnect(struct inpcb
*inp
)
1545 struct socket
*so
= inp
->inp_socket
;
1547 if (nstat_collect
&& SOCK_PROTO(so
) == IPPROTO_UDP
)
1548 nstat_pcb_cache(inp
);
1550 inp
->inp_faddr
.s_addr
= INADDR_ANY
;
1553 if (!lck_rw_try_lock_exclusive(inp
->inp_pcbinfo
->ipi_lock
)) {
1554 /* lock inversion issue, mostly with udp multicast packets */
1555 socket_unlock(so
, 0);
1556 lck_rw_lock_exclusive(inp
->inp_pcbinfo
->ipi_lock
);
1561 lck_rw_done(inp
->inp_pcbinfo
->ipi_lock
);
1563 * A multipath subflow socket would have its SS_NOFDREF set by default,
1564 * so check for SOF_MP_SUBFLOW socket flag before detaching the PCB;
1565 * when the socket is closed for real, SOF_MP_SUBFLOW would be cleared.
1567 if (!(so
->so_flags
& SOF_MP_SUBFLOW
) && (so
->so_state
& SS_NOFDREF
))
1572 in_pcbdetach(struct inpcb
*inp
)
1574 struct socket
*so
= inp
->inp_socket
;
1576 if (so
->so_pcb
== NULL
) {
1577 /* PCB has been disposed */
1578 panic("%s: inp=%p so=%p proto=%d so_pcb is null!\n", __func__
,
1579 inp
, so
, SOCK_PROTO(so
));
1584 if (inp
->inp_sp
!= NULL
) {
1585 (void) ipsec4_delete_pcbpolicy(inp
);
1590 * Let NetworkStatistics know this PCB is going away
1591 * before we detach it.
1593 if (nstat_collect
&&
1594 (SOCK_PROTO(so
) == IPPROTO_TCP
|| SOCK_PROTO(so
) == IPPROTO_UDP
))
1595 nstat_pcb_detach(inp
);
1597 /* Free memory buffer held for generating keep alives */
1598 if (inp
->inp_keepalive_data
!= NULL
) {
1599 FREE(inp
->inp_keepalive_data
, M_TEMP
);
1600 inp
->inp_keepalive_data
= NULL
;
1603 /* mark socket state as dead */
1604 if (in_pcb_checkstate(inp
, WNT_STOPUSING
, 1) != WNT_STOPUSING
) {
1605 panic("%s: so=%p proto=%d couldn't set to STOPUSING\n",
1606 __func__
, so
, SOCK_PROTO(so
));
1610 if (!(so
->so_flags
& SOF_PCBCLEARING
)) {
1611 struct ip_moptions
*imo
;
1614 if (inp
->inp_options
!= NULL
) {
1615 (void) m_free(inp
->inp_options
);
1616 inp
->inp_options
= NULL
;
1618 ROUTE_RELEASE(&inp
->inp_route
);
1619 imo
= inp
->inp_moptions
;
1620 inp
->inp_moptions
= NULL
;
1621 sofreelastref(so
, 0);
1622 inp
->inp_state
= INPCB_STATE_DEAD
;
1623 /* makes sure we're not called twice from so_close */
1624 so
->so_flags
|= SOF_PCBCLEARING
;
1626 inpcb_gc_sched(inp
->inp_pcbinfo
, INPCB_TIMER_FAST
);
1629 * See inp_join_group() for why we need to unlock
1632 socket_unlock(so
, 0);
1641 in_pcbdispose(struct inpcb
*inp
)
1643 struct socket
*so
= inp
->inp_socket
;
1644 struct inpcbinfo
*ipi
= inp
->inp_pcbinfo
;
1646 if (so
!= NULL
&& so
->so_usecount
!= 0) {
1647 panic("%s: so %p [%d,%d] usecount %d lockhistory %s\n",
1648 __func__
, so
, SOCK_DOM(so
), SOCK_TYPE(so
), so
->so_usecount
,
1649 solockhistory_nr(so
));
1651 } else if (inp
->inp_wantcnt
!= WNT_STOPUSING
) {
1653 panic_plain("%s: inp %p invalid wantcnt %d, so %p "
1654 "[%d,%d] usecount %d retaincnt %d state 0x%x "
1655 "flags 0x%x lockhistory %s\n", __func__
, inp
,
1656 inp
->inp_wantcnt
, so
, SOCK_DOM(so
), SOCK_TYPE(so
),
1657 so
->so_usecount
, so
->so_retaincnt
, so
->so_state
,
1658 so
->so_flags
, solockhistory_nr(so
));
1661 panic("%s: inp %p invalid wantcnt %d no socket\n",
1662 __func__
, inp
, inp
->inp_wantcnt
);
1667 lck_rw_assert(ipi
->ipi_lock
, LCK_RW_ASSERT_EXCLUSIVE
);
1669 inp
->inp_gencnt
= ++ipi
->ipi_gencnt
;
1670 /* access ipi in in_pcbremlists */
1671 in_pcbremlists(inp
);
1674 if (so
->so_proto
->pr_flags
& PR_PCBLOCK
) {
1675 sofreelastref(so
, 0);
1676 if (so
->so_rcv
.sb_cc
> 0 || so
->so_snd
.sb_cc
> 0) {
1678 * selthreadclear() already called
1679 * during sofreelastref() above.
1681 sbrelease(&so
->so_rcv
);
1682 sbrelease(&so
->so_snd
);
1684 if (so
->so_head
!= NULL
) {
1685 panic("%s: so=%p head still exist\n",
1689 lck_mtx_unlock(&inp
->inpcb_mtx
);
1690 lck_mtx_destroy(&inp
->inpcb_mtx
, ipi
->ipi_lock_grp
);
1692 /* makes sure we're not called twice from so_close */
1693 so
->so_flags
|= SOF_PCBCLEARING
;
1694 so
->so_saved_pcb
= (caddr_t
)inp
;
1696 inp
->inp_socket
= NULL
;
1698 mac_inpcb_label_destroy(inp
);
1699 #endif /* CONFIG_MACF_NET */
1701 necp_inpcb_dispose(inp
);
1704 * In case there a route cached after a detach (possible
1705 * in the tcp case), make sure that it is freed before
1706 * we deallocate the structure.
1708 ROUTE_RELEASE(&inp
->inp_route
);
1709 if ((so
->so_flags1
& SOF1_CACHED_IN_SOCK_LAYER
) == 0) {
1710 zfree(ipi
->ipi_zone
, inp
);
1717 * The calling convention of in_getsockaddr() and in_getpeeraddr() was
1718 * modified to match the pru_sockaddr() and pru_peeraddr() entry points
1719 * in struct pr_usrreqs, so that protocols can just reference then directly
1720 * without the need for a wrapper function.
1723 in_getsockaddr(struct socket
*so
, struct sockaddr
**nam
)
1726 struct sockaddr_in
*sin
;
1729 * Do the malloc first in case it blocks.
1731 MALLOC(sin
, struct sockaddr_in
*, sizeof (*sin
), M_SONAME
, M_WAITOK
);
1734 bzero(sin
, sizeof (*sin
));
1735 sin
->sin_family
= AF_INET
;
1736 sin
->sin_len
= sizeof (*sin
);
1738 if ((inp
= sotoinpcb(so
)) == NULL
) {
1739 FREE(sin
, M_SONAME
);
1742 sin
->sin_port
= inp
->inp_lport
;
1743 sin
->sin_addr
= inp
->inp_laddr
;
1745 *nam
= (struct sockaddr
*)sin
;
1750 in_getsockaddr_s(struct socket
*so
, struct sockaddr_storage
*ss
)
1752 struct sockaddr_in
*sin
= SIN(ss
);
1756 bzero(ss
, sizeof (*ss
));
1758 sin
->sin_family
= AF_INET
;
1759 sin
->sin_len
= sizeof (*sin
);
1761 if ((inp
= sotoinpcb(so
)) == NULL
1763 || (necp_socket_should_use_flow_divert(inp
))
1766 return (inp
== NULL
? EINVAL
: EPROTOTYPE
);
1768 sin
->sin_port
= inp
->inp_lport
;
1769 sin
->sin_addr
= inp
->inp_laddr
;
1774 in_getpeeraddr(struct socket
*so
, struct sockaddr
**nam
)
1777 struct sockaddr_in
*sin
;
1780 * Do the malloc first in case it blocks.
1782 MALLOC(sin
, struct sockaddr_in
*, sizeof (*sin
), M_SONAME
, M_WAITOK
);
1785 bzero((caddr_t
)sin
, sizeof (*sin
));
1786 sin
->sin_family
= AF_INET
;
1787 sin
->sin_len
= sizeof (*sin
);
1789 if ((inp
= sotoinpcb(so
)) == NULL
) {
1790 FREE(sin
, M_SONAME
);
1793 sin
->sin_port
= inp
->inp_fport
;
1794 sin
->sin_addr
= inp
->inp_faddr
;
1796 *nam
= (struct sockaddr
*)sin
;
1801 in_getpeeraddr_s(struct socket
*so
, struct sockaddr_storage
*ss
)
1803 struct sockaddr_in
*sin
= SIN(ss
);
1807 bzero(ss
, sizeof (*ss
));
1809 sin
->sin_family
= AF_INET
;
1810 sin
->sin_len
= sizeof (*sin
);
1812 if ((inp
= sotoinpcb(so
)) == NULL
1814 || (necp_socket_should_use_flow_divert(inp
))
1817 return (inp
== NULL
? EINVAL
: EPROTOTYPE
);
1820 sin
->sin_port
= inp
->inp_fport
;
1821 sin
->sin_addr
= inp
->inp_faddr
;
1826 in_pcbnotifyall(struct inpcbinfo
*pcbinfo
, struct in_addr faddr
,
1827 int errno
, void (*notify
)(struct inpcb
*, int))
1831 lck_rw_lock_shared(pcbinfo
->ipi_lock
);
1833 LIST_FOREACH(inp
, pcbinfo
->ipi_listhead
, inp_list
) {
1835 if (!(inp
->inp_vflag
& INP_IPV4
))
1838 if (inp
->inp_faddr
.s_addr
!= faddr
.s_addr
||
1839 inp
->inp_socket
== NULL
)
1841 if (in_pcb_checkstate(inp
, WNT_ACQUIRE
, 0) == WNT_STOPUSING
)
1843 socket_lock(inp
->inp_socket
, 1);
1844 (*notify
)(inp
, errno
);
1845 (void) in_pcb_checkstate(inp
, WNT_RELEASE
, 1);
1846 socket_unlock(inp
->inp_socket
, 1);
1848 lck_rw_done(pcbinfo
->ipi_lock
);
1852 * Check for alternatives when higher level complains
1853 * about service problems. For now, invalidate cached
1854 * routing information. If the route was created dynamically
1855 * (by a redirect), time to try a default gateway again.
1858 in_losing(struct inpcb
*inp
)
1860 boolean_t release
= FALSE
;
1863 if ((rt
= inp
->inp_route
.ro_rt
) != NULL
) {
1864 struct in_ifaddr
*ia
= NULL
;
1867 if (rt
->rt_flags
& RTF_DYNAMIC
) {
1869 * Prevent another thread from modifying rt_key,
1870 * rt_gateway via rt_setgate() after rt_lock is
1871 * dropped by marking the route as defunct.
1873 rt
->rt_flags
|= RTF_CONDEMNED
;
1875 (void) rtrequest(RTM_DELETE
, rt_key(rt
),
1876 rt
->rt_gateway
, rt_mask(rt
), rt
->rt_flags
, NULL
);
1880 /* if the address is gone keep the old route in the pcb */
1881 if (inp
->inp_laddr
.s_addr
!= INADDR_ANY
&&
1882 (ia
= ifa_foraddr(inp
->inp_laddr
.s_addr
)) != NULL
) {
1884 * Address is around; ditch the route. A new route
1885 * can be allocated the next time output is attempted.
1890 IFA_REMREF(&ia
->ia_ifa
);
1892 if (rt
== NULL
|| release
)
1893 ROUTE_RELEASE(&inp
->inp_route
);
1897 * After a routing change, flush old routing
1898 * and allocate a (hopefully) better one.
1901 in_rtchange(struct inpcb
*inp
, int errno
)
1903 #pragma unused(errno)
1904 boolean_t release
= FALSE
;
1907 if ((rt
= inp
->inp_route
.ro_rt
) != NULL
) {
1908 struct in_ifaddr
*ia
= NULL
;
1910 /* if address is gone, keep the old route */
1911 if (inp
->inp_laddr
.s_addr
!= INADDR_ANY
&&
1912 (ia
= ifa_foraddr(inp
->inp_laddr
.s_addr
)) != NULL
) {
1914 * Address is around; ditch the route. A new route
1915 * can be allocated the next time output is attempted.
1920 IFA_REMREF(&ia
->ia_ifa
);
1922 if (rt
== NULL
|| release
)
1923 ROUTE_RELEASE(&inp
->inp_route
);
1927 * Lookup a PCB based on the local address and port.
1930 in_pcblookup_local(struct inpcbinfo
*pcbinfo
, struct in_addr laddr
,
1931 unsigned int lport_arg
, int wild_okay
)
1934 int matchwild
= 3, wildcard
;
1935 u_short lport
= lport_arg
;
1937 KERNEL_DEBUG(DBG_FNC_PCB_LOOKUP
| DBG_FUNC_START
, 0, 0, 0, 0, 0);
1940 struct inpcbhead
*head
;
1942 * Look for an unconnected (wildcard foreign addr) PCB that
1943 * matches the local address and port we're looking for.
1945 head
= &pcbinfo
->ipi_hashbase
[INP_PCBHASH(INADDR_ANY
, lport
, 0,
1946 pcbinfo
->ipi_hashmask
)];
1947 LIST_FOREACH(inp
, head
, inp_hash
) {
1949 if (!(inp
->inp_vflag
& INP_IPV4
))
1952 if (inp
->inp_faddr
.s_addr
== INADDR_ANY
&&
1953 inp
->inp_laddr
.s_addr
== laddr
.s_addr
&&
1954 inp
->inp_lport
== lport
) {
1964 KERNEL_DEBUG(DBG_FNC_PCB_LOOKUP
| DBG_FUNC_END
, 0, 0, 0, 0, 0);
1967 struct inpcbporthead
*porthash
;
1968 struct inpcbport
*phd
;
1969 struct inpcb
*match
= NULL
;
1971 * Best fit PCB lookup.
1973 * First see if this local port is in use by looking on the
1976 porthash
= &pcbinfo
->ipi_porthashbase
[INP_PCBPORTHASH(lport
,
1977 pcbinfo
->ipi_porthashmask
)];
1978 LIST_FOREACH(phd
, porthash
, phd_hash
) {
1979 if (phd
->phd_port
== lport
)
1984 * Port is in use by one or more PCBs. Look for best
1987 LIST_FOREACH(inp
, &phd
->phd_pcblist
, inp_portlist
) {
1990 if (!(inp
->inp_vflag
& INP_IPV4
))
1993 if (inp
->inp_faddr
.s_addr
!= INADDR_ANY
)
1995 if (inp
->inp_laddr
.s_addr
!= INADDR_ANY
) {
1996 if (laddr
.s_addr
== INADDR_ANY
)
1998 else if (inp
->inp_laddr
.s_addr
!=
2002 if (laddr
.s_addr
!= INADDR_ANY
)
2005 if (wildcard
< matchwild
) {
2007 matchwild
= wildcard
;
2008 if (matchwild
== 0) {
2014 KERNEL_DEBUG(DBG_FNC_PCB_LOOKUP
| DBG_FUNC_END
, match
,
2021 * Check if PCB exists in hash list.
2024 in_pcblookup_hash_exists(struct inpcbinfo
*pcbinfo
, struct in_addr faddr
,
2025 u_int fport_arg
, struct in_addr laddr
, u_int lport_arg
, int wildcard
,
2026 uid_t
*uid
, gid_t
*gid
, struct ifnet
*ifp
)
2028 struct inpcbhead
*head
;
2030 u_short fport
= fport_arg
, lport
= lport_arg
;
2032 struct inpcb
*local_wild
= NULL
;
2034 struct inpcb
*local_wild_mapped
= NULL
;
2041 * We may have found the pcb in the last lookup - check this first.
2044 lck_rw_lock_shared(pcbinfo
->ipi_lock
);
2047 * First look for an exact match.
2049 head
= &pcbinfo
->ipi_hashbase
[INP_PCBHASH(faddr
.s_addr
, lport
, fport
,
2050 pcbinfo
->ipi_hashmask
)];
2051 LIST_FOREACH(inp
, head
, inp_hash
) {
2053 if (!(inp
->inp_vflag
& INP_IPV4
))
2056 if (inp_restricted_recv(inp
, ifp
))
2059 if (inp
->inp_faddr
.s_addr
== faddr
.s_addr
&&
2060 inp
->inp_laddr
.s_addr
== laddr
.s_addr
&&
2061 inp
->inp_fport
== fport
&&
2062 inp
->inp_lport
== lport
) {
2063 if ((found
= (inp
->inp_socket
!= NULL
))) {
2067 *uid
= kauth_cred_getuid(
2068 inp
->inp_socket
->so_cred
);
2069 *gid
= kauth_cred_getgid(
2070 inp
->inp_socket
->so_cred
);
2072 lck_rw_done(pcbinfo
->ipi_lock
);
2081 lck_rw_done(pcbinfo
->ipi_lock
);
2085 head
= &pcbinfo
->ipi_hashbase
[INP_PCBHASH(INADDR_ANY
, lport
, 0,
2086 pcbinfo
->ipi_hashmask
)];
2087 LIST_FOREACH(inp
, head
, inp_hash
) {
2089 if (!(inp
->inp_vflag
& INP_IPV4
))
2092 if (inp_restricted_recv(inp
, ifp
))
2095 if (inp
->inp_faddr
.s_addr
== INADDR_ANY
&&
2096 inp
->inp_lport
== lport
) {
2097 if (inp
->inp_laddr
.s_addr
== laddr
.s_addr
) {
2098 if ((found
= (inp
->inp_socket
!= NULL
))) {
2099 *uid
= kauth_cred_getuid(
2100 inp
->inp_socket
->so_cred
);
2101 *gid
= kauth_cred_getgid(
2102 inp
->inp_socket
->so_cred
);
2104 lck_rw_done(pcbinfo
->ipi_lock
);
2106 } else if (inp
->inp_laddr
.s_addr
== INADDR_ANY
) {
2108 if (inp
->inp_socket
&&
2109 SOCK_CHECK_DOM(inp
->inp_socket
, PF_INET6
))
2110 local_wild_mapped
= inp
;
2117 if (local_wild
== NULL
) {
2119 if (local_wild_mapped
!= NULL
) {
2120 if ((found
= (local_wild_mapped
->inp_socket
!= NULL
))) {
2121 *uid
= kauth_cred_getuid(
2122 local_wild_mapped
->inp_socket
->so_cred
);
2123 *gid
= kauth_cred_getgid(
2124 local_wild_mapped
->inp_socket
->so_cred
);
2126 lck_rw_done(pcbinfo
->ipi_lock
);
2130 lck_rw_done(pcbinfo
->ipi_lock
);
2133 if ((found
= (local_wild
->inp_socket
!= NULL
))) {
2134 *uid
= kauth_cred_getuid(
2135 local_wild
->inp_socket
->so_cred
);
2136 *gid
= kauth_cred_getgid(
2137 local_wild
->inp_socket
->so_cred
);
2139 lck_rw_done(pcbinfo
->ipi_lock
);
2144 * Lookup PCB in hash list.
2147 in_pcblookup_hash(struct inpcbinfo
*pcbinfo
, struct in_addr faddr
,
2148 u_int fport_arg
, struct in_addr laddr
, u_int lport_arg
, int wildcard
,
2151 struct inpcbhead
*head
;
2153 u_short fport
= fport_arg
, lport
= lport_arg
;
2154 struct inpcb
*local_wild
= NULL
;
2156 struct inpcb
*local_wild_mapped
= NULL
;
2160 * We may have found the pcb in the last lookup - check this first.
2163 lck_rw_lock_shared(pcbinfo
->ipi_lock
);
2166 * First look for an exact match.
2168 head
= &pcbinfo
->ipi_hashbase
[INP_PCBHASH(faddr
.s_addr
, lport
, fport
,
2169 pcbinfo
->ipi_hashmask
)];
2170 LIST_FOREACH(inp
, head
, inp_hash
) {
2172 if (!(inp
->inp_vflag
& INP_IPV4
))
2175 if (inp_restricted_recv(inp
, ifp
))
2178 if (inp
->inp_faddr
.s_addr
== faddr
.s_addr
&&
2179 inp
->inp_laddr
.s_addr
== laddr
.s_addr
&&
2180 inp
->inp_fport
== fport
&&
2181 inp
->inp_lport
== lport
) {
2185 if (in_pcb_checkstate(inp
, WNT_ACQUIRE
, 0) !=
2187 lck_rw_done(pcbinfo
->ipi_lock
);
2190 /* it's there but dead, say it isn't found */
2191 lck_rw_done(pcbinfo
->ipi_lock
);
2201 lck_rw_done(pcbinfo
->ipi_lock
);
2205 head
= &pcbinfo
->ipi_hashbase
[INP_PCBHASH(INADDR_ANY
, lport
, 0,
2206 pcbinfo
->ipi_hashmask
)];
2207 LIST_FOREACH(inp
, head
, inp_hash
) {
2209 if (!(inp
->inp_vflag
& INP_IPV4
))
2212 if (inp_restricted_recv(inp
, ifp
))
2215 if (inp
->inp_faddr
.s_addr
== INADDR_ANY
&&
2216 inp
->inp_lport
== lport
) {
2217 if (inp
->inp_laddr
.s_addr
== laddr
.s_addr
) {
2218 if (in_pcb_checkstate(inp
, WNT_ACQUIRE
, 0) !=
2220 lck_rw_done(pcbinfo
->ipi_lock
);
2223 /* it's dead; say it isn't found */
2224 lck_rw_done(pcbinfo
->ipi_lock
);
2227 } else if (inp
->inp_laddr
.s_addr
== INADDR_ANY
) {
2229 if (SOCK_CHECK_DOM(inp
->inp_socket
, PF_INET6
))
2230 local_wild_mapped
= inp
;
2237 if (local_wild
== NULL
) {
2239 if (local_wild_mapped
!= NULL
) {
2240 if (in_pcb_checkstate(local_wild_mapped
,
2241 WNT_ACQUIRE
, 0) != WNT_STOPUSING
) {
2242 lck_rw_done(pcbinfo
->ipi_lock
);
2243 return (local_wild_mapped
);
2245 /* it's dead; say it isn't found */
2246 lck_rw_done(pcbinfo
->ipi_lock
);
2251 lck_rw_done(pcbinfo
->ipi_lock
);
2254 if (in_pcb_checkstate(local_wild
, WNT_ACQUIRE
, 0) != WNT_STOPUSING
) {
2255 lck_rw_done(pcbinfo
->ipi_lock
);
2256 return (local_wild
);
2259 * It's either not found or is already dead.
2261 lck_rw_done(pcbinfo
->ipi_lock
);
2266 * @brief Insert PCB onto various hash lists.
2268 * @param inp Pointer to internet protocol control block
2269 * @param locked Implies if ipi_lock (protecting pcb list)
2270 * is already locked or not.
2272 * @return int error on failure and 0 on success
2275 in_pcbinshash(struct inpcb
*inp
, int locked
)
2277 struct inpcbhead
*pcbhash
;
2278 struct inpcbporthead
*pcbporthash
;
2279 struct inpcbinfo
*pcbinfo
= inp
->inp_pcbinfo
;
2280 struct inpcbport
*phd
;
2281 u_int32_t hashkey_faddr
;
2284 if (!lck_rw_try_lock_exclusive(pcbinfo
->ipi_lock
)) {
2286 * Lock inversion issue, mostly with udp
2289 socket_unlock(inp
->inp_socket
, 0);
2290 lck_rw_lock_exclusive(pcbinfo
->ipi_lock
);
2291 socket_lock(inp
->inp_socket
, 0);
2296 * This routine or its caller may have given up
2297 * socket's protocol lock briefly.
2298 * During that time the socket may have been dropped.
2299 * Safe-guarding against that.
2301 if (inp
->inp_state
== INPCB_STATE_DEAD
) {
2303 lck_rw_done(pcbinfo
->ipi_lock
);
2305 return (ECONNABORTED
);
2310 if (inp
->inp_vflag
& INP_IPV6
)
2311 hashkey_faddr
= inp
->in6p_faddr
.s6_addr32
[3] /* XXX */;
2314 hashkey_faddr
= inp
->inp_faddr
.s_addr
;
2316 inp
->inp_hash_element
= INP_PCBHASH(hashkey_faddr
, inp
->inp_lport
,
2317 inp
->inp_fport
, pcbinfo
->ipi_hashmask
);
2319 pcbhash
= &pcbinfo
->ipi_hashbase
[inp
->inp_hash_element
];
2321 pcbporthash
= &pcbinfo
->ipi_porthashbase
[INP_PCBPORTHASH(inp
->inp_lport
,
2322 pcbinfo
->ipi_porthashmask
)];
2325 * Go through port list and look for a head for this lport.
2327 LIST_FOREACH(phd
, pcbporthash
, phd_hash
) {
2328 if (phd
->phd_port
== inp
->inp_lport
)
2333 * If none exists, malloc one and tack it on.
2336 MALLOC(phd
, struct inpcbport
*, sizeof (struct inpcbport
),
2340 lck_rw_done(pcbinfo
->ipi_lock
);
2341 return (ENOBUFS
); /* XXX */
2343 phd
->phd_port
= inp
->inp_lport
;
2344 LIST_INIT(&phd
->phd_pcblist
);
2345 LIST_INSERT_HEAD(pcbporthash
, phd
, phd_hash
);
2348 VERIFY(!(inp
->inp_flags2
& INP2_INHASHLIST
));
2350 LIST_INSERT_HEAD(&phd
->phd_pcblist
, inp
, inp_portlist
);
2351 LIST_INSERT_HEAD(pcbhash
, inp
, inp_hash
);
2352 inp
->inp_flags2
|= INP2_INHASHLIST
;
2355 lck_rw_done(pcbinfo
->ipi_lock
);
2358 // This call catches the original setting of the local address
2359 inp_update_necp_policy(inp
, NULL
, NULL
, 0);
2366 * Move PCB to the proper hash bucket when { faddr, fport } have been
2367 * changed. NOTE: This does not handle the case of the lport changing (the
2368 * hashed port list would have to be updated as well), so the lport must
2369 * not change after in_pcbinshash() has been called.
2372 in_pcbrehash(struct inpcb
*inp
)
2374 struct inpcbhead
*head
;
2375 u_int32_t hashkey_faddr
;
2378 if (inp
->inp_vflag
& INP_IPV6
)
2379 hashkey_faddr
= inp
->in6p_faddr
.s6_addr32
[3] /* XXX */;
2382 hashkey_faddr
= inp
->inp_faddr
.s_addr
;
2384 inp
->inp_hash_element
= INP_PCBHASH(hashkey_faddr
, inp
->inp_lport
,
2385 inp
->inp_fport
, inp
->inp_pcbinfo
->ipi_hashmask
);
2386 head
= &inp
->inp_pcbinfo
->ipi_hashbase
[inp
->inp_hash_element
];
2388 if (inp
->inp_flags2
& INP2_INHASHLIST
) {
2389 LIST_REMOVE(inp
, inp_hash
);
2390 inp
->inp_flags2
&= ~INP2_INHASHLIST
;
2393 VERIFY(!(inp
->inp_flags2
& INP2_INHASHLIST
));
2394 LIST_INSERT_HEAD(head
, inp
, inp_hash
);
2395 inp
->inp_flags2
|= INP2_INHASHLIST
;
2398 // This call catches updates to the remote addresses
2399 inp_update_necp_policy(inp
, NULL
, NULL
, 0);
2404 * Remove PCB from various lists.
2405 * Must be called pcbinfo lock is held in exclusive mode.
2408 in_pcbremlists(struct inpcb
*inp
)
2410 inp
->inp_gencnt
= ++inp
->inp_pcbinfo
->ipi_gencnt
;
2413 * Check if it's in hashlist -- an inp is placed in hashlist when
2414 * it's local port gets assigned. So it should also be present
2417 if (inp
->inp_flags2
& INP2_INHASHLIST
) {
2418 struct inpcbport
*phd
= inp
->inp_phd
;
2420 VERIFY(phd
!= NULL
&& inp
->inp_lport
> 0);
2422 LIST_REMOVE(inp
, inp_hash
);
2423 inp
->inp_hash
.le_next
= NULL
;
2424 inp
->inp_hash
.le_prev
= NULL
;
2426 LIST_REMOVE(inp
, inp_portlist
);
2427 inp
->inp_portlist
.le_next
= NULL
;
2428 inp
->inp_portlist
.le_prev
= NULL
;
2429 if (LIST_EMPTY(&phd
->phd_pcblist
)) {
2430 LIST_REMOVE(phd
, phd_hash
);
2433 inp
->inp_phd
= NULL
;
2434 inp
->inp_flags2
&= ~INP2_INHASHLIST
;
2436 VERIFY(!(inp
->inp_flags2
& INP2_INHASHLIST
));
2438 if (inp
->inp_flags2
& INP2_TIMEWAIT
) {
2439 /* Remove from time-wait queue */
2440 tcp_remove_from_time_wait(inp
);
2441 inp
->inp_flags2
&= ~INP2_TIMEWAIT
;
2442 VERIFY(inp
->inp_pcbinfo
->ipi_twcount
!= 0);
2443 inp
->inp_pcbinfo
->ipi_twcount
--;
2445 /* Remove from global inp list if it is not time-wait */
2446 LIST_REMOVE(inp
, inp_list
);
2449 if (inp
->inp_flags2
& INP2_IN_FCTREE
) {
2450 inp_fc_getinp(inp
->inp_flowhash
, (INPFC_SOLOCKED
|INPFC_REMOVE
));
2451 VERIFY(!(inp
->inp_flags2
& INP2_IN_FCTREE
));
2454 inp
->inp_pcbinfo
->ipi_count
--;
2458 * Mechanism used to defer the memory release of PCBs
2459 * The pcb list will contain the pcb until the reaper can clean it up if
2460 * the following conditions are met:
2462 * 2) wantcnt is STOPUSING
2464 * This function will be called to either mark the pcb as
2467 in_pcb_checkstate(struct inpcb
*pcb
, int mode
, int locked
)
2469 volatile UInt32
*wantcnt
= (volatile UInt32
*)&pcb
->inp_wantcnt
;
2476 * Try to mark the pcb as ready for recycling. CAS with
2477 * STOPUSING, if success we're good, if it's in use, will
2481 socket_lock(pcb
->inp_socket
, 1);
2482 pcb
->inp_state
= INPCB_STATE_DEAD
;
2485 if (pcb
->inp_socket
->so_usecount
< 0) {
2486 panic("%s: pcb=%p so=%p usecount is negative\n",
2487 __func__
, pcb
, pcb
->inp_socket
);
2491 socket_unlock(pcb
->inp_socket
, 1);
2493 inpcb_gc_sched(pcb
->inp_pcbinfo
, INPCB_TIMER_FAST
);
2495 origwant
= *wantcnt
;
2496 if ((UInt16
) origwant
== 0xffff) /* should stop using */
2497 return (WNT_STOPUSING
);
2499 if ((UInt16
) origwant
== 0) {
2500 /* try to mark it as unsuable now */
2501 OSCompareAndSwap(origwant
, newwant
, wantcnt
);
2503 return (WNT_STOPUSING
);
2507 * Try to increase reference to pcb. If WNT_STOPUSING
2508 * should bail out. If socket state DEAD, try to set count
2509 * to STOPUSING, return failed otherwise increase cnt.
2512 origwant
= *wantcnt
;
2513 if ((UInt16
) origwant
== 0xffff) {
2514 /* should stop using */
2515 return (WNT_STOPUSING
);
2517 newwant
= origwant
+ 1;
2518 } while (!OSCompareAndSwap(origwant
, newwant
, wantcnt
));
2519 return (WNT_ACQUIRE
);
2523 * Release reference. If result is null and pcb state
2524 * is DEAD, set wanted bit to STOPUSING
2527 socket_lock(pcb
->inp_socket
, 1);
2530 origwant
= *wantcnt
;
2531 if ((UInt16
) origwant
== 0x0) {
2532 panic("%s: pcb=%p release with zero count",
2536 if ((UInt16
) origwant
== 0xffff) {
2537 /* should stop using */
2539 socket_unlock(pcb
->inp_socket
, 1);
2540 return (WNT_STOPUSING
);
2542 newwant
= origwant
- 1;
2543 } while (!OSCompareAndSwap(origwant
, newwant
, wantcnt
));
2545 if (pcb
->inp_state
== INPCB_STATE_DEAD
)
2547 if (pcb
->inp_socket
->so_usecount
< 0) {
2548 panic("%s: RELEASE pcb=%p so=%p usecount is negative\n",
2549 __func__
, pcb
, pcb
->inp_socket
);
2554 socket_unlock(pcb
->inp_socket
, 1);
2555 return (WNT_RELEASE
);
2558 panic("%s: so=%p not a valid state =%x\n", __func__
,
2559 pcb
->inp_socket
, mode
);
2568 * inpcb_to_compat copies specific bits of an inpcb to a inpcb_compat.
2569 * The inpcb_compat data structure is passed to user space and must
2570 * not change. We intentionally avoid copying pointers.
2573 inpcb_to_compat(struct inpcb
*inp
, struct inpcb_compat
*inp_compat
)
2575 bzero(inp_compat
, sizeof (*inp_compat
));
2576 inp_compat
->inp_fport
= inp
->inp_fport
;
2577 inp_compat
->inp_lport
= inp
->inp_lport
;
2578 inp_compat
->nat_owner
= 0;
2579 inp_compat
->nat_cookie
= 0;
2580 inp_compat
->inp_gencnt
= inp
->inp_gencnt
;
2581 inp_compat
->inp_flags
= inp
->inp_flags
;
2582 inp_compat
->inp_flow
= inp
->inp_flow
;
2583 inp_compat
->inp_vflag
= inp
->inp_vflag
;
2584 inp_compat
->inp_ip_ttl
= inp
->inp_ip_ttl
;
2585 inp_compat
->inp_ip_p
= inp
->inp_ip_p
;
2586 inp_compat
->inp_dependfaddr
.inp6_foreign
=
2587 inp
->inp_dependfaddr
.inp6_foreign
;
2588 inp_compat
->inp_dependladdr
.inp6_local
=
2589 inp
->inp_dependladdr
.inp6_local
;
2590 inp_compat
->inp_depend4
.inp4_ip_tos
= inp
->inp_depend4
.inp4_ip_tos
;
2591 inp_compat
->inp_depend6
.inp6_hlim
= 0;
2592 inp_compat
->inp_depend6
.inp6_cksum
= inp
->inp_depend6
.inp6_cksum
;
2593 inp_compat
->inp_depend6
.inp6_ifindex
= 0;
2594 inp_compat
->inp_depend6
.inp6_hops
= inp
->inp_depend6
.inp6_hops
;
2598 inpcb_to_xinpcb64(struct inpcb
*inp
, struct xinpcb64
*xinp
)
2600 xinp
->inp_fport
= inp
->inp_fport
;
2601 xinp
->inp_lport
= inp
->inp_lport
;
2602 xinp
->inp_gencnt
= inp
->inp_gencnt
;
2603 xinp
->inp_flags
= inp
->inp_flags
;
2604 xinp
->inp_flow
= inp
->inp_flow
;
2605 xinp
->inp_vflag
= inp
->inp_vflag
;
2606 xinp
->inp_ip_ttl
= inp
->inp_ip_ttl
;
2607 xinp
->inp_ip_p
= inp
->inp_ip_p
;
2608 xinp
->inp_dependfaddr
.inp6_foreign
= inp
->inp_dependfaddr
.inp6_foreign
;
2609 xinp
->inp_dependladdr
.inp6_local
= inp
->inp_dependladdr
.inp6_local
;
2610 xinp
->inp_depend4
.inp4_ip_tos
= inp
->inp_depend4
.inp4_ip_tos
;
2611 xinp
->inp_depend6
.inp6_hlim
= 0;
2612 xinp
->inp_depend6
.inp6_cksum
= inp
->inp_depend6
.inp6_cksum
;
2613 xinp
->inp_depend6
.inp6_ifindex
= 0;
2614 xinp
->inp_depend6
.inp6_hops
= inp
->inp_depend6
.inp6_hops
;
2618 * The following routines implement this scheme:
2620 * Callers of ip_output() that intend to cache the route in the inpcb pass
2621 * a local copy of the struct route to ip_output(). Using a local copy of
2622 * the cached route significantly simplifies things as IP no longer has to
2623 * worry about having exclusive access to the passed in struct route, since
2624 * it's defined in the caller's stack; in essence, this allows for a lock-
2625 * less operation when updating the struct route at the IP level and below,
2626 * whenever necessary. The scheme works as follows:
2628 * Prior to dropping the socket's lock and calling ip_output(), the caller
2629 * copies the struct route from the inpcb into its stack, and adds a reference
2630 * to the cached route entry, if there was any. The socket's lock is then
2631 * dropped and ip_output() is called with a pointer to the copy of struct
2632 * route defined on the stack (not to the one in the inpcb.)
2634 * Upon returning from ip_output(), the caller then acquires the socket's
2635 * lock and synchronizes the cache; if there is no route cached in the inpcb,
2636 * it copies the local copy of struct route (which may or may not contain any
2637 * route) back into the cache; otherwise, if the inpcb has a route cached in
2638 * it, the one in the local copy will be freed, if there's any. Trashing the
2639 * cached route in the inpcb can be avoided because ip_output() is single-
2640 * threaded per-PCB (i.e. multiple transmits on a PCB are always serialized
2641 * by the socket/transport layer.)
2644 inp_route_copyout(struct inpcb
*inp
, struct route
*dst
)
2646 struct route
*src
= &inp
->inp_route
;
2648 lck_mtx_assert(&inp
->inpcb_mtx
, LCK_MTX_ASSERT_OWNED
);
2651 * If the route in the PCB is stale or not for IPv4, blow it away;
2652 * this is possible in the case of IPv4-mapped address case.
2654 if (ROUTE_UNUSABLE(src
) || rt_key(src
->ro_rt
)->sa_family
!= AF_INET
)
2657 route_copyout(dst
, src
, sizeof (*dst
));
2661 inp_route_copyin(struct inpcb
*inp
, struct route
*src
)
2663 struct route
*dst
= &inp
->inp_route
;
2665 lck_mtx_assert(&inp
->inpcb_mtx
, LCK_MTX_ASSERT_OWNED
);
2667 /* Minor sanity check */
2668 if (src
->ro_rt
!= NULL
&& rt_key(src
->ro_rt
)->sa_family
!= AF_INET
)
2669 panic("%s: wrong or corrupted route: %p", __func__
, src
);
2671 route_copyin(src
, dst
, sizeof (*src
));
2675 * Handler for setting IP_BOUND_IF/IPV6_BOUND_IF socket option.
2678 inp_bindif(struct inpcb
*inp
, unsigned int ifscope
, struct ifnet
**pifp
)
2680 struct ifnet
*ifp
= NULL
;
2682 ifnet_head_lock_shared();
2683 if ((ifscope
> (unsigned)if_index
) || (ifscope
!= IFSCOPE_NONE
&&
2684 (ifp
= ifindex2ifnet
[ifscope
]) == NULL
)) {
2690 VERIFY(ifp
!= NULL
|| ifscope
== IFSCOPE_NONE
);
2693 * A zero interface scope value indicates an "unbind".
2694 * Otherwise, take in whatever value the app desires;
2695 * the app may already know the scope (or force itself
2696 * to such a scope) ahead of time before the interface
2697 * gets attached. It doesn't matter either way; any
2698 * route lookup from this point on will require an
2699 * exact match for the embedded interface scope.
2701 inp
->inp_boundifp
= ifp
;
2702 if (inp
->inp_boundifp
== NULL
)
2703 inp
->inp_flags
&= ~INP_BOUND_IF
;
2705 inp
->inp_flags
|= INP_BOUND_IF
;
2707 /* Blow away any cached route in the PCB */
2708 ROUTE_RELEASE(&inp
->inp_route
);
2717 * Handler for setting IP_NO_IFT_CELLULAR/IPV6_NO_IFT_CELLULAR socket option,
2718 * as well as for setting PROC_UUID_NO_CELLULAR policy.
2721 inp_set_nocellular(struct inpcb
*inp
)
2723 inp
->inp_flags
|= INP_NO_IFT_CELLULAR
;
2725 /* Blow away any cached route in the PCB */
2726 ROUTE_RELEASE(&inp
->inp_route
);
2730 * Handler for clearing IP_NO_IFT_CELLULAR/IPV6_NO_IFT_CELLULAR socket option,
2731 * as well as for clearing PROC_UUID_NO_CELLULAR policy.
2734 inp_clear_nocellular(struct inpcb
*inp
)
2736 struct socket
*so
= inp
->inp_socket
;
2739 * SO_RESTRICT_DENY_CELLULAR socket restriction issued on the socket
2740 * has a higher precendence than INP_NO_IFT_CELLULAR. Clear the flag
2741 * if and only if the socket is unrestricted.
2743 if (so
!= NULL
&& !(so
->so_restrictions
& SO_RESTRICT_DENY_CELLULAR
)) {
2744 inp
->inp_flags
&= ~INP_NO_IFT_CELLULAR
;
2746 /* Blow away any cached route in the PCB */
2747 ROUTE_RELEASE(&inp
->inp_route
);
2752 inp_set_noexpensive(struct inpcb
*inp
)
2754 inp
->inp_flags2
|= INP2_NO_IFF_EXPENSIVE
;
2756 /* Blow away any cached route in the PCB */
2757 ROUTE_RELEASE(&inp
->inp_route
);
2761 inp_set_awdl_unrestricted(struct inpcb
*inp
)
2763 inp
->inp_flags2
|= INP2_AWDL_UNRESTRICTED
;
2765 /* Blow away any cached route in the PCB */
2766 ROUTE_RELEASE(&inp
->inp_route
);
2770 inp_get_awdl_unrestricted(struct inpcb
*inp
)
2772 return (inp
->inp_flags2
& INP2_AWDL_UNRESTRICTED
) ? TRUE
: FALSE
;
2776 inp_clear_awdl_unrestricted(struct inpcb
*inp
)
2778 inp
->inp_flags2
&= ~INP2_AWDL_UNRESTRICTED
;
2780 /* Blow away any cached route in the PCB */
2781 ROUTE_RELEASE(&inp
->inp_route
);
2785 inp_set_intcoproc_allowed(struct inpcb
*inp
)
2787 inp
->inp_flags2
|= INP2_INTCOPROC_ALLOWED
;
2789 /* Blow away any cached route in the PCB */
2790 ROUTE_RELEASE(&inp
->inp_route
);
2794 inp_get_intcoproc_allowed(struct inpcb
*inp
)
2796 return (inp
->inp_flags2
& INP2_INTCOPROC_ALLOWED
) ? TRUE
: FALSE
;
2800 inp_clear_intcoproc_allowed(struct inpcb
*inp
)
2802 inp
->inp_flags2
&= ~INP2_INTCOPROC_ALLOWED
;
2804 /* Blow away any cached route in the PCB */
2805 ROUTE_RELEASE(&inp
->inp_route
);
2810 * Called when PROC_UUID_NECP_APP_POLICY is set.
2813 inp_set_want_app_policy(struct inpcb
*inp
)
2815 inp
->inp_flags2
|= INP2_WANT_APP_POLICY
;
2819 * Called when PROC_UUID_NECP_APP_POLICY is cleared.
2822 inp_clear_want_app_policy(struct inpcb
*inp
)
2824 inp
->inp_flags2
&= ~INP2_WANT_APP_POLICY
;
2829 * Calculate flow hash for an inp, used by an interface to identify a
2830 * flow. When an interface provides flow control advisory, this flow
2831 * hash is used as an identifier.
2834 inp_calc_flowhash(struct inpcb
*inp
)
2836 struct inp_flowhash_key fh
__attribute__((aligned(8)));
2837 u_int32_t flowhash
= 0;
2838 struct inpcb
*tmp_inp
= NULL
;
2840 if (inp_hash_seed
== 0)
2841 inp_hash_seed
= RandomULong();
2843 bzero(&fh
, sizeof (fh
));
2845 bcopy(&inp
->inp_dependladdr
, &fh
.infh_laddr
, sizeof (fh
.infh_laddr
));
2846 bcopy(&inp
->inp_dependfaddr
, &fh
.infh_faddr
, sizeof (fh
.infh_faddr
));
2848 fh
.infh_lport
= inp
->inp_lport
;
2849 fh
.infh_fport
= inp
->inp_fport
;
2850 fh
.infh_af
= (inp
->inp_vflag
& INP_IPV6
) ? AF_INET6
: AF_INET
;
2851 fh
.infh_proto
= inp
->inp_ip_p
;
2852 fh
.infh_rand1
= RandomULong();
2853 fh
.infh_rand2
= RandomULong();
2856 flowhash
= net_flowhash(&fh
, sizeof (fh
), inp_hash_seed
);
2857 if (flowhash
== 0) {
2858 /* try to get a non-zero flowhash */
2859 inp_hash_seed
= RandomULong();
2863 inp
->inp_flowhash
= flowhash
;
2865 /* Insert the inp into inp_fc_tree */
2866 lck_mtx_lock_spin(&inp_fc_lck
);
2867 tmp_inp
= RB_FIND(inp_fc_tree
, &inp_fc_tree
, inp
);
2868 if (tmp_inp
!= NULL
) {
2870 * There is a different inp with the same flowhash.
2871 * There can be a collision on flow hash but the
2872 * probability is low. Let's recompute the
2875 lck_mtx_unlock(&inp_fc_lck
);
2876 /* recompute hash seed */
2877 inp_hash_seed
= RandomULong();
2881 RB_INSERT(inp_fc_tree
, &inp_fc_tree
, inp
);
2882 inp
->inp_flags2
|= INP2_IN_FCTREE
;
2883 lck_mtx_unlock(&inp_fc_lck
);
2889 inp_flowadv(uint32_t flowhash
)
2893 inp
= inp_fc_getinp(flowhash
, 0);
2897 inp_fc_feedback(inp
);
2901 * Function to compare inp_fc_entries in inp flow control tree
2904 infc_cmp(const struct inpcb
*inp1
, const struct inpcb
*inp2
)
2906 return (memcmp(&(inp1
->inp_flowhash
), &(inp2
->inp_flowhash
),
2907 sizeof(inp1
->inp_flowhash
)));
2910 static struct inpcb
*
2911 inp_fc_getinp(u_int32_t flowhash
, u_int32_t flags
)
2913 struct inpcb
*inp
= NULL
;
2914 int locked
= (flags
& INPFC_SOLOCKED
) ? 1 : 0;
2916 lck_mtx_lock_spin(&inp_fc_lck
);
2917 key_inp
.inp_flowhash
= flowhash
;
2918 inp
= RB_FIND(inp_fc_tree
, &inp_fc_tree
, &key_inp
);
2920 /* inp is not present, return */
2921 lck_mtx_unlock(&inp_fc_lck
);
2925 if (flags
& INPFC_REMOVE
) {
2926 RB_REMOVE(inp_fc_tree
, &inp_fc_tree
, inp
);
2927 lck_mtx_unlock(&inp_fc_lck
);
2929 bzero(&(inp
->infc_link
), sizeof (inp
->infc_link
));
2930 inp
->inp_flags2
&= ~INP2_IN_FCTREE
;
2934 if (in_pcb_checkstate(inp
, WNT_ACQUIRE
, locked
) == WNT_STOPUSING
)
2936 lck_mtx_unlock(&inp_fc_lck
);
2942 inp_fc_feedback(struct inpcb
*inp
)
2944 struct socket
*so
= inp
->inp_socket
;
2946 /* we already hold a want_cnt on this inp, socket can't be null */
2950 if (in_pcb_checkstate(inp
, WNT_RELEASE
, 1) == WNT_STOPUSING
) {
2951 socket_unlock(so
, 1);
2955 if (inp
->inp_sndinprog_cnt
> 0)
2956 inp
->inp_flags
|= INP_FC_FEEDBACK
;
2959 * Return if the connection is not in flow-controlled state.
2960 * This can happen if the connection experienced
2961 * loss while it was in flow controlled state
2963 if (!INP_WAIT_FOR_IF_FEEDBACK(inp
)) {
2964 socket_unlock(so
, 1);
2967 inp_reset_fc_state(inp
);
2969 if (SOCK_TYPE(so
) == SOCK_STREAM
)
2970 inp_fc_unthrottle_tcp(inp
);
2972 socket_unlock(so
, 1);
2976 inp_reset_fc_state(struct inpcb
*inp
)
2978 struct socket
*so
= inp
->inp_socket
;
2979 int suspended
= (INP_IS_FLOW_SUSPENDED(inp
)) ? 1 : 0;
2980 int needwakeup
= (INP_WAIT_FOR_IF_FEEDBACK(inp
)) ? 1 : 0;
2982 inp
->inp_flags
&= ~(INP_FLOW_CONTROLLED
| INP_FLOW_SUSPENDED
);
2985 so
->so_flags
&= ~(SOF_SUSPENDED
);
2986 soevent(so
, (SO_FILT_HINT_LOCKED
| SO_FILT_HINT_RESUME
));
2989 /* Give a write wakeup to unblock the socket */
2995 inp_set_fc_state(struct inpcb
*inp
, int advcode
)
2997 struct inpcb
*tmp_inp
= NULL
;
2999 * If there was a feedback from the interface when
3000 * send operation was in progress, we should ignore
3001 * this flow advisory to avoid a race between setting
3002 * flow controlled state and receiving feedback from
3005 if (inp
->inp_flags
& INP_FC_FEEDBACK
)
3008 inp
->inp_flags
&= ~(INP_FLOW_CONTROLLED
| INP_FLOW_SUSPENDED
);
3009 if ((tmp_inp
= inp_fc_getinp(inp
->inp_flowhash
,
3010 INPFC_SOLOCKED
)) != NULL
) {
3011 if (in_pcb_checkstate(tmp_inp
, WNT_RELEASE
, 1) == WNT_STOPUSING
)
3013 VERIFY(tmp_inp
== inp
);
3015 case FADV_FLOW_CONTROLLED
:
3016 inp
->inp_flags
|= INP_FLOW_CONTROLLED
;
3018 case FADV_SUSPENDED
:
3019 inp
->inp_flags
|= INP_FLOW_SUSPENDED
;
3020 soevent(inp
->inp_socket
,
3021 (SO_FILT_HINT_LOCKED
| SO_FILT_HINT_SUSPEND
));
3023 /* Record the fact that suspend event was sent */
3024 inp
->inp_socket
->so_flags
|= SOF_SUSPENDED
;
3033 * Handler for SO_FLUSH socket option.
3036 inp_flush(struct inpcb
*inp
, int optval
)
3038 u_int32_t flowhash
= inp
->inp_flowhash
;
3039 struct ifnet
*rtifp
, *oifp
;
3041 /* Either all classes or one of the valid ones */
3042 if (optval
!= SO_TC_ALL
&& !SO_VALID_TC(optval
))
3045 /* We need a flow hash for identification */
3049 /* Grab the interfaces from the route and pcb */
3050 rtifp
= ((inp
->inp_route
.ro_rt
!= NULL
) ?
3051 inp
->inp_route
.ro_rt
->rt_ifp
: NULL
);
3052 oifp
= inp
->inp_last_outifp
;
3055 if_qflush_sc(rtifp
, so_tc2msc(optval
), flowhash
, NULL
, NULL
, 0);
3056 if (oifp
!= NULL
&& oifp
!= rtifp
)
3057 if_qflush_sc(oifp
, so_tc2msc(optval
), flowhash
, NULL
, NULL
, 0);
3063 * Clear the INP_INADDR_ANY flag (special case for PPP only)
3066 inp_clear_INP_INADDR_ANY(struct socket
*so
)
3068 struct inpcb
*inp
= NULL
;
3071 inp
= sotoinpcb(so
);
3073 inp
->inp_flags
&= ~INP_INADDR_ANY
;
3075 socket_unlock(so
, 1);
3079 inp_get_soprocinfo(struct inpcb
*inp
, struct so_procinfo
*soprocinfo
)
3081 struct socket
*so
= inp
->inp_socket
;
3083 soprocinfo
->spi_pid
= so
->last_pid
;
3084 if (so
->last_pid
!= 0)
3085 uuid_copy(soprocinfo
->spi_uuid
, so
->last_uuid
);
3087 * When not delegated, the effective pid is the same as the real pid
3089 if (so
->so_flags
& SOF_DELEGATED
) {
3090 soprocinfo
->spi_delegated
= 1;
3091 soprocinfo
->spi_epid
= so
->e_pid
;
3092 uuid_copy(soprocinfo
->spi_euuid
, so
->e_uuid
);
3094 soprocinfo
->spi_delegated
= 0;
3095 soprocinfo
->spi_epid
= so
->last_pid
;
3100 inp_findinpcb_procinfo(struct inpcbinfo
*pcbinfo
, uint32_t flowhash
,
3101 struct so_procinfo
*soprocinfo
)
3103 struct inpcb
*inp
= NULL
;
3106 bzero(soprocinfo
, sizeof (struct so_procinfo
));
3111 lck_rw_lock_shared(pcbinfo
->ipi_lock
);
3112 LIST_FOREACH(inp
, pcbinfo
->ipi_listhead
, inp_list
) {
3113 if (inp
->inp_state
!= INPCB_STATE_DEAD
&&
3114 inp
->inp_socket
!= NULL
&&
3115 inp
->inp_flowhash
== flowhash
) {
3117 inp_get_soprocinfo(inp
, soprocinfo
);
3121 lck_rw_done(pcbinfo
->ipi_lock
);
3126 #if CONFIG_PROC_UUID_POLICY
3128 inp_update_cellular_policy(struct inpcb
*inp
, boolean_t set
)
3130 struct socket
*so
= inp
->inp_socket
;
3134 VERIFY(inp
->inp_state
!= INPCB_STATE_DEAD
);
3136 before
= INP_NO_CELLULAR(inp
);
3138 inp_set_nocellular(inp
);
3140 inp_clear_nocellular(inp
);
3142 after
= INP_NO_CELLULAR(inp
);
3143 if (net_io_policy_log
&& (before
!= after
)) {
3144 static const char *ok
= "OK";
3145 static const char *nok
= "NOACCESS";
3146 uuid_string_t euuid_buf
;
3149 if (so
->so_flags
& SOF_DELEGATED
) {
3150 uuid_unparse(so
->e_uuid
, euuid_buf
);
3153 uuid_unparse(so
->last_uuid
, euuid_buf
);
3154 epid
= so
->last_pid
;
3157 /* allow this socket to generate another notification event */
3158 so
->so_ifdenied_notifies
= 0;
3160 log(LOG_DEBUG
, "%s: so 0x%llx [%d,%d] epid %d "
3161 "euuid %s%s %s->%s\n", __func__
,
3162 (uint64_t)VM_KERNEL_ADDRPERM(so
), SOCK_DOM(so
),
3163 SOCK_TYPE(so
), epid
, euuid_buf
,
3164 (so
->so_flags
& SOF_DELEGATED
) ?
3165 " [delegated]" : "",
3166 ((before
< after
) ? ok
: nok
),
3167 ((before
< after
) ? nok
: ok
));
3173 inp_update_necp_want_app_policy(struct inpcb
*inp
, boolean_t set
)
3175 struct socket
*so
= inp
->inp_socket
;
3179 VERIFY(inp
->inp_state
!= INPCB_STATE_DEAD
);
3181 before
= (inp
->inp_flags2
& INP2_WANT_APP_POLICY
);
3183 inp_set_want_app_policy(inp
);
3185 inp_clear_want_app_policy(inp
);
3187 after
= (inp
->inp_flags2
& INP2_WANT_APP_POLICY
);
3188 if (net_io_policy_log
&& (before
!= after
)) {
3189 static const char *wanted
= "WANTED";
3190 static const char *unwanted
= "UNWANTED";
3191 uuid_string_t euuid_buf
;
3194 if (so
->so_flags
& SOF_DELEGATED
) {
3195 uuid_unparse(so
->e_uuid
, euuid_buf
);
3198 uuid_unparse(so
->last_uuid
, euuid_buf
);
3199 epid
= so
->last_pid
;
3202 log(LOG_DEBUG
, "%s: so 0x%llx [%d,%d] epid %d "
3203 "euuid %s%s %s->%s\n", __func__
,
3204 (uint64_t)VM_KERNEL_ADDRPERM(so
), SOCK_DOM(so
),
3205 SOCK_TYPE(so
), epid
, euuid_buf
,
3206 (so
->so_flags
& SOF_DELEGATED
) ?
3207 " [delegated]" : "",
3208 ((before
< after
) ? unwanted
: wanted
),
3209 ((before
< after
) ? wanted
: unwanted
));
3213 #endif /* !CONFIG_PROC_UUID_POLICY */
3217 inp_update_necp_policy(struct inpcb
*inp
, struct sockaddr
*override_local_addr
, struct sockaddr
*override_remote_addr
, u_int override_bound_interface
)
3219 necp_socket_find_policy_match(inp
, override_local_addr
, override_remote_addr
, override_bound_interface
);
3220 if (necp_socket_should_rescope(inp
) &&
3221 inp
->inp_lport
== 0 &&
3222 inp
->inp_laddr
.s_addr
== INADDR_ANY
&&
3223 IN6_IS_ADDR_UNSPECIFIED(&inp
->in6p_laddr
)) {
3224 // If we should rescope, and the socket is not yet bound
3225 inp_bindif(inp
, necp_socket_get_rescope_if_index(inp
), NULL
);
3231 inp_update_policy(struct inpcb
*inp
)
3233 #if CONFIG_PROC_UUID_POLICY
3234 struct socket
*so
= inp
->inp_socket
;
3235 uint32_t pflags
= 0;
3239 if (!net_io_policy_uuid
||
3240 so
== NULL
|| inp
->inp_state
== INPCB_STATE_DEAD
)
3244 * Kernel-created sockets that aren't delegating other sockets
3245 * are currently exempted from UUID policy checks.
3247 if (so
->last_pid
== 0 && !(so
->so_flags
& SOF_DELEGATED
))
3250 ogencnt
= so
->so_policy_gencnt
;
3251 err
= proc_uuid_policy_lookup(((so
->so_flags
& SOF_DELEGATED
) ?
3252 so
->e_uuid
: so
->last_uuid
), &pflags
, &so
->so_policy_gencnt
);
3255 * Discard cached generation count if the entry is gone (ENOENT),
3256 * so that we go thru the checks below.
3258 if (err
== ENOENT
&& ogencnt
!= 0)
3259 so
->so_policy_gencnt
= 0;
3262 * If the generation count has changed, inspect the policy flags
3263 * and act accordingly. If a policy flag was previously set and
3264 * the UUID is no longer present in the table (ENOENT), treat it
3265 * as if the flag has been cleared.
3267 if ((err
== 0 || err
== ENOENT
) && ogencnt
!= so
->so_policy_gencnt
) {
3268 /* update cellular policy for this socket */
3269 if (err
== 0 && (pflags
& PROC_UUID_NO_CELLULAR
)) {
3270 inp_update_cellular_policy(inp
, TRUE
);
3271 } else if (!(pflags
& PROC_UUID_NO_CELLULAR
)) {
3272 inp_update_cellular_policy(inp
, FALSE
);
3275 /* update necp want app policy for this socket */
3276 if (err
== 0 && (pflags
& PROC_UUID_NECP_APP_POLICY
)) {
3277 inp_update_necp_want_app_policy(inp
, TRUE
);
3278 } else if (!(pflags
& PROC_UUID_NECP_APP_POLICY
)) {
3279 inp_update_necp_want_app_policy(inp
, FALSE
);
3284 return ((err
== ENOENT
) ? 0 : err
);
3285 #else /* !CONFIG_PROC_UUID_POLICY */
3288 #endif /* !CONFIG_PROC_UUID_POLICY */
3291 static unsigned int log_restricted
;
3292 SYSCTL_DECL(_net_inet
);
3293 SYSCTL_INT(_net_inet
, OID_AUTO
, log_restricted
,
3294 CTLFLAG_RW
| CTLFLAG_LOCKED
, &log_restricted
, 0,
3295 "Log network restrictions");
3297 * Called when we need to enforce policy restrictions in the input path.
3299 * Returns TRUE if we're not allowed to receive data, otherwise FALSE.
3302 _inp_restricted_recv(struct inpcb
*inp
, struct ifnet
*ifp
)
3304 VERIFY(inp
!= NULL
);
3307 * Inbound restrictions.
3309 if (!sorestrictrecv
)
3315 if (IFNET_IS_CELLULAR(ifp
) && INP_NO_CELLULAR(inp
))
3318 if (IFNET_IS_EXPENSIVE(ifp
) && INP_NO_EXPENSIVE(inp
))
3321 if (IFNET_IS_AWDL_RESTRICTED(ifp
) && !INP_AWDL_UNRESTRICTED(inp
))
3324 if (!(ifp
->if_eflags
& IFEF_RESTRICTED_RECV
))
3327 if (inp
->inp_flags
& INP_RECV_ANYIF
)
3330 if ((inp
->inp_flags
& INP_BOUND_IF
) && inp
->inp_boundifp
== ifp
)
3333 if (IFNET_IS_INTCOPROC(ifp
) && !INP_INTCOPROC_ALLOWED(inp
))
3340 inp_restricted_recv(struct inpcb
*inp
, struct ifnet
*ifp
)
3344 ret
= _inp_restricted_recv(inp
, ifp
);
3345 if (ret
== TRUE
&& log_restricted
) {
3346 printf("pid %d (%s) is unable to receive packets on %s\n",
3347 current_proc()->p_pid
, proc_best_name(current_proc()),
3354 * Called when we need to enforce policy restrictions in the output path.
3356 * Returns TRUE if we're not allowed to send data out, otherwise FALSE.
3359 _inp_restricted_send(struct inpcb
*inp
, struct ifnet
*ifp
)
3361 VERIFY(inp
!= NULL
);
3364 * Outbound restrictions.
3366 if (!sorestrictsend
)
3372 if (IFNET_IS_CELLULAR(ifp
) && INP_NO_CELLULAR(inp
))
3375 if (IFNET_IS_EXPENSIVE(ifp
) && INP_NO_EXPENSIVE(inp
))
3378 if (IFNET_IS_AWDL_RESTRICTED(ifp
) && !INP_AWDL_UNRESTRICTED(inp
))
3381 if (IFNET_IS_INTCOPROC(ifp
) && !INP_INTCOPROC_ALLOWED(inp
))
3388 inp_restricted_send(struct inpcb
*inp
, struct ifnet
*ifp
)
3392 ret
= _inp_restricted_send(inp
, ifp
);
3393 if (ret
== TRUE
&& log_restricted
) {
3394 printf("pid %d (%s) is unable to transmit packets on %s\n",
3395 current_proc()->p_pid
, proc_best_name(current_proc()),
3402 inp_count_sndbytes(struct inpcb
*inp
, u_int32_t th_ack
)
3404 struct ifnet
*ifp
= inp
->inp_last_outifp
;
3405 struct socket
*so
= inp
->inp_socket
;
3406 if (ifp
!= NULL
&& !(so
->so_flags
& SOF_MP_SUBFLOW
) &&
3407 (ifp
->if_type
== IFT_CELLULAR
||
3408 ifp
->if_subfamily
== IFNET_SUBFAMILY_WIFI
)) {
3411 so
->so_snd
.sb_flags
|= SB_SNDBYTE_CNT
;
3414 * There can be data outstanding before the connection
3415 * becomes established -- TFO case
3417 if (so
->so_snd
.sb_cc
> 0)
3418 inp_incr_sndbytes_total(so
, so
->so_snd
.sb_cc
);
3420 unsent
= inp_get_sndbytes_allunsent(so
, th_ack
);
3422 inp_incr_sndbytes_unsent(so
, unsent
);
3427 inp_incr_sndbytes_total(struct socket
*so
, int32_t len
)
3429 struct inpcb
*inp
= (struct inpcb
*)so
->so_pcb
;
3430 struct ifnet
*ifp
= inp
->inp_last_outifp
;
3433 VERIFY(ifp
->if_sndbyte_total
>= 0);
3434 OSAddAtomic64(len
, &ifp
->if_sndbyte_total
);
3439 inp_decr_sndbytes_total(struct socket
*so
, int32_t len
)
3441 struct inpcb
*inp
= (struct inpcb
*)so
->so_pcb
;
3442 struct ifnet
*ifp
= inp
->inp_last_outifp
;
3445 VERIFY(ifp
->if_sndbyte_total
>= len
);
3446 OSAddAtomic64(-len
, &ifp
->if_sndbyte_total
);
3451 inp_incr_sndbytes_unsent(struct socket
*so
, int32_t len
)
3453 struct inpcb
*inp
= (struct inpcb
*)so
->so_pcb
;
3454 struct ifnet
*ifp
= inp
->inp_last_outifp
;
3457 VERIFY(ifp
->if_sndbyte_unsent
>= 0);
3458 OSAddAtomic64(len
, &ifp
->if_sndbyte_unsent
);
3463 inp_decr_sndbytes_unsent(struct socket
*so
, int32_t len
)
3465 struct inpcb
*inp
= (struct inpcb
*)so
->so_pcb
;
3466 struct ifnet
*ifp
= inp
->inp_last_outifp
;
3468 if (so
== NULL
|| !(so
->so_snd
.sb_flags
& SB_SNDBYTE_CNT
))
3472 if (ifp
->if_sndbyte_unsent
>= len
)
3473 OSAddAtomic64(-len
, &ifp
->if_sndbyte_unsent
);
3475 ifp
->if_sndbyte_unsent
= 0;
3480 inp_decr_sndbytes_allunsent(struct socket
*so
, u_int32_t th_ack
)
3484 if (so
== NULL
|| !(so
->so_snd
.sb_flags
& SB_SNDBYTE_CNT
))
3487 len
= inp_get_sndbytes_allunsent(so
, th_ack
);
3488 inp_decr_sndbytes_unsent(so
, len
);