]> git.saurik.com Git - apple/xnu.git/blob - bsd/net/dlil.c
d679efc8b1401919f88395666ee0ad3374b6e801
[apple/xnu.git] / bsd / net / dlil.c
1 /*
2 * Copyright (c) 1999-2008 Apple Inc. All rights reserved.
3 *
4 * @APPLE_OSREFERENCE_LICENSE_HEADER_START@
5 *
6 * This file contains Original Code and/or Modifications of Original Code
7 * as defined in and that are subject to the Apple Public Source License
8 * Version 2.0 (the 'License'). You may not use this file except in
9 * compliance with the License. The rights granted to you under the License
10 * may not be used to create, or enable the creation or redistribution of,
11 * unlawful or unlicensed copies of an Apple operating system, or to
12 * circumvent, violate, or enable the circumvention or violation of, any
13 * terms of an Apple operating system software license agreement.
14 *
15 * Please obtain a copy of the License at
16 * http://www.opensource.apple.com/apsl/ and read it before using this file.
17 *
18 * The Original Code and all software distributed under the License are
19 * distributed on an 'AS IS' basis, WITHOUT WARRANTY OF ANY KIND, EITHER
20 * EXPRESS OR IMPLIED, AND APPLE HEREBY DISCLAIMS ALL SUCH WARRANTIES,
21 * INCLUDING WITHOUT LIMITATION, ANY WARRANTIES OF MERCHANTABILITY,
22 * FITNESS FOR A PARTICULAR PURPOSE, QUIET ENJOYMENT OR NON-INFRINGEMENT.
23 * Please see the License for the specific language governing rights and
24 * limitations under the License.
25 *
26 * @APPLE_OSREFERENCE_LICENSE_HEADER_END@
27 */
28 /*
29 * Data Link Inteface Layer
30 * Author: Ted Walker
31 */
32 /*
33 * NOTICE: This file was modified by SPARTA, Inc. in 2005 to introduce
34 * support for mandatory and extensible security protections. This notice
35 * is included in support of clause 2.2 (b) of the Apple Public License,
36 * Version 2.0.
37 */
38
39 #include <sys/param.h>
40 #include <sys/systm.h>
41 #include <sys/kernel.h>
42 #include <sys/malloc.h>
43 #include <sys/mbuf.h>
44 #include <sys/socket.h>
45 #include <sys/domain.h>
46 #include <sys/user.h>
47 #include <sys/random.h>
48 #include <net/if_dl.h>
49 #include <net/if.h>
50 #include <net/route.h>
51 #include <net/if_var.h>
52 #include <net/dlil.h>
53 #include <net/if_arp.h>
54 #include <sys/kern_event.h>
55 #include <sys/kdebug.h>
56
57 #include <kern/assert.h>
58 #include <kern/task.h>
59 #include <kern/thread.h>
60 #include <kern/sched_prim.h>
61 #include <kern/locks.h>
62 #include <net/kpi_protocol.h>
63
64 #include <net/if_types.h>
65 #include <net/kpi_interfacefilter.h>
66
67 #include <libkern/OSAtomic.h>
68
69 #include <machine/machine_routines.h>
70
71 #include <mach/thread_act.h>
72
73 #if CONFIG_MACF_NET
74 #include <security/mac_framework.h>
75 #endif /* MAC_NET */
76
77 #if PF
78 #include <net/pfvar.h>
79 #endif /* PF */
80
81 #define DBG_LAYER_BEG DLILDBG_CODE(DBG_DLIL_STATIC, 0)
82 #define DBG_LAYER_END DLILDBG_CODE(DBG_DLIL_STATIC, 2)
83 #define DBG_FNC_DLIL_INPUT DLILDBG_CODE(DBG_DLIL_STATIC, (1 << 8))
84 #define DBG_FNC_DLIL_OUTPUT DLILDBG_CODE(DBG_DLIL_STATIC, (2 << 8))
85 #define DBG_FNC_DLIL_IFOUT DLILDBG_CODE(DBG_DLIL_STATIC, (3 << 8))
86
87
88 #define MAX_FRAME_TYPE_SIZE 4 /* LONGWORDS */
89 #define MAX_LINKADDR 4 /* LONGWORDS */
90 #define M_NKE M_IFADDR
91
92 #if 1
93 #define DLIL_PRINTF printf
94 #else
95 #define DLIL_PRINTF kprintf
96 #endif
97
98
99 enum {
100 kProtoKPI_v1 = 1,
101 kProtoKPI_v2 = 2
102 };
103
104 struct if_proto {
105 SLIST_ENTRY(if_proto) next_hash;
106 int refcount;
107 int detaching;
108 struct ifnet *ifp;
109 struct domain *dl_domain;
110 protocol_family_t protocol_family;
111 int proto_kpi;
112 union {
113 struct {
114 proto_media_input input;
115 proto_media_preout pre_output;
116 proto_media_event event;
117 proto_media_ioctl ioctl;
118 proto_media_detached detached;
119 proto_media_resolve_multi resolve_multi;
120 proto_media_send_arp send_arp;
121 } v1;
122 struct {
123 proto_media_input_v2 input;
124 proto_media_preout pre_output;
125 proto_media_event event;
126 proto_media_ioctl ioctl;
127 proto_media_detached detached;
128 proto_media_resolve_multi resolve_multi;
129 proto_media_send_arp send_arp;
130 } v2;
131 } kpi;
132 };
133
134 SLIST_HEAD(proto_hash_entry, if_proto);
135
136
137 struct dlil_ifnet {
138 /* ifnet and drvr_ext are used by the stack and drivers
139 drvr_ext extends the public ifnet and must follow dl_if */
140 struct ifnet dl_if; /* public ifnet */
141
142 /* dlil private fields */
143 TAILQ_ENTRY(dlil_ifnet) dl_if_link; /* dlil_ifnet are link together */
144 /* it is not the ifnet list */
145 void *if_uniqueid; /* unique id identifying the interface */
146 size_t if_uniqueid_len;/* length of the unique id */
147 char if_namestorage[IFNAMSIZ]; /* interface name storage */
148 };
149
150 struct ifnet_filter {
151 TAILQ_ENTRY(ifnet_filter) filt_next;
152 ifnet_t filt_ifp;
153 int filt_detaching;
154
155 const char *filt_name;
156 void *filt_cookie;
157 protocol_family_t filt_protocol;
158 iff_input_func filt_input;
159 iff_output_func filt_output;
160 iff_event_func filt_event;
161 iff_ioctl_func filt_ioctl;
162 iff_detached_func filt_detached;
163 };
164
165 struct proto_input_entry;
166
167 static TAILQ_HEAD(, dlil_ifnet) dlil_ifnet_head;
168 static lck_grp_t *dlil_lock_group;
169 static lck_grp_t *ifnet_lock_group;
170 static lck_grp_t *ifnet_head_lock_group;
171 static lck_attr_t *ifnet_lock_attr;
172 static lck_rw_t *ifnet_head_mutex;
173 static lck_mtx_t *dlil_ifnet_mutex;
174 static lck_mtx_t *dlil_mutex;
175 static u_int32_t dlil_read_count = 0;
176 static u_int32_t dlil_detach_waiting = 0;
177 u_int32_t dlil_filter_count = 0;
178 extern u_int32_t ipv4_ll_arp_aware;
179
180 static struct dlil_threading_info dlil_lo_thread;
181 __private_extern__ struct dlil_threading_info *dlil_lo_thread_ptr = &dlil_lo_thread;
182
183 static struct mbuf *dlil_lo_input_mbuf_head = NULL;
184 static struct mbuf *dlil_lo_input_mbuf_tail = NULL;
185
186 #if IFNET_INPUT_SANITY_CHK
187 static int dlil_lo_input_mbuf_count = 0;
188 int dlil_input_sanity_check = 0; /* sanity checking of input packet lists received */
189 #endif
190 int dlil_multithreaded_input = 1;
191 static int cur_dlil_input_threads = 0;
192
193 static int dlil_event_internal(struct ifnet *ifp, struct kev_msg *msg);
194 static int dlil_detach_filter_internal(interface_filter_t filter, int detached);
195 static void dlil_call_delayed_detach_thread(void);
196
197 static void dlil_read_begin(void);
198 static __inline__ void dlil_read_end(void);
199 static int dlil_write_begin(void);
200 static void dlil_write_end(void);
201
202 #if DEBUG
203 __private_extern__ int dlil_verbose = 1;
204 #else
205 __private_extern__ int dlil_verbose = 0;
206 #endif /* DEBUG */
207
208 unsigned int net_affinity = 1;
209 static kern_return_t dlil_affinity_set(struct thread *, u_int32_t);
210
211 extern void bpfdetach(struct ifnet*);
212 extern void proto_input_run(void); // new run_netisr
213
214 void dlil_input_packet_list(struct ifnet *ifp, struct mbuf *m);
215 static void dlil_input_thread_func(struct dlil_threading_info *inpthread);
216 __private_extern__ int dlil_create_input_thread(
217 ifnet_t, struct dlil_threading_info *);
218 __private_extern__ void dlil_terminate_input_thread(
219 struct dlil_threading_info *);
220
221 __private_extern__ void link_rtrequest(int, struct rtentry *, struct sockaddr *);
222
223 int dlil_expand_mcl;
224
225 extern u_int32_t inject_buckets;
226
227 static const u_int32_t dlil_writer_waiting = 0x80000000;
228 static lck_grp_attr_t *dlil_grp_attributes = NULL;
229 static lck_attr_t *dlil_lck_attributes = NULL;
230 static lck_grp_t *dlil_input_lock_grp = NULL;
231
232 static inline void*
233 _cast_non_const(const void * ptr) {
234 union {
235 const void* cval;
236 void* val;
237 } ret;
238
239 ret.cval = ptr;
240 return (ret.val);
241 }
242
243 /* Should these be inline? */
244 static void
245 dlil_read_begin(void)
246 {
247 u_int32_t new_value;
248 u_int32_t old_value;
249 struct uthread *uth = get_bsdthread_info(current_thread());
250
251 if (uth->dlil_incremented_read == dlil_writer_waiting)
252 panic("dlil_read_begin - thread is already a writer");
253
254 do {
255 again:
256 old_value = dlil_read_count;
257
258 if ((old_value & dlil_writer_waiting) != 0 && uth->dlil_incremented_read == 0)
259 {
260 tsleep(&dlil_read_count, PRIBIO, "dlil_read_count", 1);
261 goto again;
262 }
263
264 new_value = old_value + 1;
265 } while (!OSCompareAndSwap((UInt32)old_value, (UInt32)new_value, (UInt32*)&dlil_read_count));
266
267 uth->dlil_incremented_read++;
268 }
269
270 static void
271 dlil_read_end(void)
272 {
273 struct uthread *uth = get_bsdthread_info(current_thread());
274
275 OSDecrementAtomic(&dlil_read_count);
276 uth->dlil_incremented_read--;
277 if (dlil_read_count == dlil_writer_waiting)
278 wakeup(_cast_non_const(&dlil_writer_waiting));
279 }
280
281 static int
282 dlil_write_begin(void)
283 {
284 struct uthread *uth = get_bsdthread_info(current_thread());
285
286 if (uth->dlil_incremented_read != 0) {
287 return EDEADLK;
288 }
289 lck_mtx_lock(dlil_mutex);
290 OSBitOrAtomic((UInt32)dlil_writer_waiting, &dlil_read_count);
291 again:
292 if (dlil_read_count == dlil_writer_waiting) {
293 uth->dlil_incremented_read = dlil_writer_waiting;
294 return 0;
295 }
296 else {
297 tsleep(_cast_non_const(&dlil_writer_waiting), PRIBIO, "dlil_writer_waiting", 1);
298 goto again;
299 }
300 }
301
302 static void
303 dlil_write_end(void)
304 {
305 struct uthread *uth = get_bsdthread_info(current_thread());
306
307 if (uth->dlil_incremented_read != dlil_writer_waiting)
308 panic("dlil_write_end - thread is not a writer");
309 OSBitAndAtomic((UInt32)~dlil_writer_waiting, &dlil_read_count);
310 lck_mtx_unlock(dlil_mutex);
311 uth->dlil_incremented_read = 0;
312 wakeup(&dlil_read_count);
313 }
314
315 #define PROTO_HASH_SLOTS 0x5
316
317 /*
318 * Internal functions.
319 */
320
321 static int
322 proto_hash_value(u_int32_t protocol_family)
323 {
324 /*
325 * dlil_proto_unplumb_all() depends on the mapping between
326 * the hash bucket index and the protocol family defined
327 * here; future changes must be applied there as well.
328 */
329 switch(protocol_family) {
330 case PF_INET:
331 return 0;
332 case PF_INET6:
333 return 1;
334 case PF_APPLETALK:
335 return 2;
336 case PF_VLAN:
337 return 3;
338 default:
339 return 4;
340 }
341 }
342
343 static struct if_proto*
344 find_attached_proto(struct ifnet *ifp, u_int32_t protocol_family)
345 {
346 struct if_proto *proto = NULL;
347 u_int32_t i = proto_hash_value(protocol_family);
348 if (ifp->if_proto_hash) {
349 proto = SLIST_FIRST(&ifp->if_proto_hash[i]);
350 }
351
352 while(proto && proto->protocol_family != protocol_family) {
353 proto = SLIST_NEXT(proto, next_hash);
354 }
355
356 return proto;
357 }
358
359 static void
360 if_proto_ref(struct if_proto *proto)
361 {
362 OSAddAtomic(1, &proto->refcount);
363 }
364
365 static void
366 if_proto_free(struct if_proto *proto)
367 {
368 int oldval = OSAddAtomic(-1, &proto->refcount);
369
370 if (oldval == 1) { /* This was the last reference */
371 FREE(proto, M_IFADDR);
372 }
373 }
374
375 __private_extern__ void
376 ifnet_lock_assert(
377 __unused struct ifnet *ifp,
378 __unused int what)
379 {
380 #if IFNET_RW_LOCK
381 /*
382 * Not implemented for rw locks.
383 *
384 * Function exists so when/if we use mutex we can
385 * enable this check.
386 */
387 #else
388 lck_mtx_assert(ifp->if_lock, what);
389 #endif
390 }
391
392 __private_extern__ void
393 ifnet_lock_shared(
394 struct ifnet *ifp)
395 {
396 #if IFNET_RW_LOCK
397 lck_rw_lock_shared(ifp->if_lock);
398 #else
399 lck_mtx_assert(ifp->if_lock, LCK_MTX_ASSERT_NOTOWNED);
400 lck_mtx_lock(ifp->if_lock);
401 #endif
402 }
403
404 __private_extern__ void
405 ifnet_lock_exclusive(
406 struct ifnet *ifp)
407 {
408 #if IFNET_RW_LOCK
409 lck_rw_lock_exclusive(ifp->if_lock);
410 #else
411 lck_mtx_assert(ifp->if_lock, LCK_MTX_ASSERT_NOTOWNED);
412 lck_mtx_lock(ifp->if_lock);
413 #endif
414 }
415
416 __private_extern__ void
417 ifnet_lock_done(
418 struct ifnet *ifp)
419 {
420 #if IFNET_RW_LOCK
421 lck_rw_done(ifp->if_lock);
422 #else
423 lck_mtx_assert(ifp->if_lock, LCK_MTX_ASSERT_OWNED);
424 lck_mtx_unlock(ifp->if_lock);
425 #endif
426 }
427
428 __private_extern__ void
429 ifnet_head_lock_shared(void)
430 {
431 lck_rw_lock_shared(ifnet_head_mutex);
432 }
433
434 __private_extern__ void
435 ifnet_head_lock_exclusive(void)
436 {
437 lck_rw_lock_exclusive(ifnet_head_mutex);
438 }
439
440 __private_extern__ void
441 ifnet_head_done(void)
442 {
443 lck_rw_done(ifnet_head_mutex);
444 }
445
446 static int dlil_ifp_proto_count(struct ifnet * ifp)
447 {
448 int count = 0;
449 int i;
450
451 if (ifp->if_proto_hash != NULL) {
452 for (i = 0; i < PROTO_HASH_SLOTS; i++) {
453 struct if_proto *proto;
454 SLIST_FOREACH(proto, &ifp->if_proto_hash[i], next_hash) {
455 count++;
456 }
457 }
458 }
459
460 return count;
461 }
462
463 __private_extern__ void
464 dlil_post_msg(struct ifnet *ifp, u_int32_t event_subclass, u_int32_t event_code,
465 struct net_event_data *event_data, u_int32_t event_data_len)
466 {
467 struct net_event_data ev_data;
468 struct kev_msg ev_msg;
469
470 /*
471 * a net event always starts with a net_event_data structure
472 * but the caller can generate a simple net event or
473 * provide a longer event structure to post
474 */
475
476 ev_msg.vendor_code = KEV_VENDOR_APPLE;
477 ev_msg.kev_class = KEV_NETWORK_CLASS;
478 ev_msg.kev_subclass = event_subclass;
479 ev_msg.event_code = event_code;
480
481 if (event_data == 0) {
482 event_data = &ev_data;
483 event_data_len = sizeof(struct net_event_data);
484 }
485
486 strncpy(&event_data->if_name[0], ifp->if_name, IFNAMSIZ);
487 event_data->if_family = ifp->if_family;
488 event_data->if_unit = (u_int32_t) ifp->if_unit;
489
490 ev_msg.dv[0].data_length = event_data_len;
491 ev_msg.dv[0].data_ptr = event_data;
492 ev_msg.dv[1].data_length = 0;
493
494 dlil_event_internal(ifp, &ev_msg);
495 }
496
497 __private_extern__ int
498 dlil_create_input_thread(
499 ifnet_t ifp, struct dlil_threading_info *inputthread)
500 {
501 int error;
502
503 bzero(inputthread, sizeof(*inputthread));
504 // loopback ifp may not be configured at dlil_init time.
505 if (ifp == lo_ifp)
506 strlcat(inputthread->input_name, "dlil_input_main_thread_mtx", 32);
507 else
508 snprintf(inputthread->input_name, 32, "dlil_input_%s%d_mtx", ifp->if_name, ifp->if_unit);
509
510 inputthread->lck_grp = lck_grp_alloc_init(inputthread->input_name, dlil_grp_attributes);
511 inputthread->input_lck = lck_mtx_alloc_init(inputthread->lck_grp, dlil_lck_attributes);
512
513 error= kernel_thread_start((thread_continue_t)dlil_input_thread_func, inputthread, &inputthread->input_thread);
514 if (error == 0) {
515 ml_thread_policy(inputthread->input_thread, MACHINE_GROUP,
516 (MACHINE_NETWORK_GROUP|MACHINE_NETWORK_NETISR));
517 /*
518 * Except for the loopback dlil input thread, we create
519 * an affinity set so that the matching workloop thread
520 * can be scheduled on the same processor set.
521 */
522 if (net_affinity && inputthread != dlil_lo_thread_ptr) {
523 struct thread *tp = inputthread->input_thread;
524 u_int32_t tag;
525 /*
526 * Randomize to reduce the probability
527 * of affinity tag namespace collision.
528 */
529 read_random(&tag, sizeof (tag));
530 if (dlil_affinity_set(tp, tag) == KERN_SUCCESS) {
531 thread_reference(tp);
532 inputthread->tag = tag;
533 inputthread->net_affinity = TRUE;
534 }
535 }
536 } else {
537 panic("dlil_create_input_thread: couldn't create thread\n");
538 }
539 OSAddAtomic(1, &cur_dlil_input_threads);
540 #if DLIL_DEBUG
541 printf("dlil_create_input_thread: threadinfo: %p input_thread=%p threads: cur=%d max=%d\n",
542 inputthread, inputthread->input_thread, dlil_multithreaded_input, cur_dlil_input_threads);
543 #endif
544 return error;
545 }
546 __private_extern__ void
547 dlil_terminate_input_thread(
548 struct dlil_threading_info *inputthread)
549 {
550 OSAddAtomic(-1, &cur_dlil_input_threads);
551
552 lck_mtx_unlock(inputthread->input_lck);
553 lck_mtx_free(inputthread->input_lck, inputthread->lck_grp);
554 lck_grp_free(inputthread->lck_grp);
555
556 FREE(inputthread, M_NKE);
557
558 /* For the extra reference count from kernel_thread_start() */
559 thread_deallocate(current_thread());
560
561 thread_terminate(current_thread());
562 }
563
564 static kern_return_t
565 dlil_affinity_set(struct thread *tp, u_int32_t tag)
566 {
567 thread_affinity_policy_data_t policy;
568
569 bzero(&policy, sizeof (policy));
570 policy.affinity_tag = tag;
571 return (thread_policy_set(tp, THREAD_AFFINITY_POLICY,
572 (thread_policy_t)&policy, THREAD_AFFINITY_POLICY_COUNT));
573 }
574
575 void
576 dlil_init(void)
577 {
578 thread_t thread = THREAD_NULL;
579
580 PE_parse_boot_argn("net_affinity", &net_affinity, sizeof (net_affinity));
581
582 TAILQ_INIT(&dlil_ifnet_head);
583 TAILQ_INIT(&ifnet_head);
584
585 /* Setup the lock groups we will use */
586 dlil_grp_attributes = lck_grp_attr_alloc_init();
587
588 dlil_lock_group = lck_grp_alloc_init("dlil internal locks", dlil_grp_attributes);
589 ifnet_lock_group = lck_grp_alloc_init("ifnet locks", dlil_grp_attributes);
590 ifnet_head_lock_group = lck_grp_alloc_init("ifnet head lock", dlil_grp_attributes);
591 dlil_input_lock_grp = lck_grp_alloc_init("dlil input lock", dlil_grp_attributes);
592
593 /* Setup the lock attributes we will use */
594 dlil_lck_attributes = lck_attr_alloc_init();
595
596 ifnet_lock_attr = lck_attr_alloc_init();
597
598
599 ifnet_head_mutex = lck_rw_alloc_init(ifnet_head_lock_group, dlil_lck_attributes);
600 dlil_ifnet_mutex = lck_mtx_alloc_init(dlil_lock_group, dlil_lck_attributes);
601 dlil_mutex = lck_mtx_alloc_init(dlil_lock_group, dlil_lck_attributes);
602
603 lck_attr_free(dlil_lck_attributes);
604 dlil_lck_attributes = NULL;
605
606 /*
607 * Create and start up the first dlil input thread once everything is initialized
608 */
609 dlil_create_input_thread(0, dlil_lo_thread_ptr);
610
611 (void) kernel_thread_start((thread_continue_t)dlil_call_delayed_detach_thread, NULL, &thread);
612 thread_deallocate(thread);
613 #if PF
614 /* Initialize the packet filter */
615 pfinit();
616 #endif /* PF */
617 }
618
619 __private_extern__ int
620 dlil_attach_filter(
621 struct ifnet *ifp,
622 const struct iff_filter *if_filter,
623 interface_filter_t *filter_ref)
624 {
625 int retval = 0;
626 struct ifnet_filter *filter;
627
628 MALLOC(filter, struct ifnet_filter *, sizeof(*filter), M_NKE, M_WAITOK);
629 if (filter == NULL)
630 return ENOMEM;
631 bzero(filter, sizeof(*filter));
632
633
634 filter->filt_ifp = ifp;
635 filter->filt_cookie = if_filter->iff_cookie;
636 filter->filt_name = if_filter->iff_name;
637 filter->filt_protocol = if_filter->iff_protocol;
638 filter->filt_input = if_filter->iff_input;
639 filter->filt_output = if_filter->iff_output;
640 filter->filt_event = if_filter->iff_event;
641 filter->filt_ioctl = if_filter->iff_ioctl;
642 filter->filt_detached = if_filter->iff_detached;
643
644 if ((retval = dlil_write_begin()) != 0) {
645 /* Failed to acquire the write lock */
646 FREE(filter, M_NKE);
647 return retval;
648 }
649 TAILQ_INSERT_TAIL(&ifp->if_flt_head, filter, filt_next);
650 dlil_write_end();
651 *filter_ref = filter;
652
653 /*
654 * Bump filter count and route_generation ID to let TCP
655 * know it shouldn't do TSO on this connection
656 */
657 OSAddAtomic(1, &dlil_filter_count);
658 if (use_routegenid)
659 routegenid_update();
660
661 return retval;
662 }
663
664 static int
665 dlil_detach_filter_internal(
666 interface_filter_t filter,
667 int detached)
668 {
669 int retval = 0;
670
671 if (detached == 0) {
672 ifnet_t ifp = NULL;
673 interface_filter_t entry = NULL;
674
675 /* Take the write lock */
676 retval = dlil_write_begin();
677 if (retval != 0 && retval != EDEADLK)
678 return retval;
679
680 /*
681 * At this point either we have the write lock (retval == 0)
682 * or we couldn't get it (retval == EDEADLK) because someone
683 * else up the stack is holding the read lock. It is safe to
684 * read, either the read or write is held. Verify the filter
685 * parameter before proceeding.
686 */
687 ifnet_head_lock_shared();
688 TAILQ_FOREACH(ifp, &ifnet_head, if_link) {
689 TAILQ_FOREACH(entry, &ifp->if_flt_head, filt_next) {
690 if (entry == filter)
691 break;
692 }
693 if (entry == filter)
694 break;
695 }
696 ifnet_head_done();
697
698 if (entry != filter) {
699 /* filter parameter is not a valid filter ref */
700 if (retval == 0) {
701 dlil_write_end();
702 }
703 return EINVAL;
704 }
705
706 if (retval == EDEADLK) {
707 /* Perform a delayed detach */
708 filter->filt_detaching = 1;
709 dlil_detach_waiting = 1;
710 wakeup(&dlil_detach_waiting);
711 return 0;
712 }
713
714 /* Remove the filter from the list */
715 TAILQ_REMOVE(&ifp->if_flt_head, filter, filt_next);
716 dlil_write_end();
717 }
718
719 /* Call the detached funciton if there is one */
720 if (filter->filt_detached)
721 filter->filt_detached(filter->filt_cookie, filter->filt_ifp);
722
723 /* Free the filter */
724 FREE(filter, M_NKE);
725
726 /*
727 * Decrease filter count and route_generation ID to let TCP
728 * know it should reevalute doing TSO or not
729 */
730 OSAddAtomic(-1, &dlil_filter_count);
731 if (use_routegenid)
732 routegenid_update();
733
734 return retval;
735 }
736
737 __private_extern__ void
738 dlil_detach_filter(interface_filter_t filter)
739 {
740 if (filter == NULL)
741 return;
742 dlil_detach_filter_internal(filter, 0);
743 }
744
745 static void
746 dlil_input_thread_func(
747 struct dlil_threading_info *inputthread)
748 {
749 while (1) {
750 struct mbuf *m = NULL, *m_loop = NULL;
751 #if IFNET_INPUT_SANITY_CHK
752 int loop_cnt = 0, mbuf_cnt;
753 int count;
754 struct mbuf *m1;
755 #endif /* IFNET_INPUT_SANITY_CHK */
756
757 lck_mtx_lock(inputthread->input_lck);
758
759 /* Wait until there is work to be done */
760 while ((inputthread->input_waiting & ~DLIL_INPUT_RUNNING) == 0) {
761 inputthread->input_waiting &= ~DLIL_INPUT_RUNNING;
762 msleep(&inputthread->input_waiting, inputthread->input_lck, 0, inputthread->input_name, 0);
763 }
764
765
766 lck_mtx_assert(inputthread->input_lck, LCK_MTX_ASSERT_OWNED);
767
768 m = inputthread->mbuf_head;
769 inputthread->mbuf_head = NULL;
770 inputthread->mbuf_tail = NULL;
771
772 if (inputthread->input_waiting & DLIL_INPUT_TERMINATE) {
773 if (m)
774 mbuf_freem_list(m);
775 /* this is the end */
776 dlil_terminate_input_thread(inputthread);
777 return;
778 }
779
780 inputthread->input_waiting |= DLIL_INPUT_RUNNING;
781 inputthread->input_waiting &= ~DLIL_INPUT_WAITING;
782
783 if (inputthread == dlil_lo_thread_ptr) {
784 m_loop = dlil_lo_input_mbuf_head;
785 dlil_lo_input_mbuf_head = NULL;
786 dlil_lo_input_mbuf_tail = NULL;
787 }
788
789 #if IFNET_INPUT_SANITY_CHK
790 if (dlil_input_sanity_check != 0) {
791 mbuf_cnt = inputthread->mbuf_count;
792 inputthread->mbuf_count = 0;
793 if (inputthread == dlil_lo_thread_ptr) {
794 loop_cnt = dlil_lo_input_mbuf_count;
795 dlil_lo_input_mbuf_count = 0;
796 }
797
798 lck_mtx_unlock(inputthread->input_lck);
799
800 for (m1 = m, count = 0; m1; m1 = mbuf_nextpkt(m1)) {
801 count++;
802 }
803 if (count != mbuf_cnt) {
804 panic("dlil_input_func - thread=%p reg. loop queue has %d packets, should have %d\n",
805 inputthread, count, mbuf_cnt);
806 }
807
808 if (inputthread == dlil_lo_thread_ptr) {
809 for (m1 = m_loop, count = 0; m1; m1 = mbuf_nextpkt(m1)) {
810 count++;
811 }
812 if (count != loop_cnt) {
813 panic("dlil_input_func - thread=%p loop queue has %d packets, should have %d\n",
814 inputthread, count, loop_cnt);
815 }
816 }
817 } else
818 #endif /* IFNET_INPUT_SANITY_CHK */
819 {
820 lck_mtx_unlock(inputthread->input_lck);
821 }
822
823
824 /*
825 * NOTE warning %%% attention !!!!
826 * We should think about putting some thread starvation safeguards if
827 * we deal with long chains of packets.
828 */
829 if (m_loop) {
830 if (inputthread == dlil_lo_thread_ptr)
831 dlil_input_packet_list(lo_ifp, m_loop);
832 #if IFNET_INPUT_SANITY_CHK
833 else
834 panic("dlil_input_func - thread=%p loop queue has %d packets, should have none!\n",
835 inputthread, loop_cnt);
836 #endif /* IFNET_INPUT_SANITY_CHK */
837 }
838
839
840 if (m)
841 dlil_input_packet_list(0, m);
842
843
844 lck_mtx_lock(inputthread->input_lck);
845
846 if ((inputthread->input_waiting & (DLIL_PROTO_WAITING | DLIL_PROTO_REGISTER)) != 0) {
847 lck_mtx_unlock(inputthread->input_lck);
848 proto_input_run();
849 }
850 else
851 lck_mtx_unlock(inputthread->input_lck);
852 }
853 }
854
855 errno_t
856 ifnet_input(
857 ifnet_t ifp,
858 mbuf_t m_head,
859 const struct ifnet_stat_increment_param *stats)
860 {
861 struct thread *tp = current_thread();
862 mbuf_t m_tail;
863 struct dlil_threading_info *inp;
864 #if IFNET_INPUT_SANITY_CHK
865 u_int32_t pkt_count = 0;
866 #endif /* IFNET_INPUT_SANITY_CHK */
867
868 if (ifp == NULL || m_head == NULL) {
869 if (m_head)
870 mbuf_freem_list(m_head);
871 return EINVAL;
872 }
873
874 m_tail = m_head;
875 while (1) {
876 #if IFNET_INPUT_SANITY_CHK
877 if (dlil_input_sanity_check != 0) {
878 ifnet_t rcvif;
879
880 rcvif = mbuf_pkthdr_rcvif(m_tail);
881 pkt_count++;
882
883 if (rcvif == NULL ||
884 (ifp->if_type != IFT_LOOP && rcvif != ifp) ||
885 (mbuf_flags(m_head) & MBUF_PKTHDR) == 0) {
886 panic("ifnet_input - invalid mbuf %p\n", m_tail);
887 }
888 }
889 #endif /* IFNET_INPUT_SANITY_CHK */
890 if (mbuf_nextpkt(m_tail) == NULL)
891 break;
892 m_tail = mbuf_nextpkt(m_tail);
893 }
894
895 inp = ifp->if_input_thread;
896
897 if (dlil_multithreaded_input == 0 || inp == NULL)
898 inp = dlil_lo_thread_ptr;
899
900 /*
901 * If there is a matching dlil input thread associated with an
902 * affinity set, associate this workloop thread with the same set.
903 * We will only do this once.
904 */
905 lck_mtx_lock(inp->input_lck);
906 if (inp->net_affinity && inp->workloop_thread == NULL) {
907 u_int32_t tag = inp->tag;
908 inp->workloop_thread = tp;
909 lck_mtx_unlock(inp->input_lck);
910
911 /* Associated the current thread with the new affinity tag */
912 (void) dlil_affinity_set(tp, tag);
913
914 /*
915 * Take a reference on the workloop (current) thread; during
916 * detach, we will need to refer to it in order ot tear down
917 * its affinity.
918 */
919 thread_reference(tp);
920 lck_mtx_lock(inp->input_lck);
921 }
922
923 /* WARNING
924 * Because of loopbacked multicast we cannot stuff the ifp in
925 * the rcvif of the packet header: loopback has its own dlil
926 * input queue
927 */
928
929 if (inp == dlil_lo_thread_ptr && ifp->if_type == IFT_LOOP) {
930 if (dlil_lo_input_mbuf_head == NULL)
931 dlil_lo_input_mbuf_head = m_head;
932 else if (dlil_lo_input_mbuf_tail != NULL)
933 dlil_lo_input_mbuf_tail->m_nextpkt = m_head;
934 dlil_lo_input_mbuf_tail = m_tail;
935 #if IFNET_INPUT_SANITY_CHK
936 if (dlil_input_sanity_check != 0) {
937 dlil_lo_input_mbuf_count += pkt_count;
938 inp->input_mbuf_cnt += pkt_count;
939 inp->input_wake_cnt++;
940
941 lck_mtx_assert(inp->input_lck, LCK_MTX_ASSERT_OWNED);
942 }
943 #endif
944 }
945 else {
946 if (inp->mbuf_head == NULL)
947 inp->mbuf_head = m_head;
948 else if (inp->mbuf_tail != NULL)
949 inp->mbuf_tail->m_nextpkt = m_head;
950 inp->mbuf_tail = m_tail;
951 #if IFNET_INPUT_SANITY_CHK
952 if (dlil_input_sanity_check != 0) {
953 inp->mbuf_count += pkt_count;
954 inp->input_mbuf_cnt += pkt_count;
955 inp->input_wake_cnt++;
956
957 lck_mtx_assert(inp->input_lck, LCK_MTX_ASSERT_OWNED);
958 }
959 #endif
960 }
961
962
963 inp->input_waiting |= DLIL_INPUT_WAITING;
964 if ((inp->input_waiting & DLIL_INPUT_RUNNING) == 0) {
965 wakeup((caddr_t)&inp->input_waiting);
966 }
967 if (stats) {
968 ifp->if_data.ifi_ipackets += stats->packets_in;
969 ifp->if_data.ifi_ibytes += stats->bytes_in;
970 ifp->if_data.ifi_ierrors += stats->errors_in;
971
972 ifp->if_data.ifi_opackets += stats->packets_out;
973 ifp->if_data.ifi_obytes += stats->bytes_out;
974 ifp->if_data.ifi_oerrors += stats->errors_out;
975
976 ifp->if_data.ifi_collisions += stats->collisions;
977 ifp->if_data.ifi_iqdrops += stats->dropped;
978 }
979
980 lck_mtx_unlock(inp->input_lck);
981
982 return 0;
983 }
984
985 static int
986 dlil_interface_filters_input(struct ifnet * ifp, struct mbuf * * m_p,
987 char * * frame_header_p,
988 protocol_family_t protocol_family)
989 {
990 struct ifnet_filter * filter;
991
992 TAILQ_FOREACH(filter, &ifp->if_flt_head, filt_next) {
993 int result;
994
995 if (filter->filt_input
996 && (filter->filt_protocol == 0
997 || filter->filt_protocol == protocol_family)) {
998 result = (*filter->filt_input)(filter->filt_cookie,
999 ifp, protocol_family,
1000 m_p, frame_header_p);
1001 if (result != 0) {
1002 return (result);
1003 }
1004 }
1005 }
1006 return (0);
1007 }
1008
1009 static void
1010 dlil_ifproto_input(struct if_proto * ifproto, mbuf_t m)
1011 {
1012 int error;
1013
1014 if (ifproto->proto_kpi == kProtoKPI_v1) {
1015 /* Version 1 protocols get one packet at a time */
1016 while (m != NULL) {
1017 char * frame_header;
1018 mbuf_t next_packet;
1019
1020 next_packet = m->m_nextpkt;
1021 m->m_nextpkt = NULL;
1022 frame_header = m->m_pkthdr.header;
1023 m->m_pkthdr.header = NULL;
1024 error = (*ifproto->kpi.v1.input)(ifproto->ifp,
1025 ifproto->protocol_family,
1026 m, frame_header);
1027 if (error != 0 && error != EJUSTRETURN)
1028 m_freem(m);
1029 m = next_packet;
1030 }
1031 }
1032 else if (ifproto->proto_kpi == kProtoKPI_v2) {
1033 /* Version 2 protocols support packet lists */
1034 error = (*ifproto->kpi.v2.input)(ifproto->ifp,
1035 ifproto->protocol_family,
1036 m);
1037 if (error != 0 && error != EJUSTRETURN)
1038 m_freem_list(m);
1039 }
1040 return;
1041 }
1042
1043 __private_extern__ void
1044 dlil_input_packet_list(struct ifnet * ifp_param, struct mbuf *m)
1045 {
1046 int error = 0;
1047 int locked = 0;
1048 protocol_family_t protocol_family;
1049 mbuf_t next_packet;
1050 ifnet_t ifp = ifp_param;
1051 char * frame_header;
1052 struct if_proto * last_ifproto = NULL;
1053 mbuf_t pkt_first = NULL;
1054 mbuf_t * pkt_next = NULL;
1055
1056 KERNEL_DEBUG(DBG_FNC_DLIL_INPUT | DBG_FUNC_START,0,0,0,0,0);
1057
1058 while (m != NULL) {
1059 struct if_proto * ifproto = NULL;
1060
1061 next_packet = m->m_nextpkt;
1062 m->m_nextpkt = NULL;
1063 if (ifp_param == NULL)
1064 ifp = m->m_pkthdr.rcvif;
1065 frame_header = m->m_pkthdr.header;
1066 m->m_pkthdr.header = NULL;
1067
1068 if (locked == 0) {
1069 /* dlil lock protects the demux and interface filters */
1070 locked = 1;
1071 dlil_read_begin();
1072 }
1073 /* find which protocol family this packet is for */
1074 error = (*ifp->if_demux)(ifp, m, frame_header,
1075 &protocol_family);
1076 if (error != 0) {
1077 if (error == EJUSTRETURN) {
1078 goto next;
1079 }
1080 protocol_family = 0;
1081 }
1082
1083 /* DANGER!!! */
1084 if (m->m_flags & (M_BCAST|M_MCAST))
1085 ifp->if_imcasts++;
1086
1087 /* run interface filters, exclude VLAN packets PR-3586856 */
1088 if ((m->m_pkthdr.csum_flags & CSUM_VLAN_TAG_VALID) == 0) {
1089 int filter_result;
1090
1091 filter_result = dlil_interface_filters_input(ifp, &m,
1092 &frame_header,
1093 protocol_family);
1094 if (filter_result != 0) {
1095 if (filter_result != EJUSTRETURN) {
1096 m_freem(m);
1097 }
1098 goto next;
1099 }
1100 }
1101 if (error != 0 || ((m->m_flags & M_PROMISC) != 0) ) {
1102 m_freem(m);
1103 goto next;
1104 }
1105
1106 /* Lookup the protocol attachment to this interface */
1107 if (protocol_family == 0) {
1108 ifproto = NULL;
1109 }
1110 else if (last_ifproto != NULL
1111 && last_ifproto->ifp == ifp
1112 && (last_ifproto->protocol_family
1113 == protocol_family)) {
1114 ifproto = last_ifproto;
1115 }
1116 else {
1117 ifproto = find_attached_proto(ifp, protocol_family);
1118 }
1119 if (ifproto == NULL) {
1120 /* no protocol for this packet, discard */
1121 m_freem(m);
1122 goto next;
1123 }
1124 if (ifproto != last_ifproto) {
1125 /* make sure ifproto can't go away during input */
1126 if_proto_ref(ifproto);
1127 if (last_ifproto != NULL) {
1128 /* pass up the list for the previous protocol */
1129 dlil_read_end();
1130
1131 dlil_ifproto_input(last_ifproto, pkt_first);
1132 pkt_first = NULL;
1133 if_proto_free(last_ifproto);
1134 dlil_read_begin();
1135 }
1136 last_ifproto = ifproto;
1137 }
1138 /* extend the list */
1139 m->m_pkthdr.header = frame_header;
1140 if (pkt_first == NULL) {
1141 pkt_first = m;
1142 } else {
1143 *pkt_next = m;
1144 }
1145 pkt_next = &m->m_nextpkt;
1146
1147 next:
1148 if (next_packet == NULL && last_ifproto != NULL) {
1149 /* pass up the last list of packets */
1150 dlil_read_end();
1151
1152 dlil_ifproto_input(last_ifproto, pkt_first);
1153 if_proto_free(last_ifproto);
1154 locked = 0;
1155 }
1156 m = next_packet;
1157
1158 }
1159 if (locked != 0) {
1160 dlil_read_end();
1161 }
1162 KERNEL_DEBUG(DBG_FNC_DLIL_INPUT | DBG_FUNC_END,0,0,0,0,0);
1163 return;
1164 }
1165
1166 static int
1167 dlil_event_internal(struct ifnet *ifp, struct kev_msg *event)
1168 {
1169 struct ifnet_filter *filter;
1170
1171 if (ifp_use(ifp, kIfNetUseCount_MustNotBeZero) == 0) {
1172 dlil_read_begin();
1173
1174 /* Pass the event to the interface filters */
1175 TAILQ_FOREACH(filter, &ifp->if_flt_head, filt_next) {
1176 if (filter->filt_event)
1177 filter->filt_event(filter->filt_cookie, ifp, filter->filt_protocol, event);
1178 }
1179
1180 if (ifp->if_proto_hash) {
1181 int i;
1182
1183 for (i = 0; i < PROTO_HASH_SLOTS; i++) {
1184 struct if_proto *proto;
1185
1186 SLIST_FOREACH(proto, &ifp->if_proto_hash[i], next_hash) {
1187 proto_media_event eventp = proto->proto_kpi == kProtoKPI_v1
1188 ? proto->kpi.v1.event : proto->kpi.v2.event;
1189
1190 if (eventp)
1191 eventp(ifp, proto->protocol_family, event);
1192 }
1193 }
1194 }
1195
1196 dlil_read_end();
1197
1198 /* Pass the event to the interface */
1199 if (ifp->if_event)
1200 ifp->if_event(ifp, event);
1201
1202 if (ifp_unuse(ifp))
1203 ifp_use_reached_zero(ifp);
1204 }
1205
1206 return kev_post_msg(event);
1207 }
1208
1209 errno_t
1210 ifnet_event(
1211 ifnet_t ifp,
1212 struct kern_event_msg *event)
1213 {
1214 struct kev_msg kev_msg;
1215 int result = 0;
1216
1217 if (ifp == NULL || event == NULL) return EINVAL;
1218
1219 kev_msg.vendor_code = event->vendor_code;
1220 kev_msg.kev_class = event->kev_class;
1221 kev_msg.kev_subclass = event->kev_subclass;
1222 kev_msg.event_code = event->event_code;
1223 kev_msg.dv[0].data_ptr = &event->event_data[0];
1224 kev_msg.dv[0].data_length = event->total_size - KEV_MSG_HEADER_SIZE;
1225 kev_msg.dv[1].data_length = 0;
1226
1227 result = dlil_event_internal(ifp, &kev_msg);
1228
1229 return result;
1230 }
1231
1232 #if CONFIG_MACF_NET
1233 #include <netinet/ip6.h>
1234 #include <netinet/ip.h>
1235 static int dlil_get_socket_type(struct mbuf **mp, int family, int raw)
1236 {
1237 struct mbuf *m;
1238 struct ip *ip;
1239 struct ip6_hdr *ip6;
1240 int type = SOCK_RAW;
1241
1242 if (!raw) {
1243 switch (family) {
1244 case PF_INET:
1245 m = m_pullup(*mp, sizeof(struct ip));
1246 if (m == NULL)
1247 break;
1248 *mp = m;
1249 ip = mtod(m, struct ip *);
1250 if (ip->ip_p == IPPROTO_TCP)
1251 type = SOCK_STREAM;
1252 else if (ip->ip_p == IPPROTO_UDP)
1253 type = SOCK_DGRAM;
1254 break;
1255 case PF_INET6:
1256 m = m_pullup(*mp, sizeof(struct ip6_hdr));
1257 if (m == NULL)
1258 break;
1259 *mp = m;
1260 ip6 = mtod(m, struct ip6_hdr *);
1261 if (ip6->ip6_nxt == IPPROTO_TCP)
1262 type = SOCK_STREAM;
1263 else if (ip6->ip6_nxt == IPPROTO_UDP)
1264 type = SOCK_DGRAM;
1265 break;
1266 }
1267 }
1268
1269 return (type);
1270 }
1271 #endif
1272
1273 #if 0
1274 int
1275 dlil_output_list(
1276 struct ifnet* ifp,
1277 u_long proto_family,
1278 struct mbuf *packetlist,
1279 caddr_t route,
1280 const struct sockaddr *dest,
1281 int raw)
1282 {
1283 char *frame_type = NULL;
1284 char *dst_linkaddr = NULL;
1285 int retval = 0;
1286 char frame_type_buffer[MAX_FRAME_TYPE_SIZE * 4];
1287 char dst_linkaddr_buffer[MAX_LINKADDR * 4];
1288 struct ifnet_filter *filter;
1289 struct if_proto *proto = 0;
1290 mbuf_t m;
1291 mbuf_t send_head = NULL;
1292 mbuf_t *send_tail = &send_head;
1293
1294 KERNEL_DEBUG(DBG_FNC_DLIL_OUTPUT | DBG_FUNC_START,0,0,0,0,0);
1295
1296 dlil_read_begin();
1297
1298 frame_type = frame_type_buffer;
1299 dst_linkaddr = dst_linkaddr_buffer;
1300
1301 if (raw == 0) {
1302 proto = find_attached_proto(ifp, proto_family);
1303 if (proto == NULL) {
1304 retval = ENXIO;
1305 goto cleanup;
1306 }
1307 }
1308
1309 preout_again:
1310 if (packetlist == NULL)
1311 goto cleanup;
1312 m = packetlist;
1313 packetlist = packetlist->m_nextpkt;
1314 m->m_nextpkt = NULL;
1315
1316 if (raw == 0) {
1317 proto_media_preout preoutp = proto->proto_kpi == kProtoKPI_v1
1318 ? proto->kpi.v1.pre_output : proto->kpi.v2.pre_output;
1319 retval = 0;
1320 if (preoutp)
1321 retval = preoutp(ifp, proto_family, &m, dest, route, frame_type, dst_linkaddr);
1322
1323 if (retval) {
1324 if (retval == EJUSTRETURN) {
1325 goto preout_again;
1326 }
1327
1328 m_freem(m);
1329 goto cleanup;
1330 }
1331 }
1332
1333 do {
1334 #if CONFIG_MACF_NET
1335 retval = mac_ifnet_check_transmit(ifp, m, proto_family,
1336 dlil_get_socket_type(&m, proto_family, raw));
1337 if (retval) {
1338 m_freem(m);
1339 goto cleanup;
1340 }
1341 #endif
1342
1343 if (raw == 0 && ifp->if_framer) {
1344 retval = ifp->if_framer(ifp, &m, dest, dst_linkaddr, frame_type);
1345 if (retval) {
1346 if (retval != EJUSTRETURN) {
1347 m_freem(m);
1348 }
1349 goto next;
1350 }
1351 }
1352
1353 #if BRIDGE
1354 /* !!!LOCKING!!!
1355 *
1356 * Need to consider how to handle this.
1357 * Also note that return should be a goto cleanup
1358 */
1359 broken-locking
1360 if (do_bridge) {
1361 struct mbuf *m0 = m;
1362 struct ether_header *eh = mtod(m, struct ether_header *);
1363
1364 if (m->m_pkthdr.rcvif)
1365 m->m_pkthdr.rcvif = NULL;
1366 ifp = bridge_dst_lookup(eh);
1367 bdg_forward(&m0, ifp);
1368 if (m0)
1369 m_freem(m0);
1370
1371 return 0 - should be goto cleanup?
1372 }
1373 #endif
1374
1375 /*
1376 * Let interface filters (if any) do their thing ...
1377 */
1378 /* Do not pass VLAN tagged packets to filters PR-3586856 */
1379 if ((m->m_pkthdr.csum_flags & CSUM_VLAN_TAG_VALID) == 0) {
1380 TAILQ_FOREACH(filter, &ifp->if_flt_head, filt_next) {
1381 if ((filter->filt_protocol == 0 || (filter->filt_protocol == proto_family)) &&
1382 filter->filt_output) {
1383 retval = filter->filt_output(filter->filt_cookie, ifp, proto_family, &m);
1384 if (retval) {
1385 if (retval != EJUSTRETURN)
1386 m_freem(m);
1387 goto next;
1388 }
1389 }
1390 }
1391 }
1392
1393 /*
1394 * Finally, call the driver.
1395 */
1396
1397 if ((ifp->if_eflags & IFEF_SENDLIST) != 0) {
1398 *send_tail = m;
1399 send_tail = &m->m_nextpkt;
1400 }
1401 else {
1402 KERNEL_DEBUG(DBG_FNC_DLIL_IFOUT | DBG_FUNC_START, 0,0,0,0,0);
1403 retval = ifp->if_output(ifp, m);
1404 if (retval && dlil_verbose) {
1405 printf("dlil_output: output error on %s%d retval = %d\n",
1406 ifp->if_name, ifp->if_unit, retval);
1407 }
1408 KERNEL_DEBUG(DBG_FNC_DLIL_IFOUT | DBG_FUNC_END, 0,0,0,0,0);
1409 }
1410 KERNEL_DEBUG(DBG_FNC_DLIL_IFOUT | DBG_FUNC_END, 0,0,0,0,0);
1411
1412 next:
1413 m = packetlist;
1414 if (m) {
1415 packetlist = packetlist->m_nextpkt;
1416 m->m_nextpkt = NULL;
1417 }
1418 } while (m);
1419
1420 if (send_head) {
1421 KERNEL_DEBUG(DBG_FNC_DLIL_IFOUT | DBG_FUNC_START, 0,0,0,0,0);
1422 retval = ifp->if_output(ifp, send_head);
1423 if (retval && dlil_verbose) {
1424 printf("dlil_output: output error on %s%d retval = %d\n",
1425 ifp->if_name, ifp->if_unit, retval);
1426 }
1427 KERNEL_DEBUG(DBG_FNC_DLIL_IFOUT | DBG_FUNC_END, 0,0,0,0,0);
1428 }
1429
1430 KERNEL_DEBUG(DBG_FNC_DLIL_OUTPUT | DBG_FUNC_END,0,0,0,0,0);
1431
1432 cleanup:
1433 dlil_read_end();
1434 if (packetlist) /* if any packet left, clean up */
1435 mbuf_freem_list(packetlist);
1436 if (retval == EJUSTRETURN)
1437 retval = 0;
1438 return retval;
1439 }
1440 #endif
1441
1442 /*
1443 * dlil_output
1444 *
1445 * Caller should have a lock on the protocol domain if the protocol
1446 * doesn't support finer grained locking. In most cases, the lock
1447 * will be held from the socket layer and won't be released until
1448 * we return back to the socket layer.
1449 *
1450 * This does mean that we must take a protocol lock before we take
1451 * an interface lock if we're going to take both. This makes sense
1452 * because a protocol is likely to interact with an ifp while it
1453 * is under the protocol lock.
1454 */
1455 __private_extern__ errno_t
1456 dlil_output(
1457 ifnet_t ifp,
1458 protocol_family_t proto_family,
1459 mbuf_t packetlist,
1460 void *route,
1461 const struct sockaddr *dest,
1462 int raw)
1463 {
1464 char *frame_type = NULL;
1465 char *dst_linkaddr = NULL;
1466 int retval = 0;
1467 char frame_type_buffer[MAX_FRAME_TYPE_SIZE * 4];
1468 char dst_linkaddr_buffer[MAX_LINKADDR * 4];
1469 struct ifnet_filter *filter;
1470 struct if_proto *proto = 0;
1471 mbuf_t m;
1472 mbuf_t send_head = NULL;
1473 mbuf_t *send_tail = &send_head;
1474
1475 KERNEL_DEBUG(DBG_FNC_DLIL_OUTPUT | DBG_FUNC_START,0,0,0,0,0);
1476
1477 dlil_read_begin();
1478
1479 frame_type = frame_type_buffer;
1480 dst_linkaddr = dst_linkaddr_buffer;
1481
1482 if (raw == 0) {
1483 proto = find_attached_proto(ifp, proto_family);
1484 if (proto == NULL) {
1485 retval = ENXIO;
1486 goto cleanup;
1487 }
1488 }
1489
1490 preout_again:
1491 if (packetlist == NULL)
1492 goto cleanup;
1493 m = packetlist;
1494 packetlist = packetlist->m_nextpkt;
1495 m->m_nextpkt = NULL;
1496
1497 if (raw == 0) {
1498 proto_media_preout preoutp = proto->proto_kpi == kProtoKPI_v1
1499 ? proto->kpi.v1.pre_output : proto->kpi.v2.pre_output;
1500 retval = 0;
1501 if (preoutp)
1502 retval = preoutp(ifp, proto_family, &m, dest, route, frame_type, dst_linkaddr);
1503
1504 if (retval) {
1505 if (retval == EJUSTRETURN) {
1506 goto preout_again;
1507 }
1508
1509 m_freem(m);
1510 goto cleanup;
1511 }
1512 }
1513
1514 #if CONFIG_MACF_NET
1515 retval = mac_ifnet_check_transmit(ifp, m, proto_family,
1516 dlil_get_socket_type(&m, proto_family, raw));
1517 if (retval) {
1518 m_freem(m);
1519 goto cleanup;
1520 }
1521 #endif
1522
1523 do {
1524 if (raw == 0 && ifp->if_framer) {
1525 retval = ifp->if_framer(ifp, &m, dest, dst_linkaddr, frame_type);
1526 if (retval) {
1527 if (retval != EJUSTRETURN) {
1528 m_freem(m);
1529 }
1530 goto next;
1531 }
1532 }
1533
1534 #if BRIDGE
1535 /* !!!LOCKING!!!
1536 *
1537 * Need to consider how to handle this.
1538 * Also note that return should be a goto cleanup
1539 */
1540 broken-locking
1541 if (do_bridge) {
1542 struct mbuf *m0 = m;
1543 struct ether_header *eh = mtod(m, struct ether_header *);
1544
1545 if (m->m_pkthdr.rcvif)
1546 m->m_pkthdr.rcvif = NULL;
1547 ifp = bridge_dst_lookup(eh);
1548 bdg_forward(&m0, ifp);
1549 if (m0)
1550 m_freem(m0);
1551
1552 return 0 - should be goto cleanup?
1553 }
1554 #endif
1555
1556 /*
1557 * Let interface filters (if any) do their thing ...
1558 */
1559 /* Do not pass VLAN tagged packets to filters PR-3586856 */
1560 if ((m->m_pkthdr.csum_flags & CSUM_VLAN_TAG_VALID) == 0) {
1561 TAILQ_FOREACH(filter, &ifp->if_flt_head, filt_next) {
1562 if ((filter->filt_protocol == 0 || (filter->filt_protocol == proto_family)) &&
1563 filter->filt_output) {
1564 retval = filter->filt_output(filter->filt_cookie, ifp, proto_family, &m);
1565 if (retval) {
1566 if (retval != EJUSTRETURN)
1567 m_freem(m);
1568 goto next;
1569 }
1570 }
1571 }
1572 }
1573
1574 /*
1575 * If the underlying interface is not capable of handling a
1576 * packet whose data portion spans across physically disjoint
1577 * pages, we need to "normalize" the packet so that we pass
1578 * down a chain of mbufs where each mbuf points to a span that
1579 * resides in the system page boundary. If the packet does
1580 * not cross page(s), the following is a no-op.
1581 */
1582 if (!(ifp->if_hwassist & IFNET_MULTIPAGES)) {
1583 if ((m = m_normalize(m)) == NULL)
1584 goto next;
1585 }
1586
1587 /*
1588 * If this is a TSO packet, make sure the interface still advertise TSO capability
1589 */
1590
1591 if ((m->m_pkthdr.csum_flags & CSUM_TSO_IPV4) && !(ifp->if_hwassist & IFNET_TSO_IPV4)) {
1592 retval = EMSGSIZE;
1593 m_freem(m);
1594 goto cleanup;
1595 }
1596
1597 if ((m->m_pkthdr.csum_flags & CSUM_TSO_IPV6) && !(ifp->if_hwassist & IFNET_TSO_IPV6)) {
1598 retval = EMSGSIZE;
1599 m_freem(m);
1600 goto cleanup;
1601 }
1602 /*
1603 * Finally, call the driver.
1604 */
1605
1606 if ((ifp->if_eflags & IFEF_SENDLIST) != 0) {
1607 *send_tail = m;
1608 send_tail = &m->m_nextpkt;
1609 }
1610 else {
1611 KERNEL_DEBUG(DBG_FNC_DLIL_IFOUT | DBG_FUNC_START, 0,0,0,0,0);
1612 retval = ifp->if_output(ifp, m);
1613 if (retval && dlil_verbose) {
1614 printf("dlil_output: output error on %s%d retval = %d\n",
1615 ifp->if_name, ifp->if_unit, retval);
1616 }
1617 KERNEL_DEBUG(DBG_FNC_DLIL_IFOUT | DBG_FUNC_END, 0,0,0,0,0);
1618 }
1619 KERNEL_DEBUG(DBG_FNC_DLIL_IFOUT | DBG_FUNC_END, 0,0,0,0,0);
1620
1621 next:
1622 m = packetlist;
1623 if (m) {
1624 packetlist = packetlist->m_nextpkt;
1625 m->m_nextpkt = NULL;
1626 }
1627 } while (m);
1628
1629 if (send_head) {
1630 KERNEL_DEBUG(DBG_FNC_DLIL_IFOUT | DBG_FUNC_START, 0,0,0,0,0);
1631 retval = ifp->if_output(ifp, send_head);
1632 if (retval && dlil_verbose) {
1633 printf("dlil_output: output error on %s%d retval = %d\n",
1634 ifp->if_name, ifp->if_unit, retval);
1635 }
1636 KERNEL_DEBUG(DBG_FNC_DLIL_IFOUT | DBG_FUNC_END, 0,0,0,0,0);
1637 }
1638
1639 KERNEL_DEBUG(DBG_FNC_DLIL_OUTPUT | DBG_FUNC_END,0,0,0,0,0);
1640
1641 cleanup:
1642 dlil_read_end();
1643 if (packetlist) /* if any packet left, clean up */
1644 mbuf_freem_list(packetlist);
1645 if (retval == EJUSTRETURN)
1646 retval = 0;
1647 return retval;
1648 }
1649
1650 errno_t
1651 ifnet_ioctl(
1652 ifnet_t ifp,
1653 protocol_family_t proto_fam,
1654 u_long ioctl_code,
1655 void *ioctl_arg)
1656 {
1657 struct ifnet_filter *filter;
1658 int retval = EOPNOTSUPP;
1659 int result = 0;
1660 int holding_read = 0;
1661
1662 if (ifp == NULL || ioctl_code == 0)
1663 return EINVAL;
1664
1665 /* Attempt to increment the use count. If it's zero, bail out, the ifp is invalid */
1666 result = ifp_use(ifp, kIfNetUseCount_MustNotBeZero);
1667 if (result != 0)
1668 return EOPNOTSUPP;
1669
1670 dlil_read_begin();
1671 holding_read = 1;
1672
1673 /* Run the interface filters first.
1674 * We want to run all filters before calling the protocol,
1675 * interface family, or interface.
1676 */
1677 TAILQ_FOREACH(filter, &ifp->if_flt_head, filt_next) {
1678 if ((filter->filt_protocol == 0 || (filter->filt_protocol == proto_fam)) &&
1679 filter->filt_ioctl != NULL) {
1680 result = filter->filt_ioctl(filter->filt_cookie, ifp, proto_fam, ioctl_code, ioctl_arg);
1681 /* Only update retval if no one has handled the ioctl */
1682 if (retval == EOPNOTSUPP || result == EJUSTRETURN) {
1683 if (result == ENOTSUP)
1684 result = EOPNOTSUPP;
1685 retval = result;
1686 if (retval && retval != EOPNOTSUPP) {
1687 goto cleanup;
1688 }
1689 }
1690 }
1691 }
1692
1693 /* Allow the protocol to handle the ioctl */
1694 if (proto_fam) {
1695 struct if_proto *proto = find_attached_proto(ifp, proto_fam);
1696
1697 if (proto != 0) {
1698 proto_media_ioctl ioctlp = proto->proto_kpi == kProtoKPI_v1
1699 ? proto->kpi.v1.ioctl : proto->kpi.v2.ioctl;
1700 result = EOPNOTSUPP;
1701 if (ioctlp)
1702 result = ioctlp(ifp, proto_fam, ioctl_code, ioctl_arg);
1703
1704 /* Only update retval if no one has handled the ioctl */
1705 if (retval == EOPNOTSUPP || result == EJUSTRETURN) {
1706 if (result == ENOTSUP)
1707 result = EOPNOTSUPP;
1708 retval = result;
1709 if (retval && retval != EOPNOTSUPP) {
1710 goto cleanup;
1711 }
1712 }
1713 }
1714 }
1715
1716 /*
1717 * Since we have incremented the use count on the ifp, we are guaranteed
1718 * that the ifp will not go away (the function pointers may not be changed).
1719 * We release the dlil read lock so the interface ioctl may trigger a
1720 * protocol attach. This happens with vlan and may occur with other virtual
1721 * interfaces.
1722 */
1723 dlil_read_end();
1724 holding_read = 0;
1725
1726 /* retval is either 0 or EOPNOTSUPP */
1727
1728 /*
1729 * Let the interface handle this ioctl.
1730 * If it returns EOPNOTSUPP, ignore that, we may have
1731 * already handled this in the protocol or family.
1732 */
1733 if (ifp->if_ioctl)
1734 result = (*ifp->if_ioctl)(ifp, ioctl_code, ioctl_arg);
1735
1736 /* Only update retval if no one has handled the ioctl */
1737 if (retval == EOPNOTSUPP || result == EJUSTRETURN) {
1738 if (result == ENOTSUP)
1739 result = EOPNOTSUPP;
1740 retval = result;
1741 if (retval && retval != EOPNOTSUPP) {
1742 goto cleanup;
1743 }
1744 }
1745
1746 cleanup:
1747 if (holding_read)
1748 dlil_read_end();
1749 if (ifp_unuse(ifp))
1750 ifp_use_reached_zero(ifp);
1751
1752 if (retval == EJUSTRETURN)
1753 retval = 0;
1754 return retval;
1755 }
1756
1757 __private_extern__ errno_t
1758 dlil_set_bpf_tap(
1759 ifnet_t ifp,
1760 bpf_tap_mode mode,
1761 bpf_packet_func callback)
1762 {
1763 errno_t error = 0;
1764
1765 dlil_read_begin();
1766 if (ifp->if_set_bpf_tap)
1767 error = ifp->if_set_bpf_tap(ifp, mode, callback);
1768 dlil_read_end();
1769
1770 return error;
1771 }
1772
1773 errno_t
1774 dlil_resolve_multi(
1775 struct ifnet *ifp,
1776 const struct sockaddr *proto_addr,
1777 struct sockaddr *ll_addr,
1778 size_t ll_len)
1779 {
1780 errno_t result = EOPNOTSUPP;
1781 struct if_proto *proto;
1782 const struct sockaddr *verify;
1783 proto_media_resolve_multi resolvep;
1784
1785 dlil_read_begin();
1786
1787 bzero(ll_addr, ll_len);
1788
1789 /* Call the protocol first */
1790 proto = find_attached_proto(ifp, proto_addr->sa_family);
1791 if (proto != NULL) {
1792 resolvep = proto->proto_kpi == kProtoKPI_v1
1793 ? proto->kpi.v1.resolve_multi : proto->kpi.v2.resolve_multi;
1794 if (resolvep != NULL)
1795 result = resolvep(ifp, proto_addr,(struct sockaddr_dl*)ll_addr,
1796 ll_len);
1797 }
1798
1799 /* Let the interface verify the multicast address */
1800 if ((result == EOPNOTSUPP || result == 0) && ifp->if_check_multi) {
1801 if (result == 0)
1802 verify = ll_addr;
1803 else
1804 verify = proto_addr;
1805 result = ifp->if_check_multi(ifp, verify);
1806 }
1807
1808 dlil_read_end();
1809
1810 return result;
1811 }
1812
1813 __private_extern__ errno_t
1814 dlil_send_arp_internal(
1815 ifnet_t ifp,
1816 u_short arpop,
1817 const struct sockaddr_dl* sender_hw,
1818 const struct sockaddr* sender_proto,
1819 const struct sockaddr_dl* target_hw,
1820 const struct sockaddr* target_proto)
1821 {
1822 struct if_proto *proto;
1823 errno_t result = 0;
1824
1825 dlil_read_begin();
1826
1827 proto = find_attached_proto(ifp, target_proto->sa_family);
1828 if (proto == NULL) {
1829 result = ENOTSUP;
1830 }
1831 else {
1832 proto_media_send_arp arpp;
1833 arpp = proto->proto_kpi == kProtoKPI_v1
1834 ? proto->kpi.v1.send_arp : proto->kpi.v2.send_arp;
1835 if (arpp == NULL)
1836 result = ENOTSUP;
1837 else
1838 result = arpp(ifp, arpop, sender_hw, sender_proto, target_hw,
1839 target_proto);
1840 }
1841
1842 dlil_read_end();
1843
1844 return result;
1845 }
1846
1847 static __inline__ int
1848 _is_announcement(const struct sockaddr_in * sender_sin,
1849 const struct sockaddr_in * target_sin)
1850 {
1851 if (sender_sin == NULL) {
1852 return FALSE;
1853 }
1854 return (sender_sin->sin_addr.s_addr == target_sin->sin_addr.s_addr);
1855 }
1856
1857 __private_extern__ errno_t
1858 dlil_send_arp(
1859 ifnet_t ifp,
1860 u_short arpop,
1861 const struct sockaddr_dl* sender_hw,
1862 const struct sockaddr* sender_proto,
1863 const struct sockaddr_dl* target_hw,
1864 const struct sockaddr* target_proto)
1865 {
1866 errno_t result = 0;
1867 const struct sockaddr_in * sender_sin;
1868 const struct sockaddr_in * target_sin;
1869
1870 if (target_proto == NULL || (sender_proto &&
1871 sender_proto->sa_family != target_proto->sa_family))
1872 return EINVAL;
1873
1874 /*
1875 * If this is an ARP request and the target IP is IPv4LL,
1876 * send the request on all interfaces. The exception is
1877 * an announcement, which must only appear on the specific
1878 * interface.
1879 */
1880 sender_sin = (const struct sockaddr_in *)sender_proto;
1881 target_sin = (const struct sockaddr_in *)target_proto;
1882 if (target_proto->sa_family == AF_INET
1883 && IN_LINKLOCAL(ntohl(target_sin->sin_addr.s_addr))
1884 && ipv4_ll_arp_aware != 0
1885 && arpop == ARPOP_REQUEST
1886 && !_is_announcement(target_sin, sender_sin)) {
1887 ifnet_t *ifp_list;
1888 u_int32_t count;
1889 u_int32_t ifp_on;
1890
1891 result = ENOTSUP;
1892
1893 if (ifnet_list_get(IFNET_FAMILY_ANY, &ifp_list, &count) == 0) {
1894 for (ifp_on = 0; ifp_on < count; ifp_on++) {
1895 errno_t new_result;
1896 ifaddr_t source_hw = NULL;
1897 ifaddr_t source_ip = NULL;
1898 struct sockaddr_in source_ip_copy;
1899
1900 /*
1901 * Only arp on interfaces marked for IPv4LL ARPing. This may
1902 * mean that we don't ARP on the interface the subnet route
1903 * points to.
1904 */
1905 if ((ifp_list[ifp_on]->if_eflags & IFEF_ARPLL) == 0) {
1906 continue;
1907 }
1908
1909 /* Find the source IP address */
1910 ifnet_lock_shared(ifp_list[ifp_on]);
1911 source_hw = TAILQ_FIRST(&ifp_list[ifp_on]->if_addrhead);
1912 TAILQ_FOREACH(source_ip, &ifp_list[ifp_on]->if_addrhead,
1913 ifa_link) {
1914 if (source_ip->ifa_addr &&
1915 source_ip->ifa_addr->sa_family == AF_INET) {
1916 break;
1917 }
1918 }
1919
1920 /* No IP Source, don't arp */
1921 if (source_ip == NULL) {
1922 ifnet_lock_done(ifp_list[ifp_on]);
1923 continue;
1924 }
1925
1926 /* Copy the source IP address */
1927 source_ip_copy = *(struct sockaddr_in*)source_ip->ifa_addr;
1928 ifaref(source_hw);
1929 ifnet_lock_done(ifp_list[ifp_on]);
1930
1931 /* Send the ARP */
1932 new_result = dlil_send_arp_internal(ifp_list[ifp_on], arpop,
1933 (struct sockaddr_dl*)source_hw->ifa_addr,
1934 (struct sockaddr*)&source_ip_copy, NULL,
1935 target_proto);
1936
1937 ifafree(source_hw);
1938 if (result == ENOTSUP) {
1939 result = new_result;
1940 }
1941 }
1942 }
1943
1944 ifnet_list_free(ifp_list);
1945 }
1946 else {
1947 result = dlil_send_arp_internal(ifp, arpop, sender_hw, sender_proto,
1948 target_hw, target_proto);
1949 }
1950
1951 return result;
1952 }
1953
1954 __private_extern__ int
1955 ifp_use(
1956 struct ifnet *ifp,
1957 int handle_zero)
1958 {
1959 int old_value;
1960 int retval = 0;
1961
1962 do {
1963 old_value = ifp->if_usecnt;
1964 if (old_value == 0 && handle_zero == kIfNetUseCount_MustNotBeZero) {
1965 retval = ENXIO; // ifp is invalid
1966 break;
1967 }
1968 } while (!OSCompareAndSwap((UInt32)old_value, (UInt32)old_value + 1, (UInt32*)&ifp->if_usecnt));
1969
1970 return retval;
1971 }
1972
1973 /* ifp_unuse is broken into two pieces.
1974 *
1975 * ifp_use and ifp_unuse must be called between when the caller calls
1976 * dlil_write_begin and dlil_write_end. ifp_unuse needs to perform some
1977 * operations after dlil_write_end has been called. For this reason,
1978 * anyone calling ifp_unuse must call ifp_use_reached_zero if ifp_unuse
1979 * returns a non-zero value. The caller must call ifp_use_reached_zero
1980 * after the caller has called dlil_write_end.
1981 */
1982 __private_extern__ void
1983 ifp_use_reached_zero(
1984 struct ifnet *ifp)
1985 {
1986 ifnet_detached_func free_func;
1987
1988 dlil_read_begin();
1989
1990 if (ifp->if_usecnt != 0)
1991 panic("ifp_use_reached_zero: ifp->if_usecnt != 0");
1992
1993 ifnet_head_lock_exclusive();
1994 ifnet_lock_exclusive(ifp);
1995
1996 /* Remove ourselves from the list */
1997 TAILQ_REMOVE(&ifnet_head, ifp, if_link);
1998 ifnet_addrs[ifp->if_index - 1] = NULL;
1999
2000 /* ifp should be removed from the interface list */
2001 while (ifp->if_multiaddrs.lh_first) {
2002 struct ifmultiaddr *ifma = ifp->if_multiaddrs.lh_first;
2003
2004 /*
2005 * When the interface is gone, we will no longer
2006 * be listening on these multicasts. Various bits
2007 * of the stack may be referencing these multicasts,
2008 * release only our reference.
2009 */
2010 LIST_REMOVE(ifma, ifma_link);
2011 ifma->ifma_ifp = NULL;
2012 ifma_release(ifma);
2013 }
2014
2015 ifp->if_eflags &= ~IFEF_DETACHING; // clear the detaching flag
2016 ifnet_lock_done(ifp);
2017 ifnet_head_done();
2018
2019 free_func = ifp->if_free;
2020 dlil_read_end();
2021 dlil_post_msg(ifp, KEV_DL_SUBCLASS, KEV_DL_IF_DETACHED, NULL, 0);
2022
2023 if (free_func)
2024 free_func(ifp);
2025 }
2026
2027 __private_extern__ int
2028 ifp_unuse(
2029 struct ifnet *ifp)
2030 {
2031 int oldval;
2032 oldval = OSDecrementAtomic(&ifp->if_usecnt);
2033 if (oldval == 0)
2034 panic("ifp_unuse: ifp(%s%d)->if_usecnt was zero\n", ifp->if_name, ifp->if_unit);
2035
2036 if (oldval > 1)
2037 return 0;
2038
2039 if ((ifp->if_eflags & IFEF_DETACHING) == 0)
2040 panic("ifp_unuse: use count reached zero but detching flag is not set!");
2041
2042 return 1; /* caller must call ifp_use_reached_zero */
2043 }
2044
2045 extern lck_mtx_t *domain_proto_mtx;
2046
2047 static errno_t
2048 dlil_attach_protocol_internal(
2049 struct if_proto *proto,
2050 const struct ifnet_demux_desc *demux_list,
2051 u_int32_t demux_count)
2052 {
2053 struct kev_dl_proto_data ev_pr_data;
2054 struct ifnet *ifp = proto->ifp;
2055 int retval = 0;
2056 u_int32_t hash_value = proto_hash_value(proto->protocol_family);
2057
2058 /* setup some of the common values */
2059 {
2060 struct domain *dp;
2061 lck_mtx_lock(domain_proto_mtx);
2062 dp = domains;
2063 while (dp && (protocol_family_t)dp->dom_family != proto->protocol_family)
2064 dp = dp->dom_next;
2065 proto->dl_domain = dp;
2066 lck_mtx_unlock(domain_proto_mtx);
2067 }
2068
2069 /*
2070 * Take the write lock to protect readers and exclude other writers.
2071 */
2072 if ((retval = dlil_write_begin()) != 0) {
2073 printf("dlil_attach_protocol_internal - dlil_write_begin returned %d\n", retval);
2074 return retval;
2075 }
2076
2077 /* Check that the interface isn't currently detaching */
2078 ifnet_lock_shared(ifp);
2079 if ((ifp->if_eflags & IFEF_DETACHING) != 0) {
2080 ifnet_lock_done(ifp);
2081 dlil_write_end();
2082 return ENXIO;
2083 }
2084 ifnet_lock_done(ifp);
2085
2086 if (find_attached_proto(ifp, proto->protocol_family) != NULL) {
2087 dlil_write_end();
2088 return EEXIST;
2089 }
2090
2091 /*
2092 * Call family module add_proto routine so it can refine the
2093 * demux descriptors as it wishes.
2094 */
2095 retval = ifp->if_add_proto(ifp, proto->protocol_family, demux_list, demux_count);
2096 if (retval) {
2097 dlil_write_end();
2098 return retval;
2099 }
2100
2101 /*
2102 * We can't fail from this point on.
2103 * Increment the number of uses (protocol attachments + interface attached).
2104 */
2105 ifp_use(ifp, kIfNetUseCount_MustNotBeZero);
2106
2107 /*
2108 * Insert the protocol in the hash
2109 */
2110 {
2111 struct if_proto* prev_proto = SLIST_FIRST(&ifp->if_proto_hash[hash_value]);
2112 while (prev_proto && SLIST_NEXT(prev_proto, next_hash) != NULL)
2113 prev_proto = SLIST_NEXT(prev_proto, next_hash);
2114 if (prev_proto)
2115 SLIST_INSERT_AFTER(prev_proto, proto, next_hash);
2116 else
2117 SLIST_INSERT_HEAD(&ifp->if_proto_hash[hash_value], proto, next_hash);
2118 }
2119
2120 /*
2121 * Add to if_proto list for this interface
2122 */
2123 if_proto_ref(proto);
2124 dlil_write_end();
2125
2126 /* the reserved field carries the number of protocol still attached (subject to change) */
2127 ev_pr_data.proto_family = proto->protocol_family;
2128 ev_pr_data.proto_remaining_count = dlil_ifp_proto_count(ifp);
2129 dlil_post_msg(ifp, KEV_DL_SUBCLASS, KEV_DL_PROTO_ATTACHED,
2130 (struct net_event_data *)&ev_pr_data,
2131 sizeof(struct kev_dl_proto_data));
2132 #if 0
2133 DLIL_PRINTF("dlil. Attached protocol %d to %s%d - %d\n", proto->protocol_family,
2134 ifp->if_name, ifp->if_unit, retval);
2135 #endif
2136 return retval;
2137 }
2138
2139 errno_t
2140 ifnet_attach_protocol(ifnet_t ifp, protocol_family_t protocol,
2141 const struct ifnet_attach_proto_param *proto_details)
2142 {
2143 int retval = 0;
2144 struct if_proto *ifproto = NULL;
2145
2146 if (ifp == NULL || protocol == 0 || proto_details == NULL)
2147 return EINVAL;
2148
2149 ifproto = _MALLOC(sizeof(struct if_proto), M_IFADDR, M_WAITOK);
2150 if (ifproto == 0) {
2151 DLIL_PRINTF("ERROR - dlil failed if_proto allocation\n");
2152 retval = ENOMEM;
2153 goto end;
2154 }
2155 bzero(ifproto, sizeof(*ifproto));
2156
2157 ifproto->ifp = ifp;
2158 ifproto->protocol_family = protocol;
2159 ifproto->proto_kpi = kProtoKPI_v1;
2160 ifproto->kpi.v1.input = proto_details->input;
2161 ifproto->kpi.v1.pre_output = proto_details->pre_output;
2162 ifproto->kpi.v1.event = proto_details->event;
2163 ifproto->kpi.v1.ioctl = proto_details->ioctl;
2164 ifproto->kpi.v1.detached = proto_details->detached;
2165 ifproto->kpi.v1.resolve_multi = proto_details->resolve;
2166 ifproto->kpi.v1.send_arp = proto_details->send_arp;
2167
2168 retval = dlil_attach_protocol_internal(ifproto,
2169 proto_details->demux_list, proto_details->demux_count);
2170
2171 end:
2172 if (retval && ifproto)
2173 FREE(ifproto, M_IFADDR);
2174 return retval;
2175 }
2176
2177 errno_t
2178 ifnet_attach_protocol_v2(ifnet_t ifp, protocol_family_t protocol,
2179 const struct ifnet_attach_proto_param_v2 *proto_details)
2180 {
2181 int retval = 0;
2182 struct if_proto *ifproto = NULL;
2183
2184 if (ifp == NULL || protocol == 0 || proto_details == NULL)
2185 return EINVAL;
2186
2187 ifproto = _MALLOC(sizeof(struct if_proto), M_IFADDR, M_WAITOK);
2188 if (ifproto == 0) {
2189 DLIL_PRINTF("ERROR - dlil failed if_proto allocation\n");
2190 retval = ENOMEM;
2191 goto end;
2192 }
2193 bzero(ifproto, sizeof(*ifproto));
2194
2195 ifproto->ifp = ifp;
2196 ifproto->protocol_family = protocol;
2197 ifproto->proto_kpi = kProtoKPI_v2;
2198 ifproto->kpi.v2.input = proto_details->input;
2199 ifproto->kpi.v2.pre_output = proto_details->pre_output;
2200 ifproto->kpi.v2.event = proto_details->event;
2201 ifproto->kpi.v2.ioctl = proto_details->ioctl;
2202 ifproto->kpi.v2.detached = proto_details->detached;
2203 ifproto->kpi.v2.resolve_multi = proto_details->resolve;
2204 ifproto->kpi.v2.send_arp = proto_details->send_arp;
2205
2206 retval = dlil_attach_protocol_internal(ifproto,
2207 proto_details->demux_list, proto_details->demux_count);
2208
2209 end:
2210 if (retval && ifproto)
2211 FREE(ifproto, M_IFADDR);
2212 return retval;
2213 }
2214
2215 extern void if_rtproto_del(struct ifnet *ifp, int protocol);
2216
2217 static int
2218 dlil_detach_protocol_internal(
2219 struct if_proto *proto)
2220 {
2221 struct ifnet *ifp = proto->ifp;
2222 u_int32_t proto_family = proto->protocol_family;
2223 struct kev_dl_proto_data ev_pr_data;
2224
2225 if (proto->proto_kpi == kProtoKPI_v1) {
2226 if (proto->kpi.v1.detached)
2227 proto->kpi.v1.detached(ifp, proto->protocol_family);
2228 }
2229 if (proto->proto_kpi == kProtoKPI_v2) {
2230 if (proto->kpi.v2.detached)
2231 proto->kpi.v2.detached(ifp, proto->protocol_family);
2232 }
2233 if_proto_free(proto);
2234
2235 /*
2236 * Cleanup routes that may still be in the routing table for that interface/protocol pair.
2237 */
2238
2239 if_rtproto_del(ifp, proto_family);
2240
2241 /* the reserved field carries the number of protocol still attached (subject to change) */
2242 ev_pr_data.proto_family = proto_family;
2243 ev_pr_data.proto_remaining_count = dlil_ifp_proto_count(ifp);
2244 dlil_post_msg(ifp, KEV_DL_SUBCLASS, KEV_DL_PROTO_DETACHED,
2245 (struct net_event_data *)&ev_pr_data,
2246 sizeof(struct kev_dl_proto_data));
2247 return 0;
2248 }
2249
2250 errno_t
2251 ifnet_detach_protocol(ifnet_t ifp, protocol_family_t proto_family)
2252 {
2253 struct if_proto *proto = NULL;
2254 int retval = 0;
2255 int use_reached_zero = 0;
2256
2257 if (ifp == NULL || proto_family == 0) return EINVAL;
2258
2259 if ((retval = dlil_write_begin()) != 0) {
2260 if (retval == EDEADLK) {
2261 retval = 0;
2262 dlil_read_begin();
2263 proto = find_attached_proto(ifp, proto_family);
2264 if (proto == 0) {
2265 retval = ENXIO;
2266 }
2267 else {
2268 proto->detaching = 1;
2269 dlil_detach_waiting = 1;
2270 wakeup(&dlil_detach_waiting);
2271 }
2272 dlil_read_end();
2273 }
2274 goto end;
2275 }
2276
2277 proto = find_attached_proto(ifp, proto_family);
2278
2279 if (proto == NULL) {
2280 retval = ENXIO;
2281 dlil_write_end();
2282 goto end;
2283 }
2284
2285 /*
2286 * Call family module del_proto
2287 */
2288
2289 if (ifp->if_del_proto)
2290 ifp->if_del_proto(ifp, proto->protocol_family);
2291
2292 SLIST_REMOVE(&ifp->if_proto_hash[proto_hash_value(proto_family)], proto, if_proto, next_hash);
2293
2294 /*
2295 * We can do the rest of the work outside of the write lock.
2296 */
2297 use_reached_zero = ifp_unuse(ifp);
2298 dlil_write_end();
2299
2300 dlil_detach_protocol_internal(proto);
2301
2302 /*
2303 * Only handle the case where the interface will go away after
2304 * we've sent the message. This way post message can send the
2305 * message to the interface safely.
2306 */
2307
2308 if (use_reached_zero)
2309 ifp_use_reached_zero(ifp);
2310
2311 end:
2312 return retval;
2313 }
2314
2315 /*
2316 * dlil_delayed_detach_thread is responsible for detaching
2317 * protocols, protocol filters, and interface filters after
2318 * an attempt was made to detach one of those items while
2319 * it was not safe to do so (i.e. called dlil_read_begin).
2320 *
2321 * This function will take the dlil write lock and walk
2322 * through each of the interfaces looking for items with
2323 * the detaching flag set. When an item is found, it is
2324 * detached from the interface and placed on a local list.
2325 * After all of the items have been collected, we drop the
2326 * write lock and performed the post detach. This is done
2327 * so we only have to take the write lock once.
2328 *
2329 * When detaching a protocol filter, if we find that we
2330 * have detached the very last protocol and we need to call
2331 * ifp_use_reached_zero, we have to break out of our work
2332 * to drop the write lock so we can call ifp_use_reached_zero.
2333 */
2334
2335 static void
2336 dlil_delayed_detach_thread(__unused void* foo, __unused wait_result_t wait)
2337 {
2338 thread_t self = current_thread();
2339 int asserted = 0;
2340
2341 ml_thread_policy(self, MACHINE_GROUP,
2342 (MACHINE_NETWORK_GROUP|MACHINE_NETWORK_NETISR));
2343
2344
2345 while (1) {
2346 if (dlil_detach_waiting != 0 && dlil_write_begin() == 0) {
2347 struct ifnet *ifp;
2348 struct proto_hash_entry detached_protos;
2349 struct ifnet_filter_head detached_filters;
2350 struct if_proto *proto;
2351 struct if_proto *next_proto;
2352 struct ifnet_filter *filt;
2353 struct ifnet_filter *next_filt;
2354 int reached_zero;
2355
2356 reached_zero = 0;
2357
2358 /* Clear the detach waiting flag */
2359 dlil_detach_waiting = 0;
2360 TAILQ_INIT(&detached_filters);
2361 SLIST_INIT(&detached_protos);
2362
2363 ifnet_head_lock_shared();
2364 TAILQ_FOREACH(ifp, &ifnet_head, if_link) {
2365 int i;
2366
2367 // Look for protocols and protocol filters
2368 for (i = 0; i < PROTO_HASH_SLOTS && !reached_zero; i++) {
2369 struct if_proto **prev_nextptr = &SLIST_FIRST(&ifp->if_proto_hash[i]);
2370 for (proto = *prev_nextptr; proto; proto = *prev_nextptr) {
2371
2372 // Detach this protocol
2373 if (proto->detaching) {
2374 if (ifp->if_del_proto)
2375 ifp->if_del_proto(ifp, proto->protocol_family);
2376 *prev_nextptr = SLIST_NEXT(proto, next_hash);
2377 SLIST_INSERT_HEAD(&detached_protos, proto, next_hash);
2378 reached_zero = ifp_unuse(ifp);
2379 if (reached_zero) {
2380 break;
2381 }
2382 }
2383 else {
2384 // Update prev_nextptr to point to our next ptr
2385 prev_nextptr = &SLIST_NEXT(proto, next_hash);
2386 }
2387 }
2388 }
2389
2390 // look for interface filters that need to be detached
2391 for (filt = TAILQ_FIRST(&ifp->if_flt_head); filt; filt = next_filt) {
2392 next_filt = TAILQ_NEXT(filt, filt_next);
2393 if (filt->filt_detaching != 0) {
2394 // take this interface filter off the interface filter list
2395 TAILQ_REMOVE(&ifp->if_flt_head, filt, filt_next);
2396
2397 // put this interface filter on the detached filters list
2398 TAILQ_INSERT_TAIL(&detached_filters, filt, filt_next);
2399 }
2400 }
2401
2402 if (ifp->if_delayed_detach) {
2403 ifp->if_delayed_detach = 0;
2404 reached_zero = ifp_unuse(ifp);
2405 }
2406
2407 if (reached_zero)
2408 break;
2409 }
2410 ifnet_head_done();
2411 dlil_write_end();
2412
2413 for (filt = TAILQ_FIRST(&detached_filters); filt; filt = next_filt) {
2414 next_filt = TAILQ_NEXT(filt, filt_next);
2415 /*
2416 * dlil_detach_filter_internal won't remove an item from
2417 * the list if it is already detached (second parameter).
2418 * The item will be freed though.
2419 */
2420 dlil_detach_filter_internal(filt, 1);
2421 }
2422
2423 for (proto = SLIST_FIRST(&detached_protos); proto; proto = next_proto) {
2424 next_proto = SLIST_NEXT(proto, next_hash);
2425 dlil_detach_protocol_internal(proto);
2426 }
2427
2428 if (reached_zero) {
2429 ifp_use_reached_zero(ifp);
2430 dlil_detach_waiting = 1; // we may have missed something
2431 }
2432 }
2433
2434 if (!asserted && dlil_detach_waiting == 0) {
2435 asserted = 1;
2436 assert_wait(&dlil_detach_waiting, THREAD_UNINT);
2437 }
2438
2439 if (dlil_detach_waiting == 0) {
2440 asserted = 0;
2441 thread_block(dlil_delayed_detach_thread);
2442 }
2443 }
2444 }
2445
2446 static void
2447 dlil_call_delayed_detach_thread(void) {
2448 dlil_delayed_detach_thread(NULL, THREAD_RESTART);
2449 }
2450
2451 extern int if_next_index(void);
2452
2453 errno_t
2454 ifnet_attach(
2455 ifnet_t ifp,
2456 const struct sockaddr_dl *ll_addr)
2457 {
2458 u_int32_t interface_family;
2459 struct ifnet *tmp_if;
2460 struct proto_hash_entry *new_proto_list = NULL;
2461 int locked = 0;
2462
2463 if (ifp == NULL) return EINVAL;
2464 if (ll_addr && ifp->if_addrlen == 0) {
2465 ifp->if_addrlen = ll_addr->sdl_alen;
2466 }
2467 else if (ll_addr && ll_addr->sdl_alen != ifp->if_addrlen) {
2468 return EINVAL;
2469 }
2470
2471 interface_family = ifp->if_family;
2472
2473 ifnet_head_lock_shared();
2474
2475 /* Verify we aren't already on the list */
2476 TAILQ_FOREACH(tmp_if, &ifnet_head, if_link) {
2477 if (tmp_if == ifp) {
2478 ifnet_head_done();
2479 return EEXIST;
2480 }
2481 }
2482
2483 ifnet_head_done();
2484
2485 if ((ifp->if_eflags & IFEF_REUSE) == 0 || ifp->if_lock == 0)
2486 #if IFNET_RW_LOCK
2487 ifp->if_lock = lck_rw_alloc_init(ifnet_lock_group, ifnet_lock_attr);
2488 #else
2489 ifp->if_lock = lck_mtx_alloc_init(ifnet_lock_group, ifnet_lock_attr);
2490 #endif
2491
2492 if (ifp->if_lock == 0) {
2493 return ENOMEM;
2494 }
2495
2496 if (!(ifp->if_eflags & IFEF_REUSE) || ifp->if_fwd_route_lock == NULL) {
2497 if (ifp->if_fwd_route_lock == NULL)
2498 ifp->if_fwd_route_lock = lck_mtx_alloc_init(
2499 ifnet_lock_group, ifnet_lock_attr);
2500
2501 if (ifp->if_fwd_route_lock == NULL) {
2502 #if IFNET_RW_LOCK
2503 lck_rw_free(ifp->if_lock, ifnet_lock_group);
2504 #else
2505 lck_mtx_free(ifp->if_lock, ifnet_lock_group);
2506 #endif
2507 ifp->if_lock = NULL;
2508 return (ENOMEM);
2509 }
2510 }
2511
2512 /*
2513 * Allow interfaces without protocol families to attach
2514 * only if they have the necessary fields filled out.
2515 */
2516
2517 if (ifp->if_add_proto == 0 || ifp->if_del_proto == 0) {
2518 DLIL_PRINTF("dlil Attempt to attach interface without family module - %d\n",
2519 interface_family);
2520 return ENODEV;
2521 }
2522
2523 if ((ifp->if_eflags & IFEF_REUSE) == 0 || ifp->if_proto_hash == NULL) {
2524 MALLOC(new_proto_list, struct proto_hash_entry*, sizeof(struct proto_hash_entry) * PROTO_HASH_SLOTS,
2525 M_NKE, M_WAITOK);
2526
2527 if (new_proto_list == 0) {
2528 return ENOBUFS;
2529 }
2530 }
2531
2532 dlil_write_begin();
2533 locked = 1;
2534
2535 TAILQ_INIT(&ifp->if_flt_head);
2536
2537
2538 if (new_proto_list) {
2539 bzero(new_proto_list, (PROTO_HASH_SLOTS * sizeof(struct proto_hash_entry)));
2540 ifp->if_proto_hash = new_proto_list;
2541 new_proto_list = NULL;
2542 }
2543
2544 /* old_if_attach */
2545 {
2546 char workbuf[64];
2547 int namelen, masklen, socksize, ifasize;
2548 struct ifaddr *ifa = NULL;
2549
2550 if (ifp->if_snd.ifq_maxlen == 0)
2551 ifp->if_snd.ifq_maxlen = ifqmaxlen;
2552 TAILQ_INIT(&ifp->if_prefixhead);
2553 LIST_INIT(&ifp->if_multiaddrs);
2554 ifnet_touch_lastchange(ifp);
2555
2556 /* usecount to track attachment to the ifnet list */
2557 ifp_use(ifp, kIfNetUseCount_MayBeZero);
2558
2559 /* Lock the list of interfaces */
2560 ifnet_head_lock_exclusive();
2561 ifnet_lock_exclusive(ifp);
2562
2563 if ((ifp->if_eflags & IFEF_REUSE) == 0 || ifp->if_index == 0) {
2564 int idx = if_next_index();
2565
2566 if (idx == -1) {
2567 ifnet_lock_done(ifp);
2568 ifnet_head_done();
2569 ifp_unuse(ifp);
2570 dlil_write_end();
2571
2572 return ENOBUFS;
2573 }
2574 ifp->if_index = idx;
2575 } else {
2576 ifa = TAILQ_FIRST(&ifp->if_addrhead);
2577 }
2578 namelen = snprintf(workbuf, sizeof(workbuf), "%s%d", ifp->if_name, ifp->if_unit);
2579 #define _offsetof(t, m) ((uintptr_t)((caddr_t)&((t *)0)->m))
2580 masklen = _offsetof(struct sockaddr_dl, sdl_data[0]) + namelen;
2581 socksize = masklen + ifp->if_addrlen;
2582 #define ROUNDUP(a) (1 + (((a) - 1) | (sizeof(long) - 1)))
2583 if ((u_int32_t)socksize < sizeof(struct sockaddr_dl))
2584 socksize = sizeof(struct sockaddr_dl);
2585 socksize = ROUNDUP(socksize);
2586 ifasize = sizeof(struct ifaddr) + 2 * socksize;
2587
2588 /*
2589 * Allocate a new ifa if we don't have one
2590 * or the old one is too small.
2591 */
2592 if (ifa == NULL || socksize > ifa->ifa_addr->sa_len) {
2593 if (ifa)
2594 if_detach_ifa(ifp, ifa);
2595 ifa = (struct ifaddr*)_MALLOC(ifasize, M_IFADDR, M_WAITOK);
2596 }
2597
2598 if (ifa) {
2599 struct sockaddr_dl *sdl = (struct sockaddr_dl *)(ifa + 1);
2600 ifnet_addrs[ifp->if_index - 1] = ifa;
2601 bzero(ifa, ifasize);
2602 ifa->ifa_debug |= IFD_ALLOC;
2603 sdl->sdl_len = socksize;
2604 sdl->sdl_family = AF_LINK;
2605 bcopy(workbuf, sdl->sdl_data, namelen);
2606 sdl->sdl_nlen = namelen;
2607 sdl->sdl_index = ifp->if_index;
2608 sdl->sdl_type = ifp->if_type;
2609 if (ll_addr) {
2610 sdl->sdl_alen = ll_addr->sdl_alen;
2611 if (ll_addr->sdl_alen != ifp->if_addrlen)
2612 panic("ifnet_attach - ll_addr->sdl_alen != ifp->if_addrlen");
2613 bcopy(CONST_LLADDR(ll_addr), LLADDR(sdl), sdl->sdl_alen);
2614 }
2615 ifa->ifa_ifp = ifp;
2616 ifa->ifa_rtrequest = link_rtrequest;
2617 ifa->ifa_addr = (struct sockaddr*)sdl;
2618 sdl = (struct sockaddr_dl*)(socksize + (caddr_t)sdl);
2619 ifa->ifa_netmask = (struct sockaddr*)sdl;
2620 sdl->sdl_len = masklen;
2621 while (namelen != 0)
2622 sdl->sdl_data[--namelen] = 0xff;
2623 }
2624
2625 TAILQ_INIT(&ifp->if_addrhead);
2626 ifa = ifnet_addrs[ifp->if_index - 1];
2627
2628 if (ifa) {
2629 /*
2630 * We don't use if_attach_ifa because we want
2631 * this address to be first on the list.
2632 */
2633 ifaref(ifa);
2634 ifa->ifa_debug |= IFD_ATTACHED;
2635 TAILQ_INSERT_HEAD(&ifp->if_addrhead, ifa, ifa_link);
2636 }
2637 #if CONFIG_MACF_NET
2638 mac_ifnet_label_associate(ifp);
2639 #endif
2640
2641 TAILQ_INSERT_TAIL(&ifnet_head, ifp, if_link);
2642 ifindex2ifnet[ifp->if_index] = ifp;
2643 }
2644
2645 /*
2646 * A specific dlil input thread is created per Ethernet/PDP interface.
2647 * pseudo interfaces or other types of interfaces use the main ("loopback") thread.
2648 * If the sysctl "net.link.generic.system.multi_threaded_input" is set to zero, all packets will
2649 * be handled by the main loopback thread, reverting to 10.4.x behaviour.
2650 *
2651 */
2652
2653 if (ifp->if_type == IFT_ETHER || ifp->if_type == IFT_PDP) {
2654 int err;
2655
2656 if (dlil_multithreaded_input > 0) {
2657 ifp->if_input_thread = _MALLOC(sizeof(struct dlil_threading_info), M_NKE, M_WAITOK);
2658 if (ifp->if_input_thread == NULL)
2659 panic("ifnet_attach ifp=%p couldn't alloc threading\n", ifp);
2660 if ((err = dlil_create_input_thread(ifp, ifp->if_input_thread)) != 0)
2661 panic("ifnet_attach ifp=%p couldn't get a thread. err=%d\n", ifp, err);
2662 #ifdef DLIL_DEBUG
2663 printf("ifnet_attach: dlil thread for ifp=%p if_index=%d\n", ifp, ifp->if_index);
2664 #endif
2665 }
2666 }
2667 ifnet_lock_done(ifp);
2668 ifnet_head_done();
2669 #if PF
2670 /*
2671 * Attach packet filter to this interface, if enaled.
2672 */
2673 pf_ifnet_hook(ifp, 1);
2674 #endif /* PF */
2675 dlil_write_end();
2676
2677 dlil_post_msg(ifp, KEV_DL_SUBCLASS, KEV_DL_IF_ATTACHED, NULL, 0);
2678
2679 return 0;
2680 }
2681
2682 errno_t
2683 ifnet_detach(
2684 ifnet_t ifp)
2685 {
2686 struct ifnet_filter *filter;
2687 struct ifnet_filter *filter_next;
2688 int zeroed = 0;
2689 int retval = 0;
2690 struct ifnet_filter_head fhead;
2691 struct dlil_threading_info *inputthread;
2692
2693 if (ifp == NULL) return EINVAL;
2694
2695 ifnet_lock_exclusive(ifp);
2696
2697 if ((ifp->if_eflags & IFEF_DETACHING) != 0) {
2698 /* Interface has already been detached */
2699 ifnet_lock_done(ifp);
2700 return ENXIO;
2701 }
2702
2703 /*
2704 * Indicate this interface is being detached.
2705 *
2706 * This should prevent protocols from attaching
2707 * from this point on. Interface will remain on
2708 * the list until all of the protocols are detached.
2709 */
2710 ifp->if_eflags |= IFEF_DETACHING;
2711 ifnet_lock_done(ifp);
2712
2713 dlil_post_msg(ifp, KEV_DL_SUBCLASS, KEV_DL_IF_DETACHING, NULL, 0);
2714
2715 /* Let BPF know we're detaching */
2716 bpfdetach(ifp);
2717
2718 if ((retval = dlil_write_begin()) != 0) {
2719 if (retval == EDEADLK) {
2720 retval = 0;
2721
2722 /* We need to perform a delayed detach */
2723 ifp->if_delayed_detach = 1;
2724 dlil_detach_waiting = 1;
2725 wakeup(&dlil_detach_waiting);
2726 }
2727 return retval;
2728 }
2729
2730 #if PF
2731 /*
2732 * Detach this interface from packet filter, if enabled.
2733 */
2734 pf_ifnet_hook(ifp, 0);
2735 #endif /* PF */
2736
2737 /* Steal the list of interface filters */
2738 fhead = ifp->if_flt_head;
2739 TAILQ_INIT(&ifp->if_flt_head);
2740
2741 /* unuse the interface */
2742 zeroed = ifp_unuse(ifp);
2743
2744 /*
2745 * If thread affinity was set for the workloop thread, we will need
2746 * to tear down the affinity and release the extra reference count
2747 * taken at attach time;
2748 */
2749 if ((inputthread = ifp->if_input_thread) != NULL) {
2750 if (inputthread->net_affinity) {
2751 struct thread *tp;
2752
2753 if (inputthread == dlil_lo_thread_ptr)
2754 panic("Thread affinity should not be enabled "
2755 "on the loopback dlil input thread\n");
2756
2757 lck_mtx_lock(inputthread->input_lck);
2758 tp = inputthread->workloop_thread;
2759 inputthread->workloop_thread = NULL;
2760 inputthread->tag = 0;
2761 inputthread->net_affinity = FALSE;
2762 lck_mtx_unlock(inputthread->input_lck);
2763
2764 /* Tear down workloop thread affinity */
2765 if (tp != NULL) {
2766 (void) dlil_affinity_set(tp,
2767 THREAD_AFFINITY_TAG_NULL);
2768 thread_deallocate(tp);
2769 }
2770
2771 /* Tear down dlil input thread affinity */
2772 tp = inputthread->input_thread;
2773 (void) dlil_affinity_set(tp, THREAD_AFFINITY_TAG_NULL);
2774 thread_deallocate(tp);
2775 }
2776
2777 /* cleanup ifp dlil input thread, if any */
2778 ifp->if_input_thread = NULL;
2779
2780 if (inputthread != dlil_lo_thread_ptr) {
2781 #ifdef DLIL_DEBUG
2782 printf("ifnet_detach: wakeup thread threadinfo: %p "
2783 "input_thread=%p threads: cur=%d max=%d\n",
2784 inputthread, inputthread->input_thread,
2785 dlil_multithreaded_input, cur_dlil_input_threads);
2786 #endif
2787 lck_mtx_lock(inputthread->input_lck);
2788
2789 inputthread->input_waiting |= DLIL_INPUT_TERMINATE;
2790 if ((inputthread->input_waiting & DLIL_INPUT_RUNNING) == 0) {
2791 wakeup((caddr_t)&inputthread->input_waiting);
2792 }
2793 lck_mtx_unlock(inputthread->input_lck);
2794 }
2795 }
2796 /* last chance to clean up IPv4 forwarding cached route */
2797 lck_mtx_lock(ifp->if_fwd_route_lock);
2798 if (ifp->if_fwd_route.ro_rt != NULL) {
2799 rtfree(ifp->if_fwd_route.ro_rt);
2800 ifp->if_fwd_route.ro_rt = NULL;
2801 }
2802 lck_mtx_unlock(ifp->if_fwd_route_lock);
2803 dlil_write_end();
2804
2805 for (filter = TAILQ_FIRST(&fhead); filter; filter = filter_next) {
2806 filter_next = TAILQ_NEXT(filter, filt_next);
2807 dlil_detach_filter_internal(filter, 1);
2808 }
2809
2810 if (zeroed != 0) {
2811 ifp_use_reached_zero(ifp);
2812 }
2813
2814 return retval;
2815 }
2816
2817 static errno_t
2818 dlil_recycle_ioctl(
2819 __unused ifnet_t ifnet_ptr,
2820 __unused u_long ioctl_code,
2821 __unused void *ioctl_arg)
2822 {
2823 return EOPNOTSUPP;
2824 }
2825
2826 static int
2827 dlil_recycle_output(
2828 __unused struct ifnet *ifnet_ptr,
2829 struct mbuf *m)
2830 {
2831 m_freem(m);
2832 return 0;
2833 }
2834
2835 static void
2836 dlil_recycle_free(
2837 __unused ifnet_t ifnet_ptr)
2838 {
2839 }
2840
2841 static errno_t
2842 dlil_recycle_set_bpf_tap(
2843 __unused ifnet_t ifp,
2844 __unused bpf_tap_mode mode,
2845 __unused bpf_packet_func callback)
2846 {
2847 /* XXX not sure what to do here */
2848 return 0;
2849 }
2850
2851 __private_extern__
2852 int dlil_if_acquire(
2853 u_int32_t family,
2854 const void *uniqueid,
2855 size_t uniqueid_len,
2856 struct ifnet **ifp)
2857 {
2858 struct ifnet *ifp1 = NULL;
2859 struct dlil_ifnet *dlifp1 = NULL;
2860 int ret = 0;
2861
2862 lck_mtx_lock(dlil_ifnet_mutex);
2863 TAILQ_FOREACH(dlifp1, &dlil_ifnet_head, dl_if_link) {
2864
2865 ifp1 = (struct ifnet *)dlifp1;
2866
2867 if (ifp1->if_family == family) {
2868
2869 /* same uniqueid and same len or no unique id specified */
2870 if ((uniqueid_len == dlifp1->if_uniqueid_len)
2871 && !bcmp(uniqueid, dlifp1->if_uniqueid, uniqueid_len)) {
2872
2873 /* check for matching interface in use */
2874 if (ifp1->if_eflags & IFEF_INUSE) {
2875 if (uniqueid_len) {
2876 ret = EBUSY;
2877 goto end;
2878 }
2879 }
2880 else {
2881 if (!ifp1->if_lock)
2882 panic("ifp's lock is gone\n");
2883 ifnet_lock_exclusive(ifp1);
2884 ifp1->if_eflags |= (IFEF_INUSE | IFEF_REUSE);
2885 ifnet_lock_done(ifp1);
2886 *ifp = ifp1;
2887 goto end;
2888 }
2889 }
2890 }
2891 }
2892
2893 /* no interface found, allocate a new one */
2894 MALLOC(dlifp1, struct dlil_ifnet *, sizeof(*dlifp1), M_NKE, M_WAITOK);
2895 if (dlifp1 == 0) {
2896 ret = ENOMEM;
2897 goto end;
2898 }
2899
2900 bzero(dlifp1, sizeof(*dlifp1));
2901
2902 if (uniqueid_len) {
2903 MALLOC(dlifp1->if_uniqueid, void *, uniqueid_len, M_NKE, M_WAITOK);
2904 if (dlifp1->if_uniqueid == 0) {
2905 FREE(dlifp1, M_NKE);
2906 ret = ENOMEM;
2907 goto end;
2908 }
2909 bcopy(uniqueid, dlifp1->if_uniqueid, uniqueid_len);
2910 dlifp1->if_uniqueid_len = uniqueid_len;
2911 }
2912
2913 ifp1 = (struct ifnet *)dlifp1;
2914 ifp1->if_eflags |= IFEF_INUSE;
2915 ifp1->if_name = dlifp1->if_namestorage;
2916 #if CONFIG_MACF_NET
2917 mac_ifnet_label_init(ifp1);
2918 #endif
2919
2920 TAILQ_INSERT_TAIL(&dlil_ifnet_head, dlifp1, dl_if_link);
2921
2922 *ifp = ifp1;
2923
2924 end:
2925 lck_mtx_unlock(dlil_ifnet_mutex);
2926
2927 return ret;
2928 }
2929
2930 __private_extern__ void
2931 dlil_if_release(
2932 ifnet_t ifp)
2933 {
2934 struct dlil_ifnet *dlifp = (struct dlil_ifnet *)ifp;
2935
2936 /* Interface does not have a lock until it is attached - radar 3713951 */
2937 if (ifp->if_lock)
2938 ifnet_lock_exclusive(ifp);
2939 ifp->if_eflags &= ~IFEF_INUSE;
2940 ifp->if_ioctl = dlil_recycle_ioctl;
2941 ifp->if_output = dlil_recycle_output;
2942 ifp->if_free = dlil_recycle_free;
2943 ifp->if_set_bpf_tap = dlil_recycle_set_bpf_tap;
2944
2945 strncpy(dlifp->if_namestorage, ifp->if_name, IFNAMSIZ);
2946 ifp->if_name = dlifp->if_namestorage;
2947 #if CONFIG_MACF_NET
2948 /*
2949 * We can either recycle the MAC label here or in dlil_if_acquire().
2950 * It seems logical to do it here but this means that anything that
2951 * still has a handle on ifp will now see it as unlabeled.
2952 * Since the interface is "dead" that may be OK. Revisit later.
2953 */
2954 mac_ifnet_label_recycle(ifp);
2955 #endif
2956 if (ifp->if_lock)
2957 ifnet_lock_done(ifp);
2958
2959 }
2960
2961 __private_extern__ void
2962 dlil_proto_unplumb_all(struct ifnet *ifp)
2963 {
2964 /*
2965 * if_proto_hash[0-3] are for PF_INET, PF_INET6, PF_APPLETALK
2966 * and PF_VLAN, where each bucket contains exactly one entry;
2967 * PF_VLAN does not need an explicit unplumb.
2968 *
2969 * if_proto_hash[4] is for other protocols; we expect anything
2970 * in this bucket to respond to the DETACHING event (which would
2971 * have happened by now) and do the unplumb then.
2972 */
2973 (void) proto_unplumb(PF_INET, ifp);
2974 #if INET6
2975 (void) proto_unplumb(PF_INET6, ifp);
2976 #endif /* INET6 */
2977 #if NETAT
2978 (void) proto_unplumb(PF_APPLETALK, ifp);
2979 #endif /* NETAT */
2980 }