/*
- * Copyright (c) 2000-2018 Apple Inc. All rights reserved.
+ * Copyright (c) 2000-2020 Apple Inc. All rights reserved.
*
* @APPLE_OSREFERENCE_LICENSE_HEADER_START@
*
#include "bpf.h"
#ifndef __GNUC__
-#define inline
+#define inline
#else
-#define inline __inline
+#define inline __inline
#endif
#include <sys/param.h>
#include <kern/thread_call.h>
#include <libkern/section_keywords.h>
-#if CONFIG_MACF_NET
-#include <security/mac_framework.h>
-#endif /* MAC_NET */
-
#include <os/log.h>
extern int tvtohz(struct timeval *);
-#define BPF_BUFSIZE 4096
-#define UIOMOVE(cp, len, code, uio) uiomove(cp, len, uio)
+#define BPF_BUFSIZE 4096
+#define UIOMOVE(cp, len, code, uio) uiomove(cp, len, uio)
-#define PRINET 26 /* interruptible */
+#define PRINET 26 /* interruptible */
#define ISAKMP_HDR_SIZE (sizeof(struct isakmp) + sizeof(struct isakmp_gen))
#define ESP_HDR_SIZE sizeof(struct newesp)
*/
static unsigned int bpf_bufsize = BPF_BUFSIZE;
SYSCTL_INT(_debug, OID_AUTO, bpf_bufsize, CTLFLAG_RW | CTLFLAG_LOCKED,
- &bpf_bufsize, 0, "");
+ &bpf_bufsize, 0, "");
+
+static int sysctl_bpf_maxbufsize SYSCTL_HANDLER_ARGS;
+extern const int copysize_limit_panic;
+#define BPF_MAXSIZE_CAP (copysize_limit_panic >> 1)
__private_extern__ unsigned int bpf_maxbufsize = BPF_MAXBUFSIZE;
-SYSCTL_INT(_debug, OID_AUTO, bpf_maxbufsize, CTLFLAG_RW | CTLFLAG_LOCKED,
- &bpf_maxbufsize, 0, "");
+SYSCTL_PROC(_debug, OID_AUTO, bpf_maxbufsize, CTLTYPE_INT | CTLFLAG_RW | CTLFLAG_LOCKED,
+ &bpf_maxbufsize, 0,
+ sysctl_bpf_maxbufsize, "I", "Default BPF max buffer size");
+
static unsigned int bpf_maxdevices = 256;
SYSCTL_UINT(_debug, OID_AUTO, bpf_maxdevices, CTLFLAG_RW | CTLFLAG_LOCKED,
- &bpf_maxdevices, 0, "");
+ &bpf_maxdevices, 0, "");
/*
* bpf_wantpktap controls the defaul visibility of DLT_PKTAP
* For OS X is off by default so process need to use the ioctl BPF_WANT_PKTAP
* explicitly to be able to use DLT_PKTAP.
*/
-#if CONFIG_EMBEDDED
+#if !XNU_TARGET_OS_OSX
static unsigned int bpf_wantpktap = 1;
-#else
+#else /* XNU_TARGET_OS_OSX */
static unsigned int bpf_wantpktap = 0;
-#endif
+#endif /* XNU_TARGET_OS_OSX */
SYSCTL_UINT(_debug, OID_AUTO, bpf_wantpktap, CTLFLAG_RW | CTLFLAG_LOCKED,
- &bpf_wantpktap, 0, "");
+ &bpf_wantpktap, 0, "");
static int bpf_debug = 0;
SYSCTL_INT(_debug, OID_AUTO, bpf_debug, CTLFLAG_RW | CTLFLAG_LOCKED,
- &bpf_debug, 0, "");
+ &bpf_debug, 0, "");
/*
* bpf_iflist is the list of interfaces; each corresponds to an ifnet
* bpf_dtab holds pointer to the descriptors, indexed by minor device #
*/
-static struct bpf_if *bpf_iflist;
+static struct bpf_if *bpf_iflist;
#ifdef __APPLE__
/*
* BSD now stores the bpf_d in the dev_t which is a struct
* BPF_DEV_RESERVED: device opening or closing
* other: device <n> opened with pointer to storage
*/
-#define BPF_DEV_RESERVED ((struct bpf_d *)(uintptr_t)1)
-static struct bpf_d **bpf_dtab = NULL;
+#define BPF_DEV_RESERVED ((struct bpf_d *)(uintptr_t)1)
+static struct bpf_d **bpf_dtab = NULL;
static unsigned int bpf_dtab_size = 0;
-static unsigned int nbpfilter = 0;
+static unsigned int nbpfilter = 0;
decl_lck_mtx_data(static, bpf_mlock_data);
-static lck_mtx_t *bpf_mlock = &bpf_mlock_data;
-static lck_grp_t *bpf_mlock_grp;
-static lck_grp_attr_t *bpf_mlock_grp_attr;
-static lck_attr_t *bpf_mlock_attr;
+static lck_mtx_t *bpf_mlock = &bpf_mlock_data;
+static lck_grp_t *bpf_mlock_grp;
+static lck_grp_attr_t *bpf_mlock_grp_attr;
+static lck_attr_t *bpf_mlock_attr;
#endif /* __APPLE__ */
-static int bpf_allocbufs(struct bpf_d *);
-static errno_t bpf_attachd(struct bpf_d *d, struct bpf_if *bp);
-static int bpf_detachd(struct bpf_d *d, int);
-static void bpf_freed(struct bpf_d *);
-static int bpf_movein(struct uio *, int,
- struct mbuf **, struct sockaddr *, int *);
-static int bpf_setif(struct bpf_d *, ifnet_t ifp, bool, bool);
-static void bpf_timed_out(void *, void *);
-static void bpf_wakeup(struct bpf_d *);
-static u_int get_pkt_trunc_len(u_char *, u_int);
-static void catchpacket(struct bpf_d *, struct bpf_packet *, u_int, int);
-static void reset_d(struct bpf_d *);
-static int bpf_setf(struct bpf_d *, u_int, user_addr_t, u_long);
-static int bpf_getdltlist(struct bpf_d *, caddr_t, struct proc *);
-static int bpf_setdlt(struct bpf_d *, u_int);
-static int bpf_set_traffic_class(struct bpf_d *, int);
-static void bpf_set_packet_service_class(struct mbuf *, int);
-
-static void bpf_acquire_d(struct bpf_d *);
-static void bpf_release_d(struct bpf_d *);
-
-static int bpf_devsw_installed;
+static int bpf_allocbufs(struct bpf_d *);
+static errno_t bpf_attachd(struct bpf_d *d, struct bpf_if *bp);
+static int bpf_detachd(struct bpf_d *d, int);
+static void bpf_freed(struct bpf_d *);
+static int bpf_movein(struct uio *, int,
+ struct mbuf **, struct sockaddr *, int *);
+static int bpf_setif(struct bpf_d *, ifnet_t ifp, bool, bool);
+static void bpf_timed_out(void *, void *);
+static void bpf_wakeup(struct bpf_d *);
+static u_int get_pkt_trunc_len(u_char *, u_int);
+static void catchpacket(struct bpf_d *, struct bpf_packet *, u_int, int);
+static void reset_d(struct bpf_d *);
+static int bpf_setf(struct bpf_d *, u_int, user_addr_t, u_long);
+static int bpf_getdltlist(struct bpf_d *, caddr_t, struct proc *);
+static int bpf_setdlt(struct bpf_d *, u_int);
+static int bpf_set_traffic_class(struct bpf_d *, int);
+static void bpf_set_packet_service_class(struct mbuf *, int);
+
+static void bpf_acquire_d(struct bpf_d *);
+static void bpf_release_d(struct bpf_d *);
+
+static int bpf_devsw_installed;
void bpf_init(void *unused);
static int bpf_tap_callback(struct ifnet *ifp, struct mbuf *m);
* Darwin differs from BSD here, the following are static
* on BSD and not static on Darwin.
*/
- d_open_t bpfopen;
- d_close_t bpfclose;
- d_read_t bpfread;
- d_write_t bpfwrite;
- ioctl_fcn_t bpfioctl;
- select_fcn_t bpfselect;
+d_open_t bpfopen;
+d_close_t bpfclose;
+d_read_t bpfread;
+d_write_t bpfwrite;
+ioctl_fcn_t bpfioctl;
+select_fcn_t bpfselect;
/* Darwin's cdevsw struct differs slightly from BSDs */
-#define CDEV_MAJOR 23
-static struct cdevsw bpf_cdevsw = {
- /* open */ bpfopen,
- /* close */ bpfclose,
- /* read */ bpfread,
- /* write */ bpfwrite,
- /* ioctl */ bpfioctl,
- /* stop */ eno_stop,
- /* reset */ eno_reset,
- /* tty */ NULL,
- /* select */ bpfselect,
- /* mmap */ eno_mmap,
- /* strategy */ eno_strat,
- /* getc */ eno_getc,
- /* putc */ eno_putc,
- /* type */ 0
+#define CDEV_MAJOR 23
+static const struct cdevsw bpf_cdevsw = {
+ .d_open = bpfopen,
+ .d_close = bpfclose,
+ .d_read = bpfread,
+ .d_write = bpfwrite,
+ .d_ioctl = bpfioctl,
+ .d_stop = eno_stop,
+ .d_reset = eno_reset,
+ .d_ttys = NULL,
+ .d_select = bpfselect,
+ .d_mmap = eno_mmap,
+ .d_strategy = eno_strat,
+ .d_reserved_1 = eno_getc,
+ .d_reserved_2 = eno_putc,
+ .d_type = 0
};
-#define SOCKADDR_HDR_LEN offsetof(struct sockaddr, sa_data)
+#define SOCKADDR_HDR_LEN offsetof(struct sockaddr, sa_data)
static int
bpf_movein(struct uio *uio, int linktype, struct mbuf **mp,
int hlen;
switch (linktype) {
-
#if SLIP
case DLT_SLIP:
sa_family = AF_INET;
* specified anyway.
*/
sa_family = AF_UNSPEC;
- hlen = 12; /* XXX 4(ATM_PH) + 3(LLC) + 5(SNAP) */
+ hlen = 12; /* XXX 4(ATM_PH) + 3(LLC) + 5(SNAP) */
break;
#endif
case DLT_PPP:
sa_family = AF_UNSPEC;
- hlen = 4; /* This should match PPP_HDRLEN */
+ hlen = 4; /* This should match PPP_HDRLEN */
break;
case DLT_APPLE_IP_OVER_IEEE1394:
hlen = sizeof(struct firewire_header);
break;
- case DLT_IEEE802_11: /* IEEE 802.11 wireless */
+ case DLT_IEEE802_11: /* IEEE 802.11 wireless */
sa_family = AF_IEEE80211;
hlen = 0;
break;
break;
default:
- return (EIO);
+ return EIO;
}
// LP64todo - fix this!
len = uio_resid(uio);
*datlen = len - hlen;
- if ((unsigned)len > MCLBYTES)
- return (EIO);
+ if ((unsigned)len > MCLBYTES) {
+ return EIO;
+ }
if (sockp) {
/*
* for the link level header.
*/
if ((hlen + SOCKADDR_HDR_LEN) > sockp->sa_len) {
- return (EIO);
+ return EIO;
}
sockp->sa_family = sa_family;
} else {
}
MGETHDR(m, M_WAIT, MT_DATA);
- if (m == 0)
- return (ENOBUFS);
+ if (m == 0) {
+ return ENOBUFS;
+ }
if ((unsigned)len > MHLEN) {
MCLGET(m, M_WAIT);
if ((m->m_flags & M_EXT) == 0) {
m->m_len -= hlen;
m->m_data += hlen; /* XXX */
error = UIOMOVE((caddr_t)sockp->sa_data, hlen, UIO_WRITE, uio);
- if (error)
+ if (error) {
goto bad;
+ }
}
error = UIOMOVE(mtod(m, caddr_t), len - hlen, UIO_WRITE, uio);
- if (error)
+ if (error) {
goto bad;
+ }
/* Check for multicast destination */
switch (linktype) {
- case DLT_EN10MB: {
- struct ether_header *eh;
-
- eh = mtod(m, struct ether_header *);
- if (ETHER_IS_MULTICAST(eh->ether_dhost)) {
- if (_ether_cmp(etherbroadcastaddr,
- eh->ether_dhost) == 0) {
- m->m_flags |= M_BCAST;
- } else {
- m->m_flags |= M_MCAST;
- }
+ case DLT_EN10MB: {
+ struct ether_header *eh;
+
+ eh = mtod(m, struct ether_header *);
+ if (ETHER_IS_MULTICAST(eh->ether_dhost)) {
+ if (_ether_cmp(etherbroadcastaddr,
+ eh->ether_dhost) == 0) {
+ m->m_flags |= M_BCAST;
+ } else {
+ m->m_flags |= M_MCAST;
}
- break;
}
+ break;
+ }
}
- return (0);
+ return 0;
bad:
m_freem(m);
- return (error);
+ return error;
}
#ifdef __APPLE__
static void
bpf_make_dev_t(int maj)
{
- static int bpf_growing = 0;
- unsigned int cur_size = nbpfilter, i;
+ static int bpf_growing = 0;
+ unsigned int cur_size = nbpfilter, i;
- if (nbpfilter >= bpf_maxdevices)
+ if (nbpfilter >= bpf_maxdevices) {
return;
+ }
while (bpf_growing) {
/* Wait until new device has been created */
new_dtab_size = bpf_dtab_size + NBPFILTER;
new_dtab = (struct bpf_d **)_MALLOC(
- sizeof(struct bpf_d *) * new_dtab_size, M_DEVBUF, M_WAIT);
+ sizeof(struct bpf_d *) * new_dtab_size, M_DEVBUF, M_WAIT);
if (new_dtab == 0) {
printf("bpf_make_dev_t: malloc bpf_dtab failed\n");
goto done;
old_dtab = bpf_dtab;
bpf_dtab = new_dtab;
bpf_dtab_size = new_dtab_size;
- if (old_dtab != NULL)
+ if (old_dtab != NULL) {
_FREE(old_dtab, M_DEVBUF);
+ }
}
i = nbpfilter++;
(void) devfs_make_node(makedev(maj, i),
- DEVFS_CHAR, UID_ROOT, GID_WHEEL, 0600,
- "bpf%d", i);
+ DEVFS_CHAR, UID_ROOT, GID_WHEEL, 0600,
+ "bpf%d", i);
done:
bpf_growing = 0;
wakeup((caddr_t)&bpf_growing);
bpf_attachd(struct bpf_d *d, struct bpf_if *bp)
{
int first = bp->bif_dlist == NULL;
- int error = 0;
+ int error = 0;
/*
* Point d at bp, and add d to the interface's list of listeners.
if (first) {
/* Find the default bpf entry for this ifp */
if (bp->bif_ifp->if_bpf == NULL) {
- struct bpf_if *tmp, *primary = NULL;
+ struct bpf_if *tmp, *primary = NULL;
for (tmp = bpf_iflist; tmp; tmp = tmp->bif_next) {
if (tmp->bif_ifp == bp->bif_ifp) {
bp->bif_ifp->if_bpf = primary;
}
/* Only call dlil_set_bpf_tap for primary dlt */
- if (bp->bif_ifp->if_bpf == bp)
+ if (bp->bif_ifp->if_bpf == bp) {
dlil_set_bpf_tap(bp->bif_ifp, BPF_TAP_INPUT_OUTPUT,
bpf_tap_callback);
+ }
- if (bp->bif_tap != NULL)
+ if (bp->bif_tap != NULL) {
error = bp->bif_tap(bp->bif_ifp, bp->bif_dlt,
BPF_TAP_INPUT_OUTPUT);
+ }
}
/*
} else {
d->bd_flags &= ~BPF_FINALIZE_PKTAP;
}
- return (error);
+ return error;
}
/*
/*
* Some other thread already detached
*/
- if ((d->bd_flags & (BPF_DETACHED | BPF_DETACHING)) != 0)
+ if ((d->bd_flags & (BPF_DETACHED | BPF_DETACHING)) != 0) {
goto done;
+ }
/*
* This thread is doing the detach
*/
ifp = d->bd_bif->bif_ifp;
bp = d->bd_bif;
- if (bpf_debug != 0)
+ if (bpf_debug != 0) {
printf("%s: %llx %s%s\n",
__func__, (uint64_t)VM_KERNEL_ADDRPERM(d),
if_name(ifp), closing ? " closing" : "");
+ }
/* Remove d from the interface's descriptor list. */
p = &bp->bif_dlist;
while (*p != d) {
p = &(*p)->bd_next;
- if (*p == 0)
+ if (*p == 0) {
panic("bpf_detachd: descriptor not in list");
+ }
}
*p = (*p)->bd_next;
if (bp->bif_dlist == 0) {
* Let the driver know that there are no more listeners.
*/
/* Only call dlil_set_bpf_tap for primary dlt */
- if (bp->bif_ifp->if_bpf == bp)
+ if (bp->bif_ifp->if_bpf == bp) {
dlil_set_bpf_tap(ifp, BPF_TAP_DISABLE, NULL);
- if (bp->bif_tap)
+ }
+ if (bp->bif_tap) {
bp->bif_tap(ifp, bp->bif_dlt, BPF_TAP_DISABLE);
+ }
- for (bp = bpf_iflist; bp; bp = bp->bif_next)
- if (bp->bif_ifp == ifp && bp->bif_dlist != 0)
+ for (bp = bpf_iflist; bp; bp = bp->bif_next) {
+ if (bp->bif_ifp == ifp && bp->bif_dlist != 0) {
break;
- if (bp == NULL)
+ }
+ }
+ if (bp == NULL) {
ifp->if_bpf = NULL;
+ }
}
d->bd_bif = NULL;
/*
/*
* When closing makes sure no other thread refer to the bpf_d
*/
- if (bpf_debug != 0)
+ if (bpf_debug != 0) {
printf("%s: %llx done\n",
__func__, (uint64_t)VM_KERNEL_ADDRPERM(d));
+ }
/*
* Let the caller know the bpf_d is closed
*/
- if (bpf_closed)
- return (1);
- else
- return (0);
+ if (bpf_closed) {
+ return 1;
+ } else {
+ return 0;
+ }
}
/*
tv.tv_usec = (d->bd_rtout % hz) * tick;
clock_interval_to_deadline(
- (uint64_t)tv.tv_sec * USEC_PER_SEC + tv.tv_usec,
- NSEC_PER_USEC, &deadline);
+ (uint64_t)tv.tv_sec * USEC_PER_SEC + tv.tv_usec,
+ NSEC_PER_USEC, &deadline);
/*
* The state is BPF_IDLE, so the timer hasn't
* been started yet, and hasn't gone off yet;
* find the state is BPF_IDLE, and just release the
* lock and return.
*/
- return (thread_call_cancel(d->bd_thread_call));
+ return thread_call_cancel(d->bd_thread_call);
}
void
LCK_MTX_ASSERT(bpf_mlock, LCK_MTX_ASSERT_OWNED);
- if (d->bd_refcnt <= 0)
+ if (d->bd_refcnt <= 0) {
panic("%s: %p refcnt <= 0", __func__, d);
+ }
d->bd_refcnt -= 1;
if (d->bd_refcnt == 0) {
/* Assert the device is detached */
- if ((d->bd_flags & BPF_DETACHED) == 0)
+ if ((d->bd_flags & BPF_DETACHED) == 0) {
panic("%s: %p BPF_DETACHED not set", __func__, d);
+ }
_FREE(d, M_DEVBUF);
}
/* ARGSUSED */
int
bpfopen(dev_t dev, int flags, __unused int fmt,
- struct proc *p)
+ struct proc *p)
{
struct bpf_d *d;
lck_mtx_lock(bpf_mlock);
if ((unsigned int) minor(dev) >= nbpfilter) {
lck_mtx_unlock(bpf_mlock);
- return (ENXIO);
+ return ENXIO;
}
/*
* New device nodes are created on demand when opening the last one.
* last node. If not all processes are blocked, they could unexpectedly
* get ENOENT and abort their opening loop.
*/
- if ((unsigned int) minor(dev) == (nbpfilter - 1))
+ if ((unsigned int) minor(dev) == (nbpfilter - 1)) {
bpf_make_dev_t(major(dev));
+ }
/*
* Each minor can be opened by only one process. If the requested
bpf_dtab[minor(dev)] = BPF_DEV_RESERVED;
} else {
lck_mtx_unlock(bpf_mlock);
- return (EBUSY);
+ return EBUSY;
}
d = (struct bpf_d *)_MALLOC(sizeof(struct bpf_d), M_DEVBUF,
M_WAIT | M_ZERO);
printf("bpfopen: malloc bpf_d failed\n");
bpf_dtab[minor(dev)] = NULL;
lck_mtx_unlock(bpf_mlock);
- return (ENOMEM);
+ return ENOMEM;
}
/* Mark "in use" and do most initialization. */
d->bd_state = BPF_IDLE;
d->bd_traffic_class = SO_TC_BE;
d->bd_flags |= BPF_DETACHED;
- if (bpf_wantpktap)
+ if (bpf_wantpktap) {
d->bd_flags |= BPF_WANT_PKTAP;
- else
+ } else {
d->bd_flags &= ~BPF_WANT_PKTAP;
+ }
d->bd_thread_call = thread_call_allocate(bpf_timed_out, d);
if (d->bd_thread_call == NULL) {
printf("bpfopen: malloc thread call failed\n");
bpf_release_d(d);
lck_mtx_unlock(bpf_mlock);
- return (ENOMEM);
+ return ENOMEM;
}
d->bd_opened_by = p;
uuid_generate(d->bd_uuid);
-#if CONFIG_MACF_NET
- mac_bpfdesc_label_init(d);
- mac_bpfdesc_label_associate(kauth_cred_get(), d);
-#endif
bpf_dtab[minor(dev)] = d; /* Mark opened */
lck_mtx_unlock(bpf_mlock);
- return (0);
+ return 0;
}
/*
d = bpf_dtab[minor(dev)];
if (d == NULL || d == BPF_DEV_RESERVED) {
lck_mtx_unlock(bpf_mlock);
- return (ENXIO);
+ return ENXIO;
}
/*
*/
d->bd_flags |= BPF_CLOSING;
- if (bpf_debug != 0)
+ if (bpf_debug != 0) {
printf("%s: %llx\n",
__func__, (uint64_t)VM_KERNEL_ADDRPERM(d));
+ }
bpf_dtab[minor(dev)] = BPF_DEV_RESERVED; /* Reserve while closing */
* Deal with any in-progress timeouts.
*/
switch (d->bd_state) {
- case BPF_IDLE:
- /*
- * Not waiting for a timeout, and no timeout happened.
- */
- break;
+ case BPF_IDLE:
+ /*
+ * Not waiting for a timeout, and no timeout happened.
+ */
+ break;
- case BPF_WAITING:
+ case BPF_WAITING:
+ /*
+ * Waiting for a timeout.
+ * Cancel any timer that has yet to go off,
+ * and mark the state as "closing".
+ * Then drop the lock to allow any timers that
+ * *have* gone off to run to completion, and wait
+ * for them to finish.
+ */
+ if (!bpf_stop_timer(d)) {
/*
- * Waiting for a timeout.
- * Cancel any timer that has yet to go off,
- * and mark the state as "closing".
- * Then drop the lock to allow any timers that
- * *have* gone off to run to completion, and wait
- * for them to finish.
+ * There was no pending call, so the call must
+ * have been in progress. Wait for the call to
+ * complete; we have to drop the lock while
+ * waiting. to let the in-progrss call complete
*/
- if (!bpf_stop_timer(d)) {
- /*
- * There was no pending call, so the call must
- * have been in progress. Wait for the call to
- * complete; we have to drop the lock while
- * waiting. to let the in-progrss call complete
- */
- d->bd_state = BPF_DRAINING;
- while (d->bd_state == BPF_DRAINING)
- msleep((caddr_t)d, bpf_mlock, PRINET,
- "bpfdraining", NULL);
+ d->bd_state = BPF_DRAINING;
+ while (d->bd_state == BPF_DRAINING) {
+ msleep((caddr_t)d, bpf_mlock, PRINET,
+ "bpfdraining", NULL);
}
- d->bd_state = BPF_IDLE;
- break;
+ }
+ d->bd_state = BPF_IDLE;
+ break;
- case BPF_TIMED_OUT:
- /*
- * Timer went off, and the timeout routine finished.
- */
- d->bd_state = BPF_IDLE;
- break;
+ case BPF_TIMED_OUT:
+ /*
+ * Timer went off, and the timeout routine finished.
+ */
+ d->bd_state = BPF_IDLE;
+ break;
- case BPF_DRAINING:
- /*
- * Another thread is blocked on a close waiting for
- * a timeout to finish.
- * This "shouldn't happen", as the first thread to enter
- * bpfclose() will set bpf_dtab[minor(dev)] to 1, and
- * all subsequent threads should see that and fail with
- * ENXIO.
- */
- panic("Two threads blocked in a BPF close");
- break;
+ case BPF_DRAINING:
+ /*
+ * Another thread is blocked on a close waiting for
+ * a timeout to finish.
+ * This "shouldn't happen", as the first thread to enter
+ * bpfclose() will set bpf_dtab[minor(dev)] to 1, and
+ * all subsequent threads should see that and fail with
+ * ENXIO.
+ */
+ panic("Two threads blocked in a BPF close");
+ break;
}
- if (d->bd_bif)
+ if (d->bd_bif) {
bpf_detachd(d, 1);
+ }
selthreadclear(&d->bd_sel);
-#if CONFIG_MACF_NET
- mac_bpfdesc_label_destroy(d);
-#endif
thread_call_free(d->bd_thread_call);
- while (d->bd_hbuf_read != 0)
+ while (d->bd_hbuf_read != 0) {
msleep((caddr_t)d, bpf_mlock, PRINET, "bpf_reading", NULL);
+ }
bpf_freed(d);
/* Mark free in same context as bpfopen comes to check */
- bpf_dtab[minor(dev)] = NULL; /* Mark closed */
+ bpf_dtab[minor(dev)] = NULL; /* Mark closed */
bpf_release_d(d);
lck_mtx_unlock(bpf_mlock);
- return (0);
+ return 0;
}
-#define BPF_SLEEP bpf_sleep
+#define BPF_SLEEP bpf_sleep
static int
bpf_sleep(struct bpf_d *d, int pri, const char *wmesg, int timo)
{
u_int64_t abstime = 0;
- if (timo != 0)
+ if (timo != 0) {
clock_interval_to_deadline(timo, NSEC_PER_SEC / hz, &abstime);
+ }
- return (msleep1((caddr_t)d, bpf_mlock, pri, wmesg, abstime));
+ return msleep1((caddr_t)d, bpf_mlock, pri, wmesg, abstime);
}
static void
pktap_v2_hdr = (struct pktap_v2_hdr *)pktaphdr;
- if (pktap_v2_hdr->pth_flags & PTH_FLAG_DELAY_PKTAP)
+ if (pktap_v2_hdr->pth_flags & PTH_FLAG_DELAY_PKTAP) {
pktap_v2_finalize_proc_info(pktap_v2_hdr);
+ }
} else {
- if (pktaphdr->pth_flags & PTH_FLAG_DELAY_PKTAP)
+ if (pktaphdr->pth_flags & PTH_FLAG_DELAY_PKTAP) {
pktap_finalize_proc_info(pktaphdr);
+ }
if (pktaphdr->pth_flags & PTH_FLAG_TSTAMP) {
hp->bh_tstamp.tv_sec = pktaphdr->pth_tstamp.tv_sec;
* into the hold slot, and the free buffer into the store slot.
* Zero the length of the new store buffer.
*/
-#define ROTATE_BUFFERS(d) \
+#define ROTATE_BUFFERS(d) \
if (d->bd_hbuf_read != 0) \
- panic("rotating bpf buffers during read"); \
+ panic("rotating bpf buffers during read"); \
(d)->bd_hbuf = (d)->bd_sbuf; \
(d)->bd_hlen = (d)->bd_slen; \
(d)->bd_hcnt = (d)->bd_scnt; \
if (d == NULL || d == BPF_DEV_RESERVED ||
(d->bd_flags & BPF_CLOSING) != 0) {
lck_mtx_unlock(bpf_mlock);
- return (ENXIO);
+ return ENXIO;
}
bpf_acquire_d(d);
if (uio_resid(uio) != d->bd_bufsize) {
bpf_release_d(d);
lck_mtx_unlock(bpf_mlock);
- return (EINVAL);
+ return EINVAL;
}
- if (d->bd_state == BPF_WAITING)
+ if (d->bd_state == BPF_WAITING) {
bpf_stop_timer(d);
+ }
timed_out = (d->bd_state == BPF_TIMED_OUT);
d->bd_state = BPF_IDLE;
- while (d->bd_hbuf_read != 0)
+ while (d->bd_hbuf_read != 0) {
msleep((caddr_t)d, bpf_mlock, PRINET, "bpf_reading", NULL);
+ }
if ((d->bd_flags & BPF_CLOSING) != 0) {
bpf_release_d(d);
lck_mtx_unlock(bpf_mlock);
- return (ENXIO);
+ return ENXIO;
}
/*
* If the hold buffer is empty, then do a timed sleep, which
if (d->bd_bif == NULL) {
bpf_release_d(d);
lck_mtx_unlock(bpf_mlock);
- return (ENXIO);
+ return ENXIO;
}
if (ioflag & IO_NDELAY) {
bpf_release_d(d);
lck_mtx_unlock(bpf_mlock);
- return (EWOULDBLOCK);
+ return EWOULDBLOCK;
}
- error = BPF_SLEEP(d, PRINET|PCATCH, "bpf", d->bd_rtout);
+ error = BPF_SLEEP(d, PRINET | PCATCH, "bpf", d->bd_rtout);
/*
* Make sure device is still opened
*/
if ((d->bd_flags & BPF_CLOSING) != 0) {
bpf_release_d(d);
lck_mtx_unlock(bpf_mlock);
- return (ENXIO);
+ return ENXIO;
}
- while (d->bd_hbuf_read != 0)
+ while (d->bd_hbuf_read != 0) {
msleep((caddr_t)d, bpf_mlock, PRINET, "bpf_reading",
NULL);
+ }
if ((d->bd_flags & BPF_CLOSING) != 0) {
bpf_release_d(d);
lck_mtx_unlock(bpf_mlock);
- return (ENXIO);
+ return ENXIO;
}
if (error == EINTR || error == ERESTART) {
__func__, (uint64_t)VM_KERNEL_ADDRPERM(d));
error = EINTR;
}
- return (error);
+ return error;
}
if (error == EWOULDBLOCK) {
/*
* which may be nothing. If there is something
* in the store buffer, we can rotate the buffers.
*/
- if (d->bd_hbuf)
+ if (d->bd_hbuf) {
/*
* We filled up the buffer in between
* getting the timeout and arriving
* here, so we don't need to rotate.
*/
break;
+ }
if (d->bd_slen == 0) {
bpf_release_d(d);
lck_mtx_unlock(bpf_mlock);
- return (0);
+ return 0;
}
ROTATE_BUFFERS(d);
break;
ehp = (struct bpf_hdr_ext *)(void *)p;
if ((flowid = ehp->bh_flowid) != 0) {
- if (ehp->bh_proto == IPPROTO_TCP)
+ if (ehp->bh_proto == IPPROTO_TCP) {
found = inp_findinpcb_procinfo(&tcbinfo,
flowid, &soprocinfo);
- else if (ehp->bh_proto == IPPROTO_UDP)
+ } else if (ehp->bh_proto == IPPROTO_UDP) {
found = inp_findinpcb_procinfo(&udbinfo,
flowid, &soprocinfo);
+ }
if (found == 1) {
ehp->bh_pid = soprocinfo.spi_pid;
- proc_name(ehp->bh_pid, ehp->bh_comm,
- MAXCOMLEN);
+ strlcpy(&ehp->bh_comm[0], &soprocinfo.spi_proc_name[0], sizeof(ehp->bh_comm));
}
ehp->bh_flowid = 0;
}
if ((d->bd_flags & BPF_CLOSING) != 0) {
bpf_release_d(d);
lck_mtx_unlock(bpf_mlock);
- return (ENXIO);
+ return ENXIO;
}
d->bd_hbuf_read = 0;
bpf_release_d(d);
lck_mtx_unlock(bpf_mlock);
- return (error);
-
+ return error;
}
/*
d->bd_state = BPF_IDLE;
}
wakeup((caddr_t)d);
- if (d->bd_async && d->bd_sig && d->bd_sigio)
+ if (d->bd_async && d->bd_sig && d->bd_sigio) {
pgsigio(d->bd_sigio, d->bd_sig);
+ }
selwakeup(&d->bd_sel);
- if ((d->bd_flags & BPF_KNOTE))
+ if ((d->bd_flags & BPF_KNOTE)) {
KNOTE(&d->bd_sel.si_note, 1);
+ }
}
static void
* now stuff to read, wake it up.
*/
d->bd_state = BPF_TIMED_OUT;
- if (d->bd_slen != 0)
+ if (d->bd_slen != 0) {
bpf_wakeup(d);
+ }
} else if (d->bd_state == BPF_DRAINING) {
/*
* A close is waiting for this to finish.
}
/* keep in sync with bpf_movein above: */
-#define MAX_DATALINK_HDR_LEN (sizeof(struct firewire_header))
+#define MAX_DATALINK_HDR_LEN (sizeof(struct firewire_header))
int
bpfwrite(dev_t dev, struct uio *uio, __unused int ioflag)
struct ifnet *ifp;
struct mbuf *m = NULL;
int error;
- char dst_buf[SOCKADDR_HDR_LEN + MAX_DATALINK_HDR_LEN];
+ char dst_buf[SOCKADDR_HDR_LEN + MAX_DATALINK_HDR_LEN];
int datlen = 0;
int bif_dlt;
int bd_hdrcmplt;
if (d == NULL || d == BPF_DEV_RESERVED ||
(d->bd_flags & BPF_CLOSING) != 0) {
lck_mtx_unlock(bpf_mlock);
- return (ENXIO);
+ return ENXIO;
}
bpf_acquire_d(d);
if (d->bd_bif == 0) {
bpf_release_d(d);
lck_mtx_unlock(bpf_mlock);
- return (ENXIO);
+ return ENXIO;
}
ifp = d->bd_bif->bif_ifp;
if ((ifp->if_flags & IFF_UP) == 0) {
bpf_release_d(d);
lck_mtx_unlock(bpf_mlock);
- return (ENETDOWN);
+ return ENETDOWN;
}
if (uio_resid(uio) == 0) {
bpf_release_d(d);
lck_mtx_unlock(bpf_mlock);
- return (0);
+ return 0;
}
((struct sockaddr *)dst_buf)->sa_len = sizeof(dst_buf);
lck_mtx_unlock(bpf_mlock);
error = bpf_movein(uio, bif_dlt, &m,
- bd_hdrcmplt ? NULL : (struct sockaddr *)dst_buf,
- &datlen);
+ bd_hdrcmplt ? NULL : (struct sockaddr *)dst_buf,
+ &datlen);
/* take the lock again */
lck_mtx_lock(bpf_mlock);
if (error) {
bpf_release_d(d);
lck_mtx_unlock(bpf_mlock);
- return (error);
+ return error;
}
/* verify the device is still open */
bpf_release_d(d);
lck_mtx_unlock(bpf_mlock);
m_freem(m);
- return (ENXIO);
+ return ENXIO;
}
if (d->bd_bif == NULL) {
bpf_release_d(d);
lck_mtx_unlock(bpf_mlock);
m_free(m);
- return (ENXIO);
+ return ENXIO;
}
if ((unsigned)datlen > ifp->if_mtu) {
bpf_release_d(d);
lck_mtx_unlock(bpf_mlock);
m_freem(m);
- return (EMSGSIZE);
+ return EMSGSIZE;
}
-#if CONFIG_MACF_NET
- mac_mbuf_label_associate_bpfdesc(d, m);
-#endif
-
bpf_set_packet_service_class(m, d->bd_traffic_class);
lck_mtx_unlock(bpf_mlock);
* The driver frees the mbuf.
*/
if (d->bd_hdrcmplt) {
- if (d->bd_bif->bif_send)
+ if (d->bd_bif->bif_send) {
error = d->bd_bif->bif_send(ifp, d->bd_bif->bif_dlt, m);
- else
+ } else {
error = dlil_output(ifp, 0, m, NULL, NULL, 1, NULL);
+ }
} else {
error = dlil_output(ifp, PF_INET, m, NULL,
(struct sockaddr *)dst_buf, 0, NULL);
bpf_release_d(d);
lck_mtx_unlock(bpf_mlock);
- return (error);
+ return error;
}
/*
static void
reset_d(struct bpf_d *d)
{
- if (d->bd_hbuf_read != 0)
+ if (d->bd_hbuf_read != 0) {
panic("resetting buffers during read");
+ }
if (d->bd_hbuf) {
/* Free the hold buffer. */
struct bpf_d *d = bpf_dtab[i];
if (d == NULL || d == BPF_DEV_RESERVED ||
- (d->bd_flags & BPF_CLOSING) != 0)
+ (d->bd_flags & BPF_CLOSING) != 0) {
continue;
- if (uuid_compare(uuid, d->bd_uuid) == 0)
- return (d);
+ }
+ if (uuid_compare(uuid, d->bd_uuid) == 0) {
+ return d;
+ }
}
- return (NULL);
+ return NULL;
}
/*
os_log_info(OS_LOG_DEFAULT,
"%s: uuids not found error %d",
__func__, error);
- return (error);
+ return error;
}
if (d_from->bd_opened_by != d_to->bd_opened_by) {
error = EACCES;
os_log_info(OS_LOG_DEFAULT,
"%s: processes not matching error %d",
__func__, error);
- return (error);
+ return error;
}
/*
* Prevent any read while copying
*/
- while (d_to->bd_hbuf_read != 0)
+ while (d_to->bd_hbuf_read != 0) {
msleep((caddr_t)d_to, bpf_mlock, PRINET, __func__, NULL);
+ }
d_to->bd_hbuf_read = 1;
- while (d_from->bd_hbuf_read != 0)
+ while (d_from->bd_hbuf_read != 0) {
msleep((caddr_t)d_from, bpf_mlock, PRINET, __func__, NULL);
+ }
d_from->bd_hbuf_read = 1;
/*
d_to->bd_hbuf_read = 0;
wakeup((caddr_t)d_to);
- return (error);
+ return error;
}
/*
if (d == NULL || d == BPF_DEV_RESERVED ||
(d->bd_flags & BPF_CLOSING) != 0) {
lck_mtx_unlock(bpf_mlock);
- return (ENXIO);
+ return ENXIO;
}
bpf_acquire_d(d);
- if (d->bd_state == BPF_WAITING)
+ if (d->bd_state == BPF_WAITING) {
bpf_stop_timer(d);
+ }
d->bd_state = BPF_IDLE;
switch (cmd) {
-
default:
error = EINVAL;
break;
/*
* Check for read packet available.
*/
- case FIONREAD: /* int */
- {
- int n;
-
- n = d->bd_slen;
- if (d->bd_hbuf && d->bd_hbuf_read == 0)
- n += d->bd_hlen;
+ case FIONREAD: /* int */
+ {
+ int n;
- bcopy(&n, addr, sizeof (n));
- break;
+ n = d->bd_slen;
+ if (d->bd_hbuf && d->bd_hbuf_read == 0) {
+ n += d->bd_hlen;
}
- case SIOCGIFADDR: /* struct ifreq */
- {
- struct ifnet *ifp;
+ bcopy(&n, addr, sizeof(n));
+ break;
+ }
- if (d->bd_bif == 0)
- error = EINVAL;
- else {
- ifp = d->bd_bif->bif_ifp;
- error = ifnet_ioctl(ifp, 0, cmd, addr);
- }
- break;
+ case SIOCGIFADDR: /* struct ifreq */
+ {
+ struct ifnet *ifp;
+
+ if (d->bd_bif == 0) {
+ error = EINVAL;
+ } else {
+ ifp = d->bd_bif->bif_ifp;
+ error = ifnet_ioctl(ifp, 0, cmd, addr);
}
+ break;
+ }
/*
* Get buffer len [for read()].
*/
- case BIOCGBLEN: /* u_int */
- bcopy(&d->bd_bufsize, addr, sizeof (u_int));
+ case BIOCGBLEN: /* u_int */
+ bcopy(&d->bd_bufsize, addr, sizeof(u_int));
break;
/*
* Set buffer length.
*/
- case BIOCSBLEN: { /* u_int */
+ case BIOCSBLEN: { /* u_int */
u_int size;
unsigned int maxbufsize = bpf_maxbufsize;
error = EINVAL;
break;
}
- bcopy(addr, &size, sizeof (size));
+ bcopy(addr, &size, sizeof(size));
if (size > maxbufsize) {
d->bd_bufsize = maxbufsize;
}
/* It's a read/write ioctl */
- bcopy(&d->bd_bufsize, addr, sizeof (u_int));
+ bcopy(&d->bd_bufsize, addr, sizeof(u_int));
break;
}
/*
* Set link layer read filter.
*/
case BIOCSETF32:
- case BIOCSETFNR32: { /* struct bpf_program32 */
+ case BIOCSETFNR32: { /* struct bpf_program32 */
struct bpf_program32 prg32;
- bcopy(addr, &prg32, sizeof (prg32));
+ bcopy(addr, &prg32, sizeof(prg32));
error = bpf_setf(d, prg32.bf_len,
CAST_USER_ADDR_T(prg32.bf_insns), cmd);
break;
}
case BIOCSETF64:
- case BIOCSETFNR64: { /* struct bpf_program64 */
+ case BIOCSETFNR64: { /* struct bpf_program64 */
struct bpf_program64 prg64;
- bcopy(addr, &prg64, sizeof (prg64));
+ bcopy(addr, &prg64, sizeof(prg64));
error = bpf_setf(d, prg64.bf_len, prg64.bf_insns, cmd);
break;
}
lck_mtx_unlock(bpf_mlock);
error = ifnet_set_promiscuous(d->bd_bif->bif_ifp, 1);
lck_mtx_lock(bpf_mlock);
- if (error == 0)
+ if (error == 0) {
d->bd_promisc = 1;
+ }
}
break;
/*
* Get device parameters.
*/
- case BIOCGDLT: /* u_int */
- if (d->bd_bif == 0)
+ case BIOCGDLT: /* u_int */
+ if (d->bd_bif == 0) {
error = EINVAL;
- else
- bcopy(&d->bd_bif->bif_dlt, addr, sizeof (u_int));
+ } else {
+ bcopy(&d->bd_bif->bif_dlt, addr, sizeof(u_int));
+ }
break;
/*
* Get a list of supported data link types.
*/
- case BIOCGDLTLIST: /* struct bpf_dltlist */
+ case BIOCGDLTLIST: /* struct bpf_dltlist */
if (d->bd_bif == NULL) {
error = EINVAL;
} else {
/*
* Set data link type.
*/
- case BIOCSDLT: /* u_int */
+ case BIOCSDLT: /* u_int */
if (d->bd_bif == NULL) {
error = EINVAL;
} else {
u_int dlt;
- bcopy(addr, &dlt, sizeof (dlt));
+ bcopy(addr, &dlt, sizeof(dlt));
if (dlt == DLT_PKTAP &&
!(d->bd_flags & BPF_WANT_PKTAP)) {
/*
* Get interface name.
*/
- case BIOCGETIF: /* struct ifreq */
- if (d->bd_bif == 0)
+ case BIOCGETIF: /* struct ifreq */
+ if (d->bd_bif == 0) {
error = EINVAL;
- else {
+ } else {
struct ifnet *const ifp = d->bd_bif->bif_ifp;
snprintf(((struct ifreq *)(void *)addr)->ifr_name,
- sizeof (ifr.ifr_name), "%s", if_name(ifp));
+ sizeof(ifr.ifr_name), "%s", if_name(ifp));
}
break;
/*
* Set interface.
*/
- case BIOCSETIF: { /* struct ifreq */
- ifnet_t ifp;
+ case BIOCSETIF: { /* struct ifreq */
+ ifnet_t ifp;
- bcopy(addr, &ifr, sizeof (ifr));
+ bcopy(addr, &ifr, sizeof(ifr));
ifr.ifr_name[IFNAMSIZ - 1] = '\0';
ifp = ifunit(ifr.ifr_name);
- if (ifp == NULL)
+ if (ifp == NULL) {
error = ENXIO;
- else
+ } else {
error = bpf_setif(d, ifp, true, false);
+ }
break;
}
/*
* Set read timeout.
*/
- case BIOCSRTIMEOUT32: { /* struct user32_timeval */
+ case BIOCSRTIMEOUT32: { /* struct user32_timeval */
struct user32_timeval _tv;
struct timeval tv;
- bcopy(addr, &_tv, sizeof (_tv));
+ bcopy(addr, &_tv, sizeof(_tv));
tv.tv_sec = _tv.tv_sec;
tv.tv_usec = _tv.tv_usec;
* Subtract 1 tick from tvtohz() since this isn't
* a one-shot timer.
*/
- if ((error = itimerfix(&tv)) == 0)
+ if ((error = itimerfix(&tv)) == 0) {
d->bd_rtout = tvtohz(&tv) - 1;
+ }
break;
}
- case BIOCSRTIMEOUT64: { /* struct user64_timeval */
+ case BIOCSRTIMEOUT64: { /* struct user64_timeval */
struct user64_timeval _tv;
struct timeval tv;
- bcopy(addr, &_tv, sizeof (_tv));
+ bcopy(addr, &_tv, sizeof(_tv));
tv.tv_sec = _tv.tv_sec;
tv.tv_usec = _tv.tv_usec;
* Subtract 1 tick from tvtohz() since this isn't
* a one-shot timer.
*/
- if ((error = itimerfix(&tv)) == 0)
+ if ((error = itimerfix(&tv)) == 0) {
d->bd_rtout = tvtohz(&tv) - 1;
+ }
break;
}
/*
* Get read timeout.
*/
- case BIOCGRTIMEOUT32: { /* struct user32_timeval */
+ case BIOCGRTIMEOUT32: { /* struct user32_timeval */
struct user32_timeval tv;
- bzero(&tv, sizeof (tv));
+ bzero(&tv, sizeof(tv));
tv.tv_sec = d->bd_rtout / hz;
tv.tv_usec = (d->bd_rtout % hz) * tick;
- bcopy(&tv, addr, sizeof (tv));
+ bcopy(&tv, addr, sizeof(tv));
break;
}
- case BIOCGRTIMEOUT64: { /* struct user64_timeval */
+ case BIOCGRTIMEOUT64: { /* struct user64_timeval */
struct user64_timeval tv;
- bzero(&tv, sizeof (tv));
+ bzero(&tv, sizeof(tv));
tv.tv_sec = d->bd_rtout / hz;
tv.tv_usec = (d->bd_rtout % hz) * tick;
- bcopy(&tv, addr, sizeof (tv));
+ bcopy(&tv, addr, sizeof(tv));
break;
}
/*
* Get packet stats.
*/
- case BIOCGSTATS: { /* struct bpf_stat */
+ case BIOCGSTATS: { /* struct bpf_stat */
struct bpf_stat bs;
- bzero(&bs, sizeof (bs));
+ bzero(&bs, sizeof(bs));
bs.bs_recv = d->bd_rcount;
bs.bs_drop = d->bd_dcount;
- bcopy(&bs, addr, sizeof (bs));
+ bcopy(&bs, addr, sizeof(bs));
break;
}
/*
* Set immediate mode.
*/
- case BIOCIMMEDIATE: /* u_int */
+ case BIOCIMMEDIATE: /* u_int */
d->bd_immediate = *(u_int *)(void *)addr;
break;
- case BIOCVERSION: { /* struct bpf_version */
+ case BIOCVERSION: { /* struct bpf_version */
struct bpf_version bv;
- bzero(&bv, sizeof (bv));
+ bzero(&bv, sizeof(bv));
bv.bv_major = BPF_MAJOR_VERSION;
bv.bv_minor = BPF_MINOR_VERSION;
- bcopy(&bv, addr, sizeof (bv));
+ bcopy(&bv, addr, sizeof(bv));
break;
}
/*
* Get "header already complete" flag
*/
- case BIOCGHDRCMPLT: /* u_int */
- bcopy(&d->bd_hdrcmplt, addr, sizeof (u_int));
+ case BIOCGHDRCMPLT: /* u_int */
+ bcopy(&d->bd_hdrcmplt, addr, sizeof(u_int));
break;
/*
* Set "header already complete" flag
*/
- case BIOCSHDRCMPLT: /* u_int */
- bcopy(addr, &int_arg, sizeof (int_arg));
+ case BIOCSHDRCMPLT: /* u_int */
+ bcopy(addr, &int_arg, sizeof(int_arg));
d->bd_hdrcmplt = int_arg ? 1 : 0;
break;
/*
* Get "see sent packets" flag
*/
- case BIOCGSEESENT: /* u_int */
- bcopy(&d->bd_seesent, addr, sizeof (u_int));
+ case BIOCGSEESENT: /* u_int */
+ bcopy(&d->bd_seesent, addr, sizeof(u_int));
break;
/*
* Set "see sent packets" flag
*/
- case BIOCSSEESENT: /* u_int */
- bcopy(addr, &d->bd_seesent, sizeof (u_int));
+ case BIOCSSEESENT: /* u_int */
+ bcopy(addr, &d->bd_seesent, sizeof(u_int));
break;
/*
* Set traffic service class
*/
- case BIOCSETTC: { /* int */
+ case BIOCSETTC: { /* int */
int tc;
- bcopy(addr, &tc, sizeof (int));
+ bcopy(addr, &tc, sizeof(int));
error = bpf_set_traffic_class(d, tc);
break;
}
/*
* Get traffic service class
*/
- case BIOCGETTC: /* int */
- bcopy(&d->bd_traffic_class, addr, sizeof (int));
+ case BIOCGETTC: /* int */
+ bcopy(&d->bd_traffic_class, addr, sizeof(int));
break;
- case FIONBIO: /* Non-blocking I/O; int */
+ case FIONBIO: /* Non-blocking I/O; int */
break;
- case FIOASYNC: /* Send signal on receive packets; int */
- bcopy(addr, &d->bd_async, sizeof (int));
+ case FIOASYNC: /* Send signal on receive packets; int */
+ bcopy(addr, &d->bd_async, sizeof(int));
break;
#ifndef __APPLE__
case FIOSETOWN:
*(int *)addr = -fgetown(d->bd_sigio);
break;
#endif
- case BIOCSRSIG: { /* Set receive signal; u_int */
+ case BIOCSRSIG: { /* Set receive signal; u_int */
u_int sig;
- bcopy(addr, &sig, sizeof (u_int));
+ bcopy(addr, &sig, sizeof(u_int));
- if (sig >= NSIG)
+ if (sig >= NSIG) {
error = EINVAL;
- else
+ } else {
d->bd_sig = sig;
+ }
break;
}
- case BIOCGRSIG: /* u_int */
- bcopy(&d->bd_sig, addr, sizeof (u_int));
+ case BIOCGRSIG: /* u_int */
+ bcopy(&d->bd_sig, addr, sizeof(u_int));
break;
#ifdef __APPLE__
- case BIOCSEXTHDR: /* u_int */
- bcopy(addr, &int_arg, sizeof (int_arg));
- if (int_arg)
+ case BIOCSEXTHDR: /* u_int */
+ bcopy(addr, &int_arg, sizeof(int_arg));
+ if (int_arg) {
d->bd_flags |= BPF_EXTENDED_HDR;
- else
+ } else {
d->bd_flags &= ~BPF_EXTENDED_HDR;
+ }
break;
- case BIOCGIFATTACHCOUNT: { /* struct ifreq */
- ifnet_t ifp;
+ case BIOCGIFATTACHCOUNT: { /* struct ifreq */
+ ifnet_t ifp;
struct bpf_if *bp;
- bcopy(addr, &ifr, sizeof (ifr));
+ bcopy(addr, &ifr, sizeof(ifr));
ifr.ifr_name[IFNAMSIZ - 1] = '\0';
ifp = ifunit(ifr.ifr_name);
if (ifp == NULL) {
for (bp = bpf_iflist; bp != 0; bp = bp->bif_next) {
struct bpf_d *bpf_d;
- if (bp->bif_ifp == NULL || bp->bif_ifp != ifp)
+ if (bp->bif_ifp == NULL || bp->bif_ifp != ifp) {
continue;
+ }
for (bpf_d = bp->bif_dlist; bpf_d;
bpf_d = bpf_d->bd_next) {
ifr.ifr_intval += 1;
}
}
- bcopy(&ifr, addr, sizeof (ifr));
+ bcopy(&ifr, addr, sizeof(ifr));
break;
}
- case BIOCGWANTPKTAP: /* u_int */
+ case BIOCGWANTPKTAP: /* u_int */
int_arg = d->bd_flags & BPF_WANT_PKTAP ? 1 : 0;
- bcopy(&int_arg, addr, sizeof (int_arg));
+ bcopy(&int_arg, addr, sizeof(int_arg));
break;
- case BIOCSWANTPKTAP: /* u_int */
- bcopy(addr, &int_arg, sizeof (int_arg));
- if (int_arg)
+ case BIOCSWANTPKTAP: /* u_int */
+ bcopy(addr, &int_arg, sizeof(int_arg));
+ if (int_arg) {
d->bd_flags |= BPF_WANT_PKTAP;
- else
+ } else {
d->bd_flags &= ~BPF_WANT_PKTAP;
+ }
break;
#endif
case BIOCSHEADDROP:
- bcopy(addr, &int_arg, sizeof (int_arg));
+ bcopy(addr, &int_arg, sizeof(int_arg));
d->bd_headdrop = int_arg ? 1 : 0;
break;
case BIOCGHEADDROP:
- bcopy(&d->bd_headdrop, addr, sizeof (int));
+ bcopy(&d->bd_headdrop, addr, sizeof(int));
break;
case BIOCSTRUNCATE:
bcopy(addr, &int_arg, sizeof(int_arg));
- if (int_arg)
- d->bd_flags |= BPF_TRUNCATE;
- else
+ if (int_arg) {
+ d->bd_flags |= BPF_TRUNCATE;
+ } else {
d->bd_flags &= ~BPF_TRUNCATE;
+ }
break;
case BIOCGETUUID:
- bcopy(&d->bd_uuid, addr, sizeof (uuid_t));
+ bcopy(&d->bd_uuid, addr, sizeof(uuid_t));
break;
case BIOCSETUP: {
struct bpf_setup_args bsa;
- ifnet_t ifp;
+ ifnet_t ifp;
- bcopy(addr, &bsa, sizeof (struct bpf_setup_args));
+ bcopy(addr, &bsa, sizeof(struct bpf_setup_args));
bsa.bsa_ifname[IFNAMSIZ - 1] = 0;
ifp = ifunit(bsa.bsa_ifname);
if (ifp == NULL) {
"%s: ifnet not found for %s error %d",
__func__, bsa.bsa_ifname, error);
break;
- }
+ }
error = bpf_setup(d, bsa.bsa_uuid, ifp);
break;
}
case BIOCSPKTHDRV2:
bcopy(addr, &int_arg, sizeof(int_arg));
- if (int_arg != 0)
+ if (int_arg != 0) {
d->bd_flags |= BPF_PKTHDRV2;
- else
+ } else {
d->bd_flags &= ~BPF_PKTHDRV2;
+ }
break;
case BIOCGPKTHDRV2:
int_arg = d->bd_flags & BPF_PKTHDRV2 ? 1 : 0;
- bcopy(&int_arg, addr, sizeof (int));
+ bcopy(&int_arg, addr, sizeof(int));
break;
}
bpf_release_d(d);
lck_mtx_unlock(bpf_mlock);
- return (error);
+ return error;
}
/*
struct bpf_insn *fcode, *old;
u_int flen, size;
- while (d->bd_hbuf_read != 0)
+ while (d->bd_hbuf_read != 0) {
msleep((caddr_t)d, bpf_mlock, PRINET, "bpf_reading", NULL);
+ }
- if ((d->bd_flags & BPF_CLOSING) != 0)
- return (ENXIO);
+ if ((d->bd_flags & BPF_CLOSING) != 0) {
+ return ENXIO;
+ }
old = d->bd_filter;
if (bf_insns == USER_ADDR_NULL) {
- if (bf_len != 0)
- return (EINVAL);
+ if (bf_len != 0) {
+ return EINVAL;
+ }
d->bd_filter = NULL;
reset_d(d);
- if (old != 0)
- FREE((caddr_t)old, M_DEVBUF);
- return (0);
+ if (old != 0) {
+ FREE(old, M_DEVBUF);
+ }
+ return 0;
}
flen = bf_len;
- if (flen > BPF_MAXINSNS)
- return (EINVAL);
+ if (flen > BPF_MAXINSNS) {
+ return EINVAL;
+ }
size = flen * sizeof(struct bpf_insn);
fcode = (struct bpf_insn *) _MALLOC(size, M_DEVBUF, M_WAIT);
#ifdef __APPLE__
- if (fcode == NULL)
- return (ENOBUFS);
+ if (fcode == NULL) {
+ return ENOBUFS;
+ }
#endif
if (copyin(bf_insns, (caddr_t)fcode, size) == 0 &&
bpf_validate(fcode, (int)flen)) {
d->bd_filter = fcode;
- if (cmd == BIOCSETF32 || cmd == BIOCSETF64)
+ if (cmd == BIOCSETF32 || cmd == BIOCSETF64) {
reset_d(d);
+ }
- if (old != 0)
- FREE((caddr_t)old, M_DEVBUF);
+ if (old != 0) {
+ FREE(old, M_DEVBUF);
+ }
- return (0);
+ return 0;
}
- FREE((caddr_t)fcode, M_DEVBUF);
- return (EINVAL);
+ FREE(fcode, M_DEVBUF);
+ return EINVAL;
}
/*
struct bpf_if *bp;
int error;
- while (d->bd_hbuf_read != 0 && !has_hbuf_read)
+ while (d->bd_hbuf_read != 0 && !has_hbuf_read) {
msleep((caddr_t)d, bpf_mlock, PRINET, "bpf_reading", NULL);
+ }
- if ((d->bd_flags & BPF_CLOSING) != 0)
- return (ENXIO);
+ if ((d->bd_flags & BPF_CLOSING) != 0) {
+ return ENXIO;
+ }
/*
* Look through attached interfaces for the named one.
for (bp = bpf_iflist; bp != 0; bp = bp->bif_next) {
struct ifnet *ifp = bp->bif_ifp;
- if (ifp == 0 || ifp != theywant)
+ if (ifp == 0 || ifp != theywant) {
continue;
+ }
/*
* Do not use DLT_PKTAP, unless requested explicitly
*/
- if (bp->bif_dlt == DLT_PKTAP && !(d->bd_flags & BPF_WANT_PKTAP))
+ if (bp->bif_dlt == DLT_PKTAP && !(d->bd_flags & BPF_WANT_PKTAP)) {
continue;
+ }
/*
* Skip the coprocessor interface
*/
- if (!intcoproc_unrestricted && IFNET_IS_INTCOPROC(ifp))
+ if (!intcoproc_unrestricted && IFNET_IS_INTCOPROC(ifp)) {
continue;
+ }
/*
* We found the requested interface.
* Allocate the packet buffers.
*/
error = bpf_allocbufs(d);
- if (error != 0)
- return (error);
+ if (error != 0) {
+ return error;
+ }
/*
* Detach if attached to something else.
*/
if (bp != d->bd_bif) {
if (d->bd_bif != NULL) {
- if (bpf_detachd(d, 0) != 0)
- return (ENXIO);
+ if (bpf_detachd(d, 0) != 0) {
+ return ENXIO;
+ }
+ }
+ if (bpf_attachd(d, bp) != 0) {
+ return ENXIO;
}
- if (bpf_attachd(d, bp) != 0)
- return (ENXIO);
}
if (do_reset) {
- reset_d(d);
+ reset_d(d);
}
- return (0);
+ return 0;
}
/* Not found. */
- return (ENXIO);
+ return ENXIO;
}
/*
static int
bpf_getdltlist(struct bpf_d *d, caddr_t addr, struct proc *p)
{
- u_int n;
- int error;
- struct ifnet *ifp;
- struct bpf_if *bp;
- user_addr_t dlist;
+ u_int n;
+ int error;
+ struct ifnet *ifp;
+ struct bpf_if *bp;
+ user_addr_t dlist;
struct bpf_dltlist bfl;
- bcopy(addr, &bfl, sizeof (bfl));
+ bcopy(addr, &bfl, sizeof(bfl));
if (proc_is64bit(p)) {
dlist = (user_addr_t)bfl.bfl_u.bflu_pad;
} else {
error = 0;
for (bp = bpf_iflist; bp; bp = bp->bif_next) {
- if (bp->bif_ifp != ifp)
+ if (bp->bif_ifp != ifp) {
continue;
+ }
/*
* Do not use DLT_PKTAP, unless requested explicitly
*/
- if (bp->bif_dlt == DLT_PKTAP && !(d->bd_flags & BPF_WANT_PKTAP))
+ if (bp->bif_dlt == DLT_PKTAP && !(d->bd_flags & BPF_WANT_PKTAP)) {
continue;
+ }
if (dlist != USER_ADDR_NULL) {
if (n >= bfl.bfl_len) {
- return (ENOMEM);
+ return ENOMEM;
}
error = copyout(&bp->bif_dlt, dlist,
- sizeof (bp->bif_dlt));
- if (error != 0)
+ sizeof(bp->bif_dlt));
+ if (error != 0) {
break;
- dlist += sizeof (bp->bif_dlt);
+ }
+ dlist += sizeof(bp->bif_dlt);
}
n++;
}
bfl.bfl_len = n;
- bcopy(&bfl, addr, sizeof (bfl));
+ bcopy(&bfl, addr, sizeof(bfl));
- return (error);
+ return error;
}
/*
struct ifnet *ifp;
struct bpf_if *bp;
- if (d->bd_bif->bif_dlt == dlt)
- return (0);
+ if (d->bd_bif->bif_dlt == dlt) {
+ return 0;
+ }
- while (d->bd_hbuf_read != 0)
+ while (d->bd_hbuf_read != 0) {
msleep((caddr_t)d, bpf_mlock, PRINET, "bpf_reading", NULL);
+ }
- if ((d->bd_flags & BPF_CLOSING) != 0)
- return (ENXIO);
+ if ((d->bd_flags & BPF_CLOSING) != 0) {
+ return ENXIO;
+ }
ifp = d->bd_bif->bif_ifp;
for (bp = bpf_iflist; bp; bp = bp->bif_next) {
}
if (bp != NULL) {
opromisc = d->bd_promisc;
- if (bpf_detachd(d, 0) != 0)
- return (ENXIO);
+ if (bpf_detachd(d, 0) != 0) {
+ return ENXIO;
+ }
error = bpf_attachd(d, bp);
if (error) {
printf("bpf_setdlt: bpf_attachd %s%d failed (%d)\n",
ifnet_name(bp->bif_ifp), ifnet_unit(bp->bif_ifp),
error);
- return (error);
+ return error;
}
reset_d(d);
if (opromisc) {
}
}
}
- return (bp == NULL ? EINVAL : 0);
+ return bp == NULL ? EINVAL : 0;
}
static int
{
int error = 0;
- if (!SO_VALID_TC(tc))
+ if (!SO_VALID_TC(tc)) {
error = EINVAL;
- else
+ } else {
d->bd_traffic_class = tc;
+ }
- return (error);
+ return error;
}
static void
bpf_set_packet_service_class(struct mbuf *m, int tc)
{
- if (!(m->m_flags & M_PKTHDR))
+ if (!(m->m_flags & M_PKTHDR)) {
return;
+ }
VERIFY(SO_VALID_TC(tc));
(void) m_set_service_class(m, so_tc2msc(tc));
if (d == NULL || d == BPF_DEV_RESERVED ||
(d->bd_flags & BPF_CLOSING) != 0) {
lck_mtx_unlock(bpf_mlock);
- return (ENXIO);
+ return ENXIO;
}
bpf_acquire_d(d);
if (d->bd_bif == NULL) {
bpf_release_d(d);
lck_mtx_unlock(bpf_mlock);
- return (ENXIO);
+ return ENXIO;
}
- while (d->bd_hbuf_read != 0)
+ while (d->bd_hbuf_read != 0) {
msleep((caddr_t)d, bpf_mlock, PRINET, "bpf_reading", NULL);
+ }
if ((d->bd_flags & BPF_CLOSING) != 0) {
bpf_release_d(d);
lck_mtx_unlock(bpf_mlock);
- return (ENXIO);
+ return ENXIO;
}
switch (which) {
- case FREAD:
- if (d->bd_hlen != 0 ||
- ((d->bd_immediate ||
- d->bd_state == BPF_TIMED_OUT) && d->bd_slen != 0))
- ret = 1; /* read has data to return */
- else {
- /*
- * Read has no data to return.
- * Make the select wait, and start a timer if
- * necessary.
- */
- selrecord(p, &d->bd_sel, wql);
- bpf_start_timer(d);
- }
- break;
+ case FREAD:
+ if (d->bd_hlen != 0 ||
+ ((d->bd_immediate ||
+ d->bd_state == BPF_TIMED_OUT) && d->bd_slen != 0)) {
+ ret = 1; /* read has data to return */
+ } else {
+ /*
+ * Read has no data to return.
+ * Make the select wait, and start a timer if
+ * necessary.
+ */
+ selrecord(p, &d->bd_sel, wql);
+ bpf_start_timer(d);
+ }
+ break;
- case FWRITE:
- /* can't determine whether a write would block */
- ret = 1;
- break;
+ case FWRITE:
+ /* can't determine whether a write would block */
+ ret = 1;
+ break;
}
bpf_release_d(d);
lck_mtx_unlock(bpf_mlock);
- return (ret);
+ return ret;
}
/*
int bpfkqfilter(dev_t dev, struct knote *kn);
static void filt_bpfdetach(struct knote *);
static int filt_bpfread(struct knote *, long);
-static int filt_bpftouch(struct knote *kn, struct kevent_internal_s *kev);
-static int filt_bpfprocess(struct knote *kn, struct filt_process_s *data,
- struct kevent_internal_s *kev);
+static int filt_bpftouch(struct knote *kn, struct kevent_qos_s *kev);
+static int filt_bpfprocess(struct knote *kn, struct kevent_qos_s *kev);
SECURITY_READ_ONLY_EARLY(struct filterops) bpfread_filtops = {
.f_isfd = 1,
};
static int
-filt_bpfread_common(struct knote *kn, struct bpf_d *d)
+filt_bpfread_common(struct knote *kn, struct kevent_qos_s *kev, struct bpf_d *d)
{
int ready = 0;
+ int64_t data = 0;
if (d->bd_immediate) {
/*
* If there's no data in either buffer, we're not
* ready to read.
*/
- kn->kn_data = (d->bd_hlen == 0 || d->bd_hbuf_read != 0 ?
+ data = (d->bd_hlen == 0 || d->bd_hbuf_read != 0 ?
d->bd_slen : d->bd_hlen);
- int64_t lowwat = 1;
- if (kn->kn_sfflags & NOTE_LOWAT) {
- if (kn->kn_sdata > d->bd_bufsize)
- lowwat = d->bd_bufsize;
- else if (kn->kn_sdata > lowwat)
- lowwat = kn->kn_sdata;
- }
- ready = (kn->kn_data >= lowwat);
+ int64_t lowwat = knote_low_watermark(kn);
+ if (lowwat > d->bd_bufsize) {
+ lowwat = d->bd_bufsize;
+ }
+ ready = (data >= lowwat);
} else {
/*
* If there's data in the hold buffer, it's the
* no data in the hold buffer and the timer hasn't
* expired, we're not ready to read.
*/
- kn->kn_data = ((d->bd_hlen == 0 || d->bd_hbuf_read != 0) &&
+ data = ((d->bd_hlen == 0 || d->bd_hbuf_read != 0) &&
d->bd_state == BPF_TIMED_OUT ? d->bd_slen : d->bd_hlen);
- ready = (kn->kn_data > 0);
+ ready = (data > 0);
}
- if (!ready)
+ if (!ready) {
bpf_start_timer(d);
+ } else if (kev) {
+ knote_fill_kevent(kn, kev, data);
+ }
- return (ready);
+ return ready;
}
int
/*
* Is this device a bpf?
*/
- if (major(dev) != CDEV_MAJOR ||
- kn->kn_filter != EVFILT_READ) {
- kn->kn_flags = EV_ERROR;
- kn->kn_data = EINVAL;
- return (0);
+ if (major(dev) != CDEV_MAJOR || kn->kn_filter != EVFILT_READ) {
+ knote_set_error(kn, EINVAL);
+ return 0;
}
lck_mtx_lock(bpf_mlock);
(d->bd_flags & BPF_CLOSING) != 0 ||
d->bd_bif == NULL) {
lck_mtx_unlock(bpf_mlock);
- kn->kn_flags = EV_ERROR;
- kn->kn_data = ENXIO;
- return (0);
+ knote_set_error(kn, ENXIO);
+ return 0;
}
kn->kn_hook = d;
d->bd_flags |= BPF_KNOTE;
/* capture the current state */
- res = filt_bpfread_common(kn, d);
+ res = filt_bpfread_common(kn, NULL, d);
lck_mtx_unlock(bpf_mlock);
- return (res);
+ return res;
}
static void
#pragma unused(hint)
struct bpf_d *d = (struct bpf_d *)kn->kn_hook;
- return (filt_bpfread_common(kn, d));
+ return filt_bpfread_common(kn, NULL, d);
}
static int
-filt_bpftouch(struct knote *kn, struct kevent_internal_s *kev)
+filt_bpftouch(struct knote *kn, struct kevent_qos_s *kev)
{
struct bpf_d *d = (struct bpf_d *)kn->kn_hook;
int res;
kn->kn_sfflags = kev->fflags;
/* output data will be re-generated here */
- res = filt_bpfread_common(kn, d);
+ res = filt_bpfread_common(kn, NULL, d);
lck_mtx_unlock(bpf_mlock);
- return (res);
+ return res;
}
static int
-filt_bpfprocess(struct knote *kn, struct filt_process_s *data,
- struct kevent_internal_s *kev)
+filt_bpfprocess(struct knote *kn, struct kevent_qos_s *kev)
{
-#pragma unused(data)
struct bpf_d *d = (struct bpf_d *)kn->kn_hook;
int res;
lck_mtx_lock(bpf_mlock);
- res = filt_bpfread_common(kn, d);
- if (res) {
- *kev = kn->kn_kevent;
- }
+ res = filt_bpfread_common(kn, kev, d);
lck_mtx_unlock(bpf_mlock);
- return (res);
+ return res;
}
/*
dst = dst_arg;
while (len > 0) {
- if (m == 0)
+ if (m == 0) {
panic("bpf_mcopy");
+ }
count = min(m->m_len, len);
bcopy(mbuf_data(m), dst, count);
m = m->m_next;
static inline void
bpf_tap_imp(
- ifnet_t ifp,
- u_int32_t dlt,
+ ifnet_t ifp,
+ u_int32_t dlt,
struct bpf_packet *bpf_pkt,
- int outbound)
+ int outbound)
{
- struct bpf_d *d;
+ struct bpf_d *d;
u_int slen;
struct bpf_if *bp;
struct bpf_packet bpf_pkt_tmp;
struct pktap_header_buffer bpfp_header_tmp;
- if (outbound && !d->bd_seesent)
+ if (outbound && !d->bd_seesent) {
continue;
+ }
++d->bd_rcount;
slen = bpf_filter(d->bd_filter, (u_char *)bpf_pkt,
!!(d->bd_flags & BPF_TRUNCATE));
}
- if (d->bd_flags & BPF_TRUNCATE) {
+ if (d->bd_flags & BPF_TRUNCATE) {
slen = min(slen,
get_pkt_trunc_len((u_char *)bpf_pkt,
- bpf_pkt->bpfp_total_length));
- }
+ bpf_pkt->bpfp_total_length));
+ }
}
if (slen != 0) {
-#if CONFIG_MACF_NET
- if (mac_bpfdesc_check_receive(d, bp->bif_ifp) != 0)
- continue;
-#endif
catchpacket(d, bpf_pkt, slen, outbound);
}
bpf_pkt = bpf_pkt_saved;
static inline void
bpf_tap_mbuf(
- ifnet_t ifp,
- u_int32_t dlt,
- mbuf_t m,
- void* hdr,
- size_t hlen,
- int outbound)
+ ifnet_t ifp,
+ u_int32_t dlt,
+ mbuf_t m,
+ void* hdr,
+ size_t hlen,
+ int outbound)
{
struct bpf_packet bpf_pkt;
struct mbuf *m0;
bpf_pkt.bpfp_type = BPF_PACKET_TYPE_MBUF;
bpf_pkt.bpfp_mbuf = m;
bpf_pkt.bpfp_total_length = 0;
- for (m0 = m; m0 != NULL; m0 = m0->m_next)
+ for (m0 = m; m0 != NULL; m0 = m0->m_next) {
bpf_pkt.bpfp_total_length += m0->m_len;
+ }
bpf_pkt.bpfp_header = hdr;
if (hdr != NULL) {
bpf_pkt.bpfp_total_length += hlen;
void
bpf_tap_out(
- ifnet_t ifp,
- u_int32_t dlt,
- mbuf_t m,
- void* hdr,
- size_t hlen)
+ ifnet_t ifp,
+ u_int32_t dlt,
+ mbuf_t m,
+ void* hdr,
+ size_t hlen)
{
bpf_tap_mbuf(ifp, dlt, m, hdr, hlen, 1);
}
void
bpf_tap_in(
- ifnet_t ifp,
- u_int32_t dlt,
- mbuf_t m,
- void* hdr,
- size_t hlen)
+ ifnet_t ifp,
+ u_int32_t dlt,
+ mbuf_t m,
+ void* hdr,
+ size_t hlen)
{
bpf_tap_mbuf(ifp, dlt, m, hdr, hlen, 0);
}
/* Callback registered with Ethernet driver. */
-static int bpf_tap_callback(struct ifnet *ifp, struct mbuf *m)
+static int
+bpf_tap_callback(struct ifnet *ifp, struct mbuf *m)
{
bpf_tap_mbuf(ifp, 0, m, NULL, 0, mbuf_pkthdr_rcvif(m) == NULL);
- return (0);
+ return 0;
}
err = EINVAL;
}
- return (err);
+ return err;
}
static void
{
/* copy the optional header */
if (pkt->bpfp_header_length != 0) {
- size_t count = min(len, pkt->bpfp_header_length);
+ size_t count = min(len, pkt->bpfp_header_length);
bcopy(pkt->bpfp_header, dst, count);
len -= count;
dst += count;
*/
uint16_t trunc_len = ESP_HDR_SIZE + 1;
- if (trunc_len > remaining_caplen)
- return (remaining_caplen);
+ if (trunc_len > remaining_caplen) {
+ return remaining_caplen;
+ }
- return (trunc_len);
+ return trunc_len;
}
static uint16_t
*/
uint16_t trunc_len = ISAKMP_HDR_SIZE;
- if (trunc_len > remaining_caplen)
- return (remaining_caplen);
+ if (trunc_len > remaining_caplen) {
+ return remaining_caplen;
+ }
- return (trunc_len);
+ return trunc_len;
}
static uint16_t
char payload[remaining_caplen];
err = bpf_copydata(pkt, off, remaining_caplen, payload);
- if (err != 0)
- return (remaining_caplen);
+ if (err != 0) {
+ return remaining_caplen;
+ }
/*
* They are three cases:
* - IKE: payload start with 4 bytes header set to zero before ISAKMP header
* - otherwise it's ESP
*/
if (remaining_caplen >= 4 &&
- payload[0] == 0 && payload[1] == 0 &&
- payload[2] == 0 && payload[3] == 0) {
+ payload[0] == 0 && payload[1] == 0 &&
+ payload[2] == 0 && payload[3] == 0) {
trunc_len = 4 + get_isakmp_trunc_len(pkt, off + 4, remaining_caplen - 4);
} else if (remaining_caplen == 1) {
trunc_len = 1;
trunc_len = get_esp_trunc_len(pkt, off, remaining_caplen);
}
- if (trunc_len > remaining_caplen)
- return (remaining_caplen);
-
- return (trunc_len);
+ if (trunc_len > remaining_caplen) {
+ return remaining_caplen;
+ }
+ return trunc_len;
}
static uint16_t
int err = 0;
uint16_t trunc_len = sizeof(struct udphdr); /* By default no UDP payload */
- if (trunc_len >= remaining_caplen)
- return (remaining_caplen);
+ if (trunc_len >= remaining_caplen) {
+ return remaining_caplen;
+ }
struct udphdr udphdr;
err = bpf_copydata(pkt, off, sizeof(struct udphdr), &udphdr);
- if (err != 0)
- return (remaining_caplen);
+ if (err != 0) {
+ return remaining_caplen;
+ }
u_short sport, dport;
*/
trunc_len = remaining_caplen;
} else if ((sport == PORT_BOOTPS && dport == PORT_BOOTPC) ||
- (sport == PORT_BOOTPC && dport == PORT_BOOTPS)) {
+ (sport == PORT_BOOTPC && dport == PORT_BOOTPS)) {
/*
* Full UDP payload for BOOTP and DHCP
*/
trunc_len += get_isakmp_natt_trunc_len(pkt, off + sizeof(struct udphdr),
remaining_caplen - sizeof(struct udphdr));
}
- if (trunc_len >= remaining_caplen)
- return (remaining_caplen);
+ if (trunc_len >= remaining_caplen) {
+ return remaining_caplen;
+ }
- return (trunc_len);
+ return trunc_len;
}
static uint16_t
{
int err = 0;
uint16_t trunc_len = sizeof(struct tcphdr); /* By default no TCP payload */
- if (trunc_len >= remaining_caplen)
- return (remaining_caplen);
+ if (trunc_len >= remaining_caplen) {
+ return remaining_caplen;
+ }
struct tcphdr tcphdr;
err = bpf_copydata(pkt, off, sizeof(struct tcphdr), &tcphdr);
- if (err != 0)
- return (remaining_caplen);
+ if (err != 0) {
+ return remaining_caplen;
+ }
u_short sport, dport;
sport = EXTRACT_SHORT(&tcphdr.th_sport);
} else {
trunc_len = tcphdr.th_off << 2;
}
- if (trunc_len >= remaining_caplen)
- return (remaining_caplen);
+ if (trunc_len >= remaining_caplen) {
+ return remaining_caplen;
+ }
- return (trunc_len);
+ return trunc_len;
}
static uint16_t
break;
}
}
- if (trunc_len >= remaining_caplen)
- return (remaining_caplen);
+ if (trunc_len >= remaining_caplen) {
+ return remaining_caplen;
+ }
- return (trunc_len);
+ return trunc_len;
}
static uint16_t
{
int err = 0;
uint16_t iplen = sizeof(struct ip);
- if (iplen >= remaining_caplen)
- return (remaining_caplen);
+ if (iplen >= remaining_caplen) {
+ return remaining_caplen;
+ }
struct ip iphdr;
err = bpf_copydata(pkt, off, sizeof(struct ip), &iphdr);
- if (err != 0)
- return (remaining_caplen);
+ if (err != 0) {
+ return remaining_caplen;
+ }
uint8_t proto = 0;
iplen = iphdr.ip_hl << 2;
- if (iplen >= remaining_caplen)
- return (remaining_caplen);
+ if (iplen >= remaining_caplen) {
+ return remaining_caplen;
+ }
proto = iphdr.ip_p;
iplen += get_proto_trunc_len(proto, pkt, off + iplen, remaining_caplen - iplen);
- if (iplen >= remaining_caplen)
- return (remaining_caplen);
+ if (iplen >= remaining_caplen) {
+ return remaining_caplen;
+ }
- return (iplen);
+ return iplen;
}
static uint16_t
{
int err = 0;
uint16_t iplen = sizeof(struct ip6_hdr);
- if (iplen >= remaining_caplen)
- return (remaining_caplen);
+ if (iplen >= remaining_caplen) {
+ return remaining_caplen;
+ }
struct ip6_hdr ip6hdr;
err = bpf_copydata(pkt, off, sizeof(struct ip6_hdr), &ip6hdr);
- if (err != 0)
- return (remaining_caplen);
+ if (err != 0) {
+ return remaining_caplen;
+ }
uint8_t proto = 0;
proto = ip6hdr.ip6_nxt;
iplen += get_proto_trunc_len(proto, pkt, off + iplen, remaining_caplen - iplen);
- if (iplen >= remaining_caplen)
- return (remaining_caplen);
+ if (iplen >= remaining_caplen) {
+ return remaining_caplen;
+ }
- return (iplen);
+ return iplen;
}
static uint16_t
{
int err = 0;
uint16_t ethlen = sizeof(struct ether_header);
- if (ethlen >= remaining_caplen)
- return (remaining_caplen);
+ if (ethlen >= remaining_caplen) {
+ return remaining_caplen;
+ }
struct ether_header eh;
u_short type;
err = bpf_copydata(pkt, off, sizeof(struct ether_header), &eh);
- if (err != 0)
- return (remaining_caplen);
+ if (err != 0) {
+ return remaining_caplen;
+ }
type = EXTRACT_SHORT(&eh.ether_type);
/* Include full ARP */
remaining_caplen);
} else if (type == ETHERTYPE_IPV6) {
ethlen += get_ip6_trunc_len(pkt, sizeof(struct ether_header),
- remaining_caplen);
+ remaining_caplen);
}
}
- return (ethlen);
+ return ethlen;
}
static uint32_t
* pkt->bpfp_header_length is (pktap->pth_length + pre_adjust)
* pre is the offset to the L3 header after the bpfp_header, or length
* of L2 header after bpfp_header, if present.
- */
- uint32_t pre = pktap->pth_frame_pre_length -
+ */
+ int32_t pre = pktap->pth_frame_pre_length -
(pkt->bpfp_header_length - pktap->pth_length);
/* Length of the input packet starting from L3 header */
pktap->pth_protocol_family == AF_INET6) {
/* Contains L2 header */
if (pre > 0) {
- if (pre < sizeof(struct ether_header))
+ if (pre < (int32_t)sizeof(struct ether_header)) {
goto too_short;
+ }
out_pkt_len = get_ether_trunc_len(pkt, 0, in_pkt_len);
} else if (pre == 0) {
}
done:
tlen = pkt->bpfp_header_length + out_pkt_len + pre;
- return (tlen);
+ return tlen;
too_short:
out_pkt_len = in_pkt_len;
goto done;
*/
static void
catchpacket(struct bpf_d *d, struct bpf_packet * pkt,
- u_int snaplen, int outbound)
+ u_int snaplen, int outbound)
{
struct bpf_hdr *hp;
struct bpf_hdr_ext *ehp;
* we hit the buffer size limit).
*/
totlen = hdrlen + min(snaplen, pkt->bpfp_total_length);
- if (totlen > d->bd_bufsize)
+ if (totlen > d->bd_bufsize) {
totlen = d->bd_bufsize;
+ }
- if (hdrlen > totlen)
+ if (hdrlen > totlen) {
return;
+ }
/*
* Round up the end of the previous packet to the next longword.
}
do_wakeup = 1;
curlen = 0;
- } else if (d->bd_immediate || d->bd_state == BPF_TIMED_OUT)
+ } else if (d->bd_immediate || d->bd_state == BPF_TIMED_OUT) {
/*
* Immediate mode is set, or the read timeout has
* already expired during a select call. A packet
* arrived, so the reader should be woken up.
*/
do_wakeup = 1;
+ }
/*
* Append the bpf header.
struct mbuf *m;
m = (pkt->bpfp_type == BPF_PACKET_TYPE_MBUF)
- ? pkt->bpfp_mbuf : NULL;
+ ? pkt->bpfp_mbuf : NULL;
ehp = (struct bpf_hdr_ext *)(void *)(d->bd_sbuf + curlen);
memset(ehp, 0, sizeof(*ehp));
ehp->bh_tstamp.tv_sec = tv.tv_sec;
ehp->bh_flags |= BPF_HDR_EXT_FLAGS_DIR_OUT;
/* only do lookups on non-raw INPCB */
- if ((m->m_pkthdr.pkt_flags & (PKTF_FLOW_ID|
- PKTF_FLOW_LOCALSRC|PKTF_FLOW_RAWSOCK)) ==
- (PKTF_FLOW_ID|PKTF_FLOW_LOCALSRC) &&
+ if ((m->m_pkthdr.pkt_flags & (PKTF_FLOW_ID |
+ PKTF_FLOW_LOCALSRC | PKTF_FLOW_RAWSOCK)) ==
+ (PKTF_FLOW_ID | PKTF_FLOW_LOCALSRC) &&
m->m_pkthdr.pkt_flowsrc == FLOWSRC_INPCB) {
ehp->bh_flowid = m->m_pkthdr.pkt_flowid;
ehp->bh_proto = m->m_pkthdr.pkt_proto;
}
ehp->bh_svc = so_svc2tc(m->m_pkthdr.pkt_svc);
- if (m->m_pkthdr.pkt_flags & PKTF_TCP_REXMT)
+ if (m->m_pkthdr.pkt_flags & PKTF_TCP_REXMT) {
ehp->bh_pktflags |= BPF_PKTFLAGS_TCP_REXMT;
- if (m->m_pkthdr.pkt_flags & PKTF_START_SEQ)
+ }
+ if (m->m_pkthdr.pkt_flags & PKTF_START_SEQ) {
ehp->bh_pktflags |= BPF_PKTFLAGS_START_SEQ;
- if (m->m_pkthdr.pkt_flags & PKTF_LAST_PKT)
+ }
+ if (m->m_pkthdr.pkt_flags & PKTF_LAST_PKT) {
ehp->bh_pktflags |= BPF_PKTFLAGS_LAST_PKT;
+ }
if (m->m_pkthdr.pkt_flags & PKTF_VALID_UNSENT_DATA) {
ehp->bh_unsent_bytes =
m->m_pkthdr.bufstatus_if;
ehp->bh_unsent_snd =
m->m_pkthdr.bufstatus_sndbuf;
}
- } else
+ } else {
ehp->bh_flags |= BPF_HDR_EXT_FLAGS_DIR_IN;
+ }
payload = (u_char *)ehp + hdrlen;
} else {
hp = (struct bpf_hdr *)(void *)(d->bd_sbuf + curlen);
d->bd_slen = curlen + totlen;
d->bd_scnt += 1;
- if (do_wakeup)
+ if (do_wakeup) {
bpf_wakeup(d);
+ }
}
/*
}
d->bd_fbuf = (caddr_t) _MALLOC(d->bd_bufsize, M_DEVBUF, M_WAIT);
- if (d->bd_fbuf == NULL)
- return (ENOBUFS);
+ if (d->bd_fbuf == NULL) {
+ return ENOBUFS;
+ }
d->bd_sbuf = (caddr_t) _MALLOC(d->bd_bufsize, M_DEVBUF, M_WAIT);
if (d->bd_sbuf == NULL) {
FREE(d->bd_fbuf, M_DEVBUF);
d->bd_fbuf = NULL;
- return (ENOBUFS);
+ return ENOBUFS;
}
d->bd_slen = 0;
d->bd_hlen = 0;
d->bd_scnt = 0;
d->bd_hcnt = 0;
- return (0);
+ return 0;
}
/*
* been detached from its interface and it yet hasn't been marked
* free.
*/
- if (d->bd_hbuf_read != 0)
+ if (d->bd_hbuf_read != 0) {
panic("bpf buffer freed during read");
+ }
if (d->bd_sbuf != 0) {
FREE(d->bd_sbuf, M_DEVBUF);
- if (d->bd_hbuf != 0)
+ if (d->bd_hbuf != 0) {
FREE(d->bd_hbuf, M_DEVBUF);
- if (d->bd_fbuf != 0)
+ }
+ if (d->bd_fbuf != 0) {
FREE(d->bd_fbuf, M_DEVBUF);
+ }
+ }
+ if (d->bd_filter) {
+ FREE(d->bd_filter, M_DEVBUF);
}
- if (d->bd_filter)
- FREE((caddr_t)d->bd_filter, M_DEVBUF);
}
/*
errno_t
bpf_attach(
- ifnet_t ifp,
- u_int32_t dlt,
- u_int32_t hdrlen,
- bpf_send_func send,
- bpf_tap_func tap)
+ ifnet_t ifp,
+ u_int32_t dlt,
+ u_int32_t hdrlen,
+ bpf_send_func send,
+ bpf_tap_func tap)
{
struct bpf_if *bp;
struct bpf_if *bp_new;
bp_new = (struct bpf_if *) _MALLOC(sizeof(*bp_new), M_DEVBUF,
M_WAIT | M_ZERO);
- if (bp_new == 0)
+ if (bp_new == 0) {
panic("bpfattach");
+ }
lck_mtx_lock(bpf_mlock);
if (found) {
lck_mtx_unlock(bpf_mlock);
printf("bpfattach - %s with dlt %d is already attached\n",
- if_name(ifp), dlt);
+ if_name(ifp), dlt);
FREE(bp_new, M_DEVBUF);
- return (EEXIST);
+ return EEXIST;
}
bp_new->bif_ifp = ifp;
lck_mtx_unlock(bpf_mlock);
#ifndef __APPLE__
- if (bootverbose)
+ if (bootverbose) {
printf("bpf: %s attached\n", if_name(ifp));
+ }
#endif
- return (0);
+ return 0;
}
/*
void
bpfdetach(struct ifnet *ifp)
{
- struct bpf_if *bp, *bp_prev, *bp_next;
- struct bpf_d *d;
+ struct bpf_if *bp, *bp_prev, *bp_next;
+ struct bpf_d *d;
- if (bpf_debug != 0)
+ if (bpf_debug != 0) {
printf("%s: %s\n", __func__, if_name(ifp));
+ }
lck_mtx_lock(bpf_mlock);
continue;
}
/* Unlink from the interface list */
- if (bp_prev)
+ if (bp_prev) {
bp_prev->bif_next = bp->bif_next;
- else
+ } else {
bpf_iflist = bp->bif_next;
+ }
/* Detach the devices attached to the interface */
while ((d = bp->bif_dlist) != NULL) {
bpf_init(__unused void *unused)
{
#ifdef __APPLE__
- int i;
- int maj;
+ int i;
+ int maj;
if (bpf_devsw_installed == 0) {
bpf_devsw_installed = 1;
lck_mtx_init(bpf_mlock, bpf_mlock_grp, bpf_mlock_attr);
maj = cdevsw_add(CDEV_MAJOR, &bpf_cdevsw);
if (maj == -1) {
- if (bpf_mlock_attr)
+ if (bpf_mlock_attr) {
lck_attr_free(bpf_mlock_attr);
- if (bpf_mlock_grp)
+ }
+ if (bpf_mlock_grp) {
lck_grp_free(bpf_mlock_grp);
- if (bpf_mlock_grp_attr)
+ }
+ if (bpf_mlock_grp_attr) {
lck_grp_attr_free(bpf_mlock_grp_attr);
+ }
bpf_mlock = NULL;
bpf_mlock_attr = NULL;
return;
}
- for (i = 0; i < NBPFILTER; i++)
+ for (i = 0; i < NBPFILTER; i++) {
bpf_make_dev_t(maj);
+ }
}
#else
cdevsw_add(&bpf_cdevsw);
}
#ifndef __APPLE__
-SYSINIT(bpfdev, SI_SUB_DRIVERS, SI_ORDER_MIDDLE+CDEV_MAJOR, bpf_drvinit, NULL)
+SYSINIT(bpfdev, SI_SUB_DRIVERS, SI_ORDER_MIDDLE + CDEV_MAJOR, bpf_drvinit, NULL);
#endif
-#if CONFIG_MACF_NET
-struct label *
-mac_bpfdesc_label_get(struct bpf_d *d)
+static int
+sysctl_bpf_maxbufsize SYSCTL_HANDLER_ARGS
{
+#pragma unused(arg1, arg2)
+ int i, err;
- return (d->bd_label);
-}
+ i = bpf_maxbufsize;
-void
-mac_bpfdesc_label_set(struct bpf_d *d, struct label *label)
-{
+ err = sysctl_handle_int(oidp, &i, 0, req);
+ if (err != 0 || req->newptr == USER_ADDR_NULL) {
+ return err;
+ }
- d->bd_label = label;
+ if (i < 0 || i > BPF_MAXSIZE_CAP) {
+ i = BPF_MAXSIZE_CAP;
+ }
+
+ bpf_maxbufsize = i;
+ return err;
}
-#endif