]>
git.saurik.com Git - apple/xnu.git/blob - bsd/netinet/in_cksum.c
bc302ae30e799d4b08f3ddbc197691f1aeaef390
2 * Copyright (c) 2000-2012 Apple Inc. All rights reserved.
4 * @APPLE_OSREFERENCE_LICENSE_HEADER_START@
6 * This file contains Original Code and/or Modifications of Original Code
7 * as defined in and that are subject to the Apple Public Source License
8 * Version 2.0 (the 'License'). You may not use this file except in
9 * compliance with the License. The rights granted to you under the License
10 * may not be used to create, or enable the creation or redistribution of,
11 * unlawful or unlicensed copies of an Apple operating system, or to
12 * circumvent, violate, or enable the circumvention or violation of, any
13 * terms of an Apple operating system software license agreement.
15 * Please obtain a copy of the License at
16 * http://www.opensource.apple.com/apsl/ and read it before using this file.
18 * The Original Code and all software distributed under the License are
19 * distributed on an 'AS IS' basis, WITHOUT WARRANTY OF ANY KIND, EITHER
20 * EXPRESS OR IMPLIED, AND APPLE HEREBY DISCLAIMS ALL SUCH WARRANTIES,
21 * INCLUDING WITHOUT LIMITATION, ANY WARRANTIES OF MERCHANTABILITY,
22 * FITNESS FOR A PARTICULAR PURPOSE, QUIET ENJOYMENT OR NON-INFRINGEMENT.
23 * Please see the License for the specific language governing rights and
24 * limitations under the License.
26 * @APPLE_OSREFERENCE_LICENSE_HEADER_END@
29 * Copyright (c) 1988, 1992, 1993
30 * The Regents of the University of California. All rights reserved.
32 * Redistribution and use in source and binary forms, with or without
33 * modification, are permitted provided that the following conditions
35 * 1. Redistributions of source code must retain the above copyright
36 * notice, this list of conditions and the following disclaimer.
37 * 2. Redistributions in binary form must reproduce the above copyright
38 * notice, this list of conditions and the following disclaimer in the
39 * documentation and/or other materials provided with the distribution.
40 * 3. All advertising materials mentioning features or use of this software
41 * must display the following acknowledgement:
42 * This product includes software developed by the University of
43 * California, Berkeley and its contributors.
44 * 4. Neither the name of the University nor the names of its contributors
45 * may be used to endorse or promote products derived from this software
46 * without specific prior written permission.
48 * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
49 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
50 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
51 * ARE DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
52 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
53 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
54 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
55 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
56 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
57 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
60 * @(#)in_cksum.c 8.1 (Berkeley) 6/10/93
63 #include <sys/param.h>
64 #include <machine/endian.h>
66 #include <kern/debug.h>
68 #include <netinet/in.h>
70 #include <netinet/ip.h>
71 #include <netinet/ip_var.h>
74 * Checksum routine for Internet Protocol family headers (Portable Version).
76 * This routine is very heavily used in the network
77 * code and should be modified for each CPU to be as fast as possible.
81 l_util.l = q_util.s[0] + q_util.s[1] + q_util.s[2] + q_util.s[3]; \
82 sum = l_util.s[0] + l_util.s[1]; \
97 #define PREDICT_FALSE(_exp) __builtin_expect((_exp), 0)
99 static uint16_t in_cksumdata(const void *buf
, int len
);
102 * Portable version of 16-bit 1's complement sum function that works
103 * on a contiguous buffer. This is used mainly for instances where
104 * the caller is certain about the buffer requirements, e.g. for IP
105 * header checksum calculation, though it is capable of being used
106 * on any arbitrary data span. The platform-specific cpu_in_cksum()
107 * routine might be better-optmized, so use that instead for large
110 * The logic is borrowed from <bsd/netinet/cpu_in_cksum.c>
113 #if ULONG_MAX == 0xffffffffUL
116 in_cksumdata(const void *buf
, int mlen
)
118 uint32_t sum
, partial
;
119 unsigned int final_acc
;
120 const uint8_t *data
= (const uint8_t *)buf
;
121 boolean_t needs_swap
, started_on_odd
;
126 started_on_odd
= FALSE
;
131 if ((uintptr_t)data
& 1) {
132 /* Align on word boundary */
133 started_on_odd
= !started_on_odd
;
134 #if BYTE_ORDER == LITTLE_ENDIAN
135 partial
= *data
<< 8;
142 needs_swap
= started_on_odd
;
144 __builtin_prefetch(data
+ 32);
145 partial
+= *(const uint16_t *)(const void *)data
;
146 partial
+= *(const uint16_t *)(const void *)(data
+ 2);
147 partial
+= *(const uint16_t *)(const void *)(data
+ 4);
148 partial
+= *(const uint16_t *)(const void *)(data
+ 6);
149 partial
+= *(const uint16_t *)(const void *)(data
+ 8);
150 partial
+= *(const uint16_t *)(const void *)(data
+ 10);
151 partial
+= *(const uint16_t *)(const void *)(data
+ 12);
152 partial
+= *(const uint16_t *)(const void *)(data
+ 14);
153 partial
+= *(const uint16_t *)(const void *)(data
+ 16);
154 partial
+= *(const uint16_t *)(const void *)(data
+ 18);
155 partial
+= *(const uint16_t *)(const void *)(data
+ 20);
156 partial
+= *(const uint16_t *)(const void *)(data
+ 22);
157 partial
+= *(const uint16_t *)(const void *)(data
+ 24);
158 partial
+= *(const uint16_t *)(const void *)(data
+ 26);
159 partial
+= *(const uint16_t *)(const void *)(data
+ 28);
160 partial
+= *(const uint16_t *)(const void *)(data
+ 30);
163 if (PREDICT_FALSE(partial
& 0xc0000000)) {
165 partial
= (partial
<< 8) +
167 sum
+= (partial
>> 16);
168 sum
+= (partial
& 0xffff);
173 partial
+= *(const uint16_t *)(const void *)data
;
174 partial
+= *(const uint16_t *)(const void *)(data
+ 2);
175 partial
+= *(const uint16_t *)(const void *)(data
+ 4);
176 partial
+= *(const uint16_t *)(const void *)(data
+ 6);
177 partial
+= *(const uint16_t *)(const void *)(data
+ 8);
178 partial
+= *(const uint16_t *)(const void *)(data
+ 10);
179 partial
+= *(const uint16_t *)(const void *)(data
+ 12);
180 partial
+= *(const uint16_t *)(const void *)(data
+ 14);
185 * mlen is not updated below as the remaining tests
186 * are using bit masks, which are not affected.
189 partial
+= *(const uint16_t *)(const void *)data
;
190 partial
+= *(const uint16_t *)(const void *)(data
+ 2);
191 partial
+= *(const uint16_t *)(const void *)(data
+ 4);
192 partial
+= *(const uint16_t *)(const void *)(data
+ 6);
196 partial
+= *(const uint16_t *)(const void *)data
;
197 partial
+= *(const uint16_t *)(const void *)(data
+ 2);
201 partial
+= *(const uint16_t *)(const void *)data
;
205 #if BYTE_ORDER == LITTLE_ENDIAN
208 partial
+= *data
<< 8;
210 started_on_odd
= !started_on_odd
;
214 partial
= (partial
<< 8) + (partial
>> 24);
215 sum
+= (partial
>> 16) + (partial
& 0xffff);
216 sum
= (sum
>> 16) + (sum
& 0xffff);
218 final_acc
= ((sum
>> 16) & 0xffff) + (sum
& 0xffff);
219 final_acc
= (final_acc
>> 16) + (final_acc
& 0xffff);
227 in_cksumdata(const void *buf
, int mlen
)
229 uint64_t sum
, partial
;
230 unsigned int final_acc
;
231 const uint8_t *data
= (const uint8_t *)buf
;
232 boolean_t needs_swap
, started_on_odd
;
237 started_on_odd
= FALSE
;
242 if ((uintptr_t)data
& 1) {
243 /* Align on word boundary */
244 started_on_odd
= !started_on_odd
;
245 #if BYTE_ORDER == LITTLE_ENDIAN
246 partial
= *data
<< 8;
253 needs_swap
= started_on_odd
;
254 if ((uintptr_t)data
& 2) {
257 partial
+= *(const uint16_t *)(const void *)data
;
262 __builtin_prefetch(data
+ 32);
263 __builtin_prefetch(data
+ 64);
264 partial
+= *(const uint32_t *)(const void *)data
;
265 partial
+= *(const uint32_t *)(const void *)(data
+ 4);
266 partial
+= *(const uint32_t *)(const void *)(data
+ 8);
267 partial
+= *(const uint32_t *)(const void *)(data
+ 12);
268 partial
+= *(const uint32_t *)(const void *)(data
+ 16);
269 partial
+= *(const uint32_t *)(const void *)(data
+ 20);
270 partial
+= *(const uint32_t *)(const void *)(data
+ 24);
271 partial
+= *(const uint32_t *)(const void *)(data
+ 28);
272 partial
+= *(const uint32_t *)(const void *)(data
+ 32);
273 partial
+= *(const uint32_t *)(const void *)(data
+ 36);
274 partial
+= *(const uint32_t *)(const void *)(data
+ 40);
275 partial
+= *(const uint32_t *)(const void *)(data
+ 44);
276 partial
+= *(const uint32_t *)(const void *)(data
+ 48);
277 partial
+= *(const uint32_t *)(const void *)(data
+ 52);
278 partial
+= *(const uint32_t *)(const void *)(data
+ 56);
279 partial
+= *(const uint32_t *)(const void *)(data
+ 60);
282 if (PREDICT_FALSE(partial
& (3ULL << 62))) {
284 partial
= (partial
<< 8) +
286 sum
+= (partial
>> 32);
287 sum
+= (partial
& 0xffffffff);
292 * mlen is not updated below as the remaining tests
293 * are using bit masks, which are not affected.
296 partial
+= *(const uint32_t *)(const void *)data
;
297 partial
+= *(const uint32_t *)(const void *)(data
+ 4);
298 partial
+= *(const uint32_t *)(const void *)(data
+ 8);
299 partial
+= *(const uint32_t *)(const void *)(data
+ 12);
300 partial
+= *(const uint32_t *)(const void *)(data
+ 16);
301 partial
+= *(const uint32_t *)(const void *)(data
+ 20);
302 partial
+= *(const uint32_t *)(const void *)(data
+ 24);
303 partial
+= *(const uint32_t *)(const void *)(data
+ 28);
307 partial
+= *(const uint32_t *)(const void *)data
;
308 partial
+= *(const uint32_t *)(const void *)(data
+ 4);
309 partial
+= *(const uint32_t *)(const void *)(data
+ 8);
310 partial
+= *(const uint32_t *)(const void *)(data
+ 12);
314 partial
+= *(const uint32_t *)(const void *)data
;
315 partial
+= *(const uint32_t *)(const void *)(data
+ 4);
319 partial
+= *(const uint32_t *)(const void *)data
;
323 partial
+= *(const uint16_t *)(const void *)data
;
328 #if BYTE_ORDER == LITTLE_ENDIAN
331 partial
+= *data
<< 8;
333 started_on_odd
= !started_on_odd
;
337 partial
= (partial
<< 8) + (partial
>> 56);
338 sum
+= (partial
>> 32) + (partial
& 0xffffffff);
339 sum
= (sum
>> 32) + (sum
& 0xffffffff);
341 final_acc
= (sum
>> 48) + ((sum
>> 32) & 0xffff) +
342 ((sum
>> 16) & 0xffff) + (sum
& 0xffff);
343 final_acc
= (final_acc
>> 16) + (final_acc
& 0xffff);
344 final_acc
= (final_acc
>> 16) + (final_acc
& 0xffff);
348 #endif /* ULONG_MAX != 0xffffffffUL */
351 * Perform 16-bit 1's complement sum on a contiguous span.
354 b_sum16(const void *buf
, int len
)
356 return (in_cksumdata(buf
, len
));
359 uint16_t inet_cksum_simple(struct mbuf
*, int);
361 * For the exported _in_cksum symbol in BSDKernel symbol set.
364 inet_cksum_simple(struct mbuf
*m
, int len
)
366 return (inet_cksum(m
, 0, 0, len
));
370 in_addword(uint16_t a
, uint16_t b
)
372 uint64_t sum
= a
+ b
;
379 in_pseudo(uint32_t a
, uint32_t b
, uint32_t c
)
385 sum
= (uint64_t)a
+ b
+ c
;
391 in_pseudo64(uint64_t a
, uint64_t b
, uint64_t c
)
403 * May be used on IP header with options.
406 in_cksum_hdr_opt(const struct ip
*ip
)
408 return (~b_sum16(ip
, (IP_VHL_HL(ip
->ip_vhl
) << 2)) & 0xffff);
412 * A wrapper around the simple in_cksum_hdr() and the more complicated
413 * inet_cksum(); the former is chosen if the IP header is simple,
414 * contiguous and 32-bit aligned. Also does some stats accounting.
417 ip_cksum_hdr_dir(struct mbuf
*m
, uint32_t hlen
, int out
)
419 struct ip
*ip
= mtod(m
, struct ip
*);
422 ipstat
.ips_snd_swcsum
++;
423 ipstat
.ips_snd_swcsum_bytes
+= hlen
;
425 ipstat
.ips_rcv_swcsum
++;
426 ipstat
.ips_rcv_swcsum_bytes
+= hlen
;
429 if (hlen
== sizeof (*ip
) &&
430 m
->m_len
>= sizeof (*ip
) && IP_HDR_ALIGNED_P(ip
))
431 return (in_cksum_hdr(ip
));
433 return (inet_cksum(m
, 0, 0, hlen
));
437 * m MUST contain at least an IP header, if nxt is specified;
438 * nxt is the upper layer protocol number;
439 * off is an offset where TCP/UDP/ICMP header starts;
440 * len is a total length of a transport segment (e.g. TCP header + TCP payload)
443 inet_cksum(struct mbuf
*m
, uint32_t nxt
, uint32_t off
, uint32_t len
)
447 sum
= m_sum16(m
, off
, len
);
449 /* include pseudo header checksum? */
452 unsigned char buf
[sizeof ((*ip
))] __attribute__((aligned(8)));
458 * Use m_length2() instead of m_length(), as we cannot rely on
459 * the caller setting m_pkthdr.len correctly, if the mbuf is
462 if ((mlen
= m_length2(m
, NULL
)) < sizeof (*ip
)) {
463 panic("%s: mbuf %p too short (%d) for IPv4 header",
469 * In case the IP header is not contiguous, or not 32-bit
470 * aligned, copy it to a local buffer. Note here that we
471 * expect the data pointer to point to the IP header.
473 if ((sizeof (*ip
) > m
->m_len
) ||
474 !IP_HDR_ALIGNED_P(mtod(m
, caddr_t
))) {
475 m_copydata(m
, 0, sizeof (*ip
), (caddr_t
)buf
);
476 ip
= (struct ip
*)(void *)buf
;
478 ip
= (struct ip
*)(void *)(m
->m_data
);
481 /* add pseudo header checksum */
482 sum
+= in_pseudo(ip
->ip_src
.s_addr
, ip
->ip_dst
.s_addr
,
485 /* fold in carry bits */
489 return (~sum
& 0xffff);