^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1) /* SPDX-License-Identifier: GPL-2.0-or-later */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2) /*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3) * INET An implementation of the TCP/IP protocol suite for the LINUX
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4) * operating system. INET is implemented using the BSD Socket
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 5) * interface as the means of communication with the user level.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 6) *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 7) * Checksumming functions for IP, TCP, UDP and so on
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 8) *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 9) * Authors: Jorge Cwik, <jorge@laser.satlink.net>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 10) * Arnt Gulbrandsen, <agulbra@nvg.unit.no>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 11) * Borrows very liberally from tcp.c and ip.c, see those
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 12) * files for more names.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 13) */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 14)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 15) #ifndef _CHECKSUM_H
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 16) #define _CHECKSUM_H
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 17)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 18) #include <linux/errno.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 19) #include <asm/types.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 20) #include <asm/byteorder.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 21) #include <linux/uaccess.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 22) #include <asm/checksum.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 23)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 24) #ifndef _HAVE_ARCH_COPY_AND_CSUM_FROM_USER
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 25) static __always_inline
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 26) __wsum csum_and_copy_from_user (const void __user *src, void *dst,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 27) int len)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 28) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 29) if (copy_from_user(dst, src, len))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 30) return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 31) return csum_partial(dst, len, ~0U);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 32) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 33) #endif
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 34)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 35) #ifndef HAVE_CSUM_COPY_USER
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 36) static __always_inline __wsum csum_and_copy_to_user
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 37) (const void *src, void __user *dst, int len)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 38) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 39) __wsum sum = csum_partial(src, len, ~0U);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 40)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 41) if (copy_to_user(dst, src, len) == 0)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 42) return sum;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 43) return 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 44) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 45) #endif
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 46)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 47) #ifndef _HAVE_ARCH_CSUM_AND_COPY
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 48) static __always_inline __wsum
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 49) csum_partial_copy_nocheck(const void *src, void *dst, int len)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 50) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 51) memcpy(dst, src, len);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 52) return csum_partial(dst, len, 0);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 53) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 54) #endif
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 55)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 56) #ifndef HAVE_ARCH_CSUM_ADD
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 57) static __always_inline __wsum csum_add(__wsum csum, __wsum addend)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 58) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 59) u32 res = (__force u32)csum;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 60) res += (__force u32)addend;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 61) return (__force __wsum)(res + (res < (__force u32)addend));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 62) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 63) #endif
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 64)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 65) static __always_inline __wsum csum_sub(__wsum csum, __wsum addend)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 66) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 67) return csum_add(csum, ~addend);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 68) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 69)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 70) static __always_inline __sum16 csum16_add(__sum16 csum, __be16 addend)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 71) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 72) u16 res = (__force u16)csum;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 73)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 74) res += (__force u16)addend;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 75) return (__force __sum16)(res + (res < (__force u16)addend));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 76) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 77)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 78) static __always_inline __sum16 csum16_sub(__sum16 csum, __be16 addend)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 79) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 80) return csum16_add(csum, ~addend);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 81) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 82)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 83) static __always_inline __wsum
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 84) csum_block_add(__wsum csum, __wsum csum2, int offset)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 85) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 86) u32 sum = (__force u32)csum2;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 87)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 88) /* rotate sum to align it with a 16b boundary */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 89) if (offset & 1)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 90) sum = ror32(sum, 8);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 91)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 92) return csum_add(csum, (__force __wsum)sum);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 93) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 94)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 95) static __always_inline __wsum
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 96) csum_block_add_ext(__wsum csum, __wsum csum2, int offset, int len)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 97) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 98) return csum_block_add(csum, csum2, offset);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 99) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 100)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 101) static __always_inline __wsum
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 102) csum_block_sub(__wsum csum, __wsum csum2, int offset)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 103) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 104) return csum_block_add(csum, ~csum2, offset);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 105) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 106)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 107) static __always_inline __wsum csum_unfold(__sum16 n)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 108) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 109) return (__force __wsum)n;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 110) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 111)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 112) static __always_inline
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 113) __wsum csum_partial_ext(const void *buff, int len, __wsum sum)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 114) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 115) return csum_partial(buff, len, sum);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 116) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 117)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 118) #define CSUM_MANGLED_0 ((__force __sum16)0xffff)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 119)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 120) static __always_inline void csum_replace_by_diff(__sum16 *sum, __wsum diff)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 121) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 122) *sum = csum_fold(csum_add(diff, ~csum_unfold(*sum)));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 123) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 124)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 125) static __always_inline void csum_replace4(__sum16 *sum, __be32 from, __be32 to)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 126) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 127) __wsum tmp = csum_sub(~csum_unfold(*sum), (__force __wsum)from);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 128)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 129) *sum = csum_fold(csum_add(tmp, (__force __wsum)to));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 130) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 131)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 132) /* Implements RFC 1624 (Incremental Internet Checksum)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 133) * 3. Discussion states :
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 134) * HC' = ~(~HC + ~m + m')
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 135) * m : old value of a 16bit field
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 136) * m' : new value of a 16bit field
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 137) */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 138) static __always_inline void csum_replace2(__sum16 *sum, __be16 old, __be16 new)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 139) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 140) *sum = ~csum16_add(csum16_sub(~(*sum), old), new);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 141) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 142)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 143) static inline void csum_replace(__wsum *csum, __wsum old, __wsum new)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 144) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 145) *csum = csum_add(csum_sub(*csum, old), new);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 146) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 147)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 148) struct sk_buff;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 149) void inet_proto_csum_replace4(__sum16 *sum, struct sk_buff *skb,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 150) __be32 from, __be32 to, bool pseudohdr);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 151) void inet_proto_csum_replace16(__sum16 *sum, struct sk_buff *skb,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 152) const __be32 *from, const __be32 *to,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 153) bool pseudohdr);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 154) void inet_proto_csum_replace_by_diff(__sum16 *sum, struct sk_buff *skb,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 155) __wsum diff, bool pseudohdr);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 156)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 157) static __always_inline
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 158) void inet_proto_csum_replace2(__sum16 *sum, struct sk_buff *skb,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 159) __be16 from, __be16 to, bool pseudohdr)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 160) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 161) inet_proto_csum_replace4(sum, skb, (__force __be32)from,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 162) (__force __be32)to, pseudohdr);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 163) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 164)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 165) static __always_inline __wsum remcsum_adjust(void *ptr, __wsum csum,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 166) int start, int offset)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 167) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 168) __sum16 *psum = (__sum16 *)(ptr + offset);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 169) __wsum delta;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 170)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 171) /* Subtract out checksum up to start */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 172) csum = csum_sub(csum, csum_partial(ptr, start, 0));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 173)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 174) /* Set derived checksum in packet */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 175) delta = csum_sub((__force __wsum)csum_fold(csum),
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 176) (__force __wsum)*psum);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 177) *psum = csum_fold(csum);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 178)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 179) return delta;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 180) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 181)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 182) static __always_inline void remcsum_unadjust(__sum16 *psum, __wsum delta)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 183) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 184) *psum = csum_fold(csum_sub(delta, (__force __wsum)*psum));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 185) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 186)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 187) #endif