^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1) /* SPDX-License-Identifier: GPL-2.0 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2) #ifndef _TOOLS_LINUX_BITOPS_H_
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3) #define _TOOLS_LINUX_BITOPS_H_
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 5) #include <asm/types.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 6) #include <limits.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 7) #ifndef __WORDSIZE
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 8) #define __WORDSIZE (__SIZEOF_LONG__ * 8)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 9) #endif
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 10)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 11) #ifndef BITS_PER_LONG
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 12) # define BITS_PER_LONG __WORDSIZE
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 13) #endif
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 14) #include <linux/bits.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 15) #include <linux/compiler.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 16)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 17) #define BITS_PER_TYPE(type) (sizeof(type) * BITS_PER_BYTE)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 18) #define BITS_TO_LONGS(nr) DIV_ROUND_UP(nr, BITS_PER_TYPE(long))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 19) #define BITS_TO_U64(nr) DIV_ROUND_UP(nr, BITS_PER_TYPE(u64))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 20) #define BITS_TO_U32(nr) DIV_ROUND_UP(nr, BITS_PER_TYPE(u32))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 21) #define BITS_TO_BYTES(nr) DIV_ROUND_UP(nr, BITS_PER_TYPE(char))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 22)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 23) extern unsigned int __sw_hweight8(unsigned int w);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 24) extern unsigned int __sw_hweight16(unsigned int w);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 25) extern unsigned int __sw_hweight32(unsigned int w);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 26) extern unsigned long __sw_hweight64(__u64 w);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 27)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 28) /*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 29) * Include this here because some architectures need generic_ffs/fls in
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 30) * scope
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 31) *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 32) * XXX: this needs to be asm/bitops.h, when we get to per arch optimizations
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 33) */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 34) #include <asm-generic/bitops.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 35)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 36) #define for_each_set_bit(bit, addr, size) \
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 37) for ((bit) = find_first_bit((addr), (size)); \
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 38) (bit) < (size); \
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 39) (bit) = find_next_bit((addr), (size), (bit) + 1))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 40)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 41) #define for_each_clear_bit(bit, addr, size) \
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 42) for ((bit) = find_first_zero_bit((addr), (size)); \
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 43) (bit) < (size); \
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 44) (bit) = find_next_zero_bit((addr), (size), (bit) + 1))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 45)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 46) /* same as for_each_set_bit() but use bit as value to start with */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 47) #define for_each_set_bit_from(bit, addr, size) \
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 48) for ((bit) = find_next_bit((addr), (size), (bit)); \
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 49) (bit) < (size); \
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 50) (bit) = find_next_bit((addr), (size), (bit) + 1))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 51)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 52) static inline unsigned long hweight_long(unsigned long w)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 53) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 54) return sizeof(w) == 4 ? hweight32(w) : hweight64(w);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 55) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 56)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 57) static inline unsigned fls_long(unsigned long l)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 58) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 59) if (sizeof(l) == 4)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 60) return fls(l);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 61) return fls64(l);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 62) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 63)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 64) /**
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 65) * rol32 - rotate a 32-bit value left
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 66) * @word: value to rotate
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 67) * @shift: bits to roll
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 68) */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 69) static inline __u32 rol32(__u32 word, unsigned int shift)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 70) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 71) return (word << shift) | (word >> ((-shift) & 31));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 72) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 73)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 74) #endif