^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 1) /* SPDX-License-Identifier: GPL-2.0-or-later */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 2) /*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 3) * PowerPC atomic bit operations.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 4) *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 5) * Merged version by David Gibson <david@gibson.dropbear.id.au>.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 6) * Based on ppc64 versions by: Dave Engebretsen, Todd Inglett, Don
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 7) * Reed, Pat McCarthy, Peter Bergner, Anton Blanchard. They
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 8) * originally took it from the ppc32 code.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 9) *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 10) * Within a word, bits are numbered LSB first. Lot's of places make
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 11) * this assumption by directly testing bits with (val & (1<<nr)).
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 12) * This can cause confusion for large (> 1 word) bitmaps on a
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 13) * big-endian system because, unlike little endian, the number of each
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 14) * bit depends on the word size.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 15) *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 16) * The bitop functions are defined to work on unsigned longs, so for a
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 17) * ppc64 system the bits end up numbered:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 18) * |63..............0|127............64|191...........128|255...........192|
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 19) * and on ppc32:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 20) * |31.....0|63....32|95....64|127...96|159..128|191..160|223..192|255..224|
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 21) *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 22) * There are a few little-endian macros used mostly for filesystem
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 23) * bitmaps, these work on similar bit arrays layouts, but
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 24) * byte-oriented:
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 25) * |7...0|15...8|23...16|31...24|39...32|47...40|55...48|63...56|
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 26) *
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 27) * The main difference is that bit 3-5 (64b) or 3-4 (32b) in the bit
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 28) * number field needs to be reversed compared to the big-endian bit
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 29) * fields. This can be achieved by XOR with 0x38 (64b) or 0x18 (32b).
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 30) */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 31)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 32) #ifndef _ASM_POWERPC_BITOPS_H
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 33) #define _ASM_POWERPC_BITOPS_H
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 34)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 35) #ifdef __KERNEL__
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 36)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 37) #ifndef _LINUX_BITOPS_H
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 38) #error only <linux/bitops.h> can be included directly
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 39) #endif
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 40)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 41) #include <linux/compiler.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 42) #include <asm/asm-compat.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 43) #include <asm/synch.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 44)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 45) /* PPC bit number conversion */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 46) #define PPC_BITLSHIFT(be) (BITS_PER_LONG - 1 - (be))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 47) #define PPC_BIT(bit) (1UL << PPC_BITLSHIFT(bit))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 48) #define PPC_BITMASK(bs, be) ((PPC_BIT(bs) - PPC_BIT(be)) | PPC_BIT(bs))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 49)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 50) /* Put a PPC bit into a "normal" bit position */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 51) #define PPC_BITEXTRACT(bits, ppc_bit, dst_bit) \
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 52) ((((bits) >> PPC_BITLSHIFT(ppc_bit)) & 1) << (dst_bit))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 53)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 54) #define PPC_BITLSHIFT32(be) (32 - 1 - (be))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 55) #define PPC_BIT32(bit) (1UL << PPC_BITLSHIFT32(bit))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 56) #define PPC_BITMASK32(bs, be) ((PPC_BIT32(bs) - PPC_BIT32(be))|PPC_BIT32(bs))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 57)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 58) #define PPC_BITLSHIFT8(be) (8 - 1 - (be))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 59) #define PPC_BIT8(bit) (1UL << PPC_BITLSHIFT8(bit))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 60) #define PPC_BITMASK8(bs, be) ((PPC_BIT8(bs) - PPC_BIT8(be))|PPC_BIT8(bs))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 61)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 62) #include <asm/barrier.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 63)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 64) /* Macro for generating the ***_bits() functions */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 65) #define DEFINE_BITOP(fn, op, prefix) \
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 66) static inline void fn(unsigned long mask, \
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 67) volatile unsigned long *_p) \
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 68) { \
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 69) unsigned long old; \
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 70) unsigned long *p = (unsigned long *)_p; \
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 71) __asm__ __volatile__ ( \
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 72) prefix \
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 73) "1:" PPC_LLARX(%0,0,%3,0) "\n" \
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 74) stringify_in_c(op) "%0,%0,%2\n" \
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 75) PPC_STLCX "%0,0,%3\n" \
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 76) "bne- 1b\n" \
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 77) : "=&r" (old), "+m" (*p) \
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 78) : "r" (mask), "r" (p) \
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 79) : "cc", "memory"); \
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 80) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 81)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 82) DEFINE_BITOP(set_bits, or, "")
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 83) DEFINE_BITOP(clear_bits, andc, "")
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 84) DEFINE_BITOP(clear_bits_unlock, andc, PPC_RELEASE_BARRIER)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 85) DEFINE_BITOP(change_bits, xor, "")
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 86)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 87) static inline void arch_set_bit(int nr, volatile unsigned long *addr)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 88) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 89) set_bits(BIT_MASK(nr), addr + BIT_WORD(nr));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 90) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 91)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 92) static inline void arch_clear_bit(int nr, volatile unsigned long *addr)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 93) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 94) clear_bits(BIT_MASK(nr), addr + BIT_WORD(nr));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 95) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 96)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 97) static inline void arch_clear_bit_unlock(int nr, volatile unsigned long *addr)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 98) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 99) clear_bits_unlock(BIT_MASK(nr), addr + BIT_WORD(nr));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 100) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 101)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 102) static inline void arch_change_bit(int nr, volatile unsigned long *addr)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 103) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 104) change_bits(BIT_MASK(nr), addr + BIT_WORD(nr));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 105) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 106)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 107) /* Like DEFINE_BITOP(), with changes to the arguments to 'op' and the output
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 108) * operands. */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 109) #define DEFINE_TESTOP(fn, op, prefix, postfix, eh) \
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 110) static inline unsigned long fn( \
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 111) unsigned long mask, \
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 112) volatile unsigned long *_p) \
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 113) { \
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 114) unsigned long old, t; \
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 115) unsigned long *p = (unsigned long *)_p; \
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 116) __asm__ __volatile__ ( \
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 117) prefix \
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 118) "1:" PPC_LLARX(%0,0,%3,eh) "\n" \
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 119) stringify_in_c(op) "%1,%0,%2\n" \
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 120) PPC_STLCX "%1,0,%3\n" \
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 121) "bne- 1b\n" \
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 122) postfix \
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 123) : "=&r" (old), "=&r" (t) \
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 124) : "r" (mask), "r" (p) \
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 125) : "cc", "memory"); \
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 126) return (old & mask); \
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 127) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 128)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 129) DEFINE_TESTOP(test_and_set_bits, or, PPC_ATOMIC_ENTRY_BARRIER,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 130) PPC_ATOMIC_EXIT_BARRIER, 0)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 131) DEFINE_TESTOP(test_and_set_bits_lock, or, "",
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 132) PPC_ACQUIRE_BARRIER, 1)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 133) DEFINE_TESTOP(test_and_clear_bits, andc, PPC_ATOMIC_ENTRY_BARRIER,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 134) PPC_ATOMIC_EXIT_BARRIER, 0)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 135) DEFINE_TESTOP(test_and_change_bits, xor, PPC_ATOMIC_ENTRY_BARRIER,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 136) PPC_ATOMIC_EXIT_BARRIER, 0)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 137)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 138) static inline int arch_test_and_set_bit(unsigned long nr,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 139) volatile unsigned long *addr)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 140) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 141) return test_and_set_bits(BIT_MASK(nr), addr + BIT_WORD(nr)) != 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 142) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 143)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 144) static inline int arch_test_and_set_bit_lock(unsigned long nr,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 145) volatile unsigned long *addr)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 146) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 147) return test_and_set_bits_lock(BIT_MASK(nr),
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 148) addr + BIT_WORD(nr)) != 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 149) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 150)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 151) static inline int arch_test_and_clear_bit(unsigned long nr,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 152) volatile unsigned long *addr)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 153) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 154) return test_and_clear_bits(BIT_MASK(nr), addr + BIT_WORD(nr)) != 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 155) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 156)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 157) static inline int arch_test_and_change_bit(unsigned long nr,
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 158) volatile unsigned long *addr)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 159) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 160) return test_and_change_bits(BIT_MASK(nr), addr + BIT_WORD(nr)) != 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 161) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 162)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 163) #ifdef CONFIG_PPC64
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 164) static inline unsigned long
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 165) clear_bit_unlock_return_word(int nr, volatile unsigned long *addr)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 166) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 167) unsigned long old, t;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 168) unsigned long *p = (unsigned long *)addr + BIT_WORD(nr);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 169) unsigned long mask = BIT_MASK(nr);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 170)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 171) __asm__ __volatile__ (
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 172) PPC_RELEASE_BARRIER
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 173) "1:" PPC_LLARX(%0,0,%3,0) "\n"
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 174) "andc %1,%0,%2\n"
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 175) PPC_STLCX "%1,0,%3\n"
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 176) "bne- 1b\n"
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 177) : "=&r" (old), "=&r" (t)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 178) : "r" (mask), "r" (p)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 179) : "cc", "memory");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 180)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 181) return old;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 182) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 183)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 184) /*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 185) * This is a special function for mm/filemap.c
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 186) * Bit 7 corresponds to PG_waiters.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 187) */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 188) #define arch_clear_bit_unlock_is_negative_byte(nr, addr) \
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 189) (clear_bit_unlock_return_word(nr, addr) & BIT_MASK(7))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 190)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 191) #endif /* CONFIG_PPC64 */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 192)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 193) #include <asm-generic/bitops/non-atomic.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 194)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 195) static inline void arch___clear_bit_unlock(int nr, volatile unsigned long *addr)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 196) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 197) __asm__ __volatile__(PPC_RELEASE_BARRIER "" ::: "memory");
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 198) __clear_bit(nr, addr);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 199) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 200)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 201) /*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 202) * Return the zero-based bit position (LE, not IBM bit numbering) of
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 203) * the most significant 1-bit in a double word.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 204) */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 205) #define __ilog2(x) ilog2(x)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 206)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 207) #include <asm-generic/bitops/ffz.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 208)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 209) #include <asm-generic/bitops/builtin-__ffs.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 210)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 211) #include <asm-generic/bitops/builtin-ffs.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 212)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 213) /*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 214) * fls: find last (most-significant) bit set.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 215) * Note fls(0) = 0, fls(1) = 1, fls(0x80000000) = 32.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 216) */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 217) static inline int fls(unsigned int x)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 218) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 219) int lz;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 220)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 221) if (__builtin_constant_p(x))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 222) return x ? 32 - __builtin_clz(x) : 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 223) asm("cntlzw %0,%1" : "=r" (lz) : "r" (x));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 224) return 32 - lz;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 225) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 226)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 227) #include <asm-generic/bitops/builtin-__fls.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 228)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 229) /*
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 230) * 64-bit can do this using one cntlzd (count leading zeroes doubleword)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 231) * instruction; for 32-bit we use the generic version, which does two
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 232) * 32-bit fls calls.
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 233) */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 234) #ifdef CONFIG_PPC64
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 235) static inline int fls64(__u64 x)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 236) {
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 237) int lz;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 238)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 239) if (__builtin_constant_p(x))
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 240) return x ? 64 - __builtin_clzll(x) : 0;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 241) asm("cntlzd %0,%1" : "=r" (lz) : "r" (x));
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 242) return 64 - lz;
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 243) }
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 244) #else
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 245) #include <asm-generic/bitops/fls64.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 246) #endif
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 247)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 248) #ifdef CONFIG_PPC64
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 249) unsigned int __arch_hweight8(unsigned int w);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 250) unsigned int __arch_hweight16(unsigned int w);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 251) unsigned int __arch_hweight32(unsigned int w);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 252) unsigned long __arch_hweight64(__u64 w);
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 253) #include <asm-generic/bitops/const_hweight.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 254) #else
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 255) #include <asm-generic/bitops/hweight.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 256) #endif
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 257)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 258) #include <asm-generic/bitops/find.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 259)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 260) /* wrappers that deal with KASAN instrumentation */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 261) #include <asm-generic/bitops/instrumented-atomic.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 262) #include <asm-generic/bitops/instrumented-lock.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 263)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 264) /* Little-endian versions */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 265) #include <asm-generic/bitops/le.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 266)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 267) /* Bitmap functions for the ext2 filesystem */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 268)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 269) #include <asm-generic/bitops/ext2-atomic-setbit.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 270)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 271) #include <asm-generic/bitops/sched.h>
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 272)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 273) #endif /* __KERNEL__ */
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 274)
^8f3ce5b39 (kx 2023-10-28 12:00:06 +0300 275) #endif /* _ASM_POWERPC_BITOPS_H */