]> www.infradead.org Git - users/willy/pagecache.git/commitdiff
powerpc: Implement arch_xor_unlock_is_negative_byte on 32-bit
authorMatthew Wilcox (Oracle) <willy@infradead.org>
Wed, 19 Jul 2023 19:50:04 +0000 (15:50 -0400)
committerMatthew Wilcox (Oracle) <willy@infradead.org>
Wed, 4 Oct 2023 01:18:59 +0000 (21:18 -0400)
Simply remove the ifdef.  The assembly is identical to that in the
non-optimised case of test_and_clear_bits() on PPC32, and it's not clear
to me how the PPC32 optimisation works, nor whether it would work for
arch_xor_unlock_is_negative_byte().  If that optimisation would work,
someone can implement it later, but this is more efficient than the
implementation in filemap.c.

Signed-off-by: Matthew Wilcox (Oracle) <willy@infradead.org>
arch/powerpc/include/asm/bitops.h

index 40cc3ded60cb994ba6d81f361d52bcdf5643bea1..671ecc6711e366a4067c8aebe38e7807a922e019 100644 (file)
@@ -233,7 +233,6 @@ static inline int arch_test_and_change_bit(unsigned long nr,
        return test_and_change_bits(BIT_MASK(nr), addr + BIT_WORD(nr)) != 0;
 }
 
-#ifdef CONFIG_PPC64
 static inline bool arch_xor_unlock_is_negative_byte(unsigned long mask,
                volatile unsigned long *p)
 {
@@ -251,11 +250,8 @@ static inline bool arch_xor_unlock_is_negative_byte(unsigned long mask,
 
        return (old & BIT_MASK(7)) != 0;
 }
-
 #define arch_xor_unlock_is_negative_byte arch_xor_unlock_is_negative_byte
 
-#endif /* CONFIG_PPC64 */
-
 #include <asm-generic/bitops/non-atomic.h>
 
 static inline void arch___clear_bit_unlock(int nr, volatile unsigned long *addr)