aboutsummaryrefslogtreecommitdiff
path: root/arch/parisc
diff options
context:
space:
mode:
authorGravatar Al Viro <viro@zeniv.linux.org.uk> 2024-04-01 22:22:50 -0400
committerGravatar Paul E. McKenney <paulmck@kernel.org> 2024-04-09 22:06:00 -0700
commit29b8e53c1274f0ffda915ac6c0e5c59d14ee208e (patch)
tree818340c32db544c3c02c863ba6b248082ae9afa6 /arch/parisc
parentsparc32: add __cmpxchg_u{8,16}() and teach __cmpxchg() to handle those sizes (diff)
downloadlinux-29b8e53c1274f0ffda915ac6c0e5c59d14ee208e.tar.gz
linux-29b8e53c1274f0ffda915ac6c0e5c59d14ee208e.tar.bz2
linux-29b8e53c1274f0ffda915ac6c0e5c59d14ee208e.zip
parisc: __cmpxchg_u32(): lift conversion into the callers
__cmpxchg_u32() return value is unsigned int explicitly cast to unsigned long. Both callers are returns from functions that return unsigned long; might as well have __cmpxchg_u32() return that unsigned int (aka u32) and let the callers convert implicitly. Signed-off-by: Al Viro <viro@zeniv.linux.org.uk> Signed-off-by: Paul E. McKenney <paulmck@kernel.org>
Diffstat (limited to 'arch/parisc')
-rw-r--r--arch/parisc/include/asm/cmpxchg.h3
-rw-r--r--arch/parisc/lib/bitops.c6
2 files changed, 4 insertions, 5 deletions
diff --git a/arch/parisc/include/asm/cmpxchg.h b/arch/parisc/include/asm/cmpxchg.h
index c1d776bb16b4..0924ebc576d2 100644
--- a/arch/parisc/include/asm/cmpxchg.h
+++ b/arch/parisc/include/asm/cmpxchg.h
@@ -57,8 +57,7 @@ __arch_xchg(unsigned long x, volatile void *ptr, int size)
extern void __cmpxchg_called_with_bad_pointer(void);
/* __cmpxchg_u32/u64 defined in arch/parisc/lib/bitops.c */
-extern unsigned long __cmpxchg_u32(volatile unsigned int *m, unsigned int old,
- unsigned int new_);
+extern u32 __cmpxchg_u32(volatile u32 *m, u32 old, u32 new_);
extern u64 __cmpxchg_u64(volatile u64 *ptr, u64 old, u64 new_);
extern u8 __cmpxchg_u8(volatile u8 *ptr, u8 old, u8 new_);
diff --git a/arch/parisc/lib/bitops.c b/arch/parisc/lib/bitops.c
index 36a314199074..ae2231d92198 100644
--- a/arch/parisc/lib/bitops.c
+++ b/arch/parisc/lib/bitops.c
@@ -68,16 +68,16 @@ u64 notrace __cmpxchg_u64(volatile u64 *ptr, u64 old, u64 new)
return prev;
}
-unsigned long notrace __cmpxchg_u32(volatile unsigned int *ptr, unsigned int old, unsigned int new)
+u32 notrace __cmpxchg_u32(volatile u32 *ptr, u32 old, u32 new)
{
unsigned long flags;
- unsigned int prev;
+ u32 prev;
_atomic_spin_lock_irqsave(ptr, flags);
if ((prev = *ptr) == old)
*ptr = new;
_atomic_spin_unlock_irqrestore(ptr, flags);
- return (unsigned long)prev;
+ return prev;
}
u8 notrace __cmpxchg_u8(volatile u8 *ptr, u8 old, u8 new)