diff options
author | David S. Miller <davem@davemloft.net> | 2009-11-27 02:28:13 +0300 |
---|---|---|
committer | David S. Miller <davem@davemloft.net> | 2009-11-27 02:28:13 +0300 |
commit | 55f532ec3e84410a9393d1a21b1f58518a0958bc (patch) | |
tree | 0a4f8f00bf1de07619752f58af2d7c724344b163 /arch/sparc | |
parent | 0448f5e88443667d6a2e809eb470960bdbb07657 (diff) | |
download | linux-55f532ec3e84410a9393d1a21b1f58518a0958bc.tar.xz |
Revert "sparc: Make atomic locks raw"
This reverts commit 4df286e52917c95c415400367cfd523dfbb0f93a.
Breaks the build as reported by Stephen Rothwell.
Signed-off-by: David S. Miller <davem@davemloft.net>
Diffstat (limited to 'arch/sparc')
-rw-r--r-- | arch/sparc/lib/atomic32.c | 36 |
1 files changed, 18 insertions, 18 deletions
diff --git a/arch/sparc/lib/atomic32.c b/arch/sparc/lib/atomic32.c index 080b7c26e0fd..cbddeb38ffda 100644 --- a/arch/sparc/lib/atomic32.c +++ b/arch/sparc/lib/atomic32.c @@ -15,8 +15,8 @@ #define ATOMIC_HASH_SIZE 4 #define ATOMIC_HASH(a) (&__atomic_hash[(((unsigned long)a)>>8) & (ATOMIC_HASH_SIZE-1)]) -static raw_spinlock_t __atomic_hash[ATOMIC_HASH_SIZE] = { - [0 ... (ATOMIC_HASH_SIZE-1)] = __RAW_SPIN_LOCK_UNLOCKED +spinlock_t __atomic_hash[ATOMIC_HASH_SIZE] = { + [0 ... (ATOMIC_HASH_SIZE-1)] = SPIN_LOCK_UNLOCKED }; #else /* SMP */ @@ -31,11 +31,11 @@ int __atomic_add_return(int i, atomic_t *v) { int ret; unsigned long flags; - __raw_spin_lock_irqsave(ATOMIC_HASH(v), flags); + spin_lock_irqsave(ATOMIC_HASH(v), flags); ret = (v->counter += i); - __raw_spin_unlock_irqrestore(ATOMIC_HASH(v), flags); + spin_unlock_irqrestore(ATOMIC_HASH(v), flags); return ret; } EXPORT_SYMBOL(__atomic_add_return); @@ -45,12 +45,12 @@ int atomic_cmpxchg(atomic_t *v, int old, int new) int ret; unsigned long flags; - __raw_spin_lock_irqsave(ATOMIC_HASH(v), flags); + spin_lock_irqsave(ATOMIC_HASH(v), flags); ret = v->counter; if (likely(ret == old)) v->counter = new; - __raw_spin_unlock_irqrestore(ATOMIC_HASH(v), flags); + spin_unlock_irqrestore(ATOMIC_HASH(v), flags); return ret; } EXPORT_SYMBOL(atomic_cmpxchg); @@ -60,11 +60,11 @@ int atomic_add_unless(atomic_t *v, int a, int u) int ret; unsigned long flags; - __raw_spin_lock_irqsave(ATOMIC_HASH(v), flags); + spin_lock_irqsave(ATOMIC_HASH(v), flags); ret = v->counter; if (ret != u) v->counter += a; - __raw_spin_unlock_irqrestore(ATOMIC_HASH(v), flags); + spin_unlock_irqrestore(ATOMIC_HASH(v), flags); return ret != u; } EXPORT_SYMBOL(atomic_add_unless); @@ -74,9 +74,9 @@ void atomic_set(atomic_t *v, int i) { unsigned long flags; - __raw_spin_lock_irqsave(ATOMIC_HASH(v), flags); + spin_lock_irqsave(ATOMIC_HASH(v), flags); v->counter = i; - __raw_spin_unlock_irqrestore(ATOMIC_HASH(v), flags); + spin_unlock_irqrestore(ATOMIC_HASH(v), flags); } EXPORT_SYMBOL(atomic_set); @@ -84,10 +84,10 @@ unsigned long ___set_bit(unsigned long *addr, unsigned long mask) { unsigned long old, flags; - __raw_spin_lock_irqsave(ATOMIC_HASH(addr), flags); + spin_lock_irqsave(ATOMIC_HASH(addr), flags); old = *addr; *addr = old | mask; - __raw_spin_unlock_irqrestore(ATOMIC_HASH(addr), flags); + spin_unlock_irqrestore(ATOMIC_HASH(addr), flags); return old & mask; } @@ -97,10 +97,10 @@ unsigned long ___clear_bit(unsigned long *addr, unsigned long mask) { unsigned long old, flags; - __raw_spin_lock_irqsave(ATOMIC_HASH(addr), flags); + spin_lock_irqsave(ATOMIC_HASH(addr), flags); old = *addr; *addr = old & ~mask; - __raw_spin_unlock_irqrestore(ATOMIC_HASH(addr), flags); + spin_unlock_irqrestore(ATOMIC_HASH(addr), flags); return old & mask; } @@ -110,10 +110,10 @@ unsigned long ___change_bit(unsigned long *addr, unsigned long mask) { unsigned long old, flags; - __raw_spin_lock_irqsave(ATOMIC_HASH(addr), flags); + spin_lock_irqsave(ATOMIC_HASH(addr), flags); old = *addr; *addr = old ^ mask; - __raw_spin_unlock_irqrestore(ATOMIC_HASH(addr), flags); + spin_unlock_irqrestore(ATOMIC_HASH(addr), flags); return old & mask; } @@ -124,10 +124,10 @@ unsigned long __cmpxchg_u32(volatile u32 *ptr, u32 old, u32 new) unsigned long flags; u32 prev; - __raw_spin_lock_irqsave(ATOMIC_HASH(ptr), flags); + spin_lock_irqsave(ATOMIC_HASH(ptr), flags); if ((prev = *ptr) == old) *ptr = new; - __raw_spin_unlock_irqrestore(ATOMIC_HASH(ptr), flags); + spin_unlock_irqrestore(ATOMIC_HASH(ptr), flags); return (unsigned long)prev; } |