1*4882a593Smuzhiyun /* SPDX-License-Identifier: GPL-2.0 */ 2*4882a593Smuzhiyun /* atomic.h: These still suck, but the I-cache hit rate is higher. 3*4882a593Smuzhiyun * 4*4882a593Smuzhiyun * Copyright (C) 1996 David S. Miller (davem@davemloft.net) 5*4882a593Smuzhiyun * Copyright (C) 2000 Anton Blanchard (anton@linuxcare.com.au) 6*4882a593Smuzhiyun * Copyright (C) 2007 Kyle McMartin (kyle@parisc-linux.org) 7*4882a593Smuzhiyun * 8*4882a593Smuzhiyun * Additions by Keith M Wesolowski (wesolows@foobazco.org) based 9*4882a593Smuzhiyun * on asm-parisc/atomic.h Copyright (C) 2000 Philipp Rumpf <prumpf@tux.org>. 10*4882a593Smuzhiyun */ 11*4882a593Smuzhiyun 12*4882a593Smuzhiyun #ifndef __ARCH_SPARC_ATOMIC__ 13*4882a593Smuzhiyun #define __ARCH_SPARC_ATOMIC__ 14*4882a593Smuzhiyun 15*4882a593Smuzhiyun #include <linux/types.h> 16*4882a593Smuzhiyun 17*4882a593Smuzhiyun #include <asm/cmpxchg.h> 18*4882a593Smuzhiyun #include <asm/barrier.h> 19*4882a593Smuzhiyun #include <asm-generic/atomic64.h> 20*4882a593Smuzhiyun 21*4882a593Smuzhiyun int atomic_add_return(int, atomic_t *); 22*4882a593Smuzhiyun int atomic_fetch_add(int, atomic_t *); 23*4882a593Smuzhiyun int atomic_fetch_and(int, atomic_t *); 24*4882a593Smuzhiyun int atomic_fetch_or(int, atomic_t *); 25*4882a593Smuzhiyun int atomic_fetch_xor(int, atomic_t *); 26*4882a593Smuzhiyun int atomic_cmpxchg(atomic_t *, int, int); 27*4882a593Smuzhiyun int atomic_xchg(atomic_t *, int); 28*4882a593Smuzhiyun int atomic_fetch_add_unless(atomic_t *, int, int); 29*4882a593Smuzhiyun void atomic_set(atomic_t *, int); 30*4882a593Smuzhiyun 31*4882a593Smuzhiyun #define atomic_fetch_add_unless atomic_fetch_add_unless 32*4882a593Smuzhiyun 33*4882a593Smuzhiyun #define atomic_set_release(v, i) atomic_set((v), (i)) 34*4882a593Smuzhiyun 35*4882a593Smuzhiyun #define atomic_read(v) READ_ONCE((v)->counter) 36*4882a593Smuzhiyun 37*4882a593Smuzhiyun #define atomic_add(i, v) ((void)atomic_add_return( (int)(i), (v))) 38*4882a593Smuzhiyun #define atomic_sub(i, v) ((void)atomic_add_return(-(int)(i), (v))) 39*4882a593Smuzhiyun 40*4882a593Smuzhiyun #define atomic_and(i, v) ((void)atomic_fetch_and((i), (v))) 41*4882a593Smuzhiyun #define atomic_or(i, v) ((void)atomic_fetch_or((i), (v))) 42*4882a593Smuzhiyun #define atomic_xor(i, v) ((void)atomic_fetch_xor((i), (v))) 43*4882a593Smuzhiyun 44*4882a593Smuzhiyun #define atomic_sub_return(i, v) (atomic_add_return(-(int)(i), (v))) 45*4882a593Smuzhiyun #define atomic_fetch_sub(i, v) (atomic_fetch_add (-(int)(i), (v))) 46*4882a593Smuzhiyun 47*4882a593Smuzhiyun #endif /* !(__ARCH_SPARC_ATOMIC__) */ 48