xref: /OK3568_Linux_fs/kernel/lib/atomic64.c (revision 4882a59341e53eb6f0b4789bf948001014eff981)
1*4882a593Smuzhiyun // SPDX-License-Identifier: GPL-2.0-or-later
2*4882a593Smuzhiyun /*
3*4882a593Smuzhiyun  * Generic implementation of 64-bit atomics using spinlocks,
4*4882a593Smuzhiyun  * useful on processors that don't have 64-bit atomic instructions.
5*4882a593Smuzhiyun  *
6*4882a593Smuzhiyun  * Copyright © 2009 Paul Mackerras, IBM Corp. <paulus@au1.ibm.com>
7*4882a593Smuzhiyun  */
8*4882a593Smuzhiyun #include <linux/types.h>
9*4882a593Smuzhiyun #include <linux/cache.h>
10*4882a593Smuzhiyun #include <linux/spinlock.h>
11*4882a593Smuzhiyun #include <linux/init.h>
12*4882a593Smuzhiyun #include <linux/export.h>
13*4882a593Smuzhiyun #include <linux/atomic.h>
14*4882a593Smuzhiyun 
15*4882a593Smuzhiyun /*
16*4882a593Smuzhiyun  * We use a hashed array of spinlocks to provide exclusive access
17*4882a593Smuzhiyun  * to each atomic64_t variable.  Since this is expected to used on
18*4882a593Smuzhiyun  * systems with small numbers of CPUs (<= 4 or so), we use a
19*4882a593Smuzhiyun  * relatively small array of 16 spinlocks to avoid wasting too much
20*4882a593Smuzhiyun  * memory on the spinlock array.
21*4882a593Smuzhiyun  */
22*4882a593Smuzhiyun #define NR_LOCKS	16
23*4882a593Smuzhiyun 
24*4882a593Smuzhiyun /*
25*4882a593Smuzhiyun  * Ensure each lock is in a separate cacheline.
26*4882a593Smuzhiyun  */
27*4882a593Smuzhiyun static union {
28*4882a593Smuzhiyun 	raw_spinlock_t lock;
29*4882a593Smuzhiyun 	char pad[L1_CACHE_BYTES];
30*4882a593Smuzhiyun } atomic64_lock[NR_LOCKS] __cacheline_aligned_in_smp = {
31*4882a593Smuzhiyun 	[0 ... (NR_LOCKS - 1)] = {
32*4882a593Smuzhiyun 		.lock =  __RAW_SPIN_LOCK_UNLOCKED(atomic64_lock.lock),
33*4882a593Smuzhiyun 	},
34*4882a593Smuzhiyun };
35*4882a593Smuzhiyun 
lock_addr(const atomic64_t * v)36*4882a593Smuzhiyun static inline raw_spinlock_t *lock_addr(const atomic64_t *v)
37*4882a593Smuzhiyun {
38*4882a593Smuzhiyun 	unsigned long addr = (unsigned long) v;
39*4882a593Smuzhiyun 
40*4882a593Smuzhiyun 	addr >>= L1_CACHE_SHIFT;
41*4882a593Smuzhiyun 	addr ^= (addr >> 8) ^ (addr >> 16);
42*4882a593Smuzhiyun 	return &atomic64_lock[addr & (NR_LOCKS - 1)].lock;
43*4882a593Smuzhiyun }
44*4882a593Smuzhiyun 
atomic64_read(const atomic64_t * v)45*4882a593Smuzhiyun s64 atomic64_read(const atomic64_t *v)
46*4882a593Smuzhiyun {
47*4882a593Smuzhiyun 	unsigned long flags;
48*4882a593Smuzhiyun 	raw_spinlock_t *lock = lock_addr(v);
49*4882a593Smuzhiyun 	s64 val;
50*4882a593Smuzhiyun 
51*4882a593Smuzhiyun 	raw_spin_lock_irqsave(lock, flags);
52*4882a593Smuzhiyun 	val = v->counter;
53*4882a593Smuzhiyun 	raw_spin_unlock_irqrestore(lock, flags);
54*4882a593Smuzhiyun 	return val;
55*4882a593Smuzhiyun }
56*4882a593Smuzhiyun EXPORT_SYMBOL(atomic64_read);
57*4882a593Smuzhiyun 
atomic64_set(atomic64_t * v,s64 i)58*4882a593Smuzhiyun void atomic64_set(atomic64_t *v, s64 i)
59*4882a593Smuzhiyun {
60*4882a593Smuzhiyun 	unsigned long flags;
61*4882a593Smuzhiyun 	raw_spinlock_t *lock = lock_addr(v);
62*4882a593Smuzhiyun 
63*4882a593Smuzhiyun 	raw_spin_lock_irqsave(lock, flags);
64*4882a593Smuzhiyun 	v->counter = i;
65*4882a593Smuzhiyun 	raw_spin_unlock_irqrestore(lock, flags);
66*4882a593Smuzhiyun }
67*4882a593Smuzhiyun EXPORT_SYMBOL(atomic64_set);
68*4882a593Smuzhiyun 
69*4882a593Smuzhiyun #define ATOMIC64_OP(op, c_op)						\
70*4882a593Smuzhiyun void atomic64_##op(s64 a, atomic64_t *v)				\
71*4882a593Smuzhiyun {									\
72*4882a593Smuzhiyun 	unsigned long flags;						\
73*4882a593Smuzhiyun 	raw_spinlock_t *lock = lock_addr(v);				\
74*4882a593Smuzhiyun 									\
75*4882a593Smuzhiyun 	raw_spin_lock_irqsave(lock, flags);				\
76*4882a593Smuzhiyun 	v->counter c_op a;						\
77*4882a593Smuzhiyun 	raw_spin_unlock_irqrestore(lock, flags);			\
78*4882a593Smuzhiyun }									\
79*4882a593Smuzhiyun EXPORT_SYMBOL(atomic64_##op);
80*4882a593Smuzhiyun 
81*4882a593Smuzhiyun #define ATOMIC64_OP_RETURN(op, c_op)					\
82*4882a593Smuzhiyun s64 atomic64_##op##_return(s64 a, atomic64_t *v)			\
83*4882a593Smuzhiyun {									\
84*4882a593Smuzhiyun 	unsigned long flags;						\
85*4882a593Smuzhiyun 	raw_spinlock_t *lock = lock_addr(v);				\
86*4882a593Smuzhiyun 	s64 val;							\
87*4882a593Smuzhiyun 									\
88*4882a593Smuzhiyun 	raw_spin_lock_irqsave(lock, flags);				\
89*4882a593Smuzhiyun 	val = (v->counter c_op a);					\
90*4882a593Smuzhiyun 	raw_spin_unlock_irqrestore(lock, flags);			\
91*4882a593Smuzhiyun 	return val;							\
92*4882a593Smuzhiyun }									\
93*4882a593Smuzhiyun EXPORT_SYMBOL(atomic64_##op##_return);
94*4882a593Smuzhiyun 
95*4882a593Smuzhiyun #define ATOMIC64_FETCH_OP(op, c_op)					\
96*4882a593Smuzhiyun s64 atomic64_fetch_##op(s64 a, atomic64_t *v)				\
97*4882a593Smuzhiyun {									\
98*4882a593Smuzhiyun 	unsigned long flags;						\
99*4882a593Smuzhiyun 	raw_spinlock_t *lock = lock_addr(v);				\
100*4882a593Smuzhiyun 	s64 val;							\
101*4882a593Smuzhiyun 									\
102*4882a593Smuzhiyun 	raw_spin_lock_irqsave(lock, flags);				\
103*4882a593Smuzhiyun 	val = v->counter;						\
104*4882a593Smuzhiyun 	v->counter c_op a;						\
105*4882a593Smuzhiyun 	raw_spin_unlock_irqrestore(lock, flags);			\
106*4882a593Smuzhiyun 	return val;							\
107*4882a593Smuzhiyun }									\
108*4882a593Smuzhiyun EXPORT_SYMBOL(atomic64_fetch_##op);
109*4882a593Smuzhiyun 
110*4882a593Smuzhiyun #define ATOMIC64_OPS(op, c_op)						\
111*4882a593Smuzhiyun 	ATOMIC64_OP(op, c_op)						\
112*4882a593Smuzhiyun 	ATOMIC64_OP_RETURN(op, c_op)					\
113*4882a593Smuzhiyun 	ATOMIC64_FETCH_OP(op, c_op)
114*4882a593Smuzhiyun 
115*4882a593Smuzhiyun ATOMIC64_OPS(add, +=)
116*4882a593Smuzhiyun ATOMIC64_OPS(sub, -=)
117*4882a593Smuzhiyun 
118*4882a593Smuzhiyun #undef ATOMIC64_OPS
119*4882a593Smuzhiyun #define ATOMIC64_OPS(op, c_op)						\
120*4882a593Smuzhiyun 	ATOMIC64_OP(op, c_op)						\
121*4882a593Smuzhiyun 	ATOMIC64_OP_RETURN(op, c_op)					\
122*4882a593Smuzhiyun 	ATOMIC64_FETCH_OP(op, c_op)
123*4882a593Smuzhiyun 
124*4882a593Smuzhiyun ATOMIC64_OPS(and, &=)
125*4882a593Smuzhiyun ATOMIC64_OPS(or, |=)
126*4882a593Smuzhiyun ATOMIC64_OPS(xor, ^=)
127*4882a593Smuzhiyun 
128*4882a593Smuzhiyun #undef ATOMIC64_OPS
129*4882a593Smuzhiyun #undef ATOMIC64_FETCH_OP
130*4882a593Smuzhiyun #undef ATOMIC64_OP_RETURN
131*4882a593Smuzhiyun #undef ATOMIC64_OP
132*4882a593Smuzhiyun 
atomic64_dec_if_positive(atomic64_t * v)133*4882a593Smuzhiyun s64 atomic64_dec_if_positive(atomic64_t *v)
134*4882a593Smuzhiyun {
135*4882a593Smuzhiyun 	unsigned long flags;
136*4882a593Smuzhiyun 	raw_spinlock_t *lock = lock_addr(v);
137*4882a593Smuzhiyun 	s64 val;
138*4882a593Smuzhiyun 
139*4882a593Smuzhiyun 	raw_spin_lock_irqsave(lock, flags);
140*4882a593Smuzhiyun 	val = v->counter - 1;
141*4882a593Smuzhiyun 	if (val >= 0)
142*4882a593Smuzhiyun 		v->counter = val;
143*4882a593Smuzhiyun 	raw_spin_unlock_irqrestore(lock, flags);
144*4882a593Smuzhiyun 	return val;
145*4882a593Smuzhiyun }
146*4882a593Smuzhiyun EXPORT_SYMBOL(atomic64_dec_if_positive);
147*4882a593Smuzhiyun 
atomic64_cmpxchg(atomic64_t * v,s64 o,s64 n)148*4882a593Smuzhiyun s64 atomic64_cmpxchg(atomic64_t *v, s64 o, s64 n)
149*4882a593Smuzhiyun {
150*4882a593Smuzhiyun 	unsigned long flags;
151*4882a593Smuzhiyun 	raw_spinlock_t *lock = lock_addr(v);
152*4882a593Smuzhiyun 	s64 val;
153*4882a593Smuzhiyun 
154*4882a593Smuzhiyun 	raw_spin_lock_irqsave(lock, flags);
155*4882a593Smuzhiyun 	val = v->counter;
156*4882a593Smuzhiyun 	if (val == o)
157*4882a593Smuzhiyun 		v->counter = n;
158*4882a593Smuzhiyun 	raw_spin_unlock_irqrestore(lock, flags);
159*4882a593Smuzhiyun 	return val;
160*4882a593Smuzhiyun }
161*4882a593Smuzhiyun EXPORT_SYMBOL(atomic64_cmpxchg);
162*4882a593Smuzhiyun 
atomic64_xchg(atomic64_t * v,s64 new)163*4882a593Smuzhiyun s64 atomic64_xchg(atomic64_t *v, s64 new)
164*4882a593Smuzhiyun {
165*4882a593Smuzhiyun 	unsigned long flags;
166*4882a593Smuzhiyun 	raw_spinlock_t *lock = lock_addr(v);
167*4882a593Smuzhiyun 	s64 val;
168*4882a593Smuzhiyun 
169*4882a593Smuzhiyun 	raw_spin_lock_irqsave(lock, flags);
170*4882a593Smuzhiyun 	val = v->counter;
171*4882a593Smuzhiyun 	v->counter = new;
172*4882a593Smuzhiyun 	raw_spin_unlock_irqrestore(lock, flags);
173*4882a593Smuzhiyun 	return val;
174*4882a593Smuzhiyun }
175*4882a593Smuzhiyun EXPORT_SYMBOL(atomic64_xchg);
176*4882a593Smuzhiyun 
atomic64_fetch_add_unless(atomic64_t * v,s64 a,s64 u)177*4882a593Smuzhiyun s64 atomic64_fetch_add_unless(atomic64_t *v, s64 a, s64 u)
178*4882a593Smuzhiyun {
179*4882a593Smuzhiyun 	unsigned long flags;
180*4882a593Smuzhiyun 	raw_spinlock_t *lock = lock_addr(v);
181*4882a593Smuzhiyun 	s64 val;
182*4882a593Smuzhiyun 
183*4882a593Smuzhiyun 	raw_spin_lock_irqsave(lock, flags);
184*4882a593Smuzhiyun 	val = v->counter;
185*4882a593Smuzhiyun 	if (val != u)
186*4882a593Smuzhiyun 		v->counter += a;
187*4882a593Smuzhiyun 	raw_spin_unlock_irqrestore(lock, flags);
188*4882a593Smuzhiyun 
189*4882a593Smuzhiyun 	return val;
190*4882a593Smuzhiyun }
191*4882a593Smuzhiyun EXPORT_SYMBOL(atomic64_fetch_add_unless);
192