1*4882a593Smuzhiyun // SPDX-License-Identifier: GPL-2.0-or-later
2*4882a593Smuzhiyun /*
3*4882a593Smuzhiyun * Generic implementation of 64-bit atomics using spinlocks,
4*4882a593Smuzhiyun * useful on processors that don't have 64-bit atomic instructions.
5*4882a593Smuzhiyun *
6*4882a593Smuzhiyun * Copyright © 2009 Paul Mackerras, IBM Corp. <paulus@au1.ibm.com>
7*4882a593Smuzhiyun */
8*4882a593Smuzhiyun #include <linux/types.h>
9*4882a593Smuzhiyun #include <linux/cache.h>
10*4882a593Smuzhiyun #include <linux/spinlock.h>
11*4882a593Smuzhiyun #include <linux/init.h>
12*4882a593Smuzhiyun #include <linux/export.h>
13*4882a593Smuzhiyun #include <linux/atomic.h>
14*4882a593Smuzhiyun
15*4882a593Smuzhiyun /*
16*4882a593Smuzhiyun * We use a hashed array of spinlocks to provide exclusive access
17*4882a593Smuzhiyun * to each atomic64_t variable. Since this is expected to used on
18*4882a593Smuzhiyun * systems with small numbers of CPUs (<= 4 or so), we use a
19*4882a593Smuzhiyun * relatively small array of 16 spinlocks to avoid wasting too much
20*4882a593Smuzhiyun * memory on the spinlock array.
21*4882a593Smuzhiyun */
22*4882a593Smuzhiyun #define NR_LOCKS 16
23*4882a593Smuzhiyun
24*4882a593Smuzhiyun /*
25*4882a593Smuzhiyun * Ensure each lock is in a separate cacheline.
26*4882a593Smuzhiyun */
27*4882a593Smuzhiyun static union {
28*4882a593Smuzhiyun raw_spinlock_t lock;
29*4882a593Smuzhiyun char pad[L1_CACHE_BYTES];
30*4882a593Smuzhiyun } atomic64_lock[NR_LOCKS] __cacheline_aligned_in_smp = {
31*4882a593Smuzhiyun [0 ... (NR_LOCKS - 1)] = {
32*4882a593Smuzhiyun .lock = __RAW_SPIN_LOCK_UNLOCKED(atomic64_lock.lock),
33*4882a593Smuzhiyun },
34*4882a593Smuzhiyun };
35*4882a593Smuzhiyun
lock_addr(const atomic64_t * v)36*4882a593Smuzhiyun static inline raw_spinlock_t *lock_addr(const atomic64_t *v)
37*4882a593Smuzhiyun {
38*4882a593Smuzhiyun unsigned long addr = (unsigned long) v;
39*4882a593Smuzhiyun
40*4882a593Smuzhiyun addr >>= L1_CACHE_SHIFT;
41*4882a593Smuzhiyun addr ^= (addr >> 8) ^ (addr >> 16);
42*4882a593Smuzhiyun return &atomic64_lock[addr & (NR_LOCKS - 1)].lock;
43*4882a593Smuzhiyun }
44*4882a593Smuzhiyun
atomic64_read(const atomic64_t * v)45*4882a593Smuzhiyun s64 atomic64_read(const atomic64_t *v)
46*4882a593Smuzhiyun {
47*4882a593Smuzhiyun unsigned long flags;
48*4882a593Smuzhiyun raw_spinlock_t *lock = lock_addr(v);
49*4882a593Smuzhiyun s64 val;
50*4882a593Smuzhiyun
51*4882a593Smuzhiyun raw_spin_lock_irqsave(lock, flags);
52*4882a593Smuzhiyun val = v->counter;
53*4882a593Smuzhiyun raw_spin_unlock_irqrestore(lock, flags);
54*4882a593Smuzhiyun return val;
55*4882a593Smuzhiyun }
56*4882a593Smuzhiyun EXPORT_SYMBOL(atomic64_read);
57*4882a593Smuzhiyun
atomic64_set(atomic64_t * v,s64 i)58*4882a593Smuzhiyun void atomic64_set(atomic64_t *v, s64 i)
59*4882a593Smuzhiyun {
60*4882a593Smuzhiyun unsigned long flags;
61*4882a593Smuzhiyun raw_spinlock_t *lock = lock_addr(v);
62*4882a593Smuzhiyun
63*4882a593Smuzhiyun raw_spin_lock_irqsave(lock, flags);
64*4882a593Smuzhiyun v->counter = i;
65*4882a593Smuzhiyun raw_spin_unlock_irqrestore(lock, flags);
66*4882a593Smuzhiyun }
67*4882a593Smuzhiyun EXPORT_SYMBOL(atomic64_set);
68*4882a593Smuzhiyun
69*4882a593Smuzhiyun #define ATOMIC64_OP(op, c_op) \
70*4882a593Smuzhiyun void atomic64_##op(s64 a, atomic64_t *v) \
71*4882a593Smuzhiyun { \
72*4882a593Smuzhiyun unsigned long flags; \
73*4882a593Smuzhiyun raw_spinlock_t *lock = lock_addr(v); \
74*4882a593Smuzhiyun \
75*4882a593Smuzhiyun raw_spin_lock_irqsave(lock, flags); \
76*4882a593Smuzhiyun v->counter c_op a; \
77*4882a593Smuzhiyun raw_spin_unlock_irqrestore(lock, flags); \
78*4882a593Smuzhiyun } \
79*4882a593Smuzhiyun EXPORT_SYMBOL(atomic64_##op);
80*4882a593Smuzhiyun
81*4882a593Smuzhiyun #define ATOMIC64_OP_RETURN(op, c_op) \
82*4882a593Smuzhiyun s64 atomic64_##op##_return(s64 a, atomic64_t *v) \
83*4882a593Smuzhiyun { \
84*4882a593Smuzhiyun unsigned long flags; \
85*4882a593Smuzhiyun raw_spinlock_t *lock = lock_addr(v); \
86*4882a593Smuzhiyun s64 val; \
87*4882a593Smuzhiyun \
88*4882a593Smuzhiyun raw_spin_lock_irqsave(lock, flags); \
89*4882a593Smuzhiyun val = (v->counter c_op a); \
90*4882a593Smuzhiyun raw_spin_unlock_irqrestore(lock, flags); \
91*4882a593Smuzhiyun return val; \
92*4882a593Smuzhiyun } \
93*4882a593Smuzhiyun EXPORT_SYMBOL(atomic64_##op##_return);
94*4882a593Smuzhiyun
95*4882a593Smuzhiyun #define ATOMIC64_FETCH_OP(op, c_op) \
96*4882a593Smuzhiyun s64 atomic64_fetch_##op(s64 a, atomic64_t *v) \
97*4882a593Smuzhiyun { \
98*4882a593Smuzhiyun unsigned long flags; \
99*4882a593Smuzhiyun raw_spinlock_t *lock = lock_addr(v); \
100*4882a593Smuzhiyun s64 val; \
101*4882a593Smuzhiyun \
102*4882a593Smuzhiyun raw_spin_lock_irqsave(lock, flags); \
103*4882a593Smuzhiyun val = v->counter; \
104*4882a593Smuzhiyun v->counter c_op a; \
105*4882a593Smuzhiyun raw_spin_unlock_irqrestore(lock, flags); \
106*4882a593Smuzhiyun return val; \
107*4882a593Smuzhiyun } \
108*4882a593Smuzhiyun EXPORT_SYMBOL(atomic64_fetch_##op);
109*4882a593Smuzhiyun
110*4882a593Smuzhiyun #define ATOMIC64_OPS(op, c_op) \
111*4882a593Smuzhiyun ATOMIC64_OP(op, c_op) \
112*4882a593Smuzhiyun ATOMIC64_OP_RETURN(op, c_op) \
113*4882a593Smuzhiyun ATOMIC64_FETCH_OP(op, c_op)
114*4882a593Smuzhiyun
115*4882a593Smuzhiyun ATOMIC64_OPS(add, +=)
116*4882a593Smuzhiyun ATOMIC64_OPS(sub, -=)
117*4882a593Smuzhiyun
118*4882a593Smuzhiyun #undef ATOMIC64_OPS
119*4882a593Smuzhiyun #define ATOMIC64_OPS(op, c_op) \
120*4882a593Smuzhiyun ATOMIC64_OP(op, c_op) \
121*4882a593Smuzhiyun ATOMIC64_OP_RETURN(op, c_op) \
122*4882a593Smuzhiyun ATOMIC64_FETCH_OP(op, c_op)
123*4882a593Smuzhiyun
124*4882a593Smuzhiyun ATOMIC64_OPS(and, &=)
125*4882a593Smuzhiyun ATOMIC64_OPS(or, |=)
126*4882a593Smuzhiyun ATOMIC64_OPS(xor, ^=)
127*4882a593Smuzhiyun
128*4882a593Smuzhiyun #undef ATOMIC64_OPS
129*4882a593Smuzhiyun #undef ATOMIC64_FETCH_OP
130*4882a593Smuzhiyun #undef ATOMIC64_OP_RETURN
131*4882a593Smuzhiyun #undef ATOMIC64_OP
132*4882a593Smuzhiyun
atomic64_dec_if_positive(atomic64_t * v)133*4882a593Smuzhiyun s64 atomic64_dec_if_positive(atomic64_t *v)
134*4882a593Smuzhiyun {
135*4882a593Smuzhiyun unsigned long flags;
136*4882a593Smuzhiyun raw_spinlock_t *lock = lock_addr(v);
137*4882a593Smuzhiyun s64 val;
138*4882a593Smuzhiyun
139*4882a593Smuzhiyun raw_spin_lock_irqsave(lock, flags);
140*4882a593Smuzhiyun val = v->counter - 1;
141*4882a593Smuzhiyun if (val >= 0)
142*4882a593Smuzhiyun v->counter = val;
143*4882a593Smuzhiyun raw_spin_unlock_irqrestore(lock, flags);
144*4882a593Smuzhiyun return val;
145*4882a593Smuzhiyun }
146*4882a593Smuzhiyun EXPORT_SYMBOL(atomic64_dec_if_positive);
147*4882a593Smuzhiyun
atomic64_cmpxchg(atomic64_t * v,s64 o,s64 n)148*4882a593Smuzhiyun s64 atomic64_cmpxchg(atomic64_t *v, s64 o, s64 n)
149*4882a593Smuzhiyun {
150*4882a593Smuzhiyun unsigned long flags;
151*4882a593Smuzhiyun raw_spinlock_t *lock = lock_addr(v);
152*4882a593Smuzhiyun s64 val;
153*4882a593Smuzhiyun
154*4882a593Smuzhiyun raw_spin_lock_irqsave(lock, flags);
155*4882a593Smuzhiyun val = v->counter;
156*4882a593Smuzhiyun if (val == o)
157*4882a593Smuzhiyun v->counter = n;
158*4882a593Smuzhiyun raw_spin_unlock_irqrestore(lock, flags);
159*4882a593Smuzhiyun return val;
160*4882a593Smuzhiyun }
161*4882a593Smuzhiyun EXPORT_SYMBOL(atomic64_cmpxchg);
162*4882a593Smuzhiyun
atomic64_xchg(atomic64_t * v,s64 new)163*4882a593Smuzhiyun s64 atomic64_xchg(atomic64_t *v, s64 new)
164*4882a593Smuzhiyun {
165*4882a593Smuzhiyun unsigned long flags;
166*4882a593Smuzhiyun raw_spinlock_t *lock = lock_addr(v);
167*4882a593Smuzhiyun s64 val;
168*4882a593Smuzhiyun
169*4882a593Smuzhiyun raw_spin_lock_irqsave(lock, flags);
170*4882a593Smuzhiyun val = v->counter;
171*4882a593Smuzhiyun v->counter = new;
172*4882a593Smuzhiyun raw_spin_unlock_irqrestore(lock, flags);
173*4882a593Smuzhiyun return val;
174*4882a593Smuzhiyun }
175*4882a593Smuzhiyun EXPORT_SYMBOL(atomic64_xchg);
176*4882a593Smuzhiyun
atomic64_fetch_add_unless(atomic64_t * v,s64 a,s64 u)177*4882a593Smuzhiyun s64 atomic64_fetch_add_unless(atomic64_t *v, s64 a, s64 u)
178*4882a593Smuzhiyun {
179*4882a593Smuzhiyun unsigned long flags;
180*4882a593Smuzhiyun raw_spinlock_t *lock = lock_addr(v);
181*4882a593Smuzhiyun s64 val;
182*4882a593Smuzhiyun
183*4882a593Smuzhiyun raw_spin_lock_irqsave(lock, flags);
184*4882a593Smuzhiyun val = v->counter;
185*4882a593Smuzhiyun if (val != u)
186*4882a593Smuzhiyun v->counter += a;
187*4882a593Smuzhiyun raw_spin_unlock_irqrestore(lock, flags);
188*4882a593Smuzhiyun
189*4882a593Smuzhiyun return val;
190*4882a593Smuzhiyun }
191*4882a593Smuzhiyun EXPORT_SYMBOL(atomic64_fetch_add_unless);
192