1*4882a593Smuzhiyun // SPDX-License-Identifier: GPL-2.0
2*4882a593Smuzhiyun /*
3*4882a593Smuzhiyun * Out of line spinlock code.
4*4882a593Smuzhiyun *
5*4882a593Smuzhiyun * Copyright IBM Corp. 2004, 2006
6*4882a593Smuzhiyun * Author(s): Martin Schwidefsky (schwidefsky@de.ibm.com)
7*4882a593Smuzhiyun */
8*4882a593Smuzhiyun
9*4882a593Smuzhiyun #include <linux/types.h>
10*4882a593Smuzhiyun #include <linux/export.h>
11*4882a593Smuzhiyun #include <linux/spinlock.h>
12*4882a593Smuzhiyun #include <linux/jiffies.h>
13*4882a593Smuzhiyun #include <linux/init.h>
14*4882a593Smuzhiyun #include <linux/smp.h>
15*4882a593Smuzhiyun #include <linux/percpu.h>
16*4882a593Smuzhiyun #include <asm/alternative.h>
17*4882a593Smuzhiyun #include <asm/io.h>
18*4882a593Smuzhiyun
19*4882a593Smuzhiyun int spin_retry = -1;
20*4882a593Smuzhiyun
spin_retry_init(void)21*4882a593Smuzhiyun static int __init spin_retry_init(void)
22*4882a593Smuzhiyun {
23*4882a593Smuzhiyun if (spin_retry < 0)
24*4882a593Smuzhiyun spin_retry = 1000;
25*4882a593Smuzhiyun return 0;
26*4882a593Smuzhiyun }
27*4882a593Smuzhiyun early_initcall(spin_retry_init);
28*4882a593Smuzhiyun
29*4882a593Smuzhiyun /**
30*4882a593Smuzhiyun * spin_retry= parameter
31*4882a593Smuzhiyun */
spin_retry_setup(char * str)32*4882a593Smuzhiyun static int __init spin_retry_setup(char *str)
33*4882a593Smuzhiyun {
34*4882a593Smuzhiyun spin_retry = simple_strtoul(str, &str, 0);
35*4882a593Smuzhiyun return 1;
36*4882a593Smuzhiyun }
37*4882a593Smuzhiyun __setup("spin_retry=", spin_retry_setup);
38*4882a593Smuzhiyun
39*4882a593Smuzhiyun struct spin_wait {
40*4882a593Smuzhiyun struct spin_wait *next, *prev;
41*4882a593Smuzhiyun int node_id;
42*4882a593Smuzhiyun } __aligned(32);
43*4882a593Smuzhiyun
44*4882a593Smuzhiyun static DEFINE_PER_CPU_ALIGNED(struct spin_wait, spin_wait[4]);
45*4882a593Smuzhiyun
46*4882a593Smuzhiyun #define _Q_LOCK_CPU_OFFSET 0
47*4882a593Smuzhiyun #define _Q_LOCK_STEAL_OFFSET 16
48*4882a593Smuzhiyun #define _Q_TAIL_IDX_OFFSET 18
49*4882a593Smuzhiyun #define _Q_TAIL_CPU_OFFSET 20
50*4882a593Smuzhiyun
51*4882a593Smuzhiyun #define _Q_LOCK_CPU_MASK 0x0000ffff
52*4882a593Smuzhiyun #define _Q_LOCK_STEAL_ADD 0x00010000
53*4882a593Smuzhiyun #define _Q_LOCK_STEAL_MASK 0x00030000
54*4882a593Smuzhiyun #define _Q_TAIL_IDX_MASK 0x000c0000
55*4882a593Smuzhiyun #define _Q_TAIL_CPU_MASK 0xfff00000
56*4882a593Smuzhiyun
57*4882a593Smuzhiyun #define _Q_LOCK_MASK (_Q_LOCK_CPU_MASK | _Q_LOCK_STEAL_MASK)
58*4882a593Smuzhiyun #define _Q_TAIL_MASK (_Q_TAIL_IDX_MASK | _Q_TAIL_CPU_MASK)
59*4882a593Smuzhiyun
arch_spin_lock_setup(int cpu)60*4882a593Smuzhiyun void arch_spin_lock_setup(int cpu)
61*4882a593Smuzhiyun {
62*4882a593Smuzhiyun struct spin_wait *node;
63*4882a593Smuzhiyun int ix;
64*4882a593Smuzhiyun
65*4882a593Smuzhiyun node = per_cpu_ptr(&spin_wait[0], cpu);
66*4882a593Smuzhiyun for (ix = 0; ix < 4; ix++, node++) {
67*4882a593Smuzhiyun memset(node, 0, sizeof(*node));
68*4882a593Smuzhiyun node->node_id = ((cpu + 1) << _Q_TAIL_CPU_OFFSET) +
69*4882a593Smuzhiyun (ix << _Q_TAIL_IDX_OFFSET);
70*4882a593Smuzhiyun }
71*4882a593Smuzhiyun }
72*4882a593Smuzhiyun
arch_load_niai4(int * lock)73*4882a593Smuzhiyun static inline int arch_load_niai4(int *lock)
74*4882a593Smuzhiyun {
75*4882a593Smuzhiyun int owner;
76*4882a593Smuzhiyun
77*4882a593Smuzhiyun asm_inline volatile(
78*4882a593Smuzhiyun ALTERNATIVE("", ".long 0xb2fa0040", 49) /* NIAI 4 */
79*4882a593Smuzhiyun " l %0,%1\n"
80*4882a593Smuzhiyun : "=d" (owner) : "Q" (*lock) : "memory");
81*4882a593Smuzhiyun return owner;
82*4882a593Smuzhiyun }
83*4882a593Smuzhiyun
arch_cmpxchg_niai8(int * lock,int old,int new)84*4882a593Smuzhiyun static inline int arch_cmpxchg_niai8(int *lock, int old, int new)
85*4882a593Smuzhiyun {
86*4882a593Smuzhiyun int expected = old;
87*4882a593Smuzhiyun
88*4882a593Smuzhiyun asm_inline volatile(
89*4882a593Smuzhiyun ALTERNATIVE("", ".long 0xb2fa0080", 49) /* NIAI 8 */
90*4882a593Smuzhiyun " cs %0,%3,%1\n"
91*4882a593Smuzhiyun : "=d" (old), "=Q" (*lock)
92*4882a593Smuzhiyun : "0" (old), "d" (new), "Q" (*lock)
93*4882a593Smuzhiyun : "cc", "memory");
94*4882a593Smuzhiyun return expected == old;
95*4882a593Smuzhiyun }
96*4882a593Smuzhiyun
arch_spin_decode_tail(int lock)97*4882a593Smuzhiyun static inline struct spin_wait *arch_spin_decode_tail(int lock)
98*4882a593Smuzhiyun {
99*4882a593Smuzhiyun int ix, cpu;
100*4882a593Smuzhiyun
101*4882a593Smuzhiyun ix = (lock & _Q_TAIL_IDX_MASK) >> _Q_TAIL_IDX_OFFSET;
102*4882a593Smuzhiyun cpu = (lock & _Q_TAIL_CPU_MASK) >> _Q_TAIL_CPU_OFFSET;
103*4882a593Smuzhiyun return per_cpu_ptr(&spin_wait[ix], cpu - 1);
104*4882a593Smuzhiyun }
105*4882a593Smuzhiyun
arch_spin_yield_target(int lock,struct spin_wait * node)106*4882a593Smuzhiyun static inline int arch_spin_yield_target(int lock, struct spin_wait *node)
107*4882a593Smuzhiyun {
108*4882a593Smuzhiyun if (lock & _Q_LOCK_CPU_MASK)
109*4882a593Smuzhiyun return lock & _Q_LOCK_CPU_MASK;
110*4882a593Smuzhiyun if (node == NULL || node->prev == NULL)
111*4882a593Smuzhiyun return 0; /* 0 -> no target cpu */
112*4882a593Smuzhiyun while (node->prev)
113*4882a593Smuzhiyun node = node->prev;
114*4882a593Smuzhiyun return node->node_id >> _Q_TAIL_CPU_OFFSET;
115*4882a593Smuzhiyun }
116*4882a593Smuzhiyun
arch_spin_lock_queued(arch_spinlock_t * lp)117*4882a593Smuzhiyun static inline void arch_spin_lock_queued(arch_spinlock_t *lp)
118*4882a593Smuzhiyun {
119*4882a593Smuzhiyun struct spin_wait *node, *next;
120*4882a593Smuzhiyun int lockval, ix, node_id, tail_id, old, new, owner, count;
121*4882a593Smuzhiyun
122*4882a593Smuzhiyun ix = S390_lowcore.spinlock_index++;
123*4882a593Smuzhiyun barrier();
124*4882a593Smuzhiyun lockval = SPINLOCK_LOCKVAL; /* cpu + 1 */
125*4882a593Smuzhiyun node = this_cpu_ptr(&spin_wait[ix]);
126*4882a593Smuzhiyun node->prev = node->next = NULL;
127*4882a593Smuzhiyun node_id = node->node_id;
128*4882a593Smuzhiyun
129*4882a593Smuzhiyun /* Enqueue the node for this CPU in the spinlock wait queue */
130*4882a593Smuzhiyun while (1) {
131*4882a593Smuzhiyun old = READ_ONCE(lp->lock);
132*4882a593Smuzhiyun if ((old & _Q_LOCK_CPU_MASK) == 0 &&
133*4882a593Smuzhiyun (old & _Q_LOCK_STEAL_MASK) != _Q_LOCK_STEAL_MASK) {
134*4882a593Smuzhiyun /*
135*4882a593Smuzhiyun * The lock is free but there may be waiters.
136*4882a593Smuzhiyun * With no waiters simply take the lock, if there
137*4882a593Smuzhiyun * are waiters try to steal the lock. The lock may
138*4882a593Smuzhiyun * be stolen three times before the next queued
139*4882a593Smuzhiyun * waiter will get the lock.
140*4882a593Smuzhiyun */
141*4882a593Smuzhiyun new = (old ? (old + _Q_LOCK_STEAL_ADD) : 0) | lockval;
142*4882a593Smuzhiyun if (__atomic_cmpxchg_bool(&lp->lock, old, new))
143*4882a593Smuzhiyun /* Got the lock */
144*4882a593Smuzhiyun goto out;
145*4882a593Smuzhiyun /* lock passing in progress */
146*4882a593Smuzhiyun continue;
147*4882a593Smuzhiyun }
148*4882a593Smuzhiyun /* Make the node of this CPU the new tail. */
149*4882a593Smuzhiyun new = node_id | (old & _Q_LOCK_MASK);
150*4882a593Smuzhiyun if (__atomic_cmpxchg_bool(&lp->lock, old, new))
151*4882a593Smuzhiyun break;
152*4882a593Smuzhiyun }
153*4882a593Smuzhiyun /* Set the 'next' pointer of the tail node in the queue */
154*4882a593Smuzhiyun tail_id = old & _Q_TAIL_MASK;
155*4882a593Smuzhiyun if (tail_id != 0) {
156*4882a593Smuzhiyun node->prev = arch_spin_decode_tail(tail_id);
157*4882a593Smuzhiyun WRITE_ONCE(node->prev->next, node);
158*4882a593Smuzhiyun }
159*4882a593Smuzhiyun
160*4882a593Smuzhiyun /* Pass the virtual CPU to the lock holder if it is not running */
161*4882a593Smuzhiyun owner = arch_spin_yield_target(old, node);
162*4882a593Smuzhiyun if (owner && arch_vcpu_is_preempted(owner - 1))
163*4882a593Smuzhiyun smp_yield_cpu(owner - 1);
164*4882a593Smuzhiyun
165*4882a593Smuzhiyun /* Spin on the CPU local node->prev pointer */
166*4882a593Smuzhiyun if (tail_id != 0) {
167*4882a593Smuzhiyun count = spin_retry;
168*4882a593Smuzhiyun while (READ_ONCE(node->prev) != NULL) {
169*4882a593Smuzhiyun if (count-- >= 0)
170*4882a593Smuzhiyun continue;
171*4882a593Smuzhiyun count = spin_retry;
172*4882a593Smuzhiyun /* Query running state of lock holder again. */
173*4882a593Smuzhiyun owner = arch_spin_yield_target(old, node);
174*4882a593Smuzhiyun if (owner && arch_vcpu_is_preempted(owner - 1))
175*4882a593Smuzhiyun smp_yield_cpu(owner - 1);
176*4882a593Smuzhiyun }
177*4882a593Smuzhiyun }
178*4882a593Smuzhiyun
179*4882a593Smuzhiyun /* Spin on the lock value in the spinlock_t */
180*4882a593Smuzhiyun count = spin_retry;
181*4882a593Smuzhiyun while (1) {
182*4882a593Smuzhiyun old = READ_ONCE(lp->lock);
183*4882a593Smuzhiyun owner = old & _Q_LOCK_CPU_MASK;
184*4882a593Smuzhiyun if (!owner) {
185*4882a593Smuzhiyun tail_id = old & _Q_TAIL_MASK;
186*4882a593Smuzhiyun new = ((tail_id != node_id) ? tail_id : 0) | lockval;
187*4882a593Smuzhiyun if (__atomic_cmpxchg_bool(&lp->lock, old, new))
188*4882a593Smuzhiyun /* Got the lock */
189*4882a593Smuzhiyun break;
190*4882a593Smuzhiyun continue;
191*4882a593Smuzhiyun }
192*4882a593Smuzhiyun if (count-- >= 0)
193*4882a593Smuzhiyun continue;
194*4882a593Smuzhiyun count = spin_retry;
195*4882a593Smuzhiyun if (!MACHINE_IS_LPAR || arch_vcpu_is_preempted(owner - 1))
196*4882a593Smuzhiyun smp_yield_cpu(owner - 1);
197*4882a593Smuzhiyun }
198*4882a593Smuzhiyun
199*4882a593Smuzhiyun /* Pass lock_spin job to next CPU in the queue */
200*4882a593Smuzhiyun if (node_id && tail_id != node_id) {
201*4882a593Smuzhiyun /* Wait until the next CPU has set up the 'next' pointer */
202*4882a593Smuzhiyun while ((next = READ_ONCE(node->next)) == NULL)
203*4882a593Smuzhiyun ;
204*4882a593Smuzhiyun next->prev = NULL;
205*4882a593Smuzhiyun }
206*4882a593Smuzhiyun
207*4882a593Smuzhiyun out:
208*4882a593Smuzhiyun S390_lowcore.spinlock_index--;
209*4882a593Smuzhiyun }
210*4882a593Smuzhiyun
arch_spin_lock_classic(arch_spinlock_t * lp)211*4882a593Smuzhiyun static inline void arch_spin_lock_classic(arch_spinlock_t *lp)
212*4882a593Smuzhiyun {
213*4882a593Smuzhiyun int lockval, old, new, owner, count;
214*4882a593Smuzhiyun
215*4882a593Smuzhiyun lockval = SPINLOCK_LOCKVAL; /* cpu + 1 */
216*4882a593Smuzhiyun
217*4882a593Smuzhiyun /* Pass the virtual CPU to the lock holder if it is not running */
218*4882a593Smuzhiyun owner = arch_spin_yield_target(READ_ONCE(lp->lock), NULL);
219*4882a593Smuzhiyun if (owner && arch_vcpu_is_preempted(owner - 1))
220*4882a593Smuzhiyun smp_yield_cpu(owner - 1);
221*4882a593Smuzhiyun
222*4882a593Smuzhiyun count = spin_retry;
223*4882a593Smuzhiyun while (1) {
224*4882a593Smuzhiyun old = arch_load_niai4(&lp->lock);
225*4882a593Smuzhiyun owner = old & _Q_LOCK_CPU_MASK;
226*4882a593Smuzhiyun /* Try to get the lock if it is free. */
227*4882a593Smuzhiyun if (!owner) {
228*4882a593Smuzhiyun new = (old & _Q_TAIL_MASK) | lockval;
229*4882a593Smuzhiyun if (arch_cmpxchg_niai8(&lp->lock, old, new)) {
230*4882a593Smuzhiyun /* Got the lock */
231*4882a593Smuzhiyun return;
232*4882a593Smuzhiyun }
233*4882a593Smuzhiyun continue;
234*4882a593Smuzhiyun }
235*4882a593Smuzhiyun if (count-- >= 0)
236*4882a593Smuzhiyun continue;
237*4882a593Smuzhiyun count = spin_retry;
238*4882a593Smuzhiyun if (!MACHINE_IS_LPAR || arch_vcpu_is_preempted(owner - 1))
239*4882a593Smuzhiyun smp_yield_cpu(owner - 1);
240*4882a593Smuzhiyun }
241*4882a593Smuzhiyun }
242*4882a593Smuzhiyun
arch_spin_lock_wait(arch_spinlock_t * lp)243*4882a593Smuzhiyun void arch_spin_lock_wait(arch_spinlock_t *lp)
244*4882a593Smuzhiyun {
245*4882a593Smuzhiyun if (test_cpu_flag(CIF_DEDICATED_CPU))
246*4882a593Smuzhiyun arch_spin_lock_queued(lp);
247*4882a593Smuzhiyun else
248*4882a593Smuzhiyun arch_spin_lock_classic(lp);
249*4882a593Smuzhiyun }
250*4882a593Smuzhiyun EXPORT_SYMBOL(arch_spin_lock_wait);
251*4882a593Smuzhiyun
arch_spin_trylock_retry(arch_spinlock_t * lp)252*4882a593Smuzhiyun int arch_spin_trylock_retry(arch_spinlock_t *lp)
253*4882a593Smuzhiyun {
254*4882a593Smuzhiyun int cpu = SPINLOCK_LOCKVAL;
255*4882a593Smuzhiyun int owner, count;
256*4882a593Smuzhiyun
257*4882a593Smuzhiyun for (count = spin_retry; count > 0; count--) {
258*4882a593Smuzhiyun owner = READ_ONCE(lp->lock);
259*4882a593Smuzhiyun /* Try to get the lock if it is free. */
260*4882a593Smuzhiyun if (!owner) {
261*4882a593Smuzhiyun if (__atomic_cmpxchg_bool(&lp->lock, 0, cpu))
262*4882a593Smuzhiyun return 1;
263*4882a593Smuzhiyun }
264*4882a593Smuzhiyun }
265*4882a593Smuzhiyun return 0;
266*4882a593Smuzhiyun }
267*4882a593Smuzhiyun EXPORT_SYMBOL(arch_spin_trylock_retry);
268*4882a593Smuzhiyun
arch_read_lock_wait(arch_rwlock_t * rw)269*4882a593Smuzhiyun void arch_read_lock_wait(arch_rwlock_t *rw)
270*4882a593Smuzhiyun {
271*4882a593Smuzhiyun if (unlikely(in_interrupt())) {
272*4882a593Smuzhiyun while (READ_ONCE(rw->cnts) & 0x10000)
273*4882a593Smuzhiyun barrier();
274*4882a593Smuzhiyun return;
275*4882a593Smuzhiyun }
276*4882a593Smuzhiyun
277*4882a593Smuzhiyun /* Remove this reader again to allow recursive read locking */
278*4882a593Smuzhiyun __atomic_add_const(-1, &rw->cnts);
279*4882a593Smuzhiyun /* Put the reader into the wait queue */
280*4882a593Smuzhiyun arch_spin_lock(&rw->wait);
281*4882a593Smuzhiyun /* Now add this reader to the count value again */
282*4882a593Smuzhiyun __atomic_add_const(1, &rw->cnts);
283*4882a593Smuzhiyun /* Loop until the writer is done */
284*4882a593Smuzhiyun while (READ_ONCE(rw->cnts) & 0x10000)
285*4882a593Smuzhiyun barrier();
286*4882a593Smuzhiyun arch_spin_unlock(&rw->wait);
287*4882a593Smuzhiyun }
288*4882a593Smuzhiyun EXPORT_SYMBOL(arch_read_lock_wait);
289*4882a593Smuzhiyun
arch_write_lock_wait(arch_rwlock_t * rw)290*4882a593Smuzhiyun void arch_write_lock_wait(arch_rwlock_t *rw)
291*4882a593Smuzhiyun {
292*4882a593Smuzhiyun int old;
293*4882a593Smuzhiyun
294*4882a593Smuzhiyun /* Add this CPU to the write waiters */
295*4882a593Smuzhiyun __atomic_add(0x20000, &rw->cnts);
296*4882a593Smuzhiyun
297*4882a593Smuzhiyun /* Put the writer into the wait queue */
298*4882a593Smuzhiyun arch_spin_lock(&rw->wait);
299*4882a593Smuzhiyun
300*4882a593Smuzhiyun while (1) {
301*4882a593Smuzhiyun old = READ_ONCE(rw->cnts);
302*4882a593Smuzhiyun if ((old & 0x1ffff) == 0 &&
303*4882a593Smuzhiyun __atomic_cmpxchg_bool(&rw->cnts, old, old | 0x10000))
304*4882a593Smuzhiyun /* Got the lock */
305*4882a593Smuzhiyun break;
306*4882a593Smuzhiyun barrier();
307*4882a593Smuzhiyun }
308*4882a593Smuzhiyun
309*4882a593Smuzhiyun arch_spin_unlock(&rw->wait);
310*4882a593Smuzhiyun }
311*4882a593Smuzhiyun EXPORT_SYMBOL(arch_write_lock_wait);
312*4882a593Smuzhiyun
arch_spin_relax(arch_spinlock_t * lp)313*4882a593Smuzhiyun void arch_spin_relax(arch_spinlock_t *lp)
314*4882a593Smuzhiyun {
315*4882a593Smuzhiyun int cpu;
316*4882a593Smuzhiyun
317*4882a593Smuzhiyun cpu = READ_ONCE(lp->lock) & _Q_LOCK_CPU_MASK;
318*4882a593Smuzhiyun if (!cpu)
319*4882a593Smuzhiyun return;
320*4882a593Smuzhiyun if (MACHINE_IS_LPAR && !arch_vcpu_is_preempted(cpu - 1))
321*4882a593Smuzhiyun return;
322*4882a593Smuzhiyun smp_yield_cpu(cpu - 1);
323*4882a593Smuzhiyun }
324*4882a593Smuzhiyun EXPORT_SYMBOL(arch_spin_relax);
325