1*4882a593Smuzhiyun /*
2*4882a593Smuzhiyun * bitops.h: Bit string operations on the ppc
3*4882a593Smuzhiyun */
4*4882a593Smuzhiyun
5*4882a593Smuzhiyun #ifndef _PPC_BITOPS_H
6*4882a593Smuzhiyun #define _PPC_BITOPS_H
7*4882a593Smuzhiyun
8*4882a593Smuzhiyun #include <asm/byteorder.h>
9*4882a593Smuzhiyun #include <asm-generic/bitops/__ffs.h>
10*4882a593Smuzhiyun
11*4882a593Smuzhiyun /*
12*4882a593Smuzhiyun * Arguably these bit operations don't imply any memory barrier or
13*4882a593Smuzhiyun * SMP ordering, but in fact a lot of drivers expect them to imply
14*4882a593Smuzhiyun * both, since they do on x86 cpus.
15*4882a593Smuzhiyun */
16*4882a593Smuzhiyun #ifdef CONFIG_SMP
17*4882a593Smuzhiyun #define SMP_WMB "eieio\n"
18*4882a593Smuzhiyun #define SMP_MB "\nsync"
19*4882a593Smuzhiyun #else
20*4882a593Smuzhiyun #define SMP_WMB
21*4882a593Smuzhiyun #define SMP_MB
22*4882a593Smuzhiyun #endif /* CONFIG_SMP */
23*4882a593Smuzhiyun
24*4882a593Smuzhiyun #define __INLINE_BITOPS 1
25*4882a593Smuzhiyun
26*4882a593Smuzhiyun #if __INLINE_BITOPS
27*4882a593Smuzhiyun /*
28*4882a593Smuzhiyun * These used to be if'd out here because using : "cc" as a constraint
29*4882a593Smuzhiyun * resulted in errors from egcs. Things may be OK with gcc-2.95.
30*4882a593Smuzhiyun */
set_bit(int nr,volatile void * addr)31*4882a593Smuzhiyun static __inline__ void set_bit(int nr, volatile void * addr)
32*4882a593Smuzhiyun {
33*4882a593Smuzhiyun unsigned long old;
34*4882a593Smuzhiyun unsigned long mask = 1 << (nr & 0x1f);
35*4882a593Smuzhiyun unsigned long *p = ((unsigned long *)addr) + (nr >> 5);
36*4882a593Smuzhiyun
37*4882a593Smuzhiyun __asm__ __volatile__(SMP_WMB "\
38*4882a593Smuzhiyun 1: lwarx %0,0,%3\n\
39*4882a593Smuzhiyun or %0,%0,%2\n\
40*4882a593Smuzhiyun stwcx. %0,0,%3\n\
41*4882a593Smuzhiyun bne 1b"
42*4882a593Smuzhiyun SMP_MB
43*4882a593Smuzhiyun : "=&r" (old), "=m" (*p)
44*4882a593Smuzhiyun : "r" (mask), "r" (p), "m" (*p)
45*4882a593Smuzhiyun : "cc" );
46*4882a593Smuzhiyun }
47*4882a593Smuzhiyun
clear_bit(int nr,volatile void * addr)48*4882a593Smuzhiyun static __inline__ void clear_bit(int nr, volatile void *addr)
49*4882a593Smuzhiyun {
50*4882a593Smuzhiyun unsigned long old;
51*4882a593Smuzhiyun unsigned long mask = 1 << (nr & 0x1f);
52*4882a593Smuzhiyun unsigned long *p = ((unsigned long *)addr) + (nr >> 5);
53*4882a593Smuzhiyun
54*4882a593Smuzhiyun __asm__ __volatile__(SMP_WMB "\
55*4882a593Smuzhiyun 1: lwarx %0,0,%3\n\
56*4882a593Smuzhiyun andc %0,%0,%2\n\
57*4882a593Smuzhiyun stwcx. %0,0,%3\n\
58*4882a593Smuzhiyun bne 1b"
59*4882a593Smuzhiyun SMP_MB
60*4882a593Smuzhiyun : "=&r" (old), "=m" (*p)
61*4882a593Smuzhiyun : "r" (mask), "r" (p), "m" (*p)
62*4882a593Smuzhiyun : "cc");
63*4882a593Smuzhiyun }
64*4882a593Smuzhiyun
change_bit(int nr,volatile void * addr)65*4882a593Smuzhiyun static __inline__ void change_bit(int nr, volatile void *addr)
66*4882a593Smuzhiyun {
67*4882a593Smuzhiyun unsigned long old;
68*4882a593Smuzhiyun unsigned long mask = 1 << (nr & 0x1f);
69*4882a593Smuzhiyun unsigned long *p = ((unsigned long *)addr) + (nr >> 5);
70*4882a593Smuzhiyun
71*4882a593Smuzhiyun __asm__ __volatile__(SMP_WMB "\
72*4882a593Smuzhiyun 1: lwarx %0,0,%3\n\
73*4882a593Smuzhiyun xor %0,%0,%2\n\
74*4882a593Smuzhiyun stwcx. %0,0,%3\n\
75*4882a593Smuzhiyun bne 1b"
76*4882a593Smuzhiyun SMP_MB
77*4882a593Smuzhiyun : "=&r" (old), "=m" (*p)
78*4882a593Smuzhiyun : "r" (mask), "r" (p), "m" (*p)
79*4882a593Smuzhiyun : "cc");
80*4882a593Smuzhiyun }
81*4882a593Smuzhiyun
test_and_set_bit(int nr,volatile void * addr)82*4882a593Smuzhiyun static __inline__ int test_and_set_bit(int nr, volatile void *addr)
83*4882a593Smuzhiyun {
84*4882a593Smuzhiyun unsigned int old, t;
85*4882a593Smuzhiyun unsigned int mask = 1 << (nr & 0x1f);
86*4882a593Smuzhiyun volatile unsigned int *p = ((volatile unsigned int *)addr) + (nr >> 5);
87*4882a593Smuzhiyun
88*4882a593Smuzhiyun __asm__ __volatile__(SMP_WMB "\
89*4882a593Smuzhiyun 1: lwarx %0,0,%4\n\
90*4882a593Smuzhiyun or %1,%0,%3\n\
91*4882a593Smuzhiyun stwcx. %1,0,%4\n\
92*4882a593Smuzhiyun bne 1b"
93*4882a593Smuzhiyun SMP_MB
94*4882a593Smuzhiyun : "=&r" (old), "=&r" (t), "=m" (*p)
95*4882a593Smuzhiyun : "r" (mask), "r" (p), "m" (*p)
96*4882a593Smuzhiyun : "cc");
97*4882a593Smuzhiyun
98*4882a593Smuzhiyun return (old & mask) != 0;
99*4882a593Smuzhiyun }
100*4882a593Smuzhiyun
test_and_clear_bit(int nr,volatile void * addr)101*4882a593Smuzhiyun static __inline__ int test_and_clear_bit(int nr, volatile void *addr)
102*4882a593Smuzhiyun {
103*4882a593Smuzhiyun unsigned int old, t;
104*4882a593Smuzhiyun unsigned int mask = 1 << (nr & 0x1f);
105*4882a593Smuzhiyun volatile unsigned int *p = ((volatile unsigned int *)addr) + (nr >> 5);
106*4882a593Smuzhiyun
107*4882a593Smuzhiyun __asm__ __volatile__(SMP_WMB "\
108*4882a593Smuzhiyun 1: lwarx %0,0,%4\n\
109*4882a593Smuzhiyun andc %1,%0,%3\n\
110*4882a593Smuzhiyun stwcx. %1,0,%4\n\
111*4882a593Smuzhiyun bne 1b"
112*4882a593Smuzhiyun SMP_MB
113*4882a593Smuzhiyun : "=&r" (old), "=&r" (t), "=m" (*p)
114*4882a593Smuzhiyun : "r" (mask), "r" (p), "m" (*p)
115*4882a593Smuzhiyun : "cc");
116*4882a593Smuzhiyun
117*4882a593Smuzhiyun return (old & mask) != 0;
118*4882a593Smuzhiyun }
119*4882a593Smuzhiyun
test_and_change_bit(int nr,volatile void * addr)120*4882a593Smuzhiyun static __inline__ int test_and_change_bit(int nr, volatile void *addr)
121*4882a593Smuzhiyun {
122*4882a593Smuzhiyun unsigned int old, t;
123*4882a593Smuzhiyun unsigned int mask = 1 << (nr & 0x1f);
124*4882a593Smuzhiyun volatile unsigned int *p = ((volatile unsigned int *)addr) + (nr >> 5);
125*4882a593Smuzhiyun
126*4882a593Smuzhiyun __asm__ __volatile__(SMP_WMB "\
127*4882a593Smuzhiyun 1: lwarx %0,0,%4\n\
128*4882a593Smuzhiyun xor %1,%0,%3\n\
129*4882a593Smuzhiyun stwcx. %1,0,%4\n\
130*4882a593Smuzhiyun bne 1b"
131*4882a593Smuzhiyun SMP_MB
132*4882a593Smuzhiyun : "=&r" (old), "=&r" (t), "=m" (*p)
133*4882a593Smuzhiyun : "r" (mask), "r" (p), "m" (*p)
134*4882a593Smuzhiyun : "cc");
135*4882a593Smuzhiyun
136*4882a593Smuzhiyun return (old & mask) != 0;
137*4882a593Smuzhiyun }
138*4882a593Smuzhiyun #endif /* __INLINE_BITOPS */
139*4882a593Smuzhiyun
test_bit(int nr,__const__ volatile void * addr)140*4882a593Smuzhiyun static __inline__ int test_bit(int nr, __const__ volatile void *addr)
141*4882a593Smuzhiyun {
142*4882a593Smuzhiyun __const__ unsigned int *p = (__const__ unsigned int *) addr;
143*4882a593Smuzhiyun
144*4882a593Smuzhiyun return ((p[nr >> 5] >> (nr & 0x1f)) & 1) != 0;
145*4882a593Smuzhiyun }
146*4882a593Smuzhiyun
147*4882a593Smuzhiyun /* Return the bit position of the most significant 1 bit in a word */
148*4882a593Smuzhiyun /* - the result is undefined when x == 0 */
__ilog2(unsigned int x)149*4882a593Smuzhiyun static __inline__ int __ilog2(unsigned int x)
150*4882a593Smuzhiyun {
151*4882a593Smuzhiyun int lz;
152*4882a593Smuzhiyun
153*4882a593Smuzhiyun asm ("cntlzw %0,%1" : "=r" (lz) : "r" (x));
154*4882a593Smuzhiyun return 31 - lz;
155*4882a593Smuzhiyun }
156*4882a593Smuzhiyun
ffz(unsigned int x)157*4882a593Smuzhiyun static __inline__ int ffz(unsigned int x)
158*4882a593Smuzhiyun {
159*4882a593Smuzhiyun if ((x = ~x) == 0)
160*4882a593Smuzhiyun return 32;
161*4882a593Smuzhiyun return __ilog2(x & -x);
162*4882a593Smuzhiyun }
163*4882a593Smuzhiyun
164*4882a593Smuzhiyun /*
165*4882a593Smuzhiyun * fls: find last (most-significant) bit set.
166*4882a593Smuzhiyun * Note fls(0) = 0, fls(1) = 1, fls(0x80000000) = 32.
167*4882a593Smuzhiyun *
168*4882a593Smuzhiyun * On powerpc, __ilog2(0) returns -1, but this is not safe in general
169*4882a593Smuzhiyun */
fls(unsigned int x)170*4882a593Smuzhiyun static __inline__ int fls(unsigned int x)
171*4882a593Smuzhiyun {
172*4882a593Smuzhiyun return __ilog2(x) + 1;
173*4882a593Smuzhiyun }
174*4882a593Smuzhiyun #define PLATFORM_FLS
175*4882a593Smuzhiyun
176*4882a593Smuzhiyun /**
177*4882a593Smuzhiyun * fls64 - find last set bit in a 64-bit word
178*4882a593Smuzhiyun * @x: the word to search
179*4882a593Smuzhiyun *
180*4882a593Smuzhiyun * This is defined in a similar way as the libc and compiler builtin
181*4882a593Smuzhiyun * ffsll, but returns the position of the most significant set bit.
182*4882a593Smuzhiyun *
183*4882a593Smuzhiyun * fls64(value) returns 0 if value is 0 or the position of the last
184*4882a593Smuzhiyun * set bit if value is nonzero. The last (most significant) bit is
185*4882a593Smuzhiyun * at position 64.
186*4882a593Smuzhiyun */
187*4882a593Smuzhiyun #if BITS_PER_LONG == 32
fls64(__u64 x)188*4882a593Smuzhiyun static inline int fls64(__u64 x)
189*4882a593Smuzhiyun {
190*4882a593Smuzhiyun __u32 h = x >> 32;
191*4882a593Smuzhiyun if (h)
192*4882a593Smuzhiyun return fls(h) + 32;
193*4882a593Smuzhiyun return fls(x);
194*4882a593Smuzhiyun }
195*4882a593Smuzhiyun #elif BITS_PER_LONG == 64
fls64(__u64 x)196*4882a593Smuzhiyun static inline int fls64(__u64 x)
197*4882a593Smuzhiyun {
198*4882a593Smuzhiyun if (x == 0)
199*4882a593Smuzhiyun return 0;
200*4882a593Smuzhiyun return __ilog2(x) + 1;
201*4882a593Smuzhiyun }
202*4882a593Smuzhiyun #else
203*4882a593Smuzhiyun #error BITS_PER_LONG not 32 or 64
204*4882a593Smuzhiyun #endif
205*4882a593Smuzhiyun
206*4882a593Smuzhiyun #ifdef __KERNEL__
207*4882a593Smuzhiyun
208*4882a593Smuzhiyun /*
209*4882a593Smuzhiyun * ffs: find first bit set. This is defined the same way as
210*4882a593Smuzhiyun * the libc and compiler builtin ffs routines, therefore
211*4882a593Smuzhiyun * differs in spirit from the above ffz (man ffs).
212*4882a593Smuzhiyun */
ffs(int x)213*4882a593Smuzhiyun static __inline__ int ffs(int x)
214*4882a593Smuzhiyun {
215*4882a593Smuzhiyun return __ilog2(x & -x) + 1;
216*4882a593Smuzhiyun }
217*4882a593Smuzhiyun #define PLATFORM_FFS
218*4882a593Smuzhiyun
219*4882a593Smuzhiyun /*
220*4882a593Smuzhiyun * hweightN: returns the hamming weight (i.e. the number
221*4882a593Smuzhiyun * of bits set) of a N-bit word
222*4882a593Smuzhiyun */
223*4882a593Smuzhiyun
224*4882a593Smuzhiyun #define hweight32(x) generic_hweight32(x)
225*4882a593Smuzhiyun #define hweight16(x) generic_hweight16(x)
226*4882a593Smuzhiyun #define hweight8(x) generic_hweight8(x)
227*4882a593Smuzhiyun
228*4882a593Smuzhiyun #endif /* __KERNEL__ */
229*4882a593Smuzhiyun
230*4882a593Smuzhiyun /*
231*4882a593Smuzhiyun * This implementation of find_{first,next}_zero_bit was stolen from
232*4882a593Smuzhiyun * Linus' asm-alpha/bitops.h.
233*4882a593Smuzhiyun */
234*4882a593Smuzhiyun #define find_first_zero_bit(addr, size) \
235*4882a593Smuzhiyun find_next_zero_bit((addr), (size), 0)
236*4882a593Smuzhiyun
find_next_zero_bit(void * addr,unsigned long size,unsigned long offset)237*4882a593Smuzhiyun static __inline__ unsigned long find_next_zero_bit(void * addr,
238*4882a593Smuzhiyun unsigned long size, unsigned long offset)
239*4882a593Smuzhiyun {
240*4882a593Smuzhiyun unsigned int * p = ((unsigned int *) addr) + (offset >> 5);
241*4882a593Smuzhiyun unsigned int result = offset & ~31UL;
242*4882a593Smuzhiyun unsigned int tmp;
243*4882a593Smuzhiyun
244*4882a593Smuzhiyun if (offset >= size)
245*4882a593Smuzhiyun return size;
246*4882a593Smuzhiyun size -= result;
247*4882a593Smuzhiyun offset &= 31UL;
248*4882a593Smuzhiyun if (offset) {
249*4882a593Smuzhiyun tmp = *p++;
250*4882a593Smuzhiyun tmp |= ~0UL >> (32-offset);
251*4882a593Smuzhiyun if (size < 32)
252*4882a593Smuzhiyun goto found_first;
253*4882a593Smuzhiyun if (tmp != ~0U)
254*4882a593Smuzhiyun goto found_middle;
255*4882a593Smuzhiyun size -= 32;
256*4882a593Smuzhiyun result += 32;
257*4882a593Smuzhiyun }
258*4882a593Smuzhiyun while (size >= 32) {
259*4882a593Smuzhiyun if ((tmp = *p++) != ~0U)
260*4882a593Smuzhiyun goto found_middle;
261*4882a593Smuzhiyun result += 32;
262*4882a593Smuzhiyun size -= 32;
263*4882a593Smuzhiyun }
264*4882a593Smuzhiyun if (!size)
265*4882a593Smuzhiyun return result;
266*4882a593Smuzhiyun tmp = *p;
267*4882a593Smuzhiyun found_first:
268*4882a593Smuzhiyun tmp |= ~0UL << size;
269*4882a593Smuzhiyun found_middle:
270*4882a593Smuzhiyun return result + ffz(tmp);
271*4882a593Smuzhiyun }
272*4882a593Smuzhiyun
273*4882a593Smuzhiyun
274*4882a593Smuzhiyun #define _EXT2_HAVE_ASM_BITOPS_
275*4882a593Smuzhiyun
276*4882a593Smuzhiyun #ifdef __KERNEL__
277*4882a593Smuzhiyun /*
278*4882a593Smuzhiyun * test_and_{set,clear}_bit guarantee atomicity without
279*4882a593Smuzhiyun * disabling interrupts.
280*4882a593Smuzhiyun */
281*4882a593Smuzhiyun #define ext2_set_bit(nr, addr) test_and_set_bit((nr) ^ 0x18, addr)
282*4882a593Smuzhiyun #define ext2_clear_bit(nr, addr) test_and_clear_bit((nr) ^ 0x18, addr)
283*4882a593Smuzhiyun
284*4882a593Smuzhiyun #else
ext2_set_bit(int nr,void * addr)285*4882a593Smuzhiyun static __inline__ int ext2_set_bit(int nr, void * addr)
286*4882a593Smuzhiyun {
287*4882a593Smuzhiyun int mask;
288*4882a593Smuzhiyun unsigned char *ADDR = (unsigned char *) addr;
289*4882a593Smuzhiyun int oldbit;
290*4882a593Smuzhiyun
291*4882a593Smuzhiyun ADDR += nr >> 3;
292*4882a593Smuzhiyun mask = 1 << (nr & 0x07);
293*4882a593Smuzhiyun oldbit = (*ADDR & mask) ? 1 : 0;
294*4882a593Smuzhiyun *ADDR |= mask;
295*4882a593Smuzhiyun return oldbit;
296*4882a593Smuzhiyun }
297*4882a593Smuzhiyun
ext2_clear_bit(int nr,void * addr)298*4882a593Smuzhiyun static __inline__ int ext2_clear_bit(int nr, void * addr)
299*4882a593Smuzhiyun {
300*4882a593Smuzhiyun int mask;
301*4882a593Smuzhiyun unsigned char *ADDR = (unsigned char *) addr;
302*4882a593Smuzhiyun int oldbit;
303*4882a593Smuzhiyun
304*4882a593Smuzhiyun ADDR += nr >> 3;
305*4882a593Smuzhiyun mask = 1 << (nr & 0x07);
306*4882a593Smuzhiyun oldbit = (*ADDR & mask) ? 1 : 0;
307*4882a593Smuzhiyun *ADDR = *ADDR & ~mask;
308*4882a593Smuzhiyun return oldbit;
309*4882a593Smuzhiyun }
310*4882a593Smuzhiyun #endif /* __KERNEL__ */
311*4882a593Smuzhiyun
ext2_test_bit(int nr,__const__ void * addr)312*4882a593Smuzhiyun static __inline__ int ext2_test_bit(int nr, __const__ void * addr)
313*4882a593Smuzhiyun {
314*4882a593Smuzhiyun __const__ unsigned char *ADDR = (__const__ unsigned char *) addr;
315*4882a593Smuzhiyun
316*4882a593Smuzhiyun return (ADDR[nr >> 3] >> (nr & 7)) & 1;
317*4882a593Smuzhiyun }
318*4882a593Smuzhiyun
319*4882a593Smuzhiyun /*
320*4882a593Smuzhiyun * This implementation of ext2_find_{first,next}_zero_bit was stolen from
321*4882a593Smuzhiyun * Linus' asm-alpha/bitops.h and modified for a big-endian machine.
322*4882a593Smuzhiyun */
323*4882a593Smuzhiyun
324*4882a593Smuzhiyun #define ext2_find_first_zero_bit(addr, size) \
325*4882a593Smuzhiyun ext2_find_next_zero_bit((addr), (size), 0)
326*4882a593Smuzhiyun
ext2_find_next_zero_bit(void * addr,unsigned long size,unsigned long offset)327*4882a593Smuzhiyun static __inline__ unsigned long ext2_find_next_zero_bit(void *addr,
328*4882a593Smuzhiyun unsigned long size, unsigned long offset)
329*4882a593Smuzhiyun {
330*4882a593Smuzhiyun unsigned int *p = ((unsigned int *) addr) + (offset >> 5);
331*4882a593Smuzhiyun unsigned int result = offset & ~31UL;
332*4882a593Smuzhiyun unsigned int tmp;
333*4882a593Smuzhiyun
334*4882a593Smuzhiyun if (offset >= size)
335*4882a593Smuzhiyun return size;
336*4882a593Smuzhiyun size -= result;
337*4882a593Smuzhiyun offset &= 31UL;
338*4882a593Smuzhiyun if (offset) {
339*4882a593Smuzhiyun tmp = cpu_to_le32p(p++);
340*4882a593Smuzhiyun tmp |= ~0UL >> (32-offset);
341*4882a593Smuzhiyun if (size < 32)
342*4882a593Smuzhiyun goto found_first;
343*4882a593Smuzhiyun if (tmp != ~0U)
344*4882a593Smuzhiyun goto found_middle;
345*4882a593Smuzhiyun size -= 32;
346*4882a593Smuzhiyun result += 32;
347*4882a593Smuzhiyun }
348*4882a593Smuzhiyun while (size >= 32) {
349*4882a593Smuzhiyun if ((tmp = cpu_to_le32p(p++)) != ~0U)
350*4882a593Smuzhiyun goto found_middle;
351*4882a593Smuzhiyun result += 32;
352*4882a593Smuzhiyun size -= 32;
353*4882a593Smuzhiyun }
354*4882a593Smuzhiyun if (!size)
355*4882a593Smuzhiyun return result;
356*4882a593Smuzhiyun tmp = cpu_to_le32p(p);
357*4882a593Smuzhiyun found_first:
358*4882a593Smuzhiyun tmp |= ~0U << size;
359*4882a593Smuzhiyun found_middle:
360*4882a593Smuzhiyun return result + ffz(tmp);
361*4882a593Smuzhiyun }
362*4882a593Smuzhiyun
363*4882a593Smuzhiyun /* Bitmap functions for the minix filesystem. */
364*4882a593Smuzhiyun #define minix_test_and_set_bit(nr,addr) ext2_set_bit(nr,addr)
365*4882a593Smuzhiyun #define minix_set_bit(nr,addr) ((void)ext2_set_bit(nr,addr))
366*4882a593Smuzhiyun #define minix_test_and_clear_bit(nr,addr) ext2_clear_bit(nr,addr)
367*4882a593Smuzhiyun #define minix_test_bit(nr,addr) ext2_test_bit(nr,addr)
368*4882a593Smuzhiyun #define minix_find_first_zero_bit(addr,size) ext2_find_first_zero_bit(addr,size)
369*4882a593Smuzhiyun
370*4882a593Smuzhiyun #endif /* _PPC_BITOPS_H */
371