xref: /openbmc/u-boot/arch/powerpc/include/asm/bitops.h (revision 20c700f8)
1 /*
2  * bitops.h: Bit string operations on the ppc
3  */
4 
5 #ifndef _PPC_BITOPS_H
6 #define _PPC_BITOPS_H
7 
8 #include <asm/byteorder.h>
9 #include <asm-generic/bitops/__ffs.h>
10 
11 /*
12  * Arguably these bit operations don't imply any memory barrier or
13  * SMP ordering, but in fact a lot of drivers expect them to imply
14  * both, since they do on x86 cpus.
15  */
16 #ifdef CONFIG_SMP
17 #define SMP_WMB		"eieio\n"
18 #define SMP_MB		"\nsync"
19 #else
20 #define SMP_WMB
21 #define SMP_MB
22 #endif /* CONFIG_SMP */
23 
24 #define __INLINE_BITOPS	1
25 
26 #if __INLINE_BITOPS
27 /*
28  * These used to be if'd out here because using : "cc" as a constraint
29  * resulted in errors from egcs.  Things may be OK with gcc-2.95.
30  */
31 static __inline__ void set_bit(int nr, volatile void * addr)
32 {
33 	unsigned long old;
34 	unsigned long mask = 1 << (nr & 0x1f);
35 	unsigned long *p = ((unsigned long *)addr) + (nr >> 5);
36 
37 	__asm__ __volatile__(SMP_WMB "\
38 1:	lwarx	%0,0,%3\n\
39 	or	%0,%0,%2\n\
40 	stwcx.	%0,0,%3\n\
41 	bne	1b"
42 	SMP_MB
43 	: "=&r" (old), "=m" (*p)
44 	: "r" (mask), "r" (p), "m" (*p)
45 	: "cc" );
46 }
47 
48 static __inline__ void clear_bit(int nr, volatile void *addr)
49 {
50 	unsigned long old;
51 	unsigned long mask = 1 << (nr & 0x1f);
52 	unsigned long *p = ((unsigned long *)addr) + (nr >> 5);
53 
54 	__asm__ __volatile__(SMP_WMB "\
55 1:	lwarx	%0,0,%3\n\
56 	andc	%0,%0,%2\n\
57 	stwcx.	%0,0,%3\n\
58 	bne	1b"
59 	SMP_MB
60 	: "=&r" (old), "=m" (*p)
61 	: "r" (mask), "r" (p), "m" (*p)
62 	: "cc");
63 }
64 
65 static __inline__ void change_bit(int nr, volatile void *addr)
66 {
67 	unsigned long old;
68 	unsigned long mask = 1 << (nr & 0x1f);
69 	unsigned long *p = ((unsigned long *)addr) + (nr >> 5);
70 
71 	__asm__ __volatile__(SMP_WMB "\
72 1:	lwarx	%0,0,%3\n\
73 	xor	%0,%0,%2\n\
74 	stwcx.	%0,0,%3\n\
75 	bne	1b"
76 	SMP_MB
77 	: "=&r" (old), "=m" (*p)
78 	: "r" (mask), "r" (p), "m" (*p)
79 	: "cc");
80 }
81 
82 static __inline__ int test_and_set_bit(int nr, volatile void *addr)
83 {
84 	unsigned int old, t;
85 	unsigned int mask = 1 << (nr & 0x1f);
86 	volatile unsigned int *p = ((volatile unsigned int *)addr) + (nr >> 5);
87 
88 	__asm__ __volatile__(SMP_WMB "\
89 1:	lwarx	%0,0,%4\n\
90 	or	%1,%0,%3\n\
91 	stwcx.	%1,0,%4\n\
92 	bne	1b"
93 	SMP_MB
94 	: "=&r" (old), "=&r" (t), "=m" (*p)
95 	: "r" (mask), "r" (p), "m" (*p)
96 	: "cc");
97 
98 	return (old & mask) != 0;
99 }
100 
101 static __inline__ int test_and_clear_bit(int nr, volatile void *addr)
102 {
103 	unsigned int old, t;
104 	unsigned int mask = 1 << (nr & 0x1f);
105 	volatile unsigned int *p = ((volatile unsigned int *)addr) + (nr >> 5);
106 
107 	__asm__ __volatile__(SMP_WMB "\
108 1:	lwarx	%0,0,%4\n\
109 	andc	%1,%0,%3\n\
110 	stwcx.	%1,0,%4\n\
111 	bne	1b"
112 	SMP_MB
113 	: "=&r" (old), "=&r" (t), "=m" (*p)
114 	: "r" (mask), "r" (p), "m" (*p)
115 	: "cc");
116 
117 	return (old & mask) != 0;
118 }
119 
120 static __inline__ int test_and_change_bit(int nr, volatile void *addr)
121 {
122 	unsigned int old, t;
123 	unsigned int mask = 1 << (nr & 0x1f);
124 	volatile unsigned int *p = ((volatile unsigned int *)addr) + (nr >> 5);
125 
126 	__asm__ __volatile__(SMP_WMB "\
127 1:	lwarx	%0,0,%4\n\
128 	xor	%1,%0,%3\n\
129 	stwcx.	%1,0,%4\n\
130 	bne	1b"
131 	SMP_MB
132 	: "=&r" (old), "=&r" (t), "=m" (*p)
133 	: "r" (mask), "r" (p), "m" (*p)
134 	: "cc");
135 
136 	return (old & mask) != 0;
137 }
138 #endif /* __INLINE_BITOPS */
139 
140 static __inline__ int test_bit(int nr, __const__ volatile void *addr)
141 {
142 	__const__ unsigned int *p = (__const__ unsigned int *) addr;
143 
144 	return ((p[nr >> 5] >> (nr & 0x1f)) & 1) != 0;
145 }
146 
147 /* Return the bit position of the most significant 1 bit in a word */
148 /* - the result is undefined when x == 0 */
149 static __inline__ int __ilog2(unsigned int x)
150 {
151 	int lz;
152 
153 	asm ("cntlzw %0,%1" : "=r" (lz) : "r" (x));
154 	return 31 - lz;
155 }
156 
157 static __inline__ int ffz(unsigned int x)
158 {
159 	if ((x = ~x) == 0)
160 		return 32;
161 	return __ilog2(x & -x);
162 }
163 
164 /*
165  * fls: find last (most-significant) bit set.
166  * Note fls(0) = 0, fls(1) = 1, fls(0x80000000) = 32.
167  *
168  * On powerpc, __ilog2(0) returns -1, but this is not safe in general
169  */
170 static __inline__ int fls(unsigned int x)
171 {
172 	return __ilog2(x) + 1;
173 }
174 #define PLATFORM_FLS
175 
176 /**
177  * fls64 - find last set bit in a 64-bit word
178  * @x: the word to search
179  *
180  * This is defined in a similar way as the libc and compiler builtin
181  * ffsll, but returns the position of the most significant set bit.
182  *
183  * fls64(value) returns 0 if value is 0 or the position of the last
184  * set bit if value is nonzero. The last (most significant) bit is
185  * at position 64.
186  */
187 #if BITS_PER_LONG == 32
188 static inline int fls64(__u64 x)
189 {
190 	__u32 h = x >> 32;
191 	if (h)
192 		return fls(h) + 32;
193 	return fls(x);
194 }
195 #elif BITS_PER_LONG == 64
196 static inline int fls64(__u64 x)
197 {
198 	if (x == 0)
199 		return 0;
200 	return __ilog2(x) + 1;
201 }
202 #else
203 #error BITS_PER_LONG not 32 or 64
204 #endif
205 
206 #ifdef __KERNEL__
207 
208 /*
209  * ffs: find first bit set. This is defined the same way as
210  * the libc and compiler builtin ffs routines, therefore
211  * differs in spirit from the above ffz (man ffs).
212  */
213 static __inline__ int ffs(int x)
214 {
215 	return __ilog2(x & -x) + 1;
216 }
217 #define PLATFORM_FFS
218 
219 /*
220  * hweightN: returns the hamming weight (i.e. the number
221  * of bits set) of a N-bit word
222  */
223 
224 #define hweight32(x) generic_hweight32(x)
225 #define hweight16(x) generic_hweight16(x)
226 #define hweight8(x) generic_hweight8(x)
227 
228 #endif /* __KERNEL__ */
229 
230 /*
231  * This implementation of find_{first,next}_zero_bit was stolen from
232  * Linus' asm-alpha/bitops.h.
233  */
234 #define find_first_zero_bit(addr, size) \
235 	find_next_zero_bit((addr), (size), 0)
236 
237 static __inline__ unsigned long find_next_zero_bit(void * addr,
238 	unsigned long size, unsigned long offset)
239 {
240 	unsigned int * p = ((unsigned int *) addr) + (offset >> 5);
241 	unsigned int result = offset & ~31UL;
242 	unsigned int tmp;
243 
244 	if (offset >= size)
245 		return size;
246 	size -= result;
247 	offset &= 31UL;
248 	if (offset) {
249 		tmp = *p++;
250 		tmp |= ~0UL >> (32-offset);
251 		if (size < 32)
252 			goto found_first;
253 		if (tmp != ~0U)
254 			goto found_middle;
255 		size -= 32;
256 		result += 32;
257 	}
258 	while (size >= 32) {
259 		if ((tmp = *p++) != ~0U)
260 			goto found_middle;
261 		result += 32;
262 		size -= 32;
263 	}
264 	if (!size)
265 		return result;
266 	tmp = *p;
267 found_first:
268 	tmp |= ~0UL << size;
269 found_middle:
270 	return result + ffz(tmp);
271 }
272 
273 
274 #define _EXT2_HAVE_ASM_BITOPS_
275 
276 #ifdef __KERNEL__
277 /*
278  * test_and_{set,clear}_bit guarantee atomicity without
279  * disabling interrupts.
280  */
281 #define ext2_set_bit(nr, addr)		test_and_set_bit((nr) ^ 0x18, addr)
282 #define ext2_clear_bit(nr, addr)	test_and_clear_bit((nr) ^ 0x18, addr)
283 
284 #else
285 static __inline__ int ext2_set_bit(int nr, void * addr)
286 {
287 	int		mask;
288 	unsigned char	*ADDR = (unsigned char *) addr;
289 	int oldbit;
290 
291 	ADDR += nr >> 3;
292 	mask = 1 << (nr & 0x07);
293 	oldbit = (*ADDR & mask) ? 1 : 0;
294 	*ADDR |= mask;
295 	return oldbit;
296 }
297 
298 static __inline__ int ext2_clear_bit(int nr, void * addr)
299 {
300 	int		mask;
301 	unsigned char	*ADDR = (unsigned char *) addr;
302 	int oldbit;
303 
304 	ADDR += nr >> 3;
305 	mask = 1 << (nr & 0x07);
306 	oldbit = (*ADDR & mask) ? 1 : 0;
307 	*ADDR = *ADDR & ~mask;
308 	return oldbit;
309 }
310 #endif	/* __KERNEL__ */
311 
312 static __inline__ int ext2_test_bit(int nr, __const__ void * addr)
313 {
314 	__const__ unsigned char	*ADDR = (__const__ unsigned char *) addr;
315 
316 	return (ADDR[nr >> 3] >> (nr & 7)) & 1;
317 }
318 
319 /*
320  * This implementation of ext2_find_{first,next}_zero_bit was stolen from
321  * Linus' asm-alpha/bitops.h and modified for a big-endian machine.
322  */
323 
324 #define ext2_find_first_zero_bit(addr, size) \
325 	ext2_find_next_zero_bit((addr), (size), 0)
326 
327 static __inline__ unsigned long ext2_find_next_zero_bit(void *addr,
328 	unsigned long size, unsigned long offset)
329 {
330 	unsigned int *p = ((unsigned int *) addr) + (offset >> 5);
331 	unsigned int result = offset & ~31UL;
332 	unsigned int tmp;
333 
334 	if (offset >= size)
335 		return size;
336 	size -= result;
337 	offset &= 31UL;
338 	if (offset) {
339 		tmp = cpu_to_le32p(p++);
340 		tmp |= ~0UL >> (32-offset);
341 		if (size < 32)
342 			goto found_first;
343 		if (tmp != ~0U)
344 			goto found_middle;
345 		size -= 32;
346 		result += 32;
347 	}
348 	while (size >= 32) {
349 		if ((tmp = cpu_to_le32p(p++)) != ~0U)
350 			goto found_middle;
351 		result += 32;
352 		size -= 32;
353 	}
354 	if (!size)
355 		return result;
356 	tmp = cpu_to_le32p(p);
357 found_first:
358 	tmp |= ~0U << size;
359 found_middle:
360 	return result + ffz(tmp);
361 }
362 
363 /* Bitmap functions for the minix filesystem.  */
364 #define minix_test_and_set_bit(nr,addr) ext2_set_bit(nr,addr)
365 #define minix_set_bit(nr,addr) ((void)ext2_set_bit(nr,addr))
366 #define minix_test_and_clear_bit(nr,addr) ext2_clear_bit(nr,addr)
367 #define minix_test_bit(nr,addr) ext2_test_bit(nr,addr)
368 #define minix_find_first_zero_bit(addr,size) ext2_find_first_zero_bit(addr,size)
369 
370 #endif /* _PPC_BITOPS_H */
371