xref: /openbmc/linux/arch/sh/include/asm/atomic-irq.h (revision 25b892b5)
1 /* SPDX-License-Identifier: GPL-2.0 */
2 #ifndef __ASM_SH_ATOMIC_IRQ_H
3 #define __ASM_SH_ATOMIC_IRQ_H
4 
5 #include <linux/irqflags.h>
6 
7 /*
8  * To get proper branch prediction for the main line, we must branch
9  * forward to code at the end of this object's .text section, then
10  * branch back to restart the operation.
11  */
12 
13 #define ATOMIC_OP(op, c_op)						\
14 static inline void arch_atomic_##op(int i, atomic_t *v)			\
15 {									\
16 	unsigned long flags;						\
17 									\
18 	raw_local_irq_save(flags);					\
19 	v->counter c_op i;						\
20 	raw_local_irq_restore(flags);					\
21 }
22 
23 #define ATOMIC_OP_RETURN(op, c_op)					\
24 static inline int arch_atomic_##op##_return(int i, atomic_t *v)		\
25 {									\
26 	unsigned long temp, flags;					\
27 									\
28 	raw_local_irq_save(flags);					\
29 	temp = v->counter;						\
30 	temp c_op i;							\
31 	v->counter = temp;						\
32 	raw_local_irq_restore(flags);					\
33 									\
34 	return temp;							\
35 }
36 
37 #define ATOMIC_FETCH_OP(op, c_op)					\
38 static inline int arch_atomic_fetch_##op(int i, atomic_t *v)		\
39 {									\
40 	unsigned long temp, flags;					\
41 									\
42 	raw_local_irq_save(flags);					\
43 	temp = v->counter;						\
44 	v->counter c_op i;						\
45 	raw_local_irq_restore(flags);					\
46 									\
47 	return temp;							\
48 }
49 
50 #define ATOMIC_OPS(op, c_op)						\
51 	ATOMIC_OP(op, c_op)						\
52 	ATOMIC_OP_RETURN(op, c_op)					\
53 	ATOMIC_FETCH_OP(op, c_op)
54 
55 ATOMIC_OPS(add, +=)
56 ATOMIC_OPS(sub, -=)
57 
58 #undef ATOMIC_OPS
59 #define ATOMIC_OPS(op, c_op)						\
60 	ATOMIC_OP(op, c_op)						\
61 	ATOMIC_FETCH_OP(op, c_op)
62 
63 ATOMIC_OPS(and, &=)
64 ATOMIC_OPS(or, |=)
65 ATOMIC_OPS(xor, ^=)
66 
67 #undef ATOMIC_OPS
68 #undef ATOMIC_FETCH_OP
69 #undef ATOMIC_OP_RETURN
70 #undef ATOMIC_OP
71 
72 #endif /* __ASM_SH_ATOMIC_IRQ_H */
73