xref: /openbmc/linux/arch/x86/include/asm/irqflags.h (revision c4e78957)
1 /* SPDX-License-Identifier: GPL-2.0 */
2 #ifndef _X86_IRQFLAGS_H_
3 #define _X86_IRQFLAGS_H_
4 
5 #include <asm/processor-flags.h>
6 
7 #ifndef __ASSEMBLY__
8 
9 #include <asm/nospec-branch.h>
10 
11 /* Provide __cpuidle; we can't safely include <linux/cpu.h> */
12 #define __cpuidle __section(".cpuidle.text")
13 
14 /*
15  * Interrupt control:
16  */
17 
18 /* Declaration required for gcc < 4.9 to prevent -Werror=missing-prototypes */
19 extern inline unsigned long native_save_fl(void);
20 extern __always_inline unsigned long native_save_fl(void)
21 {
22 	unsigned long flags;
23 
24 	/*
25 	 * "=rm" is safe here, because "pop" adjusts the stack before
26 	 * it evaluates its effective address -- this is part of the
27 	 * documented behavior of the "pop" instruction.
28 	 */
29 	asm volatile("# __raw_save_flags\n\t"
30 		     "pushf ; pop %0"
31 		     : "=rm" (flags)
32 		     : /* no input */
33 		     : "memory");
34 
35 	return flags;
36 }
37 
38 static __always_inline void native_irq_disable(void)
39 {
40 	asm volatile("cli": : :"memory");
41 }
42 
43 static __always_inline void native_irq_enable(void)
44 {
45 	asm volatile("sti": : :"memory");
46 }
47 
48 static inline __cpuidle void native_safe_halt(void)
49 {
50 	mds_idle_clear_cpu_buffers();
51 	asm volatile("sti; hlt": : :"memory");
52 }
53 
54 static inline __cpuidle void native_halt(void)
55 {
56 	mds_idle_clear_cpu_buffers();
57 	asm volatile("hlt": : :"memory");
58 }
59 
60 #endif
61 
62 #ifdef CONFIG_PARAVIRT_XXL
63 #include <asm/paravirt.h>
64 #else
65 #ifndef __ASSEMBLY__
66 #include <linux/types.h>
67 
68 static __always_inline unsigned long arch_local_save_flags(void)
69 {
70 	return native_save_fl();
71 }
72 
73 static __always_inline void arch_local_irq_disable(void)
74 {
75 	native_irq_disable();
76 }
77 
78 static __always_inline void arch_local_irq_enable(void)
79 {
80 	native_irq_enable();
81 }
82 
83 /*
84  * Used in the idle loop; sti takes one instruction cycle
85  * to complete:
86  */
87 static inline __cpuidle void arch_safe_halt(void)
88 {
89 	native_safe_halt();
90 }
91 
92 /*
93  * Used when interrupts are already enabled or to
94  * shutdown the processor:
95  */
96 static inline __cpuidle void halt(void)
97 {
98 	native_halt();
99 }
100 
101 /*
102  * For spinlocks, etc:
103  */
104 static __always_inline unsigned long arch_local_irq_save(void)
105 {
106 	unsigned long flags = arch_local_save_flags();
107 	arch_local_irq_disable();
108 	return flags;
109 }
110 #else
111 
112 #ifdef CONFIG_X86_64
113 #ifdef CONFIG_DEBUG_ENTRY
114 #define SAVE_FLAGS		pushfq; popq %rax
115 #endif
116 
117 #endif
118 
119 #endif /* __ASSEMBLY__ */
120 #endif /* CONFIG_PARAVIRT_XXL */
121 
122 #ifndef __ASSEMBLY__
123 static __always_inline int arch_irqs_disabled_flags(unsigned long flags)
124 {
125 	return !(flags & X86_EFLAGS_IF);
126 }
127 
128 static __always_inline int arch_irqs_disabled(void)
129 {
130 	unsigned long flags = arch_local_save_flags();
131 
132 	return arch_irqs_disabled_flags(flags);
133 }
134 
135 static __always_inline void arch_local_irq_restore(unsigned long flags)
136 {
137 	if (!arch_irqs_disabled_flags(flags))
138 		arch_local_irq_enable();
139 }
140 #endif /* !__ASSEMBLY__ */
141 
142 #endif
143