xref: /openbmc/linux/arch/x86/include/asm/debugreg.h (revision 2f164822)
1 /* SPDX-License-Identifier: GPL-2.0 */
2 #ifndef _ASM_X86_DEBUGREG_H
3 #define _ASM_X86_DEBUGREG_H
4 
5 #include <linux/bug.h>
6 #include <linux/percpu.h>
7 #include <uapi/asm/debugreg.h>
8 
9 DECLARE_PER_CPU(unsigned long, cpu_dr7);
10 
11 #ifndef CONFIG_PARAVIRT_XXL
12 /*
13  * These special macros can be used to get or set a debugging register
14  */
15 #define get_debugreg(var, register)				\
16 	(var) = native_get_debugreg(register)
17 #define set_debugreg(value, register)				\
18 	native_set_debugreg(register, value)
19 #endif
20 
21 static __always_inline unsigned long native_get_debugreg(int regno)
22 {
23 	unsigned long val = 0;	/* Damn you, gcc! */
24 
25 	switch (regno) {
26 	case 0:
27 		asm("mov %%db0, %0" :"=r" (val));
28 		break;
29 	case 1:
30 		asm("mov %%db1, %0" :"=r" (val));
31 		break;
32 	case 2:
33 		asm("mov %%db2, %0" :"=r" (val));
34 		break;
35 	case 3:
36 		asm("mov %%db3, %0" :"=r" (val));
37 		break;
38 	case 6:
39 		asm("mov %%db6, %0" :"=r" (val));
40 		break;
41 	case 7:
42 		/*
43 		 * Apply __FORCE_ORDER to DR7 reads to forbid re-ordering them
44 		 * with other code.
45 		 *
46 		 * This is needed because a DR7 access can cause a #VC exception
47 		 * when running under SEV-ES. Taking a #VC exception is not a
48 		 * safe thing to do just anywhere in the entry code and
49 		 * re-ordering might place the access into an unsafe location.
50 		 *
51 		 * This happened in the NMI handler, where the DR7 read was
52 		 * re-ordered to happen before the call to sev_es_ist_enter(),
53 		 * causing stack recursion.
54 		 */
55 		asm volatile("mov %%db7, %0" : "=r" (val) : __FORCE_ORDER);
56 		break;
57 	default:
58 		BUG();
59 	}
60 	return val;
61 }
62 
63 static __always_inline void native_set_debugreg(int regno, unsigned long value)
64 {
65 	switch (regno) {
66 	case 0:
67 		asm("mov %0, %%db0"	::"r" (value));
68 		break;
69 	case 1:
70 		asm("mov %0, %%db1"	::"r" (value));
71 		break;
72 	case 2:
73 		asm("mov %0, %%db2"	::"r" (value));
74 		break;
75 	case 3:
76 		asm("mov %0, %%db3"	::"r" (value));
77 		break;
78 	case 6:
79 		asm("mov %0, %%db6"	::"r" (value));
80 		break;
81 	case 7:
82 		/*
83 		 * Apply __FORCE_ORDER to DR7 writes to forbid re-ordering them
84 		 * with other code.
85 		 *
86 		 * While is didn't happen with a DR7 write (see the DR7 read
87 		 * comment above which explains where it happened), add the
88 		 * __FORCE_ORDER here too to avoid similar problems in the
89 		 * future.
90 		 */
91 		asm volatile("mov %0, %%db7"	::"r" (value), __FORCE_ORDER);
92 		break;
93 	default:
94 		BUG();
95 	}
96 }
97 
98 static inline void hw_breakpoint_disable(void)
99 {
100 	/* Zero the control register for HW Breakpoint */
101 	set_debugreg(0UL, 7);
102 
103 	/* Zero-out the individual HW breakpoint address registers */
104 	set_debugreg(0UL, 0);
105 	set_debugreg(0UL, 1);
106 	set_debugreg(0UL, 2);
107 	set_debugreg(0UL, 3);
108 }
109 
110 static __always_inline bool hw_breakpoint_active(void)
111 {
112 	return __this_cpu_read(cpu_dr7) & DR_GLOBAL_ENABLE_MASK;
113 }
114 
115 extern void hw_breakpoint_restore(void);
116 
117 static __always_inline unsigned long local_db_save(void)
118 {
119 	unsigned long dr7;
120 
121 	if (static_cpu_has(X86_FEATURE_HYPERVISOR) && !hw_breakpoint_active())
122 		return 0;
123 
124 	get_debugreg(dr7, 7);
125 	dr7 &= ~0x400; /* architecturally set bit */
126 	if (dr7)
127 		set_debugreg(0, 7);
128 	/*
129 	 * Ensure the compiler doesn't lower the above statements into
130 	 * the critical section; disabling breakpoints late would not
131 	 * be good.
132 	 */
133 	barrier();
134 
135 	return dr7;
136 }
137 
138 static __always_inline void local_db_restore(unsigned long dr7)
139 {
140 	/*
141 	 * Ensure the compiler doesn't raise this statement into
142 	 * the critical section; enabling breakpoints early would
143 	 * not be good.
144 	 */
145 	barrier();
146 	if (dr7)
147 		set_debugreg(dr7, 7);
148 }
149 
150 #ifdef CONFIG_CPU_SUP_AMD
151 extern void amd_set_dr_addr_mask(unsigned long mask, unsigned int dr);
152 extern unsigned long amd_get_dr_addr_mask(unsigned int dr);
153 #else
154 static inline void amd_set_dr_addr_mask(unsigned long mask, unsigned int dr) { }
155 static inline unsigned long amd_get_dr_addr_mask(unsigned int dr)
156 {
157 	return 0;
158 }
159 #endif
160 
161 #endif /* _ASM_X86_DEBUGREG_H */
162