xref: /openbmc/qemu/target/arm/debug_helper.c (revision f7160f32)
1 /*
2  * ARM debug helpers.
3  *
4  * This code is licensed under the GNU GPL v2 or later.
5  *
6  * SPDX-License-Identifier: GPL-2.0-or-later
7  */
8 #include "qemu/osdep.h"
9 #include "cpu.h"
10 #include "internals.h"
11 #include "exec/exec-all.h"
12 #include "exec/helper-proto.h"
13 
14 /* Return true if the linked breakpoint entry lbn passes its checks */
15 static bool linked_bp_matches(ARMCPU *cpu, int lbn)
16 {
17     CPUARMState *env = &cpu->env;
18     uint64_t bcr = env->cp15.dbgbcr[lbn];
19     int brps = arm_num_brps(cpu);
20     int ctx_cmps = arm_num_ctx_cmps(cpu);
21     int bt;
22     uint32_t contextidr;
23     uint64_t hcr_el2;
24 
25     /*
26      * Links to unimplemented or non-context aware breakpoints are
27      * CONSTRAINED UNPREDICTABLE: either behave as if disabled, or
28      * as if linked to an UNKNOWN context-aware breakpoint (in which
29      * case DBGWCR<n>_EL1.LBN must indicate that breakpoint).
30      * We choose the former.
31      */
32     if (lbn >= brps || lbn < (brps - ctx_cmps)) {
33         return false;
34     }
35 
36     bcr = env->cp15.dbgbcr[lbn];
37 
38     if (extract64(bcr, 0, 1) == 0) {
39         /* Linked breakpoint disabled : generate no events */
40         return false;
41     }
42 
43     bt = extract64(bcr, 20, 4);
44     hcr_el2 = arm_hcr_el2_eff(env);
45 
46     switch (bt) {
47     case 3: /* linked context ID match */
48         switch (arm_current_el(env)) {
49         default:
50             /* Context matches never fire in AArch64 EL3 */
51             return false;
52         case 2:
53             if (!(hcr_el2 & HCR_E2H)) {
54                 /* Context matches never fire in EL2 without E2H enabled. */
55                 return false;
56             }
57             contextidr = env->cp15.contextidr_el[2];
58             break;
59         case 1:
60             contextidr = env->cp15.contextidr_el[1];
61             break;
62         case 0:
63             if ((hcr_el2 & (HCR_E2H | HCR_TGE)) == (HCR_E2H | HCR_TGE)) {
64                 contextidr = env->cp15.contextidr_el[2];
65             } else {
66                 contextidr = env->cp15.contextidr_el[1];
67             }
68             break;
69         }
70         break;
71 
72     case 7:  /* linked contextidr_el1 match */
73         contextidr = env->cp15.contextidr_el[1];
74         break;
75     case 13: /* linked contextidr_el2 match */
76         contextidr = env->cp15.contextidr_el[2];
77         break;
78 
79     case 9: /* linked VMID match (reserved if no EL2) */
80     case 11: /* linked context ID and VMID match (reserved if no EL2) */
81     case 15: /* linked full context ID match */
82     default:
83         /*
84          * Links to Unlinked context breakpoints must generate no
85          * events; we choose to do the same for reserved values too.
86          */
87         return false;
88     }
89 
90     /*
91      * We match the whole register even if this is AArch32 using the
92      * short descriptor format (in which case it holds both PROCID and ASID),
93      * since we don't implement the optional v7 context ID masking.
94      */
95     return contextidr == (uint32_t)env->cp15.dbgbvr[lbn];
96 }
97 
98 static bool bp_wp_matches(ARMCPU *cpu, int n, bool is_wp)
99 {
100     CPUARMState *env = &cpu->env;
101     uint64_t cr;
102     int pac, hmc, ssc, wt, lbn;
103     /*
104      * Note that for watchpoints the check is against the CPU security
105      * state, not the S/NS attribute on the offending data access.
106      */
107     bool is_secure = arm_is_secure(env);
108     int access_el = arm_current_el(env);
109 
110     if (is_wp) {
111         CPUWatchpoint *wp = env->cpu_watchpoint[n];
112 
113         if (!wp || !(wp->flags & BP_WATCHPOINT_HIT)) {
114             return false;
115         }
116         cr = env->cp15.dbgwcr[n];
117         if (wp->hitattrs.user) {
118             /*
119              * The LDRT/STRT/LDT/STT "unprivileged access" instructions should
120              * match watchpoints as if they were accesses done at EL0, even if
121              * the CPU is at EL1 or higher.
122              */
123             access_el = 0;
124         }
125     } else {
126         uint64_t pc = is_a64(env) ? env->pc : env->regs[15];
127 
128         if (!env->cpu_breakpoint[n] || env->cpu_breakpoint[n]->pc != pc) {
129             return false;
130         }
131         cr = env->cp15.dbgbcr[n];
132     }
133     /*
134      * The WATCHPOINT_HIT flag guarantees us that the watchpoint is
135      * enabled and that the address and access type match; for breakpoints
136      * we know the address matched; check the remaining fields, including
137      * linked breakpoints. We rely on WCR and BCR having the same layout
138      * for the LBN, SSC, HMC, PAC/PMC and is-linked fields.
139      * Note that some combinations of {PAC, HMC, SSC} are reserved and
140      * must act either like some valid combination or as if the watchpoint
141      * were disabled. We choose the former, and use this together with
142      * the fact that EL3 must always be Secure and EL2 must always be
143      * Non-Secure to simplify the code slightly compared to the full
144      * table in the ARM ARM.
145      */
146     pac = extract64(cr, 1, 2);
147     hmc = extract64(cr, 13, 1);
148     ssc = extract64(cr, 14, 2);
149 
150     switch (ssc) {
151     case 0:
152         break;
153     case 1:
154     case 3:
155         if (is_secure) {
156             return false;
157         }
158         break;
159     case 2:
160         if (!is_secure) {
161             return false;
162         }
163         break;
164     }
165 
166     switch (access_el) {
167     case 3:
168     case 2:
169         if (!hmc) {
170             return false;
171         }
172         break;
173     case 1:
174         if (extract32(pac, 0, 1) == 0) {
175             return false;
176         }
177         break;
178     case 0:
179         if (extract32(pac, 1, 1) == 0) {
180             return false;
181         }
182         break;
183     default:
184         g_assert_not_reached();
185     }
186 
187     wt = extract64(cr, 20, 1);
188     lbn = extract64(cr, 16, 4);
189 
190     if (wt && !linked_bp_matches(cpu, lbn)) {
191         return false;
192     }
193 
194     return true;
195 }
196 
197 static bool check_watchpoints(ARMCPU *cpu)
198 {
199     CPUARMState *env = &cpu->env;
200     int n;
201 
202     /*
203      * If watchpoints are disabled globally or we can't take debug
204      * exceptions here then watchpoint firings are ignored.
205      */
206     if (extract32(env->cp15.mdscr_el1, 15, 1) == 0
207         || !arm_generate_debug_exceptions(env)) {
208         return false;
209     }
210 
211     for (n = 0; n < ARRAY_SIZE(env->cpu_watchpoint); n++) {
212         if (bp_wp_matches(cpu, n, true)) {
213             return true;
214         }
215     }
216     return false;
217 }
218 
219 static bool check_breakpoints(ARMCPU *cpu)
220 {
221     CPUARMState *env = &cpu->env;
222     int n;
223 
224     /*
225      * If breakpoints are disabled globally or we can't take debug
226      * exceptions here then breakpoint firings are ignored.
227      */
228     if (extract32(env->cp15.mdscr_el1, 15, 1) == 0
229         || !arm_generate_debug_exceptions(env)) {
230         return false;
231     }
232 
233     for (n = 0; n < ARRAY_SIZE(env->cpu_breakpoint); n++) {
234         if (bp_wp_matches(cpu, n, false)) {
235             return true;
236         }
237     }
238     return false;
239 }
240 
241 void HELPER(check_breakpoints)(CPUARMState *env)
242 {
243     ARMCPU *cpu = env_archcpu(env);
244 
245     if (check_breakpoints(cpu)) {
246         HELPER(exception_internal(env, EXCP_DEBUG));
247     }
248 }
249 
250 bool arm_debug_check_watchpoint(CPUState *cs, CPUWatchpoint *wp)
251 {
252     /*
253      * Called by core code when a CPU watchpoint fires; need to check if this
254      * is also an architectural watchpoint match.
255      */
256     ARMCPU *cpu = ARM_CPU(cs);
257 
258     return check_watchpoints(cpu);
259 }
260 
261 void arm_debug_excp_handler(CPUState *cs)
262 {
263     /*
264      * Called by core code when a watchpoint or breakpoint fires;
265      * need to check which one and raise the appropriate exception.
266      */
267     ARMCPU *cpu = ARM_CPU(cs);
268     CPUARMState *env = &cpu->env;
269     CPUWatchpoint *wp_hit = cs->watchpoint_hit;
270 
271     if (wp_hit) {
272         if (wp_hit->flags & BP_CPU) {
273             bool wnr = (wp_hit->flags & BP_WATCHPOINT_HIT_WRITE) != 0;
274             bool same_el = arm_debug_target_el(env) == arm_current_el(env);
275 
276             cs->watchpoint_hit = NULL;
277 
278             env->exception.fsr = arm_debug_exception_fsr(env);
279             env->exception.vaddress = wp_hit->hitaddr;
280             raise_exception(env, EXCP_DATA_ABORT,
281                     syn_watchpoint(same_el, 0, wnr),
282                     arm_debug_target_el(env));
283         }
284     } else {
285         uint64_t pc = is_a64(env) ? env->pc : env->regs[15];
286         bool same_el = (arm_debug_target_el(env) == arm_current_el(env));
287 
288         /*
289          * (1) GDB breakpoints should be handled first.
290          * (2) Do not raise a CPU exception if no CPU breakpoint has fired,
291          * since singlestep is also done by generating a debug internal
292          * exception.
293          */
294         if (cpu_breakpoint_test(cs, pc, BP_GDB)
295             || !cpu_breakpoint_test(cs, pc, BP_CPU)) {
296             return;
297         }
298 
299         env->exception.fsr = arm_debug_exception_fsr(env);
300         /*
301          * FAR is UNKNOWN: clear vaddress to avoid potentially exposing
302          * values to the guest that it shouldn't be able to see at its
303          * exception/security level.
304          */
305         env->exception.vaddress = 0;
306         raise_exception(env, EXCP_PREFETCH_ABORT,
307                         syn_breakpoint(same_el),
308                         arm_debug_target_el(env));
309     }
310 }
311 
312 #if !defined(CONFIG_USER_ONLY)
313 
314 vaddr arm_adjust_watchpoint_address(CPUState *cs, vaddr addr, int len)
315 {
316     ARMCPU *cpu = ARM_CPU(cs);
317     CPUARMState *env = &cpu->env;
318 
319     /*
320      * In BE32 system mode, target memory is stored byteswapped (on a
321      * little-endian host system), and by the time we reach here (via an
322      * opcode helper) the addresses of subword accesses have been adjusted
323      * to account for that, which means that watchpoints will not match.
324      * Undo the adjustment here.
325      */
326     if (arm_sctlr_b(env)) {
327         if (len == 1) {
328             addr ^= 3;
329         } else if (len == 2) {
330             addr ^= 2;
331         }
332     }
333 
334     return addr;
335 }
336 
337 #endif
338