xref: /openbmc/linux/arch/arm64/kvm/hyp/nvhe/host.S (revision 53f9cd5c)
1/* SPDX-License-Identifier: GPL-2.0-only */
2/*
3 * Copyright (C) 2020 - Google Inc
4 * Author: Andrew Scull <ascull@google.com>
5 */
6
7#include <linux/linkage.h>
8
9#include <asm/assembler.h>
10#include <asm/kvm_arm.h>
11#include <asm/kvm_asm.h>
12#include <asm/kvm_mmu.h>
13
14	.text
15
16SYM_FUNC_START(__host_exit)
17	get_host_ctxt	x0, x1
18
19	/* Store the host regs x2 and x3 */
20	stp	x2, x3,   [x0, #CPU_XREG_OFFSET(2)]
21
22	/* Retrieve the host regs x0-x1 from the stack */
23	ldp	x2, x3, [sp], #16	// x0, x1
24
25	/* Store the host regs x0-x1 and x4-x17 */
26	stp	x2, x3,   [x0, #CPU_XREG_OFFSET(0)]
27	stp	x4, x5,   [x0, #CPU_XREG_OFFSET(4)]
28	stp	x6, x7,   [x0, #CPU_XREG_OFFSET(6)]
29	stp	x8, x9,   [x0, #CPU_XREG_OFFSET(8)]
30	stp	x10, x11, [x0, #CPU_XREG_OFFSET(10)]
31	stp	x12, x13, [x0, #CPU_XREG_OFFSET(12)]
32	stp	x14, x15, [x0, #CPU_XREG_OFFSET(14)]
33	stp	x16, x17, [x0, #CPU_XREG_OFFSET(16)]
34
35	/* Store the host regs x18-x29, lr */
36	save_callee_saved_regs x0
37
38	/* Save the host context pointer in x29 across the function call */
39	mov	x29, x0
40	bl	handle_trap
41
42	/* Restore host regs x0-x17 */
43__host_enter_restore_full:
44	ldp	x0, x1,   [x29, #CPU_XREG_OFFSET(0)]
45	ldp	x2, x3,   [x29, #CPU_XREG_OFFSET(2)]
46	ldp	x4, x5,   [x29, #CPU_XREG_OFFSET(4)]
47	ldp	x6, x7,   [x29, #CPU_XREG_OFFSET(6)]
48
49	/* x0-7 are use for panic arguments */
50__host_enter_for_panic:
51	ldp	x8, x9,   [x29, #CPU_XREG_OFFSET(8)]
52	ldp	x10, x11, [x29, #CPU_XREG_OFFSET(10)]
53	ldp	x12, x13, [x29, #CPU_XREG_OFFSET(12)]
54	ldp	x14, x15, [x29, #CPU_XREG_OFFSET(14)]
55	ldp	x16, x17, [x29, #CPU_XREG_OFFSET(16)]
56
57	/* Restore host regs x18-x29, lr */
58	restore_callee_saved_regs x29
59
60	/* Do not touch any register after this! */
61__host_enter_without_restoring:
62	eret
63	sb
64SYM_FUNC_END(__host_exit)
65
66/*
67 * void __noreturn __host_enter(struct kvm_cpu_context *host_ctxt);
68 */
69SYM_FUNC_START(__host_enter)
70	mov	x29, x0
71	b	__host_enter_restore_full
72SYM_FUNC_END(__host_enter)
73
74/*
75 * void __noreturn __hyp_do_panic(struct kvm_cpu_context *host_ctxt, u64 spsr,
76 * 				  u64 elr, u64 par);
77 */
78SYM_FUNC_START(__hyp_do_panic)
79	/* Prepare and exit to the host's panic funciton. */
80	mov	lr, #(PSR_F_BIT | PSR_I_BIT | PSR_A_BIT | PSR_D_BIT |\
81		      PSR_MODE_EL1h)
82	msr	spsr_el2, lr
83	adr_l	lr, nvhe_hyp_panic_handler
84	hyp_kimg_va lr, x6
85	msr	elr_el2, lr
86
87	mov	x29, x0
88
89#ifdef CONFIG_NVHE_EL2_DEBUG
90	/* Ensure host stage-2 is disabled */
91	mrs	x0, hcr_el2
92	bic	x0, x0, #HCR_VM
93	msr	hcr_el2, x0
94	isb
95	tlbi	vmalls12e1
96	dsb	nsh
97#endif
98
99	/* Load the panic arguments into x0-7 */
100	mrs	x0, esr_el2
101	mov	x4, x3
102	mov	x3, x2
103	hyp_pa	x3, x6
104	get_vcpu_ptr x5, x6
105	mrs	x6, far_el2
106	mrs	x7, hpfar_el2
107
108	/* Enter the host, conditionally restoring the host context. */
109	cbz	x29, __host_enter_without_restoring
110	b	__host_enter_for_panic
111SYM_FUNC_END(__hyp_do_panic)
112
113SYM_FUNC_START(__host_hvc)
114	ldp	x0, x1, [sp]		// Don't fixup the stack yet
115
116	/* No stub for you, sonny Jim */
117alternative_if ARM64_KVM_PROTECTED_MODE
118	b	__host_exit
119alternative_else_nop_endif
120
121	/* Check for a stub HVC call */
122	cmp	x0, #HVC_STUB_HCALL_NR
123	b.hs	__host_exit
124
125	add	sp, sp, #16
126	/*
127	 * Compute the idmap address of __kvm_handle_stub_hvc and
128	 * jump there.
129	 *
130	 * Preserve x0-x4, which may contain stub parameters.
131	 */
132	adr_l	x5, __kvm_handle_stub_hvc
133	hyp_pa	x5, x6
134	br	x5
135SYM_FUNC_END(__host_hvc)
136
137.macro host_el1_sync_vect
138	.align 7
139.L__vect_start\@:
140	stp	x0, x1, [sp, #-16]!
141	mrs	x0, esr_el2
142	ubfx	x0, x0, #ESR_ELx_EC_SHIFT, #ESR_ELx_EC_WIDTH
143	cmp	x0, #ESR_ELx_EC_HVC64
144	b.eq	__host_hvc
145	b	__host_exit
146.L__vect_end\@:
147.if ((.L__vect_end\@ - .L__vect_start\@) > 0x80)
148	.error "host_el1_sync_vect larger than vector entry"
149.endif
150.endm
151
152.macro invalid_host_el2_vect
153	.align 7
154
155	/*
156	 * Test whether the SP has overflowed, without corrupting a GPR.
157	 * nVHE hypervisor stacks are aligned so that the PAGE_SHIFT bit
158	 * of SP should always be 1.
159	 */
160	add	sp, sp, x0			// sp' = sp + x0
161	sub	x0, sp, x0			// x0' = sp' - x0 = (sp + x0) - x0 = sp
162	tbz	x0, #PAGE_SHIFT, .L__hyp_sp_overflow\@
163	sub	x0, sp, x0			// x0'' = sp' - x0' = (sp + x0) - sp = x0
164	sub	sp, sp, x0			// sp'' = sp' - x0 = (sp + x0) - x0 = sp
165
166	/* If a guest is loaded, panic out of it. */
167	stp	x0, x1, [sp, #-16]!
168	get_loaded_vcpu x0, x1
169	cbnz	x0, __guest_exit_panic
170	add	sp, sp, #16
171
172	/*
173	 * The panic may not be clean if the exception is taken before the host
174	 * context has been saved by __host_exit or after the hyp context has
175	 * been partially clobbered by __host_enter.
176	 */
177	b	hyp_panic
178
179.L__hyp_sp_overflow\@:
180	/* Switch to the overflow stack */
181	adr_this_cpu sp, overflow_stack + OVERFLOW_STACK_SIZE, x0
182
183	b	hyp_panic_bad_stack
184	ASM_BUG()
185.endm
186
187.macro invalid_host_el1_vect
188	.align 7
189	mov	x0, xzr		/* restore_host = false */
190	mrs	x1, spsr_el2
191	mrs	x2, elr_el2
192	mrs	x3, par_el1
193	b	__hyp_do_panic
194.endm
195
196/*
197 * The host vector does not use an ESB instruction in order to avoid consuming
198 * SErrors that should only be consumed by the host. Guest entry is deferred by
199 * __guest_enter if there are any pending asynchronous exceptions so hyp will
200 * always return to the host without having consumerd host SErrors.
201 *
202 * CONFIG_KVM_INDIRECT_VECTORS is not applied to the host vectors because the
203 * host knows about the EL2 vectors already, and there is no point in hiding
204 * them.
205 */
206	.align 11
207SYM_CODE_START(__kvm_hyp_host_vector)
208	invalid_host_el2_vect			// Synchronous EL2t
209	invalid_host_el2_vect			// IRQ EL2t
210	invalid_host_el2_vect			// FIQ EL2t
211	invalid_host_el2_vect			// Error EL2t
212
213	invalid_host_el2_vect			// Synchronous EL2h
214	invalid_host_el2_vect			// IRQ EL2h
215	invalid_host_el2_vect			// FIQ EL2h
216	invalid_host_el2_vect			// Error EL2h
217
218	host_el1_sync_vect			// Synchronous 64-bit EL1/EL0
219	invalid_host_el1_vect			// IRQ 64-bit EL1/EL0
220	invalid_host_el1_vect			// FIQ 64-bit EL1/EL0
221	invalid_host_el1_vect			// Error 64-bit EL1/EL0
222
223	host_el1_sync_vect			// Synchronous 32-bit EL1/EL0
224	invalid_host_el1_vect			// IRQ 32-bit EL1/EL0
225	invalid_host_el1_vect			// FIQ 32-bit EL1/EL0
226	invalid_host_el1_vect			// Error 32-bit EL1/EL0
227SYM_CODE_END(__kvm_hyp_host_vector)
228
229/*
230 * Forward SMC with arguments in struct kvm_cpu_context, and
231 * store the result into the same struct. Assumes SMCCC 1.2 or older.
232 *
233 * x0: struct kvm_cpu_context*
234 */
235SYM_CODE_START(__kvm_hyp_host_forward_smc)
236	/*
237	 * Use x18 to keep the pointer to the host context because
238	 * x18 is callee-saved in SMCCC but not in AAPCS64.
239	 */
240	mov	x18, x0
241
242	ldp	x0, x1,   [x18, #CPU_XREG_OFFSET(0)]
243	ldp	x2, x3,   [x18, #CPU_XREG_OFFSET(2)]
244	ldp	x4, x5,   [x18, #CPU_XREG_OFFSET(4)]
245	ldp	x6, x7,   [x18, #CPU_XREG_OFFSET(6)]
246	ldp	x8, x9,   [x18, #CPU_XREG_OFFSET(8)]
247	ldp	x10, x11, [x18, #CPU_XREG_OFFSET(10)]
248	ldp	x12, x13, [x18, #CPU_XREG_OFFSET(12)]
249	ldp	x14, x15, [x18, #CPU_XREG_OFFSET(14)]
250	ldp	x16, x17, [x18, #CPU_XREG_OFFSET(16)]
251
252	smc	#0
253
254	stp	x0, x1,   [x18, #CPU_XREG_OFFSET(0)]
255	stp	x2, x3,   [x18, #CPU_XREG_OFFSET(2)]
256	stp	x4, x5,   [x18, #CPU_XREG_OFFSET(4)]
257	stp	x6, x7,   [x18, #CPU_XREG_OFFSET(6)]
258	stp	x8, x9,   [x18, #CPU_XREG_OFFSET(8)]
259	stp	x10, x11, [x18, #CPU_XREG_OFFSET(10)]
260	stp	x12, x13, [x18, #CPU_XREG_OFFSET(12)]
261	stp	x14, x15, [x18, #CPU_XREG_OFFSET(14)]
262	stp	x16, x17, [x18, #CPU_XREG_OFFSET(16)]
263
264	ret
265SYM_CODE_END(__kvm_hyp_host_forward_smc)
266