1 // SPDX-License-Identifier: GPL-2.0-only
2 /*
3  * psci_cpu_on_test - Test that the observable state of a vCPU targeted by the
4  * CPU_ON PSCI call matches what the caller requested.
5  *
6  * Copyright (c) 2021 Google LLC.
7  *
8  * This is a regression test for a race between KVM servicing the PSCI call and
9  * userspace reading the vCPUs registers.
10  */
11 
12 #define _GNU_SOURCE
13 
14 #include <linux/psci.h>
15 
16 #include "kvm_util.h"
17 #include "processor.h"
18 #include "test_util.h"
19 
20 #define VCPU_ID_SOURCE 0
21 #define VCPU_ID_TARGET 1
22 
23 #define CPU_ON_ENTRY_ADDR 0xfeedf00dul
24 #define CPU_ON_CONTEXT_ID 0xdeadc0deul
25 
26 static uint64_t psci_cpu_on(uint64_t target_cpu, uint64_t entry_addr,
27 			    uint64_t context_id)
28 {
29 	struct arm_smccc_res res;
30 
31 	smccc_hvc(PSCI_0_2_FN64_CPU_ON, target_cpu, entry_addr, context_id,
32 		  0, 0, 0, 0, &res);
33 
34 	return res.a0;
35 }
36 
37 static uint64_t psci_affinity_info(uint64_t target_affinity,
38 				   uint64_t lowest_affinity_level)
39 {
40 	struct arm_smccc_res res;
41 
42 	smccc_hvc(PSCI_0_2_FN64_AFFINITY_INFO, target_affinity, lowest_affinity_level,
43 		  0, 0, 0, 0, 0, &res);
44 
45 	return res.a0;
46 }
47 
48 static void guest_main(uint64_t target_cpu)
49 {
50 	GUEST_ASSERT(!psci_cpu_on(target_cpu, CPU_ON_ENTRY_ADDR, CPU_ON_CONTEXT_ID));
51 	uint64_t target_state;
52 
53 	do {
54 		target_state = psci_affinity_info(target_cpu, 0);
55 
56 		GUEST_ASSERT((target_state == PSCI_0_2_AFFINITY_LEVEL_ON) ||
57 			     (target_state == PSCI_0_2_AFFINITY_LEVEL_OFF));
58 	} while (target_state != PSCI_0_2_AFFINITY_LEVEL_ON);
59 
60 	GUEST_DONE();
61 }
62 
63 static void vcpu_power_off(struct kvm_vm *vm, uint32_t vcpuid)
64 {
65 	struct kvm_mp_state mp_state = {
66 		.mp_state = KVM_MP_STATE_STOPPED,
67 	};
68 
69 	vcpu_set_mp_state(vm, vcpuid, &mp_state);
70 }
71 
72 int main(void)
73 {
74 	uint64_t target_mpidr, obs_pc, obs_x0;
75 	struct kvm_vcpu_init init;
76 	struct kvm_vm *vm;
77 	struct ucall uc;
78 
79 	vm = vm_create(VM_MODE_DEFAULT, DEFAULT_GUEST_PHY_PAGES, O_RDWR);
80 	kvm_vm_elf_load(vm, program_invocation_name);
81 	ucall_init(vm, NULL);
82 
83 	vm_ioctl(vm, KVM_ARM_PREFERRED_TARGET, &init);
84 	init.features[0] |= (1 << KVM_ARM_VCPU_PSCI_0_2);
85 
86 	aarch64_vcpu_add_default(vm, VCPU_ID_SOURCE, &init, guest_main);
87 	aarch64_vcpu_add_default(vm, VCPU_ID_TARGET, &init, guest_main);
88 
89 	/*
90 	 * make sure the target is already off when executing the test.
91 	 */
92 	vcpu_power_off(vm, VCPU_ID_TARGET);
93 
94 	get_reg(vm, VCPU_ID_TARGET, KVM_ARM64_SYS_REG(SYS_MPIDR_EL1), &target_mpidr);
95 	vcpu_args_set(vm, VCPU_ID_SOURCE, 1, target_mpidr & MPIDR_HWID_BITMASK);
96 	vcpu_run(vm, VCPU_ID_SOURCE);
97 
98 	switch (get_ucall(vm, VCPU_ID_SOURCE, &uc)) {
99 	case UCALL_DONE:
100 		break;
101 	case UCALL_ABORT:
102 		TEST_FAIL("%s at %s:%ld", (const char *)uc.args[0], __FILE__,
103 			  uc.args[1]);
104 		break;
105 	default:
106 		TEST_FAIL("Unhandled ucall: %lu", uc.cmd);
107 	}
108 
109 	get_reg(vm, VCPU_ID_TARGET, ARM64_CORE_REG(regs.pc), &obs_pc);
110 	get_reg(vm, VCPU_ID_TARGET, ARM64_CORE_REG(regs.regs[0]), &obs_x0);
111 
112 	TEST_ASSERT(obs_pc == CPU_ON_ENTRY_ADDR,
113 		    "unexpected target cpu pc: %lx (expected: %lx)",
114 		    obs_pc, CPU_ON_ENTRY_ADDR);
115 	TEST_ASSERT(obs_x0 == CPU_ON_CONTEXT_ID,
116 		    "unexpected target context id: %lx (expected: %lx)",
117 		    obs_x0, CPU_ON_CONTEXT_ID);
118 
119 	kvm_vm_free(vm);
120 	return 0;
121 }
122