xref: /openbmc/linux/arch/powerpc/kvm/e500mc.c (revision 79f08d9e)
1 /*
2  * Copyright (C) 2010,2012 Freescale Semiconductor, Inc. All rights reserved.
3  *
4  * Author: Varun Sethi, <varun.sethi@freescale.com>
5  *
6  * Description:
7  * This file is derived from arch/powerpc/kvm/e500.c,
8  * by Yu Liu <yu.liu@freescale.com>.
9  *
10  * This program is free software; you can redistribute it and/or modify
11  * it under the terms of the GNU General Public License, version 2, as
12  * published by the Free Software Foundation.
13  */
14 
15 #include <linux/kvm_host.h>
16 #include <linux/slab.h>
17 #include <linux/err.h>
18 #include <linux/export.h>
19 
20 #include <asm/reg.h>
21 #include <asm/cputable.h>
22 #include <asm/tlbflush.h>
23 #include <asm/kvm_ppc.h>
24 #include <asm/dbell.h>
25 
26 #include "booke.h"
27 #include "e500.h"
28 
29 void kvmppc_set_pending_interrupt(struct kvm_vcpu *vcpu, enum int_class type)
30 {
31 	enum ppc_dbell dbell_type;
32 	unsigned long tag;
33 
34 	switch (type) {
35 	case INT_CLASS_NONCRIT:
36 		dbell_type = PPC_G_DBELL;
37 		break;
38 	case INT_CLASS_CRIT:
39 		dbell_type = PPC_G_DBELL_CRIT;
40 		break;
41 	case INT_CLASS_MC:
42 		dbell_type = PPC_G_DBELL_MC;
43 		break;
44 	default:
45 		WARN_ONCE(1, "%s: unknown int type %d\n", __func__, type);
46 		return;
47 	}
48 
49 
50 	tag = PPC_DBELL_LPID(vcpu->kvm->arch.lpid) | vcpu->vcpu_id;
51 	mb();
52 	ppc_msgsnd(dbell_type, 0, tag);
53 }
54 
55 /* gtlbe must not be mapped by more than one host tlb entry */
56 void kvmppc_e500_tlbil_one(struct kvmppc_vcpu_e500 *vcpu_e500,
57 			   struct kvm_book3e_206_tlb_entry *gtlbe)
58 {
59 	unsigned int tid, ts;
60 	gva_t eaddr;
61 	u32 val, lpid;
62 	unsigned long flags;
63 
64 	ts = get_tlb_ts(gtlbe);
65 	tid = get_tlb_tid(gtlbe);
66 	lpid = vcpu_e500->vcpu.kvm->arch.lpid;
67 
68 	/* We search the host TLB to invalidate its shadow TLB entry */
69 	val = (tid << 16) | ts;
70 	eaddr = get_tlb_eaddr(gtlbe);
71 
72 	local_irq_save(flags);
73 
74 	mtspr(SPRN_MAS6, val);
75 	mtspr(SPRN_MAS5, MAS5_SGS | lpid);
76 
77 	asm volatile("tlbsx 0, %[eaddr]\n" : : [eaddr] "r" (eaddr));
78 	val = mfspr(SPRN_MAS1);
79 	if (val & MAS1_VALID) {
80 		mtspr(SPRN_MAS1, val & ~MAS1_VALID);
81 		asm volatile("tlbwe");
82 	}
83 	mtspr(SPRN_MAS5, 0);
84 	/* NOTE: tlbsx also updates mas8, so clear it for host tlbwe */
85 	mtspr(SPRN_MAS8, 0);
86 	isync();
87 
88 	local_irq_restore(flags);
89 }
90 
91 void kvmppc_e500_tlbil_all(struct kvmppc_vcpu_e500 *vcpu_e500)
92 {
93 	unsigned long flags;
94 
95 	local_irq_save(flags);
96 	mtspr(SPRN_MAS5, MAS5_SGS | vcpu_e500->vcpu.kvm->arch.lpid);
97 	asm volatile("tlbilxlpid");
98 	mtspr(SPRN_MAS5, 0);
99 	local_irq_restore(flags);
100 }
101 
102 void kvmppc_set_pid(struct kvm_vcpu *vcpu, u32 pid)
103 {
104 	vcpu->arch.pid = pid;
105 }
106 
107 void kvmppc_mmu_msr_notify(struct kvm_vcpu *vcpu, u32 old_msr)
108 {
109 }
110 
111 static DEFINE_PER_CPU(struct kvm_vcpu *, last_vcpu_on_cpu);
112 
113 static void kvmppc_core_vcpu_load_e500mc(struct kvm_vcpu *vcpu, int cpu)
114 {
115 	struct kvmppc_vcpu_e500 *vcpu_e500 = to_e500(vcpu);
116 
117 	kvmppc_booke_vcpu_load(vcpu, cpu);
118 
119 	mtspr(SPRN_LPID, vcpu->kvm->arch.lpid);
120 	mtspr(SPRN_EPCR, vcpu->arch.shadow_epcr);
121 	mtspr(SPRN_GPIR, vcpu->vcpu_id);
122 	mtspr(SPRN_MSRP, vcpu->arch.shadow_msrp);
123 	mtspr(SPRN_EPLC, vcpu->arch.eplc);
124 	mtspr(SPRN_EPSC, vcpu->arch.epsc);
125 
126 	mtspr(SPRN_GIVPR, vcpu->arch.ivpr);
127 	mtspr(SPRN_GIVOR2, vcpu->arch.ivor[BOOKE_IRQPRIO_DATA_STORAGE]);
128 	mtspr(SPRN_GIVOR8, vcpu->arch.ivor[BOOKE_IRQPRIO_SYSCALL]);
129 	mtspr(SPRN_GSPRG0, (unsigned long)vcpu->arch.shared->sprg0);
130 	mtspr(SPRN_GSPRG1, (unsigned long)vcpu->arch.shared->sprg1);
131 	mtspr(SPRN_GSPRG2, (unsigned long)vcpu->arch.shared->sprg2);
132 	mtspr(SPRN_GSPRG3, (unsigned long)vcpu->arch.shared->sprg3);
133 
134 	mtspr(SPRN_GSRR0, vcpu->arch.shared->srr0);
135 	mtspr(SPRN_GSRR1, vcpu->arch.shared->srr1);
136 
137 	mtspr(SPRN_GEPR, vcpu->arch.epr);
138 	mtspr(SPRN_GDEAR, vcpu->arch.shared->dar);
139 	mtspr(SPRN_GESR, vcpu->arch.shared->esr);
140 
141 	if (vcpu->arch.oldpir != mfspr(SPRN_PIR) ||
142 	    __get_cpu_var(last_vcpu_on_cpu) != vcpu) {
143 		kvmppc_e500_tlbil_all(vcpu_e500);
144 		__get_cpu_var(last_vcpu_on_cpu) = vcpu;
145 	}
146 
147 	kvmppc_load_guest_fp(vcpu);
148 }
149 
150 static void kvmppc_core_vcpu_put_e500mc(struct kvm_vcpu *vcpu)
151 {
152 	vcpu->arch.eplc = mfspr(SPRN_EPLC);
153 	vcpu->arch.epsc = mfspr(SPRN_EPSC);
154 
155 	vcpu->arch.shared->sprg0 = mfspr(SPRN_GSPRG0);
156 	vcpu->arch.shared->sprg1 = mfspr(SPRN_GSPRG1);
157 	vcpu->arch.shared->sprg2 = mfspr(SPRN_GSPRG2);
158 	vcpu->arch.shared->sprg3 = mfspr(SPRN_GSPRG3);
159 
160 	vcpu->arch.shared->srr0 = mfspr(SPRN_GSRR0);
161 	vcpu->arch.shared->srr1 = mfspr(SPRN_GSRR1);
162 
163 	vcpu->arch.epr = mfspr(SPRN_GEPR);
164 	vcpu->arch.shared->dar = mfspr(SPRN_GDEAR);
165 	vcpu->arch.shared->esr = mfspr(SPRN_GESR);
166 
167 	vcpu->arch.oldpir = mfspr(SPRN_PIR);
168 
169 	kvmppc_booke_vcpu_put(vcpu);
170 }
171 
172 int kvmppc_core_check_processor_compat(void)
173 {
174 	int r;
175 
176 	if (strcmp(cur_cpu_spec->cpu_name, "e500mc") == 0)
177 		r = 0;
178 	else if (strcmp(cur_cpu_spec->cpu_name, "e5500") == 0)
179 		r = 0;
180 	else
181 		r = -ENOTSUPP;
182 
183 	return r;
184 }
185 
186 int kvmppc_core_vcpu_setup(struct kvm_vcpu *vcpu)
187 {
188 	struct kvmppc_vcpu_e500 *vcpu_e500 = to_e500(vcpu);
189 
190 	vcpu->arch.shadow_epcr = SPRN_EPCR_DSIGS | SPRN_EPCR_DGTMI | \
191 				 SPRN_EPCR_DUVD;
192 #ifdef CONFIG_64BIT
193 	vcpu->arch.shadow_epcr |= SPRN_EPCR_ICM;
194 #endif
195 	vcpu->arch.shadow_msrp = MSRP_UCLEP | MSRP_DEP | MSRP_PMMP;
196 	vcpu->arch.eplc = EPC_EGS | (vcpu->kvm->arch.lpid << EPC_ELPID_SHIFT);
197 	vcpu->arch.epsc = vcpu->arch.eplc;
198 
199 	vcpu->arch.pvr = mfspr(SPRN_PVR);
200 	vcpu_e500->svr = mfspr(SPRN_SVR);
201 
202 	vcpu->arch.cpu_type = KVM_CPU_E500MC;
203 
204 	return 0;
205 }
206 
207 static int kvmppc_core_get_sregs_e500mc(struct kvm_vcpu *vcpu,
208 					struct kvm_sregs *sregs)
209 {
210 	struct kvmppc_vcpu_e500 *vcpu_e500 = to_e500(vcpu);
211 
212 	sregs->u.e.features |= KVM_SREGS_E_ARCH206_MMU | KVM_SREGS_E_PM |
213 			       KVM_SREGS_E_PC;
214 	sregs->u.e.impl_id = KVM_SREGS_E_IMPL_FSL;
215 
216 	sregs->u.e.impl.fsl.features = 0;
217 	sregs->u.e.impl.fsl.svr = vcpu_e500->svr;
218 	sregs->u.e.impl.fsl.hid0 = vcpu_e500->hid0;
219 	sregs->u.e.impl.fsl.mcar = vcpu_e500->mcar;
220 
221 	kvmppc_get_sregs_e500_tlb(vcpu, sregs);
222 
223 	sregs->u.e.ivor_high[3] =
224 		vcpu->arch.ivor[BOOKE_IRQPRIO_PERFORMANCE_MONITOR];
225 	sregs->u.e.ivor_high[4] = vcpu->arch.ivor[BOOKE_IRQPRIO_DBELL];
226 	sregs->u.e.ivor_high[5] = vcpu->arch.ivor[BOOKE_IRQPRIO_DBELL_CRIT];
227 
228 	return kvmppc_get_sregs_ivor(vcpu, sregs);
229 }
230 
231 static int kvmppc_core_set_sregs_e500mc(struct kvm_vcpu *vcpu,
232 					struct kvm_sregs *sregs)
233 {
234 	struct kvmppc_vcpu_e500 *vcpu_e500 = to_e500(vcpu);
235 	int ret;
236 
237 	if (sregs->u.e.impl_id == KVM_SREGS_E_IMPL_FSL) {
238 		vcpu_e500->svr = sregs->u.e.impl.fsl.svr;
239 		vcpu_e500->hid0 = sregs->u.e.impl.fsl.hid0;
240 		vcpu_e500->mcar = sregs->u.e.impl.fsl.mcar;
241 	}
242 
243 	ret = kvmppc_set_sregs_e500_tlb(vcpu, sregs);
244 	if (ret < 0)
245 		return ret;
246 
247 	if (!(sregs->u.e.features & KVM_SREGS_E_IVOR))
248 		return 0;
249 
250 	if (sregs->u.e.features & KVM_SREGS_E_PM) {
251 		vcpu->arch.ivor[BOOKE_IRQPRIO_PERFORMANCE_MONITOR] =
252 			sregs->u.e.ivor_high[3];
253 	}
254 
255 	if (sregs->u.e.features & KVM_SREGS_E_PC) {
256 		vcpu->arch.ivor[BOOKE_IRQPRIO_DBELL] =
257 			sregs->u.e.ivor_high[4];
258 		vcpu->arch.ivor[BOOKE_IRQPRIO_DBELL_CRIT] =
259 			sregs->u.e.ivor_high[5];
260 	}
261 
262 	return kvmppc_set_sregs_ivor(vcpu, sregs);
263 }
264 
265 static int kvmppc_get_one_reg_e500mc(struct kvm_vcpu *vcpu, u64 id,
266 			      union kvmppc_one_reg *val)
267 {
268 	int r = kvmppc_get_one_reg_e500_tlb(vcpu, id, val);
269 	return r;
270 }
271 
272 static int kvmppc_set_one_reg_e500mc(struct kvm_vcpu *vcpu, u64 id,
273 			      union kvmppc_one_reg *val)
274 {
275 	int r = kvmppc_set_one_reg_e500_tlb(vcpu, id, val);
276 	return r;
277 }
278 
279 static struct kvm_vcpu *kvmppc_core_vcpu_create_e500mc(struct kvm *kvm,
280 						       unsigned int id)
281 {
282 	struct kvmppc_vcpu_e500 *vcpu_e500;
283 	struct kvm_vcpu *vcpu;
284 	int err;
285 
286 	vcpu_e500 = kmem_cache_zalloc(kvm_vcpu_cache, GFP_KERNEL);
287 	if (!vcpu_e500) {
288 		err = -ENOMEM;
289 		goto out;
290 	}
291 	vcpu = &vcpu_e500->vcpu;
292 
293 	/* Invalid PIR value -- this LPID dosn't have valid state on any cpu */
294 	vcpu->arch.oldpir = 0xffffffff;
295 
296 	err = kvm_vcpu_init(vcpu, kvm, id);
297 	if (err)
298 		goto free_vcpu;
299 
300 	err = kvmppc_e500_tlb_init(vcpu_e500);
301 	if (err)
302 		goto uninit_vcpu;
303 
304 	vcpu->arch.shared = (void *)__get_free_page(GFP_KERNEL | __GFP_ZERO);
305 	if (!vcpu->arch.shared)
306 		goto uninit_tlb;
307 
308 	return vcpu;
309 
310 uninit_tlb:
311 	kvmppc_e500_tlb_uninit(vcpu_e500);
312 uninit_vcpu:
313 	kvm_vcpu_uninit(vcpu);
314 
315 free_vcpu:
316 	kmem_cache_free(kvm_vcpu_cache, vcpu_e500);
317 out:
318 	return ERR_PTR(err);
319 }
320 
321 static void kvmppc_core_vcpu_free_e500mc(struct kvm_vcpu *vcpu)
322 {
323 	struct kvmppc_vcpu_e500 *vcpu_e500 = to_e500(vcpu);
324 
325 	free_page((unsigned long)vcpu->arch.shared);
326 	kvmppc_e500_tlb_uninit(vcpu_e500);
327 	kvm_vcpu_uninit(vcpu);
328 	kmem_cache_free(kvm_vcpu_cache, vcpu_e500);
329 }
330 
331 static int kvmppc_core_init_vm_e500mc(struct kvm *kvm)
332 {
333 	int lpid;
334 
335 	lpid = kvmppc_alloc_lpid();
336 	if (lpid < 0)
337 		return lpid;
338 
339 	kvm->arch.lpid = lpid;
340 	return 0;
341 }
342 
343 static void kvmppc_core_destroy_vm_e500mc(struct kvm *kvm)
344 {
345 	kvmppc_free_lpid(kvm->arch.lpid);
346 }
347 
348 static struct kvmppc_ops kvm_ops_e500mc = {
349 	.get_sregs = kvmppc_core_get_sregs_e500mc,
350 	.set_sregs = kvmppc_core_set_sregs_e500mc,
351 	.get_one_reg = kvmppc_get_one_reg_e500mc,
352 	.set_one_reg = kvmppc_set_one_reg_e500mc,
353 	.vcpu_load   = kvmppc_core_vcpu_load_e500mc,
354 	.vcpu_put    = kvmppc_core_vcpu_put_e500mc,
355 	.vcpu_create = kvmppc_core_vcpu_create_e500mc,
356 	.vcpu_free   = kvmppc_core_vcpu_free_e500mc,
357 	.mmu_destroy  = kvmppc_mmu_destroy_e500,
358 	.init_vm = kvmppc_core_init_vm_e500mc,
359 	.destroy_vm = kvmppc_core_destroy_vm_e500mc,
360 	.emulate_op = kvmppc_core_emulate_op_e500,
361 	.emulate_mtspr = kvmppc_core_emulate_mtspr_e500,
362 	.emulate_mfspr = kvmppc_core_emulate_mfspr_e500,
363 };
364 
365 static int __init kvmppc_e500mc_init(void)
366 {
367 	int r;
368 
369 	r = kvmppc_booke_init();
370 	if (r)
371 		goto err_out;
372 
373 	kvmppc_init_lpid(64);
374 	kvmppc_claim_lpid(0); /* host */
375 
376 	r = kvm_init(NULL, sizeof(struct kvmppc_vcpu_e500), 0, THIS_MODULE);
377 	if (r)
378 		goto err_out;
379 	kvm_ops_e500mc.owner = THIS_MODULE;
380 	kvmppc_pr_ops = &kvm_ops_e500mc;
381 
382 err_out:
383 	return r;
384 }
385 
386 static void __exit kvmppc_e500mc_exit(void)
387 {
388 	kvmppc_pr_ops = NULL;
389 	kvmppc_booke_exit();
390 }
391 
392 module_init(kvmppc_e500mc_init);
393 module_exit(kvmppc_e500mc_exit);
394