1 // SPDX-License-Identifier: GPL-2.0
2 /*
3 * Copyright (C) 2021 Western Digital Corporation or its affiliates.
4 * Copyright (C) 2022 Ventana Micro Systems Inc.
5 *
6 * Authors:
7 * Anup Patel <apatel@ventanamicro.com>
8 */
9
10 #include <linux/bits.h>
11 #include <linux/kvm_host.h>
12 #include <linux/uaccess.h>
13 #include <asm/kvm_aia_imsic.h>
14
unlock_vcpus(struct kvm * kvm,int vcpu_lock_idx)15 static void unlock_vcpus(struct kvm *kvm, int vcpu_lock_idx)
16 {
17 struct kvm_vcpu *tmp_vcpu;
18
19 for (; vcpu_lock_idx >= 0; vcpu_lock_idx--) {
20 tmp_vcpu = kvm_get_vcpu(kvm, vcpu_lock_idx);
21 mutex_unlock(&tmp_vcpu->mutex);
22 }
23 }
24
unlock_all_vcpus(struct kvm * kvm)25 static void unlock_all_vcpus(struct kvm *kvm)
26 {
27 unlock_vcpus(kvm, atomic_read(&kvm->online_vcpus) - 1);
28 }
29
lock_all_vcpus(struct kvm * kvm)30 static bool lock_all_vcpus(struct kvm *kvm)
31 {
32 struct kvm_vcpu *tmp_vcpu;
33 unsigned long c;
34
35 kvm_for_each_vcpu(c, tmp_vcpu, kvm) {
36 if (!mutex_trylock(&tmp_vcpu->mutex)) {
37 unlock_vcpus(kvm, c - 1);
38 return false;
39 }
40 }
41
42 return true;
43 }
44
aia_create(struct kvm_device * dev,u32 type)45 static int aia_create(struct kvm_device *dev, u32 type)
46 {
47 int ret;
48 unsigned long i;
49 struct kvm *kvm = dev->kvm;
50 struct kvm_vcpu *vcpu;
51
52 if (irqchip_in_kernel(kvm))
53 return -EEXIST;
54
55 ret = -EBUSY;
56 if (!lock_all_vcpus(kvm))
57 return ret;
58
59 kvm_for_each_vcpu(i, vcpu, kvm) {
60 if (vcpu->arch.ran_atleast_once)
61 goto out_unlock;
62 }
63 ret = 0;
64
65 kvm->arch.aia.in_kernel = true;
66
67 out_unlock:
68 unlock_all_vcpus(kvm);
69 return ret;
70 }
71
aia_destroy(struct kvm_device * dev)72 static void aia_destroy(struct kvm_device *dev)
73 {
74 kfree(dev);
75 }
76
aia_config(struct kvm * kvm,unsigned long type,u32 * nr,bool write)77 static int aia_config(struct kvm *kvm, unsigned long type,
78 u32 *nr, bool write)
79 {
80 struct kvm_aia *aia = &kvm->arch.aia;
81
82 /* Writes can only be done before irqchip is initialized */
83 if (write && kvm_riscv_aia_initialized(kvm))
84 return -EBUSY;
85
86 switch (type) {
87 case KVM_DEV_RISCV_AIA_CONFIG_MODE:
88 if (write) {
89 switch (*nr) {
90 case KVM_DEV_RISCV_AIA_MODE_EMUL:
91 break;
92 case KVM_DEV_RISCV_AIA_MODE_HWACCEL:
93 case KVM_DEV_RISCV_AIA_MODE_AUTO:
94 /*
95 * HW Acceleration and Auto modes only
96 * supported on host with non-zero guest
97 * external interrupts (i.e. non-zero
98 * VS-level IMSIC pages).
99 */
100 if (!kvm_riscv_aia_nr_hgei)
101 return -EINVAL;
102 break;
103 default:
104 return -EINVAL;
105 }
106 aia->mode = *nr;
107 } else
108 *nr = aia->mode;
109 break;
110 case KVM_DEV_RISCV_AIA_CONFIG_IDS:
111 if (write) {
112 if ((*nr < KVM_DEV_RISCV_AIA_IDS_MIN) ||
113 (*nr >= KVM_DEV_RISCV_AIA_IDS_MAX) ||
114 ((*nr & KVM_DEV_RISCV_AIA_IDS_MIN) !=
115 KVM_DEV_RISCV_AIA_IDS_MIN) ||
116 (kvm_riscv_aia_max_ids <= *nr))
117 return -EINVAL;
118 aia->nr_ids = *nr;
119 } else
120 *nr = aia->nr_ids;
121 break;
122 case KVM_DEV_RISCV_AIA_CONFIG_SRCS:
123 if (write) {
124 if ((*nr >= KVM_DEV_RISCV_AIA_SRCS_MAX) ||
125 (*nr >= kvm_riscv_aia_max_ids))
126 return -EINVAL;
127 aia->nr_sources = *nr;
128 } else
129 *nr = aia->nr_sources;
130 break;
131 case KVM_DEV_RISCV_AIA_CONFIG_GROUP_BITS:
132 if (write) {
133 if (*nr >= KVM_DEV_RISCV_AIA_GROUP_BITS_MAX)
134 return -EINVAL;
135 aia->nr_group_bits = *nr;
136 } else
137 *nr = aia->nr_group_bits;
138 break;
139 case KVM_DEV_RISCV_AIA_CONFIG_GROUP_SHIFT:
140 if (write) {
141 if ((*nr < KVM_DEV_RISCV_AIA_GROUP_SHIFT_MIN) ||
142 (*nr >= KVM_DEV_RISCV_AIA_GROUP_SHIFT_MAX))
143 return -EINVAL;
144 aia->nr_group_shift = *nr;
145 } else
146 *nr = aia->nr_group_shift;
147 break;
148 case KVM_DEV_RISCV_AIA_CONFIG_HART_BITS:
149 if (write) {
150 if (*nr >= KVM_DEV_RISCV_AIA_HART_BITS_MAX)
151 return -EINVAL;
152 aia->nr_hart_bits = *nr;
153 } else
154 *nr = aia->nr_hart_bits;
155 break;
156 case KVM_DEV_RISCV_AIA_CONFIG_GUEST_BITS:
157 if (write) {
158 if (*nr >= KVM_DEV_RISCV_AIA_GUEST_BITS_MAX)
159 return -EINVAL;
160 aia->nr_guest_bits = *nr;
161 } else
162 *nr = aia->nr_guest_bits;
163 break;
164 default:
165 return -ENXIO;
166 }
167
168 return 0;
169 }
170
aia_aplic_addr(struct kvm * kvm,u64 * addr,bool write)171 static int aia_aplic_addr(struct kvm *kvm, u64 *addr, bool write)
172 {
173 struct kvm_aia *aia = &kvm->arch.aia;
174
175 if (write) {
176 /* Writes can only be done before irqchip is initialized */
177 if (kvm_riscv_aia_initialized(kvm))
178 return -EBUSY;
179
180 if (*addr & (KVM_DEV_RISCV_APLIC_ALIGN - 1))
181 return -EINVAL;
182
183 aia->aplic_addr = *addr;
184 } else
185 *addr = aia->aplic_addr;
186
187 return 0;
188 }
189
aia_imsic_addr(struct kvm * kvm,u64 * addr,unsigned long vcpu_idx,bool write)190 static int aia_imsic_addr(struct kvm *kvm, u64 *addr,
191 unsigned long vcpu_idx, bool write)
192 {
193 struct kvm_vcpu *vcpu;
194 struct kvm_vcpu_aia *vcpu_aia;
195
196 vcpu = kvm_get_vcpu(kvm, vcpu_idx);
197 if (!vcpu)
198 return -EINVAL;
199 vcpu_aia = &vcpu->arch.aia_context;
200
201 if (write) {
202 /* Writes can only be done before irqchip is initialized */
203 if (kvm_riscv_aia_initialized(kvm))
204 return -EBUSY;
205
206 if (*addr & (KVM_DEV_RISCV_IMSIC_ALIGN - 1))
207 return -EINVAL;
208 }
209
210 mutex_lock(&vcpu->mutex);
211 if (write)
212 vcpu_aia->imsic_addr = *addr;
213 else
214 *addr = vcpu_aia->imsic_addr;
215 mutex_unlock(&vcpu->mutex);
216
217 return 0;
218 }
219
aia_imsic_ppn(struct kvm_aia * aia,gpa_t addr)220 static gpa_t aia_imsic_ppn(struct kvm_aia *aia, gpa_t addr)
221 {
222 u32 h, l;
223 gpa_t mask = 0;
224
225 h = aia->nr_hart_bits + aia->nr_guest_bits +
226 IMSIC_MMIO_PAGE_SHIFT - 1;
227 mask = GENMASK_ULL(h, 0);
228
229 if (aia->nr_group_bits) {
230 h = aia->nr_group_bits + aia->nr_group_shift - 1;
231 l = aia->nr_group_shift;
232 mask |= GENMASK_ULL(h, l);
233 }
234
235 return (addr & ~mask) >> IMSIC_MMIO_PAGE_SHIFT;
236 }
237
aia_imsic_hart_index(struct kvm_aia * aia,gpa_t addr)238 static u32 aia_imsic_hart_index(struct kvm_aia *aia, gpa_t addr)
239 {
240 u32 hart = 0, group = 0;
241
242 if (aia->nr_hart_bits)
243 hart = (addr >> (aia->nr_guest_bits + IMSIC_MMIO_PAGE_SHIFT)) &
244 GENMASK_ULL(aia->nr_hart_bits - 1, 0);
245 if (aia->nr_group_bits)
246 group = (addr >> aia->nr_group_shift) &
247 GENMASK_ULL(aia->nr_group_bits - 1, 0);
248
249 return (group << aia->nr_hart_bits) | hart;
250 }
251
aia_init(struct kvm * kvm)252 static int aia_init(struct kvm *kvm)
253 {
254 int ret, i;
255 unsigned long idx;
256 struct kvm_vcpu *vcpu;
257 struct kvm_vcpu_aia *vaia;
258 struct kvm_aia *aia = &kvm->arch.aia;
259 gpa_t base_ppn = KVM_RISCV_AIA_UNDEF_ADDR;
260
261 /* Irqchip can be initialized only once */
262 if (kvm_riscv_aia_initialized(kvm))
263 return -EBUSY;
264
265 /* We might be in the middle of creating a VCPU? */
266 if (kvm->created_vcpus != atomic_read(&kvm->online_vcpus))
267 return -EBUSY;
268
269 /* Number of sources should be less than or equals number of IDs */
270 if (aia->nr_ids < aia->nr_sources)
271 return -EINVAL;
272
273 /* APLIC base is required for non-zero number of sources */
274 if (aia->nr_sources && aia->aplic_addr == KVM_RISCV_AIA_UNDEF_ADDR)
275 return -EINVAL;
276
277 /* Initialize APLIC */
278 ret = kvm_riscv_aia_aplic_init(kvm);
279 if (ret)
280 return ret;
281
282 /* Iterate over each VCPU */
283 kvm_for_each_vcpu(idx, vcpu, kvm) {
284 vaia = &vcpu->arch.aia_context;
285
286 /* IMSIC base is required */
287 if (vaia->imsic_addr == KVM_RISCV_AIA_UNDEF_ADDR) {
288 ret = -EINVAL;
289 goto fail_cleanup_imsics;
290 }
291
292 /* All IMSICs should have matching base PPN */
293 if (base_ppn == KVM_RISCV_AIA_UNDEF_ADDR)
294 base_ppn = aia_imsic_ppn(aia, vaia->imsic_addr);
295 if (base_ppn != aia_imsic_ppn(aia, vaia->imsic_addr)) {
296 ret = -EINVAL;
297 goto fail_cleanup_imsics;
298 }
299
300 /* Update HART index of the IMSIC based on IMSIC base */
301 vaia->hart_index = aia_imsic_hart_index(aia,
302 vaia->imsic_addr);
303
304 /* Initialize IMSIC for this VCPU */
305 ret = kvm_riscv_vcpu_aia_imsic_init(vcpu);
306 if (ret)
307 goto fail_cleanup_imsics;
308 }
309
310 /* Set the initialized flag */
311 kvm->arch.aia.initialized = true;
312
313 return 0;
314
315 fail_cleanup_imsics:
316 for (i = idx - 1; i >= 0; i--) {
317 vcpu = kvm_get_vcpu(kvm, i);
318 if (!vcpu)
319 continue;
320 kvm_riscv_vcpu_aia_imsic_cleanup(vcpu);
321 }
322 kvm_riscv_aia_aplic_cleanup(kvm);
323 return ret;
324 }
325
aia_set_attr(struct kvm_device * dev,struct kvm_device_attr * attr)326 static int aia_set_attr(struct kvm_device *dev, struct kvm_device_attr *attr)
327 {
328 u32 nr;
329 u64 addr;
330 int nr_vcpus, r = -ENXIO;
331 unsigned long v, type = (unsigned long)attr->attr;
332 void __user *uaddr = (void __user *)(long)attr->addr;
333
334 switch (attr->group) {
335 case KVM_DEV_RISCV_AIA_GRP_CONFIG:
336 if (copy_from_user(&nr, uaddr, sizeof(nr)))
337 return -EFAULT;
338
339 mutex_lock(&dev->kvm->lock);
340 r = aia_config(dev->kvm, type, &nr, true);
341 mutex_unlock(&dev->kvm->lock);
342
343 break;
344
345 case KVM_DEV_RISCV_AIA_GRP_ADDR:
346 if (copy_from_user(&addr, uaddr, sizeof(addr)))
347 return -EFAULT;
348
349 nr_vcpus = atomic_read(&dev->kvm->online_vcpus);
350 mutex_lock(&dev->kvm->lock);
351 if (type == KVM_DEV_RISCV_AIA_ADDR_APLIC)
352 r = aia_aplic_addr(dev->kvm, &addr, true);
353 else if (type < KVM_DEV_RISCV_AIA_ADDR_IMSIC(nr_vcpus))
354 r = aia_imsic_addr(dev->kvm, &addr,
355 type - KVM_DEV_RISCV_AIA_ADDR_IMSIC(0), true);
356 mutex_unlock(&dev->kvm->lock);
357
358 break;
359
360 case KVM_DEV_RISCV_AIA_GRP_CTRL:
361 switch (type) {
362 case KVM_DEV_RISCV_AIA_CTRL_INIT:
363 mutex_lock(&dev->kvm->lock);
364 r = aia_init(dev->kvm);
365 mutex_unlock(&dev->kvm->lock);
366 break;
367 }
368
369 break;
370 case KVM_DEV_RISCV_AIA_GRP_APLIC:
371 if (copy_from_user(&nr, uaddr, sizeof(nr)))
372 return -EFAULT;
373
374 mutex_lock(&dev->kvm->lock);
375 r = kvm_riscv_aia_aplic_set_attr(dev->kvm, type, nr);
376 mutex_unlock(&dev->kvm->lock);
377
378 break;
379 case KVM_DEV_RISCV_AIA_GRP_IMSIC:
380 if (copy_from_user(&v, uaddr, sizeof(v)))
381 return -EFAULT;
382
383 mutex_lock(&dev->kvm->lock);
384 r = kvm_riscv_aia_imsic_rw_attr(dev->kvm, type, true, &v);
385 mutex_unlock(&dev->kvm->lock);
386
387 break;
388 }
389
390 return r;
391 }
392
aia_get_attr(struct kvm_device * dev,struct kvm_device_attr * attr)393 static int aia_get_attr(struct kvm_device *dev, struct kvm_device_attr *attr)
394 {
395 u32 nr;
396 u64 addr;
397 int nr_vcpus, r = -ENXIO;
398 void __user *uaddr = (void __user *)(long)attr->addr;
399 unsigned long v, type = (unsigned long)attr->attr;
400
401 switch (attr->group) {
402 case KVM_DEV_RISCV_AIA_GRP_CONFIG:
403 if (copy_from_user(&nr, uaddr, sizeof(nr)))
404 return -EFAULT;
405
406 mutex_lock(&dev->kvm->lock);
407 r = aia_config(dev->kvm, type, &nr, false);
408 mutex_unlock(&dev->kvm->lock);
409 if (r)
410 return r;
411
412 if (copy_to_user(uaddr, &nr, sizeof(nr)))
413 return -EFAULT;
414
415 break;
416 case KVM_DEV_RISCV_AIA_GRP_ADDR:
417 if (copy_from_user(&addr, uaddr, sizeof(addr)))
418 return -EFAULT;
419
420 nr_vcpus = atomic_read(&dev->kvm->online_vcpus);
421 mutex_lock(&dev->kvm->lock);
422 if (type == KVM_DEV_RISCV_AIA_ADDR_APLIC)
423 r = aia_aplic_addr(dev->kvm, &addr, false);
424 else if (type < KVM_DEV_RISCV_AIA_ADDR_IMSIC(nr_vcpus))
425 r = aia_imsic_addr(dev->kvm, &addr,
426 type - KVM_DEV_RISCV_AIA_ADDR_IMSIC(0), false);
427 mutex_unlock(&dev->kvm->lock);
428 if (r)
429 return r;
430
431 if (copy_to_user(uaddr, &addr, sizeof(addr)))
432 return -EFAULT;
433
434 break;
435 case KVM_DEV_RISCV_AIA_GRP_APLIC:
436 if (copy_from_user(&nr, uaddr, sizeof(nr)))
437 return -EFAULT;
438
439 mutex_lock(&dev->kvm->lock);
440 r = kvm_riscv_aia_aplic_get_attr(dev->kvm, type, &nr);
441 mutex_unlock(&dev->kvm->lock);
442 if (r)
443 return r;
444
445 if (copy_to_user(uaddr, &nr, sizeof(nr)))
446 return -EFAULT;
447
448 break;
449 case KVM_DEV_RISCV_AIA_GRP_IMSIC:
450 if (copy_from_user(&v, uaddr, sizeof(v)))
451 return -EFAULT;
452
453 mutex_lock(&dev->kvm->lock);
454 r = kvm_riscv_aia_imsic_rw_attr(dev->kvm, type, false, &v);
455 mutex_unlock(&dev->kvm->lock);
456 if (r)
457 return r;
458
459 if (copy_to_user(uaddr, &v, sizeof(v)))
460 return -EFAULT;
461
462 break;
463 }
464
465 return r;
466 }
467
aia_has_attr(struct kvm_device * dev,struct kvm_device_attr * attr)468 static int aia_has_attr(struct kvm_device *dev, struct kvm_device_attr *attr)
469 {
470 int nr_vcpus;
471
472 switch (attr->group) {
473 case KVM_DEV_RISCV_AIA_GRP_CONFIG:
474 switch (attr->attr) {
475 case KVM_DEV_RISCV_AIA_CONFIG_MODE:
476 case KVM_DEV_RISCV_AIA_CONFIG_IDS:
477 case KVM_DEV_RISCV_AIA_CONFIG_SRCS:
478 case KVM_DEV_RISCV_AIA_CONFIG_GROUP_BITS:
479 case KVM_DEV_RISCV_AIA_CONFIG_GROUP_SHIFT:
480 case KVM_DEV_RISCV_AIA_CONFIG_HART_BITS:
481 case KVM_DEV_RISCV_AIA_CONFIG_GUEST_BITS:
482 return 0;
483 }
484 break;
485 case KVM_DEV_RISCV_AIA_GRP_ADDR:
486 nr_vcpus = atomic_read(&dev->kvm->online_vcpus);
487 if (attr->attr == KVM_DEV_RISCV_AIA_ADDR_APLIC)
488 return 0;
489 else if (attr->attr < KVM_DEV_RISCV_AIA_ADDR_IMSIC(nr_vcpus))
490 return 0;
491 break;
492 case KVM_DEV_RISCV_AIA_GRP_CTRL:
493 switch (attr->attr) {
494 case KVM_DEV_RISCV_AIA_CTRL_INIT:
495 return 0;
496 }
497 break;
498 case KVM_DEV_RISCV_AIA_GRP_APLIC:
499 return kvm_riscv_aia_aplic_has_attr(dev->kvm, attr->attr);
500 case KVM_DEV_RISCV_AIA_GRP_IMSIC:
501 return kvm_riscv_aia_imsic_has_attr(dev->kvm, attr->attr);
502 }
503
504 return -ENXIO;
505 }
506
507 struct kvm_device_ops kvm_riscv_aia_device_ops = {
508 .name = "kvm-riscv-aia",
509 .create = aia_create,
510 .destroy = aia_destroy,
511 .set_attr = aia_set_attr,
512 .get_attr = aia_get_attr,
513 .has_attr = aia_has_attr,
514 };
515
kvm_riscv_vcpu_aia_update(struct kvm_vcpu * vcpu)516 int kvm_riscv_vcpu_aia_update(struct kvm_vcpu *vcpu)
517 {
518 /* Proceed only if AIA was initialized successfully */
519 if (!kvm_riscv_aia_initialized(vcpu->kvm))
520 return 1;
521
522 /* Update the IMSIC HW state before entering guest mode */
523 return kvm_riscv_vcpu_aia_imsic_update(vcpu);
524 }
525
kvm_riscv_vcpu_aia_reset(struct kvm_vcpu * vcpu)526 void kvm_riscv_vcpu_aia_reset(struct kvm_vcpu *vcpu)
527 {
528 struct kvm_vcpu_aia_csr *csr = &vcpu->arch.aia_context.guest_csr;
529 struct kvm_vcpu_aia_csr *reset_csr =
530 &vcpu->arch.aia_context.guest_reset_csr;
531
532 if (!kvm_riscv_aia_available())
533 return;
534 memcpy(csr, reset_csr, sizeof(*csr));
535
536 /* Proceed only if AIA was initialized successfully */
537 if (!kvm_riscv_aia_initialized(vcpu->kvm))
538 return;
539
540 /* Reset the IMSIC context */
541 kvm_riscv_vcpu_aia_imsic_reset(vcpu);
542 }
543
kvm_riscv_vcpu_aia_init(struct kvm_vcpu * vcpu)544 int kvm_riscv_vcpu_aia_init(struct kvm_vcpu *vcpu)
545 {
546 struct kvm_vcpu_aia *vaia = &vcpu->arch.aia_context;
547
548 if (!kvm_riscv_aia_available())
549 return 0;
550
551 /*
552 * We don't do any memory allocations over here because these
553 * will be done after AIA device is initialized by the user-space.
554 *
555 * Refer, aia_init() implementation for more details.
556 */
557
558 /* Initialize default values in AIA vcpu context */
559 vaia->imsic_addr = KVM_RISCV_AIA_UNDEF_ADDR;
560 vaia->hart_index = vcpu->vcpu_idx;
561
562 return 0;
563 }
564
kvm_riscv_vcpu_aia_deinit(struct kvm_vcpu * vcpu)565 void kvm_riscv_vcpu_aia_deinit(struct kvm_vcpu *vcpu)
566 {
567 /* Proceed only if AIA was initialized successfully */
568 if (!kvm_riscv_aia_initialized(vcpu->kvm))
569 return;
570
571 /* Cleanup IMSIC context */
572 kvm_riscv_vcpu_aia_imsic_cleanup(vcpu);
573 }
574
kvm_riscv_aia_inject_msi_by_id(struct kvm * kvm,u32 hart_index,u32 guest_index,u32 iid)575 int kvm_riscv_aia_inject_msi_by_id(struct kvm *kvm, u32 hart_index,
576 u32 guest_index, u32 iid)
577 {
578 unsigned long idx;
579 struct kvm_vcpu *vcpu;
580
581 /* Proceed only if AIA was initialized successfully */
582 if (!kvm_riscv_aia_initialized(kvm))
583 return -EBUSY;
584
585 /* Inject MSI to matching VCPU */
586 kvm_for_each_vcpu(idx, vcpu, kvm) {
587 if (vcpu->arch.aia_context.hart_index == hart_index)
588 return kvm_riscv_vcpu_aia_imsic_inject(vcpu,
589 guest_index,
590 0, iid);
591 }
592
593 return 0;
594 }
595
kvm_riscv_aia_inject_msi(struct kvm * kvm,struct kvm_msi * msi)596 int kvm_riscv_aia_inject_msi(struct kvm *kvm, struct kvm_msi *msi)
597 {
598 gpa_t tppn, ippn;
599 unsigned long idx;
600 struct kvm_vcpu *vcpu;
601 u32 g, toff, iid = msi->data;
602 struct kvm_aia *aia = &kvm->arch.aia;
603 gpa_t target = (((gpa_t)msi->address_hi) << 32) | msi->address_lo;
604
605 /* Proceed only if AIA was initialized successfully */
606 if (!kvm_riscv_aia_initialized(kvm))
607 return -EBUSY;
608
609 /* Convert target address to target PPN */
610 tppn = target >> IMSIC_MMIO_PAGE_SHIFT;
611
612 /* Extract and clear Guest ID from target PPN */
613 g = tppn & (BIT(aia->nr_guest_bits) - 1);
614 tppn &= ~((gpa_t)(BIT(aia->nr_guest_bits) - 1));
615
616 /* Inject MSI to matching VCPU */
617 kvm_for_each_vcpu(idx, vcpu, kvm) {
618 ippn = vcpu->arch.aia_context.imsic_addr >>
619 IMSIC_MMIO_PAGE_SHIFT;
620 if (ippn == tppn) {
621 toff = target & (IMSIC_MMIO_PAGE_SZ - 1);
622 return kvm_riscv_vcpu_aia_imsic_inject(vcpu, g,
623 toff, iid);
624 }
625 }
626
627 return 0;
628 }
629
kvm_riscv_aia_inject_irq(struct kvm * kvm,unsigned int irq,bool level)630 int kvm_riscv_aia_inject_irq(struct kvm *kvm, unsigned int irq, bool level)
631 {
632 /* Proceed only if AIA was initialized successfully */
633 if (!kvm_riscv_aia_initialized(kvm))
634 return -EBUSY;
635
636 /* Inject interrupt level change in APLIC */
637 return kvm_riscv_aia_aplic_inject(kvm, irq, level);
638 }
639
kvm_riscv_aia_init_vm(struct kvm * kvm)640 void kvm_riscv_aia_init_vm(struct kvm *kvm)
641 {
642 struct kvm_aia *aia = &kvm->arch.aia;
643
644 if (!kvm_riscv_aia_available())
645 return;
646
647 /*
648 * We don't do any memory allocations over here because these
649 * will be done after AIA device is initialized by the user-space.
650 *
651 * Refer, aia_init() implementation for more details.
652 */
653
654 /* Initialize default values in AIA global context */
655 aia->mode = (kvm_riscv_aia_nr_hgei) ?
656 KVM_DEV_RISCV_AIA_MODE_AUTO : KVM_DEV_RISCV_AIA_MODE_EMUL;
657 aia->nr_ids = kvm_riscv_aia_max_ids - 1;
658 aia->nr_sources = 0;
659 aia->nr_group_bits = 0;
660 aia->nr_group_shift = KVM_DEV_RISCV_AIA_GROUP_SHIFT_MIN;
661 aia->nr_hart_bits = 0;
662 aia->nr_guest_bits = 0;
663 aia->aplic_addr = KVM_RISCV_AIA_UNDEF_ADDR;
664 }
665
kvm_riscv_aia_destroy_vm(struct kvm * kvm)666 void kvm_riscv_aia_destroy_vm(struct kvm *kvm)
667 {
668 /* Proceed only if AIA was initialized successfully */
669 if (!kvm_riscv_aia_initialized(kvm))
670 return;
671
672 /* Cleanup APLIC context */
673 kvm_riscv_aia_aplic_cleanup(kvm);
674 }
675