1// SPDX-License-Identifier: GPL-2.0
2/*
3 * Copyright (C) 2021 Western Digital Corporation or its affiliates.
4 * Copyright (C) 2022 Ventana Micro Systems Inc.
5 *
6 * Authors:
7 * Anup Patel <apatel@ventanamicro.com>
8 */
9
10#include <linux/bits.h>
11#include <linux/kvm_host.h>
12#include <linux/uaccess.h>
13#include <asm/kvm_aia_imsic.h>
14
15static void unlock_vcpus(struct kvm *kvm, int vcpu_lock_idx)
16{
17 struct kvm_vcpu *tmp_vcpu;
18
19 for (; vcpu_lock_idx >= 0; vcpu_lock_idx--) {
20 tmp_vcpu = kvm_get_vcpu(kvm, i: vcpu_lock_idx);
21 mutex_unlock(lock: &tmp_vcpu->mutex);
22 }
23}
24
25static void unlock_all_vcpus(struct kvm *kvm)
26{
27 unlock_vcpus(kvm, vcpu_lock_idx: atomic_read(v: &kvm->online_vcpus) - 1);
28}
29
30static bool lock_all_vcpus(struct kvm *kvm)
31{
32 struct kvm_vcpu *tmp_vcpu;
33 unsigned long c;
34
35 kvm_for_each_vcpu(c, tmp_vcpu, kvm) {
36 if (!mutex_trylock(lock: &tmp_vcpu->mutex)) {
37 unlock_vcpus(kvm, vcpu_lock_idx: c - 1);
38 return false;
39 }
40 }
41
42 return true;
43}
44
45static int aia_create(struct kvm_device *dev, u32 type)
46{
47 int ret;
48 unsigned long i;
49 struct kvm *kvm = dev->kvm;
50 struct kvm_vcpu *vcpu;
51
52 if (irqchip_in_kernel(kvm))
53 return -EEXIST;
54
55 ret = -EBUSY;
56 if (!lock_all_vcpus(kvm))
57 return ret;
58
59 kvm_for_each_vcpu(i, vcpu, kvm) {
60 if (vcpu->arch.ran_atleast_once)
61 goto out_unlock;
62 }
63 ret = 0;
64
65 kvm->arch.aia.in_kernel = true;
66
67out_unlock:
68 unlock_all_vcpus(kvm);
69 return ret;
70}
71
72static void aia_destroy(struct kvm_device *dev)
73{
74 kfree(objp: dev);
75}
76
77static int aia_config(struct kvm *kvm, unsigned long type,
78 u32 *nr, bool write)
79{
80 struct kvm_aia *aia = &kvm->arch.aia;
81
82 /* Writes can only be done before irqchip is initialized */
83 if (write && kvm_riscv_aia_initialized(kvm))
84 return -EBUSY;
85
86 switch (type) {
87 case KVM_DEV_RISCV_AIA_CONFIG_MODE:
88 if (write) {
89 switch (*nr) {
90 case KVM_DEV_RISCV_AIA_MODE_EMUL:
91 break;
92 case KVM_DEV_RISCV_AIA_MODE_HWACCEL:
93 case KVM_DEV_RISCV_AIA_MODE_AUTO:
94 /*
95 * HW Acceleration and Auto modes only
96 * supported on host with non-zero guest
97 * external interrupts (i.e. non-zero
98 * VS-level IMSIC pages).
99 */
100 if (!kvm_riscv_aia_nr_hgei)
101 return -EINVAL;
102 break;
103 default:
104 return -EINVAL;
105 }
106 aia->mode = *nr;
107 } else
108 *nr = aia->mode;
109 break;
110 case KVM_DEV_RISCV_AIA_CONFIG_IDS:
111 if (write) {
112 if ((*nr < KVM_DEV_RISCV_AIA_IDS_MIN) ||
113 (*nr >= KVM_DEV_RISCV_AIA_IDS_MAX) ||
114 ((*nr & KVM_DEV_RISCV_AIA_IDS_MIN) !=
115 KVM_DEV_RISCV_AIA_IDS_MIN) ||
116 (kvm_riscv_aia_max_ids <= *nr))
117 return -EINVAL;
118 aia->nr_ids = *nr;
119 } else
120 *nr = aia->nr_ids;
121 break;
122 case KVM_DEV_RISCV_AIA_CONFIG_SRCS:
123 if (write) {
124 if ((*nr >= KVM_DEV_RISCV_AIA_SRCS_MAX) ||
125 (*nr >= kvm_riscv_aia_max_ids))
126 return -EINVAL;
127 aia->nr_sources = *nr;
128 } else
129 *nr = aia->nr_sources;
130 break;
131 case KVM_DEV_RISCV_AIA_CONFIG_GROUP_BITS:
132 if (write) {
133 if (*nr >= KVM_DEV_RISCV_AIA_GROUP_BITS_MAX)
134 return -EINVAL;
135 aia->nr_group_bits = *nr;
136 } else
137 *nr = aia->nr_group_bits;
138 break;
139 case KVM_DEV_RISCV_AIA_CONFIG_GROUP_SHIFT:
140 if (write) {
141 if ((*nr < KVM_DEV_RISCV_AIA_GROUP_SHIFT_MIN) ||
142 (*nr >= KVM_DEV_RISCV_AIA_GROUP_SHIFT_MAX))
143 return -EINVAL;
144 aia->nr_group_shift = *nr;
145 } else
146 *nr = aia->nr_group_shift;
147 break;
148 case KVM_DEV_RISCV_AIA_CONFIG_HART_BITS:
149 if (write) {
150 if (*nr >= KVM_DEV_RISCV_AIA_HART_BITS_MAX)
151 return -EINVAL;
152 aia->nr_hart_bits = *nr;
153 } else
154 *nr = aia->nr_hart_bits;
155 break;
156 case KVM_DEV_RISCV_AIA_CONFIG_GUEST_BITS:
157 if (write) {
158 if (*nr >= KVM_DEV_RISCV_AIA_GUEST_BITS_MAX)
159 return -EINVAL;
160 aia->nr_guest_bits = *nr;
161 } else
162 *nr = aia->nr_guest_bits;
163 break;
164 default:
165 return -ENXIO;
166 }
167
168 return 0;
169}
170
171static int aia_aplic_addr(struct kvm *kvm, u64 *addr, bool write)
172{
173 struct kvm_aia *aia = &kvm->arch.aia;
174
175 if (write) {
176 /* Writes can only be done before irqchip is initialized */
177 if (kvm_riscv_aia_initialized(kvm))
178 return -EBUSY;
179
180 if (*addr & (KVM_DEV_RISCV_APLIC_ALIGN - 1))
181 return -EINVAL;
182
183 aia->aplic_addr = *addr;
184 } else
185 *addr = aia->aplic_addr;
186
187 return 0;
188}
189
190static int aia_imsic_addr(struct kvm *kvm, u64 *addr,
191 unsigned long vcpu_idx, bool write)
192{
193 struct kvm_vcpu *vcpu;
194 struct kvm_vcpu_aia *vcpu_aia;
195
196 vcpu = kvm_get_vcpu(kvm, i: vcpu_idx);
197 if (!vcpu)
198 return -EINVAL;
199 vcpu_aia = &vcpu->arch.aia_context;
200
201 if (write) {
202 /* Writes can only be done before irqchip is initialized */
203 if (kvm_riscv_aia_initialized(kvm))
204 return -EBUSY;
205
206 if (*addr & (KVM_DEV_RISCV_IMSIC_ALIGN - 1))
207 return -EINVAL;
208 }
209
210 mutex_lock(&vcpu->mutex);
211 if (write)
212 vcpu_aia->imsic_addr = *addr;
213 else
214 *addr = vcpu_aia->imsic_addr;
215 mutex_unlock(lock: &vcpu->mutex);
216
217 return 0;
218}
219
220static gpa_t aia_imsic_ppn(struct kvm_aia *aia, gpa_t addr)
221{
222 u32 h, l;
223 gpa_t mask = 0;
224
225 h = aia->nr_hart_bits + aia->nr_guest_bits +
226 IMSIC_MMIO_PAGE_SHIFT - 1;
227 mask = GENMASK_ULL(h, 0);
228
229 if (aia->nr_group_bits) {
230 h = aia->nr_group_bits + aia->nr_group_shift - 1;
231 l = aia->nr_group_shift;
232 mask |= GENMASK_ULL(h, l);
233 }
234
235 return (addr & ~mask) >> IMSIC_MMIO_PAGE_SHIFT;
236}
237
238static u32 aia_imsic_hart_index(struct kvm_aia *aia, gpa_t addr)
239{
240 u32 hart, group = 0;
241
242 hart = (addr >> (aia->nr_guest_bits + IMSIC_MMIO_PAGE_SHIFT)) &
243 GENMASK_ULL(aia->nr_hart_bits - 1, 0);
244 if (aia->nr_group_bits)
245 group = (addr >> aia->nr_group_shift) &
246 GENMASK_ULL(aia->nr_group_bits - 1, 0);
247
248 return (group << aia->nr_hart_bits) | hart;
249}
250
251static int aia_init(struct kvm *kvm)
252{
253 int ret, i;
254 unsigned long idx;
255 struct kvm_vcpu *vcpu;
256 struct kvm_vcpu_aia *vaia;
257 struct kvm_aia *aia = &kvm->arch.aia;
258 gpa_t base_ppn = KVM_RISCV_AIA_UNDEF_ADDR;
259
260 /* Irqchip can be initialized only once */
261 if (kvm_riscv_aia_initialized(kvm))
262 return -EBUSY;
263
264 /* We might be in the middle of creating a VCPU? */
265 if (kvm->created_vcpus != atomic_read(v: &kvm->online_vcpus))
266 return -EBUSY;
267
268 /* Number of sources should be less than or equals number of IDs */
269 if (aia->nr_ids < aia->nr_sources)
270 return -EINVAL;
271
272 /* APLIC base is required for non-zero number of sources */
273 if (aia->nr_sources && aia->aplic_addr == KVM_RISCV_AIA_UNDEF_ADDR)
274 return -EINVAL;
275
276 /* Initialize APLIC */
277 ret = kvm_riscv_aia_aplic_init(kvm);
278 if (ret)
279 return ret;
280
281 /* Iterate over each VCPU */
282 kvm_for_each_vcpu(idx, vcpu, kvm) {
283 vaia = &vcpu->arch.aia_context;
284
285 /* IMSIC base is required */
286 if (vaia->imsic_addr == KVM_RISCV_AIA_UNDEF_ADDR) {
287 ret = -EINVAL;
288 goto fail_cleanup_imsics;
289 }
290
291 /* All IMSICs should have matching base PPN */
292 if (base_ppn == KVM_RISCV_AIA_UNDEF_ADDR)
293 base_ppn = aia_imsic_ppn(aia, addr: vaia->imsic_addr);
294 if (base_ppn != aia_imsic_ppn(aia, addr: vaia->imsic_addr)) {
295 ret = -EINVAL;
296 goto fail_cleanup_imsics;
297 }
298
299 /* Update HART index of the IMSIC based on IMSIC base */
300 vaia->hart_index = aia_imsic_hart_index(aia,
301 addr: vaia->imsic_addr);
302
303 /* Initialize IMSIC for this VCPU */
304 ret = kvm_riscv_vcpu_aia_imsic_init(vcpu);
305 if (ret)
306 goto fail_cleanup_imsics;
307 }
308
309 /* Set the initialized flag */
310 kvm->arch.aia.initialized = true;
311
312 return 0;
313
314fail_cleanup_imsics:
315 for (i = idx - 1; i >= 0; i--) {
316 vcpu = kvm_get_vcpu(kvm, i);
317 if (!vcpu)
318 continue;
319 kvm_riscv_vcpu_aia_imsic_cleanup(vcpu);
320 }
321 kvm_riscv_aia_aplic_cleanup(kvm);
322 return ret;
323}
324
325static int aia_set_attr(struct kvm_device *dev, struct kvm_device_attr *attr)
326{
327 u32 nr;
328 u64 addr;
329 int nr_vcpus, r = -ENXIO;
330 unsigned long v, type = (unsigned long)attr->attr;
331 void __user *uaddr = (void __user *)(long)attr->addr;
332
333 switch (attr->group) {
334 case KVM_DEV_RISCV_AIA_GRP_CONFIG:
335 if (copy_from_user(to: &nr, from: uaddr, n: sizeof(nr)))
336 return -EFAULT;
337
338 mutex_lock(&dev->kvm->lock);
339 r = aia_config(kvm: dev->kvm, type, nr: &nr, write: true);
340 mutex_unlock(lock: &dev->kvm->lock);
341
342 break;
343
344 case KVM_DEV_RISCV_AIA_GRP_ADDR:
345 if (copy_from_user(to: &addr, from: uaddr, n: sizeof(addr)))
346 return -EFAULT;
347
348 nr_vcpus = atomic_read(v: &dev->kvm->online_vcpus);
349 mutex_lock(&dev->kvm->lock);
350 if (type == KVM_DEV_RISCV_AIA_ADDR_APLIC)
351 r = aia_aplic_addr(kvm: dev->kvm, addr: &addr, write: true);
352 else if (type < KVM_DEV_RISCV_AIA_ADDR_IMSIC(nr_vcpus))
353 r = aia_imsic_addr(kvm: dev->kvm, addr: &addr,
354 vcpu_idx: type - KVM_DEV_RISCV_AIA_ADDR_IMSIC(0), write: true);
355 mutex_unlock(lock: &dev->kvm->lock);
356
357 break;
358
359 case KVM_DEV_RISCV_AIA_GRP_CTRL:
360 switch (type) {
361 case KVM_DEV_RISCV_AIA_CTRL_INIT:
362 mutex_lock(&dev->kvm->lock);
363 r = aia_init(kvm: dev->kvm);
364 mutex_unlock(lock: &dev->kvm->lock);
365 break;
366 }
367
368 break;
369 case KVM_DEV_RISCV_AIA_GRP_APLIC:
370 if (copy_from_user(to: &nr, from: uaddr, n: sizeof(nr)))
371 return -EFAULT;
372
373 mutex_lock(&dev->kvm->lock);
374 r = kvm_riscv_aia_aplic_set_attr(dev->kvm, type, nr);
375 mutex_unlock(lock: &dev->kvm->lock);
376
377 break;
378 case KVM_DEV_RISCV_AIA_GRP_IMSIC:
379 if (copy_from_user(to: &v, from: uaddr, n: sizeof(v)))
380 return -EFAULT;
381
382 mutex_lock(&dev->kvm->lock);
383 r = kvm_riscv_aia_imsic_rw_attr(dev->kvm, type, true, &v);
384 mutex_unlock(lock: &dev->kvm->lock);
385
386 break;
387 }
388
389 return r;
390}
391
392static int aia_get_attr(struct kvm_device *dev, struct kvm_device_attr *attr)
393{
394 u32 nr;
395 u64 addr;
396 int nr_vcpus, r = -ENXIO;
397 void __user *uaddr = (void __user *)(long)attr->addr;
398 unsigned long v, type = (unsigned long)attr->attr;
399
400 switch (attr->group) {
401 case KVM_DEV_RISCV_AIA_GRP_CONFIG:
402 if (copy_from_user(to: &nr, from: uaddr, n: sizeof(nr)))
403 return -EFAULT;
404
405 mutex_lock(&dev->kvm->lock);
406 r = aia_config(kvm: dev->kvm, type, nr: &nr, write: false);
407 mutex_unlock(lock: &dev->kvm->lock);
408 if (r)
409 return r;
410
411 if (copy_to_user(to: uaddr, from: &nr, n: sizeof(nr)))
412 return -EFAULT;
413
414 break;
415 case KVM_DEV_RISCV_AIA_GRP_ADDR:
416 if (copy_from_user(to: &addr, from: uaddr, n: sizeof(addr)))
417 return -EFAULT;
418
419 nr_vcpus = atomic_read(v: &dev->kvm->online_vcpus);
420 mutex_lock(&dev->kvm->lock);
421 if (type == KVM_DEV_RISCV_AIA_ADDR_APLIC)
422 r = aia_aplic_addr(kvm: dev->kvm, addr: &addr, write: false);
423 else if (type < KVM_DEV_RISCV_AIA_ADDR_IMSIC(nr_vcpus))
424 r = aia_imsic_addr(kvm: dev->kvm, addr: &addr,
425 vcpu_idx: type - KVM_DEV_RISCV_AIA_ADDR_IMSIC(0), write: false);
426 mutex_unlock(lock: &dev->kvm->lock);
427 if (r)
428 return r;
429
430 if (copy_to_user(to: uaddr, from: &addr, n: sizeof(addr)))
431 return -EFAULT;
432
433 break;
434 case KVM_DEV_RISCV_AIA_GRP_APLIC:
435 if (copy_from_user(to: &nr, from: uaddr, n: sizeof(nr)))
436 return -EFAULT;
437
438 mutex_lock(&dev->kvm->lock);
439 r = kvm_riscv_aia_aplic_get_attr(dev->kvm, type, &nr);
440 mutex_unlock(lock: &dev->kvm->lock);
441 if (r)
442 return r;
443
444 if (copy_to_user(to: uaddr, from: &nr, n: sizeof(nr)))
445 return -EFAULT;
446
447 break;
448 case KVM_DEV_RISCV_AIA_GRP_IMSIC:
449 if (copy_from_user(to: &v, from: uaddr, n: sizeof(v)))
450 return -EFAULT;
451
452 mutex_lock(&dev->kvm->lock);
453 r = kvm_riscv_aia_imsic_rw_attr(dev->kvm, type, false, &v);
454 mutex_unlock(lock: &dev->kvm->lock);
455 if (r)
456 return r;
457
458 if (copy_to_user(to: uaddr, from: &v, n: sizeof(v)))
459 return -EFAULT;
460
461 break;
462 }
463
464 return r;
465}
466
467static int aia_has_attr(struct kvm_device *dev, struct kvm_device_attr *attr)
468{
469 int nr_vcpus;
470
471 switch (attr->group) {
472 case KVM_DEV_RISCV_AIA_GRP_CONFIG:
473 switch (attr->attr) {
474 case KVM_DEV_RISCV_AIA_CONFIG_MODE:
475 case KVM_DEV_RISCV_AIA_CONFIG_IDS:
476 case KVM_DEV_RISCV_AIA_CONFIG_SRCS:
477 case KVM_DEV_RISCV_AIA_CONFIG_GROUP_BITS:
478 case KVM_DEV_RISCV_AIA_CONFIG_GROUP_SHIFT:
479 case KVM_DEV_RISCV_AIA_CONFIG_HART_BITS:
480 case KVM_DEV_RISCV_AIA_CONFIG_GUEST_BITS:
481 return 0;
482 }
483 break;
484 case KVM_DEV_RISCV_AIA_GRP_ADDR:
485 nr_vcpus = atomic_read(v: &dev->kvm->online_vcpus);
486 if (attr->attr == KVM_DEV_RISCV_AIA_ADDR_APLIC)
487 return 0;
488 else if (attr->attr < KVM_DEV_RISCV_AIA_ADDR_IMSIC(nr_vcpus))
489 return 0;
490 break;
491 case KVM_DEV_RISCV_AIA_GRP_CTRL:
492 switch (attr->attr) {
493 case KVM_DEV_RISCV_AIA_CTRL_INIT:
494 return 0;
495 }
496 break;
497 case KVM_DEV_RISCV_AIA_GRP_APLIC:
498 return kvm_riscv_aia_aplic_has_attr(dev->kvm, attr->attr);
499 case KVM_DEV_RISCV_AIA_GRP_IMSIC:
500 return kvm_riscv_aia_imsic_has_attr(dev->kvm, attr->attr);
501 }
502
503 return -ENXIO;
504}
505
506struct kvm_device_ops kvm_riscv_aia_device_ops = {
507 .name = "kvm-riscv-aia",
508 .create = aia_create,
509 .destroy = aia_destroy,
510 .set_attr = aia_set_attr,
511 .get_attr = aia_get_attr,
512 .has_attr = aia_has_attr,
513};
514
515int kvm_riscv_vcpu_aia_update(struct kvm_vcpu *vcpu)
516{
517 /* Proceed only if AIA was initialized successfully */
518 if (!kvm_riscv_aia_initialized(vcpu->kvm))
519 return 1;
520
521 /* Update the IMSIC HW state before entering guest mode */
522 return kvm_riscv_vcpu_aia_imsic_update(vcpu);
523}
524
525void kvm_riscv_vcpu_aia_reset(struct kvm_vcpu *vcpu)
526{
527 struct kvm_vcpu_aia_csr *csr = &vcpu->arch.aia_context.guest_csr;
528 struct kvm_vcpu_aia_csr *reset_csr =
529 &vcpu->arch.aia_context.guest_reset_csr;
530
531 if (!kvm_riscv_aia_available())
532 return;
533 memcpy(csr, reset_csr, sizeof(*csr));
534
535 /* Proceed only if AIA was initialized successfully */
536 if (!kvm_riscv_aia_initialized(vcpu->kvm))
537 return;
538
539 /* Reset the IMSIC context */
540 kvm_riscv_vcpu_aia_imsic_reset(vcpu);
541}
542
543int kvm_riscv_vcpu_aia_init(struct kvm_vcpu *vcpu)
544{
545 struct kvm_vcpu_aia *vaia = &vcpu->arch.aia_context;
546
547 if (!kvm_riscv_aia_available())
548 return 0;
549
550 /*
551 * We don't do any memory allocations over here because these
552 * will be done after AIA device is initialized by the user-space.
553 *
554 * Refer, aia_init() implementation for more details.
555 */
556
557 /* Initialize default values in AIA vcpu context */
558 vaia->imsic_addr = KVM_RISCV_AIA_UNDEF_ADDR;
559 vaia->hart_index = vcpu->vcpu_idx;
560
561 return 0;
562}
563
564void kvm_riscv_vcpu_aia_deinit(struct kvm_vcpu *vcpu)
565{
566 /* Proceed only if AIA was initialized successfully */
567 if (!kvm_riscv_aia_initialized(vcpu->kvm))
568 return;
569
570 /* Cleanup IMSIC context */
571 kvm_riscv_vcpu_aia_imsic_cleanup(vcpu);
572}
573
574int kvm_riscv_aia_inject_msi_by_id(struct kvm *kvm, u32 hart_index,
575 u32 guest_index, u32 iid)
576{
577 unsigned long idx;
578 struct kvm_vcpu *vcpu;
579
580 /* Proceed only if AIA was initialized successfully */
581 if (!kvm_riscv_aia_initialized(kvm))
582 return -EBUSY;
583
584 /* Inject MSI to matching VCPU */
585 kvm_for_each_vcpu(idx, vcpu, kvm) {
586 if (vcpu->arch.aia_context.hart_index == hart_index)
587 return kvm_riscv_vcpu_aia_imsic_inject(vcpu,
588 guest_index,
589 0, iid);
590 }
591
592 return 0;
593}
594
595int kvm_riscv_aia_inject_msi(struct kvm *kvm, struct kvm_msi *msi)
596{
597 gpa_t tppn, ippn;
598 unsigned long idx;
599 struct kvm_vcpu *vcpu;
600 u32 g, toff, iid = msi->data;
601 struct kvm_aia *aia = &kvm->arch.aia;
602 gpa_t target = (((gpa_t)msi->address_hi) << 32) | msi->address_lo;
603
604 /* Proceed only if AIA was initialized successfully */
605 if (!kvm_riscv_aia_initialized(kvm))
606 return -EBUSY;
607
608 /* Convert target address to target PPN */
609 tppn = target >> IMSIC_MMIO_PAGE_SHIFT;
610
611 /* Extract and clear Guest ID from target PPN */
612 g = tppn & (BIT(aia->nr_guest_bits) - 1);
613 tppn &= ~((gpa_t)(BIT(aia->nr_guest_bits) - 1));
614
615 /* Inject MSI to matching VCPU */
616 kvm_for_each_vcpu(idx, vcpu, kvm) {
617 ippn = vcpu->arch.aia_context.imsic_addr >>
618 IMSIC_MMIO_PAGE_SHIFT;
619 if (ippn == tppn) {
620 toff = target & (IMSIC_MMIO_PAGE_SZ - 1);
621 return kvm_riscv_vcpu_aia_imsic_inject(vcpu, g,
622 toff, iid);
623 }
624 }
625
626 return 0;
627}
628
629int kvm_riscv_aia_inject_irq(struct kvm *kvm, unsigned int irq, bool level)
630{
631 /* Proceed only if AIA was initialized successfully */
632 if (!kvm_riscv_aia_initialized(kvm))
633 return -EBUSY;
634
635 /* Inject interrupt level change in APLIC */
636 return kvm_riscv_aia_aplic_inject(kvm, irq, level);
637}
638
639void kvm_riscv_aia_init_vm(struct kvm *kvm)
640{
641 struct kvm_aia *aia = &kvm->arch.aia;
642
643 if (!kvm_riscv_aia_available())
644 return;
645
646 /*
647 * We don't do any memory allocations over here because these
648 * will be done after AIA device is initialized by the user-space.
649 *
650 * Refer, aia_init() implementation for more details.
651 */
652
653 /* Initialize default values in AIA global context */
654 aia->mode = (kvm_riscv_aia_nr_hgei) ?
655 KVM_DEV_RISCV_AIA_MODE_AUTO : KVM_DEV_RISCV_AIA_MODE_EMUL;
656 aia->nr_ids = kvm_riscv_aia_max_ids - 1;
657 aia->nr_sources = 0;
658 aia->nr_group_bits = 0;
659 aia->nr_group_shift = KVM_DEV_RISCV_AIA_GROUP_SHIFT_MIN;
660 aia->nr_hart_bits = 0;
661 aia->nr_guest_bits = 0;
662 aia->aplic_addr = KVM_RISCV_AIA_UNDEF_ADDR;
663}
664
665void kvm_riscv_aia_destroy_vm(struct kvm *kvm)
666{
667 /* Proceed only if AIA was initialized successfully */
668 if (!kvm_riscv_aia_initialized(kvm))
669 return;
670
671 /* Cleanup APLIC context */
672 kvm_riscv_aia_aplic_cleanup(kvm);
673}
674

source code of linux/arch/riscv/kvm/aia_device.c