Merge tag 'hardening-v6.5-rc1-fixes' of git://git.kernel.org/pub/scm/linux/kernel...
[linux-block.git] / arch / riscv / kvm / aia_device.c
1 // SPDX-License-Identifier: GPL-2.0
2 /*
3  * Copyright (C) 2021 Western Digital Corporation or its affiliates.
4  * Copyright (C) 2022 Ventana Micro Systems Inc.
5  *
6  * Authors:
7  *      Anup Patel <apatel@ventanamicro.com>
8  */
9
10 #include <linux/bits.h>
11 #include <linux/kvm_host.h>
12 #include <linux/uaccess.h>
13 #include <asm/kvm_aia_imsic.h>
14
15 static void unlock_vcpus(struct kvm *kvm, int vcpu_lock_idx)
16 {
17         struct kvm_vcpu *tmp_vcpu;
18
19         for (; vcpu_lock_idx >= 0; vcpu_lock_idx--) {
20                 tmp_vcpu = kvm_get_vcpu(kvm, vcpu_lock_idx);
21                 mutex_unlock(&tmp_vcpu->mutex);
22         }
23 }
24
25 static void unlock_all_vcpus(struct kvm *kvm)
26 {
27         unlock_vcpus(kvm, atomic_read(&kvm->online_vcpus) - 1);
28 }
29
30 static bool lock_all_vcpus(struct kvm *kvm)
31 {
32         struct kvm_vcpu *tmp_vcpu;
33         unsigned long c;
34
35         kvm_for_each_vcpu(c, tmp_vcpu, kvm) {
36                 if (!mutex_trylock(&tmp_vcpu->mutex)) {
37                         unlock_vcpus(kvm, c - 1);
38                         return false;
39                 }
40         }
41
42         return true;
43 }
44
45 static int aia_create(struct kvm_device *dev, u32 type)
46 {
47         int ret;
48         unsigned long i;
49         struct kvm *kvm = dev->kvm;
50         struct kvm_vcpu *vcpu;
51
52         if (irqchip_in_kernel(kvm))
53                 return -EEXIST;
54
55         ret = -EBUSY;
56         if (!lock_all_vcpus(kvm))
57                 return ret;
58
59         kvm_for_each_vcpu(i, vcpu, kvm) {
60                 if (vcpu->arch.ran_atleast_once)
61                         goto out_unlock;
62         }
63         ret = 0;
64
65         kvm->arch.aia.in_kernel = true;
66
67 out_unlock:
68         unlock_all_vcpus(kvm);
69         return ret;
70 }
71
72 static void aia_destroy(struct kvm_device *dev)
73 {
74         kfree(dev);
75 }
76
77 static int aia_config(struct kvm *kvm, unsigned long type,
78                       u32 *nr, bool write)
79 {
80         struct kvm_aia *aia = &kvm->arch.aia;
81
82         /* Writes can only be done before irqchip is initialized */
83         if (write && kvm_riscv_aia_initialized(kvm))
84                 return -EBUSY;
85
86         switch (type) {
87         case KVM_DEV_RISCV_AIA_CONFIG_MODE:
88                 if (write) {
89                         switch (*nr) {
90                         case KVM_DEV_RISCV_AIA_MODE_EMUL:
91                                 break;
92                         case KVM_DEV_RISCV_AIA_MODE_HWACCEL:
93                         case KVM_DEV_RISCV_AIA_MODE_AUTO:
94                                 /*
95                                  * HW Acceleration and Auto modes only
96                                  * supported on host with non-zero guest
97                                  * external interrupts (i.e. non-zero
98                                  * VS-level IMSIC pages).
99                                  */
100                                 if (!kvm_riscv_aia_nr_hgei)
101                                         return -EINVAL;
102                                 break;
103                         default:
104                                 return -EINVAL;
105                         }
106                         aia->mode = *nr;
107                 } else
108                         *nr = aia->mode;
109                 break;
110         case KVM_DEV_RISCV_AIA_CONFIG_IDS:
111                 if (write) {
112                         if ((*nr < KVM_DEV_RISCV_AIA_IDS_MIN) ||
113                             (*nr >= KVM_DEV_RISCV_AIA_IDS_MAX) ||
114                             ((*nr & KVM_DEV_RISCV_AIA_IDS_MIN) !=
115                              KVM_DEV_RISCV_AIA_IDS_MIN) ||
116                             (kvm_riscv_aia_max_ids <= *nr))
117                                 return -EINVAL;
118                         aia->nr_ids = *nr;
119                 } else
120                         *nr = aia->nr_ids;
121                 break;
122         case KVM_DEV_RISCV_AIA_CONFIG_SRCS:
123                 if (write) {
124                         if ((*nr >= KVM_DEV_RISCV_AIA_SRCS_MAX) ||
125                             (*nr >= kvm_riscv_aia_max_ids))
126                                 return -EINVAL;
127                         aia->nr_sources = *nr;
128                 } else
129                         *nr = aia->nr_sources;
130                 break;
131         case KVM_DEV_RISCV_AIA_CONFIG_GROUP_BITS:
132                 if (write) {
133                         if (*nr >= KVM_DEV_RISCV_AIA_GROUP_BITS_MAX)
134                                 return -EINVAL;
135                         aia->nr_group_bits = *nr;
136                 } else
137                         *nr = aia->nr_group_bits;
138                 break;
139         case KVM_DEV_RISCV_AIA_CONFIG_GROUP_SHIFT:
140                 if (write) {
141                         if ((*nr < KVM_DEV_RISCV_AIA_GROUP_SHIFT_MIN) ||
142                             (*nr >= KVM_DEV_RISCV_AIA_GROUP_SHIFT_MAX))
143                                 return -EINVAL;
144                         aia->nr_group_shift = *nr;
145                 } else
146                         *nr = aia->nr_group_shift;
147                 break;
148         case KVM_DEV_RISCV_AIA_CONFIG_HART_BITS:
149                 if (write) {
150                         if (*nr >= KVM_DEV_RISCV_AIA_HART_BITS_MAX)
151                                 return -EINVAL;
152                         aia->nr_hart_bits = *nr;
153                 } else
154                         *nr = aia->nr_hart_bits;
155                 break;
156         case KVM_DEV_RISCV_AIA_CONFIG_GUEST_BITS:
157                 if (write) {
158                         if (*nr >= KVM_DEV_RISCV_AIA_GUEST_BITS_MAX)
159                                 return -EINVAL;
160                         aia->nr_guest_bits = *nr;
161                 } else
162                         *nr = aia->nr_guest_bits;
163                 break;
164         default:
165                 return -ENXIO;
166         }
167
168         return 0;
169 }
170
171 static int aia_aplic_addr(struct kvm *kvm, u64 *addr, bool write)
172 {
173         struct kvm_aia *aia = &kvm->arch.aia;
174
175         if (write) {
176                 /* Writes can only be done before irqchip is initialized */
177                 if (kvm_riscv_aia_initialized(kvm))
178                         return -EBUSY;
179
180                 if (*addr & (KVM_DEV_RISCV_APLIC_ALIGN - 1))
181                         return -EINVAL;
182
183                 aia->aplic_addr = *addr;
184         } else
185                 *addr = aia->aplic_addr;
186
187         return 0;
188 }
189
190 static int aia_imsic_addr(struct kvm *kvm, u64 *addr,
191                           unsigned long vcpu_idx, bool write)
192 {
193         struct kvm_vcpu *vcpu;
194         struct kvm_vcpu_aia *vcpu_aia;
195
196         vcpu = kvm_get_vcpu(kvm, vcpu_idx);
197         if (!vcpu)
198                 return -EINVAL;
199         vcpu_aia = &vcpu->arch.aia_context;
200
201         if (write) {
202                 /* Writes can only be done before irqchip is initialized */
203                 if (kvm_riscv_aia_initialized(kvm))
204                         return -EBUSY;
205
206                 if (*addr & (KVM_DEV_RISCV_IMSIC_ALIGN - 1))
207                         return -EINVAL;
208         }
209
210         mutex_lock(&vcpu->mutex);
211         if (write)
212                 vcpu_aia->imsic_addr = *addr;
213         else
214                 *addr = vcpu_aia->imsic_addr;
215         mutex_unlock(&vcpu->mutex);
216
217         return 0;
218 }
219
220 static gpa_t aia_imsic_ppn(struct kvm_aia *aia, gpa_t addr)
221 {
222         u32 h, l;
223         gpa_t mask = 0;
224
225         h = aia->nr_hart_bits + aia->nr_guest_bits +
226             IMSIC_MMIO_PAGE_SHIFT - 1;
227         mask = GENMASK_ULL(h, 0);
228
229         if (aia->nr_group_bits) {
230                 h = aia->nr_group_bits + aia->nr_group_shift - 1;
231                 l = aia->nr_group_shift;
232                 mask |= GENMASK_ULL(h, l);
233         }
234
235         return (addr & ~mask) >> IMSIC_MMIO_PAGE_SHIFT;
236 }
237
238 static u32 aia_imsic_hart_index(struct kvm_aia *aia, gpa_t addr)
239 {
240         u32 hart, group = 0;
241
242         hart = (addr >> (aia->nr_guest_bits + IMSIC_MMIO_PAGE_SHIFT)) &
243                 GENMASK_ULL(aia->nr_hart_bits - 1, 0);
244         if (aia->nr_group_bits)
245                 group = (addr >> aia->nr_group_shift) &
246                         GENMASK_ULL(aia->nr_group_bits - 1, 0);
247
248         return (group << aia->nr_hart_bits) | hart;
249 }
250
251 static int aia_init(struct kvm *kvm)
252 {
253         int ret, i;
254         unsigned long idx;
255         struct kvm_vcpu *vcpu;
256         struct kvm_vcpu_aia *vaia;
257         struct kvm_aia *aia = &kvm->arch.aia;
258         gpa_t base_ppn = KVM_RISCV_AIA_UNDEF_ADDR;
259
260         /* Irqchip can be initialized only once */
261         if (kvm_riscv_aia_initialized(kvm))
262                 return -EBUSY;
263
264         /* We might be in the middle of creating a VCPU? */
265         if (kvm->created_vcpus != atomic_read(&kvm->online_vcpus))
266                 return -EBUSY;
267
268         /* Number of sources should be less than or equals number of IDs */
269         if (aia->nr_ids < aia->nr_sources)
270                 return -EINVAL;
271
272         /* APLIC base is required for non-zero number of sources */
273         if (aia->nr_sources && aia->aplic_addr == KVM_RISCV_AIA_UNDEF_ADDR)
274                 return -EINVAL;
275
276         /* Initialize APLIC */
277         ret = kvm_riscv_aia_aplic_init(kvm);
278         if (ret)
279                 return ret;
280
281         /* Iterate over each VCPU */
282         kvm_for_each_vcpu(idx, vcpu, kvm) {
283                 vaia = &vcpu->arch.aia_context;
284
285                 /* IMSIC base is required */
286                 if (vaia->imsic_addr == KVM_RISCV_AIA_UNDEF_ADDR) {
287                         ret = -EINVAL;
288                         goto fail_cleanup_imsics;
289                 }
290
291                 /* All IMSICs should have matching base PPN */
292                 if (base_ppn == KVM_RISCV_AIA_UNDEF_ADDR)
293                         base_ppn = aia_imsic_ppn(aia, vaia->imsic_addr);
294                 if (base_ppn != aia_imsic_ppn(aia, vaia->imsic_addr)) {
295                         ret = -EINVAL;
296                         goto fail_cleanup_imsics;
297                 }
298
299                 /* Update HART index of the IMSIC based on IMSIC base */
300                 vaia->hart_index = aia_imsic_hart_index(aia,
301                                                         vaia->imsic_addr);
302
303                 /* Initialize IMSIC for this VCPU */
304                 ret = kvm_riscv_vcpu_aia_imsic_init(vcpu);
305                 if (ret)
306                         goto fail_cleanup_imsics;
307         }
308
309         /* Set the initialized flag */
310         kvm->arch.aia.initialized = true;
311
312         return 0;
313
314 fail_cleanup_imsics:
315         for (i = idx - 1; i >= 0; i--) {
316                 vcpu = kvm_get_vcpu(kvm, i);
317                 if (!vcpu)
318                         continue;
319                 kvm_riscv_vcpu_aia_imsic_cleanup(vcpu);
320         }
321         kvm_riscv_aia_aplic_cleanup(kvm);
322         return ret;
323 }
324
325 static int aia_set_attr(struct kvm_device *dev, struct kvm_device_attr *attr)
326 {
327         u32 nr;
328         u64 addr;
329         int nr_vcpus, r = -ENXIO;
330         unsigned long v, type = (unsigned long)attr->attr;
331         void __user *uaddr = (void __user *)(long)attr->addr;
332
333         switch (attr->group) {
334         case KVM_DEV_RISCV_AIA_GRP_CONFIG:
335                 if (copy_from_user(&nr, uaddr, sizeof(nr)))
336                         return -EFAULT;
337
338                 mutex_lock(&dev->kvm->lock);
339                 r = aia_config(dev->kvm, type, &nr, true);
340                 mutex_unlock(&dev->kvm->lock);
341
342                 break;
343
344         case KVM_DEV_RISCV_AIA_GRP_ADDR:
345                 if (copy_from_user(&addr, uaddr, sizeof(addr)))
346                         return -EFAULT;
347
348                 nr_vcpus = atomic_read(&dev->kvm->online_vcpus);
349                 mutex_lock(&dev->kvm->lock);
350                 if (type == KVM_DEV_RISCV_AIA_ADDR_APLIC)
351                         r = aia_aplic_addr(dev->kvm, &addr, true);
352                 else if (type < KVM_DEV_RISCV_AIA_ADDR_IMSIC(nr_vcpus))
353                         r = aia_imsic_addr(dev->kvm, &addr,
354                             type - KVM_DEV_RISCV_AIA_ADDR_IMSIC(0), true);
355                 mutex_unlock(&dev->kvm->lock);
356
357                 break;
358
359         case KVM_DEV_RISCV_AIA_GRP_CTRL:
360                 switch (type) {
361                 case KVM_DEV_RISCV_AIA_CTRL_INIT:
362                         mutex_lock(&dev->kvm->lock);
363                         r = aia_init(dev->kvm);
364                         mutex_unlock(&dev->kvm->lock);
365                         break;
366                 }
367
368                 break;
369         case KVM_DEV_RISCV_AIA_GRP_APLIC:
370                 if (copy_from_user(&nr, uaddr, sizeof(nr)))
371                         return -EFAULT;
372
373                 mutex_lock(&dev->kvm->lock);
374                 r = kvm_riscv_aia_aplic_set_attr(dev->kvm, type, nr);
375                 mutex_unlock(&dev->kvm->lock);
376
377                 break;
378         case KVM_DEV_RISCV_AIA_GRP_IMSIC:
379                 if (copy_from_user(&v, uaddr, sizeof(v)))
380                         return -EFAULT;
381
382                 mutex_lock(&dev->kvm->lock);
383                 r = kvm_riscv_aia_imsic_rw_attr(dev->kvm, type, true, &v);
384                 mutex_unlock(&dev->kvm->lock);
385
386                 break;
387         }
388
389         return r;
390 }
391
392 static int aia_get_attr(struct kvm_device *dev, struct kvm_device_attr *attr)
393 {
394         u32 nr;
395         u64 addr;
396         int nr_vcpus, r = -ENXIO;
397         void __user *uaddr = (void __user *)(long)attr->addr;
398         unsigned long v, type = (unsigned long)attr->attr;
399
400         switch (attr->group) {
401         case KVM_DEV_RISCV_AIA_GRP_CONFIG:
402                 if (copy_from_user(&nr, uaddr, sizeof(nr)))
403                         return -EFAULT;
404
405                 mutex_lock(&dev->kvm->lock);
406                 r = aia_config(dev->kvm, type, &nr, false);
407                 mutex_unlock(&dev->kvm->lock);
408                 if (r)
409                         return r;
410
411                 if (copy_to_user(uaddr, &nr, sizeof(nr)))
412                         return -EFAULT;
413
414                 break;
415         case KVM_DEV_RISCV_AIA_GRP_ADDR:
416                 if (copy_from_user(&addr, uaddr, sizeof(addr)))
417                         return -EFAULT;
418
419                 nr_vcpus = atomic_read(&dev->kvm->online_vcpus);
420                 mutex_lock(&dev->kvm->lock);
421                 if (type == KVM_DEV_RISCV_AIA_ADDR_APLIC)
422                         r = aia_aplic_addr(dev->kvm, &addr, false);
423                 else if (type < KVM_DEV_RISCV_AIA_ADDR_IMSIC(nr_vcpus))
424                         r = aia_imsic_addr(dev->kvm, &addr,
425                             type - KVM_DEV_RISCV_AIA_ADDR_IMSIC(0), false);
426                 mutex_unlock(&dev->kvm->lock);
427                 if (r)
428                         return r;
429
430                 if (copy_to_user(uaddr, &addr, sizeof(addr)))
431                         return -EFAULT;
432
433                 break;
434         case KVM_DEV_RISCV_AIA_GRP_APLIC:
435                 if (copy_from_user(&nr, uaddr, sizeof(nr)))
436                         return -EFAULT;
437
438                 mutex_lock(&dev->kvm->lock);
439                 r = kvm_riscv_aia_aplic_get_attr(dev->kvm, type, &nr);
440                 mutex_unlock(&dev->kvm->lock);
441                 if (r)
442                         return r;
443
444                 if (copy_to_user(uaddr, &nr, sizeof(nr)))
445                         return -EFAULT;
446
447                 break;
448         case KVM_DEV_RISCV_AIA_GRP_IMSIC:
449                 if (copy_from_user(&v, uaddr, sizeof(v)))
450                         return -EFAULT;
451
452                 mutex_lock(&dev->kvm->lock);
453                 r = kvm_riscv_aia_imsic_rw_attr(dev->kvm, type, false, &v);
454                 mutex_unlock(&dev->kvm->lock);
455                 if (r)
456                         return r;
457
458                 if (copy_to_user(uaddr, &v, sizeof(v)))
459                         return -EFAULT;
460
461                 break;
462         }
463
464         return r;
465 }
466
467 static int aia_has_attr(struct kvm_device *dev, struct kvm_device_attr *attr)
468 {
469         int nr_vcpus;
470
471         switch (attr->group) {
472         case KVM_DEV_RISCV_AIA_GRP_CONFIG:
473                 switch (attr->attr) {
474                 case KVM_DEV_RISCV_AIA_CONFIG_MODE:
475                 case KVM_DEV_RISCV_AIA_CONFIG_IDS:
476                 case KVM_DEV_RISCV_AIA_CONFIG_SRCS:
477                 case KVM_DEV_RISCV_AIA_CONFIG_GROUP_BITS:
478                 case KVM_DEV_RISCV_AIA_CONFIG_GROUP_SHIFT:
479                 case KVM_DEV_RISCV_AIA_CONFIG_HART_BITS:
480                 case KVM_DEV_RISCV_AIA_CONFIG_GUEST_BITS:
481                         return 0;
482                 }
483                 break;
484         case KVM_DEV_RISCV_AIA_GRP_ADDR:
485                 nr_vcpus = atomic_read(&dev->kvm->online_vcpus);
486                 if (attr->attr == KVM_DEV_RISCV_AIA_ADDR_APLIC)
487                         return 0;
488                 else if (attr->attr < KVM_DEV_RISCV_AIA_ADDR_IMSIC(nr_vcpus))
489                         return 0;
490                 break;
491         case KVM_DEV_RISCV_AIA_GRP_CTRL:
492                 switch (attr->attr) {
493                 case KVM_DEV_RISCV_AIA_CTRL_INIT:
494                         return 0;
495                 }
496                 break;
497         case KVM_DEV_RISCV_AIA_GRP_APLIC:
498                 return kvm_riscv_aia_aplic_has_attr(dev->kvm, attr->attr);
499         case KVM_DEV_RISCV_AIA_GRP_IMSIC:
500                 return kvm_riscv_aia_imsic_has_attr(dev->kvm, attr->attr);
501         }
502
503         return -ENXIO;
504 }
505
506 struct kvm_device_ops kvm_riscv_aia_device_ops = {
507         .name = "kvm-riscv-aia",
508         .create = aia_create,
509         .destroy = aia_destroy,
510         .set_attr = aia_set_attr,
511         .get_attr = aia_get_attr,
512         .has_attr = aia_has_attr,
513 };
514
515 int kvm_riscv_vcpu_aia_update(struct kvm_vcpu *vcpu)
516 {
517         /* Proceed only if AIA was initialized successfully */
518         if (!kvm_riscv_aia_initialized(vcpu->kvm))
519                 return 1;
520
521         /* Update the IMSIC HW state before entering guest mode */
522         return kvm_riscv_vcpu_aia_imsic_update(vcpu);
523 }
524
525 void kvm_riscv_vcpu_aia_reset(struct kvm_vcpu *vcpu)
526 {
527         struct kvm_vcpu_aia_csr *csr = &vcpu->arch.aia_context.guest_csr;
528         struct kvm_vcpu_aia_csr *reset_csr =
529                                 &vcpu->arch.aia_context.guest_reset_csr;
530
531         if (!kvm_riscv_aia_available())
532                 return;
533         memcpy(csr, reset_csr, sizeof(*csr));
534
535         /* Proceed only if AIA was initialized successfully */
536         if (!kvm_riscv_aia_initialized(vcpu->kvm))
537                 return;
538
539         /* Reset the IMSIC context */
540         kvm_riscv_vcpu_aia_imsic_reset(vcpu);
541 }
542
543 int kvm_riscv_vcpu_aia_init(struct kvm_vcpu *vcpu)
544 {
545         struct kvm_vcpu_aia *vaia = &vcpu->arch.aia_context;
546
547         if (!kvm_riscv_aia_available())
548                 return 0;
549
550         /*
551          * We don't do any memory allocations over here because these
552          * will be done after AIA device is initialized by the user-space.
553          *
554          * Refer, aia_init() implementation for more details.
555          */
556
557         /* Initialize default values in AIA vcpu context */
558         vaia->imsic_addr = KVM_RISCV_AIA_UNDEF_ADDR;
559         vaia->hart_index = vcpu->vcpu_idx;
560
561         return 0;
562 }
563
564 void kvm_riscv_vcpu_aia_deinit(struct kvm_vcpu *vcpu)
565 {
566         /* Proceed only if AIA was initialized successfully */
567         if (!kvm_riscv_aia_initialized(vcpu->kvm))
568                 return;
569
570         /* Cleanup IMSIC context */
571         kvm_riscv_vcpu_aia_imsic_cleanup(vcpu);
572 }
573
574 int kvm_riscv_aia_inject_msi_by_id(struct kvm *kvm, u32 hart_index,
575                                    u32 guest_index, u32 iid)
576 {
577         unsigned long idx;
578         struct kvm_vcpu *vcpu;
579
580         /* Proceed only if AIA was initialized successfully */
581         if (!kvm_riscv_aia_initialized(kvm))
582                 return -EBUSY;
583
584         /* Inject MSI to matching VCPU */
585         kvm_for_each_vcpu(idx, vcpu, kvm) {
586                 if (vcpu->arch.aia_context.hart_index == hart_index)
587                         return kvm_riscv_vcpu_aia_imsic_inject(vcpu,
588                                                                guest_index,
589                                                                0, iid);
590         }
591
592         return 0;
593 }
594
595 int kvm_riscv_aia_inject_msi(struct kvm *kvm, struct kvm_msi *msi)
596 {
597         gpa_t tppn, ippn;
598         unsigned long idx;
599         struct kvm_vcpu *vcpu;
600         u32 g, toff, iid = msi->data;
601         struct kvm_aia *aia = &kvm->arch.aia;
602         gpa_t target = (((gpa_t)msi->address_hi) << 32) | msi->address_lo;
603
604         /* Proceed only if AIA was initialized successfully */
605         if (!kvm_riscv_aia_initialized(kvm))
606                 return -EBUSY;
607
608         /* Convert target address to target PPN */
609         tppn = target >> IMSIC_MMIO_PAGE_SHIFT;
610
611         /* Extract and clear Guest ID from target PPN */
612         g = tppn & (BIT(aia->nr_guest_bits) - 1);
613         tppn &= ~((gpa_t)(BIT(aia->nr_guest_bits) - 1));
614
615         /* Inject MSI to matching VCPU */
616         kvm_for_each_vcpu(idx, vcpu, kvm) {
617                 ippn = vcpu->arch.aia_context.imsic_addr >>
618                                         IMSIC_MMIO_PAGE_SHIFT;
619                 if (ippn == tppn) {
620                         toff = target & (IMSIC_MMIO_PAGE_SZ - 1);
621                         return kvm_riscv_vcpu_aia_imsic_inject(vcpu, g,
622                                                                toff, iid);
623                 }
624         }
625
626         return 0;
627 }
628
629 int kvm_riscv_aia_inject_irq(struct kvm *kvm, unsigned int irq, bool level)
630 {
631         /* Proceed only if AIA was initialized successfully */
632         if (!kvm_riscv_aia_initialized(kvm))
633                 return -EBUSY;
634
635         /* Inject interrupt level change in APLIC */
636         return kvm_riscv_aia_aplic_inject(kvm, irq, level);
637 }
638
639 void kvm_riscv_aia_init_vm(struct kvm *kvm)
640 {
641         struct kvm_aia *aia = &kvm->arch.aia;
642
643         if (!kvm_riscv_aia_available())
644                 return;
645
646         /*
647          * We don't do any memory allocations over here because these
648          * will be done after AIA device is initialized by the user-space.
649          *
650          * Refer, aia_init() implementation for more details.
651          */
652
653         /* Initialize default values in AIA global context */
654         aia->mode = (kvm_riscv_aia_nr_hgei) ?
655                 KVM_DEV_RISCV_AIA_MODE_AUTO : KVM_DEV_RISCV_AIA_MODE_EMUL;
656         aia->nr_ids = kvm_riscv_aia_max_ids - 1;
657         aia->nr_sources = 0;
658         aia->nr_group_bits = 0;
659         aia->nr_group_shift = KVM_DEV_RISCV_AIA_GROUP_SHIFT_MIN;
660         aia->nr_hart_bits = 0;
661         aia->nr_guest_bits = 0;
662         aia->aplic_addr = KVM_RISCV_AIA_UNDEF_ADDR;
663 }
664
665 void kvm_riscv_aia_destroy_vm(struct kvm *kvm)
666 {
667         /* Proceed only if AIA was initialized successfully */
668         if (!kvm_riscv_aia_initialized(kvm))
669                 return;
670
671         /* Cleanup APLIC context */
672         kvm_riscv_aia_aplic_cleanup(kvm);
673 }