~ [ source navigation ] ~ [ diff markup ] ~ [ identifier search ] ~

TOMOYO Linux Cross Reference
Linux/arch/riscv/kvm/aia_device.c

Version: ~ [ linux-6.11.5 ] ~ [ linux-6.10.14 ] ~ [ linux-6.9.12 ] ~ [ linux-6.8.12 ] ~ [ linux-6.7.12 ] ~ [ linux-6.6.58 ] ~ [ linux-6.5.13 ] ~ [ linux-6.4.16 ] ~ [ linux-6.3.13 ] ~ [ linux-6.2.16 ] ~ [ linux-6.1.114 ] ~ [ linux-6.0.19 ] ~ [ linux-5.19.17 ] ~ [ linux-5.18.19 ] ~ [ linux-5.17.15 ] ~ [ linux-5.16.20 ] ~ [ linux-5.15.169 ] ~ [ linux-5.14.21 ] ~ [ linux-5.13.19 ] ~ [ linux-5.12.19 ] ~ [ linux-5.11.22 ] ~ [ linux-5.10.228 ] ~ [ linux-5.9.16 ] ~ [ linux-5.8.18 ] ~ [ linux-5.7.19 ] ~ [ linux-5.6.19 ] ~ [ linux-5.5.19 ] ~ [ linux-5.4.284 ] ~ [ linux-5.3.18 ] ~ [ linux-5.2.21 ] ~ [ linux-5.1.21 ] ~ [ linux-5.0.21 ] ~ [ linux-4.20.17 ] ~ [ linux-4.19.322 ] ~ [ linux-4.18.20 ] ~ [ linux-4.17.19 ] ~ [ linux-4.16.18 ] ~ [ linux-4.15.18 ] ~ [ linux-4.14.336 ] ~ [ linux-4.13.16 ] ~ [ linux-4.12.14 ] ~ [ linux-4.11.12 ] ~ [ linux-4.10.17 ] ~ [ linux-4.9.337 ] ~ [ linux-4.4.302 ] ~ [ linux-3.10.108 ] ~ [ linux-2.6.32.71 ] ~ [ linux-2.6.0 ] ~ [ linux-2.4.37.11 ] ~ [ unix-v6-master ] ~ [ ccs-tools-1.8.9 ] ~ [ policy-sample ] ~
Architecture: ~ [ i386 ] ~ [ alpha ] ~ [ m68k ] ~ [ mips ] ~ [ ppc ] ~ [ sparc ] ~ [ sparc64 ] ~

  1 // SPDX-License-Identifier: GPL-2.0
  2 /*
  3  * Copyright (C) 2021 Western Digital Corporation or its affiliates.
  4  * Copyright (C) 2022 Ventana Micro Systems Inc.
  5  *
  6  * Authors:
  7  *      Anup Patel <apatel@ventanamicro.com>
  8  */
  9 
 10 #include <linux/bits.h>
 11 #include <linux/irqchip/riscv-imsic.h>
 12 #include <linux/kvm_host.h>
 13 #include <linux/uaccess.h>
 14 
 15 static void unlock_vcpus(struct kvm *kvm, int vcpu_lock_idx)
 16 {
 17         struct kvm_vcpu *tmp_vcpu;
 18 
 19         for (; vcpu_lock_idx >= 0; vcpu_lock_idx--) {
 20                 tmp_vcpu = kvm_get_vcpu(kvm, vcpu_lock_idx);
 21                 mutex_unlock(&tmp_vcpu->mutex);
 22         }
 23 }
 24 
 25 static void unlock_all_vcpus(struct kvm *kvm)
 26 {
 27         unlock_vcpus(kvm, atomic_read(&kvm->online_vcpus) - 1);
 28 }
 29 
 30 static bool lock_all_vcpus(struct kvm *kvm)
 31 {
 32         struct kvm_vcpu *tmp_vcpu;
 33         unsigned long c;
 34 
 35         kvm_for_each_vcpu(c, tmp_vcpu, kvm) {
 36                 if (!mutex_trylock(&tmp_vcpu->mutex)) {
 37                         unlock_vcpus(kvm, c - 1);
 38                         return false;
 39                 }
 40         }
 41 
 42         return true;
 43 }
 44 
 45 static int aia_create(struct kvm_device *dev, u32 type)
 46 {
 47         int ret;
 48         unsigned long i;
 49         struct kvm *kvm = dev->kvm;
 50         struct kvm_vcpu *vcpu;
 51 
 52         if (irqchip_in_kernel(kvm))
 53                 return -EEXIST;
 54 
 55         ret = -EBUSY;
 56         if (!lock_all_vcpus(kvm))
 57                 return ret;
 58 
 59         kvm_for_each_vcpu(i, vcpu, kvm) {
 60                 if (vcpu->arch.ran_atleast_once)
 61                         goto out_unlock;
 62         }
 63         ret = 0;
 64 
 65         kvm->arch.aia.in_kernel = true;
 66 
 67 out_unlock:
 68         unlock_all_vcpus(kvm);
 69         return ret;
 70 }
 71 
 72 static void aia_destroy(struct kvm_device *dev)
 73 {
 74         kfree(dev);
 75 }
 76 
 77 static int aia_config(struct kvm *kvm, unsigned long type,
 78                       u32 *nr, bool write)
 79 {
 80         struct kvm_aia *aia = &kvm->arch.aia;
 81 
 82         /* Writes can only be done before irqchip is initialized */
 83         if (write && kvm_riscv_aia_initialized(kvm))
 84                 return -EBUSY;
 85 
 86         switch (type) {
 87         case KVM_DEV_RISCV_AIA_CONFIG_MODE:
 88                 if (write) {
 89                         switch (*nr) {
 90                         case KVM_DEV_RISCV_AIA_MODE_EMUL:
 91                                 break;
 92                         case KVM_DEV_RISCV_AIA_MODE_HWACCEL:
 93                         case KVM_DEV_RISCV_AIA_MODE_AUTO:
 94                                 /*
 95                                  * HW Acceleration and Auto modes only
 96                                  * supported on host with non-zero guest
 97                                  * external interrupts (i.e. non-zero
 98                                  * VS-level IMSIC pages).
 99                                  */
100                                 if (!kvm_riscv_aia_nr_hgei)
101                                         return -EINVAL;
102                                 break;
103                         default:
104                                 return -EINVAL;
105                         }
106                         aia->mode = *nr;
107                 } else
108                         *nr = aia->mode;
109                 break;
110         case KVM_DEV_RISCV_AIA_CONFIG_IDS:
111                 if (write) {
112                         if ((*nr < KVM_DEV_RISCV_AIA_IDS_MIN) ||
113                             (*nr >= KVM_DEV_RISCV_AIA_IDS_MAX) ||
114                             ((*nr & KVM_DEV_RISCV_AIA_IDS_MIN) !=
115                              KVM_DEV_RISCV_AIA_IDS_MIN) ||
116                             (kvm_riscv_aia_max_ids <= *nr))
117                                 return -EINVAL;
118                         aia->nr_ids = *nr;
119                 } else
120                         *nr = aia->nr_ids;
121                 break;
122         case KVM_DEV_RISCV_AIA_CONFIG_SRCS:
123                 if (write) {
124                         if ((*nr >= KVM_DEV_RISCV_AIA_SRCS_MAX) ||
125                             (*nr >= kvm_riscv_aia_max_ids))
126                                 return -EINVAL;
127                         aia->nr_sources = *nr;
128                 } else
129                         *nr = aia->nr_sources;
130                 break;
131         case KVM_DEV_RISCV_AIA_CONFIG_GROUP_BITS:
132                 if (write) {
133                         if (*nr >= KVM_DEV_RISCV_AIA_GROUP_BITS_MAX)
134                                 return -EINVAL;
135                         aia->nr_group_bits = *nr;
136                 } else
137                         *nr = aia->nr_group_bits;
138                 break;
139         case KVM_DEV_RISCV_AIA_CONFIG_GROUP_SHIFT:
140                 if (write) {
141                         if ((*nr < KVM_DEV_RISCV_AIA_GROUP_SHIFT_MIN) ||
142                             (*nr >= KVM_DEV_RISCV_AIA_GROUP_SHIFT_MAX))
143                                 return -EINVAL;
144                         aia->nr_group_shift = *nr;
145                 } else
146                         *nr = aia->nr_group_shift;
147                 break;
148         case KVM_DEV_RISCV_AIA_CONFIG_HART_BITS:
149                 if (write) {
150                         if (*nr >= KVM_DEV_RISCV_AIA_HART_BITS_MAX)
151                                 return -EINVAL;
152                         aia->nr_hart_bits = *nr;
153                 } else
154                         *nr = aia->nr_hart_bits;
155                 break;
156         case KVM_DEV_RISCV_AIA_CONFIG_GUEST_BITS:
157                 if (write) {
158                         if (*nr >= KVM_DEV_RISCV_AIA_GUEST_BITS_MAX)
159                                 return -EINVAL;
160                         aia->nr_guest_bits = *nr;
161                 } else
162                         *nr = aia->nr_guest_bits;
163                 break;
164         default:
165                 return -ENXIO;
166         }
167 
168         return 0;
169 }
170 
171 static int aia_aplic_addr(struct kvm *kvm, u64 *addr, bool write)
172 {
173         struct kvm_aia *aia = &kvm->arch.aia;
174 
175         if (write) {
176                 /* Writes can only be done before irqchip is initialized */
177                 if (kvm_riscv_aia_initialized(kvm))
178                         return -EBUSY;
179 
180                 if (*addr & (KVM_DEV_RISCV_APLIC_ALIGN - 1))
181                         return -EINVAL;
182 
183                 aia->aplic_addr = *addr;
184         } else
185                 *addr = aia->aplic_addr;
186 
187         return 0;
188 }
189 
190 static int aia_imsic_addr(struct kvm *kvm, u64 *addr,
191                           unsigned long vcpu_idx, bool write)
192 {
193         struct kvm_vcpu *vcpu;
194         struct kvm_vcpu_aia *vcpu_aia;
195 
196         vcpu = kvm_get_vcpu(kvm, vcpu_idx);
197         if (!vcpu)
198                 return -EINVAL;
199         vcpu_aia = &vcpu->arch.aia_context;
200 
201         if (write) {
202                 /* Writes can only be done before irqchip is initialized */
203                 if (kvm_riscv_aia_initialized(kvm))
204                         return -EBUSY;
205 
206                 if (*addr & (KVM_DEV_RISCV_IMSIC_ALIGN - 1))
207                         return -EINVAL;
208         }
209 
210         mutex_lock(&vcpu->mutex);
211         if (write)
212                 vcpu_aia->imsic_addr = *addr;
213         else
214                 *addr = vcpu_aia->imsic_addr;
215         mutex_unlock(&vcpu->mutex);
216 
217         return 0;
218 }
219 
220 static gpa_t aia_imsic_ppn(struct kvm_aia *aia, gpa_t addr)
221 {
222         u32 h, l;
223         gpa_t mask = 0;
224 
225         h = aia->nr_hart_bits + aia->nr_guest_bits +
226             IMSIC_MMIO_PAGE_SHIFT - 1;
227         mask = GENMASK_ULL(h, 0);
228 
229         if (aia->nr_group_bits) {
230                 h = aia->nr_group_bits + aia->nr_group_shift - 1;
231                 l = aia->nr_group_shift;
232                 mask |= GENMASK_ULL(h, l);
233         }
234 
235         return (addr & ~mask) >> IMSIC_MMIO_PAGE_SHIFT;
236 }
237 
238 static u32 aia_imsic_hart_index(struct kvm_aia *aia, gpa_t addr)
239 {
240         u32 hart = 0, group = 0;
241 
242         if (aia->nr_hart_bits)
243                 hart = (addr >> (aia->nr_guest_bits + IMSIC_MMIO_PAGE_SHIFT)) &
244                        GENMASK_ULL(aia->nr_hart_bits - 1, 0);
245         if (aia->nr_group_bits)
246                 group = (addr >> aia->nr_group_shift) &
247                         GENMASK_ULL(aia->nr_group_bits - 1, 0);
248 
249         return (group << aia->nr_hart_bits) | hart;
250 }
251 
252 static int aia_init(struct kvm *kvm)
253 {
254         int ret, i;
255         unsigned long idx;
256         struct kvm_vcpu *vcpu;
257         struct kvm_vcpu_aia *vaia;
258         struct kvm_aia *aia = &kvm->arch.aia;
259         gpa_t base_ppn = KVM_RISCV_AIA_UNDEF_ADDR;
260 
261         /* Irqchip can be initialized only once */
262         if (kvm_riscv_aia_initialized(kvm))
263                 return -EBUSY;
264 
265         /* We might be in the middle of creating a VCPU? */
266         if (kvm->created_vcpus != atomic_read(&kvm->online_vcpus))
267                 return -EBUSY;
268 
269         /* Number of sources should be less than or equals number of IDs */
270         if (aia->nr_ids < aia->nr_sources)
271                 return -EINVAL;
272 
273         /* APLIC base is required for non-zero number of sources */
274         if (aia->nr_sources && aia->aplic_addr == KVM_RISCV_AIA_UNDEF_ADDR)
275                 return -EINVAL;
276 
277         /* Initialize APLIC */
278         ret = kvm_riscv_aia_aplic_init(kvm);
279         if (ret)
280                 return ret;
281 
282         /* Iterate over each VCPU */
283         kvm_for_each_vcpu(idx, vcpu, kvm) {
284                 vaia = &vcpu->arch.aia_context;
285 
286                 /* IMSIC base is required */
287                 if (vaia->imsic_addr == KVM_RISCV_AIA_UNDEF_ADDR) {
288                         ret = -EINVAL;
289                         goto fail_cleanup_imsics;
290                 }
291 
292                 /* All IMSICs should have matching base PPN */
293                 if (base_ppn == KVM_RISCV_AIA_UNDEF_ADDR)
294                         base_ppn = aia_imsic_ppn(aia, vaia->imsic_addr);
295                 if (base_ppn != aia_imsic_ppn(aia, vaia->imsic_addr)) {
296                         ret = -EINVAL;
297                         goto fail_cleanup_imsics;
298                 }
299 
300                 /* Update HART index of the IMSIC based on IMSIC base */
301                 vaia->hart_index = aia_imsic_hart_index(aia,
302                                                         vaia->imsic_addr);
303 
304                 /* Initialize IMSIC for this VCPU */
305                 ret = kvm_riscv_vcpu_aia_imsic_init(vcpu);
306                 if (ret)
307                         goto fail_cleanup_imsics;
308         }
309 
310         /* Set the initialized flag */
311         kvm->arch.aia.initialized = true;
312 
313         return 0;
314 
315 fail_cleanup_imsics:
316         for (i = idx - 1; i >= 0; i--) {
317                 vcpu = kvm_get_vcpu(kvm, i);
318                 if (!vcpu)
319                         continue;
320                 kvm_riscv_vcpu_aia_imsic_cleanup(vcpu);
321         }
322         kvm_riscv_aia_aplic_cleanup(kvm);
323         return ret;
324 }
325 
326 static int aia_set_attr(struct kvm_device *dev, struct kvm_device_attr *attr)
327 {
328         u32 nr;
329         u64 addr;
330         int nr_vcpus, r = -ENXIO;
331         unsigned long v, type = (unsigned long)attr->attr;
332         void __user *uaddr = (void __user *)(long)attr->addr;
333 
334         switch (attr->group) {
335         case KVM_DEV_RISCV_AIA_GRP_CONFIG:
336                 if (copy_from_user(&nr, uaddr, sizeof(nr)))
337                         return -EFAULT;
338 
339                 mutex_lock(&dev->kvm->lock);
340                 r = aia_config(dev->kvm, type, &nr, true);
341                 mutex_unlock(&dev->kvm->lock);
342 
343                 break;
344 
345         case KVM_DEV_RISCV_AIA_GRP_ADDR:
346                 if (copy_from_user(&addr, uaddr, sizeof(addr)))
347                         return -EFAULT;
348 
349                 nr_vcpus = atomic_read(&dev->kvm->online_vcpus);
350                 mutex_lock(&dev->kvm->lock);
351                 if (type == KVM_DEV_RISCV_AIA_ADDR_APLIC)
352                         r = aia_aplic_addr(dev->kvm, &addr, true);
353                 else if (type < KVM_DEV_RISCV_AIA_ADDR_IMSIC(nr_vcpus))
354                         r = aia_imsic_addr(dev->kvm, &addr,
355                             type - KVM_DEV_RISCV_AIA_ADDR_IMSIC(0), true);
356                 mutex_unlock(&dev->kvm->lock);
357 
358                 break;
359 
360         case KVM_DEV_RISCV_AIA_GRP_CTRL:
361                 switch (type) {
362                 case KVM_DEV_RISCV_AIA_CTRL_INIT:
363                         mutex_lock(&dev->kvm->lock);
364                         r = aia_init(dev->kvm);
365                         mutex_unlock(&dev->kvm->lock);
366                         break;
367                 }
368 
369                 break;
370         case KVM_DEV_RISCV_AIA_GRP_APLIC:
371                 if (copy_from_user(&nr, uaddr, sizeof(nr)))
372                         return -EFAULT;
373 
374                 mutex_lock(&dev->kvm->lock);
375                 r = kvm_riscv_aia_aplic_set_attr(dev->kvm, type, nr);
376                 mutex_unlock(&dev->kvm->lock);
377 
378                 break;
379         case KVM_DEV_RISCV_AIA_GRP_IMSIC:
380                 if (copy_from_user(&v, uaddr, sizeof(v)))
381                         return -EFAULT;
382 
383                 mutex_lock(&dev->kvm->lock);
384                 r = kvm_riscv_aia_imsic_rw_attr(dev->kvm, type, true, &v);
385                 mutex_unlock(&dev->kvm->lock);
386 
387                 break;
388         }
389 
390         return r;
391 }
392 
393 static int aia_get_attr(struct kvm_device *dev, struct kvm_device_attr *attr)
394 {
395         u32 nr;
396         u64 addr;
397         int nr_vcpus, r = -ENXIO;
398         void __user *uaddr = (void __user *)(long)attr->addr;
399         unsigned long v, type = (unsigned long)attr->attr;
400 
401         switch (attr->group) {
402         case KVM_DEV_RISCV_AIA_GRP_CONFIG:
403                 if (copy_from_user(&nr, uaddr, sizeof(nr)))
404                         return -EFAULT;
405 
406                 mutex_lock(&dev->kvm->lock);
407                 r = aia_config(dev->kvm, type, &nr, false);
408                 mutex_unlock(&dev->kvm->lock);
409                 if (r)
410                         return r;
411 
412                 if (copy_to_user(uaddr, &nr, sizeof(nr)))
413                         return -EFAULT;
414 
415                 break;
416         case KVM_DEV_RISCV_AIA_GRP_ADDR:
417                 if (copy_from_user(&addr, uaddr, sizeof(addr)))
418                         return -EFAULT;
419 
420                 nr_vcpus = atomic_read(&dev->kvm->online_vcpus);
421                 mutex_lock(&dev->kvm->lock);
422                 if (type == KVM_DEV_RISCV_AIA_ADDR_APLIC)
423                         r = aia_aplic_addr(dev->kvm, &addr, false);
424                 else if (type < KVM_DEV_RISCV_AIA_ADDR_IMSIC(nr_vcpus))
425                         r = aia_imsic_addr(dev->kvm, &addr,
426                             type - KVM_DEV_RISCV_AIA_ADDR_IMSIC(0), false);
427                 mutex_unlock(&dev->kvm->lock);
428                 if (r)
429                         return r;
430 
431                 if (copy_to_user(uaddr, &addr, sizeof(addr)))
432                         return -EFAULT;
433 
434                 break;
435         case KVM_DEV_RISCV_AIA_GRP_APLIC:
436                 if (copy_from_user(&nr, uaddr, sizeof(nr)))
437                         return -EFAULT;
438 
439                 mutex_lock(&dev->kvm->lock);
440                 r = kvm_riscv_aia_aplic_get_attr(dev->kvm, type, &nr);
441                 mutex_unlock(&dev->kvm->lock);
442                 if (r)
443                         return r;
444 
445                 if (copy_to_user(uaddr, &nr, sizeof(nr)))
446                         return -EFAULT;
447 
448                 break;
449         case KVM_DEV_RISCV_AIA_GRP_IMSIC:
450                 if (copy_from_user(&v, uaddr, sizeof(v)))
451                         return -EFAULT;
452 
453                 mutex_lock(&dev->kvm->lock);
454                 r = kvm_riscv_aia_imsic_rw_attr(dev->kvm, type, false, &v);
455                 mutex_unlock(&dev->kvm->lock);
456                 if (r)
457                         return r;
458 
459                 if (copy_to_user(uaddr, &v, sizeof(v)))
460                         return -EFAULT;
461 
462                 break;
463         }
464 
465         return r;
466 }
467 
468 static int aia_has_attr(struct kvm_device *dev, struct kvm_device_attr *attr)
469 {
470         int nr_vcpus;
471 
472         switch (attr->group) {
473         case KVM_DEV_RISCV_AIA_GRP_CONFIG:
474                 switch (attr->attr) {
475                 case KVM_DEV_RISCV_AIA_CONFIG_MODE:
476                 case KVM_DEV_RISCV_AIA_CONFIG_IDS:
477                 case KVM_DEV_RISCV_AIA_CONFIG_SRCS:
478                 case KVM_DEV_RISCV_AIA_CONFIG_GROUP_BITS:
479                 case KVM_DEV_RISCV_AIA_CONFIG_GROUP_SHIFT:
480                 case KVM_DEV_RISCV_AIA_CONFIG_HART_BITS:
481                 case KVM_DEV_RISCV_AIA_CONFIG_GUEST_BITS:
482                         return 0;
483                 }
484                 break;
485         case KVM_DEV_RISCV_AIA_GRP_ADDR:
486                 nr_vcpus = atomic_read(&dev->kvm->online_vcpus);
487                 if (attr->attr == KVM_DEV_RISCV_AIA_ADDR_APLIC)
488                         return 0;
489                 else if (attr->attr < KVM_DEV_RISCV_AIA_ADDR_IMSIC(nr_vcpus))
490                         return 0;
491                 break;
492         case KVM_DEV_RISCV_AIA_GRP_CTRL:
493                 switch (attr->attr) {
494                 case KVM_DEV_RISCV_AIA_CTRL_INIT:
495                         return 0;
496                 }
497                 break;
498         case KVM_DEV_RISCV_AIA_GRP_APLIC:
499                 return kvm_riscv_aia_aplic_has_attr(dev->kvm, attr->attr);
500         case KVM_DEV_RISCV_AIA_GRP_IMSIC:
501                 return kvm_riscv_aia_imsic_has_attr(dev->kvm, attr->attr);
502         }
503 
504         return -ENXIO;
505 }
506 
507 struct kvm_device_ops kvm_riscv_aia_device_ops = {
508         .name = "kvm-riscv-aia",
509         .create = aia_create,
510         .destroy = aia_destroy,
511         .set_attr = aia_set_attr,
512         .get_attr = aia_get_attr,
513         .has_attr = aia_has_attr,
514 };
515 
516 int kvm_riscv_vcpu_aia_update(struct kvm_vcpu *vcpu)
517 {
518         /* Proceed only if AIA was initialized successfully */
519         if (!kvm_riscv_aia_initialized(vcpu->kvm))
520                 return 1;
521 
522         /* Update the IMSIC HW state before entering guest mode */
523         return kvm_riscv_vcpu_aia_imsic_update(vcpu);
524 }
525 
526 void kvm_riscv_vcpu_aia_reset(struct kvm_vcpu *vcpu)
527 {
528         struct kvm_vcpu_aia_csr *csr = &vcpu->arch.aia_context.guest_csr;
529         struct kvm_vcpu_aia_csr *reset_csr =
530                                 &vcpu->arch.aia_context.guest_reset_csr;
531 
532         if (!kvm_riscv_aia_available())
533                 return;
534         memcpy(csr, reset_csr, sizeof(*csr));
535 
536         /* Proceed only if AIA was initialized successfully */
537         if (!kvm_riscv_aia_initialized(vcpu->kvm))
538                 return;
539 
540         /* Reset the IMSIC context */
541         kvm_riscv_vcpu_aia_imsic_reset(vcpu);
542 }
543 
544 int kvm_riscv_vcpu_aia_init(struct kvm_vcpu *vcpu)
545 {
546         struct kvm_vcpu_aia *vaia = &vcpu->arch.aia_context;
547 
548         if (!kvm_riscv_aia_available())
549                 return 0;
550 
551         /*
552          * We don't do any memory allocations over here because these
553          * will be done after AIA device is initialized by the user-space.
554          *
555          * Refer, aia_init() implementation for more details.
556          */
557 
558         /* Initialize default values in AIA vcpu context */
559         vaia->imsic_addr = KVM_RISCV_AIA_UNDEF_ADDR;
560         vaia->hart_index = vcpu->vcpu_idx;
561 
562         return 0;
563 }
564 
565 void kvm_riscv_vcpu_aia_deinit(struct kvm_vcpu *vcpu)
566 {
567         /* Proceed only if AIA was initialized successfully */
568         if (!kvm_riscv_aia_initialized(vcpu->kvm))
569                 return;
570 
571         /* Cleanup IMSIC context */
572         kvm_riscv_vcpu_aia_imsic_cleanup(vcpu);
573 }
574 
575 int kvm_riscv_aia_inject_msi_by_id(struct kvm *kvm, u32 hart_index,
576                                    u32 guest_index, u32 iid)
577 {
578         unsigned long idx;
579         struct kvm_vcpu *vcpu;
580 
581         /* Proceed only if AIA was initialized successfully */
582         if (!kvm_riscv_aia_initialized(kvm))
583                 return -EBUSY;
584 
585         /* Inject MSI to matching VCPU */
586         kvm_for_each_vcpu(idx, vcpu, kvm) {
587                 if (vcpu->arch.aia_context.hart_index == hart_index)
588                         return kvm_riscv_vcpu_aia_imsic_inject(vcpu,
589                                                                guest_index,
590                                                                0, iid);
591         }
592 
593         return 0;
594 }
595 
596 int kvm_riscv_aia_inject_msi(struct kvm *kvm, struct kvm_msi *msi)
597 {
598         gpa_t tppn, ippn;
599         unsigned long idx;
600         struct kvm_vcpu *vcpu;
601         u32 g, toff, iid = msi->data;
602         struct kvm_aia *aia = &kvm->arch.aia;
603         gpa_t target = (((gpa_t)msi->address_hi) << 32) | msi->address_lo;
604 
605         /* Proceed only if AIA was initialized successfully */
606         if (!kvm_riscv_aia_initialized(kvm))
607                 return -EBUSY;
608 
609         /* Convert target address to target PPN */
610         tppn = target >> IMSIC_MMIO_PAGE_SHIFT;
611 
612         /* Extract and clear Guest ID from target PPN */
613         g = tppn & (BIT(aia->nr_guest_bits) - 1);
614         tppn &= ~((gpa_t)(BIT(aia->nr_guest_bits) - 1));
615 
616         /* Inject MSI to matching VCPU */
617         kvm_for_each_vcpu(idx, vcpu, kvm) {
618                 ippn = vcpu->arch.aia_context.imsic_addr >>
619                                         IMSIC_MMIO_PAGE_SHIFT;
620                 if (ippn == tppn) {
621                         toff = target & (IMSIC_MMIO_PAGE_SZ - 1);
622                         return kvm_riscv_vcpu_aia_imsic_inject(vcpu, g,
623                                                                toff, iid);
624                 }
625         }
626 
627         return 0;
628 }
629 
630 int kvm_riscv_aia_inject_irq(struct kvm *kvm, unsigned int irq, bool level)
631 {
632         /* Proceed only if AIA was initialized successfully */
633         if (!kvm_riscv_aia_initialized(kvm))
634                 return -EBUSY;
635 
636         /* Inject interrupt level change in APLIC */
637         return kvm_riscv_aia_aplic_inject(kvm, irq, level);
638 }
639 
640 void kvm_riscv_aia_init_vm(struct kvm *kvm)
641 {
642         struct kvm_aia *aia = &kvm->arch.aia;
643 
644         if (!kvm_riscv_aia_available())
645                 return;
646 
647         /*
648          * We don't do any memory allocations over here because these
649          * will be done after AIA device is initialized by the user-space.
650          *
651          * Refer, aia_init() implementation for more details.
652          */
653 
654         /* Initialize default values in AIA global context */
655         aia->mode = (kvm_riscv_aia_nr_hgei) ?
656                 KVM_DEV_RISCV_AIA_MODE_AUTO : KVM_DEV_RISCV_AIA_MODE_EMUL;
657         aia->nr_ids = kvm_riscv_aia_max_ids - 1;
658         aia->nr_sources = 0;
659         aia->nr_group_bits = 0;
660         aia->nr_group_shift = KVM_DEV_RISCV_AIA_GROUP_SHIFT_MIN;
661         aia->nr_hart_bits = 0;
662         aia->nr_guest_bits = 0;
663         aia->aplic_addr = KVM_RISCV_AIA_UNDEF_ADDR;
664 }
665 
666 void kvm_riscv_aia_destroy_vm(struct kvm *kvm)
667 {
668         /* Proceed only if AIA was initialized successfully */
669         if (!kvm_riscv_aia_initialized(kvm))
670                 return;
671 
672         /* Cleanup APLIC context */
673         kvm_riscv_aia_aplic_cleanup(kvm);
674 }
675 

~ [ source navigation ] ~ [ diff markup ] ~ [ identifier search ] ~

kernel.org | git.kernel.org | LWN.net | Project Home | SVN repository | Mail admin

Linux® is a registered trademark of Linus Torvalds in the United States and other countries.
TOMOYO® is a registered trademark of NTT DATA CORPORATION.

sflogo.php