aboutsummaryrefslogtreecommitdiffstats
path: root/virt/kvm
diff options
context:
space:
mode:
authorAndre Przywara <andre.przywara@arm.com>2014-06-06 18:53:08 -0400
committerChristoffer Dall <christoffer.dall@linaro.org>2015-01-20 12:25:30 -0500
commit1d916229e348c628ddc9cf97528e76d13f52c122 (patch)
treeca7881a221f0a5ffb40367fd3445d38e0b239bb1 /virt/kvm
parent832158125d2ef30b364f21e1616495c40c286a4a (diff)
arm/arm64: KVM: split GICv2 specific emulation code from vgic.c
vgic.c is currently a mixture of generic vGIC emulation code and functions specific to emulating a GICv2. To ease the addition of GICv3, split off strictly v2 specific parts into a new file vgic-v2-emul.c. Signed-off-by: Andre Przywara <andre.przywara@arm.com> Acked-by: Christoffer Dall <christoffer.dall@linaro.org> ------- As the diff isn't always obvious here (and to aid eventual rebases), here is a list of high-level changes done to the code: * added new file to respective arm/arm64 Makefiles * moved GICv2 specific functions to vgic-v2-emul.c: - handle_mmio_misc() - handle_mmio_set_enable_reg() - handle_mmio_clear_enable_reg() - handle_mmio_set_pending_reg() - handle_mmio_clear_pending_reg() - handle_mmio_priority_reg() - vgic_get_target_reg() - vgic_set_target_reg() - handle_mmio_target_reg() - handle_mmio_cfg_reg() - handle_mmio_sgi_reg() - vgic_v2_unqueue_sgi() - read_set_clear_sgi_pend_reg() - write_set_clear_sgi_pend_reg() - handle_mmio_sgi_set() - handle_mmio_sgi_clear() - vgic_v2_handle_mmio() - vgic_get_sgi_sources() - vgic_dispatch_sgi() - vgic_v2_queue_sgi() - vgic_v2_map_resources() - vgic_v2_init() - vgic_v2_add_sgi_source() - vgic_v2_init_model() - vgic_v2_init_emulation() - handle_cpu_mmio_misc() - handle_mmio_abpr() - handle_cpu_mmio_ident() - vgic_attr_regs_access() - vgic_create() (renamed to vgic_v2_create()) - vgic_destroy() (renamed to vgic_v2_destroy()) - vgic_has_attr() (renamed to vgic_v2_has_attr()) - vgic_set_attr() (renamed to vgic_v2_set_attr()) - vgic_get_attr() (renamed to vgic_v2_get_attr()) - struct kvm_mmio_range vgic_dist_ranges[] - struct kvm_mmio_range vgic_cpu_ranges[] - struct kvm_device_ops kvm_arm_vgic_v2_ops {} Signed-off-by: Christoffer Dall <christoffer.dall@linaro.org>
Diffstat (limited to 'virt/kvm')
-rw-r--r--virt/kvm/arm/vgic-v2-emul.c847
-rw-r--r--virt/kvm/arm/vgic.c806
2 files changed, 848 insertions, 805 deletions
diff --git a/virt/kvm/arm/vgic-v2-emul.c b/virt/kvm/arm/vgic-v2-emul.c
new file mode 100644
index 000000000000..19c6210f02cf
--- /dev/null
+++ b/virt/kvm/arm/vgic-v2-emul.c
@@ -0,0 +1,847 @@
1/*
2 * Contains GICv2 specific emulation code, was in vgic.c before.
3 *
4 * Copyright (C) 2012 ARM Ltd.
5 * Author: Marc Zyngier <marc.zyngier@arm.com>
6 *
7 * This program is free software; you can redistribute it and/or modify
8 * it under the terms of the GNU General Public License version 2 as
9 * published by the Free Software Foundation.
10 *
11 * This program is distributed in the hope that it will be useful,
12 * but WITHOUT ANY WARRANTY; without even the implied warranty of
13 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
14 * GNU General Public License for more details.
15 *
16 * You should have received a copy of the GNU General Public License
17 * along with this program. If not, see <http://www.gnu.org/licenses/>.
18 */
19
20#include <linux/cpu.h>
21#include <linux/kvm.h>
22#include <linux/kvm_host.h>
23#include <linux/interrupt.h>
24#include <linux/io.h>
25#include <linux/uaccess.h>
26
27#include <linux/irqchip/arm-gic.h>
28
29#include <asm/kvm_emulate.h>
30#include <asm/kvm_arm.h>
31#include <asm/kvm_mmu.h>
32
33#include "vgic.h"
34
35#define GICC_ARCH_VERSION_V2 0x2
36
37static void vgic_dispatch_sgi(struct kvm_vcpu *vcpu, u32 reg);
38static u8 *vgic_get_sgi_sources(struct vgic_dist *dist, int vcpu_id, int sgi)
39{
40 return dist->irq_sgi_sources + vcpu_id * VGIC_NR_SGIS + sgi;
41}
42
43static bool handle_mmio_misc(struct kvm_vcpu *vcpu,
44 struct kvm_exit_mmio *mmio, phys_addr_t offset)
45{
46 u32 reg;
47 u32 word_offset = offset & 3;
48
49 switch (offset & ~3) {
50 case 0: /* GICD_CTLR */
51 reg = vcpu->kvm->arch.vgic.enabled;
52 vgic_reg_access(mmio, &reg, word_offset,
53 ACCESS_READ_VALUE | ACCESS_WRITE_VALUE);
54 if (mmio->is_write) {
55 vcpu->kvm->arch.vgic.enabled = reg & 1;
56 vgic_update_state(vcpu->kvm);
57 return true;
58 }
59 break;
60
61 case 4: /* GICD_TYPER */
62 reg = (atomic_read(&vcpu->kvm->online_vcpus) - 1) << 5;
63 reg |= (vcpu->kvm->arch.vgic.nr_irqs >> 5) - 1;
64 vgic_reg_access(mmio, &reg, word_offset,
65 ACCESS_READ_VALUE | ACCESS_WRITE_IGNORED);
66 break;
67
68 case 8: /* GICD_IIDR */
69 reg = (PRODUCT_ID_KVM << 24) | (IMPLEMENTER_ARM << 0);
70 vgic_reg_access(mmio, &reg, word_offset,
71 ACCESS_READ_VALUE | ACCESS_WRITE_IGNORED);
72 break;
73 }
74
75 return false;
76}
77
78static bool handle_mmio_set_enable_reg(struct kvm_vcpu *vcpu,
79 struct kvm_exit_mmio *mmio,
80 phys_addr_t offset)
81{
82 return vgic_handle_enable_reg(vcpu->kvm, mmio, offset,
83 vcpu->vcpu_id, ACCESS_WRITE_SETBIT);
84}
85
86static bool handle_mmio_clear_enable_reg(struct kvm_vcpu *vcpu,
87 struct kvm_exit_mmio *mmio,
88 phys_addr_t offset)
89{
90 return vgic_handle_enable_reg(vcpu->kvm, mmio, offset,
91 vcpu->vcpu_id, ACCESS_WRITE_CLEARBIT);
92}
93
94static bool handle_mmio_set_pending_reg(struct kvm_vcpu *vcpu,
95 struct kvm_exit_mmio *mmio,
96 phys_addr_t offset)
97{
98 return vgic_handle_set_pending_reg(vcpu->kvm, mmio, offset,
99 vcpu->vcpu_id);
100}
101
102static bool handle_mmio_clear_pending_reg(struct kvm_vcpu *vcpu,
103 struct kvm_exit_mmio *mmio,
104 phys_addr_t offset)
105{
106 return vgic_handle_clear_pending_reg(vcpu->kvm, mmio, offset,
107 vcpu->vcpu_id);
108}
109
110static bool handle_mmio_priority_reg(struct kvm_vcpu *vcpu,
111 struct kvm_exit_mmio *mmio,
112 phys_addr_t offset)
113{
114 u32 *reg = vgic_bytemap_get_reg(&vcpu->kvm->arch.vgic.irq_priority,
115 vcpu->vcpu_id, offset);
116 vgic_reg_access(mmio, reg, offset,
117 ACCESS_READ_VALUE | ACCESS_WRITE_VALUE);
118 return false;
119}
120
121#define GICD_ITARGETSR_SIZE 32
122#define GICD_CPUTARGETS_BITS 8
123#define GICD_IRQS_PER_ITARGETSR (GICD_ITARGETSR_SIZE / GICD_CPUTARGETS_BITS)
124static u32 vgic_get_target_reg(struct kvm *kvm, int irq)
125{
126 struct vgic_dist *dist = &kvm->arch.vgic;
127 int i;
128 u32 val = 0;
129
130 irq -= VGIC_NR_PRIVATE_IRQS;
131
132 for (i = 0; i < GICD_IRQS_PER_ITARGETSR; i++)
133 val |= 1 << (dist->irq_spi_cpu[irq + i] + i * 8);
134
135 return val;
136}
137
138static void vgic_set_target_reg(struct kvm *kvm, u32 val, int irq)
139{
140 struct vgic_dist *dist = &kvm->arch.vgic;
141 struct kvm_vcpu *vcpu;
142 int i, c;
143 unsigned long *bmap;
144 u32 target;
145
146 irq -= VGIC_NR_PRIVATE_IRQS;
147
148 /*
149 * Pick the LSB in each byte. This ensures we target exactly
150 * one vcpu per IRQ. If the byte is null, assume we target
151 * CPU0.
152 */
153 for (i = 0; i < GICD_IRQS_PER_ITARGETSR; i++) {
154 int shift = i * GICD_CPUTARGETS_BITS;
155
156 target = ffs((val >> shift) & 0xffU);
157 target = target ? (target - 1) : 0;
158 dist->irq_spi_cpu[irq + i] = target;
159 kvm_for_each_vcpu(c, vcpu, kvm) {
160 bmap = vgic_bitmap_get_shared_map(&dist->irq_spi_target[c]);
161 if (c == target)
162 set_bit(irq + i, bmap);
163 else
164 clear_bit(irq + i, bmap);
165 }
166 }
167}
168
169static bool handle_mmio_target_reg(struct kvm_vcpu *vcpu,
170 struct kvm_exit_mmio *mmio,
171 phys_addr_t offset)
172{
173 u32 reg;
174
175 /* We treat the banked interrupts targets as read-only */
176 if (offset < 32) {
177 u32 roreg;
178
179 roreg = 1 << vcpu->vcpu_id;
180 roreg |= roreg << 8;
181 roreg |= roreg << 16;
182
183 vgic_reg_access(mmio, &roreg, offset,
184 ACCESS_READ_VALUE | ACCESS_WRITE_IGNORED);
185 return false;
186 }
187
188 reg = vgic_get_target_reg(vcpu->kvm, offset & ~3U);
189 vgic_reg_access(mmio, &reg, offset,
190 ACCESS_READ_VALUE | ACCESS_WRITE_VALUE);
191 if (mmio->is_write) {
192 vgic_set_target_reg(vcpu->kvm, reg, offset & ~3U);
193 vgic_update_state(vcpu->kvm);
194 return true;
195 }
196
197 return false;
198}
199
200static bool handle_mmio_cfg_reg(struct kvm_vcpu *vcpu,
201 struct kvm_exit_mmio *mmio, phys_addr_t offset)
202{
203 u32 *reg;
204
205 reg = vgic_bitmap_get_reg(&vcpu->kvm->arch.vgic.irq_cfg,
206 vcpu->vcpu_id, offset >> 1);
207
208 return vgic_handle_cfg_reg(reg, mmio, offset);
209}
210
211static bool handle_mmio_sgi_reg(struct kvm_vcpu *vcpu,
212 struct kvm_exit_mmio *mmio, phys_addr_t offset)
213{
214 u32 reg;
215
216 vgic_reg_access(mmio, &reg, offset,
217 ACCESS_READ_RAZ | ACCESS_WRITE_VALUE);
218 if (mmio->is_write) {
219 vgic_dispatch_sgi(vcpu, reg);
220 vgic_update_state(vcpu->kvm);
221 return true;
222 }
223
224 return false;
225}
226
227/* Handle reads of GICD_CPENDSGIRn and GICD_SPENDSGIRn */
228static bool read_set_clear_sgi_pend_reg(struct kvm_vcpu *vcpu,
229 struct kvm_exit_mmio *mmio,
230 phys_addr_t offset)
231{
232 struct vgic_dist *dist = &vcpu->kvm->arch.vgic;
233 int sgi;
234 int min_sgi = (offset & ~0x3);
235 int max_sgi = min_sgi + 3;
236 int vcpu_id = vcpu->vcpu_id;
237 u32 reg = 0;
238
239 /* Copy source SGIs from distributor side */
240 for (sgi = min_sgi; sgi <= max_sgi; sgi++) {
241 u8 sources = *vgic_get_sgi_sources(dist, vcpu_id, sgi);
242
243 reg |= ((u32)sources) << (8 * (sgi - min_sgi));
244 }
245
246 mmio_data_write(mmio, ~0, reg);
247 return false;
248}
249
250static bool write_set_clear_sgi_pend_reg(struct kvm_vcpu *vcpu,
251 struct kvm_exit_mmio *mmio,
252 phys_addr_t offset, bool set)
253{
254 struct vgic_dist *dist = &vcpu->kvm->arch.vgic;
255 int sgi;
256 int min_sgi = (offset & ~0x3);
257 int max_sgi = min_sgi + 3;
258 int vcpu_id = vcpu->vcpu_id;
259 u32 reg;
260 bool updated = false;
261
262 reg = mmio_data_read(mmio, ~0);
263
264 /* Clear pending SGIs on the distributor */
265 for (sgi = min_sgi; sgi <= max_sgi; sgi++) {
266 u8 mask = reg >> (8 * (sgi - min_sgi));
267 u8 *src = vgic_get_sgi_sources(dist, vcpu_id, sgi);
268
269 if (set) {
270 if ((*src & mask) != mask)
271 updated = true;
272 *src |= mask;
273 } else {
274 if (*src & mask)
275 updated = true;
276 *src &= ~mask;
277 }
278 }
279
280 if (updated)
281 vgic_update_state(vcpu->kvm);
282
283 return updated;
284}
285
286static bool handle_mmio_sgi_set(struct kvm_vcpu *vcpu,
287 struct kvm_exit_mmio *mmio,
288 phys_addr_t offset)
289{
290 if (!mmio->is_write)
291 return read_set_clear_sgi_pend_reg(vcpu, mmio, offset);
292 else
293 return write_set_clear_sgi_pend_reg(vcpu, mmio, offset, true);
294}
295
296static bool handle_mmio_sgi_clear(struct kvm_vcpu *vcpu,
297 struct kvm_exit_mmio *mmio,
298 phys_addr_t offset)
299{
300 if (!mmio->is_write)
301 return read_set_clear_sgi_pend_reg(vcpu, mmio, offset);
302 else
303 return write_set_clear_sgi_pend_reg(vcpu, mmio, offset, false);
304}
305
306static const struct kvm_mmio_range vgic_dist_ranges[] = {
307 {
308 .base = GIC_DIST_CTRL,
309 .len = 12,
310 .bits_per_irq = 0,
311 .handle_mmio = handle_mmio_misc,
312 },
313 {
314 .base = GIC_DIST_IGROUP,
315 .len = VGIC_MAX_IRQS / 8,
316 .bits_per_irq = 1,
317 .handle_mmio = handle_mmio_raz_wi,
318 },
319 {
320 .base = GIC_DIST_ENABLE_SET,
321 .len = VGIC_MAX_IRQS / 8,
322 .bits_per_irq = 1,
323 .handle_mmio = handle_mmio_set_enable_reg,
324 },
325 {
326 .base = GIC_DIST_ENABLE_CLEAR,
327 .len = VGIC_MAX_IRQS / 8,
328 .bits_per_irq = 1,
329 .handle_mmio = handle_mmio_clear_enable_reg,
330 },
331 {
332 .base = GIC_DIST_PENDING_SET,
333 .len = VGIC_MAX_IRQS / 8,
334 .bits_per_irq = 1,
335 .handle_mmio = handle_mmio_set_pending_reg,
336 },
337 {
338 .base = GIC_DIST_PENDING_CLEAR,
339 .len = VGIC_MAX_IRQS / 8,
340 .bits_per_irq = 1,
341 .handle_mmio = handle_mmio_clear_pending_reg,
342 },
343 {
344 .base = GIC_DIST_ACTIVE_SET,
345 .len = VGIC_MAX_IRQS / 8,
346 .bits_per_irq = 1,
347 .handle_mmio = handle_mmio_raz_wi,
348 },
349 {
350 .base = GIC_DIST_ACTIVE_CLEAR,
351 .len = VGIC_MAX_IRQS / 8,
352 .bits_per_irq = 1,
353 .handle_mmio = handle_mmio_raz_wi,
354 },
355 {
356 .base = GIC_DIST_PRI,
357 .len = VGIC_MAX_IRQS,
358 .bits_per_irq = 8,
359 .handle_mmio = handle_mmio_priority_reg,
360 },
361 {
362 .base = GIC_DIST_TARGET,
363 .len = VGIC_MAX_IRQS,
364 .bits_per_irq = 8,
365 .handle_mmio = handle_mmio_target_reg,
366 },
367 {
368 .base = GIC_DIST_CONFIG,
369 .len = VGIC_MAX_IRQS / 4,
370 .bits_per_irq = 2,
371 .handle_mmio = handle_mmio_cfg_reg,
372 },
373 {
374 .base = GIC_DIST_SOFTINT,
375 .len = 4,
376 .handle_mmio = handle_mmio_sgi_reg,
377 },
378 {
379 .base = GIC_DIST_SGI_PENDING_CLEAR,
380 .len = VGIC_NR_SGIS,
381 .handle_mmio = handle_mmio_sgi_clear,
382 },
383 {
384 .base = GIC_DIST_SGI_PENDING_SET,
385 .len = VGIC_NR_SGIS,
386 .handle_mmio = handle_mmio_sgi_set,
387 },
388 {}
389};
390
391static bool vgic_v2_handle_mmio(struct kvm_vcpu *vcpu, struct kvm_run *run,
392 struct kvm_exit_mmio *mmio)
393{
394 unsigned long base = vcpu->kvm->arch.vgic.vgic_dist_base;
395
396 if (!is_in_range(mmio->phys_addr, mmio->len, base,
397 KVM_VGIC_V2_DIST_SIZE))
398 return false;
399
400 /* GICv2 does not support accesses wider than 32 bits */
401 if (mmio->len > 4) {
402 kvm_inject_dabt(vcpu, mmio->phys_addr);
403 return true;
404 }
405
406 return vgic_handle_mmio_range(vcpu, run, mmio, vgic_dist_ranges, base);
407}
408
409static void vgic_dispatch_sgi(struct kvm_vcpu *vcpu, u32 reg)
410{
411 struct kvm *kvm = vcpu->kvm;
412 struct vgic_dist *dist = &kvm->arch.vgic;
413 int nrcpus = atomic_read(&kvm->online_vcpus);
414 u8 target_cpus;
415 int sgi, mode, c, vcpu_id;
416
417 vcpu_id = vcpu->vcpu_id;
418
419 sgi = reg & 0xf;
420 target_cpus = (reg >> 16) & 0xff;
421 mode = (reg >> 24) & 3;
422
423 switch (mode) {
424 case 0:
425 if (!target_cpus)
426 return;
427 break;
428
429 case 1:
430 target_cpus = ((1 << nrcpus) - 1) & ~(1 << vcpu_id) & 0xff;
431 break;
432
433 case 2:
434 target_cpus = 1 << vcpu_id;
435 break;
436 }
437
438 kvm_for_each_vcpu(c, vcpu, kvm) {
439 if (target_cpus & 1) {
440 /* Flag the SGI as pending */
441 vgic_dist_irq_set_pending(vcpu, sgi);
442 *vgic_get_sgi_sources(dist, c, sgi) |= 1 << vcpu_id;
443 kvm_debug("SGI%d from CPU%d to CPU%d\n",
444 sgi, vcpu_id, c);
445 }
446
447 target_cpus >>= 1;
448 }
449}
450
451static bool vgic_v2_queue_sgi(struct kvm_vcpu *vcpu, int irq)
452{
453 struct vgic_dist *dist = &vcpu->kvm->arch.vgic;
454 unsigned long sources;
455 int vcpu_id = vcpu->vcpu_id;
456 int c;
457
458 sources = *vgic_get_sgi_sources(dist, vcpu_id, irq);
459
460 for_each_set_bit(c, &sources, dist->nr_cpus) {
461 if (vgic_queue_irq(vcpu, c, irq))
462 clear_bit(c, &sources);
463 }
464
465 *vgic_get_sgi_sources(dist, vcpu_id, irq) = sources;
466
467 /*
468 * If the sources bitmap has been cleared it means that we
469 * could queue all the SGIs onto link registers (see the
470 * clear_bit above), and therefore we are done with them in
471 * our emulated gic and can get rid of them.
472 */
473 if (!sources) {
474 vgic_dist_irq_clear_pending(vcpu, irq);
475 vgic_cpu_irq_clear(vcpu, irq);
476 return true;
477 }
478
479 return false;
480}
481
482/**
483 * kvm_vgic_map_resources - Configure global VGIC state before running any VCPUs
484 * @kvm: pointer to the kvm struct
485 *
486 * Map the virtual CPU interface into the VM before running any VCPUs. We
487 * can't do this at creation time, because user space must first set the
488 * virtual CPU interface address in the guest physical address space.
489 */
490static int vgic_v2_map_resources(struct kvm *kvm,
491 const struct vgic_params *params)
492{
493 int ret = 0;
494
495 if (!irqchip_in_kernel(kvm))
496 return 0;
497
498 mutex_lock(&kvm->lock);
499
500 if (vgic_ready(kvm))
501 goto out;
502
503 if (IS_VGIC_ADDR_UNDEF(kvm->arch.vgic.vgic_dist_base) ||
504 IS_VGIC_ADDR_UNDEF(kvm->arch.vgic.vgic_cpu_base)) {
505 kvm_err("Need to set vgic cpu and dist addresses first\n");
506 ret = -ENXIO;
507 goto out;
508 }
509
510 /*
511 * Initialize the vgic if this hasn't already been done on demand by
512 * accessing the vgic state from userspace.
513 */
514 ret = vgic_init(kvm);
515 if (ret) {
516 kvm_err("Unable to allocate maps\n");
517 goto out;
518 }
519
520 ret = kvm_phys_addr_ioremap(kvm, kvm->arch.vgic.vgic_cpu_base,
521 params->vcpu_base, KVM_VGIC_V2_CPU_SIZE,
522 true);
523 if (ret) {
524 kvm_err("Unable to remap VGIC CPU to VCPU\n");
525 goto out;
526 }
527
528 kvm->arch.vgic.ready = true;
529out:
530 if (ret)
531 kvm_vgic_destroy(kvm);
532 mutex_unlock(&kvm->lock);
533 return ret;
534}
535
536static void vgic_v2_add_sgi_source(struct kvm_vcpu *vcpu, int irq, int source)
537{
538 struct vgic_dist *dist = &vcpu->kvm->arch.vgic;
539
540 *vgic_get_sgi_sources(dist, vcpu->vcpu_id, irq) |= 1 << source;
541}
542
543static int vgic_v2_init_model(struct kvm *kvm)
544{
545 int i;
546
547 for (i = VGIC_NR_PRIVATE_IRQS; i < kvm->arch.vgic.nr_irqs; i += 4)
548 vgic_set_target_reg(kvm, 0, i);
549
550 return 0;
551}
552
553void vgic_v2_init_emulation(struct kvm *kvm)
554{
555 struct vgic_dist *dist = &kvm->arch.vgic;
556
557 dist->vm_ops.handle_mmio = vgic_v2_handle_mmio;
558 dist->vm_ops.queue_sgi = vgic_v2_queue_sgi;
559 dist->vm_ops.add_sgi_source = vgic_v2_add_sgi_source;
560 dist->vm_ops.init_model = vgic_v2_init_model;
561 dist->vm_ops.map_resources = vgic_v2_map_resources;
562
563 kvm->arch.max_vcpus = VGIC_V2_MAX_CPUS;
564}
565
566static bool handle_cpu_mmio_misc(struct kvm_vcpu *vcpu,
567 struct kvm_exit_mmio *mmio, phys_addr_t offset)
568{
569 bool updated = false;
570 struct vgic_vmcr vmcr;
571 u32 *vmcr_field;
572 u32 reg;
573
574 vgic_get_vmcr(vcpu, &vmcr);
575
576 switch (offset & ~0x3) {
577 case GIC_CPU_CTRL:
578 vmcr_field = &vmcr.ctlr;
579 break;
580 case GIC_CPU_PRIMASK:
581 vmcr_field = &vmcr.pmr;
582 break;
583 case GIC_CPU_BINPOINT:
584 vmcr_field = &vmcr.bpr;
585 break;
586 case GIC_CPU_ALIAS_BINPOINT:
587 vmcr_field = &vmcr.abpr;
588 break;
589 default:
590 BUG();
591 }
592
593 if (!mmio->is_write) {
594 reg = *vmcr_field;
595 mmio_data_write(mmio, ~0, reg);
596 } else {
597 reg = mmio_data_read(mmio, ~0);
598 if (reg != *vmcr_field) {
599 *vmcr_field = reg;
600 vgic_set_vmcr(vcpu, &vmcr);
601 updated = true;
602 }
603 }
604 return updated;
605}
606
607static bool handle_mmio_abpr(struct kvm_vcpu *vcpu,
608 struct kvm_exit_mmio *mmio, phys_addr_t offset)
609{
610 return handle_cpu_mmio_misc(vcpu, mmio, GIC_CPU_ALIAS_BINPOINT);
611}
612
613static bool handle_cpu_mmio_ident(struct kvm_vcpu *vcpu,
614 struct kvm_exit_mmio *mmio,
615 phys_addr_t offset)
616{
617 u32 reg;
618
619 if (mmio->is_write)
620 return false;
621
622 /* GICC_IIDR */
623 reg = (PRODUCT_ID_KVM << 20) |
624 (GICC_ARCH_VERSION_V2 << 16) |
625 (IMPLEMENTER_ARM << 0);
626 mmio_data_write(mmio, ~0, reg);
627 return false;
628}
629
630/*
631 * CPU Interface Register accesses - these are not accessed by the VM, but by
632 * user space for saving and restoring VGIC state.
633 */
634static const struct kvm_mmio_range vgic_cpu_ranges[] = {
635 {
636 .base = GIC_CPU_CTRL,
637 .len = 12,
638 .handle_mmio = handle_cpu_mmio_misc,
639 },
640 {
641 .base = GIC_CPU_ALIAS_BINPOINT,
642 .len = 4,
643 .handle_mmio = handle_mmio_abpr,
644 },
645 {
646 .base = GIC_CPU_ACTIVEPRIO,
647 .len = 16,
648 .handle_mmio = handle_mmio_raz_wi,
649 },
650 {
651 .base = GIC_CPU_IDENT,
652 .len = 4,
653 .handle_mmio = handle_cpu_mmio_ident,
654 },
655};
656
657static int vgic_attr_regs_access(struct kvm_device *dev,
658 struct kvm_device_attr *attr,
659 u32 *reg, bool is_write)
660{
661 const struct kvm_mmio_range *r = NULL, *ranges;
662 phys_addr_t offset;
663 int ret, cpuid, c;
664 struct kvm_vcpu *vcpu, *tmp_vcpu;
665 struct vgic_dist *vgic;
666 struct kvm_exit_mmio mmio;
667
668 offset = attr->attr & KVM_DEV_ARM_VGIC_OFFSET_MASK;
669 cpuid = (attr->attr & KVM_DEV_ARM_VGIC_CPUID_MASK) >>
670 KVM_DEV_ARM_VGIC_CPUID_SHIFT;
671
672 mutex_lock(&dev->kvm->lock);
673
674 ret = vgic_init(dev->kvm);
675 if (ret)
676 goto out;
677
678 if (cpuid >= atomic_read(&dev->kvm->online_vcpus)) {
679 ret = -EINVAL;
680 goto out;
681 }
682
683 vcpu = kvm_get_vcpu(dev->kvm, cpuid);
684 vgic = &dev->kvm->arch.vgic;
685
686 mmio.len = 4;
687 mmio.is_write = is_write;
688 if (is_write)
689 mmio_data_write(&mmio, ~0, *reg);
690 switch (attr->group) {
691 case KVM_DEV_ARM_VGIC_GRP_DIST_REGS:
692 mmio.phys_addr = vgic->vgic_dist_base + offset;
693 ranges = vgic_dist_ranges;
694 break;
695 case KVM_DEV_ARM_VGIC_GRP_CPU_REGS:
696 mmio.phys_addr = vgic->vgic_cpu_base + offset;
697 ranges = vgic_cpu_ranges;
698 break;
699 default:
700 BUG();
701 }
702 r = vgic_find_range(ranges, &mmio, offset);
703
704 if (unlikely(!r || !r->handle_mmio)) {
705 ret = -ENXIO;
706 goto out;
707 }
708
709
710 spin_lock(&vgic->lock);
711
712 /*
713 * Ensure that no other VCPU is running by checking the vcpu->cpu
714 * field. If no other VPCUs are running we can safely access the VGIC
715 * state, because even if another VPU is run after this point, that
716 * VCPU will not touch the vgic state, because it will block on
717 * getting the vgic->lock in kvm_vgic_sync_hwstate().
718 */
719 kvm_for_each_vcpu(c, tmp_vcpu, dev->kvm) {
720 if (unlikely(tmp_vcpu->cpu != -1)) {
721 ret = -EBUSY;
722 goto out_vgic_unlock;
723 }
724 }
725
726 /*
727 * Move all pending IRQs from the LRs on all VCPUs so the pending
728 * state can be properly represented in the register state accessible
729 * through this API.
730 */
731 kvm_for_each_vcpu(c, tmp_vcpu, dev->kvm)
732 vgic_unqueue_irqs(tmp_vcpu);
733
734 offset -= r->base;
735 r->handle_mmio(vcpu, &mmio, offset);
736
737 if (!is_write)
738 *reg = mmio_data_read(&mmio, ~0);
739
740 ret = 0;
741out_vgic_unlock:
742 spin_unlock(&vgic->lock);
743out:
744 mutex_unlock(&dev->kvm->lock);
745 return ret;
746}
747
748static int vgic_v2_create(struct kvm_device *dev, u32 type)
749{
750 return kvm_vgic_create(dev->kvm, type);
751}
752
753static void vgic_v2_destroy(struct kvm_device *dev)
754{
755 kfree(dev);
756}
757
758static int vgic_v2_set_attr(struct kvm_device *dev,
759 struct kvm_device_attr *attr)
760{
761 int ret;
762
763 ret = vgic_set_common_attr(dev, attr);
764 if (ret != -ENXIO)
765 return ret;
766
767 switch (attr->group) {
768 case KVM_DEV_ARM_VGIC_GRP_DIST_REGS:
769 case KVM_DEV_ARM_VGIC_GRP_CPU_REGS: {
770 u32 __user *uaddr = (u32 __user *)(long)attr->addr;
771 u32 reg;
772
773 if (get_user(reg, uaddr))
774 return -EFAULT;
775
776 return vgic_attr_regs_access(dev, attr, &reg, true);
777 }
778
779 }
780
781 return -ENXIO;
782}
783
784static int vgic_v2_get_attr(struct kvm_device *dev,
785 struct kvm_device_attr *attr)
786{
787 int ret;
788
789 ret = vgic_get_common_attr(dev, attr);
790 if (ret != -ENXIO)
791 return ret;
792
793 switch (attr->group) {
794 case KVM_DEV_ARM_VGIC_GRP_DIST_REGS:
795 case KVM_DEV_ARM_VGIC_GRP_CPU_REGS: {
796 u32 __user *uaddr = (u32 __user *)(long)attr->addr;
797 u32 reg = 0;
798
799 ret = vgic_attr_regs_access(dev, attr, &reg, false);
800 if (ret)
801 return ret;
802 return put_user(reg, uaddr);
803 }
804
805 }
806
807 return -ENXIO;
808}
809
810static int vgic_v2_has_attr(struct kvm_device *dev,
811 struct kvm_device_attr *attr)
812{
813 phys_addr_t offset;
814
815 switch (attr->group) {
816 case KVM_DEV_ARM_VGIC_GRP_ADDR:
817 switch (attr->attr) {
818 case KVM_VGIC_V2_ADDR_TYPE_DIST:
819 case KVM_VGIC_V2_ADDR_TYPE_CPU:
820 return 0;
821 }
822 break;
823 case KVM_DEV_ARM_VGIC_GRP_DIST_REGS:
824 offset = attr->attr & KVM_DEV_ARM_VGIC_OFFSET_MASK;
825 return vgic_has_attr_regs(vgic_dist_ranges, offset);
826 case KVM_DEV_ARM_VGIC_GRP_CPU_REGS:
827 offset = attr->attr & KVM_DEV_ARM_VGIC_OFFSET_MASK;
828 return vgic_has_attr_regs(vgic_cpu_ranges, offset);
829 case KVM_DEV_ARM_VGIC_GRP_NR_IRQS:
830 return 0;
831 case KVM_DEV_ARM_VGIC_GRP_CTRL:
832 switch (attr->attr) {
833 case KVM_DEV_ARM_VGIC_CTRL_INIT:
834 return 0;
835 }
836 }
837 return -ENXIO;
838}
839
840struct kvm_device_ops kvm_arm_vgic_v2_ops = {
841 .name = "kvm-arm-vgic-v2",
842 .create = vgic_v2_create,
843 .destroy = vgic_v2_destroy,
844 .set_attr = vgic_v2_set_attr,
845 .get_attr = vgic_v2_get_attr,
846 .has_attr = vgic_v2_has_attr,
847};
diff --git a/virt/kvm/arm/vgic.c b/virt/kvm/arm/vgic.c
index ce6c998a3f8d..0e84292c2197 100644
--- a/virt/kvm/arm/vgic.c
+++ b/virt/kvm/arm/vgic.c
@@ -77,12 +77,8 @@
77 77
78#include "vgic.h" 78#include "vgic.h"
79 79
80#define GICC_ARCH_VERSION_V2 0x2
81
82static void vgic_retire_disabled_irqs(struct kvm_vcpu *vcpu); 80static void vgic_retire_disabled_irqs(struct kvm_vcpu *vcpu);
83static void vgic_retire_lr(int lr_nr, int irq, struct kvm_vcpu *vcpu); 81static void vgic_retire_lr(int lr_nr, int irq, struct kvm_vcpu *vcpu);
84static u8 *vgic_get_sgi_sources(struct vgic_dist *dist, int vcpu_id, int sgi);
85static void vgic_dispatch_sgi(struct kvm_vcpu *vcpu, u32 reg);
86static struct vgic_lr vgic_get_lr(const struct kvm_vcpu *vcpu, int lr); 82static struct vgic_lr vgic_get_lr(const struct kvm_vcpu *vcpu, int lr);
87static void vgic_set_lr(struct kvm_vcpu *vcpu, int lr, struct vgic_lr lr_desc); 83static void vgic_set_lr(struct kvm_vcpu *vcpu, int lr, struct vgic_lr lr_desc);
88 84
@@ -421,41 +417,6 @@ void vgic_reg_access(struct kvm_exit_mmio *mmio, u32 *reg,
421 } 417 }
422} 418}
423 419
424static bool handle_mmio_misc(struct kvm_vcpu *vcpu,
425 struct kvm_exit_mmio *mmio, phys_addr_t offset)
426{
427 u32 reg;
428 u32 word_offset = offset & 3;
429
430 switch (offset & ~3) {
431 case 0: /* GICD_CTLR */
432 reg = vcpu->kvm->arch.vgic.enabled;
433 vgic_reg_access(mmio, &reg, word_offset,
434 ACCESS_READ_VALUE | ACCESS_WRITE_VALUE);
435 if (mmio->is_write) {
436 vcpu->kvm->arch.vgic.enabled = reg & 1;
437 vgic_update_state(vcpu->kvm);
438 return true;
439 }
440 break;
441
442 case 4: /* GICD_TYPER */
443 reg = (atomic_read(&vcpu->kvm->online_vcpus) - 1) << 5;
444 reg |= (vcpu->kvm->arch.vgic.nr_irqs >> 5) - 1;
445 vgic_reg_access(mmio, &reg, word_offset,
446 ACCESS_READ_VALUE | ACCESS_WRITE_IGNORED);
447 break;
448
449 case 8: /* GICD_IIDR */
450 reg = (PRODUCT_ID_KVM << 24) | (IMPLEMENTER_ARM << 0);
451 vgic_reg_access(mmio, &reg, word_offset,
452 ACCESS_READ_VALUE | ACCESS_WRITE_IGNORED);
453 break;
454 }
455
456 return false;
457}
458
459bool handle_mmio_raz_wi(struct kvm_vcpu *vcpu, struct kvm_exit_mmio *mmio, 420bool handle_mmio_raz_wi(struct kvm_vcpu *vcpu, struct kvm_exit_mmio *mmio,
460 phys_addr_t offset) 421 phys_addr_t offset)
461{ 422{
@@ -486,22 +447,6 @@ bool vgic_handle_enable_reg(struct kvm *kvm, struct kvm_exit_mmio *mmio,
486 return false; 447 return false;
487} 448}
488 449
489static bool handle_mmio_set_enable_reg(struct kvm_vcpu *vcpu,
490 struct kvm_exit_mmio *mmio,
491 phys_addr_t offset)
492{
493 return vgic_handle_enable_reg(vcpu->kvm, mmio, offset,
494 vcpu->vcpu_id, ACCESS_WRITE_SETBIT);
495}
496
497static bool handle_mmio_clear_enable_reg(struct kvm_vcpu *vcpu,
498 struct kvm_exit_mmio *mmio,
499 phys_addr_t offset)
500{
501 return vgic_handle_enable_reg(vcpu->kvm, mmio, offset,
502 vcpu->vcpu_id, ACCESS_WRITE_CLEARBIT);
503}
504
505bool vgic_handle_set_pending_reg(struct kvm *kvm, 450bool vgic_handle_set_pending_reg(struct kvm *kvm,
506 struct kvm_exit_mmio *mmio, 451 struct kvm_exit_mmio *mmio,
507 phys_addr_t offset, int vcpu_id) 452 phys_addr_t offset, int vcpu_id)
@@ -575,109 +520,6 @@ bool vgic_handle_clear_pending_reg(struct kvm *kvm,
575 return false; 520 return false;
576} 521}
577 522
578static bool handle_mmio_set_pending_reg(struct kvm_vcpu *vcpu,
579 struct kvm_exit_mmio *mmio,
580 phys_addr_t offset)
581{
582 return vgic_handle_set_pending_reg(vcpu->kvm, mmio, offset,
583 vcpu->vcpu_id);
584}
585
586static bool handle_mmio_clear_pending_reg(struct kvm_vcpu *vcpu,
587 struct kvm_exit_mmio *mmio,
588 phys_addr_t offset)
589{
590 return vgic_handle_clear_pending_reg(vcpu->kvm, mmio, offset,
591 vcpu->vcpu_id);
592}
593
594static bool handle_mmio_priority_reg(struct kvm_vcpu *vcpu,
595 struct kvm_exit_mmio *mmio,
596 phys_addr_t offset)
597{
598 u32 *reg = vgic_bytemap_get_reg(&vcpu->kvm->arch.vgic.irq_priority,
599 vcpu->vcpu_id, offset);
600 vgic_reg_access(mmio, reg, offset,
601 ACCESS_READ_VALUE | ACCESS_WRITE_VALUE);
602 return false;
603}
604
605#define GICD_ITARGETSR_SIZE 32
606#define GICD_CPUTARGETS_BITS 8
607#define GICD_IRQS_PER_ITARGETSR (GICD_ITARGETSR_SIZE / GICD_CPUTARGETS_BITS)
608static u32 vgic_get_target_reg(struct kvm *kvm, int irq)
609{
610 struct vgic_dist *dist = &kvm->arch.vgic;
611 int i;
612 u32 val = 0;
613
614 irq -= VGIC_NR_PRIVATE_IRQS;
615
616 for (i = 0; i < GICD_IRQS_PER_ITARGETSR; i++)
617 val |= 1 << (dist->irq_spi_cpu[irq + i] + i * 8);
618
619 return val;
620}
621
622static void vgic_set_target_reg(struct kvm *kvm, u32 val, int irq)
623{
624 struct vgic_dist *dist = &kvm->arch.vgic;
625 struct kvm_vcpu *vcpu;
626 int i, c;
627 unsigned long *bmap;
628 u32 target;
629
630 irq -= VGIC_NR_PRIVATE_IRQS;
631
632 /*
633 * Pick the LSB in each byte. This ensures we target exactly
634 * one vcpu per IRQ. If the byte is null, assume we target
635 * CPU0.
636 */
637 for (i = 0; i < GICD_IRQS_PER_ITARGETSR; i++) {
638 int shift = i * GICD_CPUTARGETS_BITS;
639 target = ffs((val >> shift) & 0xffU);
640 target = target ? (target - 1) : 0;
641 dist->irq_spi_cpu[irq + i] = target;
642 kvm_for_each_vcpu(c, vcpu, kvm) {
643 bmap = vgic_bitmap_get_shared_map(&dist->irq_spi_target[c]);
644 if (c == target)
645 set_bit(irq + i, bmap);
646 else
647 clear_bit(irq + i, bmap);
648 }
649 }
650}
651
652static bool handle_mmio_target_reg(struct kvm_vcpu *vcpu,
653 struct kvm_exit_mmio *mmio,
654 phys_addr_t offset)
655{
656 u32 reg;
657
658 /* We treat the banked interrupts targets as read-only */
659 if (offset < 32) {
660 u32 roreg = 1 << vcpu->vcpu_id;
661 roreg |= roreg << 8;
662 roreg |= roreg << 16;
663
664 vgic_reg_access(mmio, &roreg, offset,
665 ACCESS_READ_VALUE | ACCESS_WRITE_IGNORED);
666 return false;
667 }
668
669 reg = vgic_get_target_reg(vcpu->kvm, offset & ~3U);
670 vgic_reg_access(mmio, &reg, offset,
671 ACCESS_READ_VALUE | ACCESS_WRITE_VALUE);
672 if (mmio->is_write) {
673 vgic_set_target_reg(vcpu->kvm, reg, offset & ~3U);
674 vgic_update_state(vcpu->kvm);
675 return true;
676 }
677
678 return false;
679}
680
681static u32 vgic_cfg_expand(u16 val) 523static u32 vgic_cfg_expand(u16 val)
682{ 524{
683 u32 res = 0; 525 u32 res = 0;
@@ -745,39 +587,6 @@ bool vgic_handle_cfg_reg(u32 *reg, struct kvm_exit_mmio *mmio,
745 return false; 587 return false;
746} 588}
747 589
748static bool handle_mmio_cfg_reg(struct kvm_vcpu *vcpu,
749 struct kvm_exit_mmio *mmio, phys_addr_t offset)
750{
751 u32 *reg;
752
753 reg = vgic_bitmap_get_reg(&vcpu->kvm->arch.vgic.irq_cfg,
754 vcpu->vcpu_id, offset >> 1);
755
756 return vgic_handle_cfg_reg(reg, mmio, offset);
757}
758
759static bool handle_mmio_sgi_reg(struct kvm_vcpu *vcpu,
760 struct kvm_exit_mmio *mmio, phys_addr_t offset)
761{
762 u32 reg;
763 vgic_reg_access(mmio, &reg, offset,
764 ACCESS_READ_RAZ | ACCESS_WRITE_VALUE);
765 if (mmio->is_write) {
766 vgic_dispatch_sgi(vcpu, reg);
767 vgic_update_state(vcpu->kvm);
768 return true;
769 }
770
771 return false;
772}
773
774static void vgic_v2_add_sgi_source(struct kvm_vcpu *vcpu, int irq, int source)
775{
776 struct vgic_dist *dist = &vcpu->kvm->arch.vgic;
777
778 *vgic_get_sgi_sources(dist, vcpu->vcpu_id, irq) |= 1 << source;
779}
780
781/** 590/**
782 * vgic_unqueue_irqs - move pending IRQs from LRs to the distributor 591 * vgic_unqueue_irqs - move pending IRQs from LRs to the distributor
783 * @vgic_cpu: Pointer to the vgic_cpu struct holding the LRs 592 * @vgic_cpu: Pointer to the vgic_cpu struct holding the LRs
@@ -838,168 +647,6 @@ void vgic_unqueue_irqs(struct kvm_vcpu *vcpu)
838 } 647 }
839} 648}
840 649
841/* Handle reads of GICD_CPENDSGIRn and GICD_SPENDSGIRn */
842static bool read_set_clear_sgi_pend_reg(struct kvm_vcpu *vcpu,
843 struct kvm_exit_mmio *mmio,
844 phys_addr_t offset)
845{
846 struct vgic_dist *dist = &vcpu->kvm->arch.vgic;
847 int sgi;
848 int min_sgi = (offset & ~0x3);
849 int max_sgi = min_sgi + 3;
850 int vcpu_id = vcpu->vcpu_id;
851 u32 reg = 0;
852
853 /* Copy source SGIs from distributor side */
854 for (sgi = min_sgi; sgi <= max_sgi; sgi++) {
855 int shift = 8 * (sgi - min_sgi);
856 reg |= ((u32)*vgic_get_sgi_sources(dist, vcpu_id, sgi)) << shift;
857 }
858
859 mmio_data_write(mmio, ~0, reg);
860 return false;
861}
862
863static bool write_set_clear_sgi_pend_reg(struct kvm_vcpu *vcpu,
864 struct kvm_exit_mmio *mmio,
865 phys_addr_t offset, bool set)
866{
867 struct vgic_dist *dist = &vcpu->kvm->arch.vgic;
868 int sgi;
869 int min_sgi = (offset & ~0x3);
870 int max_sgi = min_sgi + 3;
871 int vcpu_id = vcpu->vcpu_id;
872 u32 reg;
873 bool updated = false;
874
875 reg = mmio_data_read(mmio, ~0);
876
877 /* Clear pending SGIs on the distributor */
878 for (sgi = min_sgi; sgi <= max_sgi; sgi++) {
879 u8 mask = reg >> (8 * (sgi - min_sgi));
880 u8 *src = vgic_get_sgi_sources(dist, vcpu_id, sgi);
881 if (set) {
882 if ((*src & mask) != mask)
883 updated = true;
884 *src |= mask;
885 } else {
886 if (*src & mask)
887 updated = true;
888 *src &= ~mask;
889 }
890 }
891
892 if (updated)
893 vgic_update_state(vcpu->kvm);
894
895 return updated;
896}
897
898static bool handle_mmio_sgi_set(struct kvm_vcpu *vcpu,
899 struct kvm_exit_mmio *mmio,
900 phys_addr_t offset)
901{
902 if (!mmio->is_write)
903 return read_set_clear_sgi_pend_reg(vcpu, mmio, offset);
904 else
905 return write_set_clear_sgi_pend_reg(vcpu, mmio, offset, true);
906}
907
908static bool handle_mmio_sgi_clear(struct kvm_vcpu *vcpu,
909 struct kvm_exit_mmio *mmio,
910 phys_addr_t offset)
911{
912 if (!mmio->is_write)
913 return read_set_clear_sgi_pend_reg(vcpu, mmio, offset);
914 else
915 return write_set_clear_sgi_pend_reg(vcpu, mmio, offset, false);
916}
917
918static const struct kvm_mmio_range vgic_dist_ranges[] = {
919 {
920 .base = GIC_DIST_CTRL,
921 .len = 12,
922 .bits_per_irq = 0,
923 .handle_mmio = handle_mmio_misc,
924 },
925 {
926 .base = GIC_DIST_IGROUP,
927 .len = VGIC_MAX_IRQS / 8,
928 .bits_per_irq = 1,
929 .handle_mmio = handle_mmio_raz_wi,
930 },
931 {
932 .base = GIC_DIST_ENABLE_SET,
933 .len = VGIC_MAX_IRQS / 8,
934 .bits_per_irq = 1,
935 .handle_mmio = handle_mmio_set_enable_reg,
936 },
937 {
938 .base = GIC_DIST_ENABLE_CLEAR,
939 .len = VGIC_MAX_IRQS / 8,
940 .bits_per_irq = 1,
941 .handle_mmio = handle_mmio_clear_enable_reg,
942 },
943 {
944 .base = GIC_DIST_PENDING_SET,
945 .len = VGIC_MAX_IRQS / 8,
946 .bits_per_irq = 1,
947 .handle_mmio = handle_mmio_set_pending_reg,
948 },
949 {
950 .base = GIC_DIST_PENDING_CLEAR,
951 .len = VGIC_MAX_IRQS / 8,
952 .bits_per_irq = 1,
953 .handle_mmio = handle_mmio_clear_pending_reg,
954 },
955 {
956 .base = GIC_DIST_ACTIVE_SET,
957 .len = VGIC_MAX_IRQS / 8,
958 .bits_per_irq = 1,
959 .handle_mmio = handle_mmio_raz_wi,
960 },
961 {
962 .base = GIC_DIST_ACTIVE_CLEAR,
963 .len = VGIC_MAX_IRQS / 8,
964 .bits_per_irq = 1,
965 .handle_mmio = handle_mmio_raz_wi,
966 },
967 {
968 .base = GIC_DIST_PRI,
969 .len = VGIC_MAX_IRQS,
970 .bits_per_irq = 8,
971 .handle_mmio = handle_mmio_priority_reg,
972 },
973 {
974 .base = GIC_DIST_TARGET,
975 .len = VGIC_MAX_IRQS,
976 .bits_per_irq = 8,
977 .handle_mmio = handle_mmio_target_reg,
978 },
979 {
980 .base = GIC_DIST_CONFIG,
981 .len = VGIC_MAX_IRQS / 4,
982 .bits_per_irq = 2,
983 .handle_mmio = handle_mmio_cfg_reg,
984 },
985 {
986 .base = GIC_DIST_SOFTINT,
987 .len = 4,
988 .handle_mmio = handle_mmio_sgi_reg,
989 },
990 {
991 .base = GIC_DIST_SGI_PENDING_CLEAR,
992 .len = VGIC_NR_SGIS,
993 .handle_mmio = handle_mmio_sgi_clear,
994 },
995 {
996 .base = GIC_DIST_SGI_PENDING_SET,
997 .len = VGIC_NR_SGIS,
998 .handle_mmio = handle_mmio_sgi_set,
999 },
1000 {}
1001};
1002
1003const 650const
1004struct kvm_mmio_range *vgic_find_range(const struct kvm_mmio_range *ranges, 651struct kvm_mmio_range *vgic_find_range(const struct kvm_mmio_range *ranges,
1005 struct kvm_exit_mmio *mmio, 652 struct kvm_exit_mmio *mmio,
@@ -1127,24 +774,6 @@ bool vgic_handle_mmio_range(struct kvm_vcpu *vcpu, struct kvm_run *run,
1127 return true; 774 return true;
1128} 775}
1129 776
1130static bool vgic_v2_handle_mmio(struct kvm_vcpu *vcpu, struct kvm_run *run,
1131 struct kvm_exit_mmio *mmio)
1132{
1133 unsigned long base = vcpu->kvm->arch.vgic.vgic_dist_base;
1134
1135 if (!is_in_range(mmio->phys_addr, mmio->len, base,
1136 KVM_VGIC_V2_DIST_SIZE))
1137 return false;
1138
1139 /* GICv2 does not support accesses wider than 32 bits */
1140 if (mmio->len > 4) {
1141 kvm_inject_dabt(vcpu, mmio->phys_addr);
1142 return true;
1143 }
1144
1145 return vgic_handle_mmio_range(vcpu, run, mmio, vgic_dist_ranges, base);
1146}
1147
1148/** 777/**
1149 * vgic_handle_mmio - handle an in-kernel MMIO access for the GIC emulation 778 * vgic_handle_mmio - handle an in-kernel MMIO access for the GIC emulation
1150 * @vcpu: pointer to the vcpu performing the access 779 * @vcpu: pointer to the vcpu performing the access
@@ -1169,52 +798,6 @@ bool vgic_handle_mmio(struct kvm_vcpu *vcpu, struct kvm_run *run,
1169 return vcpu->kvm->arch.vgic.vm_ops.handle_mmio(vcpu, run, mmio); 798 return vcpu->kvm->arch.vgic.vm_ops.handle_mmio(vcpu, run, mmio);
1170} 799}
1171 800
1172static u8 *vgic_get_sgi_sources(struct vgic_dist *dist, int vcpu_id, int sgi)
1173{
1174 return dist->irq_sgi_sources + vcpu_id * VGIC_NR_SGIS + sgi;
1175}
1176
1177static void vgic_dispatch_sgi(struct kvm_vcpu *vcpu, u32 reg)
1178{
1179 struct kvm *kvm = vcpu->kvm;
1180 struct vgic_dist *dist = &kvm->arch.vgic;
1181 int nrcpus = atomic_read(&kvm->online_vcpus);
1182 u8 target_cpus;
1183 int sgi, mode, c, vcpu_id;
1184
1185 vcpu_id = vcpu->vcpu_id;
1186
1187 sgi = reg & 0xf;
1188 target_cpus = (reg >> 16) & 0xff;
1189 mode = (reg >> 24) & 3;
1190
1191 switch (mode) {
1192 case 0:
1193 if (!target_cpus)
1194 return;
1195 break;
1196
1197 case 1:
1198 target_cpus = ((1 << nrcpus) - 1) & ~(1 << vcpu_id) & 0xff;
1199 break;
1200
1201 case 2:
1202 target_cpus = 1 << vcpu_id;
1203 break;
1204 }
1205
1206 kvm_for_each_vcpu(c, vcpu, kvm) {
1207 if (target_cpus & 1) {
1208 /* Flag the SGI as pending */
1209 vgic_dist_irq_set_pending(vcpu, sgi);
1210 *vgic_get_sgi_sources(dist, c, sgi) |= 1 << vcpu_id;
1211 kvm_debug("SGI%d from CPU%d to CPU%d\n", sgi, vcpu_id, c);
1212 }
1213
1214 target_cpus >>= 1;
1215 }
1216}
1217
1218static int vgic_nr_shared_irqs(struct vgic_dist *dist) 801static int vgic_nr_shared_irqs(struct vgic_dist *dist)
1219{ 802{
1220 return dist->nr_irqs - VGIC_NR_PRIVATE_IRQS; 803 return dist->nr_irqs - VGIC_NR_PRIVATE_IRQS;
@@ -1368,6 +951,7 @@ static void vgic_retire_disabled_irqs(struct kvm_vcpu *vcpu)
1368/* 951/*
1369 * Queue an interrupt to a CPU virtual interface. Return true on success, 952 * Queue an interrupt to a CPU virtual interface. Return true on success,
1370 * or false if it wasn't possible to queue it. 953 * or false if it wasn't possible to queue it.
954 * sgi_source must be zero for any non-SGI interrupts.
1371 */ 955 */
1372bool vgic_queue_irq(struct kvm_vcpu *vcpu, u8 sgi_source_id, int irq) 956bool vgic_queue_irq(struct kvm_vcpu *vcpu, u8 sgi_source_id, int irq)
1373{ 957{
@@ -1418,37 +1002,6 @@ bool vgic_queue_irq(struct kvm_vcpu *vcpu, u8 sgi_source_id, int irq)
1418 return true; 1002 return true;
1419} 1003}
1420 1004
1421static bool vgic_v2_queue_sgi(struct kvm_vcpu *vcpu, int irq)
1422{
1423 struct vgic_dist *dist = &vcpu->kvm->arch.vgic;
1424 unsigned long sources;
1425 int vcpu_id = vcpu->vcpu_id;
1426 int c;
1427
1428 sources = *vgic_get_sgi_sources(dist, vcpu_id, irq);
1429
1430 for_each_set_bit(c, &sources, dist->nr_cpus) {
1431 if (vgic_queue_irq(vcpu, c, irq))
1432 clear_bit(c, &sources);
1433 }
1434
1435 *vgic_get_sgi_sources(dist, vcpu_id, irq) = sources;
1436
1437 /*
1438 * If the sources bitmap has been cleared it means that we
1439 * could queue all the SGIs onto link registers (see the
1440 * clear_bit above), and therefore we are done with them in
1441 * our emulated gic and can get rid of them.
1442 */
1443 if (!sources) {
1444 vgic_dist_irq_clear_pending(vcpu, irq);
1445 vgic_cpu_irq_clear(vcpu, irq);
1446 return true;
1447 }
1448
1449 return false;
1450}
1451
1452static bool vgic_queue_hwirq(struct kvm_vcpu *vcpu, int irq) 1005static bool vgic_queue_hwirq(struct kvm_vcpu *vcpu, int irq)
1453{ 1006{
1454 if (!vgic_can_sample_irq(vcpu, irq)) 1007 if (!vgic_can_sample_irq(vcpu, irq))
@@ -1894,16 +1447,6 @@ void kvm_vgic_destroy(struct kvm *kvm)
1894 dist->nr_cpus = 0; 1447 dist->nr_cpus = 0;
1895} 1448}
1896 1449
1897static int vgic_v2_init_model(struct kvm *kvm)
1898{
1899 int i;
1900
1901 for (i = VGIC_NR_PRIVATE_IRQS; i < kvm->arch.vgic.nr_irqs; i += 4)
1902 vgic_set_target_reg(kvm, 0, i);
1903
1904 return 0;
1905}
1906
1907/* 1450/*
1908 * Allocate and initialize the various data structures. Must be called 1451 * Allocate and initialize the various data structures. Must be called
1909 * with kvm->lock held! 1452 * with kvm->lock held!
@@ -1994,73 +1537,6 @@ out:
1994 return ret; 1537 return ret;
1995} 1538}
1996 1539
1997/**
1998 * kvm_vgic_map_resources - Configure global VGIC state before running any VCPUs
1999 * @kvm: pointer to the kvm struct
2000 *
2001 * Map the virtual CPU interface into the VM before running any VCPUs. We
2002 * can't do this at creation time, because user space must first set the
2003 * virtual CPU interface address in the guest physical address space.
2004 */
2005static int vgic_v2_map_resources(struct kvm *kvm,
2006 const struct vgic_params *params)
2007{
2008 int ret = 0;
2009
2010 if (!irqchip_in_kernel(kvm))
2011 return 0;
2012
2013 mutex_lock(&kvm->lock);
2014
2015 if (vgic_ready(kvm))
2016 goto out;
2017
2018 if (IS_VGIC_ADDR_UNDEF(kvm->arch.vgic.vgic_dist_base) ||
2019 IS_VGIC_ADDR_UNDEF(kvm->arch.vgic.vgic_cpu_base)) {
2020 kvm_err("Need to set vgic cpu and dist addresses first\n");
2021 ret = -ENXIO;
2022 goto out;
2023 }
2024
2025 /*
2026 * Initialize the vgic if this hasn't already been done on demand by
2027 * accessing the vgic state from userspace.
2028 */
2029 ret = vgic_init(kvm);
2030 if (ret) {
2031 kvm_err("Unable to allocate maps\n");
2032 goto out;
2033 }
2034
2035 ret = kvm_phys_addr_ioremap(kvm, kvm->arch.vgic.vgic_cpu_base,
2036 params->vcpu_base, KVM_VGIC_V2_CPU_SIZE,
2037 true);
2038 if (ret) {
2039 kvm_err("Unable to remap VGIC CPU to VCPU\n");
2040 goto out;
2041 }
2042
2043 kvm->arch.vgic.ready = true;
2044out:
2045 if (ret)
2046 kvm_vgic_destroy(kvm);
2047 mutex_unlock(&kvm->lock);
2048 return ret;
2049}
2050
2051void vgic_v2_init_emulation(struct kvm *kvm)
2052{
2053 struct vgic_dist *dist = &kvm->arch.vgic;
2054
2055 dist->vm_ops.handle_mmio = vgic_v2_handle_mmio;
2056 dist->vm_ops.queue_sgi = vgic_v2_queue_sgi;
2057 dist->vm_ops.add_sgi_source = vgic_v2_add_sgi_source;
2058 dist->vm_ops.init_model = vgic_v2_init_model;
2059 dist->vm_ops.map_resources = vgic_v2_map_resources;
2060
2061 kvm->arch.max_vcpus = VGIC_V2_MAX_CPUS;
2062}
2063
2064static int init_vgic_model(struct kvm *kvm, int type) 1540static int init_vgic_model(struct kvm *kvm, int type)
2065{ 1541{
2066 switch (type) { 1542 switch (type) {
@@ -2210,188 +1686,6 @@ int kvm_vgic_addr(struct kvm *kvm, unsigned long type, u64 *addr, bool write)
2210 return r; 1686 return r;
2211} 1687}
2212 1688
2213static bool handle_cpu_mmio_misc(struct kvm_vcpu *vcpu,
2214 struct kvm_exit_mmio *mmio, phys_addr_t offset)
2215{
2216 bool updated = false;
2217 struct vgic_vmcr vmcr;
2218 u32 *vmcr_field;
2219 u32 reg;
2220
2221 vgic_get_vmcr(vcpu, &vmcr);
2222
2223 switch (offset & ~0x3) {
2224 case GIC_CPU_CTRL:
2225 vmcr_field = &vmcr.ctlr;
2226 break;
2227 case GIC_CPU_PRIMASK:
2228 vmcr_field = &vmcr.pmr;
2229 break;
2230 case GIC_CPU_BINPOINT:
2231 vmcr_field = &vmcr.bpr;
2232 break;
2233 case GIC_CPU_ALIAS_BINPOINT:
2234 vmcr_field = &vmcr.abpr;
2235 break;
2236 default:
2237 BUG();
2238 }
2239
2240 if (!mmio->is_write) {
2241 reg = *vmcr_field;
2242 mmio_data_write(mmio, ~0, reg);
2243 } else {
2244 reg = mmio_data_read(mmio, ~0);
2245 if (reg != *vmcr_field) {
2246 *vmcr_field = reg;
2247 vgic_set_vmcr(vcpu, &vmcr);
2248 updated = true;
2249 }
2250 }
2251 return updated;
2252}
2253
2254static bool handle_mmio_abpr(struct kvm_vcpu *vcpu,
2255 struct kvm_exit_mmio *mmio, phys_addr_t offset)
2256{
2257 return handle_cpu_mmio_misc(vcpu, mmio, GIC_CPU_ALIAS_BINPOINT);
2258}
2259
2260static bool handle_cpu_mmio_ident(struct kvm_vcpu *vcpu,
2261 struct kvm_exit_mmio *mmio,
2262 phys_addr_t offset)
2263{
2264 u32 reg;
2265
2266 if (mmio->is_write)
2267 return false;
2268
2269 /* GICC_IIDR */
2270 reg = (PRODUCT_ID_KVM << 20) |
2271 (GICC_ARCH_VERSION_V2 << 16) |
2272 (IMPLEMENTER_ARM << 0);
2273 mmio_data_write(mmio, ~0, reg);
2274 return false;
2275}
2276
2277/*
2278 * CPU Interface Register accesses - these are not accessed by the VM, but by
2279 * user space for saving and restoring VGIC state.
2280 */
2281static const struct kvm_mmio_range vgic_cpu_ranges[] = {
2282 {
2283 .base = GIC_CPU_CTRL,
2284 .len = 12,
2285 .handle_mmio = handle_cpu_mmio_misc,
2286 },
2287 {
2288 .base = GIC_CPU_ALIAS_BINPOINT,
2289 .len = 4,
2290 .handle_mmio = handle_mmio_abpr,
2291 },
2292 {
2293 .base = GIC_CPU_ACTIVEPRIO,
2294 .len = 16,
2295 .handle_mmio = handle_mmio_raz_wi,
2296 },
2297 {
2298 .base = GIC_CPU_IDENT,
2299 .len = 4,
2300 .handle_mmio = handle_cpu_mmio_ident,
2301 },
2302};
2303
2304static int vgic_attr_regs_access(struct kvm_device *dev,
2305 struct kvm_device_attr *attr,
2306 u32 *reg, bool is_write)
2307{
2308 const struct kvm_mmio_range *r = NULL, *ranges;
2309 phys_addr_t offset;
2310 int ret, cpuid, c;
2311 struct kvm_vcpu *vcpu, *tmp_vcpu;
2312 struct vgic_dist *vgic;
2313 struct kvm_exit_mmio mmio;
2314
2315 offset = attr->attr & KVM_DEV_ARM_VGIC_OFFSET_MASK;
2316 cpuid = (attr->attr & KVM_DEV_ARM_VGIC_CPUID_MASK) >>
2317 KVM_DEV_ARM_VGIC_CPUID_SHIFT;
2318
2319 mutex_lock(&dev->kvm->lock);
2320
2321 ret = vgic_init(dev->kvm);
2322 if (ret)
2323 goto out;
2324
2325 if (cpuid >= atomic_read(&dev->kvm->online_vcpus)) {
2326 ret = -EINVAL;
2327 goto out;
2328 }
2329
2330 vcpu = kvm_get_vcpu(dev->kvm, cpuid);
2331 vgic = &dev->kvm->arch.vgic;
2332
2333 mmio.len = 4;
2334 mmio.is_write = is_write;
2335 if (is_write)
2336 mmio_data_write(&mmio, ~0, *reg);
2337 switch (attr->group) {
2338 case KVM_DEV_ARM_VGIC_GRP_DIST_REGS:
2339 mmio.phys_addr = vgic->vgic_dist_base + offset;
2340 ranges = vgic_dist_ranges;
2341 break;
2342 case KVM_DEV_ARM_VGIC_GRP_CPU_REGS:
2343 mmio.phys_addr = vgic->vgic_cpu_base + offset;
2344 ranges = vgic_cpu_ranges;
2345 break;
2346 default:
2347 BUG();
2348 }
2349 r = vgic_find_range(ranges, &mmio, offset);
2350
2351 if (unlikely(!r || !r->handle_mmio)) {
2352 ret = -ENXIO;
2353 goto out;
2354 }
2355
2356
2357 spin_lock(&vgic->lock);
2358
2359 /*
2360 * Ensure that no other VCPU is running by checking the vcpu->cpu
2361 * field. If no other VPCUs are running we can safely access the VGIC
2362 * state, because even if another VPU is run after this point, that
2363 * VCPU will not touch the vgic state, because it will block on
2364 * getting the vgic->lock in kvm_vgic_sync_hwstate().
2365 */
2366 kvm_for_each_vcpu(c, tmp_vcpu, dev->kvm) {
2367 if (unlikely(tmp_vcpu->cpu != -1)) {
2368 ret = -EBUSY;
2369 goto out_vgic_unlock;
2370 }
2371 }
2372
2373 /*
2374 * Move all pending IRQs from the LRs on all VCPUs so the pending
2375 * state can be properly represented in the register state accessible
2376 * through this API.
2377 */
2378 kvm_for_each_vcpu(c, tmp_vcpu, dev->kvm)
2379 vgic_unqueue_irqs(tmp_vcpu);
2380
2381 offset -= r->base;
2382 r->handle_mmio(vcpu, &mmio, offset);
2383
2384 if (!is_write)
2385 *reg = mmio_data_read(&mmio, ~0);
2386
2387 ret = 0;
2388out_vgic_unlock:
2389 spin_unlock(&vgic->lock);
2390out:
2391 mutex_unlock(&dev->kvm->lock);
2392 return ret;
2393}
2394
2395int vgic_set_common_attr(struct kvm_device *dev, struct kvm_device_attr *attr) 1689int vgic_set_common_attr(struct kvm_device *dev, struct kvm_device_attr *attr)
2396{ 1690{
2397 int r; 1691 int r;
@@ -2451,31 +1745,6 @@ int vgic_set_common_attr(struct kvm_device *dev, struct kvm_device_attr *attr)
2451 return -ENXIO; 1745 return -ENXIO;
2452} 1746}
2453 1747
2454static int vgic_set_attr(struct kvm_device *dev, struct kvm_device_attr *attr)
2455{
2456 int ret;
2457
2458 ret = vgic_set_common_attr(dev, attr);
2459 if (ret != -ENXIO)
2460 return ret;
2461
2462 switch (attr->group) {
2463 case KVM_DEV_ARM_VGIC_GRP_DIST_REGS:
2464 case KVM_DEV_ARM_VGIC_GRP_CPU_REGS: {
2465 u32 __user *uaddr = (u32 __user *)(long)attr->addr;
2466 u32 reg;
2467
2468 if (get_user(reg, uaddr))
2469 return -EFAULT;
2470
2471 return vgic_attr_regs_access(dev, attr, &reg, true);
2472 }
2473
2474 }
2475
2476 return -ENXIO;
2477}
2478
2479int vgic_get_common_attr(struct kvm_device *dev, struct kvm_device_attr *attr) 1748int vgic_get_common_attr(struct kvm_device *dev, struct kvm_device_attr *attr)
2480{ 1749{
2481 int r = -ENXIO; 1750 int r = -ENXIO;
@@ -2506,31 +1775,6 @@ int vgic_get_common_attr(struct kvm_device *dev, struct kvm_device_attr *attr)
2506 return r; 1775 return r;
2507} 1776}
2508 1777
2509static int vgic_get_attr(struct kvm_device *dev, struct kvm_device_attr *attr)
2510{
2511 int ret;
2512
2513 ret = vgic_get_common_attr(dev, attr);
2514 if (ret != -ENXIO)
2515 return ret;
2516
2517 switch (attr->group) {
2518 case KVM_DEV_ARM_VGIC_GRP_DIST_REGS:
2519 case KVM_DEV_ARM_VGIC_GRP_CPU_REGS: {
2520 u32 __user *uaddr = (u32 __user *)(long)attr->addr;
2521 u32 reg = 0;
2522
2523 ret = vgic_attr_regs_access(dev, attr, &reg, false);
2524 if (ret)
2525 return ret;
2526 return put_user(reg, uaddr);
2527 }
2528
2529 }
2530
2531 return -ENXIO;
2532}
2533
2534int vgic_has_attr_regs(const struct kvm_mmio_range *ranges, phys_addr_t offset) 1778int vgic_has_attr_regs(const struct kvm_mmio_range *ranges, phys_addr_t offset)
2535{ 1779{
2536 struct kvm_exit_mmio dev_attr_mmio; 1780 struct kvm_exit_mmio dev_attr_mmio;
@@ -2542,54 +1786,6 @@ int vgic_has_attr_regs(const struct kvm_mmio_range *ranges, phys_addr_t offset)
2542 return -ENXIO; 1786 return -ENXIO;
2543} 1787}
2544 1788
2545static int vgic_has_attr(struct kvm_device *dev, struct kvm_device_attr *attr)
2546{
2547 phys_addr_t offset;
2548
2549 switch (attr->group) {
2550 case KVM_DEV_ARM_VGIC_GRP_ADDR:
2551 switch (attr->attr) {
2552 case KVM_VGIC_V2_ADDR_TYPE_DIST:
2553 case KVM_VGIC_V2_ADDR_TYPE_CPU:
2554 return 0;
2555 }
2556 break;
2557 case KVM_DEV_ARM_VGIC_GRP_DIST_REGS:
2558 offset = attr->attr & KVM_DEV_ARM_VGIC_OFFSET_MASK;
2559 return vgic_has_attr_regs(vgic_dist_ranges, offset);
2560 case KVM_DEV_ARM_VGIC_GRP_CPU_REGS:
2561 offset = attr->attr & KVM_DEV_ARM_VGIC_OFFSET_MASK;
2562 return vgic_has_attr_regs(vgic_cpu_ranges, offset);
2563 case KVM_DEV_ARM_VGIC_GRP_NR_IRQS:
2564 return 0;
2565 case KVM_DEV_ARM_VGIC_GRP_CTRL:
2566 switch (attr->attr) {
2567 case KVM_DEV_ARM_VGIC_CTRL_INIT:
2568 return 0;
2569 }
2570 }
2571 return -ENXIO;
2572}
2573
2574void vgic_destroy(struct kvm_device *dev)
2575{
2576 kfree(dev);
2577}
2578
2579int vgic_create(struct kvm_device *dev, u32 type)
2580{
2581 return kvm_vgic_create(dev->kvm, type);
2582}
2583
2584struct kvm_device_ops kvm_arm_vgic_v2_ops = {
2585 .name = "kvm-arm-vgic",
2586 .create = vgic_create,
2587 .destroy = vgic_destroy,
2588 .set_attr = vgic_set_attr,
2589 .get_attr = vgic_get_attr,
2590 .has_attr = vgic_has_attr,
2591};
2592
2593static void vgic_init_maintenance_interrupt(void *info) 1789static void vgic_init_maintenance_interrupt(void *info)
2594{ 1790{
2595 enable_percpu_irq(vgic->maint_irq, 0); 1791 enable_percpu_irq(vgic->maint_irq, 0);