diff options
author | Gleb Natapov <gleb@redhat.com> | 2013-04-28 05:50:07 -0400 |
---|---|---|
committer | Gleb Natapov <gleb@redhat.com> | 2013-04-28 05:50:07 -0400 |
commit | 064d1afaa5a60fc391d0b4b77599fc8f63f99cd3 (patch) | |
tree | 2e640cdfa50b0048c52e021f07a8b24560251b26 /arch/powerpc/kvm/book3s_xics.c | |
parent | 730dca42c1d363c939da18c1499c7327c66e2b37 (diff) | |
parent | 8b78645c93b5d469e8006d68dbc92edc2640c654 (diff) |
Merge git://github.com/agraf/linux-2.6.git kvm-ppc-next into queue
Diffstat (limited to 'arch/powerpc/kvm/book3s_xics.c')
-rw-r--r-- | arch/powerpc/kvm/book3s_xics.c | 1130 |
1 files changed, 1130 insertions, 0 deletions
diff --git a/arch/powerpc/kvm/book3s_xics.c b/arch/powerpc/kvm/book3s_xics.c new file mode 100644 index 000000000000..ee841ed8a690 --- /dev/null +++ b/arch/powerpc/kvm/book3s_xics.c | |||
@@ -0,0 +1,1130 @@ | |||
1 | /* | ||
2 | * Copyright 2012 Michael Ellerman, IBM Corporation. | ||
3 | * Copyright 2012 Benjamin Herrenschmidt, IBM Corporation. | ||
4 | * | ||
5 | * This program is free software; you can redistribute it and/or modify | ||
6 | * it under the terms of the GNU General Public License, version 2, as | ||
7 | * published by the Free Software Foundation. | ||
8 | */ | ||
9 | |||
10 | #include <linux/kernel.h> | ||
11 | #include <linux/kvm_host.h> | ||
12 | #include <linux/err.h> | ||
13 | #include <linux/gfp.h> | ||
14 | |||
15 | #include <asm/uaccess.h> | ||
16 | #include <asm/kvm_book3s.h> | ||
17 | #include <asm/kvm_ppc.h> | ||
18 | #include <asm/hvcall.h> | ||
19 | #include <asm/xics.h> | ||
20 | #include <asm/debug.h> | ||
21 | |||
22 | #include <linux/debugfs.h> | ||
23 | #include <linux/seq_file.h> | ||
24 | |||
25 | #include "book3s_xics.h" | ||
26 | |||
27 | #if 1 | ||
28 | #define XICS_DBG(fmt...) do { } while (0) | ||
29 | #else | ||
30 | #define XICS_DBG(fmt...) trace_printk(fmt) | ||
31 | #endif | ||
32 | |||
33 | #define ENABLE_REALMODE true | ||
34 | #define DEBUG_REALMODE false | ||
35 | |||
36 | /* | ||
37 | * LOCKING | ||
38 | * ======= | ||
39 | * | ||
40 | * Each ICS has a mutex protecting the information about the IRQ | ||
41 | * sources and avoiding simultaneous deliveries if the same interrupt. | ||
42 | * | ||
43 | * ICP operations are done via a single compare & swap transaction | ||
44 | * (most ICP state fits in the union kvmppc_icp_state) | ||
45 | */ | ||
46 | |||
47 | /* | ||
48 | * TODO | ||
49 | * ==== | ||
50 | * | ||
51 | * - To speed up resends, keep a bitmap of "resend" set bits in the | ||
52 | * ICS | ||
53 | * | ||
54 | * - Speed up server# -> ICP lookup (array ? hash table ?) | ||
55 | * | ||
56 | * - Make ICS lockless as well, or at least a per-interrupt lock or hashed | ||
57 | * locks array to improve scalability | ||
58 | * | ||
59 | * - ioctl's to save/restore the entire state for snapshot & migration | ||
60 | */ | ||
61 | |||
62 | /* -- ICS routines -- */ | ||
63 | |||
64 | static void icp_deliver_irq(struct kvmppc_xics *xics, struct kvmppc_icp *icp, | ||
65 | u32 new_irq); | ||
66 | |||
67 | static int ics_deliver_irq(struct kvmppc_xics *xics, u32 irq, u32 level) | ||
68 | { | ||
69 | struct ics_irq_state *state; | ||
70 | struct kvmppc_ics *ics; | ||
71 | u16 src; | ||
72 | |||
73 | XICS_DBG("ics deliver %#x (level: %d)\n", irq, level); | ||
74 | |||
75 | ics = kvmppc_xics_find_ics(xics, irq, &src); | ||
76 | if (!ics) { | ||
77 | XICS_DBG("ics_deliver_irq: IRQ 0x%06x not found !\n", irq); | ||
78 | return -EINVAL; | ||
79 | } | ||
80 | state = &ics->irq_state[src]; | ||
81 | if (!state->exists) | ||
82 | return -EINVAL; | ||
83 | |||
84 | /* | ||
85 | * We set state->asserted locklessly. This should be fine as | ||
86 | * we are the only setter, thus concurrent access is undefined | ||
87 | * to begin with. | ||
88 | */ | ||
89 | if (level == KVM_INTERRUPT_SET_LEVEL) | ||
90 | state->asserted = 1; | ||
91 | else if (level == KVM_INTERRUPT_UNSET) { | ||
92 | state->asserted = 0; | ||
93 | return 0; | ||
94 | } | ||
95 | |||
96 | /* Attempt delivery */ | ||
97 | icp_deliver_irq(xics, NULL, irq); | ||
98 | |||
99 | return 0; | ||
100 | } | ||
101 | |||
102 | static void ics_check_resend(struct kvmppc_xics *xics, struct kvmppc_ics *ics, | ||
103 | struct kvmppc_icp *icp) | ||
104 | { | ||
105 | int i; | ||
106 | |||
107 | mutex_lock(&ics->lock); | ||
108 | |||
109 | for (i = 0; i < KVMPPC_XICS_IRQ_PER_ICS; i++) { | ||
110 | struct ics_irq_state *state = &ics->irq_state[i]; | ||
111 | |||
112 | if (!state->resend) | ||
113 | continue; | ||
114 | |||
115 | XICS_DBG("resend %#x prio %#x\n", state->number, | ||
116 | state->priority); | ||
117 | |||
118 | mutex_unlock(&ics->lock); | ||
119 | icp_deliver_irq(xics, icp, state->number); | ||
120 | mutex_lock(&ics->lock); | ||
121 | } | ||
122 | |||
123 | mutex_unlock(&ics->lock); | ||
124 | } | ||
125 | |||
126 | static bool write_xive(struct kvmppc_xics *xics, struct kvmppc_ics *ics, | ||
127 | struct ics_irq_state *state, | ||
128 | u32 server, u32 priority, u32 saved_priority) | ||
129 | { | ||
130 | bool deliver; | ||
131 | |||
132 | mutex_lock(&ics->lock); | ||
133 | |||
134 | state->server = server; | ||
135 | state->priority = priority; | ||
136 | state->saved_priority = saved_priority; | ||
137 | deliver = false; | ||
138 | if ((state->masked_pending || state->resend) && priority != MASKED) { | ||
139 | state->masked_pending = 0; | ||
140 | deliver = true; | ||
141 | } | ||
142 | |||
143 | mutex_unlock(&ics->lock); | ||
144 | |||
145 | return deliver; | ||
146 | } | ||
147 | |||
148 | int kvmppc_xics_set_xive(struct kvm *kvm, u32 irq, u32 server, u32 priority) | ||
149 | { | ||
150 | struct kvmppc_xics *xics = kvm->arch.xics; | ||
151 | struct kvmppc_icp *icp; | ||
152 | struct kvmppc_ics *ics; | ||
153 | struct ics_irq_state *state; | ||
154 | u16 src; | ||
155 | |||
156 | if (!xics) | ||
157 | return -ENODEV; | ||
158 | |||
159 | ics = kvmppc_xics_find_ics(xics, irq, &src); | ||
160 | if (!ics) | ||
161 | return -EINVAL; | ||
162 | state = &ics->irq_state[src]; | ||
163 | |||
164 | icp = kvmppc_xics_find_server(kvm, server); | ||
165 | if (!icp) | ||
166 | return -EINVAL; | ||
167 | |||
168 | XICS_DBG("set_xive %#x server %#x prio %#x MP:%d RS:%d\n", | ||
169 | irq, server, priority, | ||
170 | state->masked_pending, state->resend); | ||
171 | |||
172 | if (write_xive(xics, ics, state, server, priority, priority)) | ||
173 | icp_deliver_irq(xics, icp, irq); | ||
174 | |||
175 | return 0; | ||
176 | } | ||
177 | |||
178 | int kvmppc_xics_get_xive(struct kvm *kvm, u32 irq, u32 *server, u32 *priority) | ||
179 | { | ||
180 | struct kvmppc_xics *xics = kvm->arch.xics; | ||
181 | struct kvmppc_ics *ics; | ||
182 | struct ics_irq_state *state; | ||
183 | u16 src; | ||
184 | |||
185 | if (!xics) | ||
186 | return -ENODEV; | ||
187 | |||
188 | ics = kvmppc_xics_find_ics(xics, irq, &src); | ||
189 | if (!ics) | ||
190 | return -EINVAL; | ||
191 | state = &ics->irq_state[src]; | ||
192 | |||
193 | mutex_lock(&ics->lock); | ||
194 | *server = state->server; | ||
195 | *priority = state->priority; | ||
196 | mutex_unlock(&ics->lock); | ||
197 | |||
198 | return 0; | ||
199 | } | ||
200 | |||
201 | int kvmppc_xics_int_on(struct kvm *kvm, u32 irq) | ||
202 | { | ||
203 | struct kvmppc_xics *xics = kvm->arch.xics; | ||
204 | struct kvmppc_icp *icp; | ||
205 | struct kvmppc_ics *ics; | ||
206 | struct ics_irq_state *state; | ||
207 | u16 src; | ||
208 | |||
209 | if (!xics) | ||
210 | return -ENODEV; | ||
211 | |||
212 | ics = kvmppc_xics_find_ics(xics, irq, &src); | ||
213 | if (!ics) | ||
214 | return -EINVAL; | ||
215 | state = &ics->irq_state[src]; | ||
216 | |||
217 | icp = kvmppc_xics_find_server(kvm, state->server); | ||
218 | if (!icp) | ||
219 | return -EINVAL; | ||
220 | |||
221 | if (write_xive(xics, ics, state, state->server, state->saved_priority, | ||
222 | state->saved_priority)) | ||
223 | icp_deliver_irq(xics, icp, irq); | ||
224 | |||
225 | return 0; | ||
226 | } | ||
227 | |||
228 | int kvmppc_xics_int_off(struct kvm *kvm, u32 irq) | ||
229 | { | ||
230 | struct kvmppc_xics *xics = kvm->arch.xics; | ||
231 | struct kvmppc_ics *ics; | ||
232 | struct ics_irq_state *state; | ||
233 | u16 src; | ||
234 | |||
235 | if (!xics) | ||
236 | return -ENODEV; | ||
237 | |||
238 | ics = kvmppc_xics_find_ics(xics, irq, &src); | ||
239 | if (!ics) | ||
240 | return -EINVAL; | ||
241 | state = &ics->irq_state[src]; | ||
242 | |||
243 | write_xive(xics, ics, state, state->server, MASKED, state->priority); | ||
244 | |||
245 | return 0; | ||
246 | } | ||
247 | |||
248 | /* -- ICP routines, including hcalls -- */ | ||
249 | |||
250 | static inline bool icp_try_update(struct kvmppc_icp *icp, | ||
251 | union kvmppc_icp_state old, | ||
252 | union kvmppc_icp_state new, | ||
253 | bool change_self) | ||
254 | { | ||
255 | bool success; | ||
256 | |||
257 | /* Calculate new output value */ | ||
258 | new.out_ee = (new.xisr && (new.pending_pri < new.cppr)); | ||
259 | |||
260 | /* Attempt atomic update */ | ||
261 | success = cmpxchg64(&icp->state.raw, old.raw, new.raw) == old.raw; | ||
262 | if (!success) | ||
263 | goto bail; | ||
264 | |||
265 | XICS_DBG("UPD [%04x] - C:%02x M:%02x PP: %02x PI:%06x R:%d O:%d\n", | ||
266 | icp->server_num, | ||
267 | old.cppr, old.mfrr, old.pending_pri, old.xisr, | ||
268 | old.need_resend, old.out_ee); | ||
269 | XICS_DBG("UPD - C:%02x M:%02x PP: %02x PI:%06x R:%d O:%d\n", | ||
270 | new.cppr, new.mfrr, new.pending_pri, new.xisr, | ||
271 | new.need_resend, new.out_ee); | ||
272 | /* | ||
273 | * Check for output state update | ||
274 | * | ||
275 | * Note that this is racy since another processor could be updating | ||
276 | * the state already. This is why we never clear the interrupt output | ||
277 | * here, we only ever set it. The clear only happens prior to doing | ||
278 | * an update and only by the processor itself. Currently we do it | ||
279 | * in Accept (H_XIRR) and Up_Cppr (H_XPPR). | ||
280 | * | ||
281 | * We also do not try to figure out whether the EE state has changed, | ||
282 | * we unconditionally set it if the new state calls for it. The reason | ||
283 | * for that is that we opportunistically remove the pending interrupt | ||
284 | * flag when raising CPPR, so we need to set it back here if an | ||
285 | * interrupt is still pending. | ||
286 | */ | ||
287 | if (new.out_ee) { | ||
288 | kvmppc_book3s_queue_irqprio(icp->vcpu, | ||
289 | BOOK3S_INTERRUPT_EXTERNAL_LEVEL); | ||
290 | if (!change_self) | ||
291 | kvmppc_fast_vcpu_kick(icp->vcpu); | ||
292 | } | ||
293 | bail: | ||
294 | return success; | ||
295 | } | ||
296 | |||
297 | static void icp_check_resend(struct kvmppc_xics *xics, | ||
298 | struct kvmppc_icp *icp) | ||
299 | { | ||
300 | u32 icsid; | ||
301 | |||
302 | /* Order this load with the test for need_resend in the caller */ | ||
303 | smp_rmb(); | ||
304 | for_each_set_bit(icsid, icp->resend_map, xics->max_icsid + 1) { | ||
305 | struct kvmppc_ics *ics = xics->ics[icsid]; | ||
306 | |||
307 | if (!test_and_clear_bit(icsid, icp->resend_map)) | ||
308 | continue; | ||
309 | if (!ics) | ||
310 | continue; | ||
311 | ics_check_resend(xics, ics, icp); | ||
312 | } | ||
313 | } | ||
314 | |||
315 | static bool icp_try_to_deliver(struct kvmppc_icp *icp, u32 irq, u8 priority, | ||
316 | u32 *reject) | ||
317 | { | ||
318 | union kvmppc_icp_state old_state, new_state; | ||
319 | bool success; | ||
320 | |||
321 | XICS_DBG("try deliver %#x(P:%#x) to server %#x\n", irq, priority, | ||
322 | icp->server_num); | ||
323 | |||
324 | do { | ||
325 | old_state = new_state = ACCESS_ONCE(icp->state); | ||
326 | |||
327 | *reject = 0; | ||
328 | |||
329 | /* See if we can deliver */ | ||
330 | success = new_state.cppr > priority && | ||
331 | new_state.mfrr > priority && | ||
332 | new_state.pending_pri > priority; | ||
333 | |||
334 | /* | ||
335 | * If we can, check for a rejection and perform the | ||
336 | * delivery | ||
337 | */ | ||
338 | if (success) { | ||
339 | *reject = new_state.xisr; | ||
340 | new_state.xisr = irq; | ||
341 | new_state.pending_pri = priority; | ||
342 | } else { | ||
343 | /* | ||
344 | * If we failed to deliver we set need_resend | ||
345 | * so a subsequent CPPR state change causes us | ||
346 | * to try a new delivery. | ||
347 | */ | ||
348 | new_state.need_resend = true; | ||
349 | } | ||
350 | |||
351 | } while (!icp_try_update(icp, old_state, new_state, false)); | ||
352 | |||
353 | return success; | ||
354 | } | ||
355 | |||
356 | static void icp_deliver_irq(struct kvmppc_xics *xics, struct kvmppc_icp *icp, | ||
357 | u32 new_irq) | ||
358 | { | ||
359 | struct ics_irq_state *state; | ||
360 | struct kvmppc_ics *ics; | ||
361 | u32 reject; | ||
362 | u16 src; | ||
363 | |||
364 | /* | ||
365 | * This is used both for initial delivery of an interrupt and | ||
366 | * for subsequent rejection. | ||
367 | * | ||
368 | * Rejection can be racy vs. resends. We have evaluated the | ||
369 | * rejection in an atomic ICP transaction which is now complete, | ||
370 | * so potentially the ICP can already accept the interrupt again. | ||
371 | * | ||
372 | * So we need to retry the delivery. Essentially the reject path | ||
373 | * boils down to a failed delivery. Always. | ||
374 | * | ||
375 | * Now the interrupt could also have moved to a different target, | ||
376 | * thus we may need to re-do the ICP lookup as well | ||
377 | */ | ||
378 | |||
379 | again: | ||
380 | /* Get the ICS state and lock it */ | ||
381 | ics = kvmppc_xics_find_ics(xics, new_irq, &src); | ||
382 | if (!ics) { | ||
383 | XICS_DBG("icp_deliver_irq: IRQ 0x%06x not found !\n", new_irq); | ||
384 | return; | ||
385 | } | ||
386 | state = &ics->irq_state[src]; | ||
387 | |||
388 | /* Get a lock on the ICS */ | ||
389 | mutex_lock(&ics->lock); | ||
390 | |||
391 | /* Get our server */ | ||
392 | if (!icp || state->server != icp->server_num) { | ||
393 | icp = kvmppc_xics_find_server(xics->kvm, state->server); | ||
394 | if (!icp) { | ||
395 | pr_warn("icp_deliver_irq: IRQ 0x%06x server 0x%x not found !\n", | ||
396 | new_irq, state->server); | ||
397 | goto out; | ||
398 | } | ||
399 | } | ||
400 | |||
401 | /* Clear the resend bit of that interrupt */ | ||
402 | state->resend = 0; | ||
403 | |||
404 | /* | ||
405 | * If masked, bail out | ||
406 | * | ||
407 | * Note: PAPR doesn't mention anything about masked pending | ||
408 | * when doing a resend, only when doing a delivery. | ||
409 | * | ||
410 | * However that would have the effect of losing a masked | ||
411 | * interrupt that was rejected and isn't consistent with | ||
412 | * the whole masked_pending business which is about not | ||
413 | * losing interrupts that occur while masked. | ||
414 | * | ||
415 | * I don't differenciate normal deliveries and resends, this | ||
416 | * implementation will differ from PAPR and not lose such | ||
417 | * interrupts. | ||
418 | */ | ||
419 | if (state->priority == MASKED) { | ||
420 | XICS_DBG("irq %#x masked pending\n", new_irq); | ||
421 | state->masked_pending = 1; | ||
422 | goto out; | ||
423 | } | ||
424 | |||
425 | /* | ||
426 | * Try the delivery, this will set the need_resend flag | ||
427 | * in the ICP as part of the atomic transaction if the | ||
428 | * delivery is not possible. | ||
429 | * | ||
430 | * Note that if successful, the new delivery might have itself | ||
431 | * rejected an interrupt that was "delivered" before we took the | ||
432 | * icp mutex. | ||
433 | * | ||
434 | * In this case we do the whole sequence all over again for the | ||
435 | * new guy. We cannot assume that the rejected interrupt is less | ||
436 | * favored than the new one, and thus doesn't need to be delivered, | ||
437 | * because by the time we exit icp_try_to_deliver() the target | ||
438 | * processor may well have alrady consumed & completed it, and thus | ||
439 | * the rejected interrupt might actually be already acceptable. | ||
440 | */ | ||
441 | if (icp_try_to_deliver(icp, new_irq, state->priority, &reject)) { | ||
442 | /* | ||
443 | * Delivery was successful, did we reject somebody else ? | ||
444 | */ | ||
445 | if (reject && reject != XICS_IPI) { | ||
446 | mutex_unlock(&ics->lock); | ||
447 | new_irq = reject; | ||
448 | goto again; | ||
449 | } | ||
450 | } else { | ||
451 | /* | ||
452 | * We failed to deliver the interrupt we need to set the | ||
453 | * resend map bit and mark the ICS state as needing a resend | ||
454 | */ | ||
455 | set_bit(ics->icsid, icp->resend_map); | ||
456 | state->resend = 1; | ||
457 | |||
458 | /* | ||
459 | * If the need_resend flag got cleared in the ICP some time | ||
460 | * between icp_try_to_deliver() atomic update and now, then | ||
461 | * we know it might have missed the resend_map bit. So we | ||
462 | * retry | ||
463 | */ | ||
464 | smp_mb(); | ||
465 | if (!icp->state.need_resend) { | ||
466 | mutex_unlock(&ics->lock); | ||
467 | goto again; | ||
468 | } | ||
469 | } | ||
470 | out: | ||
471 | mutex_unlock(&ics->lock); | ||
472 | } | ||
473 | |||
474 | static void icp_down_cppr(struct kvmppc_xics *xics, struct kvmppc_icp *icp, | ||
475 | u8 new_cppr) | ||
476 | { | ||
477 | union kvmppc_icp_state old_state, new_state; | ||
478 | bool resend; | ||
479 | |||
480 | /* | ||
481 | * This handles several related states in one operation: | ||
482 | * | ||
483 | * ICP State: Down_CPPR | ||
484 | * | ||
485 | * Load CPPR with new value and if the XISR is 0 | ||
486 | * then check for resends: | ||
487 | * | ||
488 | * ICP State: Resend | ||
489 | * | ||
490 | * If MFRR is more favored than CPPR, check for IPIs | ||
491 | * and notify ICS of a potential resend. This is done | ||
492 | * asynchronously (when used in real mode, we will have | ||
493 | * to exit here). | ||
494 | * | ||
495 | * We do not handle the complete Check_IPI as documented | ||
496 | * here. In the PAPR, this state will be used for both | ||
497 | * Set_MFRR and Down_CPPR. However, we know that we aren't | ||
498 | * changing the MFRR state here so we don't need to handle | ||
499 | * the case of an MFRR causing a reject of a pending irq, | ||
500 | * this will have been handled when the MFRR was set in the | ||
501 | * first place. | ||
502 | * | ||
503 | * Thus we don't have to handle rejects, only resends. | ||
504 | * | ||
505 | * When implementing real mode for HV KVM, resend will lead to | ||
506 | * a H_TOO_HARD return and the whole transaction will be handled | ||
507 | * in virtual mode. | ||
508 | */ | ||
509 | do { | ||
510 | old_state = new_state = ACCESS_ONCE(icp->state); | ||
511 | |||
512 | /* Down_CPPR */ | ||
513 | new_state.cppr = new_cppr; | ||
514 | |||
515 | /* | ||
516 | * Cut down Resend / Check_IPI / IPI | ||
517 | * | ||
518 | * The logic is that we cannot have a pending interrupt | ||
519 | * trumped by an IPI at this point (see above), so we | ||
520 | * know that either the pending interrupt is already an | ||
521 | * IPI (in which case we don't care to override it) or | ||
522 | * it's either more favored than us or non existent | ||
523 | */ | ||
524 | if (new_state.mfrr < new_cppr && | ||
525 | new_state.mfrr <= new_state.pending_pri) { | ||
526 | WARN_ON(new_state.xisr != XICS_IPI && | ||
527 | new_state.xisr != 0); | ||
528 | new_state.pending_pri = new_state.mfrr; | ||
529 | new_state.xisr = XICS_IPI; | ||
530 | } | ||
531 | |||
532 | /* Latch/clear resend bit */ | ||
533 | resend = new_state.need_resend; | ||
534 | new_state.need_resend = 0; | ||
535 | |||
536 | } while (!icp_try_update(icp, old_state, new_state, true)); | ||
537 | |||
538 | /* | ||
539 | * Now handle resend checks. Those are asynchronous to the ICP | ||
540 | * state update in HW (ie bus transactions) so we can handle them | ||
541 | * separately here too | ||
542 | */ | ||
543 | if (resend) | ||
544 | icp_check_resend(xics, icp); | ||
545 | } | ||
546 | |||
547 | static noinline unsigned long kvmppc_h_xirr(struct kvm_vcpu *vcpu) | ||
548 | { | ||
549 | union kvmppc_icp_state old_state, new_state; | ||
550 | struct kvmppc_icp *icp = vcpu->arch.icp; | ||
551 | u32 xirr; | ||
552 | |||
553 | /* First, remove EE from the processor */ | ||
554 | kvmppc_book3s_dequeue_irqprio(icp->vcpu, | ||
555 | BOOK3S_INTERRUPT_EXTERNAL_LEVEL); | ||
556 | |||
557 | /* | ||
558 | * ICP State: Accept_Interrupt | ||
559 | * | ||
560 | * Return the pending interrupt (if any) along with the | ||
561 | * current CPPR, then clear the XISR & set CPPR to the | ||
562 | * pending priority | ||
563 | */ | ||
564 | do { | ||
565 | old_state = new_state = ACCESS_ONCE(icp->state); | ||
566 | |||
567 | xirr = old_state.xisr | (((u32)old_state.cppr) << 24); | ||
568 | if (!old_state.xisr) | ||
569 | break; | ||
570 | new_state.cppr = new_state.pending_pri; | ||
571 | new_state.pending_pri = 0xff; | ||
572 | new_state.xisr = 0; | ||
573 | |||
574 | } while (!icp_try_update(icp, old_state, new_state, true)); | ||
575 | |||
576 | XICS_DBG("h_xirr vcpu %d xirr %#x\n", vcpu->vcpu_id, xirr); | ||
577 | |||
578 | return xirr; | ||
579 | } | ||
580 | |||
581 | static noinline int kvmppc_h_ipi(struct kvm_vcpu *vcpu, unsigned long server, | ||
582 | unsigned long mfrr) | ||
583 | { | ||
584 | union kvmppc_icp_state old_state, new_state; | ||
585 | struct kvmppc_xics *xics = vcpu->kvm->arch.xics; | ||
586 | struct kvmppc_icp *icp; | ||
587 | u32 reject; | ||
588 | bool resend; | ||
589 | bool local; | ||
590 | |||
591 | XICS_DBG("h_ipi vcpu %d to server %lu mfrr %#lx\n", | ||
592 | vcpu->vcpu_id, server, mfrr); | ||
593 | |||
594 | icp = vcpu->arch.icp; | ||
595 | local = icp->server_num == server; | ||
596 | if (!local) { | ||
597 | icp = kvmppc_xics_find_server(vcpu->kvm, server); | ||
598 | if (!icp) | ||
599 | return H_PARAMETER; | ||
600 | } | ||
601 | |||
602 | /* | ||
603 | * ICP state: Set_MFRR | ||
604 | * | ||
605 | * If the CPPR is more favored than the new MFRR, then | ||
606 | * nothing needs to be rejected as there can be no XISR to | ||
607 | * reject. If the MFRR is being made less favored then | ||
608 | * there might be a previously-rejected interrupt needing | ||
609 | * to be resent. | ||
610 | * | ||
611 | * If the CPPR is less favored, then we might be replacing | ||
612 | * an interrupt, and thus need to possibly reject it as in | ||
613 | * | ||
614 | * ICP state: Check_IPI | ||
615 | */ | ||
616 | do { | ||
617 | old_state = new_state = ACCESS_ONCE(icp->state); | ||
618 | |||
619 | /* Set_MFRR */ | ||
620 | new_state.mfrr = mfrr; | ||
621 | |||
622 | /* Check_IPI */ | ||
623 | reject = 0; | ||
624 | resend = false; | ||
625 | if (mfrr < new_state.cppr) { | ||
626 | /* Reject a pending interrupt if not an IPI */ | ||
627 | if (mfrr <= new_state.pending_pri) | ||
628 | reject = new_state.xisr; | ||
629 | new_state.pending_pri = mfrr; | ||
630 | new_state.xisr = XICS_IPI; | ||
631 | } | ||
632 | |||
633 | if (mfrr > old_state.mfrr && mfrr > new_state.cppr) { | ||
634 | resend = new_state.need_resend; | ||
635 | new_state.need_resend = 0; | ||
636 | } | ||
637 | } while (!icp_try_update(icp, old_state, new_state, local)); | ||
638 | |||
639 | /* Handle reject */ | ||
640 | if (reject && reject != XICS_IPI) | ||
641 | icp_deliver_irq(xics, icp, reject); | ||
642 | |||
643 | /* Handle resend */ | ||
644 | if (resend) | ||
645 | icp_check_resend(xics, icp); | ||
646 | |||
647 | return H_SUCCESS; | ||
648 | } | ||
649 | |||
650 | static noinline void kvmppc_h_cppr(struct kvm_vcpu *vcpu, unsigned long cppr) | ||
651 | { | ||
652 | union kvmppc_icp_state old_state, new_state; | ||
653 | struct kvmppc_xics *xics = vcpu->kvm->arch.xics; | ||
654 | struct kvmppc_icp *icp = vcpu->arch.icp; | ||
655 | u32 reject; | ||
656 | |||
657 | XICS_DBG("h_cppr vcpu %d cppr %#lx\n", vcpu->vcpu_id, cppr); | ||
658 | |||
659 | /* | ||
660 | * ICP State: Set_CPPR | ||
661 | * | ||
662 | * We can safely compare the new value with the current | ||
663 | * value outside of the transaction as the CPPR is only | ||
664 | * ever changed by the processor on itself | ||
665 | */ | ||
666 | if (cppr > icp->state.cppr) | ||
667 | icp_down_cppr(xics, icp, cppr); | ||
668 | else if (cppr == icp->state.cppr) | ||
669 | return; | ||
670 | |||
671 | /* | ||
672 | * ICP State: Up_CPPR | ||
673 | * | ||
674 | * The processor is raising its priority, this can result | ||
675 | * in a rejection of a pending interrupt: | ||
676 | * | ||
677 | * ICP State: Reject_Current | ||
678 | * | ||
679 | * We can remove EE from the current processor, the update | ||
680 | * transaction will set it again if needed | ||
681 | */ | ||
682 | kvmppc_book3s_dequeue_irqprio(icp->vcpu, | ||
683 | BOOK3S_INTERRUPT_EXTERNAL_LEVEL); | ||
684 | |||
685 | do { | ||
686 | old_state = new_state = ACCESS_ONCE(icp->state); | ||
687 | |||
688 | reject = 0; | ||
689 | new_state.cppr = cppr; | ||
690 | |||
691 | if (cppr <= new_state.pending_pri) { | ||
692 | reject = new_state.xisr; | ||
693 | new_state.xisr = 0; | ||
694 | new_state.pending_pri = 0xff; | ||
695 | } | ||
696 | |||
697 | } while (!icp_try_update(icp, old_state, new_state, true)); | ||
698 | |||
699 | /* | ||
700 | * Check for rejects. They are handled by doing a new delivery | ||
701 | * attempt (see comments in icp_deliver_irq). | ||
702 | */ | ||
703 | if (reject && reject != XICS_IPI) | ||
704 | icp_deliver_irq(xics, icp, reject); | ||
705 | } | ||
706 | |||
707 | static noinline int kvmppc_h_eoi(struct kvm_vcpu *vcpu, unsigned long xirr) | ||
708 | { | ||
709 | struct kvmppc_xics *xics = vcpu->kvm->arch.xics; | ||
710 | struct kvmppc_icp *icp = vcpu->arch.icp; | ||
711 | struct kvmppc_ics *ics; | ||
712 | struct ics_irq_state *state; | ||
713 | u32 irq = xirr & 0x00ffffff; | ||
714 | u16 src; | ||
715 | |||
716 | XICS_DBG("h_eoi vcpu %d eoi %#lx\n", vcpu->vcpu_id, xirr); | ||
717 | |||
718 | /* | ||
719 | * ICP State: EOI | ||
720 | * | ||
721 | * Note: If EOI is incorrectly used by SW to lower the CPPR | ||
722 | * value (ie more favored), we do not check for rejection of | ||
723 | * a pending interrupt, this is a SW error and PAPR sepcifies | ||
724 | * that we don't have to deal with it. | ||
725 | * | ||
726 | * The sending of an EOI to the ICS is handled after the | ||
727 | * CPPR update | ||
728 | * | ||
729 | * ICP State: Down_CPPR which we handle | ||
730 | * in a separate function as it's shared with H_CPPR. | ||
731 | */ | ||
732 | icp_down_cppr(xics, icp, xirr >> 24); | ||
733 | |||
734 | /* IPIs have no EOI */ | ||
735 | if (irq == XICS_IPI) | ||
736 | return H_SUCCESS; | ||
737 | /* | ||
738 | * EOI handling: If the interrupt is still asserted, we need to | ||
739 | * resend it. We can take a lockless "peek" at the ICS state here. | ||
740 | * | ||
741 | * "Message" interrupts will never have "asserted" set | ||
742 | */ | ||
743 | ics = kvmppc_xics_find_ics(xics, irq, &src); | ||
744 | if (!ics) { | ||
745 | XICS_DBG("h_eoi: IRQ 0x%06x not found !\n", irq); | ||
746 | return H_PARAMETER; | ||
747 | } | ||
748 | state = &ics->irq_state[src]; | ||
749 | |||
750 | /* Still asserted, resend it */ | ||
751 | if (state->asserted) | ||
752 | icp_deliver_irq(xics, icp, irq); | ||
753 | |||
754 | return H_SUCCESS; | ||
755 | } | ||
756 | |||
757 | static noinline int kvmppc_xics_rm_complete(struct kvm_vcpu *vcpu, u32 hcall) | ||
758 | { | ||
759 | struct kvmppc_xics *xics = vcpu->kvm->arch.xics; | ||
760 | struct kvmppc_icp *icp = vcpu->arch.icp; | ||
761 | |||
762 | XICS_DBG("XICS_RM: H_%x completing, act: %x state: %lx tgt: %p\n", | ||
763 | hcall, icp->rm_action, icp->rm_dbgstate.raw, icp->rm_dbgtgt); | ||
764 | |||
765 | if (icp->rm_action & XICS_RM_KICK_VCPU) | ||
766 | kvmppc_fast_vcpu_kick(icp->rm_kick_target); | ||
767 | if (icp->rm_action & XICS_RM_CHECK_RESEND) | ||
768 | icp_check_resend(xics, icp); | ||
769 | if (icp->rm_action & XICS_RM_REJECT) | ||
770 | icp_deliver_irq(xics, icp, icp->rm_reject); | ||
771 | |||
772 | icp->rm_action = 0; | ||
773 | |||
774 | return H_SUCCESS; | ||
775 | } | ||
776 | |||
777 | int kvmppc_xics_hcall(struct kvm_vcpu *vcpu, u32 req) | ||
778 | { | ||
779 | struct kvmppc_xics *xics = vcpu->kvm->arch.xics; | ||
780 | unsigned long res; | ||
781 | int rc = H_SUCCESS; | ||
782 | |||
783 | /* Check if we have an ICP */ | ||
784 | if (!xics || !vcpu->arch.icp) | ||
785 | return H_HARDWARE; | ||
786 | |||
787 | /* Check for real mode returning too hard */ | ||
788 | if (xics->real_mode) | ||
789 | return kvmppc_xics_rm_complete(vcpu, req); | ||
790 | |||
791 | switch (req) { | ||
792 | case H_XIRR: | ||
793 | res = kvmppc_h_xirr(vcpu); | ||
794 | kvmppc_set_gpr(vcpu, 4, res); | ||
795 | break; | ||
796 | case H_CPPR: | ||
797 | kvmppc_h_cppr(vcpu, kvmppc_get_gpr(vcpu, 4)); | ||
798 | break; | ||
799 | case H_EOI: | ||
800 | rc = kvmppc_h_eoi(vcpu, kvmppc_get_gpr(vcpu, 4)); | ||
801 | break; | ||
802 | case H_IPI: | ||
803 | rc = kvmppc_h_ipi(vcpu, kvmppc_get_gpr(vcpu, 4), | ||
804 | kvmppc_get_gpr(vcpu, 5)); | ||
805 | break; | ||
806 | } | ||
807 | |||
808 | return rc; | ||
809 | } | ||
810 | |||
811 | |||
812 | /* -- Initialisation code etc. -- */ | ||
813 | |||
814 | static int xics_debug_show(struct seq_file *m, void *private) | ||
815 | { | ||
816 | struct kvmppc_xics *xics = m->private; | ||
817 | struct kvm *kvm = xics->kvm; | ||
818 | struct kvm_vcpu *vcpu; | ||
819 | int icsid, i; | ||
820 | |||
821 | if (!kvm) | ||
822 | return 0; | ||
823 | |||
824 | seq_printf(m, "=========\nICP state\n=========\n"); | ||
825 | |||
826 | kvm_for_each_vcpu(i, vcpu, kvm) { | ||
827 | struct kvmppc_icp *icp = vcpu->arch.icp; | ||
828 | union kvmppc_icp_state state; | ||
829 | |||
830 | if (!icp) | ||
831 | continue; | ||
832 | |||
833 | state.raw = ACCESS_ONCE(icp->state.raw); | ||
834 | seq_printf(m, "cpu server %#lx XIRR:%#x PPRI:%#x CPPR:%#x MFRR:%#x OUT:%d NR:%d\n", | ||
835 | icp->server_num, state.xisr, | ||
836 | state.pending_pri, state.cppr, state.mfrr, | ||
837 | state.out_ee, state.need_resend); | ||
838 | } | ||
839 | |||
840 | for (icsid = 0; icsid <= KVMPPC_XICS_MAX_ICS_ID; icsid++) { | ||
841 | struct kvmppc_ics *ics = xics->ics[icsid]; | ||
842 | |||
843 | if (!ics) | ||
844 | continue; | ||
845 | |||
846 | seq_printf(m, "=========\nICS state for ICS 0x%x\n=========\n", | ||
847 | icsid); | ||
848 | |||
849 | mutex_lock(&ics->lock); | ||
850 | |||
851 | for (i = 0; i < KVMPPC_XICS_IRQ_PER_ICS; i++) { | ||
852 | struct ics_irq_state *irq = &ics->irq_state[i]; | ||
853 | |||
854 | seq_printf(m, "irq 0x%06x: server %#x prio %#x save prio %#x asserted %d resend %d masked pending %d\n", | ||
855 | irq->number, irq->server, irq->priority, | ||
856 | irq->saved_priority, irq->asserted, | ||
857 | irq->resend, irq->masked_pending); | ||
858 | |||
859 | } | ||
860 | mutex_unlock(&ics->lock); | ||
861 | } | ||
862 | return 0; | ||
863 | } | ||
864 | |||
865 | static int xics_debug_open(struct inode *inode, struct file *file) | ||
866 | { | ||
867 | return single_open(file, xics_debug_show, inode->i_private); | ||
868 | } | ||
869 | |||
870 | static const struct file_operations xics_debug_fops = { | ||
871 | .open = xics_debug_open, | ||
872 | .read = seq_read, | ||
873 | .llseek = seq_lseek, | ||
874 | .release = single_release, | ||
875 | }; | ||
876 | |||
877 | static void xics_debugfs_init(struct kvmppc_xics *xics) | ||
878 | { | ||
879 | char *name; | ||
880 | |||
881 | name = kasprintf(GFP_KERNEL, "kvm-xics-%p", xics); | ||
882 | if (!name) { | ||
883 | pr_err("%s: no memory for name\n", __func__); | ||
884 | return; | ||
885 | } | ||
886 | |||
887 | xics->dentry = debugfs_create_file(name, S_IRUGO, powerpc_debugfs_root, | ||
888 | xics, &xics_debug_fops); | ||
889 | |||
890 | pr_debug("%s: created %s\n", __func__, name); | ||
891 | kfree(name); | ||
892 | } | ||
893 | |||
894 | struct kvmppc_ics *kvmppc_xics_create_ics(struct kvm *kvm, | ||
895 | struct kvmppc_xics *xics, int irq) | ||
896 | { | ||
897 | struct kvmppc_ics *ics; | ||
898 | int i, icsid; | ||
899 | |||
900 | icsid = irq >> KVMPPC_XICS_ICS_SHIFT; | ||
901 | |||
902 | mutex_lock(&kvm->lock); | ||
903 | |||
904 | /* ICS already exists - somebody else got here first */ | ||
905 | if (xics->ics[icsid]) | ||
906 | goto out; | ||
907 | |||
908 | /* Create the ICS */ | ||
909 | ics = kzalloc(sizeof(struct kvmppc_ics), GFP_KERNEL); | ||
910 | if (!ics) | ||
911 | goto out; | ||
912 | |||
913 | mutex_init(&ics->lock); | ||
914 | ics->icsid = icsid; | ||
915 | |||
916 | for (i = 0; i < KVMPPC_XICS_IRQ_PER_ICS; i++) { | ||
917 | ics->irq_state[i].number = (icsid << KVMPPC_XICS_ICS_SHIFT) | i; | ||
918 | ics->irq_state[i].priority = MASKED; | ||
919 | ics->irq_state[i].saved_priority = MASKED; | ||
920 | } | ||
921 | smp_wmb(); | ||
922 | xics->ics[icsid] = ics; | ||
923 | |||
924 | if (icsid > xics->max_icsid) | ||
925 | xics->max_icsid = icsid; | ||
926 | |||
927 | out: | ||
928 | mutex_unlock(&kvm->lock); | ||
929 | return xics->ics[icsid]; | ||
930 | } | ||
931 | |||
932 | int kvmppc_xics_create_icp(struct kvm_vcpu *vcpu, unsigned long server_num) | ||
933 | { | ||
934 | struct kvmppc_icp *icp; | ||
935 | |||
936 | if (!vcpu->kvm->arch.xics) | ||
937 | return -ENODEV; | ||
938 | |||
939 | if (kvmppc_xics_find_server(vcpu->kvm, server_num)) | ||
940 | return -EEXIST; | ||
941 | |||
942 | icp = kzalloc(sizeof(struct kvmppc_icp), GFP_KERNEL); | ||
943 | if (!icp) | ||
944 | return -ENOMEM; | ||
945 | |||
946 | icp->vcpu = vcpu; | ||
947 | icp->server_num = server_num; | ||
948 | icp->state.mfrr = MASKED; | ||
949 | icp->state.pending_pri = MASKED; | ||
950 | vcpu->arch.icp = icp; | ||
951 | |||
952 | XICS_DBG("created server for vcpu %d\n", vcpu->vcpu_id); | ||
953 | |||
954 | return 0; | ||
955 | } | ||
956 | |||
957 | u64 kvmppc_xics_get_icp(struct kvm_vcpu *vcpu) | ||
958 | { | ||
959 | struct kvmppc_icp *icp = vcpu->arch.icp; | ||
960 | union kvmppc_icp_state state; | ||
961 | |||
962 | if (!icp) | ||
963 | return 0; | ||
964 | state = icp->state; | ||
965 | return ((u64)state.cppr << KVM_REG_PPC_ICP_CPPR_SHIFT) | | ||
966 | ((u64)state.xisr << KVM_REG_PPC_ICP_XISR_SHIFT) | | ||
967 | ((u64)state.mfrr << KVM_REG_PPC_ICP_MFRR_SHIFT) | | ||
968 | ((u64)state.pending_pri << KVM_REG_PPC_ICP_PPRI_SHIFT); | ||
969 | } | ||
970 | |||
971 | int kvmppc_xics_set_icp(struct kvm_vcpu *vcpu, u64 icpval) | ||
972 | { | ||
973 | struct kvmppc_icp *icp = vcpu->arch.icp; | ||
974 | struct kvmppc_xics *xics = vcpu->kvm->arch.xics; | ||
975 | union kvmppc_icp_state old_state, new_state; | ||
976 | struct kvmppc_ics *ics; | ||
977 | u8 cppr, mfrr, pending_pri; | ||
978 | u32 xisr; | ||
979 | u16 src; | ||
980 | bool resend; | ||
981 | |||
982 | if (!icp || !xics) | ||
983 | return -ENOENT; | ||
984 | |||
985 | cppr = icpval >> KVM_REG_PPC_ICP_CPPR_SHIFT; | ||
986 | xisr = (icpval >> KVM_REG_PPC_ICP_XISR_SHIFT) & | ||
987 | KVM_REG_PPC_ICP_XISR_MASK; | ||
988 | mfrr = icpval >> KVM_REG_PPC_ICP_MFRR_SHIFT; | ||
989 | pending_pri = icpval >> KVM_REG_PPC_ICP_PPRI_SHIFT; | ||
990 | |||
991 | /* Require the new state to be internally consistent */ | ||
992 | if (xisr == 0) { | ||
993 | if (pending_pri != 0xff) | ||
994 | return -EINVAL; | ||
995 | } else if (xisr == XICS_IPI) { | ||
996 | if (pending_pri != mfrr || pending_pri >= cppr) | ||
997 | return -EINVAL; | ||
998 | } else { | ||
999 | if (pending_pri >= mfrr || pending_pri >= cppr) | ||
1000 | return -EINVAL; | ||
1001 | ics = kvmppc_xics_find_ics(xics, xisr, &src); | ||
1002 | if (!ics) | ||
1003 | return -EINVAL; | ||
1004 | } | ||
1005 | |||
1006 | new_state.raw = 0; | ||
1007 | new_state.cppr = cppr; | ||
1008 | new_state.xisr = xisr; | ||
1009 | new_state.mfrr = mfrr; | ||
1010 | new_state.pending_pri = pending_pri; | ||
1011 | |||
1012 | /* | ||
1013 | * Deassert the CPU interrupt request. | ||
1014 | * icp_try_update will reassert it if necessary. | ||
1015 | */ | ||
1016 | kvmppc_book3s_dequeue_irqprio(icp->vcpu, | ||
1017 | BOOK3S_INTERRUPT_EXTERNAL_LEVEL); | ||
1018 | |||
1019 | /* | ||
1020 | * Note that if we displace an interrupt from old_state.xisr, | ||
1021 | * we don't mark it as rejected. We expect userspace to set | ||
1022 | * the state of the interrupt sources to be consistent with | ||
1023 | * the ICP states (either before or afterwards, which doesn't | ||
1024 | * matter). We do handle resends due to CPPR becoming less | ||
1025 | * favoured because that is necessary to end up with a | ||
1026 | * consistent state in the situation where userspace restores | ||
1027 | * the ICS states before the ICP states. | ||
1028 | */ | ||
1029 | do { | ||
1030 | old_state = ACCESS_ONCE(icp->state); | ||
1031 | |||
1032 | if (new_state.mfrr <= old_state.mfrr) { | ||
1033 | resend = false; | ||
1034 | new_state.need_resend = old_state.need_resend; | ||
1035 | } else { | ||
1036 | resend = old_state.need_resend; | ||
1037 | new_state.need_resend = 0; | ||
1038 | } | ||
1039 | } while (!icp_try_update(icp, old_state, new_state, false)); | ||
1040 | |||
1041 | if (resend) | ||
1042 | icp_check_resend(xics, icp); | ||
1043 | |||
1044 | return 0; | ||
1045 | } | ||
1046 | |||
1047 | /* -- ioctls -- */ | ||
1048 | |||
1049 | int kvm_vm_ioctl_xics_irq(struct kvm *kvm, struct kvm_irq_level *args) | ||
1050 | { | ||
1051 | struct kvmppc_xics *xics; | ||
1052 | int r; | ||
1053 | |||
1054 | /* locking against multiple callers? */ | ||
1055 | |||
1056 | xics = kvm->arch.xics; | ||
1057 | if (!xics) | ||
1058 | return -ENODEV; | ||
1059 | |||
1060 | switch (args->level) { | ||
1061 | case KVM_INTERRUPT_SET: | ||
1062 | case KVM_INTERRUPT_SET_LEVEL: | ||
1063 | case KVM_INTERRUPT_UNSET: | ||
1064 | r = ics_deliver_irq(xics, args->irq, args->level); | ||
1065 | break; | ||
1066 | default: | ||
1067 | r = -EINVAL; | ||
1068 | } | ||
1069 | |||
1070 | return r; | ||
1071 | } | ||
1072 | |||
1073 | void kvmppc_xics_free(struct kvmppc_xics *xics) | ||
1074 | { | ||
1075 | int i; | ||
1076 | struct kvm *kvm = xics->kvm; | ||
1077 | |||
1078 | debugfs_remove(xics->dentry); | ||
1079 | |||
1080 | if (kvm) | ||
1081 | kvm->arch.xics = NULL; | ||
1082 | |||
1083 | for (i = 0; i <= xics->max_icsid; i++) | ||
1084 | kfree(xics->ics[i]); | ||
1085 | kfree(xics); | ||
1086 | } | ||
1087 | |||
1088 | int kvm_xics_create(struct kvm *kvm, u32 type) | ||
1089 | { | ||
1090 | struct kvmppc_xics *xics; | ||
1091 | int ret = 0; | ||
1092 | |||
1093 | xics = kzalloc(sizeof(*xics), GFP_KERNEL); | ||
1094 | if (!xics) | ||
1095 | return -ENOMEM; | ||
1096 | |||
1097 | xics->kvm = kvm; | ||
1098 | |||
1099 | /* Already there ? */ | ||
1100 | mutex_lock(&kvm->lock); | ||
1101 | if (kvm->arch.xics) | ||
1102 | ret = -EEXIST; | ||
1103 | else | ||
1104 | kvm->arch.xics = xics; | ||
1105 | mutex_unlock(&kvm->lock); | ||
1106 | |||
1107 | if (ret) | ||
1108 | return ret; | ||
1109 | |||
1110 | xics_debugfs_init(xics); | ||
1111 | |||
1112 | #ifdef CONFIG_KVM_BOOK3S_64_HV | ||
1113 | if (cpu_has_feature(CPU_FTR_ARCH_206)) { | ||
1114 | /* Enable real mode support */ | ||
1115 | xics->real_mode = ENABLE_REALMODE; | ||
1116 | xics->real_mode_dbg = DEBUG_REALMODE; | ||
1117 | } | ||
1118 | #endif /* CONFIG_KVM_BOOK3S_64_HV */ | ||
1119 | |||
1120 | return 0; | ||
1121 | } | ||
1122 | |||
1123 | void kvmppc_xics_free_icp(struct kvm_vcpu *vcpu) | ||
1124 | { | ||
1125 | if (!vcpu->arch.icp) | ||
1126 | return; | ||
1127 | kfree(vcpu->arch.icp); | ||
1128 | vcpu->arch.icp = NULL; | ||
1129 | vcpu->arch.irq_type = KVMPPC_IRQ_DEFAULT; | ||
1130 | } | ||