diff options
author | Glauber de Oliveira Costa <gcosta@redhat.com> | 2008-01-07 08:05:33 -0500 |
---|---|---|
committer | Rusty Russell <rusty@rustcorp.com.au> | 2008-01-30 06:50:11 -0500 |
commit | fc708b3e407dfd2e12ba9a6cf35bd0bffad1796d (patch) | |
tree | e9a6df9c9b8cf4077c98198c3f5d3bc6dc991c0f | |
parent | a53a35a8b485b9c16b73e5177bddaa4321971199 (diff) |
lguest: replace lguest_arch with lg_cpu_arch.
The fields found in lguest_arch are not really per-guest,
but per-cpu (gdt, idt, etc). So this patch turns lguest_arch
into lg_cpu_arch.
It makes sense to have a per-guest per-arch struct, but this
can be addressed later, when the need arrives.
Signed-off-by: Glauber de Oliveira Costa <gcosta@redhat.com>
Signed-off-by: Rusty Russell <rusty@rustcorp.com.au>
-rw-r--r-- | drivers/lguest/interrupts_and_traps.c | 28 | ||||
-rw-r--r-- | drivers/lguest/lg.h | 19 | ||||
-rw-r--r-- | drivers/lguest/segments.c | 42 | ||||
-rw-r--r-- | drivers/lguest/x86/core.c | 24 | ||||
-rw-r--r-- | include/asm-x86/lguest.h | 2 |
5 files changed, 58 insertions, 57 deletions
diff --git a/drivers/lguest/interrupts_and_traps.c b/drivers/lguest/interrupts_and_traps.c index 468faf8233d6..306b93c71dcc 100644 --- a/drivers/lguest/interrupts_and_traps.c +++ b/drivers/lguest/interrupts_and_traps.c | |||
@@ -178,7 +178,7 @@ void maybe_do_interrupt(struct lg_cpu *cpu) | |||
178 | /* Look at the IDT entry the Guest gave us for this interrupt. The | 178 | /* Look at the IDT entry the Guest gave us for this interrupt. The |
179 | * first 32 (FIRST_EXTERNAL_VECTOR) entries are for traps, so we skip | 179 | * first 32 (FIRST_EXTERNAL_VECTOR) entries are for traps, so we skip |
180 | * over them. */ | 180 | * over them. */ |
181 | idt = &lg->arch.idt[FIRST_EXTERNAL_VECTOR+irq]; | 181 | idt = &cpu->arch.idt[FIRST_EXTERNAL_VECTOR+irq]; |
182 | /* If they don't have a handler (yet?), we just ignore it */ | 182 | /* If they don't have a handler (yet?), we just ignore it */ |
183 | if (idt_present(idt->a, idt->b)) { | 183 | if (idt_present(idt->a, idt->b)) { |
184 | /* OK, mark it no longer pending and deliver it. */ | 184 | /* OK, mark it no longer pending and deliver it. */ |
@@ -251,15 +251,15 @@ int deliver_trap(struct lg_cpu *cpu, unsigned int num) | |||
251 | { | 251 | { |
252 | /* Trap numbers are always 8 bit, but we set an impossible trap number | 252 | /* Trap numbers are always 8 bit, but we set an impossible trap number |
253 | * for traps inside the Switcher, so check that here. */ | 253 | * for traps inside the Switcher, so check that here. */ |
254 | if (num >= ARRAY_SIZE(cpu->lg->arch.idt)) | 254 | if (num >= ARRAY_SIZE(cpu->arch.idt)) |
255 | return 0; | 255 | return 0; |
256 | 256 | ||
257 | /* Early on the Guest hasn't set the IDT entries (or maybe it put a | 257 | /* Early on the Guest hasn't set the IDT entries (or maybe it put a |
258 | * bogus one in): if we fail here, the Guest will be killed. */ | 258 | * bogus one in): if we fail here, the Guest will be killed. */ |
259 | if (!idt_present(cpu->lg->arch.idt[num].a, cpu->lg->arch.idt[num].b)) | 259 | if (!idt_present(cpu->arch.idt[num].a, cpu->arch.idt[num].b)) |
260 | return 0; | 260 | return 0; |
261 | set_guest_interrupt(cpu, cpu->lg->arch.idt[num].a, | 261 | set_guest_interrupt(cpu, cpu->arch.idt[num].a, |
262 | cpu->lg->arch.idt[num].b, has_err(num)); | 262 | cpu->arch.idt[num].b, has_err(num)); |
263 | return 1; | 263 | return 1; |
264 | } | 264 | } |
265 | 265 | ||
@@ -385,7 +385,7 @@ static void set_trap(struct lguest *lg, struct desc_struct *trap, | |||
385 | * | 385 | * |
386 | * We saw the Guest setting Interrupt Descriptor Table (IDT) entries with the | 386 | * We saw the Guest setting Interrupt Descriptor Table (IDT) entries with the |
387 | * LHCALL_LOAD_IDT_ENTRY hypercall before: that comes here. */ | 387 | * LHCALL_LOAD_IDT_ENTRY hypercall before: that comes here. */ |
388 | void load_guest_idt_entry(struct lguest *lg, unsigned int num, u32 lo, u32 hi) | 388 | void load_guest_idt_entry(struct lg_cpu *cpu, unsigned int num, u32 lo, u32 hi) |
389 | { | 389 | { |
390 | /* Guest never handles: NMI, doublefault, spurious interrupt or | 390 | /* Guest never handles: NMI, doublefault, spurious interrupt or |
391 | * hypercall. We ignore when it tries to set them. */ | 391 | * hypercall. We ignore when it tries to set them. */ |
@@ -394,13 +394,13 @@ void load_guest_idt_entry(struct lguest *lg, unsigned int num, u32 lo, u32 hi) | |||
394 | 394 | ||
395 | /* Mark the IDT as changed: next time the Guest runs we'll know we have | 395 | /* Mark the IDT as changed: next time the Guest runs we'll know we have |
396 | * to copy this again. */ | 396 | * to copy this again. */ |
397 | lg->changed |= CHANGED_IDT; | 397 | cpu->lg->changed |= CHANGED_IDT; |
398 | 398 | ||
399 | /* Check that the Guest doesn't try to step outside the bounds. */ | 399 | /* Check that the Guest doesn't try to step outside the bounds. */ |
400 | if (num >= ARRAY_SIZE(lg->arch.idt)) | 400 | if (num >= ARRAY_SIZE(cpu->arch.idt)) |
401 | kill_guest(lg, "Setting idt entry %u", num); | 401 | kill_guest(cpu->lg, "Setting idt entry %u", num); |
402 | else | 402 | else |
403 | set_trap(lg, &lg->arch.idt[num], num, lo, hi); | 403 | set_trap(cpu->lg, &cpu->arch.idt[num], num, lo, hi); |
404 | } | 404 | } |
405 | 405 | ||
406 | /* The default entry for each interrupt points into the Switcher routines which | 406 | /* The default entry for each interrupt points into the Switcher routines which |
@@ -436,14 +436,14 @@ void setup_default_idt_entries(struct lguest_ro_state *state, | |||
436 | /*H:240 We don't use the IDT entries in the "struct lguest" directly, instead | 436 | /*H:240 We don't use the IDT entries in the "struct lguest" directly, instead |
437 | * we copy them into the IDT which we've set up for Guests on this CPU, just | 437 | * we copy them into the IDT which we've set up for Guests on this CPU, just |
438 | * before we run the Guest. This routine does that copy. */ | 438 | * before we run the Guest. This routine does that copy. */ |
439 | void copy_traps(const struct lguest *lg, struct desc_struct *idt, | 439 | void copy_traps(const struct lg_cpu *cpu, struct desc_struct *idt, |
440 | const unsigned long *def) | 440 | const unsigned long *def) |
441 | { | 441 | { |
442 | unsigned int i; | 442 | unsigned int i; |
443 | 443 | ||
444 | /* We can simply copy the direct traps, otherwise we use the default | 444 | /* We can simply copy the direct traps, otherwise we use the default |
445 | * ones in the Switcher: they will return to the Host. */ | 445 | * ones in the Switcher: they will return to the Host. */ |
446 | for (i = 0; i < ARRAY_SIZE(lg->arch.idt); i++) { | 446 | for (i = 0; i < ARRAY_SIZE(cpu->arch.idt); i++) { |
447 | /* If no Guest can ever override this trap, leave it alone. */ | 447 | /* If no Guest can ever override this trap, leave it alone. */ |
448 | if (!direct_trap(i)) | 448 | if (!direct_trap(i)) |
449 | continue; | 449 | continue; |
@@ -452,8 +452,8 @@ void copy_traps(const struct lguest *lg, struct desc_struct *idt, | |||
452 | * Interrupt gates (type 14) disable interrupts as they are | 452 | * Interrupt gates (type 14) disable interrupts as they are |
453 | * entered, which we never let the Guest do. Not present | 453 | * entered, which we never let the Guest do. Not present |
454 | * entries (type 0x0) also can't go direct, of course. */ | 454 | * entries (type 0x0) also can't go direct, of course. */ |
455 | if (idt_type(lg->arch.idt[i].a, lg->arch.idt[i].b) == 0xF) | 455 | if (idt_type(cpu->arch.idt[i].a, cpu->arch.idt[i].b) == 0xF) |
456 | idt[i] = lg->arch.idt[i]; | 456 | idt[i] = cpu->arch.idt[i]; |
457 | else | 457 | else |
458 | /* Reset it to the default. */ | 458 | /* Reset it to the default. */ |
459 | default_idt_entry(&idt[i], i, def[i]); | 459 | default_idt_entry(&idt[i], i, def[i]); |
diff --git a/drivers/lguest/lg.h b/drivers/lguest/lg.h index 35b331230c55..d08b85342b92 100644 --- a/drivers/lguest/lg.h +++ b/drivers/lguest/lg.h | |||
@@ -57,6 +57,8 @@ struct lg_cpu { | |||
57 | 57 | ||
58 | /* Pending virtual interrupts */ | 58 | /* Pending virtual interrupts */ |
59 | DECLARE_BITMAP(irqs_pending, LGUEST_IRQS); | 59 | DECLARE_BITMAP(irqs_pending, LGUEST_IRQS); |
60 | |||
61 | struct lg_cpu_arch arch; | ||
60 | }; | 62 | }; |
61 | 63 | ||
62 | /* The private info the thread maintains about the guest. */ | 64 | /* The private info the thread maintains about the guest. */ |
@@ -99,8 +101,6 @@ struct lguest | |||
99 | 101 | ||
100 | /* Dead? */ | 102 | /* Dead? */ |
101 | const char *dead; | 103 | const char *dead; |
102 | |||
103 | struct lguest_arch arch; | ||
104 | }; | 104 | }; |
105 | 105 | ||
106 | extern struct mutex lguest_lock; | 106 | extern struct mutex lguest_lock; |
@@ -139,12 +139,13 @@ int run_guest(struct lg_cpu *cpu, unsigned long __user *user); | |||
139 | /* interrupts_and_traps.c: */ | 139 | /* interrupts_and_traps.c: */ |
140 | void maybe_do_interrupt(struct lg_cpu *cpu); | 140 | void maybe_do_interrupt(struct lg_cpu *cpu); |
141 | int deliver_trap(struct lg_cpu *cpu, unsigned int num); | 141 | int deliver_trap(struct lg_cpu *cpu, unsigned int num); |
142 | void load_guest_idt_entry(struct lguest *lg, unsigned int i, u32 low, u32 hi); | 142 | void load_guest_idt_entry(struct lg_cpu *cpu, unsigned int i, |
143 | u32 low, u32 hi); | ||
143 | void guest_set_stack(struct lguest *lg, u32 seg, u32 esp, unsigned int pages); | 144 | void guest_set_stack(struct lguest *lg, u32 seg, u32 esp, unsigned int pages); |
144 | void pin_stack_pages(struct lguest *lg); | 145 | void pin_stack_pages(struct lguest *lg); |
145 | void setup_default_idt_entries(struct lguest_ro_state *state, | 146 | void setup_default_idt_entries(struct lguest_ro_state *state, |
146 | const unsigned long *def); | 147 | const unsigned long *def); |
147 | void copy_traps(const struct lguest *lg, struct desc_struct *idt, | 148 | void copy_traps(const struct lg_cpu *cpu, struct desc_struct *idt, |
148 | const unsigned long *def); | 149 | const unsigned long *def); |
149 | void guest_set_clockevent(struct lg_cpu *cpu, unsigned long delta); | 150 | void guest_set_clockevent(struct lg_cpu *cpu, unsigned long delta); |
150 | void init_clockdev(struct lg_cpu *cpu); | 151 | void init_clockdev(struct lg_cpu *cpu); |
@@ -154,11 +155,11 @@ void free_interrupts(void); | |||
154 | 155 | ||
155 | /* segments.c: */ | 156 | /* segments.c: */ |
156 | void setup_default_gdt_entries(struct lguest_ro_state *state); | 157 | void setup_default_gdt_entries(struct lguest_ro_state *state); |
157 | void setup_guest_gdt(struct lguest *lg); | 158 | void setup_guest_gdt(struct lg_cpu *cpu); |
158 | void load_guest_gdt(struct lguest *lg, unsigned long table, u32 num); | 159 | void load_guest_gdt(struct lg_cpu *cpu, unsigned long table, u32 num); |
159 | void guest_load_tls(struct lguest *lg, unsigned long tls_array); | 160 | void guest_load_tls(struct lg_cpu *cpu, unsigned long tls_array); |
160 | void copy_gdt(const struct lguest *lg, struct desc_struct *gdt); | 161 | void copy_gdt(const struct lg_cpu *cpu, struct desc_struct *gdt); |
161 | void copy_gdt_tls(const struct lguest *lg, struct desc_struct *gdt); | 162 | void copy_gdt_tls(const struct lg_cpu *cpu, struct desc_struct *gdt); |
162 | 163 | ||
163 | /* page_tables.c: */ | 164 | /* page_tables.c: */ |
164 | int init_guest_pagetable(struct lguest *lg, unsigned long pgtable); | 165 | int init_guest_pagetable(struct lguest *lg, unsigned long pgtable); |
diff --git a/drivers/lguest/segments.c b/drivers/lguest/segments.c index 9e189cbec7dd..02138450ecf5 100644 --- a/drivers/lguest/segments.c +++ b/drivers/lguest/segments.c | |||
@@ -58,7 +58,7 @@ static int ignored_gdt(unsigned int num) | |||
58 | * Protection Fault in the Switcher when it restores a Guest segment register | 58 | * Protection Fault in the Switcher when it restores a Guest segment register |
59 | * which tries to use that entry. Then we kill the Guest for causing such a | 59 | * which tries to use that entry. Then we kill the Guest for causing such a |
60 | * mess: the message will be "unhandled trap 256". */ | 60 | * mess: the message will be "unhandled trap 256". */ |
61 | static void fixup_gdt_table(struct lguest *lg, unsigned start, unsigned end) | 61 | static void fixup_gdt_table(struct lg_cpu *cpu, unsigned start, unsigned end) |
62 | { | 62 | { |
63 | unsigned int i; | 63 | unsigned int i; |
64 | 64 | ||
@@ -71,14 +71,14 @@ static void fixup_gdt_table(struct lguest *lg, unsigned start, unsigned end) | |||
71 | /* Segment descriptors contain a privilege level: the Guest is | 71 | /* Segment descriptors contain a privilege level: the Guest is |
72 | * sometimes careless and leaves this as 0, even though it's | 72 | * sometimes careless and leaves this as 0, even though it's |
73 | * running at privilege level 1. If so, we fix it here. */ | 73 | * running at privilege level 1. If so, we fix it here. */ |
74 | if ((lg->arch.gdt[i].b & 0x00006000) == 0) | 74 | if ((cpu->arch.gdt[i].b & 0x00006000) == 0) |
75 | lg->arch.gdt[i].b |= (GUEST_PL << 13); | 75 | cpu->arch.gdt[i].b |= (GUEST_PL << 13); |
76 | 76 | ||
77 | /* Each descriptor has an "accessed" bit. If we don't set it | 77 | /* Each descriptor has an "accessed" bit. If we don't set it |
78 | * now, the CPU will try to set it when the Guest first loads | 78 | * now, the CPU will try to set it when the Guest first loads |
79 | * that entry into a segment register. But the GDT isn't | 79 | * that entry into a segment register. But the GDT isn't |
80 | * writable by the Guest, so bad things can happen. */ | 80 | * writable by the Guest, so bad things can happen. */ |
81 | lg->arch.gdt[i].b |= 0x00000100; | 81 | cpu->arch.gdt[i].b |= 0x00000100; |
82 | } | 82 | } |
83 | } | 83 | } |
84 | 84 | ||
@@ -109,31 +109,31 @@ void setup_default_gdt_entries(struct lguest_ro_state *state) | |||
109 | 109 | ||
110 | /* This routine sets up the initial Guest GDT for booting. All entries start | 110 | /* This routine sets up the initial Guest GDT for booting. All entries start |
111 | * as 0 (unusable). */ | 111 | * as 0 (unusable). */ |
112 | void setup_guest_gdt(struct lguest *lg) | 112 | void setup_guest_gdt(struct lg_cpu *cpu) |
113 | { | 113 | { |
114 | /* Start with full 0-4G segments... */ | 114 | /* Start with full 0-4G segments... */ |
115 | lg->arch.gdt[GDT_ENTRY_KERNEL_CS] = FULL_EXEC_SEGMENT; | 115 | cpu->arch.gdt[GDT_ENTRY_KERNEL_CS] = FULL_EXEC_SEGMENT; |
116 | lg->arch.gdt[GDT_ENTRY_KERNEL_DS] = FULL_SEGMENT; | 116 | cpu->arch.gdt[GDT_ENTRY_KERNEL_DS] = FULL_SEGMENT; |
117 | /* ...except the Guest is allowed to use them, so set the privilege | 117 | /* ...except the Guest is allowed to use them, so set the privilege |
118 | * level appropriately in the flags. */ | 118 | * level appropriately in the flags. */ |
119 | lg->arch.gdt[GDT_ENTRY_KERNEL_CS].b |= (GUEST_PL << 13); | 119 | cpu->arch.gdt[GDT_ENTRY_KERNEL_CS].b |= (GUEST_PL << 13); |
120 | lg->arch.gdt[GDT_ENTRY_KERNEL_DS].b |= (GUEST_PL << 13); | 120 | cpu->arch.gdt[GDT_ENTRY_KERNEL_DS].b |= (GUEST_PL << 13); |
121 | } | 121 | } |
122 | 122 | ||
123 | /*H:650 An optimization of copy_gdt(), for just the three "thead-local storage" | 123 | /*H:650 An optimization of copy_gdt(), for just the three "thead-local storage" |
124 | * entries. */ | 124 | * entries. */ |
125 | void copy_gdt_tls(const struct lguest *lg, struct desc_struct *gdt) | 125 | void copy_gdt_tls(const struct lg_cpu *cpu, struct desc_struct *gdt) |
126 | { | 126 | { |
127 | unsigned int i; | 127 | unsigned int i; |
128 | 128 | ||
129 | for (i = GDT_ENTRY_TLS_MIN; i <= GDT_ENTRY_TLS_MAX; i++) | 129 | for (i = GDT_ENTRY_TLS_MIN; i <= GDT_ENTRY_TLS_MAX; i++) |
130 | gdt[i] = lg->arch.gdt[i]; | 130 | gdt[i] = cpu->arch.gdt[i]; |
131 | } | 131 | } |
132 | 132 | ||
133 | /*H:640 When the Guest is run on a different CPU, or the GDT entries have | 133 | /*H:640 When the Guest is run on a different CPU, or the GDT entries have |
134 | * changed, copy_gdt() is called to copy the Guest's GDT entries across to this | 134 | * changed, copy_gdt() is called to copy the Guest's GDT entries across to this |
135 | * CPU's GDT. */ | 135 | * CPU's GDT. */ |
136 | void copy_gdt(const struct lguest *lg, struct desc_struct *gdt) | 136 | void copy_gdt(const struct lg_cpu *cpu, struct desc_struct *gdt) |
137 | { | 137 | { |
138 | unsigned int i; | 138 | unsigned int i; |
139 | 139 | ||
@@ -141,21 +141,22 @@ void copy_gdt(const struct lguest *lg, struct desc_struct *gdt) | |||
141 | * replaced. See ignored_gdt() above. */ | 141 | * replaced. See ignored_gdt() above. */ |
142 | for (i = 0; i < GDT_ENTRIES; i++) | 142 | for (i = 0; i < GDT_ENTRIES; i++) |
143 | if (!ignored_gdt(i)) | 143 | if (!ignored_gdt(i)) |
144 | gdt[i] = lg->arch.gdt[i]; | 144 | gdt[i] = cpu->arch.gdt[i]; |
145 | } | 145 | } |
146 | 146 | ||
147 | /*H:620 This is where the Guest asks us to load a new GDT (LHCALL_LOAD_GDT). | 147 | /*H:620 This is where the Guest asks us to load a new GDT (LHCALL_LOAD_GDT). |
148 | * We copy it from the Guest and tweak the entries. */ | 148 | * We copy it from the Guest and tweak the entries. */ |
149 | void load_guest_gdt(struct lguest *lg, unsigned long table, u32 num) | 149 | void load_guest_gdt(struct lg_cpu *cpu, unsigned long table, u32 num) |
150 | { | 150 | { |
151 | struct lguest *lg = cpu->lg; | ||
151 | /* We assume the Guest has the same number of GDT entries as the | 152 | /* We assume the Guest has the same number of GDT entries as the |
152 | * Host, otherwise we'd have to dynamically allocate the Guest GDT. */ | 153 | * Host, otherwise we'd have to dynamically allocate the Guest GDT. */ |
153 | if (num > ARRAY_SIZE(lg->arch.gdt)) | 154 | if (num > ARRAY_SIZE(cpu->arch.gdt)) |
154 | kill_guest(lg, "too many gdt entries %i", num); | 155 | kill_guest(lg, "too many gdt entries %i", num); |
155 | 156 | ||
156 | /* We read the whole thing in, then fix it up. */ | 157 | /* We read the whole thing in, then fix it up. */ |
157 | __lgread(lg, lg->arch.gdt, table, num * sizeof(lg->arch.gdt[0])); | 158 | __lgread(lg, cpu->arch.gdt, table, num * sizeof(cpu->arch.gdt[0])); |
158 | fixup_gdt_table(lg, 0, ARRAY_SIZE(lg->arch.gdt)); | 159 | fixup_gdt_table(cpu, 0, ARRAY_SIZE(cpu->arch.gdt)); |
159 | /* Mark that the GDT changed so the core knows it has to copy it again, | 160 | /* Mark that the GDT changed so the core knows it has to copy it again, |
160 | * even if the Guest is run on the same CPU. */ | 161 | * even if the Guest is run on the same CPU. */ |
161 | lg->changed |= CHANGED_GDT; | 162 | lg->changed |= CHANGED_GDT; |
@@ -165,12 +166,13 @@ void load_guest_gdt(struct lguest *lg, unsigned long table, u32 num) | |||
165 | * Remember that this happens on every context switch, so it's worth | 166 | * Remember that this happens on every context switch, so it's worth |
166 | * optimizing. But wouldn't it be neater to have a single hypercall to cover | 167 | * optimizing. But wouldn't it be neater to have a single hypercall to cover |
167 | * both cases? */ | 168 | * both cases? */ |
168 | void guest_load_tls(struct lguest *lg, unsigned long gtls) | 169 | void guest_load_tls(struct lg_cpu *cpu, unsigned long gtls) |
169 | { | 170 | { |
170 | struct desc_struct *tls = &lg->arch.gdt[GDT_ENTRY_TLS_MIN]; | 171 | struct desc_struct *tls = &cpu->arch.gdt[GDT_ENTRY_TLS_MIN]; |
172 | struct lguest *lg = cpu->lg; | ||
171 | 173 | ||
172 | __lgread(lg, tls, gtls, sizeof(*tls)*GDT_ENTRY_TLS_ENTRIES); | 174 | __lgread(lg, tls, gtls, sizeof(*tls)*GDT_ENTRY_TLS_ENTRIES); |
173 | fixup_gdt_table(lg, GDT_ENTRY_TLS_MIN, GDT_ENTRY_TLS_MAX+1); | 175 | fixup_gdt_table(cpu, GDT_ENTRY_TLS_MIN, GDT_ENTRY_TLS_MAX+1); |
174 | /* Note that just the TLS entries have changed. */ | 176 | /* Note that just the TLS entries have changed. */ |
175 | lg->changed |= CHANGED_GDT_TLS; | 177 | lg->changed |= CHANGED_GDT_TLS; |
176 | } | 178 | } |
diff --git a/drivers/lguest/x86/core.c b/drivers/lguest/x86/core.c index d96a93d95aea..e989b8358864 100644 --- a/drivers/lguest/x86/core.c +++ b/drivers/lguest/x86/core.c | |||
@@ -100,14 +100,14 @@ static void copy_in_guest_info(struct lg_cpu *cpu, struct lguest_pages *pages) | |||
100 | 100 | ||
101 | /* Copy direct-to-Guest trap entries. */ | 101 | /* Copy direct-to-Guest trap entries. */ |
102 | if (lg->changed & CHANGED_IDT) | 102 | if (lg->changed & CHANGED_IDT) |
103 | copy_traps(lg, pages->state.guest_idt, default_idt_entries); | 103 | copy_traps(cpu, pages->state.guest_idt, default_idt_entries); |
104 | 104 | ||
105 | /* Copy all GDT entries which the Guest can change. */ | 105 | /* Copy all GDT entries which the Guest can change. */ |
106 | if (lg->changed & CHANGED_GDT) | 106 | if (lg->changed & CHANGED_GDT) |
107 | copy_gdt(lg, pages->state.guest_gdt); | 107 | copy_gdt(cpu, pages->state.guest_gdt); |
108 | /* If only the TLS entries have changed, copy them. */ | 108 | /* If only the TLS entries have changed, copy them. */ |
109 | else if (lg->changed & CHANGED_GDT_TLS) | 109 | else if (lg->changed & CHANGED_GDT_TLS) |
110 | copy_gdt_tls(lg, pages->state.guest_gdt); | 110 | copy_gdt_tls(cpu, pages->state.guest_gdt); |
111 | 111 | ||
112 | /* Mark the Guest as unchanged for next time. */ | 112 | /* Mark the Guest as unchanged for next time. */ |
113 | lg->changed = 0; | 113 | lg->changed = 0; |
@@ -196,7 +196,7 @@ void lguest_arch_run_guest(struct lg_cpu *cpu) | |||
196 | * re-enable interrupts an interrupt could fault and thus overwrite | 196 | * re-enable interrupts an interrupt could fault and thus overwrite |
197 | * cr2, or we could even move off to a different CPU. */ | 197 | * cr2, or we could even move off to a different CPU. */ |
198 | if (cpu->regs->trapnum == 14) | 198 | if (cpu->regs->trapnum == 14) |
199 | lg->arch.last_pagefault = read_cr2(); | 199 | cpu->arch.last_pagefault = read_cr2(); |
200 | /* Similarly, if we took a trap because the Guest used the FPU, | 200 | /* Similarly, if we took a trap because the Guest used the FPU, |
201 | * we have to restore the FPU it expects to see. */ | 201 | * we have to restore the FPU it expects to see. */ |
202 | else if (cpu->regs->trapnum == 7) | 202 | else if (cpu->regs->trapnum == 7) |
@@ -307,7 +307,7 @@ void lguest_arch_handle_trap(struct lg_cpu *cpu) | |||
307 | * | 307 | * |
308 | * The errcode tells whether this was a read or a write, and | 308 | * The errcode tells whether this was a read or a write, and |
309 | * whether kernel or userspace code. */ | 309 | * whether kernel or userspace code. */ |
310 | if (demand_page(lg, lg->arch.last_pagefault, cpu->regs->errcode)) | 310 | if (demand_page(lg,cpu->arch.last_pagefault,cpu->regs->errcode)) |
311 | return; | 311 | return; |
312 | 312 | ||
313 | /* OK, it's really not there (or not OK): the Guest needs to | 313 | /* OK, it's really not there (or not OK): the Guest needs to |
@@ -318,7 +318,7 @@ void lguest_arch_handle_trap(struct lg_cpu *cpu) | |||
318 | * happen before it's done the LHCALL_LGUEST_INIT hypercall, so | 318 | * happen before it's done the LHCALL_LGUEST_INIT hypercall, so |
319 | * lg->lguest_data could be NULL */ | 319 | * lg->lguest_data could be NULL */ |
320 | if (lg->lguest_data && | 320 | if (lg->lguest_data && |
321 | put_user(lg->arch.last_pagefault, &lg->lguest_data->cr2)) | 321 | put_user(cpu->arch.last_pagefault, &lg->lguest_data->cr2)) |
322 | kill_guest(lg, "Writing cr2"); | 322 | kill_guest(lg, "Writing cr2"); |
323 | break; | 323 | break; |
324 | case 7: /* We've intercepted a Device Not Available fault. */ | 324 | case 7: /* We've intercepted a Device Not Available fault. */ |
@@ -349,7 +349,7 @@ void lguest_arch_handle_trap(struct lg_cpu *cpu) | |||
349 | * it handle), it dies with a cryptic error message. */ | 349 | * it handle), it dies with a cryptic error message. */ |
350 | kill_guest(lg, "unhandled trap %li at %#lx (%#lx)", | 350 | kill_guest(lg, "unhandled trap %li at %#lx (%#lx)", |
351 | cpu->regs->trapnum, cpu->regs->eip, | 351 | cpu->regs->trapnum, cpu->regs->eip, |
352 | cpu->regs->trapnum == 14 ? lg->arch.last_pagefault | 352 | cpu->regs->trapnum == 14 ? cpu->arch.last_pagefault |
353 | : cpu->regs->errcode); | 353 | : cpu->regs->errcode); |
354 | } | 354 | } |
355 | 355 | ||
@@ -495,17 +495,15 @@ void __exit lguest_arch_host_fini(void) | |||
495 | /*H:122 The i386-specific hypercalls simply farm out to the right functions. */ | 495 | /*H:122 The i386-specific hypercalls simply farm out to the right functions. */ |
496 | int lguest_arch_do_hcall(struct lg_cpu *cpu, struct hcall_args *args) | 496 | int lguest_arch_do_hcall(struct lg_cpu *cpu, struct hcall_args *args) |
497 | { | 497 | { |
498 | struct lguest *lg = cpu->lg; | ||
499 | |||
500 | switch (args->arg0) { | 498 | switch (args->arg0) { |
501 | case LHCALL_LOAD_GDT: | 499 | case LHCALL_LOAD_GDT: |
502 | load_guest_gdt(lg, args->arg1, args->arg2); | 500 | load_guest_gdt(cpu, args->arg1, args->arg2); |
503 | break; | 501 | break; |
504 | case LHCALL_LOAD_IDT_ENTRY: | 502 | case LHCALL_LOAD_IDT_ENTRY: |
505 | load_guest_idt_entry(lg, args->arg1, args->arg2, args->arg3); | 503 | load_guest_idt_entry(cpu, args->arg1, args->arg2, args->arg3); |
506 | break; | 504 | break; |
507 | case LHCALL_LOAD_TLS: | 505 | case LHCALL_LOAD_TLS: |
508 | guest_load_tls(lg, args->arg1); | 506 | guest_load_tls(cpu, args->arg1); |
509 | break; | 507 | break; |
510 | default: | 508 | default: |
511 | /* Bad Guest. Bad! */ | 509 | /* Bad Guest. Bad! */ |
@@ -586,5 +584,5 @@ void lguest_arch_setup_regs(struct lg_cpu *cpu, unsigned long start) | |||
586 | 584 | ||
587 | /* There are a couple of GDT entries the Guest expects when first | 585 | /* There are a couple of GDT entries the Guest expects when first |
588 | * booting. */ | 586 | * booting. */ |
589 | setup_guest_gdt(cpu->lg); | 587 | setup_guest_gdt(cpu); |
590 | } | 588 | } |
diff --git a/include/asm-x86/lguest.h b/include/asm-x86/lguest.h index ccd338460811..b9d003b8005e 100644 --- a/include/asm-x86/lguest.h +++ b/include/asm-x86/lguest.h | |||
@@ -56,7 +56,7 @@ struct lguest_ro_state | |||
56 | struct desc_struct guest_gdt[GDT_ENTRIES]; | 56 | struct desc_struct guest_gdt[GDT_ENTRIES]; |
57 | }; | 57 | }; |
58 | 58 | ||
59 | struct lguest_arch | 59 | struct lg_cpu_arch |
60 | { | 60 | { |
61 | /* The GDT entries copied into lguest_ro_state when running. */ | 61 | /* The GDT entries copied into lguest_ro_state when running. */ |
62 | struct desc_struct gdt[GDT_ENTRIES]; | 62 | struct desc_struct gdt[GDT_ENTRIES]; |