diff options
author | Linus Torvalds <torvalds@linux-foundation.org> | 2012-03-29 17:28:26 -0400 |
---|---|---|
committer | Linus Torvalds <torvalds@linux-foundation.org> | 2012-03-29 17:28:26 -0400 |
commit | 6b8212a313dae341ef3a2e413dfec5c4dea59617 (patch) | |
tree | bbca09d88f61f999c7714fe82710bdfe6ee0e98b /arch/x86/kernel | |
parent | bcd550745fc54f789c14e7526e0633222c505faa (diff) | |
parent | 8abc3122aa02567bfe626cd13f4d34853c9b1225 (diff) |
Merge branch 'x86-urgent-for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/tip/tip
Pull x86 updates from Ingo Molnar.
This touches some non-x86 files due to the sanitized INLINE_SPIN_UNLOCK
config usage.
Fixed up trivial conflicts due to just header include changes (removing
headers due to cpu_idle() merge clashing with the <asm/system.h> split).
* 'x86-urgent-for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/tip/tip:
x86/apic/amd: Be more verbose about LVT offset assignments
x86, tls: Off by one limit check
x86/ioapic: Add io_apic_ops driver layer to allow interception
x86/olpc: Add debugfs interface for EC commands
x86: Merge the x86_32 and x86_64 cpu_idle() functions
x86/kconfig: Remove CONFIG_TR=y from the defconfigs
x86: Stop recursive fault in print_context_stack after stack overflow
x86/io_apic: Move and reenable irq only when CONFIG_GENERIC_PENDING_IRQ=y
x86/apic: Add separate apic_id_valid() functions for selected apic drivers
locking/kconfig: Simplify INLINE_SPIN_UNLOCK usage
x86/kconfig: Update defconfigs
x86: Fix excessive MSR print out when show_msr is not specified
Diffstat (limited to 'arch/x86/kernel')
-rw-r--r-- | arch/x86/kernel/acpi/boot.c | 2 | ||||
-rw-r--r-- | arch/x86/kernel/apic/apic.c | 13 | ||||
-rw-r--r-- | arch/x86/kernel/apic/apic_numachip.c | 3 | ||||
-rw-r--r-- | arch/x86/kernel/apic/io_apic.c | 159 | ||||
-rw-r--r-- | arch/x86/kernel/apic/x2apic_cluster.c | 2 | ||||
-rw-r--r-- | arch/x86/kernel/apic/x2apic_phys.c | 2 | ||||
-rw-r--r-- | arch/x86/kernel/apic/x2apic_uv_x.c | 7 | ||||
-rw-r--r-- | arch/x86/kernel/cpu/common.c | 2 | ||||
-rw-r--r-- | arch/x86/kernel/dumpstack.c | 7 | ||||
-rw-r--r-- | arch/x86/kernel/process.c | 114 | ||||
-rw-r--r-- | arch/x86/kernel/process_32.c | 58 | ||||
-rw-r--r-- | arch/x86/kernel/process_64.c | 107 | ||||
-rw-r--r-- | arch/x86/kernel/tls.c | 4 |
13 files changed, 253 insertions, 227 deletions
diff --git a/arch/x86/kernel/acpi/boot.c b/arch/x86/kernel/acpi/boot.c index 406ed77216d..0f42c2f4431 100644 --- a/arch/x86/kernel/acpi/boot.c +++ b/arch/x86/kernel/acpi/boot.c | |||
@@ -239,7 +239,7 @@ acpi_parse_x2apic(struct acpi_subtable_header *header, const unsigned long end) | |||
239 | * to not preallocating memory for all NR_CPUS | 239 | * to not preallocating memory for all NR_CPUS |
240 | * when we use CPU hotplug. | 240 | * when we use CPU hotplug. |
241 | */ | 241 | */ |
242 | if (!cpu_has_x2apic && (apic_id >= 0xff) && enabled) | 242 | if (!apic->apic_id_valid(apic_id) && enabled) |
243 | printk(KERN_WARNING PREFIX "x2apic entry ignored\n"); | 243 | printk(KERN_WARNING PREFIX "x2apic entry ignored\n"); |
244 | else | 244 | else |
245 | acpi_register_lapic(apic_id, enabled); | 245 | acpi_register_lapic(apic_id, enabled); |
diff --git a/arch/x86/kernel/apic/apic.c b/arch/x86/kernel/apic/apic.c index 2eec05b6d1b..11544d8f1e9 100644 --- a/arch/x86/kernel/apic/apic.c +++ b/arch/x86/kernel/apic/apic.c | |||
@@ -383,20 +383,25 @@ static inline int eilvt_entry_is_changeable(unsigned int old, unsigned int new) | |||
383 | 383 | ||
384 | static unsigned int reserve_eilvt_offset(int offset, unsigned int new) | 384 | static unsigned int reserve_eilvt_offset(int offset, unsigned int new) |
385 | { | 385 | { |
386 | unsigned int rsvd; /* 0: uninitialized */ | 386 | unsigned int rsvd, vector; |
387 | 387 | ||
388 | if (offset >= APIC_EILVT_NR_MAX) | 388 | if (offset >= APIC_EILVT_NR_MAX) |
389 | return ~0; | 389 | return ~0; |
390 | 390 | ||
391 | rsvd = atomic_read(&eilvt_offsets[offset]) & ~APIC_EILVT_MASKED; | 391 | rsvd = atomic_read(&eilvt_offsets[offset]); |
392 | do { | 392 | do { |
393 | if (rsvd && | 393 | vector = rsvd & ~APIC_EILVT_MASKED; /* 0: unassigned */ |
394 | !eilvt_entry_is_changeable(rsvd, new)) | 394 | if (vector && !eilvt_entry_is_changeable(vector, new)) |
395 | /* may not change if vectors are different */ | 395 | /* may not change if vectors are different */ |
396 | return rsvd; | 396 | return rsvd; |
397 | rsvd = atomic_cmpxchg(&eilvt_offsets[offset], rsvd, new); | 397 | rsvd = atomic_cmpxchg(&eilvt_offsets[offset], rsvd, new); |
398 | } while (rsvd != new); | 398 | } while (rsvd != new); |
399 | 399 | ||
400 | rsvd &= ~APIC_EILVT_MASKED; | ||
401 | if (rsvd && rsvd != vector) | ||
402 | pr_info("LVT offset %d assigned for vector 0x%02x\n", | ||
403 | offset, rsvd); | ||
404 | |||
400 | return new; | 405 | return new; |
401 | } | 406 | } |
402 | 407 | ||
diff --git a/arch/x86/kernel/apic/apic_numachip.c b/arch/x86/kernel/apic/apic_numachip.c index d9ea5f331ac..899803e0321 100644 --- a/arch/x86/kernel/apic/apic_numachip.c +++ b/arch/x86/kernel/apic/apic_numachip.c | |||
@@ -229,11 +229,10 @@ static int __init numachip_system_init(void) | |||
229 | } | 229 | } |
230 | early_initcall(numachip_system_init); | 230 | early_initcall(numachip_system_init); |
231 | 231 | ||
232 | static int __cpuinit numachip_acpi_madt_oem_check(char *oem_id, char *oem_table_id) | 232 | static int numachip_acpi_madt_oem_check(char *oem_id, char *oem_table_id) |
233 | { | 233 | { |
234 | if (!strncmp(oem_id, "NUMASC", 6)) { | 234 | if (!strncmp(oem_id, "NUMASC", 6)) { |
235 | numachip_system = 1; | 235 | numachip_system = 1; |
236 | setup_force_cpu_cap(X86_FEATURE_X2APIC); | ||
237 | return 1; | 236 | return 1; |
238 | } | 237 | } |
239 | 238 | ||
diff --git a/arch/x86/kernel/apic/io_apic.c b/arch/x86/kernel/apic/io_apic.c index 6d10a66fc5a..e88300d8e80 100644 --- a/arch/x86/kernel/apic/io_apic.c +++ b/arch/x86/kernel/apic/io_apic.c | |||
@@ -64,9 +64,28 @@ | |||
64 | #include <asm/apic.h> | 64 | #include <asm/apic.h> |
65 | 65 | ||
66 | #define __apicdebuginit(type) static type __init | 66 | #define __apicdebuginit(type) static type __init |
67 | |||
67 | #define for_each_irq_pin(entry, head) \ | 68 | #define for_each_irq_pin(entry, head) \ |
68 | for (entry = head; entry; entry = entry->next) | 69 | for (entry = head; entry; entry = entry->next) |
69 | 70 | ||
71 | static void __init __ioapic_init_mappings(void); | ||
72 | |||
73 | static unsigned int __io_apic_read (unsigned int apic, unsigned int reg); | ||
74 | static void __io_apic_write (unsigned int apic, unsigned int reg, unsigned int val); | ||
75 | static void __io_apic_modify(unsigned int apic, unsigned int reg, unsigned int val); | ||
76 | |||
77 | static struct io_apic_ops io_apic_ops = { | ||
78 | .init = __ioapic_init_mappings, | ||
79 | .read = __io_apic_read, | ||
80 | .write = __io_apic_write, | ||
81 | .modify = __io_apic_modify, | ||
82 | }; | ||
83 | |||
84 | void __init set_io_apic_ops(const struct io_apic_ops *ops) | ||
85 | { | ||
86 | io_apic_ops = *ops; | ||
87 | } | ||
88 | |||
70 | /* | 89 | /* |
71 | * Is the SiS APIC rmw bug present ? | 90 | * Is the SiS APIC rmw bug present ? |
72 | * -1 = don't know, 0 = no, 1 = yes | 91 | * -1 = don't know, 0 = no, 1 = yes |
@@ -294,6 +313,22 @@ static void free_irq_at(unsigned int at, struct irq_cfg *cfg) | |||
294 | irq_free_desc(at); | 313 | irq_free_desc(at); |
295 | } | 314 | } |
296 | 315 | ||
316 | static inline unsigned int io_apic_read(unsigned int apic, unsigned int reg) | ||
317 | { | ||
318 | return io_apic_ops.read(apic, reg); | ||
319 | } | ||
320 | |||
321 | static inline void io_apic_write(unsigned int apic, unsigned int reg, unsigned int value) | ||
322 | { | ||
323 | io_apic_ops.write(apic, reg, value); | ||
324 | } | ||
325 | |||
326 | static inline void io_apic_modify(unsigned int apic, unsigned int reg, unsigned int value) | ||
327 | { | ||
328 | io_apic_ops.modify(apic, reg, value); | ||
329 | } | ||
330 | |||
331 | |||
297 | struct io_apic { | 332 | struct io_apic { |
298 | unsigned int index; | 333 | unsigned int index; |
299 | unsigned int unused[3]; | 334 | unsigned int unused[3]; |
@@ -314,16 +349,17 @@ static inline void io_apic_eoi(unsigned int apic, unsigned int vector) | |||
314 | writel(vector, &io_apic->eoi); | 349 | writel(vector, &io_apic->eoi); |
315 | } | 350 | } |
316 | 351 | ||
317 | static inline unsigned int io_apic_read(unsigned int apic, unsigned int reg) | 352 | static unsigned int __io_apic_read(unsigned int apic, unsigned int reg) |
318 | { | 353 | { |
319 | struct io_apic __iomem *io_apic = io_apic_base(apic); | 354 | struct io_apic __iomem *io_apic = io_apic_base(apic); |
320 | writel(reg, &io_apic->index); | 355 | writel(reg, &io_apic->index); |
321 | return readl(&io_apic->data); | 356 | return readl(&io_apic->data); |
322 | } | 357 | } |
323 | 358 | ||
324 | static inline void io_apic_write(unsigned int apic, unsigned int reg, unsigned int value) | 359 | static void __io_apic_write(unsigned int apic, unsigned int reg, unsigned int value) |
325 | { | 360 | { |
326 | struct io_apic __iomem *io_apic = io_apic_base(apic); | 361 | struct io_apic __iomem *io_apic = io_apic_base(apic); |
362 | |||
327 | writel(reg, &io_apic->index); | 363 | writel(reg, &io_apic->index); |
328 | writel(value, &io_apic->data); | 364 | writel(value, &io_apic->data); |
329 | } | 365 | } |
@@ -334,7 +370,7 @@ static inline void io_apic_write(unsigned int apic, unsigned int reg, unsigned i | |||
334 | * | 370 | * |
335 | * Older SiS APIC requires we rewrite the index register | 371 | * Older SiS APIC requires we rewrite the index register |
336 | */ | 372 | */ |
337 | static inline void io_apic_modify(unsigned int apic, unsigned int reg, unsigned int value) | 373 | static void __io_apic_modify(unsigned int apic, unsigned int reg, unsigned int value) |
338 | { | 374 | { |
339 | struct io_apic __iomem *io_apic = io_apic_base(apic); | 375 | struct io_apic __iomem *io_apic = io_apic_base(apic); |
340 | 376 | ||
@@ -377,6 +413,7 @@ static struct IO_APIC_route_entry __ioapic_read_entry(int apic, int pin) | |||
377 | 413 | ||
378 | eu.w1 = io_apic_read(apic, 0x10 + 2 * pin); | 414 | eu.w1 = io_apic_read(apic, 0x10 + 2 * pin); |
379 | eu.w2 = io_apic_read(apic, 0x11 + 2 * pin); | 415 | eu.w2 = io_apic_read(apic, 0x11 + 2 * pin); |
416 | |||
380 | return eu.entry; | 417 | return eu.entry; |
381 | } | 418 | } |
382 | 419 | ||
@@ -384,9 +421,11 @@ static struct IO_APIC_route_entry ioapic_read_entry(int apic, int pin) | |||
384 | { | 421 | { |
385 | union entry_union eu; | 422 | union entry_union eu; |
386 | unsigned long flags; | 423 | unsigned long flags; |
424 | |||
387 | raw_spin_lock_irqsave(&ioapic_lock, flags); | 425 | raw_spin_lock_irqsave(&ioapic_lock, flags); |
388 | eu.entry = __ioapic_read_entry(apic, pin); | 426 | eu.entry = __ioapic_read_entry(apic, pin); |
389 | raw_spin_unlock_irqrestore(&ioapic_lock, flags); | 427 | raw_spin_unlock_irqrestore(&ioapic_lock, flags); |
428 | |||
390 | return eu.entry; | 429 | return eu.entry; |
391 | } | 430 | } |
392 | 431 | ||
@@ -396,8 +435,7 @@ static struct IO_APIC_route_entry ioapic_read_entry(int apic, int pin) | |||
396 | * the interrupt, and we need to make sure the entry is fully populated | 435 | * the interrupt, and we need to make sure the entry is fully populated |
397 | * before that happens. | 436 | * before that happens. |
398 | */ | 437 | */ |
399 | static void | 438 | static void __ioapic_write_entry(int apic, int pin, struct IO_APIC_route_entry e) |
400 | __ioapic_write_entry(int apic, int pin, struct IO_APIC_route_entry e) | ||
401 | { | 439 | { |
402 | union entry_union eu = {{0, 0}}; | 440 | union entry_union eu = {{0, 0}}; |
403 | 441 | ||
@@ -409,6 +447,7 @@ __ioapic_write_entry(int apic, int pin, struct IO_APIC_route_entry e) | |||
409 | static void ioapic_write_entry(int apic, int pin, struct IO_APIC_route_entry e) | 447 | static void ioapic_write_entry(int apic, int pin, struct IO_APIC_route_entry e) |
410 | { | 448 | { |
411 | unsigned long flags; | 449 | unsigned long flags; |
450 | |||
412 | raw_spin_lock_irqsave(&ioapic_lock, flags); | 451 | raw_spin_lock_irqsave(&ioapic_lock, flags); |
413 | __ioapic_write_entry(apic, pin, e); | 452 | __ioapic_write_entry(apic, pin, e); |
414 | raw_spin_unlock_irqrestore(&ioapic_lock, flags); | 453 | raw_spin_unlock_irqrestore(&ioapic_lock, flags); |
@@ -435,8 +474,7 @@ static void ioapic_mask_entry(int apic, int pin) | |||
435 | * shared ISA-space IRQs, so we have to support them. We are super | 474 | * shared ISA-space IRQs, so we have to support them. We are super |
436 | * fast in the common case, and fast for shared ISA-space IRQs. | 475 | * fast in the common case, and fast for shared ISA-space IRQs. |
437 | */ | 476 | */ |
438 | static int | 477 | static int __add_pin_to_irq_node(struct irq_cfg *cfg, int node, int apic, int pin) |
439 | __add_pin_to_irq_node(struct irq_cfg *cfg, int node, int apic, int pin) | ||
440 | { | 478 | { |
441 | struct irq_pin_list **last, *entry; | 479 | struct irq_pin_list **last, *entry; |
442 | 480 | ||
@@ -521,6 +559,7 @@ static void io_apic_sync(struct irq_pin_list *entry) | |||
521 | * a dummy read from the IO-APIC | 559 | * a dummy read from the IO-APIC |
522 | */ | 560 | */ |
523 | struct io_apic __iomem *io_apic; | 561 | struct io_apic __iomem *io_apic; |
562 | |||
524 | io_apic = io_apic_base(entry->apic); | 563 | io_apic = io_apic_base(entry->apic); |
525 | readl(&io_apic->data); | 564 | readl(&io_apic->data); |
526 | } | 565 | } |
@@ -2512,21 +2551,73 @@ static void ack_apic_edge(struct irq_data *data) | |||
2512 | 2551 | ||
2513 | atomic_t irq_mis_count; | 2552 | atomic_t irq_mis_count; |
2514 | 2553 | ||
2515 | static void ack_apic_level(struct irq_data *data) | ||
2516 | { | ||
2517 | struct irq_cfg *cfg = data->chip_data; | ||
2518 | int i, do_unmask_irq = 0, irq = data->irq; | ||
2519 | unsigned long v; | ||
2520 | |||
2521 | irq_complete_move(cfg); | ||
2522 | #ifdef CONFIG_GENERIC_PENDING_IRQ | 2554 | #ifdef CONFIG_GENERIC_PENDING_IRQ |
2555 | static inline bool ioapic_irqd_mask(struct irq_data *data, struct irq_cfg *cfg) | ||
2556 | { | ||
2523 | /* If we are moving the irq we need to mask it */ | 2557 | /* If we are moving the irq we need to mask it */ |
2524 | if (unlikely(irqd_is_setaffinity_pending(data))) { | 2558 | if (unlikely(irqd_is_setaffinity_pending(data))) { |
2525 | do_unmask_irq = 1; | ||
2526 | mask_ioapic(cfg); | 2559 | mask_ioapic(cfg); |
2560 | return true; | ||
2527 | } | 2561 | } |
2562 | return false; | ||
2563 | } | ||
2564 | |||
2565 | static inline void ioapic_irqd_unmask(struct irq_data *data, | ||
2566 | struct irq_cfg *cfg, bool masked) | ||
2567 | { | ||
2568 | if (unlikely(masked)) { | ||
2569 | /* Only migrate the irq if the ack has been received. | ||
2570 | * | ||
2571 | * On rare occasions the broadcast level triggered ack gets | ||
2572 | * delayed going to ioapics, and if we reprogram the | ||
2573 | * vector while Remote IRR is still set the irq will never | ||
2574 | * fire again. | ||
2575 | * | ||
2576 | * To prevent this scenario we read the Remote IRR bit | ||
2577 | * of the ioapic. This has two effects. | ||
2578 | * - On any sane system the read of the ioapic will | ||
2579 | * flush writes (and acks) going to the ioapic from | ||
2580 | * this cpu. | ||
2581 | * - We get to see if the ACK has actually been delivered. | ||
2582 | * | ||
2583 | * Based on failed experiments of reprogramming the | ||
2584 | * ioapic entry from outside of irq context starting | ||
2585 | * with masking the ioapic entry and then polling until | ||
2586 | * Remote IRR was clear before reprogramming the | ||
2587 | * ioapic I don't trust the Remote IRR bit to be | ||
2588 | * completey accurate. | ||
2589 | * | ||
2590 | * However there appears to be no other way to plug | ||
2591 | * this race, so if the Remote IRR bit is not | ||
2592 | * accurate and is causing problems then it is a hardware bug | ||
2593 | * and you can go talk to the chipset vendor about it. | ||
2594 | */ | ||
2595 | if (!io_apic_level_ack_pending(cfg)) | ||
2596 | irq_move_masked_irq(data); | ||
2597 | unmask_ioapic(cfg); | ||
2598 | } | ||
2599 | } | ||
2600 | #else | ||
2601 | static inline bool ioapic_irqd_mask(struct irq_data *data, struct irq_cfg *cfg) | ||
2602 | { | ||
2603 | return false; | ||
2604 | } | ||
2605 | static inline void ioapic_irqd_unmask(struct irq_data *data, | ||
2606 | struct irq_cfg *cfg, bool masked) | ||
2607 | { | ||
2608 | } | ||
2528 | #endif | 2609 | #endif |
2529 | 2610 | ||
2611 | static void ack_apic_level(struct irq_data *data) | ||
2612 | { | ||
2613 | struct irq_cfg *cfg = data->chip_data; | ||
2614 | int i, irq = data->irq; | ||
2615 | unsigned long v; | ||
2616 | bool masked; | ||
2617 | |||
2618 | irq_complete_move(cfg); | ||
2619 | masked = ioapic_irqd_mask(data, cfg); | ||
2620 | |||
2530 | /* | 2621 | /* |
2531 | * It appears there is an erratum which affects at least version 0x11 | 2622 | * It appears there is an erratum which affects at least version 0x11 |
2532 | * of I/O APIC (that's the 82093AA and cores integrated into various | 2623 | * of I/O APIC (that's the 82093AA and cores integrated into various |
@@ -2581,38 +2672,7 @@ static void ack_apic_level(struct irq_data *data) | |||
2581 | eoi_ioapic_irq(irq, cfg); | 2672 | eoi_ioapic_irq(irq, cfg); |
2582 | } | 2673 | } |
2583 | 2674 | ||
2584 | /* Now we can move and renable the irq */ | 2675 | ioapic_irqd_unmask(data, cfg, masked); |
2585 | if (unlikely(do_unmask_irq)) { | ||
2586 | /* Only migrate the irq if the ack has been received. | ||
2587 | * | ||
2588 | * On rare occasions the broadcast level triggered ack gets | ||
2589 | * delayed going to ioapics, and if we reprogram the | ||
2590 | * vector while Remote IRR is still set the irq will never | ||
2591 | * fire again. | ||
2592 | * | ||
2593 | * To prevent this scenario we read the Remote IRR bit | ||
2594 | * of the ioapic. This has two effects. | ||
2595 | * - On any sane system the read of the ioapic will | ||
2596 | * flush writes (and acks) going to the ioapic from | ||
2597 | * this cpu. | ||
2598 | * - We get to see if the ACK has actually been delivered. | ||
2599 | * | ||
2600 | * Based on failed experiments of reprogramming the | ||
2601 | * ioapic entry from outside of irq context starting | ||
2602 | * with masking the ioapic entry and then polling until | ||
2603 | * Remote IRR was clear before reprogramming the | ||
2604 | * ioapic I don't trust the Remote IRR bit to be | ||
2605 | * completey accurate. | ||
2606 | * | ||
2607 | * However there appears to be no other way to plug | ||
2608 | * this race, so if the Remote IRR bit is not | ||
2609 | * accurate and is causing problems then it is a hardware bug | ||
2610 | * and you can go talk to the chipset vendor about it. | ||
2611 | */ | ||
2612 | if (!io_apic_level_ack_pending(cfg)) | ||
2613 | irq_move_masked_irq(data); | ||
2614 | unmask_ioapic(cfg); | ||
2615 | } | ||
2616 | } | 2676 | } |
2617 | 2677 | ||
2618 | #ifdef CONFIG_IRQ_REMAP | 2678 | #ifdef CONFIG_IRQ_REMAP |
@@ -3873,6 +3933,11 @@ static struct resource * __init ioapic_setup_resources(int nr_ioapics) | |||
3873 | 3933 | ||
3874 | void __init ioapic_and_gsi_init(void) | 3934 | void __init ioapic_and_gsi_init(void) |
3875 | { | 3935 | { |
3936 | io_apic_ops.init(); | ||
3937 | } | ||
3938 | |||
3939 | static void __init __ioapic_init_mappings(void) | ||
3940 | { | ||
3876 | unsigned long ioapic_phys, idx = FIX_IO_APIC_BASE_0; | 3941 | unsigned long ioapic_phys, idx = FIX_IO_APIC_BASE_0; |
3877 | struct resource *ioapic_res; | 3942 | struct resource *ioapic_res; |
3878 | int i; | 3943 | int i; |
diff --git a/arch/x86/kernel/apic/x2apic_cluster.c b/arch/x86/kernel/apic/x2apic_cluster.c index 9193713060a..48f3103b3c9 100644 --- a/arch/x86/kernel/apic/x2apic_cluster.c +++ b/arch/x86/kernel/apic/x2apic_cluster.c | |||
@@ -213,7 +213,7 @@ static struct apic apic_x2apic_cluster = { | |||
213 | .name = "cluster x2apic", | 213 | .name = "cluster x2apic", |
214 | .probe = x2apic_cluster_probe, | 214 | .probe = x2apic_cluster_probe, |
215 | .acpi_madt_oem_check = x2apic_acpi_madt_oem_check, | 215 | .acpi_madt_oem_check = x2apic_acpi_madt_oem_check, |
216 | .apic_id_valid = default_apic_id_valid, | 216 | .apic_id_valid = x2apic_apic_id_valid, |
217 | .apic_id_registered = x2apic_apic_id_registered, | 217 | .apic_id_registered = x2apic_apic_id_registered, |
218 | 218 | ||
219 | .irq_delivery_mode = dest_LowestPrio, | 219 | .irq_delivery_mode = dest_LowestPrio, |
diff --git a/arch/x86/kernel/apic/x2apic_phys.c b/arch/x86/kernel/apic/x2apic_phys.c index bcd1db6eaca..8a778db45e3 100644 --- a/arch/x86/kernel/apic/x2apic_phys.c +++ b/arch/x86/kernel/apic/x2apic_phys.c | |||
@@ -119,7 +119,7 @@ static struct apic apic_x2apic_phys = { | |||
119 | .name = "physical x2apic", | 119 | .name = "physical x2apic", |
120 | .probe = x2apic_phys_probe, | 120 | .probe = x2apic_phys_probe, |
121 | .acpi_madt_oem_check = x2apic_acpi_madt_oem_check, | 121 | .acpi_madt_oem_check = x2apic_acpi_madt_oem_check, |
122 | .apic_id_valid = default_apic_id_valid, | 122 | .apic_id_valid = x2apic_apic_id_valid, |
123 | .apic_id_registered = x2apic_apic_id_registered, | 123 | .apic_id_registered = x2apic_apic_id_registered, |
124 | 124 | ||
125 | .irq_delivery_mode = dest_Fixed, | 125 | .irq_delivery_mode = dest_Fixed, |
diff --git a/arch/x86/kernel/apic/x2apic_uv_x.c b/arch/x86/kernel/apic/x2apic_uv_x.c index fc477142585..87bfa69e216 100644 --- a/arch/x86/kernel/apic/x2apic_uv_x.c +++ b/arch/x86/kernel/apic/x2apic_uv_x.c | |||
@@ -266,6 +266,11 @@ static void uv_send_IPI_all(int vector) | |||
266 | uv_send_IPI_mask(cpu_online_mask, vector); | 266 | uv_send_IPI_mask(cpu_online_mask, vector); |
267 | } | 267 | } |
268 | 268 | ||
269 | static int uv_apic_id_valid(int apicid) | ||
270 | { | ||
271 | return 1; | ||
272 | } | ||
273 | |||
269 | static int uv_apic_id_registered(void) | 274 | static int uv_apic_id_registered(void) |
270 | { | 275 | { |
271 | return 1; | 276 | return 1; |
@@ -351,7 +356,7 @@ static struct apic __refdata apic_x2apic_uv_x = { | |||
351 | .name = "UV large system", | 356 | .name = "UV large system", |
352 | .probe = uv_probe, | 357 | .probe = uv_probe, |
353 | .acpi_madt_oem_check = uv_acpi_madt_oem_check, | 358 | .acpi_madt_oem_check = uv_acpi_madt_oem_check, |
354 | .apic_id_valid = default_apic_id_valid, | 359 | .apic_id_valid = uv_apic_id_valid, |
355 | .apic_id_registered = uv_apic_id_registered, | 360 | .apic_id_registered = uv_apic_id_registered, |
356 | 361 | ||
357 | .irq_delivery_mode = dest_Fixed, | 362 | .irq_delivery_mode = dest_Fixed, |
diff --git a/arch/x86/kernel/cpu/common.c b/arch/x86/kernel/cpu/common.c index e49477444ff..67e258362a3 100644 --- a/arch/x86/kernel/cpu/common.c +++ b/arch/x86/kernel/cpu/common.c | |||
@@ -999,7 +999,7 @@ void __cpuinit print_cpu_info(struct cpuinfo_x86 *c) | |||
999 | else | 999 | else |
1000 | printk(KERN_CONT "\n"); | 1000 | printk(KERN_CONT "\n"); |
1001 | 1001 | ||
1002 | __print_cpu_msr(); | 1002 | print_cpu_msr(c); |
1003 | } | 1003 | } |
1004 | 1004 | ||
1005 | void __cpuinit print_cpu_msr(struct cpuinfo_x86 *c) | 1005 | void __cpuinit print_cpu_msr(struct cpuinfo_x86 *c) |
diff --git a/arch/x86/kernel/dumpstack.c b/arch/x86/kernel/dumpstack.c index 4025fe4f928..90bf130f09b 100644 --- a/arch/x86/kernel/dumpstack.c +++ b/arch/x86/kernel/dumpstack.c | |||
@@ -37,13 +37,16 @@ print_ftrace_graph_addr(unsigned long addr, void *data, | |||
37 | const struct stacktrace_ops *ops, | 37 | const struct stacktrace_ops *ops, |
38 | struct thread_info *tinfo, int *graph) | 38 | struct thread_info *tinfo, int *graph) |
39 | { | 39 | { |
40 | struct task_struct *task = tinfo->task; | 40 | struct task_struct *task; |
41 | unsigned long ret_addr; | 41 | unsigned long ret_addr; |
42 | int index = task->curr_ret_stack; | 42 | int index; |
43 | 43 | ||
44 | if (addr != (unsigned long)return_to_handler) | 44 | if (addr != (unsigned long)return_to_handler) |
45 | return; | 45 | return; |
46 | 46 | ||
47 | task = tinfo->task; | ||
48 | index = task->curr_ret_stack; | ||
49 | |||
47 | if (!task->ret_stack || index < *graph) | 50 | if (!task->ret_stack || index < *graph) |
48 | return; | 51 | return; |
49 | 52 | ||
diff --git a/arch/x86/kernel/process.c b/arch/x86/kernel/process.c index 9b24f36eb55..a33afaa5ddb 100644 --- a/arch/x86/kernel/process.c +++ b/arch/x86/kernel/process.c | |||
@@ -12,6 +12,9 @@ | |||
12 | #include <linux/user-return-notifier.h> | 12 | #include <linux/user-return-notifier.h> |
13 | #include <linux/dmi.h> | 13 | #include <linux/dmi.h> |
14 | #include <linux/utsname.h> | 14 | #include <linux/utsname.h> |
15 | #include <linux/stackprotector.h> | ||
16 | #include <linux/tick.h> | ||
17 | #include <linux/cpuidle.h> | ||
15 | #include <trace/events/power.h> | 18 | #include <trace/events/power.h> |
16 | #include <linux/hw_breakpoint.h> | 19 | #include <linux/hw_breakpoint.h> |
17 | #include <asm/cpu.h> | 20 | #include <asm/cpu.h> |
@@ -22,6 +25,24 @@ | |||
22 | #include <asm/i387.h> | 25 | #include <asm/i387.h> |
23 | #include <asm/fpu-internal.h> | 26 | #include <asm/fpu-internal.h> |
24 | #include <asm/debugreg.h> | 27 | #include <asm/debugreg.h> |
28 | #include <asm/nmi.h> | ||
29 | |||
30 | #ifdef CONFIG_X86_64 | ||
31 | static DEFINE_PER_CPU(unsigned char, is_idle); | ||
32 | static ATOMIC_NOTIFIER_HEAD(idle_notifier); | ||
33 | |||
34 | void idle_notifier_register(struct notifier_block *n) | ||
35 | { | ||
36 | atomic_notifier_chain_register(&idle_notifier, n); | ||
37 | } | ||
38 | EXPORT_SYMBOL_GPL(idle_notifier_register); | ||
39 | |||
40 | void idle_notifier_unregister(struct notifier_block *n) | ||
41 | { | ||
42 | atomic_notifier_chain_unregister(&idle_notifier, n); | ||
43 | } | ||
44 | EXPORT_SYMBOL_GPL(idle_notifier_unregister); | ||
45 | #endif | ||
25 | 46 | ||
26 | struct kmem_cache *task_xstate_cachep; | 47 | struct kmem_cache *task_xstate_cachep; |
27 | EXPORT_SYMBOL_GPL(task_xstate_cachep); | 48 | EXPORT_SYMBOL_GPL(task_xstate_cachep); |
@@ -370,6 +391,99 @@ static inline int hlt_use_halt(void) | |||
370 | } | 391 | } |
371 | #endif | 392 | #endif |
372 | 393 | ||
394 | #ifndef CONFIG_SMP | ||
395 | static inline void play_dead(void) | ||
396 | { | ||
397 | BUG(); | ||
398 | } | ||
399 | #endif | ||
400 | |||
401 | #ifdef CONFIG_X86_64 | ||
402 | void enter_idle(void) | ||
403 | { | ||
404 | percpu_write(is_idle, 1); | ||
405 | atomic_notifier_call_chain(&idle_notifier, IDLE_START, NULL); | ||
406 | } | ||
407 | |||
408 | static void __exit_idle(void) | ||
409 | { | ||
410 | if (x86_test_and_clear_bit_percpu(0, is_idle) == 0) | ||
411 | return; | ||
412 | atomic_notifier_call_chain(&idle_notifier, IDLE_END, NULL); | ||
413 | } | ||
414 | |||
415 | /* Called from interrupts to signify idle end */ | ||
416 | void exit_idle(void) | ||
417 | { | ||
418 | /* idle loop has pid 0 */ | ||
419 | if (current->pid) | ||
420 | return; | ||
421 | __exit_idle(); | ||
422 | } | ||
423 | #endif | ||
424 | |||
425 | /* | ||
426 | * The idle thread. There's no useful work to be | ||
427 | * done, so just try to conserve power and have a | ||
428 | * low exit latency (ie sit in a loop waiting for | ||
429 | * somebody to say that they'd like to reschedule) | ||
430 | */ | ||
431 | void cpu_idle(void) | ||
432 | { | ||
433 | /* | ||
434 | * If we're the non-boot CPU, nothing set the stack canary up | ||
435 | * for us. CPU0 already has it initialized but no harm in | ||
436 | * doing it again. This is a good place for updating it, as | ||
437 | * we wont ever return from this function (so the invalid | ||
438 | * canaries already on the stack wont ever trigger). | ||
439 | */ | ||
440 | boot_init_stack_canary(); | ||
441 | current_thread_info()->status |= TS_POLLING; | ||
442 | |||
443 | while (1) { | ||
444 | tick_nohz_idle_enter(); | ||
445 | |||
446 | while (!need_resched()) { | ||
447 | rmb(); | ||
448 | |||
449 | if (cpu_is_offline(smp_processor_id())) | ||
450 | play_dead(); | ||
451 | |||
452 | /* | ||
453 | * Idle routines should keep interrupts disabled | ||
454 | * from here on, until they go to idle. | ||
455 | * Otherwise, idle callbacks can misfire. | ||
456 | */ | ||
457 | local_touch_nmi(); | ||
458 | local_irq_disable(); | ||
459 | |||
460 | enter_idle(); | ||
461 | |||
462 | /* Don't trace irqs off for idle */ | ||
463 | stop_critical_timings(); | ||
464 | |||
465 | /* enter_idle() needs rcu for notifiers */ | ||
466 | rcu_idle_enter(); | ||
467 | |||
468 | if (cpuidle_idle_call()) | ||
469 | pm_idle(); | ||
470 | |||
471 | rcu_idle_exit(); | ||
472 | start_critical_timings(); | ||
473 | |||
474 | /* In many cases the interrupt that ended idle | ||
475 | has already called exit_idle. But some idle | ||
476 | loops can be woken up without interrupt. */ | ||
477 | __exit_idle(); | ||
478 | } | ||
479 | |||
480 | tick_nohz_idle_exit(); | ||
481 | preempt_enable_no_resched(); | ||
482 | schedule(); | ||
483 | preempt_disable(); | ||
484 | } | ||
485 | } | ||
486 | |||
373 | /* | 487 | /* |
374 | * We use this if we don't have any better | 488 | * We use this if we don't have any better |
375 | * idle routine.. | 489 | * idle routine.. |
diff --git a/arch/x86/kernel/process_32.c b/arch/x86/kernel/process_32.c index aae4f4bbbe8..ae6847303e2 100644 --- a/arch/x86/kernel/process_32.c +++ b/arch/x86/kernel/process_32.c | |||
@@ -9,7 +9,6 @@ | |||
9 | * This file handles the architecture-dependent parts of process handling.. | 9 | * This file handles the architecture-dependent parts of process handling.. |
10 | */ | 10 | */ |
11 | 11 | ||
12 | #include <linux/stackprotector.h> | ||
13 | #include <linux/cpu.h> | 12 | #include <linux/cpu.h> |
14 | #include <linux/errno.h> | 13 | #include <linux/errno.h> |
15 | #include <linux/sched.h> | 14 | #include <linux/sched.h> |
@@ -31,14 +30,12 @@ | |||
31 | #include <linux/kallsyms.h> | 30 | #include <linux/kallsyms.h> |
32 | #include <linux/ptrace.h> | 31 | #include <linux/ptrace.h> |
33 | #include <linux/personality.h> | 32 | #include <linux/personality.h> |
34 | #include <linux/tick.h> | ||
35 | #include <linux/percpu.h> | 33 | #include <linux/percpu.h> |
36 | #include <linux/prctl.h> | 34 | #include <linux/prctl.h> |
37 | #include <linux/ftrace.h> | 35 | #include <linux/ftrace.h> |
38 | #include <linux/uaccess.h> | 36 | #include <linux/uaccess.h> |
39 | #include <linux/io.h> | 37 | #include <linux/io.h> |
40 | #include <linux/kdebug.h> | 38 | #include <linux/kdebug.h> |
41 | #include <linux/cpuidle.h> | ||
42 | 39 | ||
43 | #include <asm/pgtable.h> | 40 | #include <asm/pgtable.h> |
44 | #include <asm/ldt.h> | 41 | #include <asm/ldt.h> |
@@ -57,7 +54,6 @@ | |||
57 | #include <asm/idle.h> | 54 | #include <asm/idle.h> |
58 | #include <asm/syscalls.h> | 55 | #include <asm/syscalls.h> |
59 | #include <asm/debugreg.h> | 56 | #include <asm/debugreg.h> |
60 | #include <asm/nmi.h> | ||
61 | #include <asm/switch_to.h> | 57 | #include <asm/switch_to.h> |
62 | 58 | ||
63 | asmlinkage void ret_from_fork(void) __asm__("ret_from_fork"); | 59 | asmlinkage void ret_from_fork(void) __asm__("ret_from_fork"); |
@@ -70,60 +66,6 @@ unsigned long thread_saved_pc(struct task_struct *tsk) | |||
70 | return ((unsigned long *)tsk->thread.sp)[3]; | 66 | return ((unsigned long *)tsk->thread.sp)[3]; |
71 | } | 67 | } |
72 | 68 | ||
73 | #ifndef CONFIG_SMP | ||
74 | static inline void play_dead(void) | ||
75 | { | ||
76 | BUG(); | ||
77 | } | ||
78 | #endif | ||
79 | |||
80 | /* | ||
81 | * The idle thread. There's no useful work to be | ||
82 | * done, so just try to conserve power and have a | ||
83 | * low exit latency (ie sit in a loop waiting for | ||
84 | * somebody to say that they'd like to reschedule) | ||
85 | */ | ||
86 | void cpu_idle(void) | ||
87 | { | ||
88 | int cpu = smp_processor_id(); | ||
89 | |||
90 | /* | ||
91 | * If we're the non-boot CPU, nothing set the stack canary up | ||
92 | * for us. CPU0 already has it initialized but no harm in | ||
93 | * doing it again. This is a good place for updating it, as | ||
94 | * we wont ever return from this function (so the invalid | ||
95 | * canaries already on the stack wont ever trigger). | ||
96 | */ | ||
97 | boot_init_stack_canary(); | ||
98 | |||
99 | current_thread_info()->status |= TS_POLLING; | ||
100 | |||
101 | /* endless idle loop with no priority at all */ | ||
102 | while (1) { | ||
103 | tick_nohz_idle_enter(); | ||
104 | rcu_idle_enter(); | ||
105 | while (!need_resched()) { | ||
106 | |||
107 | check_pgt_cache(); | ||
108 | rmb(); | ||
109 | |||
110 | if (cpu_is_offline(cpu)) | ||
111 | play_dead(); | ||
112 | |||
113 | local_touch_nmi(); | ||
114 | local_irq_disable(); | ||
115 | /* Don't trace irqs off for idle */ | ||
116 | stop_critical_timings(); | ||
117 | if (cpuidle_idle_call()) | ||
118 | pm_idle(); | ||
119 | start_critical_timings(); | ||
120 | } | ||
121 | rcu_idle_exit(); | ||
122 | tick_nohz_idle_exit(); | ||
123 | schedule_preempt_disabled(); | ||
124 | } | ||
125 | } | ||
126 | |||
127 | void __show_regs(struct pt_regs *regs, int all) | 69 | void __show_regs(struct pt_regs *regs, int all) |
128 | { | 70 | { |
129 | unsigned long cr0 = 0L, cr2 = 0L, cr3 = 0L, cr4 = 0L; | 71 | unsigned long cr0 = 0L, cr2 = 0L, cr3 = 0L, cr4 = 0L; |
diff --git a/arch/x86/kernel/process_64.c b/arch/x86/kernel/process_64.c index 61270e8d428..2b154da0b6d 100644 --- a/arch/x86/kernel/process_64.c +++ b/arch/x86/kernel/process_64.c | |||
@@ -14,7 +14,6 @@ | |||
14 | * This file handles the architecture-dependent parts of process handling.. | 14 | * This file handles the architecture-dependent parts of process handling.. |
15 | */ | 15 | */ |
16 | 16 | ||
17 | #include <linux/stackprotector.h> | ||
18 | #include <linux/cpu.h> | 17 | #include <linux/cpu.h> |
19 | #include <linux/errno.h> | 18 | #include <linux/errno.h> |
20 | #include <linux/sched.h> | 19 | #include <linux/sched.h> |
@@ -32,12 +31,10 @@ | |||
32 | #include <linux/notifier.h> | 31 | #include <linux/notifier.h> |
33 | #include <linux/kprobes.h> | 32 | #include <linux/kprobes.h> |
34 | #include <linux/kdebug.h> | 33 | #include <linux/kdebug.h> |
35 | #include <linux/tick.h> | ||
36 | #include <linux/prctl.h> | 34 | #include <linux/prctl.h> |
37 | #include <linux/uaccess.h> | 35 | #include <linux/uaccess.h> |
38 | #include <linux/io.h> | 36 | #include <linux/io.h> |
39 | #include <linux/ftrace.h> | 37 | #include <linux/ftrace.h> |
40 | #include <linux/cpuidle.h> | ||
41 | 38 | ||
42 | #include <asm/pgtable.h> | 39 | #include <asm/pgtable.h> |
43 | #include <asm/processor.h> | 40 | #include <asm/processor.h> |
@@ -51,115 +48,11 @@ | |||
51 | #include <asm/idle.h> | 48 | #include <asm/idle.h> |
52 | #include <asm/syscalls.h> | 49 | #include <asm/syscalls.h> |
53 | #include <asm/debugreg.h> | 50 | #include <asm/debugreg.h> |
54 | #include <asm/nmi.h> | ||
55 | #include <asm/switch_to.h> | 51 | #include <asm/switch_to.h> |
56 | 52 | ||
57 | asmlinkage extern void ret_from_fork(void); | 53 | asmlinkage extern void ret_from_fork(void); |
58 | 54 | ||
59 | DEFINE_PER_CPU(unsigned long, old_rsp); | 55 | DEFINE_PER_CPU(unsigned long, old_rsp); |
60 | static DEFINE_PER_CPU(unsigned char, is_idle); | ||
61 | |||
62 | static ATOMIC_NOTIFIER_HEAD(idle_notifier); | ||
63 | |||
64 | void idle_notifier_register(struct notifier_block *n) | ||
65 | { | ||
66 | atomic_notifier_chain_register(&idle_notifier, n); | ||
67 | } | ||
68 | EXPORT_SYMBOL_GPL(idle_notifier_register); | ||
69 | |||
70 | void idle_notifier_unregister(struct notifier_block *n) | ||
71 | { | ||
72 | atomic_notifier_chain_unregister(&idle_notifier, n); | ||
73 | } | ||
74 | EXPORT_SYMBOL_GPL(idle_notifier_unregister); | ||
75 | |||
76 | void enter_idle(void) | ||
77 | { | ||
78 | percpu_write(is_idle, 1); | ||
79 | atomic_notifier_call_chain(&idle_notifier, IDLE_START, NULL); | ||
80 | } | ||
81 | |||
82 | static void __exit_idle(void) | ||
83 | { | ||
84 | if (x86_test_and_clear_bit_percpu(0, is_idle) == 0) | ||
85 | return; | ||
86 | atomic_notifier_call_chain(&idle_notifier, IDLE_END, NULL); | ||
87 | } | ||
88 | |||
89 | /* Called from interrupts to signify idle end */ | ||
90 | void exit_idle(void) | ||
91 | { | ||
92 | /* idle loop has pid 0 */ | ||
93 | if (current->pid) | ||
94 | return; | ||
95 | __exit_idle(); | ||
96 | } | ||
97 | |||
98 | #ifndef CONFIG_SMP | ||
99 | static inline void play_dead(void) | ||
100 | { | ||
101 | BUG(); | ||
102 | } | ||
103 | #endif | ||
104 | |||
105 | /* | ||
106 | * The idle thread. There's no useful work to be | ||
107 | * done, so just try to conserve power and have a | ||
108 | * low exit latency (ie sit in a loop waiting for | ||
109 | * somebody to say that they'd like to reschedule) | ||
110 | */ | ||
111 | void cpu_idle(void) | ||
112 | { | ||
113 | current_thread_info()->status |= TS_POLLING; | ||
114 | |||
115 | /* | ||
116 | * If we're the non-boot CPU, nothing set the stack canary up | ||
117 | * for us. CPU0 already has it initialized but no harm in | ||
118 | * doing it again. This is a good place for updating it, as | ||
119 | * we wont ever return from this function (so the invalid | ||
120 | * canaries already on the stack wont ever trigger). | ||
121 | */ | ||
122 | boot_init_stack_canary(); | ||
123 | |||
124 | /* endless idle loop with no priority at all */ | ||
125 | while (1) { | ||
126 | tick_nohz_idle_enter(); | ||
127 | while (!need_resched()) { | ||
128 | |||
129 | rmb(); | ||
130 | |||
131 | if (cpu_is_offline(smp_processor_id())) | ||
132 | play_dead(); | ||
133 | /* | ||
134 | * Idle routines should keep interrupts disabled | ||
135 | * from here on, until they go to idle. | ||
136 | * Otherwise, idle callbacks can misfire. | ||
137 | */ | ||
138 | local_touch_nmi(); | ||
139 | local_irq_disable(); | ||
140 | enter_idle(); | ||
141 | /* Don't trace irqs off for idle */ | ||
142 | stop_critical_timings(); | ||
143 | |||
144 | /* enter_idle() needs rcu for notifiers */ | ||
145 | rcu_idle_enter(); | ||
146 | |||
147 | if (cpuidle_idle_call()) | ||
148 | pm_idle(); | ||
149 | |||
150 | rcu_idle_exit(); | ||
151 | start_critical_timings(); | ||
152 | |||
153 | /* In many cases the interrupt that ended idle | ||
154 | has already called exit_idle. But some idle | ||
155 | loops can be woken up without interrupt. */ | ||
156 | __exit_idle(); | ||
157 | } | ||
158 | |||
159 | tick_nohz_idle_exit(); | ||
160 | schedule_preempt_disabled(); | ||
161 | } | ||
162 | } | ||
163 | 56 | ||
164 | /* Prints also some state that isn't saved in the pt_regs */ | 57 | /* Prints also some state that isn't saved in the pt_regs */ |
165 | void __show_regs(struct pt_regs *regs, int all) | 58 | void __show_regs(struct pt_regs *regs, int all) |
diff --git a/arch/x86/kernel/tls.c b/arch/x86/kernel/tls.c index 73920e4c6dc..9d9d2f9e77a 100644 --- a/arch/x86/kernel/tls.c +++ b/arch/x86/kernel/tls.c | |||
@@ -162,7 +162,7 @@ int regset_tls_get(struct task_struct *target, const struct user_regset *regset, | |||
162 | { | 162 | { |
163 | const struct desc_struct *tls; | 163 | const struct desc_struct *tls; |
164 | 164 | ||
165 | if (pos > GDT_ENTRY_TLS_ENTRIES * sizeof(struct user_desc) || | 165 | if (pos >= GDT_ENTRY_TLS_ENTRIES * sizeof(struct user_desc) || |
166 | (pos % sizeof(struct user_desc)) != 0 || | 166 | (pos % sizeof(struct user_desc)) != 0 || |
167 | (count % sizeof(struct user_desc)) != 0) | 167 | (count % sizeof(struct user_desc)) != 0) |
168 | return -EINVAL; | 168 | return -EINVAL; |
@@ -197,7 +197,7 @@ int regset_tls_set(struct task_struct *target, const struct user_regset *regset, | |||
197 | struct user_desc infobuf[GDT_ENTRY_TLS_ENTRIES]; | 197 | struct user_desc infobuf[GDT_ENTRY_TLS_ENTRIES]; |
198 | const struct user_desc *info; | 198 | const struct user_desc *info; |
199 | 199 | ||
200 | if (pos > GDT_ENTRY_TLS_ENTRIES * sizeof(struct user_desc) || | 200 | if (pos >= GDT_ENTRY_TLS_ENTRIES * sizeof(struct user_desc) || |
201 | (pos % sizeof(struct user_desc)) != 0 || | 201 | (pos % sizeof(struct user_desc)) != 0 || |
202 | (count % sizeof(struct user_desc)) != 0) | 202 | (count % sizeof(struct user_desc)) != 0) |
203 | return -EINVAL; | 203 | return -EINVAL; |