diff options
40 files changed, 283 insertions, 213 deletions
diff --git a/arch/m32r/Kconfig b/arch/m32r/Kconfig index c41234f1b825..3a9319f93e89 100644 --- a/arch/m32r/Kconfig +++ b/arch/m32r/Kconfig | |||
| @@ -11,6 +11,9 @@ config M32R | |||
| 11 | select HAVE_IDE | 11 | select HAVE_IDE |
| 12 | select HAVE_OPROFILE | 12 | select HAVE_OPROFILE |
| 13 | select INIT_ALL_POSSIBLE | 13 | select INIT_ALL_POSSIBLE |
| 14 | select HAVE_KERNEL_GZIP | ||
| 15 | select HAVE_KERNEL_BZIP2 | ||
| 16 | select HAVE_KERNEL_LZMA | ||
| 14 | 17 | ||
| 15 | config SBUS | 18 | config SBUS |
| 16 | bool | 19 | bool |
diff --git a/arch/m32r/boot/compressed/Makefile b/arch/m32r/boot/compressed/Makefile index 560484ae35ec..1003880d0dfd 100644 --- a/arch/m32r/boot/compressed/Makefile +++ b/arch/m32r/boot/compressed/Makefile | |||
| @@ -4,8 +4,8 @@ | |||
| 4 | # create a compressed vmlinux image from the original vmlinux | 4 | # create a compressed vmlinux image from the original vmlinux |
| 5 | # | 5 | # |
| 6 | 6 | ||
| 7 | targets := vmlinux vmlinux.bin vmlinux.bin.gz head.o misc.o \ | 7 | targets := vmlinux vmlinux.bin vmlinux.bin.gz vmlinux.bin.bz2 \ |
| 8 | piggy.o vmlinux.lds | 8 | vmlinux.bin.lzma head.o misc.o piggy.o vmlinux.lds |
| 9 | 9 | ||
| 10 | OBJECTS = $(obj)/head.o $(obj)/misc.o | 10 | OBJECTS = $(obj)/head.o $(obj)/misc.o |
| 11 | 11 | ||
| @@ -27,6 +27,12 @@ $(obj)/vmlinux.bin: vmlinux FORCE | |||
| 27 | $(obj)/vmlinux.bin.gz: $(obj)/vmlinux.bin FORCE | 27 | $(obj)/vmlinux.bin.gz: $(obj)/vmlinux.bin FORCE |
| 28 | $(call if_changed,gzip) | 28 | $(call if_changed,gzip) |
| 29 | 29 | ||
| 30 | $(obj)/vmlinux.bin.bz2: $(obj)/vmlinux.bin FORCE | ||
| 31 | $(call if_changed,bzip2) | ||
| 32 | |||
| 33 | $(obj)/vmlinux.bin.lzma: $(obj)/vmlinux.bin FORCE | ||
| 34 | $(call if_changed,lzma) | ||
| 35 | |||
| 30 | CFLAGS_misc.o += -fpic | 36 | CFLAGS_misc.o += -fpic |
| 31 | 37 | ||
| 32 | ifdef CONFIG_MMU | 38 | ifdef CONFIG_MMU |
| @@ -37,5 +43,9 @@ endif | |||
| 37 | 43 | ||
| 38 | OBJCOPYFLAGS += -R .empty_zero_page | 44 | OBJCOPYFLAGS += -R .empty_zero_page |
| 39 | 45 | ||
| 40 | $(obj)/piggy.o: $(obj)/vmlinux.scr $(obj)/vmlinux.bin.gz FORCE | 46 | suffix_$(CONFIG_KERNEL_GZIP) = gz |
| 47 | suffix_$(CONFIG_KERNEL_BZIP2) = bz2 | ||
| 48 | suffix_$(CONFIG_KERNEL_LZMA) = lzma | ||
| 49 | |||
| 50 | $(obj)/piggy.o: $(obj)/vmlinux.scr $(obj)/vmlinux.bin.$(suffix_y) FORCE | ||
| 41 | $(call if_changed,ld) | 51 | $(call if_changed,ld) |
diff --git a/arch/m32r/boot/compressed/misc.c b/arch/m32r/boot/compressed/misc.c index d394292498c0..370d60881977 100644 --- a/arch/m32r/boot/compressed/misc.c +++ b/arch/m32r/boot/compressed/misc.c | |||
| @@ -9,140 +9,49 @@ | |||
| 9 | * Adapted for SH by Stuart Menefy, Aug 1999 | 9 | * Adapted for SH by Stuart Menefy, Aug 1999 |
| 10 | * | 10 | * |
| 11 | * 2003-02-12: Support M32R by Takeo Takahashi | 11 | * 2003-02-12: Support M32R by Takeo Takahashi |
| 12 | * This is based on arch/sh/boot/compressed/misc.c. | ||
| 13 | */ | 12 | */ |
| 14 | 13 | ||
| 15 | #include <linux/string.h> | ||
| 16 | |||
| 17 | /* | 14 | /* |
| 18 | * gzip declarations | 15 | * gzip declarations |
| 19 | */ | 16 | */ |
| 20 | |||
| 21 | #define OF(args) args | ||
| 22 | #define STATIC static | 17 | #define STATIC static |
| 23 | 18 | ||
| 24 | #undef memset | 19 | #undef memset |
| 25 | #undef memcpy | 20 | #undef memcpy |
| 26 | #define memzero(s, n) memset ((s), 0, (n)) | 21 | #define memzero(s, n) memset ((s), 0, (n)) |
| 27 | 22 | ||
| 28 | typedef unsigned char uch; | ||
| 29 | typedef unsigned short ush; | ||
| 30 | typedef unsigned long ulg; | ||
| 31 | |||
| 32 | #define WSIZE 0x8000 /* Window size must be at least 32k, */ | ||
| 33 | /* and a power of two */ | ||
| 34 | |||
| 35 | static uch *inbuf; /* input buffer */ | ||
| 36 | static uch window[WSIZE]; /* Sliding window buffer */ | ||
| 37 | |||
| 38 | static unsigned insize = 0; /* valid bytes in inbuf */ | ||
| 39 | static unsigned inptr = 0; /* index of next byte to be processed in inbuf */ | ||
| 40 | static unsigned outcnt = 0; /* bytes in output buffer */ | ||
| 41 | |||
| 42 | /* gzip flag byte */ | ||
| 43 | #define ASCII_FLAG 0x01 /* bit 0 set: file probably ASCII text */ | ||
| 44 | #define CONTINUATION 0x02 /* bit 1 set: continuation of multi-part gzip file */ | ||
| 45 | #define EXTRA_FIELD 0x04 /* bit 2 set: extra field present */ | ||
| 46 | #define ORIG_NAME 0x08 /* bit 3 set: original file name present */ | ||
| 47 | #define COMMENT 0x10 /* bit 4 set: file comment present */ | ||
| 48 | #define ENCRYPTED 0x20 /* bit 5 set: file is encrypted */ | ||
| 49 | #define RESERVED 0xC0 /* bit 6,7: reserved */ | ||
| 50 | |||
| 51 | #define get_byte() (inptr < insize ? inbuf[inptr++] : fill_inbuf()) | ||
| 52 | |||
| 53 | /* Diagnostic functions */ | ||
| 54 | #ifdef DEBUG | ||
| 55 | # define Assert(cond,msg) {if(!(cond)) error(msg);} | ||
| 56 | # define Trace(x) fprintf x | ||
| 57 | # define Tracev(x) {if (verbose) fprintf x ;} | ||
| 58 | # define Tracevv(x) {if (verbose>1) fprintf x ;} | ||
| 59 | # define Tracec(c,x) {if (verbose && (c)) fprintf x ;} | ||
| 60 | # define Tracecv(c,x) {if (verbose>1 && (c)) fprintf x ;} | ||
| 61 | #else | ||
| 62 | # define Assert(cond,msg) | ||
| 63 | # define Trace(x) | ||
| 64 | # define Tracev(x) | ||
| 65 | # define Tracevv(x) | ||
| 66 | # define Tracec(c,x) | ||
| 67 | # define Tracecv(c,x) | ||
| 68 | #endif | ||
| 69 | |||
| 70 | static int fill_inbuf(void); | ||
| 71 | static void flush_window(void); | ||
| 72 | static void error(char *m); | 23 | static void error(char *m); |
| 73 | 24 | ||
| 74 | static unsigned char *input_data; | ||
| 75 | static int input_len; | ||
| 76 | |||
| 77 | static long bytes_out = 0; | ||
| 78 | static uch *output_data; | ||
| 79 | static unsigned long output_ptr = 0; | ||
| 80 | |||
| 81 | #include "m32r_sio.c" | 25 | #include "m32r_sio.c" |
| 82 | 26 | ||
| 83 | static unsigned long free_mem_ptr; | 27 | static unsigned long free_mem_ptr; |
| 84 | static unsigned long free_mem_end_ptr; | 28 | static unsigned long free_mem_end_ptr; |
| 85 | 29 | ||
| 86 | #define HEAP_SIZE 0x10000 | 30 | #ifdef CONFIG_KERNEL_BZIP2 |
| 87 | 31 | static void *memset(void *s, int c, size_t n) | |
| 88 | #include "../../../../lib/inflate.c" | ||
| 89 | |||
| 90 | void* memset(void* s, int c, size_t n) | ||
| 91 | { | 32 | { |
| 92 | int i; | 33 | char *ss = s; |
| 93 | char *ss = (char*)s; | ||
| 94 | 34 | ||
| 95 | for (i=0;i<n;i++) ss[i] = c; | 35 | while (n--) |
| 36 | *ss++ = c; | ||
| 96 | return s; | 37 | return s; |
| 97 | } | 38 | } |
| 39 | #endif | ||
| 98 | 40 | ||
| 99 | void* memcpy(void* __dest, __const void* __src, | 41 | #ifdef CONFIG_KERNEL_GZIP |
| 100 | size_t __n) | 42 | #define BOOT_HEAP_SIZE 0x10000 |
| 101 | { | 43 | #include "../../../../lib/decompress_inflate.c" |
| 102 | int i; | 44 | #endif |
| 103 | char *d = (char *)__dest, *s = (char *)__src; | ||
| 104 | |||
| 105 | for (i=0;i<__n;i++) d[i] = s[i]; | ||
| 106 | return __dest; | ||
| 107 | } | ||
| 108 | |||
| 109 | /* =========================================================================== | ||
| 110 | * Fill the input buffer. This is called only when the buffer is empty | ||
| 111 | * and at least one byte is really needed. | ||
| 112 | */ | ||
| 113 | static int fill_inbuf(void) | ||
| 114 | { | ||
| 115 | if (insize != 0) { | ||
| 116 | error("ran out of input data"); | ||
| 117 | } | ||
| 118 | |||
| 119 | inbuf = input_data; | ||
| 120 | insize = input_len; | ||
| 121 | inptr = 1; | ||
| 122 | return inbuf[0]; | ||
| 123 | } | ||
| 124 | 45 | ||
| 125 | /* =========================================================================== | 46 | #ifdef CONFIG_KERNEL_BZIP2 |
| 126 | * Write the output window window[0..outcnt-1] and update crc and bytes_out. | 47 | #define BOOT_HEAP_SIZE 0x400000 |
| 127 | * (Used for the decompressed data only.) | 48 | #include "../../../../lib/decompress_bunzip2.c" |
| 128 | */ | 49 | #endif |
| 129 | static void flush_window(void) | ||
| 130 | { | ||
| 131 | ulg c = crc; /* temporary variable */ | ||
| 132 | unsigned n; | ||
| 133 | uch *in, *out, ch; | ||
| 134 | 50 | ||
| 135 | in = window; | 51 | #ifdef CONFIG_KERNEL_LZMA |
| 136 | out = &output_data[output_ptr]; | 52 | #define BOOT_HEAP_SIZE 0x10000 |
| 137 | for (n = 0; n < outcnt; n++) { | 53 | #include "../../../../lib/decompress_unlzma.c" |
| 138 | ch = *out++ = *in++; | 54 | #endif |
| 139 | c = crc_32_tab[((int)c ^ ch) & 0xff] ^ (c >> 8); | ||
| 140 | } | ||
| 141 | crc = c; | ||
| 142 | bytes_out += (ulg)outcnt; | ||
| 143 | output_ptr += (ulg)outcnt; | ||
| 144 | outcnt = 0; | ||
| 145 | } | ||
| 146 | 55 | ||
| 147 | static void error(char *x) | 56 | static void error(char *x) |
| 148 | { | 57 | { |
| @@ -153,20 +62,20 @@ static void error(char *x) | |||
| 153 | while(1); /* Halt */ | 62 | while(1); /* Halt */ |
| 154 | } | 63 | } |
| 155 | 64 | ||
| 156 | /* return decompressed size */ | ||
| 157 | void | 65 | void |
| 158 | decompress_kernel(int mmu_on, unsigned char *zimage_data, | 66 | decompress_kernel(int mmu_on, unsigned char *zimage_data, |
| 159 | unsigned int zimage_len, unsigned long heap) | 67 | unsigned int zimage_len, unsigned long heap) |
| 160 | { | 68 | { |
| 69 | unsigned char *input_data = zimage_data; | ||
| 70 | int input_len = zimage_len; | ||
| 71 | unsigned char *output_data; | ||
| 72 | |||
| 161 | output_data = (unsigned char *)CONFIG_MEMORY_START + 0x2000 | 73 | output_data = (unsigned char *)CONFIG_MEMORY_START + 0x2000 |
| 162 | + (mmu_on ? 0x80000000 : 0); | 74 | + (mmu_on ? 0x80000000 : 0); |
| 163 | free_mem_ptr = heap; | 75 | free_mem_ptr = heap; |
| 164 | free_mem_end_ptr = free_mem_ptr + HEAP_SIZE; | 76 | free_mem_end_ptr = free_mem_ptr + BOOT_HEAP_SIZE; |
| 165 | input_data = zimage_data; | ||
| 166 | input_len = zimage_len; | ||
| 167 | 77 | ||
| 168 | makecrc(); | 78 | puts("\nDecompressing Linux... "); |
| 169 | puts("Uncompressing Linux... "); | 79 | decompress(input_data, input_len, NULL, NULL, output_data, NULL, error); |
| 170 | gunzip(); | 80 | puts("done.\nBooting the kernel.\n"); |
| 171 | puts("Ok, booting the kernel.\n"); | ||
| 172 | } | 81 | } |
diff --git a/arch/m32r/kernel/smp.c b/arch/m32r/kernel/smp.c index 8a88f1f0a3e2..31cef20b2996 100644 --- a/arch/m32r/kernel/smp.c +++ b/arch/m32r/kernel/smp.c | |||
| @@ -806,7 +806,7 @@ unsigned long send_IPI_mask_phys(cpumask_t physid_mask, int ipi_num, | |||
| 806 | 806 | ||
| 807 | if (mask & ~physids_coerce(phys_cpu_present_map)) | 807 | if (mask & ~physids_coerce(phys_cpu_present_map)) |
| 808 | BUG(); | 808 | BUG(); |
| 809 | if (ipi_num >= NR_IPIS) | 809 | if (ipi_num >= NR_IPIS || ipi_num < 0) |
| 810 | BUG(); | 810 | BUG(); |
| 811 | 811 | ||
| 812 | mask <<= IPI_SHIFT; | 812 | mask <<= IPI_SHIFT; |
diff --git a/arch/m32r/kernel/time.c b/arch/m32r/kernel/time.c index e7fee0f198d5..9cedcef11575 100644 --- a/arch/m32r/kernel/time.c +++ b/arch/m32r/kernel/time.c | |||
| @@ -75,7 +75,7 @@ u32 arch_gettimeoffset(void) | |||
| 75 | count = 0; | 75 | count = 0; |
| 76 | 76 | ||
| 77 | count = (latch - count) * TICK_SIZE; | 77 | count = (latch - count) * TICK_SIZE; |
| 78 | elapsed_time = (count + latch / 2) / latch; | 78 | elapsed_time = DIV_ROUND_CLOSEST(count, latch); |
| 79 | /* NOTE: LATCH is equal to the "interval" value (= reload count). */ | 79 | /* NOTE: LATCH is equal to the "interval" value (= reload count). */ |
| 80 | 80 | ||
| 81 | #else /* CONFIG_SMP */ | 81 | #else /* CONFIG_SMP */ |
| @@ -93,7 +93,7 @@ u32 arch_gettimeoffset(void) | |||
| 93 | p_count = count; | 93 | p_count = count; |
| 94 | 94 | ||
| 95 | count = (latch - count) * TICK_SIZE; | 95 | count = (latch - count) * TICK_SIZE; |
| 96 | elapsed_time = (count + latch / 2) / latch; | 96 | elapsed_time = DIV_ROUND_CLOSEST(count, latch); |
| 97 | /* NOTE: LATCH is equal to the "interval" value (= reload count). */ | 97 | /* NOTE: LATCH is equal to the "interval" value (= reload count). */ |
| 98 | #endif /* CONFIG_SMP */ | 98 | #endif /* CONFIG_SMP */ |
| 99 | #elif defined(CONFIG_CHIP_M32310) | 99 | #elif defined(CONFIG_CHIP_M32310) |
| @@ -211,7 +211,7 @@ void __init time_init(void) | |||
| 211 | 211 | ||
| 212 | bus_clock = boot_cpu_data.bus_clock; | 212 | bus_clock = boot_cpu_data.bus_clock; |
| 213 | divide = boot_cpu_data.timer_divide; | 213 | divide = boot_cpu_data.timer_divide; |
| 214 | latch = (bus_clock/divide + HZ / 2) / HZ; | 214 | latch = DIV_ROUND_CLOSEST(bus_clock/divide, HZ); |
| 215 | 215 | ||
| 216 | printk("Timer start : latch = %ld\n", latch); | 216 | printk("Timer start : latch = %ld\n", latch); |
| 217 | 217 | ||
diff --git a/arch/m32r/kernel/vmlinux.lds.S b/arch/m32r/kernel/vmlinux.lds.S index 8ceb6181d805..7da94eaa082b 100644 --- a/arch/m32r/kernel/vmlinux.lds.S +++ b/arch/m32r/kernel/vmlinux.lds.S | |||
| @@ -42,6 +42,8 @@ SECTIONS | |||
| 42 | _etext = .; /* End of text section */ | 42 | _etext = .; /* End of text section */ |
| 43 | 43 | ||
| 44 | EXCEPTION_TABLE(16) | 44 | EXCEPTION_TABLE(16) |
| 45 | NOTES | ||
| 46 | |||
| 45 | RODATA | 47 | RODATA |
| 46 | RW_DATA_SECTION(32, PAGE_SIZE, THREAD_SIZE) | 48 | RW_DATA_SECTION(32, PAGE_SIZE, THREAD_SIZE) |
| 47 | _edata = .; /* End of data section */ | 49 | _edata = .; /* End of data section */ |
diff --git a/arch/powerpc/Kconfig b/arch/powerpc/Kconfig index 10a0a5488a44..2ba14e77296c 100644 --- a/arch/powerpc/Kconfig +++ b/arch/powerpc/Kconfig | |||
| @@ -414,6 +414,10 @@ config ARCH_SPARSEMEM_DEFAULT | |||
| 414 | config ARCH_POPULATES_NODE_MAP | 414 | config ARCH_POPULATES_NODE_MAP |
| 415 | def_bool y | 415 | def_bool y |
| 416 | 416 | ||
| 417 | config SYS_SUPPORTS_HUGETLBFS | ||
| 418 | def_bool y | ||
| 419 | depends on PPC_BOOK3S_64 | ||
| 420 | |||
| 417 | source "mm/Kconfig" | 421 | source "mm/Kconfig" |
| 418 | 422 | ||
| 419 | config ARCH_MEMORY_PROBE | 423 | config ARCH_MEMORY_PROBE |
diff --git a/arch/powerpc/kernel/time.c b/arch/powerpc/kernel/time.c index 92dc844299b6..a136a11c490d 100644 --- a/arch/powerpc/kernel/time.c +++ b/arch/powerpc/kernel/time.c | |||
| @@ -777,7 +777,7 @@ int update_persistent_clock(struct timespec now) | |||
| 777 | return ppc_md.set_rtc_time(&tm); | 777 | return ppc_md.set_rtc_time(&tm); |
| 778 | } | 778 | } |
| 779 | 779 | ||
| 780 | void read_persistent_clock(struct timespec *ts) | 780 | static void __read_persistent_clock(struct timespec *ts) |
| 781 | { | 781 | { |
| 782 | struct rtc_time tm; | 782 | struct rtc_time tm; |
| 783 | static int first = 1; | 783 | static int first = 1; |
| @@ -800,10 +800,23 @@ void read_persistent_clock(struct timespec *ts) | |||
| 800 | return; | 800 | return; |
| 801 | } | 801 | } |
| 802 | ppc_md.get_rtc_time(&tm); | 802 | ppc_md.get_rtc_time(&tm); |
| 803 | |||
| 803 | ts->tv_sec = mktime(tm.tm_year+1900, tm.tm_mon+1, tm.tm_mday, | 804 | ts->tv_sec = mktime(tm.tm_year+1900, tm.tm_mon+1, tm.tm_mday, |
| 804 | tm.tm_hour, tm.tm_min, tm.tm_sec); | 805 | tm.tm_hour, tm.tm_min, tm.tm_sec); |
| 805 | } | 806 | } |
| 806 | 807 | ||
| 808 | void read_persistent_clock(struct timespec *ts) | ||
| 809 | { | ||
| 810 | __read_persistent_clock(ts); | ||
| 811 | |||
| 812 | /* Sanitize it in case real time clock is set below EPOCH */ | ||
| 813 | if (ts->tv_sec < 0) { | ||
| 814 | ts->tv_sec = 0; | ||
| 815 | ts->tv_nsec = 0; | ||
| 816 | } | ||
| 817 | |||
| 818 | } | ||
| 819 | |||
| 807 | /* clocksource code */ | 820 | /* clocksource code */ |
| 808 | static cycle_t rtc_read(struct clocksource *cs) | 821 | static cycle_t rtc_read(struct clocksource *cs) |
| 809 | { | 822 | { |
diff --git a/arch/powerpc/kvm/timing.h b/arch/powerpc/kvm/timing.h index bb13b1f3cd5a..806ef67868bd 100644 --- a/arch/powerpc/kvm/timing.h +++ b/arch/powerpc/kvm/timing.h | |||
| @@ -48,7 +48,11 @@ static inline void kvmppc_set_exit_type(struct kvm_vcpu *vcpu, int type) {} | |||
| 48 | static inline void kvmppc_account_exit_stat(struct kvm_vcpu *vcpu, int type) | 48 | static inline void kvmppc_account_exit_stat(struct kvm_vcpu *vcpu, int type) |
| 49 | { | 49 | { |
| 50 | /* type has to be known at build time for optimization */ | 50 | /* type has to be known at build time for optimization */ |
| 51 | |||
| 52 | /* The BUILD_BUG_ON below breaks in funny ways, commented out | ||
| 53 | * for now ... -BenH | ||
| 51 | BUILD_BUG_ON(__builtin_constant_p(type)); | 54 | BUILD_BUG_ON(__builtin_constant_p(type)); |
| 55 | */ | ||
| 52 | switch (type) { | 56 | switch (type) { |
| 53 | case EXT_INTR_EXITS: | 57 | case EXT_INTR_EXITS: |
| 54 | vcpu->stat.ext_intr_exits++; | 58 | vcpu->stat.ext_intr_exits++; |
diff --git a/arch/powerpc/mm/mmu_context_nohash.c b/arch/powerpc/mm/mmu_context_nohash.c index c2f93dc470e6..be4f34c30a0b 100644 --- a/arch/powerpc/mm/mmu_context_nohash.c +++ b/arch/powerpc/mm/mmu_context_nohash.c | |||
| @@ -25,8 +25,8 @@ | |||
| 25 | * also clear mm->cpu_vm_mask bits when processes are migrated | 25 | * also clear mm->cpu_vm_mask bits when processes are migrated |
| 26 | */ | 26 | */ |
| 27 | 27 | ||
| 28 | #define DEBUG_MAP_CONSISTENCY | 28 | //#define DEBUG_MAP_CONSISTENCY |
| 29 | #define DEBUG_CLAMP_LAST_CONTEXT 31 | 29 | //#define DEBUG_CLAMP_LAST_CONTEXT 31 |
| 30 | //#define DEBUG_HARDER | 30 | //#define DEBUG_HARDER |
| 31 | 31 | ||
| 32 | /* We don't use DEBUG because it tends to be compiled in always nowadays | 32 | /* We don't use DEBUG because it tends to be compiled in always nowadays |
diff --git a/arch/powerpc/platforms/pseries/msi.c b/arch/powerpc/platforms/pseries/msi.c index bf2e1ac41308..1164c3430f2c 100644 --- a/arch/powerpc/platforms/pseries/msi.c +++ b/arch/powerpc/platforms/pseries/msi.c | |||
| @@ -432,8 +432,6 @@ static int rtas_setup_msi_irqs(struct pci_dev *pdev, int nvec, int type) | |||
| 432 | /* Read config space back so we can restore after reset */ | 432 | /* Read config space back so we can restore after reset */ |
| 433 | read_msi_msg(virq, &msg); | 433 | read_msi_msg(virq, &msg); |
| 434 | entry->msg = msg; | 434 | entry->msg = msg; |
| 435 | |||
| 436 | unmask_msi_irq(virq); | ||
| 437 | } | 435 | } |
| 438 | 436 | ||
| 439 | return 0; | 437 | return 0; |
diff --git a/arch/powerpc/platforms/pseries/xics.c b/arch/powerpc/platforms/pseries/xics.c index 419f8a637ffe..b9bf0eedccf2 100644 --- a/arch/powerpc/platforms/pseries/xics.c +++ b/arch/powerpc/platforms/pseries/xics.c | |||
| @@ -18,6 +18,7 @@ | |||
| 18 | #include <linux/init.h> | 18 | #include <linux/init.h> |
| 19 | #include <linux/radix-tree.h> | 19 | #include <linux/radix-tree.h> |
| 20 | #include <linux/cpu.h> | 20 | #include <linux/cpu.h> |
| 21 | #include <linux/msi.h> | ||
| 21 | #include <linux/of.h> | 22 | #include <linux/of.h> |
| 22 | 23 | ||
| 23 | #include <asm/firmware.h> | 24 | #include <asm/firmware.h> |
| @@ -219,6 +220,14 @@ static void xics_unmask_irq(unsigned int virq) | |||
| 219 | 220 | ||
| 220 | static unsigned int xics_startup(unsigned int virq) | 221 | static unsigned int xics_startup(unsigned int virq) |
| 221 | { | 222 | { |
| 223 | /* | ||
| 224 | * The generic MSI code returns with the interrupt disabled on the | ||
| 225 | * card, using the MSI mask bits. Firmware doesn't appear to unmask | ||
| 226 | * at that level, so we do it here by hand. | ||
| 227 | */ | ||
| 228 | if (irq_to_desc(virq)->msi_desc) | ||
| 229 | unmask_msi_irq(virq); | ||
| 230 | |||
| 222 | /* unmask it */ | 231 | /* unmask it */ |
| 223 | xics_unmask_irq(virq); | 232 | xics_unmask_irq(virq); |
| 224 | return 0; | 233 | return 0; |
diff --git a/arch/x86/include/asm/amd_iommu.h b/arch/x86/include/asm/amd_iommu.h index ac95995b7bad..4b180897e6b5 100644 --- a/arch/x86/include/asm/amd_iommu.h +++ b/arch/x86/include/asm/amd_iommu.h | |||
| @@ -31,6 +31,7 @@ extern irqreturn_t amd_iommu_int_handler(int irq, void *data); | |||
| 31 | extern void amd_iommu_flush_all_domains(void); | 31 | extern void amd_iommu_flush_all_domains(void); |
| 32 | extern void amd_iommu_flush_all_devices(void); | 32 | extern void amd_iommu_flush_all_devices(void); |
| 33 | extern void amd_iommu_shutdown(void); | 33 | extern void amd_iommu_shutdown(void); |
| 34 | extern void amd_iommu_apply_erratum_63(u16 devid); | ||
| 34 | #else | 35 | #else |
| 35 | static inline int amd_iommu_init(void) { return -ENODEV; } | 36 | static inline int amd_iommu_init(void) { return -ENODEV; } |
| 36 | static inline void amd_iommu_detect(void) { } | 37 | static inline void amd_iommu_detect(void) { } |
diff --git a/arch/x86/include/asm/desc.h b/arch/x86/include/asm/desc.h index e8de2f6f5ca5..617bd56b3070 100644 --- a/arch/x86/include/asm/desc.h +++ b/arch/x86/include/asm/desc.h | |||
| @@ -288,7 +288,7 @@ static inline void load_LDT(mm_context_t *pc) | |||
| 288 | 288 | ||
| 289 | static inline unsigned long get_desc_base(const struct desc_struct *desc) | 289 | static inline unsigned long get_desc_base(const struct desc_struct *desc) |
| 290 | { | 290 | { |
| 291 | return desc->base0 | ((desc->base1) << 16) | ((desc->base2) << 24); | 291 | return (unsigned)(desc->base0 | ((desc->base1) << 16) | ((desc->base2) << 24)); |
| 292 | } | 292 | } |
| 293 | 293 | ||
| 294 | static inline void set_desc_base(struct desc_struct *desc, unsigned long base) | 294 | static inline void set_desc_base(struct desc_struct *desc, unsigned long base) |
diff --git a/arch/x86/include/asm/processor.h b/arch/x86/include/asm/processor.h index c3429e8b2424..c9786480f0fe 100644 --- a/arch/x86/include/asm/processor.h +++ b/arch/x86/include/asm/processor.h | |||
| @@ -1000,7 +1000,7 @@ extern unsigned long thread_saved_pc(struct task_struct *tsk); | |||
| 1000 | #define thread_saved_pc(t) (*(unsigned long *)((t)->thread.sp - 8)) | 1000 | #define thread_saved_pc(t) (*(unsigned long *)((t)->thread.sp - 8)) |
| 1001 | 1001 | ||
| 1002 | #define task_pt_regs(tsk) ((struct pt_regs *)(tsk)->thread.sp0 - 1) | 1002 | #define task_pt_regs(tsk) ((struct pt_regs *)(tsk)->thread.sp0 - 1) |
| 1003 | #define KSTK_ESP(tsk) -1 /* sorry. doesn't work for syscall. */ | 1003 | extern unsigned long KSTK_ESP(struct task_struct *task); |
| 1004 | #endif /* CONFIG_X86_64 */ | 1004 | #endif /* CONFIG_X86_64 */ |
| 1005 | 1005 | ||
| 1006 | extern void start_thread(struct pt_regs *regs, unsigned long new_ip, | 1006 | extern void start_thread(struct pt_regs *regs, unsigned long new_ip, |
diff --git a/arch/x86/include/asm/topology.h b/arch/x86/include/asm/topology.h index d823c245f63b..40e37b10c6c0 100644 --- a/arch/x86/include/asm/topology.h +++ b/arch/x86/include/asm/topology.h | |||
| @@ -143,7 +143,7 @@ extern unsigned long node_remap_size[]; | |||
| 143 | | 1*SD_BALANCE_FORK \ | 143 | | 1*SD_BALANCE_FORK \ |
| 144 | | 0*SD_BALANCE_WAKE \ | 144 | | 0*SD_BALANCE_WAKE \ |
| 145 | | 1*SD_WAKE_AFFINE \ | 145 | | 1*SD_WAKE_AFFINE \ |
| 146 | | 1*SD_PREFER_LOCAL \ | 146 | | 0*SD_PREFER_LOCAL \ |
| 147 | | 0*SD_SHARE_CPUPOWER \ | 147 | | 0*SD_SHARE_CPUPOWER \ |
| 148 | | 0*SD_POWERSAVINGS_BALANCE \ | 148 | | 0*SD_POWERSAVINGS_BALANCE \ |
| 149 | | 0*SD_SHARE_PKG_RESOURCES \ | 149 | | 0*SD_SHARE_PKG_RESOURCES \ |
diff --git a/arch/x86/kernel/amd_iommu.c b/arch/x86/kernel/amd_iommu.c index 98f230f6a28d..0285521e0a99 100644 --- a/arch/x86/kernel/amd_iommu.c +++ b/arch/x86/kernel/amd_iommu.c | |||
| @@ -1220,6 +1220,8 @@ static void __detach_device(struct protection_domain *domain, u16 devid) | |||
| 1220 | amd_iommu_dev_table[devid].data[1] = 0; | 1220 | amd_iommu_dev_table[devid].data[1] = 0; |
| 1221 | amd_iommu_dev_table[devid].data[2] = 0; | 1221 | amd_iommu_dev_table[devid].data[2] = 0; |
| 1222 | 1222 | ||
| 1223 | amd_iommu_apply_erratum_63(devid); | ||
| 1224 | |||
| 1223 | /* decrease reference counter */ | 1225 | /* decrease reference counter */ |
| 1224 | domain->dev_cnt -= 1; | 1226 | domain->dev_cnt -= 1; |
| 1225 | 1227 | ||
diff --git a/arch/x86/kernel/amd_iommu_init.c b/arch/x86/kernel/amd_iommu_init.c index b4b61d462dcc..c20001e4f556 100644 --- a/arch/x86/kernel/amd_iommu_init.c +++ b/arch/x86/kernel/amd_iommu_init.c | |||
| @@ -240,7 +240,7 @@ static void iommu_feature_enable(struct amd_iommu *iommu, u8 bit) | |||
| 240 | writel(ctrl, iommu->mmio_base + MMIO_CONTROL_OFFSET); | 240 | writel(ctrl, iommu->mmio_base + MMIO_CONTROL_OFFSET); |
| 241 | } | 241 | } |
| 242 | 242 | ||
| 243 | static void __init iommu_feature_disable(struct amd_iommu *iommu, u8 bit) | 243 | static void iommu_feature_disable(struct amd_iommu *iommu, u8 bit) |
| 244 | { | 244 | { |
| 245 | u32 ctrl; | 245 | u32 ctrl; |
| 246 | 246 | ||
| @@ -519,6 +519,26 @@ static void set_dev_entry_bit(u16 devid, u8 bit) | |||
| 519 | amd_iommu_dev_table[devid].data[i] |= (1 << _bit); | 519 | amd_iommu_dev_table[devid].data[i] |= (1 << _bit); |
| 520 | } | 520 | } |
| 521 | 521 | ||
| 522 | static int get_dev_entry_bit(u16 devid, u8 bit) | ||
| 523 | { | ||
| 524 | int i = (bit >> 5) & 0x07; | ||
| 525 | int _bit = bit & 0x1f; | ||
| 526 | |||
| 527 | return (amd_iommu_dev_table[devid].data[i] & (1 << _bit)) >> _bit; | ||
| 528 | } | ||
| 529 | |||
| 530 | |||
| 531 | void amd_iommu_apply_erratum_63(u16 devid) | ||
| 532 | { | ||
| 533 | int sysmgt; | ||
| 534 | |||
| 535 | sysmgt = get_dev_entry_bit(devid, DEV_ENTRY_SYSMGT1) | | ||
| 536 | (get_dev_entry_bit(devid, DEV_ENTRY_SYSMGT2) << 1); | ||
| 537 | |||
| 538 | if (sysmgt == 0x01) | ||
| 539 | set_dev_entry_bit(devid, DEV_ENTRY_IW); | ||
| 540 | } | ||
| 541 | |||
| 522 | /* Writes the specific IOMMU for a device into the rlookup table */ | 542 | /* Writes the specific IOMMU for a device into the rlookup table */ |
| 523 | static void __init set_iommu_for_device(struct amd_iommu *iommu, u16 devid) | 543 | static void __init set_iommu_for_device(struct amd_iommu *iommu, u16 devid) |
| 524 | { | 544 | { |
| @@ -547,6 +567,8 @@ static void __init set_dev_entry_from_acpi(struct amd_iommu *iommu, | |||
| 547 | if (flags & ACPI_DEVFLAG_LINT1) | 567 | if (flags & ACPI_DEVFLAG_LINT1) |
| 548 | set_dev_entry_bit(devid, DEV_ENTRY_LINT1_PASS); | 568 | set_dev_entry_bit(devid, DEV_ENTRY_LINT1_PASS); |
| 549 | 569 | ||
| 570 | amd_iommu_apply_erratum_63(devid); | ||
| 571 | |||
| 550 | set_iommu_for_device(iommu, devid); | 572 | set_iommu_for_device(iommu, devid); |
| 551 | } | 573 | } |
| 552 | 574 | ||
diff --git a/arch/x86/kernel/cpu/mtrr/cleanup.c b/arch/x86/kernel/cpu/mtrr/cleanup.c index 315738c74aad..73c86db5acbe 100644 --- a/arch/x86/kernel/cpu/mtrr/cleanup.c +++ b/arch/x86/kernel/cpu/mtrr/cleanup.c | |||
| @@ -846,7 +846,7 @@ int __init mtrr_cleanup(unsigned address_bits) | |||
| 846 | sort(range, nr_range, sizeof(struct res_range), cmp_range, NULL); | 846 | sort(range, nr_range, sizeof(struct res_range), cmp_range, NULL); |
| 847 | 847 | ||
| 848 | range_sums = sum_ranges(range, nr_range); | 848 | range_sums = sum_ranges(range, nr_range); |
| 849 | printk(KERN_INFO "total RAM coverred: %ldM\n", | 849 | printk(KERN_INFO "total RAM covered: %ldM\n", |
| 850 | range_sums >> (20 - PAGE_SHIFT)); | 850 | range_sums >> (20 - PAGE_SHIFT)); |
| 851 | 851 | ||
| 852 | if (mtrr_chunk_size && mtrr_gran_size) { | 852 | if (mtrr_chunk_size && mtrr_gran_size) { |
diff --git a/arch/x86/kernel/process_64.c b/arch/x86/kernel/process_64.c index ad535b683170..eb62cbcaa490 100644 --- a/arch/x86/kernel/process_64.c +++ b/arch/x86/kernel/process_64.c | |||
| @@ -664,3 +664,8 @@ long sys_arch_prctl(int code, unsigned long addr) | |||
| 664 | return do_arch_prctl(current, code, addr); | 664 | return do_arch_prctl(current, code, addr); |
| 665 | } | 665 | } |
| 666 | 666 | ||
| 667 | unsigned long KSTK_ESP(struct task_struct *task) | ||
| 668 | { | ||
| 669 | return (test_tsk_thread_flag(task, TIF_IA32)) ? | ||
| 670 | (task_pt_regs(task)->sp) : ((task)->thread.usersp); | ||
| 671 | } | ||
diff --git a/arch/x86/kernel/reboot.c b/arch/x86/kernel/reboot.c index a1a3cdda06e1..f93078746e00 100644 --- a/arch/x86/kernel/reboot.c +++ b/arch/x86/kernel/reboot.c | |||
| @@ -436,6 +436,14 @@ static struct dmi_system_id __initdata pci_reboot_dmi_table[] = { | |||
| 436 | DMI_MATCH(DMI_PRODUCT_NAME, "MacBookPro5"), | 436 | DMI_MATCH(DMI_PRODUCT_NAME, "MacBookPro5"), |
| 437 | }, | 437 | }, |
| 438 | }, | 438 | }, |
| 439 | { /* Handle problems with rebooting on Apple Macmini3,1 */ | ||
| 440 | .callback = set_pci_reboot, | ||
| 441 | .ident = "Apple Macmini3,1", | ||
| 442 | .matches = { | ||
| 443 | DMI_MATCH(DMI_SYS_VENDOR, "Apple Inc."), | ||
| 444 | DMI_MATCH(DMI_PRODUCT_NAME, "Macmini3,1"), | ||
| 445 | }, | ||
| 446 | }, | ||
| 439 | { } | 447 | { } |
| 440 | }; | 448 | }; |
| 441 | 449 | ||
diff --git a/arch/x86/kvm/x86.c b/arch/x86/kvm/x86.c index 9b9695322f56..ae07d261527c 100644 --- a/arch/x86/kvm/x86.c +++ b/arch/x86/kvm/x86.c | |||
| @@ -1692,7 +1692,7 @@ static int kvm_vcpu_ioctl_x86_setup_mce(struct kvm_vcpu *vcpu, | |||
| 1692 | unsigned bank_num = mcg_cap & 0xff, bank; | 1692 | unsigned bank_num = mcg_cap & 0xff, bank; |
| 1693 | 1693 | ||
| 1694 | r = -EINVAL; | 1694 | r = -EINVAL; |
| 1695 | if (!bank_num) | 1695 | if (!bank_num || bank_num >= KVM_MAX_MCE_BANKS) |
| 1696 | goto out; | 1696 | goto out; |
| 1697 | if (mcg_cap & ~(KVM_MCE_CAP_SUPPORTED | 0xff | 0xff0000)) | 1697 | if (mcg_cap & ~(KVM_MCE_CAP_SUPPORTED | 0xff | 0xff0000)) |
| 1698 | goto out; | 1698 | goto out; |
| @@ -4051,7 +4051,7 @@ static int save_guest_segment_descriptor(struct kvm_vcpu *vcpu, u16 selector, | |||
| 4051 | return kvm_write_guest_virt(dtable.base + index*8, seg_desc, sizeof(*seg_desc), vcpu); | 4051 | return kvm_write_guest_virt(dtable.base + index*8, seg_desc, sizeof(*seg_desc), vcpu); |
| 4052 | } | 4052 | } |
| 4053 | 4053 | ||
| 4054 | static u32 get_tss_base_addr(struct kvm_vcpu *vcpu, | 4054 | static gpa_t get_tss_base_addr(struct kvm_vcpu *vcpu, |
| 4055 | struct desc_struct *seg_desc) | 4055 | struct desc_struct *seg_desc) |
| 4056 | { | 4056 | { |
| 4057 | u32 base_addr = get_desc_base(seg_desc); | 4057 | u32 base_addr = get_desc_base(seg_desc); |
diff --git a/arch/x86/xen/enlighten.c b/arch/x86/xen/enlighten.c index 23a4d80fb39e..dfbf70e65860 100644 --- a/arch/x86/xen/enlighten.c +++ b/arch/x86/xen/enlighten.c | |||
| @@ -178,6 +178,7 @@ static __read_mostly unsigned int cpuid_leaf1_ecx_mask = ~0; | |||
| 178 | static void xen_cpuid(unsigned int *ax, unsigned int *bx, | 178 | static void xen_cpuid(unsigned int *ax, unsigned int *bx, |
| 179 | unsigned int *cx, unsigned int *dx) | 179 | unsigned int *cx, unsigned int *dx) |
| 180 | { | 180 | { |
| 181 | unsigned maskebx = ~0; | ||
| 181 | unsigned maskecx = ~0; | 182 | unsigned maskecx = ~0; |
| 182 | unsigned maskedx = ~0; | 183 | unsigned maskedx = ~0; |
| 183 | 184 | ||
| @@ -185,9 +186,16 @@ static void xen_cpuid(unsigned int *ax, unsigned int *bx, | |||
| 185 | * Mask out inconvenient features, to try and disable as many | 186 | * Mask out inconvenient features, to try and disable as many |
| 186 | * unsupported kernel subsystems as possible. | 187 | * unsupported kernel subsystems as possible. |
| 187 | */ | 188 | */ |
| 188 | if (*ax == 1) { | 189 | switch (*ax) { |
| 190 | case 1: | ||
| 189 | maskecx = cpuid_leaf1_ecx_mask; | 191 | maskecx = cpuid_leaf1_ecx_mask; |
| 190 | maskedx = cpuid_leaf1_edx_mask; | 192 | maskedx = cpuid_leaf1_edx_mask; |
| 193 | break; | ||
| 194 | |||
| 195 | case 0xb: | ||
| 196 | /* Suppress extended topology stuff */ | ||
| 197 | maskebx = 0; | ||
| 198 | break; | ||
| 191 | } | 199 | } |
| 192 | 200 | ||
| 193 | asm(XEN_EMULATE_PREFIX "cpuid" | 201 | asm(XEN_EMULATE_PREFIX "cpuid" |
| @@ -197,6 +205,7 @@ static void xen_cpuid(unsigned int *ax, unsigned int *bx, | |||
| 197 | "=d" (*dx) | 205 | "=d" (*dx) |
| 198 | : "0" (*ax), "2" (*cx)); | 206 | : "0" (*ax), "2" (*cx)); |
| 199 | 207 | ||
| 208 | *bx &= maskebx; | ||
| 200 | *cx &= maskecx; | 209 | *cx &= maskecx; |
| 201 | *dx &= maskedx; | 210 | *dx &= maskedx; |
| 202 | } | 211 | } |
diff --git a/drivers/char/hvc_xen.c b/drivers/char/hvc_xen.c index eba999f8598d..a6ee32b599a8 100644 --- a/drivers/char/hvc_xen.c +++ b/drivers/char/hvc_xen.c | |||
| @@ -55,7 +55,7 @@ static inline void notify_daemon(void) | |||
| 55 | notify_remote_via_evtchn(xen_start_info->console.domU.evtchn); | 55 | notify_remote_via_evtchn(xen_start_info->console.domU.evtchn); |
| 56 | } | 56 | } |
| 57 | 57 | ||
| 58 | static int write_console(uint32_t vtermno, const char *data, int len) | 58 | static int __write_console(const char *data, int len) |
| 59 | { | 59 | { |
| 60 | struct xencons_interface *intf = xencons_interface(); | 60 | struct xencons_interface *intf = xencons_interface(); |
| 61 | XENCONS_RING_IDX cons, prod; | 61 | XENCONS_RING_IDX cons, prod; |
| @@ -76,6 +76,29 @@ static int write_console(uint32_t vtermno, const char *data, int len) | |||
| 76 | return sent; | 76 | return sent; |
| 77 | } | 77 | } |
| 78 | 78 | ||
| 79 | static int write_console(uint32_t vtermno, const char *data, int len) | ||
| 80 | { | ||
| 81 | int ret = len; | ||
| 82 | |||
| 83 | /* | ||
| 84 | * Make sure the whole buffer is emitted, polling if | ||
| 85 | * necessary. We don't ever want to rely on the hvc daemon | ||
| 86 | * because the most interesting console output is when the | ||
| 87 | * kernel is crippled. | ||
| 88 | */ | ||
| 89 | while (len) { | ||
| 90 | int sent = __write_console(data, len); | ||
| 91 | |||
| 92 | data += sent; | ||
| 93 | len -= sent; | ||
| 94 | |||
| 95 | if (unlikely(len)) | ||
| 96 | HYPERVISOR_sched_op(SCHEDOP_yield, NULL); | ||
| 97 | } | ||
| 98 | |||
| 99 | return ret; | ||
| 100 | } | ||
| 101 | |||
| 79 | static int read_console(uint32_t vtermno, char *buf, int len) | 102 | static int read_console(uint32_t vtermno, char *buf, int len) |
| 80 | { | 103 | { |
| 81 | struct xencons_interface *intf = xencons_interface(); | 104 | struct xencons_interface *intf = xencons_interface(); |
diff --git a/drivers/edac/amd64_edac.c b/drivers/edac/amd64_edac.c index d4560d9d5a83..a38831c82649 100644 --- a/drivers/edac/amd64_edac.c +++ b/drivers/edac/amd64_edac.c | |||
| @@ -2254,7 +2254,7 @@ static inline void __amd64_decode_bus_error(struct mem_ctl_info *mci, | |||
| 2254 | { | 2254 | { |
| 2255 | u32 ec = ERROR_CODE(info->nbsl); | 2255 | u32 ec = ERROR_CODE(info->nbsl); |
| 2256 | u32 xec = EXT_ERROR_CODE(info->nbsl); | 2256 | u32 xec = EXT_ERROR_CODE(info->nbsl); |
| 2257 | int ecc_type = info->nbsh & (0x3 << 13); | 2257 | int ecc_type = (info->nbsh >> 13) & 0x3; |
| 2258 | 2258 | ||
| 2259 | /* Bail early out if this was an 'observed' error */ | 2259 | /* Bail early out if this was an 'observed' error */ |
| 2260 | if (PP(ec) == K8_NBSL_PP_OBS) | 2260 | if (PP(ec) == K8_NBSL_PP_OBS) |
| @@ -3163,7 +3163,7 @@ static int __init amd64_edac_init(void) | |||
| 3163 | opstate_init(); | 3163 | opstate_init(); |
| 3164 | 3164 | ||
| 3165 | if (cache_k8_northbridges() < 0) | 3165 | if (cache_k8_northbridges() < 0) |
| 3166 | goto err_exit; | 3166 | return err; |
| 3167 | 3167 | ||
| 3168 | err = pci_register_driver(&amd64_pci_driver); | 3168 | err = pci_register_driver(&amd64_pci_driver); |
| 3169 | if (err) | 3169 | if (err) |
| @@ -3189,8 +3189,6 @@ static int __init amd64_edac_init(void) | |||
| 3189 | 3189 | ||
| 3190 | err_2nd_stage: | 3190 | err_2nd_stage: |
| 3191 | debugf0("2nd stage failed\n"); | 3191 | debugf0("2nd stage failed\n"); |
| 3192 | |||
| 3193 | err_exit: | ||
| 3194 | pci_unregister_driver(&amd64_pci_driver); | 3192 | pci_unregister_driver(&amd64_pci_driver); |
| 3195 | 3193 | ||
| 3196 | return err; | 3194 | return err; |
diff --git a/drivers/misc/sgi-gru/gruprocfs.c b/drivers/misc/sgi-gru/gruprocfs.c index ccd4408a26c7..3f2375c5ba5b 100644 --- a/drivers/misc/sgi-gru/gruprocfs.c +++ b/drivers/misc/sgi-gru/gruprocfs.c | |||
| @@ -161,14 +161,15 @@ static int options_show(struct seq_file *s, void *p) | |||
| 161 | static ssize_t options_write(struct file *file, const char __user *userbuf, | 161 | static ssize_t options_write(struct file *file, const char __user *userbuf, |
| 162 | size_t count, loff_t *data) | 162 | size_t count, loff_t *data) |
| 163 | { | 163 | { |
| 164 | unsigned long val; | 164 | char buf[20]; |
| 165 | char buf[80]; | ||
| 166 | 165 | ||
| 167 | if (strncpy_from_user(buf, userbuf, sizeof(buf) - 1) < 0) | 166 | if (count >= sizeof(buf)) |
| 167 | return -EINVAL; | ||
| 168 | if (copy_from_user(buf, userbuf, count)) | ||
| 168 | return -EFAULT; | 169 | return -EFAULT; |
| 169 | buf[count - 1] = '\0'; | 170 | buf[count] = '\0'; |
| 170 | if (!strict_strtoul(buf, 10, &val)) | 171 | if (strict_strtoul(buf, 0, &gru_options)) |
| 171 | gru_options = val; | 172 | return -EINVAL; |
| 172 | 173 | ||
| 173 | return count; | 174 | return count; |
| 174 | } | 175 | } |
diff --git a/drivers/mtd/maps/Kconfig b/drivers/mtd/maps/Kconfig index 841e085ab74a..14be0755d7cd 100644 --- a/drivers/mtd/maps/Kconfig +++ b/drivers/mtd/maps/Kconfig | |||
| @@ -486,6 +486,7 @@ config MTD_BFIN_ASYNC | |||
| 486 | 486 | ||
| 487 | config MTD_GPIO_ADDR | 487 | config MTD_GPIO_ADDR |
| 488 | tristate "GPIO-assisted Flash Chip Support" | 488 | tristate "GPIO-assisted Flash Chip Support" |
| 489 | depends on GENERIC_GPIO || GPIOLIB | ||
| 489 | depends on MTD_COMPLEX_MAPPINGS | 490 | depends on MTD_COMPLEX_MAPPINGS |
| 490 | select MTD_PARTITIONS | 491 | select MTD_PARTITIONS |
| 491 | help | 492 | help |
diff --git a/drivers/mtd/maps/gpio-addr-flash.c b/drivers/mtd/maps/gpio-addr-flash.c index 44ef9a49a860..1ad5caf9fe69 100644 --- a/drivers/mtd/maps/gpio-addr-flash.c +++ b/drivers/mtd/maps/gpio-addr-flash.c | |||
| @@ -13,7 +13,9 @@ | |||
| 13 | * Licensed under the GPL-2 or later. | 13 | * Licensed under the GPL-2 or later. |
| 14 | */ | 14 | */ |
| 15 | 15 | ||
| 16 | #include <linux/gpio.h> | ||
| 16 | #include <linux/init.h> | 17 | #include <linux/init.h> |
| 18 | #include <linux/io.h> | ||
| 17 | #include <linux/kernel.h> | 19 | #include <linux/kernel.h> |
| 18 | #include <linux/module.h> | 20 | #include <linux/module.h> |
| 19 | #include <linux/mtd/mtd.h> | 21 | #include <linux/mtd/mtd.h> |
| @@ -23,9 +25,6 @@ | |||
| 23 | #include <linux/platform_device.h> | 25 | #include <linux/platform_device.h> |
| 24 | #include <linux/types.h> | 26 | #include <linux/types.h> |
| 25 | 27 | ||
| 26 | #include <asm/gpio.h> | ||
| 27 | #include <asm/io.h> | ||
| 28 | |||
| 29 | #define pr_devinit(fmt, args...) ({ static const __devinitconst char __fmt[] = fmt; printk(__fmt, ## args); }) | 28 | #define pr_devinit(fmt, args...) ({ static const __devinitconst char __fmt[] = fmt; printk(__fmt, ## args); }) |
| 30 | 29 | ||
| 31 | #define DRIVER_NAME "gpio-addr-flash" | 30 | #define DRIVER_NAME "gpio-addr-flash" |
diff --git a/drivers/mtd/nand/nand_base.c b/drivers/mtd/nand/nand_base.c index 22113865438b..2957cc70da3d 100644 --- a/drivers/mtd/nand/nand_base.c +++ b/drivers/mtd/nand/nand_base.c | |||
| @@ -761,6 +761,7 @@ static int nand_wait(struct mtd_info *mtd, struct nand_chip *chip) | |||
| 761 | * @mtd: mtd info structure | 761 | * @mtd: mtd info structure |
| 762 | * @chip: nand chip info structure | 762 | * @chip: nand chip info structure |
| 763 | * @buf: buffer to store read data | 763 | * @buf: buffer to store read data |
| 764 | * @page: page number to read | ||
| 764 | * | 765 | * |
| 765 | * Not for syndrome calculating ecc controllers, which use a special oob layout | 766 | * Not for syndrome calculating ecc controllers, which use a special oob layout |
| 766 | */ | 767 | */ |
| @@ -777,6 +778,7 @@ static int nand_read_page_raw(struct mtd_info *mtd, struct nand_chip *chip, | |||
| 777 | * @mtd: mtd info structure | 778 | * @mtd: mtd info structure |
| 778 | * @chip: nand chip info structure | 779 | * @chip: nand chip info structure |
| 779 | * @buf: buffer to store read data | 780 | * @buf: buffer to store read data |
| 781 | * @page: page number to read | ||
| 780 | * | 782 | * |
| 781 | * We need a special oob layout and handling even when OOB isn't used. | 783 | * We need a special oob layout and handling even when OOB isn't used. |
| 782 | */ | 784 | */ |
| @@ -818,6 +820,7 @@ static int nand_read_page_raw_syndrome(struct mtd_info *mtd, struct nand_chip *c | |||
| 818 | * @mtd: mtd info structure | 820 | * @mtd: mtd info structure |
| 819 | * @chip: nand chip info structure | 821 | * @chip: nand chip info structure |
| 820 | * @buf: buffer to store read data | 822 | * @buf: buffer to store read data |
| 823 | * @page: page number to read | ||
| 821 | */ | 824 | */ |
| 822 | static int nand_read_page_swecc(struct mtd_info *mtd, struct nand_chip *chip, | 825 | static int nand_read_page_swecc(struct mtd_info *mtd, struct nand_chip *chip, |
| 823 | uint8_t *buf, int page) | 826 | uint8_t *buf, int page) |
| @@ -939,6 +942,7 @@ static int nand_read_subpage(struct mtd_info *mtd, struct nand_chip *chip, uint3 | |||
| 939 | * @mtd: mtd info structure | 942 | * @mtd: mtd info structure |
| 940 | * @chip: nand chip info structure | 943 | * @chip: nand chip info structure |
| 941 | * @buf: buffer to store read data | 944 | * @buf: buffer to store read data |
| 945 | * @page: page number to read | ||
| 942 | * | 946 | * |
| 943 | * Not for syndrome calculating ecc controllers which need a special oob layout | 947 | * Not for syndrome calculating ecc controllers which need a special oob layout |
| 944 | */ | 948 | */ |
| @@ -983,6 +987,7 @@ static int nand_read_page_hwecc(struct mtd_info *mtd, struct nand_chip *chip, | |||
| 983 | * @mtd: mtd info structure | 987 | * @mtd: mtd info structure |
| 984 | * @chip: nand chip info structure | 988 | * @chip: nand chip info structure |
| 985 | * @buf: buffer to store read data | 989 | * @buf: buffer to store read data |
| 990 | * @page: page number to read | ||
| 986 | * | 991 | * |
| 987 | * Hardware ECC for large page chips, require OOB to be read first. | 992 | * Hardware ECC for large page chips, require OOB to be read first. |
| 988 | * For this ECC mode, the write_page method is re-used from ECC_HW. | 993 | * For this ECC mode, the write_page method is re-used from ECC_HW. |
| @@ -1031,6 +1036,7 @@ static int nand_read_page_hwecc_oob_first(struct mtd_info *mtd, | |||
| 1031 | * @mtd: mtd info structure | 1036 | * @mtd: mtd info structure |
| 1032 | * @chip: nand chip info structure | 1037 | * @chip: nand chip info structure |
| 1033 | * @buf: buffer to store read data | 1038 | * @buf: buffer to store read data |
| 1039 | * @page: page number to read | ||
| 1034 | * | 1040 | * |
| 1035 | * The hw generator calculates the error syndrome automatically. Therefor | 1041 | * The hw generator calculates the error syndrome automatically. Therefor |
| 1036 | * we need a special oob layout and handling. | 1042 | * we need a special oob layout and handling. |
diff --git a/fs/Kconfig b/fs/Kconfig index 2126078a38ed..64d44efad7a5 100644 --- a/fs/Kconfig +++ b/fs/Kconfig | |||
| @@ -135,7 +135,7 @@ config TMPFS_POSIX_ACL | |||
| 135 | 135 | ||
| 136 | config HUGETLBFS | 136 | config HUGETLBFS |
| 137 | bool "HugeTLB file system support" | 137 | bool "HugeTLB file system support" |
| 138 | depends on X86 || IA64 || PPC_BOOK3S_64 || SPARC64 || (S390 && 64BIT) || \ | 138 | depends on X86 || IA64 || SPARC64 || (S390 && 64BIT) || \ |
| 139 | SYS_SUPPORTS_HUGETLBFS || BROKEN | 139 | SYS_SUPPORTS_HUGETLBFS || BROKEN |
| 140 | help | 140 | help |
| 141 | hugetlbfs is a filesystem backing for HugeTLB pages, based on | 141 | hugetlbfs is a filesystem backing for HugeTLB pages, based on |
diff --git a/fs/compat.c b/fs/compat.c index d576b552e8e2..6c19040ffeef 100644 --- a/fs/compat.c +++ b/fs/compat.c | |||
| @@ -1532,6 +1532,8 @@ int compat_do_execve(char * filename, | |||
| 1532 | if (retval < 0) | 1532 | if (retval < 0) |
| 1533 | goto out; | 1533 | goto out; |
| 1534 | 1534 | ||
| 1535 | current->stack_start = current->mm->start_stack; | ||
| 1536 | |||
| 1535 | /* execve succeeded */ | 1537 | /* execve succeeded */ |
| 1536 | current->fs->in_exec = 0; | 1538 | current->fs->in_exec = 0; |
| 1537 | current->in_execve = 0; | 1539 | current->in_execve = 0; |
diff --git a/fs/fuse/dir.c b/fs/fuse/dir.c index 992f6c9410bb..8ada78aade58 100644 --- a/fs/fuse/dir.c +++ b/fs/fuse/dir.c | |||
| @@ -712,8 +712,10 @@ static int fuse_rename(struct inode *olddir, struct dentry *oldent, | |||
| 712 | fuse_invalidate_attr(newdir); | 712 | fuse_invalidate_attr(newdir); |
| 713 | 713 | ||
| 714 | /* newent will end up negative */ | 714 | /* newent will end up negative */ |
| 715 | if (newent->d_inode) | 715 | if (newent->d_inode) { |
| 716 | fuse_invalidate_attr(newent->d_inode); | ||
| 716 | fuse_invalidate_entry_cache(newent); | 717 | fuse_invalidate_entry_cache(newent); |
| 718 | } | ||
| 717 | } else if (err == -EINTR) { | 719 | } else if (err == -EINTR) { |
| 718 | /* If request was interrupted, DEITY only knows if the | 720 | /* If request was interrupted, DEITY only knows if the |
| 719 | rename actually took place. If the invalidation | 721 | rename actually took place. If the invalidation |
diff --git a/fs/fuse/file.c b/fs/fuse/file.c index a3492f7d207c..c18913a777ae 100644 --- a/fs/fuse/file.c +++ b/fs/fuse/file.c | |||
| @@ -1063,7 +1063,8 @@ ssize_t fuse_direct_io(struct file *file, const char __user *buf, | |||
| 1063 | break; | 1063 | break; |
| 1064 | } | 1064 | } |
| 1065 | } | 1065 | } |
| 1066 | fuse_put_request(fc, req); | 1066 | if (!IS_ERR(req)) |
| 1067 | fuse_put_request(fc, req); | ||
| 1067 | if (res > 0) | 1068 | if (res > 0) |
| 1068 | *ppos = pos; | 1069 | *ppos = pos; |
| 1069 | 1070 | ||
| @@ -1599,7 +1600,7 @@ static int fuse_ioctl_copy_user(struct page **pages, struct iovec *iov, | |||
| 1599 | kaddr += copy; | 1600 | kaddr += copy; |
| 1600 | } | 1601 | } |
| 1601 | 1602 | ||
| 1602 | kunmap(map); | 1603 | kunmap(page); |
| 1603 | } | 1604 | } |
| 1604 | 1605 | ||
| 1605 | return 0; | 1606 | return 0; |
diff --git a/fs/sysfs/dir.c b/fs/sysfs/dir.c index 5fad489ce5bc..e0201837d244 100644 --- a/fs/sysfs/dir.c +++ b/fs/sysfs/dir.c | |||
| @@ -21,6 +21,7 @@ | |||
| 21 | #include <linux/completion.h> | 21 | #include <linux/completion.h> |
| 22 | #include <linux/mutex.h> | 22 | #include <linux/mutex.h> |
| 23 | #include <linux/slab.h> | 23 | #include <linux/slab.h> |
| 24 | #include <linux/security.h> | ||
| 24 | #include "sysfs.h" | 25 | #include "sysfs.h" |
| 25 | 26 | ||
| 26 | DEFINE_MUTEX(sysfs_mutex); | 27 | DEFINE_MUTEX(sysfs_mutex); |
| @@ -285,6 +286,9 @@ void release_sysfs_dirent(struct sysfs_dirent * sd) | |||
| 285 | sysfs_put(sd->s_symlink.target_sd); | 286 | sysfs_put(sd->s_symlink.target_sd); |
| 286 | if (sysfs_type(sd) & SYSFS_COPY_NAME) | 287 | if (sysfs_type(sd) & SYSFS_COPY_NAME) |
| 287 | kfree(sd->s_name); | 288 | kfree(sd->s_name); |
| 289 | if (sd->s_iattr && sd->s_iattr->ia_secdata) | ||
| 290 | security_release_secctx(sd->s_iattr->ia_secdata, | ||
| 291 | sd->s_iattr->ia_secdata_len); | ||
| 288 | kfree(sd->s_iattr); | 292 | kfree(sd->s_iattr); |
| 289 | sysfs_free_ino(sd->s_ino); | 293 | sysfs_free_ino(sd->s_ino); |
| 290 | kmem_cache_free(sysfs_dir_cachep, sd); | 294 | kmem_cache_free(sysfs_dir_cachep, sd); |
diff --git a/kernel/kthread.c b/kernel/kthread.c index 5fe709982caa..ab7ae57773e1 100644 --- a/kernel/kthread.c +++ b/kernel/kthread.c | |||
| @@ -150,29 +150,6 @@ struct task_struct *kthread_create(int (*threadfn)(void *data), | |||
| 150 | EXPORT_SYMBOL(kthread_create); | 150 | EXPORT_SYMBOL(kthread_create); |
| 151 | 151 | ||
| 152 | /** | 152 | /** |
| 153 | * kthread_bind - bind a just-created kthread to a cpu. | ||
| 154 | * @k: thread created by kthread_create(). | ||
| 155 | * @cpu: cpu (might not be online, must be possible) for @k to run on. | ||
| 156 | * | ||
| 157 | * Description: This function is equivalent to set_cpus_allowed(), | ||
| 158 | * except that @cpu doesn't need to be online, and the thread must be | ||
| 159 | * stopped (i.e., just returned from kthread_create()). | ||
| 160 | */ | ||
| 161 | void kthread_bind(struct task_struct *k, unsigned int cpu) | ||
| 162 | { | ||
| 163 | /* Must have done schedule() in kthread() before we set_task_cpu */ | ||
| 164 | if (!wait_task_inactive(k, TASK_UNINTERRUPTIBLE)) { | ||
| 165 | WARN_ON(1); | ||
| 166 | return; | ||
| 167 | } | ||
| 168 | set_task_cpu(k, cpu); | ||
| 169 | k->cpus_allowed = cpumask_of_cpu(cpu); | ||
| 170 | k->rt.nr_cpus_allowed = 1; | ||
| 171 | k->flags |= PF_THREAD_BOUND; | ||
| 172 | } | ||
| 173 | EXPORT_SYMBOL(kthread_bind); | ||
| 174 | |||
| 175 | /** | ||
| 176 | * kthread_stop - stop a thread created by kthread_create(). | 153 | * kthread_stop - stop a thread created by kthread_create(). |
| 177 | * @k: thread created by kthread_create(). | 154 | * @k: thread created by kthread_create(). |
| 178 | * | 155 | * |
diff --git a/kernel/sched.c b/kernel/sched.c index a455dca884a6..28dd4f490bfc 100644 --- a/kernel/sched.c +++ b/kernel/sched.c | |||
| @@ -1992,6 +1992,38 @@ static inline void check_class_changed(struct rq *rq, struct task_struct *p, | |||
| 1992 | p->sched_class->prio_changed(rq, p, oldprio, running); | 1992 | p->sched_class->prio_changed(rq, p, oldprio, running); |
| 1993 | } | 1993 | } |
| 1994 | 1994 | ||
| 1995 | /** | ||
| 1996 | * kthread_bind - bind a just-created kthread to a cpu. | ||
| 1997 | * @k: thread created by kthread_create(). | ||
| 1998 | * @cpu: cpu (might not be online, must be possible) for @k to run on. | ||
| 1999 | * | ||
| 2000 | * Description: This function is equivalent to set_cpus_allowed(), | ||
| 2001 | * except that @cpu doesn't need to be online, and the thread must be | ||
| 2002 | * stopped (i.e., just returned from kthread_create()). | ||
| 2003 | * | ||
| 2004 | * Function lives here instead of kthread.c because it messes with | ||
| 2005 | * scheduler internals which require locking. | ||
| 2006 | */ | ||
| 2007 | void kthread_bind(struct task_struct *p, unsigned int cpu) | ||
| 2008 | { | ||
| 2009 | struct rq *rq = cpu_rq(cpu); | ||
| 2010 | unsigned long flags; | ||
| 2011 | |||
| 2012 | /* Must have done schedule() in kthread() before we set_task_cpu */ | ||
| 2013 | if (!wait_task_inactive(p, TASK_UNINTERRUPTIBLE)) { | ||
| 2014 | WARN_ON(1); | ||
| 2015 | return; | ||
| 2016 | } | ||
| 2017 | |||
| 2018 | spin_lock_irqsave(&rq->lock, flags); | ||
| 2019 | set_task_cpu(p, cpu); | ||
| 2020 | p->cpus_allowed = cpumask_of_cpu(cpu); | ||
| 2021 | p->rt.nr_cpus_allowed = 1; | ||
| 2022 | p->flags |= PF_THREAD_BOUND; | ||
| 2023 | spin_unlock_irqrestore(&rq->lock, flags); | ||
| 2024 | } | ||
| 2025 | EXPORT_SYMBOL(kthread_bind); | ||
| 2026 | |||
| 1995 | #ifdef CONFIG_SMP | 2027 | #ifdef CONFIG_SMP |
| 1996 | /* | 2028 | /* |
| 1997 | * Is this task likely cache-hot: | 2029 | * Is this task likely cache-hot: |
| @@ -2004,7 +2036,7 @@ task_hot(struct task_struct *p, u64 now, struct sched_domain *sd) | |||
| 2004 | /* | 2036 | /* |
| 2005 | * Buddy candidates are cache hot: | 2037 | * Buddy candidates are cache hot: |
| 2006 | */ | 2038 | */ |
| 2007 | if (sched_feat(CACHE_HOT_BUDDY) && | 2039 | if (sched_feat(CACHE_HOT_BUDDY) && this_rq()->nr_running && |
| 2008 | (&p->se == cfs_rq_of(&p->se)->next || | 2040 | (&p->se == cfs_rq_of(&p->se)->next || |
| 2009 | &p->se == cfs_rq_of(&p->se)->last)) | 2041 | &p->se == cfs_rq_of(&p->se)->last)) |
| 2010 | return 1; | 2042 | return 1; |
| @@ -9532,13 +9564,13 @@ void __init sched_init(void) | |||
| 9532 | current->sched_class = &fair_sched_class; | 9564 | current->sched_class = &fair_sched_class; |
| 9533 | 9565 | ||
| 9534 | /* Allocate the nohz_cpu_mask if CONFIG_CPUMASK_OFFSTACK */ | 9566 | /* Allocate the nohz_cpu_mask if CONFIG_CPUMASK_OFFSTACK */ |
| 9535 | alloc_cpumask_var(&nohz_cpu_mask, GFP_NOWAIT); | 9567 | zalloc_cpumask_var(&nohz_cpu_mask, GFP_NOWAIT); |
| 9536 | #ifdef CONFIG_SMP | 9568 | #ifdef CONFIG_SMP |
| 9537 | #ifdef CONFIG_NO_HZ | 9569 | #ifdef CONFIG_NO_HZ |
| 9538 | alloc_cpumask_var(&nohz.cpu_mask, GFP_NOWAIT); | 9570 | zalloc_cpumask_var(&nohz.cpu_mask, GFP_NOWAIT); |
| 9539 | alloc_cpumask_var(&nohz.ilb_grp_nohz_mask, GFP_NOWAIT); | 9571 | alloc_cpumask_var(&nohz.ilb_grp_nohz_mask, GFP_NOWAIT); |
| 9540 | #endif | 9572 | #endif |
| 9541 | alloc_cpumask_var(&cpu_isolated_map, GFP_NOWAIT); | 9573 | zalloc_cpumask_var(&cpu_isolated_map, GFP_NOWAIT); |
| 9542 | #endif /* SMP */ | 9574 | #endif /* SMP */ |
| 9543 | 9575 | ||
| 9544 | perf_event_init(); | 9576 | perf_event_init(); |
diff --git a/kernel/sched_fair.c b/kernel/sched_fair.c index c32c3e643daa..37087a7fac22 100644 --- a/kernel/sched_fair.c +++ b/kernel/sched_fair.c | |||
| @@ -822,6 +822,26 @@ check_preempt_tick(struct cfs_rq *cfs_rq, struct sched_entity *curr) | |||
| 822 | * re-elected due to buddy favours. | 822 | * re-elected due to buddy favours. |
| 823 | */ | 823 | */ |
| 824 | clear_buddies(cfs_rq, curr); | 824 | clear_buddies(cfs_rq, curr); |
| 825 | return; | ||
| 826 | } | ||
| 827 | |||
| 828 | /* | ||
| 829 | * Ensure that a task that missed wakeup preemption by a | ||
| 830 | * narrow margin doesn't have to wait for a full slice. | ||
| 831 | * This also mitigates buddy induced latencies under load. | ||
| 832 | */ | ||
| 833 | if (!sched_feat(WAKEUP_PREEMPT)) | ||
| 834 | return; | ||
| 835 | |||
| 836 | if (delta_exec < sysctl_sched_min_granularity) | ||
| 837 | return; | ||
| 838 | |||
| 839 | if (cfs_rq->nr_running > 1) { | ||
| 840 | struct sched_entity *se = __pick_next_entity(cfs_rq); | ||
| 841 | s64 delta = curr->vruntime - se->vruntime; | ||
| 842 | |||
| 843 | if (delta > ideal_runtime) | ||
| 844 | resched_task(rq_of(cfs_rq)->curr); | ||
| 825 | } | 845 | } |
| 826 | } | 846 | } |
| 827 | 847 | ||
| @@ -861,21 +881,18 @@ wakeup_preempt_entity(struct sched_entity *curr, struct sched_entity *se); | |||
| 861 | static struct sched_entity *pick_next_entity(struct cfs_rq *cfs_rq) | 881 | static struct sched_entity *pick_next_entity(struct cfs_rq *cfs_rq) |
| 862 | { | 882 | { |
| 863 | struct sched_entity *se = __pick_next_entity(cfs_rq); | 883 | struct sched_entity *se = __pick_next_entity(cfs_rq); |
| 864 | struct sched_entity *buddy; | 884 | struct sched_entity *left = se; |
| 865 | 885 | ||
| 866 | if (cfs_rq->next) { | 886 | if (cfs_rq->next && wakeup_preempt_entity(cfs_rq->next, left) < 1) |
| 867 | buddy = cfs_rq->next; | 887 | se = cfs_rq->next; |
| 868 | cfs_rq->next = NULL; | ||
| 869 | if (wakeup_preempt_entity(buddy, se) < 1) | ||
| 870 | return buddy; | ||
| 871 | } | ||
| 872 | 888 | ||
| 873 | if (cfs_rq->last) { | 889 | /* |
| 874 | buddy = cfs_rq->last; | 890 | * Prefer last buddy, try to return the CPU to a preempted task. |
| 875 | cfs_rq->last = NULL; | 891 | */ |
| 876 | if (wakeup_preempt_entity(buddy, se) < 1) | 892 | if (cfs_rq->last && wakeup_preempt_entity(cfs_rq->last, left) < 1) |
| 877 | return buddy; | 893 | se = cfs_rq->last; |
| 878 | } | 894 | |
| 895 | clear_buddies(cfs_rq, se); | ||
| 879 | 896 | ||
| 880 | return se; | 897 | return se; |
| 881 | } | 898 | } |
| @@ -1577,6 +1594,7 @@ static void check_preempt_wakeup(struct rq *rq, struct task_struct *p, int wake_ | |||
| 1577 | struct sched_entity *se = &curr->se, *pse = &p->se; | 1594 | struct sched_entity *se = &curr->se, *pse = &p->se; |
| 1578 | struct cfs_rq *cfs_rq = task_cfs_rq(curr); | 1595 | struct cfs_rq *cfs_rq = task_cfs_rq(curr); |
| 1579 | int sync = wake_flags & WF_SYNC; | 1596 | int sync = wake_flags & WF_SYNC; |
| 1597 | int scale = cfs_rq->nr_running >= sched_nr_latency; | ||
| 1580 | 1598 | ||
| 1581 | update_curr(cfs_rq); | 1599 | update_curr(cfs_rq); |
| 1582 | 1600 | ||
| @@ -1591,18 +1609,7 @@ static void check_preempt_wakeup(struct rq *rq, struct task_struct *p, int wake_ | |||
| 1591 | if (unlikely(se == pse)) | 1609 | if (unlikely(se == pse)) |
| 1592 | return; | 1610 | return; |
| 1593 | 1611 | ||
| 1594 | /* | 1612 | if (sched_feat(NEXT_BUDDY) && scale && !(wake_flags & WF_FORK)) |
| 1595 | * Only set the backward buddy when the current task is still on the | ||
| 1596 | * rq. This can happen when a wakeup gets interleaved with schedule on | ||
| 1597 | * the ->pre_schedule() or idle_balance() point, either of which can | ||
| 1598 | * drop the rq lock. | ||
| 1599 | * | ||
| 1600 | * Also, during early boot the idle thread is in the fair class, for | ||
| 1601 | * obvious reasons its a bad idea to schedule back to the idle thread. | ||
| 1602 | */ | ||
| 1603 | if (sched_feat(LAST_BUDDY) && likely(se->on_rq && curr != rq->idle)) | ||
| 1604 | set_last_buddy(se); | ||
| 1605 | if (sched_feat(NEXT_BUDDY) && !(wake_flags & WF_FORK)) | ||
| 1606 | set_next_buddy(pse); | 1613 | set_next_buddy(pse); |
| 1607 | 1614 | ||
| 1608 | /* | 1615 | /* |
| @@ -1648,8 +1655,22 @@ static void check_preempt_wakeup(struct rq *rq, struct task_struct *p, int wake_ | |||
| 1648 | 1655 | ||
| 1649 | BUG_ON(!pse); | 1656 | BUG_ON(!pse); |
| 1650 | 1657 | ||
| 1651 | if (wakeup_preempt_entity(se, pse) == 1) | 1658 | if (wakeup_preempt_entity(se, pse) == 1) { |
| 1652 | resched_task(curr); | 1659 | resched_task(curr); |
| 1660 | /* | ||
| 1661 | * Only set the backward buddy when the current task is still | ||
| 1662 | * on the rq. This can happen when a wakeup gets interleaved | ||
| 1663 | * with schedule on the ->pre_schedule() or idle_balance() | ||
| 1664 | * point, either of which can * drop the rq lock. | ||
| 1665 | * | ||
| 1666 | * Also, during early boot the idle thread is in the fair class, | ||
| 1667 | * for obvious reasons its a bad idea to schedule back to it. | ||
| 1668 | */ | ||
| 1669 | if (unlikely(!se->on_rq || curr == rq->idle)) | ||
| 1670 | return; | ||
| 1671 | if (sched_feat(LAST_BUDDY) && scale && entity_is_task(se)) | ||
| 1672 | set_last_buddy(se); | ||
| 1673 | } | ||
| 1653 | } | 1674 | } |
| 1654 | 1675 | ||
| 1655 | static struct task_struct *pick_next_task_fair(struct rq *rq) | 1676 | static struct task_struct *pick_next_task_fair(struct rq *rq) |
diff --git a/kernel/trace/ftrace.c b/kernel/trace/ftrace.c index 9c451a1930b6..6dc4e5ef7a01 100644 --- a/kernel/trace/ftrace.c +++ b/kernel/trace/ftrace.c | |||
| @@ -2222,15 +2222,15 @@ ftrace_regex_write(struct file *file, const char __user *ubuf, | |||
| 2222 | ret = ftrace_process_regex(parser->buffer, | 2222 | ret = ftrace_process_regex(parser->buffer, |
| 2223 | parser->idx, enable); | 2223 | parser->idx, enable); |
| 2224 | if (ret) | 2224 | if (ret) |
| 2225 | goto out; | 2225 | goto out_unlock; |
| 2226 | 2226 | ||
| 2227 | trace_parser_clear(parser); | 2227 | trace_parser_clear(parser); |
| 2228 | } | 2228 | } |
| 2229 | 2229 | ||
| 2230 | ret = read; | 2230 | ret = read; |
| 2231 | 2231 | out_unlock: | |
| 2232 | mutex_unlock(&ftrace_regex_lock); | 2232 | mutex_unlock(&ftrace_regex_lock); |
| 2233 | out: | 2233 | |
| 2234 | return ret; | 2234 | return ret; |
| 2235 | } | 2235 | } |
| 2236 | 2236 | ||
diff --git a/kernel/trace/ring_buffer.c b/kernel/trace/ring_buffer.c index 3ffa502fb243..5dd017fea6f5 100644 --- a/kernel/trace/ring_buffer.c +++ b/kernel/trace/ring_buffer.c | |||
| @@ -1193,6 +1193,7 @@ rb_remove_pages(struct ring_buffer_per_cpu *cpu_buffer, unsigned nr_pages) | |||
| 1193 | atomic_inc(&cpu_buffer->record_disabled); | 1193 | atomic_inc(&cpu_buffer->record_disabled); |
| 1194 | synchronize_sched(); | 1194 | synchronize_sched(); |
| 1195 | 1195 | ||
| 1196 | spin_lock_irq(&cpu_buffer->reader_lock); | ||
| 1196 | rb_head_page_deactivate(cpu_buffer); | 1197 | rb_head_page_deactivate(cpu_buffer); |
| 1197 | 1198 | ||
| 1198 | for (i = 0; i < nr_pages; i++) { | 1199 | for (i = 0; i < nr_pages; i++) { |
| @@ -1207,6 +1208,7 @@ rb_remove_pages(struct ring_buffer_per_cpu *cpu_buffer, unsigned nr_pages) | |||
| 1207 | return; | 1208 | return; |
| 1208 | 1209 | ||
| 1209 | rb_reset_cpu(cpu_buffer); | 1210 | rb_reset_cpu(cpu_buffer); |
| 1211 | spin_unlock_irq(&cpu_buffer->reader_lock); | ||
| 1210 | 1212 | ||
| 1211 | rb_check_pages(cpu_buffer); | 1213 | rb_check_pages(cpu_buffer); |
| 1212 | 1214 | ||
diff --git a/lib/dma-debug.c b/lib/dma-debug.c index 58a9f9fc609a..ce6b7eabf674 100644 --- a/lib/dma-debug.c +++ b/lib/dma-debug.c | |||
| @@ -819,9 +819,11 @@ static void check_unmap(struct dma_debug_entry *ref) | |||
| 819 | err_printk(ref->dev, entry, "DMA-API: device driver frees " | 819 | err_printk(ref->dev, entry, "DMA-API: device driver frees " |
| 820 | "DMA memory with different CPU address " | 820 | "DMA memory with different CPU address " |
| 821 | "[device address=0x%016llx] [size=%llu bytes] " | 821 | "[device address=0x%016llx] [size=%llu bytes] " |
| 822 | "[cpu alloc address=%p] [cpu free address=%p]", | 822 | "[cpu alloc address=0x%016llx] " |
| 823 | "[cpu free address=0x%016llx]", | ||
| 823 | ref->dev_addr, ref->size, | 824 | ref->dev_addr, ref->size, |
| 824 | (void *)entry->paddr, (void *)ref->paddr); | 825 | (unsigned long long)entry->paddr, |
| 826 | (unsigned long long)ref->paddr); | ||
| 825 | } | 827 | } |
| 826 | 828 | ||
| 827 | if (ref->sg_call_ents && ref->type == dma_debug_sg && | 829 | if (ref->sg_call_ents && ref->type == dma_debug_sg && |
