diff options
author | Andreas Herrmann <andreas.herrmann3@amd.com> | 2009-06-08 13:09:39 -0400 |
---|---|---|
committer | Ingo Molnar <mingo@elte.hu> | 2009-06-08 13:18:25 -0400 |
commit | c9690998ef48ffefeccb91c70a7739eebdea57f9 (patch) | |
tree | 01d628fd94a943ee6fe046257aa3a6be1b05390f /arch/x86/mm/memtest.c | |
parent | c4ed3f04ba9defe22aa729d1646f970f791c03d7 (diff) |
x86: memtest: remove 64-bit division
Using gcc 3.3.5 a "make allmodconfig" + "CONFIG_KVM=n"
triggers a build error:
arch/x86/mm/built-in.o(.init.text+0x43f7): In function `__change_page_attr':
arch/x86/mm/pageattr.c:114: undefined reference to `__udivdi3'
make: *** [.tmp_vmlinux1] Error 1
The culprit turned out to be a division in arch/x86/mm/memtest.c
For more info see this thread:
http://marc.info/?l=linux-kernel&m=124416232620683
The patch entirely removes the division that caused the build
error.
[ Impact: build fix with certain GCC versions ]
Reported-by: Tetsuo Handa <penguin-kernel@i-love.sakura.ne.jp>
Signed-off-by: Andreas Herrmann <andreas.herrmann3@amd.com>
Cc: Yinghai Lu <yinghai@kernel.org>
Cc: xiyou.wangcong@gmail.com
Cc: Andrew Morton <akpm@linux-foundation.org>
Cc: <stable@kernel.org>
LKML-Reference: <20090608170939.GB12431@alberich.amd.com>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
Diffstat (limited to 'arch/x86/mm/memtest.c')
-rw-r--r-- | arch/x86/mm/memtest.c | 14 |
1 files changed, 7 insertions, 7 deletions
diff --git a/arch/x86/mm/memtest.c b/arch/x86/mm/memtest.c index 605c8be06217..c0bedcd10f97 100644 --- a/arch/x86/mm/memtest.c +++ b/arch/x86/mm/memtest.c | |||
@@ -40,23 +40,23 @@ static void __init reserve_bad_mem(u64 pattern, u64 start_bad, u64 end_bad) | |||
40 | 40 | ||
41 | static void __init memtest(u64 pattern, u64 start_phys, u64 size) | 41 | static void __init memtest(u64 pattern, u64 start_phys, u64 size) |
42 | { | 42 | { |
43 | u64 i, count; | 43 | u64 *p; |
44 | u64 *start; | 44 | void *start, *end; |
45 | u64 start_bad, last_bad; | 45 | u64 start_bad, last_bad; |
46 | u64 start_phys_aligned; | 46 | u64 start_phys_aligned; |
47 | size_t incr; | 47 | size_t incr; |
48 | 48 | ||
49 | incr = sizeof(pattern); | 49 | incr = sizeof(pattern); |
50 | start_phys_aligned = ALIGN(start_phys, incr); | 50 | start_phys_aligned = ALIGN(start_phys, incr); |
51 | count = (size - (start_phys_aligned - start_phys))/incr; | ||
52 | start = __va(start_phys_aligned); | 51 | start = __va(start_phys_aligned); |
52 | end = start + size - (start_phys_aligned - start_phys); | ||
53 | start_bad = 0; | 53 | start_bad = 0; |
54 | last_bad = 0; | 54 | last_bad = 0; |
55 | 55 | ||
56 | for (i = 0; i < count; i++) | 56 | for (p = start; p < end; p++) |
57 | start[i] = pattern; | 57 | *p = pattern; |
58 | for (i = 0; i < count; i++, start++, start_phys_aligned += incr) { | 58 | for (p = start; p < end; p++, start_phys_aligned += incr) { |
59 | if (*start == pattern) | 59 | if (*p == pattern) |
60 | continue; | 60 | continue; |
61 | if (start_phys_aligned == last_bad + incr) { | 61 | if (start_phys_aligned == last_bad + incr) { |
62 | last_bad += incr; | 62 | last_bad += incr; |