diff options
author | David Howells <dhowells@redhat.com> | 2010-10-27 12:28:46 -0400 |
---|---|---|
committer | David Howells <dhowells@redhat.com> | 2010-10-27 12:28:46 -0400 |
commit | b478491f2628114b2eae76587f22ce3789b66012 (patch) | |
tree | 038580a05fa1a3c8c9cbee0fa8743af2bba650b9 /arch/mn10300 | |
parent | 9731d23710736b96786d68c2e63148ff3f22e6eb (diff) |
MN10300: Allow some cacheflushes to be avoided if cache snooping is available
The AM34 core is able to do cache snooping, and so can skip some of the cache
flushing.
Signed-off-by: David Howells <dhowells@redhat.com>
Diffstat (limited to 'arch/mn10300')
-rw-r--r-- | arch/mn10300/Kconfig | 1 | ||||
-rw-r--r-- | arch/mn10300/include/asm/cacheflush.h | 16 | ||||
-rw-r--r-- | arch/mn10300/kernel/kprobes.c | 4 | ||||
-rw-r--r-- | arch/mn10300/kernel/traps.c | 2 | ||||
-rw-r--r-- | arch/mn10300/mm/Kconfig.cache | 34 | ||||
-rw-r--r-- | arch/mn10300/mm/Makefile | 2 | ||||
-rw-r--r-- | arch/mn10300/mm/cache-flush-icache.c | 137 | ||||
-rw-r--r-- | arch/mn10300/mm/cache-inv-icache.c | 119 | ||||
-rw-r--r-- | arch/mn10300/mm/cache.c | 90 |
9 files changed, 309 insertions, 96 deletions
diff --git a/arch/mn10300/Kconfig b/arch/mn10300/Kconfig index 069e34d4c4ac..21e2a534d98e 100644 --- a/arch/mn10300/Kconfig +++ b/arch/mn10300/Kconfig | |||
@@ -18,6 +18,7 @@ config AM33_3 | |||
18 | config AM34_2 | 18 | config AM34_2 |
19 | def_bool n | 19 | def_bool n |
20 | select MN10300_HAS_ATOMIC_OPS_UNIT | 20 | select MN10300_HAS_ATOMIC_OPS_UNIT |
21 | select MN10300_HAS_CACHE_SNOOP | ||
21 | 22 | ||
22 | config MMU | 23 | config MMU |
23 | def_bool y | 24 | def_bool y |
diff --git a/arch/mn10300/include/asm/cacheflush.h b/arch/mn10300/include/asm/cacheflush.h index 748143f65418..faed90240ded 100644 --- a/arch/mn10300/include/asm/cacheflush.h +++ b/arch/mn10300/include/asm/cacheflush.h | |||
@@ -131,18 +131,22 @@ extern void mn10300_dcache_flush_inv_range2(unsigned long start, unsigned long s | |||
131 | /* | 131 | /* |
132 | * Physically-indexed cache management | 132 | * Physically-indexed cache management |
133 | */ | 133 | */ |
134 | #ifdef CONFIG_MN10300_CACHE_ENABLED | 134 | #if defined(CONFIG_MN10300_CACHE_FLUSH_ICACHE) |
135 | 135 | extern void flush_icache_page(struct vm_area_struct *vma, struct page *page); | |
136 | extern void flush_icache_range(unsigned long start, unsigned long end); | ||
137 | #elif defined(CONFIG_MN10300_CACHE_INV_ICACHE) | ||
138 | static inline void flush_icache_page(struct vm_area_struct *vma, | ||
139 | struct page *page) | ||
140 | { | ||
141 | mn10300_icache_inv_page(page_to_phys(page)); | ||
142 | } | ||
136 | extern void flush_icache_range(unsigned long start, unsigned long end); | 143 | extern void flush_icache_range(unsigned long start, unsigned long end); |
137 | extern void flush_icache_page(struct vm_area_struct *vma, struct page *pg); | ||
138 | |||
139 | #else | 144 | #else |
140 | |||
141 | #define flush_icache_range(start, end) do {} while (0) | 145 | #define flush_icache_range(start, end) do {} while (0) |
142 | #define flush_icache_page(vma, pg) do {} while (0) | 146 | #define flush_icache_page(vma, pg) do {} while (0) |
143 | |||
144 | #endif | 147 | #endif |
145 | 148 | ||
149 | |||
146 | #define flush_icache_user_range(vma, pg, adr, len) \ | 150 | #define flush_icache_user_range(vma, pg, adr, len) \ |
147 | flush_icache_range(adr, adr + len) | 151 | flush_icache_range(adr, adr + len) |
148 | 152 | ||
diff --git a/arch/mn10300/kernel/kprobes.c b/arch/mn10300/kernel/kprobes.c index 67e6389d625a..0311a7fcea16 100644 --- a/arch/mn10300/kernel/kprobes.c +++ b/arch/mn10300/kernel/kprobes.c | |||
@@ -377,8 +377,10 @@ void __kprobes arch_arm_kprobe(struct kprobe *p) | |||
377 | 377 | ||
378 | void __kprobes arch_disarm_kprobe(struct kprobe *p) | 378 | void __kprobes arch_disarm_kprobe(struct kprobe *p) |
379 | { | 379 | { |
380 | #ifndef CONFIG_MN10300_CACHE_SNOOP | ||
380 | mn10300_dcache_flush(); | 381 | mn10300_dcache_flush(); |
381 | mn10300_icache_inv(); | 382 | mn10300_icache_inv(); |
383 | #endif | ||
382 | } | 384 | } |
383 | 385 | ||
384 | void arch_remove_kprobe(struct kprobe *p) | 386 | void arch_remove_kprobe(struct kprobe *p) |
@@ -390,8 +392,10 @@ void __kprobes disarm_kprobe(struct kprobe *p, struct pt_regs *regs) | |||
390 | { | 392 | { |
391 | *p->addr = p->opcode; | 393 | *p->addr = p->opcode; |
392 | regs->pc = (unsigned long) p->addr; | 394 | regs->pc = (unsigned long) p->addr; |
395 | #ifndef CONFIG_MN10300_CACHE_SNOOP | ||
393 | mn10300_dcache_flush(); | 396 | mn10300_dcache_flush(); |
394 | mn10300_icache_inv(); | 397 | mn10300_icache_inv(); |
398 | #endif | ||
395 | } | 399 | } |
396 | 400 | ||
397 | static inline | 401 | static inline |
diff --git a/arch/mn10300/kernel/traps.c b/arch/mn10300/kernel/traps.c index a64604b512d5..c7257a1304a9 100644 --- a/arch/mn10300/kernel/traps.c +++ b/arch/mn10300/kernel/traps.c | |||
@@ -533,8 +533,10 @@ void __init set_intr_stub(enum exception_code code, void *handler) | |||
533 | vector[6] = 0xcb; | 533 | vector[6] = 0xcb; |
534 | vector[7] = 0xcb; | 534 | vector[7] = 0xcb; |
535 | 535 | ||
536 | #ifndef CONFIG_MN10300_CACHE_SNOOP | ||
536 | mn10300_dcache_flush_inv(); | 537 | mn10300_dcache_flush_inv(); |
537 | mn10300_icache_inv(); | 538 | mn10300_icache_inv(); |
539 | #endif | ||
538 | } | 540 | } |
539 | 541 | ||
540 | /* | 542 | /* |
diff --git a/arch/mn10300/mm/Kconfig.cache b/arch/mn10300/mm/Kconfig.cache index 97adc06e7128..653254a34f88 100644 --- a/arch/mn10300/mm/Kconfig.cache +++ b/arch/mn10300/mm/Kconfig.cache | |||
@@ -22,12 +22,26 @@ choice | |||
22 | 22 | ||
23 | config MN10300_CACHE_WBACK | 23 | config MN10300_CACHE_WBACK |
24 | bool "Write-Back" | 24 | bool "Write-Back" |
25 | help | ||
26 | The dcache operates in delayed write-back mode. It must be manually | ||
27 | flushed if writes are made that subsequently need to be executed or | ||
28 | to be DMA'd by a device. | ||
25 | 29 | ||
26 | config MN10300_CACHE_WTHRU | 30 | config MN10300_CACHE_WTHRU |
27 | bool "Write-Through" | 31 | bool "Write-Through" |
32 | help | ||
33 | The dcache operates in immediate write-through mode. Writes are | ||
34 | committed to RAM immediately in addition to being stored in the | ||
35 | cache. This means that the written data is immediately available for | ||
36 | execution or DMA. | ||
37 | |||
38 | This is not available for use with an SMP kernel if cache flushing | ||
39 | and invalidation by automatic purge register is not selected. | ||
28 | 40 | ||
29 | config MN10300_CACHE_DISABLED | 41 | config MN10300_CACHE_DISABLED |
30 | bool "Disabled" | 42 | bool "Disabled" |
43 | help | ||
44 | The icache and dcache are disabled. | ||
31 | 45 | ||
32 | endchoice | 46 | endchoice |
33 | 47 | ||
@@ -64,3 +78,23 @@ config MN10300_CACHE_FLUSH_BY_TAG | |||
64 | 78 | ||
65 | config MN10300_CACHE_FLUSH_BY_REG | 79 | config MN10300_CACHE_FLUSH_BY_REG |
66 | def_bool y if MN10300_CACHE_MANAGE_BY_REG && MN10300_CACHE_WBACK | 80 | def_bool y if MN10300_CACHE_MANAGE_BY_REG && MN10300_CACHE_WBACK |
81 | |||
82 | |||
83 | config MN10300_HAS_CACHE_SNOOP | ||
84 | def_bool n | ||
85 | |||
86 | config MN10300_CACHE_SNOOP | ||
87 | bool "Use CPU Cache Snooping" | ||
88 | depends on MN10300_CACHE_ENABLED && MN10300_HAS_CACHE_SNOOP | ||
89 | default y | ||
90 | |||
91 | config MN10300_CACHE_FLUSH_ICACHE | ||
92 | def_bool y if MN10300_CACHE_WBACK && !MN10300_CACHE_SNOOP | ||
93 | help | ||
94 | Set if we need the dcache flushing before the icache is invalidated. | ||
95 | |||
96 | config MN10300_CACHE_INV_ICACHE | ||
97 | def_bool y if MN10300_CACHE_WTHRU && !MN10300_CACHE_SNOOP | ||
98 | help | ||
99 | Set if we need the icache to be invalidated, even if the dcache is in | ||
100 | write-through mode and doesn't need flushing. | ||
diff --git a/arch/mn10300/mm/Makefile b/arch/mn10300/mm/Makefile index 7b997236ed20..56c5af83151b 100644 --- a/arch/mn10300/mm/Makefile +++ b/arch/mn10300/mm/Makefile | |||
@@ -3,6 +3,8 @@ | |||
3 | # | 3 | # |
4 | 4 | ||
5 | cacheflush-y := cache.o | 5 | cacheflush-y := cache.o |
6 | cacheflush-$(CONFIG_MN10300_CACHE_INV_ICACHE) += cache-inv-icache.o | ||
7 | cacheflush-$(CONFIG_MN10300_CACHE_FLUSH_ICACHE) += cache-flush-icache.o | ||
6 | cacheflush-$(CONFIG_MN10300_CACHE_INV_BY_TAG) += cache-inv-by-tag.o | 8 | cacheflush-$(CONFIG_MN10300_CACHE_INV_BY_TAG) += cache-inv-by-tag.o |
7 | cacheflush-$(CONFIG_MN10300_CACHE_INV_BY_REG) += cache-inv-by-reg.o | 9 | cacheflush-$(CONFIG_MN10300_CACHE_INV_BY_REG) += cache-inv-by-reg.o |
8 | cacheflush-$(CONFIG_MN10300_CACHE_FLUSH_BY_TAG) += cache-flush-by-tag.o | 10 | cacheflush-$(CONFIG_MN10300_CACHE_FLUSH_BY_TAG) += cache-flush-by-tag.o |
diff --git a/arch/mn10300/mm/cache-flush-icache.c b/arch/mn10300/mm/cache-flush-icache.c new file mode 100644 index 000000000000..0e471e1cb2da --- /dev/null +++ b/arch/mn10300/mm/cache-flush-icache.c | |||
@@ -0,0 +1,137 @@ | |||
1 | /* Flush dcache and invalidate icache when the dcache is in writeback mode | ||
2 | * | ||
3 | * Copyright (C) 2010 Red Hat, Inc. All Rights Reserved. | ||
4 | * Written by David Howells (dhowells@redhat.com) | ||
5 | * | ||
6 | * This program is free software; you can redistribute it and/or | ||
7 | * modify it under the terms of the GNU General Public Licence | ||
8 | * as published by the Free Software Foundation; either version | ||
9 | * 2 of the Licence, or (at your option) any later version. | ||
10 | */ | ||
11 | #include <linux/module.h> | ||
12 | #include <linux/mm.h> | ||
13 | #include <asm/cacheflush.h> | ||
14 | /** | ||
15 | * flush_icache_page - Flush a page from the dcache and invalidate the icache | ||
16 | * @vma: The VMA the page is part of. | ||
17 | * @page: The page to be flushed. | ||
18 | * | ||
19 | * Write a page back from the dcache and invalidate the icache so that we can | ||
20 | * run code from it that we've just written into it | ||
21 | */ | ||
22 | void flush_icache_page(struct vm_area_struct *vma, struct page *page) | ||
23 | { | ||
24 | unsigned long start = page_to_phys(page); | ||
25 | |||
26 | mn10300_dcache_flush_page(start); | ||
27 | mn10300_icache_inv_page(start); | ||
28 | } | ||
29 | EXPORT_SYMBOL(flush_icache_page); | ||
30 | |||
31 | /** | ||
32 | * flush_icache_page_range - Flush dcache and invalidate icache for part of a | ||
33 | * single page | ||
34 | * @start: The starting virtual address of the page part. | ||
35 | * @end: The ending virtual address of the page part. | ||
36 | * | ||
37 | * Flush the dcache and invalidate the icache for part of a single page, as | ||
38 | * determined by the virtual addresses given. The page must be in the paged | ||
39 | * area. | ||
40 | */ | ||
41 | static void flush_icache_page_range(unsigned long start, unsigned long end) | ||
42 | { | ||
43 | unsigned long addr, size, off; | ||
44 | struct page *page; | ||
45 | pgd_t *pgd; | ||
46 | pud_t *pud; | ||
47 | pmd_t *pmd; | ||
48 | pte_t *ppte, pte; | ||
49 | |||
50 | /* work out how much of the page to flush */ | ||
51 | off = start & ~PAGE_MASK; | ||
52 | size = end - start; | ||
53 | |||
54 | /* get the physical address the page is mapped to from the page | ||
55 | * tables */ | ||
56 | pgd = pgd_offset(current->mm, start); | ||
57 | if (!pgd || !pgd_val(*pgd)) | ||
58 | return; | ||
59 | |||
60 | pud = pud_offset(pgd, start); | ||
61 | if (!pud || !pud_val(*pud)) | ||
62 | return; | ||
63 | |||
64 | pmd = pmd_offset(pud, start); | ||
65 | if (!pmd || !pmd_val(*pmd)) | ||
66 | return; | ||
67 | |||
68 | ppte = pte_offset_map(pmd, start); | ||
69 | if (!ppte) | ||
70 | return; | ||
71 | pte = *ppte; | ||
72 | pte_unmap(ppte); | ||
73 | |||
74 | if (pte_none(pte)) | ||
75 | return; | ||
76 | |||
77 | page = pte_page(pte); | ||
78 | if (!page) | ||
79 | return; | ||
80 | |||
81 | addr = page_to_phys(page); | ||
82 | |||
83 | /* flush the dcache and invalidate the icache coverage on that | ||
84 | * region */ | ||
85 | mn10300_dcache_flush_range2(addr + off, size); | ||
86 | mn10300_icache_inv_range2(addr + off, size); | ||
87 | } | ||
88 | |||
89 | /** | ||
90 | * flush_icache_range - Globally flush dcache and invalidate icache for region | ||
91 | * @start: The starting virtual address of the region. | ||
92 | * @end: The ending virtual address of the region. | ||
93 | * | ||
94 | * This is used by the kernel to globally flush some code it has just written | ||
95 | * from the dcache back to RAM and then to globally invalidate the icache over | ||
96 | * that region so that that code can be run on all CPUs in the system. | ||
97 | */ | ||
98 | void flush_icache_range(unsigned long start, unsigned long end) | ||
99 | { | ||
100 | unsigned long start_page, end_page; | ||
101 | |||
102 | if (end > 0x80000000UL) { | ||
103 | /* addresses above 0xa0000000 do not go through the cache */ | ||
104 | if (end > 0xa0000000UL) { | ||
105 | end = 0xa0000000UL; | ||
106 | if (start >= end) | ||
107 | return; | ||
108 | } | ||
109 | |||
110 | /* kernel addresses between 0x80000000 and 0x9fffffff do not | ||
111 | * require page tables, so we just map such addresses | ||
112 | * directly */ | ||
113 | start_page = (start >= 0x80000000UL) ? start : 0x80000000UL; | ||
114 | mn10300_dcache_flush_range(start_page, end); | ||
115 | mn10300_icache_inv_range(start_page, end); | ||
116 | if (start_page == start) | ||
117 | return; | ||
118 | end = start_page; | ||
119 | } | ||
120 | |||
121 | start_page = start & PAGE_MASK; | ||
122 | end_page = end & PAGE_MASK; | ||
123 | |||
124 | if (start_page == end_page) { | ||
125 | /* the first and last bytes are on the same page */ | ||
126 | flush_icache_page_range(start, end); | ||
127 | } else if (start_page + 1 == end_page) { | ||
128 | /* split over two virtually contiguous pages */ | ||
129 | flush_icache_page_range(start, end_page); | ||
130 | flush_icache_page_range(end_page, end); | ||
131 | } else { | ||
132 | /* more than 2 pages; just flush the entire cache */ | ||
133 | mn10300_dcache_flush(); | ||
134 | mn10300_icache_inv(); | ||
135 | } | ||
136 | } | ||
137 | EXPORT_SYMBOL(flush_icache_range); | ||
diff --git a/arch/mn10300/mm/cache-inv-icache.c b/arch/mn10300/mm/cache-inv-icache.c new file mode 100644 index 000000000000..4a3f7afcfe53 --- /dev/null +++ b/arch/mn10300/mm/cache-inv-icache.c | |||
@@ -0,0 +1,119 @@ | |||
1 | /* Invalidate icache when dcache doesn't need invalidation as it's in | ||
2 | * write-through mode | ||
3 | * | ||
4 | * Copyright (C) 2010 Red Hat, Inc. All Rights Reserved. | ||
5 | * Written by David Howells (dhowells@redhat.com) | ||
6 | * | ||
7 | * This program is free software; you can redistribute it and/or | ||
8 | * modify it under the terms of the GNU General Public Licence | ||
9 | * as published by the Free Software Foundation; either version | ||
10 | * 2 of the Licence, or (at your option) any later version. | ||
11 | */ | ||
12 | #include <linux/module.h> | ||
13 | #include <linux/mm.h> | ||
14 | #include <asm/cacheflush.h> | ||
15 | |||
16 | /** | ||
17 | * flush_icache_page_range - Flush dcache and invalidate icache for part of a | ||
18 | * single page | ||
19 | * @start: The starting virtual address of the page part. | ||
20 | * @end: The ending virtual address of the page part. | ||
21 | * | ||
22 | * Invalidate the icache for part of a single page, as determined by the | ||
23 | * virtual addresses given. The page must be in the paged area. The dcache is | ||
24 | * not flushed as the cache must be in write-through mode to get here. | ||
25 | */ | ||
26 | static void flush_icache_page_range(unsigned long start, unsigned long end) | ||
27 | { | ||
28 | unsigned long addr, size, off; | ||
29 | struct page *page; | ||
30 | pgd_t *pgd; | ||
31 | pud_t *pud; | ||
32 | pmd_t *pmd; | ||
33 | pte_t *ppte, pte; | ||
34 | |||
35 | /* work out how much of the page to flush */ | ||
36 | off = start & ~PAGE_MASK; | ||
37 | size = end - start; | ||
38 | |||
39 | /* get the physical address the page is mapped to from the page | ||
40 | * tables */ | ||
41 | pgd = pgd_offset(current->mm, start); | ||
42 | if (!pgd || !pgd_val(*pgd)) | ||
43 | return; | ||
44 | |||
45 | pud = pud_offset(pgd, start); | ||
46 | if (!pud || !pud_val(*pud)) | ||
47 | return; | ||
48 | |||
49 | pmd = pmd_offset(pud, start); | ||
50 | if (!pmd || !pmd_val(*pmd)) | ||
51 | return; | ||
52 | |||
53 | ppte = pte_offset_map(pmd, start); | ||
54 | if (!ppte) | ||
55 | return; | ||
56 | pte = *ppte; | ||
57 | pte_unmap(ppte); | ||
58 | |||
59 | if (pte_none(pte)) | ||
60 | return; | ||
61 | |||
62 | page = pte_page(pte); | ||
63 | if (!page) | ||
64 | return; | ||
65 | |||
66 | addr = page_to_phys(page); | ||
67 | |||
68 | /* invalidate the icache coverage on that region */ | ||
69 | mn10300_icache_inv_range2(addr + off, size); | ||
70 | } | ||
71 | |||
72 | /** | ||
73 | * flush_icache_range - Globally flush dcache and invalidate icache for region | ||
74 | * @start: The starting virtual address of the region. | ||
75 | * @end: The ending virtual address of the region. | ||
76 | * | ||
77 | * This is used by the kernel to globally flush some code it has just written | ||
78 | * from the dcache back to RAM and then to globally invalidate the icache over | ||
79 | * that region so that that code can be run on all CPUs in the system. | ||
80 | */ | ||
81 | void flush_icache_range(unsigned long start, unsigned long end) | ||
82 | { | ||
83 | unsigned long start_page, end_page; | ||
84 | |||
85 | if (end > 0x80000000UL) { | ||
86 | /* addresses above 0xa0000000 do not go through the cache */ | ||
87 | if (end > 0xa0000000UL) { | ||
88 | end = 0xa0000000UL; | ||
89 | if (start >= end) | ||
90 | return; | ||
91 | } | ||
92 | |||
93 | /* kernel addresses between 0x80000000 and 0x9fffffff do not | ||
94 | * require page tables, so we just map such addresses | ||
95 | * directly */ | ||
96 | start_page = (start >= 0x80000000UL) ? start : 0x80000000UL; | ||
97 | mn10300_dcache_flush_range(start_page, end); | ||
98 | mn10300_icache_inv_range(start_page, end); | ||
99 | if (start_page == start) | ||
100 | return; | ||
101 | end = start_page; | ||
102 | } | ||
103 | |||
104 | start_page = start & PAGE_MASK; | ||
105 | end_page = end & PAGE_MASK; | ||
106 | |||
107 | if (start_page == end_page) { | ||
108 | /* the first and last bytes are on the same page */ | ||
109 | flush_icache_page_range(start, end); | ||
110 | } else if (start_page + 1 == end_page) { | ||
111 | /* split over two virtually contiguous pages */ | ||
112 | flush_icache_page_range(start, end_page); | ||
113 | flush_icache_page_range(end_page, end); | ||
114 | } else { | ||
115 | /* more than 2 pages; just flush the entire cache */ | ||
116 | mn10300_icache_inv(); | ||
117 | } | ||
118 | } | ||
119 | EXPORT_SYMBOL(flush_icache_range); | ||
diff --git a/arch/mn10300/mm/cache.c b/arch/mn10300/mm/cache.c index 9261217e8d2c..bc35826f1357 100644 --- a/arch/mn10300/mm/cache.c +++ b/arch/mn10300/mm/cache.c | |||
@@ -37,96 +37,6 @@ EXPORT_SYMBOL(mn10300_dcache_flush_page); | |||
37 | #endif | 37 | #endif |
38 | 38 | ||
39 | /* | 39 | /* |
40 | * write a page back from the dcache and invalidate the icache so that we can | ||
41 | * run code from it that we've just written into it | ||
42 | */ | ||
43 | void flush_icache_page(struct vm_area_struct *vma, struct page *page) | ||
44 | { | ||
45 | mn10300_dcache_flush_page(page_to_phys(page)); | ||
46 | mn10300_icache_inv(); | ||
47 | } | ||
48 | EXPORT_SYMBOL(flush_icache_page); | ||
49 | |||
50 | /* | ||
51 | * write some code we've just written back from the dcache and invalidate the | ||
52 | * icache so that we can run that code | ||
53 | */ | ||
54 | void flush_icache_range(unsigned long start, unsigned long end) | ||
55 | { | ||
56 | #ifdef CONFIG_MN10300_CACHE_WBACK | ||
57 | unsigned long addr, size, base, off; | ||
58 | struct page *page; | ||
59 | pgd_t *pgd; | ||
60 | pud_t *pud; | ||
61 | pmd_t *pmd; | ||
62 | pte_t *ppte, pte; | ||
63 | |||
64 | if (end > 0x80000000UL) { | ||
65 | /* addresses above 0xa0000000 do not go through the cache */ | ||
66 | if (end > 0xa0000000UL) { | ||
67 | end = 0xa0000000UL; | ||
68 | if (start >= end) | ||
69 | return; | ||
70 | } | ||
71 | |||
72 | /* kernel addresses between 0x80000000 and 0x9fffffff do not | ||
73 | * require page tables, so we just map such addresses directly */ | ||
74 | base = (start >= 0x80000000UL) ? start : 0x80000000UL; | ||
75 | mn10300_dcache_flush_range(base, end); | ||
76 | if (base == start) | ||
77 | goto invalidate; | ||
78 | end = base; | ||
79 | } | ||
80 | |||
81 | for (; start < end; start += size) { | ||
82 | /* work out how much of the page to flush */ | ||
83 | off = start & (PAGE_SIZE - 1); | ||
84 | |||
85 | size = end - start; | ||
86 | if (size > PAGE_SIZE - off) | ||
87 | size = PAGE_SIZE - off; | ||
88 | |||
89 | /* get the physical address the page is mapped to from the page | ||
90 | * tables */ | ||
91 | pgd = pgd_offset(current->mm, start); | ||
92 | if (!pgd || !pgd_val(*pgd)) | ||
93 | continue; | ||
94 | |||
95 | pud = pud_offset(pgd, start); | ||
96 | if (!pud || !pud_val(*pud)) | ||
97 | continue; | ||
98 | |||
99 | pmd = pmd_offset(pud, start); | ||
100 | if (!pmd || !pmd_val(*pmd)) | ||
101 | continue; | ||
102 | |||
103 | ppte = pte_offset_map(pmd, start); | ||
104 | if (!ppte) | ||
105 | continue; | ||
106 | pte = *ppte; | ||
107 | pte_unmap(ppte); | ||
108 | |||
109 | if (pte_none(pte)) | ||
110 | continue; | ||
111 | |||
112 | page = pte_page(pte); | ||
113 | if (!page) | ||
114 | continue; | ||
115 | |||
116 | addr = page_to_phys(page); | ||
117 | |||
118 | /* flush the dcache and invalidate the icache coverage on that | ||
119 | * region */ | ||
120 | mn10300_dcache_flush_range2(addr + off, size); | ||
121 | } | ||
122 | #endif | ||
123 | |||
124 | invalidate: | ||
125 | mn10300_icache_inv(); | ||
126 | } | ||
127 | EXPORT_SYMBOL(flush_icache_range); | ||
128 | |||
129 | /* | ||
130 | * allow userspace to flush the instruction cache | 40 | * allow userspace to flush the instruction cache |
131 | */ | 41 | */ |
132 | asmlinkage long sys_cacheflush(unsigned long start, unsigned long end) | 42 | asmlinkage long sys_cacheflush(unsigned long start, unsigned long end) |