diff options
Diffstat (limited to 'arch/microblaze/kernel/cpu/cache.c')
| -rw-r--r-- | arch/microblaze/kernel/cpu/cache.c | 258 |
1 files changed, 258 insertions, 0 deletions
diff --git a/arch/microblaze/kernel/cpu/cache.c b/arch/microblaze/kernel/cpu/cache.c new file mode 100644 index 000000000000..be9fecca4f91 --- /dev/null +++ b/arch/microblaze/kernel/cpu/cache.c | |||
| @@ -0,0 +1,258 @@ | |||
| 1 | /* | ||
| 2 | * Cache control for MicroBlaze cache memories | ||
| 3 | * | ||
| 4 | * Copyright (C) 2007-2009 Michal Simek <monstr@monstr.eu> | ||
| 5 | * Copyright (C) 2007-2009 PetaLogix | ||
| 6 | * Copyright (C) 2007 John Williams <john.williams@petalogix.com> | ||
| 7 | * | ||
| 8 | * This file is subject to the terms and conditions of the GNU General | ||
| 9 | * Public License. See the file COPYING in the main directory of this | ||
| 10 | * archive for more details. | ||
| 11 | */ | ||
| 12 | |||
| 13 | #include <asm/cacheflush.h> | ||
| 14 | #include <linux/cache.h> | ||
| 15 | #include <asm/cpuinfo.h> | ||
| 16 | |||
| 17 | /* Exported functions */ | ||
| 18 | |||
| 19 | void _enable_icache(void) | ||
| 20 | { | ||
| 21 | if (cpuinfo.use_icache) { | ||
| 22 | #if CONFIG_XILINX_MICROBLAZE0_USE_MSR_INSTR | ||
| 23 | __asm__ __volatile__ (" \ | ||
| 24 | msrset r0, %0; \ | ||
| 25 | nop; " \ | ||
| 26 | : \ | ||
| 27 | : "i" (MSR_ICE) \ | ||
| 28 | : "memory"); | ||
| 29 | #else | ||
| 30 | __asm__ __volatile__ (" \ | ||
| 31 | mfs r12, rmsr; \ | ||
| 32 | nop; \ | ||
| 33 | ori r12, r12, %0; \ | ||
| 34 | mts rmsr, r12; \ | ||
| 35 | nop; " \ | ||
| 36 | : \ | ||
| 37 | : "i" (MSR_ICE) \ | ||
| 38 | : "memory", "r12"); | ||
| 39 | #endif | ||
| 40 | } | ||
| 41 | } | ||
| 42 | |||
| 43 | void _disable_icache(void) | ||
| 44 | { | ||
| 45 | if (cpuinfo.use_icache) { | ||
| 46 | #if CONFIG_XILINX_MICROBLAZE0_USE_MSR_INSTR | ||
| 47 | __asm__ __volatile__ (" \ | ||
| 48 | msrclr r0, %0; \ | ||
| 49 | nop; " \ | ||
| 50 | : \ | ||
| 51 | : "i" (MSR_ICE) \ | ||
| 52 | : "memory"); | ||
| 53 | #else | ||
| 54 | __asm__ __volatile__ (" \ | ||
| 55 | mfs r12, rmsr; \ | ||
| 56 | nop; \ | ||
| 57 | andi r12, r12, ~%0; \ | ||
| 58 | mts rmsr, r12; \ | ||
| 59 | nop; " \ | ||
| 60 | : \ | ||
| 61 | : "i" (MSR_ICE) \ | ||
| 62 | : "memory", "r12"); | ||
| 63 | #endif | ||
| 64 | } | ||
| 65 | } | ||
| 66 | |||
| 67 | void _invalidate_icache(unsigned int addr) | ||
| 68 | { | ||
| 69 | if (cpuinfo.use_icache) { | ||
| 70 | __asm__ __volatile__ (" \ | ||
| 71 | wic %0, r0" \ | ||
| 72 | : \ | ||
| 73 | : "r" (addr)); | ||
| 74 | } | ||
| 75 | } | ||
| 76 | |||
| 77 | void _enable_dcache(void) | ||
| 78 | { | ||
| 79 | if (cpuinfo.use_dcache) { | ||
| 80 | #if CONFIG_XILINX_MICROBLAZE0_USE_MSR_INSTR | ||
| 81 | __asm__ __volatile__ (" \ | ||
| 82 | msrset r0, %0; \ | ||
| 83 | nop; " \ | ||
| 84 | : \ | ||
| 85 | : "i" (MSR_DCE) \ | ||
| 86 | : "memory"); | ||
| 87 | #else | ||
| 88 | __asm__ __volatile__ (" \ | ||
| 89 | mfs r12, rmsr; \ | ||
| 90 | nop; \ | ||
| 91 | ori r12, r12, %0; \ | ||
| 92 | mts rmsr, r12; \ | ||
| 93 | nop; " \ | ||
| 94 | : \ | ||
| 95 | : "i" (MSR_DCE) \ | ||
| 96 | : "memory", "r12"); | ||
| 97 | #endif | ||
| 98 | } | ||
| 99 | } | ||
| 100 | |||
| 101 | void _disable_dcache(void) | ||
| 102 | { | ||
| 103 | if (cpuinfo.use_dcache) { | ||
| 104 | #if CONFIG_XILINX_MICROBLAZE0_USE_MSR_INSTR | ||
| 105 | __asm__ __volatile__ (" \ | ||
| 106 | msrclr r0, %0; \ | ||
| 107 | nop; " \ | ||
| 108 | : \ | ||
| 109 | : "i" (MSR_DCE) \ | ||
| 110 | : "memory"); | ||
| 111 | #else | ||
| 112 | __asm__ __volatile__ (" \ | ||
| 113 | mfs r12, rmsr; \ | ||
| 114 | nop; \ | ||
| 115 | andi r12, r12, ~%0; \ | ||
| 116 | mts rmsr, r12; \ | ||
| 117 | nop; " \ | ||
| 118 | : \ | ||
| 119 | : "i" (MSR_DCE) \ | ||
| 120 | : "memory", "r12"); | ||
| 121 | #endif | ||
| 122 | } | ||
| 123 | } | ||
| 124 | |||
| 125 | void _invalidate_dcache(unsigned int addr) | ||
| 126 | { | ||
| 127 | if (cpuinfo.use_dcache) | ||
| 128 | __asm__ __volatile__ (" \ | ||
| 129 | wdc %0, r0" \ | ||
| 130 | : \ | ||
| 131 | : "r" (addr)); | ||
| 132 | } | ||
| 133 | |||
| 134 | void __invalidate_icache_all(void) | ||
| 135 | { | ||
| 136 | unsigned int i; | ||
| 137 | unsigned flags; | ||
| 138 | |||
| 139 | if (cpuinfo.use_icache) { | ||
| 140 | local_irq_save(flags); | ||
| 141 | __disable_icache(); | ||
| 142 | |||
| 143 | /* Just loop through cache size and invalidate, no need to add | ||
| 144 | CACHE_BASE address */ | ||
| 145 | for (i = 0; i < cpuinfo.icache_size; | ||
| 146 | i += cpuinfo.icache_line) | ||
| 147 | __invalidate_icache(i); | ||
| 148 | |||
| 149 | __enable_icache(); | ||
| 150 | local_irq_restore(flags); | ||
| 151 | } | ||
| 152 | } | ||
| 153 | |||
| 154 | void __invalidate_icache_range(unsigned long start, unsigned long end) | ||
| 155 | { | ||
| 156 | unsigned int i; | ||
| 157 | unsigned flags; | ||
| 158 | unsigned int align; | ||
| 159 | |||
| 160 | if (cpuinfo.use_icache) { | ||
| 161 | /* | ||
| 162 | * No need to cover entire cache range, | ||
| 163 | * just cover cache footprint | ||
| 164 | */ | ||
| 165 | end = min(start + cpuinfo.icache_size, end); | ||
| 166 | align = ~(cpuinfo.icache_line - 1); | ||
| 167 | start &= align; /* Make sure we are aligned */ | ||
| 168 | /* Push end up to the next cache line */ | ||
| 169 | end = ((end & align) + cpuinfo.icache_line); | ||
| 170 | |||
| 171 | local_irq_save(flags); | ||
| 172 | __disable_icache(); | ||
| 173 | |||
| 174 | for (i = start; i < end; i += cpuinfo.icache_line) | ||
| 175 | __invalidate_icache(i); | ||
| 176 | |||
| 177 | __enable_icache(); | ||
| 178 | local_irq_restore(flags); | ||
| 179 | } | ||
| 180 | } | ||
| 181 | |||
| 182 | void __invalidate_icache_page(struct vm_area_struct *vma, struct page *page) | ||
| 183 | { | ||
| 184 | __invalidate_icache_all(); | ||
| 185 | } | ||
| 186 | |||
| 187 | void __invalidate_icache_user_range(struct vm_area_struct *vma, | ||
| 188 | struct page *page, unsigned long adr, | ||
| 189 | int len) | ||
| 190 | { | ||
| 191 | __invalidate_icache_all(); | ||
| 192 | } | ||
| 193 | |||
| 194 | void __invalidate_cache_sigtramp(unsigned long addr) | ||
| 195 | { | ||
| 196 | __invalidate_icache_range(addr, addr + 8); | ||
| 197 | } | ||
| 198 | |||
| 199 | void __invalidate_dcache_all(void) | ||
| 200 | { | ||
| 201 | unsigned int i; | ||
| 202 | unsigned flags; | ||
| 203 | |||
| 204 | if (cpuinfo.use_dcache) { | ||
| 205 | local_irq_save(flags); | ||
| 206 | __disable_dcache(); | ||
| 207 | |||
| 208 | /* | ||
| 209 | * Just loop through cache size and invalidate, | ||
| 210 | * no need to add CACHE_BASE address | ||
| 211 | */ | ||
| 212 | for (i = 0; i < cpuinfo.dcache_size; | ||
| 213 | i += cpuinfo.dcache_line) | ||
| 214 | __invalidate_dcache(i); | ||
| 215 | |||
| 216 | __enable_dcache(); | ||
| 217 | local_irq_restore(flags); | ||
| 218 | } | ||
| 219 | } | ||
| 220 | |||
| 221 | void __invalidate_dcache_range(unsigned long start, unsigned long end) | ||
| 222 | { | ||
| 223 | unsigned int i; | ||
| 224 | unsigned flags; | ||
| 225 | unsigned int align; | ||
| 226 | |||
| 227 | if (cpuinfo.use_dcache) { | ||
| 228 | /* | ||
| 229 | * No need to cover entire cache range, | ||
| 230 | * just cover cache footprint | ||
| 231 | */ | ||
| 232 | end = min(start + cpuinfo.dcache_size, end); | ||
| 233 | align = ~(cpuinfo.dcache_line - 1); | ||
| 234 | start &= align; /* Make sure we are aligned */ | ||
| 235 | /* Push end up to the next cache line */ | ||
| 236 | end = ((end & align) + cpuinfo.dcache_line); | ||
| 237 | local_irq_save(flags); | ||
| 238 | __disable_dcache(); | ||
| 239 | |||
| 240 | for (i = start; i < end; i += cpuinfo.dcache_line) | ||
| 241 | __invalidate_dcache(i); | ||
| 242 | |||
| 243 | __enable_dcache(); | ||
| 244 | local_irq_restore(flags); | ||
| 245 | } | ||
| 246 | } | ||
| 247 | |||
| 248 | void __invalidate_dcache_page(struct vm_area_struct *vma, struct page *page) | ||
| 249 | { | ||
| 250 | __invalidate_dcache_all(); | ||
| 251 | } | ||
| 252 | |||
| 253 | void __invalidate_dcache_user_range(struct vm_area_struct *vma, | ||
| 254 | struct page *page, unsigned long adr, | ||
| 255 | int len) | ||
| 256 | { | ||
| 257 | __invalidate_dcache_all(); | ||
| 258 | } | ||
