diff options
author | Russell King <rmk+kernel@arm.linux.org.uk> | 2014-06-30 11:29:12 -0400 |
---|---|---|
committer | Russell King <rmk+kernel@arm.linux.org.uk> | 2014-07-18 07:29:04 -0400 |
commit | 6ebbf2ce437b33022d30badd49dc94d33ecfa498 (patch) | |
tree | bc015e35b456a28bb0e501803a454dc0c0d3291a /arch/arm/mm/proc-v6.S | |
parent | af040ffc9ba1e079ee4c0748aff64fa3d4716fa5 (diff) |
ARM: convert all "mov.* pc, reg" to "bx reg" for ARMv6+
ARMv6 and greater introduced a new instruction ("bx") which can be used
to return from function calls. Recent CPUs perform better when the
"bx lr" instruction is used rather than the "mov pc, lr" instruction,
and this sequence is strongly recommended to be used by the ARM
architecture manual (section A.4.1.1).
We provide a new macro "ret" with all its variants for the condition
code which will resolve to the appropriate instruction.
Rather than doing this piecemeal, and miss some instances, change all
the "mov pc" instances to use the new macro, with the exception of
the "movs" instruction and the kprobes code. This allows us to detect
the "mov pc, lr" case and fix it up - and also gives us the possibility
of deploying this for other registers depending on the CPU selection.
Reported-by: Will Deacon <will.deacon@arm.com>
Tested-by: Stephen Warren <swarren@nvidia.com> # Tegra Jetson TK1
Tested-by: Robert Jarzmik <robert.jarzmik@free.fr> # mioa701_bootresume.S
Tested-by: Andrew Lunn <andrew@lunn.ch> # Kirkwood
Tested-by: Shawn Guo <shawn.guo@freescale.com>
Tested-by: Tony Lindgren <tony@atomide.com> # OMAPs
Tested-by: Gregory CLEMENT <gregory.clement@free-electrons.com> # Armada XP, 375, 385
Acked-by: Sekhar Nori <nsekhar@ti.com> # DaVinci
Acked-by: Christoffer Dall <christoffer.dall@linaro.org> # kvm/hyp
Acked-by: Haojian Zhuang <haojian.zhuang@gmail.com> # PXA3xx
Acked-by: Stefano Stabellini <stefano.stabellini@eu.citrix.com> # Xen
Tested-by: Uwe Kleine-König <u.kleine-koenig@pengutronix.de> # ARMv7M
Tested-by: Simon Horman <horms+renesas@verge.net.au> # Shmobile
Signed-off-by: Russell King <rmk+kernel@arm.linux.org.uk>
Diffstat (limited to 'arch/arm/mm/proc-v6.S')
-rw-r--r-- | arch/arm/mm/proc-v6.S | 16 |
1 files changed, 8 insertions, 8 deletions
diff --git a/arch/arm/mm/proc-v6.S b/arch/arm/mm/proc-v6.S index 32b3558321c4..d0390f4b3f18 100644 --- a/arch/arm/mm/proc-v6.S +++ b/arch/arm/mm/proc-v6.S | |||
@@ -36,14 +36,14 @@ | |||
36 | #define PMD_FLAGS_SMP PMD_SECT_WBWA|PMD_SECT_S | 36 | #define PMD_FLAGS_SMP PMD_SECT_WBWA|PMD_SECT_S |
37 | 37 | ||
38 | ENTRY(cpu_v6_proc_init) | 38 | ENTRY(cpu_v6_proc_init) |
39 | mov pc, lr | 39 | ret lr |
40 | 40 | ||
41 | ENTRY(cpu_v6_proc_fin) | 41 | ENTRY(cpu_v6_proc_fin) |
42 | mrc p15, 0, r0, c1, c0, 0 @ ctrl register | 42 | mrc p15, 0, r0, c1, c0, 0 @ ctrl register |
43 | bic r0, r0, #0x1000 @ ...i............ | 43 | bic r0, r0, #0x1000 @ ...i............ |
44 | bic r0, r0, #0x0006 @ .............ca. | 44 | bic r0, r0, #0x0006 @ .............ca. |
45 | mcr p15, 0, r0, c1, c0, 0 @ disable caches | 45 | mcr p15, 0, r0, c1, c0, 0 @ disable caches |
46 | mov pc, lr | 46 | ret lr |
47 | 47 | ||
48 | /* | 48 | /* |
49 | * cpu_v6_reset(loc) | 49 | * cpu_v6_reset(loc) |
@@ -62,7 +62,7 @@ ENTRY(cpu_v6_reset) | |||
62 | mcr p15, 0, r1, c1, c0, 0 @ disable MMU | 62 | mcr p15, 0, r1, c1, c0, 0 @ disable MMU |
63 | mov r1, #0 | 63 | mov r1, #0 |
64 | mcr p15, 0, r1, c7, c5, 4 @ ISB | 64 | mcr p15, 0, r1, c7, c5, 4 @ ISB |
65 | mov pc, r0 | 65 | ret r0 |
66 | ENDPROC(cpu_v6_reset) | 66 | ENDPROC(cpu_v6_reset) |
67 | .popsection | 67 | .popsection |
68 | 68 | ||
@@ -77,14 +77,14 @@ ENTRY(cpu_v6_do_idle) | |||
77 | mov r1, #0 | 77 | mov r1, #0 |
78 | mcr p15, 0, r1, c7, c10, 4 @ DWB - WFI may enter a low-power mode | 78 | mcr p15, 0, r1, c7, c10, 4 @ DWB - WFI may enter a low-power mode |
79 | mcr p15, 0, r1, c7, c0, 4 @ wait for interrupt | 79 | mcr p15, 0, r1, c7, c0, 4 @ wait for interrupt |
80 | mov pc, lr | 80 | ret lr |
81 | 81 | ||
82 | ENTRY(cpu_v6_dcache_clean_area) | 82 | ENTRY(cpu_v6_dcache_clean_area) |
83 | 1: mcr p15, 0, r0, c7, c10, 1 @ clean D entry | 83 | 1: mcr p15, 0, r0, c7, c10, 1 @ clean D entry |
84 | add r0, r0, #D_CACHE_LINE_SIZE | 84 | add r0, r0, #D_CACHE_LINE_SIZE |
85 | subs r1, r1, #D_CACHE_LINE_SIZE | 85 | subs r1, r1, #D_CACHE_LINE_SIZE |
86 | bhi 1b | 86 | bhi 1b |
87 | mov pc, lr | 87 | ret lr |
88 | 88 | ||
89 | /* | 89 | /* |
90 | * cpu_v6_switch_mm(pgd_phys, tsk) | 90 | * cpu_v6_switch_mm(pgd_phys, tsk) |
@@ -113,7 +113,7 @@ ENTRY(cpu_v6_switch_mm) | |||
113 | #endif | 113 | #endif |
114 | mcr p15, 0, r1, c13, c0, 1 @ set context ID | 114 | mcr p15, 0, r1, c13, c0, 1 @ set context ID |
115 | #endif | 115 | #endif |
116 | mov pc, lr | 116 | ret lr |
117 | 117 | ||
118 | /* | 118 | /* |
119 | * cpu_v6_set_pte_ext(ptep, pte, ext) | 119 | * cpu_v6_set_pte_ext(ptep, pte, ext) |
@@ -131,7 +131,7 @@ ENTRY(cpu_v6_set_pte_ext) | |||
131 | #ifdef CONFIG_MMU | 131 | #ifdef CONFIG_MMU |
132 | armv6_set_pte_ext cpu_v6 | 132 | armv6_set_pte_ext cpu_v6 |
133 | #endif | 133 | #endif |
134 | mov pc, lr | 134 | ret lr |
135 | 135 | ||
136 | /* Suspend/resume support: taken from arch/arm/mach-s3c64xx/sleep.S */ | 136 | /* Suspend/resume support: taken from arch/arm/mach-s3c64xx/sleep.S */ |
137 | .globl cpu_v6_suspend_size | 137 | .globl cpu_v6_suspend_size |
@@ -241,7 +241,7 @@ __v6_setup: | |||
241 | mcreq p15, 0, r5, c1, c0, 1 @ write aux control reg | 241 | mcreq p15, 0, r5, c1, c0, 1 @ write aux control reg |
242 | orreq r0, r0, #(1 << 21) @ low interrupt latency configuration | 242 | orreq r0, r0, #(1 << 21) @ low interrupt latency configuration |
243 | #endif | 243 | #endif |
244 | mov pc, lr @ return to head.S:__ret | 244 | ret lr @ return to head.S:__ret |
245 | 245 | ||
246 | /* | 246 | /* |
247 | * V X F I D LR | 247 | * V X F I D LR |