diff options
author | Russell King <rmk@dyn-67.arm.linux.org.uk> | 2007-07-18 04:37:10 -0400 |
---|---|---|
committer | Russell King <rmk+kernel@arm.linux.org.uk> | 2007-07-20 04:39:57 -0400 |
commit | 228adef16d6e7b7725ef6b9ba760810d5966afa5 (patch) | |
tree | f7473090e2284a7f3b2933d97e684e4b2445d79c /arch/arm/vfp/vfpmodule.c | |
parent | 21d1ca04532005c50ed57c2b2948e465b2e90720 (diff) |
[ARM] vfp: make fpexc bit names less verbose
Use the fpexc abbreviated names instead of long verbose names
for fpexc bits.
Signed-off-by: Russell King <rmk+kernel@arm.linux.org.uk>
Diffstat (limited to 'arch/arm/vfp/vfpmodule.c')
-rw-r--r-- | arch/arm/vfp/vfpmodule.c | 12 |
1 files changed, 6 insertions, 6 deletions
diff --git a/arch/arm/vfp/vfpmodule.c b/arch/arm/vfp/vfpmodule.c index 1106b5f9cf19..04ddab2bd876 100644 --- a/arch/arm/vfp/vfpmodule.c +++ b/arch/arm/vfp/vfpmodule.c | |||
@@ -53,7 +53,7 @@ static int vfp_notifier(struct notifier_block *self, unsigned long cmd, void *v) | |||
53 | * case the thread migrates to a different CPU. The | 53 | * case the thread migrates to a different CPU. The |
54 | * restoring is done lazily. | 54 | * restoring is done lazily. |
55 | */ | 55 | */ |
56 | if ((fpexc & FPEXC_ENABLE) && last_VFP_context[cpu]) { | 56 | if ((fpexc & FPEXC_EN) && last_VFP_context[cpu]) { |
57 | vfp_save_state(last_VFP_context[cpu], fpexc); | 57 | vfp_save_state(last_VFP_context[cpu], fpexc); |
58 | last_VFP_context[cpu]->hard.cpu = cpu; | 58 | last_VFP_context[cpu]->hard.cpu = cpu; |
59 | } | 59 | } |
@@ -70,7 +70,7 @@ static int vfp_notifier(struct notifier_block *self, unsigned long cmd, void *v) | |||
70 | * Always disable VFP so we can lazily save/restore the | 70 | * Always disable VFP so we can lazily save/restore the |
71 | * old state. | 71 | * old state. |
72 | */ | 72 | */ |
73 | fmxr(FPEXC, fpexc & ~FPEXC_ENABLE); | 73 | fmxr(FPEXC, fpexc & ~FPEXC_EN); |
74 | return NOTIFY_DONE; | 74 | return NOTIFY_DONE; |
75 | } | 75 | } |
76 | 76 | ||
@@ -81,13 +81,13 @@ static int vfp_notifier(struct notifier_block *self, unsigned long cmd, void *v) | |||
81 | */ | 81 | */ |
82 | memset(vfp, 0, sizeof(union vfp_state)); | 82 | memset(vfp, 0, sizeof(union vfp_state)); |
83 | 83 | ||
84 | vfp->hard.fpexc = FPEXC_ENABLE; | 84 | vfp->hard.fpexc = FPEXC_EN; |
85 | vfp->hard.fpscr = FPSCR_ROUND_NEAREST; | 85 | vfp->hard.fpscr = FPSCR_ROUND_NEAREST; |
86 | 86 | ||
87 | /* | 87 | /* |
88 | * Disable VFP to ensure we initialise it first. | 88 | * Disable VFP to ensure we initialise it first. |
89 | */ | 89 | */ |
90 | fmxr(FPEXC, fmrx(FPEXC) & ~FPEXC_ENABLE); | 90 | fmxr(FPEXC, fmrx(FPEXC) & ~FPEXC_EN); |
91 | } | 91 | } |
92 | 92 | ||
93 | /* flush and release case: Per-thread VFP cleanup. */ | 93 | /* flush and release case: Per-thread VFP cleanup. */ |
@@ -229,7 +229,7 @@ void VFP9_bounce(u32 trigger, u32 fpexc, struct pt_regs *regs) | |||
229 | /* | 229 | /* |
230 | * Enable access to the VFP so we can handle the bounce. | 230 | * Enable access to the VFP so we can handle the bounce. |
231 | */ | 231 | */ |
232 | fmxr(FPEXC, fpexc & ~(FPEXC_EXCEPTION|FPEXC_INV|FPEXC_UFC|FPEXC_IOC)); | 232 | fmxr(FPEXC, fpexc & ~(FPEXC_EX|FPEXC_INV|FPEXC_UFC|FPEXC_IOC)); |
233 | 233 | ||
234 | orig_fpscr = fpscr = fmrx(FPSCR); | 234 | orig_fpscr = fpscr = fmrx(FPSCR); |
235 | 235 | ||
@@ -248,7 +248,7 @@ void VFP9_bounce(u32 trigger, u32 fpexc, struct pt_regs *regs) | |||
248 | /* | 248 | /* |
249 | * Modify fpscr to indicate the number of iterations remaining | 249 | * Modify fpscr to indicate the number of iterations remaining |
250 | */ | 250 | */ |
251 | if (fpexc & FPEXC_EXCEPTION) { | 251 | if (fpexc & FPEXC_EX) { |
252 | u32 len; | 252 | u32 len; |
253 | 253 | ||
254 | len = fpexc + (1 << FPEXC_LENGTH_BIT); | 254 | len = fpexc + (1 << FPEXC_LENGTH_BIT); |