diff options
| author | Lucas De Marchi <lucas.demarchi@profusion.mobi> | 2011-03-30 21:57:33 -0400 |
|---|---|---|
| committer | Lucas De Marchi <lucas.demarchi@profusion.mobi> | 2011-03-31 10:26:23 -0400 |
| commit | 25985edcedea6396277003854657b5f3cb31a628 (patch) | |
| tree | f026e810210a2ee7290caeb737c23cb6472b7c38 /arch/powerpc/kernel | |
| parent | 6aba74f2791287ec407e0f92487a725a25908067 (diff) | |
Fix common misspellings
Fixes generated by 'codespell' and manually reviewed.
Signed-off-by: Lucas De Marchi <lucas.demarchi@profusion.mobi>
Diffstat (limited to 'arch/powerpc/kernel')
| -rw-r--r-- | arch/powerpc/kernel/btext.c | 2 | ||||
| -rw-r--r-- | arch/powerpc/kernel/exceptions-64e.S | 2 | ||||
| -rw-r--r-- | arch/powerpc/kernel/exceptions-64s.S | 2 | ||||
| -rw-r--r-- | arch/powerpc/kernel/head_40x.S | 2 | ||||
| -rw-r--r-- | arch/powerpc/kernel/head_44x.S | 2 | ||||
| -rw-r--r-- | arch/powerpc/kernel/head_64.S | 2 | ||||
| -rw-r--r-- | arch/powerpc/kernel/head_fsl_booke.S | 2 | ||||
| -rw-r--r-- | arch/powerpc/kernel/l2cr_6xx.S | 2 | ||||
| -rw-r--r-- | arch/powerpc/kernel/lparcfg.c | 2 | ||||
| -rw-r--r-- | arch/powerpc/kernel/perf_event.c | 2 | ||||
| -rw-r--r-- | arch/powerpc/kernel/ppc_save_regs.S | 2 | ||||
| -rw-r--r-- | arch/powerpc/kernel/prom.c | 4 | ||||
| -rw-r--r-- | arch/powerpc/kernel/ptrace.c | 2 | ||||
| -rw-r--r-- | arch/powerpc/kernel/rtasd.c | 2 | ||||
| -rw-r--r-- | arch/powerpc/kernel/swsusp_32.S | 2 | ||||
| -rw-r--r-- | arch/powerpc/kernel/traps.c | 2 | ||||
| -rw-r--r-- | arch/powerpc/kernel/udbg_16550.c | 2 | ||||
| -rw-r--r-- | arch/powerpc/kernel/vdso32/sigtramp.S | 2 | ||||
| -rw-r--r-- | arch/powerpc/kernel/vdso64/sigtramp.S | 2 |
19 files changed, 20 insertions, 20 deletions
diff --git a/arch/powerpc/kernel/btext.c b/arch/powerpc/kernel/btext.c index 625942ae5585..60b3e377b1e4 100644 --- a/arch/powerpc/kernel/btext.c +++ b/arch/powerpc/kernel/btext.c | |||
| @@ -99,7 +99,7 @@ void __init btext_prepare_BAT(void) | |||
| 99 | 99 | ||
| 100 | /* This function can be used to enable the early boot text when doing | 100 | /* This function can be used to enable the early boot text when doing |
| 101 | * OF booting or within bootx init. It must be followed by a btext_unmap() | 101 | * OF booting or within bootx init. It must be followed by a btext_unmap() |
| 102 | * call before the logical address becomes unuseable | 102 | * call before the logical address becomes unusable |
| 103 | */ | 103 | */ |
| 104 | void __init btext_setup_display(int width, int height, int depth, int pitch, | 104 | void __init btext_setup_display(int width, int height, int depth, int pitch, |
| 105 | unsigned long address) | 105 | unsigned long address) |
diff --git a/arch/powerpc/kernel/exceptions-64e.S b/arch/powerpc/kernel/exceptions-64e.S index 5c43063d2506..9651acc3504a 100644 --- a/arch/powerpc/kernel/exceptions-64e.S +++ b/arch/powerpc/kernel/exceptions-64e.S | |||
| @@ -379,7 +379,7 @@ interrupt_end_book3e: | |||
| 379 | mfspr r13,SPRN_SPRG_PACA /* get our PACA */ | 379 | mfspr r13,SPRN_SPRG_PACA /* get our PACA */ |
| 380 | b system_call_common | 380 | b system_call_common |
| 381 | 381 | ||
| 382 | /* Auxillary Processor Unavailable Interrupt */ | 382 | /* Auxiliary Processor Unavailable Interrupt */ |
| 383 | START_EXCEPTION(ap_unavailable); | 383 | START_EXCEPTION(ap_unavailable); |
| 384 | NORMAL_EXCEPTION_PROLOG(0xf20, PROLOG_ADDITION_NONE) | 384 | NORMAL_EXCEPTION_PROLOG(0xf20, PROLOG_ADDITION_NONE) |
| 385 | EXCEPTION_COMMON(0xf20, PACA_EXGEN, INTS_KEEP) | 385 | EXCEPTION_COMMON(0xf20, PACA_EXGEN, INTS_KEEP) |
diff --git a/arch/powerpc/kernel/exceptions-64s.S b/arch/powerpc/kernel/exceptions-64s.S index c532cb2c927a..aeb739e18769 100644 --- a/arch/powerpc/kernel/exceptions-64s.S +++ b/arch/powerpc/kernel/exceptions-64s.S | |||
| @@ -5,7 +5,7 @@ | |||
| 5 | * handling and other fixed offset specific things. | 5 | * handling and other fixed offset specific things. |
| 6 | * | 6 | * |
| 7 | * This file is meant to be #included from head_64.S due to | 7 | * This file is meant to be #included from head_64.S due to |
| 8 | * position dependant assembly. | 8 | * position dependent assembly. |
| 9 | * | 9 | * |
| 10 | * Most of this originates from head_64.S and thus has the same | 10 | * Most of this originates from head_64.S and thus has the same |
| 11 | * copyright history. | 11 | * copyright history. |
diff --git a/arch/powerpc/kernel/head_40x.S b/arch/powerpc/kernel/head_40x.S index 9dd21a8c4d52..a91626d87fc9 100644 --- a/arch/powerpc/kernel/head_40x.S +++ b/arch/powerpc/kernel/head_40x.S | |||
| @@ -766,7 +766,7 @@ DataAccess: | |||
| 766 | * miss get to this point to load the TLB. | 766 | * miss get to this point to load the TLB. |
| 767 | * r10 - TLB_TAG value | 767 | * r10 - TLB_TAG value |
| 768 | * r11 - Linux PTE | 768 | * r11 - Linux PTE |
| 769 | * r12, r9 - avilable to use | 769 | * r12, r9 - available to use |
| 770 | * PID - loaded with proper value when we get here | 770 | * PID - loaded with proper value when we get here |
| 771 | * Upon exit, we reload everything and RFI. | 771 | * Upon exit, we reload everything and RFI. |
| 772 | * Actually, it will fit now, but oh well.....a common place | 772 | * Actually, it will fit now, but oh well.....a common place |
diff --git a/arch/powerpc/kernel/head_44x.S b/arch/powerpc/kernel/head_44x.S index cbb3436b592d..5e12b741ba5f 100644 --- a/arch/powerpc/kernel/head_44x.S +++ b/arch/powerpc/kernel/head_44x.S | |||
| @@ -178,7 +178,7 @@ interrupt_base: | |||
| 178 | NORMAL_EXCEPTION_PROLOG | 178 | NORMAL_EXCEPTION_PROLOG |
| 179 | EXC_XFER_EE_LITE(0x0c00, DoSyscall) | 179 | EXC_XFER_EE_LITE(0x0c00, DoSyscall) |
| 180 | 180 | ||
| 181 | /* Auxillary Processor Unavailable Interrupt */ | 181 | /* Auxiliary Processor Unavailable Interrupt */ |
| 182 | EXCEPTION(0x2020, AuxillaryProcessorUnavailable, unknown_exception, EXC_XFER_EE) | 182 | EXCEPTION(0x2020, AuxillaryProcessorUnavailable, unknown_exception, EXC_XFER_EE) |
| 183 | 183 | ||
| 184 | /* Decrementer Interrupt */ | 184 | /* Decrementer Interrupt */ |
diff --git a/arch/powerpc/kernel/head_64.S b/arch/powerpc/kernel/head_64.S index 782f23df7c85..285e6f775bdf 100644 --- a/arch/powerpc/kernel/head_64.S +++ b/arch/powerpc/kernel/head_64.S | |||
| @@ -40,7 +40,7 @@ | |||
| 40 | #include <asm/kvm_book3s_asm.h> | 40 | #include <asm/kvm_book3s_asm.h> |
| 41 | #include <asm/ptrace.h> | 41 | #include <asm/ptrace.h> |
| 42 | 42 | ||
| 43 | /* The physical memory is layed out such that the secondary processor | 43 | /* The physical memory is laid out such that the secondary processor |
| 44 | * spin code sits at 0x0000...0x00ff. On server, the vectors follow | 44 | * spin code sits at 0x0000...0x00ff. On server, the vectors follow |
| 45 | * using the layout described in exceptions-64s.S | 45 | * using the layout described in exceptions-64s.S |
| 46 | */ | 46 | */ |
diff --git a/arch/powerpc/kernel/head_fsl_booke.S b/arch/powerpc/kernel/head_fsl_booke.S index 3e02710d9562..5ecf54cfa7d4 100644 --- a/arch/powerpc/kernel/head_fsl_booke.S +++ b/arch/powerpc/kernel/head_fsl_booke.S | |||
| @@ -326,7 +326,7 @@ interrupt_base: | |||
| 326 | NORMAL_EXCEPTION_PROLOG | 326 | NORMAL_EXCEPTION_PROLOG |
| 327 | EXC_XFER_EE_LITE(0x0c00, DoSyscall) | 327 | EXC_XFER_EE_LITE(0x0c00, DoSyscall) |
| 328 | 328 | ||
| 329 | /* Auxillary Processor Unavailable Interrupt */ | 329 | /* Auxiliary Processor Unavailable Interrupt */ |
| 330 | EXCEPTION(0x2900, AuxillaryProcessorUnavailable, unknown_exception, EXC_XFER_EE) | 330 | EXCEPTION(0x2900, AuxillaryProcessorUnavailable, unknown_exception, EXC_XFER_EE) |
| 331 | 331 | ||
| 332 | /* Decrementer Interrupt */ | 332 | /* Decrementer Interrupt */ |
diff --git a/arch/powerpc/kernel/l2cr_6xx.S b/arch/powerpc/kernel/l2cr_6xx.S index 2a2f3c3f6d80..97ec8557f974 100644 --- a/arch/powerpc/kernel/l2cr_6xx.S +++ b/arch/powerpc/kernel/l2cr_6xx.S | |||
| @@ -151,7 +151,7 @@ END_FTR_SECTION_IFSET(CPU_FTR_ALTIVEC) | |||
| 151 | /**** Might be a good idea to set L2DO here - to prevent instructions | 151 | /**** Might be a good idea to set L2DO here - to prevent instructions |
| 152 | from getting into the cache. But since we invalidate | 152 | from getting into the cache. But since we invalidate |
| 153 | the next time we enable the cache it doesn't really matter. | 153 | the next time we enable the cache it doesn't really matter. |
| 154 | Don't do this unless you accomodate all processor variations. | 154 | Don't do this unless you accommodate all processor variations. |
| 155 | The bit moved on the 7450..... | 155 | The bit moved on the 7450..... |
| 156 | ****/ | 156 | ****/ |
| 157 | 157 | ||
diff --git a/arch/powerpc/kernel/lparcfg.c b/arch/powerpc/kernel/lparcfg.c index 16468362ad57..301db65f05a1 100644 --- a/arch/powerpc/kernel/lparcfg.c +++ b/arch/powerpc/kernel/lparcfg.c | |||
| @@ -262,7 +262,7 @@ static void parse_ppp_data(struct seq_file *m) | |||
| 262 | seq_printf(m, "system_active_processors=%d\n", | 262 | seq_printf(m, "system_active_processors=%d\n", |
| 263 | ppp_data.active_system_procs); | 263 | ppp_data.active_system_procs); |
| 264 | 264 | ||
| 265 | /* pool related entries are apropriate for shared configs */ | 265 | /* pool related entries are appropriate for shared configs */ |
| 266 | if (lppaca_of(0).shared_proc) { | 266 | if (lppaca_of(0).shared_proc) { |
| 267 | unsigned long pool_idle_time, pool_procs; | 267 | unsigned long pool_idle_time, pool_procs; |
| 268 | 268 | ||
diff --git a/arch/powerpc/kernel/perf_event.c b/arch/powerpc/kernel/perf_event.c index 97e0ae414940..c4063b7f49a0 100644 --- a/arch/powerpc/kernel/perf_event.c +++ b/arch/powerpc/kernel/perf_event.c | |||
| @@ -759,7 +759,7 @@ static int power_pmu_add(struct perf_event *event, int ef_flags) | |||
| 759 | 759 | ||
| 760 | /* | 760 | /* |
| 761 | * If group events scheduling transaction was started, | 761 | * If group events scheduling transaction was started, |
| 762 | * skip the schedulability test here, it will be peformed | 762 | * skip the schedulability test here, it will be performed |
| 763 | * at commit time(->commit_txn) as a whole | 763 | * at commit time(->commit_txn) as a whole |
| 764 | */ | 764 | */ |
| 765 | if (cpuhw->group_flag & PERF_EVENT_TXN) | 765 | if (cpuhw->group_flag & PERF_EVENT_TXN) |
diff --git a/arch/powerpc/kernel/ppc_save_regs.S b/arch/powerpc/kernel/ppc_save_regs.S index e83ba3f078e4..1b1787d52896 100644 --- a/arch/powerpc/kernel/ppc_save_regs.S +++ b/arch/powerpc/kernel/ppc_save_regs.S | |||
| @@ -15,7 +15,7 @@ | |||
| 15 | 15 | ||
| 16 | /* | 16 | /* |
| 17 | * Grab the register values as they are now. | 17 | * Grab the register values as they are now. |
| 18 | * This won't do a particularily good job because we really | 18 | * This won't do a particularly good job because we really |
| 19 | * want our caller's caller's registers, and our caller has | 19 | * want our caller's caller's registers, and our caller has |
| 20 | * already executed its prologue. | 20 | * already executed its prologue. |
| 21 | * ToDo: We could reach back into the caller's save area to do | 21 | * ToDo: We could reach back into the caller's save area to do |
diff --git a/arch/powerpc/kernel/prom.c b/arch/powerpc/kernel/prom.c index 05b7139d6a27..e74fa12afc82 100644 --- a/arch/powerpc/kernel/prom.c +++ b/arch/powerpc/kernel/prom.c | |||
| @@ -683,7 +683,7 @@ void __init early_init_devtree(void *params) | |||
| 683 | #endif | 683 | #endif |
| 684 | 684 | ||
| 685 | #ifdef CONFIG_PHYP_DUMP | 685 | #ifdef CONFIG_PHYP_DUMP |
| 686 | /* scan tree to see if dump occured during last boot */ | 686 | /* scan tree to see if dump occurred during last boot */ |
| 687 | of_scan_flat_dt(early_init_dt_scan_phyp_dump, NULL); | 687 | of_scan_flat_dt(early_init_dt_scan_phyp_dump, NULL); |
| 688 | #endif | 688 | #endif |
| 689 | 689 | ||
| @@ -739,7 +739,7 @@ void __init early_init_devtree(void *params) | |||
| 739 | 739 | ||
| 740 | DBG("Scanning CPUs ...\n"); | 740 | DBG("Scanning CPUs ...\n"); |
| 741 | 741 | ||
| 742 | /* Retreive CPU related informations from the flat tree | 742 | /* Retrieve CPU related informations from the flat tree |
| 743 | * (altivec support, boot CPU ID, ...) | 743 | * (altivec support, boot CPU ID, ...) |
| 744 | */ | 744 | */ |
| 745 | of_scan_flat_dt(early_init_dt_scan_cpus, NULL); | 745 | of_scan_flat_dt(early_init_dt_scan_cpus, NULL); |
diff --git a/arch/powerpc/kernel/ptrace.c b/arch/powerpc/kernel/ptrace.c index 895b082f1e48..55613e33e263 100644 --- a/arch/powerpc/kernel/ptrace.c +++ b/arch/powerpc/kernel/ptrace.c | |||
| @@ -463,7 +463,7 @@ static int vr_set(struct task_struct *target, const struct user_regset *regset, | |||
| 463 | #ifdef CONFIG_VSX | 463 | #ifdef CONFIG_VSX |
| 464 | /* | 464 | /* |
| 465 | * Currently to set and and get all the vsx state, you need to call | 465 | * Currently to set and and get all the vsx state, you need to call |
| 466 | * the fp and VMX calls aswell. This only get/sets the lower 32 | 466 | * the fp and VMX calls as well. This only get/sets the lower 32 |
| 467 | * 128bit VSX registers. | 467 | * 128bit VSX registers. |
| 468 | */ | 468 | */ |
| 469 | 469 | ||
diff --git a/arch/powerpc/kernel/rtasd.c b/arch/powerpc/kernel/rtasd.c index 7980ec0e1e1a..67f6c3b51357 100644 --- a/arch/powerpc/kernel/rtasd.c +++ b/arch/powerpc/kernel/rtasd.c | |||
| @@ -465,7 +465,7 @@ static void start_event_scan(void) | |||
| 465 | pr_debug("rtasd: will sleep for %d milliseconds\n", | 465 | pr_debug("rtasd: will sleep for %d milliseconds\n", |
| 466 | (30000 / rtas_event_scan_rate)); | 466 | (30000 / rtas_event_scan_rate)); |
| 467 | 467 | ||
| 468 | /* Retreive errors from nvram if any */ | 468 | /* Retrieve errors from nvram if any */ |
| 469 | retreive_nvram_error_log(); | 469 | retreive_nvram_error_log(); |
| 470 | 470 | ||
| 471 | schedule_delayed_work_on(cpumask_first(cpu_online_mask), | 471 | schedule_delayed_work_on(cpumask_first(cpu_online_mask), |
diff --git a/arch/powerpc/kernel/swsusp_32.S b/arch/powerpc/kernel/swsusp_32.S index b0754e237438..ba4dee3d233f 100644 --- a/arch/powerpc/kernel/swsusp_32.S +++ b/arch/powerpc/kernel/swsusp_32.S | |||
| @@ -143,7 +143,7 @@ END_FTR_SECTION_IFSET(CPU_FTR_ALTIVEC) | |||
| 143 | 143 | ||
| 144 | /* Disable MSR:DR to make sure we don't take a TLB or | 144 | /* Disable MSR:DR to make sure we don't take a TLB or |
| 145 | * hash miss during the copy, as our hash table will | 145 | * hash miss during the copy, as our hash table will |
| 146 | * for a while be unuseable. For .text, we assume we are | 146 | * for a while be unusable. For .text, we assume we are |
| 147 | * covered by a BAT. This works only for non-G5 at this | 147 | * covered by a BAT. This works only for non-G5 at this |
| 148 | * point. G5 will need a better approach, possibly using | 148 | * point. G5 will need a better approach, possibly using |
| 149 | * a small temporary hash table filled with large mappings, | 149 | * a small temporary hash table filled with large mappings, |
diff --git a/arch/powerpc/kernel/traps.c b/arch/powerpc/kernel/traps.c index bd74fac169be..5ddb801bc154 100644 --- a/arch/powerpc/kernel/traps.c +++ b/arch/powerpc/kernel/traps.c | |||
| @@ -959,7 +959,7 @@ void __kprobes program_check_exception(struct pt_regs *regs) | |||
| 959 | * ESR_DST (!?) or 0. In the process of chasing this with the | 959 | * ESR_DST (!?) or 0. In the process of chasing this with the |
| 960 | * hardware people - not sure if it can happen on any illegal | 960 | * hardware people - not sure if it can happen on any illegal |
| 961 | * instruction or only on FP instructions, whether there is a | 961 | * instruction or only on FP instructions, whether there is a |
| 962 | * pattern to occurences etc. -dgibson 31/Mar/2003 */ | 962 | * pattern to occurrences etc. -dgibson 31/Mar/2003 */ |
| 963 | switch (do_mathemu(regs)) { | 963 | switch (do_mathemu(regs)) { |
| 964 | case 0: | 964 | case 0: |
| 965 | emulate_single_step(regs); | 965 | emulate_single_step(regs); |
diff --git a/arch/powerpc/kernel/udbg_16550.c b/arch/powerpc/kernel/udbg_16550.c index b4b167b33643..baa33a7517bc 100644 --- a/arch/powerpc/kernel/udbg_16550.c +++ b/arch/powerpc/kernel/udbg_16550.c | |||
| @@ -1,5 +1,5 @@ | |||
| 1 | /* | 1 | /* |
| 2 | * udbg for NS16550 compatable serial ports | 2 | * udbg for NS16550 compatible serial ports |
| 3 | * | 3 | * |
| 4 | * Copyright (C) 2001-2005 PPC 64 Team, IBM Corp | 4 | * Copyright (C) 2001-2005 PPC 64 Team, IBM Corp |
| 5 | * | 5 | * |
diff --git a/arch/powerpc/kernel/vdso32/sigtramp.S b/arch/powerpc/kernel/vdso32/sigtramp.S index 68d49dd71dcc..cf0c9c9c24f9 100644 --- a/arch/powerpc/kernel/vdso32/sigtramp.S +++ b/arch/powerpc/kernel/vdso32/sigtramp.S | |||
| @@ -19,7 +19,7 @@ | |||
| 19 | 19 | ||
| 20 | /* The nop here is a hack. The dwarf2 unwind routines subtract 1 from | 20 | /* The nop here is a hack. The dwarf2 unwind routines subtract 1 from |
| 21 | the return address to get an address in the middle of the presumed | 21 | the return address to get an address in the middle of the presumed |
| 22 | call instruction. Since we don't have a call here, we artifically | 22 | call instruction. Since we don't have a call here, we artificially |
| 23 | extend the range covered by the unwind info by adding a nop before | 23 | extend the range covered by the unwind info by adding a nop before |
| 24 | the real start. */ | 24 | the real start. */ |
| 25 | nop | 25 | nop |
diff --git a/arch/powerpc/kernel/vdso64/sigtramp.S b/arch/powerpc/kernel/vdso64/sigtramp.S index 59eb59bb4082..45ea281e9a21 100644 --- a/arch/powerpc/kernel/vdso64/sigtramp.S +++ b/arch/powerpc/kernel/vdso64/sigtramp.S | |||
| @@ -20,7 +20,7 @@ | |||
| 20 | 20 | ||
| 21 | /* The nop here is a hack. The dwarf2 unwind routines subtract 1 from | 21 | /* The nop here is a hack. The dwarf2 unwind routines subtract 1 from |
| 22 | the return address to get an address in the middle of the presumed | 22 | the return address to get an address in the middle of the presumed |
| 23 | call instruction. Since we don't have a call here, we artifically | 23 | call instruction. Since we don't have a call here, we artificially |
| 24 | extend the range covered by the unwind info by padding before the | 24 | extend the range covered by the unwind info by padding before the |
| 25 | real start. */ | 25 | real start. */ |
| 26 | nop | 26 | nop |
