aboutsummaryrefslogtreecommitdiffstats
path: root/arch
diff options
context:
space:
mode:
authorLinus Torvalds <torvalds@linux-foundation.org>2013-07-02 19:17:25 -0400
committerLinus Torvalds <torvalds@linux-foundation.org>2013-07-02 19:17:25 -0400
commit2d722f6d5671794c0de0e29e3da75006ac086718 (patch)
tree042c6320cfd771798f985e79aa0422cbe8096258 /arch
parentf0bb4c0ab064a8aeeffbda1cee380151a594eaab (diff)
parent2fd1b487884310d0aa0c0640179dc7490ad86313 (diff)
Merge branch 'sched-core-for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/tip/tip
Pull scheduler updates from Ingo Molnar: "The main changes: - load-calculation cleanups and improvements, by Alex Shi - various nohz related tidying up of statisics, by Frederic Weisbecker - factor out /proc functions to kernel/sched/proc.c, by Paul Gortmaker - simplify the RT policy scheduler, by Kirill Tkhai - various fixes and cleanups" * 'sched-core-for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/tip/tip: (42 commits) sched/debug: Remove CONFIG_FAIR_GROUP_SCHED mask sched/debug: Fix formatting of /proc/<PID>/sched sched: Fix typo in struct sched_avg member description sched/fair: Fix typo describing flags in enqueue_entity sched/debug: Add load-tracking statistics to task sched: Change get_rq_runnable_load() to static and inline sched/tg: Remove tg.load_weight sched/cfs_rq: Change atomic64_t removed_load to atomic_long_t sched/tg: Use 'unsigned long' for load variable in task group sched: Change cfs_rq load avg to unsigned long sched: Consider runnable load average in move_tasks() sched: Compute runnable load avg in cpu_load and cpu_avg_load_per_task sched: Update cpu load after task_tick sched: Fix sleep time double accounting in enqueue entity sched: Set an initial value of runnable avg for new forked task sched: Move a few runnable tg variables into CONFIG_SMP Revert "sched: Introduce temporary FAIR_GROUP_SCHED dependency for load-tracking" sched: Don't mix use of typedef ctl_table and struct ctl_table sched: Remove WARN_ON(!sd) from init_sched_groups_power() sched: Fix memory leakage in build_sched_groups() ...
Diffstat (limited to 'arch')
-rw-r--r--arch/avr32/kernel/process.c2
-rw-r--r--arch/cris/include/arch-v10/arch/bitops.h2
-rw-r--r--arch/ia64/kernel/head.S2
-rw-r--r--arch/mips/kernel/mips-mt-fpaff.c4
-rw-r--r--arch/mips/kernel/scall32-o32.S5
-rw-r--r--arch/powerpc/include/asm/mmu_context.h2
-rw-r--r--arch/tile/include/asm/processor.h2
-rw-r--r--arch/tile/kernel/stack.c2
-rw-r--r--arch/um/kernel/sysrq.c2
9 files changed, 12 insertions, 11 deletions
diff --git a/arch/avr32/kernel/process.c b/arch/avr32/kernel/process.c
index e7b61494c312..c2731003edef 100644
--- a/arch/avr32/kernel/process.c
+++ b/arch/avr32/kernel/process.c
@@ -341,7 +341,7 @@ unsigned long get_wchan(struct task_struct *p)
341 * is actually quite ugly. It might be possible to 341 * is actually quite ugly. It might be possible to
342 * determine the frame size automatically at build 342 * determine the frame size automatically at build
343 * time by doing this: 343 * time by doing this:
344 * - compile sched.c 344 * - compile sched/core.c
345 * - disassemble the resulting sched.o 345 * - disassemble the resulting sched.o
346 * - look for 'sub sp,??' shortly after '<schedule>:' 346 * - look for 'sub sp,??' shortly after '<schedule>:'
347 */ 347 */
diff --git a/arch/cris/include/arch-v10/arch/bitops.h b/arch/cris/include/arch-v10/arch/bitops.h
index be85f6de25d3..03d9cfd92c8a 100644
--- a/arch/cris/include/arch-v10/arch/bitops.h
+++ b/arch/cris/include/arch-v10/arch/bitops.h
@@ -17,7 +17,7 @@ static inline unsigned long cris_swapnwbrlz(unsigned long w)
17 in another register: 17 in another register:
18 ! __asm__ ("swapnwbr %2\n\tlz %2,%0" 18 ! __asm__ ("swapnwbr %2\n\tlz %2,%0"
19 ! : "=r,r" (res), "=r,X" (dummy) : "1,0" (w)); 19 ! : "=r,r" (res), "=r,X" (dummy) : "1,0" (w));
20 confuses gcc (sched.c, gcc from cris-dist-1.14). */ 20 confuses gcc (core.c, gcc from cris-dist-1.14). */
21 21
22 unsigned long res; 22 unsigned long res;
23 __asm__ ("swapnwbr %0 \n\t" 23 __asm__ ("swapnwbr %0 \n\t"
diff --git a/arch/ia64/kernel/head.S b/arch/ia64/kernel/head.S
index 9be4e497f3d3..991ca336b8a2 100644
--- a/arch/ia64/kernel/head.S
+++ b/arch/ia64/kernel/head.S
@@ -1035,7 +1035,7 @@ END(ia64_delay_loop)
1035 * Return a CPU-local timestamp in nano-seconds. This timestamp is 1035 * Return a CPU-local timestamp in nano-seconds. This timestamp is
1036 * NOT synchronized across CPUs its return value must never be 1036 * NOT synchronized across CPUs its return value must never be
1037 * compared against the values returned on another CPU. The usage in 1037 * compared against the values returned on another CPU. The usage in
1038 * kernel/sched.c ensures that. 1038 * kernel/sched/core.c ensures that.
1039 * 1039 *
1040 * The return-value of sched_clock() is NOT supposed to wrap-around. 1040 * The return-value of sched_clock() is NOT supposed to wrap-around.
1041 * If it did, it would cause some scheduling hiccups (at the worst). 1041 * If it did, it would cause some scheduling hiccups (at the worst).
diff --git a/arch/mips/kernel/mips-mt-fpaff.c b/arch/mips/kernel/mips-mt-fpaff.c
index fd814e08c945..cb098628aee8 100644
--- a/arch/mips/kernel/mips-mt-fpaff.c
+++ b/arch/mips/kernel/mips-mt-fpaff.c
@@ -27,12 +27,12 @@ unsigned long mt_fpemul_threshold;
27 * FPU affinity with the user's requested processor affinity. 27 * FPU affinity with the user's requested processor affinity.
28 * This code is 98% identical with the sys_sched_setaffinity() 28 * This code is 98% identical with the sys_sched_setaffinity()
29 * and sys_sched_getaffinity() system calls, and should be 29 * and sys_sched_getaffinity() system calls, and should be
30 * updated when kernel/sched.c changes. 30 * updated when kernel/sched/core.c changes.
31 */ 31 */
32 32
33/* 33/*
34 * find_process_by_pid - find a process with a matching PID value. 34 * find_process_by_pid - find a process with a matching PID value.
35 * used in sys_sched_set/getaffinity() in kernel/sched.c, so 35 * used in sys_sched_set/getaffinity() in kernel/sched/core.c, so
36 * cloned here. 36 * cloned here.
37 */ 37 */
38static inline struct task_struct *find_process_by_pid(pid_t pid) 38static inline struct task_struct *find_process_by_pid(pid_t pid)
diff --git a/arch/mips/kernel/scall32-o32.S b/arch/mips/kernel/scall32-o32.S
index 9b36424b03c5..e9127ec612ef 100644
--- a/arch/mips/kernel/scall32-o32.S
+++ b/arch/mips/kernel/scall32-o32.S
@@ -476,8 +476,9 @@ einval: li v0, -ENOSYS
476 /* 476 /*
477 * For FPU affinity scheduling on MIPS MT processors, we need to 477 * For FPU affinity scheduling on MIPS MT processors, we need to
478 * intercept sys_sched_xxxaffinity() calls until we get a proper hook 478 * intercept sys_sched_xxxaffinity() calls until we get a proper hook
479 * in kernel/sched.c. Considered only temporary we only support these 479 * in kernel/sched/core.c. Considered only temporary we only support
480 * hooks for the 32-bit kernel - there is no MIPS64 MT processor atm. 480 * these hooks for the 32-bit kernel - there is no MIPS64 MT processor
481 * atm.
481 */ 482 */
482 sys mipsmt_sys_sched_setaffinity 3 483 sys mipsmt_sys_sched_setaffinity 3
483 sys mipsmt_sys_sched_getaffinity 3 484 sys mipsmt_sys_sched_getaffinity 3
diff --git a/arch/powerpc/include/asm/mmu_context.h b/arch/powerpc/include/asm/mmu_context.h
index a73668a5f30d..b467530e2485 100644
--- a/arch/powerpc/include/asm/mmu_context.h
+++ b/arch/powerpc/include/asm/mmu_context.h
@@ -38,7 +38,7 @@ extern void drop_cop(unsigned long acop, struct mm_struct *mm);
38 38
39/* 39/*
40 * switch_mm is the entry point called from the architecture independent 40 * switch_mm is the entry point called from the architecture independent
41 * code in kernel/sched.c 41 * code in kernel/sched/core.c
42 */ 42 */
43static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, 43static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next,
44 struct task_struct *tsk) 44 struct task_struct *tsk)
diff --git a/arch/tile/include/asm/processor.h b/arch/tile/include/asm/processor.h
index 2b70dfb1442e..b3f104953da2 100644
--- a/arch/tile/include/asm/processor.h
+++ b/arch/tile/include/asm/processor.h
@@ -225,7 +225,7 @@ extern int do_work_pending(struct pt_regs *regs, u32 flags);
225 225
226/* 226/*
227 * Return saved (kernel) PC of a blocked thread. 227 * Return saved (kernel) PC of a blocked thread.
228 * Only used in a printk() in kernel/sched.c, so don't work too hard. 228 * Only used in a printk() in kernel/sched/core.c, so don't work too hard.
229 */ 229 */
230#define thread_saved_pc(t) ((t)->thread.pc) 230#define thread_saved_pc(t) ((t)->thread.pc)
231 231
diff --git a/arch/tile/kernel/stack.c b/arch/tile/kernel/stack.c
index ed258b8ae320..af8dfc9665f6 100644
--- a/arch/tile/kernel/stack.c
+++ b/arch/tile/kernel/stack.c
@@ -442,7 +442,7 @@ void _KBacktraceIterator_init_current(struct KBacktraceIterator *kbt, ulong pc,
442 regs_to_pt_regs(&regs, pc, lr, sp, r52)); 442 regs_to_pt_regs(&regs, pc, lr, sp, r52));
443} 443}
444 444
445/* This is called only from kernel/sched.c, with esp == NULL */ 445/* This is called only from kernel/sched/core.c, with esp == NULL */
446void show_stack(struct task_struct *task, unsigned long *esp) 446void show_stack(struct task_struct *task, unsigned long *esp)
447{ 447{
448 struct KBacktraceIterator kbt; 448 struct KBacktraceIterator kbt;
diff --git a/arch/um/kernel/sysrq.c b/arch/um/kernel/sysrq.c
index 7d101a2a1541..0dc4d1c6f98a 100644
--- a/arch/um/kernel/sysrq.c
+++ b/arch/um/kernel/sysrq.c
@@ -39,7 +39,7 @@ void show_trace(struct task_struct *task, unsigned long * stack)
39static const int kstack_depth_to_print = 24; 39static const int kstack_depth_to_print = 24;
40 40
41/* This recently started being used in arch-independent code too, as in 41/* This recently started being used in arch-independent code too, as in
42 * kernel/sched.c.*/ 42 * kernel/sched/core.c.*/
43void show_stack(struct task_struct *task, unsigned long *esp) 43void show_stack(struct task_struct *task, unsigned long *esp)
44{ 44{
45 unsigned long *stack; 45 unsigned long *stack;