aboutsummaryrefslogtreecommitdiffstats
path: root/include/linux/sched.h
diff options
context:
space:
mode:
authorLinus Torvalds <torvalds@linux-foundation.org>2009-12-12 14:34:10 -0500
committerLinus Torvalds <torvalds@linux-foundation.org>2009-12-12 14:34:10 -0500
commit702a7c7609bec3a940b6a46b0d6ab9ce45274580 (patch)
tree6c169691449259410b9b51a146acb0e837dae96a /include/linux/sched.h
parent053fe57ac249a9531c396175778160d9e9509399 (diff)
parentb9889ed1ddeca5a3f3569c8de7354e9e97d803ae (diff)
Merge branch 'sched-fixes-for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/tip/linux-2.6-tip
* 'sched-fixes-for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/tip/linux-2.6-tip: (21 commits) sched: Remove forced2_migrations stats sched: Fix memory leak in two error corner cases sched: Fix build warning in get_update_sysctl_factor() sched: Update normalized values on user updates via proc sched: Make tunable scaling style configurable sched: Fix missing sched tunable recalculation on cpu add/remove sched: Fix task priority bug sched: cgroup: Implement different treatment for idle shares sched: Remove unnecessary RCU exclusion sched: Discard some old bits sched: Clean up check_preempt_wakeup() sched: Move update_curr() in check_preempt_wakeup() to avoid redundant call sched: Sanitize fork() handling sched: Clean up ttwu() rq locking sched: Remove rq->clock coupling from set_task_cpu() sched: Consolidate select_task_rq() callers sched: Remove sysctl.sched_features sched: Protect sched_rr_get_param() access to task->sched_class sched: Protect task->cpus_allowed access in sched_getaffinity() sched: Fix balance vs hotplug race ... Fixed up conflicts in kernel/sysctl.c (due to sysctl cleanup)
Diffstat (limited to 'include/linux/sched.h')
-rw-r--r--include/linux/sched.h20
1 files changed, 13 insertions, 7 deletions
diff --git a/include/linux/sched.h b/include/linux/sched.h
index 3f4fa73b512a..294eb2f80144 100644
--- a/include/linux/sched.h
+++ b/include/linux/sched.h
@@ -1102,7 +1102,7 @@ struct sched_class {
1102 1102
1103 void (*set_curr_task) (struct rq *rq); 1103 void (*set_curr_task) (struct rq *rq);
1104 void (*task_tick) (struct rq *rq, struct task_struct *p, int queued); 1104 void (*task_tick) (struct rq *rq, struct task_struct *p, int queued);
1105 void (*task_new) (struct rq *rq, struct task_struct *p); 1105 void (*task_fork) (struct task_struct *p);
1106 1106
1107 void (*switched_from) (struct rq *this_rq, struct task_struct *task, 1107 void (*switched_from) (struct rq *this_rq, struct task_struct *task,
1108 int running); 1108 int running);
@@ -1111,7 +1111,8 @@ struct sched_class {
1111 void (*prio_changed) (struct rq *this_rq, struct task_struct *task, 1111 void (*prio_changed) (struct rq *this_rq, struct task_struct *task,
1112 int oldprio, int running); 1112 int oldprio, int running);
1113 1113
1114 unsigned int (*get_rr_interval) (struct task_struct *task); 1114 unsigned int (*get_rr_interval) (struct rq *rq,
1115 struct task_struct *task);
1115 1116
1116#ifdef CONFIG_FAIR_GROUP_SCHED 1117#ifdef CONFIG_FAIR_GROUP_SCHED
1117 void (*moved_group) (struct task_struct *p); 1118 void (*moved_group) (struct task_struct *p);
@@ -1151,8 +1152,6 @@ struct sched_entity {
1151 u64 start_runtime; 1152 u64 start_runtime;
1152 u64 avg_wakeup; 1153 u64 avg_wakeup;
1153 1154
1154 u64 avg_running;
1155
1156#ifdef CONFIG_SCHEDSTATS 1155#ifdef CONFIG_SCHEDSTATS
1157 u64 wait_start; 1156 u64 wait_start;
1158 u64 wait_max; 1157 u64 wait_max;
@@ -1175,7 +1174,6 @@ struct sched_entity {
1175 u64 nr_failed_migrations_running; 1174 u64 nr_failed_migrations_running;
1176 u64 nr_failed_migrations_hot; 1175 u64 nr_failed_migrations_hot;
1177 u64 nr_forced_migrations; 1176 u64 nr_forced_migrations;
1178 u64 nr_forced2_migrations;
1179 1177
1180 u64 nr_wakeups; 1178 u64 nr_wakeups;
1181 u64 nr_wakeups_sync; 1179 u64 nr_wakeups_sync;
@@ -1904,14 +1902,22 @@ extern unsigned int sysctl_sched_wakeup_granularity;
1904extern unsigned int sysctl_sched_shares_ratelimit; 1902extern unsigned int sysctl_sched_shares_ratelimit;
1905extern unsigned int sysctl_sched_shares_thresh; 1903extern unsigned int sysctl_sched_shares_thresh;
1906extern unsigned int sysctl_sched_child_runs_first; 1904extern unsigned int sysctl_sched_child_runs_first;
1905
1906enum sched_tunable_scaling {
1907 SCHED_TUNABLESCALING_NONE,
1908 SCHED_TUNABLESCALING_LOG,
1909 SCHED_TUNABLESCALING_LINEAR,
1910 SCHED_TUNABLESCALING_END,
1911};
1912extern enum sched_tunable_scaling sysctl_sched_tunable_scaling;
1913
1907#ifdef CONFIG_SCHED_DEBUG 1914#ifdef CONFIG_SCHED_DEBUG
1908extern unsigned int sysctl_sched_features;
1909extern unsigned int sysctl_sched_migration_cost; 1915extern unsigned int sysctl_sched_migration_cost;
1910extern unsigned int sysctl_sched_nr_migrate; 1916extern unsigned int sysctl_sched_nr_migrate;
1911extern unsigned int sysctl_sched_time_avg; 1917extern unsigned int sysctl_sched_time_avg;
1912extern unsigned int sysctl_timer_migration; 1918extern unsigned int sysctl_timer_migration;
1913 1919
1914int sched_nr_latency_handler(struct ctl_table *table, int write, 1920int sched_proc_update_handler(struct ctl_table *table, int write,
1915 void __user *buffer, size_t *length, 1921 void __user *buffer, size_t *length,
1916 loff_t *ppos); 1922 loff_t *ppos);
1917#endif 1923#endif