diff options
| author | Clark Williams <williams@redhat.com> | 2013-02-07 10:47:07 -0500 |
|---|---|---|
| committer | Ingo Molnar <mingo@kernel.org> | 2013-02-07 14:51:08 -0500 |
| commit | 8bd75c77b7c6a3954140dd2e20346aef3efe4a35 (patch) | |
| tree | 10e0d451a58aeb6c8f48b871a848276bf3a8a359 /include/linux | |
| parent | ce0dbbbb30aee6a835511d5be446462388ba9eee (diff) | |
sched/rt: Move rt specific bits into new header file
Move rt scheduler definitions out of include/linux/sched.h into
new file include/linux/sched/rt.h
Signed-off-by: Clark Williams <williams@redhat.com>
Cc: Peter Zijlstra <peterz@infradead.org>
Cc: Steven Rostedt <rostedt@goodmis.org>
Link: http://lkml.kernel.org/r/20130207094707.7b9f825f@riff.lan
Signed-off-by: Ingo Molnar <mingo@kernel.org>
Diffstat (limited to 'include/linux')
| -rw-r--r-- | include/linux/sched.h | 55 | ||||
| -rw-r--r-- | include/linux/sched/rt.h | 58 |
2 files changed, 60 insertions, 53 deletions
diff --git a/include/linux/sched.h b/include/linux/sched.h index 8fc9b2710a80..33cc42130371 100644 --- a/include/linux/sched.h +++ b/include/linux/sched.h | |||
| @@ -1164,6 +1164,7 @@ struct sched_entity { | |||
| 1164 | /* rq "owned" by this entity/group: */ | 1164 | /* rq "owned" by this entity/group: */ |
| 1165 | struct cfs_rq *my_q; | 1165 | struct cfs_rq *my_q; |
| 1166 | #endif | 1166 | #endif |
| 1167 | |||
| 1167 | /* | 1168 | /* |
| 1168 | * Load-tracking only depends on SMP, FAIR_GROUP_SCHED dependency below may be | 1169 | * Load-tracking only depends on SMP, FAIR_GROUP_SCHED dependency below may be |
| 1169 | * removed when useful for applications beyond shares distribution (e.g. | 1170 | * removed when useful for applications beyond shares distribution (e.g. |
| @@ -1191,6 +1192,7 @@ struct sched_rt_entity { | |||
| 1191 | #endif | 1192 | #endif |
| 1192 | }; | 1193 | }; |
| 1193 | 1194 | ||
| 1195 | |||
| 1194 | struct rcu_node; | 1196 | struct rcu_node; |
| 1195 | 1197 | ||
| 1196 | enum perf_event_task_context { | 1198 | enum perf_event_task_context { |
| @@ -1596,37 +1598,6 @@ static inline void set_numabalancing_state(bool enabled) | |||
| 1596 | } | 1598 | } |
| 1597 | #endif | 1599 | #endif |
| 1598 | 1600 | ||
| 1599 | /* | ||
| 1600 | * Priority of a process goes from 0..MAX_PRIO-1, valid RT | ||
| 1601 | * priority is 0..MAX_RT_PRIO-1, and SCHED_NORMAL/SCHED_BATCH | ||
| 1602 | * tasks are in the range MAX_RT_PRIO..MAX_PRIO-1. Priority | ||
| 1603 | * values are inverted: lower p->prio value means higher priority. | ||
| 1604 | * | ||
| 1605 | * The MAX_USER_RT_PRIO value allows the actual maximum | ||
| 1606 | * RT priority to be separate from the value exported to | ||
| 1607 | * user-space. This allows kernel threads to set their | ||
| 1608 | * priority to a value higher than any user task. Note: | ||
| 1609 | * MAX_RT_PRIO must not be smaller than MAX_USER_RT_PRIO. | ||
| 1610 | */ | ||
| 1611 | |||
| 1612 | #define MAX_USER_RT_PRIO 100 | ||
| 1613 | #define MAX_RT_PRIO MAX_USER_RT_PRIO | ||
| 1614 | |||
| 1615 | #define MAX_PRIO (MAX_RT_PRIO + 40) | ||
| 1616 | #define DEFAULT_PRIO (MAX_RT_PRIO + 20) | ||
| 1617 | |||
| 1618 | static inline int rt_prio(int prio) | ||
| 1619 | { | ||
| 1620 | if (unlikely(prio < MAX_RT_PRIO)) | ||
| 1621 | return 1; | ||
| 1622 | return 0; | ||
| 1623 | } | ||
| 1624 | |||
| 1625 | static inline int rt_task(struct task_struct *p) | ||
| 1626 | { | ||
| 1627 | return rt_prio(p->prio); | ||
| 1628 | } | ||
| 1629 | |||
| 1630 | static inline struct pid *task_pid(struct task_struct *task) | 1601 | static inline struct pid *task_pid(struct task_struct *task) |
| 1631 | { | 1602 | { |
| 1632 | return task->pids[PIDTYPE_PID].pid; | 1603 | return task->pids[PIDTYPE_PID].pid; |
| @@ -2054,26 +2025,6 @@ static inline void sched_autogroup_fork(struct signal_struct *sig) { } | |||
| 2054 | static inline void sched_autogroup_exit(struct signal_struct *sig) { } | 2025 | static inline void sched_autogroup_exit(struct signal_struct *sig) { } |
| 2055 | #endif | 2026 | #endif |
| 2056 | 2027 | ||
| 2057 | #ifdef CONFIG_RT_MUTEXES | ||
| 2058 | extern int rt_mutex_getprio(struct task_struct *p); | ||
| 2059 | extern void rt_mutex_setprio(struct task_struct *p, int prio); | ||
| 2060 | extern void rt_mutex_adjust_pi(struct task_struct *p); | ||
| 2061 | static inline bool tsk_is_pi_blocked(struct task_struct *tsk) | ||
| 2062 | { | ||
| 2063 | return tsk->pi_blocked_on != NULL; | ||
| 2064 | } | ||
| 2065 | #else | ||
| 2066 | static inline int rt_mutex_getprio(struct task_struct *p) | ||
| 2067 | { | ||
| 2068 | return p->normal_prio; | ||
| 2069 | } | ||
| 2070 | # define rt_mutex_adjust_pi(p) do { } while (0) | ||
| 2071 | static inline bool tsk_is_pi_blocked(struct task_struct *tsk) | ||
| 2072 | { | ||
| 2073 | return false; | ||
| 2074 | } | ||
| 2075 | #endif | ||
| 2076 | |||
| 2077 | extern bool yield_to(struct task_struct *p, bool preempt); | 2028 | extern bool yield_to(struct task_struct *p, bool preempt); |
| 2078 | extern void set_user_nice(struct task_struct *p, long nice); | 2029 | extern void set_user_nice(struct task_struct *p, long nice); |
| 2079 | extern int task_prio(const struct task_struct *p); | 2030 | extern int task_prio(const struct task_struct *p); |
| @@ -2703,8 +2654,6 @@ static inline void set_task_cpu(struct task_struct *p, unsigned int cpu) | |||
| 2703 | extern long sched_setaffinity(pid_t pid, const struct cpumask *new_mask); | 2654 | extern long sched_setaffinity(pid_t pid, const struct cpumask *new_mask); |
| 2704 | extern long sched_getaffinity(pid_t pid, struct cpumask *mask); | 2655 | extern long sched_getaffinity(pid_t pid, struct cpumask *mask); |
| 2705 | 2656 | ||
| 2706 | extern void normalize_rt_tasks(void); | ||
| 2707 | |||
| 2708 | #ifdef CONFIG_CGROUP_SCHED | 2657 | #ifdef CONFIG_CGROUP_SCHED |
| 2709 | 2658 | ||
| 2710 | extern struct task_group root_task_group; | 2659 | extern struct task_group root_task_group; |
diff --git a/include/linux/sched/rt.h b/include/linux/sched/rt.h new file mode 100644 index 000000000000..94e19ea28fc3 --- /dev/null +++ b/include/linux/sched/rt.h | |||
| @@ -0,0 +1,58 @@ | |||
| 1 | #ifndef _SCHED_RT_H | ||
| 2 | #define _SCHED_RT_H | ||
| 3 | |||
| 4 | /* | ||
| 5 | * Priority of a process goes from 0..MAX_PRIO-1, valid RT | ||
| 6 | * priority is 0..MAX_RT_PRIO-1, and SCHED_NORMAL/SCHED_BATCH | ||
| 7 | * tasks are in the range MAX_RT_PRIO..MAX_PRIO-1. Priority | ||
| 8 | * values are inverted: lower p->prio value means higher priority. | ||
| 9 | * | ||
| 10 | * The MAX_USER_RT_PRIO value allows the actual maximum | ||
| 11 | * RT priority to be separate from the value exported to | ||
| 12 | * user-space. This allows kernel threads to set their | ||
| 13 | * priority to a value higher than any user task. Note: | ||
| 14 | * MAX_RT_PRIO must not be smaller than MAX_USER_RT_PRIO. | ||
| 15 | */ | ||
| 16 | |||
| 17 | #define MAX_USER_RT_PRIO 100 | ||
| 18 | #define MAX_RT_PRIO MAX_USER_RT_PRIO | ||
| 19 | |||
| 20 | #define MAX_PRIO (MAX_RT_PRIO + 40) | ||
| 21 | #define DEFAULT_PRIO (MAX_RT_PRIO + 20) | ||
| 22 | |||
| 23 | static inline int rt_prio(int prio) | ||
| 24 | { | ||
| 25 | if (unlikely(prio < MAX_RT_PRIO)) | ||
| 26 | return 1; | ||
| 27 | return 0; | ||
| 28 | } | ||
| 29 | |||
| 30 | static inline int rt_task(struct task_struct *p) | ||
| 31 | { | ||
| 32 | return rt_prio(p->prio); | ||
| 33 | } | ||
| 34 | |||
| 35 | #ifdef CONFIG_RT_MUTEXES | ||
| 36 | extern int rt_mutex_getprio(struct task_struct *p); | ||
| 37 | extern void rt_mutex_setprio(struct task_struct *p, int prio); | ||
| 38 | extern void rt_mutex_adjust_pi(struct task_struct *p); | ||
| 39 | static inline bool tsk_is_pi_blocked(struct task_struct *tsk) | ||
| 40 | { | ||
| 41 | return tsk->pi_blocked_on != NULL; | ||
| 42 | } | ||
| 43 | #else | ||
| 44 | static inline int rt_mutex_getprio(struct task_struct *p) | ||
| 45 | { | ||
| 46 | return p->normal_prio; | ||
| 47 | } | ||
| 48 | # define rt_mutex_adjust_pi(p) do { } while (0) | ||
| 49 | static inline bool tsk_is_pi_blocked(struct task_struct *tsk) | ||
| 50 | { | ||
| 51 | return false; | ||
| 52 | } | ||
| 53 | #endif | ||
| 54 | |||
| 55 | extern void normalize_rt_tasks(void); | ||
| 56 | |||
| 57 | |||
| 58 | #endif /* _SCHED_RT_H */ | ||
