aboutsummaryrefslogtreecommitdiffstats
path: root/include/linux/sched.h
diff options
context:
space:
mode:
Diffstat (limited to 'include/linux/sched.h')
-rw-r--r--include/linux/sched.h33
1 files changed, 4 insertions, 29 deletions
diff --git a/include/linux/sched.h b/include/linux/sched.h
index af2ed4bae678..1aa1cfa63b37 100644
--- a/include/linux/sched.h
+++ b/include/linux/sched.h
@@ -894,34 +894,6 @@ struct sched_entity {
894#endif 894#endif
895}; 895};
896 896
897#ifdef CONFIG_CGROUPS
898
899#define SUBSYS(_x) _x ## _subsys_id,
900enum cgroup_subsys_id {
901#include <linux/cgroup_subsys.h>
902 CGROUP_SUBSYS_COUNT
903};
904#undef SUBSYS
905
906/* A css_set is a structure holding pointers to a set of
907 * cgroup_subsys_state objects.
908 */
909
910struct css_set {
911
912 /* Set of subsystem states, one for each subsystem. NULL for
913 * subsystems that aren't part of this hierarchy. These
914 * pointers reduce the number of dereferences required to get
915 * from a task to its state for a given cgroup, but result
916 * in increased space usage if tasks are in wildly different
917 * groupings across different hierarchies. This array is
918 * immutable after creation */
919 struct cgroup_subsys_state *subsys[CGROUP_SUBSYS_COUNT];
920
921};
922
923#endif /* CONFIG_CGROUPS */
924
925struct task_struct { 897struct task_struct {
926 volatile long state; /* -1 unrunnable, 0 runnable, >0 stopped */ 898 volatile long state; /* -1 unrunnable, 0 runnable, >0 stopped */
927 void *stack; 899 void *stack;
@@ -1159,7 +1131,10 @@ struct task_struct {
1159 int cpuset_mem_spread_rotor; 1131 int cpuset_mem_spread_rotor;
1160#endif 1132#endif
1161#ifdef CONFIG_CGROUPS 1133#ifdef CONFIG_CGROUPS
1162 struct css_set cgroups; 1134 /* Control Group info protected by css_set_lock */
1135 struct css_set *cgroups;
1136 /* cg_list protected by css_set_lock and tsk->alloc_lock */
1137 struct list_head cg_list;
1163#endif 1138#endif
1164#ifdef CONFIG_FUTEX 1139#ifdef CONFIG_FUTEX
1165 struct robust_list_head __user *robust_list; 1140 struct robust_list_head __user *robust_list;