diff options
| author | Dhaval Giani <dhaval.giani@gmail.com> | 2010-01-20 07:26:18 -0500 |
|---|---|---|
| committer | Ingo Molnar <mingo@elte.hu> | 2010-01-21 07:40:18 -0500 |
| commit | 7c9414385ebfdd87cc542d4e7e3bb0dbb2d3ce25 (patch) | |
| tree | beb51d7d0d543d72e2754cff807df5c399f2d376 | |
| parent | 871e35bc9733f273eaf5ceb69bbd0423b58e5285 (diff) | |
sched: Remove USER_SCHED
Remove the USER_SCHED feature. It has been scheduled to be removed in
2.6.34 as per http://marc.info/?l=linux-kernel&m=125728479022976&w=2
Signed-off-by: Dhaval Giani <dhaval.giani@gmail.com>
Signed-off-by: Peter Zijlstra <a.p.zijlstra@chello.nl>
LKML-Reference: <1263990378.24844.3.camel@localhost>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
| -rw-r--r-- | Documentation/feature-removal-schedule.txt | 15 | ||||
| -rw-r--r-- | include/linux/sched.h | 14 | ||||
| -rw-r--r-- | init/Kconfig | 81 | ||||
| -rw-r--r-- | kernel/ksysfs.c | 8 | ||||
| -rw-r--r-- | kernel/sched.c | 114 | ||||
| -rw-r--r-- | kernel/sys.c | 5 | ||||
| -rw-r--r-- | kernel/user.c | 305 |
7 files changed, 38 insertions, 504 deletions
diff --git a/Documentation/feature-removal-schedule.txt b/Documentation/feature-removal-schedule.txt index 870d190fe617..04a3fc3d139b 100644 --- a/Documentation/feature-removal-schedule.txt +++ b/Documentation/feature-removal-schedule.txt | |||
| @@ -6,21 +6,6 @@ be removed from this file. | |||
| 6 | 6 | ||
| 7 | --------------------------- | 7 | --------------------------- |
| 8 | 8 | ||
| 9 | What: USER_SCHED | ||
| 10 | When: 2.6.34 | ||
| 11 | |||
| 12 | Why: USER_SCHED was implemented as a proof of concept for group scheduling. | ||
| 13 | The effect of USER_SCHED can already be achieved from userspace with | ||
| 14 | the help of libcgroup. The removal of USER_SCHED will also simplify | ||
| 15 | the scheduler code with the removal of one major ifdef. There are also | ||
| 16 | issues USER_SCHED has with USER_NS. A decision was taken not to fix | ||
| 17 | those and instead remove USER_SCHED. Also new group scheduling | ||
| 18 | features will not be implemented for USER_SCHED. | ||
| 19 | |||
| 20 | Who: Dhaval Giani <dhaval@linux.vnet.ibm.com> | ||
| 21 | |||
| 22 | --------------------------- | ||
| 23 | |||
| 24 | What: PRISM54 | 9 | What: PRISM54 |
| 25 | When: 2.6.34 | 10 | When: 2.6.34 |
| 26 | 11 | ||
diff --git a/include/linux/sched.h b/include/linux/sched.h index 50d685cde70e..8b079735ae5f 100644 --- a/include/linux/sched.h +++ b/include/linux/sched.h | |||
| @@ -731,14 +731,6 @@ struct user_struct { | |||
| 731 | uid_t uid; | 731 | uid_t uid; |
| 732 | struct user_namespace *user_ns; | 732 | struct user_namespace *user_ns; |
| 733 | 733 | ||
| 734 | #ifdef CONFIG_USER_SCHED | ||
| 735 | struct task_group *tg; | ||
| 736 | #ifdef CONFIG_SYSFS | ||
| 737 | struct kobject kobj; | ||
| 738 | struct delayed_work work; | ||
| 739 | #endif | ||
| 740 | #endif | ||
| 741 | |||
| 742 | #ifdef CONFIG_PERF_EVENTS | 734 | #ifdef CONFIG_PERF_EVENTS |
| 743 | atomic_long_t locked_vm; | 735 | atomic_long_t locked_vm; |
| 744 | #endif | 736 | #endif |
| @@ -2502,13 +2494,9 @@ extern long sched_getaffinity(pid_t pid, struct cpumask *mask); | |||
| 2502 | 2494 | ||
| 2503 | extern void normalize_rt_tasks(void); | 2495 | extern void normalize_rt_tasks(void); |
| 2504 | 2496 | ||
| 2505 | #ifdef CONFIG_GROUP_SCHED | 2497 | #ifdef CONFIG_CGROUP_SCHED |
| 2506 | 2498 | ||
| 2507 | extern struct task_group init_task_group; | 2499 | extern struct task_group init_task_group; |
| 2508 | #ifdef CONFIG_USER_SCHED | ||
| 2509 | extern struct task_group root_task_group; | ||
| 2510 | extern void set_tg_uid(struct user_struct *user); | ||
| 2511 | #endif | ||
| 2512 | 2500 | ||
| 2513 | extern struct task_group *sched_create_group(struct task_group *parent); | 2501 | extern struct task_group *sched_create_group(struct task_group *parent); |
| 2514 | extern void sched_destroy_group(struct task_group *tg); | 2502 | extern void sched_destroy_group(struct task_group *tg); |
diff --git a/init/Kconfig b/init/Kconfig index a23da9f01803..e9fa3007a6fc 100644 --- a/init/Kconfig +++ b/init/Kconfig | |||
| @@ -435,57 +435,6 @@ config LOG_BUF_SHIFT | |||
| 435 | config HAVE_UNSTABLE_SCHED_CLOCK | 435 | config HAVE_UNSTABLE_SCHED_CLOCK |
| 436 | bool | 436 | bool |
| 437 | 437 | ||
| 438 | config GROUP_SCHED | ||
| 439 | bool "Group CPU scheduler" | ||
| 440 | depends on EXPERIMENTAL | ||
| 441 | default n | ||
| 442 | help | ||
| 443 | This feature lets CPU scheduler recognize task groups and control CPU | ||
| 444 | bandwidth allocation to such task groups. | ||
| 445 | In order to create a group from arbitrary set of processes, use | ||
| 446 | CONFIG_CGROUPS. (See Control Group support.) | ||
| 447 | |||
| 448 | config FAIR_GROUP_SCHED | ||
| 449 | bool "Group scheduling for SCHED_OTHER" | ||
| 450 | depends on GROUP_SCHED | ||
| 451 | default GROUP_SCHED | ||
| 452 | |||
| 453 | config RT_GROUP_SCHED | ||
| 454 | bool "Group scheduling for SCHED_RR/FIFO" | ||
| 455 | depends on EXPERIMENTAL | ||
| 456 | depends on GROUP_SCHED | ||
| 457 | default n | ||
| 458 | help | ||
| 459 | This feature lets you explicitly allocate real CPU bandwidth | ||
| 460 | to users or control groups (depending on the "Basis for grouping tasks" | ||
| 461 | setting below. If enabled, it will also make it impossible to | ||
| 462 | schedule realtime tasks for non-root users until you allocate | ||
| 463 | realtime bandwidth for them. | ||
| 464 | See Documentation/scheduler/sched-rt-group.txt for more information. | ||
| 465 | |||
| 466 | choice | ||
| 467 | depends on GROUP_SCHED | ||
| 468 | prompt "Basis for grouping tasks" | ||
| 469 | default USER_SCHED | ||
| 470 | |||
| 471 | config USER_SCHED | ||
| 472 | bool "user id" | ||
| 473 | help | ||
| 474 | This option will choose userid as the basis for grouping | ||
| 475 | tasks, thus providing equal CPU bandwidth to each user. | ||
| 476 | |||
| 477 | config CGROUP_SCHED | ||
| 478 | bool "Control groups" | ||
| 479 | depends on CGROUPS | ||
| 480 | help | ||
| 481 | This option allows you to create arbitrary task groups | ||
| 482 | using the "cgroup" pseudo filesystem and control | ||
| 483 | the cpu bandwidth allocated to each such task group. | ||
| 484 | Refer to Documentation/cgroups/cgroups.txt for more | ||
| 485 | information on "cgroup" pseudo filesystem. | ||
| 486 | |||
| 487 | endchoice | ||
| 488 | |||
| 489 | menuconfig CGROUPS | 438 | menuconfig CGROUPS |
| 490 | boolean "Control Group support" | 439 | boolean "Control Group support" |
| 491 | help | 440 | help |
| @@ -606,6 +555,36 @@ config CGROUP_MEM_RES_CTLR_SWAP | |||
| 606 | Now, memory usage of swap_cgroup is 2 bytes per entry. If swap page | 555 | Now, memory usage of swap_cgroup is 2 bytes per entry. If swap page |
| 607 | size is 4096bytes, 512k per 1Gbytes of swap. | 556 | size is 4096bytes, 512k per 1Gbytes of swap. |
| 608 | 557 | ||
| 558 | menuconfig CGROUP_SCHED | ||
| 559 | bool "Group CPU scheduler" | ||
| 560 | depends on EXPERIMENTAL && CGROUPS | ||
| 561 | default n | ||
| 562 | help | ||
| 563 | This feature lets CPU scheduler recognize task groups and control CPU | ||
| 564 | bandwidth allocation to such task groups. It uses cgroups to group | ||
| 565 | tasks. | ||
| 566 | |||
| 567 | if CGROUP_SCHED | ||
| 568 | config FAIR_GROUP_SCHED | ||
| 569 | bool "Group scheduling for SCHED_OTHER" | ||
| 570 | depends on CGROUP_SCHED | ||
| 571 | default CGROUP_SCHED | ||
| 572 | |||
| 573 | config RT_GROUP_SCHED | ||
| 574 | bool "Group scheduling for SCHED_RR/FIFO" | ||
| 575 | depends on EXPERIMENTAL | ||
| 576 | depends on CGROUP_SCHED | ||
| 577 | default n | ||
| 578 | help | ||
| 579 | This feature lets you explicitly allocate real CPU bandwidth | ||
| 580 | to users or control groups (depending on the "Basis for grouping tasks" | ||
| 581 | setting below. If enabled, it will also make it impossible to | ||
| 582 | schedule realtime tasks for non-root users until you allocate | ||
| 583 | realtime bandwidth for them. | ||
| 584 | See Documentation/scheduler/sched-rt-group.txt for more information. | ||
| 585 | |||
| 586 | endif #CGROUP_SCHED | ||
| 587 | |||
| 609 | endif # CGROUPS | 588 | endif # CGROUPS |
| 610 | 589 | ||
| 611 | config MM_OWNER | 590 | config MM_OWNER |
diff --git a/kernel/ksysfs.c b/kernel/ksysfs.c index 3feaf5a74514..6b1ccc3f0205 100644 --- a/kernel/ksysfs.c +++ b/kernel/ksysfs.c | |||
| @@ -197,16 +197,8 @@ static int __init ksysfs_init(void) | |||
| 197 | goto group_exit; | 197 | goto group_exit; |
| 198 | } | 198 | } |
| 199 | 199 | ||
| 200 | /* create the /sys/kernel/uids/ directory */ | ||
| 201 | error = uids_sysfs_init(); | ||
| 202 | if (error) | ||
| 203 | goto notes_exit; | ||
| 204 | |||
| 205 | return 0; | 200 | return 0; |
| 206 | 201 | ||
| 207 | notes_exit: | ||
| 208 | if (notes_size > 0) | ||
| 209 | sysfs_remove_bin_file(kernel_kobj, ¬es_attr); | ||
| 210 | group_exit: | 202 | group_exit: |
| 211 | sysfs_remove_group(kernel_kobj, &kernel_attr_group); | 203 | sysfs_remove_group(kernel_kobj, &kernel_attr_group); |
| 212 | kset_exit: | 204 | kset_exit: |
diff --git a/kernel/sched.c b/kernel/sched.c index c0be07932a8d..41e76d325648 100644 --- a/kernel/sched.c +++ b/kernel/sched.c | |||
| @@ -233,7 +233,7 @@ static void destroy_rt_bandwidth(struct rt_bandwidth *rt_b) | |||
| 233 | */ | 233 | */ |
| 234 | static DEFINE_MUTEX(sched_domains_mutex); | 234 | static DEFINE_MUTEX(sched_domains_mutex); |
| 235 | 235 | ||
| 236 | #ifdef CONFIG_GROUP_SCHED | 236 | #ifdef CONFIG_CGROUP_SCHED |
| 237 | 237 | ||
| 238 | #include <linux/cgroup.h> | 238 | #include <linux/cgroup.h> |
| 239 | 239 | ||
| @@ -243,13 +243,7 @@ static LIST_HEAD(task_groups); | |||
| 243 | 243 | ||
| 244 | /* task group related information */ | 244 | /* task group related information */ |
| 245 | struct task_group { | 245 | struct task_group { |
| 246 | #ifdef CONFIG_CGROUP_SCHED | ||
| 247 | struct cgroup_subsys_state css; | 246 | struct cgroup_subsys_state css; |
| 248 | #endif | ||
| 249 | |||
| 250 | #ifdef CONFIG_USER_SCHED | ||
| 251 | uid_t uid; | ||
| 252 | #endif | ||
| 253 | 247 | ||
| 254 | #ifdef CONFIG_FAIR_GROUP_SCHED | 248 | #ifdef CONFIG_FAIR_GROUP_SCHED |
| 255 | /* schedulable entities of this group on each cpu */ | 249 | /* schedulable entities of this group on each cpu */ |
| @@ -274,35 +268,7 @@ struct task_group { | |||
| 274 | struct list_head children; | 268 | struct list_head children; |
| 275 | }; | 269 | }; |
| 276 | 270 | ||
| 277 | #ifdef CONFIG_USER_SCHED | ||
| 278 | |||
| 279 | /* Helper function to pass uid information to create_sched_user() */ | ||
| 280 | void set_tg_uid(struct user_struct *user) | ||
| 281 | { | ||
| 282 | user->tg->uid = user->uid; | ||
| 283 | } | ||
| 284 | |||
| 285 | /* | ||
| 286 | * Root task group. | ||
| 287 | * Every UID task group (including init_task_group aka UID-0) will | ||
| 288 | * be a child to this group. | ||
| 289 | */ | ||
| 290 | struct task_group root_task_group; | ||
| 291 | |||
| 292 | #ifdef CONFIG_FAIR_GROUP_SCHED | ||
| 293 | /* Default task group's sched entity on each cpu */ | ||
| 294 | static DEFINE_PER_CPU(struct sched_entity, init_sched_entity); | ||
| 295 | /* Default task group's cfs_rq on each cpu */ | ||
| 296 | static DEFINE_PER_CPU_SHARED_ALIGNED(struct cfs_rq, init_tg_cfs_rq); | ||
| 297 | #endif /* CONFIG_FAIR_GROUP_SCHED */ | ||
| 298 | |||
| 299 | #ifdef CONFIG_RT_GROUP_SCHED | ||
| 300 | static DEFINE_PER_CPU(struct sched_rt_entity, init_sched_rt_entity); | ||
| 301 | static DEFINE_PER_CPU_SHARED_ALIGNED(struct rt_rq, init_rt_rq_var); | ||
| 302 | #endif /* CONFIG_RT_GROUP_SCHED */ | ||
| 303 | #else /* !CONFIG_USER_SCHED */ | ||
| 304 | #define root_task_group init_task_group | 271 | #define root_task_group init_task_group |
| 305 | #endif /* CONFIG_USER_SCHED */ | ||
| 306 | 272 | ||
| 307 | /* task_group_lock serializes add/remove of task groups and also changes to | 273 | /* task_group_lock serializes add/remove of task groups and also changes to |
| 308 | * a task group's cpu shares. | 274 | * a task group's cpu shares. |
| @@ -318,11 +284,7 @@ static int root_task_group_empty(void) | |||
| 318 | } | 284 | } |
| 319 | #endif | 285 | #endif |
| 320 | 286 | ||
| 321 | #ifdef CONFIG_USER_SCHED | ||
| 322 | # define INIT_TASK_GROUP_LOAD (2*NICE_0_LOAD) | ||
| 323 | #else /* !CONFIG_USER_SCHED */ | ||
| 324 | # define INIT_TASK_GROUP_LOAD NICE_0_LOAD | 287 | # define INIT_TASK_GROUP_LOAD NICE_0_LOAD |
| 325 | #endif /* CONFIG_USER_SCHED */ | ||
| 326 | 288 | ||
| 327 | /* | 289 | /* |
| 328 | * A weight of 0 or 1 can cause arithmetics problems. | 290 | * A weight of 0 or 1 can cause arithmetics problems. |
| @@ -348,11 +310,7 @@ static inline struct task_group *task_group(struct task_struct *p) | |||
| 348 | { | 310 | { |
| 349 | struct task_group *tg; | 311 | struct task_group *tg; |
| 350 | 312 | ||
| 351 | #ifdef CONFIG_USER_SCHED | 313 | #ifdef CONFIG_CGROUP_SCHED |
| 352 | rcu_read_lock(); | ||
| 353 | tg = __task_cred(p)->user->tg; | ||
| 354 | rcu_read_unlock(); | ||
| 355 | #elif defined(CONFIG_CGROUP_SCHED) | ||
| 356 | tg = container_of(task_subsys_state(p, cpu_cgroup_subsys_id), | 314 | tg = container_of(task_subsys_state(p, cpu_cgroup_subsys_id), |
| 357 | struct task_group, css); | 315 | struct task_group, css); |
| 358 | #else | 316 | #else |
| @@ -383,7 +341,7 @@ static inline struct task_group *task_group(struct task_struct *p) | |||
| 383 | return NULL; | 341 | return NULL; |
| 384 | } | 342 | } |
| 385 | 343 | ||
| 386 | #endif /* CONFIG_GROUP_SCHED */ | 344 | #endif /* CONFIG_CGROUP_SCHED */ |
| 387 | 345 | ||
| 388 | /* CFS-related fields in a runqueue */ | 346 | /* CFS-related fields in a runqueue */ |
| 389 | struct cfs_rq { | 347 | struct cfs_rq { |
| @@ -7678,9 +7636,6 @@ void __init sched_init(void) | |||
| 7678 | #ifdef CONFIG_RT_GROUP_SCHED | 7636 | #ifdef CONFIG_RT_GROUP_SCHED |
| 7679 | alloc_size += 2 * nr_cpu_ids * sizeof(void **); | 7637 | alloc_size += 2 * nr_cpu_ids * sizeof(void **); |
| 7680 | #endif | 7638 | #endif |
| 7681 | #ifdef CONFIG_USER_SCHED | ||
| 7682 | alloc_size *= 2; | ||
| 7683 | #endif | ||
| 7684 | #ifdef CONFIG_CPUMASK_OFFSTACK | 7639 | #ifdef CONFIG_CPUMASK_OFFSTACK |
| 7685 | alloc_size += num_possible_cpus() * cpumask_size(); | 7640 | alloc_size += num_possible_cpus() * cpumask_size(); |
| 7686 | #endif | 7641 | #endif |
| @@ -7694,13 +7649,6 @@ void __init sched_init(void) | |||
| 7694 | init_task_group.cfs_rq = (struct cfs_rq **)ptr; | 7649 | init_task_group.cfs_rq = (struct cfs_rq **)ptr; |
| 7695 | ptr += nr_cpu_ids * sizeof(void **); | 7650 | ptr += nr_cpu_ids * sizeof(void **); |
| 7696 | 7651 | ||
| 7697 | #ifdef CONFIG_USER_SCHED | ||
| 7698 | root_task_group.se = (struct sched_entity **)ptr; | ||
| 7699 | ptr += nr_cpu_ids * sizeof(void **); | ||
| 7700 | |||
| 7701 | root_task_group.cfs_rq = (struct cfs_rq **)ptr; | ||
| 7702 | ptr += nr_cpu_ids * sizeof(void **); | ||
| 7703 | #endif /* CONFIG_USER_SCHED */ | ||
| 7704 | #endif /* CONFIG_FAIR_GROUP_SCHED */ | 7652 | #endif /* CONFIG_FAIR_GROUP_SCHED */ |
| 7705 | #ifdef CONFIG_RT_GROUP_SCHED | 7653 | #ifdef CONFIG_RT_GROUP_SCHED |
| 7706 | init_task_group.rt_se = (struct sched_rt_entity **)ptr; | 7654 | init_task_group.rt_se = (struct sched_rt_entity **)ptr; |
| @@ -7709,13 +7657,6 @@ void __init sched_init(void) | |||
| 7709 | init_task_group.rt_rq = (struct rt_rq **)ptr; | 7657 | init_task_group.rt_rq = (struct rt_rq **)ptr; |
| 7710 | ptr += nr_cpu_ids * sizeof(void **); | 7658 | ptr += nr_cpu_ids * sizeof(void **); |
| 7711 | 7659 | ||
| 7712 | #ifdef CONFIG_USER_SCHED | ||
| 7713 | root_task_group.rt_se = (struct sched_rt_entity **)ptr; | ||
| 7714 | ptr += nr_cpu_ids * sizeof(void **); | ||
| 7715 | |||
| 7716 | root_task_group.rt_rq = (struct rt_rq **)ptr; | ||
| 7717 | ptr += nr_cpu_ids * sizeof(void **); | ||
| 7718 | #endif /* CONFIG_USER_SCHED */ | ||
| 7719 | #endif /* CONFIG_RT_GROUP_SCHED */ | 7660 | #endif /* CONFIG_RT_GROUP_SCHED */ |
| 7720 | #ifdef CONFIG_CPUMASK_OFFSTACK | 7661 | #ifdef CONFIG_CPUMASK_OFFSTACK |
| 7721 | for_each_possible_cpu(i) { | 7662 | for_each_possible_cpu(i) { |
| @@ -7735,22 +7676,13 @@ void __init sched_init(void) | |||
| 7735 | #ifdef CONFIG_RT_GROUP_SCHED | 7676 | #ifdef CONFIG_RT_GROUP_SCHED |
| 7736 | init_rt_bandwidth(&init_task_group.rt_bandwidth, | 7677 | init_rt_bandwidth(&init_task_group.rt_bandwidth, |
| 7737 | global_rt_period(), global_rt_runtime()); | 7678 | global_rt_period(), global_rt_runtime()); |
| 7738 | #ifdef CONFIG_USER_SCHED | ||
| 7739 | init_rt_bandwidth(&root_task_group.rt_bandwidth, | ||
| 7740 | global_rt_period(), RUNTIME_INF); | ||
| 7741 | #endif /* CONFIG_USER_SCHED */ | ||
| 7742 | #endif /* CONFIG_RT_GROUP_SCHED */ | 7679 | #endif /* CONFIG_RT_GROUP_SCHED */ |
| 7743 | 7680 | ||
| 7744 | #ifdef CONFIG_GROUP_SCHED | 7681 | #ifdef CONFIG_CGROUP_SCHED |
| 7745 | list_add(&init_task_group.list, &task_groups); | 7682 | list_add(&init_task_group.list, &task_groups); |
| 7746 | INIT_LIST_HEAD(&init_task_group.children); | 7683 | INIT_LIST_HEAD(&init_task_group.children); |
| 7747 | 7684 | ||
| 7748 | #ifdef CONFIG_USER_SCHED | 7685 | #endif /* CONFIG_CGROUP_SCHED */ |
| 7749 | INIT_LIST_HEAD(&root_task_group.children); | ||
| 7750 | init_task_group.parent = &root_task_group; | ||
| 7751 | list_add(&init_task_group.siblings, &root_task_group.children); | ||
| 7752 | #endif /* CONFIG_USER_SCHED */ | ||
| 7753 | #endif /* CONFIG_GROUP_SCHED */ | ||
| 7754 | 7686 | ||
| 7755 | #if defined CONFIG_FAIR_GROUP_SCHED && defined CONFIG_SMP | 7687 | #if defined CONFIG_FAIR_GROUP_SCHED && defined CONFIG_SMP |
| 7756 | update_shares_data = __alloc_percpu(nr_cpu_ids * sizeof(unsigned long), | 7688 | update_shares_data = __alloc_percpu(nr_cpu_ids * sizeof(unsigned long), |
| @@ -7790,25 +7722,6 @@ void __init sched_init(void) | |||
| 7790 | * directly in rq->cfs (i.e init_task_group->se[] = NULL). | 7722 | * directly in rq->cfs (i.e init_task_group->se[] = NULL). |
| 7791 | */ | 7723 | */ |
| 7792 | init_tg_cfs_entry(&init_task_group, &rq->cfs, NULL, i, 1, NULL); | 7724 | init_tg_cfs_entry(&init_task_group, &rq->cfs, NULL, i, 1, NULL); |
| 7793 | #elif defined CONFIG_USER_SCHED | ||
| 7794 | root_task_group.shares = NICE_0_LOAD; | ||
| 7795 | init_tg_cfs_entry(&root_task_group, &rq->cfs, NULL, i, 0, NULL); | ||
| 7796 | /* | ||
| 7797 | * In case of task-groups formed thr' the user id of tasks, | ||
| 7798 | * init_task_group represents tasks belonging to root user. | ||
| 7799 | * Hence it forms a sibling of all subsequent groups formed. | ||
| 7800 | * In this case, init_task_group gets only a fraction of overall | ||
| 7801 | * system cpu resource, based on the weight assigned to root | ||
| 7802 | * user's cpu share (INIT_TASK_GROUP_LOAD). This is accomplished | ||
| 7803 | * by letting tasks of init_task_group sit in a separate cfs_rq | ||
| 7804 | * (init_tg_cfs_rq) and having one entity represent this group of | ||
| 7805 | * tasks in rq->cfs (i.e init_task_group->se[] != NULL). | ||
| 7806 | */ | ||
| 7807 | init_tg_cfs_entry(&init_task_group, | ||
| 7808 | &per_cpu(init_tg_cfs_rq, i), | ||
| 7809 | &per_cpu(init_sched_entity, i), i, 1, | ||
| 7810 | root_task_group.se[i]); | ||
| 7811 | |||
| 7812 | #endif | 7725 | #endif |
| 7813 | #endif /* CONFIG_FAIR_GROUP_SCHED */ | 7726 | #endif /* CONFIG_FAIR_GROUP_SCHED */ |
| 7814 | 7727 | ||
| @@ -7817,12 +7730,6 @@ void __init sched_init(void) | |||
| 7817 | INIT_LIST_HEAD(&rq->leaf_rt_rq_list); | 7730 | INIT_LIST_HEAD(&rq->leaf_rt_rq_list); |
| 7818 | #ifdef CONFIG_CGROUP_SCHED | 7731 | #ifdef CONFIG_CGROUP_SCHED |
| 7819 | init_tg_rt_entry(&init_task_group, &rq->rt, NULL, i, 1, NULL); | 7732 | init_tg_rt_entry(&init_task_group, &rq->rt, NULL, i, 1, NULL); |
| 7820 | #elif defined CONFIG_USER_SCHED | ||
| 7821 | init_tg_rt_entry(&root_task_group, &rq->rt, NULL, i, 0, NULL); | ||
| 7822 | init_tg_rt_entry(&init_task_group, | ||
| 7823 | &per_cpu(init_rt_rq_var, i), | ||
| 7824 | &per_cpu(init_sched_rt_entity, i), i, 1, | ||
| 7825 | root_task_group.rt_se[i]); | ||
| 7826 | #endif | 7733 | #endif |
| 7827 | #endif | 7734 | #endif |
| 7828 | 7735 | ||
| @@ -8218,7 +8125,7 @@ static inline void unregister_rt_sched_group(struct task_group *tg, int cpu) | |||
| 8218 | } | 8125 | } |
| 8219 | #endif /* CONFIG_RT_GROUP_SCHED */ | 8126 | #endif /* CONFIG_RT_GROUP_SCHED */ |
| 8220 | 8127 | ||
| 8221 | #ifdef CONFIG_GROUP_SCHED | 8128 | #ifdef CONFIG_CGROUP_SCHED |
| 8222 | static void free_sched_group(struct task_group *tg) | 8129 | static void free_sched_group(struct task_group *tg) |
| 8223 | { | 8130 | { |
| 8224 | free_fair_sched_group(tg); | 8131 | free_fair_sched_group(tg); |
| @@ -8327,7 +8234,7 @@ void sched_move_task(struct task_struct *tsk) | |||
| 8327 | 8234 | ||
| 8328 | task_rq_unlock(rq, &flags); | 8235 | task_rq_unlock(rq, &flags); |
| 8329 | } | 8236 | } |
| 8330 | #endif /* CONFIG_GROUP_SCHED */ | 8237 | #endif /* CONFIG_CGROUP_SCHED */ |
| 8331 | 8238 | ||
| 8332 | #ifdef CONFIG_FAIR_GROUP_SCHED | 8239 | #ifdef CONFIG_FAIR_GROUP_SCHED |
| 8333 | static void __set_se_shares(struct sched_entity *se, unsigned long shares) | 8240 | static void __set_se_shares(struct sched_entity *se, unsigned long shares) |
| @@ -8469,13 +8376,6 @@ static int tg_schedulable(struct task_group *tg, void *data) | |||
| 8469 | runtime = d->rt_runtime; | 8376 | runtime = d->rt_runtime; |
| 8470 | } | 8377 | } |
| 8471 | 8378 | ||
| 8472 | #ifdef CONFIG_USER_SCHED | ||
| 8473 | if (tg == &root_task_group) { | ||
| 8474 | period = global_rt_period(); | ||
| 8475 | runtime = global_rt_runtime(); | ||
| 8476 | } | ||
| 8477 | #endif | ||
| 8478 | |||
| 8479 | /* | 8379 | /* |
| 8480 | * Cannot have more runtime than the period. | 8380 | * Cannot have more runtime than the period. |
| 8481 | */ | 8381 | */ |
diff --git a/kernel/sys.c b/kernel/sys.c index 26a6b73a6b85..f75bf0936f47 100644 --- a/kernel/sys.c +++ b/kernel/sys.c | |||
| @@ -569,11 +569,6 @@ static int set_user(struct cred *new) | |||
| 569 | if (!new_user) | 569 | if (!new_user) |
| 570 | return -EAGAIN; | 570 | return -EAGAIN; |
| 571 | 571 | ||
| 572 | if (!task_can_switch_user(new_user, current)) { | ||
| 573 | free_uid(new_user); | ||
| 574 | return -EINVAL; | ||
| 575 | } | ||
| 576 | |||
| 577 | if (atomic_read(&new_user->processes) >= | 572 | if (atomic_read(&new_user->processes) >= |
| 578 | current->signal->rlim[RLIMIT_NPROC].rlim_cur && | 573 | current->signal->rlim[RLIMIT_NPROC].rlim_cur && |
| 579 | new_user != INIT_USER) { | 574 | new_user != INIT_USER) { |
diff --git a/kernel/user.c b/kernel/user.c index 46d0165ca70c..766467b3bcb7 100644 --- a/kernel/user.c +++ b/kernel/user.c | |||
| @@ -56,9 +56,6 @@ struct user_struct root_user = { | |||
| 56 | .sigpending = ATOMIC_INIT(0), | 56 | .sigpending = ATOMIC_INIT(0), |
| 57 | .locked_shm = 0, | 57 | .locked_shm = 0, |
| 58 | .user_ns = &init_user_ns, | 58 | .user_ns = &init_user_ns, |
| 59 | #ifdef CONFIG_USER_SCHED | ||
| 60 | .tg = &init_task_group, | ||
| 61 | #endif | ||
| 62 | }; | 59 | }; |
| 63 | 60 | ||
| 64 | /* | 61 | /* |
| @@ -75,268 +72,6 @@ static void uid_hash_remove(struct user_struct *up) | |||
| 75 | put_user_ns(up->user_ns); | 72 | put_user_ns(up->user_ns); |
| 76 | } | 73 | } |
| 77 | 74 | ||
| 78 | #ifdef CONFIG_USER_SCHED | ||
| 79 | |||
| 80 | static void sched_destroy_user(struct user_struct *up) | ||
| 81 | { | ||
| 82 | sched_destroy_group(up->tg); | ||
| 83 | } | ||
| 84 | |||
| 85 | static int sched_create_user(struct user_struct *up) | ||
| 86 | { | ||
| 87 | int rc = 0; | ||
| 88 | |||
| 89 | up->tg = sched_create_group(&root_task_group); | ||
| 90 | if (IS_ERR(up->tg)) | ||
| 91 | rc = -ENOMEM; | ||
| 92 | |||
| 93 | set_tg_uid(up); | ||
| 94 | |||
| 95 | return rc; | ||
| 96 | } | ||
| 97 | |||
| 98 | #else /* CONFIG_USER_SCHED */ | ||
| 99 | |||
| 100 | static void sched_destroy_user(struct user_struct *up) { } | ||
| 101 | static int sched_create_user(struct user_struct *up) { return 0; } | ||
| 102 | |||
| 103 | #endif /* CONFIG_USER_SCHED */ | ||
| 104 | |||
| 105 | #if defined(CONFIG_USER_SCHED) && defined(CONFIG_SYSFS) | ||
| 106 | |||
| 107 | static struct user_struct *uid_hash_find(uid_t uid, struct hlist_head *hashent) | ||
| 108 | { | ||
| 109 | struct user_struct *user; | ||
| 110 | struct hlist_node *h; | ||
| 111 | |||
| 112 | hlist_for_each_entry(user, h, hashent, uidhash_node) { | ||
| 113 | if (user->uid == uid) { | ||
| 114 | /* possibly resurrect an "almost deleted" object */ | ||
| 115 | if (atomic_inc_return(&user->__count) == 1) | ||
| 116 | cancel_delayed_work(&user->work); | ||
| 117 | return user; | ||
| 118 | } | ||
| 119 | } | ||
| 120 | |||
| 121 | return NULL; | ||
| 122 | } | ||
| 123 | |||
| 124 | static struct kset *uids_kset; /* represents the /sys/kernel/uids/ directory */ | ||
| 125 | static DEFINE_MUTEX(uids_mutex); | ||
| 126 | |||
| 127 | static inline void uids_mutex_lock(void) | ||
| 128 | { | ||
| 129 | mutex_lock(&uids_mutex); | ||
| 130 | } | ||
| 131 | |||
| 132 | static inline void uids_mutex_unlock(void) | ||
| 133 | { | ||
| 134 | mutex_unlock(&uids_mutex); | ||
| 135 | } | ||
| 136 | |||
| 137 | /* uid directory attributes */ | ||
| 138 | #ifdef CONFIG_FAIR_GROUP_SCHED | ||
| 139 | static ssize_t cpu_shares_show(struct kobject *kobj, | ||
| 140 | struct kobj_attribute *attr, | ||
| 141 | char *buf) | ||
| 142 | { | ||
| 143 | struct user_struct *up = container_of(kobj, struct user_struct, kobj); | ||
| 144 | |||
| 145 | return sprintf(buf, "%lu\n", sched_group_shares(up->tg)); | ||
| 146 | } | ||
| 147 | |||
| 148 | static ssize_t cpu_shares_store(struct kobject *kobj, | ||
| 149 | struct kobj_attribute *attr, | ||
| 150 | const char *buf, size_t size) | ||
| 151 | { | ||
| 152 | struct user_struct *up = container_of(kobj, struct user_struct, kobj); | ||
| 153 | unsigned long shares; | ||
| 154 | int rc; | ||
| 155 | |||
| 156 | sscanf(buf, "%lu", &shares); | ||
| 157 | |||
| 158 | rc = sched_group_set_shares(up->tg, shares); | ||
| 159 | |||
| 160 | return (rc ? rc : size); | ||
| 161 | } | ||
| 162 | |||
| 163 | static struct kobj_attribute cpu_share_attr = | ||
| 164 | __ATTR(cpu_share, 0644, cpu_shares_show, cpu_shares_store); | ||
| 165 | #endif | ||
| 166 | |||
| 167 | #ifdef CONFIG_RT_GROUP_SCHED | ||
| 168 | static ssize_t cpu_rt_runtime_show(struct kobject *kobj, | ||
| 169 | struct kobj_attribute *attr, | ||
| 170 | char *buf) | ||
| 171 | { | ||
| 172 | struct user_struct *up = container_of(kobj, struct user_struct, kobj); | ||
| 173 | |||
| 174 | return sprintf(buf, "%ld\n", sched_group_rt_runtime(up->tg)); | ||
| 175 | } | ||
| 176 | |||
| 177 | static ssize_t cpu_rt_runtime_store(struct kobject *kobj, | ||
| 178 | struct kobj_attribute *attr, | ||
| 179 | const char *buf, size_t size) | ||
| 180 | { | ||
| 181 | struct user_struct *up = container_of(kobj, struct user_struct, kobj); | ||
| 182 | unsigned long rt_runtime; | ||
| 183 | int rc; | ||
| 184 | |||
| 185 | sscanf(buf, "%ld", &rt_runtime); | ||
| 186 | |||
| 187 | rc = sched_group_set_rt_runtime(up->tg, rt_runtime); | ||
| 188 | |||
| 189 | return (rc ? rc : size); | ||
| 190 | } | ||
| 191 | |||
| 192 | static struct kobj_attribute cpu_rt_runtime_attr = | ||
| 193 | __ATTR(cpu_rt_runtime, 0644, cpu_rt_runtime_show, cpu_rt_runtime_store); | ||
| 194 | |||
| 195 | static ssize_t cpu_rt_period_show(struct kobject *kobj, | ||
| 196 | struct kobj_attribute *attr, | ||
| 197 | char *buf) | ||
| 198 | { | ||
| 199 | struct user_struct *up = container_of(kobj, struct user_struct, kobj); | ||
| 200 | |||
| 201 | return sprintf(buf, "%lu\n", sched_group_rt_period(up->tg)); | ||
| 202 | } | ||
| 203 | |||
| 204 | static ssize_t cpu_rt_period_store(struct kobject *kobj, | ||
| 205 | struct kobj_attribute *attr, | ||
| 206 | const char *buf, size_t size) | ||
| 207 | { | ||
| 208 | struct user_struct *up = container_of(kobj, struct user_struct, kobj); | ||
| 209 | unsigned long rt_period; | ||
| 210 | int rc; | ||
| 211 | |||
| 212 | sscanf(buf, "%lu", &rt_period); | ||
| 213 | |||
| 214 | rc = sched_group_set_rt_period(up->tg, rt_period); | ||
| 215 | |||
| 216 | return (rc ? rc : size); | ||
| 217 | } | ||
| 218 | |||
| 219 | static struct kobj_attribute cpu_rt_period_attr = | ||
| 220 | __ATTR(cpu_rt_period, 0644, cpu_rt_period_show, cpu_rt_period_store); | ||
| 221 | #endif | ||
| 222 | |||
| 223 | /* default attributes per uid directory */ | ||
| 224 | static struct attribute *uids_attributes[] = { | ||
| 225 | #ifdef CONFIG_FAIR_GROUP_SCHED | ||
| 226 | &cpu_share_attr.attr, | ||
| 227 | #endif | ||
| 228 | #ifdef CONFIG_RT_GROUP_SCHED | ||
| 229 | &cpu_rt_runtime_attr.attr, | ||
| 230 | &cpu_rt_period_attr.attr, | ||
| 231 | #endif | ||
| 232 | NULL | ||
| 233 | }; | ||
| 234 | |||
| 235 | /* the lifetime of user_struct is not managed by the core (now) */ | ||
| 236 | static void uids_release(struct kobject *kobj) | ||
| 237 | { | ||
| 238 | return; | ||
| 239 | } | ||
| 240 | |||
| 241 | static struct kobj_type uids_ktype = { | ||
| 242 | .sysfs_ops = &kobj_sysfs_ops, | ||
| 243 | .default_attrs = uids_attributes, | ||
| 244 | .release = uids_release, | ||
| 245 | }; | ||
| 246 | |||
| 247 | /* | ||
| 248 | * Create /sys/kernel/uids/<uid>/cpu_share file for this user | ||
| 249 | * We do not create this file for users in a user namespace (until | ||
| 250 | * sysfs tagging is implemented). | ||
| 251 | * | ||
| 252 | * See Documentation/scheduler/sched-design-CFS.txt for ramifications. | ||
| 253 | */ | ||
| 254 | static int uids_user_create(struct user_struct *up) | ||
| 255 | { | ||
| 256 | struct kobject *kobj = &up->kobj; | ||
| 257 | int error; | ||
| 258 | |||
| 259 | memset(kobj, 0, sizeof(struct kobject)); | ||
| 260 | if (up->user_ns != &init_user_ns) | ||
| 261 | return 0; | ||
| 262 | kobj->kset = uids_kset; | ||
| 263 | error = kobject_init_and_add(kobj, &uids_ktype, NULL, "%d", up->uid); | ||
| 264 | if (error) { | ||
| 265 | kobject_put(kobj); | ||
| 266 | goto done; | ||
| 267 | } | ||
| 268 | |||
| 269 | kobject_uevent(kobj, KOBJ_ADD); | ||
| 270 | done: | ||
| 271 | return error; | ||
| 272 | } | ||
| 273 | |||
| 274 | /* create these entries in sysfs: | ||
| 275 | * "/sys/kernel/uids" directory | ||
| 276 | * "/sys/kernel/uids/0" directory (for root user) | ||
| 277 | * "/sys/kernel/uids/0/cpu_share" file (for root user) | ||
| 278 | */ | ||
| 279 | int __init uids_sysfs_init(void) | ||
| 280 | { | ||
| 281 | uids_kset = kset_create_and_add("uids", NULL, kernel_kobj); | ||
| 282 | if (!uids_kset) | ||
| 283 | return -ENOMEM; | ||
| 284 | |||
| 285 | return uids_user_create(&root_user); | ||
| 286 | } | ||
| 287 | |||
| 288 | /* delayed work function to remove sysfs directory for a user and free up | ||
| 289 | * corresponding structures. | ||
| 290 | */ | ||
| 291 | static void cleanup_user_struct(struct work_struct *w) | ||
| 292 | { | ||
| 293 | struct user_struct *up = container_of(w, struct user_struct, work.work); | ||
| 294 | unsigned long flags; | ||
| 295 | int remove_user = 0; | ||
| 296 | |||
| 297 | /* Make uid_hash_remove() + sysfs_remove_file() + kobject_del() | ||
| 298 | * atomic. | ||
| 299 | */ | ||
| 300 | uids_mutex_lock(); | ||
| 301 | |||
| 302 | spin_lock_irqsave(&uidhash_lock, flags); | ||
| 303 | if (atomic_read(&up->__count) == 0) { | ||
| 304 | uid_hash_remove(up); | ||
| 305 | remove_user = 1; | ||
| 306 | } | ||
| 307 | spin_unlock_irqrestore(&uidhash_lock, flags); | ||
| 308 | |||
| 309 | if (!remove_user) | ||
| 310 | goto done; | ||
| 311 | |||
| 312 | if (up->user_ns == &init_user_ns) { | ||
| 313 | kobject_uevent(&up->kobj, KOBJ_REMOVE); | ||
| 314 | kobject_del(&up->kobj); | ||
| 315 | kobject_put(&up->kobj); | ||
| 316 | } | ||
| 317 | |||
| 318 | sched_destroy_user(up); | ||
| 319 | key_put(up->uid_keyring); | ||
| 320 | key_put(up->session_keyring); | ||
| 321 | kmem_cache_free(uid_cachep, up); | ||
| 322 | |||
| 323 | done: | ||
| 324 | uids_mutex_unlock(); | ||
| 325 | } | ||
| 326 | |||
| 327 | /* IRQs are disabled and uidhash_lock is held upon function entry. | ||
| 328 | * IRQ state (as stored in flags) is restored and uidhash_lock released | ||
| 329 | * upon function exit. | ||
| 330 | */ | ||
| 331 | static void free_user(struct user_struct *up, unsigned long flags) | ||
| 332 | { | ||
| 333 | INIT_DELAYED_WORK(&up->work, cleanup_user_struct); | ||
| 334 | schedule_delayed_work(&up->work, msecs_to_jiffies(1000)); | ||
| 335 | spin_unlock_irqrestore(&uidhash_lock, flags); | ||
| 336 | } | ||
| 337 | |||
| 338 | #else /* CONFIG_USER_SCHED && CONFIG_SYSFS */ | ||
| 339 | |||
| 340 | static struct user_struct *uid_hash_find(uid_t uid, struct hlist_head *hashent) | 75 | static struct user_struct *uid_hash_find(uid_t uid, struct hlist_head *hashent) |
| 341 | { | 76 | { |
| 342 | struct user_struct *user; | 77 | struct user_struct *user; |
| @@ -352,11 +87,6 @@ static struct user_struct *uid_hash_find(uid_t uid, struct hlist_head *hashent) | |||
| 352 | return NULL; | 87 | return NULL; |
| 353 | } | 88 | } |
| 354 | 89 | ||
| 355 | int uids_sysfs_init(void) { return 0; } | ||
| 356 | static inline int uids_user_create(struct user_struct *up) { return 0; } | ||
| 357 | static inline void uids_mutex_lock(void) { } | ||
| 358 | static inline void uids_mutex_unlock(void) { } | ||
| 359 | |||
| 360 | /* IRQs are disabled and uidhash_lock is held upon function entry. | 90 | /* IRQs are disabled and uidhash_lock is held upon function entry. |
| 361 | * IRQ state (as stored in flags) is restored and uidhash_lock released | 91 | * IRQ state (as stored in flags) is restored and uidhash_lock released |
| 362 | * upon function exit. | 92 | * upon function exit. |
| @@ -365,32 +95,11 @@ static void free_user(struct user_struct *up, unsigned long flags) | |||
| 365 | { | 95 | { |
| 366 | uid_hash_remove(up); | 96 | uid_hash_remove(up); |
| 367 | spin_unlock_irqrestore(&uidhash_lock, flags); | 97 | spin_unlock_irqrestore(&uidhash_lock, flags); |
| 368 | sched_destroy_user(up); | ||
| 369 | key_put(up->uid_keyring); | 98 | key_put(up->uid_keyring); |
| 370 | key_put(up->session_keyring); | 99 | key_put(up->session_keyring); |
| 371 | kmem_cache_free(uid_cachep, up); | 100 | kmem_cache_free(uid_cachep, up); |
| 372 | } | 101 | } |
| 373 | 102 | ||
| 374 | #endif | ||
| 375 | |||
| 376 | #if defined(CONFIG_RT_GROUP_SCHED) && defined(CONFIG_USER_SCHED) | ||
| 377 | /* | ||
| 378 | * We need to check if a setuid can take place. This function should be called | ||
| 379 | * before successfully completing the setuid. | ||
| 380 | */ | ||
| 381 | int task_can_switch_user(struct user_struct *up, struct task_struct *tsk) | ||
| 382 | { | ||
| 383 | |||
| 384 | return sched_rt_can_attach(up->tg, tsk); | ||
| 385 | |||
| 386 | } | ||
| 387 | #else | ||
| 388 | int task_can_switch_user(struct user_struct *up, struct task_struct *tsk) | ||
| 389 | { | ||
| 390 | return 1; | ||
| 391 | } | ||
| 392 | #endif | ||
| 393 | |||
| 394 | /* | 103 | /* |
| 395 | * Locate the user_struct for the passed UID. If found, take a ref on it. The | 104 | * Locate the user_struct for the passed UID. If found, take a ref on it. The |
| 396 | * caller must undo that ref with free_uid(). | 105 | * caller must undo that ref with free_uid(). |
| @@ -431,8 +140,6 @@ struct user_struct *alloc_uid(struct user_namespace *ns, uid_t uid) | |||
| 431 | /* Make uid_hash_find() + uids_user_create() + uid_hash_insert() | 140 | /* Make uid_hash_find() + uids_user_create() + uid_hash_insert() |
| 432 | * atomic. | 141 | * atomic. |
| 433 | */ | 142 | */ |
| 434 | uids_mutex_lock(); | ||
| 435 | |||
| 436 | spin_lock_irq(&uidhash_lock); | 143 | spin_lock_irq(&uidhash_lock); |
| 437 | up = uid_hash_find(uid, hashent); | 144 | up = uid_hash_find(uid, hashent); |
| 438 | spin_unlock_irq(&uidhash_lock); | 145 | spin_unlock_irq(&uidhash_lock); |
| @@ -445,14 +152,8 @@ struct user_struct *alloc_uid(struct user_namespace *ns, uid_t uid) | |||
| 445 | new->uid = uid; | 152 | new->uid = uid; |
| 446 | atomic_set(&new->__count, 1); | 153 | atomic_set(&new->__count, 1); |
| 447 | 154 | ||
| 448 | if (sched_create_user(new) < 0) | ||
| 449 | goto out_free_user; | ||
| 450 | |||
| 451 | new->user_ns = get_user_ns(ns); | 155 | new->user_ns = get_user_ns(ns); |
| 452 | 156 | ||
| 453 | if (uids_user_create(new)) | ||
| 454 | goto out_destoy_sched; | ||
| 455 | |||
| 456 | /* | 157 | /* |
| 457 | * Before adding this, check whether we raced | 158 | * Before adding this, check whether we raced |
| 458 | * on adding the same user already.. | 159 | * on adding the same user already.. |
| @@ -475,17 +176,11 @@ struct user_struct *alloc_uid(struct user_namespace *ns, uid_t uid) | |||
| 475 | spin_unlock_irq(&uidhash_lock); | 176 | spin_unlock_irq(&uidhash_lock); |
| 476 | } | 177 | } |
| 477 | 178 | ||
| 478 | uids_mutex_unlock(); | ||
| 479 | |||
| 480 | return up; | 179 | return up; |
| 481 | 180 | ||
| 482 | out_destoy_sched: | ||
| 483 | sched_destroy_user(new); | ||
| 484 | put_user_ns(new->user_ns); | 181 | put_user_ns(new->user_ns); |
| 485 | out_free_user: | ||
| 486 | kmem_cache_free(uid_cachep, new); | 182 | kmem_cache_free(uid_cachep, new); |
| 487 | out_unlock: | 183 | out_unlock: |
| 488 | uids_mutex_unlock(); | ||
| 489 | return NULL; | 184 | return NULL; |
| 490 | } | 185 | } |
| 491 | 186 | ||
