diff options
Diffstat (limited to 'kernel')
| -rw-r--r-- | kernel/irq/migration.c | 14 | ||||
| -rw-r--r-- | kernel/sched_fair.c | 13 | ||||
| -rw-r--r-- | kernel/sched_rt.c | 2 | ||||
| -rw-r--r-- | kernel/sys.c | 3 |
4 files changed, 19 insertions, 13 deletions
diff --git a/kernel/irq/migration.c b/kernel/irq/migration.c index 1d254194048..441fd629ff0 100644 --- a/kernel/irq/migration.c +++ b/kernel/irq/migration.c | |||
| @@ -56,6 +56,7 @@ void move_masked_irq(int irq) | |||
| 56 | void move_native_irq(int irq) | 56 | void move_native_irq(int irq) |
| 57 | { | 57 | { |
| 58 | struct irq_desc *desc = irq_to_desc(irq); | 58 | struct irq_desc *desc = irq_to_desc(irq); |
| 59 | bool masked; | ||
| 59 | 60 | ||
| 60 | if (likely(!(desc->status & IRQ_MOVE_PENDING))) | 61 | if (likely(!(desc->status & IRQ_MOVE_PENDING))) |
| 61 | return; | 62 | return; |
| @@ -63,8 +64,15 @@ void move_native_irq(int irq) | |||
| 63 | if (unlikely(desc->status & IRQ_DISABLED)) | 64 | if (unlikely(desc->status & IRQ_DISABLED)) |
| 64 | return; | 65 | return; |
| 65 | 66 | ||
| 66 | desc->irq_data.chip->irq_mask(&desc->irq_data); | 67 | /* |
| 68 | * Be careful vs. already masked interrupts. If this is a | ||
| 69 | * threaded interrupt with ONESHOT set, we can end up with an | ||
| 70 | * interrupt storm. | ||
| 71 | */ | ||
| 72 | masked = desc->status & IRQ_MASKED; | ||
| 73 | if (!masked) | ||
| 74 | desc->irq_data.chip->irq_mask(&desc->irq_data); | ||
| 67 | move_masked_irq(irq); | 75 | move_masked_irq(irq); |
| 68 | desc->irq_data.chip->irq_unmask(&desc->irq_data); | 76 | if (!masked) |
| 77 | desc->irq_data.chip->irq_unmask(&desc->irq_data); | ||
| 69 | } | 78 | } |
| 70 | |||
diff --git a/kernel/sched_fair.c b/kernel/sched_fair.c index 354769979c0..0c26e2df450 100644 --- a/kernel/sched_fair.c +++ b/kernel/sched_fair.c | |||
| @@ -722,10 +722,10 @@ static void update_cfs_load(struct cfs_rq *cfs_rq, int global_update) | |||
| 722 | u64 now, delta; | 722 | u64 now, delta; |
| 723 | unsigned long load = cfs_rq->load.weight; | 723 | unsigned long load = cfs_rq->load.weight; |
| 724 | 724 | ||
| 725 | if (!cfs_rq) | 725 | if (cfs_rq->tg == &root_task_group) |
| 726 | return; | 726 | return; |
| 727 | 727 | ||
| 728 | now = rq_of(cfs_rq)->clock; | 728 | now = rq_of(cfs_rq)->clock_task; |
| 729 | delta = now - cfs_rq->load_stamp; | 729 | delta = now - cfs_rq->load_stamp; |
| 730 | 730 | ||
| 731 | /* truncate load history at 4 idle periods */ | 731 | /* truncate load history at 4 idle periods */ |
| @@ -830,9 +830,6 @@ static void update_cfs_shares(struct cfs_rq *cfs_rq, long weight_delta) | |||
| 830 | struct sched_entity *se; | 830 | struct sched_entity *se; |
| 831 | long shares; | 831 | long shares; |
| 832 | 832 | ||
| 833 | if (!cfs_rq) | ||
| 834 | return; | ||
| 835 | |||
| 836 | tg = cfs_rq->tg; | 833 | tg = cfs_rq->tg; |
| 837 | se = tg->se[cpu_of(rq_of(cfs_rq))]; | 834 | se = tg->se[cpu_of(rq_of(cfs_rq))]; |
| 838 | if (!se) | 835 | if (!se) |
| @@ -1432,7 +1429,7 @@ static inline unsigned long effective_load(struct task_group *tg, int cpu, | |||
| 1432 | 1429 | ||
| 1433 | static int wake_affine(struct sched_domain *sd, struct task_struct *p, int sync) | 1430 | static int wake_affine(struct sched_domain *sd, struct task_struct *p, int sync) |
| 1434 | { | 1431 | { |
| 1435 | unsigned long this_load, load; | 1432 | s64 this_load, load; |
| 1436 | int idx, this_cpu, prev_cpu; | 1433 | int idx, this_cpu, prev_cpu; |
| 1437 | unsigned long tl_per_task; | 1434 | unsigned long tl_per_task; |
| 1438 | struct task_group *tg; | 1435 | struct task_group *tg; |
| @@ -1471,8 +1468,8 @@ static int wake_affine(struct sched_domain *sd, struct task_struct *p, int sync) | |||
| 1471 | * Otherwise check if either cpus are near enough in load to allow this | 1468 | * Otherwise check if either cpus are near enough in load to allow this |
| 1472 | * task to be woken on this_cpu. | 1469 | * task to be woken on this_cpu. |
| 1473 | */ | 1470 | */ |
| 1474 | if (this_load) { | 1471 | if (this_load > 0) { |
| 1475 | unsigned long this_eff_load, prev_eff_load; | 1472 | s64 this_eff_load, prev_eff_load; |
| 1476 | 1473 | ||
| 1477 | this_eff_load = 100; | 1474 | this_eff_load = 100; |
| 1478 | this_eff_load *= power_of(prev_cpu); | 1475 | this_eff_load *= power_of(prev_cpu); |
diff --git a/kernel/sched_rt.c b/kernel/sched_rt.c index c914ec747ca..ad6267714c8 100644 --- a/kernel/sched_rt.c +++ b/kernel/sched_rt.c | |||
| @@ -625,7 +625,7 @@ static void update_curr_rt(struct rq *rq) | |||
| 625 | struct rt_rq *rt_rq = rt_rq_of_se(rt_se); | 625 | struct rt_rq *rt_rq = rt_rq_of_se(rt_se); |
| 626 | u64 delta_exec; | 626 | u64 delta_exec; |
| 627 | 627 | ||
| 628 | if (!task_has_rt_policy(curr)) | 628 | if (curr->sched_class != &rt_sched_class) |
| 629 | return; | 629 | return; |
| 630 | 630 | ||
| 631 | delta_exec = rq->clock_task - curr->se.exec_start; | 631 | delta_exec = rq->clock_task - curr->se.exec_start; |
diff --git a/kernel/sys.c b/kernel/sys.c index 31b71a276b4..18da702ec81 100644 --- a/kernel/sys.c +++ b/kernel/sys.c | |||
| @@ -1385,7 +1385,8 @@ static int check_prlimit_permission(struct task_struct *task) | |||
| 1385 | const struct cred *cred = current_cred(), *tcred; | 1385 | const struct cred *cred = current_cred(), *tcred; |
| 1386 | 1386 | ||
| 1387 | tcred = __task_cred(task); | 1387 | tcred = __task_cred(task); |
| 1388 | if ((cred->uid != tcred->euid || | 1388 | if (current != task && |
| 1389 | (cred->uid != tcred->euid || | ||
| 1389 | cred->uid != tcred->suid || | 1390 | cred->uid != tcred->suid || |
| 1390 | cred->uid != tcred->uid || | 1391 | cred->uid != tcred->uid || |
| 1391 | cred->gid != tcred->egid || | 1392 | cred->gid != tcred->egid || |
