diff options
Diffstat (limited to 'kernel/sched')
| -rw-r--r-- | kernel/sched/clock.c | 26 | ||||
| -rw-r--r-- | kernel/sched/core.c | 8 | ||||
| -rw-r--r-- | kernel/sched/cputime.c | 2 |
3 files changed, 32 insertions, 4 deletions
diff --git a/kernel/sched/clock.c b/kernel/sched/clock.c index c685e31492df..c3ae1446461c 100644 --- a/kernel/sched/clock.c +++ b/kernel/sched/clock.c | |||
| @@ -176,10 +176,36 @@ static u64 sched_clock_remote(struct sched_clock_data *scd) | |||
| 176 | u64 this_clock, remote_clock; | 176 | u64 this_clock, remote_clock; |
| 177 | u64 *ptr, old_val, val; | 177 | u64 *ptr, old_val, val; |
| 178 | 178 | ||
| 179 | #if BITS_PER_LONG != 64 | ||
| 180 | again: | ||
| 181 | /* | ||
| 182 | * Careful here: The local and the remote clock values need to | ||
| 183 | * be read out atomic as we need to compare the values and | ||
| 184 | * then update either the local or the remote side. So the | ||
| 185 | * cmpxchg64 below only protects one readout. | ||
| 186 | * | ||
| 187 | * We must reread via sched_clock_local() in the retry case on | ||
| 188 | * 32bit as an NMI could use sched_clock_local() via the | ||
| 189 | * tracer and hit between the readout of | ||
| 190 | * the low32bit and the high 32bit portion. | ||
| 191 | */ | ||
| 192 | this_clock = sched_clock_local(my_scd); | ||
| 193 | /* | ||
| 194 | * We must enforce atomic readout on 32bit, otherwise the | ||
| 195 | * update on the remote cpu can hit inbetween the readout of | ||
| 196 | * the low32bit and the high 32bit portion. | ||
| 197 | */ | ||
| 198 | remote_clock = cmpxchg64(&scd->clock, 0, 0); | ||
| 199 | #else | ||
| 200 | /* | ||
| 201 | * On 64bit the read of [my]scd->clock is atomic versus the | ||
| 202 | * update, so we can avoid the above 32bit dance. | ||
| 203 | */ | ||
| 179 | sched_clock_local(my_scd); | 204 | sched_clock_local(my_scd); |
| 180 | again: | 205 | again: |
| 181 | this_clock = my_scd->clock; | 206 | this_clock = my_scd->clock; |
| 182 | remote_clock = scd->clock; | 207 | remote_clock = scd->clock; |
| 208 | #endif | ||
| 183 | 209 | ||
| 184 | /* | 210 | /* |
| 185 | * Use the opportunity that we have both locks | 211 | * Use the opportunity that we have both locks |
diff --git a/kernel/sched/core.c b/kernel/sched/core.c index 7f12624a393c..67d04651f44b 100644 --- a/kernel/sched/core.c +++ b/kernel/sched/core.c | |||
| @@ -1498,8 +1498,10 @@ static void try_to_wake_up_local(struct task_struct *p) | |||
| 1498 | { | 1498 | { |
| 1499 | struct rq *rq = task_rq(p); | 1499 | struct rq *rq = task_rq(p); |
| 1500 | 1500 | ||
| 1501 | BUG_ON(rq != this_rq()); | 1501 | if (WARN_ON_ONCE(rq != this_rq()) || |
| 1502 | BUG_ON(p == current); | 1502 | WARN_ON_ONCE(p == current)) |
| 1503 | return; | ||
| 1504 | |||
| 1503 | lockdep_assert_held(&rq->lock); | 1505 | lockdep_assert_held(&rq->lock); |
| 1504 | 1506 | ||
| 1505 | if (!raw_spin_trylock(&p->pi_lock)) { | 1507 | if (!raw_spin_trylock(&p->pi_lock)) { |
| @@ -4999,7 +5001,7 @@ static void sd_free_ctl_entry(struct ctl_table **tablep) | |||
| 4999 | } | 5001 | } |
| 5000 | 5002 | ||
| 5001 | static int min_load_idx = 0; | 5003 | static int min_load_idx = 0; |
| 5002 | static int max_load_idx = CPU_LOAD_IDX_MAX; | 5004 | static int max_load_idx = CPU_LOAD_IDX_MAX-1; |
| 5003 | 5005 | ||
| 5004 | static void | 5006 | static void |
| 5005 | set_table_entry(struct ctl_table *entry, | 5007 | set_table_entry(struct ctl_table *entry, |
diff --git a/kernel/sched/cputime.c b/kernel/sched/cputime.c index ed12cbb135f4..e93cca92f38b 100644 --- a/kernel/sched/cputime.c +++ b/kernel/sched/cputime.c | |||
| @@ -310,7 +310,7 @@ void thread_group_cputime(struct task_struct *tsk, struct task_cputime *times) | |||
| 310 | 310 | ||
| 311 | t = tsk; | 311 | t = tsk; |
| 312 | do { | 312 | do { |
| 313 | task_cputime(tsk, &utime, &stime); | 313 | task_cputime(t, &utime, &stime); |
| 314 | times->utime += utime; | 314 | times->utime += utime; |
| 315 | times->stime += stime; | 315 | times->stime += stime; |
| 316 | times->sum_exec_runtime += task_sched_runtime(t); | 316 | times->sum_exec_runtime += task_sched_runtime(t); |
