diff options
Diffstat (limited to 'kernel')
-rw-r--r-- | kernel/perf_event.c | 27 | ||||
-rw-r--r-- | kernel/sched.c | 6 |
2 files changed, 17 insertions, 16 deletions
diff --git a/kernel/perf_event.c b/kernel/perf_event.c index 03cc061398d1..099bd662daa6 100644 --- a/kernel/perf_event.c +++ b/kernel/perf_event.c | |||
@@ -1170,9 +1170,9 @@ static void perf_event_sync_stat(struct perf_event_context *ctx, | |||
1170 | * not restart the event. | 1170 | * not restart the event. |
1171 | */ | 1171 | */ |
1172 | void perf_event_task_sched_out(struct task_struct *task, | 1172 | void perf_event_task_sched_out(struct task_struct *task, |
1173 | struct task_struct *next, int cpu) | 1173 | struct task_struct *next) |
1174 | { | 1174 | { |
1175 | struct perf_cpu_context *cpuctx = &per_cpu(perf_cpu_context, cpu); | 1175 | struct perf_cpu_context *cpuctx = &__get_cpu_var(perf_cpu_context); |
1176 | struct perf_event_context *ctx = task->perf_event_ctxp; | 1176 | struct perf_event_context *ctx = task->perf_event_ctxp; |
1177 | struct perf_event_context *next_ctx; | 1177 | struct perf_event_context *next_ctx; |
1178 | struct perf_event_context *parent; | 1178 | struct perf_event_context *parent; |
@@ -1252,8 +1252,9 @@ static void perf_event_cpu_sched_out(struct perf_cpu_context *cpuctx) | |||
1252 | 1252 | ||
1253 | static void | 1253 | static void |
1254 | __perf_event_sched_in(struct perf_event_context *ctx, | 1254 | __perf_event_sched_in(struct perf_event_context *ctx, |
1255 | struct perf_cpu_context *cpuctx, int cpu) | 1255 | struct perf_cpu_context *cpuctx) |
1256 | { | 1256 | { |
1257 | int cpu = smp_processor_id(); | ||
1257 | struct perf_event *event; | 1258 | struct perf_event *event; |
1258 | int can_add_hw = 1; | 1259 | int can_add_hw = 1; |
1259 | 1260 | ||
@@ -1326,24 +1327,24 @@ __perf_event_sched_in(struct perf_event_context *ctx, | |||
1326 | * accessing the event control register. If a NMI hits, then it will | 1327 | * accessing the event control register. If a NMI hits, then it will |
1327 | * keep the event running. | 1328 | * keep the event running. |
1328 | */ | 1329 | */ |
1329 | void perf_event_task_sched_in(struct task_struct *task, int cpu) | 1330 | void perf_event_task_sched_in(struct task_struct *task) |
1330 | { | 1331 | { |
1331 | struct perf_cpu_context *cpuctx = &per_cpu(perf_cpu_context, cpu); | 1332 | struct perf_cpu_context *cpuctx = &__get_cpu_var(perf_cpu_context); |
1332 | struct perf_event_context *ctx = task->perf_event_ctxp; | 1333 | struct perf_event_context *ctx = task->perf_event_ctxp; |
1333 | 1334 | ||
1334 | if (likely(!ctx)) | 1335 | if (likely(!ctx)) |
1335 | return; | 1336 | return; |
1336 | if (cpuctx->task_ctx == ctx) | 1337 | if (cpuctx->task_ctx == ctx) |
1337 | return; | 1338 | return; |
1338 | __perf_event_sched_in(ctx, cpuctx, cpu); | 1339 | __perf_event_sched_in(ctx, cpuctx); |
1339 | cpuctx->task_ctx = ctx; | 1340 | cpuctx->task_ctx = ctx; |
1340 | } | 1341 | } |
1341 | 1342 | ||
1342 | static void perf_event_cpu_sched_in(struct perf_cpu_context *cpuctx, int cpu) | 1343 | static void perf_event_cpu_sched_in(struct perf_cpu_context *cpuctx) |
1343 | { | 1344 | { |
1344 | struct perf_event_context *ctx = &cpuctx->ctx; | 1345 | struct perf_event_context *ctx = &cpuctx->ctx; |
1345 | 1346 | ||
1346 | __perf_event_sched_in(ctx, cpuctx, cpu); | 1347 | __perf_event_sched_in(ctx, cpuctx); |
1347 | } | 1348 | } |
1348 | 1349 | ||
1349 | #define MAX_INTERRUPTS (~0ULL) | 1350 | #define MAX_INTERRUPTS (~0ULL) |
@@ -1461,7 +1462,7 @@ static void rotate_ctx(struct perf_event_context *ctx) | |||
1461 | raw_spin_unlock(&ctx->lock); | 1462 | raw_spin_unlock(&ctx->lock); |
1462 | } | 1463 | } |
1463 | 1464 | ||
1464 | void perf_event_task_tick(struct task_struct *curr, int cpu) | 1465 | void perf_event_task_tick(struct task_struct *curr) |
1465 | { | 1466 | { |
1466 | struct perf_cpu_context *cpuctx; | 1467 | struct perf_cpu_context *cpuctx; |
1467 | struct perf_event_context *ctx; | 1468 | struct perf_event_context *ctx; |
@@ -1469,7 +1470,7 @@ void perf_event_task_tick(struct task_struct *curr, int cpu) | |||
1469 | if (!atomic_read(&nr_events)) | 1470 | if (!atomic_read(&nr_events)) |
1470 | return; | 1471 | return; |
1471 | 1472 | ||
1472 | cpuctx = &per_cpu(perf_cpu_context, cpu); | 1473 | cpuctx = &__get_cpu_var(perf_cpu_context); |
1473 | ctx = curr->perf_event_ctxp; | 1474 | ctx = curr->perf_event_ctxp; |
1474 | 1475 | ||
1475 | perf_ctx_adjust_freq(&cpuctx->ctx); | 1476 | perf_ctx_adjust_freq(&cpuctx->ctx); |
@@ -1484,9 +1485,9 @@ void perf_event_task_tick(struct task_struct *curr, int cpu) | |||
1484 | if (ctx) | 1485 | if (ctx) |
1485 | rotate_ctx(ctx); | 1486 | rotate_ctx(ctx); |
1486 | 1487 | ||
1487 | perf_event_cpu_sched_in(cpuctx, cpu); | 1488 | perf_event_cpu_sched_in(cpuctx); |
1488 | if (ctx) | 1489 | if (ctx) |
1489 | perf_event_task_sched_in(curr, cpu); | 1490 | perf_event_task_sched_in(curr); |
1490 | } | 1491 | } |
1491 | 1492 | ||
1492 | /* | 1493 | /* |
@@ -1527,7 +1528,7 @@ static void perf_event_enable_on_exec(struct task_struct *task) | |||
1527 | 1528 | ||
1528 | raw_spin_unlock(&ctx->lock); | 1529 | raw_spin_unlock(&ctx->lock); |
1529 | 1530 | ||
1530 | perf_event_task_sched_in(task, smp_processor_id()); | 1531 | perf_event_task_sched_in(task); |
1531 | out: | 1532 | out: |
1532 | local_irq_restore(flags); | 1533 | local_irq_restore(flags); |
1533 | } | 1534 | } |
diff --git a/kernel/sched.c b/kernel/sched.c index 18cceeecce35..d6527ac0f6e7 100644 --- a/kernel/sched.c +++ b/kernel/sched.c | |||
@@ -2752,7 +2752,7 @@ static void finish_task_switch(struct rq *rq, struct task_struct *prev) | |||
2752 | */ | 2752 | */ |
2753 | prev_state = prev->state; | 2753 | prev_state = prev->state; |
2754 | finish_arch_switch(prev); | 2754 | finish_arch_switch(prev); |
2755 | perf_event_task_sched_in(current, cpu_of(rq)); | 2755 | perf_event_task_sched_in(current); |
2756 | finish_lock_switch(rq, prev); | 2756 | finish_lock_switch(rq, prev); |
2757 | 2757 | ||
2758 | fire_sched_in_preempt_notifiers(current); | 2758 | fire_sched_in_preempt_notifiers(current); |
@@ -5266,7 +5266,7 @@ void scheduler_tick(void) | |||
5266 | curr->sched_class->task_tick(rq, curr, 0); | 5266 | curr->sched_class->task_tick(rq, curr, 0); |
5267 | raw_spin_unlock(&rq->lock); | 5267 | raw_spin_unlock(&rq->lock); |
5268 | 5268 | ||
5269 | perf_event_task_tick(curr, cpu); | 5269 | perf_event_task_tick(curr); |
5270 | 5270 | ||
5271 | #ifdef CONFIG_SMP | 5271 | #ifdef CONFIG_SMP |
5272 | rq->idle_at_tick = idle_cpu(cpu); | 5272 | rq->idle_at_tick = idle_cpu(cpu); |
@@ -5480,7 +5480,7 @@ need_resched_nonpreemptible: | |||
5480 | 5480 | ||
5481 | if (likely(prev != next)) { | 5481 | if (likely(prev != next)) { |
5482 | sched_info_switch(prev, next); | 5482 | sched_info_switch(prev, next); |
5483 | perf_event_task_sched_out(prev, next, cpu); | 5483 | perf_event_task_sched_out(prev, next); |
5484 | 5484 | ||
5485 | rq->nr_switches++; | 5485 | rq->nr_switches++; |
5486 | rq->curr = next; | 5486 | rq->curr = next; |