diff options
Diffstat (limited to 'arch/ia64/kernel/perfmon.c')
-rw-r--r-- | arch/ia64/kernel/perfmon.c | 18 |
1 files changed, 9 insertions, 9 deletions
diff --git a/arch/ia64/kernel/perfmon.c b/arch/ia64/kernel/perfmon.c index e7191ca30b16..b7133cabdbea 100644 --- a/arch/ia64/kernel/perfmon.c +++ b/arch/ia64/kernel/perfmon.c | |||
@@ -1318,7 +1318,7 @@ pfm_reserve_session(struct task_struct *task, int is_syswide, unsigned int cpu) | |||
1318 | { | 1318 | { |
1319 | unsigned long flags; | 1319 | unsigned long flags; |
1320 | /* | 1320 | /* |
1321 | * validy checks on cpu_mask have been done upstream | 1321 | * validity checks on cpu_mask have been done upstream |
1322 | */ | 1322 | */ |
1323 | LOCK_PFS(flags); | 1323 | LOCK_PFS(flags); |
1324 | 1324 | ||
@@ -1384,7 +1384,7 @@ pfm_unreserve_session(pfm_context_t *ctx, int is_syswide, unsigned int cpu) | |||
1384 | { | 1384 | { |
1385 | unsigned long flags; | 1385 | unsigned long flags; |
1386 | /* | 1386 | /* |
1387 | * validy checks on cpu_mask have been done upstream | 1387 | * validity checks on cpu_mask have been done upstream |
1388 | */ | 1388 | */ |
1389 | LOCK_PFS(flags); | 1389 | LOCK_PFS(flags); |
1390 | 1390 | ||
@@ -1835,7 +1835,7 @@ pfm_flush(struct file *filp, fl_owner_t id) | |||
1835 | /* | 1835 | /* |
1836 | * remove our file from the async queue, if we use this mode. | 1836 | * remove our file from the async queue, if we use this mode. |
1837 | * This can be done without the context being protected. We come | 1837 | * This can be done without the context being protected. We come |
1838 | * here when the context has become unreacheable by other tasks. | 1838 | * here when the context has become unreachable by other tasks. |
1839 | * | 1839 | * |
1840 | * We may still have active monitoring at this point and we may | 1840 | * We may still have active monitoring at this point and we may |
1841 | * end up in pfm_overflow_handler(). However, fasync_helper() | 1841 | * end up in pfm_overflow_handler(). However, fasync_helper() |
@@ -2132,7 +2132,7 @@ doit: | |||
2132 | filp->private_data = NULL; | 2132 | filp->private_data = NULL; |
2133 | 2133 | ||
2134 | /* | 2134 | /* |
2135 | * if we free on the spot, the context is now completely unreacheable | 2135 | * if we free on the spot, the context is now completely unreachable |
2136 | * from the callers side. The monitored task side is also cut, so we | 2136 | * from the callers side. The monitored task side is also cut, so we |
2137 | * can freely cut. | 2137 | * can freely cut. |
2138 | * | 2138 | * |
@@ -2562,7 +2562,7 @@ pfm_reset_pmu_state(pfm_context_t *ctx) | |||
2562 | ctx->ctx_all_pmcs[0] = pmu_conf->impl_pmcs[0] & ~0x1; | 2562 | ctx->ctx_all_pmcs[0] = pmu_conf->impl_pmcs[0] & ~0x1; |
2563 | 2563 | ||
2564 | /* | 2564 | /* |
2565 | * bitmask of all PMDs that are accesible to this context | 2565 | * bitmask of all PMDs that are accessible to this context |
2566 | */ | 2566 | */ |
2567 | ctx->ctx_all_pmds[0] = pmu_conf->impl_pmds[0]; | 2567 | ctx->ctx_all_pmds[0] = pmu_conf->impl_pmds[0]; |
2568 | 2568 | ||
@@ -3395,7 +3395,7 @@ pfm_read_pmds(pfm_context_t *ctx, void *arg, int count, struct pt_regs *regs) | |||
3395 | if (unlikely(!PMD_IS_IMPL(cnum))) goto error; | 3395 | if (unlikely(!PMD_IS_IMPL(cnum))) goto error; |
3396 | /* | 3396 | /* |
3397 | * we can only read the register that we use. That includes | 3397 | * we can only read the register that we use. That includes |
3398 | * the one we explicitely initialize AND the one we want included | 3398 | * the one we explicitly initialize AND the one we want included |
3399 | * in the sampling buffer (smpl_regs). | 3399 | * in the sampling buffer (smpl_regs). |
3400 | * | 3400 | * |
3401 | * Having this restriction allows optimization in the ctxsw routine | 3401 | * Having this restriction allows optimization in the ctxsw routine |
@@ -3715,7 +3715,7 @@ pfm_restart(pfm_context_t *ctx, void *arg, int count, struct pt_regs *regs) | |||
3715 | * if non-blocking, then we ensure that the task will go into | 3715 | * if non-blocking, then we ensure that the task will go into |
3716 | * pfm_handle_work() before returning to user mode. | 3716 | * pfm_handle_work() before returning to user mode. |
3717 | * | 3717 | * |
3718 | * We cannot explicitely reset another task, it MUST always | 3718 | * We cannot explicitly reset another task, it MUST always |
3719 | * be done by the task itself. This works for system wide because | 3719 | * be done by the task itself. This works for system wide because |
3720 | * the tool that is controlling the session is logically doing | 3720 | * the tool that is controlling the session is logically doing |
3721 | * "self-monitoring". | 3721 | * "self-monitoring". |
@@ -4644,7 +4644,7 @@ pfm_exit_thread(struct task_struct *task) | |||
4644 | switch(state) { | 4644 | switch(state) { |
4645 | case PFM_CTX_UNLOADED: | 4645 | case PFM_CTX_UNLOADED: |
4646 | /* | 4646 | /* |
4647 | * only comes to thios function if pfm_context is not NULL, i.e., cannot | 4647 | * only comes to this function if pfm_context is not NULL, i.e., cannot |
4648 | * be in unloaded state | 4648 | * be in unloaded state |
4649 | */ | 4649 | */ |
4650 | printk(KERN_ERR "perfmon: pfm_exit_thread [%d] ctx unloaded\n", task->pid); | 4650 | printk(KERN_ERR "perfmon: pfm_exit_thread [%d] ctx unloaded\n", task->pid); |
@@ -5247,7 +5247,7 @@ pfm_end_notify_user(pfm_context_t *ctx) | |||
5247 | 5247 | ||
5248 | /* | 5248 | /* |
5249 | * main overflow processing routine. | 5249 | * main overflow processing routine. |
5250 | * it can be called from the interrupt path or explicitely during the context switch code | 5250 | * it can be called from the interrupt path or explicitly during the context switch code |
5251 | */ | 5251 | */ |
5252 | static void | 5252 | static void |
5253 | pfm_overflow_handler(struct task_struct *task, pfm_context_t *ctx, u64 pmc0, struct pt_regs *regs) | 5253 | pfm_overflow_handler(struct task_struct *task, pfm_context_t *ctx, u64 pmc0, struct pt_regs *regs) |