diff options
| author | Keshavamurthy Anil S <anil.s.keshavamurthy@intel.com> | 2006-09-26 15:03:13 -0400 |
|---|---|---|
| committer | Tony Luck <tony.luck@intel.com> | 2006-09-26 15:03:13 -0400 |
| commit | 35589a8fa8138244e7f2ef9317c440aa580c9335 (patch) | |
| tree | c108ecc979e4b95ee5b6d617790212426b80d372 | |
| parent | dd562c05410e13e878a3ee0deb8ac06db2e132c7 (diff) | |
[IA64] Move perfmon tables from thread_struct to pfm_context
This patch renders thread_struct->pmcs[] and thread_struct->pmds[]
OBSOLETE. The actual table is moved to pfm_context structure which
saves space in thread_struct (in turn saving space in task_struct
which frees up more space for kernel stacks).
Signed-off-by: Stephane Eranian <eranian@hpl.hp.com>
Signed-off-by: Anil S Keshavamurthy <anil.s.keshavamurthy@intel.com>
Signed-off-by: Tony Luck <tony.luck@intel.com>
| -rw-r--r-- | arch/ia64/kernel/perfmon.c | 113 | ||||
| -rw-r--r-- | include/asm-ia64/processor.h | 12 |
2 files changed, 50 insertions, 75 deletions
diff --git a/arch/ia64/kernel/perfmon.c b/arch/ia64/kernel/perfmon.c index 84a7e52f56f6..45000d5d0cfa 100644 --- a/arch/ia64/kernel/perfmon.c +++ b/arch/ia64/kernel/perfmon.c | |||
| @@ -62,6 +62,9 @@ | |||
| 62 | 62 | ||
| 63 | #define PFM_INVALID_ACTIVATION (~0UL) | 63 | #define PFM_INVALID_ACTIVATION (~0UL) |
| 64 | 64 | ||
| 65 | #define PFM_NUM_PMC_REGS 64 /* PMC save area for ctxsw */ | ||
| 66 | #define PFM_NUM_PMD_REGS 64 /* PMD save area for ctxsw */ | ||
| 67 | |||
| 65 | /* | 68 | /* |
| 66 | * depth of message queue | 69 | * depth of message queue |
| 67 | */ | 70 | */ |
| @@ -296,14 +299,17 @@ typedef struct pfm_context { | |||
| 296 | unsigned long ctx_reload_pmcs[4]; /* bitmask of force reload PMC on ctxsw in */ | 299 | unsigned long ctx_reload_pmcs[4]; /* bitmask of force reload PMC on ctxsw in */ |
| 297 | unsigned long ctx_used_monitors[4]; /* bitmask of monitor PMC being used */ | 300 | unsigned long ctx_used_monitors[4]; /* bitmask of monitor PMC being used */ |
| 298 | 301 | ||
| 299 | unsigned long ctx_pmcs[IA64_NUM_PMC_REGS]; /* saved copies of PMC values */ | 302 | unsigned long ctx_pmcs[PFM_NUM_PMC_REGS]; /* saved copies of PMC values */ |
| 300 | 303 | ||
| 301 | unsigned int ctx_used_ibrs[1]; /* bitmask of used IBR (speedup ctxsw in) */ | 304 | unsigned int ctx_used_ibrs[1]; /* bitmask of used IBR (speedup ctxsw in) */ |
| 302 | unsigned int ctx_used_dbrs[1]; /* bitmask of used DBR (speedup ctxsw in) */ | 305 | unsigned int ctx_used_dbrs[1]; /* bitmask of used DBR (speedup ctxsw in) */ |
| 303 | unsigned long ctx_dbrs[IA64_NUM_DBG_REGS]; /* DBR values (cache) when not loaded */ | 306 | unsigned long ctx_dbrs[IA64_NUM_DBG_REGS]; /* DBR values (cache) when not loaded */ |
| 304 | unsigned long ctx_ibrs[IA64_NUM_DBG_REGS]; /* IBR values (cache) when not loaded */ | 307 | unsigned long ctx_ibrs[IA64_NUM_DBG_REGS]; /* IBR values (cache) when not loaded */ |
| 305 | 308 | ||
| 306 | pfm_counter_t ctx_pmds[IA64_NUM_PMD_REGS]; /* software state for PMDS */ | 309 | pfm_counter_t ctx_pmds[PFM_NUM_PMD_REGS]; /* software state for PMDS */ |
| 310 | |||
| 311 | unsigned long th_pmcs[PFM_NUM_PMC_REGS]; /* PMC thread save state */ | ||
| 312 | unsigned long th_pmds[PFM_NUM_PMD_REGS]; /* PMD thread save state */ | ||
| 307 | 313 | ||
| 308 | u64 ctx_saved_psr_up; /* only contains psr.up value */ | 314 | u64 ctx_saved_psr_up; /* only contains psr.up value */ |
| 309 | 315 | ||
| @@ -867,7 +873,6 @@ static void | |||
| 867 | pfm_mask_monitoring(struct task_struct *task) | 873 | pfm_mask_monitoring(struct task_struct *task) |
| 868 | { | 874 | { |
| 869 | pfm_context_t *ctx = PFM_GET_CTX(task); | 875 | pfm_context_t *ctx = PFM_GET_CTX(task); |
| 870 | struct thread_struct *th = &task->thread; | ||
| 871 | unsigned long mask, val, ovfl_mask; | 876 | unsigned long mask, val, ovfl_mask; |
| 872 | int i; | 877 | int i; |
| 873 | 878 | ||
| @@ -888,7 +893,7 @@ pfm_mask_monitoring(struct task_struct *task) | |||
| 888 | * So in both cases, the live register contains the owner's | 893 | * So in both cases, the live register contains the owner's |
| 889 | * state. We can ONLY touch the PMU registers and NOT the PSR. | 894 | * state. We can ONLY touch the PMU registers and NOT the PSR. |
| 890 | * | 895 | * |
| 891 | * As a consequence to this call, the thread->pmds[] array | 896 | * As a consequence to this call, the ctx->th_pmds[] array |
| 892 | * contains stale information which must be ignored | 897 | * contains stale information which must be ignored |
| 893 | * when context is reloaded AND monitoring is active (see | 898 | * when context is reloaded AND monitoring is active (see |
| 894 | * pfm_restart). | 899 | * pfm_restart). |
| @@ -923,9 +928,9 @@ pfm_mask_monitoring(struct task_struct *task) | |||
| 923 | mask = ctx->ctx_used_monitors[0] >> PMU_FIRST_COUNTER; | 928 | mask = ctx->ctx_used_monitors[0] >> PMU_FIRST_COUNTER; |
| 924 | for(i= PMU_FIRST_COUNTER; mask; i++, mask>>=1) { | 929 | for(i= PMU_FIRST_COUNTER; mask; i++, mask>>=1) { |
| 925 | if ((mask & 0x1) == 0UL) continue; | 930 | if ((mask & 0x1) == 0UL) continue; |
| 926 | ia64_set_pmc(i, th->pmcs[i] & ~0xfUL); | 931 | ia64_set_pmc(i, ctx->th_pmcs[i] & ~0xfUL); |
| 927 | th->pmcs[i] &= ~0xfUL; | 932 | ctx->th_pmcs[i] &= ~0xfUL; |
| 928 | DPRINT_ovfl(("pmc[%d]=0x%lx\n", i, th->pmcs[i])); | 933 | DPRINT_ovfl(("pmc[%d]=0x%lx\n", i, ctx->th_pmcs[i])); |
| 929 | } | 934 | } |
| 930 | /* | 935 | /* |
| 931 | * make all of this visible | 936 | * make all of this visible |
| @@ -942,7 +947,6 @@ static void | |||
| 942 | pfm_restore_monitoring(struct task_struct *task) | 947 | pfm_restore_monitoring(struct task_struct *task) |
| 943 | { | 948 | { |
| 944 | pfm_context_t *ctx = PFM_GET_CTX(task); | 949 | pfm_context_t *ctx = PFM_GET_CTX(task); |
| 945 | struct thread_struct *th = &task->thread; | ||
| 946 | unsigned long mask, ovfl_mask; | 950 | unsigned long mask, ovfl_mask; |
| 947 | unsigned long psr, val; | 951 | unsigned long psr, val; |
| 948 | int i, is_system; | 952 | int i, is_system; |
| @@ -1008,9 +1012,9 @@ pfm_restore_monitoring(struct task_struct *task) | |||
| 1008 | mask = ctx->ctx_used_monitors[0] >> PMU_FIRST_COUNTER; | 1012 | mask = ctx->ctx_used_monitors[0] >> PMU_FIRST_COUNTER; |
| 1009 | for(i= PMU_FIRST_COUNTER; mask; i++, mask>>=1) { | 1013 | for(i= PMU_FIRST_COUNTER; mask; i++, mask>>=1) { |
| 1010 | if ((mask & 0x1) == 0UL) continue; | 1014 | if ((mask & 0x1) == 0UL) continue; |
| 1011 | th->pmcs[i] = ctx->ctx_pmcs[i]; | 1015 | ctx->th_pmcs[i] = ctx->ctx_pmcs[i]; |
| 1012 | ia64_set_pmc(i, th->pmcs[i]); | 1016 | ia64_set_pmc(i, ctx->th_pmcs[i]); |
| 1013 | DPRINT(("[%d] pmc[%d]=0x%lx\n", task->pid, i, th->pmcs[i])); | 1017 | DPRINT(("[%d] pmc[%d]=0x%lx\n", task->pid, i, ctx->th_pmcs[i])); |
| 1014 | } | 1018 | } |
| 1015 | ia64_srlz_d(); | 1019 | ia64_srlz_d(); |
| 1016 | 1020 | ||
| @@ -1069,7 +1073,6 @@ pfm_restore_pmds(unsigned long *pmds, unsigned long mask) | |||
| 1069 | static inline void | 1073 | static inline void |
| 1070 | pfm_copy_pmds(struct task_struct *task, pfm_context_t *ctx) | 1074 | pfm_copy_pmds(struct task_struct *task, pfm_context_t *ctx) |
| 1071 | { | 1075 | { |
| 1072 | struct thread_struct *thread = &task->thread; | ||
| 1073 | unsigned long ovfl_val = pmu_conf->ovfl_val; | 1076 | unsigned long ovfl_val = pmu_conf->ovfl_val; |
| 1074 | unsigned long mask = ctx->ctx_all_pmds[0]; | 1077 | unsigned long mask = ctx->ctx_all_pmds[0]; |
| 1075 | unsigned long val; | 1078 | unsigned long val; |
| @@ -1091,11 +1094,11 @@ pfm_copy_pmds(struct task_struct *task, pfm_context_t *ctx) | |||
| 1091 | ctx->ctx_pmds[i].val = val & ~ovfl_val; | 1094 | ctx->ctx_pmds[i].val = val & ~ovfl_val; |
| 1092 | val &= ovfl_val; | 1095 | val &= ovfl_val; |
| 1093 | } | 1096 | } |
| 1094 | thread->pmds[i] = val; | 1097 | ctx->th_pmds[i] = val; |
| 1095 | 1098 | ||
| 1096 | DPRINT(("pmd[%d]=0x%lx soft_val=0x%lx\n", | 1099 | DPRINT(("pmd[%d]=0x%lx soft_val=0x%lx\n", |
| 1097 | i, | 1100 | i, |
| 1098 | thread->pmds[i], | 1101 | ctx->th_pmds[i], |
| 1099 | ctx->ctx_pmds[i].val)); | 1102 | ctx->ctx_pmds[i].val)); |
| 1100 | } | 1103 | } |
| 1101 | } | 1104 | } |
| @@ -1106,7 +1109,6 @@ pfm_copy_pmds(struct task_struct *task, pfm_context_t *ctx) | |||
| 1106 | static inline void | 1109 | static inline void |
| 1107 | pfm_copy_pmcs(struct task_struct *task, pfm_context_t *ctx) | 1110 | pfm_copy_pmcs(struct task_struct *task, pfm_context_t *ctx) |
| 1108 | { | 1111 | { |
| 1109 | struct thread_struct *thread = &task->thread; | ||
| 1110 | unsigned long mask = ctx->ctx_all_pmcs[0]; | 1112 | unsigned long mask = ctx->ctx_all_pmcs[0]; |
| 1111 | int i; | 1113 | int i; |
| 1112 | 1114 | ||
| @@ -1114,8 +1116,8 @@ pfm_copy_pmcs(struct task_struct *task, pfm_context_t *ctx) | |||
| 1114 | 1116 | ||
| 1115 | for (i=0; mask; i++, mask>>=1) { | 1117 | for (i=0; mask; i++, mask>>=1) { |
| 1116 | /* masking 0 with ovfl_val yields 0 */ | 1118 | /* masking 0 with ovfl_val yields 0 */ |
| 1117 | thread->pmcs[i] = ctx->ctx_pmcs[i]; | 1119 | ctx->th_pmcs[i] = ctx->ctx_pmcs[i]; |
| 1118 | DPRINT(("pmc[%d]=0x%lx\n", i, thread->pmcs[i])); | 1120 | DPRINT(("pmc[%d]=0x%lx\n", i, ctx->th_pmcs[i])); |
| 1119 | } | 1121 | } |
| 1120 | } | 1122 | } |
| 1121 | 1123 | ||
| @@ -2859,7 +2861,6 @@ pfm_reset_regs(pfm_context_t *ctx, unsigned long *ovfl_regs, int is_long_reset) | |||
| 2859 | static int | 2861 | static int |
| 2860 | pfm_write_pmcs(pfm_context_t *ctx, void *arg, int count, struct pt_regs *regs) | 2862 | pfm_write_pmcs(pfm_context_t *ctx, void *arg, int count, struct pt_regs *regs) |
| 2861 | { | 2863 | { |
| 2862 | struct thread_struct *thread = NULL; | ||
| 2863 | struct task_struct *task; | 2864 | struct task_struct *task; |
| 2864 | pfarg_reg_t *req = (pfarg_reg_t *)arg; | 2865 | pfarg_reg_t *req = (pfarg_reg_t *)arg; |
| 2865 | unsigned long value, pmc_pm; | 2866 | unsigned long value, pmc_pm; |
| @@ -2880,7 +2881,6 @@ pfm_write_pmcs(pfm_context_t *ctx, void *arg, int count, struct pt_regs *regs) | |||
| 2880 | if (state == PFM_CTX_ZOMBIE) return -EINVAL; | 2881 | if (state == PFM_CTX_ZOMBIE) return -EINVAL; |
| 2881 | 2882 | ||
| 2882 | if (is_loaded) { | 2883 | if (is_loaded) { |
| 2883 | thread = &task->thread; | ||
| 2884 | /* | 2884 | /* |
| 2885 | * In system wide and when the context is loaded, access can only happen | 2885 | * In system wide and when the context is loaded, access can only happen |
| 2886 | * when the caller is running on the CPU being monitored by the session. | 2886 | * when the caller is running on the CPU being monitored by the session. |
| @@ -3035,7 +3035,7 @@ pfm_write_pmcs(pfm_context_t *ctx, void *arg, int count, struct pt_regs *regs) | |||
| 3035 | * | 3035 | * |
| 3036 | * The value in ctx_pmcs[] can only be changed in pfm_write_pmcs(). | 3036 | * The value in ctx_pmcs[] can only be changed in pfm_write_pmcs(). |
| 3037 | * | 3037 | * |
| 3038 | * The value in thread->pmcs[] may be modified on overflow, i.e., when | 3038 | * The value in th_pmcs[] may be modified on overflow, i.e., when |
| 3039 | * monitoring needs to be stopped. | 3039 | * monitoring needs to be stopped. |
| 3040 | */ | 3040 | */ |
| 3041 | if (is_monitor) CTX_USED_MONITOR(ctx, 1UL << cnum); | 3041 | if (is_monitor) CTX_USED_MONITOR(ctx, 1UL << cnum); |
| @@ -3049,7 +3049,7 @@ pfm_write_pmcs(pfm_context_t *ctx, void *arg, int count, struct pt_regs *regs) | |||
| 3049 | /* | 3049 | /* |
| 3050 | * write thread state | 3050 | * write thread state |
| 3051 | */ | 3051 | */ |
| 3052 | if (is_system == 0) thread->pmcs[cnum] = value; | 3052 | if (is_system == 0) ctx->th_pmcs[cnum] = value; |
| 3053 | 3053 | ||
| 3054 | /* | 3054 | /* |
| 3055 | * write hardware register if we can | 3055 | * write hardware register if we can |
| @@ -3101,7 +3101,6 @@ error: | |||
| 3101 | static int | 3101 | static int |
| 3102 | pfm_write_pmds(pfm_context_t *ctx, void *arg, int count, struct pt_regs *regs) | 3102 | pfm_write_pmds(pfm_context_t *ctx, void *arg, int count, struct pt_regs *regs) |
| 3103 | { | 3103 | { |
| 3104 | struct thread_struct *thread = NULL; | ||
| 3105 | struct task_struct *task; | 3104 | struct task_struct *task; |
| 3106 | pfarg_reg_t *req = (pfarg_reg_t *)arg; | 3105 | pfarg_reg_t *req = (pfarg_reg_t *)arg; |
| 3107 | unsigned long value, hw_value, ovfl_mask; | 3106 | unsigned long value, hw_value, ovfl_mask; |
| @@ -3125,7 +3124,6 @@ pfm_write_pmds(pfm_context_t *ctx, void *arg, int count, struct pt_regs *regs) | |||
| 3125 | * the owner of the local PMU. | 3124 | * the owner of the local PMU. |
| 3126 | */ | 3125 | */ |
| 3127 | if (likely(is_loaded)) { | 3126 | if (likely(is_loaded)) { |
| 3128 | thread = &task->thread; | ||
| 3129 | /* | 3127 | /* |
| 3130 | * In system wide and when the context is loaded, access can only happen | 3128 | * In system wide and when the context is loaded, access can only happen |
| 3131 | * when the caller is running on the CPU being monitored by the session. | 3129 | * when the caller is running on the CPU being monitored by the session. |
| @@ -3233,7 +3231,7 @@ pfm_write_pmds(pfm_context_t *ctx, void *arg, int count, struct pt_regs *regs) | |||
| 3233 | /* | 3231 | /* |
| 3234 | * write thread state | 3232 | * write thread state |
| 3235 | */ | 3233 | */ |
| 3236 | if (is_system == 0) thread->pmds[cnum] = hw_value; | 3234 | if (is_system == 0) ctx->th_pmds[cnum] = hw_value; |
| 3237 | 3235 | ||
| 3238 | /* | 3236 | /* |
| 3239 | * write hardware register if we can | 3237 | * write hardware register if we can |
| @@ -3299,7 +3297,6 @@ abort_mission: | |||
| 3299 | static int | 3297 | static int |
| 3300 | pfm_read_pmds(pfm_context_t *ctx, void *arg, int count, struct pt_regs *regs) | 3298 | pfm_read_pmds(pfm_context_t *ctx, void *arg, int count, struct pt_regs *regs) |
| 3301 | { | 3299 | { |
| 3302 | struct thread_struct *thread = NULL; | ||
| 3303 | struct task_struct *task; | 3300 | struct task_struct *task; |
| 3304 | unsigned long val = 0UL, lval, ovfl_mask, sval; | 3301 | unsigned long val = 0UL, lval, ovfl_mask, sval; |
| 3305 | pfarg_reg_t *req = (pfarg_reg_t *)arg; | 3302 | pfarg_reg_t *req = (pfarg_reg_t *)arg; |
| @@ -3323,7 +3320,6 @@ pfm_read_pmds(pfm_context_t *ctx, void *arg, int count, struct pt_regs *regs) | |||
| 3323 | if (state == PFM_CTX_ZOMBIE) return -EINVAL; | 3320 | if (state == PFM_CTX_ZOMBIE) return -EINVAL; |
| 3324 | 3321 | ||
| 3325 | if (likely(is_loaded)) { | 3322 | if (likely(is_loaded)) { |
| 3326 | thread = &task->thread; | ||
| 3327 | /* | 3323 | /* |
| 3328 | * In system wide and when the context is loaded, access can only happen | 3324 | * In system wide and when the context is loaded, access can only happen |
| 3329 | * when the caller is running on the CPU being monitored by the session. | 3325 | * when the caller is running on the CPU being monitored by the session. |
| @@ -3385,7 +3381,7 @@ pfm_read_pmds(pfm_context_t *ctx, void *arg, int count, struct pt_regs *regs) | |||
| 3385 | * if context is zombie, then task does not exist anymore. | 3381 | * if context is zombie, then task does not exist anymore. |
| 3386 | * In this case, we use the full value saved in the context (pfm_flush_regs()). | 3382 | * In this case, we use the full value saved in the context (pfm_flush_regs()). |
| 3387 | */ | 3383 | */ |
| 3388 | val = is_loaded ? thread->pmds[cnum] : 0UL; | 3384 | val = is_loaded ? ctx->th_pmds[cnum] : 0UL; |
| 3389 | } | 3385 | } |
| 3390 | rd_func = pmu_conf->pmd_desc[cnum].read_check; | 3386 | rd_func = pmu_conf->pmd_desc[cnum].read_check; |
| 3391 | 3387 | ||
| @@ -4354,8 +4350,8 @@ pfm_context_load(pfm_context_t *ctx, void *arg, int count, struct pt_regs *regs) | |||
| 4354 | pfm_copy_pmds(task, ctx); | 4350 | pfm_copy_pmds(task, ctx); |
| 4355 | pfm_copy_pmcs(task, ctx); | 4351 | pfm_copy_pmcs(task, ctx); |
| 4356 | 4352 | ||
| 4357 | pmcs_source = thread->pmcs; | 4353 | pmcs_source = ctx->th_pmcs; |
| 4358 | pmds_source = thread->pmds; | 4354 | pmds_source = ctx->th_pmds; |
| 4359 | 4355 | ||
| 4360 | /* | 4356 | /* |
| 4361 | * always the case for system-wide | 4357 | * always the case for system-wide |
| @@ -5864,14 +5860,12 @@ void | |||
| 5864 | pfm_save_regs(struct task_struct *task) | 5860 | pfm_save_regs(struct task_struct *task) |
| 5865 | { | 5861 | { |
| 5866 | pfm_context_t *ctx; | 5862 | pfm_context_t *ctx; |
| 5867 | struct thread_struct *t; | ||
| 5868 | unsigned long flags; | 5863 | unsigned long flags; |
| 5869 | u64 psr; | 5864 | u64 psr; |
| 5870 | 5865 | ||
| 5871 | 5866 | ||
| 5872 | ctx = PFM_GET_CTX(task); | 5867 | ctx = PFM_GET_CTX(task); |
| 5873 | if (ctx == NULL) return; | 5868 | if (ctx == NULL) return; |
| 5874 | t = &task->thread; | ||
| 5875 | 5869 | ||
| 5876 | /* | 5870 | /* |
| 5877 | * we always come here with interrupts ALREADY disabled by | 5871 | * we always come here with interrupts ALREADY disabled by |
| @@ -5929,19 +5923,19 @@ pfm_save_regs(struct task_struct *task) | |||
| 5929 | * guarantee we will be schedule at that same | 5923 | * guarantee we will be schedule at that same |
| 5930 | * CPU again. | 5924 | * CPU again. |
| 5931 | */ | 5925 | */ |
| 5932 | pfm_save_pmds(t->pmds, ctx->ctx_used_pmds[0]); | 5926 | pfm_save_pmds(ctx->th_pmds, ctx->ctx_used_pmds[0]); |
| 5933 | 5927 | ||
| 5934 | /* | 5928 | /* |
| 5935 | * save pmc0 ia64_srlz_d() done in pfm_save_pmds() | 5929 | * save pmc0 ia64_srlz_d() done in pfm_save_pmds() |
| 5936 | * we will need it on the restore path to check | 5930 | * we will need it on the restore path to check |
| 5937 | * for pending overflow. | 5931 | * for pending overflow. |
| 5938 | */ | 5932 | */ |
| 5939 | t->pmcs[0] = ia64_get_pmc(0); | 5933 | ctx->th_pmcs[0] = ia64_get_pmc(0); |
| 5940 | 5934 | ||
| 5941 | /* | 5935 | /* |
| 5942 | * unfreeze PMU if had pending overflows | 5936 | * unfreeze PMU if had pending overflows |
| 5943 | */ | 5937 | */ |
| 5944 | if (t->pmcs[0] & ~0x1UL) pfm_unfreeze_pmu(); | 5938 | if (ctx->th_pmcs[0] & ~0x1UL) pfm_unfreeze_pmu(); |
| 5945 | 5939 | ||
| 5946 | /* | 5940 | /* |
| 5947 | * finally, allow context access. | 5941 | * finally, allow context access. |
| @@ -5986,7 +5980,6 @@ static void | |||
| 5986 | pfm_lazy_save_regs (struct task_struct *task) | 5980 | pfm_lazy_save_regs (struct task_struct *task) |
| 5987 | { | 5981 | { |
| 5988 | pfm_context_t *ctx; | 5982 | pfm_context_t *ctx; |
| 5989 | struct thread_struct *t; | ||
| 5990 | unsigned long flags; | 5983 | unsigned long flags; |
| 5991 | 5984 | ||
| 5992 | { u64 psr = pfm_get_psr(); | 5985 | { u64 psr = pfm_get_psr(); |
| @@ -5994,7 +5987,6 @@ pfm_lazy_save_regs (struct task_struct *task) | |||
| 5994 | } | 5987 | } |
| 5995 | 5988 | ||
| 5996 | ctx = PFM_GET_CTX(task); | 5989 | ctx = PFM_GET_CTX(task); |
| 5997 | t = &task->thread; | ||
| 5998 | 5990 | ||
| 5999 | /* | 5991 | /* |
| 6000 | * we need to mask PMU overflow here to | 5992 | * we need to mask PMU overflow here to |
| @@ -6019,19 +6011,19 @@ pfm_lazy_save_regs (struct task_struct *task) | |||
| 6019 | /* | 6011 | /* |
| 6020 | * save all the pmds we use | 6012 | * save all the pmds we use |
| 6021 | */ | 6013 | */ |
| 6022 | pfm_save_pmds(t->pmds, ctx->ctx_used_pmds[0]); | 6014 | pfm_save_pmds(ctx->th_pmds, ctx->ctx_used_pmds[0]); |
| 6023 | 6015 | ||
| 6024 | /* | 6016 | /* |
| 6025 | * save pmc0 ia64_srlz_d() done in pfm_save_pmds() | 6017 | * save pmc0 ia64_srlz_d() done in pfm_save_pmds() |
| 6026 | * it is needed to check for pended overflow | 6018 | * it is needed to check for pended overflow |
| 6027 | * on the restore path | 6019 | * on the restore path |
| 6028 | */ | 6020 | */ |
| 6029 | t->pmcs[0] = ia64_get_pmc(0); | 6021 | ctx->th_pmcs[0] = ia64_get_pmc(0); |
| 6030 | 6022 | ||
| 6031 | /* | 6023 | /* |
| 6032 | * unfreeze PMU if had pending overflows | 6024 | * unfreeze PMU if had pending overflows |
| 6033 | */ | 6025 | */ |
| 6034 | if (t->pmcs[0] & ~0x1UL) pfm_unfreeze_pmu(); | 6026 | if (ctx->th_pmcs[0] & ~0x1UL) pfm_unfreeze_pmu(); |
| 6035 | 6027 | ||
| 6036 | /* | 6028 | /* |
| 6037 | * now get can unmask PMU interrupts, they will | 6029 | * now get can unmask PMU interrupts, they will |
| @@ -6050,7 +6042,6 @@ void | |||
| 6050 | pfm_load_regs (struct task_struct *task) | 6042 | pfm_load_regs (struct task_struct *task) |
| 6051 | { | 6043 | { |
| 6052 | pfm_context_t *ctx; | 6044 | pfm_context_t *ctx; |
| 6053 | struct thread_struct *t; | ||
| 6054 | unsigned long pmc_mask = 0UL, pmd_mask = 0UL; | 6045 | unsigned long pmc_mask = 0UL, pmd_mask = 0UL; |
| 6055 | unsigned long flags; | 6046 | unsigned long flags; |
| 6056 | u64 psr, psr_up; | 6047 | u64 psr, psr_up; |
| @@ -6061,11 +6052,10 @@ pfm_load_regs (struct task_struct *task) | |||
| 6061 | 6052 | ||
| 6062 | BUG_ON(GET_PMU_OWNER()); | 6053 | BUG_ON(GET_PMU_OWNER()); |
| 6063 | 6054 | ||
| 6064 | t = &task->thread; | ||
| 6065 | /* | 6055 | /* |
| 6066 | * possible on unload | 6056 | * possible on unload |
| 6067 | */ | 6057 | */ |
| 6068 | if (unlikely((t->flags & IA64_THREAD_PM_VALID) == 0)) return; | 6058 | if (unlikely((task->thread.flags & IA64_THREAD_PM_VALID) == 0)) return; |
| 6069 | 6059 | ||
| 6070 | /* | 6060 | /* |
| 6071 | * we always come here with interrupts ALREADY disabled by | 6061 | * we always come here with interrupts ALREADY disabled by |
| @@ -6147,21 +6137,21 @@ pfm_load_regs (struct task_struct *task) | |||
| 6147 | * | 6137 | * |
| 6148 | * XXX: optimize here | 6138 | * XXX: optimize here |
| 6149 | */ | 6139 | */ |
| 6150 | if (pmd_mask) pfm_restore_pmds(t->pmds, pmd_mask); | 6140 | if (pmd_mask) pfm_restore_pmds(ctx->th_pmds, pmd_mask); |
| 6151 | if (pmc_mask) pfm_restore_pmcs(t->pmcs, pmc_mask); | 6141 | if (pmc_mask) pfm_restore_pmcs(ctx->th_pmcs, pmc_mask); |
| 6152 | 6142 | ||
| 6153 | /* | 6143 | /* |
| 6154 | * check for pending overflow at the time the state | 6144 | * check for pending overflow at the time the state |
| 6155 | * was saved. | 6145 | * was saved. |
| 6156 | */ | 6146 | */ |
| 6157 | if (unlikely(PMC0_HAS_OVFL(t->pmcs[0]))) { | 6147 | if (unlikely(PMC0_HAS_OVFL(ctx->th_pmcs[0]))) { |
| 6158 | /* | 6148 | /* |
| 6159 | * reload pmc0 with the overflow information | 6149 | * reload pmc0 with the overflow information |
| 6160 | * On McKinley PMU, this will trigger a PMU interrupt | 6150 | * On McKinley PMU, this will trigger a PMU interrupt |
| 6161 | */ | 6151 | */ |
| 6162 | ia64_set_pmc(0, t->pmcs[0]); | 6152 | ia64_set_pmc(0, ctx->th_pmcs[0]); |
| 6163 | ia64_srlz_d(); | 6153 | ia64_srlz_d(); |
| 6164 | t->pmcs[0] = 0UL; | 6154 | ctx->th_pmcs[0] = 0UL; |
| 6165 | 6155 | ||
| 6166 | /* | 6156 | /* |
| 6167 | * will replay the PMU interrupt | 6157 | * will replay the PMU interrupt |
| @@ -6214,7 +6204,6 @@ pfm_load_regs (struct task_struct *task) | |||
| 6214 | void | 6204 | void |
| 6215 | pfm_load_regs (struct task_struct *task) | 6205 | pfm_load_regs (struct task_struct *task) |
| 6216 | { | 6206 | { |
| 6217 | struct thread_struct *t; | ||
| 6218 | pfm_context_t *ctx; | 6207 | pfm_context_t *ctx; |
| 6219 | struct task_struct *owner; | 6208 | struct task_struct *owner; |
| 6220 | unsigned long pmd_mask, pmc_mask; | 6209 | unsigned long pmd_mask, pmc_mask; |
| @@ -6223,7 +6212,6 @@ pfm_load_regs (struct task_struct *task) | |||
| 6223 | 6212 | ||
| 6224 | owner = GET_PMU_OWNER(); | 6213 | owner = GET_PMU_OWNER(); |
| 6225 | ctx = PFM_GET_CTX(task); | 6214 | ctx = PFM_GET_CTX(task); |
| 6226 | t = &task->thread; | ||
| 6227 | psr = pfm_get_psr(); | 6215 | psr = pfm_get_psr(); |
| 6228 | 6216 | ||
| 6229 | BUG_ON(psr & (IA64_PSR_UP|IA64_PSR_PP)); | 6217 | BUG_ON(psr & (IA64_PSR_UP|IA64_PSR_PP)); |
| @@ -6286,22 +6274,22 @@ pfm_load_regs (struct task_struct *task) | |||
| 6286 | */ | 6274 | */ |
| 6287 | pmc_mask = ctx->ctx_all_pmcs[0]; | 6275 | pmc_mask = ctx->ctx_all_pmcs[0]; |
| 6288 | 6276 | ||
| 6289 | pfm_restore_pmds(t->pmds, pmd_mask); | 6277 | pfm_restore_pmds(ctx->th_pmds, pmd_mask); |
| 6290 | pfm_restore_pmcs(t->pmcs, pmc_mask); | 6278 | pfm_restore_pmcs(ctx->th_pmcs, pmc_mask); |
| 6291 | 6279 | ||
| 6292 | /* | 6280 | /* |
| 6293 | * check for pending overflow at the time the state | 6281 | * check for pending overflow at the time the state |
| 6294 | * was saved. | 6282 | * was saved. |
| 6295 | */ | 6283 | */ |
| 6296 | if (unlikely(PMC0_HAS_OVFL(t->pmcs[0]))) { | 6284 | if (unlikely(PMC0_HAS_OVFL(ctx->th_pmcs[0]))) { |
| 6297 | /* | 6285 | /* |
| 6298 | * reload pmc0 with the overflow information | 6286 | * reload pmc0 with the overflow information |
| 6299 | * On McKinley PMU, this will trigger a PMU interrupt | 6287 | * On McKinley PMU, this will trigger a PMU interrupt |
| 6300 | */ | 6288 | */ |
| 6301 | ia64_set_pmc(0, t->pmcs[0]); | 6289 | ia64_set_pmc(0, ctx->th_pmcs[0]); |
| 6302 | ia64_srlz_d(); | 6290 | ia64_srlz_d(); |
| 6303 | 6291 | ||
| 6304 | t->pmcs[0] = 0UL; | 6292 | ctx->th_pmcs[0] = 0UL; |
| 6305 | 6293 | ||
| 6306 | /* | 6294 | /* |
| 6307 | * will replay the PMU interrupt | 6295 | * will replay the PMU interrupt |
| @@ -6376,11 +6364,11 @@ pfm_flush_pmds(struct task_struct *task, pfm_context_t *ctx) | |||
| 6376 | */ | 6364 | */ |
| 6377 | pfm_unfreeze_pmu(); | 6365 | pfm_unfreeze_pmu(); |
| 6378 | } else { | 6366 | } else { |
| 6379 | pmc0 = task->thread.pmcs[0]; | 6367 | pmc0 = ctx->th_pmcs[0]; |
| 6380 | /* | 6368 | /* |
| 6381 | * clear whatever overflow status bits there were | 6369 | * clear whatever overflow status bits there were |
| 6382 | */ | 6370 | */ |
| 6383 | task->thread.pmcs[0] = 0; | 6371 | ctx->th_pmcs[0] = 0; |
| 6384 | } | 6372 | } |
| 6385 | ovfl_val = pmu_conf->ovfl_val; | 6373 | ovfl_val = pmu_conf->ovfl_val; |
| 6386 | /* | 6374 | /* |
| @@ -6401,7 +6389,7 @@ pfm_flush_pmds(struct task_struct *task, pfm_context_t *ctx) | |||
| 6401 | /* | 6389 | /* |
| 6402 | * can access PMU always true in system wide mode | 6390 | * can access PMU always true in system wide mode |
| 6403 | */ | 6391 | */ |
| 6404 | val = pmd_val = can_access_pmu ? ia64_get_pmd(i) : task->thread.pmds[i]; | 6392 | val = pmd_val = can_access_pmu ? ia64_get_pmd(i) : ctx->th_pmds[i]; |
| 6405 | 6393 | ||
| 6406 | if (PMD_IS_COUNTING(i)) { | 6394 | if (PMD_IS_COUNTING(i)) { |
| 6407 | DPRINT(("[%d] pmd[%d] ctx_pmd=0x%lx hw_pmd=0x%lx\n", | 6395 | DPRINT(("[%d] pmd[%d] ctx_pmd=0x%lx hw_pmd=0x%lx\n", |
| @@ -6433,7 +6421,7 @@ pfm_flush_pmds(struct task_struct *task, pfm_context_t *ctx) | |||
| 6433 | 6421 | ||
| 6434 | DPRINT(("[%d] ctx_pmd[%d]=0x%lx pmd_val=0x%lx\n", task->pid, i, val, pmd_val)); | 6422 | DPRINT(("[%d] ctx_pmd[%d]=0x%lx pmd_val=0x%lx\n", task->pid, i, val, pmd_val)); |
| 6435 | 6423 | ||
| 6436 | if (is_self) task->thread.pmds[i] = pmd_val; | 6424 | if (is_self) ctx->th_pmds[i] = pmd_val; |
| 6437 | 6425 | ||
| 6438 | ctx->ctx_pmds[i].val = val; | 6426 | ctx->ctx_pmds[i].val = val; |
| 6439 | } | 6427 | } |
| @@ -6677,7 +6665,7 @@ pfm_init(void) | |||
| 6677 | ffz(pmu_conf->ovfl_val)); | 6665 | ffz(pmu_conf->ovfl_val)); |
| 6678 | 6666 | ||
| 6679 | /* sanity check */ | 6667 | /* sanity check */ |
| 6680 | if (pmu_conf->num_pmds >= IA64_NUM_PMD_REGS || pmu_conf->num_pmcs >= IA64_NUM_PMC_REGS) { | 6668 | if (pmu_conf->num_pmds >= PFM_NUM_PMD_REGS || pmu_conf->num_pmcs >= PFM_NUM_PMC_REGS) { |
| 6681 | printk(KERN_ERR "perfmon: not enough pmc/pmd, perfmon disabled\n"); | 6669 | printk(KERN_ERR "perfmon: not enough pmc/pmd, perfmon disabled\n"); |
| 6682 | pmu_conf = NULL; | 6670 | pmu_conf = NULL; |
| 6683 | return -1; | 6671 | return -1; |
| @@ -6752,7 +6740,6 @@ void | |||
| 6752 | dump_pmu_state(const char *from) | 6740 | dump_pmu_state(const char *from) |
| 6753 | { | 6741 | { |
| 6754 | struct task_struct *task; | 6742 | struct task_struct *task; |
| 6755 | struct thread_struct *t; | ||
| 6756 | struct pt_regs *regs; | 6743 | struct pt_regs *regs; |
| 6757 | pfm_context_t *ctx; | 6744 | pfm_context_t *ctx; |
| 6758 | unsigned long psr, dcr, info, flags; | 6745 | unsigned long psr, dcr, info, flags; |
| @@ -6797,16 +6784,14 @@ dump_pmu_state(const char *from) | |||
| 6797 | ia64_psr(regs)->up = 0; | 6784 | ia64_psr(regs)->up = 0; |
| 6798 | ia64_psr(regs)->pp = 0; | 6785 | ia64_psr(regs)->pp = 0; |
| 6799 | 6786 | ||
| 6800 | t = ¤t->thread; | ||
| 6801 | |||
| 6802 | for (i=1; PMC_IS_LAST(i) == 0; i++) { | 6787 | for (i=1; PMC_IS_LAST(i) == 0; i++) { |
| 6803 | if (PMC_IS_IMPL(i) == 0) continue; | 6788 | if (PMC_IS_IMPL(i) == 0) continue; |
| 6804 | printk("->CPU%d pmc[%d]=0x%lx thread_pmc[%d]=0x%lx\n", this_cpu, i, ia64_get_pmc(i), i, t->pmcs[i]); | 6789 | printk("->CPU%d pmc[%d]=0x%lx thread_pmc[%d]=0x%lx\n", this_cpu, i, ia64_get_pmc(i), i, ctx->th_pmcs[i]); |
| 6805 | } | 6790 | } |
| 6806 | 6791 | ||
| 6807 | for (i=1; PMD_IS_LAST(i) == 0; i++) { | 6792 | for (i=1; PMD_IS_LAST(i) == 0; i++) { |
| 6808 | if (PMD_IS_IMPL(i) == 0) continue; | 6793 | if (PMD_IS_IMPL(i) == 0) continue; |
| 6809 | printk("->CPU%d pmd[%d]=0x%lx thread_pmd[%d]=0x%lx\n", this_cpu, i, ia64_get_pmd(i), i, t->pmds[i]); | 6794 | printk("->CPU%d pmd[%d]=0x%lx thread_pmd[%d]=0x%lx\n", this_cpu, i, ia64_get_pmd(i), i, ctx->th_pmds[i]); |
| 6810 | } | 6795 | } |
| 6811 | 6796 | ||
| 6812 | if (ctx) { | 6797 | if (ctx) { |
diff --git a/include/asm-ia64/processor.h b/include/asm-ia64/processor.h index ead68f8eb883..5830d36fd8e6 100644 --- a/include/asm-ia64/processor.h +++ b/include/asm-ia64/processor.h | |||
| @@ -20,12 +20,6 @@ | |||
| 20 | #include <asm/ustack.h> | 20 | #include <asm/ustack.h> |
| 21 | 21 | ||
| 22 | #define IA64_NUM_DBG_REGS 8 | 22 | #define IA64_NUM_DBG_REGS 8 |
| 23 | /* | ||
| 24 | * Limits for PMC and PMD are set to less than maximum architected values | ||
| 25 | * but should be sufficient for a while | ||
| 26 | */ | ||
| 27 | #define IA64_NUM_PMC_REGS 64 | ||
| 28 | #define IA64_NUM_PMD_REGS 64 | ||
| 29 | 23 | ||
| 30 | #define DEFAULT_MAP_BASE __IA64_UL_CONST(0x2000000000000000) | 24 | #define DEFAULT_MAP_BASE __IA64_UL_CONST(0x2000000000000000) |
| 31 | #define DEFAULT_TASK_SIZE __IA64_UL_CONST(0xa000000000000000) | 25 | #define DEFAULT_TASK_SIZE __IA64_UL_CONST(0xa000000000000000) |
| @@ -263,13 +257,9 @@ struct thread_struct { | |||
| 263 | # define INIT_THREAD_IA32 | 257 | # define INIT_THREAD_IA32 |
| 264 | #endif /* CONFIG_IA32_SUPPORT */ | 258 | #endif /* CONFIG_IA32_SUPPORT */ |
| 265 | #ifdef CONFIG_PERFMON | 259 | #ifdef CONFIG_PERFMON |
| 266 | __u64 pmcs[IA64_NUM_PMC_REGS]; | ||
| 267 | __u64 pmds[IA64_NUM_PMD_REGS]; | ||
| 268 | void *pfm_context; /* pointer to detailed PMU context */ | 260 | void *pfm_context; /* pointer to detailed PMU context */ |
| 269 | unsigned long pfm_needs_checking; /* when >0, pending perfmon work on kernel exit */ | 261 | unsigned long pfm_needs_checking; /* when >0, pending perfmon work on kernel exit */ |
| 270 | # define INIT_THREAD_PM .pmcs = {0UL, }, \ | 262 | # define INIT_THREAD_PM .pfm_context = NULL, \ |
| 271 | .pmds = {0UL, }, \ | ||
| 272 | .pfm_context = NULL, \ | ||
| 273 | .pfm_needs_checking = 0UL, | 263 | .pfm_needs_checking = 0UL, |
| 274 | #else | 264 | #else |
| 275 | # define INIT_THREAD_PM | 265 | # define INIT_THREAD_PM |
