diff options
| author | Linus Torvalds <torvalds@linux-foundation.org> | 2010-10-27 21:48:00 -0400 |
|---|---|---|
| committer | Linus Torvalds <torvalds@linux-foundation.org> | 2010-10-27 21:48:00 -0400 |
| commit | a042e26137d7674ac04b1cd2d5c06b9ebc1ee2d5 (patch) | |
| tree | c1a7a8bda41b99caa4b4a0fe320fc73278879f7d | |
| parent | f66dd539feb849a3a00f7fac67c026e0935e373a (diff) | |
| parent | e25804a0327dad954f7d43803178fdef2fd35b4e (diff) | |
Merge branch 'perf-fixes-for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/tip/linux-2.6-tip
* 'perf-fixes-for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/tip/linux-2.6-tip: (50 commits)
perf python scripting: Add futex-contention script
perf python scripting: Fixup cut'n'paste error in sctop script
perf scripting: Shut up 'perf record' final status
perf record: Remove newline character from perror() argument
perf python scripting: Support fedora 11 (audit 1.7.17)
perf python scripting: Improve the syscalls-by-pid script
perf python scripting: print the syscall name on sctop
perf python scripting: Improve the syscalls-counts script
perf python scripting: Improve the failed-syscalls-by-pid script
kprobes: Remove redundant text_mutex lock in optimize
x86/oprofile: Fix uninitialized variable use in debug printk
tracing: Fix 'faild' -> 'failed' typo
perf probe: Fix format specified for Dwarf_Off parameter
perf trace: Fix detection of script extension
perf trace: Use $PERF_EXEC_PATH in canned report scripts
perf tools: Document event modifiers
perf tools: Remove direct slang.h include
perf_events: Fix for transaction recovery in group_sched_in()
perf_events: Revert: Fix transaction recovery in group_sched_in()
perf, x86: Use NUMA aware allocations for PEBS/BTS/DS allocations
...
48 files changed, 1501 insertions, 670 deletions
diff --git a/arch/x86/include/asm/msr-index.h b/arch/x86/include/asm/msr-index.h index 83c4bb1d917d..3ea3dc487047 100644 --- a/arch/x86/include/asm/msr-index.h +++ b/arch/x86/include/asm/msr-index.h | |||
| @@ -121,6 +121,7 @@ | |||
| 121 | #define MSR_AMD64_IBSDCLINAD 0xc0011038 | 121 | #define MSR_AMD64_IBSDCLINAD 0xc0011038 |
| 122 | #define MSR_AMD64_IBSDCPHYSAD 0xc0011039 | 122 | #define MSR_AMD64_IBSDCPHYSAD 0xc0011039 |
| 123 | #define MSR_AMD64_IBSCTL 0xc001103a | 123 | #define MSR_AMD64_IBSCTL 0xc001103a |
| 124 | #define MSR_AMD64_IBSBRTARGET 0xc001103b | ||
| 124 | 125 | ||
| 125 | /* Fam 10h MSRs */ | 126 | /* Fam 10h MSRs */ |
| 126 | #define MSR_FAM10H_MMIO_CONF_BASE 0xc0010058 | 127 | #define MSR_FAM10H_MMIO_CONF_BASE 0xc0010058 |
diff --git a/arch/x86/include/asm/perf_event.h b/arch/x86/include/asm/perf_event.h index 6e742cc4251b..550e26b1dbb3 100644 --- a/arch/x86/include/asm/perf_event.h +++ b/arch/x86/include/asm/perf_event.h | |||
| @@ -111,17 +111,18 @@ union cpuid10_edx { | |||
| 111 | #define X86_PMC_IDX_FIXED_BTS (X86_PMC_IDX_FIXED + 16) | 111 | #define X86_PMC_IDX_FIXED_BTS (X86_PMC_IDX_FIXED + 16) |
| 112 | 112 | ||
| 113 | /* IbsFetchCtl bits/masks */ | 113 | /* IbsFetchCtl bits/masks */ |
| 114 | #define IBS_FETCH_RAND_EN (1ULL<<57) | 114 | #define IBS_FETCH_RAND_EN (1ULL<<57) |
| 115 | #define IBS_FETCH_VAL (1ULL<<49) | 115 | #define IBS_FETCH_VAL (1ULL<<49) |
| 116 | #define IBS_FETCH_ENABLE (1ULL<<48) | 116 | #define IBS_FETCH_ENABLE (1ULL<<48) |
| 117 | #define IBS_FETCH_CNT 0xFFFF0000ULL | 117 | #define IBS_FETCH_CNT 0xFFFF0000ULL |
| 118 | #define IBS_FETCH_MAX_CNT 0x0000FFFFULL | 118 | #define IBS_FETCH_MAX_CNT 0x0000FFFFULL |
| 119 | 119 | ||
| 120 | /* IbsOpCtl bits */ | 120 | /* IbsOpCtl bits */ |
| 121 | #define IBS_OP_CNT_CTL (1ULL<<19) | 121 | #define IBS_OP_CNT_CTL (1ULL<<19) |
| 122 | #define IBS_OP_VAL (1ULL<<18) | 122 | #define IBS_OP_VAL (1ULL<<18) |
| 123 | #define IBS_OP_ENABLE (1ULL<<17) | 123 | #define IBS_OP_ENABLE (1ULL<<17) |
| 124 | #define IBS_OP_MAX_CNT 0x0000FFFFULL | 124 | #define IBS_OP_MAX_CNT 0x0000FFFFULL |
| 125 | #define IBS_OP_MAX_CNT_EXT 0x007FFFFFULL /* not a register bit mask */ | ||
| 125 | 126 | ||
| 126 | #ifdef CONFIG_PERF_EVENTS | 127 | #ifdef CONFIG_PERF_EVENTS |
| 127 | extern void init_hw_perf_events(void); | 128 | extern void init_hw_perf_events(void); |
diff --git a/arch/x86/kernel/cpu/perf_event.c b/arch/x86/kernel/cpu/perf_event.c index c1e8c7a51164..ed6310183efb 100644 --- a/arch/x86/kernel/cpu/perf_event.c +++ b/arch/x86/kernel/cpu/perf_event.c | |||
| @@ -237,6 +237,7 @@ struct x86_pmu { | |||
| 237 | * Intel DebugStore bits | 237 | * Intel DebugStore bits |
| 238 | */ | 238 | */ |
| 239 | int bts, pebs; | 239 | int bts, pebs; |
| 240 | int bts_active, pebs_active; | ||
| 240 | int pebs_record_size; | 241 | int pebs_record_size; |
| 241 | void (*drain_pebs)(struct pt_regs *regs); | 242 | void (*drain_pebs)(struct pt_regs *regs); |
| 242 | struct event_constraint *pebs_constraints; | 243 | struct event_constraint *pebs_constraints; |
| @@ -380,7 +381,7 @@ static void release_pmc_hardware(void) {} | |||
| 380 | 381 | ||
| 381 | #endif | 382 | #endif |
| 382 | 383 | ||
| 383 | static int reserve_ds_buffers(void); | 384 | static void reserve_ds_buffers(void); |
| 384 | static void release_ds_buffers(void); | 385 | static void release_ds_buffers(void); |
| 385 | 386 | ||
| 386 | static void hw_perf_event_destroy(struct perf_event *event) | 387 | static void hw_perf_event_destroy(struct perf_event *event) |
| @@ -477,7 +478,7 @@ static int x86_setup_perfctr(struct perf_event *event) | |||
| 477 | if ((attr->config == PERF_COUNT_HW_BRANCH_INSTRUCTIONS) && | 478 | if ((attr->config == PERF_COUNT_HW_BRANCH_INSTRUCTIONS) && |
| 478 | (hwc->sample_period == 1)) { | 479 | (hwc->sample_period == 1)) { |
| 479 | /* BTS is not supported by this architecture. */ | 480 | /* BTS is not supported by this architecture. */ |
| 480 | if (!x86_pmu.bts) | 481 | if (!x86_pmu.bts_active) |
| 481 | return -EOPNOTSUPP; | 482 | return -EOPNOTSUPP; |
| 482 | 483 | ||
| 483 | /* BTS is currently only allowed for user-mode. */ | 484 | /* BTS is currently only allowed for user-mode. */ |
| @@ -496,12 +497,13 @@ static int x86_pmu_hw_config(struct perf_event *event) | |||
| 496 | int precise = 0; | 497 | int precise = 0; |
| 497 | 498 | ||
| 498 | /* Support for constant skid */ | 499 | /* Support for constant skid */ |
| 499 | if (x86_pmu.pebs) | 500 | if (x86_pmu.pebs_active) { |
| 500 | precise++; | 501 | precise++; |
| 501 | 502 | ||
| 502 | /* Support for IP fixup */ | 503 | /* Support for IP fixup */ |
| 503 | if (x86_pmu.lbr_nr) | 504 | if (x86_pmu.lbr_nr) |
| 504 | precise++; | 505 | precise++; |
| 506 | } | ||
| 505 | 507 | ||
| 506 | if (event->attr.precise_ip > precise) | 508 | if (event->attr.precise_ip > precise) |
| 507 | return -EOPNOTSUPP; | 509 | return -EOPNOTSUPP; |
| @@ -543,11 +545,8 @@ static int __x86_pmu_event_init(struct perf_event *event) | |||
| 543 | if (atomic_read(&active_events) == 0) { | 545 | if (atomic_read(&active_events) == 0) { |
| 544 | if (!reserve_pmc_hardware()) | 546 | if (!reserve_pmc_hardware()) |
| 545 | err = -EBUSY; | 547 | err = -EBUSY; |
| 546 | else { | 548 | else |
| 547 | err = reserve_ds_buffers(); | 549 | reserve_ds_buffers(); |
| 548 | if (err) | ||
| 549 | release_pmc_hardware(); | ||
| 550 | } | ||
| 551 | } | 550 | } |
| 552 | if (!err) | 551 | if (!err) |
| 553 | atomic_inc(&active_events); | 552 | atomic_inc(&active_events); |
diff --git a/arch/x86/kernel/cpu/perf_event_intel_ds.c b/arch/x86/kernel/cpu/perf_event_intel_ds.c index 4977f9c400e5..b7dcd9f2b8a0 100644 --- a/arch/x86/kernel/cpu/perf_event_intel_ds.c +++ b/arch/x86/kernel/cpu/perf_event_intel_ds.c | |||
| @@ -74,6 +74,107 @@ static void fini_debug_store_on_cpu(int cpu) | |||
| 74 | wrmsr_on_cpu(cpu, MSR_IA32_DS_AREA, 0, 0); | 74 | wrmsr_on_cpu(cpu, MSR_IA32_DS_AREA, 0, 0); |
| 75 | } | 75 | } |
| 76 | 76 | ||
| 77 | static int alloc_pebs_buffer(int cpu) | ||
| 78 | { | ||
| 79 | struct debug_store *ds = per_cpu(cpu_hw_events, cpu).ds; | ||
| 80 | int node = cpu_to_node(cpu); | ||
| 81 | int max, thresh = 1; /* always use a single PEBS record */ | ||
| 82 | void *buffer; | ||
| 83 | |||
| 84 | if (!x86_pmu.pebs) | ||
| 85 | return 0; | ||
| 86 | |||
| 87 | buffer = kmalloc_node(PEBS_BUFFER_SIZE, GFP_KERNEL | __GFP_ZERO, node); | ||
| 88 | if (unlikely(!buffer)) | ||
| 89 | return -ENOMEM; | ||
| 90 | |||
| 91 | max = PEBS_BUFFER_SIZE / x86_pmu.pebs_record_size; | ||
| 92 | |||
| 93 | ds->pebs_buffer_base = (u64)(unsigned long)buffer; | ||
| 94 | ds->pebs_index = ds->pebs_buffer_base; | ||
| 95 | ds->pebs_absolute_maximum = ds->pebs_buffer_base + | ||
| 96 | max * x86_pmu.pebs_record_size; | ||
| 97 | |||
| 98 | ds->pebs_interrupt_threshold = ds->pebs_buffer_base + | ||
| 99 | thresh * x86_pmu.pebs_record_size; | ||
| 100 | |||
| 101 | return 0; | ||
| 102 | } | ||
| 103 | |||
| 104 | static void release_pebs_buffer(int cpu) | ||
| 105 | { | ||
| 106 | struct debug_store *ds = per_cpu(cpu_hw_events, cpu).ds; | ||
| 107 | |||
| 108 | if (!ds || !x86_pmu.pebs) | ||
| 109 | return; | ||
| 110 | |||
| 111 | kfree((void *)(unsigned long)ds->pebs_buffer_base); | ||
| 112 | ds->pebs_buffer_base = 0; | ||
| 113 | } | ||
| 114 | |||
| 115 | static int alloc_bts_buffer(int cpu) | ||
| 116 | { | ||
| 117 | struct debug_store *ds = per_cpu(cpu_hw_events, cpu).ds; | ||
| 118 | int node = cpu_to_node(cpu); | ||
| 119 | int max, thresh; | ||
| 120 | void *buffer; | ||
| 121 | |||
| 122 | if (!x86_pmu.bts) | ||
| 123 | return 0; | ||
| 124 | |||
| 125 | buffer = kmalloc_node(BTS_BUFFER_SIZE, GFP_KERNEL | __GFP_ZERO, node); | ||
| 126 | if (unlikely(!buffer)) | ||
| 127 | return -ENOMEM; | ||
| 128 | |||
| 129 | max = BTS_BUFFER_SIZE / BTS_RECORD_SIZE; | ||
| 130 | thresh = max / 16; | ||
| 131 | |||
| 132 | ds->bts_buffer_base = (u64)(unsigned long)buffer; | ||
| 133 | ds->bts_index = ds->bts_buffer_base; | ||
| 134 | ds->bts_absolute_maximum = ds->bts_buffer_base + | ||
| 135 | max * BTS_RECORD_SIZE; | ||
| 136 | ds->bts_interrupt_threshold = ds->bts_absolute_maximum - | ||
| 137 | thresh * BTS_RECORD_SIZE; | ||
| 138 | |||
| 139 | return 0; | ||
| 140 | } | ||
| 141 | |||
| 142 | static void release_bts_buffer(int cpu) | ||
| 143 | { | ||
| 144 | struct debug_store *ds = per_cpu(cpu_hw_events, cpu).ds; | ||
| 145 | |||
| 146 | if (!ds || !x86_pmu.bts) | ||
| 147 | return; | ||
| 148 | |||
| 149 | kfree((void *)(unsigned long)ds->bts_buffer_base); | ||
| 150 | ds->bts_buffer_base = 0; | ||
| 151 | } | ||
| 152 | |||
| 153 | static int alloc_ds_buffer(int cpu) | ||
| 154 | { | ||
| 155 | int node = cpu_to_node(cpu); | ||
| 156 | struct debug_store *ds; | ||
| 157 | |||
| 158 | ds = kmalloc_node(sizeof(*ds), GFP_KERNEL | __GFP_ZERO, node); | ||
| 159 | if (unlikely(!ds)) | ||
| 160 | return -ENOMEM; | ||
| 161 | |||
| 162 | per_cpu(cpu_hw_events, cpu).ds = ds; | ||
| 163 | |||
| 164 | return 0; | ||
| 165 | } | ||
| 166 | |||
| 167 | static void release_ds_buffer(int cpu) | ||
| 168 | { | ||
| 169 | struct debug_store *ds = per_cpu(cpu_hw_events, cpu).ds; | ||
| 170 | |||
| 171 | if (!ds) | ||
| 172 | return; | ||
| 173 | |||
| 174 | per_cpu(cpu_hw_events, cpu).ds = NULL; | ||
| 175 | kfree(ds); | ||
| 176 | } | ||
| 177 | |||
| 77 | static void release_ds_buffers(void) | 178 | static void release_ds_buffers(void) |
| 78 | { | 179 | { |
| 79 | int cpu; | 180 | int cpu; |
| @@ -82,93 +183,77 @@ static void release_ds_buffers(void) | |||
| 82 | return; | 183 | return; |
| 83 | 184 | ||
| 84 | get_online_cpus(); | 185 | get_online_cpus(); |
| 85 | |||
| 86 | for_each_online_cpu(cpu) | 186 | for_each_online_cpu(cpu) |
| 87 | fini_debug_store_on_cpu(cpu); | 187 | fini_debug_store_on_cpu(cpu); |
| 88 | 188 | ||
| 89 | for_each_possible_cpu(cpu) { | 189 | for_each_possible_cpu(cpu) { |
| 90 | struct debug_store *ds = per_cpu(cpu_hw_events, cpu).ds; | 190 | release_pebs_buffer(cpu); |
| 91 | 191 | release_bts_buffer(cpu); | |
| 92 | if (!ds) | 192 | release_ds_buffer(cpu); |
| 93 | continue; | ||
| 94 | |||
| 95 | per_cpu(cpu_hw_events, cpu).ds = NULL; | ||
| 96 | |||
| 97 | kfree((void *)(unsigned long)ds->pebs_buffer_base); | ||
| 98 | kfree((void *)(unsigned long)ds->bts_buffer_base); | ||
| 99 | kfree(ds); | ||
| 100 | } | 193 | } |
| 101 | |||
| 102 | put_online_cpus(); | 194 | put_online_cpus(); |
| 103 | } | 195 | } |
| 104 | 196 | ||
| 105 | static int reserve_ds_buffers(void) | 197 | static void reserve_ds_buffers(void) |
| 106 | { | 198 | { |
| 107 | int cpu, err = 0; | 199 | int bts_err = 0, pebs_err = 0; |
| 200 | int cpu; | ||
| 201 | |||
| 202 | x86_pmu.bts_active = 0; | ||
| 203 | x86_pmu.pebs_active = 0; | ||
| 108 | 204 | ||
| 109 | if (!x86_pmu.bts && !x86_pmu.pebs) | 205 | if (!x86_pmu.bts && !x86_pmu.pebs) |
| 110 | return 0; | 206 | return; |
| 207 | |||
| 208 | if (!x86_pmu.bts) | ||
| 209 | bts_err = 1; | ||
| 210 | |||
| 211 | if (!x86_pmu.pebs) | ||
| 212 | pebs_err = 1; | ||
| 111 | 213 | ||
| 112 | get_online_cpus(); | 214 | get_online_cpus(); |
| 113 | 215 | ||
| 114 | for_each_possible_cpu(cpu) { | 216 | for_each_possible_cpu(cpu) { |
| 115 | struct debug_store *ds; | 217 | if (alloc_ds_buffer(cpu)) { |
| 116 | void *buffer; | 218 | bts_err = 1; |
| 117 | int max, thresh; | 219 | pebs_err = 1; |
| 220 | } | ||
| 221 | |||
| 222 | if (!bts_err && alloc_bts_buffer(cpu)) | ||
| 223 | bts_err = 1; | ||
| 118 | 224 | ||
| 119 | err = -ENOMEM; | 225 | if (!pebs_err && alloc_pebs_buffer(cpu)) |
| 120 | ds = kzalloc(sizeof(*ds), GFP_KERNEL); | 226 | pebs_err = 1; |
| 121 | if (unlikely(!ds)) | 227 | |
| 228 | if (bts_err && pebs_err) | ||
| 122 | break; | 229 | break; |
| 123 | per_cpu(cpu_hw_events, cpu).ds = ds; | 230 | } |
| 124 | |||
| 125 | if (x86_pmu.bts) { | ||
| 126 | buffer = kzalloc(BTS_BUFFER_SIZE, GFP_KERNEL); | ||
| 127 | if (unlikely(!buffer)) | ||
| 128 | break; | ||
| 129 | |||
| 130 | max = BTS_BUFFER_SIZE / BTS_RECORD_SIZE; | ||
| 131 | thresh = max / 16; | ||
| 132 | |||
| 133 | ds->bts_buffer_base = (u64)(unsigned long)buffer; | ||
| 134 | ds->bts_index = ds->bts_buffer_base; | ||
| 135 | ds->bts_absolute_maximum = ds->bts_buffer_base + | ||
| 136 | max * BTS_RECORD_SIZE; | ||
| 137 | ds->bts_interrupt_threshold = ds->bts_absolute_maximum - | ||
| 138 | thresh * BTS_RECORD_SIZE; | ||
| 139 | } | ||
| 140 | 231 | ||
| 141 | if (x86_pmu.pebs) { | 232 | if (bts_err) { |
| 142 | buffer = kzalloc(PEBS_BUFFER_SIZE, GFP_KERNEL); | 233 | for_each_possible_cpu(cpu) |
| 143 | if (unlikely(!buffer)) | 234 | release_bts_buffer(cpu); |
| 144 | break; | 235 | } |
| 145 | |||
| 146 | max = PEBS_BUFFER_SIZE / x86_pmu.pebs_record_size; | ||
| 147 | |||
| 148 | ds->pebs_buffer_base = (u64)(unsigned long)buffer; | ||
| 149 | ds->pebs_index = ds->pebs_buffer_base; | ||
| 150 | ds->pebs_absolute_maximum = ds->pebs_buffer_base + | ||
| 151 | max * x86_pmu.pebs_record_size; | ||
| 152 | /* | ||
| 153 | * Always use single record PEBS | ||
| 154 | */ | ||
| 155 | ds->pebs_interrupt_threshold = ds->pebs_buffer_base + | ||
| 156 | x86_pmu.pebs_record_size; | ||
| 157 | } | ||
| 158 | 236 | ||
| 159 | err = 0; | 237 | if (pebs_err) { |
| 238 | for_each_possible_cpu(cpu) | ||
| 239 | release_pebs_buffer(cpu); | ||
| 160 | } | 240 | } |
| 161 | 241 | ||
| 162 | if (err) | 242 | if (bts_err && pebs_err) { |
| 163 | release_ds_buffers(); | 243 | for_each_possible_cpu(cpu) |
| 164 | else { | 244 | release_ds_buffer(cpu); |
| 245 | } else { | ||
| 246 | if (x86_pmu.bts && !bts_err) | ||
| 247 | x86_pmu.bts_active = 1; | ||
| 248 | |||
| 249 | if (x86_pmu.pebs && !pebs_err) | ||
| 250 | x86_pmu.pebs_active = 1; | ||
| 251 | |||
| 165 | for_each_online_cpu(cpu) | 252 | for_each_online_cpu(cpu) |
| 166 | init_debug_store_on_cpu(cpu); | 253 | init_debug_store_on_cpu(cpu); |
| 167 | } | 254 | } |
| 168 | 255 | ||
| 169 | put_online_cpus(); | 256 | put_online_cpus(); |
| 170 | |||
| 171 | return err; | ||
| 172 | } | 257 | } |
| 173 | 258 | ||
| 174 | /* | 259 | /* |
| @@ -233,7 +318,7 @@ static int intel_pmu_drain_bts_buffer(void) | |||
| 233 | if (!event) | 318 | if (!event) |
| 234 | return 0; | 319 | return 0; |
| 235 | 320 | ||
| 236 | if (!ds) | 321 | if (!x86_pmu.bts_active) |
| 237 | return 0; | 322 | return 0; |
| 238 | 323 | ||
| 239 | at = (struct bts_record *)(unsigned long)ds->bts_buffer_base; | 324 | at = (struct bts_record *)(unsigned long)ds->bts_buffer_base; |
| @@ -503,7 +588,7 @@ static void intel_pmu_drain_pebs_core(struct pt_regs *iregs) | |||
| 503 | struct pebs_record_core *at, *top; | 588 | struct pebs_record_core *at, *top; |
| 504 | int n; | 589 | int n; |
| 505 | 590 | ||
| 506 | if (!ds || !x86_pmu.pebs) | 591 | if (!x86_pmu.pebs_active) |
| 507 | return; | 592 | return; |
| 508 | 593 | ||
| 509 | at = (struct pebs_record_core *)(unsigned long)ds->pebs_buffer_base; | 594 | at = (struct pebs_record_core *)(unsigned long)ds->pebs_buffer_base; |
| @@ -545,7 +630,7 @@ static void intel_pmu_drain_pebs_nhm(struct pt_regs *iregs) | |||
| 545 | u64 status = 0; | 630 | u64 status = 0; |
| 546 | int bit, n; | 631 | int bit, n; |
| 547 | 632 | ||
| 548 | if (!ds || !x86_pmu.pebs) | 633 | if (!x86_pmu.pebs_active) |
| 549 | return; | 634 | return; |
| 550 | 635 | ||
| 551 | at = (struct pebs_record_nhm *)(unsigned long)ds->pebs_buffer_base; | 636 | at = (struct pebs_record_nhm *)(unsigned long)ds->pebs_buffer_base; |
| @@ -630,9 +715,8 @@ static void intel_ds_init(void) | |||
| 630 | 715 | ||
| 631 | #else /* CONFIG_CPU_SUP_INTEL */ | 716 | #else /* CONFIG_CPU_SUP_INTEL */ |
| 632 | 717 | ||
| 633 | static int reserve_ds_buffers(void) | 718 | static void reserve_ds_buffers(void) |
| 634 | { | 719 | { |
| 635 | return 0; | ||
| 636 | } | 720 | } |
| 637 | 721 | ||
| 638 | static void release_ds_buffers(void) | 722 | static void release_ds_buffers(void) |
diff --git a/arch/x86/oprofile/nmi_int.c b/arch/x86/oprofile/nmi_int.c index bd1489c3ce09..4e8baad36d37 100644 --- a/arch/x86/oprofile/nmi_int.c +++ b/arch/x86/oprofile/nmi_int.c | |||
| @@ -726,6 +726,12 @@ int __init op_nmi_init(struct oprofile_operations *ops) | |||
| 726 | case 0x11: | 726 | case 0x11: |
| 727 | cpu_type = "x86-64/family11h"; | 727 | cpu_type = "x86-64/family11h"; |
| 728 | break; | 728 | break; |
| 729 | case 0x12: | ||
| 730 | cpu_type = "x86-64/family12h"; | ||
| 731 | break; | ||
| 732 | case 0x14: | ||
| 733 | cpu_type = "x86-64/family14h"; | ||
| 734 | break; | ||
| 729 | default: | 735 | default: |
| 730 | return -ENODEV; | 736 | return -ENODEV; |
| 731 | } | 737 | } |
diff --git a/arch/x86/oprofile/op_model_amd.c b/arch/x86/oprofile/op_model_amd.c index 42fb46f83883..a011bcc0f943 100644 --- a/arch/x86/oprofile/op_model_amd.c +++ b/arch/x86/oprofile/op_model_amd.c | |||
| @@ -48,17 +48,24 @@ static unsigned long reset_value[NUM_VIRT_COUNTERS]; | |||
| 48 | 48 | ||
| 49 | static u32 ibs_caps; | 49 | static u32 ibs_caps; |
| 50 | 50 | ||
| 51 | struct op_ibs_config { | 51 | struct ibs_config { |
| 52 | unsigned long op_enabled; | 52 | unsigned long op_enabled; |
| 53 | unsigned long fetch_enabled; | 53 | unsigned long fetch_enabled; |
| 54 | unsigned long max_cnt_fetch; | 54 | unsigned long max_cnt_fetch; |
| 55 | unsigned long max_cnt_op; | 55 | unsigned long max_cnt_op; |
| 56 | unsigned long rand_en; | 56 | unsigned long rand_en; |
| 57 | unsigned long dispatched_ops; | 57 | unsigned long dispatched_ops; |
| 58 | unsigned long branch_target; | ||
| 58 | }; | 59 | }; |
| 59 | 60 | ||
| 60 | static struct op_ibs_config ibs_config; | 61 | struct ibs_state { |
| 61 | static u64 ibs_op_ctl; | 62 | u64 ibs_op_ctl; |
| 63 | int branch_target; | ||
| 64 | unsigned long sample_size; | ||
| 65 | }; | ||
| 66 | |||
| 67 | static struct ibs_config ibs_config; | ||
| 68 | static struct ibs_state ibs_state; | ||
| 62 | 69 | ||
| 63 | /* | 70 | /* |
| 64 | * IBS cpuid feature detection | 71 | * IBS cpuid feature detection |
| @@ -71,8 +78,16 @@ static u64 ibs_op_ctl; | |||
| 71 | * bit 0 is used to indicate the existence of IBS. | 78 | * bit 0 is used to indicate the existence of IBS. |
| 72 | */ | 79 | */ |
| 73 | #define IBS_CAPS_AVAIL (1U<<0) | 80 | #define IBS_CAPS_AVAIL (1U<<0) |
| 81 | #define IBS_CAPS_FETCHSAM (1U<<1) | ||
| 82 | #define IBS_CAPS_OPSAM (1U<<2) | ||
| 74 | #define IBS_CAPS_RDWROPCNT (1U<<3) | 83 | #define IBS_CAPS_RDWROPCNT (1U<<3) |
| 75 | #define IBS_CAPS_OPCNT (1U<<4) | 84 | #define IBS_CAPS_OPCNT (1U<<4) |
| 85 | #define IBS_CAPS_BRNTRGT (1U<<5) | ||
| 86 | #define IBS_CAPS_OPCNTEXT (1U<<6) | ||
| 87 | |||
| 88 | #define IBS_CAPS_DEFAULT (IBS_CAPS_AVAIL \ | ||
| 89 | | IBS_CAPS_FETCHSAM \ | ||
| 90 | | IBS_CAPS_OPSAM) | ||
| 76 | 91 | ||
| 77 | /* | 92 | /* |
| 78 | * IBS APIC setup | 93 | * IBS APIC setup |
| @@ -99,12 +114,12 @@ static u32 get_ibs_caps(void) | |||
| 99 | /* check IBS cpuid feature flags */ | 114 | /* check IBS cpuid feature flags */ |
| 100 | max_level = cpuid_eax(0x80000000); | 115 | max_level = cpuid_eax(0x80000000); |
| 101 | if (max_level < IBS_CPUID_FEATURES) | 116 | if (max_level < IBS_CPUID_FEATURES) |
| 102 | return IBS_CAPS_AVAIL; | 117 | return IBS_CAPS_DEFAULT; |
| 103 | 118 | ||
| 104 | ibs_caps = cpuid_eax(IBS_CPUID_FEATURES); | 119 | ibs_caps = cpuid_eax(IBS_CPUID_FEATURES); |
| 105 | if (!(ibs_caps & IBS_CAPS_AVAIL)) | 120 | if (!(ibs_caps & IBS_CAPS_AVAIL)) |
| 106 | /* cpuid flags not valid */ | 121 | /* cpuid flags not valid */ |
| 107 | return IBS_CAPS_AVAIL; | 122 | return IBS_CAPS_DEFAULT; |
| 108 | 123 | ||
| 109 | return ibs_caps; | 124 | return ibs_caps; |
| 110 | } | 125 | } |
| @@ -197,8 +212,8 @@ op_amd_handle_ibs(struct pt_regs * const regs, | |||
| 197 | rdmsrl(MSR_AMD64_IBSOPCTL, ctl); | 212 | rdmsrl(MSR_AMD64_IBSOPCTL, ctl); |
| 198 | if (ctl & IBS_OP_VAL) { | 213 | if (ctl & IBS_OP_VAL) { |
| 199 | rdmsrl(MSR_AMD64_IBSOPRIP, val); | 214 | rdmsrl(MSR_AMD64_IBSOPRIP, val); |
| 200 | oprofile_write_reserve(&entry, regs, val, | 215 | oprofile_write_reserve(&entry, regs, val, IBS_OP_CODE, |
| 201 | IBS_OP_CODE, IBS_OP_SIZE); | 216 | ibs_state.sample_size); |
| 202 | oprofile_add_data64(&entry, val); | 217 | oprofile_add_data64(&entry, val); |
| 203 | rdmsrl(MSR_AMD64_IBSOPDATA, val); | 218 | rdmsrl(MSR_AMD64_IBSOPDATA, val); |
| 204 | oprofile_add_data64(&entry, val); | 219 | oprofile_add_data64(&entry, val); |
| @@ -210,10 +225,14 @@ op_amd_handle_ibs(struct pt_regs * const regs, | |||
| 210 | oprofile_add_data64(&entry, val); | 225 | oprofile_add_data64(&entry, val); |
| 211 | rdmsrl(MSR_AMD64_IBSDCPHYSAD, val); | 226 | rdmsrl(MSR_AMD64_IBSDCPHYSAD, val); |
| 212 | oprofile_add_data64(&entry, val); | 227 | oprofile_add_data64(&entry, val); |
| 228 | if (ibs_state.branch_target) { | ||
| 229 | rdmsrl(MSR_AMD64_IBSBRTARGET, val); | ||
| 230 | oprofile_add_data(&entry, (unsigned long)val); | ||
| 231 | } | ||
| 213 | oprofile_write_commit(&entry); | 232 | oprofile_write_commit(&entry); |
| 214 | 233 | ||
| 215 | /* reenable the IRQ */ | 234 | /* reenable the IRQ */ |
| 216 | ctl = op_amd_randomize_ibs_op(ibs_op_ctl); | 235 | ctl = op_amd_randomize_ibs_op(ibs_state.ibs_op_ctl); |
| 217 | wrmsrl(MSR_AMD64_IBSOPCTL, ctl); | 236 | wrmsrl(MSR_AMD64_IBSOPCTL, ctl); |
| 218 | } | 237 | } |
| 219 | } | 238 | } |
| @@ -226,21 +245,32 @@ static inline void op_amd_start_ibs(void) | |||
| 226 | if (!ibs_caps) | 245 | if (!ibs_caps) |
| 227 | return; | 246 | return; |
| 228 | 247 | ||
| 248 | memset(&ibs_state, 0, sizeof(ibs_state)); | ||
| 249 | |||
| 250 | /* | ||
| 251 | * Note: Since the max count settings may out of range we | ||
| 252 | * write back the actual used values so that userland can read | ||
| 253 | * it. | ||
| 254 | */ | ||
| 255 | |||
| 229 | if (ibs_config.fetch_enabled) { | 256 | if (ibs_config.fetch_enabled) { |
| 230 | val = (ibs_config.max_cnt_fetch >> 4) & IBS_FETCH_MAX_CNT; | 257 | val = ibs_config.max_cnt_fetch >> 4; |
| 258 | val = min(val, IBS_FETCH_MAX_CNT); | ||
| 259 | ibs_config.max_cnt_fetch = val << 4; | ||
| 231 | val |= ibs_config.rand_en ? IBS_FETCH_RAND_EN : 0; | 260 | val |= ibs_config.rand_en ? IBS_FETCH_RAND_EN : 0; |
| 232 | val |= IBS_FETCH_ENABLE; | 261 | val |= IBS_FETCH_ENABLE; |
| 233 | wrmsrl(MSR_AMD64_IBSFETCHCTL, val); | 262 | wrmsrl(MSR_AMD64_IBSFETCHCTL, val); |
| 234 | } | 263 | } |
| 235 | 264 | ||
| 236 | if (ibs_config.op_enabled) { | 265 | if (ibs_config.op_enabled) { |
| 237 | ibs_op_ctl = ibs_config.max_cnt_op >> 4; | 266 | val = ibs_config.max_cnt_op >> 4; |
| 238 | if (!(ibs_caps & IBS_CAPS_RDWROPCNT)) { | 267 | if (!(ibs_caps & IBS_CAPS_RDWROPCNT)) { |
| 239 | /* | 268 | /* |
| 240 | * IbsOpCurCnt not supported. See | 269 | * IbsOpCurCnt not supported. See |
| 241 | * op_amd_randomize_ibs_op() for details. | 270 | * op_amd_randomize_ibs_op() for details. |
| 242 | */ | 271 | */ |
| 243 | ibs_op_ctl = clamp(ibs_op_ctl, 0x0081ULL, 0xFF80ULL); | 272 | val = clamp(val, 0x0081ULL, 0xFF80ULL); |
| 273 | ibs_config.max_cnt_op = val << 4; | ||
| 244 | } else { | 274 | } else { |
| 245 | /* | 275 | /* |
| 246 | * The start value is randomized with a | 276 | * The start value is randomized with a |
| @@ -248,13 +278,24 @@ static inline void op_amd_start_ibs(void) | |||
| 248 | * with the half of the randomized range. Also | 278 | * with the half of the randomized range. Also |
| 249 | * avoid underflows. | 279 | * avoid underflows. |
| 250 | */ | 280 | */ |
| 251 | ibs_op_ctl = min(ibs_op_ctl + IBS_RANDOM_MAXCNT_OFFSET, | 281 | val += IBS_RANDOM_MAXCNT_OFFSET; |
| 252 | IBS_OP_MAX_CNT); | 282 | if (ibs_caps & IBS_CAPS_OPCNTEXT) |
| 283 | val = min(val, IBS_OP_MAX_CNT_EXT); | ||
| 284 | else | ||
| 285 | val = min(val, IBS_OP_MAX_CNT); | ||
| 286 | ibs_config.max_cnt_op = | ||
| 287 | (val - IBS_RANDOM_MAXCNT_OFFSET) << 4; | ||
| 288 | } | ||
| 289 | val = ((val & ~IBS_OP_MAX_CNT) << 4) | (val & IBS_OP_MAX_CNT); | ||
| 290 | val |= ibs_config.dispatched_ops ? IBS_OP_CNT_CTL : 0; | ||
| 291 | val |= IBS_OP_ENABLE; | ||
| 292 | ibs_state.ibs_op_ctl = val; | ||
| 293 | ibs_state.sample_size = IBS_OP_SIZE; | ||
| 294 | if (ibs_config.branch_target) { | ||
| 295 | ibs_state.branch_target = 1; | ||
| 296 | ibs_state.sample_size++; | ||
| 253 | } | 297 | } |
| 254 | if (ibs_caps & IBS_CAPS_OPCNT && ibs_config.dispatched_ops) | 298 | val = op_amd_randomize_ibs_op(ibs_state.ibs_op_ctl); |
| 255 | ibs_op_ctl |= IBS_OP_CNT_CTL; | ||
| 256 | ibs_op_ctl |= IBS_OP_ENABLE; | ||
| 257 | val = op_amd_randomize_ibs_op(ibs_op_ctl); | ||
| 258 | wrmsrl(MSR_AMD64_IBSOPCTL, val); | 299 | wrmsrl(MSR_AMD64_IBSOPCTL, val); |
| 259 | } | 300 | } |
| 260 | } | 301 | } |
| @@ -281,29 +322,25 @@ static inline int eilvt_is_available(int offset) | |||
| 281 | 322 | ||
| 282 | static inline int ibs_eilvt_valid(void) | 323 | static inline int ibs_eilvt_valid(void) |
| 283 | { | 324 | { |
| 284 | u64 val; | ||
| 285 | int offset; | 325 | int offset; |
| 326 | u64 val; | ||
| 286 | 327 | ||
| 287 | rdmsrl(MSR_AMD64_IBSCTL, val); | 328 | rdmsrl(MSR_AMD64_IBSCTL, val); |
| 329 | offset = val & IBSCTL_LVT_OFFSET_MASK; | ||
| 330 | |||
| 288 | if (!(val & IBSCTL_LVT_OFFSET_VALID)) { | 331 | if (!(val & IBSCTL_LVT_OFFSET_VALID)) { |
| 289 | pr_err(FW_BUG "cpu %d, invalid IBS " | 332 | pr_err(FW_BUG "cpu %d, invalid IBS interrupt offset %d (MSR%08X=0x%016llx)\n", |
| 290 | "interrupt offset %d (MSR%08X=0x%016llx)", | 333 | smp_processor_id(), offset, MSR_AMD64_IBSCTL, val); |
| 291 | smp_processor_id(), offset, | ||
| 292 | MSR_AMD64_IBSCTL, val); | ||
| 293 | return 0; | 334 | return 0; |
| 294 | } | 335 | } |
| 295 | 336 | ||
| 296 | offset = val & IBSCTL_LVT_OFFSET_MASK; | 337 | if (!eilvt_is_available(offset)) { |
| 297 | 338 | pr_err(FW_BUG "cpu %d, IBS interrupt offset %d not available (MSR%08X=0x%016llx)\n", | |
| 298 | if (eilvt_is_available(offset)) | 339 | smp_processor_id(), offset, MSR_AMD64_IBSCTL, val); |
| 299 | return !0; | 340 | return 0; |
| 300 | 341 | } | |
| 301 | pr_err(FW_BUG "cpu %d, IBS interrupt offset %d " | ||
| 302 | "not available (MSR%08X=0x%016llx)", | ||
| 303 | smp_processor_id(), offset, | ||
| 304 | MSR_AMD64_IBSCTL, val); | ||
| 305 | 342 | ||
| 306 | return 0; | 343 | return 1; |
| 307 | } | 344 | } |
| 308 | 345 | ||
| 309 | static inline int get_ibs_offset(void) | 346 | static inline int get_ibs_offset(void) |
| @@ -630,28 +667,33 @@ static int setup_ibs_files(struct super_block *sb, struct dentry *root) | |||
| 630 | /* model specific files */ | 667 | /* model specific files */ |
| 631 | 668 | ||
| 632 | /* setup some reasonable defaults */ | 669 | /* setup some reasonable defaults */ |
| 670 | memset(&ibs_config, 0, sizeof(ibs_config)); | ||
| 633 | ibs_config.max_cnt_fetch = 250000; | 671 | ibs_config.max_cnt_fetch = 250000; |
| 634 | ibs_config.fetch_enabled = 0; | ||
| 635 | ibs_config.max_cnt_op = 250000; | 672 | ibs_config.max_cnt_op = 250000; |
| 636 | ibs_config.op_enabled = 0; | 673 | |
| 637 | ibs_config.dispatched_ops = 0; | 674 | if (ibs_caps & IBS_CAPS_FETCHSAM) { |
| 638 | 675 | dir = oprofilefs_mkdir(sb, root, "ibs_fetch"); | |
| 639 | dir = oprofilefs_mkdir(sb, root, "ibs_fetch"); | 676 | oprofilefs_create_ulong(sb, dir, "enable", |
| 640 | oprofilefs_create_ulong(sb, dir, "enable", | 677 | &ibs_config.fetch_enabled); |
| 641 | &ibs_config.fetch_enabled); | 678 | oprofilefs_create_ulong(sb, dir, "max_count", |
| 642 | oprofilefs_create_ulong(sb, dir, "max_count", | 679 | &ibs_config.max_cnt_fetch); |
| 643 | &ibs_config.max_cnt_fetch); | 680 | oprofilefs_create_ulong(sb, dir, "rand_enable", |
| 644 | oprofilefs_create_ulong(sb, dir, "rand_enable", | 681 | &ibs_config.rand_en); |
| 645 | &ibs_config.rand_en); | 682 | } |
| 646 | 683 | ||
| 647 | dir = oprofilefs_mkdir(sb, root, "ibs_op"); | 684 | if (ibs_caps & IBS_CAPS_OPSAM) { |
| 648 | oprofilefs_create_ulong(sb, dir, "enable", | 685 | dir = oprofilefs_mkdir(sb, root, "ibs_op"); |
| 649 | &ibs_config.op_enabled); | 686 | oprofilefs_create_ulong(sb, dir, "enable", |
| 650 | oprofilefs_create_ulong(sb, dir, "max_count", | 687 | &ibs_config.op_enabled); |
| 651 | &ibs_config.max_cnt_op); | 688 | oprofilefs_create_ulong(sb, dir, "max_count", |
| 652 | if (ibs_caps & IBS_CAPS_OPCNT) | 689 | &ibs_config.max_cnt_op); |
| 653 | oprofilefs_create_ulong(sb, dir, "dispatched_ops", | 690 | if (ibs_caps & IBS_CAPS_OPCNT) |
| 654 | &ibs_config.dispatched_ops); | 691 | oprofilefs_create_ulong(sb, dir, "dispatched_ops", |
| 692 | &ibs_config.dispatched_ops); | ||
| 693 | if (ibs_caps & IBS_CAPS_BRNTRGT) | ||
| 694 | oprofilefs_create_ulong(sb, dir, "branch_target", | ||
| 695 | &ibs_config.branch_target); | ||
| 696 | } | ||
| 655 | 697 | ||
| 656 | return 0; | 698 | return 0; |
| 657 | } | 699 | } |
diff --git a/include/linux/interrupt.h b/include/linux/interrupt.h index 01b281646251..79d0c4f6d071 100644 --- a/include/linux/interrupt.h +++ b/include/linux/interrupt.h | |||
| @@ -410,7 +410,7 @@ extern void open_softirq(int nr, void (*action)(struct softirq_action *)); | |||
| 410 | extern void softirq_init(void); | 410 | extern void softirq_init(void); |
| 411 | static inline void __raise_softirq_irqoff(unsigned int nr) | 411 | static inline void __raise_softirq_irqoff(unsigned int nr) |
| 412 | { | 412 | { |
| 413 | trace_softirq_raise((struct softirq_action *)(unsigned long)nr, NULL); | 413 | trace_softirq_raise(nr); |
| 414 | or_softirq_pending(1UL << nr); | 414 | or_softirq_pending(1UL << nr); |
| 415 | } | 415 | } |
| 416 | 416 | ||
diff --git a/include/linux/ring_buffer.h b/include/linux/ring_buffer.h index 25b4f686d918..8d3a2486544d 100644 --- a/include/linux/ring_buffer.h +++ b/include/linux/ring_buffer.h | |||
| @@ -62,18 +62,6 @@ enum ring_buffer_type { | |||
| 62 | unsigned ring_buffer_event_length(struct ring_buffer_event *event); | 62 | unsigned ring_buffer_event_length(struct ring_buffer_event *event); |
| 63 | void *ring_buffer_event_data(struct ring_buffer_event *event); | 63 | void *ring_buffer_event_data(struct ring_buffer_event *event); |
| 64 | 64 | ||
| 65 | /** | ||
| 66 | * ring_buffer_event_time_delta - return the delta timestamp of the event | ||
| 67 | * @event: the event to get the delta timestamp of | ||
| 68 | * | ||
| 69 | * The delta timestamp is the 27 bit timestamp since the last event. | ||
| 70 | */ | ||
| 71 | static inline unsigned | ||
| 72 | ring_buffer_event_time_delta(struct ring_buffer_event *event) | ||
| 73 | { | ||
| 74 | return event->time_delta; | ||
| 75 | } | ||
| 76 | |||
| 77 | /* | 65 | /* |
| 78 | * ring_buffer_discard_commit will remove an event that has not | 66 | * ring_buffer_discard_commit will remove an event that has not |
| 79 | * ben committed yet. If this is used, then ring_buffer_unlock_commit | 67 | * ben committed yet. If this is used, then ring_buffer_unlock_commit |
diff --git a/include/trace/events/irq.h b/include/trace/events/irq.h index 6fa7cbab7d93..1c09820df585 100644 --- a/include/trace/events/irq.h +++ b/include/trace/events/irq.h | |||
| @@ -86,76 +86,62 @@ TRACE_EVENT(irq_handler_exit, | |||
| 86 | 86 | ||
| 87 | DECLARE_EVENT_CLASS(softirq, | 87 | DECLARE_EVENT_CLASS(softirq, |
| 88 | 88 | ||
| 89 | TP_PROTO(struct softirq_action *h, struct softirq_action *vec), | 89 | TP_PROTO(unsigned int vec_nr), |
| 90 | 90 | ||
| 91 | TP_ARGS(h, vec), | 91 | TP_ARGS(vec_nr), |
| 92 | 92 | ||
| 93 | TP_STRUCT__entry( | 93 | TP_STRUCT__entry( |
| 94 | __field( int, vec ) | 94 | __field( unsigned int, vec ) |
| 95 | ), | 95 | ), |
| 96 | 96 | ||
| 97 | TP_fast_assign( | 97 | TP_fast_assign( |
| 98 | if (vec) | 98 | __entry->vec = vec_nr; |
| 99 | __entry->vec = (int)(h - vec); | ||
| 100 | else | ||
| 101 | __entry->vec = (int)(long)h; | ||
| 102 | ), | 99 | ), |
| 103 | 100 | ||
| 104 | TP_printk("vec=%d [action=%s]", __entry->vec, | 101 | TP_printk("vec=%u [action=%s]", __entry->vec, |
| 105 | show_softirq_name(__entry->vec)) | 102 | show_softirq_name(__entry->vec)) |
| 106 | ); | 103 | ); |
| 107 | 104 | ||
| 108 | /** | 105 | /** |
| 109 | * softirq_entry - called immediately before the softirq handler | 106 | * softirq_entry - called immediately before the softirq handler |
| 110 | * @h: pointer to struct softirq_action | 107 | * @vec_nr: softirq vector number |
| 111 | * @vec: pointer to first struct softirq_action in softirq_vec array | ||
| 112 | * | 108 | * |
| 113 | * The @h parameter, contains a pointer to the struct softirq_action | 109 | * When used in combination with the softirq_exit tracepoint |
| 114 | * which has a pointer to the action handler that is called. By subtracting | 110 | * we can determine the softirq handler runtine. |
| 115 | * the @vec pointer from the @h pointer, we can determine the softirq | ||
| 116 | * number. Also, when used in combination with the softirq_exit tracepoint | ||
| 117 | * we can determine the softirq latency. | ||
| 118 | */ | 111 | */ |
| 119 | DEFINE_EVENT(softirq, softirq_entry, | 112 | DEFINE_EVENT(softirq, softirq_entry, |
| 120 | 113 | ||
| 121 | TP_PROTO(struct softirq_action *h, struct softirq_action *vec), | 114 | TP_PROTO(unsigned int vec_nr), |
| 122 | 115 | ||
| 123 | TP_ARGS(h, vec) | 116 | TP_ARGS(vec_nr) |
| 124 | ); | 117 | ); |
| 125 | 118 | ||
| 126 | /** | 119 | /** |
| 127 | * softirq_exit - called immediately after the softirq handler returns | 120 | * softirq_exit - called immediately after the softirq handler returns |
| 128 | * @h: pointer to struct softirq_action | 121 | * @vec_nr: softirq vector number |
| 129 | * @vec: pointer to first struct softirq_action in softirq_vec array | ||
| 130 | * | 122 | * |
| 131 | * The @h parameter contains a pointer to the struct softirq_action | 123 | * When used in combination with the softirq_entry tracepoint |
| 132 | * that has handled the softirq. By subtracting the @vec pointer from | 124 | * we can determine the softirq handler runtine. |
| 133 | * the @h pointer, we can determine the softirq number. Also, when used in | ||
| 134 | * combination with the softirq_entry tracepoint we can determine the softirq | ||
| 135 | * latency. | ||
| 136 | */ | 125 | */ |
| 137 | DEFINE_EVENT(softirq, softirq_exit, | 126 | DEFINE_EVENT(softirq, softirq_exit, |
| 138 | 127 | ||
| 139 | TP_PROTO(struct softirq_action *h, struct softirq_action *vec), | 128 | TP_PROTO(unsigned int vec_nr), |
| 140 | 129 | ||
| 141 | TP_ARGS(h, vec) | 130 | TP_ARGS(vec_nr) |
| 142 | ); | 131 | ); |
| 143 | 132 | ||
| 144 | /** | 133 | /** |
| 145 | * softirq_raise - called immediately when a softirq is raised | 134 | * softirq_raise - called immediately when a softirq is raised |
| 146 | * @h: pointer to struct softirq_action | 135 | * @vec_nr: softirq vector number |
| 147 | * @vec: pointer to first struct softirq_action in softirq_vec array | ||
| 148 | * | 136 | * |
| 149 | * The @h parameter contains a pointer to the softirq vector number which is | 137 | * When used in combination with the softirq_entry tracepoint |
| 150 | * raised. @vec is NULL and it means @h includes vector number not | 138 | * we can determine the softirq raise to run latency. |
| 151 | * softirq_action. When used in combination with the softirq_entry tracepoint | ||
| 152 | * we can determine the softirq raise latency. | ||
| 153 | */ | 139 | */ |
| 154 | DEFINE_EVENT(softirq, softirq_raise, | 140 | DEFINE_EVENT(softirq, softirq_raise, |
| 155 | 141 | ||
| 156 | TP_PROTO(struct softirq_action *h, struct softirq_action *vec), | 142 | TP_PROTO(unsigned int vec_nr), |
| 157 | 143 | ||
| 158 | TP_ARGS(h, vec) | 144 | TP_ARGS(vec_nr) |
| 159 | ); | 145 | ); |
| 160 | 146 | ||
| 161 | #endif /* _TRACE_IRQ_H */ | 147 | #endif /* _TRACE_IRQ_H */ |
diff --git a/kernel/kprobes.c b/kernel/kprobes.c index 56a891914273..99865c33a60d 100644 --- a/kernel/kprobes.c +++ b/kernel/kprobes.c | |||
| @@ -74,7 +74,8 @@ static struct hlist_head kretprobe_inst_table[KPROBE_TABLE_SIZE]; | |||
| 74 | /* NOTE: change this value only with kprobe_mutex held */ | 74 | /* NOTE: change this value only with kprobe_mutex held */ |
| 75 | static bool kprobes_all_disarmed; | 75 | static bool kprobes_all_disarmed; |
| 76 | 76 | ||
| 77 | static DEFINE_MUTEX(kprobe_mutex); /* Protects kprobe_table */ | 77 | /* This protects kprobe_table and optimizing_list */ |
| 78 | static DEFINE_MUTEX(kprobe_mutex); | ||
| 78 | static DEFINE_PER_CPU(struct kprobe *, kprobe_instance) = NULL; | 79 | static DEFINE_PER_CPU(struct kprobe *, kprobe_instance) = NULL; |
| 79 | static struct { | 80 | static struct { |
| 80 | spinlock_t lock ____cacheline_aligned_in_smp; | 81 | spinlock_t lock ____cacheline_aligned_in_smp; |
| @@ -595,6 +596,7 @@ static __kprobes void try_to_optimize_kprobe(struct kprobe *p) | |||
| 595 | } | 596 | } |
| 596 | 597 | ||
| 597 | #ifdef CONFIG_SYSCTL | 598 | #ifdef CONFIG_SYSCTL |
| 599 | /* This should be called with kprobe_mutex locked */ | ||
| 598 | static void __kprobes optimize_all_kprobes(void) | 600 | static void __kprobes optimize_all_kprobes(void) |
| 599 | { | 601 | { |
| 600 | struct hlist_head *head; | 602 | struct hlist_head *head; |
| @@ -607,17 +609,16 @@ static void __kprobes optimize_all_kprobes(void) | |||
| 607 | return; | 609 | return; |
| 608 | 610 | ||
| 609 | kprobes_allow_optimization = true; | 611 | kprobes_allow_optimization = true; |
| 610 | mutex_lock(&text_mutex); | ||
| 611 | for (i = 0; i < KPROBE_TABLE_SIZE; i++) { | 612 | for (i = 0; i < KPROBE_TABLE_SIZE; i++) { |
| 612 | head = &kprobe_table[i]; | 613 | head = &kprobe_table[i]; |
| 613 | hlist_for_each_entry_rcu(p, node, head, hlist) | 614 | hlist_for_each_entry_rcu(p, node, head, hlist) |
| 614 | if (!kprobe_disabled(p)) | 615 | if (!kprobe_disabled(p)) |
| 615 | optimize_kprobe(p); | 616 | optimize_kprobe(p); |
| 616 | } | 617 | } |
| 617 | mutex_unlock(&text_mutex); | ||
| 618 | printk(KERN_INFO "Kprobes globally optimized\n"); | 618 | printk(KERN_INFO "Kprobes globally optimized\n"); |
| 619 | } | 619 | } |
| 620 | 620 | ||
| 621 | /* This should be called with kprobe_mutex locked */ | ||
| 621 | static void __kprobes unoptimize_all_kprobes(void) | 622 | static void __kprobes unoptimize_all_kprobes(void) |
| 622 | { | 623 | { |
| 623 | struct hlist_head *head; | 624 | struct hlist_head *head; |
diff --git a/kernel/perf_event.c b/kernel/perf_event.c index f309e8014c78..517d827f4982 100644 --- a/kernel/perf_event.c +++ b/kernel/perf_event.c | |||
| @@ -417,8 +417,8 @@ event_filter_match(struct perf_event *event) | |||
| 417 | return event->cpu == -1 || event->cpu == smp_processor_id(); | 417 | return event->cpu == -1 || event->cpu == smp_processor_id(); |
| 418 | } | 418 | } |
| 419 | 419 | ||
| 420 | static int | 420 | static void |
| 421 | __event_sched_out(struct perf_event *event, | 421 | event_sched_out(struct perf_event *event, |
| 422 | struct perf_cpu_context *cpuctx, | 422 | struct perf_cpu_context *cpuctx, |
| 423 | struct perf_event_context *ctx) | 423 | struct perf_event_context *ctx) |
| 424 | { | 424 | { |
| @@ -437,13 +437,14 @@ __event_sched_out(struct perf_event *event, | |||
| 437 | } | 437 | } |
| 438 | 438 | ||
| 439 | if (event->state != PERF_EVENT_STATE_ACTIVE) | 439 | if (event->state != PERF_EVENT_STATE_ACTIVE) |
| 440 | return 0; | 440 | return; |
| 441 | 441 | ||
| 442 | event->state = PERF_EVENT_STATE_INACTIVE; | 442 | event->state = PERF_EVENT_STATE_INACTIVE; |
| 443 | if (event->pending_disable) { | 443 | if (event->pending_disable) { |
| 444 | event->pending_disable = 0; | 444 | event->pending_disable = 0; |
| 445 | event->state = PERF_EVENT_STATE_OFF; | 445 | event->state = PERF_EVENT_STATE_OFF; |
| 446 | } | 446 | } |
| 447 | event->tstamp_stopped = ctx->time; | ||
| 447 | event->pmu->del(event, 0); | 448 | event->pmu->del(event, 0); |
| 448 | event->oncpu = -1; | 449 | event->oncpu = -1; |
| 449 | 450 | ||
| @@ -452,19 +453,6 @@ __event_sched_out(struct perf_event *event, | |||
| 452 | ctx->nr_active--; | 453 | ctx->nr_active--; |
| 453 | if (event->attr.exclusive || !cpuctx->active_oncpu) | 454 | if (event->attr.exclusive || !cpuctx->active_oncpu) |
| 454 | cpuctx->exclusive = 0; | 455 | cpuctx->exclusive = 0; |
| 455 | return 1; | ||
| 456 | } | ||
| 457 | |||
| 458 | static void | ||
| 459 | event_sched_out(struct perf_event *event, | ||
| 460 | struct perf_cpu_context *cpuctx, | ||
| 461 | struct perf_event_context *ctx) | ||
| 462 | { | ||
| 463 | int ret; | ||
| 464 | |||
| 465 | ret = __event_sched_out(event, cpuctx, ctx); | ||
| 466 | if (ret) | ||
| 467 | event->tstamp_stopped = ctx->time; | ||
| 468 | } | 456 | } |
| 469 | 457 | ||
| 470 | static void | 458 | static void |
| @@ -664,7 +652,7 @@ retry: | |||
| 664 | } | 652 | } |
| 665 | 653 | ||
| 666 | static int | 654 | static int |
| 667 | __event_sched_in(struct perf_event *event, | 655 | event_sched_in(struct perf_event *event, |
| 668 | struct perf_cpu_context *cpuctx, | 656 | struct perf_cpu_context *cpuctx, |
| 669 | struct perf_event_context *ctx) | 657 | struct perf_event_context *ctx) |
| 670 | { | 658 | { |
| @@ -684,6 +672,8 @@ __event_sched_in(struct perf_event *event, | |||
| 684 | return -EAGAIN; | 672 | return -EAGAIN; |
| 685 | } | 673 | } |
| 686 | 674 | ||
| 675 | event->tstamp_running += ctx->time - event->tstamp_stopped; | ||
| 676 | |||
| 687 | if (!is_software_event(event)) | 677 | if (!is_software_event(event)) |
| 688 | cpuctx->active_oncpu++; | 678 | cpuctx->active_oncpu++; |
| 689 | ctx->nr_active++; | 679 | ctx->nr_active++; |
| @@ -694,35 +684,6 @@ __event_sched_in(struct perf_event *event, | |||
| 694 | return 0; | 684 | return 0; |
| 695 | } | 685 | } |
| 696 | 686 | ||
| 697 | static inline int | ||
| 698 | event_sched_in(struct perf_event *event, | ||
| 699 | struct perf_cpu_context *cpuctx, | ||
| 700 | struct perf_event_context *ctx) | ||
| 701 | { | ||
| 702 | int ret = __event_sched_in(event, cpuctx, ctx); | ||
| 703 | if (ret) | ||
| 704 | return ret; | ||
| 705 | event->tstamp_running += ctx->time - event->tstamp_stopped; | ||
| 706 | return 0; | ||
| 707 | } | ||
| 708 | |||
| 709 | static void | ||
| 710 | group_commit_event_sched_in(struct perf_event *group_event, | ||
| 711 | struct perf_cpu_context *cpuctx, | ||
| 712 | struct perf_event_context *ctx) | ||
| 713 | { | ||
| 714 | struct perf_event *event; | ||
| 715 | u64 now = ctx->time; | ||
| 716 | |||
| 717 | group_event->tstamp_running += now - group_event->tstamp_stopped; | ||
| 718 | /* | ||
| 719 | * Schedule in siblings as one group (if any): | ||
| 720 | */ | ||
| 721 | list_for_each_entry(event, &group_event->sibling_list, group_entry) { | ||
| 722 | event->tstamp_running += now - event->tstamp_stopped; | ||
| 723 | } | ||
| 724 | } | ||
| 725 | |||
| 726 | static int | 687 | static int |
| 727 | group_sched_in(struct perf_event *group_event, | 688 | group_sched_in(struct perf_event *group_event, |
| 728 | struct perf_cpu_context *cpuctx, | 689 | struct perf_cpu_context *cpuctx, |
| @@ -730,19 +691,15 @@ group_sched_in(struct perf_event *group_event, | |||
| 730 | { | 691 | { |
| 731 | struct perf_event *event, *partial_group = NULL; | 692 | struct perf_event *event, *partial_group = NULL; |
| 732 | struct pmu *pmu = group_event->pmu; | 693 | struct pmu *pmu = group_event->pmu; |
| 694 | u64 now = ctx->time; | ||
| 695 | bool simulate = false; | ||
| 733 | 696 | ||
| 734 | if (group_event->state == PERF_EVENT_STATE_OFF) | 697 | if (group_event->state == PERF_EVENT_STATE_OFF) |
| 735 | return 0; | 698 | return 0; |
| 736 | 699 | ||
| 737 | pmu->start_txn(pmu); | 700 | pmu->start_txn(pmu); |
| 738 | 701 | ||
| 739 | /* | 702 | if (event_sched_in(group_event, cpuctx, ctx)) { |
| 740 | * use __event_sched_in() to delay updating tstamp_running | ||
| 741 | * until the transaction is committed. In case of failure | ||
| 742 | * we will keep an unmodified tstamp_running which is a | ||
| 743 | * requirement to get correct timing information | ||
| 744 | */ | ||
| 745 | if (__event_sched_in(group_event, cpuctx, ctx)) { | ||
| 746 | pmu->cancel_txn(pmu); | 703 | pmu->cancel_txn(pmu); |
| 747 | return -EAGAIN; | 704 | return -EAGAIN; |
| 748 | } | 705 | } |
| @@ -751,31 +708,42 @@ group_sched_in(struct perf_event *group_event, | |||
| 751 | * Schedule in siblings as one group (if any): | 708 | * Schedule in siblings as one group (if any): |
| 752 | */ | 709 | */ |
| 753 | list_for_each_entry(event, &group_event->sibling_list, group_entry) { | 710 | list_for_each_entry(event, &group_event->sibling_list, group_entry) { |
| 754 | if (__event_sched_in(event, cpuctx, ctx)) { | 711 | if (event_sched_in(event, cpuctx, ctx)) { |
| 755 | partial_group = event; | 712 | partial_group = event; |
| 756 | goto group_error; | 713 | goto group_error; |
| 757 | } | 714 | } |
| 758 | } | 715 | } |
| 759 | 716 | ||
| 760 | if (!pmu->commit_txn(pmu)) { | 717 | if (!pmu->commit_txn(pmu)) |
| 761 | /* commit tstamp_running */ | ||
| 762 | group_commit_event_sched_in(group_event, cpuctx, ctx); | ||
| 763 | return 0; | 718 | return 0; |
| 764 | } | 719 | |
| 765 | group_error: | 720 | group_error: |
| 766 | /* | 721 | /* |
| 767 | * Groups can be scheduled in as one unit only, so undo any | 722 | * Groups can be scheduled in as one unit only, so undo any |
| 768 | * partial group before returning: | 723 | * partial group before returning: |
| 724 | * The events up to the failed event are scheduled out normally, | ||
| 725 | * tstamp_stopped will be updated. | ||
| 769 | * | 726 | * |
| 770 | * use __event_sched_out() to avoid updating tstamp_stopped | 727 | * The failed events and the remaining siblings need to have |
| 771 | * because the event never actually ran | 728 | * their timings updated as if they had gone thru event_sched_in() |
| 729 | * and event_sched_out(). This is required to get consistent timings | ||
| 730 | * across the group. This also takes care of the case where the group | ||
| 731 | * could never be scheduled by ensuring tstamp_stopped is set to mark | ||
| 732 | * the time the event was actually stopped, such that time delta | ||
| 733 | * calculation in update_event_times() is correct. | ||
| 772 | */ | 734 | */ |
| 773 | list_for_each_entry(event, &group_event->sibling_list, group_entry) { | 735 | list_for_each_entry(event, &group_event->sibling_list, group_entry) { |
| 774 | if (event == partial_group) | 736 | if (event == partial_group) |
| 775 | break; | 737 | simulate = true; |
| 776 | __event_sched_out(event, cpuctx, ctx); | 738 | |
| 739 | if (simulate) { | ||
| 740 | event->tstamp_running += now - event->tstamp_stopped; | ||
| 741 | event->tstamp_stopped = now; | ||
| 742 | } else { | ||
| 743 | event_sched_out(event, cpuctx, ctx); | ||
| 744 | } | ||
| 777 | } | 745 | } |
| 778 | __event_sched_out(group_event, cpuctx, ctx); | 746 | event_sched_out(group_event, cpuctx, ctx); |
| 779 | 747 | ||
| 780 | pmu->cancel_txn(pmu); | 748 | pmu->cancel_txn(pmu); |
| 781 | 749 | ||
diff --git a/kernel/softirq.c b/kernel/softirq.c index f02a9dfa19bc..18f4be0d5fe0 100644 --- a/kernel/softirq.c +++ b/kernel/softirq.c | |||
| @@ -229,18 +229,20 @@ restart: | |||
| 229 | 229 | ||
| 230 | do { | 230 | do { |
| 231 | if (pending & 1) { | 231 | if (pending & 1) { |
| 232 | unsigned int vec_nr = h - softirq_vec; | ||
| 232 | int prev_count = preempt_count(); | 233 | int prev_count = preempt_count(); |
| 233 | kstat_incr_softirqs_this_cpu(h - softirq_vec); | ||
| 234 | 234 | ||
| 235 | trace_softirq_entry(h, softirq_vec); | 235 | kstat_incr_softirqs_this_cpu(vec_nr); |
| 236 | |||
| 237 | trace_softirq_entry(vec_nr); | ||
| 236 | h->action(h); | 238 | h->action(h); |
| 237 | trace_softirq_exit(h, softirq_vec); | 239 | trace_softirq_exit(vec_nr); |
| 238 | if (unlikely(prev_count != preempt_count())) { | 240 | if (unlikely(prev_count != preempt_count())) { |
| 239 | printk(KERN_ERR "huh, entered softirq %td %s %p" | 241 | printk(KERN_ERR "huh, entered softirq %u %s %p" |
| 240 | "with preempt_count %08x," | 242 | "with preempt_count %08x," |
| 241 | " exited with %08x?\n", h - softirq_vec, | 243 | " exited with %08x?\n", vec_nr, |
| 242 | softirq_to_name[h - softirq_vec], | 244 | softirq_to_name[vec_nr], h->action, |
| 243 | h->action, prev_count, preempt_count()); | 245 | prev_count, preempt_count()); |
| 244 | preempt_count() = prev_count; | 246 | preempt_count() = prev_count; |
| 245 | } | 247 | } |
| 246 | 248 | ||
diff --git a/kernel/trace/ring_buffer.c b/kernel/trace/ring_buffer.c index c3dab054d18e..9ed509a015d8 100644 --- a/kernel/trace/ring_buffer.c +++ b/kernel/trace/ring_buffer.c | |||
| @@ -224,6 +224,9 @@ enum { | |||
| 224 | RB_LEN_TIME_STAMP = 16, | 224 | RB_LEN_TIME_STAMP = 16, |
| 225 | }; | 225 | }; |
| 226 | 226 | ||
| 227 | #define skip_time_extend(event) \ | ||
| 228 | ((struct ring_buffer_event *)((char *)event + RB_LEN_TIME_EXTEND)) | ||
| 229 | |||
| 227 | static inline int rb_null_event(struct ring_buffer_event *event) | 230 | static inline int rb_null_event(struct ring_buffer_event *event) |
| 228 | { | 231 | { |
| 229 | return event->type_len == RINGBUF_TYPE_PADDING && !event->time_delta; | 232 | return event->type_len == RINGBUF_TYPE_PADDING && !event->time_delta; |
| @@ -248,8 +251,12 @@ rb_event_data_length(struct ring_buffer_event *event) | |||
| 248 | return length + RB_EVNT_HDR_SIZE; | 251 | return length + RB_EVNT_HDR_SIZE; |
| 249 | } | 252 | } |
| 250 | 253 | ||
| 251 | /* inline for ring buffer fast paths */ | 254 | /* |
| 252 | static unsigned | 255 | * Return the length of the given event. Will return |
| 256 | * the length of the time extend if the event is a | ||
| 257 | * time extend. | ||
| 258 | */ | ||
| 259 | static inline unsigned | ||
| 253 | rb_event_length(struct ring_buffer_event *event) | 260 | rb_event_length(struct ring_buffer_event *event) |
| 254 | { | 261 | { |
| 255 | switch (event->type_len) { | 262 | switch (event->type_len) { |
| @@ -274,13 +281,41 @@ rb_event_length(struct ring_buffer_event *event) | |||
| 274 | return 0; | 281 | return 0; |
| 275 | } | 282 | } |
| 276 | 283 | ||
| 284 | /* | ||
| 285 | * Return total length of time extend and data, | ||
| 286 | * or just the event length for all other events. | ||
| 287 | */ | ||
| 288 | static inline unsigned | ||
| 289 | rb_event_ts_length(struct ring_buffer_event *event) | ||
| 290 | { | ||
| 291 | unsigned len = 0; | ||
| 292 | |||
| 293 | if (event->type_len == RINGBUF_TYPE_TIME_EXTEND) { | ||
| 294 | /* time extends include the data event after it */ | ||
| 295 | len = RB_LEN_TIME_EXTEND; | ||
| 296 | event = skip_time_extend(event); | ||
| 297 | } | ||
| 298 | return len + rb_event_length(event); | ||
| 299 | } | ||
| 300 | |||
| 277 | /** | 301 | /** |
| 278 | * ring_buffer_event_length - return the length of the event | 302 | * ring_buffer_event_length - return the length of the event |
| 279 | * @event: the event to get the length of | 303 | * @event: the event to get the length of |
| 304 | * | ||
| 305 | * Returns the size of the data load of a data event. | ||
| 306 | * If the event is something other than a data event, it | ||
| 307 | * returns the size of the event itself. With the exception | ||
| 308 | * of a TIME EXTEND, where it still returns the size of the | ||
| 309 | * data load of the data event after it. | ||
| 280 | */ | 310 | */ |
| 281 | unsigned ring_buffer_event_length(struct ring_buffer_event *event) | 311 | unsigned ring_buffer_event_length(struct ring_buffer_event *event) |
| 282 | { | 312 | { |
| 283 | unsigned length = rb_event_length(event); | 313 | unsigned length; |
| 314 | |||
| 315 | if (event->type_len == RINGBUF_TYPE_TIME_EXTEND) | ||
| 316 | event = skip_time_extend(event); | ||
| 317 | |||
| 318 | length = rb_event_length(event); | ||
| 284 | if (event->type_len > RINGBUF_TYPE_DATA_TYPE_LEN_MAX) | 319 | if (event->type_len > RINGBUF_TYPE_DATA_TYPE_LEN_MAX) |
| 285 | return length; | 320 | return length; |
| 286 | length -= RB_EVNT_HDR_SIZE; | 321 | length -= RB_EVNT_HDR_SIZE; |
| @@ -294,6 +329,8 @@ EXPORT_SYMBOL_GPL(ring_buffer_event_length); | |||
| 294 | static void * | 329 | static void * |
| 295 | rb_event_data(struct ring_buffer_event *event) | 330 | rb_event_data(struct ring_buffer_event *event) |
| 296 | { | 331 | { |
| 332 | if (event->type_len == RINGBUF_TYPE_TIME_EXTEND) | ||
| 333 | event = skip_time_extend(event); | ||
| 297 | BUG_ON(event->type_len > RINGBUF_TYPE_DATA_TYPE_LEN_MAX); | 334 | BUG_ON(event->type_len > RINGBUF_TYPE_DATA_TYPE_LEN_MAX); |
| 298 | /* If length is in len field, then array[0] has the data */ | 335 | /* If length is in len field, then array[0] has the data */ |
| 299 | if (event->type_len) | 336 | if (event->type_len) |
| @@ -404,9 +441,6 @@ static inline int test_time_stamp(u64 delta) | |||
| 404 | /* Max payload is BUF_PAGE_SIZE - header (8bytes) */ | 441 | /* Max payload is BUF_PAGE_SIZE - header (8bytes) */ |
| 405 | #define BUF_MAX_DATA_SIZE (BUF_PAGE_SIZE - (sizeof(u32) * 2)) | 442 | #define BUF_MAX_DATA_SIZE (BUF_PAGE_SIZE - (sizeof(u32) * 2)) |
| 406 | 443 | ||
| 407 | /* Max number of timestamps that can fit on a page */ | ||
| 408 | #define RB_TIMESTAMPS_PER_PAGE (BUF_PAGE_SIZE / RB_LEN_TIME_EXTEND) | ||
| 409 | |||
| 410 | int ring_buffer_print_page_header(struct trace_seq *s) | 444 | int ring_buffer_print_page_header(struct trace_seq *s) |
| 411 | { | 445 | { |
| 412 | struct buffer_data_page field; | 446 | struct buffer_data_page field; |
| @@ -1546,6 +1580,25 @@ static void rb_inc_iter(struct ring_buffer_iter *iter) | |||
| 1546 | iter->head = 0; | 1580 | iter->head = 0; |
| 1547 | } | 1581 | } |
| 1548 | 1582 | ||
| 1583 | /* Slow path, do not inline */ | ||
| 1584 | static noinline struct ring_buffer_event * | ||
| 1585 | rb_add_time_stamp(struct ring_buffer_event *event, u64 delta) | ||
| 1586 | { | ||
| 1587 | event->type_len = RINGBUF_TYPE_TIME_EXTEND; | ||
| 1588 | |||
| 1589 | /* Not the first event on the page? */ | ||
| 1590 | if (rb_event_index(event)) { | ||
| 1591 | event->time_delta = delta & TS_MASK; | ||
| 1592 | event->array[0] = delta >> TS_SHIFT; | ||
| 1593 | } else { | ||
| 1594 | /* nope, just zero it */ | ||
| 1595 | event->time_delta = 0; | ||
| 1596 | event->array[0] = 0; | ||
| 1597 | } | ||
| 1598 | |||
| 1599 | return skip_time_extend(event); | ||
| 1600 | } | ||
| 1601 | |||
| 1549 | /** | 1602 | /** |
| 1550 | * ring_buffer_update_event - update event type and data | 1603 | * ring_buffer_update_event - update event type and data |
| 1551 | * @event: the even to update | 1604 | * @event: the even to update |
| @@ -1558,28 +1611,31 @@ static void rb_inc_iter(struct ring_buffer_iter *iter) | |||
| 1558 | * data field. | 1611 | * data field. |
| 1559 | */ | 1612 | */ |
| 1560 | static void | 1613 | static void |
| 1561 | rb_update_event(struct ring_buffer_event *event, | 1614 | rb_update_event(struct ring_buffer_per_cpu *cpu_buffer, |
| 1562 | unsigned type, unsigned length) | 1615 | struct ring_buffer_event *event, unsigned length, |
| 1616 | int add_timestamp, u64 delta) | ||
| 1563 | { | 1617 | { |
| 1564 | event->type_len = type; | 1618 | /* Only a commit updates the timestamp */ |
| 1565 | 1619 | if (unlikely(!rb_event_is_commit(cpu_buffer, event))) | |
| 1566 | switch (type) { | 1620 | delta = 0; |
| 1567 | |||
| 1568 | case RINGBUF_TYPE_PADDING: | ||
| 1569 | case RINGBUF_TYPE_TIME_EXTEND: | ||
| 1570 | case RINGBUF_TYPE_TIME_STAMP: | ||
| 1571 | break; | ||
| 1572 | 1621 | ||
| 1573 | case 0: | 1622 | /* |
| 1574 | length -= RB_EVNT_HDR_SIZE; | 1623 | * If we need to add a timestamp, then we |
| 1575 | if (length > RB_MAX_SMALL_DATA || RB_FORCE_8BYTE_ALIGNMENT) | 1624 | * add it to the start of the resevered space. |
| 1576 | event->array[0] = length; | 1625 | */ |
| 1577 | else | 1626 | if (unlikely(add_timestamp)) { |
| 1578 | event->type_len = DIV_ROUND_UP(length, RB_ALIGNMENT); | 1627 | event = rb_add_time_stamp(event, delta); |
| 1579 | break; | 1628 | length -= RB_LEN_TIME_EXTEND; |
| 1580 | default: | 1629 | delta = 0; |
| 1581 | BUG(); | ||
| 1582 | } | 1630 | } |
| 1631 | |||
| 1632 | event->time_delta = delta; | ||
| 1633 | length -= RB_EVNT_HDR_SIZE; | ||
| 1634 | if (length > RB_MAX_SMALL_DATA || RB_FORCE_8BYTE_ALIGNMENT) { | ||
| 1635 | event->type_len = 0; | ||
| 1636 | event->array[0] = length; | ||
| 1637 | } else | ||
| 1638 | event->type_len = DIV_ROUND_UP(length, RB_ALIGNMENT); | ||
| 1583 | } | 1639 | } |
| 1584 | 1640 | ||
| 1585 | /* | 1641 | /* |
| @@ -1823,10 +1879,13 @@ rb_reset_tail(struct ring_buffer_per_cpu *cpu_buffer, | |||
| 1823 | local_sub(length, &tail_page->write); | 1879 | local_sub(length, &tail_page->write); |
| 1824 | } | 1880 | } |
| 1825 | 1881 | ||
| 1826 | static struct ring_buffer_event * | 1882 | /* |
| 1883 | * This is the slow path, force gcc not to inline it. | ||
| 1884 | */ | ||
| 1885 | static noinline struct ring_buffer_event * | ||
| 1827 | rb_move_tail(struct ring_buffer_per_cpu *cpu_buffer, | 1886 | rb_move_tail(struct ring_buffer_per_cpu *cpu_buffer, |
| 1828 | unsigned long length, unsigned long tail, | 1887 | unsigned long length, unsigned long tail, |
| 1829 | struct buffer_page *tail_page, u64 *ts) | 1888 | struct buffer_page *tail_page, u64 ts) |
| 1830 | { | 1889 | { |
| 1831 | struct buffer_page *commit_page = cpu_buffer->commit_page; | 1890 | struct buffer_page *commit_page = cpu_buffer->commit_page; |
| 1832 | struct ring_buffer *buffer = cpu_buffer->buffer; | 1891 | struct ring_buffer *buffer = cpu_buffer->buffer; |
| @@ -1909,8 +1968,8 @@ rb_move_tail(struct ring_buffer_per_cpu *cpu_buffer, | |||
| 1909 | * Nested commits always have zero deltas, so | 1968 | * Nested commits always have zero deltas, so |
| 1910 | * just reread the time stamp | 1969 | * just reread the time stamp |
| 1911 | */ | 1970 | */ |
| 1912 | *ts = rb_time_stamp(buffer); | 1971 | ts = rb_time_stamp(buffer); |
| 1913 | next_page->page->time_stamp = *ts; | 1972 | next_page->page->time_stamp = ts; |
| 1914 | } | 1973 | } |
| 1915 | 1974 | ||
| 1916 | out_again: | 1975 | out_again: |
| @@ -1929,12 +1988,21 @@ rb_move_tail(struct ring_buffer_per_cpu *cpu_buffer, | |||
| 1929 | 1988 | ||
| 1930 | static struct ring_buffer_event * | 1989 | static struct ring_buffer_event * |
| 1931 | __rb_reserve_next(struct ring_buffer_per_cpu *cpu_buffer, | 1990 | __rb_reserve_next(struct ring_buffer_per_cpu *cpu_buffer, |
| 1932 | unsigned type, unsigned long length, u64 *ts) | 1991 | unsigned long length, u64 ts, |
| 1992 | u64 delta, int add_timestamp) | ||
| 1933 | { | 1993 | { |
| 1934 | struct buffer_page *tail_page; | 1994 | struct buffer_page *tail_page; |
| 1935 | struct ring_buffer_event *event; | 1995 | struct ring_buffer_event *event; |
| 1936 | unsigned long tail, write; | 1996 | unsigned long tail, write; |
| 1937 | 1997 | ||
| 1998 | /* | ||
| 1999 | * If the time delta since the last event is too big to | ||
| 2000 | * hold in the time field of the event, then we append a | ||
| 2001 | * TIME EXTEND event ahead of the data event. | ||
| 2002 | */ | ||
| 2003 | if (unlikely(add_timestamp)) | ||
| 2004 | length += RB_LEN_TIME_EXTEND; | ||
| 2005 | |||
| 1938 | tail_page = cpu_buffer->tail_page; | 2006 | tail_page = cpu_buffer->tail_page; |
| 1939 | write = local_add_return(length, &tail_page->write); | 2007 | write = local_add_return(length, &tail_page->write); |
| 1940 | 2008 | ||
| @@ -1943,7 +2011,7 @@ __rb_reserve_next(struct ring_buffer_per_cpu *cpu_buffer, | |||
| 1943 | tail = write - length; | 2011 | tail = write - length; |
| 1944 | 2012 | ||
| 1945 | /* See if we shot pass the end of this buffer page */ | 2013 | /* See if we shot pass the end of this buffer page */ |
| 1946 | if (write > BUF_PAGE_SIZE) | 2014 | if (unlikely(write > BUF_PAGE_SIZE)) |
| 1947 | return rb_move_tail(cpu_buffer, length, tail, | 2015 | return rb_move_tail(cpu_buffer, length, tail, |
| 1948 | tail_page, ts); | 2016 | tail_page, ts); |
| 1949 | 2017 | ||
| @@ -1951,18 +2019,16 @@ __rb_reserve_next(struct ring_buffer_per_cpu *cpu_buffer, | |||
| 1951 | 2019 | ||
| 1952 | event = __rb_page_index(tail_page, tail); | 2020 | event = __rb_page_index(tail_page, tail); |
| 1953 | kmemcheck_annotate_bitfield(event, bitfield); | 2021 | kmemcheck_annotate_bitfield(event, bitfield); |
| 1954 | rb_update_event(event, type, length); | 2022 | rb_update_event(cpu_buffer, event, length, add_timestamp, delta); |
| 1955 | 2023 | ||
| 1956 | /* The passed in type is zero for DATA */ | 2024 | local_inc(&tail_page->entries); |
| 1957 | if (likely(!type)) | ||
| 1958 | local_inc(&tail_page->entries); | ||
| 1959 | 2025 | ||
| 1960 | /* | 2026 | /* |
| 1961 | * If this is the first commit on the page, then update | 2027 | * If this is the first commit on the page, then update |
| 1962 | * its timestamp. | 2028 | * its timestamp. |
| 1963 | */ | 2029 | */ |
| 1964 | if (!tail) | 2030 | if (!tail) |
| 1965 | tail_page->page->time_stamp = *ts; | 2031 | tail_page->page->time_stamp = ts; |
| 1966 | 2032 | ||
| 1967 | return event; | 2033 | return event; |
| 1968 | } | 2034 | } |
| @@ -1977,7 +2043,7 @@ rb_try_to_discard(struct ring_buffer_per_cpu *cpu_buffer, | |||
| 1977 | unsigned long addr; | 2043 | unsigned long addr; |
| 1978 | 2044 | ||
| 1979 | new_index = rb_event_index(event); | 2045 | new_index = rb_event_index(event); |
| 1980 | old_index = new_index + rb_event_length(event); | 2046 | old_index = new_index + rb_event_ts_length(event); |
| 1981 | addr = (unsigned long)event; | 2047 | addr = (unsigned long)event; |
| 1982 | addr &= PAGE_MASK; | 2048 | addr &= PAGE_MASK; |
| 1983 | 2049 | ||
| @@ -2003,76 +2069,13 @@ rb_try_to_discard(struct ring_buffer_per_cpu *cpu_buffer, | |||
| 2003 | return 0; | 2069 | return 0; |
| 2004 | } | 2070 | } |
| 2005 | 2071 | ||
| 2006 | static int | ||
| 2007 | rb_add_time_stamp(struct ring_buffer_per_cpu *cpu_buffer, | ||
| 2008 | u64 *ts, u64 *delta) | ||
| 2009 | { | ||
| 2010 | struct ring_buffer_event *event; | ||
| 2011 | int ret; | ||
| 2012 | |||
| 2013 | WARN_ONCE(*delta > (1ULL << 59), | ||
| 2014 | KERN_WARNING "Delta way too big! %llu ts=%llu write stamp = %llu\n", | ||
| 2015 | (unsigned long long)*delta, | ||
| 2016 | (unsigned long long)*ts, | ||
| 2017 | (unsigned long long)cpu_buffer->write_stamp); | ||
| 2018 | |||
| 2019 | /* | ||
| 2020 | * The delta is too big, we to add a | ||
| 2021 | * new timestamp. | ||
| 2022 | */ | ||
| 2023 | event = __rb_reserve_next(cpu_buffer, | ||
| 2024 | RINGBUF_TYPE_TIME_EXTEND, | ||
| 2025 | RB_LEN_TIME_EXTEND, | ||
| 2026 | ts); | ||
| 2027 | if (!event) | ||
| 2028 | return -EBUSY; | ||
| 2029 | |||
| 2030 | if (PTR_ERR(event) == -EAGAIN) | ||
| 2031 | return -EAGAIN; | ||
| 2032 | |||
| 2033 | /* Only a commited time event can update the write stamp */ | ||
| 2034 | if (rb_event_is_commit(cpu_buffer, event)) { | ||
| 2035 | /* | ||
| 2036 | * If this is the first on the page, then it was | ||
| 2037 | * updated with the page itself. Try to discard it | ||
| 2038 | * and if we can't just make it zero. | ||
| 2039 | */ | ||
| 2040 | if (rb_event_index(event)) { | ||
| 2041 | event->time_delta = *delta & TS_MASK; | ||
| 2042 | event->array[0] = *delta >> TS_SHIFT; | ||
| 2043 | } else { | ||
| 2044 | /* try to discard, since we do not need this */ | ||
| 2045 | if (!rb_try_to_discard(cpu_buffer, event)) { | ||
| 2046 | /* nope, just zero it */ | ||
| 2047 | event->time_delta = 0; | ||
| 2048 | event->array[0] = 0; | ||
| 2049 | } | ||
| 2050 | } | ||
| 2051 | cpu_buffer->write_stamp = *ts; | ||
| 2052 | /* let the caller know this was the commit */ | ||
| 2053 | ret = 1; | ||
| 2054 | } else { | ||
| 2055 | /* Try to discard the event */ | ||
| 2056 | if (!rb_try_to_discard(cpu_buffer, event)) { | ||
| 2057 | /* Darn, this is just wasted space */ | ||
| 2058 | event->time_delta = 0; | ||
| 2059 | event->array[0] = 0; | ||
| 2060 | } | ||
| 2061 | ret = 0; | ||
| 2062 | } | ||
| 2063 | |||
| 2064 | *delta = 0; | ||
| 2065 | |||
| 2066 | return ret; | ||
| 2067 | } | ||
| 2068 | |||
| 2069 | static void rb_start_commit(struct ring_buffer_per_cpu *cpu_buffer) | 2072 | static void rb_start_commit(struct ring_buffer_per_cpu *cpu_buffer) |
| 2070 | { | 2073 | { |
| 2071 | local_inc(&cpu_buffer->committing); | 2074 | local_inc(&cpu_buffer->committing); |
| 2072 | local_inc(&cpu_buffer->commits); | 2075 | local_inc(&cpu_buffer->commits); |
| 2073 | } | 2076 | } |
| 2074 | 2077 | ||
| 2075 | static void rb_end_commit(struct ring_buffer_per_cpu *cpu_buffer) | 2078 | static inline void rb_end_commit(struct ring_buffer_per_cpu *cpu_buffer) |
| 2076 | { | 2079 | { |
| 2077 | unsigned long commits; | 2080 | unsigned long commits; |
| 2078 | 2081 | ||
| @@ -2110,9 +2113,10 @@ rb_reserve_next_event(struct ring_buffer *buffer, | |||
| 2110 | unsigned long length) | 2113 | unsigned long length) |
| 2111 | { | 2114 | { |
| 2112 | struct ring_buffer_event *event; | 2115 | struct ring_buffer_event *event; |
| 2113 | u64 ts, delta = 0; | 2116 | u64 ts, delta; |
| 2114 | int commit = 0; | ||
| 2115 | int nr_loops = 0; | 2117 | int nr_loops = 0; |
| 2118 | int add_timestamp; | ||
| 2119 | u64 diff; | ||
| 2116 | 2120 | ||
| 2117 | rb_start_commit(cpu_buffer); | 2121 | rb_start_commit(cpu_buffer); |
| 2118 | 2122 | ||
| @@ -2133,6 +2137,9 @@ rb_reserve_next_event(struct ring_buffer *buffer, | |||
| 2133 | 2137 | ||
| 2134 | length = rb_calculate_event_length(length); | 2138 | length = rb_calculate_event_length(length); |
| 2135 | again: | 2139 | again: |
| 2140 | add_timestamp = 0; | ||
| 2141 | delta = 0; | ||
| 2142 | |||
| 2136 | /* | 2143 | /* |
| 2137 | * We allow for interrupts to reenter here and do a trace. | 2144 | * We allow for interrupts to reenter here and do a trace. |
| 2138 | * If one does, it will cause this original code to loop | 2145 | * If one does, it will cause this original code to loop |
| @@ -2146,56 +2153,32 @@ rb_reserve_next_event(struct ring_buffer *buffer, | |||
| 2146 | goto out_fail; | 2153 | goto out_fail; |
| 2147 | 2154 | ||
| 2148 | ts = rb_time_stamp(cpu_buffer->buffer); | 2155 | ts = rb_time_stamp(cpu_buffer->buffer); |
| 2156 | diff = ts - cpu_buffer->write_stamp; | ||
| 2149 | 2157 | ||
| 2150 | /* | 2158 | /* make sure this diff is calculated here */ |
| 2151 | * Only the first commit can update the timestamp. | 2159 | barrier(); |
| 2152 | * Yes there is a race here. If an interrupt comes in | ||
| 2153 | * just after the conditional and it traces too, then it | ||
| 2154 | * will also check the deltas. More than one timestamp may | ||
| 2155 | * also be made. But only the entry that did the actual | ||
| 2156 | * commit will be something other than zero. | ||
| 2157 | */ | ||
| 2158 | if (likely(cpu_buffer->tail_page == cpu_buffer->commit_page && | ||
| 2159 | rb_page_write(cpu_buffer->tail_page) == | ||
| 2160 | rb_commit_index(cpu_buffer))) { | ||
| 2161 | u64 diff; | ||
| 2162 | |||
| 2163 | diff = ts - cpu_buffer->write_stamp; | ||
| 2164 | |||
| 2165 | /* make sure this diff is calculated here */ | ||
| 2166 | barrier(); | ||
| 2167 | |||
| 2168 | /* Did the write stamp get updated already? */ | ||
| 2169 | if (unlikely(ts < cpu_buffer->write_stamp)) | ||
| 2170 | goto get_event; | ||
| 2171 | 2160 | ||
| 2161 | /* Did the write stamp get updated already? */ | ||
| 2162 | if (likely(ts >= cpu_buffer->write_stamp)) { | ||
| 2172 | delta = diff; | 2163 | delta = diff; |
| 2173 | if (unlikely(test_time_stamp(delta))) { | 2164 | if (unlikely(test_time_stamp(delta))) { |
| 2174 | 2165 | WARN_ONCE(delta > (1ULL << 59), | |
| 2175 | commit = rb_add_time_stamp(cpu_buffer, &ts, &delta); | 2166 | KERN_WARNING "Delta way too big! %llu ts=%llu write stamp = %llu\n", |
| 2176 | if (commit == -EBUSY) | 2167 | (unsigned long long)delta, |
| 2177 | goto out_fail; | 2168 | (unsigned long long)ts, |
| 2178 | 2169 | (unsigned long long)cpu_buffer->write_stamp); | |
| 2179 | if (commit == -EAGAIN) | 2170 | add_timestamp = 1; |
| 2180 | goto again; | ||
| 2181 | |||
| 2182 | RB_WARN_ON(cpu_buffer, commit < 0); | ||
| 2183 | } | 2171 | } |
| 2184 | } | 2172 | } |
| 2185 | 2173 | ||
| 2186 | get_event: | 2174 | event = __rb_reserve_next(cpu_buffer, length, ts, |
| 2187 | event = __rb_reserve_next(cpu_buffer, 0, length, &ts); | 2175 | delta, add_timestamp); |
| 2188 | if (unlikely(PTR_ERR(event) == -EAGAIN)) | 2176 | if (unlikely(PTR_ERR(event) == -EAGAIN)) |
| 2189 | goto again; | 2177 | goto again; |
| 2190 | 2178 | ||
| 2191 | if (!event) | 2179 | if (!event) |
| 2192 | goto out_fail; | 2180 | goto out_fail; |
| 2193 | 2181 | ||
| 2194 | if (!rb_event_is_commit(cpu_buffer, event)) | ||
| 2195 | delta = 0; | ||
| 2196 | |||
| 2197 | event->time_delta = delta; | ||
| 2198 | |||
| 2199 | return event; | 2182 | return event; |
| 2200 | 2183 | ||
| 2201 | out_fail: | 2184 | out_fail: |
| @@ -2207,13 +2190,9 @@ rb_reserve_next_event(struct ring_buffer *buffer, | |||
| 2207 | 2190 | ||
| 2208 | #define TRACE_RECURSIVE_DEPTH 16 | 2191 | #define TRACE_RECURSIVE_DEPTH 16 |
| 2209 | 2192 | ||
| 2210 | static int trace_recursive_lock(void) | 2193 | /* Keep this code out of the fast path cache */ |
| 2194 | static noinline void trace_recursive_fail(void) | ||
| 2211 | { | 2195 | { |
| 2212 | current->trace_recursion++; | ||
| 2213 | |||
| 2214 | if (likely(current->trace_recursion < TRACE_RECURSIVE_DEPTH)) | ||
| 2215 | return 0; | ||
| 2216 | |||
| 2217 | /* Disable all tracing before we do anything else */ | 2196 | /* Disable all tracing before we do anything else */ |
| 2218 | tracing_off_permanent(); | 2197 | tracing_off_permanent(); |
| 2219 | 2198 | ||
| @@ -2225,10 +2204,21 @@ static int trace_recursive_lock(void) | |||
| 2225 | in_nmi()); | 2204 | in_nmi()); |
| 2226 | 2205 | ||
| 2227 | WARN_ON_ONCE(1); | 2206 | WARN_ON_ONCE(1); |
| 2207 | } | ||
| 2208 | |||
| 2209 | static inline int trace_recursive_lock(void) | ||
| 2210 | { | ||
| 2211 | current->trace_recursion++; | ||
| 2212 | |||
| 2213 | if (likely(current->trace_recursion < TRACE_RECURSIVE_DEPTH)) | ||
| 2214 | return 0; | ||
| 2215 | |||
| 2216 | trace_recursive_fail(); | ||
| 2217 | |||
| 2228 | return -1; | 2218 | return -1; |
| 2229 | } | 2219 | } |
| 2230 | 2220 | ||
| 2231 | static void trace_recursive_unlock(void) | 2221 | static inline void trace_recursive_unlock(void) |
| 2232 | { | 2222 | { |
| 2233 | WARN_ON_ONCE(!current->trace_recursion); | 2223 | WARN_ON_ONCE(!current->trace_recursion); |
| 2234 | 2224 | ||
| @@ -2308,12 +2298,28 @@ static void | |||
| 2308 | rb_update_write_stamp(struct ring_buffer_per_cpu *cpu_buffer, | 2298 | rb_update_write_stamp(struct ring_buffer_per_cpu *cpu_buffer, |
| 2309 | struct ring_buffer_event *event) | 2299 | struct ring_buffer_event *event) |
| 2310 | { | 2300 | { |
| 2301 | u64 delta; | ||
| 2302 | |||
| 2311 | /* | 2303 | /* |
| 2312 | * The event first in the commit queue updates the | 2304 | * The event first in the commit queue updates the |
| 2313 | * time stamp. | 2305 | * time stamp. |
| 2314 | */ | 2306 | */ |
| 2315 | if (rb_event_is_commit(cpu_buffer, event)) | 2307 | if (rb_event_is_commit(cpu_buffer, event)) { |
| 2316 | cpu_buffer->write_stamp += event->time_delta; | 2308 | /* |
| 2309 | * A commit event that is first on a page | ||
| 2310 | * updates the write timestamp with the page stamp | ||
| 2311 | */ | ||
| 2312 | if (!rb_event_index(event)) | ||
| 2313 | cpu_buffer->write_stamp = | ||
| 2314 | cpu_buffer->commit_page->page->time_stamp; | ||
| 2315 | else if (event->type_len == RINGBUF_TYPE_TIME_EXTEND) { | ||
| 2316 | delta = event->array[0]; | ||
| 2317 | delta <<= TS_SHIFT; | ||
| 2318 | delta += event->time_delta; | ||
| 2319 | cpu_buffer->write_stamp += delta; | ||
| 2320 | } else | ||
| 2321 | cpu_buffer->write_stamp += event->time_delta; | ||
| 2322 | } | ||
| 2317 | } | 2323 | } |
| 2318 | 2324 | ||
| 2319 | static void rb_commit(struct ring_buffer_per_cpu *cpu_buffer, | 2325 | static void rb_commit(struct ring_buffer_per_cpu *cpu_buffer, |
| @@ -2353,6 +2359,9 @@ EXPORT_SYMBOL_GPL(ring_buffer_unlock_commit); | |||
| 2353 | 2359 | ||
| 2354 | static inline void rb_event_discard(struct ring_buffer_event *event) | 2360 | static inline void rb_event_discard(struct ring_buffer_event *event) |
| 2355 | { | 2361 | { |
| 2362 | if (event->type_len == RINGBUF_TYPE_TIME_EXTEND) | ||
| 2363 | event = skip_time_extend(event); | ||
| 2364 | |||
| 2356 | /* array[0] holds the actual length for the discarded event */ | 2365 | /* array[0] holds the actual length for the discarded event */ |
| 2357 | event->array[0] = rb_event_data_length(event) - RB_EVNT_HDR_SIZE; | 2366 | event->array[0] = rb_event_data_length(event) - RB_EVNT_HDR_SIZE; |
| 2358 | event->type_len = RINGBUF_TYPE_PADDING; | 2367 | event->type_len = RINGBUF_TYPE_PADDING; |
| @@ -3049,12 +3058,12 @@ rb_buffer_peek(struct ring_buffer_per_cpu *cpu_buffer, u64 *ts, | |||
| 3049 | 3058 | ||
| 3050 | again: | 3059 | again: |
| 3051 | /* | 3060 | /* |
| 3052 | * We repeat when a timestamp is encountered. It is possible | 3061 | * We repeat when a time extend is encountered. |
| 3053 | * to get multiple timestamps from an interrupt entering just | 3062 | * Since the time extend is always attached to a data event, |
| 3054 | * as one timestamp is about to be written, or from discarded | 3063 | * we should never loop more than once. |
| 3055 | * commits. The most that we can have is the number on a single page. | 3064 | * (We never hit the following condition more than twice). |
| 3056 | */ | 3065 | */ |
| 3057 | if (RB_WARN_ON(cpu_buffer, ++nr_loops > RB_TIMESTAMPS_PER_PAGE)) | 3066 | if (RB_WARN_ON(cpu_buffer, ++nr_loops > 2)) |
| 3058 | return NULL; | 3067 | return NULL; |
| 3059 | 3068 | ||
| 3060 | reader = rb_get_reader_page(cpu_buffer); | 3069 | reader = rb_get_reader_page(cpu_buffer); |
| @@ -3130,14 +3139,12 @@ rb_iter_peek(struct ring_buffer_iter *iter, u64 *ts) | |||
| 3130 | return NULL; | 3139 | return NULL; |
| 3131 | 3140 | ||
| 3132 | /* | 3141 | /* |
| 3133 | * We repeat when a timestamp is encountered. | 3142 | * We repeat when a time extend is encountered. |
| 3134 | * We can get multiple timestamps by nested interrupts or also | 3143 | * Since the time extend is always attached to a data event, |
| 3135 | * if filtering is on (discarding commits). Since discarding | 3144 | * we should never loop more than once. |
| 3136 | * commits can be frequent we can get a lot of timestamps. | 3145 | * (We never hit the following condition more than twice). |
| 3137 | * But we limit them by not adding timestamps if they begin | ||
| 3138 | * at the start of a page. | ||
| 3139 | */ | 3146 | */ |
| 3140 | if (RB_WARN_ON(cpu_buffer, ++nr_loops > RB_TIMESTAMPS_PER_PAGE)) | 3147 | if (RB_WARN_ON(cpu_buffer, ++nr_loops > 2)) |
| 3141 | return NULL; | 3148 | return NULL; |
| 3142 | 3149 | ||
| 3143 | if (rb_per_cpu_empty(cpu_buffer)) | 3150 | if (rb_per_cpu_empty(cpu_buffer)) |
| @@ -3835,7 +3842,8 @@ int ring_buffer_read_page(struct ring_buffer *buffer, | |||
| 3835 | if (len > (commit - read)) | 3842 | if (len > (commit - read)) |
| 3836 | len = (commit - read); | 3843 | len = (commit - read); |
| 3837 | 3844 | ||
| 3838 | size = rb_event_length(event); | 3845 | /* Always keep the time extend and data together */ |
| 3846 | size = rb_event_ts_length(event); | ||
| 3839 | 3847 | ||
| 3840 | if (len < size) | 3848 | if (len < size) |
| 3841 | goto out_unlock; | 3849 | goto out_unlock; |
| @@ -3857,7 +3865,8 @@ int ring_buffer_read_page(struct ring_buffer *buffer, | |||
| 3857 | break; | 3865 | break; |
| 3858 | 3866 | ||
| 3859 | event = rb_reader_event(cpu_buffer); | 3867 | event = rb_reader_event(cpu_buffer); |
| 3860 | size = rb_event_length(event); | 3868 | /* Always keep the time extend and data together */ |
| 3869 | size = rb_event_ts_length(event); | ||
| 3861 | } while (len > size); | 3870 | } while (len > size); |
| 3862 | 3871 | ||
| 3863 | /* update bpage */ | 3872 | /* update bpage */ |
diff --git a/kernel/trace/trace.c b/kernel/trace/trace.c index 001bcd2ccf4a..82d9b8106cd0 100644 --- a/kernel/trace/trace.c +++ b/kernel/trace/trace.c | |||
| @@ -3996,13 +3996,9 @@ static void tracing_init_debugfs_percpu(long cpu) | |||
| 3996 | { | 3996 | { |
| 3997 | struct dentry *d_percpu = tracing_dentry_percpu(); | 3997 | struct dentry *d_percpu = tracing_dentry_percpu(); |
| 3998 | struct dentry *d_cpu; | 3998 | struct dentry *d_cpu; |
| 3999 | /* strlen(cpu) + MAX(log10(cpu)) + '\0' */ | 3999 | char cpu_dir[30]; /* 30 characters should be more than enough */ |
| 4000 | char cpu_dir[7]; | ||
| 4001 | 4000 | ||
| 4002 | if (cpu > 999 || cpu < 0) | 4001 | snprintf(cpu_dir, 30, "cpu%ld", cpu); |
| 4003 | return; | ||
| 4004 | |||
| 4005 | sprintf(cpu_dir, "cpu%ld", cpu); | ||
| 4006 | d_cpu = debugfs_create_dir(cpu_dir, d_percpu); | 4002 | d_cpu = debugfs_create_dir(cpu_dir, d_percpu); |
| 4007 | if (!d_cpu) { | 4003 | if (!d_cpu) { |
| 4008 | pr_warning("Could not create debugfs '%s' entry\n", cpu_dir); | 4004 | pr_warning("Could not create debugfs '%s' entry\n", cpu_dir); |
diff --git a/tools/perf/Documentation/perf-list.txt b/tools/perf/Documentation/perf-list.txt index 43e3dd284b90..399751befeed 100644 --- a/tools/perf/Documentation/perf-list.txt +++ b/tools/perf/Documentation/perf-list.txt | |||
| @@ -15,6 +15,23 @@ DESCRIPTION | |||
| 15 | This command displays the symbolic event types which can be selected in the | 15 | This command displays the symbolic event types which can be selected in the |
| 16 | various perf commands with the -e option. | 16 | various perf commands with the -e option. |
| 17 | 17 | ||
| 18 | EVENT MODIFIERS | ||
| 19 | --------------- | ||
| 20 | |||
| 21 | Events can optionally have a modifer by appending a colon and one or | ||
| 22 | more modifiers. Modifiers allow the user to restrict when events are | ||
| 23 | counted with 'u' for user-space, 'k' for kernel, 'h' for hypervisor. | ||
| 24 | |||
| 25 | The 'p' modifier can be used for specifying how precise the instruction | ||
| 26 | address should be. The 'p' modifier is currently only implemented for | ||
| 27 | Intel PEBS and can be specified multiple times: | ||
| 28 | 0 - SAMPLE_IP can have arbitrary skid | ||
| 29 | 1 - SAMPLE_IP must have constant skid | ||
| 30 | 2 - SAMPLE_IP requested to have 0 skid | ||
| 31 | 3 - SAMPLE_IP must have 0 skid | ||
| 32 | |||
| 33 | The PEBS implementation now supports up to 2. | ||
| 34 | |||
| 18 | RAW HARDWARE EVENT DESCRIPTOR | 35 | RAW HARDWARE EVENT DESCRIPTOR |
| 19 | ----------------------------- | 36 | ----------------------------- |
| 20 | Even when an event is not available in a symbolic form within perf right now, | 37 | Even when an event is not available in a symbolic form within perf right now, |
diff --git a/tools/perf/Documentation/perf-probe.txt b/tools/perf/Documentation/perf-probe.txt index 27d52dae5a43..62de1b7f4e76 100644 --- a/tools/perf/Documentation/perf-probe.txt +++ b/tools/perf/Documentation/perf-probe.txt | |||
| @@ -16,7 +16,9 @@ or | |||
| 16 | or | 16 | or |
| 17 | 'perf probe' --list | 17 | 'perf probe' --list |
| 18 | or | 18 | or |
| 19 | 'perf probe' --line='FUNC[:RLN[+NUM|:RLN2]]|SRC:ALN[+NUM|:ALN2]' | 19 | 'perf probe' [options] --line='FUNC[:RLN[+NUM|:RLN2]]|SRC:ALN[+NUM|:ALN2]' |
| 20 | or | ||
| 21 | 'perf probe' [options] --vars='PROBEPOINT' | ||
| 20 | 22 | ||
| 21 | DESCRIPTION | 23 | DESCRIPTION |
| 22 | ----------- | 24 | ----------- |
| @@ -31,6 +33,11 @@ OPTIONS | |||
| 31 | --vmlinux=PATH:: | 33 | --vmlinux=PATH:: |
| 32 | Specify vmlinux path which has debuginfo (Dwarf binary). | 34 | Specify vmlinux path which has debuginfo (Dwarf binary). |
| 33 | 35 | ||
| 36 | -m:: | ||
| 37 | --module=MODNAME:: | ||
| 38 | Specify module name in which perf-probe searches probe points | ||
| 39 | or lines. | ||
| 40 | |||
| 34 | -s:: | 41 | -s:: |
| 35 | --source=PATH:: | 42 | --source=PATH:: |
| 36 | Specify path to kernel source. | 43 | Specify path to kernel source. |
| @@ -57,6 +64,15 @@ OPTIONS | |||
| 57 | Show source code lines which can be probed. This needs an argument | 64 | Show source code lines which can be probed. This needs an argument |
| 58 | which specifies a range of the source code. (see LINE SYNTAX for detail) | 65 | which specifies a range of the source code. (see LINE SYNTAX for detail) |
| 59 | 66 | ||
| 67 | -V:: | ||
| 68 | --vars=:: | ||
| 69 | Show available local variables at given probe point. The argument | ||
| 70 | syntax is same as PROBE SYNTAX, but NO ARGs. | ||
| 71 | |||
| 72 | --externs:: | ||
| 73 | (Only for --vars) Show external defined variables in addition to local | ||
| 74 | variables. | ||
| 75 | |||
| 60 | -f:: | 76 | -f:: |
| 61 | --force:: | 77 | --force:: |
| 62 | Forcibly add events with existing name. | 78 | Forcibly add events with existing name. |
diff --git a/tools/perf/Documentation/perf-record.txt b/tools/perf/Documentation/perf-record.txt index 3ee27dccfde9..a91f9f9e6e5c 100644 --- a/tools/perf/Documentation/perf-record.txt +++ b/tools/perf/Documentation/perf-record.txt | |||
| @@ -83,6 +83,10 @@ OPTIONS | |||
| 83 | --call-graph:: | 83 | --call-graph:: |
| 84 | Do call-graph (stack chain/backtrace) recording. | 84 | Do call-graph (stack chain/backtrace) recording. |
| 85 | 85 | ||
| 86 | -q:: | ||
| 87 | --quiet:: | ||
| 88 | Don't print any message, useful for scripting. | ||
| 89 | |||
| 86 | -v:: | 90 | -v:: |
| 87 | --verbose:: | 91 | --verbose:: |
| 88 | Be more verbose (show counter open errors, etc). | 92 | Be more verbose (show counter open errors, etc). |
diff --git a/tools/perf/builtin-probe.c b/tools/perf/builtin-probe.c index 199d5e19554f..2e000c068cc5 100644 --- a/tools/perf/builtin-probe.c +++ b/tools/perf/builtin-probe.c | |||
| @@ -50,14 +50,17 @@ static struct { | |||
| 50 | bool list_events; | 50 | bool list_events; |
| 51 | bool force_add; | 51 | bool force_add; |
| 52 | bool show_lines; | 52 | bool show_lines; |
| 53 | bool show_vars; | ||
| 54 | bool show_ext_vars; | ||
| 55 | bool mod_events; | ||
| 53 | int nevents; | 56 | int nevents; |
| 54 | struct perf_probe_event events[MAX_PROBES]; | 57 | struct perf_probe_event events[MAX_PROBES]; |
| 55 | struct strlist *dellist; | 58 | struct strlist *dellist; |
| 56 | struct line_range line_range; | 59 | struct line_range line_range; |
| 60 | const char *target_module; | ||
| 57 | int max_probe_points; | 61 | int max_probe_points; |
| 58 | } params; | 62 | } params; |
| 59 | 63 | ||
| 60 | |||
| 61 | /* Parse an event definition. Note that any error must die. */ | 64 | /* Parse an event definition. Note that any error must die. */ |
| 62 | static int parse_probe_event(const char *str) | 65 | static int parse_probe_event(const char *str) |
| 63 | { | 66 | { |
| @@ -92,6 +95,7 @@ static int parse_probe_event_argv(int argc, const char **argv) | |||
| 92 | len = 0; | 95 | len = 0; |
| 93 | for (i = 0; i < argc; i++) | 96 | for (i = 0; i < argc; i++) |
| 94 | len += sprintf(&buf[len], "%s ", argv[i]); | 97 | len += sprintf(&buf[len], "%s ", argv[i]); |
| 98 | params.mod_events = true; | ||
| 95 | ret = parse_probe_event(buf); | 99 | ret = parse_probe_event(buf); |
| 96 | free(buf); | 100 | free(buf); |
| 97 | return ret; | 101 | return ret; |
| @@ -100,9 +104,10 @@ static int parse_probe_event_argv(int argc, const char **argv) | |||
| 100 | static int opt_add_probe_event(const struct option *opt __used, | 104 | static int opt_add_probe_event(const struct option *opt __used, |
| 101 | const char *str, int unset __used) | 105 | const char *str, int unset __used) |
| 102 | { | 106 | { |
| 103 | if (str) | 107 | if (str) { |
| 108 | params.mod_events = true; | ||
| 104 | return parse_probe_event(str); | 109 | return parse_probe_event(str); |
| 105 | else | 110 | } else |
| 106 | return 0; | 111 | return 0; |
| 107 | } | 112 | } |
| 108 | 113 | ||
| @@ -110,6 +115,7 @@ static int opt_del_probe_event(const struct option *opt __used, | |||
| 110 | const char *str, int unset __used) | 115 | const char *str, int unset __used) |
| 111 | { | 116 | { |
| 112 | if (str) { | 117 | if (str) { |
| 118 | params.mod_events = true; | ||
| 113 | if (!params.dellist) | 119 | if (!params.dellist) |
| 114 | params.dellist = strlist__new(true, NULL); | 120 | params.dellist = strlist__new(true, NULL); |
| 115 | strlist__add(params.dellist, str); | 121 | strlist__add(params.dellist, str); |
| @@ -130,6 +136,25 @@ static int opt_show_lines(const struct option *opt __used, | |||
| 130 | 136 | ||
| 131 | return ret; | 137 | return ret; |
| 132 | } | 138 | } |
| 139 | |||
| 140 | static int opt_show_vars(const struct option *opt __used, | ||
| 141 | const char *str, int unset __used) | ||
| 142 | { | ||
| 143 | struct perf_probe_event *pev = ¶ms.events[params.nevents]; | ||
| 144 | int ret; | ||
| 145 | |||
| 146 | if (!str) | ||
| 147 | return 0; | ||
| 148 | |||
| 149 | ret = parse_probe_event(str); | ||
| 150 | if (!ret && pev->nargs != 0) { | ||
| 151 | pr_err(" Error: '--vars' doesn't accept arguments.\n"); | ||
| 152 | return -EINVAL; | ||
| 153 | } | ||
| 154 | params.show_vars = true; | ||
| 155 | |||
| 156 | return ret; | ||
| 157 | } | ||
| 133 | #endif | 158 | #endif |
| 134 | 159 | ||
| 135 | static const char * const probe_usage[] = { | 160 | static const char * const probe_usage[] = { |
| @@ -138,7 +163,8 @@ static const char * const probe_usage[] = { | |||
| 138 | "perf probe [<options>] --del '[GROUP:]EVENT' ...", | 163 | "perf probe [<options>] --del '[GROUP:]EVENT' ...", |
| 139 | "perf probe --list", | 164 | "perf probe --list", |
| 140 | #ifdef DWARF_SUPPORT | 165 | #ifdef DWARF_SUPPORT |
| 141 | "perf probe --line 'LINEDESC'", | 166 | "perf probe [<options>] --line 'LINEDESC'", |
| 167 | "perf probe [<options>] --vars 'PROBEPOINT'", | ||
| 142 | #endif | 168 | #endif |
| 143 | NULL | 169 | NULL |
| 144 | }; | 170 | }; |
| @@ -180,10 +206,17 @@ static const struct option options[] = { | |||
| 180 | OPT_CALLBACK('L', "line", NULL, | 206 | OPT_CALLBACK('L', "line", NULL, |
| 181 | "FUNC[:RLN[+NUM|-RLN2]]|SRC:ALN[+NUM|-ALN2]", | 207 | "FUNC[:RLN[+NUM|-RLN2]]|SRC:ALN[+NUM|-ALN2]", |
| 182 | "Show source code lines.", opt_show_lines), | 208 | "Show source code lines.", opt_show_lines), |
| 209 | OPT_CALLBACK('V', "vars", NULL, | ||
| 210 | "FUNC[@SRC][+OFF|%return|:RL|;PT]|SRC:AL|SRC;PT", | ||
| 211 | "Show accessible variables on PROBEDEF", opt_show_vars), | ||
| 212 | OPT_BOOLEAN('\0', "externs", ¶ms.show_ext_vars, | ||
| 213 | "Show external variables too (with --vars only)"), | ||
| 183 | OPT_STRING('k', "vmlinux", &symbol_conf.vmlinux_name, | 214 | OPT_STRING('k', "vmlinux", &symbol_conf.vmlinux_name, |
| 184 | "file", "vmlinux pathname"), | 215 | "file", "vmlinux pathname"), |
| 185 | OPT_STRING('s', "source", &symbol_conf.source_prefix, | 216 | OPT_STRING('s', "source", &symbol_conf.source_prefix, |
| 186 | "directory", "path to kernel source"), | 217 | "directory", "path to kernel source"), |
| 218 | OPT_STRING('m', "module", ¶ms.target_module, | ||
| 219 | "modname", "target module name"), | ||
| 187 | #endif | 220 | #endif |
| 188 | OPT__DRY_RUN(&probe_event_dry_run), | 221 | OPT__DRY_RUN(&probe_event_dry_run), |
| 189 | OPT_INTEGER('\0', "max-probes", ¶ms.max_probe_points, | 222 | OPT_INTEGER('\0', "max-probes", ¶ms.max_probe_points, |
| @@ -217,7 +250,7 @@ int cmd_probe(int argc, const char **argv, const char *prefix __used) | |||
| 217 | usage_with_options(probe_usage, options); | 250 | usage_with_options(probe_usage, options); |
| 218 | 251 | ||
| 219 | if (params.list_events) { | 252 | if (params.list_events) { |
| 220 | if (params.nevents != 0 || params.dellist) { | 253 | if (params.mod_events) { |
| 221 | pr_err(" Error: Don't use --list with --add/--del.\n"); | 254 | pr_err(" Error: Don't use --list with --add/--del.\n"); |
| 222 | usage_with_options(probe_usage, options); | 255 | usage_with_options(probe_usage, options); |
| 223 | } | 256 | } |
| @@ -225,6 +258,10 @@ int cmd_probe(int argc, const char **argv, const char *prefix __used) | |||
| 225 | pr_err(" Error: Don't use --list with --line.\n"); | 258 | pr_err(" Error: Don't use --list with --line.\n"); |
| 226 | usage_with_options(probe_usage, options); | 259 | usage_with_options(probe_usage, options); |
| 227 | } | 260 | } |
| 261 | if (params.show_vars) { | ||
| 262 | pr_err(" Error: Don't use --list with --vars.\n"); | ||
| 263 | usage_with_options(probe_usage, options); | ||
| 264 | } | ||
| 228 | ret = show_perf_probe_events(); | 265 | ret = show_perf_probe_events(); |
| 229 | if (ret < 0) | 266 | if (ret < 0) |
| 230 | pr_err(" Error: Failed to show event list. (%d)\n", | 267 | pr_err(" Error: Failed to show event list. (%d)\n", |
| @@ -234,17 +271,35 @@ int cmd_probe(int argc, const char **argv, const char *prefix __used) | |||
| 234 | 271 | ||
| 235 | #ifdef DWARF_SUPPORT | 272 | #ifdef DWARF_SUPPORT |
| 236 | if (params.show_lines) { | 273 | if (params.show_lines) { |
| 237 | if (params.nevents != 0 || params.dellist) { | 274 | if (params.mod_events) { |
| 238 | pr_warning(" Error: Don't use --line with" | 275 | pr_err(" Error: Don't use --line with" |
| 239 | " --add/--del.\n"); | 276 | " --add/--del.\n"); |
| 277 | usage_with_options(probe_usage, options); | ||
| 278 | } | ||
| 279 | if (params.show_vars) { | ||
| 280 | pr_err(" Error: Don't use --line with --vars.\n"); | ||
| 240 | usage_with_options(probe_usage, options); | 281 | usage_with_options(probe_usage, options); |
| 241 | } | 282 | } |
| 242 | 283 | ||
| 243 | ret = show_line_range(¶ms.line_range); | 284 | ret = show_line_range(¶ms.line_range, params.target_module); |
| 244 | if (ret < 0) | 285 | if (ret < 0) |
| 245 | pr_err(" Error: Failed to show lines. (%d)\n", ret); | 286 | pr_err(" Error: Failed to show lines. (%d)\n", ret); |
| 246 | return ret; | 287 | return ret; |
| 247 | } | 288 | } |
| 289 | if (params.show_vars) { | ||
| 290 | if (params.mod_events) { | ||
| 291 | pr_err(" Error: Don't use --vars with" | ||
| 292 | " --add/--del.\n"); | ||
| 293 | usage_with_options(probe_usage, options); | ||
| 294 | } | ||
| 295 | ret = show_available_vars(params.events, params.nevents, | ||
| 296 | params.max_probe_points, | ||
| 297 | params.target_module, | ||
| 298 | params.show_ext_vars); | ||
| 299 | if (ret < 0) | ||
| 300 | pr_err(" Error: Failed to show vars. (%d)\n", ret); | ||
| 301 | return ret; | ||
| 302 | } | ||
| 248 | #endif | 303 | #endif |
| 249 | 304 | ||
| 250 | if (params.dellist) { | 305 | if (params.dellist) { |
| @@ -258,8 +313,9 @@ int cmd_probe(int argc, const char **argv, const char *prefix __used) | |||
| 258 | 313 | ||
| 259 | if (params.nevents) { | 314 | if (params.nevents) { |
| 260 | ret = add_perf_probe_events(params.events, params.nevents, | 315 | ret = add_perf_probe_events(params.events, params.nevents, |
| 261 | params.force_add, | 316 | params.max_probe_points, |
| 262 | params.max_probe_points); | 317 | params.target_module, |
| 318 | params.force_add); | ||
| 263 | if (ret < 0) { | 319 | if (ret < 0) { |
| 264 | pr_err(" Error: Failed to add events. (%d)\n", ret); | 320 | pr_err(" Error: Failed to add events. (%d)\n", ret); |
| 265 | return ret; | 321 | return ret; |
diff --git a/tools/perf/builtin-record.c b/tools/perf/builtin-record.c index ff77b805de71..4e75583ddd6d 100644 --- a/tools/perf/builtin-record.c +++ b/tools/perf/builtin-record.c | |||
| @@ -353,7 +353,7 @@ try_again: | |||
| 353 | } | 353 | } |
| 354 | 354 | ||
| 355 | if (read(fd[nr_cpu][counter][thread_index], &read_data, sizeof(read_data)) == -1) { | 355 | if (read(fd[nr_cpu][counter][thread_index], &read_data, sizeof(read_data)) == -1) { |
| 356 | perror("Unable to read perf file descriptor\n"); | 356 | perror("Unable to read perf file descriptor"); |
| 357 | exit(-1); | 357 | exit(-1); |
| 358 | } | 358 | } |
| 359 | 359 | ||
| @@ -626,7 +626,7 @@ static int __cmd_record(int argc, const char **argv) | |||
| 626 | 626 | ||
| 627 | nr_cpus = read_cpu_map(cpu_list); | 627 | nr_cpus = read_cpu_map(cpu_list); |
| 628 | if (nr_cpus < 1) { | 628 | if (nr_cpus < 1) { |
| 629 | perror("failed to collect number of CPUs\n"); | 629 | perror("failed to collect number of CPUs"); |
| 630 | return -1; | 630 | return -1; |
| 631 | } | 631 | } |
| 632 | 632 | ||
| @@ -761,6 +761,9 @@ static int __cmd_record(int argc, const char **argv) | |||
| 761 | } | 761 | } |
| 762 | } | 762 | } |
| 763 | 763 | ||
| 764 | if (quiet) | ||
| 765 | return 0; | ||
| 766 | |||
| 764 | fprintf(stderr, "[ perf record: Woken up %ld times to write data ]\n", waking); | 767 | fprintf(stderr, "[ perf record: Woken up %ld times to write data ]\n", waking); |
| 765 | 768 | ||
| 766 | /* | 769 | /* |
| @@ -820,6 +823,7 @@ static const struct option options[] = { | |||
| 820 | "do call-graph (stack chain/backtrace) recording"), | 823 | "do call-graph (stack chain/backtrace) recording"), |
| 821 | OPT_INCR('v', "verbose", &verbose, | 824 | OPT_INCR('v', "verbose", &verbose, |
| 822 | "be more verbose (show counter open errors, etc)"), | 825 | "be more verbose (show counter open errors, etc)"), |
| 826 | OPT_BOOLEAN('q', "quiet", &quiet, "don't print any message"), | ||
| 823 | OPT_BOOLEAN('s', "stat", &inherit_stat, | 827 | OPT_BOOLEAN('s', "stat", &inherit_stat, |
| 824 | "per thread counts"), | 828 | "per thread counts"), |
| 825 | OPT_BOOLEAN('d', "data", &sample_address, | 829 | OPT_BOOLEAN('d', "data", &sample_address, |
diff --git a/tools/perf/builtin-trace.c b/tools/perf/builtin-trace.c index 40a6a2992d15..2f8df45c4dcb 100644 --- a/tools/perf/builtin-trace.c +++ b/tools/perf/builtin-trace.c | |||
| @@ -46,9 +46,6 @@ static struct scripting_ops *scripting_ops; | |||
| 46 | 46 | ||
| 47 | static void setup_scripting(void) | 47 | static void setup_scripting(void) |
| 48 | { | 48 | { |
| 49 | /* make sure PERF_EXEC_PATH is set for scripts */ | ||
| 50 | perf_set_argv_exec_path(perf_exec_path()); | ||
| 51 | |||
| 52 | setup_perl_scripting(); | 49 | setup_perl_scripting(); |
| 53 | setup_python_scripting(); | 50 | setup_python_scripting(); |
| 54 | 51 | ||
| @@ -285,7 +282,7 @@ static int parse_scriptname(const struct option *opt __used, | |||
| 285 | script++; | 282 | script++; |
| 286 | } else { | 283 | } else { |
| 287 | script = str; | 284 | script = str; |
| 288 | ext = strchr(script, '.'); | 285 | ext = strrchr(script, '.'); |
| 289 | if (!ext) { | 286 | if (!ext) { |
| 290 | fprintf(stderr, "invalid script extension"); | 287 | fprintf(stderr, "invalid script extension"); |
| 291 | return -1; | 288 | return -1; |
| @@ -593,6 +590,9 @@ int cmd_trace(int argc, const char **argv, const char *prefix __used) | |||
| 593 | suffix = REPORT_SUFFIX; | 590 | suffix = REPORT_SUFFIX; |
| 594 | } | 591 | } |
| 595 | 592 | ||
| 593 | /* make sure PERF_EXEC_PATH is set for scripts */ | ||
| 594 | perf_set_argv_exec_path(perf_exec_path()); | ||
| 595 | |||
| 596 | if (!suffix && argc >= 2 && strncmp(argv[1], "-", strlen("-")) != 0) { | 596 | if (!suffix && argc >= 2 && strncmp(argv[1], "-", strlen("-")) != 0) { |
| 597 | char *record_script_path, *report_script_path; | 597 | char *record_script_path, *report_script_path; |
| 598 | int live_pipe[2]; | 598 | int live_pipe[2]; |
| @@ -625,12 +625,13 @@ int cmd_trace(int argc, const char **argv, const char *prefix __used) | |||
| 625 | dup2(live_pipe[1], 1); | 625 | dup2(live_pipe[1], 1); |
| 626 | close(live_pipe[0]); | 626 | close(live_pipe[0]); |
| 627 | 627 | ||
| 628 | __argv = malloc(5 * sizeof(const char *)); | 628 | __argv = malloc(6 * sizeof(const char *)); |
| 629 | __argv[0] = "/bin/sh"; | 629 | __argv[0] = "/bin/sh"; |
| 630 | __argv[1] = record_script_path; | 630 | __argv[1] = record_script_path; |
| 631 | __argv[2] = "-o"; | 631 | __argv[2] = "-q"; |
| 632 | __argv[3] = "-"; | 632 | __argv[3] = "-o"; |
| 633 | __argv[4] = NULL; | 633 | __argv[4] = "-"; |
| 634 | __argv[5] = NULL; | ||
| 634 | 635 | ||
| 635 | execvp("/bin/sh", (char **)__argv); | 636 | execvp("/bin/sh", (char **)__argv); |
| 636 | exit(-1); | 637 | exit(-1); |
diff --git a/tools/perf/scripts/perl/bin/failed-syscalls-report b/tools/perf/scripts/perl/bin/failed-syscalls-report index e3a5e55d54ff..4028d92dc4ae 100644 --- a/tools/perf/scripts/perl/bin/failed-syscalls-report +++ b/tools/perf/scripts/perl/bin/failed-syscalls-report | |||
| @@ -7,4 +7,4 @@ if [ $# -gt 0 ] ; then | |||
| 7 | shift | 7 | shift |
| 8 | fi | 8 | fi |
| 9 | fi | 9 | fi |
| 10 | perf trace $@ -s ~/libexec/perf-core/scripts/perl/failed-syscalls.pl $comm | 10 | perf trace $@ -s "$PERF_EXEC_PATH"/scripts/perl/failed-syscalls.pl $comm |
diff --git a/tools/perf/scripts/perl/bin/rw-by-file-report b/tools/perf/scripts/perl/bin/rw-by-file-report index d83070b7eeb5..ba25f4d41fb0 100644 --- a/tools/perf/scripts/perl/bin/rw-by-file-report +++ b/tools/perf/scripts/perl/bin/rw-by-file-report | |||
| @@ -7,7 +7,7 @@ if [ $# -lt 1 ] ; then | |||
| 7 | fi | 7 | fi |
| 8 | comm=$1 | 8 | comm=$1 |
| 9 | shift | 9 | shift |
| 10 | perf trace $@ -s ~/libexec/perf-core/scripts/perl/rw-by-file.pl $comm | 10 | perf trace $@ -s "$PERF_EXEC_PATH"/scripts/perl/rw-by-file.pl $comm |
| 11 | 11 | ||
| 12 | 12 | ||
| 13 | 13 | ||
diff --git a/tools/perf/scripts/perl/bin/rw-by-pid-report b/tools/perf/scripts/perl/bin/rw-by-pid-report index 7ef46983f62f..641a3f5d085c 100644 --- a/tools/perf/scripts/perl/bin/rw-by-pid-report +++ b/tools/perf/scripts/perl/bin/rw-by-pid-report | |||
| @@ -1,6 +1,6 @@ | |||
| 1 | #!/bin/bash | 1 | #!/bin/bash |
| 2 | # description: system-wide r/w activity | 2 | # description: system-wide r/w activity |
| 3 | perf trace $@ -s ~/libexec/perf-core/scripts/perl/rw-by-pid.pl | 3 | perf trace $@ -s "$PERF_EXEC_PATH"/scripts/perl/rw-by-pid.pl |
| 4 | 4 | ||
| 5 | 5 | ||
| 6 | 6 | ||
diff --git a/tools/perf/scripts/perl/bin/rwtop-report b/tools/perf/scripts/perl/bin/rwtop-report index 93e698cd3f38..4918dba77021 100644 --- a/tools/perf/scripts/perl/bin/rwtop-report +++ b/tools/perf/scripts/perl/bin/rwtop-report | |||
| @@ -17,7 +17,7 @@ if [ "$n_args" -gt 0 ] ; then | |||
| 17 | interval=$1 | 17 | interval=$1 |
| 18 | shift | 18 | shift |
| 19 | fi | 19 | fi |
| 20 | perf trace $@ -s ~/libexec/perf-core/scripts/perl/rwtop.pl $interval | 20 | perf trace $@ -s "$PERF_EXEC_PATH"/scripts/perl/rwtop.pl $interval |
| 21 | 21 | ||
| 22 | 22 | ||
| 23 | 23 | ||
diff --git a/tools/perf/scripts/perl/bin/wakeup-latency-report b/tools/perf/scripts/perl/bin/wakeup-latency-report index a0d898f9ca1d..49052ebcb632 100644 --- a/tools/perf/scripts/perl/bin/wakeup-latency-report +++ b/tools/perf/scripts/perl/bin/wakeup-latency-report | |||
| @@ -1,6 +1,6 @@ | |||
| 1 | #!/bin/bash | 1 | #!/bin/bash |
| 2 | # description: system-wide min/max/avg wakeup latency | 2 | # description: system-wide min/max/avg wakeup latency |
| 3 | perf trace $@ -s ~/libexec/perf-core/scripts/perl/wakeup-latency.pl | 3 | perf trace $@ -s "$PERF_EXEC_PATH"/scripts/perl/wakeup-latency.pl |
| 4 | 4 | ||
| 5 | 5 | ||
| 6 | 6 | ||
diff --git a/tools/perf/scripts/perl/bin/workqueue-stats-report b/tools/perf/scripts/perl/bin/workqueue-stats-report index 35081132ef97..df0c65f4ca93 100644 --- a/tools/perf/scripts/perl/bin/workqueue-stats-report +++ b/tools/perf/scripts/perl/bin/workqueue-stats-report | |||
| @@ -1,6 +1,6 @@ | |||
| 1 | #!/bin/bash | 1 | #!/bin/bash |
| 2 | # description: workqueue stats (ins/exe/create/destroy) | 2 | # description: workqueue stats (ins/exe/create/destroy) |
| 3 | perf trace $@ -s ~/libexec/perf-core/scripts/perl/workqueue-stats.pl | 3 | perf trace $@ -s "$PERF_EXEC_PATH"/scripts/perl/workqueue-stats.pl |
| 4 | 4 | ||
| 5 | 5 | ||
| 6 | 6 | ||
diff --git a/tools/perf/scripts/python/Perf-Trace-Util/lib/Perf/Trace/Util.py b/tools/perf/scripts/python/Perf-Trace-Util/lib/Perf/Trace/Util.py index 9689bc0acd9f..13cc02b5893a 100644 --- a/tools/perf/scripts/python/Perf-Trace-Util/lib/Perf/Trace/Util.py +++ b/tools/perf/scripts/python/Perf-Trace-Util/lib/Perf/Trace/Util.py | |||
| @@ -6,6 +6,14 @@ | |||
| 6 | # Public License ("GPL") version 2 as published by the Free Software | 6 | # Public License ("GPL") version 2 as published by the Free Software |
| 7 | # Foundation. | 7 | # Foundation. |
| 8 | 8 | ||
| 9 | import errno, os | ||
| 10 | |||
| 11 | FUTEX_WAIT = 0 | ||
| 12 | FUTEX_WAKE = 1 | ||
| 13 | FUTEX_PRIVATE_FLAG = 128 | ||
| 14 | FUTEX_CLOCK_REALTIME = 256 | ||
| 15 | FUTEX_CMD_MASK = ~(FUTEX_PRIVATE_FLAG | FUTEX_CLOCK_REALTIME) | ||
| 16 | |||
| 9 | NSECS_PER_SEC = 1000000000 | 17 | NSECS_PER_SEC = 1000000000 |
| 10 | 18 | ||
| 11 | def avg(total, n): | 19 | def avg(total, n): |
| @@ -24,5 +32,55 @@ def nsecs_str(nsecs): | |||
| 24 | str = "%5u.%09u" % (nsecs_secs(nsecs), nsecs_nsecs(nsecs)), | 32 | str = "%5u.%09u" % (nsecs_secs(nsecs), nsecs_nsecs(nsecs)), |
| 25 | return str | 33 | return str |
| 26 | 34 | ||
| 35 | def add_stats(dict, key, value): | ||
| 36 | if not dict.has_key(key): | ||
| 37 | dict[key] = (value, value, value, 1) | ||
| 38 | else: | ||
| 39 | min, max, avg, count = dict[key] | ||
| 40 | if value < min: | ||
| 41 | min = value | ||
| 42 | if value > max: | ||
| 43 | max = value | ||
| 44 | avg = (avg + value) / 2 | ||
| 45 | dict[key] = (min, max, avg, count + 1) | ||
| 46 | |||
| 27 | def clear_term(): | 47 | def clear_term(): |
| 28 | print("\x1b[H\x1b[2J") | 48 | print("\x1b[H\x1b[2J") |
| 49 | |||
| 50 | audit_package_warned = False | ||
| 51 | |||
| 52 | try: | ||
| 53 | import audit | ||
| 54 | machine_to_id = { | ||
| 55 | 'x86_64': audit.MACH_86_64, | ||
| 56 | 'alpha' : audit.MACH_ALPHA, | ||
| 57 | 'ia64' : audit.MACH_IA64, | ||
| 58 | 'ppc' : audit.MACH_PPC, | ||
| 59 | 'ppc64' : audit.MACH_PPC64, | ||
| 60 | 's390' : audit.MACH_S390, | ||
| 61 | 's390x' : audit.MACH_S390X, | ||
| 62 | 'i386' : audit.MACH_X86, | ||
| 63 | 'i586' : audit.MACH_X86, | ||
| 64 | 'i686' : audit.MACH_X86, | ||
| 65 | } | ||
| 66 | try: | ||
| 67 | machine_to_id['armeb'] = audit.MACH_ARMEB | ||
| 68 | except: | ||
| 69 | pass | ||
| 70 | machine_id = machine_to_id[os.uname()[4]] | ||
| 71 | except: | ||
| 72 | if not audit_package_warned: | ||
| 73 | audit_package_warned = True | ||
| 74 | print "Install the audit-libs-python package to get syscall names" | ||
| 75 | |||
| 76 | def syscall_name(id): | ||
| 77 | try: | ||
| 78 | return audit.audit_syscall_to_name(id, machine_id) | ||
| 79 | except: | ||
| 80 | return str(id) | ||
| 81 | |||
| 82 | def strerror(nr): | ||
| 83 | try: | ||
| 84 | return errno.errorcode[abs(nr)] | ||
| 85 | except: | ||
| 86 | return "Unknown %d errno" % nr | ||
diff --git a/tools/perf/scripts/python/bin/failed-syscalls-by-pid-report b/tools/perf/scripts/python/bin/failed-syscalls-by-pid-report index 30293545fcc2..03587021463d 100644 --- a/tools/perf/scripts/python/bin/failed-syscalls-by-pid-report +++ b/tools/perf/scripts/python/bin/failed-syscalls-by-pid-report | |||
| @@ -7,4 +7,4 @@ if [ $# -gt 0 ] ; then | |||
| 7 | shift | 7 | shift |
| 8 | fi | 8 | fi |
| 9 | fi | 9 | fi |
| 10 | perf trace $@ -s ~/libexec/perf-core/scripts/python/failed-syscalls-by-pid.py $comm | 10 | perf trace $@ -s "$PERF_EXEC_PATH"/scripts/python/failed-syscalls-by-pid.py $comm |
diff --git a/tools/perf/scripts/python/bin/futex-contention-record b/tools/perf/scripts/python/bin/futex-contention-record new file mode 100644 index 000000000000..5ecbb433caf4 --- /dev/null +++ b/tools/perf/scripts/python/bin/futex-contention-record | |||
| @@ -0,0 +1,2 @@ | |||
| 1 | #!/bin/bash | ||
| 2 | perf record -a -e syscalls:sys_enter_futex -e syscalls:sys_exit_futex $@ | ||
diff --git a/tools/perf/scripts/python/bin/futex-contention-report b/tools/perf/scripts/python/bin/futex-contention-report new file mode 100644 index 000000000000..c8268138fb7e --- /dev/null +++ b/tools/perf/scripts/python/bin/futex-contention-report | |||
| @@ -0,0 +1,4 @@ | |||
| 1 | #!/bin/bash | ||
| 2 | # description: futext contention measurement | ||
| 3 | |||
| 4 | perf trace $@ -s "$PERF_EXEC_PATH"/scripts/python/futex-contention.py | ||
diff --git a/tools/perf/scripts/python/bin/netdev-times-report b/tools/perf/scripts/python/bin/netdev-times-report index c3d0a638123d..4ad361b31249 100644 --- a/tools/perf/scripts/python/bin/netdev-times-report +++ b/tools/perf/scripts/python/bin/netdev-times-report | |||
| @@ -2,4 +2,4 @@ | |||
| 2 | # description: display a process of packet and processing time | 2 | # description: display a process of packet and processing time |
| 3 | # args: [tx] [rx] [dev=] [debug] | 3 | # args: [tx] [rx] [dev=] [debug] |
| 4 | 4 | ||
| 5 | perf trace -s ~/libexec/perf-core/scripts/python/netdev-times.py $@ | 5 | perf trace -s "$PERF_EXEC_PATH"/scripts/python/netdev-times.py $@ |
diff --git a/tools/perf/scripts/python/bin/sched-migration-report b/tools/perf/scripts/python/bin/sched-migration-report index 61d05f72e443..df1791f07c24 100644 --- a/tools/perf/scripts/python/bin/sched-migration-report +++ b/tools/perf/scripts/python/bin/sched-migration-report | |||
| @@ -1,3 +1,3 @@ | |||
| 1 | #!/bin/bash | 1 | #!/bin/bash |
| 2 | # description: sched migration overview | 2 | # description: sched migration overview |
| 3 | perf trace $@ -s ~/libexec/perf-core/scripts/python/sched-migration.py | 3 | perf trace $@ -s "$PERF_EXEC_PATH"/scripts/python/sched-migration.py |
diff --git a/tools/perf/scripts/python/bin/sctop-report b/tools/perf/scripts/python/bin/sctop-report index b01c842ae7b4..36b409c05e50 100644 --- a/tools/perf/scripts/python/bin/sctop-report +++ b/tools/perf/scripts/python/bin/sctop-report | |||
| @@ -21,4 +21,4 @@ elif [ "$n_args" -gt 0 ] ; then | |||
| 21 | interval=$1 | 21 | interval=$1 |
| 22 | shift | 22 | shift |
| 23 | fi | 23 | fi |
| 24 | perf trace $@ -s ~/libexec/perf-core/scripts/python/sctop.py $comm $interval | 24 | perf trace $@ -s "$PERF_EXEC_PATH"/scripts/python/sctop.py $comm $interval |
diff --git a/tools/perf/scripts/python/bin/syscall-counts-by-pid-report b/tools/perf/scripts/python/bin/syscall-counts-by-pid-report index 9e9d8ddd72ce..4eb88c9fc83c 100644 --- a/tools/perf/scripts/python/bin/syscall-counts-by-pid-report +++ b/tools/perf/scripts/python/bin/syscall-counts-by-pid-report | |||
| @@ -7,4 +7,4 @@ if [ $# -gt 0 ] ; then | |||
| 7 | shift | 7 | shift |
| 8 | fi | 8 | fi |
| 9 | fi | 9 | fi |
| 10 | perf trace $@ -s ~/libexec/perf-core/scripts/python/syscall-counts-by-pid.py $comm | 10 | perf trace $@ -s "$PERF_EXEC_PATH"/scripts/python/syscall-counts-by-pid.py $comm |
diff --git a/tools/perf/scripts/python/bin/syscall-counts-report b/tools/perf/scripts/python/bin/syscall-counts-report index dc076b618796..cb2f9c5cf17e 100644 --- a/tools/perf/scripts/python/bin/syscall-counts-report +++ b/tools/perf/scripts/python/bin/syscall-counts-report | |||
| @@ -7,4 +7,4 @@ if [ $# -gt 0 ] ; then | |||
| 7 | shift | 7 | shift |
| 8 | fi | 8 | fi |
| 9 | fi | 9 | fi |
| 10 | perf trace $@ -s ~/libexec/perf-core/scripts/python/syscall-counts.py $comm | 10 | perf trace $@ -s "$PERF_EXEC_PATH"/scripts/python/syscall-counts.py $comm |
diff --git a/tools/perf/scripts/python/failed-syscalls-by-pid.py b/tools/perf/scripts/python/failed-syscalls-by-pid.py index 0ca02278fe69..acd7848717b3 100644 --- a/tools/perf/scripts/python/failed-syscalls-by-pid.py +++ b/tools/perf/scripts/python/failed-syscalls-by-pid.py | |||
| @@ -13,21 +13,26 @@ sys.path.append(os.environ['PERF_EXEC_PATH'] + \ | |||
| 13 | 13 | ||
| 14 | from perf_trace_context import * | 14 | from perf_trace_context import * |
| 15 | from Core import * | 15 | from Core import * |
| 16 | from Util import * | ||
| 16 | 17 | ||
| 17 | usage = "perf trace -s syscall-counts-by-pid.py [comm]\n"; | 18 | usage = "perf trace -s syscall-counts-by-pid.py [comm|pid]\n"; |
| 18 | 19 | ||
| 19 | for_comm = None | 20 | for_comm = None |
| 21 | for_pid = None | ||
| 20 | 22 | ||
| 21 | if len(sys.argv) > 2: | 23 | if len(sys.argv) > 2: |
| 22 | sys.exit(usage) | 24 | sys.exit(usage) |
| 23 | 25 | ||
| 24 | if len(sys.argv) > 1: | 26 | if len(sys.argv) > 1: |
| 25 | for_comm = sys.argv[1] | 27 | try: |
| 28 | for_pid = int(sys.argv[1]) | ||
| 29 | except: | ||
| 30 | for_comm = sys.argv[1] | ||
| 26 | 31 | ||
| 27 | syscalls = autodict() | 32 | syscalls = autodict() |
| 28 | 33 | ||
| 29 | def trace_begin(): | 34 | def trace_begin(): |
| 30 | pass | 35 | print "Press control+C to stop and show the summary" |
| 31 | 36 | ||
| 32 | def trace_end(): | 37 | def trace_end(): |
| 33 | print_error_totals() | 38 | print_error_totals() |
| @@ -35,9 +40,9 @@ def trace_end(): | |||
| 35 | def raw_syscalls__sys_exit(event_name, context, common_cpu, | 40 | def raw_syscalls__sys_exit(event_name, context, common_cpu, |
| 36 | common_secs, common_nsecs, common_pid, common_comm, | 41 | common_secs, common_nsecs, common_pid, common_comm, |
| 37 | id, ret): | 42 | id, ret): |
| 38 | if for_comm is not None: | 43 | if (for_comm and common_comm != for_comm) or \ |
| 39 | if common_comm != for_comm: | 44 | (for_pid and common_pid != for_pid ): |
| 40 | return | 45 | return |
| 41 | 46 | ||
| 42 | if ret < 0: | 47 | if ret < 0: |
| 43 | try: | 48 | try: |
| @@ -62,7 +67,7 @@ def print_error_totals(): | |||
| 62 | print "\n%s [%d]\n" % (comm, pid), | 67 | print "\n%s [%d]\n" % (comm, pid), |
| 63 | id_keys = syscalls[comm][pid].keys() | 68 | id_keys = syscalls[comm][pid].keys() |
| 64 | for id in id_keys: | 69 | for id in id_keys: |
| 65 | print " syscall: %-16d\n" % (id), | 70 | print " syscall: %-16s\n" % syscall_name(id), |
| 66 | ret_keys = syscalls[comm][pid][id].keys() | 71 | ret_keys = syscalls[comm][pid][id].keys() |
| 67 | for ret, val in sorted(syscalls[comm][pid][id].iteritems(), key = lambda(k, v): (v, k), reverse = True): | 72 | for ret, val in sorted(syscalls[comm][pid][id].iteritems(), key = lambda(k, v): (v, k), reverse = True): |
| 68 | print " err = %-20d %10d\n" % (ret, val), | 73 | print " err = %-20s %10d\n" % (strerror(ret), val), |
diff --git a/tools/perf/scripts/python/futex-contention.py b/tools/perf/scripts/python/futex-contention.py new file mode 100644 index 000000000000..11e70a388d41 --- /dev/null +++ b/tools/perf/scripts/python/futex-contention.py | |||
| @@ -0,0 +1,50 @@ | |||
| 1 | # futex contention | ||
| 2 | # (c) 2010, Arnaldo Carvalho de Melo <acme@redhat.com> | ||
| 3 | # Licensed under the terms of the GNU GPL License version 2 | ||
| 4 | # | ||
| 5 | # Translation of: | ||
| 6 | # | ||
| 7 | # http://sourceware.org/systemtap/wiki/WSFutexContention | ||
| 8 | # | ||
| 9 | # to perf python scripting. | ||
| 10 | # | ||
| 11 | # Measures futex contention | ||
| 12 | |||
| 13 | import os, sys | ||
| 14 | sys.path.append(os.environ['PERF_EXEC_PATH'] + '/scripts/python/Perf-Trace-Util/lib/Perf/Trace') | ||
| 15 | from Util import * | ||
| 16 | |||
| 17 | process_names = {} | ||
| 18 | thread_thislock = {} | ||
| 19 | thread_blocktime = {} | ||
| 20 | |||
| 21 | lock_waits = {} # long-lived stats on (tid,lock) blockage elapsed time | ||
| 22 | process_names = {} # long-lived pid-to-execname mapping | ||
| 23 | |||
| 24 | def syscalls__sys_enter_futex(event, ctxt, cpu, s, ns, tid, comm, | ||
| 25 | nr, uaddr, op, val, utime, uaddr2, val3): | ||
| 26 | cmd = op & FUTEX_CMD_MASK | ||
| 27 | if cmd != FUTEX_WAIT: | ||
| 28 | return # we don't care about originators of WAKE events | ||
| 29 | |||
| 30 | process_names[tid] = comm | ||
| 31 | thread_thislock[tid] = uaddr | ||
| 32 | thread_blocktime[tid] = nsecs(s, ns) | ||
| 33 | |||
| 34 | def syscalls__sys_exit_futex(event, ctxt, cpu, s, ns, tid, comm, | ||
| 35 | nr, ret): | ||
| 36 | if thread_blocktime.has_key(tid): | ||
| 37 | elapsed = nsecs(s, ns) - thread_blocktime[tid] | ||
| 38 | add_stats(lock_waits, (tid, thread_thislock[tid]), elapsed) | ||
| 39 | del thread_blocktime[tid] | ||
| 40 | del thread_thislock[tid] | ||
| 41 | |||
| 42 | def trace_begin(): | ||
| 43 | print "Press control+C to stop and show the summary" | ||
| 44 | |||
| 45 | def trace_end(): | ||
| 46 | for (tid, lock) in lock_waits: | ||
| 47 | min, max, avg, count = lock_waits[tid, lock] | ||
| 48 | print "%s[%d] lock %x contended %d times, %d avg ns" % \ | ||
| 49 | (process_names[tid], tid, lock, count, avg) | ||
| 50 | |||
diff --git a/tools/perf/scripts/python/sctop.py b/tools/perf/scripts/python/sctop.py index 6cafad40c296..7a6ec2c7d8ab 100644 --- a/tools/perf/scripts/python/sctop.py +++ b/tools/perf/scripts/python/sctop.py | |||
| @@ -8,10 +8,7 @@ | |||
| 8 | # will be refreshed every [interval] seconds. The default interval is | 8 | # will be refreshed every [interval] seconds. The default interval is |
| 9 | # 3 seconds. | 9 | # 3 seconds. |
| 10 | 10 | ||
| 11 | import thread | 11 | import os, sys, thread, time |
| 12 | import time | ||
| 13 | import os | ||
| 14 | import sys | ||
| 15 | 12 | ||
| 16 | sys.path.append(os.environ['PERF_EXEC_PATH'] + \ | 13 | sys.path.append(os.environ['PERF_EXEC_PATH'] + \ |
| 17 | '/scripts/python/Perf-Trace-Util/lib/Perf/Trace') | 14 | '/scripts/python/Perf-Trace-Util/lib/Perf/Trace') |
| @@ -20,7 +17,7 @@ from perf_trace_context import * | |||
| 20 | from Core import * | 17 | from Core import * |
| 21 | from Util import * | 18 | from Util import * |
| 22 | 19 | ||
| 23 | usage = "perf trace -s syscall-counts.py [comm] [interval]\n"; | 20 | usage = "perf trace -s sctop.py [comm] [interval]\n"; |
| 24 | 21 | ||
| 25 | for_comm = None | 22 | for_comm = None |
| 26 | default_interval = 3 | 23 | default_interval = 3 |
| @@ -71,7 +68,7 @@ def print_syscall_totals(interval): | |||
| 71 | for id, val in sorted(syscalls.iteritems(), key = lambda(k, v): (v, k), \ | 68 | for id, val in sorted(syscalls.iteritems(), key = lambda(k, v): (v, k), \ |
| 72 | reverse = True): | 69 | reverse = True): |
| 73 | try: | 70 | try: |
| 74 | print "%-40d %10d\n" % (id, val), | 71 | print "%-40s %10d\n" % (syscall_name(id), val), |
| 75 | except TypeError: | 72 | except TypeError: |
| 76 | pass | 73 | pass |
| 77 | syscalls.clear() | 74 | syscalls.clear() |
diff --git a/tools/perf/scripts/python/syscall-counts-by-pid.py b/tools/perf/scripts/python/syscall-counts-by-pid.py index af722d6a4b3f..d1ee3ec10cf2 100644 --- a/tools/perf/scripts/python/syscall-counts-by-pid.py +++ b/tools/perf/scripts/python/syscall-counts-by-pid.py | |||
| @@ -5,29 +5,33 @@ | |||
| 5 | # Displays system-wide system call totals, broken down by syscall. | 5 | # Displays system-wide system call totals, broken down by syscall. |
| 6 | # If a [comm] arg is specified, only syscalls called by [comm] are displayed. | 6 | # If a [comm] arg is specified, only syscalls called by [comm] are displayed. |
| 7 | 7 | ||
| 8 | import os | 8 | import os, sys |
| 9 | import sys | ||
| 10 | 9 | ||
| 11 | sys.path.append(os.environ['PERF_EXEC_PATH'] + \ | 10 | sys.path.append(os.environ['PERF_EXEC_PATH'] + \ |
| 12 | '/scripts/python/Perf-Trace-Util/lib/Perf/Trace') | 11 | '/scripts/python/Perf-Trace-Util/lib/Perf/Trace') |
| 13 | 12 | ||
| 14 | from perf_trace_context import * | 13 | from perf_trace_context import * |
| 15 | from Core import * | 14 | from Core import * |
| 15 | from Util import syscall_name | ||
| 16 | 16 | ||
| 17 | usage = "perf trace -s syscall-counts-by-pid.py [comm]\n"; | 17 | usage = "perf trace -s syscall-counts-by-pid.py [comm]\n"; |
| 18 | 18 | ||
| 19 | for_comm = None | 19 | for_comm = None |
| 20 | for_pid = None | ||
| 20 | 21 | ||
| 21 | if len(sys.argv) > 2: | 22 | if len(sys.argv) > 2: |
| 22 | sys.exit(usage) | 23 | sys.exit(usage) |
| 23 | 24 | ||
| 24 | if len(sys.argv) > 1: | 25 | if len(sys.argv) > 1: |
| 25 | for_comm = sys.argv[1] | 26 | try: |
| 27 | for_pid = int(sys.argv[1]) | ||
| 28 | except: | ||
| 29 | for_comm = sys.argv[1] | ||
| 26 | 30 | ||
| 27 | syscalls = autodict() | 31 | syscalls = autodict() |
| 28 | 32 | ||
| 29 | def trace_begin(): | 33 | def trace_begin(): |
| 30 | pass | 34 | print "Press control+C to stop and show the summary" |
| 31 | 35 | ||
| 32 | def trace_end(): | 36 | def trace_end(): |
| 33 | print_syscall_totals() | 37 | print_syscall_totals() |
| @@ -35,9 +39,10 @@ def trace_end(): | |||
| 35 | def raw_syscalls__sys_enter(event_name, context, common_cpu, | 39 | def raw_syscalls__sys_enter(event_name, context, common_cpu, |
| 36 | common_secs, common_nsecs, common_pid, common_comm, | 40 | common_secs, common_nsecs, common_pid, common_comm, |
| 37 | id, args): | 41 | id, args): |
| 38 | if for_comm is not None: | 42 | |
| 39 | if common_comm != for_comm: | 43 | if (for_comm and common_comm != for_comm) or \ |
| 40 | return | 44 | (for_pid and common_pid != for_pid ): |
| 45 | return | ||
| 41 | try: | 46 | try: |
| 42 | syscalls[common_comm][common_pid][id] += 1 | 47 | syscalls[common_comm][common_pid][id] += 1 |
| 43 | except TypeError: | 48 | except TypeError: |
| @@ -61,4 +66,4 @@ def print_syscall_totals(): | |||
| 61 | id_keys = syscalls[comm][pid].keys() | 66 | id_keys = syscalls[comm][pid].keys() |
| 62 | for id, val in sorted(syscalls[comm][pid].iteritems(), \ | 67 | for id, val in sorted(syscalls[comm][pid].iteritems(), \ |
| 63 | key = lambda(k, v): (v, k), reverse = True): | 68 | key = lambda(k, v): (v, k), reverse = True): |
| 64 | print " %-38d %10d\n" % (id, val), | 69 | print " %-38s %10d\n" % (syscall_name(id), val), |
diff --git a/tools/perf/scripts/python/syscall-counts.py b/tools/perf/scripts/python/syscall-counts.py index f977e85ff049..ea183dc82d29 100644 --- a/tools/perf/scripts/python/syscall-counts.py +++ b/tools/perf/scripts/python/syscall-counts.py | |||
| @@ -13,6 +13,7 @@ sys.path.append(os.environ['PERF_EXEC_PATH'] + \ | |||
| 13 | 13 | ||
| 14 | from perf_trace_context import * | 14 | from perf_trace_context import * |
| 15 | from Core import * | 15 | from Core import * |
| 16 | from Util import syscall_name | ||
| 16 | 17 | ||
| 17 | usage = "perf trace -s syscall-counts.py [comm]\n"; | 18 | usage = "perf trace -s syscall-counts.py [comm]\n"; |
| 18 | 19 | ||
| @@ -27,7 +28,7 @@ if len(sys.argv) > 1: | |||
| 27 | syscalls = autodict() | 28 | syscalls = autodict() |
| 28 | 29 | ||
| 29 | def trace_begin(): | 30 | def trace_begin(): |
| 30 | pass | 31 | print "Press control+C to stop and show the summary" |
| 31 | 32 | ||
| 32 | def trace_end(): | 33 | def trace_end(): |
| 33 | print_syscall_totals() | 34 | print_syscall_totals() |
| @@ -55,4 +56,4 @@ def print_syscall_totals(): | |||
| 55 | 56 | ||
| 56 | for id, val in sorted(syscalls.iteritems(), key = lambda(k, v): (v, k), \ | 57 | for id, val in sorted(syscalls.iteritems(), key = lambda(k, v): (v, k), \ |
| 57 | reverse = True): | 58 | reverse = True): |
| 58 | print "%-40d %10d\n" % (id, val), | 59 | print "%-40s %10d\n" % (syscall_name(id), val), |
diff --git a/tools/perf/util/debug.c b/tools/perf/util/debug.c index f9c7e3ad1aa7..c8d81b00089d 100644 --- a/tools/perf/util/debug.c +++ b/tools/perf/util/debug.c | |||
| @@ -12,8 +12,8 @@ | |||
| 12 | #include "debug.h" | 12 | #include "debug.h" |
| 13 | #include "util.h" | 13 | #include "util.h" |
| 14 | 14 | ||
| 15 | int verbose = 0; | 15 | int verbose; |
| 16 | bool dump_trace = false; | 16 | bool dump_trace = false, quiet = false; |
| 17 | 17 | ||
| 18 | int eprintf(int level, const char *fmt, ...) | 18 | int eprintf(int level, const char *fmt, ...) |
| 19 | { | 19 | { |
diff --git a/tools/perf/util/debug.h b/tools/perf/util/debug.h index 7a17ee061bcb..7b514082bbaf 100644 --- a/tools/perf/util/debug.h +++ b/tools/perf/util/debug.h | |||
| @@ -6,7 +6,7 @@ | |||
| 6 | #include "event.h" | 6 | #include "event.h" |
| 7 | 7 | ||
| 8 | extern int verbose; | 8 | extern int verbose; |
| 9 | extern bool dump_trace; | 9 | extern bool quiet, dump_trace; |
| 10 | 10 | ||
| 11 | int dump_printf(const char *fmt, ...) __attribute__((format(printf, 1, 2))); | 11 | int dump_printf(const char *fmt, ...) __attribute__((format(printf, 1, 2))); |
| 12 | void trace_event(event_t *event); | 12 | void trace_event(event_t *event); |
diff --git a/tools/perf/util/map.h b/tools/perf/util/map.h index 78575796d5f3..b397c0383728 100644 --- a/tools/perf/util/map.h +++ b/tools/perf/util/map.h | |||
| @@ -215,6 +215,16 @@ struct symbol *map_groups__find_function_by_name(struct map_groups *self, | |||
| 215 | return map_groups__find_symbol_by_name(self, MAP__FUNCTION, name, mapp, filter); | 215 | return map_groups__find_symbol_by_name(self, MAP__FUNCTION, name, mapp, filter); |
| 216 | } | 216 | } |
| 217 | 217 | ||
| 218 | static inline | ||
| 219 | struct symbol *machine__find_kernel_function_by_name(struct machine *self, | ||
| 220 | const char *name, | ||
| 221 | struct map **mapp, | ||
| 222 | symbol_filter_t filter) | ||
| 223 | { | ||
| 224 | return map_groups__find_function_by_name(&self->kmaps, name, mapp, | ||
| 225 | filter); | ||
| 226 | } | ||
| 227 | |||
| 218 | int map_groups__fixup_overlappings(struct map_groups *self, struct map *map, | 228 | int map_groups__fixup_overlappings(struct map_groups *self, struct map *map, |
| 219 | int verbose, FILE *fp); | 229 | int verbose, FILE *fp); |
| 220 | 230 | ||
diff --git a/tools/perf/util/probe-event.c b/tools/perf/util/probe-event.c index fcc16e4349df..3b6a5297bf16 100644 --- a/tools/perf/util/probe-event.c +++ b/tools/perf/util/probe-event.c | |||
| @@ -74,10 +74,9 @@ static int e_snprintf(char *str, size_t size, const char *format, ...) | |||
| 74 | static char *synthesize_perf_probe_point(struct perf_probe_point *pp); | 74 | static char *synthesize_perf_probe_point(struct perf_probe_point *pp); |
| 75 | static struct machine machine; | 75 | static struct machine machine; |
| 76 | 76 | ||
| 77 | /* Initialize symbol maps and path of vmlinux */ | 77 | /* Initialize symbol maps and path of vmlinux/modules */ |
| 78 | static int init_vmlinux(void) | 78 | static int init_vmlinux(void) |
| 79 | { | 79 | { |
| 80 | struct dso *kernel; | ||
| 81 | int ret; | 80 | int ret; |
| 82 | 81 | ||
| 83 | symbol_conf.sort_by_name = true; | 82 | symbol_conf.sort_by_name = true; |
| @@ -91,33 +90,61 @@ static int init_vmlinux(void) | |||
| 91 | goto out; | 90 | goto out; |
| 92 | } | 91 | } |
| 93 | 92 | ||
| 94 | ret = machine__init(&machine, "/", 0); | 93 | ret = machine__init(&machine, "", HOST_KERNEL_ID); |
| 95 | if (ret < 0) | 94 | if (ret < 0) |
| 96 | goto out; | 95 | goto out; |
| 97 | 96 | ||
| 98 | kernel = dso__new_kernel(symbol_conf.vmlinux_name); | 97 | if (machine__create_kernel_maps(&machine) < 0) { |
| 99 | if (kernel == NULL) | 98 | pr_debug("machine__create_kernel_maps "); |
| 100 | die("Failed to create kernel dso."); | 99 | goto out; |
| 101 | 100 | } | |
| 102 | ret = __machine__create_kernel_maps(&machine, kernel); | ||
| 103 | if (ret < 0) | ||
| 104 | pr_debug("Failed to create kernel maps.\n"); | ||
| 105 | |||
| 106 | out: | 101 | out: |
| 107 | if (ret < 0) | 102 | if (ret < 0) |
| 108 | pr_warning("Failed to init vmlinux path.\n"); | 103 | pr_warning("Failed to init vmlinux path.\n"); |
| 109 | return ret; | 104 | return ret; |
| 110 | } | 105 | } |
| 111 | 106 | ||
| 107 | static struct symbol *__find_kernel_function_by_name(const char *name, | ||
| 108 | struct map **mapp) | ||
| 109 | { | ||
| 110 | return machine__find_kernel_function_by_name(&machine, name, mapp, | ||
| 111 | NULL); | ||
| 112 | } | ||
| 113 | |||
| 114 | const char *kernel_get_module_path(const char *module) | ||
| 115 | { | ||
| 116 | struct dso *dso; | ||
| 117 | |||
| 118 | if (module) { | ||
| 119 | list_for_each_entry(dso, &machine.kernel_dsos, node) { | ||
| 120 | if (strncmp(dso->short_name + 1, module, | ||
| 121 | dso->short_name_len - 2) == 0) | ||
| 122 | goto found; | ||
| 123 | } | ||
| 124 | pr_debug("Failed to find module %s.\n", module); | ||
| 125 | return NULL; | ||
| 126 | } else { | ||
| 127 | dso = machine.vmlinux_maps[MAP__FUNCTION]->dso; | ||
| 128 | if (dso__load_vmlinux_path(dso, | ||
| 129 | machine.vmlinux_maps[MAP__FUNCTION], NULL) < 0) { | ||
| 130 | pr_debug("Failed to load kernel map.\n"); | ||
| 131 | return NULL; | ||
| 132 | } | ||
| 133 | } | ||
| 134 | found: | ||
| 135 | return dso->long_name; | ||
| 136 | } | ||
| 137 | |||
| 112 | #ifdef DWARF_SUPPORT | 138 | #ifdef DWARF_SUPPORT |
| 113 | static int open_vmlinux(void) | 139 | static int open_vmlinux(const char *module) |
| 114 | { | 140 | { |
| 115 | if (map__load(machine.vmlinux_maps[MAP__FUNCTION], NULL) < 0) { | 141 | const char *path = kernel_get_module_path(module); |
| 116 | pr_debug("Failed to load kernel map.\n"); | 142 | if (!path) { |
| 117 | return -EINVAL; | 143 | pr_err("Failed to find path of %s module", module ?: "kernel"); |
| 144 | return -ENOENT; | ||
| 118 | } | 145 | } |
| 119 | pr_debug("Try to open %s\n", machine.vmlinux_maps[MAP__FUNCTION]->dso->long_name); | 146 | pr_debug("Try to open %s\n", path); |
| 120 | return open(machine.vmlinux_maps[MAP__FUNCTION]->dso->long_name, O_RDONLY); | 147 | return open(path, O_RDONLY); |
| 121 | } | 148 | } |
| 122 | 149 | ||
| 123 | /* | 150 | /* |
| @@ -125,20 +152,19 @@ static int open_vmlinux(void) | |||
| 125 | * Currently only handles kprobes. | 152 | * Currently only handles kprobes. |
| 126 | */ | 153 | */ |
| 127 | static int kprobe_convert_to_perf_probe(struct probe_trace_point *tp, | 154 | static int kprobe_convert_to_perf_probe(struct probe_trace_point *tp, |
| 128 | struct perf_probe_point *pp) | 155 | struct perf_probe_point *pp) |
| 129 | { | 156 | { |
| 130 | struct symbol *sym; | 157 | struct symbol *sym; |
| 131 | int fd, ret = -ENOENT; | 158 | struct map *map; |
| 159 | u64 addr; | ||
| 160 | int ret = -ENOENT; | ||
| 132 | 161 | ||
| 133 | sym = map__find_symbol_by_name(machine.vmlinux_maps[MAP__FUNCTION], | 162 | sym = __find_kernel_function_by_name(tp->symbol, &map); |
| 134 | tp->symbol, NULL); | ||
| 135 | if (sym) { | 163 | if (sym) { |
| 136 | fd = open_vmlinux(); | 164 | addr = map->unmap_ip(map, sym->start + tp->offset); |
| 137 | if (fd >= 0) { | 165 | pr_debug("try to find %s+%ld@%llx\n", tp->symbol, |
| 138 | ret = find_perf_probe_point(fd, | 166 | tp->offset, addr); |
| 139 | sym->start + tp->offset, pp); | 167 | ret = find_perf_probe_point((unsigned long)addr, pp); |
| 140 | close(fd); | ||
| 141 | } | ||
| 142 | } | 168 | } |
| 143 | if (ret <= 0) { | 169 | if (ret <= 0) { |
| 144 | pr_debug("Failed to find corresponding probes from " | 170 | pr_debug("Failed to find corresponding probes from " |
| @@ -156,12 +182,12 @@ static int kprobe_convert_to_perf_probe(struct probe_trace_point *tp, | |||
| 156 | /* Try to find perf_probe_event with debuginfo */ | 182 | /* Try to find perf_probe_event with debuginfo */ |
| 157 | static int try_to_find_probe_trace_events(struct perf_probe_event *pev, | 183 | static int try_to_find_probe_trace_events(struct perf_probe_event *pev, |
| 158 | struct probe_trace_event **tevs, | 184 | struct probe_trace_event **tevs, |
| 159 | int max_tevs) | 185 | int max_tevs, const char *module) |
| 160 | { | 186 | { |
| 161 | bool need_dwarf = perf_probe_event_need_dwarf(pev); | 187 | bool need_dwarf = perf_probe_event_need_dwarf(pev); |
| 162 | int fd, ntevs; | 188 | int fd, ntevs; |
| 163 | 189 | ||
| 164 | fd = open_vmlinux(); | 190 | fd = open_vmlinux(module); |
| 165 | if (fd < 0) { | 191 | if (fd < 0) { |
| 166 | if (need_dwarf) { | 192 | if (need_dwarf) { |
| 167 | pr_warning("Failed to open debuginfo file.\n"); | 193 | pr_warning("Failed to open debuginfo file.\n"); |
| @@ -300,7 +326,7 @@ error: | |||
| 300 | * Show line-range always requires debuginfo to find source file and | 326 | * Show line-range always requires debuginfo to find source file and |
| 301 | * line number. | 327 | * line number. |
| 302 | */ | 328 | */ |
| 303 | int show_line_range(struct line_range *lr) | 329 | int show_line_range(struct line_range *lr, const char *module) |
| 304 | { | 330 | { |
| 305 | int l = 1; | 331 | int l = 1; |
| 306 | struct line_node *ln; | 332 | struct line_node *ln; |
| @@ -313,7 +339,7 @@ int show_line_range(struct line_range *lr) | |||
| 313 | if (ret < 0) | 339 | if (ret < 0) |
| 314 | return ret; | 340 | return ret; |
| 315 | 341 | ||
| 316 | fd = open_vmlinux(); | 342 | fd = open_vmlinux(module); |
| 317 | if (fd < 0) { | 343 | if (fd < 0) { |
| 318 | pr_warning("Failed to open debuginfo file.\n"); | 344 | pr_warning("Failed to open debuginfo file.\n"); |
| 319 | return fd; | 345 | return fd; |
| @@ -378,11 +404,84 @@ end: | |||
| 378 | return ret; | 404 | return ret; |
| 379 | } | 405 | } |
| 380 | 406 | ||
| 407 | static int show_available_vars_at(int fd, struct perf_probe_event *pev, | ||
| 408 | int max_vls, bool externs) | ||
| 409 | { | ||
| 410 | char *buf; | ||
| 411 | int ret, i; | ||
| 412 | struct str_node *node; | ||
| 413 | struct variable_list *vls = NULL, *vl; | ||
| 414 | |||
| 415 | buf = synthesize_perf_probe_point(&pev->point); | ||
| 416 | if (!buf) | ||
| 417 | return -EINVAL; | ||
| 418 | pr_debug("Searching variables at %s\n", buf); | ||
| 419 | |||
| 420 | ret = find_available_vars_at(fd, pev, &vls, max_vls, externs); | ||
| 421 | if (ret > 0) { | ||
| 422 | /* Some variables were found */ | ||
| 423 | fprintf(stdout, "Available variables at %s\n", buf); | ||
| 424 | for (i = 0; i < ret; i++) { | ||
| 425 | vl = &vls[i]; | ||
| 426 | /* | ||
| 427 | * A probe point might be converted to | ||
| 428 | * several trace points. | ||
| 429 | */ | ||
| 430 | fprintf(stdout, "\t@<%s+%lu>\n", vl->point.symbol, | ||
| 431 | vl->point.offset); | ||
| 432 | free(vl->point.symbol); | ||
| 433 | if (vl->vars) { | ||
| 434 | strlist__for_each(node, vl->vars) | ||
| 435 | fprintf(stdout, "\t\t%s\n", node->s); | ||
| 436 | strlist__delete(vl->vars); | ||
| 437 | } else | ||
| 438 | fprintf(stdout, "(No variables)\n"); | ||
| 439 | } | ||
| 440 | free(vls); | ||
| 441 | } else | ||
| 442 | pr_err("Failed to find variables at %s (%d)\n", buf, ret); | ||
| 443 | |||
| 444 | free(buf); | ||
| 445 | return ret; | ||
| 446 | } | ||
| 447 | |||
| 448 | /* Show available variables on given probe point */ | ||
| 449 | int show_available_vars(struct perf_probe_event *pevs, int npevs, | ||
| 450 | int max_vls, const char *module, bool externs) | ||
| 451 | { | ||
| 452 | int i, fd, ret = 0; | ||
| 453 | |||
| 454 | ret = init_vmlinux(); | ||
| 455 | if (ret < 0) | ||
| 456 | return ret; | ||
| 457 | |||
| 458 | fd = open_vmlinux(module); | ||
| 459 | if (fd < 0) { | ||
| 460 | pr_warning("Failed to open debuginfo file.\n"); | ||
| 461 | return fd; | ||
| 462 | } | ||
| 463 | |||
| 464 | setup_pager(); | ||
| 465 | |||
| 466 | for (i = 0; i < npevs && ret >= 0; i++) | ||
| 467 | ret = show_available_vars_at(fd, &pevs[i], max_vls, externs); | ||
| 468 | |||
| 469 | close(fd); | ||
| 470 | return ret; | ||
| 471 | } | ||
| 472 | |||
| 381 | #else /* !DWARF_SUPPORT */ | 473 | #else /* !DWARF_SUPPORT */ |
| 382 | 474 | ||
| 383 | static int kprobe_convert_to_perf_probe(struct probe_trace_point *tp, | 475 | static int kprobe_convert_to_perf_probe(struct probe_trace_point *tp, |
| 384 | struct perf_probe_point *pp) | 476 | struct perf_probe_point *pp) |
| 385 | { | 477 | { |
| 478 | struct symbol *sym; | ||
| 479 | |||
| 480 | sym = __find_kernel_function_by_name(tp->symbol, NULL); | ||
| 481 | if (!sym) { | ||
| 482 | pr_err("Failed to find symbol %s in kernel.\n", tp->symbol); | ||
| 483 | return -ENOENT; | ||
| 484 | } | ||
| 386 | pp->function = strdup(tp->symbol); | 485 | pp->function = strdup(tp->symbol); |
| 387 | if (pp->function == NULL) | 486 | if (pp->function == NULL) |
| 388 | return -ENOMEM; | 487 | return -ENOMEM; |
| @@ -394,7 +493,7 @@ static int kprobe_convert_to_perf_probe(struct probe_trace_point *tp, | |||
| 394 | 493 | ||
| 395 | static int try_to_find_probe_trace_events(struct perf_probe_event *pev, | 494 | static int try_to_find_probe_trace_events(struct perf_probe_event *pev, |
| 396 | struct probe_trace_event **tevs __unused, | 495 | struct probe_trace_event **tevs __unused, |
| 397 | int max_tevs __unused) | 496 | int max_tevs __unused, const char *mod __unused) |
| 398 | { | 497 | { |
| 399 | if (perf_probe_event_need_dwarf(pev)) { | 498 | if (perf_probe_event_need_dwarf(pev)) { |
| 400 | pr_warning("Debuginfo-analysis is not supported.\n"); | 499 | pr_warning("Debuginfo-analysis is not supported.\n"); |
| @@ -403,12 +502,19 @@ static int try_to_find_probe_trace_events(struct perf_probe_event *pev, | |||
| 403 | return 0; | 502 | return 0; |
| 404 | } | 503 | } |
| 405 | 504 | ||
| 406 | int show_line_range(struct line_range *lr __unused) | 505 | int show_line_range(struct line_range *lr __unused, const char *module __unused) |
| 407 | { | 506 | { |
| 408 | pr_warning("Debuginfo-analysis is not supported.\n"); | 507 | pr_warning("Debuginfo-analysis is not supported.\n"); |
| 409 | return -ENOSYS; | 508 | return -ENOSYS; |
| 410 | } | 509 | } |
| 411 | 510 | ||
| 511 | int show_available_vars(struct perf_probe_event *pevs __unused, | ||
| 512 | int npevs __unused, int max_vls __unused, | ||
| 513 | const char *module __unused, bool externs __unused) | ||
| 514 | { | ||
| 515 | pr_warning("Debuginfo-analysis is not supported.\n"); | ||
| 516 | return -ENOSYS; | ||
| 517 | } | ||
| 412 | #endif | 518 | #endif |
| 413 | 519 | ||
| 414 | int parse_line_range_desc(const char *arg, struct line_range *lr) | 520 | int parse_line_range_desc(const char *arg, struct line_range *lr) |
| @@ -1087,7 +1193,7 @@ error: | |||
| 1087 | } | 1193 | } |
| 1088 | 1194 | ||
| 1089 | static int convert_to_perf_probe_event(struct probe_trace_event *tev, | 1195 | static int convert_to_perf_probe_event(struct probe_trace_event *tev, |
| 1090 | struct perf_probe_event *pev) | 1196 | struct perf_probe_event *pev) |
| 1091 | { | 1197 | { |
| 1092 | char buf[64] = ""; | 1198 | char buf[64] = ""; |
| 1093 | int i, ret; | 1199 | int i, ret; |
| @@ -1516,14 +1622,14 @@ static int __add_probe_trace_events(struct perf_probe_event *pev, | |||
| 1516 | 1622 | ||
| 1517 | static int convert_to_probe_trace_events(struct perf_probe_event *pev, | 1623 | static int convert_to_probe_trace_events(struct perf_probe_event *pev, |
| 1518 | struct probe_trace_event **tevs, | 1624 | struct probe_trace_event **tevs, |
| 1519 | int max_tevs) | 1625 | int max_tevs, const char *module) |
| 1520 | { | 1626 | { |
| 1521 | struct symbol *sym; | 1627 | struct symbol *sym; |
| 1522 | int ret = 0, i; | 1628 | int ret = 0, i; |
| 1523 | struct probe_trace_event *tev; | 1629 | struct probe_trace_event *tev; |
| 1524 | 1630 | ||
| 1525 | /* Convert perf_probe_event with debuginfo */ | 1631 | /* Convert perf_probe_event with debuginfo */ |
| 1526 | ret = try_to_find_probe_trace_events(pev, tevs, max_tevs); | 1632 | ret = try_to_find_probe_trace_events(pev, tevs, max_tevs, module); |
| 1527 | if (ret != 0) | 1633 | if (ret != 0) |
| 1528 | return ret; | 1634 | return ret; |
| 1529 | 1635 | ||
| @@ -1572,8 +1678,7 @@ static int convert_to_probe_trace_events(struct perf_probe_event *pev, | |||
| 1572 | } | 1678 | } |
| 1573 | 1679 | ||
| 1574 | /* Currently just checking function name from symbol map */ | 1680 | /* Currently just checking function name from symbol map */ |
| 1575 | sym = map__find_symbol_by_name(machine.vmlinux_maps[MAP__FUNCTION], | 1681 | sym = __find_kernel_function_by_name(tev->point.symbol, NULL); |
| 1576 | tev->point.symbol, NULL); | ||
| 1577 | if (!sym) { | 1682 | if (!sym) { |
| 1578 | pr_warning("Kernel symbol \'%s\' not found.\n", | 1683 | pr_warning("Kernel symbol \'%s\' not found.\n", |
| 1579 | tev->point.symbol); | 1684 | tev->point.symbol); |
| @@ -1596,7 +1701,7 @@ struct __event_package { | |||
| 1596 | }; | 1701 | }; |
| 1597 | 1702 | ||
| 1598 | int add_perf_probe_events(struct perf_probe_event *pevs, int npevs, | 1703 | int add_perf_probe_events(struct perf_probe_event *pevs, int npevs, |
| 1599 | bool force_add, int max_tevs) | 1704 | int max_tevs, const char *module, bool force_add) |
| 1600 | { | 1705 | { |
| 1601 | int i, j, ret; | 1706 | int i, j, ret; |
| 1602 | struct __event_package *pkgs; | 1707 | struct __event_package *pkgs; |
| @@ -1617,7 +1722,9 @@ int add_perf_probe_events(struct perf_probe_event *pevs, int npevs, | |||
| 1617 | pkgs[i].pev = &pevs[i]; | 1722 | pkgs[i].pev = &pevs[i]; |
| 1618 | /* Convert with or without debuginfo */ | 1723 | /* Convert with or without debuginfo */ |
| 1619 | ret = convert_to_probe_trace_events(pkgs[i].pev, | 1724 | ret = convert_to_probe_trace_events(pkgs[i].pev, |
| 1620 | &pkgs[i].tevs, max_tevs); | 1725 | &pkgs[i].tevs, |
| 1726 | max_tevs, | ||
| 1727 | module); | ||
| 1621 | if (ret < 0) | 1728 | if (ret < 0) |
| 1622 | goto end; | 1729 | goto end; |
| 1623 | pkgs[i].ntevs = ret; | 1730 | pkgs[i].ntevs = ret; |
diff --git a/tools/perf/util/probe-event.h b/tools/perf/util/probe-event.h index 5af39243a25b..5accbedfea37 100644 --- a/tools/perf/util/probe-event.h +++ b/tools/perf/util/probe-event.h | |||
| @@ -90,6 +90,12 @@ struct line_range { | |||
| 90 | struct list_head line_list; /* Visible lines */ | 90 | struct list_head line_list; /* Visible lines */ |
| 91 | }; | 91 | }; |
| 92 | 92 | ||
| 93 | /* List of variables */ | ||
| 94 | struct variable_list { | ||
| 95 | struct probe_trace_point point; /* Actual probepoint */ | ||
| 96 | struct strlist *vars; /* Available variables */ | ||
| 97 | }; | ||
| 98 | |||
| 93 | /* Command string to events */ | 99 | /* Command string to events */ |
| 94 | extern int parse_perf_probe_command(const char *cmd, | 100 | extern int parse_perf_probe_command(const char *cmd, |
| 95 | struct perf_probe_event *pev); | 101 | struct perf_probe_event *pev); |
| @@ -109,12 +115,18 @@ extern void clear_perf_probe_event(struct perf_probe_event *pev); | |||
| 109 | /* Command string to line-range */ | 115 | /* Command string to line-range */ |
| 110 | extern int parse_line_range_desc(const char *cmd, struct line_range *lr); | 116 | extern int parse_line_range_desc(const char *cmd, struct line_range *lr); |
| 111 | 117 | ||
| 118 | /* Internal use: Return kernel/module path */ | ||
| 119 | extern const char *kernel_get_module_path(const char *module); | ||
| 112 | 120 | ||
| 113 | extern int add_perf_probe_events(struct perf_probe_event *pevs, int npevs, | 121 | extern int add_perf_probe_events(struct perf_probe_event *pevs, int npevs, |
| 114 | bool force_add, int max_probe_points); | 122 | int max_probe_points, const char *module, |
| 123 | bool force_add); | ||
| 115 | extern int del_perf_probe_events(struct strlist *dellist); | 124 | extern int del_perf_probe_events(struct strlist *dellist); |
| 116 | extern int show_perf_probe_events(void); | 125 | extern int show_perf_probe_events(void); |
| 117 | extern int show_line_range(struct line_range *lr); | 126 | extern int show_line_range(struct line_range *lr, const char *module); |
| 127 | extern int show_available_vars(struct perf_probe_event *pevs, int npevs, | ||
| 128 | int max_probe_points, const char *module, | ||
| 129 | bool externs); | ||
| 118 | 130 | ||
| 119 | 131 | ||
| 120 | /* Maximum index number of event-name postfix */ | 132 | /* Maximum index number of event-name postfix */ |
diff --git a/tools/perf/util/probe-finder.c b/tools/perf/util/probe-finder.c index 32b81f707ff5..3991d73d1cff 100644 --- a/tools/perf/util/probe-finder.c +++ b/tools/perf/util/probe-finder.c | |||
| @@ -116,6 +116,101 @@ static void line_list__free(struct list_head *head) | |||
| 116 | } | 116 | } |
| 117 | } | 117 | } |
| 118 | 118 | ||
| 119 | /* Dwarf FL wrappers */ | ||
| 120 | |||
| 121 | static int __linux_kernel_find_elf(Dwfl_Module *mod, | ||
| 122 | void **userdata, | ||
| 123 | const char *module_name, | ||
| 124 | Dwarf_Addr base, | ||
| 125 | char **file_name, Elf **elfp) | ||
| 126 | { | ||
| 127 | int fd; | ||
| 128 | const char *path = kernel_get_module_path(module_name); | ||
| 129 | |||
| 130 | if (path) { | ||
| 131 | fd = open(path, O_RDONLY); | ||
| 132 | if (fd >= 0) { | ||
| 133 | *file_name = strdup(path); | ||
| 134 | return fd; | ||
| 135 | } | ||
| 136 | } | ||
| 137 | /* If failed, try to call standard method */ | ||
| 138 | return dwfl_linux_kernel_find_elf(mod, userdata, module_name, base, | ||
| 139 | file_name, elfp); | ||
| 140 | } | ||
| 141 | |||
| 142 | static char *debuginfo_path; /* Currently dummy */ | ||
| 143 | |||
| 144 | static const Dwfl_Callbacks offline_callbacks = { | ||
| 145 | .find_debuginfo = dwfl_standard_find_debuginfo, | ||
| 146 | .debuginfo_path = &debuginfo_path, | ||
| 147 | |||
| 148 | .section_address = dwfl_offline_section_address, | ||
| 149 | |||
| 150 | /* We use this table for core files too. */ | ||
| 151 | .find_elf = dwfl_build_id_find_elf, | ||
| 152 | }; | ||
| 153 | |||
| 154 | static const Dwfl_Callbacks kernel_callbacks = { | ||
| 155 | .find_debuginfo = dwfl_standard_find_debuginfo, | ||
| 156 | .debuginfo_path = &debuginfo_path, | ||
| 157 | |||
| 158 | .find_elf = __linux_kernel_find_elf, | ||
| 159 | .section_address = dwfl_linux_kernel_module_section_address, | ||
| 160 | }; | ||
| 161 | |||
| 162 | /* Get a Dwarf from offline image */ | ||
| 163 | static Dwarf *dwfl_init_offline_dwarf(int fd, Dwfl **dwflp, Dwarf_Addr *bias) | ||
| 164 | { | ||
| 165 | Dwfl_Module *mod; | ||
| 166 | Dwarf *dbg = NULL; | ||
| 167 | |||
| 168 | if (!dwflp) | ||
| 169 | return NULL; | ||
| 170 | |||
| 171 | *dwflp = dwfl_begin(&offline_callbacks); | ||
| 172 | if (!*dwflp) | ||
| 173 | return NULL; | ||
| 174 | |||
| 175 | mod = dwfl_report_offline(*dwflp, "", "", fd); | ||
| 176 | if (!mod) | ||
| 177 | goto error; | ||
| 178 | |||
| 179 | dbg = dwfl_module_getdwarf(mod, bias); | ||
| 180 | if (!dbg) { | ||
| 181 | error: | ||
| 182 | dwfl_end(*dwflp); | ||
| 183 | *dwflp = NULL; | ||
| 184 | } | ||
| 185 | return dbg; | ||
| 186 | } | ||
| 187 | |||
| 188 | /* Get a Dwarf from live kernel image */ | ||
| 189 | static Dwarf *dwfl_init_live_kernel_dwarf(Dwarf_Addr addr, Dwfl **dwflp, | ||
| 190 | Dwarf_Addr *bias) | ||
| 191 | { | ||
| 192 | Dwarf *dbg; | ||
| 193 | |||
| 194 | if (!dwflp) | ||
| 195 | return NULL; | ||
| 196 | |||
| 197 | *dwflp = dwfl_begin(&kernel_callbacks); | ||
| 198 | if (!*dwflp) | ||
| 199 | return NULL; | ||
| 200 | |||
| 201 | /* Load the kernel dwarves: Don't care the result here */ | ||
| 202 | dwfl_linux_kernel_report_kernel(*dwflp); | ||
| 203 | dwfl_linux_kernel_report_modules(*dwflp); | ||
| 204 | |||
| 205 | dbg = dwfl_addrdwarf(*dwflp, addr, bias); | ||
| 206 | /* Here, check whether we could get a real dwarf */ | ||
| 207 | if (!dbg) { | ||
| 208 | dwfl_end(*dwflp); | ||
| 209 | *dwflp = NULL; | ||
| 210 | } | ||
| 211 | return dbg; | ||
| 212 | } | ||
| 213 | |||
| 119 | /* Dwarf wrappers */ | 214 | /* Dwarf wrappers */ |
| 120 | 215 | ||
| 121 | /* Find the realpath of the target file. */ | 216 | /* Find the realpath of the target file. */ |
| @@ -160,26 +255,44 @@ static bool die_compare_name(Dwarf_Die *dw_die, const char *tname) | |||
| 160 | return name ? (strcmp(tname, name) == 0) : false; | 255 | return name ? (strcmp(tname, name) == 0) : false; |
| 161 | } | 256 | } |
| 162 | 257 | ||
| 163 | /* Get type die, but skip qualifiers and typedef */ | 258 | /* Get type die */ |
| 164 | static Dwarf_Die *die_get_real_type(Dwarf_Die *vr_die, Dwarf_Die *die_mem) | 259 | static Dwarf_Die *die_get_type(Dwarf_Die *vr_die, Dwarf_Die *die_mem) |
| 165 | { | 260 | { |
| 166 | Dwarf_Attribute attr; | 261 | Dwarf_Attribute attr; |
| 262 | |||
| 263 | if (dwarf_attr_integrate(vr_die, DW_AT_type, &attr) && | ||
| 264 | dwarf_formref_die(&attr, die_mem)) | ||
| 265 | return die_mem; | ||
| 266 | else | ||
| 267 | return NULL; | ||
| 268 | } | ||
| 269 | |||
| 270 | /* Get a type die, but skip qualifiers */ | ||
| 271 | static Dwarf_Die *__die_get_real_type(Dwarf_Die *vr_die, Dwarf_Die *die_mem) | ||
| 272 | { | ||
| 167 | int tag; | 273 | int tag; |
| 168 | 274 | ||
| 169 | do { | 275 | do { |
| 170 | if (dwarf_attr(vr_die, DW_AT_type, &attr) == NULL || | 276 | vr_die = die_get_type(vr_die, die_mem); |
| 171 | dwarf_formref_die(&attr, die_mem) == NULL) | 277 | if (!vr_die) |
| 172 | return NULL; | 278 | break; |
| 173 | 279 | tag = dwarf_tag(vr_die); | |
| 174 | tag = dwarf_tag(die_mem); | ||
| 175 | vr_die = die_mem; | ||
| 176 | } while (tag == DW_TAG_const_type || | 280 | } while (tag == DW_TAG_const_type || |
| 177 | tag == DW_TAG_restrict_type || | 281 | tag == DW_TAG_restrict_type || |
| 178 | tag == DW_TAG_volatile_type || | 282 | tag == DW_TAG_volatile_type || |
| 179 | tag == DW_TAG_shared_type || | 283 | tag == DW_TAG_shared_type); |
| 180 | tag == DW_TAG_typedef); | 284 | |
| 285 | return vr_die; | ||
| 286 | } | ||
| 181 | 287 | ||
| 182 | return die_mem; | 288 | /* Get a type die, but skip qualifiers and typedef */ |
| 289 | static Dwarf_Die *die_get_real_type(Dwarf_Die *vr_die, Dwarf_Die *die_mem) | ||
| 290 | { | ||
| 291 | do { | ||
| 292 | vr_die = __die_get_real_type(vr_die, die_mem); | ||
| 293 | } while (vr_die && dwarf_tag(vr_die) == DW_TAG_typedef); | ||
| 294 | |||
| 295 | return vr_die; | ||
| 183 | } | 296 | } |
| 184 | 297 | ||
| 185 | static bool die_is_signed_type(Dwarf_Die *tp_die) | 298 | static bool die_is_signed_type(Dwarf_Die *tp_die) |
| @@ -320,25 +433,35 @@ static Dwarf_Die *die_find_inlinefunc(Dwarf_Die *sp_die, Dwarf_Addr addr, | |||
| 320 | return die_find_child(sp_die, __die_find_inline_cb, &addr, die_mem); | 433 | return die_find_child(sp_die, __die_find_inline_cb, &addr, die_mem); |
| 321 | } | 434 | } |
| 322 | 435 | ||
| 436 | struct __find_variable_param { | ||
| 437 | const char *name; | ||
| 438 | Dwarf_Addr addr; | ||
| 439 | }; | ||
| 440 | |||
| 323 | static int __die_find_variable_cb(Dwarf_Die *die_mem, void *data) | 441 | static int __die_find_variable_cb(Dwarf_Die *die_mem, void *data) |
| 324 | { | 442 | { |
| 325 | const char *name = data; | 443 | struct __find_variable_param *fvp = data; |
| 326 | int tag; | 444 | int tag; |
| 327 | 445 | ||
| 328 | tag = dwarf_tag(die_mem); | 446 | tag = dwarf_tag(die_mem); |
| 329 | if ((tag == DW_TAG_formal_parameter || | 447 | if ((tag == DW_TAG_formal_parameter || |
| 330 | tag == DW_TAG_variable) && | 448 | tag == DW_TAG_variable) && |
| 331 | die_compare_name(die_mem, name)) | 449 | die_compare_name(die_mem, fvp->name)) |
| 332 | return DIE_FIND_CB_FOUND; | 450 | return DIE_FIND_CB_FOUND; |
| 333 | 451 | ||
| 334 | return DIE_FIND_CB_CONTINUE; | 452 | if (dwarf_haspc(die_mem, fvp->addr)) |
| 453 | return DIE_FIND_CB_CONTINUE; | ||
| 454 | else | ||
| 455 | return DIE_FIND_CB_SIBLING; | ||
| 335 | } | 456 | } |
| 336 | 457 | ||
| 337 | /* Find a variable called 'name' */ | 458 | /* Find a variable called 'name' at given address */ |
| 338 | static Dwarf_Die *die_find_variable(Dwarf_Die *sp_die, const char *name, | 459 | static Dwarf_Die *die_find_variable_at(Dwarf_Die *sp_die, const char *name, |
| 339 | Dwarf_Die *die_mem) | 460 | Dwarf_Addr addr, Dwarf_Die *die_mem) |
| 340 | { | 461 | { |
| 341 | return die_find_child(sp_die, __die_find_variable_cb, (void *)name, | 462 | struct __find_variable_param fvp = { .name = name, .addr = addr}; |
| 463 | |||
| 464 | return die_find_child(sp_die, __die_find_variable_cb, (void *)&fvp, | ||
| 342 | die_mem); | 465 | die_mem); |
| 343 | } | 466 | } |
| 344 | 467 | ||
| @@ -361,6 +484,60 @@ static Dwarf_Die *die_find_member(Dwarf_Die *st_die, const char *name, | |||
| 361 | die_mem); | 484 | die_mem); |
| 362 | } | 485 | } |
| 363 | 486 | ||
| 487 | /* Get the name of given variable DIE */ | ||
| 488 | static int die_get_typename(Dwarf_Die *vr_die, char *buf, int len) | ||
| 489 | { | ||
| 490 | Dwarf_Die type; | ||
| 491 | int tag, ret, ret2; | ||
| 492 | const char *tmp = ""; | ||
| 493 | |||
| 494 | if (__die_get_real_type(vr_die, &type) == NULL) | ||
| 495 | return -ENOENT; | ||
| 496 | |||
| 497 | tag = dwarf_tag(&type); | ||
| 498 | if (tag == DW_TAG_array_type || tag == DW_TAG_pointer_type) | ||
| 499 | tmp = "*"; | ||
| 500 | else if (tag == DW_TAG_subroutine_type) { | ||
| 501 | /* Function pointer */ | ||
| 502 | ret = snprintf(buf, len, "(function_type)"); | ||
| 503 | return (ret >= len) ? -E2BIG : ret; | ||
| 504 | } else { | ||
| 505 | if (!dwarf_diename(&type)) | ||
| 506 | return -ENOENT; | ||
| 507 | if (tag == DW_TAG_union_type) | ||
| 508 | tmp = "union "; | ||
| 509 | else if (tag == DW_TAG_structure_type) | ||
| 510 | tmp = "struct "; | ||
| 511 | /* Write a base name */ | ||
| 512 | ret = snprintf(buf, len, "%s%s", tmp, dwarf_diename(&type)); | ||
| 513 | return (ret >= len) ? -E2BIG : ret; | ||
| 514 | } | ||
| 515 | ret = die_get_typename(&type, buf, len); | ||
| 516 | if (ret > 0) { | ||
| 517 | ret2 = snprintf(buf + ret, len - ret, "%s", tmp); | ||
| 518 | ret = (ret2 >= len - ret) ? -E2BIG : ret2 + ret; | ||
| 519 | } | ||
| 520 | return ret; | ||
| 521 | } | ||
| 522 | |||
| 523 | /* Get the name and type of given variable DIE, stored as "type\tname" */ | ||
| 524 | static int die_get_varname(Dwarf_Die *vr_die, char *buf, int len) | ||
| 525 | { | ||
| 526 | int ret, ret2; | ||
| 527 | |||
| 528 | ret = die_get_typename(vr_die, buf, len); | ||
| 529 | if (ret < 0) { | ||
| 530 | pr_debug("Failed to get type, make it unknown.\n"); | ||
| 531 | ret = snprintf(buf, len, "(unknown_type)"); | ||
| 532 | } | ||
| 533 | if (ret > 0) { | ||
| 534 | ret2 = snprintf(buf + ret, len - ret, "\t%s", | ||
| 535 | dwarf_diename(vr_die)); | ||
| 536 | ret = (ret2 >= len - ret) ? -E2BIG : ret2 + ret; | ||
| 537 | } | ||
| 538 | return ret; | ||
| 539 | } | ||
| 540 | |||
| 364 | /* | 541 | /* |
| 365 | * Probe finder related functions | 542 | * Probe finder related functions |
| 366 | */ | 543 | */ |
| @@ -374,8 +551,13 @@ static struct probe_trace_arg_ref *alloc_trace_arg_ref(long offs) | |||
| 374 | return ref; | 551 | return ref; |
| 375 | } | 552 | } |
| 376 | 553 | ||
| 377 | /* Show a location */ | 554 | /* |
| 378 | static int convert_variable_location(Dwarf_Die *vr_die, struct probe_finder *pf) | 555 | * Convert a location into trace_arg. |
| 556 | * If tvar == NULL, this just checks variable can be converted. | ||
| 557 | */ | ||
| 558 | static int convert_variable_location(Dwarf_Die *vr_die, Dwarf_Addr addr, | ||
| 559 | Dwarf_Op *fb_ops, | ||
| 560 | struct probe_trace_arg *tvar) | ||
| 379 | { | 561 | { |
| 380 | Dwarf_Attribute attr; | 562 | Dwarf_Attribute attr; |
| 381 | Dwarf_Op *op; | 563 | Dwarf_Op *op; |
| @@ -384,20 +566,23 @@ static int convert_variable_location(Dwarf_Die *vr_die, struct probe_finder *pf) | |||
| 384 | Dwarf_Word offs = 0; | 566 | Dwarf_Word offs = 0; |
| 385 | bool ref = false; | 567 | bool ref = false; |
| 386 | const char *regs; | 568 | const char *regs; |
| 387 | struct probe_trace_arg *tvar = pf->tvar; | ||
| 388 | int ret; | 569 | int ret; |
| 389 | 570 | ||
| 571 | if (dwarf_attr(vr_die, DW_AT_external, &attr) != NULL) | ||
| 572 | goto static_var; | ||
| 573 | |||
| 390 | /* TODO: handle more than 1 exprs */ | 574 | /* TODO: handle more than 1 exprs */ |
| 391 | if (dwarf_attr(vr_die, DW_AT_location, &attr) == NULL || | 575 | if (dwarf_attr(vr_die, DW_AT_location, &attr) == NULL || |
| 392 | dwarf_getlocation_addr(&attr, pf->addr, &op, &nops, 1) <= 0 || | 576 | dwarf_getlocation_addr(&attr, addr, &op, &nops, 1) <= 0 || |
| 393 | nops == 0) { | 577 | nops == 0) { |
| 394 | /* TODO: Support const_value */ | 578 | /* TODO: Support const_value */ |
| 395 | pr_err("Failed to find the location of %s at this address.\n" | ||
| 396 | " Perhaps, it has been optimized out.\n", pf->pvar->var); | ||
| 397 | return -ENOENT; | 579 | return -ENOENT; |
| 398 | } | 580 | } |
| 399 | 581 | ||
| 400 | if (op->atom == DW_OP_addr) { | 582 | if (op->atom == DW_OP_addr) { |
| 583 | static_var: | ||
| 584 | if (!tvar) | ||
| 585 | return 0; | ||
| 401 | /* Static variables on memory (not stack), make @varname */ | 586 | /* Static variables on memory (not stack), make @varname */ |
| 402 | ret = strlen(dwarf_diename(vr_die)); | 587 | ret = strlen(dwarf_diename(vr_die)); |
| 403 | tvar->value = zalloc(ret + 2); | 588 | tvar->value = zalloc(ret + 2); |
| @@ -412,14 +597,11 @@ static int convert_variable_location(Dwarf_Die *vr_die, struct probe_finder *pf) | |||
| 412 | 597 | ||
| 413 | /* If this is based on frame buffer, set the offset */ | 598 | /* If this is based on frame buffer, set the offset */ |
| 414 | if (op->atom == DW_OP_fbreg) { | 599 | if (op->atom == DW_OP_fbreg) { |
| 415 | if (pf->fb_ops == NULL) { | 600 | if (fb_ops == NULL) |
| 416 | pr_warning("The attribute of frame base is not " | ||
| 417 | "supported.\n"); | ||
| 418 | return -ENOTSUP; | 601 | return -ENOTSUP; |
| 419 | } | ||
| 420 | ref = true; | 602 | ref = true; |
| 421 | offs = op->number; | 603 | offs = op->number; |
| 422 | op = &pf->fb_ops[0]; | 604 | op = &fb_ops[0]; |
| 423 | } | 605 | } |
| 424 | 606 | ||
| 425 | if (op->atom >= DW_OP_breg0 && op->atom <= DW_OP_breg31) { | 607 | if (op->atom >= DW_OP_breg0 && op->atom <= DW_OP_breg31) { |
| @@ -435,13 +617,18 @@ static int convert_variable_location(Dwarf_Die *vr_die, struct probe_finder *pf) | |||
| 435 | } else if (op->atom == DW_OP_regx) { | 617 | } else if (op->atom == DW_OP_regx) { |
| 436 | regn = op->number; | 618 | regn = op->number; |
| 437 | } else { | 619 | } else { |
| 438 | pr_warning("DW_OP %x is not supported.\n", op->atom); | 620 | pr_debug("DW_OP %x is not supported.\n", op->atom); |
| 439 | return -ENOTSUP; | 621 | return -ENOTSUP; |
| 440 | } | 622 | } |
| 441 | 623 | ||
| 624 | if (!tvar) | ||
| 625 | return 0; | ||
| 626 | |||
| 442 | regs = get_arch_regstr(regn); | 627 | regs = get_arch_regstr(regn); |
| 443 | if (!regs) { | 628 | if (!regs) { |
| 444 | pr_warning("Mapping for DWARF register number %u missing on this architecture.", regn); | 629 | /* This should be a bug in DWARF or this tool */ |
| 630 | pr_warning("Mapping for DWARF register number %u " | ||
| 631 | "missing on this architecture.", regn); | ||
| 445 | return -ERANGE; | 632 | return -ERANGE; |
| 446 | } | 633 | } |
| 447 | 634 | ||
| @@ -666,8 +853,14 @@ static int convert_variable(Dwarf_Die *vr_die, struct probe_finder *pf) | |||
| 666 | pr_debug("Converting variable %s into trace event.\n", | 853 | pr_debug("Converting variable %s into trace event.\n", |
| 667 | dwarf_diename(vr_die)); | 854 | dwarf_diename(vr_die)); |
| 668 | 855 | ||
| 669 | ret = convert_variable_location(vr_die, pf); | 856 | ret = convert_variable_location(vr_die, pf->addr, pf->fb_ops, |
| 670 | if (ret == 0 && pf->pvar->field) { | 857 | pf->tvar); |
| 858 | if (ret == -ENOENT) | ||
| 859 | pr_err("Failed to find the location of %s at this address.\n" | ||
| 860 | " Perhaps, it has been optimized out.\n", pf->pvar->var); | ||
| 861 | else if (ret == -ENOTSUP) | ||
| 862 | pr_err("Sorry, we don't support this variable location yet.\n"); | ||
| 863 | else if (pf->pvar->field) { | ||
| 671 | ret = convert_variable_fields(vr_die, pf->pvar->var, | 864 | ret = convert_variable_fields(vr_die, pf->pvar->var, |
| 672 | pf->pvar->field, &pf->tvar->ref, | 865 | pf->pvar->field, &pf->tvar->ref, |
| 673 | &die_mem); | 866 | &die_mem); |
| @@ -722,56 +915,39 @@ static int find_variable(Dwarf_Die *sp_die, struct probe_finder *pf) | |||
| 722 | pr_debug("Searching '%s' variable in context.\n", | 915 | pr_debug("Searching '%s' variable in context.\n", |
| 723 | pf->pvar->var); | 916 | pf->pvar->var); |
| 724 | /* Search child die for local variables and parameters. */ | 917 | /* Search child die for local variables and parameters. */ |
| 725 | if (die_find_variable(sp_die, pf->pvar->var, &vr_die)) | 918 | if (die_find_variable_at(sp_die, pf->pvar->var, pf->addr, &vr_die)) |
| 726 | ret = convert_variable(&vr_die, pf); | 919 | ret = convert_variable(&vr_die, pf); |
| 727 | else { | 920 | else { |
| 728 | /* Search upper class */ | 921 | /* Search upper class */ |
| 729 | nscopes = dwarf_getscopes_die(sp_die, &scopes); | 922 | nscopes = dwarf_getscopes_die(sp_die, &scopes); |
| 730 | if (nscopes > 0) { | 923 | while (nscopes-- > 1) { |
| 731 | ret = dwarf_getscopevar(scopes, nscopes, pf->pvar->var, | 924 | pr_debug("Searching variables in %s\n", |
| 732 | 0, NULL, 0, 0, &vr_die); | 925 | dwarf_diename(&scopes[nscopes])); |
| 733 | if (ret >= 0) | 926 | /* We should check this scope, so give dummy address */ |
| 927 | if (die_find_variable_at(&scopes[nscopes], | ||
| 928 | pf->pvar->var, 0, | ||
| 929 | &vr_die)) { | ||
| 734 | ret = convert_variable(&vr_die, pf); | 930 | ret = convert_variable(&vr_die, pf); |
| 735 | else | 931 | goto found; |
| 736 | ret = -ENOENT; | 932 | } |
| 933 | } | ||
| 934 | if (scopes) | ||
| 737 | free(scopes); | 935 | free(scopes); |
| 738 | } else | 936 | ret = -ENOENT; |
| 739 | ret = -ENOENT; | ||
| 740 | } | 937 | } |
| 938 | found: | ||
| 741 | if (ret < 0) | 939 | if (ret < 0) |
| 742 | pr_warning("Failed to find '%s' in this function.\n", | 940 | pr_warning("Failed to find '%s' in this function.\n", |
| 743 | pf->pvar->var); | 941 | pf->pvar->var); |
| 744 | return ret; | 942 | return ret; |
| 745 | } | 943 | } |
| 746 | 944 | ||
| 747 | /* Show a probe point to output buffer */ | 945 | /* Convert subprogram DIE to trace point */ |
| 748 | static int convert_probe_point(Dwarf_Die *sp_die, struct probe_finder *pf) | 946 | static int convert_to_trace_point(Dwarf_Die *sp_die, Dwarf_Addr paddr, |
| 947 | bool retprobe, struct probe_trace_point *tp) | ||
| 749 | { | 948 | { |
| 750 | struct probe_trace_event *tev; | ||
| 751 | Dwarf_Addr eaddr; | 949 | Dwarf_Addr eaddr; |
| 752 | Dwarf_Die die_mem; | ||
| 753 | const char *name; | 950 | const char *name; |
| 754 | int ret, i; | ||
| 755 | Dwarf_Attribute fb_attr; | ||
| 756 | size_t nops; | ||
| 757 | |||
| 758 | if (pf->ntevs == pf->max_tevs) { | ||
| 759 | pr_warning("Too many( > %d) probe point found.\n", | ||
| 760 | pf->max_tevs); | ||
| 761 | return -ERANGE; | ||
| 762 | } | ||
| 763 | tev = &pf->tevs[pf->ntevs++]; | ||
| 764 | |||
| 765 | /* If no real subprogram, find a real one */ | ||
| 766 | if (!sp_die || dwarf_tag(sp_die) != DW_TAG_subprogram) { | ||
| 767 | sp_die = die_find_real_subprogram(&pf->cu_die, | ||
| 768 | pf->addr, &die_mem); | ||
| 769 | if (!sp_die) { | ||
| 770 | pr_warning("Failed to find probe point in any " | ||
| 771 | "functions.\n"); | ||
| 772 | return -ENOENT; | ||
| 773 | } | ||
| 774 | } | ||
| 775 | 951 | ||
| 776 | /* Copy the name of probe point */ | 952 | /* Copy the name of probe point */ |
| 777 | name = dwarf_diename(sp_die); | 953 | name = dwarf_diename(sp_die); |
| @@ -781,26 +957,45 @@ static int convert_probe_point(Dwarf_Die *sp_die, struct probe_finder *pf) | |||
| 781 | dwarf_diename(sp_die)); | 957 | dwarf_diename(sp_die)); |
| 782 | return -ENOENT; | 958 | return -ENOENT; |
| 783 | } | 959 | } |
| 784 | tev->point.symbol = strdup(name); | 960 | tp->symbol = strdup(name); |
| 785 | if (tev->point.symbol == NULL) | 961 | if (tp->symbol == NULL) |
| 786 | return -ENOMEM; | 962 | return -ENOMEM; |
| 787 | tev->point.offset = (unsigned long)(pf->addr - eaddr); | 963 | tp->offset = (unsigned long)(paddr - eaddr); |
| 788 | } else | 964 | } else |
| 789 | /* This function has no name. */ | 965 | /* This function has no name. */ |
| 790 | tev->point.offset = (unsigned long)pf->addr; | 966 | tp->offset = (unsigned long)paddr; |
| 791 | 967 | ||
| 792 | /* Return probe must be on the head of a subprogram */ | 968 | /* Return probe must be on the head of a subprogram */ |
| 793 | if (pf->pev->point.retprobe) { | 969 | if (retprobe) { |
| 794 | if (tev->point.offset != 0) { | 970 | if (eaddr != paddr) { |
| 795 | pr_warning("Return probe must be on the head of" | 971 | pr_warning("Return probe must be on the head of" |
| 796 | " a real function\n"); | 972 | " a real function\n"); |
| 797 | return -EINVAL; | 973 | return -EINVAL; |
| 798 | } | 974 | } |
| 799 | tev->point.retprobe = true; | 975 | tp->retprobe = true; |
| 800 | } | 976 | } |
| 801 | 977 | ||
| 802 | pr_debug("Probe point found: %s+%lu\n", tev->point.symbol, | 978 | return 0; |
| 803 | tev->point.offset); | 979 | } |
| 980 | |||
| 981 | /* Call probe_finder callback with real subprogram DIE */ | ||
| 982 | static int call_probe_finder(Dwarf_Die *sp_die, struct probe_finder *pf) | ||
| 983 | { | ||
| 984 | Dwarf_Die die_mem; | ||
| 985 | Dwarf_Attribute fb_attr; | ||
| 986 | size_t nops; | ||
| 987 | int ret; | ||
| 988 | |||
| 989 | /* If no real subprogram, find a real one */ | ||
| 990 | if (!sp_die || dwarf_tag(sp_die) != DW_TAG_subprogram) { | ||
| 991 | sp_die = die_find_real_subprogram(&pf->cu_die, | ||
| 992 | pf->addr, &die_mem); | ||
| 993 | if (!sp_die) { | ||
| 994 | pr_warning("Failed to find probe point in any " | ||
| 995 | "functions.\n"); | ||
| 996 | return -ENOENT; | ||
| 997 | } | ||
| 998 | } | ||
| 804 | 999 | ||
| 805 | /* Get the frame base attribute/ops */ | 1000 | /* Get the frame base attribute/ops */ |
| 806 | dwarf_attr(sp_die, DW_AT_frame_base, &fb_attr); | 1001 | dwarf_attr(sp_die, DW_AT_frame_base, &fb_attr); |
| @@ -820,22 +1015,13 @@ static int convert_probe_point(Dwarf_Die *sp_die, struct probe_finder *pf) | |||
| 820 | #endif | 1015 | #endif |
| 821 | } | 1016 | } |
| 822 | 1017 | ||
| 823 | /* Find each argument */ | 1018 | /* Call finder's callback handler */ |
| 824 | tev->nargs = pf->pev->nargs; | 1019 | ret = pf->callback(sp_die, pf); |
| 825 | tev->args = zalloc(sizeof(struct probe_trace_arg) * tev->nargs); | ||
| 826 | if (tev->args == NULL) | ||
| 827 | return -ENOMEM; | ||
| 828 | for (i = 0; i < pf->pev->nargs; i++) { | ||
| 829 | pf->pvar = &pf->pev->args[i]; | ||
| 830 | pf->tvar = &tev->args[i]; | ||
| 831 | ret = find_variable(sp_die, pf); | ||
| 832 | if (ret != 0) | ||
| 833 | return ret; | ||
| 834 | } | ||
| 835 | 1020 | ||
| 836 | /* *pf->fb_ops will be cached in libdw. Don't free it. */ | 1021 | /* *pf->fb_ops will be cached in libdw. Don't free it. */ |
| 837 | pf->fb_ops = NULL; | 1022 | pf->fb_ops = NULL; |
| 838 | return 0; | 1023 | |
| 1024 | return ret; | ||
| 839 | } | 1025 | } |
| 840 | 1026 | ||
| 841 | /* Find probe point from its line number */ | 1027 | /* Find probe point from its line number */ |
| @@ -871,7 +1057,7 @@ static int find_probe_point_by_line(struct probe_finder *pf) | |||
| 871 | (int)i, lineno, (uintmax_t)addr); | 1057 | (int)i, lineno, (uintmax_t)addr); |
| 872 | pf->addr = addr; | 1058 | pf->addr = addr; |
| 873 | 1059 | ||
| 874 | ret = convert_probe_point(NULL, pf); | 1060 | ret = call_probe_finder(NULL, pf); |
| 875 | /* Continuing, because target line might be inlined. */ | 1061 | /* Continuing, because target line might be inlined. */ |
| 876 | } | 1062 | } |
| 877 | return ret; | 1063 | return ret; |
| @@ -984,7 +1170,7 @@ static int find_probe_point_lazy(Dwarf_Die *sp_die, struct probe_finder *pf) | |||
| 984 | (int)i, lineno, (unsigned long long)addr); | 1170 | (int)i, lineno, (unsigned long long)addr); |
| 985 | pf->addr = addr; | 1171 | pf->addr = addr; |
| 986 | 1172 | ||
| 987 | ret = convert_probe_point(sp_die, pf); | 1173 | ret = call_probe_finder(sp_die, pf); |
| 988 | /* Continuing, because target line might be inlined. */ | 1174 | /* Continuing, because target line might be inlined. */ |
| 989 | } | 1175 | } |
| 990 | /* TODO: deallocate lines, but how? */ | 1176 | /* TODO: deallocate lines, but how? */ |
| @@ -1019,7 +1205,7 @@ static int probe_point_inline_cb(Dwarf_Die *in_die, void *data) | |||
| 1019 | pr_debug("found inline addr: 0x%jx\n", | 1205 | pr_debug("found inline addr: 0x%jx\n", |
| 1020 | (uintmax_t)pf->addr); | 1206 | (uintmax_t)pf->addr); |
| 1021 | 1207 | ||
| 1022 | param->retval = convert_probe_point(in_die, pf); | 1208 | param->retval = call_probe_finder(in_die, pf); |
| 1023 | if (param->retval < 0) | 1209 | if (param->retval < 0) |
| 1024 | return DWARF_CB_ABORT; | 1210 | return DWARF_CB_ABORT; |
| 1025 | } | 1211 | } |
| @@ -1057,7 +1243,7 @@ static int probe_point_search_cb(Dwarf_Die *sp_die, void *data) | |||
| 1057 | } | 1243 | } |
| 1058 | pf->addr += pp->offset; | 1244 | pf->addr += pp->offset; |
| 1059 | /* TODO: Check the address in this function */ | 1245 | /* TODO: Check the address in this function */ |
| 1060 | param->retval = convert_probe_point(sp_die, pf); | 1246 | param->retval = call_probe_finder(sp_die, pf); |
| 1061 | } | 1247 | } |
| 1062 | } else { | 1248 | } else { |
| 1063 | struct dwarf_callback_param _param = {.data = (void *)pf, | 1249 | struct dwarf_callback_param _param = {.data = (void *)pf, |
| @@ -1079,90 +1265,276 @@ static int find_probe_point_by_func(struct probe_finder *pf) | |||
| 1079 | return _param.retval; | 1265 | return _param.retval; |
| 1080 | } | 1266 | } |
| 1081 | 1267 | ||
| 1082 | /* Find probe_trace_events specified by perf_probe_event from debuginfo */ | 1268 | /* Find probe points from debuginfo */ |
| 1083 | int find_probe_trace_events(int fd, struct perf_probe_event *pev, | 1269 | static int find_probes(int fd, struct probe_finder *pf) |
| 1084 | struct probe_trace_event **tevs, int max_tevs) | ||
| 1085 | { | 1270 | { |
| 1086 | struct probe_finder pf = {.pev = pev, .max_tevs = max_tevs}; | 1271 | struct perf_probe_point *pp = &pf->pev->point; |
| 1087 | struct perf_probe_point *pp = &pev->point; | ||
| 1088 | Dwarf_Off off, noff; | 1272 | Dwarf_Off off, noff; |
| 1089 | size_t cuhl; | 1273 | size_t cuhl; |
| 1090 | Dwarf_Die *diep; | 1274 | Dwarf_Die *diep; |
| 1091 | Dwarf *dbg; | 1275 | Dwarf *dbg = NULL; |
| 1276 | Dwfl *dwfl; | ||
| 1277 | Dwarf_Addr bias; /* Currently ignored */ | ||
| 1092 | int ret = 0; | 1278 | int ret = 0; |
| 1093 | 1279 | ||
| 1094 | pf.tevs = zalloc(sizeof(struct probe_trace_event) * max_tevs); | 1280 | dbg = dwfl_init_offline_dwarf(fd, &dwfl, &bias); |
| 1095 | if (pf.tevs == NULL) | ||
| 1096 | return -ENOMEM; | ||
| 1097 | *tevs = pf.tevs; | ||
| 1098 | pf.ntevs = 0; | ||
| 1099 | |||
| 1100 | dbg = dwarf_begin(fd, DWARF_C_READ); | ||
| 1101 | if (!dbg) { | 1281 | if (!dbg) { |
| 1102 | pr_warning("No dwarf info found in the vmlinux - " | 1282 | pr_warning("No dwarf info found in the vmlinux - " |
| 1103 | "please rebuild with CONFIG_DEBUG_INFO=y.\n"); | 1283 | "please rebuild with CONFIG_DEBUG_INFO=y.\n"); |
| 1104 | free(pf.tevs); | ||
| 1105 | *tevs = NULL; | ||
| 1106 | return -EBADF; | 1284 | return -EBADF; |
| 1107 | } | 1285 | } |
| 1108 | 1286 | ||
| 1109 | #if _ELFUTILS_PREREQ(0, 142) | 1287 | #if _ELFUTILS_PREREQ(0, 142) |
| 1110 | /* Get the call frame information from this dwarf */ | 1288 | /* Get the call frame information from this dwarf */ |
| 1111 | pf.cfi = dwarf_getcfi(dbg); | 1289 | pf->cfi = dwarf_getcfi(dbg); |
| 1112 | #endif | 1290 | #endif |
| 1113 | 1291 | ||
| 1114 | off = 0; | 1292 | off = 0; |
| 1115 | line_list__init(&pf.lcache); | 1293 | line_list__init(&pf->lcache); |
| 1116 | /* Loop on CUs (Compilation Unit) */ | 1294 | /* Loop on CUs (Compilation Unit) */ |
| 1117 | while (!dwarf_nextcu(dbg, off, &noff, &cuhl, NULL, NULL, NULL) && | 1295 | while (!dwarf_nextcu(dbg, off, &noff, &cuhl, NULL, NULL, NULL) && |
| 1118 | ret >= 0) { | 1296 | ret >= 0) { |
| 1119 | /* Get the DIE(Debugging Information Entry) of this CU */ | 1297 | /* Get the DIE(Debugging Information Entry) of this CU */ |
| 1120 | diep = dwarf_offdie(dbg, off + cuhl, &pf.cu_die); | 1298 | diep = dwarf_offdie(dbg, off + cuhl, &pf->cu_die); |
| 1121 | if (!diep) | 1299 | if (!diep) |
| 1122 | continue; | 1300 | continue; |
| 1123 | 1301 | ||
| 1124 | /* Check if target file is included. */ | 1302 | /* Check if target file is included. */ |
| 1125 | if (pp->file) | 1303 | if (pp->file) |
| 1126 | pf.fname = cu_find_realpath(&pf.cu_die, pp->file); | 1304 | pf->fname = cu_find_realpath(&pf->cu_die, pp->file); |
| 1127 | else | 1305 | else |
| 1128 | pf.fname = NULL; | 1306 | pf->fname = NULL; |
| 1129 | 1307 | ||
| 1130 | if (!pp->file || pf.fname) { | 1308 | if (!pp->file || pf->fname) { |
| 1131 | if (pp->function) | 1309 | if (pp->function) |
| 1132 | ret = find_probe_point_by_func(&pf); | 1310 | ret = find_probe_point_by_func(pf); |
| 1133 | else if (pp->lazy_line) | 1311 | else if (pp->lazy_line) |
| 1134 | ret = find_probe_point_lazy(NULL, &pf); | 1312 | ret = find_probe_point_lazy(NULL, pf); |
| 1135 | else { | 1313 | else { |
| 1136 | pf.lno = pp->line; | 1314 | pf->lno = pp->line; |
| 1137 | ret = find_probe_point_by_line(&pf); | 1315 | ret = find_probe_point_by_line(pf); |
| 1138 | } | 1316 | } |
| 1139 | } | 1317 | } |
| 1140 | off = noff; | 1318 | off = noff; |
| 1141 | } | 1319 | } |
| 1142 | line_list__free(&pf.lcache); | 1320 | line_list__free(&pf->lcache); |
| 1143 | dwarf_end(dbg); | 1321 | if (dwfl) |
| 1322 | dwfl_end(dwfl); | ||
| 1144 | 1323 | ||
| 1145 | return (ret < 0) ? ret : pf.ntevs; | 1324 | return ret; |
| 1325 | } | ||
| 1326 | |||
| 1327 | /* Add a found probe point into trace event list */ | ||
| 1328 | static int add_probe_trace_event(Dwarf_Die *sp_die, struct probe_finder *pf) | ||
| 1329 | { | ||
| 1330 | struct trace_event_finder *tf = | ||
| 1331 | container_of(pf, struct trace_event_finder, pf); | ||
| 1332 | struct probe_trace_event *tev; | ||
| 1333 | int ret, i; | ||
| 1334 | |||
| 1335 | /* Check number of tevs */ | ||
| 1336 | if (tf->ntevs == tf->max_tevs) { | ||
| 1337 | pr_warning("Too many( > %d) probe point found.\n", | ||
| 1338 | tf->max_tevs); | ||
| 1339 | return -ERANGE; | ||
| 1340 | } | ||
| 1341 | tev = &tf->tevs[tf->ntevs++]; | ||
| 1342 | |||
| 1343 | ret = convert_to_trace_point(sp_die, pf->addr, pf->pev->point.retprobe, | ||
| 1344 | &tev->point); | ||
| 1345 | if (ret < 0) | ||
| 1346 | return ret; | ||
| 1347 | |||
| 1348 | pr_debug("Probe point found: %s+%lu\n", tev->point.symbol, | ||
| 1349 | tev->point.offset); | ||
| 1350 | |||
| 1351 | /* Find each argument */ | ||
| 1352 | tev->nargs = pf->pev->nargs; | ||
| 1353 | tev->args = zalloc(sizeof(struct probe_trace_arg) * tev->nargs); | ||
| 1354 | if (tev->args == NULL) | ||
| 1355 | return -ENOMEM; | ||
| 1356 | for (i = 0; i < pf->pev->nargs; i++) { | ||
| 1357 | pf->pvar = &pf->pev->args[i]; | ||
| 1358 | pf->tvar = &tev->args[i]; | ||
| 1359 | ret = find_variable(sp_die, pf); | ||
| 1360 | if (ret != 0) | ||
| 1361 | return ret; | ||
| 1362 | } | ||
| 1363 | |||
| 1364 | return 0; | ||
| 1365 | } | ||
| 1366 | |||
| 1367 | /* Find probe_trace_events specified by perf_probe_event from debuginfo */ | ||
| 1368 | int find_probe_trace_events(int fd, struct perf_probe_event *pev, | ||
| 1369 | struct probe_trace_event **tevs, int max_tevs) | ||
| 1370 | { | ||
| 1371 | struct trace_event_finder tf = { | ||
| 1372 | .pf = {.pev = pev, .callback = add_probe_trace_event}, | ||
| 1373 | .max_tevs = max_tevs}; | ||
| 1374 | int ret; | ||
| 1375 | |||
| 1376 | /* Allocate result tevs array */ | ||
| 1377 | *tevs = zalloc(sizeof(struct probe_trace_event) * max_tevs); | ||
| 1378 | if (*tevs == NULL) | ||
| 1379 | return -ENOMEM; | ||
| 1380 | |||
| 1381 | tf.tevs = *tevs; | ||
| 1382 | tf.ntevs = 0; | ||
| 1383 | |||
| 1384 | ret = find_probes(fd, &tf.pf); | ||
| 1385 | if (ret < 0) { | ||
| 1386 | free(*tevs); | ||
| 1387 | *tevs = NULL; | ||
| 1388 | return ret; | ||
| 1389 | } | ||
| 1390 | |||
| 1391 | return (ret < 0) ? ret : tf.ntevs; | ||
| 1392 | } | ||
| 1393 | |||
| 1394 | #define MAX_VAR_LEN 64 | ||
| 1395 | |||
| 1396 | /* Collect available variables in this scope */ | ||
| 1397 | static int collect_variables_cb(Dwarf_Die *die_mem, void *data) | ||
| 1398 | { | ||
| 1399 | struct available_var_finder *af = data; | ||
| 1400 | struct variable_list *vl; | ||
| 1401 | char buf[MAX_VAR_LEN]; | ||
| 1402 | int tag, ret; | ||
| 1403 | |||
| 1404 | vl = &af->vls[af->nvls - 1]; | ||
| 1405 | |||
| 1406 | tag = dwarf_tag(die_mem); | ||
| 1407 | if (tag == DW_TAG_formal_parameter || | ||
| 1408 | tag == DW_TAG_variable) { | ||
| 1409 | ret = convert_variable_location(die_mem, af->pf.addr, | ||
| 1410 | af->pf.fb_ops, NULL); | ||
| 1411 | if (ret == 0) { | ||
| 1412 | ret = die_get_varname(die_mem, buf, MAX_VAR_LEN); | ||
| 1413 | pr_debug2("Add new var: %s\n", buf); | ||
| 1414 | if (ret > 0) | ||
| 1415 | strlist__add(vl->vars, buf); | ||
| 1416 | } | ||
| 1417 | } | ||
| 1418 | |||
| 1419 | if (af->child && dwarf_haspc(die_mem, af->pf.addr)) | ||
| 1420 | return DIE_FIND_CB_CONTINUE; | ||
| 1421 | else | ||
| 1422 | return DIE_FIND_CB_SIBLING; | ||
| 1423 | } | ||
| 1424 | |||
| 1425 | /* Add a found vars into available variables list */ | ||
| 1426 | static int add_available_vars(Dwarf_Die *sp_die, struct probe_finder *pf) | ||
| 1427 | { | ||
| 1428 | struct available_var_finder *af = | ||
| 1429 | container_of(pf, struct available_var_finder, pf); | ||
| 1430 | struct variable_list *vl; | ||
| 1431 | Dwarf_Die die_mem, *scopes = NULL; | ||
| 1432 | int ret, nscopes; | ||
| 1433 | |||
| 1434 | /* Check number of tevs */ | ||
| 1435 | if (af->nvls == af->max_vls) { | ||
| 1436 | pr_warning("Too many( > %d) probe point found.\n", af->max_vls); | ||
| 1437 | return -ERANGE; | ||
| 1438 | } | ||
| 1439 | vl = &af->vls[af->nvls++]; | ||
| 1440 | |||
| 1441 | ret = convert_to_trace_point(sp_die, pf->addr, pf->pev->point.retprobe, | ||
| 1442 | &vl->point); | ||
| 1443 | if (ret < 0) | ||
| 1444 | return ret; | ||
| 1445 | |||
| 1446 | pr_debug("Probe point found: %s+%lu\n", vl->point.symbol, | ||
| 1447 | vl->point.offset); | ||
| 1448 | |||
| 1449 | /* Find local variables */ | ||
| 1450 | vl->vars = strlist__new(true, NULL); | ||
| 1451 | if (vl->vars == NULL) | ||
| 1452 | return -ENOMEM; | ||
| 1453 | af->child = true; | ||
| 1454 | die_find_child(sp_die, collect_variables_cb, (void *)af, &die_mem); | ||
| 1455 | |||
| 1456 | /* Find external variables */ | ||
| 1457 | if (!af->externs) | ||
| 1458 | goto out; | ||
| 1459 | /* Don't need to search child DIE for externs. */ | ||
| 1460 | af->child = false; | ||
| 1461 | nscopes = dwarf_getscopes_die(sp_die, &scopes); | ||
| 1462 | while (nscopes-- > 1) | ||
| 1463 | die_find_child(&scopes[nscopes], collect_variables_cb, | ||
| 1464 | (void *)af, &die_mem); | ||
| 1465 | if (scopes) | ||
| 1466 | free(scopes); | ||
| 1467 | |||
| 1468 | out: | ||
| 1469 | if (strlist__empty(vl->vars)) { | ||
| 1470 | strlist__delete(vl->vars); | ||
| 1471 | vl->vars = NULL; | ||
| 1472 | } | ||
| 1473 | |||
| 1474 | return ret; | ||
| 1475 | } | ||
| 1476 | |||
| 1477 | /* Find available variables at given probe point */ | ||
| 1478 | int find_available_vars_at(int fd, struct perf_probe_event *pev, | ||
| 1479 | struct variable_list **vls, int max_vls, | ||
| 1480 | bool externs) | ||
| 1481 | { | ||
| 1482 | struct available_var_finder af = { | ||
| 1483 | .pf = {.pev = pev, .callback = add_available_vars}, | ||
| 1484 | .max_vls = max_vls, .externs = externs}; | ||
| 1485 | int ret; | ||
| 1486 | |||
| 1487 | /* Allocate result vls array */ | ||
| 1488 | *vls = zalloc(sizeof(struct variable_list) * max_vls); | ||
| 1489 | if (*vls == NULL) | ||
| 1490 | return -ENOMEM; | ||
| 1491 | |||
| 1492 | af.vls = *vls; | ||
| 1493 | af.nvls = 0; | ||
| 1494 | |||
| 1495 | ret = find_probes(fd, &af.pf); | ||
| 1496 | if (ret < 0) { | ||
| 1497 | /* Free vlist for error */ | ||
| 1498 | while (af.nvls--) { | ||
| 1499 | if (af.vls[af.nvls].point.symbol) | ||
| 1500 | free(af.vls[af.nvls].point.symbol); | ||
| 1501 | if (af.vls[af.nvls].vars) | ||
| 1502 | strlist__delete(af.vls[af.nvls].vars); | ||
| 1503 | } | ||
| 1504 | free(af.vls); | ||
| 1505 | *vls = NULL; | ||
| 1506 | return ret; | ||
| 1507 | } | ||
| 1508 | |||
| 1509 | return (ret < 0) ? ret : af.nvls; | ||
| 1146 | } | 1510 | } |
| 1147 | 1511 | ||
| 1148 | /* Reverse search */ | 1512 | /* Reverse search */ |
| 1149 | int find_perf_probe_point(int fd, unsigned long addr, | 1513 | int find_perf_probe_point(unsigned long addr, struct perf_probe_point *ppt) |
| 1150 | struct perf_probe_point *ppt) | ||
| 1151 | { | 1514 | { |
| 1152 | Dwarf_Die cudie, spdie, indie; | 1515 | Dwarf_Die cudie, spdie, indie; |
| 1153 | Dwarf *dbg; | 1516 | Dwarf *dbg = NULL; |
| 1517 | Dwfl *dwfl = NULL; | ||
| 1154 | Dwarf_Line *line; | 1518 | Dwarf_Line *line; |
| 1155 | Dwarf_Addr laddr, eaddr; | 1519 | Dwarf_Addr laddr, eaddr, bias = 0; |
| 1156 | const char *tmp; | 1520 | const char *tmp; |
| 1157 | int lineno, ret = 0; | 1521 | int lineno, ret = 0; |
| 1158 | bool found = false; | 1522 | bool found = false; |
| 1159 | 1523 | ||
| 1160 | dbg = dwarf_begin(fd, DWARF_C_READ); | 1524 | /* Open the live linux kernel */ |
| 1161 | if (!dbg) | 1525 | dbg = dwfl_init_live_kernel_dwarf(addr, &dwfl, &bias); |
| 1162 | return -EBADF; | 1526 | if (!dbg) { |
| 1527 | pr_warning("No dwarf info found in the vmlinux - " | ||
| 1528 | "please rebuild with CONFIG_DEBUG_INFO=y.\n"); | ||
| 1529 | ret = -EINVAL; | ||
| 1530 | goto end; | ||
| 1531 | } | ||
| 1163 | 1532 | ||
| 1533 | /* Adjust address with bias */ | ||
| 1534 | addr += bias; | ||
| 1164 | /* Find cu die */ | 1535 | /* Find cu die */ |
| 1165 | if (!dwarf_addrdie(dbg, (Dwarf_Addr)addr, &cudie)) { | 1536 | if (!dwarf_addrdie(dbg, (Dwarf_Addr)addr - bias, &cudie)) { |
| 1537 | pr_warning("No CU DIE is found at %lx\n", addr); | ||
| 1166 | ret = -EINVAL; | 1538 | ret = -EINVAL; |
| 1167 | goto end; | 1539 | goto end; |
| 1168 | } | 1540 | } |
| @@ -1225,7 +1597,8 @@ found: | |||
| 1225 | } | 1597 | } |
| 1226 | 1598 | ||
| 1227 | end: | 1599 | end: |
| 1228 | dwarf_end(dbg); | 1600 | if (dwfl) |
| 1601 | dwfl_end(dwfl); | ||
| 1229 | if (ret >= 0) | 1602 | if (ret >= 0) |
| 1230 | ret = found ? 1 : 0; | 1603 | ret = found ? 1 : 0; |
| 1231 | return ret; | 1604 | return ret; |
| @@ -1358,6 +1731,9 @@ static int line_range_search_cb(Dwarf_Die *sp_die, void *data) | |||
| 1358 | struct line_finder *lf = param->data; | 1731 | struct line_finder *lf = param->data; |
| 1359 | struct line_range *lr = lf->lr; | 1732 | struct line_range *lr = lf->lr; |
| 1360 | 1733 | ||
| 1734 | pr_debug("find (%llx) %s\n", | ||
| 1735 | (unsigned long long)dwarf_dieoffset(sp_die), | ||
| 1736 | dwarf_diename(sp_die)); | ||
| 1361 | if (dwarf_tag(sp_die) == DW_TAG_subprogram && | 1737 | if (dwarf_tag(sp_die) == DW_TAG_subprogram && |
| 1362 | die_compare_name(sp_die, lr->function)) { | 1738 | die_compare_name(sp_die, lr->function)) { |
| 1363 | lf->fname = dwarf_decl_file(sp_die); | 1739 | lf->fname = dwarf_decl_file(sp_die); |
| @@ -1401,10 +1777,12 @@ int find_line_range(int fd, struct line_range *lr) | |||
| 1401 | Dwarf_Off off = 0, noff; | 1777 | Dwarf_Off off = 0, noff; |
| 1402 | size_t cuhl; | 1778 | size_t cuhl; |
| 1403 | Dwarf_Die *diep; | 1779 | Dwarf_Die *diep; |
| 1404 | Dwarf *dbg; | 1780 | Dwarf *dbg = NULL; |
| 1781 | Dwfl *dwfl; | ||
| 1782 | Dwarf_Addr bias; /* Currently ignored */ | ||
| 1405 | const char *comp_dir; | 1783 | const char *comp_dir; |
| 1406 | 1784 | ||
| 1407 | dbg = dwarf_begin(fd, DWARF_C_READ); | 1785 | dbg = dwfl_init_offline_dwarf(fd, &dwfl, &bias); |
| 1408 | if (!dbg) { | 1786 | if (!dbg) { |
| 1409 | pr_warning("No dwarf info found in the vmlinux - " | 1787 | pr_warning("No dwarf info found in the vmlinux - " |
| 1410 | "please rebuild with CONFIG_DEBUG_INFO=y.\n"); | 1788 | "please rebuild with CONFIG_DEBUG_INFO=y.\n"); |
| @@ -1450,8 +1828,7 @@ int find_line_range(int fd, struct line_range *lr) | |||
| 1450 | } | 1828 | } |
| 1451 | 1829 | ||
| 1452 | pr_debug("path: %s\n", lr->path); | 1830 | pr_debug("path: %s\n", lr->path); |
| 1453 | dwarf_end(dbg); | 1831 | dwfl_end(dwfl); |
| 1454 | |||
| 1455 | return (ret < 0) ? ret : lf.found; | 1832 | return (ret < 0) ? ret : lf.found; |
| 1456 | } | 1833 | } |
| 1457 | 1834 | ||
diff --git a/tools/perf/util/probe-finder.h b/tools/perf/util/probe-finder.h index 4507d519f183..bba69d455699 100644 --- a/tools/perf/util/probe-finder.h +++ b/tools/perf/util/probe-finder.h | |||
| @@ -22,20 +22,27 @@ extern int find_probe_trace_events(int fd, struct perf_probe_event *pev, | |||
| 22 | int max_tevs); | 22 | int max_tevs); |
| 23 | 23 | ||
| 24 | /* Find a perf_probe_point from debuginfo */ | 24 | /* Find a perf_probe_point from debuginfo */ |
| 25 | extern int find_perf_probe_point(int fd, unsigned long addr, | 25 | extern int find_perf_probe_point(unsigned long addr, |
| 26 | struct perf_probe_point *ppt); | 26 | struct perf_probe_point *ppt); |
| 27 | 27 | ||
| 28 | /* Find a line range */ | ||
| 28 | extern int find_line_range(int fd, struct line_range *lr); | 29 | extern int find_line_range(int fd, struct line_range *lr); |
| 29 | 30 | ||
| 31 | /* Find available variables */ | ||
| 32 | extern int find_available_vars_at(int fd, struct perf_probe_event *pev, | ||
| 33 | struct variable_list **vls, int max_points, | ||
| 34 | bool externs); | ||
| 35 | |||
| 30 | #include <dwarf.h> | 36 | #include <dwarf.h> |
| 31 | #include <libdw.h> | 37 | #include <libdw.h> |
| 38 | #include <libdwfl.h> | ||
| 32 | #include <version.h> | 39 | #include <version.h> |
| 33 | 40 | ||
| 34 | struct probe_finder { | 41 | struct probe_finder { |
| 35 | struct perf_probe_event *pev; /* Target probe event */ | 42 | struct perf_probe_event *pev; /* Target probe event */ |
| 36 | struct probe_trace_event *tevs; /* Result trace events */ | 43 | |
| 37 | int ntevs; /* Number of trace events */ | 44 | /* Callback when a probe point is found */ |
| 38 | int max_tevs; /* Max number of trace events */ | 45 | int (*callback)(Dwarf_Die *sp_die, struct probe_finder *pf); |
| 39 | 46 | ||
| 40 | /* For function searching */ | 47 | /* For function searching */ |
| 41 | int lno; /* Line number */ | 48 | int lno; /* Line number */ |
| @@ -53,6 +60,22 @@ struct probe_finder { | |||
| 53 | struct probe_trace_arg *tvar; /* Current result variable */ | 60 | struct probe_trace_arg *tvar; /* Current result variable */ |
| 54 | }; | 61 | }; |
| 55 | 62 | ||
| 63 | struct trace_event_finder { | ||
| 64 | struct probe_finder pf; | ||
| 65 | struct probe_trace_event *tevs; /* Found trace events */ | ||
| 66 | int ntevs; /* Number of trace events */ | ||
| 67 | int max_tevs; /* Max number of trace events */ | ||
| 68 | }; | ||
| 69 | |||
| 70 | struct available_var_finder { | ||
| 71 | struct probe_finder pf; | ||
| 72 | struct variable_list *vls; /* Found variable lists */ | ||
| 73 | int nvls; /* Number of variable lists */ | ||
| 74 | int max_vls; /* Max no. of variable lists */ | ||
| 75 | bool externs; /* Find external vars too */ | ||
| 76 | bool child; /* Search child scopes */ | ||
| 77 | }; | ||
| 78 | |||
| 56 | struct line_finder { | 79 | struct line_finder { |
| 57 | struct line_range *lr; /* Target line range */ | 80 | struct line_range *lr; /* Target line range */ |
| 58 | 81 | ||
diff --git a/tools/perf/util/ui/browser.c b/tools/perf/util/ui/browser.c index 6d0df809a2ed..8bc010edca25 100644 --- a/tools/perf/util/ui/browser.c +++ b/tools/perf/util/ui/browser.c | |||
| @@ -1,4 +1,3 @@ | |||
| 1 | #include <slang.h> | ||
| 2 | #include "libslang.h" | 1 | #include "libslang.h" |
| 3 | #include <linux/compiler.h> | 2 | #include <linux/compiler.h> |
| 4 | #include <linux/list.h> | 3 | #include <linux/list.h> |
