diff options
| author | Linus Torvalds <torvalds@linux-foundation.org> | 2011-07-22 19:44:39 -0400 |
|---|---|---|
| committer | Linus Torvalds <torvalds@linux-foundation.org> | 2011-07-22 19:44:39 -0400 |
| commit | 4d4abdcb1dee03a4f9d6d2021622ed07e14dfd17 (patch) | |
| tree | 4ed4c74b70240451065165fda5fb2059f8c6b1e5 | |
| parent | 0342cbcfced2ee937d7c8e1c63f3d3082da7c7dc (diff) | |
| parent | 7fcfd1abd6480d3b9ef17f5759c175e036e835cf (diff) | |
Merge branch 'perf-core-for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/tip/linux-2.6-tip
* 'perf-core-for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/tip/linux-2.6-tip: (123 commits)
perf: Remove the nmi parameter from the oprofile_perf backend
x86, perf: Make copy_from_user_nmi() a library function
perf: Remove perf_event_attr::type check
x86, perf: P4 PMU - Fix typos in comments and style cleanup
perf tools: Make test use the preset debugfs path
perf tools: Add automated tests for events parsing
perf tools: De-opt the parse_events function
perf script: Fix display of IP address for non-callchain path
perf tools: Fix endian conversion reading event attr from file header
perf tools: Add missing 'node' alias to the hw_cache[] array
perf probe: Support adding probes on offline kernel modules
perf probe: Add probed module in front of function
perf probe: Introduce debuginfo to encapsulate dwarf information
perf-probe: Move dwarf library routines to dwarf-aux.{c, h}
perf probe: Remove redundant dwarf functions
perf probe: Move strtailcmp to string.c
perf probe: Rename DIE_FIND_CB_FOUND to DIE_FIND_CB_END
tracing/kprobe: Update symbol reference when loading module
tracing/kprobes: Support module init function probing
kprobes: Return -ENOENT if probe point doesn't exist
...
140 files changed, 4650 insertions, 2661 deletions
diff --git a/Documentation/trace/kprobetrace.txt b/Documentation/trace/kprobetrace.txt index c83bd6b4e6e..d0d0bb9e3e2 100644 --- a/Documentation/trace/kprobetrace.txt +++ b/Documentation/trace/kprobetrace.txt | |||
| @@ -22,14 +22,15 @@ current_tracer. Instead of that, add probe points via | |||
| 22 | 22 | ||
| 23 | Synopsis of kprobe_events | 23 | Synopsis of kprobe_events |
| 24 | ------------------------- | 24 | ------------------------- |
| 25 | p[:[GRP/]EVENT] SYMBOL[+offs]|MEMADDR [FETCHARGS] : Set a probe | 25 | p[:[GRP/]EVENT] [MOD:]SYM[+offs]|MEMADDR [FETCHARGS] : Set a probe |
| 26 | r[:[GRP/]EVENT] SYMBOL[+0] [FETCHARGS] : Set a return probe | 26 | r[:[GRP/]EVENT] [MOD:]SYM[+0] [FETCHARGS] : Set a return probe |
| 27 | -:[GRP/]EVENT : Clear a probe | 27 | -:[GRP/]EVENT : Clear a probe |
| 28 | 28 | ||
| 29 | GRP : Group name. If omitted, use "kprobes" for it. | 29 | GRP : Group name. If omitted, use "kprobes" for it. |
| 30 | EVENT : Event name. If omitted, the event name is generated | 30 | EVENT : Event name. If omitted, the event name is generated |
| 31 | based on SYMBOL+offs or MEMADDR. | 31 | based on SYM+offs or MEMADDR. |
| 32 | SYMBOL[+offs] : Symbol+offset where the probe is inserted. | 32 | MOD : Module name which has given SYM. |
| 33 | SYM[+offs] : Symbol+offset where the probe is inserted. | ||
| 33 | MEMADDR : Address where the probe is inserted. | 34 | MEMADDR : Address where the probe is inserted. |
| 34 | 35 | ||
| 35 | FETCHARGS : Arguments. Each probe can have up to 128 args. | 36 | FETCHARGS : Arguments. Each probe can have up to 128 args. |
| @@ -1290,6 +1290,7 @@ help: | |||
| 1290 | @echo ' make O=dir [targets] Locate all output files in "dir", including .config' | 1290 | @echo ' make O=dir [targets] Locate all output files in "dir", including .config' |
| 1291 | @echo ' make C=1 [targets] Check all c source with $$CHECK (sparse by default)' | 1291 | @echo ' make C=1 [targets] Check all c source with $$CHECK (sparse by default)' |
| 1292 | @echo ' make C=2 [targets] Force check of all c source with $$CHECK' | 1292 | @echo ' make C=2 [targets] Force check of all c source with $$CHECK' |
| 1293 | @echo ' make RECORDMCOUNT_WARN=1 [targets] Warn about ignored mcount sections' | ||
| 1293 | @echo ' make W=n [targets] Enable extra gcc checks, n=1,2,3 where' | 1294 | @echo ' make W=n [targets] Enable extra gcc checks, n=1,2,3 where' |
| 1294 | @echo ' 1: warnings which may be relevant and do not occur too often' | 1295 | @echo ' 1: warnings which may be relevant and do not occur too often' |
| 1295 | @echo ' 2: warnings which occur quite often but may still be relevant' | 1296 | @echo ' 2: warnings which occur quite often but may still be relevant' |
diff --git a/arch/alpha/kernel/perf_event.c b/arch/alpha/kernel/perf_event.c index 90561c45e7d..8e47709160f 100644 --- a/arch/alpha/kernel/perf_event.c +++ b/arch/alpha/kernel/perf_event.c | |||
| @@ -847,7 +847,7 @@ static void alpha_perf_event_irq_handler(unsigned long la_ptr, | |||
| 847 | data.period = event->hw.last_period; | 847 | data.period = event->hw.last_period; |
| 848 | 848 | ||
| 849 | if (alpha_perf_event_set_period(event, hwc, idx)) { | 849 | if (alpha_perf_event_set_period(event, hwc, idx)) { |
| 850 | if (perf_event_overflow(event, 1, &data, regs)) { | 850 | if (perf_event_overflow(event, &data, regs)) { |
| 851 | /* Interrupts coming too quickly; "throttle" the | 851 | /* Interrupts coming too quickly; "throttle" the |
| 852 | * counter, i.e., disable it for a little while. | 852 | * counter, i.e., disable it for a little while. |
| 853 | */ | 853 | */ |
diff --git a/arch/alpha/kernel/time.c b/arch/alpha/kernel/time.c index 818e74ed45d..f20d1b5396b 100644 --- a/arch/alpha/kernel/time.c +++ b/arch/alpha/kernel/time.c | |||
| @@ -91,7 +91,7 @@ DEFINE_PER_CPU(u8, irq_work_pending); | |||
| 91 | #define test_irq_work_pending() __get_cpu_var(irq_work_pending) | 91 | #define test_irq_work_pending() __get_cpu_var(irq_work_pending) |
| 92 | #define clear_irq_work_pending() __get_cpu_var(irq_work_pending) = 0 | 92 | #define clear_irq_work_pending() __get_cpu_var(irq_work_pending) = 0 |
| 93 | 93 | ||
| 94 | void set_irq_work_pending(void) | 94 | void arch_irq_work_raise(void) |
| 95 | { | 95 | { |
| 96 | set_irq_work_pending_flag(); | 96 | set_irq_work_pending_flag(); |
| 97 | } | 97 | } |
diff --git a/arch/arm/kernel/perf_event_v6.c b/arch/arm/kernel/perf_event_v6.c index f1e8dd94afe..dd7f3b9f4cb 100644 --- a/arch/arm/kernel/perf_event_v6.c +++ b/arch/arm/kernel/perf_event_v6.c | |||
| @@ -173,6 +173,20 @@ static const unsigned armv6_perf_cache_map[PERF_COUNT_HW_CACHE_MAX] | |||
| 173 | [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED, | 173 | [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED, |
| 174 | }, | 174 | }, |
| 175 | }, | 175 | }, |
| 176 | [C(NODE)] = { | ||
| 177 | [C(OP_READ)] = { | ||
| 178 | [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED, | ||
| 179 | [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED, | ||
| 180 | }, | ||
| 181 | [C(OP_WRITE)] = { | ||
| 182 | [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED, | ||
| 183 | [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED, | ||
| 184 | }, | ||
| 185 | [C(OP_PREFETCH)] = { | ||
| 186 | [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED, | ||
| 187 | [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED, | ||
| 188 | }, | ||
| 189 | }, | ||
| 176 | }; | 190 | }; |
| 177 | 191 | ||
| 178 | enum armv6mpcore_perf_types { | 192 | enum armv6mpcore_perf_types { |
| @@ -310,6 +324,20 @@ static const unsigned armv6mpcore_perf_cache_map[PERF_COUNT_HW_CACHE_MAX] | |||
| 310 | [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED, | 324 | [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED, |
| 311 | }, | 325 | }, |
| 312 | }, | 326 | }, |
| 327 | [C(NODE)] = { | ||
| 328 | [C(OP_READ)] = { | ||
| 329 | [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED, | ||
| 330 | [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED, | ||
| 331 | }, | ||
| 332 | [C(OP_WRITE)] = { | ||
| 333 | [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED, | ||
| 334 | [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED, | ||
| 335 | }, | ||
| 336 | [C(OP_PREFETCH)] = { | ||
| 337 | [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED, | ||
| 338 | [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED, | ||
| 339 | }, | ||
| 340 | }, | ||
| 313 | }; | 341 | }; |
| 314 | 342 | ||
| 315 | static inline unsigned long | 343 | static inline unsigned long |
| @@ -479,7 +507,7 @@ armv6pmu_handle_irq(int irq_num, | |||
| 479 | if (!armpmu_event_set_period(event, hwc, idx)) | 507 | if (!armpmu_event_set_period(event, hwc, idx)) |
| 480 | continue; | 508 | continue; |
| 481 | 509 | ||
| 482 | if (perf_event_overflow(event, 0, &data, regs)) | 510 | if (perf_event_overflow(event, &data, regs)) |
| 483 | armpmu->disable(hwc, idx); | 511 | armpmu->disable(hwc, idx); |
| 484 | } | 512 | } |
| 485 | 513 | ||
diff --git a/arch/arm/kernel/perf_event_v7.c b/arch/arm/kernel/perf_event_v7.c index 4960686afb5..e20ca9cafef 100644 --- a/arch/arm/kernel/perf_event_v7.c +++ b/arch/arm/kernel/perf_event_v7.c | |||
| @@ -255,6 +255,20 @@ static const unsigned armv7_a8_perf_cache_map[PERF_COUNT_HW_CACHE_MAX] | |||
| 255 | [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED, | 255 | [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED, |
| 256 | }, | 256 | }, |
| 257 | }, | 257 | }, |
| 258 | [C(NODE)] = { | ||
| 259 | [C(OP_READ)] = { | ||
| 260 | [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED, | ||
| 261 | [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED, | ||
| 262 | }, | ||
| 263 | [C(OP_WRITE)] = { | ||
| 264 | [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED, | ||
| 265 | [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED, | ||
| 266 | }, | ||
| 267 | [C(OP_PREFETCH)] = { | ||
| 268 | [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED, | ||
| 269 | [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED, | ||
| 270 | }, | ||
| 271 | }, | ||
| 258 | }; | 272 | }; |
| 259 | 273 | ||
| 260 | /* | 274 | /* |
| @@ -371,6 +385,20 @@ static const unsigned armv7_a9_perf_cache_map[PERF_COUNT_HW_CACHE_MAX] | |||
| 371 | [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED, | 385 | [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED, |
| 372 | }, | 386 | }, |
| 373 | }, | 387 | }, |
| 388 | [C(NODE)] = { | ||
| 389 | [C(OP_READ)] = { | ||
| 390 | [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED, | ||
| 391 | [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED, | ||
| 392 | }, | ||
| 393 | [C(OP_WRITE)] = { | ||
| 394 | [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED, | ||
| 395 | [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED, | ||
| 396 | }, | ||
| 397 | [C(OP_PREFETCH)] = { | ||
| 398 | [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED, | ||
| 399 | [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED, | ||
| 400 | }, | ||
| 401 | }, | ||
| 374 | }; | 402 | }; |
| 375 | 403 | ||
| 376 | /* | 404 | /* |
| @@ -787,7 +815,7 @@ static irqreturn_t armv7pmu_handle_irq(int irq_num, void *dev) | |||
| 787 | if (!armpmu_event_set_period(event, hwc, idx)) | 815 | if (!armpmu_event_set_period(event, hwc, idx)) |
| 788 | continue; | 816 | continue; |
| 789 | 817 | ||
| 790 | if (perf_event_overflow(event, 0, &data, regs)) | 818 | if (perf_event_overflow(event, &data, regs)) |
| 791 | armpmu->disable(hwc, idx); | 819 | armpmu->disable(hwc, idx); |
| 792 | } | 820 | } |
| 793 | 821 | ||
diff --git a/arch/arm/kernel/perf_event_xscale.c b/arch/arm/kernel/perf_event_xscale.c index 39affbe4fdb..3c4397491d0 100644 --- a/arch/arm/kernel/perf_event_xscale.c +++ b/arch/arm/kernel/perf_event_xscale.c | |||
| @@ -144,6 +144,20 @@ static const unsigned xscale_perf_cache_map[PERF_COUNT_HW_CACHE_MAX] | |||
| 144 | [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED, | 144 | [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED, |
| 145 | }, | 145 | }, |
| 146 | }, | 146 | }, |
| 147 | [C(NODE)] = { | ||
| 148 | [C(OP_READ)] = { | ||
| 149 | [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED, | ||
| 150 | [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED, | ||
| 151 | }, | ||
| 152 | [C(OP_WRITE)] = { | ||
| 153 | [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED, | ||
| 154 | [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED, | ||
| 155 | }, | ||
| 156 | [C(OP_PREFETCH)] = { | ||
| 157 | [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED, | ||
| 158 | [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED, | ||
| 159 | }, | ||
| 160 | }, | ||
| 147 | }; | 161 | }; |
| 148 | 162 | ||
| 149 | #define XSCALE_PMU_ENABLE 0x001 | 163 | #define XSCALE_PMU_ENABLE 0x001 |
| @@ -251,7 +265,7 @@ xscale1pmu_handle_irq(int irq_num, void *dev) | |||
| 251 | if (!armpmu_event_set_period(event, hwc, idx)) | 265 | if (!armpmu_event_set_period(event, hwc, idx)) |
| 252 | continue; | 266 | continue; |
| 253 | 267 | ||
| 254 | if (perf_event_overflow(event, 0, &data, regs)) | 268 | if (perf_event_overflow(event, &data, regs)) |
| 255 | armpmu->disable(hwc, idx); | 269 | armpmu->disable(hwc, idx); |
| 256 | } | 270 | } |
| 257 | 271 | ||
| @@ -583,7 +597,7 @@ xscale2pmu_handle_irq(int irq_num, void *dev) | |||
| 583 | if (!armpmu_event_set_period(event, hwc, idx)) | 597 | if (!armpmu_event_set_period(event, hwc, idx)) |
| 584 | continue; | 598 | continue; |
| 585 | 599 | ||
| 586 | if (perf_event_overflow(event, 0, &data, regs)) | 600 | if (perf_event_overflow(event, &data, regs)) |
| 587 | armpmu->disable(hwc, idx); | 601 | armpmu->disable(hwc, idx); |
| 588 | } | 602 | } |
| 589 | 603 | ||
diff --git a/arch/arm/kernel/ptrace.c b/arch/arm/kernel/ptrace.c index 97260060bf2..5c199610719 100644 --- a/arch/arm/kernel/ptrace.c +++ b/arch/arm/kernel/ptrace.c | |||
| @@ -396,7 +396,7 @@ static long ptrace_hbp_idx_to_num(int idx) | |||
| 396 | /* | 396 | /* |
| 397 | * Handle hitting a HW-breakpoint. | 397 | * Handle hitting a HW-breakpoint. |
| 398 | */ | 398 | */ |
| 399 | static void ptrace_hbptriggered(struct perf_event *bp, int unused, | 399 | static void ptrace_hbptriggered(struct perf_event *bp, |
| 400 | struct perf_sample_data *data, | 400 | struct perf_sample_data *data, |
| 401 | struct pt_regs *regs) | 401 | struct pt_regs *regs) |
| 402 | { | 402 | { |
| @@ -479,7 +479,8 @@ static struct perf_event *ptrace_hbp_create(struct task_struct *tsk, int type) | |||
| 479 | attr.bp_type = type; | 479 | attr.bp_type = type; |
| 480 | attr.disabled = 1; | 480 | attr.disabled = 1; |
| 481 | 481 | ||
| 482 | return register_user_hw_breakpoint(&attr, ptrace_hbptriggered, tsk); | 482 | return register_user_hw_breakpoint(&attr, ptrace_hbptriggered, NULL, |
| 483 | tsk); | ||
| 483 | } | 484 | } |
| 484 | 485 | ||
| 485 | static int ptrace_gethbpregs(struct task_struct *tsk, long num, | 486 | static int ptrace_gethbpregs(struct task_struct *tsk, long num, |
diff --git a/arch/arm/kernel/swp_emulate.c b/arch/arm/kernel/swp_emulate.c index 40ee7e5045e..5f452f8fde0 100644 --- a/arch/arm/kernel/swp_emulate.c +++ b/arch/arm/kernel/swp_emulate.c | |||
| @@ -183,7 +183,7 @@ static int swp_handler(struct pt_regs *regs, unsigned int instr) | |||
| 183 | unsigned int address, destreg, data, type; | 183 | unsigned int address, destreg, data, type; |
| 184 | unsigned int res = 0; | 184 | unsigned int res = 0; |
| 185 | 185 | ||
| 186 | perf_sw_event(PERF_COUNT_SW_EMULATION_FAULTS, 1, 0, regs, regs->ARM_pc); | 186 | perf_sw_event(PERF_COUNT_SW_EMULATION_FAULTS, 1, regs, regs->ARM_pc); |
| 187 | 187 | ||
| 188 | if (current->pid != previous_pid) { | 188 | if (current->pid != previous_pid) { |
| 189 | pr_debug("\"%s\" (%ld) uses deprecated SWP{B} instruction\n", | 189 | pr_debug("\"%s\" (%ld) uses deprecated SWP{B} instruction\n", |
diff --git a/arch/arm/mm/fault.c b/arch/arm/mm/fault.c index bc0e1d88fd3..9ea4f7ddd66 100644 --- a/arch/arm/mm/fault.c +++ b/arch/arm/mm/fault.c | |||
| @@ -318,11 +318,11 @@ do_page_fault(unsigned long addr, unsigned int fsr, struct pt_regs *regs) | |||
| 318 | fault = __do_page_fault(mm, addr, fsr, tsk); | 318 | fault = __do_page_fault(mm, addr, fsr, tsk); |
| 319 | up_read(&mm->mmap_sem); | 319 | up_read(&mm->mmap_sem); |
| 320 | 320 | ||
| 321 | perf_sw_event(PERF_COUNT_SW_PAGE_FAULTS, 1, 0, regs, addr); | 321 | perf_sw_event(PERF_COUNT_SW_PAGE_FAULTS, 1, regs, addr); |
| 322 | if (fault & VM_FAULT_MAJOR) | 322 | if (fault & VM_FAULT_MAJOR) |
| 323 | perf_sw_event(PERF_COUNT_SW_PAGE_FAULTS_MAJ, 1, 0, regs, addr); | 323 | perf_sw_event(PERF_COUNT_SW_PAGE_FAULTS_MAJ, 1, regs, addr); |
| 324 | else if (fault & VM_FAULT_MINOR) | 324 | else if (fault & VM_FAULT_MINOR) |
| 325 | perf_sw_event(PERF_COUNT_SW_PAGE_FAULTS_MIN, 1, 0, regs, addr); | 325 | perf_sw_event(PERF_COUNT_SW_PAGE_FAULTS_MIN, 1, regs, addr); |
| 326 | 326 | ||
| 327 | /* | 327 | /* |
| 328 | * Handle the "normal" case first - VM_FAULT_MAJOR / VM_FAULT_MINOR | 328 | * Handle the "normal" case first - VM_FAULT_MAJOR / VM_FAULT_MINOR |
diff --git a/arch/mips/include/asm/stacktrace.h b/arch/mips/include/asm/stacktrace.h index 0bf82818aa5..780ee2c2a2a 100644 --- a/arch/mips/include/asm/stacktrace.h +++ b/arch/mips/include/asm/stacktrace.h | |||
| @@ -7,6 +7,10 @@ | |||
| 7 | extern int raw_show_trace; | 7 | extern int raw_show_trace; |
| 8 | extern unsigned long unwind_stack(struct task_struct *task, unsigned long *sp, | 8 | extern unsigned long unwind_stack(struct task_struct *task, unsigned long *sp, |
| 9 | unsigned long pc, unsigned long *ra); | 9 | unsigned long pc, unsigned long *ra); |
| 10 | extern unsigned long unwind_stack_by_address(unsigned long stack_page, | ||
| 11 | unsigned long *sp, | ||
| 12 | unsigned long pc, | ||
| 13 | unsigned long *ra); | ||
| 10 | #else | 14 | #else |
| 11 | #define raw_show_trace 1 | 15 | #define raw_show_trace 1 |
| 12 | static inline unsigned long unwind_stack(struct task_struct *task, | 16 | static inline unsigned long unwind_stack(struct task_struct *task, |
diff --git a/arch/mips/kernel/perf_event.c b/arch/mips/kernel/perf_event.c index a8244854d3d..d0deaab9ace 100644 --- a/arch/mips/kernel/perf_event.c +++ b/arch/mips/kernel/perf_event.c | |||
| @@ -527,7 +527,7 @@ handle_associated_event(struct cpu_hw_events *cpuc, | |||
| 527 | if (!mipspmu_event_set_period(event, hwc, idx)) | 527 | if (!mipspmu_event_set_period(event, hwc, idx)) |
| 528 | return; | 528 | return; |
| 529 | 529 | ||
| 530 | if (perf_event_overflow(event, 0, data, regs)) | 530 | if (perf_event_overflow(event, data, regs)) |
| 531 | mipspmu->disable_event(idx); | 531 | mipspmu->disable_event(idx); |
| 532 | } | 532 | } |
| 533 | 533 | ||
diff --git a/arch/mips/kernel/perf_event_mipsxx.c b/arch/mips/kernel/perf_event_mipsxx.c index 75266ff4cc3..e5ad09a9baf 100644 --- a/arch/mips/kernel/perf_event_mipsxx.c +++ b/arch/mips/kernel/perf_event_mipsxx.c | |||
| @@ -377,6 +377,20 @@ static const struct mips_perf_event mipsxxcore_cache_map | |||
| 377 | [C(RESULT_MISS)] = { UNSUPPORTED_PERF_EVENT_ID }, | 377 | [C(RESULT_MISS)] = { UNSUPPORTED_PERF_EVENT_ID }, |
| 378 | }, | 378 | }, |
| 379 | }, | 379 | }, |
| 380 | [C(NODE)] = { | ||
| 381 | [C(OP_READ)] = { | ||
| 382 | [C(RESULT_ACCESS)] = { UNSUPPORTED_PERF_EVENT_ID }, | ||
| 383 | [C(RESULT_MISS)] = { UNSUPPORTED_PERF_EVENT_ID }, | ||
| 384 | }, | ||
| 385 | [C(OP_WRITE)] = { | ||
| 386 | [C(RESULT_ACCESS)] = { UNSUPPORTED_PERF_EVENT_ID }, | ||
| 387 | [C(RESULT_MISS)] = { UNSUPPORTED_PERF_EVENT_ID }, | ||
| 388 | }, | ||
| 389 | [C(OP_PREFETCH)] = { | ||
| 390 | [C(RESULT_ACCESS)] = { UNSUPPORTED_PERF_EVENT_ID }, | ||
| 391 | [C(RESULT_MISS)] = { UNSUPPORTED_PERF_EVENT_ID }, | ||
| 392 | }, | ||
| 393 | }, | ||
| 380 | }; | 394 | }; |
| 381 | 395 | ||
| 382 | /* 74K core has completely different cache event map. */ | 396 | /* 74K core has completely different cache event map. */ |
| @@ -480,6 +494,20 @@ static const struct mips_perf_event mipsxx74Kcore_cache_map | |||
| 480 | [C(RESULT_MISS)] = { UNSUPPORTED_PERF_EVENT_ID }, | 494 | [C(RESULT_MISS)] = { UNSUPPORTED_PERF_EVENT_ID }, |
| 481 | }, | 495 | }, |
| 482 | }, | 496 | }, |
| 497 | [C(NODE)] = { | ||
| 498 | [C(OP_READ)] = { | ||
| 499 | [C(RESULT_ACCESS)] = { UNSUPPORTED_PERF_EVENT_ID }, | ||
| 500 | [C(RESULT_MISS)] = { UNSUPPORTED_PERF_EVENT_ID }, | ||
| 501 | }, | ||
| 502 | [C(OP_WRITE)] = { | ||
| 503 | [C(RESULT_ACCESS)] = { UNSUPPORTED_PERF_EVENT_ID }, | ||
| 504 | [C(RESULT_MISS)] = { UNSUPPORTED_PERF_EVENT_ID }, | ||
| 505 | }, | ||
| 506 | [C(OP_PREFETCH)] = { | ||
| 507 | [C(RESULT_ACCESS)] = { UNSUPPORTED_PERF_EVENT_ID }, | ||
| 508 | [C(RESULT_MISS)] = { UNSUPPORTED_PERF_EVENT_ID }, | ||
| 509 | }, | ||
| 510 | }, | ||
| 483 | }; | 511 | }; |
| 484 | 512 | ||
| 485 | #ifdef CONFIG_MIPS_MT_SMP | 513 | #ifdef CONFIG_MIPS_MT_SMP |
diff --git a/arch/mips/kernel/process.c b/arch/mips/kernel/process.c index d2112d3cf11..c28fbe6107b 100644 --- a/arch/mips/kernel/process.c +++ b/arch/mips/kernel/process.c | |||
| @@ -373,18 +373,18 @@ unsigned long thread_saved_pc(struct task_struct *tsk) | |||
| 373 | 373 | ||
| 374 | 374 | ||
| 375 | #ifdef CONFIG_KALLSYMS | 375 | #ifdef CONFIG_KALLSYMS |
| 376 | /* used by show_backtrace() */ | 376 | /* generic stack unwinding function */ |
| 377 | unsigned long unwind_stack(struct task_struct *task, unsigned long *sp, | 377 | unsigned long notrace unwind_stack_by_address(unsigned long stack_page, |
| 378 | unsigned long pc, unsigned long *ra) | 378 | unsigned long *sp, |
| 379 | unsigned long pc, | ||
| 380 | unsigned long *ra) | ||
| 379 | { | 381 | { |
| 380 | unsigned long stack_page; | ||
| 381 | struct mips_frame_info info; | 382 | struct mips_frame_info info; |
| 382 | unsigned long size, ofs; | 383 | unsigned long size, ofs; |
| 383 | int leaf; | 384 | int leaf; |
| 384 | extern void ret_from_irq(void); | 385 | extern void ret_from_irq(void); |
| 385 | extern void ret_from_exception(void); | 386 | extern void ret_from_exception(void); |
| 386 | 387 | ||
| 387 | stack_page = (unsigned long)task_stack_page(task); | ||
| 388 | if (!stack_page) | 388 | if (!stack_page) |
| 389 | return 0; | 389 | return 0; |
| 390 | 390 | ||
| @@ -443,6 +443,15 @@ unsigned long unwind_stack(struct task_struct *task, unsigned long *sp, | |||
| 443 | *ra = 0; | 443 | *ra = 0; |
| 444 | return __kernel_text_address(pc) ? pc : 0; | 444 | return __kernel_text_address(pc) ? pc : 0; |
| 445 | } | 445 | } |
| 446 | EXPORT_SYMBOL(unwind_stack_by_address); | ||
| 447 | |||
| 448 | /* used by show_backtrace() */ | ||
| 449 | unsigned long unwind_stack(struct task_struct *task, unsigned long *sp, | ||
| 450 | unsigned long pc, unsigned long *ra) | ||
| 451 | { | ||
| 452 | unsigned long stack_page = (unsigned long)task_stack_page(task); | ||
| 453 | return unwind_stack_by_address(stack_page, sp, pc, ra); | ||
| 454 | } | ||
| 446 | #endif | 455 | #endif |
| 447 | 456 | ||
| 448 | /* | 457 | /* |
diff --git a/arch/mips/kernel/traps.c b/arch/mips/kernel/traps.c index e9b3af27d84..b7517e3abc8 100644 --- a/arch/mips/kernel/traps.c +++ b/arch/mips/kernel/traps.c | |||
| @@ -578,12 +578,12 @@ static int simulate_llsc(struct pt_regs *regs, unsigned int opcode) | |||
| 578 | { | 578 | { |
| 579 | if ((opcode & OPCODE) == LL) { | 579 | if ((opcode & OPCODE) == LL) { |
| 580 | perf_sw_event(PERF_COUNT_SW_EMULATION_FAULTS, | 580 | perf_sw_event(PERF_COUNT_SW_EMULATION_FAULTS, |
| 581 | 1, 0, regs, 0); | 581 | 1, regs, 0); |
| 582 | return simulate_ll(regs, opcode); | 582 | return simulate_ll(regs, opcode); |
| 583 | } | 583 | } |
| 584 | if ((opcode & OPCODE) == SC) { | 584 | if ((opcode & OPCODE) == SC) { |
| 585 | perf_sw_event(PERF_COUNT_SW_EMULATION_FAULTS, | 585 | perf_sw_event(PERF_COUNT_SW_EMULATION_FAULTS, |
| 586 | 1, 0, regs, 0); | 586 | 1, regs, 0); |
| 587 | return simulate_sc(regs, opcode); | 587 | return simulate_sc(regs, opcode); |
| 588 | } | 588 | } |
| 589 | 589 | ||
| @@ -602,7 +602,7 @@ static int simulate_rdhwr(struct pt_regs *regs, unsigned int opcode) | |||
| 602 | int rd = (opcode & RD) >> 11; | 602 | int rd = (opcode & RD) >> 11; |
| 603 | int rt = (opcode & RT) >> 16; | 603 | int rt = (opcode & RT) >> 16; |
| 604 | perf_sw_event(PERF_COUNT_SW_EMULATION_FAULTS, | 604 | perf_sw_event(PERF_COUNT_SW_EMULATION_FAULTS, |
| 605 | 1, 0, regs, 0); | 605 | 1, regs, 0); |
| 606 | switch (rd) { | 606 | switch (rd) { |
| 607 | case 0: /* CPU number */ | 607 | case 0: /* CPU number */ |
| 608 | regs->regs[rt] = smp_processor_id(); | 608 | regs->regs[rt] = smp_processor_id(); |
| @@ -640,7 +640,7 @@ static int simulate_sync(struct pt_regs *regs, unsigned int opcode) | |||
| 640 | { | 640 | { |
| 641 | if ((opcode & OPCODE) == SPEC0 && (opcode & FUNC) == SYNC) { | 641 | if ((opcode & OPCODE) == SPEC0 && (opcode & FUNC) == SYNC) { |
| 642 | perf_sw_event(PERF_COUNT_SW_EMULATION_FAULTS, | 642 | perf_sw_event(PERF_COUNT_SW_EMULATION_FAULTS, |
| 643 | 1, 0, regs, 0); | 643 | 1, regs, 0); |
| 644 | return 0; | 644 | return 0; |
| 645 | } | 645 | } |
| 646 | 646 | ||
diff --git a/arch/mips/kernel/unaligned.c b/arch/mips/kernel/unaligned.c index cfea1adfa15..eb319b58035 100644 --- a/arch/mips/kernel/unaligned.c +++ b/arch/mips/kernel/unaligned.c | |||
| @@ -111,8 +111,7 @@ static void emulate_load_store_insn(struct pt_regs *regs, | |||
| 111 | unsigned long value; | 111 | unsigned long value; |
| 112 | unsigned int res; | 112 | unsigned int res; |
| 113 | 113 | ||
| 114 | perf_sw_event(PERF_COUNT_SW_EMULATION_FAULTS, | 114 | perf_sw_event(PERF_COUNT_SW_EMULATION_FAULTS, 1, regs, 0); |
| 115 | 1, 0, regs, 0); | ||
| 116 | 115 | ||
| 117 | /* | 116 | /* |
| 118 | * This load never faults. | 117 | * This load never faults. |
| @@ -517,7 +516,7 @@ asmlinkage void do_ade(struct pt_regs *regs) | |||
| 517 | mm_segment_t seg; | 516 | mm_segment_t seg; |
| 518 | 517 | ||
| 519 | perf_sw_event(PERF_COUNT_SW_ALIGNMENT_FAULTS, | 518 | perf_sw_event(PERF_COUNT_SW_ALIGNMENT_FAULTS, |
| 520 | 1, 0, regs, regs->cp0_badvaddr); | 519 | 1, regs, regs->cp0_badvaddr); |
| 521 | /* | 520 | /* |
| 522 | * Did we catch a fault trying to load an instruction? | 521 | * Did we catch a fault trying to load an instruction? |
| 523 | * Or are we running in MIPS16 mode? | 522 | * Or are we running in MIPS16 mode? |
diff --git a/arch/mips/math-emu/cp1emu.c b/arch/mips/math-emu/cp1emu.c index d32cb050311..dbf2f93a509 100644 --- a/arch/mips/math-emu/cp1emu.c +++ b/arch/mips/math-emu/cp1emu.c | |||
| @@ -272,8 +272,7 @@ static int cop1Emulate(struct pt_regs *xcp, struct mips_fpu_struct *ctx, | |||
| 272 | } | 272 | } |
| 273 | 273 | ||
| 274 | emul: | 274 | emul: |
| 275 | perf_sw_event(PERF_COUNT_SW_EMULATION_FAULTS, | 275 | perf_sw_event(PERF_COUNT_SW_EMULATION_FAULTS, 1, xcp, 0); |
| 276 | 1, 0, xcp, 0); | ||
| 277 | MIPS_FPU_EMU_INC_STATS(emulated); | 276 | MIPS_FPU_EMU_INC_STATS(emulated); |
| 278 | switch (MIPSInst_OPCODE(ir)) { | 277 | switch (MIPSInst_OPCODE(ir)) { |
| 279 | case ldc1_op:{ | 278 | case ldc1_op:{ |
diff --git a/arch/mips/mm/fault.c b/arch/mips/mm/fault.c index 137ee76a004..937cf336816 100644 --- a/arch/mips/mm/fault.c +++ b/arch/mips/mm/fault.c | |||
| @@ -145,7 +145,7 @@ good_area: | |||
| 145 | * the fault. | 145 | * the fault. |
| 146 | */ | 146 | */ |
| 147 | fault = handle_mm_fault(mm, vma, address, write ? FAULT_FLAG_WRITE : 0); | 147 | fault = handle_mm_fault(mm, vma, address, write ? FAULT_FLAG_WRITE : 0); |
| 148 | perf_sw_event(PERF_COUNT_SW_PAGE_FAULTS, 1, 0, regs, address); | 148 | perf_sw_event(PERF_COUNT_SW_PAGE_FAULTS, 1, regs, address); |
| 149 | if (unlikely(fault & VM_FAULT_ERROR)) { | 149 | if (unlikely(fault & VM_FAULT_ERROR)) { |
| 150 | if (fault & VM_FAULT_OOM) | 150 | if (fault & VM_FAULT_OOM) |
| 151 | goto out_of_memory; | 151 | goto out_of_memory; |
| @@ -154,12 +154,10 @@ good_area: | |||
| 154 | BUG(); | 154 | BUG(); |
| 155 | } | 155 | } |
| 156 | if (fault & VM_FAULT_MAJOR) { | 156 | if (fault & VM_FAULT_MAJOR) { |
| 157 | perf_sw_event(PERF_COUNT_SW_PAGE_FAULTS_MAJ, | 157 | perf_sw_event(PERF_COUNT_SW_PAGE_FAULTS_MAJ, 1, regs, address); |
| 158 | 1, 0, regs, address); | ||
| 159 | tsk->maj_flt++; | 158 | tsk->maj_flt++; |
| 160 | } else { | 159 | } else { |
| 161 | perf_sw_event(PERF_COUNT_SW_PAGE_FAULTS_MIN, | 160 | perf_sw_event(PERF_COUNT_SW_PAGE_FAULTS_MIN, 1, regs, address); |
| 162 | 1, 0, regs, address); | ||
| 163 | tsk->min_flt++; | 161 | tsk->min_flt++; |
| 164 | } | 162 | } |
| 165 | 163 | ||
diff --git a/arch/mips/oprofile/Makefile b/arch/mips/oprofile/Makefile index 4b9d7044e26..29f2f13eb31 100644 --- a/arch/mips/oprofile/Makefile +++ b/arch/mips/oprofile/Makefile | |||
| @@ -8,7 +8,7 @@ DRIVER_OBJS = $(addprefix ../../../drivers/oprofile/, \ | |||
| 8 | oprofilefs.o oprofile_stats.o \ | 8 | oprofilefs.o oprofile_stats.o \ |
| 9 | timer_int.o ) | 9 | timer_int.o ) |
| 10 | 10 | ||
| 11 | oprofile-y := $(DRIVER_OBJS) common.o | 11 | oprofile-y := $(DRIVER_OBJS) common.o backtrace.o |
| 12 | 12 | ||
| 13 | oprofile-$(CONFIG_CPU_MIPS32) += op_model_mipsxx.o | 13 | oprofile-$(CONFIG_CPU_MIPS32) += op_model_mipsxx.o |
| 14 | oprofile-$(CONFIG_CPU_MIPS64) += op_model_mipsxx.o | 14 | oprofile-$(CONFIG_CPU_MIPS64) += op_model_mipsxx.o |
diff --git a/arch/mips/oprofile/backtrace.c b/arch/mips/oprofile/backtrace.c new file mode 100644 index 00000000000..6854ed5097d --- /dev/null +++ b/arch/mips/oprofile/backtrace.c | |||
| @@ -0,0 +1,175 @@ | |||
| 1 | #include <linux/oprofile.h> | ||
| 2 | #include <linux/sched.h> | ||
| 3 | #include <linux/mm.h> | ||
| 4 | #include <linux/uaccess.h> | ||
| 5 | #include <asm/ptrace.h> | ||
| 6 | #include <asm/stacktrace.h> | ||
| 7 | #include <linux/stacktrace.h> | ||
| 8 | #include <linux/kernel.h> | ||
| 9 | #include <asm/sections.h> | ||
| 10 | #include <asm/inst.h> | ||
| 11 | |||
| 12 | struct stackframe { | ||
| 13 | unsigned long sp; | ||
| 14 | unsigned long pc; | ||
| 15 | unsigned long ra; | ||
| 16 | }; | ||
| 17 | |||
| 18 | static inline int get_mem(unsigned long addr, unsigned long *result) | ||
| 19 | { | ||
| 20 | unsigned long *address = (unsigned long *) addr; | ||
| 21 | if (!access_ok(VERIFY_READ, addr, sizeof(unsigned long))) | ||
| 22 | return -1; | ||
| 23 | if (__copy_from_user_inatomic(result, address, sizeof(unsigned long))) | ||
| 24 | return -3; | ||
| 25 | return 0; | ||
| 26 | } | ||
| 27 | |||
| 28 | /* | ||
| 29 | * These two instruction helpers were taken from process.c | ||
| 30 | */ | ||
| 31 | static inline int is_ra_save_ins(union mips_instruction *ip) | ||
| 32 | { | ||
| 33 | /* sw / sd $ra, offset($sp) */ | ||
| 34 | return (ip->i_format.opcode == sw_op || ip->i_format.opcode == sd_op) | ||
| 35 | && ip->i_format.rs == 29 && ip->i_format.rt == 31; | ||
| 36 | } | ||
| 37 | |||
| 38 | static inline int is_sp_move_ins(union mips_instruction *ip) | ||
| 39 | { | ||
| 40 | /* addiu/daddiu sp,sp,-imm */ | ||
| 41 | if (ip->i_format.rs != 29 || ip->i_format.rt != 29) | ||
| 42 | return 0; | ||
| 43 | if (ip->i_format.opcode == addiu_op || ip->i_format.opcode == daddiu_op) | ||
| 44 | return 1; | ||
| 45 | return 0; | ||
| 46 | } | ||
| 47 | |||
| 48 | /* | ||
| 49 | * Looks for specific instructions that mark the end of a function. | ||
| 50 | * This usually means we ran into the code area of the previous function. | ||
| 51 | */ | ||
| 52 | static inline int is_end_of_function_marker(union mips_instruction *ip) | ||
| 53 | { | ||
| 54 | /* jr ra */ | ||
| 55 | if (ip->r_format.func == jr_op && ip->r_format.rs == 31) | ||
| 56 | return 1; | ||
| 57 | /* lui gp */ | ||
| 58 | if (ip->i_format.opcode == lui_op && ip->i_format.rt == 28) | ||
| 59 | return 1; | ||
| 60 | return 0; | ||
| 61 | } | ||
| 62 | |||
| 63 | /* | ||
| 64 | * TODO for userspace stack unwinding: | ||
| 65 | * - handle cases where the stack is adjusted inside a function | ||
| 66 | * (generally doesn't happen) | ||
| 67 | * - find optimal value for max_instr_check | ||
| 68 | * - try to find a way to handle leaf functions | ||
| 69 | */ | ||
| 70 | |||
| 71 | static inline int unwind_user_frame(struct stackframe *old_frame, | ||
| 72 | const unsigned int max_instr_check) | ||
| 73 | { | ||
| 74 | struct stackframe new_frame = *old_frame; | ||
| 75 | off_t ra_offset = 0; | ||
| 76 | size_t stack_size = 0; | ||
| 77 | unsigned long addr; | ||
| 78 | |||
| 79 | if (old_frame->pc == 0 || old_frame->sp == 0 || old_frame->ra == 0) | ||
| 80 | return -9; | ||
| 81 | |||
| 82 | for (addr = new_frame.pc; (addr + max_instr_check > new_frame.pc) | ||
| 83 | && (!ra_offset || !stack_size); --addr) { | ||
| 84 | union mips_instruction ip; | ||
| 85 | |||
| 86 | if (get_mem(addr, (unsigned long *) &ip)) | ||
| 87 | return -11; | ||
| 88 | |||
| 89 | if (is_sp_move_ins(&ip)) { | ||
| 90 | int stack_adjustment = ip.i_format.simmediate; | ||
| 91 | if (stack_adjustment > 0) | ||
| 92 | /* This marks the end of the previous function, | ||
| 93 | which means we overran. */ | ||
| 94 | break; | ||
| 95 | stack_size = (unsigned) stack_adjustment; | ||
| 96 | } else if (is_ra_save_ins(&ip)) { | ||
| 97 | int ra_slot = ip.i_format.simmediate; | ||
| 98 | if (ra_slot < 0) | ||
| 99 | /* This shouldn't happen. */ | ||
| 100 | break; | ||
| 101 | ra_offset = ra_slot; | ||
| 102 | } else if (is_end_of_function_marker(&ip)) | ||
| 103 | break; | ||
| 104 | } | ||
| 105 | |||
| 106 | if (!ra_offset || !stack_size) | ||
| 107 | return -1; | ||
| 108 | |||
| 109 | if (ra_offset) { | ||
| 110 | new_frame.ra = old_frame->sp + ra_offset; | ||
| 111 | if (get_mem(new_frame.ra, &(new_frame.ra))) | ||
| 112 | return -13; | ||
| 113 | } | ||
| 114 | |||
| 115 | if (stack_size) { | ||
| 116 | new_frame.sp = old_frame->sp + stack_size; | ||
| 117 | if (get_mem(new_frame.sp, &(new_frame.sp))) | ||
| 118 | return -14; | ||
| 119 | } | ||
| 120 | |||
| 121 | if (new_frame.sp > old_frame->sp) | ||
| 122 | return -2; | ||
| 123 | |||
| 124 | new_frame.pc = old_frame->ra; | ||
| 125 | *old_frame = new_frame; | ||
| 126 | |||
| 127 | return 0; | ||
| 128 | } | ||
| 129 | |||
| 130 | static inline void do_user_backtrace(unsigned long low_addr, | ||
| 131 | struct stackframe *frame, | ||
| 132 | unsigned int depth) | ||
| 133 | { | ||
| 134 | const unsigned int max_instr_check = 512; | ||
| 135 | const unsigned long high_addr = low_addr + THREAD_SIZE; | ||
| 136 | |||
| 137 | while (depth-- && !unwind_user_frame(frame, max_instr_check)) { | ||
| 138 | oprofile_add_trace(frame->ra); | ||
| 139 | if (frame->sp < low_addr || frame->sp > high_addr) | ||
| 140 | break; | ||
| 141 | } | ||
| 142 | } | ||
| 143 | |||
| 144 | #ifndef CONFIG_KALLSYMS | ||
| 145 | static inline void do_kernel_backtrace(unsigned long low_addr, | ||
| 146 | struct stackframe *frame, | ||
| 147 | unsigned int depth) { } | ||
| 148 | #else | ||
| 149 | static inline void do_kernel_backtrace(unsigned long low_addr, | ||
| 150 | struct stackframe *frame, | ||
| 151 | unsigned int depth) | ||
| 152 | { | ||
| 153 | while (depth-- && frame->pc) { | ||
| 154 | frame->pc = unwind_stack_by_address(low_addr, | ||
| 155 | &(frame->sp), | ||
| 156 | frame->pc, | ||
| 157 | &(frame->ra)); | ||
| 158 | oprofile_add_trace(frame->ra); | ||
| 159 | } | ||
| 160 | } | ||
| 161 | #endif | ||
| 162 | |||
| 163 | void notrace op_mips_backtrace(struct pt_regs *const regs, unsigned int depth) | ||
| 164 | { | ||
| 165 | struct stackframe frame = { .sp = regs->regs[29], | ||
| 166 | .pc = regs->cp0_epc, | ||
| 167 | .ra = regs->regs[31] }; | ||
| 168 | const int userspace = user_mode(regs); | ||
| 169 | const unsigned long low_addr = ALIGN(frame.sp, THREAD_SIZE); | ||
| 170 | |||
| 171 | if (userspace) | ||
| 172 | do_user_backtrace(low_addr, &frame, depth); | ||
| 173 | else | ||
| 174 | do_kernel_backtrace(low_addr, &frame, depth); | ||
| 175 | } | ||
diff --git a/arch/mips/oprofile/common.c b/arch/mips/oprofile/common.c index f9eb1aba634..d1f2d4c52d4 100644 --- a/arch/mips/oprofile/common.c +++ b/arch/mips/oprofile/common.c | |||
| @@ -115,6 +115,7 @@ int __init oprofile_arch_init(struct oprofile_operations *ops) | |||
| 115 | ops->start = op_mips_start; | 115 | ops->start = op_mips_start; |
| 116 | ops->stop = op_mips_stop; | 116 | ops->stop = op_mips_stop; |
| 117 | ops->cpu_type = lmodel->cpu_type; | 117 | ops->cpu_type = lmodel->cpu_type; |
| 118 | ops->backtrace = op_mips_backtrace; | ||
| 118 | 119 | ||
| 119 | printk(KERN_INFO "oprofile: using %s performance monitoring.\n", | 120 | printk(KERN_INFO "oprofile: using %s performance monitoring.\n", |
| 120 | lmodel->cpu_type); | 121 | lmodel->cpu_type); |
diff --git a/arch/mips/oprofile/op_impl.h b/arch/mips/oprofile/op_impl.h index f04b54fb37d..7c2da27ece0 100644 --- a/arch/mips/oprofile/op_impl.h +++ b/arch/mips/oprofile/op_impl.h | |||
| @@ -36,4 +36,6 @@ struct op_mips_model { | |||
| 36 | unsigned char num_counters; | 36 | unsigned char num_counters; |
| 37 | }; | 37 | }; |
| 38 | 38 | ||
| 39 | void op_mips_backtrace(struct pt_regs * const regs, unsigned int depth); | ||
| 40 | |||
| 39 | #endif | 41 | #endif |
diff --git a/arch/powerpc/include/asm/emulated_ops.h b/arch/powerpc/include/asm/emulated_ops.h index 45921672b97..2cc41c715d2 100644 --- a/arch/powerpc/include/asm/emulated_ops.h +++ b/arch/powerpc/include/asm/emulated_ops.h | |||
| @@ -78,14 +78,14 @@ extern void ppc_warn_emulated_print(const char *type); | |||
| 78 | #define PPC_WARN_EMULATED(type, regs) \ | 78 | #define PPC_WARN_EMULATED(type, regs) \ |
| 79 | do { \ | 79 | do { \ |
| 80 | perf_sw_event(PERF_COUNT_SW_EMULATION_FAULTS, \ | 80 | perf_sw_event(PERF_COUNT_SW_EMULATION_FAULTS, \ |
| 81 | 1, 0, regs, 0); \ | 81 | 1, regs, 0); \ |
| 82 | __PPC_WARN_EMULATED(type); \ | 82 | __PPC_WARN_EMULATED(type); \ |
| 83 | } while (0) | 83 | } while (0) |
| 84 | 84 | ||
| 85 | #define PPC_WARN_ALIGNMENT(type, regs) \ | 85 | #define PPC_WARN_ALIGNMENT(type, regs) \ |
| 86 | do { \ | 86 | do { \ |
| 87 | perf_sw_event(PERF_COUNT_SW_ALIGNMENT_FAULTS, \ | 87 | perf_sw_event(PERF_COUNT_SW_ALIGNMENT_FAULTS, \ |
| 88 | 1, 0, regs, regs->dar); \ | 88 | 1, regs, regs->dar); \ |
| 89 | __PPC_WARN_EMULATED(type); \ | 89 | __PPC_WARN_EMULATED(type); \ |
| 90 | } while (0) | 90 | } while (0) |
| 91 | 91 | ||
diff --git a/arch/powerpc/include/asm/hw_breakpoint.h b/arch/powerpc/include/asm/hw_breakpoint.h index 1c33ec17ca3..80fd4d2b4a6 100644 --- a/arch/powerpc/include/asm/hw_breakpoint.h +++ b/arch/powerpc/include/asm/hw_breakpoint.h | |||
| @@ -57,7 +57,7 @@ void hw_breakpoint_pmu_read(struct perf_event *bp); | |||
| 57 | extern void flush_ptrace_hw_breakpoint(struct task_struct *tsk); | 57 | extern void flush_ptrace_hw_breakpoint(struct task_struct *tsk); |
| 58 | 58 | ||
| 59 | extern struct pmu perf_ops_bp; | 59 | extern struct pmu perf_ops_bp; |
| 60 | extern void ptrace_triggered(struct perf_event *bp, int nmi, | 60 | extern void ptrace_triggered(struct perf_event *bp, |
| 61 | struct perf_sample_data *data, struct pt_regs *regs); | 61 | struct perf_sample_data *data, struct pt_regs *regs); |
| 62 | static inline void hw_breakpoint_disable(void) | 62 | static inline void hw_breakpoint_disable(void) |
| 63 | { | 63 | { |
diff --git a/arch/powerpc/kernel/e500-pmu.c b/arch/powerpc/kernel/e500-pmu.c index b150b510510..cb2e2949c8d 100644 --- a/arch/powerpc/kernel/e500-pmu.c +++ b/arch/powerpc/kernel/e500-pmu.c | |||
| @@ -75,6 +75,11 @@ static int e500_cache_events[C(MAX)][C(OP_MAX)][C(RESULT_MAX)] = { | |||
| 75 | [C(OP_WRITE)] = { -1, -1 }, | 75 | [C(OP_WRITE)] = { -1, -1 }, |
| 76 | [C(OP_PREFETCH)] = { -1, -1 }, | 76 | [C(OP_PREFETCH)] = { -1, -1 }, |
| 77 | }, | 77 | }, |
| 78 | [C(NODE)] = { /* RESULT_ACCESS RESULT_MISS */ | ||
| 79 | [C(OP_READ)] = { -1, -1 }, | ||
| 80 | [C(OP_WRITE)] = { -1, -1 }, | ||
| 81 | [C(OP_PREFETCH)] = { -1, -1 }, | ||
| 82 | }, | ||
| 78 | }; | 83 | }; |
| 79 | 84 | ||
| 80 | static int num_events = 128; | 85 | static int num_events = 128; |
diff --git a/arch/powerpc/kernel/mpc7450-pmu.c b/arch/powerpc/kernel/mpc7450-pmu.c index 2cc5e0301d0..845a5847889 100644 --- a/arch/powerpc/kernel/mpc7450-pmu.c +++ b/arch/powerpc/kernel/mpc7450-pmu.c | |||
| @@ -388,6 +388,11 @@ static int mpc7450_cache_events[C(MAX)][C(OP_MAX)][C(RESULT_MAX)] = { | |||
| 388 | [C(OP_WRITE)] = { -1, -1 }, | 388 | [C(OP_WRITE)] = { -1, -1 }, |
| 389 | [C(OP_PREFETCH)] = { -1, -1 }, | 389 | [C(OP_PREFETCH)] = { -1, -1 }, |
| 390 | }, | 390 | }, |
| 391 | [C(NODE)] = { /* RESULT_ACCESS RESULT_MISS */ | ||
| 392 | [C(OP_READ)] = { -1, -1 }, | ||
| 393 | [C(OP_WRITE)] = { -1, -1 }, | ||
| 394 | [C(OP_PREFETCH)] = { -1, -1 }, | ||
| 395 | }, | ||
| 391 | }; | 396 | }; |
| 392 | 397 | ||
| 393 | struct power_pmu mpc7450_pmu = { | 398 | struct power_pmu mpc7450_pmu = { |
diff --git a/arch/powerpc/kernel/perf_event.c b/arch/powerpc/kernel/perf_event.c index 822f63008ae..14967de9887 100644 --- a/arch/powerpc/kernel/perf_event.c +++ b/arch/powerpc/kernel/perf_event.c | |||
| @@ -1207,7 +1207,7 @@ struct pmu power_pmu = { | |||
| 1207 | * here so there is no possibility of being interrupted. | 1207 | * here so there is no possibility of being interrupted. |
| 1208 | */ | 1208 | */ |
| 1209 | static void record_and_restart(struct perf_event *event, unsigned long val, | 1209 | static void record_and_restart(struct perf_event *event, unsigned long val, |
| 1210 | struct pt_regs *regs, int nmi) | 1210 | struct pt_regs *regs) |
| 1211 | { | 1211 | { |
| 1212 | u64 period = event->hw.sample_period; | 1212 | u64 period = event->hw.sample_period; |
| 1213 | s64 prev, delta, left; | 1213 | s64 prev, delta, left; |
| @@ -1258,7 +1258,7 @@ static void record_and_restart(struct perf_event *event, unsigned long val, | |||
| 1258 | if (event->attr.sample_type & PERF_SAMPLE_ADDR) | 1258 | if (event->attr.sample_type & PERF_SAMPLE_ADDR) |
| 1259 | perf_get_data_addr(regs, &data.addr); | 1259 | perf_get_data_addr(regs, &data.addr); |
| 1260 | 1260 | ||
| 1261 | if (perf_event_overflow(event, nmi, &data, regs)) | 1261 | if (perf_event_overflow(event, &data, regs)) |
| 1262 | power_pmu_stop(event, 0); | 1262 | power_pmu_stop(event, 0); |
| 1263 | } | 1263 | } |
| 1264 | } | 1264 | } |
| @@ -1346,7 +1346,7 @@ static void perf_event_interrupt(struct pt_regs *regs) | |||
| 1346 | if ((int)val < 0) { | 1346 | if ((int)val < 0) { |
| 1347 | /* event has overflowed */ | 1347 | /* event has overflowed */ |
| 1348 | found = 1; | 1348 | found = 1; |
| 1349 | record_and_restart(event, val, regs, nmi); | 1349 | record_and_restart(event, val, regs); |
| 1350 | } | 1350 | } |
| 1351 | } | 1351 | } |
| 1352 | 1352 | ||
diff --git a/arch/powerpc/kernel/perf_event_fsl_emb.c b/arch/powerpc/kernel/perf_event_fsl_emb.c index b0dc8f7069c..0a6d2a9d569 100644 --- a/arch/powerpc/kernel/perf_event_fsl_emb.c +++ b/arch/powerpc/kernel/perf_event_fsl_emb.c | |||
| @@ -568,7 +568,7 @@ static struct pmu fsl_emb_pmu = { | |||
| 568 | * here so there is no possibility of being interrupted. | 568 | * here so there is no possibility of being interrupted. |
| 569 | */ | 569 | */ |
| 570 | static void record_and_restart(struct perf_event *event, unsigned long val, | 570 | static void record_and_restart(struct perf_event *event, unsigned long val, |
| 571 | struct pt_regs *regs, int nmi) | 571 | struct pt_regs *regs) |
| 572 | { | 572 | { |
| 573 | u64 period = event->hw.sample_period; | 573 | u64 period = event->hw.sample_period; |
| 574 | s64 prev, delta, left; | 574 | s64 prev, delta, left; |
| @@ -616,7 +616,7 @@ static void record_and_restart(struct perf_event *event, unsigned long val, | |||
| 616 | perf_sample_data_init(&data, 0); | 616 | perf_sample_data_init(&data, 0); |
| 617 | data.period = event->hw.last_period; | 617 | data.period = event->hw.last_period; |
| 618 | 618 | ||
| 619 | if (perf_event_overflow(event, nmi, &data, regs)) | 619 | if (perf_event_overflow(event, &data, regs)) |
| 620 | fsl_emb_pmu_stop(event, 0); | 620 | fsl_emb_pmu_stop(event, 0); |
| 621 | } | 621 | } |
| 622 | } | 622 | } |
| @@ -644,7 +644,7 @@ static void perf_event_interrupt(struct pt_regs *regs) | |||
| 644 | if (event) { | 644 | if (event) { |
| 645 | /* event has overflowed */ | 645 | /* event has overflowed */ |
| 646 | found = 1; | 646 | found = 1; |
| 647 | record_and_restart(event, val, regs, nmi); | 647 | record_and_restart(event, val, regs); |
| 648 | } else { | 648 | } else { |
| 649 | /* | 649 | /* |
| 650 | * Disabled counter is negative, | 650 | * Disabled counter is negative, |
diff --git a/arch/powerpc/kernel/power4-pmu.c b/arch/powerpc/kernel/power4-pmu.c index ead8b3c2649..e9dbc2d35c9 100644 --- a/arch/powerpc/kernel/power4-pmu.c +++ b/arch/powerpc/kernel/power4-pmu.c | |||
| @@ -587,6 +587,11 @@ static int power4_cache_events[C(MAX)][C(OP_MAX)][C(RESULT_MAX)] = { | |||
| 587 | [C(OP_WRITE)] = { -1, -1 }, | 587 | [C(OP_WRITE)] = { -1, -1 }, |
| 588 | [C(OP_PREFETCH)] = { -1, -1 }, | 588 | [C(OP_PREFETCH)] = { -1, -1 }, |
| 589 | }, | 589 | }, |
| 590 | [C(NODE)] = { /* RESULT_ACCESS RESULT_MISS */ | ||
| 591 | [C(OP_READ)] = { -1, -1 }, | ||
| 592 | [C(OP_WRITE)] = { -1, -1 }, | ||
| 593 | [C(OP_PREFETCH)] = { -1, -1 }, | ||
| 594 | }, | ||
| 590 | }; | 595 | }; |
| 591 | 596 | ||
| 592 | static struct power_pmu power4_pmu = { | 597 | static struct power_pmu power4_pmu = { |
diff --git a/arch/powerpc/kernel/power5+-pmu.c b/arch/powerpc/kernel/power5+-pmu.c index eca0ac595cb..f58a2bd41b5 100644 --- a/arch/powerpc/kernel/power5+-pmu.c +++ b/arch/powerpc/kernel/power5+-pmu.c | |||
| @@ -653,6 +653,11 @@ static int power5p_cache_events[C(MAX)][C(OP_MAX)][C(RESULT_MAX)] = { | |||
| 653 | [C(OP_WRITE)] = { -1, -1 }, | 653 | [C(OP_WRITE)] = { -1, -1 }, |
| 654 | [C(OP_PREFETCH)] = { -1, -1 }, | 654 | [C(OP_PREFETCH)] = { -1, -1 }, |
| 655 | }, | 655 | }, |
| 656 | [C(NODE)] = { /* RESULT_ACCESS RESULT_MISS */ | ||
| 657 | [C(OP_READ)] = { -1, -1 }, | ||
| 658 | [C(OP_WRITE)] = { -1, -1 }, | ||
| 659 | [C(OP_PREFETCH)] = { -1, -1 }, | ||
| 660 | }, | ||
| 656 | }; | 661 | }; |
| 657 | 662 | ||
| 658 | static struct power_pmu power5p_pmu = { | 663 | static struct power_pmu power5p_pmu = { |
diff --git a/arch/powerpc/kernel/power5-pmu.c b/arch/powerpc/kernel/power5-pmu.c index d5ff0f64a5e..b1acab68414 100644 --- a/arch/powerpc/kernel/power5-pmu.c +++ b/arch/powerpc/kernel/power5-pmu.c | |||
| @@ -595,6 +595,11 @@ static int power5_cache_events[C(MAX)][C(OP_MAX)][C(RESULT_MAX)] = { | |||
| 595 | [C(OP_WRITE)] = { -1, -1 }, | 595 | [C(OP_WRITE)] = { -1, -1 }, |
| 596 | [C(OP_PREFETCH)] = { -1, -1 }, | 596 | [C(OP_PREFETCH)] = { -1, -1 }, |
| 597 | }, | 597 | }, |
| 598 | [C(NODE)] = { /* RESULT_ACCESS RESULT_MISS */ | ||
| 599 | [C(OP_READ)] = { -1, -1 }, | ||
| 600 | [C(OP_WRITE)] = { -1, -1 }, | ||
| 601 | [C(OP_PREFETCH)] = { -1, -1 }, | ||
| 602 | }, | ||
| 598 | }; | 603 | }; |
| 599 | 604 | ||
| 600 | static struct power_pmu power5_pmu = { | 605 | static struct power_pmu power5_pmu = { |
diff --git a/arch/powerpc/kernel/power6-pmu.c b/arch/powerpc/kernel/power6-pmu.c index 31603927e37..b24a3a23d07 100644 --- a/arch/powerpc/kernel/power6-pmu.c +++ b/arch/powerpc/kernel/power6-pmu.c | |||
| @@ -516,6 +516,11 @@ static int power6_cache_events[C(MAX)][C(OP_MAX)][C(RESULT_MAX)] = { | |||
| 516 | [C(OP_WRITE)] = { -1, -1 }, | 516 | [C(OP_WRITE)] = { -1, -1 }, |
| 517 | [C(OP_PREFETCH)] = { -1, -1 }, | 517 | [C(OP_PREFETCH)] = { -1, -1 }, |
| 518 | }, | 518 | }, |
| 519 | [C(NODE)] = { /* RESULT_ACCESS RESULT_MISS */ | ||
| 520 | [C(OP_READ)] = { -1, -1 }, | ||
| 521 | [C(OP_WRITE)] = { -1, -1 }, | ||
| 522 | [C(OP_PREFETCH)] = { -1, -1 }, | ||
| 523 | }, | ||
| 519 | }; | 524 | }; |
| 520 | 525 | ||
| 521 | static struct power_pmu power6_pmu = { | 526 | static struct power_pmu power6_pmu = { |
diff --git a/arch/powerpc/kernel/power7-pmu.c b/arch/powerpc/kernel/power7-pmu.c index 593740fcb79..6d9dccb2ea5 100644 --- a/arch/powerpc/kernel/power7-pmu.c +++ b/arch/powerpc/kernel/power7-pmu.c | |||
| @@ -342,6 +342,11 @@ static int power7_cache_events[C(MAX)][C(OP_MAX)][C(RESULT_MAX)] = { | |||
| 342 | [C(OP_WRITE)] = { -1, -1 }, | 342 | [C(OP_WRITE)] = { -1, -1 }, |
| 343 | [C(OP_PREFETCH)] = { -1, -1 }, | 343 | [C(OP_PREFETCH)] = { -1, -1 }, |
| 344 | }, | 344 | }, |
| 345 | [C(NODE)] = { /* RESULT_ACCESS RESULT_MISS */ | ||
| 346 | [C(OP_READ)] = { -1, -1 }, | ||
| 347 | [C(OP_WRITE)] = { -1, -1 }, | ||
| 348 | [C(OP_PREFETCH)] = { -1, -1 }, | ||
| 349 | }, | ||
| 345 | }; | 350 | }; |
| 346 | 351 | ||
| 347 | static struct power_pmu power7_pmu = { | 352 | static struct power_pmu power7_pmu = { |
diff --git a/arch/powerpc/kernel/ppc970-pmu.c b/arch/powerpc/kernel/ppc970-pmu.c index 9a6e093858f..b121de9658e 100644 --- a/arch/powerpc/kernel/ppc970-pmu.c +++ b/arch/powerpc/kernel/ppc970-pmu.c | |||
| @@ -467,6 +467,11 @@ static int ppc970_cache_events[C(MAX)][C(OP_MAX)][C(RESULT_MAX)] = { | |||
| 467 | [C(OP_WRITE)] = { -1, -1 }, | 467 | [C(OP_WRITE)] = { -1, -1 }, |
| 468 | [C(OP_PREFETCH)] = { -1, -1 }, | 468 | [C(OP_PREFETCH)] = { -1, -1 }, |
| 469 | }, | 469 | }, |
| 470 | [C(NODE)] = { /* RESULT_ACCESS RESULT_MISS */ | ||
| 471 | [C(OP_READ)] = { -1, -1 }, | ||
| 472 | [C(OP_WRITE)] = { -1, -1 }, | ||
| 473 | [C(OP_PREFETCH)] = { -1, -1 }, | ||
| 474 | }, | ||
| 470 | }; | 475 | }; |
| 471 | 476 | ||
| 472 | static struct power_pmu ppc970_pmu = { | 477 | static struct power_pmu ppc970_pmu = { |
diff --git a/arch/powerpc/kernel/ptrace.c b/arch/powerpc/kernel/ptrace.c index cb22024f2b4..05b7dd217f6 100644 --- a/arch/powerpc/kernel/ptrace.c +++ b/arch/powerpc/kernel/ptrace.c | |||
| @@ -882,7 +882,7 @@ void user_disable_single_step(struct task_struct *task) | |||
| 882 | } | 882 | } |
| 883 | 883 | ||
| 884 | #ifdef CONFIG_HAVE_HW_BREAKPOINT | 884 | #ifdef CONFIG_HAVE_HW_BREAKPOINT |
| 885 | void ptrace_triggered(struct perf_event *bp, int nmi, | 885 | void ptrace_triggered(struct perf_event *bp, |
| 886 | struct perf_sample_data *data, struct pt_regs *regs) | 886 | struct perf_sample_data *data, struct pt_regs *regs) |
| 887 | { | 887 | { |
| 888 | struct perf_event_attr attr; | 888 | struct perf_event_attr attr; |
| @@ -973,7 +973,7 @@ int ptrace_set_debugreg(struct task_struct *task, unsigned long addr, | |||
| 973 | &attr.bp_type); | 973 | &attr.bp_type); |
| 974 | 974 | ||
| 975 | thread->ptrace_bps[0] = bp = register_user_hw_breakpoint(&attr, | 975 | thread->ptrace_bps[0] = bp = register_user_hw_breakpoint(&attr, |
| 976 | ptrace_triggered, task); | 976 | ptrace_triggered, NULL, task); |
| 977 | if (IS_ERR(bp)) { | 977 | if (IS_ERR(bp)) { |
| 978 | thread->ptrace_bps[0] = NULL; | 978 | thread->ptrace_bps[0] = NULL; |
| 979 | ptrace_put_breakpoints(task); | 979 | ptrace_put_breakpoints(task); |
diff --git a/arch/powerpc/kernel/time.c b/arch/powerpc/kernel/time.c index f33acfd872a..03b29a6759a 100644 --- a/arch/powerpc/kernel/time.c +++ b/arch/powerpc/kernel/time.c | |||
| @@ -544,7 +544,7 @@ DEFINE_PER_CPU(u8, irq_work_pending); | |||
| 544 | 544 | ||
| 545 | #endif /* 32 vs 64 bit */ | 545 | #endif /* 32 vs 64 bit */ |
| 546 | 546 | ||
| 547 | void set_irq_work_pending(void) | 547 | void arch_irq_work_raise(void) |
| 548 | { | 548 | { |
| 549 | preempt_disable(); | 549 | preempt_disable(); |
| 550 | set_irq_work_pending_flag(); | 550 | set_irq_work_pending_flag(); |
diff --git a/arch/powerpc/mm/fault.c b/arch/powerpc/mm/fault.c index ad35f66c69e..5efe8c96d37 100644 --- a/arch/powerpc/mm/fault.c +++ b/arch/powerpc/mm/fault.c | |||
| @@ -174,7 +174,7 @@ int __kprobes do_page_fault(struct pt_regs *regs, unsigned long address, | |||
| 174 | die("Weird page fault", regs, SIGSEGV); | 174 | die("Weird page fault", regs, SIGSEGV); |
| 175 | } | 175 | } |
| 176 | 176 | ||
| 177 | perf_sw_event(PERF_COUNT_SW_PAGE_FAULTS, 1, 0, regs, address); | 177 | perf_sw_event(PERF_COUNT_SW_PAGE_FAULTS, 1, regs, address); |
| 178 | 178 | ||
| 179 | /* When running in the kernel we expect faults to occur only to | 179 | /* When running in the kernel we expect faults to occur only to |
| 180 | * addresses in user space. All other faults represent errors in the | 180 | * addresses in user space. All other faults represent errors in the |
| @@ -320,7 +320,7 @@ good_area: | |||
| 320 | } | 320 | } |
| 321 | if (ret & VM_FAULT_MAJOR) { | 321 | if (ret & VM_FAULT_MAJOR) { |
| 322 | current->maj_flt++; | 322 | current->maj_flt++; |
| 323 | perf_sw_event(PERF_COUNT_SW_PAGE_FAULTS_MAJ, 1, 0, | 323 | perf_sw_event(PERF_COUNT_SW_PAGE_FAULTS_MAJ, 1, |
| 324 | regs, address); | 324 | regs, address); |
| 325 | #ifdef CONFIG_PPC_SMLPAR | 325 | #ifdef CONFIG_PPC_SMLPAR |
| 326 | if (firmware_has_feature(FW_FEATURE_CMO)) { | 326 | if (firmware_has_feature(FW_FEATURE_CMO)) { |
| @@ -331,7 +331,7 @@ good_area: | |||
| 331 | #endif | 331 | #endif |
| 332 | } else { | 332 | } else { |
| 333 | current->min_flt++; | 333 | current->min_flt++; |
| 334 | perf_sw_event(PERF_COUNT_SW_PAGE_FAULTS_MIN, 1, 0, | 334 | perf_sw_event(PERF_COUNT_SW_PAGE_FAULTS_MIN, 1, |
| 335 | regs, address); | 335 | regs, address); |
| 336 | } | 336 | } |
| 337 | up_read(&mm->mmap_sem); | 337 | up_read(&mm->mmap_sem); |
diff --git a/arch/s390/mm/fault.c b/arch/s390/mm/fault.c index fe103e891e7..095f782a551 100644 --- a/arch/s390/mm/fault.c +++ b/arch/s390/mm/fault.c | |||
| @@ -299,7 +299,7 @@ static inline int do_exception(struct pt_regs *regs, int access, | |||
| 299 | goto out; | 299 | goto out; |
| 300 | 300 | ||
| 301 | address = trans_exc_code & __FAIL_ADDR_MASK; | 301 | address = trans_exc_code & __FAIL_ADDR_MASK; |
| 302 | perf_sw_event(PERF_COUNT_SW_PAGE_FAULTS, 1, 0, regs, address); | 302 | perf_sw_event(PERF_COUNT_SW_PAGE_FAULTS, 1, regs, address); |
| 303 | flags = FAULT_FLAG_ALLOW_RETRY; | 303 | flags = FAULT_FLAG_ALLOW_RETRY; |
| 304 | if (access == VM_WRITE || (trans_exc_code & store_indication) == 0x400) | 304 | if (access == VM_WRITE || (trans_exc_code & store_indication) == 0x400) |
| 305 | flags |= FAULT_FLAG_WRITE; | 305 | flags |= FAULT_FLAG_WRITE; |
| @@ -345,11 +345,11 @@ retry: | |||
| 345 | if (flags & FAULT_FLAG_ALLOW_RETRY) { | 345 | if (flags & FAULT_FLAG_ALLOW_RETRY) { |
| 346 | if (fault & VM_FAULT_MAJOR) { | 346 | if (fault & VM_FAULT_MAJOR) { |
| 347 | tsk->maj_flt++; | 347 | tsk->maj_flt++; |
| 348 | perf_sw_event(PERF_COUNT_SW_PAGE_FAULTS_MAJ, 1, 0, | 348 | perf_sw_event(PERF_COUNT_SW_PAGE_FAULTS_MAJ, 1, |
| 349 | regs, address); | 349 | regs, address); |
| 350 | } else { | 350 | } else { |
| 351 | tsk->min_flt++; | 351 | tsk->min_flt++; |
| 352 | perf_sw_event(PERF_COUNT_SW_PAGE_FAULTS_MIN, 1, 0, | 352 | perf_sw_event(PERF_COUNT_SW_PAGE_FAULTS_MIN, 1, |
| 353 | regs, address); | 353 | regs, address); |
| 354 | } | 354 | } |
| 355 | if (fault & VM_FAULT_RETRY) { | 355 | if (fault & VM_FAULT_RETRY) { |
diff --git a/arch/sh/kernel/cpu/sh4/perf_event.c b/arch/sh/kernel/cpu/sh4/perf_event.c index 748955df018..fa4f724b295 100644 --- a/arch/sh/kernel/cpu/sh4/perf_event.c +++ b/arch/sh/kernel/cpu/sh4/perf_event.c | |||
| @@ -180,6 +180,21 @@ static const int sh7750_cache_events | |||
| 180 | [ C(RESULT_MISS) ] = -1, | 180 | [ C(RESULT_MISS) ] = -1, |
| 181 | }, | 181 | }, |
| 182 | }, | 182 | }, |
| 183 | |||
| 184 | [ C(NODE) ] = { | ||
| 185 | [ C(OP_READ) ] = { | ||
| 186 | [ C(RESULT_ACCESS) ] = -1, | ||
| 187 | [ C(RESULT_MISS) ] = -1, | ||
| 188 | }, | ||
| 189 | [ C(OP_WRITE) ] = { | ||
| 190 | [ C(RESULT_ACCESS) ] = -1, | ||
| 191 | [ C(RESULT_MISS) ] = -1, | ||
| 192 | }, | ||
| 193 | [ C(OP_PREFETCH) ] = { | ||
| 194 | [ C(RESULT_ACCESS) ] = -1, | ||
| 195 | [ C(RESULT_MISS) ] = -1, | ||
| 196 | }, | ||
| 197 | }, | ||
| 183 | }; | 198 | }; |
| 184 | 199 | ||
| 185 | static int sh7750_event_map(int event) | 200 | static int sh7750_event_map(int event) |
diff --git a/arch/sh/kernel/cpu/sh4a/perf_event.c b/arch/sh/kernel/cpu/sh4a/perf_event.c index 17e6bebfede..84a2c396cee 100644 --- a/arch/sh/kernel/cpu/sh4a/perf_event.c +++ b/arch/sh/kernel/cpu/sh4a/perf_event.c | |||
| @@ -205,6 +205,21 @@ static const int sh4a_cache_events | |||
| 205 | [ C(RESULT_MISS) ] = -1, | 205 | [ C(RESULT_MISS) ] = -1, |
| 206 | }, | 206 | }, |
| 207 | }, | 207 | }, |
| 208 | |||
| 209 | [ C(NODE) ] = { | ||
| 210 | [ C(OP_READ) ] = { | ||
| 211 | [ C(RESULT_ACCESS) ] = -1, | ||
| 212 | [ C(RESULT_MISS) ] = -1, | ||
| 213 | }, | ||
| 214 | [ C(OP_WRITE) ] = { | ||
| 215 | [ C(RESULT_ACCESS) ] = -1, | ||
| 216 | [ C(RESULT_MISS) ] = -1, | ||
| 217 | }, | ||
| 218 | [ C(OP_PREFETCH) ] = { | ||
| 219 | [ C(RESULT_ACCESS) ] = -1, | ||
| 220 | [ C(RESULT_MISS) ] = -1, | ||
| 221 | }, | ||
| 222 | }, | ||
| 208 | }; | 223 | }; |
| 209 | 224 | ||
| 210 | static int sh4a_event_map(int event) | 225 | static int sh4a_event_map(int event) |
diff --git a/arch/sh/kernel/ptrace_32.c b/arch/sh/kernel/ptrace_32.c index 3d7b209b217..92b3c276339 100644 --- a/arch/sh/kernel/ptrace_32.c +++ b/arch/sh/kernel/ptrace_32.c | |||
| @@ -63,7 +63,7 @@ static inline int put_stack_long(struct task_struct *task, int offset, | |||
| 63 | return 0; | 63 | return 0; |
| 64 | } | 64 | } |
| 65 | 65 | ||
| 66 | void ptrace_triggered(struct perf_event *bp, int nmi, | 66 | void ptrace_triggered(struct perf_event *bp, |
| 67 | struct perf_sample_data *data, struct pt_regs *regs) | 67 | struct perf_sample_data *data, struct pt_regs *regs) |
| 68 | { | 68 | { |
| 69 | struct perf_event_attr attr; | 69 | struct perf_event_attr attr; |
| @@ -91,7 +91,8 @@ static int set_single_step(struct task_struct *tsk, unsigned long addr) | |||
| 91 | attr.bp_len = HW_BREAKPOINT_LEN_2; | 91 | attr.bp_len = HW_BREAKPOINT_LEN_2; |
| 92 | attr.bp_type = HW_BREAKPOINT_R; | 92 | attr.bp_type = HW_BREAKPOINT_R; |
| 93 | 93 | ||
| 94 | bp = register_user_hw_breakpoint(&attr, ptrace_triggered, tsk); | 94 | bp = register_user_hw_breakpoint(&attr, ptrace_triggered, |
| 95 | NULL, tsk); | ||
| 95 | if (IS_ERR(bp)) | 96 | if (IS_ERR(bp)) |
| 96 | return PTR_ERR(bp); | 97 | return PTR_ERR(bp); |
| 97 | 98 | ||
diff --git a/arch/sh/kernel/traps_32.c b/arch/sh/kernel/traps_32.c index b51a17104b5..d9006f8ffc1 100644 --- a/arch/sh/kernel/traps_32.c +++ b/arch/sh/kernel/traps_32.c | |||
| @@ -393,7 +393,7 @@ int handle_unaligned_access(insn_size_t instruction, struct pt_regs *regs, | |||
| 393 | */ | 393 | */ |
| 394 | if (!expected) { | 394 | if (!expected) { |
| 395 | unaligned_fixups_notify(current, instruction, regs); | 395 | unaligned_fixups_notify(current, instruction, regs); |
| 396 | perf_sw_event(PERF_COUNT_SW_ALIGNMENT_FAULTS, 1, 0, | 396 | perf_sw_event(PERF_COUNT_SW_ALIGNMENT_FAULTS, 1, |
| 397 | regs, address); | 397 | regs, address); |
| 398 | } | 398 | } |
| 399 | 399 | ||
diff --git a/arch/sh/kernel/traps_64.c b/arch/sh/kernel/traps_64.c index 6713ca97e55..67110be83fd 100644 --- a/arch/sh/kernel/traps_64.c +++ b/arch/sh/kernel/traps_64.c | |||
| @@ -434,7 +434,7 @@ static int misaligned_load(struct pt_regs *regs, | |||
| 434 | return error; | 434 | return error; |
| 435 | } | 435 | } |
| 436 | 436 | ||
| 437 | perf_sw_event(PERF_COUNT_SW_ALIGNMENT_FAULTS, 1, 0, regs, address); | 437 | perf_sw_event(PERF_COUNT_SW_ALIGNMENT_FAULTS, 1, regs, address); |
| 438 | 438 | ||
| 439 | destreg = (opcode >> 4) & 0x3f; | 439 | destreg = (opcode >> 4) & 0x3f; |
| 440 | if (user_mode(regs)) { | 440 | if (user_mode(regs)) { |
| @@ -512,7 +512,7 @@ static int misaligned_store(struct pt_regs *regs, | |||
| 512 | return error; | 512 | return error; |
| 513 | } | 513 | } |
| 514 | 514 | ||
| 515 | perf_sw_event(PERF_COUNT_SW_ALIGNMENT_FAULTS, 1, 0, regs, address); | 515 | perf_sw_event(PERF_COUNT_SW_ALIGNMENT_FAULTS, 1, regs, address); |
| 516 | 516 | ||
| 517 | srcreg = (opcode >> 4) & 0x3f; | 517 | srcreg = (opcode >> 4) & 0x3f; |
| 518 | if (user_mode(regs)) { | 518 | if (user_mode(regs)) { |
| @@ -588,7 +588,7 @@ static int misaligned_fpu_load(struct pt_regs *regs, | |||
| 588 | return error; | 588 | return error; |
| 589 | } | 589 | } |
| 590 | 590 | ||
| 591 | perf_sw_event(PERF_COUNT_SW_EMULATION_FAULTS, 1, 0, regs, address); | 591 | perf_sw_event(PERF_COUNT_SW_EMULATION_FAULTS, 1, regs, address); |
| 592 | 592 | ||
| 593 | destreg = (opcode >> 4) & 0x3f; | 593 | destreg = (opcode >> 4) & 0x3f; |
| 594 | if (user_mode(regs)) { | 594 | if (user_mode(regs)) { |
| @@ -665,7 +665,7 @@ static int misaligned_fpu_store(struct pt_regs *regs, | |||
| 665 | return error; | 665 | return error; |
| 666 | } | 666 | } |
| 667 | 667 | ||
| 668 | perf_sw_event(PERF_COUNT_SW_EMULATION_FAULTS, 1, 0, regs, address); | 668 | perf_sw_event(PERF_COUNT_SW_EMULATION_FAULTS, 1, regs, address); |
| 669 | 669 | ||
| 670 | srcreg = (opcode >> 4) & 0x3f; | 670 | srcreg = (opcode >> 4) & 0x3f; |
| 671 | if (user_mode(regs)) { | 671 | if (user_mode(regs)) { |
diff --git a/arch/sh/math-emu/math.c b/arch/sh/math-emu/math.c index f76a5090d5d..97719521065 100644 --- a/arch/sh/math-emu/math.c +++ b/arch/sh/math-emu/math.c | |||
| @@ -620,7 +620,7 @@ int do_fpu_inst(unsigned short inst, struct pt_regs *regs) | |||
| 620 | struct task_struct *tsk = current; | 620 | struct task_struct *tsk = current; |
| 621 | struct sh_fpu_soft_struct *fpu = &(tsk->thread.xstate->softfpu); | 621 | struct sh_fpu_soft_struct *fpu = &(tsk->thread.xstate->softfpu); |
| 622 | 622 | ||
| 623 | perf_sw_event(PERF_COUNT_SW_EMULATION_FAULTS, 1, 0, regs, 0); | 623 | perf_sw_event(PERF_COUNT_SW_EMULATION_FAULTS, 1, regs, 0); |
| 624 | 624 | ||
| 625 | if (!(task_thread_info(tsk)->status & TS_USEDFPU)) { | 625 | if (!(task_thread_info(tsk)->status & TS_USEDFPU)) { |
| 626 | /* initialize once. */ | 626 | /* initialize once. */ |
diff --git a/arch/sh/mm/fault_32.c b/arch/sh/mm/fault_32.c index d4c34d757f0..7bebd044f2a 100644 --- a/arch/sh/mm/fault_32.c +++ b/arch/sh/mm/fault_32.c | |||
| @@ -160,7 +160,7 @@ asmlinkage void __kprobes do_page_fault(struct pt_regs *regs, | |||
| 160 | if ((regs->sr & SR_IMASK) != SR_IMASK) | 160 | if ((regs->sr & SR_IMASK) != SR_IMASK) |
| 161 | local_irq_enable(); | 161 | local_irq_enable(); |
| 162 | 162 | ||
| 163 | perf_sw_event(PERF_COUNT_SW_PAGE_FAULTS, 1, 0, regs, address); | 163 | perf_sw_event(PERF_COUNT_SW_PAGE_FAULTS, 1, regs, address); |
| 164 | 164 | ||
| 165 | /* | 165 | /* |
| 166 | * If we're in an interrupt, have no user context or are running | 166 | * If we're in an interrupt, have no user context or are running |
| @@ -210,11 +210,11 @@ good_area: | |||
| 210 | } | 210 | } |
| 211 | if (fault & VM_FAULT_MAJOR) { | 211 | if (fault & VM_FAULT_MAJOR) { |
| 212 | tsk->maj_flt++; | 212 | tsk->maj_flt++; |
| 213 | perf_sw_event(PERF_COUNT_SW_PAGE_FAULTS_MAJ, 1, 0, | 213 | perf_sw_event(PERF_COUNT_SW_PAGE_FAULTS_MAJ, 1, |
| 214 | regs, address); | 214 | regs, address); |
| 215 | } else { | 215 | } else { |
| 216 | tsk->min_flt++; | 216 | tsk->min_flt++; |
| 217 | perf_sw_event(PERF_COUNT_SW_PAGE_FAULTS_MIN, 1, 0, | 217 | perf_sw_event(PERF_COUNT_SW_PAGE_FAULTS_MIN, 1, |
| 218 | regs, address); | 218 | regs, address); |
| 219 | } | 219 | } |
| 220 | 220 | ||
diff --git a/arch/sh/mm/tlbflush_64.c b/arch/sh/mm/tlbflush_64.c index 7f5810f5dfd..e3430e093d4 100644 --- a/arch/sh/mm/tlbflush_64.c +++ b/arch/sh/mm/tlbflush_64.c | |||
| @@ -116,7 +116,7 @@ asmlinkage void do_page_fault(struct pt_regs *regs, unsigned long writeaccess, | |||
| 116 | /* Not an IO address, so reenable interrupts */ | 116 | /* Not an IO address, so reenable interrupts */ |
| 117 | local_irq_enable(); | 117 | local_irq_enable(); |
| 118 | 118 | ||
| 119 | perf_sw_event(PERF_COUNT_SW_PAGE_FAULTS, 1, 0, regs, address); | 119 | perf_sw_event(PERF_COUNT_SW_PAGE_FAULTS, 1, regs, address); |
| 120 | 120 | ||
| 121 | /* | 121 | /* |
| 122 | * If we're in an interrupt or have no user | 122 | * If we're in an interrupt or have no user |
| @@ -200,11 +200,11 @@ good_area: | |||
| 200 | 200 | ||
| 201 | if (fault & VM_FAULT_MAJOR) { | 201 | if (fault & VM_FAULT_MAJOR) { |
| 202 | tsk->maj_flt++; | 202 | tsk->maj_flt++; |
| 203 | perf_sw_event(PERF_COUNT_SW_PAGE_FAULTS_MAJ, 1, 0, | 203 | perf_sw_event(PERF_COUNT_SW_PAGE_FAULTS_MAJ, 1, |
| 204 | regs, address); | 204 | regs, address); |
| 205 | } else { | 205 | } else { |
| 206 | tsk->min_flt++; | 206 | tsk->min_flt++; |
| 207 | perf_sw_event(PERF_COUNT_SW_PAGE_FAULTS_MIN, 1, 0, | 207 | perf_sw_event(PERF_COUNT_SW_PAGE_FAULTS_MIN, 1, |
| 208 | regs, address); | 208 | regs, address); |
| 209 | } | 209 | } |
| 210 | 210 | ||
diff --git a/arch/sparc/kernel/perf_event.c b/arch/sparc/kernel/perf_event.c index 2cb0e1c001e..62a034318b1 100644 --- a/arch/sparc/kernel/perf_event.c +++ b/arch/sparc/kernel/perf_event.c | |||
| @@ -246,6 +246,20 @@ static const cache_map_t ultra3_cache_map = { | |||
| 246 | [ C(RESULT_MISS) ] = { CACHE_OP_UNSUPPORTED }, | 246 | [ C(RESULT_MISS) ] = { CACHE_OP_UNSUPPORTED }, |
| 247 | }, | 247 | }, |
| 248 | }, | 248 | }, |
| 249 | [C(NODE)] = { | ||
| 250 | [C(OP_READ)] = { | ||
| 251 | [C(RESULT_ACCESS)] = { CACHE_OP_UNSUPPORTED }, | ||
| 252 | [C(RESULT_MISS) ] = { CACHE_OP_UNSUPPORTED }, | ||
| 253 | }, | ||
| 254 | [ C(OP_WRITE) ] = { | ||
| 255 | [ C(RESULT_ACCESS) ] = { CACHE_OP_UNSUPPORTED }, | ||
| 256 | [ C(RESULT_MISS) ] = { CACHE_OP_UNSUPPORTED }, | ||
| 257 | }, | ||
| 258 | [ C(OP_PREFETCH) ] = { | ||
| 259 | [ C(RESULT_ACCESS) ] = { CACHE_OP_UNSUPPORTED }, | ||
| 260 | [ C(RESULT_MISS) ] = { CACHE_OP_UNSUPPORTED }, | ||
| 261 | }, | ||
| 262 | }, | ||
| 249 | }; | 263 | }; |
| 250 | 264 | ||
| 251 | static const struct sparc_pmu ultra3_pmu = { | 265 | static const struct sparc_pmu ultra3_pmu = { |
| @@ -361,6 +375,20 @@ static const cache_map_t niagara1_cache_map = { | |||
| 361 | [ C(RESULT_MISS) ] = { CACHE_OP_UNSUPPORTED }, | 375 | [ C(RESULT_MISS) ] = { CACHE_OP_UNSUPPORTED }, |
| 362 | }, | 376 | }, |
| 363 | }, | 377 | }, |
| 378 | [C(NODE)] = { | ||
| 379 | [C(OP_READ)] = { | ||
| 380 | [C(RESULT_ACCESS)] = { CACHE_OP_UNSUPPORTED }, | ||
| 381 | [C(RESULT_MISS) ] = { CACHE_OP_UNSUPPORTED }, | ||
| 382 | }, | ||
| 383 | [ C(OP_WRITE) ] = { | ||
| 384 | [ C(RESULT_ACCESS) ] = { CACHE_OP_UNSUPPORTED }, | ||
| 385 | [ C(RESULT_MISS) ] = { CACHE_OP_UNSUPPORTED }, | ||
| 386 | }, | ||
| 387 | [ C(OP_PREFETCH) ] = { | ||
| 388 | [ C(RESULT_ACCESS) ] = { CACHE_OP_UNSUPPORTED }, | ||
| 389 | [ C(RESULT_MISS) ] = { CACHE_OP_UNSUPPORTED }, | ||
| 390 | }, | ||
| 391 | }, | ||
| 364 | }; | 392 | }; |
| 365 | 393 | ||
| 366 | static const struct sparc_pmu niagara1_pmu = { | 394 | static const struct sparc_pmu niagara1_pmu = { |
| @@ -473,6 +501,20 @@ static const cache_map_t niagara2_cache_map = { | |||
| 473 | [ C(RESULT_MISS) ] = { CACHE_OP_UNSUPPORTED }, | 501 | [ C(RESULT_MISS) ] = { CACHE_OP_UNSUPPORTED }, |
| 474 | }, | 502 | }, |
| 475 | }, | 503 | }, |
| 504 | [C(NODE)] = { | ||
| 505 | [C(OP_READ)] = { | ||
| 506 | [C(RESULT_ACCESS)] = { CACHE_OP_UNSUPPORTED }, | ||
| 507 | [C(RESULT_MISS) ] = { CACHE_OP_UNSUPPORTED }, | ||
| 508 | }, | ||
| 509 | [ C(OP_WRITE) ] = { | ||
| 510 | [ C(RESULT_ACCESS) ] = { CACHE_OP_UNSUPPORTED }, | ||
| 511 | [ C(RESULT_MISS) ] = { CACHE_OP_UNSUPPORTED }, | ||
| 512 | }, | ||
| 513 | [ C(OP_PREFETCH) ] = { | ||
| 514 | [ C(RESULT_ACCESS) ] = { CACHE_OP_UNSUPPORTED }, | ||
| 515 | [ C(RESULT_MISS) ] = { CACHE_OP_UNSUPPORTED }, | ||
| 516 | }, | ||
| 517 | }, | ||
| 476 | }; | 518 | }; |
| 477 | 519 | ||
| 478 | static const struct sparc_pmu niagara2_pmu = { | 520 | static const struct sparc_pmu niagara2_pmu = { |
| @@ -1277,7 +1319,7 @@ static int __kprobes perf_event_nmi_handler(struct notifier_block *self, | |||
| 1277 | if (!sparc_perf_event_set_period(event, hwc, idx)) | 1319 | if (!sparc_perf_event_set_period(event, hwc, idx)) |
| 1278 | continue; | 1320 | continue; |
| 1279 | 1321 | ||
| 1280 | if (perf_event_overflow(event, 1, &data, regs)) | 1322 | if (perf_event_overflow(event, &data, regs)) |
| 1281 | sparc_pmu_stop(event, 0); | 1323 | sparc_pmu_stop(event, 0); |
| 1282 | } | 1324 | } |
| 1283 | 1325 | ||
diff --git a/arch/sparc/kernel/unaligned_32.c b/arch/sparc/kernel/unaligned_32.c index 4491f4cb269..7efbb2f9e77 100644 --- a/arch/sparc/kernel/unaligned_32.c +++ b/arch/sparc/kernel/unaligned_32.c | |||
| @@ -247,7 +247,7 @@ asmlinkage void kernel_unaligned_trap(struct pt_regs *regs, unsigned int insn) | |||
| 247 | unsigned long addr = compute_effective_address(regs, insn); | 247 | unsigned long addr = compute_effective_address(regs, insn); |
| 248 | int err; | 248 | int err; |
| 249 | 249 | ||
| 250 | perf_sw_event(PERF_COUNT_SW_ALIGNMENT_FAULTS, 1, 0, regs, addr); | 250 | perf_sw_event(PERF_COUNT_SW_ALIGNMENT_FAULTS, 1, regs, addr); |
| 251 | switch (dir) { | 251 | switch (dir) { |
| 252 | case load: | 252 | case load: |
| 253 | err = do_int_load(fetch_reg_addr(((insn>>25)&0x1f), | 253 | err = do_int_load(fetch_reg_addr(((insn>>25)&0x1f), |
| @@ -338,7 +338,7 @@ asmlinkage void user_unaligned_trap(struct pt_regs *regs, unsigned int insn) | |||
| 338 | } | 338 | } |
| 339 | 339 | ||
| 340 | addr = compute_effective_address(regs, insn); | 340 | addr = compute_effective_address(regs, insn); |
| 341 | perf_sw_event(PERF_COUNT_SW_ALIGNMENT_FAULTS, 1, 0, regs, addr); | 341 | perf_sw_event(PERF_COUNT_SW_ALIGNMENT_FAULTS, 1, regs, addr); |
| 342 | switch(dir) { | 342 | switch(dir) { |
| 343 | case load: | 343 | case load: |
| 344 | err = do_int_load(fetch_reg_addr(((insn>>25)&0x1f), | 344 | err = do_int_load(fetch_reg_addr(((insn>>25)&0x1f), |
diff --git a/arch/sparc/kernel/unaligned_64.c b/arch/sparc/kernel/unaligned_64.c index b2b019ea8ca..35cff1673aa 100644 --- a/arch/sparc/kernel/unaligned_64.c +++ b/arch/sparc/kernel/unaligned_64.c | |||
| @@ -317,7 +317,7 @@ asmlinkage void kernel_unaligned_trap(struct pt_regs *regs, unsigned int insn) | |||
| 317 | 317 | ||
| 318 | addr = compute_effective_address(regs, insn, | 318 | addr = compute_effective_address(regs, insn, |
| 319 | ((insn >> 25) & 0x1f)); | 319 | ((insn >> 25) & 0x1f)); |
| 320 | perf_sw_event(PERF_COUNT_SW_ALIGNMENT_FAULTS, 1, 0, regs, addr); | 320 | perf_sw_event(PERF_COUNT_SW_ALIGNMENT_FAULTS, 1, regs, addr); |
| 321 | switch (asi) { | 321 | switch (asi) { |
| 322 | case ASI_NL: | 322 | case ASI_NL: |
| 323 | case ASI_AIUPL: | 323 | case ASI_AIUPL: |
| @@ -384,7 +384,7 @@ int handle_popc(u32 insn, struct pt_regs *regs) | |||
| 384 | int ret, i, rd = ((insn >> 25) & 0x1f); | 384 | int ret, i, rd = ((insn >> 25) & 0x1f); |
| 385 | int from_kernel = (regs->tstate & TSTATE_PRIV) != 0; | 385 | int from_kernel = (regs->tstate & TSTATE_PRIV) != 0; |
| 386 | 386 | ||
| 387 | perf_sw_event(PERF_COUNT_SW_EMULATION_FAULTS, 1, 0, regs, 0); | 387 | perf_sw_event(PERF_COUNT_SW_EMULATION_FAULTS, 1, regs, 0); |
| 388 | if (insn & 0x2000) { | 388 | if (insn & 0x2000) { |
| 389 | maybe_flush_windows(0, 0, rd, from_kernel); | 389 | maybe_flush_windows(0, 0, rd, from_kernel); |
| 390 | value = sign_extend_imm13(insn); | 390 | value = sign_extend_imm13(insn); |
| @@ -431,7 +431,7 @@ int handle_ldf_stq(u32 insn, struct pt_regs *regs) | |||
| 431 | int asi = decode_asi(insn, regs); | 431 | int asi = decode_asi(insn, regs); |
| 432 | int flag = (freg < 32) ? FPRS_DL : FPRS_DU; | 432 | int flag = (freg < 32) ? FPRS_DL : FPRS_DU; |
| 433 | 433 | ||
| 434 | perf_sw_event(PERF_COUNT_SW_EMULATION_FAULTS, 1, 0, regs, 0); | 434 | perf_sw_event(PERF_COUNT_SW_EMULATION_FAULTS, 1, regs, 0); |
| 435 | 435 | ||
| 436 | save_and_clear_fpu(); | 436 | save_and_clear_fpu(); |
| 437 | current_thread_info()->xfsr[0] &= ~0x1c000; | 437 | current_thread_info()->xfsr[0] &= ~0x1c000; |
| @@ -554,7 +554,7 @@ void handle_ld_nf(u32 insn, struct pt_regs *regs) | |||
| 554 | int from_kernel = (regs->tstate & TSTATE_PRIV) != 0; | 554 | int from_kernel = (regs->tstate & TSTATE_PRIV) != 0; |
| 555 | unsigned long *reg; | 555 | unsigned long *reg; |
| 556 | 556 | ||
| 557 | perf_sw_event(PERF_COUNT_SW_EMULATION_FAULTS, 1, 0, regs, 0); | 557 | perf_sw_event(PERF_COUNT_SW_EMULATION_FAULTS, 1, regs, 0); |
| 558 | 558 | ||
| 559 | maybe_flush_windows(0, 0, rd, from_kernel); | 559 | maybe_flush_windows(0, 0, rd, from_kernel); |
| 560 | reg = fetch_reg_addr(rd, regs); | 560 | reg = fetch_reg_addr(rd, regs); |
| @@ -586,7 +586,7 @@ void handle_lddfmna(struct pt_regs *regs, unsigned long sfar, unsigned long sfsr | |||
| 586 | 586 | ||
| 587 | if (tstate & TSTATE_PRIV) | 587 | if (tstate & TSTATE_PRIV) |
| 588 | die_if_kernel("lddfmna from kernel", regs); | 588 | die_if_kernel("lddfmna from kernel", regs); |
| 589 | perf_sw_event(PERF_COUNT_SW_ALIGNMENT_FAULTS, 1, 0, regs, sfar); | 589 | perf_sw_event(PERF_COUNT_SW_ALIGNMENT_FAULTS, 1, regs, sfar); |
| 590 | if (test_thread_flag(TIF_32BIT)) | 590 | if (test_thread_flag(TIF_32BIT)) |
| 591 | pc = (u32)pc; | 591 | pc = (u32)pc; |
| 592 | if (get_user(insn, (u32 __user *) pc) != -EFAULT) { | 592 | if (get_user(insn, (u32 __user *) pc) != -EFAULT) { |
| @@ -647,7 +647,7 @@ void handle_stdfmna(struct pt_regs *regs, unsigned long sfar, unsigned long sfsr | |||
| 647 | 647 | ||
| 648 | if (tstate & TSTATE_PRIV) | 648 | if (tstate & TSTATE_PRIV) |
| 649 | die_if_kernel("stdfmna from kernel", regs); | 649 | die_if_kernel("stdfmna from kernel", regs); |
| 650 | perf_sw_event(PERF_COUNT_SW_ALIGNMENT_FAULTS, 1, 0, regs, sfar); | 650 | perf_sw_event(PERF_COUNT_SW_ALIGNMENT_FAULTS, 1, regs, sfar); |
| 651 | if (test_thread_flag(TIF_32BIT)) | 651 | if (test_thread_flag(TIF_32BIT)) |
| 652 | pc = (u32)pc; | 652 | pc = (u32)pc; |
| 653 | if (get_user(insn, (u32 __user *) pc) != -EFAULT) { | 653 | if (get_user(insn, (u32 __user *) pc) != -EFAULT) { |
diff --git a/arch/sparc/kernel/visemul.c b/arch/sparc/kernel/visemul.c index 36357717d69..32b626c9d81 100644 --- a/arch/sparc/kernel/visemul.c +++ b/arch/sparc/kernel/visemul.c | |||
| @@ -802,7 +802,7 @@ int vis_emul(struct pt_regs *regs, unsigned int insn) | |||
| 802 | 802 | ||
| 803 | BUG_ON(regs->tstate & TSTATE_PRIV); | 803 | BUG_ON(regs->tstate & TSTATE_PRIV); |
| 804 | 804 | ||
| 805 | perf_sw_event(PERF_COUNT_SW_EMULATION_FAULTS, 1, 0, regs, 0); | 805 | perf_sw_event(PERF_COUNT_SW_EMULATION_FAULTS, 1, regs, 0); |
| 806 | 806 | ||
| 807 | if (test_thread_flag(TIF_32BIT)) | 807 | if (test_thread_flag(TIF_32BIT)) |
| 808 | pc = (u32)pc; | 808 | pc = (u32)pc; |
diff --git a/arch/sparc/math-emu/math_32.c b/arch/sparc/math-emu/math_32.c index a3fccde894e..aa4d55b0bdf 100644 --- a/arch/sparc/math-emu/math_32.c +++ b/arch/sparc/math-emu/math_32.c | |||
| @@ -164,7 +164,7 @@ int do_mathemu(struct pt_regs *regs, struct task_struct *fpt) | |||
| 164 | int retcode = 0; /* assume all succeed */ | 164 | int retcode = 0; /* assume all succeed */ |
| 165 | unsigned long insn; | 165 | unsigned long insn; |
| 166 | 166 | ||
| 167 | perf_sw_event(PERF_COUNT_SW_EMULATION_FAULTS, 1, 0, regs, 0); | 167 | perf_sw_event(PERF_COUNT_SW_EMULATION_FAULTS, 1, regs, 0); |
| 168 | 168 | ||
| 169 | #ifdef DEBUG_MATHEMU | 169 | #ifdef DEBUG_MATHEMU |
| 170 | printk("In do_mathemu()... pc is %08lx\n", regs->pc); | 170 | printk("In do_mathemu()... pc is %08lx\n", regs->pc); |
diff --git a/arch/sparc/math-emu/math_64.c b/arch/sparc/math-emu/math_64.c index 56d2c44747b..e575bd2fe38 100644 --- a/arch/sparc/math-emu/math_64.c +++ b/arch/sparc/math-emu/math_64.c | |||
| @@ -184,7 +184,7 @@ int do_mathemu(struct pt_regs *regs, struct fpustate *f) | |||
| 184 | 184 | ||
| 185 | if (tstate & TSTATE_PRIV) | 185 | if (tstate & TSTATE_PRIV) |
| 186 | die_if_kernel("unfinished/unimplemented FPop from kernel", regs); | 186 | die_if_kernel("unfinished/unimplemented FPop from kernel", regs); |
| 187 | perf_sw_event(PERF_COUNT_SW_EMULATION_FAULTS, 1, 0, regs, 0); | 187 | perf_sw_event(PERF_COUNT_SW_EMULATION_FAULTS, 1, regs, 0); |
| 188 | if (test_thread_flag(TIF_32BIT)) | 188 | if (test_thread_flag(TIF_32BIT)) |
| 189 | pc = (u32)pc; | 189 | pc = (u32)pc; |
| 190 | if (get_user(insn, (u32 __user *) pc) != -EFAULT) { | 190 | if (get_user(insn, (u32 __user *) pc) != -EFAULT) { |
diff --git a/arch/sparc/mm/fault_32.c b/arch/sparc/mm/fault_32.c index 7543ddbdadb..aa1c1b1ce5c 100644 --- a/arch/sparc/mm/fault_32.c +++ b/arch/sparc/mm/fault_32.c | |||
| @@ -251,7 +251,7 @@ asmlinkage void do_sparc_fault(struct pt_regs *regs, int text_fault, int write, | |||
| 251 | if (in_atomic() || !mm) | 251 | if (in_atomic() || !mm) |
| 252 | goto no_context; | 252 | goto no_context; |
| 253 | 253 | ||
| 254 | perf_sw_event(PERF_COUNT_SW_PAGE_FAULTS, 1, 0, regs, address); | 254 | perf_sw_event(PERF_COUNT_SW_PAGE_FAULTS, 1, regs, address); |
| 255 | 255 | ||
| 256 | down_read(&mm->mmap_sem); | 256 | down_read(&mm->mmap_sem); |
| 257 | 257 | ||
| @@ -301,12 +301,10 @@ good_area: | |||
| 301 | } | 301 | } |
| 302 | if (fault & VM_FAULT_MAJOR) { | 302 | if (fault & VM_FAULT_MAJOR) { |
| 303 | current->maj_flt++; | 303 | current->maj_flt++; |
| 304 | perf_sw_event(PERF_COUNT_SW_PAGE_FAULTS_MAJ, 1, 0, | 304 | perf_sw_event(PERF_COUNT_SW_PAGE_FAULTS_MAJ, 1, regs, address); |
| 305 | regs, address); | ||
| 306 | } else { | 305 | } else { |
| 307 | current->min_flt++; | 306 | current->min_flt++; |
| 308 | perf_sw_event(PERF_COUNT_SW_PAGE_FAULTS_MIN, 1, 0, | 307 | perf_sw_event(PERF_COUNT_SW_PAGE_FAULTS_MIN, 1, regs, address); |
| 309 | regs, address); | ||
| 310 | } | 308 | } |
| 311 | up_read(&mm->mmap_sem); | 309 | up_read(&mm->mmap_sem); |
| 312 | return; | 310 | return; |
diff --git a/arch/sparc/mm/fault_64.c b/arch/sparc/mm/fault_64.c index f92ce56a8b2..504c0622f72 100644 --- a/arch/sparc/mm/fault_64.c +++ b/arch/sparc/mm/fault_64.c | |||
| @@ -325,7 +325,7 @@ asmlinkage void __kprobes do_sparc64_fault(struct pt_regs *regs) | |||
| 325 | if (in_atomic() || !mm) | 325 | if (in_atomic() || !mm) |
| 326 | goto intr_or_no_mm; | 326 | goto intr_or_no_mm; |
| 327 | 327 | ||
| 328 | perf_sw_event(PERF_COUNT_SW_PAGE_FAULTS, 1, 0, regs, address); | 328 | perf_sw_event(PERF_COUNT_SW_PAGE_FAULTS, 1, regs, address); |
| 329 | 329 | ||
| 330 | if (!down_read_trylock(&mm->mmap_sem)) { | 330 | if (!down_read_trylock(&mm->mmap_sem)) { |
| 331 | if ((regs->tstate & TSTATE_PRIV) && | 331 | if ((regs->tstate & TSTATE_PRIV) && |
| @@ -433,12 +433,10 @@ good_area: | |||
| 433 | } | 433 | } |
| 434 | if (fault & VM_FAULT_MAJOR) { | 434 | if (fault & VM_FAULT_MAJOR) { |
| 435 | current->maj_flt++; | 435 | current->maj_flt++; |
| 436 | perf_sw_event(PERF_COUNT_SW_PAGE_FAULTS_MAJ, 1, 0, | 436 | perf_sw_event(PERF_COUNT_SW_PAGE_FAULTS_MAJ, 1, regs, address); |
| 437 | regs, address); | ||
| 438 | } else { | 437 | } else { |
| 439 | current->min_flt++; | 438 | current->min_flt++; |
| 440 | perf_sw_event(PERF_COUNT_SW_PAGE_FAULTS_MIN, 1, 0, | 439 | perf_sw_event(PERF_COUNT_SW_PAGE_FAULTS_MIN, 1, regs, address); |
| 441 | regs, address); | ||
| 442 | } | 440 | } |
| 443 | up_read(&mm->mmap_sem); | 441 | up_read(&mm->mmap_sem); |
| 444 | 442 | ||
diff --git a/arch/x86/include/asm/irqflags.h b/arch/x86/include/asm/irqflags.h index 5745ce8bf10..bba3cf88e62 100644 --- a/arch/x86/include/asm/irqflags.h +++ b/arch/x86/include/asm/irqflags.h | |||
| @@ -60,23 +60,24 @@ static inline void native_halt(void) | |||
| 60 | #include <asm/paravirt.h> | 60 | #include <asm/paravirt.h> |
| 61 | #else | 61 | #else |
| 62 | #ifndef __ASSEMBLY__ | 62 | #ifndef __ASSEMBLY__ |
| 63 | #include <linux/types.h> | ||
| 63 | 64 | ||
| 64 | static inline unsigned long arch_local_save_flags(void) | 65 | static inline notrace unsigned long arch_local_save_flags(void) |
| 65 | { | 66 | { |
| 66 | return native_save_fl(); | 67 | return native_save_fl(); |
| 67 | } | 68 | } |
| 68 | 69 | ||
| 69 | static inline void arch_local_irq_restore(unsigned long flags) | 70 | static inline notrace void arch_local_irq_restore(unsigned long flags) |
| 70 | { | 71 | { |
| 71 | native_restore_fl(flags); | 72 | native_restore_fl(flags); |
| 72 | } | 73 | } |
| 73 | 74 | ||
| 74 | static inline void arch_local_irq_disable(void) | 75 | static inline notrace void arch_local_irq_disable(void) |
| 75 | { | 76 | { |
| 76 | native_irq_disable(); | 77 | native_irq_disable(); |
| 77 | } | 78 | } |
| 78 | 79 | ||
| 79 | static inline void arch_local_irq_enable(void) | 80 | static inline notrace void arch_local_irq_enable(void) |
| 80 | { | 81 | { |
| 81 | native_irq_enable(); | 82 | native_irq_enable(); |
| 82 | } | 83 | } |
| @@ -102,7 +103,7 @@ static inline void halt(void) | |||
| 102 | /* | 103 | /* |
| 103 | * For spinlocks, etc: | 104 | * For spinlocks, etc: |
| 104 | */ | 105 | */ |
| 105 | static inline unsigned long arch_local_irq_save(void) | 106 | static inline notrace unsigned long arch_local_irq_save(void) |
| 106 | { | 107 | { |
| 107 | unsigned long flags = arch_local_save_flags(); | 108 | unsigned long flags = arch_local_save_flags(); |
| 108 | arch_local_irq_disable(); | 109 | arch_local_irq_disable(); |
diff --git a/arch/x86/include/asm/perf_event.h b/arch/x86/include/asm/perf_event.h index d9d4dae305f..094fb30817a 100644 --- a/arch/x86/include/asm/perf_event.h +++ b/arch/x86/include/asm/perf_event.h | |||
| @@ -152,6 +152,11 @@ extern unsigned long perf_misc_flags(struct pt_regs *regs); | |||
| 152 | (regs)->bp = caller_frame_pointer(); \ | 152 | (regs)->bp = caller_frame_pointer(); \ |
| 153 | (regs)->cs = __KERNEL_CS; \ | 153 | (regs)->cs = __KERNEL_CS; \ |
| 154 | regs->flags = 0; \ | 154 | regs->flags = 0; \ |
| 155 | asm volatile( \ | ||
| 156 | _ASM_MOV "%%"_ASM_SP ", %0\n" \ | ||
| 157 | : "=m" ((regs)->sp) \ | ||
| 158 | :: "memory" \ | ||
| 159 | ); \ | ||
| 155 | } | 160 | } |
| 156 | 161 | ||
| 157 | #else | 162 | #else |
diff --git a/arch/x86/include/asm/perf_event_p4.h b/arch/x86/include/asm/perf_event_p4.h index 56fd9e3abbd..4f7e67e2345 100644 --- a/arch/x86/include/asm/perf_event_p4.h +++ b/arch/x86/include/asm/perf_event_p4.h | |||
| @@ -102,6 +102,14 @@ | |||
| 102 | #define P4_CONFIG_HT (1ULL << P4_CONFIG_HT_SHIFT) | 102 | #define P4_CONFIG_HT (1ULL << P4_CONFIG_HT_SHIFT) |
| 103 | 103 | ||
| 104 | /* | 104 | /* |
| 105 | * If an event has alias it should be marked | ||
| 106 | * with a special bit. (Don't forget to check | ||
| 107 | * P4_PEBS_CONFIG_MASK and related bits on | ||
| 108 | * modification.) | ||
| 109 | */ | ||
| 110 | #define P4_CONFIG_ALIASABLE (1 << 9) | ||
| 111 | |||
| 112 | /* | ||
| 105 | * The bits we allow to pass for RAW events | 113 | * The bits we allow to pass for RAW events |
| 106 | */ | 114 | */ |
| 107 | #define P4_CONFIG_MASK_ESCR \ | 115 | #define P4_CONFIG_MASK_ESCR \ |
| @@ -123,6 +131,31 @@ | |||
| 123 | (p4_config_pack_escr(P4_CONFIG_MASK_ESCR)) | \ | 131 | (p4_config_pack_escr(P4_CONFIG_MASK_ESCR)) | \ |
| 124 | (p4_config_pack_cccr(P4_CONFIG_MASK_CCCR)) | 132 | (p4_config_pack_cccr(P4_CONFIG_MASK_CCCR)) |
| 125 | 133 | ||
| 134 | /* | ||
| 135 | * In case of event aliasing we need to preserve some | ||
| 136 | * caller bits, otherwise the mapping won't be complete. | ||
| 137 | */ | ||
| 138 | #define P4_CONFIG_EVENT_ALIAS_MASK \ | ||
| 139 | (p4_config_pack_escr(P4_CONFIG_MASK_ESCR) | \ | ||
| 140 | p4_config_pack_cccr(P4_CCCR_EDGE | \ | ||
| 141 | P4_CCCR_THRESHOLD_MASK | \ | ||
| 142 | P4_CCCR_COMPLEMENT | \ | ||
| 143 | P4_CCCR_COMPARE)) | ||
| 144 | |||
| 145 | #define P4_CONFIG_EVENT_ALIAS_IMMUTABLE_BITS \ | ||
| 146 | ((P4_CONFIG_HT) | \ | ||
| 147 | p4_config_pack_escr(P4_ESCR_T0_OS | \ | ||
| 148 | P4_ESCR_T0_USR | \ | ||
| 149 | P4_ESCR_T1_OS | \ | ||
| 150 | P4_ESCR_T1_USR) | \ | ||
| 151 | p4_config_pack_cccr(P4_CCCR_OVF | \ | ||
| 152 | P4_CCCR_CASCADE | \ | ||
| 153 | P4_CCCR_FORCE_OVF | \ | ||
| 154 | P4_CCCR_THREAD_ANY | \ | ||
| 155 | P4_CCCR_OVF_PMI_T0 | \ | ||
| 156 | P4_CCCR_OVF_PMI_T1 | \ | ||
| 157 | P4_CONFIG_ALIASABLE)) | ||
| 158 | |||
| 126 | static inline bool p4_is_event_cascaded(u64 config) | 159 | static inline bool p4_is_event_cascaded(u64 config) |
| 127 | { | 160 | { |
| 128 | u32 cccr = p4_config_unpack_cccr(config); | 161 | u32 cccr = p4_config_unpack_cccr(config); |
diff --git a/arch/x86/include/asm/uaccess.h b/arch/x86/include/asm/uaccess.h index 99ddd148a76..36361bf6fdd 100644 --- a/arch/x86/include/asm/uaccess.h +++ b/arch/x86/include/asm/uaccess.h | |||
| @@ -555,6 +555,9 @@ struct __large_struct { unsigned long buf[100]; }; | |||
| 555 | 555 | ||
| 556 | #endif /* CONFIG_X86_WP_WORKS_OK */ | 556 | #endif /* CONFIG_X86_WP_WORKS_OK */ |
| 557 | 557 | ||
| 558 | extern unsigned long | ||
| 559 | copy_from_user_nmi(void *to, const void __user *from, unsigned long n); | ||
| 560 | |||
| 558 | /* | 561 | /* |
| 559 | * movsl can be slow when source and dest are not both 8-byte aligned | 562 | * movsl can be slow when source and dest are not both 8-byte aligned |
| 560 | */ | 563 | */ |
diff --git a/arch/x86/kernel/cpu/perf_event.c b/arch/x86/kernel/cpu/perf_event.c index 3a0338b4b17..4ee3abf20ed 100644 --- a/arch/x86/kernel/cpu/perf_event.c +++ b/arch/x86/kernel/cpu/perf_event.c | |||
| @@ -22,7 +22,6 @@ | |||
| 22 | #include <linux/sched.h> | 22 | #include <linux/sched.h> |
| 23 | #include <linux/uaccess.h> | 23 | #include <linux/uaccess.h> |
| 24 | #include <linux/slab.h> | 24 | #include <linux/slab.h> |
| 25 | #include <linux/highmem.h> | ||
| 26 | #include <linux/cpu.h> | 25 | #include <linux/cpu.h> |
| 27 | #include <linux/bitops.h> | 26 | #include <linux/bitops.h> |
| 28 | 27 | ||
| @@ -45,38 +44,27 @@ do { \ | |||
| 45 | #endif | 44 | #endif |
| 46 | 45 | ||
| 47 | /* | 46 | /* |
| 48 | * best effort, GUP based copy_from_user() that assumes IRQ or NMI context | 47 | * | NHM/WSM | SNB | |
| 48 | * register ------------------------------- | ||
| 49 | * | HT | no HT | HT | no HT | | ||
| 50 | *----------------------------------------- | ||
| 51 | * offcore | core | core | cpu | core | | ||
| 52 | * lbr_sel | core | core | cpu | core | | ||
| 53 | * ld_lat | cpu | core | cpu | core | | ||
| 54 | *----------------------------------------- | ||
| 55 | * | ||
| 56 | * Given that there is a small number of shared regs, | ||
| 57 | * we can pre-allocate their slot in the per-cpu | ||
| 58 | * per-core reg tables. | ||
| 49 | */ | 59 | */ |
| 50 | static unsigned long | 60 | enum extra_reg_type { |
| 51 | copy_from_user_nmi(void *to, const void __user *from, unsigned long n) | 61 | EXTRA_REG_NONE = -1, /* not used */ |
| 52 | { | ||
| 53 | unsigned long offset, addr = (unsigned long)from; | ||
| 54 | unsigned long size, len = 0; | ||
| 55 | struct page *page; | ||
| 56 | void *map; | ||
| 57 | int ret; | ||
| 58 | |||
| 59 | do { | ||
| 60 | ret = __get_user_pages_fast(addr, 1, 0, &page); | ||
| 61 | if (!ret) | ||
| 62 | break; | ||
| 63 | |||
| 64 | offset = addr & (PAGE_SIZE - 1); | ||
| 65 | size = min(PAGE_SIZE - offset, n - len); | ||
| 66 | |||
| 67 | map = kmap_atomic(page); | ||
| 68 | memcpy(to, map+offset, size); | ||
| 69 | kunmap_atomic(map); | ||
| 70 | put_page(page); | ||
| 71 | 62 | ||
| 72 | len += size; | 63 | EXTRA_REG_RSP_0 = 0, /* offcore_response_0 */ |
| 73 | to += size; | 64 | EXTRA_REG_RSP_1 = 1, /* offcore_response_1 */ |
| 74 | addr += size; | ||
| 75 | 65 | ||
| 76 | } while (len < n); | 66 | EXTRA_REG_MAX /* number of entries needed */ |
| 77 | 67 | }; | |
| 78 | return len; | ||
| 79 | } | ||
| 80 | 68 | ||
| 81 | struct event_constraint { | 69 | struct event_constraint { |
| 82 | union { | 70 | union { |
| @@ -132,11 +120,10 @@ struct cpu_hw_events { | |||
| 132 | struct perf_branch_entry lbr_entries[MAX_LBR_ENTRIES]; | 120 | struct perf_branch_entry lbr_entries[MAX_LBR_ENTRIES]; |
| 133 | 121 | ||
| 134 | /* | 122 | /* |
| 135 | * Intel percore register state. | 123 | * manage shared (per-core, per-cpu) registers |
| 136 | * Coordinate shared resources between HT threads. | 124 | * used on Intel NHM/WSM/SNB |
| 137 | */ | 125 | */ |
| 138 | int percore_used; /* Used by this CPU? */ | 126 | struct intel_shared_regs *shared_regs; |
| 139 | struct intel_percore *per_core; | ||
| 140 | 127 | ||
| 141 | /* | 128 | /* |
| 142 | * AMD specific bits | 129 | * AMD specific bits |
| @@ -187,26 +174,45 @@ struct cpu_hw_events { | |||
| 187 | for ((e) = (c); (e)->weight; (e)++) | 174 | for ((e) = (c); (e)->weight; (e)++) |
| 188 | 175 | ||
| 189 | /* | 176 | /* |
| 177 | * Per register state. | ||
| 178 | */ | ||
| 179 | struct er_account { | ||
| 180 | raw_spinlock_t lock; /* per-core: protect structure */ | ||
| 181 | u64 config; /* extra MSR config */ | ||
| 182 | u64 reg; /* extra MSR number */ | ||
| 183 | atomic_t ref; /* reference count */ | ||
| 184 | }; | ||
| 185 | |||
| 186 | /* | ||
| 190 | * Extra registers for specific events. | 187 | * Extra registers for specific events. |
| 188 | * | ||
| 191 | * Some events need large masks and require external MSRs. | 189 | * Some events need large masks and require external MSRs. |
| 192 | * Define a mapping to these extra registers. | 190 | * Those extra MSRs end up being shared for all events on |
| 191 | * a PMU and sometimes between PMU of sibling HT threads. | ||
| 192 | * In either case, the kernel needs to handle conflicting | ||
| 193 | * accesses to those extra, shared, regs. The data structure | ||
| 194 | * to manage those registers is stored in cpu_hw_event. | ||
| 193 | */ | 195 | */ |
| 194 | struct extra_reg { | 196 | struct extra_reg { |
| 195 | unsigned int event; | 197 | unsigned int event; |
| 196 | unsigned int msr; | 198 | unsigned int msr; |
| 197 | u64 config_mask; | 199 | u64 config_mask; |
| 198 | u64 valid_mask; | 200 | u64 valid_mask; |
| 201 | int idx; /* per_xxx->regs[] reg index */ | ||
| 199 | }; | 202 | }; |
| 200 | 203 | ||
| 201 | #define EVENT_EXTRA_REG(e, ms, m, vm) { \ | 204 | #define EVENT_EXTRA_REG(e, ms, m, vm, i) { \ |
| 202 | .event = (e), \ | 205 | .event = (e), \ |
| 203 | .msr = (ms), \ | 206 | .msr = (ms), \ |
| 204 | .config_mask = (m), \ | 207 | .config_mask = (m), \ |
| 205 | .valid_mask = (vm), \ | 208 | .valid_mask = (vm), \ |
| 209 | .idx = EXTRA_REG_##i \ | ||
| 206 | } | 210 | } |
| 207 | #define INTEL_EVENT_EXTRA_REG(event, msr, vm) \ | 211 | |
| 208 | EVENT_EXTRA_REG(event, msr, ARCH_PERFMON_EVENTSEL_EVENT, vm) | 212 | #define INTEL_EVENT_EXTRA_REG(event, msr, vm, idx) \ |
| 209 | #define EVENT_EXTRA_END EVENT_EXTRA_REG(0, 0, 0, 0) | 213 | EVENT_EXTRA_REG(event, msr, ARCH_PERFMON_EVENTSEL_EVENT, vm, idx) |
| 214 | |||
| 215 | #define EVENT_EXTRA_END EVENT_EXTRA_REG(0, 0, 0, 0, RSP_0) | ||
| 210 | 216 | ||
| 211 | union perf_capabilities { | 217 | union perf_capabilities { |
| 212 | struct { | 218 | struct { |
| @@ -252,7 +258,6 @@ struct x86_pmu { | |||
| 252 | void (*put_event_constraints)(struct cpu_hw_events *cpuc, | 258 | void (*put_event_constraints)(struct cpu_hw_events *cpuc, |
| 253 | struct perf_event *event); | 259 | struct perf_event *event); |
| 254 | struct event_constraint *event_constraints; | 260 | struct event_constraint *event_constraints; |
| 255 | struct event_constraint *percore_constraints; | ||
| 256 | void (*quirks)(void); | 261 | void (*quirks)(void); |
| 257 | int perfctr_second_write; | 262 | int perfctr_second_write; |
| 258 | 263 | ||
| @@ -286,8 +291,12 @@ struct x86_pmu { | |||
| 286 | * Extra registers for events | 291 | * Extra registers for events |
| 287 | */ | 292 | */ |
| 288 | struct extra_reg *extra_regs; | 293 | struct extra_reg *extra_regs; |
| 294 | unsigned int er_flags; | ||
| 289 | }; | 295 | }; |
| 290 | 296 | ||
| 297 | #define ERF_NO_HT_SHARING 1 | ||
| 298 | #define ERF_HAS_RSP_1 2 | ||
| 299 | |||
| 291 | static struct x86_pmu x86_pmu __read_mostly; | 300 | static struct x86_pmu x86_pmu __read_mostly; |
| 292 | 301 | ||
| 293 | static DEFINE_PER_CPU(struct cpu_hw_events, cpu_hw_events) = { | 302 | static DEFINE_PER_CPU(struct cpu_hw_events, cpu_hw_events) = { |
| @@ -393,10 +402,10 @@ static inline unsigned int x86_pmu_event_addr(int index) | |||
| 393 | */ | 402 | */ |
| 394 | static int x86_pmu_extra_regs(u64 config, struct perf_event *event) | 403 | static int x86_pmu_extra_regs(u64 config, struct perf_event *event) |
| 395 | { | 404 | { |
| 405 | struct hw_perf_event_extra *reg; | ||
| 396 | struct extra_reg *er; | 406 | struct extra_reg *er; |
| 397 | 407 | ||
| 398 | event->hw.extra_reg = 0; | 408 | reg = &event->hw.extra_reg; |
| 399 | event->hw.extra_config = 0; | ||
| 400 | 409 | ||
| 401 | if (!x86_pmu.extra_regs) | 410 | if (!x86_pmu.extra_regs) |
| 402 | return 0; | 411 | return 0; |
| @@ -406,8 +415,10 @@ static int x86_pmu_extra_regs(u64 config, struct perf_event *event) | |||
| 406 | continue; | 415 | continue; |
| 407 | if (event->attr.config1 & ~er->valid_mask) | 416 | if (event->attr.config1 & ~er->valid_mask) |
| 408 | return -EINVAL; | 417 | return -EINVAL; |
| 409 | event->hw.extra_reg = er->msr; | 418 | |
| 410 | event->hw.extra_config = event->attr.config1; | 419 | reg->idx = er->idx; |
| 420 | reg->config = event->attr.config1; | ||
| 421 | reg->reg = er->msr; | ||
| 411 | break; | 422 | break; |
| 412 | } | 423 | } |
| 413 | return 0; | 424 | return 0; |
| @@ -706,6 +717,9 @@ static int __x86_pmu_event_init(struct perf_event *event) | |||
| 706 | event->hw.last_cpu = -1; | 717 | event->hw.last_cpu = -1; |
| 707 | event->hw.last_tag = ~0ULL; | 718 | event->hw.last_tag = ~0ULL; |
| 708 | 719 | ||
| 720 | /* mark unused */ | ||
| 721 | event->hw.extra_reg.idx = EXTRA_REG_NONE; | ||
| 722 | |||
| 709 | return x86_pmu.hw_config(event); | 723 | return x86_pmu.hw_config(event); |
| 710 | } | 724 | } |
| 711 | 725 | ||
| @@ -747,8 +761,8 @@ static void x86_pmu_disable(struct pmu *pmu) | |||
| 747 | static inline void __x86_pmu_enable_event(struct hw_perf_event *hwc, | 761 | static inline void __x86_pmu_enable_event(struct hw_perf_event *hwc, |
| 748 | u64 enable_mask) | 762 | u64 enable_mask) |
| 749 | { | 763 | { |
| 750 | if (hwc->extra_reg) | 764 | if (hwc->extra_reg.reg) |
| 751 | wrmsrl(hwc->extra_reg, hwc->extra_config); | 765 | wrmsrl(hwc->extra_reg.reg, hwc->extra_reg.config); |
| 752 | wrmsrl(hwc->config_base, hwc->config | enable_mask); | 766 | wrmsrl(hwc->config_base, hwc->config | enable_mask); |
| 753 | } | 767 | } |
| 754 | 768 | ||
| @@ -1332,7 +1346,7 @@ static int x86_pmu_handle_irq(struct pt_regs *regs) | |||
| 1332 | if (!x86_perf_event_set_period(event)) | 1346 | if (!x86_perf_event_set_period(event)) |
| 1333 | continue; | 1347 | continue; |
| 1334 | 1348 | ||
| 1335 | if (perf_event_overflow(event, 1, &data, regs)) | 1349 | if (perf_event_overflow(event, &data, regs)) |
| 1336 | x86_pmu_stop(event, 0); | 1350 | x86_pmu_stop(event, 0); |
| 1337 | } | 1351 | } |
| 1338 | 1352 | ||
| @@ -1637,6 +1651,40 @@ static int x86_pmu_commit_txn(struct pmu *pmu) | |||
| 1637 | perf_pmu_enable(pmu); | 1651 | perf_pmu_enable(pmu); |
| 1638 | return 0; | 1652 | return 0; |
| 1639 | } | 1653 | } |
| 1654 | /* | ||
| 1655 | * a fake_cpuc is used to validate event groups. Due to | ||
| 1656 | * the extra reg logic, we need to also allocate a fake | ||
| 1657 | * per_core and per_cpu structure. Otherwise, group events | ||
| 1658 | * using extra reg may conflict without the kernel being | ||
| 1659 | * able to catch this when the last event gets added to | ||
| 1660 | * the group. | ||
| 1661 | */ | ||
| 1662 | static void free_fake_cpuc(struct cpu_hw_events *cpuc) | ||
| 1663 | { | ||
| 1664 | kfree(cpuc->shared_regs); | ||
| 1665 | kfree(cpuc); | ||
| 1666 | } | ||
| 1667 | |||
| 1668 | static struct cpu_hw_events *allocate_fake_cpuc(void) | ||
| 1669 | { | ||
| 1670 | struct cpu_hw_events *cpuc; | ||
| 1671 | int cpu = raw_smp_processor_id(); | ||
| 1672 | |||
| 1673 | cpuc = kzalloc(sizeof(*cpuc), GFP_KERNEL); | ||
| 1674 | if (!cpuc) | ||
| 1675 | return ERR_PTR(-ENOMEM); | ||
| 1676 | |||
| 1677 | /* only needed, if we have extra_regs */ | ||
| 1678 | if (x86_pmu.extra_regs) { | ||
| 1679 | cpuc->shared_regs = allocate_shared_regs(cpu); | ||
| 1680 | if (!cpuc->shared_regs) | ||
| 1681 | goto error; | ||
| 1682 | } | ||
| 1683 | return cpuc; | ||
| 1684 | error: | ||
| 1685 | free_fake_cpuc(cpuc); | ||
| 1686 | return ERR_PTR(-ENOMEM); | ||
| 1687 | } | ||
| 1640 | 1688 | ||
| 1641 | /* | 1689 | /* |
| 1642 | * validate that we can schedule this event | 1690 | * validate that we can schedule this event |
| @@ -1647,9 +1695,9 @@ static int validate_event(struct perf_event *event) | |||
| 1647 | struct event_constraint *c; | 1695 | struct event_constraint *c; |
| 1648 | int ret = 0; | 1696 | int ret = 0; |
| 1649 | 1697 | ||
| 1650 | fake_cpuc = kmalloc(sizeof(*fake_cpuc), GFP_KERNEL | __GFP_ZERO); | 1698 | fake_cpuc = allocate_fake_cpuc(); |
| 1651 | if (!fake_cpuc) | 1699 | if (IS_ERR(fake_cpuc)) |
| 1652 | return -ENOMEM; | 1700 | return PTR_ERR(fake_cpuc); |
| 1653 | 1701 | ||
| 1654 | c = x86_pmu.get_event_constraints(fake_cpuc, event); | 1702 | c = x86_pmu.get_event_constraints(fake_cpuc, event); |
| 1655 | 1703 | ||
| @@ -1659,7 +1707,7 @@ static int validate_event(struct perf_event *event) | |||
| 1659 | if (x86_pmu.put_event_constraints) | 1707 | if (x86_pmu.put_event_constraints) |
| 1660 | x86_pmu.put_event_constraints(fake_cpuc, event); | 1708 | x86_pmu.put_event_constraints(fake_cpuc, event); |
| 1661 | 1709 | ||
| 1662 | kfree(fake_cpuc); | 1710 | free_fake_cpuc(fake_cpuc); |
| 1663 | 1711 | ||
| 1664 | return ret; | 1712 | return ret; |
| 1665 | } | 1713 | } |
| @@ -1679,36 +1727,32 @@ static int validate_group(struct perf_event *event) | |||
| 1679 | { | 1727 | { |
| 1680 | struct perf_event *leader = event->group_leader; | 1728 | struct perf_event *leader = event->group_leader; |
| 1681 | struct cpu_hw_events *fake_cpuc; | 1729 | struct cpu_hw_events *fake_cpuc; |
| 1682 | int ret, n; | 1730 | int ret = -ENOSPC, n; |
| 1683 | |||
| 1684 | ret = -ENOMEM; | ||
| 1685 | fake_cpuc = kmalloc(sizeof(*fake_cpuc), GFP_KERNEL | __GFP_ZERO); | ||
| 1686 | if (!fake_cpuc) | ||
| 1687 | goto out; | ||
| 1688 | 1731 | ||
| 1732 | fake_cpuc = allocate_fake_cpuc(); | ||
| 1733 | if (IS_ERR(fake_cpuc)) | ||
| 1734 | return PTR_ERR(fake_cpuc); | ||
| 1689 | /* | 1735 | /* |
| 1690 | * the event is not yet connected with its | 1736 | * the event is not yet connected with its |
| 1691 | * siblings therefore we must first collect | 1737 | * siblings therefore we must first collect |
| 1692 | * existing siblings, then add the new event | 1738 | * existing siblings, then add the new event |
| 1693 | * before we can simulate the scheduling | 1739 | * before we can simulate the scheduling |
| 1694 | */ | 1740 | */ |
| 1695 | ret = -ENOSPC; | ||
| 1696 | n = collect_events(fake_cpuc, leader, true); | 1741 | n = collect_events(fake_cpuc, leader, true); |
| 1697 | if (n < 0) | 1742 | if (n < 0) |
| 1698 | goto out_free; | 1743 | goto out; |
| 1699 | 1744 | ||
| 1700 | fake_cpuc->n_events = n; | 1745 | fake_cpuc->n_events = n; |
| 1701 | n = collect_events(fake_cpuc, event, false); | 1746 | n = collect_events(fake_cpuc, event, false); |
| 1702 | if (n < 0) | 1747 | if (n < 0) |
| 1703 | goto out_free; | 1748 | goto out; |
| 1704 | 1749 | ||
| 1705 | fake_cpuc->n_events = n; | 1750 | fake_cpuc->n_events = n; |
| 1706 | 1751 | ||
| 1707 | ret = x86_pmu.schedule_events(fake_cpuc, n, NULL); | 1752 | ret = x86_pmu.schedule_events(fake_cpuc, n, NULL); |
| 1708 | 1753 | ||
| 1709 | out_free: | ||
| 1710 | kfree(fake_cpuc); | ||
| 1711 | out: | 1754 | out: |
| 1755 | free_fake_cpuc(fake_cpuc); | ||
| 1712 | return ret; | 1756 | return ret; |
| 1713 | } | 1757 | } |
| 1714 | 1758 | ||
diff --git a/arch/x86/kernel/cpu/perf_event_amd.c b/arch/x86/kernel/cpu/perf_event_amd.c index fe29c1d2219..941caa2e449 100644 --- a/arch/x86/kernel/cpu/perf_event_amd.c +++ b/arch/x86/kernel/cpu/perf_event_amd.c | |||
| @@ -89,6 +89,20 @@ static __initconst const u64 amd_hw_cache_event_ids | |||
| 89 | [ C(RESULT_MISS) ] = -1, | 89 | [ C(RESULT_MISS) ] = -1, |
| 90 | }, | 90 | }, |
| 91 | }, | 91 | }, |
| 92 | [ C(NODE) ] = { | ||
| 93 | [ C(OP_READ) ] = { | ||
| 94 | [ C(RESULT_ACCESS) ] = 0xb8e9, /* CPU Request to Memory, l+r */ | ||
| 95 | [ C(RESULT_MISS) ] = 0x98e9, /* CPU Request to Memory, r */ | ||
| 96 | }, | ||
| 97 | [ C(OP_WRITE) ] = { | ||
| 98 | [ C(RESULT_ACCESS) ] = -1, | ||
| 99 | [ C(RESULT_MISS) ] = -1, | ||
| 100 | }, | ||
| 101 | [ C(OP_PREFETCH) ] = { | ||
| 102 | [ C(RESULT_ACCESS) ] = -1, | ||
| 103 | [ C(RESULT_MISS) ] = -1, | ||
| 104 | }, | ||
| 105 | }, | ||
| 92 | }; | 106 | }; |
| 93 | 107 | ||
| 94 | /* | 108 | /* |
diff --git a/arch/x86/kernel/cpu/perf_event_intel.c b/arch/x86/kernel/cpu/perf_event_intel.c index 41178c826c4..45fbb8f7f54 100644 --- a/arch/x86/kernel/cpu/perf_event_intel.c +++ b/arch/x86/kernel/cpu/perf_event_intel.c | |||
| @@ -1,25 +1,15 @@ | |||
| 1 | #ifdef CONFIG_CPU_SUP_INTEL | 1 | #ifdef CONFIG_CPU_SUP_INTEL |
| 2 | 2 | ||
| 3 | #define MAX_EXTRA_REGS 2 | ||
| 4 | |||
| 5 | /* | ||
| 6 | * Per register state. | ||
| 7 | */ | ||
| 8 | struct er_account { | ||
| 9 | int ref; /* reference count */ | ||
| 10 | unsigned int extra_reg; /* extra MSR number */ | ||
| 11 | u64 extra_config; /* extra MSR config */ | ||
| 12 | }; | ||
| 13 | |||
| 14 | /* | 3 | /* |
| 15 | * Per core state | 4 | * Per core/cpu state |
| 16 | * This used to coordinate shared registers for HT threads. | 5 | * |
| 6 | * Used to coordinate shared registers between HT threads or | ||
| 7 | * among events on a single PMU. | ||
| 17 | */ | 8 | */ |
| 18 | struct intel_percore { | 9 | struct intel_shared_regs { |
| 19 | raw_spinlock_t lock; /* protect structure */ | 10 | struct er_account regs[EXTRA_REG_MAX]; |
| 20 | struct er_account regs[MAX_EXTRA_REGS]; | 11 | int refcnt; /* per-core: #HT threads */ |
| 21 | int refcnt; /* number of threads */ | 12 | unsigned core_id; /* per-core: core id */ |
| 22 | unsigned core_id; | ||
| 23 | }; | 13 | }; |
| 24 | 14 | ||
| 25 | /* | 15 | /* |
| @@ -88,16 +78,10 @@ static struct event_constraint intel_nehalem_event_constraints[] __read_mostly = | |||
| 88 | 78 | ||
| 89 | static struct extra_reg intel_nehalem_extra_regs[] __read_mostly = | 79 | static struct extra_reg intel_nehalem_extra_regs[] __read_mostly = |
| 90 | { | 80 | { |
| 91 | INTEL_EVENT_EXTRA_REG(0xb7, MSR_OFFCORE_RSP_0, 0xffff), | 81 | INTEL_EVENT_EXTRA_REG(0xb7, MSR_OFFCORE_RSP_0, 0xffff, RSP_0), |
| 92 | EVENT_EXTRA_END | 82 | EVENT_EXTRA_END |
| 93 | }; | 83 | }; |
| 94 | 84 | ||
| 95 | static struct event_constraint intel_nehalem_percore_constraints[] __read_mostly = | ||
| 96 | { | ||
| 97 | INTEL_EVENT_CONSTRAINT(0xb7, 0), | ||
| 98 | EVENT_CONSTRAINT_END | ||
| 99 | }; | ||
| 100 | |||
| 101 | static struct event_constraint intel_westmere_event_constraints[] __read_mostly = | 85 | static struct event_constraint intel_westmere_event_constraints[] __read_mostly = |
| 102 | { | 86 | { |
| 103 | FIXED_EVENT_CONSTRAINT(0x00c0, 0), /* INST_RETIRED.ANY */ | 87 | FIXED_EVENT_CONSTRAINT(0x00c0, 0), /* INST_RETIRED.ANY */ |
| @@ -116,8 +100,6 @@ static struct event_constraint intel_snb_event_constraints[] __read_mostly = | |||
| 116 | FIXED_EVENT_CONSTRAINT(0x003c, 1), /* CPU_CLK_UNHALTED.CORE */ | 100 | FIXED_EVENT_CONSTRAINT(0x003c, 1), /* CPU_CLK_UNHALTED.CORE */ |
| 117 | /* FIXED_EVENT_CONSTRAINT(0x013c, 2), CPU_CLK_UNHALTED.REF */ | 101 | /* FIXED_EVENT_CONSTRAINT(0x013c, 2), CPU_CLK_UNHALTED.REF */ |
| 118 | INTEL_EVENT_CONSTRAINT(0x48, 0x4), /* L1D_PEND_MISS.PENDING */ | 102 | INTEL_EVENT_CONSTRAINT(0x48, 0x4), /* L1D_PEND_MISS.PENDING */ |
| 119 | INTEL_EVENT_CONSTRAINT(0xb7, 0x1), /* OFF_CORE_RESPONSE_0 */ | ||
| 120 | INTEL_EVENT_CONSTRAINT(0xbb, 0x8), /* OFF_CORE_RESPONSE_1 */ | ||
| 121 | INTEL_UEVENT_CONSTRAINT(0x01c0, 0x2), /* INST_RETIRED.PREC_DIST */ | 103 | INTEL_UEVENT_CONSTRAINT(0x01c0, 0x2), /* INST_RETIRED.PREC_DIST */ |
| 122 | INTEL_EVENT_CONSTRAINT(0xcd, 0x8), /* MEM_TRANS_RETIRED.LOAD_LATENCY */ | 104 | INTEL_EVENT_CONSTRAINT(0xcd, 0x8), /* MEM_TRANS_RETIRED.LOAD_LATENCY */ |
| 123 | EVENT_CONSTRAINT_END | 105 | EVENT_CONSTRAINT_END |
| @@ -125,15 +107,13 @@ static struct event_constraint intel_snb_event_constraints[] __read_mostly = | |||
| 125 | 107 | ||
| 126 | static struct extra_reg intel_westmere_extra_regs[] __read_mostly = | 108 | static struct extra_reg intel_westmere_extra_regs[] __read_mostly = |
| 127 | { | 109 | { |
| 128 | INTEL_EVENT_EXTRA_REG(0xb7, MSR_OFFCORE_RSP_0, 0xffff), | 110 | INTEL_EVENT_EXTRA_REG(0xb7, MSR_OFFCORE_RSP_0, 0xffff, RSP_0), |
| 129 | INTEL_EVENT_EXTRA_REG(0xbb, MSR_OFFCORE_RSP_1, 0xffff), | 111 | INTEL_EVENT_EXTRA_REG(0xbb, MSR_OFFCORE_RSP_1, 0xffff, RSP_1), |
| 130 | EVENT_EXTRA_END | 112 | EVENT_EXTRA_END |
| 131 | }; | 113 | }; |
| 132 | 114 | ||
| 133 | static struct event_constraint intel_westmere_percore_constraints[] __read_mostly = | 115 | static struct event_constraint intel_v1_event_constraints[] __read_mostly = |
| 134 | { | 116 | { |
| 135 | INTEL_EVENT_CONSTRAINT(0xb7, 0), | ||
| 136 | INTEL_EVENT_CONSTRAINT(0xbb, 0), | ||
| 137 | EVENT_CONSTRAINT_END | 117 | EVENT_CONSTRAINT_END |
| 138 | }; | 118 | }; |
| 139 | 119 | ||
| @@ -145,6 +125,12 @@ static struct event_constraint intel_gen_event_constraints[] __read_mostly = | |||
| 145 | EVENT_CONSTRAINT_END | 125 | EVENT_CONSTRAINT_END |
| 146 | }; | 126 | }; |
| 147 | 127 | ||
| 128 | static struct extra_reg intel_snb_extra_regs[] __read_mostly = { | ||
| 129 | INTEL_EVENT_EXTRA_REG(0xb7, MSR_OFFCORE_RSP_0, 0x3fffffffffull, RSP_0), | ||
| 130 | INTEL_EVENT_EXTRA_REG(0xbb, MSR_OFFCORE_RSP_1, 0x3fffffffffull, RSP_1), | ||
| 131 | EVENT_EXTRA_END | ||
| 132 | }; | ||
| 133 | |||
| 148 | static u64 intel_pmu_event_map(int hw_event) | 134 | static u64 intel_pmu_event_map(int hw_event) |
| 149 | { | 135 | { |
| 150 | return intel_perfmon_event_map[hw_event]; | 136 | return intel_perfmon_event_map[hw_event]; |
| @@ -245,6 +231,21 @@ static __initconst const u64 snb_hw_cache_event_ids | |||
| 245 | [ C(RESULT_MISS) ] = -1, | 231 | [ C(RESULT_MISS) ] = -1, |
| 246 | }, | 232 | }, |
| 247 | }, | 233 | }, |
| 234 | [ C(NODE) ] = { | ||
| 235 | [ C(OP_READ) ] = { | ||
| 236 | [ C(RESULT_ACCESS) ] = -1, | ||
| 237 | [ C(RESULT_MISS) ] = -1, | ||
| 238 | }, | ||
| 239 | [ C(OP_WRITE) ] = { | ||
| 240 | [ C(RESULT_ACCESS) ] = -1, | ||
| 241 | [ C(RESULT_MISS) ] = -1, | ||
| 242 | }, | ||
| 243 | [ C(OP_PREFETCH) ] = { | ||
| 244 | [ C(RESULT_ACCESS) ] = -1, | ||
| 245 | [ C(RESULT_MISS) ] = -1, | ||
| 246 | }, | ||
| 247 | }, | ||
| 248 | |||
| 248 | }; | 249 | }; |
| 249 | 250 | ||
| 250 | static __initconst const u64 westmere_hw_cache_event_ids | 251 | static __initconst const u64 westmere_hw_cache_event_ids |
| @@ -346,6 +347,20 @@ static __initconst const u64 westmere_hw_cache_event_ids | |||
| 346 | [ C(RESULT_MISS) ] = -1, | 347 | [ C(RESULT_MISS) ] = -1, |
| 347 | }, | 348 | }, |
| 348 | }, | 349 | }, |
| 350 | [ C(NODE) ] = { | ||
| 351 | [ C(OP_READ) ] = { | ||
| 352 | [ C(RESULT_ACCESS) ] = 0x01b7, | ||
| 353 | [ C(RESULT_MISS) ] = 0x01b7, | ||
| 354 | }, | ||
| 355 | [ C(OP_WRITE) ] = { | ||
| 356 | [ C(RESULT_ACCESS) ] = 0x01b7, | ||
| 357 | [ C(RESULT_MISS) ] = 0x01b7, | ||
| 358 | }, | ||
| 359 | [ C(OP_PREFETCH) ] = { | ||
| 360 | [ C(RESULT_ACCESS) ] = 0x01b7, | ||
| 361 | [ C(RESULT_MISS) ] = 0x01b7, | ||
| 362 | }, | ||
| 363 | }, | ||
| 349 | }; | 364 | }; |
| 350 | 365 | ||
| 351 | /* | 366 | /* |
| @@ -398,7 +413,21 @@ static __initconst const u64 nehalem_hw_cache_extra_regs | |||
| 398 | [ C(RESULT_ACCESS) ] = NHM_DMND_PREFETCH|NHM_L3_ACCESS, | 413 | [ C(RESULT_ACCESS) ] = NHM_DMND_PREFETCH|NHM_L3_ACCESS, |
| 399 | [ C(RESULT_MISS) ] = NHM_DMND_PREFETCH|NHM_L3_MISS, | 414 | [ C(RESULT_MISS) ] = NHM_DMND_PREFETCH|NHM_L3_MISS, |
| 400 | }, | 415 | }, |
| 401 | } | 416 | }, |
| 417 | [ C(NODE) ] = { | ||
| 418 | [ C(OP_READ) ] = { | ||
| 419 | [ C(RESULT_ACCESS) ] = NHM_DMND_READ|NHM_ALL_DRAM, | ||
| 420 | [ C(RESULT_MISS) ] = NHM_DMND_READ|NHM_REMOTE_DRAM, | ||
| 421 | }, | ||
| 422 | [ C(OP_WRITE) ] = { | ||
| 423 | [ C(RESULT_ACCESS) ] = NHM_DMND_WRITE|NHM_ALL_DRAM, | ||
| 424 | [ C(RESULT_MISS) ] = NHM_DMND_WRITE|NHM_REMOTE_DRAM, | ||
| 425 | }, | ||
| 426 | [ C(OP_PREFETCH) ] = { | ||
| 427 | [ C(RESULT_ACCESS) ] = NHM_DMND_PREFETCH|NHM_ALL_DRAM, | ||
| 428 | [ C(RESULT_MISS) ] = NHM_DMND_PREFETCH|NHM_REMOTE_DRAM, | ||
| 429 | }, | ||
| 430 | }, | ||
| 402 | }; | 431 | }; |
| 403 | 432 | ||
| 404 | static __initconst const u64 nehalem_hw_cache_event_ids | 433 | static __initconst const u64 nehalem_hw_cache_event_ids |
| @@ -500,6 +529,20 @@ static __initconst const u64 nehalem_hw_cache_event_ids | |||
| 500 | [ C(RESULT_MISS) ] = -1, | 529 | [ C(RESULT_MISS) ] = -1, |
| 501 | }, | 530 | }, |
| 502 | }, | 531 | }, |
| 532 | [ C(NODE) ] = { | ||
| 533 | [ C(OP_READ) ] = { | ||
| 534 | [ C(RESULT_ACCESS) ] = 0x01b7, | ||
| 535 | [ C(RESULT_MISS) ] = 0x01b7, | ||
| 536 | }, | ||
| 537 | [ C(OP_WRITE) ] = { | ||
| 538 | [ C(RESULT_ACCESS) ] = 0x01b7, | ||
| 539 | [ C(RESULT_MISS) ] = 0x01b7, | ||
| 540 | }, | ||
| 541 | [ C(OP_PREFETCH) ] = { | ||
| 542 | [ C(RESULT_ACCESS) ] = 0x01b7, | ||
| 543 | [ C(RESULT_MISS) ] = 0x01b7, | ||
| 544 | }, | ||
| 545 | }, | ||
| 503 | }; | 546 | }; |
| 504 | 547 | ||
| 505 | static __initconst const u64 core2_hw_cache_event_ids | 548 | static __initconst const u64 core2_hw_cache_event_ids |
| @@ -1003,7 +1046,7 @@ again: | |||
| 1003 | 1046 | ||
| 1004 | data.period = event->hw.last_period; | 1047 | data.period = event->hw.last_period; |
| 1005 | 1048 | ||
| 1006 | if (perf_event_overflow(event, 1, &data, regs)) | 1049 | if (perf_event_overflow(event, &data, regs)) |
| 1007 | x86_pmu_stop(event, 0); | 1050 | x86_pmu_stop(event, 0); |
| 1008 | } | 1051 | } |
| 1009 | 1052 | ||
| @@ -1037,65 +1080,121 @@ intel_bts_constraints(struct perf_event *event) | |||
| 1037 | return NULL; | 1080 | return NULL; |
| 1038 | } | 1081 | } |
| 1039 | 1082 | ||
| 1083 | static bool intel_try_alt_er(struct perf_event *event, int orig_idx) | ||
| 1084 | { | ||
| 1085 | if (!(x86_pmu.er_flags & ERF_HAS_RSP_1)) | ||
| 1086 | return false; | ||
| 1087 | |||
| 1088 | if (event->hw.extra_reg.idx == EXTRA_REG_RSP_0) { | ||
| 1089 | event->hw.config &= ~INTEL_ARCH_EVENT_MASK; | ||
| 1090 | event->hw.config |= 0x01bb; | ||
| 1091 | event->hw.extra_reg.idx = EXTRA_REG_RSP_1; | ||
| 1092 | event->hw.extra_reg.reg = MSR_OFFCORE_RSP_1; | ||
| 1093 | } else if (event->hw.extra_reg.idx == EXTRA_REG_RSP_1) { | ||
| 1094 | event->hw.config &= ~INTEL_ARCH_EVENT_MASK; | ||
| 1095 | event->hw.config |= 0x01b7; | ||
| 1096 | event->hw.extra_reg.idx = EXTRA_REG_RSP_0; | ||
| 1097 | event->hw.extra_reg.reg = MSR_OFFCORE_RSP_0; | ||
| 1098 | } | ||
| 1099 | |||
| 1100 | if (event->hw.extra_reg.idx == orig_idx) | ||
| 1101 | return false; | ||
| 1102 | |||
| 1103 | return true; | ||
| 1104 | } | ||
| 1105 | |||
| 1106 | /* | ||
| 1107 | * manage allocation of shared extra msr for certain events | ||
| 1108 | * | ||
| 1109 | * sharing can be: | ||
| 1110 | * per-cpu: to be shared between the various events on a single PMU | ||
| 1111 | * per-core: per-cpu + shared by HT threads | ||
| 1112 | */ | ||
| 1040 | static struct event_constraint * | 1113 | static struct event_constraint * |
| 1041 | intel_percore_constraints(struct cpu_hw_events *cpuc, struct perf_event *event) | 1114 | __intel_shared_reg_get_constraints(struct cpu_hw_events *cpuc, |
| 1115 | struct perf_event *event) | ||
| 1042 | { | 1116 | { |
| 1043 | struct hw_perf_event *hwc = &event->hw; | 1117 | struct event_constraint *c = &emptyconstraint; |
| 1044 | unsigned int e = hwc->config & ARCH_PERFMON_EVENTSEL_EVENT; | 1118 | struct hw_perf_event_extra *reg = &event->hw.extra_reg; |
| 1045 | struct event_constraint *c; | ||
| 1046 | struct intel_percore *pc; | ||
| 1047 | struct er_account *era; | 1119 | struct er_account *era; |
| 1048 | int i; | 1120 | unsigned long flags; |
| 1049 | int free_slot; | 1121 | int orig_idx = reg->idx; |
| 1050 | int found; | ||
| 1051 | 1122 | ||
| 1052 | if (!x86_pmu.percore_constraints || hwc->extra_alloc) | 1123 | /* already allocated shared msr */ |
| 1053 | return NULL; | 1124 | if (reg->alloc) |
| 1125 | return &unconstrained; | ||
| 1054 | 1126 | ||
| 1055 | for (c = x86_pmu.percore_constraints; c->cmask; c++) { | 1127 | again: |
| 1056 | if (e != c->code) | 1128 | era = &cpuc->shared_regs->regs[reg->idx]; |
| 1057 | continue; | 1129 | /* |
| 1130 | * we use spin_lock_irqsave() to avoid lockdep issues when | ||
| 1131 | * passing a fake cpuc | ||
| 1132 | */ | ||
| 1133 | raw_spin_lock_irqsave(&era->lock, flags); | ||
| 1134 | |||
| 1135 | if (!atomic_read(&era->ref) || era->config == reg->config) { | ||
| 1136 | |||
| 1137 | /* lock in msr value */ | ||
| 1138 | era->config = reg->config; | ||
| 1139 | era->reg = reg->reg; | ||
| 1140 | |||
| 1141 | /* one more user */ | ||
| 1142 | atomic_inc(&era->ref); | ||
| 1143 | |||
| 1144 | /* no need to reallocate during incremental event scheduling */ | ||
| 1145 | reg->alloc = 1; | ||
| 1058 | 1146 | ||
| 1059 | /* | 1147 | /* |
| 1060 | * Allocate resource per core. | 1148 | * All events using extra_reg are unconstrained. |
| 1149 | * Avoids calling x86_get_event_constraints() | ||
| 1150 | * | ||
| 1151 | * Must revisit if extra_reg controlling events | ||
| 1152 | * ever have constraints. Worst case we go through | ||
| 1153 | * the regular event constraint table. | ||
| 1061 | */ | 1154 | */ |
| 1062 | pc = cpuc->per_core; | 1155 | c = &unconstrained; |
| 1063 | if (!pc) | 1156 | } else if (intel_try_alt_er(event, orig_idx)) { |
| 1064 | break; | 1157 | raw_spin_unlock(&era->lock); |
| 1065 | c = &emptyconstraint; | 1158 | goto again; |
| 1066 | raw_spin_lock(&pc->lock); | ||
| 1067 | free_slot = -1; | ||
| 1068 | found = 0; | ||
| 1069 | for (i = 0; i < MAX_EXTRA_REGS; i++) { | ||
| 1070 | era = &pc->regs[i]; | ||
| 1071 | if (era->ref > 0 && hwc->extra_reg == era->extra_reg) { | ||
| 1072 | /* Allow sharing same config */ | ||
| 1073 | if (hwc->extra_config == era->extra_config) { | ||
| 1074 | era->ref++; | ||
| 1075 | cpuc->percore_used = 1; | ||
| 1076 | hwc->extra_alloc = 1; | ||
| 1077 | c = NULL; | ||
| 1078 | } | ||
| 1079 | /* else conflict */ | ||
| 1080 | found = 1; | ||
| 1081 | break; | ||
| 1082 | } else if (era->ref == 0 && free_slot == -1) | ||
| 1083 | free_slot = i; | ||
| 1084 | } | ||
| 1085 | if (!found && free_slot != -1) { | ||
| 1086 | era = &pc->regs[free_slot]; | ||
| 1087 | era->ref = 1; | ||
| 1088 | era->extra_reg = hwc->extra_reg; | ||
| 1089 | era->extra_config = hwc->extra_config; | ||
| 1090 | cpuc->percore_used = 1; | ||
| 1091 | hwc->extra_alloc = 1; | ||
| 1092 | c = NULL; | ||
| 1093 | } | ||
| 1094 | raw_spin_unlock(&pc->lock); | ||
| 1095 | return c; | ||
| 1096 | } | 1159 | } |
| 1160 | raw_spin_unlock_irqrestore(&era->lock, flags); | ||
| 1097 | 1161 | ||
| 1098 | return NULL; | 1162 | return c; |
| 1163 | } | ||
| 1164 | |||
| 1165 | static void | ||
| 1166 | __intel_shared_reg_put_constraints(struct cpu_hw_events *cpuc, | ||
| 1167 | struct hw_perf_event_extra *reg) | ||
| 1168 | { | ||
| 1169 | struct er_account *era; | ||
| 1170 | |||
| 1171 | /* | ||
| 1172 | * only put constraint if extra reg was actually | ||
| 1173 | * allocated. Also takes care of event which do | ||
| 1174 | * not use an extra shared reg | ||
| 1175 | */ | ||
| 1176 | if (!reg->alloc) | ||
| 1177 | return; | ||
| 1178 | |||
| 1179 | era = &cpuc->shared_regs->regs[reg->idx]; | ||
| 1180 | |||
| 1181 | /* one fewer user */ | ||
| 1182 | atomic_dec(&era->ref); | ||
| 1183 | |||
| 1184 | /* allocate again next time */ | ||
| 1185 | reg->alloc = 0; | ||
| 1186 | } | ||
| 1187 | |||
| 1188 | static struct event_constraint * | ||
| 1189 | intel_shared_regs_constraints(struct cpu_hw_events *cpuc, | ||
| 1190 | struct perf_event *event) | ||
| 1191 | { | ||
| 1192 | struct event_constraint *c = NULL; | ||
| 1193 | |||
| 1194 | if (event->hw.extra_reg.idx != EXTRA_REG_NONE) | ||
| 1195 | c = __intel_shared_reg_get_constraints(cpuc, event); | ||
| 1196 | |||
| 1197 | return c; | ||
| 1099 | } | 1198 | } |
| 1100 | 1199 | ||
| 1101 | static struct event_constraint * | 1200 | static struct event_constraint * |
| @@ -1111,49 +1210,28 @@ intel_get_event_constraints(struct cpu_hw_events *cpuc, struct perf_event *event | |||
| 1111 | if (c) | 1210 | if (c) |
| 1112 | return c; | 1211 | return c; |
| 1113 | 1212 | ||
| 1114 | c = intel_percore_constraints(cpuc, event); | 1213 | c = intel_shared_regs_constraints(cpuc, event); |
| 1115 | if (c) | 1214 | if (c) |
| 1116 | return c; | 1215 | return c; |
| 1117 | 1216 | ||
| 1118 | return x86_get_event_constraints(cpuc, event); | 1217 | return x86_get_event_constraints(cpuc, event); |
| 1119 | } | 1218 | } |
| 1120 | 1219 | ||
| 1121 | static void intel_put_event_constraints(struct cpu_hw_events *cpuc, | 1220 | static void |
| 1221 | intel_put_shared_regs_event_constraints(struct cpu_hw_events *cpuc, | ||
| 1122 | struct perf_event *event) | 1222 | struct perf_event *event) |
| 1123 | { | 1223 | { |
| 1124 | struct extra_reg *er; | 1224 | struct hw_perf_event_extra *reg; |
| 1125 | struct intel_percore *pc; | ||
| 1126 | struct er_account *era; | ||
| 1127 | struct hw_perf_event *hwc = &event->hw; | ||
| 1128 | int i, allref; | ||
| 1129 | 1225 | ||
| 1130 | if (!cpuc->percore_used) | 1226 | reg = &event->hw.extra_reg; |
| 1131 | return; | 1227 | if (reg->idx != EXTRA_REG_NONE) |
| 1132 | 1228 | __intel_shared_reg_put_constraints(cpuc, reg); | |
| 1133 | for (er = x86_pmu.extra_regs; er->msr; er++) { | 1229 | } |
| 1134 | if (er->event != (hwc->config & er->config_mask)) | ||
| 1135 | continue; | ||
| 1136 | 1230 | ||
| 1137 | pc = cpuc->per_core; | 1231 | static void intel_put_event_constraints(struct cpu_hw_events *cpuc, |
| 1138 | raw_spin_lock(&pc->lock); | 1232 | struct perf_event *event) |
| 1139 | for (i = 0; i < MAX_EXTRA_REGS; i++) { | 1233 | { |
| 1140 | era = &pc->regs[i]; | 1234 | intel_put_shared_regs_event_constraints(cpuc, event); |
| 1141 | if (era->ref > 0 && | ||
| 1142 | era->extra_config == hwc->extra_config && | ||
| 1143 | era->extra_reg == er->msr) { | ||
| 1144 | era->ref--; | ||
| 1145 | hwc->extra_alloc = 0; | ||
| 1146 | break; | ||
| 1147 | } | ||
| 1148 | } | ||
| 1149 | allref = 0; | ||
| 1150 | for (i = 0; i < MAX_EXTRA_REGS; i++) | ||
| 1151 | allref += pc->regs[i].ref; | ||
| 1152 | if (allref == 0) | ||
| 1153 | cpuc->percore_used = 0; | ||
| 1154 | raw_spin_unlock(&pc->lock); | ||
| 1155 | break; | ||
| 1156 | } | ||
| 1157 | } | 1235 | } |
| 1158 | 1236 | ||
| 1159 | static int intel_pmu_hw_config(struct perf_event *event) | 1237 | static int intel_pmu_hw_config(struct perf_event *event) |
| @@ -1231,20 +1309,36 @@ static __initconst const struct x86_pmu core_pmu = { | |||
| 1231 | .event_constraints = intel_core_event_constraints, | 1309 | .event_constraints = intel_core_event_constraints, |
| 1232 | }; | 1310 | }; |
| 1233 | 1311 | ||
| 1312 | static struct intel_shared_regs *allocate_shared_regs(int cpu) | ||
| 1313 | { | ||
| 1314 | struct intel_shared_regs *regs; | ||
| 1315 | int i; | ||
| 1316 | |||
| 1317 | regs = kzalloc_node(sizeof(struct intel_shared_regs), | ||
| 1318 | GFP_KERNEL, cpu_to_node(cpu)); | ||
| 1319 | if (regs) { | ||
| 1320 | /* | ||
| 1321 | * initialize the locks to keep lockdep happy | ||
| 1322 | */ | ||
| 1323 | for (i = 0; i < EXTRA_REG_MAX; i++) | ||
| 1324 | raw_spin_lock_init(®s->regs[i].lock); | ||
| 1325 | |||
| 1326 | regs->core_id = -1; | ||
| 1327 | } | ||
| 1328 | return regs; | ||
| 1329 | } | ||
| 1330 | |||
| 1234 | static int intel_pmu_cpu_prepare(int cpu) | 1331 | static int intel_pmu_cpu_prepare(int cpu) |
| 1235 | { | 1332 | { |
| 1236 | struct cpu_hw_events *cpuc = &per_cpu(cpu_hw_events, cpu); | 1333 | struct cpu_hw_events *cpuc = &per_cpu(cpu_hw_events, cpu); |
| 1237 | 1334 | ||
| 1238 | if (!cpu_has_ht_siblings()) | 1335 | if (!x86_pmu.extra_regs) |
| 1239 | return NOTIFY_OK; | 1336 | return NOTIFY_OK; |
| 1240 | 1337 | ||
| 1241 | cpuc->per_core = kzalloc_node(sizeof(struct intel_percore), | 1338 | cpuc->shared_regs = allocate_shared_regs(cpu); |
| 1242 | GFP_KERNEL, cpu_to_node(cpu)); | 1339 | if (!cpuc->shared_regs) |
| 1243 | if (!cpuc->per_core) | ||
| 1244 | return NOTIFY_BAD; | 1340 | return NOTIFY_BAD; |
| 1245 | 1341 | ||
| 1246 | raw_spin_lock_init(&cpuc->per_core->lock); | ||
| 1247 | cpuc->per_core->core_id = -1; | ||
| 1248 | return NOTIFY_OK; | 1342 | return NOTIFY_OK; |
| 1249 | } | 1343 | } |
| 1250 | 1344 | ||
| @@ -1260,32 +1354,34 @@ static void intel_pmu_cpu_starting(int cpu) | |||
| 1260 | */ | 1354 | */ |
| 1261 | intel_pmu_lbr_reset(); | 1355 | intel_pmu_lbr_reset(); |
| 1262 | 1356 | ||
| 1263 | if (!cpu_has_ht_siblings()) | 1357 | if (!cpuc->shared_regs || (x86_pmu.er_flags & ERF_NO_HT_SHARING)) |
| 1264 | return; | 1358 | return; |
| 1265 | 1359 | ||
| 1266 | for_each_cpu(i, topology_thread_cpumask(cpu)) { | 1360 | for_each_cpu(i, topology_thread_cpumask(cpu)) { |
| 1267 | struct intel_percore *pc = per_cpu(cpu_hw_events, i).per_core; | 1361 | struct intel_shared_regs *pc; |
| 1268 | 1362 | ||
| 1363 | pc = per_cpu(cpu_hw_events, i).shared_regs; | ||
| 1269 | if (pc && pc->core_id == core_id) { | 1364 | if (pc && pc->core_id == core_id) { |
| 1270 | kfree(cpuc->per_core); | 1365 | kfree(cpuc->shared_regs); |
| 1271 | cpuc->per_core = pc; | 1366 | cpuc->shared_regs = pc; |
| 1272 | break; | 1367 | break; |
| 1273 | } | 1368 | } |
| 1274 | } | 1369 | } |
| 1275 | 1370 | ||
| 1276 | cpuc->per_core->core_id = core_id; | 1371 | cpuc->shared_regs->core_id = core_id; |
| 1277 | cpuc->per_core->refcnt++; | 1372 | cpuc->shared_regs->refcnt++; |
| 1278 | } | 1373 | } |
| 1279 | 1374 | ||
| 1280 | static void intel_pmu_cpu_dying(int cpu) | 1375 | static void intel_pmu_cpu_dying(int cpu) |
| 1281 | { | 1376 | { |
| 1282 | struct cpu_hw_events *cpuc = &per_cpu(cpu_hw_events, cpu); | 1377 | struct cpu_hw_events *cpuc = &per_cpu(cpu_hw_events, cpu); |
| 1283 | struct intel_percore *pc = cpuc->per_core; | 1378 | struct intel_shared_regs *pc; |
| 1284 | 1379 | ||
| 1380 | pc = cpuc->shared_regs; | ||
| 1285 | if (pc) { | 1381 | if (pc) { |
| 1286 | if (pc->core_id == -1 || --pc->refcnt == 0) | 1382 | if (pc->core_id == -1 || --pc->refcnt == 0) |
| 1287 | kfree(pc); | 1383 | kfree(pc); |
| 1288 | cpuc->per_core = NULL; | 1384 | cpuc->shared_regs = NULL; |
| 1289 | } | 1385 | } |
| 1290 | 1386 | ||
| 1291 | fini_debug_store_on_cpu(cpu); | 1387 | fini_debug_store_on_cpu(cpu); |
| @@ -1436,7 +1532,6 @@ static __init int intel_pmu_init(void) | |||
| 1436 | 1532 | ||
| 1437 | x86_pmu.event_constraints = intel_nehalem_event_constraints; | 1533 | x86_pmu.event_constraints = intel_nehalem_event_constraints; |
| 1438 | x86_pmu.pebs_constraints = intel_nehalem_pebs_event_constraints; | 1534 | x86_pmu.pebs_constraints = intel_nehalem_pebs_event_constraints; |
| 1439 | x86_pmu.percore_constraints = intel_nehalem_percore_constraints; | ||
| 1440 | x86_pmu.enable_all = intel_pmu_nhm_enable_all; | 1535 | x86_pmu.enable_all = intel_pmu_nhm_enable_all; |
| 1441 | x86_pmu.extra_regs = intel_nehalem_extra_regs; | 1536 | x86_pmu.extra_regs = intel_nehalem_extra_regs; |
| 1442 | 1537 | ||
| @@ -1481,10 +1576,10 @@ static __init int intel_pmu_init(void) | |||
| 1481 | intel_pmu_lbr_init_nhm(); | 1576 | intel_pmu_lbr_init_nhm(); |
| 1482 | 1577 | ||
| 1483 | x86_pmu.event_constraints = intel_westmere_event_constraints; | 1578 | x86_pmu.event_constraints = intel_westmere_event_constraints; |
| 1484 | x86_pmu.percore_constraints = intel_westmere_percore_constraints; | ||
| 1485 | x86_pmu.enable_all = intel_pmu_nhm_enable_all; | 1579 | x86_pmu.enable_all = intel_pmu_nhm_enable_all; |
| 1486 | x86_pmu.pebs_constraints = intel_westmere_pebs_event_constraints; | 1580 | x86_pmu.pebs_constraints = intel_westmere_pebs_event_constraints; |
| 1487 | x86_pmu.extra_regs = intel_westmere_extra_regs; | 1581 | x86_pmu.extra_regs = intel_westmere_extra_regs; |
| 1582 | x86_pmu.er_flags |= ERF_HAS_RSP_1; | ||
| 1488 | 1583 | ||
| 1489 | /* UOPS_ISSUED.STALLED_CYCLES */ | 1584 | /* UOPS_ISSUED.STALLED_CYCLES */ |
| 1490 | intel_perfmon_event_map[PERF_COUNT_HW_STALLED_CYCLES_FRONTEND] = 0x180010e; | 1585 | intel_perfmon_event_map[PERF_COUNT_HW_STALLED_CYCLES_FRONTEND] = 0x180010e; |
| @@ -1502,6 +1597,10 @@ static __init int intel_pmu_init(void) | |||
| 1502 | 1597 | ||
| 1503 | x86_pmu.event_constraints = intel_snb_event_constraints; | 1598 | x86_pmu.event_constraints = intel_snb_event_constraints; |
| 1504 | x86_pmu.pebs_constraints = intel_snb_pebs_events; | 1599 | x86_pmu.pebs_constraints = intel_snb_pebs_events; |
| 1600 | x86_pmu.extra_regs = intel_snb_extra_regs; | ||
| 1601 | /* all extra regs are per-cpu when HT is on */ | ||
| 1602 | x86_pmu.er_flags |= ERF_HAS_RSP_1; | ||
| 1603 | x86_pmu.er_flags |= ERF_NO_HT_SHARING; | ||
| 1505 | 1604 | ||
| 1506 | /* UOPS_ISSUED.ANY,c=1,i=1 to count stall cycles */ | 1605 | /* UOPS_ISSUED.ANY,c=1,i=1 to count stall cycles */ |
| 1507 | intel_perfmon_event_map[PERF_COUNT_HW_STALLED_CYCLES_FRONTEND] = 0x180010e; | 1606 | intel_perfmon_event_map[PERF_COUNT_HW_STALLED_CYCLES_FRONTEND] = 0x180010e; |
| @@ -1512,11 +1611,19 @@ static __init int intel_pmu_init(void) | |||
| 1512 | break; | 1611 | break; |
| 1513 | 1612 | ||
| 1514 | default: | 1613 | default: |
| 1515 | /* | 1614 | switch (x86_pmu.version) { |
| 1516 | * default constraints for v2 and up | 1615 | case 1: |
| 1517 | */ | 1616 | x86_pmu.event_constraints = intel_v1_event_constraints; |
| 1518 | x86_pmu.event_constraints = intel_gen_event_constraints; | 1617 | pr_cont("generic architected perfmon v1, "); |
| 1519 | pr_cont("generic architected perfmon, "); | 1618 | break; |
| 1619 | default: | ||
| 1620 | /* | ||
| 1621 | * default constraints for v2 and up | ||
| 1622 | */ | ||
| 1623 | x86_pmu.event_constraints = intel_gen_event_constraints; | ||
| 1624 | pr_cont("generic architected perfmon, "); | ||
| 1625 | break; | ||
| 1626 | } | ||
| 1520 | } | 1627 | } |
| 1521 | return 0; | 1628 | return 0; |
| 1522 | } | 1629 | } |
| @@ -1528,4 +1635,8 @@ static int intel_pmu_init(void) | |||
| 1528 | return 0; | 1635 | return 0; |
| 1529 | } | 1636 | } |
| 1530 | 1637 | ||
| 1638 | static struct intel_shared_regs *allocate_shared_regs(int cpu) | ||
| 1639 | { | ||
| 1640 | return NULL; | ||
| 1641 | } | ||
| 1531 | #endif /* CONFIG_CPU_SUP_INTEL */ | 1642 | #endif /* CONFIG_CPU_SUP_INTEL */ |
diff --git a/arch/x86/kernel/cpu/perf_event_intel_ds.c b/arch/x86/kernel/cpu/perf_event_intel_ds.c index bab491b8ee2..1b1ef3addcf 100644 --- a/arch/x86/kernel/cpu/perf_event_intel_ds.c +++ b/arch/x86/kernel/cpu/perf_event_intel_ds.c | |||
| @@ -340,7 +340,7 @@ static int intel_pmu_drain_bts_buffer(void) | |||
| 340 | */ | 340 | */ |
| 341 | perf_prepare_sample(&header, &data, event, ®s); | 341 | perf_prepare_sample(&header, &data, event, ®s); |
| 342 | 342 | ||
| 343 | if (perf_output_begin(&handle, event, header.size * (top - at), 1, 1)) | 343 | if (perf_output_begin(&handle, event, header.size * (top - at))) |
| 344 | return 1; | 344 | return 1; |
| 345 | 345 | ||
| 346 | for (; at < top; at++) { | 346 | for (; at < top; at++) { |
| @@ -616,7 +616,7 @@ static void __intel_pmu_pebs_event(struct perf_event *event, | |||
| 616 | else | 616 | else |
| 617 | regs.flags &= ~PERF_EFLAGS_EXACT; | 617 | regs.flags &= ~PERF_EFLAGS_EXACT; |
| 618 | 618 | ||
| 619 | if (perf_event_overflow(event, 1, &data, ®s)) | 619 | if (perf_event_overflow(event, &data, ®s)) |
| 620 | x86_pmu_stop(event, 0); | 620 | x86_pmu_stop(event, 0); |
| 621 | } | 621 | } |
| 622 | 622 | ||
diff --git a/arch/x86/kernel/cpu/perf_event_p4.c b/arch/x86/kernel/cpu/perf_event_p4.c index ead584fb6a7..7809d2bcb20 100644 --- a/arch/x86/kernel/cpu/perf_event_p4.c +++ b/arch/x86/kernel/cpu/perf_event_p4.c | |||
| @@ -554,13 +554,102 @@ static __initconst const u64 p4_hw_cache_event_ids | |||
| 554 | [ C(RESULT_MISS) ] = -1, | 554 | [ C(RESULT_MISS) ] = -1, |
| 555 | }, | 555 | }, |
| 556 | }, | 556 | }, |
| 557 | [ C(NODE) ] = { | ||
| 558 | [ C(OP_READ) ] = { | ||
| 559 | [ C(RESULT_ACCESS) ] = -1, | ||
| 560 | [ C(RESULT_MISS) ] = -1, | ||
| 561 | }, | ||
| 562 | [ C(OP_WRITE) ] = { | ||
| 563 | [ C(RESULT_ACCESS) ] = -1, | ||
| 564 | [ C(RESULT_MISS) ] = -1, | ||
| 565 | }, | ||
| 566 | [ C(OP_PREFETCH) ] = { | ||
| 567 | [ C(RESULT_ACCESS) ] = -1, | ||
| 568 | [ C(RESULT_MISS) ] = -1, | ||
| 569 | }, | ||
| 570 | }, | ||
| 557 | }; | 571 | }; |
| 558 | 572 | ||
| 573 | /* | ||
| 574 | * Because of Netburst being quite restricted in how many | ||
| 575 | * identical events may run simultaneously, we introduce event aliases, | ||
| 576 | * ie the different events which have the same functionality but | ||
| 577 | * utilize non-intersected resources (ESCR/CCCR/counter registers). | ||
| 578 | * | ||
| 579 | * This allow us to relax restrictions a bit and run two or more | ||
| 580 | * identical events together. | ||
| 581 | * | ||
| 582 | * Never set any custom internal bits such as P4_CONFIG_HT, | ||
| 583 | * P4_CONFIG_ALIASABLE or bits for P4_PEBS_METRIC, they are | ||
| 584 | * either up to date automatically or not applicable at all. | ||
| 585 | */ | ||
| 586 | struct p4_event_alias { | ||
| 587 | u64 original; | ||
| 588 | u64 alternative; | ||
| 589 | } p4_event_aliases[] = { | ||
| 590 | { | ||
| 591 | /* | ||
| 592 | * Non-halted cycles can be substituted with non-sleeping cycles (see | ||
| 593 | * Intel SDM Vol3b for details). We need this alias to be able | ||
| 594 | * to run nmi-watchdog and 'perf top' (or any other user space tool | ||
| 595 | * which is interested in running PERF_COUNT_HW_CPU_CYCLES) | ||
| 596 | * simultaneously. | ||
| 597 | */ | ||
| 598 | .original = | ||
| 599 | p4_config_pack_escr(P4_ESCR_EVENT(P4_EVENT_GLOBAL_POWER_EVENTS) | | ||
| 600 | P4_ESCR_EMASK_BIT(P4_EVENT_GLOBAL_POWER_EVENTS, RUNNING)), | ||
| 601 | .alternative = | ||
| 602 | p4_config_pack_escr(P4_ESCR_EVENT(P4_EVENT_EXECUTION_EVENT) | | ||
| 603 | P4_ESCR_EMASK_BIT(P4_EVENT_EXECUTION_EVENT, NBOGUS0)| | ||
| 604 | P4_ESCR_EMASK_BIT(P4_EVENT_EXECUTION_EVENT, NBOGUS1)| | ||
| 605 | P4_ESCR_EMASK_BIT(P4_EVENT_EXECUTION_EVENT, NBOGUS2)| | ||
| 606 | P4_ESCR_EMASK_BIT(P4_EVENT_EXECUTION_EVENT, NBOGUS3)| | ||
| 607 | P4_ESCR_EMASK_BIT(P4_EVENT_EXECUTION_EVENT, BOGUS0) | | ||
| 608 | P4_ESCR_EMASK_BIT(P4_EVENT_EXECUTION_EVENT, BOGUS1) | | ||
| 609 | P4_ESCR_EMASK_BIT(P4_EVENT_EXECUTION_EVENT, BOGUS2) | | ||
| 610 | P4_ESCR_EMASK_BIT(P4_EVENT_EXECUTION_EVENT, BOGUS3))| | ||
| 611 | p4_config_pack_cccr(P4_CCCR_THRESHOLD(15) | P4_CCCR_COMPLEMENT | | ||
| 612 | P4_CCCR_COMPARE), | ||
| 613 | }, | ||
| 614 | }; | ||
| 615 | |||
| 616 | static u64 p4_get_alias_event(u64 config) | ||
| 617 | { | ||
| 618 | u64 config_match; | ||
| 619 | int i; | ||
| 620 | |||
| 621 | /* | ||
| 622 | * Only event with special mark is allowed, | ||
| 623 | * we're to be sure it didn't come as malformed | ||
| 624 | * RAW event. | ||
| 625 | */ | ||
| 626 | if (!(config & P4_CONFIG_ALIASABLE)) | ||
| 627 | return 0; | ||
| 628 | |||
| 629 | config_match = config & P4_CONFIG_EVENT_ALIAS_MASK; | ||
| 630 | |||
| 631 | for (i = 0; i < ARRAY_SIZE(p4_event_aliases); i++) { | ||
| 632 | if (config_match == p4_event_aliases[i].original) { | ||
| 633 | config_match = p4_event_aliases[i].alternative; | ||
| 634 | break; | ||
| 635 | } else if (config_match == p4_event_aliases[i].alternative) { | ||
| 636 | config_match = p4_event_aliases[i].original; | ||
| 637 | break; | ||
| 638 | } | ||
| 639 | } | ||
| 640 | |||
| 641 | if (i >= ARRAY_SIZE(p4_event_aliases)) | ||
| 642 | return 0; | ||
| 643 | |||
| 644 | return config_match | (config & P4_CONFIG_EVENT_ALIAS_IMMUTABLE_BITS); | ||
| 645 | } | ||
| 646 | |||
| 559 | static u64 p4_general_events[PERF_COUNT_HW_MAX] = { | 647 | static u64 p4_general_events[PERF_COUNT_HW_MAX] = { |
| 560 | /* non-halted CPU clocks */ | 648 | /* non-halted CPU clocks */ |
| 561 | [PERF_COUNT_HW_CPU_CYCLES] = | 649 | [PERF_COUNT_HW_CPU_CYCLES] = |
| 562 | p4_config_pack_escr(P4_ESCR_EVENT(P4_EVENT_GLOBAL_POWER_EVENTS) | | 650 | p4_config_pack_escr(P4_ESCR_EVENT(P4_EVENT_GLOBAL_POWER_EVENTS) | |
| 563 | P4_ESCR_EMASK_BIT(P4_EVENT_GLOBAL_POWER_EVENTS, RUNNING)), | 651 | P4_ESCR_EMASK_BIT(P4_EVENT_GLOBAL_POWER_EVENTS, RUNNING)) | |
| 652 | P4_CONFIG_ALIASABLE, | ||
| 564 | 653 | ||
| 565 | /* | 654 | /* |
| 566 | * retired instructions | 655 | * retired instructions |
| @@ -945,7 +1034,7 @@ static int p4_pmu_handle_irq(struct pt_regs *regs) | |||
| 945 | 1034 | ||
| 946 | if (!x86_perf_event_set_period(event)) | 1035 | if (!x86_perf_event_set_period(event)) |
| 947 | continue; | 1036 | continue; |
| 948 | if (perf_event_overflow(event, 1, &data, regs)) | 1037 | if (perf_event_overflow(event, &data, regs)) |
| 949 | x86_pmu_stop(event, 0); | 1038 | x86_pmu_stop(event, 0); |
| 950 | } | 1039 | } |
| 951 | 1040 | ||
| @@ -1120,6 +1209,8 @@ static int p4_pmu_schedule_events(struct cpu_hw_events *cpuc, int n, int *assign | |||
| 1120 | struct p4_event_bind *bind; | 1209 | struct p4_event_bind *bind; |
| 1121 | unsigned int i, thread, num; | 1210 | unsigned int i, thread, num; |
| 1122 | int cntr_idx, escr_idx; | 1211 | int cntr_idx, escr_idx; |
| 1212 | u64 config_alias; | ||
| 1213 | int pass; | ||
| 1123 | 1214 | ||
| 1124 | bitmap_zero(used_mask, X86_PMC_IDX_MAX); | 1215 | bitmap_zero(used_mask, X86_PMC_IDX_MAX); |
| 1125 | bitmap_zero(escr_mask, P4_ESCR_MSR_TABLE_SIZE); | 1216 | bitmap_zero(escr_mask, P4_ESCR_MSR_TABLE_SIZE); |
| @@ -1128,6 +1219,17 @@ static int p4_pmu_schedule_events(struct cpu_hw_events *cpuc, int n, int *assign | |||
| 1128 | 1219 | ||
| 1129 | hwc = &cpuc->event_list[i]->hw; | 1220 | hwc = &cpuc->event_list[i]->hw; |
| 1130 | thread = p4_ht_thread(cpu); | 1221 | thread = p4_ht_thread(cpu); |
| 1222 | pass = 0; | ||
| 1223 | |||
| 1224 | again: | ||
| 1225 | /* | ||
| 1226 | * It's possible to hit a circular lock | ||
| 1227 | * between original and alternative events | ||
| 1228 | * if both are scheduled already. | ||
| 1229 | */ | ||
| 1230 | if (pass > 2) | ||
| 1231 | goto done; | ||
| 1232 | |||
| 1131 | bind = p4_config_get_bind(hwc->config); | 1233 | bind = p4_config_get_bind(hwc->config); |
| 1132 | escr_idx = p4_get_escr_idx(bind->escr_msr[thread]); | 1234 | escr_idx = p4_get_escr_idx(bind->escr_msr[thread]); |
| 1133 | if (unlikely(escr_idx == -1)) | 1235 | if (unlikely(escr_idx == -1)) |
| @@ -1141,8 +1243,17 @@ static int p4_pmu_schedule_events(struct cpu_hw_events *cpuc, int n, int *assign | |||
| 1141 | } | 1243 | } |
| 1142 | 1244 | ||
| 1143 | cntr_idx = p4_next_cntr(thread, used_mask, bind); | 1245 | cntr_idx = p4_next_cntr(thread, used_mask, bind); |
| 1144 | if (cntr_idx == -1 || test_bit(escr_idx, escr_mask)) | 1246 | if (cntr_idx == -1 || test_bit(escr_idx, escr_mask)) { |
| 1145 | goto done; | 1247 | /* |
| 1248 | * Check whether an event alias is still available. | ||
| 1249 | */ | ||
| 1250 | config_alias = p4_get_alias_event(hwc->config); | ||
| 1251 | if (!config_alias) | ||
| 1252 | goto done; | ||
| 1253 | hwc->config = config_alias; | ||
| 1254 | pass++; | ||
| 1255 | goto again; | ||
| 1256 | } | ||
| 1146 | 1257 | ||
| 1147 | p4_pmu_swap_config_ts(hwc, cpu); | 1258 | p4_pmu_swap_config_ts(hwc, cpu); |
| 1148 | if (assign) | 1259 | if (assign) |
diff --git a/arch/x86/kernel/dumpstack_64.c b/arch/x86/kernel/dumpstack_64.c index e71c98d3c0d..19853ad8afc 100644 --- a/arch/x86/kernel/dumpstack_64.c +++ b/arch/x86/kernel/dumpstack_64.c | |||
| @@ -105,34 +105,6 @@ in_irq_stack(unsigned long *stack, unsigned long *irq_stack, | |||
| 105 | } | 105 | } |
| 106 | 106 | ||
| 107 | /* | 107 | /* |
| 108 | * We are returning from the irq stack and go to the previous one. | ||
| 109 | * If the previous stack is also in the irq stack, then bp in the first | ||
| 110 | * frame of the irq stack points to the previous, interrupted one. | ||
| 111 | * Otherwise we have another level of indirection: We first save | ||
| 112 | * the bp of the previous stack, then we switch the stack to the irq one | ||
| 113 | * and save a new bp that links to the previous one. | ||
| 114 | * (See save_args()) | ||
| 115 | */ | ||
| 116 | static inline unsigned long | ||
| 117 | fixup_bp_irq_link(unsigned long bp, unsigned long *stack, | ||
| 118 | unsigned long *irq_stack, unsigned long *irq_stack_end) | ||
| 119 | { | ||
| 120 | #ifdef CONFIG_FRAME_POINTER | ||
| 121 | struct stack_frame *frame = (struct stack_frame *)bp; | ||
| 122 | unsigned long next; | ||
| 123 | |||
| 124 | if (!in_irq_stack(stack, irq_stack, irq_stack_end)) { | ||
| 125 | if (!probe_kernel_address(&frame->next_frame, next)) | ||
| 126 | return next; | ||
| 127 | else | ||
| 128 | WARN_ONCE(1, "Perf: bad frame pointer = %p in " | ||
| 129 | "callchain\n", &frame->next_frame); | ||
| 130 | } | ||
| 131 | #endif | ||
| 132 | return bp; | ||
| 133 | } | ||
| 134 | |||
| 135 | /* | ||
| 136 | * x86-64 can have up to three kernel stacks: | 108 | * x86-64 can have up to three kernel stacks: |
| 137 | * process stack | 109 | * process stack |
| 138 | * interrupt stack | 110 | * interrupt stack |
| @@ -155,9 +127,12 @@ void dump_trace(struct task_struct *task, struct pt_regs *regs, | |||
| 155 | task = current; | 127 | task = current; |
| 156 | 128 | ||
| 157 | if (!stack) { | 129 | if (!stack) { |
| 158 | stack = &dummy; | 130 | if (regs) |
| 159 | if (task && task != current) | 131 | stack = (unsigned long *)regs->sp; |
| 132 | else if (task && task != current) | ||
| 160 | stack = (unsigned long *)task->thread.sp; | 133 | stack = (unsigned long *)task->thread.sp; |
| 134 | else | ||
| 135 | stack = &dummy; | ||
| 161 | } | 136 | } |
| 162 | 137 | ||
| 163 | if (!bp) | 138 | if (!bp) |
| @@ -205,8 +180,6 @@ void dump_trace(struct task_struct *task, struct pt_regs *regs, | |||
| 205 | * pointer (index -1 to end) in the IRQ stack: | 180 | * pointer (index -1 to end) in the IRQ stack: |
| 206 | */ | 181 | */ |
| 207 | stack = (unsigned long *) (irq_stack_end[-1]); | 182 | stack = (unsigned long *) (irq_stack_end[-1]); |
| 208 | bp = fixup_bp_irq_link(bp, stack, irq_stack, | ||
| 209 | irq_stack_end); | ||
| 210 | irq_stack_end = NULL; | 183 | irq_stack_end = NULL; |
| 211 | ops->stack(data, "EOI"); | 184 | ops->stack(data, "EOI"); |
| 212 | continue; | 185 | continue; |
diff --git a/arch/x86/kernel/entry_64.S b/arch/x86/kernel/entry_64.S index 8a445a0c989..d656f68371a 100644 --- a/arch/x86/kernel/entry_64.S +++ b/arch/x86/kernel/entry_64.S | |||
| @@ -297,27 +297,26 @@ ENDPROC(native_usergs_sysret64) | |||
| 297 | .endm | 297 | .endm |
| 298 | 298 | ||
| 299 | /* save partial stack frame */ | 299 | /* save partial stack frame */ |
| 300 | .pushsection .kprobes.text, "ax" | 300 | .macro SAVE_ARGS_IRQ |
| 301 | ENTRY(save_args) | ||
| 302 | XCPT_FRAME | ||
| 303 | cld | 301 | cld |
| 304 | /* | 302 | /* start from rbp in pt_regs and jump over */ |
| 305 | * start from rbp in pt_regs and jump over | 303 | movq_cfi rdi, RDI-RBP |
| 306 | * return address. | 304 | movq_cfi rsi, RSI-RBP |
| 307 | */ | 305 | movq_cfi rdx, RDX-RBP |
| 308 | movq_cfi rdi, RDI+8-RBP | 306 | movq_cfi rcx, RCX-RBP |
| 309 | movq_cfi rsi, RSI+8-RBP | 307 | movq_cfi rax, RAX-RBP |
| 310 | movq_cfi rdx, RDX+8-RBP | 308 | movq_cfi r8, R8-RBP |
| 311 | movq_cfi rcx, RCX+8-RBP | 309 | movq_cfi r9, R9-RBP |
| 312 | movq_cfi rax, RAX+8-RBP | 310 | movq_cfi r10, R10-RBP |
| 313 | movq_cfi r8, R8+8-RBP | 311 | movq_cfi r11, R11-RBP |
| 314 | movq_cfi r9, R9+8-RBP | 312 | |
| 315 | movq_cfi r10, R10+8-RBP | 313 | /* Save rbp so that we can unwind from get_irq_regs() */ |
| 316 | movq_cfi r11, R11+8-RBP | 314 | movq_cfi rbp, 0 |
| 317 | 315 | ||
| 318 | leaq -RBP+8(%rsp),%rdi /* arg1 for handler */ | 316 | /* Save previous stack value */ |
| 319 | movq_cfi rbp, 8 /* push %rbp */ | 317 | movq %rsp, %rsi |
| 320 | leaq 8(%rsp), %rbp /* mov %rsp, %ebp */ | 318 | |
| 319 | leaq -RBP(%rsp),%rdi /* arg1 for handler */ | ||
| 321 | testl $3, CS(%rdi) | 320 | testl $3, CS(%rdi) |
| 322 | je 1f | 321 | je 1f |
| 323 | SWAPGS | 322 | SWAPGS |
| @@ -329,19 +328,14 @@ ENTRY(save_args) | |||
| 329 | */ | 328 | */ |
| 330 | 1: incl PER_CPU_VAR(irq_count) | 329 | 1: incl PER_CPU_VAR(irq_count) |
| 331 | jne 2f | 330 | jne 2f |
| 332 | popq_cfi %rax /* move return address... */ | ||
| 333 | mov PER_CPU_VAR(irq_stack_ptr),%rsp | 331 | mov PER_CPU_VAR(irq_stack_ptr),%rsp |
| 334 | EMPTY_FRAME 0 | 332 | EMPTY_FRAME 0 |
| 335 | pushq_cfi %rbp /* backlink for unwinder */ | 333 | |
| 336 | pushq_cfi %rax /* ... to the new stack */ | 334 | 2: /* Store previous stack value */ |
| 337 | /* | 335 | pushq %rsi |
| 338 | * We entered an interrupt context - irqs are off: | 336 | /* We entered an interrupt context - irqs are off: */ |
| 339 | */ | 337 | TRACE_IRQS_OFF |
| 340 | 2: TRACE_IRQS_OFF | 338 | .endm |
| 341 | ret | ||
| 342 | CFI_ENDPROC | ||
| 343 | END(save_args) | ||
| 344 | .popsection | ||
| 345 | 339 | ||
| 346 | ENTRY(save_rest) | 340 | ENTRY(save_rest) |
| 347 | PARTIAL_FRAME 1 REST_SKIP+8 | 341 | PARTIAL_FRAME 1 REST_SKIP+8 |
| @@ -791,7 +785,7 @@ END(interrupt) | |||
| 791 | /* reserve pt_regs for scratch regs and rbp */ | 785 | /* reserve pt_regs for scratch regs and rbp */ |
| 792 | subq $ORIG_RAX-RBP, %rsp | 786 | subq $ORIG_RAX-RBP, %rsp |
| 793 | CFI_ADJUST_CFA_OFFSET ORIG_RAX-RBP | 787 | CFI_ADJUST_CFA_OFFSET ORIG_RAX-RBP |
| 794 | call save_args | 788 | SAVE_ARGS_IRQ |
| 795 | PARTIAL_FRAME 0 | 789 | PARTIAL_FRAME 0 |
| 796 | call \func | 790 | call \func |
| 797 | .endm | 791 | .endm |
| @@ -814,15 +808,14 @@ ret_from_intr: | |||
| 814 | DISABLE_INTERRUPTS(CLBR_NONE) | 808 | DISABLE_INTERRUPTS(CLBR_NONE) |
| 815 | TRACE_IRQS_OFF | 809 | TRACE_IRQS_OFF |
| 816 | decl PER_CPU_VAR(irq_count) | 810 | decl PER_CPU_VAR(irq_count) |
| 817 | leaveq | ||
| 818 | 811 | ||
| 819 | CFI_RESTORE rbp | 812 | /* Restore saved previous stack */ |
| 813 | popq %rsi | ||
| 814 | leaq 16(%rsi), %rsp | ||
| 815 | |||
| 820 | CFI_DEF_CFA_REGISTER rsp | 816 | CFI_DEF_CFA_REGISTER rsp |
| 821 | CFI_ADJUST_CFA_OFFSET -8 | 817 | CFI_ADJUST_CFA_OFFSET -16 |
| 822 | 818 | ||
| 823 | /* we did not save rbx, restore only from ARGOFFSET */ | ||
| 824 | addq $8, %rsp | ||
| 825 | CFI_ADJUST_CFA_OFFSET -8 | ||
| 826 | exit_intr: | 819 | exit_intr: |
| 827 | GET_THREAD_INFO(%rcx) | 820 | GET_THREAD_INFO(%rcx) |
| 828 | testl $3,CS-ARGOFFSET(%rsp) | 821 | testl $3,CS-ARGOFFSET(%rsp) |
diff --git a/arch/x86/kernel/kgdb.c b/arch/x86/kernel/kgdb.c index 5f9ecff328b..00354d4919a 100644 --- a/arch/x86/kernel/kgdb.c +++ b/arch/x86/kernel/kgdb.c | |||
| @@ -608,7 +608,7 @@ int kgdb_arch_init(void) | |||
| 608 | return register_die_notifier(&kgdb_notifier); | 608 | return register_die_notifier(&kgdb_notifier); |
| 609 | } | 609 | } |
| 610 | 610 | ||
| 611 | static void kgdb_hw_overflow_handler(struct perf_event *event, int nmi, | 611 | static void kgdb_hw_overflow_handler(struct perf_event *event, |
| 612 | struct perf_sample_data *data, struct pt_regs *regs) | 612 | struct perf_sample_data *data, struct pt_regs *regs) |
| 613 | { | 613 | { |
| 614 | struct task_struct *tsk = current; | 614 | struct task_struct *tsk = current; |
| @@ -638,7 +638,7 @@ void kgdb_arch_late(void) | |||
| 638 | for (i = 0; i < HBP_NUM; i++) { | 638 | for (i = 0; i < HBP_NUM; i++) { |
| 639 | if (breakinfo[i].pev) | 639 | if (breakinfo[i].pev) |
| 640 | continue; | 640 | continue; |
| 641 | breakinfo[i].pev = register_wide_hw_breakpoint(&attr, NULL); | 641 | breakinfo[i].pev = register_wide_hw_breakpoint(&attr, NULL, NULL); |
| 642 | if (IS_ERR((void * __force)breakinfo[i].pev)) { | 642 | if (IS_ERR((void * __force)breakinfo[i].pev)) { |
| 643 | printk(KERN_ERR "kgdb: Could not allocate hw" | 643 | printk(KERN_ERR "kgdb: Could not allocate hw" |
| 644 | "breakpoints\nDisabling the kernel debugger\n"); | 644 | "breakpoints\nDisabling the kernel debugger\n"); |
diff --git a/arch/x86/kernel/ptrace.c b/arch/x86/kernel/ptrace.c index 807c2a2b80f..82528799c5d 100644 --- a/arch/x86/kernel/ptrace.c +++ b/arch/x86/kernel/ptrace.c | |||
| @@ -528,7 +528,7 @@ static int genregs_set(struct task_struct *target, | |||
| 528 | return ret; | 528 | return ret; |
| 529 | } | 529 | } |
| 530 | 530 | ||
| 531 | static void ptrace_triggered(struct perf_event *bp, int nmi, | 531 | static void ptrace_triggered(struct perf_event *bp, |
| 532 | struct perf_sample_data *data, | 532 | struct perf_sample_data *data, |
| 533 | struct pt_regs *regs) | 533 | struct pt_regs *regs) |
| 534 | { | 534 | { |
| @@ -715,7 +715,8 @@ static int ptrace_set_breakpoint_addr(struct task_struct *tsk, int nr, | |||
| 715 | attr.bp_type = HW_BREAKPOINT_W; | 715 | attr.bp_type = HW_BREAKPOINT_W; |
| 716 | attr.disabled = 1; | 716 | attr.disabled = 1; |
| 717 | 717 | ||
| 718 | bp = register_user_hw_breakpoint(&attr, ptrace_triggered, tsk); | 718 | bp = register_user_hw_breakpoint(&attr, ptrace_triggered, |
| 719 | NULL, tsk); | ||
| 719 | 720 | ||
| 720 | /* | 721 | /* |
| 721 | * CHECKME: the previous code returned -EIO if the addr wasn't | 722 | * CHECKME: the previous code returned -EIO if the addr wasn't |
diff --git a/arch/x86/kernel/stacktrace.c b/arch/x86/kernel/stacktrace.c index 55d9bc03f69..fdd0c6430e5 100644 --- a/arch/x86/kernel/stacktrace.c +++ b/arch/x86/kernel/stacktrace.c | |||
| @@ -66,7 +66,7 @@ void save_stack_trace(struct stack_trace *trace) | |||
| 66 | } | 66 | } |
| 67 | EXPORT_SYMBOL_GPL(save_stack_trace); | 67 | EXPORT_SYMBOL_GPL(save_stack_trace); |
| 68 | 68 | ||
| 69 | void save_stack_trace_regs(struct stack_trace *trace, struct pt_regs *regs) | 69 | void save_stack_trace_regs(struct pt_regs *regs, struct stack_trace *trace) |
| 70 | { | 70 | { |
| 71 | dump_trace(current, regs, NULL, 0, &save_stack_ops, trace); | 71 | dump_trace(current, regs, NULL, 0, &save_stack_ops, trace); |
| 72 | if (trace->nr_entries < trace->max_entries) | 72 | if (trace->nr_entries < trace->max_entries) |
diff --git a/arch/x86/lib/Makefile b/arch/x86/lib/Makefile index f2479f19ddd..6ba477342b8 100644 --- a/arch/x86/lib/Makefile +++ b/arch/x86/lib/Makefile | |||
| @@ -18,7 +18,7 @@ obj-$(CONFIG_SMP) += msr-smp.o cache-smp.o | |||
| 18 | 18 | ||
| 19 | lib-y := delay.o | 19 | lib-y := delay.o |
| 20 | lib-y += thunk_$(BITS).o | 20 | lib-y += thunk_$(BITS).o |
| 21 | lib-y += usercopy_$(BITS).o getuser.o putuser.o | 21 | lib-y += usercopy_$(BITS).o usercopy.o getuser.o putuser.o |
| 22 | lib-y += memcpy_$(BITS).o | 22 | lib-y += memcpy_$(BITS).o |
| 23 | lib-$(CONFIG_INSTRUCTION_DECODER) += insn.o inat.o | 23 | lib-$(CONFIG_INSTRUCTION_DECODER) += insn.o inat.o |
| 24 | 24 | ||
diff --git a/arch/x86/lib/usercopy.c b/arch/x86/lib/usercopy.c new file mode 100644 index 00000000000..97be9cb5448 --- /dev/null +++ b/arch/x86/lib/usercopy.c | |||
| @@ -0,0 +1,43 @@ | |||
| 1 | /* | ||
| 2 | * User address space access functions. | ||
| 3 | * | ||
| 4 | * For licencing details see kernel-base/COPYING | ||
| 5 | */ | ||
| 6 | |||
| 7 | #include <linux/highmem.h> | ||
| 8 | #include <linux/module.h> | ||
| 9 | |||
| 10 | /* | ||
| 11 | * best effort, GUP based copy_from_user() that is NMI-safe | ||
| 12 | */ | ||
| 13 | unsigned long | ||
| 14 | copy_from_user_nmi(void *to, const void __user *from, unsigned long n) | ||
| 15 | { | ||
| 16 | unsigned long offset, addr = (unsigned long)from; | ||
| 17 | unsigned long size, len = 0; | ||
| 18 | struct page *page; | ||
| 19 | void *map; | ||
| 20 | int ret; | ||
| 21 | |||
| 22 | do { | ||
| 23 | ret = __get_user_pages_fast(addr, 1, 0, &page); | ||
| 24 | if (!ret) | ||
| 25 | break; | ||
| 26 | |||
| 27 | offset = addr & (PAGE_SIZE - 1); | ||
| 28 | size = min(PAGE_SIZE - offset, n - len); | ||
| 29 | |||
| 30 | map = kmap_atomic(page); | ||
| 31 | memcpy(to, map+offset, size); | ||
| 32 | kunmap_atomic(map); | ||
| 33 | put_page(page); | ||
| 34 | |||
| 35 | len += size; | ||
| 36 | to += size; | ||
| 37 | addr += size; | ||
| 38 | |||
| 39 | } while (len < n); | ||
| 40 | |||
| 41 | return len; | ||
| 42 | } | ||
| 43 | EXPORT_SYMBOL_GPL(copy_from_user_nmi); | ||
diff --git a/arch/x86/mm/fault.c b/arch/x86/mm/fault.c index 2dbf6bf4c7e..4d09df054e3 100644 --- a/arch/x86/mm/fault.c +++ b/arch/x86/mm/fault.c | |||
| @@ -1059,7 +1059,7 @@ do_page_fault(struct pt_regs *regs, unsigned long error_code) | |||
| 1059 | if (unlikely(error_code & PF_RSVD)) | 1059 | if (unlikely(error_code & PF_RSVD)) |
| 1060 | pgtable_bad(regs, error_code, address); | 1060 | pgtable_bad(regs, error_code, address); |
| 1061 | 1061 | ||
| 1062 | perf_sw_event(PERF_COUNT_SW_PAGE_FAULTS, 1, 0, regs, address); | 1062 | perf_sw_event(PERF_COUNT_SW_PAGE_FAULTS, 1, regs, address); |
| 1063 | 1063 | ||
| 1064 | /* | 1064 | /* |
| 1065 | * If we're in an interrupt, have no user context or are running | 1065 | * If we're in an interrupt, have no user context or are running |
| @@ -1161,11 +1161,11 @@ good_area: | |||
| 1161 | if (flags & FAULT_FLAG_ALLOW_RETRY) { | 1161 | if (flags & FAULT_FLAG_ALLOW_RETRY) { |
| 1162 | if (fault & VM_FAULT_MAJOR) { | 1162 | if (fault & VM_FAULT_MAJOR) { |
| 1163 | tsk->maj_flt++; | 1163 | tsk->maj_flt++; |
| 1164 | perf_sw_event(PERF_COUNT_SW_PAGE_FAULTS_MAJ, 1, 0, | 1164 | perf_sw_event(PERF_COUNT_SW_PAGE_FAULTS_MAJ, 1, |
| 1165 | regs, address); | 1165 | regs, address); |
| 1166 | } else { | 1166 | } else { |
| 1167 | tsk->min_flt++; | 1167 | tsk->min_flt++; |
| 1168 | perf_sw_event(PERF_COUNT_SW_PAGE_FAULTS_MIN, 1, 0, | 1168 | perf_sw_event(PERF_COUNT_SW_PAGE_FAULTS_MIN, 1, |
| 1169 | regs, address); | 1169 | regs, address); |
| 1170 | } | 1170 | } |
| 1171 | if (fault & VM_FAULT_RETRY) { | 1171 | if (fault & VM_FAULT_RETRY) { |
diff --git a/arch/x86/mm/kmemcheck/error.c b/arch/x86/mm/kmemcheck/error.c index 704a37ceddd..dab41876cdd 100644 --- a/arch/x86/mm/kmemcheck/error.c +++ b/arch/x86/mm/kmemcheck/error.c | |||
| @@ -185,7 +185,7 @@ void kmemcheck_error_save(enum kmemcheck_shadow state, | |||
| 185 | e->trace.entries = e->trace_entries; | 185 | e->trace.entries = e->trace_entries; |
| 186 | e->trace.max_entries = ARRAY_SIZE(e->trace_entries); | 186 | e->trace.max_entries = ARRAY_SIZE(e->trace_entries); |
| 187 | e->trace.skip = 0; | 187 | e->trace.skip = 0; |
| 188 | save_stack_trace_regs(&e->trace, regs); | 188 | save_stack_trace_regs(regs, &e->trace); |
| 189 | 189 | ||
| 190 | /* Round address down to nearest 16 bytes */ | 190 | /* Round address down to nearest 16 bytes */ |
| 191 | shadow_copy = kmemcheck_shadow_lookup(address | 191 | shadow_copy = kmemcheck_shadow_lookup(address |
diff --git a/arch/x86/oprofile/backtrace.c b/arch/x86/oprofile/backtrace.c index a5b64ab4cd6..bff89dfe361 100644 --- a/arch/x86/oprofile/backtrace.c +++ b/arch/x86/oprofile/backtrace.c | |||
| @@ -11,10 +11,11 @@ | |||
| 11 | #include <linux/oprofile.h> | 11 | #include <linux/oprofile.h> |
| 12 | #include <linux/sched.h> | 12 | #include <linux/sched.h> |
| 13 | #include <linux/mm.h> | 13 | #include <linux/mm.h> |
| 14 | #include <linux/compat.h> | ||
| 15 | #include <linux/uaccess.h> | ||
| 16 | |||
| 14 | #include <asm/ptrace.h> | 17 | #include <asm/ptrace.h> |
| 15 | #include <asm/uaccess.h> | ||
| 16 | #include <asm/stacktrace.h> | 18 | #include <asm/stacktrace.h> |
| 17 | #include <linux/compat.h> | ||
| 18 | 19 | ||
| 19 | static int backtrace_stack(void *data, char *name) | 20 | static int backtrace_stack(void *data, char *name) |
| 20 | { | 21 | { |
| @@ -40,13 +41,13 @@ static struct stacktrace_ops backtrace_ops = { | |||
| 40 | static struct stack_frame_ia32 * | 41 | static struct stack_frame_ia32 * |
| 41 | dump_user_backtrace_32(struct stack_frame_ia32 *head) | 42 | dump_user_backtrace_32(struct stack_frame_ia32 *head) |
| 42 | { | 43 | { |
| 44 | /* Also check accessibility of one struct frame_head beyond: */ | ||
| 43 | struct stack_frame_ia32 bufhead[2]; | 45 | struct stack_frame_ia32 bufhead[2]; |
| 44 | struct stack_frame_ia32 *fp; | 46 | struct stack_frame_ia32 *fp; |
| 47 | unsigned long bytes; | ||
| 45 | 48 | ||
| 46 | /* Also check accessibility of one struct frame_head beyond */ | 49 | bytes = copy_from_user_nmi(bufhead, head, sizeof(bufhead)); |
| 47 | if (!access_ok(VERIFY_READ, head, sizeof(bufhead))) | 50 | if (bytes != sizeof(bufhead)) |
| 48 | return NULL; | ||
| 49 | if (__copy_from_user_inatomic(bufhead, head, sizeof(bufhead))) | ||
| 50 | return NULL; | 51 | return NULL; |
| 51 | 52 | ||
| 52 | fp = (struct stack_frame_ia32 *) compat_ptr(bufhead[0].next_frame); | 53 | fp = (struct stack_frame_ia32 *) compat_ptr(bufhead[0].next_frame); |
| @@ -87,12 +88,12 @@ x86_backtrace_32(struct pt_regs * const regs, unsigned int depth) | |||
| 87 | 88 | ||
| 88 | static struct stack_frame *dump_user_backtrace(struct stack_frame *head) | 89 | static struct stack_frame *dump_user_backtrace(struct stack_frame *head) |
| 89 | { | 90 | { |
| 91 | /* Also check accessibility of one struct frame_head beyond: */ | ||
| 90 | struct stack_frame bufhead[2]; | 92 | struct stack_frame bufhead[2]; |
| 93 | unsigned long bytes; | ||
| 91 | 94 | ||
| 92 | /* Also check accessibility of one struct stack_frame beyond */ | 95 | bytes = copy_from_user_nmi(bufhead, head, sizeof(bufhead)); |
| 93 | if (!access_ok(VERIFY_READ, head, sizeof(bufhead))) | 96 | if (bytes != sizeof(bufhead)) |
| 94 | return NULL; | ||
| 95 | if (__copy_from_user_inatomic(bufhead, head, sizeof(bufhead))) | ||
| 96 | return NULL; | 97 | return NULL; |
| 97 | 98 | ||
| 98 | oprofile_add_trace(bufhead[0].return_address); | 99 | oprofile_add_trace(bufhead[0].return_address); |
diff --git a/drivers/oprofile/oprofile_perf.c b/drivers/oprofile/oprofile_perf.c index 9046f7b2ed7..94796f39bc4 100644 --- a/drivers/oprofile/oprofile_perf.c +++ b/drivers/oprofile/oprofile_perf.c | |||
| @@ -31,7 +31,7 @@ static int num_counters; | |||
| 31 | /* | 31 | /* |
| 32 | * Overflow callback for oprofile. | 32 | * Overflow callback for oprofile. |
| 33 | */ | 33 | */ |
| 34 | static void op_overflow_handler(struct perf_event *event, int unused, | 34 | static void op_overflow_handler(struct perf_event *event, |
| 35 | struct perf_sample_data *data, struct pt_regs *regs) | 35 | struct perf_sample_data *data, struct pt_regs *regs) |
| 36 | { | 36 | { |
| 37 | int id; | 37 | int id; |
| @@ -79,7 +79,7 @@ static int op_create_counter(int cpu, int event) | |||
| 79 | 79 | ||
| 80 | pevent = perf_event_create_kernel_counter(&counter_config[event].attr, | 80 | pevent = perf_event_create_kernel_counter(&counter_config[event].attr, |
| 81 | cpu, NULL, | 81 | cpu, NULL, |
| 82 | op_overflow_handler); | 82 | op_overflow_handler, NULL); |
| 83 | 83 | ||
| 84 | if (IS_ERR(pevent)) | 84 | if (IS_ERR(pevent)) |
| 85 | return PTR_ERR(pevent); | 85 | return PTR_ERR(pevent); |
diff --git a/include/linux/ftrace.h b/include/linux/ftrace.h index 9d88e1cb5db..f0c0e8a47ae 100644 --- a/include/linux/ftrace.h +++ b/include/linux/ftrace.h | |||
| @@ -19,6 +19,8 @@ | |||
| 19 | 19 | ||
| 20 | #include <asm/ftrace.h> | 20 | #include <asm/ftrace.h> |
| 21 | 21 | ||
| 22 | struct ftrace_hash; | ||
| 23 | |||
| 22 | #ifdef CONFIG_FUNCTION_TRACER | 24 | #ifdef CONFIG_FUNCTION_TRACER |
| 23 | 25 | ||
| 24 | extern int ftrace_enabled; | 26 | extern int ftrace_enabled; |
| @@ -29,8 +31,6 @@ ftrace_enable_sysctl(struct ctl_table *table, int write, | |||
| 29 | 31 | ||
| 30 | typedef void (*ftrace_func_t)(unsigned long ip, unsigned long parent_ip); | 32 | typedef void (*ftrace_func_t)(unsigned long ip, unsigned long parent_ip); |
| 31 | 33 | ||
| 32 | struct ftrace_hash; | ||
| 33 | |||
| 34 | enum { | 34 | enum { |
| 35 | FTRACE_OPS_FL_ENABLED = 1 << 0, | 35 | FTRACE_OPS_FL_ENABLED = 1 << 0, |
| 36 | FTRACE_OPS_FL_GLOBAL = 1 << 1, | 36 | FTRACE_OPS_FL_GLOBAL = 1 << 1, |
| @@ -123,7 +123,8 @@ stack_trace_sysctl(struct ctl_table *table, int write, | |||
| 123 | struct ftrace_func_command { | 123 | struct ftrace_func_command { |
| 124 | struct list_head list; | 124 | struct list_head list; |
| 125 | char *name; | 125 | char *name; |
| 126 | int (*func)(char *func, char *cmd, | 126 | int (*func)(struct ftrace_hash *hash, |
| 127 | char *func, char *cmd, | ||
| 127 | char *params, int enable); | 128 | char *params, int enable); |
| 128 | }; | 129 | }; |
| 129 | 130 | ||
diff --git a/include/linux/ftrace_event.h b/include/linux/ftrace_event.h index 59d3ef100eb..96efa6794ea 100644 --- a/include/linux/ftrace_event.h +++ b/include/linux/ftrace_event.h | |||
| @@ -76,6 +76,7 @@ struct trace_iterator { | |||
| 76 | struct trace_entry *ent; | 76 | struct trace_entry *ent; |
| 77 | unsigned long lost_events; | 77 | unsigned long lost_events; |
| 78 | int leftover; | 78 | int leftover; |
| 79 | int ent_size; | ||
| 79 | int cpu; | 80 | int cpu; |
| 80 | u64 ts; | 81 | u64 ts; |
| 81 | 82 | ||
| @@ -129,6 +130,10 @@ void trace_current_buffer_unlock_commit(struct ring_buffer *buffer, | |||
| 129 | void trace_nowake_buffer_unlock_commit(struct ring_buffer *buffer, | 130 | void trace_nowake_buffer_unlock_commit(struct ring_buffer *buffer, |
| 130 | struct ring_buffer_event *event, | 131 | struct ring_buffer_event *event, |
| 131 | unsigned long flags, int pc); | 132 | unsigned long flags, int pc); |
| 133 | void trace_nowake_buffer_unlock_commit_regs(struct ring_buffer *buffer, | ||
| 134 | struct ring_buffer_event *event, | ||
| 135 | unsigned long flags, int pc, | ||
| 136 | struct pt_regs *regs); | ||
| 132 | void trace_current_buffer_discard_commit(struct ring_buffer *buffer, | 137 | void trace_current_buffer_discard_commit(struct ring_buffer *buffer, |
| 133 | struct ring_buffer_event *event); | 138 | struct ring_buffer_event *event); |
| 134 | 139 | ||
diff --git a/include/linux/hw_breakpoint.h b/include/linux/hw_breakpoint.h index d1e55fed2c7..6ae9c631a1b 100644 --- a/include/linux/hw_breakpoint.h +++ b/include/linux/hw_breakpoint.h | |||
| @@ -73,6 +73,7 @@ static inline unsigned long hw_breakpoint_len(struct perf_event *bp) | |||
| 73 | extern struct perf_event * | 73 | extern struct perf_event * |
| 74 | register_user_hw_breakpoint(struct perf_event_attr *attr, | 74 | register_user_hw_breakpoint(struct perf_event_attr *attr, |
| 75 | perf_overflow_handler_t triggered, | 75 | perf_overflow_handler_t triggered, |
| 76 | void *context, | ||
| 76 | struct task_struct *tsk); | 77 | struct task_struct *tsk); |
| 77 | 78 | ||
| 78 | /* FIXME: only change from the attr, and don't unregister */ | 79 | /* FIXME: only change from the attr, and don't unregister */ |
| @@ -85,11 +86,13 @@ modify_user_hw_breakpoint(struct perf_event *bp, struct perf_event_attr *attr); | |||
| 85 | extern struct perf_event * | 86 | extern struct perf_event * |
| 86 | register_wide_hw_breakpoint_cpu(struct perf_event_attr *attr, | 87 | register_wide_hw_breakpoint_cpu(struct perf_event_attr *attr, |
| 87 | perf_overflow_handler_t triggered, | 88 | perf_overflow_handler_t triggered, |
| 89 | void *context, | ||
| 88 | int cpu); | 90 | int cpu); |
| 89 | 91 | ||
| 90 | extern struct perf_event * __percpu * | 92 | extern struct perf_event * __percpu * |
| 91 | register_wide_hw_breakpoint(struct perf_event_attr *attr, | 93 | register_wide_hw_breakpoint(struct perf_event_attr *attr, |
| 92 | perf_overflow_handler_t triggered); | 94 | perf_overflow_handler_t triggered, |
| 95 | void *context); | ||
| 93 | 96 | ||
| 94 | extern int register_perf_hw_breakpoint(struct perf_event *bp); | 97 | extern int register_perf_hw_breakpoint(struct perf_event *bp); |
| 95 | extern int __register_perf_hw_breakpoint(struct perf_event *bp); | 98 | extern int __register_perf_hw_breakpoint(struct perf_event *bp); |
| @@ -115,6 +118,7 @@ static inline int __init init_hw_breakpoint(void) { return 0; } | |||
| 115 | static inline struct perf_event * | 118 | static inline struct perf_event * |
| 116 | register_user_hw_breakpoint(struct perf_event_attr *attr, | 119 | register_user_hw_breakpoint(struct perf_event_attr *attr, |
| 117 | perf_overflow_handler_t triggered, | 120 | perf_overflow_handler_t triggered, |
| 121 | void *context, | ||
| 118 | struct task_struct *tsk) { return NULL; } | 122 | struct task_struct *tsk) { return NULL; } |
| 119 | static inline int | 123 | static inline int |
| 120 | modify_user_hw_breakpoint(struct perf_event *bp, | 124 | modify_user_hw_breakpoint(struct perf_event *bp, |
| @@ -122,10 +126,12 @@ modify_user_hw_breakpoint(struct perf_event *bp, | |||
| 122 | static inline struct perf_event * | 126 | static inline struct perf_event * |
| 123 | register_wide_hw_breakpoint_cpu(struct perf_event_attr *attr, | 127 | register_wide_hw_breakpoint_cpu(struct perf_event_attr *attr, |
| 124 | perf_overflow_handler_t triggered, | 128 | perf_overflow_handler_t triggered, |
| 129 | void *context, | ||
| 125 | int cpu) { return NULL; } | 130 | int cpu) { return NULL; } |
| 126 | static inline struct perf_event * __percpu * | 131 | static inline struct perf_event * __percpu * |
| 127 | register_wide_hw_breakpoint(struct perf_event_attr *attr, | 132 | register_wide_hw_breakpoint(struct perf_event_attr *attr, |
| 128 | perf_overflow_handler_t triggered) { return NULL; } | 133 | perf_overflow_handler_t triggered, |
| 134 | void *context) { return NULL; } | ||
| 129 | static inline int | 135 | static inline int |
| 130 | register_perf_hw_breakpoint(struct perf_event *bp) { return -ENOSYS; } | 136 | register_perf_hw_breakpoint(struct perf_event *bp) { return -ENOSYS; } |
| 131 | static inline int | 137 | static inline int |
diff --git a/include/linux/perf_event.h b/include/linux/perf_event.h index e0786e35f24..3f2711ccf91 100644 --- a/include/linux/perf_event.h +++ b/include/linux/perf_event.h | |||
| @@ -61,7 +61,7 @@ enum perf_hw_id { | |||
| 61 | /* | 61 | /* |
| 62 | * Generalized hardware cache events: | 62 | * Generalized hardware cache events: |
| 63 | * | 63 | * |
| 64 | * { L1-D, L1-I, LLC, ITLB, DTLB, BPU } x | 64 | * { L1-D, L1-I, LLC, ITLB, DTLB, BPU, NODE } x |
| 65 | * { read, write, prefetch } x | 65 | * { read, write, prefetch } x |
| 66 | * { accesses, misses } | 66 | * { accesses, misses } |
| 67 | */ | 67 | */ |
| @@ -72,6 +72,7 @@ enum perf_hw_cache_id { | |||
| 72 | PERF_COUNT_HW_CACHE_DTLB = 3, | 72 | PERF_COUNT_HW_CACHE_DTLB = 3, |
| 73 | PERF_COUNT_HW_CACHE_ITLB = 4, | 73 | PERF_COUNT_HW_CACHE_ITLB = 4, |
| 74 | PERF_COUNT_HW_CACHE_BPU = 5, | 74 | PERF_COUNT_HW_CACHE_BPU = 5, |
| 75 | PERF_COUNT_HW_CACHE_NODE = 6, | ||
| 75 | 76 | ||
| 76 | PERF_COUNT_HW_CACHE_MAX, /* non-ABI */ | 77 | PERF_COUNT_HW_CACHE_MAX, /* non-ABI */ |
| 77 | }; | 78 | }; |
| @@ -536,6 +537,16 @@ struct perf_branch_stack { | |||
| 536 | 537 | ||
| 537 | struct task_struct; | 538 | struct task_struct; |
| 538 | 539 | ||
| 540 | /* | ||
| 541 | * extra PMU register associated with an event | ||
| 542 | */ | ||
| 543 | struct hw_perf_event_extra { | ||
| 544 | u64 config; /* register value */ | ||
| 545 | unsigned int reg; /* register address or index */ | ||
| 546 | int alloc; /* extra register already allocated */ | ||
| 547 | int idx; /* index in shared_regs->regs[] */ | ||
| 548 | }; | ||
| 549 | |||
| 539 | /** | 550 | /** |
| 540 | * struct hw_perf_event - performance event hardware details: | 551 | * struct hw_perf_event - performance event hardware details: |
| 541 | */ | 552 | */ |
| @@ -549,9 +560,7 @@ struct hw_perf_event { | |||
| 549 | unsigned long event_base; | 560 | unsigned long event_base; |
| 550 | int idx; | 561 | int idx; |
| 551 | int last_cpu; | 562 | int last_cpu; |
| 552 | unsigned int extra_reg; | 563 | struct hw_perf_event_extra extra_reg; |
| 553 | u64 extra_config; | ||
| 554 | int extra_alloc; | ||
| 555 | }; | 564 | }; |
| 556 | struct { /* software */ | 565 | struct { /* software */ |
| 557 | struct hrtimer hrtimer; | 566 | struct hrtimer hrtimer; |
| @@ -680,36 +689,9 @@ enum perf_event_active_state { | |||
| 680 | }; | 689 | }; |
| 681 | 690 | ||
| 682 | struct file; | 691 | struct file; |
| 683 | |||
| 684 | #define PERF_BUFFER_WRITABLE 0x01 | ||
| 685 | |||
| 686 | struct perf_buffer { | ||
| 687 | atomic_t refcount; | ||
| 688 | struct rcu_head rcu_head; | ||
| 689 | #ifdef CONFIG_PERF_USE_VMALLOC | ||
| 690 | struct work_struct work; | ||
| 691 | int page_order; /* allocation order */ | ||
| 692 | #endif | ||
| 693 | int nr_pages; /* nr of data pages */ | ||
| 694 | int writable; /* are we writable */ | ||
| 695 | |||
| 696 | atomic_t poll; /* POLL_ for wakeups */ | ||
| 697 | |||
| 698 | local_t head; /* write position */ | ||
| 699 | local_t nest; /* nested writers */ | ||
| 700 | local_t events; /* event limit */ | ||
| 701 | local_t wakeup; /* wakeup stamp */ | ||
| 702 | local_t lost; /* nr records lost */ | ||
| 703 | |||
| 704 | long watermark; /* wakeup watermark */ | ||
| 705 | |||
| 706 | struct perf_event_mmap_page *user_page; | ||
| 707 | void *data_pages[0]; | ||
| 708 | }; | ||
| 709 | |||
| 710 | struct perf_sample_data; | 692 | struct perf_sample_data; |
| 711 | 693 | ||
| 712 | typedef void (*perf_overflow_handler_t)(struct perf_event *, int, | 694 | typedef void (*perf_overflow_handler_t)(struct perf_event *, |
| 713 | struct perf_sample_data *, | 695 | struct perf_sample_data *, |
| 714 | struct pt_regs *regs); | 696 | struct pt_regs *regs); |
| 715 | 697 | ||
| @@ -745,6 +727,8 @@ struct perf_cgroup { | |||
| 745 | }; | 727 | }; |
| 746 | #endif | 728 | #endif |
| 747 | 729 | ||
| 730 | struct ring_buffer; | ||
| 731 | |||
| 748 | /** | 732 | /** |
| 749 | * struct perf_event - performance event kernel representation: | 733 | * struct perf_event - performance event kernel representation: |
| 750 | */ | 734 | */ |
| @@ -834,7 +818,7 @@ struct perf_event { | |||
| 834 | atomic_t mmap_count; | 818 | atomic_t mmap_count; |
| 835 | int mmap_locked; | 819 | int mmap_locked; |
| 836 | struct user_struct *mmap_user; | 820 | struct user_struct *mmap_user; |
| 837 | struct perf_buffer *buffer; | 821 | struct ring_buffer *rb; |
| 838 | 822 | ||
| 839 | /* poll related */ | 823 | /* poll related */ |
| 840 | wait_queue_head_t waitq; | 824 | wait_queue_head_t waitq; |
| @@ -855,6 +839,7 @@ struct perf_event { | |||
| 855 | u64 id; | 839 | u64 id; |
| 856 | 840 | ||
| 857 | perf_overflow_handler_t overflow_handler; | 841 | perf_overflow_handler_t overflow_handler; |
| 842 | void *overflow_handler_context; | ||
| 858 | 843 | ||
| 859 | #ifdef CONFIG_EVENT_TRACING | 844 | #ifdef CONFIG_EVENT_TRACING |
| 860 | struct ftrace_event_call *tp_event; | 845 | struct ftrace_event_call *tp_event; |
| @@ -919,8 +904,8 @@ struct perf_event_context { | |||
| 919 | u64 parent_gen; | 904 | u64 parent_gen; |
| 920 | u64 generation; | 905 | u64 generation; |
| 921 | int pin_count; | 906 | int pin_count; |
| 922 | struct rcu_head rcu_head; | ||
| 923 | int nr_cgroups; /* cgroup events present */ | 907 | int nr_cgroups; /* cgroup events present */ |
| 908 | struct rcu_head rcu_head; | ||
| 924 | }; | 909 | }; |
| 925 | 910 | ||
| 926 | /* | 911 | /* |
| @@ -945,13 +930,11 @@ struct perf_cpu_context { | |||
| 945 | 930 | ||
| 946 | struct perf_output_handle { | 931 | struct perf_output_handle { |
| 947 | struct perf_event *event; | 932 | struct perf_event *event; |
| 948 | struct perf_buffer *buffer; | 933 | struct ring_buffer *rb; |
| 949 | unsigned long wakeup; | 934 | unsigned long wakeup; |
| 950 | unsigned long size; | 935 | unsigned long size; |
| 951 | void *addr; | 936 | void *addr; |
| 952 | int page; | 937 | int page; |
| 953 | int nmi; | ||
| 954 | int sample; | ||
| 955 | }; | 938 | }; |
| 956 | 939 | ||
| 957 | #ifdef CONFIG_PERF_EVENTS | 940 | #ifdef CONFIG_PERF_EVENTS |
| @@ -972,13 +955,15 @@ extern void perf_pmu_disable(struct pmu *pmu); | |||
| 972 | extern void perf_pmu_enable(struct pmu *pmu); | 955 | extern void perf_pmu_enable(struct pmu *pmu); |
| 973 | extern int perf_event_task_disable(void); | 956 | extern int perf_event_task_disable(void); |
| 974 | extern int perf_event_task_enable(void); | 957 | extern int perf_event_task_enable(void); |
| 958 | extern int perf_event_refresh(struct perf_event *event, int refresh); | ||
| 975 | extern void perf_event_update_userpage(struct perf_event *event); | 959 | extern void perf_event_update_userpage(struct perf_event *event); |
| 976 | extern int perf_event_release_kernel(struct perf_event *event); | 960 | extern int perf_event_release_kernel(struct perf_event *event); |
| 977 | extern struct perf_event * | 961 | extern struct perf_event * |
| 978 | perf_event_create_kernel_counter(struct perf_event_attr *attr, | 962 | perf_event_create_kernel_counter(struct perf_event_attr *attr, |
| 979 | int cpu, | 963 | int cpu, |
| 980 | struct task_struct *task, | 964 | struct task_struct *task, |
| 981 | perf_overflow_handler_t callback); | 965 | perf_overflow_handler_t callback, |
| 966 | void *context); | ||
| 982 | extern u64 perf_event_read_value(struct perf_event *event, | 967 | extern u64 perf_event_read_value(struct perf_event *event, |
| 983 | u64 *enabled, u64 *running); | 968 | u64 *enabled, u64 *running); |
| 984 | 969 | ||
| @@ -1018,7 +1003,7 @@ extern void perf_prepare_sample(struct perf_event_header *header, | |||
| 1018 | struct perf_event *event, | 1003 | struct perf_event *event, |
| 1019 | struct pt_regs *regs); | 1004 | struct pt_regs *regs); |
| 1020 | 1005 | ||
| 1021 | extern int perf_event_overflow(struct perf_event *event, int nmi, | 1006 | extern int perf_event_overflow(struct perf_event *event, |
| 1022 | struct perf_sample_data *data, | 1007 | struct perf_sample_data *data, |
| 1023 | struct pt_regs *regs); | 1008 | struct pt_regs *regs); |
| 1024 | 1009 | ||
| @@ -1037,7 +1022,7 @@ static inline int is_software_event(struct perf_event *event) | |||
| 1037 | 1022 | ||
| 1038 | extern struct jump_label_key perf_swevent_enabled[PERF_COUNT_SW_MAX]; | 1023 | extern struct jump_label_key perf_swevent_enabled[PERF_COUNT_SW_MAX]; |
| 1039 | 1024 | ||
| 1040 | extern void __perf_sw_event(u32, u64, int, struct pt_regs *, u64); | 1025 | extern void __perf_sw_event(u32, u64, struct pt_regs *, u64); |
| 1041 | 1026 | ||
| 1042 | #ifndef perf_arch_fetch_caller_regs | 1027 | #ifndef perf_arch_fetch_caller_regs |
| 1043 | static inline void perf_arch_fetch_caller_regs(struct pt_regs *regs, unsigned long ip) { } | 1028 | static inline void perf_arch_fetch_caller_regs(struct pt_regs *regs, unsigned long ip) { } |
| @@ -1059,7 +1044,7 @@ static inline void perf_fetch_caller_regs(struct pt_regs *regs) | |||
| 1059 | } | 1044 | } |
| 1060 | 1045 | ||
| 1061 | static __always_inline void | 1046 | static __always_inline void |
| 1062 | perf_sw_event(u32 event_id, u64 nr, int nmi, struct pt_regs *regs, u64 addr) | 1047 | perf_sw_event(u32 event_id, u64 nr, struct pt_regs *regs, u64 addr) |
| 1063 | { | 1048 | { |
| 1064 | struct pt_regs hot_regs; | 1049 | struct pt_regs hot_regs; |
| 1065 | 1050 | ||
| @@ -1068,7 +1053,7 @@ perf_sw_event(u32 event_id, u64 nr, int nmi, struct pt_regs *regs, u64 addr) | |||
| 1068 | perf_fetch_caller_regs(&hot_regs); | 1053 | perf_fetch_caller_regs(&hot_regs); |
| 1069 | regs = &hot_regs; | 1054 | regs = &hot_regs; |
| 1070 | } | 1055 | } |
| 1071 | __perf_sw_event(event_id, nr, nmi, regs, addr); | 1056 | __perf_sw_event(event_id, nr, regs, addr); |
| 1072 | } | 1057 | } |
| 1073 | } | 1058 | } |
| 1074 | 1059 | ||
| @@ -1082,7 +1067,7 @@ static inline void perf_event_task_sched_in(struct task_struct *task) | |||
| 1082 | 1067 | ||
| 1083 | static inline void perf_event_task_sched_out(struct task_struct *task, struct task_struct *next) | 1068 | static inline void perf_event_task_sched_out(struct task_struct *task, struct task_struct *next) |
| 1084 | { | 1069 | { |
| 1085 | perf_sw_event(PERF_COUNT_SW_CONTEXT_SWITCHES, 1, 1, NULL, 0); | 1070 | perf_sw_event(PERF_COUNT_SW_CONTEXT_SWITCHES, 1, NULL, 0); |
| 1086 | 1071 | ||
| 1087 | __perf_event_task_sched_out(task, next); | 1072 | __perf_event_task_sched_out(task, next); |
| 1088 | } | 1073 | } |
| @@ -1143,8 +1128,7 @@ extern void perf_bp_event(struct perf_event *event, void *data); | |||
| 1143 | #endif | 1128 | #endif |
| 1144 | 1129 | ||
| 1145 | extern int perf_output_begin(struct perf_output_handle *handle, | 1130 | extern int perf_output_begin(struct perf_output_handle *handle, |
| 1146 | struct perf_event *event, unsigned int size, | 1131 | struct perf_event *event, unsigned int size); |
| 1147 | int nmi, int sample); | ||
| 1148 | extern void perf_output_end(struct perf_output_handle *handle); | 1132 | extern void perf_output_end(struct perf_output_handle *handle); |
| 1149 | extern void perf_output_copy(struct perf_output_handle *handle, | 1133 | extern void perf_output_copy(struct perf_output_handle *handle, |
| 1150 | const void *buf, unsigned int len); | 1134 | const void *buf, unsigned int len); |
| @@ -1166,10 +1150,13 @@ static inline void perf_event_delayed_put(struct task_struct *task) { } | |||
| 1166 | static inline void perf_event_print_debug(void) { } | 1150 | static inline void perf_event_print_debug(void) { } |
| 1167 | static inline int perf_event_task_disable(void) { return -EINVAL; } | 1151 | static inline int perf_event_task_disable(void) { return -EINVAL; } |
| 1168 | static inline int perf_event_task_enable(void) { return -EINVAL; } | 1152 | static inline int perf_event_task_enable(void) { return -EINVAL; } |
| 1153 | static inline int perf_event_refresh(struct perf_event *event, int refresh) | ||
| 1154 | { | ||
| 1155 | return -EINVAL; | ||
| 1156 | } | ||
| 1169 | 1157 | ||
| 1170 | static inline void | 1158 | static inline void |
| 1171 | perf_sw_event(u32 event_id, u64 nr, int nmi, | 1159 | perf_sw_event(u32 event_id, u64 nr, struct pt_regs *regs, u64 addr) { } |
| 1172 | struct pt_regs *regs, u64 addr) { } | ||
| 1173 | static inline void | 1160 | static inline void |
| 1174 | perf_bp_event(struct perf_event *event, void *data) { } | 1161 | perf_bp_event(struct perf_event *event, void *data) { } |
| 1175 | 1162 | ||
diff --git a/include/linux/ring_buffer.h b/include/linux/ring_buffer.h index ab38ac80b0f..b891de96000 100644 --- a/include/linux/ring_buffer.h +++ b/include/linux/ring_buffer.h | |||
| @@ -169,7 +169,7 @@ void ring_buffer_set_clock(struct ring_buffer *buffer, | |||
| 169 | size_t ring_buffer_page_len(void *page); | 169 | size_t ring_buffer_page_len(void *page); |
| 170 | 170 | ||
| 171 | 171 | ||
| 172 | void *ring_buffer_alloc_read_page(struct ring_buffer *buffer); | 172 | void *ring_buffer_alloc_read_page(struct ring_buffer *buffer, int cpu); |
| 173 | void ring_buffer_free_read_page(struct ring_buffer *buffer, void *data); | 173 | void ring_buffer_free_read_page(struct ring_buffer *buffer, void *data); |
| 174 | int ring_buffer_read_page(struct ring_buffer *buffer, void **data_page, | 174 | int ring_buffer_read_page(struct ring_buffer *buffer, void **data_page, |
| 175 | size_t len, int cpu, int full); | 175 | size_t len, int cpu, int full); |
diff --git a/include/linux/stacktrace.h b/include/linux/stacktrace.h index 25310f1d7f3..115b570e3bf 100644 --- a/include/linux/stacktrace.h +++ b/include/linux/stacktrace.h | |||
| @@ -14,8 +14,8 @@ struct stack_trace { | |||
| 14 | }; | 14 | }; |
| 15 | 15 | ||
| 16 | extern void save_stack_trace(struct stack_trace *trace); | 16 | extern void save_stack_trace(struct stack_trace *trace); |
| 17 | extern void save_stack_trace_regs(struct stack_trace *trace, | 17 | extern void save_stack_trace_regs(struct pt_regs *regs, |
| 18 | struct pt_regs *regs); | 18 | struct stack_trace *trace); |
| 19 | extern void save_stack_trace_tsk(struct task_struct *tsk, | 19 | extern void save_stack_trace_tsk(struct task_struct *tsk, |
| 20 | struct stack_trace *trace); | 20 | struct stack_trace *trace); |
| 21 | 21 | ||
diff --git a/kernel/async.c b/kernel/async.c index cd9dbb913c7..d5fe7af0de2 100644 --- a/kernel/async.c +++ b/kernel/async.c | |||
| @@ -49,12 +49,13 @@ asynchronous and synchronous parts of the kernel. | |||
| 49 | */ | 49 | */ |
| 50 | 50 | ||
| 51 | #include <linux/async.h> | 51 | #include <linux/async.h> |
| 52 | #include <linux/atomic.h> | ||
| 53 | #include <linux/ktime.h> | ||
| 52 | #include <linux/module.h> | 54 | #include <linux/module.h> |
| 53 | #include <linux/wait.h> | 55 | #include <linux/wait.h> |
| 54 | #include <linux/sched.h> | 56 | #include <linux/sched.h> |
| 55 | #include <linux/slab.h> | 57 | #include <linux/slab.h> |
| 56 | #include <linux/workqueue.h> | 58 | #include <linux/workqueue.h> |
| 57 | #include <asm/atomic.h> | ||
| 58 | 59 | ||
| 59 | static async_cookie_t next_cookie = 1; | 60 | static async_cookie_t next_cookie = 1; |
| 60 | 61 | ||
| @@ -128,7 +129,8 @@ static void async_run_entry_fn(struct work_struct *work) | |||
| 128 | 129 | ||
| 129 | /* 2) run (and print duration) */ | 130 | /* 2) run (and print duration) */ |
| 130 | if (initcall_debug && system_state == SYSTEM_BOOTING) { | 131 | if (initcall_debug && system_state == SYSTEM_BOOTING) { |
| 131 | printk("calling %lli_%pF @ %i\n", (long long)entry->cookie, | 132 | printk(KERN_DEBUG "calling %lli_%pF @ %i\n", |
| 133 | (long long)entry->cookie, | ||
| 132 | entry->func, task_pid_nr(current)); | 134 | entry->func, task_pid_nr(current)); |
| 133 | calltime = ktime_get(); | 135 | calltime = ktime_get(); |
| 134 | } | 136 | } |
| @@ -136,7 +138,7 @@ static void async_run_entry_fn(struct work_struct *work) | |||
| 136 | if (initcall_debug && system_state == SYSTEM_BOOTING) { | 138 | if (initcall_debug && system_state == SYSTEM_BOOTING) { |
| 137 | rettime = ktime_get(); | 139 | rettime = ktime_get(); |
| 138 | delta = ktime_sub(rettime, calltime); | 140 | delta = ktime_sub(rettime, calltime); |
| 139 | printk("initcall %lli_%pF returned 0 after %lld usecs\n", | 141 | printk(KERN_DEBUG "initcall %lli_%pF returned 0 after %lld usecs\n", |
| 140 | (long long)entry->cookie, | 142 | (long long)entry->cookie, |
| 141 | entry->func, | 143 | entry->func, |
| 142 | (long long)ktime_to_ns(delta) >> 10); | 144 | (long long)ktime_to_ns(delta) >> 10); |
| @@ -270,7 +272,7 @@ void async_synchronize_cookie_domain(async_cookie_t cookie, | |||
| 270 | ktime_t starttime, delta, endtime; | 272 | ktime_t starttime, delta, endtime; |
| 271 | 273 | ||
| 272 | if (initcall_debug && system_state == SYSTEM_BOOTING) { | 274 | if (initcall_debug && system_state == SYSTEM_BOOTING) { |
| 273 | printk("async_waiting @ %i\n", task_pid_nr(current)); | 275 | printk(KERN_DEBUG "async_waiting @ %i\n", task_pid_nr(current)); |
| 274 | starttime = ktime_get(); | 276 | starttime = ktime_get(); |
| 275 | } | 277 | } |
| 276 | 278 | ||
| @@ -280,7 +282,7 @@ void async_synchronize_cookie_domain(async_cookie_t cookie, | |||
| 280 | endtime = ktime_get(); | 282 | endtime = ktime_get(); |
| 281 | delta = ktime_sub(endtime, starttime); | 283 | delta = ktime_sub(endtime, starttime); |
| 282 | 284 | ||
| 283 | printk("async_continuing @ %i after %lli usec\n", | 285 | printk(KERN_DEBUG "async_continuing @ %i after %lli usec\n", |
| 284 | task_pid_nr(current), | 286 | task_pid_nr(current), |
| 285 | (long long)ktime_to_ns(delta) >> 10); | 287 | (long long)ktime_to_ns(delta) >> 10); |
| 286 | } | 288 | } |
diff --git a/kernel/events/Makefile b/kernel/events/Makefile index 1ce23d3d839..89e5e8aa4c3 100644 --- a/kernel/events/Makefile +++ b/kernel/events/Makefile | |||
| @@ -2,5 +2,5 @@ ifdef CONFIG_FUNCTION_TRACER | |||
| 2 | CFLAGS_REMOVE_core.o = -pg | 2 | CFLAGS_REMOVE_core.o = -pg |
| 3 | endif | 3 | endif |
| 4 | 4 | ||
| 5 | obj-y := core.o | 5 | obj-y := core.o ring_buffer.o |
| 6 | obj-$(CONFIG_HAVE_HW_BREAKPOINT) += hw_breakpoint.o | 6 | obj-$(CONFIG_HAVE_HW_BREAKPOINT) += hw_breakpoint.o |
diff --git a/kernel/events/core.c b/kernel/events/core.c index 9efe7108cca..b8785e26ee1 100644 --- a/kernel/events/core.c +++ b/kernel/events/core.c | |||
| @@ -36,6 +36,8 @@ | |||
| 36 | #include <linux/ftrace_event.h> | 36 | #include <linux/ftrace_event.h> |
| 37 | #include <linux/hw_breakpoint.h> | 37 | #include <linux/hw_breakpoint.h> |
| 38 | 38 | ||
| 39 | #include "internal.h" | ||
| 40 | |||
| 39 | #include <asm/irq_regs.h> | 41 | #include <asm/irq_regs.h> |
| 40 | 42 | ||
| 41 | struct remote_function_call { | 43 | struct remote_function_call { |
| @@ -200,6 +202,22 @@ __get_cpu_context(struct perf_event_context *ctx) | |||
| 200 | return this_cpu_ptr(ctx->pmu->pmu_cpu_context); | 202 | return this_cpu_ptr(ctx->pmu->pmu_cpu_context); |
| 201 | } | 203 | } |
| 202 | 204 | ||
| 205 | static void perf_ctx_lock(struct perf_cpu_context *cpuctx, | ||
| 206 | struct perf_event_context *ctx) | ||
| 207 | { | ||
| 208 | raw_spin_lock(&cpuctx->ctx.lock); | ||
| 209 | if (ctx) | ||
| 210 | raw_spin_lock(&ctx->lock); | ||
| 211 | } | ||
| 212 | |||
| 213 | static void perf_ctx_unlock(struct perf_cpu_context *cpuctx, | ||
| 214 | struct perf_event_context *ctx) | ||
| 215 | { | ||
| 216 | if (ctx) | ||
| 217 | raw_spin_unlock(&ctx->lock); | ||
| 218 | raw_spin_unlock(&cpuctx->ctx.lock); | ||
| 219 | } | ||
| 220 | |||
| 203 | #ifdef CONFIG_CGROUP_PERF | 221 | #ifdef CONFIG_CGROUP_PERF |
| 204 | 222 | ||
| 205 | /* | 223 | /* |
| @@ -340,11 +358,8 @@ void perf_cgroup_switch(struct task_struct *task, int mode) | |||
| 340 | rcu_read_lock(); | 358 | rcu_read_lock(); |
| 341 | 359 | ||
| 342 | list_for_each_entry_rcu(pmu, &pmus, entry) { | 360 | list_for_each_entry_rcu(pmu, &pmus, entry) { |
| 343 | |||
| 344 | cpuctx = this_cpu_ptr(pmu->pmu_cpu_context); | 361 | cpuctx = this_cpu_ptr(pmu->pmu_cpu_context); |
| 345 | 362 | ||
| 346 | perf_pmu_disable(cpuctx->ctx.pmu); | ||
| 347 | |||
| 348 | /* | 363 | /* |
| 349 | * perf_cgroup_events says at least one | 364 | * perf_cgroup_events says at least one |
| 350 | * context on this CPU has cgroup events. | 365 | * context on this CPU has cgroup events. |
| @@ -353,6 +368,8 @@ void perf_cgroup_switch(struct task_struct *task, int mode) | |||
| 353 | * events for a context. | 368 | * events for a context. |
| 354 | */ | 369 | */ |
| 355 | if (cpuctx->ctx.nr_cgroups > 0) { | 370 | if (cpuctx->ctx.nr_cgroups > 0) { |
| 371 | perf_ctx_lock(cpuctx, cpuctx->task_ctx); | ||
| 372 | perf_pmu_disable(cpuctx->ctx.pmu); | ||
| 356 | 373 | ||
| 357 | if (mode & PERF_CGROUP_SWOUT) { | 374 | if (mode & PERF_CGROUP_SWOUT) { |
| 358 | cpu_ctx_sched_out(cpuctx, EVENT_ALL); | 375 | cpu_ctx_sched_out(cpuctx, EVENT_ALL); |
| @@ -372,9 +389,9 @@ void perf_cgroup_switch(struct task_struct *task, int mode) | |||
| 372 | cpuctx->cgrp = perf_cgroup_from_task(task); | 389 | cpuctx->cgrp = perf_cgroup_from_task(task); |
| 373 | cpu_ctx_sched_in(cpuctx, EVENT_ALL, task); | 390 | cpu_ctx_sched_in(cpuctx, EVENT_ALL, task); |
| 374 | } | 391 | } |
| 392 | perf_pmu_enable(cpuctx->ctx.pmu); | ||
| 393 | perf_ctx_unlock(cpuctx, cpuctx->task_ctx); | ||
| 375 | } | 394 | } |
| 376 | |||
| 377 | perf_pmu_enable(cpuctx->ctx.pmu); | ||
| 378 | } | 395 | } |
| 379 | 396 | ||
| 380 | rcu_read_unlock(); | 397 | rcu_read_unlock(); |
| @@ -731,6 +748,7 @@ static u64 perf_event_time(struct perf_event *event) | |||
| 731 | 748 | ||
| 732 | /* | 749 | /* |
| 733 | * Update the total_time_enabled and total_time_running fields for a event. | 750 | * Update the total_time_enabled and total_time_running fields for a event. |
| 751 | * The caller of this function needs to hold the ctx->lock. | ||
| 734 | */ | 752 | */ |
| 735 | static void update_event_times(struct perf_event *event) | 753 | static void update_event_times(struct perf_event *event) |
| 736 | { | 754 | { |
| @@ -1105,6 +1123,10 @@ static int __perf_remove_from_context(void *info) | |||
| 1105 | raw_spin_lock(&ctx->lock); | 1123 | raw_spin_lock(&ctx->lock); |
| 1106 | event_sched_out(event, cpuctx, ctx); | 1124 | event_sched_out(event, cpuctx, ctx); |
| 1107 | list_del_event(event, ctx); | 1125 | list_del_event(event, ctx); |
| 1126 | if (!ctx->nr_events && cpuctx->task_ctx == ctx) { | ||
| 1127 | ctx->is_active = 0; | ||
| 1128 | cpuctx->task_ctx = NULL; | ||
| 1129 | } | ||
| 1108 | raw_spin_unlock(&ctx->lock); | 1130 | raw_spin_unlock(&ctx->lock); |
| 1109 | 1131 | ||
| 1110 | return 0; | 1132 | return 0; |
| @@ -1454,8 +1476,24 @@ static void add_event_to_ctx(struct perf_event *event, | |||
| 1454 | event->tstamp_stopped = tstamp; | 1476 | event->tstamp_stopped = tstamp; |
| 1455 | } | 1477 | } |
| 1456 | 1478 | ||
| 1457 | static void perf_event_context_sched_in(struct perf_event_context *ctx, | 1479 | static void task_ctx_sched_out(struct perf_event_context *ctx); |
| 1458 | struct task_struct *tsk); | 1480 | static void |
| 1481 | ctx_sched_in(struct perf_event_context *ctx, | ||
| 1482 | struct perf_cpu_context *cpuctx, | ||
| 1483 | enum event_type_t event_type, | ||
| 1484 | struct task_struct *task); | ||
| 1485 | |||
| 1486 | static void perf_event_sched_in(struct perf_cpu_context *cpuctx, | ||
| 1487 | struct perf_event_context *ctx, | ||
| 1488 | struct task_struct *task) | ||
| 1489 | { | ||
| 1490 | cpu_ctx_sched_in(cpuctx, EVENT_PINNED, task); | ||
| 1491 | if (ctx) | ||
| 1492 | ctx_sched_in(ctx, cpuctx, EVENT_PINNED, task); | ||
| 1493 | cpu_ctx_sched_in(cpuctx, EVENT_FLEXIBLE, task); | ||
| 1494 | if (ctx) | ||
| 1495 | ctx_sched_in(ctx, cpuctx, EVENT_FLEXIBLE, task); | ||
| 1496 | } | ||
| 1459 | 1497 | ||
| 1460 | /* | 1498 | /* |
| 1461 | * Cross CPU call to install and enable a performance event | 1499 | * Cross CPU call to install and enable a performance event |
| @@ -1466,20 +1504,37 @@ static int __perf_install_in_context(void *info) | |||
| 1466 | { | 1504 | { |
| 1467 | struct perf_event *event = info; | 1505 | struct perf_event *event = info; |
| 1468 | struct perf_event_context *ctx = event->ctx; | 1506 | struct perf_event_context *ctx = event->ctx; |
| 1469 | struct perf_event *leader = event->group_leader; | ||
| 1470 | struct perf_cpu_context *cpuctx = __get_cpu_context(ctx); | 1507 | struct perf_cpu_context *cpuctx = __get_cpu_context(ctx); |
| 1471 | int err; | 1508 | struct perf_event_context *task_ctx = cpuctx->task_ctx; |
| 1509 | struct task_struct *task = current; | ||
| 1510 | |||
| 1511 | perf_ctx_lock(cpuctx, task_ctx); | ||
| 1512 | perf_pmu_disable(cpuctx->ctx.pmu); | ||
| 1472 | 1513 | ||
| 1473 | /* | 1514 | /* |
| 1474 | * In case we're installing a new context to an already running task, | 1515 | * If there was an active task_ctx schedule it out. |
| 1475 | * could also happen before perf_event_task_sched_in() on architectures | ||
| 1476 | * which do context switches with IRQs enabled. | ||
| 1477 | */ | 1516 | */ |
| 1478 | if (ctx->task && !cpuctx->task_ctx) | 1517 | if (task_ctx) |
| 1479 | perf_event_context_sched_in(ctx, ctx->task); | 1518 | task_ctx_sched_out(task_ctx); |
| 1519 | |||
| 1520 | /* | ||
| 1521 | * If the context we're installing events in is not the | ||
| 1522 | * active task_ctx, flip them. | ||
| 1523 | */ | ||
| 1524 | if (ctx->task && task_ctx != ctx) { | ||
| 1525 | if (task_ctx) | ||
| 1526 | raw_spin_unlock(&task_ctx->lock); | ||
| 1527 | raw_spin_lock(&ctx->lock); | ||
| 1528 | task_ctx = ctx; | ||
| 1529 | } | ||
| 1530 | |||
| 1531 | if (task_ctx) { | ||
| 1532 | cpuctx->task_ctx = task_ctx; | ||
| 1533 | task = task_ctx->task; | ||
| 1534 | } | ||
| 1535 | |||
| 1536 | cpu_ctx_sched_out(cpuctx, EVENT_ALL); | ||
| 1480 | 1537 | ||
| 1481 | raw_spin_lock(&ctx->lock); | ||
| 1482 | ctx->is_active = 1; | ||
| 1483 | update_context_time(ctx); | 1538 | update_context_time(ctx); |
| 1484 | /* | 1539 | /* |
| 1485 | * update cgrp time only if current cgrp | 1540 | * update cgrp time only if current cgrp |
| @@ -1490,43 +1545,13 @@ static int __perf_install_in_context(void *info) | |||
| 1490 | 1545 | ||
| 1491 | add_event_to_ctx(event, ctx); | 1546 | add_event_to_ctx(event, ctx); |
| 1492 | 1547 | ||
| 1493 | if (!event_filter_match(event)) | ||
| 1494 | goto unlock; | ||
| 1495 | |||
| 1496 | /* | ||
| 1497 | * Don't put the event on if it is disabled or if | ||
| 1498 | * it is in a group and the group isn't on. | ||
| 1499 | */ | ||
| 1500 | if (event->state != PERF_EVENT_STATE_INACTIVE || | ||
| 1501 | (leader != event && leader->state != PERF_EVENT_STATE_ACTIVE)) | ||
| 1502 | goto unlock; | ||
| 1503 | |||
| 1504 | /* | 1548 | /* |
| 1505 | * An exclusive event can't go on if there are already active | 1549 | * Schedule everything back in |
| 1506 | * hardware events, and no hardware event can go on if there | ||
| 1507 | * is already an exclusive event on. | ||
| 1508 | */ | 1550 | */ |
| 1509 | if (!group_can_go_on(event, cpuctx, 1)) | 1551 | perf_event_sched_in(cpuctx, task_ctx, task); |
| 1510 | err = -EEXIST; | ||
| 1511 | else | ||
| 1512 | err = event_sched_in(event, cpuctx, ctx); | ||
| 1513 | |||
| 1514 | if (err) { | ||
| 1515 | /* | ||
| 1516 | * This event couldn't go on. If it is in a group | ||
| 1517 | * then we have to pull the whole group off. | ||
| 1518 | * If the event group is pinned then put it in error state. | ||
| 1519 | */ | ||
| 1520 | if (leader != event) | ||
| 1521 | group_sched_out(leader, cpuctx, ctx); | ||
| 1522 | if (leader->attr.pinned) { | ||
| 1523 | update_group_times(leader); | ||
| 1524 | leader->state = PERF_EVENT_STATE_ERROR; | ||
| 1525 | } | ||
| 1526 | } | ||
| 1527 | 1552 | ||
| 1528 | unlock: | 1553 | perf_pmu_enable(cpuctx->ctx.pmu); |
| 1529 | raw_spin_unlock(&ctx->lock); | 1554 | perf_ctx_unlock(cpuctx, task_ctx); |
| 1530 | 1555 | ||
| 1531 | return 0; | 1556 | return 0; |
| 1532 | } | 1557 | } |
| @@ -1739,7 +1764,7 @@ out: | |||
| 1739 | raw_spin_unlock_irq(&ctx->lock); | 1764 | raw_spin_unlock_irq(&ctx->lock); |
| 1740 | } | 1765 | } |
| 1741 | 1766 | ||
| 1742 | static int perf_event_refresh(struct perf_event *event, int refresh) | 1767 | int perf_event_refresh(struct perf_event *event, int refresh) |
| 1743 | { | 1768 | { |
| 1744 | /* | 1769 | /* |
| 1745 | * not supported on inherited events | 1770 | * not supported on inherited events |
| @@ -1752,36 +1777,35 @@ static int perf_event_refresh(struct perf_event *event, int refresh) | |||
| 1752 | 1777 | ||
| 1753 | return 0; | 1778 | return 0; |
| 1754 | } | 1779 | } |
| 1780 | EXPORT_SYMBOL_GPL(perf_event_refresh); | ||
| 1755 | 1781 | ||
| 1756 | static void ctx_sched_out(struct perf_event_context *ctx, | 1782 | static void ctx_sched_out(struct perf_event_context *ctx, |
| 1757 | struct perf_cpu_context *cpuctx, | 1783 | struct perf_cpu_context *cpuctx, |
| 1758 | enum event_type_t event_type) | 1784 | enum event_type_t event_type) |
| 1759 | { | 1785 | { |
| 1760 | struct perf_event *event; | 1786 | struct perf_event *event; |
| 1787 | int is_active = ctx->is_active; | ||
| 1761 | 1788 | ||
| 1762 | raw_spin_lock(&ctx->lock); | 1789 | ctx->is_active &= ~event_type; |
| 1763 | perf_pmu_disable(ctx->pmu); | ||
| 1764 | ctx->is_active = 0; | ||
| 1765 | if (likely(!ctx->nr_events)) | 1790 | if (likely(!ctx->nr_events)) |
| 1766 | goto out; | 1791 | return; |
| 1792 | |||
| 1767 | update_context_time(ctx); | 1793 | update_context_time(ctx); |
| 1768 | update_cgrp_time_from_cpuctx(cpuctx); | 1794 | update_cgrp_time_from_cpuctx(cpuctx); |
| 1769 | |||
| 1770 | if (!ctx->nr_active) | 1795 | if (!ctx->nr_active) |
| 1771 | goto out; | 1796 | return; |
| 1772 | 1797 | ||
| 1773 | if (event_type & EVENT_PINNED) { | 1798 | perf_pmu_disable(ctx->pmu); |
| 1799 | if ((is_active & EVENT_PINNED) && (event_type & EVENT_PINNED)) { | ||
| 1774 | list_for_each_entry(event, &ctx->pinned_groups, group_entry) | 1800 | list_for_each_entry(event, &ctx->pinned_groups, group_entry) |
| 1775 | group_sched_out(event, cpuctx, ctx); | 1801 | group_sched_out(event, cpuctx, ctx); |
| 1776 | } | 1802 | } |
| 1777 | 1803 | ||
| 1778 | if (event_type & EVENT_FLEXIBLE) { | 1804 | if ((is_active & EVENT_FLEXIBLE) && (event_type & EVENT_FLEXIBLE)) { |
| 1779 | list_for_each_entry(event, &ctx->flexible_groups, group_entry) | 1805 | list_for_each_entry(event, &ctx->flexible_groups, group_entry) |
| 1780 | group_sched_out(event, cpuctx, ctx); | 1806 | group_sched_out(event, cpuctx, ctx); |
| 1781 | } | 1807 | } |
| 1782 | out: | ||
| 1783 | perf_pmu_enable(ctx->pmu); | 1808 | perf_pmu_enable(ctx->pmu); |
| 1784 | raw_spin_unlock(&ctx->lock); | ||
| 1785 | } | 1809 | } |
| 1786 | 1810 | ||
| 1787 | /* | 1811 | /* |
| @@ -1929,8 +1953,10 @@ static void perf_event_context_sched_out(struct task_struct *task, int ctxn, | |||
| 1929 | rcu_read_unlock(); | 1953 | rcu_read_unlock(); |
| 1930 | 1954 | ||
| 1931 | if (do_switch) { | 1955 | if (do_switch) { |
| 1956 | raw_spin_lock(&ctx->lock); | ||
| 1932 | ctx_sched_out(ctx, cpuctx, EVENT_ALL); | 1957 | ctx_sched_out(ctx, cpuctx, EVENT_ALL); |
| 1933 | cpuctx->task_ctx = NULL; | 1958 | cpuctx->task_ctx = NULL; |
| 1959 | raw_spin_unlock(&ctx->lock); | ||
| 1934 | } | 1960 | } |
| 1935 | } | 1961 | } |
| 1936 | 1962 | ||
| @@ -1965,8 +1991,7 @@ void __perf_event_task_sched_out(struct task_struct *task, | |||
| 1965 | perf_cgroup_sched_out(task); | 1991 | perf_cgroup_sched_out(task); |
| 1966 | } | 1992 | } |
| 1967 | 1993 | ||
| 1968 | static void task_ctx_sched_out(struct perf_event_context *ctx, | 1994 | static void task_ctx_sched_out(struct perf_event_context *ctx) |
| 1969 | enum event_type_t event_type) | ||
| 1970 | { | 1995 | { |
| 1971 | struct perf_cpu_context *cpuctx = __get_cpu_context(ctx); | 1996 | struct perf_cpu_context *cpuctx = __get_cpu_context(ctx); |
| 1972 | 1997 | ||
| @@ -1976,7 +2001,7 @@ static void task_ctx_sched_out(struct perf_event_context *ctx, | |||
| 1976 | if (WARN_ON_ONCE(ctx != cpuctx->task_ctx)) | 2001 | if (WARN_ON_ONCE(ctx != cpuctx->task_ctx)) |
| 1977 | return; | 2002 | return; |
| 1978 | 2003 | ||
| 1979 | ctx_sched_out(ctx, cpuctx, event_type); | 2004 | ctx_sched_out(ctx, cpuctx, EVENT_ALL); |
| 1980 | cpuctx->task_ctx = NULL; | 2005 | cpuctx->task_ctx = NULL; |
| 1981 | } | 2006 | } |
| 1982 | 2007 | ||
| @@ -2055,11 +2080,11 @@ ctx_sched_in(struct perf_event_context *ctx, | |||
| 2055 | struct task_struct *task) | 2080 | struct task_struct *task) |
| 2056 | { | 2081 | { |
| 2057 | u64 now; | 2082 | u64 now; |
| 2083 | int is_active = ctx->is_active; | ||
| 2058 | 2084 | ||
| 2059 | raw_spin_lock(&ctx->lock); | 2085 | ctx->is_active |= event_type; |
| 2060 | ctx->is_active = 1; | ||
| 2061 | if (likely(!ctx->nr_events)) | 2086 | if (likely(!ctx->nr_events)) |
| 2062 | goto out; | 2087 | return; |
| 2063 | 2088 | ||
| 2064 | now = perf_clock(); | 2089 | now = perf_clock(); |
| 2065 | ctx->timestamp = now; | 2090 | ctx->timestamp = now; |
| @@ -2068,15 +2093,12 @@ ctx_sched_in(struct perf_event_context *ctx, | |||
| 2068 | * First go through the list and put on any pinned groups | 2093 | * First go through the list and put on any pinned groups |
| 2069 | * in order to give them the best chance of going on. | 2094 | * in order to give them the best chance of going on. |
| 2070 | */ | 2095 | */ |
| 2071 | if (event_type & EVENT_PINNED) | 2096 | if (!(is_active & EVENT_PINNED) && (event_type & EVENT_PINNED)) |
| 2072 | ctx_pinned_sched_in(ctx, cpuctx); | 2097 | ctx_pinned_sched_in(ctx, cpuctx); |
| 2073 | 2098 | ||
| 2074 | /* Then walk through the lower prio flexible groups */ | 2099 | /* Then walk through the lower prio flexible groups */ |
| 2075 | if (event_type & EVENT_FLEXIBLE) | 2100 | if (!(is_active & EVENT_FLEXIBLE) && (event_type & EVENT_FLEXIBLE)) |
| 2076 | ctx_flexible_sched_in(ctx, cpuctx); | 2101 | ctx_flexible_sched_in(ctx, cpuctx); |
| 2077 | |||
| 2078 | out: | ||
| 2079 | raw_spin_unlock(&ctx->lock); | ||
| 2080 | } | 2102 | } |
| 2081 | 2103 | ||
| 2082 | static void cpu_ctx_sched_in(struct perf_cpu_context *cpuctx, | 2104 | static void cpu_ctx_sched_in(struct perf_cpu_context *cpuctx, |
| @@ -2088,19 +2110,6 @@ static void cpu_ctx_sched_in(struct perf_cpu_context *cpuctx, | |||
| 2088 | ctx_sched_in(ctx, cpuctx, event_type, task); | 2110 | ctx_sched_in(ctx, cpuctx, event_type, task); |
| 2089 | } | 2111 | } |
| 2090 | 2112 | ||
| 2091 | static void task_ctx_sched_in(struct perf_event_context *ctx, | ||
| 2092 | enum event_type_t event_type) | ||
| 2093 | { | ||
| 2094 | struct perf_cpu_context *cpuctx; | ||
| 2095 | |||
| 2096 | cpuctx = __get_cpu_context(ctx); | ||
| 2097 | if (cpuctx->task_ctx == ctx) | ||
| 2098 | return; | ||
| 2099 | |||
| 2100 | ctx_sched_in(ctx, cpuctx, event_type, NULL); | ||
| 2101 | cpuctx->task_ctx = ctx; | ||
| 2102 | } | ||
| 2103 | |||
| 2104 | static void perf_event_context_sched_in(struct perf_event_context *ctx, | 2113 | static void perf_event_context_sched_in(struct perf_event_context *ctx, |
| 2105 | struct task_struct *task) | 2114 | struct task_struct *task) |
| 2106 | { | 2115 | { |
| @@ -2110,6 +2119,7 @@ static void perf_event_context_sched_in(struct perf_event_context *ctx, | |||
| 2110 | if (cpuctx->task_ctx == ctx) | 2119 | if (cpuctx->task_ctx == ctx) |
| 2111 | return; | 2120 | return; |
| 2112 | 2121 | ||
| 2122 | perf_ctx_lock(cpuctx, ctx); | ||
| 2113 | perf_pmu_disable(ctx->pmu); | 2123 | perf_pmu_disable(ctx->pmu); |
| 2114 | /* | 2124 | /* |
| 2115 | * We want to keep the following priority order: | 2125 | * We want to keep the following priority order: |
| @@ -2118,18 +2128,18 @@ static void perf_event_context_sched_in(struct perf_event_context *ctx, | |||
| 2118 | */ | 2128 | */ |
| 2119 | cpu_ctx_sched_out(cpuctx, EVENT_FLEXIBLE); | 2129 | cpu_ctx_sched_out(cpuctx, EVENT_FLEXIBLE); |
| 2120 | 2130 | ||
| 2121 | ctx_sched_in(ctx, cpuctx, EVENT_PINNED, task); | 2131 | perf_event_sched_in(cpuctx, ctx, task); |
| 2122 | cpu_ctx_sched_in(cpuctx, EVENT_FLEXIBLE, task); | ||
| 2123 | ctx_sched_in(ctx, cpuctx, EVENT_FLEXIBLE, task); | ||
| 2124 | 2132 | ||
| 2125 | cpuctx->task_ctx = ctx; | 2133 | cpuctx->task_ctx = ctx; |
| 2126 | 2134 | ||
| 2135 | perf_pmu_enable(ctx->pmu); | ||
| 2136 | perf_ctx_unlock(cpuctx, ctx); | ||
| 2137 | |||
| 2127 | /* | 2138 | /* |
| 2128 | * Since these rotations are per-cpu, we need to ensure the | 2139 | * Since these rotations are per-cpu, we need to ensure the |
| 2129 | * cpu-context we got scheduled on is actually rotating. | 2140 | * cpu-context we got scheduled on is actually rotating. |
| 2130 | */ | 2141 | */ |
| 2131 | perf_pmu_rotate_start(ctx->pmu); | 2142 | perf_pmu_rotate_start(ctx->pmu); |
| 2132 | perf_pmu_enable(ctx->pmu); | ||
| 2133 | } | 2143 | } |
| 2134 | 2144 | ||
| 2135 | /* | 2145 | /* |
| @@ -2269,7 +2279,6 @@ static void perf_ctx_adjust_freq(struct perf_event_context *ctx, u64 period) | |||
| 2269 | u64 interrupts, now; | 2279 | u64 interrupts, now; |
| 2270 | s64 delta; | 2280 | s64 delta; |
| 2271 | 2281 | ||
| 2272 | raw_spin_lock(&ctx->lock); | ||
| 2273 | list_for_each_entry_rcu(event, &ctx->event_list, event_entry) { | 2282 | list_for_each_entry_rcu(event, &ctx->event_list, event_entry) { |
| 2274 | if (event->state != PERF_EVENT_STATE_ACTIVE) | 2283 | if (event->state != PERF_EVENT_STATE_ACTIVE) |
| 2275 | continue; | 2284 | continue; |
| @@ -2301,7 +2310,6 @@ static void perf_ctx_adjust_freq(struct perf_event_context *ctx, u64 period) | |||
| 2301 | if (delta > 0) | 2310 | if (delta > 0) |
| 2302 | perf_adjust_period(event, period, delta); | 2311 | perf_adjust_period(event, period, delta); |
| 2303 | } | 2312 | } |
| 2304 | raw_spin_unlock(&ctx->lock); | ||
| 2305 | } | 2313 | } |
| 2306 | 2314 | ||
| 2307 | /* | 2315 | /* |
| @@ -2309,16 +2317,12 @@ static void perf_ctx_adjust_freq(struct perf_event_context *ctx, u64 period) | |||
| 2309 | */ | 2317 | */ |
| 2310 | static void rotate_ctx(struct perf_event_context *ctx) | 2318 | static void rotate_ctx(struct perf_event_context *ctx) |
| 2311 | { | 2319 | { |
| 2312 | raw_spin_lock(&ctx->lock); | ||
| 2313 | |||
| 2314 | /* | 2320 | /* |
| 2315 | * Rotate the first entry last of non-pinned groups. Rotation might be | 2321 | * Rotate the first entry last of non-pinned groups. Rotation might be |
| 2316 | * disabled by the inheritance code. | 2322 | * disabled by the inheritance code. |
| 2317 | */ | 2323 | */ |
| 2318 | if (!ctx->rotate_disable) | 2324 | if (!ctx->rotate_disable) |
| 2319 | list_rotate_left(&ctx->flexible_groups); | 2325 | list_rotate_left(&ctx->flexible_groups); |
| 2320 | |||
| 2321 | raw_spin_unlock(&ctx->lock); | ||
| 2322 | } | 2326 | } |
| 2323 | 2327 | ||
| 2324 | /* | 2328 | /* |
| @@ -2345,6 +2349,7 @@ static void perf_rotate_context(struct perf_cpu_context *cpuctx) | |||
| 2345 | rotate = 1; | 2349 | rotate = 1; |
| 2346 | } | 2350 | } |
| 2347 | 2351 | ||
| 2352 | perf_ctx_lock(cpuctx, cpuctx->task_ctx); | ||
| 2348 | perf_pmu_disable(cpuctx->ctx.pmu); | 2353 | perf_pmu_disable(cpuctx->ctx.pmu); |
| 2349 | perf_ctx_adjust_freq(&cpuctx->ctx, interval); | 2354 | perf_ctx_adjust_freq(&cpuctx->ctx, interval); |
| 2350 | if (ctx) | 2355 | if (ctx) |
| @@ -2355,21 +2360,20 @@ static void perf_rotate_context(struct perf_cpu_context *cpuctx) | |||
| 2355 | 2360 | ||
| 2356 | cpu_ctx_sched_out(cpuctx, EVENT_FLEXIBLE); | 2361 | cpu_ctx_sched_out(cpuctx, EVENT_FLEXIBLE); |
| 2357 | if (ctx) | 2362 | if (ctx) |
| 2358 | task_ctx_sched_out(ctx, EVENT_FLEXIBLE); | 2363 | ctx_sched_out(ctx, cpuctx, EVENT_FLEXIBLE); |
| 2359 | 2364 | ||
| 2360 | rotate_ctx(&cpuctx->ctx); | 2365 | rotate_ctx(&cpuctx->ctx); |
| 2361 | if (ctx) | 2366 | if (ctx) |
| 2362 | rotate_ctx(ctx); | 2367 | rotate_ctx(ctx); |
| 2363 | 2368 | ||
| 2364 | cpu_ctx_sched_in(cpuctx, EVENT_FLEXIBLE, current); | 2369 | perf_event_sched_in(cpuctx, ctx, current); |
| 2365 | if (ctx) | ||
| 2366 | task_ctx_sched_in(ctx, EVENT_FLEXIBLE); | ||
| 2367 | 2370 | ||
| 2368 | done: | 2371 | done: |
| 2369 | if (remove) | 2372 | if (remove) |
| 2370 | list_del_init(&cpuctx->rotation_list); | 2373 | list_del_init(&cpuctx->rotation_list); |
| 2371 | 2374 | ||
| 2372 | perf_pmu_enable(cpuctx->ctx.pmu); | 2375 | perf_pmu_enable(cpuctx->ctx.pmu); |
| 2376 | perf_ctx_unlock(cpuctx, cpuctx->task_ctx); | ||
| 2373 | } | 2377 | } |
| 2374 | 2378 | ||
| 2375 | void perf_event_task_tick(void) | 2379 | void perf_event_task_tick(void) |
| @@ -2424,9 +2428,9 @@ static void perf_event_enable_on_exec(struct perf_event_context *ctx) | |||
| 2424 | * in. | 2428 | * in. |
| 2425 | */ | 2429 | */ |
| 2426 | perf_cgroup_sched_out(current); | 2430 | perf_cgroup_sched_out(current); |
| 2427 | task_ctx_sched_out(ctx, EVENT_ALL); | ||
| 2428 | 2431 | ||
| 2429 | raw_spin_lock(&ctx->lock); | 2432 | raw_spin_lock(&ctx->lock); |
| 2433 | task_ctx_sched_out(ctx); | ||
| 2430 | 2434 | ||
| 2431 | list_for_each_entry(event, &ctx->pinned_groups, group_entry) { | 2435 | list_for_each_entry(event, &ctx->pinned_groups, group_entry) { |
| 2432 | ret = event_enable_on_exec(event, ctx); | 2436 | ret = event_enable_on_exec(event, ctx); |
| @@ -2835,16 +2839,12 @@ retry: | |||
| 2835 | unclone_ctx(ctx); | 2839 | unclone_ctx(ctx); |
| 2836 | ++ctx->pin_count; | 2840 | ++ctx->pin_count; |
| 2837 | raw_spin_unlock_irqrestore(&ctx->lock, flags); | 2841 | raw_spin_unlock_irqrestore(&ctx->lock, flags); |
| 2838 | } | 2842 | } else { |
| 2839 | |||
| 2840 | if (!ctx) { | ||
| 2841 | ctx = alloc_perf_context(pmu, task); | 2843 | ctx = alloc_perf_context(pmu, task); |
| 2842 | err = -ENOMEM; | 2844 | err = -ENOMEM; |
| 2843 | if (!ctx) | 2845 | if (!ctx) |
| 2844 | goto errout; | 2846 | goto errout; |
| 2845 | 2847 | ||
| 2846 | get_ctx(ctx); | ||
| 2847 | |||
| 2848 | err = 0; | 2848 | err = 0; |
| 2849 | mutex_lock(&task->perf_event_mutex); | 2849 | mutex_lock(&task->perf_event_mutex); |
| 2850 | /* | 2850 | /* |
| @@ -2856,14 +2856,14 @@ retry: | |||
| 2856 | else if (task->perf_event_ctxp[ctxn]) | 2856 | else if (task->perf_event_ctxp[ctxn]) |
| 2857 | err = -EAGAIN; | 2857 | err = -EAGAIN; |
| 2858 | else { | 2858 | else { |
| 2859 | get_ctx(ctx); | ||
| 2859 | ++ctx->pin_count; | 2860 | ++ctx->pin_count; |
| 2860 | rcu_assign_pointer(task->perf_event_ctxp[ctxn], ctx); | 2861 | rcu_assign_pointer(task->perf_event_ctxp[ctxn], ctx); |
| 2861 | } | 2862 | } |
| 2862 | mutex_unlock(&task->perf_event_mutex); | 2863 | mutex_unlock(&task->perf_event_mutex); |
| 2863 | 2864 | ||
| 2864 | if (unlikely(err)) { | 2865 | if (unlikely(err)) { |
| 2865 | put_task_struct(task); | 2866 | put_ctx(ctx); |
| 2866 | kfree(ctx); | ||
| 2867 | 2867 | ||
| 2868 | if (err == -EAGAIN) | 2868 | if (err == -EAGAIN) |
| 2869 | goto retry; | 2869 | goto retry; |
| @@ -2890,7 +2890,7 @@ static void free_event_rcu(struct rcu_head *head) | |||
| 2890 | kfree(event); | 2890 | kfree(event); |
| 2891 | } | 2891 | } |
| 2892 | 2892 | ||
| 2893 | static void perf_buffer_put(struct perf_buffer *buffer); | 2893 | static void ring_buffer_put(struct ring_buffer *rb); |
| 2894 | 2894 | ||
| 2895 | static void free_event(struct perf_event *event) | 2895 | static void free_event(struct perf_event *event) |
| 2896 | { | 2896 | { |
| @@ -2913,9 +2913,9 @@ static void free_event(struct perf_event *event) | |||
| 2913 | } | 2913 | } |
| 2914 | } | 2914 | } |
| 2915 | 2915 | ||
| 2916 | if (event->buffer) { | 2916 | if (event->rb) { |
| 2917 | perf_buffer_put(event->buffer); | 2917 | ring_buffer_put(event->rb); |
| 2918 | event->buffer = NULL; | 2918 | event->rb = NULL; |
| 2919 | } | 2919 | } |
| 2920 | 2920 | ||
| 2921 | if (is_cgroup_event(event)) | 2921 | if (is_cgroup_event(event)) |
| @@ -2934,12 +2934,6 @@ int perf_event_release_kernel(struct perf_event *event) | |||
| 2934 | { | 2934 | { |
| 2935 | struct perf_event_context *ctx = event->ctx; | 2935 | struct perf_event_context *ctx = event->ctx; |
| 2936 | 2936 | ||
| 2937 | /* | ||
| 2938 | * Remove from the PMU, can't get re-enabled since we got | ||
| 2939 | * here because the last ref went. | ||
| 2940 | */ | ||
| 2941 | perf_event_disable(event); | ||
| 2942 | |||
| 2943 | WARN_ON_ONCE(ctx->parent_ctx); | 2937 | WARN_ON_ONCE(ctx->parent_ctx); |
| 2944 | /* | 2938 | /* |
| 2945 | * There are two ways this annotation is useful: | 2939 | * There are two ways this annotation is useful: |
| @@ -2956,8 +2950,8 @@ int perf_event_release_kernel(struct perf_event *event) | |||
| 2956 | mutex_lock_nested(&ctx->mutex, SINGLE_DEPTH_NESTING); | 2950 | mutex_lock_nested(&ctx->mutex, SINGLE_DEPTH_NESTING); |
| 2957 | raw_spin_lock_irq(&ctx->lock); | 2951 | raw_spin_lock_irq(&ctx->lock); |
| 2958 | perf_group_detach(event); | 2952 | perf_group_detach(event); |
| 2959 | list_del_event(event, ctx); | ||
| 2960 | raw_spin_unlock_irq(&ctx->lock); | 2953 | raw_spin_unlock_irq(&ctx->lock); |
| 2954 | perf_remove_from_context(event); | ||
| 2961 | mutex_unlock(&ctx->mutex); | 2955 | mutex_unlock(&ctx->mutex); |
| 2962 | 2956 | ||
| 2963 | free_event(event); | 2957 | free_event(event); |
| @@ -3149,13 +3143,13 @@ perf_read(struct file *file, char __user *buf, size_t count, loff_t *ppos) | |||
| 3149 | static unsigned int perf_poll(struct file *file, poll_table *wait) | 3143 | static unsigned int perf_poll(struct file *file, poll_table *wait) |
| 3150 | { | 3144 | { |
| 3151 | struct perf_event *event = file->private_data; | 3145 | struct perf_event *event = file->private_data; |
| 3152 | struct perf_buffer *buffer; | 3146 | struct ring_buffer *rb; |
| 3153 | unsigned int events = POLL_HUP; | 3147 | unsigned int events = POLL_HUP; |
| 3154 | 3148 | ||
| 3155 | rcu_read_lock(); | 3149 | rcu_read_lock(); |
| 3156 | buffer = rcu_dereference(event->buffer); | 3150 | rb = rcu_dereference(event->rb); |
| 3157 | if (buffer) | 3151 | if (rb) |
| 3158 | events = atomic_xchg(&buffer->poll, 0); | 3152 | events = atomic_xchg(&rb->poll, 0); |
| 3159 | rcu_read_unlock(); | 3153 | rcu_read_unlock(); |
| 3160 | 3154 | ||
| 3161 | poll_wait(file, &event->waitq, wait); | 3155 | poll_wait(file, &event->waitq, wait); |
| @@ -3358,6 +3352,18 @@ static int perf_event_index(struct perf_event *event) | |||
| 3358 | return event->hw.idx + 1 - PERF_EVENT_INDEX_OFFSET; | 3352 | return event->hw.idx + 1 - PERF_EVENT_INDEX_OFFSET; |
| 3359 | } | 3353 | } |
| 3360 | 3354 | ||
| 3355 | static void calc_timer_values(struct perf_event *event, | ||
| 3356 | u64 *running, | ||
| 3357 | u64 *enabled) | ||
| 3358 | { | ||
| 3359 | u64 now, ctx_time; | ||
| 3360 | |||
| 3361 | now = perf_clock(); | ||
| 3362 | ctx_time = event->shadow_ctx_time + now; | ||
| 3363 | *enabled = ctx_time - event->tstamp_enabled; | ||
| 3364 | *running = ctx_time - event->tstamp_running; | ||
| 3365 | } | ||
| 3366 | |||
| 3361 | /* | 3367 | /* |
| 3362 | * Callers need to ensure there can be no nesting of this function, otherwise | 3368 | * Callers need to ensure there can be no nesting of this function, otherwise |
| 3363 | * the seqlock logic goes bad. We can not serialize this because the arch | 3369 | * the seqlock logic goes bad. We can not serialize this because the arch |
| @@ -3366,14 +3372,25 @@ static int perf_event_index(struct perf_event *event) | |||
| 3366 | void perf_event_update_userpage(struct perf_event *event) | 3372 | void perf_event_update_userpage(struct perf_event *event) |
| 3367 | { | 3373 | { |
| 3368 | struct perf_event_mmap_page *userpg; | 3374 | struct perf_event_mmap_page *userpg; |
| 3369 | struct perf_buffer *buffer; | 3375 | struct ring_buffer *rb; |
| 3376 | u64 enabled, running; | ||
| 3370 | 3377 | ||
| 3371 | rcu_read_lock(); | 3378 | rcu_read_lock(); |
| 3372 | buffer = rcu_dereference(event->buffer); | 3379 | /* |
| 3373 | if (!buffer) | 3380 | * compute total_time_enabled, total_time_running |
| 3381 | * based on snapshot values taken when the event | ||
| 3382 | * was last scheduled in. | ||
| 3383 | * | ||
| 3384 | * we cannot simply called update_context_time() | ||
| 3385 | * because of locking issue as we can be called in | ||
| 3386 | * NMI context | ||
| 3387 | */ | ||
| 3388 | calc_timer_values(event, &enabled, &running); | ||
| 3389 | rb = rcu_dereference(event->rb); | ||
| 3390 | if (!rb) | ||
| 3374 | goto unlock; | 3391 | goto unlock; |
| 3375 | 3392 | ||
| 3376 | userpg = buffer->user_page; | 3393 | userpg = rb->user_page; |
| 3377 | 3394 | ||
| 3378 | /* | 3395 | /* |
| 3379 | * Disable preemption so as to not let the corresponding user-space | 3396 | * Disable preemption so as to not let the corresponding user-space |
| @@ -3387,10 +3404,10 @@ void perf_event_update_userpage(struct perf_event *event) | |||
| 3387 | if (event->state == PERF_EVENT_STATE_ACTIVE) | 3404 | if (event->state == PERF_EVENT_STATE_ACTIVE) |
| 3388 | userpg->offset -= local64_read(&event->hw.prev_count); | 3405 | userpg->offset -= local64_read(&event->hw.prev_count); |
| 3389 | 3406 | ||
| 3390 | userpg->time_enabled = event->total_time_enabled + | 3407 | userpg->time_enabled = enabled + |
| 3391 | atomic64_read(&event->child_total_time_enabled); | 3408 | atomic64_read(&event->child_total_time_enabled); |
| 3392 | 3409 | ||
| 3393 | userpg->time_running = event->total_time_running + | 3410 | userpg->time_running = running + |
| 3394 | atomic64_read(&event->child_total_time_running); | 3411 | atomic64_read(&event->child_total_time_running); |
| 3395 | 3412 | ||
| 3396 | barrier(); | 3413 | barrier(); |
| @@ -3400,220 +3417,10 @@ unlock: | |||
| 3400 | rcu_read_unlock(); | 3417 | rcu_read_unlock(); |
| 3401 | } | 3418 | } |
| 3402 | 3419 | ||
| 3403 | static unsigned long perf_data_size(struct perf_buffer *buffer); | ||
| 3404 | |||
| 3405 | static void | ||
| 3406 | perf_buffer_init(struct perf_buffer *buffer, long watermark, int flags) | ||
| 3407 | { | ||
| 3408 | long max_size = perf_data_size(buffer); | ||
| 3409 | |||
| 3410 | if (watermark) | ||
| 3411 | buffer->watermark = min(max_size, watermark); | ||
| 3412 | |||
| 3413 | if (!buffer->watermark) | ||
| 3414 | buffer->watermark = max_size / 2; | ||
| 3415 | |||
| 3416 | if (flags & PERF_BUFFER_WRITABLE) | ||
| 3417 | buffer->writable = 1; | ||
| 3418 | |||
| 3419 | atomic_set(&buffer->refcount, 1); | ||
| 3420 | } | ||
| 3421 | |||
| 3422 | #ifndef CONFIG_PERF_USE_VMALLOC | ||
| 3423 | |||
| 3424 | /* | ||
| 3425 | * Back perf_mmap() with regular GFP_KERNEL-0 pages. | ||
| 3426 | */ | ||
| 3427 | |||
| 3428 | static struct page * | ||
| 3429 | perf_mmap_to_page(struct perf_buffer *buffer, unsigned long pgoff) | ||
| 3430 | { | ||
| 3431 | if (pgoff > buffer->nr_pages) | ||
| 3432 | return NULL; | ||
| 3433 | |||
| 3434 | if (pgoff == 0) | ||
| 3435 | return virt_to_page(buffer->user_page); | ||
| 3436 | |||
| 3437 | return virt_to_page(buffer->data_pages[pgoff - 1]); | ||
| 3438 | } | ||
| 3439 | |||
| 3440 | static void *perf_mmap_alloc_page(int cpu) | ||
| 3441 | { | ||
| 3442 | struct page *page; | ||
| 3443 | int node; | ||
| 3444 | |||
| 3445 | node = (cpu == -1) ? cpu : cpu_to_node(cpu); | ||
| 3446 | page = alloc_pages_node(node, GFP_KERNEL | __GFP_ZERO, 0); | ||
| 3447 | if (!page) | ||
| 3448 | return NULL; | ||
| 3449 | |||
| 3450 | return page_address(page); | ||
| 3451 | } | ||
| 3452 | |||
| 3453 | static struct perf_buffer * | ||
| 3454 | perf_buffer_alloc(int nr_pages, long watermark, int cpu, int flags) | ||
| 3455 | { | ||
| 3456 | struct perf_buffer *buffer; | ||
| 3457 | unsigned long size; | ||
| 3458 | int i; | ||
| 3459 | |||
| 3460 | size = sizeof(struct perf_buffer); | ||
| 3461 | size += nr_pages * sizeof(void *); | ||
| 3462 | |||
| 3463 | buffer = kzalloc(size, GFP_KERNEL); | ||
| 3464 | if (!buffer) | ||
| 3465 | goto fail; | ||
| 3466 | |||
| 3467 | buffer->user_page = perf_mmap_alloc_page(cpu); | ||
| 3468 | if (!buffer->user_page) | ||
| 3469 | goto fail_user_page; | ||
| 3470 | |||
| 3471 | for (i = 0; i < nr_pages; i++) { | ||
| 3472 | buffer->data_pages[i] = perf_mmap_alloc_page(cpu); | ||
| 3473 | if (!buffer->data_pages[i]) | ||
| 3474 | goto fail_data_pages; | ||
| 3475 | } | ||
| 3476 | |||
| 3477 | buffer->nr_pages = nr_pages; | ||
| 3478 | |||
| 3479 | perf_buffer_init(buffer, watermark, flags); | ||
| 3480 | |||
| 3481 | return buffer; | ||
| 3482 | |||
| 3483 | fail_data_pages: | ||
| 3484 | for (i--; i >= 0; i--) | ||
| 3485 | free_page((unsigned long)buffer->data_pages[i]); | ||
| 3486 | |||
| 3487 | free_page((unsigned long)buffer->user_page); | ||
| 3488 | |||
| 3489 | fail_user_page: | ||
| 3490 | kfree(buffer); | ||
| 3491 | |||
| 3492 | fail: | ||
| 3493 | return NULL; | ||
| 3494 | } | ||
| 3495 | |||
| 3496 | static void perf_mmap_free_page(unsigned long addr) | ||
| 3497 | { | ||
| 3498 | struct page *page = virt_to_page((void *)addr); | ||
| 3499 | |||
| 3500 | page->mapping = NULL; | ||
| 3501 | __free_page(page); | ||
| 3502 | } | ||
| 3503 | |||
| 3504 | static void perf_buffer_free(struct perf_buffer *buffer) | ||
| 3505 | { | ||
| 3506 | int i; | ||
| 3507 | |||
| 3508 | perf_mmap_free_page((unsigned long)buffer->user_page); | ||
| 3509 | for (i = 0; i < buffer->nr_pages; i++) | ||
| 3510 | perf_mmap_free_page((unsigned long)buffer->data_pages[i]); | ||
| 3511 | kfree(buffer); | ||
| 3512 | } | ||
| 3513 | |||
| 3514 | static inline int page_order(struct perf_buffer *buffer) | ||
| 3515 | { | ||
| 3516 | return 0; | ||
| 3517 | } | ||
| 3518 | |||
| 3519 | #else | ||
| 3520 | |||
| 3521 | /* | ||
| 3522 | * Back perf_mmap() with vmalloc memory. | ||
| 3523 | * | ||
| 3524 | * Required for architectures that have d-cache aliasing issues. | ||
| 3525 | */ | ||
| 3526 | |||
| 3527 | static inline int page_order(struct perf_buffer *buffer) | ||
| 3528 | { | ||
| 3529 | return buffer->page_order; | ||
| 3530 | } | ||
| 3531 | |||
| 3532 | static struct page * | ||
| 3533 | perf_mmap_to_page(struct perf_buffer *buffer, unsigned long pgoff) | ||
| 3534 | { | ||
| 3535 | if (pgoff > (1UL << page_order(buffer))) | ||
| 3536 | return NULL; | ||
| 3537 | |||
| 3538 | return vmalloc_to_page((void *)buffer->user_page + pgoff * PAGE_SIZE); | ||
| 3539 | } | ||
| 3540 | |||
| 3541 | static void perf_mmap_unmark_page(void *addr) | ||
| 3542 | { | ||
| 3543 | struct page *page = vmalloc_to_page(addr); | ||
| 3544 | |||
| 3545 | page->mapping = NULL; | ||
| 3546 | } | ||
| 3547 | |||
| 3548 | static void perf_buffer_free_work(struct work_struct *work) | ||
| 3549 | { | ||
| 3550 | struct perf_buffer *buffer; | ||
| 3551 | void *base; | ||
| 3552 | int i, nr; | ||
| 3553 | |||
| 3554 | buffer = container_of(work, struct perf_buffer, work); | ||
| 3555 | nr = 1 << page_order(buffer); | ||
| 3556 | |||
| 3557 | base = buffer->user_page; | ||
| 3558 | for (i = 0; i < nr + 1; i++) | ||
| 3559 | perf_mmap_unmark_page(base + (i * PAGE_SIZE)); | ||
| 3560 | |||
| 3561 | vfree(base); | ||
| 3562 | kfree(buffer); | ||
| 3563 | } | ||
| 3564 | |||
| 3565 | static void perf_buffer_free(struct perf_buffer *buffer) | ||
| 3566 | { | ||
| 3567 | schedule_work(&buffer->work); | ||
| 3568 | } | ||
| 3569 | |||
| 3570 | static struct perf_buffer * | ||
| 3571 | perf_buffer_alloc(int nr_pages, long watermark, int cpu, int flags) | ||
| 3572 | { | ||
| 3573 | struct perf_buffer *buffer; | ||
| 3574 | unsigned long size; | ||
| 3575 | void *all_buf; | ||
| 3576 | |||
| 3577 | size = sizeof(struct perf_buffer); | ||
| 3578 | size += sizeof(void *); | ||
| 3579 | |||
| 3580 | buffer = kzalloc(size, GFP_KERNEL); | ||
| 3581 | if (!buffer) | ||
| 3582 | goto fail; | ||
| 3583 | |||
| 3584 | INIT_WORK(&buffer->work, perf_buffer_free_work); | ||
| 3585 | |||
| 3586 | all_buf = vmalloc_user((nr_pages + 1) * PAGE_SIZE); | ||
| 3587 | if (!all_buf) | ||
| 3588 | goto fail_all_buf; | ||
| 3589 | |||
| 3590 | buffer->user_page = all_buf; | ||
| 3591 | buffer->data_pages[0] = all_buf + PAGE_SIZE; | ||
| 3592 | buffer->page_order = ilog2(nr_pages); | ||
| 3593 | buffer->nr_pages = 1; | ||
| 3594 | |||
| 3595 | perf_buffer_init(buffer, watermark, flags); | ||
| 3596 | |||
| 3597 | return buffer; | ||
| 3598 | |||
| 3599 | fail_all_buf: | ||
| 3600 | kfree(buffer); | ||
| 3601 | |||
| 3602 | fail: | ||
| 3603 | return NULL; | ||
| 3604 | } | ||
| 3605 | |||
| 3606 | #endif | ||
| 3607 | |||
| 3608 | static unsigned long perf_data_size(struct perf_buffer *buffer) | ||
| 3609 | { | ||
| 3610 | return buffer->nr_pages << (PAGE_SHIFT + page_order(buffer)); | ||
| 3611 | } | ||
| 3612 | |||
| 3613 | static int perf_mmap_fault(struct vm_area_struct *vma, struct vm_fault *vmf) | 3420 | static int perf_mmap_fault(struct vm_area_struct *vma, struct vm_fault *vmf) |
| 3614 | { | 3421 | { |
| 3615 | struct perf_event *event = vma->vm_file->private_data; | 3422 | struct perf_event *event = vma->vm_file->private_data; |
| 3616 | struct perf_buffer *buffer; | 3423 | struct ring_buffer *rb; |
| 3617 | int ret = VM_FAULT_SIGBUS; | 3424 | int ret = VM_FAULT_SIGBUS; |
| 3618 | 3425 | ||
| 3619 | if (vmf->flags & FAULT_FLAG_MKWRITE) { | 3426 | if (vmf->flags & FAULT_FLAG_MKWRITE) { |
| @@ -3623,14 +3430,14 @@ static int perf_mmap_fault(struct vm_area_struct *vma, struct vm_fault *vmf) | |||
| 3623 | } | 3430 | } |
| 3624 | 3431 | ||
| 3625 | rcu_read_lock(); | 3432 | rcu_read_lock(); |
| 3626 | buffer = rcu_dereference(event->buffer); | 3433 | rb = rcu_dereference(event->rb); |
| 3627 | if (!buffer) | 3434 | if (!rb) |
| 3628 | goto unlock; | 3435 | goto unlock; |
| 3629 | 3436 | ||
| 3630 | if (vmf->pgoff && (vmf->flags & FAULT_FLAG_WRITE)) | 3437 | if (vmf->pgoff && (vmf->flags & FAULT_FLAG_WRITE)) |
| 3631 | goto unlock; | 3438 | goto unlock; |
| 3632 | 3439 | ||
| 3633 | vmf->page = perf_mmap_to_page(buffer, vmf->pgoff); | 3440 | vmf->page = perf_mmap_to_page(rb, vmf->pgoff); |
| 3634 | if (!vmf->page) | 3441 | if (!vmf->page) |
| 3635 | goto unlock; | 3442 | goto unlock; |
| 3636 | 3443 | ||
| @@ -3645,35 +3452,35 @@ unlock: | |||
| 3645 | return ret; | 3452 | return ret; |
| 3646 | } | 3453 | } |
| 3647 | 3454 | ||
| 3648 | static void perf_buffer_free_rcu(struct rcu_head *rcu_head) | 3455 | static void rb_free_rcu(struct rcu_head *rcu_head) |
| 3649 | { | 3456 | { |
| 3650 | struct perf_buffer *buffer; | 3457 | struct ring_buffer *rb; |
| 3651 | 3458 | ||
| 3652 | buffer = container_of(rcu_head, struct perf_buffer, rcu_head); | 3459 | rb = container_of(rcu_head, struct ring_buffer, rcu_head); |
| 3653 | perf_buffer_free(buffer); | 3460 | rb_free(rb); |
| 3654 | } | 3461 | } |
| 3655 | 3462 | ||
| 3656 | static struct perf_buffer *perf_buffer_get(struct perf_event *event) | 3463 | static struct ring_buffer *ring_buffer_get(struct perf_event *event) |
| 3657 | { | 3464 | { |
| 3658 | struct perf_buffer *buffer; | 3465 | struct ring_buffer *rb; |
| 3659 | 3466 | ||
| 3660 | rcu_read_lock(); | 3467 | rcu_read_lock(); |
| 3661 | buffer = rcu_dereference(event->buffer); | 3468 | rb = rcu_dereference(event->rb); |
| 3662 | if (buffer) { | 3469 | if (rb) { |
| 3663 | if (!atomic_inc_not_zero(&buffer->refcount)) | 3470 | if (!atomic_inc_not_zero(&rb->refcount)) |
| 3664 | buffer = NULL; | 3471 | rb = NULL; |
| 3665 | } | 3472 | } |
| 3666 | rcu_read_unlock(); | 3473 | rcu_read_unlock(); |
| 3667 | 3474 | ||
| 3668 | return buffer; | 3475 | return rb; |
| 3669 | } | 3476 | } |
| 3670 | 3477 | ||
| 3671 | static void perf_buffer_put(struct perf_buffer *buffer) | 3478 | static void ring_buffer_put(struct ring_buffer *rb) |
| 3672 | { | 3479 | { |
| 3673 | if (!atomic_dec_and_test(&buffer->refcount)) | 3480 | if (!atomic_dec_and_test(&rb->refcount)) |
| 3674 | return; | 3481 | return; |
| 3675 | 3482 | ||
| 3676 | call_rcu(&buffer->rcu_head, perf_buffer_free_rcu); | 3483 | call_rcu(&rb->rcu_head, rb_free_rcu); |
| 3677 | } | 3484 | } |
| 3678 | 3485 | ||
| 3679 | static void perf_mmap_open(struct vm_area_struct *vma) | 3486 | static void perf_mmap_open(struct vm_area_struct *vma) |
| @@ -3688,16 +3495,16 @@ static void perf_mmap_close(struct vm_area_struct *vma) | |||
| 3688 | struct perf_event *event = vma->vm_file->private_data; | 3495 | struct perf_event *event = vma->vm_file->private_data; |
| 3689 | 3496 | ||
| 3690 | if (atomic_dec_and_mutex_lock(&event->mmap_count, &event->mmap_mutex)) { | 3497 | if (atomic_dec_and_mutex_lock(&event->mmap_count, &event->mmap_mutex)) { |
| 3691 | unsigned long size = perf_data_size(event->buffer); | 3498 | unsigned long size = perf_data_size(event->rb); |
| 3692 | struct user_struct *user = event->mmap_user; | 3499 | struct user_struct *user = event->mmap_user; |
| 3693 | struct perf_buffer *buffer = event->buffer; | 3500 | struct ring_buffer *rb = event->rb; |
| 3694 | 3501 | ||
| 3695 | atomic_long_sub((size >> PAGE_SHIFT) + 1, &user->locked_vm); | 3502 | atomic_long_sub((size >> PAGE_SHIFT) + 1, &user->locked_vm); |
| 3696 | vma->vm_mm->locked_vm -= event->mmap_locked; | 3503 | vma->vm_mm->locked_vm -= event->mmap_locked; |
| 3697 | rcu_assign_pointer(event->buffer, NULL); | 3504 | rcu_assign_pointer(event->rb, NULL); |
| 3698 | mutex_unlock(&event->mmap_mutex); | 3505 | mutex_unlock(&event->mmap_mutex); |
| 3699 | 3506 | ||
| 3700 | perf_buffer_put(buffer); | 3507 | ring_buffer_put(rb); |
| 3701 | free_uid(user); | 3508 | free_uid(user); |
| 3702 | } | 3509 | } |
| 3703 | } | 3510 | } |
| @@ -3715,7 +3522,7 @@ static int perf_mmap(struct file *file, struct vm_area_struct *vma) | |||
| 3715 | unsigned long user_locked, user_lock_limit; | 3522 | unsigned long user_locked, user_lock_limit; |
| 3716 | struct user_struct *user = current_user(); | 3523 | struct user_struct *user = current_user(); |
| 3717 | unsigned long locked, lock_limit; | 3524 | unsigned long locked, lock_limit; |
| 3718 | struct perf_buffer *buffer; | 3525 | struct ring_buffer *rb; |
| 3719 | unsigned long vma_size; | 3526 | unsigned long vma_size; |
| 3720 | unsigned long nr_pages; | 3527 | unsigned long nr_pages; |
| 3721 | long user_extra, extra; | 3528 | long user_extra, extra; |
| @@ -3724,7 +3531,7 @@ static int perf_mmap(struct file *file, struct vm_area_struct *vma) | |||
| 3724 | /* | 3531 | /* |
| 3725 | * Don't allow mmap() of inherited per-task counters. This would | 3532 | * Don't allow mmap() of inherited per-task counters. This would |
| 3726 | * create a performance issue due to all children writing to the | 3533 | * create a performance issue due to all children writing to the |
| 3727 | * same buffer. | 3534 | * same rb. |
| 3728 | */ | 3535 | */ |
| 3729 | if (event->cpu == -1 && event->attr.inherit) | 3536 | if (event->cpu == -1 && event->attr.inherit) |
| 3730 | return -EINVAL; | 3537 | return -EINVAL; |
| @@ -3736,7 +3543,7 @@ static int perf_mmap(struct file *file, struct vm_area_struct *vma) | |||
| 3736 | nr_pages = (vma_size / PAGE_SIZE) - 1; | 3543 | nr_pages = (vma_size / PAGE_SIZE) - 1; |
| 3737 | 3544 | ||
| 3738 | /* | 3545 | /* |
| 3739 | * If we have buffer pages ensure they're a power-of-two number, so we | 3546 | * If we have rb pages ensure they're a power-of-two number, so we |
| 3740 | * can do bitmasks instead of modulo. | 3547 | * can do bitmasks instead of modulo. |
| 3741 | */ | 3548 | */ |
| 3742 | if (nr_pages != 0 && !is_power_of_2(nr_pages)) | 3549 | if (nr_pages != 0 && !is_power_of_2(nr_pages)) |
| @@ -3750,9 +3557,9 @@ static int perf_mmap(struct file *file, struct vm_area_struct *vma) | |||
| 3750 | 3557 | ||
| 3751 | WARN_ON_ONCE(event->ctx->parent_ctx); | 3558 | WARN_ON_ONCE(event->ctx->parent_ctx); |
| 3752 | mutex_lock(&event->mmap_mutex); | 3559 | mutex_lock(&event->mmap_mutex); |
| 3753 | if (event->buffer) { | 3560 | if (event->rb) { |
| 3754 | if (event->buffer->nr_pages == nr_pages) | 3561 | if (event->rb->nr_pages == nr_pages) |
| 3755 | atomic_inc(&event->buffer->refcount); | 3562 | atomic_inc(&event->rb->refcount); |
| 3756 | else | 3563 | else |
| 3757 | ret = -EINVAL; | 3564 | ret = -EINVAL; |
| 3758 | goto unlock; | 3565 | goto unlock; |
| @@ -3782,18 +3589,20 @@ static int perf_mmap(struct file *file, struct vm_area_struct *vma) | |||
| 3782 | goto unlock; | 3589 | goto unlock; |
| 3783 | } | 3590 | } |
| 3784 | 3591 | ||
| 3785 | WARN_ON(event->buffer); | 3592 | WARN_ON(event->rb); |
| 3786 | 3593 | ||
| 3787 | if (vma->vm_flags & VM_WRITE) | 3594 | if (vma->vm_flags & VM_WRITE) |
| 3788 | flags |= PERF_BUFFER_WRITABLE; | 3595 | flags |= RING_BUFFER_WRITABLE; |
| 3789 | 3596 | ||
| 3790 | buffer = perf_buffer_alloc(nr_pages, event->attr.wakeup_watermark, | 3597 | rb = rb_alloc(nr_pages, |
| 3791 | event->cpu, flags); | 3598 | event->attr.watermark ? event->attr.wakeup_watermark : 0, |
| 3792 | if (!buffer) { | 3599 | event->cpu, flags); |
| 3600 | |||
| 3601 | if (!rb) { | ||
| 3793 | ret = -ENOMEM; | 3602 | ret = -ENOMEM; |
| 3794 | goto unlock; | 3603 | goto unlock; |
| 3795 | } | 3604 | } |
| 3796 | rcu_assign_pointer(event->buffer, buffer); | 3605 | rcu_assign_pointer(event->rb, rb); |
| 3797 | 3606 | ||
| 3798 | atomic_long_add(user_extra, &user->locked_vm); | 3607 | atomic_long_add(user_extra, &user->locked_vm); |
| 3799 | event->mmap_locked = extra; | 3608 | event->mmap_locked = extra; |
| @@ -3892,117 +3701,6 @@ int perf_unregister_guest_info_callbacks(struct perf_guest_info_callbacks *cbs) | |||
| 3892 | } | 3701 | } |
| 3893 | EXPORT_SYMBOL_GPL(perf_unregister_guest_info_callbacks); | 3702 | EXPORT_SYMBOL_GPL(perf_unregister_guest_info_callbacks); |
| 3894 | 3703 | ||
| 3895 | /* | ||
| 3896 | * Output | ||
| 3897 | */ | ||
| 3898 | static bool perf_output_space(struct perf_buffer *buffer, unsigned long tail, | ||
| 3899 | unsigned long offset, unsigned long head) | ||
| 3900 | { | ||
| 3901 | unsigned long mask; | ||
| 3902 | |||
| 3903 | if (!buffer->writable) | ||
| 3904 | return true; | ||
| 3905 | |||
| 3906 | mask = perf_data_size(buffer) - 1; | ||
| 3907 | |||
| 3908 | offset = (offset - tail) & mask; | ||
| 3909 | head = (head - tail) & mask; | ||
| 3910 | |||
| 3911 | if ((int)(head - offset) < 0) | ||
| 3912 | return false; | ||
| 3913 | |||
| 3914 | return true; | ||
| 3915 | } | ||
| 3916 | |||
| 3917 | static void perf_output_wakeup(struct perf_output_handle *handle) | ||
| 3918 | { | ||
| 3919 | atomic_set(&handle->buffer->poll, POLL_IN); | ||
| 3920 | |||
| 3921 | if (handle->nmi) { | ||
| 3922 | handle->event->pending_wakeup = 1; | ||
| 3923 | irq_work_queue(&handle->event->pending); | ||
| 3924 | } else | ||
| 3925 | perf_event_wakeup(handle->event); | ||
| 3926 | } | ||
| 3927 | |||
| 3928 | /* | ||
| 3929 | * We need to ensure a later event_id doesn't publish a head when a former | ||
| 3930 | * event isn't done writing. However since we need to deal with NMIs we | ||
| 3931 | * cannot fully serialize things. | ||
| 3932 | * | ||
| 3933 | * We only publish the head (and generate a wakeup) when the outer-most | ||
| 3934 | * event completes. | ||
| 3935 | */ | ||
| 3936 | static void perf_output_get_handle(struct perf_output_handle *handle) | ||
| 3937 | { | ||
| 3938 | struct perf_buffer *buffer = handle->buffer; | ||
| 3939 | |||
| 3940 | preempt_disable(); | ||
| 3941 | local_inc(&buffer->nest); | ||
| 3942 | handle->wakeup = local_read(&buffer->wakeup); | ||
| 3943 | } | ||
| 3944 | |||
| 3945 | static void perf_output_put_handle(struct perf_output_handle *handle) | ||
| 3946 | { | ||
| 3947 | struct perf_buffer *buffer = handle->buffer; | ||
| 3948 | unsigned long head; | ||
| 3949 | |||
| 3950 | again: | ||
| 3951 | head = local_read(&buffer->head); | ||
| 3952 | |||
| 3953 | /* | ||
| 3954 | * IRQ/NMI can happen here, which means we can miss a head update. | ||
| 3955 | */ | ||
| 3956 | |||
| 3957 | if (!local_dec_and_test(&buffer->nest)) | ||
| 3958 | goto out; | ||
| 3959 | |||
| 3960 | /* | ||
| 3961 | * Publish the known good head. Rely on the full barrier implied | ||
| 3962 | * by atomic_dec_and_test() order the buffer->head read and this | ||
| 3963 | * write. | ||
| 3964 | */ | ||
| 3965 | buffer->user_page->data_head = head; | ||
| 3966 | |||
| 3967 | /* | ||
| 3968 | * Now check if we missed an update, rely on the (compiler) | ||
| 3969 | * barrier in atomic_dec_and_test() to re-read buffer->head. | ||
| 3970 | */ | ||
| 3971 | if (unlikely(head != local_read(&buffer->head))) { | ||
| 3972 | local_inc(&buffer->nest); | ||
| 3973 | goto again; | ||
| 3974 | } | ||
| 3975 | |||
| 3976 | if (handle->wakeup != local_read(&buffer->wakeup)) | ||
| 3977 | perf_output_wakeup(handle); | ||
| 3978 | |||
| 3979 | out: | ||
| 3980 | preempt_enable(); | ||
| 3981 | } | ||
| 3982 | |||
| 3983 | __always_inline void perf_output_copy(struct perf_output_handle *handle, | ||
| 3984 | const void *buf, unsigned int len) | ||
| 3985 | { | ||
| 3986 | do { | ||
| 3987 | unsigned long size = min_t(unsigned long, handle->size, len); | ||
| 3988 | |||
| 3989 | memcpy(handle->addr, buf, size); | ||
| 3990 | |||
| 3991 | len -= size; | ||
| 3992 | handle->addr += size; | ||
| 3993 | buf += size; | ||
| 3994 | handle->size -= size; | ||
| 3995 | if (!handle->size) { | ||
| 3996 | struct perf_buffer *buffer = handle->buffer; | ||
| 3997 | |||
| 3998 | handle->page++; | ||
| 3999 | handle->page &= buffer->nr_pages - 1; | ||
| 4000 | handle->addr = buffer->data_pages[handle->page]; | ||
| 4001 | handle->size = PAGE_SIZE << page_order(buffer); | ||
| 4002 | } | ||
| 4003 | } while (len); | ||
| 4004 | } | ||
| 4005 | |||
| 4006 | static void __perf_event_header__init_id(struct perf_event_header *header, | 3704 | static void __perf_event_header__init_id(struct perf_event_header *header, |
| 4007 | struct perf_sample_data *data, | 3705 | struct perf_sample_data *data, |
| 4008 | struct perf_event *event) | 3706 | struct perf_event *event) |
| @@ -4033,9 +3731,9 @@ static void __perf_event_header__init_id(struct perf_event_header *header, | |||
| 4033 | } | 3731 | } |
| 4034 | } | 3732 | } |
| 4035 | 3733 | ||
| 4036 | static void perf_event_header__init_id(struct perf_event_header *header, | 3734 | void perf_event_header__init_id(struct perf_event_header *header, |
| 4037 | struct perf_sample_data *data, | 3735 | struct perf_sample_data *data, |
| 4038 | struct perf_event *event) | 3736 | struct perf_event *event) |
| 4039 | { | 3737 | { |
| 4040 | if (event->attr.sample_id_all) | 3738 | if (event->attr.sample_id_all) |
| 4041 | __perf_event_header__init_id(header, data, event); | 3739 | __perf_event_header__init_id(header, data, event); |
| @@ -4062,121 +3760,14 @@ static void __perf_event__output_id_sample(struct perf_output_handle *handle, | |||
| 4062 | perf_output_put(handle, data->cpu_entry); | 3760 | perf_output_put(handle, data->cpu_entry); |
| 4063 | } | 3761 | } |
| 4064 | 3762 | ||
| 4065 | static void perf_event__output_id_sample(struct perf_event *event, | 3763 | void perf_event__output_id_sample(struct perf_event *event, |
| 4066 | struct perf_output_handle *handle, | 3764 | struct perf_output_handle *handle, |
| 4067 | struct perf_sample_data *sample) | 3765 | struct perf_sample_data *sample) |
| 4068 | { | 3766 | { |
| 4069 | if (event->attr.sample_id_all) | 3767 | if (event->attr.sample_id_all) |
| 4070 | __perf_event__output_id_sample(handle, sample); | 3768 | __perf_event__output_id_sample(handle, sample); |
| 4071 | } | 3769 | } |
| 4072 | 3770 | ||
| 4073 | int perf_output_begin(struct perf_output_handle *handle, | ||
| 4074 | struct perf_event *event, unsigned int size, | ||
| 4075 | int nmi, int sample) | ||
| 4076 | { | ||
| 4077 | struct perf_buffer *buffer; | ||
| 4078 | unsigned long tail, offset, head; | ||
| 4079 | int have_lost; | ||
| 4080 | struct perf_sample_data sample_data; | ||
| 4081 | struct { | ||
| 4082 | struct perf_event_header header; | ||
| 4083 | u64 id; | ||
| 4084 | u64 lost; | ||
| 4085 | } lost_event; | ||
| 4086 | |||
| 4087 | rcu_read_lock(); | ||
| 4088 | /* | ||
| 4089 | * For inherited events we send all the output towards the parent. | ||
| 4090 | */ | ||
| 4091 | if (event->parent) | ||
| 4092 | event = event->parent; | ||
| 4093 | |||
| 4094 | buffer = rcu_dereference(event->buffer); | ||
| 4095 | if (!buffer) | ||
| 4096 | goto out; | ||
| 4097 | |||
| 4098 | handle->buffer = buffer; | ||
| 4099 | handle->event = event; | ||
| 4100 | handle->nmi = nmi; | ||
| 4101 | handle->sample = sample; | ||
| 4102 | |||
| 4103 | if (!buffer->nr_pages) | ||
| 4104 | goto out; | ||
| 4105 | |||
| 4106 | have_lost = local_read(&buffer->lost); | ||
| 4107 | if (have_lost) { | ||
| 4108 | lost_event.header.size = sizeof(lost_event); | ||
| 4109 | perf_event_header__init_id(&lost_event.header, &sample_data, | ||
| 4110 | event); | ||
| 4111 | size += lost_event.header.size; | ||
| 4112 | } | ||
| 4113 | |||
| 4114 | perf_output_get_handle(handle); | ||
| 4115 | |||
| 4116 | do { | ||
| 4117 | /* | ||
| 4118 | * Userspace could choose to issue a mb() before updating the | ||
| 4119 | * tail pointer. So that all reads will be completed before the | ||
| 4120 | * write is issued. | ||
| 4121 | */ | ||
| 4122 | tail = ACCESS_ONCE(buffer->user_page->data_tail); | ||
| 4123 | smp_rmb(); | ||
| 4124 | offset = head = local_read(&buffer->head); | ||
| 4125 | head += size; | ||
| 4126 | if (unlikely(!perf_output_space(buffer, tail, offset, head))) | ||
| 4127 | goto fail; | ||
| 4128 | } while (local_cmpxchg(&buffer->head, offset, head) != offset); | ||
| 4129 | |||
| 4130 | if (head - local_read(&buffer->wakeup) > buffer->watermark) | ||
| 4131 | local_add(buffer->watermark, &buffer->wakeup); | ||
| 4132 | |||
| 4133 | handle->page = offset >> (PAGE_SHIFT + page_order(buffer)); | ||
| 4134 | handle->page &= buffer->nr_pages - 1; | ||
| 4135 | handle->size = offset & ((PAGE_SIZE << page_order(buffer)) - 1); | ||
| 4136 | handle->addr = buffer->data_pages[handle->page]; | ||
| 4137 | handle->addr += handle->size; | ||
| 4138 | handle->size = (PAGE_SIZE << page_order(buffer)) - handle->size; | ||
| 4139 | |||
| 4140 | if (have_lost) { | ||
| 4141 | lost_event.header.type = PERF_RECORD_LOST; | ||
| 4142 | lost_event.header.misc = 0; | ||
| 4143 | lost_event.id = event->id; | ||
| 4144 | lost_event.lost = local_xchg(&buffer->lost, 0); | ||
| 4145 | |||
| 4146 | perf_output_put(handle, lost_event); | ||
| 4147 | perf_event__output_id_sample(event, handle, &sample_data); | ||
| 4148 | } | ||
| 4149 | |||
| 4150 | return 0; | ||
| 4151 | |||
| 4152 | fail: | ||
| 4153 | local_inc(&buffer->lost); | ||
| 4154 | perf_output_put_handle(handle); | ||
| 4155 | out: | ||
| 4156 | rcu_read_unlock(); | ||
| 4157 | |||
| 4158 | return -ENOSPC; | ||
| 4159 | } | ||
| 4160 | |||
| 4161 | void perf_output_end(struct perf_output_handle *handle) | ||
| 4162 | { | ||
| 4163 | struct perf_event *event = handle->event; | ||
| 4164 | struct perf_buffer *buffer = handle->buffer; | ||
| 4165 | |||
| 4166 | int wakeup_events = event->attr.wakeup_events; | ||
| 4167 | |||
| 4168 | if (handle->sample && wakeup_events) { | ||
| 4169 | int events = local_inc_return(&buffer->events); | ||
| 4170 | if (events >= wakeup_events) { | ||
| 4171 | local_sub(wakeup_events, &buffer->events); | ||
| 4172 | local_inc(&buffer->wakeup); | ||
| 4173 | } | ||
| 4174 | } | ||
| 4175 | |||
| 4176 | perf_output_put_handle(handle); | ||
| 4177 | rcu_read_unlock(); | ||
| 4178 | } | ||
| 4179 | |||
| 4180 | static void perf_output_read_one(struct perf_output_handle *handle, | 3771 | static void perf_output_read_one(struct perf_output_handle *handle, |
| 4181 | struct perf_event *event, | 3772 | struct perf_event *event, |
| 4182 | u64 enabled, u64 running) | 3773 | u64 enabled, u64 running) |
| @@ -4197,7 +3788,7 @@ static void perf_output_read_one(struct perf_output_handle *handle, | |||
| 4197 | if (read_format & PERF_FORMAT_ID) | 3788 | if (read_format & PERF_FORMAT_ID) |
| 4198 | values[n++] = primary_event_id(event); | 3789 | values[n++] = primary_event_id(event); |
| 4199 | 3790 | ||
| 4200 | perf_output_copy(handle, values, n * sizeof(u64)); | 3791 | __output_copy(handle, values, n * sizeof(u64)); |
| 4201 | } | 3792 | } |
| 4202 | 3793 | ||
| 4203 | /* | 3794 | /* |
| @@ -4227,7 +3818,7 @@ static void perf_output_read_group(struct perf_output_handle *handle, | |||
| 4227 | if (read_format & PERF_FORMAT_ID) | 3818 | if (read_format & PERF_FORMAT_ID) |
| 4228 | values[n++] = primary_event_id(leader); | 3819 | values[n++] = primary_event_id(leader); |
| 4229 | 3820 | ||
| 4230 | perf_output_copy(handle, values, n * sizeof(u64)); | 3821 | __output_copy(handle, values, n * sizeof(u64)); |
| 4231 | 3822 | ||
| 4232 | list_for_each_entry(sub, &leader->sibling_list, group_entry) { | 3823 | list_for_each_entry(sub, &leader->sibling_list, group_entry) { |
| 4233 | n = 0; | 3824 | n = 0; |
| @@ -4239,7 +3830,7 @@ static void perf_output_read_group(struct perf_output_handle *handle, | |||
| 4239 | if (read_format & PERF_FORMAT_ID) | 3830 | if (read_format & PERF_FORMAT_ID) |
| 4240 | values[n++] = primary_event_id(sub); | 3831 | values[n++] = primary_event_id(sub); |
| 4241 | 3832 | ||
| 4242 | perf_output_copy(handle, values, n * sizeof(u64)); | 3833 | __output_copy(handle, values, n * sizeof(u64)); |
| 4243 | } | 3834 | } |
| 4244 | } | 3835 | } |
| 4245 | 3836 | ||
| @@ -4249,7 +3840,7 @@ static void perf_output_read_group(struct perf_output_handle *handle, | |||
| 4249 | static void perf_output_read(struct perf_output_handle *handle, | 3840 | static void perf_output_read(struct perf_output_handle *handle, |
| 4250 | struct perf_event *event) | 3841 | struct perf_event *event) |
| 4251 | { | 3842 | { |
| 4252 | u64 enabled = 0, running = 0, now, ctx_time; | 3843 | u64 enabled = 0, running = 0; |
| 4253 | u64 read_format = event->attr.read_format; | 3844 | u64 read_format = event->attr.read_format; |
| 4254 | 3845 | ||
| 4255 | /* | 3846 | /* |
| @@ -4261,12 +3852,8 @@ static void perf_output_read(struct perf_output_handle *handle, | |||
| 4261 | * because of locking issue as we are called in | 3852 | * because of locking issue as we are called in |
| 4262 | * NMI context | 3853 | * NMI context |
| 4263 | */ | 3854 | */ |
| 4264 | if (read_format & PERF_FORMAT_TOTAL_TIMES) { | 3855 | if (read_format & PERF_FORMAT_TOTAL_TIMES) |
| 4265 | now = perf_clock(); | 3856 | calc_timer_values(event, &enabled, &running); |
| 4266 | ctx_time = event->shadow_ctx_time + now; | ||
| 4267 | enabled = ctx_time - event->tstamp_enabled; | ||
| 4268 | running = ctx_time - event->tstamp_running; | ||
| 4269 | } | ||
| 4270 | 3857 | ||
| 4271 | if (event->attr.read_format & PERF_FORMAT_GROUP) | 3858 | if (event->attr.read_format & PERF_FORMAT_GROUP) |
| 4272 | perf_output_read_group(handle, event, enabled, running); | 3859 | perf_output_read_group(handle, event, enabled, running); |
| @@ -4319,7 +3906,7 @@ void perf_output_sample(struct perf_output_handle *handle, | |||
| 4319 | 3906 | ||
| 4320 | size *= sizeof(u64); | 3907 | size *= sizeof(u64); |
| 4321 | 3908 | ||
| 4322 | perf_output_copy(handle, data->callchain, size); | 3909 | __output_copy(handle, data->callchain, size); |
| 4323 | } else { | 3910 | } else { |
| 4324 | u64 nr = 0; | 3911 | u64 nr = 0; |
| 4325 | perf_output_put(handle, nr); | 3912 | perf_output_put(handle, nr); |
| @@ -4329,8 +3916,8 @@ void perf_output_sample(struct perf_output_handle *handle, | |||
| 4329 | if (sample_type & PERF_SAMPLE_RAW) { | 3916 | if (sample_type & PERF_SAMPLE_RAW) { |
| 4330 | if (data->raw) { | 3917 | if (data->raw) { |
| 4331 | perf_output_put(handle, data->raw->size); | 3918 | perf_output_put(handle, data->raw->size); |
| 4332 | perf_output_copy(handle, data->raw->data, | 3919 | __output_copy(handle, data->raw->data, |
| 4333 | data->raw->size); | 3920 | data->raw->size); |
| 4334 | } else { | 3921 | } else { |
| 4335 | struct { | 3922 | struct { |
| 4336 | u32 size; | 3923 | u32 size; |
| @@ -4342,6 +3929,20 @@ void perf_output_sample(struct perf_output_handle *handle, | |||
| 4342 | perf_output_put(handle, raw); | 3929 | perf_output_put(handle, raw); |
| 4343 | } | 3930 | } |
| 4344 | } | 3931 | } |
| 3932 | |||
| 3933 | if (!event->attr.watermark) { | ||
| 3934 | int wakeup_events = event->attr.wakeup_events; | ||
| 3935 | |||
| 3936 | if (wakeup_events) { | ||
| 3937 | struct ring_buffer *rb = handle->rb; | ||
| 3938 | int events = local_inc_return(&rb->events); | ||
| 3939 | |||
| 3940 | if (events >= wakeup_events) { | ||
| 3941 | local_sub(wakeup_events, &rb->events); | ||
| 3942 | local_inc(&rb->wakeup); | ||
| 3943 | } | ||
| 3944 | } | ||
| 3945 | } | ||
| 4345 | } | 3946 | } |
| 4346 | 3947 | ||
| 4347 | void perf_prepare_sample(struct perf_event_header *header, | 3948 | void perf_prepare_sample(struct perf_event_header *header, |
| @@ -4386,7 +3987,7 @@ void perf_prepare_sample(struct perf_event_header *header, | |||
| 4386 | } | 3987 | } |
| 4387 | } | 3988 | } |
| 4388 | 3989 | ||
| 4389 | static void perf_event_output(struct perf_event *event, int nmi, | 3990 | static void perf_event_output(struct perf_event *event, |
| 4390 | struct perf_sample_data *data, | 3991 | struct perf_sample_data *data, |
| 4391 | struct pt_regs *regs) | 3992 | struct pt_regs *regs) |
| 4392 | { | 3993 | { |
| @@ -4398,7 +3999,7 @@ static void perf_event_output(struct perf_event *event, int nmi, | |||
| 4398 | 3999 | ||
| 4399 | perf_prepare_sample(&header, data, event, regs); | 4000 | perf_prepare_sample(&header, data, event, regs); |
| 4400 | 4001 | ||
| 4401 | if (perf_output_begin(&handle, event, header.size, nmi, 1)) | 4002 | if (perf_output_begin(&handle, event, header.size)) |
| 4402 | goto exit; | 4003 | goto exit; |
| 4403 | 4004 | ||
| 4404 | perf_output_sample(&handle, &header, data, event); | 4005 | perf_output_sample(&handle, &header, data, event); |
| @@ -4438,7 +4039,7 @@ perf_event_read_event(struct perf_event *event, | |||
| 4438 | int ret; | 4039 | int ret; |
| 4439 | 4040 | ||
| 4440 | perf_event_header__init_id(&read_event.header, &sample, event); | 4041 | perf_event_header__init_id(&read_event.header, &sample, event); |
| 4441 | ret = perf_output_begin(&handle, event, read_event.header.size, 0, 0); | 4042 | ret = perf_output_begin(&handle, event, read_event.header.size); |
| 4442 | if (ret) | 4043 | if (ret) |
| 4443 | return; | 4044 | return; |
| 4444 | 4045 | ||
| @@ -4481,7 +4082,7 @@ static void perf_event_task_output(struct perf_event *event, | |||
| 4481 | perf_event_header__init_id(&task_event->event_id.header, &sample, event); | 4082 | perf_event_header__init_id(&task_event->event_id.header, &sample, event); |
| 4482 | 4083 | ||
| 4483 | ret = perf_output_begin(&handle, event, | 4084 | ret = perf_output_begin(&handle, event, |
| 4484 | task_event->event_id.header.size, 0, 0); | 4085 | task_event->event_id.header.size); |
| 4485 | if (ret) | 4086 | if (ret) |
| 4486 | goto out; | 4087 | goto out; |
| 4487 | 4088 | ||
| @@ -4618,7 +4219,7 @@ static void perf_event_comm_output(struct perf_event *event, | |||
| 4618 | 4219 | ||
| 4619 | perf_event_header__init_id(&comm_event->event_id.header, &sample, event); | 4220 | perf_event_header__init_id(&comm_event->event_id.header, &sample, event); |
| 4620 | ret = perf_output_begin(&handle, event, | 4221 | ret = perf_output_begin(&handle, event, |
| 4621 | comm_event->event_id.header.size, 0, 0); | 4222 | comm_event->event_id.header.size); |
| 4622 | 4223 | ||
| 4623 | if (ret) | 4224 | if (ret) |
| 4624 | goto out; | 4225 | goto out; |
| @@ -4627,7 +4228,7 @@ static void perf_event_comm_output(struct perf_event *event, | |||
| 4627 | comm_event->event_id.tid = perf_event_tid(event, comm_event->task); | 4228 | comm_event->event_id.tid = perf_event_tid(event, comm_event->task); |
| 4628 | 4229 | ||
| 4629 | perf_output_put(&handle, comm_event->event_id); | 4230 | perf_output_put(&handle, comm_event->event_id); |
| 4630 | perf_output_copy(&handle, comm_event->comm, | 4231 | __output_copy(&handle, comm_event->comm, |
| 4631 | comm_event->comm_size); | 4232 | comm_event->comm_size); |
| 4632 | 4233 | ||
| 4633 | perf_event__output_id_sample(event, &handle, &sample); | 4234 | perf_event__output_id_sample(event, &handle, &sample); |
| @@ -4765,7 +4366,7 @@ static void perf_event_mmap_output(struct perf_event *event, | |||
| 4765 | 4366 | ||
| 4766 | perf_event_header__init_id(&mmap_event->event_id.header, &sample, event); | 4367 | perf_event_header__init_id(&mmap_event->event_id.header, &sample, event); |
| 4767 | ret = perf_output_begin(&handle, event, | 4368 | ret = perf_output_begin(&handle, event, |
| 4768 | mmap_event->event_id.header.size, 0, 0); | 4369 | mmap_event->event_id.header.size); |
| 4769 | if (ret) | 4370 | if (ret) |
| 4770 | goto out; | 4371 | goto out; |
| 4771 | 4372 | ||
| @@ -4773,7 +4374,7 @@ static void perf_event_mmap_output(struct perf_event *event, | |||
| 4773 | mmap_event->event_id.tid = perf_event_tid(event, current); | 4374 | mmap_event->event_id.tid = perf_event_tid(event, current); |
| 4774 | 4375 | ||
| 4775 | perf_output_put(&handle, mmap_event->event_id); | 4376 | perf_output_put(&handle, mmap_event->event_id); |
| 4776 | perf_output_copy(&handle, mmap_event->file_name, | 4377 | __output_copy(&handle, mmap_event->file_name, |
| 4777 | mmap_event->file_size); | 4378 | mmap_event->file_size); |
| 4778 | 4379 | ||
| 4779 | perf_event__output_id_sample(event, &handle, &sample); | 4380 | perf_event__output_id_sample(event, &handle, &sample); |
| @@ -4829,7 +4430,7 @@ static void perf_event_mmap_event(struct perf_mmap_event *mmap_event) | |||
| 4829 | 4430 | ||
| 4830 | if (file) { | 4431 | if (file) { |
| 4831 | /* | 4432 | /* |
| 4832 | * d_path works from the end of the buffer backwards, so we | 4433 | * d_path works from the end of the rb backwards, so we |
| 4833 | * need to add enough zero bytes after the string to handle | 4434 | * need to add enough zero bytes after the string to handle |
| 4834 | * the 64bit alignment we do later. | 4435 | * the 64bit alignment we do later. |
| 4835 | */ | 4436 | */ |
| @@ -4960,7 +4561,7 @@ static void perf_log_throttle(struct perf_event *event, int enable) | |||
| 4960 | perf_event_header__init_id(&throttle_event.header, &sample, event); | 4561 | perf_event_header__init_id(&throttle_event.header, &sample, event); |
| 4961 | 4562 | ||
| 4962 | ret = perf_output_begin(&handle, event, | 4563 | ret = perf_output_begin(&handle, event, |
| 4963 | throttle_event.header.size, 1, 0); | 4564 | throttle_event.header.size); |
| 4964 | if (ret) | 4565 | if (ret) |
| 4965 | return; | 4566 | return; |
| 4966 | 4567 | ||
| @@ -4973,7 +4574,7 @@ static void perf_log_throttle(struct perf_event *event, int enable) | |||
| 4973 | * Generic event overflow handling, sampling. | 4574 | * Generic event overflow handling, sampling. |
| 4974 | */ | 4575 | */ |
| 4975 | 4576 | ||
| 4976 | static int __perf_event_overflow(struct perf_event *event, int nmi, | 4577 | static int __perf_event_overflow(struct perf_event *event, |
| 4977 | int throttle, struct perf_sample_data *data, | 4578 | int throttle, struct perf_sample_data *data, |
| 4978 | struct pt_regs *regs) | 4579 | struct pt_regs *regs) |
| 4979 | { | 4580 | { |
| @@ -5016,34 +4617,28 @@ static int __perf_event_overflow(struct perf_event *event, int nmi, | |||
| 5016 | if (events && atomic_dec_and_test(&event->event_limit)) { | 4617 | if (events && atomic_dec_and_test(&event->event_limit)) { |
| 5017 | ret = 1; | 4618 | ret = 1; |
| 5018 | event->pending_kill = POLL_HUP; | 4619 | event->pending_kill = POLL_HUP; |
| 5019 | if (nmi) { | 4620 | event->pending_disable = 1; |
| 5020 | event->pending_disable = 1; | 4621 | irq_work_queue(&event->pending); |
| 5021 | irq_work_queue(&event->pending); | ||
| 5022 | } else | ||
| 5023 | perf_event_disable(event); | ||
| 5024 | } | 4622 | } |
| 5025 | 4623 | ||
| 5026 | if (event->overflow_handler) | 4624 | if (event->overflow_handler) |
| 5027 | event->overflow_handler(event, nmi, data, regs); | 4625 | event->overflow_handler(event, data, regs); |
| 5028 | else | 4626 | else |
| 5029 | perf_event_output(event, nmi, data, regs); | 4627 | perf_event_output(event, data, regs); |
| 5030 | 4628 | ||
| 5031 | if (event->fasync && event->pending_kill) { | 4629 | if (event->fasync && event->pending_kill) { |
| 5032 | if (nmi) { | 4630 | event->pending_wakeup = 1; |
| 5033 | event->pending_wakeup = 1; | 4631 | irq_work_queue(&event->pending); |
| 5034 | irq_work_queue(&event->pending); | ||
| 5035 | } else | ||
| 5036 | perf_event_wakeup(event); | ||
| 5037 | } | 4632 | } |
| 5038 | 4633 | ||
| 5039 | return ret; | 4634 | return ret; |
| 5040 | } | 4635 | } |
| 5041 | 4636 | ||
| 5042 | int perf_event_overflow(struct perf_event *event, int nmi, | 4637 | int perf_event_overflow(struct perf_event *event, |
| 5043 | struct perf_sample_data *data, | 4638 | struct perf_sample_data *data, |
| 5044 | struct pt_regs *regs) | 4639 | struct pt_regs *regs) |
| 5045 | { | 4640 | { |
| 5046 | return __perf_event_overflow(event, nmi, 1, data, regs); | 4641 | return __perf_event_overflow(event, 1, data, regs); |
| 5047 | } | 4642 | } |
| 5048 | 4643 | ||
| 5049 | /* | 4644 | /* |
| @@ -5092,7 +4687,7 @@ again: | |||
| 5092 | } | 4687 | } |
| 5093 | 4688 | ||
| 5094 | static void perf_swevent_overflow(struct perf_event *event, u64 overflow, | 4689 | static void perf_swevent_overflow(struct perf_event *event, u64 overflow, |
| 5095 | int nmi, struct perf_sample_data *data, | 4690 | struct perf_sample_data *data, |
| 5096 | struct pt_regs *regs) | 4691 | struct pt_regs *regs) |
| 5097 | { | 4692 | { |
| 5098 | struct hw_perf_event *hwc = &event->hw; | 4693 | struct hw_perf_event *hwc = &event->hw; |
| @@ -5106,7 +4701,7 @@ static void perf_swevent_overflow(struct perf_event *event, u64 overflow, | |||
| 5106 | return; | 4701 | return; |
| 5107 | 4702 | ||
| 5108 | for (; overflow; overflow--) { | 4703 | for (; overflow; overflow--) { |
| 5109 | if (__perf_event_overflow(event, nmi, throttle, | 4704 | if (__perf_event_overflow(event, throttle, |
| 5110 | data, regs)) { | 4705 | data, regs)) { |
| 5111 | /* | 4706 | /* |
| 5112 | * We inhibit the overflow from happening when | 4707 | * We inhibit the overflow from happening when |
| @@ -5119,7 +4714,7 @@ static void perf_swevent_overflow(struct perf_event *event, u64 overflow, | |||
| 5119 | } | 4714 | } |
| 5120 | 4715 | ||
| 5121 | static void perf_swevent_event(struct perf_event *event, u64 nr, | 4716 | static void perf_swevent_event(struct perf_event *event, u64 nr, |
| 5122 | int nmi, struct perf_sample_data *data, | 4717 | struct perf_sample_data *data, |
| 5123 | struct pt_regs *regs) | 4718 | struct pt_regs *regs) |
| 5124 | { | 4719 | { |
| 5125 | struct hw_perf_event *hwc = &event->hw; | 4720 | struct hw_perf_event *hwc = &event->hw; |
| @@ -5133,12 +4728,12 @@ static void perf_swevent_event(struct perf_event *event, u64 nr, | |||
| 5133 | return; | 4728 | return; |
| 5134 | 4729 | ||
| 5135 | if (nr == 1 && hwc->sample_period == 1 && !event->attr.freq) | 4730 | if (nr == 1 && hwc->sample_period == 1 && !event->attr.freq) |
| 5136 | return perf_swevent_overflow(event, 1, nmi, data, regs); | 4731 | return perf_swevent_overflow(event, 1, data, regs); |
| 5137 | 4732 | ||
| 5138 | if (local64_add_negative(nr, &hwc->period_left)) | 4733 | if (local64_add_negative(nr, &hwc->period_left)) |
| 5139 | return; | 4734 | return; |
| 5140 | 4735 | ||
| 5141 | perf_swevent_overflow(event, 0, nmi, data, regs); | 4736 | perf_swevent_overflow(event, 0, data, regs); |
| 5142 | } | 4737 | } |
| 5143 | 4738 | ||
| 5144 | static int perf_exclude_event(struct perf_event *event, | 4739 | static int perf_exclude_event(struct perf_event *event, |
| @@ -5226,7 +4821,7 @@ find_swevent_head(struct swevent_htable *swhash, struct perf_event *event) | |||
| 5226 | } | 4821 | } |
| 5227 | 4822 | ||
| 5228 | static void do_perf_sw_event(enum perf_type_id type, u32 event_id, | 4823 | static void do_perf_sw_event(enum perf_type_id type, u32 event_id, |
| 5229 | u64 nr, int nmi, | 4824 | u64 nr, |
| 5230 | struct perf_sample_data *data, | 4825 | struct perf_sample_data *data, |
| 5231 | struct pt_regs *regs) | 4826 | struct pt_regs *regs) |
| 5232 | { | 4827 | { |
| @@ -5242,7 +4837,7 @@ static void do_perf_sw_event(enum perf_type_id type, u32 event_id, | |||
| 5242 | 4837 | ||
| 5243 | hlist_for_each_entry_rcu(event, node, head, hlist_entry) { | 4838 | hlist_for_each_entry_rcu(event, node, head, hlist_entry) { |
| 5244 | if (perf_swevent_match(event, type, event_id, data, regs)) | 4839 | if (perf_swevent_match(event, type, event_id, data, regs)) |
| 5245 | perf_swevent_event(event, nr, nmi, data, regs); | 4840 | perf_swevent_event(event, nr, data, regs); |
| 5246 | } | 4841 | } |
| 5247 | end: | 4842 | end: |
| 5248 | rcu_read_unlock(); | 4843 | rcu_read_unlock(); |
| @@ -5263,8 +4858,7 @@ inline void perf_swevent_put_recursion_context(int rctx) | |||
| 5263 | put_recursion_context(swhash->recursion, rctx); | 4858 | put_recursion_context(swhash->recursion, rctx); |
| 5264 | } | 4859 | } |
| 5265 | 4860 | ||
| 5266 | void __perf_sw_event(u32 event_id, u64 nr, int nmi, | 4861 | void __perf_sw_event(u32 event_id, u64 nr, struct pt_regs *regs, u64 addr) |
| 5267 | struct pt_regs *regs, u64 addr) | ||
| 5268 | { | 4862 | { |
| 5269 | struct perf_sample_data data; | 4863 | struct perf_sample_data data; |
| 5270 | int rctx; | 4864 | int rctx; |
| @@ -5276,7 +4870,7 @@ void __perf_sw_event(u32 event_id, u64 nr, int nmi, | |||
| 5276 | 4870 | ||
| 5277 | perf_sample_data_init(&data, addr); | 4871 | perf_sample_data_init(&data, addr); |
| 5278 | 4872 | ||
| 5279 | do_perf_sw_event(PERF_TYPE_SOFTWARE, event_id, nr, nmi, &data, regs); | 4873 | do_perf_sw_event(PERF_TYPE_SOFTWARE, event_id, nr, &data, regs); |
| 5280 | 4874 | ||
| 5281 | perf_swevent_put_recursion_context(rctx); | 4875 | perf_swevent_put_recursion_context(rctx); |
| 5282 | preempt_enable_notrace(); | 4876 | preempt_enable_notrace(); |
| @@ -5524,7 +5118,7 @@ void perf_tp_event(u64 addr, u64 count, void *record, int entry_size, | |||
| 5524 | 5118 | ||
| 5525 | hlist_for_each_entry_rcu(event, node, head, hlist_entry) { | 5119 | hlist_for_each_entry_rcu(event, node, head, hlist_entry) { |
| 5526 | if (perf_tp_event_match(event, &data, regs)) | 5120 | if (perf_tp_event_match(event, &data, regs)) |
| 5527 | perf_swevent_event(event, count, 1, &data, regs); | 5121 | perf_swevent_event(event, count, &data, regs); |
| 5528 | } | 5122 | } |
| 5529 | 5123 | ||
| 5530 | perf_swevent_put_recursion_context(rctx); | 5124 | perf_swevent_put_recursion_context(rctx); |
| @@ -5617,7 +5211,7 @@ void perf_bp_event(struct perf_event *bp, void *data) | |||
| 5617 | perf_sample_data_init(&sample, bp->attr.bp_addr); | 5211 | perf_sample_data_init(&sample, bp->attr.bp_addr); |
| 5618 | 5212 | ||
| 5619 | if (!bp->hw.state && !perf_exclude_event(bp, regs)) | 5213 | if (!bp->hw.state && !perf_exclude_event(bp, regs)) |
| 5620 | perf_swevent_event(bp, 1, 1, &sample, regs); | 5214 | perf_swevent_event(bp, 1, &sample, regs); |
| 5621 | } | 5215 | } |
| 5622 | #endif | 5216 | #endif |
| 5623 | 5217 | ||
| @@ -5646,7 +5240,7 @@ static enum hrtimer_restart perf_swevent_hrtimer(struct hrtimer *hrtimer) | |||
| 5646 | 5240 | ||
| 5647 | if (regs && !perf_exclude_event(event, regs)) { | 5241 | if (regs && !perf_exclude_event(event, regs)) { |
| 5648 | if (!(event->attr.exclude_idle && current->pid == 0)) | 5242 | if (!(event->attr.exclude_idle && current->pid == 0)) |
| 5649 | if (perf_event_overflow(event, 0, &data, regs)) | 5243 | if (perf_event_overflow(event, &data, regs)) |
| 5650 | ret = HRTIMER_NORESTART; | 5244 | ret = HRTIMER_NORESTART; |
| 5651 | } | 5245 | } |
| 5652 | 5246 | ||
| @@ -5986,6 +5580,7 @@ free_dev: | |||
| 5986 | } | 5580 | } |
| 5987 | 5581 | ||
| 5988 | static struct lock_class_key cpuctx_mutex; | 5582 | static struct lock_class_key cpuctx_mutex; |
| 5583 | static struct lock_class_key cpuctx_lock; | ||
| 5989 | 5584 | ||
| 5990 | int perf_pmu_register(struct pmu *pmu, char *name, int type) | 5585 | int perf_pmu_register(struct pmu *pmu, char *name, int type) |
| 5991 | { | 5586 | { |
| @@ -6036,6 +5631,7 @@ skip_type: | |||
| 6036 | cpuctx = per_cpu_ptr(pmu->pmu_cpu_context, cpu); | 5631 | cpuctx = per_cpu_ptr(pmu->pmu_cpu_context, cpu); |
| 6037 | __perf_event_init_context(&cpuctx->ctx); | 5632 | __perf_event_init_context(&cpuctx->ctx); |
| 6038 | lockdep_set_class(&cpuctx->ctx.mutex, &cpuctx_mutex); | 5633 | lockdep_set_class(&cpuctx->ctx.mutex, &cpuctx_mutex); |
| 5634 | lockdep_set_class(&cpuctx->ctx.lock, &cpuctx_lock); | ||
| 6039 | cpuctx->ctx.type = cpu_context; | 5635 | cpuctx->ctx.type = cpu_context; |
| 6040 | cpuctx->ctx.pmu = pmu; | 5636 | cpuctx->ctx.pmu = pmu; |
| 6041 | cpuctx->jiffies_interval = 1; | 5637 | cpuctx->jiffies_interval = 1; |
| @@ -6150,7 +5746,8 @@ perf_event_alloc(struct perf_event_attr *attr, int cpu, | |||
| 6150 | struct task_struct *task, | 5746 | struct task_struct *task, |
| 6151 | struct perf_event *group_leader, | 5747 | struct perf_event *group_leader, |
| 6152 | struct perf_event *parent_event, | 5748 | struct perf_event *parent_event, |
| 6153 | perf_overflow_handler_t overflow_handler) | 5749 | perf_overflow_handler_t overflow_handler, |
| 5750 | void *context) | ||
| 6154 | { | 5751 | { |
| 6155 | struct pmu *pmu; | 5752 | struct pmu *pmu; |
| 6156 | struct perf_event *event; | 5753 | struct perf_event *event; |
| @@ -6208,10 +5805,13 @@ perf_event_alloc(struct perf_event_attr *attr, int cpu, | |||
| 6208 | #endif | 5805 | #endif |
| 6209 | } | 5806 | } |
| 6210 | 5807 | ||
| 6211 | if (!overflow_handler && parent_event) | 5808 | if (!overflow_handler && parent_event) { |
| 6212 | overflow_handler = parent_event->overflow_handler; | 5809 | overflow_handler = parent_event->overflow_handler; |
| 5810 | context = parent_event->overflow_handler_context; | ||
| 5811 | } | ||
| 6213 | 5812 | ||
| 6214 | event->overflow_handler = overflow_handler; | 5813 | event->overflow_handler = overflow_handler; |
| 5814 | event->overflow_handler_context = context; | ||
| 6215 | 5815 | ||
| 6216 | if (attr->disabled) | 5816 | if (attr->disabled) |
| 6217 | event->state = PERF_EVENT_STATE_OFF; | 5817 | event->state = PERF_EVENT_STATE_OFF; |
| @@ -6326,13 +5926,6 @@ static int perf_copy_attr(struct perf_event_attr __user *uattr, | |||
| 6326 | if (ret) | 5926 | if (ret) |
| 6327 | return -EFAULT; | 5927 | return -EFAULT; |
| 6328 | 5928 | ||
| 6329 | /* | ||
| 6330 | * If the type exists, the corresponding creation will verify | ||
| 6331 | * the attr->config. | ||
| 6332 | */ | ||
| 6333 | if (attr->type >= PERF_TYPE_MAX) | ||
| 6334 | return -EINVAL; | ||
| 6335 | |||
| 6336 | if (attr->__reserved_1) | 5929 | if (attr->__reserved_1) |
| 6337 | return -EINVAL; | 5930 | return -EINVAL; |
| 6338 | 5931 | ||
| @@ -6354,7 +5947,7 @@ err_size: | |||
| 6354 | static int | 5947 | static int |
| 6355 | perf_event_set_output(struct perf_event *event, struct perf_event *output_event) | 5948 | perf_event_set_output(struct perf_event *event, struct perf_event *output_event) |
| 6356 | { | 5949 | { |
| 6357 | struct perf_buffer *buffer = NULL, *old_buffer = NULL; | 5950 | struct ring_buffer *rb = NULL, *old_rb = NULL; |
| 6358 | int ret = -EINVAL; | 5951 | int ret = -EINVAL; |
| 6359 | 5952 | ||
| 6360 | if (!output_event) | 5953 | if (!output_event) |
| @@ -6371,7 +5964,7 @@ perf_event_set_output(struct perf_event *event, struct perf_event *output_event) | |||
| 6371 | goto out; | 5964 | goto out; |
| 6372 | 5965 | ||
| 6373 | /* | 5966 | /* |
| 6374 | * If its not a per-cpu buffer, it must be the same task. | 5967 | * If its not a per-cpu rb, it must be the same task. |
| 6375 | */ | 5968 | */ |
| 6376 | if (output_event->cpu == -1 && output_event->ctx != event->ctx) | 5969 | if (output_event->cpu == -1 && output_event->ctx != event->ctx) |
| 6377 | goto out; | 5970 | goto out; |
| @@ -6383,20 +5976,20 @@ set: | |||
| 6383 | goto unlock; | 5976 | goto unlock; |
| 6384 | 5977 | ||
| 6385 | if (output_event) { | 5978 | if (output_event) { |
| 6386 | /* get the buffer we want to redirect to */ | 5979 | /* get the rb we want to redirect to */ |
| 6387 | buffer = perf_buffer_get(output_event); | 5980 | rb = ring_buffer_get(output_event); |
| 6388 | if (!buffer) | 5981 | if (!rb) |
| 6389 | goto unlock; | 5982 | goto unlock; |
| 6390 | } | 5983 | } |
| 6391 | 5984 | ||
| 6392 | old_buffer = event->buffer; | 5985 | old_rb = event->rb; |
| 6393 | rcu_assign_pointer(event->buffer, buffer); | 5986 | rcu_assign_pointer(event->rb, rb); |
| 6394 | ret = 0; | 5987 | ret = 0; |
| 6395 | unlock: | 5988 | unlock: |
| 6396 | mutex_unlock(&event->mmap_mutex); | 5989 | mutex_unlock(&event->mmap_mutex); |
| 6397 | 5990 | ||
| 6398 | if (old_buffer) | 5991 | if (old_rb) |
| 6399 | perf_buffer_put(old_buffer); | 5992 | ring_buffer_put(old_rb); |
| 6400 | out: | 5993 | out: |
| 6401 | return ret; | 5994 | return ret; |
| 6402 | } | 5995 | } |
| @@ -6478,7 +6071,8 @@ SYSCALL_DEFINE5(perf_event_open, | |||
| 6478 | } | 6071 | } |
| 6479 | } | 6072 | } |
| 6480 | 6073 | ||
| 6481 | event = perf_event_alloc(&attr, cpu, task, group_leader, NULL, NULL); | 6074 | event = perf_event_alloc(&attr, cpu, task, group_leader, NULL, |
| 6075 | NULL, NULL); | ||
| 6482 | if (IS_ERR(event)) { | 6076 | if (IS_ERR(event)) { |
| 6483 | err = PTR_ERR(event); | 6077 | err = PTR_ERR(event); |
| 6484 | goto err_task; | 6078 | goto err_task; |
| @@ -6663,7 +6257,8 @@ err_fd: | |||
| 6663 | struct perf_event * | 6257 | struct perf_event * |
| 6664 | perf_event_create_kernel_counter(struct perf_event_attr *attr, int cpu, | 6258 | perf_event_create_kernel_counter(struct perf_event_attr *attr, int cpu, |
| 6665 | struct task_struct *task, | 6259 | struct task_struct *task, |
| 6666 | perf_overflow_handler_t overflow_handler) | 6260 | perf_overflow_handler_t overflow_handler, |
| 6261 | void *context) | ||
| 6667 | { | 6262 | { |
| 6668 | struct perf_event_context *ctx; | 6263 | struct perf_event_context *ctx; |
| 6669 | struct perf_event *event; | 6264 | struct perf_event *event; |
| @@ -6673,7 +6268,8 @@ perf_event_create_kernel_counter(struct perf_event_attr *attr, int cpu, | |||
| 6673 | * Get the target context (task or percpu): | 6268 | * Get the target context (task or percpu): |
| 6674 | */ | 6269 | */ |
| 6675 | 6270 | ||
| 6676 | event = perf_event_alloc(attr, cpu, task, NULL, NULL, overflow_handler); | 6271 | event = perf_event_alloc(attr, cpu, task, NULL, NULL, |
| 6272 | overflow_handler, context); | ||
| 6677 | if (IS_ERR(event)) { | 6273 | if (IS_ERR(event)) { |
| 6678 | err = PTR_ERR(event); | 6274 | err = PTR_ERR(event); |
| 6679 | goto err; | 6275 | goto err; |
| @@ -6780,7 +6376,6 @@ static void perf_event_exit_task_context(struct task_struct *child, int ctxn) | |||
| 6780 | * our context. | 6376 | * our context. |
| 6781 | */ | 6377 | */ |
| 6782 | child_ctx = rcu_dereference_raw(child->perf_event_ctxp[ctxn]); | 6378 | child_ctx = rcu_dereference_raw(child->perf_event_ctxp[ctxn]); |
| 6783 | task_ctx_sched_out(child_ctx, EVENT_ALL); | ||
| 6784 | 6379 | ||
| 6785 | /* | 6380 | /* |
| 6786 | * Take the context lock here so that if find_get_context is | 6381 | * Take the context lock here so that if find_get_context is |
| @@ -6788,6 +6383,7 @@ static void perf_event_exit_task_context(struct task_struct *child, int ctxn) | |||
| 6788 | * incremented the context's refcount before we do put_ctx below. | 6383 | * incremented the context's refcount before we do put_ctx below. |
| 6789 | */ | 6384 | */ |
| 6790 | raw_spin_lock(&child_ctx->lock); | 6385 | raw_spin_lock(&child_ctx->lock); |
| 6386 | task_ctx_sched_out(child_ctx); | ||
| 6791 | child->perf_event_ctxp[ctxn] = NULL; | 6387 | child->perf_event_ctxp[ctxn] = NULL; |
| 6792 | /* | 6388 | /* |
| 6793 | * If this context is a clone; unclone it so it can't get | 6389 | * If this context is a clone; unclone it so it can't get |
| @@ -6957,7 +6553,7 @@ inherit_event(struct perf_event *parent_event, | |||
| 6957 | parent_event->cpu, | 6553 | parent_event->cpu, |
| 6958 | child, | 6554 | child, |
| 6959 | group_leader, parent_event, | 6555 | group_leader, parent_event, |
| 6960 | NULL); | 6556 | NULL, NULL); |
| 6961 | if (IS_ERR(child_event)) | 6557 | if (IS_ERR(child_event)) |
| 6962 | return child_event; | 6558 | return child_event; |
| 6963 | get_ctx(child_ctx); | 6559 | get_ctx(child_ctx); |
| @@ -6984,6 +6580,8 @@ inherit_event(struct perf_event *parent_event, | |||
| 6984 | 6580 | ||
| 6985 | child_event->ctx = child_ctx; | 6581 | child_event->ctx = child_ctx; |
| 6986 | child_event->overflow_handler = parent_event->overflow_handler; | 6582 | child_event->overflow_handler = parent_event->overflow_handler; |
| 6583 | child_event->overflow_handler_context | ||
| 6584 | = parent_event->overflow_handler_context; | ||
| 6987 | 6585 | ||
| 6988 | /* | 6586 | /* |
| 6989 | * Precalculate sample_data sizes | 6587 | * Precalculate sample_data sizes |
diff --git a/kernel/events/hw_breakpoint.c b/kernel/events/hw_breakpoint.c index 086adf25a55..b7971d6f38b 100644 --- a/kernel/events/hw_breakpoint.c +++ b/kernel/events/hw_breakpoint.c | |||
| @@ -431,9 +431,11 @@ int register_perf_hw_breakpoint(struct perf_event *bp) | |||
| 431 | struct perf_event * | 431 | struct perf_event * |
| 432 | register_user_hw_breakpoint(struct perf_event_attr *attr, | 432 | register_user_hw_breakpoint(struct perf_event_attr *attr, |
| 433 | perf_overflow_handler_t triggered, | 433 | perf_overflow_handler_t triggered, |
| 434 | void *context, | ||
| 434 | struct task_struct *tsk) | 435 | struct task_struct *tsk) |
| 435 | { | 436 | { |
| 436 | return perf_event_create_kernel_counter(attr, -1, tsk, triggered); | 437 | return perf_event_create_kernel_counter(attr, -1, tsk, triggered, |
| 438 | context); | ||
| 437 | } | 439 | } |
| 438 | EXPORT_SYMBOL_GPL(register_user_hw_breakpoint); | 440 | EXPORT_SYMBOL_GPL(register_user_hw_breakpoint); |
| 439 | 441 | ||
| @@ -502,7 +504,8 @@ EXPORT_SYMBOL_GPL(unregister_hw_breakpoint); | |||
| 502 | */ | 504 | */ |
| 503 | struct perf_event * __percpu * | 505 | struct perf_event * __percpu * |
| 504 | register_wide_hw_breakpoint(struct perf_event_attr *attr, | 506 | register_wide_hw_breakpoint(struct perf_event_attr *attr, |
| 505 | perf_overflow_handler_t triggered) | 507 | perf_overflow_handler_t triggered, |
| 508 | void *context) | ||
| 506 | { | 509 | { |
| 507 | struct perf_event * __percpu *cpu_events, **pevent, *bp; | 510 | struct perf_event * __percpu *cpu_events, **pevent, *bp; |
| 508 | long err; | 511 | long err; |
| @@ -515,7 +518,8 @@ register_wide_hw_breakpoint(struct perf_event_attr *attr, | |||
| 515 | get_online_cpus(); | 518 | get_online_cpus(); |
| 516 | for_each_online_cpu(cpu) { | 519 | for_each_online_cpu(cpu) { |
| 517 | pevent = per_cpu_ptr(cpu_events, cpu); | 520 | pevent = per_cpu_ptr(cpu_events, cpu); |
| 518 | bp = perf_event_create_kernel_counter(attr, cpu, NULL, triggered); | 521 | bp = perf_event_create_kernel_counter(attr, cpu, NULL, |
| 522 | triggered, context); | ||
| 519 | 523 | ||
| 520 | *pevent = bp; | 524 | *pevent = bp; |
| 521 | 525 | ||
diff --git a/kernel/events/internal.h b/kernel/events/internal.h new file mode 100644 index 00000000000..09097dd8116 --- /dev/null +++ b/kernel/events/internal.h | |||
| @@ -0,0 +1,96 @@ | |||
| 1 | #ifndef _KERNEL_EVENTS_INTERNAL_H | ||
| 2 | #define _KERNEL_EVENTS_INTERNAL_H | ||
| 3 | |||
| 4 | #define RING_BUFFER_WRITABLE 0x01 | ||
| 5 | |||
| 6 | struct ring_buffer { | ||
| 7 | atomic_t refcount; | ||
| 8 | struct rcu_head rcu_head; | ||
| 9 | #ifdef CONFIG_PERF_USE_VMALLOC | ||
| 10 | struct work_struct work; | ||
| 11 | int page_order; /* allocation order */ | ||
| 12 | #endif | ||
| 13 | int nr_pages; /* nr of data pages */ | ||
| 14 | int writable; /* are we writable */ | ||
| 15 | |||
| 16 | atomic_t poll; /* POLL_ for wakeups */ | ||
| 17 | |||
| 18 | local_t head; /* write position */ | ||
| 19 | local_t nest; /* nested writers */ | ||
| 20 | local_t events; /* event limit */ | ||
| 21 | local_t wakeup; /* wakeup stamp */ | ||
| 22 | local_t lost; /* nr records lost */ | ||
| 23 | |||
| 24 | long watermark; /* wakeup watermark */ | ||
| 25 | |||
| 26 | struct perf_event_mmap_page *user_page; | ||
| 27 | void *data_pages[0]; | ||
| 28 | }; | ||
| 29 | |||
| 30 | extern void rb_free(struct ring_buffer *rb); | ||
| 31 | extern struct ring_buffer * | ||
| 32 | rb_alloc(int nr_pages, long watermark, int cpu, int flags); | ||
| 33 | extern void perf_event_wakeup(struct perf_event *event); | ||
| 34 | |||
| 35 | extern void | ||
| 36 | perf_event_header__init_id(struct perf_event_header *header, | ||
| 37 | struct perf_sample_data *data, | ||
| 38 | struct perf_event *event); | ||
| 39 | extern void | ||
| 40 | perf_event__output_id_sample(struct perf_event *event, | ||
| 41 | struct perf_output_handle *handle, | ||
| 42 | struct perf_sample_data *sample); | ||
| 43 | |||
| 44 | extern struct page * | ||
| 45 | perf_mmap_to_page(struct ring_buffer *rb, unsigned long pgoff); | ||
| 46 | |||
| 47 | #ifdef CONFIG_PERF_USE_VMALLOC | ||
| 48 | /* | ||
| 49 | * Back perf_mmap() with vmalloc memory. | ||
| 50 | * | ||
| 51 | * Required for architectures that have d-cache aliasing issues. | ||
| 52 | */ | ||
| 53 | |||
| 54 | static inline int page_order(struct ring_buffer *rb) | ||
| 55 | { | ||
| 56 | return rb->page_order; | ||
| 57 | } | ||
| 58 | |||
| 59 | #else | ||
| 60 | |||
| 61 | static inline int page_order(struct ring_buffer *rb) | ||
| 62 | { | ||
| 63 | return 0; | ||
| 64 | } | ||
| 65 | #endif | ||
| 66 | |||
| 67 | static unsigned long perf_data_size(struct ring_buffer *rb) | ||
| 68 | { | ||
| 69 | return rb->nr_pages << (PAGE_SHIFT + page_order(rb)); | ||
| 70 | } | ||
| 71 | |||
| 72 | static inline void | ||
| 73 | __output_copy(struct perf_output_handle *handle, | ||
| 74 | const void *buf, unsigned int len) | ||
| 75 | { | ||
| 76 | do { | ||
| 77 | unsigned long size = min_t(unsigned long, handle->size, len); | ||
| 78 | |||
| 79 | memcpy(handle->addr, buf, size); | ||
| 80 | |||
| 81 | len -= size; | ||
| 82 | handle->addr += size; | ||
| 83 | buf += size; | ||
| 84 | handle->size -= size; | ||
| 85 | if (!handle->size) { | ||
| 86 | struct ring_buffer *rb = handle->rb; | ||
| 87 | |||
| 88 | handle->page++; | ||
| 89 | handle->page &= rb->nr_pages - 1; | ||
| 90 | handle->addr = rb->data_pages[handle->page]; | ||
| 91 | handle->size = PAGE_SIZE << page_order(rb); | ||
| 92 | } | ||
| 93 | } while (len); | ||
| 94 | } | ||
| 95 | |||
| 96 | #endif /* _KERNEL_EVENTS_INTERNAL_H */ | ||
diff --git a/kernel/events/ring_buffer.c b/kernel/events/ring_buffer.c new file mode 100644 index 00000000000..a2a29205cc0 --- /dev/null +++ b/kernel/events/ring_buffer.c | |||
| @@ -0,0 +1,380 @@ | |||
| 1 | /* | ||
| 2 | * Performance events ring-buffer code: | ||
| 3 | * | ||
| 4 | * Copyright (C) 2008 Thomas Gleixner <tglx@linutronix.de> | ||
| 5 | * Copyright (C) 2008-2011 Red Hat, Inc., Ingo Molnar | ||
| 6 | * Copyright (C) 2008-2011 Red Hat, Inc., Peter Zijlstra <pzijlstr@redhat.com> | ||
| 7 | * Copyright © 2009 Paul Mackerras, IBM Corp. <paulus@au1.ibm.com> | ||
| 8 | * | ||
| 9 | * For licensing details see kernel-base/COPYING | ||
| 10 | */ | ||
| 11 | |||
| 12 | #include <linux/perf_event.h> | ||
| 13 | #include <linux/vmalloc.h> | ||
| 14 | #include <linux/slab.h> | ||
| 15 | |||
| 16 | #include "internal.h" | ||
| 17 | |||
| 18 | static bool perf_output_space(struct ring_buffer *rb, unsigned long tail, | ||
| 19 | unsigned long offset, unsigned long head) | ||
| 20 | { | ||
| 21 | unsigned long mask; | ||
| 22 | |||
| 23 | if (!rb->writable) | ||
| 24 | return true; | ||
| 25 | |||
| 26 | mask = perf_data_size(rb) - 1; | ||
| 27 | |||
| 28 | offset = (offset - tail) & mask; | ||
| 29 | head = (head - tail) & mask; | ||
| 30 | |||
| 31 | if ((int)(head - offset) < 0) | ||
| 32 | return false; | ||
| 33 | |||
| 34 | return true; | ||
| 35 | } | ||
| 36 | |||
| 37 | static void perf_output_wakeup(struct perf_output_handle *handle) | ||
| 38 | { | ||
| 39 | atomic_set(&handle->rb->poll, POLL_IN); | ||
| 40 | |||
| 41 | handle->event->pending_wakeup = 1; | ||
| 42 | irq_work_queue(&handle->event->pending); | ||
| 43 | } | ||
| 44 | |||
| 45 | /* | ||
| 46 | * We need to ensure a later event_id doesn't publish a head when a former | ||
| 47 | * event isn't done writing. However since we need to deal with NMIs we | ||
| 48 | * cannot fully serialize things. | ||
| 49 | * | ||
| 50 | * We only publish the head (and generate a wakeup) when the outer-most | ||
| 51 | * event completes. | ||
| 52 | */ | ||
| 53 | static void perf_output_get_handle(struct perf_output_handle *handle) | ||
| 54 | { | ||
| 55 | struct ring_buffer *rb = handle->rb; | ||
| 56 | |||
| 57 | preempt_disable(); | ||
| 58 | local_inc(&rb->nest); | ||
| 59 | handle->wakeup = local_read(&rb->wakeup); | ||
| 60 | } | ||
| 61 | |||
| 62 | static void perf_output_put_handle(struct perf_output_handle *handle) | ||
| 63 | { | ||
| 64 | struct ring_buffer *rb = handle->rb; | ||
| 65 | unsigned long head; | ||
| 66 | |||
| 67 | again: | ||
| 68 | head = local_read(&rb->head); | ||
| 69 | |||
| 70 | /* | ||
| 71 | * IRQ/NMI can happen here, which means we can miss a head update. | ||
| 72 | */ | ||
| 73 | |||
| 74 | if (!local_dec_and_test(&rb->nest)) | ||
| 75 | goto out; | ||
| 76 | |||
| 77 | /* | ||
| 78 | * Publish the known good head. Rely on the full barrier implied | ||
| 79 | * by atomic_dec_and_test() order the rb->head read and this | ||
| 80 | * write. | ||
| 81 | */ | ||
| 82 | rb->user_page->data_head = head; | ||
| 83 | |||
| 84 | /* | ||
| 85 | * Now check if we missed an update, rely on the (compiler) | ||
| 86 | * barrier in atomic_dec_and_test() to re-read rb->head. | ||
| 87 | */ | ||
| 88 | if (unlikely(head != local_read(&rb->head))) { | ||
| 89 | local_inc(&rb->nest); | ||
| 90 | goto again; | ||
| 91 | } | ||
| 92 | |||
| 93 | if (handle->wakeup != local_read(&rb->wakeup)) | ||
| 94 | perf_output_wakeup(handle); | ||
| 95 | |||
| 96 | out: | ||
| 97 | preempt_enable(); | ||
| 98 | } | ||
| 99 | |||
| 100 | int perf_output_begin(struct perf_output_handle *handle, | ||
| 101 | struct perf_event *event, unsigned int size) | ||
| 102 | { | ||
| 103 | struct ring_buffer *rb; | ||
| 104 | unsigned long tail, offset, head; | ||
| 105 | int have_lost; | ||
| 106 | struct perf_sample_data sample_data; | ||
| 107 | struct { | ||
| 108 | struct perf_event_header header; | ||
| 109 | u64 id; | ||
| 110 | u64 lost; | ||
| 111 | } lost_event; | ||
| 112 | |||
| 113 | rcu_read_lock(); | ||
| 114 | /* | ||
| 115 | * For inherited events we send all the output towards the parent. | ||
| 116 | */ | ||
| 117 | if (event->parent) | ||
| 118 | event = event->parent; | ||
| 119 | |||
| 120 | rb = rcu_dereference(event->rb); | ||
| 121 | if (!rb) | ||
| 122 | goto out; | ||
| 123 | |||
| 124 | handle->rb = rb; | ||
| 125 | handle->event = event; | ||
| 126 | |||
| 127 | if (!rb->nr_pages) | ||
| 128 | goto out; | ||
| 129 | |||
| 130 | have_lost = local_read(&rb->lost); | ||
| 131 | if (have_lost) { | ||
| 132 | lost_event.header.size = sizeof(lost_event); | ||
| 133 | perf_event_header__init_id(&lost_event.header, &sample_data, | ||
| 134 | event); | ||
| 135 | size += lost_event.header.size; | ||
| 136 | } | ||
| 137 | |||
| 138 | perf_output_get_handle(handle); | ||
| 139 | |||
| 140 | do { | ||
| 141 | /* | ||
| 142 | * Userspace could choose to issue a mb() before updating the | ||
| 143 | * tail pointer. So that all reads will be completed before the | ||
| 144 | * write is issued. | ||
| 145 | */ | ||
| 146 | tail = ACCESS_ONCE(rb->user_page->data_tail); | ||
| 147 | smp_rmb(); | ||
| 148 | offset = head = local_read(&rb->head); | ||
| 149 | head += size; | ||
| 150 | if (unlikely(!perf_output_space(rb, tail, offset, head))) | ||
| 151 | goto fail; | ||
| 152 | } while (local_cmpxchg(&rb->head, offset, head) != offset); | ||
| 153 | |||
| 154 | if (head - local_read(&rb->wakeup) > rb->watermark) | ||
| 155 | local_add(rb->watermark, &rb->wakeup); | ||
| 156 | |||
| 157 | handle->page = offset >> (PAGE_SHIFT + page_order(rb)); | ||
| 158 | handle->page &= rb->nr_pages - 1; | ||
| 159 | handle->size = offset & ((PAGE_SIZE << page_order(rb)) - 1); | ||
| 160 | handle->addr = rb->data_pages[handle->page]; | ||
| 161 | handle->addr += handle->size; | ||
| 162 | handle->size = (PAGE_SIZE << page_order(rb)) - handle->size; | ||
| 163 | |||
| 164 | if (have_lost) { | ||
| 165 | lost_event.header.type = PERF_RECORD_LOST; | ||
| 166 | lost_event.header.misc = 0; | ||
| 167 | lost_event.id = event->id; | ||
| 168 | lost_event.lost = local_xchg(&rb->lost, 0); | ||
| 169 | |||
| 170 | perf_output_put(handle, lost_event); | ||
| 171 | perf_event__output_id_sample(event, handle, &sample_data); | ||
| 172 | } | ||
| 173 | |||
| 174 | return 0; | ||
| 175 | |||
| 176 | fail: | ||
| 177 | local_inc(&rb->lost); | ||
| 178 | perf_output_put_handle(handle); | ||
| 179 | out: | ||
| 180 | rcu_read_unlock(); | ||
| 181 | |||
| 182 | return -ENOSPC; | ||
| 183 | } | ||
| 184 | |||
| 185 | void perf_output_copy(struct perf_output_handle *handle, | ||
| 186 | const void *buf, unsigned int len) | ||
| 187 | { | ||
| 188 | __output_copy(handle, buf, len); | ||
| 189 | } | ||
| 190 | |||
| 191 | void perf_output_end(struct perf_output_handle *handle) | ||
| 192 | { | ||
| 193 | perf_output_put_handle(handle); | ||
| 194 | rcu_read_unlock(); | ||
| 195 | } | ||
| 196 | |||
| 197 | static void | ||
| 198 | ring_buffer_init(struct ring_buffer *rb, long watermark, int flags) | ||
| 199 | { | ||
| 200 | long max_size = perf_data_size(rb); | ||
| 201 | |||
| 202 | if (watermark) | ||
| 203 | rb->watermark = min(max_size, watermark); | ||
| 204 | |||
| 205 | if (!rb->watermark) | ||
| 206 | rb->watermark = max_size / 2; | ||
| 207 | |||
| 208 | if (flags & RING_BUFFER_WRITABLE) | ||
| 209 | rb->writable = 1; | ||
| 210 | |||
| 211 | atomic_set(&rb->refcount, 1); | ||
| 212 | } | ||
| 213 | |||
| 214 | #ifndef CONFIG_PERF_USE_VMALLOC | ||
| 215 | |||
| 216 | /* | ||
| 217 | * Back perf_mmap() with regular GFP_KERNEL-0 pages. | ||
| 218 | */ | ||
| 219 | |||
| 220 | struct page * | ||
| 221 | perf_mmap_to_page(struct ring_buffer *rb, unsigned long pgoff) | ||
| 222 | { | ||
| 223 | if (pgoff > rb->nr_pages) | ||
| 224 | return NULL; | ||
| 225 | |||
| 226 | if (pgoff == 0) | ||
| 227 | return virt_to_page(rb->user_page); | ||
| 228 | |||
| 229 | return virt_to_page(rb->data_pages[pgoff - 1]); | ||
| 230 | } | ||
| 231 | |||
| 232 | static void *perf_mmap_alloc_page(int cpu) | ||
| 233 | { | ||
| 234 | struct page *page; | ||
| 235 | int node; | ||
| 236 | |||
| 237 | node = (cpu == -1) ? cpu : cpu_to_node(cpu); | ||
| 238 | page = alloc_pages_node(node, GFP_KERNEL | __GFP_ZERO, 0); | ||
| 239 | if (!page) | ||
| 240 | return NULL; | ||
| 241 | |||
| 242 | return page_address(page); | ||
| 243 | } | ||
| 244 | |||
| 245 | struct ring_buffer *rb_alloc(int nr_pages, long watermark, int cpu, int flags) | ||
| 246 | { | ||
| 247 | struct ring_buffer *rb; | ||
| 248 | unsigned long size; | ||
| 249 | int i; | ||
| 250 | |||
| 251 | size = sizeof(struct ring_buffer); | ||
| 252 | size += nr_pages * sizeof(void *); | ||
| 253 | |||
| 254 | rb = kzalloc(size, GFP_KERNEL); | ||
| 255 | if (!rb) | ||
| 256 | goto fail; | ||
| 257 | |||
| 258 | rb->user_page = perf_mmap_alloc_page(cpu); | ||
| 259 | if (!rb->user_page) | ||
| 260 | goto fail_user_page; | ||
| 261 | |||
| 262 | for (i = 0; i < nr_pages; i++) { | ||
| 263 | rb->data_pages[i] = perf_mmap_alloc_page(cpu); | ||
| 264 | if (!rb->data_pages[i]) | ||
| 265 | goto fail_data_pages; | ||
| 266 | } | ||
| 267 | |||
| 268 | rb->nr_pages = nr_pages; | ||
| 269 | |||
| 270 | ring_buffer_init(rb, watermark, flags); | ||
| 271 | |||
| 272 | return rb; | ||
| 273 | |||
| 274 | fail_data_pages: | ||
| 275 | for (i--; i >= 0; i--) | ||
| 276 | free_page((unsigned long)rb->data_pages[i]); | ||
| 277 | |||
| 278 | free_page((unsigned long)rb->user_page); | ||
| 279 | |||
| 280 | fail_user_page: | ||
| 281 | kfree(rb); | ||
| 282 | |||
| 283 | fail: | ||
| 284 | return NULL; | ||
| 285 | } | ||
| 286 | |||
| 287 | static void perf_mmap_free_page(unsigned long addr) | ||
| 288 | { | ||
| 289 | struct page *page = virt_to_page((void *)addr); | ||
| 290 | |||
| 291 | page->mapping = NULL; | ||
| 292 | __free_page(page); | ||
| 293 | } | ||
| 294 | |||
| 295 | void rb_free(struct ring_buffer *rb) | ||
| 296 | { | ||
| 297 | int i; | ||
| 298 | |||
| 299 | perf_mmap_free_page((unsigned long)rb->user_page); | ||
| 300 | for (i = 0; i < rb->nr_pages; i++) | ||
| 301 | perf_mmap_free_page((unsigned long)rb->data_pages[i]); | ||
| 302 | kfree(rb); | ||
| 303 | } | ||
| 304 | |||
| 305 | #else | ||
| 306 | |||
| 307 | struct page * | ||
| 308 | perf_mmap_to_page(struct ring_buffer *rb, unsigned long pgoff) | ||
| 309 | { | ||
| 310 | if (pgoff > (1UL << page_order(rb))) | ||
| 311 | return NULL; | ||
| 312 | |||
| 313 | return vmalloc_to_page((void *)rb->user_page + pgoff * PAGE_SIZE); | ||
| 314 | } | ||
| 315 | |||
| 316 | static void perf_mmap_unmark_page(void *addr) | ||
| 317 | { | ||
| 318 | struct page *page = vmalloc_to_page(addr); | ||
| 319 | |||
| 320 | page->mapping = NULL; | ||
| 321 | } | ||
| 322 | |||
| 323 | static void rb_free_work(struct work_struct *work) | ||
| 324 | { | ||
| 325 | struct ring_buffer *rb; | ||
| 326 | void *base; | ||
| 327 | int i, nr; | ||
| 328 | |||
| 329 | rb = container_of(work, struct ring_buffer, work); | ||
| 330 | nr = 1 << page_order(rb); | ||
| 331 | |||
| 332 | base = rb->user_page; | ||
| 333 | for (i = 0; i < nr + 1; i++) | ||
| 334 | perf_mmap_unmark_page(base + (i * PAGE_SIZE)); | ||
| 335 | |||
| 336 | vfree(base); | ||
| 337 | kfree(rb); | ||
| 338 | } | ||
| 339 | |||
| 340 | void rb_free(struct ring_buffer *rb) | ||
| 341 | { | ||
| 342 | schedule_work(&rb->work); | ||
| 343 | } | ||
| 344 | |||
| 345 | struct ring_buffer *rb_alloc(int nr_pages, long watermark, int cpu, int flags) | ||
| 346 | { | ||
| 347 | struct ring_buffer *rb; | ||
| 348 | unsigned long size; | ||
| 349 | void *all_buf; | ||
| 350 | |||
| 351 | size = sizeof(struct ring_buffer); | ||
| 352 | size += sizeof(void *); | ||
| 353 | |||
| 354 | rb = kzalloc(size, GFP_KERNEL); | ||
| 355 | if (!rb) | ||
| 356 | goto fail; | ||
| 357 | |||
| 358 | INIT_WORK(&rb->work, rb_free_work); | ||
| 359 | |||
| 360 | all_buf = vmalloc_user((nr_pages + 1) * PAGE_SIZE); | ||
| 361 | if (!all_buf) | ||
| 362 | goto fail_all_buf; | ||
| 363 | |||
| 364 | rb->user_page = all_buf; | ||
| 365 | rb->data_pages[0] = all_buf + PAGE_SIZE; | ||
| 366 | rb->page_order = ilog2(nr_pages); | ||
| 367 | rb->nr_pages = 1; | ||
| 368 | |||
| 369 | ring_buffer_init(rb, watermark, flags); | ||
| 370 | |||
| 371 | return rb; | ||
| 372 | |||
| 373 | fail_all_buf: | ||
| 374 | kfree(rb); | ||
| 375 | |||
| 376 | fail: | ||
| 377 | return NULL; | ||
| 378 | } | ||
| 379 | |||
| 380 | #endif | ||
diff --git a/kernel/kprobes.c b/kernel/kprobes.c index 77981813a1e..b30fd54eb98 100644 --- a/kernel/kprobes.c +++ b/kernel/kprobes.c | |||
| @@ -1255,19 +1255,29 @@ static int __kprobes in_kprobes_functions(unsigned long addr) | |||
| 1255 | /* | 1255 | /* |
| 1256 | * If we have a symbol_name argument, look it up and add the offset field | 1256 | * If we have a symbol_name argument, look it up and add the offset field |
| 1257 | * to it. This way, we can specify a relative address to a symbol. | 1257 | * to it. This way, we can specify a relative address to a symbol. |
| 1258 | * This returns encoded errors if it fails to look up symbol or invalid | ||
| 1259 | * combination of parameters. | ||
| 1258 | */ | 1260 | */ |
| 1259 | static kprobe_opcode_t __kprobes *kprobe_addr(struct kprobe *p) | 1261 | static kprobe_opcode_t __kprobes *kprobe_addr(struct kprobe *p) |
| 1260 | { | 1262 | { |
| 1261 | kprobe_opcode_t *addr = p->addr; | 1263 | kprobe_opcode_t *addr = p->addr; |
| 1264 | |||
| 1265 | if ((p->symbol_name && p->addr) || | ||
| 1266 | (!p->symbol_name && !p->addr)) | ||
| 1267 | goto invalid; | ||
| 1268 | |||
| 1262 | if (p->symbol_name) { | 1269 | if (p->symbol_name) { |
| 1263 | if (addr) | ||
| 1264 | return NULL; | ||
| 1265 | kprobe_lookup_name(p->symbol_name, addr); | 1270 | kprobe_lookup_name(p->symbol_name, addr); |
| 1271 | if (!addr) | ||
| 1272 | return ERR_PTR(-ENOENT); | ||
| 1266 | } | 1273 | } |
| 1267 | 1274 | ||
| 1268 | if (!addr) | 1275 | addr = (kprobe_opcode_t *)(((char *)addr) + p->offset); |
| 1269 | return NULL; | 1276 | if (addr) |
| 1270 | return (kprobe_opcode_t *)(((char *)addr) + p->offset); | 1277 | return addr; |
| 1278 | |||
| 1279 | invalid: | ||
| 1280 | return ERR_PTR(-EINVAL); | ||
| 1271 | } | 1281 | } |
| 1272 | 1282 | ||
| 1273 | /* Check passed kprobe is valid and return kprobe in kprobe_table. */ | 1283 | /* Check passed kprobe is valid and return kprobe in kprobe_table. */ |
| @@ -1311,8 +1321,8 @@ int __kprobes register_kprobe(struct kprobe *p) | |||
| 1311 | kprobe_opcode_t *addr; | 1321 | kprobe_opcode_t *addr; |
| 1312 | 1322 | ||
| 1313 | addr = kprobe_addr(p); | 1323 | addr = kprobe_addr(p); |
| 1314 | if (!addr) | 1324 | if (IS_ERR(addr)) |
| 1315 | return -EINVAL; | 1325 | return PTR_ERR(addr); |
| 1316 | p->addr = addr; | 1326 | p->addr = addr; |
| 1317 | 1327 | ||
| 1318 | ret = check_kprobe_rereg(p); | 1328 | ret = check_kprobe_rereg(p); |
| @@ -1335,6 +1345,8 @@ int __kprobes register_kprobe(struct kprobe *p) | |||
| 1335 | */ | 1345 | */ |
| 1336 | probed_mod = __module_text_address((unsigned long) p->addr); | 1346 | probed_mod = __module_text_address((unsigned long) p->addr); |
| 1337 | if (probed_mod) { | 1347 | if (probed_mod) { |
| 1348 | /* Return -ENOENT if fail. */ | ||
| 1349 | ret = -ENOENT; | ||
| 1338 | /* | 1350 | /* |
| 1339 | * We must hold a refcount of the probed module while updating | 1351 | * We must hold a refcount of the probed module while updating |
| 1340 | * its code to prohibit unexpected unloading. | 1352 | * its code to prohibit unexpected unloading. |
| @@ -1351,6 +1363,7 @@ int __kprobes register_kprobe(struct kprobe *p) | |||
| 1351 | module_put(probed_mod); | 1363 | module_put(probed_mod); |
| 1352 | goto fail_with_jump_label; | 1364 | goto fail_with_jump_label; |
| 1353 | } | 1365 | } |
| 1366 | /* ret will be updated by following code */ | ||
| 1354 | } | 1367 | } |
| 1355 | preempt_enable(); | 1368 | preempt_enable(); |
| 1356 | jump_label_unlock(); | 1369 | jump_label_unlock(); |
| @@ -1399,7 +1412,7 @@ out: | |||
| 1399 | fail_with_jump_label: | 1412 | fail_with_jump_label: |
| 1400 | preempt_enable(); | 1413 | preempt_enable(); |
| 1401 | jump_label_unlock(); | 1414 | jump_label_unlock(); |
| 1402 | return -EINVAL; | 1415 | return ret; |
| 1403 | } | 1416 | } |
| 1404 | EXPORT_SYMBOL_GPL(register_kprobe); | 1417 | EXPORT_SYMBOL_GPL(register_kprobe); |
| 1405 | 1418 | ||
| @@ -1686,8 +1699,8 @@ int __kprobes register_kretprobe(struct kretprobe *rp) | |||
| 1686 | 1699 | ||
| 1687 | if (kretprobe_blacklist_size) { | 1700 | if (kretprobe_blacklist_size) { |
| 1688 | addr = kprobe_addr(&rp->kp); | 1701 | addr = kprobe_addr(&rp->kp); |
| 1689 | if (!addr) | 1702 | if (IS_ERR(addr)) |
| 1690 | return -EINVAL; | 1703 | return PTR_ERR(addr); |
| 1691 | 1704 | ||
| 1692 | for (i = 0; kretprobe_blacklist[i].name != NULL; i++) { | 1705 | for (i = 0; kretprobe_blacklist[i].name != NULL; i++) { |
| 1693 | if (kretprobe_blacklist[i].addr == addr) | 1706 | if (kretprobe_blacklist[i].addr == addr) |
diff --git a/kernel/sched.c b/kernel/sched.c index c518b05fd06..84b9e076812 100644 --- a/kernel/sched.c +++ b/kernel/sched.c | |||
| @@ -2220,7 +2220,7 @@ void set_task_cpu(struct task_struct *p, unsigned int new_cpu) | |||
| 2220 | 2220 | ||
| 2221 | if (task_cpu(p) != new_cpu) { | 2221 | if (task_cpu(p) != new_cpu) { |
| 2222 | p->se.nr_migrations++; | 2222 | p->se.nr_migrations++; |
| 2223 | perf_sw_event(PERF_COUNT_SW_CPU_MIGRATIONS, 1, 1, NULL, 0); | 2223 | perf_sw_event(PERF_COUNT_SW_CPU_MIGRATIONS, 1, NULL, 0); |
| 2224 | } | 2224 | } |
| 2225 | 2225 | ||
| 2226 | __set_task_cpu(p, new_cpu); | 2226 | __set_task_cpu(p, new_cpu); |
diff --git a/kernel/stacktrace.c b/kernel/stacktrace.c index eb212f8f8bc..d20c6983aad 100644 --- a/kernel/stacktrace.c +++ b/kernel/stacktrace.c | |||
| @@ -26,12 +26,18 @@ void print_stack_trace(struct stack_trace *trace, int spaces) | |||
| 26 | EXPORT_SYMBOL_GPL(print_stack_trace); | 26 | EXPORT_SYMBOL_GPL(print_stack_trace); |
| 27 | 27 | ||
| 28 | /* | 28 | /* |
| 29 | * Architectures that do not implement save_stack_trace_tsk get this | 29 | * Architectures that do not implement save_stack_trace_tsk or |
| 30 | * weak alias and a once-per-bootup warning (whenever this facility | 30 | * save_stack_trace_regs get this weak alias and a once-per-bootup warning |
| 31 | * is utilized - for example by procfs): | 31 | * (whenever this facility is utilized - for example by procfs): |
| 32 | */ | 32 | */ |
| 33 | __weak void | 33 | __weak void |
| 34 | save_stack_trace_tsk(struct task_struct *tsk, struct stack_trace *trace) | 34 | save_stack_trace_tsk(struct task_struct *tsk, struct stack_trace *trace) |
| 35 | { | 35 | { |
| 36 | WARN_ONCE(1, KERN_INFO "save_stack_trace_tsk() not implemented yet.\n"); | 36 | WARN_ONCE(1, KERN_INFO "save_stack_trace_tsk() not implemented yet.\n"); |
| 37 | } | 37 | } |
| 38 | |||
| 39 | __weak void | ||
| 40 | save_stack_trace_regs(struct pt_regs *regs, struct stack_trace *trace) | ||
| 41 | { | ||
| 42 | WARN_ONCE(1, KERN_INFO "save_stack_trace_regs() not implemented yet.\n"); | ||
| 43 | } | ||
diff --git a/kernel/trace/ftrace.c b/kernel/trace/ftrace.c index 908038f5744..c3e4575e782 100644 --- a/kernel/trace/ftrace.c +++ b/kernel/trace/ftrace.c | |||
| @@ -32,7 +32,6 @@ | |||
| 32 | 32 | ||
| 33 | #include <trace/events/sched.h> | 33 | #include <trace/events/sched.h> |
| 34 | 34 | ||
| 35 | #include <asm/ftrace.h> | ||
| 36 | #include <asm/setup.h> | 35 | #include <asm/setup.h> |
| 37 | 36 | ||
| 38 | #include "trace_output.h" | 37 | #include "trace_output.h" |
| @@ -82,14 +81,14 @@ static int ftrace_disabled __read_mostly; | |||
| 82 | 81 | ||
| 83 | static DEFINE_MUTEX(ftrace_lock); | 82 | static DEFINE_MUTEX(ftrace_lock); |
| 84 | 83 | ||
| 85 | static struct ftrace_ops ftrace_list_end __read_mostly = | 84 | static struct ftrace_ops ftrace_list_end __read_mostly = { |
| 86 | { | ||
| 87 | .func = ftrace_stub, | 85 | .func = ftrace_stub, |
| 88 | }; | 86 | }; |
| 89 | 87 | ||
| 90 | static struct ftrace_ops *ftrace_global_list __read_mostly = &ftrace_list_end; | 88 | static struct ftrace_ops *ftrace_global_list __read_mostly = &ftrace_list_end; |
| 91 | static struct ftrace_ops *ftrace_ops_list __read_mostly = &ftrace_list_end; | 89 | static struct ftrace_ops *ftrace_ops_list __read_mostly = &ftrace_list_end; |
| 92 | ftrace_func_t ftrace_trace_function __read_mostly = ftrace_stub; | 90 | ftrace_func_t ftrace_trace_function __read_mostly = ftrace_stub; |
| 91 | static ftrace_func_t __ftrace_trace_function_delay __read_mostly = ftrace_stub; | ||
| 93 | ftrace_func_t __ftrace_trace_function __read_mostly = ftrace_stub; | 92 | ftrace_func_t __ftrace_trace_function __read_mostly = ftrace_stub; |
| 94 | ftrace_func_t ftrace_pid_function __read_mostly = ftrace_stub; | 93 | ftrace_func_t ftrace_pid_function __read_mostly = ftrace_stub; |
| 95 | static struct ftrace_ops global_ops; | 94 | static struct ftrace_ops global_ops; |
| @@ -148,9 +147,11 @@ void clear_ftrace_function(void) | |||
| 148 | { | 147 | { |
| 149 | ftrace_trace_function = ftrace_stub; | 148 | ftrace_trace_function = ftrace_stub; |
| 150 | __ftrace_trace_function = ftrace_stub; | 149 | __ftrace_trace_function = ftrace_stub; |
| 150 | __ftrace_trace_function_delay = ftrace_stub; | ||
| 151 | ftrace_pid_function = ftrace_stub; | 151 | ftrace_pid_function = ftrace_stub; |
| 152 | } | 152 | } |
| 153 | 153 | ||
| 154 | #undef CONFIG_HAVE_FUNCTION_TRACE_MCOUNT_TEST | ||
| 154 | #ifndef CONFIG_HAVE_FUNCTION_TRACE_MCOUNT_TEST | 155 | #ifndef CONFIG_HAVE_FUNCTION_TRACE_MCOUNT_TEST |
| 155 | /* | 156 | /* |
| 156 | * For those archs that do not test ftrace_trace_stop in their | 157 | * For those archs that do not test ftrace_trace_stop in their |
| @@ -210,7 +211,12 @@ static void update_ftrace_function(void) | |||
| 210 | #ifdef CONFIG_HAVE_FUNCTION_TRACE_MCOUNT_TEST | 211 | #ifdef CONFIG_HAVE_FUNCTION_TRACE_MCOUNT_TEST |
| 211 | ftrace_trace_function = func; | 212 | ftrace_trace_function = func; |
| 212 | #else | 213 | #else |
| 214 | #ifdef CONFIG_DYNAMIC_FTRACE | ||
| 215 | /* do not update till all functions have been modified */ | ||
| 216 | __ftrace_trace_function_delay = func; | ||
| 217 | #else | ||
| 213 | __ftrace_trace_function = func; | 218 | __ftrace_trace_function = func; |
| 219 | #endif | ||
| 214 | ftrace_trace_function = ftrace_test_stop_func; | 220 | ftrace_trace_function = ftrace_test_stop_func; |
| 215 | #endif | 221 | #endif |
| 216 | } | 222 | } |
| @@ -785,8 +791,7 @@ static void unregister_ftrace_profiler(void) | |||
| 785 | unregister_ftrace_graph(); | 791 | unregister_ftrace_graph(); |
| 786 | } | 792 | } |
| 787 | #else | 793 | #else |
| 788 | static struct ftrace_ops ftrace_profile_ops __read_mostly = | 794 | static struct ftrace_ops ftrace_profile_ops __read_mostly = { |
| 789 | { | ||
| 790 | .func = function_profile_call, | 795 | .func = function_profile_call, |
| 791 | }; | 796 | }; |
| 792 | 797 | ||
| @@ -806,19 +811,10 @@ ftrace_profile_write(struct file *filp, const char __user *ubuf, | |||
| 806 | size_t cnt, loff_t *ppos) | 811 | size_t cnt, loff_t *ppos) |
| 807 | { | 812 | { |
| 808 | unsigned long val; | 813 | unsigned long val; |
| 809 | char buf[64]; /* big enough to hold a number */ | ||
| 810 | int ret; | 814 | int ret; |
| 811 | 815 | ||
| 812 | if (cnt >= sizeof(buf)) | 816 | ret = kstrtoul_from_user(ubuf, cnt, 10, &val); |
| 813 | return -EINVAL; | 817 | if (ret) |
| 814 | |||
| 815 | if (copy_from_user(&buf, ubuf, cnt)) | ||
| 816 | return -EFAULT; | ||
| 817 | |||
| 818 | buf[cnt] = 0; | ||
| 819 | |||
| 820 | ret = strict_strtoul(buf, 10, &val); | ||
| 821 | if (ret < 0) | ||
| 822 | return ret; | 818 | return ret; |
| 823 | 819 | ||
| 824 | val = !!val; | 820 | val = !!val; |
| @@ -1182,8 +1178,14 @@ alloc_and_copy_ftrace_hash(int size_bits, struct ftrace_hash *hash) | |||
| 1182 | return NULL; | 1178 | return NULL; |
| 1183 | } | 1179 | } |
| 1184 | 1180 | ||
| 1181 | static void | ||
| 1182 | ftrace_hash_rec_disable(struct ftrace_ops *ops, int filter_hash); | ||
| 1183 | static void | ||
| 1184 | ftrace_hash_rec_enable(struct ftrace_ops *ops, int filter_hash); | ||
| 1185 | |||
| 1185 | static int | 1186 | static int |
| 1186 | ftrace_hash_move(struct ftrace_hash **dst, struct ftrace_hash *src) | 1187 | ftrace_hash_move(struct ftrace_ops *ops, int enable, |
| 1188 | struct ftrace_hash **dst, struct ftrace_hash *src) | ||
| 1187 | { | 1189 | { |
| 1188 | struct ftrace_func_entry *entry; | 1190 | struct ftrace_func_entry *entry; |
| 1189 | struct hlist_node *tp, *tn; | 1191 | struct hlist_node *tp, *tn; |
| @@ -1193,9 +1195,16 @@ ftrace_hash_move(struct ftrace_hash **dst, struct ftrace_hash *src) | |||
| 1193 | unsigned long key; | 1195 | unsigned long key; |
| 1194 | int size = src->count; | 1196 | int size = src->count; |
| 1195 | int bits = 0; | 1197 | int bits = 0; |
| 1198 | int ret; | ||
| 1196 | int i; | 1199 | int i; |
| 1197 | 1200 | ||
| 1198 | /* | 1201 | /* |
| 1202 | * Remove the current set, update the hash and add | ||
| 1203 | * them back. | ||
| 1204 | */ | ||
| 1205 | ftrace_hash_rec_disable(ops, enable); | ||
| 1206 | |||
| 1207 | /* | ||
| 1199 | * If the new source is empty, just free dst and assign it | 1208 | * If the new source is empty, just free dst and assign it |
| 1200 | * the empty_hash. | 1209 | * the empty_hash. |
| 1201 | */ | 1210 | */ |
| @@ -1215,9 +1224,10 @@ ftrace_hash_move(struct ftrace_hash **dst, struct ftrace_hash *src) | |||
| 1215 | if (bits > FTRACE_HASH_MAX_BITS) | 1224 | if (bits > FTRACE_HASH_MAX_BITS) |
| 1216 | bits = FTRACE_HASH_MAX_BITS; | 1225 | bits = FTRACE_HASH_MAX_BITS; |
| 1217 | 1226 | ||
| 1227 | ret = -ENOMEM; | ||
| 1218 | new_hash = alloc_ftrace_hash(bits); | 1228 | new_hash = alloc_ftrace_hash(bits); |
| 1219 | if (!new_hash) | 1229 | if (!new_hash) |
| 1220 | return -ENOMEM; | 1230 | goto out; |
| 1221 | 1231 | ||
| 1222 | size = 1 << src->size_bits; | 1232 | size = 1 << src->size_bits; |
| 1223 | for (i = 0; i < size; i++) { | 1233 | for (i = 0; i < size; i++) { |
| @@ -1236,7 +1246,16 @@ ftrace_hash_move(struct ftrace_hash **dst, struct ftrace_hash *src) | |||
| 1236 | rcu_assign_pointer(*dst, new_hash); | 1246 | rcu_assign_pointer(*dst, new_hash); |
| 1237 | free_ftrace_hash_rcu(old_hash); | 1247 | free_ftrace_hash_rcu(old_hash); |
| 1238 | 1248 | ||
| 1239 | return 0; | 1249 | ret = 0; |
| 1250 | out: | ||
| 1251 | /* | ||
| 1252 | * Enable regardless of ret: | ||
| 1253 | * On success, we enable the new hash. | ||
| 1254 | * On failure, we re-enable the original hash. | ||
| 1255 | */ | ||
| 1256 | ftrace_hash_rec_enable(ops, enable); | ||
| 1257 | |||
| 1258 | return ret; | ||
| 1240 | } | 1259 | } |
| 1241 | 1260 | ||
| 1242 | /* | 1261 | /* |
| @@ -1596,6 +1615,12 @@ static int __ftrace_modify_code(void *data) | |||
| 1596 | { | 1615 | { |
| 1597 | int *command = data; | 1616 | int *command = data; |
| 1598 | 1617 | ||
| 1618 | /* | ||
| 1619 | * Do not call function tracer while we update the code. | ||
| 1620 | * We are in stop machine, no worrying about races. | ||
| 1621 | */ | ||
| 1622 | function_trace_stop++; | ||
| 1623 | |||
| 1599 | if (*command & FTRACE_ENABLE_CALLS) | 1624 | if (*command & FTRACE_ENABLE_CALLS) |
| 1600 | ftrace_replace_code(1); | 1625 | ftrace_replace_code(1); |
| 1601 | else if (*command & FTRACE_DISABLE_CALLS) | 1626 | else if (*command & FTRACE_DISABLE_CALLS) |
| @@ -1609,6 +1634,18 @@ static int __ftrace_modify_code(void *data) | |||
| 1609 | else if (*command & FTRACE_STOP_FUNC_RET) | 1634 | else if (*command & FTRACE_STOP_FUNC_RET) |
| 1610 | ftrace_disable_ftrace_graph_caller(); | 1635 | ftrace_disable_ftrace_graph_caller(); |
| 1611 | 1636 | ||
| 1637 | #ifndef CONFIG_HAVE_FUNCTION_TRACE_MCOUNT_TEST | ||
| 1638 | /* | ||
| 1639 | * For archs that call ftrace_test_stop_func(), we must | ||
| 1640 | * wait till after we update all the function callers | ||
| 1641 | * before we update the callback. This keeps different | ||
| 1642 | * ops that record different functions from corrupting | ||
| 1643 | * each other. | ||
| 1644 | */ | ||
| 1645 | __ftrace_trace_function = __ftrace_trace_function_delay; | ||
| 1646 | #endif | ||
| 1647 | function_trace_stop--; | ||
| 1648 | |||
| 1612 | return 0; | 1649 | return 0; |
| 1613 | } | 1650 | } |
| 1614 | 1651 | ||
| @@ -1744,10 +1781,36 @@ static cycle_t ftrace_update_time; | |||
| 1744 | static unsigned long ftrace_update_cnt; | 1781 | static unsigned long ftrace_update_cnt; |
| 1745 | unsigned long ftrace_update_tot_cnt; | 1782 | unsigned long ftrace_update_tot_cnt; |
| 1746 | 1783 | ||
| 1784 | static int ops_traces_mod(struct ftrace_ops *ops) | ||
| 1785 | { | ||
| 1786 | struct ftrace_hash *hash; | ||
| 1787 | |||
| 1788 | hash = ops->filter_hash; | ||
| 1789 | return !!(!hash || !hash->count); | ||
| 1790 | } | ||
| 1791 | |||
| 1747 | static int ftrace_update_code(struct module *mod) | 1792 | static int ftrace_update_code(struct module *mod) |
| 1748 | { | 1793 | { |
| 1749 | struct dyn_ftrace *p; | 1794 | struct dyn_ftrace *p; |
| 1750 | cycle_t start, stop; | 1795 | cycle_t start, stop; |
| 1796 | unsigned long ref = 0; | ||
| 1797 | |||
| 1798 | /* | ||
| 1799 | * When adding a module, we need to check if tracers are | ||
| 1800 | * currently enabled and if they are set to trace all functions. | ||
| 1801 | * If they are, we need to enable the module functions as well | ||
| 1802 | * as update the reference counts for those function records. | ||
| 1803 | */ | ||
| 1804 | if (mod) { | ||
| 1805 | struct ftrace_ops *ops; | ||
| 1806 | |||
| 1807 | for (ops = ftrace_ops_list; | ||
| 1808 | ops != &ftrace_list_end; ops = ops->next) { | ||
| 1809 | if (ops->flags & FTRACE_OPS_FL_ENABLED && | ||
| 1810 | ops_traces_mod(ops)) | ||
| 1811 | ref++; | ||
| 1812 | } | ||
| 1813 | } | ||
| 1751 | 1814 | ||
| 1752 | start = ftrace_now(raw_smp_processor_id()); | 1815 | start = ftrace_now(raw_smp_processor_id()); |
| 1753 | ftrace_update_cnt = 0; | 1816 | ftrace_update_cnt = 0; |
| @@ -1760,7 +1823,7 @@ static int ftrace_update_code(struct module *mod) | |||
| 1760 | 1823 | ||
| 1761 | p = ftrace_new_addrs; | 1824 | p = ftrace_new_addrs; |
| 1762 | ftrace_new_addrs = p->newlist; | 1825 | ftrace_new_addrs = p->newlist; |
| 1763 | p->flags = 0L; | 1826 | p->flags = ref; |
| 1764 | 1827 | ||
| 1765 | /* | 1828 | /* |
| 1766 | * Do the initial record conversion from mcount jump | 1829 | * Do the initial record conversion from mcount jump |
| @@ -1783,7 +1846,7 @@ static int ftrace_update_code(struct module *mod) | |||
| 1783 | * conversion puts the module to the correct state, thus | 1846 | * conversion puts the module to the correct state, thus |
| 1784 | * passing the ftrace_make_call check. | 1847 | * passing the ftrace_make_call check. |
| 1785 | */ | 1848 | */ |
| 1786 | if (ftrace_start_up) { | 1849 | if (ftrace_start_up && ref) { |
| 1787 | int failed = __ftrace_replace_code(p, 1); | 1850 | int failed = __ftrace_replace_code(p, 1); |
| 1788 | if (failed) { | 1851 | if (failed) { |
| 1789 | ftrace_bug(failed, p->ip); | 1852 | ftrace_bug(failed, p->ip); |
| @@ -2407,10 +2470,9 @@ ftrace_match_module_records(struct ftrace_hash *hash, char *buff, char *mod) | |||
| 2407 | */ | 2470 | */ |
| 2408 | 2471 | ||
| 2409 | static int | 2472 | static int |
| 2410 | ftrace_mod_callback(char *func, char *cmd, char *param, int enable) | 2473 | ftrace_mod_callback(struct ftrace_hash *hash, |
| 2474 | char *func, char *cmd, char *param, int enable) | ||
| 2411 | { | 2475 | { |
| 2412 | struct ftrace_ops *ops = &global_ops; | ||
| 2413 | struct ftrace_hash *hash; | ||
| 2414 | char *mod; | 2476 | char *mod; |
| 2415 | int ret = -EINVAL; | 2477 | int ret = -EINVAL; |
| 2416 | 2478 | ||
| @@ -2430,11 +2492,6 @@ ftrace_mod_callback(char *func, char *cmd, char *param, int enable) | |||
| 2430 | if (!strlen(mod)) | 2492 | if (!strlen(mod)) |
| 2431 | return ret; | 2493 | return ret; |
| 2432 | 2494 | ||
| 2433 | if (enable) | ||
| 2434 | hash = ops->filter_hash; | ||
| 2435 | else | ||
| 2436 | hash = ops->notrace_hash; | ||
| 2437 | |||
| 2438 | ret = ftrace_match_module_records(hash, func, mod); | 2495 | ret = ftrace_match_module_records(hash, func, mod); |
| 2439 | if (!ret) | 2496 | if (!ret) |
| 2440 | ret = -EINVAL; | 2497 | ret = -EINVAL; |
| @@ -2760,7 +2817,7 @@ static int ftrace_process_regex(struct ftrace_hash *hash, | |||
| 2760 | mutex_lock(&ftrace_cmd_mutex); | 2817 | mutex_lock(&ftrace_cmd_mutex); |
| 2761 | list_for_each_entry(p, &ftrace_commands, list) { | 2818 | list_for_each_entry(p, &ftrace_commands, list) { |
| 2762 | if (strcmp(p->name, command) == 0) { | 2819 | if (strcmp(p->name, command) == 0) { |
| 2763 | ret = p->func(func, command, next, enable); | 2820 | ret = p->func(hash, func, command, next, enable); |
| 2764 | goto out_unlock; | 2821 | goto out_unlock; |
| 2765 | } | 2822 | } |
| 2766 | } | 2823 | } |
| @@ -2857,7 +2914,11 @@ ftrace_set_regex(struct ftrace_ops *ops, unsigned char *buf, int len, | |||
| 2857 | ftrace_match_records(hash, buf, len); | 2914 | ftrace_match_records(hash, buf, len); |
| 2858 | 2915 | ||
| 2859 | mutex_lock(&ftrace_lock); | 2916 | mutex_lock(&ftrace_lock); |
| 2860 | ret = ftrace_hash_move(orig_hash, hash); | 2917 | ret = ftrace_hash_move(ops, enable, orig_hash, hash); |
| 2918 | if (!ret && ops->flags & FTRACE_OPS_FL_ENABLED | ||
| 2919 | && ftrace_enabled) | ||
| 2920 | ftrace_run_update_code(FTRACE_ENABLE_CALLS); | ||
| 2921 | |||
| 2861 | mutex_unlock(&ftrace_lock); | 2922 | mutex_unlock(&ftrace_lock); |
| 2862 | 2923 | ||
| 2863 | mutex_unlock(&ftrace_regex_lock); | 2924 | mutex_unlock(&ftrace_regex_lock); |
| @@ -3040,18 +3101,12 @@ ftrace_regex_release(struct inode *inode, struct file *file) | |||
| 3040 | orig_hash = &iter->ops->notrace_hash; | 3101 | orig_hash = &iter->ops->notrace_hash; |
| 3041 | 3102 | ||
| 3042 | mutex_lock(&ftrace_lock); | 3103 | mutex_lock(&ftrace_lock); |
| 3043 | /* | 3104 | ret = ftrace_hash_move(iter->ops, filter_hash, |
| 3044 | * Remove the current set, update the hash and add | 3105 | orig_hash, iter->hash); |
| 3045 | * them back. | 3106 | if (!ret && (iter->ops->flags & FTRACE_OPS_FL_ENABLED) |
| 3046 | */ | 3107 | && ftrace_enabled) |
| 3047 | ftrace_hash_rec_disable(iter->ops, filter_hash); | 3108 | ftrace_run_update_code(FTRACE_ENABLE_CALLS); |
| 3048 | ret = ftrace_hash_move(orig_hash, iter->hash); | 3109 | |
| 3049 | if (!ret) { | ||
| 3050 | ftrace_hash_rec_enable(iter->ops, filter_hash); | ||
| 3051 | if (iter->ops->flags & FTRACE_OPS_FL_ENABLED | ||
| 3052 | && ftrace_enabled) | ||
| 3053 | ftrace_run_update_code(FTRACE_ENABLE_CALLS); | ||
| 3054 | } | ||
| 3055 | mutex_unlock(&ftrace_lock); | 3110 | mutex_unlock(&ftrace_lock); |
| 3056 | } | 3111 | } |
| 3057 | free_ftrace_hash(iter->hash); | 3112 | free_ftrace_hash(iter->hash); |
| @@ -3330,7 +3385,7 @@ static int ftrace_process_locs(struct module *mod, | |||
| 3330 | { | 3385 | { |
| 3331 | unsigned long *p; | 3386 | unsigned long *p; |
| 3332 | unsigned long addr; | 3387 | unsigned long addr; |
| 3333 | unsigned long flags; | 3388 | unsigned long flags = 0; /* Shut up gcc */ |
| 3334 | 3389 | ||
| 3335 | mutex_lock(&ftrace_lock); | 3390 | mutex_lock(&ftrace_lock); |
| 3336 | p = start; | 3391 | p = start; |
| @@ -3348,12 +3403,18 @@ static int ftrace_process_locs(struct module *mod, | |||
| 3348 | } | 3403 | } |
| 3349 | 3404 | ||
| 3350 | /* | 3405 | /* |
| 3351 | * Disable interrupts to prevent interrupts from executing | 3406 | * We only need to disable interrupts on start up |
| 3352 | * code that is being modified. | 3407 | * because we are modifying code that an interrupt |
| 3408 | * may execute, and the modification is not atomic. | ||
| 3409 | * But for modules, nothing runs the code we modify | ||
| 3410 | * until we are finished with it, and there's no | ||
| 3411 | * reason to cause large interrupt latencies while we do it. | ||
| 3353 | */ | 3412 | */ |
| 3354 | local_irq_save(flags); | 3413 | if (!mod) |
| 3414 | local_irq_save(flags); | ||
| 3355 | ftrace_update_code(mod); | 3415 | ftrace_update_code(mod); |
| 3356 | local_irq_restore(flags); | 3416 | if (!mod) |
| 3417 | local_irq_restore(flags); | ||
| 3357 | mutex_unlock(&ftrace_lock); | 3418 | mutex_unlock(&ftrace_lock); |
| 3358 | 3419 | ||
| 3359 | return 0; | 3420 | return 0; |
diff --git a/kernel/trace/ring_buffer.c b/kernel/trace/ring_buffer.c index b0c7aa40794..731201bf4ac 100644 --- a/kernel/trace/ring_buffer.c +++ b/kernel/trace/ring_buffer.c | |||
| @@ -997,15 +997,21 @@ static int rb_allocate_pages(struct ring_buffer_per_cpu *cpu_buffer, | |||
| 997 | unsigned nr_pages) | 997 | unsigned nr_pages) |
| 998 | { | 998 | { |
| 999 | struct buffer_page *bpage, *tmp; | 999 | struct buffer_page *bpage, *tmp; |
| 1000 | unsigned long addr; | ||
| 1001 | LIST_HEAD(pages); | 1000 | LIST_HEAD(pages); |
| 1002 | unsigned i; | 1001 | unsigned i; |
| 1003 | 1002 | ||
| 1004 | WARN_ON(!nr_pages); | 1003 | WARN_ON(!nr_pages); |
| 1005 | 1004 | ||
| 1006 | for (i = 0; i < nr_pages; i++) { | 1005 | for (i = 0; i < nr_pages; i++) { |
| 1006 | struct page *page; | ||
| 1007 | /* | ||
| 1008 | * __GFP_NORETRY flag makes sure that the allocation fails | ||
| 1009 | * gracefully without invoking oom-killer and the system is | ||
| 1010 | * not destabilized. | ||
| 1011 | */ | ||
| 1007 | bpage = kzalloc_node(ALIGN(sizeof(*bpage), cache_line_size()), | 1012 | bpage = kzalloc_node(ALIGN(sizeof(*bpage), cache_line_size()), |
| 1008 | GFP_KERNEL, cpu_to_node(cpu_buffer->cpu)); | 1013 | GFP_KERNEL | __GFP_NORETRY, |
| 1014 | cpu_to_node(cpu_buffer->cpu)); | ||
| 1009 | if (!bpage) | 1015 | if (!bpage) |
| 1010 | goto free_pages; | 1016 | goto free_pages; |
| 1011 | 1017 | ||
| @@ -1013,10 +1019,11 @@ static int rb_allocate_pages(struct ring_buffer_per_cpu *cpu_buffer, | |||
| 1013 | 1019 | ||
| 1014 | list_add(&bpage->list, &pages); | 1020 | list_add(&bpage->list, &pages); |
| 1015 | 1021 | ||
| 1016 | addr = __get_free_page(GFP_KERNEL); | 1022 | page = alloc_pages_node(cpu_to_node(cpu_buffer->cpu), |
| 1017 | if (!addr) | 1023 | GFP_KERNEL | __GFP_NORETRY, 0); |
| 1024 | if (!page) | ||
| 1018 | goto free_pages; | 1025 | goto free_pages; |
| 1019 | bpage->page = (void *)addr; | 1026 | bpage->page = page_address(page); |
| 1020 | rb_init_page(bpage->page); | 1027 | rb_init_page(bpage->page); |
| 1021 | } | 1028 | } |
| 1022 | 1029 | ||
| @@ -1045,7 +1052,7 @@ rb_allocate_cpu_buffer(struct ring_buffer *buffer, int cpu) | |||
| 1045 | { | 1052 | { |
| 1046 | struct ring_buffer_per_cpu *cpu_buffer; | 1053 | struct ring_buffer_per_cpu *cpu_buffer; |
| 1047 | struct buffer_page *bpage; | 1054 | struct buffer_page *bpage; |
| 1048 | unsigned long addr; | 1055 | struct page *page; |
| 1049 | int ret; | 1056 | int ret; |
| 1050 | 1057 | ||
| 1051 | cpu_buffer = kzalloc_node(ALIGN(sizeof(*cpu_buffer), cache_line_size()), | 1058 | cpu_buffer = kzalloc_node(ALIGN(sizeof(*cpu_buffer), cache_line_size()), |
| @@ -1067,10 +1074,10 @@ rb_allocate_cpu_buffer(struct ring_buffer *buffer, int cpu) | |||
| 1067 | rb_check_bpage(cpu_buffer, bpage); | 1074 | rb_check_bpage(cpu_buffer, bpage); |
| 1068 | 1075 | ||
| 1069 | cpu_buffer->reader_page = bpage; | 1076 | cpu_buffer->reader_page = bpage; |
| 1070 | addr = __get_free_page(GFP_KERNEL); | 1077 | page = alloc_pages_node(cpu_to_node(cpu), GFP_KERNEL, 0); |
| 1071 | if (!addr) | 1078 | if (!page) |
| 1072 | goto fail_free_reader; | 1079 | goto fail_free_reader; |
| 1073 | bpage->page = (void *)addr; | 1080 | bpage->page = page_address(page); |
| 1074 | rb_init_page(bpage->page); | 1081 | rb_init_page(bpage->page); |
| 1075 | 1082 | ||
| 1076 | INIT_LIST_HEAD(&cpu_buffer->reader_page->list); | 1083 | INIT_LIST_HEAD(&cpu_buffer->reader_page->list); |
| @@ -1314,7 +1321,6 @@ int ring_buffer_resize(struct ring_buffer *buffer, unsigned long size) | |||
| 1314 | unsigned nr_pages, rm_pages, new_pages; | 1321 | unsigned nr_pages, rm_pages, new_pages; |
| 1315 | struct buffer_page *bpage, *tmp; | 1322 | struct buffer_page *bpage, *tmp; |
| 1316 | unsigned long buffer_size; | 1323 | unsigned long buffer_size; |
| 1317 | unsigned long addr; | ||
| 1318 | LIST_HEAD(pages); | 1324 | LIST_HEAD(pages); |
| 1319 | int i, cpu; | 1325 | int i, cpu; |
| 1320 | 1326 | ||
| @@ -1375,16 +1381,24 @@ int ring_buffer_resize(struct ring_buffer *buffer, unsigned long size) | |||
| 1375 | 1381 | ||
| 1376 | for_each_buffer_cpu(buffer, cpu) { | 1382 | for_each_buffer_cpu(buffer, cpu) { |
| 1377 | for (i = 0; i < new_pages; i++) { | 1383 | for (i = 0; i < new_pages; i++) { |
| 1384 | struct page *page; | ||
| 1385 | /* | ||
| 1386 | * __GFP_NORETRY flag makes sure that the allocation | ||
| 1387 | * fails gracefully without invoking oom-killer and | ||
| 1388 | * the system is not destabilized. | ||
| 1389 | */ | ||
| 1378 | bpage = kzalloc_node(ALIGN(sizeof(*bpage), | 1390 | bpage = kzalloc_node(ALIGN(sizeof(*bpage), |
| 1379 | cache_line_size()), | 1391 | cache_line_size()), |
| 1380 | GFP_KERNEL, cpu_to_node(cpu)); | 1392 | GFP_KERNEL | __GFP_NORETRY, |
| 1393 | cpu_to_node(cpu)); | ||
| 1381 | if (!bpage) | 1394 | if (!bpage) |
| 1382 | goto free_pages; | 1395 | goto free_pages; |
| 1383 | list_add(&bpage->list, &pages); | 1396 | list_add(&bpage->list, &pages); |
| 1384 | addr = __get_free_page(GFP_KERNEL); | 1397 | page = alloc_pages_node(cpu_to_node(cpu), |
| 1385 | if (!addr) | 1398 | GFP_KERNEL | __GFP_NORETRY, 0); |
| 1399 | if (!page) | ||
| 1386 | goto free_pages; | 1400 | goto free_pages; |
| 1387 | bpage->page = (void *)addr; | 1401 | bpage->page = page_address(page); |
| 1388 | rb_init_page(bpage->page); | 1402 | rb_init_page(bpage->page); |
| 1389 | } | 1403 | } |
| 1390 | } | 1404 | } |
| @@ -3730,16 +3744,17 @@ EXPORT_SYMBOL_GPL(ring_buffer_swap_cpu); | |||
| 3730 | * Returns: | 3744 | * Returns: |
| 3731 | * The page allocated, or NULL on error. | 3745 | * The page allocated, or NULL on error. |
| 3732 | */ | 3746 | */ |
| 3733 | void *ring_buffer_alloc_read_page(struct ring_buffer *buffer) | 3747 | void *ring_buffer_alloc_read_page(struct ring_buffer *buffer, int cpu) |
| 3734 | { | 3748 | { |
| 3735 | struct buffer_data_page *bpage; | 3749 | struct buffer_data_page *bpage; |
| 3736 | unsigned long addr; | 3750 | struct page *page; |
| 3737 | 3751 | ||
| 3738 | addr = __get_free_page(GFP_KERNEL); | 3752 | page = alloc_pages_node(cpu_to_node(cpu), |
| 3739 | if (!addr) | 3753 | GFP_KERNEL | __GFP_NORETRY, 0); |
| 3754 | if (!page) | ||
| 3740 | return NULL; | 3755 | return NULL; |
| 3741 | 3756 | ||
| 3742 | bpage = (void *)addr; | 3757 | bpage = page_address(page); |
| 3743 | 3758 | ||
| 3744 | rb_init_page(bpage); | 3759 | rb_init_page(bpage); |
| 3745 | 3760 | ||
| @@ -3978,20 +3993,11 @@ rb_simple_write(struct file *filp, const char __user *ubuf, | |||
| 3978 | size_t cnt, loff_t *ppos) | 3993 | size_t cnt, loff_t *ppos) |
| 3979 | { | 3994 | { |
| 3980 | unsigned long *p = filp->private_data; | 3995 | unsigned long *p = filp->private_data; |
| 3981 | char buf[64]; | ||
| 3982 | unsigned long val; | 3996 | unsigned long val; |
| 3983 | int ret; | 3997 | int ret; |
| 3984 | 3998 | ||
| 3985 | if (cnt >= sizeof(buf)) | 3999 | ret = kstrtoul_from_user(ubuf, cnt, 10, &val); |
| 3986 | return -EINVAL; | 4000 | if (ret) |
| 3987 | |||
| 3988 | if (copy_from_user(&buf, ubuf, cnt)) | ||
| 3989 | return -EFAULT; | ||
| 3990 | |||
| 3991 | buf[cnt] = 0; | ||
| 3992 | |||
| 3993 | ret = strict_strtoul(buf, 10, &val); | ||
| 3994 | if (ret < 0) | ||
| 3995 | return ret; | 4001 | return ret; |
| 3996 | 4002 | ||
| 3997 | if (val) | 4003 | if (val) |
diff --git a/kernel/trace/ring_buffer_benchmark.c b/kernel/trace/ring_buffer_benchmark.c index 302f8a61463..a5457d577b9 100644 --- a/kernel/trace/ring_buffer_benchmark.c +++ b/kernel/trace/ring_buffer_benchmark.c | |||
| @@ -106,7 +106,7 @@ static enum event_status read_page(int cpu) | |||
| 106 | int inc; | 106 | int inc; |
| 107 | int i; | 107 | int i; |
| 108 | 108 | ||
| 109 | bpage = ring_buffer_alloc_read_page(buffer); | 109 | bpage = ring_buffer_alloc_read_page(buffer, cpu); |
| 110 | if (!bpage) | 110 | if (!bpage) |
| 111 | return EVENT_DROPPED; | 111 | return EVENT_DROPPED; |
| 112 | 112 | ||
diff --git a/kernel/trace/trace.c b/kernel/trace/trace.c index ee9c921d7f2..e5df02c69b1 100644 --- a/kernel/trace/trace.c +++ b/kernel/trace/trace.c | |||
| @@ -343,26 +343,27 @@ unsigned long trace_flags = TRACE_ITER_PRINT_PARENT | TRACE_ITER_PRINTK | | |||
| 343 | static int trace_stop_count; | 343 | static int trace_stop_count; |
| 344 | static DEFINE_SPINLOCK(tracing_start_lock); | 344 | static DEFINE_SPINLOCK(tracing_start_lock); |
| 345 | 345 | ||
| 346 | static void wakeup_work_handler(struct work_struct *work) | ||
| 347 | { | ||
| 348 | wake_up(&trace_wait); | ||
| 349 | } | ||
| 350 | |||
| 351 | static DECLARE_DELAYED_WORK(wakeup_work, wakeup_work_handler); | ||
| 352 | |||
| 346 | /** | 353 | /** |
| 347 | * trace_wake_up - wake up tasks waiting for trace input | 354 | * trace_wake_up - wake up tasks waiting for trace input |
| 348 | * | 355 | * |
| 349 | * Simply wakes up any task that is blocked on the trace_wait | 356 | * Schedules a delayed work to wake up any task that is blocked on the |
| 350 | * queue. These is used with trace_poll for tasks polling the trace. | 357 | * trace_wait queue. These is used with trace_poll for tasks polling the |
| 358 | * trace. | ||
| 351 | */ | 359 | */ |
| 352 | void trace_wake_up(void) | 360 | void trace_wake_up(void) |
| 353 | { | 361 | { |
| 354 | int cpu; | 362 | const unsigned long delay = msecs_to_jiffies(2); |
| 355 | 363 | ||
| 356 | if (trace_flags & TRACE_ITER_BLOCK) | 364 | if (trace_flags & TRACE_ITER_BLOCK) |
| 357 | return; | 365 | return; |
| 358 | /* | 366 | schedule_delayed_work(&wakeup_work, delay); |
| 359 | * The runqueue_is_locked() can fail, but this is the best we | ||
| 360 | * have for now: | ||
| 361 | */ | ||
| 362 | cpu = get_cpu(); | ||
| 363 | if (!runqueue_is_locked(cpu)) | ||
| 364 | wake_up(&trace_wait); | ||
| 365 | put_cpu(); | ||
| 366 | } | 367 | } |
| 367 | 368 | ||
| 368 | static int __init set_buf_size(char *str) | 369 | static int __init set_buf_size(char *str) |
| @@ -424,6 +425,7 @@ static const char *trace_options[] = { | |||
| 424 | "graph-time", | 425 | "graph-time", |
| 425 | "record-cmd", | 426 | "record-cmd", |
| 426 | "overwrite", | 427 | "overwrite", |
| 428 | "disable_on_free", | ||
| 427 | NULL | 429 | NULL |
| 428 | }; | 430 | }; |
| 429 | 431 | ||
| @@ -1191,6 +1193,18 @@ void trace_nowake_buffer_unlock_commit(struct ring_buffer *buffer, | |||
| 1191 | } | 1193 | } |
| 1192 | EXPORT_SYMBOL_GPL(trace_nowake_buffer_unlock_commit); | 1194 | EXPORT_SYMBOL_GPL(trace_nowake_buffer_unlock_commit); |
| 1193 | 1195 | ||
| 1196 | void trace_nowake_buffer_unlock_commit_regs(struct ring_buffer *buffer, | ||
| 1197 | struct ring_buffer_event *event, | ||
| 1198 | unsigned long flags, int pc, | ||
| 1199 | struct pt_regs *regs) | ||
| 1200 | { | ||
| 1201 | ring_buffer_unlock_commit(buffer, event); | ||
| 1202 | |||
| 1203 | ftrace_trace_stack_regs(buffer, flags, 0, pc, regs); | ||
| 1204 | ftrace_trace_userstack(buffer, flags, pc); | ||
| 1205 | } | ||
| 1206 | EXPORT_SYMBOL_GPL(trace_nowake_buffer_unlock_commit_regs); | ||
| 1207 | |||
| 1194 | void trace_current_buffer_discard_commit(struct ring_buffer *buffer, | 1208 | void trace_current_buffer_discard_commit(struct ring_buffer *buffer, |
| 1195 | struct ring_buffer_event *event) | 1209 | struct ring_buffer_event *event) |
| 1196 | { | 1210 | { |
| @@ -1234,30 +1248,103 @@ ftrace(struct trace_array *tr, struct trace_array_cpu *data, | |||
| 1234 | } | 1248 | } |
| 1235 | 1249 | ||
| 1236 | #ifdef CONFIG_STACKTRACE | 1250 | #ifdef CONFIG_STACKTRACE |
| 1251 | |||
| 1252 | #define FTRACE_STACK_MAX_ENTRIES (PAGE_SIZE / sizeof(unsigned long)) | ||
| 1253 | struct ftrace_stack { | ||
| 1254 | unsigned long calls[FTRACE_STACK_MAX_ENTRIES]; | ||
| 1255 | }; | ||
| 1256 | |||
| 1257 | static DEFINE_PER_CPU(struct ftrace_stack, ftrace_stack); | ||
| 1258 | static DEFINE_PER_CPU(int, ftrace_stack_reserve); | ||
| 1259 | |||
| 1237 | static void __ftrace_trace_stack(struct ring_buffer *buffer, | 1260 | static void __ftrace_trace_stack(struct ring_buffer *buffer, |
| 1238 | unsigned long flags, | 1261 | unsigned long flags, |
| 1239 | int skip, int pc) | 1262 | int skip, int pc, struct pt_regs *regs) |
| 1240 | { | 1263 | { |
| 1241 | struct ftrace_event_call *call = &event_kernel_stack; | 1264 | struct ftrace_event_call *call = &event_kernel_stack; |
| 1242 | struct ring_buffer_event *event; | 1265 | struct ring_buffer_event *event; |
| 1243 | struct stack_entry *entry; | 1266 | struct stack_entry *entry; |
| 1244 | struct stack_trace trace; | 1267 | struct stack_trace trace; |
| 1268 | int use_stack; | ||
| 1269 | int size = FTRACE_STACK_ENTRIES; | ||
| 1270 | |||
| 1271 | trace.nr_entries = 0; | ||
| 1272 | trace.skip = skip; | ||
| 1273 | |||
| 1274 | /* | ||
| 1275 | * Since events can happen in NMIs there's no safe way to | ||
| 1276 | * use the per cpu ftrace_stacks. We reserve it and if an interrupt | ||
| 1277 | * or NMI comes in, it will just have to use the default | ||
| 1278 | * FTRACE_STACK_SIZE. | ||
| 1279 | */ | ||
| 1280 | preempt_disable_notrace(); | ||
| 1281 | |||
| 1282 | use_stack = ++__get_cpu_var(ftrace_stack_reserve); | ||
| 1283 | /* | ||
| 1284 | * We don't need any atomic variables, just a barrier. | ||
| 1285 | * If an interrupt comes in, we don't care, because it would | ||
| 1286 | * have exited and put the counter back to what we want. | ||
| 1287 | * We just need a barrier to keep gcc from moving things | ||
| 1288 | * around. | ||
| 1289 | */ | ||
| 1290 | barrier(); | ||
| 1291 | if (use_stack == 1) { | ||
| 1292 | trace.entries = &__get_cpu_var(ftrace_stack).calls[0]; | ||
| 1293 | trace.max_entries = FTRACE_STACK_MAX_ENTRIES; | ||
| 1294 | |||
| 1295 | if (regs) | ||
| 1296 | save_stack_trace_regs(regs, &trace); | ||
| 1297 | else | ||
| 1298 | save_stack_trace(&trace); | ||
| 1299 | |||
| 1300 | if (trace.nr_entries > size) | ||
| 1301 | size = trace.nr_entries; | ||
| 1302 | } else | ||
| 1303 | /* From now on, use_stack is a boolean */ | ||
| 1304 | use_stack = 0; | ||
| 1305 | |||
| 1306 | size *= sizeof(unsigned long); | ||
| 1245 | 1307 | ||
| 1246 | event = trace_buffer_lock_reserve(buffer, TRACE_STACK, | 1308 | event = trace_buffer_lock_reserve(buffer, TRACE_STACK, |
| 1247 | sizeof(*entry), flags, pc); | 1309 | sizeof(*entry) + size, flags, pc); |
| 1248 | if (!event) | 1310 | if (!event) |
| 1249 | return; | 1311 | goto out; |
| 1250 | entry = ring_buffer_event_data(event); | 1312 | entry = ring_buffer_event_data(event); |
| 1251 | memset(&entry->caller, 0, sizeof(entry->caller)); | ||
| 1252 | 1313 | ||
| 1253 | trace.nr_entries = 0; | 1314 | memset(&entry->caller, 0, size); |
| 1254 | trace.max_entries = FTRACE_STACK_ENTRIES; | 1315 | |
| 1255 | trace.skip = skip; | 1316 | if (use_stack) |
| 1256 | trace.entries = entry->caller; | 1317 | memcpy(&entry->caller, trace.entries, |
| 1318 | trace.nr_entries * sizeof(unsigned long)); | ||
| 1319 | else { | ||
| 1320 | trace.max_entries = FTRACE_STACK_ENTRIES; | ||
| 1321 | trace.entries = entry->caller; | ||
| 1322 | if (regs) | ||
| 1323 | save_stack_trace_regs(regs, &trace); | ||
| 1324 | else | ||
| 1325 | save_stack_trace(&trace); | ||
| 1326 | } | ||
| 1327 | |||
| 1328 | entry->size = trace.nr_entries; | ||
| 1257 | 1329 | ||
| 1258 | save_stack_trace(&trace); | ||
| 1259 | if (!filter_check_discard(call, entry, buffer, event)) | 1330 | if (!filter_check_discard(call, entry, buffer, event)) |
| 1260 | ring_buffer_unlock_commit(buffer, event); | 1331 | ring_buffer_unlock_commit(buffer, event); |
| 1332 | |||
| 1333 | out: | ||
| 1334 | /* Again, don't let gcc optimize things here */ | ||
| 1335 | barrier(); | ||
| 1336 | __get_cpu_var(ftrace_stack_reserve)--; | ||
| 1337 | preempt_enable_notrace(); | ||
| 1338 | |||
| 1339 | } | ||
| 1340 | |||
| 1341 | void ftrace_trace_stack_regs(struct ring_buffer *buffer, unsigned long flags, | ||
| 1342 | int skip, int pc, struct pt_regs *regs) | ||
| 1343 | { | ||
| 1344 | if (!(trace_flags & TRACE_ITER_STACKTRACE)) | ||
| 1345 | return; | ||
| 1346 | |||
| 1347 | __ftrace_trace_stack(buffer, flags, skip, pc, regs); | ||
| 1261 | } | 1348 | } |
| 1262 | 1349 | ||
| 1263 | void ftrace_trace_stack(struct ring_buffer *buffer, unsigned long flags, | 1350 | void ftrace_trace_stack(struct ring_buffer *buffer, unsigned long flags, |
| @@ -1266,13 +1353,13 @@ void ftrace_trace_stack(struct ring_buffer *buffer, unsigned long flags, | |||
| 1266 | if (!(trace_flags & TRACE_ITER_STACKTRACE)) | 1353 | if (!(trace_flags & TRACE_ITER_STACKTRACE)) |
| 1267 | return; | 1354 | return; |
| 1268 | 1355 | ||
| 1269 | __ftrace_trace_stack(buffer, flags, skip, pc); | 1356 | __ftrace_trace_stack(buffer, flags, skip, pc, NULL); |
| 1270 | } | 1357 | } |
| 1271 | 1358 | ||
| 1272 | void __trace_stack(struct trace_array *tr, unsigned long flags, int skip, | 1359 | void __trace_stack(struct trace_array *tr, unsigned long flags, int skip, |
| 1273 | int pc) | 1360 | int pc) |
| 1274 | { | 1361 | { |
| 1275 | __ftrace_trace_stack(tr->buffer, flags, skip, pc); | 1362 | __ftrace_trace_stack(tr->buffer, flags, skip, pc, NULL); |
| 1276 | } | 1363 | } |
| 1277 | 1364 | ||
| 1278 | /** | 1365 | /** |
| @@ -1288,7 +1375,7 @@ void trace_dump_stack(void) | |||
| 1288 | local_save_flags(flags); | 1375 | local_save_flags(flags); |
| 1289 | 1376 | ||
| 1290 | /* skipping 3 traces, seems to get us at the caller of this function */ | 1377 | /* skipping 3 traces, seems to get us at the caller of this function */ |
| 1291 | __ftrace_trace_stack(global_trace.buffer, flags, 3, preempt_count()); | 1378 | __ftrace_trace_stack(global_trace.buffer, flags, 3, preempt_count(), NULL); |
| 1292 | } | 1379 | } |
| 1293 | 1380 | ||
| 1294 | static DEFINE_PER_CPU(int, user_stack_count); | 1381 | static DEFINE_PER_CPU(int, user_stack_count); |
| @@ -1536,7 +1623,12 @@ peek_next_entry(struct trace_iterator *iter, int cpu, u64 *ts, | |||
| 1536 | 1623 | ||
| 1537 | ftrace_enable_cpu(); | 1624 | ftrace_enable_cpu(); |
| 1538 | 1625 | ||
| 1539 | return event ? ring_buffer_event_data(event) : NULL; | 1626 | if (event) { |
| 1627 | iter->ent_size = ring_buffer_event_length(event); | ||
| 1628 | return ring_buffer_event_data(event); | ||
| 1629 | } | ||
| 1630 | iter->ent_size = 0; | ||
| 1631 | return NULL; | ||
| 1540 | } | 1632 | } |
| 1541 | 1633 | ||
| 1542 | static struct trace_entry * | 1634 | static struct trace_entry * |
| @@ -2051,6 +2143,9 @@ void trace_default_header(struct seq_file *m) | |||
| 2051 | { | 2143 | { |
| 2052 | struct trace_iterator *iter = m->private; | 2144 | struct trace_iterator *iter = m->private; |
| 2053 | 2145 | ||
| 2146 | if (!(trace_flags & TRACE_ITER_CONTEXT_INFO)) | ||
| 2147 | return; | ||
| 2148 | |||
| 2054 | if (iter->iter_flags & TRACE_FILE_LAT_FMT) { | 2149 | if (iter->iter_flags & TRACE_FILE_LAT_FMT) { |
| 2055 | /* print nothing if the buffers are empty */ | 2150 | /* print nothing if the buffers are empty */ |
| 2056 | if (trace_empty(iter)) | 2151 | if (trace_empty(iter)) |
| @@ -2701,20 +2796,11 @@ tracing_ctrl_write(struct file *filp, const char __user *ubuf, | |||
| 2701 | size_t cnt, loff_t *ppos) | 2796 | size_t cnt, loff_t *ppos) |
| 2702 | { | 2797 | { |
| 2703 | struct trace_array *tr = filp->private_data; | 2798 | struct trace_array *tr = filp->private_data; |
| 2704 | char buf[64]; | ||
| 2705 | unsigned long val; | 2799 | unsigned long val; |
| 2706 | int ret; | 2800 | int ret; |
| 2707 | 2801 | ||
| 2708 | if (cnt >= sizeof(buf)) | 2802 | ret = kstrtoul_from_user(ubuf, cnt, 10, &val); |
| 2709 | return -EINVAL; | 2803 | if (ret) |
| 2710 | |||
| 2711 | if (copy_from_user(&buf, ubuf, cnt)) | ||
| 2712 | return -EFAULT; | ||
| 2713 | |||
| 2714 | buf[cnt] = 0; | ||
| 2715 | |||
| 2716 | ret = strict_strtoul(buf, 10, &val); | ||
| 2717 | if (ret < 0) | ||
| 2718 | return ret; | 2804 | return ret; |
| 2719 | 2805 | ||
| 2720 | val = !!val; | 2806 | val = !!val; |
| @@ -2767,7 +2853,7 @@ int tracer_init(struct tracer *t, struct trace_array *tr) | |||
| 2767 | return t->init(tr); | 2853 | return t->init(tr); |
| 2768 | } | 2854 | } |
| 2769 | 2855 | ||
| 2770 | static int tracing_resize_ring_buffer(unsigned long size) | 2856 | static int __tracing_resize_ring_buffer(unsigned long size) |
| 2771 | { | 2857 | { |
| 2772 | int ret; | 2858 | int ret; |
| 2773 | 2859 | ||
| @@ -2819,6 +2905,41 @@ static int tracing_resize_ring_buffer(unsigned long size) | |||
| 2819 | return ret; | 2905 | return ret; |
| 2820 | } | 2906 | } |
| 2821 | 2907 | ||
| 2908 | static ssize_t tracing_resize_ring_buffer(unsigned long size) | ||
| 2909 | { | ||
| 2910 | int cpu, ret = size; | ||
| 2911 | |||
| 2912 | mutex_lock(&trace_types_lock); | ||
| 2913 | |||
| 2914 | tracing_stop(); | ||
| 2915 | |||
| 2916 | /* disable all cpu buffers */ | ||
| 2917 | for_each_tracing_cpu(cpu) { | ||
| 2918 | if (global_trace.data[cpu]) | ||
| 2919 | atomic_inc(&global_trace.data[cpu]->disabled); | ||
| 2920 | if (max_tr.data[cpu]) | ||
| 2921 | atomic_inc(&max_tr.data[cpu]->disabled); | ||
| 2922 | } | ||
| 2923 | |||
| 2924 | if (size != global_trace.entries) | ||
| 2925 | ret = __tracing_resize_ring_buffer(size); | ||
| 2926 | |||
| 2927 | if (ret < 0) | ||
| 2928 | ret = -ENOMEM; | ||
| 2929 | |||
| 2930 | for_each_tracing_cpu(cpu) { | ||
| 2931 | if (global_trace.data[cpu]) | ||
| 2932 | atomic_dec(&global_trace.data[cpu]->disabled); | ||
| 2933 | if (max_tr.data[cpu]) | ||
| 2934 | atomic_dec(&max_tr.data[cpu]->disabled); | ||
| 2935 | } | ||
| 2936 | |||
| 2937 | tracing_start(); | ||
| 2938 | mutex_unlock(&trace_types_lock); | ||
| 2939 | |||
| 2940 | return ret; | ||
| 2941 | } | ||
| 2942 | |||
| 2822 | 2943 | ||
| 2823 | /** | 2944 | /** |
| 2824 | * tracing_update_buffers - used by tracing facility to expand ring buffers | 2945 | * tracing_update_buffers - used by tracing facility to expand ring buffers |
| @@ -2836,7 +2957,7 @@ int tracing_update_buffers(void) | |||
| 2836 | 2957 | ||
| 2837 | mutex_lock(&trace_types_lock); | 2958 | mutex_lock(&trace_types_lock); |
| 2838 | if (!ring_buffer_expanded) | 2959 | if (!ring_buffer_expanded) |
| 2839 | ret = tracing_resize_ring_buffer(trace_buf_size); | 2960 | ret = __tracing_resize_ring_buffer(trace_buf_size); |
| 2840 | mutex_unlock(&trace_types_lock); | 2961 | mutex_unlock(&trace_types_lock); |
| 2841 | 2962 | ||
| 2842 | return ret; | 2963 | return ret; |
| @@ -2860,7 +2981,7 @@ static int tracing_set_tracer(const char *buf) | |||
| 2860 | mutex_lock(&trace_types_lock); | 2981 | mutex_lock(&trace_types_lock); |
| 2861 | 2982 | ||
| 2862 | if (!ring_buffer_expanded) { | 2983 | if (!ring_buffer_expanded) { |
| 2863 | ret = tracing_resize_ring_buffer(trace_buf_size); | 2984 | ret = __tracing_resize_ring_buffer(trace_buf_size); |
| 2864 | if (ret < 0) | 2985 | if (ret < 0) |
| 2865 | goto out; | 2986 | goto out; |
| 2866 | ret = 0; | 2987 | ret = 0; |
| @@ -2966,20 +3087,11 @@ tracing_max_lat_write(struct file *filp, const char __user *ubuf, | |||
| 2966 | size_t cnt, loff_t *ppos) | 3087 | size_t cnt, loff_t *ppos) |
| 2967 | { | 3088 | { |
| 2968 | unsigned long *ptr = filp->private_data; | 3089 | unsigned long *ptr = filp->private_data; |
| 2969 | char buf[64]; | ||
| 2970 | unsigned long val; | 3090 | unsigned long val; |
| 2971 | int ret; | 3091 | int ret; |
| 2972 | 3092 | ||
| 2973 | if (cnt >= sizeof(buf)) | 3093 | ret = kstrtoul_from_user(ubuf, cnt, 10, &val); |
| 2974 | return -EINVAL; | 3094 | if (ret) |
| 2975 | |||
| 2976 | if (copy_from_user(&buf, ubuf, cnt)) | ||
| 2977 | return -EFAULT; | ||
| 2978 | |||
| 2979 | buf[cnt] = 0; | ||
| 2980 | |||
| 2981 | ret = strict_strtoul(buf, 10, &val); | ||
| 2982 | if (ret < 0) | ||
| 2983 | return ret; | 3095 | return ret; |
| 2984 | 3096 | ||
| 2985 | *ptr = val * 1000; | 3097 | *ptr = val * 1000; |
| @@ -3434,67 +3546,54 @@ tracing_entries_write(struct file *filp, const char __user *ubuf, | |||
| 3434 | size_t cnt, loff_t *ppos) | 3546 | size_t cnt, loff_t *ppos) |
| 3435 | { | 3547 | { |
| 3436 | unsigned long val; | 3548 | unsigned long val; |
| 3437 | char buf[64]; | 3549 | int ret; |
| 3438 | int ret, cpu; | ||
| 3439 | |||
| 3440 | if (cnt >= sizeof(buf)) | ||
| 3441 | return -EINVAL; | ||
| 3442 | |||
| 3443 | if (copy_from_user(&buf, ubuf, cnt)) | ||
| 3444 | return -EFAULT; | ||
| 3445 | |||
| 3446 | buf[cnt] = 0; | ||
| 3447 | 3550 | ||
| 3448 | ret = strict_strtoul(buf, 10, &val); | 3551 | ret = kstrtoul_from_user(ubuf, cnt, 10, &val); |
| 3449 | if (ret < 0) | 3552 | if (ret) |
| 3450 | return ret; | 3553 | return ret; |
| 3451 | 3554 | ||
| 3452 | /* must have at least 1 entry */ | 3555 | /* must have at least 1 entry */ |
| 3453 | if (!val) | 3556 | if (!val) |
| 3454 | return -EINVAL; | 3557 | return -EINVAL; |
| 3455 | 3558 | ||
| 3456 | mutex_lock(&trace_types_lock); | ||
| 3457 | |||
| 3458 | tracing_stop(); | ||
| 3459 | |||
| 3460 | /* disable all cpu buffers */ | ||
| 3461 | for_each_tracing_cpu(cpu) { | ||
| 3462 | if (global_trace.data[cpu]) | ||
| 3463 | atomic_inc(&global_trace.data[cpu]->disabled); | ||
| 3464 | if (max_tr.data[cpu]) | ||
| 3465 | atomic_inc(&max_tr.data[cpu]->disabled); | ||
| 3466 | } | ||
| 3467 | |||
| 3468 | /* value is in KB */ | 3559 | /* value is in KB */ |
| 3469 | val <<= 10; | 3560 | val <<= 10; |
| 3470 | 3561 | ||
| 3471 | if (val != global_trace.entries) { | 3562 | ret = tracing_resize_ring_buffer(val); |
| 3472 | ret = tracing_resize_ring_buffer(val); | 3563 | if (ret < 0) |
| 3473 | if (ret < 0) { | 3564 | return ret; |
| 3474 | cnt = ret; | ||
| 3475 | goto out; | ||
| 3476 | } | ||
| 3477 | } | ||
| 3478 | 3565 | ||
| 3479 | *ppos += cnt; | 3566 | *ppos += cnt; |
| 3480 | 3567 | ||
| 3481 | /* If check pages failed, return ENOMEM */ | 3568 | return cnt; |
| 3482 | if (tracing_disabled) | 3569 | } |
| 3483 | cnt = -ENOMEM; | ||
| 3484 | out: | ||
| 3485 | for_each_tracing_cpu(cpu) { | ||
| 3486 | if (global_trace.data[cpu]) | ||
| 3487 | atomic_dec(&global_trace.data[cpu]->disabled); | ||
| 3488 | if (max_tr.data[cpu]) | ||
| 3489 | atomic_dec(&max_tr.data[cpu]->disabled); | ||
| 3490 | } | ||
| 3491 | 3570 | ||
| 3492 | tracing_start(); | 3571 | static ssize_t |
| 3493 | mutex_unlock(&trace_types_lock); | 3572 | tracing_free_buffer_write(struct file *filp, const char __user *ubuf, |
| 3573 | size_t cnt, loff_t *ppos) | ||
| 3574 | { | ||
| 3575 | /* | ||
| 3576 | * There is no need to read what the user has written, this function | ||
| 3577 | * is just to make sure that there is no error when "echo" is used | ||
| 3578 | */ | ||
| 3579 | |||
| 3580 | *ppos += cnt; | ||
| 3494 | 3581 | ||
| 3495 | return cnt; | 3582 | return cnt; |
| 3496 | } | 3583 | } |
| 3497 | 3584 | ||
| 3585 | static int | ||
| 3586 | tracing_free_buffer_release(struct inode *inode, struct file *filp) | ||
| 3587 | { | ||
| 3588 | /* disable tracing ? */ | ||
| 3589 | if (trace_flags & TRACE_ITER_STOP_ON_FREE) | ||
| 3590 | tracing_off(); | ||
| 3591 | /* resize the ring buffer to 0 */ | ||
| 3592 | tracing_resize_ring_buffer(0); | ||
| 3593 | |||
| 3594 | return 0; | ||
| 3595 | } | ||
| 3596 | |||
| 3498 | static int mark_printk(const char *fmt, ...) | 3597 | static int mark_printk(const char *fmt, ...) |
| 3499 | { | 3598 | { |
| 3500 | int ret; | 3599 | int ret; |
| @@ -3640,6 +3739,11 @@ static const struct file_operations tracing_entries_fops = { | |||
| 3640 | .llseek = generic_file_llseek, | 3739 | .llseek = generic_file_llseek, |
| 3641 | }; | 3740 | }; |
| 3642 | 3741 | ||
| 3742 | static const struct file_operations tracing_free_buffer_fops = { | ||
| 3743 | .write = tracing_free_buffer_write, | ||
| 3744 | .release = tracing_free_buffer_release, | ||
| 3745 | }; | ||
| 3746 | |||
| 3643 | static const struct file_operations tracing_mark_fops = { | 3747 | static const struct file_operations tracing_mark_fops = { |
| 3644 | .open = tracing_open_generic, | 3748 | .open = tracing_open_generic, |
| 3645 | .write = tracing_mark_write, | 3749 | .write = tracing_mark_write, |
| @@ -3696,7 +3800,7 @@ tracing_buffers_read(struct file *filp, char __user *ubuf, | |||
| 3696 | return 0; | 3800 | return 0; |
| 3697 | 3801 | ||
| 3698 | if (!info->spare) | 3802 | if (!info->spare) |
| 3699 | info->spare = ring_buffer_alloc_read_page(info->tr->buffer); | 3803 | info->spare = ring_buffer_alloc_read_page(info->tr->buffer, info->cpu); |
| 3700 | if (!info->spare) | 3804 | if (!info->spare) |
| 3701 | return -ENOMEM; | 3805 | return -ENOMEM; |
| 3702 | 3806 | ||
| @@ -3853,7 +3957,7 @@ tracing_buffers_splice_read(struct file *file, loff_t *ppos, | |||
| 3853 | 3957 | ||
| 3854 | ref->ref = 1; | 3958 | ref->ref = 1; |
| 3855 | ref->buffer = info->tr->buffer; | 3959 | ref->buffer = info->tr->buffer; |
| 3856 | ref->page = ring_buffer_alloc_read_page(ref->buffer); | 3960 | ref->page = ring_buffer_alloc_read_page(ref->buffer, info->cpu); |
| 3857 | if (!ref->page) { | 3961 | if (!ref->page) { |
| 3858 | kfree(ref); | 3962 | kfree(ref); |
| 3859 | break; | 3963 | break; |
| @@ -3862,8 +3966,7 @@ tracing_buffers_splice_read(struct file *file, loff_t *ppos, | |||
| 3862 | r = ring_buffer_read_page(ref->buffer, &ref->page, | 3966 | r = ring_buffer_read_page(ref->buffer, &ref->page, |
| 3863 | len, info->cpu, 1); | 3967 | len, info->cpu, 1); |
| 3864 | if (r < 0) { | 3968 | if (r < 0) { |
| 3865 | ring_buffer_free_read_page(ref->buffer, | 3969 | ring_buffer_free_read_page(ref->buffer, ref->page); |
| 3866 | ref->page); | ||
| 3867 | kfree(ref); | 3970 | kfree(ref); |
| 3868 | break; | 3971 | break; |
| 3869 | } | 3972 | } |
| @@ -4099,19 +4202,10 @@ trace_options_write(struct file *filp, const char __user *ubuf, size_t cnt, | |||
| 4099 | { | 4202 | { |
| 4100 | struct trace_option_dentry *topt = filp->private_data; | 4203 | struct trace_option_dentry *topt = filp->private_data; |
| 4101 | unsigned long val; | 4204 | unsigned long val; |
| 4102 | char buf[64]; | ||
| 4103 | int ret; | 4205 | int ret; |
| 4104 | 4206 | ||
| 4105 | if (cnt >= sizeof(buf)) | 4207 | ret = kstrtoul_from_user(ubuf, cnt, 10, &val); |
| 4106 | return -EINVAL; | 4208 | if (ret) |
| 4107 | |||
| 4108 | if (copy_from_user(&buf, ubuf, cnt)) | ||
| 4109 | return -EFAULT; | ||
| 4110 | |||
| 4111 | buf[cnt] = 0; | ||
| 4112 | |||
| 4113 | ret = strict_strtoul(buf, 10, &val); | ||
| 4114 | if (ret < 0) | ||
| 4115 | return ret; | 4209 | return ret; |
| 4116 | 4210 | ||
| 4117 | if (val != 0 && val != 1) | 4211 | if (val != 0 && val != 1) |
| @@ -4159,20 +4253,11 @@ trace_options_core_write(struct file *filp, const char __user *ubuf, size_t cnt, | |||
| 4159 | loff_t *ppos) | 4253 | loff_t *ppos) |
| 4160 | { | 4254 | { |
| 4161 | long index = (long)filp->private_data; | 4255 | long index = (long)filp->private_data; |
| 4162 | char buf[64]; | ||
| 4163 | unsigned long val; | 4256 | unsigned long val; |
| 4164 | int ret; | 4257 | int ret; |
| 4165 | 4258 | ||
| 4166 | if (cnt >= sizeof(buf)) | 4259 | ret = kstrtoul_from_user(ubuf, cnt, 10, &val); |
| 4167 | return -EINVAL; | 4260 | if (ret) |
| 4168 | |||
| 4169 | if (copy_from_user(&buf, ubuf, cnt)) | ||
| 4170 | return -EFAULT; | ||
| 4171 | |||
| 4172 | buf[cnt] = 0; | ||
| 4173 | |||
| 4174 | ret = strict_strtoul(buf, 10, &val); | ||
| 4175 | if (ret < 0) | ||
| 4176 | return ret; | 4261 | return ret; |
| 4177 | 4262 | ||
| 4178 | if (val != 0 && val != 1) | 4263 | if (val != 0 && val != 1) |
| @@ -4365,6 +4450,9 @@ static __init int tracer_init_debugfs(void) | |||
| 4365 | trace_create_file("buffer_size_kb", 0644, d_tracer, | 4450 | trace_create_file("buffer_size_kb", 0644, d_tracer, |
| 4366 | &global_trace, &tracing_entries_fops); | 4451 | &global_trace, &tracing_entries_fops); |
| 4367 | 4452 | ||
| 4453 | trace_create_file("free_buffer", 0644, d_tracer, | ||
| 4454 | &global_trace, &tracing_free_buffer_fops); | ||
| 4455 | |||
| 4368 | trace_create_file("trace_marker", 0220, d_tracer, | 4456 | trace_create_file("trace_marker", 0220, d_tracer, |
| 4369 | NULL, &tracing_mark_fops); | 4457 | NULL, &tracing_mark_fops); |
| 4370 | 4458 | ||
diff --git a/kernel/trace/trace.h b/kernel/trace/trace.h index 229f8591f61..3f381d0b20a 100644 --- a/kernel/trace/trace.h +++ b/kernel/trace/trace.h | |||
| @@ -278,6 +278,29 @@ struct tracer { | |||
| 278 | }; | 278 | }; |
| 279 | 279 | ||
| 280 | 280 | ||
| 281 | /* Only current can touch trace_recursion */ | ||
| 282 | #define trace_recursion_inc() do { (current)->trace_recursion++; } while (0) | ||
| 283 | #define trace_recursion_dec() do { (current)->trace_recursion--; } while (0) | ||
| 284 | |||
| 285 | /* Ring buffer has the 10 LSB bits to count */ | ||
| 286 | #define trace_recursion_buffer() ((current)->trace_recursion & 0x3ff) | ||
| 287 | |||
| 288 | /* for function tracing recursion */ | ||
| 289 | #define TRACE_INTERNAL_BIT (1<<11) | ||
| 290 | #define TRACE_GLOBAL_BIT (1<<12) | ||
| 291 | /* | ||
| 292 | * Abuse of the trace_recursion. | ||
| 293 | * As we need a way to maintain state if we are tracing the function | ||
| 294 | * graph in irq because we want to trace a particular function that | ||
| 295 | * was called in irq context but we have irq tracing off. Since this | ||
| 296 | * can only be modified by current, we can reuse trace_recursion. | ||
| 297 | */ | ||
| 298 | #define TRACE_IRQ_BIT (1<<13) | ||
| 299 | |||
| 300 | #define trace_recursion_set(bit) do { (current)->trace_recursion |= (bit); } while (0) | ||
| 301 | #define trace_recursion_clear(bit) do { (current)->trace_recursion &= ~(bit); } while (0) | ||
| 302 | #define trace_recursion_test(bit) ((current)->trace_recursion & (bit)) | ||
| 303 | |||
| 281 | #define TRACE_PIPE_ALL_CPU -1 | 304 | #define TRACE_PIPE_ALL_CPU -1 |
| 282 | 305 | ||
| 283 | int tracer_init(struct tracer *t, struct trace_array *tr); | 306 | int tracer_init(struct tracer *t, struct trace_array *tr); |
| @@ -389,6 +412,9 @@ void update_max_tr_single(struct trace_array *tr, | |||
| 389 | void ftrace_trace_stack(struct ring_buffer *buffer, unsigned long flags, | 412 | void ftrace_trace_stack(struct ring_buffer *buffer, unsigned long flags, |
| 390 | int skip, int pc); | 413 | int skip, int pc); |
| 391 | 414 | ||
| 415 | void ftrace_trace_stack_regs(struct ring_buffer *buffer, unsigned long flags, | ||
| 416 | int skip, int pc, struct pt_regs *regs); | ||
| 417 | |||
| 392 | void ftrace_trace_userstack(struct ring_buffer *buffer, unsigned long flags, | 418 | void ftrace_trace_userstack(struct ring_buffer *buffer, unsigned long flags, |
| 393 | int pc); | 419 | int pc); |
| 394 | 420 | ||
| @@ -400,6 +426,12 @@ static inline void ftrace_trace_stack(struct ring_buffer *buffer, | |||
| 400 | { | 426 | { |
| 401 | } | 427 | } |
| 402 | 428 | ||
| 429 | static inline void ftrace_trace_stack_regs(struct ring_buffer *buffer, | ||
| 430 | unsigned long flags, int skip, | ||
| 431 | int pc, struct pt_regs *regs) | ||
| 432 | { | ||
| 433 | } | ||
| 434 | |||
| 403 | static inline void ftrace_trace_userstack(struct ring_buffer *buffer, | 435 | static inline void ftrace_trace_userstack(struct ring_buffer *buffer, |
| 404 | unsigned long flags, int pc) | 436 | unsigned long flags, int pc) |
| 405 | { | 437 | { |
| @@ -507,8 +539,18 @@ static inline int ftrace_graph_addr(unsigned long addr) | |||
| 507 | return 1; | 539 | return 1; |
| 508 | 540 | ||
| 509 | for (i = 0; i < ftrace_graph_count; i++) { | 541 | for (i = 0; i < ftrace_graph_count; i++) { |
| 510 | if (addr == ftrace_graph_funcs[i]) | 542 | if (addr == ftrace_graph_funcs[i]) { |
| 543 | /* | ||
| 544 | * If no irqs are to be traced, but a set_graph_function | ||
| 545 | * is set, and called by an interrupt handler, we still | ||
| 546 | * want to trace it. | ||
| 547 | */ | ||
| 548 | if (in_irq()) | ||
| 549 | trace_recursion_set(TRACE_IRQ_BIT); | ||
| 550 | else | ||
| 551 | trace_recursion_clear(TRACE_IRQ_BIT); | ||
| 511 | return 1; | 552 | return 1; |
| 553 | } | ||
| 512 | } | 554 | } |
| 513 | 555 | ||
| 514 | return 0; | 556 | return 0; |
| @@ -609,6 +651,7 @@ enum trace_iterator_flags { | |||
| 609 | TRACE_ITER_GRAPH_TIME = 0x80000, | 651 | TRACE_ITER_GRAPH_TIME = 0x80000, |
| 610 | TRACE_ITER_RECORD_CMD = 0x100000, | 652 | TRACE_ITER_RECORD_CMD = 0x100000, |
| 611 | TRACE_ITER_OVERWRITE = 0x200000, | 653 | TRACE_ITER_OVERWRITE = 0x200000, |
| 654 | TRACE_ITER_STOP_ON_FREE = 0x400000, | ||
| 612 | }; | 655 | }; |
| 613 | 656 | ||
| 614 | /* | 657 | /* |
| @@ -677,6 +720,7 @@ struct event_subsystem { | |||
| 677 | struct dentry *entry; | 720 | struct dentry *entry; |
| 678 | struct event_filter *filter; | 721 | struct event_filter *filter; |
| 679 | int nr_events; | 722 | int nr_events; |
| 723 | int ref_count; | ||
| 680 | }; | 724 | }; |
| 681 | 725 | ||
| 682 | #define FILTER_PRED_INVALID ((unsigned short)-1) | 726 | #define FILTER_PRED_INVALID ((unsigned short)-1) |
| @@ -784,19 +828,4 @@ extern const char *__stop___trace_bprintk_fmt[]; | |||
| 784 | FTRACE_ENTRY(call, struct_name, id, PARAMS(tstruct), PARAMS(print)) | 828 | FTRACE_ENTRY(call, struct_name, id, PARAMS(tstruct), PARAMS(print)) |
| 785 | #include "trace_entries.h" | 829 | #include "trace_entries.h" |
| 786 | 830 | ||
| 787 | /* Only current can touch trace_recursion */ | ||
| 788 | #define trace_recursion_inc() do { (current)->trace_recursion++; } while (0) | ||
| 789 | #define trace_recursion_dec() do { (current)->trace_recursion--; } while (0) | ||
| 790 | |||
| 791 | /* Ring buffer has the 10 LSB bits to count */ | ||
| 792 | #define trace_recursion_buffer() ((current)->trace_recursion & 0x3ff) | ||
| 793 | |||
| 794 | /* for function tracing recursion */ | ||
| 795 | #define TRACE_INTERNAL_BIT (1<<11) | ||
| 796 | #define TRACE_GLOBAL_BIT (1<<12) | ||
| 797 | |||
| 798 | #define trace_recursion_set(bit) do { (current)->trace_recursion |= (bit); } while (0) | ||
| 799 | #define trace_recursion_clear(bit) do { (current)->trace_recursion &= ~(bit); } while (0) | ||
| 800 | #define trace_recursion_test(bit) ((current)->trace_recursion & (bit)) | ||
| 801 | |||
| 802 | #endif /* _LINUX_KERNEL_TRACE_H */ | 831 | #endif /* _LINUX_KERNEL_TRACE_H */ |
diff --git a/kernel/trace/trace_entries.h b/kernel/trace/trace_entries.h index e32744c84d9..93365907f21 100644 --- a/kernel/trace/trace_entries.h +++ b/kernel/trace/trace_entries.h | |||
| @@ -161,7 +161,8 @@ FTRACE_ENTRY(kernel_stack, stack_entry, | |||
| 161 | TRACE_STACK, | 161 | TRACE_STACK, |
| 162 | 162 | ||
| 163 | F_STRUCT( | 163 | F_STRUCT( |
| 164 | __array( unsigned long, caller, FTRACE_STACK_ENTRIES ) | 164 | __field( int, size ) |
| 165 | __dynamic_array(unsigned long, caller ) | ||
| 165 | ), | 166 | ), |
| 166 | 167 | ||
| 167 | F_printk("\t=> (%08lx)\n\t=> (%08lx)\n\t=> (%08lx)\n\t=> (%08lx)\n" | 168 | F_printk("\t=> (%08lx)\n\t=> (%08lx)\n\t=> (%08lx)\n\t=> (%08lx)\n" |
diff --git a/kernel/trace/trace_events.c b/kernel/trace/trace_events.c index 686ec399f2a..581876f9f38 100644 --- a/kernel/trace/trace_events.c +++ b/kernel/trace/trace_events.c | |||
| @@ -244,6 +244,35 @@ static void ftrace_clear_events(void) | |||
| 244 | mutex_unlock(&event_mutex); | 244 | mutex_unlock(&event_mutex); |
| 245 | } | 245 | } |
| 246 | 246 | ||
| 247 | static void __put_system(struct event_subsystem *system) | ||
| 248 | { | ||
| 249 | struct event_filter *filter = system->filter; | ||
| 250 | |||
| 251 | WARN_ON_ONCE(system->ref_count == 0); | ||
| 252 | if (--system->ref_count) | ||
| 253 | return; | ||
| 254 | |||
| 255 | if (filter) { | ||
| 256 | kfree(filter->filter_string); | ||
| 257 | kfree(filter); | ||
| 258 | } | ||
| 259 | kfree(system->name); | ||
| 260 | kfree(system); | ||
| 261 | } | ||
| 262 | |||
| 263 | static void __get_system(struct event_subsystem *system) | ||
| 264 | { | ||
| 265 | WARN_ON_ONCE(system->ref_count == 0); | ||
| 266 | system->ref_count++; | ||
| 267 | } | ||
| 268 | |||
| 269 | static void put_system(struct event_subsystem *system) | ||
| 270 | { | ||
| 271 | mutex_lock(&event_mutex); | ||
| 272 | __put_system(system); | ||
| 273 | mutex_unlock(&event_mutex); | ||
| 274 | } | ||
| 275 | |||
| 247 | /* | 276 | /* |
| 248 | * __ftrace_set_clr_event(NULL, NULL, NULL, set) will set/unset all events. | 277 | * __ftrace_set_clr_event(NULL, NULL, NULL, set) will set/unset all events. |
| 249 | */ | 278 | */ |
| @@ -486,20 +515,11 @@ event_enable_write(struct file *filp, const char __user *ubuf, size_t cnt, | |||
| 486 | loff_t *ppos) | 515 | loff_t *ppos) |
| 487 | { | 516 | { |
| 488 | struct ftrace_event_call *call = filp->private_data; | 517 | struct ftrace_event_call *call = filp->private_data; |
| 489 | char buf[64]; | ||
| 490 | unsigned long val; | 518 | unsigned long val; |
| 491 | int ret; | 519 | int ret; |
| 492 | 520 | ||
| 493 | if (cnt >= sizeof(buf)) | 521 | ret = kstrtoul_from_user(ubuf, cnt, 10, &val); |
| 494 | return -EINVAL; | 522 | if (ret) |
| 495 | |||
| 496 | if (copy_from_user(&buf, ubuf, cnt)) | ||
| 497 | return -EFAULT; | ||
| 498 | |||
| 499 | buf[cnt] = 0; | ||
| 500 | |||
| 501 | ret = strict_strtoul(buf, 10, &val); | ||
| 502 | if (ret < 0) | ||
| 503 | return ret; | 523 | return ret; |
| 504 | 524 | ||
| 505 | ret = tracing_update_buffers(); | 525 | ret = tracing_update_buffers(); |
| @@ -528,7 +548,7 @@ system_enable_read(struct file *filp, char __user *ubuf, size_t cnt, | |||
| 528 | loff_t *ppos) | 548 | loff_t *ppos) |
| 529 | { | 549 | { |
| 530 | const char set_to_char[4] = { '?', '0', '1', 'X' }; | 550 | const char set_to_char[4] = { '?', '0', '1', 'X' }; |
| 531 | const char *system = filp->private_data; | 551 | struct event_subsystem *system = filp->private_data; |
| 532 | struct ftrace_event_call *call; | 552 | struct ftrace_event_call *call; |
| 533 | char buf[2]; | 553 | char buf[2]; |
| 534 | int set = 0; | 554 | int set = 0; |
| @@ -539,7 +559,7 @@ system_enable_read(struct file *filp, char __user *ubuf, size_t cnt, | |||
| 539 | if (!call->name || !call->class || !call->class->reg) | 559 | if (!call->name || !call->class || !call->class->reg) |
| 540 | continue; | 560 | continue; |
| 541 | 561 | ||
| 542 | if (system && strcmp(call->class->system, system) != 0) | 562 | if (system && strcmp(call->class->system, system->name) != 0) |
| 543 | continue; | 563 | continue; |
| 544 | 564 | ||
| 545 | /* | 565 | /* |
| @@ -569,21 +589,13 @@ static ssize_t | |||
| 569 | system_enable_write(struct file *filp, const char __user *ubuf, size_t cnt, | 589 | system_enable_write(struct file *filp, const char __user *ubuf, size_t cnt, |
| 570 | loff_t *ppos) | 590 | loff_t *ppos) |
| 571 | { | 591 | { |
| 572 | const char *system = filp->private_data; | 592 | struct event_subsystem *system = filp->private_data; |
| 593 | const char *name = NULL; | ||
| 573 | unsigned long val; | 594 | unsigned long val; |
| 574 | char buf[64]; | ||
| 575 | ssize_t ret; | 595 | ssize_t ret; |
| 576 | 596 | ||
| 577 | if (cnt >= sizeof(buf)) | 597 | ret = kstrtoul_from_user(ubuf, cnt, 10, &val); |
| 578 | return -EINVAL; | 598 | if (ret) |
| 579 | |||
| 580 | if (copy_from_user(&buf, ubuf, cnt)) | ||
| 581 | return -EFAULT; | ||
| 582 | |||
| 583 | buf[cnt] = 0; | ||
| 584 | |||
| 585 | ret = strict_strtoul(buf, 10, &val); | ||
| 586 | if (ret < 0) | ||
| 587 | return ret; | 599 | return ret; |
| 588 | 600 | ||
| 589 | ret = tracing_update_buffers(); | 601 | ret = tracing_update_buffers(); |
| @@ -593,7 +605,14 @@ system_enable_write(struct file *filp, const char __user *ubuf, size_t cnt, | |||
| 593 | if (val != 0 && val != 1) | 605 | if (val != 0 && val != 1) |
| 594 | return -EINVAL; | 606 | return -EINVAL; |
| 595 | 607 | ||
| 596 | ret = __ftrace_set_clr_event(NULL, system, NULL, val); | 608 | /* |
| 609 | * Opening of "enable" adds a ref count to system, | ||
| 610 | * so the name is safe to use. | ||
| 611 | */ | ||
| 612 | if (system) | ||
| 613 | name = system->name; | ||
| 614 | |||
| 615 | ret = __ftrace_set_clr_event(NULL, name, NULL, val); | ||
| 597 | if (ret) | 616 | if (ret) |
| 598 | goto out; | 617 | goto out; |
| 599 | 618 | ||
| @@ -826,6 +845,52 @@ event_filter_write(struct file *filp, const char __user *ubuf, size_t cnt, | |||
| 826 | return cnt; | 845 | return cnt; |
| 827 | } | 846 | } |
| 828 | 847 | ||
| 848 | static LIST_HEAD(event_subsystems); | ||
| 849 | |||
| 850 | static int subsystem_open(struct inode *inode, struct file *filp) | ||
| 851 | { | ||
| 852 | struct event_subsystem *system = NULL; | ||
| 853 | int ret; | ||
| 854 | |||
| 855 | if (!inode->i_private) | ||
| 856 | goto skip_search; | ||
| 857 | |||
| 858 | /* Make sure the system still exists */ | ||
| 859 | mutex_lock(&event_mutex); | ||
| 860 | list_for_each_entry(system, &event_subsystems, list) { | ||
| 861 | if (system == inode->i_private) { | ||
| 862 | /* Don't open systems with no events */ | ||
| 863 | if (!system->nr_events) { | ||
| 864 | system = NULL; | ||
| 865 | break; | ||
| 866 | } | ||
| 867 | __get_system(system); | ||
| 868 | break; | ||
| 869 | } | ||
| 870 | } | ||
| 871 | mutex_unlock(&event_mutex); | ||
| 872 | |||
| 873 | if (system != inode->i_private) | ||
| 874 | return -ENODEV; | ||
| 875 | |||
| 876 | skip_search: | ||
| 877 | ret = tracing_open_generic(inode, filp); | ||
| 878 | if (ret < 0 && system) | ||
| 879 | put_system(system); | ||
| 880 | |||
| 881 | return ret; | ||
| 882 | } | ||
| 883 | |||
| 884 | static int subsystem_release(struct inode *inode, struct file *file) | ||
| 885 | { | ||
| 886 | struct event_subsystem *system = inode->i_private; | ||
| 887 | |||
| 888 | if (system) | ||
| 889 | put_system(system); | ||
| 890 | |||
| 891 | return 0; | ||
| 892 | } | ||
| 893 | |||
| 829 | static ssize_t | 894 | static ssize_t |
| 830 | subsystem_filter_read(struct file *filp, char __user *ubuf, size_t cnt, | 895 | subsystem_filter_read(struct file *filp, char __user *ubuf, size_t cnt, |
| 831 | loff_t *ppos) | 896 | loff_t *ppos) |
| @@ -963,17 +1028,19 @@ static const struct file_operations ftrace_event_filter_fops = { | |||
| 963 | }; | 1028 | }; |
| 964 | 1029 | ||
| 965 | static const struct file_operations ftrace_subsystem_filter_fops = { | 1030 | static const struct file_operations ftrace_subsystem_filter_fops = { |
| 966 | .open = tracing_open_generic, | 1031 | .open = subsystem_open, |
| 967 | .read = subsystem_filter_read, | 1032 | .read = subsystem_filter_read, |
| 968 | .write = subsystem_filter_write, | 1033 | .write = subsystem_filter_write, |
| 969 | .llseek = default_llseek, | 1034 | .llseek = default_llseek, |
| 1035 | .release = subsystem_release, | ||
| 970 | }; | 1036 | }; |
| 971 | 1037 | ||
| 972 | static const struct file_operations ftrace_system_enable_fops = { | 1038 | static const struct file_operations ftrace_system_enable_fops = { |
| 973 | .open = tracing_open_generic, | 1039 | .open = subsystem_open, |
| 974 | .read = system_enable_read, | 1040 | .read = system_enable_read, |
| 975 | .write = system_enable_write, | 1041 | .write = system_enable_write, |
| 976 | .llseek = default_llseek, | 1042 | .llseek = default_llseek, |
| 1043 | .release = subsystem_release, | ||
| 977 | }; | 1044 | }; |
| 978 | 1045 | ||
| 979 | static const struct file_operations ftrace_show_header_fops = { | 1046 | static const struct file_operations ftrace_show_header_fops = { |
| @@ -1002,8 +1069,6 @@ static struct dentry *event_trace_events_dir(void) | |||
| 1002 | return d_events; | 1069 | return d_events; |
| 1003 | } | 1070 | } |
| 1004 | 1071 | ||
| 1005 | static LIST_HEAD(event_subsystems); | ||
| 1006 | |||
| 1007 | static struct dentry * | 1072 | static struct dentry * |
| 1008 | event_subsystem_dir(const char *name, struct dentry *d_events) | 1073 | event_subsystem_dir(const char *name, struct dentry *d_events) |
| 1009 | { | 1074 | { |
| @@ -1013,6 +1078,7 @@ event_subsystem_dir(const char *name, struct dentry *d_events) | |||
| 1013 | /* First see if we did not already create this dir */ | 1078 | /* First see if we did not already create this dir */ |
| 1014 | list_for_each_entry(system, &event_subsystems, list) { | 1079 | list_for_each_entry(system, &event_subsystems, list) { |
| 1015 | if (strcmp(system->name, name) == 0) { | 1080 | if (strcmp(system->name, name) == 0) { |
| 1081 | __get_system(system); | ||
| 1016 | system->nr_events++; | 1082 | system->nr_events++; |
| 1017 | return system->entry; | 1083 | return system->entry; |
| 1018 | } | 1084 | } |
| @@ -1035,6 +1101,7 @@ event_subsystem_dir(const char *name, struct dentry *d_events) | |||
| 1035 | } | 1101 | } |
| 1036 | 1102 | ||
| 1037 | system->nr_events = 1; | 1103 | system->nr_events = 1; |
| 1104 | system->ref_count = 1; | ||
| 1038 | system->name = kstrdup(name, GFP_KERNEL); | 1105 | system->name = kstrdup(name, GFP_KERNEL); |
| 1039 | if (!system->name) { | 1106 | if (!system->name) { |
| 1040 | debugfs_remove(system->entry); | 1107 | debugfs_remove(system->entry); |
| @@ -1062,8 +1129,7 @@ event_subsystem_dir(const char *name, struct dentry *d_events) | |||
| 1062 | "'%s/filter' entry\n", name); | 1129 | "'%s/filter' entry\n", name); |
| 1063 | } | 1130 | } |
| 1064 | 1131 | ||
| 1065 | trace_create_file("enable", 0644, system->entry, | 1132 | trace_create_file("enable", 0644, system->entry, system, |
| 1066 | (void *)system->name, | ||
| 1067 | &ftrace_system_enable_fops); | 1133 | &ftrace_system_enable_fops); |
| 1068 | 1134 | ||
| 1069 | return system->entry; | 1135 | return system->entry; |
| @@ -1184,16 +1250,9 @@ static void remove_subsystem_dir(const char *name) | |||
| 1184 | list_for_each_entry(system, &event_subsystems, list) { | 1250 | list_for_each_entry(system, &event_subsystems, list) { |
| 1185 | if (strcmp(system->name, name) == 0) { | 1251 | if (strcmp(system->name, name) == 0) { |
| 1186 | if (!--system->nr_events) { | 1252 | if (!--system->nr_events) { |
| 1187 | struct event_filter *filter = system->filter; | ||
| 1188 | |||
| 1189 | debugfs_remove_recursive(system->entry); | 1253 | debugfs_remove_recursive(system->entry); |
| 1190 | list_del(&system->list); | 1254 | list_del(&system->list); |
| 1191 | if (filter) { | 1255 | __put_system(system); |
| 1192 | kfree(filter->filter_string); | ||
| 1193 | kfree(filter); | ||
| 1194 | } | ||
| 1195 | kfree(system->name); | ||
| 1196 | kfree(system); | ||
| 1197 | } | 1256 | } |
| 1198 | break; | 1257 | break; |
| 1199 | } | 1258 | } |
diff --git a/kernel/trace/trace_events_filter.c b/kernel/trace/trace_events_filter.c index 8008ddcfbf2..256764ecccd 100644 --- a/kernel/trace/trace_events_filter.c +++ b/kernel/trace/trace_events_filter.c | |||
| @@ -1886,6 +1886,12 @@ int apply_subsystem_event_filter(struct event_subsystem *system, | |||
| 1886 | 1886 | ||
| 1887 | mutex_lock(&event_mutex); | 1887 | mutex_lock(&event_mutex); |
| 1888 | 1888 | ||
| 1889 | /* Make sure the system still has events */ | ||
| 1890 | if (!system->nr_events) { | ||
| 1891 | err = -ENODEV; | ||
| 1892 | goto out_unlock; | ||
| 1893 | } | ||
| 1894 | |||
| 1889 | if (!strcmp(strstrip(filter_string), "0")) { | 1895 | if (!strcmp(strstrip(filter_string), "0")) { |
| 1890 | filter_free_subsystem_preds(system); | 1896 | filter_free_subsystem_preds(system); |
| 1891 | remove_filter_string(system->filter); | 1897 | remove_filter_string(system->filter); |
diff --git a/kernel/trace/trace_functions.c b/kernel/trace/trace_functions.c index 8d0e1cc4e97..c7b0c6a7db0 100644 --- a/kernel/trace/trace_functions.c +++ b/kernel/trace/trace_functions.c | |||
| @@ -324,7 +324,8 @@ ftrace_trace_onoff_unreg(char *glob, char *cmd, char *param) | |||
| 324 | } | 324 | } |
| 325 | 325 | ||
| 326 | static int | 326 | static int |
| 327 | ftrace_trace_onoff_callback(char *glob, char *cmd, char *param, int enable) | 327 | ftrace_trace_onoff_callback(struct ftrace_hash *hash, |
| 328 | char *glob, char *cmd, char *param, int enable) | ||
| 328 | { | 329 | { |
| 329 | struct ftrace_probe_ops *ops; | 330 | struct ftrace_probe_ops *ops; |
| 330 | void *count = (void *)-1; | 331 | void *count = (void *)-1; |
diff --git a/kernel/trace/trace_functions_graph.c b/kernel/trace/trace_functions_graph.c index 962cdb24ed8..a7d2a4c653d 100644 --- a/kernel/trace/trace_functions_graph.c +++ b/kernel/trace/trace_functions_graph.c | |||
| @@ -74,6 +74,20 @@ static struct tracer_flags tracer_flags = { | |||
| 74 | 74 | ||
| 75 | static struct trace_array *graph_array; | 75 | static struct trace_array *graph_array; |
| 76 | 76 | ||
| 77 | /* | ||
| 78 | * DURATION column is being also used to display IRQ signs, | ||
| 79 | * following values are used by print_graph_irq and others | ||
| 80 | * to fill in space into DURATION column. | ||
| 81 | */ | ||
| 82 | enum { | ||
| 83 | DURATION_FILL_FULL = -1, | ||
| 84 | DURATION_FILL_START = -2, | ||
| 85 | DURATION_FILL_END = -3, | ||
| 86 | }; | ||
| 87 | |||
| 88 | static enum print_line_t | ||
| 89 | print_graph_duration(unsigned long long duration, struct trace_seq *s, | ||
| 90 | u32 flags); | ||
| 77 | 91 | ||
| 78 | /* Add a function return address to the trace stack on thread info.*/ | 92 | /* Add a function return address to the trace stack on thread info.*/ |
| 79 | int | 93 | int |
| @@ -213,7 +227,7 @@ int __trace_graph_entry(struct trace_array *tr, | |||
| 213 | 227 | ||
| 214 | static inline int ftrace_graph_ignore_irqs(void) | 228 | static inline int ftrace_graph_ignore_irqs(void) |
| 215 | { | 229 | { |
| 216 | if (!ftrace_graph_skip_irqs) | 230 | if (!ftrace_graph_skip_irqs || trace_recursion_test(TRACE_IRQ_BIT)) |
| 217 | return 0; | 231 | return 0; |
| 218 | 232 | ||
| 219 | return in_irq(); | 233 | return in_irq(); |
| @@ -577,32 +591,6 @@ get_return_for_leaf(struct trace_iterator *iter, | |||
| 577 | return next; | 591 | return next; |
| 578 | } | 592 | } |
| 579 | 593 | ||
| 580 | /* Signal a overhead of time execution to the output */ | ||
| 581 | static int | ||
| 582 | print_graph_overhead(unsigned long long duration, struct trace_seq *s, | ||
| 583 | u32 flags) | ||
| 584 | { | ||
| 585 | /* If duration disappear, we don't need anything */ | ||
| 586 | if (!(flags & TRACE_GRAPH_PRINT_DURATION)) | ||
| 587 | return 1; | ||
| 588 | |||
| 589 | /* Non nested entry or return */ | ||
| 590 | if (duration == -1) | ||
| 591 | return trace_seq_printf(s, " "); | ||
| 592 | |||
| 593 | if (flags & TRACE_GRAPH_PRINT_OVERHEAD) { | ||
| 594 | /* Duration exceeded 100 msecs */ | ||
| 595 | if (duration > 100000ULL) | ||
| 596 | return trace_seq_printf(s, "! "); | ||
| 597 | |||
| 598 | /* Duration exceeded 10 msecs */ | ||
| 599 | if (duration > 10000ULL) | ||
| 600 | return trace_seq_printf(s, "+ "); | ||
| 601 | } | ||
| 602 | |||
| 603 | return trace_seq_printf(s, " "); | ||
| 604 | } | ||
| 605 | |||
| 606 | static int print_graph_abs_time(u64 t, struct trace_seq *s) | 594 | static int print_graph_abs_time(u64 t, struct trace_seq *s) |
| 607 | { | 595 | { |
| 608 | unsigned long usecs_rem; | 596 | unsigned long usecs_rem; |
| @@ -625,34 +613,36 @@ print_graph_irq(struct trace_iterator *iter, unsigned long addr, | |||
| 625 | addr >= (unsigned long)__irqentry_text_end) | 613 | addr >= (unsigned long)__irqentry_text_end) |
| 626 | return TRACE_TYPE_UNHANDLED; | 614 | return TRACE_TYPE_UNHANDLED; |
| 627 | 615 | ||
| 628 | /* Absolute time */ | 616 | if (trace_flags & TRACE_ITER_CONTEXT_INFO) { |
| 629 | if (flags & TRACE_GRAPH_PRINT_ABS_TIME) { | 617 | /* Absolute time */ |
| 630 | ret = print_graph_abs_time(iter->ts, s); | 618 | if (flags & TRACE_GRAPH_PRINT_ABS_TIME) { |
| 631 | if (!ret) | 619 | ret = print_graph_abs_time(iter->ts, s); |
| 632 | return TRACE_TYPE_PARTIAL_LINE; | 620 | if (!ret) |
| 633 | } | 621 | return TRACE_TYPE_PARTIAL_LINE; |
| 622 | } | ||
| 634 | 623 | ||
| 635 | /* Cpu */ | 624 | /* Cpu */ |
| 636 | if (flags & TRACE_GRAPH_PRINT_CPU) { | 625 | if (flags & TRACE_GRAPH_PRINT_CPU) { |
| 637 | ret = print_graph_cpu(s, cpu); | 626 | ret = print_graph_cpu(s, cpu); |
| 638 | if (ret == TRACE_TYPE_PARTIAL_LINE) | 627 | if (ret == TRACE_TYPE_PARTIAL_LINE) |
| 639 | return TRACE_TYPE_PARTIAL_LINE; | 628 | return TRACE_TYPE_PARTIAL_LINE; |
| 640 | } | 629 | } |
| 641 | 630 | ||
| 642 | /* Proc */ | 631 | /* Proc */ |
| 643 | if (flags & TRACE_GRAPH_PRINT_PROC) { | 632 | if (flags & TRACE_GRAPH_PRINT_PROC) { |
| 644 | ret = print_graph_proc(s, pid); | 633 | ret = print_graph_proc(s, pid); |
| 645 | if (ret == TRACE_TYPE_PARTIAL_LINE) | 634 | if (ret == TRACE_TYPE_PARTIAL_LINE) |
| 646 | return TRACE_TYPE_PARTIAL_LINE; | 635 | return TRACE_TYPE_PARTIAL_LINE; |
| 647 | ret = trace_seq_printf(s, " | "); | 636 | ret = trace_seq_printf(s, " | "); |
| 648 | if (!ret) | 637 | if (!ret) |
| 649 | return TRACE_TYPE_PARTIAL_LINE; | 638 | return TRACE_TYPE_PARTIAL_LINE; |
| 639 | } | ||
| 650 | } | 640 | } |
| 651 | 641 | ||
| 652 | /* No overhead */ | 642 | /* No overhead */ |
| 653 | ret = print_graph_overhead(-1, s, flags); | 643 | ret = print_graph_duration(DURATION_FILL_START, s, flags); |
| 654 | if (!ret) | 644 | if (ret != TRACE_TYPE_HANDLED) |
| 655 | return TRACE_TYPE_PARTIAL_LINE; | 645 | return ret; |
| 656 | 646 | ||
| 657 | if (type == TRACE_GRAPH_ENT) | 647 | if (type == TRACE_GRAPH_ENT) |
| 658 | ret = trace_seq_printf(s, "==========>"); | 648 | ret = trace_seq_printf(s, "==========>"); |
| @@ -662,9 +652,10 @@ print_graph_irq(struct trace_iterator *iter, unsigned long addr, | |||
| 662 | if (!ret) | 652 | if (!ret) |
| 663 | return TRACE_TYPE_PARTIAL_LINE; | 653 | return TRACE_TYPE_PARTIAL_LINE; |
| 664 | 654 | ||
| 665 | /* Don't close the duration column if haven't one */ | 655 | ret = print_graph_duration(DURATION_FILL_END, s, flags); |
| 666 | if (flags & TRACE_GRAPH_PRINT_DURATION) | 656 | if (ret != TRACE_TYPE_HANDLED) |
| 667 | trace_seq_printf(s, " |"); | 657 | return ret; |
| 658 | |||
| 668 | ret = trace_seq_printf(s, "\n"); | 659 | ret = trace_seq_printf(s, "\n"); |
| 669 | 660 | ||
| 670 | if (!ret) | 661 | if (!ret) |
| @@ -716,9 +707,49 @@ trace_print_graph_duration(unsigned long long duration, struct trace_seq *s) | |||
| 716 | } | 707 | } |
| 717 | 708 | ||
| 718 | static enum print_line_t | 709 | static enum print_line_t |
| 719 | print_graph_duration(unsigned long long duration, struct trace_seq *s) | 710 | print_graph_duration(unsigned long long duration, struct trace_seq *s, |
| 711 | u32 flags) | ||
| 720 | { | 712 | { |
| 721 | int ret; | 713 | int ret = -1; |
| 714 | |||
| 715 | if (!(flags & TRACE_GRAPH_PRINT_DURATION) || | ||
| 716 | !(trace_flags & TRACE_ITER_CONTEXT_INFO)) | ||
| 717 | return TRACE_TYPE_HANDLED; | ||
| 718 | |||
| 719 | /* No real adata, just filling the column with spaces */ | ||
| 720 | switch (duration) { | ||
| 721 | case DURATION_FILL_FULL: | ||
| 722 | ret = trace_seq_printf(s, " | "); | ||
| 723 | return ret ? TRACE_TYPE_HANDLED : TRACE_TYPE_PARTIAL_LINE; | ||
| 724 | case DURATION_FILL_START: | ||
| 725 | ret = trace_seq_printf(s, " "); | ||
| 726 | return ret ? TRACE_TYPE_HANDLED : TRACE_TYPE_PARTIAL_LINE; | ||
| 727 | case DURATION_FILL_END: | ||
| 728 | ret = trace_seq_printf(s, " |"); | ||
| 729 | return ret ? TRACE_TYPE_HANDLED : TRACE_TYPE_PARTIAL_LINE; | ||
| 730 | } | ||
| 731 | |||
| 732 | /* Signal a overhead of time execution to the output */ | ||
| 733 | if (flags & TRACE_GRAPH_PRINT_OVERHEAD) { | ||
| 734 | /* Duration exceeded 100 msecs */ | ||
| 735 | if (duration > 100000ULL) | ||
| 736 | ret = trace_seq_printf(s, "! "); | ||
| 737 | /* Duration exceeded 10 msecs */ | ||
| 738 | else if (duration > 10000ULL) | ||
| 739 | ret = trace_seq_printf(s, "+ "); | ||
| 740 | } | ||
| 741 | |||
| 742 | /* | ||
| 743 | * The -1 means we either did not exceed the duration tresholds | ||
| 744 | * or we dont want to print out the overhead. Either way we need | ||
| 745 | * to fill out the space. | ||
| 746 | */ | ||
| 747 | if (ret == -1) | ||
| 748 | ret = trace_seq_printf(s, " "); | ||
| 749 | |||
| 750 | /* Catching here any failure happenned above */ | ||
| 751 | if (!ret) | ||
| 752 | return TRACE_TYPE_PARTIAL_LINE; | ||
| 722 | 753 | ||
| 723 | ret = trace_print_graph_duration(duration, s); | 754 | ret = trace_print_graph_duration(duration, s); |
| 724 | if (ret != TRACE_TYPE_HANDLED) | 755 | if (ret != TRACE_TYPE_HANDLED) |
| @@ -767,18 +798,11 @@ print_graph_entry_leaf(struct trace_iterator *iter, | |||
| 767 | cpu_data->enter_funcs[call->depth] = 0; | 798 | cpu_data->enter_funcs[call->depth] = 0; |
| 768 | } | 799 | } |
| 769 | 800 | ||
| 770 | /* Overhead */ | 801 | /* Overhead and duration */ |
| 771 | ret = print_graph_overhead(duration, s, flags); | 802 | ret = print_graph_duration(duration, s, flags); |
| 772 | if (!ret) | 803 | if (ret == TRACE_TYPE_PARTIAL_LINE) |
| 773 | return TRACE_TYPE_PARTIAL_LINE; | 804 | return TRACE_TYPE_PARTIAL_LINE; |
| 774 | 805 | ||
| 775 | /* Duration */ | ||
| 776 | if (flags & TRACE_GRAPH_PRINT_DURATION) { | ||
| 777 | ret = print_graph_duration(duration, s); | ||
| 778 | if (ret == TRACE_TYPE_PARTIAL_LINE) | ||
| 779 | return TRACE_TYPE_PARTIAL_LINE; | ||
| 780 | } | ||
| 781 | |||
| 782 | /* Function */ | 806 | /* Function */ |
| 783 | for (i = 0; i < call->depth * TRACE_GRAPH_INDENT; i++) { | 807 | for (i = 0; i < call->depth * TRACE_GRAPH_INDENT; i++) { |
| 784 | ret = trace_seq_printf(s, " "); | 808 | ret = trace_seq_printf(s, " "); |
| @@ -815,17 +839,10 @@ print_graph_entry_nested(struct trace_iterator *iter, | |||
| 815 | cpu_data->enter_funcs[call->depth] = call->func; | 839 | cpu_data->enter_funcs[call->depth] = call->func; |
| 816 | } | 840 | } |
| 817 | 841 | ||
| 818 | /* No overhead */ | ||
| 819 | ret = print_graph_overhead(-1, s, flags); | ||
| 820 | if (!ret) | ||
| 821 | return TRACE_TYPE_PARTIAL_LINE; | ||
| 822 | |||
| 823 | /* No time */ | 842 | /* No time */ |
| 824 | if (flags & TRACE_GRAPH_PRINT_DURATION) { | 843 | ret = print_graph_duration(DURATION_FILL_FULL, s, flags); |
| 825 | ret = trace_seq_printf(s, " | "); | 844 | if (ret != TRACE_TYPE_HANDLED) |
| 826 | if (!ret) | 845 | return ret; |
| 827 | return TRACE_TYPE_PARTIAL_LINE; | ||
| 828 | } | ||
| 829 | 846 | ||
| 830 | /* Function */ | 847 | /* Function */ |
| 831 | for (i = 0; i < call->depth * TRACE_GRAPH_INDENT; i++) { | 848 | for (i = 0; i < call->depth * TRACE_GRAPH_INDENT; i++) { |
| @@ -865,6 +882,9 @@ print_graph_prologue(struct trace_iterator *iter, struct trace_seq *s, | |||
| 865 | return TRACE_TYPE_PARTIAL_LINE; | 882 | return TRACE_TYPE_PARTIAL_LINE; |
| 866 | } | 883 | } |
| 867 | 884 | ||
| 885 | if (!(trace_flags & TRACE_ITER_CONTEXT_INFO)) | ||
| 886 | return 0; | ||
| 887 | |||
| 868 | /* Absolute time */ | 888 | /* Absolute time */ |
| 869 | if (flags & TRACE_GRAPH_PRINT_ABS_TIME) { | 889 | if (flags & TRACE_GRAPH_PRINT_ABS_TIME) { |
| 870 | ret = print_graph_abs_time(iter->ts, s); | 890 | ret = print_graph_abs_time(iter->ts, s); |
| @@ -1078,18 +1098,11 @@ print_graph_return(struct ftrace_graph_ret *trace, struct trace_seq *s, | |||
| 1078 | if (print_graph_prologue(iter, s, 0, 0, flags)) | 1098 | if (print_graph_prologue(iter, s, 0, 0, flags)) |
| 1079 | return TRACE_TYPE_PARTIAL_LINE; | 1099 | return TRACE_TYPE_PARTIAL_LINE; |
| 1080 | 1100 | ||
| 1081 | /* Overhead */ | 1101 | /* Overhead and duration */ |
| 1082 | ret = print_graph_overhead(duration, s, flags); | 1102 | ret = print_graph_duration(duration, s, flags); |
| 1083 | if (!ret) | 1103 | if (ret == TRACE_TYPE_PARTIAL_LINE) |
| 1084 | return TRACE_TYPE_PARTIAL_LINE; | 1104 | return TRACE_TYPE_PARTIAL_LINE; |
| 1085 | 1105 | ||
| 1086 | /* Duration */ | ||
| 1087 | if (flags & TRACE_GRAPH_PRINT_DURATION) { | ||
| 1088 | ret = print_graph_duration(duration, s); | ||
| 1089 | if (ret == TRACE_TYPE_PARTIAL_LINE) | ||
| 1090 | return TRACE_TYPE_PARTIAL_LINE; | ||
| 1091 | } | ||
| 1092 | |||
| 1093 | /* Closing brace */ | 1106 | /* Closing brace */ |
| 1094 | for (i = 0; i < trace->depth * TRACE_GRAPH_INDENT; i++) { | 1107 | for (i = 0; i < trace->depth * TRACE_GRAPH_INDENT; i++) { |
| 1095 | ret = trace_seq_printf(s, " "); | 1108 | ret = trace_seq_printf(s, " "); |
| @@ -1146,17 +1159,10 @@ print_graph_comment(struct trace_seq *s, struct trace_entry *ent, | |||
| 1146 | if (print_graph_prologue(iter, s, 0, 0, flags)) | 1159 | if (print_graph_prologue(iter, s, 0, 0, flags)) |
| 1147 | return TRACE_TYPE_PARTIAL_LINE; | 1160 | return TRACE_TYPE_PARTIAL_LINE; |
| 1148 | 1161 | ||
| 1149 | /* No overhead */ | ||
| 1150 | ret = print_graph_overhead(-1, s, flags); | ||
| 1151 | if (!ret) | ||
| 1152 | return TRACE_TYPE_PARTIAL_LINE; | ||
| 1153 | |||
| 1154 | /* No time */ | 1162 | /* No time */ |
| 1155 | if (flags & TRACE_GRAPH_PRINT_DURATION) { | 1163 | ret = print_graph_duration(DURATION_FILL_FULL, s, flags); |
| 1156 | ret = trace_seq_printf(s, " | "); | 1164 | if (ret != TRACE_TYPE_HANDLED) |
| 1157 | if (!ret) | 1165 | return ret; |
| 1158 | return TRACE_TYPE_PARTIAL_LINE; | ||
| 1159 | } | ||
| 1160 | 1166 | ||
| 1161 | /* Indentation */ | 1167 | /* Indentation */ |
| 1162 | if (depth > 0) | 1168 | if (depth > 0) |
| @@ -1207,7 +1213,7 @@ print_graph_comment(struct trace_seq *s, struct trace_entry *ent, | |||
| 1207 | 1213 | ||
| 1208 | 1214 | ||
| 1209 | enum print_line_t | 1215 | enum print_line_t |
| 1210 | __print_graph_function_flags(struct trace_iterator *iter, u32 flags) | 1216 | print_graph_function_flags(struct trace_iterator *iter, u32 flags) |
| 1211 | { | 1217 | { |
| 1212 | struct ftrace_graph_ent_entry *field; | 1218 | struct ftrace_graph_ent_entry *field; |
| 1213 | struct fgraph_data *data = iter->private; | 1219 | struct fgraph_data *data = iter->private; |
| @@ -1270,18 +1276,7 @@ __print_graph_function_flags(struct trace_iterator *iter, u32 flags) | |||
| 1270 | static enum print_line_t | 1276 | static enum print_line_t |
| 1271 | print_graph_function(struct trace_iterator *iter) | 1277 | print_graph_function(struct trace_iterator *iter) |
| 1272 | { | 1278 | { |
| 1273 | return __print_graph_function_flags(iter, tracer_flags.val); | 1279 | return print_graph_function_flags(iter, tracer_flags.val); |
| 1274 | } | ||
| 1275 | |||
| 1276 | enum print_line_t print_graph_function_flags(struct trace_iterator *iter, | ||
| 1277 | u32 flags) | ||
| 1278 | { | ||
| 1279 | if (trace_flags & TRACE_ITER_LATENCY_FMT) | ||
| 1280 | flags |= TRACE_GRAPH_PRINT_DURATION; | ||
| 1281 | else | ||
| 1282 | flags |= TRACE_GRAPH_PRINT_ABS_TIME; | ||
| 1283 | |||
| 1284 | return __print_graph_function_flags(iter, flags); | ||
| 1285 | } | 1280 | } |
| 1286 | 1281 | ||
| 1287 | static enum print_line_t | 1282 | static enum print_line_t |
| @@ -1309,8 +1304,7 @@ static void print_lat_header(struct seq_file *s, u32 flags) | |||
| 1309 | seq_printf(s, "#%.*s / _----=> need-resched \n", size, spaces); | 1304 | seq_printf(s, "#%.*s / _----=> need-resched \n", size, spaces); |
| 1310 | seq_printf(s, "#%.*s| / _---=> hardirq/softirq \n", size, spaces); | 1305 | seq_printf(s, "#%.*s| / _---=> hardirq/softirq \n", size, spaces); |
| 1311 | seq_printf(s, "#%.*s|| / _--=> preempt-depth \n", size, spaces); | 1306 | seq_printf(s, "#%.*s|| / _--=> preempt-depth \n", size, spaces); |
| 1312 | seq_printf(s, "#%.*s||| / _-=> lock-depth \n", size, spaces); | 1307 | seq_printf(s, "#%.*s||| / \n", size, spaces); |
| 1313 | seq_printf(s, "#%.*s|||| / \n", size, spaces); | ||
| 1314 | } | 1308 | } |
| 1315 | 1309 | ||
| 1316 | static void __print_graph_headers_flags(struct seq_file *s, u32 flags) | 1310 | static void __print_graph_headers_flags(struct seq_file *s, u32 flags) |
| @@ -1329,7 +1323,7 @@ static void __print_graph_headers_flags(struct seq_file *s, u32 flags) | |||
| 1329 | if (flags & TRACE_GRAPH_PRINT_PROC) | 1323 | if (flags & TRACE_GRAPH_PRINT_PROC) |
| 1330 | seq_printf(s, " TASK/PID "); | 1324 | seq_printf(s, " TASK/PID "); |
| 1331 | if (lat) | 1325 | if (lat) |
| 1332 | seq_printf(s, "|||||"); | 1326 | seq_printf(s, "||||"); |
| 1333 | if (flags & TRACE_GRAPH_PRINT_DURATION) | 1327 | if (flags & TRACE_GRAPH_PRINT_DURATION) |
| 1334 | seq_printf(s, " DURATION "); | 1328 | seq_printf(s, " DURATION "); |
| 1335 | seq_printf(s, " FUNCTION CALLS\n"); | 1329 | seq_printf(s, " FUNCTION CALLS\n"); |
| @@ -1343,7 +1337,7 @@ static void __print_graph_headers_flags(struct seq_file *s, u32 flags) | |||
| 1343 | if (flags & TRACE_GRAPH_PRINT_PROC) | 1337 | if (flags & TRACE_GRAPH_PRINT_PROC) |
| 1344 | seq_printf(s, " | | "); | 1338 | seq_printf(s, " | | "); |
| 1345 | if (lat) | 1339 | if (lat) |
| 1346 | seq_printf(s, "|||||"); | 1340 | seq_printf(s, "||||"); |
| 1347 | if (flags & TRACE_GRAPH_PRINT_DURATION) | 1341 | if (flags & TRACE_GRAPH_PRINT_DURATION) |
| 1348 | seq_printf(s, " | | "); | 1342 | seq_printf(s, " | | "); |
| 1349 | seq_printf(s, " | | | |\n"); | 1343 | seq_printf(s, " | | | |\n"); |
| @@ -1358,15 +1352,16 @@ void print_graph_headers_flags(struct seq_file *s, u32 flags) | |||
| 1358 | { | 1352 | { |
| 1359 | struct trace_iterator *iter = s->private; | 1353 | struct trace_iterator *iter = s->private; |
| 1360 | 1354 | ||
| 1355 | if (!(trace_flags & TRACE_ITER_CONTEXT_INFO)) | ||
| 1356 | return; | ||
| 1357 | |||
| 1361 | if (trace_flags & TRACE_ITER_LATENCY_FMT) { | 1358 | if (trace_flags & TRACE_ITER_LATENCY_FMT) { |
| 1362 | /* print nothing if the buffers are empty */ | 1359 | /* print nothing if the buffers are empty */ |
| 1363 | if (trace_empty(iter)) | 1360 | if (trace_empty(iter)) |
| 1364 | return; | 1361 | return; |
| 1365 | 1362 | ||
| 1366 | print_trace_header(s, iter); | 1363 | print_trace_header(s, iter); |
| 1367 | flags |= TRACE_GRAPH_PRINT_DURATION; | 1364 | } |
| 1368 | } else | ||
| 1369 | flags |= TRACE_GRAPH_PRINT_ABS_TIME; | ||
| 1370 | 1365 | ||
| 1371 | __print_graph_headers_flags(s, flags); | 1366 | __print_graph_headers_flags(s, flags); |
| 1372 | } | 1367 | } |
diff --git a/kernel/trace/trace_irqsoff.c b/kernel/trace/trace_irqsoff.c index c77424be284..667aa8cc0cf 100644 --- a/kernel/trace/trace_irqsoff.c +++ b/kernel/trace/trace_irqsoff.c | |||
| @@ -226,7 +226,9 @@ static void irqsoff_trace_close(struct trace_iterator *iter) | |||
| 226 | } | 226 | } |
| 227 | 227 | ||
| 228 | #define GRAPH_TRACER_FLAGS (TRACE_GRAPH_PRINT_CPU | \ | 228 | #define GRAPH_TRACER_FLAGS (TRACE_GRAPH_PRINT_CPU | \ |
| 229 | TRACE_GRAPH_PRINT_PROC) | 229 | TRACE_GRAPH_PRINT_PROC | \ |
| 230 | TRACE_GRAPH_PRINT_ABS_TIME | \ | ||
| 231 | TRACE_GRAPH_PRINT_DURATION) | ||
| 230 | 232 | ||
| 231 | static enum print_line_t irqsoff_print_line(struct trace_iterator *iter) | 233 | static enum print_line_t irqsoff_print_line(struct trace_iterator *iter) |
| 232 | { | 234 | { |
diff --git a/kernel/trace/trace_kprobe.c b/kernel/trace/trace_kprobe.c index 27d13b36b8b..5fb3697bf0e 100644 --- a/kernel/trace/trace_kprobe.c +++ b/kernel/trace/trace_kprobe.c | |||
| @@ -343,6 +343,14 @@ DEFINE_BASIC_FETCH_FUNCS(deref) | |||
| 343 | DEFINE_FETCH_deref(string) | 343 | DEFINE_FETCH_deref(string) |
| 344 | DEFINE_FETCH_deref(string_size) | 344 | DEFINE_FETCH_deref(string_size) |
| 345 | 345 | ||
| 346 | static __kprobes void update_deref_fetch_param(struct deref_fetch_param *data) | ||
| 347 | { | ||
| 348 | if (CHECK_FETCH_FUNCS(deref, data->orig.fn)) | ||
| 349 | update_deref_fetch_param(data->orig.data); | ||
| 350 | else if (CHECK_FETCH_FUNCS(symbol, data->orig.fn)) | ||
| 351 | update_symbol_cache(data->orig.data); | ||
| 352 | } | ||
| 353 | |||
| 346 | static __kprobes void free_deref_fetch_param(struct deref_fetch_param *data) | 354 | static __kprobes void free_deref_fetch_param(struct deref_fetch_param *data) |
| 347 | { | 355 | { |
| 348 | if (CHECK_FETCH_FUNCS(deref, data->orig.fn)) | 356 | if (CHECK_FETCH_FUNCS(deref, data->orig.fn)) |
| @@ -377,6 +385,19 @@ DEFINE_BASIC_FETCH_FUNCS(bitfield) | |||
| 377 | #define fetch_bitfield_string_size NULL | 385 | #define fetch_bitfield_string_size NULL |
| 378 | 386 | ||
| 379 | static __kprobes void | 387 | static __kprobes void |
| 388 | update_bitfield_fetch_param(struct bitfield_fetch_param *data) | ||
| 389 | { | ||
| 390 | /* | ||
| 391 | * Don't check the bitfield itself, because this must be the | ||
| 392 | * last fetch function. | ||
| 393 | */ | ||
| 394 | if (CHECK_FETCH_FUNCS(deref, data->orig.fn)) | ||
| 395 | update_deref_fetch_param(data->orig.data); | ||
| 396 | else if (CHECK_FETCH_FUNCS(symbol, data->orig.fn)) | ||
| 397 | update_symbol_cache(data->orig.data); | ||
| 398 | } | ||
| 399 | |||
| 400 | static __kprobes void | ||
| 380 | free_bitfield_fetch_param(struct bitfield_fetch_param *data) | 401 | free_bitfield_fetch_param(struct bitfield_fetch_param *data) |
| 381 | { | 402 | { |
| 382 | /* | 403 | /* |
| @@ -389,6 +410,7 @@ free_bitfield_fetch_param(struct bitfield_fetch_param *data) | |||
| 389 | free_symbol_cache(data->orig.data); | 410 | free_symbol_cache(data->orig.data); |
| 390 | kfree(data); | 411 | kfree(data); |
| 391 | } | 412 | } |
| 413 | |||
| 392 | /* Default (unsigned long) fetch type */ | 414 | /* Default (unsigned long) fetch type */ |
| 393 | #define __DEFAULT_FETCH_TYPE(t) u##t | 415 | #define __DEFAULT_FETCH_TYPE(t) u##t |
| 394 | #define _DEFAULT_FETCH_TYPE(t) __DEFAULT_FETCH_TYPE(t) | 416 | #define _DEFAULT_FETCH_TYPE(t) __DEFAULT_FETCH_TYPE(t) |
| @@ -536,6 +558,7 @@ struct probe_arg { | |||
| 536 | /* Flags for trace_probe */ | 558 | /* Flags for trace_probe */ |
| 537 | #define TP_FLAG_TRACE 1 | 559 | #define TP_FLAG_TRACE 1 |
| 538 | #define TP_FLAG_PROFILE 2 | 560 | #define TP_FLAG_PROFILE 2 |
| 561 | #define TP_FLAG_REGISTERED 4 | ||
| 539 | 562 | ||
| 540 | struct trace_probe { | 563 | struct trace_probe { |
| 541 | struct list_head list; | 564 | struct list_head list; |
| @@ -555,16 +578,49 @@ struct trace_probe { | |||
| 555 | (sizeof(struct probe_arg) * (n))) | 578 | (sizeof(struct probe_arg) * (n))) |
| 556 | 579 | ||
| 557 | 580 | ||
| 558 | static __kprobes int probe_is_return(struct trace_probe *tp) | 581 | static __kprobes int trace_probe_is_return(struct trace_probe *tp) |
| 559 | { | 582 | { |
| 560 | return tp->rp.handler != NULL; | 583 | return tp->rp.handler != NULL; |
| 561 | } | 584 | } |
| 562 | 585 | ||
| 563 | static __kprobes const char *probe_symbol(struct trace_probe *tp) | 586 | static __kprobes const char *trace_probe_symbol(struct trace_probe *tp) |
| 564 | { | 587 | { |
| 565 | return tp->symbol ? tp->symbol : "unknown"; | 588 | return tp->symbol ? tp->symbol : "unknown"; |
| 566 | } | 589 | } |
| 567 | 590 | ||
| 591 | static __kprobes unsigned long trace_probe_offset(struct trace_probe *tp) | ||
| 592 | { | ||
| 593 | return tp->rp.kp.offset; | ||
| 594 | } | ||
| 595 | |||
| 596 | static __kprobes bool trace_probe_is_enabled(struct trace_probe *tp) | ||
| 597 | { | ||
| 598 | return !!(tp->flags & (TP_FLAG_TRACE | TP_FLAG_PROFILE)); | ||
| 599 | } | ||
| 600 | |||
| 601 | static __kprobes bool trace_probe_is_registered(struct trace_probe *tp) | ||
| 602 | { | ||
| 603 | return !!(tp->flags & TP_FLAG_REGISTERED); | ||
| 604 | } | ||
| 605 | |||
| 606 | static __kprobes bool trace_probe_has_gone(struct trace_probe *tp) | ||
| 607 | { | ||
| 608 | return !!(kprobe_gone(&tp->rp.kp)); | ||
| 609 | } | ||
| 610 | |||
| 611 | static __kprobes bool trace_probe_within_module(struct trace_probe *tp, | ||
| 612 | struct module *mod) | ||
| 613 | { | ||
| 614 | int len = strlen(mod->name); | ||
| 615 | const char *name = trace_probe_symbol(tp); | ||
| 616 | return strncmp(mod->name, name, len) == 0 && name[len] == ':'; | ||
| 617 | } | ||
| 618 | |||
| 619 | static __kprobes bool trace_probe_is_on_module(struct trace_probe *tp) | ||
| 620 | { | ||
| 621 | return !!strchr(trace_probe_symbol(tp), ':'); | ||
| 622 | } | ||
| 623 | |||
| 568 | static int register_probe_event(struct trace_probe *tp); | 624 | static int register_probe_event(struct trace_probe *tp); |
| 569 | static void unregister_probe_event(struct trace_probe *tp); | 625 | static void unregister_probe_event(struct trace_probe *tp); |
| 570 | 626 | ||
| @@ -646,6 +702,16 @@ error: | |||
| 646 | return ERR_PTR(ret); | 702 | return ERR_PTR(ret); |
| 647 | } | 703 | } |
| 648 | 704 | ||
| 705 | static void update_probe_arg(struct probe_arg *arg) | ||
| 706 | { | ||
| 707 | if (CHECK_FETCH_FUNCS(bitfield, arg->fetch.fn)) | ||
| 708 | update_bitfield_fetch_param(arg->fetch.data); | ||
| 709 | else if (CHECK_FETCH_FUNCS(deref, arg->fetch.fn)) | ||
| 710 | update_deref_fetch_param(arg->fetch.data); | ||
| 711 | else if (CHECK_FETCH_FUNCS(symbol, arg->fetch.fn)) | ||
| 712 | update_symbol_cache(arg->fetch.data); | ||
| 713 | } | ||
| 714 | |||
| 649 | static void free_probe_arg(struct probe_arg *arg) | 715 | static void free_probe_arg(struct probe_arg *arg) |
| 650 | { | 716 | { |
| 651 | if (CHECK_FETCH_FUNCS(bitfield, arg->fetch.fn)) | 717 | if (CHECK_FETCH_FUNCS(bitfield, arg->fetch.fn)) |
| @@ -671,7 +737,7 @@ static void free_trace_probe(struct trace_probe *tp) | |||
| 671 | kfree(tp); | 737 | kfree(tp); |
| 672 | } | 738 | } |
| 673 | 739 | ||
| 674 | static struct trace_probe *find_probe_event(const char *event, | 740 | static struct trace_probe *find_trace_probe(const char *event, |
| 675 | const char *group) | 741 | const char *group) |
| 676 | { | 742 | { |
| 677 | struct trace_probe *tp; | 743 | struct trace_probe *tp; |
| @@ -683,13 +749,96 @@ static struct trace_probe *find_probe_event(const char *event, | |||
| 683 | return NULL; | 749 | return NULL; |
| 684 | } | 750 | } |
| 685 | 751 | ||
| 752 | /* Enable trace_probe - @flag must be TP_FLAG_TRACE or TP_FLAG_PROFILE */ | ||
| 753 | static int enable_trace_probe(struct trace_probe *tp, int flag) | ||
| 754 | { | ||
| 755 | int ret = 0; | ||
| 756 | |||
| 757 | tp->flags |= flag; | ||
| 758 | if (trace_probe_is_enabled(tp) && trace_probe_is_registered(tp) && | ||
| 759 | !trace_probe_has_gone(tp)) { | ||
| 760 | if (trace_probe_is_return(tp)) | ||
| 761 | ret = enable_kretprobe(&tp->rp); | ||
| 762 | else | ||
| 763 | ret = enable_kprobe(&tp->rp.kp); | ||
| 764 | } | ||
| 765 | |||
| 766 | return ret; | ||
| 767 | } | ||
| 768 | |||
| 769 | /* Disable trace_probe - @flag must be TP_FLAG_TRACE or TP_FLAG_PROFILE */ | ||
| 770 | static void disable_trace_probe(struct trace_probe *tp, int flag) | ||
| 771 | { | ||
| 772 | tp->flags &= ~flag; | ||
| 773 | if (!trace_probe_is_enabled(tp) && trace_probe_is_registered(tp)) { | ||
| 774 | if (trace_probe_is_return(tp)) | ||
| 775 | disable_kretprobe(&tp->rp); | ||
| 776 | else | ||
| 777 | disable_kprobe(&tp->rp.kp); | ||
| 778 | } | ||
| 779 | } | ||
| 780 | |||
| 781 | /* Internal register function - just handle k*probes and flags */ | ||
| 782 | static int __register_trace_probe(struct trace_probe *tp) | ||
| 783 | { | ||
| 784 | int i, ret; | ||
| 785 | |||
| 786 | if (trace_probe_is_registered(tp)) | ||
| 787 | return -EINVAL; | ||
| 788 | |||
| 789 | for (i = 0; i < tp->nr_args; i++) | ||
| 790 | update_probe_arg(&tp->args[i]); | ||
| 791 | |||
| 792 | /* Set/clear disabled flag according to tp->flag */ | ||
| 793 | if (trace_probe_is_enabled(tp)) | ||
| 794 | tp->rp.kp.flags &= ~KPROBE_FLAG_DISABLED; | ||
| 795 | else | ||
| 796 | tp->rp.kp.flags |= KPROBE_FLAG_DISABLED; | ||
| 797 | |||
| 798 | if (trace_probe_is_return(tp)) | ||
| 799 | ret = register_kretprobe(&tp->rp); | ||
| 800 | else | ||
| 801 | ret = register_kprobe(&tp->rp.kp); | ||
| 802 | |||
| 803 | if (ret == 0) | ||
| 804 | tp->flags |= TP_FLAG_REGISTERED; | ||
| 805 | else { | ||
| 806 | pr_warning("Could not insert probe at %s+%lu: %d\n", | ||
| 807 | trace_probe_symbol(tp), trace_probe_offset(tp), ret); | ||
| 808 | if (ret == -ENOENT && trace_probe_is_on_module(tp)) { | ||
| 809 | pr_warning("This probe might be able to register after" | ||
| 810 | "target module is loaded. Continue.\n"); | ||
| 811 | ret = 0; | ||
| 812 | } else if (ret == -EILSEQ) { | ||
| 813 | pr_warning("Probing address(0x%p) is not an " | ||
| 814 | "instruction boundary.\n", | ||
| 815 | tp->rp.kp.addr); | ||
| 816 | ret = -EINVAL; | ||
| 817 | } | ||
| 818 | } | ||
| 819 | |||
| 820 | return ret; | ||
| 821 | } | ||
| 822 | |||
| 823 | /* Internal unregister function - just handle k*probes and flags */ | ||
| 824 | static void __unregister_trace_probe(struct trace_probe *tp) | ||
| 825 | { | ||
| 826 | if (trace_probe_is_registered(tp)) { | ||
| 827 | if (trace_probe_is_return(tp)) | ||
| 828 | unregister_kretprobe(&tp->rp); | ||
| 829 | else | ||
| 830 | unregister_kprobe(&tp->rp.kp); | ||
| 831 | tp->flags &= ~TP_FLAG_REGISTERED; | ||
| 832 | /* Cleanup kprobe for reuse */ | ||
| 833 | if (tp->rp.kp.symbol_name) | ||
| 834 | tp->rp.kp.addr = NULL; | ||
| 835 | } | ||
| 836 | } | ||
| 837 | |||
| 686 | /* Unregister a trace_probe and probe_event: call with locking probe_lock */ | 838 | /* Unregister a trace_probe and probe_event: call with locking probe_lock */ |
| 687 | static void unregister_trace_probe(struct trace_probe *tp) | 839 | static void unregister_trace_probe(struct trace_probe *tp) |
| 688 | { | 840 | { |
| 689 | if (probe_is_return(tp)) | 841 | __unregister_trace_probe(tp); |
| 690 | unregister_kretprobe(&tp->rp); | ||
| 691 | else | ||
| 692 | unregister_kprobe(&tp->rp.kp); | ||
| 693 | list_del(&tp->list); | 842 | list_del(&tp->list); |
| 694 | unregister_probe_event(tp); | 843 | unregister_probe_event(tp); |
| 695 | } | 844 | } |
| @@ -702,41 +851,65 @@ static int register_trace_probe(struct trace_probe *tp) | |||
| 702 | 851 | ||
| 703 | mutex_lock(&probe_lock); | 852 | mutex_lock(&probe_lock); |
| 704 | 853 | ||
| 705 | /* register as an event */ | 854 | /* Delete old (same name) event if exist */ |
| 706 | old_tp = find_probe_event(tp->call.name, tp->call.class->system); | 855 | old_tp = find_trace_probe(tp->call.name, tp->call.class->system); |
| 707 | if (old_tp) { | 856 | if (old_tp) { |
| 708 | /* delete old event */ | ||
| 709 | unregister_trace_probe(old_tp); | 857 | unregister_trace_probe(old_tp); |
| 710 | free_trace_probe(old_tp); | 858 | free_trace_probe(old_tp); |
| 711 | } | 859 | } |
| 860 | |||
| 861 | /* Register new event */ | ||
| 712 | ret = register_probe_event(tp); | 862 | ret = register_probe_event(tp); |
| 713 | if (ret) { | 863 | if (ret) { |
| 714 | pr_warning("Failed to register probe event(%d)\n", ret); | 864 | pr_warning("Failed to register probe event(%d)\n", ret); |
| 715 | goto end; | 865 | goto end; |
| 716 | } | 866 | } |
| 717 | 867 | ||
| 718 | tp->rp.kp.flags |= KPROBE_FLAG_DISABLED; | 868 | /* Register k*probe */ |
| 719 | if (probe_is_return(tp)) | 869 | ret = __register_trace_probe(tp); |
| 720 | ret = register_kretprobe(&tp->rp); | 870 | if (ret < 0) |
| 721 | else | ||
| 722 | ret = register_kprobe(&tp->rp.kp); | ||
| 723 | |||
| 724 | if (ret) { | ||
| 725 | pr_warning("Could not insert probe(%d)\n", ret); | ||
| 726 | if (ret == -EILSEQ) { | ||
| 727 | pr_warning("Probing address(0x%p) is not an " | ||
| 728 | "instruction boundary.\n", | ||
| 729 | tp->rp.kp.addr); | ||
| 730 | ret = -EINVAL; | ||
| 731 | } | ||
| 732 | unregister_probe_event(tp); | 871 | unregister_probe_event(tp); |
| 733 | } else | 872 | else |
| 734 | list_add_tail(&tp->list, &probe_list); | 873 | list_add_tail(&tp->list, &probe_list); |
| 874 | |||
| 735 | end: | 875 | end: |
| 736 | mutex_unlock(&probe_lock); | 876 | mutex_unlock(&probe_lock); |
| 737 | return ret; | 877 | return ret; |
| 738 | } | 878 | } |
| 739 | 879 | ||
| 880 | /* Module notifier call back, checking event on the module */ | ||
| 881 | static int trace_probe_module_callback(struct notifier_block *nb, | ||
| 882 | unsigned long val, void *data) | ||
| 883 | { | ||
| 884 | struct module *mod = data; | ||
| 885 | struct trace_probe *tp; | ||
| 886 | int ret; | ||
| 887 | |||
| 888 | if (val != MODULE_STATE_COMING) | ||
| 889 | return NOTIFY_DONE; | ||
| 890 | |||
| 891 | /* Update probes on coming module */ | ||
| 892 | mutex_lock(&probe_lock); | ||
| 893 | list_for_each_entry(tp, &probe_list, list) { | ||
| 894 | if (trace_probe_within_module(tp, mod)) { | ||
| 895 | __unregister_trace_probe(tp); | ||
| 896 | ret = __register_trace_probe(tp); | ||
| 897 | if (ret) | ||
| 898 | pr_warning("Failed to re-register probe %s on" | ||
| 899 | "%s: %d\n", | ||
| 900 | tp->call.name, mod->name, ret); | ||
| 901 | } | ||
| 902 | } | ||
| 903 | mutex_unlock(&probe_lock); | ||
| 904 | |||
| 905 | return NOTIFY_DONE; | ||
| 906 | } | ||
| 907 | |||
| 908 | static struct notifier_block trace_probe_module_nb = { | ||
| 909 | .notifier_call = trace_probe_module_callback, | ||
| 910 | .priority = 1 /* Invoked after kprobe module callback */ | ||
| 911 | }; | ||
| 912 | |||
| 740 | /* Split symbol and offset. */ | 913 | /* Split symbol and offset. */ |
| 741 | static int split_symbol_offset(char *symbol, unsigned long *offset) | 914 | static int split_symbol_offset(char *symbol, unsigned long *offset) |
| 742 | { | 915 | { |
| @@ -962,8 +1135,8 @@ static int create_trace_probe(int argc, char **argv) | |||
| 962 | { | 1135 | { |
| 963 | /* | 1136 | /* |
| 964 | * Argument syntax: | 1137 | * Argument syntax: |
| 965 | * - Add kprobe: p[:[GRP/]EVENT] KSYM[+OFFS]|KADDR [FETCHARGS] | 1138 | * - Add kprobe: p[:[GRP/]EVENT] [MOD:]KSYM[+OFFS]|KADDR [FETCHARGS] |
| 966 | * - Add kretprobe: r[:[GRP/]EVENT] KSYM[+0] [FETCHARGS] | 1139 | * - Add kretprobe: r[:[GRP/]EVENT] [MOD:]KSYM[+0] [FETCHARGS] |
| 967 | * Fetch args: | 1140 | * Fetch args: |
| 968 | * $retval : fetch return value | 1141 | * $retval : fetch return value |
| 969 | * $stack : fetch stack address | 1142 | * $stack : fetch stack address |
| @@ -1025,7 +1198,7 @@ static int create_trace_probe(int argc, char **argv) | |||
| 1025 | return -EINVAL; | 1198 | return -EINVAL; |
| 1026 | } | 1199 | } |
| 1027 | mutex_lock(&probe_lock); | 1200 | mutex_lock(&probe_lock); |
| 1028 | tp = find_probe_event(event, group); | 1201 | tp = find_trace_probe(event, group); |
| 1029 | if (!tp) { | 1202 | if (!tp) { |
| 1030 | mutex_unlock(&probe_lock); | 1203 | mutex_unlock(&probe_lock); |
| 1031 | pr_info("Event %s/%s doesn't exist.\n", group, event); | 1204 | pr_info("Event %s/%s doesn't exist.\n", group, event); |
| @@ -1144,7 +1317,7 @@ error: | |||
| 1144 | return ret; | 1317 | return ret; |
| 1145 | } | 1318 | } |
| 1146 | 1319 | ||
| 1147 | static void cleanup_all_probes(void) | 1320 | static void release_all_trace_probes(void) |
| 1148 | { | 1321 | { |
| 1149 | struct trace_probe *tp; | 1322 | struct trace_probe *tp; |
| 1150 | 1323 | ||
| @@ -1158,7 +1331,6 @@ static void cleanup_all_probes(void) | |||
| 1158 | mutex_unlock(&probe_lock); | 1331 | mutex_unlock(&probe_lock); |
| 1159 | } | 1332 | } |
| 1160 | 1333 | ||
| 1161 | |||
| 1162 | /* Probes listing interfaces */ | 1334 | /* Probes listing interfaces */ |
| 1163 | static void *probes_seq_start(struct seq_file *m, loff_t *pos) | 1335 | static void *probes_seq_start(struct seq_file *m, loff_t *pos) |
| 1164 | { | 1336 | { |
| @@ -1181,15 +1353,16 @@ static int probes_seq_show(struct seq_file *m, void *v) | |||
| 1181 | struct trace_probe *tp = v; | 1353 | struct trace_probe *tp = v; |
| 1182 | int i; | 1354 | int i; |
| 1183 | 1355 | ||
| 1184 | seq_printf(m, "%c", probe_is_return(tp) ? 'r' : 'p'); | 1356 | seq_printf(m, "%c", trace_probe_is_return(tp) ? 'r' : 'p'); |
| 1185 | seq_printf(m, ":%s/%s", tp->call.class->system, tp->call.name); | 1357 | seq_printf(m, ":%s/%s", tp->call.class->system, tp->call.name); |
| 1186 | 1358 | ||
| 1187 | if (!tp->symbol) | 1359 | if (!tp->symbol) |
| 1188 | seq_printf(m, " 0x%p", tp->rp.kp.addr); | 1360 | seq_printf(m, " 0x%p", tp->rp.kp.addr); |
| 1189 | else if (tp->rp.kp.offset) | 1361 | else if (tp->rp.kp.offset) |
| 1190 | seq_printf(m, " %s+%u", probe_symbol(tp), tp->rp.kp.offset); | 1362 | seq_printf(m, " %s+%u", trace_probe_symbol(tp), |
| 1363 | tp->rp.kp.offset); | ||
| 1191 | else | 1364 | else |
| 1192 | seq_printf(m, " %s", probe_symbol(tp)); | 1365 | seq_printf(m, " %s", trace_probe_symbol(tp)); |
| 1193 | 1366 | ||
| 1194 | for (i = 0; i < tp->nr_args; i++) | 1367 | for (i = 0; i < tp->nr_args; i++) |
| 1195 | seq_printf(m, " %s=%s", tp->args[i].name, tp->args[i].comm); | 1368 | seq_printf(m, " %s=%s", tp->args[i].name, tp->args[i].comm); |
| @@ -1209,7 +1382,7 @@ static int probes_open(struct inode *inode, struct file *file) | |||
| 1209 | { | 1382 | { |
| 1210 | if ((file->f_mode & FMODE_WRITE) && | 1383 | if ((file->f_mode & FMODE_WRITE) && |
| 1211 | (file->f_flags & O_TRUNC)) | 1384 | (file->f_flags & O_TRUNC)) |
| 1212 | cleanup_all_probes(); | 1385 | release_all_trace_probes(); |
| 1213 | 1386 | ||
| 1214 | return seq_open(file, &probes_seq_op); | 1387 | return seq_open(file, &probes_seq_op); |
| 1215 | } | 1388 | } |
| @@ -1397,7 +1570,8 @@ static __kprobes void kprobe_trace_func(struct kprobe *kp, struct pt_regs *regs) | |||
| 1397 | store_trace_args(sizeof(*entry), tp, regs, (u8 *)&entry[1], dsize); | 1570 | store_trace_args(sizeof(*entry), tp, regs, (u8 *)&entry[1], dsize); |
| 1398 | 1571 | ||
| 1399 | if (!filter_current_check_discard(buffer, call, entry, event)) | 1572 | if (!filter_current_check_discard(buffer, call, entry, event)) |
| 1400 | trace_nowake_buffer_unlock_commit(buffer, event, irq_flags, pc); | 1573 | trace_nowake_buffer_unlock_commit_regs(buffer, event, |
| 1574 | irq_flags, pc, regs); | ||
| 1401 | } | 1575 | } |
| 1402 | 1576 | ||
| 1403 | /* Kretprobe handler */ | 1577 | /* Kretprobe handler */ |
| @@ -1429,7 +1603,8 @@ static __kprobes void kretprobe_trace_func(struct kretprobe_instance *ri, | |||
| 1429 | store_trace_args(sizeof(*entry), tp, regs, (u8 *)&entry[1], dsize); | 1603 | store_trace_args(sizeof(*entry), tp, regs, (u8 *)&entry[1], dsize); |
| 1430 | 1604 | ||
| 1431 | if (!filter_current_check_discard(buffer, call, entry, event)) | 1605 | if (!filter_current_check_discard(buffer, call, entry, event)) |
| 1432 | trace_nowake_buffer_unlock_commit(buffer, event, irq_flags, pc); | 1606 | trace_nowake_buffer_unlock_commit_regs(buffer, event, |
| 1607 | irq_flags, pc, regs); | ||
| 1433 | } | 1608 | } |
| 1434 | 1609 | ||
| 1435 | /* Event entry printers */ | 1610 | /* Event entry printers */ |
| @@ -1511,30 +1686,6 @@ partial: | |||
| 1511 | return TRACE_TYPE_PARTIAL_LINE; | 1686 | return TRACE_TYPE_PARTIAL_LINE; |
| 1512 | } | 1687 | } |
| 1513 | 1688 | ||
| 1514 | static int probe_event_enable(struct ftrace_event_call *call) | ||
| 1515 | { | ||
| 1516 | struct trace_probe *tp = (struct trace_probe *)call->data; | ||
| 1517 | |||
| 1518 | tp->flags |= TP_FLAG_TRACE; | ||
| 1519 | if (probe_is_return(tp)) | ||
| 1520 | return enable_kretprobe(&tp->rp); | ||
| 1521 | else | ||
| 1522 | return enable_kprobe(&tp->rp.kp); | ||
| 1523 | } | ||
| 1524 | |||
| 1525 | static void probe_event_disable(struct ftrace_event_call *call) | ||
| 1526 | { | ||
| 1527 | struct trace_probe *tp = (struct trace_probe *)call->data; | ||
| 1528 | |||
| 1529 | tp->flags &= ~TP_FLAG_TRACE; | ||
| 1530 | if (!(tp->flags & (TP_FLAG_TRACE | TP_FLAG_PROFILE))) { | ||
| 1531 | if (probe_is_return(tp)) | ||
| 1532 | disable_kretprobe(&tp->rp); | ||
| 1533 | else | ||
| 1534 | disable_kprobe(&tp->rp.kp); | ||
| 1535 | } | ||
| 1536 | } | ||
| 1537 | |||
| 1538 | #undef DEFINE_FIELD | 1689 | #undef DEFINE_FIELD |
| 1539 | #define DEFINE_FIELD(type, item, name, is_signed) \ | 1690 | #define DEFINE_FIELD(type, item, name, is_signed) \ |
| 1540 | do { \ | 1691 | do { \ |
| @@ -1596,7 +1747,7 @@ static int __set_print_fmt(struct trace_probe *tp, char *buf, int len) | |||
| 1596 | 1747 | ||
| 1597 | const char *fmt, *arg; | 1748 | const char *fmt, *arg; |
| 1598 | 1749 | ||
| 1599 | if (!probe_is_return(tp)) { | 1750 | if (!trace_probe_is_return(tp)) { |
| 1600 | fmt = "(%lx)"; | 1751 | fmt = "(%lx)"; |
| 1601 | arg = "REC->" FIELD_STRING_IP; | 1752 | arg = "REC->" FIELD_STRING_IP; |
| 1602 | } else { | 1753 | } else { |
| @@ -1713,49 +1864,25 @@ static __kprobes void kretprobe_perf_func(struct kretprobe_instance *ri, | |||
| 1713 | head = this_cpu_ptr(call->perf_events); | 1864 | head = this_cpu_ptr(call->perf_events); |
| 1714 | perf_trace_buf_submit(entry, size, rctx, entry->ret_ip, 1, regs, head); | 1865 | perf_trace_buf_submit(entry, size, rctx, entry->ret_ip, 1, regs, head); |
| 1715 | } | 1866 | } |
| 1716 | |||
| 1717 | static int probe_perf_enable(struct ftrace_event_call *call) | ||
| 1718 | { | ||
| 1719 | struct trace_probe *tp = (struct trace_probe *)call->data; | ||
| 1720 | |||
| 1721 | tp->flags |= TP_FLAG_PROFILE; | ||
| 1722 | |||
| 1723 | if (probe_is_return(tp)) | ||
| 1724 | return enable_kretprobe(&tp->rp); | ||
| 1725 | else | ||
| 1726 | return enable_kprobe(&tp->rp.kp); | ||
| 1727 | } | ||
| 1728 | |||
| 1729 | static void probe_perf_disable(struct ftrace_event_call *call) | ||
| 1730 | { | ||
| 1731 | struct trace_probe *tp = (struct trace_probe *)call->data; | ||
| 1732 | |||
| 1733 | tp->flags &= ~TP_FLAG_PROFILE; | ||
| 1734 | |||
| 1735 | if (!(tp->flags & TP_FLAG_TRACE)) { | ||
| 1736 | if (probe_is_return(tp)) | ||
| 1737 | disable_kretprobe(&tp->rp); | ||
| 1738 | else | ||
| 1739 | disable_kprobe(&tp->rp.kp); | ||
| 1740 | } | ||
| 1741 | } | ||
| 1742 | #endif /* CONFIG_PERF_EVENTS */ | 1867 | #endif /* CONFIG_PERF_EVENTS */ |
| 1743 | 1868 | ||
| 1744 | static __kprobes | 1869 | static __kprobes |
| 1745 | int kprobe_register(struct ftrace_event_call *event, enum trace_reg type) | 1870 | int kprobe_register(struct ftrace_event_call *event, enum trace_reg type) |
| 1746 | { | 1871 | { |
| 1872 | struct trace_probe *tp = (struct trace_probe *)event->data; | ||
| 1873 | |||
| 1747 | switch (type) { | 1874 | switch (type) { |
| 1748 | case TRACE_REG_REGISTER: | 1875 | case TRACE_REG_REGISTER: |
| 1749 | return probe_event_enable(event); | 1876 | return enable_trace_probe(tp, TP_FLAG_TRACE); |
| 1750 | case TRACE_REG_UNREGISTER: | 1877 | case TRACE_REG_UNREGISTER: |
| 1751 | probe_event_disable(event); | 1878 | disable_trace_probe(tp, TP_FLAG_TRACE); |
| 1752 | return 0; | 1879 | return 0; |
| 1753 | 1880 | ||
| 1754 | #ifdef CONFIG_PERF_EVENTS | 1881 | #ifdef CONFIG_PERF_EVENTS |
| 1755 | case TRACE_REG_PERF_REGISTER: | 1882 | case TRACE_REG_PERF_REGISTER: |
| 1756 | return probe_perf_enable(event); | 1883 | return enable_trace_probe(tp, TP_FLAG_PROFILE); |
| 1757 | case TRACE_REG_PERF_UNREGISTER: | 1884 | case TRACE_REG_PERF_UNREGISTER: |
| 1758 | probe_perf_disable(event); | 1885 | disable_trace_probe(tp, TP_FLAG_PROFILE); |
| 1759 | return 0; | 1886 | return 0; |
| 1760 | #endif | 1887 | #endif |
| 1761 | } | 1888 | } |
| @@ -1805,7 +1932,7 @@ static int register_probe_event(struct trace_probe *tp) | |||
| 1805 | 1932 | ||
| 1806 | /* Initialize ftrace_event_call */ | 1933 | /* Initialize ftrace_event_call */ |
| 1807 | INIT_LIST_HEAD(&call->class->fields); | 1934 | INIT_LIST_HEAD(&call->class->fields); |
| 1808 | if (probe_is_return(tp)) { | 1935 | if (trace_probe_is_return(tp)) { |
| 1809 | call->event.funcs = &kretprobe_funcs; | 1936 | call->event.funcs = &kretprobe_funcs; |
| 1810 | call->class->define_fields = kretprobe_event_define_fields; | 1937 | call->class->define_fields = kretprobe_event_define_fields; |
| 1811 | } else { | 1938 | } else { |
| @@ -1844,6 +1971,9 @@ static __init int init_kprobe_trace(void) | |||
| 1844 | struct dentry *d_tracer; | 1971 | struct dentry *d_tracer; |
| 1845 | struct dentry *entry; | 1972 | struct dentry *entry; |
| 1846 | 1973 | ||
| 1974 | if (register_module_notifier(&trace_probe_module_nb)) | ||
| 1975 | return -EINVAL; | ||
| 1976 | |||
| 1847 | d_tracer = tracing_init_dentry(); | 1977 | d_tracer = tracing_init_dentry(); |
| 1848 | if (!d_tracer) | 1978 | if (!d_tracer) |
| 1849 | return 0; | 1979 | return 0; |
| @@ -1897,12 +2027,12 @@ static __init int kprobe_trace_self_tests_init(void) | |||
| 1897 | warn++; | 2027 | warn++; |
| 1898 | } else { | 2028 | } else { |
| 1899 | /* Enable trace point */ | 2029 | /* Enable trace point */ |
| 1900 | tp = find_probe_event("testprobe", KPROBE_EVENT_SYSTEM); | 2030 | tp = find_trace_probe("testprobe", KPROBE_EVENT_SYSTEM); |
| 1901 | if (WARN_ON_ONCE(tp == NULL)) { | 2031 | if (WARN_ON_ONCE(tp == NULL)) { |
| 1902 | pr_warning("error on getting new probe.\n"); | 2032 | pr_warning("error on getting new probe.\n"); |
| 1903 | warn++; | 2033 | warn++; |
| 1904 | } else | 2034 | } else |
| 1905 | probe_event_enable(&tp->call); | 2035 | enable_trace_probe(tp, TP_FLAG_TRACE); |
| 1906 | } | 2036 | } |
| 1907 | 2037 | ||
| 1908 | ret = command_trace_probe("r:testprobe2 kprobe_trace_selftest_target " | 2038 | ret = command_trace_probe("r:testprobe2 kprobe_trace_selftest_target " |
| @@ -1912,12 +2042,12 @@ static __init int kprobe_trace_self_tests_init(void) | |||
| 1912 | warn++; | 2042 | warn++; |
| 1913 | } else { | 2043 | } else { |
| 1914 | /* Enable trace point */ | 2044 | /* Enable trace point */ |
| 1915 | tp = find_probe_event("testprobe2", KPROBE_EVENT_SYSTEM); | 2045 | tp = find_trace_probe("testprobe2", KPROBE_EVENT_SYSTEM); |
| 1916 | if (WARN_ON_ONCE(tp == NULL)) { | 2046 | if (WARN_ON_ONCE(tp == NULL)) { |
| 1917 | pr_warning("error on getting new probe.\n"); | 2047 | pr_warning("error on getting new probe.\n"); |
| 1918 | warn++; | 2048 | warn++; |
| 1919 | } else | 2049 | } else |
| 1920 | probe_event_enable(&tp->call); | 2050 | enable_trace_probe(tp, TP_FLAG_TRACE); |
| 1921 | } | 2051 | } |
| 1922 | 2052 | ||
| 1923 | if (warn) | 2053 | if (warn) |
| @@ -1938,7 +2068,7 @@ static __init int kprobe_trace_self_tests_init(void) | |||
| 1938 | } | 2068 | } |
| 1939 | 2069 | ||
| 1940 | end: | 2070 | end: |
| 1941 | cleanup_all_probes(); | 2071 | release_all_trace_probes(); |
| 1942 | if (warn) | 2072 | if (warn) |
| 1943 | pr_cont("NG: Some tests are failed. Please check them.\n"); | 2073 | pr_cont("NG: Some tests are failed. Please check them.\n"); |
| 1944 | else | 2074 | else |
diff --git a/kernel/trace/trace_output.c b/kernel/trace/trace_output.c index e37de492a9e..51999309a6c 100644 --- a/kernel/trace/trace_output.c +++ b/kernel/trace/trace_output.c | |||
| @@ -1107,19 +1107,20 @@ static enum print_line_t trace_stack_print(struct trace_iterator *iter, | |||
| 1107 | { | 1107 | { |
| 1108 | struct stack_entry *field; | 1108 | struct stack_entry *field; |
| 1109 | struct trace_seq *s = &iter->seq; | 1109 | struct trace_seq *s = &iter->seq; |
| 1110 | int i; | 1110 | unsigned long *p; |
| 1111 | unsigned long *end; | ||
| 1111 | 1112 | ||
| 1112 | trace_assign_type(field, iter->ent); | 1113 | trace_assign_type(field, iter->ent); |
| 1114 | end = (unsigned long *)((long)iter->ent + iter->ent_size); | ||
| 1113 | 1115 | ||
| 1114 | if (!trace_seq_puts(s, "<stack trace>\n")) | 1116 | if (!trace_seq_puts(s, "<stack trace>\n")) |
| 1115 | goto partial; | 1117 | goto partial; |
| 1116 | for (i = 0; i < FTRACE_STACK_ENTRIES; i++) { | 1118 | |
| 1117 | if (!field->caller[i] || (field->caller[i] == ULONG_MAX)) | 1119 | for (p = field->caller; p && *p != ULONG_MAX && p < end; p++) { |
| 1118 | break; | ||
| 1119 | if (!trace_seq_puts(s, " => ")) | 1120 | if (!trace_seq_puts(s, " => ")) |
| 1120 | goto partial; | 1121 | goto partial; |
| 1121 | 1122 | ||
| 1122 | if (!seq_print_ip_sym(s, field->caller[i], flags)) | 1123 | if (!seq_print_ip_sym(s, *p, flags)) |
| 1123 | goto partial; | 1124 | goto partial; |
| 1124 | if (!trace_seq_puts(s, "\n")) | 1125 | if (!trace_seq_puts(s, "\n")) |
| 1125 | goto partial; | 1126 | goto partial; |
diff --git a/kernel/trace/trace_sched_wakeup.c b/kernel/trace/trace_sched_wakeup.c index f029dd4fd2c..e4a70c0c71b 100644 --- a/kernel/trace/trace_sched_wakeup.c +++ b/kernel/trace/trace_sched_wakeup.c | |||
| @@ -227,7 +227,9 @@ static void wakeup_trace_close(struct trace_iterator *iter) | |||
| 227 | graph_trace_close(iter); | 227 | graph_trace_close(iter); |
| 228 | } | 228 | } |
| 229 | 229 | ||
| 230 | #define GRAPH_TRACER_FLAGS (TRACE_GRAPH_PRINT_PROC) | 230 | #define GRAPH_TRACER_FLAGS (TRACE_GRAPH_PRINT_PROC | \ |
| 231 | TRACE_GRAPH_PRINT_ABS_TIME | \ | ||
| 232 | TRACE_GRAPH_PRINT_DURATION) | ||
| 231 | 233 | ||
| 232 | static enum print_line_t wakeup_print_line(struct trace_iterator *iter) | 234 | static enum print_line_t wakeup_print_line(struct trace_iterator *iter) |
| 233 | { | 235 | { |
diff --git a/kernel/trace/trace_stack.c b/kernel/trace/trace_stack.c index b0b53b8e4c2..77575b386d9 100644 --- a/kernel/trace/trace_stack.c +++ b/kernel/trace/trace_stack.c | |||
| @@ -156,20 +156,11 @@ stack_max_size_write(struct file *filp, const char __user *ubuf, | |||
| 156 | { | 156 | { |
| 157 | long *ptr = filp->private_data; | 157 | long *ptr = filp->private_data; |
| 158 | unsigned long val, flags; | 158 | unsigned long val, flags; |
| 159 | char buf[64]; | ||
| 160 | int ret; | 159 | int ret; |
| 161 | int cpu; | 160 | int cpu; |
| 162 | 161 | ||
| 163 | if (count >= sizeof(buf)) | 162 | ret = kstrtoul_from_user(ubuf, count, 10, &val); |
| 164 | return -EINVAL; | 163 | if (ret) |
| 165 | |||
| 166 | if (copy_from_user(&buf, ubuf, count)) | ||
| 167 | return -EFAULT; | ||
| 168 | |||
| 169 | buf[count] = 0; | ||
| 170 | |||
| 171 | ret = strict_strtoul(buf, 10, &val); | ||
| 172 | if (ret < 0) | ||
| 173 | return ret; | 164 | return ret; |
| 174 | 165 | ||
| 175 | local_irq_save(flags); | 166 | local_irq_save(flags); |
diff --git a/kernel/watchdog.c b/kernel/watchdog.c index 3d0c56ad479..36491cd5b7d 100644 --- a/kernel/watchdog.c +++ b/kernel/watchdog.c | |||
| @@ -200,6 +200,7 @@ static int is_softlockup(unsigned long touch_ts) | |||
| 200 | } | 200 | } |
| 201 | 201 | ||
| 202 | #ifdef CONFIG_HARDLOCKUP_DETECTOR | 202 | #ifdef CONFIG_HARDLOCKUP_DETECTOR |
| 203 | |||
| 203 | static struct perf_event_attr wd_hw_attr = { | 204 | static struct perf_event_attr wd_hw_attr = { |
| 204 | .type = PERF_TYPE_HARDWARE, | 205 | .type = PERF_TYPE_HARDWARE, |
| 205 | .config = PERF_COUNT_HW_CPU_CYCLES, | 206 | .config = PERF_COUNT_HW_CPU_CYCLES, |
| @@ -209,7 +210,7 @@ static struct perf_event_attr wd_hw_attr = { | |||
| 209 | }; | 210 | }; |
| 210 | 211 | ||
| 211 | /* Callback function for perf event subsystem */ | 212 | /* Callback function for perf event subsystem */ |
| 212 | static void watchdog_overflow_callback(struct perf_event *event, int nmi, | 213 | static void watchdog_overflow_callback(struct perf_event *event, |
| 213 | struct perf_sample_data *data, | 214 | struct perf_sample_data *data, |
| 214 | struct pt_regs *regs) | 215 | struct pt_regs *regs) |
| 215 | { | 216 | { |
| @@ -368,10 +369,11 @@ static int watchdog_nmi_enable(int cpu) | |||
| 368 | if (event != NULL) | 369 | if (event != NULL) |
| 369 | goto out_enable; | 370 | goto out_enable; |
| 370 | 371 | ||
| 371 | /* Try to register using hardware perf events */ | ||
| 372 | wd_attr = &wd_hw_attr; | 372 | wd_attr = &wd_hw_attr; |
| 373 | wd_attr->sample_period = hw_nmi_get_sample_period(watchdog_thresh); | 373 | wd_attr->sample_period = hw_nmi_get_sample_period(watchdog_thresh); |
| 374 | event = perf_event_create_kernel_counter(wd_attr, cpu, NULL, watchdog_overflow_callback); | 374 | |
| 375 | /* Try to register using hardware perf events */ | ||
| 376 | event = perf_event_create_kernel_counter(wd_attr, cpu, NULL, watchdog_overflow_callback, NULL); | ||
| 375 | if (!IS_ERR(event)) { | 377 | if (!IS_ERR(event)) { |
| 376 | printk(KERN_INFO "NMI watchdog enabled, takes one hw-pmu counter.\n"); | 378 | printk(KERN_INFO "NMI watchdog enabled, takes one hw-pmu counter.\n"); |
| 377 | goto out_save; | 379 | goto out_save; |
diff --git a/samples/hw_breakpoint/data_breakpoint.c b/samples/hw_breakpoint/data_breakpoint.c index 063653955f9..ef7f3229185 100644 --- a/samples/hw_breakpoint/data_breakpoint.c +++ b/samples/hw_breakpoint/data_breakpoint.c | |||
| @@ -41,7 +41,7 @@ module_param_string(ksym, ksym_name, KSYM_NAME_LEN, S_IRUGO); | |||
| 41 | MODULE_PARM_DESC(ksym, "Kernel symbol to monitor; this module will report any" | 41 | MODULE_PARM_DESC(ksym, "Kernel symbol to monitor; this module will report any" |
| 42 | " write operations on the kernel symbol"); | 42 | " write operations on the kernel symbol"); |
| 43 | 43 | ||
| 44 | static void sample_hbp_handler(struct perf_event *bp, int nmi, | 44 | static void sample_hbp_handler(struct perf_event *bp, |
| 45 | struct perf_sample_data *data, | 45 | struct perf_sample_data *data, |
| 46 | struct pt_regs *regs) | 46 | struct pt_regs *regs) |
| 47 | { | 47 | { |
| @@ -60,7 +60,7 @@ static int __init hw_break_module_init(void) | |||
| 60 | attr.bp_len = HW_BREAKPOINT_LEN_4; | 60 | attr.bp_len = HW_BREAKPOINT_LEN_4; |
| 61 | attr.bp_type = HW_BREAKPOINT_W | HW_BREAKPOINT_R; | 61 | attr.bp_type = HW_BREAKPOINT_W | HW_BREAKPOINT_R; |
| 62 | 62 | ||
| 63 | sample_hbp = register_wide_hw_breakpoint(&attr, sample_hbp_handler); | 63 | sample_hbp = register_wide_hw_breakpoint(&attr, sample_hbp_handler, NULL); |
| 64 | if (IS_ERR((void __force *)sample_hbp)) { | 64 | if (IS_ERR((void __force *)sample_hbp)) { |
| 65 | ret = PTR_ERR((void __force *)sample_hbp); | 65 | ret = PTR_ERR((void __force *)sample_hbp); |
| 66 | goto fail; | 66 | goto fail; |
diff --git a/tools/perf/Documentation/perf-annotate.txt b/tools/perf/Documentation/perf-annotate.txt index 6f5a498608b..85c5f026930 100644 --- a/tools/perf/Documentation/perf-annotate.txt +++ b/tools/perf/Documentation/perf-annotate.txt | |||
| @@ -66,6 +66,12 @@ OPTIONS | |||
| 66 | used. This interfaces starts by centering on the line with more | 66 | used. This interfaces starts by centering on the line with more |
| 67 | samples, TAB/UNTAB cycles through the lines with more samples. | 67 | samples, TAB/UNTAB cycles through the lines with more samples. |
| 68 | 68 | ||
| 69 | -c:: | ||
| 70 | --cpu:: Only report samples for the list of CPUs provided. Multiple CPUs can | ||
| 71 | be provided as a comma-separated list with no space: 0,1. Ranges of | ||
| 72 | CPUs are specified with -: 0-2. Default is to report samples on all | ||
| 73 | CPUs. | ||
| 74 | |||
| 69 | SEE ALSO | 75 | SEE ALSO |
| 70 | -------- | 76 | -------- |
| 71 | linkperf:perf-record[1], linkperf:perf-report[1] | 77 | linkperf:perf-record[1], linkperf:perf-report[1] |
diff --git a/tools/perf/Documentation/perf-probe.txt b/tools/perf/Documentation/perf-probe.txt index 02bafce4b34..2780d9ce48b 100644 --- a/tools/perf/Documentation/perf-probe.txt +++ b/tools/perf/Documentation/perf-probe.txt | |||
| @@ -34,9 +34,11 @@ OPTIONS | |||
| 34 | Specify vmlinux path which has debuginfo (Dwarf binary). | 34 | Specify vmlinux path which has debuginfo (Dwarf binary). |
| 35 | 35 | ||
| 36 | -m:: | 36 | -m:: |
| 37 | --module=MODNAME:: | 37 | --module=MODNAME|PATH:: |
| 38 | Specify module name in which perf-probe searches probe points | 38 | Specify module name in which perf-probe searches probe points |
| 39 | or lines. | 39 | or lines. If a path of module file is passed, perf-probe |
| 40 | treat it as an offline module (this means you can add a probe on | ||
| 41 | a module which has not been loaded yet). | ||
| 40 | 42 | ||
| 41 | -s:: | 43 | -s:: |
| 42 | --source=PATH:: | 44 | --source=PATH:: |
diff --git a/tools/perf/Documentation/perf-report.txt b/tools/perf/Documentation/perf-report.txt index 8ba03d6e539..04253c07d19 100644 --- a/tools/perf/Documentation/perf-report.txt +++ b/tools/perf/Documentation/perf-report.txt | |||
| @@ -80,15 +80,24 @@ OPTIONS | |||
| 80 | --dump-raw-trace:: | 80 | --dump-raw-trace:: |
| 81 | Dump raw trace in ASCII. | 81 | Dump raw trace in ASCII. |
| 82 | 82 | ||
| 83 | -g [type,min]:: | 83 | -g [type,min,order]:: |
| 84 | --call-graph:: | 84 | --call-graph:: |
| 85 | Display call chains using type and min percent threshold. | 85 | Display call chains using type, min percent threshold and order. |
| 86 | type can be either: | 86 | type can be either: |
| 87 | - flat: single column, linear exposure of call chains. | 87 | - flat: single column, linear exposure of call chains. |
| 88 | - graph: use a graph tree, displaying absolute overhead rates. | 88 | - graph: use a graph tree, displaying absolute overhead rates. |
| 89 | - fractal: like graph, but displays relative rates. Each branch of | 89 | - fractal: like graph, but displays relative rates. Each branch of |
| 90 | the tree is considered as a new profiled object. + | 90 | the tree is considered as a new profiled object. + |
| 91 | Default: fractal,0.5. | 91 | |
| 92 | order can be either: | ||
| 93 | - callee: callee based call graph. | ||
| 94 | - caller: inverted caller based call graph. | ||
| 95 | |||
| 96 | Default: fractal,0.5,callee. | ||
| 97 | |||
| 98 | -G:: | ||
| 99 | --inverted:: | ||
| 100 | alias for inverted caller based call graph. | ||
| 92 | 101 | ||
| 93 | --pretty=<key>:: | 102 | --pretty=<key>:: |
| 94 | Pretty printing style. key: normal, raw | 103 | Pretty printing style. key: normal, raw |
| @@ -119,6 +128,12 @@ OPTIONS | |||
| 119 | --symfs=<directory>:: | 128 | --symfs=<directory>:: |
| 120 | Look for files with symbols relative to this directory. | 129 | Look for files with symbols relative to this directory. |
| 121 | 130 | ||
| 131 | -c:: | ||
| 132 | --cpu:: Only report samples for the list of CPUs provided. Multiple CPUs can | ||
| 133 | be provided as a comma-separated list with no space: 0,1. Ranges of | ||
| 134 | CPUs are specified with -: 0-2. Default is to report samples on all | ||
| 135 | CPUs. | ||
| 136 | |||
| 122 | SEE ALSO | 137 | SEE ALSO |
| 123 | -------- | 138 | -------- |
| 124 | linkperf:perf-stat[1] | 139 | linkperf:perf-stat[1] |
diff --git a/tools/perf/Documentation/perf-script.txt b/tools/perf/Documentation/perf-script.txt index 86c87e214b1..db017867d9e 100644 --- a/tools/perf/Documentation/perf-script.txt +++ b/tools/perf/Documentation/perf-script.txt | |||
| @@ -115,10 +115,10 @@ OPTIONS | |||
| 115 | -f:: | 115 | -f:: |
| 116 | --fields:: | 116 | --fields:: |
| 117 | Comma separated list of fields to print. Options are: | 117 | Comma separated list of fields to print. Options are: |
| 118 | comm, tid, pid, time, cpu, event, trace, sym. Field | 118 | comm, tid, pid, time, cpu, event, trace, ip, sym, dso, addr. |
| 119 | list can be prepended with the type, trace, sw or hw, | 119 | Field list can be prepended with the type, trace, sw or hw, |
| 120 | to indicate to which event type the field list applies. | 120 | to indicate to which event type the field list applies. |
| 121 | e.g., -f sw:comm,tid,time,sym and -f trace:time,cpu,trace | 121 | e.g., -f sw:comm,tid,time,ip,sym and -f trace:time,cpu,trace |
| 122 | 122 | ||
| 123 | perf script -f <fields> | 123 | perf script -f <fields> |
| 124 | 124 | ||
| @@ -132,17 +132,17 @@ OPTIONS | |||
| 132 | The arguments are processed in the order received. A later usage can | 132 | The arguments are processed in the order received. A later usage can |
| 133 | reset a prior request. e.g.: | 133 | reset a prior request. e.g.: |
| 134 | 134 | ||
| 135 | -f trace: -f comm,tid,time,sym | 135 | -f trace: -f comm,tid,time,ip,sym |
| 136 | 136 | ||
| 137 | The first -f suppresses trace events (field list is ""), but then the | 137 | The first -f suppresses trace events (field list is ""), but then the |
| 138 | second invocation sets the fields to comm,tid,time,sym. In this case a | 138 | second invocation sets the fields to comm,tid,time,ip,sym. In this case a |
| 139 | warning is given to the user: | 139 | warning is given to the user: |
| 140 | 140 | ||
| 141 | "Overriding previous field request for all events." | 141 | "Overriding previous field request for all events." |
| 142 | 142 | ||
| 143 | Alternativey, consider the order: | 143 | Alternativey, consider the order: |
| 144 | 144 | ||
| 145 | -f comm,tid,time,sym -f trace: | 145 | -f comm,tid,time,ip,sym -f trace: |
| 146 | 146 | ||
| 147 | The first -f sets the fields for all events and the second -f | 147 | The first -f sets the fields for all events and the second -f |
| 148 | suppresses trace events. The user is given a warning message about | 148 | suppresses trace events. The user is given a warning message about |
| @@ -182,6 +182,12 @@ OPTIONS | |||
| 182 | --hide-call-graph:: | 182 | --hide-call-graph:: |
| 183 | When printing symbols do not display call chain. | 183 | When printing symbols do not display call chain. |
| 184 | 184 | ||
| 185 | -c:: | ||
| 186 | --cpu:: Only report samples for the list of CPUs provided. Multiple CPUs can | ||
| 187 | be provided as a comma-separated list with no space: 0,1. Ranges of | ||
| 188 | CPUs are specified with -: 0-2. Default is to report samples on all | ||
| 189 | CPUs. | ||
| 190 | |||
| 185 | SEE ALSO | 191 | SEE ALSO |
| 186 | -------- | 192 | -------- |
| 187 | linkperf:perf-record[1], linkperf:perf-script-perl[1], | 193 | linkperf:perf-record[1], linkperf:perf-script-perl[1], |
diff --git a/tools/perf/Makefile b/tools/perf/Makefile index 940257b5774..d0861bbd1d9 100644 --- a/tools/perf/Makefile +++ b/tools/perf/Makefile | |||
| @@ -279,6 +279,7 @@ LIB_H += util/thread.h | |||
| 279 | LIB_H += util/thread_map.h | 279 | LIB_H += util/thread_map.h |
| 280 | LIB_H += util/trace-event.h | 280 | LIB_H += util/trace-event.h |
| 281 | LIB_H += util/probe-finder.h | 281 | LIB_H += util/probe-finder.h |
| 282 | LIB_H += util/dwarf-aux.h | ||
| 282 | LIB_H += util/probe-event.h | 283 | LIB_H += util/probe-event.h |
| 283 | LIB_H += util/pstack.h | 284 | LIB_H += util/pstack.h |
| 284 | LIB_H += util/cpumap.h | 285 | LIB_H += util/cpumap.h |
| @@ -435,6 +436,7 @@ else | |||
| 435 | BASIC_CFLAGS += -DDWARF_SUPPORT | 436 | BASIC_CFLAGS += -DDWARF_SUPPORT |
| 436 | EXTLIBS += -lelf -ldw | 437 | EXTLIBS += -lelf -ldw |
| 437 | LIB_OBJS += $(OUTPUT)util/probe-finder.o | 438 | LIB_OBJS += $(OUTPUT)util/probe-finder.o |
| 439 | LIB_OBJS += $(OUTPUT)util/dwarf-aux.o | ||
| 438 | endif # PERF_HAVE_DWARF_REGS | 440 | endif # PERF_HAVE_DWARF_REGS |
| 439 | endif # NO_DWARF | 441 | endif # NO_DWARF |
| 440 | 442 | ||
diff --git a/tools/perf/builtin-annotate.c b/tools/perf/builtin-annotate.c index 7b139e1e7e8..555aefd7fe0 100644 --- a/tools/perf/builtin-annotate.c +++ b/tools/perf/builtin-annotate.c | |||
| @@ -28,6 +28,8 @@ | |||
| 28 | #include "util/hist.h" | 28 | #include "util/hist.h" |
| 29 | #include "util/session.h" | 29 | #include "util/session.h" |
| 30 | 30 | ||
| 31 | #include <linux/bitmap.h> | ||
| 32 | |||
| 31 | static char const *input_name = "perf.data"; | 33 | static char const *input_name = "perf.data"; |
| 32 | 34 | ||
| 33 | static bool force, use_tui, use_stdio; | 35 | static bool force, use_tui, use_stdio; |
| @@ -38,6 +40,9 @@ static bool print_line; | |||
| 38 | 40 | ||
| 39 | static const char *sym_hist_filter; | 41 | static const char *sym_hist_filter; |
| 40 | 42 | ||
| 43 | static const char *cpu_list; | ||
| 44 | static DECLARE_BITMAP(cpu_bitmap, MAX_NR_CPUS); | ||
| 45 | |||
| 41 | static int perf_evlist__add_sample(struct perf_evlist *evlist, | 46 | static int perf_evlist__add_sample(struct perf_evlist *evlist, |
| 42 | struct perf_sample *sample, | 47 | struct perf_sample *sample, |
| 43 | struct perf_evsel *evsel, | 48 | struct perf_evsel *evsel, |
| @@ -90,6 +95,9 @@ static int process_sample_event(union perf_event *event, | |||
| 90 | return -1; | 95 | return -1; |
| 91 | } | 96 | } |
| 92 | 97 | ||
| 98 | if (cpu_list && !test_bit(sample->cpu, cpu_bitmap)) | ||
| 99 | return 0; | ||
| 100 | |||
| 93 | if (!al.filtered && | 101 | if (!al.filtered && |
| 94 | perf_evlist__add_sample(session->evlist, sample, evsel, &al)) { | 102 | perf_evlist__add_sample(session->evlist, sample, evsel, &al)) { |
| 95 | pr_warning("problem incrementing symbol count, " | 103 | pr_warning("problem incrementing symbol count, " |
| @@ -177,6 +185,12 @@ static int __cmd_annotate(void) | |||
| 177 | if (session == NULL) | 185 | if (session == NULL) |
| 178 | return -ENOMEM; | 186 | return -ENOMEM; |
| 179 | 187 | ||
| 188 | if (cpu_list) { | ||
| 189 | ret = perf_session__cpu_bitmap(session, cpu_list, cpu_bitmap); | ||
| 190 | if (ret) | ||
| 191 | goto out_delete; | ||
| 192 | } | ||
| 193 | |||
| 180 | ret = perf_session__process_events(session, &event_ops); | 194 | ret = perf_session__process_events(session, &event_ops); |
| 181 | if (ret) | 195 | if (ret) |
| 182 | goto out_delete; | 196 | goto out_delete; |
| @@ -252,6 +266,7 @@ static const struct option options[] = { | |||
| 252 | "print matching source lines (may be slow)"), | 266 | "print matching source lines (may be slow)"), |
| 253 | OPT_BOOLEAN('P', "full-paths", &full_paths, | 267 | OPT_BOOLEAN('P', "full-paths", &full_paths, |
| 254 | "Don't shorten the displayed pathnames"), | 268 | "Don't shorten the displayed pathnames"), |
| 269 | OPT_STRING('c', "cpu", &cpu_list, "cpu", "list of cpus to profile"), | ||
| 255 | OPT_END() | 270 | OPT_END() |
| 256 | }; | 271 | }; |
| 257 | 272 | ||
diff --git a/tools/perf/builtin-probe.c b/tools/perf/builtin-probe.c index 2c0e64d0b4a..5f2a5c7046d 100644 --- a/tools/perf/builtin-probe.c +++ b/tools/perf/builtin-probe.c | |||
| @@ -242,7 +242,8 @@ static const struct option options[] = { | |||
| 242 | OPT_STRING('s', "source", &symbol_conf.source_prefix, | 242 | OPT_STRING('s', "source", &symbol_conf.source_prefix, |
| 243 | "directory", "path to kernel source"), | 243 | "directory", "path to kernel source"), |
| 244 | OPT_STRING('m', "module", ¶ms.target_module, | 244 | OPT_STRING('m', "module", ¶ms.target_module, |
| 245 | "modname", "target module name"), | 245 | "modname|path", |
| 246 | "target module name (for online) or path (for offline)"), | ||
| 246 | #endif | 247 | #endif |
| 247 | OPT__DRY_RUN(&probe_event_dry_run), | 248 | OPT__DRY_RUN(&probe_event_dry_run), |
| 248 | OPT_INTEGER('\0', "max-probes", ¶ms.max_probe_points, | 249 | OPT_INTEGER('\0', "max-probes", ¶ms.max_probe_points, |
diff --git a/tools/perf/builtin-record.c b/tools/perf/builtin-record.c index 8e2c8579818..80dc5b790e4 100644 --- a/tools/perf/builtin-record.c +++ b/tools/perf/builtin-record.c | |||
| @@ -740,7 +740,7 @@ static bool force, append_file; | |||
| 740 | const struct option record_options[] = { | 740 | const struct option record_options[] = { |
| 741 | OPT_CALLBACK('e', "event", &evsel_list, "event", | 741 | OPT_CALLBACK('e', "event", &evsel_list, "event", |
| 742 | "event selector. use 'perf list' to list available events", | 742 | "event selector. use 'perf list' to list available events", |
| 743 | parse_events), | 743 | parse_events_option), |
| 744 | OPT_CALLBACK(0, "filter", &evsel_list, "filter", | 744 | OPT_CALLBACK(0, "filter", &evsel_list, "filter", |
| 745 | "event filter", parse_filter), | 745 | "event filter", parse_filter), |
| 746 | OPT_INTEGER('p', "pid", &target_pid, | 746 | OPT_INTEGER('p', "pid", &target_pid, |
diff --git a/tools/perf/builtin-report.c b/tools/perf/builtin-report.c index 287a173523a..f854efda768 100644 --- a/tools/perf/builtin-report.c +++ b/tools/perf/builtin-report.c | |||
| @@ -33,6 +33,8 @@ | |||
| 33 | #include "util/sort.h" | 33 | #include "util/sort.h" |
| 34 | #include "util/hist.h" | 34 | #include "util/hist.h" |
| 35 | 35 | ||
| 36 | #include <linux/bitmap.h> | ||
| 37 | |||
| 36 | static char const *input_name = "perf.data"; | 38 | static char const *input_name = "perf.data"; |
| 37 | 39 | ||
| 38 | static bool force, use_tui, use_stdio; | 40 | static bool force, use_tui, use_stdio; |
| @@ -45,9 +47,13 @@ static struct perf_read_values show_threads_values; | |||
| 45 | static const char default_pretty_printing_style[] = "normal"; | 47 | static const char default_pretty_printing_style[] = "normal"; |
| 46 | static const char *pretty_printing_style = default_pretty_printing_style; | 48 | static const char *pretty_printing_style = default_pretty_printing_style; |
| 47 | 49 | ||
| 48 | static char callchain_default_opt[] = "fractal,0.5"; | 50 | static char callchain_default_opt[] = "fractal,0.5,callee"; |
| 51 | static bool inverted_callchain; | ||
| 49 | static symbol_filter_t annotate_init; | 52 | static symbol_filter_t annotate_init; |
| 50 | 53 | ||
| 54 | static const char *cpu_list; | ||
| 55 | static DECLARE_BITMAP(cpu_bitmap, MAX_NR_CPUS); | ||
| 56 | |||
| 51 | static int perf_session__add_hist_entry(struct perf_session *session, | 57 | static int perf_session__add_hist_entry(struct perf_session *session, |
| 52 | struct addr_location *al, | 58 | struct addr_location *al, |
| 53 | struct perf_sample *sample, | 59 | struct perf_sample *sample, |
| @@ -116,6 +122,9 @@ static int process_sample_event(union perf_event *event, | |||
| 116 | if (al.filtered || (hide_unresolved && al.sym == NULL)) | 122 | if (al.filtered || (hide_unresolved && al.sym == NULL)) |
| 117 | return 0; | 123 | return 0; |
| 118 | 124 | ||
| 125 | if (cpu_list && !test_bit(sample->cpu, cpu_bitmap)) | ||
| 126 | return 0; | ||
| 127 | |||
| 119 | if (al.map != NULL) | 128 | if (al.map != NULL) |
| 120 | al.map->dso->hit = 1; | 129 | al.map->dso->hit = 1; |
| 121 | 130 | ||
| @@ -262,6 +271,12 @@ static int __cmd_report(void) | |||
| 262 | if (session == NULL) | 271 | if (session == NULL) |
| 263 | return -ENOMEM; | 272 | return -ENOMEM; |
| 264 | 273 | ||
| 274 | if (cpu_list) { | ||
| 275 | ret = perf_session__cpu_bitmap(session, cpu_list, cpu_bitmap); | ||
| 276 | if (ret) | ||
| 277 | goto out_delete; | ||
| 278 | } | ||
| 279 | |||
| 265 | if (show_threads) | 280 | if (show_threads) |
| 266 | perf_read_values_init(&show_threads_values); | 281 | perf_read_values_init(&show_threads_values); |
| 267 | 282 | ||
| @@ -386,13 +401,29 @@ parse_callchain_opt(const struct option *opt __used, const char *arg, | |||
| 386 | if (!tok) | 401 | if (!tok) |
| 387 | goto setup; | 402 | goto setup; |
| 388 | 403 | ||
| 389 | tok2 = strtok(NULL, ","); | ||
| 390 | callchain_param.min_percent = strtod(tok, &endptr); | 404 | callchain_param.min_percent = strtod(tok, &endptr); |
| 391 | if (tok == endptr) | 405 | if (tok == endptr) |
| 392 | return -1; | 406 | return -1; |
| 393 | 407 | ||
| 394 | if (tok2) | 408 | /* get the print limit */ |
| 409 | tok2 = strtok(NULL, ","); | ||
| 410 | if (!tok2) | ||
| 411 | goto setup; | ||
| 412 | |||
| 413 | if (tok2[0] != 'c') { | ||
| 395 | callchain_param.print_limit = strtod(tok2, &endptr); | 414 | callchain_param.print_limit = strtod(tok2, &endptr); |
| 415 | tok2 = strtok(NULL, ","); | ||
| 416 | if (!tok2) | ||
| 417 | goto setup; | ||
| 418 | } | ||
| 419 | |||
| 420 | /* get the call chain order */ | ||
| 421 | if (!strcmp(tok2, "caller")) | ||
| 422 | callchain_param.order = ORDER_CALLER; | ||
| 423 | else if (!strcmp(tok2, "callee")) | ||
| 424 | callchain_param.order = ORDER_CALLEE; | ||
| 425 | else | ||
| 426 | return -1; | ||
| 396 | setup: | 427 | setup: |
| 397 | if (callchain_register_param(&callchain_param) < 0) { | 428 | if (callchain_register_param(&callchain_param) < 0) { |
| 398 | fprintf(stderr, "Can't register callchain params\n"); | 429 | fprintf(stderr, "Can't register callchain params\n"); |
| @@ -436,9 +467,10 @@ static const struct option options[] = { | |||
| 436 | "regex filter to identify parent, see: '--sort parent'"), | 467 | "regex filter to identify parent, see: '--sort parent'"), |
| 437 | OPT_BOOLEAN('x', "exclude-other", &symbol_conf.exclude_other, | 468 | OPT_BOOLEAN('x', "exclude-other", &symbol_conf.exclude_other, |
| 438 | "Only display entries with parent-match"), | 469 | "Only display entries with parent-match"), |
| 439 | OPT_CALLBACK_DEFAULT('g', "call-graph", NULL, "output_type,min_percent", | 470 | OPT_CALLBACK_DEFAULT('g', "call-graph", NULL, "output_type,min_percent, call_order", |
| 440 | "Display callchains using output_type (graph, flat, fractal, or none) and min percent threshold. " | 471 | "Display callchains using output_type (graph, flat, fractal, or none) , min percent threshold and callchain order. " |
| 441 | "Default: fractal,0.5", &parse_callchain_opt, callchain_default_opt), | 472 | "Default: fractal,0.5,callee", &parse_callchain_opt, callchain_default_opt), |
| 473 | OPT_BOOLEAN('G', "inverted", &inverted_callchain, "alias for inverted call graph"), | ||
| 442 | OPT_STRING('d', "dsos", &symbol_conf.dso_list_str, "dso[,dso...]", | 474 | OPT_STRING('d', "dsos", &symbol_conf.dso_list_str, "dso[,dso...]", |
| 443 | "only consider symbols in these dsos"), | 475 | "only consider symbols in these dsos"), |
| 444 | OPT_STRING('C', "comms", &symbol_conf.comm_list_str, "comm[,comm...]", | 476 | OPT_STRING('C', "comms", &symbol_conf.comm_list_str, "comm[,comm...]", |
| @@ -455,6 +487,7 @@ static const struct option options[] = { | |||
| 455 | "Only display entries resolved to a symbol"), | 487 | "Only display entries resolved to a symbol"), |
| 456 | OPT_STRING(0, "symfs", &symbol_conf.symfs, "directory", | 488 | OPT_STRING(0, "symfs", &symbol_conf.symfs, "directory", |
| 457 | "Look for files with symbols relative to this directory"), | 489 | "Look for files with symbols relative to this directory"), |
| 490 | OPT_STRING('c', "cpu", &cpu_list, "cpu", "list of cpus to profile"), | ||
| 458 | OPT_END() | 491 | OPT_END() |
| 459 | }; | 492 | }; |
| 460 | 493 | ||
| @@ -467,6 +500,9 @@ int cmd_report(int argc, const char **argv, const char *prefix __used) | |||
| 467 | else if (use_tui) | 500 | else if (use_tui) |
| 468 | use_browser = 1; | 501 | use_browser = 1; |
| 469 | 502 | ||
| 503 | if (inverted_callchain) | ||
| 504 | callchain_param.order = ORDER_CALLER; | ||
| 505 | |||
| 470 | if (strcmp(input_name, "-") != 0) | 506 | if (strcmp(input_name, "-") != 0) |
| 471 | setup_browser(true); | 507 | setup_browser(true); |
| 472 | else | 508 | else |
| @@ -504,7 +540,14 @@ int cmd_report(int argc, const char **argv, const char *prefix __used) | |||
| 504 | if (parent_pattern != default_parent_pattern) { | 540 | if (parent_pattern != default_parent_pattern) { |
| 505 | if (sort_dimension__add("parent") < 0) | 541 | if (sort_dimension__add("parent") < 0) |
| 506 | return -1; | 542 | return -1; |
| 507 | sort_parent.elide = 1; | 543 | |
| 544 | /* | ||
| 545 | * Only show the parent fields if we explicitly | ||
| 546 | * sort that way. If we only use parent machinery | ||
| 547 | * for filtering, we don't want it. | ||
| 548 | */ | ||
| 549 | if (!strstr(sort_order, "parent")) | ||
| 550 | sort_parent.elide = 1; | ||
| 508 | } else | 551 | } else |
| 509 | symbol_conf.exclude_other = false; | 552 | symbol_conf.exclude_other = false; |
| 510 | 553 | ||
diff --git a/tools/perf/builtin-script.c b/tools/perf/builtin-script.c index 22747de7234..09024ec2ab2 100644 --- a/tools/perf/builtin-script.c +++ b/tools/perf/builtin-script.c | |||
| @@ -13,6 +13,7 @@ | |||
| 13 | #include "util/util.h" | 13 | #include "util/util.h" |
| 14 | #include "util/evlist.h" | 14 | #include "util/evlist.h" |
| 15 | #include "util/evsel.h" | 15 | #include "util/evsel.h" |
| 16 | #include <linux/bitmap.h> | ||
| 16 | 17 | ||
| 17 | static char const *script_name; | 18 | static char const *script_name; |
| 18 | static char const *generate_script_lang; | 19 | static char const *generate_script_lang; |
| @@ -21,6 +22,8 @@ static u64 last_timestamp; | |||
| 21 | static u64 nr_unordered; | 22 | static u64 nr_unordered; |
| 22 | extern const struct option record_options[]; | 23 | extern const struct option record_options[]; |
| 23 | static bool no_callchain; | 24 | static bool no_callchain; |
| 25 | static const char *cpu_list; | ||
| 26 | static DECLARE_BITMAP(cpu_bitmap, MAX_NR_CPUS); | ||
| 24 | 27 | ||
| 25 | enum perf_output_field { | 28 | enum perf_output_field { |
| 26 | PERF_OUTPUT_COMM = 1U << 0, | 29 | PERF_OUTPUT_COMM = 1U << 0, |
| @@ -30,7 +33,10 @@ enum perf_output_field { | |||
| 30 | PERF_OUTPUT_CPU = 1U << 4, | 33 | PERF_OUTPUT_CPU = 1U << 4, |
| 31 | PERF_OUTPUT_EVNAME = 1U << 5, | 34 | PERF_OUTPUT_EVNAME = 1U << 5, |
| 32 | PERF_OUTPUT_TRACE = 1U << 6, | 35 | PERF_OUTPUT_TRACE = 1U << 6, |
| 33 | PERF_OUTPUT_SYM = 1U << 7, | 36 | PERF_OUTPUT_IP = 1U << 7, |
| 37 | PERF_OUTPUT_SYM = 1U << 8, | ||
| 38 | PERF_OUTPUT_DSO = 1U << 9, | ||
| 39 | PERF_OUTPUT_ADDR = 1U << 10, | ||
| 34 | }; | 40 | }; |
| 35 | 41 | ||
| 36 | struct output_option { | 42 | struct output_option { |
| @@ -44,7 +50,10 @@ struct output_option { | |||
| 44 | {.str = "cpu", .field = PERF_OUTPUT_CPU}, | 50 | {.str = "cpu", .field = PERF_OUTPUT_CPU}, |
| 45 | {.str = "event", .field = PERF_OUTPUT_EVNAME}, | 51 | {.str = "event", .field = PERF_OUTPUT_EVNAME}, |
| 46 | {.str = "trace", .field = PERF_OUTPUT_TRACE}, | 52 | {.str = "trace", .field = PERF_OUTPUT_TRACE}, |
| 53 | {.str = "ip", .field = PERF_OUTPUT_IP}, | ||
| 47 | {.str = "sym", .field = PERF_OUTPUT_SYM}, | 54 | {.str = "sym", .field = PERF_OUTPUT_SYM}, |
| 55 | {.str = "dso", .field = PERF_OUTPUT_DSO}, | ||
| 56 | {.str = "addr", .field = PERF_OUTPUT_ADDR}, | ||
| 48 | }; | 57 | }; |
| 49 | 58 | ||
| 50 | /* default set to maintain compatibility with current format */ | 59 | /* default set to maintain compatibility with current format */ |
| @@ -60,7 +69,8 @@ static struct { | |||
| 60 | 69 | ||
| 61 | .fields = PERF_OUTPUT_COMM | PERF_OUTPUT_TID | | 70 | .fields = PERF_OUTPUT_COMM | PERF_OUTPUT_TID | |
| 62 | PERF_OUTPUT_CPU | PERF_OUTPUT_TIME | | 71 | PERF_OUTPUT_CPU | PERF_OUTPUT_TIME | |
| 63 | PERF_OUTPUT_EVNAME | PERF_OUTPUT_SYM, | 72 | PERF_OUTPUT_EVNAME | PERF_OUTPUT_IP | |
| 73 | PERF_OUTPUT_SYM | PERF_OUTPUT_DSO, | ||
| 64 | 74 | ||
| 65 | .invalid_fields = PERF_OUTPUT_TRACE, | 75 | .invalid_fields = PERF_OUTPUT_TRACE, |
| 66 | }, | 76 | }, |
| @@ -70,7 +80,8 @@ static struct { | |||
| 70 | 80 | ||
| 71 | .fields = PERF_OUTPUT_COMM | PERF_OUTPUT_TID | | 81 | .fields = PERF_OUTPUT_COMM | PERF_OUTPUT_TID | |
| 72 | PERF_OUTPUT_CPU | PERF_OUTPUT_TIME | | 82 | PERF_OUTPUT_CPU | PERF_OUTPUT_TIME | |
| 73 | PERF_OUTPUT_EVNAME | PERF_OUTPUT_SYM, | 83 | PERF_OUTPUT_EVNAME | PERF_OUTPUT_IP | |
| 84 | PERF_OUTPUT_SYM | PERF_OUTPUT_DSO, | ||
| 74 | 85 | ||
| 75 | .invalid_fields = PERF_OUTPUT_TRACE, | 86 | .invalid_fields = PERF_OUTPUT_TRACE, |
| 76 | }, | 87 | }, |
| @@ -88,7 +99,8 @@ static struct { | |||
| 88 | 99 | ||
| 89 | .fields = PERF_OUTPUT_COMM | PERF_OUTPUT_TID | | 100 | .fields = PERF_OUTPUT_COMM | PERF_OUTPUT_TID | |
| 90 | PERF_OUTPUT_CPU | PERF_OUTPUT_TIME | | 101 | PERF_OUTPUT_CPU | PERF_OUTPUT_TIME | |
| 91 | PERF_OUTPUT_EVNAME | PERF_OUTPUT_SYM, | 102 | PERF_OUTPUT_EVNAME | PERF_OUTPUT_IP | |
| 103 | PERF_OUTPUT_SYM | PERF_OUTPUT_DSO, | ||
| 92 | 104 | ||
| 93 | .invalid_fields = PERF_OUTPUT_TRACE, | 105 | .invalid_fields = PERF_OUTPUT_TRACE, |
| 94 | }, | 106 | }, |
| @@ -157,9 +169,9 @@ static int perf_evsel__check_attr(struct perf_evsel *evsel, | |||
| 157 | !perf_session__has_traces(session, "record -R")) | 169 | !perf_session__has_traces(session, "record -R")) |
| 158 | return -EINVAL; | 170 | return -EINVAL; |
| 159 | 171 | ||
| 160 | if (PRINT_FIELD(SYM)) { | 172 | if (PRINT_FIELD(IP)) { |
| 161 | if (perf_event_attr__check_stype(attr, PERF_SAMPLE_IP, "IP", | 173 | if (perf_event_attr__check_stype(attr, PERF_SAMPLE_IP, "IP", |
| 162 | PERF_OUTPUT_SYM)) | 174 | PERF_OUTPUT_IP)) |
| 163 | return -EINVAL; | 175 | return -EINVAL; |
| 164 | 176 | ||
| 165 | if (!no_callchain && | 177 | if (!no_callchain && |
| @@ -167,6 +179,24 @@ static int perf_evsel__check_attr(struct perf_evsel *evsel, | |||
| 167 | symbol_conf.use_callchain = false; | 179 | symbol_conf.use_callchain = false; |
| 168 | } | 180 | } |
| 169 | 181 | ||
| 182 | if (PRINT_FIELD(ADDR) && | ||
| 183 | perf_event_attr__check_stype(attr, PERF_SAMPLE_ADDR, "ADDR", | ||
| 184 | PERF_OUTPUT_ADDR)) | ||
| 185 | return -EINVAL; | ||
| 186 | |||
| 187 | if (PRINT_FIELD(SYM) && !PRINT_FIELD(IP) && !PRINT_FIELD(ADDR)) { | ||
| 188 | pr_err("Display of symbols requested but neither sample IP nor " | ||
| 189 | "sample address\nis selected. Hence, no addresses to convert " | ||
| 190 | "to symbols.\n"); | ||
| 191 | return -EINVAL; | ||
| 192 | } | ||
| 193 | if (PRINT_FIELD(DSO) && !PRINT_FIELD(IP) && !PRINT_FIELD(ADDR)) { | ||
| 194 | pr_err("Display of DSO requested but neither sample IP nor " | ||
| 195 | "sample address\nis selected. Hence, no addresses to convert " | ||
| 196 | "to DSO.\n"); | ||
| 197 | return -EINVAL; | ||
| 198 | } | ||
| 199 | |||
| 170 | if ((PRINT_FIELD(PID) || PRINT_FIELD(TID)) && | 200 | if ((PRINT_FIELD(PID) || PRINT_FIELD(TID)) && |
| 171 | perf_event_attr__check_stype(attr, PERF_SAMPLE_TID, "TID", | 201 | perf_event_attr__check_stype(attr, PERF_SAMPLE_TID, "TID", |
| 172 | PERF_OUTPUT_TID|PERF_OUTPUT_PID)) | 202 | PERF_OUTPUT_TID|PERF_OUTPUT_PID)) |
| @@ -230,7 +260,7 @@ static void print_sample_start(struct perf_sample *sample, | |||
| 230 | if (PRINT_FIELD(COMM)) { | 260 | if (PRINT_FIELD(COMM)) { |
| 231 | if (latency_format) | 261 | if (latency_format) |
| 232 | printf("%8.8s ", thread->comm); | 262 | printf("%8.8s ", thread->comm); |
| 233 | else if (PRINT_FIELD(SYM) && symbol_conf.use_callchain) | 263 | else if (PRINT_FIELD(IP) && symbol_conf.use_callchain) |
| 234 | printf("%s ", thread->comm); | 264 | printf("%s ", thread->comm); |
| 235 | else | 265 | else |
| 236 | printf("%16s ", thread->comm); | 266 | printf("%16s ", thread->comm); |
| @@ -271,6 +301,63 @@ static void print_sample_start(struct perf_sample *sample, | |||
| 271 | } | 301 | } |
| 272 | } | 302 | } |
| 273 | 303 | ||
| 304 | static bool sample_addr_correlates_sym(struct perf_event_attr *attr) | ||
| 305 | { | ||
| 306 | if ((attr->type == PERF_TYPE_SOFTWARE) && | ||
| 307 | ((attr->config == PERF_COUNT_SW_PAGE_FAULTS) || | ||
| 308 | (attr->config == PERF_COUNT_SW_PAGE_FAULTS_MIN) || | ||
| 309 | (attr->config == PERF_COUNT_SW_PAGE_FAULTS_MAJ))) | ||
| 310 | return true; | ||
| 311 | |||
| 312 | return false; | ||
| 313 | } | ||
| 314 | |||
| 315 | static void print_sample_addr(union perf_event *event, | ||
| 316 | struct perf_sample *sample, | ||
| 317 | struct perf_session *session, | ||
| 318 | struct thread *thread, | ||
| 319 | struct perf_event_attr *attr) | ||
| 320 | { | ||
| 321 | struct addr_location al; | ||
| 322 | u8 cpumode = event->header.misc & PERF_RECORD_MISC_CPUMODE_MASK; | ||
| 323 | const char *symname, *dsoname; | ||
| 324 | |||
| 325 | printf("%16" PRIx64, sample->addr); | ||
| 326 | |||
| 327 | if (!sample_addr_correlates_sym(attr)) | ||
| 328 | return; | ||
| 329 | |||
| 330 | thread__find_addr_map(thread, session, cpumode, MAP__FUNCTION, | ||
| 331 | event->ip.pid, sample->addr, &al); | ||
| 332 | if (!al.map) | ||
| 333 | thread__find_addr_map(thread, session, cpumode, MAP__VARIABLE, | ||
| 334 | event->ip.pid, sample->addr, &al); | ||
| 335 | |||
| 336 | al.cpu = sample->cpu; | ||
| 337 | al.sym = NULL; | ||
| 338 | |||
| 339 | if (al.map) | ||
| 340 | al.sym = map__find_symbol(al.map, al.addr, NULL); | ||
| 341 | |||
| 342 | if (PRINT_FIELD(SYM)) { | ||
| 343 | if (al.sym && al.sym->name) | ||
| 344 | symname = al.sym->name; | ||
| 345 | else | ||
| 346 | symname = ""; | ||
| 347 | |||
| 348 | printf(" %16s", symname); | ||
| 349 | } | ||
| 350 | |||
| 351 | if (PRINT_FIELD(DSO)) { | ||
| 352 | if (al.map && al.map->dso && al.map->dso->name) | ||
| 353 | dsoname = al.map->dso->name; | ||
| 354 | else | ||
| 355 | dsoname = ""; | ||
| 356 | |||
| 357 | printf(" (%s)", dsoname); | ||
| 358 | } | ||
| 359 | } | ||
| 360 | |||
| 274 | static void process_event(union perf_event *event __unused, | 361 | static void process_event(union perf_event *event __unused, |
| 275 | struct perf_sample *sample, | 362 | struct perf_sample *sample, |
| 276 | struct perf_evsel *evsel, | 363 | struct perf_evsel *evsel, |
| @@ -288,12 +375,16 @@ static void process_event(union perf_event *event __unused, | |||
| 288 | print_trace_event(sample->cpu, sample->raw_data, | 375 | print_trace_event(sample->cpu, sample->raw_data, |
| 289 | sample->raw_size); | 376 | sample->raw_size); |
| 290 | 377 | ||
| 291 | if (PRINT_FIELD(SYM)) { | 378 | if (PRINT_FIELD(ADDR)) |
| 379 | print_sample_addr(event, sample, session, thread, attr); | ||
| 380 | |||
| 381 | if (PRINT_FIELD(IP)) { | ||
| 292 | if (!symbol_conf.use_callchain) | 382 | if (!symbol_conf.use_callchain) |
| 293 | printf(" "); | 383 | printf(" "); |
| 294 | else | 384 | else |
| 295 | printf("\n"); | 385 | printf("\n"); |
| 296 | perf_session__print_symbols(event, sample, session); | 386 | perf_session__print_ip(event, sample, session, |
| 387 | PRINT_FIELD(SYM), PRINT_FIELD(DSO)); | ||
| 297 | } | 388 | } |
| 298 | 389 | ||
| 299 | printf("\n"); | 390 | printf("\n"); |
| @@ -365,6 +456,10 @@ static int process_sample_event(union perf_event *event, | |||
| 365 | last_timestamp = sample->time; | 456 | last_timestamp = sample->time; |
| 366 | return 0; | 457 | return 0; |
| 367 | } | 458 | } |
| 459 | |||
| 460 | if (cpu_list && !test_bit(sample->cpu, cpu_bitmap)) | ||
| 461 | return 0; | ||
| 462 | |||
| 368 | scripting_ops->process_event(event, sample, evsel, session, thread); | 463 | scripting_ops->process_event(event, sample, evsel, session, thread); |
| 369 | 464 | ||
| 370 | session->hists.stats.total_period += sample->period; | 465 | session->hists.stats.total_period += sample->period; |
| @@ -985,8 +1080,9 @@ static const struct option options[] = { | |||
| 985 | OPT_STRING(0, "symfs", &symbol_conf.symfs, "directory", | 1080 | OPT_STRING(0, "symfs", &symbol_conf.symfs, "directory", |
| 986 | "Look for files with symbols relative to this directory"), | 1081 | "Look for files with symbols relative to this directory"), |
| 987 | OPT_CALLBACK('f', "fields", NULL, "str", | 1082 | OPT_CALLBACK('f', "fields", NULL, "str", |
| 988 | "comma separated output fields prepend with 'type:'. Valid types: hw,sw,trace,raw. Fields: comm,tid,pid,time,cpu,event,trace,sym", | 1083 | "comma separated output fields prepend with 'type:'. Valid types: hw,sw,trace,raw. Fields: comm,tid,pid,time,cpu,event,trace,ip,sym,dso,addr", |
| 989 | parse_output_fields), | 1084 | parse_output_fields), |
| 1085 | OPT_STRING('c', "cpu", &cpu_list, "cpu", "list of cpus to profile"), | ||
| 990 | 1086 | ||
| 991 | OPT_END() | 1087 | OPT_END() |
| 992 | }; | 1088 | }; |
| @@ -1167,6 +1263,11 @@ int cmd_script(int argc, const char **argv, const char *prefix __used) | |||
| 1167 | if (session == NULL) | 1263 | if (session == NULL) |
| 1168 | return -ENOMEM; | 1264 | return -ENOMEM; |
| 1169 | 1265 | ||
| 1266 | if (cpu_list) { | ||
| 1267 | if (perf_session__cpu_bitmap(session, cpu_list, cpu_bitmap)) | ||
| 1268 | return -1; | ||
| 1269 | } | ||
| 1270 | |||
| 1170 | if (!no_callchain) | 1271 | if (!no_callchain) |
| 1171 | symbol_conf.use_callchain = true; | 1272 | symbol_conf.use_callchain = true; |
| 1172 | else | 1273 | else |
diff --git a/tools/perf/builtin-stat.c b/tools/perf/builtin-stat.c index a9f06715e44..1ad04ce29c3 100644 --- a/tools/perf/builtin-stat.c +++ b/tools/perf/builtin-stat.c | |||
| @@ -61,6 +61,8 @@ | |||
| 61 | #include <locale.h> | 61 | #include <locale.h> |
| 62 | 62 | ||
| 63 | #define DEFAULT_SEPARATOR " " | 63 | #define DEFAULT_SEPARATOR " " |
| 64 | #define CNTR_NOT_SUPPORTED "<not supported>" | ||
| 65 | #define CNTR_NOT_COUNTED "<not counted>" | ||
| 64 | 66 | ||
| 65 | static struct perf_event_attr default_attrs[] = { | 67 | static struct perf_event_attr default_attrs[] = { |
| 66 | 68 | ||
| @@ -448,6 +450,7 @@ static int run_perf_stat(int argc __used, const char **argv) | |||
| 448 | if (verbose) | 450 | if (verbose) |
| 449 | ui__warning("%s event is not supported by the kernel.\n", | 451 | ui__warning("%s event is not supported by the kernel.\n", |
| 450 | event_name(counter)); | 452 | event_name(counter)); |
| 453 | counter->supported = false; | ||
| 451 | continue; | 454 | continue; |
| 452 | } | 455 | } |
| 453 | 456 | ||
| @@ -466,6 +469,7 @@ static int run_perf_stat(int argc __used, const char **argv) | |||
| 466 | die("Not all events could be opened.\n"); | 469 | die("Not all events could be opened.\n"); |
| 467 | return -1; | 470 | return -1; |
| 468 | } | 471 | } |
| 472 | counter->supported = true; | ||
| 469 | } | 473 | } |
| 470 | 474 | ||
| 471 | if (perf_evlist__set_filters(evsel_list)) { | 475 | if (perf_evlist__set_filters(evsel_list)) { |
| @@ -513,7 +517,10 @@ static void print_noise_pct(double total, double avg) | |||
| 513 | if (avg) | 517 | if (avg) |
| 514 | pct = 100.0*total/avg; | 518 | pct = 100.0*total/avg; |
| 515 | 519 | ||
| 516 | fprintf(stderr, " ( +-%6.2f%% )", pct); | 520 | if (csv_output) |
| 521 | fprintf(stderr, "%s%.2f%%", csv_sep, pct); | ||
| 522 | else | ||
| 523 | fprintf(stderr, " ( +-%6.2f%% )", pct); | ||
| 517 | } | 524 | } |
| 518 | 525 | ||
| 519 | static void print_noise(struct perf_evsel *evsel, double avg) | 526 | static void print_noise(struct perf_evsel *evsel, double avg) |
| @@ -861,7 +868,7 @@ static void print_counter_aggr(struct perf_evsel *counter) | |||
| 861 | if (scaled == -1) { | 868 | if (scaled == -1) { |
| 862 | fprintf(stderr, "%*s%s%*s", | 869 | fprintf(stderr, "%*s%s%*s", |
| 863 | csv_output ? 0 : 18, | 870 | csv_output ? 0 : 18, |
| 864 | "<not counted>", | 871 | counter->supported ? CNTR_NOT_COUNTED : CNTR_NOT_SUPPORTED, |
| 865 | csv_sep, | 872 | csv_sep, |
| 866 | csv_output ? 0 : -24, | 873 | csv_output ? 0 : -24, |
| 867 | event_name(counter)); | 874 | event_name(counter)); |
| @@ -878,13 +885,13 @@ static void print_counter_aggr(struct perf_evsel *counter) | |||
| 878 | else | 885 | else |
| 879 | abs_printout(-1, counter, avg); | 886 | abs_printout(-1, counter, avg); |
| 880 | 887 | ||
| 888 | print_noise(counter, avg); | ||
| 889 | |||
| 881 | if (csv_output) { | 890 | if (csv_output) { |
| 882 | fputc('\n', stderr); | 891 | fputc('\n', stderr); |
| 883 | return; | 892 | return; |
| 884 | } | 893 | } |
| 885 | 894 | ||
| 886 | print_noise(counter, avg); | ||
| 887 | |||
| 888 | if (scaled) { | 895 | if (scaled) { |
| 889 | double avg_enabled, avg_running; | 896 | double avg_enabled, avg_running; |
| 890 | 897 | ||
| @@ -914,7 +921,8 @@ static void print_counter(struct perf_evsel *counter) | |||
| 914 | csv_output ? 0 : -4, | 921 | csv_output ? 0 : -4, |
| 915 | evsel_list->cpus->map[cpu], csv_sep, | 922 | evsel_list->cpus->map[cpu], csv_sep, |
| 916 | csv_output ? 0 : 18, | 923 | csv_output ? 0 : 18, |
| 917 | "<not counted>", csv_sep, | 924 | counter->supported ? CNTR_NOT_COUNTED : CNTR_NOT_SUPPORTED, |
| 925 | csv_sep, | ||
| 918 | csv_output ? 0 : -24, | 926 | csv_output ? 0 : -24, |
| 919 | event_name(counter)); | 927 | event_name(counter)); |
| 920 | 928 | ||
| @@ -1024,7 +1032,7 @@ static int stat__set_big_num(const struct option *opt __used, | |||
| 1024 | static const struct option options[] = { | 1032 | static const struct option options[] = { |
| 1025 | OPT_CALLBACK('e', "event", &evsel_list, "event", | 1033 | OPT_CALLBACK('e', "event", &evsel_list, "event", |
| 1026 | "event selector. use 'perf list' to list available events", | 1034 | "event selector. use 'perf list' to list available events", |
| 1027 | parse_events), | 1035 | parse_events_option), |
| 1028 | OPT_CALLBACK(0, "filter", &evsel_list, "filter", | 1036 | OPT_CALLBACK(0, "filter", &evsel_list, "filter", |
| 1029 | "event filter", parse_filter), | 1037 | "event filter", parse_filter), |
| 1030 | OPT_BOOLEAN('i', "no-inherit", &no_inherit, | 1038 | OPT_BOOLEAN('i', "no-inherit", &no_inherit, |
diff --git a/tools/perf/builtin-test.c b/tools/perf/builtin-test.c index 2da9162262b..55f4c76f282 100644 --- a/tools/perf/builtin-test.c +++ b/tools/perf/builtin-test.c | |||
| @@ -12,6 +12,7 @@ | |||
| 12 | #include "util/parse-events.h" | 12 | #include "util/parse-events.h" |
| 13 | #include "util/symbol.h" | 13 | #include "util/symbol.h" |
| 14 | #include "util/thread_map.h" | 14 | #include "util/thread_map.h" |
| 15 | #include "../../include/linux/hw_breakpoint.h" | ||
| 15 | 16 | ||
| 16 | static long page_size; | 17 | static long page_size; |
| 17 | 18 | ||
| @@ -245,8 +246,8 @@ static int trace_event__id(const char *evname) | |||
| 245 | int err = -1, fd; | 246 | int err = -1, fd; |
| 246 | 247 | ||
| 247 | if (asprintf(&filename, | 248 | if (asprintf(&filename, |
| 248 | "/sys/kernel/debug/tracing/events/syscalls/%s/id", | 249 | "%s/syscalls/%s/id", |
| 249 | evname) < 0) | 250 | debugfs_path, evname) < 0) |
| 250 | return -1; | 251 | return -1; |
| 251 | 252 | ||
| 252 | fd = open(filename, O_RDONLY); | 253 | fd = open(filename, O_RDONLY); |
| @@ -600,6 +601,246 @@ out_free_threads: | |||
| 600 | #undef nsyscalls | 601 | #undef nsyscalls |
| 601 | } | 602 | } |
| 602 | 603 | ||
| 604 | #define TEST_ASSERT_VAL(text, cond) \ | ||
| 605 | do { \ | ||
| 606 | if (!cond) { \ | ||
| 607 | pr_debug("FAILED %s:%d %s\n", __FILE__, __LINE__, text); \ | ||
| 608 | return -1; \ | ||
| 609 | } \ | ||
| 610 | } while (0) | ||
| 611 | |||
| 612 | static int test__checkevent_tracepoint(struct perf_evlist *evlist) | ||
| 613 | { | ||
| 614 | struct perf_evsel *evsel = list_entry(evlist->entries.next, | ||
| 615 | struct perf_evsel, node); | ||
| 616 | |||
| 617 | TEST_ASSERT_VAL("wrong number of entries", 1 == evlist->nr_entries); | ||
| 618 | TEST_ASSERT_VAL("wrong type", PERF_TYPE_TRACEPOINT == evsel->attr.type); | ||
| 619 | TEST_ASSERT_VAL("wrong sample_type", | ||
| 620 | (PERF_SAMPLE_RAW | PERF_SAMPLE_TIME | PERF_SAMPLE_CPU) == | ||
| 621 | evsel->attr.sample_type); | ||
| 622 | TEST_ASSERT_VAL("wrong sample_period", 1 == evsel->attr.sample_period); | ||
| 623 | return 0; | ||
| 624 | } | ||
| 625 | |||
| 626 | static int test__checkevent_tracepoint_multi(struct perf_evlist *evlist) | ||
| 627 | { | ||
| 628 | struct perf_evsel *evsel; | ||
| 629 | |||
| 630 | TEST_ASSERT_VAL("wrong number of entries", evlist->nr_entries > 1); | ||
| 631 | |||
| 632 | list_for_each_entry(evsel, &evlist->entries, node) { | ||
| 633 | TEST_ASSERT_VAL("wrong type", | ||
| 634 | PERF_TYPE_TRACEPOINT == evsel->attr.type); | ||
| 635 | TEST_ASSERT_VAL("wrong sample_type", | ||
| 636 | (PERF_SAMPLE_RAW | PERF_SAMPLE_TIME | PERF_SAMPLE_CPU) | ||
| 637 | == evsel->attr.sample_type); | ||
| 638 | TEST_ASSERT_VAL("wrong sample_period", | ||
| 639 | 1 == evsel->attr.sample_period); | ||
| 640 | } | ||
| 641 | return 0; | ||
| 642 | } | ||
| 643 | |||
| 644 | static int test__checkevent_raw(struct perf_evlist *evlist) | ||
| 645 | { | ||
| 646 | struct perf_evsel *evsel = list_entry(evlist->entries.next, | ||
| 647 | struct perf_evsel, node); | ||
| 648 | |||
| 649 | TEST_ASSERT_VAL("wrong number of entries", 1 == evlist->nr_entries); | ||
| 650 | TEST_ASSERT_VAL("wrong type", PERF_TYPE_RAW == evsel->attr.type); | ||
| 651 | TEST_ASSERT_VAL("wrong config", 1 == evsel->attr.config); | ||
| 652 | return 0; | ||
| 653 | } | ||
| 654 | |||
| 655 | static int test__checkevent_numeric(struct perf_evlist *evlist) | ||
| 656 | { | ||
| 657 | struct perf_evsel *evsel = list_entry(evlist->entries.next, | ||
| 658 | struct perf_evsel, node); | ||
| 659 | |||
| 660 | TEST_ASSERT_VAL("wrong number of entries", 1 == evlist->nr_entries); | ||
| 661 | TEST_ASSERT_VAL("wrong type", 1 == evsel->attr.type); | ||
| 662 | TEST_ASSERT_VAL("wrong config", 1 == evsel->attr.config); | ||
| 663 | return 0; | ||
| 664 | } | ||
| 665 | |||
| 666 | static int test__checkevent_symbolic_name(struct perf_evlist *evlist) | ||
| 667 | { | ||
| 668 | struct perf_evsel *evsel = list_entry(evlist->entries.next, | ||
| 669 | struct perf_evsel, node); | ||
| 670 | |||
| 671 | TEST_ASSERT_VAL("wrong number of entries", 1 == evlist->nr_entries); | ||
| 672 | TEST_ASSERT_VAL("wrong type", PERF_TYPE_HARDWARE == evsel->attr.type); | ||
| 673 | TEST_ASSERT_VAL("wrong config", | ||
| 674 | PERF_COUNT_HW_INSTRUCTIONS == evsel->attr.config); | ||
| 675 | return 0; | ||
| 676 | } | ||
| 677 | |||
| 678 | static int test__checkevent_symbolic_alias(struct perf_evlist *evlist) | ||
| 679 | { | ||
| 680 | struct perf_evsel *evsel = list_entry(evlist->entries.next, | ||
| 681 | struct perf_evsel, node); | ||
| 682 | |||
| 683 | TEST_ASSERT_VAL("wrong number of entries", 1 == evlist->nr_entries); | ||
| 684 | TEST_ASSERT_VAL("wrong type", PERF_TYPE_SOFTWARE == evsel->attr.type); | ||
| 685 | TEST_ASSERT_VAL("wrong config", | ||
| 686 | PERF_COUNT_SW_PAGE_FAULTS == evsel->attr.config); | ||
| 687 | return 0; | ||
| 688 | } | ||
| 689 | |||
| 690 | static int test__checkevent_genhw(struct perf_evlist *evlist) | ||
| 691 | { | ||
| 692 | struct perf_evsel *evsel = list_entry(evlist->entries.next, | ||
| 693 | struct perf_evsel, node); | ||
| 694 | |||
| 695 | TEST_ASSERT_VAL("wrong number of entries", 1 == evlist->nr_entries); | ||
| 696 | TEST_ASSERT_VAL("wrong type", PERF_TYPE_HW_CACHE == evsel->attr.type); | ||
| 697 | TEST_ASSERT_VAL("wrong config", (1 << 16) == evsel->attr.config); | ||
| 698 | return 0; | ||
| 699 | } | ||
| 700 | |||
| 701 | static int test__checkevent_breakpoint(struct perf_evlist *evlist) | ||
| 702 | { | ||
| 703 | struct perf_evsel *evsel = list_entry(evlist->entries.next, | ||
| 704 | struct perf_evsel, node); | ||
| 705 | |||
| 706 | TEST_ASSERT_VAL("wrong number of entries", 1 == evlist->nr_entries); | ||
| 707 | TEST_ASSERT_VAL("wrong type", PERF_TYPE_BREAKPOINT == evsel->attr.type); | ||
| 708 | TEST_ASSERT_VAL("wrong config", 0 == evsel->attr.config); | ||
| 709 | TEST_ASSERT_VAL("wrong bp_type", (HW_BREAKPOINT_R | HW_BREAKPOINT_W) == | ||
| 710 | evsel->attr.bp_type); | ||
| 711 | TEST_ASSERT_VAL("wrong bp_len", HW_BREAKPOINT_LEN_4 == | ||
| 712 | evsel->attr.bp_len); | ||
| 713 | return 0; | ||
| 714 | } | ||
| 715 | |||
| 716 | static int test__checkevent_breakpoint_x(struct perf_evlist *evlist) | ||
| 717 | { | ||
| 718 | struct perf_evsel *evsel = list_entry(evlist->entries.next, | ||
| 719 | struct perf_evsel, node); | ||
| 720 | |||
| 721 | TEST_ASSERT_VAL("wrong number of entries", 1 == evlist->nr_entries); | ||
| 722 | TEST_ASSERT_VAL("wrong type", PERF_TYPE_BREAKPOINT == evsel->attr.type); | ||
| 723 | TEST_ASSERT_VAL("wrong config", 0 == evsel->attr.config); | ||
| 724 | TEST_ASSERT_VAL("wrong bp_type", | ||
| 725 | HW_BREAKPOINT_X == evsel->attr.bp_type); | ||
| 726 | TEST_ASSERT_VAL("wrong bp_len", sizeof(long) == evsel->attr.bp_len); | ||
| 727 | return 0; | ||
| 728 | } | ||
| 729 | |||
| 730 | static int test__checkevent_breakpoint_r(struct perf_evlist *evlist) | ||
| 731 | { | ||
| 732 | struct perf_evsel *evsel = list_entry(evlist->entries.next, | ||
| 733 | struct perf_evsel, node); | ||
| 734 | |||
| 735 | TEST_ASSERT_VAL("wrong number of entries", 1 == evlist->nr_entries); | ||
| 736 | TEST_ASSERT_VAL("wrong type", | ||
| 737 | PERF_TYPE_BREAKPOINT == evsel->attr.type); | ||
| 738 | TEST_ASSERT_VAL("wrong config", 0 == evsel->attr.config); | ||
| 739 | TEST_ASSERT_VAL("wrong bp_type", | ||
| 740 | HW_BREAKPOINT_R == evsel->attr.bp_type); | ||
| 741 | TEST_ASSERT_VAL("wrong bp_len", | ||
| 742 | HW_BREAKPOINT_LEN_4 == evsel->attr.bp_len); | ||
| 743 | return 0; | ||
| 744 | } | ||
| 745 | |||
| 746 | static int test__checkevent_breakpoint_w(struct perf_evlist *evlist) | ||
| 747 | { | ||
| 748 | struct perf_evsel *evsel = list_entry(evlist->entries.next, | ||
| 749 | struct perf_evsel, node); | ||
| 750 | |||
| 751 | TEST_ASSERT_VAL("wrong number of entries", 1 == evlist->nr_entries); | ||
| 752 | TEST_ASSERT_VAL("wrong type", | ||
| 753 | PERF_TYPE_BREAKPOINT == evsel->attr.type); | ||
| 754 | TEST_ASSERT_VAL("wrong config", 0 == evsel->attr.config); | ||
| 755 | TEST_ASSERT_VAL("wrong bp_type", | ||
| 756 | HW_BREAKPOINT_W == evsel->attr.bp_type); | ||
| 757 | TEST_ASSERT_VAL("wrong bp_len", | ||
| 758 | HW_BREAKPOINT_LEN_4 == evsel->attr.bp_len); | ||
| 759 | return 0; | ||
| 760 | } | ||
| 761 | |||
| 762 | static struct test__event_st { | ||
| 763 | const char *name; | ||
| 764 | __u32 type; | ||
| 765 | int (*check)(struct perf_evlist *evlist); | ||
| 766 | } test__events[] = { | ||
| 767 | { | ||
| 768 | .name = "syscalls:sys_enter_open", | ||
| 769 | .check = test__checkevent_tracepoint, | ||
| 770 | }, | ||
| 771 | { | ||
| 772 | .name = "syscalls:*", | ||
| 773 | .check = test__checkevent_tracepoint_multi, | ||
| 774 | }, | ||
| 775 | { | ||
| 776 | .name = "r1", | ||
| 777 | .check = test__checkevent_raw, | ||
| 778 | }, | ||
| 779 | { | ||
| 780 | .name = "1:1", | ||
| 781 | .check = test__checkevent_numeric, | ||
| 782 | }, | ||
| 783 | { | ||
| 784 | .name = "instructions", | ||
| 785 | .check = test__checkevent_symbolic_name, | ||
| 786 | }, | ||
| 787 | { | ||
| 788 | .name = "faults", | ||
| 789 | .check = test__checkevent_symbolic_alias, | ||
| 790 | }, | ||
| 791 | { | ||
| 792 | .name = "L1-dcache-load-miss", | ||
| 793 | .check = test__checkevent_genhw, | ||
| 794 | }, | ||
| 795 | { | ||
| 796 | .name = "mem:0", | ||
| 797 | .check = test__checkevent_breakpoint, | ||
| 798 | }, | ||
| 799 | { | ||
| 800 | .name = "mem:0:x", | ||
| 801 | .check = test__checkevent_breakpoint_x, | ||
| 802 | }, | ||
| 803 | { | ||
| 804 | .name = "mem:0:r", | ||
| 805 | .check = test__checkevent_breakpoint_r, | ||
| 806 | }, | ||
| 807 | { | ||
| 808 | .name = "mem:0:w", | ||
| 809 | .check = test__checkevent_breakpoint_w, | ||
| 810 | }, | ||
| 811 | }; | ||
| 812 | |||
| 813 | #define TEST__EVENTS_CNT (sizeof(test__events) / sizeof(struct test__event_st)) | ||
| 814 | |||
| 815 | static int test__parse_events(void) | ||
| 816 | { | ||
| 817 | struct perf_evlist *evlist; | ||
| 818 | u_int i; | ||
| 819 | int ret = 0; | ||
| 820 | |||
| 821 | for (i = 0; i < TEST__EVENTS_CNT; i++) { | ||
| 822 | struct test__event_st *e = &test__events[i]; | ||
| 823 | |||
| 824 | evlist = perf_evlist__new(NULL, NULL); | ||
| 825 | if (evlist == NULL) | ||
| 826 | break; | ||
| 827 | |||
| 828 | ret = parse_events(evlist, e->name, 0); | ||
| 829 | if (ret) { | ||
| 830 | pr_debug("failed to parse event '%s', err %d\n", | ||
| 831 | e->name, ret); | ||
| 832 | break; | ||
| 833 | } | ||
| 834 | |||
| 835 | ret = e->check(evlist); | ||
| 836 | if (ret) | ||
| 837 | break; | ||
| 838 | |||
| 839 | perf_evlist__delete(evlist); | ||
| 840 | } | ||
| 841 | |||
| 842 | return ret; | ||
| 843 | } | ||
| 603 | static struct test { | 844 | static struct test { |
| 604 | const char *desc; | 845 | const char *desc; |
| 605 | int (*func)(void); | 846 | int (*func)(void); |
| @@ -621,6 +862,10 @@ static struct test { | |||
| 621 | .func = test__basic_mmap, | 862 | .func = test__basic_mmap, |
| 622 | }, | 863 | }, |
| 623 | { | 864 | { |
| 865 | .desc = "parse events tests", | ||
| 866 | .func = test__parse_events, | ||
| 867 | }, | ||
| 868 | { | ||
| 624 | .func = NULL, | 869 | .func = NULL, |
| 625 | }, | 870 | }, |
| 626 | }; | 871 | }; |
diff --git a/tools/perf/builtin-top.c b/tools/perf/builtin-top.c index f2f3f4937aa..a43433f0830 100644 --- a/tools/perf/builtin-top.c +++ b/tools/perf/builtin-top.c | |||
| @@ -990,7 +990,7 @@ static const char * const top_usage[] = { | |||
| 990 | static const struct option options[] = { | 990 | static const struct option options[] = { |
| 991 | OPT_CALLBACK('e', "event", &top.evlist, "event", | 991 | OPT_CALLBACK('e', "event", &top.evlist, "event", |
| 992 | "event selector. use 'perf list' to list available events", | 992 | "event selector. use 'perf list' to list available events", |
| 993 | parse_events), | 993 | parse_events_option), |
| 994 | OPT_INTEGER('c', "count", &default_interval, | 994 | OPT_INTEGER('c', "count", &default_interval, |
| 995 | "event period to sample"), | 995 | "event period to sample"), |
| 996 | OPT_INTEGER('p', "pid", &top.target_pid, | 996 | OPT_INTEGER('p', "pid", &top.target_pid, |
diff --git a/tools/perf/util/callchain.h b/tools/perf/util/callchain.h index 1a79df9f739..9b4ff16cac9 100644 --- a/tools/perf/util/callchain.h +++ b/tools/perf/util/callchain.h | |||
| @@ -14,6 +14,11 @@ enum chain_mode { | |||
| 14 | CHAIN_GRAPH_REL | 14 | CHAIN_GRAPH_REL |
| 15 | }; | 15 | }; |
| 16 | 16 | ||
| 17 | enum chain_order { | ||
| 18 | ORDER_CALLER, | ||
| 19 | ORDER_CALLEE | ||
| 20 | }; | ||
| 21 | |||
| 17 | struct callchain_node { | 22 | struct callchain_node { |
| 18 | struct callchain_node *parent; | 23 | struct callchain_node *parent; |
| 19 | struct list_head siblings; | 24 | struct list_head siblings; |
| @@ -41,6 +46,7 @@ struct callchain_param { | |||
| 41 | u32 print_limit; | 46 | u32 print_limit; |
| 42 | double min_percent; | 47 | double min_percent; |
| 43 | sort_chain_func_t sort; | 48 | sort_chain_func_t sort; |
| 49 | enum chain_order order; | ||
| 44 | }; | 50 | }; |
| 45 | 51 | ||
| 46 | struct callchain_list { | 52 | struct callchain_list { |
diff --git a/tools/perf/util/dwarf-aux.c b/tools/perf/util/dwarf-aux.c new file mode 100644 index 00000000000..fddf40f30d3 --- /dev/null +++ b/tools/perf/util/dwarf-aux.c | |||
| @@ -0,0 +1,663 @@ | |||
| 1 | /* | ||
| 2 | * dwarf-aux.c : libdw auxiliary interfaces | ||
| 3 | * | ||
| 4 | * This program is free software; you can redistribute it and/or modify | ||
| 5 | * it under the terms of the GNU General Public License as published by | ||
| 6 | * the Free Software Foundation; either version 2 of the License, or | ||
| 7 | * (at your option) any later version. | ||
| 8 | * | ||
| 9 | * This program is distributed in the hope that it will be useful, | ||
| 10 | * but WITHOUT ANY WARRANTY; without even the implied warranty of | ||
| 11 | * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the | ||
| 12 | * GNU General Public License for more details. | ||
| 13 | * | ||
| 14 | * You should have received a copy of the GNU General Public License | ||
| 15 | * along with this program; if not, write to the Free Software | ||
| 16 | * Foundation, Inc., 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA. | ||
| 17 | * | ||
| 18 | */ | ||
| 19 | |||
| 20 | #include <stdbool.h> | ||
| 21 | #include "util.h" | ||
| 22 | #include "debug.h" | ||
| 23 | #include "dwarf-aux.h" | ||
| 24 | |||
| 25 | /** | ||
| 26 | * cu_find_realpath - Find the realpath of the target file | ||
| 27 | * @cu_die: A DIE(dwarf information entry) of CU(compilation Unit) | ||
| 28 | * @fname: The tail filename of the target file | ||
| 29 | * | ||
| 30 | * Find the real(long) path of @fname in @cu_die. | ||
| 31 | */ | ||
| 32 | const char *cu_find_realpath(Dwarf_Die *cu_die, const char *fname) | ||
| 33 | { | ||
| 34 | Dwarf_Files *files; | ||
| 35 | size_t nfiles, i; | ||
| 36 | const char *src = NULL; | ||
| 37 | int ret; | ||
| 38 | |||
| 39 | if (!fname) | ||
| 40 | return NULL; | ||
| 41 | |||
| 42 | ret = dwarf_getsrcfiles(cu_die, &files, &nfiles); | ||
| 43 | if (ret != 0) | ||
| 44 | return NULL; | ||
| 45 | |||
| 46 | for (i = 0; i < nfiles; i++) { | ||
| 47 | src = dwarf_filesrc(files, i, NULL, NULL); | ||
| 48 | if (strtailcmp(src, fname) == 0) | ||
| 49 | break; | ||
| 50 | } | ||
| 51 | if (i == nfiles) | ||
| 52 | return NULL; | ||
| 53 | return src; | ||
| 54 | } | ||
| 55 | |||
| 56 | /** | ||
| 57 | * cu_get_comp_dir - Get the path of compilation directory | ||
| 58 | * @cu_die: a CU DIE | ||
| 59 | * | ||
| 60 | * Get the path of compilation directory of given @cu_die. | ||
| 61 | * Since this depends on DW_AT_comp_dir, older gcc will not | ||
| 62 | * embedded it. In that case, this returns NULL. | ||
| 63 | */ | ||
| 64 | const char *cu_get_comp_dir(Dwarf_Die *cu_die) | ||
| 65 | { | ||
| 66 | Dwarf_Attribute attr; | ||
| 67 | if (dwarf_attr(cu_die, DW_AT_comp_dir, &attr) == NULL) | ||
| 68 | return NULL; | ||
| 69 | return dwarf_formstring(&attr); | ||
| 70 | } | ||
| 71 | |||
| 72 | /** | ||
| 73 | * cu_find_lineinfo - Get a line number and file name for given address | ||
| 74 | * @cu_die: a CU DIE | ||
| 75 | * @addr: An address | ||
| 76 | * @fname: a pointer which returns the file name string | ||
| 77 | * @lineno: a pointer which returns the line number | ||
| 78 | * | ||
| 79 | * Find a line number and file name for @addr in @cu_die. | ||
| 80 | */ | ||
| 81 | int cu_find_lineinfo(Dwarf_Die *cu_die, unsigned long addr, | ||
| 82 | const char **fname, int *lineno) | ||
| 83 | { | ||
| 84 | Dwarf_Line *line; | ||
| 85 | Dwarf_Addr laddr; | ||
| 86 | |||
| 87 | line = dwarf_getsrc_die(cu_die, (Dwarf_Addr)addr); | ||
| 88 | if (line && dwarf_lineaddr(line, &laddr) == 0 && | ||
| 89 | addr == (unsigned long)laddr && dwarf_lineno(line, lineno) == 0) { | ||
| 90 | *fname = dwarf_linesrc(line, NULL, NULL); | ||
| 91 | if (!*fname) | ||
| 92 | /* line number is useless without filename */ | ||
| 93 | *lineno = 0; | ||
| 94 | } | ||
| 95 | |||
| 96 | return *lineno ?: -ENOENT; | ||
| 97 | } | ||
| 98 | |||
| 99 | /** | ||
| 100 | * die_compare_name - Compare diename and tname | ||
| 101 | * @dw_die: a DIE | ||
| 102 | * @tname: a string of target name | ||
| 103 | * | ||
| 104 | * Compare the name of @dw_die and @tname. Return false if @dw_die has no name. | ||
| 105 | */ | ||
| 106 | bool die_compare_name(Dwarf_Die *dw_die, const char *tname) | ||
| 107 | { | ||
| 108 | const char *name; | ||
| 109 | name = dwarf_diename(dw_die); | ||
| 110 | return name ? (strcmp(tname, name) == 0) : false; | ||
| 111 | } | ||
| 112 | |||
| 113 | /** | ||
| 114 | * die_get_call_lineno - Get callsite line number of inline-function instance | ||
| 115 | * @in_die: a DIE of an inlined function instance | ||
| 116 | * | ||
| 117 | * Get call-site line number of @in_die. This means from where the inline | ||
| 118 | * function is called. | ||
| 119 | */ | ||
| 120 | int die_get_call_lineno(Dwarf_Die *in_die) | ||
| 121 | { | ||
| 122 | Dwarf_Attribute attr; | ||
| 123 | Dwarf_Word ret; | ||
| 124 | |||
| 125 | if (!dwarf_attr(in_die, DW_AT_call_line, &attr)) | ||
| 126 | return -ENOENT; | ||
| 127 | |||
| 128 | dwarf_formudata(&attr, &ret); | ||
| 129 | return (int)ret; | ||
| 130 | } | ||
| 131 | |||
| 132 | /** | ||
| 133 | * die_get_type - Get type DIE | ||
| 134 | * @vr_die: a DIE of a variable | ||
| 135 | * @die_mem: where to store a type DIE | ||
| 136 | * | ||
| 137 | * Get a DIE of the type of given variable (@vr_die), and store | ||
| 138 | * it to die_mem. Return NULL if fails to get a type DIE. | ||
| 139 | */ | ||
| 140 | Dwarf_Die *die_get_type(Dwarf_Die *vr_die, Dwarf_Die *die_mem) | ||
| 141 | { | ||
| 142 | Dwarf_Attribute attr; | ||
| 143 | |||
| 144 | if (dwarf_attr_integrate(vr_die, DW_AT_type, &attr) && | ||
| 145 | dwarf_formref_die(&attr, die_mem)) | ||
| 146 | return die_mem; | ||
| 147 | else | ||
| 148 | return NULL; | ||
| 149 | } | ||
| 150 | |||
| 151 | /* Get a type die, but skip qualifiers */ | ||
| 152 | static Dwarf_Die *__die_get_real_type(Dwarf_Die *vr_die, Dwarf_Die *die_mem) | ||
| 153 | { | ||
| 154 | int tag; | ||
| 155 | |||
| 156 | do { | ||
| 157 | vr_die = die_get_type(vr_die, die_mem); | ||
| 158 | if (!vr_die) | ||
| 159 | break; | ||
| 160 | tag = dwarf_tag(vr_die); | ||
| 161 | } while (tag == DW_TAG_const_type || | ||
| 162 | tag == DW_TAG_restrict_type || | ||
| 163 | tag == DW_TAG_volatile_type || | ||
| 164 | tag == DW_TAG_shared_type); | ||
| 165 | |||
| 166 | return vr_die; | ||
| 167 | } | ||
| 168 | |||
| 169 | /** | ||
| 170 | * die_get_real_type - Get a type die, but skip qualifiers and typedef | ||
| 171 | * @vr_die: a DIE of a variable | ||
| 172 | * @die_mem: where to store a type DIE | ||
| 173 | * | ||
| 174 | * Get a DIE of the type of given variable (@vr_die), and store | ||
| 175 | * it to die_mem. Return NULL if fails to get a type DIE. | ||
| 176 | * If the type is qualifiers (e.g. const) or typedef, this skips it | ||
| 177 | * and tries to find real type (structure or basic types, e.g. int). | ||
| 178 | */ | ||
| 179 | Dwarf_Die *die_get_real_type(Dwarf_Die *vr_die, Dwarf_Die *die_mem) | ||
| 180 | { | ||
| 181 | do { | ||
| 182 | vr_die = __die_get_real_type(vr_die, die_mem); | ||
| 183 | } while (vr_die && dwarf_tag(vr_die) == DW_TAG_typedef); | ||
| 184 | |||
| 185 | return vr_die; | ||
| 186 | } | ||
| 187 | |||
| 188 | /* Get attribute and translate it as a udata */ | ||
| 189 | static int die_get_attr_udata(Dwarf_Die *tp_die, unsigned int attr_name, | ||
| 190 | Dwarf_Word *result) | ||
| 191 | { | ||
| 192 | Dwarf_Attribute attr; | ||
| 193 | |||
| 194 | if (dwarf_attr(tp_die, attr_name, &attr) == NULL || | ||
| 195 | dwarf_formudata(&attr, result) != 0) | ||
| 196 | return -ENOENT; | ||
| 197 | |||
| 198 | return 0; | ||
| 199 | } | ||
| 200 | |||
| 201 | /** | ||
| 202 | * die_is_signed_type - Check whether a type DIE is signed or not | ||
| 203 | * @tp_die: a DIE of a type | ||
| 204 | * | ||
| 205 | * Get the encoding of @tp_die and return true if the encoding | ||
| 206 | * is signed. | ||
| 207 | */ | ||
| 208 | bool die_is_signed_type(Dwarf_Die *tp_die) | ||
| 209 | { | ||
| 210 | Dwarf_Word ret; | ||
| 211 | |||
| 212 | if (die_get_attr_udata(tp_die, DW_AT_encoding, &ret)) | ||
| 213 | return false; | ||
| 214 | |||
| 215 | return (ret == DW_ATE_signed_char || ret == DW_ATE_signed || | ||
| 216 | ret == DW_ATE_signed_fixed); | ||
| 217 | } | ||
| 218 | |||
| 219 | /** | ||
| 220 | * die_get_data_member_location - Get the data-member offset | ||
| 221 | * @mb_die: a DIE of a member of a data structure | ||
| 222 | * @offs: The offset of the member in the data structure | ||
| 223 | * | ||
| 224 | * Get the offset of @mb_die in the data structure including @mb_die, and | ||
| 225 | * stores result offset to @offs. If any error occurs this returns errno. | ||
| 226 | */ | ||
| 227 | int die_get_data_member_location(Dwarf_Die *mb_die, Dwarf_Word *offs) | ||
| 228 | { | ||
| 229 | Dwarf_Attribute attr; | ||
| 230 | Dwarf_Op *expr; | ||
| 231 | size_t nexpr; | ||
| 232 | int ret; | ||
| 233 | |||
| 234 | if (dwarf_attr(mb_die, DW_AT_data_member_location, &attr) == NULL) | ||
| 235 | return -ENOENT; | ||
| 236 | |||
| 237 | if (dwarf_formudata(&attr, offs) != 0) { | ||
| 238 | /* DW_AT_data_member_location should be DW_OP_plus_uconst */ | ||
| 239 | ret = dwarf_getlocation(&attr, &expr, &nexpr); | ||
| 240 | if (ret < 0 || nexpr == 0) | ||
| 241 | return -ENOENT; | ||
| 242 | |||
| 243 | if (expr[0].atom != DW_OP_plus_uconst || nexpr != 1) { | ||
| 244 | pr_debug("Unable to get offset:Unexpected OP %x (%zd)\n", | ||
| 245 | expr[0].atom, nexpr); | ||
| 246 | return -ENOTSUP; | ||
| 247 | } | ||
| 248 | *offs = (Dwarf_Word)expr[0].number; | ||
| 249 | } | ||
| 250 | return 0; | ||
| 251 | } | ||
| 252 | |||
| 253 | /** | ||
| 254 | * die_find_child - Generic DIE search function in DIE tree | ||
| 255 | * @rt_die: a root DIE | ||
| 256 | * @callback: a callback function | ||
| 257 | * @data: a user data passed to the callback function | ||
| 258 | * @die_mem: a buffer for result DIE | ||
| 259 | * | ||
| 260 | * Trace DIE tree from @rt_die and call @callback for each child DIE. | ||
| 261 | * If @callback returns DIE_FIND_CB_END, this stores the DIE into | ||
| 262 | * @die_mem and returns it. If @callback returns DIE_FIND_CB_CONTINUE, | ||
| 263 | * this continues to trace the tree. Optionally, @callback can return | ||
| 264 | * DIE_FIND_CB_CHILD and DIE_FIND_CB_SIBLING, those means trace only | ||
| 265 | * the children and trace only the siblings respectively. | ||
| 266 | * Returns NULL if @callback can't find any appropriate DIE. | ||
| 267 | */ | ||
| 268 | Dwarf_Die *die_find_child(Dwarf_Die *rt_die, | ||
| 269 | int (*callback)(Dwarf_Die *, void *), | ||
| 270 | void *data, Dwarf_Die *die_mem) | ||
| 271 | { | ||
| 272 | Dwarf_Die child_die; | ||
| 273 | int ret; | ||
| 274 | |||
| 275 | ret = dwarf_child(rt_die, die_mem); | ||
| 276 | if (ret != 0) | ||
| 277 | return NULL; | ||
| 278 | |||
| 279 | do { | ||
| 280 | ret = callback(die_mem, data); | ||
| 281 | if (ret == DIE_FIND_CB_END) | ||
| 282 | return die_mem; | ||
| 283 | |||
| 284 | if ((ret & DIE_FIND_CB_CHILD) && | ||
| 285 | die_find_child(die_mem, callback, data, &child_die)) { | ||
| 286 | memcpy(die_mem, &child_die, sizeof(Dwarf_Die)); | ||
| 287 | return die_mem; | ||
| 288 | } | ||
| 289 | } while ((ret & DIE_FIND_CB_SIBLING) && | ||
| 290 | dwarf_siblingof(die_mem, die_mem) == 0); | ||
| 291 | |||
| 292 | return NULL; | ||
| 293 | } | ||
| 294 | |||
| 295 | struct __addr_die_search_param { | ||
| 296 | Dwarf_Addr addr; | ||
| 297 | Dwarf_Die *die_mem; | ||
| 298 | }; | ||
| 299 | |||
| 300 | /* die_find callback for non-inlined function search */ | ||
| 301 | static int __die_search_func_cb(Dwarf_Die *fn_die, void *data) | ||
| 302 | { | ||
| 303 | struct __addr_die_search_param *ad = data; | ||
| 304 | |||
| 305 | if (dwarf_tag(fn_die) == DW_TAG_subprogram && | ||
| 306 | dwarf_haspc(fn_die, ad->addr)) { | ||
| 307 | memcpy(ad->die_mem, fn_die, sizeof(Dwarf_Die)); | ||
| 308 | return DWARF_CB_ABORT; | ||
| 309 | } | ||
| 310 | return DWARF_CB_OK; | ||
| 311 | } | ||
| 312 | |||
| 313 | /** | ||
| 314 | * die_find_realfunc - Search a non-inlined function at given address | ||
| 315 | * @cu_die: a CU DIE which including @addr | ||
| 316 | * @addr: target address | ||
| 317 | * @die_mem: a buffer for result DIE | ||
| 318 | * | ||
| 319 | * Search a non-inlined function DIE which includes @addr. Stores the | ||
| 320 | * DIE to @die_mem and returns it if found. Returns NULl if failed. | ||
| 321 | */ | ||
| 322 | Dwarf_Die *die_find_realfunc(Dwarf_Die *cu_die, Dwarf_Addr addr, | ||
| 323 | Dwarf_Die *die_mem) | ||
| 324 | { | ||
| 325 | struct __addr_die_search_param ad; | ||
| 326 | ad.addr = addr; | ||
| 327 | ad.die_mem = die_mem; | ||
| 328 | /* dwarf_getscopes can't find subprogram. */ | ||
| 329 | if (!dwarf_getfuncs(cu_die, __die_search_func_cb, &ad, 0)) | ||
| 330 | return NULL; | ||
| 331 | else | ||
| 332 | return die_mem; | ||
| 333 | } | ||
| 334 | |||
| 335 | /* die_find callback for inline function search */ | ||
| 336 | static int __die_find_inline_cb(Dwarf_Die *die_mem, void *data) | ||
| 337 | { | ||
| 338 | Dwarf_Addr *addr = data; | ||
| 339 | |||
| 340 | if (dwarf_tag(die_mem) == DW_TAG_inlined_subroutine && | ||
| 341 | dwarf_haspc(die_mem, *addr)) | ||
| 342 | return DIE_FIND_CB_END; | ||
| 343 | |||
| 344 | return DIE_FIND_CB_CONTINUE; | ||
| 345 | } | ||
| 346 | |||
| 347 | /** | ||
| 348 | * die_find_inlinefunc - Search an inlined function at given address | ||
| 349 | * @cu_die: a CU DIE which including @addr | ||
| 350 | * @addr: target address | ||
| 351 | * @die_mem: a buffer for result DIE | ||
| 352 | * | ||
| 353 | * Search an inlined function DIE which includes @addr. Stores the | ||
| 354 | * DIE to @die_mem and returns it if found. Returns NULl if failed. | ||
| 355 | * If several inlined functions are expanded recursively, this trace | ||
| 356 | * it and returns deepest one. | ||
| 357 | */ | ||
| 358 | Dwarf_Die *die_find_inlinefunc(Dwarf_Die *sp_die, Dwarf_Addr addr, | ||
| 359 | Dwarf_Die *die_mem) | ||
| 360 | { | ||
| 361 | Dwarf_Die tmp_die; | ||
| 362 | |||
| 363 | sp_die = die_find_child(sp_die, __die_find_inline_cb, &addr, &tmp_die); | ||
| 364 | if (!sp_die) | ||
| 365 | return NULL; | ||
| 366 | |||
| 367 | /* Inlined function could be recursive. Trace it until fail */ | ||
| 368 | while (sp_die) { | ||
| 369 | memcpy(die_mem, sp_die, sizeof(Dwarf_Die)); | ||
| 370 | sp_die = die_find_child(sp_die, __die_find_inline_cb, &addr, | ||
| 371 | &tmp_die); | ||
| 372 | } | ||
| 373 | |||
| 374 | return die_mem; | ||
| 375 | } | ||
| 376 | |||
| 377 | /* Line walker internal parameters */ | ||
| 378 | struct __line_walk_param { | ||
| 379 | const char *fname; | ||
| 380 | line_walk_callback_t callback; | ||
| 381 | void *data; | ||
| 382 | int retval; | ||
| 383 | }; | ||
| 384 | |||
| 385 | static int __die_walk_funclines_cb(Dwarf_Die *in_die, void *data) | ||
| 386 | { | ||
| 387 | struct __line_walk_param *lw = data; | ||
| 388 | Dwarf_Addr addr; | ||
| 389 | int lineno; | ||
| 390 | |||
| 391 | if (dwarf_tag(in_die) == DW_TAG_inlined_subroutine) { | ||
| 392 | lineno = die_get_call_lineno(in_die); | ||
| 393 | if (lineno > 0 && dwarf_entrypc(in_die, &addr) == 0) { | ||
| 394 | lw->retval = lw->callback(lw->fname, lineno, addr, | ||
| 395 | lw->data); | ||
| 396 | if (lw->retval != 0) | ||
| 397 | return DIE_FIND_CB_END; | ||
| 398 | } | ||
| 399 | } | ||
| 400 | return DIE_FIND_CB_SIBLING; | ||
| 401 | } | ||
| 402 | |||
| 403 | /* Walk on lines of blocks included in given DIE */ | ||
| 404 | static int __die_walk_funclines(Dwarf_Die *sp_die, | ||
| 405 | line_walk_callback_t callback, void *data) | ||
| 406 | { | ||
| 407 | struct __line_walk_param lw = { | ||
| 408 | .callback = callback, | ||
| 409 | .data = data, | ||
| 410 | .retval = 0, | ||
| 411 | }; | ||
| 412 | Dwarf_Die die_mem; | ||
| 413 | Dwarf_Addr addr; | ||
| 414 | int lineno; | ||
| 415 | |||
| 416 | /* Handle function declaration line */ | ||
| 417 | lw.fname = dwarf_decl_file(sp_die); | ||
| 418 | if (lw.fname && dwarf_decl_line(sp_die, &lineno) == 0 && | ||
| 419 | dwarf_entrypc(sp_die, &addr) == 0) { | ||
| 420 | lw.retval = callback(lw.fname, lineno, addr, data); | ||
| 421 | if (lw.retval != 0) | ||
| 422 | goto done; | ||
| 423 | } | ||
| 424 | die_find_child(sp_die, __die_walk_funclines_cb, &lw, &die_mem); | ||
| 425 | done: | ||
| 426 | return lw.retval; | ||
| 427 | } | ||
| 428 | |||
| 429 | static int __die_walk_culines_cb(Dwarf_Die *sp_die, void *data) | ||
| 430 | { | ||
| 431 | struct __line_walk_param *lw = data; | ||
| 432 | |||
| 433 | lw->retval = __die_walk_funclines(sp_die, lw->callback, lw->data); | ||
| 434 | if (lw->retval != 0) | ||
| 435 | return DWARF_CB_ABORT; | ||
| 436 | |||
| 437 | return DWARF_CB_OK; | ||
| 438 | } | ||
| 439 | |||
| 440 | /** | ||
| 441 | * die_walk_lines - Walk on lines inside given DIE | ||
| 442 | * @rt_die: a root DIE (CU or subprogram) | ||
| 443 | * @callback: callback routine | ||
| 444 | * @data: user data | ||
| 445 | * | ||
| 446 | * Walk on all lines inside given @rt_die and call @callback on each line. | ||
| 447 | * If the @rt_die is a function, walk only on the lines inside the function, | ||
| 448 | * otherwise @rt_die must be a CU DIE. | ||
| 449 | * Note that this walks not only dwarf line list, but also function entries | ||
| 450 | * and inline call-site. | ||
| 451 | */ | ||
| 452 | int die_walk_lines(Dwarf_Die *rt_die, line_walk_callback_t callback, void *data) | ||
| 453 | { | ||
| 454 | Dwarf_Lines *lines; | ||
| 455 | Dwarf_Line *line; | ||
| 456 | Dwarf_Addr addr; | ||
| 457 | const char *fname; | ||
| 458 | int lineno, ret = 0; | ||
| 459 | Dwarf_Die die_mem, *cu_die; | ||
| 460 | size_t nlines, i; | ||
| 461 | |||
| 462 | /* Get the CU die */ | ||
| 463 | if (dwarf_tag(rt_die) == DW_TAG_subprogram) | ||
| 464 | cu_die = dwarf_diecu(rt_die, &die_mem, NULL, NULL); | ||
| 465 | else | ||
| 466 | cu_die = rt_die; | ||
| 467 | if (!cu_die) { | ||
| 468 | pr_debug2("Failed to get CU from subprogram\n"); | ||
| 469 | return -EINVAL; | ||
| 470 | } | ||
| 471 | |||
| 472 | /* Get lines list in the CU */ | ||
| 473 | if (dwarf_getsrclines(cu_die, &lines, &nlines) != 0) { | ||
| 474 | pr_debug2("Failed to get source lines on this CU.\n"); | ||
| 475 | return -ENOENT; | ||
| 476 | } | ||
| 477 | pr_debug2("Get %zd lines from this CU\n", nlines); | ||
| 478 | |||
| 479 | /* Walk on the lines on lines list */ | ||
| 480 | for (i = 0; i < nlines; i++) { | ||
| 481 | line = dwarf_onesrcline(lines, i); | ||
| 482 | if (line == NULL || | ||
| 483 | dwarf_lineno(line, &lineno) != 0 || | ||
| 484 | dwarf_lineaddr(line, &addr) != 0) { | ||
| 485 | pr_debug2("Failed to get line info. " | ||
| 486 | "Possible error in debuginfo.\n"); | ||
| 487 | continue; | ||
| 488 | } | ||
| 489 | /* Filter lines based on address */ | ||
| 490 | if (rt_die != cu_die) | ||
| 491 | /* | ||
| 492 | * Address filtering | ||
| 493 | * The line is included in given function, and | ||
| 494 | * no inline block includes it. | ||
| 495 | */ | ||
| 496 | if (!dwarf_haspc(rt_die, addr) || | ||
| 497 | die_find_inlinefunc(rt_die, addr, &die_mem)) | ||
| 498 | continue; | ||
| 499 | /* Get source line */ | ||
| 500 | fname = dwarf_linesrc(line, NULL, NULL); | ||
| 501 | |||
| 502 | ret = callback(fname, lineno, addr, data); | ||
| 503 | if (ret != 0) | ||
| 504 | return ret; | ||
| 505 | } | ||
| 506 | |||
| 507 | /* | ||
| 508 | * Dwarf lines doesn't include function declarations and inlined | ||
| 509 | * subroutines. We have to check functions list or given function. | ||
| 510 | */ | ||
| 511 | if (rt_die != cu_die) | ||
| 512 | ret = __die_walk_funclines(rt_die, callback, data); | ||
| 513 | else { | ||
| 514 | struct __line_walk_param param = { | ||
| 515 | .callback = callback, | ||
| 516 | .data = data, | ||
| 517 | .retval = 0, | ||
| 518 | }; | ||
| 519 | dwarf_getfuncs(cu_die, __die_walk_culines_cb, ¶m, 0); | ||
| 520 | ret = param.retval; | ||
| 521 | } | ||
| 522 | |||
| 523 | return ret; | ||
| 524 | } | ||
| 525 | |||
| 526 | struct __find_variable_param { | ||
| 527 | const char *name; | ||
| 528 | Dwarf_Addr addr; | ||
| 529 | }; | ||
| 530 | |||
| 531 | static int __die_find_variable_cb(Dwarf_Die *die_mem, void *data) | ||
| 532 | { | ||
| 533 | struct __find_variable_param *fvp = data; | ||
| 534 | int tag; | ||
| 535 | |||
| 536 | tag = dwarf_tag(die_mem); | ||
| 537 | if ((tag == DW_TAG_formal_parameter || | ||
| 538 | tag == DW_TAG_variable) && | ||
| 539 | die_compare_name(die_mem, fvp->name)) | ||
| 540 | return DIE_FIND_CB_END; | ||
| 541 | |||
| 542 | if (dwarf_haspc(die_mem, fvp->addr)) | ||
| 543 | return DIE_FIND_CB_CONTINUE; | ||
| 544 | else | ||
| 545 | return DIE_FIND_CB_SIBLING; | ||
| 546 | } | ||
| 547 | |||
| 548 | /** | ||
| 549 | * die_find_variable_at - Find a given name variable at given address | ||
| 550 | * @sp_die: a function DIE | ||
| 551 | * @name: variable name | ||
| 552 | * @addr: address | ||
| 553 | * @die_mem: a buffer for result DIE | ||
| 554 | * | ||
| 555 | * Find a variable DIE called @name at @addr in @sp_die. | ||
| 556 | */ | ||
| 557 | Dwarf_Die *die_find_variable_at(Dwarf_Die *sp_die, const char *name, | ||
| 558 | Dwarf_Addr addr, Dwarf_Die *die_mem) | ||
| 559 | { | ||
| 560 | struct __find_variable_param fvp = { .name = name, .addr = addr}; | ||
| 561 | |||
| 562 | return die_find_child(sp_die, __die_find_variable_cb, (void *)&fvp, | ||
| 563 | die_mem); | ||
| 564 | } | ||
| 565 | |||
| 566 | static int __die_find_member_cb(Dwarf_Die *die_mem, void *data) | ||
| 567 | { | ||
| 568 | const char *name = data; | ||
| 569 | |||
| 570 | if ((dwarf_tag(die_mem) == DW_TAG_member) && | ||
| 571 | die_compare_name(die_mem, name)) | ||
| 572 | return DIE_FIND_CB_END; | ||
| 573 | |||
| 574 | return DIE_FIND_CB_SIBLING; | ||
| 575 | } | ||
| 576 | |||
| 577 | /** | ||
| 578 | * die_find_member - Find a given name member in a data structure | ||
| 579 | * @st_die: a data structure type DIE | ||
| 580 | * @name: member name | ||
| 581 | * @die_mem: a buffer for result DIE | ||
| 582 | * | ||
| 583 | * Find a member DIE called @name in @st_die. | ||
| 584 | */ | ||
| 585 | Dwarf_Die *die_find_member(Dwarf_Die *st_die, const char *name, | ||
| 586 | Dwarf_Die *die_mem) | ||
| 587 | { | ||
| 588 | return die_find_child(st_die, __die_find_member_cb, (void *)name, | ||
| 589 | die_mem); | ||
| 590 | } | ||
| 591 | |||
| 592 | /** | ||
| 593 | * die_get_typename - Get the name of given variable DIE | ||
| 594 | * @vr_die: a variable DIE | ||
| 595 | * @buf: a buffer for result type name | ||
| 596 | * @len: a max-length of @buf | ||
| 597 | * | ||
| 598 | * Get the name of @vr_die and stores it to @buf. Return the actual length | ||
| 599 | * of type name if succeeded. Return -E2BIG if @len is not enough long, and | ||
| 600 | * Return -ENOENT if failed to find type name. | ||
| 601 | * Note that the result will stores typedef name if possible, and stores | ||
| 602 | * "*(function_type)" if the type is a function pointer. | ||
| 603 | */ | ||
| 604 | int die_get_typename(Dwarf_Die *vr_die, char *buf, int len) | ||
| 605 | { | ||
| 606 | Dwarf_Die type; | ||
| 607 | int tag, ret, ret2; | ||
| 608 | const char *tmp = ""; | ||
| 609 | |||
| 610 | if (__die_get_real_type(vr_die, &type) == NULL) | ||
| 611 | return -ENOENT; | ||
| 612 | |||
| 613 | tag = dwarf_tag(&type); | ||
| 614 | if (tag == DW_TAG_array_type || tag == DW_TAG_pointer_type) | ||
| 615 | tmp = "*"; | ||
| 616 | else if (tag == DW_TAG_subroutine_type) { | ||
| 617 | /* Function pointer */ | ||
| 618 | ret = snprintf(buf, len, "(function_type)"); | ||
| 619 | return (ret >= len) ? -E2BIG : ret; | ||
| 620 | } else { | ||
| 621 | if (!dwarf_diename(&type)) | ||
| 622 | return -ENOENT; | ||
| 623 | if (tag == DW_TAG_union_type) | ||
| 624 | tmp = "union "; | ||
| 625 | else if (tag == DW_TAG_structure_type) | ||
| 626 | tmp = "struct "; | ||
| 627 | /* Write a base name */ | ||
| 628 | ret = snprintf(buf, len, "%s%s", tmp, dwarf_diename(&type)); | ||
| 629 | return (ret >= len) ? -E2BIG : ret; | ||
| 630 | } | ||
| 631 | ret = die_get_typename(&type, buf, len); | ||
| 632 | if (ret > 0) { | ||
| 633 | ret2 = snprintf(buf + ret, len - ret, "%s", tmp); | ||
| 634 | ret = (ret2 >= len - ret) ? -E2BIG : ret2 + ret; | ||
| 635 | } | ||
| 636 | return ret; | ||
| 637 | } | ||
| 638 | |||
| 639 | /** | ||
| 640 | * die_get_varname - Get the name and type of given variable DIE | ||
| 641 | * @vr_die: a variable DIE | ||
| 642 | * @buf: a buffer for type and variable name | ||
| 643 | * @len: the max-length of @buf | ||
| 644 | * | ||
| 645 | * Get the name and type of @vr_die and stores it in @buf as "type\tname". | ||
| 646 | */ | ||
| 647 | int die_get_varname(Dwarf_Die *vr_die, char *buf, int len) | ||
| 648 | { | ||
| 649 | int ret, ret2; | ||
| 650 | |||
| 651 | ret = die_get_typename(vr_die, buf, len); | ||
| 652 | if (ret < 0) { | ||
| 653 | pr_debug("Failed to get type, make it unknown.\n"); | ||
| 654 | ret = snprintf(buf, len, "(unknown_type)"); | ||
| 655 | } | ||
| 656 | if (ret > 0) { | ||
| 657 | ret2 = snprintf(buf + ret, len - ret, "\t%s", | ||
| 658 | dwarf_diename(vr_die)); | ||
| 659 | ret = (ret2 >= len - ret) ? -E2BIG : ret2 + ret; | ||
| 660 | } | ||
| 661 | return ret; | ||
| 662 | } | ||
| 663 | |||
diff --git a/tools/perf/util/dwarf-aux.h b/tools/perf/util/dwarf-aux.h new file mode 100644 index 00000000000..bc3b21167e7 --- /dev/null +++ b/tools/perf/util/dwarf-aux.h | |||
| @@ -0,0 +1,100 @@ | |||
| 1 | #ifndef _DWARF_AUX_H | ||
| 2 | #define _DWARF_AUX_H | ||
| 3 | /* | ||
| 4 | * dwarf-aux.h : libdw auxiliary interfaces | ||
| 5 | * | ||
| 6 | * This program is free software; you can redistribute it and/or modify | ||
| 7 | * it under the terms of the GNU General Public License as published by | ||
| 8 | * the Free Software Foundation; either version 2 of the License, or | ||
| 9 | * (at your option) any later version. | ||
| 10 | * | ||
| 11 | * This program is distributed in the hope that it will be useful, | ||
| 12 | * but WITHOUT ANY WARRANTY; without even the implied warranty of | ||
| 13 | * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the | ||
| 14 | * GNU General Public License for more details. | ||
| 15 | * | ||
| 16 | * You should have received a copy of the GNU General Public License | ||
| 17 | * along with this program; if not, write to the Free Software | ||
| 18 | * Foundation, Inc., 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA. | ||
| 19 | * | ||
| 20 | */ | ||
| 21 | |||
| 22 | #include <dwarf.h> | ||
| 23 | #include <elfutils/libdw.h> | ||
| 24 | #include <elfutils/libdwfl.h> | ||
| 25 | #include <elfutils/version.h> | ||
| 26 | |||
| 27 | /* Find the realpath of the target file */ | ||
| 28 | extern const char *cu_find_realpath(Dwarf_Die *cu_die, const char *fname); | ||
| 29 | |||
| 30 | /* Get DW_AT_comp_dir (should be NULL with older gcc) */ | ||
| 31 | extern const char *cu_get_comp_dir(Dwarf_Die *cu_die); | ||
| 32 | |||
| 33 | /* Get a line number and file name for given address */ | ||
| 34 | extern int cu_find_lineinfo(Dwarf_Die *cudie, unsigned long addr, | ||
| 35 | const char **fname, int *lineno); | ||
| 36 | |||
| 37 | /* Compare diename and tname */ | ||
| 38 | extern bool die_compare_name(Dwarf_Die *dw_die, const char *tname); | ||
| 39 | |||
| 40 | /* Get callsite line number of inline-function instance */ | ||
| 41 | extern int die_get_call_lineno(Dwarf_Die *in_die); | ||
| 42 | |||
| 43 | /* Get type die */ | ||
| 44 | extern Dwarf_Die *die_get_type(Dwarf_Die *vr_die, Dwarf_Die *die_mem); | ||
| 45 | |||
| 46 | /* Get a type die, but skip qualifiers and typedef */ | ||
| 47 | extern Dwarf_Die *die_get_real_type(Dwarf_Die *vr_die, Dwarf_Die *die_mem); | ||
| 48 | |||
| 49 | /* Check whether the DIE is signed or not */ | ||
| 50 | extern bool die_is_signed_type(Dwarf_Die *tp_die); | ||
| 51 | |||
| 52 | /* Get data_member_location offset */ | ||
| 53 | extern int die_get_data_member_location(Dwarf_Die *mb_die, Dwarf_Word *offs); | ||
| 54 | |||
| 55 | /* Return values for die_find_child() callbacks */ | ||
| 56 | enum { | ||
| 57 | DIE_FIND_CB_END = 0, /* End of Search */ | ||
| 58 | DIE_FIND_CB_CHILD = 1, /* Search only children */ | ||
| 59 | DIE_FIND_CB_SIBLING = 2, /* Search only siblings */ | ||
| 60 | DIE_FIND_CB_CONTINUE = 3, /* Search children and siblings */ | ||
| 61 | }; | ||
| 62 | |||
| 63 | /* Search child DIEs */ | ||
| 64 | extern Dwarf_Die *die_find_child(Dwarf_Die *rt_die, | ||
| 65 | int (*callback)(Dwarf_Die *, void *), | ||
| 66 | void *data, Dwarf_Die *die_mem); | ||
| 67 | |||
| 68 | /* Search a non-inlined function including given address */ | ||
| 69 | extern Dwarf_Die *die_find_realfunc(Dwarf_Die *cu_die, Dwarf_Addr addr, | ||
| 70 | Dwarf_Die *die_mem); | ||
| 71 | |||
| 72 | /* Search an inlined function including given address */ | ||
| 73 | extern Dwarf_Die *die_find_inlinefunc(Dwarf_Die *sp_die, Dwarf_Addr addr, | ||
| 74 | Dwarf_Die *die_mem); | ||
| 75 | |||
| 76 | /* Walker on lines (Note: line number will not be sorted) */ | ||
| 77 | typedef int (* line_walk_callback_t) (const char *fname, int lineno, | ||
| 78 | Dwarf_Addr addr, void *data); | ||
| 79 | |||
| 80 | /* | ||
| 81 | * Walk on lines inside given DIE. If the DIE is a subprogram, walk only on | ||
| 82 | * the lines inside the subprogram, otherwise the DIE must be a CU DIE. | ||
| 83 | */ | ||
| 84 | extern int die_walk_lines(Dwarf_Die *rt_die, line_walk_callback_t callback, | ||
| 85 | void *data); | ||
| 86 | |||
| 87 | /* Find a variable called 'name' at given address */ | ||
| 88 | extern Dwarf_Die *die_find_variable_at(Dwarf_Die *sp_die, const char *name, | ||
| 89 | Dwarf_Addr addr, Dwarf_Die *die_mem); | ||
| 90 | |||
| 91 | /* Find a member called 'name' */ | ||
| 92 | extern Dwarf_Die *die_find_member(Dwarf_Die *st_die, const char *name, | ||
| 93 | Dwarf_Die *die_mem); | ||
| 94 | |||
| 95 | /* Get the name of given variable DIE */ | ||
| 96 | extern int die_get_typename(Dwarf_Die *vr_die, char *buf, int len); | ||
| 97 | |||
| 98 | /* Get the name and type of given variable DIE, stored as "type\tname" */ | ||
| 99 | extern int die_get_varname(Dwarf_Die *vr_die, char *buf, int len); | ||
| 100 | #endif | ||
diff --git a/tools/perf/util/evsel.c b/tools/perf/util/evsel.c index 0239eb87b23..a03a36b7908 100644 --- a/tools/perf/util/evsel.c +++ b/tools/perf/util/evsel.c | |||
| @@ -377,6 +377,7 @@ int perf_event__parse_sample(const union perf_event *event, u64 type, | |||
| 377 | array++; | 377 | array++; |
| 378 | } | 378 | } |
| 379 | 379 | ||
| 380 | data->addr = 0; | ||
| 380 | if (type & PERF_SAMPLE_ADDR) { | 381 | if (type & PERF_SAMPLE_ADDR) { |
| 381 | data->addr = *array; | 382 | data->addr = *array; |
| 382 | array++; | 383 | array++; |
diff --git a/tools/perf/util/evsel.h b/tools/perf/util/evsel.h index 7e9366e4490..e9a31554e26 100644 --- a/tools/perf/util/evsel.h +++ b/tools/perf/util/evsel.h | |||
| @@ -61,6 +61,7 @@ struct perf_evsel { | |||
| 61 | off_t id_offset; | 61 | off_t id_offset; |
| 62 | }; | 62 | }; |
| 63 | struct cgroup_sel *cgrp; | 63 | struct cgroup_sel *cgrp; |
| 64 | bool supported; | ||
| 64 | }; | 65 | }; |
| 65 | 66 | ||
| 66 | struct cpu_map; | 67 | struct cpu_map; |
diff --git a/tools/perf/util/header.c b/tools/perf/util/header.c index afb0849fe53..cb2959a3fb4 100644 --- a/tools/perf/util/header.c +++ b/tools/perf/util/header.c | |||
| @@ -877,9 +877,12 @@ int perf_session__read_header(struct perf_session *session, int fd) | |||
| 877 | struct perf_evsel *evsel; | 877 | struct perf_evsel *evsel; |
| 878 | off_t tmp; | 878 | off_t tmp; |
| 879 | 879 | ||
| 880 | if (perf_header__getbuffer64(header, fd, &f_attr, sizeof(f_attr))) | 880 | if (readn(fd, &f_attr, sizeof(f_attr)) <= 0) |
| 881 | goto out_errno; | 881 | goto out_errno; |
| 882 | 882 | ||
| 883 | if (header->needs_swap) | ||
| 884 | perf_event__attr_swap(&f_attr.attr); | ||
| 885 | |||
| 883 | tmp = lseek(fd, 0, SEEK_CUR); | 886 | tmp = lseek(fd, 0, SEEK_CUR); |
| 884 | evsel = perf_evsel__new(&f_attr.attr, i); | 887 | evsel = perf_evsel__new(&f_attr.attr, i); |
| 885 | 888 | ||
diff --git a/tools/perf/util/hist.c b/tools/perf/util/hist.c index 627a02e03c5..677e1da6bb3 100644 --- a/tools/perf/util/hist.c +++ b/tools/perf/util/hist.c | |||
| @@ -14,7 +14,8 @@ enum hist_filter { | |||
| 14 | 14 | ||
| 15 | struct callchain_param callchain_param = { | 15 | struct callchain_param callchain_param = { |
| 16 | .mode = CHAIN_GRAPH_REL, | 16 | .mode = CHAIN_GRAPH_REL, |
| 17 | .min_percent = 0.5 | 17 | .min_percent = 0.5, |
| 18 | .order = ORDER_CALLEE | ||
| 18 | }; | 19 | }; |
| 19 | 20 | ||
| 20 | u16 hists__col_len(struct hists *self, enum hist_column col) | 21 | u16 hists__col_len(struct hists *self, enum hist_column col) |
| @@ -846,6 +847,9 @@ print_entries: | |||
| 846 | for (nd = rb_first(&self->entries); nd; nd = rb_next(nd)) { | 847 | for (nd = rb_first(&self->entries); nd; nd = rb_next(nd)) { |
| 847 | struct hist_entry *h = rb_entry(nd, struct hist_entry, rb_node); | 848 | struct hist_entry *h = rb_entry(nd, struct hist_entry, rb_node); |
| 848 | 849 | ||
| 850 | if (h->filtered) | ||
| 851 | continue; | ||
| 852 | |||
| 849 | if (show_displacement) { | 853 | if (show_displacement) { |
| 850 | if (h->pair != NULL) | 854 | if (h->pair != NULL) |
| 851 | displacement = ((long)h->pair->position - | 855 | displacement = ((long)h->pair->position - |
diff --git a/tools/perf/util/parse-events.c b/tools/perf/util/parse-events.c index 41982c373fa..4ea7e19f525 100644 --- a/tools/perf/util/parse-events.c +++ b/tools/perf/util/parse-events.c | |||
| @@ -86,22 +86,24 @@ static const char *sw_event_names[PERF_COUNT_SW_MAX] = { | |||
| 86 | 86 | ||
| 87 | #define MAX_ALIASES 8 | 87 | #define MAX_ALIASES 8 |
| 88 | 88 | ||
| 89 | static const char *hw_cache[][MAX_ALIASES] = { | 89 | static const char *hw_cache[PERF_COUNT_HW_CACHE_MAX][MAX_ALIASES] = { |
| 90 | { "L1-dcache", "l1-d", "l1d", "L1-data", }, | 90 | { "L1-dcache", "l1-d", "l1d", "L1-data", }, |
| 91 | { "L1-icache", "l1-i", "l1i", "L1-instruction", }, | 91 | { "L1-icache", "l1-i", "l1i", "L1-instruction", }, |
| 92 | { "LLC", "L2" }, | 92 | { "LLC", "L2", }, |
| 93 | { "dTLB", "d-tlb", "Data-TLB", }, | 93 | { "dTLB", "d-tlb", "Data-TLB", }, |
| 94 | { "iTLB", "i-tlb", "Instruction-TLB", }, | 94 | { "iTLB", "i-tlb", "Instruction-TLB", }, |
| 95 | { "branch", "branches", "bpu", "btb", "bpc", }, | 95 | { "branch", "branches", "bpu", "btb", "bpc", }, |
| 96 | { "node", }, | ||
| 96 | }; | 97 | }; |
| 97 | 98 | ||
| 98 | static const char *hw_cache_op[][MAX_ALIASES] = { | 99 | static const char *hw_cache_op[PERF_COUNT_HW_CACHE_OP_MAX][MAX_ALIASES] = { |
| 99 | { "load", "loads", "read", }, | 100 | { "load", "loads", "read", }, |
| 100 | { "store", "stores", "write", }, | 101 | { "store", "stores", "write", }, |
| 101 | { "prefetch", "prefetches", "speculative-read", "speculative-load", }, | 102 | { "prefetch", "prefetches", "speculative-read", "speculative-load", }, |
| 102 | }; | 103 | }; |
| 103 | 104 | ||
| 104 | static const char *hw_cache_result[][MAX_ALIASES] = { | 105 | static const char *hw_cache_result[PERF_COUNT_HW_CACHE_RESULT_MAX] |
| 106 | [MAX_ALIASES] = { | ||
| 105 | { "refs", "Reference", "ops", "access", }, | 107 | { "refs", "Reference", "ops", "access", }, |
| 106 | { "misses", "miss", }, | 108 | { "misses", "miss", }, |
| 107 | }; | 109 | }; |
| @@ -124,6 +126,7 @@ static unsigned long hw_cache_stat[C(MAX)] = { | |||
| 124 | [C(DTLB)] = (CACHE_READ | CACHE_WRITE | CACHE_PREFETCH), | 126 | [C(DTLB)] = (CACHE_READ | CACHE_WRITE | CACHE_PREFETCH), |
| 125 | [C(ITLB)] = (CACHE_READ), | 127 | [C(ITLB)] = (CACHE_READ), |
| 126 | [C(BPU)] = (CACHE_READ), | 128 | [C(BPU)] = (CACHE_READ), |
| 129 | [C(NODE)] = (CACHE_READ | CACHE_WRITE | CACHE_PREFETCH), | ||
| 127 | }; | 130 | }; |
| 128 | 131 | ||
| 129 | #define for_each_subsystem(sys_dir, sys_dirent, sys_next) \ | 132 | #define for_each_subsystem(sys_dir, sys_dirent, sys_next) \ |
| @@ -393,7 +396,7 @@ parse_generic_hw_event(const char **str, struct perf_event_attr *attr) | |||
| 393 | PERF_COUNT_HW_CACHE_OP_MAX); | 396 | PERF_COUNT_HW_CACHE_OP_MAX); |
| 394 | if (cache_op >= 0) { | 397 | if (cache_op >= 0) { |
| 395 | if (!is_cache_op_valid(cache_type, cache_op)) | 398 | if (!is_cache_op_valid(cache_type, cache_op)) |
| 396 | return 0; | 399 | return EVT_FAILED; |
| 397 | continue; | 400 | continue; |
| 398 | } | 401 | } |
| 399 | } | 402 | } |
| @@ -475,7 +478,7 @@ parse_single_tracepoint_event(char *sys_name, | |||
| 475 | /* sys + ':' + event + ':' + flags*/ | 478 | /* sys + ':' + event + ':' + flags*/ |
| 476 | #define MAX_EVOPT_LEN (MAX_EVENT_LENGTH * 2 + 2 + 128) | 479 | #define MAX_EVOPT_LEN (MAX_EVENT_LENGTH * 2 + 2 + 128) |
| 477 | static enum event_result | 480 | static enum event_result |
| 478 | parse_multiple_tracepoint_event(const struct option *opt, char *sys_name, | 481 | parse_multiple_tracepoint_event(struct perf_evlist *evlist, char *sys_name, |
| 479 | const char *evt_exp, char *flags) | 482 | const char *evt_exp, char *flags) |
| 480 | { | 483 | { |
| 481 | char evt_path[MAXPATHLEN]; | 484 | char evt_path[MAXPATHLEN]; |
| @@ -509,7 +512,7 @@ parse_multiple_tracepoint_event(const struct option *opt, char *sys_name, | |||
| 509 | if (len < 0) | 512 | if (len < 0) |
| 510 | return EVT_FAILED; | 513 | return EVT_FAILED; |
| 511 | 514 | ||
| 512 | if (parse_events(opt, event_opt, 0)) | 515 | if (parse_events(evlist, event_opt, 0)) |
| 513 | return EVT_FAILED; | 516 | return EVT_FAILED; |
| 514 | } | 517 | } |
| 515 | 518 | ||
| @@ -517,7 +520,7 @@ parse_multiple_tracepoint_event(const struct option *opt, char *sys_name, | |||
| 517 | } | 520 | } |
| 518 | 521 | ||
| 519 | static enum event_result | 522 | static enum event_result |
| 520 | parse_tracepoint_event(const struct option *opt, const char **strp, | 523 | parse_tracepoint_event(struct perf_evlist *evlist, const char **strp, |
| 521 | struct perf_event_attr *attr) | 524 | struct perf_event_attr *attr) |
| 522 | { | 525 | { |
| 523 | const char *evt_name; | 526 | const char *evt_name; |
| @@ -557,8 +560,8 @@ parse_tracepoint_event(const struct option *opt, const char **strp, | |||
| 557 | return EVT_FAILED; | 560 | return EVT_FAILED; |
| 558 | if (strpbrk(evt_name, "*?")) { | 561 | if (strpbrk(evt_name, "*?")) { |
| 559 | *strp += strlen(sys_name) + evt_length + 1; /* 1 == the ':' */ | 562 | *strp += strlen(sys_name) + evt_length + 1; /* 1 == the ':' */ |
| 560 | return parse_multiple_tracepoint_event(opt, sys_name, evt_name, | 563 | return parse_multiple_tracepoint_event(evlist, sys_name, |
| 561 | flags); | 564 | evt_name, flags); |
| 562 | } else { | 565 | } else { |
| 563 | return parse_single_tracepoint_event(sys_name, evt_name, | 566 | return parse_single_tracepoint_event(sys_name, evt_name, |
| 564 | evt_length, attr, strp); | 567 | evt_length, attr, strp); |
| @@ -778,12 +781,12 @@ parse_event_modifier(const char **strp, struct perf_event_attr *attr) | |||
| 778 | * Symbolic names are (almost) exactly matched. | 781 | * Symbolic names are (almost) exactly matched. |
| 779 | */ | 782 | */ |
| 780 | static enum event_result | 783 | static enum event_result |
| 781 | parse_event_symbols(const struct option *opt, const char **str, | 784 | parse_event_symbols(struct perf_evlist *evlist, const char **str, |
| 782 | struct perf_event_attr *attr) | 785 | struct perf_event_attr *attr) |
| 783 | { | 786 | { |
| 784 | enum event_result ret; | 787 | enum event_result ret; |
| 785 | 788 | ||
| 786 | ret = parse_tracepoint_event(opt, str, attr); | 789 | ret = parse_tracepoint_event(evlist, str, attr); |
| 787 | if (ret != EVT_FAILED) | 790 | if (ret != EVT_FAILED) |
| 788 | goto modifier; | 791 | goto modifier; |
| 789 | 792 | ||
| @@ -822,9 +825,8 @@ modifier: | |||
| 822 | return ret; | 825 | return ret; |
| 823 | } | 826 | } |
| 824 | 827 | ||
| 825 | int parse_events(const struct option *opt, const char *str, int unset __used) | 828 | int parse_events(struct perf_evlist *evlist , const char *str, int unset __used) |
| 826 | { | 829 | { |
| 827 | struct perf_evlist *evlist = *(struct perf_evlist **)opt->value; | ||
| 828 | struct perf_event_attr attr; | 830 | struct perf_event_attr attr; |
| 829 | enum event_result ret; | 831 | enum event_result ret; |
| 830 | const char *ostr; | 832 | const char *ostr; |
| @@ -832,7 +834,7 @@ int parse_events(const struct option *opt, const char *str, int unset __used) | |||
| 832 | for (;;) { | 834 | for (;;) { |
| 833 | ostr = str; | 835 | ostr = str; |
| 834 | memset(&attr, 0, sizeof(attr)); | 836 | memset(&attr, 0, sizeof(attr)); |
| 835 | ret = parse_event_symbols(opt, &str, &attr); | 837 | ret = parse_event_symbols(evlist, &str, &attr); |
| 836 | if (ret == EVT_FAILED) | 838 | if (ret == EVT_FAILED) |
| 837 | return -1; | 839 | return -1; |
| 838 | 840 | ||
| @@ -863,6 +865,13 @@ int parse_events(const struct option *opt, const char *str, int unset __used) | |||
| 863 | return 0; | 865 | return 0; |
| 864 | } | 866 | } |
| 865 | 867 | ||
| 868 | int parse_events_option(const struct option *opt, const char *str, | ||
| 869 | int unset __used) | ||
| 870 | { | ||
| 871 | struct perf_evlist *evlist = *(struct perf_evlist **)opt->value; | ||
| 872 | return parse_events(evlist, str, unset); | ||
| 873 | } | ||
| 874 | |||
| 866 | int parse_filter(const struct option *opt, const char *str, | 875 | int parse_filter(const struct option *opt, const char *str, |
| 867 | int unset __used) | 876 | int unset __used) |
| 868 | { | 877 | { |
diff --git a/tools/perf/util/parse-events.h b/tools/perf/util/parse-events.h index 746d3fcbfc2..2f8e375e038 100644 --- a/tools/perf/util/parse-events.h +++ b/tools/perf/util/parse-events.h | |||
| @@ -8,6 +8,7 @@ | |||
| 8 | 8 | ||
| 9 | struct list_head; | 9 | struct list_head; |
| 10 | struct perf_evsel; | 10 | struct perf_evsel; |
| 11 | struct perf_evlist; | ||
| 11 | 12 | ||
| 12 | struct option; | 13 | struct option; |
| 13 | 14 | ||
| @@ -24,7 +25,10 @@ const char *event_type(int type); | |||
| 24 | const char *event_name(struct perf_evsel *event); | 25 | const char *event_name(struct perf_evsel *event); |
| 25 | extern const char *__event_name(int type, u64 config); | 26 | extern const char *__event_name(int type, u64 config); |
| 26 | 27 | ||
| 27 | extern int parse_events(const struct option *opt, const char *str, int unset); | 28 | extern int parse_events_option(const struct option *opt, const char *str, |
| 29 | int unset); | ||
| 30 | extern int parse_events(struct perf_evlist *evlist, const char *str, | ||
| 31 | int unset); | ||
| 28 | extern int parse_filter(const struct option *opt, const char *str, int unset); | 32 | extern int parse_filter(const struct option *opt, const char *str, int unset); |
| 29 | 33 | ||
| 30 | #define EVENTS_HELP_MAX (128*1024) | 34 | #define EVENTS_HELP_MAX (128*1024) |
diff --git a/tools/perf/util/probe-event.c b/tools/perf/util/probe-event.c index f0223166e76..b82d54fa2c5 100644 --- a/tools/perf/util/probe-event.c +++ b/tools/perf/util/probe-event.c | |||
| @@ -117,6 +117,10 @@ static struct map *kernel_get_module_map(const char *module) | |||
| 117 | struct rb_node *nd; | 117 | struct rb_node *nd; |
| 118 | struct map_groups *grp = &machine.kmaps; | 118 | struct map_groups *grp = &machine.kmaps; |
| 119 | 119 | ||
| 120 | /* A file path -- this is an offline module */ | ||
| 121 | if (module && strchr(module, '/')) | ||
| 122 | return machine__new_module(&machine, 0, module); | ||
| 123 | |||
| 120 | if (!module) | 124 | if (!module) |
| 121 | module = "kernel"; | 125 | module = "kernel"; |
| 122 | 126 | ||
| @@ -170,16 +174,24 @@ const char *kernel_get_module_path(const char *module) | |||
| 170 | } | 174 | } |
| 171 | 175 | ||
| 172 | #ifdef DWARF_SUPPORT | 176 | #ifdef DWARF_SUPPORT |
| 173 | static int open_vmlinux(const char *module) | 177 | /* Open new debuginfo of given module */ |
| 178 | static struct debuginfo *open_debuginfo(const char *module) | ||
| 174 | { | 179 | { |
| 175 | const char *path = kernel_get_module_path(module); | 180 | const char *path; |
| 176 | if (!path) { | 181 | |
| 177 | pr_err("Failed to find path of %s module.\n", | 182 | /* A file path -- this is an offline module */ |
| 178 | module ?: "kernel"); | 183 | if (module && strchr(module, '/')) |
| 179 | return -ENOENT; | 184 | path = module; |
| 185 | else { | ||
| 186 | path = kernel_get_module_path(module); | ||
| 187 | |||
| 188 | if (!path) { | ||
| 189 | pr_err("Failed to find path of %s module.\n", | ||
| 190 | module ?: "kernel"); | ||
| 191 | return NULL; | ||
| 192 | } | ||
| 180 | } | 193 | } |
| 181 | pr_debug("Try to open %s\n", path); | 194 | return debuginfo__new(path); |
| 182 | return open(path, O_RDONLY); | ||
| 183 | } | 195 | } |
| 184 | 196 | ||
| 185 | /* | 197 | /* |
| @@ -193,13 +205,24 @@ static int kprobe_convert_to_perf_probe(struct probe_trace_point *tp, | |||
| 193 | struct map *map; | 205 | struct map *map; |
| 194 | u64 addr; | 206 | u64 addr; |
| 195 | int ret = -ENOENT; | 207 | int ret = -ENOENT; |
| 208 | struct debuginfo *dinfo; | ||
| 196 | 209 | ||
| 197 | sym = __find_kernel_function_by_name(tp->symbol, &map); | 210 | sym = __find_kernel_function_by_name(tp->symbol, &map); |
| 198 | if (sym) { | 211 | if (sym) { |
| 199 | addr = map->unmap_ip(map, sym->start + tp->offset); | 212 | addr = map->unmap_ip(map, sym->start + tp->offset); |
| 200 | pr_debug("try to find %s+%ld@%" PRIx64 "\n", tp->symbol, | 213 | pr_debug("try to find %s+%ld@%" PRIx64 "\n", tp->symbol, |
| 201 | tp->offset, addr); | 214 | tp->offset, addr); |
| 202 | ret = find_perf_probe_point((unsigned long)addr, pp); | 215 | |
| 216 | dinfo = debuginfo__new_online_kernel(addr); | ||
| 217 | if (dinfo) { | ||
| 218 | ret = debuginfo__find_probe_point(dinfo, | ||
| 219 | (unsigned long)addr, pp); | ||
| 220 | debuginfo__delete(dinfo); | ||
| 221 | } else { | ||
| 222 | pr_debug("Failed to open debuginfo at 0x%" PRIx64 "\n", | ||
| 223 | addr); | ||
| 224 | ret = -ENOENT; | ||
| 225 | } | ||
| 203 | } | 226 | } |
| 204 | if (ret <= 0) { | 227 | if (ret <= 0) { |
| 205 | pr_debug("Failed to find corresponding probes from " | 228 | pr_debug("Failed to find corresponding probes from " |
| @@ -214,30 +237,70 @@ static int kprobe_convert_to_perf_probe(struct probe_trace_point *tp, | |||
| 214 | return 0; | 237 | return 0; |
| 215 | } | 238 | } |
| 216 | 239 | ||
| 240 | static int add_module_to_probe_trace_events(struct probe_trace_event *tevs, | ||
| 241 | int ntevs, const char *module) | ||
| 242 | { | ||
| 243 | int i, ret = 0; | ||
| 244 | char *tmp; | ||
| 245 | |||
| 246 | if (!module) | ||
| 247 | return 0; | ||
| 248 | |||
| 249 | tmp = strrchr(module, '/'); | ||
| 250 | if (tmp) { | ||
| 251 | /* This is a module path -- get the module name */ | ||
| 252 | module = strdup(tmp + 1); | ||
| 253 | if (!module) | ||
| 254 | return -ENOMEM; | ||
| 255 | tmp = strchr(module, '.'); | ||
| 256 | if (tmp) | ||
| 257 | *tmp = '\0'; | ||
| 258 | tmp = (char *)module; /* For free() */ | ||
| 259 | } | ||
| 260 | |||
| 261 | for (i = 0; i < ntevs; i++) { | ||
| 262 | tevs[i].point.module = strdup(module); | ||
| 263 | if (!tevs[i].point.module) { | ||
| 264 | ret = -ENOMEM; | ||
| 265 | break; | ||
| 266 | } | ||
| 267 | } | ||
| 268 | |||
| 269 | if (tmp) | ||
| 270 | free(tmp); | ||
| 271 | |||
| 272 | return ret; | ||
| 273 | } | ||
| 274 | |||
| 217 | /* Try to find perf_probe_event with debuginfo */ | 275 | /* Try to find perf_probe_event with debuginfo */ |
| 218 | static int try_to_find_probe_trace_events(struct perf_probe_event *pev, | 276 | static int try_to_find_probe_trace_events(struct perf_probe_event *pev, |
| 219 | struct probe_trace_event **tevs, | 277 | struct probe_trace_event **tevs, |
| 220 | int max_tevs, const char *module) | 278 | int max_tevs, const char *module) |
| 221 | { | 279 | { |
| 222 | bool need_dwarf = perf_probe_event_need_dwarf(pev); | 280 | bool need_dwarf = perf_probe_event_need_dwarf(pev); |
| 223 | int fd, ntevs; | 281 | struct debuginfo *dinfo = open_debuginfo(module); |
| 282 | int ntevs, ret = 0; | ||
| 224 | 283 | ||
| 225 | fd = open_vmlinux(module); | 284 | if (!dinfo) { |
| 226 | if (fd < 0) { | ||
| 227 | if (need_dwarf) { | 285 | if (need_dwarf) { |
| 228 | pr_warning("Failed to open debuginfo file.\n"); | 286 | pr_warning("Failed to open debuginfo file.\n"); |
| 229 | return fd; | 287 | return -ENOENT; |
| 230 | } | 288 | } |
| 231 | pr_debug("Could not open vmlinux. Try to use symbols.\n"); | 289 | pr_debug("Could not open debuginfo. Try to use symbols.\n"); |
| 232 | return 0; | 290 | return 0; |
| 233 | } | 291 | } |
| 234 | 292 | ||
| 235 | /* Searching trace events corresponding to probe event */ | 293 | /* Searching trace events corresponding to a probe event */ |
| 236 | ntevs = find_probe_trace_events(fd, pev, tevs, max_tevs); | 294 | ntevs = debuginfo__find_trace_events(dinfo, pev, tevs, max_tevs); |
| 295 | |||
| 296 | debuginfo__delete(dinfo); | ||
| 237 | 297 | ||
| 238 | if (ntevs > 0) { /* Succeeded to find trace events */ | 298 | if (ntevs > 0) { /* Succeeded to find trace events */ |
| 239 | pr_debug("find %d probe_trace_events.\n", ntevs); | 299 | pr_debug("find %d probe_trace_events.\n", ntevs); |
| 240 | return ntevs; | 300 | if (module) |
| 301 | ret = add_module_to_probe_trace_events(*tevs, ntevs, | ||
| 302 | module); | ||
| 303 | return ret < 0 ? ret : ntevs; | ||
| 241 | } | 304 | } |
| 242 | 305 | ||
| 243 | if (ntevs == 0) { /* No error but failed to find probe point. */ | 306 | if (ntevs == 0) { /* No error but failed to find probe point. */ |
| @@ -371,8 +434,9 @@ int show_line_range(struct line_range *lr, const char *module) | |||
| 371 | { | 434 | { |
| 372 | int l = 1; | 435 | int l = 1; |
| 373 | struct line_node *ln; | 436 | struct line_node *ln; |
| 437 | struct debuginfo *dinfo; | ||
| 374 | FILE *fp; | 438 | FILE *fp; |
| 375 | int fd, ret; | 439 | int ret; |
| 376 | char *tmp; | 440 | char *tmp; |
| 377 | 441 | ||
| 378 | /* Search a line range */ | 442 | /* Search a line range */ |
| @@ -380,13 +444,14 @@ int show_line_range(struct line_range *lr, const char *module) | |||
| 380 | if (ret < 0) | 444 | if (ret < 0) |
| 381 | return ret; | 445 | return ret; |
| 382 | 446 | ||
| 383 | fd = open_vmlinux(module); | 447 | dinfo = open_debuginfo(module); |
| 384 | if (fd < 0) { | 448 | if (!dinfo) { |
| 385 | pr_warning("Failed to open debuginfo file.\n"); | 449 | pr_warning("Failed to open debuginfo file.\n"); |
| 386 | return fd; | 450 | return -ENOENT; |
| 387 | } | 451 | } |
| 388 | 452 | ||
| 389 | ret = find_line_range(fd, lr); | 453 | ret = debuginfo__find_line_range(dinfo, lr); |
| 454 | debuginfo__delete(dinfo); | ||
| 390 | if (ret == 0) { | 455 | if (ret == 0) { |
| 391 | pr_warning("Specified source line is not found.\n"); | 456 | pr_warning("Specified source line is not found.\n"); |
| 392 | return -ENOENT; | 457 | return -ENOENT; |
| @@ -448,7 +513,8 @@ end: | |||
| 448 | return ret; | 513 | return ret; |
| 449 | } | 514 | } |
| 450 | 515 | ||
| 451 | static int show_available_vars_at(int fd, struct perf_probe_event *pev, | 516 | static int show_available_vars_at(struct debuginfo *dinfo, |
| 517 | struct perf_probe_event *pev, | ||
| 452 | int max_vls, struct strfilter *_filter, | 518 | int max_vls, struct strfilter *_filter, |
| 453 | bool externs) | 519 | bool externs) |
| 454 | { | 520 | { |
| @@ -463,7 +529,8 @@ static int show_available_vars_at(int fd, struct perf_probe_event *pev, | |||
| 463 | return -EINVAL; | 529 | return -EINVAL; |
| 464 | pr_debug("Searching variables at %s\n", buf); | 530 | pr_debug("Searching variables at %s\n", buf); |
| 465 | 531 | ||
| 466 | ret = find_available_vars_at(fd, pev, &vls, max_vls, externs); | 532 | ret = debuginfo__find_available_vars_at(dinfo, pev, &vls, |
| 533 | max_vls, externs); | ||
| 467 | if (ret <= 0) { | 534 | if (ret <= 0) { |
| 468 | pr_err("Failed to find variables at %s (%d)\n", buf, ret); | 535 | pr_err("Failed to find variables at %s (%d)\n", buf, ret); |
| 469 | goto end; | 536 | goto end; |
| @@ -504,24 +571,26 @@ int show_available_vars(struct perf_probe_event *pevs, int npevs, | |||
| 504 | int max_vls, const char *module, | 571 | int max_vls, const char *module, |
| 505 | struct strfilter *_filter, bool externs) | 572 | struct strfilter *_filter, bool externs) |
| 506 | { | 573 | { |
| 507 | int i, fd, ret = 0; | 574 | int i, ret = 0; |
| 575 | struct debuginfo *dinfo; | ||
| 508 | 576 | ||
| 509 | ret = init_vmlinux(); | 577 | ret = init_vmlinux(); |
| 510 | if (ret < 0) | 578 | if (ret < 0) |
| 511 | return ret; | 579 | return ret; |
| 512 | 580 | ||
| 581 | dinfo = open_debuginfo(module); | ||
| 582 | if (!dinfo) { | ||
| 583 | pr_warning("Failed to open debuginfo file.\n"); | ||
| 584 | return -ENOENT; | ||
| 585 | } | ||
| 586 | |||
| 513 | setup_pager(); | 587 | setup_pager(); |
| 514 | 588 | ||
| 515 | for (i = 0; i < npevs && ret >= 0; i++) { | 589 | for (i = 0; i < npevs && ret >= 0; i++) |
| 516 | fd = open_vmlinux(module); | 590 | ret = show_available_vars_at(dinfo, &pevs[i], max_vls, _filter, |
| 517 | if (fd < 0) { | ||
| 518 | pr_warning("Failed to open debug information file.\n"); | ||
| 519 | ret = fd; | ||
| 520 | break; | ||
| 521 | } | ||
| 522 | ret = show_available_vars_at(fd, &pevs[i], max_vls, _filter, | ||
| 523 | externs); | 591 | externs); |
| 524 | } | 592 | |
| 593 | debuginfo__delete(dinfo); | ||
| 525 | return ret; | 594 | return ret; |
| 526 | } | 595 | } |
| 527 | 596 | ||
| @@ -990,7 +1059,7 @@ bool perf_probe_event_need_dwarf(struct perf_probe_event *pev) | |||
| 990 | 1059 | ||
| 991 | /* Parse probe_events event into struct probe_point */ | 1060 | /* Parse probe_events event into struct probe_point */ |
| 992 | static int parse_probe_trace_command(const char *cmd, | 1061 | static int parse_probe_trace_command(const char *cmd, |
| 993 | struct probe_trace_event *tev) | 1062 | struct probe_trace_event *tev) |
| 994 | { | 1063 | { |
| 995 | struct probe_trace_point *tp = &tev->point; | 1064 | struct probe_trace_point *tp = &tev->point; |
| 996 | char pr; | 1065 | char pr; |
| @@ -1023,8 +1092,14 @@ static int parse_probe_trace_command(const char *cmd, | |||
| 1023 | 1092 | ||
| 1024 | tp->retprobe = (pr == 'r'); | 1093 | tp->retprobe = (pr == 'r'); |
| 1025 | 1094 | ||
| 1026 | /* Scan function name and offset */ | 1095 | /* Scan module name(if there), function name and offset */ |
| 1027 | ret = sscanf(argv[1], "%a[^+]+%lu", (float *)(void *)&tp->symbol, | 1096 | p = strchr(argv[1], ':'); |
| 1097 | if (p) { | ||
| 1098 | tp->module = strndup(argv[1], p - argv[1]); | ||
| 1099 | p++; | ||
| 1100 | } else | ||
| 1101 | p = argv[1]; | ||
| 1102 | ret = sscanf(p, "%a[^+]+%lu", (float *)(void *)&tp->symbol, | ||
| 1028 | &tp->offset); | 1103 | &tp->offset); |
| 1029 | if (ret == 1) | 1104 | if (ret == 1) |
| 1030 | tp->offset = 0; | 1105 | tp->offset = 0; |
| @@ -1269,9 +1344,10 @@ char *synthesize_probe_trace_command(struct probe_trace_event *tev) | |||
| 1269 | if (buf == NULL) | 1344 | if (buf == NULL) |
| 1270 | return NULL; | 1345 | return NULL; |
| 1271 | 1346 | ||
| 1272 | len = e_snprintf(buf, MAX_CMDLEN, "%c:%s/%s %s+%lu", | 1347 | len = e_snprintf(buf, MAX_CMDLEN, "%c:%s/%s %s%s%s+%lu", |
| 1273 | tp->retprobe ? 'r' : 'p', | 1348 | tp->retprobe ? 'r' : 'p', |
| 1274 | tev->group, tev->event, | 1349 | tev->group, tev->event, |
| 1350 | tp->module ?: "", tp->module ? ":" : "", | ||
| 1275 | tp->symbol, tp->offset); | 1351 | tp->symbol, tp->offset); |
| 1276 | if (len <= 0) | 1352 | if (len <= 0) |
| 1277 | goto error; | 1353 | goto error; |
| @@ -1378,6 +1454,8 @@ static void clear_probe_trace_event(struct probe_trace_event *tev) | |||
| 1378 | free(tev->group); | 1454 | free(tev->group); |
| 1379 | if (tev->point.symbol) | 1455 | if (tev->point.symbol) |
| 1380 | free(tev->point.symbol); | 1456 | free(tev->point.symbol); |
| 1457 | if (tev->point.module) | ||
| 1458 | free(tev->point.module); | ||
| 1381 | for (i = 0; i < tev->nargs; i++) { | 1459 | for (i = 0; i < tev->nargs; i++) { |
| 1382 | if (tev->args[i].name) | 1460 | if (tev->args[i].name) |
| 1383 | free(tev->args[i].name); | 1461 | free(tev->args[i].name); |
| @@ -1729,7 +1807,7 @@ static int convert_to_probe_trace_events(struct perf_probe_event *pev, | |||
| 1729 | /* Convert perf_probe_event with debuginfo */ | 1807 | /* Convert perf_probe_event with debuginfo */ |
| 1730 | ret = try_to_find_probe_trace_events(pev, tevs, max_tevs, module); | 1808 | ret = try_to_find_probe_trace_events(pev, tevs, max_tevs, module); |
| 1731 | if (ret != 0) | 1809 | if (ret != 0) |
| 1732 | return ret; | 1810 | return ret; /* Found in debuginfo or got an error */ |
| 1733 | 1811 | ||
| 1734 | /* Allocate trace event buffer */ | 1812 | /* Allocate trace event buffer */ |
| 1735 | tev = *tevs = zalloc(sizeof(struct probe_trace_event)); | 1813 | tev = *tevs = zalloc(sizeof(struct probe_trace_event)); |
| @@ -1742,6 +1820,11 @@ static int convert_to_probe_trace_events(struct perf_probe_event *pev, | |||
| 1742 | ret = -ENOMEM; | 1820 | ret = -ENOMEM; |
| 1743 | goto error; | 1821 | goto error; |
| 1744 | } | 1822 | } |
| 1823 | tev->point.module = strdup(module); | ||
| 1824 | if (tev->point.module == NULL) { | ||
| 1825 | ret = -ENOMEM; | ||
| 1826 | goto error; | ||
| 1827 | } | ||
| 1745 | tev->point.offset = pev->point.offset; | 1828 | tev->point.offset = pev->point.offset; |
| 1746 | tev->point.retprobe = pev->point.retprobe; | 1829 | tev->point.retprobe = pev->point.retprobe; |
| 1747 | tev->nargs = pev->nargs; | 1830 | tev->nargs = pev->nargs; |
diff --git a/tools/perf/util/probe-event.h b/tools/perf/util/probe-event.h index 3434fc9d79d..a7dee835f49 100644 --- a/tools/perf/util/probe-event.h +++ b/tools/perf/util/probe-event.h | |||
| @@ -10,6 +10,7 @@ extern bool probe_event_dry_run; | |||
| 10 | /* kprobe-tracer tracing point */ | 10 | /* kprobe-tracer tracing point */ |
| 11 | struct probe_trace_point { | 11 | struct probe_trace_point { |
| 12 | char *symbol; /* Base symbol */ | 12 | char *symbol; /* Base symbol */ |
| 13 | char *module; /* Module name */ | ||
| 13 | unsigned long offset; /* Offset from symbol */ | 14 | unsigned long offset; /* Offset from symbol */ |
| 14 | bool retprobe; /* Return probe flag */ | 15 | bool retprobe; /* Return probe flag */ |
| 15 | }; | 16 | }; |
diff --git a/tools/perf/util/probe-finder.c b/tools/perf/util/probe-finder.c index 3b9d0b800d5..3e44a3e3651 100644 --- a/tools/perf/util/probe-finder.c +++ b/tools/perf/util/probe-finder.c | |||
| @@ -43,21 +43,6 @@ | |||
| 43 | /* Kprobe tracer basic type is up to u64 */ | 43 | /* Kprobe tracer basic type is up to u64 */ |
| 44 | #define MAX_BASIC_TYPE_BITS 64 | 44 | #define MAX_BASIC_TYPE_BITS 64 |
| 45 | 45 | ||
| 46 | /* | ||
| 47 | * Compare the tail of two strings. | ||
| 48 | * Return 0 if whole of either string is same as another's tail part. | ||
| 49 | */ | ||
| 50 | static int strtailcmp(const char *s1, const char *s2) | ||
| 51 | { | ||
| 52 | int i1 = strlen(s1); | ||
| 53 | int i2 = strlen(s2); | ||
| 54 | while (--i1 >= 0 && --i2 >= 0) { | ||
| 55 | if (s1[i1] != s2[i2]) | ||
| 56 | return s1[i1] - s2[i2]; | ||
| 57 | } | ||
| 58 | return 0; | ||
| 59 | } | ||
| 60 | |||
| 61 | /* Line number list operations */ | 46 | /* Line number list operations */ |
| 62 | 47 | ||
| 63 | /* Add a line to line number list */ | 48 | /* Add a line to line number list */ |
| @@ -131,29 +116,37 @@ static const Dwfl_Callbacks offline_callbacks = { | |||
| 131 | }; | 116 | }; |
| 132 | 117 | ||
| 133 | /* Get a Dwarf from offline image */ | 118 | /* Get a Dwarf from offline image */ |
| 134 | static Dwarf *dwfl_init_offline_dwarf(int fd, Dwfl **dwflp, Dwarf_Addr *bias) | 119 | static int debuginfo__init_offline_dwarf(struct debuginfo *self, |
| 120 | const char *path) | ||
| 135 | { | 121 | { |
| 136 | Dwfl_Module *mod; | 122 | Dwfl_Module *mod; |
| 137 | Dwarf *dbg = NULL; | 123 | int fd; |
| 138 | 124 | ||
| 139 | if (!dwflp) | 125 | fd = open(path, O_RDONLY); |
| 140 | return NULL; | 126 | if (fd < 0) |
| 127 | return fd; | ||
| 141 | 128 | ||
| 142 | *dwflp = dwfl_begin(&offline_callbacks); | 129 | self->dwfl = dwfl_begin(&offline_callbacks); |
| 143 | if (!*dwflp) | 130 | if (!self->dwfl) |
| 144 | return NULL; | 131 | goto error; |
| 145 | 132 | ||
| 146 | mod = dwfl_report_offline(*dwflp, "", "", fd); | 133 | mod = dwfl_report_offline(self->dwfl, "", "", fd); |
| 147 | if (!mod) | 134 | if (!mod) |
| 148 | goto error; | 135 | goto error; |
| 149 | 136 | ||
| 150 | dbg = dwfl_module_getdwarf(mod, bias); | 137 | self->dbg = dwfl_module_getdwarf(mod, &self->bias); |
| 151 | if (!dbg) { | 138 | if (!self->dbg) |
| 139 | goto error; | ||
| 140 | |||
| 141 | return 0; | ||
| 152 | error: | 142 | error: |
| 153 | dwfl_end(*dwflp); | 143 | if (self->dwfl) |
| 154 | *dwflp = NULL; | 144 | dwfl_end(self->dwfl); |
| 155 | } | 145 | else |
| 156 | return dbg; | 146 | close(fd); |
| 147 | memset(self, 0, sizeof(*self)); | ||
| 148 | |||
| 149 | return -ENOENT; | ||
| 157 | } | 150 | } |
| 158 | 151 | ||
| 159 | #if _ELFUTILS_PREREQ(0, 148) | 152 | #if _ELFUTILS_PREREQ(0, 148) |
| @@ -189,597 +182,81 @@ static const Dwfl_Callbacks kernel_callbacks = { | |||
| 189 | }; | 182 | }; |
| 190 | 183 | ||
| 191 | /* Get a Dwarf from live kernel image */ | 184 | /* Get a Dwarf from live kernel image */ |
| 192 | static Dwarf *dwfl_init_live_kernel_dwarf(Dwarf_Addr addr, Dwfl **dwflp, | 185 | static int debuginfo__init_online_kernel_dwarf(struct debuginfo *self, |
| 193 | Dwarf_Addr *bias) | 186 | Dwarf_Addr addr) |
| 194 | { | 187 | { |
| 195 | Dwarf *dbg; | 188 | self->dwfl = dwfl_begin(&kernel_callbacks); |
| 196 | 189 | if (!self->dwfl) | |
| 197 | if (!dwflp) | 190 | return -EINVAL; |
| 198 | return NULL; | ||
| 199 | |||
| 200 | *dwflp = dwfl_begin(&kernel_callbacks); | ||
| 201 | if (!*dwflp) | ||
| 202 | return NULL; | ||
| 203 | 191 | ||
| 204 | /* Load the kernel dwarves: Don't care the result here */ | 192 | /* Load the kernel dwarves: Don't care the result here */ |
| 205 | dwfl_linux_kernel_report_kernel(*dwflp); | 193 | dwfl_linux_kernel_report_kernel(self->dwfl); |
| 206 | dwfl_linux_kernel_report_modules(*dwflp); | 194 | dwfl_linux_kernel_report_modules(self->dwfl); |
| 207 | 195 | ||
| 208 | dbg = dwfl_addrdwarf(*dwflp, addr, bias); | 196 | self->dbg = dwfl_addrdwarf(self->dwfl, addr, &self->bias); |
| 209 | /* Here, check whether we could get a real dwarf */ | 197 | /* Here, check whether we could get a real dwarf */ |
| 210 | if (!dbg) { | 198 | if (!self->dbg) { |
| 211 | pr_debug("Failed to find kernel dwarf at %lx\n", | 199 | pr_debug("Failed to find kernel dwarf at %lx\n", |
| 212 | (unsigned long)addr); | 200 | (unsigned long)addr); |
| 213 | dwfl_end(*dwflp); | 201 | dwfl_end(self->dwfl); |
| 214 | *dwflp = NULL; | 202 | memset(self, 0, sizeof(*self)); |
| 203 | return -ENOENT; | ||
| 215 | } | 204 | } |
| 216 | return dbg; | 205 | |
| 206 | return 0; | ||
| 217 | } | 207 | } |
| 218 | #else | 208 | #else |
| 219 | /* With older elfutils, this just support kernel module... */ | 209 | /* With older elfutils, this just support kernel module... */ |
| 220 | static Dwarf *dwfl_init_live_kernel_dwarf(Dwarf_Addr addr __used, Dwfl **dwflp, | 210 | static int debuginfo__init_online_kernel_dwarf(struct debuginfo *self, |
| 221 | Dwarf_Addr *bias) | 211 | Dwarf_Addr addr __used) |
| 222 | { | 212 | { |
| 223 | int fd; | ||
| 224 | const char *path = kernel_get_module_path("kernel"); | 213 | const char *path = kernel_get_module_path("kernel"); |
| 225 | 214 | ||
| 226 | if (!path) { | 215 | if (!path) { |
| 227 | pr_err("Failed to find vmlinux path\n"); | 216 | pr_err("Failed to find vmlinux path\n"); |
| 228 | return NULL; | 217 | return -ENOENT; |
| 229 | } | 218 | } |
| 230 | 219 | ||
| 231 | pr_debug2("Use file %s for debuginfo\n", path); | 220 | pr_debug2("Use file %s for debuginfo\n", path); |
| 232 | fd = open(path, O_RDONLY); | 221 | return debuginfo__init_offline_dwarf(self, path); |
| 233 | if (fd < 0) | ||
| 234 | return NULL; | ||
| 235 | |||
| 236 | return dwfl_init_offline_dwarf(fd, dwflp, bias); | ||
| 237 | } | 222 | } |
| 238 | #endif | 223 | #endif |
| 239 | 224 | ||
| 240 | /* Dwarf wrappers */ | 225 | struct debuginfo *debuginfo__new(const char *path) |
| 241 | |||
| 242 | /* Find the realpath of the target file. */ | ||
| 243 | static const char *cu_find_realpath(Dwarf_Die *cu_die, const char *fname) | ||
| 244 | { | ||
| 245 | Dwarf_Files *files; | ||
| 246 | size_t nfiles, i; | ||
| 247 | const char *src = NULL; | ||
| 248 | int ret; | ||
| 249 | |||
| 250 | if (!fname) | ||
| 251 | return NULL; | ||
| 252 | |||
| 253 | ret = dwarf_getsrcfiles(cu_die, &files, &nfiles); | ||
| 254 | if (ret != 0) | ||
| 255 | return NULL; | ||
| 256 | |||
| 257 | for (i = 0; i < nfiles; i++) { | ||
| 258 | src = dwarf_filesrc(files, i, NULL, NULL); | ||
| 259 | if (strtailcmp(src, fname) == 0) | ||
| 260 | break; | ||
| 261 | } | ||
| 262 | if (i == nfiles) | ||
| 263 | return NULL; | ||
| 264 | return src; | ||
| 265 | } | ||
| 266 | |||
| 267 | /* Get DW_AT_comp_dir (should be NULL with older gcc) */ | ||
| 268 | static const char *cu_get_comp_dir(Dwarf_Die *cu_die) | ||
| 269 | { | ||
| 270 | Dwarf_Attribute attr; | ||
| 271 | if (dwarf_attr(cu_die, DW_AT_comp_dir, &attr) == NULL) | ||
| 272 | return NULL; | ||
| 273 | return dwarf_formstring(&attr); | ||
| 274 | } | ||
| 275 | |||
| 276 | /* Get a line number and file name for given address */ | ||
| 277 | static int cu_find_lineinfo(Dwarf_Die *cudie, unsigned long addr, | ||
| 278 | const char **fname, int *lineno) | ||
| 279 | { | ||
| 280 | Dwarf_Line *line; | ||
| 281 | Dwarf_Addr laddr; | ||
| 282 | |||
| 283 | line = dwarf_getsrc_die(cudie, (Dwarf_Addr)addr); | ||
| 284 | if (line && dwarf_lineaddr(line, &laddr) == 0 && | ||
| 285 | addr == (unsigned long)laddr && dwarf_lineno(line, lineno) == 0) { | ||
| 286 | *fname = dwarf_linesrc(line, NULL, NULL); | ||
| 287 | if (!*fname) | ||
| 288 | /* line number is useless without filename */ | ||
| 289 | *lineno = 0; | ||
| 290 | } | ||
| 291 | |||
| 292 | return *lineno ?: -ENOENT; | ||
| 293 | } | ||
| 294 | |||
| 295 | /* Compare diename and tname */ | ||
| 296 | static bool die_compare_name(Dwarf_Die *dw_die, const char *tname) | ||
| 297 | { | ||
| 298 | const char *name; | ||
| 299 | name = dwarf_diename(dw_die); | ||
| 300 | return name ? (strcmp(tname, name) == 0) : false; | ||
| 301 | } | ||
| 302 | |||
| 303 | /* Get callsite line number of inline-function instance */ | ||
| 304 | static int die_get_call_lineno(Dwarf_Die *in_die) | ||
| 305 | { | ||
| 306 | Dwarf_Attribute attr; | ||
| 307 | Dwarf_Word ret; | ||
| 308 | |||
| 309 | if (!dwarf_attr(in_die, DW_AT_call_line, &attr)) | ||
| 310 | return -ENOENT; | ||
| 311 | |||
| 312 | dwarf_formudata(&attr, &ret); | ||
| 313 | return (int)ret; | ||
| 314 | } | ||
| 315 | |||
| 316 | /* Get type die */ | ||
| 317 | static Dwarf_Die *die_get_type(Dwarf_Die *vr_die, Dwarf_Die *die_mem) | ||
| 318 | { | ||
| 319 | Dwarf_Attribute attr; | ||
| 320 | |||
| 321 | if (dwarf_attr_integrate(vr_die, DW_AT_type, &attr) && | ||
| 322 | dwarf_formref_die(&attr, die_mem)) | ||
| 323 | return die_mem; | ||
| 324 | else | ||
| 325 | return NULL; | ||
| 326 | } | ||
| 327 | |||
| 328 | /* Get a type die, but skip qualifiers */ | ||
| 329 | static Dwarf_Die *__die_get_real_type(Dwarf_Die *vr_die, Dwarf_Die *die_mem) | ||
| 330 | { | ||
| 331 | int tag; | ||
| 332 | |||
| 333 | do { | ||
| 334 | vr_die = die_get_type(vr_die, die_mem); | ||
| 335 | if (!vr_die) | ||
| 336 | break; | ||
| 337 | tag = dwarf_tag(vr_die); | ||
| 338 | } while (tag == DW_TAG_const_type || | ||
| 339 | tag == DW_TAG_restrict_type || | ||
| 340 | tag == DW_TAG_volatile_type || | ||
| 341 | tag == DW_TAG_shared_type); | ||
| 342 | |||
| 343 | return vr_die; | ||
| 344 | } | ||
| 345 | |||
| 346 | /* Get a type die, but skip qualifiers and typedef */ | ||
| 347 | static Dwarf_Die *die_get_real_type(Dwarf_Die *vr_die, Dwarf_Die *die_mem) | ||
| 348 | { | ||
| 349 | do { | ||
| 350 | vr_die = __die_get_real_type(vr_die, die_mem); | ||
| 351 | } while (vr_die && dwarf_tag(vr_die) == DW_TAG_typedef); | ||
| 352 | |||
| 353 | return vr_die; | ||
| 354 | } | ||
| 355 | |||
| 356 | static int die_get_attr_udata(Dwarf_Die *tp_die, unsigned int attr_name, | ||
| 357 | Dwarf_Word *result) | ||
| 358 | { | ||
| 359 | Dwarf_Attribute attr; | ||
| 360 | |||
| 361 | if (dwarf_attr(tp_die, attr_name, &attr) == NULL || | ||
| 362 | dwarf_formudata(&attr, result) != 0) | ||
| 363 | return -ENOENT; | ||
| 364 | |||
| 365 | return 0; | ||
| 366 | } | ||
| 367 | |||
| 368 | static bool die_is_signed_type(Dwarf_Die *tp_die) | ||
| 369 | { | ||
| 370 | Dwarf_Word ret; | ||
| 371 | |||
| 372 | if (die_get_attr_udata(tp_die, DW_AT_encoding, &ret)) | ||
| 373 | return false; | ||
| 374 | |||
| 375 | return (ret == DW_ATE_signed_char || ret == DW_ATE_signed || | ||
| 376 | ret == DW_ATE_signed_fixed); | ||
| 377 | } | ||
| 378 | |||
| 379 | static int die_get_byte_size(Dwarf_Die *tp_die) | ||
| 380 | { | ||
| 381 | Dwarf_Word ret; | ||
| 382 | |||
| 383 | if (die_get_attr_udata(tp_die, DW_AT_byte_size, &ret)) | ||
| 384 | return 0; | ||
| 385 | |||
| 386 | return (int)ret; | ||
| 387 | } | ||
| 388 | |||
| 389 | static int die_get_bit_size(Dwarf_Die *tp_die) | ||
| 390 | { | ||
| 391 | Dwarf_Word ret; | ||
| 392 | |||
| 393 | if (die_get_attr_udata(tp_die, DW_AT_bit_size, &ret)) | ||
| 394 | return 0; | ||
| 395 | |||
| 396 | return (int)ret; | ||
| 397 | } | ||
| 398 | |||
| 399 | static int die_get_bit_offset(Dwarf_Die *tp_die) | ||
| 400 | { | ||
| 401 | Dwarf_Word ret; | ||
| 402 | |||
| 403 | if (die_get_attr_udata(tp_die, DW_AT_bit_offset, &ret)) | ||
| 404 | return 0; | ||
| 405 | |||
| 406 | return (int)ret; | ||
| 407 | } | ||
| 408 | |||
| 409 | /* Get data_member_location offset */ | ||
| 410 | static int die_get_data_member_location(Dwarf_Die *mb_die, Dwarf_Word *offs) | ||
| 411 | { | ||
| 412 | Dwarf_Attribute attr; | ||
| 413 | Dwarf_Op *expr; | ||
| 414 | size_t nexpr; | ||
| 415 | int ret; | ||
| 416 | |||
| 417 | if (dwarf_attr(mb_die, DW_AT_data_member_location, &attr) == NULL) | ||
| 418 | return -ENOENT; | ||
| 419 | |||
| 420 | if (dwarf_formudata(&attr, offs) != 0) { | ||
| 421 | /* DW_AT_data_member_location should be DW_OP_plus_uconst */ | ||
| 422 | ret = dwarf_getlocation(&attr, &expr, &nexpr); | ||
| 423 | if (ret < 0 || nexpr == 0) | ||
| 424 | return -ENOENT; | ||
| 425 | |||
| 426 | if (expr[0].atom != DW_OP_plus_uconst || nexpr != 1) { | ||
| 427 | pr_debug("Unable to get offset:Unexpected OP %x (%zd)\n", | ||
| 428 | expr[0].atom, nexpr); | ||
| 429 | return -ENOTSUP; | ||
| 430 | } | ||
| 431 | *offs = (Dwarf_Word)expr[0].number; | ||
| 432 | } | ||
| 433 | return 0; | ||
| 434 | } | ||
| 435 | |||
| 436 | /* Return values for die_find callbacks */ | ||
| 437 | enum { | ||
| 438 | DIE_FIND_CB_FOUND = 0, /* End of Search */ | ||
| 439 | DIE_FIND_CB_CHILD = 1, /* Search only children */ | ||
| 440 | DIE_FIND_CB_SIBLING = 2, /* Search only siblings */ | ||
| 441 | DIE_FIND_CB_CONTINUE = 3, /* Search children and siblings */ | ||
| 442 | }; | ||
| 443 | |||
| 444 | /* Search a child die */ | ||
| 445 | static Dwarf_Die *die_find_child(Dwarf_Die *rt_die, | ||
| 446 | int (*callback)(Dwarf_Die *, void *), | ||
| 447 | void *data, Dwarf_Die *die_mem) | ||
| 448 | { | 226 | { |
| 449 | Dwarf_Die child_die; | 227 | struct debuginfo *self = zalloc(sizeof(struct debuginfo)); |
| 450 | int ret; | 228 | if (!self) |
| 451 | |||
| 452 | ret = dwarf_child(rt_die, die_mem); | ||
| 453 | if (ret != 0) | ||
| 454 | return NULL; | 229 | return NULL; |
| 455 | 230 | ||
| 456 | do { | 231 | if (debuginfo__init_offline_dwarf(self, path) < 0) { |
| 457 | ret = callback(die_mem, data); | 232 | free(self); |
| 458 | if (ret == DIE_FIND_CB_FOUND) | 233 | self = NULL; |
| 459 | return die_mem; | ||
| 460 | |||
| 461 | if ((ret & DIE_FIND_CB_CHILD) && | ||
| 462 | die_find_child(die_mem, callback, data, &child_die)) { | ||
| 463 | memcpy(die_mem, &child_die, sizeof(Dwarf_Die)); | ||
| 464 | return die_mem; | ||
| 465 | } | ||
| 466 | } while ((ret & DIE_FIND_CB_SIBLING) && | ||
| 467 | dwarf_siblingof(die_mem, die_mem) == 0); | ||
| 468 | |||
| 469 | return NULL; | ||
| 470 | } | ||
| 471 | |||
| 472 | struct __addr_die_search_param { | ||
| 473 | Dwarf_Addr addr; | ||
| 474 | Dwarf_Die *die_mem; | ||
| 475 | }; | ||
| 476 | |||
| 477 | static int __die_search_func_cb(Dwarf_Die *fn_die, void *data) | ||
| 478 | { | ||
| 479 | struct __addr_die_search_param *ad = data; | ||
| 480 | |||
| 481 | if (dwarf_tag(fn_die) == DW_TAG_subprogram && | ||
| 482 | dwarf_haspc(fn_die, ad->addr)) { | ||
| 483 | memcpy(ad->die_mem, fn_die, sizeof(Dwarf_Die)); | ||
| 484 | return DWARF_CB_ABORT; | ||
| 485 | } | 234 | } |
| 486 | return DWARF_CB_OK; | ||
| 487 | } | ||
| 488 | |||
| 489 | /* Search a real subprogram including this line, */ | ||
| 490 | static Dwarf_Die *die_find_real_subprogram(Dwarf_Die *cu_die, Dwarf_Addr addr, | ||
| 491 | Dwarf_Die *die_mem) | ||
| 492 | { | ||
| 493 | struct __addr_die_search_param ad; | ||
| 494 | ad.addr = addr; | ||
| 495 | ad.die_mem = die_mem; | ||
| 496 | /* dwarf_getscopes can't find subprogram. */ | ||
| 497 | if (!dwarf_getfuncs(cu_die, __die_search_func_cb, &ad, 0)) | ||
| 498 | return NULL; | ||
| 499 | else | ||
| 500 | return die_mem; | ||
| 501 | } | ||
| 502 | |||
| 503 | /* die_find callback for inline function search */ | ||
| 504 | static int __die_find_inline_cb(Dwarf_Die *die_mem, void *data) | ||
| 505 | { | ||
| 506 | Dwarf_Addr *addr = data; | ||
| 507 | |||
| 508 | if (dwarf_tag(die_mem) == DW_TAG_inlined_subroutine && | ||
| 509 | dwarf_haspc(die_mem, *addr)) | ||
| 510 | return DIE_FIND_CB_FOUND; | ||
| 511 | 235 | ||
| 512 | return DIE_FIND_CB_CONTINUE; | 236 | return self; |
| 513 | } | 237 | } |
| 514 | 238 | ||
| 515 | /* Similar to dwarf_getfuncs, but returns inlined_subroutine if exists. */ | 239 | struct debuginfo *debuginfo__new_online_kernel(unsigned long addr) |
| 516 | static Dwarf_Die *die_find_inlinefunc(Dwarf_Die *sp_die, Dwarf_Addr addr, | ||
| 517 | Dwarf_Die *die_mem) | ||
| 518 | { | 240 | { |
| 519 | Dwarf_Die tmp_die; | 241 | struct debuginfo *self = zalloc(sizeof(struct debuginfo)); |
| 520 | 242 | if (!self) | |
| 521 | sp_die = die_find_child(sp_die, __die_find_inline_cb, &addr, &tmp_die); | ||
| 522 | if (!sp_die) | ||
| 523 | return NULL; | 243 | return NULL; |
| 524 | 244 | ||
| 525 | /* Inlined function could be recursive. Trace it until fail */ | 245 | if (debuginfo__init_online_kernel_dwarf(self, (Dwarf_Addr)addr) < 0) { |
| 526 | while (sp_die) { | 246 | free(self); |
| 527 | memcpy(die_mem, sp_die, sizeof(Dwarf_Die)); | 247 | self = NULL; |
| 528 | sp_die = die_find_child(sp_die, __die_find_inline_cb, &addr, | ||
| 529 | &tmp_die); | ||
| 530 | } | ||
| 531 | |||
| 532 | return die_mem; | ||
| 533 | } | ||
| 534 | |||
| 535 | /* Walker on lines (Note: line number will not be sorted) */ | ||
| 536 | typedef int (* line_walk_handler_t) (const char *fname, int lineno, | ||
| 537 | Dwarf_Addr addr, void *data); | ||
| 538 | |||
| 539 | struct __line_walk_param { | ||
| 540 | const char *fname; | ||
| 541 | line_walk_handler_t handler; | ||
| 542 | void *data; | ||
| 543 | int retval; | ||
| 544 | }; | ||
| 545 | |||
| 546 | static int __die_walk_funclines_cb(Dwarf_Die *in_die, void *data) | ||
| 547 | { | ||
| 548 | struct __line_walk_param *lw = data; | ||
| 549 | Dwarf_Addr addr; | ||
| 550 | int lineno; | ||
| 551 | |||
| 552 | if (dwarf_tag(in_die) == DW_TAG_inlined_subroutine) { | ||
| 553 | lineno = die_get_call_lineno(in_die); | ||
| 554 | if (lineno > 0 && dwarf_entrypc(in_die, &addr) == 0) { | ||
| 555 | lw->retval = lw->handler(lw->fname, lineno, addr, | ||
| 556 | lw->data); | ||
| 557 | if (lw->retval != 0) | ||
| 558 | return DIE_FIND_CB_FOUND; | ||
| 559 | } | ||
| 560 | } | ||
| 561 | return DIE_FIND_CB_SIBLING; | ||
| 562 | } | ||
| 563 | |||
| 564 | /* Walk on lines of blocks included in given DIE */ | ||
| 565 | static int __die_walk_funclines(Dwarf_Die *sp_die, | ||
| 566 | line_walk_handler_t handler, void *data) | ||
| 567 | { | ||
| 568 | struct __line_walk_param lw = { | ||
| 569 | .handler = handler, | ||
| 570 | .data = data, | ||
| 571 | .retval = 0, | ||
| 572 | }; | ||
| 573 | Dwarf_Die die_mem; | ||
| 574 | Dwarf_Addr addr; | ||
| 575 | int lineno; | ||
| 576 | |||
| 577 | /* Handle function declaration line */ | ||
| 578 | lw.fname = dwarf_decl_file(sp_die); | ||
| 579 | if (lw.fname && dwarf_decl_line(sp_die, &lineno) == 0 && | ||
| 580 | dwarf_entrypc(sp_die, &addr) == 0) { | ||
| 581 | lw.retval = handler(lw.fname, lineno, addr, data); | ||
| 582 | if (lw.retval != 0) | ||
| 583 | goto done; | ||
| 584 | } | ||
| 585 | die_find_child(sp_die, __die_walk_funclines_cb, &lw, &die_mem); | ||
| 586 | done: | ||
| 587 | return lw.retval; | ||
| 588 | } | ||
| 589 | |||
| 590 | static int __die_walk_culines_cb(Dwarf_Die *sp_die, void *data) | ||
| 591 | { | ||
| 592 | struct __line_walk_param *lw = data; | ||
| 593 | |||
| 594 | lw->retval = __die_walk_funclines(sp_die, lw->handler, lw->data); | ||
| 595 | if (lw->retval != 0) | ||
| 596 | return DWARF_CB_ABORT; | ||
| 597 | |||
| 598 | return DWARF_CB_OK; | ||
| 599 | } | ||
| 600 | |||
| 601 | /* | ||
| 602 | * Walk on lines inside given PDIE. If the PDIE is subprogram, walk only on | ||
| 603 | * the lines inside the subprogram, otherwise PDIE must be a CU DIE. | ||
| 604 | */ | ||
| 605 | static int die_walk_lines(Dwarf_Die *pdie, line_walk_handler_t handler, | ||
| 606 | void *data) | ||
| 607 | { | ||
| 608 | Dwarf_Lines *lines; | ||
| 609 | Dwarf_Line *line; | ||
| 610 | Dwarf_Addr addr; | ||
| 611 | const char *fname; | ||
| 612 | int lineno, ret = 0; | ||
| 613 | Dwarf_Die die_mem, *cu_die; | ||
| 614 | size_t nlines, i; | ||
| 615 | |||
| 616 | /* Get the CU die */ | ||
| 617 | if (dwarf_tag(pdie) == DW_TAG_subprogram) | ||
| 618 | cu_die = dwarf_diecu(pdie, &die_mem, NULL, NULL); | ||
| 619 | else | ||
| 620 | cu_die = pdie; | ||
| 621 | if (!cu_die) { | ||
| 622 | pr_debug2("Failed to get CU from subprogram\n"); | ||
| 623 | return -EINVAL; | ||
| 624 | } | ||
| 625 | |||
| 626 | /* Get lines list in the CU */ | ||
| 627 | if (dwarf_getsrclines(cu_die, &lines, &nlines) != 0) { | ||
| 628 | pr_debug2("Failed to get source lines on this CU.\n"); | ||
| 629 | return -ENOENT; | ||
| 630 | } | ||
| 631 | pr_debug2("Get %zd lines from this CU\n", nlines); | ||
| 632 | |||
| 633 | /* Walk on the lines on lines list */ | ||
| 634 | for (i = 0; i < nlines; i++) { | ||
| 635 | line = dwarf_onesrcline(lines, i); | ||
| 636 | if (line == NULL || | ||
| 637 | dwarf_lineno(line, &lineno) != 0 || | ||
| 638 | dwarf_lineaddr(line, &addr) != 0) { | ||
| 639 | pr_debug2("Failed to get line info. " | ||
| 640 | "Possible error in debuginfo.\n"); | ||
| 641 | continue; | ||
| 642 | } | ||
| 643 | /* Filter lines based on address */ | ||
| 644 | if (pdie != cu_die) | ||
| 645 | /* | ||
| 646 | * Address filtering | ||
| 647 | * The line is included in given function, and | ||
| 648 | * no inline block includes it. | ||
| 649 | */ | ||
| 650 | if (!dwarf_haspc(pdie, addr) || | ||
| 651 | die_find_inlinefunc(pdie, addr, &die_mem)) | ||
| 652 | continue; | ||
| 653 | /* Get source line */ | ||
| 654 | fname = dwarf_linesrc(line, NULL, NULL); | ||
| 655 | |||
| 656 | ret = handler(fname, lineno, addr, data); | ||
| 657 | if (ret != 0) | ||
| 658 | return ret; | ||
| 659 | } | ||
| 660 | |||
| 661 | /* | ||
| 662 | * Dwarf lines doesn't include function declarations and inlined | ||
| 663 | * subroutines. We have to check functions list or given function. | ||
| 664 | */ | ||
| 665 | if (pdie != cu_die) | ||
| 666 | ret = __die_walk_funclines(pdie, handler, data); | ||
| 667 | else { | ||
| 668 | struct __line_walk_param param = { | ||
| 669 | .handler = handler, | ||
| 670 | .data = data, | ||
| 671 | .retval = 0, | ||
| 672 | }; | ||
| 673 | dwarf_getfuncs(cu_die, __die_walk_culines_cb, ¶m, 0); | ||
| 674 | ret = param.retval; | ||
| 675 | } | 248 | } |
| 676 | 249 | ||
| 677 | return ret; | 250 | return self; |
| 678 | } | ||
| 679 | |||
| 680 | struct __find_variable_param { | ||
| 681 | const char *name; | ||
| 682 | Dwarf_Addr addr; | ||
| 683 | }; | ||
| 684 | |||
| 685 | static int __die_find_variable_cb(Dwarf_Die *die_mem, void *data) | ||
| 686 | { | ||
| 687 | struct __find_variable_param *fvp = data; | ||
| 688 | int tag; | ||
| 689 | |||
| 690 | tag = dwarf_tag(die_mem); | ||
| 691 | if ((tag == DW_TAG_formal_parameter || | ||
| 692 | tag == DW_TAG_variable) && | ||
| 693 | die_compare_name(die_mem, fvp->name)) | ||
| 694 | return DIE_FIND_CB_FOUND; | ||
| 695 | |||
| 696 | if (dwarf_haspc(die_mem, fvp->addr)) | ||
| 697 | return DIE_FIND_CB_CONTINUE; | ||
| 698 | else | ||
| 699 | return DIE_FIND_CB_SIBLING; | ||
| 700 | } | ||
| 701 | |||
| 702 | /* Find a variable called 'name' at given address */ | ||
| 703 | static Dwarf_Die *die_find_variable_at(Dwarf_Die *sp_die, const char *name, | ||
| 704 | Dwarf_Addr addr, Dwarf_Die *die_mem) | ||
| 705 | { | ||
| 706 | struct __find_variable_param fvp = { .name = name, .addr = addr}; | ||
| 707 | |||
| 708 | return die_find_child(sp_die, __die_find_variable_cb, (void *)&fvp, | ||
| 709 | die_mem); | ||
| 710 | } | ||
| 711 | |||
| 712 | static int __die_find_member_cb(Dwarf_Die *die_mem, void *data) | ||
| 713 | { | ||
| 714 | const char *name = data; | ||
| 715 | |||
| 716 | if ((dwarf_tag(die_mem) == DW_TAG_member) && | ||
| 717 | die_compare_name(die_mem, name)) | ||
| 718 | return DIE_FIND_CB_FOUND; | ||
| 719 | |||
| 720 | return DIE_FIND_CB_SIBLING; | ||
| 721 | } | ||
| 722 | |||
| 723 | /* Find a member called 'name' */ | ||
| 724 | static Dwarf_Die *die_find_member(Dwarf_Die *st_die, const char *name, | ||
| 725 | Dwarf_Die *die_mem) | ||
| 726 | { | ||
| 727 | return die_find_child(st_die, __die_find_member_cb, (void *)name, | ||
| 728 | die_mem); | ||
| 729 | } | ||
| 730 | |||
| 731 | /* Get the name of given variable DIE */ | ||
| 732 | static int die_get_typename(Dwarf_Die *vr_die, char *buf, int len) | ||
| 733 | { | ||
| 734 | Dwarf_Die type; | ||
| 735 | int tag, ret, ret2; | ||
| 736 | const char *tmp = ""; | ||
| 737 | |||
| 738 | if (__die_get_real_type(vr_die, &type) == NULL) | ||
| 739 | return -ENOENT; | ||
| 740 | |||
| 741 | tag = dwarf_tag(&type); | ||
| 742 | if (tag == DW_TAG_array_type || tag == DW_TAG_pointer_type) | ||
| 743 | tmp = "*"; | ||
| 744 | else if (tag == DW_TAG_subroutine_type) { | ||
| 745 | /* Function pointer */ | ||
| 746 | ret = snprintf(buf, len, "(function_type)"); | ||
| 747 | return (ret >= len) ? -E2BIG : ret; | ||
| 748 | } else { | ||
| 749 | if (!dwarf_diename(&type)) | ||
| 750 | return -ENOENT; | ||
| 751 | if (tag == DW_TAG_union_type) | ||
| 752 | tmp = "union "; | ||
| 753 | else if (tag == DW_TAG_structure_type) | ||
| 754 | tmp = "struct "; | ||
| 755 | /* Write a base name */ | ||
| 756 | ret = snprintf(buf, len, "%s%s", tmp, dwarf_diename(&type)); | ||
| 757 | return (ret >= len) ? -E2BIG : ret; | ||
| 758 | } | ||
| 759 | ret = die_get_typename(&type, buf, len); | ||
| 760 | if (ret > 0) { | ||
| 761 | ret2 = snprintf(buf + ret, len - ret, "%s", tmp); | ||
| 762 | ret = (ret2 >= len - ret) ? -E2BIG : ret2 + ret; | ||
| 763 | } | ||
| 764 | return ret; | ||
| 765 | } | 251 | } |
| 766 | 252 | ||
| 767 | /* Get the name and type of given variable DIE, stored as "type\tname" */ | 253 | void debuginfo__delete(struct debuginfo *self) |
| 768 | static int die_get_varname(Dwarf_Die *vr_die, char *buf, int len) | ||
| 769 | { | 254 | { |
| 770 | int ret, ret2; | 255 | if (self) { |
| 771 | 256 | if (self->dwfl) | |
| 772 | ret = die_get_typename(vr_die, buf, len); | 257 | dwfl_end(self->dwfl); |
| 773 | if (ret < 0) { | 258 | free(self); |
| 774 | pr_debug("Failed to get type, make it unknown.\n"); | ||
| 775 | ret = snprintf(buf, len, "(unknown_type)"); | ||
| 776 | } | 259 | } |
| 777 | if (ret > 0) { | ||
| 778 | ret2 = snprintf(buf + ret, len - ret, "\t%s", | ||
| 779 | dwarf_diename(vr_die)); | ||
| 780 | ret = (ret2 >= len - ret) ? -E2BIG : ret2 + ret; | ||
| 781 | } | ||
| 782 | return ret; | ||
| 783 | } | 260 | } |
| 784 | 261 | ||
| 785 | /* | 262 | /* |
| @@ -897,6 +374,7 @@ static int convert_variable_type(Dwarf_Die *vr_die, | |||
| 897 | struct probe_trace_arg_ref **ref_ptr = &tvar->ref; | 374 | struct probe_trace_arg_ref **ref_ptr = &tvar->ref; |
| 898 | Dwarf_Die type; | 375 | Dwarf_Die type; |
| 899 | char buf[16]; | 376 | char buf[16]; |
| 377 | int bsize, boffs, total; | ||
| 900 | int ret; | 378 | int ret; |
| 901 | 379 | ||
| 902 | /* TODO: check all types */ | 380 | /* TODO: check all types */ |
| @@ -906,11 +384,15 @@ static int convert_variable_type(Dwarf_Die *vr_die, | |||
| 906 | return (tvar->type == NULL) ? -ENOMEM : 0; | 384 | return (tvar->type == NULL) ? -ENOMEM : 0; |
| 907 | } | 385 | } |
| 908 | 386 | ||
| 909 | if (die_get_bit_size(vr_die) != 0) { | 387 | bsize = dwarf_bitsize(vr_die); |
| 388 | if (bsize > 0) { | ||
| 910 | /* This is a bitfield */ | 389 | /* This is a bitfield */ |
| 911 | ret = snprintf(buf, 16, "b%d@%d/%zd", die_get_bit_size(vr_die), | 390 | boffs = dwarf_bitoffset(vr_die); |
| 912 | die_get_bit_offset(vr_die), | 391 | total = dwarf_bytesize(vr_die); |
| 913 | BYTES_TO_BITS(die_get_byte_size(vr_die))); | 392 | if (boffs < 0 || total < 0) |
| 393 | return -ENOENT; | ||
| 394 | ret = snprintf(buf, 16, "b%d@%d/%zd", bsize, boffs, | ||
| 395 | BYTES_TO_BITS(total)); | ||
| 914 | goto formatted; | 396 | goto formatted; |
| 915 | } | 397 | } |
| 916 | 398 | ||
| @@ -958,10 +440,11 @@ static int convert_variable_type(Dwarf_Die *vr_die, | |||
| 958 | return (tvar->type == NULL) ? -ENOMEM : 0; | 440 | return (tvar->type == NULL) ? -ENOMEM : 0; |
| 959 | } | 441 | } |
| 960 | 442 | ||
| 961 | ret = BYTES_TO_BITS(die_get_byte_size(&type)); | 443 | ret = dwarf_bytesize(&type); |
| 962 | if (!ret) | 444 | if (ret <= 0) |
| 963 | /* No size ... try to use default type */ | 445 | /* No size ... try to use default type */ |
| 964 | return 0; | 446 | return 0; |
| 447 | ret = BYTES_TO_BITS(ret); | ||
| 965 | 448 | ||
| 966 | /* Check the bitwidth */ | 449 | /* Check the bitwidth */ |
| 967 | if (ret > MAX_BASIC_TYPE_BITS) { | 450 | if (ret > MAX_BASIC_TYPE_BITS) { |
| @@ -1025,7 +508,7 @@ static int convert_variable_fields(Dwarf_Die *vr_die, const char *varname, | |||
| 1025 | else | 508 | else |
| 1026 | *ref_ptr = ref; | 509 | *ref_ptr = ref; |
| 1027 | } | 510 | } |
| 1028 | ref->offset += die_get_byte_size(&type) * field->index; | 511 | ref->offset += dwarf_bytesize(&type) * field->index; |
| 1029 | if (!field->next) | 512 | if (!field->next) |
| 1030 | /* Save vr_die for converting types */ | 513 | /* Save vr_die for converting types */ |
| 1031 | memcpy(die_mem, vr_die, sizeof(*die_mem)); | 514 | memcpy(die_mem, vr_die, sizeof(*die_mem)); |
| @@ -1245,8 +728,7 @@ static int call_probe_finder(Dwarf_Die *sp_die, struct probe_finder *pf) | |||
| 1245 | 728 | ||
| 1246 | /* If no real subprogram, find a real one */ | 729 | /* If no real subprogram, find a real one */ |
| 1247 | if (!sp_die || dwarf_tag(sp_die) != DW_TAG_subprogram) { | 730 | if (!sp_die || dwarf_tag(sp_die) != DW_TAG_subprogram) { |
| 1248 | sp_die = die_find_real_subprogram(&pf->cu_die, | 731 | sp_die = die_find_realfunc(&pf->cu_die, pf->addr, &die_mem); |
| 1249 | pf->addr, &die_mem); | ||
| 1250 | if (!sp_die) { | 732 | if (!sp_die) { |
| 1251 | pr_warning("Failed to find probe point in any " | 733 | pr_warning("Failed to find probe point in any " |
| 1252 | "functions.\n"); | 734 | "functions.\n"); |
| @@ -1504,28 +986,18 @@ static int pubname_search_cb(Dwarf *dbg, Dwarf_Global *gl, void *data) | |||
| 1504 | } | 986 | } |
| 1505 | 987 | ||
| 1506 | /* Find probe points from debuginfo */ | 988 | /* Find probe points from debuginfo */ |
| 1507 | static int find_probes(int fd, struct probe_finder *pf) | 989 | static int debuginfo__find_probes(struct debuginfo *self, |
| 990 | struct probe_finder *pf) | ||
| 1508 | { | 991 | { |
| 1509 | struct perf_probe_point *pp = &pf->pev->point; | 992 | struct perf_probe_point *pp = &pf->pev->point; |
| 1510 | Dwarf_Off off, noff; | 993 | Dwarf_Off off, noff; |
| 1511 | size_t cuhl; | 994 | size_t cuhl; |
| 1512 | Dwarf_Die *diep; | 995 | Dwarf_Die *diep; |
| 1513 | Dwarf *dbg = NULL; | ||
| 1514 | Dwfl *dwfl; | ||
| 1515 | Dwarf_Addr bias; /* Currently ignored */ | ||
| 1516 | int ret = 0; | 996 | int ret = 0; |
| 1517 | 997 | ||
| 1518 | dbg = dwfl_init_offline_dwarf(fd, &dwfl, &bias); | ||
| 1519 | if (!dbg) { | ||
| 1520 | pr_warning("No debug information found in the vmlinux - " | ||
| 1521 | "please rebuild with CONFIG_DEBUG_INFO=y.\n"); | ||
| 1522 | close(fd); /* Without dwfl_end(), fd isn't closed. */ | ||
| 1523 | return -EBADF; | ||
| 1524 | } | ||
| 1525 | |||
| 1526 | #if _ELFUTILS_PREREQ(0, 142) | 998 | #if _ELFUTILS_PREREQ(0, 142) |
| 1527 | /* Get the call frame information from this dwarf */ | 999 | /* Get the call frame information from this dwarf */ |
| 1528 | pf->cfi = dwarf_getcfi(dbg); | 1000 | pf->cfi = dwarf_getcfi(self->dbg); |
| 1529 | #endif | 1001 | #endif |
| 1530 | 1002 | ||
| 1531 | off = 0; | 1003 | off = 0; |
| @@ -1544,7 +1016,8 @@ static int find_probes(int fd, struct probe_finder *pf) | |||
| 1544 | .data = pf, | 1016 | .data = pf, |
| 1545 | }; | 1017 | }; |
| 1546 | 1018 | ||
| 1547 | dwarf_getpubnames(dbg, pubname_search_cb, &pubname_param, 0); | 1019 | dwarf_getpubnames(self->dbg, pubname_search_cb, |
| 1020 | &pubname_param, 0); | ||
| 1548 | if (pubname_param.found) { | 1021 | if (pubname_param.found) { |
| 1549 | ret = probe_point_search_cb(&pf->sp_die, &probe_param); | 1022 | ret = probe_point_search_cb(&pf->sp_die, &probe_param); |
| 1550 | if (ret) | 1023 | if (ret) |
| @@ -1553,9 +1026,9 @@ static int find_probes(int fd, struct probe_finder *pf) | |||
| 1553 | } | 1026 | } |
| 1554 | 1027 | ||
| 1555 | /* Loop on CUs (Compilation Unit) */ | 1028 | /* Loop on CUs (Compilation Unit) */ |
| 1556 | while (!dwarf_nextcu(dbg, off, &noff, &cuhl, NULL, NULL, NULL)) { | 1029 | while (!dwarf_nextcu(self->dbg, off, &noff, &cuhl, NULL, NULL, NULL)) { |
| 1557 | /* Get the DIE(Debugging Information Entry) of this CU */ | 1030 | /* Get the DIE(Debugging Information Entry) of this CU */ |
| 1558 | diep = dwarf_offdie(dbg, off + cuhl, &pf->cu_die); | 1031 | diep = dwarf_offdie(self->dbg, off + cuhl, &pf->cu_die); |
| 1559 | if (!diep) | 1032 | if (!diep) |
| 1560 | continue; | 1033 | continue; |
| 1561 | 1034 | ||
| @@ -1582,8 +1055,6 @@ static int find_probes(int fd, struct probe_finder *pf) | |||
| 1582 | 1055 | ||
| 1583 | found: | 1056 | found: |
| 1584 | line_list__free(&pf->lcache); | 1057 | line_list__free(&pf->lcache); |
| 1585 | if (dwfl) | ||
| 1586 | dwfl_end(dwfl); | ||
| 1587 | 1058 | ||
| 1588 | return ret; | 1059 | return ret; |
| 1589 | } | 1060 | } |
| @@ -1629,8 +1100,9 @@ static int add_probe_trace_event(Dwarf_Die *sp_die, struct probe_finder *pf) | |||
| 1629 | } | 1100 | } |
| 1630 | 1101 | ||
| 1631 | /* Find probe_trace_events specified by perf_probe_event from debuginfo */ | 1102 | /* Find probe_trace_events specified by perf_probe_event from debuginfo */ |
| 1632 | int find_probe_trace_events(int fd, struct perf_probe_event *pev, | 1103 | int debuginfo__find_trace_events(struct debuginfo *self, |
| 1633 | struct probe_trace_event **tevs, int max_tevs) | 1104 | struct perf_probe_event *pev, |
| 1105 | struct probe_trace_event **tevs, int max_tevs) | ||
| 1634 | { | 1106 | { |
| 1635 | struct trace_event_finder tf = { | 1107 | struct trace_event_finder tf = { |
| 1636 | .pf = {.pev = pev, .callback = add_probe_trace_event}, | 1108 | .pf = {.pev = pev, .callback = add_probe_trace_event}, |
| @@ -1645,7 +1117,7 @@ int find_probe_trace_events(int fd, struct perf_probe_event *pev, | |||
| 1645 | tf.tevs = *tevs; | 1117 | tf.tevs = *tevs; |
| 1646 | tf.ntevs = 0; | 1118 | tf.ntevs = 0; |
| 1647 | 1119 | ||
| 1648 | ret = find_probes(fd, &tf.pf); | 1120 | ret = debuginfo__find_probes(self, &tf.pf); |
| 1649 | if (ret < 0) { | 1121 | if (ret < 0) { |
| 1650 | free(*tevs); | 1122 | free(*tevs); |
| 1651 | *tevs = NULL; | 1123 | *tevs = NULL; |
| @@ -1739,9 +1211,10 @@ out: | |||
| 1739 | } | 1211 | } |
| 1740 | 1212 | ||
| 1741 | /* Find available variables at given probe point */ | 1213 | /* Find available variables at given probe point */ |
| 1742 | int find_available_vars_at(int fd, struct perf_probe_event *pev, | 1214 | int debuginfo__find_available_vars_at(struct debuginfo *self, |
| 1743 | struct variable_list **vls, int max_vls, | 1215 | struct perf_probe_event *pev, |
| 1744 | bool externs) | 1216 | struct variable_list **vls, |
| 1217 | int max_vls, bool externs) | ||
| 1745 | { | 1218 | { |
| 1746 | struct available_var_finder af = { | 1219 | struct available_var_finder af = { |
| 1747 | .pf = {.pev = pev, .callback = add_available_vars}, | 1220 | .pf = {.pev = pev, .callback = add_available_vars}, |
| @@ -1756,7 +1229,7 @@ int find_available_vars_at(int fd, struct perf_probe_event *pev, | |||
| 1756 | af.vls = *vls; | 1229 | af.vls = *vls; |
| 1757 | af.nvls = 0; | 1230 | af.nvls = 0; |
| 1758 | 1231 | ||
| 1759 | ret = find_probes(fd, &af.pf); | 1232 | ret = debuginfo__find_probes(self, &af.pf); |
| 1760 | if (ret < 0) { | 1233 | if (ret < 0) { |
| 1761 | /* Free vlist for error */ | 1234 | /* Free vlist for error */ |
| 1762 | while (af.nvls--) { | 1235 | while (af.nvls--) { |
| @@ -1774,28 +1247,19 @@ int find_available_vars_at(int fd, struct perf_probe_event *pev, | |||
| 1774 | } | 1247 | } |
| 1775 | 1248 | ||
| 1776 | /* Reverse search */ | 1249 | /* Reverse search */ |
| 1777 | int find_perf_probe_point(unsigned long addr, struct perf_probe_point *ppt) | 1250 | int debuginfo__find_probe_point(struct debuginfo *self, unsigned long addr, |
| 1251 | struct perf_probe_point *ppt) | ||
| 1778 | { | 1252 | { |
| 1779 | Dwarf_Die cudie, spdie, indie; | 1253 | Dwarf_Die cudie, spdie, indie; |
| 1780 | Dwarf *dbg = NULL; | 1254 | Dwarf_Addr _addr, baseaddr; |
| 1781 | Dwfl *dwfl = NULL; | ||
| 1782 | Dwarf_Addr _addr, baseaddr, bias = 0; | ||
| 1783 | const char *fname = NULL, *func = NULL, *tmp; | 1255 | const char *fname = NULL, *func = NULL, *tmp; |
| 1784 | int baseline = 0, lineno = 0, ret = 0; | 1256 | int baseline = 0, lineno = 0, ret = 0; |
| 1785 | 1257 | ||
| 1786 | /* Open the live linux kernel */ | ||
| 1787 | dbg = dwfl_init_live_kernel_dwarf(addr, &dwfl, &bias); | ||
| 1788 | if (!dbg) { | ||
| 1789 | pr_warning("No debug information found in the vmlinux - " | ||
| 1790 | "please rebuild with CONFIG_DEBUG_INFO=y.\n"); | ||
| 1791 | ret = -EINVAL; | ||
| 1792 | goto end; | ||
| 1793 | } | ||
| 1794 | |||
| 1795 | /* Adjust address with bias */ | 1258 | /* Adjust address with bias */ |
| 1796 | addr += bias; | 1259 | addr += self->bias; |
| 1260 | |||
| 1797 | /* Find cu die */ | 1261 | /* Find cu die */ |
| 1798 | if (!dwarf_addrdie(dbg, (Dwarf_Addr)addr - bias, &cudie)) { | 1262 | if (!dwarf_addrdie(self->dbg, (Dwarf_Addr)addr - self->bias, &cudie)) { |
| 1799 | pr_warning("Failed to find debug information for address %lx\n", | 1263 | pr_warning("Failed to find debug information for address %lx\n", |
| 1800 | addr); | 1264 | addr); |
| 1801 | ret = -EINVAL; | 1265 | ret = -EINVAL; |
| @@ -1807,7 +1271,7 @@ int find_perf_probe_point(unsigned long addr, struct perf_probe_point *ppt) | |||
| 1807 | /* Don't care whether it failed or not */ | 1271 | /* Don't care whether it failed or not */ |
| 1808 | 1272 | ||
| 1809 | /* Find a corresponding function (name, baseline and baseaddr) */ | 1273 | /* Find a corresponding function (name, baseline and baseaddr) */ |
| 1810 | if (die_find_real_subprogram(&cudie, (Dwarf_Addr)addr, &spdie)) { | 1274 | if (die_find_realfunc(&cudie, (Dwarf_Addr)addr, &spdie)) { |
| 1811 | /* Get function entry information */ | 1275 | /* Get function entry information */ |
| 1812 | tmp = dwarf_diename(&spdie); | 1276 | tmp = dwarf_diename(&spdie); |
| 1813 | if (!tmp || | 1277 | if (!tmp || |
| @@ -1871,8 +1335,6 @@ post: | |||
| 1871 | } | 1335 | } |
| 1872 | } | 1336 | } |
| 1873 | end: | 1337 | end: |
| 1874 | if (dwfl) | ||
| 1875 | dwfl_end(dwfl); | ||
| 1876 | if (ret == 0 && (fname || func)) | 1338 | if (ret == 0 && (fname || func)) |
| 1877 | ret = 1; /* Found a point */ | 1339 | ret = 1; /* Found a point */ |
| 1878 | return ret; | 1340 | return ret; |
| @@ -1982,26 +1444,15 @@ static int find_line_range_by_func(struct line_finder *lf) | |||
| 1982 | return param.retval; | 1444 | return param.retval; |
| 1983 | } | 1445 | } |
| 1984 | 1446 | ||
| 1985 | int find_line_range(int fd, struct line_range *lr) | 1447 | int debuginfo__find_line_range(struct debuginfo *self, struct line_range *lr) |
| 1986 | { | 1448 | { |
| 1987 | struct line_finder lf = {.lr = lr, .found = 0}; | 1449 | struct line_finder lf = {.lr = lr, .found = 0}; |
| 1988 | int ret = 0; | 1450 | int ret = 0; |
| 1989 | Dwarf_Off off = 0, noff; | 1451 | Dwarf_Off off = 0, noff; |
| 1990 | size_t cuhl; | 1452 | size_t cuhl; |
| 1991 | Dwarf_Die *diep; | 1453 | Dwarf_Die *diep; |
| 1992 | Dwarf *dbg = NULL; | ||
| 1993 | Dwfl *dwfl; | ||
| 1994 | Dwarf_Addr bias; /* Currently ignored */ | ||
| 1995 | const char *comp_dir; | 1454 | const char *comp_dir; |
| 1996 | 1455 | ||
| 1997 | dbg = dwfl_init_offline_dwarf(fd, &dwfl, &bias); | ||
| 1998 | if (!dbg) { | ||
| 1999 | pr_warning("No debug information found in the vmlinux - " | ||
| 2000 | "please rebuild with CONFIG_DEBUG_INFO=y.\n"); | ||
| 2001 | close(fd); /* Without dwfl_end(), fd isn't closed. */ | ||
| 2002 | return -EBADF; | ||
| 2003 | } | ||
| 2004 | |||
| 2005 | /* Fastpath: lookup by function name from .debug_pubnames section */ | 1456 | /* Fastpath: lookup by function name from .debug_pubnames section */ |
| 2006 | if (lr->function) { | 1457 | if (lr->function) { |
| 2007 | struct pubname_callback_param pubname_param = { | 1458 | struct pubname_callback_param pubname_param = { |
| @@ -2010,7 +1461,8 @@ int find_line_range(int fd, struct line_range *lr) | |||
| 2010 | struct dwarf_callback_param line_range_param = { | 1461 | struct dwarf_callback_param line_range_param = { |
| 2011 | .data = (void *)&lf, .retval = 0}; | 1462 | .data = (void *)&lf, .retval = 0}; |
| 2012 | 1463 | ||
| 2013 | dwarf_getpubnames(dbg, pubname_search_cb, &pubname_param, 0); | 1464 | dwarf_getpubnames(self->dbg, pubname_search_cb, |
| 1465 | &pubname_param, 0); | ||
| 2014 | if (pubname_param.found) { | 1466 | if (pubname_param.found) { |
| 2015 | line_range_search_cb(&lf.sp_die, &line_range_param); | 1467 | line_range_search_cb(&lf.sp_die, &line_range_param); |
| 2016 | if (lf.found) | 1468 | if (lf.found) |
| @@ -2020,11 +1472,12 @@ int find_line_range(int fd, struct line_range *lr) | |||
| 2020 | 1472 | ||
| 2021 | /* Loop on CUs (Compilation Unit) */ | 1473 | /* Loop on CUs (Compilation Unit) */ |
| 2022 | while (!lf.found && ret >= 0) { | 1474 | while (!lf.found && ret >= 0) { |
| 2023 | if (dwarf_nextcu(dbg, off, &noff, &cuhl, NULL, NULL, NULL) != 0) | 1475 | if (dwarf_nextcu(self->dbg, off, &noff, &cuhl, |
| 1476 | NULL, NULL, NULL) != 0) | ||
| 2024 | break; | 1477 | break; |
| 2025 | 1478 | ||
| 2026 | /* Get the DIE(Debugging Information Entry) of this CU */ | 1479 | /* Get the DIE(Debugging Information Entry) of this CU */ |
| 2027 | diep = dwarf_offdie(dbg, off + cuhl, &lf.cu_die); | 1480 | diep = dwarf_offdie(self->dbg, off + cuhl, &lf.cu_die); |
| 2028 | if (!diep) | 1481 | if (!diep) |
| 2029 | continue; | 1482 | continue; |
| 2030 | 1483 | ||
| @@ -2058,7 +1511,6 @@ found: | |||
| 2058 | } | 1511 | } |
| 2059 | 1512 | ||
| 2060 | pr_debug("path: %s\n", lr->path); | 1513 | pr_debug("path: %s\n", lr->path); |
| 2061 | dwfl_end(dwfl); | ||
| 2062 | return (ret < 0) ? ret : lf.found; | 1514 | return (ret < 0) ? ret : lf.found; |
| 2063 | } | 1515 | } |
| 2064 | 1516 | ||
diff --git a/tools/perf/util/probe-finder.h b/tools/perf/util/probe-finder.h index 605730a366d..c478b42a247 100644 --- a/tools/perf/util/probe-finder.h +++ b/tools/perf/util/probe-finder.h | |||
| @@ -16,27 +16,42 @@ static inline int is_c_varname(const char *name) | |||
| 16 | } | 16 | } |
| 17 | 17 | ||
| 18 | #ifdef DWARF_SUPPORT | 18 | #ifdef DWARF_SUPPORT |
| 19 | |||
| 20 | #include "dwarf-aux.h" | ||
| 21 | |||
| 22 | /* TODO: export debuginfo data structure even if no dwarf support */ | ||
| 23 | |||
| 24 | /* debug information structure */ | ||
| 25 | struct debuginfo { | ||
| 26 | Dwarf *dbg; | ||
| 27 | Dwfl *dwfl; | ||
| 28 | Dwarf_Addr bias; | ||
| 29 | }; | ||
| 30 | |||
| 31 | extern struct debuginfo *debuginfo__new(const char *path); | ||
| 32 | extern struct debuginfo *debuginfo__new_online_kernel(unsigned long addr); | ||
| 33 | extern void debuginfo__delete(struct debuginfo *self); | ||
| 34 | |||
| 19 | /* Find probe_trace_events specified by perf_probe_event from debuginfo */ | 35 | /* Find probe_trace_events specified by perf_probe_event from debuginfo */ |
| 20 | extern int find_probe_trace_events(int fd, struct perf_probe_event *pev, | 36 | extern int debuginfo__find_trace_events(struct debuginfo *self, |
| 21 | struct probe_trace_event **tevs, | 37 | struct perf_probe_event *pev, |
| 22 | int max_tevs); | 38 | struct probe_trace_event **tevs, |
| 39 | int max_tevs); | ||
| 23 | 40 | ||
| 24 | /* Find a perf_probe_point from debuginfo */ | 41 | /* Find a perf_probe_point from debuginfo */ |
| 25 | extern int find_perf_probe_point(unsigned long addr, | 42 | extern int debuginfo__find_probe_point(struct debuginfo *self, |
| 26 | struct perf_probe_point *ppt); | 43 | unsigned long addr, |
| 44 | struct perf_probe_point *ppt); | ||
| 27 | 45 | ||
| 28 | /* Find a line range */ | 46 | /* Find a line range */ |
| 29 | extern int find_line_range(int fd, struct line_range *lr); | 47 | extern int debuginfo__find_line_range(struct debuginfo *self, |
| 48 | struct line_range *lr); | ||
| 30 | 49 | ||
| 31 | /* Find available variables */ | 50 | /* Find available variables */ |
| 32 | extern int find_available_vars_at(int fd, struct perf_probe_event *pev, | 51 | extern int debuginfo__find_available_vars_at(struct debuginfo *self, |
| 33 | struct variable_list **vls, int max_points, | 52 | struct perf_probe_event *pev, |
| 34 | bool externs); | 53 | struct variable_list **vls, |
| 35 | 54 | int max_points, bool externs); | |
| 36 | #include <dwarf.h> | ||
| 37 | #include <elfutils/libdw.h> | ||
| 38 | #include <elfutils/libdwfl.h> | ||
| 39 | #include <elfutils/version.h> | ||
| 40 | 55 | ||
| 41 | struct probe_finder { | 56 | struct probe_finder { |
| 42 | struct perf_probe_event *pev; /* Target probe event */ | 57 | struct perf_probe_event *pev; /* Target probe event */ |
diff --git a/tools/perf/util/python.c b/tools/perf/util/python.c index a9ac0504aab..8e0b5a39d8a 100644 --- a/tools/perf/util/python.c +++ b/tools/perf/util/python.c | |||
| @@ -247,7 +247,7 @@ struct pyrf_cpu_map { | |||
| 247 | static int pyrf_cpu_map__init(struct pyrf_cpu_map *pcpus, | 247 | static int pyrf_cpu_map__init(struct pyrf_cpu_map *pcpus, |
| 248 | PyObject *args, PyObject *kwargs) | 248 | PyObject *args, PyObject *kwargs) |
| 249 | { | 249 | { |
| 250 | static char *kwlist[] = { "cpustr", NULL, NULL, }; | 250 | static char *kwlist[] = { "cpustr", NULL }; |
| 251 | char *cpustr = NULL; | 251 | char *cpustr = NULL; |
| 252 | 252 | ||
| 253 | if (!PyArg_ParseTupleAndKeywords(args, kwargs, "|s", | 253 | if (!PyArg_ParseTupleAndKeywords(args, kwargs, "|s", |
| @@ -316,7 +316,7 @@ struct pyrf_thread_map { | |||
| 316 | static int pyrf_thread_map__init(struct pyrf_thread_map *pthreads, | 316 | static int pyrf_thread_map__init(struct pyrf_thread_map *pthreads, |
| 317 | PyObject *args, PyObject *kwargs) | 317 | PyObject *args, PyObject *kwargs) |
| 318 | { | 318 | { |
| 319 | static char *kwlist[] = { "pid", "tid", NULL, NULL, }; | 319 | static char *kwlist[] = { "pid", "tid", NULL }; |
| 320 | int pid = -1, tid = -1; | 320 | int pid = -1, tid = -1; |
| 321 | 321 | ||
| 322 | if (!PyArg_ParseTupleAndKeywords(args, kwargs, "|ii", | 322 | if (!PyArg_ParseTupleAndKeywords(args, kwargs, "|ii", |
| @@ -418,7 +418,9 @@ static int pyrf_evsel__init(struct pyrf_evsel *pevsel, | |||
| 418 | "wakeup_events", | 418 | "wakeup_events", |
| 419 | "bp_type", | 419 | "bp_type", |
| 420 | "bp_addr", | 420 | "bp_addr", |
| 421 | "bp_len", NULL, NULL, }; | 421 | "bp_len", |
| 422 | NULL | ||
| 423 | }; | ||
| 422 | u64 sample_period = 0; | 424 | u64 sample_period = 0; |
| 423 | u32 disabled = 0, | 425 | u32 disabled = 0, |
| 424 | inherit = 0, | 426 | inherit = 0, |
| @@ -499,7 +501,7 @@ static PyObject *pyrf_evsel__open(struct pyrf_evsel *pevsel, | |||
| 499 | struct thread_map *threads = NULL; | 501 | struct thread_map *threads = NULL; |
| 500 | PyObject *pcpus = NULL, *pthreads = NULL; | 502 | PyObject *pcpus = NULL, *pthreads = NULL; |
| 501 | int group = 0, inherit = 0; | 503 | int group = 0, inherit = 0; |
| 502 | static char *kwlist[] = {"cpus", "threads", "group", "inherit", NULL, NULL}; | 504 | static char *kwlist[] = { "cpus", "threads", "group", "inherit", NULL }; |
| 503 | 505 | ||
| 504 | if (!PyArg_ParseTupleAndKeywords(args, kwargs, "|OOii", kwlist, | 506 | if (!PyArg_ParseTupleAndKeywords(args, kwargs, "|OOii", kwlist, |
| 505 | &pcpus, &pthreads, &group, &inherit)) | 507 | &pcpus, &pthreads, &group, &inherit)) |
| @@ -582,8 +584,7 @@ static PyObject *pyrf_evlist__mmap(struct pyrf_evlist *pevlist, | |||
| 582 | PyObject *args, PyObject *kwargs) | 584 | PyObject *args, PyObject *kwargs) |
| 583 | { | 585 | { |
| 584 | struct perf_evlist *evlist = &pevlist->evlist; | 586 | struct perf_evlist *evlist = &pevlist->evlist; |
| 585 | static char *kwlist[] = {"pages", "overwrite", | 587 | static char *kwlist[] = { "pages", "overwrite", NULL }; |
| 586 | NULL, NULL}; | ||
| 587 | int pages = 128, overwrite = false; | 588 | int pages = 128, overwrite = false; |
| 588 | 589 | ||
| 589 | if (!PyArg_ParseTupleAndKeywords(args, kwargs, "|ii", kwlist, | 590 | if (!PyArg_ParseTupleAndKeywords(args, kwargs, "|ii", kwlist, |
| @@ -603,7 +604,7 @@ static PyObject *pyrf_evlist__poll(struct pyrf_evlist *pevlist, | |||
| 603 | PyObject *args, PyObject *kwargs) | 604 | PyObject *args, PyObject *kwargs) |
| 604 | { | 605 | { |
| 605 | struct perf_evlist *evlist = &pevlist->evlist; | 606 | struct perf_evlist *evlist = &pevlist->evlist; |
| 606 | static char *kwlist[] = {"timeout", NULL, NULL}; | 607 | static char *kwlist[] = { "timeout", NULL }; |
| 607 | int timeout = -1, n; | 608 | int timeout = -1, n; |
| 608 | 609 | ||
| 609 | if (!PyArg_ParseTupleAndKeywords(args, kwargs, "|i", kwlist, &timeout)) | 610 | if (!PyArg_ParseTupleAndKeywords(args, kwargs, "|i", kwlist, &timeout)) |
| @@ -674,7 +675,7 @@ static PyObject *pyrf_evlist__read_on_cpu(struct pyrf_evlist *pevlist, | |||
| 674 | struct perf_evlist *evlist = &pevlist->evlist; | 675 | struct perf_evlist *evlist = &pevlist->evlist; |
| 675 | union perf_event *event; | 676 | union perf_event *event; |
| 676 | int sample_id_all = 1, cpu; | 677 | int sample_id_all = 1, cpu; |
| 677 | static char *kwlist[] = {"cpu", "sample_id_all", NULL, NULL}; | 678 | static char *kwlist[] = { "cpu", "sample_id_all", NULL }; |
| 678 | int err; | 679 | int err; |
| 679 | 680 | ||
| 680 | if (!PyArg_ParseTupleAndKeywords(args, kwargs, "i|i", kwlist, | 681 | if (!PyArg_ParseTupleAndKeywords(args, kwargs, "i|i", kwlist, |
diff --git a/tools/perf/util/session.c b/tools/perf/util/session.c index f5a8fbdd3f7..72458d9da5b 100644 --- a/tools/perf/util/session.c +++ b/tools/perf/util/session.c | |||
| @@ -12,6 +12,7 @@ | |||
| 12 | #include "session.h" | 12 | #include "session.h" |
| 13 | #include "sort.h" | 13 | #include "sort.h" |
| 14 | #include "util.h" | 14 | #include "util.h" |
| 15 | #include "cpumap.h" | ||
| 15 | 16 | ||
| 16 | static int perf_session__open(struct perf_session *self, bool force) | 17 | static int perf_session__open(struct perf_session *self, bool force) |
| 17 | { | 18 | { |
| @@ -247,9 +248,14 @@ int perf_session__resolve_callchain(struct perf_session *self, | |||
| 247 | callchain_cursor_reset(&self->callchain_cursor); | 248 | callchain_cursor_reset(&self->callchain_cursor); |
| 248 | 249 | ||
| 249 | for (i = 0; i < chain->nr; i++) { | 250 | for (i = 0; i < chain->nr; i++) { |
| 250 | u64 ip = chain->ips[i]; | 251 | u64 ip; |
| 251 | struct addr_location al; | 252 | struct addr_location al; |
| 252 | 253 | ||
| 254 | if (callchain_param.order == ORDER_CALLEE) | ||
| 255 | ip = chain->ips[i]; | ||
| 256 | else | ||
| 257 | ip = chain->ips[chain->nr - i - 1]; | ||
| 258 | |||
| 253 | if (ip >= PERF_CONTEXT_MAX) { | 259 | if (ip >= PERF_CONTEXT_MAX) { |
| 254 | switch (ip) { | 260 | switch (ip) { |
| 255 | case PERF_CONTEXT_HV: | 261 | case PERF_CONTEXT_HV: |
| @@ -407,20 +413,26 @@ static void perf_event__read_swap(union perf_event *event) | |||
| 407 | event->read.id = bswap_64(event->read.id); | 413 | event->read.id = bswap_64(event->read.id); |
| 408 | } | 414 | } |
| 409 | 415 | ||
| 410 | static void perf_event__attr_swap(union perf_event *event) | 416 | /* exported for swapping attributes in file header */ |
| 417 | void perf_event__attr_swap(struct perf_event_attr *attr) | ||
| 418 | { | ||
| 419 | attr->type = bswap_32(attr->type); | ||
| 420 | attr->size = bswap_32(attr->size); | ||
| 421 | attr->config = bswap_64(attr->config); | ||
| 422 | attr->sample_period = bswap_64(attr->sample_period); | ||
| 423 | attr->sample_type = bswap_64(attr->sample_type); | ||
| 424 | attr->read_format = bswap_64(attr->read_format); | ||
| 425 | attr->wakeup_events = bswap_32(attr->wakeup_events); | ||
| 426 | attr->bp_type = bswap_32(attr->bp_type); | ||
| 427 | attr->bp_addr = bswap_64(attr->bp_addr); | ||
| 428 | attr->bp_len = bswap_64(attr->bp_len); | ||
| 429 | } | ||
| 430 | |||
| 431 | static void perf_event__hdr_attr_swap(union perf_event *event) | ||
| 411 | { | 432 | { |
| 412 | size_t size; | 433 | size_t size; |
| 413 | 434 | ||
| 414 | event->attr.attr.type = bswap_32(event->attr.attr.type); | 435 | perf_event__attr_swap(&event->attr.attr); |
| 415 | event->attr.attr.size = bswap_32(event->attr.attr.size); | ||
| 416 | event->attr.attr.config = bswap_64(event->attr.attr.config); | ||
| 417 | event->attr.attr.sample_period = bswap_64(event->attr.attr.sample_period); | ||
| 418 | event->attr.attr.sample_type = bswap_64(event->attr.attr.sample_type); | ||
| 419 | event->attr.attr.read_format = bswap_64(event->attr.attr.read_format); | ||
| 420 | event->attr.attr.wakeup_events = bswap_32(event->attr.attr.wakeup_events); | ||
| 421 | event->attr.attr.bp_type = bswap_32(event->attr.attr.bp_type); | ||
| 422 | event->attr.attr.bp_addr = bswap_64(event->attr.attr.bp_addr); | ||
| 423 | event->attr.attr.bp_len = bswap_64(event->attr.attr.bp_len); | ||
| 424 | 436 | ||
| 425 | size = event->header.size; | 437 | size = event->header.size; |
| 426 | size -= (void *)&event->attr.id - (void *)event; | 438 | size -= (void *)&event->attr.id - (void *)event; |
| @@ -448,7 +460,7 @@ static perf_event__swap_op perf_event__swap_ops[] = { | |||
| 448 | [PERF_RECORD_LOST] = perf_event__all64_swap, | 460 | [PERF_RECORD_LOST] = perf_event__all64_swap, |
| 449 | [PERF_RECORD_READ] = perf_event__read_swap, | 461 | [PERF_RECORD_READ] = perf_event__read_swap, |
| 450 | [PERF_RECORD_SAMPLE] = perf_event__all64_swap, | 462 | [PERF_RECORD_SAMPLE] = perf_event__all64_swap, |
| 451 | [PERF_RECORD_HEADER_ATTR] = perf_event__attr_swap, | 463 | [PERF_RECORD_HEADER_ATTR] = perf_event__hdr_attr_swap, |
| 452 | [PERF_RECORD_HEADER_EVENT_TYPE] = perf_event__event_type_swap, | 464 | [PERF_RECORD_HEADER_EVENT_TYPE] = perf_event__event_type_swap, |
| 453 | [PERF_RECORD_HEADER_TRACING_DATA] = perf_event__tracing_data_swap, | 465 | [PERF_RECORD_HEADER_TRACING_DATA] = perf_event__tracing_data_swap, |
| 454 | [PERF_RECORD_HEADER_BUILD_ID] = NULL, | 466 | [PERF_RECORD_HEADER_BUILD_ID] = NULL, |
| @@ -708,9 +720,9 @@ static void dump_sample(struct perf_session *session, union perf_event *event, | |||
| 708 | if (!dump_trace) | 720 | if (!dump_trace) |
| 709 | return; | 721 | return; |
| 710 | 722 | ||
| 711 | printf("(IP, %d): %d/%d: %#" PRIx64 " period: %" PRIu64 "\n", | 723 | printf("(IP, %d): %d/%d: %#" PRIx64 " period: %" PRIu64 " addr: %#" PRIx64 "\n", |
| 712 | event->header.misc, sample->pid, sample->tid, sample->ip, | 724 | event->header.misc, sample->pid, sample->tid, sample->ip, |
| 713 | sample->period); | 725 | sample->period, sample->addr); |
| 714 | 726 | ||
| 715 | if (session->sample_type & PERF_SAMPLE_CALLCHAIN) | 727 | if (session->sample_type & PERF_SAMPLE_CALLCHAIN) |
| 716 | callchain__printf(sample); | 728 | callchain__printf(sample); |
| @@ -1202,9 +1214,10 @@ struct perf_evsel *perf_session__find_first_evtype(struct perf_session *session, | |||
| 1202 | return NULL; | 1214 | return NULL; |
| 1203 | } | 1215 | } |
| 1204 | 1216 | ||
| 1205 | void perf_session__print_symbols(union perf_event *event, | 1217 | void perf_session__print_ip(union perf_event *event, |
| 1206 | struct perf_sample *sample, | 1218 | struct perf_sample *sample, |
| 1207 | struct perf_session *session) | 1219 | struct perf_session *session, |
| 1220 | int print_sym, int print_dso) | ||
| 1208 | { | 1221 | { |
| 1209 | struct addr_location al; | 1222 | struct addr_location al; |
| 1210 | const char *symname, *dsoname; | 1223 | const char *symname, *dsoname; |
| @@ -1233,32 +1246,83 @@ void perf_session__print_symbols(union perf_event *event, | |||
| 1233 | if (!node) | 1246 | if (!node) |
| 1234 | break; | 1247 | break; |
| 1235 | 1248 | ||
| 1236 | if (node->sym && node->sym->name) | 1249 | printf("\t%16" PRIx64, node->ip); |
| 1237 | symname = node->sym->name; | 1250 | if (print_sym) { |
| 1251 | if (node->sym && node->sym->name) | ||
| 1252 | symname = node->sym->name; | ||
| 1253 | else | ||
| 1254 | symname = ""; | ||
| 1255 | |||
| 1256 | printf(" %s", symname); | ||
| 1257 | } | ||
| 1258 | if (print_dso) { | ||
| 1259 | if (node->map && node->map->dso && node->map->dso->name) | ||
| 1260 | dsoname = node->map->dso->name; | ||
| 1261 | else | ||
| 1262 | dsoname = ""; | ||
| 1263 | |||
| 1264 | printf(" (%s)", dsoname); | ||
| 1265 | } | ||
| 1266 | printf("\n"); | ||
| 1267 | |||
| 1268 | callchain_cursor_advance(cursor); | ||
| 1269 | } | ||
| 1270 | |||
| 1271 | } else { | ||
| 1272 | printf("%16" PRIx64, sample->ip); | ||
| 1273 | if (print_sym) { | ||
| 1274 | if (al.sym && al.sym->name) | ||
| 1275 | symname = al.sym->name; | ||
| 1238 | else | 1276 | else |
| 1239 | symname = ""; | 1277 | symname = ""; |
| 1240 | 1278 | ||
| 1241 | if (node->map && node->map->dso && node->map->dso->name) | 1279 | printf(" %s", symname); |
| 1242 | dsoname = node->map->dso->name; | 1280 | } |
| 1281 | |||
| 1282 | if (print_dso) { | ||
| 1283 | if (al.map && al.map->dso && al.map->dso->name) | ||
| 1284 | dsoname = al.map->dso->name; | ||
| 1243 | else | 1285 | else |
| 1244 | dsoname = ""; | 1286 | dsoname = ""; |
| 1245 | 1287 | ||
| 1246 | printf("\t%16" PRIx64 " %s (%s)\n", node->ip, symname, dsoname); | 1288 | printf(" (%s)", dsoname); |
| 1289 | } | ||
| 1290 | } | ||
| 1291 | } | ||
| 1247 | 1292 | ||
| 1248 | callchain_cursor_advance(cursor); | 1293 | int perf_session__cpu_bitmap(struct perf_session *session, |
| 1294 | const char *cpu_list, unsigned long *cpu_bitmap) | ||
| 1295 | { | ||
| 1296 | int i; | ||
| 1297 | struct cpu_map *map; | ||
| 1298 | |||
| 1299 | for (i = 0; i < PERF_TYPE_MAX; ++i) { | ||
| 1300 | struct perf_evsel *evsel; | ||
| 1301 | |||
| 1302 | evsel = perf_session__find_first_evtype(session, i); | ||
| 1303 | if (!evsel) | ||
| 1304 | continue; | ||
| 1305 | |||
| 1306 | if (!(evsel->attr.sample_type & PERF_SAMPLE_CPU)) { | ||
| 1307 | pr_err("File does not contain CPU events. " | ||
| 1308 | "Remove -c option to proceed.\n"); | ||
| 1309 | return -1; | ||
| 1249 | } | 1310 | } |
| 1311 | } | ||
| 1250 | 1312 | ||
| 1251 | } else { | 1313 | map = cpu_map__new(cpu_list); |
| 1252 | if (al.sym && al.sym->name) | ||
| 1253 | symname = al.sym->name; | ||
| 1254 | else | ||
| 1255 | symname = ""; | ||
| 1256 | 1314 | ||
| 1257 | if (al.map && al.map->dso && al.map->dso->name) | 1315 | for (i = 0; i < map->nr; i++) { |
| 1258 | dsoname = al.map->dso->name; | 1316 | int cpu = map->map[i]; |
| 1259 | else | 1317 | |
| 1260 | dsoname = ""; | 1318 | if (cpu >= MAX_NR_CPUS) { |
| 1319 | pr_err("Requested CPU %d too large. " | ||
| 1320 | "Consider raising MAX_NR_CPUS\n", cpu); | ||
| 1321 | return -1; | ||
| 1322 | } | ||
| 1261 | 1323 | ||
| 1262 | printf("%16" PRIx64 " %s (%s)", al.addr, symname, dsoname); | 1324 | set_bit(cpu, cpu_bitmap); |
| 1263 | } | 1325 | } |
| 1326 | |||
| 1327 | return 0; | ||
| 1264 | } | 1328 | } |
diff --git a/tools/perf/util/session.h b/tools/perf/util/session.h index 66d4e149087..170601e67d6 100644 --- a/tools/perf/util/session.h +++ b/tools/perf/util/session.h | |||
| @@ -112,6 +112,7 @@ int perf_session__set_kallsyms_ref_reloc_sym(struct map **maps, | |||
| 112 | u64 addr); | 112 | u64 addr); |
| 113 | 113 | ||
| 114 | void mem_bswap_64(void *src, int byte_size); | 114 | void mem_bswap_64(void *src, int byte_size); |
| 115 | void perf_event__attr_swap(struct perf_event_attr *attr); | ||
| 115 | 116 | ||
| 116 | int perf_session__create_kernel_maps(struct perf_session *self); | 117 | int perf_session__create_kernel_maps(struct perf_session *self); |
| 117 | 118 | ||
| @@ -167,8 +168,12 @@ static inline int perf_session__parse_sample(struct perf_session *session, | |||
| 167 | struct perf_evsel *perf_session__find_first_evtype(struct perf_session *session, | 168 | struct perf_evsel *perf_session__find_first_evtype(struct perf_session *session, |
| 168 | unsigned int type); | 169 | unsigned int type); |
| 169 | 170 | ||
| 170 | void perf_session__print_symbols(union perf_event *event, | 171 | void perf_session__print_ip(union perf_event *event, |
| 171 | struct perf_sample *sample, | 172 | struct perf_sample *sample, |
| 172 | struct perf_session *session); | 173 | struct perf_session *session, |
| 174 | int print_sym, int print_dso); | ||
| 175 | |||
| 176 | int perf_session__cpu_bitmap(struct perf_session *session, | ||
| 177 | const char *cpu_list, unsigned long *cpu_bitmap); | ||
| 173 | 178 | ||
| 174 | #endif /* __PERF_SESSION_H */ | 179 | #endif /* __PERF_SESSION_H */ |
diff --git a/tools/perf/util/sort.c b/tools/perf/util/sort.c index f44fa541d56..401e220566f 100644 --- a/tools/perf/util/sort.c +++ b/tools/perf/util/sort.c | |||
| @@ -15,95 +15,6 @@ char * field_sep; | |||
| 15 | 15 | ||
| 16 | LIST_HEAD(hist_entry__sort_list); | 16 | LIST_HEAD(hist_entry__sort_list); |
| 17 | 17 | ||
| 18 | static int hist_entry__thread_snprintf(struct hist_entry *self, char *bf, | ||
| 19 | size_t size, unsigned int width); | ||
| 20 | static int hist_entry__comm_snprintf(struct hist_entry *self, char *bf, | ||
| 21 | size_t size, unsigned int width); | ||
| 22 | static int hist_entry__dso_snprintf(struct hist_entry *self, char *bf, | ||
| 23 | size_t size, unsigned int width); | ||
| 24 | static int hist_entry__sym_snprintf(struct hist_entry *self, char *bf, | ||
| 25 | size_t size, unsigned int width); | ||
| 26 | static int hist_entry__parent_snprintf(struct hist_entry *self, char *bf, | ||
| 27 | size_t size, unsigned int width); | ||
| 28 | static int hist_entry__cpu_snprintf(struct hist_entry *self, char *bf, | ||
| 29 | size_t size, unsigned int width); | ||
| 30 | |||
| 31 | struct sort_entry sort_thread = { | ||
| 32 | .se_header = "Command: Pid", | ||
| 33 | .se_cmp = sort__thread_cmp, | ||
| 34 | .se_snprintf = hist_entry__thread_snprintf, | ||
| 35 | .se_width_idx = HISTC_THREAD, | ||
| 36 | }; | ||
| 37 | |||
| 38 | struct sort_entry sort_comm = { | ||
| 39 | .se_header = "Command", | ||
| 40 | .se_cmp = sort__comm_cmp, | ||
| 41 | .se_collapse = sort__comm_collapse, | ||
| 42 | .se_snprintf = hist_entry__comm_snprintf, | ||
| 43 | .se_width_idx = HISTC_COMM, | ||
| 44 | }; | ||
| 45 | |||
| 46 | struct sort_entry sort_dso = { | ||
| 47 | .se_header = "Shared Object", | ||
| 48 | .se_cmp = sort__dso_cmp, | ||
| 49 | .se_snprintf = hist_entry__dso_snprintf, | ||
| 50 | .se_width_idx = HISTC_DSO, | ||
| 51 | }; | ||
| 52 | |||
| 53 | struct sort_entry sort_sym = { | ||
| 54 | .se_header = "Symbol", | ||
| 55 | .se_cmp = sort__sym_cmp, | ||
| 56 | .se_snprintf = hist_entry__sym_snprintf, | ||
| 57 | .se_width_idx = HISTC_SYMBOL, | ||
| 58 | }; | ||
| 59 | |||
| 60 | struct sort_entry sort_parent = { | ||
| 61 | .se_header = "Parent symbol", | ||
| 62 | .se_cmp = sort__parent_cmp, | ||
| 63 | .se_snprintf = hist_entry__parent_snprintf, | ||
| 64 | .se_width_idx = HISTC_PARENT, | ||
| 65 | }; | ||
| 66 | |||
| 67 | struct sort_entry sort_cpu = { | ||
| 68 | .se_header = "CPU", | ||
| 69 | .se_cmp = sort__cpu_cmp, | ||
| 70 | .se_snprintf = hist_entry__cpu_snprintf, | ||
| 71 | .se_width_idx = HISTC_CPU, | ||
| 72 | }; | ||
| 73 | |||
| 74 | struct sort_dimension { | ||
| 75 | const char *name; | ||
| 76 | struct sort_entry *entry; | ||
| 77 | int taken; | ||
| 78 | }; | ||
| 79 | |||
| 80 | static struct sort_dimension sort_dimensions[] = { | ||
| 81 | { .name = "pid", .entry = &sort_thread, }, | ||
| 82 | { .name = "comm", .entry = &sort_comm, }, | ||
| 83 | { .name = "dso", .entry = &sort_dso, }, | ||
| 84 | { .name = "symbol", .entry = &sort_sym, }, | ||
| 85 | { .name = "parent", .entry = &sort_parent, }, | ||
| 86 | { .name = "cpu", .entry = &sort_cpu, }, | ||
| 87 | }; | ||
| 88 | |||
| 89 | int64_t cmp_null(void *l, void *r) | ||
| 90 | { | ||
| 91 | if (!l && !r) | ||
| 92 | return 0; | ||
| 93 | else if (!l) | ||
| 94 | return -1; | ||
| 95 | else | ||
| 96 | return 1; | ||
| 97 | } | ||
| 98 | |||
| 99 | /* --sort pid */ | ||
| 100 | |||
| 101 | int64_t | ||
| 102 | sort__thread_cmp(struct hist_entry *left, struct hist_entry *right) | ||
| 103 | { | ||
| 104 | return right->thread->pid - left->thread->pid; | ||
| 105 | } | ||
| 106 | |||
| 107 | static int repsep_snprintf(char *bf, size_t size, const char *fmt, ...) | 18 | static int repsep_snprintf(char *bf, size_t size, const char *fmt, ...) |
| 108 | { | 19 | { |
| 109 | int n; | 20 | int n; |
| @@ -125,6 +36,24 @@ static int repsep_snprintf(char *bf, size_t size, const char *fmt, ...) | |||
| 125 | return n; | 36 | return n; |
| 126 | } | 37 | } |
| 127 | 38 | ||
| 39 | static int64_t cmp_null(void *l, void *r) | ||
| 40 | { | ||
| 41 | if (!l && !r) | ||
| 42 | return 0; | ||
| 43 | else if (!l) | ||
| 44 | return -1; | ||
| 45 | else | ||
| 46 | return 1; | ||
| 47 | } | ||
| 48 | |||
| 49 | /* --sort pid */ | ||
| 50 | |||
| 51 | static int64_t | ||
| 52 | sort__thread_cmp(struct hist_entry *left, struct hist_entry *right) | ||
| 53 | { | ||
| 54 | return right->thread->pid - left->thread->pid; | ||
| 55 | } | ||
| 56 | |||
| 128 | static int hist_entry__thread_snprintf(struct hist_entry *self, char *bf, | 57 | static int hist_entry__thread_snprintf(struct hist_entry *self, char *bf, |
| 129 | size_t size, unsigned int width) | 58 | size_t size, unsigned int width) |
| 130 | { | 59 | { |
| @@ -132,15 +61,50 @@ static int hist_entry__thread_snprintf(struct hist_entry *self, char *bf, | |||
| 132 | self->thread->comm ?: "", self->thread->pid); | 61 | self->thread->comm ?: "", self->thread->pid); |
| 133 | } | 62 | } |
| 134 | 63 | ||
| 64 | struct sort_entry sort_thread = { | ||
| 65 | .se_header = "Command: Pid", | ||
| 66 | .se_cmp = sort__thread_cmp, | ||
| 67 | .se_snprintf = hist_entry__thread_snprintf, | ||
| 68 | .se_width_idx = HISTC_THREAD, | ||
| 69 | }; | ||
| 70 | |||
| 71 | /* --sort comm */ | ||
| 72 | |||
| 73 | static int64_t | ||
| 74 | sort__comm_cmp(struct hist_entry *left, struct hist_entry *right) | ||
| 75 | { | ||
| 76 | return right->thread->pid - left->thread->pid; | ||
| 77 | } | ||
| 78 | |||
| 79 | static int64_t | ||
| 80 | sort__comm_collapse(struct hist_entry *left, struct hist_entry *right) | ||
| 81 | { | ||
| 82 | char *comm_l = left->thread->comm; | ||
| 83 | char *comm_r = right->thread->comm; | ||
| 84 | |||
| 85 | if (!comm_l || !comm_r) | ||
| 86 | return cmp_null(comm_l, comm_r); | ||
| 87 | |||
| 88 | return strcmp(comm_l, comm_r); | ||
| 89 | } | ||
| 90 | |||
| 135 | static int hist_entry__comm_snprintf(struct hist_entry *self, char *bf, | 91 | static int hist_entry__comm_snprintf(struct hist_entry *self, char *bf, |
| 136 | size_t size, unsigned int width) | 92 | size_t size, unsigned int width) |
| 137 | { | 93 | { |
| 138 | return repsep_snprintf(bf, size, "%*s", width, self->thread->comm); | 94 | return repsep_snprintf(bf, size, "%*s", width, self->thread->comm); |
| 139 | } | 95 | } |
| 140 | 96 | ||
| 97 | struct sort_entry sort_comm = { | ||
| 98 | .se_header = "Command", | ||
| 99 | .se_cmp = sort__comm_cmp, | ||
| 100 | .se_collapse = sort__comm_collapse, | ||
| 101 | .se_snprintf = hist_entry__comm_snprintf, | ||
| 102 | .se_width_idx = HISTC_COMM, | ||
| 103 | }; | ||
| 104 | |||
| 141 | /* --sort dso */ | 105 | /* --sort dso */ |
| 142 | 106 | ||
| 143 | int64_t | 107 | static int64_t |
| 144 | sort__dso_cmp(struct hist_entry *left, struct hist_entry *right) | 108 | sort__dso_cmp(struct hist_entry *left, struct hist_entry *right) |
| 145 | { | 109 | { |
| 146 | struct dso *dso_l = left->ms.map ? left->ms.map->dso : NULL; | 110 | struct dso *dso_l = left->ms.map ? left->ms.map->dso : NULL; |
| @@ -173,9 +137,16 @@ static int hist_entry__dso_snprintf(struct hist_entry *self, char *bf, | |||
| 173 | return repsep_snprintf(bf, size, "%-*s", width, "[unknown]"); | 137 | return repsep_snprintf(bf, size, "%-*s", width, "[unknown]"); |
| 174 | } | 138 | } |
| 175 | 139 | ||
| 140 | struct sort_entry sort_dso = { | ||
| 141 | .se_header = "Shared Object", | ||
| 142 | .se_cmp = sort__dso_cmp, | ||
| 143 | .se_snprintf = hist_entry__dso_snprintf, | ||
| 144 | .se_width_idx = HISTC_DSO, | ||
| 145 | }; | ||
| 146 | |||
| 176 | /* --sort symbol */ | 147 | /* --sort symbol */ |
| 177 | 148 | ||
| 178 | int64_t | 149 | static int64_t |
| 179 | sort__sym_cmp(struct hist_entry *left, struct hist_entry *right) | 150 | sort__sym_cmp(struct hist_entry *left, struct hist_entry *right) |
| 180 | { | 151 | { |
| 181 | u64 ip_l, ip_r; | 152 | u64 ip_l, ip_r; |
| @@ -211,29 +182,16 @@ static int hist_entry__sym_snprintf(struct hist_entry *self, char *bf, | |||
| 211 | return ret; | 182 | return ret; |
| 212 | } | 183 | } |
| 213 | 184 | ||
| 214 | /* --sort comm */ | 185 | struct sort_entry sort_sym = { |
| 215 | 186 | .se_header = "Symbol", | |
| 216 | int64_t | 187 | .se_cmp = sort__sym_cmp, |
| 217 | sort__comm_cmp(struct hist_entry *left, struct hist_entry *right) | 188 | .se_snprintf = hist_entry__sym_snprintf, |
| 218 | { | 189 | .se_width_idx = HISTC_SYMBOL, |
| 219 | return right->thread->pid - left->thread->pid; | 190 | }; |
| 220 | } | ||
| 221 | |||
| 222 | int64_t | ||
| 223 | sort__comm_collapse(struct hist_entry *left, struct hist_entry *right) | ||
| 224 | { | ||
| 225 | char *comm_l = left->thread->comm; | ||
| 226 | char *comm_r = right->thread->comm; | ||
| 227 | |||
| 228 | if (!comm_l || !comm_r) | ||
| 229 | return cmp_null(comm_l, comm_r); | ||
| 230 | |||
| 231 | return strcmp(comm_l, comm_r); | ||
| 232 | } | ||
| 233 | 191 | ||
| 234 | /* --sort parent */ | 192 | /* --sort parent */ |
| 235 | 193 | ||
| 236 | int64_t | 194 | static int64_t |
| 237 | sort__parent_cmp(struct hist_entry *left, struct hist_entry *right) | 195 | sort__parent_cmp(struct hist_entry *left, struct hist_entry *right) |
| 238 | { | 196 | { |
| 239 | struct symbol *sym_l = left->parent; | 197 | struct symbol *sym_l = left->parent; |
| @@ -252,9 +210,16 @@ static int hist_entry__parent_snprintf(struct hist_entry *self, char *bf, | |||
| 252 | self->parent ? self->parent->name : "[other]"); | 210 | self->parent ? self->parent->name : "[other]"); |
| 253 | } | 211 | } |
| 254 | 212 | ||
| 213 | struct sort_entry sort_parent = { | ||
| 214 | .se_header = "Parent symbol", | ||
| 215 | .se_cmp = sort__parent_cmp, | ||
| 216 | .se_snprintf = hist_entry__parent_snprintf, | ||
| 217 | .se_width_idx = HISTC_PARENT, | ||
| 218 | }; | ||
| 219 | |||
| 255 | /* --sort cpu */ | 220 | /* --sort cpu */ |
| 256 | 221 | ||
| 257 | int64_t | 222 | static int64_t |
| 258 | sort__cpu_cmp(struct hist_entry *left, struct hist_entry *right) | 223 | sort__cpu_cmp(struct hist_entry *left, struct hist_entry *right) |
| 259 | { | 224 | { |
| 260 | return right->cpu - left->cpu; | 225 | return right->cpu - left->cpu; |
| @@ -266,6 +231,28 @@ static int hist_entry__cpu_snprintf(struct hist_entry *self, char *bf, | |||
| 266 | return repsep_snprintf(bf, size, "%-*d", width, self->cpu); | 231 | return repsep_snprintf(bf, size, "%-*d", width, self->cpu); |
| 267 | } | 232 | } |
| 268 | 233 | ||
| 234 | struct sort_entry sort_cpu = { | ||
| 235 | .se_header = "CPU", | ||
| 236 | .se_cmp = sort__cpu_cmp, | ||
| 237 | .se_snprintf = hist_entry__cpu_snprintf, | ||
| 238 | .se_width_idx = HISTC_CPU, | ||
| 239 | }; | ||
| 240 | |||
| 241 | struct sort_dimension { | ||
| 242 | const char *name; | ||
| 243 | struct sort_entry *entry; | ||
| 244 | int taken; | ||
| 245 | }; | ||
| 246 | |||
| 247 | static struct sort_dimension sort_dimensions[] = { | ||
| 248 | { .name = "pid", .entry = &sort_thread, }, | ||
| 249 | { .name = "comm", .entry = &sort_comm, }, | ||
| 250 | { .name = "dso", .entry = &sort_dso, }, | ||
| 251 | { .name = "symbol", .entry = &sort_sym, }, | ||
| 252 | { .name = "parent", .entry = &sort_parent, }, | ||
| 253 | { .name = "cpu", .entry = &sort_cpu, }, | ||
| 254 | }; | ||
| 255 | |||
| 269 | int sort_dimension__add(const char *tok) | 256 | int sort_dimension__add(const char *tok) |
| 270 | { | 257 | { |
| 271 | unsigned int i; | 258 | unsigned int i; |
| @@ -273,15 +260,9 @@ int sort_dimension__add(const char *tok) | |||
| 273 | for (i = 0; i < ARRAY_SIZE(sort_dimensions); i++) { | 260 | for (i = 0; i < ARRAY_SIZE(sort_dimensions); i++) { |
| 274 | struct sort_dimension *sd = &sort_dimensions[i]; | 261 | struct sort_dimension *sd = &sort_dimensions[i]; |
| 275 | 262 | ||
| 276 | if (sd->taken) | ||
| 277 | continue; | ||
| 278 | |||
| 279 | if (strncasecmp(tok, sd->name, strlen(tok))) | 263 | if (strncasecmp(tok, sd->name, strlen(tok))) |
| 280 | continue; | 264 | continue; |
| 281 | 265 | ||
| 282 | if (sd->entry->se_collapse) | ||
| 283 | sort__need_collapse = 1; | ||
| 284 | |||
| 285 | if (sd->entry == &sort_parent) { | 266 | if (sd->entry == &sort_parent) { |
| 286 | int ret = regcomp(&parent_regex, parent_pattern, REG_EXTENDED); | 267 | int ret = regcomp(&parent_regex, parent_pattern, REG_EXTENDED); |
| 287 | if (ret) { | 268 | if (ret) { |
| @@ -294,6 +275,12 @@ int sort_dimension__add(const char *tok) | |||
| 294 | sort__has_parent = 1; | 275 | sort__has_parent = 1; |
| 295 | } | 276 | } |
| 296 | 277 | ||
| 278 | if (sd->taken) | ||
| 279 | return 0; | ||
| 280 | |||
| 281 | if (sd->entry->se_collapse) | ||
| 282 | sort__need_collapse = 1; | ||
| 283 | |||
| 297 | if (list_empty(&hist_entry__sort_list)) { | 284 | if (list_empty(&hist_entry__sort_list)) { |
| 298 | if (!strcmp(sd->name, "pid")) | 285 | if (!strcmp(sd->name, "pid")) |
| 299 | sort__first_dimension = SORT_PID; | 286 | sort__first_dimension = SORT_PID; |
diff --git a/tools/perf/util/sort.h b/tools/perf/util/sort.h index 0b91053a7d1..77d0388ad41 100644 --- a/tools/perf/util/sort.h +++ b/tools/perf/util/sort.h | |||
| @@ -103,20 +103,6 @@ extern struct sort_entry sort_thread; | |||
| 103 | extern struct list_head hist_entry__sort_list; | 103 | extern struct list_head hist_entry__sort_list; |
| 104 | 104 | ||
| 105 | void setup_sorting(const char * const usagestr[], const struct option *opts); | 105 | void setup_sorting(const char * const usagestr[], const struct option *opts); |
| 106 | |||
| 107 | extern size_t sort__thread_print(FILE *, struct hist_entry *, unsigned int); | ||
| 108 | extern size_t sort__comm_print(FILE *, struct hist_entry *, unsigned int); | ||
| 109 | extern size_t sort__dso_print(FILE *, struct hist_entry *, unsigned int); | ||
| 110 | extern size_t sort__sym_print(FILE *, struct hist_entry *, unsigned int __used); | ||
| 111 | extern int64_t cmp_null(void *, void *); | ||
| 112 | extern int64_t sort__thread_cmp(struct hist_entry *, struct hist_entry *); | ||
| 113 | extern int64_t sort__comm_cmp(struct hist_entry *, struct hist_entry *); | ||
| 114 | extern int64_t sort__comm_collapse(struct hist_entry *, struct hist_entry *); | ||
| 115 | extern int64_t sort__dso_cmp(struct hist_entry *, struct hist_entry *); | ||
| 116 | extern int64_t sort__sym_cmp(struct hist_entry *, struct hist_entry *); | ||
| 117 | extern int64_t sort__parent_cmp(struct hist_entry *, struct hist_entry *); | ||
| 118 | int64_t sort__cpu_cmp(struct hist_entry *left, struct hist_entry *right); | ||
| 119 | extern size_t sort__parent_print(FILE *, struct hist_entry *, unsigned int); | ||
| 120 | extern int sort_dimension__add(const char *); | 106 | extern int sort_dimension__add(const char *); |
| 121 | void sort_entry__setup_elide(struct sort_entry *self, struct strlist *list, | 107 | void sort_entry__setup_elide(struct sort_entry *self, struct strlist *list, |
| 122 | const char *list_name, FILE *fp); | 108 | const char *list_name, FILE *fp); |
diff --git a/tools/perf/util/string.c b/tools/perf/util/string.c index b9a985dadd0..d5836382ff2 100644 --- a/tools/perf/util/string.c +++ b/tools/perf/util/string.c | |||
| @@ -294,3 +294,22 @@ bool strlazymatch(const char *str, const char *pat) | |||
| 294 | { | 294 | { |
| 295 | return __match_glob(str, pat, true); | 295 | return __match_glob(str, pat, true); |
| 296 | } | 296 | } |
| 297 | |||
| 298 | /** | ||
| 299 | * strtailcmp - Compare the tail of two strings | ||
| 300 | * @s1: 1st string to be compared | ||
| 301 | * @s2: 2nd string to be compared | ||
| 302 | * | ||
| 303 | * Return 0 if whole of either string is same as another's tail part. | ||
| 304 | */ | ||
| 305 | int strtailcmp(const char *s1, const char *s2) | ||
| 306 | { | ||
| 307 | int i1 = strlen(s1); | ||
| 308 | int i2 = strlen(s2); | ||
| 309 | while (--i1 >= 0 && --i2 >= 0) { | ||
| 310 | if (s1[i1] != s2[i2]) | ||
| 311 | return s1[i1] - s2[i2]; | ||
| 312 | } | ||
| 313 | return 0; | ||
| 314 | } | ||
| 315 | |||
diff --git a/tools/perf/util/trace-event-info.c b/tools/perf/util/trace-event-info.c index 35729f4c40c..3403f814ad7 100644 --- a/tools/perf/util/trace-event-info.c +++ b/tools/perf/util/trace-event-info.c | |||
| @@ -183,106 +183,59 @@ int bigendian(void) | |||
| 183 | return *ptr == 0x01020304; | 183 | return *ptr == 0x01020304; |
| 184 | } | 184 | } |
| 185 | 185 | ||
| 186 | static unsigned long long copy_file_fd(int fd) | 186 | /* unfortunately, you can not stat debugfs or proc files for size */ |
| 187 | static void record_file(const char *file, size_t hdr_sz) | ||
| 187 | { | 188 | { |
| 188 | unsigned long long size = 0; | 189 | unsigned long long size = 0; |
| 189 | char buf[BUFSIZ]; | 190 | char buf[BUFSIZ], *sizep; |
| 190 | int r; | 191 | off_t hdr_pos = lseek(output_fd, 0, SEEK_CUR); |
| 191 | 192 | int r, fd; | |
| 192 | do { | ||
| 193 | r = read(fd, buf, BUFSIZ); | ||
| 194 | if (r > 0) { | ||
| 195 | size += r; | ||
| 196 | write_or_die(buf, r); | ||
| 197 | } | ||
| 198 | } while (r > 0); | ||
| 199 | |||
| 200 | return size; | ||
| 201 | } | ||
| 202 | |||
| 203 | static unsigned long long copy_file(const char *file) | ||
| 204 | { | ||
| 205 | unsigned long long size = 0; | ||
| 206 | int fd; | ||
| 207 | 193 | ||
| 208 | fd = open(file, O_RDONLY); | 194 | fd = open(file, O_RDONLY); |
| 209 | if (fd < 0) | 195 | if (fd < 0) |
| 210 | die("Can't read '%s'", file); | 196 | die("Can't read '%s'", file); |
| 211 | size = copy_file_fd(fd); | ||
| 212 | close(fd); | ||
| 213 | 197 | ||
| 214 | return size; | 198 | /* put in zeros for file size, then fill true size later */ |
| 215 | } | 199 | write_or_die(&size, hdr_sz); |
| 216 | |||
| 217 | static unsigned long get_size_fd(int fd) | ||
| 218 | { | ||
| 219 | unsigned long long size = 0; | ||
| 220 | char buf[BUFSIZ]; | ||
| 221 | int r; | ||
| 222 | 200 | ||
| 223 | do { | 201 | do { |
| 224 | r = read(fd, buf, BUFSIZ); | 202 | r = read(fd, buf, BUFSIZ); |
| 225 | if (r > 0) | 203 | if (r > 0) { |
| 226 | size += r; | 204 | size += r; |
| 205 | write_or_die(buf, r); | ||
| 206 | } | ||
| 227 | } while (r > 0); | 207 | } while (r > 0); |
| 228 | |||
| 229 | lseek(fd, 0, SEEK_SET); | ||
| 230 | |||
| 231 | return size; | ||
| 232 | } | ||
| 233 | |||
| 234 | static unsigned long get_size(const char *file) | ||
| 235 | { | ||
| 236 | unsigned long long size = 0; | ||
| 237 | int fd; | ||
| 238 | |||
| 239 | fd = open(file, O_RDONLY); | ||
| 240 | if (fd < 0) | ||
| 241 | die("Can't read '%s'", file); | ||
| 242 | size = get_size_fd(fd); | ||
| 243 | close(fd); | 208 | close(fd); |
| 244 | 209 | ||
| 245 | return size; | 210 | /* ugh, handle big-endian hdr_size == 4 */ |
| 211 | sizep = (char*)&size; | ||
| 212 | if (bigendian()) | ||
| 213 | sizep += sizeof(u64) - hdr_sz; | ||
| 214 | |||
| 215 | if (pwrite(output_fd, sizep, hdr_sz, hdr_pos) < 0) | ||
| 216 | die("writing to %s", output_file); | ||
| 246 | } | 217 | } |
| 247 | 218 | ||
| 248 | static void read_header_files(void) | 219 | static void read_header_files(void) |
| 249 | { | 220 | { |
| 250 | unsigned long long size, check_size; | ||
| 251 | char *path; | 221 | char *path; |
| 252 | int fd; | 222 | struct stat st; |
| 253 | 223 | ||
| 254 | path = get_tracing_file("events/header_page"); | 224 | path = get_tracing_file("events/header_page"); |
| 255 | fd = open(path, O_RDONLY); | 225 | if (stat(path, &st) < 0) |
| 256 | if (fd < 0) | ||
| 257 | die("can't read '%s'", path); | 226 | die("can't read '%s'", path); |
| 258 | 227 | ||
| 259 | /* unfortunately, you can not stat debugfs files for size */ | ||
| 260 | size = get_size_fd(fd); | ||
| 261 | |||
| 262 | write_or_die("header_page", 12); | 228 | write_or_die("header_page", 12); |
| 263 | write_or_die(&size, 8); | 229 | record_file(path, 8); |
| 264 | check_size = copy_file_fd(fd); | ||
| 265 | close(fd); | ||
| 266 | |||
| 267 | if (size != check_size) | ||
| 268 | die("wrong size for '%s' size=%lld read=%lld", | ||
| 269 | path, size, check_size); | ||
| 270 | put_tracing_file(path); | 230 | put_tracing_file(path); |
| 271 | 231 | ||
| 272 | path = get_tracing_file("events/header_event"); | 232 | path = get_tracing_file("events/header_event"); |
| 273 | fd = open(path, O_RDONLY); | 233 | if (stat(path, &st) < 0) |
| 274 | if (fd < 0) | ||
| 275 | die("can't read '%s'", path); | 234 | die("can't read '%s'", path); |
| 276 | 235 | ||
| 277 | size = get_size_fd(fd); | ||
| 278 | |||
| 279 | write_or_die("header_event", 13); | 236 | write_or_die("header_event", 13); |
| 280 | write_or_die(&size, 8); | 237 | record_file(path, 8); |
| 281 | check_size = copy_file_fd(fd); | ||
| 282 | if (size != check_size) | ||
| 283 | die("wrong size for '%s'", path); | ||
| 284 | put_tracing_file(path); | 238 | put_tracing_file(path); |
| 285 | close(fd); | ||
| 286 | } | 239 | } |
| 287 | 240 | ||
| 288 | static bool name_in_tp_list(char *sys, struct tracepoint_path *tps) | 241 | static bool name_in_tp_list(char *sys, struct tracepoint_path *tps) |
| @@ -298,7 +251,6 @@ static bool name_in_tp_list(char *sys, struct tracepoint_path *tps) | |||
| 298 | 251 | ||
| 299 | static void copy_event_system(const char *sys, struct tracepoint_path *tps) | 252 | static void copy_event_system(const char *sys, struct tracepoint_path *tps) |
| 300 | { | 253 | { |
| 301 | unsigned long long size, check_size; | ||
| 302 | struct dirent *dent; | 254 | struct dirent *dent; |
| 303 | struct stat st; | 255 | struct stat st; |
| 304 | char *format; | 256 | char *format; |
| @@ -338,14 +290,8 @@ static void copy_event_system(const char *sys, struct tracepoint_path *tps) | |||
| 338 | sprintf(format, "%s/%s/format", sys, dent->d_name); | 290 | sprintf(format, "%s/%s/format", sys, dent->d_name); |
| 339 | ret = stat(format, &st); | 291 | ret = stat(format, &st); |
| 340 | 292 | ||
| 341 | if (ret >= 0) { | 293 | if (ret >= 0) |
| 342 | /* unfortunately, you can not stat debugfs files for size */ | 294 | record_file(format, 8); |
| 343 | size = get_size(format); | ||
| 344 | write_or_die(&size, 8); | ||
| 345 | check_size = copy_file(format); | ||
| 346 | if (size != check_size) | ||
| 347 | die("error in size of file '%s'", format); | ||
| 348 | } | ||
| 349 | 295 | ||
| 350 | free(format); | 296 | free(format); |
| 351 | } | 297 | } |
| @@ -426,7 +372,7 @@ static void read_event_files(struct tracepoint_path *tps) | |||
| 426 | 372 | ||
| 427 | static void read_proc_kallsyms(void) | 373 | static void read_proc_kallsyms(void) |
| 428 | { | 374 | { |
| 429 | unsigned int size, check_size; | 375 | unsigned int size; |
| 430 | const char *path = "/proc/kallsyms"; | 376 | const char *path = "/proc/kallsyms"; |
| 431 | struct stat st; | 377 | struct stat st; |
| 432 | int ret; | 378 | int ret; |
| @@ -438,17 +384,12 @@ static void read_proc_kallsyms(void) | |||
| 438 | write_or_die(&size, 4); | 384 | write_or_die(&size, 4); |
| 439 | return; | 385 | return; |
| 440 | } | 386 | } |
| 441 | size = get_size(path); | 387 | record_file(path, 4); |
| 442 | write_or_die(&size, 4); | ||
| 443 | check_size = copy_file(path); | ||
| 444 | if (size != check_size) | ||
| 445 | die("error in size of file '%s'", path); | ||
| 446 | |||
| 447 | } | 388 | } |
| 448 | 389 | ||
| 449 | static void read_ftrace_printk(void) | 390 | static void read_ftrace_printk(void) |
| 450 | { | 391 | { |
| 451 | unsigned int size, check_size; | 392 | unsigned int size; |
| 452 | char *path; | 393 | char *path; |
| 453 | struct stat st; | 394 | struct stat st; |
| 454 | int ret; | 395 | int ret; |
| @@ -461,11 +402,8 @@ static void read_ftrace_printk(void) | |||
| 461 | write_or_die(&size, 4); | 402 | write_or_die(&size, 4); |
| 462 | goto out; | 403 | goto out; |
| 463 | } | 404 | } |
| 464 | size = get_size(path); | 405 | record_file(path, 4); |
| 465 | write_or_die(&size, 4); | 406 | |
| 466 | check_size = copy_file(path); | ||
| 467 | if (size != check_size) | ||
| 468 | die("error in size of file '%s'", path); | ||
| 469 | out: | 407 | out: |
| 470 | put_tracing_file(path); | 408 | put_tracing_file(path); |
| 471 | } | 409 | } |
diff --git a/tools/perf/util/util.h b/tools/perf/util/util.h index fc784284ac8..0128906bac8 100644 --- a/tools/perf/util/util.h +++ b/tools/perf/util/util.h | |||
| @@ -238,6 +238,7 @@ char **argv_split(const char *str, int *argcp); | |||
| 238 | void argv_free(char **argv); | 238 | void argv_free(char **argv); |
| 239 | bool strglobmatch(const char *str, const char *pat); | 239 | bool strglobmatch(const char *str, const char *pat); |
| 240 | bool strlazymatch(const char *str, const char *pat); | 240 | bool strlazymatch(const char *str, const char *pat); |
| 241 | int strtailcmp(const char *s1, const char *s2); | ||
| 241 | unsigned long convert_unit(unsigned long value, char *unit); | 242 | unsigned long convert_unit(unsigned long value, char *unit); |
| 242 | int readn(int fd, void *buf, size_t size); | 243 | int readn(int fd, void *buf, size_t size); |
| 243 | 244 | ||
