diff options
Diffstat (limited to 'kernel/trace/trace_hw_branches.c')
| -rw-r--r-- | kernel/trace/trace_hw_branches.c | 185 |
1 files changed, 146 insertions, 39 deletions
diff --git a/kernel/trace/trace_hw_branches.c b/kernel/trace/trace_hw_branches.c index 649df22d435f..7bfdf4c2347f 100644 --- a/kernel/trace/trace_hw_branches.c +++ b/kernel/trace/trace_hw_branches.c | |||
| @@ -1,30 +1,53 @@ | |||
| 1 | /* | 1 | /* |
| 2 | * h/w branch tracer for x86 based on bts | 2 | * h/w branch tracer for x86 based on bts |
| 3 | * | 3 | * |
| 4 | * Copyright (C) 2008 Markus Metzger <markus.t.metzger@gmail.com> | 4 | * Copyright (C) 2008-2009 Intel Corporation. |
| 5 | * | 5 | * Markus Metzger <markus.t.metzger@gmail.com>, 2008-2009 |
| 6 | */ | 6 | */ |
| 7 | 7 | #include <linux/spinlock.h> | |
| 8 | #include <linux/module.h> | 8 | #include <linux/kallsyms.h> |
| 9 | #include <linux/fs.h> | ||
| 10 | #include <linux/debugfs.h> | 9 | #include <linux/debugfs.h> |
| 11 | #include <linux/ftrace.h> | 10 | #include <linux/ftrace.h> |
| 12 | #include <linux/kallsyms.h> | 11 | #include <linux/module.h> |
| 12 | #include <linux/cpu.h> | ||
| 13 | #include <linux/smp.h> | ||
| 14 | #include <linux/fs.h> | ||
| 13 | 15 | ||
| 14 | #include <asm/ds.h> | 16 | #include <asm/ds.h> |
| 15 | 17 | ||
| 16 | #include "trace.h" | 18 | #include "trace.h" |
| 19 | #include "trace_output.h" | ||
| 17 | 20 | ||
| 18 | 21 | ||
| 19 | #define SIZEOF_BTS (1 << 13) | 22 | #define SIZEOF_BTS (1 << 13) |
| 20 | 23 | ||
| 24 | /* | ||
| 25 | * The tracer lock protects the below per-cpu tracer array. | ||
| 26 | * It needs to be held to: | ||
| 27 | * - start tracing on all cpus | ||
| 28 | * - stop tracing on all cpus | ||
| 29 | * - start tracing on a single hotplug cpu | ||
| 30 | * - stop tracing on a single hotplug cpu | ||
| 31 | * - read the trace from all cpus | ||
| 32 | * - read the trace from a single cpu | ||
| 33 | */ | ||
| 34 | static DEFINE_SPINLOCK(bts_tracer_lock); | ||
| 21 | static DEFINE_PER_CPU(struct bts_tracer *, tracer); | 35 | static DEFINE_PER_CPU(struct bts_tracer *, tracer); |
| 22 | static DEFINE_PER_CPU(unsigned char[SIZEOF_BTS], buffer); | 36 | static DEFINE_PER_CPU(unsigned char[SIZEOF_BTS], buffer); |
| 23 | 37 | ||
| 24 | #define this_tracer per_cpu(tracer, smp_processor_id()) | 38 | #define this_tracer per_cpu(tracer, smp_processor_id()) |
| 25 | #define this_buffer per_cpu(buffer, smp_processor_id()) | 39 | #define this_buffer per_cpu(buffer, smp_processor_id()) |
| 26 | 40 | ||
| 41 | static int __read_mostly trace_hw_branches_enabled; | ||
| 42 | static struct trace_array *hw_branch_trace __read_mostly; | ||
| 43 | |||
| 27 | 44 | ||
| 45 | /* | ||
| 46 | * Start tracing on the current cpu. | ||
| 47 | * The argument is ignored. | ||
| 48 | * | ||
| 49 | * pre: bts_tracer_lock must be locked. | ||
| 50 | */ | ||
| 28 | static void bts_trace_start_cpu(void *arg) | 51 | static void bts_trace_start_cpu(void *arg) |
| 29 | { | 52 | { |
| 30 | if (this_tracer) | 53 | if (this_tracer) |
| @@ -42,14 +65,20 @@ static void bts_trace_start_cpu(void *arg) | |||
| 42 | 65 | ||
| 43 | static void bts_trace_start(struct trace_array *tr) | 66 | static void bts_trace_start(struct trace_array *tr) |
| 44 | { | 67 | { |
| 45 | int cpu; | 68 | spin_lock(&bts_tracer_lock); |
| 46 | 69 | ||
| 47 | tracing_reset_online_cpus(tr); | 70 | on_each_cpu(bts_trace_start_cpu, NULL, 1); |
| 71 | trace_hw_branches_enabled = 1; | ||
| 48 | 72 | ||
| 49 | for_each_cpu(cpu, cpu_possible_mask) | 73 | spin_unlock(&bts_tracer_lock); |
| 50 | smp_call_function_single(cpu, bts_trace_start_cpu, NULL, 1); | ||
| 51 | } | 74 | } |
| 52 | 75 | ||
| 76 | /* | ||
| 77 | * Stop tracing on the current cpu. | ||
| 78 | * The argument is ignored. | ||
| 79 | * | ||
| 80 | * pre: bts_tracer_lock must be locked. | ||
| 81 | */ | ||
| 53 | static void bts_trace_stop_cpu(void *arg) | 82 | static void bts_trace_stop_cpu(void *arg) |
| 54 | { | 83 | { |
| 55 | if (this_tracer) { | 84 | if (this_tracer) { |
| @@ -60,26 +89,60 @@ static void bts_trace_stop_cpu(void *arg) | |||
| 60 | 89 | ||
| 61 | static void bts_trace_stop(struct trace_array *tr) | 90 | static void bts_trace_stop(struct trace_array *tr) |
| 62 | { | 91 | { |
| 63 | int cpu; | 92 | spin_lock(&bts_tracer_lock); |
| 93 | |||
| 94 | trace_hw_branches_enabled = 0; | ||
| 95 | on_each_cpu(bts_trace_stop_cpu, NULL, 1); | ||
| 96 | |||
| 97 | spin_unlock(&bts_tracer_lock); | ||
| 98 | } | ||
| 99 | |||
| 100 | static int __cpuinit bts_hotcpu_handler(struct notifier_block *nfb, | ||
| 101 | unsigned long action, void *hcpu) | ||
| 102 | { | ||
| 103 | unsigned int cpu = (unsigned long)hcpu; | ||
| 64 | 104 | ||
| 65 | for_each_cpu(cpu, cpu_possible_mask) | 105 | spin_lock(&bts_tracer_lock); |
| 106 | |||
| 107 | if (!trace_hw_branches_enabled) | ||
| 108 | goto out; | ||
| 109 | |||
| 110 | switch (action) { | ||
| 111 | case CPU_ONLINE: | ||
| 112 | case CPU_DOWN_FAILED: | ||
| 113 | smp_call_function_single(cpu, bts_trace_start_cpu, NULL, 1); | ||
| 114 | break; | ||
| 115 | case CPU_DOWN_PREPARE: | ||
| 66 | smp_call_function_single(cpu, bts_trace_stop_cpu, NULL, 1); | 116 | smp_call_function_single(cpu, bts_trace_stop_cpu, NULL, 1); |
| 117 | break; | ||
| 118 | } | ||
| 119 | |||
| 120 | out: | ||
| 121 | spin_unlock(&bts_tracer_lock); | ||
| 122 | return NOTIFY_DONE; | ||
| 67 | } | 123 | } |
| 68 | 124 | ||
| 125 | static struct notifier_block bts_hotcpu_notifier __cpuinitdata = { | ||
| 126 | .notifier_call = bts_hotcpu_handler | ||
| 127 | }; | ||
| 128 | |||
| 69 | static int bts_trace_init(struct trace_array *tr) | 129 | static int bts_trace_init(struct trace_array *tr) |
| 70 | { | 130 | { |
| 71 | tracing_reset_online_cpus(tr); | 131 | hw_branch_trace = tr; |
| 132 | |||
| 72 | bts_trace_start(tr); | 133 | bts_trace_start(tr); |
| 73 | 134 | ||
| 74 | return 0; | 135 | return 0; |
| 75 | } | 136 | } |
| 76 | 137 | ||
| 138 | static void bts_trace_reset(struct trace_array *tr) | ||
| 139 | { | ||
| 140 | bts_trace_stop(tr); | ||
| 141 | } | ||
| 142 | |||
| 77 | static void bts_trace_print_header(struct seq_file *m) | 143 | static void bts_trace_print_header(struct seq_file *m) |
| 78 | { | 144 | { |
| 79 | seq_puts(m, | 145 | seq_puts(m, "# CPU# TO <- FROM\n"); |
| 80 | "# CPU# FROM TO FUNCTION\n"); | ||
| 81 | seq_puts(m, | ||
| 82 | "# | | | |\n"); | ||
| 83 | } | 146 | } |
| 84 | 147 | ||
| 85 | static enum print_line_t bts_trace_print_line(struct trace_iterator *iter) | 148 | static enum print_line_t bts_trace_print_line(struct trace_iterator *iter) |
| @@ -87,15 +150,15 @@ static enum print_line_t bts_trace_print_line(struct trace_iterator *iter) | |||
| 87 | struct trace_entry *entry = iter->ent; | 150 | struct trace_entry *entry = iter->ent; |
| 88 | struct trace_seq *seq = &iter->seq; | 151 | struct trace_seq *seq = &iter->seq; |
| 89 | struct hw_branch_entry *it; | 152 | struct hw_branch_entry *it; |
| 153 | unsigned long symflags = TRACE_ITER_SYM_OFFSET; | ||
| 90 | 154 | ||
| 91 | trace_assign_type(it, entry); | 155 | trace_assign_type(it, entry); |
| 92 | 156 | ||
| 93 | if (entry->type == TRACE_HW_BRANCHES) { | 157 | if (entry->type == TRACE_HW_BRANCHES) { |
| 94 | if (trace_seq_printf(seq, "%4d ", entry->cpu) && | 158 | if (trace_seq_printf(seq, "%4d ", iter->cpu) && |
| 95 | trace_seq_printf(seq, "0x%016llx -> 0x%016llx ", | 159 | seq_print_ip_sym(seq, it->to, symflags) && |
| 96 | it->from, it->to) && | 160 | trace_seq_printf(seq, "\t <- ") && |
| 97 | (!it->from || | 161 | seq_print_ip_sym(seq, it->from, symflags) && |
| 98 | seq_print_ip_sym(seq, it->from, /* sym_flags = */ 0)) && | ||
| 99 | trace_seq_printf(seq, "\n")) | 162 | trace_seq_printf(seq, "\n")) |
| 100 | return TRACE_TYPE_HANDLED; | 163 | return TRACE_TYPE_HANDLED; |
| 101 | return TRACE_TYPE_PARTIAL_LINE;; | 164 | return TRACE_TYPE_PARTIAL_LINE;; |
| @@ -103,26 +166,42 @@ static enum print_line_t bts_trace_print_line(struct trace_iterator *iter) | |||
| 103 | return TRACE_TYPE_UNHANDLED; | 166 | return TRACE_TYPE_UNHANDLED; |
| 104 | } | 167 | } |
| 105 | 168 | ||
| 106 | void trace_hw_branch(struct trace_array *tr, u64 from, u64 to) | 169 | void trace_hw_branch(u64 from, u64 to) |
| 107 | { | 170 | { |
| 171 | struct trace_array *tr = hw_branch_trace; | ||
| 108 | struct ring_buffer_event *event; | 172 | struct ring_buffer_event *event; |
| 109 | struct hw_branch_entry *entry; | 173 | struct hw_branch_entry *entry; |
| 110 | unsigned long irq; | 174 | unsigned long irq1; |
| 175 | int cpu; | ||
| 111 | 176 | ||
| 112 | event = ring_buffer_lock_reserve(tr->buffer, sizeof(*entry), &irq); | 177 | if (unlikely(!tr)) |
| 113 | if (!event) | ||
| 114 | return; | 178 | return; |
| 179 | |||
| 180 | if (unlikely(!trace_hw_branches_enabled)) | ||
| 181 | return; | ||
| 182 | |||
| 183 | local_irq_save(irq1); | ||
| 184 | cpu = raw_smp_processor_id(); | ||
| 185 | if (atomic_inc_return(&tr->data[cpu]->disabled) != 1) | ||
| 186 | goto out; | ||
| 187 | |||
| 188 | event = trace_buffer_lock_reserve(tr, TRACE_HW_BRANCHES, | ||
| 189 | sizeof(*entry), 0, 0); | ||
| 190 | if (!event) | ||
| 191 | goto out; | ||
| 115 | entry = ring_buffer_event_data(event); | 192 | entry = ring_buffer_event_data(event); |
| 116 | tracing_generic_entry_update(&entry->ent, 0, from); | 193 | tracing_generic_entry_update(&entry->ent, 0, from); |
| 117 | entry->ent.type = TRACE_HW_BRANCHES; | 194 | entry->ent.type = TRACE_HW_BRANCHES; |
| 118 | entry->ent.cpu = smp_processor_id(); | ||
| 119 | entry->from = from; | 195 | entry->from = from; |
| 120 | entry->to = to; | 196 | entry->to = to; |
| 121 | ring_buffer_unlock_commit(tr->buffer, event, irq); | 197 | trace_buffer_unlock_commit(tr, event, 0, 0); |
| 198 | |||
| 199 | out: | ||
| 200 | atomic_dec(&tr->data[cpu]->disabled); | ||
| 201 | local_irq_restore(irq1); | ||
| 122 | } | 202 | } |
| 123 | 203 | ||
| 124 | static void trace_bts_at(struct trace_array *tr, | 204 | static void trace_bts_at(const struct bts_trace *trace, void *at) |
| 125 | const struct bts_trace *trace, void *at) | ||
| 126 | { | 205 | { |
| 127 | struct bts_struct bts; | 206 | struct bts_struct bts; |
| 128 | int err = 0; | 207 | int err = 0; |
| @@ -137,18 +216,29 @@ static void trace_bts_at(struct trace_array *tr, | |||
| 137 | 216 | ||
| 138 | switch (bts.qualifier) { | 217 | switch (bts.qualifier) { |
| 139 | case BTS_BRANCH: | 218 | case BTS_BRANCH: |
| 140 | trace_hw_branch(tr, bts.variant.lbr.from, bts.variant.lbr.to); | 219 | trace_hw_branch(bts.variant.lbr.from, bts.variant.lbr.to); |
| 141 | break; | 220 | break; |
| 142 | } | 221 | } |
| 143 | } | 222 | } |
| 144 | 223 | ||
| 224 | /* | ||
| 225 | * Collect the trace on the current cpu and write it into the ftrace buffer. | ||
| 226 | * | ||
| 227 | * pre: bts_tracer_lock must be locked | ||
| 228 | */ | ||
| 145 | static void trace_bts_cpu(void *arg) | 229 | static void trace_bts_cpu(void *arg) |
| 146 | { | 230 | { |
| 147 | struct trace_array *tr = (struct trace_array *) arg; | 231 | struct trace_array *tr = (struct trace_array *) arg; |
| 148 | const struct bts_trace *trace; | 232 | const struct bts_trace *trace; |
| 149 | unsigned char *at; | 233 | unsigned char *at; |
| 150 | 234 | ||
| 151 | if (!this_tracer) | 235 | if (unlikely(!tr)) |
| 236 | return; | ||
| 237 | |||
| 238 | if (unlikely(atomic_read(&tr->data[raw_smp_processor_id()]->disabled))) | ||
| 239 | return; | ||
| 240 | |||
| 241 | if (unlikely(!this_tracer)) | ||
| 152 | return; | 242 | return; |
| 153 | 243 | ||
| 154 | ds_suspend_bts(this_tracer); | 244 | ds_suspend_bts(this_tracer); |
| @@ -158,11 +248,11 @@ static void trace_bts_cpu(void *arg) | |||
| 158 | 248 | ||
| 159 | for (at = trace->ds.top; (void *)at < trace->ds.end; | 249 | for (at = trace->ds.top; (void *)at < trace->ds.end; |
| 160 | at += trace->ds.size) | 250 | at += trace->ds.size) |
| 161 | trace_bts_at(tr, trace, at); | 251 | trace_bts_at(trace, at); |
| 162 | 252 | ||
| 163 | for (at = trace->ds.begin; (void *)at < trace->ds.top; | 253 | for (at = trace->ds.begin; (void *)at < trace->ds.top; |
| 164 | at += trace->ds.size) | 254 | at += trace->ds.size) |
| 165 | trace_bts_at(tr, trace, at); | 255 | trace_bts_at(trace, at); |
| 166 | 256 | ||
| 167 | out: | 257 | out: |
| 168 | ds_resume_bts(this_tracer); | 258 | ds_resume_bts(this_tracer); |
| @@ -170,26 +260,43 @@ out: | |||
| 170 | 260 | ||
| 171 | static void trace_bts_prepare(struct trace_iterator *iter) | 261 | static void trace_bts_prepare(struct trace_iterator *iter) |
| 172 | { | 262 | { |
| 173 | int cpu; | 263 | spin_lock(&bts_tracer_lock); |
| 264 | |||
| 265 | on_each_cpu(trace_bts_cpu, iter->tr, 1); | ||
| 266 | |||
| 267 | spin_unlock(&bts_tracer_lock); | ||
| 268 | } | ||
| 269 | |||
| 270 | static void trace_bts_close(struct trace_iterator *iter) | ||
| 271 | { | ||
| 272 | tracing_reset_online_cpus(iter->tr); | ||
| 273 | } | ||
| 274 | |||
| 275 | void trace_hw_branch_oops(void) | ||
| 276 | { | ||
| 277 | spin_lock(&bts_tracer_lock); | ||
| 278 | |||
| 279 | trace_bts_cpu(hw_branch_trace); | ||
| 174 | 280 | ||
| 175 | for_each_cpu(cpu, cpu_possible_mask) | 281 | spin_unlock(&bts_tracer_lock); |
| 176 | smp_call_function_single(cpu, trace_bts_cpu, iter->tr, 1); | ||
| 177 | } | 282 | } |
| 178 | 283 | ||
| 179 | struct tracer bts_tracer __read_mostly = | 284 | struct tracer bts_tracer __read_mostly = |
| 180 | { | 285 | { |
| 181 | .name = "hw-branch-tracer", | 286 | .name = "hw-branch-tracer", |
| 182 | .init = bts_trace_init, | 287 | .init = bts_trace_init, |
| 183 | .reset = bts_trace_stop, | 288 | .reset = bts_trace_reset, |
| 184 | .print_header = bts_trace_print_header, | 289 | .print_header = bts_trace_print_header, |
| 185 | .print_line = bts_trace_print_line, | 290 | .print_line = bts_trace_print_line, |
| 186 | .start = bts_trace_start, | 291 | .start = bts_trace_start, |
| 187 | .stop = bts_trace_stop, | 292 | .stop = bts_trace_stop, |
| 188 | .open = trace_bts_prepare | 293 | .open = trace_bts_prepare, |
| 294 | .close = trace_bts_close | ||
| 189 | }; | 295 | }; |
| 190 | 296 | ||
| 191 | __init static int init_bts_trace(void) | 297 | __init static int init_bts_trace(void) |
| 192 | { | 298 | { |
| 299 | register_hotcpu_notifier(&bts_hotcpu_notifier); | ||
| 193 | return register_tracer(&bts_tracer); | 300 | return register_tracer(&bts_tracer); |
| 194 | } | 301 | } |
| 195 | device_initcall(init_bts_trace); | 302 | device_initcall(init_bts_trace); |
