diff options
-rw-r--r-- | arch/x86/include/asm/ds.h | 241 | ||||
-rw-r--r-- | arch/x86/include/asm/processor.h | 13 | ||||
-rw-r--r-- | arch/x86/include/asm/ptrace.h | 36 | ||||
-rw-r--r-- | arch/x86/include/asm/thread_info.h | 5 | ||||
-rw-r--r-- | arch/x86/kernel/cpu/intel.c | 4 | ||||
-rw-r--r-- | arch/x86/kernel/ds.c | 859 | ||||
-rw-r--r-- | arch/x86/kernel/process_32.c | 59 | ||||
-rw-r--r-- | arch/x86/kernel/process_64.c | 50 | ||||
-rw-r--r-- | arch/x86/kernel/ptrace.c | 415 | ||||
-rw-r--r-- | include/linux/sched.h | 1 | ||||
-rw-r--r-- | include/trace/boot.h | 8 | ||||
-rw-r--r-- | include/trace/sched.h | 4 | ||||
-rw-r--r-- | kernel/sched.c | 3 | ||||
-rw-r--r-- | kernel/trace/Kconfig | 4 | ||||
-rw-r--r-- | kernel/trace/Makefile | 2 | ||||
-rw-r--r-- | kernel/trace/trace.c | 2 | ||||
-rw-r--r-- | kernel/trace/trace.h | 14 | ||||
-rw-r--r-- | kernel/trace/trace_hw_branches.c (renamed from kernel/trace/trace_bts.c) | 163 |
18 files changed, 878 insertions, 1005 deletions
diff --git a/arch/x86/include/asm/ds.h b/arch/x86/include/asm/ds.h index 99b6c39774a4..ee0ea3a96c11 100644 --- a/arch/x86/include/asm/ds.h +++ b/arch/x86/include/asm/ds.h | |||
@@ -6,13 +6,13 @@ | |||
6 | * precise-event based sampling (PEBS). | 6 | * precise-event based sampling (PEBS). |
7 | * | 7 | * |
8 | * It manages: | 8 | * It manages: |
9 | * - per-thread and per-cpu allocation of BTS and PEBS | 9 | * - DS and BTS hardware configuration |
10 | * - buffer overflow handling (to be done) | 10 | * - buffer overflow handling (to be done) |
11 | * - buffer access | 11 | * - buffer access |
12 | * | 12 | * |
13 | * It assumes: | 13 | * It does not do: |
14 | * - get_task_struct on all traced tasks | 14 | * - security checking (is the caller allowed to trace the task) |
15 | * - current is allowed to trace tasks | 15 | * - buffer allocation (memory accounting) |
16 | * | 16 | * |
17 | * | 17 | * |
18 | * Copyright (C) 2007-2008 Intel Corporation. | 18 | * Copyright (C) 2007-2008 Intel Corporation. |
@@ -31,6 +31,7 @@ | |||
31 | #ifdef CONFIG_X86_DS | 31 | #ifdef CONFIG_X86_DS |
32 | 32 | ||
33 | struct task_struct; | 33 | struct task_struct; |
34 | struct ds_context; | ||
34 | struct ds_tracer; | 35 | struct ds_tracer; |
35 | struct bts_tracer; | 36 | struct bts_tracer; |
36 | struct pebs_tracer; | 37 | struct pebs_tracer; |
@@ -38,6 +39,38 @@ struct pebs_tracer; | |||
38 | typedef void (*bts_ovfl_callback_t)(struct bts_tracer *); | 39 | typedef void (*bts_ovfl_callback_t)(struct bts_tracer *); |
39 | typedef void (*pebs_ovfl_callback_t)(struct pebs_tracer *); | 40 | typedef void (*pebs_ovfl_callback_t)(struct pebs_tracer *); |
40 | 41 | ||
42 | |||
43 | /* | ||
44 | * A list of features plus corresponding macros to talk about them in | ||
45 | * the ds_request function's flags parameter. | ||
46 | * | ||
47 | * We use the enum to index an array of corresponding control bits; | ||
48 | * we use the macro to index a flags bit-vector. | ||
49 | */ | ||
50 | enum ds_feature { | ||
51 | dsf_bts = 0, | ||
52 | dsf_bts_kernel, | ||
53 | #define BTS_KERNEL (1 << dsf_bts_kernel) | ||
54 | /* trace kernel-mode branches */ | ||
55 | |||
56 | dsf_bts_user, | ||
57 | #define BTS_USER (1 << dsf_bts_user) | ||
58 | /* trace user-mode branches */ | ||
59 | |||
60 | dsf_bts_overflow, | ||
61 | dsf_bts_max, | ||
62 | dsf_pebs = dsf_bts_max, | ||
63 | |||
64 | dsf_pebs_max, | ||
65 | dsf_ctl_max = dsf_pebs_max, | ||
66 | dsf_bts_timestamps = dsf_ctl_max, | ||
67 | #define BTS_TIMESTAMPS (1 << dsf_bts_timestamps) | ||
68 | /* add timestamps into BTS trace */ | ||
69 | |||
70 | #define BTS_USER_FLAGS (BTS_KERNEL | BTS_USER | BTS_TIMESTAMPS) | ||
71 | }; | ||
72 | |||
73 | |||
41 | /* | 74 | /* |
42 | * Request BTS or PEBS | 75 | * Request BTS or PEBS |
43 | * | 76 | * |
@@ -58,92 +91,135 @@ typedef void (*pebs_ovfl_callback_t)(struct pebs_tracer *); | |||
58 | * NULL if cyclic buffer requested | 91 | * NULL if cyclic buffer requested |
59 | * th: the interrupt threshold in records from the end of the buffer; | 92 | * th: the interrupt threshold in records from the end of the buffer; |
60 | * -1 if no interrupt threshold is requested. | 93 | * -1 if no interrupt threshold is requested. |
94 | * flags: a bit-mask of the above flags | ||
61 | */ | 95 | */ |
62 | extern struct bts_tracer *ds_request_bts(struct task_struct *task, | 96 | extern struct bts_tracer *ds_request_bts(struct task_struct *task, |
63 | void *base, size_t size, | 97 | void *base, size_t size, |
64 | bts_ovfl_callback_t ovfl, size_t th); | 98 | bts_ovfl_callback_t ovfl, |
99 | size_t th, unsigned int flags); | ||
65 | extern struct pebs_tracer *ds_request_pebs(struct task_struct *task, | 100 | extern struct pebs_tracer *ds_request_pebs(struct task_struct *task, |
66 | void *base, size_t size, | 101 | void *base, size_t size, |
67 | pebs_ovfl_callback_t ovfl, | 102 | pebs_ovfl_callback_t ovfl, |
68 | size_t th); | 103 | size_t th, unsigned int flags); |
69 | 104 | ||
70 | /* | 105 | /* |
71 | * Release BTS or PEBS resources | 106 | * Release BTS or PEBS resources |
72 | * | 107 | * Suspend and resume BTS or PEBS tracing |
73 | * Returns 0 on success; -Eerrno otherwise | ||
74 | * | 108 | * |
75 | * tracer: the tracer handle returned from ds_request_~() | 109 | * tracer: the tracer handle returned from ds_request_~() |
76 | */ | 110 | */ |
77 | extern int ds_release_bts(struct bts_tracer *tracer); | 111 | extern void ds_release_bts(struct bts_tracer *tracer); |
78 | extern int ds_release_pebs(struct pebs_tracer *tracer); | 112 | extern void ds_suspend_bts(struct bts_tracer *tracer); |
113 | extern void ds_resume_bts(struct bts_tracer *tracer); | ||
114 | extern void ds_release_pebs(struct pebs_tracer *tracer); | ||
115 | extern void ds_suspend_pebs(struct pebs_tracer *tracer); | ||
116 | extern void ds_resume_pebs(struct pebs_tracer *tracer); | ||
117 | |||
79 | 118 | ||
80 | /* | 119 | /* |
81 | * Get the (array) index of the write pointer. | 120 | * The raw DS buffer state as it is used for BTS and PEBS recording. |
82 | * (assuming an array of BTS/PEBS records) | ||
83 | * | ||
84 | * Returns 0 on success; -Eerrno on error | ||
85 | * | 121 | * |
86 | * tracer: the tracer handle returned from ds_request_~() | 122 | * This is the low-level, arch-dependent interface for working |
87 | * pos (out): will hold the result | 123 | * directly on the raw trace data. |
88 | */ | 124 | */ |
89 | extern int ds_get_bts_index(struct bts_tracer *tracer, size_t *pos); | 125 | struct ds_trace { |
90 | extern int ds_get_pebs_index(struct pebs_tracer *tracer, size_t *pos); | 126 | /* the number of bts/pebs records */ |
127 | size_t n; | ||
128 | /* the size of a bts/pebs record in bytes */ | ||
129 | size_t size; | ||
130 | /* pointers into the raw buffer: | ||
131 | - to the first entry */ | ||
132 | void *begin; | ||
133 | /* - one beyond the last entry */ | ||
134 | void *end; | ||
135 | /* - one beyond the newest entry */ | ||
136 | void *top; | ||
137 | /* - the interrupt threshold */ | ||
138 | void *ith; | ||
139 | /* flags given on ds_request() */ | ||
140 | unsigned int flags; | ||
141 | }; | ||
91 | 142 | ||
92 | /* | 143 | /* |
93 | * Get the (array) index one record beyond the end of the array. | 144 | * An arch-independent view on branch trace data. |
94 | * (assuming an array of BTS/PEBS records) | ||
95 | * | ||
96 | * Returns 0 on success; -Eerrno on error | ||
97 | * | ||
98 | * tracer: the tracer handle returned from ds_request_~() | ||
99 | * pos (out): will hold the result | ||
100 | */ | 145 | */ |
101 | extern int ds_get_bts_end(struct bts_tracer *tracer, size_t *pos); | 146 | enum bts_qualifier { |
102 | extern int ds_get_pebs_end(struct pebs_tracer *tracer, size_t *pos); | 147 | bts_invalid, |
148 | #define BTS_INVALID bts_invalid | ||
149 | |||
150 | bts_branch, | ||
151 | #define BTS_BRANCH bts_branch | ||
152 | |||
153 | bts_task_arrives, | ||
154 | #define BTS_TASK_ARRIVES bts_task_arrives | ||
155 | |||
156 | bts_task_departs, | ||
157 | #define BTS_TASK_DEPARTS bts_task_departs | ||
158 | |||
159 | bts_qual_bit_size = 4, | ||
160 | bts_qual_max = (1 << bts_qual_bit_size), | ||
161 | }; | ||
162 | |||
163 | struct bts_struct { | ||
164 | __u64 qualifier; | ||
165 | union { | ||
166 | /* BTS_BRANCH */ | ||
167 | struct { | ||
168 | __u64 from; | ||
169 | __u64 to; | ||
170 | } lbr; | ||
171 | /* BTS_TASK_ARRIVES or BTS_TASK_DEPARTS */ | ||
172 | struct { | ||
173 | __u64 jiffies; | ||
174 | pid_t pid; | ||
175 | } timestamp; | ||
176 | } variant; | ||
177 | }; | ||
178 | |||
103 | 179 | ||
104 | /* | 180 | /* |
105 | * Provide a pointer to the BTS/PEBS record at parameter index. | 181 | * The BTS state. |
106 | * (assuming an array of BTS/PEBS records) | ||
107 | * | ||
108 | * The pointer points directly into the buffer. The user is | ||
109 | * responsible for copying the record. | ||
110 | * | ||
111 | * Returns the size of a single record on success; -Eerrno on error | ||
112 | * | 182 | * |
113 | * tracer: the tracer handle returned from ds_request_~() | 183 | * This gives access to the raw DS state and adds functions to provide |
114 | * index: the index of the requested record | 184 | * an arch-independent view of the BTS data. |
115 | * record (out): pointer to the requested record | ||
116 | */ | 185 | */ |
117 | extern int ds_access_bts(struct bts_tracer *tracer, | 186 | struct bts_trace { |
118 | size_t index, const void **record); | 187 | struct ds_trace ds; |
119 | extern int ds_access_pebs(struct pebs_tracer *tracer, | 188 | |
120 | size_t index, const void **record); | 189 | int (*read)(struct bts_tracer *tracer, const void *at, |
190 | struct bts_struct *out); | ||
191 | int (*write)(struct bts_tracer *tracer, const struct bts_struct *in); | ||
192 | }; | ||
193 | |||
121 | 194 | ||
122 | /* | 195 | /* |
123 | * Write one or more BTS/PEBS records at the write pointer index and | 196 | * The PEBS state. |
124 | * advance the write pointer. | ||
125 | * | 197 | * |
126 | * If size is not a multiple of the record size, trailing bytes are | 198 | * This gives access to the raw DS state and the PEBS-specific counter |
127 | * zeroed out. | 199 | * reset value. |
128 | * | 200 | */ |
129 | * May result in one or more overflow notifications. | 201 | struct pebs_trace { |
130 | * | 202 | struct ds_trace ds; |
131 | * If called during overflow handling, that is, with index >= | 203 | |
132 | * interrupt threshold, the write will wrap around. | 204 | /* the PEBS reset value */ |
205 | unsigned long long reset_value; | ||
206 | }; | ||
207 | |||
208 | |||
209 | /* | ||
210 | * Read the BTS or PEBS trace. | ||
133 | * | 211 | * |
134 | * An overflow notification is given if and when the interrupt | 212 | * Returns a view on the trace collected for the parameter tracer. |
135 | * threshold is reached during or after the write. | ||
136 | * | 213 | * |
137 | * Returns the number of bytes written or -Eerrno. | 214 | * The view remains valid as long as the traced task is not running or |
215 | * the tracer is suspended. | ||
216 | * Writes into the trace buffer are not reflected. | ||
138 | * | 217 | * |
139 | * tracer: the tracer handle returned from ds_request_~() | 218 | * tracer: the tracer handle returned from ds_request_~() |
140 | * buffer: the buffer to write | ||
141 | * size: the size of the buffer | ||
142 | */ | 219 | */ |
143 | extern int ds_write_bts(struct bts_tracer *tracer, | 220 | extern const struct bts_trace *ds_read_bts(struct bts_tracer *tracer); |
144 | const void *buffer, size_t size); | 221 | extern const struct pebs_trace *ds_read_pebs(struct pebs_tracer *tracer); |
145 | extern int ds_write_pebs(struct pebs_tracer *tracer, | 222 | |
146 | const void *buffer, size_t size); | ||
147 | 223 | ||
148 | /* | 224 | /* |
149 | * Reset the write pointer of the BTS/PEBS buffer. | 225 | * Reset the write pointer of the BTS/PEBS buffer. |
@@ -156,27 +232,6 @@ extern int ds_reset_bts(struct bts_tracer *tracer); | |||
156 | extern int ds_reset_pebs(struct pebs_tracer *tracer); | 232 | extern int ds_reset_pebs(struct pebs_tracer *tracer); |
157 | 233 | ||
158 | /* | 234 | /* |
159 | * Clear the BTS/PEBS buffer and reset the write pointer. | ||
160 | * The entire buffer will be zeroed out. | ||
161 | * | ||
162 | * Returns 0 on success; -Eerrno on error | ||
163 | * | ||
164 | * tracer: the tracer handle returned from ds_request_~() | ||
165 | */ | ||
166 | extern int ds_clear_bts(struct bts_tracer *tracer); | ||
167 | extern int ds_clear_pebs(struct pebs_tracer *tracer); | ||
168 | |||
169 | /* | ||
170 | * Provide the PEBS counter reset value. | ||
171 | * | ||
172 | * Returns 0 on success; -Eerrno on error | ||
173 | * | ||
174 | * tracer: the tracer handle returned from ds_request_pebs() | ||
175 | * value (out): the counter reset value | ||
176 | */ | ||
177 | extern int ds_get_pebs_reset(struct pebs_tracer *tracer, u64 *value); | ||
178 | |||
179 | /* | ||
180 | * Set the PEBS counter reset value. | 235 | * Set the PEBS counter reset value. |
181 | * | 236 | * |
182 | * Returns 0 on success; -Eerrno on error | 237 | * Returns 0 on success; -Eerrno on error |
@@ -192,35 +247,17 @@ extern int ds_set_pebs_reset(struct pebs_tracer *tracer, u64 value); | |||
192 | struct cpuinfo_x86; | 247 | struct cpuinfo_x86; |
193 | extern void __cpuinit ds_init_intel(struct cpuinfo_x86 *); | 248 | extern void __cpuinit ds_init_intel(struct cpuinfo_x86 *); |
194 | 249 | ||
195 | |||
196 | |||
197 | /* | 250 | /* |
198 | * The DS context - part of struct thread_struct. | 251 | * Context switch work |
199 | */ | 252 | */ |
200 | #define MAX_SIZEOF_DS (12 * 8) | 253 | extern void ds_switch_to(struct task_struct *prev, struct task_struct *next); |
201 | |||
202 | struct ds_context { | ||
203 | /* pointer to the DS configuration; goes into MSR_IA32_DS_AREA */ | ||
204 | unsigned char ds[MAX_SIZEOF_DS]; | ||
205 | /* the owner of the BTS and PEBS configuration, respectively */ | ||
206 | struct ds_tracer *owner[2]; | ||
207 | /* use count */ | ||
208 | unsigned long count; | ||
209 | /* a pointer to the context location inside the thread_struct | ||
210 | * or the per_cpu context array */ | ||
211 | struct ds_context **this; | ||
212 | /* a pointer to the task owning this context, or NULL, if the | ||
213 | * context is owned by a cpu */ | ||
214 | struct task_struct *task; | ||
215 | }; | ||
216 | |||
217 | /* called by exit_thread() to free leftover contexts */ | ||
218 | extern void ds_free(struct ds_context *context); | ||
219 | 254 | ||
220 | #else /* CONFIG_X86_DS */ | 255 | #else /* CONFIG_X86_DS */ |
221 | 256 | ||
222 | struct cpuinfo_x86; | 257 | struct cpuinfo_x86; |
223 | static inline void __cpuinit ds_init_intel(struct cpuinfo_x86 *ignored) {} | 258 | static inline void __cpuinit ds_init_intel(struct cpuinfo_x86 *ignored) {} |
259 | static inline void ds_switch_to(struct task_struct *prev, | ||
260 | struct task_struct *next) {} | ||
224 | 261 | ||
225 | #endif /* CONFIG_X86_DS */ | 262 | #endif /* CONFIG_X86_DS */ |
226 | #endif /* _ASM_X86_DS_H */ | 263 | #endif /* _ASM_X86_DS_H */ |
diff --git a/arch/x86/include/asm/processor.h b/arch/x86/include/asm/processor.h index 5ca01e383269..aa5914f8e501 100644 --- a/arch/x86/include/asm/processor.h +++ b/arch/x86/include/asm/processor.h | |||
@@ -752,6 +752,19 @@ extern void switch_to_new_gdt(void); | |||
752 | extern void cpu_init(void); | 752 | extern void cpu_init(void); |
753 | extern void init_gdt(int cpu); | 753 | extern void init_gdt(int cpu); |
754 | 754 | ||
755 | static inline unsigned long get_debugctlmsr(void) | ||
756 | { | ||
757 | unsigned long debugctlmsr = 0; | ||
758 | |||
759 | #ifndef CONFIG_X86_DEBUGCTLMSR | ||
760 | if (boot_cpu_data.x86 < 6) | ||
761 | return 0; | ||
762 | #endif | ||
763 | rdmsrl(MSR_IA32_DEBUGCTLMSR, debugctlmsr); | ||
764 | |||
765 | return debugctlmsr; | ||
766 | } | ||
767 | |||
755 | static inline void update_debugctlmsr(unsigned long debugctlmsr) | 768 | static inline void update_debugctlmsr(unsigned long debugctlmsr) |
756 | { | 769 | { |
757 | #ifndef CONFIG_X86_DEBUGCTLMSR | 770 | #ifndef CONFIG_X86_DEBUGCTLMSR |
diff --git a/arch/x86/include/asm/ptrace.h b/arch/x86/include/asm/ptrace.h index eefb0594b058..fbf744215911 100644 --- a/arch/x86/include/asm/ptrace.h +++ b/arch/x86/include/asm/ptrace.h | |||
@@ -6,7 +6,6 @@ | |||
6 | #include <asm/processor-flags.h> | 6 | #include <asm/processor-flags.h> |
7 | 7 | ||
8 | #ifdef __KERNEL__ | 8 | #ifdef __KERNEL__ |
9 | #include <asm/ds.h> /* the DS BTS struct is used for ptrace too */ | ||
10 | #include <asm/segment.h> | 9 | #include <asm/segment.h> |
11 | #endif | 10 | #endif |
12 | 11 | ||
@@ -128,34 +127,6 @@ struct pt_regs { | |||
128 | #endif /* !__i386__ */ | 127 | #endif /* !__i386__ */ |
129 | 128 | ||
130 | 129 | ||
131 | #ifdef CONFIG_X86_PTRACE_BTS | ||
132 | /* a branch trace record entry | ||
133 | * | ||
134 | * In order to unify the interface between various processor versions, | ||
135 | * we use the below data structure for all processors. | ||
136 | */ | ||
137 | enum bts_qualifier { | ||
138 | BTS_INVALID = 0, | ||
139 | BTS_BRANCH, | ||
140 | BTS_TASK_ARRIVES, | ||
141 | BTS_TASK_DEPARTS | ||
142 | }; | ||
143 | |||
144 | struct bts_struct { | ||
145 | __u64 qualifier; | ||
146 | union { | ||
147 | /* BTS_BRANCH */ | ||
148 | struct { | ||
149 | __u64 from_ip; | ||
150 | __u64 to_ip; | ||
151 | } lbr; | ||
152 | /* BTS_TASK_ARRIVES or | ||
153 | BTS_TASK_DEPARTS */ | ||
154 | __u64 jiffies; | ||
155 | } variant; | ||
156 | }; | ||
157 | #endif /* CONFIG_X86_PTRACE_BTS */ | ||
158 | |||
159 | #ifdef __KERNEL__ | 130 | #ifdef __KERNEL__ |
160 | 131 | ||
161 | #include <linux/init.h> | 132 | #include <linux/init.h> |
@@ -163,13 +134,6 @@ struct bts_struct { | |||
163 | struct cpuinfo_x86; | 134 | struct cpuinfo_x86; |
164 | struct task_struct; | 135 | struct task_struct; |
165 | 136 | ||
166 | #ifdef CONFIG_X86_PTRACE_BTS | ||
167 | extern void __cpuinit ptrace_bts_init_intel(struct cpuinfo_x86 *); | ||
168 | extern void ptrace_bts_take_timestamp(struct task_struct *, enum bts_qualifier); | ||
169 | #else | ||
170 | #define ptrace_bts_init_intel(config) do {} while (0) | ||
171 | #endif /* CONFIG_X86_PTRACE_BTS */ | ||
172 | |||
173 | extern unsigned long profile_pc(struct pt_regs *regs); | 137 | extern unsigned long profile_pc(struct pt_regs *regs); |
174 | 138 | ||
175 | extern unsigned long | 139 | extern unsigned long |
diff --git a/arch/x86/include/asm/thread_info.h b/arch/x86/include/asm/thread_info.h index 0921b4018c11..bf8113d16a33 100644 --- a/arch/x86/include/asm/thread_info.h +++ b/arch/x86/include/asm/thread_info.h | |||
@@ -93,7 +93,6 @@ struct thread_info { | |||
93 | #define TIF_FORCED_TF 24 /* true if TF in eflags artificially */ | 93 | #define TIF_FORCED_TF 24 /* true if TF in eflags artificially */ |
94 | #define TIF_DEBUGCTLMSR 25 /* uses thread_struct.debugctlmsr */ | 94 | #define TIF_DEBUGCTLMSR 25 /* uses thread_struct.debugctlmsr */ |
95 | #define TIF_DS_AREA_MSR 26 /* uses thread_struct.ds_area_msr */ | 95 | #define TIF_DS_AREA_MSR 26 /* uses thread_struct.ds_area_msr */ |
96 | #define TIF_BTS_TRACE_TS 27 /* record scheduling event timestamps */ | ||
97 | 96 | ||
98 | #define _TIF_SYSCALL_TRACE (1 << TIF_SYSCALL_TRACE) | 97 | #define _TIF_SYSCALL_TRACE (1 << TIF_SYSCALL_TRACE) |
99 | #define _TIF_NOTIFY_RESUME (1 << TIF_NOTIFY_RESUME) | 98 | #define _TIF_NOTIFY_RESUME (1 << TIF_NOTIFY_RESUME) |
@@ -115,7 +114,6 @@ struct thread_info { | |||
115 | #define _TIF_FORCED_TF (1 << TIF_FORCED_TF) | 114 | #define _TIF_FORCED_TF (1 << TIF_FORCED_TF) |
116 | #define _TIF_DEBUGCTLMSR (1 << TIF_DEBUGCTLMSR) | 115 | #define _TIF_DEBUGCTLMSR (1 << TIF_DEBUGCTLMSR) |
117 | #define _TIF_DS_AREA_MSR (1 << TIF_DS_AREA_MSR) | 116 | #define _TIF_DS_AREA_MSR (1 << TIF_DS_AREA_MSR) |
118 | #define _TIF_BTS_TRACE_TS (1 << TIF_BTS_TRACE_TS) | ||
119 | 117 | ||
120 | /* work to do in syscall_trace_enter() */ | 118 | /* work to do in syscall_trace_enter() */ |
121 | #define _TIF_WORK_SYSCALL_ENTRY \ | 119 | #define _TIF_WORK_SYSCALL_ENTRY \ |
@@ -141,8 +139,7 @@ struct thread_info { | |||
141 | 139 | ||
142 | /* flags to check in __switch_to() */ | 140 | /* flags to check in __switch_to() */ |
143 | #define _TIF_WORK_CTXSW \ | 141 | #define _TIF_WORK_CTXSW \ |
144 | (_TIF_IO_BITMAP|_TIF_DEBUGCTLMSR|_TIF_DS_AREA_MSR|_TIF_BTS_TRACE_TS| \ | 142 | (_TIF_IO_BITMAP|_TIF_DEBUGCTLMSR|_TIF_DS_AREA_MSR|_TIF_NOTSC) |
145 | _TIF_NOTSC) | ||
146 | 143 | ||
147 | #define _TIF_WORK_CTXSW_PREV _TIF_WORK_CTXSW | 144 | #define _TIF_WORK_CTXSW_PREV _TIF_WORK_CTXSW |
148 | #define _TIF_WORK_CTXSW_NEXT (_TIF_WORK_CTXSW|_TIF_DEBUG) | 145 | #define _TIF_WORK_CTXSW_NEXT (_TIF_WORK_CTXSW|_TIF_DEBUG) |
diff --git a/arch/x86/kernel/cpu/intel.c b/arch/x86/kernel/cpu/intel.c index 816f27f289b1..cd413d9a0218 100644 --- a/arch/x86/kernel/cpu/intel.c +++ b/arch/x86/kernel/cpu/intel.c | |||
@@ -11,7 +11,6 @@ | |||
11 | #include <asm/pgtable.h> | 11 | #include <asm/pgtable.h> |
12 | #include <asm/msr.h> | 12 | #include <asm/msr.h> |
13 | #include <asm/uaccess.h> | 13 | #include <asm/uaccess.h> |
14 | #include <asm/ptrace.h> | ||
15 | #include <asm/ds.h> | 14 | #include <asm/ds.h> |
16 | #include <asm/bugs.h> | 15 | #include <asm/bugs.h> |
17 | 16 | ||
@@ -309,9 +308,6 @@ static void __cpuinit init_intel(struct cpuinfo_x86 *c) | |||
309 | set_cpu_cap(c, X86_FEATURE_P3); | 308 | set_cpu_cap(c, X86_FEATURE_P3); |
310 | #endif | 309 | #endif |
311 | 310 | ||
312 | if (cpu_has_bts) | ||
313 | ptrace_bts_init_intel(c); | ||
314 | |||
315 | detect_extended_topology(c); | 311 | detect_extended_topology(c); |
316 | if (!cpu_has(c, X86_FEATURE_XTOPOLOGY)) { | 312 | if (!cpu_has(c, X86_FEATURE_XTOPOLOGY)) { |
317 | /* | 313 | /* |
diff --git a/arch/x86/kernel/ds.c b/arch/x86/kernel/ds.c index 19a8c2c0389f..dc1e7123ea4e 100644 --- a/arch/x86/kernel/ds.c +++ b/arch/x86/kernel/ds.c | |||
@@ -6,13 +6,13 @@ | |||
6 | * precise-event based sampling (PEBS). | 6 | * precise-event based sampling (PEBS). |
7 | * | 7 | * |
8 | * It manages: | 8 | * It manages: |
9 | * - per-thread and per-cpu allocation of BTS and PEBS | 9 | * - DS and BTS hardware configuration |
10 | * - buffer overflow handling (to be done) | 10 | * - buffer overflow handling (to be done) |
11 | * - buffer access | 11 | * - buffer access |
12 | * | 12 | * |
13 | * It assumes: | 13 | * It does not do: |
14 | * - get_task_struct on all traced tasks | 14 | * - security checking (is the caller allowed to trace the task) |
15 | * - current is allowed to trace tasks | 15 | * - buffer allocation (memory accounting) |
16 | * | 16 | * |
17 | * | 17 | * |
18 | * Copyright (C) 2007-2008 Intel Corporation. | 18 | * Copyright (C) 2007-2008 Intel Corporation. |
@@ -34,15 +34,30 @@ | |||
34 | * The configuration for a particular DS hardware implementation. | 34 | * The configuration for a particular DS hardware implementation. |
35 | */ | 35 | */ |
36 | struct ds_configuration { | 36 | struct ds_configuration { |
37 | /* the size of the DS structure in bytes */ | 37 | /* the name of the configuration */ |
38 | unsigned char sizeof_ds; | 38 | const char *name; |
39 | /* the size of one pointer-typed field in the DS structure in bytes; | 39 | /* the size of one pointer-typed field in the DS structure and |
40 | this covers the first 8 fields related to buffer management. */ | 40 | in the BTS and PEBS buffers in bytes; |
41 | this covers the first 8 DS fields related to buffer management. */ | ||
41 | unsigned char sizeof_field; | 42 | unsigned char sizeof_field; |
42 | /* the size of a BTS/PEBS record in bytes */ | 43 | /* the size of a BTS/PEBS record in bytes */ |
43 | unsigned char sizeof_rec[2]; | 44 | unsigned char sizeof_rec[2]; |
45 | /* a series of bit-masks to control various features indexed | ||
46 | * by enum ds_feature */ | ||
47 | unsigned long ctl[dsf_ctl_max]; | ||
44 | }; | 48 | }; |
45 | static struct ds_configuration ds_cfg; | 49 | static DEFINE_PER_CPU(struct ds_configuration, ds_cfg_array); |
50 | |||
51 | #define ds_cfg per_cpu(ds_cfg_array, smp_processor_id()) | ||
52 | |||
53 | #define MAX_SIZEOF_DS (12 * 8) /* maximal size of a DS configuration */ | ||
54 | #define MAX_SIZEOF_BTS (3 * 8) /* maximal size of a BTS record */ | ||
55 | #define DS_ALIGNMENT (1 << 3) /* BTS and PEBS buffer alignment */ | ||
56 | |||
57 | #define BTS_CONTROL \ | ||
58 | (ds_cfg.ctl[dsf_bts] | ds_cfg.ctl[dsf_bts_kernel] | ds_cfg.ctl[dsf_bts_user] |\ | ||
59 | ds_cfg.ctl[dsf_bts_overflow]) | ||
60 | |||
46 | 61 | ||
47 | /* | 62 | /* |
48 | * A BTS or PEBS tracer. | 63 | * A BTS or PEBS tracer. |
@@ -61,6 +76,8 @@ struct ds_tracer { | |||
61 | struct bts_tracer { | 76 | struct bts_tracer { |
62 | /* the common DS part */ | 77 | /* the common DS part */ |
63 | struct ds_tracer ds; | 78 | struct ds_tracer ds; |
79 | /* the trace including the DS configuration */ | ||
80 | struct bts_trace trace; | ||
64 | /* buffer overflow notification function */ | 81 | /* buffer overflow notification function */ |
65 | bts_ovfl_callback_t ovfl; | 82 | bts_ovfl_callback_t ovfl; |
66 | }; | 83 | }; |
@@ -68,6 +85,8 @@ struct bts_tracer { | |||
68 | struct pebs_tracer { | 85 | struct pebs_tracer { |
69 | /* the common DS part */ | 86 | /* the common DS part */ |
70 | struct ds_tracer ds; | 87 | struct ds_tracer ds; |
88 | /* the trace including the DS configuration */ | ||
89 | struct pebs_trace trace; | ||
71 | /* buffer overflow notification function */ | 90 | /* buffer overflow notification function */ |
72 | pebs_ovfl_callback_t ovfl; | 91 | pebs_ovfl_callback_t ovfl; |
73 | }; | 92 | }; |
@@ -134,13 +153,11 @@ static inline void ds_set(unsigned char *base, enum ds_qualifier qual, | |||
134 | (*(unsigned long *)base) = value; | 153 | (*(unsigned long *)base) = value; |
135 | } | 154 | } |
136 | 155 | ||
137 | #define DS_ALIGNMENT (1 << 3) /* BTS and PEBS buffer alignment */ | ||
138 | |||
139 | 156 | ||
140 | /* | 157 | /* |
141 | * Locking is done only for allocating BTS or PEBS resources. | 158 | * Locking is done only for allocating BTS or PEBS resources. |
142 | */ | 159 | */ |
143 | static spinlock_t ds_lock = __SPIN_LOCK_UNLOCKED(ds_lock); | 160 | static DEFINE_SPINLOCK(ds_lock); |
144 | 161 | ||
145 | 162 | ||
146 | /* | 163 | /* |
@@ -156,27 +173,32 @@ static spinlock_t ds_lock = __SPIN_LOCK_UNLOCKED(ds_lock); | |||
156 | * >0 number of per-thread tracers | 173 | * >0 number of per-thread tracers |
157 | * <0 number of per-cpu tracers | 174 | * <0 number of per-cpu tracers |
158 | * | 175 | * |
159 | * The below functions to get and put tracers and to check the | ||
160 | * allocation type require the ds_lock to be held by the caller. | ||
161 | * | ||
162 | * Tracers essentially gives the number of ds contexts for a certain | 176 | * Tracers essentially gives the number of ds contexts for a certain |
163 | * type of allocation. | 177 | * type of allocation. |
164 | */ | 178 | */ |
165 | static long tracers; | 179 | static atomic_t tracers = ATOMIC_INIT(0); |
166 | 180 | ||
167 | static inline void get_tracer(struct task_struct *task) | 181 | static inline void get_tracer(struct task_struct *task) |
168 | { | 182 | { |
169 | tracers += (task ? 1 : -1); | 183 | if (task) |
184 | atomic_inc(&tracers); | ||
185 | else | ||
186 | atomic_dec(&tracers); | ||
170 | } | 187 | } |
171 | 188 | ||
172 | static inline void put_tracer(struct task_struct *task) | 189 | static inline void put_tracer(struct task_struct *task) |
173 | { | 190 | { |
174 | tracers -= (task ? 1 : -1); | 191 | if (task) |
192 | atomic_dec(&tracers); | ||
193 | else | ||
194 | atomic_inc(&tracers); | ||
175 | } | 195 | } |
176 | 196 | ||
177 | static inline int check_tracer(struct task_struct *task) | 197 | static inline int check_tracer(struct task_struct *task) |
178 | { | 198 | { |
179 | return (task ? (tracers >= 0) : (tracers <= 0)); | 199 | return task ? |
200 | (atomic_read(&tracers) >= 0) : | ||
201 | (atomic_read(&tracers) <= 0); | ||
180 | } | 202 | } |
181 | 203 | ||
182 | 204 | ||
@@ -190,14 +212,30 @@ static inline int check_tracer(struct task_struct *task) | |||
190 | * Contexts are use-counted. They are allocated on first access and | 212 | * Contexts are use-counted. They are allocated on first access and |
191 | * deallocated when the last user puts the context. | 213 | * deallocated when the last user puts the context. |
192 | */ | 214 | */ |
193 | static DEFINE_PER_CPU(struct ds_context *, system_context); | 215 | struct ds_context { |
216 | /* pointer to the DS configuration; goes into MSR_IA32_DS_AREA */ | ||
217 | unsigned char ds[MAX_SIZEOF_DS]; | ||
218 | /* the owner of the BTS and PEBS configuration, respectively */ | ||
219 | struct bts_tracer *bts_master; | ||
220 | struct pebs_tracer *pebs_master; | ||
221 | /* use count */ | ||
222 | unsigned long count; | ||
223 | /* a pointer to the context location inside the thread_struct | ||
224 | * or the per_cpu context array */ | ||
225 | struct ds_context **this; | ||
226 | /* a pointer to the task owning this context, or NULL, if the | ||
227 | * context is owned by a cpu */ | ||
228 | struct task_struct *task; | ||
229 | }; | ||
230 | |||
231 | static DEFINE_PER_CPU(struct ds_context *, system_context_array); | ||
194 | 232 | ||
195 | #define this_system_context per_cpu(system_context, smp_processor_id()) | 233 | #define system_context per_cpu(system_context_array, smp_processor_id()) |
196 | 234 | ||
197 | static inline struct ds_context *ds_get_context(struct task_struct *task) | 235 | static struct ds_context *ds_get_context(struct task_struct *task) |
198 | { | 236 | { |
199 | struct ds_context **p_context = | 237 | struct ds_context **p_context = |
200 | (task ? &task->thread.ds_ctx : &this_system_context); | 238 | (task ? &task->thread.ds_ctx : &system_context); |
201 | struct ds_context *context = *p_context; | 239 | struct ds_context *context = *p_context; |
202 | unsigned long irq; | 240 | unsigned long irq; |
203 | 241 | ||
@@ -225,10 +263,22 @@ static inline struct ds_context *ds_get_context(struct task_struct *task) | |||
225 | wrmsrl(MSR_IA32_DS_AREA, | 263 | wrmsrl(MSR_IA32_DS_AREA, |
226 | (unsigned long)context->ds); | 264 | (unsigned long)context->ds); |
227 | } | 265 | } |
266 | |||
267 | context->count++; | ||
268 | |||
269 | spin_unlock_irqrestore(&ds_lock, irq); | ||
270 | } else { | ||
271 | spin_lock_irqsave(&ds_lock, irq); | ||
272 | |||
273 | context = *p_context; | ||
274 | if (context) | ||
275 | context->count++; | ||
276 | |||
228 | spin_unlock_irqrestore(&ds_lock, irq); | 277 | spin_unlock_irqrestore(&ds_lock, irq); |
229 | } | ||
230 | 278 | ||
231 | context->count++; | 279 | if (!context) |
280 | context = ds_get_context(task); | ||
281 | } | ||
232 | 282 | ||
233 | return context; | 283 | return context; |
234 | } | 284 | } |
@@ -242,8 +292,10 @@ static inline void ds_put_context(struct ds_context *context) | |||
242 | 292 | ||
243 | spin_lock_irqsave(&ds_lock, irq); | 293 | spin_lock_irqsave(&ds_lock, irq); |
244 | 294 | ||
245 | if (--context->count) | 295 | if (--context->count) { |
246 | goto out; | 296 | spin_unlock_irqrestore(&ds_lock, irq); |
297 | return; | ||
298 | } | ||
247 | 299 | ||
248 | *(context->this) = NULL; | 300 | *(context->this) = NULL; |
249 | 301 | ||
@@ -253,14 +305,14 @@ static inline void ds_put_context(struct ds_context *context) | |||
253 | if (!context->task || (context->task == current)) | 305 | if (!context->task || (context->task == current)) |
254 | wrmsrl(MSR_IA32_DS_AREA, 0); | 306 | wrmsrl(MSR_IA32_DS_AREA, 0); |
255 | 307 | ||
256 | kfree(context); | ||
257 | out: | ||
258 | spin_unlock_irqrestore(&ds_lock, irq); | 308 | spin_unlock_irqrestore(&ds_lock, irq); |
309 | |||
310 | kfree(context); | ||
259 | } | 311 | } |
260 | 312 | ||
261 | 313 | ||
262 | /* | 314 | /* |
263 | * Handle a buffer overflow | 315 | * Call the tracer's callback on a buffer overflow. |
264 | * | 316 | * |
265 | * context: the ds context | 317 | * context: the ds context |
266 | * qual: the buffer type | 318 | * qual: the buffer type |
@@ -268,30 +320,244 @@ static inline void ds_put_context(struct ds_context *context) | |||
268 | static void ds_overflow(struct ds_context *context, enum ds_qualifier qual) | 320 | static void ds_overflow(struct ds_context *context, enum ds_qualifier qual) |
269 | { | 321 | { |
270 | switch (qual) { | 322 | switch (qual) { |
271 | case ds_bts: { | 323 | case ds_bts: |
272 | struct bts_tracer *tracer = | 324 | if (context->bts_master && |
273 | container_of(context->owner[qual], | 325 | context->bts_master->ovfl) |
274 | struct bts_tracer, ds); | 326 | context->bts_master->ovfl(context->bts_master); |
275 | if (tracer->ovfl) | 327 | break; |
276 | tracer->ovfl(tracer); | 328 | case ds_pebs: |
277 | } | 329 | if (context->pebs_master && |
330 | context->pebs_master->ovfl) | ||
331 | context->pebs_master->ovfl(context->pebs_master); | ||
278 | break; | 332 | break; |
279 | case ds_pebs: { | ||
280 | struct pebs_tracer *tracer = | ||
281 | container_of(context->owner[qual], | ||
282 | struct pebs_tracer, ds); | ||
283 | if (tracer->ovfl) | ||
284 | tracer->ovfl(tracer); | ||
285 | } | 333 | } |
334 | } | ||
335 | |||
336 | |||
337 | /* | ||
338 | * Write raw data into the BTS or PEBS buffer. | ||
339 | * | ||
340 | * The remainder of any partially written record is zeroed out. | ||
341 | * | ||
342 | * context: the DS context | ||
343 | * qual: the buffer type | ||
344 | * record: the data to write | ||
345 | * size: the size of the data | ||
346 | */ | ||
347 | static int ds_write(struct ds_context *context, enum ds_qualifier qual, | ||
348 | const void *record, size_t size) | ||
349 | { | ||
350 | int bytes_written = 0; | ||
351 | |||
352 | if (!record) | ||
353 | return -EINVAL; | ||
354 | |||
355 | while (size) { | ||
356 | unsigned long base, index, end, write_end, int_th; | ||
357 | unsigned long write_size, adj_write_size; | ||
358 | |||
359 | /* | ||
360 | * write as much as possible without producing an | ||
361 | * overflow interrupt. | ||
362 | * | ||
363 | * interrupt_threshold must either be | ||
364 | * - bigger than absolute_maximum or | ||
365 | * - point to a record between buffer_base and absolute_maximum | ||
366 | * | ||
367 | * index points to a valid record. | ||
368 | */ | ||
369 | base = ds_get(context->ds, qual, ds_buffer_base); | ||
370 | index = ds_get(context->ds, qual, ds_index); | ||
371 | end = ds_get(context->ds, qual, ds_absolute_maximum); | ||
372 | int_th = ds_get(context->ds, qual, ds_interrupt_threshold); | ||
373 | |||
374 | write_end = min(end, int_th); | ||
375 | |||
376 | /* if we are already beyond the interrupt threshold, | ||
377 | * we fill the entire buffer */ | ||
378 | if (write_end <= index) | ||
379 | write_end = end; | ||
380 | |||
381 | if (write_end <= index) | ||
382 | break; | ||
383 | |||
384 | write_size = min((unsigned long) size, write_end - index); | ||
385 | memcpy((void *)index, record, write_size); | ||
386 | |||
387 | record = (const char *)record + write_size; | ||
388 | size -= write_size; | ||
389 | bytes_written += write_size; | ||
390 | |||
391 | adj_write_size = write_size / ds_cfg.sizeof_rec[qual]; | ||
392 | adj_write_size *= ds_cfg.sizeof_rec[qual]; | ||
393 | |||
394 | /* zero out trailing bytes */ | ||
395 | memset((char *)index + write_size, 0, | ||
396 | adj_write_size - write_size); | ||
397 | index += adj_write_size; | ||
398 | |||
399 | if (index >= end) | ||
400 | index = base; | ||
401 | ds_set(context->ds, qual, ds_index, index); | ||
402 | |||
403 | if (index >= int_th) | ||
404 | ds_overflow(context, qual); | ||
405 | } | ||
406 | |||
407 | return bytes_written; | ||
408 | } | ||
409 | |||
410 | |||
411 | /* | ||
412 | * Branch Trace Store (BTS) uses the following format. Different | ||
413 | * architectures vary in the size of those fields. | ||
414 | * - source linear address | ||
415 | * - destination linear address | ||
416 | * - flags | ||
417 | * | ||
418 | * Later architectures use 64bit pointers throughout, whereas earlier | ||
419 | * architectures use 32bit pointers in 32bit mode. | ||
420 | * | ||
421 | * We compute the base address for the first 8 fields based on: | ||
422 | * - the field size stored in the DS configuration | ||
423 | * - the relative field position | ||
424 | * | ||
425 | * In order to store additional information in the BTS buffer, we use | ||
426 | * a special source address to indicate that the record requires | ||
427 | * special interpretation. | ||
428 | * | ||
429 | * Netburst indicated via a bit in the flags field whether the branch | ||
430 | * was predicted; this is ignored. | ||
431 | * | ||
432 | * We use two levels of abstraction: | ||
433 | * - the raw data level defined here | ||
434 | * - an arch-independent level defined in ds.h | ||
435 | */ | ||
436 | |||
437 | enum bts_field { | ||
438 | bts_from, | ||
439 | bts_to, | ||
440 | bts_flags, | ||
441 | |||
442 | bts_qual = bts_from, | ||
443 | bts_jiffies = bts_to, | ||
444 | bts_pid = bts_flags, | ||
445 | |||
446 | bts_qual_mask = (bts_qual_max - 1), | ||
447 | bts_escape = ((unsigned long)-1 & ~bts_qual_mask) | ||
448 | }; | ||
449 | |||
450 | static inline unsigned long bts_get(const char *base, enum bts_field field) | ||
451 | { | ||
452 | base += (ds_cfg.sizeof_field * field); | ||
453 | return *(unsigned long *)base; | ||
454 | } | ||
455 | |||
456 | static inline void bts_set(char *base, enum bts_field field, unsigned long val) | ||
457 | { | ||
458 | base += (ds_cfg.sizeof_field * field);; | ||
459 | (*(unsigned long *)base) = val; | ||
460 | } | ||
461 | |||
462 | |||
463 | /* | ||
464 | * The raw BTS data is architecture dependent. | ||
465 | * | ||
466 | * For higher-level users, we give an arch-independent view. | ||
467 | * - ds.h defines struct bts_struct | ||
468 | * - bts_read translates one raw bts record into a bts_struct | ||
469 | * - bts_write translates one bts_struct into the raw format and | ||
470 | * writes it into the top of the parameter tracer's buffer. | ||
471 | * | ||
472 | * return: bytes read/written on success; -Eerrno, otherwise | ||
473 | */ | ||
474 | static int bts_read(struct bts_tracer *tracer, const void *at, | ||
475 | struct bts_struct *out) | ||
476 | { | ||
477 | if (!tracer) | ||
478 | return -EINVAL; | ||
479 | |||
480 | if (at < tracer->trace.ds.begin) | ||
481 | return -EINVAL; | ||
482 | |||
483 | if (tracer->trace.ds.end < (at + tracer->trace.ds.size)) | ||
484 | return -EINVAL; | ||
485 | |||
486 | memset(out, 0, sizeof(*out)); | ||
487 | if ((bts_get(at, bts_qual) & ~bts_qual_mask) == bts_escape) { | ||
488 | out->qualifier = (bts_get(at, bts_qual) & bts_qual_mask); | ||
489 | out->variant.timestamp.jiffies = bts_get(at, bts_jiffies); | ||
490 | out->variant.timestamp.pid = bts_get(at, bts_pid); | ||
491 | } else { | ||
492 | out->qualifier = bts_branch; | ||
493 | out->variant.lbr.from = bts_get(at, bts_from); | ||
494 | out->variant.lbr.to = bts_get(at, bts_to); | ||
495 | } | ||
496 | |||
497 | return ds_cfg.sizeof_rec[ds_bts]; | ||
498 | } | ||
499 | |||
500 | static int bts_write(struct bts_tracer *tracer, const struct bts_struct *in) | ||
501 | { | ||
502 | unsigned char raw[MAX_SIZEOF_BTS]; | ||
503 | |||
504 | if (!tracer) | ||
505 | return -EINVAL; | ||
506 | |||
507 | if (MAX_SIZEOF_BTS < ds_cfg.sizeof_rec[ds_bts]) | ||
508 | return -EOVERFLOW; | ||
509 | |||
510 | switch (in->qualifier) { | ||
511 | case bts_invalid: | ||
512 | bts_set(raw, bts_from, 0); | ||
513 | bts_set(raw, bts_to, 0); | ||
514 | bts_set(raw, bts_flags, 0); | ||
515 | break; | ||
516 | case bts_branch: | ||
517 | bts_set(raw, bts_from, in->variant.lbr.from); | ||
518 | bts_set(raw, bts_to, in->variant.lbr.to); | ||
519 | bts_set(raw, bts_flags, 0); | ||
520 | break; | ||
521 | case bts_task_arrives: | ||
522 | case bts_task_departs: | ||
523 | bts_set(raw, bts_qual, (bts_escape | in->qualifier)); | ||
524 | bts_set(raw, bts_jiffies, in->variant.timestamp.jiffies); | ||
525 | bts_set(raw, bts_pid, in->variant.timestamp.pid); | ||
286 | break; | 526 | break; |
527 | default: | ||
528 | return -EINVAL; | ||
287 | } | 529 | } |
530 | |||
531 | return ds_write(tracer->ds.context, ds_bts, raw, | ||
532 | ds_cfg.sizeof_rec[ds_bts]); | ||
288 | } | 533 | } |
289 | 534 | ||
290 | 535 | ||
291 | static void ds_install_ds_config(struct ds_context *context, | 536 | static void ds_write_config(struct ds_context *context, |
292 | enum ds_qualifier qual, | 537 | struct ds_trace *cfg, enum ds_qualifier qual) |
293 | void *base, size_t size, size_t ith) | 538 | { |
539 | unsigned char *ds = context->ds; | ||
540 | |||
541 | ds_set(ds, qual, ds_buffer_base, (unsigned long)cfg->begin); | ||
542 | ds_set(ds, qual, ds_index, (unsigned long)cfg->top); | ||
543 | ds_set(ds, qual, ds_absolute_maximum, (unsigned long)cfg->end); | ||
544 | ds_set(ds, qual, ds_interrupt_threshold, (unsigned long)cfg->ith); | ||
545 | } | ||
546 | |||
547 | static void ds_read_config(struct ds_context *context, | ||
548 | struct ds_trace *cfg, enum ds_qualifier qual) | ||
294 | { | 549 | { |
550 | unsigned char *ds = context->ds; | ||
551 | |||
552 | cfg->begin = (void *)ds_get(ds, qual, ds_buffer_base); | ||
553 | cfg->top = (void *)ds_get(ds, qual, ds_index); | ||
554 | cfg->end = (void *)ds_get(ds, qual, ds_absolute_maximum); | ||
555 | cfg->ith = (void *)ds_get(ds, qual, ds_interrupt_threshold); | ||
556 | } | ||
557 | |||
558 | static void ds_init_ds_trace(struct ds_trace *trace, enum ds_qualifier qual, | ||
559 | void *base, size_t size, size_t ith, | ||
560 | unsigned int flags) { | ||
295 | unsigned long buffer, adj; | 561 | unsigned long buffer, adj; |
296 | 562 | ||
297 | /* adjust the buffer address and size to meet alignment | 563 | /* adjust the buffer address and size to meet alignment |
@@ -308,32 +574,30 @@ static void ds_install_ds_config(struct ds_context *context, | |||
308 | buffer += adj; | 574 | buffer += adj; |
309 | size -= adj; | 575 | size -= adj; |
310 | 576 | ||
311 | size /= ds_cfg.sizeof_rec[qual]; | 577 | trace->n = size / ds_cfg.sizeof_rec[qual]; |
312 | size *= ds_cfg.sizeof_rec[qual]; | 578 | trace->size = ds_cfg.sizeof_rec[qual]; |
313 | 579 | ||
314 | ds_set(context->ds, qual, ds_buffer_base, buffer); | 580 | size = (trace->n * trace->size); |
315 | ds_set(context->ds, qual, ds_index, buffer); | ||
316 | ds_set(context->ds, qual, ds_absolute_maximum, buffer + size); | ||
317 | 581 | ||
582 | trace->begin = (void *)buffer; | ||
583 | trace->top = trace->begin; | ||
584 | trace->end = (void *)(buffer + size); | ||
318 | /* The value for 'no threshold' is -1, which will set the | 585 | /* The value for 'no threshold' is -1, which will set the |
319 | * threshold outside of the buffer, just like we want it. | 586 | * threshold outside of the buffer, just like we want it. |
320 | */ | 587 | */ |
321 | ds_set(context->ds, qual, | 588 | trace->ith = (void *)(buffer + size - ith); |
322 | ds_interrupt_threshold, buffer + size - ith); | 589 | |
590 | trace->flags = flags; | ||
323 | } | 591 | } |
324 | 592 | ||
325 | static int ds_request(struct ds_tracer *tracer, enum ds_qualifier qual, | 593 | |
326 | struct task_struct *task, | 594 | static int ds_request(struct ds_tracer *tracer, struct ds_trace *trace, |
327 | void *base, size_t size, size_t th) | 595 | enum ds_qualifier qual, struct task_struct *task, |
596 | void *base, size_t size, size_t th, unsigned int flags) | ||
328 | { | 597 | { |
329 | struct ds_context *context; | 598 | struct ds_context *context; |
330 | unsigned long irq; | ||
331 | int error; | 599 | int error; |
332 | 600 | ||
333 | error = -EOPNOTSUPP; | ||
334 | if (!ds_cfg.sizeof_ds) | ||
335 | goto out; | ||
336 | |||
337 | error = -EINVAL; | 601 | error = -EINVAL; |
338 | if (!base) | 602 | if (!base) |
339 | goto out; | 603 | goto out; |
@@ -360,43 +624,26 @@ static int ds_request(struct ds_tracer *tracer, enum ds_qualifier qual, | |||
360 | goto out; | 624 | goto out; |
361 | tracer->context = context; | 625 | tracer->context = context; |
362 | 626 | ||
627 | ds_init_ds_trace(trace, qual, base, size, th, flags); | ||
363 | 628 | ||
364 | spin_lock_irqsave(&ds_lock, irq); | 629 | error = 0; |
365 | |||
366 | error = -EPERM; | ||
367 | if (!check_tracer(task)) | ||
368 | goto out_unlock; | ||
369 | get_tracer(task); | ||
370 | |||
371 | error = -EPERM; | ||
372 | if (context->owner[qual]) | ||
373 | goto out_put_tracer; | ||
374 | context->owner[qual] = tracer; | ||
375 | |||
376 | spin_unlock_irqrestore(&ds_lock, irq); | ||
377 | |||
378 | |||
379 | ds_install_ds_config(context, qual, base, size, th); | ||
380 | |||
381 | return 0; | ||
382 | |||
383 | out_put_tracer: | ||
384 | put_tracer(task); | ||
385 | out_unlock: | ||
386 | spin_unlock_irqrestore(&ds_lock, irq); | ||
387 | ds_put_context(context); | ||
388 | tracer->context = NULL; | ||
389 | out: | 630 | out: |
390 | return error; | 631 | return error; |
391 | } | 632 | } |
392 | 633 | ||
393 | struct bts_tracer *ds_request_bts(struct task_struct *task, | 634 | struct bts_tracer *ds_request_bts(struct task_struct *task, |
394 | void *base, size_t size, | 635 | void *base, size_t size, |
395 | bts_ovfl_callback_t ovfl, size_t th) | 636 | bts_ovfl_callback_t ovfl, size_t th, |
637 | unsigned int flags) | ||
396 | { | 638 | { |
397 | struct bts_tracer *tracer; | 639 | struct bts_tracer *tracer; |
640 | unsigned long irq; | ||
398 | int error; | 641 | int error; |
399 | 642 | ||
643 | error = -EOPNOTSUPP; | ||
644 | if (!ds_cfg.ctl[dsf_bts]) | ||
645 | goto out; | ||
646 | |||
400 | /* buffer overflow notification is not yet implemented */ | 647 | /* buffer overflow notification is not yet implemented */ |
401 | error = -EOPNOTSUPP; | 648 | error = -EOPNOTSUPP; |
402 | if (ovfl) | 649 | if (ovfl) |
@@ -408,12 +655,40 @@ struct bts_tracer *ds_request_bts(struct task_struct *task, | |||
408 | goto out; | 655 | goto out; |
409 | tracer->ovfl = ovfl; | 656 | tracer->ovfl = ovfl; |
410 | 657 | ||
411 | error = ds_request(&tracer->ds, ds_bts, task, base, size, th); | 658 | error = ds_request(&tracer->ds, &tracer->trace.ds, |
659 | ds_bts, task, base, size, th, flags); | ||
412 | if (error < 0) | 660 | if (error < 0) |
413 | goto out_tracer; | 661 | goto out_tracer; |
414 | 662 | ||
663 | |||
664 | spin_lock_irqsave(&ds_lock, irq); | ||
665 | |||
666 | error = -EPERM; | ||
667 | if (!check_tracer(task)) | ||
668 | goto out_unlock; | ||
669 | get_tracer(task); | ||
670 | |||
671 | error = -EPERM; | ||
672 | if (tracer->ds.context->bts_master) | ||
673 | goto out_put_tracer; | ||
674 | tracer->ds.context->bts_master = tracer; | ||
675 | |||
676 | spin_unlock_irqrestore(&ds_lock, irq); | ||
677 | |||
678 | |||
679 | tracer->trace.read = bts_read; | ||
680 | tracer->trace.write = bts_write; | ||
681 | |||
682 | ds_write_config(tracer->ds.context, &tracer->trace.ds, ds_bts); | ||
683 | ds_resume_bts(tracer); | ||
684 | |||
415 | return tracer; | 685 | return tracer; |
416 | 686 | ||
687 | out_put_tracer: | ||
688 | put_tracer(task); | ||
689 | out_unlock: | ||
690 | spin_unlock_irqrestore(&ds_lock, irq); | ||
691 | ds_put_context(tracer->ds.context); | ||
417 | out_tracer: | 692 | out_tracer: |
418 | kfree(tracer); | 693 | kfree(tracer); |
419 | out: | 694 | out: |
@@ -422,9 +697,11 @@ struct bts_tracer *ds_request_bts(struct task_struct *task, | |||
422 | 697 | ||
423 | struct pebs_tracer *ds_request_pebs(struct task_struct *task, | 698 | struct pebs_tracer *ds_request_pebs(struct task_struct *task, |
424 | void *base, size_t size, | 699 | void *base, size_t size, |
425 | pebs_ovfl_callback_t ovfl, size_t th) | 700 | pebs_ovfl_callback_t ovfl, size_t th, |
701 | unsigned int flags) | ||
426 | { | 702 | { |
427 | struct pebs_tracer *tracer; | 703 | struct pebs_tracer *tracer; |
704 | unsigned long irq; | ||
428 | int error; | 705 | int error; |
429 | 706 | ||
430 | /* buffer overflow notification is not yet implemented */ | 707 | /* buffer overflow notification is not yet implemented */ |
@@ -438,300 +715,171 @@ struct pebs_tracer *ds_request_pebs(struct task_struct *task, | |||
438 | goto out; | 715 | goto out; |
439 | tracer->ovfl = ovfl; | 716 | tracer->ovfl = ovfl; |
440 | 717 | ||
441 | error = ds_request(&tracer->ds, ds_pebs, task, base, size, th); | 718 | error = ds_request(&tracer->ds, &tracer->trace.ds, |
719 | ds_pebs, task, base, size, th, flags); | ||
442 | if (error < 0) | 720 | if (error < 0) |
443 | goto out_tracer; | 721 | goto out_tracer; |
444 | 722 | ||
723 | spin_lock_irqsave(&ds_lock, irq); | ||
724 | |||
725 | error = -EPERM; | ||
726 | if (!check_tracer(task)) | ||
727 | goto out_unlock; | ||
728 | get_tracer(task); | ||
729 | |||
730 | error = -EPERM; | ||
731 | if (tracer->ds.context->pebs_master) | ||
732 | goto out_put_tracer; | ||
733 | tracer->ds.context->pebs_master = tracer; | ||
734 | |||
735 | spin_unlock_irqrestore(&ds_lock, irq); | ||
736 | |||
737 | ds_write_config(tracer->ds.context, &tracer->trace.ds, ds_bts); | ||
738 | ds_resume_pebs(tracer); | ||
739 | |||
445 | return tracer; | 740 | return tracer; |
446 | 741 | ||
742 | out_put_tracer: | ||
743 | put_tracer(task); | ||
744 | out_unlock: | ||
745 | spin_unlock_irqrestore(&ds_lock, irq); | ||
746 | ds_put_context(tracer->ds.context); | ||
447 | out_tracer: | 747 | out_tracer: |
448 | kfree(tracer); | 748 | kfree(tracer); |
449 | out: | 749 | out: |
450 | return ERR_PTR(error); | 750 | return ERR_PTR(error); |
451 | } | 751 | } |
452 | 752 | ||
453 | static void ds_release(struct ds_tracer *tracer, enum ds_qualifier qual) | 753 | void ds_release_bts(struct bts_tracer *tracer) |
454 | { | ||
455 | BUG_ON(tracer->context->owner[qual] != tracer); | ||
456 | tracer->context->owner[qual] = NULL; | ||
457 | |||
458 | put_tracer(tracer->context->task); | ||
459 | ds_put_context(tracer->context); | ||
460 | } | ||
461 | |||
462 | int ds_release_bts(struct bts_tracer *tracer) | ||
463 | { | 754 | { |
464 | if (!tracer) | 755 | if (!tracer) |
465 | return -EINVAL; | 756 | return; |
466 | 757 | ||
467 | ds_release(&tracer->ds, ds_bts); | 758 | ds_suspend_bts(tracer); |
468 | kfree(tracer); | ||
469 | 759 | ||
470 | return 0; | 760 | WARN_ON_ONCE(tracer->ds.context->bts_master != tracer); |
471 | } | 761 | tracer->ds.context->bts_master = NULL; |
472 | 762 | ||
473 | int ds_release_pebs(struct pebs_tracer *tracer) | 763 | put_tracer(tracer->ds.context->task); |
474 | { | 764 | ds_put_context(tracer->ds.context); |
475 | if (!tracer) | ||
476 | return -EINVAL; | ||
477 | 765 | ||
478 | ds_release(&tracer->ds, ds_pebs); | ||
479 | kfree(tracer); | 766 | kfree(tracer); |
480 | |||
481 | return 0; | ||
482 | } | ||
483 | |||
484 | static size_t ds_get_index(struct ds_context *context, enum ds_qualifier qual) | ||
485 | { | ||
486 | unsigned long base, index; | ||
487 | |||
488 | base = ds_get(context->ds, qual, ds_buffer_base); | ||
489 | index = ds_get(context->ds, qual, ds_index); | ||
490 | |||
491 | return (index - base) / ds_cfg.sizeof_rec[qual]; | ||
492 | } | 767 | } |
493 | 768 | ||
494 | int ds_get_bts_index(struct bts_tracer *tracer, size_t *pos) | 769 | void ds_suspend_bts(struct bts_tracer *tracer) |
495 | { | 770 | { |
496 | if (!tracer) | 771 | struct task_struct *task; |
497 | return -EINVAL; | ||
498 | 772 | ||
499 | if (!pos) | ||
500 | return -EINVAL; | ||
501 | |||
502 | *pos = ds_get_index(tracer->ds.context, ds_bts); | ||
503 | |||
504 | return 0; | ||
505 | } | ||
506 | |||
507 | int ds_get_pebs_index(struct pebs_tracer *tracer, size_t *pos) | ||
508 | { | ||
509 | if (!tracer) | 773 | if (!tracer) |
510 | return -EINVAL; | 774 | return; |
511 | 775 | ||
512 | if (!pos) | 776 | task = tracer->ds.context->task; |
513 | return -EINVAL; | ||
514 | 777 | ||
515 | *pos = ds_get_index(tracer->ds.context, ds_pebs); | 778 | if (!task || (task == current)) |
779 | update_debugctlmsr(get_debugctlmsr() & ~BTS_CONTROL); | ||
516 | 780 | ||
517 | return 0; | 781 | if (task) { |
518 | } | 782 | task->thread.debugctlmsr &= ~BTS_CONTROL; |
519 | 783 | ||
520 | static size_t ds_get_end(struct ds_context *context, enum ds_qualifier qual) | 784 | if (!task->thread.debugctlmsr) |
521 | { | 785 | clear_tsk_thread_flag(task, TIF_DEBUGCTLMSR); |
522 | unsigned long base, max; | 786 | } |
523 | |||
524 | base = ds_get(context->ds, qual, ds_buffer_base); | ||
525 | max = ds_get(context->ds, qual, ds_absolute_maximum); | ||
526 | |||
527 | return (max - base) / ds_cfg.sizeof_rec[qual]; | ||
528 | } | 787 | } |
529 | 788 | ||
530 | int ds_get_bts_end(struct bts_tracer *tracer, size_t *pos) | 789 | void ds_resume_bts(struct bts_tracer *tracer) |
531 | { | 790 | { |
532 | if (!tracer) | 791 | struct task_struct *task; |
533 | return -EINVAL; | 792 | unsigned long control; |
534 | |||
535 | if (!pos) | ||
536 | return -EINVAL; | ||
537 | |||
538 | *pos = ds_get_end(tracer->ds.context, ds_bts); | ||
539 | |||
540 | return 0; | ||
541 | } | ||
542 | 793 | ||
543 | int ds_get_pebs_end(struct pebs_tracer *tracer, size_t *pos) | ||
544 | { | ||
545 | if (!tracer) | 794 | if (!tracer) |
546 | return -EINVAL; | 795 | return; |
547 | |||
548 | if (!pos) | ||
549 | return -EINVAL; | ||
550 | |||
551 | *pos = ds_get_end(tracer->ds.context, ds_pebs); | ||
552 | |||
553 | return 0; | ||
554 | } | ||
555 | |||
556 | static int ds_access(struct ds_context *context, enum ds_qualifier qual, | ||
557 | size_t index, const void **record) | ||
558 | { | ||
559 | unsigned long base, idx; | ||
560 | |||
561 | if (!record) | ||
562 | return -EINVAL; | ||
563 | |||
564 | base = ds_get(context->ds, qual, ds_buffer_base); | ||
565 | idx = base + (index * ds_cfg.sizeof_rec[qual]); | ||
566 | |||
567 | if (idx > ds_get(context->ds, qual, ds_absolute_maximum)) | ||
568 | return -EINVAL; | ||
569 | 796 | ||
570 | *record = (const void *)idx; | 797 | task = tracer->ds.context->task; |
571 | 798 | ||
572 | return ds_cfg.sizeof_rec[qual]; | 799 | control = ds_cfg.ctl[dsf_bts]; |
573 | } | 800 | if (!(tracer->trace.ds.flags & BTS_KERNEL)) |
801 | control |= ds_cfg.ctl[dsf_bts_kernel]; | ||
802 | if (!(tracer->trace.ds.flags & BTS_USER)) | ||
803 | control |= ds_cfg.ctl[dsf_bts_user]; | ||
574 | 804 | ||
575 | int ds_access_bts(struct bts_tracer *tracer, size_t index, | 805 | if (task) { |
576 | const void **record) | 806 | task->thread.debugctlmsr |= control; |
577 | { | 807 | set_tsk_thread_flag(task, TIF_DEBUGCTLMSR); |
578 | if (!tracer) | 808 | } |
579 | return -EINVAL; | ||
580 | 809 | ||
581 | return ds_access(tracer->ds.context, ds_bts, index, record); | 810 | if (!task || (task == current)) |
811 | update_debugctlmsr(get_debugctlmsr() | control); | ||
582 | } | 812 | } |
583 | 813 | ||
584 | int ds_access_pebs(struct pebs_tracer *tracer, size_t index, | 814 | void ds_release_pebs(struct pebs_tracer *tracer) |
585 | const void **record) | ||
586 | { | 815 | { |
587 | if (!tracer) | 816 | if (!tracer) |
588 | return -EINVAL; | 817 | return; |
589 | |||
590 | return ds_access(tracer->ds.context, ds_pebs, index, record); | ||
591 | } | ||
592 | |||
593 | static int ds_write(struct ds_context *context, enum ds_qualifier qual, | ||
594 | const void *record, size_t size) | ||
595 | { | ||
596 | int bytes_written = 0; | ||
597 | |||
598 | if (!record) | ||
599 | return -EINVAL; | ||
600 | |||
601 | while (size) { | ||
602 | unsigned long base, index, end, write_end, int_th; | ||
603 | unsigned long write_size, adj_write_size; | ||
604 | |||
605 | /* | ||
606 | * write as much as possible without producing an | ||
607 | * overflow interrupt. | ||
608 | * | ||
609 | * interrupt_threshold must either be | ||
610 | * - bigger than absolute_maximum or | ||
611 | * - point to a record between buffer_base and absolute_maximum | ||
612 | * | ||
613 | * index points to a valid record. | ||
614 | */ | ||
615 | base = ds_get(context->ds, qual, ds_buffer_base); | ||
616 | index = ds_get(context->ds, qual, ds_index); | ||
617 | end = ds_get(context->ds, qual, ds_absolute_maximum); | ||
618 | int_th = ds_get(context->ds, qual, ds_interrupt_threshold); | ||
619 | |||
620 | write_end = min(end, int_th); | ||
621 | |||
622 | /* if we are already beyond the interrupt threshold, | ||
623 | * we fill the entire buffer */ | ||
624 | if (write_end <= index) | ||
625 | write_end = end; | ||
626 | |||
627 | if (write_end <= index) | ||
628 | break; | ||
629 | |||
630 | write_size = min((unsigned long) size, write_end - index); | ||
631 | memcpy((void *)index, record, write_size); | ||
632 | |||
633 | record = (const char *)record + write_size; | ||
634 | size -= write_size; | ||
635 | bytes_written += write_size; | ||
636 | |||
637 | adj_write_size = write_size / ds_cfg.sizeof_rec[qual]; | ||
638 | adj_write_size *= ds_cfg.sizeof_rec[qual]; | ||
639 | |||
640 | /* zero out trailing bytes */ | ||
641 | memset((char *)index + write_size, 0, | ||
642 | adj_write_size - write_size); | ||
643 | index += adj_write_size; | ||
644 | 818 | ||
645 | if (index >= end) | 819 | ds_suspend_pebs(tracer); |
646 | index = base; | ||
647 | ds_set(context->ds, qual, ds_index, index); | ||
648 | 820 | ||
649 | if (index >= int_th) | 821 | WARN_ON_ONCE(tracer->ds.context->pebs_master != tracer); |
650 | ds_overflow(context, qual); | 822 | tracer->ds.context->pebs_master = NULL; |
651 | } | ||
652 | 823 | ||
653 | return bytes_written; | 824 | put_tracer(tracer->ds.context->task); |
654 | } | 825 | ds_put_context(tracer->ds.context); |
655 | 826 | ||
656 | int ds_write_bts(struct bts_tracer *tracer, const void *record, size_t size) | 827 | kfree(tracer); |
657 | { | ||
658 | if (!tracer) | ||
659 | return -EINVAL; | ||
660 | |||
661 | return ds_write(tracer->ds.context, ds_bts, record, size); | ||
662 | } | 828 | } |
663 | 829 | ||
664 | int ds_write_pebs(struct pebs_tracer *tracer, const void *record, size_t size) | 830 | void ds_suspend_pebs(struct pebs_tracer *tracer) |
665 | { | 831 | { |
666 | if (!tracer) | ||
667 | return -EINVAL; | ||
668 | 832 | ||
669 | return ds_write(tracer->ds.context, ds_pebs, record, size); | ||
670 | } | 833 | } |
671 | 834 | ||
672 | static void ds_reset_or_clear(struct ds_context *context, | 835 | void ds_resume_pebs(struct pebs_tracer *tracer) |
673 | enum ds_qualifier qual, int clear) | ||
674 | { | 836 | { |
675 | unsigned long base, end; | ||
676 | |||
677 | base = ds_get(context->ds, qual, ds_buffer_base); | ||
678 | end = ds_get(context->ds, qual, ds_absolute_maximum); | ||
679 | |||
680 | if (clear) | ||
681 | memset((void *)base, 0, end - base); | ||
682 | 837 | ||
683 | ds_set(context->ds, qual, ds_index, base); | ||
684 | } | 838 | } |
685 | 839 | ||
686 | int ds_reset_bts(struct bts_tracer *tracer) | 840 | const struct bts_trace *ds_read_bts(struct bts_tracer *tracer) |
687 | { | 841 | { |
688 | if (!tracer) | 842 | if (!tracer) |
689 | return -EINVAL; | 843 | return NULL; |
690 | |||
691 | ds_reset_or_clear(tracer->ds.context, ds_bts, /* clear = */ 0); | ||
692 | 844 | ||
693 | return 0; | 845 | ds_read_config(tracer->ds.context, &tracer->trace.ds, ds_bts); |
846 | return &tracer->trace; | ||
694 | } | 847 | } |
695 | 848 | ||
696 | int ds_reset_pebs(struct pebs_tracer *tracer) | 849 | const struct pebs_trace *ds_read_pebs(struct pebs_tracer *tracer) |
697 | { | 850 | { |
698 | if (!tracer) | 851 | if (!tracer) |
699 | return -EINVAL; | 852 | return NULL; |
700 | 853 | ||
701 | ds_reset_or_clear(tracer->ds.context, ds_pebs, /* clear = */ 0); | 854 | ds_read_config(tracer->ds.context, &tracer->trace.ds, ds_pebs); |
855 | tracer->trace.reset_value = | ||
856 | *(u64 *)(tracer->ds.context->ds + (ds_cfg.sizeof_field * 8)); | ||
702 | 857 | ||
703 | return 0; | 858 | return &tracer->trace; |
704 | } | 859 | } |
705 | 860 | ||
706 | int ds_clear_bts(struct bts_tracer *tracer) | 861 | int ds_reset_bts(struct bts_tracer *tracer) |
707 | { | 862 | { |
708 | if (!tracer) | 863 | if (!tracer) |
709 | return -EINVAL; | 864 | return -EINVAL; |
710 | 865 | ||
711 | ds_reset_or_clear(tracer->ds.context, ds_bts, /* clear = */ 1); | 866 | tracer->trace.ds.top = tracer->trace.ds.begin; |
712 | |||
713 | return 0; | ||
714 | } | ||
715 | |||
716 | int ds_clear_pebs(struct pebs_tracer *tracer) | ||
717 | { | ||
718 | if (!tracer) | ||
719 | return -EINVAL; | ||
720 | 867 | ||
721 | ds_reset_or_clear(tracer->ds.context, ds_pebs, /* clear = */ 1); | 868 | ds_set(tracer->ds.context->ds, ds_bts, ds_index, |
869 | (unsigned long)tracer->trace.ds.top); | ||
722 | 870 | ||
723 | return 0; | 871 | return 0; |
724 | } | 872 | } |
725 | 873 | ||
726 | int ds_get_pebs_reset(struct pebs_tracer *tracer, u64 *value) | 874 | int ds_reset_pebs(struct pebs_tracer *tracer) |
727 | { | 875 | { |
728 | if (!tracer) | 876 | if (!tracer) |
729 | return -EINVAL; | 877 | return -EINVAL; |
730 | 878 | ||
731 | if (!value) | 879 | tracer->trace.ds.top = tracer->trace.ds.begin; |
732 | return -EINVAL; | ||
733 | 880 | ||
734 | *value = *(u64 *)(tracer->ds.context->ds + (ds_cfg.sizeof_field * 8)); | 881 | ds_set(tracer->ds.context->ds, ds_bts, ds_index, |
882 | (unsigned long)tracer->trace.ds.top); | ||
735 | 883 | ||
736 | return 0; | 884 | return 0; |
737 | } | 885 | } |
@@ -746,35 +894,59 @@ int ds_set_pebs_reset(struct pebs_tracer *tracer, u64 value) | |||
746 | return 0; | 894 | return 0; |
747 | } | 895 | } |
748 | 896 | ||
749 | static const struct ds_configuration ds_cfg_var = { | 897 | static const struct ds_configuration ds_cfg_netburst = { |
750 | .sizeof_ds = sizeof(long) * 12, | 898 | .name = "netburst", |
751 | .sizeof_field = sizeof(long), | 899 | .ctl[dsf_bts] = (1 << 2) | (1 << 3), |
752 | .sizeof_rec[ds_bts] = sizeof(long) * 3, | 900 | .ctl[dsf_bts_kernel] = (1 << 5), |
901 | .ctl[dsf_bts_user] = (1 << 6), | ||
902 | |||
903 | .sizeof_field = sizeof(long), | ||
904 | .sizeof_rec[ds_bts] = sizeof(long) * 3, | ||
753 | #ifdef __i386__ | 905 | #ifdef __i386__ |
754 | .sizeof_rec[ds_pebs] = sizeof(long) * 10 | 906 | .sizeof_rec[ds_pebs] = sizeof(long) * 10, |
755 | #else | 907 | #else |
756 | .sizeof_rec[ds_pebs] = sizeof(long) * 18 | 908 | .sizeof_rec[ds_pebs] = sizeof(long) * 18, |
757 | #endif | 909 | #endif |
758 | }; | 910 | }; |
759 | static const struct ds_configuration ds_cfg_64 = { | 911 | static const struct ds_configuration ds_cfg_pentium_m = { |
760 | .sizeof_ds = 8 * 12, | 912 | .name = "pentium m", |
761 | .sizeof_field = 8, | 913 | .ctl[dsf_bts] = (1 << 6) | (1 << 7), |
762 | .sizeof_rec[ds_bts] = 8 * 3, | 914 | |
915 | .sizeof_field = sizeof(long), | ||
916 | .sizeof_rec[ds_bts] = sizeof(long) * 3, | ||
763 | #ifdef __i386__ | 917 | #ifdef __i386__ |
764 | .sizeof_rec[ds_pebs] = 8 * 10 | 918 | .sizeof_rec[ds_pebs] = sizeof(long) * 10, |
765 | #else | 919 | #else |
766 | .sizeof_rec[ds_pebs] = 8 * 18 | 920 | .sizeof_rec[ds_pebs] = sizeof(long) * 18, |
767 | #endif | 921 | #endif |
768 | }; | 922 | }; |
923 | static const struct ds_configuration ds_cfg_core2 = { | ||
924 | .name = "core 2", | ||
925 | .ctl[dsf_bts] = (1 << 6) | (1 << 7), | ||
926 | .ctl[dsf_bts_kernel] = (1 << 9), | ||
927 | .ctl[dsf_bts_user] = (1 << 10), | ||
928 | |||
929 | .sizeof_field = 8, | ||
930 | .sizeof_rec[ds_bts] = 8 * 3, | ||
931 | .sizeof_rec[ds_pebs] = 8 * 18, | ||
932 | }; | ||
769 | 933 | ||
770 | static inline void | 934 | static void |
771 | ds_configure(const struct ds_configuration *cfg) | 935 | ds_configure(const struct ds_configuration *cfg) |
772 | { | 936 | { |
937 | memset(&ds_cfg, 0, sizeof(ds_cfg)); | ||
773 | ds_cfg = *cfg; | 938 | ds_cfg = *cfg; |
774 | 939 | ||
775 | printk(KERN_INFO "DS available\n"); | 940 | printk(KERN_INFO "[ds] using %s configuration\n", ds_cfg.name); |
941 | |||
942 | if (!cpu_has_bts) { | ||
943 | ds_cfg.ctl[dsf_bts] = 0; | ||
944 | printk(KERN_INFO "[ds] bts not available\n"); | ||
945 | } | ||
946 | if (!cpu_has_pebs) | ||
947 | printk(KERN_INFO "[ds] pebs not available\n"); | ||
776 | 948 | ||
777 | BUG_ON(MAX_SIZEOF_DS < ds_cfg.sizeof_ds); | 949 | WARN_ON_ONCE(MAX_SIZEOF_DS < (12 * ds_cfg.sizeof_field)); |
778 | } | 950 | } |
779 | 951 | ||
780 | void __cpuinit ds_init_intel(struct cpuinfo_x86 *c) | 952 | void __cpuinit ds_init_intel(struct cpuinfo_x86 *c) |
@@ -787,10 +959,10 @@ void __cpuinit ds_init_intel(struct cpuinfo_x86 *c) | |||
787 | break; | 959 | break; |
788 | case 0xD: | 960 | case 0xD: |
789 | case 0xE: /* Pentium M */ | 961 | case 0xE: /* Pentium M */ |
790 | ds_configure(&ds_cfg_var); | 962 | ds_configure(&ds_cfg_pentium_m); |
791 | break; | 963 | break; |
792 | default: /* Core2, Atom, ... */ | 964 | default: /* Core2, Atom, ... */ |
793 | ds_configure(&ds_cfg_64); | 965 | ds_configure(&ds_cfg_core2); |
794 | break; | 966 | break; |
795 | } | 967 | } |
796 | break; | 968 | break; |
@@ -799,7 +971,7 @@ void __cpuinit ds_init_intel(struct cpuinfo_x86 *c) | |||
799 | case 0x0: | 971 | case 0x0: |
800 | case 0x1: | 972 | case 0x1: |
801 | case 0x2: /* Netburst */ | 973 | case 0x2: /* Netburst */ |
802 | ds_configure(&ds_cfg_var); | 974 | ds_configure(&ds_cfg_netburst); |
803 | break; | 975 | break; |
804 | default: | 976 | default: |
805 | /* sorry, don't know about them */ | 977 | /* sorry, don't know about them */ |
@@ -812,14 +984,41 @@ void __cpuinit ds_init_intel(struct cpuinfo_x86 *c) | |||
812 | } | 984 | } |
813 | } | 985 | } |
814 | 986 | ||
815 | void ds_free(struct ds_context *context) | 987 | /* |
988 | * Change the DS configuration from tracing prev to tracing next. | ||
989 | */ | ||
990 | void ds_switch_to(struct task_struct *prev, struct task_struct *next) | ||
816 | { | 991 | { |
817 | /* This is called when the task owning the parameter context | 992 | struct ds_context *prev_ctx = prev->thread.ds_ctx; |
818 | * is dying. There should not be any user of that context left | 993 | struct ds_context *next_ctx = next->thread.ds_ctx; |
819 | * to disturb us, anymore. */ | 994 | |
820 | unsigned long leftovers = context->count; | 995 | if (prev_ctx) { |
821 | while (leftovers--) { | 996 | update_debugctlmsr(0); |
822 | put_tracer(context->task); | 997 | |
823 | ds_put_context(context); | 998 | if (prev_ctx->bts_master && |
999 | (prev_ctx->bts_master->trace.ds.flags & BTS_TIMESTAMPS)) { | ||
1000 | struct bts_struct ts = { | ||
1001 | .qualifier = bts_task_departs, | ||
1002 | .variant.timestamp.jiffies = jiffies_64, | ||
1003 | .variant.timestamp.pid = prev->pid | ||
1004 | }; | ||
1005 | bts_write(prev_ctx->bts_master, &ts); | ||
1006 | } | ||
1007 | } | ||
1008 | |||
1009 | if (next_ctx) { | ||
1010 | if (next_ctx->bts_master && | ||
1011 | (next_ctx->bts_master->trace.ds.flags & BTS_TIMESTAMPS)) { | ||
1012 | struct bts_struct ts = { | ||
1013 | .qualifier = bts_task_arrives, | ||
1014 | .variant.timestamp.jiffies = jiffies_64, | ||
1015 | .variant.timestamp.pid = next->pid | ||
1016 | }; | ||
1017 | bts_write(next_ctx->bts_master, &ts); | ||
1018 | } | ||
1019 | |||
1020 | wrmsrl(MSR_IA32_DS_AREA, (unsigned long)next_ctx->ds); | ||
824 | } | 1021 | } |
1022 | |||
1023 | update_debugctlmsr(next->thread.debugctlmsr); | ||
825 | } | 1024 | } |
diff --git a/arch/x86/kernel/process_32.c b/arch/x86/kernel/process_32.c index 24c2276aa453..605eff9a8ac0 100644 --- a/arch/x86/kernel/process_32.c +++ b/arch/x86/kernel/process_32.c | |||
@@ -252,11 +252,14 @@ void exit_thread(void) | |||
252 | put_cpu(); | 252 | put_cpu(); |
253 | } | 253 | } |
254 | #ifdef CONFIG_X86_DS | 254 | #ifdef CONFIG_X86_DS |
255 | /* Free any DS contexts that have not been properly released. */ | 255 | /* Free any BTS tracers that have not been properly released. */ |
256 | if (unlikely(current->thread.ds_ctx)) { | 256 | if (unlikely(current->bts)) { |
257 | /* we clear debugctl to make sure DS is not used. */ | 257 | ds_release_bts(current->bts); |
258 | update_debugctlmsr(0); | 258 | current->bts = NULL; |
259 | ds_free(current->thread.ds_ctx); | 259 | |
260 | kfree(current->bts_buffer); | ||
261 | current->bts_buffer = NULL; | ||
262 | current->bts_size = 0; | ||
260 | } | 263 | } |
261 | #endif /* CONFIG_X86_DS */ | 264 | #endif /* CONFIG_X86_DS */ |
262 | } | 265 | } |
@@ -420,48 +423,19 @@ int set_tsc_mode(unsigned int val) | |||
420 | return 0; | 423 | return 0; |
421 | } | 424 | } |
422 | 425 | ||
423 | #ifdef CONFIG_X86_DS | ||
424 | static int update_debugctl(struct thread_struct *prev, | ||
425 | struct thread_struct *next, unsigned long debugctl) | ||
426 | { | ||
427 | unsigned long ds_prev = 0; | ||
428 | unsigned long ds_next = 0; | ||
429 | |||
430 | if (prev->ds_ctx) | ||
431 | ds_prev = (unsigned long)prev->ds_ctx->ds; | ||
432 | if (next->ds_ctx) | ||
433 | ds_next = (unsigned long)next->ds_ctx->ds; | ||
434 | |||
435 | if (ds_next != ds_prev) { | ||
436 | /* we clear debugctl to make sure DS | ||
437 | * is not in use when we change it */ | ||
438 | debugctl = 0; | ||
439 | update_debugctlmsr(0); | ||
440 | wrmsr(MSR_IA32_DS_AREA, ds_next, 0); | ||
441 | } | ||
442 | return debugctl; | ||
443 | } | ||
444 | #else | ||
445 | static int update_debugctl(struct thread_struct *prev, | ||
446 | struct thread_struct *next, unsigned long debugctl) | ||
447 | { | ||
448 | return debugctl; | ||
449 | } | ||
450 | #endif /* CONFIG_X86_DS */ | ||
451 | |||
452 | static noinline void | 426 | static noinline void |
453 | __switch_to_xtra(struct task_struct *prev_p, struct task_struct *next_p, | 427 | __switch_to_xtra(struct task_struct *prev_p, struct task_struct *next_p, |
454 | struct tss_struct *tss) | 428 | struct tss_struct *tss) |
455 | { | 429 | { |
456 | struct thread_struct *prev, *next; | 430 | struct thread_struct *prev, *next; |
457 | unsigned long debugctl; | ||
458 | 431 | ||
459 | prev = &prev_p->thread; | 432 | prev = &prev_p->thread; |
460 | next = &next_p->thread; | 433 | next = &next_p->thread; |
461 | 434 | ||
462 | debugctl = update_debugctl(prev, next, prev->debugctlmsr); | 435 | if (test_tsk_thread_flag(next_p, TIF_DS_AREA_MSR) || |
463 | 436 | test_tsk_thread_flag(prev_p, TIF_DS_AREA_MSR)) | |
464 | if (next->debugctlmsr != debugctl) | 437 | ds_switch_to(prev_p, next_p); |
438 | else if (next->debugctlmsr != prev->debugctlmsr) | ||
465 | update_debugctlmsr(next->debugctlmsr); | 439 | update_debugctlmsr(next->debugctlmsr); |
466 | 440 | ||
467 | if (test_tsk_thread_flag(next_p, TIF_DEBUG)) { | 441 | if (test_tsk_thread_flag(next_p, TIF_DEBUG)) { |
@@ -483,15 +457,6 @@ __switch_to_xtra(struct task_struct *prev_p, struct task_struct *next_p, | |||
483 | hard_enable_TSC(); | 457 | hard_enable_TSC(); |
484 | } | 458 | } |
485 | 459 | ||
486 | #ifdef CONFIG_X86_PTRACE_BTS | ||
487 | if (test_tsk_thread_flag(prev_p, TIF_BTS_TRACE_TS)) | ||
488 | ptrace_bts_take_timestamp(prev_p, BTS_TASK_DEPARTS); | ||
489 | |||
490 | if (test_tsk_thread_flag(next_p, TIF_BTS_TRACE_TS)) | ||
491 | ptrace_bts_take_timestamp(next_p, BTS_TASK_ARRIVES); | ||
492 | #endif /* CONFIG_X86_PTRACE_BTS */ | ||
493 | |||
494 | |||
495 | if (!test_tsk_thread_flag(next_p, TIF_IO_BITMAP)) { | 460 | if (!test_tsk_thread_flag(next_p, TIF_IO_BITMAP)) { |
496 | /* | 461 | /* |
497 | * Disable the bitmap via an invalid offset. We still cache | 462 | * Disable the bitmap via an invalid offset. We still cache |
diff --git a/arch/x86/kernel/process_64.c b/arch/x86/kernel/process_64.c index fbb321d53d34..1cfd2a4bf853 100644 --- a/arch/x86/kernel/process_64.c +++ b/arch/x86/kernel/process_64.c | |||
@@ -237,11 +237,14 @@ void exit_thread(void) | |||
237 | put_cpu(); | 237 | put_cpu(); |
238 | } | 238 | } |
239 | #ifdef CONFIG_X86_DS | 239 | #ifdef CONFIG_X86_DS |
240 | /* Free any DS contexts that have not been properly released. */ | 240 | /* Free any BTS tracers that have not been properly released. */ |
241 | if (unlikely(t->ds_ctx)) { | 241 | if (unlikely(current->bts)) { |
242 | /* we clear debugctl to make sure DS is not used. */ | 242 | ds_release_bts(current->bts); |
243 | update_debugctlmsr(0); | 243 | current->bts = NULL; |
244 | ds_free(t->ds_ctx); | 244 | |
245 | kfree(current->bts_buffer); | ||
246 | current->bts_buffer = NULL; | ||
247 | current->bts_size = 0; | ||
245 | } | 248 | } |
246 | #endif /* CONFIG_X86_DS */ | 249 | #endif /* CONFIG_X86_DS */ |
247 | } | 250 | } |
@@ -471,35 +474,14 @@ static inline void __switch_to_xtra(struct task_struct *prev_p, | |||
471 | struct tss_struct *tss) | 474 | struct tss_struct *tss) |
472 | { | 475 | { |
473 | struct thread_struct *prev, *next; | 476 | struct thread_struct *prev, *next; |
474 | unsigned long debugctl; | ||
475 | 477 | ||
476 | prev = &prev_p->thread, | 478 | prev = &prev_p->thread, |
477 | next = &next_p->thread; | 479 | next = &next_p->thread; |
478 | 480 | ||
479 | debugctl = prev->debugctlmsr; | 481 | if (test_tsk_thread_flag(next_p, TIF_DS_AREA_MSR) || |
480 | 482 | test_tsk_thread_flag(prev_p, TIF_DS_AREA_MSR)) | |
481 | #ifdef CONFIG_X86_DS | 483 | ds_switch_to(prev_p, next_p); |
482 | { | 484 | else if (next->debugctlmsr != prev->debugctlmsr) |
483 | unsigned long ds_prev = 0, ds_next = 0; | ||
484 | |||
485 | if (prev->ds_ctx) | ||
486 | ds_prev = (unsigned long)prev->ds_ctx->ds; | ||
487 | if (next->ds_ctx) | ||
488 | ds_next = (unsigned long)next->ds_ctx->ds; | ||
489 | |||
490 | if (ds_next != ds_prev) { | ||
491 | /* | ||
492 | * We clear debugctl to make sure DS | ||
493 | * is not in use when we change it: | ||
494 | */ | ||
495 | debugctl = 0; | ||
496 | update_debugctlmsr(0); | ||
497 | wrmsrl(MSR_IA32_DS_AREA, ds_next); | ||
498 | } | ||
499 | } | ||
500 | #endif /* CONFIG_X86_DS */ | ||
501 | |||
502 | if (next->debugctlmsr != debugctl) | ||
503 | update_debugctlmsr(next->debugctlmsr); | 485 | update_debugctlmsr(next->debugctlmsr); |
504 | 486 | ||
505 | if (test_tsk_thread_flag(next_p, TIF_DEBUG)) { | 487 | if (test_tsk_thread_flag(next_p, TIF_DEBUG)) { |
@@ -534,14 +516,6 @@ static inline void __switch_to_xtra(struct task_struct *prev_p, | |||
534 | */ | 516 | */ |
535 | memset(tss->io_bitmap, 0xff, prev->io_bitmap_max); | 517 | memset(tss->io_bitmap, 0xff, prev->io_bitmap_max); |
536 | } | 518 | } |
537 | |||
538 | #ifdef CONFIG_X86_PTRACE_BTS | ||
539 | if (test_tsk_thread_flag(prev_p, TIF_BTS_TRACE_TS)) | ||
540 | ptrace_bts_take_timestamp(prev_p, BTS_TASK_DEPARTS); | ||
541 | |||
542 | if (test_tsk_thread_flag(next_p, TIF_BTS_TRACE_TS)) | ||
543 | ptrace_bts_take_timestamp(next_p, BTS_TASK_ARRIVES); | ||
544 | #endif /* CONFIG_X86_PTRACE_BTS */ | ||
545 | } | 519 | } |
546 | 520 | ||
547 | /* | 521 | /* |
diff --git a/arch/x86/kernel/ptrace.c b/arch/x86/kernel/ptrace.c index 2c8ec1ba75e6..45e9855da2d2 100644 --- a/arch/x86/kernel/ptrace.c +++ b/arch/x86/kernel/ptrace.c | |||
@@ -581,153 +581,73 @@ static int ioperm_get(struct task_struct *target, | |||
581 | } | 581 | } |
582 | 582 | ||
583 | #ifdef CONFIG_X86_PTRACE_BTS | 583 | #ifdef CONFIG_X86_PTRACE_BTS |
584 | /* | ||
585 | * The configuration for a particular BTS hardware implementation. | ||
586 | */ | ||
587 | struct bts_configuration { | ||
588 | /* the size of a BTS record in bytes; at most BTS_MAX_RECORD_SIZE */ | ||
589 | unsigned char sizeof_bts; | ||
590 | /* the size of a field in the BTS record in bytes */ | ||
591 | unsigned char sizeof_field; | ||
592 | /* a bitmask to enable/disable BTS in DEBUGCTL MSR */ | ||
593 | unsigned long debugctl_mask; | ||
594 | }; | ||
595 | static struct bts_configuration bts_cfg; | ||
596 | |||
597 | #define BTS_MAX_RECORD_SIZE (8 * 3) | ||
598 | |||
599 | |||
600 | /* | ||
601 | * Branch Trace Store (BTS) uses the following format. Different | ||
602 | * architectures vary in the size of those fields. | ||
603 | * - source linear address | ||
604 | * - destination linear address | ||
605 | * - flags | ||
606 | * | ||
607 | * Later architectures use 64bit pointers throughout, whereas earlier | ||
608 | * architectures use 32bit pointers in 32bit mode. | ||
609 | * | ||
610 | * We compute the base address for the first 8 fields based on: | ||
611 | * - the field size stored in the DS configuration | ||
612 | * - the relative field position | ||
613 | * | ||
614 | * In order to store additional information in the BTS buffer, we use | ||
615 | * a special source address to indicate that the record requires | ||
616 | * special interpretation. | ||
617 | * | ||
618 | * Netburst indicated via a bit in the flags field whether the branch | ||
619 | * was predicted; this is ignored. | ||
620 | */ | ||
621 | |||
622 | enum bts_field { | ||
623 | bts_from = 0, | ||
624 | bts_to, | ||
625 | bts_flags, | ||
626 | |||
627 | bts_escape = (unsigned long)-1, | ||
628 | bts_qual = bts_to, | ||
629 | bts_jiffies = bts_flags | ||
630 | }; | ||
631 | |||
632 | static inline unsigned long bts_get(const char *base, enum bts_field field) | ||
633 | { | ||
634 | base += (bts_cfg.sizeof_field * field); | ||
635 | return *(unsigned long *)base; | ||
636 | } | ||
637 | |||
638 | static inline void bts_set(char *base, enum bts_field field, unsigned long val) | ||
639 | { | ||
640 | base += (bts_cfg.sizeof_field * field);; | ||
641 | (*(unsigned long *)base) = val; | ||
642 | } | ||
643 | |||
644 | /* | ||
645 | * Translate a BTS record from the raw format into the bts_struct format | ||
646 | * | ||
647 | * out (out): bts_struct interpretation | ||
648 | * raw: raw BTS record | ||
649 | */ | ||
650 | static void ptrace_bts_translate_record(struct bts_struct *out, const void *raw) | ||
651 | { | ||
652 | memset(out, 0, sizeof(*out)); | ||
653 | if (bts_get(raw, bts_from) == bts_escape) { | ||
654 | out->qualifier = bts_get(raw, bts_qual); | ||
655 | out->variant.jiffies = bts_get(raw, bts_jiffies); | ||
656 | } else { | ||
657 | out->qualifier = BTS_BRANCH; | ||
658 | out->variant.lbr.from_ip = bts_get(raw, bts_from); | ||
659 | out->variant.lbr.to_ip = bts_get(raw, bts_to); | ||
660 | } | ||
661 | } | ||
662 | |||
663 | static int ptrace_bts_read_record(struct task_struct *child, size_t index, | 584 | static int ptrace_bts_read_record(struct task_struct *child, size_t index, |
664 | struct bts_struct __user *out) | 585 | struct bts_struct __user *out) |
665 | { | 586 | { |
666 | struct bts_struct ret; | 587 | const struct bts_trace *trace; |
667 | const void *bts_record; | 588 | struct bts_struct bts; |
668 | size_t bts_index, bts_end; | 589 | const unsigned char *at; |
669 | int error; | 590 | int error; |
670 | 591 | ||
671 | error = ds_get_bts_end(child->bts, &bts_end); | 592 | trace = ds_read_bts(child->bts); |
672 | if (error < 0) | 593 | if (!trace) |
673 | return error; | 594 | return -EPERM; |
674 | |||
675 | if (bts_end <= index) | ||
676 | return -EINVAL; | ||
677 | 595 | ||
678 | error = ds_get_bts_index(child->bts, &bts_index); | 596 | at = trace->ds.top - ((index + 1) * trace->ds.size); |
679 | if (error < 0) | 597 | if ((void *)at < trace->ds.begin) |
680 | return error; | 598 | at += (trace->ds.n * trace->ds.size); |
681 | 599 | ||
682 | /* translate the ptrace bts index into the ds bts index */ | 600 | if (!trace->read) |
683 | bts_index += bts_end - (index + 1); | 601 | return -EOPNOTSUPP; |
684 | if (bts_end <= bts_index) | ||
685 | bts_index -= bts_end; | ||
686 | 602 | ||
687 | error = ds_access_bts(child->bts, bts_index, &bts_record); | 603 | error = trace->read(child->bts, at, &bts); |
688 | if (error < 0) | 604 | if (error < 0) |
689 | return error; | 605 | return error; |
690 | 606 | ||
691 | ptrace_bts_translate_record(&ret, bts_record); | 607 | if (copy_to_user(out, &bts, sizeof(bts))) |
692 | |||
693 | if (copy_to_user(out, &ret, sizeof(ret))) | ||
694 | return -EFAULT; | 608 | return -EFAULT; |
695 | 609 | ||
696 | return sizeof(ret); | 610 | return sizeof(bts); |
697 | } | 611 | } |
698 | 612 | ||
699 | static int ptrace_bts_drain(struct task_struct *child, | 613 | static int ptrace_bts_drain(struct task_struct *child, |
700 | long size, | 614 | long size, |
701 | struct bts_struct __user *out) | 615 | struct bts_struct __user *out) |
702 | { | 616 | { |
703 | struct bts_struct ret; | 617 | const struct bts_trace *trace; |
704 | const unsigned char *raw; | 618 | const unsigned char *at; |
705 | size_t end, i; | 619 | int error, drained = 0; |
706 | int error; | ||
707 | 620 | ||
708 | error = ds_get_bts_index(child->bts, &end); | 621 | trace = ds_read_bts(child->bts); |
709 | if (error < 0) | 622 | if (!trace) |
710 | return error; | 623 | return -EPERM; |
711 | 624 | ||
712 | if (size < (end * sizeof(struct bts_struct))) | 625 | if (!trace->read) |
626 | return -EOPNOTSUPP; | ||
627 | |||
628 | if (size < (trace->ds.top - trace->ds.begin)) | ||
713 | return -EIO; | 629 | return -EIO; |
714 | 630 | ||
715 | error = ds_access_bts(child->bts, 0, (const void **)&raw); | 631 | for (at = trace->ds.begin; (void *)at < trace->ds.top; |
716 | if (error < 0) | 632 | out++, drained++, at += trace->ds.size) { |
717 | return error; | 633 | struct bts_struct bts; |
634 | int error; | ||
718 | 635 | ||
719 | for (i = 0; i < end; i++, out++, raw += bts_cfg.sizeof_bts) { | 636 | error = trace->read(child->bts, at, &bts); |
720 | ptrace_bts_translate_record(&ret, raw); | 637 | if (error < 0) |
638 | return error; | ||
721 | 639 | ||
722 | if (copy_to_user(out, &ret, sizeof(ret))) | 640 | if (copy_to_user(out, &bts, sizeof(bts))) |
723 | return -EFAULT; | 641 | return -EFAULT; |
724 | } | 642 | } |
725 | 643 | ||
726 | error = ds_clear_bts(child->bts); | 644 | memset(trace->ds.begin, 0, trace->ds.n * trace->ds.size); |
645 | |||
646 | error = ds_reset_bts(child->bts); | ||
727 | if (error < 0) | 647 | if (error < 0) |
728 | return error; | 648 | return error; |
729 | 649 | ||
730 | return end; | 650 | return drained; |
731 | } | 651 | } |
732 | 652 | ||
733 | static int ptrace_bts_config(struct task_struct *child, | 653 | static int ptrace_bts_config(struct task_struct *child, |
@@ -735,136 +655,89 @@ static int ptrace_bts_config(struct task_struct *child, | |||
735 | const struct ptrace_bts_config __user *ucfg) | 655 | const struct ptrace_bts_config __user *ucfg) |
736 | { | 656 | { |
737 | struct ptrace_bts_config cfg; | 657 | struct ptrace_bts_config cfg; |
738 | int error = 0; | 658 | unsigned int flags = 0; |
739 | |||
740 | error = -EOPNOTSUPP; | ||
741 | if (!bts_cfg.sizeof_bts) | ||
742 | goto errout; | ||
743 | 659 | ||
744 | error = -EIO; | ||
745 | if (cfg_size < sizeof(cfg)) | 660 | if (cfg_size < sizeof(cfg)) |
746 | goto errout; | 661 | return -EIO; |
747 | 662 | ||
748 | error = -EFAULT; | ||
749 | if (copy_from_user(&cfg, ucfg, sizeof(cfg))) | 663 | if (copy_from_user(&cfg, ucfg, sizeof(cfg))) |
750 | goto errout; | 664 | return -EFAULT; |
751 | |||
752 | error = -EINVAL; | ||
753 | if ((cfg.flags & PTRACE_BTS_O_SIGNAL) && | ||
754 | !(cfg.flags & PTRACE_BTS_O_ALLOC)) | ||
755 | goto errout; | ||
756 | |||
757 | if (cfg.flags & PTRACE_BTS_O_ALLOC) { | ||
758 | bts_ovfl_callback_t ovfl = NULL; | ||
759 | unsigned int sig = 0; | ||
760 | |||
761 | error = -EINVAL; | ||
762 | if (cfg.size < (10 * bts_cfg.sizeof_bts)) | ||
763 | goto errout; | ||
764 | 665 | ||
765 | if (cfg.flags & PTRACE_BTS_O_SIGNAL) { | 666 | if (child->bts) { |
766 | if (!cfg.signal) | 667 | ds_release_bts(child->bts); |
767 | goto errout; | 668 | child->bts = NULL; |
669 | } | ||
768 | 670 | ||
769 | error = -EOPNOTSUPP; | 671 | if (cfg.flags & PTRACE_BTS_O_SIGNAL) { |
770 | goto errout; | 672 | if (!cfg.signal) |
673 | return -EINVAL; | ||
771 | 674 | ||
772 | sig = cfg.signal; | 675 | return -EOPNOTSUPP; |
773 | } | ||
774 | 676 | ||
775 | if (child->bts) { | 677 | child->thread.bts_ovfl_signal = cfg.signal; |
776 | (void)ds_release_bts(child->bts); | 678 | } |
777 | kfree(child->bts_buffer); | ||
778 | 679 | ||
779 | child->bts = NULL; | 680 | if ((cfg.flags & PTRACE_BTS_O_ALLOC) && |
780 | child->bts_buffer = NULL; | 681 | (cfg.size != child->bts_size)) { |
781 | } | 682 | kfree(child->bts_buffer); |
782 | 683 | ||
783 | error = -ENOMEM; | 684 | child->bts_size = cfg.size; |
784 | child->bts_buffer = kzalloc(cfg.size, GFP_KERNEL); | 685 | child->bts_buffer = kzalloc(cfg.size, GFP_KERNEL); |
785 | if (!child->bts_buffer) | 686 | if (!child->bts_buffer) { |
786 | goto errout; | 687 | child->bts_size = 0; |
787 | 688 | return -ENOMEM; | |
788 | child->bts = ds_request_bts(child, child->bts_buffer, cfg.size, | ||
789 | ovfl, /* th = */ (size_t)-1); | ||
790 | if (IS_ERR(child->bts)) { | ||
791 | error = PTR_ERR(child->bts); | ||
792 | kfree(child->bts_buffer); | ||
793 | child->bts = NULL; | ||
794 | child->bts_buffer = NULL; | ||
795 | goto errout; | ||
796 | } | 689 | } |
797 | |||
798 | child->thread.bts_ovfl_signal = sig; | ||
799 | } | 690 | } |
800 | 691 | ||
801 | error = -EINVAL; | ||
802 | if (!child->thread.ds_ctx && cfg.flags) | ||
803 | goto errout; | ||
804 | |||
805 | if (cfg.flags & PTRACE_BTS_O_TRACE) | 692 | if (cfg.flags & PTRACE_BTS_O_TRACE) |
806 | child->thread.debugctlmsr |= bts_cfg.debugctl_mask; | 693 | flags |= BTS_USER; |
807 | else | ||
808 | child->thread.debugctlmsr &= ~bts_cfg.debugctl_mask; | ||
809 | 694 | ||
810 | if (cfg.flags & PTRACE_BTS_O_SCHED) | 695 | if (cfg.flags & PTRACE_BTS_O_SCHED) |
811 | set_tsk_thread_flag(child, TIF_BTS_TRACE_TS); | 696 | flags |= BTS_TIMESTAMPS; |
812 | else | ||
813 | clear_tsk_thread_flag(child, TIF_BTS_TRACE_TS); | ||
814 | 697 | ||
815 | error = sizeof(cfg); | 698 | child->bts = ds_request_bts(child, child->bts_buffer, child->bts_size, |
699 | /* ovfl = */ NULL, /* th = */ (size_t)-1, | ||
700 | flags); | ||
701 | if (IS_ERR(child->bts)) { | ||
702 | int error = PTR_ERR(child->bts); | ||
816 | 703 | ||
817 | out: | 704 | kfree(child->bts_buffer); |
818 | if (child->thread.debugctlmsr) | 705 | child->bts = NULL; |
819 | set_tsk_thread_flag(child, TIF_DEBUGCTLMSR); | 706 | child->bts_buffer = NULL; |
820 | else | 707 | child->bts_size = 0; |
821 | clear_tsk_thread_flag(child, TIF_DEBUGCTLMSR); | ||
822 | 708 | ||
823 | return error; | 709 | return error; |
710 | } | ||
824 | 711 | ||
825 | errout: | 712 | return sizeof(cfg); |
826 | child->thread.debugctlmsr &= ~bts_cfg.debugctl_mask; | ||
827 | clear_tsk_thread_flag(child, TIF_BTS_TRACE_TS); | ||
828 | goto out; | ||
829 | } | 713 | } |
830 | 714 | ||
831 | static int ptrace_bts_status(struct task_struct *child, | 715 | static int ptrace_bts_status(struct task_struct *child, |
832 | long cfg_size, | 716 | long cfg_size, |
833 | struct ptrace_bts_config __user *ucfg) | 717 | struct ptrace_bts_config __user *ucfg) |
834 | { | 718 | { |
719 | const struct bts_trace *trace; | ||
835 | struct ptrace_bts_config cfg; | 720 | struct ptrace_bts_config cfg; |
836 | size_t end; | ||
837 | const void *base, *max; | ||
838 | int error; | ||
839 | 721 | ||
840 | if (cfg_size < sizeof(cfg)) | 722 | if (cfg_size < sizeof(cfg)) |
841 | return -EIO; | 723 | return -EIO; |
842 | 724 | ||
843 | error = ds_get_bts_end(child->bts, &end); | 725 | trace = ds_read_bts(child->bts); |
844 | if (error < 0) | 726 | if (!trace) |
845 | return error; | 727 | return -EPERM; |
846 | |||
847 | error = ds_access_bts(child->bts, /* index = */ 0, &base); | ||
848 | if (error < 0) | ||
849 | return error; | ||
850 | |||
851 | error = ds_access_bts(child->bts, /* index = */ end, &max); | ||
852 | if (error < 0) | ||
853 | return error; | ||
854 | 728 | ||
855 | memset(&cfg, 0, sizeof(cfg)); | 729 | memset(&cfg, 0, sizeof(cfg)); |
856 | cfg.size = (max - base); | 730 | cfg.size = trace->ds.end - trace->ds.begin; |
857 | cfg.signal = child->thread.bts_ovfl_signal; | 731 | cfg.signal = child->thread.bts_ovfl_signal; |
858 | cfg.bts_size = sizeof(struct bts_struct); | 732 | cfg.bts_size = sizeof(struct bts_struct); |
859 | 733 | ||
860 | if (cfg.signal) | 734 | if (cfg.signal) |
861 | cfg.flags |= PTRACE_BTS_O_SIGNAL; | 735 | cfg.flags |= PTRACE_BTS_O_SIGNAL; |
862 | 736 | ||
863 | if (test_tsk_thread_flag(child, TIF_DEBUGCTLMSR) && | 737 | if (trace->ds.flags & BTS_USER) |
864 | child->thread.debugctlmsr & bts_cfg.debugctl_mask) | ||
865 | cfg.flags |= PTRACE_BTS_O_TRACE; | 738 | cfg.flags |= PTRACE_BTS_O_TRACE; |
866 | 739 | ||
867 | if (test_tsk_thread_flag(child, TIF_BTS_TRACE_TS)) | 740 | if (trace->ds.flags & BTS_TIMESTAMPS) |
868 | cfg.flags |= PTRACE_BTS_O_SCHED; | 741 | cfg.flags |= PTRACE_BTS_O_SCHED; |
869 | 742 | ||
870 | if (copy_to_user(ucfg, &cfg, sizeof(cfg))) | 743 | if (copy_to_user(ucfg, &cfg, sizeof(cfg))) |
@@ -873,104 +746,28 @@ static int ptrace_bts_status(struct task_struct *child, | |||
873 | return sizeof(cfg); | 746 | return sizeof(cfg); |
874 | } | 747 | } |
875 | 748 | ||
876 | static int ptrace_bts_write_record(struct task_struct *child, | 749 | static int ptrace_bts_clear(struct task_struct *child) |
877 | const struct bts_struct *in) | ||
878 | { | 750 | { |
879 | unsigned char bts_record[BTS_MAX_RECORD_SIZE]; | 751 | const struct bts_trace *trace; |
880 | 752 | ||
881 | BUG_ON(BTS_MAX_RECORD_SIZE < bts_cfg.sizeof_bts); | 753 | trace = ds_read_bts(child->bts); |
754 | if (!trace) | ||
755 | return -EPERM; | ||
882 | 756 | ||
883 | memset(bts_record, 0, bts_cfg.sizeof_bts); | 757 | memset(trace->ds.begin, 0, trace->ds.n * trace->ds.size); |
884 | switch (in->qualifier) { | ||
885 | case BTS_INVALID: | ||
886 | break; | ||
887 | 758 | ||
888 | case BTS_BRANCH: | 759 | return ds_reset_bts(child->bts); |
889 | bts_set(bts_record, bts_from, in->variant.lbr.from_ip); | ||
890 | bts_set(bts_record, bts_to, in->variant.lbr.to_ip); | ||
891 | break; | ||
892 | |||
893 | case BTS_TASK_ARRIVES: | ||
894 | case BTS_TASK_DEPARTS: | ||
895 | bts_set(bts_record, bts_from, bts_escape); | ||
896 | bts_set(bts_record, bts_qual, in->qualifier); | ||
897 | bts_set(bts_record, bts_jiffies, in->variant.jiffies); | ||
898 | break; | ||
899 | |||
900 | default: | ||
901 | return -EINVAL; | ||
902 | } | ||
903 | |||
904 | return ds_write_bts(child->bts, bts_record, bts_cfg.sizeof_bts); | ||
905 | } | 760 | } |
906 | 761 | ||
907 | void ptrace_bts_take_timestamp(struct task_struct *tsk, | 762 | static int ptrace_bts_size(struct task_struct *child) |
908 | enum bts_qualifier qualifier) | ||
909 | { | 763 | { |
910 | struct bts_struct rec = { | 764 | const struct bts_trace *trace; |
911 | .qualifier = qualifier, | ||
912 | .variant.jiffies = jiffies_64 | ||
913 | }; | ||
914 | |||
915 | ptrace_bts_write_record(tsk, &rec); | ||
916 | } | ||
917 | |||
918 | static const struct bts_configuration bts_cfg_netburst = { | ||
919 | .sizeof_bts = sizeof(long) * 3, | ||
920 | .sizeof_field = sizeof(long), | ||
921 | .debugctl_mask = (1<<2)|(1<<3)|(1<<5) | ||
922 | }; | ||
923 | 765 | ||
924 | static const struct bts_configuration bts_cfg_pentium_m = { | 766 | trace = ds_read_bts(child->bts); |
925 | .sizeof_bts = sizeof(long) * 3, | 767 | if (!trace) |
926 | .sizeof_field = sizeof(long), | 768 | return -EPERM; |
927 | .debugctl_mask = (1<<6)|(1<<7) | ||
928 | }; | ||
929 | 769 | ||
930 | static const struct bts_configuration bts_cfg_core2 = { | 770 | return (trace->ds.top - trace->ds.begin) / trace->ds.size; |
931 | .sizeof_bts = 8 * 3, | ||
932 | .sizeof_field = 8, | ||
933 | .debugctl_mask = (1<<6)|(1<<7)|(1<<9) | ||
934 | }; | ||
935 | |||
936 | static inline void bts_configure(const struct bts_configuration *cfg) | ||
937 | { | ||
938 | bts_cfg = *cfg; | ||
939 | } | ||
940 | |||
941 | void __cpuinit ptrace_bts_init_intel(struct cpuinfo_x86 *c) | ||
942 | { | ||
943 | switch (c->x86) { | ||
944 | case 0x6: | ||
945 | switch (c->x86_model) { | ||
946 | case 0 ... 0xC: | ||
947 | /* sorry, don't know about them */ | ||
948 | break; | ||
949 | case 0xD: | ||
950 | case 0xE: /* Pentium M */ | ||
951 | bts_configure(&bts_cfg_pentium_m); | ||
952 | break; | ||
953 | default: /* Core2, Atom, ... */ | ||
954 | bts_configure(&bts_cfg_core2); | ||
955 | break; | ||
956 | } | ||
957 | break; | ||
958 | case 0xF: | ||
959 | switch (c->x86_model) { | ||
960 | case 0x0: | ||
961 | case 0x1: | ||
962 | case 0x2: /* Netburst */ | ||
963 | bts_configure(&bts_cfg_netburst); | ||
964 | break; | ||
965 | default: | ||
966 | /* sorry, don't know about them */ | ||
967 | break; | ||
968 | } | ||
969 | break; | ||
970 | default: | ||
971 | /* sorry, don't know about them */ | ||
972 | break; | ||
973 | } | ||
974 | } | 771 | } |
975 | #endif /* CONFIG_X86_PTRACE_BTS */ | 772 | #endif /* CONFIG_X86_PTRACE_BTS */ |
976 | 773 | ||
@@ -987,15 +784,12 @@ void ptrace_disable(struct task_struct *child) | |||
987 | #endif | 784 | #endif |
988 | #ifdef CONFIG_X86_PTRACE_BTS | 785 | #ifdef CONFIG_X86_PTRACE_BTS |
989 | if (child->bts) { | 786 | if (child->bts) { |
990 | (void)ds_release_bts(child->bts); | 787 | ds_release_bts(child->bts); |
788 | child->bts = NULL; | ||
789 | |||
991 | kfree(child->bts_buffer); | 790 | kfree(child->bts_buffer); |
992 | child->bts_buffer = NULL; | 791 | child->bts_buffer = NULL; |
993 | 792 | child->bts_size = 0; | |
994 | child->thread.debugctlmsr &= ~bts_cfg.debugctl_mask; | ||
995 | if (!child->thread.debugctlmsr) | ||
996 | clear_tsk_thread_flag(child, TIF_DEBUGCTLMSR); | ||
997 | |||
998 | clear_tsk_thread_flag(child, TIF_BTS_TRACE_TS); | ||
999 | } | 793 | } |
1000 | #endif /* CONFIG_X86_PTRACE_BTS */ | 794 | #endif /* CONFIG_X86_PTRACE_BTS */ |
1001 | } | 795 | } |
@@ -1128,16 +922,9 @@ long arch_ptrace(struct task_struct *child, long request, long addr, long data) | |||
1128 | (child, data, (struct ptrace_bts_config __user *)addr); | 922 | (child, data, (struct ptrace_bts_config __user *)addr); |
1129 | break; | 923 | break; |
1130 | 924 | ||
1131 | case PTRACE_BTS_SIZE: { | 925 | case PTRACE_BTS_SIZE: |
1132 | size_t size; | 926 | ret = ptrace_bts_size(child); |
1133 | |||
1134 | ret = ds_get_bts_index(child->bts, &size); | ||
1135 | if (ret == 0) { | ||
1136 | BUG_ON(size != (int) size); | ||
1137 | ret = (int) size; | ||
1138 | } | ||
1139 | break; | 927 | break; |
1140 | } | ||
1141 | 928 | ||
1142 | case PTRACE_BTS_GET: | 929 | case PTRACE_BTS_GET: |
1143 | ret = ptrace_bts_read_record | 930 | ret = ptrace_bts_read_record |
@@ -1145,7 +932,7 @@ long arch_ptrace(struct task_struct *child, long request, long addr, long data) | |||
1145 | break; | 932 | break; |
1146 | 933 | ||
1147 | case PTRACE_BTS_CLEAR: | 934 | case PTRACE_BTS_CLEAR: |
1148 | ret = ds_clear_bts(child->bts); | 935 | ret = ptrace_bts_clear(child); |
1149 | break; | 936 | break; |
1150 | 937 | ||
1151 | case PTRACE_BTS_DRAIN: | 938 | case PTRACE_BTS_DRAIN: |
@@ -1408,6 +1195,14 @@ long compat_arch_ptrace(struct task_struct *child, compat_long_t request, | |||
1408 | 1195 | ||
1409 | case PTRACE_GET_THREAD_AREA: | 1196 | case PTRACE_GET_THREAD_AREA: |
1410 | case PTRACE_SET_THREAD_AREA: | 1197 | case PTRACE_SET_THREAD_AREA: |
1198 | #ifdef CONFIG_X86_PTRACE_BTS | ||
1199 | case PTRACE_BTS_CONFIG: | ||
1200 | case PTRACE_BTS_STATUS: | ||
1201 | case PTRACE_BTS_SIZE: | ||
1202 | case PTRACE_BTS_GET: | ||
1203 | case PTRACE_BTS_CLEAR: | ||
1204 | case PTRACE_BTS_DRAIN: | ||
1205 | #endif /* CONFIG_X86_PTRACE_BTS */ | ||
1411 | return arch_ptrace(child, request, addr, data); | 1206 | return arch_ptrace(child, request, addr, data); |
1412 | 1207 | ||
1413 | default: | 1208 | default: |
diff --git a/include/linux/sched.h b/include/linux/sched.h index 4b81fc5f7731..dc5ea65dc716 100644 --- a/include/linux/sched.h +++ b/include/linux/sched.h | |||
@@ -1176,6 +1176,7 @@ struct task_struct { | |||
1176 | * The buffer to hold the BTS data. | 1176 | * The buffer to hold the BTS data. |
1177 | */ | 1177 | */ |
1178 | void *bts_buffer; | 1178 | void *bts_buffer; |
1179 | size_t bts_size; | ||
1179 | #endif /* CONFIG_X86_PTRACE_BTS */ | 1180 | #endif /* CONFIG_X86_PTRACE_BTS */ |
1180 | 1181 | ||
1181 | /* PID/PID hash table linkage. */ | 1182 | /* PID/PID hash table linkage. */ |
diff --git a/include/trace/boot.h b/include/trace/boot.h index 6b54537eab02..088ea089e31d 100644 --- a/include/trace/boot.h +++ b/include/trace/boot.h | |||
@@ -1,6 +1,10 @@ | |||
1 | #ifndef _LINUX_TRACE_BOOT_H | 1 | #ifndef _LINUX_TRACE_BOOT_H |
2 | #define _LINUX_TRACE_BOOT_H | 2 | #define _LINUX_TRACE_BOOT_H |
3 | 3 | ||
4 | #include <linux/module.h> | ||
5 | #include <linux/kallsyms.h> | ||
6 | #include <linux/init.h> | ||
7 | |||
4 | /* | 8 | /* |
5 | * Structure which defines the trace of an initcall | 9 | * Structure which defines the trace of an initcall |
6 | * while it is called. | 10 | * while it is called. |
@@ -9,7 +13,7 @@ | |||
9 | */ | 13 | */ |
10 | struct boot_trace_call { | 14 | struct boot_trace_call { |
11 | pid_t caller; | 15 | pid_t caller; |
12 | char func[KSYM_NAME_LEN]; | 16 | char func[KSYM_SYMBOL_LEN]; |
13 | }; | 17 | }; |
14 | 18 | ||
15 | /* | 19 | /* |
@@ -17,7 +21,7 @@ struct boot_trace_call { | |||
17 | * while it returns. | 21 | * while it returns. |
18 | */ | 22 | */ |
19 | struct boot_trace_ret { | 23 | struct boot_trace_ret { |
20 | char func[KSYM_NAME_LEN]; | 24 | char func[KSYM_SYMBOL_LEN]; |
21 | int result; | 25 | int result; |
22 | unsigned long long duration; /* nsecs */ | 26 | unsigned long long duration; /* nsecs */ |
23 | }; | 27 | }; |
diff --git a/include/trace/sched.h b/include/trace/sched.h index 9b2854abf7e2..f4549d506b16 100644 --- a/include/trace/sched.h +++ b/include/trace/sched.h | |||
@@ -30,8 +30,8 @@ DECLARE_TRACE(sched_switch, | |||
30 | TPARGS(rq, prev, next)); | 30 | TPARGS(rq, prev, next)); |
31 | 31 | ||
32 | DECLARE_TRACE(sched_migrate_task, | 32 | DECLARE_TRACE(sched_migrate_task, |
33 | TPPROTO(struct rq *rq, struct task_struct *p, int dest_cpu), | 33 | TPPROTO(struct task_struct *p, int orig_cpu, int dest_cpu), |
34 | TPARGS(rq, p, dest_cpu)); | 34 | TPARGS(p, orig_cpu, dest_cpu)); |
35 | 35 | ||
36 | DECLARE_TRACE(sched_process_free, | 36 | DECLARE_TRACE(sched_process_free, |
37 | TPPROTO(struct task_struct *p), | 37 | TPPROTO(struct task_struct *p), |
diff --git a/kernel/sched.c b/kernel/sched.c index 7729c4bbc8ba..d377097572f9 100644 --- a/kernel/sched.c +++ b/kernel/sched.c | |||
@@ -1851,6 +1851,8 @@ void set_task_cpu(struct task_struct *p, unsigned int new_cpu) | |||
1851 | 1851 | ||
1852 | clock_offset = old_rq->clock - new_rq->clock; | 1852 | clock_offset = old_rq->clock - new_rq->clock; |
1853 | 1853 | ||
1854 | trace_sched_migrate_task(p, task_cpu(p), new_cpu); | ||
1855 | |||
1854 | #ifdef CONFIG_SCHEDSTATS | 1856 | #ifdef CONFIG_SCHEDSTATS |
1855 | if (p->se.wait_start) | 1857 | if (p->se.wait_start) |
1856 | p->se.wait_start -= clock_offset; | 1858 | p->se.wait_start -= clock_offset; |
@@ -2868,7 +2870,6 @@ static void sched_migrate_task(struct task_struct *p, int dest_cpu) | |||
2868 | || unlikely(!cpu_active(dest_cpu))) | 2870 | || unlikely(!cpu_active(dest_cpu))) |
2869 | goto out; | 2871 | goto out; |
2870 | 2872 | ||
2871 | trace_sched_migrate_task(rq, p, dest_cpu); | ||
2872 | /* force the process onto the specified CPU */ | 2873 | /* force the process onto the specified CPU */ |
2873 | if (migrate_task(p, dest_cpu, &req)) { | 2874 | if (migrate_task(p, dest_cpu, &req)) { |
2874 | /* Need to wait for migration thread (might exit: take ref). */ | 2875 | /* Need to wait for migration thread (might exit: take ref). */ |
diff --git a/kernel/trace/Kconfig b/kernel/trace/Kconfig index bde6f03512d5..d8bae6f4219e 100644 --- a/kernel/trace/Kconfig +++ b/kernel/trace/Kconfig | |||
@@ -251,9 +251,9 @@ config STACK_TRACER | |||
251 | 251 | ||
252 | Say N if unsure. | 252 | Say N if unsure. |
253 | 253 | ||
254 | config BTS_TRACER | 254 | config HW_BRANCH_TRACER |
255 | depends on HAVE_HW_BRANCH_TRACER | 255 | depends on HAVE_HW_BRANCH_TRACER |
256 | bool "Trace branches" | 256 | bool "Trace hw branches" |
257 | select TRACING | 257 | select TRACING |
258 | help | 258 | help |
259 | This tracer records all branches on the system in a circular | 259 | This tracer records all branches on the system in a circular |
diff --git a/kernel/trace/Makefile b/kernel/trace/Makefile index 62dc561b6676..349d5a93653f 100644 --- a/kernel/trace/Makefile +++ b/kernel/trace/Makefile | |||
@@ -31,7 +31,7 @@ obj-$(CONFIG_MMIOTRACE) += trace_mmiotrace.o | |||
31 | obj-$(CONFIG_BOOT_TRACER) += trace_boot.o | 31 | obj-$(CONFIG_BOOT_TRACER) += trace_boot.o |
32 | obj-$(CONFIG_FUNCTION_GRAPH_TRACER) += trace_functions_graph.o | 32 | obj-$(CONFIG_FUNCTION_GRAPH_TRACER) += trace_functions_graph.o |
33 | obj-$(CONFIG_TRACE_BRANCH_PROFILING) += trace_branch.o | 33 | obj-$(CONFIG_TRACE_BRANCH_PROFILING) += trace_branch.o |
34 | obj-$(CONFIG_BTS_TRACER) += trace_bts.o | 34 | obj-$(CONFIG_HW_BRANCH_TRACER) += trace_hw_branches.o |
35 | obj-$(CONFIG_POWER_TRACER) += trace_power.o | 35 | obj-$(CONFIG_POWER_TRACER) += trace_power.o |
36 | 36 | ||
37 | libftrace-y := ftrace.o | 37 | libftrace-y := ftrace.o |
diff --git a/kernel/trace/trace.c b/kernel/trace/trace.c index 8ebe0070c47a..639344a4d3a2 100644 --- a/kernel/trace/trace.c +++ b/kernel/trace/trace.c | |||
@@ -2425,7 +2425,7 @@ __tracing_open(struct inode *inode, struct file *file, int *ret) | |||
2425 | 2425 | ||
2426 | /* Notify the tracer early; before we stop tracing. */ | 2426 | /* Notify the tracer early; before we stop tracing. */ |
2427 | if (iter->trace && iter->trace->open) | 2427 | if (iter->trace && iter->trace->open) |
2428 | iter->trace->open(iter); | 2428 | iter->trace->open(iter); |
2429 | 2429 | ||
2430 | /* Annotate start of buffers if we had overruns */ | 2430 | /* Annotate start of buffers if we had overruns */ |
2431 | if (ring_buffer_overruns(iter->tr->buffer)) | 2431 | if (ring_buffer_overruns(iter->tr->buffer)) |
diff --git a/kernel/trace/trace.h b/kernel/trace/trace.h index 5ac697065a48..f07c246dd73d 100644 --- a/kernel/trace/trace.h +++ b/kernel/trace/trace.h | |||
@@ -28,7 +28,7 @@ enum trace_type { | |||
28 | TRACE_GRAPH_RET, | 28 | TRACE_GRAPH_RET, |
29 | TRACE_GRAPH_ENT, | 29 | TRACE_GRAPH_ENT, |
30 | TRACE_USER_STACK, | 30 | TRACE_USER_STACK, |
31 | TRACE_BTS, | 31 | TRACE_HW_BRANCHES, |
32 | TRACE_POWER, | 32 | TRACE_POWER, |
33 | 33 | ||
34 | __TRACE_LAST_TYPE | 34 | __TRACE_LAST_TYPE |
@@ -159,10 +159,10 @@ struct trace_branch { | |||
159 | char correct; | 159 | char correct; |
160 | }; | 160 | }; |
161 | 161 | ||
162 | struct bts_entry { | 162 | struct hw_branch_entry { |
163 | struct trace_entry ent; | 163 | struct trace_entry ent; |
164 | unsigned long from; | 164 | u64 from; |
165 | unsigned long to; | 165 | u64 to; |
166 | }; | 166 | }; |
167 | 167 | ||
168 | struct trace_power { | 168 | struct trace_power { |
@@ -278,7 +278,7 @@ extern void __ftrace_bad_type(void); | |||
278 | TRACE_GRAPH_ENT); \ | 278 | TRACE_GRAPH_ENT); \ |
279 | IF_ASSIGN(var, ent, struct ftrace_graph_ret_entry, \ | 279 | IF_ASSIGN(var, ent, struct ftrace_graph_ret_entry, \ |
280 | TRACE_GRAPH_RET); \ | 280 | TRACE_GRAPH_RET); \ |
281 | IF_ASSIGN(var, ent, struct bts_entry, TRACE_BTS);\ | 281 | IF_ASSIGN(var, ent, struct hw_branch_entry, TRACE_HW_BRANCHES);\ |
282 | IF_ASSIGN(var, ent, struct trace_power, TRACE_POWER); \ | 282 | IF_ASSIGN(var, ent, struct trace_power, TRACE_POWER); \ |
283 | __ftrace_bad_type(); \ | 283 | __ftrace_bad_type(); \ |
284 | } while (0) | 284 | } while (0) |
@@ -414,9 +414,7 @@ void trace_function(struct trace_array *tr, | |||
414 | 414 | ||
415 | void trace_graph_return(struct ftrace_graph_ret *trace); | 415 | void trace_graph_return(struct ftrace_graph_ret *trace); |
416 | int trace_graph_entry(struct ftrace_graph_ent *trace); | 416 | int trace_graph_entry(struct ftrace_graph_ent *trace); |
417 | void trace_bts(struct trace_array *tr, | 417 | void trace_hw_branch(struct trace_array *tr, u64 from, u64 to); |
418 | unsigned long from, | ||
419 | unsigned long to); | ||
420 | 418 | ||
421 | void tracing_start_cmdline_record(void); | 419 | void tracing_start_cmdline_record(void); |
422 | void tracing_stop_cmdline_record(void); | 420 | void tracing_stop_cmdline_record(void); |
diff --git a/kernel/trace/trace_bts.c b/kernel/trace/trace_hw_branches.c index 23b76e4690ef..ee29e012aa97 100644 --- a/kernel/trace/trace_bts.c +++ b/kernel/trace/trace_hw_branches.c | |||
@@ -1,5 +1,5 @@ | |||
1 | /* | 1 | /* |
2 | * BTS tracer | 2 | * h/w branch tracer for x86 based on bts |
3 | * | 3 | * |
4 | * Copyright (C) 2008 Markus Metzger <markus.t.metzger@gmail.com> | 4 | * Copyright (C) 2008 Markus Metzger <markus.t.metzger@gmail.com> |
5 | * | 5 | * |
@@ -25,68 +25,6 @@ static DEFINE_PER_CPU(unsigned char[SIZEOF_BTS], buffer); | |||
25 | #define this_buffer per_cpu(buffer, smp_processor_id()) | 25 | #define this_buffer per_cpu(buffer, smp_processor_id()) |
26 | 26 | ||
27 | 27 | ||
28 | /* | ||
29 | * Information to interpret a BTS record. | ||
30 | * This will go into an in-kernel BTS interface. | ||
31 | */ | ||
32 | static unsigned char sizeof_field; | ||
33 | static unsigned long debugctl_mask; | ||
34 | |||
35 | #define sizeof_bts (3 * sizeof_field) | ||
36 | |||
37 | static void bts_trace_cpuinit(struct cpuinfo_x86 *c) | ||
38 | { | ||
39 | switch (c->x86) { | ||
40 | case 0x6: | ||
41 | switch (c->x86_model) { | ||
42 | case 0x0 ... 0xC: | ||
43 | break; | ||
44 | case 0xD: | ||
45 | case 0xE: /* Pentium M */ | ||
46 | sizeof_field = sizeof(long); | ||
47 | debugctl_mask = (1<<6)|(1<<7); | ||
48 | break; | ||
49 | default: | ||
50 | sizeof_field = 8; | ||
51 | debugctl_mask = (1<<6)|(1<<7); | ||
52 | break; | ||
53 | } | ||
54 | break; | ||
55 | case 0xF: | ||
56 | switch (c->x86_model) { | ||
57 | case 0x0: | ||
58 | case 0x1: | ||
59 | case 0x2: /* Netburst */ | ||
60 | sizeof_field = sizeof(long); | ||
61 | debugctl_mask = (1<<2)|(1<<3); | ||
62 | break; | ||
63 | default: | ||
64 | /* sorry, don't know about them */ | ||
65 | break; | ||
66 | } | ||
67 | break; | ||
68 | default: | ||
69 | /* sorry, don't know about them */ | ||
70 | break; | ||
71 | } | ||
72 | } | ||
73 | |||
74 | static inline void bts_enable(void) | ||
75 | { | ||
76 | unsigned long debugctl; | ||
77 | |||
78 | rdmsrl(MSR_IA32_DEBUGCTLMSR, debugctl); | ||
79 | wrmsrl(MSR_IA32_DEBUGCTLMSR, debugctl | debugctl_mask); | ||
80 | } | ||
81 | |||
82 | static inline void bts_disable(void) | ||
83 | { | ||
84 | unsigned long debugctl; | ||
85 | |||
86 | rdmsrl(MSR_IA32_DEBUGCTLMSR, debugctl); | ||
87 | wrmsrl(MSR_IA32_DEBUGCTLMSR, debugctl & ~debugctl_mask); | ||
88 | } | ||
89 | |||
90 | static void bts_trace_reset(struct trace_array *tr) | 28 | static void bts_trace_reset(struct trace_array *tr) |
91 | { | 29 | { |
92 | int cpu; | 30 | int cpu; |
@@ -99,15 +37,17 @@ static void bts_trace_reset(struct trace_array *tr) | |||
99 | 37 | ||
100 | static void bts_trace_start_cpu(void *arg) | 38 | static void bts_trace_start_cpu(void *arg) |
101 | { | 39 | { |
40 | if (this_tracer) | ||
41 | ds_release_bts(this_tracer); | ||
42 | |||
102 | this_tracer = | 43 | this_tracer = |
103 | ds_request_bts(/* task = */ NULL, this_buffer, SIZEOF_BTS, | 44 | ds_request_bts(/* task = */ NULL, this_buffer, SIZEOF_BTS, |
104 | /* ovfl = */ NULL, /* th = */ (size_t)-1); | 45 | /* ovfl = */ NULL, /* th = */ (size_t)-1, |
46 | BTS_KERNEL); | ||
105 | if (IS_ERR(this_tracer)) { | 47 | if (IS_ERR(this_tracer)) { |
106 | this_tracer = NULL; | 48 | this_tracer = NULL; |
107 | return; | 49 | return; |
108 | } | 50 | } |
109 | |||
110 | bts_enable(); | ||
111 | } | 51 | } |
112 | 52 | ||
113 | static void bts_trace_start(struct trace_array *tr) | 53 | static void bts_trace_start(struct trace_array *tr) |
@@ -123,8 +63,6 @@ static void bts_trace_start(struct trace_array *tr) | |||
123 | static void bts_trace_stop_cpu(void *arg) | 63 | static void bts_trace_stop_cpu(void *arg) |
124 | { | 64 | { |
125 | if (this_tracer) { | 65 | if (this_tracer) { |
126 | bts_disable(); | ||
127 | |||
128 | ds_release_bts(this_tracer); | 66 | ds_release_bts(this_tracer); |
129 | this_tracer = NULL; | 67 | this_tracer = NULL; |
130 | } | 68 | } |
@@ -140,7 +78,6 @@ static void bts_trace_stop(struct trace_array *tr) | |||
140 | 78 | ||
141 | static int bts_trace_init(struct trace_array *tr) | 79 | static int bts_trace_init(struct trace_array *tr) |
142 | { | 80 | { |
143 | bts_trace_cpuinit(&boot_cpu_data); | ||
144 | bts_trace_reset(tr); | 81 | bts_trace_reset(tr); |
145 | bts_trace_start(tr); | 82 | bts_trace_start(tr); |
146 | 83 | ||
@@ -149,47 +86,37 @@ static int bts_trace_init(struct trace_array *tr) | |||
149 | 86 | ||
150 | static void bts_trace_print_header(struct seq_file *m) | 87 | static void bts_trace_print_header(struct seq_file *m) |
151 | { | 88 | { |
152 | #ifdef __i386__ | ||
153 | seq_puts(m, "# CPU# FROM TO FUNCTION\n"); | ||
154 | seq_puts(m, "# | | | |\n"); | ||
155 | #else | ||
156 | seq_puts(m, | 89 | seq_puts(m, |
157 | "# CPU# FROM TO FUNCTION\n"); | 90 | "# CPU# FROM TO FUNCTION\n"); |
158 | seq_puts(m, | 91 | seq_puts(m, |
159 | "# | | | |\n"); | 92 | "# | | | |\n"); |
160 | #endif | ||
161 | } | 93 | } |
162 | 94 | ||
163 | static enum print_line_t bts_trace_print_line(struct trace_iterator *iter) | 95 | static enum print_line_t bts_trace_print_line(struct trace_iterator *iter) |
164 | { | 96 | { |
165 | struct trace_entry *entry = iter->ent; | 97 | struct trace_entry *entry = iter->ent; |
166 | struct trace_seq *seq = &iter->seq; | 98 | struct trace_seq *seq = &iter->seq; |
167 | struct bts_entry *it; | 99 | struct hw_branch_entry *it; |
168 | 100 | ||
169 | trace_assign_type(it, entry); | 101 | trace_assign_type(it, entry); |
170 | 102 | ||
171 | if (entry->type == TRACE_BTS) { | 103 | if (entry->type == TRACE_HW_BRANCHES) { |
172 | int ret; | 104 | if (trace_seq_printf(seq, "%4d ", entry->cpu) && |
173 | #ifdef CONFIG_KALLSYMS | 105 | trace_seq_printf(seq, "0x%016llx -> 0x%016llx ", |
174 | char function[KSYM_SYMBOL_LEN]; | 106 | it->from, it->to) && |
175 | sprint_symbol(function, it->from); | 107 | (!it->from || |
176 | #else | 108 | seq_print_ip_sym(seq, it->from, /* sym_flags = */ 0)) && |
177 | char *function = "<unknown>"; | 109 | trace_seq_printf(seq, "\n")) |
178 | #endif | 110 | return TRACE_TYPE_HANDLED; |
179 | 111 | return TRACE_TYPE_PARTIAL_LINE;; | |
180 | ret = trace_seq_printf(seq, "%4d 0x%lx -> 0x%lx [%s]\n", | ||
181 | entry->cpu, it->from, it->to, function); | ||
182 | if (!ret) | ||
183 | return TRACE_TYPE_PARTIAL_LINE;; | ||
184 | return TRACE_TYPE_HANDLED; | ||
185 | } | 112 | } |
186 | return TRACE_TYPE_UNHANDLED; | 113 | return TRACE_TYPE_UNHANDLED; |
187 | } | 114 | } |
188 | 115 | ||
189 | void trace_bts(struct trace_array *tr, unsigned long from, unsigned long to) | 116 | void trace_hw_branch(struct trace_array *tr, u64 from, u64 to) |
190 | { | 117 | { |
191 | struct ring_buffer_event *event; | 118 | struct ring_buffer_event *event; |
192 | struct bts_entry *entry; | 119 | struct hw_branch_entry *entry; |
193 | unsigned long irq; | 120 | unsigned long irq; |
194 | 121 | ||
195 | event = ring_buffer_lock_reserve(tr->buffer, sizeof(*entry), &irq); | 122 | event = ring_buffer_lock_reserve(tr->buffer, sizeof(*entry), &irq); |
@@ -197,56 +124,58 @@ void trace_bts(struct trace_array *tr, unsigned long from, unsigned long to) | |||
197 | return; | 124 | return; |
198 | entry = ring_buffer_event_data(event); | 125 | entry = ring_buffer_event_data(event); |
199 | tracing_generic_entry_update(&entry->ent, 0, from); | 126 | tracing_generic_entry_update(&entry->ent, 0, from); |
200 | entry->ent.type = TRACE_BTS; | 127 | entry->ent.type = TRACE_HW_BRANCHES; |
201 | entry->ent.cpu = smp_processor_id(); | 128 | entry->ent.cpu = smp_processor_id(); |
202 | entry->from = from; | 129 | entry->from = from; |
203 | entry->to = to; | 130 | entry->to = to; |
204 | ring_buffer_unlock_commit(tr->buffer, event, irq); | 131 | ring_buffer_unlock_commit(tr->buffer, event, irq); |
205 | } | 132 | } |
206 | 133 | ||
207 | static void trace_bts_at(struct trace_array *tr, size_t index) | 134 | static void trace_bts_at(struct trace_array *tr, |
135 | const struct bts_trace *trace, void *at) | ||
208 | { | 136 | { |
209 | const void *raw = NULL; | 137 | struct bts_struct bts; |
210 | unsigned long from, to; | 138 | int err = 0; |
211 | int err; | ||
212 | 139 | ||
213 | err = ds_access_bts(this_tracer, index, &raw); | 140 | WARN_ON_ONCE(!trace->read); |
214 | if (err < 0) | 141 | if (!trace->read) |
215 | return; | 142 | return; |
216 | 143 | ||
217 | from = *(const unsigned long *)raw; | 144 | err = trace->read(this_tracer, at, &bts); |
218 | to = *(const unsigned long *)((const char *)raw + sizeof_field); | 145 | if (err < 0) |
146 | return; | ||
219 | 147 | ||
220 | trace_bts(tr, from, to); | 148 | switch (bts.qualifier) { |
149 | case BTS_BRANCH: | ||
150 | trace_hw_branch(tr, bts.variant.lbr.from, bts.variant.lbr.to); | ||
151 | break; | ||
152 | } | ||
221 | } | 153 | } |
222 | 154 | ||
223 | static void trace_bts_cpu(void *arg) | 155 | static void trace_bts_cpu(void *arg) |
224 | { | 156 | { |
225 | struct trace_array *tr = (struct trace_array *) arg; | 157 | struct trace_array *tr = (struct trace_array *) arg; |
226 | size_t index = 0, end = 0, i; | 158 | const struct bts_trace *trace; |
227 | int err; | 159 | unsigned char *at; |
228 | 160 | ||
229 | if (!this_tracer) | 161 | if (!this_tracer) |
230 | return; | 162 | return; |
231 | 163 | ||
232 | bts_disable(); | 164 | ds_suspend_bts(this_tracer); |
233 | 165 | trace = ds_read_bts(this_tracer); | |
234 | err = ds_get_bts_index(this_tracer, &index); | 166 | if (!trace) |
235 | if (err < 0) | ||
236 | goto out; | ||
237 | |||
238 | err = ds_get_bts_end(this_tracer, &end); | ||
239 | if (err < 0) | ||
240 | goto out; | 167 | goto out; |
241 | 168 | ||
242 | for (i = index; i < end; i++) | 169 | for (at = trace->ds.top; (void *)at < trace->ds.end; |
243 | trace_bts_at(tr, i); | 170 | at += trace->ds.size) |
171 | trace_bts_at(tr, trace, at); | ||
244 | 172 | ||
245 | for (i = 0; i < index; i++) | 173 | for (at = trace->ds.begin; (void *)at < trace->ds.top; |
246 | trace_bts_at(tr, i); | 174 | at += trace->ds.size) |
175 | trace_bts_at(tr, trace, at); | ||
247 | 176 | ||
248 | out: | 177 | out: |
249 | bts_enable(); | 178 | ds_resume_bts(this_tracer); |
250 | } | 179 | } |
251 | 180 | ||
252 | static void trace_bts_prepare(struct trace_iterator *iter) | 181 | static void trace_bts_prepare(struct trace_iterator *iter) |
@@ -259,7 +188,7 @@ static void trace_bts_prepare(struct trace_iterator *iter) | |||
259 | 188 | ||
260 | struct tracer bts_tracer __read_mostly = | 189 | struct tracer bts_tracer __read_mostly = |
261 | { | 190 | { |
262 | .name = "bts", | 191 | .name = "hw-branch-tracer", |
263 | .init = bts_trace_init, | 192 | .init = bts_trace_init, |
264 | .reset = bts_trace_stop, | 193 | .reset = bts_trace_stop, |
265 | .print_header = bts_trace_print_header, | 194 | .print_header = bts_trace_print_header, |