aboutsummaryrefslogtreecommitdiffstats
path: root/kernel/trace/trace_kprobe.c
diff options
context:
space:
mode:
Diffstat (limited to 'kernel/trace/trace_kprobe.c')
-rw-r--r--kernel/trace/trace_kprobe.c115
1 files changed, 73 insertions, 42 deletions
diff --git a/kernel/trace/trace_kprobe.c b/kernel/trace/trace_kprobe.c
index 09cba270392d..97309d4714f7 100644
--- a/kernel/trace/trace_kprobe.c
+++ b/kernel/trace/trace_kprobe.c
@@ -1149,35 +1149,49 @@ static __kprobes int kprobe_profile_func(struct kprobe *kp,
1149 struct trace_probe *tp = container_of(kp, struct trace_probe, rp.kp); 1149 struct trace_probe *tp = container_of(kp, struct trace_probe, rp.kp);
1150 struct ftrace_event_call *call = &tp->call; 1150 struct ftrace_event_call *call = &tp->call;
1151 struct kprobe_trace_entry *entry; 1151 struct kprobe_trace_entry *entry;
1152 int size, __size, i, pc; 1152 struct trace_entry *ent;
1153 int size, __size, i, pc, __cpu;
1153 unsigned long irq_flags; 1154 unsigned long irq_flags;
1155 char *raw_data;
1154 1156
1155 local_save_flags(irq_flags);
1156 pc = preempt_count(); 1157 pc = preempt_count();
1157
1158 __size = SIZEOF_KPROBE_TRACE_ENTRY(tp->nr_args); 1158 __size = SIZEOF_KPROBE_TRACE_ENTRY(tp->nr_args);
1159 size = ALIGN(__size + sizeof(u32), sizeof(u64)); 1159 size = ALIGN(__size + sizeof(u32), sizeof(u64));
1160 size -= sizeof(u32); 1160 size -= sizeof(u32);
1161 if (WARN_ONCE(size > FTRACE_MAX_PROFILE_SIZE,
1162 "profile buffer not large enough"))
1163 return 0;
1161 1164
1162 do { 1165 /*
1163 char raw_data[size]; 1166 * Protect the non nmi buffer
1164 struct trace_entry *ent; 1167 * This also protects the rcu read side
1165 /* 1168 */
1166 * Zero dead bytes from alignment to avoid stack leak 1169 local_irq_save(irq_flags);
1167 * to userspace 1170 __cpu = smp_processor_id();
1168 */ 1171
1169 *(u64 *)(&raw_data[size - sizeof(u64)]) = 0ULL; 1172 if (in_nmi())
1170 entry = (struct kprobe_trace_entry *)raw_data; 1173 raw_data = rcu_dereference(trace_profile_buf_nmi);
1171 ent = &entry->ent; 1174 else
1172 1175 raw_data = rcu_dereference(trace_profile_buf);
1173 tracing_generic_entry_update(ent, irq_flags, pc); 1176
1174 ent->type = call->id; 1177 if (!raw_data)
1175 entry->nargs = tp->nr_args; 1178 goto end;
1176 entry->ip = (unsigned long)kp->addr; 1179
1177 for (i = 0; i < tp->nr_args; i++) 1180 raw_data = per_cpu_ptr(raw_data, __cpu);
1178 entry->args[i] = call_fetch(&tp->args[i].fetch, regs); 1181 /* Zero dead bytes from alignment to avoid buffer leak to userspace */
1179 perf_tp_event(call->id, entry->ip, 1, entry, size); 1182 *(u64 *)(&raw_data[size - sizeof(u64)]) = 0ULL;
1180 } while (0); 1183 entry = (struct kprobe_trace_entry *)raw_data;
1184 ent = &entry->ent;
1185
1186 tracing_generic_entry_update(ent, irq_flags, pc);
1187 ent->type = call->id;
1188 entry->nargs = tp->nr_args;
1189 entry->ip = (unsigned long)kp->addr;
1190 for (i = 0; i < tp->nr_args; i++)
1191 entry->args[i] = call_fetch(&tp->args[i].fetch, regs);
1192 perf_tp_event(call->id, entry->ip, 1, entry, size);
1193end:
1194 local_irq_restore(irq_flags);
1181 return 0; 1195 return 0;
1182} 1196}
1183 1197
@@ -1188,33 +1202,50 @@ static __kprobes int kretprobe_profile_func(struct kretprobe_instance *ri,
1188 struct trace_probe *tp = container_of(ri->rp, struct trace_probe, rp); 1202 struct trace_probe *tp = container_of(ri->rp, struct trace_probe, rp);
1189 struct ftrace_event_call *call = &tp->call; 1203 struct ftrace_event_call *call = &tp->call;
1190 struct kretprobe_trace_entry *entry; 1204 struct kretprobe_trace_entry *entry;
1191 int size, __size, i, pc; 1205 struct trace_entry *ent;
1206 int size, __size, i, pc, __cpu;
1192 unsigned long irq_flags; 1207 unsigned long irq_flags;
1208 char *raw_data;
1193 1209
1194 local_save_flags(irq_flags);
1195 pc = preempt_count(); 1210 pc = preempt_count();
1196
1197 __size = SIZEOF_KRETPROBE_TRACE_ENTRY(tp->nr_args); 1211 __size = SIZEOF_KRETPROBE_TRACE_ENTRY(tp->nr_args);
1198 size = ALIGN(__size + sizeof(u32), sizeof(u64)); 1212 size = ALIGN(__size + sizeof(u32), sizeof(u64));
1199 size -= sizeof(u32); 1213 size -= sizeof(u32);
1214 if (WARN_ONCE(size > FTRACE_MAX_PROFILE_SIZE,
1215 "profile buffer not large enough"))
1216 return 0;
1217
1218 /*
1219 * Protect the non nmi buffer
1220 * This also protects the rcu read side
1221 */
1222 local_irq_save(irq_flags);
1223 __cpu = smp_processor_id();
1224
1225 if (in_nmi())
1226 raw_data = rcu_dereference(trace_profile_buf_nmi);
1227 else
1228 raw_data = rcu_dereference(trace_profile_buf);
1229
1230 if (!raw_data)
1231 goto end;
1232
1233 raw_data = per_cpu_ptr(raw_data, __cpu);
1234 /* Zero dead bytes from alignment to avoid buffer leak to userspace */
1235 *(u64 *)(&raw_data[size - sizeof(u64)]) = 0ULL;
1236 entry = (struct kretprobe_trace_entry *)raw_data;
1237 ent = &entry->ent;
1200 1238
1201 do { 1239 tracing_generic_entry_update(ent, irq_flags, pc);
1202 char raw_data[size]; 1240 ent->type = call->id;
1203 struct trace_entry *ent; 1241 entry->nargs = tp->nr_args;
1204 1242 entry->func = (unsigned long)tp->rp.kp.addr;
1205 *(u64 *)(&raw_data[size - sizeof(u64)]) = 0ULL; 1243 entry->ret_ip = (unsigned long)ri->ret_addr;
1206 entry = (struct kretprobe_trace_entry *)raw_data; 1244 for (i = 0; i < tp->nr_args; i++)
1207 ent = &entry->ent; 1245 entry->args[i] = call_fetch(&tp->args[i].fetch, regs);
1208 1246 perf_tp_event(call->id, entry->ret_ip, 1, entry, size);
1209 tracing_generic_entry_update(ent, irq_flags, pc); 1247end:
1210 ent->type = call->id; 1248 local_irq_restore(irq_flags);
1211 entry->nargs = tp->nr_args;
1212 entry->func = (unsigned long)tp->rp.kp.addr;
1213 entry->ret_ip = (unsigned long)ri->ret_addr;
1214 for (i = 0; i < tp->nr_args; i++)
1215 entry->args[i] = call_fetch(&tp->args[i].fetch, regs);
1216 perf_tp_event(call->id, entry->ret_ip, 1, entry, size);
1217 } while (0);
1218 return 0; 1249 return 0;
1219} 1250}
1220 1251