|
@@ -353,15 +353,7 @@ static inline int ftrace_get_offsets_##call( \
|
|
|
/*
|
|
|
* Generate the functions needed for tracepoint perf_counter support.
|
|
|
*
|
|
|
- * static void ftrace_profile_<call>(proto)
|
|
|
- * {
|
|
|
- * extern void perf_tpcounter_event(int, u64, u64);
|
|
|
- * u64 __addr = 0, __count = 1;
|
|
|
- *
|
|
|
- * <assign> <-- here we expand the TP_perf_assign() macro
|
|
|
- *
|
|
|
- * perf_tpcounter_event(event_<call>.id, __addr, __count);
|
|
|
- * }
|
|
|
+ * NOTE: The insertion profile callback (ftrace_profile_<call>) is defined later
|
|
|
*
|
|
|
* static int ftrace_profile_enable_<call>(struct ftrace_event_call *event_call)
|
|
|
* {
|
|
@@ -381,28 +373,10 @@ static inline int ftrace_get_offsets_##call( \
|
|
|
*
|
|
|
*/
|
|
|
|
|
|
-#undef TP_fast_assign
|
|
|
-#define TP_fast_assign(args...)
|
|
|
-
|
|
|
-#undef TP_perf_assign
|
|
|
-#define TP_perf_assign(args...) args
|
|
|
-
|
|
|
-#undef __perf_addr
|
|
|
-#define __perf_addr(a) __addr = (a)
|
|
|
-
|
|
|
-#undef __perf_count
|
|
|
-#define __perf_count(c) __count = (c)
|
|
|
-
|
|
|
#undef TRACE_EVENT
|
|
|
#define TRACE_EVENT(call, proto, args, tstruct, assign, print) \
|
|
|
\
|
|
|
-static void ftrace_profile_##call(proto) \
|
|
|
-{ \
|
|
|
- extern void perf_tpcounter_event(int, u64, u64); \
|
|
|
- u64 __addr = 0, __count = 1; \
|
|
|
- { assign; } \
|
|
|
- perf_tpcounter_event(event_##call.id, __addr, __count); \
|
|
|
-} \
|
|
|
+static void ftrace_profile_##call(proto); \
|
|
|
\
|
|
|
static int ftrace_profile_enable_##call(struct ftrace_event_call *event_call) \
|
|
|
{ \
|
|
@@ -422,12 +396,6 @@ static void ftrace_profile_disable_##call(struct ftrace_event_call *event_call)\
|
|
|
|
|
|
#include TRACE_INCLUDE(TRACE_INCLUDE_FILE)
|
|
|
|
|
|
-#undef TP_fast_assign
|
|
|
-#define TP_fast_assign(args...) args
|
|
|
-
|
|
|
-#undef TP_perf_assign
|
|
|
-#define TP_perf_assign(args...)
|
|
|
-
|
|
|
#endif
|
|
|
|
|
|
/*
|
|
@@ -647,5 +615,99 @@ __attribute__((section("_ftrace_events"))) event_##call = { \
|
|
|
|
|
|
#include TRACE_INCLUDE(TRACE_INCLUDE_FILE)
|
|
|
|
|
|
+/*
|
|
|
+ * Define the insertion callback to profile events
|
|
|
+ *
|
|
|
+ * The job is very similar to ftrace_raw_event_<call> except that we don't
|
|
|
+ * insert in the ring buffer but in a perf counter.
|
|
|
+ *
|
|
|
+ * static void ftrace_profile_<call>(proto)
|
|
|
+ * {
|
|
|
+ * struct ftrace_data_offsets_<call> __maybe_unused __data_offsets;
|
|
|
+ * struct ftrace_event_call *event_call = &event_<call>;
|
|
|
+ * extern void perf_tpcounter_event(int, u64, u64, void *, int);
|
|
|
+ * struct ftrace_raw_##call *entry;
|
|
|
+ * u64 __addr = 0, __count = 1;
|
|
|
+ * unsigned long irq_flags;
|
|
|
+ * int __entry_size;
|
|
|
+ * int __data_size;
|
|
|
+ * int pc;
|
|
|
+ *
|
|
|
+ * local_save_flags(irq_flags);
|
|
|
+ * pc = preempt_count();
|
|
|
+ *
|
|
|
+ * __data_size = ftrace_get_offsets_<call>(&__data_offsets, args);
|
|
|
+ * __entry_size = __data_size + sizeof(*entry);
|
|
|
+ *
|
|
|
+ * do {
|
|
|
+ * char raw_data[__entry_size]; <- allocate our sample in the stack
|
|
|
+ * struct trace_entry *ent;
|
|
|
+ *
|
|
|
+ * entry = (struct ftrace_raw_<call> *)raw_data;
|
|
|
+ * ent = &entry->ent;
|
|
|
+ * tracing_generic_entry_update(ent, irq_flags, pc);
|
|
|
+ * ent->type = event_call->id;
|
|
|
+ *
|
|
|
+ * <tstruct> <- do some jobs with dynamic arrays
|
|
|
+ *
|
|
|
+ * <assign> <- affect our values
|
|
|
+ *
|
|
|
+ * perf_tpcounter_event(event_call->id, __addr, __count, entry,
|
|
|
+ * __entry_size); <- submit them to perf counter
|
|
|
+ * } while (0);
|
|
|
+ *
|
|
|
+ * }
|
|
|
+ */
|
|
|
+
|
|
|
+#ifdef CONFIG_EVENT_PROFILE
|
|
|
+
|
|
|
+#undef __perf_addr
|
|
|
+#define __perf_addr(a) __addr = (a)
|
|
|
+
|
|
|
+#undef __perf_count
|
|
|
+#define __perf_count(c) __count = (c)
|
|
|
+
|
|
|
+#undef TRACE_EVENT
|
|
|
+#define TRACE_EVENT(call, proto, args, tstruct, assign, print) \
|
|
|
+static void ftrace_profile_##call(proto) \
|
|
|
+{ \
|
|
|
+ struct ftrace_data_offsets_##call __maybe_unused __data_offsets;\
|
|
|
+ struct ftrace_event_call *event_call = &event_##call; \
|
|
|
+ extern void perf_tpcounter_event(int, u64, u64, void *, int); \
|
|
|
+ struct ftrace_raw_##call *entry; \
|
|
|
+ u64 __addr = 0, __count = 1; \
|
|
|
+ unsigned long irq_flags; \
|
|
|
+ int __entry_size; \
|
|
|
+ int __data_size; \
|
|
|
+ int pc; \
|
|
|
+ \
|
|
|
+ local_save_flags(irq_flags); \
|
|
|
+ pc = preempt_count(); \
|
|
|
+ \
|
|
|
+ __data_size = ftrace_get_offsets_##call(&__data_offsets, args); \
|
|
|
+ __entry_size = ALIGN(__data_size + sizeof(*entry), sizeof(u64));\
|
|
|
+ \
|
|
|
+ do { \
|
|
|
+ char raw_data[__entry_size]; \
|
|
|
+ struct trace_entry *ent; \
|
|
|
+ \
|
|
|
+ entry = (struct ftrace_raw_##call *)raw_data; \
|
|
|
+ ent = &entry->ent; \
|
|
|
+ tracing_generic_entry_update(ent, irq_flags, pc); \
|
|
|
+ ent->type = event_call->id; \
|
|
|
+ \
|
|
|
+ tstruct \
|
|
|
+ \
|
|
|
+ { assign; } \
|
|
|
+ \
|
|
|
+ perf_tpcounter_event(event_call->id, __addr, __count, entry,\
|
|
|
+ __entry_size); \
|
|
|
+ } while (0); \
|
|
|
+ \
|
|
|
+}
|
|
|
+
|
|
|
+#include TRACE_INCLUDE(TRACE_INCLUDE_FILE)
|
|
|
+#endif /* CONFIG_EVENT_PROFILE */
|
|
|
+
|
|
|
#undef _TRACE_PROFILE_INIT
|
|
|
|