|
@@ -25,139 +25,10 @@
|
|
|
#include <asm/ptrace.h>
|
|
|
|
|
|
#ifdef CONFIG_HW_PERF_EVENTS
|
|
|
-/*
|
|
|
- * Per performance monitor configuration as set via oprofilefs.
|
|
|
- */
|
|
|
-struct op_counter_config {
|
|
|
- unsigned long count;
|
|
|
- unsigned long enabled;
|
|
|
- unsigned long event;
|
|
|
- unsigned long unit_mask;
|
|
|
- unsigned long kernel;
|
|
|
- unsigned long user;
|
|
|
- struct perf_event_attr attr;
|
|
|
-};
|
|
|
-
|
|
|
-static int op_arm_enabled;
|
|
|
-static DEFINE_MUTEX(op_arm_mutex);
|
|
|
-
|
|
|
-static struct op_counter_config *counter_config;
|
|
|
-static struct perf_event **perf_events[nr_cpumask_bits];
|
|
|
-static int perf_num_counters;
|
|
|
-
|
|
|
-/*
|
|
|
- * Overflow callback for oprofile.
|
|
|
- */
|
|
|
-static void op_overflow_handler(struct perf_event *event, int unused,
|
|
|
- struct perf_sample_data *data, struct pt_regs *regs)
|
|
|
+char *op_name_from_perf_id(void)
|
|
|
{
|
|
|
- int id;
|
|
|
- u32 cpu = smp_processor_id();
|
|
|
-
|
|
|
- for (id = 0; id < perf_num_counters; ++id)
|
|
|
- if (perf_events[cpu][id] == event)
|
|
|
- break;
|
|
|
-
|
|
|
- if (id != perf_num_counters)
|
|
|
- oprofile_add_sample(regs, id);
|
|
|
- else
|
|
|
- pr_warning("oprofile: ignoring spurious overflow "
|
|
|
- "on cpu %u\n", cpu);
|
|
|
-}
|
|
|
-
|
|
|
-/*
|
|
|
- * Called by op_arm_setup to create perf attributes to mirror the oprofile
|
|
|
- * settings in counter_config. Attributes are created as `pinned' events and
|
|
|
- * so are permanently scheduled on the PMU.
|
|
|
- */
|
|
|
-static void op_perf_setup(void)
|
|
|
-{
|
|
|
- int i;
|
|
|
- u32 size = sizeof(struct perf_event_attr);
|
|
|
- struct perf_event_attr *attr;
|
|
|
-
|
|
|
- for (i = 0; i < perf_num_counters; ++i) {
|
|
|
- attr = &counter_config[i].attr;
|
|
|
- memset(attr, 0, size);
|
|
|
- attr->type = PERF_TYPE_RAW;
|
|
|
- attr->size = size;
|
|
|
- attr->config = counter_config[i].event;
|
|
|
- attr->sample_period = counter_config[i].count;
|
|
|
- attr->pinned = 1;
|
|
|
- }
|
|
|
-}
|
|
|
-
|
|
|
-static int op_create_counter(int cpu, int event)
|
|
|
-{
|
|
|
- int ret = 0;
|
|
|
- struct perf_event *pevent;
|
|
|
-
|
|
|
- if (!counter_config[event].enabled || (perf_events[cpu][event] != NULL))
|
|
|
- return ret;
|
|
|
-
|
|
|
- pevent = perf_event_create_kernel_counter(&counter_config[event].attr,
|
|
|
- cpu, NULL,
|
|
|
- op_overflow_handler);
|
|
|
-
|
|
|
- if (IS_ERR(pevent)) {
|
|
|
- ret = PTR_ERR(pevent);
|
|
|
- } else if (pevent->state != PERF_EVENT_STATE_ACTIVE) {
|
|
|
- perf_event_release_kernel(pevent);
|
|
|
- pr_warning("oprofile: failed to enable event %d "
|
|
|
- "on CPU %d\n", event, cpu);
|
|
|
- ret = -EBUSY;
|
|
|
- } else {
|
|
|
- perf_events[cpu][event] = pevent;
|
|
|
- }
|
|
|
-
|
|
|
- return ret;
|
|
|
-}
|
|
|
+ enum arm_perf_pmu_ids id = armpmu_get_pmu_id();
|
|
|
|
|
|
-static void op_destroy_counter(int cpu, int event)
|
|
|
-{
|
|
|
- struct perf_event *pevent = perf_events[cpu][event];
|
|
|
-
|
|
|
- if (pevent) {
|
|
|
- perf_event_release_kernel(pevent);
|
|
|
- perf_events[cpu][event] = NULL;
|
|
|
- }
|
|
|
-}
|
|
|
-
|
|
|
-/*
|
|
|
- * Called by op_arm_start to create active perf events based on the
|
|
|
- * perviously configured attributes.
|
|
|
- */
|
|
|
-static int op_perf_start(void)
|
|
|
-{
|
|
|
- int cpu, event, ret = 0;
|
|
|
-
|
|
|
- for_each_online_cpu(cpu) {
|
|
|
- for (event = 0; event < perf_num_counters; ++event) {
|
|
|
- ret = op_create_counter(cpu, event);
|
|
|
- if (ret)
|
|
|
- goto out;
|
|
|
- }
|
|
|
- }
|
|
|
-
|
|
|
-out:
|
|
|
- return ret;
|
|
|
-}
|
|
|
-
|
|
|
-/*
|
|
|
- * Called by op_arm_stop at the end of a profiling run.
|
|
|
- */
|
|
|
-static void op_perf_stop(void)
|
|
|
-{
|
|
|
- int cpu, event;
|
|
|
-
|
|
|
- for_each_online_cpu(cpu)
|
|
|
- for (event = 0; event < perf_num_counters; ++event)
|
|
|
- op_destroy_counter(cpu, event);
|
|
|
-}
|
|
|
-
|
|
|
-
|
|
|
-static char *op_name_from_perf_id(enum arm_perf_pmu_ids id)
|
|
|
-{
|
|
|
switch (id) {
|
|
|
case ARM_PERF_PMU_ID_XSCALE1:
|
|
|
return "arm/xscale1";
|
|
@@ -176,116 +47,6 @@ static char *op_name_from_perf_id(enum arm_perf_pmu_ids id)
|
|
|
}
|
|
|
}
|
|
|
|
|
|
-static int op_arm_create_files(struct super_block *sb, struct dentry *root)
|
|
|
-{
|
|
|
- unsigned int i;
|
|
|
-
|
|
|
- for (i = 0; i < perf_num_counters; i++) {
|
|
|
- struct dentry *dir;
|
|
|
- char buf[4];
|
|
|
-
|
|
|
- snprintf(buf, sizeof buf, "%d", i);
|
|
|
- dir = oprofilefs_mkdir(sb, root, buf);
|
|
|
- oprofilefs_create_ulong(sb, dir, "enabled", &counter_config[i].enabled);
|
|
|
- oprofilefs_create_ulong(sb, dir, "event", &counter_config[i].event);
|
|
|
- oprofilefs_create_ulong(sb, dir, "count", &counter_config[i].count);
|
|
|
- oprofilefs_create_ulong(sb, dir, "unit_mask", &counter_config[i].unit_mask);
|
|
|
- oprofilefs_create_ulong(sb, dir, "kernel", &counter_config[i].kernel);
|
|
|
- oprofilefs_create_ulong(sb, dir, "user", &counter_config[i].user);
|
|
|
- }
|
|
|
-
|
|
|
- return 0;
|
|
|
-}
|
|
|
-
|
|
|
-static int op_arm_setup(void)
|
|
|
-{
|
|
|
- spin_lock(&oprofilefs_lock);
|
|
|
- op_perf_setup();
|
|
|
- spin_unlock(&oprofilefs_lock);
|
|
|
- return 0;
|
|
|
-}
|
|
|
-
|
|
|
-static int op_arm_start(void)
|
|
|
-{
|
|
|
- int ret = -EBUSY;
|
|
|
-
|
|
|
- mutex_lock(&op_arm_mutex);
|
|
|
- if (!op_arm_enabled) {
|
|
|
- ret = 0;
|
|
|
- op_perf_start();
|
|
|
- op_arm_enabled = 1;
|
|
|
- }
|
|
|
- mutex_unlock(&op_arm_mutex);
|
|
|
- return ret;
|
|
|
-}
|
|
|
-
|
|
|
-static void op_arm_stop(void)
|
|
|
-{
|
|
|
- mutex_lock(&op_arm_mutex);
|
|
|
- if (op_arm_enabled)
|
|
|
- op_perf_stop();
|
|
|
- op_arm_enabled = 0;
|
|
|
- mutex_unlock(&op_arm_mutex);
|
|
|
-}
|
|
|
-
|
|
|
-#ifdef CONFIG_PM
|
|
|
-static int op_arm_suspend(struct platform_device *dev, pm_message_t state)
|
|
|
-{
|
|
|
- mutex_lock(&op_arm_mutex);
|
|
|
- if (op_arm_enabled)
|
|
|
- op_perf_stop();
|
|
|
- mutex_unlock(&op_arm_mutex);
|
|
|
- return 0;
|
|
|
-}
|
|
|
-
|
|
|
-static int op_arm_resume(struct platform_device *dev)
|
|
|
-{
|
|
|
- mutex_lock(&op_arm_mutex);
|
|
|
- if (op_arm_enabled && op_perf_start())
|
|
|
- op_arm_enabled = 0;
|
|
|
- mutex_unlock(&op_arm_mutex);
|
|
|
- return 0;
|
|
|
-}
|
|
|
-
|
|
|
-static struct platform_driver oprofile_driver = {
|
|
|
- .driver = {
|
|
|
- .name = "arm-oprofile",
|
|
|
- },
|
|
|
- .resume = op_arm_resume,
|
|
|
- .suspend = op_arm_suspend,
|
|
|
-};
|
|
|
-
|
|
|
-static struct platform_device *oprofile_pdev;
|
|
|
-
|
|
|
-static int __init init_driverfs(void)
|
|
|
-{
|
|
|
- int ret;
|
|
|
-
|
|
|
- ret = platform_driver_register(&oprofile_driver);
|
|
|
- if (ret)
|
|
|
- goto out;
|
|
|
-
|
|
|
- oprofile_pdev = platform_device_register_simple(
|
|
|
- oprofile_driver.driver.name, 0, NULL, 0);
|
|
|
- if (IS_ERR(oprofile_pdev)) {
|
|
|
- ret = PTR_ERR(oprofile_pdev);
|
|
|
- platform_driver_unregister(&oprofile_driver);
|
|
|
- }
|
|
|
-
|
|
|
-out:
|
|
|
- return ret;
|
|
|
-}
|
|
|
-
|
|
|
-static void exit_driverfs(void)
|
|
|
-{
|
|
|
- platform_device_unregister(oprofile_pdev);
|
|
|
- platform_driver_unregister(&oprofile_driver);
|
|
|
-}
|
|
|
-#else
|
|
|
-static int __init init_driverfs(void) { return 0; }
|
|
|
-#define exit_driverfs() do { } while (0)
|
|
|
-#endif /* CONFIG_PM */
|
|
|
-
|
|
|
static int report_trace(struct stackframe *frame, void *d)
|
|
|
{
|
|
|
unsigned int *depth = d;
|
|
@@ -350,74 +111,14 @@ static void arm_backtrace(struct pt_regs * const regs, unsigned int depth)
|
|
|
|
|
|
int __init oprofile_arch_init(struct oprofile_operations *ops)
|
|
|
{
|
|
|
- int cpu, ret = 0;
|
|
|
-
|
|
|
- perf_num_counters = armpmu_get_max_events();
|
|
|
-
|
|
|
- counter_config = kcalloc(perf_num_counters,
|
|
|
- sizeof(struct op_counter_config), GFP_KERNEL);
|
|
|
-
|
|
|
- if (!counter_config) {
|
|
|
- pr_info("oprofile: failed to allocate %d "
|
|
|
- "counters\n", perf_num_counters);
|
|
|
- return -ENOMEM;
|
|
|
- }
|
|
|
-
|
|
|
- ret = init_driverfs();
|
|
|
- if (ret) {
|
|
|
- kfree(counter_config);
|
|
|
- counter_config = NULL;
|
|
|
- return ret;
|
|
|
- }
|
|
|
-
|
|
|
- for_each_possible_cpu(cpu) {
|
|
|
- perf_events[cpu] = kcalloc(perf_num_counters,
|
|
|
- sizeof(struct perf_event *), GFP_KERNEL);
|
|
|
- if (!perf_events[cpu]) {
|
|
|
- pr_info("oprofile: failed to allocate %d perf events "
|
|
|
- "for cpu %d\n", perf_num_counters, cpu);
|
|
|
- while (--cpu >= 0)
|
|
|
- kfree(perf_events[cpu]);
|
|
|
- return -ENOMEM;
|
|
|
- }
|
|
|
- }
|
|
|
-
|
|
|
ops->backtrace = arm_backtrace;
|
|
|
- ops->create_files = op_arm_create_files;
|
|
|
- ops->setup = op_arm_setup;
|
|
|
- ops->start = op_arm_start;
|
|
|
- ops->stop = op_arm_stop;
|
|
|
- ops->shutdown = op_arm_stop;
|
|
|
- ops->cpu_type = op_name_from_perf_id(armpmu_get_pmu_id());
|
|
|
-
|
|
|
- if (!ops->cpu_type)
|
|
|
- ret = -ENODEV;
|
|
|
- else
|
|
|
- pr_info("oprofile: using %s\n", ops->cpu_type);
|
|
|
|
|
|
- return ret;
|
|
|
+ return oprofile_perf_init(ops);
|
|
|
}
|
|
|
|
|
|
-void oprofile_arch_exit(void)
|
|
|
+void __exit oprofile_arch_exit(void)
|
|
|
{
|
|
|
- int cpu, id;
|
|
|
- struct perf_event *event;
|
|
|
-
|
|
|
- if (*perf_events) {
|
|
|
- for_each_possible_cpu(cpu) {
|
|
|
- for (id = 0; id < perf_num_counters; ++id) {
|
|
|
- event = perf_events[cpu][id];
|
|
|
- if (event != NULL)
|
|
|
- perf_event_release_kernel(event);
|
|
|
- }
|
|
|
- kfree(perf_events[cpu]);
|
|
|
- }
|
|
|
- }
|
|
|
-
|
|
|
- if (counter_config) {
|
|
|
- kfree(counter_config);
|
|
|
- exit_driverfs();
|
|
|
- }
|
|
|
+ oprofile_perf_exit();
|
|
|
}
|
|
|
#else
|
|
|
int __init oprofile_arch_init(struct oprofile_operations *ops)
|
|
@@ -425,5 +126,5 @@ int __init oprofile_arch_init(struct oprofile_operations *ops)
|
|
|
pr_info("oprofile: hardware counters not available\n");
|
|
|
return -ENODEV;
|
|
|
}
|
|
|
-void oprofile_arch_exit(void) {}
|
|
|
+void __exit oprofile_arch_exit(void) {}
|
|
|
#endif /* CONFIG_HW_PERF_EVENTS */
|