|
@@ -1979,8 +1979,7 @@ void __perf_event_task_sched_out(struct task_struct *task,
|
|
|
perf_cgroup_sched_out(task);
|
|
|
}
|
|
|
|
|
|
-static void task_ctx_sched_out(struct perf_event_context *ctx,
|
|
|
- enum event_type_t event_type)
|
|
|
+static void task_ctx_sched_out(struct perf_event_context *ctx)
|
|
|
{
|
|
|
struct perf_cpu_context *cpuctx = __get_cpu_context(ctx);
|
|
|
|
|
@@ -1990,7 +1989,7 @@ static void task_ctx_sched_out(struct perf_event_context *ctx,
|
|
|
if (WARN_ON_ONCE(ctx != cpuctx->task_ctx))
|
|
|
return;
|
|
|
|
|
|
- ctx_sched_out(ctx, cpuctx, event_type);
|
|
|
+ ctx_sched_out(ctx, cpuctx, EVENT_ALL);
|
|
|
cpuctx->task_ctx = NULL;
|
|
|
}
|
|
|
|
|
@@ -2098,19 +2097,6 @@ static void cpu_ctx_sched_in(struct perf_cpu_context *cpuctx,
|
|
|
ctx_sched_in(ctx, cpuctx, event_type, task);
|
|
|
}
|
|
|
|
|
|
-static void task_ctx_sched_in(struct perf_event_context *ctx,
|
|
|
- enum event_type_t event_type)
|
|
|
-{
|
|
|
- struct perf_cpu_context *cpuctx;
|
|
|
-
|
|
|
- cpuctx = __get_cpu_context(ctx);
|
|
|
- if (cpuctx->task_ctx == ctx)
|
|
|
- return;
|
|
|
-
|
|
|
- ctx_sched_in(ctx, cpuctx, event_type, NULL);
|
|
|
- cpuctx->task_ctx = ctx;
|
|
|
-}
|
|
|
-
|
|
|
static void perf_event_context_sched_in(struct perf_event_context *ctx,
|
|
|
struct task_struct *task)
|
|
|
{
|
|
@@ -2363,7 +2349,7 @@ static void perf_rotate_context(struct perf_cpu_context *cpuctx)
|
|
|
|
|
|
cpu_ctx_sched_out(cpuctx, EVENT_FLEXIBLE);
|
|
|
if (ctx)
|
|
|
- task_ctx_sched_out(ctx, EVENT_FLEXIBLE);
|
|
|
+ ctx_sched_out(ctx, cpuctx, EVENT_FLEXIBLE);
|
|
|
|
|
|
rotate_ctx(&cpuctx->ctx);
|
|
|
if (ctx)
|
|
@@ -2371,7 +2357,7 @@ static void perf_rotate_context(struct perf_cpu_context *cpuctx)
|
|
|
|
|
|
cpu_ctx_sched_in(cpuctx, EVENT_FLEXIBLE, current);
|
|
|
if (ctx)
|
|
|
- task_ctx_sched_in(ctx, EVENT_FLEXIBLE);
|
|
|
+ ctx_sched_in(ctx, cpuctx, EVENT_FLEXIBLE, current);
|
|
|
|
|
|
done:
|
|
|
if (remove)
|
|
@@ -2435,7 +2421,7 @@ static void perf_event_enable_on_exec(struct perf_event_context *ctx)
|
|
|
perf_cgroup_sched_out(current);
|
|
|
|
|
|
raw_spin_lock(&ctx->lock);
|
|
|
- task_ctx_sched_out(ctx, EVENT_ALL);
|
|
|
+ task_ctx_sched_out(ctx);
|
|
|
|
|
|
list_for_each_entry(event, &ctx->pinned_groups, group_entry) {
|
|
|
ret = event_enable_on_exec(event, ctx);
|
|
@@ -6794,7 +6780,7 @@ static void perf_event_exit_task_context(struct task_struct *child, int ctxn)
|
|
|
* incremented the context's refcount before we do put_ctx below.
|
|
|
*/
|
|
|
raw_spin_lock(&child_ctx->lock);
|
|
|
- task_ctx_sched_out(child_ctx, EVENT_ALL);
|
|
|
+ task_ctx_sched_out(child_ctx);
|
|
|
child->perf_event_ctxp[ctxn] = NULL;
|
|
|
/*
|
|
|
* If this context is a clone; unclone it so it can't get
|