|
@@ -83,15 +83,51 @@ static unsigned int max_task_bp_pinned(int cpu)
|
|
|
return 0;
|
|
|
}
|
|
|
|
|
|
+static int task_bp_pinned(struct task_struct *tsk)
|
|
|
+{
|
|
|
+ struct perf_event_context *ctx = tsk->perf_event_ctxp;
|
|
|
+ struct list_head *list;
|
|
|
+ struct perf_event *bp;
|
|
|
+ unsigned long flags;
|
|
|
+ int count = 0;
|
|
|
+
|
|
|
+ if (WARN_ONCE(!ctx, "No perf context for this task"))
|
|
|
+ return 0;
|
|
|
+
|
|
|
+ list = &ctx->event_list;
|
|
|
+
|
|
|
+ spin_lock_irqsave(&ctx->lock, flags);
|
|
|
+
|
|
|
+ /*
|
|
|
+ * The current breakpoint counter is not included in the list
|
|
|
+ * at the open() callback time
|
|
|
+ */
|
|
|
+ list_for_each_entry(bp, list, event_entry) {
|
|
|
+ if (bp->attr.type == PERF_TYPE_BREAKPOINT)
|
|
|
+ count++;
|
|
|
+ }
|
|
|
+
|
|
|
+ spin_unlock_irqrestore(&ctx->lock, flags);
|
|
|
+
|
|
|
+ return count;
|
|
|
+}
|
|
|
+
|
|
|
/*
|
|
|
* Report the number of pinned/un-pinned breakpoints we have in
|
|
|
* a given cpu (cpu > -1) or in all of them (cpu = -1).
|
|
|
*/
|
|
|
-static void fetch_bp_busy_slots(struct bp_busy_slots *slots, int cpu)
|
|
|
+static void
|
|
|
+fetch_bp_busy_slots(struct bp_busy_slots *slots, struct perf_event *bp)
|
|
|
{
|
|
|
+ int cpu = bp->cpu;
|
|
|
+ struct task_struct *tsk = bp->ctx->task;
|
|
|
+
|
|
|
if (cpu >= 0) {
|
|
|
slots->pinned = per_cpu(nr_cpu_bp_pinned, cpu);
|
|
|
- slots->pinned += max_task_bp_pinned(cpu);
|
|
|
+ if (!tsk)
|
|
|
+ slots->pinned += max_task_bp_pinned(cpu);
|
|
|
+ else
|
|
|
+ slots->pinned += task_bp_pinned(tsk);
|
|
|
slots->flexible = per_cpu(nr_bp_flexible, cpu);
|
|
|
|
|
|
return;
|
|
@@ -101,7 +137,10 @@ static void fetch_bp_busy_slots(struct bp_busy_slots *slots, int cpu)
|
|
|
unsigned int nr;
|
|
|
|
|
|
nr = per_cpu(nr_cpu_bp_pinned, cpu);
|
|
|
- nr += max_task_bp_pinned(cpu);
|
|
|
+ if (!tsk)
|
|
|
+ nr += max_task_bp_pinned(cpu);
|
|
|
+ else
|
|
|
+ nr += task_bp_pinned(tsk);
|
|
|
|
|
|
if (nr > slots->pinned)
|
|
|
slots->pinned = nr;
|
|
@@ -118,33 +157,10 @@ static void fetch_bp_busy_slots(struct bp_busy_slots *slots, int cpu)
|
|
|
*/
|
|
|
static void toggle_bp_task_slot(struct task_struct *tsk, int cpu, bool enable)
|
|
|
{
|
|
|
- int count = 0;
|
|
|
- struct perf_event *bp;
|
|
|
- struct perf_event_context *ctx = tsk->perf_event_ctxp;
|
|
|
unsigned int *tsk_pinned;
|
|
|
- struct list_head *list;
|
|
|
- unsigned long flags;
|
|
|
-
|
|
|
- if (WARN_ONCE(!ctx, "No perf context for this task"))
|
|
|
- return;
|
|
|
-
|
|
|
- list = &ctx->event_list;
|
|
|
-
|
|
|
- spin_lock_irqsave(&ctx->lock, flags);
|
|
|
-
|
|
|
- /*
|
|
|
- * The current breakpoint counter is not included in the list
|
|
|
- * at the open() callback time
|
|
|
- */
|
|
|
- list_for_each_entry(bp, list, event_entry) {
|
|
|
- if (bp->attr.type == PERF_TYPE_BREAKPOINT)
|
|
|
- count++;
|
|
|
- }
|
|
|
-
|
|
|
- spin_unlock_irqrestore(&ctx->lock, flags);
|
|
|
+ int count = 0;
|
|
|
|
|
|
- if (WARN_ONCE(count < 0, "No breakpoint counter found in the counter list"))
|
|
|
- return;
|
|
|
+ count = task_bp_pinned(tsk);
|
|
|
|
|
|
tsk_pinned = per_cpu(task_bp_pinned, cpu);
|
|
|
if (enable) {
|
|
@@ -233,7 +249,7 @@ int reserve_bp_slot(struct perf_event *bp)
|
|
|
|
|
|
mutex_lock(&nr_bp_mutex);
|
|
|
|
|
|
- fetch_bp_busy_slots(&slots, bp->cpu);
|
|
|
+ fetch_bp_busy_slots(&slots, bp);
|
|
|
|
|
|
/* Flexible counters need to keep at least one slot */
|
|
|
if (slots.pinned + (!!slots.flexible) == HBP_NUM) {
|